text
stringlengths 6
128k
|
---|
# Entity-aware Transformers for Entity Search
Emma J. Gerritse Radboud University<EMAIL_ADDRESS>, Faegheh Hasibi
Radboud University<EMAIL_ADDRESS>and Arjen P. de Vries Radboud
University<EMAIL_ADDRESS>
(2022)
###### Abstract.
Pre-trained language models such as BERT have been a key ingredient to achieve
state-of-the-art results on a variety of tasks in natural language processing
and, more recently, also in information retrieval. Recent research even claims
that BERT is able to capture factual knowledge about entity relations and
properties, the information that is commonly obtained from knowledge graphs.
This paper investigates the following question: Do BERT-based entity retrieval
models benefit from additional entity information stored in knowledge graphs?
To address this research question, we map entity embeddings into the same
input space as a pre-trained BERT model and inject these entity embeddings
into the BERT model. This entity-enriched language model is then employed on
the entity retrieval task. We show that the entity-enriched BERT model
improves effectiveness on entity-oriented queries over a regular BERT model,
establishing a new state-of-the-art result for the entity retrieval task, with
substantial improvements for complex natural language queries and queries
requesting a list of entities with a certain property. Additionally, we show
that the entity information provided by our entity-enriched model particularly
helps queries related to less popular entities. Last, we observe empirically
that the entity-enriched BERT models enable fine-tuning on limited training
data, which otherwise would not be feasible due to the known instabilities of
BERT in few-sample fine-tuning, thereby contributing to data-efficient
training of BERT for entity search.
Entity retrieval, transformers, BERT, entity embeddings
††journalyear: 2022††copyright: rightsretained††conference: Proceedings of the
45th International ACM SIGIR Conference on Research and Development in
Information Retrieval; July 11–15, 2022; Madrid, Spain††booktitle: Proceedings
of the 45th International ACM SIGIR Conference on Research and Development in
Information Retrieval (SIGIR ’22), July 11–15, 2022, Madrid, Spain††doi:
10.1145/3477495.3531971††isbn: 978-1-4503-8732-3/22/07††ccs: Information
systems Language models
## 1\. introduction
Pre-trained language models (LMs) such as BERT (Devlin et al., 2019) and its
successors (Raffel et al., 2020; Yang et al., 2019; Liu et al., 2019) learn
rich contextual information about words from large-scale unstructured corpora
and have achieved intriguing results on a variety of downstream tasks in
natural language processing (NLP) (Devlin et al., 2019; Joshi et al., 2019;
Joshi et al., 2020) and information retrieval (IR) (Nogueira and Cho, 2019;
Wang et al., 2019b). It is even shown that these language models have the
capability of capturing a tremendous amount of world knowledge, including
information about real-world entities otherwise found in knowledge graphs
(KGs) (Petroni et al., 2019; Wang et al., 2020). For example, language models
can predict masked objects in cloze sentences such as “The native language of
Mammootty is _____” and “The Sharon Cuneta Show was created in _____”, where
each of them demonstrates a $\langle$subject, relation, object$\rangle$ triple
in a knowledge graph.
Language models, however, fail to perform complex reasoning about entities, as
they cannot capture sparse world facts (Talmor et al., 2020; Petroni et al.,
2019; Jiang et al., 2020). In information retrieval systems, users ask complex
queries about entities, such as “What is the second highest mountain in the
world?” or “Most famous civic-military airports.” Answering such queries
requires leveraging rich human curated information about entities in the
knowledge graphs. To bridge the gap between LMs and KGs, recent works enhance
language models with rich structured entity information from knowledge graphs,
showing superior performance for knowledge-driven NLP tasks such as relation
classification, entity typing, and cloze-style question answering (Zhang et
al., 2019; Yamada et al., 2020b; Poerner et al., 2020; Jiang et al., 2020;
Wang et al., 2019a; Wang et al., 2020; Peters et al., 2019). Despite this
success, no previous study has examined the effect of entity-enriched language
models for answering entity-oriented queries in IR.
This work explores the benefit of enriching BERT-based retrieval models with
auxiliary entity information from KGs for the _entity retrieval_ task, where
users’ queries are better answered with entities rather than passages or
documents (Balog, 2018). It is shown that entity information improves
performance of document and entity retrieval tasks (Xiong et al., 2017b, a;
Dalton et al., 2014; Hasibi et al., 2016; Garigliotti et al., 2019). Yet,
these studies were performed using traditional retrieval models, without
utilizing LMs such as BERT.
The development of BERT-based models for entity retrieval faces a major
challenge: there is limited training data for the entity retrieval task, and
fine-tuning of BERT in data-constraint regimes leads to instabilities in model
performance (Dodge et al., 2020; Phang et al., 2019; Zhang et al., 2021). More
specifically, the official entity retrieval dataset, DBpedia-Entity v2 (Hasibi
et al., 2017c), contains 467 queries; considered a small dataset for fine-
tuning of a large neural language model. When fine-tuning BERT (especially its
large variant) on a target task, the training data should be large enough that
the model parameters get close to the ideal setting of the target task, or
otherwise it causes forgetting of what the model has already learned.
Against this background, our first and primary research question is _RQ1: Can
an entity-enriched BERT-based retrieval model improve the performance of
entity retrieval?_ To address this question, we propose an entity-enriched
BERT-based retrieval model, EM-BERT, where factual entity information is
injected into the monoBERT model (Nogueira and Cho, 2019) in the form of
Wikipedia2Vec (Yamada et al., 2020a) entity embeddings that are transformed
into the BERT’s word piece vector space (Poerner et al., 2020); see Figure 1.
The EM-BERT model is first trained with the MS MARCO passage dataset and
further fine-tuned on the DBpedia-Entity v2 collection (in a cross validation
setup). Our experiments indicate that our EM-BERT model improves state-of-the-
art entity retrieval results by 11% with respect to NDCG@10. We also make the
intriguing finding that while fine-tuning of the plain BERT-based retrieval
model (monoBERT) on DBpedia-Entity v2 is brittle and prone to degenerate
performance, the same fine-tuning process using the entity-enriched BERT model
(EM-BERT) results in a stable model and brings consistent and meaningful
improvements. We posit that this is attributed to direct injection of entity
information into the BERT model, which brings the input distribution of the
model close to the output label space, thereby contributing to data-efficient
training of BERT-based models for entity-oriented tasks.
After observing the effectiveness of our EM-BERT model, we focus on
understanding the whens and whys of entity-enriched BERT models for
information retrieval tasks. Our next research question is: _RQ2: When and
which queries are helped by the EM-BERT model?_ We observe that EM-BERT mainly
helps queries that are annotated with at least one entity; plain and entity-
enriched BERT models perform on par with each other for queries without linked
entities. We further examine queries with linked entities and find that EM-
BERT is most helpful when entity mentions are broken into multiple word pieces
by the BERT tokenizer (e.g. mention “Tuvalu” in Figure 1). This indicates that
directly injecting entity information into the BERT-based models is
particularly important for less popular entities, which BERT does not
recognize as a single word (or phrase) because they are less observed during
pre-training.
In our third question, we investigate _RQ3: Why does the EM-BERT model work
and what does it learn during the fine-tuning stage?_ We approach this by
comparing the embeddings produced at the final layer of the EM-BERT network
for both entities and their mention tokens. We find that the learned entity
representations, unlike their corresponding word piece representations, form
clear clusters in the embedding space, capturing the relation between
entities. We also study some examples, and show that attention weights of
entity tokens are larger than other tokens for queries requesting lists of
entities or queries related to less known entities.
Finally, we question _RQ4: How does our entity-enriched BERT-based model
perform on other ranking tasks?_ To address this research question, we apply
the EM-BERT model to the passage ranking task, where entity information is
known to be less important and a large amount of training data can be used. We
observe that the entity information that is embedded in the plain BERT-based
retrieval model is enough for addressing this task and auxiliary entity
information from knowledge graphs does not bring significant improvements.
In summary, this work makes the following contributions:
* •
We study the effect of enriching BERT with entity information for information
retrieval tasks and propose an entity-enhanced BERT-based re-ranker EM-BERT.
* •
We establish new state-of-the-art results for the entity retrieval task on
DBpedia-Entity v2 collection.
* •
We show that our entity-enriched model, unlike its equivalent model based on
plain BERT, is robust against instabilities of BERT in data-constraint regimes
and introduce EM-BERT as a data-efficient BERT-based model for entity
retrieval.
* •
We perform thorough analysis of our EM-BERT model and add to our understanding
of the model. We unfold when and for which queries this model work, what it
learns, and how it works.
The resources developed within the course of this paper are available at
https://github.com/informagi/EMBERT.
## 2\. Related work
Figure 1. Illustration of the EM-BERT model. Entity annotated query and
documents are tokenized and mapped to their corresponding vector
representations using $F_{BERT}$ and $F_{E-BERT}$ functions.
[]
#### Graph Embeddings
After the introduction of word embeddings methods like Word2Vec (Mikolov et
al., 2013), neural embedding methods became increasingly popular. Word2Vec
uses a single-layer neural network to either predict a word depending on its
context (Continuous Bag of Word) or the context depending on a word (Skip-
Gram). The resulting hidden layer of this network captures a representation of
words and can be used for other downstream tasks. After the success of these,
embedding methods were expanded to capture other types of knowledge as well,
for example knowledge graphs. These so-called graph embeddings aim to embed
the nodes of a knowledge graph in a way that entities and their relations are
mapped accordingly into the vector space. One of the earliest introduced
methods was TransE (Bordes et al., 2013), which optimizes embeddings for head-
relation-tail triples.
The graph embedding method utilized in this paper is Wikipedia2Vec (Yamada et
al., 2020a). This method is similar to Word2Vec, but is extended to be used on
text-heavy knowledge bases like Wikipedia. Instead of solely relying on words
or edges between entities, it uses a combination of words, entities, and
anchor texts. Using either Skip-Gram or Continuous Bag of Words method like in
Word2Vec, it employs a single layer neural network to predict neighboring
words to capture word similarities, entities relations (based on neighboring
entities), and word-entity relations (based on surrounding anchor-text).
Through this combination, words and entities get embedded into the same vector
space.
#### Entity Retrieval
Graph embeddings have been used to assist in _entity-oriented information
retrieval_. Gerritse et al. (2020) uses Wikipedia2Vec embeddings for entity
retrieval on the DBpedia-Entity v2 collection (Hasibi et al., 2017c). This is
done by first tagging entities in the queries by using the TAGME entity linker
(Ferragina and Scaiella, 2010). Retrieved entities are then re-ranked by
computing the cosine similarity score between the retrieved entities and the
tagged entities, improving upon the state of the art for entity retrieval.
KEWER (Nikolaev and Kotov, 2020) is another method for entity retrieval based
on graph embeddings. It learns joint embeddings for both words and entities,
combining this with entities annotated queries using SMAPH (Cornolti et al.,
2016). KEWER also improved previous state of the art for entity retrieval. The
most recent graph-based entity ranking method, ENT Rank (Dietz, 2019),
utilizes entity-neighbor-text relations in a learning-to-rank model. ENT Rank
incorporates a variety of entity features based on text, co-occurring
entities, and entity neighbor relations, and performs competitively on entity
ranking tasks. The paper, however, uses a different version of the DBpedia-
Entity v2 collection and calculates the evaluation metrics only over judged
entities, making it incomparable to our and the aforementioned entity
retrieval methods.
The BEIR (Thakur et al., 2021) benchmark provides a number of zero-shot
retrieval models for several IR databases, including the DBpedia-Entity v2
entity retrieval collection. The best performing entity retrieval model in
BIER, BM25 + CE, re-ranks the top 100 entities retrieved by BM25, and then
uses the MiniLM cross-encoder model (Wang et al., 2020), trained with the
knowledge distillation setup provided by Hofstätter et al. (2020).
#### Entity Linking
Utilizing entity information for retrieval models requires identifying entity
information from documents and queries (Hasibi et al., 2017a; Blanco et al.,
2015; Hasibi et al., 2015). This is performed by annotating text with entity
linking toolkits (Ferragina and Scaiella, 2010; Cornolti et al., 2016; van
Hulst et al., 2020; Hasibi et al., 2017b; Cao et al., 2021). In this paper, we
use the REL entity linker (van Hulst et al., 2020), which is the state-of-the-
art open source entity linking toolkit. REL detects mentions using Flair
(Akbik et al., 2018) embeddings. REL performs candidate selection based on
Wikipedia2Vec embeddings, and entity disambiguation based on latent relations
between entity mentions in the text.
#### Transformer-based Rankers
Bidirectional Encoder Representations from Transformers (BERT) (Devlin et al.,
2019) is a language representation model, pre-trained to capture bi-
directional relations between words. The pre-trained BERT model can be fine-
tuned for various NLP tasks. BERT can be used for binary relevance
classification by feeding it a sentence A and sentence B, and using the BERT
classifier token (further denoted as [CLS]) to predict relevance. This setup
has been shown to be effective for question answering.
BERT-based models are effective for _information retrieval_ as well, and due
to their relatively low number of input tokens became especially popular for
_passage retrieval_. Several BERT-based methods have made enormous
improvements on the previous state of the art (Wang et al., 2019b; Nogueira
and Cho, 2019). MonoBERT (Nogueira and Cho, 2019) is a point-wise re-ranker,
in which the BERT model is used as a binary relevance classifier. MonoBERT, at
the time of release, obtained state-of-the-art results on the MS MARCO and
TREC CAR datasets. After pre-training, the ranking is done by first making an
initial ranking with a baseline method like BM25, and then re-ranking a query
and passage by feeding a pair of sentences A (query) and B (passage) to BERT.
#### Entities and Transformers
During pre-training, BERT does not explicitly get structured data about
entities. Some researchers have claimed that adding this may not be necessary
since BERT captures this information implicitly (Wang et al., 2020; Petroni et
al., 2019). Others set out to enrich BERT with entity information. One of the
first works combining entities with transformers, ERNIE (Zhang et al., 2019),
is an enhanced language representation model, where the combination of
entities and text is used to fine-tune the model. In KnowBert (Peters et al.,
2019), contextual word embeddings are also enhanced by integrating information
about entities. This is done by introducing a Knowledge Attention and
Recontextualization component, which takes information in one layer of a BERT
network, computes the enhanced entity information of that layer, and feeds it
to the next layer. This method improves on BERT for entity-related tasks like
entity linking.
E-BERT (Poerner et al., 2020) is another method to enhance BERT with entity
information. The main point which distinguishes E-BERT from the previously
mentioned work is that it does not require any additional training of the
network; it only requires the computation of one large transformation matrix.
This allows the method to be applied to any fine-tuned transformer model.
Since Wikipedia2Vec embeds both words and entities in the same vector space,
the embeddings of words in Wikipedia2Vec can be used to align a mapping.
E-BERT has been shown to improve performance on unsupervised QA tasks like
LAMA, as well as downstream tasks like relation classification.
Other works combining language models and knowledge graphs include (Broscheit,
2019), which enhances BERT with entities by using BERT itself for entity
linking and seeing how much information of knowledge graphs is already
contained in BERT. Next, BERT classifies texts to predict whether certain
tokens belong to an entity. The authors show that specifically fine-tuning
BERT on entity information greatly improves their scores on entity linking,
thus showing that pre-trained BERT does not embed all information about
entities yet, claiming that additional entity information does not help if
either entities are too scarce in the data or if the task does not require
entity knowledge.
## 3\. Method
In this section, we first provide a brief overview of knowledge graph
embeddings and Wikipedia2Vec (Section 3.1), and then describe the EM-BERT
model, which is an entity-enhanced version of a BERT-based retrieval model
(Section 3.2), followed by a description of EM-BERT query-document input
representation.
### 3.1. Background
Knowledge graph embeddings provide vector representations of entities in a
knowledge graph, projecting entity properties and relations into a continuous
vector space. These embeddings have proven effective not only for entity-
related tasks such as entity linking (Yamada et al., 2016), but also for
general tasks such as question answering (Nikolaev and Kotov, 2020). Knowledge
graph embeddings can be constructed purely based on graph topology (i.e.,
entities and their relations) (Bordes et al., 2013), or combining graph
topology with additional textual information such as the entity descriptions
(Yamada et al., 2016). In this paper, we use Wikipedia2Vec embeddings (Yamada
et al., 2020a), which extends the skip-gram model (Mikolov et al., 2013) with
entity relations and descriptions from Wikipedia and maps words and entities
into a shared vector space. Following (Poerner et al., 2020), we transform
Wikipedia2Vec entity embeddings into the BERT word piece embedding space and
use them as input to the BERT model, similar to a BERT word piece vector.
Table 1. Statistics about document length and number of linked entities in queries and documents of the used datasets, with mean and standard deviation. Document and query length information is computed without adding entity information. Dataset | size | avg length | avg entities | total linked
---|---|---|---|---
MS MARCO dev queries | 6980 | $6.95\pm 2.54$ | $0.34\pm 0.57$ | 2392
MS MARCO passages | 8841823 | $58.27\pm 22.94$ | $2.55\pm 3.22$ | 22532200
DBpedia-Entity v2 queries | 467 | $6.55\pm 2.88$ | $0.82\pm 0.64$ | 381
DBpedia-Entity v2 short abstracts | 389622 | $55.93\pm 25.95$ | $6.88\pm 3.93$ | 2678843
Formally, Wikipedia2Vec embeds a list of words $\mathbb{L}_{word}$ and a list
of entities $\mathbb{L}_{ent}$ into a vector space
$\mathbb{R}^{d_{\mathit{Wikipedia}}}$, where $d_{\mathit{Wikipedia}}$ is
dimensions of the Wikipedia embeddings. The embeddings are trained by fitting
a one layer neural network with the loss:
$\mathcal{L}=\mathcal{L}_{w}+\mathcal{L}_{e}+\mathcal{L}_{a},$
which consist of three cross-entropy loss functions:
* $\mathcal{L}_{w}$:
predicting context words of a given word.
* $\mathcal{L}_{e}$:
predicting neighboring entities of a given entity.
* $\mathcal{L}_{a}$:
predicting anchor text of an entity to combine words and entities in a single
embedding space.
The weight layer of this neural network results in the function
$F_{\mathit{Wikipedia}}\
:\mathbb{L}_{word}\cup\mathbb{L}_{ent}\rightarrow\mathbb{R}^{d_{\mathit{Wikipedia}}}$,
which embeds both words and entities in the Wikipedia vector space. BERT, on
the other hand, tokenizes text into the word piece dictionary
$\mathbb{L}_{wordpiece}$. We use the lookup function
$F_{BERT}:\mathbb{L}_{wordpiece}\rightarrow\mathbb{R}^{d_{BERT}}$ to transform
word pieces into the $d_{BERT}$-dimensional native BERT word piece vector
space, the resulting vector being the input for the BERT model.
### 3.2. EM-BERT
Our proposed model, referred to as EM-BERT, incorporates entity embeddings
into a point-wise document ranking approach. In its essence, EM-BERT combines
E-BERT (Poerner et al., 2020) with monoBERT (Nogueira and Cho, 2019) and
predicts a retrieval score for a query-document pair, each represented by a
sequence of words and entities. The model takes as input the concatenation of
the query tokens $t_{q_{1}},...,t_{q_{n}}$ and the document tokens
$t_{d_{1}},...,t_{d_{m}}$, where each token $t_{i}$ is either a BERT native
word piece token or an entity token.
#### Entity-enriched BERT
The off-the-shelf BERT model only accepts its native word pieces. To
incorporate entity embeddings into BERT, we need to align entity vectors with
word piece vectors. Following (Poerner et al., 2020), this is done by a linear
transformation of entity vectors to BERT-like vectors. Since
$\mathbb{L}_{wordpiece}$ does not contain any entities, the Wikipedia2Vec word
dictionary $\mathbb{L}_{Word}$ is used to obtain the linear transformation
$\mathbf{W}\in\mathbb{R}^{d_{BERT}\times d_{\mathit{Wikipedia}}}$, learned
from the Wikipedia2Vec word vector space
$F_{\mathit{Wikipedia}}[\mathbb{L}_{Word}]$ and BERT word piece space
$F_{BERT}[\mathbb{L}_{wordpiece}]$:
(1)
$\mathbf{W}=argmin_{W}\sum_{\mathclap{x\in\mathbb{L}_{word}\cap\mathbb{L}_{wordpiece}}}||W\cdot
F_{\mathit{Wikipedia}}(x)-F_{BERT}(x)||^{2}_{2}.$
Here, the intersection between $\mathbb{L}_{word}$ and
$\mathbb{L}_{wordpiece}$ is taken to ensure that the selected words have
embeddings in both Wikipedia and BERT vector spaces. The equation computes
matrix $\mathbf{W}$, in which the distance between $\mathbf{W}\cdot
F_{\mathit{Wikipedia}}(x)$ and $F_{BERT}(x)$ is minimal for all $x$. Using
$\mathbf{W}$, we can then construct the function $F_{E-BERT}$, which maps both
entities and word tokens to the BERT input vector space:
(2) $F_{E-BERT}(a)=\begin{cases}\mathbf{W}\cdot
F_{\mathit{Wikipedia}}(a),&a\in\mathbb{L}_{ent}\\\
F_{\mathit{wordpiece}}(a),&\text{otherwise}.\end{cases}$
The aligned entity vectors are then fed into BERT when an entity is mentioned
in the input text. The entities are obtained by annotating the text with an
entity; e.g., given the text “Who produced films starring Natalie Portman”,
the annotated sentence becomes “Who produced films starring Natalie Portman
ENTITY/NataliePortman”. This text is then tokenized as “who produced films
starring natalie port ##man /ENTITY/Natalie_Portman,” where the entity ID (in
bold face) is a single token, embedded by $F_{E-BERT}$ (Eq. 2). We note that
the transformation matrix $\mathbf{W}$ can be fit into any pre-trained or
fine-tuned BERT model, thus making E-BERT embeddings usable for every
available BERT model.
#### Retrieval
The retrieval phase is based on monoBERT (Nogueira et al., 2019), which is a
multi-stage ranking method with BERT and has been shown to achieve competitive
results on MS MARCO passage retrieval (Nguyen et al., 2016) and TREC-CAsT
(Dalton et al., 2020). The simplicity of this model, coupled with its high
effectiveness enables us to discern the effect of entity information on BERT-
based retrieval models and obtain an understanding of the whys and whens of
entity-enriched BERT models for information retrieval tasks.
In the multi-stage ranking method, for a query $q$, documents are ranked with
an initial ranking method; e.g., BM25. The top-$k$ documents, denoted as
$D_{K}=\\{d_{1},d_{2},\dots,d_{k}\\}$, are then passed to a second ranker for
re-ranking; here a BERT-based retrieval model. Every query-document pair
$\langle q,d_{i}\in D_{K}\rangle$ is passed to BERT as two sentences $A$ and
$B$, with the separation token [SEP] in between. The BERT [CLS] classification
vector is used as input to a single layer neural network to obtain the
probability $s_{i}$ of the document $d_{i}$ being relevant to the query $q$.
To train this model, the cross-entropy loss is used:
(3) $L=-\sum\limits_{j\in J_{pos}}log(s_{j})-\sum\limits_{j\in
J_{neg}}log(1-s_{j}),$
where $J_{pos}$ and $J_{neg}$ are indexes of relevant and non-relevant
documents for all queries.
Putting all the pieces together, the training process of the EM-BERT model is
summarized as follows. First, the transformation matrix $\mathbf{W}$ is
trained using Wikipedia2Vec embeddings and BERT word pieces. All queries and
documents are annotated with an entity linker, and the tagged queries and
documents are tokenized and mapped to the corresponding vector representations
using function $F_{E-BERT}$ . This input is fed into the EM-BERT model and the
model is first fine-tuned on MS MARCO passages and then on the entity
retrieval collection.
#### Query-Document Representation
In the EM-BERT model (similar to monoBERT), queries are truncated to have a
maximum token length of 64, and the combination of the query, passage, and
separation tokens are truncated to have a maximum length of 512 tokens.
Following (Poerner et al., 2020), we concatenate mentions and entities,
separated by the token ‘/’; i.e., “mention + ‘/’ + EntityID.” The final
representation of query-document pairs is “[CLS] + Query + [SEP] + Document +
[SEP];” see Figure 1.
## 4\. Experimental setup
Table 2. Results on DBpedia-Entity v2 collection. Superscripts 1/2/3/4 denote statistically significant differences (better or worse) compared to base method/monoBERT (1st)/monoBERT/EM-BERT (1st), respectively. The highest value per column is marked with underline and in bold. The highest value per column and per block is marked with underline. NDCG | SemSearch | INEX-LD | ListSearch | QALD-2 | Total
---|---|---|---|---|---
@10 | @100 | @10 | @100 | @10 | @100 | @10 | @100 | @10 | @100
ESIMcg (Gerritse et al., 2020) | 0.417 | 0.478 | 0.217 | 0.286 | 0.211 | 0.302 | 0.212 | 0.282 | 0.262 | 0.335
KEWER (Nikolaev and Kotov, 2020) | - | - | - | - | - | - | - | - | 0.270 | 0.310
BLP-TransE (Daza et al., 2021) | 0.631 | 0.723 | 0.446 | 0.546 | 0.442 | 0.540 | 0.401 | 0.482 | 0.472 | 0.562
BM25F+KEWER (Nikolaev and Kotov, 2020) | 0.661 | 0.733 | 0.468 | 0.530 | 0.440 | 0.521 | 0.386 | 0.474 | 0.483 | 0.560
BM25 | 0.425 | 0.523 | 0.298 | 0.330 | 0.274 | 0.322 | 0.192 | 0.243 | 0.291 | 0.349
+monoBERT (1st) | $0.588^{1}$ | $0.655^{1}$ | $0.406^{1}$ | $\underline{0.459}^{1}$ | $0.422^{1}$ | $0.458^{1}$ | $0.350^{1}$ | $0.406^{1}$ | $0.437^{1}$ | $0.490^{1}$
+monoBERT | $0.577^{1}$ | $0.640^{12}$ | $\underline{0.409}^{1}$ | $0.452^{1}$ | $0.423^{1}$ | $0.449^{12}$ | $0.347^{1}$ | $0.392^{12}$ | $0.435^{1}$ | $0.479^{12}$
+EM-BERT (1st) | $0.593^{1}$ | $0.667^{13}$ | $0.395^{1}$ | $0.448^{1}$ | $0.436^{1}$ | $0.465^{1}$ | $0.351^{1}$ | $0.403^{1}$ | $0.440^{1}$ | $0.492^{13}$
+EM-BERT | $\underline{0.612}^{1}$ | $\underline{0.672}^{1}$ | $0.392^{1}$ | $0.434^{12}$ | $\underline{0.478}^{1234}$ | $\underline{0.469}^{13}$ | $\underline{0.375}^{1234}$ | $\underline{0.418}^{134}$ | $\underline{0.461}^{1234}$ | $\underline{0.495}^{13}$
BM25F-CA (Hasibi et al., 2017c) | $\underline{0.628}$ | $\underline{0.720}$ | $\underline{0.439}$ | $\underline{0.530}$ | $0.425$ | $0.511$ | $0.369$ | $0.461$ | $0.461$ | $0.551$
+monoBERT (1st) | $0.605$ | $0.682^{1}$ | $0.427$ | $0.501$ | $0.440$ | $0.510$ | $0.405$ | $0.490$ | $0.467$ | $0.544$
+monoBERT | $0.591^{2}$ | $0.663^{12}$ | $0.434$ | $0.489^{12}$ | $0.438$ | $0.496^{2}$ | $0.401$ | $0.467^{2}$ | $0.463$ | $0.526^{12}$
+EM-BERT (1st) | $0.609$ | $0.695^{3}$ | $0.418$ | $0.490^{1}$ | $0.452$ | $0.518^{3}$ | $0.398$ | $0.483$ | $0.466$ | $0.544^{3}$
+EM-BERT | $0.621$ | $0.695$ | $0.399$ | $0.469^{12}$ | $\underline{0.492}^{1234}$ | $\underline{0.535}^{23}$ | $\underline{0.435}^{1234}$ | $\underline{0.512}^{1234}$ | $\underline{0.486}^{1234}$ | $\underline{0.553}^{3}$
GEEER (Gerritse et al., 2020) | $0.660$ | $0.736$ | $0.466$ | $0.552$ | $0.452$ | $0.535$ | $0.390$ | $0.483$ | $0.487$ | $0.572$
+monoBERT (1st) | $0.643$ | $0.733$ | $\mathbf{\underline{0.486}}$ | $\mathbf{\underline{0.564}}$ | $0.493^{1}$ | $0.554^{1}$ | $0.449^{1}$ | $0.525^{1}$ | $0.515^{1}$ | $0.591^{1}$
+monoBERT | $0.633^{2}$ | $0.725^{2}$ | $0.481$ | $0.563$ | $0.492^{1}$ | $0.553^{1}$ | $0.439^{12}$ | $0.519^{12}$ | $0.508^{12}$ | $0.586^{12}$
+EM-BERT (1st) | ${{0.650}}$ | $0.743$ | $0.483$ | $0.561$ | $0.504^{1}$ | $0.562^{1}$ | $0.444^{1}$ | $0.522^{1}$ | $0.517^{13}$ | $0.594^{13}$
+EM-BERT | $\mathbf{\underline{{0.664}}}$ | $\mathbf{\underline{0.744}}$ | $0.479$ | $0.561$ | $\mathbf{\underline{0.544}}^{1234}$ | $\mathbf{\underline{0.579}}^{1234}$ | $\mathbf{\underline{0.483}}^{1234}$ | $\mathbf{\underline{0.543}}^{1234}$ | $\mathbf{\underline{0.541}}^{1234}$ | $\mathbf{\underline{0.604}}^{1234}$
To address our research question, we compare the EM-BERT model with several
baselines on the entity retrieval task and analyze the results. Due to limited
training data for entity retrieval, we perform two-stage fine-tuning: we first
fine-tune our models on the annotated MS MARCO passage dataset, and then
continue fine-tuning on DBpedia-Entity v2 collection. In the following, we
describe our experimental setup.
### 4.1. Entity Linking and Embeddings
We use the REL entity linking toolkit (van Hulst et al., 2020) to annotate
texts with entities. REL is known to have high precision compared to other
entity linking toolkits such as GENRE (Cao et al., 2021) and TAGME (Ferragina
and Scaiella, 2010), ensuring that limited noise (incorrectly annotated
entities) is added to the model. REL incorporates Wikipedia2Vec embeddings
trained on the Wikipedia link graph. To avoid a mismatch between annotated
entities and Wikipedia2Vec entity embeddings (Gerritse et al., 2020; Dietz,
2019), we follow the REL Wikipedia2Vec training procedure, using the Wikipedia
dump of 2019-07-01, setting min-entity-count parameter to zero, with a
dimensionality of 500.
### 4.2. First Stage Fine-tuning
#### Collection
For the first stage fine-tuning of EM-BERT, we use the MS MARCO passage
ranking collection, consisting of 8,841,823 passages, extracted from web
documents retrieved by Bing. MS MARCO provides a training set with
approximately 800k queries, a development set with 6980 queries with public
relevance judgments, and an evaluation set with 6837 queries and private
relevance judgments.
Although MS MARCO passage ranking is not an entity-related task, REL annotates
29% of queries in the development sets with at least one entity; e.g., “what
were the _brothers grimm_ names.” The statistics of this collection (with REL
annotations) are reported in Table 1. It shows that on average every query and
document is annotated with 0.34 and 2.55 entities respectively, indicating
that a fair amount of entities is present in this dataset.
#### Training
We fine-tune our EM-BERT model on the officially provided MS MARCO training
set. This set provides triples of a query, a positive example, and a negative
example, making the ratio of positive-negative documents equal during
training. We refer to these triples as “sample.” Starting from the monoBERT
model provided by HuggingFace,111https://huggingface.co/castorini/monobert-
large-msmarco we fine-tune EM-BERT with randomly selected 300k samples (cf.
§3.2); fine-tuning with more samples, up until 900k, did not give any further
improvements on the development set. We use an accumulated batch size of 64, a
learning rate of $10^{-6}$, 40k warm-up steps, and AdamW as optimizer. Fine-
tuning for 300k samples took around 34 hours on a single GeForce RTX 3090 with
24 GB memory.
### 4.3. Second Stage Fine-tuning
#### Collection
For the second stage fine-tuning, we use the standard entity retrieval
collection, DBpedia-Entity v2 (Hasibi et al., 2017c). The collection is based
on DBpedia 2015-10 and consists of four different categories of queries: (i)
_SemSearch_ : named entity queries looking for particular entities; e.g.,
“mario bros”, (ii) _INEX-LD_ : keyword-style entity-oriented queries; e.g.,
“bicycle benefits health”, (iii) _ListSearch_ : queries asking for a list of
entities; e.g., “Hybrid cars sold in Europe”, and (iv) _QALD-2_ : natural
language questions about entities; e.g., “Who produces Orangina?” Relevance
judgements in this dataset are graded: Highly relevant (2), Relevant (1), and
irrelevant (0).
Documents in the DBpedia-Entity v2 collection are entities, each represented
by an abstract and other metadata (stored in RDF format). These entity
abstracts provide concise descriptions about entities and are comparable in
length to MS MARCO passages. Both queries and entity abstracts are highly
entity-oriented, reflected by the average number of entity annotations for
queries and abstracts in Table 1.
#### Training
We consider entity abstracts as entity descriptions and use them as documents
in monoBERT and EM-BERT. Since the DBpedia-Entity v2 collection provides a
limited amount of training data (only 467 queries), we take the fine-tuned
model on the MS MARCO passage dataset (Section 4.2) and continue fine-tuning
for both monoBERT and EM-BERT models using the provided folds for 5-fold cross
validation, each fold containing 48k query-entity pairs with approximately 34%
relevant and 66% non-relevant entities. In this process, both models are fine-
tuned for one epoch using all training data of each fold, resulting in 5
different models each. The fine-tuned models are then used for ranking the
test queries in the corresponding folds. Note that in this fine-tuning
process, no hyper-parameter optimization is performed. In this process, both
models are fine-tuned for one epoch using the training data of each fold,
resulting in 5 different models each. The fine-tuned models are used for
ranking the test queries in the corresponding folds. Note that in this fine-
tuning process, no hyper-parameter optimization is performed. For both models
an aggregated batch size of 64, learning-rate of $10^{-6}$, and AdamW
optimizer is used. Warmup step is changed to 4000, following the 10% warmup
step rule.
### 4.4. Baselines
We compare EM-BERT with the state-of-the-art entity retrieval methods using
term-matching, neural, and BERT-based approaches. The baselines are:
Figure 2. Comparison of re-ranked BM25F-CA run with monoBERT and EM-BERT for
different mention tokenization categories.
[]
Figure 3. Mention tokenization categories for the for types of queries in
DBPedia-Entity v2 collection.
[]
* BM25: The BM25 run, based on the short abstracts of entities, with parameters $k=0.9$ and $b=0.4$. Similar to monoBERT and EM-BERT, this run uses only abstract information of entities (unlike other methods that use reach representation of entities).
* BM25F-CA (Hasibi et al., 2017c): The fielded variant of BM25, obtaining scores from 5-field representation of entities. Field weights are computed using Coordinate Ascent on each train fold, and then scored on the corresponding test fold. This is the best non-neural run and also the best baseline reported in (Hasibi et al., 2017c).
* ESIMcg (Gerritse et al., 2020): ESIMcg ranks entities based on the aggregated cosine similarity scores between the Wikipedia2Vec embedding of each linked entity in the query and the target entity. Entities in this method are linked using TAGME (Ferragina and Scaiella, 2010). The method re-ranks the BM25F-CA run.
* GEEER (Gerritse et al., 2020): GEEER linearly combines the BM25F-CA score and entity-query similarities based on ESIMcg. The weights of this linear combination are computed on each train fold using Coordinate Ascent, and scored accordingly on each test fold. Similar to ESIMcg, GEEER re-ranks the BM25F-CA run.
* KEWER (Nikolaev and Kotov, 2020): KEWER uses an embedding method with joint word and entity embeddings. A query-entity score is computed on all query terms by the weighted sum of the similarity between the target entity embedding and the term, re-ranking the BM25F-CA run.
* BM25F+KEWER (Nikolaev and Kotov, 2020): The model linearly combines the KEWER score as listed above with BM25F-CA and re-ranks the BM25F-CA run.
* BLP-TransE (Daza et al., 2021): BLP-TransE is a BERT-based method, combining a BERT cross-encoder architecture with TransE graph embeddings. It encodes queries and entities, uses the similarity between these encodings as query score, and then re-ranks and linearly combines the BM25F-CA scores.
* monoBERT: Our Pygaggle implementation of the monoBERT model. This implementation achieves slightly better performance compared to results reported in (Nogueira et al., 2019): MRR@10 of 0.379 vs 0.372 on the MS MARCO development set. Following the multi-stage re-ranking method (Nogueira et al., 2019), we re-rank three entity retrieval runs using monoBERT: BM25, BM25F-CA, and GEEER (the best baseline run). The re-ranking is performed on the top-1000 entities of the BM25 and BM25F-CA, and top-100 entities of the GEEER run (as GEEER re-ranks the BM25F-CA run).
Table 3. Comparison of results for queries without and with at least one linked entity. Superscripts 1/2/3 show statistically significant differences compared to BM25F-CA/monoBERT (1st)/monoBERT, respectively. DBpedia-Entity v2 | NDCG@10 (¿1 en) | NDCG@10 (no-en)
---|---|---
BM25F-CA | 0.481 | 0.414
+monoBERT (1st) | 0.487 | 0.422
+monoBERT | 0.482 | 0.420
+EM-BERT | $0.516^{123}$ | 0.420
### 4.5. Evaluation Metrics
For evaluation of entity retrieval, we use the commonly reported metric in
(Hasibi et al., 2017c); the Normalized Discounted Cumulative Gain (NDCG) at
ranks 10 and 100. Statistical significant differences of NDCG@10 and NDCG@100
values are determined using the two-tailed paired t-test with p-value $<0.05$.
## 5\. Results and Analysis
In the following, we evaluate our entity-enriched BERT-based retrieval model
and answer our four research questions (Sections 5.1-5.4 ) listed in Section
1.
### 5.1. Entity Retrieval Results
In this section, we answer our first research question: _RQ1: Can an entity-
enriched BERT-based retrieval model improve the performance of entity
retrieval?_
We compare retrieval performance of EM-BERT with a variety of models on the
DBpedia-Entity v2 collection in Table 2. The baseline (cf. Section 4.4)
results are presented in the top part of the table. The next three
compartments in the Table summarise results of re-ranking BM25, BM25F-CA, and
GEEER runs with the following models:
* •
monoBERT (1st): The monoBERT model after first stage fine-tuning; cf. Section
4.4.
* •
monoBERT: The monoBERT model after second stage fine-tuning on DBpedia-Entity
v2.
* •
EM-BERT (1st): The EM-BERT model fine-tuned on the MS MARCO passage
collection. Re-ranking setup is similar to the monoBERT runs; cf. Section 4.4.
* •
EM-BERT: The EM-BERT model after second stage fine-tuning. Setup is similar to
the previous run.
As shown in Table 2, re-ranking of the GEEER run with EM-BERT markedly
outperforms all baselines and establishes a new state-of-the-art result on the
DBpedia-Entity v2 collection, with 11% improvements over the best run. When
re-ranking the BM25 and BM25F-CA runs, we observe the same trend that EM-BERT
outperforms all the corresponding BERT-based baselines. Comparing different
query categories, EM-BERT improves over all runs by a large margin for
ListSearch and most importantly for QALD-2 queries; improvements over GEEER
are 20% and 24% with respect to NDCG@10 for ListSearch and QALD-2 queries,
respectively. QALD queries are complex natural language queries that are hard
to answer for most neural and non-neural models, reflected by the lowest NDCG
scores on all baseline models.
Table 2 also shows that monoBERT performance is decreased when fine-tuned with
a limited amount of training data. The drop in the performance of monoBERT is
expected due to the known instabilities of BERT with few-sample training
(Zhang et al., 2021), which can cause forgetting of what the model has already
learned. The striking observation, however, is that EM-BERT results are
improved when trained on the same data and using the same procedure. This
indicates that entity-enriched BERT models enable fine-tuning on limited
training data, thereby contributing to data-efficient training of BERT for
entity-oriented tasks.
Based on these results, we can positively answer our first research question:
_our entity-enriched BERT-based model significantly improves state-of-the-art
results on entity retrieval, and more interestingly, it is robust against
instabilities of BERT when fine-tuned with limited training data._
(a) Embeddings of entity tokens.
(b) Embeddings of entity mention tokens.
Figure 4. UMAP plots of final layer EM-BERT embeddings of entity tokens and
their corresponding mention tokens. Queries were randomly selected from
queries with between 100 and 200 relevant entities.
### 5.2. Query Analysis
In the following we answer our research question _RQ2: When and which queries
are helped by the EM-BERT model?_
#### Entity annotations of queries
Knowing that not all queries in DBpedia-Entity v2 have a linked entity, we
expect EM-BERT to perform better on queries with linked entities than on
queries without them. To study this hypothesis, we compare NDCG@10 for two
subsets of DBpedia-Entity v2 queries: with and without at least one linked
entity. Table 3 shows the results. We see substantial improvements of EM-BERT
over monoBERT and monoBERT (1st) for queries with linked entities, confirming
our hypothesis.
#### Tokenization of entity mentions
To further our understanding of helped queries, we investigate the difference
in EM-BERT performance for various types of linked queries by looking into
their mentions tokenization. The BERT tokenizer splits words into word pieces,
which can be categorized into _beginning_ word pieces, indicating either the
start of a word or a full word, and _middle/end_ word pieces starting with a
‘##’. BERT employs around 30k tokens, consisting of around 24k beginning
tokens and 6k middle/end tokens. Given that Wikipedia has around 12 million
distinct entities, we expect only the most common entities to be present as a
single word piece. We hypothesize that there is a performance difference
between these very common entities, which are often seen in their natural form
during the BERT pre-training process, and the rarer entities, where their
mentions are broken into several word pieces. We, therefore, divide entity
mentions into four categories:
* •
One token: Mentions that are considered as a single token by the BERT
tokenizer; e.g., “France” is tokenized to ‘France’.
* •
Multiple tokens, no ##: Mentions that are tokenized to more than one token,
where each token represents a single unbroken word; e.g., “Yoko Ono” is
tokenized to ‘Yoko’, ‘Ono’.
* •
One ##: Mentions that are split into multiple word pieces, of which exactly
one is a middle/end word piece, e.g., “Weser” is be tokenized to ‘wes’, ‘##er’
* •
Multiple ##: Mentions which gets split into multiple word pieces, of which
more than one is a middle/end word piece, for example “Frisian” is tokenized
to ‘fr’, ‘##isi’, ‘##an’.
We categorize each query with a linked entity in one of the previously
mentioned categories, prioritizing the most complicated entity; i.e., if a
query has two mentions belonging to “One token” and “Multiple ##” categories,
the query is categorized as “Multiple ##”. Figure 3 compares the performance
of monoBERT and EM-BERT models for these four mention tokenization categories.
It shows that gains of EM-BERT are more pronounced for queries in “Multiple
##” and “One ##” categories than other categories. This supports our claim
that queries with uncommon entity mentions are most helped by including their
entity embeddings in the EM-BERT model. Additionally, we illustrate the
relation between different tokenization categories and the query types mention
(cf. Section 4.3) in Figure 3. The plot shows that 51% of INEX-LD queries have
no linked entities, explaining marginal improvements of EM-BERT for these
queries. This plot also describes why EM-BERT performs best on ListSearch and
QALD-2 queries: most of the less known entities fall in “Multiple ##” and “One
##” categories.
These results provide an answer to RQ2: _EM-BERT helps queries that are linked
to at least one entity. Additionally, entity information provided by the EM-
BERT model is most useful for less popular entities, which BERT cannot treat
their mentions as a single token._
Table 4. Comparison between EM-BERT and monoBERT for two example queries. Each query is listed twice, first with the normal BERT tokenization, then with the EM-BERT tokenization, both with the attentions of the first layer and first attention head. The Rel. column indicates the relevance judgement for that entity. The last two columns show the corresponding ranks obtained by the monoBERT/EM-BERT models for the entity. Query | Entity abstract | Rel. | Baseline | Comparison
---|---|---|---|---
give me all movies directed by francis ford cop ##pol ##a | christopher cop ##pol ##a ( born january 25 , 1962 ) is a film director and producer | 0 | monoBERT: 2 | EM-BERT: 551
give me all movies directed by francis ford cop ##pol ##a / ENTITY/Francis_Ford_Coppola | rumble fish / ENTITY/Rumble_Fish is an american / ENTITY/United_States 1983 drama film directed by francis ford cop ##pol ##a / ENTITY/Francis_Ford_Coppola . | 2 | monoBERT: 119 | EM-BERT: 1
give me the capitals of all countries in africa | list of african depend ##encies — including the respective capitals . | 0 | monoBERT: 1 | EM-BERT: 105
give me the capitals of all countries in africa / ENTITY/Africa | da ##kar / ENTITY/Dakar is the capital and largest city of senegal . it is located on the cap - ve ##rt peninsula / ENTITY/Cap-Vert on the atlantic coast / ENTITY/East_Coast_of_the_United_States and is the western ##most city in the old world / ENTITY/Europe and on the african / ENTITY/Africa mainland . | 1 | monoBERT: 82 | EM-BERT: 1
### 5.3. Model Analysis
To investigate _RQ3: Why does the EM-BERT model work and what does it learn
during the fine-tuning stage?_, we examine the proximity of the embedding
tokens compared to the mention tokens in the final layer of the BERT network,
and also discuss several examples of helped queries.
#### Entities vs. mention embeddings
To get an understanding of what BERT learns from entity information, we
compare entity and mention embeddings in the final layer of EM-BERT. We
randomly choose eight queries with 100-200 relevant entities. For each of
those documents, we take the first four linked entities and extract their
entity and mention embeddings from the final layer of EM-BERT. The mention
embeddings are taken from the first starting token before that entity. We use
UMAP (McInnes et al., 2018) with the default settings to plot these embeddings
in 2-dimensional space; see Figure 4. As is evident from the plots, the entity
embeddings form more pronounced clusters than the entity mention embeddings.
For example, in the entity embedding plot, American bridges and rivers build a
cluster together with other entities based in the US or Canada, while being
close to the corresponding queries. This also indicates that the entity
information injected through entity embeddings prevails into the final layers
of the BERT model, keeping a better representation of entity information than
provided with the default BERT word piece tokens.
#### Insightful Examples
We discuss in detail a few (hand-picked) example queries that are especially
helped or hurt by re-ranking with the EM-BERT model. Table 4 visualizes the
attention weights of the [CLS] token for the first attention head on the first
layer, for sample query-entity pairs. In the query “Give me all movies
directed by Francis Ford Coppola”, monoBERT incorrectly ranks other filmmaker
members of the Coppola family, matching the last name only. EM-BERT, in
contrast, finds the correct entity, as movies directed by Francis Ford Coppola
are situated close to him in the embedding space (as opposed to the movies by
his daughter Sofia Coppola). Table 4 shows that EM-BERT gives high attention
to the entity tokens relevant to the entity and query. It also shows high
attention to the words “movies”, which is essential here.
We can also see that through fine-tuning on DBpedia-entity, EM-BERT learns to
retrieve specific entities for queries requesting a list. For example for the
query “Give me the capitals of all countries in Africa”, while monoBERT’s
highest ranked entities are general overview pages like ‘List of African
dependencies’, all of EM-BERT’s highest-ranked entities are capitals of
African countries like ‘Dakar’ and ‘Porto-Novo’.
### 5.4. Performance on Passage Retrieval
Finally, we turn to our last research question, _RQ4: How does our entity-
enriched BERT-based model perform on other ranking tasks?_ To further our
understanding on the effect of entity-enriched BERT for IR tasks, we compare
the monoBERT and EM-BERT models after the first fine tuning stage on the MS
MARCO collection. We observe no significant differences between the two models
for both queries with and without linked entities. A similar observation is
made for the TREC Conversation Assistant Track (TREC-CAsT) (Joko et al.,
2022). We posit that the added value of enhancing BERT-based retrieval models
with explicit entity representations is mainly pronounced for entity-oriented
tasks.
## 6\. Conclusion and future work
In this research, we investigated the value of adding entity embeddings to
BERT for entity search. We proposed an entity-enhanced BERT-based retrieval
model, EM-BERT, and compared it with a regular BERT model. We found that EM-
BERT improves substantially over the state-of-the-art results, showing that
entity-oriented tasks benefit from entity embeddings in BERT, as is
hypothesized in (Broscheit, 2019). We found three categories where entities
help the most: complex natural language queries, queries requesting a list of
entities with a certain property, and queries with less known entities, in
which their mentions get split in multiple tokens by the BERT tokenizer. We
also showed that EM-BERT, unlike BERT-based retrieval models, can be fine-
tuned with limited training data, thus being effective for lesser-resourced
entity-related tasks.
For further research, the effect of the entity linking method can be
investigated; here we used REL, which is an entity linker with high precision.
It would be also interesting to evaluate the performance of other dense and
sparse retrieval methods on entity-oriented queries and tasks. This, however,
may require additional training data for fine-tuning on entity-oriented
queries, which would create a supplementary MS Marco subset to train and
evaluate ranking systems.
## References
* (1)
* Akbik et al. (2018) Alan Akbik, Duncan Blythe, and Roland Vollgraf. 2018\. Contextual String Embeddings for Sequence Labeling. In _Proc. of 27th International Conference on Computational Linguistics_ _(COLING ’18)_. 1638–1649.
* Balog (2018) Krisztian Balog. 2018\. _Entity-Oriented Search_. The Information Retrieval Series, Vol. 39. Springer.
* Blanco et al. (2015) Roi Blanco, Giuseppe Ottaviano, and Edgar Meij. 2015\. Fast and Space-Efficient Entity Linking in Queries. _Proc. of the Eighth ACM International Conference on Web Search and Data Mining_ (2015), 179–188.
* Bordes et al. (2013) Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko. 2013\. Translating embeddings for modeling multi-relational data. In _Proc. of the 26th International Conference on Neural Information Processing Systems_ _(NeurIPS ’13)_. 2787–2795.
* Broscheit (2019) Samuel Broscheit. 2019\. Investigating Entity Knowledge in BERT with Simple Neural End-To-End Entity Linking. In _Proc. of the 23rd Conference on Computational Natural Language Learning_ _(CoNLL ’19)_. 677–685.
* Cao et al. (2021) Nicola De Cao, Gautier Izacard, Sebastian Riedel, and Fabio Petroni. 2021. Autoregressive Entity Retrieval. In _Proc. of the International Conference on Learning Representations_ _(ICLR ’21)_.
* Cornolti et al. (2016) Marco Cornolti, Paolo Ferragina, Massimiliano Ciaramita, Stefan Rüd, and Hinrich Schütze. 2016\. A Piggyback System for Joint Entity Mention Detection and Linking in Web Queries. In _Proc. of the 25th International Conference on World Wide Web_ _(WWW ’16)_. 567–578.
* Dalton et al. (2020) Jeffrey Dalton, Chenyan , and Jamie Callan. 2020. TREC CAsT 2019: The conversational assistance track overview. _arXiv preprint_ (2020). arXiv:2003.13624
* Dalton et al. (2014) Jeffrey Dalton, Laura Dietz, and James Allan. 2014\. Entity Query Feature Expansion Using Knowledge Base Links. In _Proc. of the 37th International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR ’14)_. 365–374.
* Daza et al. (2021) Daniel Daza, Michael Cochez, and Paul Groth. 2021\. Inductive Entity Representations from Text via Link Prediction. In _Proc. of the Web Conference 2021_ _(WWW ’21)_. 798–808.
* Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In _Proc. of The North American Chapter of the Association for Computational Linguistics ’19_ _(NAACL ’19)_. 4171–4186.
* Dietz (2019) Laura Dietz. 2019\. ENT Rank: Retrieving Entities for Topical Information Needs through Entity-Neighbor-Text Relations. In _Proc. of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR’19)_. 215–224.
* Dodge et al. (2020) Jesse Dodge, Gabriel Ilharco, Roy Schwartz, Ali Farhadi, Hannaneh Hajishirzi, and Noah Smith. 2020\. Fine-tuning pretrained language models: Weight initializations, data orders, and early stopping. _arXiv preprint arXiv:2002.06305_ (2020).
* Ferragina and Scaiella (2010) Paolo Ferragina and Ugo Scaiella. 2010. TAGME: On-the-Fly Annotation of Short Text Fragments (by Wikipedia Entities). In _Proc. of the International Conference on Information and Knowledge Management_ _(CIKM ’10)_. 1625–1628.
* Garigliotti et al. (2019) Darío Garigliotti, Faegheh Hasibi, and Krisztian Balog. 2019\. Identifying and Exploiting Target Entity Type Information for Ad hoc Entity Retrieval. _Information Retrieval Journal_ 22, 3 (2019), 285–323.
* Gerritse et al. (2020) Emma Gerritse, Faegheh Hasibi, and Arjen De Vries. 2020\. Graph-Embedding Empowered Entity Retrieval. In _Proc. of European Conference on Information Retrieval_ _(ECIR ’20)_.
* Hasibi et al. (2015) Faegheh Hasibi, Krisztian Balog, and Svein Erik Bratsberg. 2015\. Entity Linking in Queries: Tasks and Evaluation. In _Proc. of the 2015 International Conference on The Theory of Information Retrieval_ _(ICTIR ’15)_. 171–180.
* Hasibi et al. (2016) Faegheh Hasibi, Krisztian Balog, and Svein Erik Bratsberg. 2016\. Exploiting Entity Linking in Queries for Entity Retrieval. In _Proc. of the 2016 ACM International Conference on the Theory of Information Retrieval_ _(ICTIR ’16)_. 209–218.
* Hasibi et al. (2017a) Faegheh Hasibi, Krisztian Balog, and Svein Erik Bratsberg. 2017a. Entity Linking in Queries: Efficiency vs. Effectiveness. In _Proc. of the 39th European Conference on Information Retrieval_ _(ECIR ’17)_. 40–53.
* Hasibi et al. (2017b) Faegheh Hasibi, Krisztian Balog, Daŕio Garigliotti, and Shuo Zhang. 2017b. Nordlys: A toolkit for entity-oriented and semantic search. In _Proc. of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR ’17)_. 1289–1292.
* Hasibi et al. (2017c) Faegheh Hasibi, Fedor Nikolaev, Chenyan Xiong, Krisztian Balog, Svein Erik Bratsberg, Alexander Kotov, and Jamie Callan. 2017c. DBpedia-Entity V2: A Test Collection for Entity Search. In _Proc. of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR ’19)_. 1265–1268.
* Hofstätter et al. (2020) Sebastian Hofstätter, Sophia Althammer, Michael Schröder, Mete Sertkan, and Allan Hanbury. 2020\. Improving Efficient Neural Ranking Models with Cross-Architecture Knowledge Distillation. _arXiv preprint_ (2020). arXiv:2010.02666
* Jiang et al. (2020) Zhengbao Jiang, Antonios Anastasopoulos, Jun Araki, Haibo Ding, and Graham Neubig. 2020. X-FACTR: Multilingual Factual Knowledge Retrieval from Pretrained Language Models. In _Proc. of the 2020 Conference on Empirical Methods in Natural Language Processing_ _(EMNLP ’20)_. 5943–5959.
* Joko et al. (2022) Hideaki Joko, Emma J Gerritse, Faegheh Hasibi, and Arjen P de Vries. 2022. Radboud University at TREC CAsT 2021. In _TREC_.
* Joshi et al. (2020) Mandar Joshi, Danqi Chen, Yinhan Liu, Daniel S. Weld, Luke Zettlemoyer, and Omer Levy. 2020\. SpanBERT: Improving Pre-training by Representing and Predicting Spans. _Transactions of the Association for Computational Linguistics_ (2020), 64–77.
* Joshi et al. (2019) Mandar Joshi, Omer Levy, Luke Zettlemoyer, and Daniel Weld. 2019\. BERT for Coreference Resolution: Baselines and Analysis. In _Proc. of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing_ _(EMNLP-IJCNLP ’19)_. 5803–5808.
* Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. _arXiv preprint_ (2019). arXiv:1907.11692
* McInnes et al. (2018) Leland McInnes, John Healy, Nathaniel Saul, and Lukas Grossberger. 2018. UMAP: Uniform Manifold Approximation and Projection. _The Journal of Open Source Software_ 3, 29 (2018), 861\.
* Mikolov et al. (2013) Tomas Mikolov, Ilya Sutskever, Kai Chen, Gregory S. Corrado, and Jeffrey Dean. 2013. Distributed Representations of Words and Phrases and their Compositionality. In _Proc. of Advances in Neural Information Processing Systems_ _(NeurIPS ’13)_. 3111–3119.
* Nguyen et al. (2016) Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. 2016. MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. In _Proc. of the Workshop on Cognitive Computation: Integrating neural and symbolic approaches 2016_.
* Nikolaev and Kotov (2020) Fedor Nikolaev and Alexander Kotov. 2020. Joint Word and Entity Embeddings for Entity Retrieval from a Knowledge Graph. In _Proc. of the European Conference on Information Retrieval_ _(ECIR ’20)_. 141–155.
* Nogueira and Cho (2019) Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Re-ranking with BERT. _arXiv preprint_ (2019). arXiv:1901.04085
* Nogueira et al. (2019) Rodrigo Nogueira, Wei Yang, Kyunghyun Cho, and Jimmy Lin. 2019\. Multi-stage document ranking with BERT. _arXiv preprint_ (2019). arXiv:1910.14424
* Peters et al. (2019) Matthew E. Peters, Mark Neumann, Robert Logan, Roy Schwartz, Vidur Joshi, Sameer Singh, and Noah A. Smith. 2019. Knowledge Enhanced Contextual Word Representations. In _Proc. of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing_ _(EMNLP-IJCNLP ’19)_. 43–54.
* Petroni et al. (2019) Fabio Petroni, Tim Rocktäschel, Patrick S. H. Lewis, Anton Bakhtin, Yuxiang Wu, Alexander H. Miller, and Sebastian Riedel. 2019. Language Models as Knowledge Bases?. In _Proc. of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing_ _(EMNLP-IJCNLP ’19)_. 2463–2473.
* Phang et al. (2019) Jason Phang, Thibault Fevry, and Samuel R. Bowman. 2019\. Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks. _arXiv preprint arXiv:1811.01088_ (2019).
* Poerner et al. (2020) Nina Poerner, Ulli Waltinger, and Hinrich Schütze. 2020\. E-BERT: Efficient-Yet-Effective Entity Embeddings for BERT. In _Findings of the Association for Computational Linguistics_ _(ELMNLP ’20)_. 803–818.
* Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. _Journal of Machine Learning Research_ 21, 140 (2020), 1–67.
* Talmor et al. (2020) Alon Talmor, Yanai Elazar, Yoav Goldberg, and Jonathan Berant. 2020. oLMpics - On what Language Model Pre-training Captures. _Transactions of the Association for Computational Linguistics_ 8 (2020), 743–758.
* Thakur et al. (2021) Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, and Iryna Gurevych. 2021\. BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models. In _Proc. of 35th Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2)_ _(NeurIPS ’21)_.
* van Hulst et al. (2020) Johannes M. van Hulst, Faegheh Hasibi, Koen Dercksen, Krisztian Balog, and Arjen P. de Vries. 2020. REL: An Entity Linker Standing on the Shoulders of Giants. In _Proc. of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR ’20)_.
* Wang et al. (2020) Chenguang Wang, Xiao Liu, and Dawn Song. 2020. Language Models are Open Knowledge Graphs. _arXiv preprint_ (2020). arXiv:2010.11967
* Wang et al. (2020) Ruize Wang, Duyu Tang, Nan Duan, Zhongyu Wei, Xuanjing Huang, Jianshu ji, Guihong Cao, Daxin Jiang, and Ming Zhou. 2020. K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters. _arXiv preprint_ (2020). arXiv:2002.01808
* Wang et al. (2020) Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020. MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers. In _Proc. of Advances in Neural Information Processing Systems_ _(NeurIPS ’21)_ , H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.). 5776–5788.
* Wang et al. (2019a) Xiaozhi Wang, Tianyu Gao, Zhaocheng Zhu, Zhiyuan Liu, Juanzi Li, and Jian Tang. 2019a. KEPLER: A unified model for knowledge embedding and pre-trained language representation. _arXiv preprint_ (2019). arXiv:1911.06136
* Wang et al. (2019b) Zhiguo Wang, Patrick Ng, Xiaofei Ma, Ramesh Nallapati, and Bing Xiang. 2019b. Multi-passage BERT: A Globally Normalized BERT Model for Open-domain Question Answering. In _Proc. of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing_ _(EMNLP-IJCNLP ’19)_. 5878–5882.
* Xiong et al. (2017a) Chenyan Xiong, Jamie Callan, and Tie-Yan Liu. 2017a. Word-Entity Duet Representations for Document Ranking. In _Proc. of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR ’17)_. 763–772.
* Xiong et al. (2017b) Chenyan Xiong, Russell Power, and Jamie Callan. 2017b. Explicit Semantic Ranking for Academic Search via Knowledge Graph Embedding. In _Proc. of the 26th International Conference on World Wide Web_ _(WWW ’17)_. 1271–1279.
* Yamada et al. (2020a) Ikuya Yamada, Akari Asai, Jin Sakuma, Hiroyuki Shindo, Hideaki Takeda, Yoshiyasu Takefuji, and Yuji Matsumoto. 2020a. Wikipedia2Vec: An Efficient Toolkit for Learning and Visualizing the Embeddings of Words and Entities from Wikipedia. In _Proc. of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations_ _(EMNLP ’20)_. 23–30.
* Yamada et al. (2020b) Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, and Yuji Matsumoto. 2020b. LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention. In _Proc. of the 2020 Conference on Empirical Methods in Natural Language Processing_ _(EMNLP ’20)_. 6442–6454.
* Yamada et al. (2016) Ikuya Yamada, Hiroyuki Shindo, Hideaki Takeda, and Yoshiyasu Takefuji. 2016. Joint learning of the embedding of words and entities for named entity disambiguation. In _Proc. of The SIGNLL Conference on Computational Natural Language Learning_ _(SIGNLL ’16)_. 250–259.
* Yang et al. (2019) Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R Salakhutdinov, and Quoc V Le. 2019\. XLNet: Generalized Autoregressive Pretraining for Language Understanding. In _Advances in Neural Information Processing Systems_ _(NeurIPS ’19)_. 5753–5763.
* Zhang et al. (2021) Tianyi Zhang, Felix Wu, Arzoo Katiyar, Kilian Q Weinberger, and Yoav Artzi. 2021. Revisiting Few-sample BERT Fine-tuning. In _Proc. of 2021 International Conference on Learning Representations_ _(ICLR ’20)_.
* Zhang et al. (2019) Zhengyan Zhang, Xu Han, Zhiyuan Liu, Xin Jiang, Maosong Sun, and Qun Liu. 2019\. ERNIE: Enhanced Language Representation with Informative Entities. In _Proc. of the 57th Annual Meeting of the Association for Computational Linguistics_ _(ACL ’19)_. 1441–1451.
|
# Multiply Robust Estimation of Causal Effect Curves for Difference-in-
Differences Designs
Gary Hettinger
Department of Biostatistics, Epidemiology, and Informatics
University of Pennsylvania
Philadelphia, PA, U.S.A.
<EMAIL_ADDRESS>
&Youjin Lee
Department of Biostatistics
Brown University
Providence, RI, U.S.A.
&Nandita Mitra
Department of Biostatistics, Epidemiology, and Informatics
University of Pennsylvania
Philadelphia, PA U.S.A.
###### Abstract
Researchers commonly use difference-in-differences (DiD) designs to evaluate
public policy interventions. While established methodologies exist for
estimating effects in the context of binary interventions, policies often
result in varied exposures across regions implementing the policy. Yet,
existing approaches for incorporating continuous exposures face substantial
limitations in addressing confounding variables associated with intervention
status, exposure levels, and outcome trends. These limitations significantly
constrain policymakers’ ability to fully comprehend policy impacts and design
future interventions. In this study, we propose innovative estimators for
causal effect curves within the DiD framework, accounting for multiple sources
of confounding. Our approach accommodates misspecification of a subset of
treatment, exposure, and outcome models while avoiding any parametric
assumptions on the effect curve. We present the statistical properties of the
proposed methods and illustrate their application through simulations and a
study investigating the diverse effects of a nutritional excise tax.
_Keywords_ Dose Response $\cdot$ Doubly Robust $\cdot$ Health Policy $\cdot$
Semiparametric
## 1 Introduction
Public policies wield significant influence on a broad spectrum of population
outcomes, including those related to public health and economics.(Pollack
Porter et al., 2018) Accordingly, decision-makers require rigorous evaluation
of such policies to comprehend their effectiveness and inform the design of
subsequent policy initiatives. Complicating the comprehensive assessment of
policy effects, units within a study may be differentially exposed to a policy
for various reasons. For example, exposure to an excise tax policy may depend
on the ease in which a buyer can evade the tax (e.g., geographic proximity to
cross-border shopping) or responsive price adjustments made by the seller
which may diminish the actual tax rate seen by consumers.(Chaloupka et al.,
2019) These differential exposures can lead to effect heterogeneity that, if
understood, can help identify affected populations and inspire future
interventions that may generate a different exposure distribution.(Cintron et
al., 2022)
When evaluating effects of binary policy interventions, researchers commonly
use difference-in-differences (DiD) approaches, which exploit the so-called
counterfactual parallel trends assumption to utilize outcomes from a control
group, measured before and after the intervention, to impute what would have
happened in the treated group in the absence of intervention.(Heckman et al.,
1997; Abadie, 2005) This assumption requires that there are no confounding
differences between treated and control groups that also affect the outcome
trend between the before- and after-intervention windows. Recent work has
demonstrated that incorporating a continuous measure of exposure requires even
stricter parallel trends assumptions.(Callaway et al., 2021) As DiD methods to
incorporate continuous exposure measures are limited, many researchers
alternatively conduct subgroup analyses to explore effect heterogeneity, but
these cannot identify causal differences between subgroups as they adjust for
confounding within, rather than between, subgroups.
In the DiD setting with a binary exposure variable, substantial methodology
has been developed to relax the counterfactual parallel trends assumption by
adjusting for observed confounders between the treated and control
groups.(Abadie, 2005; Li and Li, 2019; Sant’Anna and Zhao, 2020) However, few
works have targeted confounding between different exposure levels among the
treated group when estimating effect curves in DiD studies. Han and Yu (2019)
considered DiD extensions of the approach developed by Hirano and Imbens
(2004) to estimate causal effect curves by utilizing an estimate for the
generalized propensity score with continuous exposures as a bias-adjusting
covariate in a regression model. However, this approach hinges entirely on
accurately specifying the conditional density function for the continuous
exposure given covariates and may yield unstable estimates even under correct
specification.(Austin, 2018) Beyond DiD studies, there have been notable
advancements in estimating dose-response curves in the presence of
confounding. These include doubly robust estimators that utilize both
propensity scores and outcome regression models to estimate effect curves and
generally improve upon approaches solely reliant on the propensity score in
terms of both efficiency and robustness.(Díaz and van der Laan, 2013; Kennedy
et al., 2017; Bonvini and Kennedy, 2022) Still, methods to harness these
properties have yet to be developed to account for the multiple levels of
confounding – between treated and control groups and between different
exposure levels among the treated group – and longitudinal design present in
policy evaluations that employ a control group.
In this work, we develop novel estimators for causal effect curves that
account for multiple sources of confounding under a DiD framework. Our
methodology incorporates both outcome and treatment models to rigorously
adjust for confounding effects and exploits the efficiency of semi-parametric
influence-function based estimators. Our estimators are multiply robust in the
sense that only a subset of the four required nuisance function models need to
be well-specified for consistent effect estimation. To further mitigate
potential model misspecification, our approach allows for flexible modeling
techniques to capture complex confounding relationships and non-parametric
effect curves.
The remainder of the paper is organized as follows: in Section 2, we introduce
the setting of interest, identification assumptions, and proposed estimators.
In Section 3, we demonstrate finite sample properties of our estimators under
different nuisance function model specifications through simulation studies.
We then apply our methodology to study the effects of the Philadelphia
beverage tax in Section 4 and conclude with a discussion in Section 5.
## 2 Methods
### 2.1 Setting
We assume an intervention is introduced between two time-periods, $t=0,1$.
Units, denoted with subscript $i$, are either in the intervention group,
$A_{i}=1$, or not, $A_{i}=0$, and the intervention assignment status at time
$t$ is denoted by $Z_{it}=tA_{i}\in\\{0,1\\}$. The actual exposure to the
intervention received by a unit is measured by $D_{i}\in\mathbbm{R}$, with the
exposure status given by $\zeta_{it}=tD_{i}$. Depending on our choice of
exposure, a “zero" exposure may not represent true protection from the
intervention and therefore we denote this by $D_{i}=\emptyset$ or
$\zeta_{it}=\emptyset$. We denote population-level vectors for intervention
and exposure variables as $\mathbf{A}$, $\mathbf{Z_{t}}$, $\mathbf{D}$, and
$\boldsymbol{\zeta_{t}}$, respectively. We observe outcomes of interest at
each time point, $Y_{it}$. Finally, units and their outcomes may differ
according to a vector of observed pre-intervention covariates,
$\mathbf{X_{i}}$, although we could also include time-varying covariates
assuming time-invariant effects.(Zeldow and Hatfield, 2021)
The following assumptions allow us to define potential outcomes in terms of
intervention and exposure status as $Y_{it}^{(Z_{it},\zeta_{it})}$:
###### Assumption 1 (Arrow of Time)
$Y_{i0}^{(\mathbf{Z_{0}},\boldsymbol{\zeta_{0}},\mathbf{Z_{1}},\boldsymbol{\zeta_{1}})}=Y_{i0}^{(\mathbf{Z_{0}},\boldsymbol{\zeta_{0}})}$
###### Assumption 2 (Stable Unit Treatment Value Assumption (SUTVA))
$Y_{it}^{(\mathbf{Z_{t}},\boldsymbol{\zeta_{t}})}=Y_{it}^{(Z_{it},\zeta_{it})}$
Assumption 1 states that potential outcomes do not depend on future treatment
or exposure, which would be violated if for example units adapted their
behavior in anticipation of an upcoming effect. After connecting potential
outcomes to treatment at a given time, we further assume that potential
outcomes of a given unit only depend on population intervention and exposure
status through a unit’s own intervention and exposure status (Assumption 2).
This formulation can be considered a relaxation of the traditional SUTVA from
the setting with only a binary intervention by allowing potential outcomes to
vary according to $\zeta_{it}$ in addition to $Z_{it}$.
We are primarily interested in estimating a curve for the causal effect of the
intervention at different exposure levels, $\delta$, as represented by what we
denote the Average Dose Effect on the Treated:
$ADT(\delta):=\Psi(\delta)=E[Y_{1}^{(1,\delta)}-Y_{1}^{(0,\emptyset)}|A=1]$
(1)
This estimand answers the question, what would be the average effect of a
policy on the treated group where all treated units received the exposure of
$\delta$? The Average Treatment Effect on the Treated (ATT) commonly estimated
in DiD studies under a binary treatment study has one unobserved potential
outcome for treated units – $Y_{1}^{(0,\emptyset)}$, which is unobserved for
all treated units. Additionally, standard dose response curves focus on
identifying analogies to $Y_{1}^{(1,\delta)}$, which is unobserved for treated
units $i$, where $D_{i1}\neq\delta$. To estimate (1) in our setting, we must
identify both sets of unobserved potential outcomes.
A natural extension of the $ADT(\delta)$ would consider the expected value of
(1) over a relevant distribution of doses to quantify an aggregate effect,
which we refer to as the Stochastic Average Dose Effect on the Treated:
$\begin{split}SADT&:=\Psi=E_{\mathbbm{D}}[E[Y_{1}^{(1,\delta)}-Y_{1}^{(0,\emptyset)}|A=1]|A=1]\\\
&=\int\limits_{\mathbbm{D}}ADT(\delta)df(\delta|A=1)\end{split}$ (2)
This estimand represents the average effect of an intervention that randomly
assigns treatment to the treated group based on the density of exposures, $f$,
and thus carries a different interpretation than other proposed aggregate DiD
estimands like the ATT or average slope effects from Callaway et al. (2021)
and de Chaisemartin and D’Haultfoeuille (2022). While stochastic estimands
like the $SADT$ have many potential uses (e.g., (Muñoz and van der Laan,
2012)), we primarily will use the $SADT$ for its relevance in deriving an
estimator for the $ADT(\delta)$.
### 2.2 Identification
To identify the $ADT(\delta)$ and $SADT$, we require additional assumptions to
link the unobservable potential outcomes to observable data. The strong
assumption of binary DiD approaches is that of counterfactual parallel trends:
###### Assumption 3 (Conditional Counterfactual Parallel Trends Between
Treated and Control)
$E[Y_{1}^{(0,\emptyset)}-Y_{0}^{(0,\emptyset)}|A=1,\mathbf{X}]=E[Y_{1}^{(0,\emptyset)}-Y_{0}^{(0,\emptyset)}|A=0,\mathbf{X}]$
Assumption 3 allows us to connect $E[Y_{1}^{(0,\emptyset)}|A=1]$ in (1) to
observable potential outcomes in the control group, $A=0$. While the DiD
design innately adjusts for baseline differences between the treated and
control groups by within-group differencing, it does not address confounding
between treatment group and outcome trends. Here, we have adopted the approach
several have taken to relax this assumption by assuming parallel trends holds
after conditioning on a set of observed baseline confounders.(Abadie, 2005; Li
and Li, 2019; Sant’Anna and Zhao, 2020)
To connect $E[Y_{1}^{(1,\delta)}|A=1]$ to observable potential outcomes for
all treated units, we require an additional, parallel trends assumption:
###### Assumption 4 (Conditional Counterfactual Parallel Trends Among Treated
Between Doses)
$E[Y_{1}^{(1,\delta)}-Y_{0}^{(0,\emptyset)}|A=1,D=\delta,\mathbf{X}]=E[Y_{1}^{(1,\delta)}-Y_{0}^{(0,\emptyset)}|A=1,\mathbf{X}]\text{
for all }\delta\in\mathbbm{D}$
For intuition, consider similar units, as defined by $\mathbf{X}$, between (i)
the subset of the treated group that receives a particular exposure level,
$\delta$, and (ii) the entire treated group. In principle, we then require
that the observed trends in group (i) are representative of the counterfactual
trends in group (ii) were all of group (ii) given that exposure level when
intervened on. A visual example of this assumption is presented in Figure 1.
Previous works have noted the necessity of similar assumptions in the
continuous exposure setting either unconditionally or with regards to
conditional distributions rather than expectations.(Callaway et al., 2021; Han
and Yu, 2019)
Figure 1: An example of Assumption 4 where exposure level is discretized into
high and low dose for visualization.
While assumptions regarding parallel trends are by definition untestable using
observed data, it is common for practitioners to test for counterfactual
parallel trends between the treated and control groups leading up to the
intervention when multiple pre-treatment observations are available (e.g.,
$t=-1$) under a binary treatment.(Bilinski and Hatfield, 2018) The test
derives its usefulness by assuming the fully observable relationship between
$E[Y_{0}^{(0,\emptyset)}-Y_{-1}^{(0,\emptyset)}|A=1]$ and
$E[Y_{0}^{(0,\emptyset)}-Y_{-1}^{(0,\emptyset)}|A=0]$ is a reasonable proxy
for the unobservable relationship between
$E[Y_{1}^{(0,\emptyset)}-Y_{0}^{(0,\emptyset)}|A=1]$ and
$E[Y_{1}^{(0,\emptyset)}-Y_{0}^{(0,\emptyset)}|A=0]$. Similar placebo tests
are possible for Assumption 4, however part of this assumption,
$E[Y_{t}^{(1,\delta)}|A=1]$, is unobservable even in the pre-treatment period.
Therefore, the usefulness of comparable tests assessing
$E[Y_{0}^{(0,\emptyset)}-Y_{-1}^{(0,\emptyset)}|A=1,D=\delta]=E[Y_{0}^{(0,\emptyset)}-Y_{-1}^{(0,\emptyset)}|A=1]$
also depends on how well these trends represent counterfactual trends at
different exposure levels.
The final two assumptions are extensions of standard causal inference
assumptions to the continuous exposure setting:
###### Assumption 5 (Consistency)
$Y_{it}^{(z_{it},d_{it})}=Y_{it}$ when $(Z_{it},D_{it})=(z_{it},d_{it})$
###### Assumption 6 (Positivity)
There is $\epsilon>0$ such that
(i) $\epsilon\leq\pi_{A}(\mathbf{x})<1$ for all $\mathbf{x}\in\mathbbm{X}$
and
(ii) $\epsilon\leq\pi_{D}(\delta|\mathbf{x},A=1)$ for all
$\mathbf{x}\in\mathbbm{X}|A=1,\text{ }\delta\in\mathbbm{D}$,
where $\pi_{A}(\mathbf{x})=P(A=1|\mathbf{x})$ and
$\pi_{D}(\delta|\mathbf{x},A=1)=f(\delta|A=1,\mathbf{X})$.
Here, we require a stricter positivity assumption (Assumption 6) than the
binary setting, which only requires Assumption 6(i). Assumption 6(ii) relates
to covariate overlap between the different dose levels and would violated if
certain subpopulations, defined by confounders $\mathbf{X}$, were restricted
from certain exposure levels.
After these assumptions, we then can frame (1) and (2) fully with observable
data as:
###### Theorem 1 (Identifiability)
Under Assumptions 1-6, (1) is identifiable through
$\Psi(\delta)=E[\mu_{1,\Delta}(\delta,\mathbf{X})-\mu_{0,\Delta}(\mathbf{X})|A=1]$,
where
$\mu_{1,\Delta}(\delta,\mathbf{X})=E[Y_{1}-Y_{0}|A=1,D=\delta,\mathbf{X}]$ and
$\mu_{0,\Delta}(\mathbf{X})=E[Y_{1}-Y_{0}|A=0,\mathbf{X}]$.
Accordingly, (2) is then identifiable through $\Psi=E[\Psi(D)|A=1]$.
### 2.3 Multiply-Robust Estimator
An ideal estimator for (1) would then be based on the efficient influence
function (EIF) for this estimand, as EIF-based estimators come with optimal
convergence properties and often, though not necessarily, certain robustness
to model misspecification. However, the EIF is not tractable for (1) without
requiring parametric restrictions on the form of the otherwise infinite-
dimensional curve as such a curve does not satisfy the requirement for
pathwise differentiability.(Kennedy, 2022) Parametric restrictions generally
limit method applicability and data fit, so we do not consider this a viable
path forward. However, the EIF for (2) is tractable and given by:
###### Theorem 2 (Efficient Influence Function for $\Psi$)
The efficient influence function for $\Psi$, is given by:
$\phi(\mathbf{O})=\frac{A}{P(A=1)}\xi(Y_{1},Y_{0},A,D,\mathbf{X})+\tau(Y_{1},Y_{0},A,\mathbf{X})+J(A,\mathbf{X})-\Psi$
Where
$\displaystyle\xi(Y_{1},Y_{0},A,D,\mathbf{X})$
$\displaystyle=m(D)+\frac{(Y_{1}-Y_{0})-\mu_{1,\Delta}(D,\mathbf{X})}{\pi_{D}(D|A=1,\mathbf{X})}f(D|A=1)$
$\displaystyle\tau(Y_{1},Y_{0},A,\mathbf{X})$
$\displaystyle=\frac{A}{P(A=1)}\mu_{0,\Delta}(\mathbf{X})+\frac{1-A}{P(A=1)}\frac{\pi_{A}(\mathbf{X})}{1-\pi_{A}(\mathbf{X})}[(Y_{1}-Y_{0})-\mu_{0,\Delta}(\mathbf{X})]$
$\displaystyle J(A,\mathbf{X})$
$\displaystyle=\frac{A}{P(A=1)}[\int\limits_{\mathbbm{D}}\\{\mu_{1,\Delta}(d,\mathbf{X})-m(d)\\}df(d|A=1)]$
And
$m(D)=\int\limits_{\mathbbm{X}}\mu_{1,\Delta}(D,\mathbf{x})dP(\mathbf{x}|A=1)$,
$f(D|A=1)=\int\limits_{\mathbbm{X}}\pi_{D}(D|A=1,\mathbf{x}))dP(\mathbf{x}|A=1)$
While notationally complex, the resulting influence function is largely
intuitive. Note that $\xi(\mathbf{Y_{1},Y_{0},A,D,\mathbf{X}})$ strongly
corresponds to $E[Y_{1}-Y_{0}|A=1,D=D]=E[Y_{1}(1,D)-Y_{0}(0,\emptyset)|A=1]$.
Here, $m(D)$ averages the conditional expectation,
$\mu_{1,\Delta}(D,\mathbf{X})=E[Y_{1}-Y_{0}|A=1,D,\mathbf{X}]$ across all
covariates in the treated group. The second term of $\xi$ re-weights observed
deviations from expectation by the balance-inducing generalized propensity
score weight, $w^{(1)}=f(D)/\pi_{D}(D|A=1,\mathbf{X})$, as a sort of bias
correction term when $\mu_{1,\Delta}$ is misspecified. Similarly,
$\tau(Y_{1},Y_{0},A,\mathbf{X})$ correseponds to
$E[Y_{1}-Y_{0}|A=1]=E[Y_{1}(0,\emptyset)-Y_{0}(0,\emptyset)|A=1]$, where now
we plug in treated unit covariates to
$\mu_{0,\Delta}(\mathbf{X})=E[Y_{1}-Y_{0}|A=0,\mathbf{X}]$. The second term
re-weights observed deviations from expectation for the control group by ATT
propensity score weights,
$w^{(0)}=\pi_{A}(\mathbf{X})/(1-\pi_{A}(\mathbf{X}))$ as a sort of bias
correction term when $\mu_{0}$ is misspecified. Finally, we note that
$J(\mathbf{A,\mathbf{X}})$ is independent of exposure dose and
$E[J(\mathbf{A,\mathbf{X}})|A=1]=0$, so will not factor into point estimation
for either (1) or (2). A similar decomposition for $\xi$ and $J$ was noted by
Kennedy et al. (2017) to generate their so-called pseudo-outcomes when
estimating a dose response curve for $E[Y^{(\delta)}]$ under a cross-sectional
study design.
Utilizing this influence function and subsequent decomposition, we then can
estimate (1) and (2) with an multi-step, plug-in procedure:
1. 1.
Estimate nuisance functions as
$\hat{\pi}_{D},\hat{\pi}_{A},\hat{\mu}_{1,\Delta},\hat{\mu}_{0,\Delta}$.
2. 2.
Plug in data and estimated nuisance functions to construct
$\hat{\tau}(\mathbf{O};\hat{\pi}_{A},\hat{\mu}_{0,\Delta})$ and take the
empirical mean of $\hat{\tau}$ to obtain $\hat{\theta}_{0}$.
3. 3.
Plug in data and estimated nuisance functions to construct
$\hat{\xi}(\mathbf{O};\hat{\pi}_{D},\hat{\mu}_{1,\Delta})$ and regress
$\hat{\xi}$ on dose variable $D$ using a flexible regression function
$r_{n}(\hat{\xi},D)$ to obtain $\hat{\theta}(\delta)$.
4. 4.
For (1), take the difference $\hat{\theta}(\delta)-\hat{\theta}_{0}$ as
$\hat{\Psi}(\delta)$.
5. 5.
For (2), take the empirical mean of $\frac{A}{P(A=1)}\hat{\xi}-\hat{\tau}$ as
$\hat{\Psi}$.
Practically speaking, $\hat{\pi}_{D}$ and $\hat{\mu}_{1,\Delta}$ must be
estimated only using data from the $A=1$ group whereas $\hat{\pi}_{A}$
requires data from both $A=1$ and $A=0$ groups. Estimation of
$\hat{\mu}_{0,\Delta}$ could incorporate data from both $A=1$ and $A=0$
groups, but the efficiency gained from utilizing $A=1$ data comes at the cost
of robustness in that correctly specifying $\hat{\mu}_{0,\Delta}$ would
require properly modeling treatment effect dynamics, linking its specification
to that of $\hat{\mu}_{1,\Delta}$. Another generally beneficial step in
practice is to normalize $w^{(1)}$ and $w^{(0)}$ to both have mean one as in
Hajek weighted estimators, which has previously been noted as important for
improving the stability of doubly robust DiD estimators.(Sant’Anna and Zhao,
2020)
The decomposition of the influence function $\phi$ also suggests components
related to doubly-robust estimators for both
$E[Y_{1}(1,D)-Y_{0}(0,\emptyset)|A=1]$ and
$E[Y_{1}(0,\emptyset)-Y_{0}(0,\emptyset)|A=1]$, foreshadowing the multiple
robustness properties of our proposed estimator:
###### Theorem 3 (Multiple Robustness)
Suppose that Assumptions 1-6 hold. Additionally assume that:
1. i.
At least one of the dose-specific nuisance functions {$\pi_{D}$,
$\mu_{1,\Delta}$} are correctly specified.
2. ii.
At least one of the intervention-specific nuisance functions {$\pi_{A}$,
$\mu_{0,\Delta}$} are correctly specified.
Then, $\hat{\Psi}$ is a consistent estimator for (2). Further, if
$r_{n}(\xi,d)$ is a regression estimator such that
$r_{n}(\xi,d)\overset{p}{\to}E[\xi|D=d]$, then $\hat{\Psi}(\delta)$ is a
consistent estimator for (1) when $\delta$ is an exposure value in the
interior of the compact support of $D$.
Thereby, we only require one correct model specification in each specified
pair of nuisance functions corresponding to
$E[Y_{1}(1,D)-Y_{0}(0,\emptyset)|A=1]$ and
$E[Y_{1}(0,\emptyset)-Y_{0}(0,\emptyset)|A=1]$, respectively. Accordingly, our
estimator will result in consistent effect estimation in 9 of the 16 ($56\%$)
permutations of correct/incorrect nuisance function specifications (i.e.,
{$\pi_{D}\in\\{\text{Correct,Incorrect}\\},\mu_{1,\Delta}\in\\{\text{Correct,Incorrect}\\},\pi_{A}\in\\{\text{Correct,Incorrect}\\},\mu_{0,\Delta}\in\\{\text{Correct,Incorrect}\\}$}).
Compared to standard doubly robust estimators which only require two nuisance
functions and thus achieve consistent effect estimation in 3 of the 4 ($75\%$)
specification permutations, we see the required cost of addressing multiple
levels of confounding in terms of robustness. However, if we consider the
specification of both outcome models, $\\{\mu_{1,\Delta},\mu_{0,\Delta}\\}$,
together and both propensity score models, $\\{\pi_{D},\pi_{A}\\}$, together,
then we have consistent estimates in 3 of the 4 specification permutations and
further achieve consistency under additional “partial” misspecifications.
The requirements regarding $r_{n}$ suggest the use of a flexible non-
parametric regression function to avoid misspecifying the conditional density
of $\xi$ given $D$. Following Kennedy et al. (2017), we specifically consider
a local linear kernel regression estimator for $r_{n}$, which is a non-
parametric estimator imposing relatively few regularity conditions regarding
effect curves and nuisance functions for consistency in this framework
(Appendix).
### 2.4 Uncertainty Quantification
Calculating confidence intervals for these multi-step estimators is non-
trivial, particularly for $\Psi(D)$ which does not absorb the semi-parametric
efficiency guarantees of influence-function based estimators for binary
treatments when all nuisance functions are well-specified. Since standard non-
parametric regressors are consistent but centered at a smoothed version of the
effect curve, rather than the true curve, we adapt previous works in
expressing uncertainty about this smoothed curve (Wasserman, 2006; Kennedy et
al., 2017). While our estimator is asymptotically normal under mild
restrictions on nuisance function convergence rates (Appendix), we note that
such results ignore finite-sample uncertainty in the estimation of nuisance
functions and may lead to underestimated variances (Lunceford and Davidian,
2004). Thereby, we consider two alternative approaches.
First, we follow previous approaches that represent plug-in causal inference
estimators as a series of unbiased estimating equations (i.e., score equations
for specified parametric nuisance function models, sample mean equation for
$\theta_{0}$ using $\hat{\tau}_{i}$, and kernel regression equation for
$\hat{\theta}(D)$ using $\hat{\xi}_{i}$) and apply M-estimation theory to
engineer sandwich estimators for the variance of estimated
parameters.(Stefanski and Boos, 2002; Cole et al., 2023) Concretely, this
approach first defines the estimating equation $\Gamma_{i}(\eta,\delta)$ by:
$\sum\limits_{i=1}^{n}\Gamma_{i}(\eta,\delta)=\sum\limits_{i=1}^{n}\begin{bmatrix}\theta(\delta)-\ell_{i}(\delta)\hat{\xi_{i}}\\\
\theta_{0}-\hat{\tau_{i}}\\\
A_{i}S_{\pi_{D}}(Y_{i1},Y_{i0},D_{i},\mathbf{X}_{i};\alpha^{(D)},\sigma^{2})\\\
A_{i}S_{\mu_{1}}(Y_{i1},Y_{i0},D_{i},\mathbf{X}_{i};\beta^{(D)})\\\
S_{\pi_{A}}(Y_{i1},Y_{i0},\mathbf{X}_{i};\alpha^{(A)})\\\
(1-A_{i})S_{\mu_{0}}(Y_{i1},Y_{i0},\mathbf{X}_{i};\beta^{(0)})\end{bmatrix}=0$
where
$\eta=(\theta(\delta),\tau,\alpha^{(D)},\sigma^{2},\beta^{(D)},\alpha^{(A)},\beta^{(0)})$
are estimable parameters; $S_{\pi_{D}}$, $S_{\mu_{1}}$, $S_{\pi_{A}}$, and
$S_{\mu_{0}}$ are score equations for parametric nuisance models; and
$\ell_{i}(\delta)$ are the local linear kernel regression estimation
weights.(Wasserman, 2006) We then use the sandwich variance formula as:
$\hat{V}(\hat{\eta})=\\{\sum\limits_{i=1}^{n}\frac{\partial}{\partial\eta^{\prime}}\Gamma_{i}(\hat{\eta})\\}^{-1}\\{\sum\limits_{i=1}^{n}\Gamma_{i}^{\bigotimes
2}(\hat{\eta})\\{\sum\limits_{i=1}^{n}\frac{\partial}{\partial\eta}\Gamma_{i}^{\prime}(\hat{\eta})\\}^{-1}$
Finally, we can calculate
$\hat{V}(\hat{\Psi}(\delta))=V(\hat{\theta}(\delta)-\hat{\theta}_{0})=\hat{V}_{1,1}(\hat{\eta})+\hat{V}_{2,2}(\hat{\eta})-2\hat{V}_{1,2}(\hat{\eta})$
and derive $95\%$ confidence intervals under a normal approximation for each
$\delta$. The calculated sandwich estimator under linear
($\pi_{D},\mu_{1,\Delta},\mu_{0,\Delta}$) and logistic ($\pi_{A}$) models is
provided in the Appendix.
Still, this sandwich estimator only applies for a restricted set of nuisance
function models, which may be inadequate for complex dynamics, and rely on
correct specification of all nuisance functions. Alternatively, bootstrapping
approaches can be used to express pointwise uncertainty about the smoothed
kernel estimate under certain regularity conditions like those required for
consistency of $r_{n}$ described in the Appendix.(Wasserman, 2006) Further,
bootstrap approaches have been shown to better account for finite sample
variability and model misspecification than parametric formulations of
asymptotic approximations for previous doubly robust estimators.(Funk et al.,
2011) Throughout this paper, we implement a non-parametric weighted bootstrap
where we sample unit-specific, rather than observation-specific, weights to
account for correlation between multiple observations of the same unit. After
sampling these weights from an $exponential(1)$ distribution, we then scale
the weights such that the sum of weights in each treatment group are equal to
their observed sample sizes and utilize these weights for each fitted model
and population mean required in the estimation process. Allowing for small but
non-zero weights, unlike a discrete bootstrap sampling approach, is
particularly helpful for stabilizing bootstrap sample estimates which
otherwise may only have data from a few exposure levels. We then estimate the
$2.5$ and $97.5$ percentiles among 500 bootstrap replicates as our interval
bounds.
### 2.5 Multiple Observation Times
For simplicity, we introduced methods with one observation before the
intervention ($t=0$) and one observation after ($t=1$). In practice, we often
have multiple observation times within these periods that we would like to
incorporate additional information from. Following the design of our data
example, we consider multiple post-intervention observation times,
$m=1,...,M$, where each of these observation times has a corresponding pre-
intervention observation time (e.g., corresponding months of the year before
and after a policy implementation).(Hettinger et al., 2023) Further, to
distinguish from studies involving related to staggered adoption and dynamic
treatment regimes, we consider a single intervention time for all treated
units where the exposure is constant for all times after the intervention.
Presuming our parallel trends assumptions now hold between each of the $m$
pairs of corresponding observation times, we can calculate $m$-time specific
effects as defined in the $M=1$ setting and average them over all $m$.
Alternative approaches for multiple observation times will depend on the study
and we refer the reader to Callaway and Sant’Anna (2021) for frameworks under
more complex designs. When estimating outcome-based nuisance functions under
multiple observation times, practitioners must consider efficiency and
robustness trade-offs when deciding between modeling each observation time
separately or incorporating time into a single model. For propensity score
models, we utilize a single model to mimic that interventions and exposures
are only assigned once, but time-specific weights and their interpretation
have been further explored by Stuart et al. (2014). Finally, both mentioned
sandwich and bootstrapping uncertainty approaches are readily extensible to
this setting. Sandwich estimators require stacking additional estimating
equations for each observation time into the formulation of
$\Gamma_{i}(\eta,\delta)$, while described bootstrap approaches, which re-
sample units, already account for within-unit correlation induced by multiple
observations.(Li and Li, 2019)
## 3 Simulation Studies
We conducted numerical studies to examine the finite-sample properties of our
proposed methods relative to alternative approaches under different forms of
model misspecification. Specifically, our data generating mechanism, which
roughly matches the structure of our data example in the following section,
simulated normally distributed covariates,
$\mathbf{X_{i}}=(X_{i1},X_{i2},X_{i3},X_{i4})^{T}\sim
N(\mathbf{0},\mathbf{I}_{4})$
a binary intervention,
$\displaystyle A_{i}\mid\mathbf{X_{i}}\sim
Bern(expit(-0.3+0.15X_{i1}+0.15X_{i2}-0.15X_{i3}+0.25X_{i4}))$ $\displaystyle
expit(z)=exp(z)/[1+exp(z)]$
a normally distributed exposure among treated units,
$D_{i}\mid\mathbf{X_{i}},A=1\sim
N(\mu=6+0.45X_{i1}+0.4X_{i2}-0.45X_{i3}+0.9X_{i4},\sigma^{2}=4)$
a normally distributed outcome at time $t=0$,
$Y_{i0}\mid\mathbf{X_{i}},A_{i}\sim
N(\mu=10+0.4X_{i1}-X_{i2}+0.4X_{i3}+0.3X_{i4}+2A_{i},\sigma^{2}=0.3^{2})$
and a normally distributed outcome at time $t=1$ where the expected trend is
determined by confounders, intervention group, and exposure level,
$\displaystyle Y_{i1}\mid\mathbf{X_{i}},A_{i},D_{i}\sim
N(\mu=Y_{i0}+A_{i}\lambda_{1}(\mathbf{X_{i}},D_{i})+(1-A_{i})\lambda_{0}(\mathbf{X_{i}}),\sigma^{2}=0.1^{2})$
$\displaystyle\lambda_{1}(\mathbf{X_{i}},D_{i})=1+0.2X_{i1}+0.2X_{i2}+0.3X_{i3}-0.1X_{i4}+2A_{i}+$
$\displaystyle\hskip 113.81102ptD(0.1-0.1X_{i1}+0.1X_{i3}-0.01/10D^{2})$
$\displaystyle\lambda_{0}(\mathbf{X_{i}})=-3-X_{i1}+0.7X_{i2}+0.6X_{i3}-0.6X_{i4}$
To analyze the simulated data, we compared our proposed, multiply robust
method (MR) to four alternative estimators. First, we estimated a standard
linear two-way fixed effects (TWFE) model,
$Y_{it}=\beta_{0}+\boldsymbol{\beta}^{\prime}\mathbf{X_{i}}+\gamma_{T}t+\gamma_{A}A_{i}+\gamma_{D}A_{i}D_{i}+\tau_{A}tA_{i}+\tau_{D}tA_{i}D_{i}+\epsilon_{it}$,
and calculated an effect curve as
$\hat{\Psi}^{(twfe)}(\delta)=\hat{\tau}_{A}+\hat{\tau}_{D}\delta$. This is an
analogous approach to the model studied by Callaway et al. (2021) that does
not adjust for confounding between intervention or exposure levels and outcome
trends. Second, we estimated a flexible, confounding-naive difference-in-
differences estimator as
$\hat{\Psi}^{(naive)}(\delta)=\hat{E}[Y_{1}-Y_{0}|D=\delta,A=1]-\hat{E}[Y_{1}-Y_{0}|A=0]$,
where the first expectation is estimated with a local linear kernel regression
and the second with a sample mean. Then, we considered outcome regression and
IPW analogies to our proposed estimator. The outcome regression approach
relies entirely on specification of $\mu_{1,\Delta}$ and $\mu_{0,\Delta}$ and
estimates
$\hat{\Psi}^{(or)}(\delta)=\hat{E}[\hat{\mu}_{1,\Delta}(\mathbf{X})-\hat{\mu}_{0,\Delta}(\mathbf{X})|A=1]$
with a sample mean over treated units. The IPW approach relies entirely on
specification of $\pi_{D}$ and $\pi_{A}$, estimating
$\hat{\theta}^{(ipw)}(\delta)=\hat{E}[\hat{w}^{(1)}(Y_{1}-Y_{0})|A=1,D=\delta]$
with a local linear kernel regression and
$\hat{\theta}^{(ipw)}_{0}=\hat{E}[\hat{w}^{(0)}(Y_{1}-Y_{0})|A=0]$ with a
sample mean over control units, to get
$\hat{\Psi}^{(ipw)}(\delta)=\hat{\theta}^{(ipw)}(\delta)-\hat{\theta}^{(ipw)}_{0}$.
This is a similar approach to that of Han and Yu (2019) but using the
estimated propensity scores for weights rather than covariate adjustment.
We consider nuisance function misspecification by developing models under the
specified mean functions, but using $\mathbf{X}$ for covariates in our model
under a correct specification and $\mathbf{W}$ for covariates under an
incorrect specification, where $\mathbf{W}$ is the same covariate
transformation as given in Kang and Schafer (2007). For each of the 16
permuations of nuisance model specification, we then generate 1000 simulated
datasets for each of three sample sizes: $n=200$, $1000$, and $10000$.
Potential estimators are compared using integrated absolute mean bias and root
mean squared error (RMSE) where summations are taken over the density of $D$
as estimated from a super-population with $n=1,000,000$. The $10\%$ of mass at
each boundary are excluded from evaluation as these values may not frequently
be observed in simulations. Finally, for $n=200$ datasets, we assess finite
sample inference properties by calculating the integrated coverage
probabilities and interval width of confidence intervals derived from the
described sandwich and bootstrap variance approaches for our proposed,
multiply robust estimator.
### 3.1 Results
Figure 2: An example fit for one simulated dataset ($n=1000$) showing the true effect curve and estimated effect curves using the multiply robust (MR), confounding-naive (Naive), and two-way fixed effects (TWFE) approaches. A constant difference between the MR and Naive approaches represents confounding between treated and control groups while the increasing difference over values of $\delta$ represents confounding between units at different exposure levels. Table 1: Comparison of Bias (RMSE) of effect curve estimates under multiply robust (MR), outcome regression (OR), and inverse probability weighting (IPW) approaches in simulation scenarios varying by nuisance function specification ($n=200$). Green cells represent scenarios where both sources of confounding are correctly adjusted for asymptotically, whereas orange and red cells represent scenarios where one and both sources are not correctly adjusted for, respectively. Confounding-naive and two-way fixed effects estimators have Bias (RMSE) of 0.329 (0.169) and 0.384 (0.184) in all scenarios. Incorrect Models | MR | OR | IPW
---|---|---|---
None | 0.008 (0.019) | 0.002 (0.019) | 0.015 (0.064)
$\pi_{A}$ | 0.008 (0.019) | 0.002 (0.019) | 0.117 (0.09)
$\mu_{0,\Delta}$ | 0.034 (0.03) | 0.089 (0.046) | 0.015 (0.064)
$\pi_{D}$ | 0.008 (0.019) | 0.002 (0.019) | 0.112 (0.08)
$\mu_{1,\Delta}$ | 0.028 (0.054) | 0.138 (0.056) | 0.015 (0.064)
$\pi_{A},\pi_{D}$ | 0.008 (0.019) | 0.002 (0.019) | 0.149 (0.112)
$\mu_{0,\Delta},\mu_{1,\Delta}$ | 0.033 (0.063) | 0.186 (0.095) | 0.015 (0.064)
$\mu_{0,\Delta},\pi_{D}$ | 0.034 (0.03) | 0.089 (0.046) | 0.112 (0.08)
$\pi_{A},\mu_{1,\Delta}$ | 0.028 (0.054) | 0.138 (0.056) | 0.117 (0.09)
$\pi_{A},\mu_{0,\Delta}$ | 0.132 (0.053) | 0.089 (0.046) | 0.117 (0.09)
$\pi_{D},\mu_{1,\Delta}$ | 0.119 (0.068) | 0.138 (0.056) | 0.112 (0.08)
$\pi_{A},\mu_{0,\Delta},\pi_{D}$ | 0.131 (0.053) | 0.089 (0.046) | 0.149 (0.112)
$\pi_{A},\mu_{0,\Delta},\mu_{1,\Delta}$ | 0.140 (0.088) | 0.186 (0.095) | 0.117 (0.09)
$\pi_{A},\pi_{D},\mu_{1,\Delta}$ | 0.119 (0.068) | 0.138 (0.056) | 0.149 (0.112)
$\mu_{0,\Delta},\pi_{D},\mu_{1,\Delta}$ | 0.116 (0.075) | 0.186 (0.095) | 0.112 (0.08)
$\pi_{A},\mu_{0,\Delta},\pi_{D},\mu_{1,\Delta}$ | 0.179 (0.11) | 0.186 (0.095) | 0.149 (0.112)
Table 2: A comparison of Coverage Probabilities (Interval Widths) for confidence intervals derived from sandwich and bootstrap variances for the proposed, multiply robust estimator in simulation scenarios varying by nuisance function specification ($n=200$). Incorrect Models | Sandwich | Bootstrap
---|---|---
None | 99.1 (0.77) | 94.4 (0.52)
$\pi_{A},\pi_{D}$ | 99.0 (0.77) | 94.5 (0.52)
$\mu_{0,\Delta},\mu_{1,\Delta}$ | 96.2 (0.93) | 94.3 (0.85)
$\pi_{A},\mu_{0,\Delta},\pi_{D},\mu_{1,\Delta}$ | 85.2 (0.90) | 84.4 (0.87)
Results summarizing the performance of causal effect curve point estimates are
given in Table 1. In all analyzed scenarios, $\hat{\Psi}^{(twfe)}(\delta)$ and
$\hat{\Psi}^{(naive)}(\delta)$ exhibit significant bias as they fail to
account for confounding between the treated and control groups and between the
different exposure levels within the treated group. When $\hat{\pi}_{A}$ or
$\hat{\pi}_{D}$ are misspecified ($75\%$ of scenarios),
$\hat{\Psi}^{(ipw)}(\delta)$ demonstrates bias, the severity of which depends
upon the specific models that are misspecified. Similarly,
$\hat{\Psi}^{(or)}(\delta)$ is biased whenever $\hat{\mu}_{0,\Delta}$ or
$\hat{\mu}_{1,\Delta}$ are misspecified ($75\%$ of scenarios). On the
contrary, our proposed estimator, $\hat{\Psi}^{(mr)}(\delta)$, exhibits bias
only when both $\hat{\pi}_{A}$ and $\hat{\mu}_{0,\Delta}$ or both
$\hat{\pi}_{D}$ and $\hat{\mu}_{1,\Delta}$ are misspecified ($44\%$ of
scenarios). Further, $\hat{\Psi}^{(mr)}(\delta)$ correctly accounts for at
least one of the two sources of confounding as long as one of the four
nuisance functions are well-specified.
Under correct nuisance function specification, $\hat{\Psi}^{(mr)}(\delta)$ and
$\hat{\Psi}^{(or)}(\delta)$ demonstrate the lowest RMSE and thus most
efficient estimates. The RMSE of $\hat{\Psi}^{(ipw)}(\delta)$, roughly 5x that
of $\hat{\Psi}^{(mr)}(\delta)$ and $\hat{\Psi}^{(or)}(\delta)$ in well-
specified scenarios, demonstrates the instability of approaches that rely
solely on weighting even under correct model specification. Still,
$\hat{\Psi}^{(mr)}(\delta)$ performance may be subject to similar instability
when outcome models are misspecified.
Bootstrapped confidence intervals for $\hat{\Psi}^{(mr)}$ generally achieve
the specified coverage probabilities in scenarios with consistent effect
estimates, suggesting multiply robust inference in these simulations.
Robustness of inference has not been proven theoretically for bootstrap
approaches but has been observed frequently in practice (Funk et al., 2011).
The sandwich variance estimator is conservative under correct nuisance
function specification, likely due to the sample size considered ($n=200$) and
the asymptotic reliance of similar approaches noted previously (Lunceford and
Davidian, 2004). Additionally, the sandwich variance estimator is not robust
to any nuisance function misspecification due to its fully parametric
specification and only achieves the specified coverage bounds when all models
are correctly specified.
## 4 Application to Nutritional Excise Taxes
### 4.1 Philadelphia Beverage Tax Study
We applied the proposed methodology to estimate the effects of an excise tax
on sugary and sweetened beverages implemented in January 2017 in Philadelphia,
Pennsylvania. Previous studies have largely used difference-in-differences
methods to estimate the average effect of the tax on volume sales of taxed
beverages in Philadelphia and non-taxed neighboring county stores (Roberto et
al., 2019; Hettinger et al., 2023; Cawley et al., 2019). Although these
studies assessed the impact of the tax as a binary intervention, they
identified two significant mechanisms contributing to heterogeneous exposure
to the policy. First, as the tax was imposed on the distributor, each
individual store had the discretion to determine the extent of corresponding
price adjustments. While prior commentaries have speculated that consumers
might be more responsive to the awareness of the tax than actual price
fluctuations, this hypothesis has yet to be rigorously investigated. Second,
compelling evidence suggests a pattern of cross-border shopping in that
neighboring non-taxed counties demonstrated a concurrent increase in sales
during the tax implementation period. Prior studies have established
associations between distance to the border and magnitude of tax effects, but
lacked a robust causal framework to effectively control for confounding
factors between border proximity and beverage sales (Cawley et al., 2019;
Hettinger et al., 2023). A comprehensive understanding of the heterogeneity in
tax effects related to both of these exposures is crucial for informing the
design of future policies.
In the analysis, we considered 140 pharmacies from Philadelphia as our treated
group and 123 pharmacies from Baltimore and non-neighboring Pennsylvania
counties as our control group. Data provided by Information Resources Inc.
(IRI) and described previously includes volume sales of taxed beverages
aggregated in each 4-week period in the year prior to (2016) and after (2017)
tax implementation ($m=1,...,13$) (Muth et al., 2016; Roberto et al., 2019).
To account for variations in store size, we scaled volume sales for each store
by their average 4-week sales in 2016. This scaling not only mitigates many
differences in outcome trends between stores in our study, but also
facilitates attempts to generalize effects to regions where stores may have a
different magnitude of beverage sales.
To quantify the first source of exposure heterogeneity, we assessed the
weighted price of taxed beverages per ounce. The measure for price change was
determined by calculating the difference between the average metric in the
first 4-week period of 2017 (post-tax price) and the average metric across all
of 2016 (pre-tax price). Here, we excluded prices after January 2017 from our
calculation as these may be influenced by post-tax sales and could incorrectly
associate our exposure measure with effect heterogeneity. For the second
source of exposure heterogeneity, we calculated, for each store, the
percentage of adjacent zip codes also subject to the tax as a measure of
protection from cross-border shopping. Unfortunately, the de-identification of
stores in our data besides zip code prevented the calculation of a more
detailed distance or protection metric. When evaluating effects according to
each exposure, we considered several confounding variables including the
alternative exposure measure (for models on the treated group), the pre-tax
price and sales at a given store, and both the percentage of white residents
and average income per household in the zip-code.
We used the proposed kernel smoothing approach to estimate the full effect
curve flexibly and estimated our nuisance functions using SuperLearner to
limit the potential for model misspecification (van der Laan et al., 2007).
Specifically, we included a comprehensive ensemble of generalized linear
models, generalized additive models, Lasso, and boosting algorithms as
candidate learners for the conditional mean functions of our binomial
($\hat{\pi}_{A}$) and gaussian
($\hat{\pi}_{D},\hat{\mu}_{0,\Delta},\hat{\mu}_{1,\Delta}$) exposures and
outcomes. To model the conditional density function of the continuous
exposure, we followed previous methodologies to model the squared residuals
from the mean model, $\hat{\epsilon}_{i}^{2}$, using SuperLearner under the
same set of learners and then estimate the density of
$(D_{i}-\hat{D}_{i})/\sqrt{\hat{\epsilon}_{i}^{2}}$ (Hirano and Imbens, 2004;
Kreif et al., 2015). To account for the point mass at $D=100\%$ in the second
exposure, we estimated the density at this point separately by treating the
point as a binomial dependent variable.
### 4.2 Analysis of Effect Curves
Our exposure measures exhibited substantial heterogeneity. $10\%$ of
Philadelphia stores actually decreased their price of taxed beverages
following the tax, whereas $59\%$ increased their price by more than the tax
amount ($\$0.015$/oz.). $40\%$ of Philadelphia stores were $100\%$ surrounded
by other Philadelphia zip codes and the IQR of the remaining $60\%$ of stores
was $55-92\%$ surrounded. All confounding variables were significantly
associated with the intervention, exposure, and outcome trends in generalized
linear models where all confounding variables were included in an additive
fashion.
We first used our methodology to assess for parallel trends in the pre-
treatment period as described in a previous section by testing known placebo
effects for the trends between $m=1$ and $m=2,...,13$ in 2016. In spirit, we
treat January 2016 as the “pre-intervention”, $t=0$, time period and the
subsequent 2016 periods individually as “post-intervention”, $t=1$, time
periods. In addition to only serving as a proxy for the untestable
counterfactual parallel trends assumptions, these placebo tests do not
directly assess conditional parallel trends in the pre-treatment period as
such tests would require estimation of particular parameters in parametric
models (limiting confounder adjustment) or conditional effects by $\mathbf{X}$
(an active and challenging area of research in its own right). Instead, these
placebo tests assess whether any confounding exists on the population level
after adjusting for potential confounders using our methodology. Previous work
has used this approach to assess conditional parallel trends in the binary
setting (Hettinger et al., 2023). In our setting, our methods generally
estimated placebo effects with magnitudes between $10-30\%$ those of non-
placebo effects presented in the subsequent section, suggesting observed
deviations from parallel trends would explain only a portion of our estimated
effects. The only trend exhibiting high risk for violations of parallel trends
in the pre-treatment period came between periods $m=1$ and $m=13$, which may
indicate consumers adjusting shopping behavior prior to the tax, a violation
of Assumption 1 rather than Assumption 4.
Figure 3: Causal effect curve estimates representing the effect of the tax
where all Philadelphia stores change their price by $x$ cents per ounce.
Estimates are calculated using the multiply robust (MR), inverse probability
weighted (IPW), outcome regression (OR), and confounder-naive (Naive)
approaches with pointwise confidence intervals derived for the multiply robust
estimator from bootstrapped variance estimates (shaded).
Figure 4: Causal effect curve estimates representing the effect of the tax
where all Philadelphia stores are $x\%$ surrounded by taxed zip codes.
Estimates are calculated using the multiply robust (MR), inverse probability
weighted (IPW), outcome regression (OR), and confounder-naive (Naive)
approaches with pointwise confidence intervals derived for the multiply robust
estimator from bootstrapped variance estimates (shaded).
Effect curves for both exposure measures under the described multiply robust,
IPW, outcome regression, and confounder-naive approaches are visualized with
bootstrapped confidence intervals for the multiply robust estimator in Figures
3 and 4. When assessing effect curves related to specific price changes, we
saw strong evidence that consumers are responding to specific price changes
even after adjusting for observed confounders, regardless of estimator. If all
stores were to actually decrease their price following the tax, we estimated
that the tax would actually increase sales, whereas if all stores increased
their price by $\$0.02$/oz., we estimated that the tax would decrease sales by
about $25\%$. While these estimates demonstrate substantial influences of
economic competition, we likely would require incorporation of some sort of
economic network to truly interpret such effects under counterfactual
scenarios like where all pharmacies had decreased their prices since many of
these additional sales come from Philadelphia consumers unlikely to change
shopping habits under such a scenario. Still, policymakers could utilize this
curve to determine the relative elasticity of consumer shopping habits
conditional on the observed alternative price points when designing future
policies.
Estimated curves for effects of the tax under different exposures to cross-
border shopping may assist policy evaluators in understanding effect
heterogeneity resulting from cross-border shopping in Philadelphia. Further,
policymakers may use the estimated curve to assess what would have happened
had all neighboring counties also been taxed,
$\hat{\Psi}^{(dr)}(\delta=100\%)$, compared to if only some had been taxed,
e.g., , $\hat{\Psi}^{(dr)}(\delta=80\%)$. This counterfactual scenario was
previously explored by Lee et al. (2023), wherein they estimated that the tax
would still have had a substantial impact under this implementation. Rather
than utilizing a cross-border shopping exposure measure among Philadelphia
stores, their analytical approach exploited assumptions about which sale
increases in non-taxed neighboring counties would be returned to Philadelphia
had neighboring counties also been taxed. In our analysis, we similarly
estimated a significant tax effect in this counterfactual scenario, suggesting
that the tax would have had a $13\%$ effect on store-level sales had all
Philadelphia stores been fully surrounded by taxed zip codes, which falls
between the estimated effects ignoring confounding ($9\%$) and the estimated
aggregate effect over the observed distribution of doses ($17\%$). Notably,
$\hat{\Psi}^{(ipw)}(\delta)$ is quite volatile due to large weights accruing
near exposure boundaries, supporting the need for outcome models to improve
estimate stability.
## 5 Discussion
In this paper, we introduced a novel approach for estimating causal effect
curves within the framework of difference-in-differences designs. These effect
curves enable researchers to explore questions such as, “What would be the
average effect of a policy where all units were exposed to the policy at the
$\delta$-level?”, while accounting for multiple sources of observed
confounding commonly encountered in policy evaluations. Crucially, our
approach offers a flexible and multiply robust mechanism for confounder
adjustment without requiring any parametric assumptions about the form of the
effect curve. Further, our approach can incorporate general machine learning
and non-parametric modeling techniques by either imposing a limited set of
restrictions or utilizing sample splitting techniques (Bonvini and Kennedy,
2022; Chang, 2020). This flexibility and robustness are particularly vital
given the additional modeling complexities in this setting with a continuous
exposure and two sources of confounding compared to the standard binary
exposure setting. Moreover, by harnessing information from both
treatment/exposure and outcome models through an influence function-based
estimator, our approach enhances the efficiency of existing approaches that
rely solely on the propensity score.
While designed for difference-in-differences studies like the application
presented, the fundamental concepts of our approach have broader applications.
For example, one can readily extend our methodology to accommodate alternative
study designs, such as the controlled interrupted time series, by modifying
the presented parallel trends assumptions to the new design and altering the
derived influence function according to the newly identified conditional
expectations (Lopez Bernal et al., 2018). Further, our specification of
potential outcomes under a two-dimensional exposure vector resembles the
exposure mapping framework introduced by Aronow and Samii (2017). In this
framework, our continuous exposure, $D$, is analogous to their exposure
mapping function, $h_{it}(\mathbf{A})$, which represents the amount of
exposure received through one’s neighbors. As a result, our approach holds
promise for estimation of effects specified under interference, where current
approaches are highly sensitive to confounding (Clarke, 2017; Butts, 2021). To
fully realize these applications, however, future work will need to address
uncertainty estimation under such dependencies between units.
## Acknowledgements
This material is based upon work supported by the National Science Foundation
under Grant No. 2149716.
## Available Code
Code and an example simulated dataset are provided on GitHub at
https://github.com/garyhettinger/DiD-interference.
## References
* Pollack Porter et al. [2018] Keshia M. Pollack Porter, Lainie Rutkow, and Emma E. McGinty. The Importance of Policy Change for Addressing Public Health Problems. _Public Health Reports_ , 133(1_suppl):9S–14S, 11 2018. ISSN 0033-3549. doi:10.1177/0033354918788880.
* Chaloupka et al. [2019] Frank J. Chaloupka, Lisa M. Powell, and Kenneth E. Warner. The Use of Excise Taxes to Reduce Tobacco, Alcohol, and Sugary Beverage Consumption. _Annual Review of Public Health_ , 40(1):187–201, 4 2019. ISSN 0163-7525. doi:10.1146/annurev-publhealth-040218-043816.
* Cintron et al. [2022] Dakota W. Cintron, Nancy E. Adler, Laura M. Gottlieb, Erin Hagan, May Lynn Tan, David Vlahov, Madellena Maria Glymour, and Ellicott C. Matthay. Heterogeneous treatment effects in social policy studies: An assessment of contemporary articles in the health and social sciences. _Annals of Epidemiology_ , 70:79–88, 6 2022. ISSN 10472797. doi:10.1016/j.annepidem.2022.04.009.
* Heckman et al. [1997] J. J. Heckman, H. Ichimura, and P. E. Todd. Matching As An Econometric Evaluation Estimator: Evidence from Evaluating a Job Training Programme. _The Review of Economic Studies_ , 64(4):605–654, 10 1997. ISSN 0034-6527. doi:10.2307/2971733.
* Abadie [2005] Alberto Abadie. Semiparametric Difference-in-Differences Estimators. _The Review of Economic Studies_ , 72(1):1–19, 1 2005. ISSN 1467-937X. doi:10.1111/0034-6527.00321.
* Callaway et al. [2021] Brantly Callaway, Andrew Goodman-Bacon, and Pedro H. C. Sant’Anna. Difference-in-Differences with a Continuous Treatment. _arXiv_ , 7 2021.
* Li and Li [2019] Fan Li and Fan Li. Double-Robust Estimation in Difference-in-Differences with an Application to Traffic Safety Evaluation. _Observational Studies_ , 5(1):1–23, 2019. ISSN 2767-3324. doi:10.1353/obs.2019.0009.
* Sant’Anna and Zhao [2020] Pedro H.C. Sant’Anna and Jun Zhao. Doubly robust difference-in-differences estimators. _Journal of Econometrics_ , 219(1):101–122, 11 2020. ISSN 03044076. doi:10.1016/j.jeconom.2020.06.003.
* Han and Yu [2019] Bing Han and Hao Yu. Causal difference-in-differences estimation for evaluating the impact of semi-continuous medical home scores on health care for children. _Health Services and Outcomes Research Methodology_ , 19(1):61–78, 3 2019. ISSN 1387-3741. doi:10.1007/s10742-018-00195-9.
* Hirano and Imbens [2004] Keisuke Hirano and Guido W. Imbens. The Propensity Score with Continuous Treatments. pages 73–84. 7 2004. doi:10.1002/0470090456.ch7.
* Austin [2018] Peter C. Austin. Assessing the performance of the generalized propensity score for estimating the effect of quantitative or continuous exposures on binary outcomes. _Statistics in Medicine_ , 37(11):1874–1894, 5 2018. ISSN 0277-6715. doi:10.1002/sim.7615.
* Díaz and van der Laan [2013] Iván Díaz and Mark J. van der Laan. Targeted Data Adaptive Estimation of the Causal Dose–Response Curve. _Journal of Causal Inference_ , 1(2):171–192, 12 2013. ISSN 2193-3677. doi:10.1515/jci-2012-0005.
* Kennedy et al. [2017] Edward H. Kennedy, Zongming Ma, Matthew D. McHugh, and Dylan S. Small. Non-parametric methods for doubly robust estimation of continuous treatment effects. _Journal of the Royal Statistical Society: Series B (Statistical Methodology)_ , 79(4):1229–1245, 9 2017. ISSN 1369-7412. doi:10.1111/rssb.12212.
* Bonvini and Kennedy [2022] Matteo Bonvini and Edward H. Kennedy. Fast convergence rates for dose-response estimation. _arXiv_ , 7 2022.
* Zeldow and Hatfield [2021] Bret Zeldow and Laura A. Hatfield. Confounding and regression adjustment in <scp>difference-in-differences</scp> studies. _Health Services Research_ , 56(5):932–941, 10 2021. ISSN 0017-9124. doi:10.1111/1475-6773.13666.
* de Chaisemartin and D’Haultfoeuille [2022] Clément de Chaisemartin and Xavier D’Haultfoeuille. Two-Way Fixed Effects and Differences-in-Differences with Heterogeneous Treatment Effects: A Survey. Technical report, National Bureau of Economic Research, Cambridge, MA, 1 2022.
* Muñoz and van der Laan [2012] Iván Díaz Muñoz and Mark van der Laan. Population Intervention Causal Effects Based on Stochastic Interventions. _Biometrics_ , 68(2):541–549, 6 2012. ISSN 0006-341X. doi:10.1111/j.1541-0420.2011.01685.x.
* Bilinski and Hatfield [2018] Alyssa Bilinski and Laura A. Hatfield. Nothing to see here? Non-inferiority approaches to parallel trends and other model assumptions. _arXiv_ , 5 2018.
* Kennedy [2022] Edward H. Kennedy. Semiparametric doubly robust targeted double machine learning: a review. _arXiv_ , 3 2022.
* Wasserman [2006] Larry Wasserman. _All of Nonparametric Statistics_. Springer New York, New York, NY, 2006. ISBN 978-0-387-25145-5. doi:10.1007/0-387-30623-4.
* Lunceford and Davidian [2004] Jared K. Lunceford and Marie Davidian. Stratification and weighting via the propensity score in estimation of causal treatment effects: a comparative study. _Statistics in Medicine_ , 23(19):2937–2960, 10 2004. ISSN 0277-6715. doi:10.1002/sim.1903.
* Stefanski and Boos [2002] Leonard A Stefanski and Dennis D Boos. The Calculus of M-Estimation. _The American Statistician_ , 56(1):29–38, 2 2002.
* Cole et al. [2023] Stephen R Cole, Jessie K Edwards, Alexander Breskin, Samuel Rosin, Paul N Zivich, Bonnie E Shook-Sa, and Michael G Hudgens. Illustration of 2 Fusion Designs and Estimators. _American Journal of Epidemiology_ , 192(3):467–474, 2 2023. ISSN 0002-9262. doi:10.1093/aje/kwac067.
* Funk et al. [2011] Michele Jonsson Funk, Daniel Westreich, Chris Wiesen, Til Stürmer, M. Alan Brookhart, and Marie Davidian. Doubly Robust Estimation of Causal Effects. _American Journal of Epidemiology_ , 173(7):761–767, 4 2011. ISSN 1476-6256. doi:10.1093/aje/kwq439.
* Hettinger et al. [2023] Gary Hettinger, Christina Roberto, Youjin Lee, and Nandita Mitra. Estimation of Policy-Relevant Causal Effects in the Presence of Interference with an Application to the Philadelphia Beverage Tax. _arXiv_ , 1 2023.
* Callaway and Sant’Anna [2021] Brantly Callaway and Pedro H.C. Sant’Anna. Difference-in-Differences with multiple time periods. _Journal of Econometrics_ , 225(2):200–230, 12 2021. ISSN 03044076. doi:10.1016/j.jeconom.2020.12.001.
* Stuart et al. [2014] Elizabeth A. Stuart, Haiden A. Huskamp, Kenneth Duckworth, Jeffrey Simmons, Zirui Song, Michael E. Chernew, and Colleen L. Barry. Using propensity scores in difference-in-differences models to estimate the effects of a policy change. _Health Services and Outcomes Research Methodology_ , 14(4):166–182, 12 2014. ISSN 1387-3741. doi:10.1007/s10742-014-0123-z.
* Kang and Schafer [2007] Joseph D. Y. Kang and Joseph L. Schafer. Demystifying Double Robustness: A Comparison of Alternative Strategies for Estimating a Population Mean from Incomplete Data. _Statistical Science_ , 22(4), 11 2007. ISSN 0883-4237. doi:10.1214/07-STS227.
* Roberto et al. [2019] Christina A. Roberto, Hannah G. Lawman, Michael T. LeVasseur, Nandita Mitra, Ana Peterhans, Bradley Herring, and Sara N. Bleich. Association of a Beverage Tax on Sugar-Sweetened and Artificially Sweetened Beverages With Changes in Beverage Prices and Sales at Chain Retailers in a Large Urban Setting. _JAMA_ , 321(18):1799, 5 2019. ISSN 0098-7484. doi:10.1001/jama.2019.4249.
* Cawley et al. [2019] John Cawley, David Frisvold, Anna Hill, and David Jones. The impact of the Philadelphia beverage tax on purchases and consumption by adults and children. _Journal of Health Economics_ , 67:102225, 9 2019. ISSN 01676296. doi:10.1016/j.jhealeco.2019.102225.
* Muth et al. [2016] MK Muth, M Sweitzer, D Brown, K Capogrossi, S Karns, D Levin, A Okrent, P Siegel, and C Zhen. Understanding IRI household-based and store-based scanner data, 4 2016.
* van der Laan et al. [2007] Mark J. van der Laan, Eric C Polley, and Alan E. Hubbard. Super Learner. _Statistical Applications in Genetics and Molecular Biology_ , 6(1), 1 2007. ISSN 1544-6115. doi:10.2202/1544-6115.1309.
* Kreif et al. [2015] Noémi Kreif, Richard Grieve, Iván Díaz, and David Harrison. Evaluation of the Effect of a Continuous Treatment: A Machine Learning Approach with an Application to Treatment for Traumatic Brain Injury. _Health Economics_ , 24(9):1213–1228, 9 2015. ISSN 1057-9230. doi:10.1002/hec.3189.
* Lee et al. [2023] Youjin Lee, Gary Hettinger, and Nandita Mitra. Policy effect evaluation under counterfactual neighborhood interventions in the presence of spillover. _arXiv_ , 3 2023.
* Chang [2020] Neng-Chieh Chang. Double/debiased machine learning for difference-in-differences models. _The Econometrics Journal_ , 23(2):177–191, 5 2020. ISSN 1368-4221. doi:10.1093/ectj/utaa001.
* Lopez Bernal et al. [2018] James Lopez Bernal, Steven Cummins, and Antonio Gasparrini. The use of controls in interrupted time series studies of public health interventions. _International Journal of Epidemiology_ , 47(6):2082–2093, 12 2018. ISSN 0300-5771. doi:10.1093/ije/dyy135.
* Aronow and Samii [2017] Peter M. Aronow and Cyrus Samii. Estimating average causal effects under general interference, with application to a social network experiment. _The Annals of Applied Statistics_ , 11(4), 12 2017. ISSN 1932-6157. doi:10.1214/16-AOAS1005.
* Clarke [2017] Damian Clarke. Estimating Difference-in-Differences in the Presence of Spillovers. _Munich Personal RePEc Archive_ , 2017.
* Butts [2021] Kyle Butts. Difference-in-Differences Estimation with Spatial Spillovers. _arXiv_ , 5 2021.
|
# A State-Distribution Matching Approach to Non-Episodic Reinforcement
Learning
Archit Sharma Rehaan Ahmad Chelsea Finn
###### Abstract
While reinforcement learning (RL) provides a framework for learning through
trial and error, translating RL algorithms into the real world has remained
challenging. A major hurdle to real-world application arises from the
development of algorithms in an episodic setting where the environment is
reset after every trial, in contrast with the continual and non-episodic
nature of the real-world encountered by embodied agents such as humans and
robots. Enabling agents to learn behaviors autonomously in such non-episodic
environments requires that the agent to be able to conduct its own trials.
Prior works have considered an alternating approach where a forward policy
learns to solve the task and the backward policy learns to reset the
environment, but what initial state distribution should the backward policy
reset the agent to? Assuming access to a few demonstrations, we propose a new
method, MEDAL, that trains the backward policy to match the state distribution
in the provided demonstrations. This keeps the agent close to the task-
relevant states, allowing for a mix of easy and difficult starting states for
the forward policy. Our experiments show that MEDAL matches or outperforms
prior methods on three sparse-reward continuous control tasks from the EARL
benchmark, with 40% gains on the hardest task, while making fewer assumptions
than prior works. Code and videos are at: https://sites.google.com/view/medal-
arl/home
reinforcement learning, autonomous, discriminator
## 1 Introduction
A cornerstone of human and animal intelligence is the ability to learn
autonomously through trial and error. To that extent, reinforcement learning
(RL) presents a natural framework to develop learning algorithms for embodied
agents. Unfortunately, the predominant emphasis on episodic learning
represents a departure from the continual non-episodic nature of the real-
world, which presents multiple technical challenges. First, episodic training
undermines the autonomy of the learning agent by requiring repeated extrinsic
interventions to reset the environment after every trial, which can be both
time-consuming and expensive as these interventions may have to be conducted
by a human. Second, episodic training from narrow initial state distributions
can lead to less robust policies that are reliant on environment resets to
recover; e.g. Sharma et al. (2022) show that policies learned in episodic
settings with narrow initial state distributions are more sensitive to
perturbations than those trained in non-episodic settings.
Figure 1: An overview of our proposed method MEDAL (right) contrasting it with
forward-backward RL (Han et al., 2015; Eysenbach et al., 2017) (left). MEDAL
trains a backward policy $\pi_{b}$ to pull the agent back to the state
distribution defined by the demonstrations, enabling the forward policy
$\pi_{f}$ to the learn the task efficiently in contrast to FBRL that retrieves
the agent to the initial state distribution before every trial of $\pi_{f}$.
Prior works have found that conventional RL algorithms substantially
depreciate in performance when applied in non-episodic settings (Co-Reyes et
al., 2020; Zhu et al., 2020a; Sharma et al., 2022). Why do such algorithms
struggle to learn in non-episodic, autonomous RL (ARL) settings? Resetting the
environment after every single episode allows for natural repetition: the
agent can repeatedly practice the task under a narrow set of initial
conditions to incrementally improve the policy. Critically, algorithms
developed for episodic learning do not have to learn how to reach these
initial conditions in the first place. Thus, the main additional challenge in
non-episodic, autonomous RL settings is to enable the repetitive practice that
is necessary to learn an adept policy. For example, an autonomous robot that
is practicing how to close a door will also need to learn how to open a door.
Several recent works learn a backward policy to enable the main forward policy
to practice the task: for example, Han et al. (2015); Eysenbach et al. (2017)
propose a backward policy that learns to match the initial state distribution.
However, unlike the episodic setting, the agent can practice the task from any
initial state, and not just the narrow initial state distribution that is
usually provided by resets. Can the backward policy create starting conditions
that enable the forward policy to improve efficiently? It could be useful for
the agent to try the task both from “easy” states that are close to the goal
and harder states that are representative of the starting conditions at
evaluation. Easier and harder initial conditions can be seen as a curriculum
that simplifies exploration. Kakade & Langford (2002) provide a theoretical
discussion on how the initial state distribution affects the performance of
the learned policy. One of the results show that the closer the starting state
distribution is to the state distribution of the optimal policy $\rho^{*}$,
the faster the policy moves toward the optimal policy $\pi^{*}$. While an
oracle access to $\rho^{*}$ is rarely available, we often have access to a
modest set of demonstrations. In this work, we aim to improve autonomous RL by
learning a backward policy that matches the starting state distribution to the
state distribution observed in the demonstrations. This enables the agent to
practice the task from a variety of initial states, including some that are
possibly easier to explore from. An intuitive representation of the algorithm
is shown in Figure 1.
The primary contribution of our work is an autonomous RL algorithm Matching
Expert Distributions for Autonomous Learning (MEDAL), which learns a backward
policy that matches the state distribution of a small set of demonstrations,
in conjunction with a forward policy that optimizes the task reward. We use a
classification based approach that implicitly minimizes the distance between
the state distribution of the backward policy and the state distribution in
the demonstrations without requiring the density under either distribution. In
Section 5, we empirically analyze the performance of MEDAL on the Environments
for Autonomous RL (EARL) benchmark (Sharma et al., 2022). We find that MEDAL
matches or outperforms competitive baselines in all of the sparse-reward
environments, with a more than a 40% gain in success rate on the hardest task
where all other comparisons fail completely. Our ablations additionally
indicate the importance of matching the state distribution in the
demonstrations, providing additional empirical support for the hypothesis that
the expert state distribution constitutes a good starting state distribution
for learning a task.
## 2 Related Work
Autonomous RL. Using additional policies to enable autonomous learning goes
back to the works of (Rivest & Schapire, 1993) in context of finite state
automaton, also referred to as “homing strategies” in (Even-Dar et al., 2005)
in context of POMDPs. More recently, in context of continuous control, several
works propose autonomous RL methods targeting different starting distributions
to learn from: Han et al. (2015); Eysenbach et al. (2017) match the initial
state distribution, Zhu et al. (2020a) leverage state-novelty (Burda et al.,
2018) to create new starting conditions for every trial, and Sharma et al.
(2021) create a curriculum of starting states based on the performance of the
forward policy to accelerate the learning. In addition, (Xu et al., 2020; Lu
et al., 2020) leverage ideas from unsupervised skill discovery (Gregor et al.,
2016; Eysenbach et al., 2018; Sharma et al., 2019; Hazan et al., 2019; Campos
et al., 2020), with the former using it to create an adversarial initial state
distribution and the latter to tackle non-episodic lifelong learning with a
non-stationary task-distribution. Our work proposes a novel algorithm MEDAL
that, unlike these prior works, opts to match the starting distribution to the
state distribution of demonstrations. Value-accelerated Persistent RL (VaPRL)
(Sharma et al., 2021) also considers the problem of autonomous RL with a few
initial demonstrations. Unlike VaPRL, our algorithm does not rely on
relabeling transitions with new goals (Andrychowicz et al., 2017), and thus
does not require access to the functional form of the reward function,
eliminating the need for additional hyperparameters that require task-specific
tuning. A simple and task-agnostic ARL method would accelerate the development
of autonomous robotic systems, the benefits of such autonomy being
demonstrated by several recent works (Chatzilygeroudis et al., 2018; Gupta et
al., 2021; Smith et al., 2021; Ha et al., 2020; Bloesch et al., 2022).
Distribution Matching in RL. Critical to our method is matching the state
distribution of the demonstrations. Such a distribution matching perspective
is often employed in inverse RL (Ng et al., 2000; Ziebart et al., 2008, 2010;
Finn et al., 2016) and imitation learning (Ghasemipour et al., 2020; Argall et
al., 2009) or to encourage efficient exploration (Lee et al., 2019). More
recently, several works have leveraged implicit distribution matching by
posing a classification problem, pioneered in Goodfellow et al. (2014), to
imitate demonstrations (Ho & Ermon, 2016; Baram et al., 2017; Kostrikov et
al., 2018; Rafailov et al., 2021), to imitate sequences of observations
(Torabi et al., 2019; Zhu et al., 2020b), or to learn reward functions for
goal-reaching (Fu et al., 2018; Singh et al., 2019). Our work employs a
similar discriminator-based approach to encourage the state distribution
induced by the policy to match that of the demonstrations. Importantly, our
work focuses on creating an initial state distribution that the forward policy
can learn efficiently from, as opposed to these prior works that are designed
for the episodic RL setting. As the experiments in Section 5.2 and Section 5.3
show, naïve extensions of these methods to non-episodic settings don’t fare
well.
Accelerating RL using Demonstrations. There is rich literature on using
demonstrations to speed up reinforcement learning, especially for sparse
reward problems. Prior works have considering shaping rewards using
demonstrations (Brys et al., 2015), pre-training the policy (Rajeswaran et
al., 2017), using behavior cloning loss as a regularizer for policy gradients
(Rajeswaran et al., 2017) and $Q$-learning (Nair et al., 2018), and
initializing the replay buffer (Nair et al., 2018; Vecerik et al., 2017;
Hester et al., 2018). MEDAL leverages demonstrations to accelerate non-
episodic reinforcement learning by utilizing demo distribution to create
initial conditions for the forward policy. The techniques proposed in these
prior works are complimentary to our proposal, and can be leveraged for non-
episodic RL in general as well. Indeed, for all methods in our experiments,
the replay buffer is initialized with demonstrations.
## 3 Preliminaries
Autonomous Reinforcement Learning. We use the ARL framework for non-episodic
learning defined in Sharma et al. (2022), which we briefly summarize here.
Consider a Markov decision process
${\mathcal{M}\equiv(\mathcal{S},\mathcal{A},p,r,\rho_{0})}$, where
$\mathcal{S}$ denotes the state space, $\mathcal{A}$ denotes the action space,
${p:\mathcal{S}\times\mathcal{A}\times\mathcal{S}\mapsto\mathbb{R}_{\geq 0}}$
denotes the transition dynamics,
${r:\mathcal{S}\times\mathcal{A}\mapsto\mathbb{R}}$ denotes the reward
function and $\rho_{0}$ denotes the initial state distribution. The learning
algorithm $\mathbb{A}$ is defined as
${\mathbb{A}:\\{s_{i},a_{i},s_{i+1},r_{i}\\}_{i=0}^{t}\mapsto\\{a_{t},\pi_{t}\\}}$,
which maps the transitions collected in the environment until time $t$ to an
action $a_{t}$ and its best guess at the optimal policy
${\pi_{t}:\mathcal{S}\times\mathcal{A}\mapsto\mathbb{R}_{\geq 0}}$. First, the
initial state is sampled exactly once ($s_{0}\sim\rho_{0}$) at the beginning
of the interaction and the learning algorithm interacts with the environment
through the actions $a_{t}$ till $t\to\infty$. This is the key distinction
from an episodic RL setting where the environment resets to a state from the
initial state distribution after a few steps. Second, the action taken in the
environment does not necessarily come from $\pi_{t}$, for example, a backward
policy $\pi_{b}$ may generate the action taken in the environment.
ARL defines two metrics: Continuing Policy Evaluation measures the reward
accumulated by $\mathbb{A}$ over the course of training, defined as
${\mathbb{C}(\mathbb{A})=\lim_{h\to\infty}\frac{1}{h}\mathbb{E}\left[\sum_{t=0}^{h}r(s_{t},a_{t})\right]}$
and Deployed Policy Evaluation metric measures how quickly an algorithm
improves the output policy $\pi_{t}$ at the task defined by the reward
function $r$, defined as:
$\mathbb{D}(\mathbb{A})=\sum_{t=0}^{\infty}J(\pi^{*})-J(\pi_{t}),$ (1)
where
${J(\pi)=\mathbb{E}\left[\sum_{j=0}^{\infty}\gamma^{j}r(s_{j},a_{j})\right],s_{0}\sim\rho_{0},a_{t}\sim\pi(\cdot\mid
s_{t})}$, ${s_{t+1}\sim p(\cdot\mid s_{t},a_{t})}$ and
${\pi^{*}\in\operatorname*{arg\,max}_{\pi}J(\pi)}$. The goal for an algorithm
$\mathbb{A}$ is to minimize $\mathbb{D}(\mathbb{A})$, that is to bring
$J(\pi_{t})$ close to $J(\pi^{*})$ in the least number of samples possible.
Intuitively, minimizing $\mathbb{D}(\mathbb{A})$ corresponds to maximizing the
area under the curve for $J(\pi_{t})$ versus $t$.
$\mathbb{C}(\mathbb{A})$ corresponds to the more conventional average-reward
reinforcement learning. While algorithms are able to accumulate large rewards
during training, they do not necessarily recover the optimal policy in non-
episodic settings (Zhu et al., 2020a; Co-Reyes et al., 2020; Sharma et al.,
2022). In response, Sharma et al. (2022) introduce $\mathbb{D}(\mathbb{A})$ to
explicitly encourage algorithms to learn task-solving behaviors and not just
accumulate reward through training.
Imitation Learning via Distribution Matching. Generative Adversarial Networks
(Goodfellow, 2016) pioneered implicit distribution matching for distributions
where likelihood cannot be computed explicitly. Given a dataset of samples
$\\{x_{i}\\}_{i=1}^{N}$, where $x_{i}\sim p^{*}(\cdot)$ for some target
distribution $p^{*}$ over the data space $\mathcal{X}$, generative
distribution $p_{\theta}(\cdot)$ can be learned through the following minimax
optimization:
$\min_{p_{\theta}}\max_{D}\mathbb{E}_{x\sim p^{*}}\left[\log
D(x)\right]+\mathbb{E}_{x\sim p_{\theta}}\left[\log(1-D(x))\right]$ (2)
where $D:\mathcal{X}\mapsto[0,1]$ is discriminator solving a binary
classification problem. This can be shown to minimize the Jensen-Shannon
divergence, that is $\mathcal{D}_{\textrm{JS}}(p_{\theta}\mid\mid p^{*})$
(Goodfellow et al., 2014; Nowozin et al., 2016) by observing that the Bayes-
optimal classifier satisfies
$D^{*}(x)=\frac{p^{*}(x)}{p^{*}(x)+p^{\theta}(x)}$ (assuming that prior
probability of true data and fake data is balanced). Because we do not require
an explicit density under the generative distribution and only require the
ability to sample the distribution, this allows construction of imitation
learning methods such as GAIL (Ho & Ermon, 2016) which minimizes
$\mathcal{D}_{\textrm{JS}}(\rho^{\pi}(s,a)\mid\mid\rho^{*}(s,a))$, where the
policy $\pi$ is rolled out in the environment starting from initial state
distribution $\rho_{0}$ to generate samples from the state-action distribution
$\rho^{\pi}(s,a)$ and $\rho^{*}(s,a)$ is the target state-action distribution
of the demonstrations.
## 4 Matching Expert Distributions for Autonomous Learning (MEDAL)
Several prior works demonstrate the ineffectiveness of standard RL methods in
non-episodic settings (Co-Reyes et al., 2020; Zhu et al., 2020a; Sharma et
al., 2022). Adding noise to actions, for example $\epsilon$-greedy in DQN
(Mnih et al., 2015) or Gaussian noise in SAC (Haarnoja et al., 2018)), can be
sufficient for exploration in episodic setting where every trial starts from a
narrow initial state distribution. However, such an approach becomes
ineffective in non-episodic settings because the same policy is expected to
both solve the task and be sufficiently exploratory. As a result, a common
solution in non-episodic autonomous RL settings is to learn another policy in
addition to the forward policy $\pi_{f}$ that solves the task (Han et al.,
2015; Eysenbach et al., 2017; Zhu et al., 2020a): a backward policy $\pi_{b}$
that targets a set of states to explore solving the task from. More precisely,
the forward policy $\pi_{f}$ learns to solve the task from a state sampled
from $\rho^{b}$, the marginal state distribution of $\pi^{b}$. An appropriate
$\rho^{b}$ can improve the efficiency of learning $\pi_{f}$ by creating an
effective initial state distribution for it. What should the $\pi_{b}$
optimize? We discuss this question in Section 4.1 and a practical way to
optimize the suggested objective in Section 4.2. An overview of our proposed
algorithm is given in Section 4.3.
Figure 2: Comparison of sampling initial states $s_{0}$ from the state
distribution of the optimal policy $\rho^{*}$, with sampling the initial state
from the default distribution $\rho_{0}$ in the episodic setting. The episodic
return is computed by initializing the agent at $s_{0}\sim\rho_{0}$ in both
the cases. The former improves both the sample efficiency and the performance
of the final policy.
### 4.1 Finding Better Starting States
In episodic settings, $\pi_{f}$ always starts exploring from $\rho_{0}$, which
is the same distribution from which it will be evaluated. A natural objective
for $\pi_{b}$ then is to minimize the distance between $\rho^{b}$ and
$\rho_{0}$. And indeed, prior works have proposed this approach (Han et al.,
2015; Eysenbach et al., 2017) by learning a backward controller to retrieve
the initial state distribution $\rho_{0}$. While the initial state
distribution cannot be changed in the episodic setting, $\pi_{b}$ does not
have any restriction to match $\rho_{0}$ in the autonomous RL setting. Is
there a better initial state distribution to efficiently learn $\pi_{f}$ from?
Interestingly, Kakade & Langford (2002) provide a theoretical discussion on
how the initial state distribution affects the performance. The main idea is
that learning an optimal policy often requires policy improvement at states
that are unlikely to be visited. Creating a more uniform starting state
distribution can accelerate policy improvement by encouraging policy
improvement at those unlikely states. The formal statement can be found in
(Kakade & Langford, 2002, Corollary 4.5). Informally, the result states that
the upper bound on the difference between the optimal performance and that of
policy $\pi$ is proportional to
$\lVert\frac{\rho^{*}(s)}{\mu}\rVert_{\infty}$, where $\rho^{*}$ is the state
distribution of the optimal policy and $\mu$ is the initial state
distribution. This suggests that an initial state distribution $\mu$ that is
close to the optimal state distribution $\rho^{*}$ would enable efficient
learning. Intuitively, some initial states in the optimal state distribution
would simplify the exploration by being closer to high reward states, which
can be bootstrapped upon to learn faster from the harder initial states. To
empirically verify the theoretical results, we compare the learning speed of
RL algorithm in the episodic setting on tabletop organization (environment
details in Section 5) when starting from (a) the standard initial state
distribution, that is $s_{0}\sim\rho_{0}$, versus (b) states sampled from the
stationary distribution of the optimal policy, that is $s_{0}\sim\rho^{*}(s)$.
We find in Figure 2 that the latter not only improves the learning speed, but
also improves the performance by nearly 18%.
### 4.2 Resetting to Match the Expert State Distribution
The theoretical and empirical results in the previous section suggest that
$\pi_{f}$ should attempt to solve the task from an initial state distribution
that is close to $\rho^{*}(s)$, thus implying that $\pi_{b}$ should try to
match $\rho^{*}(s)$. How do we match $\rho^{b}$ to $\rho^{*}$? We will assume
access to a small set of samples from $\rho^{*}(s)$ in the form of
demonstrations $\mathcal{D}_{f}$. Because we are limited to sampling
$\rho^{b}$ and only have a fixed set of samples from $\rho^{*}$, we consider
the following optimization problem:
$\min_{\pi_{b}}\max_{C}\mathbb{E}_{s\sim\rho^{*}}\big{[}\log
C(s)\big{]}+\mathbb{E}_{s\sim\rho^{b}}\big{[}\log(1-C(s))\big{]}$ (3)
where $C:\mathcal{S}\mapsto[0,1]$ is a state-space classifier. This
optimization is very much reminiscent of implicit distribution matching
techniques used in (Goodfellow et al., 2014; Nowozin et al., 2016; Ho & Ermon,
2016; Ghasemipour et al., 2020) when only the samples are available and
densities cannot be explicitly measured. This can be interpreted as minimizing
the Jensen-Shannon divergence
$\mathcal{D}_{\textrm{JS}}(\rho^{b}\mid\mid\rho^{*})$. Following these prior
works, $C(s)$ solves a binary classification where $s\sim\rho^{*}$ has a label
$1$ and $s\sim\rho^{b}$ has a label $0$. Further, $\pi_{b}$ solves a RL
problem to maximize ${\mathbb{E}_{s\sim\rho^{b}}[r(s,a)]}$, where the reward
function $r(s,a)=-\log(1-C(s))$. Assuming sufficiently expressive non-
parametric function classes, $(\rho^{*},0.5)$ is a saddle point for the
optimization in Equation 3.
Relationship to Prior Imitation Learning Methods. GAIL (Ho & Ermon, 2016)
proposes to match the state-action distribution $\rho^{\pi}(s,a)$ to that of
the expert $\rho^{*}(s,a)$, that is minimize
$\mathcal{D}_{\textrm{JS}}(\rho^{\pi}(s,a)\mid\mid\rho^{*}(s,a))$. Prior works
have considered the problem of imitation learning when state-only observations
are available (Torabi et al., 2019; Zhu et al., 2020b) by minimizing
$\mathcal{D}_{f}(\rho^{\pi}(s,s^{\prime})\mid\mid\rho^{*}(s,s^{\prime}))$,
where $f$-divergence enables generalized treatment of different discrepancy
measures such KL-divergence of JS-divergence used in prior work (Nowozin et
al., 2016). In contrast to these works, our work proposes to minimize
$\mathcal{D}_{\textrm{JS}}(\rho^{\pi}(s)\mid\mid\rho^{*}(s))$. Furthermore,
state distribution matching is only used for the backward policy in our
algorithm, whereas the forward policy is maximizing return, as we summarize in
the next section. Finally, unlike prior works, the motivation for matching the
state distributions is to create an effective initial state distribution for
the forward policy $\pi_{f}$. Our experimental results in Section 5.2 suggest
that naively extending GAIL to non-episodic settings is not effective,
validating the importance of these key differences.
Algorithm 1 Matching Expert Distributions for Autonomous Learning (MEDAL)
require: forward demos $\mathcal{D}_{f}$;
optional: backward demos $\mathcal{D}_{b}$;
initialize: $\mathcal{R}_{f},\pi_{f}(a\mid s),Q^{\pi_{f}}(s,a)$; // forward
policy
initialize: $\mathcal{R}_{b},\pi_{b}(a\mid s),Q^{\pi_{b}}(s,a)$; // backward
policy
initialize: $C(s)$; // state-space discriminator
$\mathcal{R}_{f}\leftarrow\mathcal{R}_{f}\cup\mathcal{D}_{f},\mathcal{R}_{b}\leftarrow\mathcal{R}_{b}\cup\mathcal{D}_{b}$;
$s\sim\rho_{0}$; // sample initial state
while not done do
// run forward policy for a fixed number of steps or until goal is reached,
otherwise run backward policy
if forward then
$a\sim\pi_{f}(\cdot\mid s)$;
$s^{\prime}\sim p(\cdot\mid s,a),r\leftarrow r(s,a)$;
$\mathcal{R}_{f}\leftarrow\mathcal{R}_{f}\cup\\{(s,a,s^{\prime},r)\\}$;
update $\pi_{f},Q^{\pi_{f}}$;
else
$a\sim\pi_{b}(\cdot\mid s)$;
$s^{\prime}\sim p(\cdot\mid s,a),r\leftarrow-\log(1-C(s^{\prime}))$;
$\mathcal{R}_{b}\leftarrow\mathcal{R}_{b}\cup\\{(s,a,s^{\prime},r)\\}$;
update $\pi_{b},Q^{\pi_{b}}$;
end if
// train disriminator every $K$ steps
if train-discriminator then
// sample a batch of positives $S_{p}$ from the forward demos
$\mathcal{D}_{f}$, and a batch of negatives $S_{n}$ from backward replay
buffer $\mathcal{R}_{b}$
$S_{p}\sim\mathcal{D}_{f},S_{n}\sim\mathcal{R}_{b}$;
update $C$ on $S_{p}\cup S_{n}$;
end if
$s\leftarrow s^{\prime}$;
end while
### 4.3 MEDAL Overview
With these components in place, we now summarize our proposed algorithm,
_Matching Expert Distributions for Autonomous Learning_ (MEDAL). We
simultaneously learn the following components: a _forward policy_ that learns
to solve the task and will also be used for evaluation, a backward policy that
learns creates the initial state distribution for the forward policy by
matching the state distribution in the demonstrations, and finally a state-
space discriminator that learns to distinguish between the states visited by
the backward policy and the states visited in the demonstrations. MEDAL
assumes access to a set of forward demonstrations $\mathcal{D}_{f}$,
completing the task from the initial state distribution, and optionally, a set
of backward demonstrations $\mathcal{D}_{b}$ undoing the task back to the
initial state distribution. The forward policy $\pi_{f}$ is trained to
maximize $\mathbb{E}[{\sum_{t=0}^{\infty}\gamma^{t}r(s_{t},a_{t})}]$ and the
replay buffer for the forward policy is initialized using $\mathcal{D}_{f}$.
The backward policy $\pi_{b}$ trains to minimize
${\mathcal{D}_{\textrm{JS}}(\rho^{b}(s)\mid\mid\rho^{*}(s))}$ which translates
into maximizing
${-\mathbb{E}[\sum_{t=0}^{\infty}\gamma^{t}\log(1-C(s_{t+1}))]}$ and the
replay buffer for the backward policy is initialized using the backward
demonstrations $\mathcal{D}_{b}$, if available. Finally, the state-space
discriminator $C(s)$ trains to classify states sampled from the forward
demonstrations $\mathcal{D}_{f}$ with label $1$ and states visited by
$\pi_{b}$ as label $0$. Note, we are trying to match the state marginal of
policy $\pi_{b}$ (i.e. $\rho^{b}(s)$) to the optimal state distribution
$\rho^{*}(s)$ (approximated via forward demonstrations $\mathcal{D}_{f}$, not
backward demonstrations), thereby motivating the classification problem for
$C(s)$.
When interacting with the environment during training, we alternate between
collecting samples using $\pi_{f}$ for a fixed number of steps and collecting
samples using $\pi_{b}$ for a fixed number of steps. The policies can be
updated using any RL algorithm. The state-space discriminator $C(s)$ is
updated every $K$ steps collected in the environment, with the states visited
by $\pi_{b}$ being labeled as $0$ and states in $\mathcal{D}_{f}$ labeled as
$1$. The minibatch for updating the parameters of $C(s)$ is balanced to ensure
equal samples from $\rho^{*}(s)$ and $\rho^{b}(s)$. The pseudocode for MEDAL
is provided in Algorithm 1, and further implementation details can be found in
Appendix A.
Figure 3: Continuous-control environments from the EARL benchmark: (_left_)
Table-top organization (TO) where a gripper is tasked with moving a mug to one
of the four goal locations, (_center_) sawyer door closing (SD) where the
sawyer robot is tasked with closing the door, (_right_) sawyer peg insertion
(SP) where the robot is tasked with picking up the peg and inserting into the
goal location.
## 5 Experiments
In this section, we empirically analyze the performance of MEDAL to answer to
following questions: (1) How does MEDAL compare to other non-episodic,
autonomous RL methods? (2) Given the demonstrations, can existing imitation
learning methods suffice? (3) How important is it for the backward controller
to match the entire state distribution, instead of just the initial state
distribution?
Environments. To analyze these questions, we consider three sparse-reward
continuous-control environments from the EARL benchmark (Sharma et al., 2022).
The _table-top organization_ is a simplified manipulation environment where a
gripper is tasked to move the mug to one of four coasters. The _sawyer door
closing_ environment requires a sawyer robot arm to learn how to close a door
starting from various initial positions. The challenge in the ARL setting
arises from the fact that the agent has to open the door to practice closing
it again. Finally, the _sawyer peg insertion_ environment requires the sawyer
robot arm to pick up a peg and insert it into a designated goal location. This
is a particularly challenging environment as the autonomously operating robot
can push the peg into places where it can be hard to retrieve it back, a
problem that is not encountered in the episodic setting as the environment is
reset to the initial state distribution every few hundred steps.
Evaluation. We follow the evaluation protocol laid down in the EARL benchmark.
All algorithms are reset to a state $s_{0}\sim\rho_{0}$ and interact with
their environments almost fully autonomously thereon, only being reset to an
initial state intermittently after several hundreds of thousands of steps of
interaction. Since our objective is to acquire task policies in a sample
efficient way, we will focus on deployed policy evaluation. Specifically, we
approximate
$J(\pi_{t})=\mathbb{E}[\sum_{t=0}^{\infty}\gamma^{t}r(s_{t},a_{t})]$ by
averaging the return of the policy over $10$ episodes starting from
$s_{0}\sim\rho_{0}$, every $10,000$ steps collected in the training
environment. Note, the trajectories collected for evaluation are not provided
to the learning algorithm $\mathbb{A}$. For all considered environments, the
reward functions are sparse in nature and correspondingly, EARL provides a
small set of demonstrations to the algorithms, that correspond to doing and
undoing the task (a total of $10$-$30$ demonstrations depending on the
environment). Environment specific details such as reward functions and
intermittent resets can be found in Appendix B.
### 5.1 Benchmarking MEDAL on EARL
Method | Tabletop | Sawyer | Sawyer
---|---|---|---
| Organization | Door | Peg
_naïve RL_ | 0.32 (0.17) | 0.00 (0.00) | 0.00 (0.00)
_FBRL_ | 0.94 (0.04) | 1.00 (0.00) | 0.00 (0.00)
_R3L_ | 0.96 (0.04) | 0.54 (0.18) | 0.00 (0.00)
_VaPRL_ | 0.98 (0.02) | 0.94 (0.05) | 0.00 (0.00)
_MEDAL_ | 0.98 (0.02) | 1.00 (0.00) | 0.40 (0.16)
_oracle RL_ | 0.80 (0.11) | 1.00 (0.00) | 1.00 (0.00)
Table 1: Average return of the final learned policy. Performance is averaged
over 5 random seeds. The mean and and the standard error are reported, with
the best performing entry in bold. For all domains, $1.0$ indicates the
maximum performance and $0.0$ indicates minimum performance.
First, we benchmark our proposed method MEDAL on the aforementioned EARL
environments against state-of-the-art non-episodic ARL methods.
Comparisons. We briefly review the methods benchmarked on EARL, which MEDAL
will be compared against: (1) forward-backward RL (FBRL) (Han et al., 2015;
Eysenbach et al., 2017), where the backward policy recovers the initial state
distribution; (2) value-accelerated persistent RL (VaPRL) (Sharma et al.,
2021), where the backward policy creates a curriculum based on the forward
policy’s performance; (3) R3L (Zhu et al., 2020a) has a backward policy that
optimizes a state-novelty reward (Burda et al., 2018) to encourage the forward
policy to solve the tasks from new states in every trial; (4) naïve RL
represents the episodic RL approach where only a forward policy optimizes the
task-reward throughout training; and finally (5) oracle RL is the same
episodic RL baseline but operating in the episodic setting. For a fair
comparison, the forward policy for all baselines use SAC (Haarnoja et al.,
2018), and the replay buffer is always initialized with the forward
demonstrations. Further, the replay buffers for backward policies in FBRL,
VaPRL is also initialized with the backward demos. The replay buffer of the
backward policy in R3L is not initialized with backward demos as it will
reduce the novelty of the states in the backward demos for the RND reward
without the backward policy ever visiting those states.
Figure 4: Performance of each method on (left) the table-top organization
environment, (center) the sawyer door closing environment, and (right) the
sawyer peg environment. Plots show learning curves with mean and standard
error over 5 random seeds.
It’s important to note that some of these comparisons make additional
assumptions compared to MEDAL:
* •
oracle RL operates in the episodic setting, that is the environment is reset
to a state from the initial state distribution every few hundred steps. The
baseline is included as a reference to compare performance of baselines in ARL
versus the conventional episodic setting. It also enables us to compare the
performance of conventional RL algorithms when moving from the episodic
setting to the ARL setting, by comparing the performance of oracle RL and
naïve RL.
* •
VaPRL relies on relabeling goals which requires the ability to query the
reward function for any arbitrary state and goal (as the VaPRL curriculum can
task the agent to reach arbitrary goals from the demonstrations).
Additionally, VaPRL has a task specific hyperparameter that controls how
quickly the curriculum moves towards the initial state distribution.
In a real-world settings, where the reward function often needs to be learned
as well (for example from images), these assumptions can be detrimental to
their practical application. While FBRL also requires an additional reward
function to reach the initial state distribution, the requirement is not as
steep. Additionally, we consider a version of FBRL that learns this reward
function in Section 5.3. However, the ability to query the reward function for
arbitrary states and goals, as is required by VaPRL, can be infeasible in
practice. The impact of these additional assumptions cannot be overstated, as
the primary motivation for the autonomous RL framework is to be representative
of real-world RL training.
Results. Table 1 shows the performance of the final forward policy, and Figure
4 shows the deployed performance of the forward policy versus the training
time for different methods. MEDAL consistently outputs the best performing
final policy, as can be seen in Table 1. Particularly notable is the
performance on sawyer peg insertion, where the final policy learned by MEDAL
gets 40% success rate on average, while all other methods fail completely.
With the exception of VaPRL on tabletop organization, MEDAL also learns more
efficiently compared to any of the prior methods. Notably, MEDAL substantially
reduces the sample efficiency gap between ARL methods and episodic methods on
sawyer door closing.
We posit two reasons for the success of MEDAL: (a) Learning a backward policy
that retrieves the agent close to the task distribution enables efficient
exploration, producing the speedup in performance. (b) Bringing the agent
closer to the state distribution implicit in the demonstrations may be easier
to maximize compared to other objectives, for example, retrieving the agent to
the initial state distribution.
Figure 5: MEDAL in comparison to imitation learning methods on tabletop
organization and sawyer door closing. Behavior cloning (BC) does not fare
well, suggesting the importance of online data collection. The success of
online imitation learning methods such as GAIL in episodic settings does not
translate to the non-episodic ARL setting, as indicated by the substantial
drop in performance of naïve GAIL compared to oracle GAIL.
### 5.2 Imitation Learning
Given that MEDAL assumes access to a set of demonstrations, a natural
alternative to consider is imitation learning. In this section, we focus our
experiments on the tabletop organization environment. We first test how a
behavior cloning fares (BC). Results in Figure 5 suggest that behavior cloning
does not do well on tabletop organization, completely failing to solve the
task and leaves substantial room for improvement on sawyer door. This is to be
expected as EARL provides only a small number of demonstrations, and errors
compounding over time from imperfect policies generally leads to poor
performance. How do imitation learning methods with online data collection
fare? We consider an off-policy version of GAIL, Discriminator Actor-Critic
(DAC) (Kostrikov et al., 2018), which matches $\rho^{\pi}(s,a)$ to
$\rho^{*}(s,a)$ with an implicit distribution matching approach similar to
ours. Assuming that $\rho^{\pi}(s,a)$ can match $\rho^{*}(s,a)$, the method
should in principle recover the optimal policy – there is nothing specific
about GAIL that restricts it to the episodic setting. However, as the results
in Figure 5 suggest, there is a substantial drop in performance when running
GAIL in episodic setting (oracle GAIL) versus the non-episodic ARL setting
(naïve GAIL). While such a distribution matching could succeed, naïvely
extending the methods to the ARL setting is not as successful, suggesting that
it may require an additional policy (similar to the backward policy) to be
more effective.
### 5.3 The Choice of State Distribution
The key element in MEDAL is matching the state distribution of the backward
policy to the states in the demonstrations. To isolate the role of our
proposed scheme of minimizing
$\mathcal{D}_{\textrm{JS}}(\rho^{b}\mid\mid\rho^{*})$, we compare it to an
alternate method that minimizes
$\mathcal{D}_{\textrm{JS}}(\rho^{b}\mid\mid\rho_{0})$, i.e., matching the
initial state distribution $\rho_{0}$ instead of $\rho^{*}$. This makes
exactly one change to MEDAL: instead of sampling positives for the
discriminator $C(s)$ from forward demonstrations $\mathcal{D}_{f}$, the
positives are sampled from $\rho_{0}$. Interestingly, this also provides a
practically realizable implementation of FBRL, as it removes the requirement
of the additional reward function required for the learning a backward policy
to reach the initial state distribution. We call this method FBRL + VICE as
VICE (Singh et al., 2019) enables learning a goal reaching reward function
using a few samples of the goal distribution, in this case the goal
distribution for $\pi_{b}$ being $\rho_{0}$. As can be seen in Figure 6, the
FBRL + VICE learns slower than MEDAL, highlighting the importance of matching
the entire state distribution as done in MEDAL.
Figure 6: Isolating the effect of matching demonstration data. The speed up of
MEDAL compared to FBRL + VICE, which matches the initial state distribution,
suggests that the performance gains of MEDAL can be attributed to the better
initial state distribution created by the backward controller.
## 6 Conclusion
We propose MEDAL, an autonomous RL algorithm that learns a backward policy to
match the expert state distribution using an implicit distribution matching
approach. Our empirical analysis indicates that this approach creates an
effective initial state distribution for the forward policy, improving both
the performance and the efficiency. The simplicity of MEDAL also makes it more
amenable for the real-world, not requiring access to additional reward
functions.
MEDAL assumes access to a (small) set of demonstrations, which may not be
feasible in several real-world scenarios. Identifying good initial state
distributions without relying on a set of demonstrations would increase the
applicability of MEDAL. Similarly, in applications where safe exploration is a
requirement, MEDAL can be adapted to constrain the forward policy such that it
stays close to the task-distribution defined by the demonstrations. While
MEDAL pushes further the improvements in ARL, as exemplified by the reduction
of sample efficiency gap on sawyer door closing results, there is still a
substantial gap in performance between ARL methods and oracle RL on sawyer
peg, motivating the search for better methods.
## References
* Andrychowicz et al. (2017) Andrychowicz, M., Wolski, F., Ray, A., Schneider, J., Fong, R., Welinder, P., McGrew, B., Tobin, J., Abbeel, P., and Zaremba, W. Hindsight experience replay. _arXiv preprint arXiv:1707.01495_ , 2017.
* Argall et al. (2009) Argall, B. D., Chernova, S., Veloso, M., and Browning, B. A survey of robot learning from demonstration. _Robotics and autonomous systems_ , 57(5):469–483, 2009.
* Baram et al. (2017) Baram, N., Anschel, O., Caspi, I., and Mannor, S. End-to-end differentiable adversarial imitation learning. In _International Conference on Machine Learning_ , pp. 390–399. PMLR, 2017.
* Bloesch et al. (2022) Bloesch, M., Humplik, J., Patraucean, V., Hafner, R., Haarnoja, T., Byravan, A., Siegel, N. Y., Tunyasuvunakool, S., Casarini, F., Batchelor, N., et al. Towards real robot learning in the wild: A case study in bipedal locomotion. In _Conference on Robot Learning_ , pp. 1502–1511. PMLR, 2022.
* Brys et al. (2015) Brys, T., Harutyunyan, A., Suay, H. B., Chernova, S., Taylor, M. E., and Nowé, A. Reinforcement learning from demonstration through shaping. In _Twenty-fourth international joint conference on artificial intelligence_ , 2015.
* Burda et al. (2018) Burda, Y., Edwards, H., Storkey, A., and Klimov, O. Exploration by random network distillation. _arXiv preprint arXiv:1810.12894_ , 2018.
* Campos et al. (2020) Campos, V., Trott, A., Xiong, C., Socher, R., Giró-i Nieto, X., and Torres, J. Explore, discover and learn: Unsupervised discovery of state-covering skills. In _International Conference on Machine Learning_ , pp. 1317–1327. PMLR, 2020.
* Chatzilygeroudis et al. (2018) Chatzilygeroudis, K., Vassiliades, V., and Mouret, J.-B. Reset-free trial-and-error learning for robot damage recovery. _Robotics and Autonomous Systems_ , 100:236–250, 2018.
* Co-Reyes et al. (2020) Co-Reyes, J. D., Sanjeev, S., Berseth, G., Gupta, A., and Levine, S. Ecological reinforcement learning. _arXiv preprint arXiv:2006.12478_ , 2020.
* Even-Dar et al. (2005) Even-Dar, E., Kakade, S. M., and Mansour, Y. Reinforcement learning in pomdps without resets. 2005\.
* Eysenbach et al. (2017) Eysenbach, B., Gu, S., Ibarz, J., and Levine, S. Leave no trace: Learning to reset for safe and autonomous reinforcement learning. _arXiv preprint arXiv:1711.06782_ , 2017.
* Eysenbach et al. (2018) Eysenbach, B., Gupta, A., Ibarz, J., and Levine, S. Diversity is all you need: Learning skills without a reward function. _arXiv preprint arXiv:1802.06070_ , 2018.
* Finn et al. (2016) Finn, C., Levine, S., and Abbeel, P. Guided cost learning: Deep inverse optimal control via policy optimization. In _International conference on machine learning_ , pp. 49–58. PMLR, 2016.
* Fu et al. (2018) Fu, J., Singh, A., Ghosh, D., Yang, L., and Levine, S. Variational inverse control with events: A general framework for data-driven reward definition. _arXiv preprint arXiv:1805.11686_ , 2018.
* Ghasemipour et al. (2020) Ghasemipour, S. K. S., Zemel, R., and Gu, S. A divergence minimization perspective on imitation learning methods. In _Conference on Robot Learning_ , pp. 1259–1277. PMLR, 2020.
* Goodfellow (2016) Goodfellow, I. Nips 2016 tutorial: Generative adversarial networks. _arXiv preprint arXiv:1701.00160_ , 2016.
* Goodfellow et al. (2014) Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. Generative adversarial nets. _Advances in neural information processing systems_ , 27, 2014.
* Gregor et al. (2016) Gregor, K., Rezende, D. J., and Wierstra, D. Variational intrinsic control. _arXiv preprint arXiv:1611.07507_ , 2016.
* Gupta et al. (2021) Gupta, A., Yu, J., Zhao, T. Z., Kumar, V., Rovinsky, A., Xu, K., Devlin, T., and Levine, S. Reset-free reinforcement learning via multi-task learning: Learning dexterous manipulation behaviors without human intervention. _arXiv preprint arXiv:2104.11203_ , 2021.
* Ha et al. (2020) Ha, S., Xu, P., Tan, Z., Levine, S., and Tan, J. Learning to walk in the real world with minimal human effort. _arXiv preprint arXiv:2002.08550_ , 2020.
* Haarnoja et al. (2018) Haarnoja, T., Zhou, A., Abbeel, P., and Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In _International conference on machine learning_ , pp. 1861–1870. PMLR, 2018.
* Han et al. (2015) Han, W., Levine, S., and Abbeel, P. Learning compound multi-step controllers under unknown dynamics. In _2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_ , pp. 6435–6442. IEEE, 2015.
* Hazan et al. (2019) Hazan, E., Kakade, S., Singh, K., and Van Soest, A. Provably efficient maximum entropy exploration. In _International Conference on Machine Learning_ , pp. 2681–2691. PMLR, 2019.
* Hester et al. (2018) Hester, T., Vecerik, M., Pietquin, O., Lanctot, M., Schaul, T., Piot, B., Horgan, D., Quan, J., Sendonaris, A., Osband, I., et al. Deep q-learning from demonstrations. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 32, 2018.
* Ho & Ermon (2016) Ho, J. and Ermon, S. Generative adversarial imitation learning. _Advances in neural information processing systems_ , 29:4565–4573, 2016.
* Kakade & Langford (2002) Kakade, S. and Langford, J. Approximately optimal approximate reinforcement learning. In _In Proc. 19th International Conference on Machine Learning_. Citeseer, 2002.
* Kostrikov et al. (2018) Kostrikov, I., Agrawal, K. K., Dwibedi, D., Levine, S., and Tompson, J. Discriminator-actor-critic: Addressing sample inefficiency and reward bias in adversarial imitation learning. _arXiv preprint arXiv:1809.02925_ , 2018.
* Lee et al. (2019) Lee, L., Eysenbach, B., Parisotto, E., Xing, E., Levine, S., and Salakhutdinov, R. Efficient exploration via state marginal matching. _arXiv preprint arXiv:1906.05274_ , 2019.
* Lu et al. (2020) Lu, K., Grover, A., Abbeel, P., and Mordatch, I. Reset-free lifelong learning with skill-space planning. _arXiv preprint arXiv:2012.03548_ , 2020.
* Mnih et al. (2015) Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., Graves, A., Riedmiller, M., Fidjeland, A. K., Ostrovski, G., et al. Human-level control through deep reinforcement learning. _nature_ , 518(7540):529–533, 2015.
* Nair et al. (2018) Nair, A., McGrew, B., Andrychowicz, M., Zaremba, W., and Abbeel, P. Overcoming exploration in reinforcement learning with demonstrations. In _2018 IEEE international conference on robotics and automation (ICRA)_ , pp. 6292–6299. IEEE, 2018.
* Ng et al. (2000) Ng, A. Y., Russell, S. J., et al. Algorithms for inverse reinforcement learning. In _Icml_ , volume 1, pp. 2, 2000.
* Nowozin et al. (2016) Nowozin, S., Cseke, B., and Tomioka, R. f-gan: Training generative neural samplers using variational divergence minimization. In _Proceedings of the 30th International Conference on Neural Information Processing Systems_ , pp. 271–279, 2016.
* Rafailov et al. (2021) Rafailov, R., Yu, T., Rajeswaran, A., and Finn, C. Visual adversarial imitation learning using variational models. _Advances in Neural Information Processing Systems_ , 34, 2021.
* Rajeswaran et al. (2017) Rajeswaran, A., Kumar, V., Gupta, A., Vezzani, G., Schulman, J., Todorov, E., and Levine, S. Learning complex dexterous manipulation with deep reinforcement learning and demonstrations. _arXiv preprint arXiv:1709.10087_ , 2017.
* Rivest & Schapire (1993) Rivest, R. L. and Schapire, R. E. Inference of finite automata using homing sequences. _Information and Computation_ , 103(2):299–347, 1993.
* Sharma et al. (2019) Sharma, A., Gu, S., Levine, S., Kumar, V., and Hausman, K. Dynamics-aware unsupervised discovery of skills. _arXiv preprint arXiv:1907.01657_ , 2019.
* Sharma et al. (2021) Sharma, A., Gupta, A., Levine, S., Hausman, K., and Finn, C. Autonomous reinforcement learning via subgoal curricula. _Advances in Neural Information Processing Systems_ , 34, 2021.
* Sharma et al. (2022) Sharma, A., Xu, K., Sardana, N., Gupta, A., Hausman, K., Levine, S., and Finn, C. Autonomous reinforcement learning: Formalism and benchmarking. _International Conference on Learning Representations_ , 2022.
* Singh et al. (2019) Singh, A., Yang, L., Hartikainen, K., Finn, C., and Levine, S. End-to-end robotic reinforcement learning without reward engineering. _arXiv preprint arXiv:1904.07854_ , 2019.
* Smith et al. (2021) Smith, L., Kew, J. C., Peng, X. B., Ha, S., Tan, J., and Levine, S. Legged robots that keep on learning: Fine-tuning locomotion policies in the real world. _arXiv preprint arXiv:2110.05457_ , 2021.
* Torabi et al. (2019) Torabi, F., Warnell, G., and Stone, P. Adversarial imitation learning from state-only demonstrations. In _Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems_ , pp. 2229–2231, 2019.
* Vecerik et al. (2017) Vecerik, M., Hester, T., Scholz, J., Wang, F., Pietquin, O., Piot, B., Heess, N., Rothörl, T., Lampe, T., and Riedmiller, M. Leveraging demonstrations for deep reinforcement learning on robotics problems with sparse rewards. _arXiv preprint arXiv:1707.08817_ , 2017.
* Xu et al. (2020) Xu, K., Verma, S., Finn, C., and Levine, S. Continual learning of control primitives: Skill discovery via reset-games. _arXiv preprint arXiv:2011.05286_ , 2020.
* Zhang et al. (2017) Zhang, H., Cisse, M., Dauphin, Y. N., and Lopez-Paz, D. mixup: Beyond empirical risk minimization. _arXiv preprint arXiv:1710.09412_ , 2017.
* Zhu et al. (2020a) Zhu, H., Yu, J., Gupta, A., Shah, D., Hartikainen, K., Singh, A., Kumar, V., and Levine, S. The ingredients of real world robotic reinforcement learning. In _International Conference on Learning Representations_ , 2020a.
* Zhu et al. (2020b) Zhu, Z., Lin, K., Dai, B., and Zhou, J. Off-policy imitation learning from observations. In _the Thirty-fourth Annual Conference on Neural Information Processing Systems (NeurIPS 2020)_ , 2020b.
* Ziebart et al. (2008) Ziebart, B. D., Maas, A. L., Bagnell, J. A., Dey, A. K., et al. Maximum entropy inverse reinforcement learning. In _Aaai_ , volume 8, pp. 1433–1438. Chicago, IL, USA, 2008.
* Ziebart et al. (2010) Ziebart, B. D., Bagnell, J. A., and Dey, A. K. Modeling interaction via the principle of maximum causal entropy. In _ICML_ , 2010.
## Appendix A MEDAL Implementation
MEDAL is implemented with TF-Agents, built on SAC as the base RL algorithm.
Hyperparameters follow the default values: initial collect steps: 10,000,
batch size sampled from replay buffer for updating policy and critic: 256,
steps collected per iteration: 1, trained per iteration: 1, discount factor:
0.99, learning rate: $3e-4$ (for critics, actors, and discriminator). The
actor and critic network were parameterized as neural networks with two hidden
layers each of size $256$. For the discriminator, it was parameterized as a
neural network with one hidden layer of size $128$. This discriminator is
updated once every 10 collection steps for all environments. Due to a small
positive dataset, mixup (Zhang et al., 2017) is used as a regularization
technique on the discriminator for all environments. Additionally, the batch
size for the discriminator is set to $800$ for all environments as this
significantly larger value was found to stabilize training.
Another choice that improved the stability was the choice of reward function
for the backward controller: both $r(s,a)=-\log(1-C(s))$ and
$r(s,a)=\log(C(s))$ preserve the saddle point $(\rho^{*},0.5)$ for the
optimization in Equation 3. However, as can be seen in Figure 7,
$r(s,a)=-\log(1-C(s))$ leads to both better and stable performance. We
hypothesize that this is due to smaller gradients of the $-\log(1-C(s))$ when
$C(s)\leq 0.5$, which is where the discriminator is expected to be for most of
the training as the discriminator can easily distinguish between expert states
and those of the backward policy to begin with.
Figure 7: Performance comparison of $r(s,a)=\log(C(s))$ and
$r(s,a)=-\log(1-C(s))$ on (left) the table-top organization environment,
(center) the sawyer door closing environment, and (right) the sawyer peg
environment. Plots show learning curves with mean and standard error over 5
random seeds.
## Appendix B Environments
The environment details can be found in (Sharma et al., 2022). We briefly
describe environments for completeness. For every environment, $H_{T}$ defines
the number of steps after which the environment is reset to a state
$s_{0}\sim\rho_{0}$, and $H_{E}$ defines the evaluation horizon over which the
return is computed for deployed policy evaluation:
table-top organization: Table-top organization is run with a training horizon
of $H_{T}=200,000$ and $H_{E}=200$. The sparse reward function is given by:
$r(s,g)=\mathbb{I}(\lVert s-g\rVert_{2}\leq 0.2),$
where $\mathbb{I}$ denotes the indicator function. The environment has $4$
possible goal locations for the mug, and goal location for the gripper is in
the center. EARL provides a total of $12$ forward demonstrations and $12$
backward demonstrations ($3$ per goal).
sawyer door closing: Sawyer door closing is run with a training horizon of
$H_{T}=200,000$ and an episode horizon of $H_{E}=300$. The sparse reward
function is:
$r(s,g)=\mathbb{I}(\lVert s-g\rVert_{2}\leq 0.02),$
where $\mathbb{I}$ again denotes the indicator function. The goal for the door
and the robot arm is the closed door position. EARL provides $5$ forward
demonstrations and $5$ backward demonstrations.
sawyer peg: Sawyer peg is run with a training horizon of $H_{T}=100,000$ and
an episode horizon of $H_{E}=200$. The sparse reward function is:
$r(s,g)=\mathbb{I}(\lVert s-g\rVert_{2}\leq 0.05),$
where $\mathbb{I}$ again denotes the indicator function. The goal for the peg
is to be placed in the goal slot. EARL provides $10$ forward demonstrations
and $20$ backward demonstrations.
|
# Cross-Entropy Method Variants for Optimization ††thanks: Code available at
https://github.com/mossr/CrossEntropyVariants.jl
Robert J. Moss Stanford University, Computer Science
Stanford, CA, 94305
<EMAIL_ADDRESS>
###### Abstract
The cross-entropy (CE) method is a popular stochastic method for optimization
due to its simplicity and effectiveness. Designed for rare-event simulations
where the probability of a target event occurring is relatively small, the CE-
method relies on enough objective function calls to accurately estimate the
optimal parameters of the underlying distribution. Certain objective functions
may be computationally expensive to evaluate, and the CE-method could
potentially get stuck in local minima. This is compounded with the need to
have an initial covariance wide enough to cover the design space of interest.
We introduce novel variants of the CE-method to address these concerns. To
mitigate expensive function calls, during optimization we use every sample to
build a surrogate model to approximate the objective function. The surrogate
model augments the belief of the objective function with less expensive
evaluations. We use a Gaussian process for our surrogate model to incorporate
uncertainty in the predictions which is especially helpful when dealing with
sparse data. To address local minima convergence, we use Gaussian mixture
models to encourage exploration of the design space. We experiment with
evaluation scheduling techniques to reallocate true objective function calls
earlier in the optimization when the covariance is the largest. To test our
approach, we created a parameterized test objective function with many local
minima and a single global minimum. Our test function can be adjusted to
control the spread and distinction of the minima. Experiments were run to
stress the cross-entropy method variants and results indicate that the
surrogate model-based approach reduces local minima convergence using the same
number of function evaluations.
## I Introduction
The cross-entropy (CE) method is a probabilistic optimization approach that
attempts to iteratively fit a distribution to elite samples from an initial
input distribution [1, 2]. The goal is to estimate a rare-event probability by
minimizing the cross-entropy between the two distributions [3]. The CE-method
has gained popularity in part due to its simplicity in implementation and
straightforward derivation. The technique uses importance sampling which
introduces a proposal distribution over the rare-events to sample from then
re-weights the posterior likelihood by the likelihood ratio of the true
distribution over the proposal distribution.
There are a few key assumptions that make the CE-method work effectively.
Through random sampling, the CE-method assumes that there are enough objective
function evaluations to accurately represent the objective. This may not be a
problem for simple applications, but can be an issue for computationally
expensive objective functions. Another assumption is that the initial
parameters of the input distribution are wide enough to cover the design space
of interest. For the case with a multivariate Gaussian distribution, this
corresponds to an appropriate mean and wide covariance. In rare-event
simulations with many local minima, the CE-method can fail to find a global
minima especially with sparse objective function evaluations.
This work aims to address the key assumptions of the CE-method. We introduce
variants of the CE-method that use surrogate modeling to approximate the
objective function, thus updating the belief of the underlying objective
through estimation. As part of this approach, we introduce evaluation
scheduling techniques to reallocate true objective function calls earlier in
the optimization when we know the covariance will be large. The evaluation
schedules can be based on a distribution (e.g., the Geometric distribution) or
can be prescribed manually depending on the problem. We also use a Gaussian
mixture model representation of the prior distribution as a method to explore
competing local optima. While the use of Gaussian mixture models in the CE-
method is not novel, we connect the use of mixture models and surrogate
modeling in the CE-method. This connection uses each elite sample as the mean
of a component distribution in the mixture, optimized through a subroutine
call to the standard CE-method using the learned surrogate model. To test our
approach, we introduce a parameterized test objective function called sierra.
The sierra function is built from a multivariate Gaussian mixture model with
many local minima and a single global minimum. Parameters for the sierra
function allow control over both the spread and distinction of the minima.
Lastly, we provide an analysis of the weak areas of the CE-method compared to
our proposed variants.
## II Related Work
The cross-entropy method is popular in the fields of operations research,
machine learning, and optimization [4, 5]. The combination of the cross-
entropy method, surrogate modeling, and mixture models has been explored in
other work [6]. The work in [6] proposed an adaptive grid approach to
accelerate Gaussian-process-based surrogate modeling using mixture models as
the prior in the cross-entropy method. They showed that a mixture model
performs better than a single Gaussian when the objective function is
multimodal. Our work differs in that we augment the “elite” samples both by an
approximate surrogate model and by a subroutine call to the CE-method using
the learned surrogate model. Other related work use Gaussian processes and a
modified cross-entropy method for receding-horizon trajectory optimization
[7]. Their cross-entropy method variant also incorporates the notion of
exploration in the context of path finding applications. An approach based on
relative entropy, described in section III-A, proposed a model-based
stochastic search that seeks to minimize the relative entropy [8]. They also
explore the use of a simple quadratic surrogate model to approximate the
objective function. Prior work that relate cross-entropy-based adaptive
importance sampling with Gaussian mixture models show that a mixture model
require less objective function calls than a naïve Monte Carlo or standard
unimodal cross-entropy-based importance sampling method [9, 10].
## III Background
This section provides necessary background on techniques used in this work. We
provide introductions to cross-entropy and the cross-entropy method, surrogate
modeling using Gaussian processes, and multivariate Gaussian mixture models.
### III-A Cross-Entropy
Before understanding the cross-entropy method, we first must understand the
notion of cross-entropy. Cross-entropy is a metric used to measure the
distance between two probability distributions, where the distance may not be
symmetric [3]. The distance used to define cross-entropy is called the
Kullback-Leibler (KL) distance or KL divergence. The KL distance is also
called the relative entropy, and we can use this to derive the cross-entropy.
Formally, for a random variable $\mathbf{X}=(X_{1},\ldots,X_{n})$ with a
support of $\mathcal{X}$, the KL distance between two continuous probability
density functions $f$ and $g$ is defined to be:
$\displaystyle\mathcal{D}(f,g)$
$\displaystyle=\mathbb{E}_{f}\left[\log\frac{f(\mathbf{X})}{g(\mathbf{X})}\right]$
$\displaystyle=\int\limits_{\mathbf{x}\in\mathcal{X}}f(\mathbf{x})\log
f(\mathbf{x})d\mathbf{x}-\int\limits_{\mathbf{x}\in\mathcal{X}}f(\mathbf{x})\log
g(\mathbf{x})d\mathbf{x}$
We denote the expectation of some function with respect to a distribution $f$
as $\mathbb{E}_{f}$. Minimizing the KL distance $\mathcal{D}$ between our true
distribution $f$ and our proposal distribution $g$ parameterized by
$\bm{\uptheta}$, is equivalent to choosing $\bm{\uptheta}$ that minimizes the
following, called the cross-entropy:
$\displaystyle H(f,g)$ $\displaystyle=H(f)+\mathcal{D}(f,g)$
$\displaystyle=-\mathbb{E}_{f}[\log g(\mathbf{X})]$ (using KL distance)
$\displaystyle=-\int\limits_{\mathbf{x}\in\mathcal{X}}f(\mathbf{x})\log
g(\mathbf{x}\mid\bm{\uptheta})d\mathbf{x}$
where $H(f)$ denotes the entropy of the distribution $f$ (where we conflate
entropy and continuous entropy for convenience). This assumes that $f$ and $g$
share the support $\mathcal{X}$ and are continuous with respect to
$\mathbf{x}$. The minimization problem then becomes:
$\displaystyle\operatornamewithlimits{minimize}_{\bm{\uptheta}}$
$\displaystyle-\int\limits_{\mathbf{x}\in\mathcal{X}}f(\mathbf{x})\log
g(\mathbf{x}\mid\bm{\uptheta})d\mathbf{x}$ (1)
Efficiently finding this minimum is the goal of the cross-entropy method
algorithm.
### III-B Cross-Entropy Method
Using the definition of cross-entropy, intuitively the cross-entropy method
(CEM or CE-method) aims to minimize the cross-entropy between the unknown true
distribution $f$ and a proposal distribution $g$ parameterized by
$\bm{\uptheta}$. This technique reformulates the minimization problem as a
probability estimation problem, and uses adaptive importance sampling to
estimate the unknown expectation [3]. The cross-entropy method has been
applied in the context of both discrete and continuous optimization problems
[2, 11].
The initial goal is to estimate the probability
$\displaystyle\ell=P_{\bm{\uptheta}}(S(\mathbf{x})\geq\gamma)$
where $S$ can the thought of as an objective function of $\mathbf{x}$, and
$\mathbf{x}$ follows a distribution defined by
$g(\mathbf{x}\mid\bm{\uptheta})$. We want to find events where our objective
function $S$ is above some threshold $\gamma$. We can express this unknown
probability as the expectation
$\displaystyle\ell=\mathbb{E}_{\bm{\uptheta}}[\mathbbm{1}_{(S(\mathbf{x})\geq\gamma)}]$
(2)
where $\mathbbm{1}$ denotes the indicator function. A straightforward way to
estimate eq. 2 can be done through Monte Carlo sampling. But for rare-event
simulations where the probability of a target event occurring is relatively
small, this estimate becomes inadequate. The challenge of the minimization in
eq. 1 then becomes choosing the density function for the true distribution
$f(\mathbf{x})$. Importance sampling tells us that the optimal importance
sampling density can be reduced to
$\displaystyle
f^{*}(\mathbf{x})=\frac{\mathbbm{1}_{(S(\mathbf{x})\geq\gamma)}g(\mathbf{x}\mid\bm{\uptheta})}{\ell}$
thus resulting in the optimization problem:
$\displaystyle\bm{\uptheta}_{g}^{*}$
$\displaystyle=\operatornamewithlimits{arg\,min}_{\bm{\uptheta}_{g}}-\int\limits_{\mathbf{x}\in\mathcal{X}}f^{*}(\mathbf{x})\log
g(\mathbf{x}\mid\bm{\uptheta}_{g})d\mathbf{x}$
$\displaystyle=\operatornamewithlimits{arg\,min}_{\bm{\uptheta}_{g}}-\int\limits_{\mathbf{x}\in\mathcal{X}}\frac{\mathbbm{1}_{(S(\mathbf{x})\geq\gamma)}g(\mathbf{x}\mid\bm{\uptheta})}{\ell}\log
g(\mathbf{x}\mid\bm{\uptheta}_{g})d\mathbf{x}$
Note that since we assume $f$ and $g$ belong to the same family of
distributions, we get that $f(\mathbf{x})=g(\mathbf{x}\mid\bm{\uptheta}_{g})$.
Now notice that $\ell$ is independent of $\bm{\uptheta}_{g}$, thus we can drop
$\ell$ and get the final optimization problem of:
$\displaystyle\bm{\uptheta}_{g}^{*}$
$\displaystyle=\operatornamewithlimits{arg\,min}_{\bm{\uptheta}_{g}}-\int\limits_{\mathbf{x}\in\mathcal{X}}\mathbbm{1}_{(S(\mathbf{x})\geq\gamma)}g(\mathbf{x}\mid\bm{\uptheta})\log
g(\mathbf{x}\mid\bm{\uptheta}_{g})d\mathbf{x}$ (3)
$\displaystyle=\operatornamewithlimits{arg\,min}_{\bm{\uptheta}_{g}}-\mathbb{E}_{\bm{\uptheta}}[\mathbbm{1}_{(S(\mathbf{x})\geq\gamma)}\log
g(\mathbf{x}\mid\bm{\uptheta}_{g})]$
The CE-method uses a multi-level algorithm to estimate $\bm{\uptheta}_{g}^{*}$
iteratively. The parameter $\bm{\uptheta}_{k}$ at iteration $k$ is used to
find new parameters $\bm{\uptheta}_{k^{\prime}}$ at the next iteration
$k^{\prime}$. The threshold $\gamma_{k}$ becomes smaller that its initial
value, thus artificially making events less rare under $\mathbf{X}\sim
g(\mathbf{x}\mid\bm{\uptheta}_{k})$.
In practice, the CE-method algorithm requires the user to specify a number of
elite samples $m_{\text{elite}}$ which are used when fitting the new
parameters for iteration $k^{\prime}$. Conveniently, if our distribution $g$
belongs to the natural exponential family then the optimal parameters can be
found analytically [5]. For a multivariate Gaussian distribution parameterized
by $\bm{\upmu}$ and $\bm{\Sigma}$, the optimal parameters for the next
iteration $k^{\prime}$ correspond to the maximum likelihood estimate (MLE):
$\displaystyle\bm{\upmu}_{k^{\prime}}$
$\displaystyle=\frac{1}{m_{\text{elite}}}\sum_{i=1}^{m_{\text{elite}}}\mathbf{x}_{i}$
$\displaystyle\bm{\Sigma}_{k^{\prime}}$
$\displaystyle=\frac{1}{m_{\text{elite}}}\sum_{i=1}^{m_{\text{elite}}}(\mathbf{x}_{i}-\bm{\upmu}_{k^{\prime}})(\mathbf{x}_{i}-\bm{\upmu}_{k^{\prime}})^{\top}$
The cross-entropy method algorithm is shown in algorithm 1. For an objective
function $S$ and input distribution $g$, the CE-method algorithm will run for
$k_{\text{max}}$ iterations. At each iteration, $m$ inputs are sampled from
$g$ and evaluated using the objective function $S$. The sampled inputs are
denoted by $\mathbf{X}$ and the evaluated values are denoted by $\mathbf{Y}$.
Next, the top $m_{\text{elite}}$ samples are stored in the elite set
$\mathbf{e}$, and the distribution $g$ is fit to the elites. This process is
repeated for $k_{\text{max}}$ iterations and the resulting parameters
$\bm{\uptheta}_{k_{\text{max}}}$ are returned. Note that a variety of input
distributions for $g$ are supported, but we focus on the multivariate Gaussian
distribution and the Gaussian mixture model in this work.
function CrossEntropyMethod($S,g,m,m_{\text{elite}},k_{\text{max}}$)
for $k\in[1,\ldots,k_{\text{max}}]$ do
$\mathbf{X}\sim g(\;\cdot\mid\bm{\uptheta}_{k})$ where
$\mathbf{X}\in\mathbb{R}^{m}$
$\mathbf{Y}\leftarrow S(\mathbf{x})$ for $\mathbf{x}\in\mathbf{X}$
$\mathbf{e}\leftarrow$ store top $m_{\text{elite}}$ from $\mathbf{Y}$
$\bm{\uptheta}_{k^{\prime}}\leftarrow\textsc{Fit}(g(\;\cdot\mid\bm{\uptheta}_{k}),\mathbf{e})$
return $g(\;\cdot\mid\bm{\uptheta}_{k_{\text{max}}})$
Algorithm 1 Cross-entropy method.
### III-C Mixture Models
A standard Gaussian distribution is unimodal and can have trouble generalizing
over data that is multimodal. A mixture model is a weighted mixture of
component distributions used to represent continuous multimodal distributions
[4]. Formally, a Gaussian mixture model (GMM) is defined by its parameters
$\bm{\upmu}$ and $\bm{\Sigma}$ and associated weights $\mathbf{w}$ where
$\sum_{i=1}^{n}w_{i}=1$. We denote that a random variable $\mathbf{X}$ is
distributed according to a mixture model as
$\mathbf{X}\sim\operatorname{Mixture}(\bm{\upmu},\bm{\Sigma},\mathbf{w})$. The
probability density of the GMM then becomes:
$\displaystyle
P(\mathbf{X}=\mathbf{x}\mid\bm{\upmu},\bm{\Sigma},\mathbf{w})=\sum_{i=1}^{n}w_{i}\operatorname{\mathcal{N}}(\mathbf{x}\mid\bm{\upmu}_{i},\bm{\Sigma}_{i})$
To fit the parameters of a Gaussian mixture model, it is well known that the
expectation-maximization (EM) algorithm can be used [12, 13]. The EM algorithm
seeks to find the maximum likelihood estimate of the hidden variable $H$ using
the observed data defined by $E$. Intuitively, the algorithm alternates
between an expectation step (E-step) and a maximization step (M-step) to
guarantee convergence to a local minima. A simplified EM algorithm is provide
in algorithm 2 for reference and we refer to [12, 13] for further reading.
function ExpectationMaximization($H,E,\bm{\uptheta}$)
for E-step do
Compute $Q(h)=P(H=h\mid E=e,\bm{\uptheta})$ for each $h$
Create weighted points: $(h,e)$ with weight $Q(h)$
for M-step do
Compute $\mathbf{\hat{\bm{\uptheta}}}_{\text{MLE}}$
Repeat until convergence.
return $\mathbf{\hat{\bm{\uptheta}}}_{\text{MLE}}$
Algorithm 2 Expectation-maximization.
### III-D Surrogate Models
In the context of optimization, a surrogate model $\hat{S}$ is used to
estimate the true objective function and provide less expensive evaluations.
Surrogate models are a popular approach and have been used to evaluate rare-
event probabilities in computationally expensive systems [14, 15]. The
simplest example of a surrogate model is linear regression. In this work, we
focus on the Gaussian process surrogate model. A Gaussian process (GP) is a
distribution over functions that predicts the underlying objective function
$S$ and captures the uncertainty of the prediction using a probability
distribution [5]. This means a GP can be sampled to generate random functions,
which can then be fit to our given data $\mathbf{X}$. A Gaussian process is
parameterized by a mean function $\mathbf{m}(\mathbf{X})$ and kernel function
$\mathbf{K}(\mathbf{X},\mathbf{X})$, which captures the relationship between
data points as covariance values. We denote a Gaussian process that produces
estimates $\hat{\mathbf{y}}$ as:
$\displaystyle\hat{\mathbf{y}}$
$\displaystyle\sim\mathcal{N}\left(\mathbf{m}(\mathbf{X}),\mathbf{K}(\mathbf{X},\mathbf{X})\right)$
$\displaystyle=\begin{bmatrix}\hat{S}(\mathbf{x}_{1}),\ldots,\hat{S}(\mathbf{x}_{n})\end{bmatrix}$
where
$\displaystyle\mathbf{m}(\mathbf{X})=\begin{bmatrix}m(\mathbf{x}_{1}),\ldots,m(\mathbf{x}_{n})\end{bmatrix}$
$\displaystyle\mathbf{K}(\mathbf{X},\mathbf{X})=\begin{bmatrix}k(\mathbf{x}_{1},\mathbf{x}_{1})&\cdots&k(\mathbf{x}_{1},\mathbf{x}_{n})\\\
\vdots&\ddots&\vdots\\\
k(\mathbf{x}_{n},\mathbf{x}_{1})&\cdots&k(\mathbf{x}_{n},\mathbf{x}_{n})\end{bmatrix}$
We use the commonly used zero-mean function $m(\mathbf{x}_{i})=\mathbf{0}$.
For the kernel function $k(\mathbf{x}_{i},\mathbf{x}_{i})$, we use the squared
exponential kernel with variance $\sigma^{2}$ and characteristic scale-length
$\ell$, where larger $\ell$ values increase the correlation between successive
data points, thus smoothing out the generated functions. The squared
exponential kernel is defined as:
$\displaystyle
k(\mathbf{x},\mathbf{x}^{\prime})=\sigma^{2}\exp\left(-\frac{(\mathbf{x}-\mathbf{x}^{\prime})^{\top}(\mathbf{x}-\mathbf{x}^{\prime})}{2\ell^{2}}\right)$
We refer to [5] for a detailed overview of Gaussian processes and different
kernel functions.
## IV Algorithms
We can now describe the cross-entropy method variants introduced in this work.
This section will first cover the main algorithm introduced, the cross-entropy
surrogate method (CE-surrogate). Then we introduce a modification to the CE-
surrogate method, namely the cross-entropy mixture method (CE-mixture).
Lastly, we describe various evaluation schedules for redistributing objective
function calls over the iterations.
### IV-A Cross-Entropy Surrogate Method
The main CE-method variant we introduce is the cross-entropy surrogate method
(CE-surrogate). The CE-surrogate method is a superset of the CE-method, where
the differences lie in the evaluation scheduling and modeling of the elite set
using a surrogate model. The goal of the CE-surrogate algorithm is to address
the shortcomings of the CE-method when the number of objective function calls
is sparse and the underlying objective function $S$ has multiple local minima.
The CE-surrogate algorithm is shown in algorithm 3. It takes as input the
objective function $S$, the distribution $\mathbf{M}$ parameterized by
$\bm{\uptheta}$, the number of samples $m$, the number of elite samples
$m_{\text{elite}}$, and the maximum iterations $k_{\text{max}}$. For each
iteration $k$, the number of samples $m$ are redistributed through a call to
EvaluationSchedule, where $m$ controls the number of true objective function
evaluations of $S$. Then, the algorithm samples from $\mathbf{M}$
parameterized by the current $\bm{\uptheta}_{k}$ given the adjusted number of
samples $m$. For each sample in $\mathbf{X}$, the objective function $S$ is
evaluated and the results are stored in $\mathbf{Y}$. The top
$m_{\text{elite}}$ evaluations from $\mathbf{Y}$ are stored in $\mathbf{e}$.
Using all of the current function evaluations $\mathbf{Y}$ from sampled inputs
$\mathbf{X}$, a modeled elite set $\mathbf{E}$ is created to augment the
sparse information provided by a low number of true objective function
evaluations. Finally, the distribution $\mathbf{M}$ is fit to the elite set
$\mathbf{E}$ and the distribution with the final parameters
$\bm{\uptheta}_{k_{\text{max}}}$ is returned.
function CE-Surrogate($S$, $\mathbf{M}$, $m$, $m_{\text{elite}}$,
$k_{\text{max}}$)
for $k\in[1,\ldots,k_{\text{max}}]$ do
$m,m_{\text{elite}}\leftarrow\textsc{EvaluationSchedule}(k,k_{\text{max}})$
$\mathbf{X}\sim\mathbf{M}(\;\cdot\mid\bm{\uptheta}_{k})$ where
$\mathbf{X}\in\mathbb{R}^{m}$
$\mathbf{Y}\leftarrow S(\mathbf{x})$ for $\mathbf{x}\in\mathbf{X}$
$\mathbf{e}\leftarrow$ store top $m_{\text{elite}}$ from $\mathbf{Y}$
$\mathbf{E}\leftarrow\textsc{ModelEliteSet}(\mathbf{X},\mathbf{Y},\mathbf{M},\mathbf{e},m,m_{\text{elite}})$
$\bm{\uptheta}_{k^{\prime}}\leftarrow\textsc{Fit}(\mathbf{M}(\;\cdot\mid\bm{\uptheta}_{k}),\mathbf{E})$
return $\mathbf{M}(\;\cdot\mid\bm{\uptheta}_{k_{\text{max}}})$
Algorithm 3 Cross-entropy surrogate method.
The main difference between the standard CE-method and the CE-surrogate
variant lies in the call to ModelEliteSet. The motivation is to use all of the
already evaluated objective function values $\mathbf{Y}$ from a set of sampled
inputs $\mathbf{X}$. This way the expensive function evaluations—otherwise
discarded—can be used to build a surrogate model of the underlying objective
function. First, a surrogate model $\hat{S}$ is constructed from the samples
$\mathbf{X}$ and true objective function values $\mathbf{Y}$. We used a
Gaussian process with a specified kernel and optimizer, but other surrogate
modeling techniques such as regression with basis functions can be used. We
chose a Gaussian process because it incorporates probabilistic uncertainty in
the predictions, which may more accurately represent our objective function,
or at least be sensitive to over-fitting to sparse data. Now we have an
approximated objective function $\hat{S}$ that we can inexpensively call. We
sample $10m$ values from the distribution $\mathbf{M}$ and evaluate them using
the surrogate model. We then store the top $10m_{\text{elite}}$ values from
the estimates $\mathbf{\hat{\mathbf{Y}}}_{\text{m}}$. We call these estimated
elite values $\mathbf{e}_{\text{model}}$ the model-elites. The surrogate model
is then passed to SubEliteSet, which returns more estimates for elite values.
Finally, the elite set $\mathbf{E}$ is built from the true-elites
$\mathbf{e}$, the model-elites $\mathbf{e}_{\text{model}}$, and the
subcomponent-elites $\mathbf{e}_{\text{sub}}$. The resulting concatenated
elite set $\mathbf{E}$ is returned.
function
ModelEliteSet($\mathbf{X},\mathbf{Y},\mathbf{M},\mathbf{e},m,m_{\text{elite}}$)
$\hat{S}\leftarrow\textsc{GaussianProcess}(\mathbf{X},\mathbf{Y},\text{kernel},\text{optimizer})$
$\mathbf{X}_{\text{m}}\sim\mathbf{M}(\;\cdot\mid\bm{\uptheta}_{k})$ where
$\mathbf{X}_{\text{m}}\in\mathbb{R}^{10m}$
$\mathbf{\hat{\mathbf{Y}}}_{\text{m}}\leftarrow\hat{S}(\mathbf{x}_{\text{m}})$
for $\mathbf{x}_{\text{m}}\in\mathbf{X}_{\text{m}}$
$\mathbf{e}_{\text{model}}\leftarrow$ store top $10m_{\text{elite}}$ from
$\mathbf{\hat{\mathbf{Y}}}_{\text{m}}$
$\mathbf{e}_{\text{sub}}\leftarrow\textsc{SubEliteSet}(\hat{S},\mathbf{M},\mathbf{e})$
$\mathbf{E}\leftarrow\\{\mathbf{e}\\}\cup\\{\mathbf{e}_{\text{model}}\\}\cup\\{\mathbf{e}_{\text{sub}}\\}$
$\triangleright$ elite set
return $\mathbf{E}$
Algorithm 4 Modeling elite set.
To encourage exploration of promising areas of the design space, the algorithm
SubEliteSet focuses on the already marked true-elites $\mathbf{e}$. Each elite
$e_{x}\in\mathbf{e}$ is used as the mean of a new multivariate Gaussian
distribution with covariance inherited from the distribution $\mathbf{M}$. The
collection of subcomponent distributions is stored in $\mathbf{m}$. The idea
is to use the information given to us by the true-elites to emphasize areas of
the design space that look promising. For each distribution
$\mathbf{m}_{i}\in\mathbf{m}$ we run a subroutine call to the standard CE-
method to fit the distribution $\mathbf{m}_{i}$ using the surrogate model
$\hat{S}$. Then the best objective function value is added to the
subcomponent-elite set $\mathbf{e}_{\text{sub}}$, and after iterating the full
set is returned. Note that we use $\theta_{\text{CE}}$ to denote the
parameters for the CE-method algorithm. In our case, we recommend using a
small $k_{\text{max}}$ of around $2$ so the subcomponent-elites do not over-
fit to the surrogate model but have enough CE-method iterations to tend
towards optimal.
function SubEliteSet($\hat{S},\mathbf{M},\mathbf{e}$)
$\mathbf{e}_{\text{sub}}\leftarrow\emptyset$
$\mathbf{m}\leftarrow\\{e_{x}\in\mathbf{e}\mid\operatorname{\mathcal{N}}(e_{x},\mathbf{M}.\Sigma)\\}$
for $\mathbf{m}_{i}\in\mathbf{m}$ do
$\mathbf{m}_{i}\leftarrow\textsc{CrossEntropyMethod}(\hat{S},\mathbf{m}_{i}\mid\theta_{\text{CE}})$
$\mathbf{e}_{\text{sub}}\leftarrow\\{\mathbf{e}_{\text{sub}}\\}\cup\\{\textsc{Best}(\mathbf{m}_{i})\\}$
return $\mathbf{e}_{\text{sub}}$
Algorithm 5 Subcomponent elite set.
### IV-B Cross-Entropy Mixture Method
We refer to the variant of our CE-surrogate method that takes an input mixture
model $\mathbf{M}$ as the cross-entropy mixture method (CE-mixture). The CE-
mixture algorithm is identical to the CE-surrogate algorithm, but calls a
custom Fit function to fit a mixture model to the elite set $\mathbf{E}$. The
input distribution $\mathbf{M}$ is cast to a mixture model using the
subcomponent distributions $\mathbf{m}$ as the components of the mixture. We
use the default uniform weighting for each mixture component. The mixture
model $\mathbf{M}$ is then fit using the expectation-maximization algorithm
shown in algorithm 2, and the resulting distribution is returned. The idea is
to use the distributions in $\mathbf{m}$ that are centered around each true-
elite as the components of the casted mixture model. Therefore, we would
expect better performance of the CE-mixture method when the objective function
has many competing local minima. Results in section V-C aim to show this
behavior.
function Fit($\mathbf{M},\mathbf{m},\mathbf{E}$)
$\mathbf{M}\leftarrow\operatorname{Mixture}(\mathbf{m})$
$\mathbf{\hat{\bm{\uptheta}}}\leftarrow\textsc{ExpectationMaximization}(\mathbf{M},\mathbf{E})$
return $\mathbf{M}(\;\cdot\mid\mathbf{\hat{\bm{\uptheta}}})$
Algorithm 6 Fitting mixture models (used by CE-mixture).
### IV-C Evaluation Scheduling
Given the nature of the CE-method, we expect the covariance to shrink over
time, thus resulting in a solution with higher confidence. Yet if each
iteration is given the same number of objective function evaluations $m$,
there is the potential for elite samples from early iterations dominating the
convergence. Therefore, we would like to redistribute the objective function
evaluations throughout the iterations to use more truth information early in
the process. We call these heuristics evaluation schedules. One way to achieve
this is to reallocate the evaluations according to a Geometric distribution.
Evaluation schedules can also be ad-hoc and manually prescribed based on the
current iteration.
We provide the evaluation schedule we use that follows a Geometric
distribution with parameter $p$ in algorithm 7. We denote
$G\sim\operatorname{Geo}(p)$ to be a random variable that follows a truncated
Geometric distribution with the probability mass function
$p_{G}(k)=p(1-p)^{k}$ for $k\in\\{0,1,2,\ldots,k_{\text{max}}\\}$. Note the
use of the integer rounding function (e.g., $\lfloor x\rceil$), which we later
have to compensate for towards the final iterations. Results in section V-C
compare values of $p$ that control the redistribution of evaluations.
function EvaluationSchedule($k,k_{\text{max}}$)
$G\sim\operatorname{Geo}(p)$
$N_{\text{max}}\leftarrow k_{\text{max}}\cdot m$
$m\leftarrow\lfloor N_{\text{max}}\cdot p_{G}(k)\rceil$
if $k=k_{\text{max}}$ then
$s\leftarrow\displaystyle\sum_{i=1}^{k_{\text{max}}-1}\lfloor
N_{\text{max}}\cdot p_{G}(i)\rceil$
$m\leftarrow\min(N_{\text{max}}-s,N_{\text{max}}-m)$
$m_{\text{elite}}\leftarrow\min(m_{\text{elite}},m)$
return ($m,m_{\text{elite}}$)
Algorithm 7 Evaluation schedule using a Geometric distr.
## V Experiments
In this section, we detail the experiments we ran to compare the CE-method
variants and evaluation schedules. We first introduce a test objective
function we created to stress the issue of converging to local minima. We then
describe the experimental setup for each of our experiments and provide an
analysis and results.
### V-A Test Objective Function Generation
$-15$$-10$$-5$$0$$5$$10$$15$$\eta=0.5$, decay=1$\eta=2.0$, decay=1$\eta=6.0$,
decay=1$-2$$-1.5$$-1$$-0.5$$0$$\cdot
10^{-2}$$-15$$-10$$-5$$0$$5$$10$$15$$-15$$-10$$-5$$0$$5$$10$$15$$\eta=0.5$,
decay=0$-15$$-10$$-5$$0$$5$$10$$15$$\eta=2.0$,
decay=0$-15$$-10$$-5$$0$$5$$10$$15$$\eta=6.0$,
decay=0$-2$$-1.5$$-1$$-0.5$$0$$\cdot 10^{-2}$
Figure 1: Example test objective functions generated using the sierra
function.
To stress the cross-entropy method and its variants, we created a test
objective function called sierra that is generated from a mixture model
comprised of $49$ multivariate Gaussian distributions. We chose this
construction so that we can use the negative peeks of the component
distributions as local minima and can force a global minimum centered at our
desired $\mathbf{\tilde{\bm{\upmu}}}$. The construction of the sierra test
function can be controlled by parameters that define the spread of the local
minima. We first start with the center defined by a mean vector
$\mathbf{\tilde{\bm{\upmu}}}$ and we use a common covariance
$\mathbf{\tilde{\bm{\Sigma}}}$:
$\displaystyle\mathbf{\tilde{\bm{\upmu}}}$
$\displaystyle=[\mu_{1},\mu_{2}],\quad\mathbf{\tilde{\bm{\Sigma}}}=\begin{bmatrix}\sigma&0\\\
0&\sigma\end{bmatrix}$
Next, we use the parameter $\delta$ that controls the clustered distance
between symmetric points:
$\displaystyle\mathbf{G}$
$\displaystyle=\left\\{[+\delta,+\delta],[+\delta,-\delta],[-\delta,+\delta],[-\delta,-\delta]\right\\}$
We chose points $\mathbf{P}$ to fan out the clustered minima relative to the
center defined by $\mathbf{\tilde{\bm{\upmu}}}$:
$\displaystyle\mathbf{P}$
$\displaystyle=\left\\{[0,0],[1,1],[2,0],[3,1],[0,2],[1,3]\right\\}$
The vector $\mathbf{s}$ is used to control the $\pm$ distance to create an ‘s’
shape comprised of minima, using the standard deviation $\sigma$:
$\mathbf{s}=\begin{bmatrix}+\sigma,-\sigma\end{bmatrix}$. We set the following
default parameters: standard deviation $\sigma=3$, spread rate $\eta=6$, and
cluster distance $\delta=2$. We can also control if the local minima clusters
“decay”, thus making those local minima less distinct (where
$\text{decay}\in\\{0,1\\})$. The parameters that define the sierra function
are collected into
$\bm{\uptheta}=\langle\mathbf{\tilde{\bm{\upmu}}},\mathbf{\tilde{\bm{\Sigma}}},\mathbf{G},\mathbf{P},\mathbf{s}\rangle$.
Using these parameters, we can define the mixture model used by the sierra
function as:
$\displaystyle\mathbf{M}_{\mathcal{S}}\sim\operatorname{Mixture}\left(\left\\{\bm{\uptheta}~{}\Big{|}~{}\operatorname{\mathcal{N}}\left(\mathbf{g}+s\mathbf{p}_{i}+\mathbf{\tilde{\bm{\upmu}}},\;\mathbf{\tilde{\bm{\Sigma}}}\cdot
i^{\text{decay}}/\eta\right)\right\\}\right)$ $\displaystyle\text{for
}(\mathbf{g},\mathbf{p}_{i},s)\in(\mathbf{G},\mathbf{P},\mathbf{s})$
We add a final component to be our global minimum centered at
$\mathbf{\tilde{\bm{\upmu}}}$ and with a covariance scaled by $\sigma\eta$.
Namely, the global minimum is
$\mathbf{x}^{*}=\mathbb{E}[\operatorname{\mathcal{N}}(\mathbf{\tilde{\bm{\upmu}}},\mathbf{\tilde{\bm{\Sigma}}}/(\sigma\eta))]=\mathbf{\tilde{\bm{\upmu}}}$.
We can now use this constant mixture model with $49$ components and define the
sierra objective function $\mathcal{S}(\mathbf{x})$ to be the negative
probability density of the mixture at input $\mathbf{x}$ with uniform weights:
$\displaystyle\mathcal{S}(\mathbf{x})$
$\displaystyle=-P(\mathbf{M}_{\mathcal{S}}=\mathbf{x})=-\frac{1}{|\mathbf{M}_{\mathcal{S}}|}\sum_{j=1}^{n}\operatorname{\mathcal{N}}(\mathbf{x}\mid\bm{\upmu}_{j},\bm{\Sigma}_{j})$
An example of six different objective functions generated using the sierra
function are shown in fig. 1, sweeping over the spread rate $\eta$, with and
without decay.
### V-B Experimental Setup
Experiments were run to stress a variety of behaviors of each CE-method
variant. The experiments are split into two categories: algorithmic and
scheduling. The algorithmic category aims to compare features of each CE-
method variant while holding common parameters constant (for a better
comparison). While the scheduling category experiments with evaluation
scheduling heuristics.
$\displaystyle-15$$\displaystyle-10$$\displaystyle-5$$\displaystyle
0$$\displaystyle 5$$\displaystyle 10$$\displaystyle
15$$\displaystyle-15$$\displaystyle-10$$\displaystyle-5$$\displaystyle
0$$\displaystyle 5$$\displaystyle 10$$\displaystyle 15$Experiment 1A: CE-
method ($\displaystyle k=5$)
(a) The cross-entropy method.
$\displaystyle-15$$\displaystyle-10$$\displaystyle-5$$\displaystyle
0$$\displaystyle 5$$\displaystyle 10$$\displaystyle
15$$\displaystyle-15$$\displaystyle-10$$\displaystyle-5$$\displaystyle
0$$\displaystyle 5$$\displaystyle 10$$\displaystyle 15$Experiment 1A: CE-
surrogate ($\displaystyle k=5$)
(b) The cross-entropy surrogate method.
$\displaystyle-15$$\displaystyle-10$$\displaystyle-5$$\displaystyle
0$$\displaystyle 5$$\displaystyle 10$$\displaystyle
15$$\displaystyle-15$$\displaystyle-10$$\displaystyle-5$$\displaystyle
0$$\displaystyle 5$$\displaystyle 10$$\displaystyle 15$Experiment 1A: CE-
mixture ($\displaystyle k=5$)
(c) The cross-entropy mixture method.
Figure 2: Iteration $k=5$ illustrated for each algorithm. The covariance is
shown by the contours.
Because the algorithms are stochastic, we run each experiment with 50
different random number generator seed values. To evaluate the performance of
the algorithms in their respective experiments, we define three metrics.
First, we define the average “optimal” value $\bar{b}_{v}$ to be the average
of the best so-far objective function value (termed “optimal” in the context
of each algorithm). Again, we emphasize that we average over the 50 seed
values to gather meaningful statistics. Another metric we monitor is the
average distance to the true global optimal
$\bar{b}_{d}=\mathinner{\\!\left\lVert\mathbf{b}_{\mathbf{x}}-\mathbf{x}^{*}\right\rVert}$,
where $\mathbf{b}_{\mathbf{x}}$ denotes the $\mathbf{x}$-value associated with
the “optimal”. We make the distinction between these metrics to show both
“closeness” in value to the global minimum and “closeness” in the design space
to the global minimum. Our final metric looks at the average runtime of each
algorithm, noting that our goal is to off-load computationally expensive
objective function calls to the surrogate model.
For all of the experiments, we use a common setting of the following
parameters for the sierra test function (shown in the top-right plot in fig.
1):
$(\mathbf{\tilde{\bm{\upmu}}}=[0,0],\;\sigma=3,\;\delta=2,\;\eta=6,\;\text{decay}=1)$
#### V-B1 Algorithmic Experiments
We run three separate algorithmic experiments, each to test a specific
feature. For our first algorithmic experiment (1A), we want to test each
algorithm when the user-defined mean is centered at the global minimum and the
covariance is arbitrarily wide enough to cover the design space. Let
$\mathbf{M}$ be a distribution parameterized by
$\bm{\uptheta}=(\bm{\upmu},\bm{\Sigma})$, and for experiment (1A) we set the
following:
$\bm{\upmu}^{(\text{1A})}=[0,0]\qquad\bm{\Sigma}^{(\text{1A})}=\begin{bmatrix}200&0\\\
0&200\end{bmatrix}$
For our second algorithmic experiment (1B), we test a mean that is far off-
centered with a wider covariance:
$\bm{\upmu}^{(\text{1B})}=[-50,-50]\qquad\bm{\Sigma}^{(\text{1B})}=\begin{bmatrix}2000&0\\\
0&2000\end{bmatrix}$
This experiment is used to test the “exploration” of the CE-method variants
introduced in this work. In experiments (1A) and (1B), we set the following
common parameters across each CE-method variant:
$(k_{\text{max}}=10,\;m=10,\;m_{\text{elite}}=5)^{(\text{1A,1B})}$
This results in $m\cdot k_{\text{max}}=100$ objective function evaluations,
which we define to be relatively low.
For our third algorithmic experiment (1C), we want to test how each variant
responds to an extremely low number of function evaluations. This sparse
experiment sets the common CE-method parameters to:
$(k_{\text{max}}=10,\;m=5,\;m_{\text{elite}}=3)^{(\text{1C})}$
This results in $m\cdot k_{\text{max}}=50$ objective function evaluations,
which we defined to be extremely low. We use the same mean and covariance
defined for experiment (1A):
$\bm{\upmu}^{(\text{1C})}=[0,0]\qquad\bm{\Sigma}^{(\text{1C})}=\begin{bmatrix}200&0\\\
0&200\end{bmatrix}$
#### V-B2 Scheduling Experiments
In our final experiment (2), we test the evaluation scheduling heuristics
which are based on the Geometric distribution. We sweep over the parameter $p$
that determines the Geometric distribution which controls the redistribution
of objective function evaluations. In this experiment, we compare the CE-
surrogate methods using the same setup as experiment (1B), namely the far off-
centered mean. We chose this setup to analyze exploration schemes when given
very little information about the true objective function.
### V-C Results and Analysis
$2$$4$$6$$8$$10$$-2$$-1$$0$$\cdot 10^{-2}$Iteration$\bar{b}_{v}$Experiment
1ACE-methodCE-surrogateCE-mixture
Figure 3: Average optimal value for experiment (1A) when the initial mean is
centered at the global minimum and the covariance sufficiently covers the
design space.
Figure 3 shows the average value of the current optimal $\bar{b}_{v}$ for the
three algorithms for experiment (1A). One standard deviation is plotted in the
shaded region. Notice that the standard CE-method converges to a local minima
before $k_{\text{max}}$ is reached. Both CE-surrogate method and CE-mixture
stay below the standard CE-method curve, highlighting the mitigation of
convergence to local minima. Minor differences can be seen between CE-
surrogate and CE-mixture, differing slightly towards the tail in favor of CE-
surrogate. The average runtime of the algorithms along with the performance
metrics are shown together for each experiment in table I.
TABLE I: Experimental results. Exper. | Algorithm | Runtime | $\bar{b}_{v}$ | $\bar{b}_{d}$
---|---|---|---|---
1A | CE-method | 0.029 $\operatorname{s}$ | $-$0.0134 | 23.48
CE-surrogate | 1.47 $\operatorname{s}$ | $-$0.0179 | 12.23
CE-mixture | 9.17 $\operatorname{s}$ | $-$0.0169 | 16.87
1B | CE-method | 0.046 $\operatorname{s}$ | $-$0.0032 | 138.87
CE-surrogate | 11.82 $\operatorname{s}$ | $-$0.0156 | 18.24
CE-mixture | 28.10 $\operatorname{s}$ | $-$0.0146 | 33.30
1C | CE-method | 0.052 $\operatorname{s}$ | $-$0.0065 | 43.14
CE-surrogate | 0.474 $\operatorname{s}$ | $-$0.0156 | 17.23
CE-mixture | 2.57 $\operatorname{s}$ | $-$0.0146 | 22.17
2 | CE-surrogate, $\operatorname{Uniform}$ | — | $-$0.0193 | 8.53
CE-surrogate, $\operatorname{Geo}(0.1)$ | — | $-$0.0115 | 25.35
CE-surrogate, $\operatorname{Geo}(0.2)$ | — | $-$0.0099 | 27.59
| CE-surrogate, $\operatorname{Geo}(0.3)$ | — | $-$0.0089 | 30.88
| | | $-\text{0.0220}\approx\mathbf{x}^{*}$
An apparent benefit of the standard CE-method is in its simplicity and speed.
As shown in table I, the CE-method is the fastest approach by about 2-3 orders
of magnitude compared to CE-surrogate and CE-mixture. The CE-mixture method is
notably the slowest approach. Although the runtime is also based on the
objective function being tested, recall that we are using the same number of
true objective function calls in each algorithm, and the metrics we are
concerned with in optimization are to minimize $\bar{b}_{v}$ and
$\bar{b}_{d}$. We can see that the CE-surrogate method consistently out
performs the other methods. Surprisingly, a uniform evaluation schedule
performs the best even in the sparse scenario where the initial mean is far
away from the global optimal.
$2$$4$$6$$8$$10$$-1.5$$-1$$-0.5$$0$$\cdot
10^{-2}$Iteration$\bar{b}_{v}$Experiment 1BCE-methodCE-surrogateCE-mixture
Figure 4: Average optimal value for experiment (1B) when the initial mean is
far from the global minimum with a wide covariance.
When the initial mean of the input distribution is placed far away from the
global optimal, the CE-method tends to converge prematurely as shown in fig.
4. This scenario is illustrated in fig. 5. We can see that both CE-surrogate
and CE-mixture perform well in this case.
$\displaystyle-100$$\displaystyle-80$$\displaystyle-60$$\displaystyle-40$$\displaystyle-20$$\displaystyle
0$$\displaystyle-100$$\displaystyle-80$$\displaystyle-60$$\displaystyle-40$$\displaystyle-20$$\displaystyle
0$Experiment 1B: $\displaystyle k=1$
Figure 5: First iteration of the scenario in experiment (1B) where the
initial distribution is far away form the global optimal. The red dots
indicate the true-elites, the black dots with white outlines indicate the
“non-elites” evaluated from the true objective function, and the white dots
with black outlines indicate the samples evaluated using the surrogate model.
$2$$4$$6$$8$$10$$-2$$-1$$0$$\cdot 10^{-2}$Iteration$\bar{b}_{v}$Experiment
1CCE-methodCE-surrogateCE-mixture
Figure 6: Average optimal value for experiment (1C) when we restrict the
number of objective function calls.
Given the same centered mean as before, when we restrict the number of
objective function calls even further to just 50 we see interesting behavior.
Notice that the results of experiment (1C) shown in fig. 6 follow a curve
closer to the far away mean from experiment (1B) than from the same setup as
experiment (1A). Also notice that the CE-surrogate results cap out at
iteration 9 due to the evaluation schedule front-loading the objective
function calls, thus leaving none for the final iteration (while still
maintaining the same total number of evaluations of 50).
## VI Conclusion
We presented variants of the popular cross-entropy method for optimization of
objective functions with multiple local minima. Using a Gaussian processes-
based surrogate model, we can use the same number of true objective function
evaluations and achieve better performance than the standard CE-method on
average. We also explored the use of a Gaussian mixture model to help find
global minimum in multimodal objective functions. We introduce a parameterized
test objective function with a controllable global minimum and spread of local
minima. Using this test function, we showed that the CE-surrogate algorithm
achieves the best performance relative to the standard CE-method, each using
the same number of true objective function evaluations.
## References
* [1] RY Rubinstein and DP Kroese “The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning” Springer, 2004
* [2] Reuven Rubinstein “The Cross-Entropy Method for Combinatorial and Continuous Optimization” In _Methodology and Computing in Applied Probability_ 1.2 Springer, 1999, pp. 127–190
* [3] Pieter-Tjerk De Boer, Dirk P Kroese, Shie Mannor and Reuven Y Rubinstein “A Tutorial on the Cross-Entropy Method” In _Annals of Operations Research_ 134.1 Springer, 2005, pp. 19–67
* [4] Mykel J. Kochenderfer “Decision Making Under Uncertainty: Theory and Application” MIT Press, 2015
* [5] Mykel J Kochenderfer and Tim A Wheeler “Algorithms for Optimization” MIT Press, 2019
* [6] Rémi Bardenet and Balázs Kégl “Surrogating the Surrogate: Accelerating Gaussian-Process-Based Global Optimization with a Mixture Cross-Entropy Algorithm” In _International Conference on Machine Learning (ICML)_ , 2010, pp. 55–62
* [7] Y.. Tan, A. Kunapareddy and M. Kobilarov “Gaussian Process Adaptive Sampling Using the Cross-Entropy Method for Environmental Sensing and Monitoring” In _International Conference on Robotics and Automation (ICRA)_ , 2018, pp. 6220–6227
* [8] Abbas Abdolmaleki et al. “Model-Based Relative Entropy Stochastic Search” In _Advances in Neural Information Processing Systems (NIPS)_ , 2015, pp. 3537–3545
* [9] Nolan Kurtz and Junho Song “Cross-entropy-based adaptive importance sampling using Gaussian mixture” In _Structural Safety_ 42 Elsevier, 2013, pp. 35–44
* [10] Ziqi Wang and Junho Song “Cross-entropy-based adaptive importance sampling using von Mises-Fisher mixture for high dimensional reliability analysis” In _Structural Safety_ 59 Elsevier, 2016, pp. 42–52
* [11] Dirk P Kroese, Sergey Porotsky and Reuven Y Rubinstein “The cross-entropy method for continuous multi-extremal optimization” In _Methodology and Computing in Applied Probability_ 8.3 Springer, 2006, pp. 383–407
* [12] Arthur P Dempster, Nan M Laird and Donald B Rubin “Maximum likelihood from incomplete data via the EM algorithm” In _Journal of the Royal Statistical Society: Series B (Methodological)_ 39.1 Wiley Online Library, 1977, pp. 1–22
* [13] Murray Aitkin and Granville Tunnicliffe Wilson “Mixture models, outliers, and the EM algorithm” In _Technometrics_ 22.3 Taylor & Francis Group, 1980, pp. 325–331
* [14] Jing Li and Dongbin Xiu “Evaluation of failure probability via surrogate models” In _Journal of Computational Physics_ 229.23 Elsevier, 2010, pp. 8966–8980
* [15] Jing Li, Jinglai Li and Dongbin Xiu “An efficient surrogate-based method for computing rare failure probability” In _Journal of Computational Physics_ 230.24 Elsevier, 2011, pp. 8683–8697
|
that the full integrand localises on the maximal non-separating boundary of
the moduli space, corresponding to a bi-nodal sphere, parametrised by four
‘loop marked points’, one pair per node,
$\displaystyle\mathfrak{I}^{(2)}_{4}$
$\displaystyle=\frac{\mathcal{K}\tilde{\mathcal{K}}}{\prod_{I}(\ell^{I})^{2}}\int_{{\mathfrak{M}}_{0,4+2g}}c^{(g)}\big{(}{\mathcal{J}}^{(g)}\mathcal{Y}^{(g)}\big{)}^{2}\prod_{{\scalebox{0.6}{$A$}}=1}^{4+4}{}^{\prime}\bar{\delta}(\mathcal{E}_{\scalebox{0.6}{$A$}})\
\,.$ (5.16)
The resulting integrand formula, presented here for $n=4$ particles for
simplicity, takes the form of an integral over the moduli space
${\mathfrak{M}}_{0,n+2g}$, fully localised on solutions to the nodal
scattering equations $\mathcal{E}_{\scalebox{0.6}{$A$}}=0$. While the
structure is reminiscent of the one-loop case, new features appear as well; in
particular the factors $c^{(g)}$ and ${\mathcal{J}}^{(g)}$ arise from the
degeneration of ${\mathfrak{M}}_{g,n}$ to ${\mathfrak{M}}_{0,n+2g}$. We
briefly discuss these ingredients below, more detailed expositions, as well as
$n$-point formulæ for supergravity and super Yang-Mills theory (constructed
again using the double copy) can be found in the original paper [108].
* •
_Moduli:_ The residue theorem localises the integrand on the non-separating
boundary, where $q_{II}=e^{i\pi\Omega_{II}}=0$. In this limit, the remaining
moduli, given by the off-diagonal components of the period matrix, become
cross-ratios of the nodal marked points $\sigma_{I^{\pm}}$,
$q_{IJ}=e^{2i\pi\Omega_{IJ}}=\frac{\sigma_{I^{+}J^{+}}\sigma_{I^{-}J^{-}}}{\sigma_{I^{+}J^{-}}\sigma_{I^{-}J^{+}}}\,.$
(5.17)
The measure is then naturally expressed in terms of all marked points
(including the nodes) modulo Möbius transformation, leading to a Jacobian
${\mathcal{J}}^{(g)}$ with
$\prod_{I<J}\frac{dq_{IJ}}{q_{IJ}}=\frac{{\mathcal{J}}^{(g)}}{\mathrm{vol\;SL}(2,\mathbb{C})}\,,\quad{\mathcal{J}}^{(g)}=\frac{1}{\sigma_{1^{+}2^{+}}\sigma_{1^{+}2^{-}}\sigma_{1^{-}2^{+}}\sigma_{1^{-}2^{-}}}\prod_{I^{\pm}}d\sigma_{I^{\pm}}\,.$
(5.18)
The analogous change of variables in the scattering equations results in
another copy of the same Jacobian factor.
The cross-ratio factor $\,c^{(2)}=1/(1-q_{12})$ originates in the degeneration
of the moduli space ${\mathfrak{M}}_{2,n}$ to ${\mathfrak{M}}_{0,n+4}$; to be
precise, from mapping the last modular parameter $q_{12}$ to the nodal sphere
[108]. It can be given a concrete physical interpretation in projecting out
unphysical poles from the integrand $\mathfrak{I}_{n}^{(2)}$.
* •
_Nodal scattering equations:_ On the nodal sphere, $P_{\mu}$ takes the
following form;
$P_{\mu}(\sigma)=\ell^{I}_{\mu}\,\omega_{I^{+}I^{-}}(\sigma)+\sum_{i}\frac{k_{i\mu}}{\sigma-\sigma_{i}}\,d\sigma\,,$
(5.19)
where $\omega_{I^{+}I^{-}}(\sigma)$ are the genus-two holomorphic Abelian
differentials in the maximal non-separating degeneration. In this limit, these
differentials acquire simple poles at the corresponding nodes:
$\omega_{I}=\frac{\omega_{I^{+}I^{-}}}{2\pi
i}\,,\quad\omega_{I^{+}I^{-}}(\sigma)=\frac{(\sigma_{I^{+}}-\sigma_{I^{-}})\,d\sigma}{(\sigma-\sigma_{I^{+}})(\sigma-\sigma_{I^{-}})}\,,$
(5.20)
The nodal scattering equations can then be compactly expressed as the
vanishing of a meromorphic quadratic differential $\mathfrak{P}^{(g)}$ with
only simple poles,
$\mathcal{E}_{\scalebox{0.6}{$A$}}=\mathrm{Res}_{\sigma_{\\!A}}\mathfrak{P}^{(g)}\,,\qquad\mathfrak{P}^{(2)}=P^{2}-(\ell^{I}\\!\omega_{I^{+}I^{-}})^{2}+(\ell_{1}^{2}+\ell_{2}^{2})\,\omega_{1^{+}1^{-}}\omega_{2^{+}2^{-}}\,.$
(5.21)
Note in particular the last term, a novel feature at genus two that plays a
crucial role in obtaining the correct loop propagators. 191919and is closely
related to the absence of a straightforward Feynman tree-theorem at two loops.
* •
_Integrand:_ The chiral integrand is defined straightforwardly by the nodal
sphere-limit of the genus-two expression,
$\mathcal{Y}=\mathcal{Y}^{(2)}\big{|}_{\mathrm{nodal}}$, and can be calculated
using (5.20).
### 5.6 Further topics
The nodal sphere formulation of loop integrands proved to be a starting point
for many further exciting avenues of research.
* •
At the level of the worldsheet model, the simple structure of one-loop
correlators, supported on a nodal sphere is reflected by the presence of a so-
called ‘gluing operator’ in the ambitwistor string [159]. This gluing operator
$\Delta$ encodes the propagator of the target-space field theory, and is thus
a BRST-invariant but non-local worldsheet operator. Genus zero correlators
with an insertion of $\Delta$ directly give the one-loop integrand formulæ
(5.2) localized on the nodal sphere, without need for further simplifications.
Extensions to higher loops are currently not known.
* •
Focusing on the integrand expressions, there has been tremendous progress on
extending many of the tree-level evaluation techniques to loop level, [160,
161, 162, 163].
* •
The nodal sphere, and in particular the forward limit structure of the
integrand, also inspired loop formulæ in the twistorial models [100]. These
were obtained from a forward limit of the 6d spinorial tree-level amplitudes
mentioned in section 3.
* •
Very recently, a proposal has also appeared for one-loop correlators in
massive $\phi^{4}$ theory on de Sitter spacetime, based on the nodal sphere
[164]. This builds on earlier work expressing (tree-level) de Sitter
‘cosmological correlators’ as worldsheet integrals, supported on so-called
cosmological scattering equations, which are now differential operators
expressed as functions of the conformal generators [165].
* •
Finally, the ambitwistor string progress at loop level has also inspired
calculations and proposals in the full superstring. In ref. [166], the authors
used an ambitwistor-string-inspired method, based on forward limits of the
moduli space integrals, to construct one-loop matrix elements with insertions
of operators $D^{2k}F^{n}$ and $D^{2k}R^{n}$ in the tree-level effective
action. Progress has also been made at higher loop orders, where the double
copy and the close relation between the ambitwistor string and the superstring
chiral integrands has been used to propose a formula for the three-loop four-
particle superstring integrand [91]. This is achieved by constructing first an
expression on the three-nodal sphere using BCJ numerators, and then lifting
this to a fully modular invariant proposal for the superstring chiral
integrand on a $g=3$ Riemann surface.
## 6 Frontiers
As we have seen, ambitwistor-strings give one of the most direct routes from a
physical theory to compact formulæ for tree amplitudes and loop integrands.
Despite these successes, many open questions remain and it is questionable as
to whether it will one day be possible to understand these worldsheet models
as providing secure basic formulations of physical theories. To consolidate
them, more work needs to be done to relate them to more standard fully
nonlinear formulations of physical theories, either via field theory, string
theory or holography. We briefly expand on these connections.
### 6.1 Curved backgrounds
Amplitudes on curved backgrounds are a relatively new subject with the
frontier being, until recently, at three points at tree level. They provide a
stepping stone to connect with conventional nonlinear field theory. Spaces of
complexified null geodesics make good sense on an analytic curved space with
metric $g(x)_{\mu\nu}$ [33] and one can ask whether ambitwistor strings can be
defined on such curved ambitwistor spaces. In particular in the RNS models of
§2 we can replace $P^{2}\rightarrow H=g^{\mu\nu}(x)P_{\mu}P_{\nu}+\ldots$. It
was shown in [167] that, in the type II case, curved analogues of the
constraints $H$, $G=\Psi\cdot P$ and $\tilde{G}=\tilde{\Psi}\cdot P$ can be
constructed so that they satisfy the flat space OPEs iff they are obtained
from a solution to the NS-NS-sector of 10d type II supergravity. This can be
used to construct vertex operators and amplitudes at three points on a plane
wave [168, 169], and one can similarly encode the Yang-Mills equations in the
heterotic model [170]. This provides a completely different perspective to
that pursued in the 1970’s and 1980’s when space-time field equations in 4d
were shown [24, 34, 171] to correspond to the existence of formal
neighbourhoods of $\mathbb{A}$ inside
$\mathbb{P}\mathbb{T}\times\mathbb{P}\mathbb{T}^{*}$ or supersymmetric
extensions [23, 172]. In the ambitwistor string, the field equations are
encoded in the quantum consistency of the worldsheet model on a curved
ambitwistor space. In all these cases and unlike the conventional string, one
can use field redefinitions to make the gauge-fixed action linear, so that
correlators are relatively easy to compute. However, the complexity of the
curved background constraints makes it problematic to identify the generic
integrated vertex operators.
The 4d twistor models and twistor-strings can also be defined on curved
backgrounds built from curved twistor spaces. These are equivalent to curved
but self-dual Yang-Mills or Einstein backgrounds, which are integrable.
Exploiting this integrability, one can construct amplitudes of arbitrary
multiplicity, [83, 173, 174] with explicit formulæ for certain classes of
backgrounds.
### 6.2 Relationship with conventional, null and chiral strings
All the amplitude formulæ we have discussed localize on the solutions to the
scattering equations. These equations were first obtained by Fairlie [175,
176] in a semiclassical study of string solutions, and made famous in the work
of Gross and Mende [177, 178], where they were shown to govern the string path
integral in the limit of high energy scattering at fixed angle ($s\gg
1/\alpha^{\prime}$). The scattering equations thus play a prominent role both
in the $\alpha^{\prime}\rightarrow\infty$ limit and in the field theory limit
$\alpha^{\prime}\rightarrow 0$ , albeit indirectly via the ambitwistor string.
This clash, with the scattering equations appearing in both the low tension
and high tension field theory limit, has so far impeded attempts to connect
the ambitwistor string directly with conventional strings. In this context, it
is interesting to note that the $\alpha^{\prime}\rightarrow\infty$ limit can
also be understood as a null limit in which the worldsheet becomes null ruled
by null geodesics [124].
More generally Siegel proposed that so-called chiral string theories could be
obtained by flipping certain worldsheet boundary conditions [123], so that
both left-moving and right-moving modes in the conventional string become
holomorphic on the worldsheet in the sectorized chiral string. A bosonic such
model had already been introduced in [179], and pure spinor versions discussed
in [180]. In these models both the left moving and right-moving Virasoro
constraints become holomorphic but commute with each other and each satisfies
the holomorphic Virasoro algebra. Ambitwistor have a similar character in the
sense that $X^{\mu}$ and $\int^{\sigma}P^{\mu}$ can be thought of as being
independent holomorphic functions on the worldsheet playing roles as different
combinations of left and right movers. However, the ambitwistor-string only
contains one holomorphic Virasoro generator, $P\cdot\partial X$. The gauged
constraint $P^{2}$ is analogous to difference between the two Virasoro
generators, but has trivial OPE with itself, even on a curved background. This
reflects the degeneration of the two copies of the Virasoro algebra to a
Galilean Conformal Algebra [124, 125]. On the other hand, in the model of
[179], the gauged constraint $P^{2}$ is replaced by a more general quadratic
expression in $P$ and $\partial X$
$P^{2}\rightarrow\mathcal{H}:=A_{\mu\nu}(X)P^{\mu}P^{\nu}+B_{\mu\nu}P^{\mu}\partial
X^{\nu}+C(x)_{\mu\nu}\partial X^{\mu}\partial X^{\nu}\,,$ (6.1)
that is constrained to obey a nontrivial OPE. The imposition of these OPEs
yields field equations for the background with a finite number of
$\alpha^{\prime}$ corrections. However, since the OPE of $\mathcal{H}$ with
itself is nontrivial, there no longer appears to be a reduction to ambitwistor
space nor localization on the scattering equations. The connections between
sectorized strings, null strings and ambitwistor strings are now well studied
in different models [124, 125, 180, 181], see also [182, 183]. For example,
versions of T-duality become possible in sectorized strings whereas they are
not in the ambitwistor-string [184].
The amplitude formulæ to which such sectorized chiral strings give rise appear
to be problematic. The Koba-Nielsen factor of the conventional string,
consisting of a product $\prod_{i<j}|\sigma_{ij}|^{\alpha^{\prime}s_{ij}}$, is
replaced by a product
$\prod_{i<j}(\sigma_{ij}/\bar{\sigma}_{ij})^{\alpha^{\prime}s_{ij}}$ so that
the branching would seem to make the contour prescription problematic. This is
resolved in the work of Mizera who uses twisted cohomology and residues to
define the amplitudes of these theories [185]. As
$\alpha^{\prime}\rightarrow\infty$ he shows localization on the scattering
equations.
### 6.3 Celestial holography and soft theorems
Celestial holography [186] seeks to understand the S-matrices of massless
theories by formulating these theories on the conformal boundary of
asymptotically flat space-times, $\mathscr{I}$, the _light-cone at infinity_.
This approach emerged from the study of connections between soft theorems for
amplitudes and asymptotic BMS symmetries of space-time [187, 188], and aims to
establish a holographic dictionary for the S-matrix from $\mathscr{I}$ or the
celestial sphere [189, 190]. In an asymptotically simple space-time, all light
rays reach $\mathscr{I}$ and we can represent ambitwistor space
$\mathbb{A}=T^{*}\mathscr{I}_{\mathbb{C}}$ as the cotangent bundle of the
complexification of null infinity. Both the RNS and 4d twistor ambitwistor
models can be expressed in this representation [191, 192] and from this
perspective, the connection between soft theorems and asymptotic symmetries
can be understood directly at the level of the vertex operators as generators
of supertranslations and super-rotations in their soft limits. The ambitwistor
strings then provide an underpinning theory for the generation of the S-matrix
now seen as being based at $\mathscr{I}$. The recently discovered
gravitational ‘$w_{1+\infty}$-symmetry’ [193] can be seen in this formulation
as arising directly from the geometry of the asymptotic twistor space at
$\mathscr{I}$ [194].
## Acknowledgements
This work was supported by the European Union’s Horizon 2020 research and
innovation programme under the Marie Skłodowska-Curie grant agreement No.
764850 “SAGEX”. YG is grateful for the support from the CUniverse research
promotion project “Toward World-class Fundamental Physics” of Chulalongkorn
University (grant reference CUAASC). The research of LJM is supported by the
STFC grant ST/T000864/1.
## Appendix A Penrose transform
The Penrose transform for a general ambitwistor space $\mathbb{P}\mathbb{A}$
maps spin $s$ fields on space-time to elements of
$H^{1}(\mathbb{P}\mathbb{A},\mathcal{O}(s-1))$, with the classes corresponding
to spin $s$ plane-waves of the form
$\epsilon_{\mu_{1}}\ldots\epsilon_{\mu_{s}}\rme^{ik\cdot X}$ mapping to
$(\epsilon\cdot P)^{s}\,\bar{\delta}(k\cdot P)\,\rme^{ik\cdot x}$.
Briefly, this follows by taking a cohomology class $\phi\in
H^{1}(\mathbb{P}\mathbb{A},\mathcal{O}(s-1))$ on $\mathbb{P}\mathbb{A}$,
pulling it back to $p^{*}\phi$ on $PT^{*}M|_{P^{2}=0}$, the lightcone inside
the projective cotangent bundle, where it becomes necessarily trivial as there
is no first cohomology on this space for the given homogeneity weights. Thus
we can find $g$ of weight $s-1$ on $T^{*}M|_{P^{2}}=0$ such that
$p^{*}\phi=\bar{\partial}g\,,\qquad\mbox{e.g.\ for $\phi$ in \eqref{drep}
}\qquad g=(\epsilon\cdot P)^{s}\frac{\rme^{ik\cdot X}}{k\cdot P}\,.$ (A.1)
Then we can obtain the field via
$(\epsilon\cdot P)^{s}\,\rme^{ik\cdot x}=P\cdot\partial_{X}g\,.$ (A.2)
Since $g$ is defined up to the gauge freedom, $\delta g=(\epsilon\cdot
P)^{s-1}\rme^{ik\cdot X}$ (or indeed some a general global holomorphic
function of $P$), we have the usual gauge freedom
$\delta(\epsilon_{\mu_{1}}\ldots\epsilon_{\mu_{s}}\rme^{ik\cdot
X})=k_{(\mu_{1}}\epsilon_{\mu_{2}}\ldots\epsilon_{\mu_{s})}\,\rme^{ik\cdot
X}.$ (A.3)
This can be proven more abstractly as follows. The Penrose transform is the
connecting map
$H^{0}(PT^{*}M_{P^{2}=0},\mathcal{O}(s))/P\cdot\partial_{X}\left(H^{0}(PT^{*}M_{P^{2}=0},\mathcal{O}(s-1))\right)\stackrel{{\scriptstyle\delta}}{{\longrightarrow}}H^{1}(\mathbb{P}\mathbb{A},\mathcal{O}(s-1)),,$
(A.4)
from the long exact sequence in cohomology arising from the short exact
sequence:
$0\rightarrow\mathcal{O}(s-1)_{\mathbb{P}\mathbb{A}}\rightarrow\mathcal{O}(s-1)_{T^{*}M_{P^{2}=0}}\xrightarrow{P\cdot\partial_{X}}\mathcal{O}(s)_{T^{*}M_{P^{2}=0}}\rightarrow
0\,,.$ (A.5)
that defines the functions on $\mathbb{P}\mathbb{A}$, see [34, 22] for a full
discussion.
## Appendix B Conformal Field Theory basics
In this appendix, we present a lightning review of some aspects of chiral two-
dimensional conformal field theories (CFTs) particularly relevant for the
ambitwistor string. For a more extensive introduction, we refer the interested
reader to the multitude of excellent textbooks, for example [28, 29, 118, 195,
196, 197].
All ambitwistor string models are two-dimensional CFTs with a local action in
terms of a set of fields defined over a closed Riemann surface $\Sigma$,
referred to as the worldsheet. These fields can be characterized by their
statistics (bosonic vs fermionic) and their conformal weight, as well as
additional quantum numbers such as the spin structure at higher genus. The
conformal weight of a field $\Phi$ is a pair of half-integers
$(h,\bar{h})\in\mathbb{Z}/2\mathbb{Z}\times\mathbb{Z}/2\mathbb{Z}$ labeling
the transformation properties of $\Phi$ under two-dimensional conformal
transformations. 202020A field is called _primary_ if its conformal weights
are well-defined. Since 2d local conformal transformations are equivalent to
holomorphic coordinate transformations, we can identify the conformal weight
with the form degree. This means that $\Phi$ is section of $(h,\bar{h})$
powers fo the holomorphic and antiholomorphic canonical bundles
$K_{\Sigma}^{h}\otimes\overline{K}_{\Sigma}^{\bar{h}}$, i.e.
$\Phi\in\Omega^{0}(\Sigma,K_{\Sigma}^{h}\otimes\overline{K}_{\Sigma}^{\bar{h}})\,.$
(B.1)
Here, negative weights are to be interpreted as sections of the respective
tangent bundles, using the isomorphism $K_{\Sigma}^{-1}\cong T_{\Sigma}$.
Equivalently, $\Phi$ may be expressed locally in affine worldsheet coordinates
$z$ as
$\Phi(z,\bar{z})=\phi_{h,\bar{h}}\,(dz)^{h}\,(d\bar{z})^{\bar{h}}\,.$ (B.2)
In the ambitwistor string, we will only encounter _chiral_ CFTs, known as
$\beta\gamma$-systems (for bosons) or $bc$-systems (for fermions). We will
describe these jointly below, keeping track of the statistics via a variable
$\epsilon$, with $\epsilon=-1$ for fermionic statistics, and $\epsilon=1$ for
bosons. A chiral CFT is then defined by the action
$S=\frac{1}{2\pi}\int b\bar{\partial}c\,,$ (B.3)
in conformal gauge. The fields $b$ has conformal weight $(h,0)$, giving
conformal weight $(1-h,0)$ to the conjugate field $c$. Fields such as these
with conformal weight $\bar{h}=0$ are often referred to as ‘left-moving’. The
OPE between these conjugate fields is
$c(z)\,b(w)\sim\frac{1}{z-w}\,,\qquad\qquad
b(z)\,c(w)\sim-\frac{\epsilon}{z-w}\,.$ (B.4)
A standard calculation gives the holomorphic stress-energy tensor
$T_{bc}=-h\,b\partial c+(1-h)\,(\partial b)c\,.$ (B.5)
From this expression, we find the central charge anomaly as (twice) the
coefficient of the fourth order pole in the $T(z)T(w)$ OPE,
$\mathfrak{c}=2\epsilon\left(6h^{2}-6h+1\right)\,.$ (B.6)
Finally we note the following useful formula for the number of zero modes
$n_{b}$ and $n_{c}$ of the two fields on a Riemann surface of genus $g$,
derived via the Riemann-Roch theorem,
$n_{c}-n_{b}=\frac{1}{2}\left(2h-1\right)\chi\,,\qquad\mathrm{with}\;\;\chi=2(1-g)\,.$
(B.7)
## Bibliography
## References
* [1] E. Witten, _Perturbative gauge theory as a string theory in twistor space_ , _Commun.Math.Phys._ 252 (2004) 189–258, [hep-th/0312171].
* [2] N. Berkovits, _An Alternative string theory in twistor space for N=4 superYang-Mills_ , _Phys.Rev.Lett._ 93 (2004) 011601, [hep-th/0402045].
* [3] N. Berkovits and E. Witten, _Conformal supergravity in twistor-string theory_ , _JHEP_ 0408 (2004) 009, [hep-th/0406051].
* [4] V. P. Nair, _A Current Algebra for Some Gauge Theory Amplitudes_ , _Phys. Lett._ B214 (1988) 215–218.
* [5] R. Roiban, M. Spradlin and A. Volovich, _On the tree level S matrix of Yang-Mills theory_ , _Phys.Rev._ D70 (2004) 026009, [hep-th/0403190].
* [6] E. Witten, _Parity invariance for strings in twistor space_ , _Adv.Theor.Math.Phys._ 8 (2004) 779–796, [hep-th/0403199].
* [7] L. Dolan and P. Goddard, _Tree and Loop Amplitudes in Open Twistor String Theory_ , _JHEP_ 0706 (2007) 005, [hep-th/0703054].
* [8] Z. Bern, J. J. M. Carrasco and H. Johansson, _Perturbative Quantum Gravity as a Double Copy of Gauge Theory_ , _Phys.Rev.Lett._ 105 (2010) 061602, [1004.0476].
* [9] Z. Bern, J. J. Carrasco, M. Chiodaroli, H. Johansson and R. Roiban, _The SAGEX Review on Scattering Amplitudes, Chapter 2: An Invitation to Color-Kinematics Duality and the Double Copy_ , 2203.13013.
* [10] H. Kawai, D. C. Lewellen and S. H. H. Tye, _A Relation Between Tree Amplitudes of Closed and Open Strings_ , _Nucl. Phys._ B269 (1986) 1.
* [11] F. Cachazo, _Fundamental BCJ Relation in N=4 SYM From The Connected Formulation_ , 1206.5970.
* [12] F. Cachazo and Y. Geyer, _A ‘Twistor String’ Inspired Formula For Tree-Level Scattering Amplitudes in N=8 SUGRA_ , 1206.6511.
* [13] A. Hodges, _A simple formula for gravitational MHV amplitudes_ , 1204.1930.
* [14] F. Cachazo and D. Skinner, _Gravity from Rational Curves in Twistor Space_ , _Phys.Rev.Lett._ 110 (2013) 161301, [1207.0741].
* [15] F. Cachazo, L. Mason and D. Skinner, _Gravity in Twistor Space and its Grassmannian Formulation_ , _SIGMA_ 10 (2014) 051, [1207.4712].
* [16] D. Skinner, _Twistor strings for $\mathcal{N}$ = 8 supergravity_, _JHEP_ 04 (2020) 047, [1301.0868].
* [17] F. Cachazo, S. He and E. Y. Yuan, _Scattering equations and Kawai-Lewellen-Tye orthogonality_ , _Phys. Rev._ D90 (2014) 065001, [1306.6575].
* [18] F. Cachazo, S. He and E. Y. Yuan, _Scattering of Massless Particles in Arbitrary Dimensions_ , _Phys.Rev.Lett._ 113 (2014) 171601, [1307.2199].
* [19] F. Cachazo, S. He and E. Y. Yuan, _Scattering of Massless Particles: Scalars, Gluons and Gravitons_ , _JHEP_ 1407 (2014) 033, [1309.0885].
* [20] F. Cachazo, S. He and E. Y. Yuan, _Einstein-Yang-Mills Scattering Amplitudes From Scattering Equations_ , _JHEP_ 01 (2015) 121, [1409.8256].
* [21] F. Cachazo, S. He and E. Y. Yuan, _Scattering Equations and Matrices: From Einstein To Yang-Mills, DBI and NLSM_ , _JHEP_ 07 (2015) 149, [1412.3479].
* [22] L. Mason and D. Skinner, _Ambitwistor strings and the scattering equations_ , _JHEP_ 1407 (2014) 048, [1311.2564].
* [23] E. Witten, _An Interpretation of Classical Yang-Mills Theory_ , _Phys.Lett._ B77 (1978) 394.
* [24] J. Isenberg, P. Yasskin and P. Green, _Non selfdual Gauge Fields_ , _Phys.Lett._ B78 (1978) 462–464.
* [25] R. S. Ward, _On Selfdual gauge fields_ , _Phys. Lett. A_ 61 (1977) 81–82.
* [26] Y. Geyer, A. E. Lipstein and L. J. Mason, _Ambitwistor Strings in Four Dimensions_ , _Phys. Rev. Lett._ 113 (2014) 081602, [1404.6219].
* [27] E. Casali, Y. Geyer, L. Mason, R. Monteiro and K. A. Roehrig, _New Ambitwistor String Theories_ , _JHEP_ 11 (2015) 038, [1506.08771].
* [28] M. B. Green, J. H. Schwarz and E. Witten, _Superstring Theory_ , vol. 1. CUP, 1987.
* [29] M. B. Green, J. H. Schwarz and E. Witten, _Superstring Theory_ , vol. 2. CUP, 1987.
* [30] E. Witten, _Superstring Perturbation Theory Revisited_ , 1209.5461.
* [31] T. Adamo and L. Mason, _Conformal and Einstein gravity from twistor actions_ , _Class.Quant.Grav._ 31 (2014) 045014, [1307.5043].
* [32] K. Ohmori, _Worldsheet Geometries of Ambitwistor String_ , _JHEP_ 06 (2015) 075, [1504.02675].
* [33] C. LeBrun, _Spaces of complex null geodesics in Complex Riemmannian Geometry_ , _Transactions of the AMS._ 278 (1983) 209–31.
* [34] R. Baston and L. Mason, _Conformal gravity, the Einstein equations and spaces of complex null geodesics_ , _Class.Quant.Grav._ 4 (1987) 815–826.
* [35] E. Witten, _Two-dimensional models with (0,2) supersymmetry: Perturbative aspects_ , _Adv. Theor. Math. Phys._ 11 (2007) 1–63, [hep-th/0504078].
* [36] R. Britto, F. Cachazo and B. Feng, _New recursion relations for tree amplitudes of gluons_ , _Nucl.Phys._ B715 (2005) 499–522, [hep-th/0412308].
* [37] R. Britto, F. Cachazo, B. Feng and E. Witten, _Direct proof of tree-level recursion relation in Yang-Mills theory_ , _Phys.Rev.Lett._ 94 (2005) 181602, [hep-th/0501052].
* [38] F. Cachazo and P. Svrcek, _Tree level recursion relations in general relativity_ , hep-th/0502160.
* [39] J. Bedford, A. Brandhuber, B. J. Spence and G. Travaglini, _A Recursion relation for gravity amplitudes_ , _Nucl.Phys._ B721 (2005) 98–110, [hep-th/0502146].
* [40] A. Brandhuber, J. Plefka and G. Travaglini, _The SAGEX Review on Scattering Amplitudes, Chapter 1: Modern Fundamentals of Amplitudes_ , 2203.13012.
* [41] N. Arkani-Hamed and J. Kaplan, _On Tree Amplitudes in Gauge Theory and Gravity_ , _JHEP_ 04 (2008) 076, [0801.2385].
* [42] N. Arkani-Hamed and J. Kaplan, _On Tree Amplitudes in Gauge Theory and Gravity_ , _JHEP_ 0804 (2008) 076, [0801.2385].
* [43] T. Cohen, H. Elvang and M. Kiermaier, _On-shell constructibility of tree amplitudes in general field theories_ , _JHEP_ 04 (2011) 053, [1010.0257].
* [44] H. Elvang and Y.-t. Huang, _Scattering Amplitudes_ , 1308.1697.
* [45] H. Elvang and Y.-t. Huang, _Scattering Amplitudes in Gauge Theory and Gravity_. Cambridge University Press, 4, 2015.
* [46] C. Cheung, _TASI Lectures on Scattering Amplitudes_ , pp. 571–623. 2018\. 1708.03872. 10.1142/9789813233348-0008.
* [47] B. Feng and M. Luo, _An Introduction to On-shell Recursion Relations_ , _Front. Phys. (Beijing)_ 7 (2012) 533–575, [1111.5759].
* [48] L. Dolan and P. Goddard, _Proof of the Formula of Cachazo, He and Yuan for Yang-Mills Tree Amplitudes in Arbitrary Dimension_ , _JHEP_ 1405 (2014) 010, [1311.5200].
* [49] P. Deligne and D. Mumford, _The irreducibility of the space of curves of given genus_ , _Publ. Math. IHES_ (1969) 75–110.
* [50] H. Johansson and J. Nohle, _Conformal Gravity from Gauge Theory_ , 1707.02965.
* [51] N. Berkovits and M. Lize, _Field theory actions for ambitwistor string and superstring_ , _JHEP_ 09 (2018) 097, [1807.07661].
* [52] Y. Geyer, L. Mason and D. Skinner, _Ambitwistor strings in six and five dimensions_ , _JHEP_ 08 (2021) 153, [2012.15172].
* [53] T. Azevedo and O. T. Engelund, _Ambitwistor formulations of R 2 gravity and (DF)2 gauge theories_, _JHEP_ 11 (2017) 052, [1707.02192].
* [54] C. Cheung, K. Kampf, J. Novotny, C.-H. Shen and J. Trnka, _On-Shell Recursion Relations for Effective Field Theories_ , _Phys. Rev. Lett._ 116 (2016) 041601, [1509.03309].
* [55] H. Luo and C. Wen, _Recursion relations from soft theorems_ , _JHEP_ 03 (2016) 088, [1512.06801].
* [56] N. Berkovits, _Infinite Tension Limit of the Pure Spinor Superstring_ , _JHEP_ 1403 (2014) 017, [1311.4156].
* [57] H. Gomez and E. Y. Yuan, _N-point tree-level scattering amplitude in the new Berkovits‘ string_ , _JHEP_ 04 (2014) 046, [1312.5485].
* [58] O. Chandia and B. C. Vallilo, _Ambitwistor superstring in the Green–Schwarz formulation_ , _Eur. Phys. J. C_ 77 (2017) 473, [1612.01806].
* [59] N. Berkovits, E. Casali, M. Guillen and L. Mason, _Notes on the $D=11$ pure spinor superparticle_, _JHEP_ 08 (2019) 178, [1905.03737].
* [60] M. Guillen, _Notes on the 11D pure spinor wordline vertex operators_ , _JHEP_ 08 (2020) 122, [2006.06022].
* [61] Z. Bern, J. Carrasco and H. Johansson, _New Relations for Gauge-Theory Amplitudes_ , _Phys.Rev._ D78 (2008) 085011, [0805.3993].
* [62] C.-H. Fu, Y.-J. Du, R. Huang and B. Feng, _Expansion of Einstein-Yang-Mills Amplitude_ , _JHEP_ 09 (2017) 021, [1702.08158].
* [63] S. Mizera, _Kinematic Jacobi Identity is a Residue Theorem: Geometry of Color-Kinematics Duality for Gauge and Gravity Amplitudes_ , _Phys. Rev. Lett._ 124 (2020) 141601, [1912.03397].
* [64] A. Edison and F. Teng, _Efficient Calculation of Crossing Symmetric BCJ Tree Numerators_ , _JHEP_ 12 (2020) 138, [2005.03638].
* [65] S. He, L. Hou, J. Tian and Y. Zhang, _Kinematic numerators from the worldsheet: cubic trees from labelled trees_ , _JHEP_ 08 (2021) 118, [2103.15810].
* [66] H. Frost and L. Mason, _Lie Polynomials and a Twistorial Correspondence for Amplitudes_ , 1912.04198.
* [67] C. R. Mafra, _Planar binary trees in scattering amplitudes_ , 11, 2020, 2011.14413, DOI.
* [68] H. Frost, C. R. Mafra and L. Mason, _A Lie bracket for the momentum kernel_ , 2012.00519.
* [69] N. Arkani-Hamed, T.-C. Huang and Y.-t. Huang, _Scattering amplitudes for all masses and spins_ , _JHEP_ 11 (2021) 070, [1709.04891].
* [70] H. Frost, _Universal aspects of perturbative gauge theory_ , Ph.D. thesis, Oxford U., 2020.
* [71] H. Frost, _The Algebraic Structure of the KLT Relations for Gauge and Gravity Tree Amplitudes_ , _SIGMA_ 17 (2021) 101, [2111.07257].
* [72] T. Adamo, _Lectures on twistor theory_ , _PoS_ Modave2017 (2018) 003, [1712.02196].
* [73] T. Adamo, M. Bullimore, L. Mason and D. Skinner, _Scattering Amplitudes and Wilson Loops in Twistor Space_ , _J.Phys.A_ A44 (2011) 454008, [1104.2890].
* [74] S. A. Huggett and K. P. Tod, _An Introduction to Twistor Theory_. Cambridge University Press, 1986, 10.1017/CBO9780511624018.
* [75] N. Woodhouse, _Real methods in twistor theory_ , _Class.Quant.Grav._ 2 (1985) 257–291.
* [76] Y. Geyer, _Ambitwistor Strings: Worldsheet Approaches to perturbative Quantum Field Theories_ , Ph.D. thesis, Oxford U., Inst. Math., 2016. 1610.04525.
* [77] D. Skinner, _A Direct Proof of BCFW Recursion for Twistor-Strings_ , _JHEP_ 1101 (2011) 072, [1007.0195].
* [78] N. Berkovits and L. Motl, _Cubic twistorial string field theory_ , _JHEP_ 0404 (2004) 056, [hep-th/0403187].
* [79] T. Adamo, _Worldsheet factorization for twistor-strings_ , _JHEP_ 04 (2014) 080, [1310.8602].
* [80] R. Penrose and W. Rindler, _Spinors and Space-Time: Spinor And Twistor Methods In Space-time Geometry_ , vol. 2. Cambridge University Press, 1986.
* [81] M. Wolf, _Self-Dual Supergravity and Twistor Theory_ , _Class.Quant.Grav._ 24 (2007) 6287–6328, [0705.1422].
* [82] T. Adamo, _Gravity with a cosmological constant from rational curves_ , _JHEP_ 11 (2015) 098, [1508.02554].
* [83] T. Adamo, L. Mason and A. Sharma, _Twistor sigma models for quaternionic geometry and graviton scattering_ , 2103.16984.
* [84] K. A. Roehrig, _Chiral splitting and $\mathcal{N}=4$ Einstein-Yang-Mills tree amplitudes in 4d_, _JHEP_ 08 (2017) 033, [1705.09315].
* [85] T. Adamo, E. Casali, K. A. Roehrig and D. Skinner, _On tree amplitudes of supersymmetric Einstein-Yang-Mills theory_ , _JHEP_ 12 (2015) 177, [1507.02207].
* [86] G. Albonico, Y. Geyer and L. Mason, _Recursion and worldsheet formulae for 6d superamplitudes_ , _JHEP_ 08 (2020) 066, [2001.05928].
* [87] S. He and Y. Zhang, _Connected formulas for amplitudes in standard model_ , _JHEP_ 03 (2017) 093, [1607.02843].
* [88] Y. Zhang, _CHY formulae in 4d_ , _JHEP_ 07 (2017) 069, [1610.05205].
* [89] Y. Geyer and L. Mason, _Polarized Scattering Equations for 6D Superamplitudes_ , _Phys. Rev. Lett._ 122 (2019) 101601, [1812.05548].
* [90] Y. Geyer and L. Mason, _Supersymmetric S-matrices from the worldsheet in 10 \ & 11d_, _Phys. Lett. B_ 804 (2020) 135361, [1901.00134].
* [91] Y. Geyer, R. Monteiro and R. Stark-Muchão, _Superstring Loop Amplitudes from the Field Theory Limit_ , _Phys. Rev. Lett._ 127 (2021) 211603, [2106.03968].
* [92] M. Heydeman, J. H. Schwarz and C. Wen, _M5-Brane and D-Brane Scattering Amplitudes_ , _JHEP_ 12 (2017) 003, [1710.02170].
* [93] M. Heydeman, J. H. Schwarz, C. Wen and S.-Q. Zhang, _All Tree Amplitudes of 6D $(2,0)$ Supergravity: Interacting Tensor Multiplets and the $K3$ Moduli Space_, _Phys. Rev. Lett._ 122 (2019) 111604, [1812.06111].
* [94] F. Cachazo, A. Guevara, M. Heydeman, S. Mizera, J. H. Schwarz and C. Wen, _The S Matrix of 6D Super Yang-Mills and Maximal Supergravity from Rational Maps_ , _JHEP_ 09 (2018) 125, [1805.11111].
* [95] N. Berkovits, M. Guillen and L. Mason, _Supertwistor description of ambitwistor strings_ , _JHEP_ 01 (2020) 020, [1908.06899].
* [96] R. A. Reid-Edwards and D. A. Riccombeni, _A Superstring Field Theory for Supergravity_ , _JHEP_ 09 (2017) 103, [1701.05495].
* [97] D. G. Sepúlveda and M. Guillen, _A Pure Spinor Twistor Description of Ambitwistor Strings_ , 2006.06025.
* [98] R. Penrose, _Twistors and Particles: An Outline_ , in _Feldafing Conference of the Max-Planck Inst. on Quantum Theory and the Structure of Space-time, 1974_ , pp. 129–145, Carl Hanser Verlag, Munich, 1975.
* [99] G. Albonico, Y. Geyer and L. Mason, _From twistor-particle models to massive amplitudes_ , 2203.08087.
* [100] C. Wen and S.-Q. Zhang, _D3-Brane Loop Amplitudes from M5-Brane Tree Amplitudes_ , _JHEP_ 07 (2020) 098, [2004.02735].
* [101] M. Bullimore, L. Mason and D. Skinner, _Twistor-Strings, Grassmannians and Leading Singularities_ , _JHEP_ 1003 (2010) 070, [0912.0539].
* [102] N. Arkani-Hamed, J. Bourjaily, F. Cachazo and J. Trnka, _Unification of Residues and Grassmannian Dualities_ , _JHEP_ 01 (2011) 049, [0912.4912].
* [103] N. Arkani-Hamed, F. Cachazo, C. Cheung and J. Kaplan, _A Duality For The S Matrix_ , _JHEP_ 1003 (2010) 020, [0907.5418].
* [104] J. H. Schwarz and C. Wen, _Unified Formalism for 6D Superamplitudes Based on a Symplectic Grassmannian_ , _JHEP_ 08 (2019) 125, [1907.03485].
* [105] T. Adamo, E. Casali and D. Skinner, _Ambitwistor strings and the scattering equations at one loop_ , _JHEP_ 1404 (2014) 104, [1312.3828].
* [106] E. Casali and P. Tourkine, _Infrared behaviour of the one-loop scattering equations and supergravity integrands_ , _JHEP_ 1504 (2015) 013, [1412.3787].
* [107] T. Adamo and E. Casali, _Scattering equations, supergravity integrands, and pure spinors_ , _JHEP_ 1505 (2015) 120, [1502.06826].
* [108] Y. Geyer and R. Monteiro, _Two-Loop Scattering Amplitudes from Ambitwistor Strings: from Genus Two to the Nodal Riemann Sphere_ , _JHEP_ 11 (2018) 008, [1805.05344].
* [109] J. A. Farrow and A. E. Lipstein, _From 4d Ambitwistor Strings to On Shell Diagrams and Back_ , _JHEP_ 07 (2017) 114, [1705.07087].
* [110] E. D’Hoker and D. H. Phong, _Two loop superstrings. 1. Main formulas_ , _Phys. Lett._ B529 (2002) 241–255, [hep-th/0110247].
* [111] E. D’Hoker and D. H. Phong, _Two loop superstrings. 2. The Chiral measure on moduli space_ , _Nucl. Phys._ B636 (2002) 3–60, [hep-th/0110283].
* [112] E. D’Hoker and D. H. Phong, _Two loop superstrings. 3. Slice independence and absence of ambiguities_ , _Nucl. Phys._ B636 (2002) 61–79, [hep-th/0111016].
* [113] E. D’Hoker and D. H. Phong, _Two loop superstrings 4: The Cosmological constant and modular forms_ , _Nucl. Phys._ B639 (2002) 129–181, [hep-th/0111040].
* [114] E. D’Hoker and D. H. Phong, _Two-loop superstrings. V. Gauge slice independence of the N-point function_ , _Nucl. Phys._ B715 (2005) 91–119, [hep-th/0501196].
* [115] E. D’Hoker and D. H. Phong, _Two-loop superstrings VI: Non-renormalization theorems and the 4-point function_ , _Nucl. Phys._ B715 (2005) 3–90, [hep-th/0501197].
* [116] E. D’Hoker and D. H. Phong, _Lectures on two loop superstrings_ , _Conf. Proc._ C0208124 (2002) 85–123, [hep-th/0211111].
* [117] E. D’Hoker and D. H. Phong, _The Geometry of String Perturbation Theory_ , _Rev. Mod. Phys._ 60 (1988) 917.
* [118] J. Polchinski, _String theory. Vol. 1: An introduction to the bosonic string_. Cambridge University Press, 2007.
* [119] N. Seiberg and E. Witten, _Spin Structures in String Theory_ , _Nucl. Phys. B_ 276 (1986) 272.
* [120] E. D’Hoker and D. H. Phong, _Conformal Scalar Fields and Chiral Splitting on Superriemann Surfaces_ , _Commun. Math. Phys._ 125 (1989) 469.
* [121] P. Tourkine, _Integrands and loop momentum in string and field theory_ , _Phys. Rev. D_ 102 (2020) 026006, [1901.02432].
* [122] N. Kalyanapuram, _On Chiral Splitting and the Ambitwistor String_ , _Phys. Rev. D_ 104 (2021) 086027, [2103.08584].
* [123] W. Siegel, _Amplitudes for Left-Handed Strings_ , 1512.02569.
* [124] E. Casali and P. Tourkine, _On the null origin of the ambitwistor string_ , _JHEP_ 11 (2016) 036, [1606.05636].
* [125] E. Casali, Y. Herfray and P. Tourkine, _The complex null string, Galilean conformal algebra and scattering equations_ , _JHEP_ 10 (2017) 164, [1707.09900].
* [126] M. Yu, C. Zhang and Y.-Z. Zhang, _One loop amplitude from null string_ , _JHEP_ 06 (2017) 051, [1704.01290].
* [127] Y. Geyer, L. Mason, R. Monteiro and P. Tourkine, _Loop Integrands for Scattering Amplitudes from the Riemann Sphere_ , _Phys. Rev. Lett._ 115 (2015) 121603, [1507.00321].
* [128] Y. Geyer, L. Mason, R. Monteiro and P. Tourkine, _One-loop amplitudes on the Riemann sphere_ , _JHEP_ 03 (2016) 114, [1511.06315].
* [129] Y. Geyer, L. Mason, R. Monteiro and P. Tourkine, _Two-Loop Scattering Amplitudes from the Riemann Sphere_ , _Phys. Rev._ D94 (2016) 125029, [1607.08887].
* [130] P. Tourkine and P. Vanhove, _One-loop four-graviton amplitudes in ${\cal N}=4$ supergravity models_, _Phys. Rev._ D87 (2013) 045001, [1208.1255].
* [131] S. He and E. Y. Yuan, _One-loop Scattering Equations and Amplitudes from Forward Limit_ , _Phys. Rev._ D92 (2015) 105004, [1508.06027].
* [132] F. Cachazo, S. He and E. Y. Yuan, _One-Loop Corrections from Higher Dimensional Tree Amplitudes_ , _JHEP_ 08 (2016) 008, [1512.05001].
* [133] B. Feng, _CHY-construction of Planar Loop Integrands of Cubic Scalar Theory_ , _JHEP_ 05 (2016) 061, [1601.05864].
* [134] B. Feng and C. Hu, _One-loop CHY-Integrand of Bi-adjoint Scalar Theory_ , _JHEP_ 02 (2020) 187, [1912.12960].
* [135] F. Porkert and O. Schlotterer, _One-loop amplitudes in Einstein-Yang-Mills from forward limits_ , 2201.12072.
* [136] R. P. Feynman, _Quantum theory of gravitation_ , _Acta Phys. Polon._ 24 (1963) 697–722.
* [137] S. Caron-Huot, _Loops and trees_ , _JHEP_ 05 (2011) 080, [1007.3224].
* [138] C. Baadsgaard, N. E. J. Bjerrum-Bohr, J. L. Bourjaily, S. Caron-Huot, P. H. Damgaard and B. Feng, _New Representations of the Perturbative S-Matrix_ , _Phys. Rev. Lett._ 116 (2016) 061601, [1509.02169].
* [139] S. Abreu, R. Britto and C. Duhr, _The SAGEX Review on Scattering Amplitudes, Chapter 3: Mathematical structures in Feynman integrals_ , 2203.13014.
* [140] H. Gomez, _Quadratic Feynman Loop Integrands From Massless Scattering Equations_ , _Phys. Rev. D_ 95 (2017) 106006, [1703.04714].
* [141] H. Gomez, C. Lopez-Arcos and P. Talavera, _One-loop Parke-Taylor factors for quadratic propagators from massless scattering equations_ , _JHEP_ 10 (2017) 175, [1707.08584].
* [142] N. Ahmadiniaz, H. Gomez and C. Lopez-Arcos, _Non-planar one-loop Parke-Taylor factors in the CHY approach for quadratic propagators_ , _JHEP_ 05 (2018) 055, [1802.00015].
* [143] J. A. Farrow, Y. Geyer, A. E. Lipstein, R. Monteiro and R. Stark-Muchão, _Propagators, BCFW recursion and new scattering equations at one loop_ , _JHEP_ 10 (2020) 074, [2007.00623].
* [144] J. J. Carrasco and H. Johansson, _Five-Point Amplitudes in N=4 Super-Yang-Mills Theory and N=8 Supergravity_ , _Phys.Rev._ D85 (2012) 025006, [1106.4711].
* [145] Z. Bern, J. J. M. Carrasco, L. J. Dixon, H. Johansson and R. Roiban, _Simplifying Multiloop Integrands and Ultraviolet Divergences of Gauge Theory and Gravity Amplitudes_ , _Phys. Rev._ D85 (2012) 105014, [1201.5366].
* [146] Z. Bern, S. Davies, T. Dennen and Y.-t. Huang, _Absence of Three-Loop Four-Point Divergences in N=4 Supergravity_ , _Phys. Rev. Lett._ 108 (2012) 201301, [1202.3423].
* [147] Z. Bern, S. Davies and T. Dennen, _The Ultraviolet Structure of Half-Maximal Supergravity with Matter Multiplets at Two and Three Loops_ , _Phys. Rev._ D88 (2013) 065007, [1305.4876].
* [148] Z. Bern, S. Davies, T. Dennen, A. V. Smirnov and V. A. Smirnov, _Ultraviolet Properties of N=4 Supergravity at Four Loops_ , _Phys. Rev. Lett._ 111 (2013) 231302, [1309.2498].
* [149] Z. Bern, S. Davies and T. Dennen, _Enhanced ultraviolet cancellations in $\mathcal{N}=5$ supergravity at four loops_, _Phys. Rev._ D90 (2014) 105011, [1409.3089].
* [150] C. R. Mafra and O. Schlotterer, _Two-loop five-point amplitudes of super Yang-Mills and supergravity in pure spinor superspace_ , _JHEP_ 10 (2015) 124, [1505.02746].
* [151] H. Johansson, G. Kalin and G. Mogull, _Two-loop supersymmetric QCD and half-maximal supergravity amplitudes_ , _JHEP_ 09 (2017) 019, [1706.09381].
* [152] S. He and O. Schlotterer, _New Relations for Gauge-Theory and Gravity Amplitudes at Loop Level_ , _Phys. Rev. Lett._ 118 (2017) 161601, [1612.00417].
* [153] S. He, O. Schlotterer and Y. Zhang, _New BCJ representations for one-loop amplitudes in gauge theories and gravity_ , _Nucl. Phys._ B930 (2018) 328–383, [1706.00640].
* [154] Y. Geyer and R. Monteiro, _Gluons and gravitons at one loop from ambitwistor strings_ , _JHEP_ 03 (2018) 068, [1711.09923].
* [155] A. Edison, S. He, O. Schlotterer and F. Teng, _One-loop Correlators and BCJ Numerators from Forward Limits_ , _JHEP_ 09 (2020) 079, [2005.03639].
* [156] Y. Geyer, R. Monteiro and R. Stark-Muchão, _Two-Loop Scattering Amplitudes: Double-Forward Limit and Colour-Kinematics Duality_ , _JHEP_ 12 (2019) 049, [1908.05221].
* [157] C. R. Mafra and O. Schlotterer, _Towards one-loop SYM amplitudes from the pure spinor BRST cohomology_ , _Fortsch. Phys._ 63 (2015) 105–131, [1410.0668].
* [158] M. Berg, I. Buchberger and O. Schlotterer, _String-motivated one-loop amplitudes in gauge theories with half-maximal supersymmetry_ , _JHEP_ 07 (2017) 138, [1611.03459].
* [159] K. A. Roehrig and D. Skinner, _A Gluing Operator for the Ambitwistor String_ , _JHEP_ 01 (2018) 069, [1709.03262].
* [160] C. Baadsgaard, N. E. J. Bjerrum-Bohr, J. L. Bourjaily, P. H. Damgaard and B. Feng, _Integration Rules for Loop Scattering Equations_ , _JHEP_ 11 (2015) 080, [1508.03627].
* [161] M. Zlotnikov, _Polynomial reduction and evaluation of tree- and loop-level CHY amplitudes_ , _JHEP_ 08 (2016) 143, [1605.08758].
* [162] H. Gomez, S. Mizera and G. Zhang, _CHY Loop Integrands from Holomorphic Forms_ , _JHEP_ 03 (2017) 092, [1612.06854].
* [163] G. Chen, Y.-K. E. Cheung, T. Wang and F. Xu, _A differential operator for integrating one-loop scattering equations_ , _JHEP_ 01 (2017) 028, [1609.07621].
* [164] H. Gomez, R. L. Jusinskas and A. Lipstein, _Cosmological Scattering Equations at Tree-level and One-loop_ , 2112.12695.
* [165] H. Gomez, R. L. Jusinskas and A. Lipstein, _Cosmological Scattering Equations_ , _Phys. Rev. Lett._ 127 (2021) 251604, [2106.11903].
* [166] A. Edison, M. Guillen, H. Johansson, O. Schlotterer and F. Teng, _One-loop matrix elements of effective superstring interactions: $\alpha$’-expanding loop integrands_, _JHEP_ 12 (2021) 007, [2107.08009].
* [167] T. Adamo, E. Casali and D. Skinner, _A Worldsheet Theory for Supergravity_ , _JHEP_ 02 (2015) 116, [1409.5656].
* [168] T. Adamo, E. Casali, L. Mason and S. Nekovar, _Amplitudes on plane waves from ambitwistor strings_ , _JHEP_ 11 (2017) 160, [1708.09249].
* [169] T. Adamo, E. Casali and S. Nekovar, _Ambitwistor string vertex operators on curved backgrounds_ , _JHEP_ 01 (2019) 213, [1809.04489].
* [170] T. Adamo, E. Casali and S. Nekovar, _Yang-Mills theory from the worldsheet_ , _Phys. Rev._ D98 (2018) 086022, [1807.09171].
* [171] C. LeBrun, _Thickenings and conformal gravity_ , _Commun.Math.Phys._ 139 (1991) 1–43.
* [172] E. Witten, _Twistor - Like Transform in Ten-Dimensions_ , _Nucl.Phys._ B266 (1986) 245.
* [173] T. Adamo, L. Mason and A. Sharma, _Gluon scattering on self-dual radiative gauge fields_ , 2010.14996.
* [174] T. Adamo, L. Mason and A. Sharma, _Graviton scattering in self-dual radiative space-times_ , 2203.02238.
* [175] D. B. Fairlie and D. E. Roberts, _Dual models without tachyons - a new approach_ , .
* [176] D. B. Fairlie, _A Coding of Real Null Four-Momenta into World-Sheet Co-ordinates_ , _Adv. Math. Phys._ 2009 (2009) 284689, [0805.2263].
* [177] D. J. Gross and P. F. Mende, _String Theory Beyond the Planck Scale_ , _Nucl.Phys._ B303 (1988) 407.
* [178] D. J. Gross and P. F. Mende, _The High-Energy Behavior of String Scattering Amplitudes_ , _Phys.Lett._ B197 (1987) 129.
* [179] O. Hohm, W. Siegel and B. Zwiebach, _Doubled $\alpha^{\prime}$-geometry_, _JHEP_ 02 (2014) 065, [1306.2970].
* [180] T. Azevedo and R. L. Jusinskas, _Connecting the ambitwistor and the sectorized heterotic strings_ , _JHEP_ 10 (2017) 216, [1707.08840].
* [181] T. Azevedo, R. L. Jusinskas and M. Lize, _Bosonic sectorized strings and the $(DF)^{2}$ theory_, _JHEP_ 01 (2020) 082, [1908.11371].
* [182] R. L. Jusinskas, _Asymmetrically twisted strings_ , 2108.13426.
* [183] M. L. Lize, B. Lyu, W. Siegel and Y.-P. Wang, _Chiral string theories as an interpolation between strings and particles_ , 2109.10401.
* [184] E. Casali and P. Tourkine, _Windings of twisted strings_ , _Phys. Rev._ D97 (2018) 061902, [1710.01241].
* [185] S. Mizera, _Aspects of Scattering Amplitudes and Moduli Space Localization_ , Ph.D. thesis, Princeton, Inst. Advanced Study, 2020. 1906.02099. 10.1007/978-3-030-53010-5.
* [186] T. McLoughlin, A. Puhm and A.-M. Raclariu, _The SAGEX Review on Scattering Amplitudes, Chapter 11: Soft Theorems and Celestial Amplitudes_ , 2203.13022.
* [187] A. Strominger, _On BMS Invariance of Gravitational Scattering_ , _JHEP_ 07 (2014) 152, [1312.2229].
* [188] A. Strominger, _Lectures on the Infrared Structure of Gravity and Gauge Theory_ , 1703.05448.
* [189] S. Pasterski and S.-H. Shao, _Conformal basis for flat space amplitudes_ , _Phys. Rev. D_ 96 (2017) 065022, [1705.01027].
* [190] S. Pasterski, _Lectures on celestial amplitudes_ , _Eur. Phys. J. C_ 81 (2021) 1062, [2108.04801].
* [191] T. Adamo, E. Casali and D. Skinner, _Perturbative gravity at null infinity_ , _Class.Quant.Grav._ 31 (2014) 225008, [1405.5122].
* [192] Y. Geyer, A. E. Lipstein and L. Mason, _Ambitwistor strings at null infinity and (subleading) soft limits_ , _Class.Quant.Grav._ 32 (2015) 055003, [1406.1462].
* [193] A. Strominger, _w(1+infinity) and the Celestial Sphere_ , 2105.14346.
* [194] T. Adamo, L. Mason and A. Sharma, _Celestial $w_{1+\infty}$ Symmetries from Twistor Space_, _SIGMA_ 18 (2022) 016, [2110.06066].
* [195] S. V. Ketov, _Conformal Field Theory_. World Scientific, Singapore, 1995.
* [196] P. Di Francesco, P. Mathieu and D. Senechal, _Conformal Field Theory_. Graduate Texts in Contemporary Physics. Springer-Verlag, New York, 1997, 10.1007/978-1-4612-2256-9.
* [197] J. Polchinski, _String theory. Vol. 2: Superstring theory and beyond_ , .
|
pnasresearcharticle Panahi Human activities are having increasingly negative
impacts on natural systems and it is of interest to understand how the “pace”
of parameter change may lead to catastrophic consequences. This work studies
the phenomenon of rate-induced tipping (R-tipping) in high-dimensional
ecological networks, where the rate of parameter change can cause the system
to undergo a tipping point from healthy survival to extinction. A quantitative
scaling law between the probability of R-tipping and the rate was uncovered,
with a striking and devastating consequence: in order to reduce the
probability, parameter changes must be slowed down to such an extent that the
rate practically reaches zero. This may pose an extremely significant
challenge in our efforts to protect and preserve the natural environment. YCL
and AH conceived the idea. SP performed computations. All analyzed data. YCL
and SP wrote the paper. YCL and AH edited the paper. The authors declare no
competing interests. 1To whom correspondence should be addressed. E-mail:
<EMAIL_ADDRESS>
# Rate-induced tipping in complex high-dimensional ecological networks
Shirin Panahi School of Electrical, Computer and Energy Engineering, Arizona
State University, Tempe, Arizona 85287, USA Younghae Do Department of
Mathematics, Nonlinear Dynamics $\&$ Mathematical Application Center,
Kyungpook National University, Daegu 41566, Republic of Korea Alan Hastings
Department of Environmental Science and Policy, University of California, One
Shields Avenue, Davis, CA 95616, USA and Santa Fe Institute, 1399 Hyde Park
Road, Santa Fe, NM 87501, USA Ying-Cheng Lai School of Electrical, Computer
and Energy Engineering, Arizona State University, Tempe, Arizona 85287, USA
Department of Physics, Arizona State University, Tempe, Arizona 85287, USA
###### Abstract
In an ecosystem, environmental changes as a result of natural and human
processes can cause some key parameters of the system to change with time.
Depending on how fast such a parameter changes, a tipping point can occur.
Existing works on rate-induced tipping, or R-tipping, offered a theoretical
way to study this phenomenon but from a local dynamical point of view,
revealing, e.g., the existence of a critical rate for some specific initial
condition above which a tipping point will occur. As ecosystems are subject to
constant disturbances and can drift away from their equilibrium point, it is
necessary to study R-tipping from a global perspective in terms of the initial
conditions in the entire relevant phase space region. In particular, we
introduce the notion of the probability of R-tipping defined for initial
conditions taken from the whole relevant phase space. Using a number of real-
world, complex mutualistic networks as a paradigm, we discover a scaling law
between this probability and the rate of parameter change and provide a
geometric theory to explain the law. The real-world implication is that even a
slow parameter change can lead to a system collapse with catastrophic
consequences. In fact, to mitigate the environmental changes by merely slowing
down the parameter drift may not always be effective: only when the rate of
parameter change is reduced to practically zero would the tipping be avoided.
Our global dynamics approach offers a more complete and physically meaningful
way to understand the important phenomenon of R-tipping.
###### keywords:
Rate-induced tipping $|$ nonlinear dynamics $|$ scaling law
###### doi:
www.pnas.org/cgi/doi/10.1073/pnas.XXXXXXXXXX
This manuscript was compiled on
In complex dynamical systems, the phenomenon of tipping point, characterized
by an abrupt transition from one type of behavior (typically normal, healthy,
or survival) to another type (e.g., catastrophic), has received growing
attention (1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19,
20, 21, 22, 23, 24, 25). A tipping point is often of significant concern
because it is a point of “no return” in the parameter space, manifested by the
collapse of the system as a parameter passes through a critical value. In
ecological systems, sudden extinction of species on a large scale can be the
result of a tipping point (1, 2, 3, 4). Tipping points can arise in diverse
contexts such as the outbreak of epidemics (26), global climate changes (27),
and the sudden transition from normal to depressed mood in bipolar patients
(28). In nonlinear dynamics, a common type of bifurcation responsible for a
tipping point is saddle-node bifurcations (forward or backward). Consider the
situation where, in the parameter regime of interest, there are two coexisting
stable steady states: a “high” state corresponding to normal or “survival”
functioning and a “low” or “extinction” state, each with its own basin of
attractor. Suppose external factors such as climate change cause a bifurcation
parameter of the system to increase. A tipping point is a backward saddle-node
bifurcation, after which the “survival” fixed point disappears, leaving the
“extinction” state as the only destination of the system, where the original
basin of the “survival” state is absorbed into the basin of the “extinction”
state.
In real-world dynamical systems, parameters are not stationary but constantly
drift in time. A known example is the slow increase in the average global
temperature with time due to human activities. In ecological systems, some key
parameters such as the carrying capacity or the species decay rate can change
with time, and there is a global tendency for such parameter changes to “speed
up.” In fact, the rate of environmental change is an important driver across
different scales in ecology (29). The behavior of the parameter variations
with time introduces another “meta” parameter into the dynamical process: the
rate of change of the parameters. About ten years ago, it was found that the
rate can induce a tipping point - the phenomenon of rate-induced tipping or
R-tipping (9), which is relevant to fields such as climate science (30, 31),
neuroscience (32, 33), vibration engineering (34), and even competitive
economy (35). Existing studies of R-tipping were for low-dimensional dynamical
systems and the analysis approaches were “local” in the sense that they
focused on the behaviors of some specific initial condition and trajectory,
addressing issues such as the critical rate for tipping (36, 37). In
particular, with respect to a specific initial condition, R-tipping in these
studies was defined as an abrupt change in the behavior of the system (or a
critical transition) taking place at a specific rate of change of a
bifurcation parameter (9).
The state that an ecological system is in depends on a combination of
deterministic dynamics, small scale stochastic influences (e.g., demographic
stochasticity (38)), and large stochastic disturbances such as drought or
other significant climatic event (39). So when considering future dynamics of
ecological systems it makes sense to consider systems that may be far from
equilibrium, but still within the basin of attraction of an equilibrium,
rather than starting at the equilibrium. High-dimensional ecological systems
are particularly likely to be found far from equilibrium. In fact, it has been
suggested that it can be common for ecosystems to be in some transient state
(40, 41).
In this paper, we focus on high-dimensional, empirical ecological networks and
investigate R-tipping with two key time-varying parameters by presenting a
computationally feasible, “global” approach to exploring the entire relevant
phase space region with analytic insights. We focus on a representative class
of such systems: mutualistic networks (42, 43, 44, 14, 45, 46, 47, 17, 21, 19,
48) that are fundamental to ecological systems, which are formed through
mutually beneficial interactions between two groups of species. In a
mutualistic network, a species in one group derives some form of benefit from
the relationship with some species in the other group. Examples include the
corals and the single-celled zooxanthellae that form the large-scale coral
reefs, and the various networks of pollinators and plants in different
geographical regions of the world. These networks influence biodiversity,
ecosystem stability, nutrient cycling, community resilience, and evolutionary
dynamics (49), and they are a key aspect of ecosystem functioning with
implications for conservation and ecosystem management. Understanding the
ecological significance of mutualistic networks is crucial for unraveling the
complexities of ecological communities and for implementing effective
strategies to safeguard biodiversity and ecosystem health (50).
Mathematically, because of the typically large number of species involved in
the mutualistic interactions, the underlying networked systems are high-
dimensional nonlinear dynamical systems.
We first ask if R-tipping can arise in such high-dimensional systems through
simulating a number of empirical pollinator-plant mutualistic networks (see
Tab. 1) and obtain an affirmative answer. Our computations reveal that the
critical rate above which a tipping point occurs depends strongly on the
initial condition. Rather than studying the critical rate for any specific
initial condition, we go beyond the existing local analysis approaches by
investigating the probability of R-tipping for a large ensemble of initial
conditions taken from the whole relevant phase space and asking how this
probability, denoted as $\Phi(r)$, depends on the rate $r$ of parameter
change. We discover a scaling law between $\Phi(r)$ and $r$: as the rate
increases, the probability first increases rapidly, then slowly, and finally
saturates. Using a universal two-dimensional (2D) effective model that was
validated to be particularly suitable for predicting and analyzing tipping
points in high-dimensional mutualistic networks (17), we analytically derive
the scaling law. Specifically, let $\kappa(t)$ be the parameter that changes
with time linearly at the rate $r$ in a finite range, denoted as $[\kappa_{\rm
min},\kappa_{\rm max}]$. The scaling law is given by
$\displaystyle\Phi(r)\sim\exp{[-C(\kappa_{\rm max}-\kappa_{\rm min})/r]},$ (1)
where $C>0$ is a constant. Our theoretical analysis indicates that, for 2D
systems, $C$ is nothing but the maximum possible unstable eigenvalue of the
mediating unstable fixed point on the boundary separating the basins of the
survival and extinction attractors when the parameter $\kappa$ varies in the
range $[\kappa_{\rm min},\kappa_{\rm max}]$. However, for high-dimensional
systems, such correspondence does not hold, but $C$ can be determined through
a numerical fitting.
The scaling law (1) has the following features. First, the probability
$\Phi(r)$ is an increasing function of $r$ for $r>0$ [$\Phi^{\prime}(r)>0$].
Second, $\Phi^{\prime}(r)$ is a decreasing function of $r$, i.e., the increase
of $\Phi(r)$ with $r$ slows down with $r$ and the rate of increase becomes
zero for $r\rightarrow\infty$. Third, the rate of increase in $\Phi(r)$ with
$r$ is the maximum for $r\gtrsim 0$, and $\Phi(r)$ becomes approximately
constant for $r>r^{*}\sim\sqrt{(\kappa_{\rm max}-\kappa_{\rm min})}$. This
third feature has a striking implication because the probability of R-tipping
will grow dramatically as soon as the rate of parameter change increases from
zero. The real-world implication is alarming because it means that even a slow
parameter change can lead to a system collapse with catastrophic consequences.
To control or mitigate the environmental changes by merely slowing down the
parameter drift may not always be effective: only when the rate of parameter
change is reduced to practically zero would the tipping be avoided!
Alternatively, the scaling law (1) can be expressed as the following explicit
formula:
$\displaystyle\Phi(r)=B\exp{[-C(\kappa_{\rm max}-\kappa_{\rm min})/r]}$ (2)
with an additional positive constant $B$. For 2D systems, $B$ is the
difference between the basin areas of the extinction state for
$\kappa=\kappa_{\rm max}$ and $\kappa=\kappa_{\rm min}$. For high-dimensional
systems, $B$ can be determined numerically.
While a comprehensive understanding of the entire parameter space as well as
the rates and directions of change associated with these parameters are worth
investigating, the picture that motivated our work was the potential impacts
of ongoing climate change on ecological systems. Under climate change, various
parameters can vary in different directions, and some changes might offset the
effects of others. Our focus is on the scenarios where the changes in
parameters align with the observed detrimental environmental impacts. For
example, consistent with environmental deterioration, in a mutualistic network
the species decay rate can increase, and/or the mutualistic interaction
strength can decrease. A simultaneous increase in the species decay rate and
mutualistic strength does not seem physically reasonable in this context. We
thus study scenarios of multiple parameter variations that align with the
realistic climate change impacts by carrying out computations with a
systematic analysis of the tipping-point transitions in the two-dimensional
parameter plane of the species decay rate and the mutualistic interaction
strength.
## Results
We consider the empirical mutualistic pollinator–plant networks from the Web
of Life database (www.Web-of-Life.es). The needs to search for the parameter
regions exhibiting R-tipping and to simulate a large number of initial
conditions in the phase space as required by our global analysis approach, as
well as the high dimensionality of the empirical mutualistic networks demand
extremely intensive computation (51). To make the computation feasible, we
select ten networks to represent a diverse range of mutualistic interactions
from different regions of the world and highlight the generality of our
approach to R-tipping and the scaling law. The basic information about these
networks such as the name of the networks, the countries where the empirical
data were collected, and the number of species in each network, is listed in
Tab. 1. The dynamics of a mutualistic network of $N_{A}$ pollinator and
$N_{P}$ plant species, taking into account the generic plant-pollinator
interactions (42), can be described by a total of $N=N_{A}+N_{P}$ nonlinear
differential equations of the Holling type (52) in terms of the species
abundances as
$\displaystyle\dot{P}_{i}=$ $\displaystyle
P_{i}\Big{(}\alpha_{i}^{P}-\sum_{l=1}^{N_{p}}\beta_{il}^{P}P_{l}+\frac{\sum_{j=1}^{N_{A}}\gamma_{ij}^{P}A_{j}}{1+h\sum_{j=1}^{N_{A}}\gamma_{ij}^{P}A_{j}}\Big{)},$
(3a) $\displaystyle\dot{A}_{j}=$ $\displaystyle
A_{j}\Big{(}\alpha_{j}^{A}-\kappa_{j}-\sum_{l=1}^{N_{A}}\beta_{jl}^{A}A_{l}+\frac{\sum_{i=1}^{N_{P}}\gamma_{ji}^{A}P_{i}}{1+h\sum_{i=1}^{N_{P}}\gamma_{ji}^{A}P_{i}}\Big{)},$
(3b)
where $P_{i}$ and $A_{j}$ are the abundances of the $i^{th}$ and $j^{th}$
plant and pollinator species, respectively, $i=1,\ldots,N_{P}$,
$j=1,\ldots,N_{A}$, $\kappa$ is the pollinator decay rate, $\alpha^{P(A)}$ is
the intrinsic growth rate in the absence of intraspecific competition and any
mutualistic effect, $h$ is the half-saturation constant. Intraspecific and
interspecific competition of the plants (pollinators) is characterized by the
parameters $\beta_{ii}^{P}$ ($\beta_{jj}^{A}$) and $\beta_{il}^{P}$
($\beta_{jl}^{A}$), respectively. The mutualistic interactions as
characterized by the parameter $\gamma_{ij}^{P}$ can be written as
$\gamma_{ij}^{P}=\xi_{ij}\gamma_{0}/K_{i}^{\tau}$, where $\xi_{ij}=1$ if there
is a mutualistic interaction between the $i^{th}$ plant and $j^{th}$
pollinator (zero otherwise), $\gamma_{0}$ is a general interaction parameter,
$K_{i}$ is the degree of the $i^{th}$ plant species in the network, and $\tau$
determines the strength of the trade-off between the interaction strength and
the number of interactions. (The expression for $\gamma_{ji}^{A}$ is similar.)
Table 1: Empirical mutualistic networks studied in this work and their corresponding range of parameter change and fitting constants in Eq. (1) Network | Country | $N_{P}$ | $N_{A}$ | $\kappa$-interval | $B$ | $C$
---|---|---|---|---|---|---
M_PL_008 | Canary Islands | $11$ | $38$ | ${[0.90,0.93]}$ | $0.36$ | $0.19$
M_PL_013 | South Africa | $9$ | $56$ | ${[0.70,0.99]}$ | $0.07$ | $0.11$
M_PL_022 | Argentina | $21$ | $45$ | ${[0.75,0.93]}$ | $0.88$ | $0.12$
M_PL_023 | Argentina | $23$ | $72$ | $[0.87,0.96]$ | $0.59$ | $0.09$
M_PL_027 | New Zealand | $18$ | $60$ | $[0.90,0.95]$ | $0.49$ | $0.09$
M_PL_032 | USA | $7$ | $33$ | $[0.82,0.99]$ | $0.55$ | $0.19$
M_PL_036 | Açores | $10$ | $12$ | $[0.74,0.88]$ | $0.39$ | $0.12$
M_PL_037 | Denmark | $10$ | $40$ | $[0.87,0.93]$ | $0.23$ | $0.15$
M_PL_038 | Denmark | $8$ | $42$ | $[0.85,0.95]$ | $0.30$ | $0.15$
M_PL_045 | Greenland | $17$ | $26$ | $[0.96,0.98]$ | $0.09$ | $0.05$
The computational setting of our study is as follows. In the network system
described by Eqs. 3a and 3b, the number of the equations determines the phase-
space dimension of the underlying nonlinear dynamical system. There are a
large number of basic parameters in the model, such as $\gamma$ that
quantifies the strength of the mutualistic interactions and $\kappa$
characterizing the species decay rate. In the context of R-tipping, while all
the parameters should be time-varying in principle, to make our study
computationally feasible, we assume that the two key parameters ($\kappa$ and
$\gamma$) are time-dependent while keeping the other parameters fixed. Since
the defining characteristic of a system exhibiting a tipping point is the
coexistence of two stable steady states: survival and extinction, we focus on
the range of parameter variations in which the network system under study
exhibits the two stable equilibria. When presenting our results (Fig. 1 below
and Figs. S4-S14 in Supplementary Information), in each case the parameter
variation is along a specific direction in the two-dimensional parameter
plane: $\kappa$ and $\gamma$ with the goal to understand how changes in these
two parameters impact the R-tipping probability.
To introduce the rate change of a parameter, we consider the scenario where
negative environmental impacts cause the species decay rate to increase
linearly with time at the rate $r$ from a minimal value $\kappa_{\rm min}$ to
a maximal value $\kappa_{\rm max}$:
$\displaystyle\dot{\kappa}_{j}=\begin{cases}r\quad{\rm if}\quad\kappa_{\rm
min}<\kappa_{j}<\kappa_{\rm max}\\\ 0\quad\quad{\rm otherwise.}\end{cases}$
(4)
(For mutualistic networks, another relevant parameter that is vulnerable to
environmental change is the mutualistic interaction strength. The pertinent
results are presented in Supplementary Information.) To calculate the
probability of R-tipping, $\Phi(r)$, we set $r=0$ so that $\kappa=\kappa_{\rm
min}$, solve Eqs. (3a) and (3b) numerically for a large number of random
initial conditions chosen uniformly from the whole high-dimensional phase
space, and determine $10^{5}$ initial conditions that approach the high stable
steady state in which no species becomes extinct. We then increase the rate
$r$ from zero. For each fixed value of $r$, we calculate, for each of the
selected $10^{5}$ initial conditions, whether or not the final state is the
high stable state. If yes, then there is no R-tipping for the particular
initial condition. However, if the final state becomes the extinction state,
R-tipping has occurred for this value of $r$. The probability $\Phi(r)$ can be
approximated by the fraction of the number of initial conditions leading to
R-tipping out of the $10^{5}$ initial conditions.
Figure 1: Probability $\Phi(r)$ of R-tipping versus the time rate $r$ of
parameter change for ten real-world mutualistic networks, together with the
corresponding bipartite structure for each network. The dots are the
probability calculated by simulating Eqs. (3a), (3b), and (4), with an
ensemble of random initial conditions. The solid curves are the analytic
results from Eq. (1), with the two fitting parameters $B$ and $C$ given in
Tab. 1. Other parameter values are $\alpha=0.3$, $\beta=1$, $h=0.4$,
$\gamma^{P}=1.93$, and $\gamma^{A}=1.77$.
Figure 1 shows $\Phi(r)$ versus $r$ for the ten empirical mutualistic networks
specified in Tab. 1, together with the bipartite network structure, which are
distinguished by different colors. For each network, the interval
$[\kappa_{\rm min},\kappa_{\rm max}]$ of the bifurcation parameter (listed in
the fifth column of Tab. 1) is chosen such that the dynamical network for
static parameter values exhibits two stable steady states and a tipping point
in this interval, as determined by a computational bifurcation analysis of the
species abundances versus $\kappa$. Despite the differences in the topology
and the specific parameter values among the empirical networks, it is
remarkable that the probability of R-tipping exhibits a characteristic
behavior common among all the networks: as the time-varying rate of the
bifurcation parameter increases from zero, the probability increases rapidly
and then saturates at an approximately finite constant value, as quantified by
the analytic scaling law (1). The final or asymptotic value of the R-tipping
probability attained in the regime of large time rate of change of the
bifurcation parameter depends on the specifics of the underlying mutualistic
network in terms of its topological structure and basic system parameters. For
example, the total number of species, or the phase-space dimension of the
underlying dynamical system, and the relative numbers of the pollinator and
plant species vary dramatically across the ten networks, as indicated by the
ten surrounding network-structure diagrams in Fig. 1. The structural and
parametric disparities among the networks lead to different relative basin
volumes of the survival and extinction stable steady states, giving rise to
distinct asymptotic probabilities of R-tipping.
The ecological interpretation of the observed different asymptotic values of
the R-tipping probability is, as follows. Previous studies revealed the
complex interplay between the structural properties of the network and
environmental factors in shaping extinction probabilities (53, 54, 55, 56, 57,
58, 59, 60). For example, it was discovered (53) that the plant species tend
to play a more significant role in the health of the network system as
compared to pollinators (53) in the sense that plant extinction due to climate
change is more likely to trigger pollinator coextinction than the other way
around. In another example, distinct topological features were found to be
associated with the networks with a higher probability of extinction (54, 55).
It was also found that robust plant-pollinator mutualistic networks tend to
exhibit a combination of compartmentalized and nested patterns (55). More
generally, mutualistic networks in the real world are quite diverse in their
structure and parameters, each possessing one or two or all the features
including lower interaction density, heightened specialization, fewer
pollinator visitors per plant species, lower nestedness, and lower modularity,
etc. (54, 55, 56, 57, 58, 59, 60). It is thus reasonable to hypothesize that
the asymptotic value of the R-tipping probability can be attributed to the
sensitivity of the underlying network to environmental changes. For example, a
higher (lower) saturation value of the extinction probability can be
associated with the networks with a large (small) number of plants and lower
(higher) nestedness, network $M\\_PL\\_22$ (network $M\\_PL\\_13$). In other
examples, network $M\\_PL\\_32$ can be categorized as “vulnerable” due to its
low modularity, despite having a small number of plants, and network
$M\\_PL\\_45$ is resilient against extinction due to its high nestedness and
high modularity.
The remarkable phenomenon is that, in spite of these differences, the rapid
initial increase in the R-tipping probability is shared by all ten networks!
That is when a parameter begins to change with time from zero, even slowly,
the probability of R-tipping increases dramatically. The practical implication
is that ecosystems with time-dependent parameter drift are highly susceptible
to R-tipping. Parameter drifting, even at a slow pace, will be detrimental.
This poses a daunting challenge to preserving ecosystems against negative
environmental changes. In particular, according to conventional wisdom,
ecosystems can be effectively protected by slowing down the environmental
changes, but our results suggest that catastrophic tipping can occur with a
finite probability unless the rate of the environmental changes is reduced to
a near zero value.
To further explore the effects of parameter changes, we study a scenario in
which negative environmental impacts cause the species decay rate ($\kappa$)
and the mutualistic interaction strength ($\gamma_{0}$) to linearly increase
and decrease, respectively. By considering the rate of changing of $\kappa$
($\gamma_{0}$) as $r_{\kappa}$ ($r_{\gamma_{0}}$), we set
$r_{\kappa}=\mathcal{C}r_{\gamma_{0}}$, which allows us to investigate three
different scenarios by varying the parameter $\mathcal{C}$: $\mathcal{C}<1$,
$\mathcal{C}=1$, and $\mathcal{C}>1$. Figure 2 shows a systematic analysis of
the tipping-point transitions in the 2D parameter plane along with the
probability of R-tipping for the three different scenarios for the network
${\rm M}\\_{\rm PL}\\_032$. (The pertinent results for the other nine networks
are presented in Sec. 4 in Supporting Information.)
Figure 2: Scaling law of probability of R-tipping in two-dimensional
parameter space. (a) Two-dimensional parameter plane of the species decay rate
($\kappa$) and the mutualistic interaction strength ($\gamma_{0}$) for the
network ${\rm M}\\_{\rm PL}\\_032$. The solid lines correspond to the three
different scenarios of the parameter $\mathcal{C}$ in calculating the
probability of the R-tipping point from $10^{7}$ initial conditions for (b)
$\mathcal{C}<1$, (c) $\mathcal{C}=1$, and (d) $\mathcal{C}>1$. Other
parameters are $t=0.5$, $\beta=1$, $\alpha=0.3$, $\mu=0$, and $h=0.4$.
To understand the behavior of the R-tipping probability in Fig. 1, we resort
to the approach of dimension reduction (61). In particular, for tipping-point
dynamics, a high-dimensional mutualistic network can be approximated by an
effective model of two dynamical variables: the respective mean abundances of
all pollinator and plant species. The effective 2D model can be written as
(17)
$\displaystyle\dot{P}$ $\displaystyle=P\Big{(}\alpha-\beta
P+\frac{\gamma^{P}A}{1+h\gamma^{P}A}\Big{)},$ (5a) $\displaystyle\dot{A}$
$\displaystyle=A\Big{(}\alpha-\kappa-\beta
A+\frac{\gamma^{A}P}{1+h\gamma^{A}P}\Big{)},$ (5b)
where $P$ and $A$ are the average abundances of all the plants and
pollinators, respectively, $\alpha$ is the effective growth rate and parameter
$\beta$ characterizes the combined effects of intraspecific and interspecific
competition. The parameters $\gamma^{P}$ and $\gamma^{A}$ are the effective
mutualistic interaction strengths that can be determined by the method of
eigenvector weighting (17) (Supplementary Note 1). Equations (5a) and (5b)
possess five possible equilibria: $\mathbf{f}_{1}\equiv(0,0)^{T}$,
$\mathbf{f}_{2}\equiv(\alpha/\beta,0)^{T}$,
$\mathbf{f}_{3}\equiv(0,(\alpha-\kappa)/\beta)^{T}$, and
$\mathbf{f}_{4,5}\equiv(g_{1},g_{2})^{T}$, where $g_{1}$ and $g_{2}$ are two
possible equilibria that depend on the values of the parameters of the model
and can be calculated by setting zero the factor in the parentheses of Eqs.
(5a) and (5b).
Figure 3: Dynamics of the 2D reduced model Eqs. (5a) and (5b). (a,b) A tipping
point occurs as the species decay rate $\kappa$ increases towards one, for the
mean plant and pollinator abundances, respectively. The green and yellow
curves correspond to the stable and unstable steady states, respectively,
where the yellow curve with a red segment is the unstable fixed point
$\mathbf{f}_{5}$. The region of bistability is indicated by the blue dot-
dashed ellipses. (c,d) Examples of basins of attraction of the 2D model whose
parameters are determined as the corresponding averages from the empirical
network ${\rm M}\\_{\rm PL}\\_036$ in Tab. 1 for $\kappa_{min}=0.74$ and
$\kappa_{max}=0.88$, respectively. Other parameters are $\alpha=0.3$,
$\beta=1$, $h=0.4$, $\gamma^{P}=1.93$, and $\gamma^{A}=1.77$. The pink and
gray regions correspond to the basins of the stable steady states $f_{2}$
(extinction) and $f_{4}$ (survival), respectively.
The first equilibrium $\mathbf{f}_{1}$, an extinction state, is at the origin
$(P^{*},\,A^{*})=(0,0)$ and is unstable. The second equilibrium
$\mathbf{f}_{2}$ is located at the $(P^{*},\,A^{*})=(\alpha/\beta,0)$ and it
can be stable or unstable depending on the parameters $\kappa$. The locations
of the remaining equilibria depend on the value of $\kappa$. In particular,
the third equilibrium $\mathbf{f}_{3}$ can be unstable or nonexistent and the
fourth equilibrium $\mathbf{f}_{4}$ is stable and coexists with stable
equilibrium $\mathbf{f}_{2}$ in some interval of $\kappa$. The fifth
equilibrium $\mathbf{f}_{5}$ is an unstable saddle fixed point in some
relevant interval of $\kappa$. Figures 3(a) and 3(b) exemplify the behaviors
of the equilibria as $\kappa$ increases from zero to one for $\alpha=0.3$,
$\beta=1$, $h=0.4$, $\gamma^{P}=1.93$, and $\gamma^{A}=1.77$, for the average
plant and pollinator abundances, respectively. In each panel, the upper green
curve is the survival fixed point $\mathbf{f}_{4}$, while the lower horizontal
green line corresponds to the extinction state $\mathbf{f}_{2}$. The blue dot-
dashed ellipse indicates the interval of $\kappa$ in which two stable
equilibria coexist (bistability), whose right edge marks a tipping point.
Figure 4: Basin boundary between the extinction and survival fixed-point
attractors for different rate of parameter change. Shown are a series of basin
boundaries for $r\in[0\,1]$. As $r$ increases from zero, the boundary moves in
the direction of large species abundances. The basin boundaries are calculated
from the 2D model of the empirical network ${\rm M}\\_{\rm PL}\\_036$ in Tab.
1. Other parameters are the same as those in Figs. 3(c) and 3(d). Figure 5:
Analysis of the dynamical mechanism responsible for R-tipping. The dynamics of
the unstable fixed point $\mathbf{f}_{5}$ together with those of a rectangle
region of the initial conditions in its neighborhood are illustrated for two
values of the rate of parameter change: (a) $r=r_{1}$ and (b) $r=r_{2}>r_{1}$.
At $t=0$ ($\kappa=\kappa_{\rm min}$, a segment of the basin boundary in the
vicinity of $\mathbf{f}_{5}$ is illustrated as the solid green line. As the
rate of parameter change begins to increase, the basin boundary moves upward
and eventually accumulates at that for $\kappa=\kappa_{\rm max}$ \- the solid
orange lines. For a small rate, the time required for the green boundary to
reach the orange boundary is longer. All the initial conditions in the light
green region above the green boundary belong to the basin of the survival
attractor for $\kappa=\kappa_{\rm min}$, whose trajectories are determined by
the local exponentially contracting and expanding dynamics of
$\mathbf{f}_{5}$. The fate of these initial conditions is the result of a
“race” to go above the orange basin boundary within the “allowed” time
duration [$T_{1}$ in (a) and $T_{2}<T_{1}$ in (b)]: those managing to go above
will approach the survival attractor, but those that stay below will lead to
extinction. Since $T_{1}>T_{2}$, the fraction of initial conditions that can
go above in (a) is larger than that in (b), leading to a higher probability of
R-tipping for case (b). The precise fraction of those initial conditions is
indicated by the red rectangular region in (b), whose initial height stretches
to the distance between the green and orange boundaries in the permissible
time. This gives the probability of R-tipping as quantified by the scaling law
(1). See text for more details.
What will happen to the dynamics when the parameter $\kappa$ becomes time
dependent? Without loss of generality, we focus on the interval of $\kappa$ as
exemplified by the horizontal range of the dot-dashed ellipse in Figs. 3(a)
and 3(b), defined as $[\kappa_{\rm min},\kappa_{\rm max}]$, in which there is
bistability in the original high-dimensional network and in the 2D reduced
model as well. Note that, for different high-dimensional empirical networks,
the values of $\kappa_{\rm min}$ and $\kappa_{\rm max}$ in the 2D effective
models are different, as illustrated in the fifth column of Tab. 1. Because of
the coexistence of two stable fixed points, for every parameter value in the
range $[\kappa_{\rm min},\kappa_{\rm max}]$, there are two basins of
attraction, as illustrated in Figs. 3(c) and 3(d) for the 2D effective model
of an empirical network ${\rm M}\\_{\rm PL}\\_036$ for $\kappa=0.74$ and
$\kappa=0.88$, respectively, where the pink and gray regions correspond to the
basins of the extinction and survival attractors, respectively. The basin
boundary is the stable manifold of the unstable fixed point $\mathbf{f}_{5}$.
It can be seen that, as $\kappa$ increases, the basin of the extinction
attractor increases, accompanied by a simultaneous decrease in the basin area
of the survival attractor. This can be understood by comparing the positions
of the equilibria in Fig. 3 where, as $\kappa$ increases, the position of
survival fixed point $\mathbf{f}_{4}$ moves towards lower plant and pollinator
abundances, but the unstable fixed point $\mathbf{f}_{5}$ moves in the
opposite direction: the direction of larger species abundances.
The boundary separating the basins of the extinction and survival fixed-point
attractors is the stable manifold of $\mathbf{f}_{5}$. As $r$ increases from
zero, $\mathbf{f}_{5}$ moves in the direction of larger plant and pollinator
abundances, so must the basin boundary, as exemplified in Fig. 4 (the various
dashed curves) for $r\in[0,1]$. Note that, since the decay parameter $\kappa$
increases from $\kappa_{\rm min}$ to $\kappa_{\rm max}$ at the linear rate
$r$, we have $\kappa(r=0)=\kappa_{\rm min}$ and $\kappa(r=\infty)=\kappa_{\rm
max}$. As $r$ increases, the basin boundaries accumulate at the one for
$\kappa_{\rm max}$.
## Theory
Figure 4 provides a physical base for deriving the scaling law (1). Consider
the small neighborhood of the unstable fixed point $\mathbf{f}_{5}$, where the
basin boundary is approximately straight, as shown schematically in Fig. 5.
Consider two cases: one with rate $r_{1}$ of parameter increase and another
with rate $r_{2}$, where $r_{2}>r_{1}$, as shown in Figs. 5(a) and 5(b),
respectively. For any given rate $r$, at the beginning and end of the
parameter variation, we have $\kappa=\kappa_{\rm min}$ and $\kappa=\kappa_{\rm
max}$, respectively, where the time it takes to complete this process is
$T=(\kappa_{\rm max}-\kappa_{\rm min})/r$. Since $r_{2}>r_{1}$, we have
$T_{2}<T_{1}$. For both Figs. 5(a) and 5(b), the unstable fixed point
$\mathbf{f}_{5}$ is marked by the filled green circles at $t=0$ and filled
orange circles at the end of the parameter variation, and the blue dashed line
with an arrow indicates the direction of change in the location of
$\mathbf{f}_{5}$ in the phase space as the parameter varies with time.
Likewise, the solid green (orange) line segments through $\mathbf{f}_{5}$
denote the boundary separating the extinction basin from the survival basin of
attraction at $t=0$ ($t=T_{1}$ or $T_{2}$). That is, before the parameter
variation is turned on for $\kappa=\kappa_{\rm min}$, the initial conditions
below (above) the solid green lines belong to the basin of the extinction
(survival) attractor. After the process of parameter variation ends so that
$\kappa=\kappa_{\rm max}$, the initial conditions below (above) the solid
orange lines belong to the basin of the extinction (survival) attractor.
During the process of parameter variation, $\mathbf{f}_{5}$ moves from the
position of the green circle to that of the orange circle, and its stable
manifold (the basin boundary) moves accordingly. Now consider the initial
conditions in the light shaded green area, which belong to the basin of the
survival attractor for $\kappa=\kappa_{\rm min}$. Without parameter variation,
as time goes, this green rectangular area will be stretched along the unstable
direction of $\mathbf{f}_{5}$ exponentially according to its unstable
eigenvalue $\lambda$ and compressed exponentially in the stable direction,
evolving into an orange rectangle that is long in the unstable direction.
Since $T_{1}>T_{2}$, the orange rectangle for $r=r_{1}$ is longer and thinner
than that for $r=r_{2}$.
The dynamical mechanism responsible for R-tipping can now be understood based
on the schematic illustration in Figs. 5(a) and 5(b). In particular, because
of the movement of $\mathbf{f}_{5}$ and the basin boundary as the parameter
variation is turned on, the dark shaded orange part of the long rectangle now
belongs to the basin of the extinction attractor. The initial conditions in
the original green rectangle which evolve into this dark shaded orange region
are nothing but the initial conditions that switch their destinations from the
survival to the extinction attractor as the result of the time variation of
the parameter. That is, these initial conditions will experience R-tipping, as
indicated by the red rectangle inside the green area in Fig. 5(b). For any
given rate $r$, the fraction of such initial conditions determines the
R-tipping probability. Let $d(0)$ denote the fraction of R-tipping initial
conditions and let $D$ be the distance between the basin boundaries at the
beginning and end of parameter variation along the unstable direction of
$\mathbf{f}_{5}$. We have $d(T)=D=d(0)\exp{(\lambda T)}$. Our argument for
$\Phi(r)\sim d(0)$ and the use of $T=(\kappa_{\rm max}-\kappa_{\rm min})/r$
lead to the scaling law (1).
## Discussion
To summarize, nonlinear dynamical systems in nature such as ecosystems and
climate systems on different scales are experiencing parameter changes due to
increasing human activities, and it is of interest to understand how the
“pace” or rate of parameter change might lead to catastrophic consequences. To
this end, we studied high-dimensional mutualistic networks, as motivated by
the following general considerations. In ecosystems, mutualistic interactions,
broadly defined as a close, interdependent, mutually beneficial relationship
between two species, are one of the most fundamental interspecific
relationships. Mutualistic networks contribute to biodiversity and ecosystem
stability. As species within these networks rely on each other for essential
services, such as pollination, seed dispersal, or nutrient exchange, they
promote species coexistence and reduce competitive exclusion. This coexistence
enhances the overall diversity of the ecosystem, making it more resilient to
disturbances and less susceptible to the dominance of a few species. They can
drive co-evolutionary processes between interacting species. As species
interact over time, they may evolve in response to each other’s adaptations,
leading to reciprocal changes that strengthen the mutualistic relationship.
Disruptions to these networks, such as the decline of pollinator populations,
can have cascading effects on ecosystem functions and the survival of
dependent species. By studying mutualistic interactions, conservationists can
design more effective strategies to protect and restore these vital
relationships and the ecosystems they support.
This paper focuses on the phenomenon of rate-induced tipping or R-tipping,
where the rate of parameter change can cause the system to experience a
tipping point from normal functioning to collapse. The main accomplishments
are three. First, we went beyond the existing local approaches to R-tipping by
taking a global approach of dynamical analysis based on consideration of
basins of attraction of coexisting attractors. This allows us to introduce the
probability of R-tipping with respect to initial conditions taken from the
whole phase space. Second, most previous works on R-tipping analyzed low-
dimensional toy models but our study focused on high-dimensional mutualistic
networks constructed from empirical data. Third, we developed a geometric
analysis and derived a scaling law governing the probability of R-tipping with
respect to the rate of parameter change. The scaling law contains two
parameters which, for two-dimensional systems, can be determined theoretically
from the dynamics. For high-dimensional systems, the two scaling parameters
can be determined through numerical fitting. For all ten empirical networks
studied, the scaling law agrees well with the results from direct numerical
simulations. To our knowledge, no such quantitative law characterizing
R-tipping has been uncovered previously.
The effective 2D systems in our study were previously derived (17), which
capture the bipartite and mutualistic nature of the ecological interactions in
the empirical, high-dimensional networks. It employs one collective variable
to account for the dynamical behavior of the pollinators and another for the
plants. The effective 2D models allow us to mathematically analyze the global
R-tipping dynamics, with results and the scaling law verified by direct
simulations of high-dimensional networks. The key reason that the dimension-
reduced 2D models can be effective lies, again, in the general setting of the
ecological systems under consideration: coexistence of a survival and an
extinction state. While our present work focused on mutualistic networks, we
expect the global approach to R-tipping and the scaling law to be generally
applicable to ecological systems in which large-scale extinction from a
survival state is possible due to even a small, nonzero rate of parameter
change.
It is worth noting that the setting under which the scaling law of the
probability of R-tipping with the rate of parameter change holds is the
coexistence of two stable steady states in the phase space, one associated
with survival and another corresponding to extinction. This setting is general
for studying tipping, system collapse and massive extinction in ecological
systems. Our theoretical analysis leading to the scaling law requires minimal
conditions: two coexisting basins of attraction separated by a basin boundary.
The scaling law is not the result of some specific parameterization of the
mutualistic systems but is a generic feature in systems with two coexisting
states. Insofar as the system can potentially undergo a transition from the
survival to the extinction state, we expect our R-tipping scaling law to hold.
In a broad context, coexisting stable steady states or attractors are
ubiquitous in nonlinear physical and biological systems.
The scaling law stipulates that, as the rate increases from zero, the
R-tipping probability increases rapidly first and then saturates. This has a
striking and potentially devastating consequence: in order to reduce the
probability of R-tipping, the parameter change must be slowed down to such an
extent that its rate of change is practically zero. This has serious
implications. For example, to avoid climate-change induced species extinction,
it would be necessary to ensure that no parameters change with time, and this
may pose an extremely significant challenge in our efforts to protect and
preserve the natural environment.
This work was supported by the Office of Naval Research under Grant No.
N00014-21-1-2323. Y. Do was supported by the National Research Foundation of
Korea (NRF) grant funded by the Korean government (MSIP) (Nos.
NRF-2022R1A5A1033624 and 2022R1A2C3011711).
## References
* Scheffer (2004) Marten Scheffer. _Ecology of Shallow Lakes_. Springer Science & Business Media, 2004.
* Scheffer et al. (2009) Marten Scheffer, Jordi Bascompte, William A Brock, Victor Brovkin, Stephen R Carpenter, Vasilis Dakos, Hermann Held, Egbert H Van Nes, Max Rietkerk, and George Sugihara. Early-warning signals for critical transitions. _Nature_ , 461(7260):53–59, 2009.
* Scheffer (2010) M. Scheffer. Complex systems: foreseeing tipping points. _Nature_ , 467:411–412, 2010.
* Wysham and Hastings (2010) D. B. Wysham and A. Hastings. Regime shifts in ecological systems can occur with no warning. _Ecol. Lett._ , 13:464–472, 2010.
* Drake and Griffen (2010) John M Drake and Blaine D Griffen. Early warning signals of extinction in deteriorating environments. _Nature_ , 467(7314):456–459, 2010.
* Chen et al. (2012) Luonan Chen, Rui Liu, Zhi-Ping Liu, Meiyi Li, and Kazuyuki Aihara. Detecting early-warning signals for sudden deterioration of complex diseases by dynamical network biomarkers. _Sci. Rep._ , 2:342, 2012.
* Boettiger and Hastings (2012) C. Boettiger and Alan Hastings. Quantifying limits to detection of early warning for critical transitions. _J. R. Soc. Interface_ , 9:2527–2539, 2012.
* Dai et al. (2012) Lei Dai, Daan Vorselen, Kirill S Korolev, and Jeff Gore. Generic indicators for loss of resilience before a tipping point leading to population collapse. _Science_ , 336(6085):1175–1177, 2012.
* Ashwin et al. (2012) Peter Ashwin, Sebastian Wieczorek, Renato Vitolo, and Peter Cox. Tipping points in open systems: bifurcation, noise-induced and rate-dependent examples in the climate system. _Philo Trans. R. Soc. A Math. Phys. Eng. Sci._ , 370(1962):1166–1184, 2012.
* Lenton et al. (2012) T. M. Lenton, V. N. Livina, V. Dakos, E. H. van Nes, and M. Scheffer. Early warning of climate tipping points from critical slowing down: comparing methods to improve robustness. _Phil. Trans. Roy. Soc. A_ , 370:1185–1204, 2012.
* Barnosky et al. (2012) A. D. Barnosky, E. A. Hadly, J. Bascompte, E. L. Berlowand J. H. Brown, M. Fortelius, W. M. Getz, J. Harte, A. Hastings, P. A. Marquet, N. D. Martinez, A. Mooers, P. Roopnarine, G. Vermeij, J. W. Williams, R. Gillespie, J. Kitzes, C. Marshall, N. Matzke, D. P. Mindell, E. Revilla, and A. B. Smith. Approaching a state shift in earth’s biosphere. _Nature_ , 486:52–58, 2012.
* Boettiger and Hastings (2013) C. Boettiger and A. Hastings. Tipping points: From patterns to predictions. _Nature_ , 493:157–158, 2013.
* Tylianakis and Coux (2014) Jason M Tylianakis and Camille Coux. Tipping points in ecological networks. _Trends. Plant. Sci._ , 19(5):281–283, 2014.
* Lever et al. (2014) J Jelle Lever, Egbert H Nes, Marten Scheffer, and Jordi Bascompte. The sudden collapse of pollinator communities. _Ecol. Lett._ , 17(3):350–359, 2014.
* Lontzek et al. (2015) T. S. Lontzek, Y.-Y. Cai, K. L. Judd, and T. M. Lenton. Stochastic integrated assessment of climate tipping points indicates the need for strict climate policy. _Nat. Clim. Change_ , 5:441–444, 2015.
* Gualdia et al. (2015) S. Gualdia, M. Tarziaa, F. Zamponic, and J.-P. Bouchaudd. Tipping points in macroeconomic agent-based models. _J. Econ. Dyn. Contr._ , 50:29–61, 2015.
* Jiang et al. (2018) Junjie Jiang, Zi-Gang Huang, Thomas P Seager, Wei Lin, Celso Grebogi, Alan Hastings, and Ying-Cheng Lai. Predicting tipping points in mutualistic networks through dimension reduction. _Proc. Nat. Acad. Sci. (UsA)_ , 115(4):E639–E647, 2018.
* Yang et al. (2018) Biwei Yang, Meiyi Li, Wenqing Tang, Si Liu, Weixinand Zhang, Luonan Chen, and Jinglin Xia. Dynamic network biomarker indicates pulmonary metastasis at the tipping point of hepatocellular carcinoma. _Nat. Commun._ , 9:678, 2018.
* Jiang et al. (2019) Junjie Jiang, Alan Hastings, and Ying-Cheng Lai. Harnessing tipping points in complex ecological networks. _J. R. Soc. Interface_ , 16(158):20190345, 2019\.
* Scheffer (2020) Marten Scheffer. _Critical Transitions in Nature and Society_ , volume 16. Princeton University Press, 2020.
* Meng et al. (2020a) Yu Meng, Junjie Jiang, Celso Grebogi, and Ying-Cheng Lai. Noise-enabled species recovery in the aftermath of a tipping point. _Phys. Rev. E_ , 101(1):012206, 2020a.
* Meng et al. (2020b) Yu Meng, Ying-Cheng Lai, and Celso Grebogi. Tipping point and noise-induced transients in ecological networks. _J. R. Soc. Interface_ , 17(171):20200645, 2020b.
* Meng and Grebogi (2021) Yu Meng and Celso Grebogi. Control of tipping points in stochastic mutualistic complex networks. _Chaos_ , 31(2):023118, 2021.
* Meng et al. (2022) Yu Meng, Ying-Cheng Lai, and Celso Grebogi. The fundamental benefits of multiplexity in ecological networks. _J. R. Soc. Interface_ , 19:20220438, 2022.
* O’Keeffe and Wieczorek (2020) Paul E O’Keeffe and Sebastian Wieczorek. Tipping phenomena and points of no return in ecosystems: beyond classical bifurcations. _SIAM J. Appl. Dyn. Syst._ , 19(4):2371–2402, 2020.
* Trefois et al. (2015) Christophe Trefois, Paul MA Antony, Jorge Goncalves, Alexander Skupin, and Rudi Balling. Critical transitions in chronic disease: transferring concepts from ecology to systems medicine. _Cur. Opin. Biotechnol._ , 34:48–55, 2015.
* Albrich et al. (2020) Katharina Albrich, Werner Rammer, and Rupert Seidl. Climate change causes critical transitions and irreversible alterations of mountain forests. _Global Change Biol._ , 26(7):4013–4027, 2020\.
* Bayani et al. (2017) Atiyeh Bayani, Fatemeh Hadaeghi, Sajad Jafari, and Greg Murray. Critical slowing down as an early warning of transitions in episodes of bipolar disorder: A simulation study based on a computational model of circadian activity rhythms. _Chronobiol. Int._ , 34(2):235–245, 2017.
* Synodinos et al. (2022) Alexis D Synodinos, Rajat Karnatak, Carlos A Aguilar-Trigueros, Pierre Gras, Tina Heger, Danny Ionescu, Stefanie Maaß, Camille L Musseau, Gabriela Onandia, Aimara Planillo, et al. The rate of environmental change as an important driver across scales in ecology. _Oikos_ , page e09616, 2022.
* Morris et al. (2002) James T Morris, PV Sundareshwar, Christopher T Nietch, Björn Kjerfve, and Donald R Cahoon. Responses of coastal wetlands to rising sea level. _Ecology_ , 83(10):2869–2877, 2002.
* Ritchie et al. (2023) Paul DL Ritchie, Hassan Alkhayuon, Peter M Cox, and Sebastian Wieczorek. Rate-induced tipping in natural and human systems. _Earth Syst. Dynam_ , 14(3):669–683, 2023.
* Wieczorek et al. (2011) Sebastian Wieczorek, Peter Ashwin, Catherine M Luke, and Peter M Cox. Excitability in ramped systems: the compost-bomb instability. _Proc. R. Soc. A Math. Phys. Eng. Sci._ , 467(2129):1243–1269, 2011.
* Mitry et al. (2013) John Mitry, Michelle McCarthy, Nancy Kopell, and Martin Wechselberger. Excitable neurons, firing threshold manifolds and canards. _J. Math. Neurosci._ , 3(1):1–32, 2013.
* Alexander et al. (2011) NA Alexander, O Oddbjornsson, CA Taylor, HM Osinga, and David E Kelly. Exploring the dynamics of a class of post-tensioned, moment resisting frames. _J. Sound Vib._ , 330(15):3710–3728, 2011.
* Hsu and Shih (2015) Sheng-Yi Hsu and Mau-Hsiang Shih. The tendency toward a moving equilibrium. _SIAM J. Appl. Dyn. Sys._ , 14(4):1699–1730, 2015\.
* Ashwin et al. (2017) Peter Ashwin, Clare Perryman, and Sebastian Wieczorek. Parameter shifts for nonautonomous systems in low dimension: bifurcation-and rate-induced tipping. _Nonlinearity_ , 30(6):2185, 2017.
* Vanselow et al. (2019) Anna Vanselow, Sebastian Wieczorek, and Ulrike Feudel. When very slow is too fast-collapse of a predator-prey system. _J. Theo. Biol._ , 479:64–72, 2019.
* Lande et al. (2003) Russell Lande, Steinar Engen, and Bernt-Erik Saether. _Stochastic population dynamics in ecology and conservation_. Oxford University Press on Demand, 2003.
* Hansen et al. (2019) Brage B Hansen, Marlène Gamelon, Steve D Albon, Aline M Lee, Audun Stien, R Justin Irvine, Bernt-Erik Sæther, Leif E Loe, Erik Ropstad, Vebjørn Veiberg, et al. More frequent extreme climate events stabilize reindeer population dynamics. _Nat. Commun_ , 10(1):1616, 2019.
* Hastings et al. (2018) Alan Hastings, Karen C Abbott, Kim Cuddington, Tessa Francis, Gabriel Gellner, Ying-Cheng Lai, Andrew Morozov, Sergei Petrovskii, Katherine Scranton, and Mary Lou Zeeman. Transient phenomena in ecology. _Science_ , 361(6406), 2018.
* Morozov et al. (2020) Andrew Morozov, Karen Abbott, Kim Cuddington, Tessa Francis, Gabriel Gellner, Alan Hastings, Ying-Cheng Lai, Sergei Petrovskii, Katherine Scranton, and Mary Lou Zeeman. Long transients in ecology: theory and applications. _Phys. Life Rev._ , 32:1–40, 2020.
* Bascompte et al. (2003) J. Bascompte, P. Jordano, C. J. Melián, and J. M. Olesen. The nested assembly of plant-animal mutualistic networks. _Proc. Natl. Acad. Sci. (USA)_ , 100:9383–9387, 2003.
* Guimaraes et al. (2011) P. R. Guimaraes, P. Jordano, and J. N. Thompson. Evolution and coevolution in mutualistic networks. _Ecol. Lett._ , 14:877–885, 2011.
* Nuismer et al. (2013) S. L. Nuismer, P. Jordano, and J. Bascompte. Coevolution and the architecture of mutualistic networks. _Evolution_ , 67:338–354, 2013.
* Rohr et al. (2014) Rudolf P Rohr, Serguei Saavedra, and Jordi Bascompte. On the structural stability of mutualistic systems. _Science_ , 345(6195):1253497, 2014.
* Dakos and Bascompte (2014) Vasilis Dakos and Jordi Bascompte. Critical slowing down as early warning for the onset of collapse in mutualistic communities. _Proc. Natl. Acad. Sci. (USA)_ , 111(49):17546–17551, 2014.
* Guimaraes et al. (2017) P. R. Guimaraes, M. M. Pires, P. Jordano, J. Bascompte, and J. N. Thompson. Indirect effects drive coevolution in mutualistic networks. _Nature_ , 550:511–514, 2017.
* Ohgushi et al. (2012) T. Ohgushi, O. Schmitz, and R. D. Holt. _Trait-Mediated Indirect Interactions: Ecological and Evolutionary Perspectives_. Cambridge Univ. Press, Cambridge UK, 2012.
* Bastolla et al. (2009) Ugo Bastolla, Miguel A Fortuna, Alberto Pascual-García, Antonio Ferrera, Bartolo Luque, and Jordi Bascompte. The architecture of mutualistic networks minimizes competition and increases biodiversity. _Nature_ , 458(7241):1018–1020, 2009.
* Bascompte and Jordano (2007) Jordi Bascompte and Pedro Jordano. Plant-animal mutualistic networks: the architecture of biodiversity. _Annu. Rev. Ecol. Evol. Syst._ , 38:567–593, 2007.
* Menck et al. (2013) Peter J Menck, Jobst Heitzig, Norbert Marwan, and Jürgen Kurths. How basin stability complements the linear-stability paradigm. _Nature physics_ , 9(2):89–92, 2013.
* Holling (1973) Crawford S Holling. Resilience and stability of ecological systems. _Ann. Rev. Ecol. Systema_ , pages 1–23, 1973.
* Schleuning et al. (2016) Matthias Schleuning, Jochen Fründ, Oliver Schweiger, Erik Welk, Jörg Albrecht, Matthias Albrecht, Marion Beil, Gita Benadi, Nico Blüthgen, Helge Bruelheide, et al. Ecological networks are more sensitive to plant than to animal extinction under climate change. _Nat. Commun_ , 7(1):13965, 2016.
* Vanbergen et al. (2017) Adam J Vanbergen, Ben A Woodcock, Matthew S Heard, and Daniel S Chapman. Network size, structure and mutualism dependence affect the propensity for plant–pollinator extinction cascades. _Funct. Ecol._ , 31(6):1285–1293, 2017.
* Lewinsohn et al. (2006) Thomas M Lewinsohn, Paulo Inácio Prado, Pedro Jordano, Jordi Bascompte, and Jens M. Olesen. Structure in plant–animal interaction assemblages. _Oikos_ , 113(1):174–184, 2006.
* Vázquez et al. (2009) Diego P. Vázquez, Nico Blüthgen, Luciano Cagnolo, and Natacha P. Chacoff. Uniting pattern and process in plant–animal mutualistic networks: a review. _Ann. Bot._ , 103(9):1445–1457, 03 2009. ISSN 0305-7364. 10.1093/aob/mcp057. URL https://doi.org/10.1093/aob/mcp057.
* Aslan et al. (2013) Clare E Aslan, Erika S Zavaleta, Bernie Tershy, and Donald Croll. Mutualism disruption threatens global plant biodiversity: a systematic review. _PLoS one_ , 8(6):e66993, 2013.
* Sheykhali et al. (2020) Somaye Sheykhali, Juan Fernández-Gracia, Anna Traveset, Maren Ziegler, Christian R Voolstra, Carlos M Duarte, and Víctor M Eguíluz. Robustness to extinction and plasticity derived from mutualistic bipartite ecological networks. _Sci. Rep._ , 10(1):9783, 2020.
* Pires et al. (2020) Mathias M Pires, James L O’Donnell, Laura A Burkle, Cecilia Díaz-Castelazo, David H Hembry, Justin D Yeakel, Erica A Newman, Lucas P Medeiros, Marcus AM de Aguiar, and Paulo R Guimarães Jr. The indirect paths to cascading effects of extinctions in mutualistic networks, 2020.
* Young et al. (2021) Jean-Gabriel Young, Fernanda S Valdovinos, and MEJ Newman. Reconstruction of plant–pollinator networks from observational data. _Nat. Commun_ , 12(1):3911, 2021.
* Gao et al. (2016) Jianxi Gao, Baruch Barzel, and Albert-László Barabási. Universal resilience patterns in complex networks. _Nature_ , 530(7590):307–312, 2016.
|
# End-to-End Optimization of JPEG-Based Deep Learning Process for Image
Classification ††thanks: This material is based upon work supported by Google
Cloud.
Siyu Qi, Lahiru D. Chamain, and Zhi Ding Department of Electrical and
Computer Engineering
University of California, Davis, CA, 95616
Email: {syqi, hdchamain<EMAIL_ADDRESS>
###### Abstract
Among major deep learning (DL) applications, distributed learning involving
image classification require effective image compression codecs deployed on
low-cost sensing devices for efficient transmission and storage. Traditional
codecs such as JPEG designed for perceptual quality are not configured for DL
tasks. This work introduces an integrative end-to-end trainable model for
image compression and classification consisting of a JPEG image codec and a
DL-based classifier. We demonstrate how this model can optimize the widely
deployed JPEG codec settings to improve classification accuracy in
consideration of bandwidth constraint. Our tests on CIFAR-100 and ImageNet
also demonstrate improved validation accuracy over preset JPEG configuration.
###### Index Terms:
JPEG, joint compression and classification, end-to-end optimization.
## I Introduction
In recent years, deep convolutional neural networks (CNNs) have demonstrated
successes in learning tasks such as image classification and recognition,
owing to their capability of extracting image features among adjacent pixels.
The emergence of residual network (ResNet)[1] further enhanced image
classification without introducing extra computational complexity.
In the era of IoT and cloud computing, many practical applications rely on
widely deployed low-cost cameras and sensors for data collection before
transmitting sensor data to powerful cloud or edge servers that host pre-
trained deep classifiers. As most (RF) network links usually are severely
band-limited and must prioritize heavy data traffic, image compression
techniques are vital for efficient and effective utilization of limited
network bandwidth and storage resources. JPEG [2] is a highly popular codec
standard for lossy image compression, widely used to conserve bandwidth in
source data transmission and storage. The JPEG encoding process includes
discrete cosine transform (DCT) and quantization. The quantized integer DCT
coefficients are encoded via run-length encoding (RLE) and Huffman coding. Due
to RLE, total bit rate of an image cannot be predicted straightforwardly[3, 4,
5]. The JPEG encoding achieves substantial image compression ratio with little
human perception quality sacrifice. These encoded bits are then transmitted
over a channel of limited capacity (bit rate) before decoding and recovery for
various applications.
Targeting human users, the parameters in JPEG configuration are selected
according to visualization subjective tests. However, in CNN-based image
classifications, naïve adoption of the lossy JPEG image encoding, designed
primarily for human visualization needs, can lead to unexpected accuracy loss
because the traditional CNN models are agnostic of the compression distortion.
To tackle this issue, this work is motivated by the obvious and important
question in distributed AI: How to optimally (re)configure standardized JPEG
for image compression to improve DL-based image classification.
Motivated by the strong need to conserve network bandwidth and local storage
for remote image classification, we present an end-to-end trainable DL model
for joint image compression and classification that can optimize the widely
deployed JPEG codec to improve classification accuracy over current JPEG
settings. We formulate this dual-objective problem as a constrained
optimization problem which maximizes classification accuracy subject to a
compression ratio constraint. We incorporate trainable JPEG compression blocks
and JPEG decoding blocks together with the trainable CNN classifier in our
end-to-end learning model. Our proposed DL model can configure JPEG encoding
parameters to achieve high classification accuracy.
We organize the rest of the paper as follows. Section II introduces the basics
of JPEG codec and summarizes related works. Section III proposes the novel
end-to-end DL architecture. We provide experimental results in Section IV,
before concluding and discussing potential future directions in Section V.
## II JPEG Codec and Learning over JPEG
### II-A JPEG Codec in View of Deep Learning
In JPEG compression with 4:2:0 chroma subsampling, an RGB source image is
first converted to YCBCR color space through linear transformations.
Chrominance channels (CB and CR) are subsampled by 2 both vertically and
horizontally. After subsampling, each of the 3 YCBCR channels is split into
non-overlapping $8\times 8$ blocks before applying blockwise DCT. The
2-dimensional (2-D) DCT of an image block $\boldsymbol{I}$ of size $N\times N$
with entries $\boldsymbol{I}(k,l)$ is defined by $N\times N$ block
$\boldsymbol{F}=\boldsymbol{DID}^{T}$, where $\boldsymbol{D}$ is a constant
matrix. The DCT is capable of compacting image features with a small number of
DCT coefficients with little perceptible loss after compression.
For compression, each block of the $8\times 8$ frequency-domain coefficients
$\boldsymbol{F}$ is quantized using pre-defined quantization matrices, or
“Q-tables” $\boldsymbol{Q}$ with entries $\boldsymbol{Q}(j,k)$ at JPEG encoder
to obtain quantized block $\boldsymbol{F}_{q}$ whose entries are
$\boldsymbol{F}_{q}(j,k)=\mbox{round}[{\boldsymbol{F}(j,k)}/{\boldsymbol{Q}(j,k)}]$.
The decoder reconstructs from the compressed block $\boldsymbol{F}_{q}$ and
the Q-table to form Hadamard product
$\boldsymbol{F}_{q}=\boldsymbol{F}\circ\boldsymbol{Q}$. Decoder would then use
inverse DCT (IDCT) to recover spatial RGB images. Parameters in
$\boldsymbol{Q}$ can be adjusted to achieve different compression levels and
visual effects. JPEG standard provides two Q-tables to adjust compression
loss, one for the Y channel and another for CB and CR channels.
In networked image applications, training using full resolution images would
make little practical sense and would be prone to accuracy loss because only
codec-compressed image data are available at the cloud/edge processing node.
Thus, deep learning networks should directly use compressed DCT coefficients
as inputs for both training and inference instead of full resolution RGB
images for training. Image classification (labeling) directly based on DCT
coefficients can further reduce decoder computation during both training and
inference by skipping the IDCT and potentially achieve better robustness under
dynamic levels of JPEG compression.
Importantly, ResNets that were successfully developed for recognition of fully
reconstructed JPEG images tend to exhibit performance loss if they are
directly used on image data in DCT domain. Motivated by the need to improve
image processing performance in networked environments under channel bandwidth
and storage constraints, this work investigates deep learning architecture
designs suitable for optimizing standard compliant JPEG configurations to
achieve high classification accuracy and low bandwidth consumption by directly
applying DCT input data. Our joint optimization of the JPEG configuration is
achieved by optimizing both the JPEG Q-table parameters and the deep learning
classifier to achieve end-to-end deep learning framework spanning from the IoT
source encoder to the cloud classifier. Our experiments include tests on the
high resolution ImageNet dataset.
### II-B Related Works
For bandwidth and storage conservation, DL architectures such as auto-encoders
have been effectively trained[6, 7, 8] for image compression with little
degradation of classification accuracy or perceptual quality. Previous works
[9, 10] also revealed direct training of DL models on the DCT coefficients
using faster CNN structure with a modified ResNet-50 architecture. The authors
of [11] developed a joint compression and classification network model based
on JPEG2000 encoding. These works suggest the benefit of DL-based end-to-end
optimization of image codecs.
Previous studies [12, 13, 14, 5] have also recognized the importance of
Q-tables in JPEG codecs and seek to optimize them for DL-based image
classification. [13, 12] propose to design JPEG Q-tables based on the
importance of DCT coefficients, evaluated by the relative frequency [13] or
the standard deviation [12] of the coefficients. Both [14, 5] offer end-to-end
DL models to estimate a set of optimized Q-tables for each input image, where
models are pre-trained to predict the bandwidth of each image. In contrast,
targeting low-cost sensing nodes, our proposed model learns a single set of
Q-tables for all images during training, which can be pre-configured within
the JPEG codec after training for inference tasks. This reduces the required
computational power at the sensing nodes. Moreover, our proposed training
tunes Q-tables and does not require a separate entropy estimation model.
To our best knowledge, there exists no published work on JPEG Q-table
optimization for distributed learning to target low-cost sensing devices.
Since JPEG continues to be a commonly used image coding methods in massive
number of low-cost devices, we focus our investigation on the rate-accuracy
trade-off to facilitate their widespread applications in distributed learning.
## III Joint DL Architectures
### III-A Wide ResNet (WRN) for CIFAR-100 and Tiny ImageNet
For CIFAR-100 and Tiny ImageNet, we propose the WRN model of Fig. 1. Following
JPEG , our preprocessing steps include level shifting, color transformation,
subsampling, and DCT. The invertible color transformation can also be trained.
Figure 1: Proposed architecture for CIFAR-100 and Tiny ImageNet. The detailed
structure in WRN-28 is omitted here.
#### III-A1 Compression Layers
The trainable quantization layer in our proposed WRN model is similar to the
“quan block” in [11], as shown in Fig. 1. JPEG can take three distinct
Q-tables $\boldsymbol{Q}_{1}$, $\boldsymbol{Q}_{2}$, and $\boldsymbol{Q}_{3}$,
respectively, for each of the YCBCR color channels, leading to 192 parameters
in this layer. Same as [11], we replace the trainable quantization parameters
by element-wise reciprocal of the Q-table entries
$\boldsymbol{q}_{i}(j,k)=\boldsymbol{Q}_{i}(j,k)^{-1}.$ The matrices
$\boldsymbol{q}_{i}$, referred to as the “compression kernels”, allows the
model to learn to discard any frequency domain information by setting the
corresponding entry $\boldsymbol{q}_{i}(j,k)=0$. Smaller $\boldsymbol{q}_{i}$
values lead to smaller range of quantized DCT coefficients and consequently
generates fewer encoded bits.
The quantization layer includes a non-differentiable rounding operation
$a(\boldsymbol{F})=\mbox{round}(\boldsymbol{F})$, which cannot be used in a
gradient-based training framework, as its activation function. Following [15],
we address this problem by substituting a smooth approximation
$\hat{a}(\boldsymbol{F})=\boldsymbol{F}$ for the rounding function in
backpropagation. Together, preprocessing and quantization layers form a JPEG-
based encoder.
The dequantization layer only needs to multiply the encoded DCT blocks
$\boldsymbol{F}_{q}$ element-wise by their respective Q-table matrices
$\boldsymbol{Q}_{1}$, $\boldsymbol{Q}_{2}$ and $\boldsymbol{Q}_{3}$
corresponding to the encoder quantization layer. The quantization and
dequantization layers jointly form a pair of “compression layers”.
#### III-A2 Reconstruction Layer
As described in Section II, the reconstruction layer performs IDCT via
$\boldsymbol{I}=\boldsymbol{D}^{T}\boldsymbol{F}_{q}\boldsymbol{D}$ for each
quantized DCT coefficient block $\boldsymbol{F}_{q}$ and rearranges the
reconstructed blocks $\boldsymbol{I}$. CB and CR channels are upsampled via
bilinear interpolation. The outputs of this layer are YCBCR spatial images.
Together, the dequantization layer and the reconstruction layer form the JPEG
decoder.
#### III-A3 Classifier
WRNs[16] achieve impressive classification performance on the CIFAR-100[17]
and Tiny ImageNet[18] datasets. Without loss of generality, we adopt a
28-layer WRN as the classifier. For CIFAR-100, we set the convolutional layer
width multiplier $k=10$, same as that used in [16]. For Tiny ImageNet, we set
$k=1$ to further simplify training.
### III-B Loss Function During Training
To jointly reconfigure the JPEG parameters in compression layers and to
optimize the deep learning classifier, we design the following loss function
during training:
$\displaystyle\mathcal{L}=\mathcal{L}_{\rm CLA}+\lambda\mathcal{L}_{\rm
Quan},\vspace*{-2mm}$
where $\mathcal{L}_{\rm CLA}$ is the cross entropy classification loss,
$\mathcal{L}_{\rm Quan}$ is a penalty term for the compression kernels and
$\lambda$ is an adjustable hyper-parameter to govern the importance of the
rate. Since there is no simple or standard metric to directly control JPEG
encoded image size, which further involves RLE and Huffman coding, we propose
the following surrogate penalty function:
$\displaystyle\mathcal{L}_{\rm
Quan}=\sum_{i=1}^{3}\sum_{j=1}^{8}\sum_{k=1}^{8}$
$\displaystyle\left[\mbox{max}(\boldsymbol{q}_{i}^{2}(j,k)-c,0)+\lambda_{1}|\boldsymbol{q}_{i}(j,k)|\right]$
where $c$ and $\lambda_{1}$ are tunable hyper-parameters. The $\ell_{1}$ loss
term promotes sparsity whereas the $\ell_{2}$ loss term regulates the
compression kernels. The hyper-parameter $c$ acts as a constraint on the
squared magnitude of $\boldsymbol{q}_{i}$, shall be appropriately selected
based on the values of $\lambda$ and $\lambda_{1}$. Larger $\lambda$ and
$\lambda_{1}$ and a smaller $c$ leads to higher compression ratio and lower
classification accuracy. We propose the current form of the surrogate penalty
function after testing both logarithm and sigmoid functions without witnessing
performance benefits.
### III-C Modified ResNet for ImageNet
For ImageNet, we adopt the same preprocessing steps and quantization layer
from III-A and utilize the Deconvolution-RFA architecture in [10], which is
inspired by ResNet-50, as the classifier. As demonstrated in Fig. 2, the
quantized DCT coefficients of CB and CR channels are augmented to the same
spatial size as Y channel by two separate transposed convolutional layers. The
three channels are concatenated as input of the deconvolution-RFA model.
Considering the higher complexity of this model, we suggest a single
$\ell_{2}$ regularization for compression kernels for optimizing the
quantization parameters via quantization loss $\mathcal{L}_{\rm
Quan}=\sum_{i=1}^{3}\sum_{j=1}^{8}\sum_{k=1}^{8}\boldsymbol{q}_{i}^{2}(j,k).$
Figure 2: Architecture of our proposed model for ImageNet. The detailed
structure of Deconvolution-RFA[10] is omitted here.
### III-D Implementation
We test the learning framework in Fig. 1 with CIFAR-100 and Tiny ImageNet
datasets. The images in CIFAR-100 dataset are of $32\times 32$ pixels, while
images in Tiny ImageNet are of $64\times 64$ pixels.
There are 192 trainable parameters in the compression kernels, all initialized
to 1. The randomly-initialized WRN-28 classifier is trained jointly with the
parameters in compression kernels, as well as the color transformation
coefficients if needed, using Adam optimizer with a batch size of 100. The
training of WRN-28 proceeds alternatively: the classifier is trained for 2
epochs while JPEG-based layers are frozen, followed by the compression layer
being trained for 1 epoch while freezing the classifier. The training takes
150 such alternations. The learning rate starts from 0.05, and is scaled by
0.1 and 0.01, at alternation 50 and 100, respectively.
We implement the larger learning framework in Fig. 2 with ImageNet in which
the images are of $224\times 224$ pixels. Similarly, color transformation
coefficients are initialized to JPEG standard and all 192 quantization
parameters are initialized to 1. The color transformation, compression
kernels, and Deconvolution-RFA classifier are trained from end to end by using
Adam optimizer with a batch size of 32. The learning rate of compression
kernels starts from $1\times 10^{-8}$ while that for other parameters starts
from 0.001. Both learning rates are scaled by 0.1, 0.01, and 0.001, at epoch
30, 60 and 80, respectively. The training takes 90 epochs.
For the quantized DCT coefficients using the new Q-tables, we customize
Huffman tables for bandwidth saving by randomly choosing 50k images from the
corresponding training set to generate Huffman tables. The bandwidth of the
validation data set is measured by combining the principles of RLE and Huffman
coding.
## IV Experiments
Our experiments are conducted on Keras and TensorFlow. We utilize the two
metrics to evaluate perceptual quality of images: peak signal-to-noise ratio
(PSNR) and structural similarity index measure (SSIM) index. We test the
proposed joint compression and classification (JCC) frameworks on three
datasets: (a) CIFAR-100 dataset based on 50k training images and 10k test
images belonging to 100 categories; (b) Tiny ImageNet dataset based on 100k
training images and 10k validation images belonging to 200 categories; (c)
ImageNet dataset based on approximately 1.3M training images and 50k test
images belonging to 1000 categories.
### IV-A JPEG Standard Baseline
We first present baseline results of images using standard JPEG algorithm with
4:2:0 chroma subsampling. We initialize compression kernels
$\boldsymbol{q}_{i}$, $i=1,\ 2,\ 3$ using the Q-tables given in JPEG standard.
In this baseline scenario, we only adapt the classifier parameters during
training.
For CIFAR-100, we consider 9 different JPEG image qualities between $12.5$%
and $100\%$. For Tiny ImageNet, we select 4 different image qualities between
$10$% and $80\%$. For ImageNet, our experiments consider 5 different image
qualities between $12.5$% and $100\%$. The classification results are shown in
Figs. 3-4. These baseline results reveal that the classification accuracy
correlates positively with image qualities and average image bandwidth (rate).
### IV-B Joint Compression and Classification (JCC)
For JCC, we train and optimize the compression kernels and the classifier. For
CIFAR-100, we selected 8 values of $\lambda$ from $10^{-6}$ to $5$ with
$\lambda_{1}=1$ and $c=0.01/\lambda$. The classification accuracy, PSNR and
SSIM results can be found in Fig. 3. Compared with the JPEG baseline, JCC
achieves clear improvement of up to 2.4% in accuracy at bandwidth between 0.75
and 1.5 KB per image. The PSNR and SSIM of JCC-compressed images are similar
to those using JPEG.
(a)
(b)
(c)
Figure 3: Comparison between JCC and JPEG standard on CIFAR-100 and Tiny
ImageNet datasets with WRN-28 Model, with respect to classification accuracy,
PSNR and SSIM versus average bandwidth.
For Tiny ImageNet, we select 6 values of $c$ from $5\times 10^{-3}$ to $0.8$.
As shown in Fig. 3, when comparing with the JPEG baseline, we observe accuracy
gain of up to 4% by the proposed JCC model at low bandwidth between 0.9 and
1.6 KB per image while maintaining similar visual quality. Overall, the PSNR
and SSIM of JCC-optimized and JPEG standard quantization tables are similar.
For ImageNet, we consider 9 values of $\lambda$ between $25$ and $100$. The
resulting top-5 classification accuracy, PSNR, and SSIM are given in Fig. 4.
For encoding rates below 11 KB per image, the JCC model outperforms the
baseline by up to $3.7\%$ in terms of classification accuracy. For bandwidths
above 11 KB per image, the classification accuracy difference between JPEG and
JCC is quite insignificant. Furthermore, PSNR and SSIM of JCC-compressed
images outperform those of standard JPEG encoded images.
(a)
(b)
(c)
Figure 4: Comparison between JCC and JPEG baseline on ImageNet dataset with
Deconvolution-RFA Model, with respect to top-5 classification accuracy, PSNR
and SSIM versus average bandwidth.
From these experimental results, we observe that the JCC model can effectively
optimize the JPEG compression kernels for better rate-accuracy trade-off,
especially at moderate image bit rates. It is intuitive that the performance
edge of JCC diminishes for very high image sizes because most image features
can be preserved when given sufficient number of bits and JCC and JPEG no
longer need to delicately balance the rate-accuracy trade-off. Similarly, for
very low image sizes, very few bits can be used to encode vital information in
DCT coefficients. Hence, the encoders have less flexibility to further
optimize the rate-accuracy trade-off, thereby making it difficult for even the
JCC model to find better parameter settings.
### IV-C JCC and Color Transformation
Considering JCC and color transformation (JCC-color), color transformation
coefficients, compression kernels and classifier are jointly trained. We
initialize the color conversion coefficients according to the settings in
JPEG. According to our experimental results, tuning the color transformation
coefficients offers no performance gain over JCC. Theoretically, invertible
color space transformation does not lead to information loss and can be
subsumed by first dense-layer in the neural network. In fact, we observe that
the resulting color transform rarely move from their initial values and
further optimization is unnecessary.
### IV-D Further Analysis and Discussions
(a) Figure 5: JCC optimized compression kernels based on CIFAR-100 for Y
(left), CB (middle) and CR (right) channels, with
$\lambda=0.001,\lambda_{1}=1$ and $c=10$.
Our experimental results suggest that trainable JCC model can extract critical
DCT features for classification among categories of CIFAR-100, Tiny ImageNet,
and ImageNet datasets. Furthermore, the perceptual quality of images are
preserved. Fig. 5 shows the resulting compression kernels with which WRN-28
achieves classification accuracy of 75.20% at an average rate of 0.979
KB/image. Darker grids imply low compression or higher importance of the
corresponding DCT coefficient. The encoder clearly favors lower frequency
bands. Since there are longer consecutive 0’s in the zig-zag order in the end-
to-end learned compression kernels, the correspondingly compressed DCT
coefficients require fewer bits via RLE. Furthermore, the trainable model
learns to discard some higher frequency DCT components less critical to
classification.
Together, these experimental results show performance enhancement on
standardized JPEG codec for cloud-based image classification. The optimized Q
tables can be distributed in pre-installed JPEG encoders of low-cost devices
through software updates for different encoding sizes.
## V Conclusions
We present an end-to-end deep learning (DL) architecture to jointly optimize
JPEG image compression and classification for low-cost sensors in distributed
learning systems. Results on CIFAR-100, Tiny ImageNet, and ImageNet datasets
demonstrate successful training of the end-to-end DL framework for better
image compression and classification performance without perceptual quality
loss. Optimized JPEG Q-tables can be readily incorporated within deployed
codecs in practice. Future works may explore the broad appeal of this end-to-
end learning principle in other bandwidth-constrained distributed DL tasks
such as object detection, segmentation, and tracking.
## References
* [1] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” pp. 770–778, 2016.
* [2] C. R. T.81, “Information technology - Digital compression and coding of continuous-tone still images - Requirements and guidelines.” Standard, 1992\.
* [3] S.-W. Wu and A. Gersho, “Rate-constrained picture-adaptive quantization for JPEG baseline coders,” in _Int. Conf. on Acoustics, Speech, and Signal Processing_. IEEE, 1993, pp. 389–392.
* [4] E. Tuba, M. Tuba, D. Simian, and R. Jovanovic, “JPEG quantization table optimization by guided fireworks algorithm,” in _Int. Workshop on Combinatorial Image Analysis_. Springer, 2017, pp. 294–307.
* [5] X. Luo, H. Talebi, F. Yang, M. Elad, and P. Milanfar, “The rate-distortion-accuracy tradeoff: JPEG case study,” _arXiv:2008.00605_ , 2020.
* [6] S. Singh, S. Abu-el Haija, N. Johnston, J. Ballé, A. Shrivastava, and G. Toderici, “End-to-end learning of compressible features,” _arXiv:2007.11797_ , 2020.
* [7] L. Theis, W. Shi, A. Cunningham, and F. Husz$\acute{a}$r, “Lossy image compression with compressive autoencoders,” 2017.
* [8] S. Diamond, V. Sitzmann, S. Boyd, G. Wetzstein, and F. Heide, “Dirty pixels: Optimizing image classification architectures for raw sensor data,” 2017.
* [9] X. Zou, X. Xu, C. Qing, and X. Xing, “High speed deep networks based on discrete cosine transformation,” in _IEEE Int. Conf. on Image Proc._ , 2014, pp. 5921–5925.
* [10] L. Gueguen, A. Sergeev, B. Kadlec, R. Liu, and J. Yosinski, “Faster neural networks straight from JPEG,” in _Advances in Neural Information Processing Systems_ , 2018, pp. 3937–3948.
* [11] L. D. Chamain, S. S. Cheung, and Z. Ding, “Quannet: Joint image compression and classification over channels with limited bandwidth,” in _2019 IEEE Int. Conf. on Multimedia and Expo (ICME)_ , 2019, pp. 338–343.
* [12] Z. Liu, T. Liu, W. Wen, L. Jiang, J. Xu, Y. Wang, and G. Quan, “DeepN-JPEG: A deep neural network favorable JPEG-based image compression framework,” _Proc. 55th Annual Design Automation Conf._ , 2018.
* [13] Z. Li, C. D. Sa, and A. Sampson, “Optimizing JPEG quantization for classification networks,” _arXiv:2003.02874_ , 2020.
* [14] J. Choi and B. Han, “Task-aware quantization network for JPEG image compression,” in _European Conf. on Computer Vision_. Springer, 2020, pp. 309–324.
* [15] L. Theis, W. Shi, A. Cunningham, and F. Huszár, “Lossy image compression with compressive autoencoders,” _arXiv:1703.00395_ , 2017.
* [16] S. Zagoruyko and N. Komodakis, “Wide residual networks,” _arXiv:1605.07146_ , 2016.
* [17] A. Krizhevsky, “Learning multiple layers of features from tiny images,” Tech. Rep., 2009.
* [18] Y. Le and X. Yang, “Tiny imagenet visual recognition challenge,” _CS 231N_ , vol. 7, 2015.
|
# Double semidirect products and
skew left braces of size $np$
Teresa Crespo(1), Daniel Gil-Muñoz(2), Anna Rio(3)
and Montserrat Vela(3)
###### Abstract
We define the double semidirect product of skew left braces and prove that if
$p$ is an odd prime and $n$ is an integer such that $p\nmid n$ and each group
of order $np$ has a unique $p$-Sylow subgroup, then any skew left brace of
size $np$ is a double semidirect product of the trivial brace of size $p$ and
a skew brace of size $n$. We develop an algorithm to obtain all braces of size
$np$ from the set of braces of size $n$ and provide a formula to count them.
We use the result to describe all braces of size $12p$ for $p\geq 7$ and prove
a conjecture of Bardakov, M.V. Neshchadim and M.K. Yadav.
(1) Departament de Matemàtiques i Informàtica, Universitat de Barcelona, Gran
Via de les Corts Catalanes 585, 08007, Barcelona (Spain)
(2) Charles University, Faculty of Mathematics and Physics, Department of
Algebra, Sokolovska 83, 18600 Praha 8, Czech Republic
(3) Departament de Matemàtiques, Universitat Politècnica de Catalunya, Edifici
Omega, Jordi Girona, 1-3, 08034, Barcelona (Spain)
Keywords: Left braces, Holomorphs, regular subgroups, Hopf Galois structures.
2020MSC: Primary: 16T05, 16T25, 12F10; Secondary: 20B35, 20B05, 20D20, 20D45.
††footnotetext: The first author was supported by grant PID2019-107297GB-I00
(Ministerio de Ciencia, Innovación y Universidades). The second author was
supported by Czech Science Foundation, grant 21-00420M, and by Charles
University Research Centre program UNCE/SCI/022. Third and fourth authors are
supported by 2021 SGR 01468. Email addresses<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>
## 1 Introduction
In [7] Rump introduced an algebraic structure called brace to study set-
theoretic solutions of the Yang-Baxter equation. Guarnieri and Vendramin in
[6] gave the more general concept of skew brace. A skew (left) brace is a set
$B$ with two operations $\cdot$ and $\circ$ such that $(B,\cdot)$ and
$(B,\circ)$ are groups and the brace relation is satisfied, namely,
$a\circ(b\cdot c)=a\circ b\cdot a^{-1}\cdot a\circ c,$
for all $a,b,c\in B$, where $a^{-1}$ denotes the inverse of $a$ with respect
to $\cdot$. We call $(B,\cdot$) the additive group and $(B,\circ)$ the
multiplicative group of the skew left brace. We say that $B$ is a classical
brace if its additive group is abelian.
Let $B_{1}$ and $B_{2}$ be skew left braces. A map $f:B_{1}\to B_{2}$ is said
to be a skew brace morphism if $f(a\cdot b)=f(a)\cdot f(b)$ and $f(a\circ
b)=f(a)\circ f(b)$ for all $a,b\in B_{1}$. If $f$ is bijective, we say that
$f$ is an isomorphism. In that case we say that the skew braces $B_{1}$ and
$B_{2}$ are isomorphic.
In [6] Guarnieri and Vendramin proved that given a group $N$, there is a
bijective correspondence between isomorphism classes of skew left braces with
additive group $N$ and conjugacy classes of regular subgroups of
$\mathrm{Hol}(N)$.
In [3], we classified classical left braces of size $8p$, for $p$ a prime
number, $p\neq 3,7$. In [4], we gave a method to determine all braces of size
$np$ from the braces of size $n$ and certain classes of morphisms from the
multiplicative group of these braces of size $n$ to $\mathbf{Z}_{p}^{*}$. For
a prime number $p\geq 7$, we applied the obtained results to classify
classical left braces of size $12p$. In particular we proved the conjectures
in [1] on the numbers $b(8p),b(12p)$ of isomorphism classes of classical left
braces of size $8p$ and $12p$, respectively.
Our aim in this paper is to generalize the method obtained in [4] to skew
nonclassical left braces of size $np$, where $p$ is a prime number and $n$ an
integer not divisible by $p$ and such that every group of order $np$ has a
normal subgroup of order $p$. We shall apply our method to classify skew left
braces of size $12p$, for $p$ a prime number, $p\geq 7$. In particular we will
prove the following conjecture in [1] on the number $s(12p)$ of isomorphism
classes of skew left braces of size $12p$, $p\geq 7$.
$s(12p)=\left\\{\begin{array}[]{ll}324&\text{\ if }p\equiv 11\pmod{12},\\\
410&\text{\ if }p\equiv 5\pmod{12},\\\ 606&\text{\ if }p\equiv 7\pmod{12},\\\
782&\text{\ if }p\equiv 1\pmod{12}.\end{array}\right.$ (1)
We note that $s(24)=855,s(36)=400$ and $s(60)=418$ (see [9]).
## 2 Skew braces of size $np$
We recall the definition of direct and semidirect product of braces as defined
in [2] and [8].
Let $A$ and $B$ be skew left braces. Then $A\times B$ together with
$(a,b)\cdot(a^{\prime},b^{\prime})=(a\cdot a^{\prime},b\cdot
b^{\prime})\quad(a,b)\circ(a^{\prime},b^{\prime})=(a\circ a^{\prime},b\circ
b^{\prime})$
is a skew left brace called the direct product of the braces $A$ and $B$.
Now, let $\tau:(B,\circ)\to\operatorname{Aut}(A,\cdot,\circ)$ be a group
homomorphism. Consider in $A\times B$ the additive structure of the direct
product $(A,\cdot)\times(B,\cdot)$
$(a,b)\cdot(a^{\prime},b^{\prime})=(a\cdot a^{\prime},b\cdot b^{\prime})$
and the multiplicative structure of the semidirect product
$(A,\circ)\rtimes_{\tau}(B,\circ)$
$(a,b)\circ(a^{\prime},b^{\prime})=(a\circ\tau(b)(a^{\prime}),b\circ
b^{\prime})$
Then, we obtain a skew left brace, which is called the semidirect product of
the skew left braces $A$ and $B$ via $\tau$.
We observe in particular that the additive group of the semidirect product of
two left braces is the direct product of the corresponding additive groups. In
[4] we proved that a classical brace of size $np$, where $p$ is a prime number
and $n$ an integer not divisible by $p$ and such that every group of order
$np$ has a normal subgroup of order $p$, is a direct or semidirect product of
the trivial brace of size $p$ and a brace of size $n$. In this case, the
additive group was a direct product of $C_{p}$ and an abelian group of order
$n$. However we are now interested in skew braces of size $np$, for which both
additive and multiplicative groups may be strictly semidirect products of
$C_{p}$ and a group of order $n$. We want then to determine the relationship
between these skew braces of order $np$ and skew braces of order $n$. To this
end we shall define the double semidirect product of two skew braces.
Let us recall first that for a skew left brace $(A,\cdot,\circ)$ and each
$a\in A$, the map
$\begin{array}[]{llll}\lambda_{a}:&A&\rightarrow&A\\\ &b&\mapsto&a^{-1}\cdot
a\circ b\end{array}$
is bijective with inverse
$\begin{array}[]{llll}\lambda_{a}^{-1}:&A&\rightarrow&A\\\
&b&\mapsto&\overline{a}\circ(a\cdot b)\end{array},$
where $\overline{a}$ denotes the inverse of $a$ with respect to $\circ$. It
follows that $a\circ b=a\cdot\lambda_{a}(b),a\cdot
b=a\circ\lambda_{a}^{-1}(b)$.
###### Proposition 1.
Let $A$ and $B$ be skew left braces and let
$\sigma:(B,\cdot)\to\operatorname{Aut}(A,\cdot,\circ),\quad\tau:(B,\circ)\to\operatorname{Aut}(A,\cdot,\circ)$
be morphisms of groups. Consider in $A\times B$ the additive structure of the
semidirect product $(A,\cdot)\rtimes_{\sigma}(B,\cdot)$
$(a,b)\cdot(a^{\prime},b^{\prime})=(a\cdot\sigma(b)(a^{\prime}),b\cdot
b^{\prime})$
and the multiplicative structure of the semidirect product
$(A,\circ)\rtimes_{\tau}(B,\circ)$
$(a,b)\circ(a^{\prime},b^{\prime})=(a\circ\tau(b)(a^{\prime}),b\circ
b^{\prime})$
We assume that the following equality holds for any $a\in A$ and all
$b_{1},b_{2}\in B$
$\sigma(b_{1}\circ b_{2}\cdot
b_{1}^{-1})\lambda_{a}=\lambda_{a}\tau(b_{1})\sigma(b_{2})\tau(\overline{b_{1}}).$
(2)
Then $(A\times B,\cdot,\circ)$ is a skew brace.
###### Proof.
We check the brace condition for $(A\times B,\cdot,\circ)$, i.e., the equality
$(a_{1},b_{1})\circ\left((a_{2},b_{2})\cdot(a_{3},b_{3})\right)=(a_{1},b_{1})\circ(a_{2},b_{2})\cdot(a_{1},b_{1})^{-1}\cdot(a_{1},b_{1})\circ(a_{3},b_{3})$
for $(a_{1},b_{1}),(a_{2},b_{2}),(a_{3},b_{3})\in A\times B$. The term on the
left is
$\begin{array}[]{lll}(a_{1},b_{1})\circ\left((a_{2},b_{2})\cdot(a_{3},b_{3})\right)&=&(a_{1},b_{1})\circ(a_{2}\cdot\sigma(b_{2})(a_{3}),b_{2}\cdot
b_{3})\\\
&=&(a_{1}\circ\tau(b_{1})(a_{2}\cdot\sigma(b_{2})(a_{3})),b_{1}\circ(b_{2}\cdot
b_{3}))\end{array}$
and the term on the right is
$\begin{array}[]{l}(a_{1},b_{1})\circ(a_{2},b_{2})\cdot(a_{1},b_{1})^{-1}\cdot(a_{1},b_{1})\circ(a_{3},b_{3})\\\
=(a_{1}\circ\tau(b_{1})(a_{2}),b_{1}\circ
b_{2})\cdot(\sigma(b_{1}^{-1})(a_{1}^{-1}),b_{1}^{-1})\cdot(a_{1}\circ\tau(b_{1})(a_{3}),b_{1}\circ
b_{3})\\\ =(a_{1}\circ\tau(b_{1})(a_{2})\cdot\sigma(b_{1}\circ
b_{2})((\sigma(b_{1}^{-1})(a_{1}^{-1})),b_{1}\circ b_{2}\cdot
b_{1}^{-1})\cdot(a_{1}\circ\tau(b_{1})(a_{3}),b_{1}\circ b_{3})\\\
=(a_{1}\circ\tau(b_{1})(a_{2})\cdot\sigma(b_{1}\circ b_{2}\cdot
b_{1}^{-1})(a_{1}^{-1})\cdot\sigma(b_{1}\circ b_{2}\cdot
b_{1}^{-1})(a_{1}\circ\tau(b_{1})(a_{3})),b_{1}\circ b_{2}\cdot
b_{1}^{-1}\cdot b_{1}\circ b_{3})\end{array}$
Since $B$ is a brace, the two second components are equal. Now, the first
component of the term on the left is
$\begin{array}[]{lll}a_{1}\circ\tau(b_{1})(a_{2}\cdot\sigma(b_{2})(a_{3}))&=&a_{1}\circ\left(\tau(b_{1})(a_{2})\cdot\tau(b_{1})(\sigma(b_{2})(a_{3})\right)\\\
&=&a_{1}\circ\tau(b_{1})(a_{2})\cdot a_{1}^{-1}\cdot
a_{1}\circ\tau(b_{1})(\sigma(b_{2})(a_{3}))\\\
&=&a_{1}\circ\tau(b_{1})(a_{2})\cdot\lambda_{a_{1}}(\tau(b_{1})(\sigma(b_{2})(a_{3})))\end{array}$
where the second equality follows from the brace condition for $A$. The first
component of the term on the right is
$\begin{array}[]{l}a_{1}\circ\tau(b_{1})(a_{2})\cdot\sigma(b_{1}\circ
b_{2}\cdot b_{1}^{-1})(a_{1}^{-1})\cdot\sigma(b_{1}\circ b_{2}\cdot
b_{1}^{-1})(a_{1}\circ\tau(b_{1})(a_{3}))\\\
=a_{1}\circ\tau(b_{1})(a_{2})\cdot\sigma(b_{1}\circ b_{2}\cdot
b_{1}^{-1})(a_{1}^{-1}\cdot a_{1}\circ\tau(b_{1})(a_{3}))\\\
=a_{1}\circ\tau(b_{1})(a_{2})\cdot\sigma(b_{1}\circ b_{2}\cdot
b_{1}^{-1})(\lambda_{a_{1}}(\tau(b_{1})(a_{3}))).\end{array}$
The equality of the first components is then equivalent to
$\lambda_{a_{1}}(\tau(b_{1})(\sigma(b_{2})(a_{3})))=\sigma(b_{1}\circ
b_{2}\cdot b_{1}^{-1})(\lambda_{a_{1}}(\tau(b_{1})(a_{3})))$
for all $a_{3}\in A$, hence to equality (2). ∎
###### Definition 2.
We call double semidirect product of the skew braces $A$ and $B$ via $\sigma$
and $\tau$ the skew brace $(A\times B,\cdot,\circ)$ in Proposition 1.
###### Remarks 3.
1. 1)
If $\sigma$ is the trivial morphism, then (2) is satisfied and we recover the
notion of semidirect product of skew braces.
2. 2)
If $A$ is a trivial brace and $\operatorname{Aut}A$ is abelian, then (2) is
equivalent to $\sigma(b_{1}\circ b_{2})=\sigma(b_{1})\sigma(b_{2})$ for all
$b_{1},b_{2}\in B$, i.e. to $\sigma$ being a group morphism with respect to
$\circ$.
Turning back to skew braces of size $np$, we establish the following
hypothesis on $n$ and $p$ which will be maintained all along the paper.
###### Hypothesis.
Let $p$ be an odd prime and $n$ an integer such that $p$ does not divide $n$
and each group of order $np$ has a normal subgroup of order $p$.
In particular, we may consider $n=12$ and $p\geq 7$, which satisfy this
hypothesis.
###### Proposition 4.
Any skew brace of size $np$ is a double semidirect product of the trivial
brace of size $p$ and a skew brace of size $n$.
###### Proof.
Let $(B_{np},\cdot,\circ)$ be a left brace of size $np$ with additive group
$N$ and multiplicative group $G$. Then, by the Schur-Zassenhaus theorem,
$N=\mathbf{Z}_{p}\rtimes_{\sigma}E$ with $(E,\cdot)$ a group of order $n$ and
$\sigma:E\to\operatorname{Aut}(\mathbf{Z}_{p})$ a group morphism and
$G=\mathbf{Z}_{p}\rtimes_{\tau}F$ with $(F,\circ)$ a group of order $n$ and
$\tau:F\to\operatorname{Aut}(\mathbf{Z}_{p})$ a group morphism. We note that
$\operatorname{Aut}(\mathbf{Z}_{p})$ is the group of automorphisms of the
trivial brace of size $p$, which is the unique brace of size $p$. We write
$\mathbf{Z}_{p}$ additively and identify $\operatorname{Aut}(\mathbf{Z}_{p})$
with $\mathbf{Z}_{p}^{*}$. As a set, $B_{np}$ is the cartesian product of
$\mathbf{Z}_{p}$ and a set $B_{n}$ with $n$ elements endowed with two
operations $\cdot$ and $\circ$ such that $(B_{n},\cdot)$ and $(B_{n},\circ)$
are groups isomorphic to $(E,\cdot)$ and $(F,\circ)$, respectively.
For $(a,b),(a^{\prime},b^{\prime})\in B_{np}$, we have then
$\begin{array}[]{rcl}(a,b)\cdot(a^{\prime},b^{\prime})&=&(a+\sigma(b)a^{\prime},bb^{\prime})\\\
(a,b)\circ(a^{\prime},b^{\prime})&=&(a+\tau(b)a^{\prime},b\circ
b^{\prime})\end{array}.$
We write the brace relation for $B_{np}$
$(a,b)\circ\left((a^{\prime},b^{\prime})\cdot(a^{\prime\prime},b^{\prime\prime})\right)=(a,b)\circ(a^{\prime},b^{\prime})\cdot(a,b)^{-1}\cdot(a,b)\circ(a^{\prime\prime},b^{\prime\prime}).$
(3)
The lefthand side is
$\begin{array}[]{rcl}(a,b)\circ\left((a^{\prime},b^{\prime})\cdot(a^{\prime\prime},b^{\prime\prime})\right)&=&(a,b)\circ\left(a^{\prime}+\sigma(b^{\prime})a^{\prime\prime},b^{\prime}\cdot
b^{\prime\prime}\right)\\\
&=&\left(a+\tau(b)(a^{\prime}+\sigma(b^{\prime})a^{\prime\prime}),b\circ(b^{\prime}\cdot
b^{\prime\prime})\right)\end{array}$
and the righthand side is
$\begin{array}[]{l}(a,b)\circ(a^{\prime},b^{\prime})\cdot(a,b)^{-1}\cdot(a,b)\circ(a^{\prime\prime},b^{\prime\prime})\\\
=(a+\tau(b)a^{\prime},b\circ
b^{\prime})\cdot(-\sigma(b^{-1})a,b^{-1})\cdot(a+\tau(b)a^{\prime\prime},b\circ
b^{\prime\prime})\\\ =(a+\tau(b)a^{\prime}-\sigma(b\circ
b^{\prime})\sigma(b^{-1})a,(b\circ b^{\prime})\cdot
b^{-1})\cdot(a+\tau(b)a^{\prime\prime},b\circ b^{\prime\prime})\\\
=(a+\tau(b)a^{\prime}-\sigma(b\circ b^{\prime})\sigma(b^{-1})a+\sigma((b\circ
b^{\prime})\cdot b^{-1})(a+\tau(b)a^{\prime\prime}),(b\circ b^{\prime})\cdot
b^{-1}\cdot(b\circ b^{\prime\prime})).\end{array}$
The equality of the second components gives a brace $B_{n}$ of size $n$ with
additive group $E$ and multiplicative group $F$. The first component in the
lefthand side is
$a+\tau(b)(a^{\prime}+\sigma(b^{\prime})a^{\prime\prime})=a+\tau(b)a^{\prime}+\tau(b)\sigma(b^{\prime})a^{\prime\prime}$
and the first component in the righthand side is
$\begin{array}[]{l}a+\tau(b)a^{\prime}-\sigma(b\circ
b^{\prime})\sigma(b^{-1})a+\sigma((b\circ b^{\prime})\cdot
b^{-1})(a+\tau(b)a^{\prime\prime})\\\ =a+\tau(b)a^{\prime}-\sigma((b\circ
b^{\prime})\cdot b^{-1})a+\sigma((b\circ b^{\prime})\cdot
b^{-1})a+\sigma((b\circ b^{\prime})\cdot b^{-1})\tau(b)a^{\prime\prime}\\\
=a+\tau(b)a^{\prime}+\sigma((b\circ b^{\prime})\cdot
b^{-1})\tau(b)a^{\prime\prime}\end{array}$
The equality of the first components is then equivalent to
$\tau(b)\sigma(b^{\prime})a^{\prime\prime}=\sigma((b\circ b^{\prime})\cdot
b^{-1})\tau(b)a^{\prime\prime}$
for all $a^{\prime\prime}\in\mathbf{Z}_{p}$, hence to
$\sigma(b^{\prime})=\sigma((b\circ b^{\prime})\cdot b^{-1})=\sigma(b\circ
b^{\prime})\sigma(b^{-1})$
which is equivalent to
$\sigma(b\circ b^{\prime})=\sigma(b)\sigma(b^{\prime}),$ (4)
for all $b,b^{\prime}\in B_{n}$, hence to $\sigma$ being also a morphism with
respect to $\circ$.
We have then obtained that $B_{np}$ is the double semidirect product of the
trivial brace $\mathbf{Z}_{p}$ and the skew brace $B_{n}$. ∎
We recall now precisely the correspondence between braces and regular
subgroups of the holomorph given by Guarnieri and Vendramin. We note that, if
$G$ is a regular subgroup of $\mathrm{Hol}(A,\cdot)$, then
$\begin{array}[]{lcll}{\pi_{1}}_{|G}:&G&\rightarrow&A\\\
&(a,f)&\mapsto&a\end{array}$
is bijective.
###### Proposition 5 ([6] th. 4.2).
Let $A$ be a skew left brace. Then
$\\{(a,\lambda_{a})\,:\,a\in A\\}$
is a regular subgroup of $\mathrm{Hol}(A,\cdot)$, isomorphic to $(A,\circ)$.
Conversely, if $(A,\cdot)$ is a group and $G$ is a regular subgroup of
$\mathrm{Hol}(A,\cdot)$, then $A$ is a skew left brace with $(A,\circ)\simeq
G$, where
$a\circ b=a\cdot f(b),$
and $({\pi_{1}}_{|G})^{-1}(a)=(a,f)\in G$.
These assignments give a bijective correspondence between the set of
isomorphism classes of skew left braces $(A,\cdot,\circ)$ and the set of
conjugation classes of regular subgroups of $\mathrm{Hol}(A,\cdot)$.
Using this correspondence we can see how braces of size $np$ give braces of
size $n$ and how to obtain braces of size $np$ from the trivial brace of size
of size $p$ and braces of size $n$.
Let $(B_{np},\cdot,\circ)$ be a brace of size $np$ with additive group
$N=\mathbf{Z}_{p}\rtimes_{\sigma}E$, as above, and multiplicative group
$G=\mathbf{Z}_{p}\rtimes_{\tau}F$. By applying Proposition 5, we obtain that
$\\{(x,\lambda_{x})\,:\,x\in N\\}$ is a regular subgroup of $\mathrm{Hol}(N)$,
isomorphic to $G$. Now, for $x=(m,a),y=(n,b)\in N$, we have
$\lambda_{x}(y)=x^{-1}\cdot(x\circ
y)=(-\sigma(a^{-1})m,a^{-1})\cdot(m+\tau(a)n,a\circ
b)=(\sigma(a^{-1})\tau(a)n,\lambda_{a}(b)),$
hence $\pi_{2}\lambda_{x}=\lambda_{a}\pi_{2}$, where
$\pi_{2}:N=\mathbf{Z}_{p}\rtimes_{\sigma}E\rightarrow E$ is the projection on
the second component.
Let us see these regular groups explicitly in $\mathrm{Hol}(N)$. Following
[5], the automorphisms of $N$ can be described as matrices
$\begin{pmatrix}k&\gamma\\\ 0&\lambda\end{pmatrix}$
with $k\in\mathbf{Z}_{p}^{*}=\operatorname{Aut}(\mathbf{Z}_{p})$,
$\lambda\in\operatorname{Aut}(E)$, such that $\sigma\lambda=\sigma$ and
$\gamma:E\to\mathbf{Z}_{p}$ a $1$-cocycle. The action on $N$ is given by
$\begin{pmatrix}k&\gamma\\\ 0&\lambda\end{pmatrix}\begin{pmatrix}m\\\
a\end{pmatrix}=\begin{pmatrix}km+\gamma(a)\\\ \lambda(a)\end{pmatrix}.$
Under our hypothesis, Schur Zassenhaus condition on subgroups of order $|E|$
being conjugates of $E$ by elements in $\mathbf{Z}_{p}$ leaves us with the
coboundaries
$\gamma_{i}(a)=i-\sigma(a)i,\quad i\in\mathbf{Z}_{p}.$
For a trivial $\sigma$, we get the direct product
$\operatorname{Aut}(\mathbf{Z}_{p})\times\operatorname{Aut}(E)$ and for a non
trivial $\sigma$ the group $\operatorname{Aut}(N)$ has order $p(p-1)s$ with
$s=\\#\\{\lambda\in\operatorname{Aut}(E):\sigma\lambda=\sigma\\}$. This
subgroup is the stabilizer $\Sigma_{\sigma}$ of $\sigma$ under the action
$\begin{array}[]{ccl}\operatorname{Aut}(E)\times\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\longrightarrow&\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})\\\
(g,\sigma)&\longrightarrow&\sigma g\end{array}.$
The orbit of $\sigma$ gives the $\sigma^{\prime}$ such that
$\mathbf{Z}_{p}\rtimes_{\sigma^{\prime}}E\simeq N$.
All together, if we write
$\operatorname{Aut}(\mathbf{Z}_{p}\rtimes_{\sigma}E)=\left\\{\begin{pmatrix}k&\gamma_{i}\\\
0&\lambda\end{pmatrix}\qquad k\in\mathbf{Z}_{p}^{*},\ i\in\mathbf{Z}_{p},\
\lambda\in\Sigma_{\sigma}\right\\}$
then
$\mathrm{Hol}(\mathbf{Z}_{p}\rtimes_{\sigma}E)=\left\\{[\begin{pmatrix}m\\\
a\end{pmatrix},M]\qquad m\in\mathbf{Z}_{p},\ a\in
E,M\in\operatorname{Aut}(\mathbf{Z}_{p}\rtimes_{\sigma}E)\right\\}$
with
$[u,M_{u}][v,M_{v}]=[u\rtimes M_{u}v,M_{u}M_{v}]$
where $\rtimes$ indicates operation in $\mathbf{Z}_{p}\rtimes_{\sigma}E$. For
a brace $B_{np}$ as before, this is the $\cdot$ operation. The set
$\\{(0,a)\mid a\in E\\}$ is stable under this operation. For the
multiplicative group $G=\\{(x,\lambda_{x})\mid\ x\in N\\}$, if $x=(m,a)$ we
have
$\lambda_{x}=\begin{pmatrix}\sigma(a)^{-1}\tau(a)&0\\\
0&\lambda_{a}\end{pmatrix}$
The set $\\{(a,\lambda_{a})\mid a\in E\\}$ is a group isomorphic to $F$, and a
regular subgroup of $\mathrm{Hol}(E)$. This describes the subbbrace $B_{n}$ of
$B_{np}$ with additive structure isomorphic to $E$ and multiplicative
structure isomorphic to $F$. Notice that in the description of automorphisms
of $N$ we got the condition $\sigma\in\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})$,
namely $\sigma\lambda_{a}=\sigma$ for all $a\in E$.
From the above, we have also obtained a way to construct a brace of size $np$
from a brace of size $n$. If $E$ is a group of order $n$,
$\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ and $F$ is a regular subgroup of
$\mathrm{Hol}(E)$ such that $\pi_{2}(F)$ is inside the stabilizer of $\sigma$
by the action of $\operatorname{Aut}(E)$, then for any
$\tau\in\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$
$G(\sigma,\tau)=\left\\{[\begin{pmatrix}m\\\
a\end{pmatrix},\begin{pmatrix}\sigma(a)^{-1}\tau(a,\lambda_{a})&0\\\
0&\lambda_{a}\end{pmatrix}]\quad m\in\mathbf{Z}_{p},a\in E\right\\}.$
is a regular subgroup of $\mathrm{Hol}(\mathbf{Z}_{p}\rtimes_{\sigma}E)$
isomorphic to $\mathbf{Z}_{p}\rtimes_{\tau}F.$
But we can also consider the opposite group operation on $N$, so that elements
in $N$ are $(\sigma(a)m,a)$, with $m\in\mathbf{Z}_{p}$ and $a\in E$. Then,
$\begin{array}[]{ccl}N&\longrightarrow&\mathrm{Hol}(N)\\\
(m,a)&\to&[\begin{pmatrix}\sigma(a)m\\\
a\end{pmatrix},\begin{pmatrix}\sigma(a)&\gamma_{-m}\\\
0&\lambda_{a}\end{pmatrix}]\end{array}$
is a regular embedding which restricted to $E$ provides an embedding of $F$:
$[\begin{pmatrix}\sigma(a)m\\\
a\end{pmatrix},\begin{pmatrix}\sigma(a)&\gamma_{-m}\\\
0&\lambda_{a}\end{pmatrix}][\begin{pmatrix}\sigma(b)n\\\
b\end{pmatrix},\begin{pmatrix}\sigma(b)&\gamma_{-n}\\\
0&\lambda_{b}\end{pmatrix}]=$ $=[\begin{pmatrix}\sigma(ab)(m+\sigma(a)n)\\\
a\lambda_{a}(b)\end{pmatrix},\begin{pmatrix}\sigma(ab)&\gamma_{-(m+\sigma(a)n)}\\\
0&\lambda_{a}\lambda_{b}\end{pmatrix}].$
Using again [6] we get the regular subgroup
$G(\sigma,\tau)^{\prime}=\left\\{[\begin{pmatrix}\sigma(a)m\\\
a\end{pmatrix},\begin{pmatrix}\tau(a,\lambda_{a})&\gamma_{-m}\\\
0&\lambda_{a}\end{pmatrix}]\quad m\in\mathbf{Z}_{p},a\in E\right\\}$
isomorphic to $\mathbf{Z}_{p}\rtimes_{\tau}F$.
###### Proposition 6.
Let $p$ be a prime and $n$ a positive integer not divisible by $p$ and such
that every group of order $np$ has a normal subgroup of order $p$. Let $B_{n}$
a brace of order $n$ with additive structure $(E,\cdot)$ and multiplicative
structure $(F,\circ)$. For every
$\sigma\in\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})$ and every
$\tau\in\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$,
$G(\sigma,\tau)=\left\\{[\begin{pmatrix}m\\\
a\end{pmatrix},\begin{pmatrix}\sigma(a)^{-1}\tau(a,\lambda_{a})&0\\\ 0&\
\lambda_{a}\end{pmatrix}]\quad m\in\mathbf{Z}_{p},(a,\lambda_{a})\in
F\right\\}$ $G(\sigma,\tau)^{\prime}=\left\\{[\begin{pmatrix}\sigma(a)m\\\
a\end{pmatrix},\begin{pmatrix}\tau(a,\lambda)&\gamma_{-m}\\\
0&\lambda_{a}\end{pmatrix}]\quad m\in\mathbf{Z}_{p},(a,\lambda_{a})\in
F\right\\}$
are regular subgroups of $\mathrm{Hol}(\mathbf{Z}_{p}\rtimes_{\sigma}E)$
isomorphic to $\mathbf{Z}_{p}\rtimes_{\tau}F$.
We have $G(1,\tau)=G(1,\tau)^{\prime}$ for any $\tau$. For a nontrivial
$\sigma$, $G(\sigma,\tau)$ and $G(\sigma,\tau)^{\prime}$ are not conjugate in
$\mathrm{Hol}(N)$.
Any regular subgroup of $\mathrm{Hol}(\mathbf{Z}_{p}\rtimes_{\sigma}N)$
isomorphic to $\mathbf{Z}_{p}\rtimes_{\tau}F$ is conjugate to $G(\sigma,\tau)$
or $G(\sigma,\tau)^{\prime}$ by an element of $\operatorname{Aut}(N)$.
When $\sigma=1$, the coboundaries are also trivial, and both subgroups are
equal. For a nontrivial $\sigma$, since $\mathbf{Z}_{p}^{*}$ is abelian, the
first coefficient of the matrix is invariant under conjugation, and this shows
that the groups are not conjugate in $\mathrm{Hol}(N)$ for any $\tau$.
Let us prove the last statement. If $G\subseteq\mathrm{Hol}(N)$ is a regular
subgroup isomorphic to $\mathbf{Z}_{p}\rtimes_{\tau}F$, then via
$\mathbf{Z}_{p}\rtimes_{\tau}F\simeq G\hookrightarrow\mathrm{Hol}(N)$, using
[6] and $\mathrm{Hom}(\mathbf{Z}_{p},\mathbf{Z}_{p}^{*})=1$, we get a
semiregular embedding of $\mathbf{Z}_{p}$ in $\mathrm{Hol}(N)$,
$\ \begin{array}[]{rcl}\quad\mathbf{Z}_{p}&\longrightarrow&\mathrm{Hol}(N)\\\
m&\to&[\begin{pmatrix}f(m)\\\ 1\end{pmatrix},\begin{pmatrix}1&\gamma_{h(m)}\\\
0&1\end{pmatrix}]\\\ \end{array}.$
so that $f\in\operatorname{Aut}(\mathbf{Z}_{p})$ and
$h\in\mathrm{Hom}(\mathbf{Z}_{p},\mathbf{Z}_{p})$, and the embedding is
$\
\begin{array}[]{rcl}\nu_{k,l}:\quad\mathbf{Z}_{p}&\longrightarrow&\mathrm{Hol}(N)\\\
m&\to&[\begin{pmatrix}km\\\ 1\end{pmatrix},\begin{pmatrix}1&\gamma_{lm}\\\
0&1\end{pmatrix}]\\\ \end{array}$
for some $k\in\mathbf{Z}_{p}^{*},\ l\in\mathbf{Z}_{p}$. It is easy to check
that conjugation by elements in $\operatorname{Aut}(N)$ splits this family
into two classes represented by $\nu_{1,0}$ and $\nu_{1,l}$ for any
$l\in\mathbf{Z}_{p}^{*}$, and that conjugation is given by an element in
$\operatorname{Aut}(\mathbf{Z}_{p})$.
On the other hand, we get a semiregular embedding of $F$ in $\mathrm{Hol}(N)$.
By [6] and using $\mathrm{Hom}(E,\mathbf{Z}_{p})=1$, it has to be
$\begin{array}[]{rcl}\Psi_{\alpha}:\quad F&\longrightarrow&\mathrm{Hol}(N)\\\
(a,\lambda_{a})&\to&[\begin{pmatrix}0\\\
a\end{pmatrix},\begin{pmatrix}\alpha(a)&0\\\ 0&\lambda_{a}\end{pmatrix}]\\\
\end{array}$
with $\alpha\in\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})$. The conjugacy class by
elements of $\operatorname{Aut}(N)$ of the image of $\Psi_{\alpha}$ is
$[\begin{pmatrix}\gamma_{i_{M}}(a)\\\
\lambda_{M}(a)\end{pmatrix},\begin{pmatrix}\alpha(a)&\gamma_{i_{M}-\alpha(a)i_{M}}\\\
0&\lambda_{M}\lambda_{a}\lambda_{M}^{-1}\end{pmatrix}]$
so that $\Psi_{\alpha}(F)$ and $\Psi_{\alpha^{\prime}}(F)$ are not conjugate
when $\alpha\neq\alpha^{\prime}$. Notice that
$\\{(\lambda_{M}(x),\lambda_{M}\lambda_{a}\lambda_{M}^{-1})\mid(a,\lambda_{a})\in
F\\}=\Phi_{\lambda_{M}}(F).$ It is equal to $F$ if and only if $\lambda_{M}$
is a brace automorphism. Notice also that the image is invariant under
conjugation by $\operatorname{Aut}(\mathbf{Z}_{p})$. This gives the following
result.
###### Proposition 7.
Let $F,F^{\prime}$ be regular subgroups of $\mathrm{Hol}(E)$ giving
multiplicative structures for a brace of order $n$. If $\Psi_{\alpha}(F)$ and
$\Psi_{\alpha}(F^{\prime})$ are conjugate in $\mathrm{Hol}(N)$ then they are
conjugate in $\mathrm{Hol}(E)$.
Back to the proof of the previous proposition, up to conjugation by elements
in $\operatorname{Aut}(\mathbf{Z}_{p})$, the group $G$ has to be the
semidirect product of $\Psi_{\alpha}(F)$ and either
$\nu_{1,0}(\mathbf{Z}_{p})$ or $\nu_{1,l}(\mathbf{Z}_{p})$. An easy
computation shows that the only possibilities are those corresponding to
$G(\sigma,\tau)$ and $G(\sigma,\tau)^{\prime}$, namely $\nu_{1,0}$ and
$\alpha(a)=\tau(a,\lambda_{a})\sigma(a)^{-1}$ or $\nu_{1,-1}$ and
$\alpha(a)=\tau(a,\lambda_{a})$. Indeed,
$[\begin{pmatrix}m\\\ a\end{pmatrix},\begin{pmatrix}\alpha(a)&0\\\
0&\lambda_{a}\end{pmatrix}][\begin{pmatrix}n\\\
b\end{pmatrix},\begin{pmatrix}\alpha(b)&0\\\
0&\lambda_{b}\end{pmatrix}]=[\begin{pmatrix}m+\sigma(a)\alpha(a)n\\\
a\lambda_{a}(b)\end{pmatrix}\begin{pmatrix}\alpha(ab)&0\\\
0&\lambda_{a}\lambda_{b}\end{pmatrix}]$
gives the image of the product
$(m+\tau(a,\lambda_{a})n,(a\lambda_{a}(b),\lambda_{a}\lambda_{b}))$ if and
only if $\alpha(a)=\tau(a,\lambda_{a})\sigma(a)^{-1}$, and
$[\begin{pmatrix}\sigma(a)m\\\
a\end{pmatrix},\begin{pmatrix}\alpha(a)&\gamma_{lm}\\\
0&\lambda_{a}\end{pmatrix}][\begin{pmatrix}\sigma(b)n\\\
b\end{pmatrix},\begin{pmatrix}\alpha(b)&\gamma_{ln}\\\
0&\lambda_{b}\end{pmatrix}]=$
$=[\begin{pmatrix}\sigma(a)(m+lm)+\sigma(ab)(-lm+\alpha(a)n)\\\
a\lambda_{a}(b)\end{pmatrix}\begin{pmatrix}\alpha(ab)&\gamma_{l(m+\alpha(a)n)}\\\
0&\lambda_{a}\lambda_{b}\end{pmatrix}]$
gives the image of the product if and only if $l=-1$ and
$\alpha(a)=\tau(a,\lambda_{a})$.
In order to count different brace structures, we have to determine conjugation
orbits in the families $\\{G(\sigma,\tau)\\}_{\sigma,\tau},$ and
$\\{G(\sigma,\tau)^{\prime}\\}_{\sigma,\tau}$ or equivalently, isomorphism
classes of braces structures. Therefore, we should take $\sigma$ up to brace
automorphisms of $(B_{n},\cdot,\circ)$ and $\tau$ up to brace automorphisms of
$(B_{n},\circ)$.
### 2.1 Algorithm
Our aim is to determine all skew braces of size $np$ from skew braces of size
$n$. Therefore we need a precomputation
Step 0
Determine isomorphism classes of groups $E$ of order $n$ and the number of
braces of size $n$ of each type $(E,F)$, identifying $F$ with a representative
for a conjugacy class of regular subgroups of $\mathrm{Hol}(E)$.
For every $E$ compute the stabilisers $\Sigma_{\sigma}$ for the action
$\begin{array}[]{ccc}\operatorname{Aut}(E)\times\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\longrightarrow&\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})\\\
(g,\sigma)&\longrightarrow&\sigma g\end{array}$
namely, $\Sigma_{\sigma}=\\{g\in\operatorname{Aut}(E):\ \sigma g=\sigma\\}$.
Next, we consider as an input a pair $(E,F)$, namely a brace $B_{n}$. The
steps to count all the braces of size $np$ having additive group isomorphic to
$\mathbf{Z}_{p}\rtimes E$ and multiplicative group isomorphic to
$\mathbf{Z}_{p}\rtimes F$ are the following:
Step 1
Determine
$\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})=\\{\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*}):\
\pi_{2}(F)\subseteq\Sigma_{\sigma}\ \\}$
Step 2
Determine $\operatorname{Aut}(B_{n})=\\{g\in\operatorname{Aut}(E):\
\Phi_{g}(F)=F\ \\}$
Step 3
Compute the number of orbits of the action
$\begin{array}[]{ccc}\operatorname{Aut}(B_{n})\times\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})&\longrightarrow&\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})\\\
(g,\sigma)&\longrightarrow&\sigma g\end{array}$
Notice that the orbit of $\sigma=1$, which corresponds to the direct product,
has a single element. The remaining orbits will give rise to two different
braces.
The number of additive structures for $F$ is the number of orbits. Since
orbits preserve order and isomorphism class of kernel, we can keep track of
the number of structures according to these parameters.
Step 4
For each $\sigma$ in a system of representatives of the above orbits, compute
orbits of the action
$\begin{array}[]{ccc}(\operatorname{Aut}(B_{n})\cap\Sigma_{\sigma})\times\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})&\longrightarrow&\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})\\\
(g,\tau)&\longrightarrow&\tau\Phi_{g}\end{array}$
Notice that $\operatorname{Aut}(B_{n})\cap\Sigma_{\sigma}$ is a subgroup of
the normaliser of $\pi_{2}(F)$ in $\Sigma_{\sigma}$.
The number of multiplicative structures for $F$ is twice the number of orbits
except for $\sigma$ trivial, when we get a single one. Since orbits preserve
order and isomorphism class of kernel, we can keep track of the number of
structures according to these parameters.
### 2.2 Number of braces of size $np$
Taking into account the description given in terms of orbits, using the
Burnside lemma we can give formulas for the number of different braces of size
$np$ in terms of the number of braces of size $n$.
###### Theorem 8.
Let $p$ be a prime and $n$ a positive integer not divisible by $p$ and such
that every group of order $np$ has a normal subgroup of order $p$. For braces
$B_{np}$ of size $np$ the number of additive structures is
${\sum}_{(B_{n},\cdot)}\,\dfrac{1}{|\operatorname{Aut}(B_{n})|}\sum_{\sigma\in\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})}|\mathrm{Stab}_{\operatorname{Aut}(B_{n})}(\sigma)|$
where $B_{n}$ runs over the braces of size $n$. The total amount of braces of
size $np$ is
${\sum}_{(B_{n},\cdot)}\,\left(\dfrac{1}{|A_{1}|}\sum_{\tau\in
H}|\mathrm{Stab}_{A_{1}}(\tau))|+2\sum_{\sigma\neq
1}\dfrac{1}{|A_{\sigma}|}\sum_{\tau\in
H}|\mathrm{Stab}_{A_{\sigma}}(\tau)|\right)$
where $H=\mathrm{Hom}((B_{n},\circ),\mathbf{Z}_{p}^{*})$,
$\sigma\in\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})$ runs over a system of
representatives for additive structures and
$A_{\sigma}=\operatorname{Aut}(B_{n})\cap\mathrm{Stab}_{\operatorname{Aut}(B_{n},\cdot)}(\sigma)$.
The case $\sigma=1$, where $A_{1}=\operatorname{Aut}(B_{n})$, corresponds to
the semidirect product of $B_{n}$ and the trivial brace of order $p$. The
remaining cases are double semidirect products of these two braces.
Equivalently, the number of additive structures can be expressed as
${\sum}_{(B_{n},\cdot)}\,\dfrac{1}{|\operatorname{Aut}(B_{n})|}\sum_{g\in\operatorname{Aut}(B_{n})}|\operatorname{Fix}_{\mathrm{Hom}(B_{n},\mathbf{Z}_{p}^{*})}(g)|$
and the total amount as
${\sum}_{(B_{n},\cdot)}\,\left(\dfrac{1}{|A_{1}|}\sum_{g\in
A_{1}}|\operatorname{Fix}_{H}(g)|+2\sum_{\sigma\neq
1}\dfrac{1}{|A_{\sigma}|}\sum_{g\in
A_{\sigma}}|\operatorname{Fix}_{H}(g)|\right),$
where $\operatorname{Fix}$ denotes the number of fixed points.
## 3 Groups of order 12 and skew braces of size 12
There are five groups of order 12, up to isomorphism, two abelian ones
$C_{12}$ and $C_{6}\times C_{2}$ and three non-abelian ones, the alternating
group $A_{4}$, the dihedral group $D_{2\cdot 6}$ and the dicyclic group
$\operatorname{Dic}_{12}$. By computation with Magma, we obtain that the
number of skew left braces with additive group $E$ and multiplicative group
$F$ is as shown in the following table.
$E\backslash F$ | $C_{12}$ | $C_{6}\times C_{2}$ | $A_{4}$ | $D_{2\cdot 6}$ | $\operatorname{Dic}_{12}$
---|---|---|---|---|---
$C_{12}$ | 1 | 1 | 0 | 2 | 1
$C_{6}\times C_{2}$ | 1 | 1 | 1 | 1 | 1
$A_{4}$ | 0 | 2 | 4 | 0 | 2
$D_{2\cdot 6}$ | 2 | 2 | 0 | 4 | 2
$\operatorname{Dic}_{12}$ | 2 | 2 | 0 | 4 | 2
We determine now for each group $E$ of order 12, its automorphism group
$\operatorname{Aut}E$, the group $\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ and
stabilizers of the corresponding group action.
Since $\mathbf{Z}_{p}^{*}=\langle\zeta\rangle$ is a cyclic group of order
$p-1$, for each $d\mid p-1$ it has a unique subgroup
$\mu_{d}=\langle\zeta^{\frac{p-1}{d}}\rangle$ of order $d$. We write
$\zeta_{d}$ for $\zeta^{\frac{p-1}{d}}$. In order to have elements in
$\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ with image $\mu_{d}$ it is necessary that
$d\mid\gcd(\exp(E),p-1)$, where $\exp(E)$ is the exponent of the group $E$.
### 3.1 $E=C_{12}$
Let $E=\langle c\rangle\simeq C_{12}$. Then, $\operatorname{Aut}E=\langle
g_{5}\rangle\times\langle
g_{7}\rangle\simeq\mathbf{Z}_{12}^{*}\simeq\mathbf{Z}_{2}\times\mathbf{Z}_{2}$
with $g_{i}(c)=c^{i}$. If $D=\gcd(12,p-1)$, then
$\begin{array}[]{ccc}\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\xrightarrow{\sim}&\mu_{D}=\langle\zeta_{D}\rangle\\\
\sigma&\to&\sigma(c)\end{array}.$
If $|\operatorname{Ker}\sigma|=k$, then the order of $\sigma$ in
$\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ is $d=\frac{12}{k}$. The Euler totient
function gives the number $\varphi(d)$ of $\sigma$ having order $d$. Such a
sigma can be identified as a pair $(d,j)$ meaning $\sigma(c)=\zeta_{d}^{j}$
for $j$ relatively prime to $d$. Elements of the same order are in the same
orbit under the action of $\operatorname{Aut}(E)$
$(d,j)g_{i}=(d,j^{\prime})\mbox{ for }i=j^{\prime}j^{-1}\bmod d$
and the size of the kernel parametrizes the orbits. As for stabilisers, since
$\sigma
g_{i}=\sigma\iff\sigma(g_{i}(c))=\sigma(c)\iff\sigma(c)^{i}=\sigma(c)\iff
i\equiv 1\bmod d,$
we have $\Sigma_{\sigma}=\\{g_{i}\in\operatorname{Aut}(E):i\equiv 1\bmod
d\\},$ which identifies as subgroup of $\mathbf{Z}_{12}^{*}=\\{1,5,7,11\\}$.
In the following table we compile all the information:
$k$ | $d$ | $\varphi(d)$ | Orbit of $\sigma=(d,j\bmod d)$ | $\Sigma_{\sigma}$
---|---|---|---|---
12 | 1 | 1 | 1 | $\operatorname{Aut}(E)$
6 | 2 | 1 | 1 | $\operatorname{Aut}(E)$
4 | 3 | 2 | $1\xrightarrow{5}2$ | $\\{1,7\\}$
3 | 4 | 2 | $1\xrightarrow{7}3$ | $\\{1,5\\}$
2 | 6 | 2 | $1\xrightarrow{5}5$ | $\\{1,7\\}$
1 | 12 | 4 | $1\xrightarrow{5}5\xrightarrow{11}7\xrightarrow{5}11$ | $\\{1\\}$
The number of isomorphism classes of semidirect products
$\mathbf{Z}_{p}\rtimes C_{12}$ is equal to the number of divisors of
$\gcd(12,p-1)$, namely $\sigma_{0}(D)$, and can be labelled by the size $k$ of
$\operatorname{Ker}\sigma$ (size 12 means direct product):
$p\bmod 12$ | $D$ | $\sigma_{0}(D)$ | $k$
---|---|---|---
1 | 12 | 6 | 12,6,4,3,2,1
5 | 4 | 3 | 12,6,3
7 | 6 | 4 | 12,6,4,2
11 | 2 | 2 | 12,6
### 3.2 $E=C_{6}\times C_{2}$
Let us write $E=\langle a\rangle\times\langle b\rangle$, with $a$ of order 6,
$b$ of order 2. The automorphism group is $\operatorname{Aut}(C_{6}\times
C_{2})\simeq D_{2\cdot 6}=\langle g_{1},\ g_{2}\rangle$, where
$\begin{array}[]{llll}g_{1}:&a&\mapsto&ab\\\
&b&\mapsto&b\end{array},\quad\begin{array}[]{llll}g_{2}:&a&\mapsto&a^{2}b\\\
&b&\mapsto&a^{3}b\end{array}$
$g_{1}$ has order 2, $g_{2}$ has order 6 and $g_{1}g_{2}g_{1}=g_{2}^{-1}$.
Since the exponent of $E$ is $6$, elements
$\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ have order dividing $6$. If
$D=\gcd(6,p-1)$, we have
$\begin{array}[]{rcccl}\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\xrightarrow{\sim}&\mu_{D}\times\mu_{2}=\langle\zeta_{D}\rangle\times\langle-1\rangle&\xrightarrow{\sim}&\mathbf{Z}_{D}\times\mathbf{Z}_{2}\\\
\end{array}$
For each $d\mid D$, if we write $(j\bmod d,i\bmod 2)$ for
$\sigma(a)=\zeta_{d}^{j},\ \sigma(b)=(-1)^{i},$
we get the following orbits for the action of $\operatorname{Aut}(E)$
$k$ | $d$ | Orbit of $\sigma=(j\bmod d,i\bmod 2)$
---|---|---
12 | 1 | $(1,0)$
6 | 2 | $(1,0)\xrightarrow{g_{2}}(0,1)\xrightarrow{g_{2}}(1,1)$
4 | 3 | $(1,0)\xrightarrow{g_{2}}(2,0)$
2 | 6 | $(1,0)\xrightarrow{g_{2}}(2,1)\xrightarrow{g_{2}}(1,1)\xrightarrow{g_{2}}(5,0)\xrightarrow{g_{2}}(4,1)\xrightarrow{g_{2}}(5,1)$
and again the order determines the orbit. Therefore, the size of kernel
parameterise isomorphism classes of semidirect products.
$p\bmod 12$ | $D$ | $k$
---|---|---
1,7 | 6 | 2,4,6,12
5,11 | 2 | 6,12
Finally, let us compute stabilizers for the first element in each orbit and
take the corresponding conjugates. If $\sigma(b)=1$, then
$\sigma(g_{1}(a))=\sigma(ab)=\sigma(a),$ $\sigma(g_{1}(b)=\sigma(b)$ and we
have $g_{1}\in\mathrm{Stab}_{\sigma}$. Since $\langle g_{1}\rangle$ has index
6, this is the stabilizer for $d=6$. For the other values, the (cyclic) orbits
above show the power of $g_{2}$ that we have to take: $g_{2}^{3}$ for $d=2$
and $g_{2}^{2}$ for $d=3$.
$k$ | $d$ | $\Sigma_{\sigma}$ | | |
---|---|---|---|---|---
12 | 1 | $\operatorname{Aut}(E)$ | | |
6 | 2 | $g_{2}^{m}\langle g_{2}^{3},g_{1}\rangle g_{2}^{-m}$ | $=\langle g_{2}^{3},\ g_{2}^{2m}g_{1}\rangle$ | $m=0,1,2$ | $C_{2}\times C_{2}$
4 | 3 | $g_{2}^{m}\langle g_{2}^{2},g_{1}\rangle g_{2}^{-m}$ | $=\langle g_{2}^{2},\ g_{2}^{2m}g_{1}\rangle$ | $m=0,1$ | $S_{3}$
2 | 6 | $g_{2}^{m}\langle g_{1}\rangle g_{2}^{-m}$ | $=\langle\ g_{2}^{2m}g_{1}\rangle$ | $m=0,1,2$ | $C_{2}$
### 3.3 $E=A_{4}$
Let $E$ be the alternating group $A_{4}$, and take generators $(1,2,3)$ and
$(1,2)(3,4)$. We have $\operatorname{Aut}A_{4}\simeq S_{4}$, the isomorphism
given by restriction to $A_{4}$ of inner automorphisms of $S_{4}$. Therefore
the action is
$\phi_{s}(1,2,3)=(s(1),s(2),s(3)),\quad\phi_{s}(1,2)(3,4)=(s(1),s(2)),(s(3),s(4))$
for any $s\in S_{4}$.
Since the exponent of $E$ is $3$, nontrivial elements
$\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ have order $3$, which forces
them to be trivial on $(1,2)(3,4)$ and every element of order 2, namely,
$\operatorname{Ker}\sigma\simeq V_{4}$. For $p\not\equiv 1\bmod 3$,
$\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})=1$ and for $p\equiv 1\bmod 3$
$\begin{array}[]{cclc}\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\xrightarrow{\sim}&\mu_{3}=\langle\zeta_{3}\rangle\simeq\mathbf{Z}_{3}\\\
\sigma_{j}&\to&\sigma_{j}(1,2,3)=\zeta_{3}^{j}\simeq j\end{array}$
For any $s\in S_{4}$ such that $(1,2,3)(s(1),s(2),s(3))$ has order 1 or 2, for
example $s=(1,2)$, we have $\sigma_{1}=\sigma_{2}\phi_{s}$ so that they are in
the same orbit for the action of $\operatorname{Aut}(E)$. The stabiliser is a
normal subgroup of $S_{4}$ of index $2$, namely
$\Sigma_{\sigma_{1}}=\Sigma_{\sigma_{2}}=A_{4}$.
$k$ | $d$ | Orbit of $\sigma=j\bmod 3$ | Stabiliser $\Sigma_{\sigma}$
---|---|---|---
12 | 1 | $0$ | $S_{4}$
4 | 3 | $1\xrightarrow{\phi_{(1,2)}}2$ | $A_{4}$
### 3.4 $E=D_{2\cdot 6}$
Let us write $E=\langle r,s\mid r^{6}=\operatorname{Id},\
s^{2}=\operatorname{Id},\ srs=r^{5}\rangle$. We note that
$\operatorname{Aut}E$ is generated by
$\begin{array}[]{llll}g_{1}:&r&\mapsto&r^{-1}\\\
&s&\mapsto&s\end{array},\quad\begin{array}[]{llll}g_{2}:&r&\mapsto&r\\\
&s&\mapsto&rs\end{array}$
$g_{1}$ has order 2, $g_{2}$ has order 6 and $g_{1}g_{2}g_{1}=g_{2}^{-1}$,
hence $\operatorname{Aut}D_{2\cdot 6}\simeq D_{2\cdot 6}$.
Since $E$ has exponent $6$ and the unique normal subgroups with cyclic kernel
are those of índex 2, we have that for every $p$,
$\begin{array}[]{ccl}\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\xrightarrow{\sim}&\mu_{2}\times\mu_{2}=\\{\pm
1\\}\times\\{\pm 1\\}\\\ \sigma&\to&(\sigma(r),\sigma(s))\end{array}$
$(1,-1)$ has cyclic kernel $\langle r\rangle$ while $(-1,-1)$ and $(-1,1)$
have kernel isomorphic to dihedral group $D_{2\cdot 3}$. Therefore, they give
at have at least two orbits under the action of $\operatorname{Aut}(E)$. Since
$(-1,-1)g_{2}(r)=(-1,-1)(r)=-1$ and $(-1,-1)g_{2}(s)=(-1,-1)(rs)=1$, namely
$(-1,-1)g_{2}=(-1,1)$, we have three orbits for the action, which means three
isomorphism classes of semidirect (or direct) products $\mathbf{Z}_{p}\rtimes
E$, parameterised by the isomorphism class of the kernel. For $\sigma=(1,-1)$,
we denote $\mathbb{Z}_{p}\rtimes_{\sigma}E=\mathbb{Z}_{p}\rtimes_{6}^{c}E$
(cyclic kernel), while for $\sigma=(-1,-1)$ we write
$\mathbb{Z}_{p}\rtimes_{\sigma}E=\mathbb{Z}_{p}\rtimes_{6}^{d}E$ (dihedral
kernel).
As for stabilisers, the size is determined by the length of the orbit and we
have
$(-1,-1)g_{1}=(-1,-1),\ (-1,-1)g_{2}^{2}=(-1,1)g_{2}=(-1,-1).$
Therefore, the action of $\operatorname{Aut}(E)$ can be described as follows:
$k$ | $d$ | Orbit of $\sigma$ | Stabiliser $\Sigma_{\sigma}$
---|---|---|---
12 | 1 | $(1,1)$ | $\operatorname{Aut}(E)$
6 | 2 | $(1,-1)$ | $\operatorname{Aut}(E)$
6 | 2 | $(-1,-1)\xrightarrow{g_{2}}(-1,1)$ | $\langle g_{1},g_{2}^{2}\rangle\simeq D_{2\cdot 3}$
### 3.5 $E=\operatorname{Dic}_{12}$
Let us write $E=\langle x,y\mid x^{3}=1,y^{4}=1,yxy^{-1}=x^{2}\rangle$. The
group $\operatorname{Aut}E$ is generated by
$\begin{array}[]{llll}g_{1}:&x&\mapsto&x^{2}\\\
&y&\mapsto&y^{3}\end{array},\quad\begin{array}[]{llll}g_{2}:&x&\mapsto&x\\\
&y&\mapsto&xy^{3}\end{array}$
$g_{1}$ has order 2, $g_{2}$ has order 6 and $g_{1}g_{2}g_{1}=g_{2}^{-1}$,
hence $\operatorname{Aut}\operatorname{Dic}_{12}\simeq D_{2\cdot 6}$.
The exponent of the dicyclic group is $6$ (the element $xy$ has order $6$) and
the normal subgroups with cyclic quotient are $\langle x^{2}\rangle$ and the
derived subgroup $\langle x\rangle$. Therefore, $\sigma(x)=1$ for all
$\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ and the image of $y$ determines
$\sigma$. If $D=\gcd(4,p-1)$, we have
$\begin{array}[]{ccl}\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})&\xrightarrow{\sim}&\mu_{D}\\\
\sigma&\to&\sigma(y)=\zeta_{d}^{j}\end{array}$
Again, we can identify the elements by a pair $(d,j)$. For $d=2$ we have a
single element, and for $d=4$, since $g_{1}(y)=y^{-1}$, we have $(4,1)$ and
$(4,3)$ in the same orbit. In this case, the stabiliser is a normal subgroup
of $\operatorname{Aut}(E)$ of index $2$. We have
$(4,1)g_{1}g_{2}=(4,3)g_{2}=(4,1),\quad(4,1)g_{2}^{2}=(4,3)g_{2}=(4,1)$
and $(4,1)g_{2}^{3}=(4,1)g_{2}=(4,3).$ Therefore, the action goes as follows:
$k$ | $d$ | Orbit of $\sigma=(d,j)$ | Stabiliser $\Sigma_{\sigma}$
---|---|---|---
12 | 1 | $(1,1)$ | $\operatorname{Aut}(E)$
6 | 2 | $(2,1)$ | $\operatorname{Aut}(E)$
3 | 4 | $(4,1)\xrightarrow[g_{2}^{3}]{g_{1}}(4,3)$ | $\langle g_{2}^{2},g_{1}g_{2}\rangle\simeq D_{2\cdot 3}$
## 4 Skew braces of size $12p$
Once we have done the precomputations for all groups $E$, we bring in the
multiplicative group $F$ of the brace $B_{12}$. In all the sub-cases we follow
the procedure described at the end of Section 2 and the notations in Section 3
for $E$, $\operatorname{Aut}(E)$ and $\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$.
### 4.1 $F=C_{12}$
According to the table in Section 3 and Proposition 4, for a brace of size 12
with multiplicative group $C_{12}$, the additive group can be any group except
for $A_{4}$. In all the sub-cases the group
$\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is as in 3.1 so that $\tau$ identifies as
a pair $(d,j)$ meaning that the given generator of $F$ has image
$\zeta_{d}^{j}.$
#### Case $E=C_{12}$
Let $E=\langle c\rangle\simeq C_{12}$. There is a unique conjugacy class of
regular subgroups of $\mathrm{Hol}(E)$ isomorphic to $C_{12}$, given by the
normal subgroup
$F=\langle\ (c,\operatorname{Id})\ \rangle\subseteq\mathrm{Hol}(E).$
We have stabilisers $\Sigma_{\sigma}=\\{g_{i}\in\operatorname{Aut}(E):i\equiv
1\bmod\mbox{ ord}(\sigma)\\}.$ Since $\pi_{2}(F)=\\{\operatorname{Id}\\}$ we
have
$\mathrm{Hom}(B_{12},\mathbf{Z}_{p}^{*})=\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$.
Since $F$ is a normal subgroup of $\mathrm{Hol}(E)$, we have $\Phi_{g}(F)=F$
for all $g\in\operatorname{Aut}(E)$ and
$\operatorname{Aut}(B_{12})=\operatorname{Aut}(E).$ Therefore, the orbits of
the action of $\operatorname{Aut}(B_{12})$ on
$\mathrm{Hom}(B_{12},\mathbf{Z}_{p}^{*})$ are those computed in 3.1 and we
have one orbit for each size of kernel. The number of addditive structures is
the number of divisors $\sigma_{0}(\gcd(12,p-1))$.
Finally we have to compute orbits of the action of $\Sigma_{\sigma}$ on
$\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$. Since
$\Phi_{g_{i}}(c,\operatorname{Id})=(g_{i}(c),g_{i}\operatorname{Id}g_{i}^{-1})=(c^{i},\operatorname{Id})$
we have $(d,j)$ and $(d,j^{\prime})$ in the same orbit if and only if $j\equiv
ij^{\prime}\bmod d$ for some $i\in\Sigma_{\sigma}$. Therefore, we have the
following representatives for the orbits
$k_{\sigma}$ | $\Sigma_{\sigma}$ | $(d,j)$
---|---|---
$12,6$ | $\langle 5,7\rangle$ | $(d,1)$, $\quad d|12$
$4,2$ | $\langle 7\rangle$ | $(1,1)$, $(2,1)$, $(4,1)$ (3,1), (3,2), (6,1), (6,5), (12,1) (12,5)
$3$ | $\langle 5\rangle$ | $(1,1)$, $(2,1)$, $(3,1)$ (4,1), (4,3), (6,1), (12,1) (12,7)
$1$ | $\\{1\\}$ | $(d,j)$, $\quad d|12$, $\quad j\in\mathbf{Z}_{d}^{*}$
If we count the number of classes according to the size $\frac{12}{d}$ of
$\operatorname{Ker}\tau$, we obtain
$k_{\sigma}$ | $\Sigma_{\sigma}$ | 12 | 6 | 4 | 3 | 2 | 1
---|---|---|---|---|---|---|---
$12,6$ | $\langle 5,7\rangle$ | 1 | 1 | 1 | 1 | 1 | 1
$4,2$ | $\langle 7\rangle$ | 1 | 1 | 2 | 1 | 2 | 2
$3$ | $\langle 5\rangle$ | 1 | 1 | 1 | 2 | 1 | 2
$1$ | $\\{1\\}$ | 1 | 1 | 2 | 2 | 2 | 4
For $k=12$, the trivial $\sigma$, we have one structure for each $\tau$,
otherwise we have two multiplicative structures for each additive one.
###### Proposition 9.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes_{k}C_{12}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes_{k^{\prime}}C_{12}$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times C_{12}$ | $\rtimes_{6}C_{12}$ | $\rtimes_{4}C_{12}$ | $\rtimes_{3}C_{12}$ | $\rtimes_{2}C_{12}$ | $\rtimes_{1}C_{12}$
---|---|---|---|---|---|---
$\mathbf{Z}_{p}\times C_{12}$ | 1 | 1 | 1 | 1 | 1 | 1
$\mathbf{Z}_{p}\rtimes_{6}C_{12}$ | 2 | 2 | 2 | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{4}C_{12}$ | 2 | 2 | 4 | 2 | 4 | 4
$\mathbf{Z}_{p}\rtimes_{3}C_{12}$ | 2 | 2 | 2 | 4 | 2 | 4
$\mathbf{Z}_{p}\rtimes_{2}C_{12}$ | 2 | 2 | 4 | 2 | 4 | 4
$\mathbf{Z}_{p}\rtimes_{1}C_{12}$ | 2 | 2 | 4 | 4 | 4 | 8
#### Case $E=C_{6}\times C_{2}$
We write $E=\langle a\rangle\times\langle b\rangle$. There is one brace with
additive group $C_{6}\times C_{2}$ and multiplicative group $C_{12}$, hence
one conjugation class of regular subgroups of $\mathrm{Hol}(E)$ isomorphic to
$C_{12}$. We may consider as representative the following regular subgroup
$F=\langle c=(a,g_{1})\rangle$
where $(a,g_{1})^{m}$ is $(a^{m}b^{\frac{m}{2}},\operatorname{Id})$ if $m$ is
even and $(a^{m}b^{\frac{m-1}{2}},g_{1})$ if $m$ is odd.
Since $\pi_{2}(F)=\langle g_{1}\rangle$, computation of stabilisers
$\Sigma_{\sigma}$ in 3.2 gives that the $\sigma$’s to consider are $(1\bmod
d,0\bmod 2)$ for every $d|6$, meaning $\sigma(a)=\zeta_{d}$, $\sigma(b)=1$.
Since there is only one for each order, they give a set of representatives for
the action of $\operatorname{Aut}(B_{12})$. Therefore, we have 4 possible
additive structures.
We consider now $\tau\in\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ and the action by
elements $\Phi_{g}$ such that $\Phi_{g}(F)=F$. We need $gg_{1}g^{-1}=g_{1}$,
that is, $g$ belongs to the centraliser of $g_{1}$, which is the Klein group
$\langle g_{2}^{3},g_{1}\rangle$. We have
$\Phi_{g_{1}}(c)=(ab,g_{1})=c^{7},\ \Phi_{g_{2}^{3}}(c)=c^{5},\
\Phi_{g_{2}^{3}g_{1}}(c)=(a^{5}b,g_{1})=c^{11}.$
For $k=12,6$ we have all these automorphisms in $\Sigma_{\sigma}$, therefore,
$(d,j)$ and $(d,j^{\prime})$ are conjugate. If $k=4,2$, the intersection
between $\Sigma_{\sigma}$ and the centralizer is $\langle g_{1}\rangle$ and
the orbit consists just in $(d,j)$ and $(d,7j\bmod d)$. Therefore, we have the
following representatives for the orbits:
$k$ | $\langle g_{2}^{3},g_{1}\rangle{\cap}\Sigma$ | $(d,j)$
---|---|---
$12$ | $\langle g_{2}^{3},g_{1}\rangle$ | $(d,1)$, $\quad d|12$
$6$ | $\langle g_{2}^{3},g_{1}\rangle$ | $(d,1)$, $\quad d|12$
$4$ | $\langle g_{1}\rangle$ | $(1,1)$, $(2,1)$, $(3,1)$, $(3,2)$ (4,1), (6,1), (6,5), (12,1) (12,5)
$2$ | $\langle g_{1}\rangle$ | $(1,1)$, $(2,1)$, $(3,1)$, $(3,2)$ (4,1), (6,1), (6,5), (12,1) (12,5)
###### Proposition 10.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes C_{12}$ is as shown in the following table, where we
need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$ to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times C_{12}$ | $\rtimes_{6}C_{12}$ | $\rtimes_{4}C_{12}$ | $\rtimes_{3}C_{12}$ | $\rtimes_{2}C_{12}$ | $\rtimes_{1}C_{12}$
---|---|---|---|---|---|---
$\mathbf{Z}_{p}\times(C_{6}\times C_{2})$ | 1 | 1 | 1 | 1 | 1 | 1
$\mathbf{Z}_{p}\rtimes_{6}(C_{6}\times C_{2})$ | 2 | 2 | 2 | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{4}(C_{6}\times C_{2})$ | 2 | 2 | 4 | 2 | 4 | 4
$\mathbf{Z}_{p}\rtimes_{2}(C_{6}\times C_{2})$ | 2 | 2 | 4 | 2 | 4 | 4
#### Case $E=D_{2\cdot 6}$
We write $E=\langle r,s\rangle$. There are two braces with additive group
$D_{2\cdot 6}$ and multiplicative group $C_{12}$, hence two conjugation
classes of regular subgroups of $\mathrm{Hol}(D_{2\cdot 6})$ isomorphic to
$C_{12}$. We may consider the following regular subgroups of $\mathrm{Hol}(E)$
isomorphic to $C_{12}$,
$F_{1}=\langle\ c_{1}=(s,g_{2})\ \rangle$
where $c_{1}^{m}$ is $(r^{6-m/2},g_{2}^{m})$ for $m$ even and
$(r^{\frac{k-1}{2}}s,g_{2}^{m})$ for $m$ odd, and
$F_{2}=\langle\ c_{2}=(rs,g_{2}g_{1})\ \rangle$
where $c_{2}^{m}$ is $(r^{m/2},\operatorname{Id})$ for $m$ even and
$(r^{\frac{k+1}{2}}s,g_{2}g_{1})$ for $m$ odd. The groups
$\pi_{2}(F_{1})=\langle g_{2}\rangle\simeq C_{6}$ and $\pi_{2}(F_{2})=\langle
g_{2}g_{1}\rangle\simeq C_{2}$ are not conjugate in $\operatorname{Aut}(E)$,
hence $F_{1}$ and $F_{2}$ are not conjugate in $\mathrm{Hol}(E)$. These groups
show also that the unique nontrivial element in
$\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ to consider is $\sigma=(1,-1)$, meaning
$\sigma(r)=1,\sigma(s)=-1$, whose stabiliser is $\operatorname{Aut}(E)$.
Therefore, we have 2 additive structures for each $F$. The nontrivial one has
cyclic kernel and we denoted it by $\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot
6}$.
Regarding orbits on $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$, since
$\pi_{2}(F_{1})=\langle g_{2}\rangle$ is a normal subgroup, we can take any
$g$ in $\operatorname{Aut}(E)$ to satisfy
$g\pi_{2}(F_{1})g^{-1}=\pi_{2}(F_{1})$. But
$\Phi_{g_{2}^{m}}(c_{1})=(r^{m}s,g_{2})\in F_{1}\iff m=0,3$
and
$\Phi_{g_{2}^{m}g_{1}}(c_{1})=(r^{m}s,g_{2}^{-1})\in F_{1}\iff m=2,5$
Therefore, we are left with the automorphisms in $\langle
g_{2}^{3},g_{2}^{2}g_{1}\rangle\simeq V_{4}$ which happens to be the
normaliser of $\pi_{2}(F_{2})=\langle g_{2}g_{1}\rangle$. For automorphisms of
this group we have
$\Phi_{g_{2}^{3}}(c_{1})=c_{1}^{7},\quad\Phi_{g_{2}^{2}g_{1}}(c_{1})=c_{1}^{5},\quad\Phi_{g_{2}^{5}g_{1}}(c_{1})=c_{1}^{11}$
$\Phi_{g_{2}^{3}}(c_{2})=c_{2}^{7},\quad\Phi_{g_{2}^{2}g_{1}}(c_{2})=c_{2}^{11},\quad\Phi_{g_{2}^{5}g_{1}}(c_{2})=c_{2}^{5}.$
Therefore, in $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ we have just orbit for each
$d$.
###### Proposition 11.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes D_{2.6}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes C_{12}$ is as shown in the following table, where we
need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$ to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times C_{12}$ | $\rtimes_{6}C_{12}$ | $\rtimes_{4}C_{12}$ | $\rtimes_{3}C_{12}$ | $\rtimes_{2}C_{12}$ | $\rtimes_{1}C_{12}$
---|---|---|---|---|---|---
$\mathbf{Z}_{p}\times D_{2\cdot 6}$ | 2 | 2 | 2 | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | 4 | 4 | 4 | 4 | 4 | 4
#### Case $E=\operatorname{Dic}_{12}$
We write $E=\langle\ x,y\ \rangle$. There are two braces with additive group
$\operatorname{Dic}_{12}$ and multiplicative group $C_{12}$, hence two
conjugation classes of regular subgroups of
$\mathrm{Hol}(\operatorname{Dic}_{12})$ isomorphic to $C_{12}$. We may
consider representatives
$\begin{array}[]{l}F_{1}=\langle\ c_{1}=(y,g_{2}g_{1})\ \rangle,\\\
F_{2}=\langle\ c_{2}=(y,g_{2}^{2})\ \rangle,\end{array}$
with $\pi_{2}(F_{1})=\langle g_{2}g_{1}\rangle\simeq C_{2}$ and
$\pi_{2}(F_{2})=\langle g_{2}^{2}\rangle\simeq C_{3}$. Since
$\pi_{2}(F_{1}),\pi_{2}(F_{2})\subseteq\langle g_{2}^{2},g_{1}g_{2}\rangle$,
we have $\pi_{2}(F_{1}),\pi_{2}(F_{2})\subseteq\Sigma_{\sigma}$ for all
$\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$.
In order to compute brace automorphisms we consider first the normalizer of
$\pi_{2}(F_{1})$ in $\operatorname{Aut}(E)$, which is $\langle
g_{2}^{3},g_{2}g_{1}\rangle$, and the normalizer of $\pi_{2}(F_{2})$, which is
$\langle g_{2}\rangle$. Since
$\Phi_{g_{2}^{3}}(F_{1})=\langle\ (y^{3},g_{2}g_{1})\ \rangle=\langle\
c_{1}^{7}\ \rangle=F_{1},$ $\Phi_{g_{2}^{3}}(F_{2})=\langle\
(y^{3},g_{2}^{2})\ \rangle=\langle\ c_{2}^{7}\ \rangle=F_{2},$
we have one orbit for each order and 3 possible additive structures for each
$F$. On the other hand, we have
$\begin{array}[]{lcl}\Phi_{g_{2}}(c_{2})=(xy^{3},g_{2})\not\in
F_{2},&&\Phi_{g_{2}^{2}}(c_{2})=(x^{2}y,g_{2}^{2})\not\in
F_{2},\\\\[2.15277pt]
\Phi_{g_{2}^{3}}(c_{2})=(y^{3},g_{2}^{2})=c_{2}^{7},&&\Phi_{g_{2}^{3}}(c_{1})=(y^{3},g_{2}g_{1})=c_{1}^{7},\\\\[2.15277pt]
\Phi_{g_{2}g_{1}}(c_{2})=(xy,g_{2}^{4})=c_{2}^{5},&&\Phi_{g_{2}g_{1}}(c_{1})=(xy,g_{2}g_{1})=c_{1}^{5},\end{array}$
and this shows that in both cases the automorphism group of the brace is the
maximal subgroup $\langle g_{2}^{3},g_{2}g_{1}\rangle\simeq V_{4}$.
For $\sigma=(1,1)$ or $(2,1)$, we have
$\operatorname{Aut}(B_{12})\cap\Sigma_{\sigma}=\operatorname{Aut}(B_{12})$ and
$\operatorname{Aut}(B_{12})$ acts on
$\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})=\\{(d,j)\\}$ giving a single orbit for
each order $d$. But, for $\sigma=(4,1)$, we have
$\operatorname{Aut}(B_{12})\cap\Sigma_{\sigma}=\langle g_{2}g_{1}\rangle$.
Then, $\tau=(d,j)$ and $\tau^{\prime}=(d,j^{\prime})$ are in the same orbit
if, and only if, $j\equiv 5j^{\prime}\bmod d$, and we have orbits represented
by $(1,1),(2,1),(3,1),(4,1),(4,3),(6,1),(12,1),(12,7).$
###### Proposition 12.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes C_{12}$ is as shown in the following table, where we
need where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$ to
occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times C_{12}$ | $\rtimes_{6}C_{12}$ | $\rtimes_{4}C_{12}$ | $\rtimes_{3}C_{12}$ | $\rtimes_{2}C_{12}$ | $\rtimes_{1}C_{12}$
---|---|---|---|---|---|---
$\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | 2 | 2 | 2 | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | 4 | 4 | 4 | 4 | 4 | 4
$\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$ | 4 | 4 | 4 | 8 | 4 | 8
### 4.2 $F=C_{6}\times C_{2}$
According to the table in Section 3 and Proposition 4, for a brace of size 12
with multiplicative group $C_{12}$, the additive group can be any group. In
all the sub-cases the group $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is as in 3.2
so that $\tau$ identifies as a pair $(j\bmod d,i\bmod 2)$ meaning
$\tau(a)=\zeta_{d}^{j},\ \tau(b)=(-1)^{i}$ for $F=\langle
a\rangle\times\langle b\rangle$.
#### Case $E=C_{12}$
There is one brace with additive group $C_{12}$ and multiplicative group
$C_{6}\times C_{2}$, hence one conjugation class of regular subgroups of
$\mathrm{Hol}(C_{12})$ isomorphic to $C_{6}\times C_{2}$. We may consider the
following representative,
$F=\langle a=(c,g_{7}),\ b=(c^{6},\operatorname{Id})\rangle,$
where $a$ has order 6, $b$ has order 2. Since $\pi_{2}(F)=\langle
g_{7}\rangle$ we cannot consider $\sigma$ of order 4 or 12. Since $F$ is a
normal subgroup of $\mathrm{Hol}(E)$ all automorphisms of $E$ are brace
automorphisms and we have a single orbit for each order. Therefore, we have 4
possible additive structures.
Regarding the action on $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$, since
$\begin{array}[]{lcl}\Phi_{g_{5}}(a)=(c^{5},g_{7})=a^{5},&&\Phi_{g_{5}}(b)=(c^{6},\operatorname{Id})=b,\\\\[2.15277pt]
\Phi_{g_{7}}(a)=(c^{7},g_{7})=ab,&&\Phi_{g_{7}}(b)=(c^{6},\operatorname{Id})=b,\\\\[2.15277pt]
\Phi_{g_{11}}(a)=(c^{11},g_{7})=a^{5}b,&&\Phi_{g_{11}}(b)=(c^{6},\operatorname{Id})=b.\end{array}$
we have the following orbits, acording to the order of $\tau$:
$\begin{array}[]{lccccc}d=1&&&(1,0)\ \bullet\circlearrowleft&&\\\\[8.61108pt]
d=2&&&(1,0)\ \bullet\circlearrowleft&&(0,1)\
\bullet\xleftrightarrow[11]{7}\bullet(1,1)\\\\[12.91663pt]
d=3&&&(1,0)\bullet\xleftrightarrow[11]{5}\bullet(2,0)&&\\\\[12.91663pt]
d=6&&&(1,0)\bullet&&(1,1)\bullet\xleftrightarrow{7}\bullet(4,1)\\\ &&&\
\qquad{{}_{5}}\Big{\updownarrow}{{}_{11}}&&{{}_{5}}\Big{\updownarrow}\quad\quad\Big{\updownarrow}{{}_{5}}\\\
&&&(5,0)\bullet&&(5,1)\bullet\xleftrightarrow{7}\bullet(2,1)\\\ \end{array}$
In the last two cases, when $\Sigma_{\sigma}=\langle g_{7}\rangle$ some points
get disconnected and we have 2 and 4 orbits, respectively.
###### Proposition 13.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes C_{12}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times(C_{6}\times C_{2})$ | $\rtimes_{6}(C_{6}\times C_{2})$ | $\rtimes_{4}(C_{6}\times C_{2})$ | $\rtimes_{2}(C_{6}\times C_{2})$
---|---|---|---|---
$\mathbf{Z}_{p}\times C_{12}$ | 1 | 2 | 1 | 2
$\mathbf{Z}_{p}\rtimes_{6}C_{12}$ | 2 | 4 | 2 | 4
$\mathbf{Z}_{p}\rtimes_{2}C_{12}$ | 2 | 4 | 4 | 8
$\mathbf{Z}_{p}\rtimes_{4}C_{12}$ | 2 | 4 | 4 | 8
#### Case $E=C_{6}\times C_{2}$
Let $E=\langle a\rangle\times\langle b\rangle$. There is one brace with
additive group $C_{6}\times C_{2}$ and multiplicative group $C_{6}\times
C_{2}$, hence one conjugation class of regular subgroups of
$\mathrm{Hol}(C_{6}\times C_{2})$ isomorphic to $C_{6}\times C_{2}$, which has
to be that of the normal subgroup
$F=\langle(a,\operatorname{Id}),(b,\operatorname{Id})\rangle.$
Clearly, we have a single orbit for each order of
$\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$ and 4 possible additive
structures.
Regarding the action on $\tau\in\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$, for
$\sigma=1$ we have one orbit for each order and for $\tau=1$ the orbit has a
single point. According to the stabilisers and orbits computed in 3.2, in the
remaining cases we have
$\Sigma_{(1,0)}$ | $d$ | Orbits
---|---|---
$\langle g_{2}^{3},g_{1}\rangle$ | 2 | $(1,0)\ \bullet\circlearrowleft\quad(0,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (1,1)$
| 3 | $(1,0)\ \bullet\xleftrightarrow{g_{2}^{3}}\bullet\ (2,0)$
| 6 | $(1,0)\ \bullet\xleftrightarrow{g_{2}^{3}}\bullet\ (5,0)$
| | $(2,1)\ \bullet\xleftrightarrow{g_{2}^{3}}\bullet\ (4,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (1,1)\ \bullet\xleftrightarrow{g_{2}^{3}}\bullet(5,1)$
$\langle g_{2}^{2},g_{1}\rangle$ | 2 | $(1,0)\ \bullet\xleftrightarrow{g_{2}}\bullet\ (1,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (0,1)$
| 3 | $(1,0)\ \bullet\circlearrowleft\qquad\bullet\ (2,0)\circlearrowleft$
| 6 | $(1,0)\ \bullet\xleftrightarrow{g_{2}^{2}}\bullet\ (1,1)\ \bullet\xleftrightarrow{g_{2}^{2}}\bullet\ (4,1)$
| | $(5,1)\ \bullet\xleftrightarrow{g_{2}^{2}}\bullet\ (2,1)\ \bullet\xleftrightarrow{g_{2}^{2}}\bullet\ (5,0)$
$\langle g_{1}\rangle$ | 2 | $(1,0)\ \bullet\circlearrowleft\quad(0,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (1,1)$
| 3 | $(1,0)\ \bullet\circlearrowleft\qquad\bullet\ (2,0)\circlearrowleft$
| 6 | $(1,0)\ \bullet\circlearrowleft\qquad\bullet\ (5,0)\circlearrowleft$
| | $(1,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (4,1)$ $(2,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,1)$
###### Proposition 14.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ is as shown in the following
table, where we need where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel
of size $k$ to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times C_{6}\times C_{2}$ | $\rtimes_{6}(C_{6}\times C_{2})$ | $\rtimes_{4}(C_{6}\times C_{2})$ | $\rtimes_{2}(C_{6}\times C_{2})$
---|---|---|---|---
$\mathbf{Z}_{p}\times C_{6}\times C_{2}$ | 1 | 1 | 1 | 1
$\mathbf{Z}_{p}\rtimes_{6}(C_{6}\times C_{2})$ | 2 | 4 | 2 | 4
$\mathbf{Z}_{p}\rtimes_{4}(C_{6}\times C_{2})$ | 2 | 2 | 4 | 4
$\mathbf{Z}_{p}\rtimes_{2}(C_{6}\times C_{2})$ | 2 | 4 | 4 | 8
#### Case $E=A_{4}$
There are two braces with additive group $A_{4}$ and multiplicative group
$C_{6}\times C_{2}$, hence two conjugation classes of regular subgroups of
$\mathrm{Hol}(A_{4})$ isomorphic to $C_{6}\times C_{2}$. We may consider the
following representatives,
$\begin{array}[]{l}F_{1}=\langle a_{1}=((1,2,3),\phi_{(1,2,4)}),\
b_{1}=((1,2)(3,4),\operatorname{Id})\rangle,\\\ F_{2}=\langle
a_{2}=((1,3,4),\phi_{(1,2)(3,4)}),\
b_{2}=((1,3)(2,4),\phi_{(1,3)(2,4)})\rangle,\end{array}$
where $a_{1},a_{2}$ and have order 6 and $b_{1},b_{2}$ have order 2. We have
$\pi_{2}(F_{1})=\langle\phi_{(1,2,4)}\rangle\simeq C_{3}$ and
$\pi_{2}(F_{1})\simeq V_{4}$. Therefore, on one hand we see that the groups
are not conjugate in $\mathrm{Hol}(A_{4})$ and one the other hand that we can
consider all $\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*}).$
Since $\pi_{2}(F_{1})=\langle\phi_{(1,2,4)}\rangle$, in order to find brace
automorphisms we need $\phi_{s}$ in its normaliser, namely
$s\in\langle(2,4),(1,2,4)\rangle\simeq S_{3}$. We have
$\begin{array}[]{lcl}\phi_{\phi_{(2,4)}}(a_{1})=a_{1}^{5},&&\phi_{(2,4)}(b_{1})=a_{1}^{3}b_{1},\\\
\phi_{\phi_{(1,2,4)}}(a_{1})=b_{1}a_{1},&&\phi_{(1,2,4)}(b_{1})=a_{1}^{3}.\end{array}$
Since $\phi_{(2,4)}(1,2,3)=(1,4,3)=(1,3,2)(1,4)(2,3)$, we have that for
$F_{1}$ it is enough to consider $\sigma_{1}$. The normaliser of
$\pi_{2}(F_{1})=V_{4}$ is all $S_{4}$ and we find
$\Phi_{\phi_{(1,2)}}(a_{2})=a_{2}^{5},\quad\phi_{(1,2)}(b_{2})=a_{2}^{3}b_{2}.$
Since $\phi_{(1,2)}(1,2,3)=(1,3,2)(1,4)$ we are also left with $\sigma_{1}$
for $F_{2}$. We have two additive structures for each $F$.
We complete the description of the group of brace automorphisms, since
$\Phi_{\phi_{(1,2,3,4)}}(a_{2})\notin
F_{2},\quad\Phi_{\phi_{(1,2,3)}}(a_{2})=a_{2}b_{2},\quad\Phi_{\phi_{(1,2,3)}}(b_{2})=a_{2}^{3}$
shows that it is not $S_{4}$ and contains the maximal group
$\langle(1,2),(1,2,3)\rangle\simeq S_{3}$. All together,
$\\{s\in
S_{4}:\Phi_{\phi_{s}}(F_{i})=F_{i}\\}=\left\\{\begin{array}[]{l}\langle(2,4),(1,2,4)\rangle\
\mbox{ for }i=1\\\ \langle(1,2),(1,2,3)\rangle\ \mbox{ for
}i=2\end{array}\right.\ \simeq S_{3}.$
The intersection with $\Sigma_{\sigma_{1}}$ is
$A_{4}\cap\\{s\in
S_{4}:\Phi_{\phi_{s}}(F_{i})=F_{i}\\}=\left\\{\begin{array}[]{l}\langle(1,2,4)\rangle\
\mbox{ for }i=1\\\ \langle(1,2,3)\rangle\ \mbox{ for }i=2.\end{array}\right.$
To get the multiplicative structures we need the orbits of
$(s,\tau)\longrightarrow\tau\Phi_{\phi_{s}}$. We have the same action on
generators $\phi_{c}(a)=ab$, $\phi_{c}(b)=a^{3}$, so that
$(j\bmod d,i\bmod 2)\Phi_{\phi_{c}}(a)=(-1)^{i}\zeta_{d}^{j}\quad(j\bmod
d,i\bmod 2)\Phi_{\phi_{c}}(b)=\zeta_{d}^{3j}$
and we obtain the following orbits:
$d=2$ | | $(1,0)\ \bullet\xleftrightarrow{}\bullet\ (1,1)\ \bullet\xleftrightarrow{}\bullet\ (0,1)$
---|---|---
$d=3$ | | $(1,0)\ \bullet\circlearrowleft\qquad\bullet\ (2,0)\circlearrowleft$
$d=6$ | | $(1,0)\ \bullet\xleftrightarrow{}\bullet\ (1,1)\ \bullet\xleftrightarrow{}\bullet\ (4,1)$
| | $(5,0)\ \bullet\xleftrightarrow{}\bullet\ (5,1)\ \bullet\xleftrightarrow{}\bullet\ (2,1)$
###### Proposition 15.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes A_{4}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times(C_{6}\times C_{2})$ | $\rtimes_{6}(C_{6}\times C_{2})$ | $\rtimes_{4}(C_{6}\times C_{2})$ | $\rtimes_{2}(C_{6}\times C_{2})$
---|---|---|---|---
$\mathbf{Z}_{p}\times A_{4}$ | 2 | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes A_{4}$ | 4 | 4 | 8 | 8
#### Case $E=D_{2\cdot 6}$
We write $E=\langle r,s\rangle$. There are two braces with additive group
$D_{2\cdot 6}$ and multiplicative group $C_{6}\times C_{2}$, hence two
conjugation classes of regular subgroups of $\mathrm{Hol}(D_{2\cdot 6})$
isomorphic to $C_{6}\times C_{2}$. We may consider the following
representatives:
$\begin{array}[]{l}F_{1}=\langle a_{1}=(r,\operatorname{Id}),\
b_{1}=(s,g_{1})\rangle,\\\ F_{2}=\langle a_{2}=(r,g_{2}^{4}),\
b_{2}=(s,\operatorname{Id})\rangle,\end{array}$
where $a_{1},a_{2}$ have order 6, and $b_{1},b_{2}$ have order 2. We have
$\pi_{2}(F_{2})=\langle g_{1}\rangle\simeq C_{2}$ and $\pi_{2}(F_{1})=\langle
g_{2}^{4}\rangle\simeq C_{3}$. This shows that they are not conjugate and that
they are subgroups of all the stabilisers $\Sigma_{\sigma}$. Moreover, since
$g_{2}^{3}$ normalizes both $F_{1}$ and $F_{2}$ we have only one orbit in
order $6$ with dihedral kernel and we have 3 additive structures.
Since the normaliser of $\langle g_{1}\rangle$ is $\langle
g_{2}^{3},g_{1}\rangle$ and
$\begin{array}[]{lll}\Phi_{g_{1}}(a_{1})=(r^{5},\operatorname{Id})=a_{1}^{5},&&\Phi_{g_{1}}(b_{1})=(s,g_{1})=b_{1},\\\
\Phi_{g_{2}^{3}}(a_{1})=(r,\operatorname{Id})=a_{1},&&\Phi_{g_{2}^{3}}(b_{1})=(r^{3},g_{1})=a_{1}^{3}b_{1}\end{array}$
we have that for $F_{1}$ the group of brace automorphisms is $\langle
g_{2}^{3},g_{1}\rangle\simeq V_{4}$. On the other hand,
$\begin{array}[]{lll}\Phi_{g_{1}}(a_{2})=(r^{5},g_{2}^{2})=a_{2}^{5},&&\Phi_{g_{1}}(b_{2})=(s,\operatorname{Id})=b_{2},\\\
\Phi_{g_{2}^{3}}(a_{2})=(r,g_{2}^{4})=a_{2},&&\Phi_{g_{2}^{3}}(b_{2})=(r^{3}s,\operatorname{Id})=a_{2}^{3}b_{2}\\\
\Phi_{g_{2}}(a_{2})=(r,g_{2}^{4})=a_{2},&&\Phi_{g_{2}}(b_{2})=(rs,\operatorname{Id})\not\in\mathbf{F}_{2}\\\
\end{array}$
shows that for $F_{2}$ the group of brace automorphisms is also the maximal
subgroup $\langle g_{2}^{3},g_{1}\rangle$. Since the action of generators is
the same, there is no difference in computation of orbits for $F_{1}$ or
$F_{2}$. We obtain
$d=2$ | | $(0,1)\ \bullet\circlearrowleft\qquad(1,0)\ \bullet\xleftrightarrow{g_{2}^{3}}\bullet\ (1,1)$ |
---|---|---|---
$d=3$ | | $(1,0)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (2,0)$ |
$d=6$ | | $(2,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (4,1)$ |
| | $(1,0)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,0)$ |
| | $\quad\ {{}_{g_{2}^{3}}}\Big{\updownarrow}\qquad\ \Big{\updownarrow}{{}_{g_{2}^{3}}}$ |
| | $(1,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,1)$ |
For $\sigma$ of order $6$ with dihedral kernel
$g_{2}^{3}\not\in\Sigma_{\sigma}$ and in first and last case orbits split.
###### Proposition 16.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes D_{2.6}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times(C_{6}\times C_{2})$ | $\rtimes_{6}(C_{6}\times C_{2})$ | $\rtimes_{4}(C_{6}\times C_{2})$ | $\rtimes_{2}(C_{6}\times C_{2})$
---|---|---|---|---
$\mathbf{Z}_{p}\times D_{2\cdot 6}$ | 2 | 4 | 2 | 4
$\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | 4 | 8 | 4 | 8
$\mathbf{Z}_{p}\rtimes_{6}^{d}D_{2\cdot 6}$ | 4 | 12 | 4 | 12
#### Case $E=\operatorname{Dic}_{12}$
We write $E=\langle x,y\rangle$. There are two braces with additive group
$\operatorname{Dic}_{12}$ and multiplicative group $C_{6}\times C_{2}$, hence
two conjugation classes of regular subgroups of
$\mathrm{Hol}(\operatorname{Dic}_{12})$ isomorphic to $C_{6}\times C_{2}$. We
may consider the following representatives
$\begin{array}[]{l}F_{1}=\langle a_{1}=(xy,g_{1}),\
\,b_{1}=(y^{2},\operatorname{Id})\rangle,\\\ F_{2}=\langle
a_{2}=(xy^{2},\operatorname{Id}),\ b_{2}=(y,g_{1})\rangle,\end{array}$
where $a_{1},a_{2}$ have order 6, $b_{1},b_{2}$ have order 2. We note that
since the second component of $a_{2}$ is $\operatorname{Id}$ and the second
components of $a_{1}$ and $a_{1}^{-1}$ are equal to $g_{1}$, they are not
conjugated. We have $\pi_{2}(F_{1})=\pi_{2}(F_{2})=\langle g_{1}\rangle$ and
therefore we cannot consider $\sigma$ order $4$ and then we have
$\Sigma_{\sigma}=\operatorname{Aut}(E)$. We need to determine brace
automorphisms in order to determine the orbits of the action on
$\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$
The normaliser of $\langle g_{1}\rangle$ in $\operatorname{Aut}(E)$ is
$\langle g_{1},g_{2}^{3}\rangle$ and we have
$\begin{array}[]{lll}\Phi_{g_{1}}(a_{1})=(x^{2}y^{3},g_{1})=a_{1}^{5}b_{1},&&\Phi_{g_{1}}(b_{1})=(y^{2},\operatorname{Id})=b_{1}\\\
\Phi_{g_{2}^{3}}(a_{1})=(xy^{3},g_{1})=a_{1}b_{1}&&\Phi_{g_{2}^{3}}(b_{1})=(y^{2},\operatorname{Id})=b_{1}\\\
\Phi_{g_{1}}(a_{2})=(x^{2}y^{2},\operatorname{Id})=a_{2}^{5}&&\Phi_{g_{1}}(b_{2})=(y^{3},g_{1})=a_{2}^{3}b_{2}\\\
\Phi_{g_{2}^{3}}(a_{2})=(xy^{2},\operatorname{Id})=a_{2}&&\Phi_{g_{2}^{3}}(b_{2})=(y^{3},g_{1})=a_{2}^{3}b_{2}.\\\
\end{array}$
This group of brace automorphism gives raise to the following orbits
$d=2$ | | $(1,0)\ \bullet\circlearrowleft\qquad(0,1)\ \bullet\xleftrightarrow[g_{1}]{g_{2}^{3}}\bullet\ (1,1)$ | | $F_{1}$
---|---|---|---|---
| | $(0,1)\ \bullet\circlearrowleft\qquad(1,0)\ \bullet\xleftrightarrow[g_{1}]{g_{2}^{3}}\bullet\ (2,0)$ | | $F_{2}$
$d=3$ | | $(1,0)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (2,0)$ | | $F_{1},F_{2}$
$d=6$ | | $(1,0)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,0)$ | $(1,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (2,1)$ | $F_{1}$
| | | $\quad\ {{}_{g_{2}^{3}}}\Big{\updownarrow}\qquad\ \Big{\updownarrow}{{}_{g_{2}^{3}}}$ |
| | | $(4,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,1)$ |
| | $(2,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (4,1)$ | $(1,0)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,1)$ | $F_{2}$
| | | $\quad\ {{}_{g_{2}^{3}}}\Big{\updownarrow}\qquad\ \Big{\updownarrow}{{}_{g_{2}^{3}}}$ |
| | | $(1,1)\ \bullet\xleftrightarrow{g_{1}}\bullet\ (5,0)$ |
and both groups give raise to the same number of multiplicative structures.
###### Proposition 17.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G=\mathbf{Z}_{p}...$ | $\times(C_{6}\times C_{2})$ | $\rtimes_{{}^{6}}(C_{6}\times C_{2})$ | $\rtimes_{4}(C_{6}\times C_{2})$ | $\rtimes_{2}(C_{6}\times C_{2})$
---|---|---|---|---
$\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | 2 | 4 | 2 | 4
$\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | 4 | 8 | 4 | 8
### 4.3 $F=A_{4}$
According to the table in Section 3 and Proposition 4, for a brace with
multiplicative group a semidirect product of $\mathbf{Z}_{p}$ and $A_{4}$, the
additive group is either a semidirect product of $\mathbf{Z}_{p}$ and
$C_{6}\times C_{2}$ or a semidirect product of $\mathbf{Z}_{p}$ and $A_{4}$.
In both sub-cases the group $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is as in 3.3
so that its elements are $\tau_{j}$, with $j=0,1,2$, defined by
$\tau_{j}(1,2,3)=\zeta_{3}^{j}$ and $\tau_{j}$ trivial on elements of order 2.
#### Case $E=C_{6}\times C_{2}$
There is one brace with additive group $E=\langle a\rangle\times\langle
b\rangle\simeq C_{6}\times C_{2}$ and multiplicative group $A_{4}$, hence one
conjugation class of regular subgroups of $\mathrm{Hol}(C_{6}\times C_{2})$
isomorphic to $A_{4}$. We may consider representative
$F=\langle(a,g_{2}^{2}),(b,\operatorname{Id})\rangle.$
Indeed, $(a,g_{2}^{2})$ has order 3, $(b,\operatorname{Id})$ has order 2, and
$(a,g_{2}^{2})(b,\operatorname{Id})=(a^{4},g_{2}^{2})$ which has order 3. The
Klein subgroup is
$\langle(b,\operatorname{Id}),(a^{3},\operatorname{Id})\rangle$.
We have $\pi_{2}(F)=<g_{2}^{2}>$ and therefore it is a subgroup of
$\Sigma_{\sigma}$ only when $\sigma$ has order $1$ or $3$. For $\sigma=(1,0)$
of order $3$, we have $(1,0)g_{1}g_{2}=(1,0)g_{2}=(2,0)$. Since
$\Phi_{g_{1}g_{2}}(a,g_{2}^{2})=(a^{2}b,g_{2}^{4})=(a,g_{2}^{2})^{2},\quad\Phi_{g_{1}g_{2}}(b,\operatorname{Id})=(a^{3},\operatorname{Id})\in
F$
there is one orbit for each order and two different additive structures.
Let us check if elements of $\Sigma_{\sigma}=\langle g_{2}^{2},g_{1}\rangle$
are brace automorphisms. Since
$\Phi_{g_{2}^{2}}(a)=(a^{2}b,g_{2}^{2})\not\in
F,\quad\Phi_{g_{1}}(a)=(ab,g_{2}^{4})\not\in F$
the action on $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is just given by the
identity and orbits consist on single elements. For a non trivial $\sigma$,
$\tau_{1}$ and $\tau_{2}$ give different multiplicative structures.
###### Proposition 18.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes A_{4}$ is as shown in the following table, where we
need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$ to occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times A_{4}$ | $\mathbf{Z}_{p}\rtimes_{4}A_{4}$
---|---|---
$\mathbf{Z}_{p}\times(C_{6}\times C_{2})$ | 1 | 1
$\mathbf{Z}_{p}\rtimes_{4}(C_{6}\times C_{2})$ | 2 | 4
#### Case $E=A_{4}$
There are four braces with additive group $A_{4}$ and multiplicative group
$A_{4}$, hence four conjugation classes of regular subgroups of
$\mathrm{Hol}(A_{4})$ isomorphic to $A_{4}$. If we write $\rho=(1,2,3)$ and
$\mu=(1,2)(3,4)$, we may consider the following representatives
$\begin{array}[]{l}F_{1}=\langle(\rho,\operatorname{Id}),(\mu,\operatorname{Id})\rangle,\quad
F_{2}=\langle(\rho,\phi_{\rho}),(\mu,\operatorname{Id})\rangle,\\\\[4.30554pt]
F_{3}=\langle(\rho,\phi_{\rho^{-1}}),(\mu,\phi_{\mu})\rangle,\quad
F_{4}=\langle(\rho,\phi_{\rho}),(\mu,\phi_{\mu})\rangle,\\\ \end{array}$
We have $\pi_{2}(F_{1})={\operatorname{Id}}$,
$\pi_{2}(F_{2})=\langle\phi_{\rho}\rangle\simeq C_{3}$, and
$\pi_{2}(F_{3})=\pi_{2}(F_{4})\simeq A_{4}$. But they are not conjugate, since
in $F_{3}$ all elements of order $3$ are of the form $(x,\phi_{x^{-1}})$. On
the other hand, we see that $\pi_{2}(F_{i})\subseteq\Sigma_{\sigma}$ for all
$i=1..4$ and all $\sigma\in\mathrm{Hom}(E,\mathbf{Z}_{p}^{*})$.
Since
$\begin{array}[]{l}\Phi_{\phi_{(1,2)}}((1,2,3),\operatorname{Id})=((1,3,2),\operatorname{Id})=((1,2,3),\operatorname{Id})^{2},\\\
\Phi_{\phi_{(1,2)}}((1,2,3),\phi_{(1,3,2)})=((1,3,2),\phi_{(1,2,3)})=((1,2,3),\phi_{(1,3,2)})^{2},\\\
\Phi_{\phi_{(1,2)}}((1,2,3),\phi_{(1,2,3)})=((1,3,2),\phi_{(1,3,2)})=((1,2,3),\phi_{(1,2,3)})^{2},\end{array}$
and $((1,2)(3,4),\operatorname{Id})$ and $((1,2)(3,4),\phi_{(1,2)(3,4)}))$ are
invariant under $\Phi_{\phi_{(1,2)}}$ we have $\Phi_{2}(F_{i})=F_{i}$ for all
$i$ and elements of order 3 of $\mathrm{Hom}(E,\mathbf{Z}_{p})^{*}$ are in the
same orbit. We have two possible additive structures for each $F_{i}$.
As for morphisms $\tau\in\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$, since
$\Sigma_{\sigma}$ stabilizes them, the action give orbits of a single elememt,
so that we have one orbit in order 1 and two orbits in order 3.
###### Proposition 19.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes A_{4}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes A_{4}$ is as shown in the following table, where we
need $p\equiv 1\pmod{3}$ for the group $\mathbf{Z}_{p}\rtimes_{4}A_{4}$ to
occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times A_{4}$ | $\mathbf{Z}_{p}\rtimes_{4}A_{4}$
---|---|---
$\mathbf{Z}_{p}\times A_{4}$ | 4 | 4
$\mathbf{Z}_{p}\rtimes_{4}A_{4}$ | 8 | 16
### 4.4 $F=D_{2\cdot 6}$
Let us assume that $F$ is isomorphic to the dihedral group $D_{2\cdot 6}$.
From the table in Section 3 we know that for a brace of size $12$ with
multiplicative group $F$, the additive group can be in any isomorphism class
of groups of order $12$ except $A_{4}$. Moreover, in all the sub-cases the
group $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is as in 3.4, so $\tau$ identifies
with any of the pairs $(\pm 1,\pm 1)$. Let us call $\tau_{1}=(1,-1)$,
$\tau_{2}=(-1,-1)$ and $\tau_{3}=(-1,1)$, so that $\tau_{1}$ has kernel
isomorphic to $C_{6}$ and $\tau_{2}$, $\tau_{3}$ have kernel isomorphic to
$D_{2\cdot 3}$.
#### Case $E=C_{12}$
There are two braces with additive group $C_{12}$ and multiplicative group
$D_{2\cdot 6}$, hence two conjugation classes of regular subgroups of
$\mathrm{Hol}(C_{12})$ isomorphic to $D_{2\cdot 6}$. These are represented by
$\begin{array}[]{l}F_{1}=\langle
a_{1}=(c^{2},\mathrm{Id}),b_{1}=(c,g_{11})\rangle,\\\ F_{2}=\langle
a_{2}=(c,g_{7}),b_{2}=(c^{3},g_{11})\rangle,\end{array}$
respectively. Here $a_{1}$, $a_{2}$ have order $6$ and $b_{1}$, $b_{2}$ have
order $2$. Since $\pi_{2}(F_{1})=\langle g_{11}\rangle$ and
$\pi_{2}(F_{2})=\mathrm{Aut}(E)$, in both cases we have
$\pi_{2}(F_{i})\subseteq\Sigma_{\sigma}$ only for $\sigma=(d,1)$, $d=1,2$. On
the other hand, conjugating $a_{1}$ and $b_{1}$ by the three nontrivial
automorphisms of $C_{12}$, we obtain
$\begin{array}[]{lcl}\Phi_{g_{5}}(a_{1})=(c^{10},\operatorname{Id})=a_{1}^{5},&&\Phi_{g_{5}}(b_{1})=(c^{5},g_{11})=a_{1}^{2}b_{1},\\\
\Phi_{g_{7}}(a_{1})=(c^{2},\operatorname{Id})=a_{1},&&\Phi_{g_{7}}(b)=(c^{7},g_{11})=a_{1}^{3}b_{1},\\\
\Phi_{g_{11}}(a_{1})=(c^{10},\operatorname{Id})=a_{1}^{5},&&\Phi_{g_{11}}bg_{11}^{-1}=(c^{11},g_{11})=a_{1}^{5}b_{1}.\end{array}$
Hence, $\Phi_{g}(F_{1})=F_{1}$ for every $g$ and
$\mathrm{Aut}(B_{12})=\mathrm{Aut}(E)$. As for $F_{2}$, the only
$g\in\mathrm{Aut}(E)$ such that $\Phi_{g}(F_{2})=F_{2}$ is $g=g_{5}$, so
$\mathrm{Aut}(B_{12})=\langle g_{5}\rangle$. In any case, each morphism
$\sigma=(1,1),(2,1)$ give rise to two orbits (the ones at the first two rows
at the first table in 3.1). For each of $F_{1}$ and $F_{2}$, the first
morphism gives an additive structure, while the second one gives two additive
structures.
Now, we consider the morphisms $\tau\colon
F\longrightarrow\mathbb{Z}_{p}^{*}$. For $F_{1}$, we have that
$\tau_{2}\Phi_{g_{7}}=\tau_{3}$ with
$g_{7}\in\Sigma_{\sigma}=\mathrm{Aut}(E)$. If we now take $F_{2}$, we need to
consider the action of $\langle g_{5}\rangle$ on
$\mathrm{Hom}(F,\mathbb{Z}_{p}^{*})$, and we see that
$\tau_{2}\Phi_{g_{5}}\neq\tau_{3}$. Therefore, for each $\sigma$, the number
of orbits of the action of $\mathrm{Aut}(B_{12})\cap\Sigma_{\sigma}$ over
$\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is as follows:
$F_{i}\backslash\operatorname{Ker}\tau$ | $C_{6}$ | $D_{2\cdot 3}$
---|---|---
$F_{1}$ | 1 | 1
$F_{2}$ | 1 | 2
###### Proposition 20.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes C_{12}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes D_{2\cdot 6}$ is as shown in the following table.
$N\backslash G$ | $\mathbf{Z}_{p}\times D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{d}D_{2\cdot 6}$
---|---|---|---
$\mathbf{Z}_{p}\times C_{12}$ | 2 | 2 | 3
$\mathbf{Z}_{p}\rtimes_{6}C_{12}$ | 4 | 4 | 6
#### Case $E=C_{6}\times C_{2}$
There is one brace with additive group $C_{6}\times C_{2}$ and multiplicative
group $D_{2\cdot 6}$, and hence one conjugacy class of regular subgroups of
$\mathrm{Hol}(C_{6}\times C_{2})$ isomorphic to $D_{2\cdot 6}$. This is
represented by
$F=\langle r=(a,\mathrm{Id}),s=(b,g_{2}^{3})\rangle\cong D_{2\cdot 6}.$
Here $r$ has order $6$, $s$ has order $2$ and $srs=r^{-1}$.
Since $\pi_{2}(F)=\langle g_{2}^{3}\rangle$, we have that
$\pi_{2}(F)=\Sigma_{\sigma}$ only for $\sigma=(0,0)$ and
$\sigma=(1,0),(0,1),(1,1)$ (corresponding to the first two rows of the final
table at 3.2). On the other hand, it is easily checked that $\Phi_{g}(F)=F$ if
and only if $g\in\langle g_{2}^{3},g_{1}g_{2}^{2}\rangle$. None of these
automorphisms satisfy that $(0,1)=(1,0)g$ or $(0,1)=(1,1)g$, while
$(1,0)=(1,1)g_{1}g_{2}^{2}$. Hence, we obtain two orbits (other than the
trivial one, given by $(0,0)$).
As for the morphisms $\tau\colon F\longrightarrow\mathbb{Z}_{p}^{*}$, we have
that $\tau_{2}=\tau_{3}\Phi_{g}$ if and only if
$g\in\\{g_{1}g_{2}^{2},g_{1}g_{2}^{5}\\}$, with
$g_{1}g_{2}^{5}\in\Sigma_{(0,1)}$, while for $\sigma=(1,0),(1,1)$, no
$g\in\mathrm{Aut}(B_{12})$ satisfies that $\tau_{2}\Phi_{g}=\tau_{3}$. We
collect the number of orbits for the action corresponding to each $\sigma$ in
the following table.
$\Sigma\backslash\operatorname{Ker}\tau$ | $C_{6}$ | $D_{2\cdot 3}$
---|---|---
$\operatorname{Aut}(E)$ | 1 | 1
$\Sigma_{(0,1)}$ | 1 | 1
$\Sigma_{(1,0)}$ | 1 | 2
$\Sigma_{(1,1)}$ | 1 | 2
###### Proposition 21.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes C_{6}\times C_{2}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes D_{2\cdot 6}$ is as shown in the following table.
$N\backslash G$ | $\mathbf{Z}_{p}\times D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{d}D_{2\cdot 6}$
---|---|---|---
$\mathbf{Z}_{p}\times C_{6}\times C_{2}$ | 1 | 1 | 1
$\mathbf{Z}_{p}\rtimes_{6}C_{6}\times C_{2}$ | 4 | 4 | 6
#### Case $E=D_{2\cdot 6}$
There are four braces with additive group $D_{2\cdot 6}$ and multiplicative
group $D_{2\cdot 6}$, and hence four conjugacy classes of regular subgroups of
$\mathrm{Hol}(D_{2\cdot 6})$ isomorphic to $D_{2\cdot 6}$. These are
represented by
$F_{1}=\langle\varphi_{1}=(r,\mathrm{Id}),\psi_{1}=(s,\mathrm{Id})\rangle,$
$F_{2}=\langle\varphi_{2}=(rs,g_{1}),\psi_{2}=(s,\mathrm{Id})\rangle,$
$F_{3}=\langle\varphi_{3}=(r,g_{2}^{4}),\psi_{3}=(s,g_{1})\rangle,$
$F_{4}=\langle\varphi_{4}=(s,g_{2}^{2}),\psi_{4}=(r^{3}s,g_{1})\rangle.$
For $j\in\\{1,2,3,4\\}$, $\varphi_{j}$ has order $6$, $\psi_{j}$ has order
$2$, and $\psi_{j}\varphi_{j}\psi_{j}^{-1}=\varphi_{j}^{-1}$.
Let us call $\sigma_{1}=(1,-1)$, $\sigma_{2}=(-1,-1)$ and $\sigma_{3}=(-1,1)$.
We see immediately that $\pi_{2}(F_{1})=\\{\mathrm{Id}\\}$,
$\pi_{2}(F_{2})=\langle g_{1}\rangle$ and
$\pi_{2}(F_{3})=\pi_{2}(F_{4})=\langle g_{1},g_{2}^{2}\rangle$. All of these
are contained in $\Sigma_{\sigma_{i}}$ for $i\in\\{1,2,3\\}$, so
$\mathrm{Aut}(B_{12})=\\{1,\sigma_{1},\sigma_{2},\sigma_{3}\\}$. On the other
hand, we find that:
* •
For $j=1,3$, $\Phi_{g}(F_{j})=F_{j}$ for every $g\in\mathrm{Aut}(E)$.
* •
$\Phi_{g}(F_{2})=F_{2}$ if and only if $g\in\langle g_{1}\rangle$.
* •
$\Phi_{g}(F_{4})=F_{4}$ if and only if $g\in\langle g_{1}g_{2}^{2}\rangle$.
Now, we have that $\sigma_{3}=\sigma_{2}g_{1}g_{2}^{i}$ if and only if $i$ is
odd, but this is a brace automorphism only for $F_{1}$ and $F_{3}$. Thus,
other than the trivial one, there are two orbits for $F_{1}$ and $F_{3}$ and
three orbits for $F_{2}$ and $F_{4}$.
We examine the classes of morphisms $\tau\colon
F_{i}\longrightarrow\mathbb{Z}_{p}^{*}$. For $F_{1}$, we have that
$\tau_{3}\Phi_{g_{1}g_{2}^{i}}=\tau_{2}$ if and only if $i$ is odd. Hence,
$\tau_{2}\simeq\tau_{3}$ under $\mathrm{Aut}(E)$ but define two different
classes under $\Sigma_{\sigma_{2}}=\Sigma_{\sigma_{3}}$. For $F_{2}$, since
$\tau_{3}\Phi_{g_{1}}\neq\tau_{2}$, we have that $\tau_{2}\not\simeq\tau_{3}$.
For $F_{3}$, we have that $\tau_{3}\Phi_{g_{2}^{i}}=\tau_{2}$ if and only if
$i$ is odd, so we obtain the same conclusions as with $F_{1}$. Finally, for
$F_{4}$ we have that $\tau_{3}\Phi_{g_{1}g_{2}^{2}}\neq\tau_{2}$, so
$\tau_{2}\not\simeq\tau_{3}$.
We collect the numbers of orbits corresponding to each $\sigma$ in the
following tables.
$\Sigma\backslash\operatorname{Ker}\tau$ | $C_{6}$ | $D_{2\cdot 3}$
---|---|---
$\operatorname{Aut}(E)$ | 1 | 1
$\Sigma_{2}=\Sigma_{3}$ | 1 | 2
$\Sigma\backslash\operatorname{Ker}\tau$ | $C_{6}$ | $D_{2\cdot 3}$
---|---|---
$\operatorname{Aut}(E)$ | 1 | 2
$\Sigma_{2}=\Sigma_{3}$ | 1 | 2
$F_{1},F_{3}\qquad\qquad\qquad\qquad\qquad\qquad F_{2},F_{4}$
###### Proposition 22.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes D_{2\cdot 6}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes D_{2\cdot 6}$ is as shown in the following table.
$N\backslash G$ | $\mathbf{Z}_{p}\times D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{d}D_{2\cdot 6}$
---|---|---|---
$\mathbf{Z}_{p}\times D_{2\cdot 6}$ | 4 | 4 | 6
$\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | 8 | 8 | 12
$\mathbf{Z}_{p}\rtimes_{6}^{d}D_{2\cdot 6}$ | 12 | 12 | 24
#### Case $E=\operatorname{Dic}_{12}$
There are four braces with additive group $\operatorname{Dic}_{12}$ and
multiplicative group $D_{2\cdot 6}$, and hence four conjugacy classes of
regular subgroups of $\mathrm{Hol}(\operatorname{Dic}_{12})$ isomorphic to
$D_{2\cdot 6}$. These are represented by
$\begin{array}[]{l}F_{1}=\langle r_{1}=(xy,g_{1}),\
\,s_{1}=(y^{2},g_{1}g_{2}^{3})\rangle,\\\ F_{2}=\langle r_{2}=(xy,g_{2}),\
\,s_{2}=(y^{3},g_{1})\rangle,\\\ F_{3}=\langle
r_{3}=(xy^{2},\operatorname{Id}),\ s_{3}=(y,g_{2}^{3})\rangle,\\\
F_{4}=\langle r_{4}=(xy^{2},g_{2}^{4}),\ \,s_{4}=(y,g_{1})\rangle.\end{array}$
We have that $\pi_{2}(F_{1})=\langle g_{1},g_{2}^{3}\rangle$,
$\pi_{2}(F_{2})=\mathrm{Aut}(E)$, $\pi_{2}(F_{3})=\langle g_{2}^{3}\rangle$
and $\pi_{2}(F_{4})=\langle g_{1},g_{2}^{2}\rangle$. Hence, for each
$j=1,2,3,4$, $\pi_{2}(F_{j})\subseteq\Sigma_{\sigma}$ only for $\sigma=(d,1)$
with $d=1,2$, in which case $\Sigma_{\sigma}=\mathrm{Aut}(E)$. On the other
hand, we find that
$\Phi_{g}(F_{1})=F_{1}\iff g\in\langle
g_{1}g_{2}^{3}\rangle,\quad\Phi_{g}(F_{2})=F_{2}\iff g\in\langle
g_{1}g_{2}^{5}\rangle$
and for $j\in\\{3,4\\}$, $\Phi_{g}(F_{j})=F_{j}$ for every
$g\in\mathrm{Aut}(E)$.
Let us examine the classes of morphisms $\tau\colon
F_{j}\longrightarrow\mathbb{Z}_{p}^{*}$ for each $j$. For $F_{1}$ we have that
$\tau_{2}(s_{1})\neq\tau_{3}\Phi_{g_{1}g_{2}^{3}}(s_{1})$, so
$\tau_{2}\not\simeq\tau_{3}$. For $F_{2}$,
$\tau_{2}(s_{2})\neq\tau_{3}\Phi_{g_{1}g_{2}^{5}}(s_{2})$, whence
$\tau_{2}\not\simeq\tau_{3}$ neither. For $F_{3}$, we see that
$\tau_{2}=\tau_{3}\Phi_{g_{1}g_{2}^{2}}$ with $g_{1}g_{2}^{2}$, hence
$\tau_{2}\simeq\tau_{3}$. Finally, for $F_{4}$, we see that
$\tau_{2}=\tau_{3}\Phi_{g_{1}}$, so $\tau_{2}\simeq\tau_{3}$. Then, the orbits
for each $\sigma$ are as in the following table.
$F_{j}\backslash\operatorname{Ker}\tau$ | $C_{6}$ | $D_{2\cdot 3}$
---|---|---
$F_{1}$ | 1 | 2
$F_{2}$ | 1 | 2
$F_{3}$ | 1 | 1
$F_{4}$ | 1 | 1
###### Proposition 23.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes Dic_{12}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes D_{2\cdot 6}$ is as shown in the following table.
$N\backslash G$ | $\mathbf{Z}_{p}\times D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | $\mathbf{Z}_{p}\rtimes_{6}^{d}D_{2\cdot 6}$
---|---|---|---
$\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | 4 | 4 | 6
$\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | 8 | 8 | 12
### 4.5 $F=\operatorname{Dic}_{12}$
We take for $F$ the dicyclic group of order 12. According to the table in
Section 3 and Proposition 4, for a brace of size 12 with multiplicative group
$F=\operatorname{Dic}_{12}$, the additive group can be any group of order 12.
In all the sub-cases the group $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ is as in
3.5 so that $\tau$ identifies as a pair $(1,1),(2,1),(4,1)$ or $(4,3)$.
#### Case $E=C_{12}$
Let $E=\langle c\rangle\simeq C_{12}$. There is one brace with additive group
$C_{12}$ and multiplicative group $\operatorname{Dic}_{12}$, hence one
conjugation class of regular subgroups of $\mathrm{Hol}(C_{12})$ isomorphic to
$\operatorname{Dic}_{12}$. We may consider representative
$F=\langle x=(c^{4},\operatorname{Id}),\ y=(c^{3},g_{5})\rangle,$
where $x$ has order 3, $y$ has order 4 and $y\,x=x^{2}\,y$.
Since $\pi_{2}(F)=\langle g_{5}\rangle$ we have
$\pi_{2}(F)\subseteq\Sigma_{\sigma}$ only for $\sigma=(d,j)$ with $d=1,2,4$.
Since $F$ is a normal subgroup of $\mathrm{Hol}(E)$, we have
$\Phi_{g}(F)=F,\,\forall g\in\operatorname{Aut}(E)$. Since $\sigma=(4,1)$ and
$\sigma^{\prime}=(4,3)$ are in the same orbit, the action of
$\operatorname{Aut}(B_{12})$ gives one orbit for each order. All together, we
have 3 additive structures for $F$ if $p\equiv 1\pmod{4}$ and 2 otherwise.
For $\sigma$ of order $1$ or two,
$\Sigma_{\sigma}=\operatorname{Aut}(E)=\operatorname{Aut}(B_{12})$ and the
action on $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$ gives one orbit for each order.
For $\sigma$ of order $4$, to consider the action of
$\operatorname{Aut}(B_{12})\cap\Sigma_{\sigma}$ we just have to check the
action of $g_{5}$ on $y$. Since
$\Phi_{g_{5}}(y)=(c^{3},g_{5})=y$
$\tau=(4,1)$ and $\tau^{\prime}=(4,3)$ are not in the same orbit and this
gives two multiplicative structures with kernel of order 3 for $\sigma$.
###### Proposition 24.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes C_{12}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$
---|---|---|---
$\mathbf{Z}_{p}\times C_{12}$ | 1 | 1 | 1
$\mathbf{Z}_{p}\rtimes_{6}C_{12}$ | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{3}C_{12}$ | 2 | 2 | 4
#### Case $E=C_{6}\times C_{2}$
We write $E=\langle a\rangle\times\langle b\rangle$. There is one brace with
additive group $C_{6}\times C_{2}$ and multiplicative group
$\operatorname{Dic}_{12}$, hence one conjugation class of regular subgroups of
$\mathrm{Hol}(C_{6}\times C_{2})$ isomorphic to $\operatorname{Dic}_{12}$ for
which we may consider as representative the following subgroup
$F=\langle x=(a^{2},\operatorname{Id}),y=(b,g_{1}g_{2})\rangle.$
Since $\pi_{2}(F)=\langle g_{1}g_{2}\rangle$ there is only one nontrivial
$\sigma$ to consider, namely that of order 2 identified as $(1,1)$ and defined
by $\sigma(a)=\sigma(b)=\sigma_{2}=-1$, which has stabiliser $\langle
g_{2}^{3},g_{2}^{4}g_{1}\rangle\simeq V_{4}$. Since we only have $1$ and
$\sigma$, it is clear that they are fixed points for the action of
$\operatorname{Aut}(B_{12)}$ and we have two orbits, namely 2 additive
structures for $F$.
Since $F$ is normal in $\mathrm{Hol}(E)$, the elements of $\langle
g_{2}^{3},g_{2}^{4}g_{1}\rangle$ are brace automorphisms and to determine its
action on $\mathrm{Hom}(F,\mathbf{Z}_{p}^{*})$, we just have to check the
action on $y$. From $\Phi_{g_{1}g_{2}}(y)=y^{3}$, we obtain that
$\tau=(4,3)=(4,1)\Phi_{g_{1}g_{2}}$ and there is just one orbit in order 4.
###### Proposition 25.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes(C_{6}\times C_{2})$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ is as shown in the following
table, where we need $p\equiv 1\pmod{4}$ for the group
$\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$ to occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$
---|---|---|---
$\mathbf{Z}_{p}\times(C_{6}\times C_{2})$ | 1 | 1 | 1
$\mathbf{Z}_{p}\rtimes_{6}(C_{6}\times C_{2})$ | 2 | 2 | 2
#### Case $E=A_{4}$
There are two braces with additive group $A_{4}$ and multiplicative group
$\operatorname{Dic}_{12}$, hence two conjugation classes of regular subgroups
of $\mathrm{Hol}(A_{4})$ isomorphic to $\operatorname{Dic}_{12}$. We may
consider the following representatives
$\begin{array}[]{l}F_{1}=\langle x_{1}=((1,3,2),\phi_{(1,2,3)}),\
y_{1}=((1,3)(2,4),\phi_{(1,2)})\rangle,\\\\[4.30554pt] F_{2}=\langle
x_{2}=((1,3,4),\operatorname{Id}),\
y_{2}=((1,3)(2,4),\phi_{(1,3,2,4)})\rangle,\end{array}$
where $x_{1},x_{2}$ have order 3, $y_{1},y_{2}$ have order 4 and
$y_{i}x_{i}=x_{i}^{2}y_{i}$ for $i=1,2$. We have
$\pi_{2}(F_{1})=\langle\phi_{(1,2,3)}\phi_{(1,2)}\rangle\simeq S_{3}$ and
$\pi_{2}(F_{1})=\langle\phi_{(1,3,2,4)}\rangle$ which shows that $F_{1},F_{2}$
are not conjugate in $\mathrm{Hol}(A_{4})$ and that their projections are not
subgroups of $A_{4}$, so that we obtain only one additive structure for each
$F_{1},F_{2}$, corresponding to the direct product.
Since $\Sigma_{1}=\operatorname{Aut}(E)$ in order to determine multiplicative
structures, we have one orbit for order 1 and one for order 2, and we only
need to check if $\tau=(4,1)$ and $\tau^{\prime}=(4,3)$ differ in a brace
automorphism. We have
$\Phi_{\phi_{(1,2)}}(x_{1})=x_{1}^{2},\quad\Phi_{\phi_{(1,2)}}(y_{1})=y_{1}^{3},\quad\Phi_{\phi_{(1,2)}}(F_{1})=F_{1},\quad(4,3)=(4,1)\Phi_{\phi_{(1,2)}}$
$\Phi_{\phi_{(3,4)}}(x_{2})=x_{2}^{2},\quad\Phi_{\phi_{(3,4)}}(y_{2})=y_{2}^{3},\quad\Phi_{\phi_{(3,4)}}(F_{2})=F_{2},\quad(4,3)=(4,1)\Phi_{\phi_{(3,4)}}.$
So, for $F_{1}$ and $F_{2}$ we obtain one class of morphisms $\tau$ with
kernel of order 6 and, if $p\equiv 1\pmod{4}$, there is another class of
morphisms $\tau$ with kernel of order 3.
###### Proposition 26.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes A_{4}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ is as shown in the following
table, where we need $p\equiv 1\pmod{4}$ for the group
$\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$ to occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$
---|---|---|---
$\mathbf{Z}_{p}\times A_{4}$ | 2 | 2 | 2
#### Case $E=D_{2\cdot 6}$
There are two braces with additive group $D_{2\cdot 6}$ and multiplicative
group $\operatorname{Dic}_{12}$, hence two conjugation classes of regular
subgroups of $\mathrm{Hol}(D_{2\cdot 6})$ isomorphic to
$\operatorname{Dic}_{12}$. We take representatives
$\begin{array}[]{l}F_{1}=\langle x_{1}=(r^{2},\operatorname{Id}),\
y_{1}=(sr^{5},g_{2}^{3})\rangle,\\\ F_{2}=\langle x_{2}=(r^{2},g_{2}^{2}),\
y_{2}=(r^{2}s,g_{2}g_{1})\rangle,\end{array}$
where $x_{1},x_{2}$ have order 3, $y_{1},y_{2}$ have order 4 and
$y_{i}x_{i}=x_{i}^{2}y_{i}$ for $i=1,2$.
Since $\pi_{2}(F_{1})=\langle g_{2}^{3}\rangle$ and $\pi_{2}(F_{2})=\langle
g_{2}^{2},g_{2}g_{1}\rangle$, we see that the groups are not conjugate and
that the only nontrivial $\sigma$ to consider is that of order $2$ with cyclic
kernel. So, the number of additive structures for each $F_{i}$ is $2$.
Again, to determine multiplicative structures it is enough to check if $(4,3)$
and $(4,1)$ differ in a brace automorphism. We have
$\Phi_{g_{2}^{3}}(x_{1})=x_{1},\quad\Phi_{g_{2}^{3}}(y_{1})=(r^{4}s,g_{2}^{3})=y_{1}^{3},\quad\Phi_{g_{2}^{3}}(F_{1})=F_{1},\quad(4,3)=(4,1)\Phi_{g_{2}^{3}}$
$\Phi_{g_{2}^{3}}(x_{2})=x_{2},\quad\Phi_{g_{2}^{3}}(y_{2})=(r^{5}s,g_{2}g_{1})=y_{2}^{3},\quad\Phi_{g_{2}^{3}}(F_{2})=F_{2},\
(4,3)=(4,1)\Phi_{g_{2}^{3}}$
So, for each $F_{1},F_{2}$ we obtain one orbit for each order.
###### Proposition 27.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N=\mathbf{Z}_{p}\rtimes D_{2\cdot 6}$ and multiplicative group
$G=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ is as shown in the following
table, where we need $p\equiv 1\pmod{4}$ for the group
$\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$ to occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$
---|---|---|---
$\mathbf{Z}_{p}\times D_{2\cdot 6}$ | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{6}^{c}D_{2\cdot 6}$ | 4 | 4 | 4
#### Case $E=\operatorname{Dic}_{12}$
There are two braces with additive group $\operatorname{Dic}_{12}$ and
multiplicative group $\operatorname{Dic}_{12}$, hence two conjugation classes
of regular subgroups of $\mathrm{Hol}(\operatorname{Dic}_{12})$ isomorphic to
$\operatorname{Dic}_{12}$. We may consider the following representatives
$\begin{array}[]{l}F_{1}=\langle x_{1}=(x,\operatorname{Id}),\
y_{1}=(y,\operatorname{Id})\rangle,\\\ F_{2}=\langle x_{2}=(x^{2},g_{2}^{2}),\
y_{2}=(x^{2}y,g_{2}g_{1})\rangle,\end{array}$
where $x_{1},x_{2}$ have order 3, $y_{1},y_{2}$ have order 4 and
$y_{i}x_{i}=x_{i}^{2}y_{i}$ for $i=1,2$. Since
$\pi_{2}(F_{1})=\\{\operatorname{Id}\\}$ and $\pi_{2}(F_{2})=\langle
g_{2}^{2},g_{2}g_{1}\rangle=\langle g_{2}^{2},g_{1}g_{2}\rangle$ we see that
groups are not conjugate and that we can consider all possible $\sigma$. We
have
$\Phi_{g_{1}}(x_{1})=x_{1}^{2},\quad\Phi_{g_{1}}(y_{1})=y_{1}^{3},\quad\Phi_{g_{1}}(F_{1})=F_{1},$
$\Phi_{g_{1}}(x_{2})=x_{2}^{2},\quad\Phi_{g_{1}}(y_{2})=y_{2}^{3},\quad\Phi_{g_{1}}(F_{2})=F_{2},$
and $(4,3)=(4,1)g_{1}\mbox{ in }\mathrm{Hom}(E,\mathbf{Z}_{p}^{*}).$ Then, the
number of additive structures for each $F_{i}$ is $3$ if $p\equiv 1\pmod{4}$
and $2$ otherwise.
For $\sigma=(2,1)$, since $\Sigma_{\sigma}=\operatorname{Aut}(E)$, the above
computation shows that $(4,3)=(4,1)\Phi_{g_{1}}$ and we have just one orbit of
order 4. But for $\sigma$ order 4, since $\Sigma_{\sigma}$ is also the
stabiliser of the $\tau$’s of order 4, we have fixed points and 2 orbits.
###### Proposition 28.
Let $p$ be a prime number, $p\geq 7$. The number of braces with additive group
$N:=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ and multiplicative group
$G:=\mathbf{Z}_{p}\rtimes\operatorname{Dic}_{12}$ is as shown in the following
table, where we need $p\equiv 1\pmod{\frac{12}{k}}$ for a kernel of size $k$
to occur.
$N\backslash G$ | $\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | $\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$
---|---|---|---
$\mathbf{Z}_{p}\times\operatorname{Dic}_{12}$ | 2 | 2 | 2
$\mathbf{Z}_{p}\rtimes_{6}\operatorname{Dic}_{12}$ | 4 | 4 | 4
$\mathbf{Z}_{p}\rtimes_{3}\operatorname{Dic}_{12}$ | 4 | 4 | 8
### 4.6 Total numbers
For a prime number $p\geq 7$ we compile in the following tables the total
number of skew left braces of size $12p$.
The additive group is a semidirect product $\mathbf{Z}_{p}\rtimes E$ and the
multiplicative group is a semidirect product $\mathbf{Z}_{p}\rtimes F$. In the
first column we have the possible $E$’s and in the first row the possible
$F$’s.
* •
If $p\equiv 11\pmod{12}$
| $C_{12}$ | $C_{6}\times C_{2}$ | $A_{4}$ | $D_{2\cdot 6}$ | $\operatorname{Dic}_{12}$ |
---|---|---|---|---|---|---
$C_{12}$ | 6 | 9 | 0 | 21 | 6 |
$C_{6}\times C_{2}$ | 6 | 8 | 1 | 17 | 6 |
$A_{4}$ | 0 | 4 | 4 | 0 | 4 |
$D_{2\cdot 6}$ | 12 | 34 | 0 | 90 | 12 |
$\operatorname{Dic}_{12}$ | 12 | 18 | 0 | 42 | 12 |
| 36 | 73 | 5 | 170 | 40 | $\mathbf{324}$
* •
If $p\equiv 5\pmod{12}$
| $C_{12}$ | $C_{6}\times C_{2}$ | $A_{4}$ | $D_{2\cdot 6}$ | $\operatorname{Dic}_{12}$ |
---|---|---|---|---|---|---
$C_{12}$ | 17 | 9 | 0 | 21 | 17 |
$C_{6}\times C_{2}$ | 9 | 8 | 1 | 17 | 9 |
$A_{4}$ | 0 | 4 | 4 | 0 | 6 |
$D_{2\cdot 6}$ | 18 | 34 | 0 | 90 | 18 |
$\operatorname{Dic}_{12}$ | 34 | 18 | 0 | 42 | 34 |
| 78 | 73 | 5 | 170 | 84 | $\mathbf{410}$
* •
If $p\equiv 7\pmod{12}$
| $C_{12}$ | $C_{6}\times C_{2}$ | $A_{4}$ | $D_{2\cdot 6}$ | $\operatorname{Dic}_{12}$ |
---|---|---|---|---|---|---
$C_{12}$ | 36 | 54 | 0 | 21 | 6 |
$C_{6}\times C_{2}$ | 36 | 46 | 8 | 17 | 6 |
$A_{4}$ | 0 | 32 | 32 | 0 | 4 |
$D_{2\cdot 6}$ | 24 | 68 | 0 | 90 | 12 |
$\operatorname{Dic}_{12}$ | 24 | 36 | 0 | 42 | 12 |
| 120 | 236 | 40 | 170 | 40 | $\mathbf{606}$
* •
If $p\equiv 1\pmod{12}$
| $C_{12}$ | $C_{6}\times C_{2}$ | $A_{4}$ | $D_{2\cdot 6}$ | $\operatorname{Dic}_{12}$ |
---|---|---|---|---|---|---
$C_{12}$ | 94 | 54 | 0 | 21 | 17 |
$C_{6}\times C_{2}$ | 54 | 46 | 8 | 17 | 9 |
$A_{4}$ | 0 | 32 | 32 | 0 | 6 |
$D_{2\cdot 6}$ | 36 | 68 | 0 | 90 | 18 |
$\operatorname{Dic}_{12}$ | 68 | 36 | 0 | 42 | 34 |
| 252 | 236 | 40 | 170 | 84 | $\mathbf{782}$
With the results summarized in the above tables, the validity of conjecture
(1) is then established.
## References
* [1] V.G. Bardakov, M.V. Neshchadim, M.K. Yadav, _Computing skew left braces of small orders_ , Internat. J. Algebra Comput. 30 (2020), no. 4, 839–851.
* [2] F. Cedó, _Left Braces: Solutions of the Yang-Baxter Equation_ , Advances in Group Theory and Applications, 5 (2018), 3–90.
* [3] T. Crespo, D. Gil-Muñoz, A. Rio, M. Vela, _Left braces of size $8p$_, Journal of Algebra Volume 617, (2023), 317-339.
* [4] T. Crespo, D. Gil-Muñoz, A. Rio, M. Vela, _Inducing braces and Hopf Galois structures_ , to appear in J. Pure Appl. Algebra (2023)
* [5] M. J. Curran, _Automorphisms of semidirect products_ , Mathematical Proceedings of the Royal Irish Academy Vol. 108A, No. 2 (2008), pp. 205-210
* [6] L. Guarnieri, L. Vendramin, _Skew braces and the Yang-Baxter equation_ , Math. Comp. 86 (2017), 2519-2534.
* [7] W. Rump, _Braces, radical rings, and the quantum Yang–Baxter equation_ , Journal of Algebra, 307 (2007), 153-170.
* [8] A. Smoktunowicz, L. Vendramin, _On skew braces (with an appendix by N. Byott and L. Vendramin)_ , J. Comb. Algebra 2 no. 1 (2018), 47–86
* [9] L. Vendramin, _Problems on skew left braces_ , Advances in Group Theory and Applications, 7 (2019), 15–37.
|
# AI-Assisted Design Concept Exploration
Through Character Space Construction
Shin Sano 1,3∗ and Seiji Yamada 1,2
1Department of Informatics, School of Multidisciplinary Sciences,
Graduate University for Advanced Studies (SOKENDAI), Hayama , Japan
2Digital Content and Media Sciences Research Division, National Institute of
Informatics , Tokyo, Japan
3Institute for Creative Integration , Oakland, California, USA
###### Abstract
We propose an AI-assisted design concept exploration tool, the “Character
Space Construction” (“CSC”). Concept designers explore and articulate the
target product aesthetics and semantics in language, which is expressed using
“Design Concept Phrases” (“DCPs”), that is, compound adjective phrases, and
contrasting terms that convey what are not their target design concepts.
Designers often utilize this dichotomy technique to communicate the nature of
their aesthetic and semantic design concepts with stakeholders, especially in
an early design development phase. The CSC assists this designers’ cognitive
activity by constructing a “Character Space” (“CS”), which is a semantic
quadrant system, in a structured manner. A CS created by designers with the
assistance of the CSC enables them to discern and explain their design
concepts in contrast with opposing terms. These terms in a CS are retrieved
and combined in the CSC by using a knowledge graph. The CSC presents terms and
phrases as lists of candidates to users from which users will choose in order
to define the target design concept, which is then visualized in a CS. The
participants in our experiment, who were in the “arts and design” profession,
were given two conditions under which to create DCPs and explain them. One
group created and explained the DCPs with the assistance of the proposed CSC,
and the other did the same task without this assistance, given the freedom to
use any publicly available web search tools instead. The result showed that
the group assisted by the CSC indicated their tasks were supported
significantly better, especially in exploration, as measured by the Creativity
Support Index (CSI).
_K_ eywords Intelligent interactive system $\cdot$ Industrial design $\cdot$
Concept design $\cdot$ Creativity support tool $\cdot$ Product semantics
$\cdot$ Design aesthetics $\cdot$ Human-computer interaction $\cdot$ Lexical
semantics
## 1 Introduction
Our research is motivated by an observation in the professional industrial
design domain and a research question derived from it: “could concept
designers’ creative activities, especially verbalizing novel design concept
phrase, be modeled and computationally supported?” The concept design used in
industrial design is primarily concerned with developing the aesthetics and
semantics of products[1]. The visual appearance and meaning behind a product
is frequently a key attribute of a product’s appeal to consumers [2, 3, 4], as
the functionality of a product is increasingly taken for granted, and users
are looking for different levels of appreciation [5]. Consequently, designers
are in charge of not only creating the visual appearance of products but also
verbally articulating the product semantics and emotional feelings attached to
them [6]. Design concepts articulated using this verbal mode often have to be
explored, communicated, understood, and approved by the project stakeholders
during a design project. Krippendorff [1] called this practice a “design
discourse”, describing, “Designers have to justify the aesthetics of mass-
produced products, products that would ideally be of use to everyone, not
individual works of art.” This characterizes the special role industrial
designers play, as distinguished from engineering designers, and this role is
most clearly observable in the automobile industry [7].
While emotional components in industrial design are important [8, 9, 10],
automotive customers especially value emotional experiences when they own and
use products, such as those related to their self-image and brand messages
reflected in vehicles [11, 12]. Designers and other stakeholders exchange
views on the form of a vehicle design as part of the design process. Tovey [7]
characterized the language they use as “idiosyncratic and atypical,” requiring
them to explore and communicate different shapes and features in words, such
as “slippery,” “exciting,” “fluid,” “tailored,” and “sheer.” Such language
used in car design studios describes particular forms or connotes a “feeling.”
Bouchard et al. [13] described these verbal expressions as “intermediate
representations” that will eventually be translated into visual aesthetic and
semantic features of product designs. Designers in practice often feel that
they run out of these words to express different design concepts in different
projects.
The present research attempts to uncover professional concept designers’
cognitive activities, that are traditionally highly empirical and
undocumented, for when they explore and form such verbal representations of
design concepts and also attempts to computationally support the process by an
AI algorithm. The contributions of this research are as follows:
* •
We defined concept designers’ cognitive activities in exploring, generating,
and explaining a design concept with a “Design Concept Phrase” (“DCP”) and
contrasting concepts, and modeled this process as constructing what we call a
“Character Space” (“CS”).
* •
We implemented this model of constructing a CS into an AI algorithm and
created an interactive system, which supports concept designers’ creative
process described above.
* •
We defined a DCP as a compound phrase using an adjective-“noun-form-of”
adjective formula.
In this paper, we will first discuss the background of the topic space
together with related works. We will highlight how professional concept
designers utilize a verbal mode of exploration to generate and communicate
novel design concepts in the early stage of the design process, especially
with an emphasis on their process of using a dichotomy technique, taking as an
example the development of an automotive design concept. Then, we propose an
AI-assisted interactive system, “Character Space Construction” (“CSC”), and
its methodologies, utilizing the quadrant system, that is, the CS, for
designers to generate and use DCPs. Last, we introduce our experiment with
participants in design profession to measure the effectiveness of the tool,
followed by discussions.
We illustrate a case study of how they use DCPs from an automotive design,
however, these adjective phrases are used in the practice of industrial design
in general [14], fashion design[15], furniture design [16], or even in the
tourism industry [17]. Therefore, the proposed methods can contribute to
various design and marketing fields.
## 2 Background
Figure 1: (A) left: Character space (CS) for “kinetic warmth” (B) right:
Toyota Concept-i design derived from “kinetic warmth” (©2021 Toyota Motor
Sales, U.S.A., Inc.).
### 2.1 Defining Design Concept Phrase and Character Space
First, we define the “Design Concept Phrase” (“DCP”) as a compound adjective
phrase that conveys product aesthetics and semantics. In the industry, such
phrases are sometimes called “keywords” [18]; however, we call them DCP in
this research since the term “keywords” is used differently in the natural
language processing (NLP) context. A DCP, “kinetic warmth,” for example, was
created by the designers at Toyota’s North American design studio for
Concept-i (Fig. 1-(B)). Note that the noun “warmth” was converted into a noun
from the adjective “warm.” Ian Cartabiano, the chief designer of Concept-i at
Toyota’s Calty Design Research studio explained, “There’s a school of design
at the moment which makes a car feel just like a laptop on wheels, an
impersonal transport unit. We wanted to do something more soulful, intriguing,
lively, and human.” When they identified “kinetic” and “warmth,” the design
team explored other terms, such as “dynamic,” “engaging,” “intriguing,”
“soulful,” “human,” “lively,” and “passionate” 111Car Design News (2017). CES
2017: Toyota Concept-i in depth. https://www.cardesignnews.com/cdn-
live/ces-2017-toyota-concept-i-in-depth/22991.article [Accessed October 22,
2021]. . Notice that Cartabiano expressed a type of aesthetics and semantics
to be avoided, that is, “a laptop on wheels, an impersonal transport unit.”
They later put those feelings into two other contrasting terms, “calm” and
“cold,” and placed them onto a quadrant system, which we define in this
research as the “Character Space” (“CS”, Fig. 1(A)). These contrasting terms
make “kinetic warmth” understanable by stating what is NOT “kinetic warmth.”
A CS visually resembles a dimensionality reduction on 2D semantic space
employing multidimensional scaling (MDS); however, the way a CS is formed in
practice fundamentally differs from how an MDS is formed. A CS is informally
utilized as a synthesis tool employing a quadrant system, on which a user
identifies each end of the axis through creative exploration and speculation.
The primary focus is the upper-right quadrant, which represents the target
concept expressed by a compound adjective; then, the other quadrants will be
defined as the user identifies opposite terms to each of the compound
adjectives (e.g., “kinetic-calm” and “warm-cold”). These polar terms on the
axes consequently form semantic differentials [19], however in a CS, the two
terms in an orthogonal relationship (e.g., “kinetic” on the vertical axis and
“warm” on the horizontal axis) are identified first before a pair of polar
terms on a semantic differential is defined. Also, that these polar terms can
be defined by the designers as a part of creation depending on what aspect of
a particular adjective they want to emphasize. For instance, an adjective,
“beautiful,” can be perceived differently depending on the term at the other
end (e.g., “beautiful”$\leftrightarrow$“ugly” or
“beautiful”$\leftrightarrow$“practical”). A CS is completed when all four ends
on the axes are defined. In this research, we label the top end of the
vertical axis as word 1 ($w_{1}$), and the rest are then labeled in clockwise
order as word 2 ($w_{2}$), word 3 ($w_{3}$), and word 4 ($w_{4}$). The upper-
right quadrant, represented by the combination of $w_{1}$ and $w_{2}$, is the
target design-concept phrase. All the other quadrants, represented by
$w_{2}$-$w_{3}$, $w_{3}$-$w_{4}$, and $w_{4}$-$w_{1}$, are contrasting
concepts to be used to explain the target DCP in terms of how and by which
attributes they differ and what already exists or what is to be avoided.
Meanwhile, a multidimensional scaling with MDSs has been widely used as an
analytical tool since first being introduced by Ramsay [20] and became a
popular data visualization tool when Zimet et al. [21] used them in the form
of the Multidimensional Scale of Perceived Social Support (MSPSS). MDSs have
been successfully implemented in new product development domains, including
concept mapping [22] and kansei engineering [23]. They visualize the spatial
distance between individual cases of a data set according to similarities. The
usage of MDSs first involves collecting data on the basis of ratings of
statements or semantic differentials for each case, followed by dimensionality
reduction done by using statistical techniques, such as principal component
analysis (PCA), to yield a scatter plot visualization of the cases on a two-
dimensional space. After a visualization with an MDS users usually have to
define the labels of the axes by interpreting the collective meanings of the
distantly-plotted clusters on an MDS. Therefore, a dimensionality reduction is
a methodologically different process from the CS, which we attempt to define
in this research.
### 2.2 Compound adjective phrase
While our goal is to address product aesthetics and semantics that evoke end
users’ emotional feelings attached to them, we focused on adjectives that
describe the emotional aspects of product designs. Hsiao and Chen [24]
conducted surveys using product images and adjectives on semantic
differentials. The factor analyses identified four main dimensions of the
participants’ emotional responses to product shapes, characterized by the
pairs of polar affective adjectives. They demonstrated the role of adjectives
in product designs, yet their focus was not to explore a variety of nuanced
adjectives to describe a unique design concept.
In order to describe distinct quality of a design aesthetics and semantics, we
chose to use adjective-adjective combinations. This is called a compound
adjective in a coordinated sequence - a sequence in which two or more
adjectives severally restrict the head noun (e.g., a tall thin woman). In this
example, the phrase therefore means a “woman who is both tall and thin.” The
criteria for being coordinated and in a sequence is that the adjectives in a
coordinated sequence may, but need not, be linked by “and.” Also, a change in
the order of adjectives in a coordinated sequence may result in a shifting of
the emphasis, or even in a stylistic flaw, but will not change the meaning of
the phrase [25]. In addition, the second adjective takes the noun form (e.g.,
“kinetic warm”$\rightarrow$“kinetic warmth,” “effortless
elegant”$\rightarrow$“effortless elegance,” “fluid
beautiful”$\rightarrow$“fluid beauty”) in order for the phrases to sound
complete by themselves. In the case of “kinetic warm(th),” the complete phrase
would be “kinetic and warm style” or “kinetic and warm car”; however, since
the head nouns “style” or “car” are obvious in the context, they are omitted.
Therefore, the type of phrase we use in this research can be defined as a
compound adjective (adjective + adjective) phrase in a coordinated sequence,
which is modified to an adjective-noun form. This type of compound phrase is
underexplored in the research topic called “combinational creativity.”
Combinational creativity attempts to produce new ideas by combining existing
ideas in unfamiliar ways [26]. A number of studies have identified the effect
of combinational creativity as a tool for generating creative design concepts
[27, 28, 29]. Nagai et al. [30], in a collaboration with designers, suggested
that a new concept can be created by synthesizing two ideas in noun-noun
combination phrases. Chiu and Shu [29] used oppositely and similarly related
pairs of words as stimuli for design-concept generation and observed that
oppositely related verb-verb combination stimuli could increase concept
creativity. The Combinator [31] imitates the way the human brain achieves
combinational creativity and suggests combined ideas in both visual and
textual representations. In these works, combinational creativity focused
either on combined product categories or objects derived from noun-noun
combinations (e.g., “desk + elevator” or “pen + ruler”) or functional features
derived from verb-verb combinations (e.g., “fill + insert”), and they did not
address product aesthetics.
### 2.3 Lexico-semantic exploration tool
In constructing a CS in our system, we use the ConceptNet knowledge graph [32]
for exploring compound phrases. ConceptNet is a knowledge graph that connects
words and phrases with labeled edges. It is designed to represent varieties of
general knowledge to allow users to make word associations and make sense of
the meanings behind them. Its knowledge is collected from a wide range of
sources, such as expert-created resources, crowd-sourcing, and games with a
purpose. Among creativity support tools, Spinneret [33] uses ConceptNet and
provides idea suggestions as nodes on a graph. It aims to support divergent
thinking, and it explores ideas and provides “suggestions” to the user to add
to a mind map. Mini-Map [34] also uses ConceptNet, and it gamifies
collaborations with an intelligent agent in creating a mind map. The both
studies chose the ConceptNet as to be suitable for enhancing explorations and
mitigating fixation during an ideation process.
### 2.4 Metrics of creativity support tools
Unlike productivity support tools that can be evaluated on the basis of
objective metrics, the evaluation of creativity support tools lacks obvious
metrics, and there is no one-size-fits-all approach [35]. In one of the
several dimensions of creativity that Sternberg [36] discussed, creativity can
be a property of people, a property of products, and a property of a set of
cognitive processes. With respect to a property of products, “originality” and
“effectiveness” are popular criteria [37, 38, 39]. Both terms have versions of
labels, such as “novelty,” “unusuality,” or “uniqueness” for the former and
“usefulness,” “practicality,” “appropriateness,” or “relevancy” for the latter
[40, 41, 42]. They are measured either on the basis of self-report evaluation
or rated by experts with the precaution of the inter-rater disagreement. With
respect to evaluating a property of people, it tends to lead to a concern with
individual differences between people.
In this research, we used the Creativity Support Index (CSI)[43], which covers
the two properties of creativity above, that are, a property of products, and
a property of a set of cognitive processes, both in terms of users’ own
perceptions. This would be suitable for the present research, especially when
the objective is to cater a tool, which is designed with certain functional
requirements in mind, to support the effectiveness of designers’ cognitive
processes. We will discuss the further details of the CSI, its benefits and
application in section 4.3.
## 3 Method
### 3.1 Preliminary study for adjective-adjective compound phrases
As discussed in section 2.2, we intend to use compound adjective phrases for
the DCP. Despite the fact that some related works demonstrated the role of
adjectives in conveying users’ emotional responses to product aesthetics[24],
or the effectiveness of combinational creativity[24], this particular type of
combinational phrase is under-researched in the context of supporting
designers’ creativity, there is no prior work that demonstrated the
effectiveness of adjective-adjective compound phrase compare to different
types with other parts-of-speech(PoS), such as noun-noun or verb-verb. Thus,
we conducted a preliminary experiment comparing phrases with different PoS to
confirm that adjective phrase combinations are suitable for communicating
product aesthetics.
We used Survey Monkey to recruit 55 participants whose job function is “arts
and design.” Six sample phrases were randomly generated for each of four
different combinations of PoS: adjective-adjective (AA), adjective-noun (AN),
noun-noun (NN), and verb-verb (VV). For the selection of these words, we
randomly extracted the words for all combinations from a corpus, which we
created with English automotive design articles containing about 1.4 million
words on Sketch Engine [44]. The following filtration was performed equally
for all PoS combinations. First, we removed non-English terms and proper nouns
and then removed words whose relative frequency per a million tokens was less
than 1.0, compared with enTenTen15 [45], a large web text corpus containing
about 1.6 billion words. This made the list of all words, regardless of PoS,
more suitable for designing concept expressions that participants can readily
recognize and make a judgment about. The participants rated 24 randomly
generated two-word combinational phrases in randomized order on a 7-point
Likert scale on the basis of the degree to which they agreed with the
statement “I can imagine the product aesthetics, characters, mood, and
emotional quality that are conveyed by the product design.” An ANOVA-Tukey
test identified homogeneous subsets as [AA/AN/NN] and [VV/NN], where the mean
score of the former was significantly higher $(p=0.05^{*})$. Therefore, using
adjective-“noun-form-of” adjective phrases, as opposed to using noun-noun or
verb-verb phrases, would yield optimized effect on addressing product
aesthetics and semantics, as well as emotional qualities attached to them we
are trying to accomplish in this research.
Further details on the individual comparisons are as follows. There was a
significant difference $(p=.038^{*})$ in mean scores between AA
$(3.890,\sigma=1.121)$ and VV $(3.300,\sigma=1.205)$, a significant difference
$(p=.040^{*})$ between AN $(3.885,\sigma=1.025)$ and VV, no significant
difference $(p=.581)$ between AA and NN $(3.612,\sigma=1.232)$, no significant
difference $(p=.599)$ between AN and NN, no significant difference $(p=.485)$
between NN and VV, and no significant difference $(p=1.000)$ between AA and
AN.
### 3.2 System overview
Figure 2: A Process forming and explaining a DCP (A), which is modeled as
constructing a CS (B) Figure 3: An example of a moodboard, which is inspired
by a DCP “flavorous tranquility”
The present system, Character Space Construction (CSC), is an AI-assisted
interactive system. A new contribution of the CSC is that it helps designers
to construct a 2-dimensional CS as a quadrant system, which yields a DCP, a
compound adjective phrase, that portrays a design concept and explains it in
semantic comparisons. The CSC is a web application that has all the functions
above in one system and is operated in an interactive manner. The algorithm of
the system, which constructs a CS by letting the user set four words
($w_{1}$-$w_{4}$), will be discussed in detail in section 3.3. The CSC
application consists of a front-end web application written in JavaScript,
HTML, and CSS and a back-end web server written in Python with a MySQL
database hosted on PythonAnywhere, which is a hosting service for Python web
applications. The MySQL database was built on PythonAnywhere and holds the
data of word embeddings from the ConceptNet Numberbatch [32].
### 3.3 Character Space Construction algorithm
As discussed in Section 2.1., some concept designers empirically use a CS as a
synthesis tool, not as an analytical tool, in a process of forming a DCP. On a
quadrant system, two terms in the upper right quadrant are identified first
then the opposite term on each vertical and horizontal axis are identified.
This unique process has not been formalized in the past to the best of our
knowledge. Fig. 2(A) illustrates the sequence of designers’ cognitive
activities. It starts with searching and selecting a few candidates of the
first words of a compound adjective, followed by searching and identifying the
second adjective to be combined with the candidates of word 1. Once word 1 and
word 2 are identified as a DCP, designers then attempt to find contrasting
words for each to explain the DCP in comparison to those terms, with or
without CS. Fig. 2(B) shows the process model of constructing a CS in
sequence, replicating the process in Fig. 2(A) in a structured manner. A
formalization of constructing a CS not only clarifies the designers’ process
of verbal design concept explorations, but also signifies the next phase of
creative activities, that is, creating a visual mood board. Although creating
a mood board is outside of this research’s scope, it should be noted that this
quadrant system on a CS will eventually help them explore the visual images
for a target design concept, followed by finding visual images that should be
avoided. Fig. 3 is an example of what a mood board would look like with a DCP,
“flavorous tranquility.”222All the images in Fig. 4 are licensed under
Creative Commons in either of the following conditions: CC PDM 1.0, CC BY-NC-
ND 2.0, CC BY 2.0, or CC BY-SA 2.0
Adhering to this model, the CSC tool consists of three steps in a UI (Fig. 4).
The top section is step 1, where users can input a design brief in sentence
form to start searching for word 1 $w_{1}$. A design brief is a written
description of a project that requires some form of design, containing a
project overview, its objectives, tasks, target audience, and expected
outcomes [46, 47]. In design competitions, design briefs are usually
explicitly written material that calls for designers’ entries. In industry
practice, design briefs may be implicitly formed through discussions between
stakeholders in a project. In the top right section above step 2, the system
provides a search window in which the participant can input queries in case
they did not find any word that they wanted to use in the Explorer, after
candidate words are displayed. The lower left space is allocated for
“Explorer,” in which ranked search results are shown either as $w_{1}$ or
combinations of $w_{1}$ and $w_{2}$, depending on the phase of the
exploration.
The middle right section is step 2, where users can choose and pool candidates
for $w_{1}$ and search for candidates for $w_{2}$ that are combined with the
$w_{1}$ candidates. The users at this point will see a variety of adjective
phrases ($w_{1}$-$w_{2}$) in the Explorer as ranked lists in columns, grouped
by $w_{1}$.
Figure 4: Web application user interface of the Character Space Construction
(CSC). The top sections are the input windows for “Design Brief” and
individual query words. The largest section is allocated to “Explorer”, where
candidates for word 1 and word1 - word 2 compound phrases are presented. On
the right sidebar, the “Candidates for Word 1” section shows the words a user
selected and the “Search for Word 2” button. “Character Space” shows a
quadrant system as a user chooses word 1 and word 2, that trigger to show the
candidates of Word 3 and Word 4 as opposing concepts.
Finally, the lower right section is the CS, which is a quadrant system with
all four words set at the ends of each axis. The first quadrant of the CS is
set by dragging and dropping an adjective phrase chosen by the user from the
lists in the Explorer. Once the user finalizes the first quadrant of the CS
with adjective phrases ($w_{1}$-$w_{2}$) and clicks the “Search for word 3/4”
button, the system will search for antonyms to $w_{1}$ and $w_{2}$, as $w_{3}$
and $w_{4}$, respectively. The candidates for $w_{3}$ and $w_{4}$ are shown in
a pull-down menu at the end of each axis so that the user can choose a
suitable contrasting word for each $w_{1}$ and $w_{2}$ to complete the CS. The
upper right quadrant, represented by the combination of $w_{1}$ and $w_{2}$,
is the target design-concept phrase. All the other quadrants, represented by
$w_{2}$-$w_{3}$, $w_{3}$-$w_{4}$, and $w_{4}$-$w_{1}$, are contrasting
concepts to be used by the participants to explain the target design concepts
in comparison with opposing concepts. After clicking on the “Finish” button,
the system generates an explanation of the DCP as a template matched sentence,
like “My design concept is $w_{1}$-$w_{2}$. It has a sense of $w_{2}$ yet is
$w_{1}$, not $w_{3}$. It is $w_{1}$ but not $w_{4}$. In this design, $w_{1}$
and $w_{2}$ can go together.” Algorithm 1 illustrates the procedure described
above. Table 1 shows brief explanations of each sub-function in Algorithm 1.
Algorithm 1 Character Space Construction
function construct_character_space
$\mathit{query\\_words}$:= GenerateQuery(design_brief)
$w_{1}\\_initial\\_\mathit{candidates}$:=[]
$\mathit{phrases}$:=[]
$phrase\\_scores$:=[]
for each $w$ in $\mathit{query\\_words}$ do
$w_{1}\\_\mathit{initial\\_candidates}$.append(SearchCandidates($w$))
sort $w_{1}\\_\mathit{initial\\_candidates}$ by CalculateWordScore($w_{1}$)
display $w_{1}\\_\mathit{initial\\_candidates}$
$w_{1}\\_\mathit{candidates}$:= let user set 5 words from
$w_{1}\\_\mathit{initial\\_candidates}$
end for
for each $w_{1}$ in $w_{1}\\_\mathit{candidate}$ do
$w_{2}\\_\mathit{candidates}$:= SearchCandidates($w_{1}$)
for each $w_{2}$ in $w_{2}\\_\mathit{candidates}(w_{1})$ do
if CalculateWordScore($w_{2})>=1.7$ then
$\mathit{phrases}$.append($w_{1},w_{2}$)
$\mathit{phrase\\_scores}$.append(CalculatePhraseScore($w_{1},w_{2}$))
end if
end for
end for
sort $\mathit{phrases}$($w_{1}$, $w_{2}$) by $phrase\\_scores$
display $\mathit{phrases}$
($w_{1},w_{2}$):= let user set a phrase from $\mathit{phrases}$
display CharacterSpace($w_{1},w_{2}$)
$w_{3}\\_candidates$:= SearchAntomnyms($w_{1}$)
$w_{4}\\_candidates$:= SearchAntomnyms($w_{2}$)
display CharacterSpace($w_{1},w_{2},w_{3},w_{4}$)
display GenerateExplanation($w_{1},w_{2},w_{3},w_{4}$)
end function
Table 1: List of functions in CSC algorithm Function | Description
---|---
GenerateQuery | Extract adjectives from design brief
SearchCandidates | Search related words from ConceptNet
CalculateWordScore | Calculate score of word according to usefulness
CalculatePhraseScore | Calculate score of adjective phrase according to creativity
SearchAntonyms | Search antonyms from ConceptNet
CharacterSpace | Draw a quadrant with set words
GenerateExplanation | Generate an explanation with template matching
For the function CalculateWordScore, the preliminary study gave us an
opportunity to observe a list of randomly chosen adjectives. Despite being
extracted from the domain-specific corpus we created, the list included a
considerable number of unusable adjectives. These unusable adjectives included
ones with negative connotations (e.g., amateurish, costly, detrimental),
temporality (e.g., first, latest, recent, subsequent), distance (e.g., far,
near, close), or utilities (e.g., guest, sourced, takeaway) as well as those
that were too concrete in terms of expressing visuals (e.g., square, yellow,
golden), to name a few.
To rule out such unusable adjectives, we used a binary labeling
(usable/unusable) for a design concept by five professional automotive
designers, who were independent from this research, and we used the count of
usable labels and the word embedding vectors from the ConceptNet Numberbatch
[32] to form a model that predicts the usefulness score of unknown adjectives
using gradient boosting [48], which is a decision-tree-based ensemble machine
learning algorithm. We used a Python software library called XGBoost
(v0.90)[49], which uses a gradient boosting framework. We implemented a linear
regression predictive model that was trained to minimize the prediction errors
(RMSE) in a 10-fold cross validation that yielded the max depth of the
decision tree and the learning rate as $6$ and $0.05$, respectively. The
average RMSE of the prediction with the best hyperparameters on test data was
$1.406$. We used this model to predict usefulness scores for new adjectives,
which were used to display the candidate adjectives for $w_{1}$ in descending
order.
For the function CalculatePhraseScore, we tried to replicate the predictor of
the degree of creativity, which is the threshold of the similarity of two
words, from the related works on combinational creativity [50, 51], which
demonstrated that noun-noun combinations that represent good (award-winning)
design concepts fell into a certain range of distance between two nouns.
However, our test with five professional automotive designers on adjective
phrases in terms of creativity with pre-calculated similarity did not show a
clear normal distribution in the scores to determine a threshold. Yet, some
brackets for similarity showed better or worse scores than the others, so we
assigned these scores to unknown adjective phrases in accordance with the
calculated similarity values for each instance. These phrase scores were used
to display the candidate phrases for $w_{1}$-$w_{2}$ in descending order.
## 4 Experiment Design
### 4.1 Participants and independent variables
The study protocols below have been approved by the Institutional Review Board
of the National Institute of Informatics, Tokyo, Japan on February 15th, 2021
(Approval number 0042).
172 participants, whose job function was arts and design and who were fluent
in English, were recruited via Prolific, an online user test participant
recruiting service. 17 duplicate participants were removed. Of the 155 non-
duplicate participants, 15 ($9.68\%$) withdrew due to system trouble, and 37
($33.55\%$) did not complete the study for unknown reasons. The withdrawal
rate was high due to the complex and time-consuming nature of the task. That
left us with a total of 103 participants (54 M, 49 F) who completed the study,
with a mean age of 28.1 years ($\sigma=8.51$).
We recruited professional designers who would utilize the tool, instead of
recruiting potential end users who would consume the product, because as
discussed in Section 2.4, the present research focuses on an evaluation of a
creative support tool (CST), most commonly defined as tools to make more
people more creative more often [52].
All of the participants who completed the study used either laptops or desktop
computers of their own. The participants who completed the study were paid
US$15. The independent variables differed by whether or not the participants
used the CSC system (Fig. 4). When not using the CSC, the participants were
given complete freedom to use any publicly available web search tools, such as
Google search, an online thesaurus, or Wikipedia. All of the participants were
asked to perform the same task twice with (experiment) and without (control)
the CSC system. The order of the tool they used in two tasks was assigned
randomly in a counterbalanced order. They were given different design briefs
(Table 2) for each task to mitigate a learning effect, however, we deemed
those two design briefs equivalent in terms of influences on the results for
the following rationals. First, both briefs share the the domain of the
products (automotive products), and the basic component of the design briefs:
target audience, the context of the product in use, and expected features.
Second, the textual stimuli in both design briefs are described in the same
abstraction levels. Gonalves et al. [53] reported although different visual
stimuli strongly provoked different outcomes, difference in textual stimuli
did not show significant difference in the outcomes, unless they are
substantially different in specificity of domain, level of abstraction, and
the distance between the baseline and the supplemental briefs.
We did not offer a blank CS for the control participants because using the CS
is a part of our proposed method. The participants in the control group were
given instructions and an example of how to generate an explanation of their
DCPs. They participated in the experiment online using Survey Monkey, an
online platform, to which the participants were redirected from the recruiting
service, Prolific. Each participant was given video instructions explaining
the goal of the task and how to use the CSC tool (experiment group only)
before the task. The CSC tool was given to the experiment group with a URL
link.
Table 2: Two variations of design brief Design brief A: | The owner of this car will be retired, married, a countryside dweller, and a dog owner. The customer’s main requirement is that the car be economical and spacious. The car should be easy to get in and out of and easy to load. It should be functional and smart.
---|---
Design brief B: | Imagine a car for a family with small kids in a suburban community of 2030, where technologies are ubiquitous. Consider three important aspects of the car: harmony with local charm, fun activities, and safety.
### 4.2 Stimuli and tasks
After reading a brief, the participants were prompted to start the task. Those
with the experiment condition (CSC tool) was instructed to start the task by
copying and pasting the design brief to the CSC tool, and as described in
section 3.3, the CSC system provides a search window in which the participant
can input queries in case they did not find any word that they wanted to use
in the Explorer.
For both experiment and control conditions, there were two tasks. The first
task was “to create a unique and original DCP as an adjective-noun phrase form
that will inspire the audience (described in the design brief) to imagine the
character of the design.” “Effortless elegance” was given to the participants
as an example of an adjective phrase in the instruction video. The second task
was to explain their own DCP with contrasting words. As an example, the
adjective “uneasy” and the noun “clumsiness” were presented as contrasting
words for “effortless” and “elegance,” respectively. An example of an
explanation sentence was also given as follows: “My design concept is
‘effortless elegance.’ It has a sense of elegance, yet it is effortless, not
uneasy. It is effortless but not clumsy. In this design, effortless and
elegance can go together.”
For the experiment condition, this explanation was generated automatically
using all $w_{1}$ through $w_{4}$ using template matching as long as they
completed the CSC properly, and the CSC provided a “copy” button that allowed
the participants to copy the automatically generated explanation to the
clipboard so that they could paste it into the survey. However, they were also
given the freedom of editing it in the survey questionnaire. The participants
with the control condition had to generate and write an explanation down in
the survey questionnaire on their own, however, they were also provided
exactly the same example paragraph described above. In the experiment
condition. participants were also given a unique session ID by the CSC system
so that we could match the system log to examine how each participant
interacted with the words and phrases.
### 4.3 Evaluation method
As discussed in section 2.4, we used the CSI [43] as a post-task psychometric
measurement to compare two conditions, with (experiment) or without (control)
the CSC, in terms of creativity support. The CSI enables us to quantify the
cognitive processes of users using psychometric scales for six factors:
Exploration, Expressiveness, Immersion, Enjoyment, Results Worth Effort, and
Collaboration (Table 4). The CSI evaluates a creativity support tool itself,
and focuses on the experience of using it to create rather than trying to
directly evaluate a property of products. Instead, the CSI evaluates the
result of a creation in relation to an effort a user made, such as “I was
satisfied with what I got out of the system or tool.” This is suitable for
tools designed for experienced users, who know what the creative outcomes are
and what the ideal experiences in creation are, as opposed to the tools
designed for novice users where creativity is not sufficient and thus needs
support. The CSI also offers flexibility, such that it can be applied to
various tools and scenarios over time and provides standardized measurement.
It is robustly developed as each of the six factors has two different
statements (Table 3) that improve the statistical power of a survey, which is
deployed in the CSI. In combination with paired factor comparisons for each of
the six factors (Table 4), it also provides insight into what aspects of
creativity support may need attention.
The CSI has a rigorous protocol that the researcher should follow in order for
the measurement and analysis to be universal and reliable. For instance, our
CSC tool is not designed for collaborative tasks; however, the CSI protocol
discourages researchers from skipping the statements in the Collaboration
factor, still allowing participants to rate the collaboration statements.
Instead, the CSI protocol allows adding “N/A” responses to statements that
belong to the collaboration factor, which we incorporated in our survey.
Table 3: The 12 Agreement Statements on the CSI. Each agreement statement is
answered on a scale of “Highly Disagree” (1) to “Highly Agree” (10). In
deployment,the factor names are not shown, and the participant does not see
the statements grouped by factor[43]. Collaboration:
---
1\. The system or tool allowed other people to work with me easily.
2\. It was really easy to share ideas and designs with other people inside
this system or tool.
Enjoyment:
1\. I would be happy to use this system or tool on a regular basis.
2\. I enjoyed using the system or tool.
Exploration:
1\. It was easy for me to explore many different ideas, options, designs,or
outcomes, using this system or tool.
2\. The system or tool was helpful in allowing me to track different ideas,
outcomes, or possibilities.
Expressiveness:
1\. I was able to be very creative while doing the activity inside this system
or tool.
2.The system or tool allowed me to be very expressive.
Immersion:
1\. My attention was fully tuned to the activity, and I forgot about the
system or tool that I was using.
2\. I became so absorbed in the activity that I forgot about the system or
tool that I was using.
Results Worth Effort:
1\. I was satisfied with what I got out of the system or tool.
2\. What I was able to produce was worth the effort I had to exert to produce
it.
Table 4: The Paired-Factor Comparison Test has 15 comparisons for each pair, a
user will choose a factor description in response to the following statement:
“When doing this task, it’s most important that I’m able to…”[43] 1\. Be
creative and expressive.
---
2\. Become immersed in the activity.
3\. Enjoy using the system or tool
4\. Explore many different ideas, outcomes, or possibilities.
5\. Produce results that are worth the effort I put in.
6\. Work with other people.
Figure 5: Equation for scoring the CSI [43]
Also, the final CSI score will be weighted by the scores of paired-factor
comparisons, so if a participant does not think of the collaboration factor as
important, it is reflected in the final score accordingly. The only other
modification we made was to replace “the system or tool” with “The CSC” in
each statement (e.g., “I enjoyed using the CSC.”) that we were able to do so
in. We asked 12 questions on how strongly they agreed or disagreed with each
statement on a 10-point Likert scale after each task was completed and 15
paired-factor comparison questions at the end, emphasizing that they were
comparing the factors that they thought were important for the tasks, not for
the different tools. To check the reliability of each factor, we examined the
similarity of the scores (Table 5) across the two different statements.
### 4.4 Participant profile
Since we designed the CSC with professional designers in mind, we were
interested in examining the CSI ratings depending on the years of experience
and how often the participants performed the particular task we attempted to
support with the CSC, that is, generating DCPs especially in the aesthetic
sense and explaining them to stakeholders. For the former, we bracketed the
years-of-experience groups as: less than 1 year, 1 to 3 years, 3 to 5 years, 5
to 10 years, and 10+ years. For the latter, the groups were divided as: less
than quarterly or never, quarterly, monthly, weekly, and daily.
## 5 Results
### 5.1 Creativity Support Index
Table 5: CSI and factor scores | Cronbach’s Alpha per Factor | Average Factor Count | Factor Score | Weighted Factor Score
---|---|---|---|---
| | | CSC (N=93) | Control (N=95) | CSC (N=93) | Control (N=95) |
| | Mean ($\mathit{\sigma}$) | Mean ($\mathit{\sigma}$) | Mean ($\mathit{\sigma}$) | Mean ($\mathit{\sigma}$) | Mean ($\mathit{\sigma}$) | $p$ (Weighted Factor Score)
Collaboration | $0.829$ | $0.38$ ($0.8$) | $4.96$ ($6.07$) | $5.37$ ($6.53$) | $2.69$ ($8.24$) | $3.06$ ($8.39$) | $0.758$
Enjoyment | $0.894$ | $2.26$ ($1.36$) | $15.48$ ($4.13$) | $14.11$ ($3.88$) | $35.63$ ($24.84$) | $30.8$ ($20.76$) | $0.149$
Exploration | $0.787$ | $3.89$ ($0.88$) | $15.87$ ($3.48$) | $13.35$ ($3.94$) | $61.91$ ($20.6$) | $51.32$ ($18.73$) | $<0.010^{**}$
Expressiveness | $0.843$ | $3.69$ ($1.04$) | $14.45$ ($3.85$) | $13.63$ ($3.25$) | $52.56$ ($19.71$) | $50.61$ ($19.48$) | $0.496$
Immersion | $0.846$ | $1.78$ ($1.07$) | $11.06$ ($4.76$) | $11.37$ ($4.41$) | $19.95$ ($16.12$) | $21.15$ ($15.7$) | $0.605$
Result Worth Effort | $0.811$ | $3.01$ ($1.25$) | $14.66$ ($3.75$) | $14.03$ ($3.11$) | $43.35$ ($20.67$) | $42.03$ ($21.9$) | $0.671$
CSI | | | | | 72.03 (16.67) | 66.32 (14.22) | $0.012^{*}$
Table 5 shows the total CSI scores and factor scores between the experiment
(CSC) group and the control group. The mean CSI score of the experiment group,
which used our proposed CSC ($72.03,\sigma=16.67)$, was significantly higher
($p=.012<.05^{*}$, Cohen’s d $=.369)$ than that of the control group
($66.32,\sigma=14.22$)(Fig. 6, error bars show standard error). Note that 4
participants out of 103 were excluded due to their DCP being invalid because
the compound phrase ($w_{1}$-$w_{2}$) was impossible to interpret, hence
leaving us with 198 cases from 99 participants. Ten cases were excluded as
outliers whose CSI scores were more than two standard deviations away from
both sides of the mean, which left us with 188 cases (93 experiments, 95
control) for the final analysis.
For the reliability of the ratings within each factor, the Cronbach’s alpha
for Exploration, Expressiveness, Immersion, Enjoyment, Results Worth Effort,
and Collaboration was $.787,.843,.846,.894,.811$, and $.829$, respectively. As
for the factors of the CSC the participants found important, we compared the
mean factor counts from paired factor comparisons, which were
$3.8876(\sigma=0.88)$, $3.6854(\sigma=1.04)$, $1.7753(\sigma=1.07)$, $2.2584$
($\sigma=1.36)$, $3.0112(\sigma=1.25)$, and $.3820(\sigma=0.80)$,
respectively.
Weighted factor scores were calculated by multiplying a participant’s factor
agreement rating on the statement that belongs to the factor by the factor
count (Fig.5) [43]. This was done in order to make the weighted factor score
more sensitive to the factors that were the most important to the given task.
The weighted factor score of Exploration for the experiment group ($61.91$,
$\sigma=20.6$) was significantly higher ($p<.001^{**}$, Cohen’s d $=.679)$
than that of the control group ($51.32,\sigma=18.73$). The other weighted
factor scores did not indicate a significant difference between the experiment
and control groups. As for the perceived importance of the factors by the
participants, the mean count of the factors for Exploration [$3.89$
($\sigma=1.04$)] was among the highest, and that of Collaboration [$0.38$
($\sigma=0.8$)] was among the lowest.
A partly-repeated two-way ANOVA was performed in order to check the order
effects. The within-participants factor is the tool they used (CSC or
control), and the between-participants factor is the order they used the tool
(CSC first or control first). We report a significant order effect
($p=.007^{**},F=7.70$) between CSC-first or control-first groups. As for the
within-participants test, there were no significant main effect or significant
interaction between the tool and the order.
### 5.2 CSI score distribution by participants profile
Fig. 7 shows the distribution of CSI score between the CSI(experiment) group
and control group by years of experience brackets in “arts and design”
profession. While the years of experience was asked as a categorical choice we
examined a Pearson rank correlation between the CSI score and the median rank
in each bracket with respect to the years of professional experiences for each
group. Weak correlations were observed between the CSI score and the years of
experience for both CSC ($r=.235,p=0.023^{*}$ and the
Control($r=.237,p=0.021^{*}$ groups. This indicates that the longer the
experience, the more they value the benefit of the tool they used, which is
consistent in both CSC and control groups. The mean CSI scores of CSC group
for: less than 1 year, 1 to 3 years, 3 to 5 years, 5 to 10 years, and 10+
years were $65.27$, $69.9$0, $69.61$, $76.41$, and $76.87$, respectively. The
mean CSI scores of the control group for the same years-of-experience brackets
were $61.54$, $63.01$, $65.54$, $69.62$, $71.98$, and $66.32$, respectively.
As for how often the participants performed the particular task, we did not
observe any such tendencies in the CSI score by the frequency of the DCP
creation task they engaged in. Table 6 shows the illustrative examples of
randomly selected DCPs generated by the participants per tools and design
briefs.
Figure 6: CSI scores between experiment (CSC) group and Control group Figure 7: CSI score distribution between groups by years of experience bracket in “arts & design” profession Table 6: Examples of DCPs generated by participants per tools Tool | Des. Brief | $w_{1}+w_{2}$ | Explanation
---|---|---|---
CSC | A | efficient intelligence | My design concept is efficient-intelligence. It has a sense of intelligence yet it is efficient, not incompetent. It is efficient, but not ignorance. In this design, efficient and intelligence can go together.
CSC | B | harmonious interaction | My design concept is harmonious interaction. It has a sense of interactivity yet it is harmonious, not incongruous. It is harmonious, but not incompatibility. In this design, harmony and interaction can go together.
Control | B | practical smart | The design is practical and smart, yet is not non-rational. It is practical, not useless.
Control | A | comfortable utility | This design focuses on comfortable utility, it foregoes the roughness and harshness of typical utility vehicles, yet also avoids the impracticality and compromises of vehicles that put aesthetics over function.
## 6 Discussion
### 6.1 Contributions and applications
The present research showed that our proposed system, CSC, demonstrated its
advantage for the goal we aimed to achieve, which is to support the creativity
of professional designers. Our research is motivated by an empirical
observation on the characteristics of the concept designer task, that is,
exploring, identifying, and communicating a design concept, especially with
respect to product aesthetics and semantics, as opposed to product features
and functionalities.
We took the example of the automotive industry to describe a case study
because the particular characteristics of the design conceptualization task
are most apparent in automotive design studios. Yet, these characteristics are
common in industrial design in general, where the skills of industrial
designers are highly specialized and focused. As Tovey [7] discussed , the
tasks of industrial designers are distinguished from those of their
engineering colleagues, and they are specialized in determining the appearance
and identity of a product. However, these practices are traditionally highly
empirical, informal, undocumented, and sometimes even secretive due to the
strict confidentiality of the industry. It also tends to be dependent upon the
individual experiences of designers. We modeled this process as generating a
compound adjective phrase that best conceptualizes the nature of the design
aesthetics and semantics and then constructing a CS to explain them in
contrast with an opposing or distinct concept in a quadrant system (Fig.2
(B)). With respect to adjective phrases, as discussed in section 1, these
adjective phrases can also be used in design practices on other industries
[15, 16, 17], in computer poetry [54], or even in the advertisements in
tourism industry [17] where the verbal expressions of aesthetics and the
semantics of the products matter. Therefore, the proposed methods can be
applied to various design and new product development fields, and are likely
to be optimized by introducing domain-specialized corpora for the function
CalculateWordScore and the function CalculatePhraseScore. In regards to the
quadrant system, there has been a vague distinction between analysis using a
dimensionality reduction on multidimensional scaling (MDS) visualization
technique and what designers use as a synthesis tool. We defined the latter as
the CS, translated it into a computational algorithm, and implemented it in an
interactive web application so that designers can effectively trace the
process in an orderly sequence. Designers know what the important factors of
this task are, and evidently, we confirmed that designers value the
Exploration factor most in this task.
In this research, we shed light on concept designers’ cognitive activities
(Fig. 2-(A)), that were previously indefinite, formalized them (Fig. 2-(B)),
translated them into the CSC algorithm (Algorithm 1), and created the
interactive user interface (Fig. 3). The theoretical contribution of the
present research is based on a formalization of a designer’s concept
exploration activities, which involves searching and selecting a compound
adjective as a DCP. This formalization of constructing a CS not only signifies
the designers’ process of verbal design concept explorations but also leads to
subsequent visual creations (Fig.3).
As for a practical implication of the present research, experienced designers
seem to have valued the structured process of forming and explaining the DCPs
in the experiment. As some of the Weighted Factor Scores of CSI, such as
Enjoyment and Immersion, did not seem to matter to them, the participants
probably appraised the CSC as to offer them a certain efficiency and
effectiveness as a tool for professional use. This factor evaluation result,
with less emphasis on Enjoyment and Immersion, came in with a bit of surprise
to us, however, it actually gave us an assurance that we should spend more
resources to improve more Exploration factors, which we put many efforts into
this research by offering a variety of words and phrases presented on the
Explorer, enabled by the ConceptNet knowledge graph. Also, while a new tool
often involves a novelty effect [55], we can claim that the CSC did not
introduce a notable novelty effect, as it did not show a significant
difference in Enjoyment and Immersion compared with the control condition,
where participants used generic tools.
Regarding the level of expertise of the target audience, about half of
creativity support tools (CSTs) have no clear target level of expertise,
followed by experts (33%) and novices (17%) according to Frich et al. [56].
Our proposed method is clearly targeted for the tasks of experienced designers
since we observed the need for support in a particular task. In that sense, as
discussed in Fig. 7, though consistent between two groups, there was a
tendency indicating that the more experienced the participants were, the more
aware they were of the importance of the task, which suggests the necessity or
usefulness of this type of creativity support tool. Meanwhile, it can be also
said that, if the CSC were successfully emulating the process that an expert
designer would use, we should have observed a bottom-up effect for the
designers with fewer experiences. In the future, we may need to promote a
method to raise the awareness of process-oriented practices for less
experienced designers.
As reported in section 5.1, we observed an order effect between the
participants who used CSC first or second. The CSI score for the control tool
rated by the participants who used the CSC first substantially dropped.
Meanwhile, the CSI score for the CSC tool rated by the participants who used
the control tool first practically stayed the same. While this may have been
cased by a “fatigue effect” on repeated task [57] as the tasks given to the
participants were fairly complex and time consuming, it is probably a positive
sign to see the CSI score which is maintained the same level when used in the
second task.
### 6.2 Limitation
As the focus of this research was to identify and implement the design concept
generation process used by designers, that is, creating DCPs and communicating
them, into an algorithm for constructing the CS, there are some limitations on
the CSC tool. First, we did not put much emphasis on generating the “good”
results achieved by some of the related works mentioned in section 2.3., which
would have been reflected in the Result Worth Effort factor or Expressiveness
of the CSI. For example, the function, CalculateWordScore and the function
CalculatePhraseScore could be iterated more in optimizing the algorithms. In
fact, how to optimize suggested words and phrases can be another research
question, while the features and characteristics of adjectives can be
identified and utilized in order to suggest words and phrases that users can
find more creative. In the course of the research we have examined several
potential predictors, such as adjective supersense [58], and psycholinguistic
attributes of words [59], however, none appeared promising. There should be
more potential predictors we can investigate in the future. Second, the CSC
tool focuses on verbal concept design activities and does not involve visual
activities at this time. Although the verbal activity can be a starting point
to conceptualize and manipulate the product aesthetics and semantics,
designers will engage in visual ideation activities that can be also
computationally supported. We would like to implement this visual process into
the system in the future, that is, creating a mood board as we previewed an
example (Fig. 4), into the CS.
As for the limitation of our experiment, the participants were not given the
identical design brief for each condition, however, we carefully composed both
design briefs in consistent manner described in the section 4.1., and deemed
those two design briefs equivalent in terms of influences on the results. Our
intention of the experiment was to compare different tools, not to compare the
different stimuli though it may leave a room for discussions. Some may argue
it might have been better to set up a between-participants arrangement, yet,
we had to take a within-participants arrangement for two reasons: first, we
were still trying to test the relative effectiveness between two conditions -
the proposed method and what they usually practice in their regular works.
Second, recruiting a sufficient body of participants who are in “arts and
design” profession for a between-participants arrangement was prohibitively
difficult.
We did not ask the participants with control conditions what web search
tool(s) they actually used in creating and explaining their DCPs, in order to
avoid overloading the participants with additional questions. Also, those
publicly available web search tools do not provide the log data of the
participants’ behaviors during the search. Thus, how many and what queries the
participants tried, or how many words the participant actually looked at
before deciding $w1$ or $w1-w2$ combinations, for example, is unknown and we
were unable to compare such interactions between the conditions.
Another factor that may affect the CSI results is the graphic user interface.
Our proposed CSC tool did not suggest novelty effect indicated by rather
humble Enjoyment and Immersion scores. However, it can also be said that there
is a room to introduce a novelty bias if we adopt a more entertaining graphic
user interface. In the future, we would also like to compare different graphic
user interfaces to examine how Enjoyment and Immersion factors would interact
with Exploration, Result Worth Effort, or overall CSI scores.
Lastly, despite the fact that the CSI is most commonly used to evaluate
creative support tools, we acknowledge that it does not evaluate the
“effectiveness” of the tool in terms of how strongly the potential end-users
are influenced by the generated design concepts. As there is no one-size-fits-
all approach [35] we may also need to develop our original tool that
appropriately measure the effectiveness of the creation while we implement the
visual component discussed earlier in this section.
## 7 Conclusion
We focused on the process of verbal design concept exploration practiced by
concept designers, that is, generating and verbally communicating product
aesthetics and semantics, and we attempted to implement them into an
algorithm. We created a system, Character Space Construction (CSC), that
enables concept designers to explore and choose words to create a compound
adjective phrase, which we call the Design Concept Phrase (DCP). This is a
part of the process of constructing a quadrant system, which we call the
Character Space (CS). In a CS, a DCP can be represented in the upper right
quadrant, whereas other quadrants can be described as what to stay away from
or avoid in a semantic space. Our experiment, which was done with 103
participants using a within-participants arrangement, compared our proposed
methods, CSC and the general process designers use, and it was shown that our
proposed method, the CSC, has a significantly positive effect using Creativity
Support Index (CSI).
## 8 list of non-standard abbreviations
* •
CSC: Character Space Construction
* •
CS: Character Space
* •
DCP: design concept phrase
* •
CSI: Creativity Support Index
* •
MDS: Multidimensional scaling
## 9 Acknowledgments
We appreciate the support of the automotive designers at an anonymous company
for cooperating with the preliminary studies on labeling adjectives and
adjective phrases. This work was partially supported by JST, CREST
(JPMJCR21D4), Japan.
## 10 Data Availability Statement
The following data, supporting the conclusions of this article, will be made
available by the authors, without undue reservation.
* •
Participants responses to the experiment.
* •
Expert labeling on adjectives and adjective phrases.
* •
The generated design corpus from which we randomly selected the words and
phrases for PoS comparisons
* •
Participants responses to the PoS ratings on the preliminary study
* •
The source code of the CSC system
The following datasets presented in this study can be found in online
repositories. The names of the repositories can be found in the article.
* •
The ConceptNet API
* •
The word embedding vectors of the ConceptNet Numberbatch
* •
EnTenTen15 English web corpus
## 11 Conflict of Interest Statement
Author Shin Sano was employed by Toyota Motor Corporation and its North
American subsidiary, Calty Design Research, Inc. between April 1992 and July
2007. Author Shin Sano was hired as a design consultant between January 2015
and July 2020 for the Concept-i project, which is introduced as an industry
case study in the introduction section of this paper. This research is
independent of any part of that project and the authors received no funding
for the cost of this research. The remaining author declared that the research
was conducted in the absence of any commercial or financial relationships that
could be construed as a potential conflict of interest.
## References
* Krippendorff [2005] Klaus Krippendorff. _The semantic turn: A new foundation for design_. crc Press, 2005.
* Bloch et al. [2003] Peter H Bloch, Frederic F Brunel, and Todd J Arnold. Individual differences in the centrality of visual product aesthetics: Concept and measurement. _Journal of consumer research_ , 29(4):551–565, 2003.
* Alcaide-Marzal et al. [2020] Jorge Alcaide-Marzal, Jose Antonio Diego-Mas, and Gonzalo Acosta-Zazueta. A 3D shape generative method for aesthetic product design. _Design Studies_ , 66:144–176, 2020.
* Han et al. [2021] Ji Han, Hannah Forbes, and Dirk Schaefer. An exploration of how creativity, functionality, and aesthetics are related in design. _Research in Engineering Design_ , pages 1–19, 2021.
* Demirbilek and Sener [2003] Oya Demirbilek and Bahar Sener. Product design, semantics and emotional response. _Ergonomics_ , 46(13-14):1346–1360, 2003.
* Dong [2005] Andy Dong. The latent semantic approach to studying design team communication. _Design studies_ , 26(5):445–461, 2005.
* Tovey [1992] Michael Tovey. Intuitive and objective processes in automotive design. _Design Studies_ , 13(1):23–41, 1992.
* Desmet and Hekkert [2002] Pieter MA Desmet and Paul Hekkert. The basis of product emotions. _Pleasure with products, beyond usability_ , pages 60–68, 2002.
* McDonagh et al. [2002] Deana McDonagh, Anne Bruseberg, and Cheryl Haslam. Visual product evaluation: exploring users’ emotional relationships with products. _Applied Ergonomics_ , 33(3):231–240, 2002.
* Norman [2004] Donald A Norman. _Emotional design: Why we love (or hate) everyday things_. Basic Civitas Books, 2004.
* Hekkert [2006] Paul Hekkert. Design aesthetics: principles of pleasure in design. _Psychology science_ , 48(2):157, 2006.
* Helander et al. [2013] Martin G Helander, Halimahtun M Khalid, Tek Yong Lim, Hong Peng, and Xi Yang. Emotional needs of car buyers and emotional intent of car designers. _Theoretical Issues in Ergonomics Science_ , 14(5):455–474, 2013.
* Bouchard et al. [2005] Carole Bouchard, Roger Camous, and Ameziane Aoussat. Nature and role of intermediate representations (ir) in the design process: Case studies in car design. _International journal of vehicle design_ , 38(1):1–25, 2005.
* Shieh et al. [2011] Meng Dar Shieh, Tsung Hsing Wang, and Chih Chieh Yang. Affective response dimension selection for product design: A comparison of cluster analysis and procrustes analysis. _International Journal of Digital Content Technology and its Applications_ , 5(1):305–318, 2011.
* Choo and Kim [2003] Sunhyung Choo and Youngin Kim. Effect of color on fashion fabric image. _Color Research & Application: Endorsed by Inter-Society Color Council, The Colour Group (Great Britain), Canadian Society for Color, Color Science Association of Japan, Dutch Society for the Study of Color, The Swedish Colour Centre Foundation, Colour Society of Australia, Centre Français de la Couleur_, 28(3):221–226, 2003.
* Khalaj and Pedgley [2014] Javad Khalaj and Owain Pedgley. Comparison of semantic intent and realization in product design: A study on high-end furniture impressions. _International Journal of Design_ , 8(3), 2014.
* Durán-Muñoz [2019] Isabel Durán-Muñoz. Adjectives and their keyness: a corpus-based analysis of tourism discourse in english. _Corpora_ , 14(3):351–378, 2019.
* Nagai and Noguchi [2003] Y Nagai and H Noguchi. An experimental study on the design thinking process started from difficult keywords: Modeling the thinking process of creative design. _Journal of engineering design_ , 14(4):429–437, 2003.
* Osgood et al. [1957] Charles Egerton Osgood, George J Suci, and Percy H Tannenbaum. _The measurement of meaning_. Number 47. University of Illinois press, 1957.
* Ramsay [1975] James O Ramsay. Solving implicit equations in psychometric data analysis. _Psychometrika_ , 40(3):337–360, 1975.
* Zimet et al. [1988] Gregory D Zimet, Nancy W Dahlem, Sara G Zimet, and Gordon K Farley. The multidimensional scale of perceived social support. _Journal of personality assessment_ , 52(1):30–41, 1988.
* Trochim [1989] William MK Trochim. An introduction to concept mapping for planning and evaluation. _Evaluation and program planning_ , 12(1):1–16, 1989.
* Nagamachi [1995] Mitsuo Nagamachi. Kansei engineering: a new ergonomic consumer-oriented technology for product development. _International Journal of industrial ergonomics_ , 15(1):3–11, 1995.
* Hsiao and Chen [2006] Kun-An Hsiao and Lin-Lin Chen. Fundamental dimensions of affective responses to product shapes. _International Journal of Industrial Ergonomics_ , 36(6):553–564, 2006.
* Sopher [1962] H Sopher. Sequence of adjectives. _ELT Journal_ , 16(4):192–198, 1962.
* Boden [2001] Margaret Boden. Creativity and knowledge. _Creativity in education_ , pages 95–102, 2001.
* Han et al. [2020] Ji Han, Min Hua, Dongmyung Park, Pan Wang, and PRN Childs. Computational conceptual distances in combinational creativity. In _Proceedings of the Design Society: DESIGN Conference_ , volume 1, pages 177–186. Cambridge University Press, 2020.
* Georgiev et al. [2010] Georgi V. Georgiev, Yukari Nagai, and Toshiharu Taura. A method for the evaluation of meaning structures and its application in conceptual design. _Journal of Design Research_ , 8(3):214–234, 2010\. ISSN 15691551. doi: 10.1504/JDR.2010.032607.
* Chiu and Shu [2012] I. Chiu and L. H. Shu. Investigating effects of oppositely related semantic stimuli on design concept creativity. _Journal of Engineering Design_ , 23(4):271–296, 2012. ISSN 09544828. doi: 10.1080/09544828.2011.603298.
* Nagai et al. [2009] Yukari Nagai, Toshiharu Taura, and Futoshi Mukai. Concept blending and dissimilarity: factors for creative concept generation process. _Design Studies_ , 30(6):648–675, 2009.
* Han et al. [2016] J Han, Feng Shi, PRN Childs, et al. The combinator: A computer-based tool for idea generation. In _DS 84: Proceedings of the DESIGN 2016 14th International Design Conference_ , pages 639–648, 2016.
* Speer et al. [2017] Robyn Speer, Joshua Chin, and Catherine Havasi. ConceptNet 5.5: An open multilingual graph of general knowledge. In _Thirty-first AAAI conference on artificial intelligence_ , 2017\.
* Bae et al. [2020] Suyun Sandra Bae, Oh-Hyun Kwon, Senthil Chandrasegaran, and Kwan-Liu Ma. Spinneret: Aiding creative ideation through non-obvious concept associations. In _Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems_ , pages 1–13, 2020.
* Chen et al. [2019] Ting-Ju Chen, Sai Ganesh Subramanian, and Vinayak R Krishnamurthy. Mini-map: Mixed-initiative mind-mapping via contextual query expansion. In _AIAA Scitech 2019 Forum_ , page 2347, 2019.
* Hewett et al. [2005] Tom Hewett, Mary Czerwinski, Michael Terry, Jay Nunamaker, Linda Candy, Bill Kules, and Elisabeth Sylvan. Creativity support tool evaluation methods and metrics. _Creativity Support Tools_ , pages 10–24, 2005.
* Sternberg [1999] Robert J Sternberg. _Handbook of creativity_. Cambridge University Press, 1999.
* Barron [1955] Frank Barron. The disposition toward originality. _The Journal of Abnormal and Social Psychology_ , 51(3):478, 1955.
* Stein [1953] Morris I Stein. Creativity and culture. _The journal of psychology_ , 36(2):311–322, 1953\.
* Runco and Jaeger [2012] Mark A Runco and Garrett J Jaeger. The standard definition of creativity. _Creativity research journal_ , 24(1):92–96, 2012\.
* Bruner [1962] Jerome S Bruner. The conditions of creativity. In _Contemporary Approaches to Creative Thinking, 1958, University of Colorado, CO, US; This paper was presented at the aforementioned symposium._ Atherton Press, 1962.
* Cropley [1967] AJ Cropley. Creativity, intelligence, and achievement. _Alberta Journal of Educational Research_ , 1967.
* Kneller [1965] George Frederick Kneller. _The art and science of creativity_. Holt, Rinehart and Winston, 1965.
* Cherry and Latulipe [2014] Erin Cherry and Celine Latulipe. Quantifying the creativity support of digital tools through the creativity support index. _ACM Transactions on Computer-Human Interaction (TOCHI)_ , 21(4):1–25, 2014.
* Kilgarriff et al. [2004] Adam Kilgarriff, Pavel Rychly, Pavel Smrz, and David Tugwell. Itri-04-08 the sketch engine. _Information Technology_ , 105:116, 2004.
* Jakubíček et al. [2013] Miloš Jakubíček, Adam Kilgarriff, Vojtěch Kovář, Pavel Rychlỳ, and Vít Suchomel. The tenten corpus family. In _7th International Corpus Linguistics Conference CL_ , pages 125–127, 2013.
* Phillips [2004] Peter L Phillips. _Creating the Perfect Design Brief: How to manage design for strategic advantage_. Skyhorse Publishing Inc., 2004.
* Koronis et al. [2018] Georgios Koronis, Arlindo Silva, Jacob Kang, et al. The impact of design briefs on creativity: A study on measuring student designers outcomes. In _DS 92: Proceedings of the DESIGN 2018 15th International Design Conference_ , pages 2461–2472, 2018.
* Friedman [2001] Jerome H Friedman. Greedy function approximation: a gradient boosting machine. _Annals of statistics_ , pages 1189–1232, 2001.
* Chen and Guestrin [2016] Tianqi Chen and Carlos Guestrin. Xgboost: A scalable tree boosting system. In _Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining_ , pages 785–794, 2016.
* Han et al. [2019] Ji Han, Dongmyung Park, Feng Shi, Liuqing Chen, Min Hua, and Peter RN Childs. Three driven approaches to combinational creativity: Problem-, similarity-and inspiration-driven. _Proceedings of the Institution of Mechanical Engineers, Part C: Journal of Mechanical Engineering Science_ , 233(2):373–384, 2019.
* Han et al. [2018] Ji Han, Feng Shi, Dongmyung Park, Liuqing Chen, Peter Childs, et al. The conceptual distances between ideas in combinational creativity. In _DS 92: Proceedings of the DESIGN 2018 15th International Design Conference_ , pages 1857–1866, 2018.
* Shneiderman [2002] Ben Shneiderman. Creativity support tools. _Communications of the ACM_ , 45(10):116–120, 2002.
* Gonalves et al. [2012] M Gonalves, Carlos Cardoso, and Petra Badke-Schaub. Find your inspiration: exploring different levels of abstraction in textual stimuli. In _DS 73-1 Proceedings of the 2nd International Conference on Design Creativity Volume 1_ , 2012.
* Toivanen et al. [2012] Jukka Toivanen, Hannu Toivonen, Alessandro Valitutti, Oskar Gross, et al. Corpus-based generation of content and form in poetry. In _Proceedings of the third international conference on computational creativity_. University College Dublin, 2012.
* Tsay et al. [2020] Crystal Han-Huei Tsay, Alexander K Kofinas, Smita K Trivedi, and Yang Yang. Overcoming the novelty effect in online gamified learning systems: An empirical evaluation of student engagement and performance. _Journal of Computer Assisted Learning_ , 36(2):128–146, 2020.
* Frich et al. [2019] Jonas Frich, Lindsay MacDonald Vermeulen, Christian Remy, Michael Mose Biskjaer, and Peter Dalsgaard. Mapping the landscape of creativity support tools in hci. In _Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems_ , pages 1–18, 2019.
* Bradley and Daly [1994] Mark Bradley and Andrew Daly. Use of the logit scaling approach to test for rank-order and fatigue effects in stated preference data. _Transportation_ , 21(2):167–184, 1994.
* Tsvetkov et al. [2014] Yulia Tsvetkov, Nathan Schneider, Dirk Hovy, Archna Bhatia, Manaal Faruqui, and Chris Dyer. Augmenting english adjective senses with supersenses. 2014\.
* Wilson [1988] Michael Wilson. Mrc psycholinguistic database: Machine-usable dictionary, version 2.00. _Behavior research methods, instruments, & computers_, 20(1):6–10, 1988.
|
# Domain-Agnostic Prior for Transfer Semantic Segmentation
Xinyue Huo1,2 Lingxi Xie2 Hengtong Hu2,3 Wengang Zhou1 Houqiang Li1 Qi Tian2
1University of Science and Technology of China 2Huawei Inc. 3Hefei University
of Technology
<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>
<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Unsupervised domain adaptation (UDA) is an important topic in the computer
vision community. The key difficulty lies in defining a common property
between the source and target domains so that the source-domain features can
align with the target-domain semantics. In this paper, we present a simple and
effective mechanism that regularizes cross-domain representation learning with
a domain-agnostic prior (DAP) that constrains the features extracted from
source and target domains to align with a domain-agnostic space. In practice,
this is easily implemented as an extra loss term that requires a little extra
costs. In the standard evaluation protocol of transferring synthesized data to
real data, we validate the effectiveness of different types of DAP, especially
that borrowed from a text embedding model that shows favorable performance
beyond the state-of-the-art UDA approaches in terms of segmentation accuracy.
Our research reveals that UDA benefits much from better proxies, possibly from
other data modalities.
## 1 Introduction
Figure 1: Top: The goal is to transfer a segmentation model trained in the
source domain to the target domain, but some semantically similar classes (see
the left of (a) for examples of bike and motorbike) are difficult to
distinguish due to domain shift. Bottom: segmentation results of the upper-
right image without and with DAP, where (b) shows incorrect segmentation
(bike$\rightarrow$motorbike and sidewalk$\rightarrow$road) of the baseline
(DACS [49]), and (c) shows how DAP improves segmentation.
In the deep learning era, the most powerful approach for visual recognition is
to train deep neural networks with abundant, labeled data. Such a data-driven
methodology suffers the difficulty of transferring across domains, which
raises an important research field named domain adaptation. This paper focuses
on the setting of unsupervised domain adaptation (UDA), which assumes that the
source domain offers full supervision but the target domain has no annotations
available. In particular, we investigate semantic segmentation – provided the
increasing amount of unlabeled data and the expensiveness of annotation, it
becomes increasingly important to gain the ability of transferring models from
a known domain (e.g., labeled or synthesized data).
The goal of semantic segmentation is to assign each pixel with a class label.
In the scenarios that annotations are absent, this is even more challenging
than image-level prediction (e.g., classification) because the subtle
differences at the pixel level can be easily affected by the change of data
distribution. Note that this factor can bring in the risk that pseudo labels
are inaccurate and thus deteriorate transfer performance, and the risk becomes
even higher when the target domain lacks sufficient data for specific classes
or class pairs. Fig 1 shows an example that such approaches [64, 49] are
difficult in distinguishing semantically similar classes (e.g., motorbike vs.
bike, road vs. sidewalk) from each other.
To alleviate the above issue, we first offer a hypothesis about the confusion
– the proportion of similar categories in the two domains varies too much or
they often appear adjacent to each other and the border is difficult to find (
there are limited pixels around the boundary). Consequently, it is difficult
for the deep network to learn the discrimination boundary based on the
transferred image features. To compensate, we propose to add a domain-agnostic
prior (DAP) to force the features from the source and target domains to align
with an auxiliary space that is individual to both domains. According to the
Bayesian theory, a properly designed prior relieves the instability of
likelihood (provided by the limited training data with similar classes co-
occurring) and leads to more accurate inference.
We implement our algorithm upon DACS, a recent approach built upon an advanced
data augmentation named ClassMix [37]. The training procedure of DACS involves
sampling a pair of source and target images and making use of pseudo labels to
generate a mixed image with partly-pseudo labels, and feeding the source and
mixed images into the deep network. We introduce DAP into the framework by
defining a high-dimensional, domain-agnostic embedding vector for each class,
and force the features extracted from both the source and mixed images to
align with the embedding vectors through an auxiliary module. We investigate
two types of domain-agnostic embedding, namely, one-hot vectors and the
word2vec features [35], and show that both of them bring consistent gain in
transferring. As a side note, DAP is efficient to carry out – the auxiliary
module is lightweight that requires around $7\%$ extra training computation
and is removed during the inference stage.
We evaluate DAP on two standard UDA segmentation benchmarks, in which the
source domain is defined by a synthesized dataset (i.e., GTAv [39] or SYNTHIA
[40]) while the target domain involves Cityscapes [10], a dataset captured
from the real world. With the word2vec features as prior, DAP achieves
segmentation mIOU scores of $55.0\%$ and $50.2\%$ from GTAv and SYNTHIA,
respectively, with absolute gains of $2.9\%$ and $1.3\%$ over the DACS
baseline, setting the new state-of-the-art among single-model, single-round
approaches for UDA segmentation.
The main contribution of this work lies in the proposal and implementation of
domain-agnostic prior for UDA segmentation. With such a simple and effective
approach, we reveal that much room is left behind UDA. We expect more
sophisticated priors and/or more effective constraints to be explored in the
future.
Figure 2: The proposed framework of that involves building DAP (the blue-
shaded region) upon DACS [49] (the yellow-shaded region). We omit the
illustration of using ClassMix [37] to generate
$\mathbf{x}_{n,m}^{\mathbb{T}}$. The details of producing the embedding map
(i.e., the $\mathrm{proj}(\cdot)$ function) is shown in Fig 3. This figure is
best viewed in color.
## 2 Related Work
Unsupervised domain adaptation (UDA) aims to transfer models trained in a
known (labeled) source domain to an unknown (unlabeled) target domain [15, 32,
64]. It differs from semi-supervised learning [63, 27] mainly in the
potentially significant difference between the labeled and unlabeled data. The
past years have witnessed a fast development of UDA , extending the studied
task from image classification [42, 5] to fine-scaled visual recognition
including detection [8, 25], segmentation [50, 4], person re-identification
[14, 45], etc. Specifically, segmentation is a good testbed for UDA for at
least two reasons. First, compared to classification, pixel-level segmentation
requires more sophisticated manipulation of domain transfer. Second, UDA can
reduce the annotation cost for segmentation which is often expensive. The
recent approaches of UDA segmentation is roughly categorized into three parts,
adversarial learning, self-training, and data generation.
$\bullet$ Adversarial learning is an important tool for domain adaptation [52,
2]. When it was applied to UDA segmentation [50, 48], the segmentation modules
are regarded as the generator that produce prediction, and an auxiliary
discriminator is trained to judge which domain the inputs are from. By
confusing the discriminator, the purpose of domain adaptation is achieved.
Extensions beyond this idea include [24] that facilitated the domain alignment
from the input image and transferred the low-level representation (e.g. the
texture and brightness) from the target domain to the source domain, [56] that
applied a Fourier transform to narrow the low-frequency gap between the two
domains, and others.
$\bullet$ The idea of self-training was borrowed from semi-supervised learning
[47, 41, 21] and few-shot learning [28, 31], and was applied well to domain
adaptation [44]. The key to self-training is to produce high-quality pseudo
labels [27], but prediction errors are inevitable especially when the domain
gap is significant. Many efforts have been made to alleviate the inaccuracy,
including entropy minimization [65, 54, 7] that tried to increase the
confidence on unlabeled data, class-balance regulation [64] that improved the
prediction probability for hard categories and thus alleviated the burden that
minor categories may be suppressed, and [17, 60] that reduced the uncertainty
of prediction to rectify pseudo labels.
$\bullet$ The data generation branch tried to integrate the advantages of
adversarial learning and self-training. The idea is to bridge the gap between
the source and target domains using a generated domain in which source and
target data are mixed. The images can be generated either by performing GAN
[62, 19] or pixel-level mixing [37, 57]. Our approach follows the second path
which is verified powerful in UDA segmentation [49, 16, 61], and we introduce
a domain-agnostic prior to constrain representation learning on the target
domain.
There have been studies of combining visual data with other types of
information, in particular, with text data [23, 36, 26]. Powered by pre-
training on a large amount of image and text data, either paired [38] or
unpaired [22], deep neural networks gain the ability of aligning visual and
linguistic features in a shared space, hence facilitating cross-modal
understanding [46, 29] and retrieval [55, 13]. Recently, the rapid development
of Transformers [53, 12] brings the possibility of unifying image and text
data within one framework. Transferring the learned knowledge from one
modality to another may cause the setting somewhat similar to zero-shot
learning [3, 59]. There are also some discussions on safe and/or efficient
cross-modal learning, including [20] that introduced knowledge distillation
into cross-modal retrieval, and [1] that introduced boundary-aware regression
and semantic consistency constraint and improved the discrimination for the
unlabeled classes.
## 3 Our Approach
### 3.1 Problem Setting and Baselines
Unsupervised domain adaptation (UDA) starts with defining two domains
corresponding to different data distributions. In our setting for semantic
segmentation, complete pixel-level annotations are available for the source
domain, $\mathbb{S}$, but unavailable for the target domain, $\mathbb{T}$, yet
we hold an assumption that the class sets on $\mathbb{S}$ and $\mathbb{T}$ are
identical. From the data perspective, let the training samples on $\mathbb{S}$
from a set of
$\mathcal{D}^{\mathbb{S}}=\\{(\mathbf{x}_{n}^{\mathbb{S}},\mathbf{y}_{n}^{\mathbb{S}})\\}_{n=1}^{N}$
where both $\mathbf{x}_{n}^{\mathbb{S}}$ is a high-resolution image and
$\mathbf{y}_{n}^{\mathbb{S}}$ of the same size offers pixel-wise semantic
labels. Similarly, we have a training set on $\mathbb{T}$ being
$\mathcal{D}^{\mathbb{T}}=\\{(\mathbf{x}_{m}^{\mathbb{T}})\\}_{m=1}^{M}$ where
ground-truth labels are not provided. The goal is to train a model
$\mathbf{y}=f(\mathbf{x};\boldsymbol{\theta})$ on both
$\mathcal{D}^{\mathbb{S}}$ and $\mathcal{D}^{\mathbb{T}}$ so that it works
well on a hidden test set $\mathcal{D}^{\mathbb{T}\prime}$ which is also
sampled from $\mathbb{T}$.
The major difficulty of UDA comes from the domain gap between $\mathbb{S}$ and
$\mathbb{T}$, e.g., in our testing environment, $\mathbb{S}$ corresponds to
synthesized data [39, 40] but $\mathbb{T}$ corresponds to real-world data
[10]. The potential differences in lighting, object styles, etc., can
downgrade prediction confidence as well as quality on the target domain. One
of the popular solutions to bridge the domain gap is to use the source model
to generate pseudo labels $\hat{\mathbf{y}}_{m}^{\mathbb{T}}$ for
$\mathbf{x}_{m}^{\mathbb{T}}$ by a mean-teacher model [47], and uses both
$\mathcal{D}^{\mathbb{S}}$ and
$\hat{\mathcal{D}}^{\mathbb{T}}=\\{(\mathbf{x}_{m}^{\mathbb{T}},\hat{\mathbf{y}}_{m}^{\mathbb{T}})\\}_{m=1}^{M}$,
the extended target set, for updating the online student model. We denote the
teacher and student models as
$f^{\mathrm{te}}(\mathbf{x};\boldsymbol{\theta}^{\mathrm{te}})$ and
$f^{\mathrm{st}}(\mathbf{x};\boldsymbol{\theta}^{\mathrm{st}})$, respectively,
where in the mean-teacher algorithm, $f^{\mathrm{te}}(\cdot)$ is the moving
average of $f^{\mathrm{st}}(\cdot)$.
The above mechanism, though simple and elegant, suffers the unreliability of
$\hat{\mathbf{y}}_{m}^{\mathbb{T}}$. To alleviate this issue, in a recently
published work named DACS [49], researchers proposed to replace the training
data from the target domain with that from a mixed domain, $\mathbb{M}$, where
the samples are generated by mixing images at the pixel level guided by class
labels [37]. In each training iteration, a pair of source and target images
with (true or pseudo) labels are sampled and cropped into the same resolution,
denoted as
$(\mathbf{x}_{n}^{\mathbb{S}},\mathbf{y}_{n}^{\mathbb{S}},\mathbf{x}_{m}^{\mathbb{T}},\hat{\mathbf{y}}_{m}^{\mathbb{T}})$.
Next, a subset of classes is randomly chosen from
$\mathbf{y}_{n}^{\mathbb{S}}$ and a binary mask $\mathbf{M}_{n,m}$ of the same
size as $\mathbf{x}_{n}^{\mathbb{S}}$ is made, with all pixels that
$\mathbf{y}_{n}^{\mathbb{S}}$ belongs to the subset being $1$ and otherwise
$0$. Upon $\mathbf{M}_{n,m}$, a mixed image with its label is defined:
$\left\\{\begin{array}[]{l}\mathbf{x}_{n,m}^{\mathbb{M}}=\mathbf{x}_{n}^{\mathbb{S}}\odot\mathbf{M}_{n,m}+\mathbf{x}_{m}^{\mathbb{T}}\odot(\mathbf{1}-\mathbf{M}_{n,m})\\\
\mathbf{y}_{n,m}^{\mathbb{M}}=\mathbf{y}_{n}^{\mathbb{S}}\odot\mathbf{M}_{n,m}+\hat{\mathbf{y}}_{m}^{\mathbb{T}}\odot(\mathbf{1}-\mathbf{M}_{n,m})\end{array}\right.,$
(1)
where $\odot$ denotes element-wise multiplication. The student model,
$f^{\mathrm{st}}(\mathbf{x};\boldsymbol{\theta}^{\mathrm{st}})$, is trained
with $(\mathbf{x}_{n}^{\mathbb{S}},\mathbf{y}_{n}^{\mathbb{S}})$ and
$(\mathbf{x}_{n}^{\mathbb{M}},\mathbf{y}_{n}^{\mathbb{M}})$:
$\mathcal{L}_{\mathrm{seg}}=\mathcal{L}_{\mathrm{CE}}(f^{\mathrm{st}}(\mathbf{x}_{n}^{\mathbb{S}};\boldsymbol{\theta}^{\mathrm{st}}),\mathbf{y}_{n}^{\mathbb{S}})+\mathcal{L}_{\mathrm{CE}}(f^{\mathrm{st}}(\mathbf{x}_{n,m}^{\mathbb{M}};\boldsymbol{\theta}^{\mathrm{st}}),\mathbf{y}_{n,m}^{\mathbb{M}}),$
(2)
where $\mathcal{L}_{\mathrm{CE}}(\cdot,\cdot)$ is the pixel-wise cross-entropy
loss. The teacher model,
$f^{\mathrm{te}}(\mathbf{x};\boldsymbol{\theta}^{\mathrm{te}})$, is updated
with the student model using the exponential moving average (EMA) mechanism,
namely,
$\boldsymbol{\theta}^{\mathrm{te}}\leftarrow\boldsymbol{\theta}^{\mathrm{te}}\cdot\lambda+\boldsymbol{\theta}^{\mathrm{st}}\cdot(1-\lambda)$,
where $\lambda$ controls the window of EMA and is often close to $1.0$. The
entire flowchart of DACS is illustrated in the yellow-shaded part of Fig 2.
### 3.2 Confusion of Semantically Similar Classes
| | $\mathrm{A}$: motor, $\mathrm{B}$: bike | $\mathrm{A}$: road, $\mathrm{B}$: sidewalk
---|---|---|---
Methods | Data | $\mathrm{cos}\langle\boldsymbol{\mu}_{\mathrm{A}}^{\cdot},\boldsymbol{\mu}_{\mathrm{B}}^{\cdot}\rangle$ | IOU (%) | $\mathrm{cos}\langle\boldsymbol{\mu}_{\mathrm{A}}^{\cdot},\boldsymbol{\mu}_{\mathrm{B}}^{\cdot}\rangle$ | IOU (%)
Source Only | $\mathbb{S}$ | 0.42 | 6.36 | 0.35 | 1.43
$\mathbb{T}$ | 0.60 | 27.76 | 0.43 | 5.53
DACS | $\mathbb{T}$ | 0.60 | 26.62 | 0.38 | 1.95
DAP | $\mathbb{T}$ | 0.56 | 22.19 | 0.23 | 1.29
Table 1: Statistics of features extracted two pairs of semantically similar
classes, where $\mathbb{S}$ and $\mathbb{T}$ corresponds to GTAv and
Cityscapes, respectively. Please refer to the main text for details.
Despite the effectiveness in stabilizing self-learning, DACS still has
difficulties in distinguishing semantically similar classes, especially when
these classes do not appear frequently in the target domain, e.g., motorbike
accounts for only $0.1\%$ of the total number of pixels. Fig 1 shows an
example that the class pair of motorbike and bike can easily confuse the
model, and so can the pair of road and sidewalk. According to experimental
results, such confusion contributes significantly to segmentation error, e.g.,
$20.8\%$ mis-classification of bike pixels goes to motorbike.
We offer a hypothesis for the above phenomenon. Since data from the target
domain, say $\mathbf{x}_{m}^{\mathbb{T}}$, are not labeled, the semantic
correspondence is learned by mapping $\mathbf{x}_{m}^{\mathbb{T}}$ to the
source domain, e.g., by image-level style transfer for GAN-based approaches
[62, 19] and by label-level simulation for DACS. Mathematically, this is to
learn a transfer function (which transfers visual features from the target
domain to the source domain) in a weakly-supervised manner. This causes
approximation of visual representation and consequently incurs inaccuracy of
recognition.
In addition, we consider two semantically similar classes denoted as
$\mathrm{A}$ and $\mathrm{B}$, and the corresponding features extracted from
these classes form two sets of $\mathcal{F}_{\mathrm{A}}^{\cdot}$ and
$\mathcal{F}_{\mathrm{B}}^{\cdot}$, respectively, where the superscript can be
either $\mathbb{S}$ or $\mathbb{T}$. Let us assume that each feature set
follows a multi-variate Gaussian distribution denoted by
$\mathcal{N}(\boldsymbol{\mu}_{\mathrm{A}}^{\cdot},\boldsymbol{\Sigma}_{\mathrm{A}}^{\cdot})$
and
$\mathcal{N}(\boldsymbol{\mu}_{\mathrm{B}}^{\cdot},\boldsymbol{\Sigma}_{\mathrm{B}}^{\cdot})$,
abbreviated as $\mathcal{N}_{\mathrm{A}}^{\cdot}$ and
$\mathcal{N}_{\mathrm{B}}^{\cdot}$, respectively. Since $\mathrm{A}$ and
$\mathrm{B}$ are semantically similar, we assume that
$\boldsymbol{\mu}_{\mathrm{A}}^{\mathbb{S}}$ and
$\boldsymbol{\mu}_{\mathrm{B}}^{\mathbb{S}}$ are close in the feature space,
and the source model learns to distinguish $\mathrm{A}$ from $\mathrm{B}$ by
reducing
$\cos\langle\boldsymbol{\mu}_{\mathrm{A}}^{\mathbb{S}},\boldsymbol{\mu}_{\mathrm{B}}^{\mathbb{S}}\rangle$
and hence the IOU between $\mathcal{N}_{\mathrm{A}}^{\mathbb{S}}$ and
$\mathcal{N}_{\mathrm{B}}^{\mathbb{S}}$ 111To calculate the IOU between
$\mathcal{N}_{\mathrm{A}}^{\mathbb{S}}$ and
$\mathcal{N}_{\mathrm{B}}^{\mathbb{S}}$, we sample an equal amount of points
from $\mathcal{N}_{\mathrm{A}}^{\mathbb{S}}$ and
$\mathcal{N}_{\mathrm{B}}^{\mathbb{S}}$, and calculate $r_{\mathrm{A}}$ as the
probability that a point sampled from $\mathcal{N}_{\mathrm{A}}^{\mathbb{S}}$
actually has a higher density at $\mathcal{N}_{\mathrm{B}}^{\mathbb{S}}$, and
$r_{\mathrm{B}}$ vice versa. The IOU is then approximated as
$(r_{\mathrm{A}}+r_{\mathrm{B}})/(2-r_{\mathrm{A}}-r_{\mathrm{B}})$.. However,
in the target domain, these conditions are not necessarily satisfied since no
strong supervision is present – the distance between
$\boldsymbol{\mu}_{\mathrm{A}}^{\mathbb{T}}$ and
$\boldsymbol{\mu}_{\mathrm{B}}^{\mathbb{T}}$ gets smaller (i.e.,
$\cos\langle\boldsymbol{\mu}_{\mathrm{A}}^{\mathbb{T}},\boldsymbol{\mu}_{\mathrm{B}}^{\mathbb{T}}\rangle$
gets larger), and, consequently, the IOU between
$\mathcal{N}_{\mathrm{A}}^{\mathbb{T}}$ and
$\mathcal{N}_{\mathrm{B}}^{\mathbb{T}}$ is larger. Tab 1 offers quantitative
results from transferring a segmentation model from GTAv to Cityscapes. The
features of two semantically similar class pairs, namely, motorbike vs. bike,
and road vs. sidewalk, are both made easier to confuse the network.
### 3.3 Domain-Agnostic Prior for UDA Segmentation
To obtain more accurate estimation for $\mathcal{N}_{\mathrm{A}}^{\mathbb{T}}$
and $\mathcal{N}_{\mathrm{B}}^{\mathbb{T}}$, we refer to the Bayesian theory
that the posterior distribution is composed of a prior and a likelihood. In
our setting, the likelihood comes from the target dataset,
$\mathcal{D}^{\mathbb{T}}$, where there are insufficient data to guarantee
accurate estimation. The solution lies in introducing an informative prior for
the elements, namely,
$\mathbf{z}_{\mathrm{A}}^{\mathbb{T}}\sim\mathcal{N}_{\mathrm{A}}^{\mathbb{T}}$
and
$\mathbf{z}_{\mathrm{B}}^{\mathbb{T}}\sim\mathcal{N}_{\mathrm{B}}^{\mathbb{T}}$.
This involves defining constraints as follows:
$\mathbf{z}_{\mathrm{A}}^{\mathbb{T}}=g(\mathbf{e}_{\mathrm{A}}),\quad\mathbf{z}_{\mathrm{B}}^{\mathbb{T}}=g(\mathbf{e}_{\mathrm{B}}),$
(3)
where $\mathbf{e}_{\mathrm{A}}$ and $\mathbf{e}_{\mathrm{B}}$ are not related
to any specific domain – we name them domain-agnostic priors (DAP), and
$g(\cdot)$ is a learnable function that projects the priors to the semantic
space. In practice, it is less likely to strictly satisfy Eqn (3), so we
implement it as a loss term to minimize.
We instantiate DAP into two examples. The first one is a set of one-hot
vectors, i.e., for a target domain with $C$ classes, the $c$-th class is
encoded into $\mathbf{I}_{c}$, a $C$-dimensional vector in which all entries
are $0$ except for the $c$-th dimension being $1$. The second one is borrowed
from word2vec [35], a pre-trained language model that represents each word
using a $300$-dimensional vector. Note that both cases are completely
independent from the vision domain, i.e., reflecting the principle of being
domain-agnostic. We denote the set of embedding vectors as
$\mathcal{E}=\\{\mathbf{e}_{c}\\}_{c=1}^{C}$, where $\mathbf{e}_{c}$ is the
embedding vector of the $c$-th class.
Back to the main story, DAP and the embedded vectors are easily integrated
into DACS, the baseline framework. Besides the segmentation loss
$\mathcal{L}_{\mathrm{seg}}$ defined in Eqn (2), we introduce another loss
term $\mathcal{L}_{\mathrm{DAP}}$ that measures the distance between the
embedded domain-agnostic priors and visual features extracted from training
images,
$\displaystyle\mathcal{L}_{\mathrm{DAP}}=$
$\displaystyle\|g_{\mathrm{vi}}(\tilde{f}^{\mathrm{st}}(\mathbf{x}_{n}^{\mathbb{S}};\boldsymbol{\theta}^{\mathrm{st}}))-g_{\mathrm{pr}}(\mathrm{proj}(\tilde{\mathbf{y}}_{n}^{\mathbb{S}};\mathcal{E}))\|_{2}^{2}+$
(4)
$\displaystyle\|g_{\mathrm{vi}}(\tilde{f}^{\mathrm{st}}(\mathbf{x}_{n,m}^{\mathbb{M}};\boldsymbol{\theta}^{\mathrm{st}}))-g_{\mathrm{pr}}(\mathrm{proj}(\tilde{\mathbf{y}}_{n,m}^{\mathbb{M}};\mathcal{E}))\|_{2}^{2},$
where $\tilde{f}^{\mathrm{st}}(\cdot)$ is the backbone part of
$f^{\mathrm{st}}(\cdot)$ that extracts mid-resolution features (i.e., visual
features), $\tilde{\mathbf{y}}_{n}^{\mathbb{S}}$ and
$\tilde{\mathbf{y}}_{n,m}^{\mathbb{M}}$ denote the supervisions adjusted to
the same size of the backbone outputs, $\mathrm{proj}(\cdot)$ projects the
domain-agnostic embedding vectors to the image plane based on the labels (to
be detailed later), and $g_{\mathrm{vi}}(\cdot)$ and $g_{\mathrm{pr}}(\cdot)$
are each a learnable convolution layer that maps initial features from the
vision and prior spaces to the common space.
We elaborate the construction of $\mathrm{proj}(\cdot)$ in Fig 3. The
computation is simple: for each position on the image plane, we refer to the
ground-truth label $\tilde{\mathbf{y}}_{n}^{\mathbb{S}}$ or partly-pseudo
label $\tilde{\mathbf{y}}_{n,m}^{\mathbb{M}}$, and directly paste the class-
wise embedding vectors to the corresponding positions. Since $\mathcal{E}$ is
fixed, this procedure does not need any gradient back-propagation. We are
aware that advanced versions of $\mathrm{proj}(\cdot)$ (e.g., performing local
smoothing) are available, and we leave these properties to be learned by
$g_{\mathrm{vi}}(\cdot)$ and $g_{\mathrm{pr}}(\cdot)$.
Figure 3: An illustration of constructing the domain-agnostic embedding map
with prior vectors. Each number in the label map stands for a class ID. The
total number of classes can be arbitrary.
Finally, the overall loss function is written as
$\mathcal{L}_{\mathrm{overall}}=\mathcal{L}_{\mathrm{seg}}+\alpha\cdot\mathcal{L}_{\mathrm{DAP}},$
(5)
where $\alpha$ is the balancing coefficient which is by default set to be
$1.0$ – an analysis of the effect of $\alpha$ is provided in the experimental
part (see Tab 6).
Back to Tab 1, one can observe how DAP reduces the IOU between
$\mathcal{N}_{\mathrm{A}}^{\mathbb{T}}$ and
$\mathcal{N}_{\mathrm{B}}^{\mathbb{T}}$, and thus alleviate the confusion
between the semantically similar classes. Interestingly, the gains of two
class pairs are consistent and significant, even though the training samples
of road vs. sidewalk being more abundant. In experiments, we shall see how the
ability of discriminating these class pairs is improved.
### 3.4 Discussions
To the best of our knowledge, this is the first work that integrates text
embedding into UDA segmentation and producing considerable accuracy gain,
which demonstrating the effectiveness of linguistic cues assisting visual
recognition. However, it is yet a preliminary solution, and some possible
directions can be discovered.
$\bullet$ Enhancing text embedding. The currently used word2vec features it
does not consider different words that correspond to the same semantics (e.g.,
person can be pedestrian). Interestingly, we tried to enhance the prior by
searching for semantically similar words, but obtained little accuracy gain.
This may call for a complicated mechanism of exploring the text world.
$\bullet$ Constructing domain-agnostic yet vision-aware priors. This is to
answer the question: what kind of image data is considered to offer domain-
free information? The answer may lie in generalized datasets like ImageNet
[11] or Conceptual Captioning [43], or even the pre-trained image-text models
such as CLIP [38] that absorbed $400$ million image-text pairs. Note that it
is a major challenge to disentangle domain-related information to avoid over-
fitting, and we will continue exploring the possibility in the future.
## 4 Experiments
GTAv$\rightarrow$Cityscapes
---
Method | Road | Sidewalk | Building | Wall | Fence | Pole | Light | Sign | Veg | Terrain | Sky | Person | Rider | Car | Truck | Bus | Train | Motor | Bike | mIOU
AdaptSegNe [50] | 86.5 | 36.0 | 79.9 | 23.4 | 23.3 | 23.9 | 35.2 | 14.8 | 83.4 | 33.3 | 75.6 | 58.5 | 27.6 | 73.7 | 32.5 | 35.4 | 3.9 | 30.1 | 28.1 | 42.4
PatchAligne [51] | 92.3 | 51.9 | 82.1 | 29.2 | 25.1 | 24.5 | 33.8 | 33.0 | 82.4 | 32.8 | 82.2 | 58.6 | 27.2 | 84.3 | 33.4 | 46.3 | 2.2 | 29.5 | 32.3 | 46.5
LTIR [24] | 92.9 | 55.0 | 85.3 | 34.2 | 31.1 | 34.9 | 40.7 | 34.0 | 85.2 | 40.1 | 87.1 | 61.0 | 31.1 | 82.5 | 32.3 | 42.9 | 0.3 | 36.4 | 46.1 | 50.2
PIT [33] | 87.5 | 43.4 | 78.8 | 31.2 | 30.2 | 36.3 | 39.9 | 42.0 | 79.2 | 37.1 | 79.3 | 65.4 | 37.5 | 83.2 | 46.0 | 45.6 | 25.7 | 23.5 | 49.9 | 50.6
FDA [56] | 92.5 | 53.3 | 82.4 | 26.5 | 27.6 | 36.4 | 40.6 | 38.9 | 82.3 | 39.8 | 78.0 | 62.6 | 34.4 | 84.9 | 34.1 | 53.1 | 16.9 | 27.7 | 46.4 | 50.5
MetaCorrect [17] | 92.8 | 58.1 | 86.2 | 39.7 | 33.1 | 36.3 | 42.0 | 38.6 | 85.5 | 37.8 | 87.6 | 62.8 | 31.7 | 84.8 | 35.7 | 50.3 | 2.0 | 36.8 | 48.0 | 52.1
DACS [49] | 89.9 | 39.7 | 87.9 | 39.7 | 39.5 | 38.5 | 46.4 | 52.8 | 88.0 | 44.0 | 88.8 | 67.2 | 35.8 | 84.5 | 45.7 | 50.2 | 0.0 | 27.3 | 34.0 | 52.1
IAST [34] | 94.1 | 58.8 | 85.4 | 39.7 | 29.2 | 25.1 | 43.1 | 34.2 | 84.8 | 34.6 | 88.7 | 62.7 | 30.3 | 87.6 | 42.3 | 50.3 | 24.7 | 35.2 | 40.2 | 52.2
DPL [9] | 92.8 | 54.4 | 86.2 | 41.6 | 32.7 | 36.4 | 49.0 | 34.0 | 85.8 | 41.3 | 86.0 | 63.2 | 34.2 | 87.2 | 39.3 | 44.5 | 18.7 | 42.6 | 43.1 | 53.3
Source only | 75.6 | 17.1 | 69.8 | 10.7 | 16.1 | 21.1 | 27.0 | 10.6 | 77.3 | 15.1 | 71.1 | 53.4 | 20.5 | 73.9 | 28.6 | 31.1 | 1.62 | 32.4 | 21.5 | 35.5
DACS (rep) | 93.1 | 48.1 | 87.3 | 36.7 | 35.1 | 38.7 | 42.5 | 49.3 | 87.5 | 41.9 | 87.9 | 64.8 | 30.7 | 88.3 | 40.2 | 51.0 | 0.0 | 25.1 | 42.6 | 52.1
DACS+DAP | 93.5 | 53.9 | 87.5 | 30.0 | 36.4 | 39.0 | 43.9 | 49.5 | 87.5 | 45.4 | 88.8 | 66.6 | 36.8 | 89.4 | 49.1 | 51.4 | 0.0 | 42.2 | 53.1 | 55.0
ProDA [58] | 87.8 | 56.0 | 79.7 | 46.3 | 44.8 | 45.6 | 53.5 | 53.5 | 88.6 | 45.2 | 82.1 | 70.7 | 39.2 | 88.8 | 45.5 | 50.4 | 1.0 | 48.9 | 56.4 | 57.5
Chao et al. [4] | 94.4 | 60.9 | 88.1 | 39.5 | 41.8 | 43.2 | 49.1 | 56.0 | 88.0 | 45.8 | 87.8 | 67.6 | 38.1 | 90.1 | 57.6 | 51.9 | 0.0 | 46.6 | 55.3 | 58.0
DAP + ProDA | 94.5 | 63.1 | 89.1 | 29.8 | 47.5 | 50.4 | 56.7 | 58.7 | 89.5 | 50.2 | 87.0 | 73.6 | 38.6 | 91.3 | 50.2 | 52.9 | 0.0 | 50.2 | 63.5 | 59.8
Table 2: Segmentation accuracy (IOU, %) of different UDA approaches from GTAv
to Cityscapes. DACS (rep) indicates our re-implementation of DACS. For each
class, we mark the highest number with bold and the second highest with
underline. The top and bottom parts are achieved without and with multi-stage
training or multi-model fusion.
SYNTHIA$\rightarrow$Cityscapes
---
Method | Road | Sidewalk | Building | Wall* | Fence* | Pole* | Light | Sign | Veg | Sky | Person | Rider | Car | Bus | Motor | Bike | mIOU | mIOU*
PatchAlign [51] | 82.4 | 38.0 | 78.6 | 8.7 | 0.6 | 26.0 | 3.9 | 11.1 | 75.5 | 84.6 | 53.5 | 21.6 | 71.4 | 32.6 | 19.3 | 31.7 | 40.0 | 46.5
AdaptSegNe [50] | 84.3 | 42.7 | 77.5 | – | – | – | 4.7 | 7.0 | 77.9 | 82.5 | 54.3 | 21.0 | 72.3 | 32.2 | 18.9 | 32.3 | – | 46.7
FDA [56] | 79.3 | 35.0 | 73.2 | – | – | – | 19.9 | 24.0 | 61.7 | 82.6 | 61.4 | 31.1 | 83.9 | 40.8 | 38.4 | 51.1 | – | 52.5
LTIR [24] | 92.6 | 53.2 | 79.2 | – | – | – | 1.6 | 7.5 | 78.6 | 84.4 | 52.6 | 20.0 | 82.1 | 34.8 | 14.6 | 39.4 | – | 49.3
PIT [33] | 83.1 | 27.6 | 81.5 | 8.9 | 0.3 | 21.8 | 26.4 | 33.8 | 76.4 | 78.8 | 64.2 | 27.6 | 79.6 | 31.2 | 31.0 | 31.3 | 44.0 | 51.8
MetaCorrect [17] | 92.6 | 52.7 | 81.3 | 8.9 | 2.4 | 28.1 | 13.0 | 7.3 | 83.5 | 85.0 | 60.1 | 19.7 | 84.8 | 37.2 | 21.5 | 43.9 | 45.1 | 52.5
DPL [9] | 87.5 | 45.7 | 82.8 | 13.3 | 0.6 | 33.2 | 22.0 | 20.1 | 83.1 | 86.0 | 56.6 | 21.9 | 83.1 | 40.3 | 29.8 | 45.7 | 47.0 | 54.2
DACS [49] | 80.6 | 25.1 | 81.9 | 21.5 | 2.9 | 37.2 | 22.7 | 24.0 | 83.7 | 90.8 | 67.6 | 38.3 | 82.9 | 38.9 | 28.5 | 47.6 | 48.3 | 54.8
IAST [34] | 81.9 | 41.5 | 83.3 | 17.7 | 4.6 | 32.3 | 30.9 | 28.8 | 83.4 | 85.0 | 65.6 | 30.8 | 86.5 | 38.2 | 33.1 | 52.7 | 49.8 | 57.0
Source only | 24.7 | 12.1 | 75.4 | 11.3 | 0.1 | 22.4 | 7.5 | 16.6 | 71.7 | 78.3 | 52.9 | 10.1 | 56.6 | 23.3 | 4.0 | 13.0 | 30.0 | 34.3
DACS (rep) | 82.1 | 31.0 | 82.4 | 22.1 | 1.2 | 33.1 | 32.7 | 25.1 | 84.4 | 88.2 | 65.2 | 34.3 | 83.4 | 42.9 | 24.1 | 50.8 | 48.9 | 55.9
DACS+DAP | 83.9 | 33.3 | 80.2 | 24.1 | 1.2 | 33.4 | 30.8 | 33.8 | 84.3 | 88.5 | 65.7 | 36.2 | 84.3 | 43.3 | 33.3 | 46.3 | 50.2 | 57.2
Chao et al. [4] | 88.7 | 46.7 | 83.8 | 22.7 | 4.1 | 35.0 | 35.9 | 36.1 | 82.8 | 81.4 | 61.6 | 32.1 | 87.9 | 52.8 | 32.0 | 57.7 | 52.6 | 60.0
ProDA [58] | 87.8 | 45.7 | 84.6 | 37.1 | 0.6 | 44.0 | 54.6 | 37.0 | 88.1 | 84.4 | 74.2 | 24.3 | 88.2 | 51.1 | 40.5 | 45.6 | 55.5 | 62.0
DAP + ProDA | 84.2 | 46.5 | 82.5 | 35.1 | 0.2 | 46.7 | 53.6 | 45.7 | 89.3 | 87.5 | 75.7 | 34.6 | 91.7 | 73.5 | 49.4 | 60.5 | 59.8 | 64.3
Table 3: Segmentation accuracy (IOU, %) of different UDA approaches from
SYNTHIA to Cityscapes. DACS (rep) indicates our re-implementation of DACS. For
each class, we mark the highest number with bold and the second highest with
underline. The top and bottom parts are achieved without and with multi-stage
training or multi-model fusion. mIOU* denotes the average of 13 classes, with
the classes marked with * not computed.
### 4.1 Datasets and Implementation Details
$\bullet$ Datasets. We evaluate our method on a popular scenario transferring
the information from a synthesis domain to a real domain. We use GTAv [39] and
SYNTHIA [40] as composite domain datasets and Cityscapes [10] as the real
domain. GTAv [39] is a synthetic dataset extracted from the game of Grand
Theft Auto V. There are $24\rm{,}966$ images with pixel-level semantic
segmentation ground truth. The resolution of these images is $1914\times 1052$
and we resize them into $1280\times 720$ in our experiments. GTAv shares $19$
common classes with Cityscapes. SYNTHIA [40] contains $9\rm{,}400$ virtual
European-style urban images whose resolution is $1280\times 760$ and we keep
the original size in our experiments. We evaluate two settings (13 and 16
categories) in SYNTHIA. Cityscapes is a large-scale dataset with a resolution
of $2048\times 1024$. There are $2\rm{,}975$ and $500$ images in the training
and validation sets, respectively.
$\bullet$ Implementation Details. To be consistent with other methods, we use
the Deeplabv2 [6] framework with a RseNet101 [18] backbone as our image
encoder and an ASPP classifier as segmentation head. The output map is up-
sampled and operated by a softmax layer to match the size of the inputs. The
pre-trained model on ImageNet [11] and MSCOCO [30] is applied to initialize
the backbone. The visual and prior feature projectors,
$g_{\mathrm{vi}}(\cdot)$ and $g_{\mathrm{pr}}(\cdot)$, are both convolution
layers with a $1\times 1$ kernel to adjust the channel to $256$. The batch
size of one GPU is set to $2$. We use SGD with Nesterov acceleration as the
optimizer which is decreased based on a polynomial decay policy with exponent
$0.9$. The initial learning rate of backbone and feature projectors is
$2.5\times 10^{-4}$, that of segmentation head is $10\times$ larger. The
momentum and weight decay of the optimizer are $0.9$ and $5\times 10^{-4}$.
During the training process, we apply color jittering and Gaussian blurring on
the mixed data and only resize operation on source and target data. Teacher
model is updated with an EMA decay and $\lambda$ equals $0.99$. We train the
model for $250\mathrm{K}$ iterations on a single NVIDIA Tesla-V100 GPU and
adopt an early stop setting. The background and invalid categories are ignored
during training. The weight of $\mathcal{L}_{\mathrm{DAP}}$, i.e. $\alpha$, is
set to be $1.0$. We adapt bilinear interpolation to down-sample the embedding
map from the input size to that of the visual feature map – this is an
important step for DAP, which is ablated in Section 4.3.
Figure 4: An example of transfer segmentation from GTAv to Cityscapes. The top
row shows the ground-truth and three segmentation results, while the input and
legend is in the bottom row. The yellow boxes indicate the regions that the
segmentation quality is largely improved, and the number corresponds to
single-image mIOU (over the existing classes). The bottom row also shows the
t-SNE of visual features colored by the predicted class. The blue boxes locate
the features of bike and motorbike, in which DAP achieves a favorable ability
to distinguish them. This figure is best viewed in color and we suggest the
reader to zoom in for details.
### 4.2 Quantitative Results and Visualization
We first evaluate DAP on the transfer segmentation task from GTAv to
Cityscapes. The comparison against recent approaches are shown in Tab 2. To
ensure reliability, we run DACS (the baseline) and DAP three times and report
the average accuracy. DAP achieves a mean IOU of $55.0\%$ over 19 classes,
which claims a $2.9\%$ gain beyond the baseline and also outperforms all other
competitors except for Chao et al. [4] and ProDA [58]. Specifically, Chao et
al. [4] used ensemble learning to integrate the prediction from four
complementarily-trained models, including DACS, but DAP used a single model;
ProDA [58] improved the segmentation accuracy significantly via multi-stage
training, yet its first stage reported a $53.7\%$ mIOU.What is more, the
results on transferring SYNTHIA to Cityscapes, as shown in Tab 3, demonstrate
the similar trend – DAP outperforms all the competitors, except for ProDA and
RED, in terms of either 13-class or 16-class mIOU. To show that DAP offers
complementary benefits, we feed the output of DAP as the pseudo labels to the
1st stage of ProDA, and the 2nd and 3rd stages remain unchanged. As shown in
Tables 2 and 3, the segmentation mIOUs of ProDA in GTAv$\rightarrow$Cityscapes
and SYNTHIA$\rightarrow$Cityscapes are improved by $2.3\%$ and $4.3\%$,
respectively, setting new records in these two scenarios.
Next, we investigate the ability of DAP in distinguishing semantically similar
classes. From GTAv to Cityscapes, the segmentation mIOUs of bike and motorbike
are improved from $42.6\%$ and $25.1\%$ to $53.1\%$ and $42.2\%$, with
absolute gains of $10.5\%$ and $17.1\%$, respectively. From SYNTHIA to
Cityscapes, the mIOU of bike drops by $4.5\%$ and that of motorbike increases
by $9.2\%$, achieving an average improvement of $2.4\%$. We visualize an
example of segmentation in Fig 4. Besides a qualitative observation on the
improvement of distinguishing bike from motorbike and road from sidewalk, we
also notice the reason behind the improvement being a scattered feature
distribution of these similar classes. This aligns with the statistics shown
in Tab 1, indicating that DAP reduces the IOU between the estimated
distributions of bike and motorbike as well as that between road from
sidewalk.
From another perspective, we study how the language-based prior assists visual
recognition. In Fig 5, we show the relationship matrix of the features learned
from the source (GTAv) and target (Cityscapes) domains as well as the word2vec
features. There is an interesting example that person and rider are
semantically similar in both GTAv and Cityscapes but not so correlated
according to word2vec. This may cause a strong yet harmful correlation of
these two feature sets, leading to confusion in segmentation. The relatively
weaker correlation of word2vec features alleviates the bias and improves the
IOU of both classes (refer to Tab 2). Similar phenomena are also observed for
the class pairs of road vs. sidewalk, and bike vs. motorbike.
Figure 5: The relationship matrix of all 19 classes, where the left and middle
are generated by the class-averaged feature vectors extracted from the models
directly trained on GTAv and Cityscapes, and the right uses the word2vec
features. Each cell is the inner product of two normalized features. The order
of class is identical to that in Tab 2 – for easier reference: 1 for road, 2
for sidewalk, 12 for person, 13 for rider, 18 for motorbike, 19 for bike.
Source Dataset | GTAv | SYNTHIA
---|---|---
mIOU (%) | gain | std(%) | mIOU (%) | gain | std(%)
Baseline | 52.1 | – | 1.6 | 48.9 | – | 0.3
w/ random vectors | 52.3 | 0.2$\uparrow$ | 2.3 | 48.0 | 0.9$\downarrow$ | 0.8
w/ one-hot vectors | 53.0 | 0.9$\uparrow$ | 0.8 | 49.8 | 0.9$\uparrow$ | 0.4
w/ CLIP [38] | 54.6 | 2.5$\uparrow$ | 0.6 | 50.6 | 1.7$\uparrow$ | 0.5
w/ word2vec [35] | 55.0 | 2.9$\uparrow$ | 0.5 | 50.2 | 1.3$\uparrow$ | 0.4
Table 4: The results of different vectors choices include the adaptation from
GTAv and SYNTHIA
### 4.3 Diagnostic Studies
$\bullet$ The choice of domain-agnostic prior. We investigate three other
options except for word2vec embedding [35], namely, (1) that generating a
$300$-dimensional, normalized random vector for each class, (2) directly
creating a one-hot vector for each class, where the dimensionality equals to
the number of classes, (3) classes embeddings generated by the language branch
of CLIP [38]. Similar to the main experiments, we run each option three times
and report the averaged accuracy. Results are summarized in Tab 4.
One can see that random vectors achieve a slight $0.2\%$ accuracy gain on
GTAv$\rightarrow$Cityscapes, but incurs a $0.9\%$ accuracy drop on
SYNTHIA$\rightarrow$Cityscapes, implying the instability. Throughout the three
individual runs on GTAv$\rightarrow$Cityscapes, the best run achieves a
$54.4\%$ mIOU, just $0.6\%$ lower than using word2vec, but the worst one
reports $49.2\%$ which is even significantly lower than the baseline.
Regarding the one-hot vectors, the results over three runs are less
diversified, and the average improvement is consistent, (i.e., $0.9\%$ on both
GTAv$\rightarrow$Cityscapes and SYNTHIA$\rightarrow$Cityscapes), though
smaller than that brought by word2vec embedding. The prior from CLIP [38]
reports 54.6% and 50.6% mIOUs on the GTAv and SYNTHIA experiments,
respectively. Despite the fact that CLIP is stronger than word2vec, the mIOUs
are just comparable to that using word2vec (55.0% and 50.2%). From these
results, we learn the lesson that (1) even a naive prior alleviates the inter-
class confusion caused by domain shift, however, (2) it would be better if the
inter-class relationship is better captured so that the model is aware of
semantically similar classes – text embedding offers a safe and effective
option. (3) The limited number and diversity of target categories may have
diminished the advantages of a stronger language model (e.g., CLIP), and we
still pursue for a vision-aware yet domain-agnostic embedding method.
$\bullet$ Different Backbones. To verify the effectiveness of DAP on different
network structures, we replace the convolution backbone (ResNet101) with a
transformer network (ViT-Base [12]). The numbers of the block layers, token
size, and heads are $12,768$ and $12$ respectively in the transformer encoder.
The input size of the training data is set $768\times 768$. And we initiate
the ViT encoder with a model pre-trained on ImageNet-21k then fine-tune the
segmentation network with a base learning rate of $0.01$ adopted with the
’poly’ learning rate decay and use SGD as the optimizer. The $g_{\mathrm{pr}}$
is one layer transformer structure and weight of $\mathcal{L}_{\mathrm{DAP}}$
is $0.25$. On GTAv, source-only, DACS, DAP report $49.4\%$, $58.4\%$, $61.1\%$
mIOUs. As for the transferring from SYNTHIA, the numbers of the three settings
are $42.7\%$, $53.2\%$, $59.1\%$ when evaluating on 16 classes and $48.1\%$,
$60.9\%$, $66.1\%$ on 13 classes. We can see that DAP still obtains consistent
accuracy gain. To the best of our knowledge, the DAP numbers are SOTA.
$\bullet$ The importance of feature interpolation. From the model that only
uses the source domain, we gradually add mixed data (by DACS [49]), introduce
DAP (using the word2vec embedding), and perform feature interpolation to down-
sample the embedding map. As shown in Tab 5, for the both two transferring
scenarios, feature interpolation contributes nearly half accuracy gain of DAP
over DACS. Intuitively, feature interpolation enables the features on small-
area classes to be preserved, yet nearest-neighbor down-sampling can cause
these features to be ignored.
Setting | source only | DACS | DAP w/o interp | DAP w/ interp
---|---|---|---|---
GTAv | 35.5 | 52.1 | 53.8 | 55.0
SYNTHIA | 30.0 | 48.9 | 49.6 | 50.2
Table 5: Ablation on the contribution of each module in both
GTAv$\rightarrow$Cityscapes and SYNTHIA$\rightarrow$Cityscapes experiments.
$\bullet$ Parameter Analysis. Lastly, we study the impact of the coefficient
$\alpha$ that balances between $\mathcal{L}_{\mathrm{seg}}$ and
$\mathcal{L}_{\mathrm{DAP}}$. The results shown in Tab 6 suggest that
$\alpha=1.0$ is the best option. In addition, increasing $\alpha$ causes a
larger accuracy drop compared to decreasing it, which may indicate that
$\mathcal{L}_{\mathrm{seg}}$ is the essential goal and
$\mathcal{L}_{\mathrm{DAP}}$ serves as an auxiliary term.
$\alpha$ | 0.50 | 0.75 | 1.00 | 1.25 | 1.50
---|---|---|---|---|---
GTAv | 54.5 | 54.1 | 55.0 | 54.3 | 54.7
SYNTHIA | 50.0 | 49.4 | 50.2 | 49.6 | 49.1
Table 6: Impact of tuning the balancing coefficient, $\alpha$. All numbers are
segmentation mIOU (%).
## 5 Conclusions
In this paper, we investigate the UDA segmentation problem and observe that
semantically similar classes are easily confused during the transfer
procedure. We formulate this problem using the Bayesian theory and owe such
confusion to the weakness of likelihood, e.g., insufficient training data. To
alleviate the issue, we introduce domain-agnostic priors to compensate the
likelihood. Experiments on two standard benchmarks for UDA segmentation in
urban scenes verify its effectiveness both quantitatively and qualitatively
Limitations of this work. Currently, the best option of DAP is to leverage the
text embedding vectors. We are looking forward to more powerful priors, e.g.,
from the cross-modal pre-trained models [38, 22]. This may call for more
sophisticated designs of prior embedding, projection, alignment, etc., which
we leave for future work.
Acknowledge This work was supported in part by the National Natural Science
Foundation of China under Contract U20A20183 and 62021001, and in part by the
Youth Innovation Promotion Association CAS under Grant 2018497. It was also
supported by the GPU cluster built by MCC Lab of Information Science and
Technology Institution, USTC.
## References
* [1] Donghyeon Baek, Youngmin Oh, and Bumsub Ham. Exploiting a joint embedding space for generalized zero-shot semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9536–9545, 2021.
* [2] Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan. Unsupervised pixel-level domain adaptation with generative adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3722–3731, 2017.
* [3] Maxime Bucher, Tuan-Hung Vu, Matthieu Cord, and Patrick Pérez. Zero-shot semantic segmentation. Advances in Neural Information Processing Systems, 32:468–479, 2019\.
* [4] Chen-Hao Chao, Bo-Wun Cheng, and Chun-Yi Lee. Rethinking ensemble-distillation for semantic segmentation based unsupervised domain adaption. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2610–2620, 2021.
* [5] Chaoqi Chen, Weiping Xie, Wenbing Huang, Yu Rong, Xinghao Ding, Yue Huang, Tingyang Xu, and Junzhou Huang. Progressive feature alignment for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 627–636, 2019.
* [6] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE transactions on pattern analysis and machine intelligence, 40(4):834–848, 2017.
* [7] Minghao Chen, Hongyang Xue, and Deng Cai. Domain adaptation for semantic segmentation with maximum squares loss. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2090–2099, 2019.
* [8] Yuhua Chen, Wen Li, Christos Sakaridis, Dengxin Dai, and Luc Van Gool. Domain adaptive faster r-cnn for object detection in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3339–3348, 2018.
* [9] Yiting Cheng, Fangyun Wei, Jianmin Bao, Dong Chen, Fang Wen, and Wenqiang Zhang. Dual path learning for domain adaptation of semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9082–9091, 2021.
* [10] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3213–3223, 2016.
* [11] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
* [12] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
* [13] Fangxiang Feng, Xiaojie Wang, and Ruifan Li. Cross-modal retrieval with correspondence autoencoder. In Proceedings of the 22nd ACM international conference on Multimedia, pages 7–16, 2014.
* [14] Yang Fu, Yunchao Wei, Guanshuo Wang, Yuqian Zhou, Honghui Shi, and Thomas S Huang. Self-similarity grouping: A simple unsupervised cross domain adaptation approach for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6112–6121, 2019.
* [15] Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation. In International conference on machine learning, pages 1180–1189. PMLR, 2015.
* [16] Li Gao, Jing Zhang, Lefei Zhang, and Dacheng Tao. Dsp: Dual soft-paste for unsupervised domain adaptive semantic segmentation. arXiv preprint arXiv:2107.09600, 2021.
* [17] Xiaoqing Guo, Chen Yang, Baopu Li, and Yixuan Yuan. Metacorrection: Domain-aware meta loss correction for unsupervised domain adaptation in semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3927–3936, 2021.
* [18] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
* [19] Judy Hoffman, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate Saenko, Alexei Efros, and Trevor Darrell. Cycada: Cycle-consistent adversarial domain adaptation. In International conference on machine learning, pages 1989–1998. PMLR, 2018.
* [20] Hengtong Hu, Lingxi Xie, Richang Hong, and Qi Tian. Creating something from nothing: Unsupervised knowledge distillation for cross-modal hashing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3123–3132, 2020.
* [21] Xinyue Huo, Lingxi Xie, Jianzhong He, Zijie Yang, Wengang Zhou, Houqiang Li, and Qi Tian. Atso: Asynchronous teacher-student optimization for semi-supervised image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1235–1244, 2021.
* [22] Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V Le, Yunhsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. arXiv preprint arXiv:2102.05918, 2021.
* [23] Aishwarya Kamath, Mannat Singh, Yann LeCun, Gabriel Synnaeve, Ishan Misra, and Nicolas Carion. Mdetr-modulated detection for end-to-end multi-modal understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1780–1790, 2021.
* [24] Myeongjin Kim and Hyeran Byun. Learning texture invariant representation for domain adaptation of semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12975–12984, 2020.
* [25] Seunghyeon Kim, Jaehoon Choi, Taekyung Kim, and Changick Kim. Self-training and adversarial background regularization for unsupervised domain adaptive one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6092–6101, 2019.
* [26] Hugo Larochelle, Dumitru Erhan, and Yoshua Bengio. Zero-data learning of new tasks. In AAAI, volume 1, page 3, 2008.
* [27] Dong-Hyun Lee et al. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, volume 3, page 896, 2013.
* [28] Xinzhe Li, Qianru Sun, Yaoyao Liu, Qin Zhou, Shibao Zheng, Tat-Seng Chua, and Bernt Schiele. Learning to self-train for semi-supervised few-shot classification. Advances in Neural Information Processing Systems, 32:10276–10286, 2019.
* [29] Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan Zhang, Xiaowei Hu, Lei Zhang, Lijuan Wang, Houdong Hu, Li Dong, Furu Wei, et al. Oscar: Object-semantics aligned pre-training for vision-language tasks. In European Conference on Computer Vision, pages 121–137. Springer, 2020.
* [30] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European conference on computer vision, pages 740–755. Springer, 2014.
* [31] Jinlu Liu, Liang Song, and Yongqiang Qin. Prototype rectification for few-shot learning. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part I 16, pages 741–756. Springer, 2020.
* [32] Mingsheng Long, Yue Cao, Jianmin Wang, and Michael Jordan. Learning transferable features with deep adaptation networks. In International conference on machine learning, pages 97–105. PMLR, 2015.
* [33] Fengmao Lv, Tao Liang, Xiang Chen, and Guosheng Lin. Cross-domain semantic segmentation via domain-invariant interactive relation transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4334–4343, 2020.
* [34] Ke Mei, Chuang Zhu, Jiaqi Zou, and Shanghang Zhang. Instance adaptive self-training for unsupervised domain adaptation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXVI 16, pages 415–430. Springer, 2020.
* [35] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems, pages 3111–3119, 2013.
* [36] Roozbeh Mottaghi, Xianjie Chen, Xiaobai Liu, Nam-Gyu Cho, Seong-Whan Lee, Sanja Fidler, Raquel Urtasun, and Alan Yuille. The role of context for object detection and semantic segmentation in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 891–898, 2014.
* [37] Viktor Olsson, Wilhelm Tranheden, Juliano Pinto, and Lennart Svensson. Classmix: Segmentation-based data augmentation for semi-supervised learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 1369–1378, 2021.
* [38] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020, 2021.
* [39] Stephan R Richter, Vibhav Vineet, Stefan Roth, and Vladlen Koltun. Playing for data: Ground truth from computer games. In European conference on computer vision, pages 102–118. Springer, 2016.
* [40] German Ros, Laura Sellart, Joanna Materzynska, David Vazquez, and Antonio M Lopez. The synthia dataset: A large collection of synthetic images for semantic segmentation of urban scenes. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3234–3243, 2016.
* [41] Chuck Rosenberg, Martial Hebert, and Henry Schneiderman. Semi-supervised self-training of object detection models. 2005\.
* [42] Kuniaki Saito, Kohei Watanabe, Yoshitaka Ushiku, and Tatsuya Harada. Maximum classifier discrepancy for unsupervised domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3723–3732, 2018.
* [43] Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2556–2565, 2018.
* [44] Inkyu Shin, Sanghyun Woo, Fei Pan, and In So Kweon. Two-phase pseudo label densification for self-training based domain adaptation. In European conference on computer vision, pages 532–548. Springer, 2020.
* [45] Liangchen Song, Cheng Wang, Lefei Zhang, Bo Du, Qian Zhang, Chang Huang, and Xinggang Wang. Unsupervised domain adaptive re-identification: Theory and practice. Pattern Recognition, 102:107173, 2020.
* [46] Hao Tan and Mohit Bansal. Lxmert: Learning cross-modality encoder representations from transformers. arXiv preprint arXiv:1908.07490, 2019.
* [47] Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. arXiv preprint arXiv:1703.01780, 2017.
* [48] Marco Toldo, Umberto Michieli, Gianluca Agresti, and Pietro Zanuttigh. Unsupervised domain adaptation for mobile semantic segmentation based on cycle consistency and feature alignment. Image and Vision Computing, 95:103889, 2020.
* [49] Wilhelm Tranheden, Viktor Olsson, Juliano Pinto, and Lennart Svensson. Dacs: Domain adaptation via cross-domain mixed sampling. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 1379–1389, 2021.
* [50] Yi-Hsuan Tsai, Wei-Chih Hung, Samuel Schulter, Kihyuk Sohn, Ming-Hsuan Yang, and Manmohan Chandraker. Learning to adapt structured output space for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7472–7481, 2018.
* [51] Yi-Hsuan Tsai, Kihyuk Sohn, Samuel Schulter, and Manmohan Chandraker. Domain adaptation for structured output via discriminative patch representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1456–1465, 2019.
* [52] Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. Adversarial discriminative domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7167–7176, 2017.
* [53] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information processing systems, pages 5998–6008, 2017.
* [54] Tuan-Hung Vu, Himalaya Jain, Maxime Bucher, Matthieu Cord, and Patrick Pérez. Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2517–2526, 2019.
* [55] Bokun Wang, Yang Yang, Xing Xu, Alan Hanjalic, and Heng Tao Shen. Adversarial cross-modal retrieval. In Proceedings of the 25th ACM international conference on Multimedia, pages 154–162, 2017.
* [56] Yanchao Yang and Stefano Soatto. Fda: Fourier domain adaptation for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4085–4095, 2020.
* [57] Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412, 2017.
* [58] Pan Zhang, Bo Zhang, Ting Zhang, Dong Chen, Yong Wang, and Fang Wen. Prototypical pseudo label denoising and target structure learning for domain adaptive semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12414–12424, 2021.
* [59] Hang Zhao, Xavier Puig, Bolei Zhou, Sanja Fidler, and Antonio Torralba. Open vocabulary scene parsing. In Proceedings of the IEEE International Conference on Computer Vision, pages 2002–2010, 2017.
* [60] Qianyu Zhou, Zhengyang Feng, Qiqi Gu, Guangliang Cheng, Xuequan Lu, Jianping Shi, and Lizhuang Ma. Uncertainty-aware consistency regularization for cross-domain semantic segmentation. arXiv preprint arXiv:2004.08878, 2020.
* [61] Qianyu Zhou, Zhengyang Feng, Qiqi Gu, Jiangmiao Pang, Guangliang Cheng, Xuequan Lu, Jianping Shi, and Lizhuang Ma. Context-aware mixup for domain adaptive semantic segmentation. arXiv preprint arXiv:2108.03557, 2021.
* [62] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, pages 2223–2232, 2017.
* [63] Xiaojin Jerry Zhu. Semi-supervised learning literature survey. 2005\.
* [64] Yang Zou, Zhiding Yu, BVK Kumar, and Jinsong Wang. Unsupervised domain adaptation for semantic segmentation via class-balanced self-training. In Proceedings of the European conference on computer vision (ECCV), pages 289–305, 2018.
* [65] Yang Zou, Zhiding Yu, Xiaofeng Liu, BVK Kumar, and Jinsong Wang. Confidence regularized self-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5982–5991, 2019.
|
# On Computing Probabilistic Abductive Explanations
Yacine Izza<EMAIL_ADDRESS>
University of Toulouse, Toulouse, France
Monash University, Melbourne, Australia Xuanxiang Huang xuanxiang.huang@univ-
toulouse.fr
University of Toulouse, Toulouse, France Alexey Ignatiev
<EMAIL_ADDRESS>
Monash University, Melbourne, Australia Nina Narodytska<EMAIL_ADDRESS>
VMware Research, Palo Alto, CA, U.S.A. Martin C. Cooper<EMAIL_ADDRESS>
IRIT, UPS, Toulouse, France Joao Marques-Silva<EMAIL_ADDRESS>
IRIT, CNRS, Toulouse, France
###### Abstract
The most widely studied explainable AI (XAI) approaches are unsound. This is
the case with well-known model-agnostic explanation approaches, and it is also
the case with approaches based on saliency maps. One solution is to consider
intrinsic interpretability, which does not exhibit the drawback of
unsoundness. Unfortunately, intrinsic interpretability can display unwieldy
explanation redundancy. Formal explainability represents the alternative to
these non-rigorous approaches, with one example being PI-explanations.
Unfortunately, PI-explanations also exhibit important drawbacks, the most
visible of which is arguably their size. Recently, it has been observed that
the (absolute) rigor of PI-explanations can be traded off for a smaller
explanation size, by computing the so-called relevant sets. Given some
positive $\delta$, a set ${\mathcal{S}}$ of features is $\delta$-relevant if,
when the features in ${\mathcal{S}}$ are fixed, the probability of getting the
target class exceeds $\delta$. However, even for very simple classifiers, the
complexity of computing relevant sets of features is prohibitive, with the
decision problem being $\textnormal{NP}^{\textnormal{PP}}$-complete for
circuit-based classifiers. In contrast with earlier negative results, this
paper investigates practical approaches for computing relevant sets for a
number of widely used classifiers that include Decision Trees (DTs), Naive
Bayes Classifiers (NBCs), and several families of classifiers obtained from
propositional languages. Moreover, the paper shows that, in practice, and for
these families of classifiers, relevant sets are easy to compute. Furthermore,
the experiments confirm that succinct sets of relevant features can be
obtained for the families of classifiers considered.
###### Contents
1. 1 Introduction
2. 2 Preliminaries
1. 2.1 Logic Foundations
2. 2.2 Classification Problems
3. 2.3 Families of Classifiers
4. 2.4 Running Examples
5. 2.5 Formal Explainability
6. 2.6 $\delta$-Relevant Sets
3. 3 Relevant Sets – Probabilistic Abductive Explanations
1. 3.1 Definitions of Probabilistic AXp’s
2. 3.2 Computing Locally-Minimal PAXp’s
4. 4 Probabilistic Explanations for Decision Trees
1. 4.1 Path Probabilities for DTs
2. 4.2 Computing Locally-Minimal PAXp’s for DT’s
3. 4.3 Computing Minimum-Size PAXp’s for DTs
4. 4.4 Deciding Whether a Locally-Minimal PAXp is a Plain PAXp for DTs
5. 4.5 Instance-Based vs. Path-Based Explanations
5. 5 Probabilistic Explanations for Naive Bayes Classifiers
1. 5.1 Explaining NBCs in Polynomial Time
2. 5.2 Counting Models of XLCs
3. 5.3 Computing Locally-Minimal PAXp’s for NBCs
6. 6 Probabilistic Explanations for Other Families of Classifiers
1. 6.1 Propositional Classifiers
2. 6.2 Graph-Based Classifiers
7. 7 Experiments
1. 7.1 Case Study 1: Decision Trees
2. 7.2 Case Study 2: Naive Bayes Classifiers
3. 7.3 Case Study 3: Graph-Based Classifiers
8. 8 Related Work
9. 9 Conclusions
## 1 Introduction
The advances in Machine Learning (ML) in recent years motivated an ever
increasing range of practical applications of systems of Artificial
Intelligence (AI). Some uses of ML models are deemed _high-risk_ given the
impact that their operation can have on people [28]. (Other authors refer to
_high-stakes_ applications [84].) In some domains, with high-risk applications
representing one key example, the deployment of AI systems is premised on the
availability of mechanisms for explaining the often opaque operation of ML
models [28]. The need for explaining the opaque operation of ML models
motivated the emergence of eXplainable AI (XAI). Moreover, the rigor of
explanations is a cornerstone to delivering trustworthy AI [64]. For example,
rigorous explanations can be mathematically proven correct, and so a human
decision maker can independently validate that an explanation is (logically)
rigorous. The rigor of explanations is even more significant in high-risk AI
systems, where human decision makers must be certain that provided
explanations are sound.
Motivated by the importance of understanding the operation of blackbox ML
models, recent years have witnessed a growing interest in XAI [72, 33, 86, 87,
71, 85]. The best-known XAI approaches can be broadly categorized as either
model-agnostic methods, that include for example LIME [81], SHAP [60] and
Anchor [82], or intrinsic interpretability [84, 71], for which the explanation
is represented by the actual (interpretable) ML model. For specific ML models,
e.g. neural networks, there are dedicated explainability approaches, including
those based on saliency maps [91, 8]; these exhibit limitations similar to
model-agnostic approaches [1, 55, 99, 92]. Moreover, intrinsic
interpretability may not represent a viable option in some uses of AI systems.
Also, it has been shown, both in theory and in practice, that intrinsic
interpretable models (such as decision trees or sets) also require being
explained [45, 41, 46]. On the other hand, model-agnostic methods, even if
locally accurate, can produce explanations that are unsound [39], in addition
to displaying several other drawbacks [16, 93, 57, 25]. Unsound explanations
are hopeless whenever rigor is a key requirement; thus, model-agnostic
explanations ought not be used in high-risk settings. Indeed, it has been
reported [39] that an explanation ${\mathcal{X}}$ can be consistent with
different predicted classes. For example, for a bank loan application,
${\mathcal{X}}$ might be consistent with an approved loan application, but
also with a declined loan application. An explanation that is consistent with
both a declined and an approved loan applications offers no insight to why one
of the loan applications was declined. There have been recent efforts on
rigorous XAI approaches [89, 43, 22, 7, 61, 13, 32], most of which are based
on feature selection, namely the computation of so-called abductive
explanations (AXp’s). However, these efforts have mostly focused on the
scalability of computing rigorous explanations, with more recent work
investigating input distributions [32]. Nevertheless, another important
limitation of rigorous XAI approaches is the often unwieldy size of
explanations. Recent work studied probabilistic explanations, as a mechanism
to reduce the size of rigorous explanations [102, 101, 100]. Some approaches
for computing probabilistic explanations have extended model-agnostic
approaches [102], and so can suffer from unsoundness. Alternatively, more
rigorous approaches to computing probabilistic explanations have been shown to
be computationally hard, concretely hard for
$\textnormal{NP}^{\textnormal{PP}}$ [101, 100], and so are all but certain to
fall beyond the reach of modern automated reasoners.
This paper builds on recent work [101, 100] on computing rigorous
probabilistic explanations, and investigates their practical scalability.
However, instead of considering classifiers represented as boolean circuits
(as in [101]), the paper studies families of classifiers that are shown to be
amenable to the practical computation of relevant sets. As the paper shows,
such examples include decision trees (DTs), naive Bayes classifiers (NBCs),
but also several propositional classifiers, graph-based classifiers, and their
multi-valued variants 111This paper aggregates and extends recent preprints
that compute relevant sets for concrete families of classifiers [48, 50]..
The paper revisits the original definition of $\delta$-relevant set, and
considers different variants, in addition to the one proposed in earlier work
[101, 100], i.e. smallest relevant sets. These include subset-minimal relevant
sets, and locally-minimal relevant sets. Throughout the paper, relevant sets
will be referred to as _probabilistic abductive explanations_ (PAXp’s).
Whereas computing a smallest and a subset-minimal PAXp’s is shown to be in NP
for decision trees and different propositional classifiers, computing locally-
minimal PAXp’s is shown to be in P for decision trees and different
propositional classifiers, and in pseudo-polynomial time for naive Bayes
classifiers. Furthermore, the experiments confirm that locally-minimal PAXp’s
most often match subset-minimal PAXp’s. As a result, locally-minimal relevant
sets are shown to represent a practically efficient approach for computing (in
polynomial or pseudo-polynomial time) PAXp’s that are most often subset-
minimal.
Complexity of
PAXpFP$\textnormal{LmPAXp}(\textnormal{d-DDNF})$$\textnormal{LmPAXp}(\textnormal{GDF})$$\textnormal{LmPAXp}(\textnormal{SDD})$$\textnormal{LmPAXp}(\textnormal{DG})$$\textnormal{LmPAXp}(\textnormal{DT})$$\textbf{FP}^{\textbf{NP}}$$\textnormal{PAXp}(\textnormal{d-DDNF})$$\textnormal{PAXp}(\textnormal{GDF})$$\textnormal{PAXp}(\textnormal{SDD})$$\textnormal{PAXp}(\textnormal{DG})$$\textnormal{PAXp}(\textnormal{DT})$$\textbf{FP}^{\textbf{NP}^{\textbf{PP}}}$$\textnormal{PAXp}(\textnormal{BC})$$\textbf{FP}^{\textbf{pseudo-
NP}}$$\textnormal{PAXp}(\textnormal{NBC})$pseudo-
FP$\textnormal{LmPAXp}(\textnormal{NBC})$ Figure 1: Summary of results
The paper’s contributions are summarized in Figure 1. Figure 1 overviews
complexity class membership results for several families of classifiers, most
of which are established in this paper. For several families of classifiers,
the paper proves that computing one locally-minimal PAXp is either in
polynomial time or pseudo-polynomial time. Similarly, computing one subset-
minimal (and also one cardinality-minimal) PAXp is shown to be in NP (or
pseudo-NP) for several families of classifiers.
The paper is organized as follows. Section 2 introduces the definitions and
notation used throughout. Section 3 offers an overview of probabilistic
abductive explanations (or relevant sets). Section 4 investigates the
computation of relevant sets in the case of decision trees. Sections 5 and 6
replicate the same exercise, respectively in the case of naive Bayes
classifiers, and also for graph-based classifiers and classifiers based on
propositional languages. Section 7 presents experimental results on computing
relevant sets for the classifiers studied in the earlier sections. The paper
concludes in Section 9.
## 2 Preliminaries
#### Complexity classes.
The paper addresses a number of well-known classes of decision and function
problems, that include P, NP, $\textnormal{FP}^{\textnormal{NP}}$,
$\textnormal{NP}^{\textnormal{PP}}$, among others. The interested reader is
referred to a standard reference on the computational complexity [5].
### 2.1 Logic Foundations
#### Propositional logic.
The paper assumes the notation and definitions that are standard when
reasoning about the decision problem for propositional logic, i.e. the Boolean
Satisfiability (SAT) problem [12]. SAT is well-known to be an NP-complete [17]
decision problem. A propositional formula $\varphi$ is defined over a finite
set of Boolean variables $X=\\{x_{1},x_{2},\ldots,x_{n}\\}$. Formulas are most
often represented in _conjunctive normal form_ (CNF). A CNF formula is a
conjunction of clauses, a clause is a disjunction of literals, and a literal
is a variable ($x_{i}$) or its negation ($\neg{x_{i}}$). A term is a
conjunction of literals. Whenever convenient, a formula is viewed as a set of
sets of literals. A Boolean interpretation $\mu$ of a formula $\varphi$ is a
total mapping of $X$ to $\\{0,1\\}$ ($0$ corresponds to _False_ and $1$
corresponds to _True_). Interpretations can be extended to literals, clauses
and formulas with the usual semantics; hence we can refer to $\mu(l)$,
$\mu(\omega)$, $\mu(\varphi)$, to denote respectively the value of a literal,
clause and formula given an interpretation. Given a formula $\varphi$, $\mu$
is a _model_ of $\varphi$ if it makes $\varphi$ _True_ , i.e. $\mu(\phi)=1$. A
formula $\varphi$ is _satisfiable_ ($\varphi\operatorname*{\nvDash}\perp$) if
it admits a model, otherwise, it is _unsatisfiable_
($\varphi\operatorname*{\vDash}\perp$). Given two formulas $\varphi$ and
$\psi$, we say that $\varphi$ entails $\psi$ (denoted
$\varphi\operatorname*{\vDash}\psi$) if all models of $\varphi$ are also
models of $\psi$. $\varphi$ and $\psi$ are equivalent (denoted
$\varphi\equiv\psi$) if $\varphi\operatorname*{\vDash}\psi$ and
$\psi\operatorname*{\vDash}\varphi$.
#### First Order Logic (FOL) and SMT.
When necessary, the paper will consider the restriction of FOL to
Satisfiability Modulo Theories (SMT). These represent restricted (and often
decidable) fragments of FOL [10]. All the definitions above apply to SMT. A
SMT-solver reports whether a formula is satisfiable, and if so, may provide a
model of this satisfaction. (Other possible features include dynamic addition
and retraction of constraints, production of proofs, and optimization.)
### 2.2 Classification Problems
This paper considers classification problems, which are defined on a set of
features (or attributes) ${\mathcal{F}}=\\{1,\ldots,m\\}$ and a set of classes
${\mathcal{K}}=\\{c_{1},c_{2},\ldots,c_{K}\\}$. Each feature
$i\in{\mathcal{F}}$ takes values from a domain $\mathbb{D}_{i}$. In general,
domains can be categorical or ordinal, with values that can be boolean or
integer. (Although real-valued could be considered for some of the classifiers
studied in the paper, we opt not to specifically address real-valued
features.) Feature space is defined as
$\mathbb{F}=\mathbb{D}_{1}\times{\mathbb{D}_{2}}\times\ldots\times{\mathbb{D}_{m}}$;
$|\mathbb{F}|$ represents the total number of points in $\mathbb{F}$. For
boolean domains, $\mathbb{D}_{i}=\\{0,1\\}=\mathbb{B}$, $i=1,\ldots,m$, and
$\mathbb{F}=\mathbb{B}^{m}$. The notation $\mathbf{x}=(x_{1},\ldots,x_{m})$
denotes an arbitrary point in feature space, where each $x_{i}$ is a variable
taking values from $\mathbb{D}_{i}$. The set of variables associated with
features is $X=\\{x_{1},\ldots,x_{m}\\}$. Moreover, the notation
$\mathbf{v}=(v_{1},\ldots,v_{m})$ represents a specific point in feature
space, where each $v_{i}$ is a constant representing one concrete value from
$\mathbb{D}_{i}$. With respect to the set of classes ${\mathcal{K}}$, the size
of ${\mathcal{K}}$ is assumed to be finite; no additional restrictions are
imposed on ${\mathcal{K}}$. Nevertheless, with the goal of simplicity, the
paper considers examples where $|{\mathcal{K}}|=2$, concretely
${\mathcal{K}}=\\{\ominus,\oplus\\}$, or alternatively
${\mathcal{K}}=\\{0,1\\}$. An ML classifier $\mathbb{M}$ is characterized by a
(non-constant) _classification function_ $\kappa$ that maps feature space
$\mathbb{F}$ into the set of classes ${\mathcal{K}}$, i.e.
$\kappa:\mathbb{F}\to{\mathcal{K}}$. An _instance_ (or observation) denotes a
pair $(\mathbf{v},c)$, where $\mathbf{v}\in\mathbb{F}$ and
$c\in{\mathcal{K}}$, with $c=\kappa(\mathbf{v})$.
### 2.3 Families of Classifiers
Throughout the paper, a number of families of classifiers will be studied in
detail. These include decision trees [14, 80], naive Bayes classifiers [29],
graph-based classifiers [76, 77, 35] and classifiers based on propositional
languages [23, 21].
#### Decision trees.
A decision tree ${\mathcal{T}}=(V,E)$ is a directed acyclic graph, with
$V=\\{1,\ldots,|V|\\}$, having at most one path between every pair of nodes.
${\mathcal{T}}$ has a root node, characterized by having no incoming edges.
All other nodes have one incoming edge. We consider univariate decision trees
where each non-terminal node is associated with a single feature $x_{i}$. Each
edge is labeled with a literal, relating a feature (associated with the edge’s
starting node) with some values (or range of values) from the feature’s
domain. We will consider literals to be of the form $x_{i}\in\mathbb{E}_{i}$.
$x_{i}$ is a variable that denotes the value taken by feature $i$, whereas
$\mathbb{E}_{i}\subseteq\mathbb{D}_{i}$ is a subset of the domain of feature
$i\in{\mathcal{F}}$. The type of literals used to label the edges of a DT
allows the representation of the DTs generated by a wide range of decision
tree learners (e.g. [96]). The set of paths of ${\mathcal{T}}$ is denoted by
${\mathcal{R}}$. $\mathrm{\Phi}(R_{k})$ denotes the set of features associated
with path $R_{k}\in{\mathcal{R}}$, one per node in the tree, with repetitions
allowed. It is assumed that for any $\mathbf{v}\in\mathbb{F}$ there exists
_exactly_ one path in ${\mathcal{T}}$ that is consistent with $\mathbf{v}$. By
_consistent_ we mean that the literals associated with the path are satisfied
(or consistent) with the feature values in $\mathbf{v}$. Given $\mathbf{v}$,
the set of paths ${\mathcal{R}}$ is partitioned into ${\mathcal{P}}$ and
${\mathcal{Q}}$, such that each of the paths in ${\mathcal{P}}$ yields the
prediction $c=\kappa(\mathbf{v}$, whereas each of the paths in ${\mathcal{Q}}$
yields a prediction in ${\mathcal{K}}\setminus\\{c\\}$. For the purposes of
this paper, the path consistent with $\mathbf{v}$ as $P_{t}\in{\mathcal{P}}$,
i.e. the target path. (A more in-depth analysis of explaining decision trees
is available in [46].)
#### Naive Bayes Classifiers (NBCs).
An NBC [26] is a Bayesian Network model [29] characterized by strong
conditional independence assumptions among the features. Given some
observation $\mathbf{x}\in\mathbb{F}$, the predicted class is given by:
$\kappa(\mathbf{x})=\operatorname*{\mathsf{argmax}}\nolimits_{c\in{\mathcal{K}}}\left(\textnormal{Pr}(c|\mathbf{x})\right)$
(1)
Using Bayes’s theorem, $\textnormal{Pr}(c|\mathbf{x})$ can be computed as
follows:
$\textnormal{Pr}(c|\mathbf{x})=\nicefrac{{\textnormal{Pr}(c,\mathbf{x})}}{{\textnormal{Pr}(\mathbf{x})}}$.
In practice, we compute only the numerator of the fraction, since the
denominator $\textnormal{Pr}(\mathbf{x})$ is constant for every
$c\in{\mathcal{K}}$. Moreover, given the conditional mutual independency of
the features, we have:
$\textnormal{Pr}(c,\mathbf{x})=\textnormal{Pr}(c)\times\prod\nolimits_{i}\textnormal{Pr}(x_{i}|c)$
Furthermore, it is also common in practice to apply logarithmic
transformations on probabilities of $\textnormal{Pr}(c,\mathbf{x})$, thus
getting:
$\log\textnormal{Pr}(c,\mathbf{x})=\log{\textnormal{Pr}(c)}+\sum\nolimits_{i}\log{\textnormal{Pr}(x_{i}|c)}$
Therefore, (1) can be rewritten as follows:
$\kappa(\mathbf{x})=\operatorname*{\mathsf{argmax}}\nolimits_{c\in{\mathcal{K}}}\left(\log{\textnormal{Pr}(c)}+\sum\nolimits_{i}\log{\textnormal{Pr}(x_{i}|c)}\right)$
(2)
For simplicity, and following the notation used in earlier work [62],
$\operatorname*{lPr}$ denotes a logarithmic probability. Thus, we get:
$\kappa(\mathbf{x})=\operatorname*{\mathsf{argmax}}\nolimits_{c\in{\mathcal{K}}}\left(\operatorname*{lPr}(c)+\sum\nolimits_{i}\operatorname*{lPr}(x_{i}|c)\right)$
(3)
(Note that also for simplicity, it is common in practice to add a sufficiently
large positive constant $T$ to each probability, which allows using only
positive values.)
#### Graph-based classifiers.
The paper considers the class of graph-based classifiers (referred to as
Decision Graphs) studied in earlier work [35]. A Decision Graph (DG) is a
Directed Acyclic Graph (DAG) consisting of two types of nodes, non-terminal
nodes and terminal nodes, and has a single root node. Each non-terminal node
is labeled with a feature and has at least one child node, each terminal node
is labeled with a class and has no child node. Moreover, for an arbitrary non-
terminal node $p$ labeled with feature $i$, we use
$\mathbb{C}_{i}\subseteq\mathbb{D}_{i}$ to denote the set of values of feature
$i$ which are consistent with any path connecting the root to $p$. Each
outgoing edge of node $p$ represents a literal of the form
$x_{i}\in{\mathbb{E}_{i}}$ ($\mathbb{E}_{i}\neq\emptyset$) where
$\mathbb{E}_{i}\subseteq\mathbb{C}_{i}$. Furthermore, the following
constraints are imposed on DGs:
1. 1.
The literals associated with the outgoing edges of $p$ represent a partition
of $\mathbb{C}_{i}$.
2. 2.
Any path connecting the root node to a terminal node contains no inconsistent
literals.
3. 3.
No dead-ends.
A DG is _read-once_ if each feature tested at most once on any path. A DG is
_ordered_ , if features are tested in the same order on all paths. An Ordered
Multi-valued Decision Diagrams (OMDD) is a _read-once_ and _ordered_ DG such
that every $\mathbb{E}_{i}$ is a singleton, which means multiple-edges between
two nodes may exist. Furthermore, the OMDDs considered in this paper are
_reduced_ [53, 94], i.e.
1. 1.
No node $p$ such that all child nodes of $p$ are isomorphic; and
2. 2.
No isomorphic subgraphs.
When the domains of features are all boolean and the set of classes is binary,
then OMDDs corresponds to Ordered Binary Decision Diagrams (OBDDs) [15].
#### Propositional languages & classifiers.
For classification problems for which the feature space $\mathbb{F}$ is
restricted to $\mathbb{B}^{m}$, then boolean circuits can be used as binary
classifiers. Each boolean circuit is a sentence of some propositional
language. We briefly review some well-known propositional languages and
queries/transformations that these languages support in polynomial time.
The language _negation normal form_ (NNF) is the set of all directed acyclic
graphs, where each terminal node is labeled with either $\top$, $\bot$,
$x_{i}$ or $\neg{x_{i}}$, for $x_{i}\in{X}$. Each non-terminal node is labeled
with either $\land$ (or AND) or $\lor$ (or OR). The language _decomposable_
NNF (DNNF) [19, 23] is the set of all NNFs, where for every node labeled with
$\land$, $\alpha=\alpha_{1}\land\cdots\land\alpha_{k}$, no variables are
shared between the conjuncts $\alpha_{j}$. A _deterministic_ DNNF (d-DNNF)
[23, 73] is a DNNF, where for every node labeled with $\lor$,
$\beta=\beta_{1}\lor\cdots\lor\beta_{k}$, each pair $\beta_{p},\beta_{q}$,
with $p\not=q$, is inconsistent, i.e.
$\beta_{p}\land\beta_{q}\operatorname*{\vDash}\bot$. A _Smooth_ d-DNNF (sd-
DNNF) [20] is a d-DNNF, where for every node labeled with $\lor$,
$\beta=\beta_{1}\lor\cdots\lor\beta_{k}$, each pair $\beta_{p},\beta_{q}$ is
defined on the same set of variables. We focus in this paper on d-DNNF, but
for simplicity of algorithms, sd-DNNF is often considered. Furthermore,
_sentential decision diagrams_ (SDDs) [21, 98] represent a well-known subset
of the d-DNNF. (Furthermore, it should be noted that OBDD is a proper subset
of SDD.) SDDs are based on a strongly deterministic decomposition [21], which
is used to decompose a Boolean function into the form: $(p_{1}\land
s_{1})\lor\dots\lor(p_{n}\land s_{n})$, where each $p_{i}$ is called a prime
and each $s_{i}$ is called a sub (both primes and subs are sub-functions).
Furthermore, the process of decomposition is governed by a variable tree
(vtree) which stipulates the variable order [21].
The languages d-DNNF, sd-DNNF and SDD satisfy the query _polytime model
counting_ (CT), and the transformation _polytime conditioning_ (CD). Let
$\Delta$ represent a propositional formula and let $\rho$ denote a consistent
term ($\rho\operatorname*{\nvDash}\bot$). The _conditioning_ [23] of $\Delta$
on $\rho$, denoted $\Delta|_{\rho}$ is the formula obtained by replacing each
variable $x_{i}$ by $\top$ (resp. $\bot$) if $x_{i}$ (resp. $\neg{x_{i}}$) is
a positive (resp. negative) literal of $\rho$. A propositional language L
satisfies CT if there exists a polynomial-time algorithm that maps every
formula $\Delta$ from L into a non-negative integer denoting the number of
models of $\Delta$. L satisfies CD iff there exists a polynomial-time
algorithm that maps every formula $\Delta$ from L and every consistent term
$\rho$ into a formula in L that is logically equivalent to $\Delta|_{\rho}$.
There are additional queries and transformations of interest [23], but these
are beyond the goals of this paper. It is important to note that OMDD and OBDD
also satisfy CT and CD [23, 75].
### 2.4 Running Examples
#### Example Decision Tree.
Figure 2 shows the example DT used throughout the paper. This example DT also
illustrates the notation used to represent DTs.
BDT/.style= for tree= l=1.35cm,s sep=1.5cm, if n children=0circle, draw, edge=
my edge , if n=1 edge+=0 my edge, , , BDT [$x_{1}$, label=1 [$x_{2}$, s
sep=0.75cm, label=2, edge label=node[near start,left,xshift=-0.75pt]
$\in\\{1\\}$ [$\ominus$, label=[xshift=0.25ex,yshift=1.875ex]4, edge
label=node[near start,left,xshift=-0.5pt] $\in\\{1\\}$] [$\oplus$,
label=[xshift=0.25ex,yshift=1.875ex]5, edge label=node[near
start,right,xshift=-1pt] $\in\\{2..4\\}$] ] [$x_{2}$, s sep=0.75cm, label=3,
edge label=node[near start,right,xshift=1pt] $\in\\{2..4\\}$ [$\oplus$,
label=[xshift=0.25ex,yshift=1.875ex]6, edge label=node[near
start,left,xshift=-1pt] $\in\\{1\\}$] [$x_{3}$, label=7, edge label=node[near
start,right,xshift=-1pt] $\in\\{2..4\\}$ [$\ominus$,
label=[xshift=0.25ex,yshift=1.875ex]8, edge
label=node[pos=0.6,left,xshift=-1pt] $\in\\{1\\}$] [$\oplus$,
label=[xshift=0.25ex,yshift=1.875ex]9, edge
label=node[pos=0.6,right,xshift=0.25pt] $\in\\{2\\}$] ] ] ]
(a) Decision tree
Paths
---
$P_{1}=\langle 1,2,5\rangle$
$P_{2}=\langle 1,3,6\rangle$
$P_{3}=\langle 1,3,7,9\rangle$
$Q_{1}=\langle 1,2,4\rangle$
$Q_{2}=\langle 1,3,7,8\rangle$
(b) Paths in the DT
Figure 2: Example DT.
The set of paths ${\mathcal{R}}$ is partitioned into two sets ${\mathcal{P}}$
and ${\mathcal{Q}}$, such that the paths in
${\mathcal{P}}=\\{P_{1},P_{2},P_{3}\\}$ yield a prediction of $\oplus$, and
such that the paths in ${\mathcal{Q}}=\\{Q_{1},Q_{2}\\}$ yield a prediction of
$\ominus$. (In general, ${\mathcal{P}}$ denotes the paths with prediction
$c\in{\mathcal{K}}$, and ${\mathcal{Q}}$ denotes the paths with prediction
other than $c$, i.e. any class in ${\mathcal{K}}\setminus\\{c\\}$.)
#### Example Naive Bayes Classifier.
Consider the NBC depicted graphically in Figure 3 222This example of an NBC is
adapted from [62], and it was first studied in [9, Ch.10]..
$G$$R_{2}$$R_{3}$$R_{1}$$R_{4}$$R_{5}$ $G$ | $\textnormal{Pr}(G)$
---|---
$\ominus$ | 0.90
$G$ | $\textnormal{Pr}(R_{1}|G)$
---|---
$\oplus$ | 0.95
$\ominus$ | 0.03
$G$ | $\textnormal{Pr}(R_{2}|G)$
---|---
$\oplus$ | 0.05
$\ominus$ | 0.95
$G$ | $\textnormal{Pr}(R_{3}|G)$
---|---
$\oplus$ | 0.02
$\ominus$ | 0.34
$G$ | $\textnormal{Pr}(R_{4}|G)$
---|---
$\oplus$ | 0.20
$\ominus$ | 0.75
$G$ | $\textnormal{Pr}(R_{5}|G)$
---|---
$\oplus$ | 0.95
$\ominus$ | 0.03
Figure 3: Example NBC.
The features are the boolean random variables $R_{1}$, $R_{2}$, $R_{3}$,
$R_{4}$ and $R_{5}$. Each $R_{i}$ can take values $\mathbf{t}$ or $\mathbf{f}$
denoting, respectively, whether a listener likes or not that radio station.
The boolean random variable $G$ corresponds to an age class: the target class
$\oplus$ denotes the prediction that the listener is young and $\ominus$
denotes the prediction that the listener is old. Thus,
${\mathcal{K}}=\\{\ominus,\oplus\\}$. Let us consider
$\mathbf{v}=(R_{1},R_{2},R_{3},R_{4},R_{5})=(\mathbf{t},\mathbf{f},\mathbf{f},\mathbf{f},\mathbf{t})$.
We associate $r_{i}$ to each literal ($R_{i}=\mathbf{t}$) and $\neg{r_{i}}$ to
literals ($R_{i}=\mathbf{f}$). Using (3), we get the values shown in Figure 4.
(Note that to use positive values, we added $T=+4$ to each
$\operatorname*{lPr}(\cdot)$.). As can be seen by comparing the values of
$\operatorname*{lPr}(\oplus|\mathbf{v})$ and
$\operatorname*{lPr}(\ominus|\mathbf{v})$, the classifier will predict
$\oplus$.
| $\textnormal{Pr}(\oplus)$ | $\textnormal{Pr}(r_{1}|\oplus)$ | $\textnormal{Pr}(\neg{r_{2}}|\oplus)$ | $\textnormal{Pr}(\neg{r_{3}}|\oplus)$ | $\textnormal{Pr}(\neg{r_{4}}|\oplus)$ | $\textnormal{Pr}(r_{5}|\oplus)$ | $\operatorname*{lPr}(\oplus|\mathbf{v})$
---|---|---|---|---|---|---|---
$\textnormal{Pr}(\cdot)$ | 0.10 | 0.95 | 0.95 | 0.98 | 0.80 | 0.95 |
$\operatorname*{lPr}(\cdot)$ | 1.70 | 3.95 | 3.95 | 3.98 | 3.78 | 3.95 | 21.31
(a) Computing $\operatorname*{lPr}(\oplus|\mathbf{v})$
| $\textnormal{Pr}({\ominus})$ | $\textnormal{Pr}(r_{1}|{\ominus})$ | $\textnormal{Pr}(\neg{r_{2}}|{\ominus})$ | $\textnormal{Pr}(\neg{r_{3}}|{\ominus})$ | $\textnormal{Pr}(\neg{r_{4}}|{\ominus})$ | $\textnormal{Pr}(r_{5}|{\ominus})$ | $\operatorname*{lPr}(\ominus|\mathbf{v})$
---|---|---|---|---|---|---|---
$\textnormal{Pr}(\cdot)$ | 0.90 | 0.03 | 0.05 | 0.66 | 0.25 | 0.03 |
$\operatorname*{lPr}(\cdot)$ | 3.89 | 0.49 | 1.00 | 3.58 | 2.61 | 0.49 | 12.06
(b) Computing $\operatorname*{lPr}(\ominus|\mathbf{v})$
Figure 4: Deciding prediction for
$\mathbf{v}=(\mathbf{t},\mathbf{f},\mathbf{f},\mathbf{f},\mathbf{t})$. (Note
that to use positive values, $T=+4$ was added to each
$\operatorname*{lPr}(\cdot)$.)
#### Example graph-based classifiers.
Figure 5 shows two examples of graph-based classifiers. Figure 5(a) shows an
OBDD, and Figure 5(b) an OMDD. Figure 5(b) represents a function defined on
${\mathcal{F}}=\\{1,2,3\\}$ and ${\mathcal{K}}=\\{\ominus,\oplus,\otimes\\}$,
with the domains of features being $\mathbb{D}_{1}=\\{0,1\\}$,
$\mathbb{D}_{2}=\mathbb{D}_{3}=\\{0,1,2\\}$. If we consider the instance
$\mathbf{v}=\\{1,1,2\\}$, the classifier predicts the class $\otimes$.
$x_{1}$$x_{2}$$x_{3}$$\ominus$$\oplus$ 1 2 3 4 5 010101
(a) An OBDD
$x_{1}$$x_{2}$$x_{3}$$x_{3}$$\oplus$$\ominus$$\otimes$ 1 2 3 4 5 6 7
01012201012
(b) An OMDD
Figure 5: Example DD.
### 2.5 Formal Explainability
In contrast with well-known model-agnostic approaches to XAI [81, 60, 82, 33],
formal explanations are model-precise, i.e. their definition reflects the
model’s computed function.
#### Abductive explanations.
Prime implicant (PI) explanations [89] denote a minimal set of literals
(relating a feature value $x_{i}$ and a constant $v_{i}\in\mathbb{D}_{i}$)
that are sufficient for the prediction. PI-explanations are related with
abduction, and so are also referred to as abductive explanations (AXp’s)
[43]333PI-explanations were first proposed in the context of boolean
classifiers based on restricted bayesian networks [89]. Independent work [43]
studied PI-explanations in the case of more general classification functions,
i.e. not necessarily boolean, and related instead explanations with abduction.
This paper follows the formalizations used in more recent work [62, 49, 41,
63, 35, 18, 34, 40, 64]. . Formally, given
$\mathbf{v}=(v_{1},\ldots,v_{m})\in\mathbb{F}$ with $\kappa(\mathbf{v})=c$, a
set of features ${\mathcal{X}}\subseteq{\mathcal{F}}$ is a _weak abductive
explanation_ [18] (or weak AXp) if the following predicate holds true444Each
predicate associated with a given concept will be noted in sans-serif
letterform. When referring to the same concept in the text, the same acronym
will be used, but in standard letterform. For example, the predicate name
$\mathsf{AXp}$ will be used in logic statements, and the acronym AXp will be
used throughout the text.:
$\begin{array}[]{lcr}\mathsf{WeakAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)&\>{:=}&\forall(\mathbf{x}\in\mathbb{F}).\left[\bigwedge\nolimits_{i\in{{\mathcal{X}}}}(x_{i}=v_{i})\right]\operatorname*{\rightarrow}(\kappa(\mathbf{x})=c)\end{array}$
(4)
Moreover, a set of features ${\mathcal{X}}\subseteq{\mathcal{F}}$ is an
_abductive explanation_ (or (plain) AXp) if the following predicate holds
true:
$\displaystyle\mathsf{AXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)\quad{:=}\quad$
$\displaystyle\mathsf{WeakAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)\leavevmode\nobreak\
\land$
$\displaystyle\forall({\mathcal{X}}^{\prime}\subsetneq{\mathcal{X}}).\neg\mathsf{WeakAXp}({\mathcal{X}}^{\prime};\mathbb{F},\kappa,\mathbf{v},c)$
(5)
Clearly, an AXp is any weak AXp that is subset-minimal (or irrreducible). It
is straightforward to observe that the definition of predicate
$\mathsf{WeakAXp}$ is monotone, and so an AXp can instead be defined as
follows:
$\displaystyle\mathsf{AXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)\quad{:=}\quad$
$\displaystyle\mathsf{WeakAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)\leavevmode\nobreak\
\land$
$\displaystyle\forall(j\in{\mathcal{X}}).\neg\mathsf{WeakAXp}({\mathcal{X}}\setminus\\{j\\};\mathbb{F},\kappa,\mathbf{v},c)$
(6)
This alternative equivalent definition of abductive explanation is at the core
of most algorithms for computing one AXp. (Throughout the paper, we will drop
the parameterization associated with each predicate, and so we will write
$\mathsf{AXp}({\mathcal{X}})$ instead of
$\mathsf{AXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)$, when the
parameters are clear from the context.)
Figure 6: A schematic representation of relationships between different types
of explanations. $E_{1}({\mathcal{X}};\ldots)\Rightarrow
E_{2}({\mathcal{X}};\ldots)$ means that if ${\mathcal{X}}$ is an $E_{1}$
explanation then ${\mathcal{X}}$ must be an $E_{2}$ explanation.
###### Example 1.
The computation of (weak) AXp’s is illustrated with the DT from Figure 2(a).
The instance considered throughout is
$\mathbf{v}=(v_{1},v_{2},v_{3})=(4,4,2)$, with $c=\kappa(\mathbf{v})=\oplus$.
The point $\mathbf{v}$ is consistent with $P_{3}$, and
$\mathrm{\Phi}(P_{3})=\\{1,2,3\\}$. Table 1 (columns 1 to 4) analyzes three
sets of features $\\{1,2,3\\}$, $\\{1,3\\}$ and $\\{3\\}$ in terms of being a
weak AXp or an AXp. The decision on whether each set is a weak AXp or an AXp
can be obtained by analyzing all the 32 points in feature space, or by using
an off-the-shelf algorithm. (The analysis of all points in feature space is
ommited for brevity.)
${\mathcal{S}}$ | ${\mathcal{U}}$ | $\mathsf{WeakAXp}$? | $\mathsf{AXp}$? | $\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c|(\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}}))$ | $\mathsf{WeakPAXp}$? | $\mathsf{PAXp}$? | $\\#({\mathcal{S}})$ | $\\#(P_{1})$ | $\\#(P_{2})$ | $\\#(P_{3})$ | $\\#(Q_{1})$ | $\\#(Q_{2})$
---|---|---|---|---|---|---|---|---|---|---|---|---
$\\{1,2,3\\}$ | $\emptyset$ | Yes | No | $1\geq\delta$ | Yes | No | 1 | 0 | 0 | 1 | 0 | 0
$\\{1,3\\}$ | $\\{2\\}$ | Yes | Yes | $1\geq\delta$ | Yes | No | 4 | 0 | 1 | 3 | 0 | 0
$\\{3\\}$ | $\\{1,2\\}$ | No | – | $\nicefrac{{15}}{{16}}=0.9375\geq\delta$ | Yes | Yes | 16 | 3 | 3 | 9 | 1 | 0
Table 1: Examples of sets of fixed features given $\mathbf{v}=(4,4,2)$ and
$\delta=0.93$
It is apparent that (4), (2.5), and (2.5) can be viewed as representing a
(logic) _rule_ of the form:
$\textbf{IF\leavevmode\nobreak\ \leavevmode\nobreak\
}\left[\land_{i\in{\mathcal{X}}}(x_{i}=v_{i})\right]\textbf{\leavevmode\nobreak\
\leavevmode\nobreak\ THEN\leavevmode\nobreak\ \leavevmode\nobreak\
}\left[\kappa(\mathbf{x})=c\right]$ (7)
Unless otherwise noted, this interpretation of explanations will be assumed
throughout the paper.
Abductive explanations can be viewed as answering a ‘Why?’ question, i.e. why
is some prediction made given some point in feature space. A different view of
explanations is a contrastive explanation [70], which answers a ‘Why Not?’
question, i.e. which features can be changed to change the prediction. The
formalization of contrastive explanations revealed a minimal hitting set
duality relationship between abductive and contrastive explanations [42]. The
paper does not detail further contrastive explanations, as the focus is solely
on abductive explanations.
Figure 6 shows relationships between different classes of explanations that we
investigate in the paper.
$\mathsf{AXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)$ and
$\mathsf{WeakAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c)$ are
deterministic classes of explanations. These classes are shown in yellow
boxes. Other classes of explanations, shown in blue boxes, represent
probabilistic counterparts of AXp’s. We study them in the following sections.
#### Progress in formal explainability.
The introduction of abductive explanations [89, 43] also revealed practical
limitations in the case of bayesian network classifiers [89, 90] and neural
networks [43]. However, since then there has been a stream of results, that
demonstrate the practical applicability of formal explainability. These
results can be broadly organized as follows (a more detailed overview is
available in [64]):
* •
Tractable explanations.
Recent work showed that computing one explanation is tractable for naive Bayes
classifiers [62], decision trees [45, 35, 46], graph-based classifiers [35],
monotonic classifiers [63, 18], and classifiers represented with well-known
classes of propositional languages [34]. Additional tractability results were
obtained in [18].
* •
Efficient explanations.
For some other families of classifiers, recent work showed that computing one
explanation is computationally hard, but it is nevertheless efficient in
practice. This is the case with decision lists and sets [41], random forests
[49], and tree ensembles in general [43, 39, 40].
* •
Explainability queries.
There has been interest in understanding the complexity of answering different
queries related with reasoning about explainability [7, 36, 35, 6]. For
example, the feature membership problem is the decision problem of deciding
whether some (possibly sensitive) feature occurs in some explanation. Although
computationally hard in general [35], it has been shown to be solved
efficiently in theory and in practice for specific families of classifiers
[35, 37]. Queries related with enumeration of explanations have been
extensively studied [62, 42, 41, 63, 35, 34, 40].
* •
Properties of explanations.
A number of works studied the connections between explanations and robustness
[44], and connections between different types of explanations [42].
Despite the observed progress, formal explainability still faces several
important challenges. First, for some widely used families of classifiers,
e.g. neural networks, formal explainability does not scale in practice [43].
Second, input distributions are not taken into account, since these are not
readily available. There is however recent work on accounting for input
constraints [18, 32, 104]. Third, the size of explanations may exceed the
cognitive limits of human decision makers [69], and computing smallest
explanations does not offer a computationally realistic alternative [43].
Recent work studied $\delta$-relevant sets [101, 100], and these are also the
focus of this paper.
Finally, we note that there have been different approaches to formal
explainability based on the study of the formal logic or the axiomatics of
explainers [103, 2, 59]. This paper studies exclusively those approaches for
which there is practical supporting evidence of observed progress, as attested
above.
### 2.6 $\delta$-Relevant Sets
$\delta$-relevant sets were proposed in more recent work [101, 100] as a
generalized formalization of PI-explanations (or AXp’s). $\delta$-relevant
sets can be viewed as _probabilistic_ PIs [100], with $\mathsf{AXp}$’s
representing a special case of $\delta$-relevant sets where $\delta=1$, i.e.
probabilistic PIs that are actual PIs. We briefly overview the definitions
related with relevant sets. The assumptions regarding the probabilities of
logical propositions are those made in earlier work [101, 100]. Let
$\textnormal{Pr}_{\mathbf{x}}(A(\mathbf{x}))$ denote the probability of some
proposition $A$ defined on the vector of variables
$\mathbf{x}=(x_{1},\ldots,x_{m})$, i.e.
$\begin{array}[]{rcl}\textnormal{Pr}_{\mathbf{x}}(A(\mathbf{x}))&=&\frac{|\\{\mathbf{x}\in\mathbb{F}:A(\mathbf{x})=1\\}|}{|\\{\mathbf{x}\in\mathbb{F}\\}|}\\\\[8.0pt]
\textnormal{Pr}_{\mathbf{x}}(A(\mathbf{x})\,|\,B(\mathbf{x}))&=&\frac{|\\{\mathbf{x}\in\mathbb{F}:A(\mathbf{x})=1\land{B(\mathbf{x})=1}\\}|}{|\\{\mathbf{x}\in\mathbb{F}:B(\mathbf{x})=1\\}|}\end{array}$
(8)
(Similar to earlier work, it is assumed that the features are independent and
uniformly distributed [101]. Moreover, the definitions above can be adapted in
case some of the features are real-valued. As noted earlier, the present paper
studies only non-continuous features.)
###### Definition 1 ($\delta$-relevant set [101]).
Consider $\kappa:\mathbb{B}^{m}\to{\mathcal{K}}=\mathbb{B}$,
$\mathbf{v}\in\mathbb{B}^{m}$, $\kappa(\mathbf{v})=c\in\mathbb{B}$, and
$\delta\in[0,1]$. ${\mathcal{S}}\subseteq{\mathcal{F}}$ is a $\delta$-relevant
set for $\kappa$ and $\mathbf{v}$ if,
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})\geq\delta$
(9)
(Where the restriction of $\mathbf{x}$ to the variables with indices in
${\mathcal{S}}$ is represented by
$\mathbf{x}_{{\mathcal{S}}}=(x_{i})_{i\in{\mathcal{S}}}$. Concretely, the
notation $\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}}$ represents
the constraint $\land_{i\in{\mathcal{S}}}x_{i}=v_{i}$.)
(Moreover, observe that
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})$
is often referred to as the _precision_ of ${\mathcal{S}}$ [82, 74].) Thus, a
$\delta$-relevant set represents a set of features which, if fixed to some
pre-defined value (taken from a reference vector $\mathbf{v}$), ensures that
the probability of the prediction being the same as the one for $\mathbf{v}$
is no less than $\delta$.
###### Definition 2 (Min-$\delta$-relevant set).
Given $\kappa$, $\mathbf{v}\in\mathbb{B}^{m}$, and $\delta\in[0,1]$, find the
smallest $k$, such that there exists ${\mathcal{S}}\subseteq{\mathcal{F}}$,
with $|{\mathcal{S}}|={k}$, and ${\mathcal{S}}$ is a $\delta$-relevant set for
$\kappa$ and $\mathbf{v}$.
With the goal of proving the computational complexity of finding a minimum-
size set of features that is a $\delta$-relevant set, earlier work [101]
restricted the definition to the case where $\kappa$ is represented as a
boolean circuit. (Boolean circuits were restricted to propositional formulas
defined using the operators $\lor$, $\land$ and $\neg$, and using a set of
variables representing the inputs; this explains the choice of _inputs_ over
_sets_ in earlier work [101].) The main complexity result from earlier work is
that the computation of $\delta$-relevant sets is hard for
$\textnormal{NP}^{\textnormal{PP}}$ [101]. Hence, as noted in earlier work
[101, 100], it is unlikely that exact computation of $\delta$-relevant sets
will be practically feasible.
## 3 Relevant Sets – Probabilistic Abductive Explanations
In contrast with Min-$\delta$-relevant sets, whose focus are smallest-size
explanations, this section investigates alternative definitions of relevant
sets (which we will also and indistinguishably refer to as _probabilistic
abductive explanations_).
### 3.1 Definitions of Probabilistic AXp’s
Conceptually, Definition 1 does not need to impose a restriction on the
classifier considered (although this is done in earlier work [101]), i.e. the
logical representation of $\kappa$ need not be a boolean circuit. As a result,
Definition 1 can also be considered in the case of multi-class classifiers
defined on categorical or ordinal (non-continuous) features.
Given the above, a _weak probabilistic_ AXp (or weak PAXp) is a pick of fixed
features for which the conditional probability of predicting the correct class
$c$ exceeds $\delta$, given $c=\kappa(\mathbf{v})$. Thus,
${\mathcal{X}}\subseteq{\mathcal{F}}$ is a weak PAXp if the following
predicate holds true,
$\displaystyle\mathsf{WeakPAXp}$
$\displaystyle({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)$
$\displaystyle:=\,\>$
$\displaystyle\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})\geq\delta$
(10) $\displaystyle:=\,\>$
$\displaystyle\frac{|\\{\mathbf{x}\in\mathbb{F}:\kappa(\mathbf{x})=c\land(\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})\\}|}{|\\{\mathbf{x}\in\mathbb{F}:(\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})\\}|}\geq\delta$
which means that the fraction of the number of points predicting the target
class and consistent with the fixed features (represented by ${\mathcal{X}}$),
given the total number of points in feature space consistent with the fixed
features, must exceed $\delta$. (Observe that the difference to (1) is solely
that features and classes are no longer required to be boolean. Hence, weak
PAXp’s can be viewed as generalized $\delta$-relevant sets.) Moreover, a set
${\mathcal{X}}\subseteq{\mathcal{F}}$ is a _probabilistic_ AXp (or (plain)
PAXp) if the following predicate holds true,
$\displaystyle\mathsf{PAXp}$
$\displaystyle({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)\>:=$
$\displaystyle\mathsf{WeakPAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)\>\>\land$
(11)
$\displaystyle\forall({\mathcal{X}}^{\prime}\subsetneq{\mathcal{X}}).\neg\mathsf{WeakPAXp}({\mathcal{X}}^{\prime};\mathbb{F},\kappa,\mathbf{v},c,\delta)$
Thus, ${\mathcal{X}}\subseteq{\mathcal{F}}$ is a PAXp if it is a weak PAXp
that is also subset-minimal,
As can be observed, the definition of weak PAXp (see (3.1)) does not guarantee
monotonicity. In turn, this makes the computation of (subset-minimal) PAXp’s
harder. With the purpose of identifiying classes of weak PAXp’s that are
easier to compute, it will be convenient to study _locally-minimal_ PAXp’s. A
set of features ${\mathcal{X}}\subseteq{\mathcal{F}}$ is a locally-minimal
PAXp if,
$\displaystyle\mathsf{LmPAXp}$
$\displaystyle({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)\>:=$
$\displaystyle\mathsf{WeakPAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)\>\>\land$
(12)
$\displaystyle\forall(j\in{\mathcal{X}}).\neg\mathsf{WeakPAXp}({\mathcal{X}}\setminus\\{j\\};\mathbb{F},\kappa,\mathbf{v},c,\delta)$
As observed earlier in Section 2.5, because the predicate $\mathsf{WeakAXp}$
is monotone, subset-minimal AXp’s match locally-minimal AXp’s. An important
practical consequence is that most algorithms for computing one subset-minimal
AXp, will instead compute a locally-minimal AXp, since these will be the same.
Nevertheless, a critical observation is that in the case of probabilistic
AXp’s (see (3.1)), the predicate $\mathsf{WeakPAXp}$ is _not_ monotone. Thus,
there can exist locally-minimal PAXp’s that are not subset-minimal PAXp’s. (As
shown in the experiments, computed locally minimal APXp’s are most often
PAXp’s. However, exceptions do exist, even though these are rarely observed).
Furthermore, the fact that a set of features
${\mathcal{X}}\subseteq{\mathcal{F}}$ may satisfy (3.1) but not (3.1) imposes
that subset-minimal PAXp’s must be computed by using (3.1); as shown later,
this requires more complex algorithms.
Finally, minimum-size PAXp’s (or a smallest PAXp’s) generalize
Min-$\delta$-relevant sets in Definition 2. A set of features
${\mathcal{X}}\subseteq{\mathcal{F}}$ is a minimum-size AXp if,
$\displaystyle\mathsf{MinPAXp}$
$\displaystyle({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)\>:=$
$\displaystyle\mathsf{WeakPAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)\>\>\land$
(13)
$\displaystyle\forall({\mathcal{X}}^{\prime}\subseteq{\mathcal{F}}).\left[(|{\mathcal{X}}^{\prime}|<|{\mathcal{X}}|)\operatorname*{\rightarrow}\neg\mathsf{WeakPAXp}({\mathcal{X}}^{\prime};\mathbb{F},\kappa,\mathbf{v},c,\delta)\right]$
(As stated earlier, throughout the paper, we will drop the parameterization
associated with each predicate, and so we will write
$\mathsf{PAXp}({\mathcal{X}})$ instead of
$\mathsf{PAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)$, when the
parameters are clear from the context. Although the parameterization on
$\delta$ is paramount, we opt instead for simpler notation.)
###### Example 2.
The computation of (probabilistic) AXp’s is illustrated with the DT from
Figure 2. The instance considered throughout is
$\mathbf{v}=(v_{1},v_{2},v_{3})=(4,4,2)$, with $c=\kappa(\mathbf{v})=\oplus$.
Clearly, $\mathbf{v}$ is consistent with $P_{3}$. The goal is to compute a
$\delta$-relevant set given $\delta=0.93$. Let $\\#(R_{k})$ denote the number
of points in feature space that are consistent with path $R_{k}$. Moreover,
let $\\#({\mathcal{X}})$ denote the total number of points in feature space
that are consistent with the set of _fixed_ features
${\mathcal{X}}\in{\mathcal{F}}$. Table 1 summarizes the computation of
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c|\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})$
for different sets ${\mathcal{S}}$. The table also includes information on
whether each set is a weak AXp, an AXp, a weak PAXp, or a PAXp. The set
$\\{1,3\\}$ represents an AXp, since for any point consistent with the
assignments $x_{1}=4$ and $x_{3}=2$, the prediction is $\oplus$. However, by
setting ${\mathcal{S}}=\\{3\\}$, the probability of predicting $\oplus$ given
a point consistent with $x_{3}=2$ still exceeds $\delta$, since
$\nicefrac{{15}}{{16}}=93.75\%$. Hence, $\\{3\\}$ is a PAXp for
$\mathbf{v}=(4,4,2)$ when $\delta=0.93$.
#### Properties of locally-minimal PAXp’s.
Let ${\mathcal{X}}$ denote an AXp. Clearly, ${\mathcal{X}}$ is also a PAXp.
Then, for any locally-minimal ${\mathcal{A}}\subseteq{\mathcal{X}}$ (i.e.
${\mathcal{A}}$ is computed using ${\mathcal{X}}$ as a seed), we have the
following properties, which follow from the definition:
1. 1.
${\mathcal{A}}\subseteq{\mathcal{X}}$ (by hypothesis);
2. 2.
${\mathcal{A}}$ is a weak PAXp (by definition); and
3. 3.
There exists at least one PAXp ${\mathcal{E}}$ such that
${\mathcal{E}}\subseteq{\mathcal{A}}$.
Thus, given some AXp ${\mathcal{X}}$, we can compute a locally-minimal PAXp
${\mathcal{A}}$ that is both a subset of ${\mathcal{X}}$ and a a superset of
some PAXp, and such that ${\mathcal{A}}$ exhibits the strong probabilistic
properties of relevant sets. Although any locally-minimal PAXp is a subset of
a weak AXp, there can exist locally-minimal PAXp’s that are not subsets of
some (plain) AXp.
### 3.2 Computing Locally-Minimal PAXp’s
Algorithm 1 shows one approach for computing a locally-minimal PAXp555This
simple algorithm is often referred to as the deletion-based algorithm, namely
in settings related with solving function problems in propositional logic and
constraint programming [66]. However, the same general algorithm can be traced
at least to the work of Valiant [97], and some authors [51] argue that it is
implicit in works from the $\text{19}^{\text{th}}$ century [68]. . As shown,
to compute one locally-minimal PAXp, one starts from
${\mathcal{F}}=\\{1,\ldots,m\\}$ and iterately removes features while it is
safe to do so, i.e. while (3.1) holds for the resulting set. Beside Algorithm
1, one could consider for example variants of the QuickXplain [52] and the
Progression [65, 66] algorithms. Both of which also allow computing preferred
(locally-minimal) sets subject to anti-lexicographic preferences [52, 67].
Furthermore, we note that the same algorithms (i.e. Deletion, Progression and
QuickXplain, among others) can also be used for computing one AXp. Moreover,
observe that these algorithms can also be applied to _any classifier_ with
respect to which we seek to compute one locally-minimal PAXp. Furthermore,
another simple observation is that explanations can be enumerated by
exploiting hitting set dualization [42], e.g. using a MARCO-like algorithm
[58].
Input: Features $\\{1,\ldots,m\\}$; feature space $\mathbb{F}$, classifier
$\kappa$, instance $(\mathbf{v},c)$, threshold $\delta$
Output: Locally-minimal PAXp ${\mathcal{S}}$
1:procedure
$\mathsf{findLmPAXp}$($\\{1,\dots,m\\};\mathbb{F},\kappa,\mathbf{v},c,\delta$)
2: ${\mathcal{S}}\leftarrow\\{1,\ldots,m\\}$
3: for $i\in\\{1,\ldots,m\\}$ do
4: if
$\mathsf{WeakPAXp}({\mathcal{S}}\setminus\\{i\\};\mathbb{F},\kappa,\mathbf{v},c,\delta)$
then
5: ${\mathcal{S}}\leftarrow{\mathcal{S}}\setminus\\{i\\}$
6: return ${\mathcal{S}}$
Algorithm 1 Computing one locally-minimal PAXp
#### Practically efficient computation of relevant sets.
Further to the computation of locally-minimal PAXp’s, the next few sections
show that the computation of relevant sets (PAXp’s) can be achieved
efficiently in practice, for several families of classifiers. Concretely, over
the next few sections we analyze decision trees, naive Bayes classifiers, but
also several families of propositional and graph-based classifiers, studied in
recent work [35, 34].
## 4 Probabilistic Explanations for Decision Trees
This section shows that the problem of deciding whether a set
${\mathcal{X}}\subseteq{\mathcal{F}}$ is a PAXp is in NP when $\kappa$ is
represented by a decision tree666As noted earlier, and for simplicity, the
paper considers the case of non-continuous features. However, in the case of
DTs, the results generalize to continuous features.. As a result, a minimum-
size PAXp can be computed with at most a logarithmic number of calls to an NP
oracle. (This is a consequence that optimizing a linear cost function, subject
to a set of constraints for which deciding satisfiability is in NP, can be
achieved with a logarithmic number of calls to a NP oracle.) An SMT
formulation of the problem is proposed and the empirical evaluation confirms
its practical effectiveness. This section also proposes a polynomial time
algorithm to compute one locally-minimal PAXp, thus offering an alternative to
computing one PAXp. The results in Section 7 confirm that in practice computed
locally-minimal PAXp’s are often subset-minimal, i.e. a locally-minimal PAXp
actually represents a (plain) PAXp.
### 4.1 Path Probabilities for DTs
This section investigates how to compute, in the case of DTs, the conditional
probability,
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})$
(14)
where ${\mathcal{X}}$ is a set of _fixed_ features (whereas the other features
are not fixed, being deemed _universal_), and $P_{t}$ is a path in the DT
consistent with the instance $(\mathbf{v},c)$. (Also, note that (14) is the
left-hand side of (9).) To motivate the proposed approach, let us first
analyze how we can compute
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c)$, where
${\mathcal{P}}\subseteq{\mathcal{R}}$ is the set of paths in the DT with
prediction $c$. Let $\mathrm{\Lambda}(R_{k})$ denote the set of literals (each
of the form $x_{i}\in\mathbb{E}_{i}$) in some path $R_{k}\in{\mathcal{R}}$. If
a feature $i$ is tested multiple times along path $R_{k}$, then
$\mathbb{E}_{i}$ is the intersection of the sets in each of the literals of
$R_{k}$ on $i$. The number of values of $\mathbb{D}_{i}$ consistent with
literal $x_{i}\in\mathbb{E}_{i}$ is $|\mathbb{E}_{i}|$. Finally, the features
_not_ tested along $R_{k}$ are denoted by $\mathrm{\Psi}(R_{k})$. For path
$R_{k}$, the probability that a randomly chosen point in feature space is
consistent with $R_{k}$ (i.e. the _path probability_ of $R_{k}$) is given by,
$\textnormal{Pr}(R_{k})=\nicefrac{{\left[\prod_{(x_{i}\in\mathbb{E}_{i})\in\mathrm{\Lambda}(R_{k})}|\mathbb{E}_{i}|\times\prod_{i\in\mathrm{\Psi}(R_{k})}|\mathbb{D}_{i}|\right]}}{{|\mathbb{F}|}}$
(15)
As a result, we get that,
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c)={\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{P}}}\textnormal{Pr}(R_{k})$
(16)
Given an instance $(\mathbf{v},c)$ and a set of fixed features ${\mathcal{X}}$
(and so a set of universal features ${\mathcal{F}}\setminus{\mathcal{X}}$), we
now detail how to compute (14). Since some features will now be declared
universal, multiple paths with possibly different conditions can become
consistent. For example, in Figure 2 if feature 1 and 2 are declared
universal, then (at least) paths $P_{1}$, $P_{2}$ and $Q_{1}$ are consistent
with some of the possible assignments. Although universal variables might seem
to complicate the computation of the conditional probability, this is not the
case.
A key observation is that the feature values that make a path consistent are
disjoint from the values that make other paths consistent. This observation
allows us to compute the models consistent with each path and, as a result, to
compute (9). Let $R_{k}\in{\mathcal{R}}$ represent some path in the decision
tree. (Recall that $P_{t}\in{\mathcal{P}}$ is the target path, which is
consistent with $\mathbf{v}$.) Let $n_{ik}$ represent the (integer) number of
assignments to feature $i$ that are consistent with path
$R_{k}\in{\mathcal{R}}$, given $\mathbf{v}\in\mathbb{F}$ and
${\mathcal{X}}\subseteq{\mathcal{F}}$. For a feature $i$, let $\mathbb{E}_{i}$
denote the set of domain values of feature $i$ that is consistent with path
$R_{k}$. Hence, for path $R_{k}$, we consider a literal
$(x_{i}\in{\mathbb{E}_{i}})$. Given the above, the value of $n_{ik}$ is
defined as follows:
1. 1.
If $i$ is fixed:
1. (a)
If $i$ is tested along $R_{k}$ and the value of $x_{i}$ is inconsistent with
$\mathbf{v}$, i.e. there exists a literal
$(x_{i}\in\mathbb{E}_{i})\in\mathrm{\Lambda}(R_{k})$ and
$\\{v_{i}\\}\cap{\mathbb{E}_{i}}=\emptyset$, then $n_{ik}=0$;
2. (b)
If $i$ is tested along $R_{k}$ and the value of $x_{i}$ is consistent with
$R_{k}$, i.e. there exists a literal
$(x_{i}\in\mathbb{E}_{i})\in\mathrm{\Lambda}(R_{k})$ and
$\\{v_{i}\\}\cap{\mathbb{E}_{i}}\not=\emptyset$, then $n_{ik}=1$;
3. (c)
If $i$ is not tested along $R_{k}$, then $n_{ik}=1$.
2. 2.
Otherwise, $i$ is universal:
1. (a)
If $i$ is tested along $R_{k}$, with some literal $x_{i}\in\mathbb{E}_{i}$,
then $n_{ik}=|\mathbb{E}_{i}|$;
2. (b)
If $i$ is not tested along $R_{k}$, then $n_{ik}=|\mathbb{D}_{i}|$.
Using the definition of $n_{ik}$, we can then compute the number of
assignments consistent with $R_{k}$ as follows:
$\\#(R_{k};\mathbf{v},{\mathcal{X}})={\textstyle\prod\nolimits}_{i\in{\mathcal{F}}}n_{ik}$
(17)
Finally, (14) is given by,
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})\,=\,\nicefrac{{\sum_{P_{k}\in{\mathcal{P}}}\\#(P_{k};\mathbf{v},{\mathcal{X}})}}{{\sum_{R_{k}\in{\mathcal{R}}}\\#(R_{k};\mathbf{v},{\mathcal{X}})}}$
(18)
As can be concluded, and in the case of a decision tree, both
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})$
and $\mathsf{WeakPAXp}({\mathcal{X}};\mathbb{F},\kappa,\mathbf{v},c,\delta)$
are computed in polynomial time on the size of the DT.
###### Example 3.
With respect to the DT in Figure 2, and given the instance $((4,4,2),\oplus)$,
the number of models for each path is shown in Table 1. For example, for set
$\\{3\\}$, we immediately get that
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})=\nicefrac{{15}}{{(15+1)}}=\nicefrac{{15}}{{16}}$.
### 4.2 Computing Locally-Minimal PAXp’s for DT’s
Recent work showed that, for DTs, one AXp can be computed in polynomial time
[45, 35, 46]. A simple polynomial-time algorithm can be summarized as follows.
The AXp ${\mathcal{X}}$ is initialized with all the features in
${\mathcal{F}}$. Pick the path consistent with a given instance
$(\mathbf{v},c)$. The features not in the path are removed from
${\mathcal{X}}$. Then, iteratively check whether
${\mathcal{X}}\setminus\\{i\\}$ guarantees that all paths to a prediction in
${\mathcal{K}}\setminus\\{c\\}$ are still inconsistent. If so, then update
${\mathcal{X}}$. As argued in Section 3, we can use a similar (deletion-based)
approach for computing one locally-minimal PAXp for DTs. Such an approach
builds on Algorithm 1. In the case of DTs, (3.1) is computed using (18) on
some given set ${\mathcal{S}}\setminus\\{i\\}$. to decide whether the
precision of the approximation ${\mathcal{S}}\setminus\\{i\\}$ is no smaller
than the threshold $\delta$. As stated earlier, (18) is computed in polynomial
time. Hence, Algorithm 1 runs in polynomial time for DTs.
### 4.3 Computing Minimum-Size PAXp’s for DTs
For computing a minimum-size PAXp, we propose two SMT encodings, thus showing
that the decision problem is in NP, and that finding a smallest set requires a
logarithmic number of calls to an NP-oracle. Regarding the two SMT encodings,
one involves the multiplication of integer variables, and so it involves non-
linear arithmetic. Given the structure of the problem, we also show that
linear arithmetic can be used, by proposing a (polynomially) larger encoding.
#### A multiplication-based SMT encoding.
Taking into account the definition of path probabilities (see Section 4.1), we
now devise a model that computes path probabilities based on the same ideas.
Let $j\in{\mathcal{F}}$ denote a given feature. Let $n_{jk}$ denote the number
of elements in $\mathbb{D}_{j}$ consistent with path $R_{k}$ (for simplicity,
we just use the path index $k$). Also, $u_{j}$ is a boolean variable that
indicates whether feature $j$ is fixed ($u_{j}=0$) or universal ($u_{j}=1$).
If feature $j$ is not tested along path $R_{k}$, then if $j$ is fixed, then
$n_{jk}=1$. If not, then $n_{jk}=|\mathbb{D}_{j}|$. Otherwise, $j$ is tested
along path $R_{k}$. $n_{jk}$ is 0 if $j$ is fixed (i.e. $u_{j}=0$) and
inconsistent with the values of $\mathbb{D}_{j}$ allowed for path $R_{k}$.
$n_{jk}$ is 1 if $j$ is fixed and consistent with the values of
$\mathbb{D}_{j}$ allowed for path $R_{k}$. If feature $j$ is not fixed (i.e.
it is deemed universal and so $u_{j}=1$), then $n_{jk}$ denotes the number of
domain values of $j$ consistent with path $R_{k}$. Let the fixed value of
$n_{jk}$ be $n_{0jk}$ and the _universal_ value of $n_{jk}$ be $n_{1jk}$.
Thus, $n_{jk}$ is defined as follows,
$n_{jk}=\textnormal{ite}(u_{j},n_{1jk},n_{0jk})$ (19)
Moreover, let $\eta_{k}$ denote the number of models of path $R_{k}$. Then,
$\eta_{k}$ is defined as follows:
$\eta_{k}={\textstyle\prod\nolimits}_{i\in\mathrm{\Phi}(k)}n_{ik}$ (20)
If the domains are boolean, then we can use a purely boolean formulation for
the problem. However, if the domains are multi-valued, then we need this
formulation.
Recall what we must ensure that (3.1) holds true. In the case of DTs, since we
can count the models associated with each path, depending on which features
are fixed or not, then the previous constraint can be translated to:
${\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{P}}}\eta_{k}\geq\delta\times{\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{P}}}\eta_{k}+\delta\times{\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{Q}}}\eta_{k}$
(21)
Recall that ${\mathcal{P}}$ are the paths with the matching prediction, and
${\mathcal{Q}}$ are the rest of the paths.
Finally, the soft constraints are of the form $(u_{i})$, one for each feature
$i\in{\mathcal{F}}\setminus\Psi(R_{k})$, i.e. for the features tested along
path $R_{k}$. (For each feature $i$ not tested along $R_{k}$, i.e.
$i\in\Psi(R_{k})$, enforce that the feature is universal by adding a hard
clause $(u_{i})$.) The solution to the optimization problem will then be a
_smallest_ weak PAXp, and so also a (plain) PAXp. (The minimum-cost solution
is well-known to be computed with a worst-case logarithmic number of calls (on
the number of features) to an SMT solver.)
Feature | Attr. | $P_{1}$ | $P_{2}$ | $P_{3}$ | $Q_{1}$ | $Q_{2}$
---|---|---|---|---|---|---
1 | $n_{01k}$ | 0 | 1 | 1 | 0 | 1
$n_{11k}$ | 1 | 3 | 3 | 1 | 3
$n_{1k}$ | $n_{1k}=\textnormal{ite}(u_{1},n_{11k},n_{01k})$
2 | $n_{02k}$ | 1 | 0 | 1 | 0 | 1
$n_{12k}$ | 3 | 1 | 3 | 1 | 3
$n_{2k}$ | $n_{2k}=\textnormal{ite}(u_{2},n_{12k},n_{02k})$
3 | $n_{03k}$ | 1 | 1 | 1 | 1 | 0
$n_{13k}$ | 2 | 2 | 1 | 2 | 1
$n_{3k}$ | $n_{3k}=\textnormal{ite}(u_{3},n_{13k},n_{03k})$
Path counts | $\eta_{k}=n_{1k}\times{n_{2k}}\times{n_{3k}}$
Table 2: SMT encoding for multiplication-based encoding
###### Example 4.
For the running example, let us consider ${\mathcal{X}}=\\{3\\}$. This means
that $u_{1}=u_{2}=1$. As a result, given the instance and the proposed
encoding, we get Table 2 and Table 3.
Path | $n_{1k}$ | $n_{2k}$ | $n_{3k}$ | $\eta_{k}$
---|---|---|---|---
$R_{1}$ | 1 | 3 | 1 | 3
$R_{2}$ | 1 | 3 | 1 | 3
$R_{3}$ | 3 | 3 | 1 | 9
$R_{4}$ | 1 | 1 | 1 | 1
$R_{5}$ | 3 | 3 | 0 | 0
Table 3: Concrete values for the multiplication-based encoding for the case
${\mathcal{X}}=\\{3\\}$, i.e. $u_{1}=u_{2}=1$ and $u_{3}=0$
Finally, by plugging into (21) the values from Table 3, we get: $15\geq
0.93\times(15+1)$. Thus, ${\mathcal{X}}$ is a weak PAXp, and we can show that
it is both a plain PAXp and a smallest PAXp. Indeed, with
${\mathcal{Y}}=\emptyset$, we get
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{Y}}}=\mathbf{v}_{{\mathcal{Y}}})=21/32=0.65625<\delta$.
Hence, ${\mathcal{X}}=\\{3\\}$ is subset-minimal. Since there can be no PAXp’s
of smaller size, then ${\mathcal{X}}$ is also a smallest PAXp.
#### An alternative addition-based SMT encoding.
A possible downside of the SMT encoding described above is the use of
multiplication of variables in (20); this causes the SMT problem formulation
to involve different theories (which may turn out to be harder to reason about
in practice). Given the problem formulation, we can use an encoding that just
uses linear arithmetic. This encoding is organized as follows. Let the order
of features be: $\langle 1,2,\ldots,m\rangle$. Define $\eta_{j,k}$ as the sum
of models of path $R_{k}$ taking into account features 1 up to $j$, with
$\eta_{0,k}=1$. Given $\eta_{{j-1},{k}}$, $\eta_{{j},{k}}$ is computed as
follows:
* •
Let the domain of feature $j$ be $\mathbb{D}_{j}=\\{v_{j1},\ldots,v_{jr}\\}$,
and let $s_{j,l,k}$ denote the number of models taking into account features 1
up to $j-1$ and domain values $v_{j1}$ up to $v_{j{l-1}}$. Also, let
$s_{j,0,k}=0$.
* •
For each value $v_{jl}$ in $\mathbb{D}_{j}$, for $l=1,\ldots,r$:
* –
If $j$ is tested along path $R_{k}$: (i) If $v_{jl}$ is inconsistent with path
$R_{k}$, then $s_{j,l,k}=s_{j,l-1,k}$; (ii) If $v_{jl}$ is consistent with
path $R_{k}$ and with $\mathbf{v}$, then
$s_{j,l,k}=s_{j,l-1,k}+\eta_{{j-1},{k}}$; (iii) If $v_{jl}$ is consistent with
path $R_{k}$ but not with $\mathbf{v}$, or if feature $j$ is not tested in
path $R_{k}$, then
$s_{j,l,k}=s_{j,l-1,k}+\textnormal{ite}(u_{j},\eta_{{j-1},{k}},0)$.
* –
If $j$ is not tested along path $R_{k}$: (i) If $v_{jl}$ is consistent with
$\mathbf{v}$, then $s_{j,l,k}=s_{j,l-1,k}+\eta_{{j-1},{k}}$; (ii) Otherwise,
$s_{j,l,k}=s_{j,l-1,k}+\textnormal{ite}(u_{j},\eta_{{j-1},{k}},0)$.
* •
Finally, define $\eta_{{j},{k}}=s_{j,r,k}$.
After considering all the features in order, $\eta_{m,k}$ represents the
number of models for path $R_{k}$ given the assigment to the $u_{j}$
variables. As a result, we can re-write (21) as follows:
${\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{P}}}\eta_{m,k}\geq\delta\times{\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{P}}}\eta_{m,k}+\delta\times{\textstyle\sum\nolimits}_{R_{k}\in{\mathcal{Q}}}\eta_{m,k}$
(22)
As with the multiplication-based encoding, the soft clauses are of the form
$(u_{i})$ for $i\in{\mathcal{F}}$.
Var. | $R_{1}\cong{P_{1}}$ | $R_{2}\cong{P_{2}}$ | $R_{3}\cong{P_{3}}$ | $R_{4}\cong{Q_{1}}$ | $R_{5}\cong{Q_{2}}$
---|---|---|---|---|---
$s_{1,0,k}$ | $s_{1,0,1}=0$ | $s_{1,0,2}=0$ | $s_{1,0,3}=0$ | $s_{1,0,4}=0$ | $s_{1,0,5}=0$
$s_{1,1,k}$ | $s_{1,0,1}+\textnormal{ite}(u_{1},\eta_{0,1},0)$ | $s_{1,0,2}$ | $s_{1,0,3}$ | $s_{1,0,4}+\textnormal{ite}(u_{1},\eta_{0,4},0)$ | $s_{1,0,5}$
$s_{1,2,k}$ | $s_{1,1,1}$ | $s_{1,1,2}+\textnormal{ite}(u_{1},\eta_{0,2},0)$ | $s_{1,1,3}+\textnormal{ite}(u_{1},\eta_{0,3},0)$ | $s_{1,1,4}$ | $s_{1,1,5}+\textnormal{ite}(u_{1},\eta_{0,5},0)$
$s_{1,3,k}$ | $s_{1,2,1}$ | $s_{1,2,2}+\textnormal{ite}(u_{1},\eta_{0,2},0)$ | $s_{1,2,3}+\textnormal{ite}(u_{1},\eta_{0,3},0)$ | $s_{1,2,4}$ | $s_{1,2,5}+\textnormal{ite}(u_{1},\eta_{0,5},0)$
$s_{1,4,k}$ | $s_{1,3,1}$ | $s_{1,3,2}+\eta_{0,2}$ | $s_{1,3,3}+\eta_{0,3}$ | $s_{1,3,4}$ | $s_{1,3,5}+\eta_{0,5}$
$\eta_{1,k}$ | $s_{1,4,1}$ | $s_{1,4,2}$ | $s_{1,4,3}$ | $s_{1,4,4}$ | $s_{1,4,5}$
$s_{2,0,k}$ | $s_{2,0,1}=0$ | $s_{2,0,2}=0$ | $s_{2,0,3}=0$ | $s_{2,0,4}=0$ | $s_{2,0,5}=0$
$s_{2,1,k}$ | $s_{2,0,1}$ | $s_{2,0,2}+\eta_{1,2}$ | $s_{2,0,3}$ | $s_{2,0,4}+\textnormal{ite}(u_{2},\eta_{1,4},0)$ | $s_{2,0,5}+\textnormal{ite}(u_{2},\eta_{1,5},0)$
$s_{2,2,k}$ | $s_{2,1,1}+\textnormal{ite}(u_{2},\eta_{1,1},0)$ | $s_{2,1,2}$ | $s_{2,1,3}+\textnormal{ite}(u_{2},\eta_{1,3},0)$ | $s_{2,1,4}$ | $s_{2,1,5}$
$s_{2,3,k}$ | $s_{2,2,1}+\textnormal{ite}(u_{2},\eta_{1,1},0)$ | $s_{2,2,2}$ | $s_{2,2,3}+\textnormal{ite}(u_{2},\eta_{1,3},0)$ | $s_{2,2,4}$ | $s_{2,2,5}$
$s_{2,4,k}$ | $s_{2,3,1}+\eta_{1,1}$ | $s_{2,3,2}$ | $s_{2,3,3}+\eta_{1,3},0$ | $s_{2,3,4}$ | $s_{2,3,5}$
$\eta_{2,k}$ | $s_{2,4,1}$ | $s_{2,4,2}$ | $s_{2,4,3}$ | $s_{2,4,4}$ | $s_{2,4,5}$
$s_{3,0,k}$ | $s_{3,0,1}=0$ | $s_{3,0,2}=0$ | $s_{3,0,3}=0$ | $s_{3,0,4}=0$ | $s_{3,0,5}=0$
$s_{3,1,k}$ | $s_{3,0,1}+\textnormal{ite}(u_{3},\eta_{2,1},0)$ | $s_{3,0,2}+\textnormal{ite}(u_{3},\eta_{2,2},0)$ | $s_{3,0,3}$ | $s_{3,0,4}+\textnormal{ite}(u_{3},\eta_{2,4},0)$ | $s_{3,0,5}+\textnormal{ite}(u_{3},\eta_{2,5},0)$
$s_{3,2,k}$ | $s_{3,1,1}+\eta_{2,1}$ | $s_{3,1,2}+\eta_{2,2}$ | $s_{3,1,3}+\eta_{2,3}$ | $s_{3,1,4}+\eta_{2,4}$ | $s_{3,1,5}$
$\eta_{3,k}$ | $s_{3,4,1}$ | $s_{3,4,2}$ | $s_{3,4,3}$ | $s_{3,2,4}$ | $s_{3,2,5}$
Table 4: Partial addition-based SMT encoding for paths with prediction
$\oplus$, with $(\mathbf{v},c)=((4,4,2),\oplus)$, and with
$\eta_{0,1}=\eta_{0,2}=\eta_{0,3}=1$
###### Example 5.
Table 4 summarizes the SMT encoding based on iterated summations for paths
with either prediction $\oplus$ or $\ominus$. The final computed values are
then used in the linear inequality (22), as follows,
$\eta_{3,1}+\eta_{3,2}+\eta_{3,2}\geq\delta\times(\eta_{3,1}+\eta_{3,2}+\eta_{3,2})+\delta\times(\eta_{3,4}+\eta_{3,5})$
The optimization problem also includes
${\mathcal{B}}=\\{(\neg{u_{1}}),(\neg{u_{2}}),(\neg{u_{3}})\\}$ as the soft
clauses. For the counting-based encoding, and from Table 4, we get the values
shown in Table 5. Moreover, we can then confirm that $15\geq 0.93\times 16$,
as intended.
Var. | $R_{1}\cong{P_{1}}$ | $R_{2}\cong{P_{2}}$ | $R_{3}\cong{P_{3}}$ | $R_{4}\cong{Q_{1}}$ | $R_{5}\cong{Q_{2}}$
---|---|---|---|---|---
$s_{1,0,k}$ | 0 | 0 | 0 | 0 | 0
$s_{1,1,k}$ | 1 | 0 | 0 | 1 | 0
$s_{1,2,k}$ | 1 | 2 | 1 | 1 | 1
$s_{1,3,k}$ | 1 | 2 | 2 | 1 | 2
$s_{1,4,k}$ | 1 | 3 | 3 | 1 | 3
$\eta_{1,k}$ | 1 | 3 | 3 | 1 | 3
$s_{2,0,k}$ | 0 | 0 | 0 | 0 | 0
$s_{2,1,k}$ | 0 | 3 | 0 | 0 | 3
$s_{2,2,k}$ | 1 | 3 | 3 | 1 | 3
$s_{2,3,k}$ | 2 | 3 | 6 | 1 | 3
$s_{2,4,k}$ | 3 | 3 | 9 | 1 | 3
$\eta_{2,k}$ | 3 | 3 | 9 | 1 | 3
$s_{3,0,k}$ | 0 | 0 | 0 | 0 | 0
$s_{3,1,k}$ | 0 | 2 | 0 | 0 | 0
$s_{3,2,k}$ | 3 | 3 | 9 | 1 | 0
$\eta_{3,k}$ | 3 | 3 | 9 | 1 | 0
Table 5: Assignment to variables of addition-based SMT encoding, given
${\mathcal{X}}=\\{3\\}$, i.e. $u_{1}=u_{2}=1$ and $u_{3}=0$
#### Discussion.
In this as in the following sections, one might consider the use of a model
counter as a possible alternative. However, a model counter would have to be
used for each pick of features. Given the complexity of exactly computing the
number of models, such approaches are all but assured to be impractical in
practice.
### 4.4 Deciding Whether a Locally-Minimal PAXp is a Plain PAXp for DTs
The problem of deciding whether a set of features ${\mathcal{X}}$,
representing an $\mathsf{LmPAXp}$, is subset-minimal can be achieved by using
one of the models above, keeping the features that are already universal, and
checking whether additional universal features can be made to exist. In
addition, we need to add constraints forcing universal features to remain
universal, and at least one of the currently fixed features to also become
universal. Thus, if ${\mathcal{X}}$ is the set of fixed features, the SMT
models proposed in earlier sections is extended with the following
constraints:
${\textstyle\bigwedge\nolimits}_{j\in{\mathcal{F}}\setminus{\mathcal{X}}}(u_{j}){\textstyle\bigwedge}\left({\textstyle\bigvee\nolimits}_{j\in{\mathcal{X}}}u_{j}\right)$
(23)
which allow checking whether some set of fixed features can be declared
universal while respecting the other constraints.
### 4.5 Instance-Based vs. Path-Based Explanations
The standard definitions of abductive explanations consider a concrete
instance $(\mathbf{v},c)$. As argued earlier (see (7)), each (weak) AXp
${\mathcal{X}}$ can then be viewed as a rule of the form:
$\textbf{IF\leavevmode\nobreak\ \leavevmode\nobreak\
}\left[\land_{i\in{\mathcal{X}}}(x_{i}=v_{i})\right]\textbf{\leavevmode\nobreak\
\leavevmode\nobreak\ THEN\leavevmode\nobreak\ \leavevmode\nobreak\
}\left[\kappa(\mathbf{x})=c\right]$
In the case of DTs, a given $\mathbf{v}$ is consistent with a concrete path
$P_{t}$. As argued in recent work [46], this enables studying instead
generalizations of AXp’s, concretely to so-called _path-based explanations_ ,
each of which can be viewed as representing instead a rule of the form:
$\textbf{IF\leavevmode\nobreak\ \leavevmode\nobreak\
}\left[\land_{i\in{\mathcal{X}}}(x_{i}\in{\mathbb{E}_{i}})\right]\textbf{\leavevmode\nobreak\
\leavevmode\nobreak\ THEN\leavevmode\nobreak\ \leavevmode\nobreak\
}\left[\kappa(\mathbf{x})=c\right]$ (24)
where $\mathbb{E}_{i}\subseteq{\mathbb{D}_{i}}$ and where each literal
$x_{i}\in{\mathbb{E}_{i}}$ is one of the literals in the path $P_{t}$
consistent with the instance $(\mathbf{v},c)$.
Clearly, the literals associated with a path $R_{k}$ offer more information
than those associated with a concrete point $\mathbf{v}$ in feature space. As
a result, in the case of DTs, we consider a generalization of the definition
of relevant set, and seek instead to compute:
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}\in\mathbb{E}_{{\mathcal{X}}})$
(25)
where the notation $\mathbf{x}_{{\mathcal{X}}}\in\mathbb{E}_{{\mathcal{X}}}$
represents the constraint $\land_{i\in{\mathcal{X}}}x_{i}\in{E_{i}}$, and
where $\mathbb{E}_{i}$ denotes the set of values consistent with feature $i$
in path $R_{k}$.) Thus, the condition of weak PAXp considers instead the
following probability:
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}\in\mathbb{E}_{{\mathcal{X}}})\geq\delta$
(26)
The rest of this section investigates the computation of path probabilities in
the case of path-based explanations. For instance-based explanations, the
definition of $n_{ik}$ needs to be adapted. Let $P_{t}\in{\mathcal{P}}$ be the
target path. (For example, $P_{t}$ can be the path consistent with
$\mathbf{v}$.) Moreover, let $R_{K}\in{\mathcal{R}}$ by some path in the
decision tree. For a feature $i$, let $E_{ik}$ denote the set of domain values
of feature $i$ that is consistent with path $R_{k}$. Hence, for path $R_{k}$,
we consider a literal $(x_{i}\in{E_{ik}})$. Similarly, let $E_{it}$ denote the
set of domain values of feature $i$ that is consistent with path $P_{t}$.
Thus, for path $P_{t}$, we consider a literal $(x_{i}\in{E_{it}})$. Given the
above, the value of $n_{ik}$ is now defined as follows:
1. 1.
If $i$ is fixed:
1. (a)
If $i$ is tested along $R_{k}$ and the value of $x_{i}$ is inconsistent with
$E_{it}$, i.e. there exists a literal
$(x_{i}\in\mathbb{E}_{i})\in\mathrm{\Lambda}(R_{k})$ and
${\mathbb{E}_{it}}\cap{\mathbb{E}_{ik}}=\emptyset$, then $n_{ik}=0$;
2. (b)
If $i$ is tested along $R_{k}$ and the value of $x_{i}$ is consistent with
$R_{k}$, i.e. there exists a literal
$(x_{i}\in\mathbb{E}_{i})\in\mathrm{\Lambda}(R_{k})$ and
${\mathbb{E}_{it}}\cap{\mathbb{E}_{ik}}\not=\emptyset$, then $n_{ik}=1$.
3. (c)
If $i$ is not tested along $R_{k}$, then $n_{ik}=1$.
2. 2.
Otherwise, $i$ is universal:
1. (a)
If $i$ is tested along $R_{k}$, with some literal $x_{i}\in\mathbb{E}_{ik}$,
then $n_{ik}=|\mathbb{E}_{ik}|$;
2. (b)
If $i$ is not tested along $R_{k}$, then $n_{ik}=|\mathbb{D}_{i}|$.
Using the modified definition of $n_{ik}$, we can now compute (25) as follows:
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}\in\mathbb{E}_{{\mathcal{X}}})\,=\,\nicefrac{{\sum_{P_{k}\in{\mathcal{P}}}\\#(P_{k};{\mathcal{F}}\setminus{\mathcal{X}},\mathbf{v})}}{{\sum_{R_{k}\in{\mathcal{R}}}\\#(R_{k};{\mathcal{F}}\setminus{\mathcal{X}},\mathbf{v})}}$
(27)
The computation of probabilistic explanations proposed in the previous
sections can either assume instance-based or path-based explanations. For
consistency with the rest of the paper, we opted to investigate instance-based
explanations. Changing the proposed algorithms to consider instead path-based
explanations would be straightforward, but that is beyond the scope of this
paper.
## 5 Probabilistic Explanations for Naive Bayes Classifiers
This section investigates the computation of relevant sets in the concrete
case of NBCs.
### 5.1 Explaining NBCs in Polynomial Time
This section overviews the approach proposed in [62] for computing AXp’s for
binary NBCs. The general idea is to reduce the NBC problem into an Extended
Linear Classifier (XLC) and then explain the resulting XLC. Our purpose is to
devise a new approach that builds on the XLC formulation to compute
$\delta$-relevant sets for NBCs. Hence, it is useful to recall first the
translation of NBCs into XLCs and the extraction of AXp’s from XLCs.
#### Extended Linear Classifiers.
We consider an XLC with categorical features. (Recall that the present paper
considers NBCs with binary classes and categorical data.) Each feature
$i\in{\mathcal{F}}$ has $x_{i}\in\\{1,\dots,d_{i}\\}$, (i.e.
$\mathbb{D}_{i}=\\{1,\dots,d_{i}\\}$). Let,
$\nu(\mathbf{x})\triangleq
w_{0}+\sum\nolimits_{i\in{\mathcal{F}}}\sigma(x_{i},v_{i}^{1},v_{i}^{2},\ldots,v_{i}^{d_{i}})$
(28)
$\sigma$ is a selector function that picks the value $v_{i}^{r}$ iff $x_{i}$
takes value $r$. Moreover, let us define the decision function,
$\kappa(\mathbf{x})=\oplus$ if $\nu(\mathbf{x})>0$ and
$\kappa(\mathbf{x})=\ominus$ if $\nu(\mathbf{x})\leq 0$.
The reduction of a binary NBC, with categorical features, to an XLC is
completed by setting:
$w_{0}\triangleq\operatorname*{lPr}(\oplus)-\operatorname*{lPr}(\ominus)$,
$v_{i}^{1}\triangleq\operatorname*{lPr}(x_{i}=1|\oplus)-\operatorname*{lPr}(x_{i}=1|\ominus)$,
$v_{i}^{2}\triangleq\operatorname*{lPr}(x_{i}=2|\oplus)-\operatorname*{lPr}(x_{i}=2|\ominus)$,
$\dots$,
$v_{i}^{d_{i}}\triangleq\operatorname*{lPr}(x_{i}=d_{i}|\oplus)-\operatorname*{lPr}(x_{i}=d_{i}|\ominus)$.
Hence, the argmax in (3) is replaced with inequality to get the following:
$\displaystyle\operatorname*{lPr}(\oplus)-\operatorname*{lPr}(\ominus)+\sum\nolimits_{i=1}^{m}\sum\nolimits_{k=1}^{k=d_{i}}$
$\displaystyle(\operatorname*{lPr}(x_{i}=k|\oplus)-\operatorname*{lPr}(x_{i}=k|\ominus))(x_{i}=k)>0$
(29)
$w_{0}$ | $v_{1}^{1}$ | $v_{1}^{2}$ | $v_{2}^{1}$ | $v_{2}^{2}$ | $v_{3}^{1}$ | $v_{3}^{2}$ | $v_{4}^{1}$ | $v_{4}^{2}$ | $v_{5}^{1}$ | $v_{5}^{2}$
---|---|---|---|---|---|---|---|---|---|---
-2.19 | -2.97 | 3.46 | 2.95 | -2.95 | 0.4 | -2.83 | 1.17 | -1.32 | -2.97 | 3.46
(a) Example reduction of NBC to XLC (Example 6)
$\Gamma$ | $\delta_{1}$ | $\delta_{5}$ | $\delta_{2}$ | $\delta_{3}$ | $\delta_{4}$ | $\Phi$
---|---|---|---|---|---|---
9.25 | 6.43 | 6.43 | 5.90 | 3.23 | 2.49 | 15.23
(b) Computing $\delta_{j}$’s for the XLC (Example 7)
Figure 7: Values used in the running example (Example 6 and Example 7)
###### Example 6.
7(a) shows the resulting XLC formulation for the example in Figure 4. We also
let $\mathbf{f}$ be associated with value 1 and $\mathbf{t}$ be associated
with value 2, and $d_{i}=2$.
#### Explaining XLCs.
We now describe how AXp’s can be computed for XLCs. For a given instance
$\mathbf{x}=\mathbf{a}$, define a _constant_ slack (or gap) value $\Gamma$
given by,
$\Gamma\triangleq\nu(\mathbf{a})=\sum\nolimits_{i\in{\mathcal{F}}}\sigma(a_{i},v_{i}^{1},v_{i}^{2},\ldots,v_{i}^{d_{i}})$
(30)
Computing an AXp corresponds to finding a subset-minimal set of literals
${\mathcal{S}}\subseteq{\mathcal{F}}$ such that (2.5) holds, or alternatively,
$\forall(\mathbf{x}\in\mathbb{F}).\bigwedge\nolimits_{i\in{\mathcal{S}}}(x_{i}=a_{i})\
\operatorname*{\rightarrow}\ \left(\nu(\mathbf{x})>0\right)$ (31)
under the assumption that $\nu(\mathbf{a})>0$. Thus, the purpose is to find
the _smallest_ slack that can be achieved by allowing the feature not in
${\mathcal{S}}$ to take any value (i.e. _universal_ /_free_ features), given
that the literals in ${\mathcal{S}}$ are fixed by $\mathbf{a}$ (i.e.
$\bigwedge\nolimits_{i\in{\mathcal{S}}}(x_{i}=a_{i})$).
Let $v_{i}^{\omega}$ denote the _smallest_ (or _worst-case_) value associated
with $x_{i}$. Then, by letting every $x_{i}$ take _any_ value, the _worst-
case_ value of $\nu(\mathbf{e})$ is,
$\Gamma^{\omega}=w_{0}+\sum\nolimits_{i\in{\mathcal{F}}}v_{i}^{\omega}$ (32)
Moreover, from (30), we have:
$\Gamma=w_{0}+\sum_{i\in{\mathcal{F}}}v_{i}^{a_{i}}$. The expression above can
be rewritten as follows,
$\begin{array}[]{rcl}\Gamma^{\omega}&=&w_{0}+\sum\nolimits_{i\in{\mathcal{F}}}v_{i}^{a_{i}}-\sum\nolimits_{i\in{\mathcal{F}}}(v_{i}^{a_{i}}-v_{i}^{\omega})\\\\[3.0pt]
&=&\Gamma-\sum\nolimits_{i\in{\mathcal{F}}}\delta_{i}=-\Phi\\\ \end{array}$
(33)
where $\delta_{i}\triangleq{v_{i}^{a_{i}}}-{v_{i}^{\omega}}$, and
$\Phi\triangleq\sum_{i\in{\mathcal{F}}}\delta_{i}-\Gamma=-\Gamma^{\omega}$.
Recall the goal is to find a subset-minimal set ${\mathcal{S}}$ such that the
prediction is still $c$ (whatever the values of the other features):
$w_{0}+\sum\nolimits_{i\in{\mathcal{S}}}v_{i}^{a_{i}}+\sum\nolimits_{i\notin{\mathcal{S}}}v_{i}^{\omega}=-\Phi+\sum\nolimits_{i\in{\mathcal{S}}}\delta_{i}>0$
(34)
In turn, (34) can be represented as the following knapsack problem [54]:
$\begin{array}[]{lcl}\textnormal{min}&&\sum_{i=1}^{m}p_{i}\\\\[4.5pt]
\textnormal{such that}&&\sum_{i=1}^{m}\delta_{i}p_{i}>\Phi\\\\[2.5pt]
&&p_{i}\in\\{0,1\\}\\\ \end{array}$ (35)
where the variables $p_{i}$ assigned value 1 denote the indices included in
${\mathcal{S}}$. Note that, the fact that the coefficients in the cost
function are all equal to 1 makes the problem solvable in log-linear time
[62].
###### Example 7.
7(b) shows the values used for computing explanations for the example in
Figure 4. For this example, the sorted $\delta_{j}$’s become
$\langle\delta_{1},\delta_{5},\delta_{2},\delta_{4},\delta_{3}\rangle$. By
picking $\delta_{1}$, $\delta_{2}$ and $\delta_{5}$, we ensure that the
prediction is $\oplus$, independently of the values assigned to features 3 and
4. Thus $\\{1,2,5\\}$ is an AXp for the NBC shown in Figure 3, with the input
instance
$(v_{1},v_{2},v_{3},v_{4},v_{5})=(\mathbf{t},\mathbf{f},\mathbf{f},\mathbf{f},\mathbf{t})$.
(It is easy to observe that
$\kappa(\mathbf{t},\mathbf{f},\mathbf{f},\mathbf{t},\mathbf{t})=\kappa(\mathbf{t},\mathbf{f},\mathbf{t},\mathbf{f},\mathbf{t})=\kappa(\mathbf{t},\mathbf{f},\mathbf{t},\mathbf{t},\mathbf{t})=\oplus$.)
The next section introduces a pseudo-polynomial time algorithm for computing
locally-minimal PAXp’s. Although locally-minimal PAXp’s are not necessarily
subset/cardinality minimal, the experiments (see Section 7) show that the
proposed approach computes (in pseudo-polynomial time) succinct [69] and
highly precise locally-minimal explanations.
### 5.2 Counting Models of XLCs
Earlier work [27, 31, 30, 83] proposed the use of dynamic programming (DP) for
approximating the number of feasible solutions of the 0-1 knapsack constraint,
i.e. the #knapsack problem. Here we propose an extension of the basic
formulation, to allow counting feasible solutions of XLCs.
We are interested in the number of solutions of,
$\sum\nolimits_{j\in{\mathcal{F}}}\sigma(x_{j},v_{j}^{1},v_{j}^{2},\ldots,v_{j}^{d_{j}})>-w_{0}$
(36)
where we assume all $v^{i}_{j}$ to be integer-valued and non-negative (e.g.
this is what our translation from NBCs to XLCs yields after scaling and
rounding). Moreover, (36) can be written as follows:
$\sum\nolimits_{j\in{\mathcal{F}}}\sigma(x_{j},-v_{j}^{1},-v_{j}^{2},\ldots,-v_{j}^{d_{j}})<{w_{0}}$
(37)
which reveals the relationship with the standard knapsack constraint.
For each $j$, let us sort the $-v_{j}^{i}$ in non-decreasing order, collapsing
duplicates, and counting the number of duplicates, obtaining two sequences:
$\begin{array}[]{l}\langle
w^{1}_{j},\ldots,w^{d^{{}^{\prime}}_{j}}_{j}\rangle\\\\[4.5pt] \langle
n^{1}_{j},\ldots,n^{d^{{}^{\prime}}_{j}}_{j}\rangle\\\ \end{array}$
such that $w^{1}_{j}<w^{2}_{j}<\ldots<w^{d^{{}^{\prime}}_{j}}_{j}$ and each
$n^{i}_{j}\geq 1$ gives the number of repetitions of weight $w^{i}_{j}$.
#### Counting.
Let $C(k,r)$ denote the number of solutions of (37) when the subset of
features considered is $\\{1,\ldots,k\\}$ and the sum of picked weights is at
most $r$. To define the solution for the first $k$ features, taking into
account the solution for the first $k-1$ features, we must consider that the
solution for $r$ can be obtained due to _any_ of the possible values of
$x_{j}$. As a result, for an XLC the general recursive definition of $C(k,r)$
becomes,
$C(k,r)=\sum\nolimits_{i=1}^{d^{{}^{\prime}}_{k}}n^{i}_{k}\times{C}(k-1,r-w^{i}_{k})$
Moreover, $C(1,r)$ is given by,
$C(1,r)=\left\\{\begin{array}[]{lcl}0&&\textnormal{if
$r<w^{1}_{1}$}\\\\[4.5pt] n^{1}_{1}&&\textnormal{if $w^{1}_{1}\leq
r<w^{2}_{1}$}\\\\[5.5pt] n^{1}_{1}+n^{2}_{1}&&\textnormal{if $w^{2}_{1}\leq
r<w^{3}_{1}$}\\\\[2.0pt] \ldots\\\\[2.0pt]
\sum\nolimits_{i=1}^{d^{{}^{\prime}}_{1}}n^{i}_{1}&&\textnormal{if
$w^{d^{{}^{\prime}}_{1}}_{1}\leq r$}\\\\[2.0pt] \end{array}\right.$
In addition, if $r<0$, then $C(k,r)=0$, for $k=1,\ldots,m$. Finally, the
dimensions of the $C(k,r)$ table are as follows:
1. 1.
The number of rows is $m$.
2. 2.
The (worst-case) number of columns is given by:
$W^{\prime}=\sum_{j\in{\mathcal{F}}}{n^{d^{\prime}_{j}}_{j}}\times{w^{d^{\prime}_{j}}_{j}}$
(38)
$W^{\prime}$ represents the largest possible value, in theory. However, in
practice, it suffices to set the number of columns to $W=w_{0}+T$, which is
often much smaller than $W^{\prime}$.
###### Example 8.
Consider the following problem. There are 4 features,
${\mathcal{F}}=\\{1,2,3,4\\}$. Each feature $j$ takes values in $\\{1,2,3\\}$,
i.e. $x_{j}\in\\{1,2,3\\}$. The prediction should be 1 when the sum of the
values of the $x_{j}$ variables is not less than 8. We set $w_{0}=-7$, and get
the formulation,
$\sum_{j\in\\{1,2,3,4\\}}\sigma(x_{j},1,2,3)>7$
where each $x_{j}$ picks value in $\\{1,2,3\\}$. We translate to the extended
knapsack formulation and obtain:
$\sum_{j\in\\{1,2,3,4\\}}\sigma(x_{j},-1,-2,-3)<-7$
We require the weights to be integer and non-negative, and so we add to each
$w_{j}^{k}$ the complement of the most negative $w_{j}^{k}$ plus 1. Therefore,
we add +4 to each $j$ and +16 to right-hand side of the inequality. Thus, we
get
$\sum_{j\in\\{1,2,3,4\\}}\sigma(x_{j},3,2,1)<9$
For this formulation, $x_{j}=1$ picks value $3$. (For example, we can pick two
(but not three) $x_{j}$ with value 1, which is as expected.)
In this case, the DP table size will be $4\times 12$, even though we are
interested in entry $C(4,8)$.
$k$ | $r$
---|---
0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12
1 | 0 | 1 | 2 | 3 | 3 | 3 | 3 | 3 | 3 | – | – | – | –
2 | 0 | 0 | 1 | 3 | 6 | 8 | 9 | 9 | 9 | – | – | – | –
3 | 0 | 0 | 0 | 1 | 4 | 10 | 17 | 23 | 16 | – | – | – | –
4 | 0 | 0 | 0 | 0 | 1 | 5 | 15 | 31 | 50 | – | – | – | –
Table 6: DP table for Example 8
Table 6 shows the DP table, and the number of solutions for the starting
problem, i.e. there are 50 combinations of values whose sum is no less than 8.
By default, the dynamic programming formulation assumes that features can take
any value. However, the same formulation can be adapted when features take a
given (fixed) value. Observe that this will be instrumental for computing
$\mathsf{LmPAXp}$’s.
Consider that feature $k$ is fixed to value $l$. Then, the formulation for
$C(k,r)$ becomes:
$C(k,r)=n^{l}_{k}\times{C}(k-1,r-w^{l}_{k})={C}(k-1,r-w^{l}_{k})$
Given that $k$ is fixed, then we have $n^{l}_{k}=1$.
###### Example 9.
For Example 8, assume that $x_{2}=1$ and $x_{4}=3$. Then, the constraint we
want to satisfy is:
$\sum_{j\in\\{1,3\\}}\sigma(x_{j},1,2,3)>3$
Following a similar transformation into knapsack formulation, we get
$\sum_{j\in\\{1,3\\}}\sigma(x_{j},3,2,1)<5$
After updating the DP table, with fixing features 2 and 4, we get the DP table
shown in Table 7. As a result, we can conclude that the number of solutions is
6.
$k$ | $r$
---|---
0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12
1 | 0 | 1 | 2 | 3 | 3 | 3 | 3 | 3 | 3 | – | – | – | –
2 | 0 | 0 | 0 | 0 | 1 | 2 | 3 | 3 | 3 | – | – | – | –
3 | 0 | 0 | 0 | 0 | 0 | 1 | 3 | 6 | 8 | – | – | – | –
4 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 3 | 6 | – | – | – | –
Table 7: DP table for Example 9
The table $C(k,r)$ can be filled out in pseudo-polynomial time. The number of
rows is $m$. The number of columns is $W$ (see (38)). Moreover, the
computation of each entry uses the values of at most $m$ other entries. Thus,
the total running time is: $\Theta(m^{2}\times{W})$.
#### From NBCs to positive integer knapsacks.
To assess heuristic explainers, we consider NBCs, and use a standard
transformation from probabilities to positive real values [78]. Afterwards, we
convert the real values to integer values by scaling the numbers. However, to
avoid building a very large DP table, we implement the following optimization.
The number of decimal places of the probabilities is reduced while there is no
decrease in the accuracy of the classifier both on training and on test data.
In our experiments, we observed that there is no loss of accuracy if four
decimal places are used, and that there is a negligible loss of accuracy with
three decimal places.
#### Assessing explanation precision.
Given a Naive Bayes classifier, expressed as an XLC, we can assess explanation
accuracy in pseudo-polynomial time. Given an instance $\mathbf{v}$, a
prediction $\kappa(\mathbf{v})=\oplus$, and an approximate explanation
$\mathbf{S}$, we can use the approach described in this section to count the
number of instances consistent with the explanation for which the prediction
remains unchanged (i.e. number of points $\mathbf{x}\in\mathbb{F}$ s.t.
$(\kappa(\mathbf{x})=\kappa(\mathbf{v})\land(\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}}))$).
Let this number be $n_{\oplus}$ (given the assumption that the prediction is
$\oplus$). Let the number of instances with a different prediction be
$n_{\ominus}$ 777Recall that we are assuming that
${\mathcal{K}}=\\{\ominus,\oplus\\}$.. Hence, the conditional probability (8)
can be defined, in the case of NBCs, as follow:
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=\oplus\,|\,\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})=\frac{n_{\oplus}}{|\\{\mathbf{x}\in\mathbb{F}:(\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})\\}|}$
Observe that the numerator
$|\\{\mathbf{x}\in\mathbb{F}:\kappa(\mathbf{x})=\oplus\land(\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})\\}|$
is expressed by the number of models $n_{\oplus}$, i.e. the points
$\mathbf{x}$ in feature space that are consistent with $\mathbf{v}$ given
${\mathcal{S}}$ and with prediction $\oplus$. Further, we have
$\displaystyle\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=\oplus\,|\,\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})=$
$\displaystyle
1-\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=\ominus\,|\,\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})$
$\displaystyle=$ $\displaystyle
1-\frac{n_{\ominus}}{|\\{\mathbf{x}\in\mathbb{F}:(\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})\\}|}$
where
$n_{\ominus}=|\\{\mathbf{x}\in\mathbb{F}:\kappa(\mathbf{x})=\ominus\land(\mathbf{x}_{{\mathcal{S}}}=\mathbf{v}_{{\mathcal{S}}})\\}|$.
### 5.3 Computing Locally-Minimal PAXp’s for NBCs
Similarly to the case of DTs, we can also use Algorithm 1 for computing
locally-minimal PAXp’s in the case of NBCs. The only difference is in the
definition of the predicate $\mathsf{WeakPAXp}$. For NBCs, the procedure
$\mathsf{isWeakPAXp}$ implements the pseudo-polynomial approach, described in
the previous section, for model counting. Hence, in the case of NBCs, it is
implicit that the DP table is updated at each iteration of the main loop of
Algorithm 1. More specifically, when a feature $i$ is just set to universal,
its associated cells $C(i,r)$ are recalculated such that
$C(k,r)=\sum\nolimits_{i=1}^{d^{{}^{\prime}}_{k}}n^{i}_{k}\times{C}(k-1,r-w^{i}_{k})$;
and when $i$ is fixed, i.e. $i\in{\mathcal{S}}$, then
$C(i,r)=C(i-1,r-v^{j}_{i})$ where
$v^{j}_{i}\triangleq\operatorname*{lPr}(v_{i}=j|c)-\operatorname*{lPr}(v_{i}=j|\neg
c)$. Furthermore, we point out that in our experiments, ${\mathcal{S}}$ is
initialized to an AXp ${\mathcal{X}}$ that we compute initially for all tested
instances using the outlined (polynomial) algorithm in Section 5.1. It is easy
to observe that features not belonging to ${\mathcal{X}}$ do not contribute in
the decision of $\kappa(\mathbf{v})$ (i.e. their removal does not change the
value of $n_{\ominus}$ that is equal to zero) and thus can be set universal at
the initialisation step, which allows us to improve the performance of
Algorithm 1.
Moreover, we apply a heuristic order over ${\mathcal{S}}$ that aims to remove
earlier less relevant features and thus to produce shorter approximate
explanations. Typically, we order ${\mathcal{S}}$ following the increasing
order of $\delta_{i}$ values, namely the reverse order applied to compute the
AXp. Preliminary experiments conducted using a (naive heuristic) lexicographic
order over the features produced less succinct explanations.
Finally, notice that Algorithm 1 can be used to compute an AXp, i.e. locally-
minimal PAXp when $\delta=1$. Nevertheless, the polynomial time algorithm for
computing AXp’s proposed in [62] remains a better choice to use in case of
AXp’s than Algorithm 1 which runs in pseudo-polynomial time.
###### Example 10.
Let us consider again the NBC of the running example (Example 6) and
$\mathbf{v}=(\mathbf{t},\mathbf{f},\mathbf{f},\mathbf{f},\mathbf{t})$. The
corresponding XLC is shown in 7(b) (Example 7). Also, consider the AXp
$\\{1,2,5\\}$ of $\mathbf{v}$ and $\delta=0.85$. The resulting DP table for
${\mathcal{S}}=\\{1,2,5\\}$ is shown in Table 8. Note that for illustrating
small tables, we set the number of decimal places to zero (greater number of
decimal places, i.e. 1,2, etc, were tested and returned the same results).
(Also, note that the DP table reports “—” if the cell is not calculated during
the running of Algorithm 1.) Moreover, we convert the probabilities into
positive integers, so we add to each $w_{j}^{k}$ the complement of the most
negative $w_{j}^{k}$ plus 1. The resulting weights are shown in Figure 8.
Thus, we get
$\sum\nolimits_{i\in\\{1,2,3,4,5\\}}\sigma(x_{i},w_{i}^{1},w_{i}^{2})<17$.
Observe that the number of models $n_{\oplus}=C(5,16)$, and $C(5,16)$ is
calculated using $C(4,16-w^{2}_{5})=C(4,15)$, i.e. $C(4,15)=C(5,16)$ (feature
5 is fixed, so it is allowed to take only the value $w^{2}_{5}=1$). Next,
$C(4,15)=C(3,15-w^{1}_{4})+C(3,15-w^{2}_{4})=C(3,12)+C(3,14)$ (feature 4 is
free, so it is allowed to take any value of $\\{w^{1}_{4},w^{2}_{4}\\}$); the
recursion ends when k=1, namely for $C(1,5)=C(2,6)=n^{2}_{1}=1$,
$C(1,7)=C(2,7)=n^{2}_{1}=1$, $C(1,8)=C(2,8)=n^{2}_{1}=1$ and
$C(1,10)=C(2,11)=n^{2}_{1}=1$ (feature 1 is fixed and takes value
$w^{2}_{1}$). Next, Table 9 (resp. Table 10 and Table 11) report the resulting
DP table for ${\mathcal{S}}=\\{2,5\\}$ (resp. ${\mathcal{S}}=\\{1,5\\}$ and
${\mathcal{S}}=\\{1\\}$). It is easy to confirm that after dropping feature 2,
the precision of ${\mathcal{S}}=\\{1,5\\}$ becomes $87.5\%$, i.e.
$\frac{7}{8}=0.875>\delta$. Furthermore, observe that the resulting
${\mathcal{S}}$ when dropping feature 1 or 2 and 5, are not weak PAXp’s,
namely, the precision of $\\{2,5\\}$ is $\frac{6}{8}=0.75<\delta$ and the
precision of $\\{1\\}$ is $\frac{9}{16}=0.5625<\delta$. In summary, Algorithm
1 starts with ${\mathcal{S}}=\\{1,2,5\\}$, then at iteration #1, feature 1 is
tested and since $\\{2,5\\}$ is not a weak PAXp then 1 is kept in
${\mathcal{S}}$; at iteration #2, feature 2 is tested and since $\\{1,5\\}$ is
a weak PAXp, then ${\mathcal{S}}$ is updated (i.e. ${\mathcal{S}}=\\{1,5\\}$);
at iteration #3, feature 5 is tested and since $\\{1\\}$ is not a weak PAXp,
then 5 is saved in ${\mathcal{S}}$. As a result, the computed locally-minimal
PAXp is $\\{1,5\\}$.
We underline that we could initialize ${\mathcal{S}}$ to ${\mathcal{F}}$, in
which case the number of models would be 1. However, we opt instead to always
start from an AXp. In the example, the AXp is $\\{1,2,5\\}$ which, because it
is an AXp, the number of models must be 4 (i.e. $2^{2}$, since two features
are free).
For any proper subset of the AXp, with $r$ free variables, it must be the case
that the number of models is strictly less than $2^{r}$. Otherwise, we would
have an AXp as a proper subset of another AXp; but this would contradict the
definition of AXp. The fact that the number of models is strictly less than
$2^{r}$ is confirmed by the examples of subsets considered. It must also be
the case that if ${\mathcal{S}}^{\prime}\subseteq{\mathcal{S}}$, then the
number of models of ${\mathcal{S}}^{\prime}$ must not exceed the number of
models of ${\mathcal{S}}$. So, we can argue that there is monotonicity in the
number of models, but not in the precision.
$W$ | $w_{1}^{1}$ | $w_{1}^{2}$ | $w_{2}^{1}$ | $w_{2}^{2}$ | $w_{3}^{1}$ | $w_{3}^{2}$ | $w_{4}^{1}$ | $w_{4}^{2}$ | $w_{5}^{1}$ | $w_{5}^{2}$
---|---|---|---|---|---|---|---|---|---|---
16 | 7 | 1 | 1 | 6 | 3 | 6 | 1 | 3 | 7 | 1
Figure 8: #knapsack problem of Example 10 $k$ | $r$
---|---
0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16
1 | 0 | — | — | — | — | 1 | — | 1 | 1 | — | 1 | — | — | — | — | — | —
2 | 0 | — | — | — | — | — | 1 | — | 1 | 1 | — | 1 | — | — | — | — | —
3 | 0 | — | — | — | — | — | — | — | — | — | — | — | 2 | — | 2 | — | —
4 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 4 | —
5 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 4
Table 8: DP table for ${\mathcal{S}}=\\{1,2,5\\}$ (Example 10) $k$ | $r$
---|---
0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16
1 | 0 | — | — | — | — | 1 | — | 1 | 2 | — | 2 | — | — | — | — | — | —
2 | 0 | — | — | — | — | — | 1 | — | 1 | 2 | — | 2 | — | — | — | — | —
3 | 0 | — | — | — | — | — | — | — | — | — | — | — | 3 | — | 3 | — | —
4 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 6 | —
5 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 6
Table 9: DP table for ${\mathcal{S}}=\\{2,5\\}$ (Example 10) $k$ | $r$
---|---
0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16
1 | 0 | — | 1 | 1 | — | 1 | — | 1 | 1 | — | 1 | — | — | — | — | — | —
2 | 0 | — | — | — | — | — | 1 | — | 2 | 2 | — | 2 | — | — | — | — | —
3 | 0 | — | — | — | — | — | — | — | — | — | — | — | 3 | — | 4 | — | —
4 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 7 | —
5 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 7
Table 10: DP table for ${\mathcal{S}}=\\{1,5\\}$ (Example 10) $k$ | $r$
---|---
0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16
1 | 0 | 0 | 1 | 1 | 1 | 1 | — | 1 | 1 | — | 1 | — | — | — | — | — | —
2 | 0 | — | 0 | 1 | — | 1 | 1 | — | 2 | 2 | — | 2 | — | — | — | — | —
3 | 0 | — | — | — | — | — | 1 | — | 1 | — | — | — | 3 | — | 4 | — | —
4 | 0 | — | — | — | — | — | — | — | — | 2 | — | — | — | — | — | 7 | —
5 | 0 | — | — | — | — | — | — | — | — | — | — | — | — | — | — | — | 9
Table 11: DP table for ${\mathcal{S}}=\\{1\\}$ (Example 10)
## 6 Probabilistic Explanations for Other Families of Classifiers
This section investigates additional families of classifiers, namely those
either based on propositional languages [34] or based on graphs [35]. It
should be noted that some families of classifiers can be viewed as both
propositional languages and as graphs.
### 6.1 Propositional Classifiers
This section considers the families of classifiers represented as
propositional languages and studied in [34]. Examples include classifiers
based on d-DNNFs and SDDs.
###### Proposition 1.
For any language that allows conditioning (CD) in polynomial time, a locally-
minimal PAXp can be found in polynomial time (for all $\delta\in[0,1]$) if and
only if the language allows _counting_ (CT) in polynomial time.
###### Proof.
If the language allows CD and CT in polynomial time, Algorithm 1 finds a
locally-minimal PAXp in polynomial time. Conversely, testing whether
${\mathcal{X}}=\emptyset$ is a locally-minimal PAXp amounts to determining
whether the number of models of the classifier $\kappa$ is at least $\delta
2^{m}$, and hence a binary search on $\delta$ computes the number of models
(i.e. CT) in polynomial time. ∎
###### Proposition 2.
For any language that allows CT and CD in polynomial time, given an integer
$k$, the problem of deciding the existence of a weak PAXp of size at most $k$
is in NP.
###### Proof.
Let ${\mathcal{P}}\subseteq{\mathcal{F}}$ denote a guessed set of picked
features of size ${k^{\prime}}\leq{k}$, to be fixed. Since CT and CD run in
polynomial time, the left-hand side of (3.1), i.e.
$\textnormal{Pr}_{\mathbf{x}}(\kappa(\mathbf{x})=c\,|\,\mathbf{x}_{{\mathcal{X}}}=\mathbf{v}_{{\mathcal{X}}})$
can be computed in polynomial time. Consequently, we can decide in polynomial
time whether or not ${\mathcal{P}}$ is a weak PAXp, and so the decision
problem is in NP. ∎
###### Corollary 1.
For any language that allows CT and CD in polynomial time, the problem of
finding a minimum-size PAXp belongs to
$\textnormal{FP}^{\textnormal{NP}}[{\mathcal{O}}(\log m)]$.
###### Proof.
It suffices to run a binary search, on the number $k$ of features, using an NP
oracle that decides the decision problem introduced in the proof of
Proposition 2. This algorithm finds a smallest weak PAXp ${\mathcal{P}}$,
which is necessarily a PAXp (and hence a smallest PAXp) since no proper subset
of ${\mathcal{P}}$ can be a weak PAXp. ∎
Corollary 1 notably applies to the following languages:
* •
the language EADT of extended affine decision trees [56] which are a strict
generalisation of decision trees.
* •
the language d-DNNF and sublanguages of d-DNNF, such as SDD and OBDD.
### 6.2 Graph-Based Classifiers
This section considers the DG classifiers introduced in Section 2.3, and in
particular two restricted types of DGs: OMDDs and OBDDs. For OMDDs and OBDDs,
it is well-known that there are poly-time algorithms for model counting [23,
75]. Hence, we can compute a minimum-size PAXp with a logarithmic number of
calls to an NP oracle. We will describe one such algorithm in this section.
Moreover, a general method for counting models of DGs will be described as
well.
#### Counting models of OBDD/OMDD.
We describe next a dedicated algorithm for counting models of an OMDD, which
can also be used with OBDDs.
We associate an indicator $n_{p}$ to each node $p$, the indicator of the root
node being $n_{1}$, such that the value of $n_{1}$ represents the number of
models consistent with the class $c$. And let $s_{i}$ be the boolean selector
of feature $i$ such that $s_{i}=1$ if feature $i$ is included in PAXp,
otherwise it is free. Besides, we assume that the feature-index order is
increasing top-down, and the leftmost outgoing edge of a non-terminal node is
consistent with the given instance $\mathbf{v}$.
1. 1.
For a terminal node $p$, $n_{p}=1$ if the label of node $p$ is consistent with
the class $c$; otherwise $n_{p}=0$.
2. 2.
For a non-terminal node $p$ label with feature $i$, that has $k$ child nodes
$q_{1},\ldots,q_{k}$, let $q$ be an arbitrary child node of $p$ labeled with
feature $j$, and edge $(p,q_{1})$ be the leftmost edge. By assumption, since
$i<j$, there may exist some untested features $i<l<j$ along the outgoing edge
$(p,q)$. Let $b_{p,q}$ be the indicator of the edge $(p,q)$, we have:
$b_{p,q}=n_{q}\times\prod_{i<l<j}ite(s_{l},1,\mathbb{D}_{l})\times\lvert(p,q)\rvert$,
so the indicator of node $p$ is $n_{p}=ite(s_{i},b_{p,q_{1}},\sum_{q_{1}\leq
l\leq q_{k}}b_{p,l})$. It should be noted that multiple edges may exist; we
use $\lvert(p,q)\rvert$ to denote the number of edges between node $p$ and
$q$.
3. 3.
The number of models consistent with the class $c$ is $n_{1}$.
In the case of $\lvert{\mathcal{K}}\rvert=2$ and $\mathbb{F}=\mathbb{B}^{m}$,
this algorithm is still applicable to OBDDs with some minor modifications.
Note that each non-terminal node of an OBDD has exactly two outgoing edges and
multiple edges between two nodes are not allowed (otherwise, the OBDD is not
_reduced_); this means that $\mathbb{D}_{l}=\mathbb{B}$, and
$\lvert(p,q)\rvert=1$.
#### Counting models for unrestricted DGs.
For unrestricted DG classifiers, given Section 2.3, any path connecting the
root to a terminal node is consistent with some input, and no two paths are
consistent with the same input. As a result, in the case of unrestricted DG
classifiers, it suffices to enumerate all paths, and count the models
associated with the path. Of course, the downside is that the number of paths
is worst-case exponential on the size of the graph. To alleviate this, an
approach similar to the one outlined above can be considered.
## 7 Experiments
This section reports the experimental results on computing relevant sets for
the classifiers studied in the earlier sections, namely: decision trees, naive
Bayes classifiers and graph-based classifiers (in our case OMDDs). For each
case study, we describe the prototype implementation and the used benchmarks;
we show a table that summarizes the results and then we discuss the results.
Furthermore, for the case of DTs, the evaluation includes a comparison with
the model-agnostic explainer Anchor [82], aiming at assessing not only the
succinctness and precision of computed explanations but also the scalability
of our solution. (Observe that for the case of NBCs, earlier work on computing
AXp’s [62] have compared their approach with the heuristic methods, e.g.
Anchor, SHAP, and show that the latter are slower and do not show a strong
correlation between features of their explanations and common features
identified from AXp’s. As a result, the comparison with Anchor is restricted
to the case of DTs.)
All experiments were conducted on a MacBook Air with a 1.1GHz Quad-Core Intel
Core i5 CPU with 16 GByte RAM running macOS Monterey.
Dataset | | | | $\bm{\mathsf{MinPAXp}}$ | $\bm{\mathsf{LmPAXp}}$ | Anchor
---|---|---|---|---|---|---
DT | Path | $\bm{\delta}$ | Length | Prec | Time | Length | Prec | m⊆ | Time | D | Length | Prec | Time
N | A | M | m | avg | | M | m | avg | avg | avg | M | m | avg | avg | | avg | | M | m | avg | F/∈P | avg | avg
| | | | | | $100$ | $11$ | $3$ | $6.8$ | 100 | $2.34$ | $11$ | $3$ | $6.9$ | 100 | 100 | $0.00$ | d | $12$ | $2$ | $7.0$ | $26.8$ | 76.8 | $0.96$
adult | $1241$ | $89$ | $14$ | $3$ | $10.7$ | $95$ | $11$ | $3$ | $6.2$ | 98.4 | $5.36$ | $11$ | $3$ | $6.3$ | 98.6 | 99.0 | $0.01$ | u | $12$ | $3$ | $10.0$ | $29.4$ | 93.7 | $2.20$
| | | | | | $90$ | $11$ | $2$ | $5.6$ | 94.6 | $4.64$ | $11$ | $2$ | $5.8$ | 95.2 | 96.4 | $0.01$ | | | | | | |
| | | | | | $100$ | $12$ | $1$ | $4.4$ | 100 | $0.35$ | $12$ | $1$ | $4.4$ | 100 | 100 | $0.00$ | d | $31$ | $1$ | $4.8$ | $58.1$ | 32.9 | $3.10$
dermatology | $71$ | $100$ | $13$ | $1$ | $5.1$ | $95$ | $12$ | $1$ | $4.1$ | 99.7 | $0.37$ | $12$ | $1$ | $4.1$ | 99.7 | 99.3 | $0.00$ | u | $34$ | $1$ | $13.1$ | $43.2$ | 87.2 | $25.13$
| | | | | | $90$ | $11$ | $1$ | $4.0$ | 98.8 | $0.35$ | $11$ | $1$ | $4.0$ | 98.8 | 100 | $0.00$ | | | | | | |
| | | | | | $100$ | $12$ | $2$ | $4.8$ | 100 | $0.93$ | $12$ | $2$ | $4.9$ | 100 | 100 | $0.00$ | d | $36$ | $2$ | $7.9$ | $44.8$ | 69.4 | $1.94$
kr-vs-kp | $231$ | $100$ | $14$ | $3$ | $6.6$ | $95$ | $11$ | $2$ | $3.9$ | 98.1 | $0.97$ | $11$ | $2$ | $4.0$ | 98.1 | 100 | $0.00$ | u | $12$ | $2$ | $3.6$ | $16.6$ | 97.3 | $1.81$
| | | | | | $90$ | $10$ | $2$ | $3.2$ | 95.4 | $0.92$ | $10$ | $2$ | $3.3$ | 95.4 | 99.0 | $0.00$ | | | | | | |
| | | | | | $100$ | $12$ | $4$ | $8.2$ | 100 | $16.06$ | $11$ | $4$ | $8.2$ | 100 | 100 | $0.00$ | d | $16$ | $3$ | $13.2$ | $43.1$ | 71.3 | $12.22$
letter | $3261$ | $93$ | $14$ | $4$ | $11.8$ | $95$ | $12$ | $4$ | $8.0$ | 99.6 | $18.28$ | $11$ | $4$ | $8.0$ | 99.5 | 100 | $0.00$ | u | $16$ | $3$ | $13.7$ | $47.3$ | 66.3 | $10.15$
| | | | | | $90$ | $12$ | $4$ | $7.7$ | 97.7 | $16.35$ | $10$ | $4$ | $7.8$ | 97.8 | 100 | $0.00$ | | | | | | |
| | | | | | $100$ | $14$ | $3$ | $6.4$ | 100 | $0.92$ | $14$ | $3$ | $6.5$ | 100 | 100 | $0.00$ | d | $35$ | $2$ | $8.6$ | $55.4$ | 33.6 | $5.43$
soybean | $219$ | $100$ | $16$ | $3$ | $7.3$ | $95$ | $14$ | $3$ | $6.4$ | 99.8 | $0.95$ | $14$ | $3$ | $6.4$ | 99.8 | 100 | $0.00$ | u | $35$ | $3$ | $19.2$ | $66.0$ | 75.0 | $38.96$
| | | | | | $90$ | $14$ | $3$ | $6.1$ | 98.1 | $0.94$ | $14$ | $3$ | $6.1$ | 98.2 | 98.5 | $0.00$ | | | | | | |
| | | | | | $0$ | $12$ | $3$ | $7.4$ | 100 | $1.23$ | $12$ | $3$ | $7.5$ | 100 | 100 | $0.01$ | d | $38$ | $2$ | $6.3$ | $65.3$ | 63.3 | $24.12$
spambase | $141$ | $99$ | $14$ | $3$ | $8.5$ | $95$ | $9$ | $1$ | $3.7$ | 96.1 | $2.16$ | $9$ | $1$ | $3.8$ | 96.5 | 100 | $0.01$ | u | $57$ | $3$ | $28.0$ | $86.2$ | 65.3 | $834.70$
| | | | | | $90$ | $6$ | $1$ | $2.4$ | 92.4 | $2.15$ | $8$ | $1$ | $2.4$ | 92.2 | 100 | $0.01$ | | | | | | |
| | | | | | $100$ | $12$ | $3$ | $6.2$ | 100 | $2.01$ | $11$ | $3$ | $6.2$ | 100 | 100 | $0.01$ | d | $40$ | $2$ | $16.5$ | $80.6$ | 32.2 | $532.42$
texture | $257$ | $100$ | $13$ | $3$ | $6.6$ | $95$ | $11$ | $3$ | $5.4$ | 99.3 | $2.19$ | $11$ | $3$ | $5.4$ | 99.4 | 100 | $0.01$ | u | $40$ | $5$ | $17.5$ | $84.4$ | 31.6 | $402.07$
| | | | | | $90$ | $11$ | $3$ | $5.4$ | 98.5 | $2.20$ | $11$ | $3$ | $5.4$ | 99.4 | 100 | $0.01$ | | | | | | |
Table 12: Assessing explanations of $\mathsf{MinPAXp}$, $\mathsf{LmPAXp}$ and
Anchor for DTs. (For each dataset, we run the explainers on 500 samples
randomly picked or all samples if there are less than 500.) In column DT, N
and A denote, resp., the number of nodes and the training accuracy of the DT.
Column $\bm{\delta}$ reports (in %) the value of the threshold $\delta$. In
column Path, avg (resp. M and m) denotes the average (resp. max. and min.)
depth of paths consistent with the instances. In column Length, avg (resp. M
and m) denotes the average (resp. max. and min.) length of the explanations;
and F/∈P denotes the avg. % of features in Anchor’s explanations that do not
belong to the consistent paths. Prec reports (in %) the average precision
(defined in (9)) of resulting explanations. m⊆ shows the number in (%) of
LmPAXp’s that are subset-minimal, i.e. PAXp’s. Time reports (in seconds) the
average runtime to compute an explanation. Finally, D indicates which
distribution is applied on data given to Anchor: either data distribution
(denoted by d) or uniform distribution (denoted by u).
Dataset | (#F #I) | NBC | $\bm{\mathsf{AXp}}$ | | $\bm{\mathsf{LmPAXp}}_{\leq 9}$ | $\bm{\mathsf{LmPAXp}}_{\leq 7}$ | $\bm{\mathsf{LmPAXp}}_{\leq 4}$
---|---|---|---|---|---|---|---
A% | Length | $\delta$ | Length | Precision | W% | Time | Length | Precision | W% | Time | Length | Precision | W% | Time
adult | (13 | 200) | 81.37 | 6.8$\pm$ 1.2 | $98$ | 6.8$\pm$ 1.1 | 100$\pm$ 0.0 | $100$ | $0.003$ | 6.3$\pm$ 0.9 | 99.61$\pm$ 0.6 | $96$ | $0.023$ | 4.8$\pm$ 1.3 | 98.73$\pm$ 0.5 | $48$ | $0.059$
$95$ | 6.8$\pm$ 1.1 | 99.99$\pm$ 0.2 | $100$ | $0.074$ | 5.9$\pm$ 1.0 | 98.87$\pm$ 1.8 | $99$ | $0.058$ | 3.9$\pm$ 1.0 | 96.93$\pm$ 1.1 | $80$ | $0.071$
$93$ | 6.8$\pm$ 1.1 | 99.97$\pm$ 0.4 | $100$ | $0.104$ | 5.7$\pm$ 1.3 | 98.34$\pm$ 2.6 | $100$ | $0.086$ | 3.4$\pm$ 0.9 | 95.21$\pm$ 1.6 | $90$ | $0.093$
$90$ | 6.8$\pm$ 1.1 | 99.95$\pm$ 0.6 | $100$ | $0.164$ | 5.5$\pm$ 1.4 | 97.86$\pm$ 3.4 | $100$ | $0.100$ | 3.0$\pm$ 0.8 | 93.46$\pm$ 1.5 | $94$ | $0.103$
agaricus | (23 | 200) | 95.41 | 10.3$\pm$ 2.5 | $98$ | 7.7$\pm$ 2.7 | 99.12$\pm$ 0.8 | $92$ | $0.593$ | 6.4$\pm$ 3.0 | 98.75$\pm$ 0.6 | $87$ | $0.763$ | 6.0$\pm$ 3.1 | 98.67$\pm$ 0.5 | $29$ | $0.870$
$95$ | 6.9$\pm$ 3.1 | 97.62$\pm$ 2.1 | $95$ | $0.954$ | 5.3$\pm$ 3.2 | 96.59$\pm$ 1.6 | $92$ | $1.273$ | 4.8$\pm$ 3.3 | 96.24$\pm$ 1.2 | $55$ | $1.217$
$93$ | 6.5$\pm$ 3.1 | 96.65$\pm$ 2.8 | $95$ | $1.112$ | 4.8$\pm$ 3.1 | 95.38$\pm$ 1.9 | $93$ | $1.309$ | 4.3$\pm$ 3.1 | 94.92$\pm$ 1.3 | $64$ | $1.390$
$90$ | 5.9$\pm$ 3.3 | 94.95$\pm$ 4.1 | $96$ | $1.332$ | 4.0$\pm$ 3.0 | 92.60$\pm$ 2.8 | $95$ | $1.598$ | 3.6$\pm$ 2.8 | 92.08$\pm$ 1.7 | $76$ | $1.830$
chess | (37 | 200) | 88.34 | 12.1$\pm$ 3.7 | $98$ | 8.1$\pm$ 4.1 | 99.27$\pm$ 0.6 | $64$ | $0.383$ | 5.9$\pm$ 4.9 | 98.70$\pm$ 0.4 | $64$ | $0.454$ | 5.7$\pm$ 5.0 | 98.65$\pm$ 0.4 | $46$ | $0.457$
$95$ | 7.7$\pm$ 3.8 | 98.51$\pm$ 1.4 | $68$ | $0.404$ | 5.5$\pm$ 4.4 | 97.90$\pm$ 0.9 | $64$ | $0.483$ | 5.3$\pm$ 4.5 | 97.85$\pm$ 0.8 | $46$ | $0.478$
$93$ | 7.3$\pm$ 3.5 | 97.56$\pm$ 2.4 | $68$ | $0.419$ | 5.0$\pm$ 4.1 | 96.26$\pm$ 2.2 | $64$ | $0.485$ | 4.8$\pm$ 4.1 | 96.21$\pm$ 2.1 | $64$ | $0.493$
$90$ | 7.3$\pm$ 3.5 | 97.29$\pm$ 2.9 | $70$ | $0.413$ | 4.9$\pm$ 4.0 | 95.99$\pm$ 2.6 | $64$ | $0.483$ | 4.8$\pm$ 4.0 | 95.93$\pm$ 2.5 | $64$ | $0.543$
vote | (17 | 81) | 89.66 | 5.3$\pm$ 1.4 | $98$ | 5.3$\pm$ 1.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.3$\pm$ 1.3 | 99.95$\pm$ 0.2 | $100$ | $0.007$ | 4.6$\pm$ 1.1 | 99.60$\pm$ 0.4 | $64$ | $0.014$
$95$ | 5.3$\pm$ 1.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.3$\pm$ 1.3 | 99.93$\pm$ 0.3 | $100$ | $0.008$ | 4.1$\pm$ 1.0 | 98.25$\pm$ 1.7 | $64$ | $0.018$
$93$ | 5.3$\pm$ 1.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.2$\pm$ 1.3 | 99.78$\pm$ 1.1 | $100$ | $0.012$ | 4.1$\pm$ 0.9 | 98.10$\pm$ 1.9 | $64$ | $0.018$
$90$ | 5.3$\pm$ 1.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.2$\pm$ 1.3 | 99.78$\pm$ 1.1 | $100$ | $0.012$ | 4.0$\pm$ 1.2 | 97.24$\pm$ 3.1 | $64$ | $0.022$
kr-vs-kp | (37 | 200) | 88.07 | 12.2$\pm$ 3.9 | $98$ | 7.8$\pm$ 4.2 | 99.19$\pm$ 0.5 | $64$ | $0.387$ | 6.5$\pm$ 4.7 | 98.99$\pm$ 0.4 | $64$ | $0.427$ | 6.1$\pm$ 4.9 | 98.88$\pm$ 0.3 | $43$ | $0.457$
$95$ | 7.3$\pm$ 3.9 | 98.29$\pm$ 1.4 | $64$ | $0.416$ | 6.0$\pm$ 4.3 | 97.89$\pm$ 1.1 | $64$ | $0.453$ | 5.5$\pm$ 4.5 | 97.79$\pm$ 0.9 | $43$ | $0.462$
$93$ | 6.9$\pm$ 3.5 | 97.21$\pm$ 2.5 | $69$ | $0.422$ | 5.6$\pm$ 3.8 | 96.82$\pm$ 2.2 | $64$ | $0.448$ | 5.2$\pm$ 4.0 | 96.71$\pm$ 2.1 | $43$ | $0.468$
$90$ | 6.8$\pm$ 3.5 | 96.65$\pm$ 3.1 | $69$ | $0.418$ | 5.4$\pm$ 3.8 | 95.69$\pm$ 3.0 | $64$ | $0.468$ | 5.0$\pm$ 4.0 | 95.59$\pm$ 2.8 | $61$ | $0.487$
mushroom | (23 | 200) | 95.51 | 10.7$\pm$ 2.3 | $98$ | 7.5$\pm$ 2.4 | 98.99$\pm$ 0.7 | $90$ | $0.641$ | 6.5$\pm$ 2.6 | 98.74$\pm$ 0.5 | $83$ | $0.751$ | 6.3$\pm$ 2.7 | 98.70$\pm$ 0.4 | $18$ | $0.828$
$95$ | 6.5$\pm$ 2.6 | 97.35$\pm$ 1.8 | $96$ | $1.011$ | 5.1$\pm$ 2.5 | 96.52$\pm$ 1.0 | $90$ | $1.130$ | 5.0$\pm$ 2.5 | 96.39$\pm$ 0.8 | $54$ | $1.113$
$93$ | 5.8$\pm$ 2.8 | 95.77$\pm$ 2.7 | $96$ | $1.257$ | 4.4$\pm$ 2.5 | 94.67$\pm$ 1.6 | $94$ | $1.297$ | 4.2$\pm$ 2.4 | 94.48$\pm$ 1.3 | $65$ | $1.324$
$90$ | 5.3$\pm$ 3.0 | 94.01$\pm$ 3.9 | $97$ | $1.455$ | 3.8$\pm$ 2.3 | 92.36$\pm$ 2.2 | $96$ | $1.543$ | 3.6$\pm$ 2.2 | 92.07$\pm$ 1.6 | $76$ | $1.650$
threeOf9 | (10 | 103) | 83.13 | 4.2$\pm$ 0.4 | $98$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $78$ | $0.001$
$95$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.0$\pm$ 0.2 | 99.23$\pm$ 1.4 | $100$ | $0.002$
$93$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 3.9$\pm$ 0.2 | 99.20$\pm$ 1.5 | $100$ | $0.002$
$90$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.2$\pm$ 0.4 | 100$\pm$ 0.0 | $100$ | $0.000$ | 3.8$\pm$ 0.4 | 98.29$\pm$ 3.3 | $100$ | $0.003$
xd6 | (10 | 176) | 81.36 | 4.5$\pm$ 0.9 | $98$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $73$ | $0.001$
$95$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $73$ | $0.001$
$93$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.3$\pm$ 0.4 | 98.30$\pm$ 2.7 | $73$ | $0.001$
$90$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.5$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.3$\pm$ 0.4 | 98.30$\pm$ 2.7 | $73$ | $0.002$
mamo | (14 | 53) | 80.21 | 4.9$\pm$ 0.8 | $98$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.6$\pm$ 0.6 | 99.66$\pm$ 0.5 | $53$ | $0.007$
$95$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 3.9$\pm$ 0.6 | 97.80$\pm$ 1.6 | $85$ | $0.009$
$93$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 3.9$\pm$ 0.6 | 97.68$\pm$ 1.7 | $85$ | $0.009$
$90$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 4.9$\pm$ 0.7 | 100$\pm$ 0.0 | $100$ | $0.000$ | 3.6$\pm$ 0.8 | 96.18$\pm$ 3.2 | $96$ | $0.011$
tumor | (16 | 104) | 83.21 | 5.3$\pm$ 0.9 | $98$ | 5.3$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.2$\pm$ 0.7 | 99.96$\pm$ 0.2 | $100$ | $0.008$ | 4.1$\pm$ 0.7 | 99.41$\pm$ 0.5 | $91$ | $0.012$
$95$ | 5.3$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.2$\pm$ 0.6 | 99.83$\pm$ 0.7 | $100$ | $0.012$ | 3.2$\pm$ 0.6 | 96.02$\pm$ 1.5 | $94$ | $0.016$
$93$ | 5.3$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.2$\pm$ 0.6 | 99.74$\pm$ 1.2 | $100$ | $0.014$ | 3.1$\pm$ 0.7 | 95.50$\pm$ 1.4 | $95$ | $0.016$
$90$ | 5.3$\pm$ 0.8 | 100$\pm$ 0.0 | $100$ | $0.000$ | 5.1$\pm$ 0.7 | 99.67$\pm$ 1.4 | $100$ | $0.016$ | 3.0$\pm$ 0.6 | 95.30$\pm$ 1.6 | $95$ | $0.017$
Table 13: Assessing $\mathsf{LmPAXp}$ explanations for NBCs. Columns #F and
#I show, respectively, number of features and tested instances in the Dataset.
Column A% reports (in %) the training accuracy of the classifier. Column
$\delta$ reports (in %) the value of the parameter $\delta$.
$\bm{\mathsf{LmPAXp}}_{\leq 9}$, $\bm{\mathsf{LmPAXp}}_{\leq 7}$ and
$\bm{\mathsf{LmPAXp}}_{\leq 4}$ denote, respectively, LmPAXp’s of (target)
length 9, 7 and 4. Columns Length and Precision report, respectively, the
average explanation length and the average explanation precision ($\pm$
denotes the standard deviation). W% shows (in %) the number of success/wins
where the explanation size is less than or equal to the target size. Finally,
the average runtime to compute an explanation is shown (in seconds) in Time.
(Note that the reported average time is the mean of runtimes for instances for
which we effectively computed an approximate explanation, namely instances
that have AXp’s of length longer than the target length; whereas for the
remaining instances the trimming process is skipped and the runtime is 0 sec,
thus we exclude them when calculating the average.)
### 7.1 Case Study 1: Decision Trees
#### Prototype implementation.
A prototype implementation888All sources implemented in these experiments will
be publicly available after the paper gets accepted. of the proposed
algorithms for DTs was developed in Python; whenever necessary, it instruments
oracle calls to the well-known SMT solver z3999https://github.com/Z3Prover/z3/
[24] as described in Section 4 . Hence, the prototype implements the
$\mathsf{LmPAXp}$ procedure outlined in Algorithm 1 and augmented with a
heuristic that orders the features in ${\mathcal{X}}$. The idea consists in
computing the precision loss of the overapproximation of each
${\mathcal{X}}\setminus\\{j\\}$ and then sorting the features from the less to
the most important one. This strategy often allows us to obtain the closest
superset to a PAXp, in contrast to the simple lexicographic order applied over
${\mathcal{X}}$. (Recall that ${\mathcal{X}}$ is initialized to the set of
features involved in the decision path.) Algorithm $\mathsf{MinPAXp}$ outlined
in Section 4.3 implements the two (multiplication- and addition-based) SMT
encodings. Nevertheless, preliminary results show that both encodings perform
similarly, with some exceptions where the addition-based encoding is much
larger and so slower. Therefore, the results reported below refer only to the
multiplication-based encoding.
#### Benchmarks.
The benchmarks used in the experiments comprise publicly available and widely
used datasets obtained from the UCI ML Repository [95]. All the DTs are
trained using the learning tool _IAI_ (_Interpretable AI_) [11, 38]. The
maximum depth parameter in IAI is set to 16. As the baseline, we ran Anchor
with the default explanation precision of 0.95. Two assessments are performed
with Anchor: (i) with the original training data101010The same training set
used to learn the model. that follows the data distribution; (ii) with using
sampled data that follows a uniform distribution. Our setup assumes that all
instances of the feature space are equally possible, and so there is no
assumed probability distribution over the features. Therefore in order to be
fair with Anchor, we further assess Anchor with uniformly sampled data. (Also,
we point out that the implementation of Anchor demonstrates that it can
generate samples that do not belong to the input distribution. Thus, there is
no guarantee that these samples come from the input distribution.) Also, the
prototype implementation was tested with varying the threshold $\delta$ while
Anchor runs guided by its own metric.
#### Results.
Table 12 summarizes the results of our experiments for the case study of DTs.
One can observe that $\mathsf{MinPAXp}$ and $\mathsf{LmPAXp}$ compute succinct
explanations (i.e. of average size $7\pm 2$ [69]), for the majority of tested
instances across all datasets, noticeably shorter than consistent-path
explanations. More importantly, the computed explanations are trustworthy and
show good quality precision, e.g. dermatology, soybean and texture show
average precisions greater than 98% for all values of $\delta$. Additionally,
the results clearly demonstrate that our proposed SMT encoding scales for deep
DTs with runtimes on average less than 20 sec for the largest encodings while
the runtimes of $\mathsf{LmPAXp}$ are negligible, never exceeding 0.01 sec.
Also, observe from the table (see column m⊆) that the over-approximations
computed by $\mathsf{LmPAXp}$ are often subset-minimal PAXp’s, and often as
short as computed MinPAXp’s. This confirms empirically the advantages of
computing LmPAXp’s, i.e. in practice one may rely on the computation of
LmPAXp’s, which pays off in terms of (1) performance, (2) sufficiently high
probabilistic guarantees of precision, and (3) good quality over-approximation
of subset-minimal PAXp’s. In contrast, Anchor is unable to provide precise and
succinct explanations in both settings of data and uniform distribution.
Moreover, we observe that Anchor’s explanations often include features that
are not involved in the consistent path, e.g. for texture less than 20% of an
explanation is shared with the consistent path. (This trend was also pointed
out by [39].) In terms of average runtime, Anchor is overall slower, being
outperformed by the computation of $\mathsf{LmPAXp}$ by several orders of
magnitude.
Overall, the experiments demonstrate that our approach efficiently computes
succinct and provably precise explanations for large DTs. The results also
substantiate the limitations of model-agnostic explainers, both in terms of
explanation quality and computation time.
### 7.2 Case Study 2: Naive Bayes Classifiers
#### Prototype implementation.
A prototype implementation of the proposed approach for computing relevant
sets for NBCs was developed in Python. To compute AXp’s, we use the Perl
script implemented by [62]111111Publicly available from:
https://github.com/jpmarquessilva/expxlc. The prototype implementation was
tested with varying thresholds $\delta\in\\{0.90,0.93,0.95,0.98\\}$. When
converting probabilities from real values to integer values, the selected
number of decimal places is 3. (As outlined earlier, we observed that there is
a negligible accuracy loss from using three decimal places). In order to
produce explanations of size admissible for the cognitive capacity of human
decision makers [69], we selected three different target sizes for the
explanations to compute: 9, 7 and 4, and we computed a LmPAXp for the input
instance when its AXp ${\mathcal{X}}$ is larger than the target size (recall
that ${\mathcal{S}}$ is initialized to ${\mathcal{X}}$); otherwise we consider
that the AXp is succinct and the explainer returns ${\mathcal{X}}$. For
example, assume the target size is 7, an instance $\mathbf{v}_{1}$ with an
$\mathsf{AXp}$ ${\mathcal{S}}_{1}$ of 5 features and a second instance
$\mathbf{v}_{2}$ with an $\mathsf{AXp}$ ${\mathcal{S}}_{2}$ of 8 features,
then for $\mathbf{v}_{1}$ the output will be ${\mathcal{S}}_{1}$ and for
$\mathbf{v}_{2}$ the output will be a subset of ${\mathcal{S}}_{2}$.
#### Benchmarks.
The benchmarks used in this evaluation originate from the UCI ML Repository
[95] and Penn ML Benchmarks [79]. The number of training data (resp. features)
in the target datasets varies from 336 to 14113 (resp. 10 to 37) and on
average is 3999.1 (resp. 20.0). All the NBCs are trained using the learning
tool _scikit-learn_ [88]. The data split for training and test data is set to
80% and 20%, respectively. Model accuracies are above 80% for the training
accuracy and above 75% for the test accuracy. For each dataset, we run the
explainer on 200 instances randomly picked from the test data or on all
instances if there are less than 200.
#### Results.
Table 13 summarizes the results of our experiments for the case study of NBCs.
For all tested values for the parameter threshold $\delta$ and target size,
the reported results show the sizes and precisions of the computed
explanations. As can be observed for all considered settings, the approximate
explanations are succinct, in particular the average sizes of the explanations
are invariably lower than the target sizes. Moreover, theses explanations
offer strong guarantees of precision, as their average precisions are strictly
greater than $\delta$ with significant gaps (e.g. above 97%, in column
$\mathsf{LmPAXp}_{\leq 7}$, for datasets _adult_ , _vote_ , _threeOf9_ , _xd6_
, _mamo_ and _tumor_ and above 95% for _chess_ and _kr-vs-kp_). An important
observation from the results, is the gain of succinctness (explanation size)
|
# Histories without collapse
Anthony Sudbery1
Department of Mathematics, University of York,
Heslington, York, England YO10 5DD
1<EMAIL_ADDRESS>
###### Abstract
This paper is a comparison of two theories of the probability of a history in
quantum mechanics. One is derived from Copenhagen quantum mechanics using the
projection postulate and is the basis of the “consistent histories”
interpretation; the other is based on a proposal by Bell, originally for the
“pilot state” theory but here applied to pure unitary quantum mechanics. The
first can be used for a wider class of histories but depends on the projection
postulate, or “collapse”, which is widely held to be an unsatisfactory feature
of the theory; the second can be used in a theory of the universal state
vector without collapse. We examine a simple model based on Wigner’s friend,
in which Bell’s model and the projection postulate give different
probabilities for the histories of a sentient system. We also examine the
Frauchiger-Renner extension of this model, in which comparison of the two
calculations of histories throws light on the contradiction found by
Frauchiger and Renner. By extending the model to equip the observer with a
memory, we reduce the probability of histories to the use of the Born rule at
a single time, and show that the Born rule, with the memory, gives the same
result as applying projection in the course of the history, because of
entanglement with the memory. Entanglement implements collapse. We discuss the
implications of this for the use of histories in quantum cosmology.
Keywords: quantum mechanics, foundations, histories
Declarations
Funding: This research was not supported by any funding.
Conflicts of interest: I am not aware of any conflict of interest.
Availabiilty of data and material: There is no data or material to be made
available.
Code availability: There is no relevant code.
## 1 Introduction
In quantum mechanics one often wants to know the probability of a sequence of
measurements, or the conditional probability of a measurement result at one
time given a result at an earlier time. In other words, one wants the
probability of a history. In the form of quantum mechanics presented in most
textbooks (the “Copenhagen interpretation”), this is managed by assuming that
the unitary evolution is interrupted by a projection, or “collapse”, at the
times of the sequence. Other interpretations of quantum mechanics, usually
associated with the name of Everett, deny that such collapses take place at a
fundamental level, and assert that all evolution follows a continuous unitary
process. If this is so, however, it is not obvious how to calculate the
probability of a history: the only point at which probability appears to enter
the theory is via the Born rule at one time. However, in 1984 John Bell [3],
in a study of quantum field theory in the context of Bohmian “pilot state”
theory, proposed a process which can be generalised, detached from the Bohmian
context of “beables” (or “hidden variables”), and applied to any history in
the different ontology of Everettian quantum mechanics. In this paper the
generalisation of Bell’s process is presented and applied to some examples
deriving from the scenario of Wigner’s friend, which bring out particularly
clearly the issues related to collapse. The probabilities derived from Bell’s
process are compared with those obtained from Copenhagen quantum mechanics.
Finally, we consider how the presence of memory makes it possible to reduce
the probability of a history to a single-time probability, using the Born
rule.
The paper is organised as follows. In Section 2 we set up the general
formalism for a discussion of histories. The next three sections set out the
two methods (Copenhagen and Bell) for calculating of the probability of a
history, with a discussion (Section 4) of the problems of histories and
probability in pure unitary quantum mechanics. In Section 5 the solution of
these problems based on Bell’s process is introduced and applied to a variant
of the Wigner’s friend scenario due to Brukner; the results are compared with
those of the Copenhagen calculation. In Section 6 the two methods are applied
to a more elaborate version of the Wigner’s friend scenario due to Frauchiger
and Renner. Section 7 shows the effects of memory in this scenario, and the
paper concludes with a discussion of the results.
## 2 Histories and probabilities
A _history_ is a sequence of events $E_{0},\ldots,E_{n}$ at times
$t_{0},\ldots,t_{n}$. In quantum mechanics, as in statistics and law (and
unlike the study of human history, relativity theory or high-energy physics)
an event is defined by a proposition and therefore, in quantum mechanics, by a
closed subspace of Hilbert space, or equivalently by the orthogonal projection
onto this subspace. A history is therefore given by a time-labelled sequence
$h=\big{(}(\Pi_{0},t_{0}),\ldots,(\Pi_{n},t_{n})\big{)}.$ (2.1)
of events $(\Pi_{j},t_{j})$ each defined by a projection $\Pi_{j}$ and a time
$t_{j}$. For simplicity, we will assume that the system starts in a pure state
$|\Psi_{0}\rangle$ at time $t=0$, so that
$\Pi_{0}=|\Psi_{0}\rangle\langle\Psi_{0}|$ and $t_{0}=0$. The theory should
give a probability for this history, i.e. for the truth of the proposition
represented by the succession of projections $\Pi_{j}$ at the times $t_{j}$. A
closely related quantity is the _conditional_ probability of an event at one
time conditional on an event at an earlier time. However, there are competing
accounts of this probability. We now consider two of these accounts.
## 3 The Copenhagen probability
In the version of quantum mechanics presented in most textbooks, the
probability of a history like $h$ in (2.1) is obtained by applying the Born
rule to successive measurements of the projectors $\Pi_{j}$, assuming
orthogonal projection of the state after each measurement and unitary
evolution according to the Schrödinger equation between the measurements. This
gives the probability
$\displaystyle P_{\text{C}}(h)$
$\displaystyle=\langle\Psi_{0}|\widetilde{\Pi}_{1}\ldots\widetilde{\Pi}_{n}\ldots\widetilde{\Pi}_{1}|\Psi_{0}\rangle$
(3.1) $\displaystyle=\langle K(h)|K(h)\rangle$ (3.2)
where
$\widetilde{\Pi}_{j}=\mathrm{e}^{\mathrm{i}Ht_{j}}\Pi_{j}\mathrm{e}^{-\mathrm{i}Ht_{j}}$
(taking $\hbar=1$) and
$|K(h)\rangle=\widetilde{\Pi}_{n}\ldots\widetilde{\Pi}_{1}|\Psi_{0}\rangle.$
(3.3)
In the consistent-histories [12] and temporal-logic [19] interpretations of
quantum mechanics the formula (3.1) is extended to include conjunctions and
disjunctions of histories (though this is not essential in the latter
interpretation [21]). Logically desirable properties of probability then
require a consistency condition on the set of histories, namely that
$|K(h_{1})\rangle$ and $|K(h_{2})\rangle$ are orthogonal if $h_{1}$ and
$h_{2}$ are distinct histories. However, if this condition is not satisfied
the formula (3.1) for single histories is still meaningful provided that the
evolution in each time interval $[t_{j},t_{j+1}]$ is purely unitary.
Recent discussions of conditional probabilities and histories ([8, 11, 2])
have used the Page-Wootters mechanism to deal with time. The resulting
calculations of conditional probabilities are either equivalent to assuming
collapse at the events of a history, or proceed by examining the contents of a
memory at a single time after all the times of the history. The latter
procedure, as will be discussed in Section 7, gives the same results as
assuming collapse.
## 4 Probability in collapse-free quantum mechanics
The derivation of $P_{\text{C}}$ assumes that, in addition to unitary
evolution according to the Schrödinger equation, the quantum state undergoes
projection at the events in the history. Such projection is denied in some
influential interpretations of quantum mechanics, in particular the relative-
state interpretation of Everett and Wheeler [9, 23]. In this interpretation
the truth about the world and its development in time is given by a universal,
time-dependent state vector $|\Psi(t)\rangle$, and nothing else. It is
possible to use the Born rule to deduce that, given a projector $\Pi$, the
probability that the corresponding proposition is true at time $t$ is
$|\langle\Psi(t)|\Pi|\Psi(t)\rangle|^{2}$, though we need to discuss what this
probability means. But if projection processes are removed from the time
evolution, the theory offers no connection between events at different times.
This might appear to suggest that the probability of the history $h$ of (2.1)
should be obtained by treating the $(\Pi_{j},t_{j})$ as independent events, so
that the probability of the history $h$ is
$P_{\text{E}}=\langle\Psi(t_{1})|\Pi_{1}|\Psi(t_{1})\rangle\langle\Psi(t_{2})|\Pi_{2}|\Psi(t_{2})\rangle\cdots\langle\Psi(t_{n})|\Pi_{n}|\Psi(t_{n})\rangle.$
(4.1)
But this makes little sense. It does not reflect the continuity of experience,
implying as it does that in a very short time interval, no matter how short,
there is a finite probability, independent of the length of the interval, of
the state changing to a very different one; and it gives a non-zero
probability for a history (decayed at $t_{1}$, undecayed at $t_{2}$) of an
unstable particle, where $t_{2}>t_{1}$. These defects are remedied [17] in the
proposal of John Bell which is explained in the next section. First, however,
we will look more closely at the meaning of probability in the Everett
interpretation.
Probability enters the formalism of quantum mechanics only in the Born rule
for the results of measurements and the projection postulate which is
associated with measurements. This is rejected in Everettian theory (and
indeed, according to Bell [5], any respectable physical theory) in which there
are no preferred physical processes which have the special status of
“measurement”. In Everettian quantum mechanics there is only deterministic
evolution of the universal state vector, so there is no absolute role for
probability. However, the universe includes sentient systems (or “observers”,
or “organisms”) and the universal state vector must incorporate the experience
of such systems, including the changing experience in time. Each possible
experience of a sentient system must be defined by a set of values of physical
quantities relating to the organism, so the experience corresponds to a set of
eigenvalues of these quantities and therefore to a subspace of the sentient
organism’s state space. The orthogonal projection onto this experience
subspace extends to a projection of the universal state space, corresponding
to the proposition “the organism has the given experience”. Now probability
applies to the change in time of the organism’s experience: while having
experience $E_{1}$ at time $t_{1}$, the organism wants to know the probability
that it will experience $E_{2}$ at a later time $t_{2}$, and then $E_{3}$ at
time $t_{3}$, and so on. Thus probability is not absolute but _relative_ to a
particular (appropriately constituted) subsystem, to a particular experience
of that subsystem, and to a particular time.
Note that “relative” does not mean “subjective”; this probability is an
objective fact about the organism’s situation and does not refer to its
beliefs, just as the velocity of a particle is relative to a frame of
reference but is an objective property of the particle.
## 5 The Bell probability
Let $\Pi_{j}$ be the projector onto the subspace $\mathcal{S}_{j}$ of the
universal state space describing experience $\eta_{j}$ of the sentient being,
or “observer”. We will call these subspaces _sensible_ subspaces because of
their relation to the sense organs of the observer; the same word will be
applied to the corresponding subspaces of the observer’s state space (a tensor
factor of the universal state space). The experience label $j$ is assumed
discrete for convenience. Adopting the observer’s perspective, we assume that
at each time $t$ the observer has a definite experience $\eta_{j}$ (this,
however, is not true in the external perspective in which the whole truth is
the universal state vector $|\Psi(t)\rangle$, a superposition of different
experience states). To describe the perceived change of this experience in
time we assume, following Bell [3], that it follows a stochastic process
governed by the universal Hamiltonian $H$ together with the universal state
vector $|\Psi(t)\rangle$, as follows. If, at time $t$, the observer’s
experience is $\eta_{j}$, then the probability that at time $t+\delta t$ their
experience has changed to $\eta_{k}$ is $w_{jk}\delta t$ where the transition
probability $w_{jk}$ is given by
$\text{\normalsize$w_{jk}$}=\begin{cases}\frac{2\text{Im}[\hbar^{-1}\langle\Psi(t)|\Pi_{k}H\Pi_{j}|\Psi(t)\rangle]}{\langle\Psi(t)|\Pi_{j}|\Psi(t)\rangle}&\text{\normalsize
if this is $\geq 0$},\\\ \text{\normalsize$0$}&\text{\normalsize if it is
negative.}\end{cases}$ (5.1)
The probabilities for the observer’s changing experience are then compatible
with the Born rule, in the following sense.
###### Theorem 1.
([3], [18] p.215)
If the experience state moves between the sensible subspaces with transition
probabilities given by (5.1), then the probability $P_{j}(t)$ that it lies in
the subspace $\mathcal{S}_{j}$ at time $t$ is given by the Born rule
$P_{j}(t)=\langle\Psi(t)|\Pi_{j}|\Psi(t)\rangle$
at all positive times $t$, if the probabilities are so given at the initial
time $t=0$.
This theorem is the justification for the transition probabilities (5.1).
Bell’s formula is not uniquely determined by the requirement that the theorem
should hold: there is a range of possible transition probabilities with the
same property [1]. However, Bell’s formula is uniquely natural in applications
to decay [17] and measurement processes [20, 13]: it ensures that the
underlying direction of change in such processes is always forwards, without
intermittent reversals (decay products, for example, recombining to
reconstitute the unstable decaying state). This means that the theory is not
invariant under time reversal. I regard this as a virtue; for those with
different predilections, the range of transition probabilities found in [1]
include one which is time-symmetric. However, it should be emphasised that in
the theory presented here, the probabilities arise only in the description of
the subjective experience of a sentient system; the existence of a preferred
direction of time for such a system has been explained within a time-symmetric
theory [7].
We use these transition probabilities to define the Bell probability
$P_{\text{B}}(h)$, relative to a particular observer, of a history of
experiences of that observer. More generally, Bell’s process will give
probabilities for any set of histories in which the projectors $\Pi_{j}$ are
chosen from a fixed complete set of projectors. There is no need for the
histories to satisfy the decoherence condition of the consistent-histories
formulation of quantum mechanics.
Note that our ontology is different from Bell’s. Bell intended the subspaces
$\mathcal{S}_{j}$ to have an absolute meaning, being the subspaces in which
certain preferred observables, called “beables”, had definite values, and that
it was absolutely true that the state of the world was in one of those
subspaces. In the theory adopted here, nothing has absolute existence or truth
except the universal state vector. The subspaces $\mathcal{S}_{j}$ are only
defined _if_ the universal state space admits a tensor product decomposition
in which some factors have the structure of sentient systems (whatever that
may be); the focus on a particular one of those systems is arbitrary. (This is
somewhat like some other recently expressed interpretations, such as those
called “convivial solipsism” [24] and “minimalism” [16].) In Bell’s exposition
of Bohmian theory these subspaces are called “beable” [3] or “viable” ([18] p.
215).
Theorem 1 usually gives the probabilities for the first stage of a history,
but often is not helpful in subsequent stages since the condition that the
initial probabilities are given by the Born rule is not satisfied. To
illustrate this, we now give the calculation of history probabilities in a
version of the “Wigner’s friend” scenario considered by Brukner [6].
In this scenario an experimenter (“Wigner”) observes a closed laboratory
containing another experimenter (“Wigner’s friend”) who makes measurements on
a quantum system $S$. We take $S$ to be a qubit (e.g. an electron spin) with
orthogonal states $|1\rangle$ and $|2\rangle$, and we simplify Wigner and his
friend so that each of them has just three orthogonal states: a ready state
$|0\rangle$ and two states registering the result of a measurement:
$|+\rangle$ and $|-\rangle$ for Wigner, $|1\rangle$ and $|2\rangle$ for his
friend (let us call her Frieda). The scenario is as follows.
At $t=0$ the qubit $S$ is in a general state $c_{1}|1\rangle+c_{2}|2\rangle$
and both Wigner and Frieda are in their ready states. The universal state is
therefore
$|\Psi(0)\rangle=\big{(}c_{1}|1\rangle+c_{2}|2\rangle\big{)}_{S}|0\rangle_{F}|0\rangle_{W}.$
Between $t=0$ and $t=1$ Frieda measures $S$ in the basis
$\\{|1\rangle_{S},|2\rangle_{S}\\}$, so that $S$ and $F$ become entangled and
the universal state at $t=1$ is
$\displaystyle|\Psi(1)\rangle$
$\displaystyle=\big{(}c_{1}|1\rangle_{S}|1\rangle_{F}+c_{2}|2\rangle_{S}|2\rangle_{F}\big{)}|0\rangle_{W}$
$\displaystyle=\big{(}c_{1}|1\rangle_{SF}+c_{2}|2\rangle_{SF}\big{)}|0\rangle_{W}\quad\text{in
an obvious notation.}$
Between $t=1$ and $t=2$ Wigner measures the state of Frieda and her laboratory
in a basis which includes the states
$|\pm\rangle_{SF}=\frac{1}{\sqrt{2}}\big{(}|1\rangle\pm|2\rangle\big{)}_{SF},$
so that the universal state at $t=2$ is
$|\Psi(2)\rangle=\frac{1}{\sqrt{2}}(c_{1}+c_{2})|+\rangle_{SF}|+\rangle_{W}+\frac{1}{\sqrt{2}}(c_{1}-c_{2})|-\rangle_{SF}|-\rangle_{W}.$
Thus the sensible state of $F$ is $|0\rangle$ at $t=0$ and either $|1\rangle$
or $|2\rangle$ at both $t=1$ and $t=2$. For $F$, there are four possible
histories
$(0jk)=\big{(}(\Pi_{0},t=0),(\Pi_{j},t=1),(\Pi_{k},t=2)\big{)}$
with $j,k\in\\{1,2\\}$. The probabilities of transition from $|0\rangle$ at
$t=0$ are given by Theorem 1 since the probabilities at $t=0$, namely
$P_{0}=1,P_{1}=P_{2}=0$, are indeed given by the Born rule; hence the
probability that Frieda finds herself in state $|j\rangle$ at $t=1$ is
$|c_{j}|^{2}$. To find the probability that she experiences a change of state
between $t=1$ and $t=2$, however, we cannot use this theorem, since her
probability for the state $|j\rangle$ at $t=1$ is either $0$ or $1$, which is
not the probability given by the Born rule: her experience of her state is at
odds with the universal state. To calculate the probability of a change of
state we need to know the Hamiltonian governing the change in the universal
state in this time interval, when Wigner is measuring Frieda and her
laboratory. Let us assume that the measurement process takes place by a simple
rotation, so that the state in the time interval $1<t<2$ is given by
$|\Psi(1+\tau)\rangle=\frac{1}{\sqrt{2}}(c_{1}+c_{2})|+\rangle_{SF}\big{(}\cos\tfrac{1}{2}\pi\tau|0\rangle_{W}+\sin\tfrac{1}{2}\pi\tau|+\rangle_{W}\big{)}\\\
+\frac{1}{\sqrt{2}}(c_{1}-c_{2})|-\rangle_{SF}\big{(}\cos\tfrac{1}{2}\pi\tau|0\rangle_{W}+\sin\tfrac{1}{2}\pi\tau|-\rangle_{W}\big{)}$
and the components in Frieda’s sensible subspaces are
$\displaystyle\Pi_{1}|\Psi(1+\tau)\rangle$
$\displaystyle=|1\rangle_{SF}\Big{(}c_{1}\cos\tfrac{1}{2}\pi\tau|0\rangle_{W}+\sin\tfrac{1}{2}\pi\tau\big{[}\tfrac{1}{2}(c_{1}+c_{2})|+\rangle+\tfrac{1}{2}(c_{1}-c_{2})|-\rangle\big{]}_{W}\Big{)},$
$\displaystyle\Pi_{2}|\Psi(1+\tau)\rangle$
$\displaystyle=|2\rangle_{SF}\Big{(}c_{2}\cos\tfrac{1}{2}\pi\tau|0\rangle_{W}+\sin\tfrac{1}{2}\pi\tau\big{[}\tfrac{1}{2}(c_{1}+c_{2})|+\rangle-\tfrac{1}{2}(c_{1}-c_{2})|-\rangle\big{]}_{W}\Big{)}.$
The Hamiltonian driving this evolution is, taking $\hbar=1$,
$H=\frac{\mathrm{i}\pi}{2}\sum_{z=\pm}|z\rangle\langle
z|_{SF}\otimes\big{[}|z\rangle\langle 0|-|0\rangle\langle z|\big{]}_{W}.$
(5.2)
The matrix element relevant to the transition
$|1\rangle_{F}\rightarrow|2\rangle_{F}$ is
$\langle\Psi(1+\tau)|\Pi_{2}H\Pi_{1}|\Psi(1+\tau)\rangle=\frac{\mathrm{i}\pi}{2}\big{(}|c_{1}|^{2}-|c_{2}|^{2}\big{)}\cos\tfrac{1}{2}\pi\tau\sin\tfrac{1}{2}\pi\tau.$
(5.3)
Suppose $|c_{1}|^{2}\geq|c_{2}|^{2}$; then the transition rate for
$|1\rangle_{F}\rightarrow|2\rangle_{F}$ at time $1+\tau$ is, according to
(5.1),
$\displaystyle w_{12}(1+\tau)$
$\displaystyle=\frac{2\text{Re}[\langle\Psi(1+\tau)|\Pi_{2}H\Pi_{1}|\Psi(1+\tau)\rangle]}{\langle\Psi(t)|\Pi_{1}|\Psi(t)\rangle}$
$\displaystyle=\frac{\pi}{2}\big{(}|c_{1}|^{2}-|c_{2}|^{2}\big{)}\frac{\cos\tfrac{1}{2}\pi\tau\sin\tfrac{1}{2}\pi\tau}{|c_{1}|^{2}+\big{(}|c_{1}|^{2}-|c_{2}|^{2}\big{)}\sin^{2}\tfrac{1}{2}\pi\tau},$
while the rate for the opposite transition is $w_{21}=0$. Hence the
probabilities $P_{1}(1+\tau),P_{2}(1+\tau)$ for Frieda’s experience at time
$1+\tau$ satisfy
$\frac{\text{d}}{\text{d}\tau}P_{1}=-w_{12}P_{1},\quad\frac{\text{d}}{\text{d}\tau}P_{2}=w_{12}P_{1},$
the solution of which is
$P_{1}(1+\tau)=C\Big{[}|c_{1}|^{2}-\tfrac{1}{2}\big{(}|c_{1}|^{2}-|c_{2}|^{2}\big{)}\sin^{2}\tfrac{1}{2}\pi\tau\big{]},\quad
P_{2}=1-P_{1}$
where $C$ is a constant of integration. By taking $C=1/|c_{1}|^{2}$ or $C=0$,
corresponding to initial conditions $P_{1}(1)=1$ and $P_{1}(1)=0$
respectively, we get the conditional probabilities for Frieda’s experience:
$\displaystyle P\big{(}|1\rangle_{F}\text{ at
}t=2\;\big{|}\;|1\rangle_{F}\text{ at }t=1)$
$\displaystyle=\frac{1}{2|c_{1}|^{2}},$ (5.4) $\displaystyle
P\big{(}|1\rangle_{F}\text{ at }t=2\;\big{|}\;|2\rangle_{F}\text{ at }t=1)$
$\displaystyle=0,$ (5.5) $\displaystyle P\big{(}|2\rangle_{F}\text{ at
}t=2\;\big{|}\;|1\rangle_{F}\text{ at }t=1)$
$\displaystyle=\frac{|c_{1}|^{2}-|c_{2}|^{2}}{2|c_{1}|^{2}},$ (5.6)
$\displaystyle P\big{(}|2\rangle_{F}\text{ at
}t=2\;\big{|}\;|2\rangle_{F}\text{ at }t=1)$ $\displaystyle=1.$ (5.7)
These, together with the probabilities $|c_{1}|^{2}$ and $|c_{2}|^{2}$ for the
results of Frieda’s measurement at $t=1$, give the non-zero Bell probabilities
for her histories as set out in Table 1, which also shows the corresponding
Copenhagen probabilities.
$h$ | $P_{B}(h)$ | $P_{C}(h)$
---|---|---
(011) | $\tfrac{1}{2}$ | $\tfrac{1}{2}|c_{1}|^{2}$
(012) | $\tfrac{1}{2}\big{(}|c_{1}|^{2}-|c_{2}|^{2})$ | $\tfrac{1}{2}|c_{1}|^{2}$
(021) | $0$ | $\tfrac{1}{2}|c_{2}|^{2}$
(022) | $|c_{2}|^{2}$ | $\tfrac{1}{2}|c_{2}|^{2}$
Table 1: Probabilities for the histories of Wigner’s friend in Brukner’s
scenario
Bell and Copenhagen agree that the probability that Frieda is in state
$|1\rangle$ at $t=1$ is $|c_{1}|^{2}$, while the probability that she is in
$|2\rangle$ is $|c_{2}|^{2}$, so she is more likely to be in the state
$|1\rangle$, while at $t=2$ she is equally likely to be in either state.
According to Bell, this change is caused entirely by a transition from
$|1\rangle$ to $|2\rangle$, so she will only feel the effect of Wigner’s
measurement if she is in the more probable state $|1\rangle$; but according to
Copenhagen there is a flow of probability in both directions.
It is illuminating to examine the two sets of probabilities in the special
case $c_{1}=c_{2}=1/\sqrt{2}$. for which Bell and Copenhagen agree that the
two states $|1\rangle_{F}$ and $|2\rangle_{F}$ each have probability
$\tfrac{1}{2}$ at both $t=1$ and $t=2$. According to Bell there are no
transitions between these two states in the time period $1\leq t\leq 2$,
whereas Copenhagen gives equiprobable transitions in both directions. For
these values of the coefficients Frieda and her laboratory are already in the
state $|+\rangle_{SF}$ at $t=1$, so Wigner’s measurement simply registers this
fact, changing Wigner’s state from $|0\rangle_{W}$ to $|+\rangle_{W}$ but
leaving Frieda’s state unchanged. This is reflected in the fact that the
Hamiltonian (5.2) has zero matrix elements (5.3) between the two components
$\Pi_{i}|\Psi(1+\tau)\rangle$ in which Frieda has a definite value $i=1$ or
$2$. The Copenhagen calculation of probabilities, however, assumes that
Frieda’s state collapses at $t=1$, despite the lack of physical mechanism for
this collapse; if this happens then Wigner’s measurement does have an effect
on the resulting states, leading to transitions between them.
In the general case the Bell calculation gives the minimal transitions
necessary for the change in probabilities from $t=1$ to $t=2$.
## 6 The Frauchiger-Renner universe
A more elaborate variant of the Wigner’s friend thought experiment has been
devised by Frauchiger and Renner [10]. It is instructive to compare the
different probabilities for this scenario. This takes place in a universe
consisting of six subsystems $C,S,F_{1},F_{2},W_{1}$ and $W_{2}$. Of these $C$
is a quantum coin with two orthogonal states $|\text{head}\rangle_{C}$ and
$|\text{tail}\rangle_{C}$; $S$ is a qubit with orthogonal states
$|\uparrow\rangle_{S}$ and $|\downarrow\rangle_{S}$; and the four systems
$F_{1},F_{2},W_{1},W_{2}$ are sentient agents, each agent $X$ having three
orthogonal states $|0\rangle_{X},|a\rangle_{X},|b\rangle_{X}$ where
$|0\rangle$ is a state of readiness before $X$ makes a measurement and
$|a\rangle$ and $|b\rangle$ are states recording the results $a,b$ of the
measurement. We refer to the agents $F_{1}$ and $F_{2}$ as “experimenters”,
since they perform operations on the physical systems $C$ and $S$, and to
$W_{1}$ and $W_{2}$ as “para-experimenters” since they operate on the
experimenters $F_{1}$ and $F_{2}$ as well as $C$ and $S$. The experiment
proceeds as follows.
Initially, at time $t=0$, the system is in the state
$|\Psi(0)\rangle=\left(\sqrt{\tfrac{1}{3}}|\text{head}\rangle+\sqrt{\tfrac{2}{3}}|\text{tail}\rangle\right)_{C}|0\rangle_{S}|0\rangle_{F_{1}}|0\rangle_{F_{2}}|0\rangle_{W_{1}}|0\rangle_{W_{2}}.$
(6.1)
Between times $t=0$ and $t=1$ the experimenter $F_{1}$ measures the state of
the coin, so that at time $t=1$ the universal state is
$|\Psi(1)\rangle=\left(\sqrt{\tfrac{1}{3}}|\text{head}\rangle+\sqrt{\tfrac{2}{3}}|\text{tail}\rangle\right)_{F_{1}C}|0\rangle_{S}|0\rangle_{F_{2}}|0\rangle_{W_{1}}|0\rangle_{W_{2}},$
(6.2)
where
$|\text{head}\rangle_{F_{1}C}=|\text{head}\rangle_{F_{1}}|\text{head}\rangle_{C}$
and similarly for “tail”.
Between $t=1$ and $t=2$ the experimenter $F_{1}$ prepares the qubit $S$ in a
way which depends on the result of their measurement of $C$: if the coin came
down showing “head”, $F_{1}$ prepares $|\downarrow\rangle_{S}$; if it showed
“tail”, $F_{1}$ prepares
$|\rightarrow\rangle_{S}=\tfrac{1}{\sqrt{2}}\big{(}|\uparrow\rangle+|\downarrow\rangle\big{)}_{S}$.
The universal state at $t=2$ is therefore
$|\Psi(2)\rangle=\left(\sqrt{\tfrac{1}{3}}|\text{head}\rangle_{F_{1}C}|\downarrow\rangle_{S}+\sqrt{\tfrac{2}{3}}|\text{tail}\rangle_{F_{1}C}|\rightarrow\rangle_{S}\right)|0\rangle_{F_{2}}|0\rangle_{W_{1}}|0\rangle_{W_{2}}.$
(6.3)
Between $t=2$ and $t=3$ the qubit $S$ is passed to the second experimenter
$F_{2}$, who measures it in the basis
$\\{|\uparrow\rangle,|\downarrow\rangle\\}$, recording the result $+,-$
respectively; we will use the same symbol $\pm$ to denote the joint state of
$F_{2}$ and $S$. The universal state becomes
$|\Psi(3)\rangle=\sqrt{\tfrac{1}{3}}\Big{(}|\text{head}\rangle_{F_{1}C}|-\rangle_{F_{2}S}+|\text{tail}\rangle_{F_{1}C}|+\rangle_{F_{2}S}+|\text{tail}\rangle_{F_{1}C}|-\rangle_{F_{2}S}\Big{)}|0\rangle_{W_{1}}|0\rangle_{W_{2}}.$
(6.4)
Between $t=3$ and $t=4$ the first para-experimenter $W_{1}$ measures the state
of $F_{1}$ and the coin $C$ in the basis
$\displaystyle|\text{ok}\rangle_{F_{1}C}$
$\displaystyle=\tfrac{1}{\sqrt{2}}\big{(}|\text{head}\rangle_{F_{1}C}-|\text{tail}\rangle_{F_{1}C}\big{)},$
$\displaystyle|\text{fail}\rangle_{F_{1}C}$
$\displaystyle=\tfrac{1}{\sqrt{2}}\big{(}|\text{head}\rangle_{F_{1}C}+|\text{tail}\rangle_{F_{1}}C\big{)},$
recording the result in their own state $|\text{ok}\rangle_{W_{1}}$ or
$|\text{fail}\rangle_{W_{1}}$. The universal state becomes, at $t=4$,
$|\Psi(4)\rangle=\Big{(}\sqrt{\tfrac{1}{6}}\big{[}-|\text{ok}\rangle_{F_{1}C}|\text{ok}\rangle_{W_{1}}+|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}\big{]}|+\rangle_{F_{2}S}\\\
+\sqrt{\tfrac{2}{3}}|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}|-\rangle_{F_{2}S}\Big{)}|0\rangle_{W_{2}}.$
Between $t=4$ and $t=5$ the second para-experimenter $W_{2}$ measures the
state of $F_{2}$ and the qubit $S$ in the basis
$\displaystyle|\text{ok}\rangle_{F_{2}S}$
$\displaystyle=\tfrac{1}{\sqrt{2}}\big{(}|-\rangle_{F_{2}S}-|+\rangle_{F_{2}S}\big{)},$
$\displaystyle|\text{fail}\rangle_{F_{2}S}$
$\displaystyle=\tfrac{1}{\sqrt{2}}\big{(}|-\rangle_{F_{2}S}+|+\rangle_{F_{2}S}\big{)},$
recording the result in their own state $|\text{ok}\rangle_{W_{2}}$ or
$|\text{fail}\rangle_{W_{2}}$. The universal state becomes, at $t=5$,
$|\Psi(5)\rangle=\sqrt{\tfrac{1}{12}}\Big{(}|\text{ok}\rangle_{F_{1}C}|\text{ok}\rangle_{W_{1}}+|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}\Big{)}|\text{ok}\rangle_{F_{2}S}|\text{ok}\rangle_{W_{2}}\\\
+\sqrt{\tfrac{1}{12}}\Big{(}-|\text{ok}\rangle_{F_{1}C}|\text{ok}\rangle_{W_{1}}+3|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}\Big{)}|\text{fail}\rangle_{F_{2}S}|\text{fail}\rangle_{W_{2}}.$
Now consider the history of this experiment as seen by the experimenter
$F_{2}$ from $t=1$ to $t=5$. The possible histories are
$(abcde)=h(a,b,c,d,e)=\big{(}(\Pi_{a},t=1),\ldots,(\Pi_{e},t=5)\big{)}$
where each of $a,b,c,d,e$ is taken from the set $\\{0,+,-\\}$, and $\Pi_{x}$
is the projector onto the 32-dimensional subspace
$\mathcal{S}_{x}=\mathcal{H}_{C}\otimes\mathcal{H}_{S}\otimes\mathcal{H}_{F_{1}}\otimes\text{span}\left(|x\rangle_{F_{2}}\right)\otimes\mathcal{H}_{W_{1}}\otimes\mathcal{H}_{W_{2}}.$
The Copenhagen probabilities $P_{\text{C}}(abcde)$ can be calculated in the
same way as in Section 3. The result is that the only non-zero probabilities
are
$P_{\text{C}}(00+++)=\tfrac{1}{6},\quad P_{\text{C}}(00--+)=\tfrac{1}{3},\quad
P_{\text{C}}(00++-)=\tfrac{1}{6},\quad P_{\text{C}}(00---)=\tfrac{1}{3}.$
To calculate the Bell probabilities, we need to know the Hamiltonian governing
the evolution of the system, i.e. the succession of measurements made by the
four agents. In each time interval $[n,n+1]$ ($n=1,\ldots,4$) the system
evolves from a product state to an entangled state of the experimenter and the
measured system. The Bell probabilities can be calculated, as in Section 5, by
assuming that the Hamiltonian in this interval is a generator of rotations in
the subspace spanned by these two states, operating for a time such that
$\mathrm{e}^{-\mathrm{i}H}$ is a rotation through $\pi/2$. The resulting non-
zero probabilities are
$P_{\text{B}}(00+++)=\tfrac{1}{3},\quad P_{\text{B}}(00--+)=\tfrac{1}{6},\quad
P_{\text{C}}(00---)=\tfrac{1}{2}.$
These Bell probabilities are set out beside the Copenhagen probabilities in
Table 2.
$abcde$ | $P_{\text{B}}(abcde)$ | $P_{\text{C}}(abcde)$
---|---|---
($00+++$) | $\tfrac{1}{3}$ | $\tfrac{1}{6}$
($00++-$) | $0$ | $\tfrac{1}{6}$
($00--+$) | $\tfrac{1}{6}$ | $\tfrac{1}{3}$
($00---$) | $\tfrac{1}{2}$ | $\tfrac{1}{3}$
Table 2: Probabilities for the histories of $F_{2}$ in the Frauchiger-Renner
experiment
The difference between the Bell and Copenhagen probabilities in Table 2 is
similar to what we have already seen in Table 1 for Brukner’s version of the
Wigner’s friend experiment.
To be relevant to the conclusion that Frauchiger and Renner drew from their
thought experiment, we must extend our concept of a sentient observer. The
Frauchiger-Renner argument assumes communication between $F_{1}$, $W_{1}$ and
$W_{2}$ and therefore requires that they should be treated as a single
observer whose experience covers the measurement results of all three. The
histories of this three-headed observer from $t=0$ to $t=5$ are of the form
$h(a\,a^{\prime}\,a^{\prime\prime},\ldots,f\,f^{\prime}\,f^{\prime\prime})=\big{(}(a,a^{\prime},a^{\prime\prime};t=0),\ldots,(f,f^{\prime},f^{\prime\prime};t=5)\big{)}$
where
$a,\ldots,f\in\\{0,\text{head},\text{tail}\\},\,a^{\prime}\ldots,f^{\prime}\in\\{0,\text{ok},\text{fail}\\}$
and
$a^{\prime\prime},\ldots,f^{\prime\prime}\in\\{0,\text{ok},\text{fail}\\}$.
However, the Bell and Copenhagen probabilities are both zero for any history
in which there is any change in the experience of $F_{1}$ between $t=1$ and
$t=3$; the relevant projection operators commute with the evolution operators
$\mathrm{e}^{-\mathrm{i}H(t-1)}$ for $1<t<3$ since all the action in this
interval involves only $S$ and $F_{2}$. We therefore only need to consider
histories
$h(a\,a^{\prime}\,a^{\prime\prime},\;b\,b^{\prime}\,b^{\prime\prime},\;c\,c^{\prime}\,c^{\prime\prime},\;d\,d^{\prime}\,d^{\prime\prime})\\\
=\big{(}(a,a^{\prime},a^{\prime\prime};t=0),(b,b^{\prime},b^{\prime\prime};t=1),(c,c^{\prime},c^{\prime\prime};t=4),(d,d^{\prime},d^{\prime\prime};t=5)\big{)}$
The two sets of probabilities for these histories are shown in Table 3, in
which “head”, “tail” and “fail” are abbreviated to their initial letters. All
histories not shown in this table have $P_{\text{B}}=P_{\text{C}}=0$.
$h$ | $P_{\text{B}}(h)$ | $P_{\text{C}}(h)$
---|---|---
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{h}\,\text{f}\,0,\;\text{h}\,\text{f}\,\text{f})$ | $\tfrac{3}{20}$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{h}\,\text{f}\,0,\;\text{h}\,\text{f}\,\,\text{ok})$ | $\tfrac{1}{60}$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{h}\,\text{ok}\,0,\;\text{h}\,\text{f}\,\text{ok})$ | $0$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{h}\,\text{ok}\,0,\;\text{h}\,\text{ok}\,\text{ok})$ | $0$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{t}\,\text{f}\,0,\;\text{t}\,\text{f}\,0$ | $\tfrac{3}{20}$ | $\tfrac{1}{24}$
$(0\,0\,0.\;\text{h}\,0\,0,\;\text{t}\,\text{f}\,0,\;\text{t}\,\text{f}\,\text{ok})$ | $\tfrac{1}{60}$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{t}\,\text{ok}\,0,\;\text{t}\,\text{ok}\,\text{f})$ | $0$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{h}\,0\,0,\;\text{t}\,\text{ok}\,0,\;\text{t}\,\text{ok}\,\text{ok})$ | $0$ | $\tfrac{1}{24}$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{h}\,\text{f}\,0,\;\text{h}\,\text{f}\,\text{f})$ | $\tfrac{9}{40}$ | $\tfrac{1}{6}$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{h}\,\text{f}\,0,\;\text{h}\,\text{f}\,\,\text{ok})$ | $\tfrac{1}{40}$ | $0$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{h}\,\text{ok}\,0,\;\text{h}\,\text{ok}\,\text{f})$ | $\tfrac{1}{24}$ | $\tfrac{1}{6}$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{h}\,\text{ok}\,0,\;\text{h}\,\text{ok}\,\text{ok})$ | $\tfrac{1}{24}$ | $0$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{t}\,\text{f}\,0,\;\text{t}\,\text{f}\,\text{f})$ | $\tfrac{9}{40}$ | $\tfrac{1}{6}$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{t}\,\text{f}\,0,\;\text{t}\,\text{f}\,\,\text{ok})$ | $\tfrac{1}{40}$ | $0$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{t}\,\text{ok}\,0,\;\text{t}\,\text{ok}\,\text{f})$ | $\tfrac{1}{24}$ | $\tfrac{1}{6}$
$(0\,0\,0,\;\text{t}\,0\,0,\;\text{t}\,\text{ok}\,0,\;\text{t}\,\text{ok}\,\text{ok})$ | $\tfrac{1}{24}$ | $0$
Table 3: Probabilities for the histories of $F_{1}$, $W_{1}$ and $W_{2}$ in
the Frauchiger-Renner experiment
As before, there are pairs of histories which are equally probable in the
Copenhagen calculation, while Bell assigns zero probability to one of them.
But now there are also histories which have $P_{\text{C}}=0$ but
$P_{\text{B}}\neq 0$, in particular those in which $F_{1}$ observes “tails”
when she measures the coin at $t=2$ and $W_{1}$ and $W_{2}$ both obtain the
result “ok” in their measurements of $F_{1}$ and $F_{2}$. This difference is
the core of the Frauchiger-Renner argument, and the substance of their claim
of a problem in the description of this thought experiment by standard quantum
mechanics. The zero probability in the Copenhagen calculation corresponds to
$F_{1}$ applying the projection postulate to the universal state after seeing
the result of her measurement. The non-zero probability in the Bell
calculation arises if $F_{1}$ applies Vedral’s dictum “Unobserved results can
affect future measurements” [22]. Knowing how the experiment is proceeding,
she knows that what she observes is registered in just one component of the
universal state. The whole of the universal state continues to evolve and
affects the future measurements, as is explicitly included in Bell’s process:
the transition probabilities at any time depend on the value of the universal
state at that time. The agents $F_{1}$, $W_{1}$ and $W_{2}$ are entitled to
regard their experience as “reality”; then in their perspective the universal
state vector is not a description of reality but an influence, or force, that
affects the development of reality.
We note that the history states $|K(h)\rangle$, as defined in (3.3), for the
histories $h$ in table 3 are not all orthogonal; in fact there are four cases
in which two different histories give rise to the same history state, for
example
$\displaystyle|K(0\,0\,0,\;\text{h}\,0\,0,\;\text{h}\,\text{f}\,0,\;\text{h}\,\text{f}\,\text{f})\rangle$
$\displaystyle=|K(0\,0\,0,\;\text{t}\,0\,0,\;\text{t}\,\text{f}\,0,\;\text{t}\,\text{f}\,\text{f})\rangle$
$\displaystyle=\frac{1}{\sqrt{6}}|\text{h}\rangle_{F_{1}C}|\text{f}\rangle_{F_{2}S}|\text{f}\rangle_{W_{1}}|\text{f}\rangle_{W_{2}}.$
It follows, as has been pointed out before [14], that the histories in Table 3
do not satisfy the consistency condition of [12]. This does not mean that
there is anything inconsistent in these histories; it only means that the
formula (3.1) cannot consistently be applied to disjunctions of them. This can
be seen, for example, by using (3.1) to calculate the Copenhagen probability
of a three-time history with times $t=0$, $t=4$ and $t=5$; the result is not
the sum of the corresponding histories in Table 3 — as is to be expected in
standard quantum mechanics when one ignores the fact that a measurement has
taken place at $t=1$. But if the histories are considered as atoms in a
probability space111Confusingly, this would normally be called an “event
space.”, the usual (Kolmogorov) rules can be applied.
## 7 Observation of History
If an agent is equipped with a memory which records their momentary experience
at each stage of a history, they can reduce their experience of history to a
single measurement at the end of the history. Let us return to $F_{2}$ and
suppose she has a memory with a basis $|x_{1}x_{2}x_{3}x_{4}x_{5}\rangle_{M}$
where $x_{t}$ can take values $0,+,-$ or $\times$, the final value $\times$
being recorded at times when $t$ lies in the future. Then the universal state
space has an extra tensor factor $\mathcal{H}_{M}$. We denote the universal
state vector in this enlarged space by $|\widetilde{\Psi}(t)\rangle$, with
$|\Psi(t)\rangle$ continuing to denote the universal state in Section 6. Then
the state with memory is of the form
$|\widetilde{\Psi}(t)\rangle=\sum_{a\in\\{0,+,-\\}}|a\rangle_{F_{2}}|x_{1}x_{2}x_{3}x_{4}x_{5}\rangle_{M}|\psi(t)\rangle\\\
\text{with
}|\psi(t)\rangle\in\mathcal{H}_{S}\otimes\mathcal{H}_{C}\otimes\mathcal{H}_{F_{1}}\otimes\mathcal{H}_{W_{1}}\otimes\mathcal{H}_{W_{2}}$
where $x_{t}=a$ and $x_{s}=\times$ if $s>t$. This develops as follows:
$\displaystyle|\widetilde{\Psi}(0)\rangle$
$\displaystyle=|\Psi(0)\rangle|0\times\times\times\times\rangle_{M},$
$\displaystyle|\widetilde{\Psi}(1)\rangle$
$\displaystyle=|\Psi(1)\rangle|00\times\times\times\rangle_{M},$
$\displaystyle|\widetilde{\Psi}(2)\rangle$
$\displaystyle=\Big{[}\sqrt{\tfrac{1}{6}}\big{(}|\text{fail}\rangle-|\text{ok}\rangle\big{)}_{F_{1}C}|+\rangle_{F_{2}S}|00+\times\times\rangle_{M}$
$\displaystyle\quad+\sqrt{\tfrac{2}{3}}|\text{fail}\rangle_{F_{1}C}|-\rangle_{F_{2}S}|00-\times\times\rangle_{M}\Big{]}|0\rangle_{W_{1}}|0\rangle_{W_{2}},$
$\displaystyle|\widetilde{\Psi}(3)\rangle$
$\displaystyle=\sqrt{\tfrac{1}{6}}|00++\times\rangle\Big{(}|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}-|\text{ok}\rangle_{F_{1}C}|\text{ok}\rangle_{W_{1}}\Big{)}|+\rangle_{F_{2}}|0\rangle_{W_{2}}$
$\displaystyle\quad+\sqrt{\tfrac{2}{3}}|00--\times\rangle|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}|-\rangle_{F_{2}}|0\rangle_{W_{2}}$
$\displaystyle|\widetilde{\Psi}(4)\rangle$
$\displaystyle=\tfrac{1}{\sqrt{24}}|00+++\rangle_{M}\left(|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}-|\text{ok}\rangle_{F_{1}C}|\text{ok}\rangle_{W_{1}}\right)|+\rangle_{F_{2}S}\big{(}|\text{fail}\rangle+|\text{ok}\rangle\big{)}_{W_{2}}$
$\displaystyle+\tfrac{1}{\sqrt{24}}|00++-\rangle_{M}\left(|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}-|\text{ok}\rangle_{F_{1}C}|\text{ok}\rangle_{W_{1}}\right)|+\rangle_{F_{2}S}\big{(}|\text{fail}\rangle-|\text{ok}\rangle\big{)}_{W_{2}}$
$\displaystyle+\tfrac{1}{\sqrt{6}}|00--+\rangle_{M}|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}|+\rangle_{F_{2}S}\big{(}|\text{fail}\rangle-|\text{ok}\rangle\big{)}_{W_{2}}$
$\displaystyle+\tfrac{1}{\sqrt{6}}|00---\rangle_{M}|\text{fail}\rangle_{F_{1}C}|\text{fail}\rangle_{W_{1}}|-\rangle_{F_{2}S}\big{(}|\text{fail}\rangle+|\text{ok}\rangle\big{)}_{W_{2}}.$
From this we can read off the probabilities of the memory readings as
$P(00+++)=P(00++-)=\tfrac{1}{6},\quad P(00--+)=P(00---)=\tfrac{1}{3},$
the same as the Copenhagen probabilities in Table 1. The reason for this is
clear: the entanglement of the original system with the memory serves to
separate the sensible states of $F_{2}$, so that the time development of each
such state proceeds as if the other states had been annihilated by a
projection operator. This prevents subsequent interference between these
components: entanglement implements collapse.
This calculation illustrates a general fact: if a system is linked to a memory
which keeps a permanent record of a set $B$ of basis states of the system,
then the probabilities which will be observed in the memory are the same as
those which would be calculated in a Copenhagen calculation assuming that the
system (without the memory) undergoes collapse at each time stage onto a state
of the basis $B$. Thus a direct attempt to establish the probabilities of
various histories of the system will lead to the Copenhagen probabilities.
This might be seen as a justification of the probabilities assumed in the
“consistent histories” version of quantum mechanics. However, this does not
rule out the possibility of indirect effects of Bell probabilities if the
system is not coupled to a memory, in a similar way to the quantum effects of
superposition which are not visible if the system undergoes measurement.
Moreover, this justification of Copenhagen probabilities is only valid if the
system whose histories are in question is not the whole universe; it depends
on there being something (the memory) external to the system. This undercuts
the claim of consistent-histories theory to be a version of quantum mechanics
which is specially adapted to cosmology.
## 8 Conclusion
We have explored an alternative theory of histories — sequences of events,
such as measurements — in quantum mechanics. These have usually been treated
by assuming the Dirac/von Neumann/Lüders projection postulate of the
Copenhagen formulation of quantum mechanics. We have shown that this is not
necessary; there is an alternative formulation in which this postulate is not
assumed, but histories can still be treated. We have shown that this gives
different results for the probability of a history. However, this difference
is not directly observable: a simple empirical procedure for observing
histories is bound to give the same results as if the projection postulate was
valid. Nevertheless, this means that any theory that assumes this postulate,
such as the “consistent histories” interpretation of quantum mechanics, cannot
be a theory of the whole universe and cannot be a basis for cosmology.
Acknowledgement
I am grateful to Flavio del Santo and his co-workers for a valuable
discussion, and to an anonymous referee for suggestions which have greatly
improved the presentation of these ideas.
## References
* [1] G. Bacciagaluppi and M. Dickson. Dynamics for density operator interpretations of quantum theory. Found. Phys. 29:1165, 1999. arXiv:quant-ph/9711048
* [2] Veronika Baumann, Flavio Del Santo, Alexander R. H. Smith, Flaminia Giacomini, Esteban Castro-Ruiz, and Caslav Brukner. Generalized probability rules from a timeless formulation of Wigner’s friend scenarios. 2019\. arXiv:1911.09696.
* [3] J. S. Bell. Beables for quantum field theory. In B. J. Hiley and F. D. Peat, editors, Quantum Implications, pages 227–234. Routledge and Kegan Paul, London, 1984. Reprinted in [4].
* [4] J. S. Bell. Speakable and unspeakable in quantum mechanics. Cambridge University Press, 1987.
* [5] J. S. Bell. Against “measurement”. Physics World, 3(8):33–40, August 1990.
* [6] C. Brukner. On the quantum measurement problem. In Quantum [Un]Speakables II: 50 Years of Bell’s Theorem, pages 95–117. Springer, 2017.
* [7] Andrea Di Biagio, Pietro Donà and Carlo Rovelli. The arrow of time in operational formulations of quantum theory. 2021\. arXiv:2010.05734.
* [8] Carl E. Dolby. The conditional probability interpretation of the Hamiltonian constraint. 2004\. arXiv:gr-qc.0406034.
* [9] Hugh Everett III. “Relative state” formulation of quantum mechanics. Rev. Mod. Phys., 29:141–153, 1957.
* [10] Daniela Frauchiger and Renato Renner. Quantum theory cannot consistently describe the use of itself. Nature Communications, 9:3711, 2018.
* [11] Vittorio Giovannetti, Seth Lloyd, and Lorenzo Maccone. Quantum time. Phys. Rev. D, 92:045033, 2015. arXiv:1504.04215.
* [12] R. B. Griffiths. Consistent Quantum Theory. Cambridge University Press, 2002.
* [13] Timothy J. Hollowood. Classical from quantum. 2018\. arXiv:1803.04700
* [14] Marcelo Losada, Roberto Laura, and Olimpia Lombardi. Frauchiger-Renner argument and quantum histories. Physical Review A, 100:052114–052119, 2019. arXiv:1907.10095.
* [15] Carlo Rovelli. Relational quantum mechanics. Int. J. Theor. Phys., 35:1637–1677, 1996.
* [16] D. Sokolovski. A minimalist’s view of quantum mechanics. Euro. Phys. Lett., 128:50001, 2019. arXiv:2005.12674.
* [17] A. Sudbery. Diese verdammte Quantenspringerei. Stud. Hist. Phil. Mod. Phys., 33B:387–411, 2002. arXiv:quant-ph/0011082.
* [18] Anthony Sudbery. Quantum Mechanics and the Particles of Nature. Cambridge University Press, 1986.
* [19] Anthony Sudbery. The logic of the future in quantum theory. Synthese, 194:4429–4453, 2017. arXiv:1409.0755.
* [20] Anthony Sudbery. Single-world theory of the extended Wigner’s friend experiment. Found. Phys., 47:658-669, 2017. arXiv:1608.05873.
* [21] Anthony Sudbery. John Bell and the great enterprise. Quanta, 7:68–73, 2018. arXiv:1808.06845.
* [22] Vlatko Vedral. Local quantum reality. 2020\. arXiv:2011.01039.
* [23] J. A. Wheeler. Assessment of Everett’s “relative state” formulation of quantum theory. Rev. Mod. Phys., 29:463–465, 1957.
* [24] Hervé Zwirn. Non locality versus modified realism. Found. Phys., 50:1–26, 2020. arXiv:1812.06451.
|
${\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>to party $\mathsf{P}_{i}$.
* •
Receive the teleportation measurement result $(z_{i},x_{i})$ from party
$\mathsf{P}_{i}$.
* •
Extract the input
$\rho^{\prime}_{i}\leftarrow\mathsf{TP.Receive}(z_{i},x_{i},\tilde{e}_{R,i})$
and send it to $\mathsf{T}$.
3. 3.
Output recovery:
* •
Upon receiving $\sigma_{j}$ from $\mathsf{T}$, teleport $\sigma_{j}$ into the
first component $\hat{e}_{S,j}^{(0)}$.
* •
Obtain
$(z^{\prime}_{j},x^{\prime}_{j})\leftarrow\mathsf{TP.Send}(\sigma_{j},\hat{e}_{S,j}^{(0)})$.
* •
Send
$P^{\prime\prime}_{j}=\hat{P}_{j}(X^{x^{\prime}_{j}}Z^{z^{\prime}_{j}}\otimes\mathbbm{1}^{(\ell-\ell^{\prime})(1+t)})$
to $\mathsf{P}_{j}$.
4. 4.
For each measurement step of $C$, do the following (with $l$ from $\ell$ to
$\ell^{\prime}+1$):
* •
Send to $\mathsf{P}_{j}$ the gate $\hat{E}_{j}^{(l)}$.
* •
Receive the measurement result $r_{j}\in\\{0,1\\}^{1+t}$ from party
$\mathsf{P}_{j}$.
* •
Measure $\hat{e}_{S,j}^{(l)}$ and compare the result
$\tilde{r}_{j}=\mathsf{\Pi}(\hat{e}_{S,j}^{(l)})$ with $r_{j}$.
* •
If $\tilde{r}_{j}\neq r_{j}$, let $\mathsf{P}_{j}$ send abort to $\mathsf{T}$.
Output: the final output of ${\mathcal{A}}$.
We prove by induction on the number of circuit steps $|C|$ that the real
protocol can be simulated by the simulator $\mathsf{Sim}^{\mathsf{RQC}}$. We
also prove in the induction that the end state of both the real and ideal
world when $\mathsf{P}_{j}$ is not being marked as corrupted are
indistinguishable from a state where ${\mathsf{cMPC}}$ holds random Clifford
keys $\\{E_{j}\\}_{j\in[n]}$ and the adversary holds
$\displaystyle A^{\Delta}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{
};\text{
}{(\underset{j\in[n]}{\otimes}E_{j})({\mathsf{QECC.Enc}}(C(\cdot,\phi_{h})),0^{lnt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
(9)
for some map $A$. For the base case $C=\mathbbm{1}$, the protocol is equal to
$(\Sigma^{\mathsf{IE}},\Pi^{\mathsf{IE}})$. From the analysis of the previous
section, it remains to show that $\mathsf{Sim}^{\mathsf{RQC}}(\mathbbm{1})$
ends in state (7). Using notation (5), the state at the end of simulator step
$2$ is
$\underset{\Delta}{\Sigma}\text{
}\big{(}{\ket{z_{i}}\\!\bra{z_{i}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}I_{i}}},{\ket{x_{i}}\\!\bra{x_{i}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}S_{i}}},{\underset{i\in[n]}{\cup}\rho_{i}^{\Delta}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R^{\prime}_{i}}},{U_{P_{I_{i}},P_{S_{i}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j},W}}({\hat{P}_{j}\cdot\hat{e}_{R,j}^{(0)}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}){U_{P^{\prime}_{I_{i}},P_{S^{\prime}_{i}}}^{\dagger}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j},W}}\big{)}$
After step $3$, the values in registers
${\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>is correctly QECC-encoded, securely authenticated, and being teleported. The
end state of the simulation becomes
$\displaystyle A^{\Delta}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{
};\text{
}{(\underset{j\in[n]}{\otimes}\hat{P}_{j}X^{x^{\prime}_{j}}Z^{z^{\prime}_{j}}E_{j})({\mathsf{QECC.Enc}}(\cdot,\phi_{h}),0^{lnt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
$\displaystyle=$ $\displaystyle
A^{\Delta}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{ };\text{
}{(\underset{j\in[n]}{\otimes}P^{\prime\prime}_{j}E_{j})({\mathsf{QECC.Enc}}(\cdot,\phi_{h}),0^{lnt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
which is equal to (7). (When $C=\mathbbm{1}$, the simulator does nothing in
step $4$.)
Suppose the parties are going to run a further computation step under our
induction hypothesis. For the protocol of performing a Clifford gate $G$,
${\mathsf{cMPC}}$ updates the keys to
$\\{E^{\prime}_{j}=E_{j}(G\otimes\mathbbm{1}^{lt})^{\dagger}\\}_{j\in[n]}$.
These keys are still random Clifford keys. We can rewrite the state of the
real-protocol adversary in terms of $\\{E^{\prime}_{j}\\}_{j\in[n]}$ as
$\displaystyle A^{\Delta}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{
};\text{
}{(\underset{j\in[n]}{\otimes}E^{\prime}_{j}(G\otimes\mathbbm{1}^{lt}))({\mathsf{QECC.Enc}}(C(\cdot,\phi_{h})),0^{lnt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
(10) $\displaystyle=$ $\displaystyle
A^{\Delta}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{ };\text{
}{(\underset{j\in[n]}{\otimes}E^{\prime}_{j})({\mathsf{QECC.Enc}}(G\circ
C(\cdot,\phi_{h})),0^{lnt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
since the QECC supports transversal computation on these Clifford gates. As
the only difference between $\mathsf{Sim}^{\mathsf{RQC}}(C)$ and
$\mathsf{Sim}^{\mathsf{RQC}}(G\circ C)$ is the ideal functionalities they
interact with, we see that $\mathsf{Sim}^{\mathsf{RQC}}(G\circ C)$ also ends
with state (10).
For the protocol of measuring a logical qubit shared over registers
${\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}Q_{j}}\subseteq{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>the adversary receives $V_{j}=(E^{\prime}_{j}\otimes
Z^{c^{z}_{j}}X^{c^{x}_{j}}\mathsf{CNOT}_{c^{t}_{j}})E_{j}^{\dagger}$, so its
state can be described as
$\displaystyle A^{\prime\Delta}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{
};\text{ }{(\underset{j\in[n]}{\otimes}(E^{\prime}_{j}\otimes
Z^{c^{z}_{j}}X^{c^{x}_{j}}\mathsf{CNOT}_{c^{t}_{j}}))({\mathsf{QECC.Enc}}(C(\cdot,\phi_{h})),0^{lnt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
for some $A^{\prime}$ that can depend on $V_{j}$. Next, the adversary measures
${\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>To analyze ${\mathsf{cMPC}}$’s examination, we decompose the attack
$A^{\prime}$ according to Pauli basis on registers
${\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>That is, we write
$A^{\prime}={\Sigma_{x_{j}}}{A^{\prime}_{x_{j}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j},W}}{X^{x_{j}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>where
$A^{\prime}_{x_{j}}={\Sigma_{z_{j}}}{A^{\prime}_{x_{j},z_{j}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>Q_{j},W}}{Z^{z_{j}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>Similar to [DGJ+20], all Pauli $P$ with $x(P)\neq 0$ would fail
${\mathsf{cMPC}}$’s test with overwhelming probability over random
$(c^{z}_{j},c^{x}_{j},c^{t}_{j})$. Thus, the residual state passing the test
would be indistinguishable from
$\displaystyle
A^{\prime\Delta}_{0}\big{(}\underset{i\in[n]}{\cup}\rho_{i}\text{ };\text{
}{(\underset{j\in[n]}{\otimes}E^{\prime}_{j})({\mathsf{QECC.Enc}}(C(\cdot,\phi_{h}))_{[:l-1]},0^{(l-1)nt})}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>Q_{j}}},{\rho_{{\mathsf{aux}}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}W}}\big{)}$
(11)
while the string $r_{j}$ provided by $\mathsf{P}_{j}$ is
$(\mathsf{\Pi}({\mathsf{QECC.Enc}}(C(\cup_{i}\rho_{i},\phi_{h})_{[l]})_{[j]})(1,c^{t}_{j})\oplus
x_{j})$. Then the bit $b$ recovered by ${\mathsf{cMPC}}$ is
$\mathsf{\Pi}(C(\cup_{i}\rho_{i},\phi_{h}){[l]})$ because measurement is
transversal on our QECC. Now we turn to the simulated end state. There are two
main differences between $\mathsf{Sim}^{\mathsf{RQC}}(C)$ and
$\mathsf{Sim}^{\mathsf{RQC}}((\mathbbm{1}\otimes\mathsf{\Pi})\circ C)$. First,
the latter simulator encrypts part of the receiving portals under key
$\hat{E}_{j}^{(l)}$. Second, the latter simulator performs an additional
measurement step. As neither simulators give $\hat{e}_{S,j}^{(l)}$ to the
adversary, both $\hat{e}_{R,j}^{(l)}$ and
$\hat{E}_{j}^{(l)}\cdot\hat{e}_{R,j}^{(l)}$ are maximally mixed states and are
indistinguishable from the view of ${\mathcal{A}}$. Next, we decompose
$A^{\prime}={\Sigma_{x_{j}}}{A^{\prime}_{x_{j}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color@gray@fill{.5}R_{j},W}}{X^{x_{j}}}^{{\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>The only possible Pauli operators on
${\color[rgb]{.5,.5,.5}\definecolor[named]{pgfstrokecolor}{rgb}{.5,.5,.5}\pgfsys@color@gray@stroke{.5}\pgfsys@color<EMAIL_ADDRESS>that can yield equal measurements on $\hat{e}_{S,j}^{(l)}$ and
$\hat{e}_{R,j}^{(l)}$ are those without any bit-flips. Hence, conditioned that
the simulator does not make $\mathsf{P}_{j}$ send abort, the attack reduces to
$A^{\prime}_{0}$, and the end state is equal to (11).
Finally, we analyze the case when some corruptions are detected. If party
$\mathsf{P}_{j}$ is identified as corrupted, the output key $E_{j}$ will be
replaced with $\bot$ in both the real and simulated worlds, so that the $j$-th
share in (9) becomes the maximally mixed state in both worlds. When
$\mathsf{P}_{j}$ is added into the list Corr because the measurement step
failed, one easily sees that both worlds have residual states
(indistinguishably) described by
$(A^{\prime\Delta}-A^{\prime\Delta}_{0})(\cdot)$.
∎
### 6.3 Multiparty Quantum Computation
The final step is to transit from RQC to MPQC. After the computation, the
evaluators must send the encrypted and encoded outputs back to their owners.
Apparently, the outputs could turn out ruined when the owner unpacks the
messages. This situation is where AQA comes into play, which guarantees public
verifiable identifiable abort when some sender modifies the message. If no
such problem happens, the parties have accomplished their MPQC task. Our MPQC
protocol will be a direct concatenation of RQC and AQA.
Protocol 4 $(\Sigma^{{{\mathsf{BoBW}}_{0}}},\Pi^{{{\mathsf{BoBW}}_{0}}})$ for
the Preprocessing Best-of-Both-Worlds Multiparty Quantum Computation with
Public Verifiable Identifiable Abort
Parameters:
$n,t,\mathsf{t},(\ell_{1},\cdots,\ell_{n}),\ell_{{\mathsf{anc}}},\ell^{\prime}$
Preprocessing $\Sigma^{{\mathsf{BoBW}}_{0}}$:
1. 1.
Run $\Sigma^{\mathsf{RQC}}$. Then run
$\Sigma^{\mathsf{AQA}}(\mathsf{P}_{j},\mathsf{P}_{i})$ for every pair of
parties $i,j\in[n]$.
Online Input:
1. 1.
Everyone holds the circuit description $C$.
2. 2.
$\mathsf{P}_{i}$ holds private input $\rho_{i}\in\mathcal{D}^{\ell_{i}}$.
Protocol $\Pi^{{\mathsf{BoBW}}_{0}}$:
1. 1.
All parties run the protocol $\Pi^{\mathsf{RQC}}$ to compute $C$.
At any time if $|{\texttt{Corr}}|>\mathsf{t}$, ${\mathsf{cMPC}}$ publicly
aborts to all players in Corr.
2. 2.
Suppose $\sigma_{j}$ that $\mathsf{P}_{j}$ obtains from step 1 is
authenticated under key $E_{j}\neq\bot$.
${\mathsf{cMPC}}$ retrieves key pair
$(E^{\prime}_{i,j},E^{\prime\prime}_{i,j})$ that was given by
$\Sigma^{\mathsf{AQA}}(\mathsf{P}_{j},\mathsf{P}_{i})$.
${\mathsf{cMPC}}$ sends $V_{j}=(E^{\prime}_{1,j}\otimes\cdots\otimes
E^{\prime}_{1,n})E^{{\dagger}}_{j}$ to $\mathsf{P}_{j}$.
$\mathsf{P}_{j}$ calculates $V_{j}\cdot\sigma_{j}$ and parses the result into
$(\sigma^{\prime}_{1,j},\cdots,\sigma^{\prime}_{n,j})$.
3. 3.
AQA from party $\mathsf{P}_{j}$ to party $\mathsf{P}_{i}$: run
$\Pi^{\mathsf{AQA}}(\mathsf{P}_{j}(\sigma^{\prime}_{i,j}),\mathsf{P}_{i},{\mathsf{cMPC}})$
for $i,j\in[n]$.
$\mathsf{P}_{i}$ collects what it received
$\\{\sigma^{\prime\prime}_{i,j}\\}_{j\in[n]}$ from the other parties.
If AQA identifies $\mathsf{P}_{j}$ as malicious, put $\mathsf{P}_{j}$ into
Corr.
4. 4.
${\mathsf{cMPC}}$ sends keys $\\{E^{\prime\prime}_{i,j}\\}_{j\in[n]}$ to
$\mathsf{P}_{i}$ and publicly outputs classical output $r^{\prime}$.
Output:
1. 1.
$\mathsf{P}_{i}$ outputs
$\rho^{\prime}_{i}\leftarrow\mathsf{QECC.Dec}\big{(}E^{\prime\prime{\dagger}}_{i,1}\cdot\sigma^{\prime\prime}_{i,1},\cdots,E^{\prime\prime{\dagger}}_{i,n}\cdot\sigma^{\prime\prime}_{i,n}\big{)}$
and $r^{\prime}$.
###### Theorem 6.3.
$(\Sigma^{{\mathsf{BoBW}}_{0}},\Pi^{{{\mathsf{BoBW}}_{0}}})$ is a best-of-
both-worlds multi-party quantum computation secure with public verifiable
identifiable abort of threshold $\mathsf{t}$ in the preprocessing MPC-hybrid
model as defined in 4.1.
###### Proof.
In short, the simulator runs a concatenation of $\mathsf{Sim}^{\mathsf{RQC}}$
with $\mathsf{Sim}^{\mathsf{AQA}}$. ${\mathsf{Ideal}}^{{\mathsf{MPQC}}}$ is
invoked in $\mathsf{Sim}^{\mathsf{RQC}}$ to get $\\{\rho^{\prime}_{i}\\}_{i\in
I}$. In the output recovery phase of $\mathsf{Sim}^{\mathsf{RQC}}$, the
simulator teleports $\mathsf{TP}^{\dagger}\cdot\tilde{e}_{S}$ prepared by
$\mathsf{Sim}^{\mathsf{AQA}}$ instead of $\hat{e}_{S}^{(0)}$. Whenever
$|{\texttt{Corr}}|>\mathsf{t}$, the simulator sends abort message from parties
in $|{\texttt{Corr}}|$ to ${\mathsf{Ideal}}^{\mathsf{MPQC}}$. In the output
recovery phase of $\mathsf{Sim}^{\mathsf{AQA}}$, the simulator authenticates
$\rho^{\prime}_{i}$ with fresh random keys, teleports the result to
$\mathsf{P}_{i}$, and sends the key out.
When $|I|\leq\mathsf{t}$, the protocol never reaches the abort condition. By
Lemma 6.2 and Lemma 5.1, the honest parties will certainly receive at least a
proportion $1-\frac{\mathsf{t}}{n}$ of their output codewords. Our requirement
on QECC then implies that the honest parties can recover their outputs. From
Lemma 6.2 and Lemma 5.1, we deduce that the protocol is indistinguishable from
simulation using ${\mathsf{Ideal}}^{{\mathsf{MPQC-Full}}}$ when
$|I|\leq\mathsf{t}$ and using ${\mathsf{Ideal}}^{\mathsf{MPQC}}$ when $|I|\leq
n-1-\mathsf{t}$.
∎
## 7 BoBW-MPQC-PVIA without Trusted Setup
So far, we have shown an MPQC-PVIA protocol in the preprocessing model. This
section presents how we discard the trusted setup and obtain an MPQC-PVIA
protocol without preprocessing. The idea here is to instantiate the
preprocessing phase with MPQC-SWIA e.g., [ACC+21].
##### Hierarchical MPQC-SWIA
We quickly recall the SWIA property of [ACC+21]. When the [ACC+21] protocol
aborts, their classical MPC outputs a partition of all parties into two
mutually distrusting groups. By observation, one of the two groups comprises
only malicious players, and the other group contains all honest players, with
some malicious players pretending to be honest. We can take advantage of this
information to get an MPQC-PVIA protocol. First, we replace the trusted setup
of our preprocessing MPQC-PVIA by running MPQC-SWIA, where it can create
preprocessing data (including portals, magic states, and ancilla) for each
party. Then, if the protocol fails, we allow those two groups to run their
MPQC-SWIA protocols independently. The parties will ultimately obtain reliable
preprocessing data within subgroups by iterating this procedure. Note that we
do not publicly identify anyone during this stage.
After generating enough reliable preprocessing data, every party can run the
protocol $\Pi^{{\mathsf{BoBW}}_{0}}$ within its group to obtain the output.
Since the honest parties are always in the same group, they will jointly
evaluate their outputs in the $\Pi^{{\mathsf{BoBW}}_{0}}$. Our approach
circumvents the false accusation problem encountered in [ACC+21] because the
parties in our protocol no longer accuse between groups. Instead, each group
runs its own $\Pi^{{\mathsf{BoBW}}_{0}}$, which only aborts dishonest members.
We emphasize that the parties may divide into groups when the protocol
executes successfully. In this situation, if the circuit of interest involves
public output, the legitimacy of each group must be determined, and the public
output selected accordingly. Our protocol does not attempt to address this. We
take a step back and consider only circuits with private outputs. At the end
of our protocol, either a public verifiable identifiable abort occurs, or each
party is satisfied with its private outcome. Dealing with circuits with public
outputs remains an intriguing open problem for one who would like to achieve a
stronger PVIA notion.
Now, we formulate our protocol. Let $\Pi^{\mathsf{SWIA}}$ be [ACC+21]’s MPQC-
SWIA protocol.
Protocol 5 $\Pi^{{\mathsf{BoBW}}_{1}}$ for the Best-of-Both-Worlds MPQC-PVIA
Input:
1. 1.
Everyone holds the circuit description $C$ which has only private outputs.
2. 2.
Party $\mathsf{P}_{i}$ holds private input
$\rho_{i}\in\mathcal{D}^{\ell_{i}}$.
Protocol:
1. 1.
Put $G=\\{\\{\mathsf{P}_{1},\cdots,\mathsf{P}_{n}\\}\\}$ as the initial
partition of parties (i.e., no partition).
Mark the set $\\{\mathsf{P}_{1},\cdots,\mathsf{P}_{n}\\}\in G$ as unfinished.
2. 2.
Repeat the following until all sets in $G$ are marked as finished:
For unfinished $S\in G$, the parties in $S$ run $\Pi^{\mathsf{SWIA}}$ together
on the circuit $\Sigma^{{\mathsf{BoBW}}_{0}}(C_{S})$, where $C_{S}$ is the
circuit that prepares default inputs for parties $\not\in S$ and runs $C$.
1. (a)
If the protocol succeeds, parties in $S$ obtain the required preprocessing
data. In this case, mark $S$ as finished.
2. (b)
Otherwise, the MPQC-SWIA protocol outputs a partition $\\{S_{0},S_{1}\\}$ of
$S$.
In this case, replace $S$ with the unfinished $S_{0}$ and $S_{1}$ in partition
$G$.
3. 3.
Run $\Pi^{{\mathsf{BoBW}}_{0}}$ within every set $S\in G$.
Abort to all identified parties if any execution of
$\Pi^{{\mathsf{BoBW}}_{0}}$ aborts.
Output:
1. 1.
Upon abort, all parties output the identities collected from all
$\Pi^{{\mathsf{BoBW}}_{0}}$.
2. 2.
Otherwise, $\mathsf{P}_{i}$ outputs the result obtained from his/her execution
of $\Pi^{{\mathsf{BoBW}}_{0}}$.
###### Theorem 7.1.
$\Pi^{{\mathsf{BoBW}}_{1}}$ is a best-of-both-worlds multi-party quantum
computation secure with public verifiable identifiable abort over circuits
with only private outputs in the MPC-hybrid model as defined in 4.1.
###### Proof.
Since our protocol makes only sequential calls to MPQC-SWIA protocols, the
security of MPQC-SWIA allows one to replace these protocol executions with the
ideal functionality of MPQC-SWIA. The resulting hybrid world $\mathcal{H}_{1}$
is indistinguishable from the real world. The ideal functionality of MPQC-SWIA
in $\mathcal{H}_{1}$ translates to the following game. At the beginning of
$\mathcal{H}_{1}$, the adversary chooses how to partition all parties step by
step, but cannot partition the honest parties. Afterward, there is a trusted
setup on $\Sigma^{{\mathsf{BoBW}}_{0}}$ followed by
$\Pi^{{\mathsf{BoBW}}_{0}}$ for each group of the partition. Since these
preprocessing BoBW protocols are independent instances, 6.3 implies that they
can be replaced with the ideal world of BoBW-MPQC-PVIA. It is easy to see that
the resulting hybrid $\mathcal{H}_{2}$ preserves the PVIA property. To argue
full security when $|I|\leq\mathsf{t}$, one observes that the group containing
all the honest parties has at least a $1-\frac{\mathsf{t}}{n}$ proportion of
honest parties, so the full security of BoBW-MPQC-PVIA is also preserved. ∎
## 8 Extentions
### 8.1 Constant Round MPQC-PVIA in the Preprocessing Model
When aiming only for MPQC-PVIA, our BoBW-MPQC-PVIA protocol in the
preprocessing model can be made round efficient. The reason is that we no
longer need to keep the data distributed across parties, which eliminates the
need for information exchange during computation. Instead, a designated party
can hold all the data, and the entire process can become constant-round using
quantum garbled circuits for computation as in [BCKM21].
###### Theorem 8.1.
There is a constant-round multi-party quantum computation secure with public
verifiable identifiable abort in the preprocessing MPC-hybrid model.
Protocol 6
$(\Sigma^{{{\mathsf{Const}}{\mathsf{MPQC}}}},\Pi^{{{\mathsf{Const}}{\mathsf{MPQC}}}})$
for the Preprocessing Constant-Round MPQC-PVIA
Preprocessing $\Sigma^{{{\mathsf{Const}}{\mathsf{MPQC}}}}$:
1. 1.
Run $\Sigma^{\mathsf{IE}}$. Then run
$\Sigma^{\mathsf{AQA}}(\mathsf{P}_{i},\mathsf{P}_{1})$ for every party
$i\in[n]$.
Protocol $\Pi^{{{\mathsf{Const}}{\mathsf{MPQC}}}}$:
1. 1.
All parties run the input encoding protocol $\Pi^{\mathsf{IE}}$.
The designated evaluator $\mathsf{P}_{1}$ obtains ciphertext $\sigma$ under
key $E$.
2. 2.
${\mathsf{cMPC}}$ generates quantum garbled circuit:
1. (a)
Retrieve key pair $(E^{\prime}_{i},E^{\prime\prime}_{i})$ that was created by
$\Sigma^{\mathsf{AQA}}(\mathsf{P}_{i},\mathsf{P}_{1})$.
2. (b)
Define the augmented circuit $C^{\prime}$ as:
* •
Measure all the traps. If any trap is non-zero, end by outputting $\bot$.
* •
Evaluate $C$ on the inputs and ancilla states, and apply
$E^{\prime}_{1}\otimes\cdots\otimes E^{\prime}_{n}$.
3. (c)
Run $(\hat{E},\hat{C})\leftarrow\mathsf{QG.Garble}(C^{\prime})$.
4. (d)
Send the gate $V=\hat{E}E^{\dagger}$ and the garbled circuit $\hat{C}$ to the
evaluator $\mathsf{P}_{1}$.
3. 3.
$\mathsf{P}_{1}$ evaluates the garbled circuit
$\sigma^{\prime}\leftarrow\mathsf{QG.Eval}(V\cdot\sigma,\hat{C})$.
$\mathsf{P}_{1}$ parses the result into
$(\sigma^{\prime}_{1},\cdots,\sigma^{\prime}_{n})$.
4. 4.
AQA from party $\mathsf{P}_{1}$ to party $\mathsf{P}_{i}$ for $i\in[n]$: run
$\Pi^{\mathsf{AQA}}(\mathsf{P}_{1}(\sigma^{\prime}_{i}),\mathsf{P}_{i},{\mathsf{cMPC}})$.
If any $\Pi^{\mathsf{AQA}}$ fails, ${\mathsf{cMPC}}$ publicly aborts to
$\mathsf{P}_{1}$.
Otherwise, $\mathsf{P}_{i}$ receives the ciphertext
$\sigma^{\prime\prime}_{i}$ successfully.
5. 5.
${\mathsf{cMPC}}$ sends key $E^{\prime\prime}_{i}$ to $\mathsf{P}_{i}$.
Output: $\mathsf{P}_{i}$ outputs
$\rho^{\prime}_{i}\leftarrow\mathsf{C.Dec}_{E^{\prime\prime}_{i}}(\sigma^{\prime\prime}_{i})$.
The security follows from adapting the proof of 6.3 with the proofs [BCKM21,
BY22] that use garbled circuits to perform the computation.
### 8.2 MPQC-PVIA from Trap-code AQA
We also discover the possibility of constructing MPQC-PVIA using Trap-code
form AQA (Figure 4). Existing MPQC protocols against a dishonest majority
[DGJ+20, ACC+21, BCKM21] only consider Clifford authentication code because of
the obstruction of the Trap authentication code’s checking step. For example,
if we replace the Clifford code in [ACC+21] with Trap code, the player will
not know if the errors are caused by the previous player or other prior
players. This is because trap code requires a permutation to hide the input
qubits’ information. By only considering the Trap authentication code, one
cannot guarantee the secrecy of the qubits’ location while still identifying
the suspect.
Thanks to our AQA, by slightly twisting Trap-code and plugging it into our
AQA, we can construct the first MPQC-PVIA protocol from Trap-code form AQA in
the preprocessing model. The protocol will be very similar to the Clifford-
form protocol (Section 6.3) but with Trap-form IE and AQA (Figure 4). Similar
to [ACC+21], we make use of a verifiable quantum fully homomorphic encryption
scheme (vQFHE) [ADSS17] for computation on Trap code.
Protocol 7
$(\Sigma^{{\mathsf{BoBW}}_{\mathsf{TP}}},\Pi^{{\mathsf{BoBW}}_{\mathsf{TP}}})$
for the Trap-form Preprocessing MPQC-PVIA
Preprocessing: $\Sigma^{{\mathsf{BoBW}}_{\mathsf{TP}}}$
1. 1.
Run $\Sigma^{\mathsf{IE}}$, distributing not only Trap-code ciphertexts on EPR
pairs and ancilla states, but also gadgets required for vQFHE computation
under secret key $\mathsf{sk}$ given to ${\mathsf{cMPC}}$.
2. 2.
Run $\Sigma^{\mathsf{AQA}}(\mathsf{P}_{j},\mathsf{P}_{i})$ for all pairs
$i,j\in[n]$.
Protocol: $\Pi^{{\mathsf{BoBW}}_{\mathsf{TP}}}$
1. 1.
All parties run $\Pi^{\mathsf{IE}}$. so that each $\mathsf{P}_{i}$ obtains
several Trap-code ciphertexts $\\{\sigma_{i,l}\\}$ corresponding to QECC
codewords of inputs, under the same permutation key $\pi_{i}$.
2. 2.
${\mathsf{cMPC}}$ initializes the set ${\texttt{Corr}}=\\{\\}$.
At any time if $|{\texttt{Corr}}|>\mathsf{t}$, ${\mathsf{cMPC}}$ publicly
abort to all parties in Corr.
3. 3.
$\mathsf{P}_{i}$ computes via
$(\\{\tilde{\sigma}_{i,l}\\},\mathsf{log}_{i})\leftarrow\mathsf{TrapTP.Eval}(C,\\{\sigma_{i,l}\\})$
and sends $\mathsf{log}_{i}$ to ${\mathsf{cMPC}}$.
4. 4.
${\mathsf{cMPC}}$ runs $\mathsf{CheckLog}(\mathsf{log}_{i},\mathsf{sk})$.If
the output is $\bot$, ${\mathsf{cMPC}}$ publicly abort $\mathsf{P}_{i}$.
5. 5.
$\mathsf{P}_{i}$ sends $\tilde{\sigma}_{i,l}$ to the respective receiver
$\mathsf{P}_{j}$ via $\Pi^{\mathsf{AQA}}$. The receiver $\mathsf{P}_{j}$ gets
$\sigma^{\prime}_{j}$.
6. 6.
${\mathsf{cMPC}}$ sends the decoding instruction $V_{\mathsf{Dec},i}$ to the
respective party $\mathsf{P}_{i}$.
Output: $\mathsf{P}_{i}$ outputs $\rho^{\prime}_{i}\leftarrow
V_{\mathsf{Dec},i}(\sigma^{\prime}_{i})$.
The Trap-form protocol
$(\Sigma^{{\mathsf{BoBW}}_{\mathsf{TP}}},\Pi^{{\mathsf{BoBW}}_{\mathsf{TP}}})$
satisfies the statement of 6.3. We omit the proof because of its similarity to
the Clifford-code form MPQC-PVIA while basing on the security guarantee of
Trap-code form AQA.
## Acknowledgement
We would like to thank Kai-Min Chung and Jiapeng Zhang for their helpful
discussions. This research is partially supported by NSTC QC project under
Grant no. NSTC 111-2119-M-001-004- and partially supported by NSF CAREER award
2141536.
## References
* [ABDR04] Andris Ambainis, Harry Buhrman, Yevgeniy Dodis, and Hein Rohrig. Multiparty quantum coin flipping. In Proceedings. 19th IEEE Annual Conference on Computational Complexity, 2004., pages 250–259. IEEE, 2004.
* [ABOEM17] Dorit Aharonov, Michael Ben-Or, Elad Eban, and Urmila Mahadev. Interactive proofs for quantum computations. arXiv preprint arXiv:1704.04487, 2017.
* [ACC+21] Bar Alon, Hao Chung, Kai-Min Chung, Mi-Ying Huang, Yi Lee, and Yu-Ching Shen. Round efficient secure multiparty quantum computation with identifiable abort. In Annual International Cryptology Conference, pages 436–466. Springer, 2021.
* [ADSS17] Gorjan Alagic, Yfke Dulek, Christian Schaffner, and Florian Speelman. Quantum fully homomorphic encryption with verification. In International Conference on the Theory and Application of Cryptology and Information Security, pages 438–467. Springer, 2017.
* [BCKM21] James Bartusek, Andrea Coladangelo, Dakshita Khurana, and Fermi Ma. On the round complexity of secure quantum computation. In Annual International Cryptology Conference, pages 406–435. Springer, 2021.
* [BDO14] Carsten Baum, Ivan Damgård, and Claudio Orlandi. Publicly auditable secure multi-party computation. In International Conference on Security and Cryptography for Networks, pages 175–196. Springer, 2014.
* [BDSW96] Charles H. Bennett, David P. DiVincenzo, John A. Smolin, and William K. Wootters. Mixed-state entanglement and quantum error correction. Phys. Rev. A, 54:3824–3851, Nov 1996.
* [BGS13] Anne Broadbent, Gus Gutoski, and Douglas Stebila. Quantum one-time programs. In Annual Cryptology Conference, pages 344–360. Springer, 2013\.
* [BJSW16] Anne Broadbent, Zhengfeng Ji, Fang Song, and John Watrous. Zero-knowledge proof systems for qma. In 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), pages 31–40. IEEE, 2016.
* [BLOO11] Amos Beimel, Yehuda Lindell, Eran Omri, and Ilan Orlov. 1/p-secure multiparty computation without honest majority and the best of both worlds. In Annual Cryptology Conference, pages 277–296. Springer, 2011\.
* [BOCG+06] Michael Ben-Or, Claude Crepeau, Daniel Gottesman, Avinatan Hassidim, and Adam Smith. Secure multiparty quantum computation with (only) a strict honest majority. In 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06). IEEE, 2006.
* [BOSSV20] Carsten Baum, Emmanuela Orsini, Peter Scholl, and Eduardo Soria-Vazquez. Efficient constant-round mpc with identifiable abort and public verifiability. In Annual International Cryptology Conference, pages 562–592. Springer, 2020.
* [BW16] Anne Broadbent and Evelyn Wainewright. Efficient simulation for quantum message authentication. In International Conference on Information Theoretic Security, pages 72–91. Springer, 2016.
* [BY22] Zvika Brakerski and Henry Yuen. Quantum garbled circuits. In Proceedings of the 54th Annual ACM SIGACT Symposium on Theory of Computing, pages 804–817, 2022.
* [CGL99] Richard Cleve, Daniel Gottesman, and Hoi-Kwong Lo. How to share a quantum secret. Physical review letters, 83(3):648, 1999.
* [CGS02] Claude Crépeau, Daniel Gottesman, and Adam Smith. Secure multi-party quantum computation. In Proceedings of the thiry-fourth annual ACM symposium on Theory of computing - STOC '02. ACM Press, 2002.
* [CLLW22] Kai-Min Chung, Yi Lee, Han-Hsuan Lin, and Xiaodi Wu. Constant-round blind classical verification of quantum sampling. In Annual International Conference on the Theory and Applications of Cryptographic Techniques, pages 707–736. Springer, 2022.
* [DGJ+20] Yfke Dulek, Alex B Grilo, Stacey Jeffery, Christian Majenz, and Christian Schaffner. Secure multi-party quantum computation with a dishonest majority. Advances in Cryptology - EUROCRYPT 2020., 2020.
* [DNS12] Frédéric Dupuis, Jesper Buus Nielsen, and Louis Salvail. Actively secure two-party evaluation of any quantum operation. In Annual Cryptology Conference, pages 794–811. Springer, 2012\.
* [IKK+11] Yuval Ishai, Jonathan Katz, Eyal Kushilevitz, Yehuda Lindell, and Erez Petrank. On achieving the “best of both worlds” in secure multiparty computation. SIAM journal on computing, 40(1):122–141, 2011.
* [IOZ14] Yuval Ishai, Rafail Ostrovsky, and Vassilis Zikas. Secure multi-party computation with identifiable abort. In Advances in Cryptology – CRYPTO 2014, pages 369–386. Springer Berlin Heidelberg, 2014.
* [Kat07] Jonathan Katz. On achieving the” best of both worlds” in secure multiparty computation. In Proceedings of the thirty-ninth annual ACM symposium on Theory of computing, pages 11–20, 2007.
* [Mah18] Urmila Mahadev. Classical verification of quantum computations. In 2018 IEEE 59th Annual Symposium on Foundations of Computer Science (FOCS), pages 259–267. IEEE, 2018.
* [MGW87] Silvio Micali, Oded Goldreich, and Avi Wigderson. How to play any mental game. In Proceedings of the Nineteenth ACM Symp. on Theory of Computing, STOC, pages 218–229. ACM, 1987.
* [RBO89] Tal Rabin and Michael Ben-Or. Verifiable secret sharing and multiparty protocols with honest majority. In Proceedings of the twenty-first annual ACM symposium on Theory of computing, pages 73–85, 1989.
* [Unr10] Dominique Unruh. Universally composable quantum multi-party computation. In Annual International Conference on the Theory and Applications of Cryptographic Techniques, pages 486–505. Springer, 2010.
* [Yao86] Andrew Chi-Chih Yao. How to generate and exchange secrets. In 27th Annual Symposium on Foundations of Computer Science (sfcs 1986), pages 162–167, 1986.
|
# Greybody Radiation of scalar and Dirac perturbations of NUT Black Holes
Ahmad Al-Badawi<EMAIL_ADDRESS>Department of Physics, Al-Hussein Bin
Talal University, P. O. Box: 20, 71111, Ma’an, Jordan. Sara Kanzi
<EMAIL_ADDRESS>İzzet Sakallı<EMAIL_ADDRESS>Physics
Department, Eastern Mediterranean University, Famagusta, North Cyprus via
Mersin 10, Turkey.
###### Abstract
We consider the spinorial wave equations, namely the Dirac and the Klein-
Gordon equations, and greybody radiation in the NUT black hole spacetime. To
this end, we first study the Dirac equation in NUT spacetime by using a null
tetrad in the Newman-Penrose (NP) formalism. Next, we separate the Dirac
equation into radial and angular sets. The angular set is solved in terms of
associated Legendre functions. With the radial set, we obtain the decoupled
radial wave equations and derive the one-dimensional Schrödinger wave
equations together with effective potentials. Then, we discuss the potentials
by plotting them as a function of radial distance in a physically acceptable
region. We also study the Klein-Gordon equation to compute the greybody
factors (GFs) for both bosons and fermions. The influence of the NUT parameter
on the GFs of the NUT spacetime is investigated in detail.
Hawking Radiation, NUT, Black Hole, Dirac Equation, Klein-Gordon Equation,
Greybody Factor, Effective Potential, Perturbation
††preprint: HEP/123-qed
###### Contents
1. I Introduction
2. II Dirac equation in NUT BH spacetime
3. III Solutions of angular and radial equations
1. III.1 Angular equations
2. III.2 Radial equations
4. IV Greybody Radiation from NUT BH
1. IV.1 Klein-Gordon equation
2. IV.2 GFs of bosons
3. IV.3 GFs of fermions
5. V Conclusion
## I Introduction
NUT spacetimes are special case of the family of electrovac spacetimes of
Petrov type-D given by the Plebański-Demiański class of solutions 1 . These
axially symmetric spacetimes are characterized by the seven parameters: mass,
rotation, acceleration, cosmological constant, magnetic charge, electric
charge, and NUT parameter. We are interested in the solution, which is
characterized by two physical parameters: the gravitational mass and the NUT
parameter. From now on, we will refer to this spacetime [see metric (1)] as
the NUT black hole (BH). In fact, the NUT BH is a solution to the vacuum
Einstein equation that was found by Newman–Unti–Tamburino (NUT) in 1963 2 .
Since the discovery of the NUT solution, its physical meaning is a matter of
debate. The interpretations of the NUT solution and thus the NUT parameter are
still a controversy and it has attracted lots of attention over the years.
Misner approach 3 to the interpretation of the NUT solution aims at
eliminating singular regions at the expense of introducing the “periodic” time
and, consequently, the $S^{3}$ topology for the hypersurfaces $r=const$. In 4
the NUT solution was interpreted as representing the exterior field of a mass
located at the origin together with a semi-infinite massless source of angular
momentum. Whereas in 5 the solution was interpreted as the presence of two
semi-infinite counter-rotating singular regions endowed with negative masses
and infinitely large angular momenta.
Regarding to the NUT parameter, in general it was interpreted as a
gravitomagnetic charge bestowed upon the central mass 6 . However, the exact
physical meaning has not been attained yet. In 7 it was explained that if the
NUT parameter dominates the rotation parameter, then this leaves the spacetime
free of curvature singularities. But if the rotation parameter dominates the
NUT parameter, the solution is Kerr-like, and a ring curvature singularity
arises. Conversely, exact interpretation of the NUT parameter becomes possible
when a static Schwarzschild mass is immersed in a stationary, source free
electromagnetic universe 8 . In this case, the NUT parameter is interpreted as
a twist parameter of the electromagnetic universe. However, in the absence of
this field, it reduces to the twist of the vacuum spacetime. A different
interpretation of the NUT spacetime has been clarified via its thermodynamics.
In 9 , it was shown that if the NUT parameter is interpreted as possessing
simultaneously rotational and electromagnetic features then the thermodynamic
quantities satisfy the Bekenstein-Smarr mass formula.
In article 10 , Misner called the NUT metric a ’counter-example to almost
anything’: it means that it has so many peculiar properties. Following this
line of thought, the aim of the present paper is to study the scalar and Dirac
perturbations of the NUT BH and analyze their corresponding greybody
radiation. Studies of this type have always been important in the field of
quantum gravity. Furthermore, an effective way to understand and analyze
characteristics of a spacetime is to study its behavior under different kinds
of perturbations including spinor and gauge fields. Also, one can pave the way
to study the GFs of the NUT spacetime in order to investigate the evolution of
perturbations in the exterior region of the BH 11 ; 12 . The analytical
expressions of the solution obtained in this paper could be useful for the
study of the thermodynamical properties of the spinor fields in NUT geometry.
In the literature, the scalar/Dirac equation and greybody radiation in BH
geometry have been extensively studied. For example, wave analysis of the
Schwarzschild BH was performed in 13 ; 14 ; 15 ; 16 ; 17 ; 18 , for the Kerr
BH in 19 ; 20 ; 21 ; 22 , for the Kerr Taub-NUT BH in 23 , for the Kerr-Newman
AdS BH in 24 , and for the Reissner-Nordsrtöm de Sitter BH in 25 . Meanwhile,
it is worth noting that the GFs measure the deviation of the thermal spectra
from the black body or the so-called Hawking radiation 26 ; 27 . Essentially,
GF is a physical quantity that relates to the quantum nature of a BH. A high
value of GF shows a high probability that Hawking radiation can reach to
spatial infinity (or an observer) 28 ; 29 . Today, there are different methods
to compute the GF such as the matching method 30 ; 31 , rigorous bound method
32 , the WKB approximation 33 ; 34 , and analytical method for the various of
spin fields 35 ; 36 ; 37 ; 38 ; 38n ; 39 ; 40 . Here, we shall employ the
general method of semi-analytical bounds for the GFs. To this end, we shall
examine the characteristic bosonic and fermionic quantum radiation (i.e., the
GF) of NUT BH based on the current studies in the literature (see for example
Harmark:2007jy ) and our recent studies 37 ; 38 ; 38n . Although the
methodology followed is parallel to the procedure seen in 37 ; 38 ; 38n , due
to the unique characteristics of each BH solution, the calculations have their
own difficulties and therefore the results obtained will differ significantly.
In those recent studies 37 ; 38 ; 38n , for example, we have revealed the
effects of quintessence, negative cosmological constant, and Lorentz symmetry
breaking parameters on the quantum thermal radiation of black holes. However,
in this study, we aim to reveal the effect of the NUT parameter on the GF and
thus present a theoretical study to the literature that will contribute to the
black hole classification studies according to GF measurements that might be
obtained in the (distant) future.
The paper is organized as follows: In the next section, we obtain the Dirac
equation in the NUT spacetime and separate the coupled equations into the
angular and radial parts. In Sect. 3, analytical solutions to the angular
equations are sought. We also study the radial wave equations and examine the
behaviors of the effective potential that emerge in the transformed radial
equations. In Sect. 4, we study the Klein–Gordon equation in the NUT BH
spacetime. Then, we derive the associated effective potential of bosons. Then,
we compute the GFs of the NUT BH spacetime for both fermions and bosons in
Sect. 5. We draw our conclusions in Sect. 6.
## II Dirac equation in NUT BH spacetime
The NUT metric is a solution to the vacuum Einstein equation that depends on
two parameters: mass $M$ and NUT parameter $l$ 1 . A detailed discussion on
the NUT spacetime can be found in the famous monograph of Griffiths and
Podolsky 41 and in the specific studies given in 42 ; 43 ; 44 . In spherical
coordinates $(t,r,\theta,\phi)$ the NUT BH metric reads
$ds^{2}=f(r)\left[dt-2l\cos\theta
d\phi\right]^{2}-\frac{1}{f(r)}dr^{2}-\left(r^{2}+l^{2}\right)\left(d\theta^{2}+\sin^{2}\theta
d\phi^{2}\right),$ (1)
where
$f(r)=1-\frac{2\left(Mr+l^{2}\right)}{r^{2}+l^{2}}>0.$ (2)
The distinctive feature of the NUT metric is the presence of the NUT parameter
$l$ that causes the spacetime to be asymptotically non-flat. The horizons are
defined by $f(r)=0$ and hence they will occur at
$r_{\pm}=M\pm\sqrt{M^{2}+l^{2}}$.
Following the NP formalism, we introduce null tetrads $(l,n,m,\overline{m})$
to satisfy orthogonality relations: $l.n=-m.\overline{m}=1.$ Thus, we write
the basis vectors of null tetrad in terms of elements of the NUT geometry as
follows
$l_{\mu}=(1,-\frac{1}{f\left(r\right)},0,-2l\cos\theta),$
$n_{\mu}=\frac{1}{2}(f\left(r\right),1,0,-2lf(r)\cos\theta),$
$m_{\mu}=\sqrt{\frac{r^{2}+l^{2}}{2}}(0,0,-1,-i\sin\theta),$
$\overline{m}_{\mu}=\sqrt{\frac{r^{2}+l^{2}}{2}}(0,0,-1,i\sin\theta),$ (3)
where a bar over a quantity denotes complex conjugation and the dual co-tetrad
of Eq. (3) are given by
$l^{\mu}=(\frac{1}{f\left(r\right)},1,0,0),$
$n^{\mu}=\frac{1}{2}(1,-f(r),0,0),$
$m^{\mu}=\frac{1}{\sqrt{2\left(r^{2}+l^{2}\right)}}(2il\cot\theta,0,1,\frac{i}{\sin\theta}),$
$\overline{m}^{\mu}=\frac{1}{\sqrt{2\left(r^{2}+l^{2}\right)}}(-2il\cot\theta,0,1,\frac{-i}{\sin\theta}).$
(4)
Using the above null tetrad one can find the non-zero spin coefficients in
terms of NUT metric:
$\epsilon=\frac{il}{2\left(r^{2}+l^{2}\right)};\gamma=\frac{1}{4}\left(f^{\prime}(r)+\frac{ilf(r)}{r^{2}+l^{2}}\right);\alpha=-\beta=\frac{-\cot\theta}{2\sqrt{2\left(r^{2}+l^{2}\right)}}$
$\rho=\frac{-r+il}{r^{2}+l^{2}};\mu=\frac{f(r)}{2}\frac{-r+il}{r^{2}+l^{2}}.$
(5)
Their corresponding directional derivatives become
$D=l^{\mu}\partial_{\mu}=\frac{1}{f(r)}\frac{\partial}{\partial
t}+\frac{\partial}{\partial r},$
$\Delta=n^{\mu}\partial_{\mu}=\frac{1}{2}\frac{\partial}{\partial
t}-\frac{f(r)}{2}\frac{\partial}{\partial r},$
$\delta=m^{\mu}\partial_{\mu}=\frac{1}{\sqrt{2\left(r^{2}+l^{2}\right)}}(2il\cot\theta\frac{\partial}{\partial
t}+\frac{\partial}{\partial\theta}+\frac{i}{\sin\theta}\frac{\partial}{\partial\phi}),$
$\bar{\delta}=\overline{m}^{\mu}\partial_{\mu}=\frac{1}{\sqrt{2\left(r^{2}+l^{2}\right)}}(-2il\cot\theta\frac{\partial}{\partial
t}+\frac{\partial}{\partial\theta}-\frac{i}{\sin\theta}\frac{\partial}{\partial\phi}),$
(6)
The Dirac equations in the curved spacetime are given by 45
$\left(D+\epsilon-\rho\right)F_{1}+\left(\overline{\delta}+\pi-\alpha\right)F_{2}=i\mu_{\ast}G_{1},$
$\left(\Delta+\mu-\gamma\right)F_{2}+\left(\delta+\beta-\tau\right)F_{1}=i\mu_{\ast}G_{2},$
$\left(D+\overline{\epsilon}-\overline{\rho}\right)G_{2}-\left(\delta+\overline{\pi}-\overline{\alpha}\right)G_{1}=i\mu_{\ast}F_{2},$
$\left(\Delta+\overline{\mu}-\overline{\gamma}\right)G_{1}-\left(\overline{\delta}+\overline{\beta}-\overline{\overline{\tau}}\right)G_{2}=i\mu_{\ast}F_{1},$
(7)
where $\mu_{\ast}$ is the mass of the Dirac particle.
The form of the Dirac equation suggests that we define the spinor fields as
follows 45 ,
$F_{1}=F_{1}\left(r,\theta\right)e^{i\left(kt+n\phi\right)},$
$F_{2}=F_{2}\left(r,\theta\right)e^{i\left(kt+n\phi\right)},$
$G_{1}=G_{1}\left(r,\theta\right)e^{i\left(kt+n\phi\right)},$
$G_{2}=G_{2}\left(r,\theta\right)e^{i\left(kt+n\phi\right)}.$ (8)
where $k$ is the frequency of the incoming wave and $n$ is the azimuthal
quantum number of the wave.
Substituting the appropriate spin coefficients (5) and the spinors (8) into
the Dirac equation (7) , we obtain
$\sqrt{r^{2}+l^{2}}\left(\emph{D}-\frac{il}{2\left(r^{2}+l^{2}\right)}\right)F_{1}+\frac{1}{\sqrt{2}}\mathit{L}F_{2}=i\mu_{0}\sqrt{r^{2}+l^{2}}G_{1},$
$\frac{f}{2}\sqrt{r^{2}+l^{2}}\left(\mathit{D}^{{\dagger}}+\frac{f^{\prime}}{2f}-\frac{il}{2\left(r^{2}+l^{2}\right)}\right)F_{2}-\frac{1}{\sqrt{2}}\mathit{L}^{{\dagger}}F_{1}=-i\mu_{0}\sqrt{r^{2}+l^{2}}G_{2},$
$\sqrt{r^{2}+l^{2}}\left(\emph{D}+\frac{il}{2\left(r^{2}+l^{2}\right)}\right)G_{2}-\frac{1}{\sqrt{2}}\mathit{L}^{{\dagger}}G_{1}=i\mu_{0}\sqrt{r^{2}+l^{2}}F_{2},$
$\frac{f}{2}\sqrt{r^{2}+l^{2}}\left(\mathit{D}^{{\dagger}}+\frac{f^{\prime}}{2f}+\frac{il}{2\left(r^{2}+l^{2}\right)}\right)G_{1}+\frac{1}{\sqrt{2}}\mathit{L}G_{2}=-i\mu_{0}\sqrt{r^{2}+l^{2}}F_{1},$
(9)
where, $\mu_{0}=\sqrt{2}\mu_{\ast}$ and the above operators are given by
$\emph{D}=\frac{\partial}{\partial r}+\frac{r}{r^{2}+l^{2}}+\frac{ik}{f},$
$\mathit{D}^{{\dagger}}=\frac{\partial}{\partial
r}+\frac{r}{r^{2}+l^{2}}-\frac{ik}{f}$
$\mathit{L}=\frac{\partial}{\partial\theta}+\left(\frac{1}{2}+2lk\right)\cot\theta+\frac{n}{\sin\theta},$
$\mathit{L}^{{\dagger}}=\frac{\partial}{\partial\theta}+\left(\frac{1}{2}-2lk\right)\cot\theta-\frac{n}{\sin\theta}.$
(10)
Further, we define
$F_{1}\left(r,\theta\right)=R_{+1/2}\left(r\right)A_{+1/2}\left(\theta\right),F_{2}\left(r,\theta\right)=R_{-1/2}\left(r\right)A_{-1/2}\left(\theta\right),G_{1}\left(r,\theta\right)=R_{-1/2}\left(r\right)A_{+1/2}\left(\theta\right)$
,$G_{2}\left(r,\theta\right)=R_{+1/2}\left(r\right)A_{-1/2}\left(\theta\right)$
hence, the set (9) can be separated into angular and radial parts as
$\sqrt{r^{2}+l^{2}}\emph{D}R_{+1/2}=\left(\lambda+i\mu_{0}\sqrt{r^{2}+l^{2}}\right)R_{-1/2},$
(11)
$\frac{f}{2}\sqrt{r^{2}+l^{2}}\mathit{D}^{{\dagger}}R_{-1/2}=\left(\lambda-i\mu_{0}\sqrt{r^{2}+l^{2}}\right)R_{+1/2},$
(12)
$\frac{1}{\sqrt{2}}\mathit{L}A_{-1/2}=\lambda A_{+1/2},$ (13)
$\frac{1}{\sqrt{2}}\mathit{L}^{{\dagger}}A_{+1/2}=-\lambda A_{-1/2},$ (14)
where $\lambda$ is the separation constant.
## III Solutions of angular and radial equations
### III.1 Angular equations
Angular Eqs. (14) and (13) can be written as
$\frac{dA_{+1/2}}{d\theta}+\left(\left(\frac{1}{2}+2lk\right)\cot\theta+\frac{n}{\sin\theta}\right)A_{+1/2}=-\lambda
A_{-1/2},$ (15)
$\frac{dA_{-1/2}}{d\theta}+\left(\left(\frac{1}{2}-2lk\right)\cot\theta-\frac{n}{\sin\theta}\right)A_{-1/2}=\lambda
A_{+1/2}.$ (16)
We note that the structure of the angular equations admits the similar
solutions for both $A_{+1/2}$ and $A_{-1/2}$. Therefore, it is sufficient to
focus on one of the angular equations. To this end, we apply the following
transformations
$A_{+1/2}\left(\theta\right)=\cos\left(\frac{\theta}{2}\right)S_{1}(\theta)+\sin\left(\frac{\theta}{2}\right)S_{2}(\theta),$
(17)
$A_{-1/2}\left(\theta\right)=-\sin\left(\frac{\theta}{2}\right)S_{1}(\theta)+\cos\left(\frac{\theta}{2}\right)S_{2}(\theta).$
(18)
By setting $x=\cos\theta$, one can write Eq. (15) as a decoupled second order
differential equation:
$\left(1-x^{2}\right)\frac{d^{2}S_{1}}{dx^{2}}-2x\frac{dS_{1}}{dx}+\left[\lambda\left(\lambda+1\right)-\frac{\left(n+\frac{1}{2}\right)^{2}}{1-x^{2}}\right]S_{1}=0,$
(19)
with $\left(n+\frac{1}{2}\right)^{2}\leq\lambda^{2}$. The solution to Eq. (19)
can be expressed in terms of the associated Legendre functions 46 :
$P_{\lambda}^{n}\left(x\right)=\left(1-x^{2}\right)^{\tau/2}\frac{d^{n}}{dx^{n}}P_{\lambda}\left(x\right).$
(20)
### III.2 Radial equations
The radial equations (11-12) can be rearranged as
$\sqrt{r^{2}+l^{2}}\left(\frac{d}{dr}+\frac{r}{r^{2}+l^{2}}+\frac{ik}{f}\right)R_{+1/2}=\left(\lambda-i\mu_{0}\sqrt{r^{2}+l^{2}}\right)R_{-1/2},$
(21)
$\frac{f}{2}\sqrt{r^{2}+l^{2}}\left(\frac{d}{dr}+\frac{r}{r^{2}+l^{2}}+\frac{f^{\prime}}{2f}-\frac{ik}{f}\right)R_{-1/2}=\left(\lambda+i\mu_{0}\sqrt{r^{2}+l^{2}}\right)R_{+1/2}.$
(22)
To obtain the radial equations in the form of one-dimensional Schrödinger like
wave equations, we perform the following transformations
$P_{+1/2}=\sqrt{r^{2}+l^{2}}R_{+1/2},\qquad
P_{-1/2}=\sqrt{\frac{f}{2}}\sqrt{r^{2}+l^{2}}R_{-1/2}.$ (23)
Then, Eqs. (21) and (22) transform to
$\left(\frac{d}{dr}+i\frac{k}{f}\right)P_{+1/2}=\sqrt{\frac{2}{f}}\left(\frac{\lambda}{\sqrt{r^{2}+l^{2}}}-i\mu_{0}\right)P_{-1/2},$
(24)
$\left(\frac{d}{dr}-i\frac{k}{f}\right)P_{-1/2}=\sqrt{\frac{2}{f}}\left(\frac{\lambda}{\sqrt{r^{2}+l^{2}}}+i\mu_{0}\right)P_{+1/2},$
(25)
and defining the tortoise coordinate $r_{\ast}$ as
$\frac{dr_{\ast}}{dr}=\frac{1}{f}$, Eqs. (24) and (25) become
$\left(\frac{d}{dr_{\ast}}+ik\right)P_{+1/2}=\sqrt{2f}\left(\frac{\lambda}{\sqrt{r^{2}+l^{2}}}-i\mu_{0}\right)P_{-1/2}$
(26)
$\left(\frac{d}{dr_{\ast}}-ik\right)P_{-1/2}=\sqrt{2f}\left(\frac{\lambda}{\sqrt{r^{2}+l^{2}}}+i\mu_{0}\right)P_{+1/2}$
(27)
In order to write Eqs. (26) and (27) in a more compact form, we combine the
solutions as $\psi_{+}=P_{+1/2}+P_{-1/2}$ and $\psi_{-}=P_{+1/2}-P_{-1/2}$ .
After making some computations, we end up with a pair of one-dimensional
Schrödinger like equations:
$\frac{d^{2}\psi_{+}}{dr_{\ast}^{2}}+\left(4k^{2}-V_{+}\right)\psi_{+}=0,$
(28)
$\frac{d^{2}\psi_{-}}{dr_{\ast}^{2}}+\left(4k^{2}-V_{-}\right)\psi_{-}=0,$
(29)
with the following effective potentials:
$V_{\pm}=\frac{2L^{3/2}\sqrt{\Delta}}{D^{2}}\left[2L^{3/2}\sqrt{\Delta}\pm
3\mu_{0}^{2}r\Delta\pm L\left(r-M\right)\mp\frac{\Delta
L}{D}\left(2\mu_{0}^{2}r\left(r^{2}+l^{2}\right)-\frac{\lambda\mu_{0}\left(r-M\right)}{2k}+2rL\right)\right],$
$L\left(r-M\right)\mp\frac{\Delta
L}{D}\left(2\mu_{0}^{2}r\left(r^{2}+l^{2}\right)-\frac{\lambda\mu_{0}\left(r-M\right)}{2k}+2rL\right)],$
(30)
where
$\Delta=r^{2}-2Mr-l^{2},L=\left(\lambda^{2}+\mu_{0}^{2}r^{2}\right),D=L\left(r^{2}+l^{2}\right)-\frac{\lambda\mu_{0}}{4k}.$
(31)
We see that the potentials (30) depend on the NUT parameter $l$. To obtain the
potentials for massless fermions (neutrino), we simply set $\mu_{0}=0$ in Eq.
(30), which yields
$V_{\pm}=\frac{2\lambda\sqrt{\Delta}}{\left(r^{2}+l^{2}\right)^{2}}\left[2\lambda\sqrt{\Delta}\pm\left(r-M\right)\mp\frac{2\Delta
r}{r^{2}+l^{2}}\right].$ (32)
To study the asymptotic behavior of the potentials (30), we can expand it up
to order $O\left(\frac{1}{r^{3}}\right)$. Thus, the potentials behave as
$V_{\pm}\simeq
4\mu_{0}^{2}+8M\mu_{0}^{2}\frac{1}{r}\pm\left(6l^{2}\mu_{0}^{2}-2\lambda^{2}+\frac{\lambda\mu_{0}}{2k}\right)\frac{1}{r^{2}}+O\left(\frac{1}{r^{3}}\right).$
(33)
From Eq. (33), we see that the first term corresponds to the constant value of
the potential at the asymptotic infinity. In the second term, the mass
produces the usual monopole type attractive potential. The NUT parameter in
the third term represents the dipole type potential.
Figure 1: Graph of $V_{+}$ (30) for various values of the NUT parameter $l$.
Here, $\lambda=k=1,\mu=0.12,$ and $M=0.5$.
We will now investigate the effect of the NUT parameter on the effective
potentials (30) by plotting them as a function of radial distance. Figs. 1 and
2 describe the effective potentials (30) for massive particles where we obtain
potential curves for some specific values of the NUT parameter. As can be seen
from Figs. 1 and 2, for sufficiently small values of the NUT parameter, the
potentials have sharp peaks in the physical region. When the NUT parameter
$l=0$, the peak is seen to be maximum. We also observe that while the NUT
parameter increases, the sharpness of the peaks decreases, and the peaks tend
to disappear after a specific value of the NUT parameter. This implies that
for a large values of the NUT parameter, a massive Dirac particle moving in
the physical region experiences a potential barrier without peaks and its
kinetic energy increases. But for a small value of the NUT parameter, it feels
a sharp potential barrier.
Figure 2: Graph of $V_{-}$ (30) for various values of NUT parameter $l$. Here,
$\lambda=k=1,\mu=0.12,$ and $M=0.5$. Figure 3: Three-dimensional plot of
$V_{+}$ for various values of NUT parameter $l$. Here, $\lambda=k=1,\mu=0.12,$
and $M=0.5$.
Explicitly, the influence of the NUT parameter is shown in the three-
dimensional plots of the potentials with respect to the NUT parameter and the
radial distance. As can be seen from Fig. (3), we observe a peak for small
values of the NUT parameter and as the value of the NUT parameter increases,
the potentials decrease. Let us note that, for massless Dirac particle
$\left(\mu_{0}=0\right)$, the potentials have similar behaviors as for the
massive case.
## IV Greybody Radiation from NUT BH
### IV.1 Klein-Gordon equation
The Klein-Gordon equation for the massless scalar field in curved coordinates
is given by isepjc
$\square\Psi=\frac{1}{\sqrt{-g}}\partial_{\mu}\sqrt{-g}g^{\mu\nu}\partial_{\nu}\Psi=0,$
(34)
where $g$ is the determinant of the spacetime metric (1):
$\sqrt{-g}=\left(r^{2}+l^{2}\right)\sin\theta$. To separate the variables in
equation (34), we assume the solutions to the wave equation in the form
$\Psi=R\left(r\right)\Theta\left(\theta\right)\exp\left(-i\omega
t\right)\exp\left(im\phi\right),$ (35)
where $\omega$ denotes the frequency of the wave and $m$ is the azimuthal
quantum number. Therefore, the Klein-Gordon equation (34) becomes
$\left(\frac{\left(r^{2}+l^{2}\right)^{2}}{\Delta}-\frac{4l^{2}\cos^{2}\theta}{\sin^{2}\theta}\right)\omega^{2}+\frac{1}{R\left(r\right)}\frac{d}{dr}\left(\Delta\frac{dR\left(r\right)}{dr}\right)+$
$\frac{1}{\Theta\left(\theta\right)}\frac{1}{\sin\theta}\frac{d}{d\theta}\left(\sin\theta\frac{\partial\Theta\left(\theta\right)}{\partial\theta}\right)-\frac{m^{2}}{\sin^{2}\theta}-\frac{4l\cos\theta}{\sin^{2}\theta}\omega
m=0.$ (36)
The angular part satisfies
$\frac{1}{\sin\theta}\frac{d}{d\theta}\left(\sin\theta\frac{\partial\Theta\left(\theta\right)}{\partial\theta}\right)-\left(\frac{4l^{2}\cos^{2}\theta+4l\omega
m\cos\theta+m^{2}}{\sin^{2}\theta}\right)\Theta\left(\theta\right)=-\lambda\Theta\left(\theta\right),$
(37)
where it has solutions in terms of the oblate spherical harmonic functions
having eigenvalues $\lambda$ 46 . Therefore, the Klein–Gordon equation (36) is
left with the radial part
$\frac{d}{dr}\left(\Delta\frac{d}{dr}R\left(r\right)\right)+\left(\left(r^{2}+l^{2}\right)^{2}\omega^{2}-\Delta\lambda\right)R\left(r\right)=0.$
(38)
By changing the variable in a new form as
$R\left(r\right)=\frac{U\left(r\right)}{\sqrt{r^{2}+l^{2}}},$ the radial wave
equation (38) recasts into a one-dimensional Schrödinger like equation as
follows
$\frac{d^{2}U}{dr_{\ast}^{2}}+\left(\omega^{2}-V_{eff}\right)U=0,$ (39)
in which $r_{\ast}$ is the tortoise coordinate:
$\frac{dr_{\ast}}{dr}=\frac{r^{2}+l^{2}}{\Delta},$ and $V_{eff}$ is the
effective potential given by
$V_{eff}=\frac{\Delta}{\left(r^{2}+l^{2}\right)^{3}}\left(\frac{-3r^{2}\Delta}{r^{2}+l^{2}}-\lambda\left(r^{2}+l^{2}\right)-3r^{2}-4Mr-l^{2}\right).$
(40)
Figure 4: Plots of $V_{eff}$ (40) versus $r$ for the spin-$0$ particles for
various values of NUT parameter $l$. The physical parameters are chosen as;
$\lambda=k=1,\mu=0.12,$ and $M=0.5$.
It is seen from (40) that the effective potential depends on the NUT parameter
$l$. Fig. 4 indicates the behavior of potential (40) under the effect of NUT
parameter for massless particles. As can be seen from Fig. (4), when the NUT
parameter increases, peaks of the potential goes down.
### IV.2 GFs of bosons
It is known that nothing can escape a BH when approaching it. But when a
quantum effect is considered, the BH can radiate. This radiation is known as
the Hawking radiation. We now assume that Hawking radiation is a massless
scalar field (bosons) that satisfies the Klein–Gordon equation. To evaluate
the GF of the NUT metric for bosons, we use 47
$\sigma_{l}\left(\omega\right)\geq\sec h^{2}\left(\int_{-\infty}^{+\infty}\wp
dr_{\ast}\right),$ (41)
in which $r_{\ast}$ is the tortoise coordinate and
$\wp=\frac{1}{2h}\sqrt{\left(\frac{dh\left(r_{\ast}\right)}{dr_{\ast}}\right)^{2}+(\omega^{2}-V_{eff}-h^{2}\left(r_{\ast}\right))^{2}},$
(42)
where $h(r\ast)$ is a positive function satisfying
$h\left(-\infty\right)=h\left(-\infty\right)=\omega$: see Refs. 47 ; 48 for
more details. Without loss of generality, we select $h=\omega$. Therefore,
$\sigma_{l}\left(\omega\right)\geq\sec
h^{2}\left(\int_{r_{h}}^{+\infty}\frac{V_{eff}}{2\omega}dr_{\ast}\right).$
(43)
We use the potential derived in (40) to obtain the GFs for bosons, namely
$\sigma^{s}_{l}\left(\omega\right)\geq\sec
h^{2}\frac{1}{2\omega}\int_{r_{h}}^{+\infty}\frac{1}{\left(r^{2}+l^{2}\right)}\left(\frac{-3r^{2}\Delta}{\left(r^{2}+l^{2}\right)^{2}}+\lambda+\frac{\Delta^{\prime}{r}+\Delta}{r^{2}+l^{2}}\right)dr.$
(44)
After integration, we obtain
$\sigma^{s}_{l}\left(\omega\right)\geq\sec
h^{2}\frac{1}{2\omega}{\left(\frac{-\lambda}{r}-\frac{M}{r^{2}}+\frac{l^{2}\left(\lambda-5\right)}{3r^{3}}+\frac{5Ml^{2}}{2r^{4}}+\frac{l^{4}\left(16-\lambda\right)}{5r^{5}}+\frac{l^{6}\left(\lambda-3\right)}{7r^{7}}\right)}.$
(45)
The behaviors of the obtained bosonic GFs of the NUT BH are depicted in Fig. 5
in which the plots of $\sigma\left(\omega\right)$ for various values of NUT
parameter. As can be seen from those graphs, the NUT has an effect on the
bosonic GFs. Remarkably, $\sigma\left(\omega\right)$ clearly decreases with
the increasing NUT parameter. In other words, the NUT plays a kind of booster
role for the GFs of the spin-0 particles.
Figure 5: Plots of $\sigma^{s}_{l}\left(\omega\right)$ (45) versus $\omega$
for the zero spin particles for various values of NUT parameter $l$. The
physical parameters are chosen as; $\lambda=2$ and $M=1$.
### IV.3 GFs of fermions
Here, we shall derive the fermionic GF of the neutrinos emitted from the NUT
BH. To this end, we consider the potentials (32). Following the procedure
described above [see Eqs. (41)-(43)], we obtain,
$\sigma_{\pm}\left(\omega\right)\geq\sec
h^{2}\frac{\lambda}{\omega}\left(-\frac{2\lambda\mp
1}{r}\mp\left(\frac{M}{r^{2}}+\frac{M^{2}-l^{2}\mp
4l^{2}\lambda}{6r^{3}}\right)\pm\frac{3M^{3}-Ml^{2}}{8r^{4}}\right).$ (46)
The fermionic GFs of the NUT BH are depicted in Figs. 6 and 7 in which the
plots of $\sigma_{\pm}\left(\omega\right)$ for various values of NUT parameter
are obtained. Similar to the bosonic GFs, the NUT has an intensifier effect on
the GFs of the spin-$\frac{1}{2}$ particles.
Figure 6: Plots of $\sigma_{+}\left(\omega\right)$ (45) versus $\omega$ for
the Non-zero spin particles for various values of NUT parameter $l$. The
physical parameters are chosen as; $\lambda=-0.5$ and $M=1$. Figure 7: Plots
of $\sigma_{-}\left(\omega\right)$ (46) versus $\omega$ for the Non-zero spin
particles for various values of NUT parameter $l$. The physical parameters are
chosen as; $\lambda=-0.5$ and $M=1$.
## V Conclusion
We have studied the perturbation and greybody radiation in the NUT spacetime.
The NUT metric describes the vacuum spacetime around a source that is
characterized by two parameters, the mass of the central object and the NUT
parameter. The perturbations studied here are for the bosons and fermions,
namely the Klein-Gordon and Dirac equations are considered. For the Dirac
equation, we explicitly work out the separability of the equations into radial
and angular parts by using a null tetrad in the NP formalism. It is shown that
solutions to the angular equations could be given in terms of the associated
Legendre functions. We also discussed the radial equations and obtain a wave
equation with an effective potential. The radial part involves both mass and
NUT parameters. It is shown in the asymptotic expansion of the potentials (33)
that the mass produces the usual $1/r$ attractive potential while the NUT
parameter manifests itself in the $1/r^{2}$ asymptotically. Thus, the effect
of gravity is as expected stronger than the effect of the NUT parameter in the
NUT geometry. To understand the physical interpretations of the potentials
(30), we make two- and three-dimensional plots of the potentials for different
values of the NUT parameter. It is seen from the plots that the potential
barriers are higher for small values of the NUT parameter. If the NUT
parameter decreases, the peak of the potential barriers distinctly increases.
Potentials become limited regardless of the value of the NUT parameter and
tend to a constant value for large values of distances.
For the GFs of bosons, we have considered the massless scalar wave equation.
To this end, we have reduced the radial part of the Klein-Gordon equation to
the one-dimensional Schrödinger like wave equation. The behaviors of the
effective potential (40) under the effect of the NUT parameter for the scalar
waves are depicted. It is seen from Fig. 4 that as the NUT parameter increases
then the peak of the barrier goes down. In the sequel, we have computed the
GFs, one of the fundamental information that can be obtained from the BHs, for
both bosons and fermions. It has been observed from Figs. 5, 6 and 7 that
increment of the NUT value significantly increases the GF radiation of both
bosons and fermions.
In the cosmology, the NUT parameter measures the amount of anisotropy that a
metric has at large times [50]. Therefore, the GF analyses of the rotating NUT
BHs, like Kerr-NUT-de Sitter BH [50], in the AdS background shall be a
significant future extension of the present work. This may also be important
to understand the AdS/CFT conjecture with the NUT similar to the quasinormal
modes [51] analysis of planar Taub-NUT BHs [52]. This will be investigated in
a near future work.
## References
* (1) J. Plebański and M. Demiański, Ann. Phys. (N.Y.) 98, 98 (1976).
* (2) Newman E T, Tamburino L and Unti T W J, J. Math. Phys. 4 915-23, (1963).
* (3) C. W. Misner, Relativity Theory and Astrophysics ed J Ehlers (Amer. Math. Soc., Providence, Rhode Island) 1, 160 (1967).
* (4) W.B. Bonnor, Proc. Camb. Philos. Soc. 66, 145 (1969).
* (5) V.S. Manko, E. Ruiz, Class. Quantum Gravity 22, 3555 (2005).
* (6) D. Lynden-Bell, M. Nouri-Zonoz, Rev. Mod. Phys. 70, 427 (1998).
* (7) Griffiths. J.B. and Padolsky.J, Class. Quant. Grav., 24, 1687. (2007).
* (8) A. Al-Badawi, M. Halilsoy, Gen. Relativ. Gravit. 38, 1729 (2006).
* (9) S.-Q. Wu, D. Wu, Phys. Rev. D 100, 101501(R) (2019).
* (10) c. Misner, Relativity Theory and Astrophysics. I. ed J Ehlers (Providence, RI: American Mathematical Society), 160 (1967).
* (11) K. D. Kokkotas and B. G. Schmidt, Living Rev. Rel. 2, 2 (1999).[gr-qc/9909058].
* (12) H. P. Nollert, Class. Quant. Grav. 16, R159 (1999).
* (13) B. Mukhopadhyay, S.K. Chakrabarti, Class. Quantum Grav. 16, 3165 (1999).
* (14) J. Jing, Phys. Rev. D 70 065004 (2004); Phys. Rev. D 71 124006 (2005).
* (15) A. Al-Badawi, M.Q. Owaidat, Gen Relativ Gravit 49,110 (2017).
* (16) I.I. Cotaescu, Mod. Phys. Lett. A 22, 2493 (2007).
* (17) W. G. Unruh, Phys. Rev. D 14, 3251 (1976).
* (18) A. Zecca, Nuovo Cimento 30, 13091315 (1998).
* (19) H. Schmid, Math. Nachr. 274275, 117129 (2004).
* (20) D. Page, Phys. Rev. D 14, 1509 (1976).
* (21) E. G. Kalnins and W. Miller, J. Math. Phys. 33, 286 (1992).
* (22) F. Belgiorno, S.L. Cacciatori, J. Phys. A 42, 135207 (2009)
* (23) H. Cebeci, N. Ozdemir, Class. Quantum Grav. 30, 175005 (2013).
* (24) F. Belgiorno, S.L. Cacciatori, J. Math. Phys. 51, 033517 (2010).
* (25) Yan Lyu and Song Cui, Phys. Scr. 79, 025001 (2009).
* (26) D.N. Page, Phys. Rev. D 13, 198 (1976).
* (27) D.N. Page, Phys. Rev. D 14, 3260 (1976).
* (28) R. Jorge, E.S. de Oliveira, J.V. Rocha, Class. Quantum Gravity 32, 065008 (2015).
* (29) J. Abedi, H. Arfaei, Class. Quantum Gravity 31, 195005 (2014).
* (30) S. Fernando, Gen. Relativ. Gravit. 37, 461 (2005).
* (31) W. Kim, J.J. Oh, JKPS 52, 986 (2008).
* (32) T. Harmark, J. Natario, R. Schiappa, Adv. Theor. Math. Phys. 14, 727 (2010).
* (33) K. Jusufi, M. Amir, M. Sabir Ali, S.D. Maharaj, Phys. Rev. D 102, 064020 (2020)
* (34) M.K. Parikh, F. Wilczek, Phys. Rev. Lett. 85, 5049 (2000).
* (35) P. Boonserm, T. Ngampitipan, P.Wongjun, Eur. Phys. J. C 79, 330 (2019).
* (36) I. Sakalli, Phys. Rev. D 94, 084040 (2016). arXiv:1606.00896 [grqc].
* (37) A. Al-Badawi, I. Sakalli, S. Kanzi, Ann. Phys. 412, 168026 (2020).
* (38) A. Al-Badawi, S. Kanzi, I. Sakalli, Eur. Phys. J. Plus 135, 219 (2020).
* (39) H. Gursel, I. Sakalli, Eur. Phys. J. C 80, 234 (2020).
* (40) S. Kanzi, S.H.Mazharimousavi, I. Sakalli, Ann. Phys. 422, 168301 (2020).
* (41) S. Kanzi, İ. Sakallı, Eur. Phys. J. C 81, 501 (2021).
* (42) T. Harmark, J. Natario, R. Schiappa, Adv. Theor. Math. Phys. 14, no.3, 727-794 (2010).
* (43) J. B. Griffiths, J. Podolsky, Exact Space-Times in Einstein’s General Relativity (Cambridge: Cambridge University Press), (2009).
* (44) Jefremov, P.I., Perlick, V.: Circular motion in NUT space-time, Class. Quantum Gravity 33, 179501 (2016).
* (45) M. Nouri-Zonoz, D. Lynden-Bell, Mon. Not. Roy. Astron. Soc. 292, 714 (1998).
* (46) M. Halla, V. Perlick, Gen. Relativ. Gravit. 52, 112 (2020).
* (47) S. Chandrasekhar, The Mathematical Theory of Black Holes Clarendon, London (1983).
* (48) M. Abramowitz, I. A. Stegun (ed.), Handbook of Mathematical Functions (Dover, New York, 1970) and [S. Detweiler, Phys. Rev. D 22, 2323 (1980)].
* (49) I. Sakalli, Eur. Phys. J. C 75, no.4, 144 (2015).
* (50) Y.G. Miao, Z.M. Xu, Phys. Lett. B 772, 542 (2017).
* (51) M. Visser, Phys. Rev. A 59, 427438 (1999). arXiv: quant-ph/9901030.
* (52) A. Anabalón, S. F. Bramberger and J. L. Lehners, JHEP 09, 096 (2019).
* (53) I. Sakalli, Mod. Phys. Lett. A 28, 1350109 (2013).
* (54) P. A. Cano and D. Pereñiguez, [arXiv:2101.10652 [hep-th]].
|
Given a conical affine special Kähler (CASK) manifold together with a compatible mutually local variation of BPS structures, one can construct a quaternionic-Kähler (QK) manifold. We call the resulting QK manifold an instanton corrected c-map space. Our main aim is to study the isometries of a subclass of instanton corrected c-map spaces associated to projective special real (PSR) manifolds with a compatible mutually local variation of BPS structures. We call the latter subclass instanton corrected q-map spaces. In the setting of Calabi-Yau compactifications of type IIB string theory, instanton corrected q-map spaces are related to the hypermultiplet moduli space metric with perturbative corrections, together with worldsheet, D(-1) and D1 instanton corrections. In the physics literature, it has been shown that the hypermultiplet metric with such corrections must have an $\mathrm{SL}(2,\mathbb{Z})$ acting by isometries, related to S-duality. We give a mathematical treatment of this result, specifying under which conditions instanton corrected q-map spaces carry an action by isometries by $\mathrm{SL}(2,\mathbb{Z})$ or some of its subgroups. We further study the universal isometries of instanton corrected q-map spaces, and compare them to the universal isometries of tree-level q-map spaces. Finally, we give an explicit example of a non-trivial instanton corrected q-map space with full $\mathrm{SL}(2,\mathbb{Z})$ acting by isometries and admitting a quotient of finite volume by a discrete group of isometries.
§ INTRODUCTION
The supergravity c-map assigns to a projective special Kähler (PSK) manifold $(\overline{M},g_{\overline{M}})$ of complex dimension $n$, a quaternionic-Kähler (QK) manifold $(\overline{N},g_{\overline{N}})$ of real dimension $4n+4$. In the context of Calabi-Yau compactifications of type IIA/B string theory on a Calabi-Yau threefold $X$, the c-map takes the vector multiplet moduli space $\mathcal{M}_{\text{VM}}^{\text{IIA/B}}(X)$ to the hypermultiplet moduli space $\mathcal{M}_{\text{HM}}^{\text{IIB/A}}(X)$ with its string-tree-level metric $g_{\text{FS}}$, also known as the Ferrara-Sabharwal metric [22, 32]. Such a construction receives quantum corrections in the string coupling $g_s$ of several kinds:
* Perturbative corrections: these produce the so-called 1-loop corrected Ferrara-Sabharwal metric on $\mathcal{M}_{\text{HM}}^{\text{IIA/B}}(X)$ [38, 7]. In a purely mathematical setting, this construction can be understood as a way of assigning to a PSK manifold $(\overline{M},g_{\overline{N}})$ a 1-parameter family of QK manifolds $\{(\overline{N}_{c_{\ell}},g_{\overline{N},c_{\ell}})\}_{c_{\ell}\in \mathbb{R}}$ [6], where $c_{\ell}\in \mathbb{R}$ corresponds to the 1-loop correction.
* Non-perturbative quantum corrections: these are divided in D-instanton and NS5-instanton corrections. These have been extensively studied in the physics literature via twistor methods, see for example the reviews [8, 10] and the references therein. The inclusion of all D-instanton corrections is understood in the physics literature [12, 13], while the NS5-instanton corrections are still work in progress (see for example [1]).
When restricted to the simpler setting of mutually local D-instanton corrections, a fairly explicit local formula for the QK metric was given in the physics literature via twistor methods in [2][Here by “fairly explicit" we mean that the expression is explicit except for a certain function $\mathcal{R}$, which is implicitly defined in terms of the field variables.]. On the other hand, in [25] a mathematical treatment (based on the geometric approach of [6]) of a class of QK metrics related to the above mutually local case was given. Namely, if $(\overline{M},g_{\overline{M}})$ is a PSK manifold and $(M,g_M,\omega_M,\nabla,\xi)$ the associated conical affine special Kähler (CASK) manifold, then one can complement this data with a mutually local variation of BPS structures $(M,\Gamma,Z,\Omega)$ to construct a new QK metric $(\overline{N},g_{\overline{N}})$ (we suppress the choice of $1$-loop parameter $c_{\ell}$ from the notation). The general notion of variation of BPS structure can be found in [17] (see also Definition <ref> below for the specific case to be used throughout this paper). Here $(M,\Gamma,Z,\Omega)$ is assumed to satisfy certain compatibility conditions with the CASK structure $(M,g_M,\omega_M,\nabla,\xi)$ (see Section <ref> below), and encodes the analog of “mutually local D-instanton" corrections when compared to the string theory setting.
On the other hand, type IIB string theory carries an $\mathrm{SL}(2,\mathbb{Z})$-symmetry called S-duality. In the physics literature, it has been shown that S-duality descends to an action by isometries on $\mathcal{M}_{\text{HM}}^{\text{IIB}}(X)$ when one includes the appropriate amount of quantum corrections for the metric. For example:
* When one drops all corrections in the string coupling $g_s$ and takes a large volume limit, one obtains the classical QK metric on $\mathcal{M}_{\text{HM}}^{\text{IIB}}(X)$. This metric has been shown to have an $\mathrm{SL}(2,\mathbb{R})$ acting by isometries [16], which comes from the $\mathrm{SL}(2,\mathbb{Z})$ S-duality action. Furthermore, it has been shown in [16] that the $\mathrm{SL}(2,\mathbb{R})$ (and even $\mathrm{SL}(2,\mathbb{Z})\subset \mathrm{SL}(2,\mathbb{R}))$ is broken when one includes world-sheet instanton corrections and/or the 1-loop correction in $g_s$.
* On the other hand, it has been shown in [15, 11] that when one includes world-sheet instanton corrections, the 1-loop correction, and D(-1) and D1 instanton corrections, one recovers again an $\mathrm{SL}(2,\mathbb{Z})$ action by isometries coming from the S-duality action. As a consequence of their result, it also follows that only including perturbative world-sheet instanton corrections, the 1-loop correction in $g_s$ and D(-1) instanton corrections also preserves the isometric S-duality action. For some other extensions of this result in the physics literature see for example [11, 1, 3, 4].
* The QK metric of $\mathcal{M}_{\text{HM}}^{\text{IIB}}(X)$ is also expected to retain the $\mathrm{SL}(2,\mathbb{Z})$ S-duality action by isometries when one includes all quantum corrections, but such a metric has not been constructed or understood (as far as the authors know) in the physics literature.
The classical metric obtained in the case when one drops all corrections in $g_s$ and take a large volume limit lies in a subset of c-map metrics called q-map metrics [27]. Mathematically, the q-map assign to a $n-1\geq 0$ dimensional projective special real (PSR) manifold a $4n+4$ dimensional QK manifold [23, 18]. A purely differential geometric proof that q-map metrics carry an $\mathrm{SL}(2,\mathbb{R})$ of isometries was given in [26], while more traditional supergravity arguments can be found for example in [29, 28]. On the other hand, including world-sheet instanton corrections and the 1-loop correction takes the q-map metric to the class of 1-loop corrected c-map metrics; while including D(-1) and D1 instanton corrections takes the 1-loop corrected c-map metric to the class of mutually local instanton corrected QK metrics studied in [2, 25].
Our main objective in this paper is to do a mathematical treatment of the S-duality results in [15, 11], and to study the universal isometries of instanton corrected q-map spaces, i.e. those isometries that are independent of the initial PSR manifold and the form of the quantum corrections that we restrict ourselves to. Namely, among the class of instanton corrected c-map metrics $(\overline{N},g_{\overline{N}})$ constructed in [25] we restrict to a subclass, which we call instanton corrected q-map metrics, and show under which conditions they carry an $\mathrm{SL}(2,\mathbb{Z})$-action (or an action by some of its subgroups) by isometries. This $\mathrm{SL}(2,\mathbb{Z})$-action is furthermore related to the S-duality symmetry in the string theory setting. Furthermore, we study how the universal group of isometries of q-map spaces described in [26] gets modified for instanton corrected q-map spaces (see Section <ref> below for more details).
§.§ Summary of main results
The main differences and new results compared to the works in the physics literature [15, 11], from which we take a lot of inspiration from, are the following:
* If the S-duality $\mathrm{SL}(2,\mathbb{Z})$ action defined in (<ref>) restricts to an action on the domain of definition of the instanton corrected q-map space, then the twistor space argument of [15, 11] follows and the action is also by isometries. However, verifying that the domain of definition actually carries such an action seems non-trivial. In Theorem <ref> (explained in more detail below) we collect results regarding this point. In particular we find that, even in a case where we do not have an $\mathrm{SL}(2,\mathbb{Z})$-invariant domain of definition, one can always find either an $S\in \mathrm{SL}(2,\mathbb{Z})$ or $T\in \mathrm{SL}(2,\mathbb{Z})$-invariant one (where $T$ and $S$ are the usual generators of $\mathrm{SL}(2,\mathbb{Z})$ given in (<ref>)). Furthermore, in Section <ref> we give an explicit non-trivial (albeit simple) example where one can find such an $\mathrm{SL}(2,\mathbb{Z})$-invariant neighborhood of definition.
* Assuming that the domain of definition of the instanton corrected q-map space carries an action by the S-duality $\mathrm{SL}(2,\mathbb{Z})$, we show that this action must be by isometries by following an argument similar but slightly different from [15]. Using their twistor description of the relevant QK manifolds, it is shown in [15] that certain “type IIA" Darboux coordinates for the contact structure of the twistor space can be Poisson resummed, and then it is shown that the later can be related via a gauge transformation to certain “type IIB" Darboux coordinates. The type IIB coordinates then make transparent that a certain lift of the $\mathrm{SL}(2,\mathbb{Z})$ action to the twistor space acts by twistor space automorphisms, and hence that $\mathrm{SL}(2,\mathbb{Z})$ acts by isometries on the QK metric. In our work, we find it simpler to do a direct Poisson resummation of the contact structure corresponding to the QK metric constructed in [25], and then use the resulting expression to show that the “type IIB" coordinates from [15] are Darboux coordinates for the contact structure.
* We further study certain universal isometry groups of instanton corrected q-map spaces, and compare with with what happens in the case where no quantum corrections are considered (see Section <ref> and Theorem <ref>). In particular, while S-duality is a universal isometry for tree-level q-map spaces, we are not able to guarantee that the the same is true for instanton corrected q-map spaces (see Remark <ref>). Furthermore, in the example of Section <ref> we use the $\mathrm{SL}(2,\mathbb{Z})$ action by isometries together with the universal isometry group to show that our example admits a quotient of finite volume by a discrete group of isometries.
In the following, we explain in more detail the setting and the aforementioned results. As previously mentioned, our main results concerns a class of QK metrics called instanton corrected q-map spaces, defined in Section <ref>. Roughly speaking, these are QK manifolds associated to a CASK manifold described by a holomorphic prepotential $(M,\mathfrak{F})$, together with a variation of mutually local BPS structures $(M,\Gamma,Z,\Omega)$ (see [17] for the general definition of variation of BPS structures) of the following form:
* The holomorphic prepotential $\mathfrak{F}:M\subset \mathbb{C}^{n+1}\to \mathbb{C}$ has the form
\begin{equation}
\mathfrak{F}(Z^i)=-\frac{1}{6}k_{abc}\frac{Z^aZ^bZ^c}{Z^0}+\chi\frac{(Z^0)^2\zeta(3)}{2(2\pi \I )^3}-\frac{(Z^0)^2}{(2\pi \I)^3}\sum_{\hat{\gamma}=q_a\gamma^a \in \Lambda^+}n_{\hat{\gamma}}\mathrm{Li}_3(e^{2\pi \I q_aZ^a/Z^0}),
\end{equation}
where $i=0,...,n$; $a,b,c=1,...,n$; $k_{abc}\in \mathbb{R}$ are symmetric in the indices, $\chi\in \mathbb{Z}$, $n_{\hat{\gamma}}\in \mathbb{Z}$, $\mathrm{Li}_n(x)$ denote the n-th polylogarithms, $\zeta(x)$ is the Riemann zeta function, and $\Lambda^{+}:=\text{span}_{\mathbb{Z}_{\geq 0}}\{\gamma^a\}_{a=1}^n-\{0\}$ is a commutative semigroup freely generated by $n$ elements $\{\gamma^a\}_{a=1}^n$.
This choice of $\mathfrak{F}$ is motivated by Calabi-Yau compactifications of string theory. Namely, if $X$ denotes a Calabi-Yau threefold, and if $k_{abc}$ are taken to be the triple intersection numbers of $X$, $\chi=\chi(X)$ the Euler characteristic, and for $\hat{\gamma}\in H_2^{+}(X,\mathbb{Z})$ we have that $n_{\hat{\gamma}}=n_{\hat{\gamma}}^{(0)}$ are the genus zero Gopakumar-Vafa invariants, then $\mathfrak{F}$ denotes the prepotential specifying the PSK geometry of $\mathcal{M}_{\text{VM}}^{\text{IIA}}(X)$ with all worldsheet corrections. Applying the 1-loop corrected c-map, one obtains from $\mathcal{M}_{\text{VM}}^{\text{IIA}}(X)$ the $1$-loop corrected metric on $\mathcal{M}_{\text{HM}}^{\text{IIB}}(X)$.
* The charge lattice $\Gamma$ and the central charge $Z$ of $(M,\Gamma,Z,\Omega)$ are canonically determined by $\mathfrak{F}$ (see Section <ref>), while the BPS indices $\Omega(\gamma)$ are also determined by $\mathfrak{F}$ as follows: with respect to a canonical Darboux basis $(\widetilde{\gamma}_i,\gamma^i)_{i=0}^n$ of $\Gamma$ with respect to its symplectic pairing, we have
\begin{equation}\label{BPSindintro}
\begin{cases}
\Omega(q_0\gamma^0)=-\chi, \quad q_0\in \mathbb{Z}-\{0\}\\
\Omega(q_0\gamma^0\pm q_a\gamma^a)=\Omega(\pm q_a\gamma^a)=n_{q_a\gamma^a} \quad \text{for $q_a\gamma^a \in \Lambda^+$, $q_0\in \mathbb{Z}$}\\
\Omega(\gamma)=0 \quad \text{else}.\\
\end{cases}
\end{equation}
The prescription (<ref>) has previously appeared in the physics literature (see for example <cit.>), and is the data required to add the D(-1) and D1 instanton corrections to $\mathcal{M}_{\text{HM}}^{\text{IIB}}(X)$ in an “S-duality compatible"-way. Furthermore, in the case of a non-compact Calabi-Yau threefold $X$ without compact divisors the same structure for the BPS indices is expected. Indeed, in [34] the appropriate $\Omega(\gamma)$ determining generalized Donaldson-Thomas invariants are constructed. It is then shown that $\Omega(q_0\gamma^0)=-\chi(X)$ <cit.>, and conjectured that $\Omega(q_0\gamma^0\pm \hat{\gamma})=n_{\hat{\gamma}}^{(0)}$<cit.>.
To the above data $(M,\mathfrak{F})$ and $(M,\Gamma,Z,\Omega)$ we can apply the construction of [25] and obtain a QK manifold $(\overline{N},g_{\overline{N}})$, which we call an instanton corrected q-map space. $(\overline{N},g_{\overline{N}})$ depends on a choice of projective special real (PSR) manifold $(\mathcal{H},g_{\mathcal{H}})$ (determining the first term in $\mathfrak{F}$), the choice of $\chi\in \mathbb{Z}$ and $n_{\hat{\gamma}} \in \mathbb{Z}$, and the choice of 1-loop parameter $c_{\ell}\in \mathbb{R}$ (see Section <ref>). Our main results concern the isometries of a lift $(\widetilde{N},g_{\overline{N}})\to (\overline{N},g_{\overline{N}})$ on which we have no periodic directions (see Definition <ref> for a more precise statement). In order to state the main results, we consider the following subgroups of the Heisenberg group $\text{Heis}_{2n+3}(\mathbb{R})$ (endowed with standard global coordinates $(\eta^i,\widetilde{\eta}_i,\kappa )$, $i=0,\ldots, n$):
\begin{equation}
\begin{split}
H_{2n+2}&:=\{ (\eta^i,\widetilde{\eta}_i,\kappa)\in \text{Heis}_{2n+3}(\mathbb{R}) \; |\quad \; \eta^0=0 \}\\
H_{2n+2,D}&:=\{ (\eta^i,\widetilde{\eta}_i,\kappa)\in H_{2n+2} \; |\quad \; \text{$\eta^a \in \mathbb{Z}$ for $a=1,...,n$}\}\,.
\end{split}
\end{equation}
The following theorem collects our main results:
Theorem <ref>: Consider an instanton corrected q-map space $(\widetilde{N},g_{\overline{N}})$ of dimension $4n+4$ as defined in Section <ref> (in particular $\widetilde{N}$ here is the maximal domain of definition). Furthermore, let $T,S\in \mathrm{SL}(2,\mathbb{Z})$ be as in (<ref>), where $\mathrm{SL}(2,\mathbb{Z})$ acts on the ambient manifold $\overline{\mathcal{N}}_{\text{IIB}}^{\text{cl}}\supset \overline{\mathcal{N}}_{\text{IIB}}\cong \overline{\mathcal{N}}_{\text{IIA}}\supset \widetilde{N}$ as described in (<ref>). Then:
* $(\widetilde{N},g_{\overline{N}})$ has a group acting by isometries of the form
\begin{equation}\label{breakisointroduction}
\langle T \rangle \ltimes (\mathbb{Z}^n \ltimes H_{2n+2,D})
\end{equation}
where $\langle T \rangle \cong \mathbb{Z}$ denotes the subgroup of $\mathrm{SL}(2,\mathbb{Z})$ generated by $T$.
* Assume that we take the one-loop parameter to be $c_{\ell}=\frac{\chi}{192\pi}$. Then we can always find a non-empty open subset $\widetilde{N}_S\subset \widetilde{N}$ where $(\widetilde{N}_S,g_{\overline{N}})$ has a group acting by isometries of the form
\begin{equation}
\langle S \rangle \ltimes (\mathbb{Z}^n \ltimes H_{2n+2,D}),
\end{equation}
where $\langle S \rangle \cong \mathbb{Z}/4\mathbb{Z}$ is the subgroup generated by $S$. Furthermore, if $\widetilde{N}_{\mathrm{SL}(2,\mathbb{Z})}\subset \widetilde{N}$
is an open subset, which is $\mathrm{SL}(2,\mathbb{Z})$-invariant under the S-duality action (<ref>), then $\mathrm{SL}(2,\mathbb{Z})$ acts by isometries on $(\widetilde{N}_{\mathrm{SL}(2,\mathbb{Z})},g_{\overline{N}})$. In particular, if $\widetilde{N}$ is already invariant under $\mathrm{SL}(2,\mathbb{Z})$ then (<ref>) can be enhanced to
\begin{equation}
\mathrm{SL}(2,\mathbb{Z})\ltimes (\mathbb{Z}^n\ltimes H_{2n+2,D})\,.
\end{equation}
* Finally, if $n_{\hat{\gamma}}=0$ for all $\hat{\gamma}\in \Lambda^{+}$, then in the previous statements we can replace $\mathbb{Z}^n$ and $H_{2n+2,D}$ by $\mathbb{R}^n$ and $H_{2n+2}$.
If furthermore we take $\chi=c_{\ell}=0$ and $n_{\hat{\gamma}}=0$ for all $\hat{\gamma}\in \Lambda^{+}$, then we return to the tree-level q-map space case, where there is a connected $3n+6$ dimensional Lie group $G$ acting by isometries on $(\widetilde{N},g_{\overline{N}})$, see <cit.>. The group $G$ in particular contains the S-duality action by $\mathrm{SL}(2,\mathbb{R})$, an action by $\mathbb{R}^{n}\ltimes H_{2n+2}$, and a dilation action by $\mathbb{R}_{>0}$.
* Note that from Theorem <ref> one finds that
$\langle T \rangle \ltimes (\mathbb{Z}^n \ltimes H_{2n+2,D})$ is a universal group of isometries, in the sense that it is always an isometry group for any instanton corrected q-map space (provided one takes the maximal domain of definition $\widetilde{N}$ of the metric $g_{\overline{N}}$). On the other hand, even in the case of $c_{\ell}=\frac{\chi}{192\pi}$, Theorem <ref> does not guarantee in general that $\mathrm{SL}(2,\mathbb{Z})$ is an isometry group for $(\widetilde{N},g_{\overline{N}})$, but rather one must first check that $\widetilde{N}$ (or an open subset) carries an action by S-duality. In particular, Theorem <ref> does not let us conclude that $\mathrm{SL}(2,\mathbb{Z})$ is a universal group of isometries. This should be contrasted to the tree-level q-map space case, where $\mathrm{SL}(2,\mathbb{R})$ is known to always act by isometries.
* We remark that the action of $S\in \mathrm{SL}(2,\mathbb{Z})$ is perhaps the most interesting and non-trivial within $\mathrm{SL}(2,\mathbb{Z})$, and corresponds to interchanging weak coupling and strong coupling in the type IIB string theory setting. On the other hand, the action by $T\in \mathrm{SL}(2,\mathbb{Z})$ generates the discrete Heisenberg isometry that is missing from $H_{2n+2,D}$.
In Section <ref> we give an explicit example where we can achieve full $\mathrm{SL}(2,\mathbb{Z})$ acting by isometries. More precisely, we consider the case where $\mathfrak{F}$ is given simply by
\begin{equation}\label{exint}
\mathfrak{F}=-\frac{1}{6}\frac{(Z^1)^3}{Z^0}+\chi \frac{(Z^0)^2 \zeta(3)}{2(2\pi \mathrm{i})^3}, \quad \chi>0\,,
\end{equation}
with variation of BPS structures having BPS indices of the form
\begin{equation}\label{exint2}
\Omega(\gamma)=\begin{cases}
\Omega(q_0\gamma^0)=-\chi, \quad q_0\neq 0 \\
\Omega(\gamma)=0 \quad \text{else}.\\
\end{cases}
\end{equation}
From this data one obtains an 8-dimensional instanton corrected q-map space. It satisfies the following:
Corollary <ref> and <ref>: let $\widetilde{N}$ be defined by (<ref>) and take $c_{\ell}=\frac{\chi}{192\pi}$. Then the instanton corrected q-map metric $g_{\overline{N}}$ associated to (<ref>) and (<ref>) is defined and positive definite on $\widetilde{N}$. Furthermore, it carries an effective action by isometries by a group of the form $\mathrm{SL}(2,\mathbb{Z})\ltimes(\mathbb{R}\ltimes H_4)$.
Theorem <ref>: Let $(\widetilde{N},g_{\overline{N}})$ be as in Corollary <ref>. Then:
* There is a free and properly discontinuous action by isometries of a discrete group of the form $\mathrm{SL}(2,\mathbb{Z})'\ltimes \Lambda$, where $\Lambda \subset \mathbb{R}\ltimes H_4$ is a lattice, $\mathrm{SL}(2,\mathbb{Z})' \subset \mathrm{SL}(2,\mathbb{Z})$ is a finite index subgroup and the QK manifold $(\widetilde{N}/(\mathrm{SL}(2,\mathbb{Z})'\ltimes \Lambda),g_{\overline{N}})$ has finite volume.
* Furthermore, there is a submanifold with boundary $\hat{N}\subset \widetilde{N}$ where $\mathrm{SL}(2,\mathbb{Z})'\ltimes \Lambda$ acts and the quotient $(\hat{N}/(\mathrm{SL}(2,\mathbb{Z})'\ltimes \Lambda),g_{\overline{N}})$ gives a complete QK manifold with boundary[Recall that a Riemannian manifold with boundary is complete if it is complete as a metric space with the induced distance function.] and of finite volume. The manifold with boundary $\hat{N}$ is of the form $\hat{N}=\widetilde{N}'\cup \partial \hat{N}$, where $\widetilde{N}'$ is defined in the second point of Remark <ref>.
We note that the example $(\widetilde{N},g_{\overline{N}})$ of Corollary <ref> is incomplete (see Remark <ref>). We do not know if the metric and the $\mathrm{SL}(2,\mathbb{Z})$-action can be extended to a complete manifold (without boundary).
At the end of Remark <ref> we comment about some expectations of a related example associated to the resolved conifold.
Finally, we make a short comment related to the swampland program in physics. Among the geometric properties expected from the moduli space $\mathcal{M}$ of a low energy effective theory consistent with quantum gravity, are that $\mathcal{M}$ should be non-compact, have finite volume, and be geodesically complete (see [37] and <cit.>). In particular, applied to type IIA/B string theory, they imply that $\mathcal{M}_{\text{HM}}^{\text{IIA/B}}(X)$ must be a non-compact complete QK manifold of finite volume (after including all quantum corrections). On the other hand, the example from Theorem <ref> produces a non-compact QK manifold of finite volume, with “partial completeness" in the sense that it has a complete end, and a boundary where the metric is geodesically incomplete. It would be interesting to see if a suitable extension of the example of Theorem <ref> would produce a QK manifold with the required geometric properties expected by the swampland conjectures.
§.§ Organization of topics
The topics are organized as follows:
* In Section <ref> we review the construction of instanton corrected c-map metrics from [25], and also discuss their twistor description. In particular, the instanton corrected c-map spaces from [25] are in the image of the HK/QK correspondence, and we want to recall a description of the QK twistor space in terms of the HK data, as done in <cit.>.
* In Section <ref> we start by specifying the class of instanton corrected q-map spaces within the class of instanton corrected c-map metrics. Following closely the work in the physics literature of [15, 11], we study when an instanton corrected q-map space carries an $\mathrm{SL}(2,\mathbb{Z})$-action by isometries, or at least an action by some of its the subgroups, like $\langle S \rangle \subset \mathrm{SL}(2,\mathbb{Z})$.
* In Section <ref> we study certain universal isometries of instanton corrected q-map spaces and how they are related to the $\mathrm{SL}(2,\mathbb{Z})$ S-duality symmetry. We collect the main results from Section <ref> and Section <ref> in Theorem <ref>.
* In Section <ref>, we give an explicit example of an instanton corrected q-map space where the full $S$-duality acting by isometries is realized. Furthermore, we show that it admits a quotient of finite volume by a discrete group of isometries.
* Finally, in Appendix <ref> we collect some useful integral identities involving Bessel functions, and in Appendix <ref> we include a rather long computation that is not really needed for the main points of the paper, but we include for completeness.
Acknowledgements: this work was supported by the Deutsche Forschungsgemeinschaft (German Research Foundation) under Germany’s Excellence Strategy – EXC 2121 “Quantum Universe” – 390833306. As with our previous related joint works [25, 26], the idea for this work originated from discussions within our Swampland seminar, which is part of the Cluster of Excellence Quantum Universe. The authors would like to thank Murad Alim, Jörg Teschner and Timo Weigand for their contributions in the aforementioned discussions.
§ INSTANTON CORRECTED QK METRICS AND THEIR TWISTOR DESCRIPTION
The main aims of this section are the following:
* On one hand, we want to recall the main results of [25], concerning the construction of QK metrics associated to certain CASK manifolds with mutually local variation of BPS structures. In the setting of Calabi-Yau compactifications of string theory, these metrics are related to the type IIA hypermultiplet metric with mutually local D-instanton corrections, studied in the physics literature in [2].
* On the other hand, we want to recall certain general facts about the twistor space of QK metrics in the image of the HK/QK correspondence. This part will be mostly based on <cit.>. In particular, the QK metrics from the previous point lie in the image of the HK/QK correspondence, and we want to write down an explicit expression for the holomorphic contact structure of its twistor space in terms of the HK data (see (<ref>) and (<ref>)). These formulas will be used throughout the rest of this work, and in particular in Section <ref>, where we study the isometries of instanton corrected q-map spaces.
* Finally, we write down certain “type IIA" Darboux coordinates for the holomorphic contact structure, see (<ref>). These have been previously written down in the physics literature [12], under slightly different conventions. Using the explicit formula for the contact structure obtained in the previous point, we will give in the Appendix <ref> a direct proof of the fact that they are Darboux coordinates. This particular result is not needed for Section <ref>, where certain "type IIB" Darboux coordinates are found, but we include it for completeness.
§.§ QK metrics associated to CASK manifolds with mutually local variations of BPS structures
We briefly recall the main ingredients in the construction of [25].
An integral conical affine special Kähler (CASK) manifold is a tuple $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ where:
* $(M,g_M,\omega_M,\nabla)$ is an affine special Kähler (ASK) manifold. Namely, $(M,g_M,\omega_M)$ is pseudo-Kähler, with the complex structure $J$ determined by the metric $g_M$ and the Kähler form $\omega_M$ by $g_M(J-,-)=\omega_M(-,-)$; $\nabla$ is a torsion-free flat connection with $\nabla \omega_M=0$; and if $d_{\nabla}:\Omega^k(M,TM)\to \Omega^{k+1}(M,TM)$ denotes the extension of $\nabla:\Omega^0(M,TM)\to \Omega^{1}(M,TM)$ to higher degree forms, then $d_{\nabla}J=0$, where we think of $J$ as an element of $\Omega^1(M,TM)$.
* $\Gamma\subset TM$ is a sub-bundle of $\nabla$-flat lattices with $\Gamma\otimes_{\mathbb{Z}}\mathbb{R}=TM$. Around any $p\in M$, we can find a local trivialization $(\widetilde{\gamma}_i,\gamma^i)$ of $\Gamma$ of Darboux frames with respect to $\omega_M$. We denote $\langle -,- \rangle:=\omega_M(-,-)|_{\Gamma\times \Gamma}$, and our conventions are that $\langle \widetilde{\gamma}_i,\gamma^j\rangle=\delta_i^j$.
* $\xi$ is a vector field on $M$ such that $\nabla \xi=D\xi=\text{Id}_{TM}$, where $D$ denotes the Levi-Civita connection of $g_M$, and $\text{Id}_{TM}$ is the identity endomorphism of $TM$. Furthermore, we assume that $g_M$ is positive definite on $\mathcal{D}:=\text{span}\{\xi,J\xi\}$ and negative definite on $\mathcal{D}^{\perp}$.
On the other hand, the data corresponding to the mutually local instanton corrections in the string theory setting was specified in terms of the notion of a mutually local variation of BPS structures (see for example [17] for the more general notion of a variation of BPS structures).
A variation of mutually-local BPS structures over the complex manifold $M'$ is a tuple $(M',\Gamma',Z,\Omega)$ where
* $\Gamma' \to M'$ is a local system of lattices with a skew-pairing $\langle - , - \rangle:\Gamma' \times \Gamma' \to \mathbb{Z}$.
* $Z$ is a holomorphic section of $(\Gamma')^*\otimes \mathbb{C}\to M'$, where $(\Gamma')^*$ denotes the dual local system of $\Gamma'$. If $\gamma$ is a local section of $\Gamma'$, then we denote by $Z_{\gamma}:=Z(\gamma)$ the corresponding local holomorphic function on $M'$.
* $\Omega: \Gamma' -\{0\}\to \mathbb{Z}$ is a function of sets satisfying $\Omega(\gamma)=\Omega(-\gamma)$ and the following properties
* Mutual-locality: if we define $\text{Supp}(\Omega):=\{\gamma \in \Gamma' -\{0\}\;\; | \;\; \Omega(\gamma)\neq 0\}$, then $\gamma_1,\gamma_2\in \Gamma'_p\cap \text{Supp}(\Omega)$ implies that $\langle \gamma_1,\gamma_2\rangle=0$.
* Support property: for any compact set $K\subset M'$ and a choice of covariantly constant norm $|\cdot |$ on $\Gamma'|_{K}\otimes_{\mathbb{Z}}\mathbb{R}$, there is a constant $C>0$ such that for all $\gamma \in \Gamma'|_{K}\cap \text{Supp}(\Omega)$
\begin{equation}
\end{equation}
* Convergence property: for any $R>0$, the series
\begin{equation}\label{convpropBPS}
\sum_{\gamma\in \Gamma'|_p}\Omega(\gamma)e^{-R|Z_{\gamma}|}
\end{equation}
converges normally over compact subsets of $M'$.
* The numbers $\Omega(\gamma)$, called BPS indices, are monodromy invariant. Namely if $\gamma$ has monodromy $\gamma \to A\gamma$ around a loop, then $\Omega(\gamma)=\Omega(A\gamma)$.
Given an integral CASK manifold $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$, we will only consider mutually local variations of BPS structures $(M',\Gamma',Z,\Omega)$ where $(M,\Gamma)=(M',\Gamma')$, $\langle -, - \rangle=\omega_M(-,-)|_{\Gamma\times \Gamma}$, and where $Z$ is the canonical central charge associated to the integral CASK manifold <cit.>. The later is determined as follows: if $\xi^{1,0}=\frac{1}{2}(\xi-\I J\xi)$, then
\begin{equation}
\end{equation}
In particular, given a local Darboux frame $(\widetilde{\gamma}_i,\gamma^i)$ of $\Gamma$, the locally defined functions $\{Z_{\widetilde{\gamma}_i}\}_{i=0}^{n}$ and $\{Z_{\gamma^i}\}_{i=0}^{n}$ give conjugate systems of holomorphic special coordinates for the CASK geometry, where $n+1=\text{dim}_{\mathbb{C}}(M)$.
§.§.§ Associated instanton corrected HK manifold
To the data $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ one can associate an “instanton-corrected" hyperkähler (HK) geometry <cit.>. This HK manifold can be thought as a deformation of the canonical HK manifold associated to $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ via the rigid c-map (also known as the associated “semi-flat" HK manifold) [22, 24, 31, 5]. In the physics literature, see [33], such instanton corrected HK metrics were studied in the context of $S^1$-compactifications of 4d $\mathcal{N}=2$ SUSY gauge theories. There the description of the HK geometry is in terms of its associated twistor space, which in turn in described in terms of a twistor family of holomorphic Darboux coordinates satisfying “TBA"-like integral equations.
In order to describe the instanton corrected HK manifold, we first let $N:=T^*M/\Gamma^*$. This can be canonically identified with
\begin{equation}
N\cong \{\zeta: \Gamma \to \mathbb{R}/ \mathbb{Z} \; | \; \zeta_{\gamma+\gamma'}=\zeta_{\gamma}+\zeta_{\gamma'}\}\,.
\end{equation}
In particular, slightly abusing notation and denoting by $\zeta$ the evaluation map on $N$, and given a local Darboux frame $(\widetilde{\gamma}_i,\gamma^i)$ of $\Gamma$, we obtain local coordinates on $N$ by $(Z_{\gamma^i},\zeta_{\widetilde{\gamma}_i},\zeta_{\gamma^i})$ (or $(Z_{\widetilde{\gamma}_i},\zeta_{\widetilde{\gamma}_i},\zeta_{\gamma^i})$). Note that $Z_{\gamma^i}$ and $Z_{\widetilde{\gamma_i}}$ are (pull-backs of local) holomorphic functions on the base manifold $M$ while $\zeta_{\widetilde{\gamma}_i}$ and $\zeta_{\gamma^i}$ are “fiber coordinates" taking values in the circle.
In the following, we will also denote by $\langle -, -\rangle$ the pairing on $\Gamma^*$ induced by the isomorphism $\gamma \mapsto \langle \gamma, - \rangle$. With this definition, the dual of a Darboux basis of $\Gamma$ is a Darboux basis of $\Gamma^*$. We will also denote by $\langle -, -\rangle$ the $\mathbb{C}$-linear extension of the pairing to $\Gamma^*\otimes \mathbb{C}$.
Finally, if $K_i: \mathbb{R}_{>0}\to \mathbb{R}$ denotes the $i$-th modified Bessel function of the second kind, and $\gamma$ is a local section of $\Gamma$ with $\gamma \in \text{Supp}(\Omega)$, we define the following local function (resp. 1-form) on $N$:
\begin{equation}\label{VAinst}
V_{\gamma}^{\text{inst}}:=\frac{1}{2\pi}\sum_{n>0}e^{2\pi\I n\zeta_{\gamma}}K_0(2\pi n|Z_{\gamma}|), \quad A_{\gamma}^{\text{inst}}:=-\frac{1}{4\pi}\sum_{n>0}e^{2\pi\I n\zeta_{\gamma}}|Z_{\gamma}|K_1(2\pi n|Z_{\gamma}|)\Big( \frac{\mathrm{d}Z_{\gamma}}{Z_{\gamma}}-\frac{\mathrm{d}\overline{Z}_{\gamma}}{\overline{Z}_{\gamma}}\Big)\,.
\end{equation}
Due to the convergence property and support property of variations of BPS structures, these expressions are well-defined local smooth functions (resp. $1$-forms) on $N$ (see <cit.>).
Finally, we will need the following compatibility condition between the data $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$:
Let $\pi:N\to M$ be the canonical projection. We will say that $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ are compatible if the tensor
\begin{equation}\label{non-deg}
T:=\pi^*g_{M}- \sum_{\gamma}\Omega(\gamma)V_{\gamma}^{\text{inst}}\pi^*|\mathrm{d}Z_{\gamma}|^2
\end{equation}
on $N$ is horizontally non-degenerate.
We then have the following:
<cit.> Let $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ be as before. Furthermore, let $\omega_i\in \Omega^2(N)$ for $i=1,2,3$ be defined by
\begin{equation} \label{holsym}
\omega_1+\mathrm{i}\omega_2:=-2\pi\left(\langle \mathrm{d}Z\wedge \mathrm{d}\zeta \rangle + \sum_{\gamma}\Omega(\gamma)\left(\mathrm{d}Z_{\gamma}\wedge A_{\gamma}^{\text{inst}} +\I V_{\gamma}^{\text{inst}}\mathrm{d}\zeta_{\gamma}\wedge \mathrm{d}Z_{\gamma}\right)\right)
\end{equation}
\begin{equation}\label{invKF}
\omega_{3}:=2\pi\left(\frac{1}{4}\langle \mathrm{d}Z\wedge \mathrm{d}\overline{Z}\rangle-\frac{1}{2} \langle \mathrm{d}\zeta\wedge \mathrm{d}\zeta \rangle -\sum_{\gamma}\Omega(\gamma)\left( \frac{\I}{2} V^{\text{inst}}_{\gamma}\mathrm{d}Z_{\gamma}\wedge \mathrm{d}\overline{Z}_{\gamma}+\mathrm{d}\zeta_{\gamma}\wedge A_{\gamma}^{\text{inst}}\right)\right)\,.
\end{equation}
Then the triple of real $2$-forms $(\omega_1,\omega_2, \omega_3)$ corresponds to the Kähler forms of a pseudo-HK structure[Our terminology is such that the signature of the metric is not assumed to be constant, in case $N$ has several components.] on $N$ if and only if $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ are compatible.
We denote the resulting instanton corrected HK manifold from the previous theorem by $(N,g_N,\omega_1,\omega_2,\omega_3)$.
* Compared to <cit.>, we have rescaled the above 2-forms $\omega_i$ by a factor of $2\pi$, and rescaled $\zeta$ by $2\pi$ (i.e. $\zeta:\Gamma \to \mathbb{R}/\mathbb{Z}$ instead of $\zeta:\Gamma \to \mathbb{R}/2\pi\mathbb{Z}$). Furthermore, we have changed by a sign the convention of how the BPS indices $\Omega$ enter into the formulas (<ref>), (<ref>) (i.e. the above formulas would correspond in [25] to the HK metric associated to the mutually local variation of BPS structures $(M,\Gamma,Z,-\Omega)$). We do this change of conventions in order to simplify the formulas taken from <cit.> below and also to be able to compare more easily with the physics literature in Section <ref> below.
* In the expressions (<ref>) and (<ref>) we are combining the wedge $\wedge$ with the pairing $\langle -, - \rangle$ on $\Gamma^*\otimes \mathbb{C}$. For example, with respect to a Darboux frame $(\widetilde{\gamma}_i,\gamma^i)$ of $\Gamma$, we have $\langle \mathrm{d}Z\wedge \D\overline{Z}\rangle=\D Z_{\widetilde{\gamma}_i}\wedge \D \overline{Z}_{\gamma^i} -\D Z_{\gamma^i}\wedge \D \overline{Z}_{\widetilde{\gamma}_i}$ and $\langle \mathrm{d}\zeta \wedge \mathrm{d}\zeta \rangle=\mathrm{d}\zeta_{\widetilde{\gamma}_i}\wedge \mathrm{d}\zeta_{\gamma^i}-\mathrm{d}\zeta_{\gamma^i}\wedge\mathrm{d}\zeta_{\widetilde{\gamma}_i}=2\mathrm{d}\zeta_{\widetilde{\gamma}_i}\wedge \mathrm{d}\zeta_{\gamma^i}$. Furthermore, the expressions in (<ref>) and (<ref>) are actually global and well-defined due to the monodromy invariance of $\Omega(\gamma)$, and the support and convergence property of the variations of BPS structures.
* $(N,g_N,\omega_1,\omega_2,\omega_3)$ carries an infinitesimal rotating circle action <cit.>. Namely, there is a vector field $V$ on $N$ such that
\begin{equation}\label{rotvec}
\mathcal{L}_V(\omega_1+\I\omega_2)=2\I(\omega_1+\I\omega_2), \quad \mathcal{L}_V\omega_3=0\,.
\end{equation}
Note that due to the factor $2$ in (<ref>) the vector field $V$ is twice the vector field denoted $V$ in [25].
* Under the mild assumption on the flow of $\xi$ that it generates a free-action on $M$ of the (multiplicative) monoid $\mathbb{R}_{\geq 1}$, we can guarantee that $g_N$ has signature $(4,4n)$ where $n+1=\text{dim}_{\mathbb{C}}(M)$ <cit.>.
* If one sets $\Omega(\gamma)=0$ for all $\gamma \in \Gamma$, then $(N,g_N,\omega_1,\omega_2,\omega_3)$ reduces to the semi-flat HK manifold obtained via the rigid c-map.
§.§.§ Associated instanton corrected QK manifold via HK/QK correspondence
The instanton corrected QK manifold $(\overline{N},g_{\overline{N}})$ associated to the data of $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ was constructed in [25] by applying the HK/QK correspondence to $(N,g_N,\omega_1,\omega_2,\omega_3)$. In order to do this, one needs the following additional data:
* A hyperholomorphic principal $S^1$-bundle $(\pi_N:P\to N,\eta)$, where $\pi_N:P\to N$ is constructed via <cit.>, and $\eta$ is a connection on $P$ having curvature
\begin{equation}
\D \eta=\pi_N^*(\omega_3 -\frac{1}{2}\D\iota_V g_N)\,.
\end{equation}
The connection $\eta$ is given by <cit.>
\begin{equation}\label{etadef}
\eta:=\Theta +\pi^*_N\Big(\frac{\pi \I}{2}\pi^*_M( \overline{\partial} r^2-\partial r^2) - \sum_{\gamma}2\pi \Omega(\gamma)\eta_{\gamma}^{\text{inst}}-\frac{1}{2}\iota_Vg_N\Big)
\end{equation}
where $r^2:=g_M(\xi,\xi)$; $V$ is the rotating vector field satisfying (<ref>); $\Theta$ is another connection on $P$ having curvature $-\pi\langle \D \zeta \wedge \D \zeta \rangle=2\pi \mathrm{d}\zeta_{\gamma^i}\wedge \mathrm{d}\zeta_{\widetilde{\gamma}_i}$,
\begin{equation}\label{etagammainst}
\eta_{\gamma}^{\text{inst}}:=\frac{\I}{8\pi^2}\sum_{n>0}\frac{e^{2\pi \I n\zeta_{\gamma}}}{n}|Z_{\gamma}|K_1(2\pi n|Z_{\gamma}|)\Big(\frac{\D Z_{\gamma}}{Z_{\gamma}}-\frac{\D \overline{Z}_{\gamma}}{\overline{Z}_{\gamma}}\Big)\,.
\end{equation}
If $\sigma$ denotes a local coordinate for the $S^1$-fiber, then one can write
\begin{equation}\label{Thetacon}
\Theta=\pi\left(\D\sigma -\pi_N^*\langle \zeta,\D\zeta \rangle\right)\,.
\end{equation}
* We need to furthermore specify a Hamiltonian for $\omega_3$ with respect to the rotating vector field $V$ satisfying (<ref>), which in this case is given by <cit.>
\begin{equation}\label{fdef}
f=2\pi (r^2- 8c_{\ell} - \sum_{\gamma}\Omega(\gamma)\iota_V\eta_{\gamma}^{\text{inst}}),\quad c_{\ell} \in \mathbb{R},
\end{equation}
together with the lift $V^P$ of $V$ to $P$ given by
\begin{equation}
V^P:=\widetilde{V}+f_3\partial_{\sigma}, \quad f_3:=f-\frac{1}{2}g_{N}(V,V)
\end{equation}
where $\widetilde{V}$ denotes the horizontal lift with respect to $\eta$ and $\partial_{\sigma}$ is the vertical vector field of $P$ generating the $S^1$-action.
* Finally, we consider the open subset $N'\subset N$ given by
\begin{equation}\label{n'def}
N'=\{p\in N \quad | \quad f(p)\neq 0, \quad f_3(p)\neq 0, \quad g_N(V_p,V_p)\neq 0\},
\end{equation}
and the $1$-forms on $P$ given by
\begin{equation}\label{thetadef}
\theta_0^P=-\frac{1}{2}\pi_N^*\mathrm{d}f, \quad \theta_3^P:=\eta +\frac{1}{2}\pi_N^*\iota_Vg_N, \quad
\theta_1^P:=\frac{1}{2}\pi_N^*\iota_V\omega_2,\quad
\theta_2^P:=-\frac{1}{2}\pi_N^*\iota_V\omega_1\,,
\end{equation}
We then have
Let $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ be an integral CASK manifold and $(M,\Gamma,Z,\Omega)$ a compatible mutually local variation of BPS structures. Furthermore, let $(N,g_N,\omega_1,\omega_2,\omega_3)$, $(\pi_N:P\to N,\eta)$, $f$, $f_3$, $\theta_i^P$, $V^P$ and $N'$ the associated data defined in the previous points. Given any submanifold $\overline{N}\subset P|_{N'}$ transverse to $V^P$, the symmetric $2$-tensor
\begin{equation}
\end{equation}
defines a pseudo-QK metric on $\overline{N}$. Furthermore, if $(N,g_N,\omega_1,\omega_2,\omega_3)$ has signature $(4,4n)$, then $g_{\overline{N}}$ is positive definite on $\overline{N}_{+}=\overline{N}\cap\{f>0,f_3<0\}$.
Recall that we can guarantee that $(N,g_N,\omega_1,\omega_2,\omega_3)$ has signature $(4,4n)$ if the flow of $\xi$ generates a free-action on $M$ of the monoid $\mathbb{R}_{\geq 1}$ (see Section <ref>).
§.§.§ The case of a CASK domain
We now specialize the previous construction to the case of a CASK domain. This will be the case of interest in the following Section <ref>.
A CASK domain is a tuple $(M,\mathfrak{F})$ where
* $M\subset \mathbb{C}^{n+1}-\{0\}$ is a $\mathbb{C}^{\times}$-invariant domain. We denote the canonical holomorphic coordinates by $Z^i$, $i=0,1,...,n$. To avoid inessential coordinate changes later on, we
will assume for simplicity that $Z^0$ does not vanish on $M$.
* $\mathfrak{F}:M\to \mathbb{C}$ is a holomorphic function, homogeneous of degree $2$ with respect to the natural $\mathbb{C}^{\times}$-action on $M$.
* The matrix
\begin{equation}
\text{Im}\left(\tau_{ij}\right), \quad \tau_{ij}:=\frac{\partial^2 \mathfrak{F}}{\partial Z^i\partial Z^j}
\end{equation}
has signature $(n,1)$, and $\text{Im}(\tau_{ij})Z^i\overline{Z}^j<0$.
A CASK domain $(M,\mathfrak{F})$ induces in the usual way a CASK manifold $(M,g_M,\omega_M,\nabla,\xi)$ [5]. With our conventions on the signature of the CASK manifold, if $Z_i:=\frac{\partial \mathfrak{F}}{\partial Z^i}$, then $\{Z^i\}$ and $\{-Z_i\}$ are a global system of conjugate conical special holomorphic coordinates. If $x^i=\text{Re}(Z^i)$ and $y_i:=\text{Re}(Z_i)$, then $\nabla$ is defined such that $\D x^i$ and $\D y_i$ are flat. Furthermore
\begin{equation}\label{CASKconvention}
g_M=-\text{Im}(\tau_{ij})\D Z^i \D\overline{Z}^j, \quad \omega_M=-\frac{\mathrm{i}}{2}\text{Im}(\tau_{ij})\D Z^i\wedge \D\overline{Z}^j=\D x^i\wedge \D y_i, \quad \xi=Z^i\partial_{Z^i} +\overline{Z}^i\partial_{\overline{Z}^i}\,.
\end{equation}
Given a CASK domain $(M,\mathfrak{F})$ we can induce a canonical integral structure on the CASK manifold by defining $\Gamma \to M$ to be $\Gamma=\text{span}_{\mathbb{Z}}\{\partial_{x^i},\partial_{y_i}\}$. In the following, we will assume that:
* $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ is an integral CASK manifold induced by a CASK domain $(M,\mathfrak{F})$ with the canonical integral structure. In this case, we will sometimes use the notation $(\partial_{x^i},\partial_{y_i})=(\widetilde{\gamma}_i,\gamma^i)$.
* Given a mutually local variation of BPS structures $(M,\Gamma,Z,\Omega)$ with $(M,\Gamma)$ as in the previous point, we assume that $Z$ is the canonical central charge, and that $\text{Supp}(\Omega)\subset \text{span}_{\mathbb{Z}}\{\partial_{y_i}\}$. In particular, the canonical central charge satisfies in this case
\begin{equation}
Z=Z_{\widetilde{\gamma}_i}\mathrm{d}x^i+Z_{\gamma^i}\mathrm{d}y_i=-Z_i\D x^i+Z^i\D y_i\,.
\end{equation}
In order to construct the associated QK metric, we need to choose $\overline{N}\subset P|_{N'}$ transverse to $V^P$. Denoting by $\pi_M:=\pi\circ \pi_N$ the composition of the projections $\pi_N:P\to N$, and $\pi:N\to M$, we have that $\pi_M(p)=(Z^0,...,Z^n)$. We define $\overline{N}$ by
\begin{equation}\label{QKCASKdomain}
\overline{N}:=\{p \in P|_{N'} \quad| \quad \text{Arg}(Z^0)=0,\quad \text{ where $\pi_M(p)=(Z^0,...,Z^n)$} \}\,.
\end{equation}
Throughout the paper, we will use coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$ on (<ref>) defined as follows:
* $16\pi\rho:=2\pi r^2 -16\pi c_{\ell}$ where $r^2=g_M(\xi,\xi) = -\text{Im}(\tau_{ij}) Z^i \overline{Z}^j$ and $c_{\ell}\in \mathbb{R}$. Note that $r^2$ is a Kähler potential for the affine special Kähler metric $g_M$.
* $z^a:=Z^a/Z^0$ for $a=1,...,n$. These are global holomorphic coordinates on the induced projective special Kähler (PSK) manifold $(\overline{M},g_{\overline{M}})$ induced by the CASK domain. In particular, we have a projection $\pi_{\overline{M}}: M\to \overline{M}$.
* $(\zeta^i,\widetilde{\zeta}_i)$ are given by $\zeta^i:=-\zeta_{\partial_{y_i}}=-\zeta_{\gamma^i}$ and $\widetilde{\zeta}_i:=\zeta_{\partial_{x^i}}=\zeta_{\widetilde{\gamma}_i}$, where the latter are the evaluation map on $N$ contracted with $\partial_{x^i}$ and $\partial_{y_i}$.
* $\sigma$ is a local coordinate for the $S^1$-fiber of $\pi_N:P\to N$ satisfying (<ref>).
In the string theory setting, the coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$ can be identified with certain fields from the type IIA hypermultiplet. Namely, $\rho$ is the 4d-dilaton, $z^a$ are coordinates for the complex moduli of the Calabi-Yau, $(\zeta^i,\widetilde{\zeta}_i)$
are the RR-axions, and $\sigma$ is the NS-axion. Furthermore, the constant $c_{\ell}$ appearing in the definition of $\rho$ is identified with the $1$-loop correction to the tree-level Ferrara-Sabharwal metric.
In <cit.> an explicit expression of the resulting QK metric $(\overline{N},g_{\overline{N}})$ in these coordinates is given (with slightly different conventions). In order to write down the formula we introduce the following notation:
* We denote by $\widetilde{Z}_{\gamma}:=Z_{\gamma}/Z^0$ the normalized central charge. In particular, we have $\widetilde{Z}_{\gamma^i}=z^i$ with $z^0=1$. If furthermore we let $\mathcal{K}=-\log(-2\text{Im}(\tau_{ij})z^i\overline{z}^j)$, then $\mathcal{K}$ is a global Kähler potential for the projective special Kähler manifold $(\overline{M},g_{\overline{M}})$ induced by the CASK domain. Note that
\begin{equation} \label{r2:eq}r^2 = |Z^0|^2\frac{e^{-\mathcal{K}}}{2}.\end{equation}
* We denote $N_{ij}=-2\text{Im}(\tau_{ij})$. If $\gamma \in \text{Supp}(\Omega)$ we write $\gamma=q_i(\gamma)\gamma^i$ (recall that we assume that $\text{Supp}(\Omega)\subset \text{span}_{\mathbb{Z}}\{\gamma^i\}$), and define
\begin{equation}
\end{equation}
* We let
\begin{equation}\label{instW}
W_i:=\D\zeta_{\widetilde{\gamma}_i}+\tau_{ij}\D\zeta_{\gamma^j}, \;\;\;\;\;\; W_i^{\text{inst}}:=-\sum_{\gamma}\Omega(\gamma)q_i(\gamma)( A_{\gamma}^{\text{inst}}-\I V_{\gamma}^{\text{inst}}\D\zeta_{\gamma}),
\end{equation}
\begin{equation}\label{etainstdef}
\eta^{\text{inst}}:=-\sum_{\gamma}\Omega(\gamma) \eta_{\gamma}^{\text{inst}}-\frac{1}{2}\iota_V\left(\frac{g_N}{2\pi}- \pi^*_Mg_{M}\right) \,.
\end{equation}
* We split $f=16\pi\rho+f^{\text{inst}}=16\pi(\rho+\rho^{\text{inst}})$ in (<ref>) using that $16\pi \rho=2\pi r^2-16\pi c_{\ell}$, and where $f^{\text{inst}}=16\pi \rho^{\text{inst}}$ contains the terms with the BPS indices $\Omega(\gamma)$, namely
\begin{equation}
\end{equation}
Finally, we denote $ f_3^{\text{inst}}=16\pi \rho_3^{\text{inst}}=2\pi \iota_V \eta^{\text{inst}}$. We remark that in the case where $\Omega(\gamma)=0$ for all $\gamma$ we have $f^{\text{inst}}=f_3^{\text{inst}}=0$, and similarly for all the other quantities with an $^\text{inst}$ superscript.
We then have
<cit.> Let $(M,\mathfrak{F})$ and $(M,\Gamma,Z,\Omega)$ be as before. By possibly restricting $M$, we assume that $M$ is the maximal open subset where $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ are compatible. Furthermore, let $\overline{N}$ be as (<ref>). Then in the coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$ from Definition <ref> the instanton corrected QK metric $(\overline{N},g_{\overline{N}})$ associated to $(M,g_M,\omega_M,\nabla,\xi,\Gamma)$ and $(M,\Gamma,Z,\Omega)$ has the form:
\begin{equation}\label{coordQKmetric2}
\begin{split}
g_{\overline{N}}=& \frac{\rho+c_{\ell}}{\rho+\rho^{\text{inst}}}\Big(g_{\overline{M}}+2e^{\mathcal{K}}\sum_{\gamma}\Omega(\gamma)V_{\gamma}^{\text{inst}}\Big|\mathrm{d}\widetilde{Z}_{\gamma} + \widetilde{Z}_{\gamma}\Big(\frac{\D\rho}{2(\rho+c_{\ell})}+\frac{\D\mathcal{K}}{2}\Big)\Big|^2\Big)\\
&+\frac{1}{2(\rho+\rho^{\text{inst}})^2}\Big(\frac{\rho +2c_{\ell} -\rho^{\text{inst}}}{2(\rho+c_{\ell})}\D\rho^2+2\D\rho \D \rho^{\text{inst}}|_{\overline{N}}+(\D \rho^{\text{inst}})^2|_{\overline{N}}\Big)\\
&+\frac{\rho+c_{\ell}+\rho_-^{\text{inst}}}{64(\rho+\rho^{\text{inst}})^2(\rho+2c_{\ell}-\rho_3^{\text{inst}})}\Big(\D \sigma -\langle \zeta,\D\zeta \rangle -4 c_{\ell}\D^c\mathcal{K}+\eta_+^{\text{inst}}|_{\overline{N}}+\frac{\rho_+^{\text{inst}}-c_{\ell}}{\rho+c_{\ell}+\rho_-^{\text{inst}}}\eta_-^{\text{inst}}|_{\overline{N}}\Big)^2\\
&-\frac{1}{4(\rho+\rho^{\text{inst}})}(W_i+W_i^{\text{inst}}|_{\overline{N}})(N+N^{\text{inst}})^{ij}(\overline{W}_j+\overline{W}_j^{\text{inst}}|_{\overline{N}}) \\
&+\frac{(\rho+c_{\ell})e^{\mathcal{K}}}{2(\rho+\rho^{\text{inst}})^2}\Big|z^i(W_i+W_i^{\text{inst}}|_{\overline{N}})- \frac{\mathrm{i}}{2}\sum_{\gamma}\Omega(\gamma) A_{\gamma}^{\text{inst}}(V)\Big(\D\widetilde{Z}_{\gamma} + \widetilde{Z}_{\gamma}\Big(\frac{\D\rho}{2(\rho+c_{\ell})}+\frac{\D \mathcal{K}}{2}\Big)\Big)\Big|^2\\
\end{split}
\end{equation}
where $\mathrm{d}^c:=\mathrm{i}(\overline{\partial}-\partial)$, $\eta^{\text{inst}}_{\pm}$ are given by
\begin{equation}
\eta_{\pm}^{\text{inst}}:=\Big( \eta^{\text{inst}}-4\rho_3^{\text{inst}}\widetilde{\eta}\Big)\pm \Big( \sum_{\gamma}\Omega(\gamma)\eta_{\gamma}^{\text{inst}}-4\rho^{\text{inst}}\widetilde{\eta}\Big)\,, \quad \widetilde{\eta}:=\mathrm{d}^c \log(r)\,,
\end{equation}
\begin{equation}
\rho_{\pm}^{\text{inst}}:=(\rho^{\text{inst}}\pm \rho_3^{\text{inst}})/2 \,.
\end{equation}
Furthermore, the open subset $\overline{N}_{+}=\overline{N}\cap\{f>0,f_3<0\}$ is non-empty, and $g_{\overline{N}}$ is positive-definite on $\overline{N}_{+}$.
* In Theorem <ref> we have relaxed a bit the possible restriction of $M$ compared to <cit.>. In [25] we assume that we can restrict to an $M$ invariant under the action of the monoid $\mathbb{R}_{\geq 1}\times S^1$ to make the CASK structure compatible with the BPS structure. This ensures that no matter the point of $\overline{M}$ the metric is defined for $\rho>K$ for some sufficiently big uniform $K$. Our weakened assumption makes it so that the constant $K$ might depend on the point $z^a\in \overline{M}$.
* When $\Omega(\gamma)=0$ for all $\gamma \in \Gamma$ the expression reduces to the 1-loop corrected Ferrara-Sabharwal metric:
\begin{equation}
\begin{split}
g_{\overline{N}}=& \frac{\rho+c_{\ell}}{\rho}g_{\overline{M}}+\frac{\rho +2c_{\ell}}{4\rho^2(\rho+c_{\ell})}\D\rho^2+\frac{\rho+c_{\ell}}{64\rho^2(\rho+2c_{\ell})}\Big(\D \sigma -\langle \zeta,\D\zeta \rangle -4 c_{\ell}\D^c\mathcal{K}\Big)^2\\
&-\frac{1}{4\rho}\left(N^{ij}-\frac{2(\rho+c_{\ell})e^{\mathcal{K}}}{\rho }z^i\overline{z}^j\right)W_i\overline{W}_j\,.
\end{split}
\end{equation}
In particular $(\overline{N},g_{\overline{N}})$ can be thought as a deformation of the $1$-loop corrected metric.
* Since the instanton corrections of the HK geometry are exponentially suppressed as $|Z_{\gamma}|\to \infty$ for $\gamma \in \text{Supp}(\Omega)$, it is easy to check that the possibly restricted $M$ from above satisfying the required conditions is never empty. Furthermore, on $\overline{N}$ the instanton corrections of the QK geometry are exponentially suppressed as $\rho \to \infty$ (due to the relation $|Z_{\gamma}|=|Z^0||\widetilde{Z}_{\gamma}|=4\sqrt{\rho+c_{\ell}}e^{\mathcal{K}/2}|\widetilde{Z}_{\gamma}|$), so we can ensure that $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)\in \overline{N}_{+}$ by taking $\rho$ sufficiently big.
* The function $f=16\pi\rho+f^{\text{inst}}$ can be thought in the string theory setting as the D-instanton corrected 4d dilaton (up to different conventions in the normalization) [9].
* When comparing (<ref>) to <cit.>, we note that here we are using different conventions for the normalization of $g_{N}$; the rotating vector field $V$; the functions $N_{ij}$, $N_{ij}^{\text{inst}}$, $e^{-\mathcal{K}}$, $\eta_{\pm}^{\text{inst}}$; the signature of $\text{Im}(\tau_{ij})$; and the coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$. Compared to [25], $g_{N}$ is scaled by $2\pi$;
$V$, $N_{ij}$, $N_{ij}^{\text{inst}}$ and $e^{-\mathcal{K}}$ are scaled by $2$; $\eta_{\pm}^{\text{inst}}$ are scaled by $\pi^{-1}$; and the signature of $\text{Im}(\tau_{ij})$ is opposite. Furthermore, the coordinates and $1$-loop constant of [25] are related to the ones in Definition <ref> by performing the scaling
\begin{equation}\label{scaling}
\rho \to 16\pi\rho,\quad c_{\ell}\to 16\pi c_{\ell}, \quad \sigma \to \pi \sigma, \quad \zeta^i\to -2\pi\zeta^i, \quad \widetilde{\zeta}_i \to 2\pi \widetilde{\zeta}_i\,.
\end{equation}
Finally, as mentioned in Remark <ref>, the sign with which the $\Omega(\gamma)$ enter the formula in (<ref>) is opposite to <cit.>. This changes of convention will make the formulas from subsequent section look more simple and more easily comparable to the physics literature.
In what follows, it will be useful to consider the following lift of $(\overline{N},g_{\overline{N}})$:
We will denote by $(\widetilde{N},g_{\overline{N}})$ the QK manifold obtained by lifting the QK metric $(\overline{N},g_{\overline{N}})$ obtained in Theorem <ref> to the open subset $\widetilde{N}\subset \mathbb{R}_{>0}\times \overline{M}\times \mathbb{R}^{2n+2}\times \mathbb{R}$ obtained by considering $(\zeta^i,\widetilde{\zeta}_i,\sigma)\in \mathbb{R}^{2n+2}\times \mathbb{R}$ as (non-periodic) global coordinates of $\mathbb{R}^{2n+2}\times \mathbb{R}$. We will call such a space an instanton corrected c-map space.
§.§ Twistor space description and Darboux coordinates
Let $(\overline{N},g_{\overline{N}},Q)$ denote a QK manifold, where $Q\to \overline{N}$ denotes the associated quaternionic structure. Namely, a parallel subbundle $Q\subset \text{End}(T\overline{N})$ admiting local trivializations $(J_1,J_2,J_3)$ by skew endomorphisms satisfying the quaternion relations. The structure of a QK manifold $(\overline{N},g_{\overline{N}},Q)$ can be encoded in a holomorphic object, known as the twistor space $(\mathcal{Z},\mathcal{I},\lambda,\tau)$ [39, 35]. Here $\mathcal{Z}\to \overline{N}$ is a sphere subbundle of $Q$ defined by
\begin{equation}
\mathcal{Z}_p:=\{J\in Q_p \; |\; J^{2}=-1\}, \quad p \in \overline{N};
\end{equation}
$\mathcal{I}$ is a canonical holomorphic structure on $\mathcal{Z}$; $\lambda \in \Omega^1(\mathcal{Z},\mathcal{L})$ defines a holomorphic contact structure on $\mathcal{Z}$, where $\mathcal{L}\to \mathcal{Z}$ is a certain holomorphic line bundle; and $\tau$ is a real structure on $\mathcal{Z}$ (i.e. an antiholomorphic involution).
In what follows, we consider $(\overline{N},g_{\overline{N}})$ built in the previous Section <ref>, and its lift $(\widetilde{N},g_{\overline{N}})$. We start by recalling the following from the discussion in <cit.>:
Let $(\widetilde{N},g_{\overline{N}})$ be the lift of an instanton corrected QK manifold associated to a CASK domain $(M,\mathfrak{F})$ and mutually local variation of BPS structures $(M,\Gamma,Z,\Omega)$. Then $\mathcal{Z}\cong \widetilde{N}\times \mathbb{C}P^1$ (non-holomorphically), and there is a holomorphic coordinate $t$ on the $\mathbb{C}P^1$ factor and a holomorphic section $s$ of the holomorphic line bundle $\mathcal{L}\to \mathcal{Z}$ vanishing at $t=0,\infty$, such that
\begin{equation}\label{contactlocal}
\lambda=\left(f\frac{\D t}{t}+t^{-1}\theta_{+}^P|_{\overline{N}}-2\mathrm{i} \theta_3^P|_{\overline{N}} +t \theta_{-}^{P}|_{\overline{N}}\right)\cdot s
\end{equation}
where $\theta_{\pm}^P:=\theta_1^P\pm \mathrm{i}\theta_2^P$, and $\theta^P_i$ for $i=1,2,3$ are defined as in (<ref>), and $f$ is defined as in (<ref>).
This follows from the discussion in <cit.>, where the twistor space of a QK manifold $(\overline{N},g_{\overline{N}})$ obtained via HK/QK correspondence is described in terms of the “HK data" given by $(N,g_{N},\omega_1,\omega_2,\omega_3)$, $(\pi_N:P\to N,\eta)$, $f$, and the associated HK cone. In particular, whenever the QK manifold admits a global chart of coordinates, which is the case of $(\widetilde{N},g_{\overline{N}})$ obtained in Section <ref>, it follows that $\mathcal{Z}\cong \widetilde{N}\times \mathbb{C}P^1$ non-holomorphically. The formula (<ref>) follows from <cit.>.
The lift to $\widetilde{N}\rightarrow \overline{N}$ is understood.
In what follows, we will be concerned in describing Darboux coordinates for the contact structure $\lambda$ expressed as (<ref>), in the case where $(\widetilde{N},g_{\overline{N}})$ is the instanton corrected QK metric obtained in Section <ref>. For this it will be important to have an explicit expression for $f$, $\theta_{+}^P|_{\overline{N}}$ and $\theta_3^P|_{\overline{N}}$ ($\theta_-^P|_{\overline{N}}$ can be obtained from $\theta_-^P=\overline{\theta_+^P}$).
Consider the CASK domain $(M,\mathfrak{F})$ and mutually local variation of BPS structures $(M,\Gamma,Z,\Omega)$ as in Section <ref>. Then $f$, $\theta_{+}^P|_{\overline{N}}$ and $\theta_3^P|_{\overline{N}}$ from (<ref>) have the following formulas with respect to the coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$:
\begin{align*}
f&=16\pi\rho + \frac{2R}{\pi}\sum_{\gamma}\Omega(\gamma)\sum_{n>0}\frac{e^{-2\pi\mathrm{i}n\zeta_{\gamma}}}{n}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\\
\theta^{P}_{+}|_{\overline{N}}&=-4\pi R\langle \widetilde{Z},\D\zeta \rangle +2\mathrm{i}R\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi\mathrm{i}n\zeta_{\gamma}}K_0(4\pi Rn|\widetilde{Z}_{\gamma}|)\D\zeta_{\gamma}\\
&\quad \quad +2R^2\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi \mathrm{i}n\zeta_{\gamma}}|\widetilde{Z}_{\gamma}|K_1(4\pi R n|\widetilde{Z}_{\gamma}|)\left(\frac{\D\widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}+\frac{\D\overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}+\frac{\D\rho}{(\rho+c_{\ell})}+\D\mathcal{K}\right)\\
\theta_3^P|_{\overline{N}}&=\pi\D\sigma -\pi \langle\zeta, \D\zeta \rangle-4\pi(\rho+c_{\ell})\D^c\mathcal{K}-\frac{\mathrm{i}R}{2\pi}\sum_{\gamma}\Omega(\gamma)\sum_{n>0}\frac{e^{-2\pi \mathrm{i}n\zeta_{\gamma}}}{n}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\left(\frac{\D\widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}-\frac{\D\overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}\right) \numberthis \label{fthetaexpres}
\end{align*}
where $R:=2\sqrt{\rho+c_{\ell}}e^{\mathcal{K}/2}$, $\widetilde{Z}_{\gamma}=q_0+q_az^a$ and $\zeta_{\gamma}=-q_i\zeta^i$ for $\gamma=q_i\gamma^i$, and in the last formula we have used $\mathrm{d}^c=\mathrm{i}(\overline{\partial}-\partial)$.
To obtain the above identities we will use that
\begin{equation}\label{z0rrel}
\end{equation}
together with $\Omega(\gamma)=\Omega(-\gamma)$ and the fact that the rotating vector field is given globally in the case of a CASK domain by
\begin{equation}\label{vcaskd}
\end{equation}
To obtain the formula for $f$ we just use (<ref>) together with (<ref>) and a relabeling of the sum variable $\gamma \to -\gamma$. To obtain the formula for $\theta_{+}^P|_{\overline{N}}$ we use the formulas (<ref>) for $\omega_1+\mathrm{i}\omega_2$, the definitions for $\theta_1^P$ and $\theta_2^P$ in (<ref>), a relabeling $\gamma \to -\gamma$ of the sums over $\gamma$, the CASK relation $Z_{\widetilde{\gamma}_i}=-Z_i=-\tau_{ij}Z^j=-\tau_{ij}Z_{\gamma^j}$, and the fact that
\begin{equation}\label{restrictionrel}
\left(\frac{\D Z_{\gamma}}{Z_{\gamma}}-\frac{\D\overline{Z}_{\gamma}}{\overline{Z}_{\gamma}}\right)\Bigg|_{\overline{N}}=\left(\frac{\D\widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}-\frac{\D\overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}\right), \quad \D Z_{\gamma}|_{\overline{N}}=|Z^0|\widetilde{Z}_{\gamma}\left(\frac{\D\widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}} +\frac{\D \rho}{2(\rho+c_{\ell})}+\frac{\D\mathcal{K}}{2}\right)\,.
\end{equation}
Finally, the equality for $\theta_3^P|_{\overline{N}}$ follows from the formulas (<ref>) for $\eta$, (<ref>) for $\theta_3^P$, the first equation from (<ref>), and again a relabeling $\gamma \to -\gamma$ in the sum over $\gamma$. We also use the fact that
\begin{equation}
\D^c r^2|_{\overline{N}}=-r^2\D^c \mathcal{K}\,,
\end{equation}
which follows from the second relation in (<ref>).
We remark that if we denote $F_i:=\partial_{Z^i}\mathfrak{F}/Z^0=Z_i/Z^0$, then in the expression (<ref>) we have
\begin{equation}
\langle \widetilde{Z},\mathrm{d}\zeta\rangle= \widetilde{Z}_{\widetilde{\gamma}_i}\D\zeta_{\gamma^i}-\widetilde{Z}_{\gamma^i}\D\zeta_{\widetilde{\gamma}_i}=(-F_i)\D(-\zeta^i) - z^i\D\widetilde{\zeta}_i=F_i\D\zeta^i-z^i\mathrm{d}\widetilde\zeta_i\,.
\end{equation}
Similarly, we have
\begin{equation}
\langle \zeta,\mathrm{d}\zeta\rangle=\zeta_{\widetilde{\gamma}_i}\D\zeta_{\gamma^i}-\zeta_{\gamma^i}\D\zeta_{\widetilde{\gamma}_i}=-\widetilde{\zeta}_i\mathrm{d}\zeta^i+\zeta^i\mathrm{d}\widetilde{\zeta}_i\,.
\end{equation}
§.§.§ Darboux coordinates for c-map spaces associated to CASK domains
In this section we focus on the easier case of a c-map space associated to a CASK domain. The following coordinates have been been previously obtained in the physics literature in [36] in the $c_{\ell}=0$ case, and in the 1-loop corrected case in [14] via slightly different methods.
Consider the QK manifold $(\widetilde{N},g_{\overline{N}})$ obtained from a CASK domain $(M,\mathfrak{F})$ via the c-map (i.e. by taking $\Omega(\gamma)=0$ for all $\gamma$ in our previous constructions). If $F_i:=Z_i/Z^0=\partial_{Z^i} \mathfrak{F}/Z^0$, $t$ denotes the twistor fiber coordinate from Proposition <ref> and $R=2\sqrt{\rho+c_{\ell}}e^{\mathcal{K}/2}$, then the functions on the twistor space given by
\begin{equation}\label{Darbouxcoords1loop}
\begin{split}
\xi^{i}&=\zeta^{i}- \mathrm{i}R(t^{-1}z^i +t\overline{z}^i)\\
\widetilde{\xi}_i&=\widetilde{\zeta}_{i}- \mathrm{i}R(t^{-1}F_{i}+t\overline{F}_i)\\
\alpha&= \sigma -\mathrm{i}R(t^{-1}\langle \widetilde{Z},\zeta \rangle +t \langle \overline{\widetilde{Z}}, \zeta \rangle)-8\mathrm{i} c_{\ell}\log(t)\,,
\end{split}
\end{equation}
define Darboux coordinates for the holomorphic contact structure $\lambda$, in the sense that
\begin{equation}
\lambda=-2\pi\mathrm{i}(\D\alpha +\widetilde{\xi}_i\D\xi^i-\xi^i\D\widetilde{\xi}_i)\cdot s\,.
\end{equation}
By Proposition <ref>, it is enough to check that
\begin{equation}
-2\pi \I(\D\alpha +\widetilde{\xi}_i\D\xi^i-\xi^i\D\widetilde{\xi}_i)=f\frac{\D t}{t}+t^{-1}\theta_{+}^P|_{\overline{N}}-2\mathrm{i} \theta_3^P|_{\overline{N}} +t \theta_{-}^{P}|_{\overline{N}}
\end{equation}
where $f$, $\theta_{+}^P=\overline{\theta_{-}^P}$ and $\theta_3^P$ are obtained by setting $\Omega(\gamma)=0$ for all $\gamma$ in (<ref>). That is:
\begin{equation}
f=16\pi \rho, \quad\quad \theta_3^P|_{\overline{N}}=\pi\D\sigma -\pi\langle\zeta, \D\zeta \rangle-4\pi(\rho+c_{\ell})\D^c\mathcal{K}, \quad\quad
\theta_{+}^P|_{\overline{N}}=-4\pi R\langle \widetilde{Z},\D\zeta \rangle\,.
\end{equation}
We now compute
\begin{equation}
\begin{split}
\widetilde{\xi}_i\D\xi^i-\xi^i\D\widetilde{\xi}_i=&-\langle \zeta, \D\zeta \rangle +\I(t^{-1}\langle\widetilde{Z},\zeta \rangle +t \langle\overline{\widetilde{Z}}, \zeta \rangle)\D R +\I R(-t^{-2}\langle\widetilde{Z},\zeta\rangle \D t+\langle \overline{\widetilde{Z}},\zeta \rangle \D t)+ 8\I(\rho+c_{\ell})\frac{\D t}{t}\\
& +\I R(t^{-1}\langle \D \widetilde{Z},\zeta \rangle + t \langle \D\overline{\widetilde{Z}},\zeta \rangle) -\I R(t^{-1}\langle \widetilde{Z},\D\zeta \rangle + t \langle \overline{\widetilde{Z}},\D\zeta \rangle)-4(\rho+c_{\ell})\D^c\mathcal{K}\,,
\end{split}
\end{equation}
where for the terms $8\I (\rho+c_{\ell})\frac{\D t}{t}$ and $-4(\rho+c_{\ell})\D^c\mathcal{K}$ we have used that the CASK relation $F_i=\tau_{ij}z^j$ implies
\begin{equation}
\begin{split}
\Big(-8(\rho+c_{\ell})e^{\mathcal{K}}F_i\overline{z}^i+8(\rho+c_{\ell})e^{\mathcal{K}}\overline{F}_iz^i\Big)\frac{\D t}{t}&= \Big(-8(\rho+c_{\ell})e^{\mathcal{K}}\tau_{ij}z^j\overline{z}^i+8(\rho+c_{\ell})e^{\mathcal{K}}\overline{\tau}_{ij}\overline{z}^jz^i\Big)\frac{\D t}{t}\\
&=\Big(8\I(\rho+c_{\ell})e^{\mathcal{K}}(-2\text{Im}(\tau_{ij}))z^j\overline{z}^i\Big)\frac{\D t}{t}=8\I(\rho+c_{\ell})\frac{\D t}{t}\\
\end{split}
\end{equation}
and by using the relation $\D F_i=\tau_{ij}\D z^i$
\begin{equation}
\begin{split}
-4(\rho+c_{\ell})e^{\mathcal{K}}F_i\D \overline{z}^i&-4(\rho+c_{\ell})e^{\mathcal{K}}\overline{F}_i\D z^i +4(\rho+c_{\ell})e^{\mathcal{K}}\D \overline{F}_iz^i+4(\rho+c_{\ell})e^{\mathcal{K}}\D F_i\overline{z}^i\\
&=-4(\rho+c_{\ell})e^{\mathcal{K}}(-2\text{Im}(\tau_{ij}))\Big(i\overline{z}^j\D z^i -iz^i\D \overline{z}^j\Big)\\
&=-4(\rho+c_{\ell})\D ^c\mathcal{K}\,.
\end{split}
\end{equation}
(Recall that $\mathcal{K}=-\log K$ where $K= -2 \mathrm{Im}(\tau_{ij}) z^i\bar z^j$).
On the other hand, we find that
\begin{equation}
\begin{split}
\D \alpha=&\D \sigma -\I(t^{-1}\langle \widetilde{Z},\zeta \rangle +t \langle \overline{\widetilde{Z}},\zeta \rangle)\D R + \I R(t^{-2}\langle \widetilde{Z},\zeta \rangle - \langle \overline{\widetilde{Z}},\zeta \rangle)\D t\\
&-\I R(t^{-1}\langle \D \widetilde{Z},\zeta \rangle +t \langle \D \overline{\widetilde{Z}},\zeta \rangle)-\I R(t^{-1}\langle \widetilde{Z},\D \zeta \rangle +t \langle \overline{\widetilde{Z}},\D \zeta \rangle)-8\I c_{\ell}\frac{\D t}{t},
\end{split}
\end{equation}
so we conclude that
\begin{equation}
\begin{split}
\D \alpha +\widetilde{\xi}_i\D \xi^i-\xi^i\D \widetilde{\xi}_i&=8\I \rho\frac{\D t}{t}+\D \sigma -\langle \zeta, \D \zeta \rangle-2\I R(t^{-1}\langle \widetilde{Z},\D \zeta \rangle + t \langle \overline{\widetilde{Z}},\D \zeta \rangle) -4(\rho+c_{\ell})\D ^c\mathcal{K}\\
\end{split}
\end{equation}
so that
\begin{equation}
\begin{split}
-2\pi \mathrm{i}\left(\D \alpha +\widetilde{\xi}_i\D \xi^i-\xi^i\D \widetilde{\xi}_i\right)&=f\frac{\D t}{t} -4\pi R(t^{-1}\langle \widetilde{Z},\D \zeta \rangle + t \langle \overline{\widetilde{Z}},\D \zeta \rangle)-2\I \left(\pi\D\sigma -\pi\langle\zeta, \D\zeta \rangle-4\pi(\rho+c_{\ell})\D^c\mathcal{K}\right)\\
&=f\frac{\D t}{t}+t^{-1}\theta_+^P|_{\overline{N}} -2\mathrm{i}\theta_3^P|_{\overline{N}} +t\theta_-^P|_{\overline{N}}
\end{split}
\end{equation}
and the result follows.
§.§.§ The case with instanton corrections
We now consider the case where the BPS indices are not all equal to $0$. We want to write down the modifications of the coordinates (<ref>), such that
\begin{equation}
\lambda=-2\pi \mathrm{i}\left(\D \alpha + \widetilde{\xi}_i\D \xi^i -\xi^i\D \widetilde{\xi}_i\right)\cdot s=\left(f\frac{\D t}{t}+t^{-1}\theta_+^P|_{\overline{N}} -2\mathrm{i}\theta_3^P|_{\overline{N}} +t\theta_-^P|_{\overline{N}}\right)\cdot s
\end{equation}
where $f$ and $\theta_{\alpha}^{P}$ are as in (<ref>).
The expressions for the coordinates below have been previously found in the physics literature (using slightly different arguments and conventions), for example in [13, 12].
Consider the instanton corrected QK manifold $(\widetilde{N},g_{\overline{N}})$ associated to a CASK domain and $(M,\Gamma,Z,\Omega)$, as before. Then the functions on the twistor space $\mathcal{Z}$ given by
\begin{equation}\label{Darbouxcoordsinst}
\begin{split}
\xi^{i}&=\zeta^{i}- \mathrm{i}R(t^{-1}z^i +t\overline{z}^i)\\
\widetilde{\xi}_i&=\widetilde{\zeta}_{i}- \mathrm{i}R(t^{-1}F_{i}+t\overline{F}_i)-\frac{1}{8\pi^2 }\sum_{\gamma}\Omega(\gamma)n_i(\gamma) \int_{l_{\gamma}}\frac{\D \zeta}{\zeta}\frac{t+\zeta}{t-\zeta}\log(1-\exp(2\pi\mathrm{i}\xi_{\gamma}(\zeta)))\\
\alpha&= \sigma -\mathrm{i}R(t^{-1}\langle \widetilde{Z},\zeta \rangle +t \langle \overline{\widetilde{Z}}, \zeta \rangle)-8\mathrm{i} c_{\ell}\log(t)+\frac{1}{4\pi^2\mathrm{i}}\left(t^{-1}\mathcal{W} +t\overline{\mathcal{W}}-\frac{1}{2\pi}\sum_{\gamma}\Omega(\gamma)\int_{l_{\gamma}}\frac{\D \zeta}{\zeta}\frac{t+\zeta}{t-\zeta}\mathrm{L}(\exp{2\pi\mathrm{i}\xi_{\gamma}(\zeta)})\right)\,,
\end{split}
\end{equation}
where the integration contours are given by $l_{\gamma}:=\mathbb{R}_{<0}\widetilde{Z}_{\gamma}$ oriented from $0$ to $\infty$, $\xi_\gamma(\zeta) = q_i \xi^i(\zeta)$ for $\gamma=q_i\gamma^i$ and where we abbreviate $\xi^i(\zeta)=\xi^i(R,z^i,\zeta^i,t=\zeta)$,
\begin{equation}\label{wdef}
\mathcal{W}:=R\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}\frac{e^{-2\pi\mathrm{i}n\zeta_{\gamma}}}{n}K_0(4\pi Rn|\widetilde{Z}_{\gamma}|)\,,
\end{equation}
\begin{equation}\label{Rogersdilogarithm}
\mathrm{L}(\exp({2\pi\mathrm{i} \xi_{\gamma}}))=\mathrm{Li}_2(\exp(2\pi\mathrm{i}\xi_{\gamma}))+\pi\mathrm{i}\xi_{\gamma}\log(1-\exp(2\pi\mathrm{i}\xi_{\gamma}));
\end{equation}
define Darboux coordinates for the contact structure $\lambda$ in the sense that
\begin{equation}
\lambda=-2\pi \I(\D \alpha +\widetilde{\xi}_i\D \xi^i-\xi^i\D \widetilde{\xi}_i)\cdot s\,.
\end{equation}
A explicit proof is given in Appendix <ref>. This result is not needed for the following sections and is only included for completeness. See also the work in the physics literature [12].
* The function $\mathrm{Li}_2(x)$ appearing in (<ref>) is the dilogarithm function, while the function $\mathrm{L}(x)=\mathrm{Li}_2(x)+\frac{\log(x)}{2}\log(1-x)$ is the Rogers dilogarithm.
* We remark that along the chosen $l_{\gamma}$ contours, the integrands are exponentially decreasing near $0$ and $\infty$. In particular, one can deform the contour $l_{\gamma}$ within the half plane centered at $l_{\gamma}$ without changing the value of the integral, as long as it does not collide with another ray of the form $l_{\gamma}$ for $\gamma \in \text{Supp}(\Omega)$.
§ S-DUALITY ON INSTANTON CORRECTED Q-MAP SPACES
The structure of the section is as follows:
* In Section <ref> we define instanton corrected q-map spaces $(\overline{N},g_{\overline{N}})$ in terms of the construction of [25] reviewed in Section <ref>,
applied to certain special pairs of compatible initial data $(M,\mathfrak{F})$ and $(M,\Gamma,Z,\Omega)$. These spaces will be the main objects of study in the rest of the section.
* In Section <ref> we define the quantum corrected mirror map, previously defined in [15], and relating the type IIA variables $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$ with the type IIB variables $(\tau_1+\I\tau_2,b^a+\I t^a, c^a,c_a,c_0,\psi)$. In terms of the type IIB variables, we define the corresponding $\mathrm{SL}(2,\mathbb{Z})$-action.
* In Section <ref> we show that the holomorphic contact structure of the twistor space associated to $(\overline{N},g_{\overline{N}})$ admits Darboux coordinates of the form studied in [15]. We show this directly by Poisson resumming the contact form (<ref>), which in turn is expressed in terms of the Bessel function expressions (<ref>) specialized to the particular form of the BPS indices below (<ref>).
* In Section <ref> we give conditions that guarantee that either the S-duality $\mathrm{SL}(2,\mathbb{Z})$-transformations or the subgroup $\langle S \rangle \subset \mathrm{SL}(2,\mathbb{Z})$ acts by isometries on an instanton corrected q-map space.
§.§ Setting
Let us first recall the notion of a PSK manifold in the image of the r-map, and the associated CASK domain.
A projective special real (PSR) manifold is a Riemannian manifold $(\mathcal{H},g_{\mathcal{H}})$ where $\mathcal{H}\subset \mathbb{R}^n$ is a hypersurface, for which there is a cubic polynomial $h:\mathbb{R}^n\to \mathbb{R}$ such that $\mathcal{H}\subset \{h(t)=1\}$ and $g_{\mathcal{H}}=-\partial^2h|_{T\mathcal{H}\times T\mathcal{H}}$.
If we denote the canonical coordinates of $\mathbb{R}^n$ by $t^a$, then we can write
\begin{equation}\label{cubicpsr}
\end{equation}
with $k_{abc}\in \mathbb{R}$ symmetric in the indices. Now let $U:=\mathbb{R}_{>0}\cdot \mathcal{H}\subset \mathbb{R}^{n}-\{0\}$ and $\overline{M}^{\text{cl}}:=\mathbb{R}^n+\I U\subset \mathbb{C}^n$ with canonical holomorphic coordinates $z^a:=b^a+\I t^a$. On $\overline{M}^{\text{cl}}$ we have a PSK metric by defining
\begin{equation}\label{8h:eq}
g_{\overline{M}^{\text{cl}}}:=\frac{\partial^2\mathcal{K}}{\partial z^a \partial \overline{z}^b}\mathrm{d}z^a\mathrm{d}\overline{z}^b\, \quad \; \mathcal{K}=-\log(8h(t))\,.
\end{equation}
The PSK manifold $(\overline{M}^{\text{cl}},g_{\overline{M}^{\text{cl}}})$ is associated to the CASK domain $(M^{\text{cl}},\mathfrak{F}^{\text{cl}})$ of signature $(n,1)$ given by
\begin{equation}
M^{\text{cl}}:=\{(Z^0,...,Z^n)=Z^0\cdot(1,z) \in \mathbb{C}^{n+1}\; | \; Z^0\in \mathbb{C}^{\times}, z\in \overline{M}^{\text{cl}}\},\, \quad \mathfrak{F}^{\text{cl}}:=-\frac{1}{6}k_{abc}\frac{Z^aZ^bZ^c}{Z^0}\,.
\end{equation}
Notice in particular the relation $Z^a/Z^0=z^a=b^a+\I t^a$, which we use repeatedly below.
The construction given above that associates the PSK manifold $(\overline{M}^{\text{cl}}, g_{\overline{M}^{\text{cl}}})$ to the PSR manifold $(\mathcal{H},g_{\mathcal{H}})$ is called the r-map. Furthermore, the QK metric obtained by applying the c-map (resp. tree-level c-map) to a PSK manifold in the image of the r-map is called a q-map space (resp. tree-level q-map space).
We now want to consider a tuple $(M,\mathfrak{F})$ of the following special type:
* We start with a PSR manifold $(\mathcal{H},g_{\mathcal{H}})$ and consider the associated CASK domain $(M^{\text{cl}},\mathfrak{F}^{\text{cl}})$. We further let $\Lambda^{+}:=\text{span}_{\mathbb{Z}_{\geq 0}}\{\gamma^a\}_{a=1}^{n}-\{0\}$ be a commutative semi-group generated by $\{\gamma^a\}_{a=1}^{n}$, where $n=\text{dim}(\mathcal{H})+1$. We want to consider a new CASK domain with a holomorphic prepotential of the form
\begin{equation}\label{prepotential}
\mathfrak{F}(Z^i)=\mathfrak{F}^{\text{cl}}(Z^i)+\mathfrak{F}^{\text{w.s.}}(Z^i)
\end{equation}
where $\mathfrak{F}^{\text{cl}}$ is as before, and
\begin{equation}\label{Fws:eq}
\mathfrak{F}^{\text{w.s.}}:=\chi\frac{(Z^0)^2\zeta(3)}{2(2\pi \I )^3}-\frac{(Z^0)^2}{(2\pi \I)^3}\sum_{\hat{\gamma}=q_a\gamma^a \in \Lambda^+}n_{\hat{\gamma}}\mathrm{Li}_3(e^{2\pi \I q_aZ^a/Z^0})\,\,.
\end{equation}
In the above expression $\chi \in \mathbb{Z}$, $n_{\hat{\gamma}}\in \mathbb{Z}$, $\mathrm{Li}_n(z)$ denotes the n-th polylogarithm, and $\zeta(x)$ denotes the Riemann zeta function.
* We assume that the $n_{\hat{\gamma}}$ are such that the $\mathbb{C}^{\times}$-invariant subset of $\mathbb{C}^{n+1}$ defined by
\begin{equation}\label{defM}
M^q:=M^{\text{cl}}\cap\{(Z^0,...,Z^n)\in \mathbb{C}^{n+1} \; | \; q_at^a>0 \; \text{for all $\hat{\gamma}=q_a\gamma^a\in \Lambda^+$ with $n_{\hat{\gamma}}\neq 0$}\}\;
\end{equation}
is open[This condition is automatic if the commutative semigroup generated by $\{\hat{\gamma}\in \Lambda^+ \; | \; n_{\hat{\gamma}}\neq 0\}$ is finitely generated.]. We further assume that the growth of the $n_{\hat{\gamma}}$ with $\hat{\gamma}$ is such that for any $\epsilon>0$ the series
\begin{equation}\label{convpropn}
\sum_{\hat{\gamma}=q_a\gamma^a\in \Lambda^+}|n_{\hat{\gamma}}|e^{-\epsilon q_at^a}
\end{equation}
has compact normal convergence on $M^q$, so that $\mathfrak{F}$ defines a holomorphic function on $M^q$.
In particular, the condition $q_at^a>0$ ensures that $\mathrm{Li}_3(e^{2\pi \I q_aZ^a/Z^0})$ is well defined and can be expressed as
\begin{equation}
\mathrm{Li}_3(e^{2\pi \I q_aZ^a/Z^0})=\sum_{k>0}\frac{e^{2\pi \I kq_aZ^a/Z^0}}{k^3}\,.
\end{equation}
* We denote by $M\subset M^q$ the maximal open subset of $M^q$ where $\text{Im}(\tau_{ij})=\text{Im}(\partial_i\partial_j\mathfrak{F})$ has signature $(n,1)$ and $\text{Im}(\tau_{ij})Z^i\overline{Z}^j<0$. These conditions are $\mathbb{C}^{\times}$-invariant, so that $(M,\mathfrak{F})$ defines a CASK domain on each connected component of $M$.
* To such a tuple $(M,\mathfrak{F})$ we associate the canonical lattice $\Gamma \to M$, together with the canonical central charge $Z$ (recall Section <ref>). Namely, if $Z_i:=\partial_{Z^i}\mathfrak{F}$ and $x^i=\text{Re}(Z^i)$, $y_i=\text{Re}(Z_i)$, then $(\partial_{x^i},\partial_{y_i})=(\widetilde{\gamma}_i,\gamma^i)$ defines a global Darboux frame for $\Gamma\to M$. For the canonical central charge we then have
\begin{equation}\label{Z:eq}
Z_{\gamma^i}=Z^i, \quad Z_{\widetilde{\gamma}_i}=-Z_i=-\frac{\partial \mathfrak{F}}{\partial Z^i}\,.
\end{equation}
We will identify the semigroups $\Lambda^+\cong \text{span}_{\mathbb{Z}_{\geq 0}}\{\partial_{y_a}\}_{a=1}^n-\{0\}$.
* We remark that given any $(Z^0,...,Z^n)=Z^0\cdot(1,b^a+\I t^a)\in M^q$, all points of the form $Z^0\cdot(1,b^a+\lambda\I t^a)$ with $\lambda>0$ must also be in $M^q$. In particular, for $\lambda>0$ sufficiently big we have $\text{Im}(\tau_{ij})\sim \text{Im}(\partial_i\partial_j\mathfrak{F}^{\text{cl}})$ due to the exponential decay of the terms with polylogarithms. Since $\text{Im}(\partial_i\partial_j\mathfrak{F}^{\text{cl}})$ has signature $(n,1)$ and $\text{Im}(\partial_i\partial_j\mathfrak{F}^{\text{cl}})Z^i\overline{Z}^j<0$, it follows that at the points $Z^0\cdot(1,b^a+\lambda\I t^a)$ for $\lambda>0$ sufficiently big we have that $\text{Im}(\tau_{ij})$ has signature $(n,1)$ and $\text{Im}(\tau_{ij})Z^i\overline{Z}^j<0$, so that the required $M$ is never empty, provided $M^q$ is not empty.
* We would also like to comment on the particular form of the prepotential (<ref>). In the setting of Calabi-Yau compactifications of type IIB string theory, the term $\mathfrak{F}^{\text{cl}}$ has $k_{abc}\in \mathbb{Z}$ equal to the triple interesection numbers of the Calabi-Yau threefold $X$, and corresponds to the holomorphic prepotential determining (via the c-map) the classical QK geometry of the hypermultiplet moduli space. On the other hand, $\mathfrak{F}^{\text{w.s.}}$
correspond to world-sheet instanton corrections. In such a setting, $\chi$ coincides with the Euler characteristic $\chi(X)$ of $X$, and the numbers $n_{\hat{\gamma}}$ are the so-called genus zero Gopakumar-Vafa invariants, and usually denoted by $n_{\hat{\gamma}}^{(0)}$. Since we are in a setting independent of string theory, we will use the simpler notation of $n_{\hat{\gamma}}$ for the coefficients appearing in (<ref>). Furthermore, it will be useful to define $n_{0}:=-\frac{\chi}{2}$. Then, using that $\mathrm{Li}_3(1)=\zeta(3)$ we can write
\begin{equation}\label{Fws}
\mathfrak{F}^{\text{w.s.}}=-\frac{(Z^0)^2}{(2\pi \I)^3}\sum_{\hat{\gamma} \in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\mathrm{Li}_3(e^{2\pi \I q_aZ^a/Z^0})\,.
\end{equation}
* In the physics literature sometimes an additional term of the form
\begin{equation}
\frac{1}{2}A_{ij}Z^iZ^j, \quad A_{ij}=A_{ji}\in \mathbb{R}
\end{equation}
is added to $\mathfrak{F}$ in (<ref>). While the addition of such a term does not alter the PSK metric $g_{\overline{M}}$ or the CASK metric $g_{M}$ (since $A_{ij}$ is real), the inclusion of such a term turns out to be important for mirror symmetry between D-branes of type IIA and type IIB string theory (see for example the review <cit.>, and the references therein). However, below we will focus on a case analogous to including D(-1) and D1 instanton corrections on type IIB, and we can safely ignore the inclusion of such a term.
Finally, after possibly restricting $M$, we assume that $M$ is the maximal open subset such that $(M,\Gamma,Z,\Omega)$ with $\Gamma$ and $Z$ as before and with
\begin{equation}\label{varBPS}
\begin{cases}
\Omega(q_0\gamma^0)=-\chi=2n_0, \quad q_0\in \mathbb{Z}-\{0\}\\
\Omega(q_0\gamma^0\pm q_a\gamma^a)=\Omega(\pm q_a\gamma^a)=n_{q_a\gamma^a} \quad \text{for $q_a\gamma^a \in \Lambda^+$, $q_0\in \mathbb{Z}$}\\
\Omega(\gamma)=0 \quad \text{else},\\
\end{cases}
\end{equation}
is a mutually local variation of BPS structures compatible (in the sense of Definition <ref>) with the CASK manifold $(M,g_M,\omega_M,\nabla,\xi)$ associated to $(M,\mathfrak{F})$. To check that $(M,\Gamma,Z,\Omega)$ defines a mutually local variation of BPS structures, one only needs to check the support property, since the convergence property follows easily from (<ref>), while the mutual locality and invariance under monodromy is obvious.
We remark that the BPS indices in (<ref>) are determined by the holomorphic prepotential $\mathfrak{F}$, see (<ref>). Such a prescription of BPS indices has previously appeared in the physics literature in [15, 11] or more explicitly in <cit.>. See also [34] (in particular, <cit.>), where such a BPS spectrum is conjectured for a non-compact Calabi-Yau 3-folds without compact divisors).
In the following, we will consider the associated instanton corrected QK manifold $(\overline{N},g_{\overline{N}})$ associated to $(M,\mathfrak{F})$ and $(M,\Gamma,Z,\Omega)$ as in Theorem <ref>. It will be convenient to lift the QK metric $(\overline{N},g_{\overline{N}})$ to $(\widetilde{N},g_{\overline{N}})$ as in Definition <ref>, where the $(\zeta^i,\widetilde{\zeta}_i,\sigma)$ directions are no longer periodic. In particular, we recall that we have the following:
* $\widetilde{N}\subset \mathbb{R}_{>0}\times \overline{M}\times \mathbb{R}^{2n+2}\times \mathbb{R}$ is an open subset, where the splitting is such that $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)\in \mathbb{R}_{>0}\times \mathbb{C}^n\times \mathbb{R}^{2n+2}\times \mathbb{R}$, where $\overline{M}$ is the associated PSK manifold.
* The twistor space $\mathcal{Z}$ of $(\widetilde{N},g_{\overline{N}})$ smoothly decomposes as $\mathcal{Z}=\widetilde{N}\times \mathbb{C}P^1$. In particular, the expression (<ref>) for the contact structure holds for the global coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma,t)$ of $\mathcal{Z}$.
Here we are slightly abusing notation by considering $t$, the identity map on $\mathbb{C}P^1$, as a global coordinate.)
If $(M,\mathfrak{F})$ and $(M,\Gamma,Z,\Omega)$ are given as before, we will call the resulting QK manifold $(\overline{N},g_{\overline{N}})$ (or its lift $(\widetilde{N},g_{\overline{N}})$) an instanton corrected q-map space.
The reason for the calling $(\overline{N},g_{\overline{N}})$ (or $(\widetilde{N},g_{\overline{N}})$) an instanton corrected q-map space is the following:
* By setting the one-loop parameter to $c_{\ell}=0$, and $n_{\hat{\gamma}}=0$ for all $\hat{\gamma}\in \Lambda^+\cup\{0\}$ (and hence, also $\Omega(\gamma)=0$ for all $\gamma$), one recovers a q-map space. That is the QK metric obtained by applying the c-map to a PSK manifold in the image of the r-map:
\begin{equation}\label{QKtree}
g_{\overline{N}}=g_{\overline{M}}+\frac{\D\rho^2}{4\rho^2}-\frac{1}{4\rho}(N^{ij}-2e^{\mathcal{K}}z^i\overline{z}^j)W_i\overline{W}_j + \frac{1}{64\rho^2}(\D\sigma + \widetilde{\zeta}_{i}\D\zeta^{i}-\zeta^{i}\D\widetilde{\zeta}_{i})^2\,.
\end{equation}
where $g_{\overline{M}}$, $N_{ij}$, $\mathcal{K}$ and $W_i$ are constructed in terms of $\mathfrak{F}^{\text{cl}}$ (see Section <ref> for the definition of the above terms).
* In the setting of Calabi-Yau compactification of type IIB string theory, the terms due to the BPS structure $(M,\Gamma,Z,\Omega)$ are thought as D(-1), D1 instanton corrections, and those due to $\mathfrak{F}^{\text{w.s.}}$ are thought as world-sheet instantons corrections. Hence, the QK metric obtained by the above $(M,\mathfrak{F})$ and $(M,\Gamma,Z,\Omega)$ can be thought as a q-map with the inclusion of the analog of the above corrections.
§.§ The quantum corrected mirror map and the S-duality action
In the sections that follow we will rescale the twistor fiber coordinate $t\to -\I t$. The contact structure on the twistor space is then expressed by (compare with (<ref>))
\begin{equation}\label{contstr}
\lambda=\left(f\frac{\D t}{t}+t^{-1}\I \theta_+^P|_{\overline{N}} -2\I \theta_3^P|_{\overline{N}} -t\I \theta_-^P|_{\overline{N}}\right)\cdot s\,.
\end{equation}
In order to define the S-duality action, we consider the following diffeomorphism, first defined (under slightly different conventions) in [15]:
Let $\overline{M}:=\{z \in \mathbb{C}^n \; | \; (1,z)\in M\}$ where $M$ was given in the previous section, and
consider the manifold $\overline{\mathcal{N}}_{\text{IIA}}:=\mathbb{R}_{>-c_{\ell}}\times \overline{M}\times \mathbb{R}^{2n+2}\times \mathbb{R}$ with global coordinates $(\rho,z^a,\zeta^i,\widetilde{\zeta}_i,\sigma)$. We will call such coordinates type IIA coordinates. On the other hand, if $H\subset \mathbb{C}$ denotes the upper half-plane, we define $\overline{\mathcal{N}}_{\text{IIB}}:=H\times \overline{M}\times \mathbb{R}^{2n}\times \mathbb{R}^2$ with global coordinates $(\tau_1+\mathrm{i}\tau_2,b^a+\I t^a,c^a,c_a,c_0,\psi)$. We call the latter type IIB coordinates. The type IIB coordinates are related to the type IIA coordinates via the diffeomorphism (see Remark <ref> below) $\mathcal{M}: \overline{\mathcal{N}}_{\text{IIB}}\to \overline{\mathcal{N}}_{\text{IIA}}$ defined by
\begin{equation}\label{MM}
\begin{split}
z^a&=b^a+\I t^a, \;\;\;\;\; \rho=\frac{\tau_2^2}{16}e^{-\mathcal{K}}-c_{\ell}, \;\;\;\;\;\;\; \zeta^0=\tau_1, \;\;\;\;\; \zeta^a=-(c^a-\tau_1b^a),\\
\widetilde{\zeta}_a&=c_a +\frac{k_{abc}}{2}b^b(c^c-\tau_1b^c)+\widetilde{\zeta}_{a}^{\text{inst}}, \;\;\;\;\;\; \widetilde{\zeta}_0=c_0-\frac{k_{abc}}{6}b^ab^b(c^c-\tau_1b^c)+\widetilde{\zeta}_{0}^{\text{inst}}\\
\sigma&= -2(\psi + \frac{1}{2}\tau_1c_0) + c_a(c^a-\tau_1b^a) -\frac{k_{abc}}{6}b^ac^b(c^c-\tau_1b^c)+\sigma^{\text{inst}}\,,
\end{split}
\end{equation}
where $\mathcal{K}=-\log(-2\text{Im}(\tau_{ij})z^i\overline{z}^j)$ with $z^0=1$, $\tau_{ij}=\frac{\partial^2 \mathfrak{F}}{\partial Z^i\partial Z^j}$, $c_{\ell}\in \mathbb{R}$ is the 1-loop parameter, and
\begin{equation}\label{typeiiacoordcor}
\begin{split}
\widetilde{\zeta}^{\text{inst}}_{a}&:=\frac{1}{8\pi^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}q_a\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\frac{m\tau_1+ n}{m|m\tau+ n|^2}e^{-S_{\hat{\gamma},m,n}}\\
\widetilde{\zeta}^{\text{inst}}_0&:=\frac{\I}{16\pi^3}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\left(\frac{(m\tau_1+ n)^2}{|m\tau + n|^3}+ 2\pi q_a\left(t^a+\I b^a\frac{m\tau_1+n}{|m\tau+ n|}\right)\right)\frac{e^{-S_{\hat{\gamma},m,n}}}{m|m\tau+ n|}\\
\sigma^{\text{inst}}&:=\tau_1\widetilde{\zeta}_0^{\text{inst}}-(c^a-\tau_1b^a)\widetilde{\zeta}_a^{\text{inst}}-\frac{\I\tau_2^2}{8\pi^2}\sum_{\hat{\gamma}\in \Lambda_{+}}n_{\hat{\gamma}}q_at^a\sum_{n\in \mathbb{Z}-\{0\}}\frac{e^{-S_{\hat{\gamma},0,n}}}{n|n|}\\
&\quad\quad +\frac{\I}{8\pi^3}\sum_{\hat{\gamma}\in \Lambda^{+}}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\left(2-\frac{(m\tau_1+n)^2}{|m\tau+ n|^2}\right)\frac{(m\tau_1+ n)e^{-S_{\hat{\gamma},m,n}}}{m^2|m\tau+ n|^2}
\end{split}
\end{equation}
\begin{equation}\label{mnaction}
S_{\hat{\gamma},m,n}:=2\pi q_a(|m\tau+ n|t^a +\I mc^a+ \I nb^a)\,.
\end{equation}
We will refer to the diffeomorphism $\mathcal{M}$ as the quantum corrected mirror map.
* Note that since $z=(z^a)\in \overline{M}$ and $M\subset M^{q}$, we have $\text{Re}(S_{\hat{\gamma},m,n})>0$ for all $\hat{\gamma}\in \Lambda^{+}$ with $n_{\hat{\gamma}}\neq 0$, so (<ref>) implies that the sums in (<ref>) have compact normal convergence on $\overline{\mathcal{N}}_{\text{IIB}}$. Furthermore, we remark that (<ref>) really defines a diffeomorphism. Indeed, since $-\text{Im}(\tau_{ij})z^i\overline{z}^j>0$ on $\overline{M}$ and $\tau_2>0$, we can always invert the relation involving $\rho$, while $\widetilde{\zeta}_i^{\text{inst}}$ and $\sigma^{\text{inst}}$ only depend on $\tau=\tau_1+\mathrm{i}\tau_2$, $z^a$ and $c^a$, so it is easy to invert all the other relations. Note, in particular, that the expressions for $\widetilde{\zeta}_i^{\text{inst}}$ and $\sigma^{\text{inst}}$ are real. Furthermore, if we set $c_{\ell}=0$, $n_{\hat{\gamma}}=\chi=0$ for all $\hat{\gamma}\in \Lambda^{+}$, then we recover the classical mirror map.
* While the expressions (<ref>) look complicated, they satisfy nice transformation properties with respect to the expected isometry groups of the metric. See Section <ref>, Lemma <ref> and Corollary <ref>.
Let $\overline{\mathcal{N}}_{\text{IIB}}^{\text{cl}}:=H\times \overline{M}^{\text{cl}}\times \mathbb{R}^{2n}\times \mathbb{R}^2$ with global type IIB variables $(\tau=\tau_1+\I \tau_2,b^a+\I t^a,c^a,c_a,c_0,\psi)$. We define an $\mathrm{SL}(2,\mathbb{Z})$-action on $\overline{\mathcal{N}}_{\text{IIB}}^{\text{cl}}$ by
\begin{equation}\label{sl2can}
\begin{split}
\tau&\to \frac{a\tau +b}{c\tau +d},\;\;\;\;\; t^a \to |c\tau +d|t^a, \;\;\;\; c_a \to c_a, \\
\begin{pmatrix}c^a\\ b^a\end{pmatrix}&\to \begin{pmatrix} a & b \\ c & d\\ \end{pmatrix}\begin{pmatrix}c^a \\ b^a\end{pmatrix}, \;\;\;\; \begin{pmatrix}c_0\\ \psi\end{pmatrix}\to \begin{pmatrix} d & -c \\ -b & a\\ \end{pmatrix}\begin{pmatrix}c_0 \\ \psi \end{pmatrix}, \;\;\;\; \begin{pmatrix} a & b \\ c & d\\ \end{pmatrix}\in \mathrm{SL}(2,\mathbb{Z})
\end{split}
\end{equation}
We call this action the S-duality action. We note that there is the possibility that S-duality does not act on $\overline{\mathcal{N}}_{\text{IIB}}\subset \overline{\mathcal{N}}_{\text{IIB}}^{\text{cl}}$, since it may happen that $\overline{M}$ is not invariant under the scaling of the $t^a$ in (<ref>).
Recall from Section <ref> that the quaternionic Kähler manifold $\overline{N}$ was constructed in [25] via HK/QK correspondence (by specializing [6]) as the hypersurface
of the bundle $P\to N$ which is
defined by the equation $\mathrm{Arg}\, Z^0 =0$.
Using the IIB variables we can now simply match $|Z^0|=\tau_2$.
§.§ Type IIB Darboux coordinates
We now want to write a distinguished set of Darboux coordinates which will be useful for studying S-duality or the action by specific elements of $\mathrm{SL}(2,\mathbb{Z})$. The coordinates below have been previously written in [15]. Nevertheless, our approach will be different from [15] in the sense that we will start from the mathematical construction of the instanton QK metric obtained in [25], and then explicitly show that (<ref>) indeed define Darboux coordinates for the contact structure on its twistor space.
Recalling (<ref>) and (<ref>), we let
\begin{equation}
\widetilde{\zeta}_i^{\text{cl}}:=\widetilde{\zeta}_i-\widetilde{\zeta}_i^{\text{inst}}, \quad \sigma^{\text{cl}}:=\sigma - \sigma^{\text{inst}}\,,
\end{equation}
and define
\begin{equation}\label{dcclassical}
\begin{split}
\xi^{i,\text{cl}}&:=\zeta^{i}+ R(t^{-1}z^i -t\overline{z}^i)\\
\widetilde{\xi}_i^{\text{cl}}&:=\widetilde{\zeta}_{i}^{\text{cl}}+ R(t^{-1}F^{\text{cl}}_{i}-t\overline{F}^{\text{cl}}_i)\\
\alpha^{\text{cl}}&:= \sigma^{\text{cl}} +R(t^{-1}\langle \widetilde{Z}^{\text{cl}},\zeta^{\text{cl}} \rangle -t \langle \overline{\widetilde{Z}}^{\text{cl}}, \zeta^{\text{cl}} \rangle)\,,
\end{split}
\end{equation}
where $\widetilde{Z}^{\text{cl}}:= Z^{\text{cl}}/Z^0$ and $Z^{\text{cl}}$ is defined replacing
$Z$ and $\mathfrak F$ in equation (<ref>) by $Z^{\text{cl}}$ and $\mathfrak F^{\text{cl}}$, and $F_i^{\text{cl}}:=\partial_{Z^i}\mathfrak{F}^{\text{cl}}/Z^0$.
The expressions (<ref>) match the coordinates (<ref>) for the case $\mathfrak{F}=\mathfrak{F}^{\text{cl}}$ after the scaling $t \to -\mathrm{i}t$ done in (<ref>) and setting $c_{\ell}=0$. In particular, they define Darboux coordinates for the contact structure of twistor space of the tree-level q-map space defined by $\mathfrak{F}^{\text{cl}}$.
Finally, if $c\in \mathbb{R}-\{0\}$ and $d\in \mathbb{R}$, we denote by $t_{\pm}^{c,d}$ the roots of $t(c\xi^{0,\text{cl}}+d)=0$ in the variable $t$. Using that $\zeta^0=\tau_1$ and $R=2\sqrt{\rho+c_{\ell}}e^{\mathcal{K}/2}=\tau_2/2$ (recall (<ref>)), we find
\begin{equation}\label{troots}
t_{\pm}^{c,d}=\frac{c\tau_1+d \pm |c\tau +d|}{c\tau_2}\,.
\end{equation}
Consider an instanton corrected q-map space $(\widetilde{N},g_{\overline{N}})$ with $1$-loop parameter $c_{\ell}=\frac{\chi}{192\pi}$. Then the functions $(\xi^i,\widetilde{\xi}_{i},\alpha)$ of the associated twistor space $\mathcal{Z}\cong \widetilde{N}\times \mathbb{C}P^1$ given by
\begin{align*}
\xi^i&=\xi^{i,\text{cl}}\\
\widetilde{\xi}_{a}&=\widetilde{\xi}_a^{\text{cl}}+\frac{\tau_2}{8\pi^2}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}q_a\left(\sum_{(m,n)\in \mathbb{Z}^2-\{0\}}\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau+ n|^2}\frac{1+t_+^{m,n}t}{t-t_{+}^{m, n}}\right)\\
\widetilde{\xi}_0&=\widetilde{\xi}_0^{\text{cl}}+\frac{\I \tau_2}{16\pi^3}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\sum_{(m,n)\in \mathbb{Z}^2-\{0\}}\left(\frac{1}{m\xi^0+ n}+\frac{m\tau_1+ n}{|m\tau+ n|^2}\right)\frac{1+t_{+}^{m, n}t}{t-t_{+}^{m, n}}\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau + n|^2}\\
&\quad-\frac{\tau_2}{8\pi^2}\sum_{\hat{\gamma}\in \Lambda^{+}\cup \{0\}}n_{\hat{\gamma}} \sum_{(m,n)\in \mathbb{Z}^2-\{0\}}\left(q_ab^a \frac{1+t_{+}^{m, n}t}{t-t_{+}^{m, n}}+\I q_at^a\frac{1-t_{+}^{m, n}t}{t-t_{+}^{m, n}}\right)\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau+ n|^2}\\
\alpha&=-\frac{1}{2}(\alpha^{\text{cl}}-\xi^{i}\widetilde{\xi}_i^{\text{cl}})+\frac{\I \tau_2^2}{32\pi^3}\sum_{\hat{\gamma}\in \Lambda^{+}\cup\{0\}}n_{\hat{\gamma}}\sum_{(m,n)\in \mathbb{Z}^2-\{0\}}\left((m\tau_1+ n)(t^{-1}-t) -2m\tau_2\right)\frac{1+t_{+}^{m, n}t}{t-t_{+}^{m, n}}\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau+n|^4}\,, \numberthis \label{typeiibdc}
\end{align*}
define Darboux coordinates for the contact structure of $\mathcal{Z}$, in the sense that
\begin{equation}\label{contdarbiib}
\lambda=4\pi \I\left(\D\alpha -\widetilde{\xi}_i\D \xi^i\right)\cdot s\,.
\end{equation}
In the above expressions, for the sums in the case where $m=0$ we set
\begin{equation}\label{m=0def}
\frac{1+t_{+}^{0, n}t}{t-t_{+}^{0, n}}:=\begin{cases}
1/t, \quad n<0\\
-t, \quad n>0
\end{cases}, \quad \quad \quad \quad \frac{1-t_{+}^{0, n}t}{t-t_{+}^{0, n}}:=\begin{cases}
1/t, \quad n<0\\
t, \quad n>0\,.
\end{cases}
\end{equation}
This will be proven in Section <ref> below.
* The definitions (<ref>) above for the case $m=0$ (where $t_{\pm}^{m,n}$ is not defined) can be motivated as follows. If $n>0$ then it is easy to check that
\begin{equation}
\lim_{c\to 0}\frac{1+t_{+}^{c, n}t}{t-t_{+}^{c, n}}=-t\,, \quad \lim_{c\to 0}\frac{1-t_{+}^{c, n}t}{t-t_{+}^{c, n}}=t.
\end{equation}
On the other hand, by using that for $c\neq 0$ we have $t_{+}^{c,d}t_{-}^{c,d}=-1$, one similarly finds for $n<0$
\begin{equation}
\lim_{c\to 0}\frac{t_-^{c,n}}{t_-^{c,n}}\left(\frac{1+t_{+}^{c, n}t}{t-t_{+}^{c, n}}\right)= \lim_{c\to 0}\frac{t_-^{c,n}-t}{t_-^{c,n}t+1}=\frac{1}{t}\,, \quad\quad \lim_{c\to 0}\frac{t_-^{c,n}}{t_-^{c,n}}\left(\frac{1-t_{+}^{c, n}t}{t-t_{+}^{c, n}}\right)=\lim_{c\to 0}\frac{t_-^{c,n}+t}{t_-^{c,n}t+1}=\frac{1}{t}.
\end{equation}
* As previously mentioned in the introduction, one can relate the Darboux coordinates (<ref>) to the Darboux coordinates (<ref>) by Poisson resumming the later and applying a contact transformation, as done in [15]. We will follow a similar approach by Poisson resumming the contact structure directly, and then checking that (<ref>) give Darboux coordinates.
Using (<ref>) it is not hard to see that the coordinates are well defined on an open dense subset of $\mathcal{Z}\cong \widetilde{N}\times \mathbb{C}P^1$, namely the points
\begin{equation}
\{(\tau,b^a+\I t^a, c^a,c_a,c_0,\psi, t)\in \mathcal{Z} \; | \; t\neq 0, \infty, \;\; t\not\in \{t_{+}^{m,n}(\tau)\}_{m\in \mathbb{Z}-\{0\}, n \in \mathbb{Z}}\}\,,
\end{equation}
where we have used the quantum corrected mirror map (<ref>) to put type IIB coordinates on $\widetilde{N}\subset \overline{\mathcal{N}}_{\text{IIA}}$.
In subsequent sections, we will also frequently use the notation
\begin{equation}
\widetilde{\xi}_i^{\text{inst}}:=\widetilde{\xi}_i-\widetilde{\xi}_i^{\text{cl}}, \quad \alpha^{\text{inst}}:=\alpha + \frac{1}{2}(\alpha^{\text{cl}}-\xi^{i}\widetilde{\xi}_i^{\text{cl}})\,.
\end{equation}
When comparing (<ref>) and (<ref>) with the formulas appearing in [15], we remark that we are using different conventions for the definition of $t_{\pm}^{c,d}$ and $S_{\hat{\gamma},m,n}$. Namely, what we call $S_{\hat{\gamma},m,n}$ in their notation is $S_{\hat{\gamma},-m,-n}$ and $t_{\pm}^{c,d}$ in their notation is $t_{\mp}^{c,d}$.
As an immediate corollary of Theorem <ref>, we obtain the following
Consider an instanton corrected q-map space $(\widetilde{N},g_{\overline{N}})$ with one-loop parameter $c_{\ell}\in \mathbb{R}$. If $(\xi^i,\widetilde{\xi}_i,\alpha)$ are the functions on $\mathcal{Z}$ defined by (<ref>), then if $\alpha_{c_{\ell}}:=\alpha+4\I \left(c_{\ell}-\frac{\chi}{192\pi}\right)\log(t)$, the functions $(\xi^i,\widetilde{\xi}_{i},\alpha_{c_{\ell}})$ of the associated twistor space are Darboux coordinates for the contact structure $\lambda$ of $\mathcal{Z}$, in the sense that
\begin{equation}
\lambda=4\pi \I\left(\D\alpha_{c_{\ell}} -\widetilde{\xi}_i\D \xi^i\right)\cdot s\,.
\end{equation}
Notice that since the one loop parameter only enters in the contact structure via $f$ by (<ref>) and $\rho=2\pi r^2 -c_{\ell}$, by writing $f=(f+16\pi (c_{\ell}-\frac{\chi}{192\pi}))-16\pi (c_{\ell}-\frac{\chi}{192\pi}) $ we have by Theorem <ref> that
\begin{equation}
4\pi \I\left(\D\alpha -\widetilde{\xi}_i\D \xi^i\right)=\left(f+16\pi \left(c_{\ell}-\frac{\chi}{192\pi}\right)\right)\frac{\D t}{t}+t^{-1}\I \theta_+^P|_{\overline{N}} -2\I \theta_3^P|_{\overline{N}} -t\I \theta_-^P|_{\overline{N}}\,.
\end{equation}
It then follows immediately that
\begin{equation}
4\pi \I\left(\D\alpha_{c_{\ell}} -\widetilde{\xi}_i\D \xi^i\right)=f\frac{\D t}{t}+t^{-1}\I \theta_+^P|_{\overline{N}} -2\I \theta_3^P|_{\overline{N}} -t\I \theta_-^P|_{\overline{N}}\,,
\end{equation}
compare (<ref>).
§.§.§ Preliminary lemmas
In this section we give some preliminary lemmas and expressions that will be useful to prove Theorem <ref>.
We will divide each of the $1$-forms $\theta_{+}^{P}|_{\overline{N}}$, $\theta_3^P|_{\overline{N}}$ appearing in the contact structure (<ref>) as follows:
\begin{equation}
\theta_{+}^P|_{\overline{N}}=\theta_{+}^{P, \text{cl}}+\theta_{+}^{P, \text{w.s.}}+\theta_{+}^{P, \text{inst}}, \quad\quad \theta_{3}^P|_{\overline{N}}=\theta_{3}^{P, \text{cl}}+\theta_{3}^{P, \text{w.s.}}+\theta_{3}^{P, \text{inst}},
\end{equation}
where using that $R=2\sqrt{\rho+c_{\ell}}e^{\mathcal{K}/2}=\tau_2/2$, the decompositions $F_i=F_i^{\text{cl}}+F_i^{\text{ws}}$ where $F_i^{\text{cl}}:=\partial_{Z^i}\mathfrak{F}^{\text{cl}}/Z^0$ and $F_i^{\text{w.s.}}:=\partial_{Z^i}\mathfrak{F}^{\text{w.s.}}/Z^0$, $\widetilde{\zeta}_i=\widetilde{\zeta}_i^{\text{cl}}+\widetilde{\zeta}_i^{\text{inst}}$ and $\sigma=\sigma^{\text{cl}}+\sigma^{\text{inst}}$, and (<ref>) we have
\begin{equation}\label{theta+split}
\begin{split}
\theta_{+}^{P,\text{cl}}&:=-2\pi\tau_2\left(F_i^{\text{cl}}\D \zeta^i-z^i\D \widetilde{\zeta}_i^{\text{cl}}\right)\\
\theta_{+}^{P,\text{w.s.}}&:=-2\pi\tau_2\left(F_i^{\text{w.s.}}\D \zeta^i-z^i\D \widetilde{\zeta}_i^{\text{inst}}\right)\\
\theta_{+}^{P,\text{inst}}&=\I \tau_2\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi\I n\zeta_{\gamma}}K_0(4\pi Rn|\widetilde{Z}_{\gamma}|)\D \zeta_{\gamma}\\
&\quad+\frac{\tau_2^2}{2}\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi \I n\zeta_{\gamma}}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\left(\frac{\D \widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}+\frac{\D \overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}+\frac{2}{\tau_2}\D \tau_2\right).\\
\end{split}
\end{equation}
For $\theta_3^P|_{\overline{N}}$ we use the relation
\begin{equation}
\D^c\mathcal{K}=e^{\mathcal{K}}(-2\text{Im}(\tau_{ij}))\Big(\I \overline{z}^j\D z^i -\I z^i\D \overline{z}^j\Big),
\end{equation}
so that
\begin{equation}
4\pi(\rho+c_{\ell})\D ^c\mathcal{K}=\pi R^2(-2\text{Im}(\tau_{ij}))\Big(\I\overline{z}^j\D z^i -\I z^i\D \overline{z}^j\Big)=\pi\tau_2^2\mathrm{Im}(\tau_{0a})\D t^a +\pi\tau_2^2\mathrm{Im}(\tau_{ab})(b^a\D t^b-t^a\D b^b),
\end{equation}
and hence we can split $\theta_3^P|_{\overline{N}}$ as follows using (<ref>)
\begin{equation}\label{theta3split}
\begin{split}
\theta_{3}^{P,\text{cl}}&:=\pi\D \sigma^{\text{cl}} +\pi\left(\widetilde{\zeta}_i^{\text{cl}}\D \zeta^i-\zeta^i\D \widetilde{\zeta}_i^{\text{cl}}\right)-\pi\tau_2^2\mathrm{Im}(\tau_{0a}^{\text{cl}})\D t^a -\pi\tau_2^2\mathrm{Im}(\tau_{ab}^{\text{cl}})(b^a\D t^b-t^a\D b^b)\\
\theta_{3}^{P,\text{w.s.}}&:=\pi\D \sigma^{\text{inst}} +\pi\left(\widetilde{\zeta}_i^{\text{inst}}\D \zeta^i-\zeta^i\D \widetilde{\zeta}_i^{\text{inst}}\right)-\pi\tau_2^2\mathrm{Im}(\tau_{0a}^{\text{w.s.}})\D t^a -\pi\tau_2^2\mathrm{Im}(\tau_{ab}^{\text{w.s.}})(b^a\D t^b-t^a\D b^b)\\
\theta_3^{P,\text{inst}}&=-\frac{\I\tau_2}{4\pi}\sum_{\gamma}\Omega(\gamma)\sum_{n>0}\frac{e^{-2\pi\I n\zeta_{\gamma}}}{n}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\left(\frac{\D \widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}-\frac{\D \overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}\right)\,,
\end{split}
\end{equation}
where $\tau_{ij}^{\text{cl}}=\partial_{Z^i}\partial_{Z^j}\mathfrak{F}^{\text{cl}}$ and $\tau_{ij}^{\text{w.s.}}=\partial_{Z^i}\partial_{Z^j}\mathfrak{F}^{\text{w.s.}}$.
Similarly, using that $f=16\pi \rho+f^{\text{inst}}$ and (using the second relation in (<ref>))
\begin{equation}\label{rhodef}
\rho=\frac{\tau_2^2}{16}e^{-\mathcal{K}}-c_{\ell}= \frac{\tau_2^2}{16}K-c_{\ell}, \quad K=-2\text{Im}(\tau_{ij})z^i\overline{z}^j\,,
\end{equation}
we can decompose $f$ as
\begin{equation}\label{fsplitting2}
f=f^{\text{cl}}+f^{\text{w.s.}}-16\pi c_{\ell}+f^{\text{inst}},
\end{equation}
\begin{equation}\label{fsplitting}
\begin{split}
f^{\text{w.s.}}&:=-2\pi\tau_2^2\text{Im}(\tau_{ij}^{\text{w.s.}})z^i\overline{z}^j=\frac{\tau_2^2}{2\pi^2}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\text{Re}\left(\mathrm{Li}_3(e^{2\pi \I q_az^a})+2\pi q_at^a\mathrm{Li}_2(e^{2\pi \I q_az^a})\right)\\
f^{\text{inst}}&=\frac{\tau_2}{\pi}\sum_{\gamma}\Omega(\gamma)\sum_{n>0}\frac{e^{-2\pi\I n\zeta_{\gamma}}}{n}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\,.\\
\end{split}
\end{equation}
In order to prove Theorem <ref>, we start with the following lemma, which will allow us to rewrite the Bessel function expressions of the instanton terms $f^{\text{inst}}$, $\theta_{\pm}^{P,\text{inst}}$ and $\theta_3^{P,\text{inst}}$ in terms of expressions similar to those appearing in (<ref>).
Given $\hat{\gamma}\in \Lambda^{+}$ with $n_{\hat{\gamma}}\neq 0$, and $\nu \in \mathbb{N}$, we define
\begin{equation}\label{Idef}
\begin{split}
\mathcal{I}_{\hat{\gamma}}^{(\nu)}&:=2\sum_{q_0\in \mathbb{Z}}\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{-2\pi\I ns\zeta_{q_0\gamma^0+\hat{\gamma}}}}{(sn)^{\nu}}K_0(4\pi Rn|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|)\\
\mathcal{I}_0^{(\nu)}&:=2\sum_{q_0\in \mathbb{Z}-\{0\}}\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{-2\pi\I ns\zeta_{q_0\gamma^0}}}{(sn)^\nu}K_0(4\pi Rn|\widetilde{Z}_{q_0\gamma^0}|)\,.
\end{split}
\end{equation}
Notice that in $\mathcal{I}_{0}^{(\nu)}$ we exclude the term $q_0=0$ from the sum, since $K_0(x)$ diverges at $x=0$. Furthermore, since $\hat{\gamma}\in \Lambda^{+}$ with $n_{\hat{\gamma}}\neq 0$ implies that $\pm (q_0\gamma^0+\hat{\gamma})\in \text{Supp}(\Omega)$ for all $q_0\in \mathbb{Z}$, by the support property of variations of BPS structures one finds that $|Z_{q_0\gamma^0+\hat{\gamma}}|\neq 0$, so that $|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|\neq 0$ on $\widetilde{N}$ (or $\overline{N})$. By the exponential decay of $K_0(x)$ as $x\to \infty$, it is easy to check that the convergence of the series $\mathcal{I}_{\hat{\gamma}}^{(\nu)}$ and $\mathcal{I}_{0}^{(\nu)}$ is compact normal on $\widetilde{N}$. In particular, they define smooth functions on $\widetilde{N}$ and we can interchange sums with differentiation.
The function $\mathcal{I}_{\hat{\gamma}}^{(\nu)}$ can be expressed as
\begin{equation}\label{Presum1}
\mathcal{I}_{\hat{\gamma}}^{(\nu)}=\sum_{n\in \mathbb{Z}}\sum_{m\in \mathbb{Z}-\{0\}}\frac{e^{-S_{\hat{\gamma},m,n}}}{m^{\nu}|m\tau + n|},
\end{equation}
where $S_{\hat{\gamma},m,n}$ was defined in (<ref>). Furthermore, we have for $\nu \geq 2$
\begin{equation}\label{I0der}
\partial_{\tau_2}\mathcal{I}_{0}^{(\nu)}=\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(\nu)}|_{\hat{\gamma}=0}+\frac{2}{\tau_2}\sum_{s=\pm 1}\sum_{n>0}\frac {1}{(sn)^{\nu}}, \quad \partial_{\tau_1}\partial_{\tau_2}\mathcal{I}_{0}^{(\nu)}=\partial_{\tau_1}\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(\nu)}|_{\hat{\gamma}=0}, \quad \partial_{\tau_1}^2\mathcal{I}_{0}^{(\nu)}=\partial_{\tau_1}^2\mathcal{I}_{\hat{\gamma}}^{(\nu)}|_{\hat{\gamma}=0}\,,
\end{equation}
where $\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(\nu)}|_{\hat{\gamma}=0}$ (and similarly for the other derivatives) means taking the derivative of (<ref>) and then evaluating at $\hat{\gamma}=0$.
We will use a similar notation to <cit.> and follow the same idea of Poisson resummation, but in an easier case than the one presented in <cit.>. We denote $x=q_ab^a$, $y=q_at^a$ and $\Theta:=q_a(\zeta^a-b^a\zeta^0)=-q_ac^a$, so that we can write
\begin{equation}\label{Poisson1}
\mathcal{I}_{\hat{\gamma}}^{(\nu)}=\sum_{q_0\in \mathbb{Z}}f(x+q_0,y,\zeta^i,R)
\end{equation}
\begin{equation}
f(x,y,\zeta^i,R):=2\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{2\pi\I ns(\Theta+x\zeta^0)}}{(sn)^{\nu}}K_0(4\pi Rn|x+\I y|)\,.
\end{equation}
We have used above that $\zeta_{q_i\gamma^i}=-q_i\zeta^i$ (recall Definition <ref>).
Equation (<ref>) makes it clear that $\mathcal{I}_{\hat{\gamma}}^{(\nu)}$ is a function that is invariant under integer shifts in the $x$ variable, and the idea is to now compute the Poisson resummation with respect to the $x$-variable. Namely,
\begin{equation}\label{Prformula}
\mathcal{I}_{\hat{\gamma}}^{(\nu)}=\sum_{q_0\in \mathbb{Z}}f(x+q_0,y,\zeta^i,R)=\sum_{q_0\in \mathbb{Z}}\hat{f}(q_0,y,\zeta^i,R)e^{2\pi \I q_0x},
\end{equation}
\begin{equation}
\hat{f}(w)=\int_{-\infty}^{\infty}\D x\; f(x)e^{-2\pi \I xw}
\end{equation}
denotes the Fourier transform (from now on, we omit the dependence of $f$ of the variables $y$, $\zeta^i$ and $R$ from the notation).
Using the integral representation (<ref>) of $K_0$ we have
\begin{equation}
\begin{split}
K_0(4\pi Rn|x+\I y|)&=\int_{0}^{\infty}\D t\exp(-4\pi Rn|x+\I y|\cosh(t))=\frac{1}{2}\int_{-\infty}^{\infty}\D t\exp(-4\pi Rn|x+\I y|\cosh(t))\\
&=\frac{1}{2}\int_{0}^{\infty}\frac{\D z}{z}\exp(-2\pi Rn|x+\I y|(z^{-1}+z))\,.
\end{split}
\end{equation}
Furthermore, since the integral exponentially
decays at both ends, provided $x\neq 0$, we can deform
the integration path by changing $z$ to $z\to z\frac{|x+\I y|}{x+\I y}\text{sign}(x)$, obtaining
\begin{equation}
K_0(4\pi Rn|x+\I y|)=\frac{1}{2}\int_{0}^{\infty}\frac{\D z}{z}\exp(-2\pi Rn\cdot\text{sign}(x)((x+\I y)z^{-1}+(x-\I y)z))\,.
\end{equation}
In particular, we have
\begin{align*}
\hat{f}(w)&=\int_{-\infty}^{\infty}\D x\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{2\pi \I ns\Theta}}{(sn)^{\nu}}\int_{0}^{\infty} \frac{\D z}{z}\exp(- 2\pi nR\cdot\text{sign}(x)((x+\I y)z^{-1}+(x-\I y)z))e^{2\pi\I x(- w+ns\zeta^0)}\\
&=\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{ 2\pi\I ns\Theta}}{(sn)^{\nu}}\int_{0}^{\infty} \frac{\D z}{z}\int_{-\infty}^{\infty}\D x\exp(- 2\pi nR\cdot\text{sign}(x)((x+\I y)z^{-1}+(x-\I y)z))e^{2\pi\I x(- w+ns\zeta^0)}\\
&=\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{ 2\pi\I ns\Theta}}{(sn)^{\nu}}\int_{0}^{\infty} \frac{\D z}{z}\int_{0}^{\infty}\D x\;e^{2\pi x(-nRz^{-1}-nRz -\I w+\I ns\zeta^0)+2\pi y(-\I nRz^{-1}+\I nRz)}\\
&\quad +\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{2\pi \I ns\Theta}}{(sn)^{\nu}}\int_{0}^{\infty} \frac{\D z}{z}\int_{0}^{\infty}\D x\;e^{2\pi x(-nRz^{-1}-nRz + \I w -\I ns\zeta^0) + 2\pi y(\I nRz^{-1}-\I nRz)}\\
&=\frac{1}{2\pi}\sum_{s=\pm 1}\sum_{n=1}^{\infty}\frac{e^{2\pi \I ns\Theta}}{(sn)^{\nu}}\left(\int_{0}^{\infty} \frac{\D z}{z}\frac{e^{2\pi y(-\I nRz^{-1}+\I nRz)}}{nRz^{-1}+nRz +\I w-\I ns\zeta^0}+\int_{0}^{\infty} \frac{\D z}{z}\frac{e^{2\pi y(\I nRz^{-1}-\I nRz)}}{nRz^{-1}+nRz -\I w+\I ns\zeta^0}\right). \numberthis \\
\end{align*}
We can combine the $s=1$ and $s=-1$ terms of the previous sum into an integral over $\mathbb{R}$ and a sum over $n\in \mathbb{Z}-\{0\}$ to obtain
\begin{equation}
\hat{f}(w)=\frac{1}{2\pi}\sum_{n\in \mathbb{Z}-\{0\}}\frac{e^{2\pi \I n\Theta}}{n^{\nu-1}|n|}\int_{-\infty}^{\infty}\frac{\D z}{z}\frac{e^{-2\pi\I nRy(z^{-1}-z)}}{nR(z^{-1}+z) + \I w-\I n\zeta^0}\,.
\end{equation}
The integrand of the $n$-th term of $\hat{f}(m)$ with $m\in \mathbb{Z}$ has poles at $z=\I t_{\pm}^{n,- m}$ where $t_{\pm}^{c,d}$ was defined before in (<ref>) as the roots in $t$ of $t(c\xi^0(t)+d)=0$. Since one of our defining conditions on the manifold $M$ is that $\text{sign}(y)=\text{sign}(q_at^a)>0$ for $n_{\hat{\gamma}}\neq 0$ (see (<ref>)), we can compute the previous integral by closing the contour in the upper half plane when $n>0$, and in the lower half plane if $n<0$. Independently of the sign of $n$, only the pole at $\mathrm{i}t_{+}^{n,-m}$ contributes to the $n$-th integral. In particular, we obtain
\begin{equation}
\begin{split}
\hat{f}(m)&= \sum_{n\in \mathbb{Z}-\{0\}}\frac{e^{2\pi \I n\Theta}}{n^{\nu}}\frac{e^{-2\pi nRy(1/t_{+}^{n,- m}+t_{+}^{n,- m})}}{nR(t_{+}^{n,- m}-t_{-}^{n,- m})}= \sum_{n\in \mathbb{Z}-\{0\}}\frac{e^{2\pi \I n\Theta}}{n^{\nu}}\frac{e^{-2\pi y|n\tau- m|}}{|n\tau- m|}\,,
\end{split}
\end{equation}
where we used that $t_{+}^{n,-m}t_{-}^{n,-m}=-1$ and the expressions (<ref>) of $t_{\pm}^{n,-m}$.
After plugging the previous result in (<ref>), using $2R=\tau_2$, (<ref>) and relabeling the summation indices we then obtain
\begin{equation}\label{resum1}
\begin{split}
\mathcal{I}_{\hat{\gamma}}^{(\nu)}&= \sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\frac{e^{2\pi \I m\Theta-2\pi \I nx}}{m^{\nu}}\frac{e^{-2\pi y|m\tau+ n|}}{|m\tau+ n|}= \sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\frac{e^{-S_{\hat{\gamma},m,n}}}{m^{\nu}|m\tau + n|}\,.\\
\end{split}
\end{equation}
Finally, we use the following identities to prove (<ref>):
\begin{equation}\label{Bessellimit}
\lim_{z\to 0}zK_1(2\pi\tau_2 nz)=\frac{1}{2\pi n\tau_2}, \quad \lim_{z\to 0}z^2K_1(2\pi\tau_2 n z)=0, \quad \lim_{z\to 0} z^2K_0(2\pi \tau_2 nz)=0\,.
\end{equation}
We show the first identity in (<ref>), with the others following similarly. By applying the first limit in (<ref>), the dominated convergence theorem (here we use that $\nu\geq 2)$ and $K_0'=-K_1$, we obtain:
\begin{equation}
\begin{split}
\partial_{\tau_2}\mathcal{I}_{0}^{(\nu)}&=\lim_{\zeta_{\hat{\gamma}}\to 0,\widetilde{Z}_{\hat{\gamma}}\to 0}\left(\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(\nu)}+2\sum_{s=\pm 1}\sum_{n>0}\frac{e^{-2\pi \mathrm{i}ns \zeta_{\hat{\gamma}}}}{(sn)^{\nu}}2\pi n|\widetilde{Z}_{\hat{\gamma}}|K_1(2\pi n\tau_2|\widetilde{Z}_{\hat{\gamma}}|)\right)\\
&=\left(\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(\nu)}\right)\Big|_{\hat{\gamma}=0}+\frac{2}{\tau_2}\sum_{s=\pm 1}\sum_{n>0}\frac{1}{(sn)^{\nu}}\,.
\end{split}
\end{equation}
§.§.§ Poisson resummation of the quantum corrections of the contact structure
In this section, we seek to rewrite the quantum corrections terms in $f$, $\theta_{+}^P$ and $\theta_3^P$ using the Poisson resummation of Lemma <ref>. This will help us show that the coordinates (<ref>) are Darboux coordinates for the contact structure describing the instanton corrected q-map metric associated to the data of Section <ref>. We start with the following proposition for $f$:
The following holds for $f^{\text{w.s.}}$ and $f^{\text{inst}}$ defined in (<ref>):
\begin{equation}\label{typeiibfquant}
f^{\text{w.s.}}+f^{\text{inst}}=\frac{\chi}{12}+\frac{\tau_2^2}{(2\pi)^2}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\sum_{(m,n)\in \mathbb{Z}^2-\{0\}}\left(\frac{1}{|m\tau+n|}+2\pi q_at^a\right)\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau + n|^2}
\end{equation}
On one hand, we have
\begin{equation}
\begin{split}
f^{\text{inst}}&=\frac{2R}{\pi}\sum_{\gamma}\Omega(\gamma)\sum_{n>0}\frac{e^{-2\pi\I n\zeta_{\gamma}}}{n}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\\
&=\frac{2R}{\pi}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{s=\pm 1}\sum_{q_0\in \mathbb{Z}}\sum_{n>0}\frac{e^{-2\pi\I ns\zeta_{q_0\gamma^0+\hat{\gamma}}}}{n}|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|K_1(4\pi Rn|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|)\\
&\quad \quad-\frac{R}{\pi}\chi\sum_{s=\pm 1}\sum_{q_0\in \mathbb{Z}-\{0\}}\sum_{n>0}\frac{e^{-2\pi\I ns\zeta_{q_0\gamma^0}}}{n}|\widetilde{Z}_{q_0\gamma^0}|K_1(4\pi Rn|\widetilde{Z}_{q_0\gamma^0}|)\\
&=-\frac{R}{2\pi^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(2)}+\frac{R}{(2\pi)^2}\chi \partial_{\tau_2}\mathcal{I}_{0}^{(2)}\,,\\
\end{split}
\end{equation}
where we have used the particular form of the BPS indices (<ref>), the expressions (<ref>) and the fact that $R=\tau_2/2$. Using Lemma <ref> we therefore obtain that
\begin{equation}\label{typeiibfinst}
\begin{split}
f^{\text{inst}}&=\frac{\tau_2^2}{(2\pi)^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\left(\frac{1}{|m\tau+ n|}+2\pi q_at^a\right)\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau + n|^2}\\
&\quad\quad -\frac{\tau_2^2}{2(2\pi)^2}\chi \sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\frac{1}{|m\tau + n|^3}+\frac{\tau_2\chi}{2(2\pi)^2}\frac{4}{\tau_2}\sum_{n>0}\frac{1}{n^2}\\
&=\frac{\chi}{12}+\frac{\tau_2^2}{(2\pi)^2}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\left(\frac{1}{|m\tau+ n|}+2\pi q_at^a\right)\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau + n|^2}\,,\\
\end{split}
\end{equation}
where in the last equality we have used that $\sum_{n>0}\frac{1}{n^2}=\frac{\pi^2}{6}$ and combined the other two sums by using the convention $n_{0}=-\frac{\chi}{2}$ from Section <ref>.
On the other hand, it is not hard to check using (<ref>) that one can write
\begin{equation}\label{typeiibfws}
f^{\text{w.s.}}=\frac{\tau_2^2}{(2\pi)^2}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\sum_{n\in \mathbb{Z}-\{0\}}\left(\frac{1}{|n|}+2\pi q_at^a\right)\frac{e^{-S_{\hat{\gamma},0,n}}}{ n^2}\\
\end{equation}
By summing (<ref>) and (<ref>) we therefore obtain (<ref>).
Combining Proposition <ref> with (<ref>) we see that
\begin{equation}\label{Poissonresumedf}
f=8\pi \tau_2^2h(t) +16\pi\left(\frac{\chi}{192\pi}-c_{\ell}\right)+ \frac{\tau_2^2}{(2\pi)^2}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}\sum_{(m,n)\in \mathbb{Z}^2-\{0\}}\left(\frac{1}{|m\tau+n|}+2\pi q_at^a\right)\frac{e^{-S_{\hat{\gamma},m,n}}}{|m\tau + n|^2}\,,
\end{equation}
where $f^{\text{cl}}=-2\pi\tau_2^2\text{Im}(\tau_{ij}^{\text{cl}})z^i\overline{z}^j=8\pi\tau_2^2h(t)$ and $h(t)$ is given in (<ref>). In particular, we see that the quantum corrected $f$ has the same transformation rule as $f^{\text{cl}}$ under $S$-duality if and only if $c_{\ell}=\frac{\chi}{192\pi}$. Namely, for this value of $c_{\ell}$, $f$ transforms under $\begin{pmatrix}
a & b\\
c & d\\ \end{pmatrix}\in \mathrm{SL}(2,\mathbb{Z})
$ as
\begin{equation}
f\to \frac{f}{|c\tau+d|}\,.
\end{equation}
This suggests that this particular value of the 1-loop constant $c_{\ell}$ could play a special role in studying $S$-duality, and indeed, we will see in Section <ref> that for this value S-duality acts by isometries on the instanton corrected q-map space (provided $S$-duality acts on the domain of definition of the metric). Furthermore, in the setting of Calabi-Yau compactification of type IIB string theory, the quantity $f$ with $c_{\ell}=\frac{\chi}{192}$ is proportional to the 4d dilaton with perturbative, world-sheet instanton, D(-1) and D1 instanton corrections, and the corresponding transformation property has been previously remarked in the physics literature (see for example <cit.> and the paragraphs below the equation).
We now continue with the Poisson resummation of the other terms of the contact form.
The 1-form $-2\I (\theta_3^{P, \text{w.s.}}+\theta_3^{P, \text{inst}})$ can be rewritten in terms of the coordinates $(\tau_2,b^a,t^a,\tau_1=\zeta^0,\zeta^a, \widetilde{\zeta}_i,\sigma)$ as follows
\begin{equation}\label{typeiibtheta3quantum}
\begin{split}
&-2\I (\theta_3^{P, \text{w.s.}}+\theta_3^{P, \text{inst}})\\
&=\sum_{\hat{\gamma}\in \Lambda^{+}\cup\{0\}}n_{\hat{\gamma}}\sum_{(m,n)\in \mathbb{Z}^{2}-\{0\}}e^{-S_{\hat{\gamma},m,n}}\Bigg[\frac{\I \tau_2^4}{2\pi}\frac{m^2}{|m\tau +n|^4}q_a\mathrm{d}b^a -\frac{\tau_2^2}{2\pi}\frac{m\tau_1+ n}{|m\tau+ n|^3}q_a\mathrm{d}t^a+\frac{\I \tau_2^2}{2\pi }\frac{m(m\tau_1+ n)}{|m\tau+ n|^4}q_a\mathrm{d}\zeta^a
\\
&\quad-\left(\frac{\tau_2^3}{\pi^2} \frac{m^2(m\tau_1+ n)}{|m\tau+ n|^6}+\frac{\tau_2}{2\pi}q_at^a\frac{(m\tau_1+ n)^3+2m^2\tau_2^2(m\tau_1+ n)}{|m\tau+ n|^5}\right)\mathrm{d}\tau_2\\
&\quad+\left(-\frac{\I \tau_2^2}{2\pi}q_ab^a \frac{m(m\tau_1+ n)}{|m\tau+ n|^4}+\frac{\tau_2^4}{2\pi}q_at^a \frac{m^3}{|m\tau+ n|^5}+\frac{\tau_2^2}{2\pi^2}\frac{m(m^2\tau_2^2-(m\tau_1+ n)^2)}{|m\tau+ n|^6}\right)\mathrm{d}\tau_1\Bigg].\\
\end{split}
\end{equation}
We follow the same idea as in the previous proposition. Namely, notice that using (<ref>) and (<ref>), $\theta_{3}^{P,\text{inst}}$ can be written as
\begin{equation}
\begin{split}
\theta_3^{P,\text{inst}}=&-\frac{\I R}{2\pi}\sum_{\gamma}\Omega(\gamma)\sum_{n>0}\frac{e^{-2\pi \I n\zeta_{\gamma}}}{n}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\left(\frac{\D \widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}-\frac{\D \overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}\right)\\
&=-\frac{\I R}{2\pi}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{s=\pm 1}\sum_{q_0\in \mathbb{Z}}\sum_{n>0}\frac{e^{-2\pi \I ns\zeta_{q_0\gamma^0+\hat{\gamma}}}}{n}|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|K_1(4\pi Rn|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|)\left(\frac{\D \widetilde{Z}_{\hat{\gamma}}}{\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}}-\frac{\D \overline{\widetilde{Z}_{\hat{\gamma}}}}{\overline{\widetilde{Z}}_{q_0\gamma^0+\hat{\gamma}}}\right)\\
&=\frac{\I }{8\pi^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\partial_{z^a}\mathcal{I}_{\hat{\gamma}}^{(2)}\D z^a-\frac{\I }{8\pi^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\partial_{\overline{z}^a}\mathcal{I}_{\hat{\gamma}}^{(2)}\D \overline{z}^a,
\end{split}
\end{equation}
where in second equality we have used that no $\hat{\gamma}=0$ terms appear due to $\mathrm{d}\widetilde{Z}_{q_0\gamma^0}=\mathrm{d}q_0=0$.
Using Lemma <ref> we then obtain the following:
\begin{equation}\label{typeiibtheta3inst}
\begin{split}
-2\I &\theta_3^{P,\text{inst}}\\ &=\frac{\I }{2\pi}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\frac{e^{-S_{\hat{\gamma},m,n}}}{m^2}q_a\D b^a+\frac{1}{2\pi}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}e^{-S_{\hat{\gamma},m,n}}\frac{m\tau_1+ n}{m^2|m\tau+ n|}q_a\D t^a\,.\\
\end{split}
\end{equation}
Observe for the calculation that the partial derivatives $\partial_{z^a}$ and $\partial_{\overline{z}^a}$ are to be taken with respect to the mixed coordinates $(\tau_2,b^a,t^a,\tau_1=\zeta^0,\zeta^a, \widetilde{\zeta}_i,\sigma)$, which are neither the IIA nor the IIB coordinates.
On the other hand, using directly the formulas (<ref>) and (<ref>) for $\sigma^{\text{inst}}$, $\widetilde{\zeta}_i^{\text{inst}}$ and $\mathfrak{F}^{\text{w.s.}}$ one can compute $-2\I \theta_3^{P,\text{w.s.}}$ in (<ref>). The $\D \zeta^a$, $\D \tau_1$ and $\D \tau_2$ components of $-2\I \theta_3^{P,\text{w.s.}}$ can be seen to match the corresponding components in the right-hand side of (<ref>). On the other hand, the $\D t^a$ and $\D b^a$ components of $-2\I \theta_3^{P,\text{w.s.}}$ are as follows
\begin{align*}\label{typeiibtheta3ws}
-2\I \theta_3^{P,\text{w.s.}}|_{\D b^a}&= -\frac{\I }{2\pi}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}q_a\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\left(2-\frac{(m\tau_1+ n)^2}{|m\tau+n|^2}\right)\frac{e^{-S_{\hat{\gamma},m,n}}(m\tau_1+ n)^2}{m^2|m\tau+ n|^2}\\
-2\I \theta_3^{P,\text{w.s.}}|_{\D t^a}&= -\frac{\tau_2^2}{2\pi}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}q_a\sum_{n\in \mathbb{Z}-\{0\}}\frac{e^{-S_{\hat{\gamma},0,n}}}{n|n|}\\
& \quad-\frac{1}{2\pi}\sum_{\hat{\gamma}\in \Lambda^+\cup\{0\}}n_{\hat{\gamma}}q_a\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\left(2-\frac{(m\tau_1+ n)^2}{|m\tau+ n|^2}\right)\frac{e^{-S_{\hat{\gamma},m,n}}(m\tau_1+ n)}{m^2|m\tau+ n|}\,.\numberthis\\
\end{align*}
One can then directly check that the $\D b^a$ and $\D t^a$ components of (<ref>) and (<ref>) combine into the corresponding components of (<ref>). We therefore obtain the required expression.
Finally, we compute a similar expression for $\theta_{+}^P$:
The 1-form $\I (\theta_{+}^{P,\text{w.s.}}+\theta_{+}^{P,\text{inst}})$ can be rewritten in terms of the coordinates $(\tau_2,b^a,t^a,\tau_1=\zeta^0,\zeta^a, \widetilde{\zeta}_i,\sigma)$ as follows
\begin{equation}\label{typeiibthetaplusquant}
\begin{split}
&\I (\theta_{+}^{P,\text{w.s.}}+\theta_{+}^{P,\text{inst}})\\
=&\sum_{\hat{\gamma}\in \Lambda^+\cup \{0\}}n_{\hat{\gamma}}\sum_{(m,n)\in \mathbb{Z}^2-\{0\}}e^{-S_{\hat{\gamma},m,n}}\Bigg[\frac{\I \tau_2^3}{4\pi}\frac{m(|m\tau+n|-(m\tau_1+n))}{|m\tau+ n|^4}q_a\mathrm{d}b^a\\
&-\frac{\tau_2^3}{4\pi } \frac{m}{|m\tau+ n|^3}q_a\mathrm{d}t^a+\frac{\I \tau_2}{4\pi}\frac{(m\tau_1+ n)(|m\tau+ n|-(m\tau_1+ n))}{|m\tau+ n|^4}q_a\mathrm{d}\zeta^a \\
&-\frac{\tau_2}{8\pi^2}\frac{2(m\tau_1+ n)(m^2\tau_2^2-(m\tau_1+ n)^2) +|m\tau+ n|(2(m\tau_1+ n)^2-m^2\tau_2^2)}{|m\tau+ n|^6}\mathrm{d}\tau_1\\
\end{split}
\end{equation}
As before, we first notice that we can write the terms of $\theta_{+}^{P,\text{inst}}$ as follows. For the first sum in (<ref>) we have
\begin{equation}
\begin{split}
&2R\I\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi\I n\zeta_{\gamma}}K_0(4\pi Rn|\widetilde{Z}_{\gamma}|)\D \zeta_{\gamma}\\
&=2R\I\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{q_0\in \mathbb{Z}}\sum_{s=\pm 1}\sum_{n>0}\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}e^{-2\pi \I ns\zeta_{q_0\gamma^0+\hat{\gamma}}}K_0(4\pi Rn|\widetilde{Z}_{q_0\gamma^0+\hat{\gamma}}|)\D \zeta_{q_0\gamma^0+\hat{\gamma}}\\
&\quad -R\I \chi\sum_{q_0\in \mathbb{Z}}\sum_{s=\pm 1}\sum_{n>0}\widetilde{Z}_{q_0\gamma^0}e^{-2\pi \I ns\zeta_{q_0\gamma^0}}K_0(4\pi Rn|\widetilde{Z}_{q_0\gamma^0}|)\D \zeta_{q_0\gamma^0}\\
&=-\frac{1}{(2\pi)^3}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\D \zeta^0\partial_{\zeta^0}\partial_{\tau_2}\partial_{\overline{z}_{\hat{\gamma}}}\mathcal{I}_{\hat{\gamma}}^{(3)}-\frac{\I }{(2\pi)^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\D \zeta^a\partial_{\tau_2}\partial_{\overline{z}^a}\mathcal{I}_{\hat{\gamma}}^{(2)}-\frac{R\I }{2(2\pi)^2}\chi \D \zeta^0\partial_{\zeta^0}^2\mathcal{I}_{0}^{(2)}\,,\\
\end{split}
\end{equation}
where $\partial_{z_{\hat{\gamma}}}:=\frac{1}{\#q}\frac{1}{q_a}\partial_{z^a}$ with the sum in the index $a$ only for the $a=1,...,n$ such that $q_a\neq0$, and $\#q$ is the number of non-zero $q_a$[It is natural to denote $z_{\hat{\gamma}}:=q_az^a$ for $\hat{\gamma}\in \Lambda^{+}$. With this notation, the differential operator $\partial_{\widetilde{z}_{\hat{\gamma}}}$ satisfies $\partial_{z_{\hat{\gamma}}}z_{\hat{\gamma}}=1$.]. By Lemma <ref>, we then find
\begin{align*}
&2R\I \sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi\I n\zeta_{\gamma}}K_0(2Rn|\widetilde{Z}_{\gamma}|)\D \zeta_{\gamma}\\ &=\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}e^{-S_{\hat{\gamma},m,n}}\Bigg[-\frac{\I \tau_2}{8\pi^2}\left(\frac{1}{|m\tau+ n|^3}\left(1-3\frac{(m\tau_1+ n)^2}{|m\tau+ n|^2}\right)-2\pi\I \frac{q_ab^a(m\tau_1+ n)}{|m\tau+ n|^3}\right)\D \zeta^0\\
&\quad-\frac{\I \tau_2}{4\pi}\frac{q_at^a}{|m\tau+ n|^4}\left(m^2\tau_2^2-|m\tau+ n|(m\tau_1+ n)-2(m\tau_1+ n)^2\right)\D \zeta^0\\
&\quad+\frac{\I \tau_2}{2}q_at^aq_b\frac{m\tau_1+ n+|m\tau+ n|}{|m\tau+ n|^2}\left(\I b^b+\frac{(m\tau_1+ n)t^b}{|m\tau+ n|}\right)\D \zeta^0\\
&\quad+\frac{\tau_2}{4\pi}\frac{q_a}{|m\tau+ n|^2}\left(2\pi q_bt^b(m\tau_1+ n+|m\tau+ n|)+\frac{m\tau_1+ n}{|m\tau + n|}\right)\mathrm{\D }\zeta^a\Bigg]\\
&\quad +\frac{\I \tau_2}{(4\pi)^2}\chi \sum_{\substack{m\in \mathbb{Z}-\{0\}\\ n\in \mathbb{Z}}}\frac{1}{|m\tau+n|^3}\left(1-3\frac{(m\tau_1+ n)^2}{|m\tau+ n|^2}\right)\D \zeta^0\,. \numberthis \label{typeiibthetaplusinst1}\\
\end{align*}
Similarly, for the remaining term of $\theta_{+}^{P,\text{inst}}$ in (<ref>) we have that
\begin{equation}
\begin{split}
&2R^2\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi \I n\zeta_{\gamma}}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\left(\frac{\D \widetilde{Z}_{\gamma}}{\widetilde{Z}_{\gamma}}+\frac{\D \overline{\widetilde{Z}}_{\gamma}}{\overline{\widetilde{Z}}_{\gamma}}+\frac{2}{\tau_2}\D \tau_2\right)\\
&=2R^2\sum_{\gamma}\Omega(\gamma)\sum_{n>0}e^{-2\pi\I n\zeta_{\gamma}}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\D \widetilde{Z}_{\gamma}+2R^2\sum_{\gamma}\Omega(\gamma)\frac{\widetilde{Z}_{\gamma}^2}{|\widetilde{Z}_{\gamma}|}\sum_{n>0}e^{-2\pi\I n\zeta_{\gamma}}K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\D \overline{\widetilde{Z}}_{\gamma}\\
&\quad +\tau_2\sum_{\gamma}\Omega(\gamma)\widetilde{Z}_{\gamma}\sum_{n>0}e^{-2\pi\I n\zeta_{\gamma}}|\widetilde{Z}_{\gamma}|K_1(4\pi Rn|\widetilde{Z}_{\gamma}|)\D \tau_2\\
&=\frac{\I \tau_2^2}{16\pi^2}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\partial_{\zeta^a}\partial_{\tau_2}\mathcal{I}_{\hat{\gamma}}^{(2)}\D z^a-\frac{1}{(2\pi)^3}\sum_{\hat{\gamma}\in \Lambda^+}n_{\hat{\gamma}}\partial_{\overline{z}^a}\partial_{\tau_2}\partial_{\overline{z}_{\hat{\gamma}}}\mathcal{I}_{\hat{\gamma}}^{(3)}\D \overline{z}^a\\
|
# Confronting the primordial black hole scenario with the gravitational-wave
events detected by LIGO-Virgo
Zu-Cheng Chen<EMAIL_ADDRESS>CAS Key Laboratory of Theoretical
Physics, Institute of Theoretical Physics, Chinese Academy of Sciences,
Beijing 100190, China School of Physical Sciences, University of Chinese
Academy of Sciences, No. 19A Yuquan Road, Beijing 100049, China Chen Yuan
<EMAIL_ADDRESS>CAS Key Laboratory of Theoretical Physics, Institute of
Theoretical Physics, Chinese Academy of Sciences, Beijing 100190, China
School of Physical Sciences, University of Chinese Academy of Sciences, No.
19A Yuquan Road, Beijing 100049, China Qing-Guo Huang Corresponding author:
<EMAIL_ADDRESS>CAS Key Laboratory of Theoretical Physics, Institute of
Theoretical Physics, Chinese Academy of Sciences, Beijing 100190, China
School of Physical Sciences, University of Chinese Academy of Sciences, No.
19A Yuquan Road, Beijing 100049, China School of Fundamental Physics and
Mathematical Sciences Hangzhou Institute for Advanced Study, UCAS, Hangzhou
310024, China Center for Gravitation and Cosmology, College of Physical
Science and Technology, Yangzhou University, Yangzhou 225009, China
###### Abstract
Adopting a binned method, we model-independently reconstruct the mass function
of primordial black holes (PBHs) from GWTC-3 and find that such a PBH mass
function can be explained by a broad red-tilted power spectrum of curvature
perturbations. Even though GW190521 with component masses in upper mass gap
$(m>65M_{\odot})$ can be naturally interpreted in the PBH scenario, the events
(including GW190814, GW190425, GW200105, and GW200115) with component masses
in the light mass range $(m<3M_{\odot})$ are quite unlikely to be explained by
binary PBHs although there are no electromagnetic counterparts because the
corresponding PBH merger rates are much smaller than those given by LIGO-
Virgo. Furthermore, we predict that both the gravitational-wave (GW)
background generated by the binary PBHs and the scalar-induced GWs
accompanying the formation of PBHs should be detected by the ground-based and
space-borne GW detectors and pulsar timing arrays in the future.
Introduction. Primordial black holes (PBHs) Carr and Hawking (1974); Carr
(1975) are formed in the very early Universe due to the collapse of over-
densed regions which are generated by the enhanced curvature power spectrum on
small scales compared to those on the cosmic microwave background (CMB)
scales. PBHs can not only represent the dark matter (DM) in the Universe, but
also may potentially provide an explanation to the merger events detected by
LIGO-Virgo Collaboration (LVC) if the fraction of the stellar mass PBHs in
cold dark matter (CDM) is $f_{\mathrm{PBH}}\sim\mathrm{few}\times 10^{-3}$
Sasaki _et al._ (2016); Chen and Huang (2018); Raidal _et al._ (2019); De
Luca _et al._ (2020a); Hall _et al._ (2020); Bhagwat _et al._ (2021); Hütsi
_et al._ (2020); Wong _et al._ (2021); De Luca _et al._ (2021a); Franciolini
_et al._ (2021).
Among all the merger events detected by LVC, some of them are likely to have
ambiguities in the astrophysical scenario. Firstly, the primary component of
GW190521 has a high probability to be within the pair-instability supernovae
mass gap Abbott _et al._ (2020a), implying that the primary black hole (BH)
might not have a stellar origin. Secondly, even though the component masses of
GW190425 lie in the mass range of $[1.12M_{\odot},2.52M_{\odot}]$ and are
consistent with the individual binary component being neutron star (NS) Abbott
_et al._ (2020b), the source-frame chirp mass and total mass are significantly
larger than any known binary NS system. Thirdly, GW190814 is reported to have
a compact object with a mass of $2.5-2.67M_{\odot}$ Abbott _et al._ (2020c),
which falls in the “lower mass gap” where no NSs or BHs have been observed in
a double compact-object system. Finally, LVC recently reported GW200105 and
GW200115 Abbott _et al._ (2021a) in which the secondary component masses are
respectively $1.9_{-0.2}^{+0.3}M_{\odot}$ and $1.5_{-0.3}^{+0.7}M_{\odot}$,
indicating that both of them are well below the maximal mass of an NS. In
addition, there are no electromagnetic counterparts to confirm and PBHs are
speculated to provide an explanation to these five gravitational-wave (GW)
events Abbott _et al._ (2020b, c, 2021a); De Luca _et al._ (2021b).
In this letter, we will give a comprehensive investigation for the possibility
that if the GW events, in particular the five events mentioned above, can be
explained in the PBH scenario. We adopt a binned method to model-independently
reconstruct the PBH mass function from GWTC-3 Abbott _et al._ (2021b), and we
find that GW190521 can be explained by a binary PBH merger. But the other four
GW events (GW190814, GW190425, GW200105, and GW200115) are quite unlikely to
be interpreted as binary PBHs because the corresponding merger rates of binary
PBHs are much smaller than those given by LVC even though the PBH mass
function around $\sim 1M_{\odot}$ is significantly enhanced due to the
softening of the equation of state during the QCD phase transition Byrnes _et
al._ (2018).
In addition, we compute the gravitational-wave background (GWB) from both the
PBH binary coalescences Chen _et al._ (2019) and the so-called scalar-induced
gravitational waves (SIGWs) generated by the curvature perturbation during the
formation of PBHs Saito and Yokoyama (2009). We find that the GWB associated
with PBHs is compatible with the current limits of observations and should be
detected by the ground-based and space-borne GW detectors and pulsar timing
arrays in the future.
A model-independent reconstruction of PBH mass function. In this work, the PBH
mass is considered in the range of $[1M_{\odot},130M_{\odot}]$. To infer the
PBH mass function from the GWTC-3, we adopt a model-independent approach using
the following binned mass function
$P(m)=\begin{cases}P_{1},&1\,M_{\odot}\leq m<3\,M_{\odot}\\\
P_{2},&3\,M_{\odot}\leq m<10\,M_{\odot}\\\ P_{3},&10\,M_{\odot}\leq
m<40\,M_{\odot}\\\ P_{4},&40\,M_{\odot}\leq m<80\,M_{\odot}\\\
P_{5},&80\,M_{\odot}\leq m\leq 130\,M_{\odot}\end{cases}$ (1)
in which the mass function $P(m)$ is normalized by $\int P(m)dm=1$. Therefore,
only four of $P_{i}(i=1,\cdots,5)$ are independent, and
$\vec{\theta}=\\{P_{1},P_{2},P_{3},P_{4}\\}$ are chosen to be the free
parameters. The merger rate density in units of
$\mathrm{Gpc}^{-3}\,\mathrm{yr}^{-1}$ for a general mass function,
$P(m|\vec{\theta})$, takes the form of Chen and Huang (2018)
$\displaystyle\mathcal{R}_{12}$ $\displaystyle(t|\vec{\theta})\approx 2.8\cdot
10^{6}\times\left({\frac{t}{t_{0}}}\right)^{-\frac{34}{37}}f_{\mathrm{PBH}}^{2}(0.7f_{\mathrm{PBH}}^{2}+\sigma_{\mathrm{eq}}^{2})^{-{21\over
74}}$ (2)
$\displaystyle\times\min\left(\frac{P(m_{1}|\vec{\theta})}{m_{1}},\frac{P(m_{2}|\vec{\theta})}{m_{2}}\right)\left({P(m_{1}|\vec{\theta})\over
m_{1}}+{P(m_{2}|\vec{\theta})\over m_{2}}\right)$
$\displaystyle\times(m_{1}m_{2})^{{3\over 37}}(m_{1}+m_{2})^{36\over 37},$
where the component masses $m_{1}$ and $m_{2}$ are in units of $M_{\odot}$,
$f_{\mathrm{PBH}}\equiv\Omega_{\mathrm{PBH}}/\Omega_{\mathrm{CDM}}$ is the
energy density fraction of PBHs in CDM, and $\sigma_{\mathrm{eq}}\approx
0.005$ Ali-Haïmoud _et al._ (2017); Chen and Huang (2018) is the variance of
density perturbations of the rest CDM on scale of order
$\mathcal{O}(10^{0}\sim 10^{3})M_{\odot}$ at radiation-matter equality.
We perform the hierarchical Bayesian inference Abbott _et al._ (2016a, b, c);
Wysocki _et al._ (2019); Fishbach _et al._ (2018); Mandel _et al._ (2019);
Thrane and Talbot (2019) to extract the population parameters
$\\{\vec{\theta},f_{\mathrm{PBH}}\\}$ from observed BBHs. Given the data of
$N$ binary BH (BBH) detections, $\vec{d}=(d_{1},\dots,d_{N})$, the likelihood
for an inhomogeneous Poisson process is Wysocki _et al._ (2019); Fishbach
_et al._ (2018); Mandel _et al._ (2019); Thrane and Talbot (2019)
$p(\vec{d}|\vec{\theta},R)\propto
e^{-\beta(\vec{\theta})}\prod_{i}^{N}\int\mathrm{d}\vec{\lambda}\
p(\vec{\lambda}|d_{i})\ \mathcal{R}_{12}(\vec{\lambda}|\vec{\theta}),$ (3)
where $\vec{\lambda}\equiv\\{m_{1},m_{2}\\}$, $p(\vec{\lambda}|d_{i})$ is the
posterior of an individual event, and
$\beta(\vec{\theta})\equiv\int\mathrm{d}\vec{\lambda}\ VT(\vec{\lambda})\
\mathcal{R}_{12}(\vec{\lambda}|\vec{\theta})$ where $VT(\vec{\lambda})$ is the
spacetime sensitivity volume of LIGO-Virgo detectors. We use the GW events
from GWTC-3 by discarding events with false alarm rate larger than $1\
\mathrm{yr}^{-1}$, and events with the secondary component mass smaller than
$3M_{\odot}$ to avoid contamination from putative events involving neutron
stars De Luca _et al._ (2021a). On the other hand, although PBHs are expected
to have negligible spin at formation De Luca _et al._ (2019a); Mirbabayi _et
al._ (2020), they might become fast rotating through accretion effects De Luca
_et al._ (2020b, c). However, given that the accretion model is very sensitive
to the cut-off points of the red-shift and there is no evidence to support the
accretion effects on PBHs so far, we consider two cases where in case I we
discard the events with non-vanishing effective spin while we keep these
events in case II.
The median value and $90\%$ equal-tailed credible intervals for the parameters
$\\{\vec{\theta},f_{\mathrm{PBH}}\\}$ are represented by crosses in Fig. 1.
For the case I, the results are $P_{1}=1.5^{+0.6}_{-0.7}\times
10^{-1}M_{\odot}^{-1}$, $P_{2}=2.5^{+1.8}_{-1.1}\times 10^{-2}M_{\odot}^{-1}$,
$P_{3}=1.4^{+0.4}_{-0.3}\times 10^{-2}M_{\odot}^{-1}$,
$P_{4}=2.0^{+0.9}_{-0.7}\times 10^{-3}M_{\odot}^{-1}$, and
$f_{\mathrm{PBH}}=2.9^{+0.7}_{-0.6}\times 10^{-3}$. Therefore, the total local
merger rate is $154^{+186}_{-88}\,\mathrm{Gpc}^{-3}\,\mathrm{yr}^{-1}$
according to Eq. (2). For the case II, the results are
$P_{1}=1.1^{+0.8}_{-0.5}\times 10^{-1}M_{\odot}^{-1}$,
$P_{2}=4.2^{+1.2}_{-1.8}\times 10^{-2}M_{\odot}^{-1}$,
$P_{3}=1.3^{+0.3}_{-0.3}\times 10^{-2}M_{\odot}^{-1}$,
$P_{4}=2.0^{+0.7}_{-0.6}\times 10^{-3}M_{\odot}^{-1}$, and
$f_{\mathrm{PBH}}=3.4^{+0.7}_{-0.5}\times 10^{-3}$. Therefore, the total local
merger rate is $175^{+144}_{-68}\,\mathrm{Gpc}^{-3}\,\mathrm{yr}^{-1}$.
Figure 1: The median values and the $90\%$ credible intervals of the
population parameters $f_{\mathrm{PBH}}P_{i}$ with $i=1,\cdots,5$. The blue
(red) curves correspond to the mass function of PBHs generated by a broad
tilted curvature power spectrum corresponding to discarding (keeping) non-
vanishing effective spin events.
From our binned reconstruction of PBH mass function, the inferred merger rates
of binary PBHs for the events with component masses in the upper mass gap
$(m>65M_{\odot})$ and the light mass range $(m<3M_{\odot})$ are listed in the
third column of Tab. 1. Compared to the merger rates given by LVC (the second
column of Tab. 1), we conclude that the upper mass gap event GW190521 can be
explained by the PBH scenario, while the events with at least one component
mass being smaller than $3M_{\odot}$ (GW190425, GW190814, GW200105, and
GW200115) are quite unlikely to be explained in the PBH scenario because the
merger rates of corresponding binary PBHs are at least one order of magnitude
smaller than those given by LVC.
Event | $R_{\mathrm{LVC}}[\mathrm{Gpc}^{-3}\,\mathrm{yr}^{-1}]$ | $R_{\mathrm{PBH}}[\mathrm{Gpc}^{-3}\,\mathrm{yr}^{-1}]$
---|---|---
case I | case II
GW190521 | $0.13^{+0.30}_{-0.11}$ Abbott _et al._ (2020a) | $0.12^{+0.11}_{-0.07}$ | $0.16^{+0.11}_{-0.08}$
GW190425 | $460^{+1050}_{-390}$ Abbott _et al._ (2020b) | $3.6^{+6.8}_{-2.8}$ | $2.5^{+6.8}_{-1.8}$
GW190814 | $7^{+16}_{-6}$ Abbott _et al._ (2020c) | $0.13^{+0.09}_{-0.07}$ | $0.12^{+0.11}_{-0.05}$
GW200105 | $16^{+38}_{-14}$ Abbott _et al._ (2021a) | $1.9^{+1.8}_{-1.0}$ | $2.8^{+2.1}_{-1.4}$
GW200115 | $36^{+82}_{-30}$ Abbott _et al._ (2021a) | $6.3^{+7.4}_{-3.8}$ | $8.7^{+8.1}_{-4.5}$
Table 1: The local merger rate inferred by LIGO-Virgo and PBH.
It is also worthy figuring out a physical explanation for the PBH mass
function reconstucted from GWTC-3 in Fig. 1. Here, we take into account a
broad tilted power spectrum for the curvature perturbations, namely
$\mathcal{P}_{\mathcal{R}}(k)=A(k/k_{\min})^{\alpha}\Theta(k-k_{\min})\Theta(k_{\max}-k),$
(4)
where $\Theta$ is the heaviside theta function, $k_{\min}=3.2\times
10^{5}\mathrm{Mpc}^{-1}$ and $k_{\max}=9.5\times 10^{6}\mathrm{Mpc}^{-1}$.
After PBHs are formed, they evolve like dust-like matter during radiation
dominated era, and then
$f_{\mathrm{PBH}}P(m)=\frac{1}{\Omega_{\mathrm{CDM}}m}\left(\frac{M_{\mathrm{eq}}}{m}\right)^{1/2}\beta\left(m\right),$
(5)
where $M_{\mathrm{eq}}\approx 2.8\times 10^{17}M_{\odot}$ is the horizon mass
at matter-radiation equality and $\beta(m)$ describes the mass fraction of the
Universe that collapse to form PBHs. Using Press-Schechter formalism Press and
Schechter (1974), $\beta(m)$ can be evaluated by integrating the probability
distribution function (PDF) $P(\delta)$ of the density contrast $\delta$ that
is larger than the threshold value, namely
$\beta(m)=\int_{\delta_{c}}^{\infty}d\delta\frac{m}{M_{H}}P(\delta),$ (6)
where
$M_{H}\simeq
17\left(\frac{g}{10.75}\right)^{-1/6}\left(\frac{k}{10^{6}\mathrm{Mpc}^{-1}}\right)^{-2}M_{\odot}$
(7)
is the horizon mass. Here $g$ is the degress of freedom of relativistic
particles, $\delta_{c}$ is the threshold value of density contrast for the
formation of PBHs,
$P(\delta)=e^{-\delta^{2}/(2\sigma_{k}^{2})}/\sqrt{2\pi\sigma_{k}^{2}}$ with
variance $\sigma_{k}$ related to the curvature power spectrum by
$\sigma^{2}_{k}=\left(\frac{4}{9}\right)^{2}\int_{0}^{\infty}\frac{dq}{q}{W}^{2}(q,R_{H})\left(\frac{q}{k}\right)^{4}T^{2}(q,R_{H})\mathcal{P}_{\mathcal{R}}(q),$
(8)
where
$T(k,R_{H})=3\frac{\sin(kR_{H}/\sqrt{3})-(kR_{H}/\sqrt{3})\cos(kR_{H}/\sqrt{3})}{(kR_{H}/\sqrt{3})^{3}}$
(9)
is the transfer function during radiation dominated era. A window function
$W(k,R_{H})$ is needed to smooth out the density contrast on a comoving length
$R\sim k^{-1}$, for which we use a top-hap window function in real space,
namely
$W(k,R_{H})=3\frac{\sin(kR_{H})-(kR_{H})\cos(kR_{H})}{(kR_{H})^{3}}.$ (10)
The PBH mass $m$ is related to the density contrast by the critical collapse,
namely $m=M_{H}\kappa\left(\delta_{m}-\delta_{c}\right)^{\gamma}$ Choptuik
(1993); Evans and Coleman (1994); Niemeyer and Jedamzik (1998) with
$\kappa=3.3$ and $\gamma=0.36$ Koike _et al._ (1995). The nonlinear relation
between the density contrast and the curvature perturbation $\delta_{\zeta}$
leads to $\delta_{m}=\delta_{\zeta}-3/8\delta_{\zeta}^{2}$ Young _et al._
(2019); De Luca _et al._ (2019b); Kawasaki and Nakatsuka (2019). For the
equation of state $w=1/3$, numerical simulation indicates $\delta_{c}\approx
0.45$ Musco _et al._ (2005, 2009). The threshold value of density contrast
$\delta_{c}$ slightly decreases due to the softening of equation of state
Musco and Miller (2013); Saikawa and Shirai (2018) during the QCD phase
transition Borsanyi _et al._ (2016); Saikawa and Shirai (2018), and therefore
the PBH mass function around $m\sim 1M_{\odot}$ should be significantly
enhanced Byrnes _et al._ (2018). In this letter, the data of equation of
state and the sound speed are adopted from Saikawa and Shirai (2018).
Here, the best-fit values of $A$ and $\alpha$ are $A=0.004647$ and
$\alpha=-0.1033$ for case I, and $A=0.004640$ and $\alpha=-0.1003$ for case
II, and then the PBH mass function generated by such a broad tilted curvature
power spectrum is illustrated by the blue (case I) and red (case II) curves in
Fig. 1. In particular, the enhancement of PBH mass function around $m\sim
1M_{\odot}$ just attributes to the softening of the equation of state during
the QCD phase transition. In a word, our results indicate that such a broad
red-tilted curvature power spectrum provides a quite reasonable explanation
for the PBH mass function implied by GWTC-3.
GWB associated with PBHs. GWB is supposed to be a superposition of incoherent
GWs that are characterized statistically. The GWB associated with PBHs
includes two main contributions: one is the coalescences of binary PBHs Chen
_et al._ (2019) (see more discussion in Mukherjee and Silk (2021); Mukherjee
_et al._ (2021)) and the other is SIGW inevitably generated by the curvature
perturbations during the formation of PBHs Saito and Yokoyama (2009).
Figure 2: The GWB associated with PBHs. The blue (red) band corresponds to
the GWB from the coalescences of binary PBHs from Case I (Case II), and the
purple curve corresponds to the SIGWs generated by the curvature perturbation
during the formation of PBHs. The two dark shaded regions are ruled out by the
LIGO O3 Abbott _et al._ (2021c) and NANOGrav 12.5-year data Arzoumanian _et
al._ (2020), respectively. We also show the sensitivity curves of FAST Nan
_et al._ (2011), LISA Audley _et al._ (2017), TianQin Luo _et al._ (2016),
Taiji Hu and Wu (2017), LIGO Design, Einstein Telescope (ET) Punturo _et al._
(2010), Cosmic Explorer (CE) Abbott _et al._ (2017) and NEMO Ackley _et al._
(2020).
For binary systems, the energy-density spectrum of a GWB which defined as the
energy density of GWs per logarithm frequency, $f=k/(2\pi)$, can be calculated
as Allen and Romano (1999); Phinney (2001); Regimbau and Mandic (2008); Zhu
_et al._ (2011, 2013)
$\Omega_{\mathrm{GW}}(f)=\frac{f}{\rho_{c}H_{0}}\int\mathrm{d}z\mathrm{d}m_{1}\mathrm{d}m_{2}\frac{\mathcal{R}_{12}(z)}{(1+z)\,E(z)}\frac{\mathrm{d}E_{\mathrm{GW}}}{\mathrm{d}f_{s}},$
(11)
where $\rho_{c}=3H_{0}^{2}/(8\pi)$ is the critical energy density of our
Universe, $f_{s}$ is frequency of GWs in the source frame, $H_{0}$ is the
Hubble constant, and
$E(z)=\sqrt{\Omega_{r}\left(1+z\right)^{4}+\Omega_{m}(1+z)^{3}+\Omega_{\Lambda}}$
accounts for the evolution of our Universe with $\Omega_{\mathrm{r}}$,
$\Omega_{\mathrm{m}}$ and $\Omega_{\mathrm{\Lambda}}$ the density parameters
for radiation, matter and dark energy. Here, we adopt the best-fit results
from Planck 2018 (Aghanim _et al._ , 2020) and approximate the energy
spectrum $\mathrm{d}E_{\mathrm{GW}}/\mathrm{d}f_{s}$ emitted by an individual
BBH using expressions from Cutler _et al._ (1993); Chernoff and Finn (1993);
Zhu _et al._ (2011). The corresponding GWBs for case I and case II are
respectively shown as the blue and red band in Fig. 2, indicating that both of
them are compatible with the current limits given by LIGO O3 Abbott _et al._
(2021c), and should be detected by the future ground-based and space-borne GW
detectors, such as Neutron Star Extreme Matter Observatory (NEMO) Ackley _et
al._ (2020), Cosmic Explorer (CE) Abbott _et al._ (2017), Einstein Telescope
(ET) Punturo _et al._ (2010), Taiji Hu and Wu (2017), TianQin Luo _et al._
(2016) and LISA Audley _et al._ (2017).
On the other hand, it is known that the GWs should be generated by the scalar
curvature perturbations at second order in perturbation theory Tomita (1967);
Matarrese _et al._ (1993, 1994, 1998); Noh and Hwang (2004); Carbone and
Matarrese (2005); Nakamura (2007); Yuan _et al._ (2020, 2019a, 2019b); Yuan
and Huang (2021). In this sense, SIGWs are inevitably produced during the
formation of PBHs. The perturbed metric in Newton gauge is given by
$\mathrm{d}s^{2}=a^{2}\left\\{-(1+2\Phi)\mathrm{d}\eta^{2}+\left[(1-2\Phi)\delta_{ij}+\frac{h_{ij}}{2}\right]\mathrm{d}x^{i}\mathrm{~{}d}x^{j}\right\\},$
(12)
where $\Phi=-2\mathcal{R}/3$ is the Bardeen potential, $h_{ij}$ is the second-
order transverse and traceless tensor mode and $\eta$ is the conformal time.
The equation of motion for $\Phi$ is govern by
$\Phi^{\prime\prime}+3\mathcal{H}(\eta)\left(1+c_{s}^{2}\right)\Phi^{\prime}+3\mathcal{H}^{2}\left(c_{s}^{2}-w\right)\Phi-
c_{s}^{2}\nabla^{2}\Phi=0,$ (13)
where $w$ and $c_{s}$ are the equation of state and the sound speed,
respectively. $\Phi_{k}(\eta)$ is related to its initial value
$\Phi_{k}\equiv\Phi_{k}(\eta\rightarrow 0)$ by
$\Phi_{k}(\eta)=\Phi_{k}T_{\Phi}(k\eta)$, where $T_{\Phi}(k\eta)$ is the
transfer function. The equation of motion for the second-order tensor modes,
$h_{ij}$, takes the form
$h_{ij}^{\prime\prime}+2\mathcal{H}h_{ij}^{\prime}-\nabla^{2}h_{ij}=-4\mathcal{T}_{ij}^{\ell
m}\mathcal{S}_{\ell m},$ (14)
with $\mathcal{T}_{ij}^{\ell
m}=e_{ij}^{(+)}(\bm{k})e^{(+)lm}(\bm{k})+e_{ij}^{(\times)}(\bm{k})e^{(\times)lm}(\bm{k})$
selects the transverse-traceless part of the source term, with $e_{ij}$ the
polarization tensor and $\mathcal{H}=a^{\prime}/a$. The source term is given
by
$S_{ij}=2\Phi\partial_{i}\partial_{j}\Phi-\frac{4}{3(1+w)}\left(\partial_{i}\Phi+\frac{\partial_{i}\Phi^{\prime}}{\mathcal{H}(\eta)}\right)\left(\partial_{j}\Phi+\frac{\partial_{j}\Phi^{\prime}}{\mathcal{H}(\eta)}\right).$
(15)
Here the prime denotes the derivative with respect to $\eta$. Following Kohri
and Terada (2018), Eq. (14) can be solved by the Green’s function and the
transfer function method, and the energy density parameter by today is given
by
$\displaystyle\Omega_{\mathrm{GW}}$
$\displaystyle=\frac{\Omega_{\mathrm{r}}}{6}\int_{0}^{\infty}\mathrm{d}u\int_{|1-u|}^{1+u}\mathrm{~{}d}v\frac{v^{2}}{u^{2}}\left[1-\left(\frac{1+v^{2}-u^{2}}{2v}\right)^{2}\right]^{2}$
(16)
$\displaystyle\qquad\qquad\times\mathcal{P}_{\mathcal{R}}(uk)\mathcal{P}_{\mathcal{R}}(vk)\overline{I^{2}(u,v)},$
The kernel function takes the form Kohri and Terada (2018)
$\displaystyle\overline{I^{2}(u,v)}=\frac{9(u^{2}+v^{2}-3)^{2}}{32u^{6}v^{6}}\Bigg{\\{}\Big{(}-4uv+(u^{2}+v^{2}-3)$
$\displaystyle\times\ln\Big{|}{3-(u+v)^{2}\over
3-(u-v)^{2}}\Big{|}\Big{)}^{2}+\pi^{2}\left(u^{2}+v^{2}-3\right)^{2}\Theta(u+v-\sqrt{3})\Bigg{\\}}.$
For the curvature power spectrum given in Eq. (4) with $A$ and $\alpha$ taking
their best-fit values, the SIGWs are illustrated as the purple curve
(coincides for case I and II) in Fig. 2. We find that the predicted SIGWs are
compatible with NANOGrav 12.5-yr data Arzoumanian _et al._ (2020), and should
be detected by FAST Nan _et al._ (2011) in the future. Note that recent
analysis implies there is no statistically significant evidence for the tensor
transverse polarization mode in the NANOGrav 12.5-yr data set Arzoumanian _et
al._ (2020); Chen _et al._ (2021a); Arzoumanian _et al._ (2021), PPTA second
data release Goncharov _et al._ (2021); Wu _et al._ (2021), IPTA second data
release Chen _et al._ (2021b), and EPTA second data release Chalumeau _et
al._ (2021).
Conclusion and Discussion. In this letter, we use a binned PBH mass function
to model-independently reconstruct the PBH mass function from GWTC-3, and find
that such a mass function can be naturally explained by a broad red-tilted
curvature power spectrum. By comparing the merger rates of binary PBHs with
those given by LVC, we conclude that GW190521 with the primary component being
within the pair-instability supernovae mass gap can be explained by the merger
of binary PBHs, but the light components (i.e. $m<3M_{\odot}$) in GW190814,
GW190425, GW200105, and GW200115 events should be NSs or other exotic compact
objects. In addition, the PBH scenario proposed in this letter can be tested
by searching for the GWB generated by the binary PBHs and the SIGW inevitably
produced by the curvature perturbations during the formation of PBHs.
Acknowledgments. We acknowledge the use of HPC Cluster of ITP-CAS and HPC
Cluster of Tianhe II in National Supercomputing Center in Guangzhou. This work
is supported by the National Key Research and Development Program of China
Grant No.2020YFC2201502, grants from NSFC (grant No. 11975019, 11991052,
12047503), Key Research Program of Frontier Sciences, CAS, Grant NO. ZDBS-
LY-7009, CAS Project for Young Scientists in Basic Research YSBR-006, the Key
Research Program of the Chinese Academy of Sciences (Grant NO. XDPB15), and
the science research grants from the China Manned Space Project with NO. CMS-
CSST-2021-B01.
## References
* Carr and Hawking (1974) Bernard J. Carr and S. W. Hawking, “Black holes in the early Universe,” Mon. Not. Roy. Astron. Soc. 168, 399–415 (1974).
* Carr (1975) Bernard J. Carr, “The Primordial black hole mass spectrum,” Astrophys. J. 201, 1–19 (1975).
* Sasaki _et al._ (2016) Misao Sasaki, Teruaki Suyama, Takahiro Tanaka, and Shuichiro Yokoyama, “Primordial Black Hole Scenario for the Gravitational-Wave Event GW150914,” Phys. Rev. Lett. 117, 061101 (2016), [erratum: Phys. Rev. Lett.121,no.5,059901(2018)], arXiv:1603.08338 [astro-ph.CO] .
* Chen and Huang (2018) Zu-Cheng Chen and Qing-Guo Huang, “Merger Rate Distribution of Primordial-Black-Hole Binaries,” Astrophys. J. 864, 61 (2018), arXiv:1801.10327 [astro-ph.CO] .
* Raidal _et al._ (2019) Martti Raidal, Christian Spethmann, Ville Vaskonen, and Hardi Veermäe, “Formation and Evolution of Primordial Black Hole Binaries in the Early Universe,” JCAP 1902, 018 (2019), arXiv:1812.01930 [astro-ph.CO] .
* De Luca _et al._ (2020a) V. De Luca, G. Franciolini, P. Pani, and A. Riotto, “Primordial Black Holes Confront LIGO/Virgo data: Current situation,” JCAP 06, 044 (2020a), arXiv:2005.05641 [astro-ph.CO] .
* Hall _et al._ (2020) Alex Hall, Andrew D. Gow, and Christian T. Byrnes, “Bayesian analysis of LIGO-Virgo mergers: Primordial vs. astrophysical black hole populations,” Phys. Rev. D 102, 123524 (2020), arXiv:2008.13704 [astro-ph.CO] .
* Bhagwat _et al._ (2021) S. Bhagwat, V. De Luca, G. Franciolini, P. Pani, and A. Riotto, “The importance of priors on LIGO-Virgo parameter estimation: the case of primordial black holes,” JCAP 01, 037 (2021), arXiv:2008.12320 [astro-ph.CO] .
* Hütsi _et al._ (2020) Gert Hütsi, Martti Raidal, Ville Vaskonen, and Hardi Veermäe, “Two populations of LIGO-Virgo black holes,” (2020), arXiv:2012.02786 [astro-ph.CO] .
* Wong _et al._ (2021) Kaze W. K. Wong, Gabriele Franciolini, Valerio De Luca, Vishal Baibhav, Emanuele Berti, Paolo Pani, and Antonio Riotto, “Constraining the primordial black hole scenario with Bayesian inference and machine learning: the GWTC-2 gravitational wave catalog,” Phys. Rev. D 103, 023026 (2021), arXiv:2011.01865 [gr-qc] .
* De Luca _et al._ (2021a) V. De Luca, G. Franciolini, P. Pani, and A. Riotto, “Bayesian Evidence for Both Astrophysical and Primordial Black Holes: Mapping the GWTC-2 Catalog to Third-Generation Detectors,” JCAP 05, 003 (2021a), arXiv:2102.03809 [astro-ph.CO] .
* Franciolini _et al._ (2021) Gabriele Franciolini, Vishal Baibhav, Valerio De Luca, Ken K. Y. Ng, Kaze W. K. Wong, Emanuele Berti, Paolo Pani, Antonio Riotto, and Salvatore Vitale, “Quantifying the evidence for primordial black holes in LIGO/Virgo gravitational-wave data,” (2021), arXiv:2105.03349 [gr-qc] .
* Abbott _et al._ (2020a) R. Abbott _et al._ (LIGO Scientific, Virgo), “GW190521: A Binary Black Hole Merger with a Total Mass of $150M_{\odot}$,” Phys. Rev. Lett. 125, 101102 (2020a), arXiv:2009.01075 [gr-qc] .
* Abbott _et al._ (2020b) B. P. Abbott _et al._ (LIGO Scientific, Virgo), “GW190425: Observation of a Compact Binary Coalescence with Total Mass $\sim 3.4M_{\odot}$,” Astrophys. J. Lett. 892, L3 (2020b), arXiv:2001.01761 [astro-ph.HE] .
* Abbott _et al._ (2020c) R. Abbott _et al._ (LIGO Scientific, Virgo), “GW190814: Gravitational Waves from the Coalescence of a 23 Solar Mass Black Hole with a 2.6 Solar Mass Compact Object,” Astrophys. J. Lett. 896, L44 (2020c), arXiv:2006.12611 [astro-ph.HE] .
* Abbott _et al._ (2021a) R. Abbott _et al._ (LIGO Scientific, KAGRA, VIRGO), “Observation of Gravitational Waves from Two Neutron Star–Black Hole Coalescences,” Astrophys. J. Lett. 915, L5 (2021a), arXiv:2106.15163 [astro-ph.HE] .
* De Luca _et al._ (2021b) V. De Luca, V. Desjacques, G. Franciolini, P. Pani, and A. Riotto, “GW190521 Mass Gap Event and the Primordial Black Hole Scenario,” Phys. Rev. Lett. 126, 051101 (2021b), arXiv:2009.01728 [astro-ph.CO] .
* Abbott _et al._ (2021b) R. Abbott _et al._ (LIGO Scientific, VIRGO, KAGRA), “GWTC-3: Compact Binary Coalescences Observed by LIGO and Virgo During the Second Part of the Third Observing Run,” (2021b), arXiv:2111.03606 [gr-qc] .
* Byrnes _et al._ (2018) Christian T. Byrnes, Mark Hindmarsh, Sam Young, and Michael R. S. Hawkins, “Primordial black holes with an accurate QCD equation of state,” JCAP 08, 041 (2018), arXiv:1801.06138 [astro-ph.CO] .
* Chen _et al._ (2019) Zu-Cheng Chen, Fan Huang, and Qing-Guo Huang, “Stochastic Gravitational-wave Background from Binary Black Holes and Binary Neutron Stars and Implications for LISA,” Astrophys. J. 871, 97 (2019), arXiv:1809.10360 [gr-qc] .
* Saito and Yokoyama (2009) Ryo Saito and Jun’ichi Yokoyama, “Gravitational wave background as a probe of the primordial black hole abundance,” Phys. Rev. Lett. 102, 161101 (2009), [Erratum: Phys. Rev. Lett.107,069901(2011)], arXiv:0812.4339 [astro-ph] .
* Ali-Haïmoud _et al._ (2017) Yacine Ali-Haïmoud, Ely D. Kovetz, and Marc Kamionkowski, “Merger rate of primordial black-hole binaries,” Phys. Rev. D96, 123523 (2017), arXiv:1709.06576 [astro-ph.CO] .
* Abbott _et al._ (2016a) B. P. Abbott _et al._ (LIGO Scientific, Virgo), “The Rate of Binary Black Hole Mergers Inferred from Advanced LIGO Observations Surrounding GW150914,” Astrophys. J. Lett. 833, L1 (2016a), arXiv:1602.03842 [astro-ph.HE] .
* Abbott _et al._ (2016b) B. P. Abbott _et al._ (LIGO Scientific, Virgo), “Supplement: The Rate of Binary Black Hole Mergers Inferred from Advanced LIGO Observations Surrounding GW150914,” Astrophys. J. Suppl. 227, 14 (2016b), arXiv:1606.03939 [astro-ph.HE] .
* Abbott _et al._ (2016c) B. P. Abbott _et al._ (LIGO Scientific, Virgo), “Binary Black Hole Mergers in the first Advanced LIGO Observing Run,” Phys. Rev. X6, 041015 (2016c), [erratum: Phys. Rev.X8,no.3,039903(2018)], arXiv:1606.04856 [gr-qc] .
* Wysocki _et al._ (2019) Daniel Wysocki, Jacob Lange, and Richard O’Shaughnessy, “Reconstructing phenomenological distributions of compact binaries via gravitational wave observations,” Phys. Rev. D 100, 043012 (2019), arXiv:1805.06442 [gr-qc] .
* Fishbach _et al._ (2018) Maya Fishbach, Daniel E. Holz, and Will M. Farr, “Does the Black Hole Merger Rate Evolve with Redshift?” Astrophys. J. Lett. 863, L41 (2018), arXiv:1805.10270 [astro-ph.HE] .
* Mandel _et al._ (2019) Ilya Mandel, Will M. Farr, and Jonathan R. Gair, “Extracting distribution parameters from multiple uncertain observations with selection biases,” Mon. Not. Roy. Astron. Soc. 486, 1086–1093 (2019), arXiv:1809.02063 [physics.data-an] .
* Thrane and Talbot (2019) Eric Thrane and Colm Talbot, “An introduction to Bayesian inference in gravitational-wave astronomy: parameter estimation, model selection, and hierarchical models,” Publ. Astron. Soc. Austral. 36, e010 (2019), [Erratum: Publ.Astron.Soc.Austral. 37, e036 (2020)], arXiv:1809.02293 [astro-ph.IM] .
* De Luca _et al._ (2019a) V. De Luca, V. Desjacques, G. Franciolini, A. Malhotra, and A. Riotto, “The initial spin probability distribution of primordial black holes,” JCAP 05, 018 (2019a), arXiv:1903.01179 [astro-ph.CO] .
* Mirbabayi _et al._ (2020) Mehrdad Mirbabayi, Andrei Gruzinov, and Jorge Noreña, “Spin of Primordial Black Holes,” JCAP 03, 017 (2020), arXiv:1901.05963 [astro-ph.CO] .
* De Luca _et al._ (2020b) V. De Luca, G. Franciolini, P. Pani, and A. Riotto, “The evolution of primordial black holes and their final observable spins,” JCAP 04, 052 (2020b), arXiv:2003.02778 [astro-ph.CO] .
* De Luca _et al._ (2020c) V. De Luca, G. Franciolini, P. Pani, and A. Riotto, “Constraints on Primordial Black Holes: the Importance of Accretion,” Phys. Rev. D 102, 043505 (2020c), arXiv:2003.12589 [astro-ph.CO] .
* Press and Schechter (1974) William H. Press and Paul Schechter, “Formation of galaxies and clusters of galaxies by selfsimilar gravitational condensation,” Astrophys. J. 187, 425–438 (1974).
* Choptuik (1993) Matthew W. Choptuik, “Universality and scaling in gravitational collapse of a massless scalar field,” Phys. Rev. Lett. 70, 9–12 (1993).
* Evans and Coleman (1994) Charles R. Evans and Jason S. Coleman, “Observation of critical phenomena and selfsimilarity in the gravitational collapse of radiation fluid,” Phys. Rev. Lett. 72, 1782–1785 (1994), arXiv:gr-qc/9402041 .
* Niemeyer and Jedamzik (1998) Jens C. Niemeyer and K. Jedamzik, “Near-critical gravitational collapse and the initial mass function of primordial black holes,” Phys. Rev. Lett. 80, 5481–5484 (1998), arXiv:astro-ph/9709072 .
* Koike _et al._ (1995) Tatsuhiko Koike, Takashi Hara, and Satoshi Adachi, “Critical behavior in gravitational collapse of radiation fluid: A Renormalization group (linear perturbation) analysis,” Phys. Rev. Lett. 74, 5170–5173 (1995), arXiv:gr-qc/9503007 .
* Young _et al._ (2019) Sam Young, Ilia Musco, and Christian T. Byrnes, “Primordial black hole formation and abundance: contribution from the non-linear relation between the density and curvature perturbation,” JCAP 11, 012 (2019), arXiv:1904.00984 [astro-ph.CO] .
* De Luca _et al._ (2019b) V. De Luca, G. Franciolini, A. Kehagias, M. Peloso, A. Riotto, and C. Ünal, “The Ineludible non-Gaussianity of the Primordial Black Hole Abundance,” JCAP 07, 048 (2019b), arXiv:1904.00970 [astro-ph.CO] .
* Kawasaki and Nakatsuka (2019) Masahiro Kawasaki and Hiromasa Nakatsuka, “Effect of nonlinearity between density and curvature perturbations on the primordial black hole formation,” Phys. Rev. D 99, 123501 (2019), arXiv:1903.02994 [astro-ph.CO] .
* Musco _et al._ (2005) Ilia Musco, John C. Miller, and Luciano Rezzolla, “Computations of primordial black hole formation,” Class. Quant. Grav. 22, 1405–1424 (2005), arXiv:gr-qc/0412063 .
* Musco _et al._ (2009) Ilia Musco, John C. Miller, and Alexander G. Polnarev, “Primordial black hole formation in the radiative era: Investigation of the critical nature of the collapse,” Class. Quant. Grav. 26, 235001 (2009), arXiv:0811.1452 [gr-qc] .
* Musco and Miller (2013) Ilia Musco and John C. Miller, “Primordial black hole formation in the early universe: critical behaviour and self-similarity,” Class. Quant. Grav. 30, 145009 (2013), arXiv:1201.2379 [gr-qc] .
* Saikawa and Shirai (2018) Ken’ichi Saikawa and Satoshi Shirai, “Primordial gravitational waves, precisely: The role of thermodynamics in the Standard Model,” JCAP 05, 035 (2018), arXiv:1803.01038 [hep-ph] .
* Borsanyi _et al._ (2016) Sz. Borsanyi _et al._ , “Calculation of the axion mass based on high-temperature lattice quantum chromodynamics,” Nature 539, 69–71 (2016), arXiv:1606.07494 [hep-lat] .
* Mukherjee and Silk (2021) Suvodip Mukherjee and Joseph Silk, “Can we distinguish astrophysical from primordial black holes via the stochastic gravitational wave background?” Mon. Not. Roy. Astron. Soc. 506, 3977–3985 (2021), arXiv:2105.11139 [gr-qc] .
* Mukherjee _et al._ (2021) Suvodip Mukherjee, Matthew S. P. Meinema, and Joseph Silk, “Prospects of discovering sub-solar primordial black holes using the stochastic gravitational wave background from third-generation detectors,” (2021), arXiv:2107.02181 [astro-ph.CO] .
* Abbott _et al._ (2021c) R. Abbott _et al._ (KAGRA, Virgo, LIGO Scientific), “Upper limits on the isotropic gravitational-wave background from Advanced LIGO and Advanced Virgo’s third observing run,” Phys. Rev. D 104, 022004 (2021c), arXiv:2101.12130 [gr-qc] .
* Arzoumanian _et al._ (2020) Zaven Arzoumanian _et al._ (NANOGrav), “The NANOGrav 12.5 yr Data Set: Search for an Isotropic Stochastic Gravitational-wave Background,” Astrophys. J. Lett. 905, L34 (2020), arXiv:2009.04496 [astro-ph.HE] .
* Nan _et al._ (2011) Rendong Nan, Di Li, Chengjin Jin, Qiming Wang, Lichun Zhu, Wenbai Zhu, Haiyan Zhang, Youling Yue, and Lei Qian, “The Five-Hundred-Meter Aperture Spherical Radio Telescope (FAST) Project,” Int. J. Mod. Phys. D 20, 989–1024 (2011), arXiv:1105.3794 [astro-ph.IM] .
* Audley _et al._ (2017) Heather Audley _et al._ (LISA), “Laser Interferometer Space Antenna,” (2017), arXiv:1702.00786 [astro-ph.IM] .
* Luo _et al._ (2016) Jun Luo _et al._ (TianQin), “TianQin: a space-borne gravitational wave detector,” Class. Quant. Grav. 33, 035010 (2016), arXiv:1512.02076 [astro-ph.IM] .
* Hu and Wu (2017) Wen-Rui Hu and Yue-Liang Wu, “The Taiji Program in Space for gravitational wave physics and the nature of gravity,” National Science Review 4, 685–686 (2017), https://academic.oup.com/nsr/article-pdf/4/5/685/31566708/nwx116.pdf .
* Punturo _et al._ (2010) M. Punturo _et al._ , “The Einstein Telescope: A third-generation gravitational wave observatory,” _Proceedings, 14th Workshop on Gravitational wave data analysis (GWDAW-14): Rome, Italy, January 26-29, 2010_ , Class. Quant. Grav. 27, 194002 (2010).
* Abbott _et al._ (2017) Benjamin P. Abbott _et al._ (LIGO Scientific), “Exploring the Sensitivity of Next Generation Gravitational Wave Detectors,” Class. Quant. Grav. 34, 044001 (2017), arXiv:1607.08697 [astro-ph.IM] .
* Ackley _et al._ (2020) K. Ackley _et al._ , “Neutron Star Extreme Matter Observatory: A kilohertz-band gravitational-wave detector in the global network,” Publ. Astron. Soc. Austral. 37, e047 (2020), arXiv:2007.03128 [astro-ph.HE] .
* Allen and Romano (1999) Bruce Allen and Joseph D. Romano, “Detecting a stochastic background of gravitational radiation: Signal processing strategies and sensitivities,” Phys. Rev. D 59, 102001 (1999), arXiv:gr-qc/9710117 .
* Phinney (2001) E. S. Phinney, “A Practical theorem on gravitational wave backgrounds,” astro-ph/0108028 (2001).
* Regimbau and Mandic (2008) T. Regimbau and V. Mandic, “Astrophysical Sources of Stochastic Gravitational-Wave Background,” _Proceedings, 12th Workshop on Gravitational wave data analysis (GWDAW-12): Cambridge, USA, December 13-16, 2007_ , Class. Quant. Grav. 25, 184018 (2008), arXiv:0806.2794 [astro-ph] .
* Zhu _et al._ (2011) Xing-Jiang Zhu, E. Howell, T. Regimbau, D. Blair, and Zong-Hong Zhu, “Stochastic Gravitational Wave Background from Coalescing Binary Black Holes,” Astrophys. J. 739, 86 (2011), arXiv:1104.3565 [gr-qc] .
* Zhu _et al._ (2013) Xing-Jiang Zhu, Eric J. Howell, David G. Blair, and Zong-Hong Zhu, “On the gravitational wave background from compact binary coalescences in the band of ground-based interferometers,” Mon. Not. Roy. Astron. Soc. 431, 882–899 (2013), arXiv:1209.0595 [gr-qc] .
* Aghanim _et al._ (2020) N. Aghanim _et al._ (Planck), “Planck 2018 results. VI. Cosmological parameters,” Astron. Astrophys. 641, A6 (2020), arXiv:1807.06209 [astro-ph.CO] .
* Cutler _et al._ (1993) C. Cutler, Eric Poisson, G. J. Sussman, and L. S. Finn, “Gravitational radiation from a particle in circular orbit around a black hole. 2: Numerical results for the nonrotating case,” Phys. Rev. D47, 1511–1518 (1993).
* Chernoff and Finn (1993) David F. Chernoff and Lee Samuel Finn, “Gravitational radiation, inspiraling binaries, and cosmology,” Astrophys. J. 411, L5–L8 (1993), arXiv:gr-qc/9304020 [gr-qc] .
* Tomita (1967) Kenji Tomita, “Non-linear theory of gravitational instability in the expanding universe,” Progress of Theoretical Physics 37, 831–846 (1967).
* Matarrese _et al._ (1993) Sabino Matarrese, Ornella Pantano, and Diego Saez, “A General relativistic approach to the nonlinear evolution of collisionless matter,” Phys. Rev. D47, 1311–1323 (1993).
* Matarrese _et al._ (1994) Sabino Matarrese, Ornella Pantano, and Diego Saez, “General relativistic dynamics of irrotational dust: Cosmological implications,” Phys. Rev. Lett. 72, 320–323 (1994), arXiv:astro-ph/9310036 [astro-ph] .
* Matarrese _et al._ (1998) Sabino Matarrese, Silvia Mollerach, and Marco Bruni, “Second order perturbations of the Einstein-de Sitter universe,” Phys. Rev. D58, 043504 (1998), arXiv:astro-ph/9707278 [astro-ph] .
* Noh and Hwang (2004) Hyerim Noh and Jai-chan Hwang, “Second-order perturbations of the Friedmann world model,” Phys. Rev. D69, 104011 (2004).
* Carbone and Matarrese (2005) Carmelita Carbone and Sabino Matarrese, “A Unified treatment of cosmological perturbations from super-horizon to small scales,” Phys. Rev. D71, 043508 (2005), arXiv:astro-ph/0407611 [astro-ph] .
* Nakamura (2007) Kouji Nakamura, “Second-order gauge invariant cosmological perturbation theory: Einstein equations in terms of gauge invariant variables,” Prog. Theor. Phys. 117, 17–74 (2007), arXiv:gr-qc/0605108 [gr-qc] .
* Yuan _et al._ (2020) Chen Yuan, Zu-Cheng Chen, and Qing-Guo Huang, “Scalar induced gravitational waves in different gauges,” Phys. Rev. D 101, 063018 (2020), arXiv:1912.00885 [astro-ph.CO] .
* Yuan _et al._ (2019a) Chen Yuan, Zu-Cheng Chen, and Qing-Guo Huang, “Probing Primordial-Black-Hole Dark Matter with Scalar Induced Gravitational Waves,” Phys. Rev. D100, 081301 (2019a), arXiv:1906.11549 [astro-ph.CO] .
* Yuan _et al._ (2019b) Chen Yuan, Zu-Cheng Chen, and Qing-Guo Huang, “Log-dependent slope of scalar induced gravitational waves in the infrared regions,” (2019b), arXiv:1910.09099 [astro-ph.CO] .
* Yuan and Huang (2021) Chen Yuan and Qing-Guo Huang, “A topic review on probing primordial black hole dark matter with scalar induced gravitational waves,” (2021), arXiv:2103.04739 [astro-ph.GA] .
* Kohri and Terada (2018) Kazunori Kohri and Takahiro Terada, “Semianalytic calculation of gravitational wave spectrum nonlinearly induced from primordial curvature perturbations,” Phys. Rev. D97, 123532 (2018), arXiv:1804.08577 [gr-qc] .
* Chen _et al._ (2021a) Zu-Cheng Chen, Chen Yuan, and Qing-Guo Huang, “Non-tensorial gravitational wave background in NANOGrav 12.5-year data set,” Sci. China Phys. Mech. Astron. 64, 120412 (2021a), arXiv:2101.06869 [astro-ph.CO] .
* Arzoumanian _et al._ (2021) Zaven Arzoumanian _et al._ (NANOGrav), “The NANOGrav 12.5-year Data Set: Search for Non-Einsteinian Polarization Modes in the Gravitational-wave Background,” Astrophys. J. Lett. 923, L22 (2021), arXiv:2109.14706 [gr-qc] .
* Goncharov _et al._ (2021) Boris Goncharov _et al._ , “On the evidence for a common-spectrum process in the search for the nanohertz gravitational-wave background with the Parkes Pulsar Timing Array,” (2021), 10.3847/2041-8213/ac17f4, arXiv:2107.12112 [astro-ph.HE] .
* Wu _et al._ (2021) Yu-Mei Wu, Zu-Cheng Chen, and Qing-Guo Huang, “Constraining the Polarization of Gravitational Waves with the Parkes Pulsar Timing Array Second Data Release,” (2021), arXiv:2108.10518 [astro-ph.CO] .
* Chen _et al._ (2021b) Zu-Cheng Chen, Yu-Mei Wu, and Qing-Guo Huang, “Searching for Isotropic Stochastic Gravitational-Wave Background in the International Pulsar Timing Array Second Data Release,” (2021b), arXiv:2109.00296 [astro-ph.CO] .
* Chalumeau _et al._ (2021) A. Chalumeau _et al._ , “Noise analysis in the European Pulsar Timing Array data release 2 and its implications on the gravitational-wave background search,” (2021), 10.1093/mnras/stab3283, arXiv:2111.05186 [astro-ph.HE] .
|
(7.22$\pm$14.25 km s-1) F38 Sinfoni 14/04/11 (-3.69$\pm$5.60 km s-1), 17/07/13
(-14.65$\pm$6.12 km s-1), 06/08/13 (-14.28$\pm$5.89 km s-1), 26/04/17
(2.17$\pm$5.99 km s-1), 30/04/17 (-7.13$\pm$5.11 km s-1), 07/05/18
(5.92$\pm$5.13 km s-1), 07/08/18 (6.23$\pm$5.40 km s-1), 12/08/18
(10.94$\pm$5.99 km s-1), 17/08/18 (low S/N), 18/08/18 (14.49$\pm$6.09 km s-1)
F40 Sinfoni 29/05/13 (-9.57$\pm$9.51 km s-1), 18/06/13 (-3.30$\pm$18.00 km
s-1), 17/07/13 (-20.06$\pm$11.19 km s-1), 03/08/13 (-2.71$\pm$10.88 km s-1),
04/08/13 (-0.87$\pm$10.08 km s-1), 06/08/13 (0.90$\pm$11.02 km s-1), 08/08/13
(-18.25$\pm$14.85 km s-1), 26/04/17 (18.17$\pm$11.69 km s-1), 30/04/17
(19.44$\pm$9.69 km s-1), 23/06/17 (-28.02$\pm$10.18 km s-1), 27/07/17
(19.52$\pm$9.07 km s-1), 07/05/18 (-5.06$\pm$10.54 km s-1), 07/08/18
(-4.57$\pm$14.84 km s-1), 12/08/18 (8.52$\pm$12.17 km s-1), 18/08/18
(25.86$\pm$11.06 km s-1)
## Appendix B The LBV phenomenon and secular variability within the Arches
As massive stars transition away from the Main Sequence they appear to
encounter an instability of unknown origin: the so-called luminous blue
variable (LBV) phase. Historically it has been thought that the extreme mass
loss that characterises LBVs helps facilitate the formation of H-depleted WRs,
although observations of Romano’s star (Clark et al. [2012], Polcaro et al.
[2016]) and the erupting component of the eclipsing hierarchical system HD5980
indicate this behaviour may also be present in WN-subtype WRs. The LBV
phenomenon appears to extend to very high masses, with the LBV component of
HD5980 having a dynamical mass of M${}_{\rm current}\sim 60M_{\odot}$,
implying M${}_{\rm initial}>80M_{\odot}$ (Koenigsberger et al. [2014]; Hillier
et al. [2019]). However, this extreme is exceeded by that of the primary in
the binary $\eta$ Carinae, which is generally supposed to be $>100M_{\odot}$
(Hillier et al. [2001]), with a mass of $\sim 40-60M_{\odot}$ suggested for
the unseen secondary (Mehner et al. [2010]). $\eta$ Carinae has been the
subject of intensive observation due to both its current exceptional
properties and its unprecedented eruption in the $19^{\rm th}$ Century, during
which a luminosity of $\gtrsim 10^{7}L_{\odot}$ and a mass loss rate of
$\gtrsim 0.5M_{\odot}$ yr-1 have been inferred (Smith et al. [2003]).
$\eta$ Carinae is located within the Trumpler 16 + Collinder 228 cluster
aggregate, for which Smith ([2006]) infers an age of 2-3Myr. As such -
although of significantly lower integrated mass - it appears directly
comparable to the Arches; indeed it hosts three WNLha stars, a single O4
supergiant and a number of mid-O giants and main sequence stars (cf. papers I
and III). Given this similarity one might ask whether the Arches hosts any
star exhibiting the LBV phenomenon? Trivially, despite the presence of
similarly massive binaries within the Arches, there appears to be no direct
analogue of $\eta$ Carinae; its K-band morphology is very different from any
cluster member (Hamann et al. [1994]), while any nebula comparable to the
Homunculus would be clearly visible (cf. the Pistol nebula; Figer et al.
[1999]).
Surprisingly, none of the stars within our sample showed any evidence of long
term changes in their spectral morphologies during this period. Simulations of
$60M_{\odot}$ stars by Groh et al. ([2014]) predict that the LBV phenomenon
occurs directly after a late-O/early-B hypergiant phase around $\gtrsim 3$Myr,
implying that the cohort of mid-O supergiants and, potentially, the mid- to
late-hypergiants have yet to reach this evolutionary juncture. Unfortunately
no predictions exist for when, or even if, more massive stars such as the
Arches WNLha cohort become LBVs, nor how long this episode endures. Given the
limited cadence of the observations, one could imagine that if the
outbursts/eruption duty cycle were long, or outbursts rapid, they could have
been missed. Indeed, while $\eta$ Carinae was highly (photometrically) active
throughout the 19th and 20th Centuries (Smith & Frew [2011]) and the
characteristic variability of LBVs occurs over months to years (e.g. Humphreys
& Davidson [1994]), observations of P Cygni reveal that LBVs may reside in a
quiescent state for centuries (cf Lamers & de Groot [1992]). Alternatively one
might suppose that WNLha stars have yet to reach the LBV phase, or that such
stars circumvent it entirely. However such a hypothesis would require the 19th
Century eruption of $\eta$ Carina to be driven by a mechanism other than the
nuclear evolution of the star, such as binary interaction/merger (Portegies
Zwart & van den Heuvel [2015]).
## Appendix C The X-ray and binary properties of WNLh stars
In order to better understand the Arches binaries it is instructive to
contrast their properties with those of other known binary populations. Young
massive clusters provide the perfect laboratory to undertake such a
comparison, since their ensemble properties allow for constraints to be placed
on their distances, extinction and ages, in turn aiding the determination of
properties such as luminosity for individual members. Moreover, selection
effects may be more easily understood and controlled in such an environment.
When assessing the binary fraction of clusters, multi-epoch spectroscopic
surveys clearly represent the most powerful methodology available, especially
if combined with photometric observations. However not all the clusters that
host WNLha stars have been subject to such an approach, and in such cases
X-ray observations - which are sensitive to the signature of shocked material
generated by wind collisions - are an excellent substitute.
In order to undertake such an assessment, the X-ray properties of both single
and binary WNLh stars must first be quantified. Clark et al. ([2009])
presented such a summary; however a reappraisal is timely, and we provide that
here. For the reasons described above, stars located within stellar aggregates
form a ‘gold standard’; moreover many more such regions have now been subject
to targeted X-ray investigation (Sect. C.1.1 and refs. therein). Nevertheless,
we supplement the resultant sample with the remaining, isolated examples for
which suitable data exist, an inclusion driven by the possibility of better
understanding the intrinsic multi-wavelength properties of single stars (Sect.
C.1.2). For the avoidance of bias, we emphasise that the isolated stellar
cohort are not included in the subsequent determination of binary fractions.
We also highlight that a large number of WNLha stars have been identified
within the the central molecular zone of the Galaxy (Clark et al. [2021]).
Unlike the cohorts listed above these were in large part identified via X-ray
observations; as a consequence of the obvious resultant bias towards detecting
colliding wind binaries we explicitly exclude them from the analysis presented
here.
### C.1 Data collation and presentation
In the following section we present the results of a literature review
undertaken to determine both the X-ray and binary properties of WNLha stars,
encompassing both isolated galactic examples and the larger population found
within star clusters and associations. In order to produce the mostly complete
dataset possible, so as to interpret their X-ray behaviour, we also present
the preliminary results of tailored model-atmosphere analysis (cf. Sect. 4) of
the X-ray bright WNLha stars found within the Galactic clusters Danks 1 and
Mercer 81 (Najarro et al. in prep.).
#### C.1.1 WNLh stars within stellar aggregates
Given their rich stellar populations the young massive cluster R136 and its
host star forming region 30 Doradus are obvious targets for such an approach.
Unfortunately, their location within the Large Magellanic Cloud leads to
significant blending in the crowded core regions of R136, and limits spectral
analysis to the brightest of the X-ray sources. As a consequence we also
review the contents of Galactic clusters, following the compilation of Rate et
al. ([2020b]). From an initial sample of 19 candidate clusters, a number were
rejected due to spectral re-classification of relevant cluster members on the
basis of published data101010 Cl 1813-178 (Messineo et al. [2011]) and VVV
CL009 (Chené et al. [2013]). and/or the cluster population suggesting an age
at which, by analogy with the Quintuplet and Westerlund 1 (Clark et al.
[2018b], [2020]), core-H burning WNLha stars should not be present111111The
Quartet (Messineo et al. [2009]), VVV CL036, VVV CL074, and VVV CL099 (Chené
et al. [2013]).. Four further clusters host WNLha stars but lack the multi-
epoch spectroscopic, X-ray and radio observations required to identify
binaries121212 [DBS2003]-179 (WR84-1, WR84-6, and WR84-7; Borissova et al.
[2012]), VVV CL041 (WR62-2; Chené et al. [2015]), VVV CL073 (WR75-26; Chené et
al. [2013]), and Mercer 23 (WR125-3; Hanson et al. [2010]). This left ten
clusters hosting WNLha stars for which the requisite observational data were
available, in addition to R136 and 30 Dor131313While 13 WNLha stars, including
the binary CXOU J1745-28, have currently been identified within the Central
Molecular Zone of the Galaxy (Mikles et al. [2006], Clark et al. [2009]) we do
not consider them here since a large number have been selected on the basis of
their X-ray properties, potentially biasing the sample towards colliding wind
binaries.. The binary and X-ray properties of the WNLh (and related) stars in
these clusters are summarised in Tables C.1 and C.2, while the thumbnail
sketches below present a synopsis of the physical properties in each
aggregate, including relevant references and conclusions derived from the
reported observations.
Table 8: The X-ray properties of WNha stars in 30 Doradus and galactic
clusters
Cluster | Dist. | ID | log | $L_{\rm x}$ | $kT_{1}$ | $kT_{2}$ | log
---|---|---|---|---|---|---|---
| (kpc) | | $(L_{\rm bol}/L_{\odot})$ | ($10^{32}$ergs-1) | (keV) | (keV) | $(L_{\rm X}/L_{\rm bol})$
R136 | $\sim 50$ | R136a1 | 6.94a | 100 | - | 1.2 | -
| | +a2 | 6.63a | | | |
| | R136a3 | 6.58a | 70 | - | 4.2 | -
| | +a6 | 6.52a | | | |
| | R136c∗ | 6.75a | 700 | - | 3.0 | -5.5
30 Dor | $\sim 50$ | Mk30∗ | 6.16a | $<10$ | - | - | $<-6.8$
| | Mk34∗ | 6.70b | 1000 | 1.2 | 4.5 | -5.3
| | Mk35 | 6.3a | 10 | - | - | -6.9
| | Mk37∗ | 6.48a | 100 | - | - | -6.1
| | Mk37Wb | 6.21a | $<10$ | - | - | $<-6.8$
| | Mk39 | 6.4a | 130 | - | 2.0 | -5.9
| | Mk51 | 6.2a | $<10$ | - | - | $<-6.8$
| | RMC134 | 6.2a | 7 | - | - | -7.0
| | RMC135∗ | 6.2a | $<6.3$ | - | - | $<-7.0$
| | RMC140b∗ | 6.4a | 30 | - | 2.5 | -6.5
| | RMC144∗ | 6.72c | 25 | - | 2.1 | -6.9
| | RMC145∗ | 6.5a | 19 | - | 1.6 | $-6.8$
| | RMC146 | 6.29a | $<10$ | - | - | $<-6.9$
| | RMC147 | 6.36a | $<10$ | - | - | $<-7.0$
| | VFTS682 | 6.51a | $<10$ | - | - | $<-7.1$
Arches | $\sim 8$ | F6∗ | 6.32d | 110 | - | 2.2 | -5.9
| | F7∗ | 6.27d | 72 | - | 1.8 | -6.0
| | F9 | 6.15e | 46 | - | 2.5 | -6.1
Bochum 7 | $\sim$6.0 | WR12∗ | 5.98f | $<0.4$ | - | - | $<-8.0$
Westerlund 2 | $\sim 4.2$ | WR20a∗ | 6.3g | 57 | 0.4 | 1.6 | -6.1
| | WR20b | - | 120 | 0.4 | 5.4 | -
| | [WR20aa | - | 24 | 0.9 | - | -]
| | [WR21a* | - | 194 | 0.8 | 3.0 | -]
Carina | $\sim 2.6$ | WR22∗ | 6.35f | 4.2 | 0.6 | 2.0 | -7.3
| | WR24 | 6.19f | 6.4 | 0.7 | 3.3 | -7.0
| | WR25∗ | 6.61f | 84 | 0.6 | 2.7 | -6.3
NGC 3603 | $\sim 7.6$ | WR43A∗ | 6.39h | 360 | 0.8 | 3.1 | -
| | +WR43B | 6.46h | | | |
| | WR43C∗ | 6.35h | 260 | 1.0 | 7.9 | -5.5
| | [WR42e | - | 2.3 | - | - | \- ]
| | [WR43-2∗ | - | 2.9 | - | - | \- ]
Danks 1 | $\sim 3.8$ | WR48-8 | 5.44e | 17 | - | - | -5.8
| | WR48-9 | 5.99e | 43 | - | - | -6.0
| | WR48-10 | 5.70e | 14 | - | - | -6.2
Mercer 81 | $\sim 10$ | WR76-3 | 6.17e | 14 | - | 2.0 | -6.6
| | WR76-7 | 6.39e | 35 | - | 2.6 | -6.4
Sco OB1 | $\sim$2 | WR78∗ | 5.8f | 6.9 | 0.6 | 2.3 | -6.6
| | WR79a∗ | 5.78i | 1.4 | 0.6 | 2.7 | -7.2
HM-1 | $\sim 3.3$ | WR87 | 6.21f | 1.6 | - | 2 | -7.6
| | WR89∗ | 6.33f | 154 | 0.6 | 2 | -5.7
W43 | $\sim 6$ | WR121a∗ | - | 154 | 1.0 | 3.6 | -
Sources in the vicinity of the clusters but for which a physical association
has yet to be proved are given in square brackets. Identifiers marked with an
asterisk are those which are identified as binaries via optical or IR
observations, non-thermal radio emission (WR78 and WR89) or periodically
modulated X-ray activity (WR121a). Variable X-ray sources have their
representative fluxes given in italics. Parameters for R136a1+a2, R136a3+a6
and WR43A+WR43B represent a blend of both sources (see text for further
details). References for the X-ray properties are given in the body of the
text, while those for the bolometric luminosities of the system primaries are
aCrowther et al. ([2016]), bTehrani et al. ([2019]), cShenar et al. ([2021]),
dthis work, eNajarro et al. (in prep.), fHamann et al. ([2019]), gRauw et al.
([2005]), hCrowther et al. ([2010]), and iCrowther & Bohannan ([1997]). We
highlight that Mk34, RMC144, and WR20a are composed of two ‘twins’ - as with
other systems the bolometric luminosities presented are those of the combined
binary; individual components in these systems will be 0.3dex fainter. Finally
we have scaled the bolometric luminosities for WR24, WR24, and WR25 (Hamann et
al. ([2019]) to a common distance of 2.6kpc for the Carina complex.
Table 9: The X-ray properties of isolated WNha stars
ID | Dist. | log | $L_{\rm x}$ | log
---|---|---|---|---
| (kpc) | $(L_{\rm bol}/L_{\odot})$ | ($10^{32}$ergs-1) | $(L_{\rm X}/L_{\rm bol})$
WR16 | 2.6 | 5.72a | 0.13 | -8.2
WR29${}^{b}*$ | 6.0 | 5.73b | 2 - 3 | -7.1 - -6.9
WR40 | 4.0 | 5.91a | $<0.7$ | $<-7.7$
WR105∗ | 1.8 | 5.89a | 4 - 12 | -6.9 - -6.4
WR148∗ | 8.3 | 6.3a,c | 2.1 | -7.6
Identifiers marked with an asterisk are those which are identified as binaries
via optical (WR29 and WR148) and radio (WR105) observations. References for
the X-ray properties are given in the body of the text, while those for the
bolometric luminosities of the system primaries are aHamann et al. ([2019]),
bNaze et al. ([2020]) and cZhekov ([2017]). Note the X-ray luminosities listed
for WR29 and WR105 represent the range allowed under the modelling assumptions
of Naze et al. ([2020]).
* •
R136: Schnurr et al. ([2009b]) undertook the most comprehensive RV survey of
the six WN5h stars within R136 to date, but report only a single potential
binary - R136c - with a tentative $\sim 8.4$d period (noting the survey was
only sensitive to periods $\lesssim 40$d). X-ray observations reveal that it
is one of the brightest objects within the LMC and demonstrates a hard
spectrum (Table 8); strongly suggestive of a colliding wind binary (see
below). Of the remaining WN5h stars within R136, Townsley et al. ([2006])
report a blend of emission from R136a1+a2 and R136a3+a6 (the latter being O2
If; Crowther et al. [2016]). Both composite sources are luminous and hard,
noting that only single temperature models were employed due to data quality
(Table 8). Finally, R136b is heavily blended with other core stars and so is
not discussed further, while no mention is made of R136a5 by these authors.
* •
30 Doradus: A further 15 WNLh and related stars are found within the wider
confines of 30 Dor (Crowther et al. [2016]). Of these, seven have been
confirmed as binaries via spectroscopic observations141414Mk30 (O2 If/WN5),
Mk34 (WN6(h)), Mk37 (O3.5 If/WN7), RMC135 (WN6h), RMC140b (WN5h), RMC144
(WN5-6h), and RMC145 (WN6h).; one is an RV variable but lacks an orbital
solution151515Mk35 (O2 If/WN5).; three appear to lack RV variability161616Mk51
(O3.5 If/WN7), RMC134 (WN6(h)), and VFTS682 (WN5h).; and appropriate datasets
are lacking for four171717Mk37Wb (O2 If/WN5), Mk39 (02.5 If/WN6), RMC146
(WN5ha), and RMC147 (WN5h). (Table 3, Schnurr et al. [2008b], Bestenlehner et
al. [2011], Tehrani et al. [2019]). X-ray data have been presented for a
number of these stars, resulting in eight detections. Of these six arise from
confirmed binaries/RV variables (Table 8). The most comprehensive analyses are
of Mk34 and RMC144. Pollock et al. ([2018]) show that the former is an
extremely luminous source (median $L_{\rm X}\sim 10^{35}$ergs-1) that is
modulated on the orbital period, with a spectrum comprising soft and hard
components. Despite having similarly luminous components RMC demonstrates a
much more modest X-ray flux, although it too displays orbital modulation and
an intrinsically hard spectrum (Townsley et al. [2006], Shenar et al. [2021]).
Two further stars - RMC140b and RMC145 - are also luminous and hard sources
($L_{\rm X}\sim 10^{33}-10^{34}$ergs-1; $kT\gtrsim 1.6$keV; noting that only
single temperature models were employed by Townsley et al. [2006]). Mk37, the
remaining binary detection, and the RV variable Mk35 are similarly luminous,
while RMC134 - which appears not to be RV variable - is comparatively faint;
unfortunately no spectral information is available for any of these three
stars (Table 8; Schnurr et al. ([2008b]). Finally, Mk39 has yet to be subject
to RV monitoring, but has X-ray properties directly comparable to confirmed
binaries (Table 8; Townsley et al. [2006]); Schnurr et al. ([2008b]) returned
individual upper limits for the last two binaries (Mk30 and RMC135; Table 8).
This leaves a total of five stars for which we adopt the generic detection
threshold of $10^{33}$ergs-1 reported by Townsley et al. ([2006]) as upper
limits, noting that in practice deviations from this are expected depending on
the local environment of individual stars.
* •
Bochum 7: Bo7 hosts a single WN8h star, WR12, for which Fahed & Moffat
([2012]) present a spectroscopic orbital solution. Pollock et al. ([1995])
report an X-ray detection, albeit of low detection significance ($(10\pm
5)\times 10^{32}$ergs-1 for a distance of 6kpc, subsequently quoted by Ignace
et al. [2000]). However more recent and sensitive observations by Nazé et al.
([2020]) failed to detect it (Table 8), the authors noting that it was
observed at a phase at which the WNLha primary and its dense stellar wind
would have (partially) obscured the secondary and any putative colliding wind
region (see also WR22 below). Clearly future observations will be required to
ascertain the nature of its (variable?) emission although we adopt the upper
limit of Nazé et al. ([2020]) in this work.
* •
Westerlund 2: Two WNLha stars have traditionally been associated with Wd2;
WR20a (WN6ha+WN6ha) and WR20b (WN6h). The former is a well known massive
eclipsing system (Rauw et al. [2005]; Table 3) and is also a luminous, hard
X-ray source that varies on the orbital period (Nazé et al. [2008]); in Table
8 we list the properties for the brightest of the epochs presented by the
authors. They further demonstrate that WR20b is a fainter, harder source
(representative $L_{\rm X}\sim 2\times 10^{33}$ergs-1, $kT\sim 3.6$keV)
although Skinner et al. ([2010]) prefer a two-temperature fit resulting in a
higher luminosity (see Table 8). As a consequence they propose that WR20b is
also a colliding wind system; however spectral monitoring by Rauw et al.
([2011]) apparently excludes short binary periods ($<50$d - subject to the
usual caveats regarding unfavourable orbital inclination). Three further WNLha
stars are found in the periphery of Wd2, for which a physical association is
currently uncertain: WR20aa (O2If∗/WN6), WR20c (O2IF∗/WN6ha), and WR21a
(O3If∗/WN6ha+O3V). The very massive short period binary WR21a (Table 3) is, as
may be anticipated, a highly luminous, hard and periodically variable X-ray
source (Gosset & Naze [2016]). WR20aa and WR20c are possible runaways
originating from Wd2 (Roman-Lopes et al. [2011]). Townsley et al. ([2019])
report that the former appears to be a moderately bright and soft X-ray source
(Table 8), while no X-ray data are available for the latter. In neither case
has RV spectral monitoring been attempted.
* •
Carina: The Carina star-forming complex, encompassing the young massive
clusters Trumpler 16 and Collinder 229, is host to three WNLha stars - WR22
(=HD92740A; WN7+abs +09), WR24 (=HD93131; WN6ha), and WR25 (=HD93162; O2.5
If∗/WN6 + O). X-ray observations have been made of all three systems (Gosset
et al. [2009], Skinner et al. [2010], Raassen et al. [2003], Pollock &
Corcoran [2006]) and are summarised in Table 8. WR22 and WR25 have already
been confirmed as binaries via RV studies, and both sources also appear to
show orbitally modulated X-ray fluxes (Gosset et al. [2009], Arora et al.
[2020]). WR22 is of particularly interest since it is observed to be
particularly faint at the point at which the WNLha and associated wind might
be expected to eclipse both the secondary and any associated wind interaction
zone ($L_{\rm X}\sim 6.3\times 10^{31}$erg-1). Finally, while spectroscopic
evidence for binarity is lacking for WR24, we highlight the close similarity
of its X-ray properties to those of WR22.
* •
NGC 3603: Three WNLha stars are found within the core of NGC3603 - WR43A (=
cl* NGC 3603 BLW A1; WN6ha+WN6ha), WR43B (= cl* NGC 3603 MDS B; WN6ha) and
WR43C (= cl* NGC3603 MDS C; WN6ha + ?), of which WR43A and WR43C are known
short period binaries (Table 3). X-ray observations are complicated by
blending due to the compact nature of NGC3603 but Moffat et al. ([2002]; their
Fig. 4) suggest fluxes of $L_{\rm x}\sim 10^{33}$ergs-1 for WR43B, $L_{\rm
x}\gtrsim 4\times 10^{34}$ergs-1 for WR43C and $L_{\rm x}\gtrsim
10^{34}$ergs-1 for the blend comprising WR43A, cl* NGC3603 BLW A2 (O3 V) and
cl* NGC3603 BLW A3 (O3 III(f∗)); unfortunately no spectral information was
presented for any source. Subsequently, Huenemoerder et al. ([2009]) analysed
the emission spectra deriving from a blend of WR43A+WR43B and from WR43C,
finding both to be highly luminous and demonstrating hard spectral components
(cf. Table 8). Of these Pollock et al. ([2018]) report WR43C as being
variable. Two further WNLha stars are located in the surrounds of NGC3603.
Displaced $\sim 6$pc from the cluster core, Roman-Lopes ([2012]) suggests that
the O2 If∗/WN6h stars WR42e could be a runaway (and hence single);
nevertheless they report a X-ray luminosity that is comparable with the binary
WR22 (Table 8) but no spectral information. Finally WR43-2 (= cl* NGC3603 MTT
58; O2-3.5 If∗/WN5-6ha + ?)) is located within a compact H ii region on the
periphery of the wind-blown bubble encompassing NGC3603, some $\sim$2pc
distant from the cluster core. Again lacking spectral information, it is of
comparable luminosity to WR42e (Roman-Lopes [2013]), but is unambiguously a
short-period binary on the basis of periodic photometric modulation (Jaque
Arancibia et al. [2015]).
* •
Mercer 30: de la Fuente et al. ([2016]) report two WNLh stars within the
cluster Mercer 30 - WN46-5 (= [DNB2016] Mc30-7; WN6h) and WN46-6 (= [DNB2016]
Mc30-8; WN7h), which modelling suggests are highly luminous, and hence massive
(log$L/L_{\odot}\gtrsim 6.0$ and $M_{\rm current}\gtrsim 50M_{\odot}$).
Despite a limited dataset comprising three epochs of spectroscopic
observations, the former is clearly a binary by virtue of $\Delta$RV$\sim
191$km s-1 between observations in 2009 and 2011; unfortunately no X-ray
observations exist for this cluster.
* •
Mercer 81: Davies et al. ([2012a]) and de la Fuente et al. ([2013]) report
five WN7-8ha stars within the distant and highly reddened cluster Mercer 81;
WR76-9 (=[DDN2012] 3), WN76-6 (=[DDN2012] 7), WR76-2 (=[DDN2012] 5), WR76-7
(=[DDN2012] 2), and WR76-3 (=[DDN2012] 8). No radial velocity surveys have
been undertaken but the X-ray survey of Norma Arm by Rahoui et al. ([2014])
report detections WR76-3 and WR76-7 (their sources #1278 and #1279
respectively), finding them to be hard, bright sources (Table 8; assuming a
single temperature fit). Bolometric luminosities for the two X-ray detections
derive from quantitative modelling to be presented in Najarro et al. (in
prep.); we find both stars to be intrinsically highly luminous (Table 8).
* •
Danks 1: In the absence of RV observations, Zhekov et al. ([2017]) presents a
preliminary analysis of X-ray data, quoting fluxes for the three WNLha stars
WR48-8 (= [DCT2012] D1-5; WN8-9ha), WR48-9 (= [DCT2012] D1-1; WN8-9ha), and
WR48-10 (= [DCT2012] D1-2; WN8-9ha), with Townsley et al. ([2019])
subsequently reporting that the first two sources are variable, with fluxes
decreasing by factors of $\sim 5$ and $\sim 10$ respectively, between two
epochs of observations separated by $\sim 4$yr. While neither author presents
spectral information for these detections both conclude that they are bona
fide binaries based on their X-ray properties (a claim we revisit below). As
with Mercer 81 we provide bolometric luminosities for the three X-ray
detections derived from modelling in Table 8 (Najarro et al. in prep.). We
highlight that while WR48-9 and 48-10 are broadly comparable to other Galactic
examples, WR48-8 is significantly fainter, while the emission lines in the
K-band spectrum appear anomalously broad and weak (Davies et al. [2012b]). As
a consequence we suspect its nature and/or evolutionary pathway differs from
other examples; we return to this topic in Najarro et al. (in prep.).
* •
Havlen Moffat 1; As with Danks 1, HM-1 as a whole has yet to be subject to a
systematic, modern RV survey, and so we are forced to fall back on
observations at other wavelengths; WR89 (WN7h) appears to be a non-thermal
radio source and hence colliding wind binary (Cappa et al. [2004]), although
inspection of archival spectra fails to reveal RV shifts (Nazé et al.
[2013])181818The lack of spectral variability is potentially explicable by
unfavourable inclination and/or a long or highly eccentric orbit..
Nevertheless, these authors reveal it to have an extreme (variable) X-ray
luminosity and a spectrum comprising both soft and hard thermal components.
Unfortunately, only X-ray observations are available to determine the nature
of WR87 (WN7h); while it is comparatively faint with respect to WR79a, it too
has a hard spectral component at least consistent with emission from a wind
collision zone (Table 8 and discussion below).
* •
Sco OB1: Hosting the young massive cluster NGC6321 at its core, the Sco OB1
association contains two WNLh stars - WR 78 (= NGC 6231 305; WN7ha) and WR79a
(= NGC6231 327; WN9ha) - that, in contrast to many other examples considered,
demonstrate relatively modest bolometric luminosities Table 8). Both stars are
detected at X-ray energies, and share similar properties, being rather faint
and hard sources (Skinner et al. [2010], [2012]). WR78 appears to demonstrate
aperiodic optical variability (Skinner et al. [2012] and refs. therein) while
Chini et al. ([2012]) lists WR79a as an SB2 system but provides no links to an
appropriate reference; as such, we refrain from assuming binarity for either
on the basis of these works. However de Becker & Ruacq ([2013]) and Cappa et
al. ([2004]) report that both stars are non-thermal radio sources and hence
colliding-wind binaries: a conclusion we adopt for the remainder of this
study.
* •
W43: The giant H ii region hosts a young, heavily embedded cluster at its
core, which in turn contains a single WN7ha star: WR121a. Arora & Pandey
([2020]) provide an analysis of time-resolved X-ray observations of this star,
finding it to be a hard and highly luminous source, which demonstrates an
apparent $\sim 4.1$d periodicity that they attribute to orbital modulation in
a massive compact binary. Examining near-IR and radio data, Luque-Escamilla et
al. ([2011]) further show that WR121a maybe resolved into two components
separated by a projected angular distance of $\sim 0.6$", suggesting the
possible presence of a further stellar companion.
* •
Quintuplet: Finally we turn to the Quintuplet, which Clark et al. ([2018b])
report contains two WN8-9ha (qF256 and qF274) and an hybrid O7-8 Ia+/WN8-9ha
star (LHO001). Given the remaining cluster population their presence is a
surprise, with the authors suggesting they could be the result of binary
interaction. Unfortunately only LHO001 has been the subject of multi-epoch
spectroscopy, which reveal it to be highly variable, although insufficient
data exists to attempt to identify periodic modulation. Wang et al. ([2006])
present X-ray observations of the Quintuplet; none of these sources are
detected and, as with the Arches (Sect. 3.2.1), a combination of potential
blending and the presence of spatial variable diffuse emission prevent the
derivation of upper limits. Given the limited observational data for these
stars as well as their uncertain provenance, we refrain from discussing them
further.
#### C.1.2 Isolated WNLh stars
We may supplement the above dataset with five isolated single and binary
stars. Of the former, Skinner et al. ([2012]) report a detection of the WN8h
star WR16; unfortunately it is too faint for spectral analysis (Table 9, with
the flux scaled to the distance adopted by Hamann et al. [2019]). Although
optically variable it currently cannot be unambiguously identified as a binary
(Antokhin et al. [1995]). Similar conclusions are also reported for WR40
(WN8h) by these authors; unfortunately Gosset et al. ([2009]) are only able to
provide a (conservative) upper limit to the X-ray flux which, suitably scaled,
is repeated in Table 9. Turning to the binaries, Nazé et al. ([2020]) find
WR29 (WN7h+O; Table 3) to be a faint X-ray source; assuming a canonical
$kT\sim 0.6+6.9$keV thermal spectrum it appears moderately luminous ($\sim
2-3\times 10^{32}$ergs-1; Table 9). WR105 (WN9h) is a non-thermal radio source
(Montes et al. [2009]) and hence a strong colliding wind binary candidate.
Although a weak X-ray detection (Nazé et al. [2020]), under the same modelling
assumptions adopted for WR29 it appears to be intrinsically luminous ($\sim
4-12\times 10^{32}$ergs-1; Table 9).
Finally we turn to the runaway binary WR148 (WN8h), the nature of which has
been the subject of some debate e.g. Drissen et al. [1986]. Recently Munoz et
al. ([2017]) have suggested an O5V classification for the secondary; however
the canonical mass of such an object - they quote $\sim 37M_{\odot}$ \- is in
tension with the inclination derived by Drissen et al. ([1986]). Upon
reappraisal of the relevant photometric and polarimetric data the former
authors suggest that the results are ‘inconclusive’. Allowing for a lower
inclination, the canonical mass of the O5V secondary implies a current mass of
$\sim 33M_{\odot}$ for the WN7ha primary - unexpectedly low in comparison to
other examples (Table 3) and for a luminosity of $\sim
10^{6.3}L_{\odot}$191919Obtained by scaling the luminosity of Hamann et al.
([2019]) to the consensus distance of 8.3kpc (Zhekov [2012]), noting that Rate
& Crowther ([2020a]) suggest $\sim 9.47^{+1.77}_{-1.49}$kpc.. Nevertheless,
Zhekov ([2012]) report a rather moderate X-ray luminosity ($\sim 2.1\times
10^{32}$ergs-1) and a thermal spectrum ($kT\sim 0.4+1.0$keV) for the system.
For completeness we briefly discuss the Galactic centre source CXO
J174536.1-285638 (WN9ha). Although excluded from the sample for the reasons
given above, we note that it is a long period binary ($\sim 189$ d; Mikles et
al. [2008]). Adopting a two-component spectral fit ($kT\sim 0.7+4.6$keV)
implies an exceptional X-ray luminosity ($L_{\rm X}\sim 10^{35}$ergs-1) that
is directly comparable to that of Mk34 (Table 8). This indicates that the
latter is not a pathological system; however, the lower luminosity of the
primary ($L_{\rm bol}\sim 10^{6.24}$; Clark et al. [2009]) leads to an even
more extreme log$(L_{\rm X}/L_{\rm bol})\sim-4.8$ for CXO J174536.1-285638
than was found for Mk34 (Table 8).
### C.2 Data interpretation
#### C.2.1 The X-ray properties of WNLh stars
Historically, and by analogy to O stars, WR binaries have been expected to be
hard, overluminous X-ray sources. However recent studies suggest that such a
picture may be overly simplistic. In particular, while a number of confirmed O
star and WR binaries within Wd1 conform to this expectation, others are found
to have comparatively soft spectra consistent with emission from single stars
(Clark et al. [2019c]; see also Oskinova [2015], [2016]). Considering a
dataset exclusively comprising WR stars allowed Skinner et al. ([2010],
[2012]) to report the ubiquity of a hard ($kT\gtrsim 2$keV) component in the
X-ray spectra of all WN stars with spectra sufficient to accommodate
modelling. Conversely, Oskinova ([2016]) state that, on average, single WN
stars display a softer ($kT\sim 0.4$keV) spectrum. Clearly, reconciling these
potentially conflicting findings requires an accurate determination of the
binary properties of the stars studied. Nevertheless, studies focusing on
individual binary systems via multi-epoch, phase-resolved observations
indicate that where present - and as with O stars - the high-temperature
spectral component originates in the wind collision zone (e.g. Gosset et al.
[2009], Montes et al. [2013], Parkin & Gosset [2011]).
Mindful of the fact that the cohorts considered by Skinner et al. ([2012]) and
Oskinova ([2015], [2016]) comprise in part H-free WNE stars - and so are not
directly comparable to our sample - we highlight that the spectral properties
of the WNLh stars considered here are uniformly characterised by the presence
of a hard thermal component (Table 8), potentially indicative of binarity.
However, while the possibility remains that single WR stars may also
demonstrate similarly hard X-ray spectra (Skinner et al. [2012]) it appears
premature to utilise this observational feature to unambiguously identify WR
binaries.
This leaves X-ray luminosity as a prospective binary diagnostic. Oskinova
([2015]) and Gagné et al. ([2012]) provide compilations of X-ray properties
for single and binary WRs, finding that the luminosities of both show an
exceptional variance, spanning $\sim$four orders of magnitude. Nevertheless,
they identify potential trends whereby (i) binaries appear to be more
prevalent amongst the brighter X-ray sources, apparently obeying the canonical
$L_{\rm X}/L_{\rm bol}\sim 10^{-7}$ relationship delineated by hot, massive
stars and (ii) the most luminous systems are biased towards the binaries with
the longest periods. However, interpretation is complicated by the
heterogenous nature of the samples appraised by these authors, comprising both
WNLh and the more evolved H-free ‘classical’ WNE stars (with the latter being
of systematically lower luminosity).
Motivated by their evolutionary uniformity we plot $L_{\rm bol}$ versus
$L_{\rm X}$ for our sample of WNLh stars in Fig. C.1 (c.f. Tables C.1 and C.2;
note that due to uncertainty regarding individual contributions, the blended
R136 and NGC3603 sources are not included). For consistency, in all instances
we plot the integrated bolometric luminosity of the stellar system in question
(with that of F9 obtained via an identical methodology to that described in
Sect. 4; Najarro et al. in prep.). It is expected that for the majority of
objects considered this will be equivalent to the output from the WNLh star
that dominates the emergent spectrum. However at least three systems - WR20a,
MK34 and RMC144 - are known binaries comprising twin WNLh stars; in such cases
the luminosity of individual components is $\sim 0.3$ dex lower than
represented in Fig. C.1. We find that X-ray luminosities of the stars range
over four orders of magnitude: from $\sim 10^{31}$ergs-1 (for the apparently
single WR16) to $\gtrsim 10^{35}$ergs-1 (the binaries Mk34 and CXO
J174536.1-285638), the latter stars being amongst the most luminous colliding-
wind systems known (Gagne et al. [2012]). Despite considerable scatter, a
positive trend between X-ray and bolometric luminosities is apparent. With
23/39 of the WNLh stars considered exceeding the empirical $L_{\rm X}/L_{\rm
bol}\sim 10^{-7}$ relationship found for O stars, it appears that the former
have a systematically greater X-ray luminosity than the latter. Moreover there
are hints that the correlation is also steeper for WNLh stars, with 12/15
examples with log$(L_{\rm bol}/L_{\odot})>6.3$ exceeding this threshold (Fig.
C.1).
Of the remaining stars, the constant RV sources WR24 and R134 essentially
straddle this division202020Based on the results presented in Fig. 4 of Moffat
et al. ([2002]), WR43B also lies on this threshold which, based on the values
in Table 8 would imply that WR43A significantly exceeds it., while the upper
limits to the X-ray flux of five stars - the binary Mk30 and apparently single
star Mk51 as well as MK37Wb, RMC146, and RMC147 (which are currently of
indeterminate nature) - render their placement uncertain. Indeed, only nine
stars are unambiguously found below this threshold; these comprise the
apparently single stars WR16, WR40 and VFTS682; the binaries WR12, WR22,
WR79a, WR148 and RMC135; and WR87, the nature of which is currently uncertain.
Of these the binaries WR12 and WR22 are of particular interest given the
stringent upper limits obtained for their X-ray fluxes ($L_{\rm X}\lesssim
10^{31.8}$ergs-1; Fig. C.1). These derive from observations of both systems
obtained at orbital phases when their WNLha primaries potentially obscure
their companions and any putative wind collision zones. Following the
arguments of Gosset et al. ([2009]) and Nazé et al. ([2020]), one might expect
the dense, metal-rich winds of the WNLh stars to efficiently absorb high
energy emission from the latter components, leaving only their own
intrinsically weak emission visible. Indeed, the observations of both WR16 and
WR40 support the hypothesis that single WNLh stars are particularly faint
X-ray sources.
A corollary of this hypothesis would be that all of the X-ray luminous WNLh
stars are colliding wind binaries. Indeed 16 of the 21 confirmed WNLh
binaries212121WR20a, WR25, WR29, WR43C, WR78, WR89, WR105, F6, F7, Mk34, Mk35,
Mk37, RMC136, RMC140b, RMC144, and RMC145 are found above the $L_{\rm
X}/L_{\rm bol}\sim 10^{-7}$ threshold; of the remaining seven co-located stars
six currently lack diagnostic observations222222WR48-8, WR48-9, WR48-10,
WR76-3, WR76-7, and Mk39 and only Arches F9 shows no evidence for binary-
induced RV variability at this time (although we note that its radio
properties are indicative of binarity; Sect. 3.2). Thus we might infer a
binary fraction of 15/16 ($\sim 94$%) amongst stars exceeding this limit at
this time (and a minimum of $\sim 70$% if the remainder are found to be
single).
Consequently, we adopt $L_{\rm X}/L_{\rm bol}\sim 10^{-7}$ as a conservative
and empirical lower limit for the identification of colliding wind binaries
with WNLh primaries via their X-ray emission. In doing so we reiterate that
five known binaries fall below this threshold; however lowering the limit to
encompass all these system at all times would potentially permit the inclusion
of genuinely single stars. This is evident upon consideration of the variable
X-ray emission from the binaries WR12 and WR24, which appear to manifest as
single stars at certain orbital phases. In addition to their intrinsic
variability, the large variance in the X-ray properties of binaries likely
results from the diversity of their physical properties - in particular the
nature of the secondary, orbital separation and eccentricity (Gagné et al.
[2012] and Table 3) - which may allow some systems to remain inherently faint.
That the absolute X-ray luminosity of binaries with WNLh primaries appears
greater than that of those containing O stars likely reflects the difference
in wind properties, with the former supporting more powerful, optically thick
winds in comparison to the optically thin outflows of the latter (e.g.
Gräfener ([2021]).
Figure 11: Plot of bolometric and X-ray luminosities for the WNLh stars given
in Tables C.1 and C.2 (black), with the addition of the Arches detections F6,
F7, and F9 (red symbols). Confirmed binaries (including the RV variable Mk35
for simplicity) are given by filled squares, stars with no indication of
binarity are depicted by open squares, and the subset for which the requisite
observations have yet to be undertaken represented by open circles. Stars
known to be X-ray variable are indicated by the nested symbols. The second,
open symbol for W22 represents the upper limit to the intrinsic emission from
the WN7h primary according to Gosset et al. ([2009]). The solid lines joining
the two points for WR29 and WR105 indicate the range of possible luminosities
for both stars (Nazé et al. [2020]). Luminosities are shifted slightly for
RMC140b (-0.003 dex), RMC144 (+0.003 dex), Mk51 (+0.02 dex) and Mk37Wb (+0.02
dex) to aid clarity. The canonical $L_{\rm X}=10^{-7}L_{\rm bol}$ relationship
that Oskinova ([2015]) suggests is followed by O and WN stars is given by the
dotted line; stars falling below this limit are individually labeled, as are
certain additional stars of interest (subject to figure clarity). Finally the
low luminosity star WR48-8 is not plotted, but lies significantly above the
threshold.
#### C.2.2 The binary fraction of WNLh stars within clusters
How may we interpret these data in terms of the incidence of binarity and the
X-ray properties of such systems? Regarding the frequency of occurrence, as a
first step we may simply consider those stars that appear unambiguously binary
via spectroscopic determination of an orbital period. Mindful of the
limitations of extant RV and X-ray surveys, we find 9/17 ($\sim 53$%)
confirmed binaries within R136/30 Dor (including the RV variable Mk35).
Likewise, recognising that clusters such as Danks 1, HM-1 and Mercer 81 have
yet to be subject to systematic RV surveys (and the limited nature of the
Mercer 30 study) we arrive at a total of 7/11 in five galactic aggregates
(Table 9; Bochum 7, Westerlund 2, Carina, NGC3603 and Mercer 30). To this
tally we may add the periodic X-ray source WR121 (W43) and the non-thermal
radio sources WR78 (Sco OB1) and WR89 (HM-1) where emission is though to arise
in shocked material in the wind collision zone, increasing the Galactic count
to 10/16. This implies a minimum binary fraction of 19/33 ($\gtrsim 58$%)
before allowance is made for the insensitivity of the extant RV surveys to
longer period systems and those seen at unfavourable orbital inclinations.
This brings us to interpretation of the X-ray properties of these stars.
Orbitally modulated X-ray emission is apparent in the spectroscopically
confirmed binaries Mk34, RMC144, WR20a, WR21a, WR22, and WR25; on this basis
the periodically variable X-ray source WR121a is also a compelling binary
candidate (see also CXO J1745-28; Table 3). However, many of the remaining
candidate and confirmed WNLh binary systems have an insufficient timeline of
X-ray observations to permit a search for periodicities. For such systems we
adopt $L_{\rm X}/L_{\rm bol}\gtrsim 10^{-7}$ as a conservative threshold for
binarity; in doing so we are mindful of (i) the relative paucity of
observational constraints for apparently bona fide single stars, (ii) that
genuine binaries may fall below this threshold (cf. Fig. C.1 and Sect. 3.2.2)
and (iii) that we may not assess binarity for a number of X-ray detected WNLh
stars that currently lack bolometric luminosity determinations232323WR20aa,
WR20b, WR21a, WR42e and WR43-2..
Nevertheless, adopting this discriminant means we may assert binarity for
Arches F9 (see Sect. 3.3.2) as well as Mk39 (within 30 Dor), WR48-8, WR48-9,
WR48-10 (within Danks 1), and WR76-3 and WR76-7 (Mercer 81). Critically, this
would mean recognising the possibility of a binary fraction approaching unity
for the most luminous subset of stars (log$(L_{\rm bol}/L_{\odot})>6.3$; Sect.
C.2.1). Such a conclusion, however, is subject to the caveat that the nature
of some members of the very luminous cohorts within R136 (a1, a2, a3, and a5)
and NGC3603 (WR43A and WR43B) remain difficult to ascertain due to blending,
although the properties of the respective composite X-ray sources are strongly
suggestive of binarity amongst one or more of the component stars (Table 8).
More provocatively, if we were to take the observations summarised in Fig. C.1
at face value, one would infer that (apparently) single WNLha stars are
intrinsically X-ray faint ($L_{\rm X}<10^{32}$ergs-1). This would lead to the
conclusion that all but one of the stars reported as X-ray detections in Table
8 \- which, with the exception of WR12, uniformly exceed this value at one or
more epochs - are binaries: a verdict consistent with their uniformly hard
X-ray spectra.
Nevertheless, under the former, more conservative X-ray cut, the binary count
for R136/30 Dor rises to 10/21 ($\sim 48$%) with the inclusion of Mk39 (noting
that the sample size has risen since the full cohort of WNLh stars has been
subject to X-ray observations). Similarly, consideration of the additional
X-ray diagnostic leads to the identification of 15/24 ($\sim 63$%) binaries
within the ten galactic aggregates considered above. Given the foregoing
discussion and the diverse and incomplete nature of the multi-wavelength
observations employed in the preceding assessment, we consider it highly
likely that the simple number counts presented above are lower limits to the
true binary fractions, but the heterogenous nature of the extant datasets
means we are unable to assess the degree of incompleteness at this time. We
discuss the Arches cluster in the context of these observations - and the
resultant conclusions - further in Sect. 4 and 5.
|
# Emergence of two-fold spectral topology through non-Abelian gauge
engineering
Ronika Sarkar<EMAIL_ADDRESS>Solid State and Structural Chemistry
Unit, Indian Institute of Science, Bangalore 560012, India Department of
Physics, Indian Institute of Science, Bangalore 560012, India Ayan Banerjee
Solid State and Structural Chemistry Unit, Indian Institute of Science,
Bangalore 560012, India Awadhesh Narayan<EMAIL_ADDRESS>Solid State and
Structural Chemistry Unit, Indian Institute of Science, Bangalore 560012,
India
###### Abstract
Non-Abelian phenomena and non-Hermitian systems have both been widely explored
in recent years. As a bridge between the two, we introduce and develop non-
Abelian gauge engineering for realizing _multi-fold_ spectral topology. As an
example of our proposal, we engineer non-Hermiticity in the paradigmatic Su-
Schrieffer-Heeger (SSH) model by introducing a generalized non-Abelian gauge,
leading to an emergent two-fold spectral topology that governs the decoupled
behaviour of the corresponding non-Hermitian skin effect. As a consequence of
the non-Abelian gauge choice, our model exhibits a rich phase diagram
consisting of distinct topological phases, which we characterize by
introducing the notion of _paired_ winding numbers, which, in turn, predict
the direction of skin localization under open boundaries. We demonstrate that
the choice of gauge parameters enables control over the directionality of the
skin effect, allowing for it to be unilateral or bilateral. Furthermore, we
discover non-dispersive flat bands emerging within the inherent SSH model
framework, arising from the non-Abelian gauge. We also introduce a simplified
toy model to capture the underlying physics, thereby giving direct physical
insights. Our findings pave way for the exploration of unconventional spectral
topology through non-Abelian gauges.
Introduction.$-$ Non-Abelian phenomena are widespread in classical and quantum
systems – prominent examples range from the Yang-Mills theory of the standard
model Yang and Mills (1954) to the emergence of anyons obeying non-Abelian
statistics Nayak _et al._ (2008). In recent years, there has been remarkable
progress in the generation and control over non-Abelian features in a variety
of platforms Yang _et al._ (2023), leading to the unveiling of intriguing
physics Sugawa _et al._ (2018); Guo _et al._ (2021); Yang _et al._ (2019);
Brosco _et al._ (2021); Zhou (2023); Pang _et al._ (2023). The intricate
interplay between non-Abelian topology and quantum condensed matter has
yielded a rich tapestry of phenomena. This includes the discovery of non-
Abelian band topology in non-interacting metals Wu _et al._ (2019), the
realization of non-Abelian topological insulators Jiang _et al._ (2021), the
emergence of non-Abelian statistics in superconducting vortices Ivanov (2001);
Sato and Fujimoto (2009), and the exploration of non-Abelian braiding Noh _et
al._ (2020); Chen _et al._ (2022); Zhang _et al._ (2022a), to name just a
few.
In parallel, the synergy of non-Hermiticity and topology has ushered in a
remarkable wave of activity Bergholtz _et al._ (2021); Ashida _et al._
(2020); Ding _et al._ (2022); Banerjee _et al._ (2023a). This has led to a
variety of intriguing discoveries, including non-Hermitian spectral topology
Okuma and Sato (2023), non-Hermitian skin effect Yao and Wang (2018); Alvarez
_et al._ (2018); Zhang _et al._ (2022b); Lin _et al._ (2023), as well as
unexpected connections to topics in modern mathematics Hu and Zhao (2021);
Jaiswal _et al._ (2023); Wang _et al._ (2022); Banerjee _et al._ (2023b);
Hu (2023). The ramified symmetries in non-Hermitian systems give rise to
distinctive topological phases with the notion of “line” and “point” gap
topology Kawabata _et al._ (2019a). The role of non-Hermitian time dynamics
in spectral topology has revealed new insights into the interplay between non-
Hermiticity and topological features Lee _et al._ (2019), providing a deeper
understanding of phenomena such as spectral singularities, i.e., exceptional
points (EPs) Heiss (2004); Ding _et al._ (2016); Miri and Alu (2019); Chen
_et al._ (2020); Kawabata _et al._ (2019b), and the emergence of novel
topological phases with exotic properties Lee (2016); Shen _et al._ (2018);
Li _et al._ (2020); Liang _et al._ (2022); Kawabata _et al._ (2022); Longhi
(2019); Kunst _et al._ (2018); Weidemann _et al._ (2020); Yokomizo and
Murakami (2019); Song _et al._ (2019); Xue _et al._ (2022). Traversing
through EPs leads to phase transitions, and unveils unconventional features
such as intersecting Riemann sheet topology Dembowski _et al._ (2004);
Doppler _et al._ (2016), manifesting as Berry phases and fractional
topological charges Zhou _et al._ (2018), which have a range of potential
applications Wiersig (2014); Özdemir _et al._ (2019); Chen _et al._ (2017).
Figure 1: Non-Abelian gauge engineering in the Su-Schrieffer-Heeger model.
(a) Non-Hermiticity introduced by generalized gauge engineering in the SSH
model. (b) A simplified schematic of the four-band model for our choice of
gauge. Non-Abelian gauge fixing associates two degrees of freedom with each
unit cell site $A$ and $B$, effectively dissociating them to $A_{1},A_{2}$ and
$B_{1},B_{2}$, respectively. Each hopping label corresponds to the same
coloured arrow. Additionally, the rightward (leftward) red arrows denote
$-it_{2}\sin{\theta_{R}}$ ($it_{2}\cos{\theta_{L}}$) and the rightward
(leftward) green arrows denote $it_{2}\sin{\theta_{R}}$
($-it_{2}\cos{\theta_{L}}$). A similar approach is directly applicable to
other models.
In this work, we introduce and develop non-Abelian gauge engineering as a
promising approach to realize non-Hermitian multi-fold spectral topology.
Using the iconic Su-Schrieffer-Heeger (SSH) model as an example, we
demonstrate a remarkably rich phase diagram and the emergence of a two-fold
spectral topology, which are a direct consequence of the underlying non-
Abelian structure. We discover a variety of distinct topological phases, which
we characterize by introducing _paired_ winding numbers. These topological
invariants enable the prediction of the skin effect under open boundary
conditions, and the selection of gauge parameters allows exquisite control
over the direction and nature of the skin effect. We show that the spectral
topology, as well as the chiral dynamics, are robust to disorder. Notably, the
phase diagram features flat bands which are accessible in a wide parameter
regime. We further introduce a toy model comprising of coupled Hatano-Nelson
chains to understand the underlying principles of such a non-Abelian gauge
engineering.
Figure 2: Phase diagram and two-fold spectral topology. The phase diagram
shows five distinct topological phases as a function of the gauge parameters,
$\theta_{L}$ and $\theta_{R}$. The red and blue lines marking the phase
boundaries denote EP2 and EP4 transition lines, respectively. The spectral
topology associated with the $E_{1},E_{2}$ energy bands are shown in purple in
each phase, while $E_{3},E_{4}$ spectra are shown in green. The corresponding
arrows illustrate the direction of spectral winding. The associated coloured
circles denote the dynamically robust chiral modes associated with each
spectral loop. The paired topological invariant for each phase is presented in
the legend. The pink dashed lines illustrate parameter ranges which result in
two non-dispersive flat bands. Here we set $t_{1}=t_{2}=1$ and $\phi=0$.
Model.$-$ Our proposal is to introduce a non-Abelian gauge to a parent
Hermitian model, thus endowing it with additional degrees of freedom. A
suitable choice of the gauge parameters in turn allows generating non-
Hermiticity and the concomitant spectral topology [Fig. 1(a)]. Originally
introduced to study solitons in polyacetylene Su _et al._ (1979), the
paradigmatic SSH model provides a fundamental framework for studying
topological systems Asbóth _et al._ (2016). We consider the Hermitian
bipartite SSH model with different intra- and inter-unit cell hopping terms.
We engineer non-Hermiticity into the model by introducing a non-Abelian gauge
to the hopping potentials. A general method of engineering a non-Hermitian
gauge to the inter-unit cell hopping $t_{2}$ is schematically shown in Fig.
1(a). The Hamiltonian consists of the intra- and inter-unit cell couplings,
i.e., $H=H_{\textup{intra}}+H_{\textup{inter}}$. Each component can be
expressed as
$\displaystyle
H_{\textup{intra}}=\sum_{i}\>t_{1}\>e^{i\phi\>\sigma_{0}}\>c^{\dagger}_{A,i}c_{B,i}+t_{1}\>e^{-i\phi\>\sigma_{0}}\>c^{\dagger}_{B,i}c_{A,i},$
(1) $\displaystyle
H_{\textup{inter}}=\sum_{i}\>[\>t_{2}\>e^{i\theta_{R}\>\vec{\nu_{R}}.\vec{\sigma}}+\eta_{R}(\theta_{R})\>]\>c^{\dagger}_{B,i}c_{A,i+1}+$
$\displaystyle[\>t_{2}\>e^{i\theta_{L}\>\vec{\nu_{L}}.\vec{\sigma}}+\eta_{L}(\theta_{L})\>]\>c^{\dagger}_{A+1,i}c_{B,i}.$
Here, $L$ and $R$ denote leftward and rightward hopping terms, respectively,
and $\vec{\sigma}$ refers to the Pauli matrices (where $\sigma_{0}$ is the
identity matrix), which form the generators of the gauge transformations.
$\theta_{R}$ and $\theta_{L}$ are parameters of this invoked local
transformation. For a non-Abelian gauge transformation, the generators
($M^{i},M^{j},M^{k}$) must form a Lie group, i.e.,
$\commutator{M^{i}}{M^{j}}=C^{ijk}M^{k}$ and satisfy
$\Tr{M^{i}M^{j}}=C^{ij}\delta_{ij}$ Jacobson (1979). In our general model, the
gauge becomes non-Abelian when at least one of the following conditions is
fulfilled: (i)
$\commutator{\>\vec{\nu_{R}}.\vec{\sigma}}{\>\vec{\nu_{L}}.\vec{\sigma}\>}\neq
0$, (ii)
$\commutator{\>\vec{\nu_{R}}.\vec{\sigma}}{\>\eta_{L}(\theta_{L})\>}\neq 0$,
(iii) $\commutator{\>\eta_{R}(\theta_{R})}{\>\vec{\nu_{L}}.\vec{\sigma}\>}\neq
0$ or (iv) $\commutator{\>\eta_{R}(\theta_{R})}{\eta_{L}(\theta_{L})\>}\neq
0$. For simplicity, we choose $\vec{\nu_{L}}.\vec{\sigma}=-i\sigma_{y}$,
$\vec{\nu_{R}}.\vec{\sigma}=\sigma_{x}$,
$\eta_{L}(\theta_{L})=-\sqrt{2}t_{2}\cos({\frac{3\pi}{4}+\theta_{L}})\sigma_{0}$
and $\eta_{R}(\theta_{R})=0$. This satisfies
$\commutator{\vec{\nu_{L}}.\vec{\sigma}}{\vec{\nu_{R}}.\vec{\sigma}}\neq 0$,
following the aforementioned conditions, and conforms to an $SU(2)$ non-
Abelian gauge transformation Kasper _et al._ (2023). A schematic for our
proposed model is illustrated in Fig. 1(b). The $\sigma$ matrices associate a
synthetic spin to each lattice site modifying the SSH model to a
_quadripartite_ lattice. The Hamiltonian for the periodic system can be
reduced to the matrix form, $H(k)=\begin{pmatrix}0&M_{1}(k)\\\
M_{2}(k)&0\end{pmatrix}$, whose components are
$\begin{split}M_{1}(k)=\begin{bmatrix}t_{1}e^{i\phi}+t_{2}e^{-ik}\cos{\theta_{L}}&-it_{2}e^{-ik}\cos{\theta_{L}}\\\
it_{2}e^{-ik}\cos{\theta_{L}}&t_{1}e^{i\phi}+t_{2}e^{-ik}\cos{\theta_{L}}\\\
\end{bmatrix},\\\ \\\
M_{2}(k)=\begin{bmatrix}t_{1}e^{-i\phi}+t_{2}e^{ik}\cos{\theta_{R}}&-it_{2}e^{ik}\sin{\theta_{R}}\\\
it_{2}e^{ik}\sin{\theta_{R}}&t_{1}e^{-i\phi}+t_{2}e^{ik}\cos{\theta_{R}}\end{bmatrix}.\end{split}$
(2)
We compute the energy eigenvalues (detailed expressions for which are
presented in the supplementary information sup ), and derive the conditions
for the occurrence of EPs Heiss (2004). For simplicity, we set $\phi=0$. The
choice of our non-Abelian gauge results in $E_{1}$ and $E_{2}$ energy bands
being independent of $\theta_{L}$, which expedites our analytical
calculations. $E_{1}$ and $E_{2}$ gap closing points occur at
$\sin{\theta_{R}}-\cos{\theta_{R}}=\pm t_{1}/t_{2}$, whereas, $E_{3}$ and
$E_{4}$ gap closings occur at $\cos{\theta_{L}}=\pm t_{1}/2t_{2}$ and
$\sin{\theta_{R}}+\cos{\theta_{R}}=\pm t_{1}/t_{2}$. To obtain EPs, the
condition $\absolutevalue{t_{1}/t_{2}}\leq 2$ must be satisfied. We set
$t_{1}=t_{2}=1$ for subsequent calculations to illustrate our results. The
nature of the energy bands around different order EPs have been examined in
detail in the supplementary information sup .
Emergence of two-fold spectral topology and flat bands.$-$ The contours of EPs
reveal the phase transition lines, partitioning the phase space
$\theta_{L}-\theta_{R}$ into topologically distinct regions. In particular,
from the general conditions for obtaining EPs, we find that EP4 (fourth-order
EP) phase transitions occur for $\theta_{R}=0,\>\pm\pi/2,\>\pm\pi$ (shown by
dark blue vertical lines in Fig. 2); whereas, the EP2 (second-order EP)
transitions occur along $\theta_{L}=\pm\pi/3,\>\pm 2\pi/3,\>\pm 4\pi/3$ (shown
by red horizontal lines in Fig. 2). These exceptional contours divide the
parameter space into different phases, which can be classified uniquely
through the spectral topology and our newly defined topological invariant –
paired winding numbers.
A striking feature of our non-Hermitian gauge-engineered model is the
emergence of a two-fold spectral topology. Spectral topology refers to the
topological structure of the complex energy spectra, usually featuring
distinct point-gaps and line-gaps Kawabata _et al._ (2019a), unique for each
phase. The complex spectra are generally arc-like or loop-like, signifying the
occurrence of the non-Hermitian skin effect for the latter under open
boundaries Zhang _et al._ (2022c). In non-Hermitian models explored so far,
each phase has one distinctive spectral topology, which is contributed by all
the underlying energy bands of the system. In our model, each pair of energy
bands ($E_{1},E_{2}$) and ($E_{3},E_{4}$) individually contribute to a
different spectral topology which can be independently characterized. The
spectral topology is thus rendered two-fold, necessitating a two-fold
topological classification of the rich phase diagram obtained. The distinct
spectral topology uncovers five different topological phases (illustrated in
Fig. 2). The purple and green curves show the complex spectra contributed by
($E_{1},E_{2})$ and ($E_{3},E_{4})$, respectively. The filled circles on top
of each complex loop denote the dynamically stable chiral modes. These are the
energy modes which persist after a non-unitary long-time evolution of the
system with lifetimes $\sim\hbar/\textup{Im}(E_{\textrm{chiral}})$ Lee _et
al._ (2019). Under adiabatic deformations, owing to its topological origin,
each spectral loop survives. Consequently, the associated chiral mode also
survives, rendering them topologically robust attributes, using which we
define a topological invariant $(\nu,\gamma)$ for each pair of energy bands,
$\begin{split}\nu=\frac{1}{2}\sum_{\begin{subarray}{c}n=1,2\\\
\textup{Im}(E)>0\end{subarray}}\text{sgn}\>[\textup{Im}\>E_{n}(k_{n\alpha})]\>\text{sgn}\>[\partial_{k}\>[\textup{Re}\>E_{n}(k_{n\alpha})]],\\\
\gamma=\frac{1}{2}\sum_{\begin{subarray}{c}n=3,4\\\
\textup{Im}(E)>0\end{subarray}}\text{sgn}\>[\textup{Im}\>E_{n}(k_{n\alpha})]\>\text{sgn}\>[\partial_{k}\>[\textup{Re}\>E_{n}(k_{n\alpha})]].\end{split}$
(3)
Here ${k_{n\alpha}:\textup{Re}\>E_{n}(k_{n\alpha})=E_{b,\alpha}}$, where
$E_{b,\alpha}$ denotes the set of point gaps of each spectral loop, which
essentially locates the corresponding chiral mode. The first sign function
takes into account the sign of the imaginary energy of the chiral mode, while
the second sign function tracks the velocity of the chiral mode. The Nielsen-
Ninomiya theorem Bessho and Sato (2021) establishes that chiral modes always
come in pairs of opposite chirality. However, the effective chirality of long-
time dynamics is characterized by $\textup{Im}(E)>0$ modes Banerjee _et al._
(2022). Thereby, the paired winding numbers $(\nu,\gamma)$ defined above track
the chirality of the dynamically robust chiral modes of each spectral loop.
Each phase is uniquely classified by its characteristic paired winding number.
Additionally, for a topological transition between any two phases, phase 1:
$(\nu_{1},\gamma_{1})$ and phase 2: $(\nu_{2},\gamma_{2})$, we introduce
$\Lambda=\absolutevalue{\absolutevalue{\nu_{1}}-\absolutevalue{\nu_{2}}}+\absolutevalue{\absolutevalue{\gamma_{1}}-\absolutevalue{\gamma_{2}}}$.
If $\Lambda=1/2$, then the phase transition must cross an EP2 line, while
$\Lambda=1$ denotes a phase transition across an EP4 line. This is also true
for non-adjacent phases. For instance, the shortest path for transiting
between the blue $(1,0)$ and grey $(1/2,-1)$ phases requires crossing one EP2
as well as one EP4 line – here $\Lambda=3/2$. The distinct characteristic
features of each phase are summarized in Table 1.
Table 1: Characterization of different phases. The paired winding number, the number of dynamically stable chiral modes, the nature of line gap, and the skin effect characterize each phase uniquely. Here, all paired quantities refer to the spectral physics of $(E_{1,2},E_{3,4})$. Here $+(-)$ in the winding number refers to counter-clockwise (clockwise) spectral winding. The indicative color corresponds to the representation in the phase diagram in Fig. 2. Re/Im imply the presence of a real/imaginary line gap in the respective spectra, whereas ‘No’ refers to no line gap. The contents in the column ‘skin effect’ indicates the direction of the skin effect due to each band-pair. Paired winding no. | Indicative colour | Chiral modes | Line gap | Skin effect
---|---|---|---|---
$(1,-1)$ | Purple | (2,2) | (Re, Re) | (Left, Right)
$(\frac{1}{2},-\frac{1}{2})$ | Red | (1,1) | (No, No) | (Left, Right)
$(1,0)$ | Blue | (2,0) | (Re, Im) | (Left, Right)
$(1,\frac{1}{2})$ | Green | (2,1) | (Re, No) | (Left, Left)
$(\frac{1}{2},-1)$ | Grey | (1,2) | (No, Re) | (Left, Right)
Another remarkable consequence of our non-Abelian gauge engineering is that by
solely tuning the gauge parameter, $\theta_{R}$, we can generate two
dynamically robust flat bands at
$\theta_{R}=\pi/4,\>3\pi/4\>(\>\forall\>\theta_{L})$. $E_{1}$ and $E_{2}$
spectral loops shrink to points on the real energy axis, resulting in two flat
bands at $\textup{Re}(E)=\pm t_{1}$ and $\textup{Im}(E)=0$. These flat bands
have infinite lifetimes and are robust to non-unitary time dynamics, thereby
rendering them accessible to experiments. Later, we further elucidate the
origin of these flat bands by mapping our model to a toy model of coupled
Hatano-Nelson chains. Moreover, as a consequence of the point-like spectra
exhibited by $E_{1},E_{2}$ under the flat band condition, the complex spectral
area of the two left-chiral bands is lost. Consequently, the skin effect
vanishes at the left edge of the lattice model. Thus, exclusively under the
flat band condition, the occurrence of the skin effect is limited solely to
the right edge of the lattice. We present further details in the supplementary
information sup .
Figure 3: Spectral winding, generalized Brillouin zone, and tunable skin
effect. (a) The spectral topology in the ‘grey’ phase showing that the
spectral winding of $E_{1},E_{2}$ (in purple) is counter-clockwise while that
of $E_{3},E_{4}$ (in green) is clockwise. This results in (b) a non-Hermitian
skin effect at both edges depicted by the local density of states (LDOS). The
inset shows the GBZ (in corresponding colours) with $|\beta(E_{1},E_{2})|<1$
and $|\beta(E_{3},E_{4})|>1$. The black dashed lines mark the unit circle. In
the green phase, (c) the direction of spectral winding is counter-clockwise
for both pairs of bands, resulting in (d) skin effect only at the left edge.
Here both $|\beta(E_{1},E_{2})|,\>|\beta(E_{3},E_{4})|<1$. (e)-(f) The real
and imaginary spectra in the ‘green’ phase as a function of $k_{x}$. A base
energy $E_{0}$ intersects $\textup{Re}(E_{1})$ and $\textup{Re}(E_{3})$ at
$k_{\alpha}$ and $k_{\beta}$, respectively. The corresponding $\textup{Im}(E)$
are marked in filled (unfilled) circles for long-time surviving (evanescent)
modes. For $E_{1}$ ($E_{3}$) the mode at $k_{\alpha}$ ($k_{\beta}$) survives
owing to higher $\textup{Im}(E)$. At $k_{\alpha}$, $E_{1}$ has a left-mover
while at $k_{\beta}$, $E_{3}$ has a left-mover. Hence, skin effect in the
green phase occurs at the left edge. (g)-(h) Behaviour of the spectra and skin
effect under disorder ($\Delta=0.1$) in the grey phase, illustrating its
robustness. Here $t_{1}=t_{2}=1,\phi=0$. For grey phase:
$\theta_{R}=3\pi/4,\theta_{L}=2\pi/5$. For green phase:
$\theta_{R}=\pi/8,\theta_{L}=2\pi/5$.
Correspondence to skin effect and chiral dynamics.$-$ We now illustrate how
our paired winding numbers ($\nu,\gamma$) can predict the skin effect under
open boundary conditions. Notably, our two-fold spectral topology also renders
the resultant non-Hermitian skin effect two-fold. We find that in all phases
(except the ‘green’ phase), $\>\nu$ and $\gamma$ have opposite signs,
signifying that the winding direction and, thereby, the chirality of the
dynamically stable modes are opposite. In contrast, in the green phase with a
paired winding number (1,1/2), the chirality of the two pairs of energy bands
is in the same direction. This peculiarity is reflected in the skin effect. To
exemplify this, we contrast the bi-chiral ‘grey’ phase and the uni-chiral
‘green’ phase in Fig. 3(a)-(d). The purple curves denote the spectral topology
of energy bands ($E_{1},E_{2}$), while the green curves show ($E_{3},E_{4}$).
The opposite signs of the winding number in the grey phase [Fig. 3(a)-(b)]
manifests as a ‘bilateral’ skin effect. The counter-clockwise propagating
energy bands ($E_{1},E_{2}$) lead to skin effect at the left edge, whereas the
clockwise propagating bands ($E_{3},E_{4}$) induce skin effect at the right
edge. The generalized Brillouin zone (GBZ) denoted by $\beta$ [inset of Fig.
3(b)] shows that $\absolutevalue{\beta(E_{1},E_{2})}<1$ (in purple) whereas,
$\absolutevalue{\beta(E_{3},E_{4})}>1$ (in green) – this establishes the
directionally decoupled skin effect. In contrast, in the uni-chiral green
phase [see Fig. 3(c)-(d)], the winding direction of both pairs of energy bands
is counter-clockwise, resulting in a ‘unilateral’ skin effect at the left
edge. The GBZ shows that both $\absolutevalue{\beta(E_{1},E_{2})}<1$ and
$\absolutevalue{\beta(E_{3},E_{4})}<1$. Detailed calculations of the GBZ are
presented in the supplementary information sup .
We highlight the physical origin of the directionally-dependent skin effect in
our model in Fig. 3(e)-(f) for the green phase, which exhibits a unilateral
skin effect. The different colours correspond to the different energy bands
specified in the legend. We consider a real excitation energy $E_{0}$, which
intersects $\textup{Re}(E_{1})$ and $\textup{Re}(E_{3})$ at $k_{\alpha}$ and
$k_{\beta}$, respectively. At $k_{\alpha}$, $E_{1}$ has a left-moving mode
while $E_{3}$ has a right-mover. However, at $k_{\beta}$, $E_{1}$ has a right-
mover while $E_{2}$ has a left-mover. The net flow of eigenstates would
counterbalance in the Hermitian regime. However, the inherent non-Hermiticity
– a consequence of non-Abelian gauge choice – associates a finite lifetime
with every eigenmode, leading to a dynamically persistent current. Owing to
the non-unitary time evolution $\sim e^{-iEt}\rightarrow
e^{-i\mathrm{Re}(E)t}\>e^{\mathrm{Im}(E)t}$, the eigenmodes with negative
imaginary energy die out over a timescale $\sim\hbar/\textup{Im}(E)$. For
energy band $E_{1}$, the left-mover at $k_{\alpha}$ survives, and for $E_{3}$,
the left-mover at $k_{\beta}$ survives. The persistent modes are denoted by
filled red circles, and the corresponding evanescent modes are shown using
unfilled red circles. Thus, the cumulative flow of states in the green phase
is leftwards, leading to the unilateral skin effect at the left edge.
Our model facilitates control over the direction of the skin effect by
suitably tuning parameters $\theta_{L}$ and $\theta_{R}$. Due to the
topological origin of the skin effect and the point-gap spectra, we expect
them to be robust to moderate amounts of disorder, which would make the
observation of the bilateral skin effect experimentally feasible. To
illustrate this robustness, we introduce random on-site disorder in our model
as,
$H_{\textup{dis}}=\sum_{i}\Delta_{i}(c^{\dagger}_{A,i}c_{A,i}+c^{\dagger}_{B,i}c_{B,i})$,
where $\Delta_{i}=\Delta[-1,1]$ is chosen randomly and the disorder strength
is $\Delta$. Fig. 3(g)-(h) show the robust nature of the spectral topology and
bilateral skin effect in the exemplary grey phase at disorder strength
$\Delta=0.1\>t_{1}$, confirming our expectations.
Figure 4: Coupled Hatano-Nelson chains as toy model. (a) Two Hatano-Nelson
chains linked using coupling strength $t_{o}$. The hopping parameters are
denoted below the schematic, mapping it to our non-Abelian SSH model. (b)
Dynamically robust ($\textup{Im}(E)=0$) flat bands at $\textup{Re}(E)=\pm
t_{o}$, in the strong coupling limit ($t_{o}/t_{1}=0.01$). (c) Bilateral skin
effect and (d) unilateral skin effect attained by tuning parameters
$\theta_{L}$ and $\theta_{R}$ in the weak coupling limit ($t_{1}/t_{o}=0.1$).
Here $t_{2}=1,\phi=0$ for all plots. We set (b) $\theta_{R,L}=\pi/4,3\pi/4$ ,
(c) $\theta_{R,L}=\pi/2,\pi/2$ , and (d) $\theta_{R,L}=-\pi/5,3\pi/4$.
Mapping to a toy model: Coupled Hatano-Nelson chains.$-$ To elucidate the
controllable nature of our proposed direction-selective skin effect, we devise
a simple toy model. We consider coupled chains of the Hatano-Nelson model,
which is the simplest model with rich non-Hermitian features Hatano and Nelson
(1996). As illustrated in Fig. 4, we couple two Hatano-Nelson chains with non-
reciprocal hopping, with coupling strength $t_{o}$. The hopping parameters are
set as represented in the schematic, in order to map it to our original non-
Abelian SSH model. In the weak coupling limit, i.e., $t_{o}/t_{1}<<1$, we
discover that by tuning parameters $\theta_{L}$ and $\theta_{R}$, we can
control the direction of the skin effect. Remarkably, we find that it is
possible to obtain the bilateral skin effect [Fig. 4(c)], as well as the
unilateral skin effect [Fig. 4(d)], even for fixed values of $t_{1}$ and
$t_{2}$. Furthermore, in the strong coupling limit, i.e., $t_{o}/t_{1}>>1$, we
can tune the free parameters such that $\theta_{R}=\pi/4$ and
$\theta_{L}=3\pi/4$, leading to the emergence of two non-dispersive flat bands
at $\textup{Re}(E)=\pm t_{o}$. Similar to our original model, these flat bands
are also dynamically robust with $\textup{Im}(E)=0$ [shown in Fig. 4(b)]. In
the strong coupling limit, there is a dissociation process, resulting in unit
cells that consist of one atom each from the upper and lower chains. In this
regime, the wavefunction within each unit cell becomes localized – this leads
to the emergence of a flat band, with the wavefunction pinned to its
respective unit cell.
Conclusions and outlook.$-$ In summary, we proposed non-Abelian gauge
engineering to realize multi-fold spectral topology and demonstrated that by
utilizing a non-Abelian $SU(2)$ gauge, a two-band model can exhibit a two-fold
spectral topology which in turn leads to a multi-chiral non-Hermitian skin
effect. We introduced a paired winding number which characterizes the distinct
spectral topology of each phase and provides insights into the nature of phase
transitions through EPs, while also capturing the chiral dynamics which
determines the direction of skin effect. We note that applying a generalized
$SU(N)$ gauge can ramify the spectral topology into an $N$-fold structure,
which may have intriguing consequences for skin effects in higher dimensions.
This multi-fold spectral topology offers the possibility of directionally-
controllable skin effects and mode selection. Importantly, our proposed models
and their associated effects are experimentally accessible, providing a
promising avenue for future studies and potential applications. As an
illustration of the experimental feasibility, we have developed a topoelectric
circuit that can faithfully emulate the underlying non-Abelian SSH
Hamiltonian. This circuit, presented in the supplementary information sup ,
provides a promising platform for realizing our predicted effects. We hope
that our proposal motivates further synergy between non-Abelian and non-
Hermitian physics.
Acknowledgments.$-$ R.S. and A.B. are supported by the Prime Minister’s
Research Fellowship (PMRF). A.N. acknowledges support from the Indian
Institute of Science.
## References
* Yang and Mills (1954) C.-N. Yang and R. L. Mills, Physical review 96, 191 (1954).
* Nayak _et al._ (2008) C. Nayak, S. H. Simon, A. Stern, M. Freedman, and S. D. Sarma, Reviews of Modern Physics 80, 1083 (2008).
* Yang _et al._ (2023) Y. Yang, B. Yang, G. Ma, J. Li, S. Zhang, and C. Chan, arXiv preprint arXiv:2305.12206 (2023).
* Sugawa _et al._ (2018) S. Sugawa, F. Salces-Carcoba, A. R. Perry, Y. Yue, and I. Spielman, Science 360, 1429 (2018).
* Guo _et al._ (2021) Q. Guo, T. Jiang, R.-Y. Zhang, L. Zhang, Z.-Q. Zhang, B. Yang, S. Zhang, and C. T. Chan, Nature 594, 195 (2021).
* Yang _et al._ (2019) Y. Yang, C. Peng, D. Zhu, H. Buljan, J. D. Joannopoulos, B. Zhen, and M. Soljačić, Science 365, 1021 (2019).
* Brosco _et al._ (2021) V. Brosco, L. Pilozzi, R. Fazio, and C. Conti, Physical Review A 103, 063518 (2021).
* Zhou (2023) L. Zhou, arXiv preprint arXiv:2302.05710 (2023).
* Pang _et al._ (2023) Z. Pang, J. Hu, and Y. Yang, arXiv preprint arXiv:2304.01876 (2023).
* Wu _et al._ (2019) Q. Wu, A. A. Soluyanov, and T. Bzdušek, Science 365, 1273 (2019).
* Jiang _et al._ (2021) T. Jiang, Q. Guo, R.-Y. Zhang, Z.-Q. Zhang, B. Yang, and C. T. Chan, Nature Communications 12, 6471 (2021).
* Ivanov (2001) D. A. Ivanov, Physical review letters 86, 268 (2001).
* Sato and Fujimoto (2009) M. Sato and S. Fujimoto, Physical Review B 79, 094504 (2009).
* Noh _et al._ (2020) J. Noh, T. Schuster, T. Iadecola, S. Huang, M. Wang, K. P. Chen, C. Chamon, and M. C. Rechtsman, Nature Physics 16, 989 (2020).
* Chen _et al._ (2022) Z.-G. Chen, R.-Y. Zhang, C. T. Chan, and G. Ma, Nature Physics 18, 179 (2022).
* Zhang _et al._ (2022a) X.-L. Zhang, F. Yu, Z.-G. Chen, Z.-N. Tian, Q.-D. Chen, H.-B. Sun, and G. Ma, Nature Photonics 16, 390 (2022a).
* Bergholtz _et al._ (2021) E. J. Bergholtz, J. C. Budich, and F. K. Kunst, Reviews of Modern Physics 93, 015005 (2021).
* Ashida _et al._ (2020) Y. Ashida, Z. Gong, and M. Ueda, Advances in Physics 69, 249 (2020).
* Ding _et al._ (2022) K. Ding, C. Fang, and G. Ma, Nature Reviews Physics 4, 745 (2022).
* Banerjee _et al._ (2023a) A. Banerjee, R. Sarkar, S. Dey, and A. Narayan, Journal of Physics: Condensed Matter (2023a).
* Okuma and Sato (2023) N. Okuma and M. Sato, Annual Review of Condensed Matter Physics 14, 83 (2023).
* Yao and Wang (2018) S. Yao and Z. Wang, Physical review letters 121, 086803 (2018).
* Alvarez _et al._ (2018) V. M. Alvarez, J. B. Vargas, and L. F. Torres, Physical Review B 97, 121401 (2018).
* Zhang _et al._ (2022b) X. Zhang, T. Zhang, M.-H. Lu, and Y.-F. Chen, Advances in Physics: X 7, 2109431 (2022b).
* Lin _et al._ (2023) R. Lin, T. Tai, L. Li, and C. H. Lee, Frontiers of Physics 18, 53605 (2023).
* Hu and Zhao (2021) H. Hu and E. Zhao, Physical Review Letters 126, 010401 (2021).
* Jaiswal _et al._ (2023) R. Jaiswal, A. Banerjee, and A. Narayan, New Journal of Physics 25, 033014 (2023).
* Wang _et al._ (2022) H.-Y. Wang, F. Song, and Z. Wang, arXiv preprint arXiv:2212.11743 (2022).
* Banerjee _et al._ (2023b) A. Banerjee, R. Jaiswal, M. Manjunath, and A. Narayan, Proceedings of the National Academy of Sciences 120, e2302572120 (2023b).
* Hu (2023) H. Hu, arXiv preprint arXiv:2306.12022 (2023).
* Kawabata _et al._ (2019a) K. Kawabata, K. Shiozaki, M. Ueda, and M. Sato, Physical Review X 9, 041015 (2019a).
* Lee _et al._ (2019) J. Y. Lee, J. Ahn, H. Zhou, and A. Vishwanath, Physical review letters 123, 206404 (2019).
* Heiss (2004) W. Heiss, Journal of Physics A: Mathematical and General 37, 2455 (2004).
* Ding _et al._ (2016) K. Ding, G. Ma, M. Xiao, Z. Zhang, and C. T. Chan, Physical Review X 6, 021007 (2016).
* Miri and Alu (2019) M.-A. Miri and A. Alu, Science 363, eaar7709 (2019).
* Chen _et al._ (2020) H.-Z. Chen, T. Liu, H.-Y. Luan, R.-J. Liu, X.-Y. Wang, X.-F. Zhu, Y.-B. Li, Z.-M. Gu, S.-J. Liang, H. Gao, _et al._ , Nature Physics 16, 571 (2020).
* Kawabata _et al._ (2019b) K. Kawabata, T. Bessho, and M. Sato, Physical review letters 123, 066405 (2019b).
* Lee (2016) T. E. Lee, Physical review letters 116, 133903 (2016).
* Shen _et al._ (2018) H. Shen, B. Zhen, and L. Fu, Physical review letters 120, 146402 (2018).
* Li _et al._ (2020) L. Li, C. H. Lee, S. Mu, and J. Gong, Nature communications 11, 5491 (2020).
* Liang _et al._ (2022) Q. Liang, D. Xie, Z. Dong, H. Li, H. Li, B. Gadway, W. Yi, and B. Yan, Physical review letters 129, 070401 (2022).
* Kawabata _et al._ (2022) K. Kawabata, K. Shiozaki, and S. Ryu, Physical Review B 105, 165137 (2022).
* Longhi (2019) S. Longhi, Physical Review Research 1, 023013 (2019).
* Kunst _et al._ (2018) F. K. Kunst, E. Edvardsson, J. C. Budich, and E. J. Bergholtz, Physical review letters 121, 026808 (2018).
* Weidemann _et al._ (2020) S. Weidemann, M. Kremer, T. Helbig, T. Hofmann, A. Stegmaier, M. Greiter, R. Thomale, and A. Szameit, Science 368, 311 (2020).
* Yokomizo and Murakami (2019) K. Yokomizo and S. Murakami, Physical review letters 123, 066404 (2019).
* Song _et al._ (2019) F. Song, S. Yao, and Z. Wang, Physical review letters 123, 170401 (2019).
* Xue _et al._ (2022) W.-T. Xue, Y.-M. Hu, F. Song, and Z. Wang, Physical Review Letters 128, 120401 (2022).
* Dembowski _et al._ (2004) C. Dembowski, B. Dietz, H.-D. Gräf, H. Harney, A. Heine, W. Heiss, and A. Richter, Physical Review E 69, 056216 (2004).
* Doppler _et al._ (2016) J. Doppler, A. A. Mailybaev, J. Böhm, U. Kuhl, A. Girschik, F. Libisch, T. J. Milburn, P. Rabl, N. Moiseyev, and S. Rotter, Nature 537, 76 (2016).
* Zhou _et al._ (2018) H. Zhou, C. Peng, Y. Yoon, C. W. Hsu, K. A. Nelson, L. Fu, J. D. Joannopoulos, M. Soljačić, and B. Zhen, Science 359, 1009 (2018).
* Wiersig (2014) J. Wiersig, Physical review letters 112, 203901 (2014).
* Özdemir _et al._ (2019) Ş. K. Özdemir, S. Rotter, F. Nori, and L. Yang, Nature materials 18, 783 (2019).
* Chen _et al._ (2017) W. Chen, Ş. Kaya Özdemir, G. Zhao, J. Wiersig, and L. Yang, Nature 548, 192 (2017).
* Su _et al._ (1979) W.-P. Su, J. R. Schrieffer, and A. J. Heeger, Physical review letters 42, 1698 (1979).
* Asbóth _et al._ (2016) J. K. Asbóth, L. Oroszlány, A. Pályi, J. K. Asbóth, L. Oroszlány, and A. Pályi, A Short Course on Topological Insulators: Band Structure and Edge States in One and Two Dimensions , 1 (2016).
* Jacobson (1979) N. Jacobson, _Lie algebras_ , 10 (Courier Corporation, 1979).
* Kasper _et al._ (2023) V. Kasper, T. V. Zache, F. Jendrzejewski, M. Lewenstein, and E. Zohar, Physical Review D 107, 014506 (2023).
* (59) “See supplementary information for energy eigenvalue expressions, discussion of band dispersions, generalized brillouin zone calculations, detailed characterization of other phases, and details of the topoelectric circuit realization.” .
* Zhang _et al._ (2022c) K. Zhang, Z. Yang, and C. Fang, Nature communications 13, 2496 (2022c).
* Bessho and Sato (2021) T. Bessho and M. Sato, Physical Review Letters 127, 196404 (2021).
* Banerjee _et al._ (2022) A. Banerjee, S. S. Hegde, A. Agarwala, and A. Narayan, Physical Review B 105, 205403 (2022).
* Hatano and Nelson (1996) N. Hatano and D. R. Nelson, Physical review letters 77, 570 (1996).
|
# Optical fields spanning the 4D space of nonparaxial polarization
David Marco 1, 2 and Miguel A. Alonso 1, 3, 4, 5
1 Aix Marseille Univ., CNRS, Centrale Marseille, Institut Fresnel, UMR 7249,
13397 Marseille Cedex 20, France
2 Instituto de Bioingeniería, Universidad Miguel Hernández de Elche, 03202
Elche, Spain
3 The Institute of Optics, University of Rochester, Rochester, NY14627, USA
4 Center for Coherence and Quantum Optics, University of Rochester, Rochester,
NY14627, USA
5 Laboratory for Laser Energetics, University of Rochester, Rochester,
NY14627, USA
E-mails<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
We present simple mathematical solutions for optical fields that span all
nonparaxial polarization ellipses. These fields, given by a superposition of
five plane waves with non-coplanar polarizations, are experimentally
realizable. By slowly varying in time the relative amplitude between two
groups of plane waves, the resulting fields take the form of a periodic
polarization lattice, where all possible nonparaxial polarization ellipses are
localized in a 4D spatio-temporal cell. The field distributions in these cells
can be regarded as 4D skyrmionic textures.
## 1 Introduction
For a monochromatic optical field, the electric field vector at any given
point traces in time an ellipse [1]. The parameters that describe the shape
and orientation of this ellipse (other than its size, which is related to the
field’s intensity) are what we call polarization. For a monochromatic paraxial
optical field, the ellipse is essentially constrained to the plane
perpendicular to the main direction of propagation, so only two parameters are
required to describe polarization: the ellipticity and the orientation angle
within the plane. These two quantities are well represented by the two
coordinates over the surface of a unit sphere known as the Poincaré sphere.
Polarization can change from point to point across (or along) a beam. Since
the Poincaré sphere is a 2D abstract space, it is possible for a beam to
contain many or even all polarization states at a given physical plane
transverse to its main direction of propagation. Simple solutions such as the
so-called full Poincaré beams [2] contain all polarization states at each
transverse plane. For these beams, the Poincaré sphere is mapped onto the
plane according to a stereographic projection. Since the sphere is covered
entirely without any singularities of the mapping (except at infinity), this
field can be regarded as a 2D skyrmionic structure. Skyrmions correspond to
topological structures in which a closed abstract parameter space (the
Poincaré sphere in this case) is mapped completely and monotonically onto the
flat physical space. They have been studied in several areas of physics,
especially in magnetic systems [3].
For monochromatic nonparaxial fields, the plane containing the polarization
ellipse at any given point is not fixed by the main direction of propagation.
Four parameters are then required to fully describe polarization, such as the
ellipticity and the three Euler angles determining the orientation of the
ellipse. These parameters define then a 4D space of nonparaxial polarization,
for which several geometric representations have been proposed in terms of two
points over the surface of a unit sphere [4, 5, 6]. For a nonparaxial field,
polarization typically changes from point to point. However, the 3D physical
space is not sufficient to fully explore the 4D polarization space, unless it
is supplemented with a four dimension, such as time.
Here we present what we refer to for brevity as full 3D polarization (F3DP)
fields, which are nonparaxial optical fields containing all possible
polarization ellipses. We describe a family of experimentally realizable
mathematical solutions of such fields in the form of optical lattices whose
polarization state changes not only in 3D space but also adiabatically in
time, so that all possible nonparaxial polarization ellipses exist inside a
cuboid in 3D space within a given temporal interval. The fields are
constructed as superpositions of five noncoplanar plane waves. The adiabatic
temporal variation of these quasimonochromatic fields is implemented by slowly
varying the relative amplitude between two sets of plane waves. Some of the
solutions present uniform intensity in space and time, and they all present a
rich polarization structure. These field can be experimentally implemented in
the focal region of a high numerical aperture lens.
Moreover, we identified 4D skyrmionic structures in these F3DP fields. Most
skyrmionic structures studied so far in vector waves are 2D skyrmions (also
known as baby skyrmions), which fully span at a plane or a section of a plane
a spherical space corresponding to an aspect of the polarization ellipse, such
as: the normalized Stokes vector for paraxial beams [2, 7, 8]; the direction
of the field vector at a fixed time for the electromagnetic evanescent waves
of surface plasmon polaritons [9], evanescent acoustic waves [10] and
traveling acoustic waves [11]; the direction of the spin vector (normal to the
ellipse) in evanescent optical fields [12] and at the focal plane of tightly
focused optical fields [13]; and the direction of the ellipse major axis for
tightly focused optical fields [13]. Recently, a 3D skyrmionic Hopfion was
found in the propagating volume of a paraxial field [14], which spans a
3-sphere where every 2D polarization state with any global phase is
represented. To study the 4D skyrmionic structures within F3DP fields, we
define a Skyrme density as the Jacobian of the mapping between the coordinates
over the unit sphere of two points representing nonparaxial polarization
states and the flat space-time. The integral of this density over a skyrmionic
spatiotemporal cell is shown to give a Skyrme number of unit size.
## 2 Constructing the optical field
In this section, we propose a temporally-variant quasi-monochromatic field
that contains all possible states of full polarization within a spatio-
temporal cell with spatial dimensions of the order of a wavelength. This field
consists of a simple combination of five plane waves. By changing the angles
between these plane waves the field can be made more directional, and the cell
containing all polarization states can be made larger.
### 2.1 The construction and laboratory reference frames
At any given point the complex electric vector $\mathbf{E}$ of a nonparaxial
(quasi)monochromatic field can be expressed as
$\mathbf{E}=A\exp(i\Phi)\left[\sin{\gamma_{1}}\left(\mathbf{x}\,\cos{\gamma_{2}}e^{-i\delta_{1}}+\mathbf{z}\,\sin{\gamma_{2}}e^{-i\delta_{2}}\right)+\mathbf{y}\,\cos{\gamma_{1}}\right],$
(1)
where $\mathbf{x}$, $\mathbf{y}$ and $\mathbf{z}$ are unit vectors in the $x$,
$y$ and $z$ directions, respectively, and the six real parameters
$A,\Phi,\gamma_{1},\gamma_{2},\delta_{1},\delta_{2}$ are functions of
position, and even time as long as their variation is negligible over the
scale of an optical cycle. The real electric field is obtained by taking the
real part of ${\bf E}\exp{(-i\omega t)}$, where $\omega$ is the carrier
temporal angular frequency of oscillation. Note that only four parameters are
required to describe a 3D polarization state, since the global amplitude $A$
and the global phase $\Phi$ do not affect the shape and orientation of the
ellipse traced by the electric field over a temporal period. In order for the
field to span all polarization states, these four parameters must span a range
of values, for example: $0\leq\gamma_{1},\gamma_{2}\leq\pi/2$, which account
for all relative amplitudes between the three field components, and
$0\leq\delta_{1},\delta_{2}<2\pi$, which account for all relative phases.
Therefore, in order to design a field that covers all states of polarization,
we must map the complete abstract four-parameter space
$(\gamma_{1},\gamma_{2},\delta_{1},\delta_{2})$ onto the physical space-time
$(x,y,z,t)$, with the constraint that the field must be a solution of
Maxwell’s equations in free space. Under the approximation that the temporal
dependence of the polarization parameters is much slower than the scale set by
$\omega$, these conditions reduce to requiring the form in Eq. (1) to satisfy
the Helmholtz equation and to have zero divergence.
It turns out that a simple solution can be found by making two assumptions: i)
the global amplitude $A$ is constant and hence can be set to unity, and ii)
each of the four polarization parameters depends linearly on one of the
spatio-temporal coordinates. The following combination, for example, achieves
the desired goals: $\gamma_{2}\rightarrow ky/\sqrt{2}$, $\delta_{1}\rightarrow
kx/\sqrt{2}$, $\delta_{2}\rightarrow kz/\sqrt{2}$, where
$k=\omega/c=2\pi/\lambda$ is the wavenumber with $\lambda$ being the
wavelength, and $\gamma_{1}\rightarrow\Omega t$, where $\Omega\ll\omega$. By
also letting $\Phi=k(x+z)/\sqrt{2}$ we arrive at the normalized field
$\displaystyle\mathbf{E}$ $\displaystyle=\left[\sin{\Omega
t}\left(\mathbf{x}\,e^{-i\frac{kx}{\sqrt{2}}}\cos{\frac{ky}{\sqrt{2}}}+\mathbf{z}\,e^{-i\frac{kz}{\sqrt{2}}}\sin{\frac{ky}{\sqrt{2}}}\right)+\mathbf{y}\,\cos{\Omega
t}\right]e^{ik\frac{x+z}{\sqrt{2}}}$
$\displaystyle=\mathbf{x}\,\frac{\sin{\Omega
t}}{2}\left(e^{i\mathbf{k}_{\mathbf{x}\mathrm{A}}\cdot{\bf
r}}+e^{i\mathbf{k}_{\mathbf{x}\mathrm{B}}\cdot{\bf
r}}\right)+\mathbf{z}\,\frac{\sin{\Omega
t}}{2i}\left(e^{i\mathbf{k}_{\mathbf{z}\mathrm{A}}\cdot{\bf
r}}-e^{i\mathbf{k}_{\mathbf{z}\mathrm{B}}\cdot{\bf
r}}\right)+\mathbf{y}\cos{\Omega t}\,e^{i\mathbf{k}_{\mathbf{y}}\cdot{\bf
r}},$ (2)
where in the second step the $y$-dependent trigonometric functions were
separated into complex exponentials to show that this field is a superposition
of five plane waves. The amplitudes of these plane waves have a slow temporal
dependence that adiabatically changes the relative weight of the first four
with respect to the fifth, as depicted in Fig. 1(a). The wavevectors of these
five plane waves are
$\displaystyle\mathbf{k}_{\mathbf{x}\mathrm{A}}$
$\displaystyle=k\left(0,1,1\right)/\sqrt{2},$ (3a)
$\displaystyle\mathbf{k}_{\mathbf{x}\mathrm{B}}$
$\displaystyle=k\left(0,-1,1\right)/\sqrt{2},$ (3b)
$\displaystyle\mathbf{k}_{\mathbf{z}\mathrm{A}}$
$\displaystyle=k\left(1,1,0\right)/\sqrt{2},$ (3c)
$\displaystyle\mathbf{k}_{\mathbf{z}\mathrm{B}}$
$\displaystyle=k\left(1,-1,0\right)/\sqrt{2},$ (3d)
$\displaystyle\mathbf{k}_{\mathbf{y}}$
$\displaystyle=k\left(1,0,1\right)/\sqrt{2},$ (3e)
each being orthogonal to the polarization of the corresponding plane wave
(indicated in the subindex). All nonparaxial polarization ellipses then exist
inside the cuboid defined by $x,z\in[0,\sqrt{2}\lambda)$ and
$y\in[0,\sqrt{2}\lambda/4]$ within a time interval $\Omega t\in[0,\pi/2]$,
referred to here as a full 3D polarization (F3DP) cell. This spatio-temporal
cell is depicted in Fig. 1(b). As mentioned earlier, while all polarization
states are explored by this field, its intensity $|{\bf E}|^{2}$ is constant
in space and time.
Figure 1: (a,c) Wavevectors and their polarization for five plane waves that
compose a nonparaxial optical field containing all possible polarization
states, in (a) the construction frame and (c) the laboratory frame. The field
is constructed by adiabatically harmonically varying in time the relative
amplitude between two sets of plane waves (green, and red and blue). The
temporal evolution of the normalized intensity of each group of plane waves is
shown. (b,d) Spatial cell where every possible nonparaxial polarization
ellipse is contained within a temporal interval (F3DP cell) $\Omega
t\in[0,\pi/2]$. The cells are aligned with the reference frame coordinates in
the (b) construction and (d) laboratory frames.
The reference frame just used (termed here the construction frame) is
convenient for understanding why the field spans all possible polarization
states. However, we now change to a reference frame that is better suited for
visualizing the polarization structure and for describing its experimental
generation. We call this new frame the laboratory frame, for which the $z$
direction is chosen to coincide with the wavevector ${\bf k}_{\bf y}$, which
bisects the other four wavevectors and can therefore be regarded as the main
propagation direction. The laboratory frame results from a rotation by
$-\pi/4$ radians around the $y$ axis of the solution in the construction
frame, as shown in Fig. 1(c). Following this rotation both for the spatial
coordinates and the polarization components, we arrive at the expression
$\displaystyle\mathbf{E}$ $\displaystyle=\left[e^{-ik(z+x)/2}\sin{\Omega
t}\left(\mathbf{p}\cos{\frac{ky}{\sqrt{2}}}+\mathbf{m}\,e^{ikx}\sin{\frac{ky}{\sqrt{2}}}\right)+\mathbf{y}\cos{\Omega
t}\right]e^{ikz}$ $\displaystyle=\mathbf{p}\,\frac{\sin{\Omega
t}}{2}\left(e^{i\mathbf{k}_{\mathbf{p}\mathrm{A}}\cdot{\bf
r}}+e^{i\mathbf{k}_{\mathbf{p}\mathrm{B}}\cdot{\bf
r}}\right)+\mathbf{m}\,\frac{\sin{\Omega
t}}{2i}\left(e^{i\mathbf{k}_{\mathbf{m}\mathrm{A}}\cdot{\bf
r}}-e^{i\mathbf{k}_{\mathbf{m}\mathrm{B}}\cdot{\bf
r}}\right)+\mathbf{y}\cos{\Omega t}\,e^{i\mathbf{k}_{\mathbf{y}}\cdot{\bf
r}},$ (4)
where the polarization components $\mathbf{x}$ and $\mathbf{z}$ turned into
the components $\mathbf{p}$ and $\mathbf{m}$, respectively, defined as
$\mathbf{p},\mathbf{m}=\frac{\pm\mathbf{x}+\mathbf{z}}{\sqrt{2}},$ (5)
and the corresponding wavevectors are given by
$\displaystyle\mathbf{k}_{\mathbf{p}\mathrm{A}}$
$\displaystyle=k\left(-1,\sqrt{2},1\right)/2,$ (6a)
$\displaystyle\mathbf{k}_{\mathbf{p}\mathrm{B}}$
$\displaystyle=k\left(-1,-\sqrt{2},1\right)/2,$ (6b)
$\displaystyle\mathbf{k}_{\mathbf{m}\mathrm{A}}$
$\displaystyle=k\left(1,\sqrt{2},1\right)/2,$ (6c)
$\displaystyle\mathbf{k}_{\mathbf{m}\mathrm{B}}$
$\displaystyle=k\left(1,-\sqrt{2},1\right)/2,$ (6d)
$\displaystyle\mathbf{k}_{\mathbf{y}}$ $\displaystyle=k\left(0,0,1\right).$
(6e)
Note that the F3DP cell in Fig. 1(b) now has sides that in the laboratory
frame are at $\pm\pi/4$ with respect to the $x$ and $z$ axes. We then choose
new F3DP cells whose boundaries are aligned with the coordinates of the
laboratory frame, by considering the range that the coordinates must sweep to
span all possible relative amplitudes and phases between the coefficients of
$\mathbf{p}$, $\mathbf{m}$ and $\mathbf{y}$. One such cell is given by
$x\in[-\lambda/2,\lambda/2)$, $y\in[0,\sqrt{2}\lambda/4]$, and
$z\in[0,2\lambda)$, for the same time interval as before. The spatial volume
of this cell is the same as for the previous one: $\lambda^{3}/\sqrt{2}$. This
cell is depicted in Fig. 1(d).
### 2.2 Increasing the size of the F3DP cell
The spatial dimensions of this F3DP cell can be increased by making the field
more directional, that is, by bringing the wavevectors of the four plane waves
with polarization states $\mathbf{p}$ and $\mathbf{m}$ (red and blue in Fig.
1(c)) closer to the main propagation direction ($z$). It is useful to regard
this transformation as a composition of the two transformations depicted in
Fig. 2(a): First, a decrease of the angle $2\alpha$ subtended by each pair of
wavevectors with the same polarization (blue-blue and red-red pairs in Fig.
2(a)). Second, a decrease in the angle $\beta$ between the projection onto the
$xz$ plane of each of these four wavevectors and the $z$ axis (so that
$2\beta$ is the angle between the projections of the blue-red pairs in Fig.
2(a)). The wavevectors then take the form
$\displaystyle\mathbf{k}_{\mathbf{p}\mathrm{A}}$
$\displaystyle=k\left(-\cos\alpha\sin\beta,\sin\alpha,\cos\alpha\cos\beta\right),$
(7a) $\displaystyle\mathbf{k}_{\mathbf{p}\mathrm{B}}$
$\displaystyle=k\left(-\cos\alpha\sin\beta,-\sin\alpha,\cos\alpha\cos\beta\right),$
(7b) $\displaystyle\mathbf{k}_{\mathbf{m}\mathrm{A}}$
$\displaystyle=k\left(\cos\alpha\sin\beta,\sin\alpha,\cos\alpha\cos\beta\right),$
(7c) $\displaystyle\mathbf{k}_{\mathbf{m}\mathrm{B}}$
$\displaystyle=k\left(\cos\alpha\sin\beta,-\sin\alpha,\cos\alpha\cos\beta\right),$
(7d) $\displaystyle\mathbf{k}_{\mathbf{y}}$
$\displaystyle=k\left(0,0,1\right),$ (7e)
where the particular case in Eq. (6) corresponds to $\alpha=\beta=\pi/4$. Note
that if $\beta=\pi/4$ these wavevectors remain perpendicular to the
corresponding polarization vectors (${\bf p}$, ${\bf m}$, or ${\bf y}$)
regardless of the value of $\alpha$, and in fact the intensity of the field
remains constant. However, changing to $\beta\neq\pi/4$ does require the
rotation of the polarization vectors for the first four plane waves according
to
$\mathbf{p},\mathbf{m}=\pm\mathbf{x}\cos{\beta}+\mathbf{z}\sin{\beta},$ (8)
in order to ensure plane-wave transversality. Since for $\beta\neq\pi/4$ these
two polarization components are not mutually orthogonal, the intensity of the
field is no longer spatio-temporally uniform. The field can now be written as
$\displaystyle\mathbf{E}$
$\displaystyle=\left[e^{-i(\kappa_{z}z+\kappa_{x}x)}\sin{\Omega
t}\left(\mathbf{p}\cos{\kappa_{y}y}+\mathbf{m}\,e^{i2\kappa_{x}x}\sin{\kappa_{y}y}\right)+\mathbf{y}\cos{\Omega
t}\right]e^{ikz}$ $\displaystyle=\mathbf{p}\,\frac{\sin{\Omega
t}}{2}\left(e^{i\mathbf{k}_{\mathbf{p}\mathrm{A}}\cdot{\bf
r}}+e^{i\mathbf{k}_{\mathbf{p}\mathrm{B}}\cdot{\bf
r}}\right)+\mathbf{m}\,\frac{\sin{\Omega
t}}{2i}\left(e^{i\mathbf{k}_{\mathbf{m}\mathrm{A}}\cdot{\bf
r}}-e^{i\mathbf{k}_{\mathbf{m}\mathrm{B}}\cdot{\bf
r}}\right)+\mathbf{y}\cos{\Omega t}\,e^{i\mathbf{k}_{\mathbf{y}}\cdot{\bf
r}},$ (9)
where the spatial frequencies between contributions are given by
$\displaystyle\kappa_{x}$ $\displaystyle=k\cos{\alpha}\sin{\beta},$ (10a)
$\displaystyle\kappa_{y}$ $\displaystyle=k\sin{\alpha},$ (10b)
$\displaystyle\kappa_{z}$ $\displaystyle=k(1-\cos{\alpha}\cos{\beta}).$ (10c)
The dimensions of the F3DP cell are inversely proportional to these spatial
frequencies, as shown in Fig. 2(b), and can be made in principle arbitrarily
large by making the angles $\alpha$ and $\beta$ smaller. Note that for
$\beta\neq\pi/4$ the polarization vectors $\mathbf{p}$, $\mathbf{m}$ and
$\mathbf{y}$ no longer define an orthogonal set, but they are still linearly
independent, and all possible relative amplitudes and phases between them are
swept in the cell, giving rise to all polarization states. However, for small
$\beta$ some polarization states are present only in regions of relatively low
intensity, and their coverage becomes less uniform.
Figure 2: (a) The dimensions of the F3DP cell can be controlled by changing
the angles $\alpha$ and $\beta$ for the wavevectors and the corresponding
polarization vectors. The angle between two wavevectors with the same
polarization state (blue-blue or red-red pairs) is $2\alpha$, and the angle
between the projections onto the $xz$ plane of these two pairs of wavevectors
(blue-red pairs) is $2\beta$. (b) The size of the F3DP cell increases as
$\alpha$ and $\beta$ decrease: the smaller cell in the figure corresponds to
$\alpha=\beta=\pi/4$, while the larger one (whose dimensions in $x$ and $y$
are twice those of the smaller one) corresponds to
$\alpha=\tilde{\alpha}=\arcsin[1/(2\sqrt{2})]=0.3614$ and
$\beta=\tilde{\beta}=\arcsin(1/\sqrt{14})=0.2706$.
This field can be implemented experimentally by focusing with a microscope
objective the light emerging from five points at the back focal plane, where
the polarization and phase of each is appropriately prepared. The separation
of the points sets the value of the angles $\alpha$ and $\beta$. The minimum
numerical aperture (in air) required to produce this field is
$(1-\cos^{2}\alpha\cos^{2}\beta)^{1/2}$.
## 3 The polarization structure of the field
We now describe the polarization structure of the field. We start by
describing the global field distribution by finding its spatial unit cell. We
then study the spatio-temporal structure of polarization singularities inside
a F3DP cell.
### 3.1 The field’s global structure
Figure 3(a) depicts a unit cell of the field, which consists of two stacked
F3DP cells (red and green cuboids). The cell dimensions are expressed in units
of $\kappa_{j}j$ ($j=x$, $y$, $z$), so the representation is valid for any
value of $\alpha$ and $\beta$. The cell A (red) is the F3DP cell described in
the previous section, which can be written as $\kappa_{x}x\in[-\pi/2,\pi/2)$,
$\kappa_{y}y\in[0,\pi/2]$ and $\kappa_{z}z\in[0,2\pi)$. The cell B (green) is
an F3DP cell displaced by $\pi/2$ in each dimensionless coordinate
$\kappa_{j}j$ with respect to A. It turns out that the polarization
distributions in A and B are identical except for a mirroring in $y$, as one
can see from applying a shift of $\pi/2$ to $\kappa_{x}x$, $\kappa_{y}y$ and
$\kappa_{z}z$ in Eq. (9). The entire field can then be constructed by stacking
these unit cells, as shown in Figs. 3(b,c). The positions where the unit cells
must be stacked are deduced from the changes in $\kappa_{j}j$ that leave the
field in Eq. (9) invariant: i) a displacement of $n\pi$ ($n\in\mathbb{Z}$) in
$\kappa_{x}x$ and $\kappa_{z}z$ (which results in the brick structure in Fig.
3(b) for each type of cell), and ii) a displacement of $n\pi$ in $\kappa_{y}y$
and a shift of $n\pi$ in $\kappa_{x}x$ or $\kappa_{z}z$ (depicted in Fig.
3(c)).
Figure 3: (a) The field unit cell consists of two F3DP cells (red and green
boxes) of dimensions $\Delta x=\pi/\kappa_{x}=\lambda/(2\cos\alpha\sin\beta)$,
$\Delta y=\pi/(2\kappa_{y})=\lambda/(4\sin\alpha)$ and $\Delta
z=2\pi/\kappa_{z}=\lambda/(1-\cos\alpha\cos\beta)$. The cell dimensions are
expressed in units of $\kappa_{j}j$ ($j=x$, $y$, $z$). Projections onto the
$xy$, $xz$ and $yz$ planes are shown in order to aid with 3D visualization.
The (green) cell B is displaced by $\pi/2$ rad in $\kappa_{x}x$, $\kappa_{y}y$
and $\kappa_{z}z$ with respect to the (red) cell A, and its polarization state
distribution is the same as in the red cell but mirrored in $y$. The global
field structure is reproduced by stacking unit cells following two rules:
stacking unit cells displaced (b) by $n\pi$ (where $n\in\mathbb{Z}$) in
$\kappa_{x}x$ and $\kappa_{z}z$ (brick structure), and (c) by $n\pi$ in
$\kappa_{y}y$ and displaced by $n\pi$ in $\kappa_{x}x$ or in $\kappa_{z}z$.
### 3.2 The skeleton of the polarization distribution: L-lines and C-lines
In order to gain intuition about the spatiotemporal distribution of the
polarization states, we now study the L-lines (lines of linear polarization)
and C-lines (lines of circular polarization) inside the F3DP cell A. Figure 4
shows the evolution of these lines for $\alpha$, $\beta=\pi/4$. (The
corresponding polarization structure for $\alpha=\tilde{\alpha}$ and
$\beta=\tilde{\beta}$ is shown in the supplementary material in section A.1.)
Once we describe the polarization state distribution inside cell A, the
corresponding distribution in cell B can be obtained by just mirroring in $y$,
as mentioned earlier.
Figure 4: Temporal evolution of the L-lines and C-lines inside the F3DP cell A
for $\alpha$, $\beta=\pi/4$.
At different times this idealized field presents L-lines, L-planes and even an
L-volume (the latter two being of course unstable under field perturbations).
For $\Omega t=0$, the polarization state is $\mathbf{y}$, so the entire 3D
space is an L-volume. On the other hand, for $\Omega t=\pi/2$, the zones of
linear polarization are two sets of L-planes, which are constant in $x$ and in
$y$, respectively. Within the interval $\Omega t\in(0,\pi/2)$, the field
presents L-lines whose positions remain constant in time. A detailed
explanation of the calculations of the L-lines is provided in the
supplementary material in subsection A.1.1.
Figure 5: Temporal evolution on the unit sphere in real space of (a,b) the
electric field vector direction for linearly polarized states, and (c,d) the
normalized spin density vector for circularly polarized states, for (a,c)
$\alpha=\beta=\pi/4$, and (b,d) $\alpha=\tilde{\alpha}$ and
$\beta=\tilde{\beta}$. Each curve correspond to one temporal instant, and the
spacing between temporal intervals is constant. The C-lines are plotted
starting from the time $t_{C}$ when they appear: $\Omega t_{C}=\pi/4$ for
$\beta=\pi/4$ and $\Omega t_{C}=0.6331$ for $\beta=\tilde{\beta}$.
Figures 5(a,b) show that, for $\Omega t\in[0,\pi/2]$, the electric field
orientation at the regions of linear polarization covers the complete unit
sphere. Note that antipodal points correspond to the same linear polarization
state. Each curve (or more correctly, each pair of curves) represents all the
values of the orientation of the linearly polarized states at a given instant.
These curves are independent of $\alpha$, and, for $\beta=\pi/4$ (Fig. 5(a)),
they are parallels. Decreasing $\beta$ (Fig. 5(b) for $\beta=\tilde{\beta}$)
distorts the curves while maintaining full directional coverage. It is clear
from Eq. (9) that, for any value of $\beta$, the $\mathbf{y}$ polarization
component of the field decreases while the $\mathbf{p}$ and $\mathbf{m}$
components grow. Eventually, all the polarization ellipses lie on the $xz$
plane at $\Omega t=\pi/2$. This fact explains why, in the unit sphere in Figs.
5(a,b), the curves move from the poles towards the equator as time increases.
The C-lines are found from the constraint $\mathbf{E}\cdot\mathbf{E}=0$ [15].
Since $\mathbf{E}\cdot\mathbf{E}$ is a complex quantity, this constraint
imposes two conditions, each defining a surface, and the intersections of
these surfaces are the C-lines. Figure 4 shows the normalized spin density
vector
$\mathbf{s}=\mathrm{Im}\>(\mathbf{E}^{*}\times\mathbf{E})/|\mathbf{E}|^{2}$
along the C-lines. This vector is normal to the plane containing the ellipse,
and its magnitude is proportional to the area of the normalized ellipse,
ranging from 0 (linear polarization) to 1 (circular polarization). The
equations describing the C-lines and $\mathbf{s}$ are inlcuded in the
supplementary material in subsection A.1.2.
There are no C-lines at $t=0$; they appear at a time $t_{C}$, which depends
only on $\beta$, and then undergo two topological transitions within the
interval $\Omega t\in[0,\pi/2]$. The first transition always occurs at the
time $\Omega t_{T1}=\pi/4$, and the other one occurs at $\Omega t_{T2}$, which
depends on $\beta$. The dependence of $t_{C}$ and $t_{T2}$ with $\beta$ is
discussed in the supplementary material. For $\beta=\pi/4$ the two topological
transitions coincide with the appearance of the C-lines, $\Omega t_{C}=\Omega
t_{T1}=\Omega t_{T2}=\pi/4$, while choosing a smaller $\beta$ causes $t_{C}$
to decrease and $t_{T1}$ to increase. Conversely, if $\alpha$ is reduced, the
C-lines get stretched out along $y$ and $z$ and shrink slightly along $x$, but
$t_{C}$ and $t_{T2}$ remain unaltered, i.e., $\alpha$ does not alter the
topological properties of the C-lines. Studying the case when $\beta=\pi/4$
and one case when $\beta<\pi/4$ (in subsection A.1.2 in the supplementary
material) is sufficient then to illustrate the topological behavior of the
C-lines. Figure 4 shows the L-lines and C-lines for $\alpha=\beta=\pi/4$.
Right after appearing at $\Omega t=\pi/4$ as interconnected straight lines,
the C-lines split into two helices within the cell. As time increases, these
helices straighten until they become parallel straight lines for $\Omega
t=\pi/2$. At this time the $\mathbf{y}$ field component vanishes, so the spin
at every point is in the $\mathbf{y}$ direction (normal to the main
propagation direction), leading to “photonic wheels” with transverse spin [16,
17].
We plot the spin directions $\mathbf{s}$ along the C-lines for a given
temporal instant as a curve on the unit sphere in Figs. 5(c,d). For $\Omega
t=\pi/4$, $\mathbf{s}$ spans all directions within the $xz$ plane for
$\beta=\pi/4$ (equator of the sphere in Fig. 5(c)). After $\Omega t=\pi/4$,
each of the two C-lines traces identical parallels, each on a different
hemisphere: the vector $\mathbf{s}$ at the C-line in the region
$-0.5\lambda<x<0$ ($0<x<0.5\lambda$) evolves in time towards the direction
$+y$ ($-y$). As was the case in Figs. 5(a,b), the curves do not depend on
$\alpha$ but they are no longer parallels for $\beta<\pi/4$ (Fig. 5(d)), while
still covering all possible spin directions along the C-lines within the time
interval $\Omega t\in[0,\pi/2]$. Note from Fig. 5(d) that the topological
transition at $t_{T2}$ ($\Omega t_{T2}=1.2094$ for $\beta=\tilde{\beta}$)
corresponds to topological transitions of the spin curves at $\pm{\bf x}$ over
the unit sphere: the curves that cycle around the $z$ axis correspond to
$t<t_{T2}$, while those that cycle around the $y$ axis correspond to
$t>t_{T2}$.
The temporal evolution of the field skeleton for $\alpha=\beta=\pi/4$ (Fig. 4)
and for $\alpha=\tilde{\alpha}$, $\beta=\tilde{\beta}$ (Fig. 8) are included
in the supplementary video 1 and video 2, respectively. The evolution of the
electric field orientation at the linearly polarized regions (Figs. 5 (a,c))
and of the vector $\mathbf{s}$ at the circularly polarized regions (Figs. 5
(b,d)) are also included in each video.
## 4 4D optical skyrmionic textures
As pointed out in the introduction, a 3D polarization ellipse can be fully
characterized by the positions of two indistinguishable points,
$\mathbf{p}_{1}$ and $\mathbf{p}_{2}$, over the unit sphere
($|\mathbf{p}_{1}|=|\mathbf{p}_{2}|=1$). Different variants of this type of
two-point construction exist: one proposed by Hannay [4] based on Majorana’s
representation of spin systems; one referred to as the Poincarana
representation that is naturally linked to geometric phase for nonparaxial
fields [5]; and a more recent one [6] defined to satisfy certain statistical
conditions. For all of them, the bisector of $\mathbf{p}_{1}$ and
$\mathbf{p}_{2}$ is parallel to $\mathbf{s}$, the line joining
$\mathbf{p}_{1}$ and $\mathbf{p}_{2}$ is parallel to the major axis of the
ellipse, and the angular separation between $\mathbf{p}_{1}$ and
$\mathbf{p}_{2}$ is a monotonic function of the ellipticity: the two points
coincide for circular polarization, and they are antipodal for linear
polarization. However, each construction encodes ellipticity in the angular
separation of the points in a different manner.
For a field to span all 3D polarization states, each of the two points in any
of these representations must span the sphere independently, so that all
combinations of the two point positions are represented. We then define a
Skyrme number for this 4D space as
$n_{S}=\iiiint\Sigma_{S}\,\mathrm{d}x\,\mathrm{d}y\,\mathrm{d}z\,\mathrm{d}t,$
(11)
where
$\Sigma_{S}=\frac{1}{2(4\pi)^{2}}\sum^{4}_{i,j,k,l=1}\epsilon_{ijkl}\;\left[\mathbf{p}_{1}\cdot\left(\partial_{i}\mathbf{p}_{1}\times\partial_{j}\mathbf{p}_{1}\right)\right]\;\left[\mathbf{p}_{2}\cdot\left(\partial_{k}\mathbf{p}_{2}\times\partial_{l}\mathbf{p}_{2}\right)\right]$
(12)
is the Skyrme density, which corresponds to the Jacobian determinant between
the coordinates in the abstract closed space (in this case the coordinates of
the two points over the sphere) and the physical space (in this case space-
time). Here $\epsilon_{ijkl}$ is the fourth-rank Levi-Civita tensor (which
differs from zero only for the 24 terms for which all indices are different,
out of the 64 in the sum), and $\partial_{i}$ for $i=1,2,3,4$ represents a
derivative in $x,y,z,t$, respectively. The factor of $1/(4\pi)^{2}$ is for
normalizing the area of the sphere for each point, and the factor $1/2$
accounts for the indistinguishability of the two points. Due to the properties
of the cross product, the 24 nonzero terms in the above expression reduce to
the 6 terms corresponding to
$(i,j,k,l)\rightarrow(x,y,z,t),(z,t,x,y),(x,z,y,t),(x,t,z,y),(z,y,x,t),(t,y,z,x)$,
multiplied by a factor of 4.
This Skyrme number can be used for any of the three constructions mentioned
earlier [4, 5, 6]. Here, we choose to use the Poincarana construction [5]
(Fig. 6), where the midpoint between the points matches exactly $\mathbf{s}$.
The Poincarana points are given by
$\displaystyle\mathbf{p}_{1,2}=\mathbf{s}\pm\sqrt{1-|\mathbf{s}|^{2}}\frac{\mathbf{a}}{|\mathbf{a}|},$
(13)
where $\mathbf{a}$ is the major semi-axis of the polarization ellipse, which
is obtained as $\mathbf{a}=\mathrm{Re}(e^{-i\phi}\mathbf{E}/|\mathbf{E}|)$
with $\phi=\mathrm{Arg}(\mathbf{E}\cdot\mathbf{E})/2$ [6]. We used this
construction because it seems to lead to the simplest expressions for the
Skyrme density in some cases.
Figure 6: Poincarana representation construction: A fully polarized 3D
polarization state is described by two indistinguishable points
$\mathbf{p}_{1,2}$ on the unit sphere. The normalized spin density vector
$\mathbf{s}$ is the centroid of the two points, and the major semi-axis vector
$\mathbf{a}$ is parallel to the point separation.
For any $\alpha$ and $\beta$, the sign of $\Sigma_{S}$ is uniform within each
cell, A or B, but it is opposite between the two types of cell (due to the
reversal in $y$ of the polarization distribution). This sign also switches
temporally every time $\Omega t$ is an integer multiple of $\pi/2$. We found
that $\Sigma_{S}$ is independent of $z$ for any $\alpha$ and $\beta$, and it
has periodic dependence on $x$ for $\beta<\pi/4$ without changing sign (the
sign depending only on $y$ and $t$). Further, for $\beta=\pi/4$, we observed
through numeric evaluation that $\Sigma_{S}$ actually does not depend on $x$,
and the long analytic expression for $\Sigma_{S}$ greatly simplifies by
setting $x=z=0$, giving
$\Sigma_{S}=\frac{k^{3}\Omega}{\pi^{2}}\cos\alpha(\sqrt{2}-\cos\alpha)\cos{\Omega
t}\sin^{3}{\Omega t}\sin{2\kappa_{y}y}.$ (14)
This expression for $\Sigma_{S}$ is plotted in Fig. 7(a) as a function of $y$
and $t$ for $\alpha=\beta=\pi/4$. For $\beta=\pi/4$, the Skyrme number inside
a cell of type A or B can then be easily calculated analytically by
integrating the Skyrme density in Eq. (14), giving the results $n_{S}=-1$ for
cell A and $n_{S}=+1$ for cell B within the interval $\Omega t\in[0,\pi/2]$.
We also verified numerically that $n_{S}=-(+)1$ in cell A(B) for several
values of $\alpha$ and $\beta$. We then refer to the F3DP cells where
$|n_{S}|=1$ as skyrmionic F3DP cells (Fig. 7(b)). Other versions of A and B
displaced in $x$ and/or $z$ are also skyrmionic F3DP cells, since these
displacements simply shift the relative phases between the components.
Note that, since $\Sigma_{S}$ does not depend on $z$ for any $\beta$, the
Skyrme number accumulates linearly and indefinitely as the integration range
in $z$ increases. As a consequence, one can construct cells where $n_{S}=\pm
1$ that are not F3DP cells. There are also F3DP cells (containing each
polarization state) whose volume is the same as that for A or B, but for which
$n_{S}=0$; we call these nonskyrmionic F3DP cells. An example corresponding to
$\kappa_{x}x\in[0,\pi]$, $\kappa_{y}y\in[0,2\pi]$ and
$\kappa_{z}z\in[0,\pi/2]$ is shown in Fig. 7(c), which is composed of segments
of cells of both types leading to cancellation of the Skyrme density
contributions. More details about the construction of F3DP nonskyrmionic cells
can be found in Section A.2 of the supplementary material.
Figure 7: (a) Skyrme density for $\alpha=\beta=\pi/4$ in units of $\Omega
k^{3}$. The Skyrme density beyond the plotted region is repeated. (b) F3DP
skyrmionic cells. The Skyrme number is $-(+)1$ in cell A (B). (c) A F3DP
nonskyrmionic cell that contains all 3D polarization ellipses and has the same
volume as the F3DP skyrmionic cells, but has Skyrme number equal to 0.
## 5 Conclusions
We proposed a simple optical field that covers all possible states of
nonparaxial polarization. This field is given by a superposition of five
linearly-polarized plane waves whose amplitudes vary slowly in time. The
polarization texture is periodic, and is composed of 4D skyrmionic spatio-
temporal cells. The polarization distribution in each cell is described in
terms of its “skeleton” given by the spatiotemporal manifolds of purely linear
or circular polarization.
While the description in this work is theoretical, the experimental
implementation of the field proposed here is probably not difficult to
achieve, and this is currently underway in our group. The main challenge is
most likely not the implementation of the field but its measurement. Current
techniques for measuring nonparaxial polarization distributions rely on
scanning a nanoparticle in order to measure one point at a time. Measuring
this field would require scanning over four dimensions. We are currently
working on an efficient measurement scheme that will allow sampling the 4D
space more efficiently.
The simple forms proposed here are most likely not the only F3DP
electromagnetic fields possible, although they might be the simplest ones. It
would be interesting to investigate whether F3DP fields exist for which the
Skyrme density does not change sign in space or time (making the skyrmionic
structure more resilient to field perturbations). These fields might be
periodic arrays of cells like the F3DP field here, or consist of a single cell
occupying a theoretically infinite volume, in analogy to the 2D skyrmionic
beams in [2]. Nevertheless, the constraints imposed by the Helmholtz equation
and the divergence condition are stronger in the nonparaxial case than they
are for paraxial beams.
## Acknowledgement
This research received funding from the Agence Nationale de Recherche (ANR)
through the project 3DPol, ANR-21-CE24-0014-01. D. M. acknowledges Ministerio
de Universidades, Spain, Universidad Miguel Hernández and the European Union
(Next generation EU fund) for a Margarita Salas grant from the program Ayudas
para la Recualificación del Sistema Universitario Español.
## References
* [1] M. Born and E. Wolf, Principles of Optics: electromagnetic theory of propagation, interference and diffraction of light (Cambridge University Press, 2001), pp. 25-33.
* [2] A.M. Beckley, T.G. Brown, and M.A. Alonso, “Full Poincaré beams,” Opt. Express 18, 10777-10785 (2010).
* [3] N. Nagaosa and Y. Tokura, “Topological properties and dynamics of magnetic skyrmions,” Nat. Nanotechnol. 8, 899-911 (2013).
* [4] J.H. Hannay, “The Majorana representation of polarization, and the Berry phase of light,” J. Mod. Opt. 45, 1001-1008 (1998).
* [5] K.Y. Bliokh, M.A. Alonso and M.R. Dennis, “Geometric phases in 2D and 3D polarized fields: geometrical, dynamical, and topological aspects,” Rep. Prog. Phys. 82, 122401 (2019).
* [6] M. A. Alonso, “Geometric descriptions for the polarization of nonparaxial light: a tutorial,” arXiv:2008.02720v2 (2022).
* [7] S. Donati, L. Dominici, G. Dagvadorj, D. Ballarini, M. De Giorgi, A. Bramati, G. Gigli, Y. G. Rubo, M. H. Szymańska and D. Sanvitto, “Twist of generalized skyrmions and spin vortices in a polariton superfluid,” Proc. Natl. Acad. Sci. U.S.A. 113, 14926–14931 (2016).
* [8] S. Gao, F. C. Speirits, F. Castellucci, S. Franke-Arnold, S.M. Barnett and J.B. Götte, “Paraxial skyrmionic beams,” Phys. Rev. A 102, 053513 (2020).
* [9] S. Tsesses, E. Ostrovsky, K. Cohen, B. Gjonaj, N.H. Lindner, and G. Bartal, “Optical skyrmion lattice in evanescent electromagnetic fields,” Science 361, 993-996 (2018).
* [10] H. Ge, X.-Y. Xu, L. Liu, R. Xu, Z.-K. Lin, S.-Y. Yu, M. Bao, J.-H. Jiang, M.-H. Lu, and Y.-F. Chen, “Observation of Acoustic Skyrmions,” Phys. Rev. Lett. 127, 144502 (2021).
* [11] R. D. Muelas-Hurtado, K. Volke-Sepulveda, J. L. Ealo, F. Nori, M. A. Alonso, K. Y. Bliokh, E. Brasselet, “Observation of polarization singularities and topological textures in sound waves,” Phys. Rev. Lett. 129, 204301 (2022).
* [12] L. Du, A. Yang, A. V. Zayats, and X. Yuan, “Deep-subwavelength features of photonic skyrmions in a confined electromagnetic field with orbital angular momentum,” Nat. Phys. 15, 650-654 (2019).
* [13] R. Gutiérrez-Cuevas and E. Pisanty, “Optical polarization skyrmionic fields in free space,” J. Opt. 23, 024004 (2021).
* [14] D. Sugic, R. Droop, E. Otte, D. Ehrmanntraut, F. Nori, J. Ruostekoski, C. Denz, and M.R. Dennis, “Particle-like topologies in light,” Nat. Commun. 12, 6785 (2021).
* [15] M.V. Berry and M.R. Dennis, “Polarization singularities in isotropic random vector waves,” Proc. R. Soc. A 457, 141-55 (2001).
* [16] K.Y. Bliokh and F. Nori, “Transverse and longitudinal angular momenta of light,” Phys. Rep. 592, 1-38 (2015).
* [17] A. Aiello, P. Banzer, M. Neugebauer, and G. Leuchs, “From transverse angular momentum to photonic wheels,” Nat. Photon. 9, 789-795 (2015).
## Appendix A Supplementary material
We include in the supplementary material details about the L-lines, C-lines,
and of the nonskyrmionic F3DP cells.
### A.1 Skeleton
In this section, we provide details of the derivation of the L-lines, and
equations for the C-lines, the normalized spin denity vector $\mathbf{s}$, and
the topological transitions that the C-lines undergo. We also study the
evolution of the L-lines and C-lines for $\beta<\pi/4$, in particular for the
case $\alpha=\tilde{\alpha}$ and $\beta=\tilde{\beta}$, shown in Fig. 8.
#### A.1.1 The L-lines within cell A
Here, we describe how to obtain the L-lines for the field defined in Eq. (9).
As mentioned in the main text, for $\Omega t=0$, the field has a polarization
state $\mathbf{y}$, which means that the 3D space itself is an L-volume. For
$\Omega t=\pi/2$, on the other hand, only the components $\mathbf{p}$ and
$\mathbf{m}$ survive. If, in addition, $\kappa_{y}y=n\pi/2$
($n\in\mathbb{Z}$), then only one linear polarization component remains, so
this condition defines planes of constant $y$ with $\mathbf{p}$ (for
$\kappa_{y}y=n\pi$) or $\mathbf{m}$ (for $\kappa_{y}y=(2n+1)\pi/2$)
polarization state. Conversely, for $\kappa_{y}y\neq n\pi/2$, the phase
between the $\mathbf{p}$ and $\mathbf{m}$ components must be $n\pi$ for the
state to be linear, which defines a set of L-planes $x=n\pi/(2\kappa_{x})$.
Within the intermediate interval $\Omega t\in(0,\pi/2)$ we find L-lines. The
first subset of L-lines is deduced for $\kappa_{y}y=n\pi/2$. Under this
spatial constraint, only two polarization components survive ($\mathbf{y}$ and
$\mathbf{p}$ or $\mathbf{y}$ and $\mathbf{m}$), and in order to have linearly
polarized light, the relative phase between these components must be $n\pi$,
which, together with the condition $\kappa_{y}y=n\pi/2$, defines a set of
L-lines that lie on the planes $y=0$ and $y=\lambda/(4\sin\alpha)$ in the F3DP
cell A. The second set of L-lines is deduced for $\kappa_{y}y\neq n\pi/2$
(i.e., when the three linear polarization components survive), imposing that
the two relative phases between the three linear polarization components must
be integer multiples of $\pi$.
#### A.1.2 The C-lines within cell A
The equations defining the C-lines for the field defined in Eq. (9) can be
found from the condition $\mathbf{E}\cdot\mathbf{E}=0$, whose real and
imaginary parts lead to
$\displaystyle\cos^{2}{\Omega t}\cos{2\kappa_{z}z}+\sin^{2}{\Omega
t}\left(\cos{2\kappa_{x}x}-\cos{2\beta}\sin{2\kappa_{y}y}\right)$
$\displaystyle=0,$ (15a) $\displaystyle\cos^{2}{\Omega
t}\sin{2\kappa_{z}z}-\sin^{2}{\Omega t}\cos{2\kappa_{y}y}\sin{2\kappa_{x}x}$
$\displaystyle=0.$ (15b)
Using
$\mathbf{s}=\mathrm{Im}\>(\mathbf{E}^{*}\times\mathbf{E})/|\mathbf{E}|^{2}$,
we obtain the normalized spin density vector $\mathbf{s}$ for the field
defined in Eq. (9):
$\mathbf{s}=\frac{1}{|\mathbf{E}|^{2}}\begin{pmatrix}-\sin{2\Omega
t}\sin{\beta}\left[\cos{\kappa_{y}y}\sin{(\kappa_{z}z+\kappa_{x}x)}+\sin{\kappa_{y}y}\sin{(\kappa_{z}z-\kappa_{x}x)}\right]\\\
-\sin^{2}{\Omega t}\sin{2\beta}\sin{2\kappa_{x}x}\sin{2\kappa_{y}y}\\\
\sin{2\Omega
t}\cos{\beta}\left[\cos{\kappa_{y}y}\sin{(\kappa_{z}z+\kappa_{x}x)}-\sin{\kappa_{y}y}\sin{(\kappa_{z}z-\kappa_{x}x)}\right]\end{pmatrix},$
(16)
where
$|\mathbf{E}|^{2}=1-\sin^{2}{\Omega
t}\cos{2\beta}\cos{(2\kappa_{x}x)\sin{(2\kappa_{y}y)}}$ (17)
is the field intensity.
For $\beta<\pi/4$, we observed for multiple values of $\beta$ that the C-lines
in the F3DP cell A always appear as points at $x=\pi/(2\kappa_{x})$,
$y=\pi/(4\kappa_{y})$, $z=\pi/\kappa_{z}$, and $x=\pi/(2\kappa_{x})$,
$y=\pi/(4\kappa_{y})$, $z=0$. (It is only for $\beta=\pi/4$ that they appear
as lines and they also appear at other regions.) Evaluating $\mathbf{s}$ at
these spatial coordinates reveals that $\mathbf{s}$ points in the $\pm z$
direction at these points, as can be seen for $\alpha=\tilde{\alpha}$ and
$\beta=\tilde{\beta}$ in Fig. 8 and on the sphere in Fig. 5(d). Setting
$|\mathbf{s}|=1$ (which implies circular polarization) at these points, we
arrive at the expression
$\Omega
t_{C}=\arccos\left(\frac{\cos\beta}{\sqrt{1+\frac{\cos{2\beta}}{2}}}\right)$
(18)
for the time $t_{C}$ when the C-lines appear: $\Omega t_{C}=\pi/4$ for
$\beta=\pi/4$ (Fig. 4) and $\Omega t_{C}=0.6331$ for $\beta=\tilde{\beta}$
(Fig. 8). As mentioned in the main text, we observed a topological transition
at $\Omega t_{T1}=\pi/4$ for any $\beta$. At a given time $t_{T2}$, there is
another topological transition for the C-lines. We observed that there are two
helical C-lines inside the F3DP cell A meeting at the points $x=0$,
$y=\pi/(4\kappa_{y})$ and $z=\pi/(2\kappa_{z})$, and $x=0$,
$y=\pi/(4\kappa_{y})$ and $z=\pi/\kappa_{z}$ for any $\beta$. Evaluating the
vector $\mathbf{s}$ in these coordinates we found that the vector $\mathbf{s}$
points along $+x$ (-$x$) for any $\beta$ at these points. Imposing that
$|\mathbf{s}|=1$, we arrive at
$\Omega
t_{T2}=\arccos\left(\frac{\sin\beta}{\sqrt{1-\frac{\cos{2\beta}}{2}}}\right).$
(19)
The topological transition is noticeable from the unit sphere in Fig. 5(d). It
occurs when $\mathbf{s}$ points along $x$ and $-x$ and it is a very fast
transition. For $\beta=\tilde{\beta}$, $\Omega t_{T2}=1.2094$, while for
$\beta=\pi/4$, $\Omega t_{T2}=\pi/4$. As pointed out in the main text, for
$\beta=\pi/4$, the topological transitions and the appearance of the C-lines
occur at the same instant $\Omega t_{C}=\Omega t_{T1}=\Omega t_{T2}=\pi/4$.
However, as $\beta$ decreases, the instants $t_{C}$ and $t_{T2}$ move away
from each other in time while $\Omega t_{T1}$ remains at $\pi/4$ as shown in
the plot in Fig. 9.
Figure 8: Temporal evolution of the L-lines and C-lines inside the F3DP cell A
for $\alpha=\tilde{\alpha}$ and $\beta=\tilde{\beta}$ (for which the
dimensions of the F3DP cell in $x$ and $y$ are twice as large as for the cell
corresponding to $\alpha=\beta=\pi/4$). The polarization ellipses are plotted
for the C-lines and L-lines, and the normalized spin density vector is plotted
for the C-lines. At $\Omega t_{C}=0.6331$, the C-lines appear as points. There
are two topological transitions, occurring at $\Omega t_{T1}=\pi/4$ and at
$\Omega t_{T2}=1.2094$.
Figure 8 shows the L-lines, the C-lines and the $\mathbf{s}$ vector inside the
F3DP cell A for $\alpha=\tilde{\alpha}$ and $\beta=\tilde{\beta}$. Here, the
C-lines appear as points at $\Omega t_{C}=0.6331$. The plot for $\Omega
t=\pi/4-\pi/50$ in Fig. 8 shows that shortly after $t_{C}$, the C-lines become
small closed loops that spill into neighboring F3DP cells (as can be deduced
from the brick-like structure in Fig. 3(b)). These loops grow until, right
before $\Omega t=\Omega t_{T1}=\pi/4$, segments of them almost lie on the
$y=0$ plane while others almost lie on the $y=\lambda/(4\sin\alpha)$ plane. At
the topological transition at $\Omega t_{T1}=\pi/4$, pairs of C-lines segments
lying on the same plane of constant $y$ get linked by straight C-lines. These
straight lines come from closed C-lines loops in the upper and lower F3DP
cells. Right after $\Omega t=\pi/4$, the segments almost lying on the $y=0$
and $y=\lambda/(4\sin\alpha)$ surfaces before $\Omega t=\pi/4$ disappear from
the cell; they move to the F3DP cells above and below to form the same
structure they form in A but displaced and reflected along $y$. This structure
consists of two tangled helical C-lines winding around each other. At the
second topological transition at $\Omega t_{T2}$, the two helical lines meet
at two points, and right after $t_{T2}$, the C-line segments within $x>0$
combine to form a new C-line, and the same for $x<0$, so there are now two
untangled helical C-lines. After $t_{T2}$, the C-lines behave as they did for
$\beta=\pi/4$: as $\Omega t$ increases the lines straighten gradually until
they become straight for $\Omega t=\pi/2$, when the ellipses are constrained
to the $xz$ plane, and pure transverse spin appears.
Figure 9: Dependence on $\beta$ of the time $t_{C}$ when the C-lines appear,
and of the topological transition times $t_{T1}$ and $t_{T2}$.
### A.2 F3DP nonskyrmionic cells
In section 4, we reported the existence of F3DP cells where $n_{S}=0$. We show
in Fig. 10 examples of regions containing all 3D polarization ellipses where
$n_{S}=0$. The regions in (a) and (b) are constructed as follows: First, we
take one or more regions in A that sum up to half of the A volume by slicing A
along $z$. Then, we take the regions in B containing all the polarization
states that are not in the chosen A regions (i.e., the regions in B that are
not a displacement of the A regions by $\pi/2$ in $\kappa_{x}x$, $\kappa_{y}y$
and $\kappa_{z}z$). Since $\Sigma_{S}$ does not depend on $z$, $n_{S}$ grows
linearly with $z$, and thus the total $n_{S}$ is $-1/2$ ($+1/2$) in the A (B)
sub-regions, so their Skyrme numbers always cancel. A simple example is
depicted in Fig. 10(a), where we took the second half of cell A. We can choose
the A volumes to be split into two parts as shown in Fig. 10(b). Following the
stacking rules to reproduce the field in subsection 3.1, one realizes that the
cell in Fig. 7(c) ($\kappa_{x}x\in[0,\pi]$, $\kappa_{y}y\in[0,2\pi]$ and
$\kappa_{z}z\in[0,\pi/2]$), which is also shown in Fig. 10(c), can be built
from Fig. 10(b), but now all the parts of the cell are connected forming a
cuboidal F3DP nonskyrmionic cell.
Figure 10: Nonskyrmionic F3DP regions. The unit cell of the polarization
structure of the field has been depicted as two dashed cuboids to better
visualize the nonskyrmionic F3DP regions.
### A.3 Videos
We show the evolution of the L-lines, C-lines and normalized spin density
vector $\mathbf{s}$ inside the F3DP cell A for $\alpha=\beta=\pi/4$ (video 1)
and for $\alpha=\tilde{\alpha}$ and $\beta=\tilde{\beta}$ (video 2). The
evolution of the electric field orientation at the regions of linear
polarization and of the $\mathbf{s}$ vector at the C-lines over the unit
sphere is included.
|
# Faster Inference of Integer SWIN Transformer by Removing the GELU Activation
Mohammadreza Tayaranian111Correspondence to
<EMAIL_ADDRESS>Seyyed Hasan Mozafari,
James J. Clark, Brett Meyer, Warren Gross
###### Abstract
SWIN transformer is a prominent vision transformer model that has state-of-
the-art accuracy in image classification tasks. Despite this success, its
unique architecture causes slower inference compared with similar deep neural
networks. Integer quantization of the model is one of the methods used to
improve its inference latency. However, state-of-the-art has not been able to
fully quantize the model. In this work, we improve upon the inference latency
of the state-of-the-art methods by removing the floating-point operations,
which are associated with the GELU activation in Swin Transformer. While
previous work proposed to replace the non-integer operations with linear
approximation functions, we propose to replace GELU with ReLU activation. The
advantage of ReLU over previous methods is its low memory and computation
complexity. We use iterative knowledge distillation to compensate for the lost
accuracy due to replacing GELU with ReLU. We quantize our GELU-less SWIN
transformer and show that on an RTX 4090 NVIDIA GPU we can improve the
inference latency of the quantized SWIN transformer by at least $11\%$ while
maintaining an accuracy drop of under $0.5\%$ on the ImageNet evaluation
dataset.
## Introduction
The attention mechanism has gained popularity in recent years after its
successful debut in transformer architecture (Vaswani et al. 2017). While the
transformer architecture has been initially used for natural language
processing (NLP) tasks, it was also brought to the computer vision domain with
the introduction of vision transformer models (Dosovitskiy et al. 2021). SWIN
transformer (Liu et al. 2021a) is a well-known vision transformer which
improves on the original design by using shifted windows in the input. It
shows state-of-the-art performance in a variety of computer vision tasks.
However, SWIN transformer’s inference latency is negatively affected due to
its use of windowed attention. The windowed attention relies on shifting of
the input activations, and the shift operation is highly memory intensive,
thus having a high impact on the inference latency. For instance, running
inference on an NVIDIA V100 GPU, SWINSMALL is shown to be $55\%$ slower
compared to ViTSMALL (Liu et al. 2022). For mobile devices, Wang et al. (2022)
demonstrated a more pronounced gap in the inference latency where SWINSMALL is
$2.2$ times slower than ViTSMALL.
Layer-normWindowShifting$Q$Fused Op #1QKVGemm$dQ$QKVBias$Q$Fused Op #2$Q\cdot
K$Gemm$dQ$Softmax &Pos. Bias$Q$Fused Op
#3$V\cdot\textrm{att}$GemmProjGemm$dQ$ProjBiasAdd(Residual)$Q$Fused Op
#4FC1Gemm$dQ$FC1 Bias & GELU activation$Q$Fused Op #5FC2Gemm$dQ$FC2 BiasAdd
&Layer-normFused Op #6 Figure 1: High-level depiction of the components of a
transformer block in the quantized SWIN-transformer. $Q$ and $dQ$ denote the
quantization and de-quantization operations, respectively.
Quantization is one of the techniques used for the improvement of the
inference latency of deep neural networks. It involves representing the values
of the neural network using data types with lower bit-widths.
Despite the theoretical possibility of using arbitrary data types and bit-
widths for quantization, the achieved inference speedup depends on hardware on
which the model is being deployed (Sun et al. 2022). For instance, consider a
quantization method which uses the 4-bit integer data type to represent
weights and activations of a deep neural network. Running this method on
hardware that doesn’t support 4-bit arithmetic operations results in a speedup
lower than the expected theoretical speedup. Besides, the overhead of
converting the quantized values to a data type which is supported by the GPU
will further reduce the speedup of the method.
Another important factor in the speedup of partially quantization models is
their non-integer components (Li and Gu 2022). Such component is the non-
linear function, e.g. Softmax, that due to its non-linearity is not easily
quantizable. Given a non-linear function $f$, its quantized input $\hat{x}$,
and the quantization scale $s$, we have $f(s\hat{x})\neq sf(\hat{x})$. As a
result, some integer implementations opt to use the floating-point data type
for these components. This enforces the inclusion of memory intensive
quantization and de-quantization functions in the inference pipeline which
results in notable overhead.
With the goal of avoiding non-integer components, previous work focused on
quantizing the non-linear operations of transformer-based models (Li and Gu
2022; Kim et al. 2021; Lin et al. 2021). The main theme of these works is to
substitute the non-linear components with a linear or piece-wise linear
version without losing accuracy. In this work, we propose to replace the GELU
activation with the piece-wise linear ReLU function (Fukushima 1975). Compared
with which needs to compute the maximum of the input tensor to approximate
GELU (Li and Gu 2022), ReLU can be simply applied with the help of a
comparator. The advantage of ReLU is its low complexity and simple logic
whereas previous work’s shift-based GELU required (Li and Gu 2022). We apply
these changes to the SWIN transformer model in a layer-by-layer fashion and
use knowledge distillation in the process to maintain the model’s accuracy.
The weights and input activations of the resulting model, which we call GELU-
less SWIN, are then quantized using post-training quantization.
The results of this comparison show that our model has a maximum accuracy drop
of $0.5\%$ while achieving more than $11\%$ inference latency reduction
compared to the FasterTransformer framework.
## Previous Work
### Quantization of Linear Components
The majority of previous work focuses on the quantization of linear components
of vision transformers, i.e. fully connected and convolutional layers. In the
quantized version of these components, either the weight, the input
activation, or both matrices are quantized. The quantization scale is obtained
by either quantization aware training or via a calibration phase in a post-
training quantization fashion.
Liu et al. (2021b) propose a mixed-precision quantization in which each linear
layer in different transformer blocks has a different bit-width. Yuan et al.
(2021) uses a Hessian-guided similarity measure for finding quantization
scales. To keep the precision in the layers that are more sensitive to the
quantization noise, they use two scale factors for each fully connected layer
with each of the scales responsible for only a part of the tensor. Li et al.
(2022a) quantizes the linear layers of vision transformers down to 2 bits.
They do so by adding trainable parameters that help the quantized weights
follow the distribution of floating-point weights. Li et al. (2022b) uses
different bit-widths and scales for each attention head. Li et al. (2022a) and
Li et al. (2022b) both show promising results in terms of accuracy when using
4-bit and 3-bit weights. Despite being able to maintain the model’s test
accuracy, all of these works lack studies of hardware performance metrics,
e.g. latency, of their quantized model and only discuss the model size, which
is not a reliable proxy for latency. In the present work, in addition to
accuracy, we measure the inference latency of our quantization method and its
speedup compared to the baseline.
### Quantization of Non-Linear Components
Another line of work tries to quantize the non-linear components of the model.
Softmax, LayerNorm, and GELU activation are the three main non-linear
components of vision transformers that are not straightforward to quantize.
Lin et al. (2020) and Kim et al. (2021) use polynomial approximations to
provide quantizable versions of the non-linear components. Although their
methods were proven successful for transformer-based language models, Li and
Gu (2022) have shown that these methods cannot be used to vision transformers.
Lin et al. (2021) propose a $\log_{2}$ quantization method which adapts the
methodology of (Kim et al. 2021) for fully integer vision transformers.
Although the authors show their method’s ability to maintain accuracy, the
performance of their method in terms of hardware metrics like latency is not
discussed.
The closest work to our work is I-ViT, which provides shift-based replacements
for the non-linear components (Li and Gu 2022). Their integer-friendly
replacement functions use the power of two approximations of the $e^{x}$
function (Stevens et al. 2021). They also improve on the integer LayerNorm
proposed by Lin et al. (2020) by using a shift-based iterative function to
compute the square root of the variance. Their experimental results show that
their quantized model has a $5.8\%$ shorter inference latency compared with
implementing quantized SWIN transformer based on NVIDIA’s FasterTransformer
framework. Despite its high accuracy, their proposed shift-based GELU
approximation has a high memory and computation complexity as it needs to
compute the maximum value of the input tensor to approximate GELU. In
comparison, our proposed method of replacing the GELU activation with ReLU has
the advantage of lower memory and computation complexity given the ReLU’s
simple logic.
## Background and Motivation
SWIN transformer addresses an architectural problem with the original vision
transformer model. It uses a windowed attention mechanism to avoid global
attention and its considerable computation. The window attention divides each
input activation into smaller windows and computes the attention over the
image patches in each window. This enables the use of SWIN transformer as a
backbone model for tasks such as semantic segmentation that have larger input
images (Liu et al. 2021a).
Despite its state-of-the-art accuracy, SWIN transformer’s use of window
shifting operations has negative effects on its hardware performance. This
negative effect is revealed when comparing the inference latency of SWIN
transformer with the original vision transformer. For instance, SWINSMALL is
$55\%$ slower in terms of inference latency when compared to DeiTSMALL which
has the same architecture as the original vision transformer (Liu et al.
2022). In the case of mobile GPU, Mehta and Rastegari (2022) demonstrated that
the window shifting operations are not supported by iPhone GPUs, making it
impossible to implement SWIN on this hardware. Nevertheless, on mobile GPUs
where SWIN can be implemented, SWINSMALL is $2.2$ times slower than DeiTSMALL
(Wang et al. 2022).
| Fused Op #
---|---
Model | 1 | 2 | 3 | 4 | 5 | 6
SWINTINY | $0.32$ | $1.51$ | $2.55$ | $0.88$ | $2.03$ | $1.16$
SWINSMALL | $0.3$ | $2.01$ | $4.1$ | $1.17$ | $2.48$ | $1.36$
SWINBASE | $0.57$ | $2.53$ | $5.44$ | $1.57$ | $3.89$ | $2.69$
SWINLARGE | $1.22$ | $3.93$ | $8.18$ | $2.92$ | $8.13$ | $3.66$
Average | $0.6$ | $2.5$ | $5.07$ | $1.64$ | $4.13$ | $2.22$
Table 1: Latency (ms) of the fused operations of the quantized SWIN
transformer, depicted in Figure 1. The latency value are measured on an NVIDIA
RTX 4090 and are averaged over 1000 inference runs.
Inspired by the gap in the inference latency, we use integer quantization to
speed up the inference of the SWIN transformer model. Since our target
hardware is the NVIDIA GPU, we start with the FasterTransformer framework’s
proposed quantized SWIN. Figure 1 depicts a high-level overview of the
components of this quantized SWIN transformer. $Q$ and $dQ$ are the
quantization and de-quantization functions that are used to convert between
the integer and floating-point data types. The quantized SWIN uses 8-bit
integer for the weights and input activations of the linear layers. It also
uses GPU’s integer tensor cores to accelerate the integer matrix
multiplication operations.
In this quantized SWIN architecture, components like biases, Softmax, layer-
norms, and the residual connections use the floating-point data type. The
quantized SWIN uses fused operations, shown in Figure 1 with dashed
rectangles, to spread the overhead of quantization and de-quantization of the
integer values over multiple functions. The functions inside each fused
operation use the shared memory of the GPU to pass values between each other.
This design minimizes the accesses to the slow global memory and thus keeps
the latency of the fused operation, and the entire model, at a minimum.
## GELU-less SWIN Transformer
The quantized SWIN transformer is depicted in Figure 1. The components that
are using the 8-bit integer GEMM, shown in light green, are already quantized
to the minimum bit-width supported by the GPU. Thus, we turn our attention to
the fused operations and measure their latencies.
The latency of each fused operation, calculated as the drop in inference
latency resulted from removing it, is provided in Table 1. Based on these
results, Softmax and GELU activation are the two non-integer components that
have the highest latency in the quantized SWIN transformer pipeline.
Furthermore, our experiments reveal that the latency of all the fused
operations are dominated by the global memory accesses of the required
quantization and de-quantization functions in a not fully quantized
implementation. As a result, instead of modifying only parts of a fused
operation, we need to remove it entirely to achieve higher inference speedup.
Considering these observations, we propose to remove the fused operation
associated with the GELU activation and substitute it with an integer
activation function. We propose to replace GELU with the piece-wise linear
ReLU activation. While the Shift-based GELU proposed by Li and Gu (2022) needs
to compute the maximum value of the input tensor, the ReLU function is a
simple activation function and thus has a lower memory complexity than its
Shift-based alternative. In order to completely remove the fused operation of
GELU, we also need to eliminate the FC1 Bias as well. We avoid changing the
Softmax fused operation as it also contains the relative position bias which
is essential to converting input image to tokens.
Algorithm 1 Our proposed GELU replacement method with knowledge distillation.
Input: SWIN, dataset
Parameter: $N$: Number of Transformer Blocks
Output: GELU-less SWIN
1: $\textrm{student}\leftarrow\operatorname{clone}(\textrm{SWIN})$
2: for $i\leftarrow 1$ to $N$ do
3: $\textrm{block}\leftarrow\textrm{student}.\textrm{blocks}[i]$
4: $\textrm{block}.\textrm{activation}\leftarrow\operatorname{ReLU}$
5: $\textrm{block}.\textrm{bias}\leftarrow 0$
6: $\operatorname{disable\\_gradient}(\textrm{block}.\textrm{bias})$
7: $\operatorname{kd\\_epoch}(\textrm{student},\textrm{SWIN})$
8: end for
9: $\textrm{GELU-less SWIN}\leftarrow\textrm{student}$
10: return GELU-less SWIN
Model | Method | Datatype | Top-1 Acc. | Latency | Speedup
---|---|---|---|---|---
(%) | (ms)
SWINTINY | Baseline | FP32 | $81.2$ | $60.27$ | $\times 1$
Half-precision | FP16 | $81.2$ | $24.96$ | $\times 2.41$
FasterTransformer | int8 | $80.1$ | $17.04$ | $\times 3.54$
Ours | int8 | $80.0$ | $15.01$ | $\times\mathbf{4.02}$
SWINSMALL | Baseline | FP32 | $83.2$ | $103.21$ | $\times 1$
Half-precision | FP16 | $83.2$ | $40.26$ | $\times 2.56$
FasterTransformer | int8 | $83.0$ | $25.05$ | $\times 4.12$
Ours | int8 | $82.5$ | $22.57$ | $\times\mathbf{4.57}$
SWINBASE | Baseline | FP32 | $83.4$ | $157.31$ | $\times 1$
Half-precision | FP16 | $83.4$ | $58.39$ | $\times 2.69$
FasterTransformer | int8 | $83.3$ | $35.55$ | $\times 4.43$
Ours | int8 | $84.6$ | $31.66$ | $\times\mathbf{4.97}$
SWINLARGE | Baseline | FP32 | $86.2$ | $284.41$ | $\times 1$
Half-precision | FP16 | $86.2$ | $104.76$ | $\times 2.71$
FasterTransformer | int8 | $85.8$ | $61.35$ | $\times 4.64$
Ours | int8 | $85.5$ | $53.22$ | $\times\mathbf{5.34}$
Table 2: Comparison of our proposed method with the floating-point baselines
and FasterTransformer’s quantizated model. The accuracy are from evaluating a
pre-trained SWIN transformer on the ImageNet evaluation dataset. The reported
inference latency is for a batch size of 128, averaged over 1000 runs.
Algorithm 1 describes our proposed method for replacing the GELU activation
with ReLU. To avoid risking model divergence, we gradually apply our changes
to the model and modify the transformer blocks one at a time. After each block
modification, we use knowledge distillation to distill the soft labels of the
fully GELU SWIN to our partially GELU and partially ReLU model. The result of
the algorithm, which we call the GELU-less SWIN, does not have a floating-
point fused operation for its activation function. As ReLU is easily
quantizable, it can be fused, as an integer operation, to the previous GEMM.
This way, ReLU’s latency will be completely masked as it will be directly
applied to the output of the GEMM. Finally, we apply the post-training
quantization method of the FasterTransformer framework to this GELU-less SWIN
transformer and quantize its weights and input activations.
## Experiments
### Experimental Setup
We study the evaluation accuracy and latency of our method by applying it to
variuos configurations of the SWIN transformer. We use a pre-trained SWIN
model which is pre-trained on the ImageNet training dataset and evaluate it on
the ImageNet evaluation dataset (Deng et al. 2009). As our methodology mainly
targets NVIDIA GPU hardware, we implement our method on top of NVIDIA’s
FasterTransformer222https://github.com/NVIDIA/FasterTransformer framework. We
compare the performance of our quantized SWIN with FasterTransformer’s
quantized SWIN and also the 32-bit and 16-bit floating-point models.
For knowledge distillation, we use the SGD optimizer with a constant learning
rate of $10^{-2}$ and a momentum of $0.9$. As described in Algorithm 1, the
number of knowledge distillation epochs is the same as the number of
transformer blocks of the SWIN model. So, SWINTINY undergoes 12 epochs of the
knowledge distillation while for other SWIN configurations this number is 24
epochs. We use $10\%$ of the ImageNet training dataset for our knowledge
distillation, with a batch size of $32$. The wall clock time for the knowledge
distillation is $74$, $203$, $213$, and $337$ minutes, respectively, for
models from SWINTINY to SWINLARGE.
For our proposed method, we apply the same post-training quantization that is
used in the FasterTransformer framework on our GELU-less SWIN. The inference
latency is an average of 1000 runs and is measured on a quantized model which
had its GELU fused operation removed. Our batch size for the evaluation
experiments is $128$. All our experiments are performed on an NVIDIA RTX 4090
GPU.
### Experimental Results
Table 2 provides the evaluation top-1 accuracy and latency of SWIN transformer
configurations using different methods. As the Table demonstrates, our
proposed GELU-less quantized SWIN has the smallest inference latency across
the SWIN configurations. Our method is able to improve the latency of the
FasterTransformer by $12\%$, $11\%$, $11\%$, and $13\%$, respectively, for
models from SWINTINY to SWINLARGE.
As an ablation study, we also apply our method without the use of knowledge
distillation which results in an evaluation accuracy of less than $0.9\%$
which shows the knowledge distillation is essential to the proposed algorithm.
## Conclusion
In this work we proposed a method to reduce the inference latency of int-8
quantized SWIN transformer model. We analyzed the latency of the operations in
the existing int-8 quantized SWIN piepline. Based on the analysis, we proposed
to replace the floating-point GELU activation with the ReLU activation. ReLU
is a piece-wise linear function which is easily quantizable and has a very low
complexity. Our proposed method replaces GELU with ReLU and removes the bias
that is fused to it. We also use knowledge distillation to maintain the
accuracy. Our experiments show that quantizing our proposed GELU-less SWIN
results in at least $11\%$ reduction of inference latency compared to the
original quantized SWIN transformer model.
## References
* Deng et al. (2009) Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei-Fei, L. 2009. Imagenet: A large-scale hierarchical image database. In _2009 IEEE conference on computer vision and pattern recognition_ , 248–255. Ieee.
* Dosovitskiy et al. (2021) Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; Uszkoreit, J.; and Houlsby, N. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In _International Conference on Learning Representations_.
* Fukushima (1975) Fukushima, K. 1975. Cognitron: A self-organizing multilayered neural network. _Biological cybernetics_ , 20(3-4): 121–136.
* Kim et al. (2021) Kim, S.; Gholami, A.; Yao, Z.; Mahoney, M. W.; and Keutzer, K. 2021. I-BERT: Integer-only BERT Quantization. In Meila, M.; and Zhang, T., eds., _Proceedings of the 38th International Conference on Machine Learning_ , volume 139 of _Proceedings of Machine Learning Research_ , 5506–5518. PMLR.
* Li et al. (2022a) Li, Y.; Xu, S.; Zhang, B.; Cao, X.; Gao, P.; and Guo, G. 2022a. Q-ViT: Accurate and Fully Quantized Low-bit Vision Transformer. _arXiv preprint arXiv:2210.06707_.
* Li and Gu (2022) Li, Z.; and Gu, Q. 2022. I-ViT: integer-only quantization for efficient vision transformer inference. _arXiv preprint arXiv:2207.01405_.
* Li et al. (2022b) Li, Z.; Yang, T.; Wang, P.; and Cheng, J. 2022b. Q-vit: Fully differentiable quantization for vision transformer. _arXiv preprint arXiv:2201.07703_.
* Lin et al. (2020) Lin, Y.; Li, Y.; Liu, T.; Xiao, T.; Liu, T.; and Zhu, J. 2020. Towards fully 8-bit integer inference for the transformer model. _arXiv preprint arXiv:2009.08034_.
* Lin et al. (2021) Lin, Y.; Zhang, T.; Sun, P.; Li, Z.; and Zhou, S. 2021. Fq-vit: Post-training quantization for fully quantized vision transformer. _arXiv preprint arXiv:2111.13824_.
* Liu et al. (2021a) Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; and Guo, B. 2021a. Swin transformer: Hierarchical vision transformer using shifted windows. In _Proceedings of the IEEE/CVF international conference on computer vision_ , 10012–10022.
* Liu et al. (2022) Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; and Xie, S. 2022. A convnet for the 2020s. In _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_ , 11976–11986.
* Liu et al. (2021b) Liu, Z.; Wang, Y.; Han, K.; Zhang, W.; Ma, S.; and Gao, W. 2021b. Post-training quantization for vision transformer. _Advances in Neural Information Processing Systems_ , 34: 28092–28103.
* Mehta and Rastegari (2022) Mehta, S.; and Rastegari, M. 2022. MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer. In _International Conference on Learning Representations_.
* Stevens et al. (2021) Stevens, J. R.; Venkatesan, R.; Dai, S.; Khailany, B.; and Raghunathan, A. 2021. Softermax: Hardware/software co-design of an efficient softmax for transformers. In _2021 58th ACM/IEEE Design Automation Conference (DAC)_ , 469–474. IEEE.
* Sun et al. (2022) Sun, M.; Ma, H.; Kang, G.; Jiang, Y.; Chen, T.; Ma, X.; Wang, Z.; and Wang, Y. 2022. VAQF: fully automatic software-hardware co-design framework for low-bit vision transformer. _arXiv preprint arXiv:2201.06618_.
* Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. _Advances in neural information processing systems_ , 30.
* Wang et al. (2022) Wang, X.; Zhang, L. L.; Wang, Y.; and Yang, M. 2022. Towards efficient vision transformer inference: A first study of transformers on mobile devices. In _Proceedings of the 23rd Annual International Workshop on Mobile Computing Systems and Applications_ , 1–7.
* Yuan et al. (2021) Yuan, Z.; Xue, C.; Chen, Y.; Wu, Q.; and Sun, G. 2021. Ptq4vit: Post-training quantization framework for vision transformers. _arXiv preprint arXiv:2111.12293_.
|
# Parametric Shape Holomorphy of Boundary Integral Operators with Applications
Jürgen Dölz Institute for Numerical Simulation, University of Bonn, Friedrich-
Hirzebruch-Allee 7, 53115 Bonn, Germany<EMAIL_ADDRESS>Fernando
Henríquez Chair of Computational Mathematics and Simulation Science, École
Polytechnique Fédérale de Lausanne, 1015 Lausanne, Switzerland ()
<EMAIL_ADDRESS>
###### Abstract
We consider a family of boundary integral operators supported on a collection
of parametrically defined bounded Lipschitz boundaries. Thus, the boundary
integral operators themselves also depend on the parametric variables, leading
to a parameter-to-operator map. One of the main results of this article is to
establish the analytic or holomorphic dependence of said boundary integral
operators upon the parametric variables, i.e., of the parameter-to-operator
map. As a direct consequence we also establish holomorphic dependence of
solutions to boundary integral equations, i.e., holomorphy of the parameter-
to-solution map. To this end, we construct a holomorphic extension to complex-
valued boundary deformations and investigate the _complex_ Fréchet
differentiability of boundary integral operators with respect to each
parametric variable. The established parametric holomorphy results have been
identified as a key property to derive best $N$-term approximation rates to
overcome the so-called curse of dimensionality in the approximation of
parametric maps with distributed, high-dimensional inputs.
To demonstrate the applicability of the derived results, we consider as a
concrete example the sound-soft Helmholtz acoustic scattering problem and its
frequency-robust boundary integral formulations. For this particular
application, we explore the consequences of our results in reduced order
modelling, Bayesian shape inversion, and the construction of efficient
surrogates using artificial neural networks.
## 1 Introduction
In this article, we focus on shape-parametric boundary integral operators
(BIO) of the form
(1.1) $\mathsf{A}_{\bm{{y}}}\colon L^{2}(\Gamma_{\bm{{y}}})\to
L^{2}(\Gamma_{\bm{{y}}}),$
where $\Gamma_{\bm{{y}}}\subset\mathbb{R}^{3}$ is assumed to be a bounded
Lipschitz boundary depending on the parametric input
${\bm{{y}}}\in\mathbb{U}\coloneqq[-1,1]^{\mathbb{N}}$. The boundary integral
operator in Eq. 1.1 is assumed to be of the form
(1.2) $(\mathsf{A}_{\bm{{y}}}\varphi)({\bf
x})\mathrel{\mathrel{\mathop{:}}=}\int\limits_{\Gamma_{\bm{{y}}}}{\sf a}({\bf
y},{\bf x}-{\bf y})\varphi({\bf y})\mathrm{ds}_{\bf y},\quad{\bf
x}\in\Gamma_{\bm{{y}}},$
for any $\varphi\in L^{2}(\Gamma_{\bm{{y}}})$, where the kernel function ${\sf
a}:\Gamma_{\bm{{y}}}\times\mathbb{R}\backslash\\{{\bf
0}\\}\rightarrow\mathbb{C}$ is bounded in the first argument, and possibly
_singular_ when ${\bf x}={\bf y}$ in the second one. A typical example
application is the $L^{2}(\Gamma_{\bm{{y}}})$-based frequency-stable
formulation for the acoustic scattering by random surfaces, wherein the
boundary’s randomness is encoded through a Karhunen-Loève-type expansion of a
given random field. This is, for example, the setting considered in [21, 5,
23, 4].
In the following, we are interested in the regularity of the _parameter-to-
operator_ map
(1.3)
$\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{A}_{\bm{{y}}}\in\mathscr{L}\left(L^{2}(\Gamma_{\bm{{y}}}),L^{2}(\Gamma_{\bm{{y}}})\right)$
and, assuming $\mathsf{A}_{\bm{{y}}}$ to be boundedly invertible for each
${\bm{{y}}}\in\mathbb{U}$, of the _parameter-to-solution_ map
(1.4) $\mathbb{U}\ni{\bm{{y}}}\mapsto
u_{\bm{{y}}}\mathrel{\mathrel{\mathop{:}}=}\mathsf{A}_{\bm{{y}}}^{-1}f_{\bm{{y}}}\in
L^{2}(\Gamma_{\bm{{y}}}),$
where $f_{\bm{{y}}}\in L^{2}(\Gamma_{\bm{{y}}})$ for each
${\bm{{y}}}\in\mathbb{U}$. For these maps, constructing efficient surrogates
is a central task in computational UQ and reduced order modelling of
parametric partial differential and boundary integral equations. However, it
is well-known that the so-called curse of dimensionality in the parameter
space renders most naïve approximation methods computationally unfeasible. To
overcome this challenge and effectively design dimension-robust algorithms for
the approximation of the maps in Eq. 1.3 and Eq. 1.4 one requires a better
understanding of the _parametric regularity_ of the BIOs upon the shape-
parametric representations.
In [10], it is proposed to construct a holomorphic extension over a tensor
product of ellipses in the complex domain as a key step to finally break the
curse of dimensionality in the sparse polynomial approximation of high-
dimensional parametric problems. Indeed, by varying the size of these ellipses
on each parameter, i.e. by exploiting the _anisotropic_ parameter dependence,
this _$(\bm{b},p,\varepsilon)$ -holomorphy_ allows to state provable
convergence rates that do not depend on the input’s parametric dimension. From
a computational point of view, $(\bm{b},p,\varepsilon)$-holomorphy constitutes
a stepping stone in the analysis and implementation of a plethora of
approximation techniques such as sparse grid interpolation and quadrature [56,
49, 41, 24], higher-order Quasi-Monte Carlo [19, 20, 17, 18] integration
(HoQMC), construction of artificial neural network (ANNs) surrogates [53, 30,
42, 29, 1], and to establish sparsity of parametric posterior densities in
Bayesian inverse problems [9, 49, 51].
The main objective of this work is to derive shape-parametric regularity
results for BIOs of the form Eq. 1.1. In addition, we aim to demonstrate the
usefulness of this property by studying its implications in reduced order
modelling, expression rates of artificial neural network (ANN) surrogates, and
Bayesian shape inversion.
### 1.1 Previous Work
The analytic dependence of the parameter-to-operator and parameter-to-solution
map is the primary object of study in many recent works tackling parametric
shape deformations. In the context of volume-based variational formulations,
this property has been established for subsurface flows [6, 25, 26], time-
harmonic electromagnetic wave scattering by perfectly conducting and
dielectric obstacles [35, 3], stationary Stokes and Navier-Stokes equation
[11], volume formulations for acoustic wave scattering by a penetrable
obstacle in a low-frequency regime [33], and high-frequency Helmholtz
formulations [54, 22].
As pointed out previously, to establish the sought parametric holomorphy
property one verifies _complex_ Fréchet differentiability of the parameter-to-
operator or parameter-to-solution map. This approach is strongly connected to
the notion of shape differentiability of the domain/boundary-to-solution map.
In [44, 46] and [8], the shape differentiability in the Fréchet sense of BIOs
for acoustic and elastic problems is studied. This result is extended in [45,
13, 14] to the BIOs that arise in electromagnetic wave scattering. The same
approach used to obtain these results is also employed in [13, 14, 55] to
establish the infinite differentiability of a collection of BIOs characterized
by a class of pseudo-homogeneous kernels. However, regardless of the specific
setting considered on the previous works, none of the imply an analytic or
holomorphic dependance upon shape perturbations. Another drawback of the
previously discussed results is that none of them are capable of
systematically account for polygons, let alone Lipschitz boundaries.
Recently, in [28, 27] the holomorphic dependence of the Calderón projector
upon boundaries of class $\mathscr{C}^{2}$ has been established. Therein, the
two-dimensional analysis for both the Laplace and Helmholtz BIOs is thoroghly
analyzed. In [43] this work has been extended to address the case of multiple
open arcs in two-dimensions. In [16], based on [36, 37], a analytic shape
differentiability result is established, and this approach is extended in [15]
to the layer potentials for the heat equation. Yet useful in the three
dimensional case, these works still rule out the possibility of polygonal
boundaries in the analysis.
### 1.2 Contributions
We establish parametric holomorphy by verifying _complex_ Fréchet
differentiability of suitable complex extensions of the parameter-to-operator
map Eq. 1.3 and paramter-to-solution map Eq. 1.4. These maps are viewed as
elements of the complex Banach space of bounded linear operators or functions,
with respect to a collection of an affinely parametrized bounded Lipschitz
boundaries.
The contribution of this article is three-fold:
1. 1.
By considering a collection of affine shape-parametric boundaries and suitable
assumptions on the integral kernel $\mathsf{a}$, we prove that the parameter-
to-operator map Eq. 1.3 is $(\bm{b},p,\varepsilon)$-holomorphic in the sense
of [10, Definition 2.1] in a $L^{2}$-based setting. We conclude that the
parameter-to-solution map Eq. 1.4 is also
$(\bm{b},p,\varepsilon)$-holomorphic. Even though we do not explore the entire
Calderón projector in a scale of fractional Sobolev spaces as in [28, 16], we
let the boundary to be globally Lipschitz, yet possibly be patchwise more
regular. This allows the possibility of polygonal domain in the analysis.
2. 2.
We demonstrate the practical applicability of our
$(\bm{b},p,\varepsilon)$-holomorphy results by exemplarily applying our result
to frequency-stable integral equations used in the boundary reduction of
acoustic wave propagation in unbounded domains. This, in turn, serves a
theoretical foundation for the analysis of the computational technique
introduced in [21].
3. 3.
We demonstrate the theoretical usability of our
$(\bm{b},p,\varepsilon)$-holomorphy results by using results from the
literature to obtain _best $n$-term_ approximation rates for Eq. 1.3 and Eq.
1.4. In addition, we explore the consequence of our result in reduced order
modelling, Bayesian shape inversion, and in the construction of efficient
surrogates using ANNs.
### 1.3 Outline
The remainder of the article is structured as follows. First, in Section 2, we
recall some preliminary facts on holomorphy in Banach spaces, parametrized
surfaces, and boundary integral operators. $(\bm{b},p,\varepsilon)$-holomorphy
of the boundary integral operators is established in Section 3, whereas the
illustration to boundary integral equations from acoustic scattering is done
in Section 4. Section 5 demonstrates how the established
$(\bm{b},p,\varepsilon)$-holomorphy leads to various _best- $n$-term_
approximation rates and is followed by concluding remarks in Section 6.
## 2 Preliminaries
Let ${\text{D}}\subset\mathbb{R}^{2}$ be a bounded Lispchitz domain [34,
Definition 3.3.1]. We denote by $L^{p}({\text{D}})$, $p\in[1,\infty)$ the set
of $p$-integrable functions over D, by $H^{s}({\text{D}})$, $s\geq 0$,
stantard Sobolev spaces in D. Let
$\mathscr{C}^{k,\lambda}(\partial{\text{D}})$ be the space of $k$ times
continuously differentiable functions in $\partial{\text{D}}$ with Hölder
continuous partial derivatives of order $\lambda$, and, in particular, by
$\mathscr{C}^{0,1}(\partial{\text{D}})$ the space of Lipschitz continuous
functions on $\partial{\text{D}}$. In addition, let
$H^{s}(\partial{\text{D}})$, for $s\in[0,1]$, be the Sobolev space of traces
on $\partial{\text{D}}$ ([48, Section 2.4], [34, Section 4.2]). As is
customary, we identify $H^{0}(\partial{\text{D}})$ with
$L^{2}({\partial{\text{D}}})$ and, for $s\in[0,1]$,
$H^{-s}(\partial{\text{D}})$ with the dual space of
$H^{s}(\partial{\text{D}})$.
For (complex) Banach spaces $X$ and $Y$, we denote by $\mathscr{L}(X,Y)$ the
space of bounded linear operators from $X$ into $Y$ and by
$\mathscr{L}_{\text{iso}}(X,Y)$ the (open) subset of isomorphisms, i.e.
bounded linear operators with a bounded inverse. Recall that
$\mathscr{L}(X,Y)$ is a (complex) Banach space equipped with the standard
operator norm [47, Theorem III.2].
Finally, we equipp $\mathbb{U}$ with the product topology, and, according to
Tychonoff’s theorem, this renders $\mathbb{U}$ compact with this topology.
### 2.1 Holomorphy in Banach Spaces
For the convenience of the reader, we recall the basic facts of holmorphy in
Banach spaces. To this end, let $E$ and $F$ be _complex_ Banach spaces
equipped with the norms $\left\lVert\cdot\right\rVert_{E}$ and
$\left\lVert\cdot\right\rVert_{F}$, respectively.
###### Definition 2.1 ([39, Definition 13.1]).
Let $U$ be an open, nonempty subset of $E$. A map $\mathcal{M}:U\rightarrow F$
is said to be _complex Fréchet differentiable_ if for each $r\in U$ there
exists a map $(\frac{d}{dr}\mathcal{M})(r,\cdot)\in\mathscr{L}(E,F)$ such that
(2.1)
$\displaystyle\left\lVert{\mathcal{M}(r+\xi)}-\mathcal{M}(r)-\left(\frac{d}{dr}\mathcal{M}\right)(r,\xi)\right\rVert_{F}=o\left(\left\lVert\xi\right\rVert_{E}\right).$
We say that $\left(\frac{d}{dr}\mathcal{M}\right)(r,\xi)$ is the _Fréchet
derivative_ of the map $\mathcal{M}:U\rightarrow F$ at $r\in U$ in the
direction $\xi\in E$.
The following results are extensions of some classical results from complex
analysis to Banach spaces.
###### Theorem 2.2 ([39, Theorem 14.7]).
Let $U$ be an open, nonempty subset of $E$. For the map
$\mathcal{M}:U\rightarrow F$ the following conditions are equivalent:
1. 1.
$\mathcal{M}$ is complex Fréchet differentiable.
2. 2.
$\mathcal{M}$ is infinitely complex Fréchet differentiable.
3. 3.
$\mathcal{M}$ is holomorphic.
###### Remark 2.3.
In view of Theorem 2.2 it sufficies to show _complex_ Fréchet
differentiability to establish an holomorphic dependence.
###### Lemma 2.4 ([31, Theorem 3.1.5, item c)]).
Let $U$ be an open, nonempty subset of $E$, and let
$\left(\mathcal{M}_{n}\right)_{n\in\mathbb{N}}$ be a sequence of holomorphic
maps from $U\subset E$ to $F$. Assume that $\mathcal{M}_{n}$ converges
uniformly to $\mathcal{M}$ in $U$. Then $\mathcal{M}:U\subset E\rightarrow F$
is holomorphic.
###### Proposition 2.5 ([28, Proposition 4.20]).
Let $X$, $Y$ be complex Banach spaces and denote by
$\mathscr{L}_{\text{iso}}(X,Y)$ the (open) subset of operators with bounded
inverse in $\mathscr{L}(X,Y)$.
1. 1.
Let $\mathsf{M}\in\mathscr{L}_{\text{iso}}(X,Y)$. Then with
(2.2)
$\displaystyle\mathcal{C}_{\mathsf{M}}\coloneqq\left\\{\mathsf{T}\in\mathscr{L}(X,Y):\;\left\lVert\mathsf{M}-\mathsf{T}\right\rVert_{\mathscr{L}(X,Y)}<\left\lVert\mathsf{M}^{-1}\right\rVert^{-1}_{\mathscr{L}(Y,X)}\right\\}$
$\mathcal{C}_{\mathsf{M}}\subset\mathscr{L}_{\text{iso}}(X,Y)$ and for all
$\mathsf{T}\in\mathcal{C}_{\mathsf{M}}$ it holds
(2.3)
$\displaystyle\left\lVert\mathsf{T}^{-1}\right\rVert_{\mathscr{L}(Y,X)}\leq\frac{\left\lVert\mathsf{M}^{-1}\right\rVert_{\mathscr{L}(Y,X)}}{1-\left\lVert\mathsf{M}-\mathsf{T}\right\rVert_{\mathscr{L}(X,Y)}\left\lVert\mathsf{M}^{-1}\right\rVert_{\mathscr{L}(Y,X)}}.$
2. 2.
The inversion map
(2.4)
$\displaystyle\text{{inv}}:\;\mathscr{L}_{\text{{iso}}}(X,Y)\mapsto\mathscr{L}_{\text{{iso}}}(Y,X):\mathsf{M}\mapsto\mathsf{M}^{-1}$
is holomorphic.
3. 3.
The application map
(2.5) $\displaystyle\text{{app}}:\;\left(\mathscr{L}(X,Y),X\right)\rightarrow
Y:\;(\mathsf{M},\mathsf{g})\mapsto\mathsf{M}\,\mathsf{g}$
is holomorphic.
### 2.2 Parametric Holomorphy
We proceed to define precisely the notion parametric holomorphy to be used
throughout this work. To this end, for $s>1$ we define the so-called Bernstein
ellipse as
(2.6)
$\displaystyle\mathcal{E}_{s}\coloneqq\left\\{\frac{z+z^{-1}}{2}:\;1\leq\left\lvert
z\right\rvert\leq s\right\\}\subset\mathbb{C}.$
This ellipse has foci at $z=\pm 1$ and semi-axes of length
$a\coloneqq(s+s^{-1})/2$ and $b\coloneqq(s-s^{-1})/2$. In addition, we define
the tensorized poly-ellipse
(2.7) $\displaystyle\mathcal{E}_{\bm{\rho}}\coloneqq\bigotimes_{j\geq
1}\mathcal{E}_{\rho_{j}}\subset\mathbb{C}^{\mathbb{N}},$
where $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ is such that $\rho_{j}>1$, for
$j\in\mathbb{N}$. We adopt the convention $\mathcal{E}_{1}\coloneqq[-1,1]$ to
include the limit case $\rho_{j}=1$.
The next definition, originally introduced in [10], constitutes the precise
the definition of parametric holomorphy. It has been considered in a variety
of previous works encompassing the approximation of high-dimensional maps, and
it is, consequently, the one to be used in this work.
###### Definition 2.6 ([10, Definition 2.1]).
Let $X$ be a complex Banach space equipped with the norm
$\left\lVert\cdot\right\rVert_{X}$. For $\varepsilon>0$ and $p\in(0,1)$, we
say that map
$\mathbb{U}\ni{\bm{{y}}}\mapsto u_{\bm{{y}}}\in X$
is $(\bm{b},p,\varepsilon)$-holomorphic if and only if
1. 1.
The map $\mathbb{U}\ni{{\bm{{y}}}}\mapsto u_{\bm{{y}}}\in X$ is uniformly
bounded, i.e.
(2.8) $\displaystyle\sup_{{\bm{{y}}}\in\mathbb{U}}\left\lVert
u_{\bm{{y}}}\right\rVert_{X}\leq C_{0},$
for some finite constant $C_{0}>0$.
2. 2.
There exists a positive sequence $\bm{b}\coloneqq(b_{j})_{j\geq
1}\in\ell^{p}(\mathbb{N})$ and a constant $C_{\varepsilon}>0$ such that for
any sequence $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ of numbers strictly
larger than one that is $(\bm{b},\varepsilon)$-admissible, i.e. satisyfing
(2.9) $\displaystyle\sum_{j\geq 1}(\rho_{j}-1)b_{j}\leq\varepsilon,$
the map ${\bm{{y}}}\mapsto u_{{\bm{{y}}}}$ admits a complex extension
${\bm{z}}\mapsto u_{{\bm{z}}}$ that is holomorphic with respect to each
variable $z_{j}$ on a set of the form
(2.10)
$\displaystyle\mathcal{O}_{\bm{\rho}}\coloneqq\displaystyle{\bigotimes_{j\geq
1}}\,\mathcal{O}_{\rho_{j}},$
where
$\mathcal{O}_{\rho_{j}}=\\{z\in\mathbb{C}\colon\operatorname{dist}(z,[-1,1])\leq\rho_{j}-1\\}.$
3. 3.
This extension is bounded on $\mathcal{E}_{\bm{\rho}}$ according to
(2.11) $\displaystyle\sup_{{\bm{z}}\in\mathcal{E}_{\bm{\rho}}}\left\lVert
u_{\bm{z}}\right\rVert_{X}\leq C_{\varepsilon}.$
According to [10, Lemma 4.4], for any $s>1$, it holds
$\mathcal{E}_{s}\subset\mathcal{O}_{s}$. This fact will be used extensively
throughout this work.
###### Remark 2.7.
Let $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ be any sequence of numbers
strictly larger than one that is $(\bm{b},\varepsilon)$-admissible. In view of
Definition 2.1 and Theorem 2.2, Item 2 in Definition 2.6 is equivalent to
stating that for each $j\in\mathbb{N}$ and
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ there exists
$\partial_{z_{j}}u_{{\bm{z}}}\in X$, referred to as the derivative in the
direction $z_{j}$, such that
(2.12) $\left\lVert
u_{{\bm{z}}+h\bm{e}_{j}}-u_{{\bm{z}}}-\partial_{z_{j}}u_{{\bm{z}}}h\right\rVert_{X}=o\left(\left\lvert
h\right\rvert\right),\quad\text{as }\left\lvert h\right\rvert\rightarrow
0^{+},\;h\in\mathbb{C},$
where $\bm{e}_{j}$ signifies the $j$-th unit vector.
### 2.3 Affine-Parametric Boundary Transformations
Let $\hat{\Gamma}\subset\mathbb{R}^{3}$, in the following referred to as the
reference boundary, be a bounded Lipschitz boundary. We refer to
$\hat{\Gamma}$ as the reference boundary. To define parametrized families of
bounded Lipschitz boundaries in $\mathbb{R}^{3}$, we consider affine-
parametric vector fields of Karhunen-Loève type. To this end, we set
${{\bm{r}}}_{\bm{{y}}}\colon\hat{\Gamma}\rightarrow\mathbb{R}^{3}$, depending
on the parametric input ${\bm{{y}}}\in\mathbb{U}$, as
(2.13) ${{\bm{r}}}_{\bm{{y}}}(\hat{{\bf x}})={\bm{\varphi}}_{0}(\hat{{\bf
x}})+\sum_{j\geq 1}y_{j}{\bm{\varphi}}_{j}(\hat{{\bf x}}),\quad\hat{{\bf
x}}\in\hat{\Gamma},\quad{\bm{{y}}}=(y_{j})_{j\geq 1}\in\mathbb{U},$
with ${\bm{\varphi}}_{j}\colon\hat{\Gamma}\to\mathbb{R}^{3}$ for
$j\in\mathbb{N}$. This gives rise to a collection of parametric boundaries
$(\Gamma_{\bm{{y}}})_{{\bm{{y}}}\in\mathbb{U}}$ of the form
(2.14) $\Gamma_{{\bm{{y}}}}\coloneqq\\{{\bf x}\in\mathbb{R}^{3}:\;{\bf
x}={{\bm{r}}}_{\bm{{y}}}(\hat{{\bf x}}),\quad\hat{{\bf x}}\in\hat{\Gamma}\\}.$
This family of affinely parametrized boundaries have recently been
investigated in the context of forward and inverse UQ. We refer to [21] for a
detailed discussion on their construction using NURBS.
We make the following assumptions on ${{\bm{r}}}_{\bm{{y}}}$:
###### Assumption 2.8.
Let $\hat{\Gamma}$ be the reference Lipschitz boundary.
1. 1.
The functions
$({\bm{\varphi}}_{i})_{i\in\mathbb{N}}\subset\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})$
are such that for each ${\bm{{y}}}\in\mathbb{U}$ one has that
${{\bm{r}}}_{\bm{{y}}}\colon\hat{\Gamma}\to\Gamma_{\bm{{y}}}$ is bijective and
bi-Lipschitz.
2. 2.
There exists $p\in(0,1)$ such that
$\bm{b}\coloneqq\left(\left\lVert{\bm{\varphi}}_{j}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma},\mathbb{R}^{3})}\right)_{j\in\mathbb{N}}\in\ell^{p}(\mathbb{N}).$
###### Remark 2.9.
2.8 has the following consequences:
1. 1.
Item 1 implies that $\Gamma_{\bm{{y}}}$ is indeed a bounded Lipschitz boundary
itself.
2. 2.
Item 2 yields absolute uniform convergence of the series Eq. 2.13 as an
element of $\mathscr{C}^{0,1}(\hat{\Gamma},\mathbb{R}^{3})$.
We are interested in extending the set of affine-parametric boundary
parametrizations to a complex set of the form $\mathcal{O}_{\bm{\rho}}$, and
establish its holomorphic dependence upon the parametric variables, in the
sense of Definition 2.6. To this end, we set
(2.15)
$\displaystyle\left(\hat{\Gamma}\times\hat{\Gamma}\right)^{\star}\coloneqq\left\\{\hat{{\bf
x}},\hat{{\bf y}}\in\hat{\Gamma}:\hat{{\bf x}}\neq\hat{{\bf y}}\right\\}.$
###### Lemma 2.10.
Let 2.8 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and $p\in(0,1)$.
There exists $\eta>0$ and $\varepsilon>0$ such that for any
$(\bm{b},\varepsilon)$–admisible sequence $\bm{\rho}\coloneqq(\rho_{j})_{j\geq
1}$ of numbers strictly larger than one it holds
(2.16) $\inf_{{\bm{z}}\in\mathcal{O}_{{\bm{\rho}}}}\inf_{(\hat{\bf
x}\times\hat{\bf
y})\in(\hat{\Gamma}\times\hat{\Gamma})^{\star}}\Re\left\\{\frac{\left(\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{{\bm{z}}}(\hat{\bf y})\right)\cdot\left(\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{{\bm{z}}}(\hat{\bf y})\right)}{\left\lVert\hat{\bf x}-\hat{\bf
y}\right\rVert^{2}}\right\\}\geq\eta>0,$
###### Proof 2.11.
The proof of this result follows the exact same steps as in the proof of [27,
Proposition 3.2]. For the sake of brevity, we skip it.
###### Lemma 2.12.
Let 2.8 be satisfied. Then, there exists $\varepsilon>0$ such that the map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\bm{r}_{\bm{{y}}}\in\mathscr{C}^{0,1}(\hat{\Gamma},\mathbb{R}^{3})$
defined in Eq. 2.13 is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 2.13.
The proof of this result follows the exact same steps as in the proof of [11,
Lemma 5.8]. For the sake of brevity, we skip it.
### 2.4 Parametric Holomorphy of Boundary Functions
We study the parametric holomorphy property for a variety of functions
appearing in the parametric holomorphy analysis of BIOs to be carried out in
Section 3.
#### 2.4.1 Parametric Holomorphy of the Gramian
Let ${\bm{{y}}}\in\mathbb{U}$, $f\colon\Gamma_{\bm{{y}}}\rightarrow\mathbb{C}$
a measurable function. Since we assume $\hat{\Gamma}$ to be Lipschitz, there
exists a decomposition $\mathcal{G}$ of $\hat{\Gamma}$ into finitely many open
and disjoint patches $\tau$, i.e.,
(2.17)
$\displaystyle\hat{\Gamma}=\bigcup_{\tau\in\mathcal{G}}\overline{\tau},$
such that each $\tau\in\mathcal{G}$ is the image of a Lipschitz diffeomorphism
$\chi_{\tau}\colon\tilde{\tau}\to\tau$ with
$\tilde{\tau}\subset\mathbb{R}^{2}$. For each $\tau\in\mathcal{G}$, we set
$\chi_{\tau,{\bm{{y}}}}={{\bm{r}}}_{\bm{{y}}}\circ\chi_{\tau}$ and
$\tilde{f}_{\tau,{\bm{{y}}}}=f\circ{{\bm{r}}}_{\bm{{y}}}\circ\chi_{\tau}$.
Then, the surface integral of $f$ over $\Gamma_{\bm{{y}}}$ can be written as
$\displaystyle\int\limits_{\Gamma_{\bm{{y}}}}f(\mathbf{x})\mathrm{d}s_{\mathbf{x}}=\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\tilde{f}_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})\sqrt{g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})}\mathrm{d}\tilde{\mathbf{u}},$
with the Gramian determinant
$g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})=\operatorname{det}G_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}}),\qquad\tilde{\mathbf{u}}\in\tilde{\tau},$
of the Gramian matrix
$G_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})=\big{(}D_{\tilde{\mathbf{u}}}\chi_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})\big{)}^{\top}\big{(}D_{\tilde{\mathbf{u}}}\chi_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})\big{)},\qquad\tilde{\mathbf{u}}\in\tilde{\tau},$
where
$D_{\tilde{\mathbf{u}}}\chi_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})\in\mathbb{R}^{3\times
2}$ signifies the Jacobian of $\chi_{\tau,{\bm{{y}}}}$.
###### Lemma 2.14.
Let 2.8 be satisfied. Then there exists $\varepsilon>0$ such that for each
$\tau\in\mathcal{G}$ the maps
(2.18) $\displaystyle\mathbb{U}\ni{\bm{{y}}}\mapsto g_{\tau,{\bm{{y}}}}\in
L^{\infty}(\widetilde{\tau})$
and
(2.19)
$\displaystyle\mathbb{U}\ni{\bm{{y}}}\mapsto\sqrt{g_{\tau,{\bm{{y}}}}}\in
L^{\infty}(\widetilde{\tau})$
are $(\bm{b},p,\varepsilon)$-holomorphic and continuous. In particular, for
the complex extension
(2.20) $\displaystyle\mathcal{O}_{\bm{\rho}}\ni{\bm{z}}\mapsto
g_{\tau,{\bm{z}}}\in L^{\infty}(\tilde{\tau})$
there exist $0<\underline{\zeta}\leq\overline{\zeta}<\infty$ such that
(2.21) $\displaystyle
0<\underline{\zeta}\leq\inf_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\inf_{\tilde{\mathbf{u}}\in\tilde{\tau}}\Re(g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}}))\leq\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\sup_{\tilde{\mathbf{u}}\in\tilde{\tau}}\Re(g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}}))\leq\overline{\zeta}<\infty.$
###### Proof 2.15.
It follows from Item 2 in 2.8 that for each $\tau\in\mathcal{G}$ we have that
$\mathbb{U}\ni{\bm{{y}}}\mapsto D_{\tilde{{\bf
x}}}\chi_{\tau,{\bm{{y}}}}\in[L^{\infty}(\tilde{\tau})]^{3\times 2}$
and, thus, also
(2.22) $\mathbb{U}\ni{\bm{{y}}}\mapsto G_{\tau,{\bf
y}}\in\left[L^{\infty}(\tilde{\tau})\right]^{2\times 2}$
is $(\bm{b},p,\varepsilon^{\prime})$-holomorphic and continuous with
$(\bm{b},p,\varepsilon^{\prime})$ as in Lemma 2.12. Since the determinant is a
polynomial, this yields $(\bm{b},p,\varepsilon^{\prime})$-holomorphy of
$g_{\tau,{\bm{z}}}$. Moreover, for each ${\bm{{y}}}\in\mathbb{U}$, the Gramian
matrix $G_{\tau,{\bm{{y}}}}$ is positive definite and symmetric. Thus, it has
real and positive eigenvalues whose product coincides with
$g_{\tau,{\bm{{y}}}}$. Since the eigenvalues depend continuosly on
$G_{\tau,{\bm{{y}}}}$, this yields that $g_{\tau,{\bm{{y}}}}$ is bounded
uniformly from below and above. Thus, there is $\varepsilon>0$ such that Eq.
2.21 holds, which implies the $(\bm{b},p,\varepsilon)$-holomorphy and
continuity of Eq. 2.19.
#### 2.4.2 The Pullback Operator
For each ${\bm{{y}}}\in\mathbb{U}$ and $\varphi\in L^{2}(\Gamma_{\bm{{y}}})$,
we set
(2.23) $\tau_{\bm{{y}}}\varphi\coloneqq\varphi\circ\bm{r}_{\bm{{y}}}\in
L^{2}(\hat{\Gamma}).$
###### Lemma 2.16.
Let 2.8 be satisfied. Then, for each ${\bm{{y}}}\in\mathbb{U}$ the map
(2.24)
$L^{2}(\Gamma_{\bm{{y}}})\ni\varphi\mapsto\hat{\varphi}\coloneqq\tau_{\bm{{y}}}\varphi\in
L^{2}(\hat{\Gamma})$
defines an isomorphism, i.e.
$\tau_{\bm{{y}}}\in\mathscr{L}_{{\text{iso}}}(L^{2}_{\bm{{y}}}(\Gamma),L^{2}(\hat{\Gamma}))$
for each ${\bm{{y}}}\in\mathbb{U}$.
###### Proof 2.17.
As a consequence of Assumptions 2.8 for each ${\bm{{y}}}\in\mathbb{U}$ we
obtain
(2.25)
$\int\limits_{\Gamma_{\bm{{y}}}}\left\lvert\varphi(\mathbf{x})\right\rvert^{2}\text{d}s_{\mathbf{x}}=\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\left\lvert\left(\tau_{\bm{{y}}}\varphi\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\right\rvert^{2}\sqrt{g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})}\text{d}\tilde{\mathbf{u}},$
and
(2.26)
$\int\limits_{\hat{\Gamma}}\left\lvert\left(\tau_{\bm{{y}}}\varphi\right)(\hat{\mathbf{x}})\right\rvert^{2}\text{d}s_{\hat{\mathbf{x}}}=\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\left\lvert\left(\tau_{\bm{{y}}}\varphi\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\right\rvert^{2}\sqrt{g_{\tau}(\tilde{\mathbf{u}})}\text{d}\tilde{\mathbf{u}},$
where $g_{\tau}$ is the Gram determinant of $\chi_{\tau}$. Using Eq. 2.21 on
$\mathbb{U}\subset\mathcal{O}_{\bm{\rho}}$ yields the assertion.
#### 2.4.3 Parametric Holomorphy of the Normal Vector
Certain integral operators, such as the double layer operator in potential
layer theory, depend on the outward pointing normal. To this end, we establish
parametric holomorphy of the outward pointing normal vector.
###### Lemma 2.18.
Let 2.8 be fulfilled and $(\bm{b},p,\varepsilon)$ as in Lemma 2.14. Then,
pullback of the outward pointing normal
(2.27)
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\mathbf{n}}_{\bm{{y}}}\mathrel{\mathrel{\mathop{:}}=}\tau_{\bm{{y}}}\mathbf{n}_{\Gamma_{\bm{{y}}}}\in
L^{\infty}(\hat{\Gamma})$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 2.19.
Let $\tau\in\mathcal{G}$,
$\chi_{\tau,{\bm{{y}}}}={{\bm{r}}}_{\bm{{y}}}\circ\chi_{\tau}$, and $\hat{{\bf
x}}=\chi_{\tau}(\tilde{\mathbf{u}})$. Then, the non-normalized outward-
pointing normal
$\hat{\mathbf{m}}_{\bm{{y}}}(\hat{{\bf
x}})=\big{(}\partial_{1}\chi_{\tau,{\bm{{y}}}}\times\partial_{2}\chi_{\tau,{\bm{{y}}}}\big{)}(\hat{\mathbf{u}})$
is $(\bm{b},p,\varepsilon)$-holomorphic due to Lemma 2.12. Moreover, due to
$\left\lVert\hat{\mathbf{m}}_{{\bm{{y}}}}(\hat{\mathbf{x}})\right\rVert=\sqrt{g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})}$
for all ${\bm{{y}}}\in\mathbb{U}$, Lemma 2.14 implies that also
$\left\lVert\hat{\mathbf{m}}_{{\bm{{y}}}}\right\rVert$ is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous. Thus, since
$\hat{\mathbf{n}}_{\bm{{y}}}(\hat{{\bf
x}})=\frac{\hat{\mathbf{m}}_{\bm{{y}}}(\hat{{\bf
x}})}{\left\lVert\hat{\mathbf{m}}_{\bm{{y}}}(\hat{{\bf
x}})\right\rVert},\quad\hat{{\bf x}}\in\hat{\Gamma},$
LABEL:{eq:gramianbound} implies the assertion.
### 2.5 Riesz-Thorin Interpolation Theorem
We briefly recall a result concerning the boundedness of BIOs in a
$L^{2}$-based setting, to be used in the parametric holomorphy analysis in
Section 3. To this end, let $\Gamma$ be a bounded Lipschitz boundary in
$\mathbb{R}^{3}$ and
(2.28) $(\mathsf{T}_{\Gamma}\varphi)({\bf
x})\mathrel{\mathrel{\mathop{:}}=}\int\limits_{\Gamma}{\sf t}({\bf y},{\bf
x}-{\bf y})\varphi({\bf y})\mathrm{ds}_{\bf y},\quad{\bf x}\in\Gamma,$
an integral operator with kernel
$\mathsf{t}:\Gamma\times\mathbb{R}^{3}\backslash\\{{\bf
0}\\}\rightarrow\mathbb{C}$. In the following, to show the mapping property
$\mathsf{T}_{\Gamma}\colon L^{2}(\Gamma)\to L^{2}(\Gamma)$, we rely on the
_Riesz-Thorin interpolation theorem_ [38, Theorem 2.b.14], which states that
the operator norm of $\mathsf{T}_{\Gamma}:L^{2}(\Gamma)\rightarrow
L^{2}(\Gamma)$ is bounded according to
(2.29)
$\displaystyle\left\lVert\mathsf{T}_{{\Gamma}}\right\rVert_{\mathscr{L}\left(L^{2}(\Gamma),L^{2}(\Gamma)\right)}\leq\left\lVert\mathsf{T}_{{\Gamma}}\right\rVert_{\mathscr{L}\left(L^{1}(\Gamma),L^{1}(\Gamma)\right)}^{\frac{1}{2}}\left\lVert\mathsf{T}_{{\Gamma}}\right\rVert_{\mathscr{L}\left(L^{\infty}(\Gamma),L^{\infty}(\Gamma)\right)}^{\frac{1}{2}}.$
To estimate the operator norms in the upper bound of Eq. 2.29 we use their
explicit characterizations
(2.30)
$\displaystyle\left\lVert\mathsf{T}_{{\Gamma}}\right\rVert_{\mathscr{L}\left(L^{1}(\Gamma),L^{1}(\Gamma)\right)}$
$\displaystyle=\underset{{\bf
y}\in{\Gamma}}{\operatorname{es}\sup}\int\limits_{\Gamma}|\mathsf{t}({\bf
y},{\bf x}-{\bf y})|\mathrm{d}s_{\bf x},$ (2.31)
$\displaystyle\left\lVert\mathsf{T}_{{\Gamma}}\right\rVert_{\mathscr{L}\left(L^{\infty}(\Gamma),L^{\infty}(\Gamma)\right)}$
$\displaystyle=\underset{{\bf
x}\in{\Gamma}}{\operatorname{es\sup}}\int\limits_{\Gamma}|\mathsf{t}({\bf
y},{\bf x}-{\bf y})|\mathrm{d}s_{\bf y}.$
## 3 Parametric Shape Holomorphy of Boundary Integral Operators
In this section, we present and prove the parametric holomorphy results of the
parameter-to-operator map Eq. 1.3 and parameter-to-solution map Eq. 1.4 in an
$L^{2}$-based setting. To this end, for each ${\bm{{y}}}\in\mathbb{U}$ and by
using the pullback operator
$\tau_{\bm{{y}}}:L^{2}(\Gamma_{\bm{{y}}})\rightarrow L^{2}(\hat{\Gamma})$, we
can transport the BIO
$\mathsf{A}_{{\bm{{y}}}}:L^{2}(\Gamma_{\bm{{y}}})\rightarrow
L^{2}(\Gamma_{\bm{{y}}})$ introduced in Eq. 1.2 to the reference boundary
$\hat{\Gamma}$ as through
(3.1)
$\displaystyle\hat{\mathsf{A}}_{\bm{{y}}}\coloneqq\tau_{\bm{{y}}}\,\mathsf{A}_{{\bm{{y}}}}\,\tau^{-1}_{\bm{{y}}}.$
As a consequence of Lemma 2.16, for each ${\bm{{y}}}\in\mathbb{U}$, one has
that ${\mathsf{A}}_{{\bm{{y}}}}:L^{2}(\Gamma_{\bm{{y}}})\rightarrow
L^{2}(\Gamma_{\bm{{y}}})$ is a bounded linear operator if and only if
$\hat{\mathsf{A}}_{\bm{{y}}}\colon L^{2}(\hat{\Gamma})\to L^{2}(\hat{\Gamma})$
is so as well.
The remainder of the section is structured as follows. In Section 3.1,
assuming that the kernel function ${\sf
a}:\Gamma_{\bm{{y}}}\times\mathbb{R}\rightarrow\mathbb{C}$ is bounded, we
prove parametric holomorphy the the map
${\bm{{y}}}\mapsto\hat{\mathsf{A}}_{\bm{{y}}}$, in the sense of Definition
2.6. Next, in Section 3.2, we no longer assume ${\sf a}(\cdot,\cdot)$ to be
bounded, but only integrable, and extend the parametric holomorphy result to
this scenario. Finally, in Section 3.3, we establish parametric holomorphy of
the parameter-to-solution map.
### 3.1 Parametric Shape Holomorphy for BIOs with Bounded Kernel
We proceed to establish parametric shape holomorphy of a class of BIOs
characterized by a bounded kernel function. More precisely, upon assuming an
affinely parametrized boundary as the one introduced in Section 2.3, we
establish $(\bm{b},p,\varepsilon)$-holomorphy for the said BIO.
###### Theorem 3.1.
Let 2.8 be fulfilled. For each $\tau\in\mathcal{G}$ we define
(3.2) $\displaystyle\mathsf{k}_{\tau,{\bm{{y}}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\coloneqq\mathsf{a}\left(\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),\bm{r}_{\bm{{y}}}(\hat{\mathbf{x}})-\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right),\quad(\hat{{\bf
x}},\tilde{\mathbf{u}})\in\hat{\Gamma}\times\tilde{\tau}.$
In addition, for each $\tau\in\mathcal{G}$ we assume that the map
(3.3) $\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{k}_{\tau,{\bm{{y}}}}\in
L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous for some
$\varepsilon>0$. Then, the map
(3.4)
$\mathcal{A}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{A}}_{{\bm{{y}}}}$
with $\hat{\mathsf{A}}_{\bm{{y}}}$ as in Eq. 3.1 is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 3.2.
Throughout, let $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ be any
$(\bm{b},\varepsilon)$-admissible sequence of numbers of numbers strictly
larger than one. For the sake of readability, we divide the proof into three
steps.
Step A: Localization of $\hat{\mathsf{A}}_{\bm{{y}}}$ to a panel. Using the
decomposition into panels $\mathcal{G}$ of the reference boundary
$\hat{\Gamma}$ as introduced in Eq. 2.17, for any $\hat{\varphi}\in
L^{2}(\hat{\Gamma})$ and ${\bm{{y}}}\in\mathbb{U}$ one has
(3.5)
$\displaystyle\left(\hat{\mathsf{A}}_{\bm{{y}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)=\sum_{\tau\in\mathcal{G}}\left(\hat{\mathsf{A}}_{\tau,{\bm{{y}}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right),\quad\hat{\bf x}\in\hat{\Gamma}.$
For each $\tau\in\mathcal{G}$, let $\mathsf{k}_{\tau,{\bm{z}}}$ denote the
holomorphic extension of $\mathsf{k}_{\tau,{\bm{{y}}}}$ to
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ and define
(3.6) $\mathsf{b}_{\tau,{\bm{z}}}(\hat{\bf x},\hat{{\bf
u}})\coloneqq\mathsf{k}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})},\quad(\hat{\bf
x},\hat{\bf u})\in\hat{\Gamma}\times\tilde{\tau}.$
For $\hat{\varphi}\in L^{2}(\hat{\Gamma})$, set
(3.7)
$\left(\hat{\mathsf{A}}_{\tau,{\bm{z}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)\coloneqq\int\limits_{\tilde{\tau}}\mathsf{b}_{\tau,{\bm{z}}}(\hat{\bf
x},\hat{{\bf
u}})\left(\hat{\varphi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\,\text{d}\tilde{\mathbf{u}}.$
The map
(3.8) $\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{b}_{\tau,{\bm{{y}}}}\in
L^{\infty}(\hat{\Gamma}\times\tilde{\tau})$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous as a consequence of
Lemma 2.14 and the properties of the map
${\bm{{y}}}\mapsto\mathsf{k}_{\tau,{\bm{{y}}}}$, stated as an assumption. For
$j,m\in\mathbb{N}$, $\tau\in\mathcal{G}$, Cauchy’s integral formula yields
(3.9)
$\displaystyle\partial^{m}_{z_{j}}\mathsf{b}_{\tau,{\bm{z}}}=\frac{m!}{2\pi\imath}\int\limits_{z^{\prime}_{j}\in\partial
B(z_{j},\zeta)}\frac{\mathsf{b}_{\tau,{\bm{z}}^{\prime}}}{(z^{\prime}_{j}-z_{j})^{m+1}}{{\text{d}}}z^{\prime}_{j}$
with ${\bm{z}}^{\prime}=(z_{1},\dots,{z}^{\prime}_{j},\dots)$. Hence, for
$\zeta>0$ such that $B(z_{j},\zeta)\subset\mathcal{O}_{\rho_{j}}$ we obtain
(3.10)
$\left\lVert\partial^{m}_{z_{j}}\mathsf{b}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)}\leq\frac{m!}{\zeta^{m}}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\mathsf{b}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)}.$
Setting
(3.11)
$\displaystyle\left(\hat{\mathsf{A}}^{(j,m)}_{\tau,{\bm{z}}}\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)\coloneqq\int\limits_{\tilde{\tau}}\left(\partial^{m}_{z_{j}}\mathsf{b}_{\tau,{\bm{z}}}\right)(\hat{\bf
x},\hat{\bf
u})\left(\hat{\varphi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\,\text{d}\tilde{\mathbf{u}},\quad\hat{\bf
x}\in\hat{\Gamma}$
for $\hat{\varphi}\in L^{2}(\hat{\Gamma})$, yields as a consequence of the
Riesz-Thorin interpolation theorem from Section 2.5
(3.12)
$\left\lVert\hat{\mathsf{A}}^{(j,m)}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma}))}\leq
C_{\tau}\left\lVert\partial^{m}_{z_{j}}\mathsf{b}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)},$
where $C_{\tau}>0$ depends only on $\tau\in\mathcal{G}$, i.e.,
$\hat{\mathsf{A}}^{(j,m)}_{\tau,{\bm{z}}}:L^{2}(\hat{\Gamma})\rightarrow
L^{2}(\hat{\Gamma})$ is linear and bounded for each
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$. As a special case, we obtain
(3.13)
$\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma}))}\leq
C_{\tau}\left\lVert\mathsf{b}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)},$
which, together with the $(\bm{b},p,\varepsilon)$-holomorphy of the map
${\bm{{y}}}\mapsto\mathsf{b}_{\tau,{\bm{{y}}}}$ from Eq. 3.8, allows to verify
Item 1 and Item 3 in the Definition 2.6 of $(\bm{b},p,\varepsilon)$-holomorphy
for the map
(3.14)
$\mathcal{A}_{\tau}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{A}}_{\tau,{\bm{{y}}}}.$
It thus only remains to verify Item 2.
Step B: Differentiability of $\mathcal{A}_{\tau}$. Let
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$, $j\in\mathbb{N}$ and consider
$h\in\mathbb{C}$ small enough such that ${\bm{z}}+\eta
h\bm{e}_{j}\in\mathcal{O}_{\bm{\rho}}$ for all $\eta\in[0,1]$. Taylor’s
expansion of the map ${\bm{z}}\mapsto\mathsf{b}_{\tau,{\bm{z}}}$ in the
direction $z_{j}$ yields
(3.15)
$\mathsf{b}_{\tau,{\bm{z}}+h\bm{e}_{j}}=\mathsf{b}_{\tau,{\bm{z}}}+h\,\partial_{{z_{j}}}\mathsf{b}_{\tau,{\bm{z}}}+h^{2}\int\limits_{0}^{1}(1-\eta)\left(\partial^{2}_{z_{j}}\mathsf{b}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\right){{\text{d}}}\eta.$
Let $\widetilde{\rho}_{j}<{\rho}_{j}$. For each
${\bm{z}}\in\mathcal{O}_{\widetilde{\bm{\rho}}}$ with
$\widetilde{\bm{\rho}}=(\rho_{1},\rho_{2},\dots,\widetilde{\rho}_{j},\dots)$
there exists ${\bm{{y}}}\in\mathbb{U}$ such that $\left\lvert
z_{j}-y_{j}\right\rvert<\widetilde{\rho}_{j}-1$. Consider $h\in\mathbb{C}$
satisfying $0<\left\lvert h\right\rvert<{\rho}_{j}-\widetilde{\rho}_{j}$. Set
(3.16) $\zeta_{h}={\rho}_{j}-\widetilde{\rho}_{j}-\left\lvert h\right\rvert$
We need to ensure $B(z_{j}+\eta h,\zeta_{h})\subset\mathcal{O}_{\rho_{j}}$.
For $\widetilde{z}_{j}\in B(z_{j}+\eta h,\zeta_{h})$ we have
(3.17) $\displaystyle\left\lvert\widetilde{z}_{j}-y_{j}\right\rvert$
$\displaystyle\leq\left\lvert\widetilde{z}_{j}-z_{j}-\eta
h\right\rvert+\left\lvert h\right\rvert+\left\lvert z_{j}-y_{j}\right\rvert$
$\displaystyle\leq\zeta_{h}+\left\lvert
h\right\rvert+\widetilde{\rho}_{j}-1\leq\rho_{j}-1.$
Thus, for $\tau\in\mathcal{G}$, $\hat{\varphi}\in L^{2}(\hat{\Gamma})$ and
${\bm{z}}\in\mathcal{O}_{\widetilde{\bm{\rho}}}$ one has
(3.18)
$\hat{\mathsf{A}}_{\tau,{\bm{z}}+h\bm{e}_{j}}\hat{\varphi}=\hat{\mathsf{A}}_{\tau,{\bm{z}}}\hat{\varphi}+h\,\hat{\mathsf{A}}^{(j,1)}_{\tau,{\bm{z}}}\hat{\varphi}+h^{2}\int\limits_{0}^{1}(1-\eta)\left(\hat{\mathsf{A}}^{(j,2)}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\hat{\varphi}\right){{\text{d}}}\eta.$
Next, we have
(3.19)
$\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}+h\bm{e}_{j}}\hat{\varphi}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}\hat{\varphi}-h\,\hat{\mathsf{A}}^{(j,1)}_{\tau,{\bm{z}}}\hat{\varphi}\right\rVert_{L^{2}\left(\hat{\Gamma}\right)}\leq\left\lvert
h\right\rvert^{2}\sup_{\eta\in[0,1]}\left\lVert\hat{\mathsf{A}}^{(j,2)}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\,\hat{\varphi}\right\rVert_{L^{2}\left(\hat{\Gamma}\right)}.$
Thus, recalling Eq. 3.12 and Eq. 3.10 one obtains
(3.20)
$\sup_{\eta\in[0,1]}\left\lVert\hat{\mathsf{A}}^{(j,2)}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\,\hat{\varphi}\right\rVert_{L^{2}\left(\hat{\Gamma}\right)}\leq\frac{2}{\zeta^{2}_{h}}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\mathsf{b}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)}.$
Hence, it holds
(3.21)
$\displaystyle\left\lVert\frac{\hat{\mathsf{A}}_{\tau,{\bm{z}}+h\bm{e}_{j}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}}{h}-\hat{\mathsf{A}}^{(j,1)}_{\tau,{\bm{z}}}\right\rVert_{{\mathscr{L}(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma}))}}\leq\left\lvert
h\right\rvert\frac{2}{\zeta^{2}_{h}}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\mathsf{b}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)},$
which shows that $\hat{\mathsf{A}}^{(j,1)}_{\tau,{\bm{z}}}$ is the derivative
of $\hat{\mathsf{A}}_{\tau,{\bm{z}}}$. Hence, the map
(3.22)
$\mathcal{A}_{\tau}:\mathcal{O}_{\widetilde{\bm{\rho}}}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{z}}\mapsto\hat{\mathsf{A}}_{\tau,{\bm{z}}}.$
is holomorphic at each ${\bm{z}}\in\mathcal{O}_{\widetilde{\bm{\rho}}}$ in
each direction $z_{j}$.
To show that $\mathcal{A}_{\tau}$ is $(\bm{b},p,\varepsilon)$-holomorphic, we
observe that for any $\rho>1$ one has
(3.23)
$\mathcal{O}_{{{\rho}}}=\bigcup_{\widetilde{\rho}:\widetilde{\rho}<\rho}\mathcal{O}_{\widetilde{{\rho}}}$
and one can extend the aforementioned property to
${\bm{z}}\in\mathcal{O}_{{\bm{\rho}}}$. Therefore, we have verified Item 2 in
Definition 2.6 and Eq. 3.14 is $(\bm{b},p,\varepsilon)$-holomorphic for each
$\tau\in\mathcal{G}$.
Step C: $(\bm{b},p,\varepsilon)$-holomorphy of $\mathcal{A}$. It remains to
conclude that $\mathcal{A}$ is $(\bm{b},p,\varepsilon)$-holomorphic. It
follows from Eq. 3.5 that
(3.24) $\mathcal{A}=\sum_{\tau\in\mathcal{G}}\mathcal{A}_{\tau}$
and since the addition of a finite number of holomorphic maps preserves
holomorphy, this yields the desired result.
### 3.2 Parametric Shape Holomorphy of BIOs with Singular Kernel
We now address the case in which the kernel function ${\sf
a}:\Gamma_{\bm{{y}}}\times\mathbb{R}^{3}\rightarrow\mathbb{C}$ of Eq. 1.2 is
allowed to have a singularity in the second variable, however we assume that
is integrable. To this end, we define
(3.25) $\displaystyle\mathfrak{U}\coloneqq\\{{\bf
z}\in\mathbb{C}^{d}:\Re\\{{\bf z}\cdot{\bf z}\\}>0\\},$
where $d\in\\{2,3\\}$.
###### Theorem 3.3.
Let 2.8 be fulfilled. In addition, assume that
1. 1.
There exists $\varepsilon>0$ such that for each $\tau\in\mathcal{G}$,
$\tilde{\mathbf{u}}\in\tilde{\tau}$ and for all ${\bf z}\in\mathfrak{U}$ the
map
(3.26)
$\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{a}\left(\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),{\bf
z}\right)\in\mathbb{C}$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
2. 2.
Let $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ be any
$(\bm{b},\varepsilon)$-admissible sequence of numbers strictly larger than
one. For any ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ the map
(3.27) $\mathfrak{U}\ni{\bf
z}\mapsto\mathsf{a}\left(\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),{\bf
z}\right)\in\mathbb{C}$
is entire.
3. 3.
Again, let $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ be any
$(\bm{b},\varepsilon)$-admissible sequence of numbers strictly larger than
one. There exists $C_{\mathsf{a}}>0$ and $\nu\in[0,2)$ such that for any
$\tau\in\mathcal{G}$, $\tilde{\mathbf{u}}\in\tilde{\tau}$, and
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ one has
(3.28)
$\left\lvert\mathsf{a}\left(\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\hat{{\bf
y}}\right),\bm{r}_{\bm{z}}(\hat{{\bf x}})-\bm{r}_{\bm{z}}(\hat{{\bf
y}})\right)\right\rvert\leq\frac{C_{\mathsf{a}}}{\left\lVert\bm{r}_{\bm{z}}(\hat{{\bf
x}})-\bm{r}_{\bm{z}}(\hat{{\bf y}})\right\rVert^{\nu}},\quad\left(\hat{{\bf
x}},\hat{{\bf
y}}\right)\in\left(\hat{\Gamma}\times\hat{\Gamma}\right)^{\star},$
Then, the map
(3.29)
$\mathcal{A}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{A}}_{{\bm{{y}}}}$
with $\hat{\mathsf{A}}_{\bm{{y}}}$ as in Eq. 3.1 is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous.
For the proof of the theorem, we follow the general strategy of the proof of
Theorem 3.1 and make the necessary modifications for singular kernel
functions.
###### Proof 3.4 (Proof of Theorem 3.3).
As per usual, throughout this proof let $\bm{\rho}\coloneqq(\rho_{j})_{j\geq
1}$ be any sequence of numbers strictly larger than one that is
$(\bm{b},\varepsilon)$-admissible for some $\varepsilon>0$.
Step A: Localization of $\hat{\mathsf{A}}_{\bm{{y}}}$ to a patch. In analogy
to the proof of Theorem 3.1 we set
(3.30) $\displaystyle\mathsf{k}_{\tau,{\bm{{y}}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\coloneqq\mathsf{a}\left(\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),\bm{r}_{\bm{{y}}}(\hat{\mathbf{x}})-\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right)$
and
(3.31) $\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\coloneqq\mathsf{k}_{\tau,{\bm{{y}}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}$
for each $(\hat{{\bf x}},\tilde{\mathbf{u}})\in\hat{\Gamma}\times\tilde{\tau}$
with $\hat{{\bf x}}\neq\chi_{\tau}(\tilde{\mathbf{u}})$, $\tau\in\mathcal{G}$,
and ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ and define
(3.32)
$\left(\hat{\mathsf{A}}_{\tau,{\bm{z}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)\coloneqq\int\limits_{\tilde{\tau}}\mathsf{b}_{\tau,{\bm{z}}}(\hat{\bf
x},\hat{{\bf
u}})\left(\hat{\varphi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\,\text{d}\tilde{\mathbf{u}},$
for $\hat{\varphi}\in L^{2}(\hat{\Gamma})$. Using the Riesz-Thorin
interpolation introduced in Section 2.5 we obtain for each
$\tau\in\mathcal{G}$
(3.33)
$\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)}\leq\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}\left(L^{1}(\hat{\Gamma}),L^{1}(\hat{\Gamma})\right)}^{\frac{1}{2}}\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}\left(L^{\infty}(\hat{\Gamma}),L^{\infty}(\hat{\Gamma})\right)}^{\frac{1}{2}}.$
Lemma 2.10 yields for each $\tau\in\mathcal{G}$ and for any
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$
(3.34) $\displaystyle\left\lVert\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{2}$
$\displaystyle\geq\Re\\{\left(\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right)\cdot\left(\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{{\bm{z}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right)\\}$
$\displaystyle\geq\eta\left\lVert\hat{\bf
x}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{2}$
$\displaystyle>0.$
This, together with Item 3 from the assumptions, yields
(3.35) $\left\lvert\mathsf{b}_{\tau,{\bm{z}}}(\hat{\bf x},\hat{\bf
u})\right\rvert\leq\frac{C_{\mathsf{a}}}{\left\lVert\bm{r}_{\bm{z}}(\hat{\bf
x})-\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\leq\frac{\widetilde{C}_{\mathsf{a}}}{\left\lVert\hat{\bf
x}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert,$
with $\widetilde{C}_{\mathsf{a}}=C_{\mathsf{a}}\eta^{-\nu/2}$ and for any
$(\hat{\bf x},\hat{\bf u})\in\hat{\Gamma}\times\tilde{\tau}$ with $\hat{\bf
x}\neq\chi_{\tau}\left(\tilde{\mathbf{u}}\right)$. Thus, for each
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$, it holds
(3.36)
$\displaystyle\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}\left(L^{1}(\hat{\Gamma}),L^{1}(\hat{\Gamma})\right)}$
$\displaystyle=\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\left\lvert\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\right\rvert\text{d}\text{s}_{\hat{\mathbf{x}}}$
$\displaystyle\leq\widetilde{C}_{\mathsf{a}}\,\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\,\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\frac{1}{\left\lVert\hat{\bf
x}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}\text{d}\text{s}_{\hat{\mathbf{x}}},$
where the latter integral is finite for any
$\tilde{\mathbf{u}}\in\tilde{\tau}$ since $\nu\in[0,2)$.
Furthermore, for each ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$, it holds
$\displaystyle\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}\right\rVert_{\mathscr{L}\left(L^{\infty}(\hat{\Gamma}),L^{\infty}(\hat{\Gamma})\right)}$
$\displaystyle=\underset{\hat{\mathbf{x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\tilde{\tau}}\left\lvert\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\right\rvert\text{d}\tilde{\mathbf{u}}$
$\displaystyle\leq\widetilde{C}_{\mathsf{a}}\underset{\hat{\bf
u}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\,\underset{\hat{\mathbf{x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\tilde{\tau}}\frac{1}{\left\lVert\hat{\bf
x}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}\text{d}\tilde{\mathbf{u}}$
$\displaystyle\leq\widetilde{C}_{\mathsf{a}}\underset{\hat{\bf
u}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\,\underset{\hat{\mathbf{x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\frac{1}{\left\lVert\hat{\bf
x}-\hat{{\bf y}}\right\rVert^{\nu}}\text{d}\text{s}_{\hat{{\bf y}}},$
where the assumptions on $\nu$ imply again that latter integral is uniformly
bounded for all $\hat{{\bf x}}\in\hat{\Gamma}$. This shows that
$\hat{\mathsf{A}}_{\tau,{\bm{z}}}\colon L^{2}(\hat{\Gamma})\to
L^{2}(\hat{\Gamma})$ is a bounded linear operator and allows us to define for
each ${\tau}\in\mathcal{G}$
(3.37)
$\mathcal{A}_{\tau}:\mathcal{O}_{\bm{\rho}}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{z}}\mapsto\hat{\mathsf{A}}_{\tau,{\bm{z}}}.$
Step A’: Regularization of localized operators
$\hat{\mathsf{A}}_{\tau,{\bm{z}}}$. Let $\phi:[0,\infty)\rightarrow\mathbb{R}$
be a $\mathscr{C}^{1}([0,\infty))$ function with
$\phi(t)\begin{cases}=0&t\in[0,\frac{1}{2}],\\\
\in[0,1]&t\in(\frac{1}{2},1),\\\ =1&t\in[1,\infty).\end{cases}$
For each $n\in\mathbb{N}$, $\tau\in\mathcal{G}$, and
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ define
(3.38) $\mathsf{b}^{(n)}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\coloneqq\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})$
and
(3.39)
$\left(\hat{\mathsf{A}}^{(n)}_{\tau,{\bm{z}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)\coloneqq\int\limits_{\tilde{\tau}}\mathsf{b}^{(n)}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\left(\hat{\varphi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\,\text{d}\tilde{\mathbf{u}},\quad\hat{\bf
x}\in\hat{\Gamma},$
for all $\hat{\varphi}\in L^{2}(\hat{\Gamma})$. Using the Riesz-Thorin
interpolation theorem Eq. 2.29 we obtain
(3.40)
$\displaystyle\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma}))}$
$\displaystyle\hskip
42.67912pt\leq\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}(L^{1}(\hat{\Gamma}),L^{1}(\hat{\Gamma}))}^{\frac{1}{2}}\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}(L^{\infty}(\hat{\Gamma}),L^{\infty}(\hat{\Gamma}))}^{\frac{1}{2}}.$
To estimate this quantity, we note that for $\hat{\bf x}\in\hat{\Gamma}$ it
holds
$\displaystyle\left(\hat{\mathsf{A}}_{\tau,{\bm{z}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)-\left(\hat{\mathsf{A}}^{(n)}_{\tau,{\bm{z}}}\,\hat{\varphi}\right)\left(\hat{{\bf
x}}\right)$
$\displaystyle\qquad\qquad\qquad=\int\limits_{\tilde{\tau}}\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\left(1-\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})\right)\left(\hat{\varphi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\,\text{d}\tilde{\mathbf{u}}.$
Therein, along the lines of step A, the integrand can be estimated as
(3.41) $\displaystyle\left\lvert\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\left(1-\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})\right)\right\rvert\leq\widetilde{C}_{\mathsf{a}}\frac{\left(1-\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})\right)}{\left\lVert\hat{\bf
x}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert,$
for $(\hat{\bf x},\hat{\bf u})\in\hat{\Gamma}\times\tilde{\tau}$ with
$\hat{\bf x}\neq\chi_{\tau}\left(\tilde{\mathbf{u}}\right)$. Thus, it holds
(3.42)
$\displaystyle\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}\left(L^{1}(\hat{\Gamma}),L^{1}(\hat{\Gamma})\right)}$
$\displaystyle\qquad\qquad=\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\left\lvert\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\right\rvert\left(1-\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})\right)\text{ds}_{\hat{\mathbf{x}}}$
$\displaystyle\qquad\qquad\leq\widetilde{C}_{\mathsf{a}}\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\,\underset{\hat{{\bf
y}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\frac{1-\phi(n\left\lVert\hat{\mathbf{x}}-\hat{\mathbf{y}}\right\rVert^{\nu})}{\left\lVert\hat{\mathbf{x}}-\hat{\mathbf{y}}\right\rVert^{\nu}}\text{ds}_{\hat{\mathbf{x}}}.$
On the other hand, it holds
(3.43)
$\displaystyle\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}\left(L^{\infty}(\hat{\Gamma}),L^{\infty}(\hat{\Gamma})\right)}$
$\displaystyle\qquad\qquad\leq\underset{\hat{{\bf
x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\tilde{\tau}}\left\lvert\mathsf{b}_{\tau,{\bm{z}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\left(1-\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})\right)\right\rvert\text{d}\tilde{\mathbf{u}}$
$\displaystyle\qquad\qquad\leq\widetilde{C}_{\mathsf{a}}\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\,\underset{\hat{{\bf
x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\tilde{\tau}}\frac{1-\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})}{\left\lVert\hat{\bf
x}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}\text{d}\tilde{\mathbf{u}}$
$\displaystyle\qquad\qquad\leq\widetilde{C}_{\mathsf{a}}\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert\underset{\hat{{\bf
x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\frac{1-\phi(n\left\lVert\hat{\mathbf{x}}-\hat{{\bf
y}}\right\rVert^{\nu})}{\left\lVert\hat{\bf x}-\hat{{\bf
y}}\right\rVert^{\nu}}\text{d}\text{s}_{\hat{{\bf y}}}.$
Upon noting that the integrals in Eq. 3.42 and Eq. 3.43 are equal, it remains
to estimate the essential supremum of this integral. Let us calculate
(3.44) $\displaystyle\underset{\hat{{\bf
x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}}\frac{1-\phi(n\left\lVert\hat{\mathbf{x}}-\hat{{\bf
y}}\right\rVert^{\nu})}{\left\lVert\hat{\bf x}-\hat{{\bf
y}}\right\rVert^{\nu}}\text{d}\text{s}_{\hat{{\bf y}}}$
$\displaystyle\leq\underset{\hat{{\bf
x}}\in\hat{\Gamma}}{\operatorname{ess}\sup}\int\limits_{\hat{\Gamma}\cap
B(\hat{{\bf x}},1/n)}\frac{1}{\left\lVert\hat{\bf x}-\hat{{\bf
y}}\right\rVert^{\nu}}\text{d}\text{s}_{\hat{{\bf y}}}$ $\displaystyle\leq
C\int_{B(0,1/n)}\frac{1}{\|\tilde{\mathbf{u}}\|^{\nu}}\text{d}\tilde{\mathbf{u}}$
$\displaystyle=\int\limits_{0}^{1/n}\int\limits_{0}^{2\pi}r^{1-\nu}\text{d}r\text{d}t$
$\displaystyle=\frac{2\pi}{2-\nu}n^{\nu-2}.$
Next, recalling Eq. 3.40 we get
(3.45)
$\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)}\leq\widetilde{C}_{\mathsf{a}}\frac{2\pi}{2-\nu}\underbrace{\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\underset{\tilde{\mathbf{u}}\in\tilde{\tau}}{\operatorname{ess}\sup}\left\lvert\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\right\rvert}_{(\spadesuit)}n^{\nu-2}.$
As a consequence of Lemma 2.14, the term $(\spadesuit)$ in (3.45) is bounded.
Recalling that we have assumed $\nu\in[0,2)$ implies
(3.46)
$\lim_{n\rightarrow\infty}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\hat{\mathsf{A}}_{\tau,{\bm{z}}}-\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}\right\rVert_{\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)}=0.$
This implies that for each $\tau\in\mathcal{G}$
(3.47)
$\mathcal{A}_{\tau}^{(n)}:\mathcal{O}_{\bm{\rho}}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{z}}\mapsto\hat{\mathsf{A}}_{\tau,{\bm{z}}}^{(n)}$
converges uniformly in $\mathcal{O}_{\bm{\rho}}$ to $\mathcal{A}_{\tau}$ from
Eq. 3.37.
Step B: $(\bm{b},p,\varepsilon)$-holomorphy of localized, regularized
operators $\hat{\mathsf{A}}_{\tau,{\bm{{y}}}}^{(n)}$. We aim to verify the
assumptions of Theorem 3.1 to show that
(3.48)
$\mathcal{A}_{\tau}^{(n)}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{A}}_{\tau,{\bm{{y}}}}^{(n)}$
is $(\bm{b},p,\varepsilon)$-holomorphic with $\varepsilon>0$ _independent_ of
$n\in\mathbb{N}$. Thus, by setting
(3.49) $\mathsf{k}_{\tau,{\bm{{y}}}}^{(n)}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\coloneqq\mathsf{k}_{\tau,{\bm{{y}}}}\left(\hat{{\bf
x}},\tilde{\mathbf{u}}\right)\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})$
with $\mathsf{k}_{\tau,{\bm{{y}}}}$ from Eq. 3.30 we need to show that
(3.50) $\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{k}_{\tau,{\bm{{y}}}}^{(n)}\in
L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)$
is $(\bm{b},p,\varepsilon)$-holomorphic with $\varepsilon>0$, again,
independent of $n$ and continuous. To this end, due to Eq. 3.34, for each
fixed $(\hat{\bf x},\hat{\bf u})\in\hat{\Gamma}\times\tilde{\tau}$ such that
${\bf x}\neq\chi_{\tau}\left(\tilde{\mathbf{u}}\right)$ and any
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ it holds
(3.51) $\left(\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right)\cdot\left(\bm{r}_{{\bm{z}}}(\hat{\bf
x})-\bm{r}_{{\bm{z}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right)\in\mathfrak{U}.$
Thus, due to Item 3 in the assumptions and Eq. 3.34, it follows that
(3.52)
$\displaystyle\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lvert\mathsf{k}^{(n)}_{\tau,{\bm{z}}}(\hat{\bf
x},\hat{\bf u})\right\rvert$ $\displaystyle\leq
C_{\mathsf{a}}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\frac{\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})}{\left\lVert\bm{r}_{\bm{z}}(\hat{\bf
x})-\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}}$
$\displaystyle\leq\widetilde{C}_{\mathsf{a}}\frac{\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})}{\left\lVert\hat{{\bf
x}}-\chi_{\tau}(\tilde{\mathbf{u}})\right\rVert^{\nu}}$
$\displaystyle\leq\widetilde{C}_{\mathsf{a}}n\sup_{t\in[0,\infty)}\left\lvert\phi^{\prime}(t)\right\rvert.$
Thus,
(3.53)
$\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\mathsf{k}^{(n)}_{\tau,{\bm{z}}}\right\rVert_{L^{\infty}(\hat{\Gamma}\times\tilde{\tau})}\leq\widetilde{C}_{\mathsf{a}}n\sup_{t\in[0,\infty)}\left\lvert\phi^{\prime}(t)\right\rvert.$
This allows to verify Item 1 and Item 3 in the Definition 2.6 of
$(\bm{b},p,\varepsilon)$-holomorphy and it only remains to verify Item 2.
To this end, we observe that the map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\bm{r}_{\bm{{y}}}(\hat{\mathbf{x}})-\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\in\mathbb{C}$
is $(\bm{b},p,\varepsilon)$-holomorphic itself as a consequence of Lemma 2.12,
with $\varepsilon>0$ independent of $(\hat{\bf x},\hat{\bf
u})\in\hat{\Gamma}\times\tilde{\tau}$. Consequently, recalling Item 1 and Item
2 in the assumptions, we may conclude that for each fixed $(\hat{\bf
x},\hat{\bf u})\in\hat{\Gamma}\times\tilde{\tau}$ such $\hat{\bf
x}\neq\chi_{\tau}\left(\tilde{\mathbf{u}}\right)$ that the map
(3.54)
$\displaystyle\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{k}_{\tau,{\bm{{y}}}}(\hat{\bf
x},\hat{\bf u})\in\mathbb{C}$
is $(\bm{b},p,\varepsilon)$-holomorphic for some $\varepsilon>0$ that does not
depend on $(\hat{\bf x},\hat{\bf u})\in\hat{\Gamma}\times\tilde{\tau}$. To
show holomorphy on $\hat{\Gamma}\times\tilde{\tau}$, let
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$, $j\in\mathbb{N}$ and consider
$h\in\mathbb{C}$ small enough such that
${\bm{z}}+h\bm{e}_{j}\in\mathcal{O}_{\bm{\rho}}$. Using Taylor’s expansion in
the direction $z_{j}$ we get
(3.55) $\displaystyle\mathsf{k}^{(n)}_{\tau,{\bm{z}}+h\bm{e}_{j}}(\hat{\bf
x},\hat{\bf u})$ $\displaystyle=\mathsf{k}^{(n)}_{\tau,{\bm{z}}}(\hat{\bf
x},\hat{\bf
u})+h\left(\partial_{z_{j}}\mathsf{k}_{\tau,{\bm{z}}}\right)\\!(\hat{\bf
x},\hat{\bf
u})\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})$
$\displaystyle\qquad+h^{2}\int\limits_{0}^{1}(1-\eta)\left(\partial^{2}_{z_{j}}\mathsf{k}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\right)\\!(\hat{\bf x},\hat{\bf
u})\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}){{\text{d}}}\eta.$
Thus, for $\zeta_{h}$ such that
$B(z_{j}+h\eta,\zeta_{h})\subset\mathcal{O}_{\rho_{j}}$, Cauchy’s integral
formula yields
(3.56) $\displaystyle\left(\partial^{2}_{z_{j}}\mathsf{k}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\right)(\hat{\bf x},\hat{\bf
u})\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})$
$\displaystyle\qquad\qquad=\frac{1}{\pi\imath}\int\limits_{z^{\prime}_{j}\in\partial
B(z_{j}+h\eta,\zeta_{{\bm{z}},h})}\frac{\mathsf{k}_{\tau,{\bm{z}}^{\prime}}(\hat{\bf
x},\hat{\bf u})}{(z^{\prime}_{j}-z_{j}-\eta
h)^{3}}\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}){{\text{d}}}z^{\prime}_{j},$
with ${\bm{z}}^{\prime}=(z_{1},\dots,z_{j}^{\prime},\dots)$, and
(3.57)
$\displaystyle\left\lvert\left(\partial^{2}_{z_{j}}\mathsf{k}_{\tau,{\bm{z}}+\eta
h\bm{e}_{j}}\right)(\hat{\bf x},\hat{\bf
u})\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu})\right\rvert\leq\frac{2}{\zeta_{h}^{2}}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lvert\mathsf{k}_{{\bm{z}}}^{(n)}(\hat{\bf
x},\hat{\bf u})\right\rvert.$
Hence, from Eq. 3.55,
(3.58)
$\left\lVert\frac{\mathsf{k}^{(n)}_{{\bm{z}}+h\bm{e}_{j}}-\mathsf{k}^{(n)}_{{\bm{z}}}}{h}-\mathsf{g}^{(n)}_{\tau,{\bf
z}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)}\leq\frac{\left\lvert
h\right\rvert}{\zeta_{h}^{2}}\sup_{{\bm{z}}\in\mathcal{O}_{\bm{\rho}}}\left\lVert\mathsf{k}_{{\bm{z}}}^{(n)}\right\rVert_{L^{\infty}(\hat{\Gamma}\times\tilde{\tau})},$
where $\mathsf{g}^{(n)}_{\tau,{\bf z}}\in
L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)$ is defined for each
$n\in\mathbb{N}$ as
(3.59) $\mathsf{g}^{(n)}_{\tau,{\bm{z}}}(\hat{\bf x},\hat{\bf
u})\coloneqq\left(\partial_{z_{j}}\mathsf{k}_{\tau,{\bm{z}}}\right)\\!(\hat{\bf
x},\hat{\bf
u})\phi(n\left\lVert\hat{\mathbf{x}}-\chi_{\tau}\left(\tilde{\mathbf{u}}\right)\right\rVert^{\nu}),$
and using Eq. 3.53 we obtain
(3.60)
$\left\lVert\frac{\mathsf{k}^{(n)}_{{\bm{z}}+h\bm{e}_{j}}-\mathsf{k}^{(n)}_{{\bm{z}}}}{h}-\mathsf{g}^{(n)}_{\tau,{\bf
z}}\right\rVert_{L^{\infty}\left(\hat{\Gamma}\times\tilde{\tau}\right)}\leq\widetilde{C}_{\mathsf{a}}\frac{\left\lvert
h\right\rvert}{\zeta^{2}_{h}}n\sup_{t\in[0,\infty)}\left\lvert\phi^{\prime}(t)\right\rvert.$
Thus, upon noting that $\zeta_{h}$ is bounded away from zero for $h\to 0$,
this shows that for each $n\in\mathbb{N}$ the map Eq. 3.50 and, due to Theorem
3.1 also Eq. 3.48, are $(\bm{b},p,\varepsilon)$-holomorphic with
$\varepsilon>0$ _independent_ of $n$.
Step C: $(\bm{b},p,\varepsilon)$-holomorphy of $\mathcal{A}$. Now, fix
$\tau\in\mathcal{G}$, and consider
$\mathcal{A}_{\tau}^{(n)}\colon\mathcal{O}_{\bm{\rho}}\to\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)$
with $\mathcal{A}_{\tau}^{(n)}$ as defined in Eq. 3.47. Then, A and A’ show
that $\mathcal{A}_{\tau}^{(n)}$ converges uniformly to $\mathcal{A}_{\tau}$ on
$\mathcal{O}_{\bm{\rho}}$ with $\mathcal{A}_{\tau}$ as in Eq. 3.37. Step B
shows that $\mathcal{A}_{\tau}^{(n)}$ is $(\bm{b},p,\varepsilon)$-holomorphic
with $\varepsilon$ independent of $n$. Thus, Lemma 2.4 yields that
$\mathcal{A}_{\tau}$ is $(\bm{b},p,\varepsilon)$-holomorphic.
In the same way as in Step C in the proof of Theorem 3.1, we now conclude that
(3.61) $\mathcal{A}=\sum_{\tau\in\mathcal{G}}\mathcal{A}_{\tau}$
with $\mathcal{A}$ as in Eq. 3.29 is $(\bm{b},p,\varepsilon)$-holomorphic for
some $\varepsilon>0$, which is the assertion.
### 3.3 Parametric Holomorphy of the Parameter–to–solution Map
A key consequence of the presently obtained parametric holomorphy result of
the parameter-to-operator map Eq. 1.3 is that of the parameter-to-solution map
Eq. 1.4.
###### Corollary 3.5.
Let 2.8 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and $p\in(0,1)$, and
let the assumptions of Theorem 3.3 be fulfilled as well. In addition, assume
that
$\hat{\mathsf{A}}_{\bm{{y}}}\in\mathscr{L}_{{\text{iso}}}(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma}))$
for each ${\bm{{y}}}\in\mathbb{U}$ and that the map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{f}_{{\bm{{y}}}}\in L^{2}(\hat{\Gamma})$ is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous for some $\varepsilon>0$.
Then there exists $\varepsilon^{\prime}>0$ such that the parameter-to-solution
map
(3.62) $\mathcal{S}:\mathbb{U}\to L^{2}(\hat{\Gamma}):{\bm{{y}}}\mapsto
u_{\bm{{y}}}\mathrel{\mathrel{\mathop{:}}=}\hat{\mathsf{A}}_{{\bm{{y}}}}^{-1}\hat{f}_{\bm{{y}}},$
is $(\bm{b},p,\varepsilon^{\prime})$-holomorphic and continuous.
###### Proof 3.6.
The assertion then follows as a direct consequence from Item 2 and Item 3 in
Proposition 2.5.
## 4 Applications to Boundary Integral Equations on Parametric Boundaries
In this section, we explore the applicability of the main result of Section 3,
i.e. Theorem 3.3, to the boundary integral operators obtained from the
boundary reduction of the sound-soft Helmholtz scattering problem.
### 4.1 Model Problem: Sound-Soft Acoustic Scattering
Let ${\text{D}}\subset\mathbb{R}^{3}$ be a bounded Lipschitz domain with
boundary $\Gamma:=\partial{\text{D}}$ and denote by
${\text{D}}^{{\text{c}}}:=\mathbb{R}^{3}\backslash\overline{{\text{D}}}$ the
corresponding exterior domain. Given $\kappa>0$ and
$\bm{\hat{d}}_{\text{inc}}\in\mathbb{S}^{2}\coloneqq\\{{\bf
x}\in\mathbb{R}^{3}:\left\lVert{\bf x}\right\rVert=1\\}$, we define an
incident plane wave $u^{\text{inc}}({\bf x})\coloneqq\exp(\imath\kappa{\bf
x}\cdot\bm{\hat{d}}_{\text{inc}})$ and aim to find the scattered wave
$u^{\text{scat}}\in H^{1}_{{\text{loc}}}({\text{D}}^{{\text{c}}})$ such that
the total field $u\coloneqq u^{\text{inc}}+u^{\text{scat}}$ satisfies
(4.1a) $\displaystyle\Delta u+\kappa^{2}u$ $\displaystyle=0,\quad\text{in
}{\text{D}}^{{\text{c}}},$ (4.1b) $\displaystyle u$
$\displaystyle=0,\quad\text{on }\Gamma,$
and the scattered field additionally satisfies the Sommerfeld radiation
conditions
(4.2) $\displaystyle\frac{\partial u^{\text{scat}}}{\partial r}({\bf
x})-\imath\kappa u^{\text{scat}}({\bf x})=o\left(r^{-1}\right)$
as $r\coloneqq\left\lVert{\bf x}\right\rVert\rightarrow\infty$, uniformly in
$\hat{{\bf x}}:={\bf x}/r$. Thus, since $u^{\text{inc}}$ satisfies Eq. 4.1a on
its own, we seek $u^{\text{scat}}\in
H_{{\text{loc}}}^{1}({\text{D}}^{{\text{c}}})$ such that
(4.3a) $\displaystyle\Delta u^{\text{scat}}+\kappa^{2}u^{\text{scat}}$
$\displaystyle=0,\quad\text{in }{\text{D}}^{{\text{c}}},$ (4.3b)
$\displaystyle u^{\text{scat}}$ $\displaystyle=-u^{\text{inc}},\quad\text{on
}\Gamma,$
and Eq. 4.2 holds.
### 4.2 Boundary Integral Operators
We proceed to perform a boundary reduction of Eq. 4.2 and Eq. 4.3 by means of
boundary potentials in order to obtain a equivalent boundary integral
formulation. To this end, we first introduce the Green’s function of the
Helmholtz operator with leading frequency $\kappa\in\mathbb{R}$ in
$\mathbb{R}^{3}$
(4.4) $\text{G}^{(\kappa)}({\bf
z})\coloneqq\frac{\exp(\imath\kappa\left\lVert{\bf
z}\right\rVert)}{4\pi\left\lVert{\bf z}\right\rVert},\quad{\bf
z}\in\mathbb{R}^{3}\backslash\\{{\bf 0}\\},$
with $\imath^{2}=-1$ being the imaginary complex unit. Next, we define the
acoustic single layer potential
$\mathcal{S}_{\Gamma}^{(\kappa)}:H^{-\frac{1}{2}}(\Gamma)\rightarrow
H^{1}_{{\text{loc}}}(\Delta,{\text{D}}^{c})$
$\left(\mathcal{S}^{(\kappa)}_{\Gamma}\varphi\right)({\bf
x})\mathrel{\mathrel{\mathop{:}}=}\int\limits_{\Gamma}\text{G}^{(\kappa)}({\bf
x}-{\bf y})\varphi({\bf y})\text{d}\text{s}_{\bf y},\quad{\bf
x}\in{\text{D}}^{{\text{c}}},$
and the double layer potential
$\mathcal{D}_{\Gamma}^{(\kappa)}:H^{\frac{1}{2}}(\Gamma)\rightarrow
H^{1}_{{\text{loc}}}(\Delta,{\text{D}}^{{\text{c}}})$
$\left(\mathcal{D}_{\Gamma}^{(\kappa)}\phi\right)({\bf
x})\mathrel{\mathrel{\mathop{:}}=}\int\limits_{\Gamma}\frac{\partial\text{G}^{(\kappa)}({\bf
x}-{\bf y})}{\partial\mathbf{n}_{\Gamma}({\bf y})}\phi({\bf
y})\text{d}\text{s}_{\bf y},\quad{\bf x}\in{\text{D}}^{{\text{c}}}.$
By $\mathbf{n}_{\Gamma}({\bf y})$ we denote the outward pointing normal on
${\bf y}\in\Gamma$. Both operators are continuous and satisfy Eq. 4.3a and the
Sommerfeld radiation condition Eq. 4.2, see, e.g., [48].
By taking exterior Dirichlet and Neumann traces
$\gamma_{\Gamma}^{{\text{c}}}\colon
H^{1}_{{\text{loc}}}({\text{D}}^{{\text{c}}})\to
H^{\frac{1}{2}}(\Gamma)\quad\text{and}\quad\frac{\partial}{\partial\mathbf{n}_{\Gamma}}\colon
H^{1}(\Delta,{\text{D}}^{{\text{c}}})\to H^{-\frac{1}{2}}(\Gamma)$
we obtain the single layer operator
(4.5)
$\displaystyle\mathsf{V}^{(\kappa)}_{\Gamma}\mathrel{\mathrel{\mathop{:}}=}\gamma_{\Gamma}^{{\text{c}}}\mathcal{S}^{(\kappa)}_{\Gamma}\colon
H^{-\frac{1}{2}}(\Gamma)\to H^{\frac{1}{2}}(\Gamma),$
the double layer operator
(4.6)
$\displaystyle\frac{1}{2}\mathsf{Id}+\mathsf{K}^{(\kappa)}_{\Gamma}\mathrel{\mathrel{\mathop{:}}=}\gamma_{\Gamma}^{{\text{c}}}\mathcal{D}^{(\kappa)}_{\Gamma}\colon
H^{\frac{1}{2}}(\Gamma)\to H^{\frac{1}{2}}(\Gamma),$
and the adjoint double layer operator
(4.7)
$\displaystyle\frac{1}{2}\mathsf{Id}-\mathsf{K}^{(\kappa)^{\prime}}_{\Gamma}\mathrel{\mathrel{\mathop{:}}=}\frac{\partial}{\partial\mathbf{n}_{\Gamma}}\mathcal{S}^{(\kappa)}_{\Gamma}\colon
H^{-\frac{1}{2}}(\Gamma)\to H^{-\frac{1}{2}}(\Gamma).$
The double layer operator and the adjoint double layer operator are indeed
adjoint operators in the bilinear $L^{2}(\Gamma)$-duality product [48].
### 4.3 Combined Boundary Integral Formulation
By means of the introduced potentials and boundary integral operators, the
sound-soft scattering problem from Section 4.1 can be cast into an equivalent
boundary integral formulation. Here, we explore two widely well-know
approaches.
#### 4.3.1 Direct Boundary Integral Formulation
According to Green’s representation formula, the total field
$u:{\text{D}}^{{\text{c}}}\rightarrow\mathbb{C}$ admits the expression
(4.8) $\displaystyle u({\bf x})=u^{\text{inc}}({\bf
x})-\mathcal{S}^{(\kappa)}_{\Gamma}\left(\frac{\partial
u}{\partial\mathbf{n}_{\Gamma}}\right)({\bf x}),\quad{\bf
x}\in{\text{D}}^{{\text{c}}}.$
The unknown boundary data $\frac{\partial u}{\partial\mathbf{n}_{\Gamma}}$
can, for example, be obtained by applying Dirichlet and Neumann traces to Eq.
4.8 and solving
(4.9a) $\displaystyle\mathsf{V}^{(\kappa)}_{\Gamma}\frac{\partial
u}{\partial\mathbf{n}_{\Gamma}}$
$\displaystyle=\gamma^{{\text{c}}}_{\Gamma}u^{\text{inc}},\quad\text{on}\;\Gamma,\quad\text{and},$
(4.9b)
$\displaystyle\left(\frac{1}{2}\mathsf{Id}+\mathsf{K}^{(\kappa)^{\prime}}_{\Gamma}\right)\frac{\partial
u}{\partial\mathbf{n}_{\Gamma}}$ $\displaystyle=\frac{\partial
u^{\text{inc}}}{\partial\mathbf{n}_{\Gamma}},\quad\text{on}\;\Gamma.$
However, well–posedness of Eq. 4.9a and Eq. 4.9b collapses when $\kappa^{2}$
corresponds to a Dirichlet or Neumann eigenvalue of the Laplacian in D,
respectively. A fix to this issue consists in taking a suitable linear
combination of Eq. 4.9a and Eq. 4.9b. Provided
$\eta\in\mathbb{R}\backslash\\{0\\}$, the so–called coupling parameter, we
define
(4.10)
$\displaystyle\mathsf{A}^{(\kappa,\eta)^{\prime}}_{\Gamma}\coloneqq\frac{1}{2}\mathsf{Id}+\mathsf{K}^{(\kappa)^{\prime}}_{\Gamma}-\imath\eta\mathsf{V}^{(\kappa)}_{\Gamma}.$
Exploiting that $\frac{\partial u}{\partial\mathbf{n}_{\Gamma}}\in
L^{2}(\Gamma)$, see, e.g., [40], a new boundary integral formulation of Eq.
4.1a using the direct approach reads as follows.
###### Problem 4.1.
Seek $\phi\coloneqq\frac{\partial u}{\partial\mathbf{n}_{\Gamma}}\in
L^{2}(\Gamma)$ such that
(4.11)
$\displaystyle\mathsf{A}^{(\kappa,\eta)^{\prime}}_{\Gamma}\phi=\frac{\partial
u^{\text{inc}}}{\partial\mathbf{n}_{\Gamma}}-\imath\eta\gamma^{{\text{c}}}_{\Gamma}u^{\text{inc}}\in
L^{2}(\Gamma).$
Thus, one may establish that for each $\kappa\in\mathbb{R}_{+}$ it holds
$\mathsf{A}^{(\kappa,\eta)^{\prime}}_{\Gamma}\in\mathscr{L}_{{\text{iso}}}(L^{2}(\Gamma),L^{2}(\Gamma)),$
thereby implying that 4.1 is well–posed as an operator equation set in
$L^{2}(\Gamma)$.
#### 4.3.2 Indirect Boundary Integral Formulation
Another approach to obtain a boundary integral formulation for the sound-soft
scattering problem consists in expressing $u$ by means of a combined single
and double layer potential, namely we use the following _ansatz_ to represent
$u\in H^{1}_{{{\text{loc}}}}({\text{D}}^{{\text{c}}})$:
(4.12) $\displaystyle u({\bf x})=u^{\text{inc}}({\bf
x})+\left(\mathcal{D}^{(\kappa)}_{\Gamma}\varphi\right)({\bf
x})-\imath\eta\,\left(\mathcal{S}^{(\kappa)}_{\Gamma}\varphi\right)({\bf
x}),\quad{\bf x}\in{\text{D}}^{{\text{c}}},$
where $\varphi\in L^{2}(\Gamma)$ is an unknown boundary density and
$\eta\in\mathbb{R}\backslash\\{0\\}$ is the coupling parameter. Application of
the exterior Dirichlet trace to Eq. 4.12 leads us to the following equivalent
boundary integral formulation for Eq. 4.1a:
###### Problem 4.2.
We seek $\varphi\in L^{2}(\Gamma)$ such that
(4.13)
$\displaystyle\mathsf{A}^{(\kappa,\eta)}_{\Gamma}\varphi=-\gamma_{\Gamma}^{{\text{c}}}u^{\text{inc}}\in
L^{2}(\Gamma),$
where
(4.14)
$\displaystyle\mathsf{A}^{(\kappa,\eta)}_{\Gamma}\coloneqq\frac{1}{2}\mathsf{Id}+\mathsf{K}^{(\kappa)}_{\Gamma}-\imath\eta\mathsf{V}^{(\kappa)}_{\Gamma}.$
We readily remark that the operators Eq. 4.10 and Eq. 4.14 are adjoint to each
other in the bilinear $L^{2}(\Gamma)$-duality pairing. Thus, using similar
tools to the ones as for the direct approach in 4.1, one can show that for
each $\kappa\in\mathbb{R}_{+}$ and $\eta\in\mathbb{R}\backslash\\{0\\}$ it
holds
$\mathsf{A}^{(\kappa,\eta)}_{\Gamma}\in\mathscr{L}_{{\text{iso}}}(L^{2}(\Gamma),L^{2}(\Gamma)).$
In the following, to show shape holomorphy of the boundary combined boundary
integral operators Eq. 4.10 and Eq. 4.14 from 4.1 and 4.2. To this end, we
first show shape holomorphy for each of the involved operators Eqs. 4.5, 4.6
and 4.7.
### 4.4 Shape Parametric Holomorphy of the Single Layer Operator
For each ${\bm{{y}}}\in\mathbb{U}$ we set
(4.15)
$\displaystyle\hat{\mathsf{V}}^{(\kappa)}_{\bm{{y}}}\coloneqq\tau_{\bm{{y}}}\,\mathsf{V}^{(\kappa)}_{\Gamma_{\bm{{y}}}}\,\tau^{-1}_{\bm{{y}}}\in\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right).$
The mapping property stated in Eq. 4.15 follows straightforwardly from Lemma
2.16 and the fact that
$\mathsf{V}^{(\kappa)}_{\Gamma_{\bm{{y}}}}:L^{2}({\Gamma}_{\bm{{y}}})\rightarrow
L^{2}({\Gamma}_{\bm{{y}}})$ defines a bounded linear operator, see e.g. [7,
Theorem 3]. The following results establishes the parametric holomorphy of the
single layer BIO.
###### Theorem 4.3.
Let 2.8 be satisfied. Then the map
(4.16)
$\displaystyle\mathcal{V}^{(\kappa)}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{V}}^{(\kappa)}_{\bm{{y}}}$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 4.4.
We first construct a complex extension of
$\hat{\mathsf{V}}^{(\kappa)}_{\bm{{y}}}$ and then show that this extension
satisfies all assumptions of Theorem 3.3. To this end, we note that the single
layer operator admits the representation Eq. 1.2 with
(4.17) $\displaystyle\mathsf{a}({\bf y},{\bf
z})=\frac{\exp(\imath\kappa\left\lVert{\bf
z}\right\rVert)}{4\pi\left\lVert{\bf z}\right\rVert},\quad{\bf
z}\in\mathbb{R}^{3}\backslash\\{{\bf 0}\\}.$
Thus, setting $\left\lVert{\bf z}\right\rVert_{\mathbb{C}}\coloneqq\sqrt{{\bf
z}\cdot{\bf z}}$ with $\sqrt{\cdot}$ signifiying the principal branch of the
square root yields a complex extension
(4.18) $\displaystyle\mathsf{a}({\bf y},{\bf
z})=\frac{\exp(\imath\kappa\left\lVert{\bf
z}\right\rVert_{\mathbb{C}})}{4\pi\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}}$
for all ${\bm{z}}\in\mathfrak{U}$. Thus, it remains to verify each item of the
assumptions of Theorem 3.3. As per usual, let
$\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ be any sequence of numbers strictly
larger than one that is $(\bm{b},\varepsilon)$-admissible.
Item 1:
Our first observation is that $\mathsf{a}(\cdot,\cdot)$ as in Eq. 4.18 is
independent of the first argument. Consequently, in this particular case, for
each fixed ${\bf z}\in\mathfrak{U}$ the map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\mathsf{a}\left(\bm{r}_{\bm{{y}}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),{\bf
z}\right)\in\mathbb{C}$ is trivially $(\bm{b},p,\varepsilon)$-holomorphic in
each direction, as it is constant.
Item 2:
We first note that the principal branch of the square root is holomorphic in
$\mathbb{C}\backslash(-\infty,0]$ and that the map ${\bf z}\mapsto{\bf
z}\cdot{\bf z}$ is a multivariate polynomial, and thus holomorphic. Thus, the
map $\mathfrak{U}\ni{\bf z}\mapsto\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}\in\mathbb{C}$ is holomorphic. Moreover, the
function $t\mapsto\exp(\imath\kappa t)$ is entire and thus holomorphic and
$t\mapsto\frac{1}{t}$ is holmorphic as well, except at $t=0$. Consequently,
for any ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ the map
(4.19) $\mathfrak{U}\ni{\bf
z}\mapsto\mathsf{a}\left(\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),{\bf
z}\right)\in\mathbb{C}$
is holomorphic.
Item 3:
For each $\tau\in\mathcal{G}$, $\tilde{\mathbf{u}}\in\tilde{\tau}$,
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$, and $\hat{{\bf x}}\in\hat{\Gamma}$ and
abbreviating ${\bf z}=\bm{r}_{\bm{z}}(\hat{{\bf
x}})-\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)$ one has
(4.20)
$\mathsf{a}\left(\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right),{\bf
z}\right)=\frac{\exp(\imath\kappa\Re\\{\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}\\}-\kappa\Im\\{\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}\\})}{4\pi\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}}.$
Observing that $\left\lvert\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}\right\rvert=\left\lVert{\bf z}\right\rVert$ yields
(4.21) $\left\lvert\mathsf{a}({\bf y},{\bf
z})\right\rvert\leq\frac{\exp(\kappa\left\lvert\Im\\{\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}\\}\right\rvert)}{4\pi\left\lvert\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}\right\rvert}\leq\frac{\exp(\kappa\left\lVert{\bf
z}\right\rVert)}{4\pi\left\lVert{\bf z}\right\rVert}.$
We remark that at this point the bound in Eq. 4.21 does not completely fulfil
Item 3 in the assumptions of Theorem 3.3. It follows from the definition of
$\mathcal{O}_{\bm{\rho}}$ that for each ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$
there exists ${\bm{{y}}}\in\mathbb{U}$ such that $\left\lvert
z_{j}-y_{j}\right\rvert\leq\rho_{j}-1$, for all $j\in\mathbb{N}$ Next, we
observe that for any $\hat{{\bf x}},\hat{{\bf y}}\in\hat{\Gamma}$ one has
(4.22) $\displaystyle\left\lVert\bm{r}_{\bm{z}}(\hat{{\bf
x}})-\bm{r}_{\bm{z}}(\hat{{\bf y}})\right\rVert\leq$
$\displaystyle\left\lVert\bm{r}_{\bm{z}}(\hat{{\bf
x}})-\bm{r}_{\bm{{y}}}(\hat{{\bf
x}})\right\rVert+\left\lVert\bm{r}_{\bm{z}}(\hat{{\bf
y}})-\bm{r}_{\bm{{y}}}(\hat{{\bf y}})\right\rVert$
$\displaystyle+\left\lVert\bm{r}_{\bm{{y}}}(\hat{{\bf
x}})-\bm{r}_{\bm{{y}}}(\hat{{\bf y}})\right\rVert.$
It follows from Eq. 2.13 that for any $\hat{{\bf x}}\in\hat{\Gamma}$
(4.23) $\displaystyle\left\lVert\bm{r}_{\bm{z}}(\hat{{\bf
x}})-\bm{r}_{\bm{{y}}}(\hat{{\bf x}})\right\rVert$
$\displaystyle\leq\sum_{j\geq 1}\left\lvert
z_{j}-y_{j}\right\rvert\left\lVert{\bm{\varphi}}_{j}\right\rVert_{L^{\infty}(\hat{\Gamma};\mathbb{R}^{3})}$
$\displaystyle\leq\sum_{j\geq
1}\left(\rho_{j}-1\right)\left\lVert{\bm{\varphi}}_{j}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})}$
$\displaystyle\leq\varepsilon,$
where the last inequality comes from the fact that the sequence
$\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ is $(\bm{b},\varepsilon)$-admissible.
In addition, for any $\hat{{\bf x}},\hat{{\bf y}}\in\hat{\Gamma}$ and each
${\bm{{y}}}\in\mathbb{U}$ one has
(4.24) $\left\lVert\bm{r}_{\bm{{y}}}(\hat{{\bf
x}})-\bm{r}_{\bm{{y}}}(\hat{{\bf y}})\right\rVert\leq\left\lVert\hat{{\bf
x}}-\hat{{\bf
y}}\right\rVert\left\lVert\bm{r}_{\bm{{y}}}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})}.$
The map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\bm{r}_{\bm{{y}}}\in\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})$
is continuous according to Lemma 2.12, thus rendering the map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\left\lVert\bm{r}_{\bm{{y}}}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})}$
continuous as well, as any norm depends continuoulsy on its argument. As
pointed out previously, $\mathbb{U}$ defines a compact subset when equipped
with the product topology. Therefore, the map
${\bm{{y}}}\mapsto\left\lVert\bm{r}_{\bm{{y}}}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})}$
attains a maximum. Consequently, by collection Eq. 4.22 Eq. 4.23, and Eq.
4.22, one has that for any $\hat{{\bf x}},\hat{{\bf y}}\in\hat{\Gamma}$ and
each ${\bm{{y}}}\in\mathbb{U}$ it holds
(4.25) $\displaystyle\left\lVert\bm{r}_{\bm{z}}(\hat{{\bf
x}})-\bm{r}_{\bm{z}}(\hat{{\bf y}})\right\rVert\leq
2\varepsilon+\text{diam}(\hat{\Gamma})\,\max_{{\bm{{y}}}\in\mathbb{U}}\left\lVert\bm{r}_{\bm{{y}}}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})}.$
Hence, with
(4.26)
$C_{\mathsf{a}}\coloneqq\exp\left(\kappa\left(2\varepsilon+\text{diam}(\hat{\Gamma})\,\max_{{\bm{{y}}}\in\mathbb{U}}\left\lVert\bm{r}_{\bm{{y}}}\right\rVert_{\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})}\right)\right)$
$\mathsf{a}(\cdot,\cdot)$ as in Eq. 4.20 satisfies Item 3 in the assumptions
of Theorem 3.3.
Therefore, it follows from Theorem 3.3 that the map in Eq. 4.16 is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous.
### 4.5 Shape Holomorphy of the Double Layer Operator
For each ${\bm{{y}}}\in\mathbb{U}$ we define
(4.27)
$\displaystyle\hat{\mathsf{K}}_{\bm{{y}}}^{(\kappa)}\coloneqq\tau_{\bm{{y}}}\;\mathsf{K}_{\Gamma_{\bm{{y}}}}^{(\kappa)}\tau^{-1}_{\bm{{y}}}\in\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right).$
Again, as argued previously for the single layer operator, the mapping
property stated in Eq. 4.27 follows straightforwardly from Lemma 2.16 and the
fact that for each ${\bm{{y}}}\in\mathbb{U}$ one has
$\mathsf{K}_{\Gamma_{\bm{{y}}}}^{(\kappa)}:L^{2}(\Gamma_{\bm{{y}}})\rightarrow
L^{2}(\Gamma_{\bm{{y}}})$ defines a bounded linear operator, see e.g. [7,
Theorem 3.5]
However, unlike in the analysis for the single layer operator, we can not
directly argue using solely 2.8 and Theorem 4.3 as the term $\frac{\partial
G^{(\kappa)}}{\partial\mathbf{n}}$ does not yield an integrable kernel when
considering the double layer operator set on a Lipschitz boundary. The latter
difficulty can be overcome by assuming that all surfaces are piecewise
$\mathscr{C}^{1,1}$ as in the following assumption.
###### Assumption 4.5.
There is a decomposition $\mathcal{G}$ of $\hat{\Gamma}$ such that for each
${\bm{{y}}}\in\mathbb{U}$ and each ${\tau}\in\mathcal{G}$ one has that
${{\bm{r}}}_{\bm{{y}}}\circ\chi_{\tau}\in\mathscr{C}^{1,1}(\tilde{\tau};\mathbb{R}^{3})$.
Next, we proceed to prove the parametric holomorphy result for the double
layer BIO.
###### Theorem 4.6.
Let 2.8 and 4.5 be fulfilled with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Then there exists $\varepsilon>0$ such that the map
(4.28)
$\displaystyle\mathcal{K}^{(\kappa)}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{K}}_{\bm{{y}}}^{(\kappa)}$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 4.7.
As in the analysis of the parametric holomorphy result for single layer
operator, we start by constructing a complex extension of
$\hat{\mathsf{K}}_{{\bm{{y}}}}$. To this end, we start by remarking that the
double layer operator has the representation Eq. 1.2 with
(4.29) $\displaystyle\mathsf{a}({\bf y},{\bf z})=\big{(}{\bf n}({\bf
y})\cdot{\bf z}\big{)}\frac{\exp(\imath\kappa\left\lVert{\bf
z}\right\rVert)}{4\pi\left\lVert{\bf
z}\right\rVert^{3}}\left(\imath\kappa\left\lVert{\bf
z}\right\rVert-1\right),\quad{\bf y}\in\hat{\Gamma},\quad{\bf
z}\in\mathbb{R}^{3}\backslash\\{{\bf 0}\\}.$
Thus, for each $\tau\in\mathcal{G}$, $\tilde{\mathbf{u}}\in\tilde{\tau}$,
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$, and $\hat{{\bf x}}\in\hat{\Gamma}$ and
abbreviating ${\bf z}=\hat{{\bf
x}}-\bm{r}_{\bm{z}}\circ\chi_{\tau}\left(\tilde{\mathbf{u}}\right)$ we obtain
the complex extension
(4.30) $\displaystyle\mathsf{a}(\bm{r}_{\bm{z}}(\hat{\mathbf{x}}),{\bf
z})=\big{(}\hat{\mathbf{n}}_{\bm{z}}(\hat{\mathbf{x}})\cdot{\bf
z}\big{)}\frac{\exp(\imath\kappa\left\lVert{\bf
z}\right\rVert_{\mathbb{C}})}{4\pi\left\lVert{\bf
z}\right\rVert^{3}_{\mathbb{C}}}\left(\imath\kappa\left\lVert{\bf
z}\right\rVert_{\mathbb{C}}-1\right),\quad{\bf z}\in\mathfrak{U}.$
As per usual, let $\bm{\rho}\coloneqq(\rho_{j})_{j\geq 1}$ be any sequence of
numbers strictly larger than one that is $(\bm{b},\varepsilon)$-admissible. We
verify each item of the assumptions of Theorem 3.3.
Item 1:
Follows directly from Lemma 2.18.
Item 2:
As in the proof of Theorem 4.3, the maps $\mathfrak{U}\ni{\bf
z}\mapsto\left\lVert{\bf z}\right\rVert_{\mathbb{C}}\in\mathbb{C}$ and
$\mathbb{C}\setminus\\{0\\}\ni t\mapsto\frac{1}{t}\in\mathbb{C}$ are
holomorphic. Consequently, for any ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ the
map $\mathfrak{U}\ni{\bf
z}\mapsto\mathsf{a}\left(\bm{r}_{\bm{z}}(\hat{\mathbf{x}}),{\bf
z}\right)\in\mathbb{C}$ is holomorphic as well.
Item 3:
We first need to address the non-integrability of $\frac{\partial
G^{(\kappa)}}{\partial\mathbf{n}}$. To this end, it is well known that for
each $\tau\in\mathcal{G}$, $\hat{\bf x},\hat{\bf y}\in\tau$, and
${\bm{{y}}}\in\mathbb{U}$ it holds
(4.31) $\displaystyle\underline{c}\left\lvert\hat{\bf n}_{\bm{{y}}}(\hat{\bf
y})\cdot\left(\bm{r}_{\bm{{y}}}(\hat{\bf x})-\bm{r}_{\bm{{y}}}(\hat{\bf
y})\right)\right\rvert\leq\left\lVert\hat{\bf x}-\hat{\bf
y}\right\rVert^{2}\leq\overline{c}\left\lVert\bm{r}_{\bm{{y}}}(\hat{\bf
x})-\bm{r}_{\bm{{y}}}(\hat{\bf y})\right\rVert^{2},$
for some $0<\underline{c}\leq\overline{c}<\infty$, see, e.g., [48, Lemma
2.2.14]. We note particularly that the proof together with our assumptions
guarantee that the constant can be chosen independently of ${\bm{{y}}}$. By a
perturbation argument we can conclude that there is $\varepsilon>0$ and two
possibly different constants $0<\underline{c}\leq\overline{c}<\infty$ such
that
(4.32) $\displaystyle\underline{c}\left\lvert\hat{\bf n}_{\bm{z}}(\hat{\bf
y})\cdot\left(\bm{r}_{\bm{z}}(\hat{\bf x})-\bm{r}_{\bm{z}}(\hat{\bf
y})\right)\right\rvert\leq\left\lVert\hat{\bf x}-\hat{\bf
y}\right\rVert^{2}\leq\overline{c}\left\lVert\bm{r}_{\bm{z}}(\hat{\bf
x})-\bm{r}_{\bm{z}}(\hat{\bf y})\right\rVert^{2},$
for all ${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$. Abbreviating ${\bf
z}=\bm{r}_{\bm{z}}(\hat{{\bf x}})-\bm{r}_{\bm{z}}(\hat{\bf y})$ and proceeding
as in the proof of Theorem 4.3 implies
$|\mathsf{a}\left(\bm{r}_{\bm{z}}(\hat{\bf y}),{\bf
z}\right)|\leq\frac{\overline{c}}{\underline{c}}\frac{\exp(\kappa\left\lVert{\bf
z}\right\rVert)}{4\pi\left\lVert{\bf
z}\right\rVert}\left(\kappa\left\lVert{\bf z}\right\rVert+1\right),$
i.e., the singularity of the complex extension of the double layer operator
behaves similarly to the single layer operator case. The conclusion follows in
complete analogy to the proof of Theorem 4.3.
Finally, recalling Theorem 3.3 one may conclude that the map introduced in Eq.
4.28 is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
### 4.6 Shape Holomorphy of the Adjoint Double Layer Operator
In order to establish the parametric holomorphy result for the adjoint double
layer operator from the double layer operator itself, we consider for each
${\bm{{y}}}\in\mathbb{U}$ the bilinear $L^{2}(\Gamma_{\bm{{y}}})$-inner
product $(\cdot,\cdot)_{L^{2}(\Gamma_{\bm{{y}}})}$. For all $\varphi,\phi\in
L^{2}(\Gamma_{\bm{{y}}})$ it holds that
(4.33)
$\displaystyle\left({\mathsf{K}}^{(\kappa)}_{\Gamma_{\bm{{y}}}}\varphi,\phi\right)_{L^{2}(\Gamma_{\bm{{y}}})}=\left(\varphi,{\mathsf{K}}^{(\kappa)^{\prime}}_{\Gamma_{\bm{{y}}}}\phi\right)_{L^{2}(\Gamma_{\bm{{y}}})}.$
For each ${\bm{{y}}}\in\mathbb{U}$ we may define
(4.34)
$\displaystyle\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{{\bm{{y}}}}\coloneqq\tau_{\bm{{y}}}\;{\mathsf{K}}^{(\kappa)^{\prime}}_{\Gamma_{\bm{{y}}}}\;\tau^{-1}_{\bm{{y}}}\in\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right),$
where the mapping properties are immediate. In fact, defining a modified,
real-valued inner product on $L^{2}(\hat{\Gamma})$ by
(4.35) $\displaystyle\left(\varphi,\phi\right)_{L^{2}(\Gamma_{\bm{{y}}})}$
$\displaystyle=\int\limits_{\Gamma_{\bm{{y}}}}\varphi({\bf y})\phi({\bf
y})\text{d}s_{{\bf y}}$
$\displaystyle=\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\left(\tau_{\bm{{y}}}\varphi\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\left(\tau_{\bm{{y}}}\phi\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\sqrt{g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})}\text{d}\tilde{\mathbf{u}}$
$\displaystyle\eqqcolon\left(\hat{\varphi},\hat{\phi}\right)_{L^{2}(\hat{\Gamma}),{\bm{{y}}}},$
with $\hat{\varphi}=\tau_{\bm{{y}}}\varphi,\hat{\phi}=\tau_{\bm{{y}}}\phi\in
L^{2}(\hat{\Gamma})$ and ${\bm{{y}}}\in\mathbb{U}$, one readily verifies that
$\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{{\bm{{y}}}}$ is indeed the adjoint
operator of $\hat{\mathsf{K}}^{(\kappa)}_{{\bm{{y}}}}$ with respect to
$\left(\cdot,\cdot\right)_{L^{2}(\hat{\Gamma}),{\bm{{y}}}}$. I.e., for each
${\bm{{y}}}\in\mathbb{U}$ and any $\hat{\varphi},\hat{\phi}\in
L^{2}(\hat{\Gamma})$ it holds
(4.36)
$\displaystyle\left(\hat{\mathsf{K}}^{(\kappa)}_{{\bm{{y}}}}\hat{\varphi},\hat{\phi}\right)_{L^{2}(\hat{\Gamma}),{\bm{{y}}}}=\left(\hat{\varphi},\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{{\bm{{y}}}}\hat{\phi}\right)_{L^{2}(\hat{\Gamma}),{\bm{{y}}}}.$
Using Eq. 4.33, parametric holomorphy of the adjoint double layer operator can
be derived from the corresponding result for the double layer operator as
throroughly described in the following result.
###### Theorem 4.8.
Let 2.8 and 4.5 be fulfilled. Then, there exists $\varepsilon>0$ such that the
map
(4.37)
$\displaystyle\mathcal{K}^{(\kappa)^{\prime}}:\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right):{\bm{{y}}}\mapsto\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{\bm{{y}}}$
is $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 4.9.
The complex extension of the modified inner product Eq. 4.35 to
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$ reads
(4.38)
$\displaystyle\left(\hat{\varphi},\hat{\phi}\right)_{L^{2}(\hat{\Gamma}),{\bm{z}}}=\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\left(\hat{\varphi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\left(\hat{\phi}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\sqrt{g_{\tau,{\bm{z}}}(\tilde{\mathbf{u}})}\text{d}\tilde{\mathbf{u}}$
and takes arguments $\hat{\varphi},\hat{\phi}\in L^{2}(\hat{\Gamma})$. Thus,
in accordance with Eq. 4.36, we can extend
$\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{{\bm{{y}}}}$ into
$\mathcal{O}_{\bm{\rho}}$ as the (unique) bounded linear operator satisfying
$\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{{\bm{z}}}:L^{2}(\hat{\Gamma})\rightarrow
L^{2}(\hat{\Gamma})$
(4.39)
$\displaystyle\left(\hat{\mathsf{K}}^{(\kappa)}_{{\bm{z}}}\hat{\varphi},\hat{\phi}\right)_{L^{2}(\hat{\Gamma}),{\bm{z}}}=\left(\hat{\varphi},\hat{\mathsf{K}}^{(\kappa)^{\prime}}_{{\bm{z}}}\hat{\phi}\right)_{L^{2}(\hat{\Gamma}),{\bm{z}}},$
for all $\hat{\varphi},\hat{\phi}\in L^{2}(\hat{\Gamma})$ and
${\bm{z}}\in\mathcal{O}_{\bm{\rho}}$. The map
$\mathbb{U}\ni{\bm{{y}}}\mapsto\left(\cdot,\cdot\right)_{{\bm{{y}}},L^{2}(\hat{\Gamma})}$
into the Banach space of complex-valued bilinear forms on
$L^{2}(\hat{\Gamma})\times L^{2}(\hat{\Gamma})$ is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous as a consequence of Lemma
2.14, and so is
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\mathsf{K}}^{(\kappa)}_{\bm{{y}}}\in\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)$
as stated in Theorem 4.6.
### 4.7 Parametric Shape holomorphy of Combined Field Integral Equations
With the shape holomorphy results of the single, double, and adjoint double
layer operator available, we can now conclude that Eq. 4.10 and Eq. 4.14,
i.e.,
$\mathsf{A}^{(\kappa,\eta)}_{\Gamma}:L^{2}(\Gamma)\rightarrow
L^{2}(\Gamma)\quad\text{and}\quad\mathsf{A}^{(\kappa,\eta)^{\prime}}_{\Gamma}:L^{2}(\Gamma)\rightarrow
L^{2}(\Gamma),$
are shape holomorphic as well.
###### Corollary 4.10.
Let 2.8 and 4.5 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Then there exists $\varepsilon>0$ such that
(4.40a)
$\displaystyle\mathcal{A}^{(\kappa,\eta)}\colon\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)$
$\displaystyle\colon{\bm{{y}}}\mapsto\hat{\mathsf{A}}_{\bm{{y}}}^{(\kappa,\eta)}\coloneqq\tau_{\bm{{y}}}\;\mathsf{A}_{\Gamma_{\bm{{y}}}}^{(\kappa,\eta)}\tau^{-1}_{\bm{{y}}},$
(4.40b)
$\displaystyle\mathcal{A}^{(\kappa,\eta)^{\prime}}\colon\mathbb{U}\rightarrow\mathscr{L}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)$
$\displaystyle\colon{\bm{{y}}}\mapsto\hat{\mathsf{A}}_{\bm{{y}}}^{(\kappa,\eta)^{\prime}}\coloneqq\tau_{\bm{{y}}}\;\mathsf{A}_{\Gamma_{\bm{{y}}}}^{(\kappa,\eta)^{\prime}}\tau^{-1}_{\bm{{y}}},$
are $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 4.11.
This result is a consequence of Theorem 4.3, Theorem 4.6, and Theorem 4.8.
### 4.8 Parametric Holomorphy of the Parameter-to-Solution Map
We are interested in the depedence of the solution to 4.1 and 4.2 upon the
parametric boundary shape. More precisely, we are interested in establishing
the holomorphic _parameter-to-solution_ maps
(4.41)
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\phi}_{\bm{{y}}}\coloneqq\left(\hat{\mathsf{A}}^{(\kappa,\eta)^{\prime}}_{{\bm{{y}}}}\right)^{-1}\hat{f}_{\bm{{y}}}\in
L^{2}(\hat{\Gamma}),$
and
(4.42)
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\varphi}_{\bm{{y}}}\coloneqq\left(\hat{\mathsf{A}}^{(\kappa,\eta)}_{{\bm{{y}}}}\right)^{-1}\hat{g}_{\bm{{y}}}\in
L^{2}(\hat{\Gamma}),$
where
(4.43)
$\displaystyle\hat{f}_{\bm{{y}}}\coloneqq\tau_{\bm{{y}}}{f}_{\Gamma_{\bm{{y}}}}\in
L^{2}(\hat{\Gamma})\quad\text{and}\quad\hat{g}_{\bm{{y}}}\coloneqq\tau_{\bm{{y}}}{g}_{\Gamma_{\bm{{y}}}}\in
L^{2}(\hat{\Gamma}).$
With the results stated in Theorem 4.3, Theorem 4.6, and Theorem 4.8, together
with Proposition 2.5, one can establish the following result.
###### Corollary 4.12.
Let 2.8 and 4.5 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Then there exists $\varepsilon>0$ such that the parameter-to-
solution maps introduced in Eq. 4.41 and Eq. 4.42 are
$(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 4.13.
The proof is in complete analogy to the proof of Corollary 3.5. To this end,
we remark that Corollary 4.10 implies that there exists an $\varepsilon>0$
such that the maps $\mathcal{A}^{(\kappa,\eta)^{\prime}}$ and
$\mathcal{A}^{(\kappa,\eta)}$ are $(\bm{b},p,\varepsilon)$-holomorphic and
continuous. Moreover, it holds
$\mathsf{A}^{(\kappa,\eta)^{\prime}}_{\Gamma},\mathsf{A}^{(\kappa,\eta)}_{\Gamma}\in\mathscr{L}_{{\text{iso}}}(\Gamma_{\bm{{y}}},\Gamma_{\bm{{y}}})$
for all ${\bm{{y}}}\in\mathbb{U}$. Thus, Lemma 2.16 implies that
$\hat{\mathsf{A}}_{\bm{{y}}}^{(\kappa,\eta)},\;\hat{\mathsf{A}}_{\bm{{y}}}^{(\kappa,\eta)^{\prime}}\in\mathscr{L}_{{\text{iso}}}\left(L^{2}(\hat{\Gamma}),L^{2}(\hat{\Gamma})\right)$
for each ${\bm{{y}}}\in\mathbb{U}$.
It only remains to show the parametric holomorphy of
${\bm{{y}}}\mapsto\hat{f}_{\bm{{y}}}$ and
${\bm{{y}}}\mapsto\hat{g}_{\bm{{y}}}$. To this end, in 4.1, the right-hand
side is given through a plane incident wave $u^{\text{i}}({\bf
x}):=\exp(\imath\kappa{\bf x}\cdot\bm{\hat{d}}_{\text{inc}})$, yielding
(4.44) $\hat{f}_{\bm{{y}}}(\hat{{\bf
x}})=\kappa\hat{\mathbf{n}}_{\bm{{y}}}(\hat{{\bf
x}})\cdot\bm{\hat{d}}_{\text{inc}}\exp\left(\imath\kappa\bm{r}_{\bm{{y}}}(\hat{{\bf
x}})\cdot\bm{\hat{d}}_{\text{inc}}\right)-\imath\eta\exp\left(\imath\kappa\bm{r}_{\bm{{y}}}(\hat{{\bf
x}})\cdot\bm{\hat{d}}_{\text{inc}}\right).$
Lemmas 2.12 and 2.18 imply that that
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{f}_{\bm{{y}}}\in L^{2}(\hat{\Gamma})$ is
indeed $(\bm{b},p,\varepsilon)$-holomorphic and continuous for some
$\varepsilon>0$. Similarly, for 4.2, we have
(4.45) $\hat{g}_{\bm{{y}}}(\hat{{\bf
x}})=-\exp\left(\imath\kappa\bm{r}_{\bm{{y}}}(\hat{{\bf
x}})\cdot\bm{\hat{d}}_{\text{inc}}\right),\quad\hat{{\bf x}}\in\hat{\Gamma},$
for which it follows with Lemma 2.12 that
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{g}_{\bm{{y}}}\in L^{2}(\hat{\Gamma})$ is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous for some $\varepsilon>0$.
### 4.9 Parametric Holomorphy of the Far-Field Pattern
For later reference, following [12, Chapter 5], we also consider the far-field
patterns of the single and double layer potentials
$\mathcal{S}^{(\kappa)}_{\Gamma}$ and $\mathcal{D}^{(\kappa)}_{\Gamma}$
introduced in Section 4.2. For $\lambda\in L^{2}(\Gamma)$ and an _observation
direction_ $\bm{d}_{o}\in\mathbb{S}^{2}$ they are defined as
(4.46) $\displaystyle\text{FF}^{(\kappa)}_{\mathcal{S},\Gamma}(\lambda)$
$\displaystyle\coloneqq-\frac{\exp(\imath\pi/4)}{\sqrt{2\pi\kappa}}\int\limits_{\Gamma}\exp(-\imath\kappa\,\bm{d}_{o}\cdot{\bf
y})\lambda({\bf y})\text{d}\text{s}_{{{\bf y}}},$ (4.47)
$\displaystyle\text{FF}^{(\kappa)}_{\mathcal{D},\Gamma}(\lambda)$
$\displaystyle\coloneqq\frac{\exp(\imath\pi/4)\sqrt{\kappa}}{\sqrt{2\pi}}\int\limits_{\Gamma}\left(\bm{d}_{o}\cdot\mathbf{n}_{\Gamma}\right)\exp(-\imath\kappa\,\bm{d}_{o}\cdot{\bf
y})\lambda({\bf y})\text{d}\text{s}_{{{\bf y}}},$
respectively. In the following, we show that these far-field patterns are
shape holomorphic as well. To this end, for each ${\bm{{y}}}\in\mathbb{U}$, we
set
(4.48)
$\displaystyle\hat{\text{FF}}^{(\kappa)}_{\mathcal{S},{\bm{{y}}}}\coloneqq\tau_{\bm{{y}}}\,{\text{FF}}^{(\kappa)}_{\mathcal{S},\Gamma_{\bm{{y}}}}\,\tau^{-1}_{\bm{{y}}}\quad\text{and}\quad\hat{\text{FF}}^{(\kappa)}_{\mathcal{D},{\bm{{y}}}}\coloneqq\tau_{\bm{{y}}}\,{\text{FF}}^{(\kappa)}_{\mathcal{D},\Gamma_{\bm{{y}}}}\,\tau^{-1}_{\bm{{y}}}.$
###### Lemma 4.14.
Let 2.8 and 4.5 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Assume that there exists $\varepsilon^{\prime}>0$ such that the
map $\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\lambda}_{\bm{{y}}}\in
L^{2}(\hat{\Gamma})$ is $(\bm{b},p,\varepsilon^{\prime})$-holomorphic and
continuous. Then there exists $\varepsilon>0$ such that
(4.49)
$\displaystyle\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\text{FF}}^{(\kappa)}_{\mathcal{S},{\bm{{y}}}}\left(\hat{\lambda}_{\bm{{y}}}\right)\in\mathbb{C}\quad\text{and}\quad\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\text{FF}}^{(\kappa)}_{\mathcal{D},{\bm{{y}}}}\left(\hat{\lambda}_{\bm{{y}}}\right)\in\mathbb{C}$
are $(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 4.15.
For each ${\bm{{y}}}\in\mathbb{U}$, the explicit representation of Eq. 4.48 is
$\displaystyle\hat{\text{FF}}^{(\kappa)}_{\mathcal{S},{\bm{{y}}}}(\hat{\lambda}_{\bm{{y}}})$
$\displaystyle=-\frac{\exp(\imath\pi/4)}{\sqrt{2\pi\kappa}}\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\hat{k}^{(\kappa)}_{\mathcal{S},{\bm{{y}}}}(\tilde{\mathbf{u}})\left(\hat{\lambda}_{\bm{{y}}}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\mathrm{d}\tilde{\mathbf{u}},$
$\displaystyle\hat{\text{FF}}^{(\kappa)}_{\mathcal{D},{\bm{{y}}}}(\hat{\lambda}_{\bm{{y}}})$
$\displaystyle=\frac{\exp(\imath\pi/4)\sqrt{\kappa}}{\sqrt{2\pi}}\sum_{\tau\in\mathcal{G}}\int\limits_{\tilde{\tau}}\hat{k}^{(\kappa)}_{\mathcal{D},{\bm{{y}}}}(\tilde{\mathbf{u}})\left(\hat{\lambda}_{\bm{{y}}}\circ\chi_{\tau}\right)(\tilde{\mathbf{u}})\mathrm{d}\tilde{\mathbf{u}},$
where $\hat{\lambda}_{\bm{{y}}}\coloneqq\tau_{\bm{{y}}}\lambda\in
L^{2}(\hat{\Gamma})$ and
$\displaystyle\hat{k}^{(\kappa)}_{\mathcal{S},{\bm{{y}}}}(\tilde{\mathbf{u}})$
$\displaystyle=\exp(-\imath\kappa\,\bm{d}_{o}\cdot\bm{r}_{\bm{{y}}}\circ\chi_{\tau}(\tilde{\mathbf{u}}))\sqrt{g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})}$
$\displaystyle\hat{k}^{(\kappa)}_{\mathcal{D},{\bm{{y}}}}(\tilde{\mathbf{u}})$
$\displaystyle=\left(\bm{d}_{o}\cdot\hat{\mathbf{n}}_{{\bm{{y}}}}\right)(\chi_{\tau}(\tilde{\mathbf{u}}))\exp(-\imath\kappa\,\bm{d}_{o}\cdot\bm{r}_{\bm{{y}}}\circ\chi_{\tau}(\tilde{\mathbf{u}}))\sqrt{g_{\tau,{\bm{{y}}}}(\tilde{\mathbf{u}})}.$
Therein, according to Lemma 2.18 and Lemma 2.14 the maps
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{\mathbf{n}}_{{\bm{{y}}}}\in
L^{\infty}(\hat{\Gamma},\mathbb{R}^{3})$ and
$\mathbb{U}\ni{\bm{{y}}}\mapsto\sqrt{g_{\tau,{\bm{{y}}}}}\in
L^{\infty}(\tilde{\tau})$ are $(\bm{b},p,\varepsilon)$-holomorphic and
continuous, respectively. Consequently, for each $\tau\in\mathcal{G}$ the maps
(4.50)
$\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{k}^{(\kappa)}_{\mathcal{S},{\bm{{y}}}}(\tilde{\mathbf{u}})\in
L^{2}(\widetilde{\tau})\quad\text{and}\quad\mathbb{U}\ni{\bm{{y}}}\mapsto\hat{k}^{(\kappa)}_{\mathcal{D},{\bm{{y}}}}(\tilde{\mathbf{u}})\in
L^{2}(\widetilde{\tau})$
are $(\bm{b},p,\varepsilon)$-holomorphic and continuous, thus rendering the
maps in Eq. 4.49 as well.
## 5 Implications in High-dimensional Approximation
In this section, we explore the implications of the presently established
parametric holomorphy result on examples in sparse polynomial approximation
and reduced order modelling, in Bayesian shape inversion, and in expression
rates of artificial neural networks.
### 5.1 Sparse Polynomial Approximation and Reduced Order Modelling
Our aim in the following is to derive best $n$-term polynomial approximation
results to the parameter-to-solution maps Eq. 4.41 and Eq. 4.42 of 4.1 and 4.2
and discuss its consequences. That is, we aim to approximate
(5.1) $\Phi:\mathbb{U}\rightarrow
L^{2}(\hat{\Gamma}):{\bm{{y}}}\mapsto\hat{\phi}_{\bm{{y}}},\hat{\varphi}_{\bm{{y}}},$
by polynomials of finite degree by truncating the infinite expansion
(5.2)
$\displaystyle\Phi({\bm{{y}}})=\sum_{\bm{\nu}\in\mathcal{F}}c_{\bm{\nu}}P_{\bm{\nu}}({\bm{{y}}})$
where
(5.3)
$\displaystyle\mathcal{F}\coloneqq\left\\{\bm{\nu}\in\mathbb{N}_{0}^{\mathbb{N}}:|\bm{\nu}|<\infty\right\\}$
and $P_{\bm{\nu}}(\bm{y})\coloneqq\prod_{j\geq
1}P_{\nu_{j}}\left(y_{j}\right)$, with $P_{n}$ denoting univariate polynomials
of degree $n$. Aiming to recall a result from [10], we consider Legendre
polynomials $P_{n}=P_{n}^{(q)}$ on the interval $[-1,1]$ with normalization
constraints $\|P_{n}^{(q)}\|_{L^{q}([-1,1])}=1$, $q\in\\{2,\infty\\}$. The
coefficients in Eq. 5.3 are denoted by
$(c_{\bm{\nu}}^{(q)})_{\bm{\nu}\in\mathcal{F}}$ accordingly.
We say that a set $\Lambda\subset\mathcal{F}$ is downward closed if
$\bm{\nu}\in\Lambda$ and $\bm{\mu}\leq\bm{\nu}$ implies $\bm{\mu}\in\Lambda$.
To state the sought result, we define for
$c:=\left(c_{\bm{\nu}}\right)_{\bm{\nu}\in\mathcal{F}}\subset\mathbb{R}$ its
downward closed envelope as
$\mathbf{c}_{\bm{\nu}}\coloneqq\sup_{\bm{\mu}\geq\bm{\nu}}\left\lVert
c_{\bm{\mu}}\right\rVert_{L^{2}(\hat{\Gamma})}$ for $\bm{\nu}\in\mathcal{F}$,
where $\bm{\mu}\geq\bm{\nu}$ signifies that the inequality is satisfied
component wise. It is the smallest monotone decreasing sequence which bounds
$\left(\left\lVert
c_{\bm{\nu}}\right\rVert_{L^{2}(\hat{\Gamma})}\right)_{\bm{\nu}\in\mathcal{F}}$
element wise. In addition, for $p>0$ we define $\ell_{m}^{p}(\mathcal{F})$ as
the space of sequences with downward closed envelope in
$\ell^{p}(\mathcal{F})$.
For $1\leq p\leq\infty$, we denote by $L^{p}(\mathbb{U};V)$ the Lebesgue-
Bochner space consisting of strongly $\varrho$-measurable functions
$f:\mathbb{U}\rightarrow{V}$ satisfying $\|f\|_{L^{p}(\mathbb{U};V)}<\infty$,
with
$\|f\|_{L^{p}(\mathbb{U};V)}:=\begin{cases}\left(\displaystyle\int_{\mathbb{U}}\|f(\bm{y})\|_{{V}}^{p}\mathrm{~{}d}\varrho(\bm{y})\right)^{1/p}&1\leq
p<\infty\\\
\underset{{\bm{{y}}}\in\mathbb{U}}{\operatorname{ess}\sup}\|f(\bm{y})\|_{V}&p=\infty,\end{cases}$
where $\mathrm{d}\varrho(\bm{y})$ signifies the tensor product of the uniform
measure in $[-1,1]$, which in turn defines a measure in $\mathbb{U}$. In
particular, we set $L^{p}(\mathbb{U})=L^{p}(\mathbb{U};\mathbb{R})$, for
$1\leq p\leq\infty$.
The next result, which is an application of the main result presented in [10],
addresses the best $n$-term polynomial approximation of the map in Eq. 5.1.
###### Theorem 5.1 ([10, Theorem 2.2, ff.]).
Let 2.8 and 4.5 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Then, $\left(\left\lVert
c_{\bm{\nu}}^{(q)}\right\rVert_{L^{2}(\hat{\Gamma})}\right)_{\bm{\nu}\in\mathcal{F}}$,
$q\in\\{2,\infty\\}$, from Eq. 5.2 belong to $\ell^{p}_{m}(\mathcal{F})$ and
there exists $C>0$ such that for each $n\in\mathbb{N}$ there are finite,
downward closed sets $\Lambda^{2}_{n},\Lambda^{\infty}_{n}\in\mathcal{F}$ of
cardinality $n$ such that
(5.4)
$\left\lVert\Phi-\sum_{\bm{\nu}\in\Lambda^{\infty}_{n}}c_{\bm{\nu}}^{(\infty)}P_{\bm{\nu}}^{(\infty)}\right\rVert_{L^{\infty}(\mathbb{U};L^{2}(\hat{\Gamma}))}\leq
C(n+1)^{-\left(\frac{1}{p}-1\right)},$
and
(5.5)
$\left\lVert\Phi-\sum_{{\bm{\nu}}\in\Lambda^{2}_{n}}c_{\bm{\nu}}^{(2)}P_{\bm{\nu}}^{(2)}\right\rVert_{L^{2}(\mathbb{U};L^{2}(\hat{\Gamma}))}\leq
C(n+1)^{-\left(\frac{1}{p}-\frac{1}{2}\right)}.$
We explore the implications of our parametric holomorphy result in the context
of reduced order modelling. There, the goal is to find a low-dimensional
affine space approximating the solution manifold
$\displaystyle\mathcal{M}\coloneqq\left\\{\Phi({\bm{{y}}}):\;{\bm{{y}}}\in\mathbb{U}\right\\}$
to Eq. 5.1 up to a given accuracy. A commonly used concept in nonlinear
approximation to quantify the approximation error to such solution manifolds
is the so-called Kolmogorov’s $n$-width. For a compact subset $\mathcal{K}$ of
a Banach space $X$ it is defined for $n\in\mathbb{N}$ as
$\displaystyle
d_{n}(\mathcal{K})_{X}\coloneqq\inf_{\begin{subarray}{c}X_{n}\subset X\\\
\operatorname{dim}\left(X_{n}\right)\leq
n\end{subarray}}\sup_{v\in\mathcal{K}}\inf_{w\in X_{n}}\left\lVert
v-w\right\rVert_{X}.$
As a direct consequence of Theorem 5.1 we obtain the following bound on
Kolmogorov’s $n$-width, giving an upper bound on the minimal number of
dimensions to reach a certain approximation accuracy to the solution manifold.
###### Corollary 5.2.
Let 2.8 and 4.5 be satisfied with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Then
$d_{n}(\mathcal{M})_{L^{2}(\hat{\Gamma})}\leq
C(n+1)^{-\left(\frac{1}{p}-1\right)}.$
### 5.2 Bayesian Shape Inversion
We proceed to discuss the significance of the parametric holomorphy results in
the context of computational Bayesian shape inversion.
#### 5.2.1 Bayesian Inverse Problems
Provided a _parameter–to–solution map_ $G:\mathbb{U}\rightarrow X$ with $X$
being a separable Banach space over $\mathbb{C}$ and a _prior_ probability
measure $\mu_{0}$ on $\mathbb{U}$, one seeks to update the prior measure from
observational data. In the following we will assume that $\mu_{0}$ is the
uniform probability measure in the sense of [52, Section 2]. To obtain
observations from solutions in $X$, an _observation operator_
$O:X\rightarrow\mathbb{R}^{K}$ is assumed to be given, allowing to introduce
the _uncertainty–to–observation_ map as
$\mathcal{J}\coloneqq O\circ G:\mathbb{U}\rightarrow\mathbb{R}^{K}.$
Further, the observations $\Upsilon\in\mathbb{R}^{K}$ to update the prior
measure are assumed to be distorted by additive Gaussian noise,
(5.6) $\displaystyle\Upsilon=\mathcal{J}({{\bm{{y}}}^{\star}})+\eta,$
where $\eta\sim\mathcal{N}(0,\Sigma)$, $\Sigma\in\mathbb{R}^{K\times
K}_{\text{sym}}$ is a symmetric, positive definite covariance matrix and
${\bm{{y}}}^{\star}$ is the _ground truth_. The updated _posterior probability
measure_ $\mu^{\Upsilon}$ can then be expressed with respect to the prior
measure as follows.
###### Theorem 5.3 ([52, Theorem 2.1]).
Assume that $\mathcal{J}:\mathbb{U}\rightarrow\mathbb{R}^{K}$ is bounded and
continuous. Then $\mu^{\Upsilon}({{\text{d}}}{\bm{{y}}})$, the distribution of
${\bm{{y}}}\in\mathbb{U}$ given the data $\Upsilon\in\mathbb{R}^{K}$, is
absolutely continuous with respect to $\mu_{0}({{\text{d}}}{\bm{{y}}})$, i.e.
there exists a parametric density $\Theta({\bm{{y}}})$ such that for each
${\bm{{y}}}\in\mathbb{U}$ the _Radon-Nikodym_ derivative is given by
(5.7)
$\displaystyle\frac{{{\text{d}}}\mu^{\Upsilon}}{{{\text{d}}}\mu_{0}}({\bm{{y}}})=\frac{1}{Z}\Theta({\bm{{y}}})$
with the posterior density
(5.8)
$\displaystyle\Theta({\bm{{y}}})\coloneqq\exp\left(-\Phi_{\Sigma}(\Upsilon,\bm{y})\right)\quad\text{and}\quad
Z\coloneqq\int\limits_{\mathbb{U}}\Theta({\bm{{y}}})\mu_{0}({{\text{d}}}\bm{y})>0,$
with
$\displaystyle\Phi_{\Sigma}(\Upsilon,\bm{y})=\frac{1}{2}\left(\Upsilon-\mathcal{J}({\bm{y}})\right)^{\top}\Sigma^{-1}\left(\Upsilon-\mathcal{J}({\bm{y}})\right).$
Given a quantity of interest $\phi\colon\mathbb{U}\to Y$, being $Y$ a
separable Banach space over $\mathbb{C}$, Theorem 5.3 allows us to compute
statistical moments of a quantity of interest with respect to the posterior
measure. For example, one can compute the mean with respect to
$\mu^{\Upsilon}({{\text{d}}}{\bm{{y}}})$ as follows
(5.9)
$\displaystyle\mathbb{E}^{\mu^{\Upsilon}}[\phi]=\int_{\mathbb{U}}\phi({\bm{{y}}})\mu^{\Upsilon}({{\text{d}}}{\bm{{y}}})=\frac{1}{Z}\int\limits_{\mathbb{U}}\phi({\bm{{y}}})\Theta({\bm{{y}}})\mu_{0}({{\text{d}}}\bm{y}),$
with $Z$ as in Eq. 5.8.
#### 5.2.2 Bayesian Shape Inversion in Acoustic Scattering
Equipped with the frequency-robust boundary integral formulations for the
sound-soft acoustic scattering problem introduced in Section 4, the
uncertainty–to–solution and observation operators of the Bayesian shape
inversion problem are
(5.10) $G\colon\mathbb{U}\to
L^{2}(\hat{\Gamma}):{\bm{{y}}}\mapsto\begin{cases}\hat{\phi}_{{\bm{{y}}}}=\left(\hat{\mathsf{A}}^{(\kappa,\eta)^{\prime}}_{{\bm{{y}}}}\right)^{-1}\hat{f}_{\bm{{y}}}&\text{for
\lx@cref{creftype~refnum}{prb:problem_Aprime_k_n}},\\\
\hat{\varphi}_{{\bm{{y}}}}=\left(\hat{\mathsf{A}}^{(\kappa,\eta)}_{{\bm{{y}}}}\right)^{-1}\hat{g}_{\bm{{y}}}&\text{for
\lx@cref{creftype~refnum}{prbm:BIE_sound_soft_A}},\end{cases}$
with $\hat{\mathsf{A}}^{(\kappa,\eta)}_{{\bm{{y}}}}$ and
$\hat{\mathsf{A}}^{(\kappa,\eta)^{\prime}}_{{\bm{{y}}}}$ as in Eq. 4.40 and
$\hat{f}_{\bm{{y}}}$ and $\hat{g}_{\bm{{y}}}$ as in Eq. 4.43 and
(5.11)
$O\colon\mathbb{U}\to\mathbb{R}^{2},\quad{\bm{{y}}}\mapsto\begin{pmatrix}\Re(u_{\bm{{y}}}^{\infty})\\\
\Im(u_{\bm{{y}}}^{\infty})\end{pmatrix}$
with
(5.12)
$u_{\bm{{y}}}^{\infty}=\begin{cases}\hat{\text{FF}}^{(\kappa)}_{\mathcal{S},\Gamma_{\bm{{y}}}}\left(\hat{\phi}_{{\bm{{y}}}}\right)&\text{for
\lx@cref{creftype~refnum}{prb:problem_Aprime_k_n}},\\\
\hat{\text{FF}}^{(\kappa)}_{\mathcal{S},\Gamma_{\bm{{y}}}}\left(\hat{\varphi}_{{\bm{{y}}}}\right)-\iota\eta\hat{\text{FF}}^{(\kappa)}_{\mathcal{D},\Gamma_{\bm{{y}}}}\left(\hat{\varphi}_{{\bm{{y}}}}\right)&\text{for
\lx@cref{creftype~refnum}{prbm:BIE_sound_soft_A}},\end{cases}$
and $\hat{\text{FF}}^{(\kappa)}_{\mathcal{S},\Gamma_{\bm{{y}}}}$ and
$\hat{\text{FF}}^{(\kappa)}_{\mathcal{D},\Gamma_{\bm{{y}}}}$ as in Eq. 4.48.
Multiple observation directions or directions of incident waves could also be
considered with the straightforward adaptions. The quantity of interest for
our Bayesian shape inversion problem is
(5.13)
$\phi\colon\mathbb{U}\to\mathscr{C}^{0,1}(\hat{\Gamma};\mathbb{R}^{3})\colon{\bm{{y}}}\mapsto\bm{r}_{\bm{{y}}},$
with $\bm{r}_{\bm{{y}}}$ is as in Eq. 2.13. Other quantities of interest could
also be considered.
#### 5.2.3 Parametric Regularity of the Posterior Distribution
In this section, we present the parametric regularity results for the
posterior density as defined in Eq. 5.8. Prior research conducted in [23] and
[50] has also investigated this topic and arrived at similar conclusions.
Specifically, in [23, Theorem 4 and Corollary 5], the authors established
parametric regularity estimates for the posterior distribution using a real-
valued differentiation approach.
In contrast, our approach centers on the sparsity of the posterior
distribution. We observe that it can be expressed as the composition of an
entire function and a $(\bm{b},p,\varepsilon)$-holomorphic map, which in turn
implies that the posterior density is also
$(\bm{b},p,\varepsilon)$-holomorphic. This finding provides a useful way to
derive the sparsity of the posterior density.
###### Corollary 5.4.
Let 2.8 and 4.5 be fulfilled with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. In addition, let $G:\mathbb{U}\rightarrow L^{2}(\hat{\Gamma})$,
$O:\mathbb{U}\rightarrow\mathbb{R}^{2}$, and
$\phi:\mathbb{U}\rightarrow\mathscr{C}^{0,1}(\hat{\Gamma},\mathbb{R}^{3})$ be
as in Eq. 5.10, Eq. 5.11, and Eq. 5.13, respectively. Then the maps
(5.14)
$\displaystyle\mathbb{U}\ni{\bm{{y}}}\mapsto\Theta({\bm{{y}}})\in\mathbb{R}\quad\text{and}\quad\mathbb{U}\ni{\bm{{y}}}\mapsto\phi({\bm{{y}}})\Theta({\bm{{y}}})\in\mathscr{C}^{0,1}(\hat{\Gamma},\mathbb{R}^{3})$
with $\Theta:\mathbb{U}\rightarrow\mathbb{C}$ as in Eq. 5.8 are
$(\bm{b},p,\varepsilon)$-holomorphic and continuous.
###### Proof 5.5.
We show that $O$ is indeed $(\bm{b},p,\varepsilon)$-holomorphic, which yields
$(\bm{b},p,\varepsilon)$-holomorphy of $\mathcal{J}=O\circ G$. The remainder
of the proof is in complete analogy to [50, Theorem 4.1]. Using Lemma 4.14,
the $(\bm{b},p,\varepsilon)$-holomorphy of $O$ is immediate, once we show that
taking the real and the imaginary part of a complex function is holomorphic.
Indeed, we provide a proof of this statement in Lemma A.1.
This result justifies the sparse polynomial interpolation results also for the
posterior distribution of the acoustic scattering problem. Moreover, along the
lines of the following section, it also allows to construct surrogates of the
posterior distribution using artificial neural networks.
### 5.3 Approximation of the Far-Field using ANNs
Let $L\in\mathbb{N}$, $\ell_{0},\dots,\ell_{L}\in\mathbb{N}$ and let
$\sigma:\mathbb{R}\rightarrow\mathbb{R}$ be a non-linear function, referred to
in the following as the _activation function_. Given $({\bf W}_{k},{\bf
b}_{k})_{k=1}^{L}$, ${\bf W}_{k}\in\mathbb{R}^{\ell_{k}\times\ell_{k-1}}$,
${\bf b}_{k}\in\mathbb{R}^{\ell_{k}}$, define the affine transformation ${\bf
A}_{k}:\mathbb{R}^{\ell_{k-1}}\rightarrow\mathbb{R}^{\ell_{k}}:{\bf
x}\mapsto{\bf W}_{k}{\bf x}+{\bf b}_{k}$ for $k\in\\{1,\ldots,L\\}$. We define
an _artificial neural network (ANN)_ with activation function $\sigma$ as a
map
$\Psi_{\mathcal{NN}}:\mathbb{R}^{\ell_{0}}\rightarrow\mathbb{R}^{\ell_{L}}$
with
(5.15) $\displaystyle\Psi_{\mathcal{NN}}({\bf x})\coloneqq\begin{cases}{\bf
A}_{1}(x),&L=1,\\\ \left({\bf A}_{L}\circ\sigma\circ{\bf
A}_{L-1}\circ\sigma\cdots\circ\sigma\circ{\bf A}_{1}\right)(x),&L\geq
2,\end{cases}$
where the activation function $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ is
applied componentwise. We define the depth and the width of an ANN as
$\text{width}(\Psi_{\mathcal{NN}})=\max\\{\ell_{1},\ldots,\ell_{L}\\}\quad\text{and}\quad\text{depth}(\Psi_{\mathcal{NN}})=L.$
In the present work, we consider as activation function the hyperbolic tangent
(5.16)
$\displaystyle\sigma(x)=\text{tanh}(x)=\frac{\exp(x)-\exp(-x)}{\exp(x)+\exp(-x)}.$
When this particular function is used, we refer to Eq. 5.15 as a tanh ANN.
However we remark that other choices exist, such as the Rectified Linear Unit
(ReLU), Rectified Power Unit (RePU), and the _swish_ activation function.
We are interested in the tanh ANN approximation of the far-field at a specific
observation direction $\bm{d}_{\text{obs}}\in\mathbb{S}^{2}$ using the
mathematical tools of [1]. To this end, we define
(5.17) $\Phi:\mathbb{U}\rightarrow\mathbb{R}:{\bm{{y}}}\mapsto\left\lvert
u_{\bm{{y}}}^{\infty}\right\rvert^{2},$
where, for each ${\bm{{y}}}\in\mathbb{U}$, $u_{\bm{{y}}}^{\infty}$ is given as
in Eq. 5.12. In the following, we present a streamlined and tailored-to-our-
purposes version of the results presented in [1] to this particular
application, which allows for a straightforward generalization to more
involved examples. To this end, we first recall the approximation of Legendre
polynomials by means of tanh ANNs.
###### Lemma 5.6 ([1, Theorem 7.4]).
Let $\Lambda\subset\mathcal{F}$ with $\mathcal{F}$ as in Eq. 5.3 be a finite
multi-index set, $m(\Lambda)=\max_{\bm{\nu}\in\Lambda}\|\bm{\nu}\|_{1}$, and
$\vartheta\subset\mathbb{N}$, $\left\lvert\vartheta\right\rvert=n$ satisfying
(5.18) $\bigcup_{\bm{\nu}\in\Lambda}\text{supp}(\bm{\nu})\subseteq\vartheta.$
Then, for any $\delta\in(0,1)$ there exists a tanh ANN
$\Psi_{\Lambda,\delta}:\mathbb{R}^{n}\rightarrow\mathbb{R}^{\left\lvert\Lambda\right\rvert}$
with
$\Psi_{\Lambda,\delta}=\left(\Psi_{\bm{\nu},\delta}\right)_{\bm{\nu}\in\Lambda}$
such that
(5.19) $\left\lVert
P_{\bm{\nu}}^{(2)}-\Psi_{\bm{\nu},\delta}\circ\mathcal{T}_{\vartheta}\right\rVert_{L^{\infty}(\mathbb{U})}\leq\delta,\quad\bm{\nu}\in\Lambda,$
with
$\text{width}\left(\Psi_{\bm{\nu},\delta}\right)=\mathcal{O}\left(\left\lvert\Lambda\right\rvert
m(\Lambda)\right)$ and
$\text{depth}\left(\Psi_{\bm{\nu},\delta}\right)=\mathcal{O}\left(\log_{2}(m(\Lambda))\right)$,
where
$\mathcal{T}_{\vartheta}:\mathbb{R}^{\mathbb{N}}\rightarrow\mathbb{R}^{\left\lvert\vartheta\right\rvert}:{\bm{{y}}}=(y_{j})_{j\in\mathbb{N}}\mapsto(y_{j})_{j\in\vartheta}.$
In combination with the sparse polynomial approximation results from Section
5.1 this result allows to derive expression rates of tanh ANNs.
###### Lemma 5.7.
Let 2.8 and 4.5 be fulfilled with $\bm{b}\in\ell^{p}(\mathbb{N})$ and
$p\in(0,1)$. Let $\Phi:\mathbb{U}\rightarrow\mathbb{R}$ be as in Eq. 5.17.
Assume that $\bm{b}$ has a minimal monotone majorante in
$\ell^{p}(\mathbb{N})$. There exists a sequence of tanh neural network
$(\Psi_{\bm{\nu},\delta}^{(n)})_{n\in\mathbb{N}}$ and $C>0$ such that
(5.20)
$\left\lVert\Phi-\Psi_{\bm{\nu},\delta}^{(n)}\circ\mathcal{T}_{\vartheta}\right\rVert_{L^{2}(\mathbb{U})}\leq
Cn^{-\left(\frac{1}{p}-\frac{1}{2}\right)}$
with
$\text{width}\left(\Psi_{\bm{\nu},\delta}^{(n)}\right)=\mathcal{O}(n^{2})$,
$\text{depth}\left(\Psi_{\bm{\nu},\delta}^{(n)}\right)=\mathcal{O}\left(\log_{2}(n)\right)$,
and $\vartheta=\\{1,\dots,n\\}$.
###### Proof 5.8.
In analogy to Corollary 5.4, it follows from 2.8, 4.5, and Lemma A.1 that the
map $\Phi:\mathbb{U}\mapsto\mathbb{R}$ from Eq. 5.17 is
$(\bm{b},p,\varepsilon)$-holomorphic and continuous. Let $\widetilde{\bm{b}}$
be the monotone majorant of $\bm{b}$. Then, as discussed in the proof of [1,
Corollary 8.2], $\Phi$ as in is also
$(\widetilde{\bm{b}},p,\varepsilon)$-holomorphic and continuous. In analogy to
Eq. 5.2, we set
$\Phi=\sum_{\bm{\nu}\in\mathcal{F}}c_{\bm{\nu}}^{(2)}P_{\bm{\nu}}^{(2)}$, with
$\left(c_{\bm{\nu}}^{(2)}\right)_{\bm{\nu}\in\mathcal{F}}$ being the Legendre
coefficients of Eq. 5.17. Set
$\widetilde{\Lambda}_{n}\coloneqq\left\\{\bm{\nu}=(\nu_{k})_{k\in\mathbb{N}}\in\mathcal{F}:\>\prod_{k:\nu_{k}\neq
0}\left(\nu_{k}+1\right)\leq n,\;\nu_{k}=0,\;k>n\right\\}.$
This set contains all _anchored_ sets of a size at most $n$ as pointed out in
[2, Proposition 2.18]. By anchored set we mean a downward closed set
satisfying for each $j\in\mathbb{N}$
(5.21)
$\bm{e}_{j}\in\widetilde{\Lambda}_{n}\Rightarrow\\{\bm{e}_{1},\dots,\bm{e}_{j}\\}\in\widetilde{\Lambda}_{n}.$
Thus, recalling that Theorem 5.1 implies that $\left(\left\lVert
c_{\bm{\nu}}^{(2)}\right\rVert_{L^{2}(\hat{\Gamma})}\right)_{\bm{\nu}\in\mathcal{F}}$
belong to $\ell^{p}_{m}(\mathcal{F})$ and according to [2, Theorem 3.33] it
holds
(5.22) $\sqrt{\sum_{\bm{\nu}\notin\widetilde{\Lambda}_{n}}\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert^{2}}\leq
C(n+1)^{-\left(\frac{1}{p}-\frac{1}{2}\right)}.$
Moreover, as a consequence of Theorem 5.1, for each $n\in\mathbb{N}$ there
exists a downward closed set $\Lambda_{n}$ of cardinality $n$ such that
(5.23)
$\left\lVert\Phi-\sum_{\bm{\nu}\in\Lambda_{n}}c_{\bm{\nu}}^{(2)}P_{\bm{\nu}}^{(2)}\right\rVert_{L^{2}(\mathbb{U})}=\sqrt{\sum_{\bm{\nu}\notin\Lambda_{n}}\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert^{2}}\leq
C(n+1)^{-\left(\frac{1}{p}-\frac{1}{2}\right)},$
and
$\left(c_{\bm{\nu}}^{(2)}\right)_{\bm{\nu}\in\mathcal{F}}\in\ell^{p}_{m}(\mathcal{F})$.
Being $\widetilde{\bm{b}}$ monotonically increasing implies that the set
$\Lambda_{n}$ is anchored as well. Set
$\Lambda=\Lambda_{n}\cap\widetilde{\Lambda}_{n}$. Then,
(5.24)
$\left\lVert\Phi-\sum_{\bm{\nu}\in\Lambda}c_{\bm{\nu}}^{(2)}\Phi_{\bm{\nu},\delta}\circ\mathcal{T}_{\vartheta}\right\rVert_{L^{2}(\mathbb{U})}\leq(\spadesuit)+(\clubsuit),$
where
(5.25)
$\displaystyle(\spadesuit)^{2}=\left\lVert\Phi-\sum_{\bm{\nu}\in\Lambda}c_{\bm{\nu}}^{(2)}P_{\bm{\nu}}^{(2)}\right\rVert^{2}_{L^{2}(\mathbb{U})}$
$\displaystyle=\sum_{\bm{\nu}\notin\Lambda}\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert^{2}$
$\displaystyle<\sum_{\bm{\nu}\notin\Lambda_{n}}\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert^{2}+\sum_{\bm{\nu}\notin\widetilde{\Lambda}_{n}}\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert^{2}\leq
C(n+1)^{-2\left(\frac{1}{p}-\frac{1}{2}\right)}$
and
(5.26)
$\displaystyle(\clubsuit)=\left\lVert\sum_{{\bm{\nu}}\in\Lambda}c_{\bm{\nu}}^{(2)}P_{\bm{\nu}}^{(2)}-\sum_{\bm{\nu}\in\Lambda}c_{\bm{\nu}}^{(2)}\Phi_{\bm{\nu},\delta}\circ\mathcal{T}_{\vartheta}\right\rVert_{L^{2}(\mathbb{U})}\leq\sum_{\bm{\nu}\in\Lambda}\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert\left\lVert
P_{\bm{\nu}}^{(2)}-\Phi_{\bm{\nu},\delta}\circ\mathcal{T}_{\vartheta}\right\rVert_{L^{\infty}(\mathbb{U})}.$
To conclude, we observe that for any
$\bm{\nu}\in\Lambda\subset\widetilde{\Lambda}_{n}$, one has
$\text{supp}(\bm{\nu})=\\{1,\dots,n\\}$. Hence, observing that
$\left(\left\lvert
c_{\bm{\nu}}^{(2)}\right\rvert\right)_{\bm{\nu}\in\mathcal{F}}$ belongs to
$\ell^{p}_{m}(\mathcal{F})\subset\ell^{1}_{m}(\mathcal{F})$, Lemma 5.6 with
$\delta=(n+1)^{-\left(\frac{1}{p}-\frac{1}{2}\right)}$ and
$\vartheta=\\{1,\dots,n\\}$ yields the final result. Also, since any
$\bm{\nu}\in\Lambda$ implies $\bm{\nu}\in\widetilde{\Lambda}_{n}$, we have
(5.27) $\displaystyle m(\Lambda)=\max_{\bm{\nu}\in\Lambda}\|\bm{\nu}\|_{1}\leq
n,$
thus yielding the width and depth estimates stated in this result.
## 6 Conclusion and Outlook
In this work, we consider a class of boundary integral operators set on a
collection of parametrically defined boundaries. In this setting, we establish
the holomorphic dependence of said operators–as an element of a Banach spaces
of bounded linear operators–upon the parametric variables, i.e., we show that
the parameter-to-operator map is holomorphic. This result allows us to
establish parametric holomorphy results for the solution of boundary integral
equations involving the aforementioned operators, provided the invertibility
of the operator for each parametric input, i.e., we show that the parameter-
to-solution map is holomorphic. Though we only consider BIOs in a
$L^{2}$-based setting, this results enables parametric holomorphy analysis in
three dimensional domains defined on affinely parametrized polygonal domains.
To show the applicability of our parametric holomorphy results, we explore the
implications on the boundary integral formulation of the sound-soft acoustic
wave scattering. In particular, herein we discussed the implications in sparse
polynomial approximation of the parameter-to-solution map, reduced order
modelling, in establishing parametric regularity of the posterior distribution
in Bayesian shape inversion, and in the construction of efficient surrogates
with artificial neural networks for the approximation of the far-field
pattern. Even though not thoroughly discussed, using the parametric regularity
and examples established in this article allows to prove dimension-independent
convergence rates for multilevel algorithms in forward and inverse acoustic
domain uncertainty quantification such as in [21], where the domain
deformations were implemented using NURBS.
Further work comprises exploring the implications of the parametric holomorphy
result for BIOs. A canonical example is the analysis and implementation of the
Galerkin POD-ANN as in [32] for the sound-soft acoustic scattering problem.
## Acknowledgement
|
]https://github.com/NThakkar-IDM/covid_and_stat_mech
# A generating function perspective on the transmission forest
Niket Thakkar<EMAIL_ADDRESS>[ Mike Famulare The
Institute for Disease Modeling
Global Health — Bill $\&$ Melinda Gates Foundation
Seattle, Washington 98109
###### Abstract
In a previous paper, we showed that a compartmental stochastic process model
of SARS-CoV-2 transmission could be fit to time series data and then
reinterpreted as a collection of interacting branching processes drawn from a
dynamic degree distribution. We called this reinterpretation a transmission
forest. This paper builds on that idea. Specifically, leveraging generating
function methods from analytic combinatorics, we develop a theory describing
the transmission forest’s properties, allowing us to show for example that
transmission tree interactions fade with increasing disease prevalence. We
then validate the theory by computing forest statistics, like the tree
survival function, which we compare to estimates based on the sampling method
developed previously. The accuracy and flexibility of the analytic approach is
clear, and it allows us to comment on multi-scale features of more general
transmission processes.
disease modeling, probability, generating function, transmission tree,
COVID-19
## I. Epidemiology and graph theory
Transmission trees are the basic graphical unit of an epidemiology. Or, said
differently, if we knew the full transmission tree, it would shed light on a
range of questions – characteristics assigned to nodes would inform our
understanding of risk, features of edges would teach us about transmission
mechanisms, and geometric changes would help us estimate the effects of
interventions.
But of course the basic epidemiological problem is that we can’t measure
transmission trees directly. Even in perfectly observed, closed populations,
assigning edges between nodes is only possible inferentially when transmission
events are sufficiently staggered. In that situation, the gold standard is
outbreak investigations conducted by specialists, where interviews and follow
ups are used to construct a plausible subtree of the full transmission tree
[1]. As a result, even in the best case, this approach cannot scale to a full
epidemiological system.
Another, modern, alternative path forward comes from work on viral
phylogenetics. In that case, genetic sequences of sampled viruses can be
arranged into a phylogenetic tree, and features of that tree can be used to
infer features of the associated transmission trees [2]. This is easier in
theory than in practice. The phylogentic tree is a product of epidemiological
and evolutionary processes, the latter of which can be very complex and system
specific, and separating signals in general is a challenge.
Within this context, in a paper last year [3], we explored a third
perspective. We demonstrated that volatility in Washington’s COVID-19
epidemiological curves contains information on the underlying transmission
trees’ degree distribution. Once specified, that time-varying distribution
could be used to grow a set of interacting trees, which we called a
transmission forest, and while those trees lack the individual-level
resolution of conventionally estimated transmission trees, they were shown to
be predictive of observations from outbreak investigations and phylogenetics.
This step towards harmony between often discordant views of a cryptic epidemic
[4], that is the time series, phylogeny, and outbreak investigations, speaks
to the transmission tree as a fundamental structure for organizing
epidemiological information. Transmission tree inference is an ambitious goal,
but we stand to learn a lot working towards it.
Along those lines, this paper contributes to the broader project of
statistically characterizing transmission trees using more readily available
epidemiological data. Our concrete goal is to more completely and rigorously
define the transmission forest, and in some sense this paper is a mathematical
supplement to the original [3]. That said, more than clarifying the work from
last year, the formalism we develop here offers a tractable connection between
individual-level behavior, the resulting tree structures, and emergent
forests, leading to broadly applicable multi-scale insights into transmission
processes.
## II. The transmission forest model
Figure 1: Sampling the forest. (a) For a given $N_{t}$ trajectory (black
curve) a transmission forest is a random graph (edges in grey) constrained to
sum to $N_{t}$. A given day’s transmission (inset) between infectious parents
(red dots) and their newly infected children (blue circles) demonstrates how
trees compete for nodes. (b) We can sample under the sum constraint by
creating transmission constellations from a fixed number of edges. (c) The
sampling approach (10k samples, mean in purple, 95$\%$ CI shaded) is able to
efficiently explore tail-events in the skewed degree distributions (black
bars) underlying transmission.
We start by defining the transmission forest mathematically and in that way
encapsulating many of the previous paper’s main ideas. Consider a discrete
time, stochastic disease transmission process between susceptible individuals,
$S_{t}$, and infectious individuals, $I_{t}$. In a closed population, we have
$\displaystyle S_{t}$ $\displaystyle=S_{t-1}-N_{t-1}$ (1) $\displaystyle
E_{t}$ $\displaystyle=\left(1-\frac{1}{d_{E}}\right)E_{t-1}+N_{t-1}$ (2)
$\displaystyle I_{t}$
$\displaystyle=\left(1-\frac{1}{d_{I}}\right)I_{t-1}+\frac{1}{d_{E}}E_{t-1},$
(3)
where newly infected individuals, $N_{t}$, become exposed but not infectious,
$E_{t}$, for $d_{E}$ days on average before becoming infectious for $d_{I}$
days on average. On the one hand, following classical models [5], we can write
$N_{t}$ as
$N_{t}=(\beta_{t}\varepsilon_{t})S_{t}I_{t},$ (4)
with transmission rate $\beta_{t}$ and log-normally distributed volatility
$\varepsilon_{t}$, capturing the idea that new infections come from a random
fraction of interacting pairs. Equivalently, inspired by branching processes
[6], we also have
$N_{t}=\sum_{i=1}^{I_{t}}T_{it},$ (5)
capturing the idea that new infections are the sum total of realized daily
transmission events, $T_{it}$, across infectious individuals indexed by $i$.
We model transmission events as independent and identically distributed with
$T_{t}\sim\text{NegBin}\left(\mu_{t},k_{t}\right),$ (6)
which is approximately entropy maximizing [7] when the mean, $\mu_{t}$, and
over-dispersion, $k_{t}$, are set to maintain Eq. 5 up to second order moments
[3].
The transmission forest is a random graph drawn from this model. We visualize
a sample in Fig. 1a. For a given $N_{t}$ trajectory (black curve), nodes are
placed in daily columns and directed edges are drawn between nodes on day $t$
(blue circles) and their infectious parents between days $t-d_{E}$ and
$t-d_{E}-d_{I}$ (red dots) according to Eq. 6. The process is repeated for all
$t$ to fill the graph (grey lines, where we’ve suppressed the nodes for visual
clarity).111This definition makes it clear that we’ve taken $d_{E}$ and
$d_{I}$ to be deterministic across nodes. Strictly speaking, that
approximation is optional. In both the sampling approach and the theory,
latent and infectious durations can be arbitrarily distributed, but this
simplifying assumption is in keeping with the previous paper, and it helps us
focus the work. In the early part of the trajectory, where $t<d_{E}+d_{I}$,
nodes have no parents, making the graph a necessarily disjoint collection of
transmission trees, inspiring the name.
Model fitting to observed time series is discussed in detail in our previous
paper – here we assume the parameters $\beta_{t}$ and $\varepsilon_{t}$ as
well as the process’s initial conditions are known. Throughout this paper, we
use the same model fit to COVID-19 data from Washington from January 2020 to
March 2021 to illustrate the theory’s application.
## III. Sampling a forest
To generate sample forests as in Fig. 1a, we draw daily transmission graphs,
shown in the inset, and stitch them together over time. The sampling approach
gives scaffolding to the theory below, and we use it to validate results
throughout, so it’s worth a brief aside.
$I_{t}$ and $N_{t}$ trajectories can be drawn from Eqs. 1–4 using standard
normal samples. Then, for a given trajectory (rounded to the nearest
integers), the number of nodes in the graph is fixed, and as a result edges
drawn from Eq. 6 are constrained to satisfy Eq. 5 and are no longer
independent. In other words, that trees compete for nodes gives rise to their
interactions in the forest.
We might consider a rejection sampler in this case, essentially drawing
$I_{t}$ negative binomial samples repeatedly until they sum to $N_{t}$ for
every $t$, but that approach is unusably inefficient. With a fixed trajectory,
the right-hand-side of Eq. 5 is asymptotically Gaussian by the central limit
theorem, implying that on average
$\mathcal{O}(I_{t}\sqrt{I_{t}\text{V}[T_{t}]})$ samples have to be drawn per
time-step to create a single forest sample. For the Washington COVID-19 model
we’re working with, with a roughly 400 day time series, that translates to
over 1 billion negative binomial samples to draw one forest.
Fig. 1b illustrates a way forward. The daily transmission graph is a
collection of star graphs, which we call a constellation, and can be
represented as a length $I_{t}$ tuple of infectious node degrees constrained
to sum to $N_{t}$. Initializing a constellation as
$\displaystyle c=(1,1,1,...,1,0,0,0,...,0),$
a tuple of $N_{t}$ ones and $I_{t}-N_{t}$ zeros, we start at $c$’s first entry
($i=0$, representing edges assigned to the first red dot) by defining the
rolling sum $C=\sum_{j=0}^{i-1}c_{j}$ and drawing a sample, $\ell$, from
$\displaystyle
p(T_{t}=\ell|N_{t},C)=\frac{p(T_{t}=\ell)}{1-\sum_{m>N_{t}-C}p(T_{t}=m)},$
the conditionally renormalized degree distribution. We then set $c_{i}=\ell$,
set the next $\ell-1$ entries to 0 (taking edges from next red dots), and then
repeat the process with $i\rightarrow i+\ell$ until $C=N_{t}$. In Fig. 1b,
this is illustrated for $\ell=3$, and in Fig. 1c, we verify that the approach
can maintain the skewed target distributions we have in mind. Finally,
completing the forest requires us to link stars over time. We do this in a
simple way, shuffling $c$ and assigning the resulting events to infectious
nodes in order, without regard to past assignments.
This edge rewiring approach requires $\mathcal{O}(N_{t})$ negative binomial
samples and is, as a result, efficient enough to be performed every day for a
given model trajectory. In the Washington example, it leads to a 3 orders of
magnitude speed up over the rejection method, allowing us to sample large sets
of forests and compute statistics empirically.
## IV. Tree interactions are weak
Setting aside the empirical approach for now, our goal is to calculate
probabilities, like the distribution of tree sizes or their chance of
extinction over time. The next sections are the core of this paper, developing
a generating function [8] approach for modeling forest growth.
The sampling algorithm above motivates an overarching strategy, first
characterizing the possible daily transmission constellations (Fig. 1a inset)
and then considering the process of assigning stars to trees. With that in
mind, the negative binomial distribution in Eq. 6 can be represented as a
probability generating function [9],
$\displaystyle f_{T_{t}}(z)$
$\displaystyle=p(T_{t}=0)+p(T_{t}=1)z+p(T_{t}=2)z^{2}+...$
$\displaystyle=\sum_{\ell\geq
0}\binom{k_{t}+\ell-1}{\ell}(p_{t}z)^{\ell}(1-p_{t})^{k_{t}}$
$\displaystyle=\left(\frac{1-p_{t}}{1-p_{t}z}\right)^{k_{t}}$
where $p_{t}\equiv\mu_{t}/(\mu_{t}+k_{t})$ and $z$ is an arbitrary complex
number. Then, for a given day’s $I_{t}$, constellations drawn from independent
samples can be organized by the total number of edges in a generating function
product,
$\displaystyle C_{t}(z)$
$\displaystyle=f_{T_{t}}(z)^{I_{t}}=\left(\frac{1-p_{t}}{1-p_{t}z}\right)^{I_{t}k_{t}}$
$\displaystyle=\sum_{n\geq 0}p(n\text{ edges})z^{n}.$
As a first illustration of a probability calculation, using $[z^{n}]$ to
signify extracting the $z^{n}$’s coefficient from a formal power series,
Newton’s binomial theorem implies
$[z^{N_{t}}]C_{t}(z)=\binom{I_{t}k_{t}+N_{t}-1}{N_{t}}p_{t}^{N_{t}}(1-p_{t})^{I_{t}k_{t}}$
(7)
is the probability of satisfying Eq. 5 for a fixed $N_{t}$ over all possible
constellations with $I_{t}$ infectious nodes.
Eq. 7 represents the subset of graphs consistent with the trajectory
constraints, that is the $I_{t}$ and $N_{t}$ node populations. To get a more
detailed view of constellation structure, we can calculate the probability of
an infectious individual infecting $m$ new people across constrained graphs.
Introducing arbitrary complex number $u$ to highlight [8] an $m$-event,
$\displaystyle C_{t,m}(z,u)=\left(p(T_{t}=m)uz^{m}+\sum_{\ell\neq
m}p(T_{t}=\ell)z^{\ell}\right)^{I_{t}}$ $\displaystyle=(1-p_{t})^{k_{t}I_{t}}$
$\displaystyle\quad\times\left[\binom{k_{t}+m-1}{m}(u-1)(p_{t}z)^{m}+(1-p_{t}z)^{-k_{t}}\right]^{I_{t}},$
gives a path towards calculating the analog of Eq. 6 with tree interactions.
Note that $C_{t,m}(z,1)=C_{t}(z)$ for any $m$, as required. Then, the expected
number of $m$-events in day $t$’s constellation, $n_{m}$, is
$\text{E}[n_{m}|N_{t},I_{t}]=\frac{[z^{N_{t}}]\frac{\partial}{\partial
u}C_{t,m}(z,u)|_{u=1}}{[z^{N_{t}}]C_{t}(z)},$
and, making use of Eq. 7, we find
$\displaystyle\frac{\text{E}[n_{m}|N_{t},I_{t}]}{I_{t}}=\frac{\binom{k_{t}+m-1}{m}\binom{k_{t}(I_{t}-1)+N_{t}-m-1}{N_{t}-m}}{\binom{k_{t}I_{t}+N_{t}-1}{N_{t}}}$
(8)
$\displaystyle\approx\binom{k_{t}+m-1}{m}p_{t}^{m}(1-p_{t})^{k_{t}}\left[1+\mathcal{O}\left(\frac{m}{I_{t}}\right)\right],$
(9)
where the first line is exact for a given trajectory, and the second line uses
Sterling’s approximation to the Gamma function in the form
$\Gamma(x+\alpha)\sim\Gamma(x)x^{\alpha}[1+(\alpha^{2}+\alpha/2+1/12)/x]$ for
$\alpha<<x$. Eq. 9 shows that the expected degree distribution of infectious
nodes is asymptotically aligned with Eq. 6 but perturbed by tree interactions
that decay with $I_{t}$.
We can prove further that the degree distribution converges to the expected
distribution as graph size grows. A similar calculation, this time leveraging
$C_{t,m}(z,u)$ to compute $\text{E}[n_{m}(n_{m}-1)|N_{t},I_{t}]$, leads to an
intuitive asymptotic variance estimate
$\displaystyle\text{V}[n_{m}|N_{t},I_{t}]\approx
I_{t}p(T_{t}=m)\left[1-p(T_{t}=m)\right],$
implying that deviations around the expected $n_{m}$ are
$\mathcal{O}(1/\sqrt{I_{t}})$, vanishing as $I_{t}$ grows. Thus, the
infectious node degree distribution converges to Eq. 8 at a square-root rate,
and $P(\ell|N_{t},I_{t})\approx\text{E}[n_{\ell}|N_{t},I_{t}]/I_{t}$ is an
approximation that improves with constellation size. In other words, as
disease prevalence increases, we expect tree interactions to fade and
constellation geometry to stabilize.
Figure 2: Interactions and stability. Empirical degree distributions from 1000
sampled constellations (black dots) have interpretable dependence on the
infectious node population. Generating function calculations (blue lines, mean
and 2 standard deviations around it) show that the degree distribution
converges to its expected value with negligible tree interactions (red dashed
line) as disease prevalence increases.
We can validate these results by comparing $P(\ell|N_{t},I_{t})$ to sampled
constellations of various size. This is shown in Fig. 2 for time-averaged
values of $\mu_{t}$ and $k_{t}$ from the Washington COVID-19 model [3]. In the
figure, dots come from the degree distribution across 1000 sampled graphs
while the blue lines show the interacting generating function results (the
mean and 2 standard deviations around it). At low $I_{t}$, tree interactions
are significant and deviations from Eq. 6 (red dashed line) are clear in both
the samples and the generating function estimates. Meanwhile, volatility
around the mean, both in the samples and in the asymptotic distribution decay
with $I_{t}$.
Fig. 2 illustrates the regime where tree interactions warp the epidemiology.
This is a key structural feature of the transmission forest, particularly
relevant in near elimination contexts, and it’s something we intend to study
further. But in the Washington COVID-19 model, from March 2020 to March 2021,
the minimum $I_{t}$ is roughly 1500, implying that for the bulk of the model
time period trees grow in functional isolation.
## V. Trees as recursive functions
Moving from constellations to transmission trees requires us to link
$P(\ell|N_{t},I_{t})$-distributed events over time, being careful to track
when nodes start and stop being infectious. As we’ll see, this process lends
itself to a family of recursively defined generating functions that can be
efficiently evaluated and analysed in reverse time.
The previous section’s results can be written concisely in terms of the
generating function for stars,
$\displaystyle s_{t}(z)$ $\displaystyle=\sum_{\ell\geq
0}P(\ell|N_{t},I_{t})z^{\ell}$ $\displaystyle\approx\sum_{\ell\geq
0}\frac{\text{E}[n_{\ell}|N_{t},I_{t}]}{I_{t}}z^{\ell}$
$\displaystyle\approx\left(\frac{1-p_{t}}{1-p_{t}z}\right)^{k_{t}},$
where the first approximation comes from neglecting volatility around the
average constellation, and the second approximation comes from neglecting tree
interactions – both valid above low prevalence. It’s notable that this likely
would’ve been our naive choice based on Eq. 6 alone, but it’s nice to have
sound theoretical footing and a deeper understanding.
In any case, for a given trajectory, we can formally define the set of all
possible transmission trees rooted at time $r$, $\mathbb{T}_{r}$, by
associating with a tree a monomial $\tau=z_{1}^{n_{1}}z_{2}^{n_{2}}\dotsm
z_{\mathcal{T}}^{n_{\mathcal{T}}}$ where $n_{t}$ is the number of nodes at
time $t$ up to final time $\mathcal{T}$. The set is finite since $0\leq
n_{t}\leq N_{t}$ for all $t$, and we can consider the generating function of
complex vector $\mathbf{z}=(z_{1},...,z_{\mathcal{T}})$
$T_{r}(\mathbf{z})=\sum_{\tau\in\mathbb{T}_{r}}p(\tau|N_{t},I_{t})z_{r}z_{r+1}^{n_{r+1}}\dotsm
z_{\mathcal{T}}^{n_{\mathcal{T}}},$ (10)
which encapsulates tree structures rooted at time $r$. This expression is more
of a formal statement than anything else since it’s not clear how sums over
$\mathbb{T}_{r}$ are executed, and $\mathbf{z}$ is $\mathcal{T}$-dimensional.
We still need a more practical method for function evaluation. Towards that
end, we define $\rho(t)$ as the probability of being infectious $t$ days after
infection, which in the deterministic case is
$\rho(t)=\begin{cases}1&d_{E}<t\leq d_{E}+d_{I}\\\
0&\text{otherwise},\end{cases}$
but clearly could be defined based on a more sophisticated pathogenesis model.
Then, for a binary process, the binomial generating function
$b_{t}(z)=1-\rho(t)+\rho(t)z$ represents infectious status.
With this book-keeping machinery in hand, we can make progress by recognizing
that all trees are star graphs where the internal node is the root and the
external nodes are replaced by trees. This is a well-known recursive idea [8],
and it implies in our case that
$T_{r}(\mathbf{z})=z_{r}\prod_{t\geq{1}}^{\mathcal{T}}\left[1-\rho(t-r)+\rho(t-r)s_{t}(T_{t}(\mathbf{z}))\right],$
(11)
linking root $z_{r}$ to new trees at the appropriate infectious times through
nested compositions of $s_{t}(z)$ and $b_{t}(z)$.222Note that an additional
function composition can be used to incorporate surveillance. If cases are
reported with probability $\pi_{t}$, the binomial generating function
$o_{t}(z)=1-\pi_{t}+\pi_{t}z$ determines if a node is observed, and we can
replace $z_{r}$ with $o_{r}(z_{r})$ directly. The product in Eq. 11 is
tractable given the boundary condition
$T_{\mathcal{T}}(\mathbf{z})=z_{\mathcal{T}}$, which captures the idea that a
tree rooted at $\mathcal{T}$ has no time to grow (that is
$\rho(t-\mathcal{T})=0$ for all $t$ under consideration).
Eq. 11 defines a family of transmission tree generating functions indexed by
their root times, but it doesn’t give us a closed form expression for
$T_{r}(\mathbf{z})$. In practice, Eq. 10 reminds us of $T_{r}(\mathbf{z})$’s
power series structure, and it motivates statistically meaningful function
evaluations. Those evaluations can then be carried out in bespoke recursive
programs based on Eq. 11, starting at time $\mathcal{T}$ and working
backwards.
## VI. Calculating forest statistics
To illustrate the theory’s application and to simultaneously validate the
sampling and recursive approaches, we take up 3 example calculations in this
section: the size distribution of trees in the forest, the tree survival
function, and the relationship between size and lifetime.
Tree size. Evaluating $T_{r}(\mathbf{z})$ at $\mathbf{z}=(z,z,...,z)$, that is
dropping all delineation between $z_{t}$ over time, organizes trees by the
number of nodes. Specifically, Eq. 10 becomes
$\displaystyle T_{r}(z)$
$\displaystyle=\sum_{\tau\in\mathbb{T}_{r}}p(\tau|N_{t},I_{t})z^{n_{r}+n_{r+1}+\dotsm+n_{\mathcal{T}}}$
$\displaystyle=\sum_{n\geq 0}p(n|N_{t},I_{t})z^{n},$
where $n=\sum_{i=1}^{\mathcal{T}}n_{i}$, and in the second line we’ve grouped
terms in the sum. Our goal is to calculate the coefficients in this now one-
dimensional power series.
Complex analysis offers an efficient and well-known path forward [8]. For any
analytic generating function $f(z)$, Cauchy’s integral theorem implies that
$[z^{n}]f(z)=\frac{1}{2\pi i}\oint\frac{f(\xi)}{\xi^{n+1}}d\xi,$
over any closed contour in the complex plane. Choosing $|\xi|=1$ as the
contour leads to
$\displaystyle[z^{n}]f(z)$
$\displaystyle=\frac{1}{2\pi}\int_{0}^{2\pi}f\left(e^{i\theta}\right)e^{in\theta}d\theta$
$\displaystyle\approx\frac{1}{M}\sum_{m=1}^{M}f\left(e^{2\pi
i\frac{m}{M}}\right)e^{2\pi i\frac{nm}{M}},$
for angle $\theta$, which we’ve discretized into $M$ points along the unit
circle in the second line. The sum above is the discrete Fourier transform of
$f(z)$, which implies that a set of coefficients can be extracted by passing
the unit circle through $f(z)$ and transforming. Eq. 11 can be used to
evaluate $T_{r}(z)$ on the unit circle directly, building function evaluations
from $\mathcal{T}$ backwards.
Figure 3: Calculating forest statistics. (a) Complex-analysis based methods
can be used to compute the size distribution of trees (yellow) which compares
favorably to samples (black) (b) Structured evaluations can be used to compute
the tree survival function (orange) which captures key features of a Kaplan-
Meier estimator applied to samples (black) and can be used to compute the
daily hazard function (grey). (c) Expectations (pink curve) and intervals
(pink region) can be computed via Eq. 11 and its derivatives, shown here for
the relationship between tree size and lifetime.
The results of this process are compared to 250k sample trees rooted in early
March 2020 in Fig. 3a, using the parameters from the Washington COVID-19 model
[3]. Both methods yield consistent estimates, but with the generating function
approach (yellow) remaining stable at very low probabilities. As a consequence
of the skewed, individual-level transmission distribution (Eq. 6), tree size
is heavy-tailed. Most trees are simply roots with little growth, but a small
set of trees capture an out-sized fraction of nodes. In other words,
intuitively, super spreaders imply the existence of super trees.
Tree survival. Spontaneous transmission tree extinction implies, in terms of
monomials $\tau=z_{1}^{n_{1}}z_{2}^{n_{2}}\dotsm
z_{\mathcal{T}}^{n_{\mathcal{T}}}$, that $n_{t}=0$ for all $t$ above
extinction time $t^{*}$. This observation motivates a systematic evaluation
method for organizing transmission trees by their lifetimes. Consider
$\mathbf{z}^{*}=(z_{1},z_{2},...,z_{t^{*}},0,0,...)$. Then,
$T_{r}(\mathbf{z}^{*})=\sum_{\tau\in\mathbb{T}_{r}}p(\tau,n_{t>t^{*}}=0|N_{t},I_{t})z_{r}z_{r+1}^{n_{r+1}}\dotsm
z_{t^{*}}^{n_{t^{*}}},$
which, upon setting all remaining $z_{t}=1$, gives
$\displaystyle T_{r}(\Theta(t^{*}))$
$\displaystyle=\sum_{\tau\in\mathbb{T}_{r}}p(\tau,n_{t>t^{*}}=0|N_{t},I_{t})$
$\displaystyle=p(L\leq t^{*}-r)$
where $\Theta(t)$ is the Heaviside step function evaluated along $\mathbf{z}$
and $L$ is the tree lifetime. The survival function,
$S_{r}(t^{*})=P(L>t^{*}-r)$ is then
$\displaystyle S_{r}(t^{*})=1-T_{r}(\Theta(t^{*}))$ (12)
which can be evaluated for all $t^{*}$ again using Eq. 11.
In Fig. 3b, we compare Eq. 12 to a Kaplan-Meier estimator applied to the same
tree samples as above. The theory (orange) is consistent with the empirical
approach (black), even in finer details. Small drops in survival correspond to
times of more concerted transmission suppression in Washington, which are
further highlighted in the hazard function (grey). We refer to past work [10,
11, 12] for epidemiological details, but these survival analysis concepts
applied to transmission trees give perspective on intervention efficacy that
complements more conventional measures like the effective reproductive number
[5].
Size vs. lifetime. Finally, to illustrate how some statistics can be related,
consider $\mathbf{z}\rightarrow u\mathbf{z}$ for complex $u$. Then,
$T_{r}(\mathbf{z})=\sum_{\tau\in\mathbb{T}_{r}}p(\tau|N_{t},I_{t})z_{r}z_{r+1}^{n_{r+1}}\dotsm
z_{\mathcal{T}}^{n_{\mathcal{T}}}u^{n}$
consolidates tree size and structure. Based on the approach above, evaluations
of the form
$\displaystyle T_{r}(u\Theta(t^{*}))-T_{r}(u\Theta(t^{*}-1))$
$\displaystyle=\sum_{\tau\in\mathbb{T}_{r}}p(\tau|L\leq
t^{*}-r)u^{n}-\sum_{\tau\in\mathbb{T}_{r}}p(\tau|L\leq t^{*}-1-r)u^{n}$
$\displaystyle=\sum_{n\geq 0}p(n|L=t^{*}-r)u^{n},$
can be used to relate tree size to the extinction time, and we can compute
statistics. For example,
$\text{E}[n|L=t^{*}-r]=\frac{\frac{\partial}{\partial
u}[T_{r}(u\Theta(t^{*}))-T_{r}(u\Theta(t^{*}-1))]_{u=1}}{T_{r}(\Theta(t^{*}))-T_{r}(\Theta(t^{*}-1))},$
is the expected size at a fixed lifetime. To evaluate the numerator, we can
differentiate Eq. 11 to derive a recursion relation for $T_{r}(\mathbf{z})$’s
partial derivatives. We find
$\frac{\partial}{\partial
u}T_{r}(u\mathbf{z})=T_{r}(u\mathbf{z})\left(\frac{1}{u}+\sum_{t=1}^{\mathcal{T}}\frac{\rho(t-r)s_{t}^{\prime}(T_{t}(u\mathbf{z}))}{1-\rho(t-r)+\rho(t-r)s_{t}(T_{t}(u\mathbf{z}))}\frac{\partial}{\partial
u}T_{t}(u\mathbf{z})\right)$ (13)
where primes denote total derivatives and the evaluation depends on the full
family of rooted generating functions $T_{r}(\mathbf{z})$. Practically, for
some specific $\mathbf{z}$, we can evaluate all $T_{r}(\mathbf{z})$ first,
which then specifies the details in Eq. 13. A similar calculation can be used
to compute higher order derivatives and corresponding higher order statistics.
We compare the generating function approach above to samples with non-zero
lifetime in Fig. 3c. Statistics based on Eq. 11 (pink) gracefully capture
discretization effects, like the necessary spacing between trees surviving one
and two generations, as well as the transition to continuous lifetimes.
Features of the relationship, like local slope changes in keeping with the
mitigation environment in Washington at that time, are clearly reflected in
both the model and the samples. That this distribution can be calculated
efficiently based on time series data alone is striking, and it illustrates
how accessible transmission tree statistics might be with the right approach.
## VII. Conclusion
The key idea in this paper is that conventional stochastic process models of
disease transmission, based on Eq. 4, can be used to estimate transmission
tree properties without additional data. The bridge between the population-
scale of trajectories and the individual-scale of trees comes from Eq. 5,
which led us to a generating function family that can be used to efficiently
compute a variety of epidemiologically relevant statistics.
The theory developed here offers new perspective on individual-level data,
like distributions across outbreak investigations or associations among
collections of genetic sequences. Those types of comparisons were made
empirically in our previous paper [3], but the analytic approach gives
additional paths towards more quantitative comparisons and potentially joint
inferences. Speaking broadly, tree structures have a long mathematical
history. Eq. 11 represents a promising and intuitive connection between that
body of work and the classical compartmental models used to describe disease
transmission.
## Acknowledgements
This work was done with many people’s support. In particular, we want to thank
Kevin McCarthy for his attention and detailed comments. His input made the
writing much more clear and it brought to light implications we hadn’t
considered.
## References
* Arunkumar _et al._ [2019] G. Arunkumar, R. Chandni, D. T. Mourya, S. K. Singh, R. Sadanandan, P. Sudan, and B. Bhargava, Outbreak investigation of nipah virus disease in Kerala, India, 2018, The Journal of Infectious Diseases 219, 1867 (2019).
* Grenfell _et al._ [2004] B. T. Grenfell, O. G. Pybus, J. R. Gog, J. L. Wood, J. M. Daly, J. A. Mumford, and E. C. Holmes, Unifying the epidemiological and evolutionary dynamics of pathogens, Science 303, 327 (2004).
* Thakkar and Famulare [2022] N. Thakkar and M. Famulare, Covid-19 epidemiology as emergent behavior on a dynamic transmission forest, arXiv preprint arXiv:2205.02150 (2022).
* Bedford _et al._ [2020] T. Bedford, A. L. Greninger, P. Roychoudhury, L. M. Starita, M. Famulare, M.-L. Huang, A. Nalla, G. Pepper, A. Reinhardt, H. Xie, _et al._ , Cryptic transmission of SARS-CoV-2 in Washington state, Science 370, 571 (2020).
* Anderson and May [1992] R. M. Anderson and R. M. May, _Infectious diseases of humans: dynamics and control_ (Oxford University Press, 1992).
* Fewster [2014] R. Fewster, Stats 325, Chapter 6: Branching Processes: The Theory of Reproduction (Lecture notes, 2014).
* Jaynes [1957] E. T. Jaynes, Information theory and statistical mechanics, Physical Review 106, 620 (1957).
* Flajolet and Sedgewick [2009] P. Flajolet and R. Sedgewick, _Analytic combinatorics_ (Cambridge University Press, 2009).
* Miller [2018] J. C. Miller, A primer on the use of probability generating functions in infectious disease modeling, Infectious Disease Modelling 3, 192 (2018).
* Thakkar _et al._ [2020a] N. Thakkar, R. Burstein, H. Hu, P. Selvaraj, and D. Klein, Social distancing and mobility reductions have reduced COVID-19 transmission in King County, WA, Institute for Disease Modeling (March 29, 2020a).
* Thakkar and Famulare [2020] N. Thakkar and M. Famulare, One state, many outbreaks: a transmission modeling perspective on current COVID-19 trends in King, Pierce, and Yakima counties, Institute for Disease Modeling (August 17, 2020).
* Thakkar _et al._ [2020b] N. Thakkar, R. Burstein, and M. Famulare, Towards robust, real-time, high-resolution COVID-19 prevalence and incidence estimation, Institute for Disease Modeling (December 13, 2020b).
|
# Multi-core parallel tempering Bayeslands for basin and landscape evolution
Rohitash Chandra<EMAIL_ADDRESS>R. Dietmar Müller Ratneel
Deo Nathaniel Butterworth Tristan Salles Sally Cripps EarthByte Group,
School of Geosciences, University of Sydney, NSW 2006, Sydney, Australia
Centre for Translational Data Science, University of Sydney, NSW 2006, Sydney,
Australia Sydney Informatics Hub, University of Sydney, NSW 2006, Sydney,
Australia School of Mathematics and Statistics, University of Sydney, NSW
2006
Sydney, Australia
###### Abstract
In recent years, Bayesian inference has become a popular methodology for the
estimation and uncertainty quantification of parameters in geological and
geophysical forward models. Badlands is a basin and landscape evolution
forward model for simulating topography evolution at a large range of spatial
and time scales. Previously, Bayesian inference has been used for parameter
estimation and uncertainty quantification in Badlands, an extension known as
Bayeslands. It was demonstrated that the posterior surface of these parameters
could exhibit highly irregular features such as multi-modality and
discontinuities making standard Markov Chain Monte Carlo (MCMC) sampling
difficult. Parallel tempering (PT) is an advanced MCMC method suited for
irregular and multi-modal distributions. Moreover, PT is more suitable for
multi-core implementations that can speed up computationally expensive
geophysical models. In this paper, we present a multi-core PT algorithm
implemented in a high performance computing architecture for enhancing
Bayeslands. The results show that PT in Bayeslands not only reduces the
computation time over a multi-core architecture, but also provides a means to
improve the sampling process in a multi-modal landscape. This motivates its
usage in larger-scale problems in basin and landscape evolution models.
###### keywords:
Landscape evolution , Bayesian inference, Parallel tempering , Badlands ,
Bayeslands
††journal: IEEE Tran. Geo. Remote Sens.
## 1 Introduction
Recent developments in Earth evolution models have the capability to link
models for dynamic and isostatic topography through time [1] with landscape
evolution models [2]. This has given the opportunity to model landscape
evolution in response to surface uplift and subsidence over a large range of
spatial scales, and track sediments from source to sink [3, 4]. Geophysical
forward models depend on uncertain initial and boundary conditions [5], but a
large variety of observations are available to optimise and ground-truth these
models. Uncertain input parameters for these models include time-dependent
topography, relative sea level change, time-varying rainfall and rock
lithology and erodibility [6, 7] . Observations useful for model ground-
truthing include digital elevation models and river geometries, sediment
thickness and stratigraphy in sedimentary basins, and other geological proxies
for landscape evolution.
Bayesian inference provides a rigorous approach for uncertainty quantification
[8, 9, 10] and emerged as a popular tool for fitting complex parameter models
to data [11]. This provides advantages over optimization methods that provide
single point solutions and need multiple experimental runs for uncertainty
quantification [12, 13]. Estimation of parameters that fit observational data
in geophysical models have shifted from optimization [14] to inference that
addresses uncertainty quantification given that the models provide an
approximation of geophysical processes [15]. Markov Chain Monte Carlo (MCMC)
sampling methods implement Bayesian inference where a sample of the desired
distribution is obtained by observing the chain after a number of steps[16,
17, 18, 19]. Bayesian methods have become popular in geophysics in the past
few decades [20, 21, 22, 23]. The convergence in canonical MCMC sampling
methods face further challenges when the size of the problem increases in
terms of the number of parameters and computational time of running the model.
This is a challenge in geophysics that use forward models such as Badlands [3,
4]. Although advanced MCMC methods such as Hamiltonian based samplers have
been proposed [24, 25, 26], they require gradient information from the models
to form proposals. Acquiring gradient information from complex geo-physical
models such as Badlands is difficult, therefore, Bayesian inference via MCMC
random-walk proposals was proposed in a framework called Bayeslands [27].
Bayeslands demonstrated that the posterior surface of selected parameters
could exhibit highly irregular features, such as multi-modality and
discontinuities which made it difficult for sampling. Hence, there is a need
to explore efficient sampling methods that feature gradient free proposals and
addresses multi-modality.
Parallel tempering [28, 29] is a MCMC sampling scheme designed for sampling
multi-modal distributions. Another feature is their feasibility of
implementation in multi-core or parallel computing architectures. Due to
massive computational requirements of geo-physical models, parallel computing
architectures have been widely utilized in geosciences [30, 31, 32]. This can
therefore help us to extend and improve our previous approach called
Bayeslands that featured Bayesian inferences via MCMC random-walk sampler for
Badlands [27]. Hence, the major limitations such as computational time and the
need for efficient multi-modal distribution given by the MCMC random-walk
sampler in Bayeslands can be overcome by multi-core parallel tempering.
In this paper, we present a multi-core parallel tempering approach for
Bayeslands that features uncertainty quantification and estimation of selected
parameters for basin and landscape evolution. We use selected problems that
feature synthetic and real-world applications for selected parameters such as
rainfall and erodibility that affect the topography development over thousands
of years. We evaluate the performance of multi-core parallel tempering
Bayeslands (PT-Bayeslands) for selected number of multiprocessing cores and
report the overall computation time. The results report the successive
predicted topographies, successive sediments given by erosion-deposition,
their prediction accuracy, posterior distribution of selected parameters for
selected problem, and also the overall computation time.
The rest of the paper is organised as follows. Section 2 provides a background
on Badlands and Bayeslands. Section 3 presents the methodology and Section 4
presents the experiments and results. Section 5 provides a discussion and
Section 6 concludes the paper with directions for future research.
## 2 Background and Related Work
### 2.1 Badlands
Over the last decades, many numerical models have been proposed to simulate
how the Earth surface has evolved over geological time scales in response to
different driving forces such as tectonics or climatic variability [33, 34,
35, 36, 37]. These models combine empirical data and conceptual methods into a
set of mathematical equations that can be used to reconstruct landscape
evolution and associated sediment fluxes [38, 39]. They are currently used in
many research fields such as hydrology, soil erosion, hillslope stability and
general landscape studies.
In this paper, we use Badlands [3, 40, 41]. This framework is intended to
simulate regional to continental sediment deposition and associated
sedimentary basin architecture [4, 42]. In its most simple formulation, the
Earth surface elevation change is related to the interaction of three types of
processes, one driven by tectonic, another describing the smoothing effects
associated to diffusive processes, and a last one representing the erosive
power of water flow:
$\frac{\partial z}{\partial t}=-\nabla\cdot\mathbf{q_{s}}+u$
where $u$ in $m\cdot yr^{-1}$ is a source term that represents tectonic
uplift. The total downhill sediment flux $\mathbf{q_{s}}$ is defined by:
$\mathbf{q_{s}}=\mathbf{q_{r}}+\mathbf{q_{d}}$
$\mathbf{q_{s}}$ is the depth-integrated, bulk volumetric soil flux per width
($m^{2}\cdot yr^{-1}$). $\mathbf{q_{r}}$ represents transport by fluvial
system and $\mathbf{q_{d}}$ hillslope processes.
Here the fluvial incision is based on the classical detachment-limited stream
power law (SPL), in which erosion rate $\dot{\epsilon}$ depends on drainage
area $A$, net precipitation $P$ and local slope $S$ and takes the form:
$\dot{\epsilon}=\kappa(PA)^{m}S^{n}$
$\kappa$ is an erodibility coefficient while $m$ and $n$ are positive
exponents characterising the contribution of rain and slope respectively on
erosion processes. Despite its simplicity, the SPL reproduces many of the
characteristic features of natural systems where detachment-limited erosion
regime dominates [33]. With this formulation, sediment deposition occurs
solely in topographically closed depression and marine offshore regions.
Constraining the exact values of $\kappa$, $m$ and $n$ from natural landscapes
is relatively complex and not easily measured from direct field observations.
Their values depend on the geomorphological, climatic and tectonic context and
their ranges are commonly admitted to be:
* 1.
$0\,<\,m\,<\,2$
* 2.
$0\,<\,n\,<\,4$
* 3.
$\kappa$ varies by several orders of magnitude not only based on lithology,
climate, sedimentary flux or river channel width but also with the chosen
values of $m$ and $n$.
In addition to overland flow, semi-continuous processes of soil displacement
are accounted for using a linear diffusion law commonly referred to as soil
creep [34]:
$\frac{\partial z}{\partial t}=\kappa_{d}\nabla^{2}z$
in which $z$ is the elevation and $\kappa_{d}$ is the diffusion coefficient.
This transport rate depends linearly on topographic gradient and encapsulates
in a simple formulation the processes operating on superficial sedimentary
layers.
### 2.2 Bayeslands
Bayelands [27] illustrates how Bayesian inference, via the posterior
distribution, can be used to estimate and quantify the uncertainty surrounding
initial conditions of factors such as rainfall and erodibility. Formally,
given some data, $\mathbf{D}$, we make inference regarding unknown parameters,
denoted by $\theta$, via the posterior distribution $p(\theta|\mathbf{D})$,
given by Bayes rule,
$p(\theta|\mathbf{D})=\frac{p(\mathbf{D}|\theta)p(\theta)}{P(\mathbf{D})}$
where $p(\mathbf{D}|\theta)$ is the likelihood of the data given the
parameters, $p(\theta)$ is the prior, and $p(\mathbf{D})$ is a normalizing
constant and equal to $\int p(\mathbf{D}|\theta)p(\theta)d\theta$. Frequently
the posterior distribution does not have a closed form solution and it is
approximated by a sampling based estimate.
Bayeslands used MCMC to obtain samples of $\theta$ from the posterior
distribution by proposing values from some known distribution $q(\theta)$. The
value of $\theta$ is then set equal to these proposed values with a
probability which ensures the detailed balance condition is met, otherwise the
chain remains in its current position, see [43]. The transition kernel which
moves the Markov chain from one point to another in the parameter space is the
product of the proposal distribution $q$ and the acceptance probability. Under
certain conditions, the draws from this transition kernel converge to draws
from the stationary distribution, $p(\theta|\mathbf{D})$. This algorithm is
known as the Metropolis-Hastings (MH) algorithm [16] which is given in
Algorithm 1
Algorithm 1 Metropolis-Hastings (MH)
Result: Drawing from $p(\theta|\mathbf{D})$
Set maximum number of iterations, $K$, and initialize $\theta=\theta^{[0]}$
Set the current value of $\theta$, $\theta^{c}=\theta^{[0]}$
for _$k=1,\ldots,K$_ do
1. 1.
Propose a new value of $\theta$, $\theta^{p}\sim q(\theta|\theta^{c})$
2. 2.
Compute acceptance probability
$\alpha=\min\left(1,\frac{p(\theta^{p}|\mathbf{D})q(\theta^{c}|\theta^{p})}{p(\theta^{c}|\mathbf{D})q(\theta^{p}|\theta^{c})}\right)$
3. 3.
Draw $u\sim U[0,1]$
if _$u <\alpha$_ then
$\theta^{[k]}=\theta^{p}$
else
$\theta^{[k]}=\theta^{c}$
end
end
The development of transitions kernels which efficiently explore posterior
distributions has been the subject of much research in sampling methods [24].
Bayeslands has major limitations since MCMC random-walk requires thousands of
samples to accurately explore the posterior distribution. Executing it in a
sequential manner that employs only one processing unit becomes a
computational challenge taking into account the time needed to run Badlands
when the application problem covers a larger region featuring millions of
years. We will address these computational challenges using multi-core
parallel tempering in this paper.
### 2.3 Parallel tempering
Posterior distributions of parameters in geophysical inversions problems are
notoriously difficult to explore. They are high dimensional, multi-modal and
very irregular, by which we mean the derivative of posterior is often
difficult to compute and/or it has many discontinuities. An example is shown
in Figure 1 that features the likelihood surface of the Continental Margin
problem used in Bayeslands [27]. While proposal distributions based on
canonical random-walk MCMC methods are gradient free, in sampling multi-modal
posterior distributions, they become very inefficient [44]. As evident, the
likelihood surface can be highly multi-modal and irregular which makes it
natural for these algorithms to be trapped in a local minimum. In such
situations, proposal distributions that feature gradients such as Hamiltonian
methods will also face difficulties since there are local convergence issues
related to gradient based methods [45]. In complex geophysical inversion
problems, proposal distributions usually need to gradient free and explore
most, if not all the posterior modes. Methods such as gradient-free
Hamiltonian Monte-Carlo [46] do not address highly irregular and complex
multi-modal distributions.
Figure 1: Likelihood surfaces of the Continental Margin topography for the
rainfall and erodibility parameters only[27].
Parallel tempering has been motivated by thermodynamics of physical systems
where the temperature of a system determines the ability to change [47].
Parallel tempering is also known as replica exchange or the Multi-Markov Chain
method [48, 49] suitable for multi-modal distributions by running multiple
MCMC chains at different temperatures in parallel. Typically, gradient free
proposals within chains are used for proposals for exploring multi-modal and
discontinuous posteriors [50, 51] .
The Markov chains in the parallel replicas have stationary distributions which
are equal to (up to a proportionality constant)
$p(\theta|\mathbf{D})^{\beta}$, where $\beta\in[0,1]$, with $\beta=0$
corresponds to a stationary distribution which is uniform and $\beta=1$
corresponds to a stationary distribution which is the posterior. The replica
with smaller values of $\beta$ are able to explore a larger regions of
distribution, while those with higher values of $\beta$ typically explore
local regions.
Communication between the parallel replicas is essential for the efficient
exploration of the posterior distribution. This is done by considering the
chain and the parameters as part of the space to be explored as highlighted in
Algorithm 2. Suppose there are $M$ chains, indexed by $m$, with corresponding
stationary distributions,
$p_{m}(\theta|\mathbf{D})=p(\theta|\mathbf{D})^{\beta_{m}}$, for,
$m=1,\ldots,M$, with $\beta_{1}=1$ and
$\beta_{M}<\beta_{M-1}<\ldots,\beta_{1}$, then the pair $(m,\theta)$ are
jointly proposed and accepted/rejected according to the MH algorithm. The
stationary distribution of this sampler is proportional to
$p(\theta|\mathbf{D})^{\beta_{m}}p(m)$. The quantity $p(m)$ must be chosen by
the user and is referred to as a pseudo-prior. For efficient mixing of the
chain we use $q(m^{p}=m^{c}\pm 1|m^{c})=1/2$, for $1<m^{c}<M$, and
$q(m^{p}=m^{c}+1|m^{c}=1)=q(m^{p}=m^{c}-1|m^{c}=M)=1$.
Algorithm 2 Metropolis-Hastings Parallel Tempering (MHPT)
Result: Drawing from $p(\theta|\mathbf{D})$
Set maximum number of iterations, $K$, and initialize $\theta=\theta^{[0]}$,
and $m=m^{[0]}$
Set the current value of $\theta$, to $\theta^{c}=\theta^{[0]}$ and $m$ to
$m^{c}=m^{[0]}$
for _$k=1,\ldots,M$_ do
1. 1.
Update $\theta^{[k]}$, from the chain with
$p(\theta|\mathbf{D})^{\beta_{m^{c}}}$ as its invariant distribution according
to Alg 1.
2. 2.
Propose a new value of $m^{p}$, from $q(m^{p}|m^{c})$.
3. 3.
Compute acceptance probability
$\alpha=\min\left(1,\frac{p(\theta^{[k]}|\mathbf{D})^{\beta_{m^{p}}}p(m^{p})q(m^{c}|m^{p})}{p(\theta^{[k]}|\mathbf{D})^{\beta_{m^{c}}}p(m^{c})q(m^{p}|m^{c})}\right)$
4. 4.
Draw $u\sim U[0,1]$
if _$u <\alpha$_ then
$m^{[k]}=m^{p}$
else
$m^{[k]}=m^{c}$
end
end
A number of attempts have been made to improve the parallel tempering
algorithm, which includes the efficient way of finding $\beta_{m}$ for
$m=1,\ldots,M$, [52, 47], exchange of solutions among the replicas [53], and
joint adaption $\beta_{m}$ and the proposal distributions for
$\mathbf{\theta}$ in the corresponding chain, [54]. The potential for parallel
tempering in geoscience problems has been demonstrated [55] with examples that
better convergence than conventional approaches for complex multi-modal
optimization problems [50, 51, 14]. A number of factors need to be considered
in multi-core implementation which specifically takes into account operating
system concepts such as interprocess communications when considering exchange
of solutions between the chains [56]. In order to address this, a
decentralized implementation of parallel tempering was presented that
eliminates global synchronization and reduces the overhead caused by
interprocess communication in exchange of solutions between the chains that
run in parallel cores [57]. Parallel tempering has also been implemented in a
distributed volunteer computing network where computers belonging to the
general public are used with help of multi-threading and graphic processing
units (GPUs) [58]. Furthermore, parallel tempering has been implemented with
Field Programmable Gate Arrays (FPGAs) that has massive parallelizing
capabilities which showed much better performance than multi-core and GPU
implementations [59]. Furthermore, in terms of applications, other studies
have efficiently implemented parallel tempering via multi-core architectures
for exploration of Earth’s resources [60].
## 3 Methodology: Multi-Core Parallel Tempering Bayeslands
The task of sampling or running geophysical models is computationally
intensive, it is worthwhile to employ parallel computing approach to parallel
tempering since thousands of samples are required for efficiently sampling a
posterior distribution. Massive parallelizing in multi-core architectures that
implements the various chains of the parallel tempering sampling method can
help in speeding up the approach while exploring multi-modal posterior
distributions. Furthermore, Parallel tempering being a method for implementing
Bayesian inference naturally accounts for uncertainty quantification.
Therefore, we present a multi-core parallel tempering approach for addressing
the limitations of Bayeslands. The overall problem remains the same as
Bayeslands, where the task is in sampling from the posterior distribution of
selected parameters such as rainfall and erodibility that contribute to the
topography development over time with Badlands. In doing so, it is essential
to define the likelihood function that will be used to determine where the
proposals are accepted.
The remainder of the section defines the model that features the likelihood
and priors for parallel tempering Bayeslands. Furthermore, the generation of
systematic topographies for the selected problems and the details of
implementation of multi-core parallel tempering for Bayeslands is also given.
### 3.1 Model and Priors
The likelihood function essentially captures the quality to the proposals by
considering the difference between the simulated final topography and the
actual one. In the problems where sediment deposition occurs, the likelihood
function also considers the difference between the simulated and real
successive topographies at selected time intervals.
Let the initial topography be denoted by $\boldsymbol{D}_{0}$, with
$\boldsymbol{D}_{0}=(D_{0,s_{1}}\ldots,D_{0,s_{n}})$, where $s_{i}$
corresponds to site $s_{i}$, with co-ordinates latitude, $u_{i}$, and
longitude, $v_{i}$. Suppose that we are interested in the topography $t$ years
into the future, we will denote this by $\boldsymbol{D}_{t}$, with
$\boldsymbol{D}_{T}$ defined to be the current topography. Our model for the
process which generates the topography is
$D_{t,s_{i}}=f_{t,s_{i}}(\boldsymbol{\theta})+\epsilon_{t,s_{i}}\;\mbox{with}\;\epsilon_{t,s_{i}}\sim(0,\tau^{2})$
(1)
for $t=0,1,\ldots,T$ and $i=1,\ldots,n$, where $\boldsymbol{\theta}$ are the
parameters of the Badlands model and $f_{t,s_{i}}(\boldsymbol{\theta})$ is the
output of the Badlands forward model. This model states that the topography is
function of the Badlands forward model given parameters $\boldsymbol{\theta}$,
plus some Gaussian noise with zero mean and variance $\tau^{2}$. The
likelihood function $L_{l}(\boldsymbol{\theta})$,is given by
$L_{l}(\boldsymbol{\theta})=\frac{1}{(2\pi\tau^{2})^{n/2}}\exp\left\\{-\frac{1}{2}\frac{\sum_{t=1}^{T}\sum_{i=1}^{n}(D_{t,s_{i}}-f_{t,s_{i}}(\boldsymbol{\theta}))^{2}}{\tau^{2}}\right\\}$
where the subscript $l$, in $L_{l}(\theta)$, denotes that it is the landscape
likelihood.
Although it may be possible to observe previous topologies, it is not likely
and therefore we assume that only the current topology $\mathbf{D}_{T}$ is
available. However it may be possible to observe other landscape features of
the past, such as sediment deposits, and we wish to incorporate the
information contained in this data to constrain landscape evolution. We allow
for this possibility by defining another random variable
$\mathbf{z}_{t}=(z_{t,s_{1}}\ldots,z_{t,s_{m}})$ which are the sediment
deposits at sites $s_{1},\ldots,s_{m}$. We assume that observed values of
$\mathbf{z}_{t}$ are a function of the Badlands forward model, with parameter
$\boldsymbol{\theta}$ and some Gaussian noise
$z_{t,s_{i}}=g_{t,s_{i}}(\boldsymbol{\theta})+\eta_{t,s_{i}}\;\mbox{with}\;\eta_{t,s_{i}}\sim(0,\sigma^{2})$
(2)
then the sediment likelihood, $L_{s}(\theta)$ is
$L_{s}(\boldsymbol{\theta})=\frac{1}{(2\pi\sigma^{2})^{mT/2}}\exp\left\\{-\frac{1}{2}\sum_{t=1}^{T}\sum_{j=1}^{m}\frac{(Z_{t,s_{j}}-g_{t,s_{j}}(\boldsymbol{\theta}))^{2}}{\sigma^{2}}\right\\}$
giving a total likelihood as a function of theta $L(\boldsymbol{\theta})$ to
be
$L(\boldsymbol{\theta})=L_{s}(\boldsymbol{\theta})\times
L_{l}(\boldsymbol{\theta}).$
The prior distribution, also known as prior, refers to one’s belief in the
distribution of the parameter without taking into account the evidence or data
[61]. The prior distribution is adjusted by sampling from the posterior with
given likelihood function that takes into account the data and the model. In
the case of the priors for Bayeslands, we focus on a subset of the parameters
of the Badlands model depending upon the problem, see Table tab:truevalues.
The complete set of unknowns in our model is $\boldsymbol{\theta}$, $\tau^{2}$
and $\sigma^{2}$. We follow the setup in [27] and place uniform priors on the
unknown parameters with limits given in Table 3.
### 3.2 Synthetic topography data
Topography | Evo.(years) | Length [km, pts] | Width [km, pts] | Res. factor | Run-time (s)
---|---|---|---|---|---
Crater [27] | 15 000 | [0.24, 123] | [0.24, 123] | 0.002 | 0.5
Crater-extended | 50 000 | [0.24, 123] | [0.24, 123] | 0.002 | 2
CM [27] | 500 000 | [136.5, 91] | [123.0, 82] | 1.5 | 0.5
CM-extended | 1 000 000 | [136.0, 136] | [123.0, 123] | 1 | 2.5
Mountain | 1 000 000 | [80,202] | [40,102] | 1 | 10
Table 1: Landscape evolution problems presented in this paper. The run-time represents the approximate length of time for one model to run. Topography | Rainfall (m/a) | Erod. | n-value | m-value | c-marine | c-surface | Uplift (mm/a)
---|---|---|---|---|---|---|---
Crater [27] | 1.5 | 5.0-e05 | 1.0 | 0.5 | - | - | -
Crater-extended | 1.5 | 5.0-e05 | 1.0 | 0.5 | - | - | -
CM [27] | 1.5 | 5.0-e06 | 1.0 | 0.5 | 0.5 | 0.8 | -
CM-extended | 1.5 | 5.0-e06 | 1.0 | 0.5 | 0.5 | 0.8 | -
Mountain | 1.5 | 5.0-e06 | 1.0 | 0.5 | - | - | 1.0
Table 2: True values of parameters Topography | Rainfall (m/a) | Erod. | n-value | m-value | c-marine | c-surface | uplift
---|---|---|---|---|---|---|---
Crater [27] | [0,3.0 ] | [3.0-e05, 7.0-e05] | - | - | - | - |
Crater-ext | [0,3.0 ] | [3.0-e05, 7.0-e05] | [0, 2.0] | [0, 2.0] | - | - |
CM [27] | [0,3.0 ] | [3.0-e06, 7.0-e06] | - | - | - | - | -
CM-ext. | [0,3.0 ] | [3.0-e06, 7.0-e06] | [0, 2.0] | [0, 2.0] | [0.3, 0.7] | [0.6, 1.0] | -
Mountain | [0,3.0 ] | [3.0-e06, 7.0-e06] | [0, 2.0] | [0, 2.0] | - | - | [0.1, 1.7]
Table 3: Prior distribution range of model parameters
Badlands is a forward stratigraphic model that requires an initial topography
and values for $\boldsymbol{\theta}$ to simulate an expected final (ground-
truth) topography at time $T$, given by $\mathbf{f}_{T}(\boldsymbol{\theta})$
in equation 1, and expected sediment deposits at several points in time, given
by $g_{t,s_{i}}(\boldsymbol{\theta})$ using Equation 2.
We consider five landscape problems from which two are taken from Bayeslands
which consist of a crater and continental margin (CM) example, where only two
parameters (rainfall and erodibility) are sampled. In this work, we present
two variants of these problems, such as crater-extended and CM-extended. The
extended problems consider additional parameters and simulation time for
topography development. Furthermore, we also use a mountain building case that
features uplift rates as a parameter which is not present in the other
problems. The initial topographies for the crater-extended and CM-extended
cases appear in Figure 2 and Figure 3, respectively. The final topography
after 50,000 years and 1,000,000 years along with erosion-deposition is also
given. The Crater and CM cases have an initial and ground-truth topography
identical to [27]. Figure 4 shows the initial and ground-truth final
topography of the Mountain landscape problem after 1,000,000 years. Note that
there is no erosion-deposition catered for this case in the likelihood
function to demonstrate that this data is unavailable. Figure 9 shows the
evolution of the PT-Bayeslands model. Tables2 and 1 shows values of parameters
used to created the synthetic ground-truth topography. The parameters for the
two crater problems are, rainfall (m/a), erodibility, m-value, and n-value.
The CM-extended problem features six free parameters, which include marine
diffusion coefficient (c-marine given by $m^{2}$/a) and surface diffusion
coefficient (c-surface given by $m^{2}$/a) as shown in Table 2. The mountain
problem features 5 free parameters, rainfall (m/a), erodibility, m-value, and
n-value, and uplift (mm/a). The priors were drawn from a uniform distribution
with lower and upper limit given in Table 3.
---
(a) Crater-extended initial topography
(b) Crater-extended synthetic ground-truth topography
Figure 2: Crater-extended: Initial and eroded ground-truth topography and
sediment deposition after 50 000 years.
---
(a) CM-extended initial topography
(b) CM-extended synthetic ground-truth topography
Figure 3: Continental Margin (CM)-extended: Initial and eroded ground-truth
topography and sediment after 1 000 000 years.
---
(a) Mountain initial topography
(b) Mountain synthetic ground truth topography
Figure 4: Mountain: Initial and eroded ground-truth topography after 1 000 000
years evolution.
### 3.3 Multi-core parallel tempering Bayeslands
Parallel tempering can be regarded as an ensemble of replicas of a MCMC
sampler. Given $N$ replicas in a ensemble defined by multiple temperature
levels, the state of the ensemble is specified by
$X={\theta_{1},\theta_{2},...,\theta_{N}}$, where $\theta_{i}$ is the replica
at temperature level $T_{i}$. A Markov chain is constructed to sample
$\theta_{i}$ at each temperature level $T_{i}$ which is used to rescale the
likelihood value for the replica in order to give variates in exploration and
exploitations during sampling. In other words, . In principle, at every
iteration, the Markov chains can feature two types of transitions that include
1) the Metropolis transition and 2) a replica transition. In the Metropolis
transition, each replica is sampled independently to perform local Monte Carlo
moves defined by the temperature which is implemented by a change in the
likelihood function for each temperature level $T_{i}$.
The replica transition considers the exchange of two neighboring replicas that
are defined by two neighboring temperature levels, $i$ and $i+1$,
$\theta_{i}\leftrightarrow\theta_{i+1}$. The exchange is accepted by the
Metropolis-Hastings criterion with probability given in Algorithm 2. The
exchange enables a replica with a low temperature level to have a chance to
reach a high temperature level which gives enhanced exploration features that
can enable to escape a local minimum. In this way, the replica exchange can
shorten the sampling time required for convergence. The temperature level is
user-defined, which needs to be determined from trial experiments, and is
highly dependent on the nature of the problem in terms of likelihood surface
and multi-modality.
Figure 5: An overview of the different replicas that are executed on a multi-
processing architecture. Note that the main process controls the given
replicas and enables them to exchange the neighboring replicass given the swap
time and probability of exchange is satisfied.
Figure 5 gives an overview of the different replicas that are executed on a
multi-processing architecture where each replica runs on a separate core with
inter-process communication for exchanging neighboring replica. The Badlands
model is executed in the same processing core where the replicas local
sampling is executed. The main process is run on a separate core which
controls the replica and enables them to exchange the neighboring replicas
given the swap time and probability of exchange is satisfied. In order to
minimize interprocess communication between processes for reducing
computational costs, we only consider replica transition at fixed intervals
that are determined by user defined number of samples used for Metropolis
samples. Each replica is allocated a fixed sampling time. The main process
waits for all samplers to complete their required sampling after which the
samplers attempt configuration exchange. The main process waits for each
replica until the swap-interval sampling limit is fulfilled. Once the replica
reaches this junction, the main process proposes configuration swaps between
adjacent replications depending on the replica exchange probability. The main
process notifies the replicas post swapping to resume sampling with latest
configurations in the chain for each replica. The process continues with the
sampling and proposing swaps until the the maximum sampling time is reached.
In the multi-core implementation, we need to consider multi-processing
software development packages so that the resources have efficient inter-
process communication [56] considering the exchange of the solutions between
neighboring replica’s.
## 4 Experiments and Results
In this section, we present the experimental design and results for five
landscape evolution problems for evaluating the performance of PT-Bayeslands.
We use the landscape problems from Table 1, that involve synthetic and real-
world topographies. We implemented multi-core parallel tempering using the
Python multiprocessing library [62] and open-source software package
111Parallel tempering Bayeslands: https://github.com/badlands-
model/paralleltemp_Bayeslands .
### 4.1 Design of Experiments
PT-Bayeslands employs a random-walk sampler for each replica chain which is
able to run on separate cores (central processing units). The random-walk was
implemented by perturbing the chain in the respective replica with a small
amount of Gaussian noise with a parameter specific step-size or standard
deviation. The step-size $\beta_{i}$ for parameter $i$ is chosen to be a
combination of a fixed step size $\phi=0.02$, common to all parameters,
multiplied by the range of possible values for parameter $i$ so that
$\beta_{i}=(a_{i}-b_{i})*\phi$ (3)
where, $a_{i}$ and $b_{i}$ represent the maximum and minimum limits of the
priors for parameter and are given in Table 2. Note that PT-Bayeslands samples
the parameters with prior distribution highlighted for the respective problems
in Table 3. The step-size ratio $\phi$ was determined by trial experimental
runs that was used by all the problems. Similarly, the temperature change for
each of the replica’s $r$ was $\delta_{r}=2.5$. In trial experiments, the
selection of these parameters depended on the prediction accuracy. We used a
fix swap ratio, $\eta=0.1$, that determines when to swap with the neighboring
replica which is incorporated into the algorithm 2 and further illustrated in
Figure 5. Therefore, the number of samples $s$ required to check when to swap
for a given replica $r$ is given as follows.
$s_{r}=\eta*(\Phi/\Omega)$ (4)
where $\Phi$ is the total number of samples and $\Omega$ is the number of
replica’s for the problem. The experiments are evaluated in terms of total
simulation time, acceptance rate of proposals, and root-mean-squared-error
(RMSE) of elevation and sediment (Pred. RMSE), which have been defined in the
methodology section.
In parallel tempering, the number of replicas needs to be tailored for the
problem. They also determine the total simulation time for the problem. It is
generally expected that increasing the number of replica’s in a multi-core
architecture will shorten the computation time; however, additional time is
taken with interprocess communication through swapping of neighboring replicas
which is also dependent on the swap interval and number of replica’s that can
contribute to the wait-time by the main process. As as shown in Figure 5, the
main process runs on a separate core. Therefore given $N$ number of replicas,
the total number of processes running will be $N+1$. The task of the main
process is to manage the ensemble of replicas. It will hold and resume the
replica sampling process during the swap period. Moreover, we need to ensure
that the quality of the topography and sediment predictions by Badlands is
retained when number of cores are increased. Hence, we need to investigate the
effect on performance accuracy and time of increasing number of replicas for
PT-Bayeslands. The experiments are designed as follows.
* 1.
Step 1: Investigate the effects on computational time and accuracy when
increasing the number of replicas;
* 2.
Step 2: Evaluate the number of samples required for convergence defined in
terms of prediction accuracy;
* 3.
Step 3: Using knowledge from above experiments, apply PT-Bayeslands to all the
given problems and report the posterior distributions, computational time,
uncertainty and accuracy in topography and sediment predictions.
We present the crater problem for Step 1 and the CM problem for Step 2. Since
the CM problem has demonstrated an irregular multi-modal surface, as shown in
Figure 1 [27], it is important to evaluate the effect of sampling time on the
posterior distribution and the accuracy of prediction.
### 4.2 Results
We evaluate the effect of the number of replicas, executed in different cores,
has on the total time for the CM-extended topography. Table 4 presents a
summary of the results that employs 100,000 samples in PT-Bayeslands for the
CM-extended experiment. We observe that increasing the number of replicas
reduces the overall computation time. However the number of replicas does not
appear to effect the RMSE. Figure 6 shows the effect of the number of cores on
the time, panel (a) and on the RMSE, panel (b).
Figure 6: The effect of the number of cores on time, panel (a) and on the
RMSE, panel (b). Note that a 95% confidence interval for the slope of the
regression line in panel (b), is [-14.2,2.6], showing that the slope is not
significantly different from zero.
Next, we evaluate the effect of the prediction accuracy given an increase in
the number of samples. Table 5 presents summary of the results on the effect
of number of samples for the CM-extended topography with 24 replicas. We
observe that increasing the number of samples increases the overall time
taken, and reduces the acceptance rate of proposals. Moreover, the interval
between (10,000 - 100,000) samples does not have a major difference in the
prediction accuracy. Hence, even with 10,000 samples, PT-Bayeslands can
provide estimation and uncertainty quantification. Due to the stochastic
nature of the method and the difficulty of the likelihood surface shown in
Figure 1, we don’t see a clear trend in convergence when the number of cores
(replicas) or samples are increased. The performance depends on initial
conditions that is given by the initial points of the chains; depending on how
far they are from one of the modes in the likelihood surface, and the tendency
to be trapped in a local minimum.
In the next set of experiments, we used 24 cores with 100,000 samples for all
the problems. Table 6 shows a summary of results for the given problems. Note
that the total prediction accuracy is given by (Pred. RMSE) which consists of
sediment prediction accuracy (Sed. RMSE) and elevation prediction accuracy
(Elev. RMSE). Pred. RMSE considers the difference between the ground-truth and
final topography given by Badlands (after maximum number of evolution time in
years). However, the Sed. RMSE considers the successive predicted sediments
taken equally over 4 different timescales given in years. A major observation
here is the overall computation time for the five problems, which is due to
the time taken by Badlands to execute each proposal. The crater and crater-
extended cases are simpler problems (with shorter evolution time and fewer
parameters), hence more proposals are accepted when compared to the CM cases.
Moreover, the overall prediction performance (Pred. RMSE) has been
significantly improved by PT-Bayeslands. A cross-section of the crater with
the uncertainty of prediction is shown in Figure 7.
We select the results for the CM-extended problem from Table 6 and present
further details which include; 1.) successive predicted topographies and
erosion-deposition at four stages of evolution (Figure 8), 2.) cross-section
that gives uncertainty of topography prediction (Figure 10), and 3.) selected
posterior distributions, and trace-plots during sampling (Figure 11). Note
that erosion is represented as positive while deposition is represented as
negative (height in meters) (Figure 8). Posterior distributions, trace plots,
and other results for all parameters and all models are available in the on-
line supplementary material. In comparison with the true values given in Table
2 that was used to generate the synthetic ground-truth topography, we notice
that the true values were not recovered in some of the posterior distributions
(e. g. Figure 11). However, as shown in the likelihood surface (Figure 1), a
number of sub-optimal values exist. Due to multi-modality, although the true
values are not recovered for certain parameters, the sub-optimal modes give
accurate predictions. Note that the likelihood surface shown in Figure 1) is
for two parameters only (rainfall and erodibility).
Replica (cores) | Time (minutes) | Pred. RMSE | Accepted %
---|---|---|---
2 | 3540 | 78.8 | 0.1
4 | 1384 | 62.7 | 0.1
8 | 614 | 71.5 | 0.4
12 | 455 | 87.0 | 0.5
16 | 375 | 72.0 | 0.6
20 | 362 | 110.6 | 0.3
24 | 270 | 65.9 | 0.4
28 | 215 | 89.7 | 0.5
32 | 310 | 205.1 | 0.4
36 | 213 | 128.0 | 0.6
40 | 187 | 107.0 | 0.6
44 | 162 | 70.0 | 0.6
48 | 142 | 80.5 | 0.8
Table 4: Effect of number replicas/cores for the CM-extended topography with 100 000 samples. Num. Samples | Time (mins.) | Pred. RMSE | Accepted %
---|---|---|---
1000 | 3 | 120.6 | 14.0
2000 | 6 | 222.5 | 15.3
4000 | 10 | 200.8 | 9.4
6000 | 18 | 98.3 | 3.7
8000 | 18 | 213.1 | 8.3
10000 | 27 | 77.3 | 2.8
50000 | 131 | 77.7 | 1.3
100000 | 258 | 95.1 | 1.0
150000 | 436 | 69.3 | 0.2
Table 5: Effect of the number of samples for the CM-extended topography running with 24 replicas on 24 cpus. Topography | Time (minutes) | Sed. RMSE | Elev. RMSE | Pred. RMSE | Accepted %
---|---|---|---|---|---
Crater | 80 | 4.2 | 1.1 | 5.3 | 12.2
Crater-extended | 229 | 0.2 | 1.0 | 1.2 | 1.9
CM | 50 | 2.6 | 19.9 | 22.5 | 1.0
CM-extended | 257 | 50.2 | 49.0 | 99.2 | 0.6
Mountain | 375 | - | 617.0 | 617.0 | 0.63
Table 6: Typical results for respective problems (24 replicas/cpus and 100 000
samples). Note that ’-’ in case of Mountain Sed. RMSE indicates that it was
not part of the likelihood function Figure 7: Crater-extended: Cross-section
of the elevation along the x-axis comparing the ground-truth of the Badlands-
model with the PT-Bayelands predictions.
---
(a) CM-extended predicted topography after 50 % evolution time
(b) CM-extended erosion-deposition after 50 % evolution time
(c) CM-extended predicted topography after 100 % evolution time
(d) CM-extended erosion-deposition after 100 % evolution time
Figure 8: CM-extended: Topography and erosion-deposition development for
selected time frames. Note that erosion (positive) and deposition (negative)
values given by the height in meters
---
(a) Mountain predicted topography after 25 % evolution time
(b) Mountain predicted topography after 50 % evolution time
(c) Mountain predicted topography after 75 % evolution time
(d) Mountain predicted topography after 100 % evolution time
Figure 9: Mountain evolution over 1 000 000 years.
---
Figure 10: CM-extended: Cross-section of the elevation along the x-axis
comparing the ground-truth of the Badlands-model with the PT-Bayelands
predictions.
---
Figure 11: Typical posterior distribution (top panel) and replica traces
(bottom panel) exploring the marine diffusion parameter-space for the CM-
extended topography. In this model run 24 replicas are represented by the
coloured lines. Parameter, c-marine, units are $m^{2}a^{-1}$.
Comparison with literature (MCMC Bayeslands [27]) is given in Figure 12. The
trace-plot in these figures show that the PT-Bayeslands have advantages in
exploration with multiple replicas. The comparison of the computational time
is given in Table 7. These results show that PT-Bayeslands has a significant
improvement in computational time and accuracy in prediction performance.
---
(a) Rainfall distribution (MCMC- Bayeslands)
(b) Rainfall distribution (PT- Bayeslands)
Figure 12: Comparison of PT-Bayeslands with MCMC-Bayeslands for the CM problem. Note that the vertical dashed lines in different colours shows sub-optimal true values or modes that correspond to the likelihood surface in Figure 1. Topography | Samples | Time(mins) | Sed RMSE | Elev RMSE | Pred RMSE | Accepted%
---|---|---|---|---|---|---
MCMC Bayeslands results from [27]
Crater | 10,000 | 136 | 8.24 | 1.06 | 9.30 | 2.35
Crater | 100,000 | 1023 | 8.24 | 1.06 | 9.30 | 2.57
CM | 10,000 | 101 | 459.85 | 67.37 | 527.22 | 0.47
CM | 100,000 | 729 | 387.92 | 10.60 | 398.53 | 0.02
PT-Bayeslands using 24 replicas (this study)
Crater | 10,000 | 8 | 4.2 | 1.0 | 5.2 | 15.0
Crater | 100,000 | 78 | 4.2 | 1.1 | 5.3 | 12.2
CM | 10,000 | 7 | 7.4 | 25.3 | 32.7 | 4.9
CM | 100,000 | 50 | 2.6 | 19.9 | 22.5 | 1.0
Table 7: Comparison of results of PT-Bayeslands with single-threaded MCMC
Bayeslands [27] shows significant difference in performance given
computational time and accuracy in prediction.
---
Figure 13: Posterior distribution (top panel) and replica traces (bottom
panel) exploring the uplift parameter-space for the Mountain topography. The
24 replicas trace distribution are represented by the coloured lines.
Parameter, uplift, units are $mm\,a^{-1}$.
---
(a) Cross-section along the x-axis.
(b) Cross-section along the y-axis
Figure 14: Typical topography for the evolved Mountain problem, comparing PT-
Bayeslands predictions and the Badlands ground-truth. Result is for 24
replicas, 100,000 samples. Figure 15: Likelihood surface of the Mountain
topography for the rainfall and uplift parameters only.
In this paper, we only show details of the posterior distribution, successive
topography and successive erosion-deposition predictions for a sample of model
runs. Detailed results for all problems and model runs are provided in
supplementary material online 222Supplementary results:
https://github.com/badlands-
model/paralleltemp_Bayeslands/tree/master/supplementary_results.
## 5 Discussion
The results have shown that computation time for a given number of iterations
does not scale linearly with the number of replicas. There is a trade-off
between proposing to swap chains and the rate of convergence of the PT-
Bayeslands. Swapping proposals between neighboring replicas, in principle,
gives better mixing but introduces computational cost because each chain needs
to wait for all the replicas to complete in order to compute the neighbor swap
probability. This computational cost can be high if the number of replicas is
large. We note that the synthetic problems considered in this paper, only
required a few seconds of simulation time for Badlands, whereas, in real-world
problems, each simulation of Badlands could take several minutes to hours. In
such cases, the trade-off between the frequency of swapping and prediction
accuracy would need to be evaluated. In this study, we observed that
increasing the number of cores does not necessarily mean that the prediction
accuracy will get better. In particular it is advantageous to use a larger
number of cores for large scale problems where the Badlands model takes hours
to evaluate a single proposal. Moreover, the replica temperature is also an
important feature that can be evaluated in future works. Since evaluation of
such parameters is computationally expensive for large scale problems,
heuristics such as geometric temperature intervals can be used [52, 47].
Figures 12 show the improvement of parallel tempering over single chain MCMC.
Panel (a) in both figures represents output from the single MCMC-Bayeslands
while panel (b), represents the output from the PT-Bayeslands. These figures
show that PT-Bayelands mixes more and discovers more of the multiple nodes
than the single chain. The degree of improvement of using PT-Bayeslands over
single core based MCMC-Bayeslands depends upon the type of topography and the
parameter. For example, the estimated posteriors distributions for all
parameters in the Crater topography contained the true value of that
parameter, while in the CM-extended topography, only some of the posterior
distributions contained the true parameter. This is not surprising when one
considers the highly irregular shape of the likelihood surface, as given in
Figure 1[27]. Despite this variability in results the predictions were
accurate which could be attributable to a multi-modal posterior. Multi-modal
distributions are well known in Earth science problems in the literature (e.g.
[63] and [64]). Multi-modality implies that there are several optimal values
of the parameters where multiple combinations of the given parameters can
plausibly predict or simulate the topography that closely resembles the
synthetic or ground-truth topography [65].
Figure 13 shows the posterior distribution of the uplift parameter for the
Mountain topography. In this case the the chains found the true value early
and did not stray far in their exploration, likely because the uplift
parameter is not sensitive to the solution at the scales explored here (0.1 –
1.7 mm/a). The likelihood surface shown in Figure 15 suggests this, as
rainfall increases and uplift decreases (with other parameters set to their
true values) the likelihood remains high. Adjusting the temperature of the
chains and swap ratios can help influence the exploration of the chains to fit
the problem more appropriately. Yet here Figure 14 shows that the ground-truth
and the PT-Bayeslands predictions fit well, indicating that the explored space
collectively for all parameters taken together was sufficient to match the
data.
Furthermore, the experiments considered a fixed value for some parameters,
such as rainfall, that can have varying effects given different regions and
geological timescales. It is well known that climate change may affect
rainfall with varying implications to the environment, and in this case the
topography in terms of elevation via affecting erosion and sedimentation.
Therefore, it would be reasonable to implement region- and time-dependent
constraints for some parameters to fully take into account the effects of
climate changes.
## 6 Conclusions and Future Work
We presented a computationally efficient implementation of multi-core parallel
tempering for improving the computational time require to estimate parameters
of the Bayeslands model. The proposed methodology provides a general
systematic approach for uncertainty quantification of free parameters in basin
and landscape dynamics models (Badlands). The results showed that the method
not only reduces the computation time, but also provides a means to explore
the parameter space in highly irregular multi-modal landscapes. This has been
demonstrated by the results that show better posterior distributions of the
parameters along with improvement in prediction accuracy of topography and
sediment deposition.
Future work will extend the method for a larger number of parameters that
includes spatio-temporal variations in rainfall and uplift for different
timescales and regions. Moreover, for large-scale or continental-scale
problems, it would be reasonable to implement further enhancements to the
method for lowering the overall computational time. This could be done through
surrogate-assisted models where at times the surrogate of Badlands implemented
via machine learning evaluates the proposals. There is scope for real-world
applications, that feature landscape evolution for geological timescales.
Further, efficient gradient free proposals need to be constructed as the
number of parameters and the complexity of the model increases.
## Acknowledgements
We would like to thank Danial Azam from School of Geoscience, the University
of Sydney for providing technical assistance. We would also like to thank the
Strategic Research Excellence Initiative (SREI) Grant from the University of
Sydney. Furthermore, we would like to acknowledge the Artemis high performance
computing infrastructure provided by Sydney Informatics Hub of the University
of Sydney.
## References
## References
* [1] N. Flament, M. Gurnis, and R. D. Müller, “A review of observations and models of dynamic topography,” _Lithosphere_ , vol. 5, no. 2, pp. 189–210, 2013.
* [2] T. Coulthard, “Landscape evolution models: a software review,” _Hydrological processes_ , vol. 15, no. 1, pp. 165–173, 2001.
* [3] T. Salles and L. Hardiman, “Badlands: An open-source, flexible and parallel framework to study landscape dynamics,” _Computers & Geosciences_, vol. 91, pp. 77–89, 2016.
* [4] T. Salles, N. Flament, and D. Müller, “Influence of mantle flow on the drainage of eastern australia since the jurassic period,” _Geochemistry, Geophysics, Geosystems_ , vol. 18, no. 1, pp. 280–305, 2017\.
* [5] L. R. Scott and S. Zhang, “Finite element interpolation of nonsmooth functions satisfying boundary conditions,” _Mathematics of Computation_ , vol. 54, no. 190, pp. 483–493, 1990.
* [6] V. Godard, J. Lavé, and R. Cattin, “Numerical modelling of erosion processes in the himalayas of nepal: Effects of spatial variations of rock strength and precipitation,” _Geological Society, London, Special Publications_ , vol. 253, no. 1, pp. 341–358, 2006.
* [7] J. Rejman, R. Turski, and J. Paluszek, “Spatial and temporal variations in erodibility of loess soil,” _Soil and Tillage Research_ , vol. 46, no. 1-2, pp. 61–68, 1998.
* [8] G. Shafer, “Belief functions and parametric models,” _Journal of the Royal Statistical Society. Series B (Methodological)_ , pp. 322–352, 1982.
* [9] S. E. Fienberg _et al._ , “When did bayesian inference become” bayesian”?” _Bayesian analysis_ , vol. 1, no. 1, pp. 1–40, 2006.
* [10] N. L. Hjort, C. Holmes, P. Müller, and S. G. Walker, _Bayesian nonparametrics_. Cambridge University Press, 2010, vol. 28.
* [11] C. Robert and G. Casella, “A short history of markov chain monte carlo: Subjective recollections from incomplete data,” _Statistical Science_ , pp. 102–115, 2011.
* [12] K. Mosegaard and P. D. Vestergaard, “A simulated annealing approach to seismic model optimization with sparse prior information,” _Geophysical Prospecting_ , vol. 39, no. 5, pp. 599–611, 1991.
* [13] P. Rocca, M. Benedetti, M. Donelli, D. Franceschini, and A. Massa, “Evolutionary optimization as applied to inverse scattering problems,” _Inverse Problems_ , vol. 25, no. 12, p. 123003, 2009.
* [14] M. K. Sen and P. L. Stoffa, _Global optimization methods in geophysical inversion_. Cambridge University Press, 2013.
* [15] K. Gallagher, K. Charvin, S. Nielsen, M. Sambridge, and J. Stephenson, “Markov chain monte carlo (mcmc) sampling methods to determine optimal models, model resolution and model choice for earth science problems,” _Marine and Petroleum Geology_ , vol. 26, no. 4, pp. 525–535, 2009.
* [16] W. K. Hastings, “Monte carlo sampling methods using markov chains and their applications,” _Biometrika_ , vol. 57, no. 1, pp. 97–109, 1970.
* [17] N. Metropolis, A. W. Rosenbluth, M. N. Rosenbluth, A. H. Teller, and E. Teller, “Equation of state calculations by fast computing machines,” _The journal of chemical physics_ , vol. 21, no. 6, pp. 1087–1092, 1953.
* [18] A. E. Raftery and S. M. Lewis, “Implementing mcmc,” _Markov chain Monte Carlo in practice_ , pp. 115–130, 1996.
* [19] D. van Ravenzwaaij, P. Cassey, and S. D. Brown, “A simple introduction to markov chain monte–carlo sampling,” _Psychonomic bulletin & review_, pp. 1–12, 2016.
* [20] A. Malinverno, “Parsimonious bayesian markov chain monte carlo inversion in a nonlinear geophysical problem,” _Geophysical Journal International_ , vol. 151, no. 3, pp. 675–688, 2002.
* [21] K. Mosegaard and A. Tarantola, “Monte carlo sampling of solutions to inverse problems,” _Journal of Geophysical Research: Solid Earth_ , vol. 100, no. B7, pp. 12 431–12 447, 1995.
* [22] M. Sambridge and K. Mosegaard, “Monte carlo methods in geophysical inverse problems,” _Reviews of Geophysics_ , vol. 40, no. 3, 2002.
* [23] M. Sambridge, “Geophysical inversion with a neighbourhood algorithm—ii. appraising the ensemble,” _Geophysical Journal International_ , vol. 138, no. 3, pp. 727–746, 1999.
* [24] R. M. Neal _et al._ , “Mcmc using hamiltonian dynamics,” _Handbook of Markov Chain Monte Carlo_ , vol. 2, no. 11, 2011.
* [25] M. D. Hoffman and A. Gelman, “The no-u-turn sampler: adaptively setting path lengths in hamiltonian monte carlo.” _Journal of Machine Learning Research_ , vol. 15, no. 1, pp. 1593–1623, 2014.
* [26] M. Girolami and B. Calderhead, “Riemann manifold langevin and hamiltonian monte carlo methods,” _Journal of the Royal Statistical Society: Series B (Statistical Methodology)_ , vol. 73, no. 2, pp. 123–214, 2011.
* [27] R. Chandra, D. Azam, D. Muller, T. Salles, and S. Cripps, “Bayeslands: A bayesian inference approach for parameter tuning for modelling basin and landscape dynamics via badlands,” _Computers and Geoscience)_ , pp. In–Review, 2018. [Online]. Available: https://github.com/rohitash-chandra/research/blob/master/2018/BayesLands.pdf
* [28] E. Marinari and G. Parisi, “Simulated tempering: a new monte carlo scheme,” _EPL (Europhysics Letters)_ , vol. 19, no. 6, p. 451, 1992.
* [29] C. J. Geyer and E. A. Thompson, “Annealing markov chain monte carlo with applications to ancestral inference,” _Journal of the American Statistical Association_ , vol. 90, no. 431, pp. 909–920, 1995.
* [30] H. Zhang, M. Liu, Y. Shi, D. A. Yuen, Z. Yan, and G. Liang, “Toward an automated parallel computing environment for geosciences,” _Physics of the Earth and Planetary Interiors_ , vol. 163, no. 1-4, pp. 2–22, 2007.
* [31] J. A. Vrugt, B. O. Nualláin, B. A. Robinson, W. Bouten, S. C. Dekker, and P. M. Sloot, “Application of parallel computing to stochastic parameter estimation in environmental models,” _Computers & Geosciences_, vol. 32, no. 8, pp. 1139–1155, 2006.
* [32] K. Mills, G. Fox, and R. Heimbach, “Implementing an intervisibility analysis model on a parallel computing system,” _Computers & Geosciences_, vol. 18, no. 8, pp. 1047–1054, 1992.
* [33] K. X. Whipple and G. E. Tucker, “Implications of sediment-flux-dependent river incision models for landscape evolution,” _Journal of Geophysical Research: Solid Earth_ , vol. 107, no. B2, pp. 1–20, 2002.
* [34] G. E. Tucker and G. R. Hancock, “Modelling landscape evolution,” _Earth Surface Processes and Landforms_ , vol. 35, no. 1, pp. 28–50, 2010.
* [35] T. Salles and G. Duclaux, “Combined hillslope diffusion and sediment transport simulation applied to landscape dynamics modelling.” _Earth Surf. Process Landf._ , vol. 40, no. 6, p. 823–39, 2015.
* [36] B. Campforts, W. Schwanghart, and G. Govers, “Accurate simulation of transient landscape evolution by eliminating numerical diffusion: the ttlem 1.0 model,” _Earth Surface Dynamics_ , vol. 5, no. 1, pp. 47–66, 2017.
* [37] J. M. Adams, N. M. Gasparini, D. E. J. Hobley, G. E. Tucker, E. W. H. Hutton, S. S. Nudurupati, and E. Istanbulluoglu, “The landlab v1.0 overlandflow component: a python tool for computing shallow-water flow across watersheds,” _Geoscientific Model Development_ , vol. 10, no. 4, pp. 1645–1663, 2017.
* [38] A. D. Howard, W. E. Dietrich, and M. A. Seidl, “Modeling fluvial erosion on regional to continental scales,” _Journal of Geophysical Research: Solid Earth_ , vol. 99, no. B7, pp. 13 971–13 986, 1994.
* [39] D. E. J. Hobley, H. D. Sinclair, S. M. Mudd, and P. A. Cowie, “Field calibration of sediment flux dependent river incision,” _Journal of Geophysical Research: Earth Surface_ , vol. 116, no. F4, 2011.
* [40] T. Salles, “Badlands: A parallel basin and landscape dynamics model,” _SoftwareX_ , vol. 5, pp. 195–202, 2016.
* [41] T. Salles, X. Ding, and G. Brocard, “pybadlands: A framework to simulate sediment transport, landscape dynamics and basin stratigraphic evolution through space and time,” _PLOS ONE_ , vol. 13, pp. 1–24, 04 2018.
* [42] T. Salles, X. Ding, J. M. Webster, A. Vila-Concejo, G. Brocard, and J. Pall, “A unified framework for modelling sediment fate from source to sink and its interactions with reef systems over geological times,” _Scientific Reports_ , vol. 8, no. 1, p. 5252, 2018.
* [43] S. Chib and E. Greenberg, “Understanding the metropolis-hastings algorithm,” _The American Statistician_ , vol. 49, no. 4, pp. 327–335, 1995.
* [44] R. M. Neal, “Sampling from multimodal distributions using tempered transitions,” _Statistics and computing_ , vol. 6, no. 4, pp. 353–366, 1996\.
* [45] P. Baldi, “Gradient descent learning algorithm overview: A general dynamical systems perspective,” _IEEE Transactions on neural networks_ , vol. 6, no. 1, pp. 182–195, 1995.
* [46] H. Strathmann, D. Sejdinovic, S. Livingstone, Z. Szabo, and A. Gretton, “Gradient-free hamiltonian monte carlo with efficient kernel exponential families,” in _Advances in Neural Information Processing Systems_ , 2015, pp. 955–963.
* [47] A. Patriksson and D. van der Spoel, “A temperature predictor for parallel tempering simulations,” _Physical Chemistry Chemical Physics_ , vol. 10, no. 15, pp. 2073–2077, 2008.
* [48] K. Hukushima and K. Nemoto, “Exchange monte carlo method and application to spin glass simulations,” _Journal of the Physical Society of Japan_ , vol. 65, no. 6, pp. 1604–1608, 1996.
* [49] U. H. Hansmann, “Parallel tempering algorithm for conformational studies of biological molecules,” _Chemical Physics Letters_ , vol. 281, no. 1-3, pp. 140–150, 1997.
* [50] M. K. Sen and P. L. Stoffa, “Bayesian inference, gibbs’ sampler and uncertainty estimation in geophysical inversion,” _Geophysical Prospecting_ , vol. 44, no. 2, pp. 313–350, 1996.
* [51] M. Maraschini and S. Foti, “A monte carlo multimodal inversion of surface waves,” _Geophysical Journal International_ , vol. 182, no. 3, pp. 1557–1566, 2010.
* [52] A. Kone and D. A. Kofke, “Selection of temperature intervals for parallel-tempering simulations,” _The Journal of chemical physics_ , vol. 122, no. 20, p. 206101, 2005.
* [53] F. Calvo, “All-exchanges parallel tempering,” _The Journal of chemical physics_ , vol. 123, no. 12, p. 124106, 2005.
* [54] B. Miasojedow, E. Moulines, and M. Vihola, “An adaptive parallel tempering algorithm,” _Journal of Computational and Graphical Statistics_ , vol. 22, no. 3, pp. 649–664, 2013.
* [55] M. Sambridge, “A parallel tempering algorithm for probabilistic sampling and multimodal optimization,” _Geophysical Journal International_ , vol. 196, no. 1, pp. 357–374, 2013.
* [56] L. Lamport, “On interprocess communication,” _Distributed computing_ , vol. 1, no. 2, pp. 86–101, 1986.
* [57] Y. Li, M. Mascagni, and A. Gorin, “A decentralized parallel implementation for parallel tempering algorithm,” _Parallel Computing_ , vol. 35, no. 5, pp. 269 – 283, 2009.
* [58] K. Karimi, N. Dickson, and F. Hamze, “High-performance physics simulations using multi-core cpus and gpgpus in a volunteer computing context,” _The International Journal of High Performance Computing Applications_ , vol. 25, no. 1, pp. 61–69, 2011.
* [59] G. Mingas, L. Bottolo, and C.-S. Bouganis, “Particle mcmc algorithms and architectures for accelerating inference in state-space models,” _International Journal of Approximate Reasoning_ , vol. 83, pp. 413–433, 2017\.
* [60] A. Reid, E. V. Bonilla, L. McCalman, T. Rawling, and F. Ramos, “Bayesian joint inversions for the exploration of Earth resources,” in _IJCAI_ , 2013, pp. 2877–2884.
* [61] H. Jeffreys, “An invariant form for the prior probability in estimation problems,” _Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences_ , pp. 453–461, 1946.
* [62] N. Singh, L.-M. Browne, and R. Butler, “Parallel astronomical data processing with python: Recipes for multicore machines,” _Astronomy and Computing_ , vol. 2, pp. 1–10, 2013.
* [63] M. Dalla Mura, S. Prasad, F. Pacifici, P. Gamba, J. Chanussot, and J. A. Benediktsson, “Challenges and opportunities of multimodality and data fusion in remote sensing,” _Proceedings of the IEEE_ , vol. 103, no. 9, pp. 1585–1601, 2015.
* [64] A. Beskos, M. Girolami, S. Lan, P. E. Farrell, and A. M. Stuart, “Geometric mcmc for infinite-dimensional inverse problems,” _Journal of Computational Physics_ , vol. 335, pp. 327–351, 2017.
* [65] T. A. Cross and M. A. Lessenger, “Construction and application of a stratigraphic inverse model,” in _Numerical Experiments in Stratigraphy: Recent Advances in Stratigraphic and Sedimentologic Computer Simulations_ , J. Harbaugh, W. Watney, E. Rankey, R. Slingerland, R. Goldstein, and E. Franseen, Eds. Special publications of Society for Sedimentary Geology, 1999, ch. 62, pp. 69–83.
|
capbtabboxtable[][] [table]capposition=top
# Rule Covering for Interpretation and Boosting
Ş. İlker Birbil
Econometric Institute
Erasmus University Rotterdam
3000 DR Rotterdam, The Netherlands
<EMAIL_ADDRESS>
Mert Edali
Department of Medicine
University of Chicago
Chicago, IL 60637, USA
<EMAIL_ADDRESS>
Birol Yüceoğlu
Migros T.A.Ş
Ataşehir, 34758, İstanbul, Turkey
<EMAIL_ADDRESS>
Corresponding author.
###### Abstract
We propose two algorithms for interpretation and boosting of tree-based
ensemble methods. Both algorithms make use of mathematical programming models
that are constructed with a set of rules extracted from an ensemble of
decision trees. The objective is to obtain the minimum total impurity with the
least number of rules that cover all the samples. The first algorithm uses the
collection of decision trees obtained from a trained random forest model. Our
numerical results show that the proposed rule covering approach selects only a
few rules that could be used for interpreting the random forest model.
Moreover, the resulting set of rules closely matches the accuracy level of the
random forest model. Inspired by the column generation algorithm in linear
programming, our second algorithm uses a rule generation scheme for boosting
decision trees. We use the dual optimal solutions of the linear programming
models as sample weights to obtain only those rules that would improve the
accuracy. With a computational study, we observe that our second algorithm
performs competitively with the other well-known boosting methods. Our
implementations also demonstrate that both algorithms can be trivially coupled
with the existing random forest and decision tree packages.
## 1 Introduction
Although their prediction performances are remarkable, tree-based ensemble
methods are difficult to interpret due to large number of trees in the trained
model. Among these methods, Random Forest algorithm is probably the most
frequently used alternative for solving classification and regression problems
arising in different domains [2]. This performance versus interpretability
trade-off leads to the following two main questions in our work: Given a
trained random forest, can we obtain an interpretable representation with
mathematical programming that shows a performance close-enough to the overall
forest? Using a similar mathematical programming model, can we come up with a
learning algorithm, which generates only those rules that improve the
performance of a base tree?
To answer these questions, we focus on the rules constituting the trees in the
forest and propose two algorithms based on mathematical programming models.
The model in the first algorithm extracts the rules corresponding to the
leaves of the forest that result in the minimum total impurity while covering
all the samples. With this algorithm, we are able to mimick the performance of
the original forest with a significantly few number of rules (most of the time
even less than the number of rules in a regular decision tree). Being
encouraged with these results, we then propose our second algorithm based on
selective generation of rules. This algorithm improves the total impurity
gradually by solving a series of linear programming problems. This approach is
a variant of the column generation algorithm in mathematical optimization. In
our scheme, we use the dual optimal solutions of the linear programs as the
sample weights. We also observe that this scheme has close ties to well-known
ideas in boosting methods [13, 14, 17]. Thus, rule extraction and boosting are
the two keywords that help us to review the related literature.
Friedman and Popescu [9] propose RuleFit method to extract rules from
ensembles. The authors use decision trees as base learners, where each node of
the decision tree corresponds to a rule. They present an ensemble generation
algorithm to obtain a rule set. Then, they solve a linear regression problem
that minimizes the loss with a lasso penalty for selecting rules. Meinshausen
[15] presents a similar approach, called Node Harvest, which relies on rules
generated by a tree ensemble. Node Harvest minimizes the quadratic loss
function by partitioning the samples and relaxing the integrality constraints
on the weights. Even though the algorithm does not constrain the number of
rules, the author observes that only a small number of rules are used.
Mashayekhi and Gras [11] propose coverage of the samples in the training set
by the rules generated by Random Forest algorithm. They create a score for
each rule by considering the percentage of correctly and incorrectly
classified samples as well as the rule length. By using a hill climbing
algorithm, they are able to obtain slightly lower accuracy with much fewer
rules than random forest. Extending this work, Mashayekhi and Gras [12]
propose three different algorithms to extract if-then rules from decision tree
ensembles. The authors then compare their algorithms with both RuleFit and
Node Harvest. All these approaches use accuracy values to construct the
objective functions. Interpretability in terms of reducing the number of rules
is achieved by using lasso penalties. Our first algorithm, on the other hand,
reduces the number of rules directly through the objective function.
Gradient boosting algorithms use a sequence of weak learners and build a model
iteratively [7, 13, 8]. At each iteration the boosting algorithm focuses on
the misclassified samples by adjusting their weights with respect to the
errors of the previous iterations. A boosting method that uses linear
programming is called LPBoost [5]. This algorithm uses weak learners to
maximize the margins separating the different classes. Like us, the authors
also use column generation to create the weak learners iteratively. The final
ensemble then becomes a weighted combination of the weak learners. The
approach can accept decision stumps or decision trees as weak learners.
However, decision trees can perform badly in terms of convergence and solution
accuracy, if the dual variables associated with margin classification are not
constrained. Unlike LPBoost, we use the duals as sample weights in our column
generation routine. Another boosting method based on mathematical programming
is called IPBoost [16]. This approach uses binary decision variables
corresponding to misclassification by a predetermined margin. Due to binary
decision variables, the authors resort to an integer programming solution
technique called branch-and-bound-and-price. Dash et al. [4] also use column
generation to learn Boolean decision rules in either disjunctive normal form
or conjunctive normal form. They solve a sequence of integer programming
problems for pricing the columns. They choose to restrict the number of
selected clauses as a user-defined hyperparameter for improving
interpretability. Bertsimas and Dunn [1] propose an integer programming
formulation to create optimal decision trees for multi-class classification.
They present two approaches with respect to node splitting. The first approach
is more general but less interpretable. The second more interpretable model
considers the values of one feature for node splitting. One advantage of their
model is that it does not rely on an assumption for the nature of the values
of the features. Günlük et al. [10] propose a binary classification tree
formulation for only categorical features. The idea can be extended to
numerical variables by using thresholding. The approach is less general than
the approach presented by Bertsimas and Dunn [1], but has some advantages like
the number of integer variables being independent of the training dataset
size. Firat et al. [6] propose a column generation based heuristic for
learning decision tree. The approach is based on generating decision paths and
can be used for solving instances with tens of thousands of observations.
In the light of this review, we make the following contributions to the
literature: We propose a new mathematical programming approach for
interpretation and boosting of random forests and decision trees,
respectively. Unlike other work in the literature, the objective function in
our models aims at minimizing both the total impurity and the number of
selected rules. When applied to a trained random forest for interpretation,
our first algorithm obtains significantly few number of rules with an accuracy
level close to the random forest model. Our second algorithm is a new boosting
method based on rule generation and linear programming. As a novel approach,
the algorithm uses dual information to weigh samples in order to increase the
coverage with less impure rules. In our computational study, we demonstrate
that both algorithms are remarkably easy to implement within a widely-used
machine learning package.111(GitHub page) –
https://github.com/sibirbil/RuleCovering Without any fine tuning, we obtain
quite promising numerical results with both algorithms showing the potential
of the rule covering for interpretation and boosting.
## 2 Minimum rule cover: interpretation of forests
Let $(\bm{x}_{i},y_{i})$, $i\in\mathcal{I}$ be the set of samples, where the
vector $\bm{x}_{i}\in\mathcal{X}$ and the scalar $y_{i}\in\mathcal{K}$ denote
the input sample and the output class, respectively. We shall assume in the
subsequent part that the learning problem at hand is a classification problem.
However, our discussion here can be extended to regression problems as well
(we elaborate on this point in Section S.3.222All cross references starting
with letter “S” refer to the supplementary document.).
Suppose that a Random Forest algorithm is trained on this dataset and a
collection of trees is grown. Given one of these trees, we can easily generate
the path $j$ from the root node to a leaf node that results in a subset of
samples $\bm{x}_{i}$, $i\in\mathcal{I}(j)\subseteq\mathcal{I}$. Actually,
these paths constitute the rules that are later used for classification with
majority voting. Each rule corresponds to a sequence of if-then clauses and it
is likely that some of these sequences may appear multiple times. We denote
all such rules in all trees by set $\mathcal{J}$. Clearly, the size of
$\mathcal{J}$ is quite large. Therefore, we next construct a mathematical
programming model that aims at selecting the minimum number of rules from the
trained forest while preserving the performance. We also make sure that all
samples are covered with the selected rules.
As rule $j\in\mathcal{J}$ corresponds to a leaf node, we can also evaluate the
node impurity, $w_{j}$. Two most common measures for evaluating the node
impurity are known as Gini and Entropy criteria. We further introduce the
binary decision variables $z_{j}$ that mark whether the corresponding rules
$j\in\mathcal{J}$ are selected or not. The cost of $w_{j}$ is incurred, when
$z_{j}=1$. Note that if we use only these impurity evaluations as the
objective function coefficients (costs), then the model would tend to select
many rules that cover only a few samples, which are likely to be _pure nodes_.
However, we want to advocate the use of fewer rules in the model so that the
resulting set of rules is easy to interpret. In other words, we also aim at
minimizing the number of rules along with the impurity cost. Therefore, we
replace the cost coefficient so that selecting excessive number of rules is
penalized. Our mathematical programming model then becomes:
$\begin{array}[]{lll}\mbox{minimize\hskip
11.38109pt}&\sum_{j\in\mathcal{J}}(1+w_{j})z_{j}&\\\\[5.69054pt] \mbox{subject
to\hskip 11.38109pt}&\sum_{j\in\mathcal{J}(i)}z_{j}\geq
1,&i\in\mathcal{I},\\\\[5.69054pt]
&z_{j}\in\\{0,1\\},&j\in\mathcal{J},\end{array}$ (1)
where $\mathcal{J}(i)$, $i\in\mathcal{I}$ is the set of rules prescribing all
the leaves involving sample $\bm{x}_{i}$. In other words, sample $i$ is
covered with the rules in subset $\mathcal{J}(i)\subseteq\mathcal{J}$.
The mathematical programming model (1) is a standard weighted set covering
formulation, where sets and items correspond to rules and samples,
respectively. Although set covering problem is NP-hard, current off-the-shelf
solvers are capable of providing optimal solutions to moderately large
instances. In case the integer programming formulation becomes difficult to
solve, there are also powerful heuristics that return approximate solutions
very quickly. One such approach is the well-known greedy algorithm proposed by
Chvatál [3]. Algorithm 3 shows the steps of this heuristic using our notation.
Recently, Chvatál’s greedy algorithm, as we implement here, has been shown to
perform quite well on a set of test problems when compared against several
more recent heuristics [18]. Thus, we have reported our results also with this
algorithm in Section 4. Clearly, this particular heuristic can be replaced
with any other approach solving the set covering problem. Since these
heuristics provide an approximate solution, we denote the resulting set of
rules by $\hat{\mathcal{J}}$, and by definition, the cardinality of set
$\hat{\mathcal{J}}$ is larger than the cardinality of the optimal set of rules
obtained by solving (1).
Algorithm 1 gives the steps of our minimum rule cover algorithm (MIRCO). The
algorithm takes as input the trained random forest model and the training
dataset. Note that after running Algorithm 1, we obtain a subset of rules
$\hat{\mathcal{J}}$ that can also be used for classifying out-of-sample points
(test set) with majority voting. If we denote the predicted class of test
sample $\bm{x}_{0}$ with $C(\bm{x}_{0},\hat{\mathcal{J}})$, then
$C(\bm{x}_{0},\hat{\mathcal{J}})=\arg\max_{k\in\mathcal{K}}\big{\\{}\sum_{j\in\hat{\mathcal{J}}}n_{jk}\mathds{L}_{j}(\bm{x}_{0})\big{\\}},$
(2)
where $n_{jk}$ stands for the number of samples from class $k$ in leaf $j$,
and $\mathds{L}_{j}(\bm{x}_{0})$ is an operator showing whether $\bm{x}_{0}$
satisfies the sequence of clauses corresponding to leaf $j$. In fact, testing
in this manner would be a natural approach for evaluating the performance of
MIRCO. Since our main objective is to interpret the underlying model obtained
with Random Forest algorithm, we hope that this subset of rules would obtain a
classification performance closer to the performance obtained with all the
rules in set $\mathcal{J}$. We show with our computational study in Section 4
that this is indeed the case.
Input : Random forest model RF; training dataset $(\bm{x}_{i},y_{i})$,
$i\in\mathcal{I}$
1 $\mathcal{J}=\emptyset$;
2 for _Tree in RF_ do
3 $\mathcal{J}\leftarrow\mathcal{J}\cup\text{(rules in {Tree})}$;
4
5 end for
6
7Evaluate impurities $w_{j}$, $j\in\mathcal{J}$ using $(\bm{x}_{i},y_{i})$,
$i\in\mathcal{I}$;
8
9Construct sets $\mathcal{J}(i)$, $i\in\mathcal{I}$ ;
10
11Solve model (1) with Algorithm 3;
12
13Return $\hat{\mathcal{J}}$;
Algorithm 1 MIRCO
Figure 1 shows an example set of of rules obtained with MIRCO on a small data
set. Clearly, the resulting set of rules does not correspond to a decision
tree. On one hand, this implies that we cannot simply select one of the trees
grown by the Random Forest algorithm and replace its leaves with the set of
rules obtained with MIRCO. Consequently, multiple rules may cover the same
test sample, in which case we can again use majority voting to determine its
class. Region A in Figure 1 shows that any test sample in that region will be
classified with two overlapping rules. On the other hand, there is also the
risk that the entire feature space is not covered by the subset of rules. In
Figure 1 such a region is marked as B, where none of the rules cover this
region. Therefore, if one decides to use only the rules in $\hat{\mathcal{J}}$
as a classifier, then some of the samples in the test set may not be
classified at all. This is an anticipated behaviour, since MIRCO guarantees to
cover only those samples that are used to train the Random Forest algorithm.
Our numerical results show that the percentages of test samples that MIRCO
fails to classify are quite low.
Figure 1: A toy example illustrating that MIRCO does not necessarily
correspond to a tree. Rectangles denote the regions obtained with the rules in
$\hat{\mathcal{J}}$. The numbers in the rectangles depict the classification
results corresponding to three classes shown with three different markers.
Notice that the entire set of rules used by MIRCO does not have to be
constructed after training a Random Forest algorithm. If there is an oracle
that provides many rules to construct the set $\mathcal{J}$, then we can still
(approximately) solve the mathematical programming model (1) and obtain the
resulting set of rules. In fact, the development of this oracle decouples our
discussion from training a forest. In the next section, we propose such an
oracle that generates the necessary rules on the fly.
## 3 RCBoost: rule cover boosting
The main difficulty with the set covering problem is the integrality condition
on the decision variables. When this condition is ignored, the resulting
problem is a simple linear programming problem that can be solved quite
efficiently. For the same reason, many approximation methods devised for the
set covering problem are based on considering the linear programming
relaxation of (1). Formally, the relaxed problem simply becomes:
$\begin{array}[]{lllc}\mbox{minimize\hskip
11.38109pt}&\sum_{j\in\mathcal{J}}(1+w_{j})z_{j}&&\\\\[5.69054pt]
\mbox{subject to\hskip 11.38109pt}&\sum_{j\in\mathcal{J}(i)}z_{j}\geq
1,&i\in\mathcal{I},\hskip 34.1433pt(\lambda_{i})\\\\[5.69054pt] &z_{j}\geq
0,&j\in\mathcal{J},&\end{array}$ (3)
where $\lambda_{i}$, $i\in\mathcal{I}$ denote the dual variables corresponding
to the constraints. After solving this problem, we obtain both the primal and
the dual optimal solutions. As we will discuss shortly, the optimal dual
variables bear important information about the sample points. Since we relax
the integrality restrictions, one suggestion could be rewriting the objective
as solely the minimization of total impurity. We refrain from such an approach
because when there are many pure leaves, then the corresponding rule does not
contribute to the objective function value ($w_{j}=0$) and the corresponding
$z_{j}$ variable becomes free to attain any value. This implies that the
primal problem has multiple optimal solutions, and hence causing degeneracy in
the dual.
We have seen in the previous section that problem (3) may easily have quite
many variables (columns). One standard approach to solve this type of linear
programming problems is based on iterative generation of only necessary
variables that would lead to an improvement in the objective function value.
This approach is known as column generation. The main idea is to start with a
set of columns (column pool) and form a linear programming problem, called the
restricted master problem (RMP). After solving the restricted master problem,
the dual optimal solution is obtained. Then using this dual solution, a
pricing subproblem is solved to find the columns with negative reduced costs.
These columns are the only candidates for improving the objective function
value when they are added to the column pool. The next iteration continues
with obtaining the optimal dual solution of the new restricted master problem
with the extended column pool. Again, the pricing subproblem is solved to find
the columns with negative reduced costs. If there is no such column, then the
column generation algorithm terminates.
Suppose that we denote the column pool in the restricted master problem at
iteration $t$ of column generation algorithm by $\mathcal{J}_{t}$. Then, we
solve the following linear programming problem:
$\begin{array}[]{lllc}\mbox{minimize\hskip
11.38109pt}&\sum_{j\in\mathcal{J}_{t}}(1+w_{j})z_{j}&&\\\\[5.69054pt]
\mbox{subject to\hskip 11.38109pt}&\sum_{j\in\mathcal{J}_{t}(i)}z_{j}\geq
1,&i\in\mathcal{I},&\hskip 11.38109pt(\lambda^{t}_{i})\\\\[5.69054pt]
&z_{j}\geq 0,&j\in\mathcal{J}_{t},&\end{array}$ (RMP$(\mathcal{J}_{t})$)
where $\lambda^{t}_{i}$ denotes the dual variable corresponding to constraint
$i\in\mathcal{I}$ at iteration $t$, and $\mathcal{J}_{t}(i)$ stands for the
rules in $\mathcal{J}_{t}$ that cover sample $i$. The reduced cost of rule
$j\in\mathcal{J}_{t}$ is simply given by
$\bar{w}_{j}=(1+w_{j})-\sum_{i\in\mathcal{I}(j)}\lambda^{t}_{i}.$ (4)
The necessary rules that would improve the objective function value in the
next iteration are the ones with negative reduced costs. In all this
discussion, the key point is, in fact, the pricing subproblem, which would
generate the rules with negative reduced costs. Recall that the columns in
this approach are constructed by the rules in our setting. Thus, in the
remaining part of this section, we shall use the term “rule pool” instead of
“column pool.”
Algorithm 2 shows the training process of the proposed rule cover boosting
(RCBoost) algorithm. The steps of the algorithm is also given in Figure 2.
Here, DecisionTree routine plays the role of the pricing subproblem. The
parameter of this routine is a vector, which is used to assign different
weights to the samples. To obtain rules with negative reduced costs, the dual
optimal solution of the restricted master problems is passed as the sample
weight vector (line 2). If there are rules with negative reduced costs (line
2), then these are added to the rule pool $\mathcal{J}^{*}$ in line 2.
Otherwise, the algorithm terminates (line 2). Note that a bound on the maximum
number of RMP calls can also be considered as a hyperparameter. The resulting
set or rules in the final pool constitute the classifier (line 2). Like in
Random Forest algorithm, a test point is classified using the majority voting
system among all the rules (if-then clauses) that are satisfied. Note that
there is no danger of failing to classify a test sample like MIRCO, since the
initial rule pool is constructed with a decision tree (line 2). The output set
of rules $\mathcal{J}^{*}$ is then used for classifying out-of-sample point
with majority voting as in (2). That is, $C(\bm{x}_{0},\mathcal{J}^{*})$ is
the predicted class for $\bm{x}_{0}$.
Input : Training data $(\bm{x}_{i},y_{i})$, $i\in\mathcal{I}$
1 $\bar{\bm{\lambda}}=(1,1,\dots,1)$;
2 $\mathcal{J}^{*}\leftarrow\text{{DecisionTree}}(\bar{\bm{\lambda}})$;
3 for _$t=1,2,\dots$_ do
4 $\bm{\lambda}^{t}\leftarrow\text{{RMP}}(\mathcal{J}^{*})$;
5 $\bar{\bm{\lambda}}\leftarrow\bar{\bm{\lambda}}+\bm{\lambda}^{t}$;
6 $\bar{\mathcal{J}}\leftarrow\text{{DecisionTree}}(\bar{\bm{\lambda}})$;
7 $\bar{\mathcal{J}}^{-}=\\{j\in\bar{\mathcal{J}}~{}:~{}\bar{w_{j}}<0\\}$;
8 if _$\bar{\mathcal{J}}^{-}=\emptyset$_ then
9 break;
10
11 end if
12 $\mathcal{J}^{*}\leftarrow\mathcal{J}^{*}\cup\bar{\mathcal{J}}^{-}$;
13
14 end for
15Return $\mathcal{J}^{*}$;
Algorithm 2 RCBoost Figure 2: Flowchart of RCBoost
As line 2 shows, we accumulate the dual vectors at every iteration. This is
because many of the dual variables quickly become zero as either the
corresponding primal constraints are inactive or the dual optimal solution
becomes degenerate. So, we make sure with the accumulation that all the sample
weights remain nonzero. We should also note that calling the decision tree
algorithm with (cumulative) sample weights is a proxy to the pricing
subproblem. Because, the actual pricing subproblem is supposed to search for
the rules with negative reduced costs. However, we argue that using such a
proxy is still sensible. We appeal to the theory of sensitivity analysis in
linear programming and the role of sample weights in boosting algorithms. An
optimal dual variable shows the change in the optimal objective function, when
the right-hand side of this constraint (all ones in our case) is slightly
adjusted. Since each constraint corresponds to the samples in (1), the
corresponding optimal dual variables show us the role of these samples for
changing the objective function value. In fact the reduced cost calculation in
(4) also bears the same intuition and promotes having a rule with low impurity
and large dual variables. Thus, if the samples with large dual values
(weights) belong to the same class, then they are more likely to be classified
within the same leaf (rule) to decrease the impurity. On the other hand, if
these samples belong to different classes, then putting them into separate
leaves also yields a decrease in impurity. This is precisely what the sample
weights try to achieve for boosting methods, where the weights for the
incorrectly classified samples are increased. For instance, let
$w^{\lambda}_{j}$ denote the weighted Gini impurity for leaf $j\in\mathcal{J}$
given by
$w^{\lambda}_{j}=1-\sum_{k\in\mathcal{K}}\left(\frac{\sum_{i\in\mathcal{I}(j)}\mathds{1}_{k}(i)+\sum_{i\in\mathcal{I}(j)}\mathds{1}_{k}(i)\sum^{t}_{s=1}\lambda_{i}^{s})}{|\mathcal{I}(j)|+\sum_{i\in\mathcal{I}(j)}\sum^{t}_{s=1}\lambda_{i}^{s}}\right)^{2}.$
where $|\mathcal{I}(j)|$ denotes the cardinality of set $\mathcal{I}(j)$ and
the indicator operator $\mathds{1}_{k}(i)$ is one when sample $i$ belongs to
class $k$; otherwise, it is zero. This weighted impurity is minimized whenever
the samples with large accumulated duals from the same class are classified
within the same leaf. This choice is likely to trigger generation of leaves
with negative reduced costs, since the reduced cost value (4) is also
minimized when the same class of samples (lower Gini impurity) with large
duals are covered by rule $j$.
Using sample weights with decision trees also has a clear advantage from
implementation point of view. Almost all existing tools available for
efficiently training decision trees admit weights for the samples so that
boosting algorithms, like the well-known AdaBoost [7], can make use of this
feature. Those algorithms increase the weights of misclassified samples in an
iterative manner and obtain a sequence of classifiers. Then, these classifiers
are further weighed according to their classification performances. RCBoost,
however, only produces a sequence of rules by increasing weights and each rule
is evaluated by its impurity contribution to the objective function through
its reduced cost. Thus, the reduced costs play the role of the partial
derivatives like in gradient boosting algorithms. These observations naturally
beg for a discussion about the relationship between our algorithm RCBoost and
the other boosting algorithms. We reserve Section S.2 to elaborate on this
point.
## 4 Numerical results
We next present our computational study with MIRCO and RCBoost algorithms on
15 datasets with varying characteristics. We have implemented both algorithms
in Python 333(GitHub page) – https://github.com/sibirbil/RuleCovering. The
details of these datasets are given in Table S.1. While most of these datasets
correspond to binary classification problems, there are also problems with
three classes (seeds and wine), six classes (glass) and eight classes (ecoli).
In the literature, mammography, diabetes, oilspill, phoneme, glass and ecoli
are considered as imbalanced datasets.
MIRCO is compared against Decision Tree (DT) and Random Forest (RF) algorithms
in terms of accuracy and the number of rules (interpretability). We also
compare results of RCBoost (RCB) against RF, AdaBoost (ADA) and Gradient
Boosting (GB). For the latter two boosting methods, we use decision trees as
base estimators. We apply $10\times 4$ nested stratified cross validation in
all experiments with the following options for hyperparameter tuning:
$\mathtt{maxdepth}\in\left\\{5,10,20\right\\}$,
$\mathtt{numberoftrees}\in\left\\{10,50,100\right\\}$. We also use
$\mathtt{numberofRMPcalls}\in\left\\{5,10,50,100,200\right\\}$ option to set a
bound on the number of RMP calls.
We first report our results with DT, RF and MIRCO algorithms in Figure 3. As
the accuracy values (average over 10 replications) in Figure 3a show, MIRCO
performs on par with RF on most of the datasets (see Table S.2 for the
details). Recall that MIRCO is not guaranteed to classify all the test
samples. Thus, we also provide the fraction of the test samples missed by
MIRCO (Figure 3b). However, we still classify such a sample by applying the
rules that have the largest fraction of accepted clauses. Therefore, the
reported comparisons against RF and DT in Figure 3a are carried on all
samples. We observe that the fraction of the missed test samples is below 6%
for all datasets but one. This shows that MIRCO does not only approximate the
RF in terms of accuracy, but also covers a very high fraction of the feature
space. Figure 3c depicts the numbers of rules in log-scale to give an
indication about the interpretability. In all but one dataset, MIRCO has
generated less number of rules than DT. Furthermore, the standard deviation
figures in Table S.2 show that MIRCO shows less variability than DT in all but
two of the datasets.
(a) (b) (c)
Figure 3: (a) Accuracy values for RF, DT, and MIRCO. (b) Average percentage of
missed samples by MIRCO. (c) Average number of rules (in logarithmic scale)
generated by RF, DT, and MIRCO.
To compare RCB, we present the average accuracy values of RF, ADA, GB in
Figure 4 (see Table S.3 for the details). We also report the results with the
initial Decision Tree (iniDT) used in RCB algorithm (see line 2 in Algorithm
2) so that we can observe how rule generation improves performance. The number
in parentheses above each dataset shows the average number of RMP calls by
RCB. These results demosntrate that RCB exhibits a quite competitive
performance. In particular, RCB outperforms both boosting algorithms, ADA and
GB in four problems. One critical point here is that RCB makes only a small
number of RMP calls. In all but one datasets, RCB improves the performance of
the base decision tree, iniDT by using the proposed rule generation.
Figure 4: RCBoost results
## 5 Conclusion
We have proposed two algorithms for interpretation and boosting. Both
algorithms are based on mathematical programming approaches for covering the
rules in decision trees. The first algorithm is used for interpreting a model
trained with the Random Forest algorithm. The second algorithm focuses on the
idea of generating necessary rules to boost an initial model trained with the
Decision Tree algorithm. Both algorithms are quite straightforward to
implement within the existing packages. Even with our first implementation
attempt, we have obtained promising results. The first algorithm has
successfully extracted much less number of rules than the random forest model
leading to better interpretability. Moreover, we have observed that the
accuracy levels obtained with the extracted rules of our first algorithm
closely follow the accuracy levels of the trained random forest model. When
compared against other boosting algorithms, our second algorithm has shown a
competitive performance. As a new boosting algorithm, we have also provided a
discussion about the relationship between our approach and other boosting
algorithms. The idea of selecting rules for interpretation or boosting leads
to a plethora of extensions that could be further studied. Therefore, we have
prepared a separate supplementary discussion on extensions in Section S.3.
This section complements our discussion as well as lists several future
research paths.
## References
* Bertsimas and Dunn [2017] D. Bertsimas and J. Dunn. Optimal classification trees. _Machine Learning_ , 106(7):1039–1082, 2017.
* Breiman [2001] L. Breiman. Random forests. _Machine Learning_ , 45(1):5–32, 2001.
* Chvatal [1979] V. Chvatal. A greedy heuristic for the set-covering problem. _Mathematics of Operations Research_ , 4(3):233–235, 1979.
* Dash et al. [2018] S. Dash, O. Gunluk, and D. Wei. Boolean decision rules via column generation. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, _Advances in Neural Information Processing Systems_ , pages 4655–4665. 2018.
* Demiriz et al. [2002] A. Demiriz, K. P. Bennett, and J. Shawe-Taylor. Linear programming boosting via column generation. _Machine Learning_ , 46(1-3):225–254, 2002.
* Firat et al. [2020] M. Firat, G. Crognier, A. F. Gabor, C. Hurkens, and Y. Zhang. Column generation based heuristic for learning classification trees. _Computers & Operations Research_, 116:104866, 2020.
* Freund and Schapire [1997] Y. Freund and R. E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. _Journal of Computer and System Sciences_ , 55(1):119 – 139, 1997.
* Friedman [2001] J. H. Friedman. Greedy function approximation: A gradient boosting machine. _Annals of Statistics_ , 29(5):1189–1232, 2001\.
* Friedman and Popescu [2008] J. H. Friedman and B. E. Popescu. Predictive learning via rule ensembles. _The Annals of Applied Statistics_ , 2(3):916–954, 2008.
* Günlük et al. [2018] O. Günlük, J. Kalagnanam, M. Menickelly, and K. Scheinberg. Optimal decision trees for categorical data via integer programming. _arXiv preprint arXiv:1612.03225_ , 2018.
* Mashayekhi and Gras [2015] M. Mashayekhi and R. Gras. Rule extraction from random forest: the rf+hc methods. In D. Barbosa and E. Milios, editors, _Advances in Artificial Intelligence_ , pages 223–237, Cham, 2015. Springer International Publishing.
* Mashayekhi and Gras [2017] M. Mashayekhi and R. Gras. Rule extraction from decision trees ensembles: New algorithms based on heuristic search and sparse group lasso methods. _International Journal of Information Technology & Decision Making_, 16(06):1707–1727, 2017.
* Mason et al. [1999] L. Mason, J. Baxter, P. Bartlett, and M. Frean. Boosting algorithms as gradient descent. In S. Solla, T. Leen, and K. Müller, editors, _Advances in Neural Information Processing Systems_ , pages 512–518. 1999.
* Mason et al. [2000] L. Mason, , P. L. Bartlett, and J. Baxter. Improved generalization through explicit optimization of margins. _Machine Learning_ , 38:243–255, 2000.
* Meinshausen [2010] N. Meinshausen. Node harvest. _The Annals of Applied Statistics_ , 4(4):2049–2072, 2010.
* Pfetsch and Pokutta [2020] M. E. Pfetsch and S. Pokutta. IPBoost–non-convex boosting via integer programming. _arXiv preprint arXiv:2002.04679_ , 2020.
* Schapire et al. [1998] R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: A new explanation for the effectiveness of voting methods. _The Annals of Statistics_ , 26(5):1651–1686, 1998.
* Vasko et al. [2016] F. J. Vasko, Y. Lu, and K. Zyma. What is the best greedy-like heuristic for the weighted set covering problem? _Operations Research Letters_ , 44:366–369, 2016.
Supplementary Material for
“Rule Covering for Interpretation and Boosting”
Note that most of the cross references in this supplementary file refer to the
original manuscript. Therefore, clicking on those references may not take you
to the desired location.
## S.1 Chvatál’s greedy algorithm
We restate the well-known heuristic of Chvatál in our own notation. Recall
that the sets $\mathcal{I}$ and $\mathcal{J}$ stand for the set of samples and
the set of rules, respectively. Algorithm 3 gives the steps of the greedy
heuristic, where $\mathcal{I}_{\mathcal{R}}(j)$ represents those samples that
are in $\mathcal{R}$ covered by rule $j\in\mathcal{J}$. In line 3, the cost of
selecting rule $j$ is $1+w_{j}$, and the operation
$|\mathcal{I}_{\mathcal{R}}(j)|$ denotes the cardinality of the set
$\mathcal{I}_{\mathcal{R}}(j)$. Removal of the redundant rules in line 3 is
done in two steps: First, the rules in $\bar{\mathcal{J}}$ are sorted in
ascending order of their costs in a list. Then, the rules at the end of this
list are removed as long as the remaining rules cover all the samples.
Result: Set of rules $\hat{\mathcal{J}}$
1
2$\mathcal{R}=\mathcal{I}$, $\hat{\mathcal{J}}=\emptyset$;
3
4while _$\mathcal{R}\neq\emptyset$_ do
5
6
$j^{*}=\arg\underset{j\in\mathcal{J}\backslash\hat{\mathcal{J}}}{\min}\left\\{\frac{1+w_{j}}{|\mathcal{I}_{\mathcal{R}}(j)|}:\mathcal{I}_{\mathcal{R}}(j)\neq\emptyset\right\\}$;
7
8
$\mathcal{R}\leftarrow\mathcal{R}\backslash\mathcal{I}_{\mathcal{R}}(j^{*})$;
9
10 $\hat{\mathcal{J}}\leftarrow\hat{\mathcal{J}}\cup j^{*}$;
11
12 end while
13
14Remove redundant columns from $\hat{\mathcal{J}}$;
15
16Return $\hat{\mathcal{J}}$;
Algorithm 3 Greedy
## S.2 Relation to other boosting algorithms
The use of sample weights and reduced costs shows a similar trait among
RCBoost and other boosting algorithms. The reduced cost of rule $j$ can be
considered as the partial derivative of the objective function with respect to
the corresponding variable $z_{j}$. Thus, rules that have negative reduced
costs are added to the rule pool. This resembles a descent iteration. To
establish the relationship between RCBoost and other boosting methods from the
gradient descent point of view [13], we first define the Lagrangean function
$\mathcal{L}(\mathbf{z},\bm{\lambda})=\sum_{j\in\mathcal{J}}\big{(}1+w_{j}\big{)}z_{j}+\sum_{i\in\mathcal{I}}\lambda_{i}\big{(}1-\sum_{j\in\mathcal{J}(i)}z_{j}\big{)},$
where $\mathbf{z}$ and $\bm{\lambda}$ denote the primal and dual vectors.
Suppose that the optimal solution pair for (RMP$(\mathcal{J}_{t})$) is given
by $\mathbf{z}^{t}$ and $\bm{\lambda}^{t}$. Then, we have
$\begin{array}[]{rl}\mathcal{L}((\mathbf{z}^{t},\mathbf{z}^{c}),\bm{\lambda}^{t})&=\sum_{i\in\mathcal{I}}\lambda^{t}_{i}+\sum_{j\in\mathcal{J}_{t}}\bar{w}_{j}z^{t}_{j}+\sum_{j\in\mathcal{J}/\mathcal{J}_{t}}\bar{w}_{j}z^{t}_{j}\\\\[5.69054pt]
&=\sum_{i\in\mathcal{I}}\lambda^{t}_{i}+\sum_{j\in\mathcal{J}/\mathcal{J}_{t}}\bar{w}_{j}z^{c}_{j},\end{array}$
where $\bar{w}_{j}$ is the reduced cost of rule $j$ as in (4) and the tuple
$(\mathbf{z}^{t},\mathbf{z}^{c})$ shows the variables in $\mathcal{J}_{t}$ and
$\mathcal{J}/\mathcal{J}_{t}$, respectively. The last equality is a direct
consequence of optimality conditions, since $\mathbf{z}^{t}$ and
$\bm{\lambda}^{t}$ are complementary solution vectors. Thus, taking the
steepest descent step for $j\in\mathcal{J}/\mathcal{J}_{t}$ simply implies
involving those rules with negative reduced costs in the next iteration. From
this perspective, RCBoost is similar to gradient boosting algorithms.
Nonetheless, unlike those algorithms RCBoost does not fit a weak-classifier
but instead obtains a larger set $\mathcal{J}_{t+1}$ by adding a subset of
rules with negative reduced costs (line 2 of Algorithm 2). Then, solving the
linear programming problem over $\mathcal{J}_{t+1}$ yields the parameters
$\mathbf{z}^{t+1}$ and $\bm{\lambda}^{t+1}$. Suppose we denote the optimal
objective function of (RMP$(\mathcal{J}_{t})$) by $\Phi(\mathcal{J}_{t})$,
which is the total impurity that we use to evaluate the training loss of our
classifier. With each iteration we improve this loss, since
$\Phi(\mathcal{J}_{t})\geq\Phi(\mathcal{J}_{t+1})$ due to our linear
programming formulation. If we further define the subset
$\partial\mathcal{J}_{t}\subseteq\\{j\in\mathcal{J}:\bar{w}_{j}<0\\},$
then
$\Phi(\mathcal{J}_{t+1})=\Phi(\mathcal{J}_{t}\cup\partial\mathcal{J}_{t})$.
There has also been a focus on the margin maximization idea to discuss the
generalization performance of boosting methods [14, 17]. Here, one of the
fundamental points is to observe that a sample with a large margin is likely
to be classified correctly. Thus, margin maximization is about assigning
larger weights to those samples with small margins. RCBoost also implicitly
aims at obtaining large margins through selection of rules with less
impurities (classification errors decrease). In other words, minimizing the
objective function of the restricted master problem promotes large margins.
Formally, we can rewrite the optimal objective function value of
(RMP$(\mathcal{J}_{t})$) as
$\sum_{j\in\mathcal{J}_{t}}(1+w_{j})z^{t}_{j}=\sum_{i\in\mathcal{I}}\sum_{j\in\mathcal{J}_{t}(i)}\mu_{j}z^{t}_{j}=\sum_{i\in\mathcal{I}}\lambda_{i}^{t},$
where $\mu_{j}=\frac{1+w_{j}}{|\mathcal{I}(j)|}$, $j\in\mathcal{J}$. Using
complementary slackness condition, we know that $\lambda_{i}^{t}>0$ only if
$\sum_{j\in\mathcal{J}_{t}(i)}z_{j}^{t}=1$. Thus,
$\sum_{j\in\mathcal{J}_{t}(i)}\mu_{j}z^{t}_{j}$ is the convex combination of
margins obtained with different rules for sample $i$. The larger the obtained
impurity, the smaller the corresponding margin. Like other boosting methods,
RCBoost also assigns larger weights to those samples with smaller margins.
## S.3 Some extensions
We have used classification for our presentation. Nonetheless, both proposed
algorithms can also be adjusted to solve regression problems. To interpret
random forest regressors with MIRCO, the first step is to use a proper
criterion, like mean squared error, for evaluating $w_{j}$, $j\in\mathcal{J}$
values. Then, MIRCO can be used as it is given in Algorithm 1. The same
criteria can also be used in RCBoost to construct the restricted master
problem. When a regression problem is solved, one needs to pay attention to
two potential caveats: (i) The two components of the objective function
coefficients in optimization model (1) or (3) may not be balanced well, since
the criteria values $w_{j}$, $j\in\mathcal{J}$ are not bounded above by one.
This could be overcome by scaling the criteria values. If scaling is done with
a multiplier, then this multiplier becomes a new hyperparameter which may
require tuning. (ii) Regression trees tend to have too many leaves. Even after
applying MIRCO, the cardinality of the resulting set of rules
$\hat{\mathcal{J}}$ may be quite large, and hence, interpreting the random
forest regressor may be difficult. Similarly, RCBoost may make too many calls
to the restricted master problem and end up with a large number of rules in
the final rule pool $\mathcal{J}^{*}$.
Many boosting algorithms use decision trees as their weak estimators. Thus, it
seems that MIRCO can also be used with those algorithms to interpret their
results. However, one needs to pay attention to the fact that boosting
algorithms assign different weights to their weak estimators and MIRCO does
not take these weights into account. Therefore, if rule covering is used for
interpretation of boosting algorithms, Algorithm 1 should be adjusted to
incorporate the estimator weights.
Recall that we use a decision tree with sample weights in RCBoost as a proxy
to an actual pricing subproblem (line 2, Algorithm 2). An alternative could be
using the impurity and the negative reduced cost together while constructing
the decision tree. In that case, splitting can be done from non-dominated bi-
criteria values; i.e., impurity and negative reduced cost values.
Clearly, we can also use a random forest or any ensemble of trees with sample
weights as a proxy to the pricing subproblem. Even further, one may even model
a separate combinatorial problem to generate many different rules. At this
point, it is important to keep in mind that considering all possible rules
with negative reduced costs could prove to be a daunting job.
In its current stage, RCBoost adds all the rules with negative reduced costs
to the rule pool. An alternative approach could be to include only the rule
with the most negative reduced cost. Along the same vein, we may also remove
those rules that have high (positive) reduced costs from the column pool.
Actually, these are typical approaches in column generation to keep the size
of the column pool manageable. As the size of the rule pool becomes large,
RCBoost also becomes less interpretable. Fortunately, one can easily apply
MIRCO or other rule extraction approaches from the literature such as the
RuleFit algorithm [9] to a trained RCBoost. In that case, Algorithm 1 can be
applied starting from line 1 with $\mathcal{J}\leftarrow\mathcal{J}^{*}$. As a
final remark, we note that RCBoost treats the rules equally when a test point
is classified with the satisfied rules in $\mathcal{J}^{*}$. An alternative
could be giving different weights to different rules. These rule weights could
be evaluated by making use of the optimal values of $z_{j}$,
$j\in\mathcal{J}^{*}$ obtained from the final restricted master problem.
However, all these changes should be carefully contemplated, since the
resulting rule pool after shrinking may not necessarily cover the feature
space as we have discussed in Section 2 (see also Figure 1).
## S.4 Details of numerical results
Table S.1: Characteristics of the datasets | Number of | Number of | Number of
---|---|---|---
Dataset | Features | Classes | Samples
banknote | 4 | 2 | 1372
ILDP | 10 | 2 | 583
ionosphere | 34 | 2 | 351
transfusion | 4 | 2 | 748
liver | 6 | 2 | 345
tic-tac-toe | 9 | 2 | 958
WDBC | 30 | 2 | 569
mammography | 6 | 2 | 11183
diabetes | 8 | 2 | 768
oilspill | 48 | 2 | 937
phoneme | 5 | 2 | 5427
seeds | 7 | 3 | 210
wine | 13 | 3 | 178
glass | 10 | 6 | 214
ecoli | 7 | 8 | 336
Table S.2: MIRCO results | DT | RF | MIRCO
---|---|---|---
Dataset | Accuracy | Accuracy | Accuracy | % of Missed Pts
banknote | 0.99 (0.02) | 0.99 (0.01) | 0.98 (0.01) | 0.01 (0.01)
ILDP | 0.66 (0.07) | 0.71 (0.06) | 0.69 (0.03) | 0.06 (0.04)
ionosphere | 0.87 (0.07) | 0.92 (0.03) | 0.91 (0.04) | 0.02 (0.03)
transfusion | 0.79 (0.04) | 0.77 (0.04) | 0.76 (0.01) | 0.01 (0.01)
liver | 0.65 (0.09) | 0.73 (0.06) | 0.66 (0.06) | 0.06 (0.05)
tic-tac-toe | 0.94 (0.02) | 0.99 (0.01) | 0.97 (0.02) | 0.02 (0.02)
WDBC | 0.93 (0.03) | 0.96 (0.02) | 0.95 (0.03) | 0.01 (0.02)
mammography | 0.98 (0.00) | 0.99 (0.00) | 0.98 (0.00) | 0.00 (0.00)
diabetes | 0.74 (0.04) | 0.77 (0.04) | 0.73 (0.04) | 0.05 (0.04)
oilspill | 0.96 (0.02) | 0.96 (0.01) | 0.96 (0.01) | 0.01 (0.01)
phoneme | 0.88 (0.01) | 0.91 (0.01) | 0.88 (0.02) | 0.03 (0.01)
seeds | 0.91 (0.06) | 0.91 (0.09) | 0.91 (0.05) | 0.02 (0.03)
wine | 0.91 (0.07) | 0.98 (0.03) | 0.91 (0.07) | 0.04 (0.04)
glass | 0.67 (0.05) | 0.79 (0.07) | 0.71 (0.07) | 0.09 (0.06)
ecoli | 0.81 (0.09) | 0.88 (0.06) | 0.76 (0.08) | 0.03 (0.03)
∗ Numbers in parantheses show the standard deviations |
Table S.3: RCBoost results | RF | ADA | GB | iniDT | RCB
---|---|---|---|---|---
Dataset | Accuracy | Accuracy | Accuracy | Accuracy | Accuracy
banknote | 0.99 (0.01) | 1.00 (0.00) | 0.99 (0.01) | 0.99 (0.01) | 0.98 (0.01)
ILDP | 0.71 (0.06) | 0.70 (0.06) | 0.69 (0.03) | 0.66 (0.07) | 0.70 (0.04)
ionosphere | 0.92 (0.03) | 0.94 (0.05) | 0.92 (0.04) | 0.88 (0.03) | 0.89 (0.05)
transfusion | 0.77 (0.04) | 0.74 (0.04) | 0.77 (0.03) | 0.78 (0.03) | 0.78 (0.04)
liver | 0.73 (0.06) | 0.68 (0.07) | 0.68 (0.06) | 0.64 (0.10) | 0.71 (0.05)
tic-tac-toe | 0.99 (0.01) | 1.00 (0.01) | 0.99 (0.01) | 0.95 (0.02) | 0.99 (0.01)
WDBC | 0.96 (0.02) | 0.97 (0.03) | 0.96 (0.03) | 0.92 (0.03) | 0.94 (0.02)
mammography | 0.99 (0.00) | 0.99 (0.00) | 0.99 (0.00) | 0.98 (0.00) | 0.99 (0.00)
diabetes | 0.77 (0.04) | 0.74 (0.04) | 0.76 (0.04) | 0.70 (0.06) | 0.75 (0.03)
oilspill | 0.96 (0.01) | 0.97 (0.01) | 0.96 (0.02) | 0.95 (0.02) | 0.96 (0.02)
phoneme | 0.91 (0.01) | 0.92 (0.01) | 0.91 (0.01) | 0.88 (0.01) | 0.90 (0.01)
seeds | 0.91 (0.09) | 0.92 (0.07) | 0.92 (0.07) | 0.91 (0.09) | 0.94 (0.06)
wine | 0.98 (0.03) | 0.92 (0.08) | 0.93 (0.05) | 0.89 (0.07) | 0.94 (0.04)
glass | 0.79 (0.07) | 0.77 (0.08) | 0.73 (0.10) | 0.67 (0.11) | 0.71 (0.09)
ecoli | 0.88 (0.06) | 0.84 (0.07) | 0.83 (0.06) | 0.82 (0.09) | 0.83 (0.05)
∗ Numbers in parantheses show the standard deviations | |
|
# Weak completions of paratopological groups
Taras Banakh T.Banakh: Ivan Franko National University of Lviv (Ukraine) and
Jan Kochanowski University in Kielce (Poland<EMAIL_ADDRESS>and
Mikhail Tkachenko M.Tkachenko: Department of Mathematics, Universidad
Autónoma Metropolitana, Av. San Rafael Atlixco 186, Col. Vicentina, C.P.
09340, Iztapalapa, Mexico City, Mexico<EMAIL_ADDRESS>
###### Abstract.
Given a $T_{0}$ paratopological group $G$ and a class $\mathcal{C}$ of
continuous homomorphisms of paratopological groups, we define the
_$\mathcal{C}$ -semicompletion_ $\mathcal{C}[G)$ and _$\mathcal{C}$
-completion_ $\mathcal{C}[G]$ of the group $G$ that contain $G$ as a dense
subgroup, satisfy the $T_{0}$-separation axiom and have certain universality
properties. For special classes $\mathcal{C}$, we present some necessary and
sufficient conditions on $G$ in order that the (semi)completions
$\mathcal{C}[G)$ and $\mathcal{C}[G]$ be Hausdorff. Also, we give an example
of a Hausdorff paratopological abelian group $G$ whose
$\mathcal{C}$-semicompletion $\mathcal{C}[G)$ fails to be a $T_{1}$-space,
where $\mathcal{C}$ is the class of continuous homomorphisms of sequentially
compact topological groups to paratopological groups. In particular, the group
$G$ contains an $\omega$-bounded sequentially compact subgroup $H$ such that
$H$ is a topological group but its closure in $G$ fails to be a subgroup.
###### Key words and phrases:
Paratopological group, pseudocompact, precompact, Raĭkov completion,
Bicompletion
###### 2010 Mathematics Subject Classification:
22A15, 54D35, 54H11
The second author was supported by grant number CAR-64356 of the Program
“Ciencia de Frontera 2019” of the CONACyT, Mexico
## 1\. Introduction
It is well known [1, §3.6] that each topological group $G$ has the Raĭkov
completion, $\varrho{G}$, which coincides with the completion of $G$ with
respect to the two-sided group uniformity of the group. The Raĭkov completion
has a nice categorial characterization. It turns out that $\varrho{G}$ is a
unique topological group, up to topological isomorphism, that contains $G$ as
a dense subgroup and has the following extension property: Every continuous
homomorphism $h\colon X\to G$ defined on a dense subgroup $X$ of a topological
group $\widetilde{X}$ admits a unique extension to a continuous homomorphism
$\tilde{h}\colon\widetilde{X}\to\varrho{G}$ (see [1, Proposition 3.6.12]). A
topological group $G$ is Raĭkov complete if and only if it is complete in its
two-sided uniformity if and only if $G=\varrho{G}$.
In [4] and [3], it is shown that a kind of the Raĭkov completion can also be
defined for paratopological groups. By a _paratopological group_ we understand
a group $G$ endowed with a topology $\tau$ making the group multiplication
$G\times G\to G$, $\langle x,y\rangle\to xy$ continuous. If, in addition, the
inversion $G\to G$, $x\mapsto x^{-1}$, is continuous, then $\langle
G,\tau\rangle$ is a topological group. A topology $\tau$ on a group $G$ is
called a topological group topology on $G$ if $\langle G,\tau\rangle$ is a
topological group.
Let us recall that a topology $\tau$ on a set $X$ satisfies the $T_{0}$
separation axiom or, equivalently, $\tau$ is a $T_{0}$-topology if for any
distinct points $x,y\in G$ there exists an open set $U\in\tau$ such that
$U\cap\\{x,y\\}$ is a singleton. It is well known, on the one hand, that each
topological group satisfying the $T_{0}$ separation axiom is automatically
Tychonoff. On the other hand, the topology
$\tau=\\{\emptyset,\mathbb{R}\\}\cup\\{(a,\infty):a\in\mathbb{R}\\}$ turns the
additive group of the reals into a $T_{0}$ paratopological group which does
not satisfy the $T_{1}$ separation axiom. So ‘$T_{0}$ does not imply $T_{1}$’
in paratopological groups. From now on we assume that all paratopological
groups considered here satisfy the $T_{0}$ separation axiom.
Each paratopological group $\langle G,\tau\rangle$ admits a stronger
topological group topology $\tau^{\sharp}$, whose neighborhood base at the
identity $e$ consists of the set $U\cap U^{-1}$ where $e\in U\in\tau$. So,
$\langle G,\tau^{\sharp}\rangle$ is topologically isomorphic to the subgroup
$\\{\langle x,x^{-1}\rangle:x\in G\\}$ of the product group $G\times G$. The
topological group $G^{\sharp}=\langle G,\tau^{\sharp}\rangle$ is called the
group coreflection of the paratopological group $G=\langle G,\tau\rangle$. If
the topology $\tau$ satisfies the $T_{0}$ separation axiom, then the
topological group topology $\tau^{\sharp}$ is Hausdorff and hence Tychonoff. A
paratopological group $\langle G,\tau\rangle$ is a topological group if and
only if $\tau=\tau^{\sharp}$. A paratopological group $G$ is called
$\sharp$-complete if the topological group $G^{\sharp}$ is Raĭkov complete.
The Sorgenfrey line $\mathbb{S}$ is an easy example of a $\sharp$-complete
paratopological group since $\mathbb{S}^{\sharp}$ is the discrete group of the
reals.
A subset $D$ of a paratopological group $\langle G,\tau\rangle$ is called
$\sharp$-dense in $G$ if $D$ is dense in $\langle G,\tau^{\sharp}\rangle$.
Since $\tau\subseteq\tau^{\sharp}$, each $\sharp$-dense subset is dense (but
not vice versa).
According to [4] or [3], each paratopological group $G$ is a $\sharp$-dense
subgroup of a $\sharp$-complete paratopological group $\breve{G}$ that has the
following extension property: any continuous homomorphism $h\colon X\to G$
defined on a $\sharp$-dense subgroup $X$ of a paratopological group
$\widetilde{X}$ admits a unique extension to a continuous homomorphism
$\tilde{h}\colon\widetilde{X}\to\breve{G}$. This extension property of
$\breve{G}$ implies that the $\sharp$-complete paratopological group
$\breve{G}$ containing $G$ as a $\sharp$-dense subgroup is unique up to a
topological isomorphism. In [3] this unique paratopological group $\breve{G}$
is called the Raĭkov completion of $G$ and in [4] it is called the
bicompletion of $G$. By [3], a neighborhood base at the identity $e$ of the
paratopological group $\breve{G}$ consists of the sets
$\overline{U}^{\sharp}$, where $U$ is a neighborhood of $e$ in $G$ and
$\overline{U}^{\sharp}$ is the closure of $U$ in the Raĭkov completion
$\varrho{G}^{\sharp}$ of the topological group $G^{\sharp}$. So as a group,
$\breve{G}$ coincides with the Raĭkov completion of $G^{\sharp}$. If $G$ is a
topological group, then the Raĭkov completion $\breve{G}$ of $G$ as a
paratopological group coincides with the usual Raĭkov completion $\varrho{G}$
of $G$.
By [3], _for any regular paratopological group $G$, its Raĭkov completion
$\breve{G}$ is regular and hence Tychonoff_ (for the latter implication, see
[2]). However, the Raĭkov completion of a Hausdorff paratopological group $G$
is not necessarily Hausdorff (see [3, Example 2.9]). To bypass this pathology
of $\breve{G}$, in this article we consider some weaker notions of completion
of paratopological groups which preserve Hausdorffness in some special cases.
Our weaker notion of completion depends on a class $\mathcal{C}$ of continuous
homomorphisms between paratopological groups. For example, $\mathcal{C}$ can
be the class of all continuous homomorphisms of topological groups (possessing
some property like precompactness, pseudocompactness, countable compactness,
sequential compactness, etc.) to paratopological groups. So, we fix a class
$\mathcal{C}$ of continuous homomorphisms between paratopological groups.
A paratopological group $\widetilde{G}$ is called a
$\mathcal{C}$-semicompletion of a paratopological group $G$ if
$G\subseteq\widetilde{G}$ and any homomorphism $h\colon X\to G$ in the class
$\mathcal{C}$ has a unique extension
$\breve{h}\colon\breve{X}\to\widetilde{G}$ to the Raĭkov completion
$\breve{X}$ of the paratopological group $X$. The extension property of the
Raĭkov completion $\breve{G}$ ensures that $\breve{G}$ is a
$\mathcal{C}$-semicompletion of $G$. So every paratopological group has at
least one $\mathcal{C}$-semicompletion.
A $\mathcal{C}$-semicompletion $\widetilde{G}$ of a paratopological group $G$
is called minimal if $\widetilde{G}=H$ for any subgroup
$H\subseteq\widetilde{G}$ which is a $\mathcal{C}$-semicompletion of $G$.
Every $\mathcal{C}$-semicompletion $\tilde{G}$ of a paratopological group $G$
contains a unique minimal $\mathcal{C}$-semicompletion of $G$, which is equal
to the intersection $\bigcap\mathcal{S}$ of the family $\mathcal{S}$ of
subgroups $H\subseteq\widetilde{G}$ such that $H$ is a
$\mathcal{C}$-semicompletion of $G$. The Hausdorffness of the group
coreflections implies that $\bigcap\mathcal{S}$ is indeed a
$\mathcal{C}$-semicompletion of $G$.
The $\mathcal{C}$-semicompletion $\mathcal{C}[G)$ of a paratopological group
$G$ is defined to be the smallest $\mathcal{C}$-semicompletion of $G$ in its
Raĭkov completion $\breve{G}$. It is equal to the intersection
$\bigcap\mathcal{S}$ of the family $\mathcal{S}$ of all groups $H$ with
$G\subseteq H\subseteq\breve{G}$ such that $H$ is a
$\mathcal{C}$-semicompletion of $G$. Let us note that the
$\mathcal{C}$-semicompletion $\mathcal{C}[G)$ of $G$ is topologically
isomorphic to any minimal $\mathcal{C}$-semicompletion of $G$. Indeed, given a
minimal $\mathcal{C}$-semicompletion $H$ of $G$, consider the Raĭkov
completion $\breve{H}$ of the paratopological group $H$. Since $G\subseteq
H\subseteq\breve{H}$, the Raĭkov completion $\breve{G}$ of $G$ can be
identified with the $\sharp$-closure of $G$ in $\breve{H}$. Now the minimality
of the $\mathcal{C}$-semicompletion $H$ implies that
$H=H\cap\breve{G}\subseteq\breve{G}$ and, finally, $H=\mathcal{C}[G)$.
Having defined the $\mathcal{C}$-semicompletions of paratopological groups, we
can pose the following two general problems.
###### Problem 1.1.
Explore the categorial properties of $\mathcal{C}$-semicompletions.
###### Problem 1.2.
Characterize the paratopological groups $G$ for which the
$\mathcal{C}$-semicompletion $\mathcal{C}[G)$ of $G$ is Hausdorff.
Some partial answers to these problems will be given in Sections 2 and 3.
Again, let $\mathcal{C}$ be a class of continuous homomorphisms of
paratopological groups. Now we introduce the notion of a
$\mathcal{C}$-completion of a paratopological group. A paratopological group
$G$ is $\mathcal{C}$-complete if $G$ is a $\mathcal{C}$-semicompletion of $G$.
Equivalently, a paratopological group $G$ is $\mathcal{C}$-complete if each
homomorphism $h\colon X\to G$ in the class $\mathcal{C}$ extends to a
continuous homomorphism $\breve{h}\colon\breve{X}\to G$ defined on the Raĭkov
completion of $X$. By the extension property of Raĭkov completions, each
$\sharp$-complete paratopological group $G$ (hence, $\breve{G}$) is
$\mathcal{C}$-complete. In particular, the Sorgenfrey line $\mathbb{S}$ is
$\mathcal{C}$-complete.
By the $\mathcal{C}$-completion, $\mathcal{C}[G]$, of a paratopological group
$G$ we understand the intersection $\bigcap\mathcal{S}$ of the family
$\mathcal{S}$ of all $\mathcal{C}$-complete subgroups $H\subseteq\breve{G}$
that contain $G$. The Hausdorffness of the groups coreflections of
paratopological groups implies that the $\mathcal{C}$-completion
$\mathcal{C}[G]$ of any paratopological group $G$ is $\mathcal{C}$-complete
and, hence, $\mathcal{C}[G)\subseteq\mathcal{C}[G]$.
It is clear from the above definitions that if $\mathcal{C}$ is the empty
class (or, more generally, every homomorphism $h\colon X\to G$ with
$h\in\mathcal{C}$, if exists, is trivial), then $G$ is $\mathcal{C}$-complete
and, therefore, $G=\mathcal{C}[G)=\mathcal{C}[G]$.
###### Problem 1.3.
Explore the categorial properties of the operation of $\mathcal{C}$-completion
in the category of paratopological groups and their continuous homomorphisms.
###### Problem 1.4.
Characterize the paratopological groups $G$ such that the
$\mathcal{C}$-completion $\mathcal{C}[G]$ of $G$ is Hausdorff.
###### Problem 1.5.
For which paratopological groups (and classes $\mathcal{C}$ of homomorphisms)
do their $\mathcal{C}$-semicompletions coincide with their
$\mathcal{C}$-completions?
## 2\. Categorial properties of $\mathcal{C}$-completions and
$\mathcal{C}$-semicompletions
A class $\mathcal{C}$ of continuous homomorphisms between paratopological
groups is called composable if for any homomorphism $f\colon X\to Y$ in the
class $\mathcal{C}$ and any continuous homomorphism $g\colon Y\to Z$ of
paratopological groups, the composition $g\circ f$ is in $\mathcal{C}$. For
example, for any class $\mathcal{P}$ of paratopological groups, the class
$\mathcal{C}$ of continuous homomorphisms $h\colon X\to Y$ between
paratopological groups with $X\in\mathcal{P}$ is composable.
The following proposition shows that for a composable class $\mathcal{C}$, the
constructions of $\mathcal{C}$-semicompletion and $\mathcal{C}$-completion are
functorial in the category of paratopological groups and their continuous
homomorphisms.
###### Proposition 2.1.
Let $\mathcal{C}$ be a composable class of continuous homomorphisms of
paratopological groups. For any continuous homomorphism $h\colon X\to Y$ of
paratopological groups, its continuous extension
$\breve{h}\colon\breve{X}\to\breve{Y}$ satisfies
$\breve{h}[\mathcal{C}[X)]\subseteq\mathcal{C}[Y)$ and
$\breve{h}[\mathcal{C}[X]]\subseteq\mathcal{C}[Y]$.
###### Proof.
To see that $\breve{h}[\mathcal{C}[X)]\subseteq\mathcal{C}[Y)$, it suffices to
check that the preimage $\breve{h}^{-1}[\mathcal{C}[Y)]$ is a
$\mathcal{C}$-semicompletion of $X$. Given any homomorphism $f\colon Z\to X$
in the class $\mathcal{C}$, consider its continuous homomorphic extension
$\breve{f}\colon\breve{Z}\to\breve{X}$. Then
$\breve{h}\circ\breve{f}\colon\breve{X}\to\breve{Y}$ is a continuous extension
of the homomorphism $h\circ f\colon Z\to Y$. Taking into account that
$\mathcal{C}[Y)$ is a $\mathcal{C}$-semicompletion of $Y$ and the topology of
the group reflection of $\breve{Y}$ is Hausdorff, we conclude that
$(\breve{h}\circ\breve{f})[\breve{Z}]\subseteq\mathcal{C}[Y)$ and hence
$\breve{f}[\breve{Z}]\subseteq\breve{h}^{-1}[\mathcal{C}[Y)]$. This implies
that $\breve{h}^{-1}[\mathcal{C}[Y)]$ is a $\mathcal{C}$-semicompletion of $X$
and $\mathcal{C}[X)\subseteq\breve{h}^{-1}[\mathcal{C}[Y)]$, by the minimality
of $\mathcal{C}[X)$.
Next, we show that the preimage $\breve{h}^{-1}[\mathcal{C}[Y]]$ is a
$\mathcal{C}$-complete paratopological group. Given any homomorphism $f\colon
Z\to\breve{h}^{-1}[\mathcal{C}[Y]]$ in the class $\mathcal{C}$, consider its
unique continuous extension $\breve{f}\colon\breve{Z}\to\breve{X}$. Then
$\breve{h}\circ\breve{f}\colon\breve{Z}\to\breve{Y}$ is a unique continuous
extension of the homomorphism $h\circ f\colon Z\to\mathcal{C}[Y]$. Taking into
account that the paratopological group $\mathcal{C}[Y]$ is
$\mathcal{C}$-complete and the topology of the group reflection of $\breve{Y}$
is Hausdorff, we conclude that
$(\breve{h}\circ\breve{f})[\breve{Z}]\subseteq\mathcal{C}[Y]$, which implies
the inclusion $\breve{f}[\breve{Z}]\subseteq\breve{h}^{-1}[\mathcal{C}[Y]]$.
Therefore, the paratopological group $\breve{h}^{-1}[\mathcal{C}[Y]]$ is
$\mathcal{C}$-complete and
$\mathcal{C}[X]\subseteq\breve{h}^{-1}[\mathcal{C}[Y]]$, by the minimality of
$\mathcal{C}[X]$. ∎
## 3\. The Hausdorff property of $\mathcal{C}$-completions of paratopological
groups
In this section, we present some necessary and some sufficient conditions on a
paratopological group $G$ in order that the $\mathcal{C}$-semicompletion
$\mathcal{C}[G)$ or $\mathcal{C}$-completion $\mathcal{C}[G]$ of $G$ be
Hausdorff. Let us recall that for a regular paratopological group $G$, its
Raĭkov completion $\breve{G}$ is regular and so are the paratopological groups
$\mathcal{C}[G)$ and $\mathcal{C}[G]$.
###### Proposition 3.1.
The $\mathcal{C}$-completion $\mathcal{C}[G]$ of a paratopological group $G$
is Hausdorff if and only if $G$ is a subgroup of a Hausdorff
$\mathcal{C}$-complete paratopological group.
###### Proof.
The “only if” part is trivial. To prove the “if” part, assume that $G$ is a
subgroup of a Hausdorff $\mathcal{C}$-complete paratopological group $H$.
Taking into account that $G\subseteq H$ and $G^{\sharp}\subseteq H^{\sharp}$,
we conclude that $\breve{G}\subseteq\breve{H}$. The $\mathcal{C}$-completeness
of the paratopological groups $\breve{G}$ and $H$ implies that the
paratopological group $\breve{G}\cap H$ is $\mathcal{C}$-complete and hence
$\mathcal{C}[G]\subseteq\breve{G}\cap H$ is Hausdorff, being a subgroup of the
Hausdorff paratopological group $H$. ∎
Let $\mathcal{P}$ be a property. A paratopological group $H$ is said to be
_projectively $\mathcal{P}$_ if every neighborhood of the identity element in
$H$ contains a closed invariant (equivalently, “normal” in the algebraic
sense) subgroup $N$ such that the quotient paratopological group $H/N$ has
$\mathcal{P}$. Similarly, a paratopological group $X$ is said to be
projectively $\Psi$ if for every neighborhood $U$ of the identity in $X$,
there exists a continuous homomorphism $h\colon X\to Y$ to a Hausdorff
paratopological group of countable pseudocharacter such that
$h^{-1}(e_{Y})\subseteq U$; here $e_{Y}$ denotes the identity of the group
$Y$.
It is easy to see that the projectively $\Psi$ paratopological groups are
exactly the projectively $\mathcal{P}$ groups, where $\mathcal{P}$ is the
property of being a Hausdorff space of countable pseudocharacter. Indeed, let
$G$ be a projectively $\Psi$ paratopological group and $U$ be a neighborhood
of the identity in $G$. By our assumption, there exists a continuous
homomorphism $h\colon G\to H$ onto a Hausdorff paratopological group
$H_{\alpha}$ of countable pseudocharacter such that $h^{-1}(e_{H})\subseteq
U$, where $e_{H}$ is the identity of $H$. Let $N=h^{-1}(e_{H})$ be the kernel
of $h$ and $p\colon G\to G/N$ be the quotient homomorphism. Clearly there
exists a continuous one-to-one homomorphism $j\colon G/N\to H$ satisfying
$h=j\circ p$. Thus $j$ is a continuous bijection of $G/N$ onto $H$ and, hence,
the quotient group $G/N$ is Hausdorff and has countable pseudocharacter.
Therefore, $G$ is projectively $\mathcal{P}$. The inverse implication is
evident.
According to [6, Proposition 2] every projectively Hausdorff paratopological
group is Hausdorff. In particular, all projectively $\Psi$ paratopological
groups are Hausdorff. This fact also follows from the characterization of
projectively $\Psi$ groups presented in the next lemma.
###### Lemma 3.2.
A paratopological group $G$ is projectively $\Psi$ if and only if $G$ is
topologically isomorphic to a subgroup of a product of Hausdorff
paratopological groups of countable pseudocharacter.
###### Proof.
The sufficiency is evident, so we verify only the necessity. Assume that $G$
is a projectively $\Psi$ paratopological group. Applying [6, Proposition 2] we
conclude that $G$ is Hausdorff. Let $\\{U_{\alpha}:\alpha\in A\\}$ be a
neighborhood base at the identity $e$ of $G$. For every $\alpha\in A$, take an
open neighborhood $V_{\alpha}$ of $e$ in $G$ such that $V_{\alpha}^{2}\subset
U_{\alpha}$. By our assumption, there exists a continuous homomorphism
$f_{\alpha}$ of $G$ onto a Hausdorff paratopological group $H_{\alpha}$ of
countable pseudocharacter such that $f_{\alpha}^{-1}(e_{\alpha})\subseteq
V_{\alpha}$, where $e_{\alpha}$ is the identity of $H_{\alpha}$. Let
$N_{\alpha}$ be the kernel of $f_{\alpha}$ and $p_{\alpha}\colon G\to
G/N_{\alpha}$ be the quotient homomorphism. Clearly there exists a continuous
one-to-one homomorphism $j_{\alpha}\colon G/N_{\alpha}\to H_{\alpha}$
satisfying $f_{\alpha}=j_{\alpha}\circ p_{\alpha}$. Thus $j_{\alpha}$ is a
continuous bijection of $G/N_{\alpha}$ onto $H_{\alpha}$ and, hence, the
quotient group $G/N_{\alpha}$ is Hausdorff and has countable pseudocharacter.
Denote by $p$ the diagonal product of the family $\\{p_{\alpha}:\alpha\in
A\\}$. Then $p$ is a continuous homomorphism of $G$ to $P=\prod_{\alpha\in
A}G/N_{\alpha}$. We claim that $p$ is an isomorphic topological embedding.
First, the kernel of $p$ is trivial. Indeed, if $x\in G$ and $x\neq e_{G}$,
take a neighborhood $U$ of $e_{G}$ such that $x\notin U$. There exists
$\alpha\in A$ such that $p_{\alpha}^{-1}(e_{\alpha})\subseteq
U_{\alpha}\subseteq U$, whence it follows that $p_{\alpha}(x)\neq e_{\alpha}$.
Hence $p(x)\neq e_{H}$ and $p$ is injective.
Further, let $U$ be an arbitrary neighborhood of $e$ in $G$. Take $\beta\in A$
with $U_{\beta}\subseteq U$. Then
$f_{\beta}^{-1}f_{\beta}(V_{\beta})=V_{\beta}N_{\beta}\subseteq
V_{\beta}^{2}\subseteq U_{\beta}\subseteq U$. Since $j_{\beta}$ is one-to-one,
it follows from $f_{\beta}=j_{\beta}\circ p_{\beta}$ that
$p_{\beta}^{-1}p_{\beta}(V_{\beta})=f_{\beta}^{-1}f_{\beta}(V_{\beta})\subseteq
U$. Thus the open neighborhood $p_{\beta}(V_{\beta})$ of the identity in
$H_{\beta}$ satisfies $p_{\beta}^{-1}p_{\beta}(V_{\beta})\subseteq U$. Denote
by $\pi_{\beta}$ the projection of $\prod_{\alpha\in A}G/N_{\alpha}$ to the
factor $G/N_{\beta}$. Applying the equality $p_{\beta}=\pi_{\beta}\circ p$ we
deduce that the open neighborhood
$W=p(G)\cap\pi_{\beta}^{-1}p_{\beta}(V_{\beta})$ of the identity in $p(G)$
satisfies $p^{-1}(W)=p_{\beta}^{-1}(p_{\beta}(V_{\beta}))\subseteq U$. We have
thus proved that for every neighborhood $U$ of the identity in $G$, there
exists a neighborhood $W$ of the identity in $p(G)$ such that
$p^{-1}(W)\subseteq U$. This property of the continuous monomorphism $p$
implies that $p$ is an isomorphic topological embedding of $G$ into
$\prod_{\alpha\in A}G/N_{\alpha}$ (see also [1, Section 3.4]). ∎
A paratopological group $G$ is called _$\omega$ -balanced_ if for any
neighborhood $U$ of the identity in $G$, there exists a countable family
$\mathcal{V}$ of neighborhoods of the identity in $G$ such that for each $x\in
G$, one can find $V\in\mathcal{V}$ satisfying $xVx^{-1}\subseteq U$.
The _Hausdorff number_ $Hs(G)$ of a Hausdorff paratopological group $G$ is the
smallest cardinal $\kappa\geq 1$ such that for every neighborhood $U\subseteq
G$ of the identity in $G$, there exists a family
$\\{V_{\alpha}\\}_{\alpha\in\kappa}$ of neighborhoods of the identity in $G$
such that $\bigcap_{\alpha\in\kappa}V_{\alpha}V_{\alpha}^{-1}\subseteq U$. The
Hausdorff number was introduced and studied in [7].
According to [7, Theorem 2.7], every $\omega$-balanced paratopological group
$G$ with $Hs(G)\leq\omega$ is a subgroup of a Tychonoff product of first-
countable Hausdorff paratopological groups. By Lemma 3.2, the latter implies
that the $\omega$-balanced paratopological groups with countable Hausdorff
number are projectively $\Psi$.
###### Theorem 3.3.
Let $\mathcal{C}$ be a subclass of the class of continuous homomorphisms from
pseudocompact topological groups to paratopological groups. If a
paratopological group $G$ is projectively $\Psi$, then its
$\mathcal{C}$-completion $\mathcal{C}[G]$ is Hausdorff. If $G$ is
$\omega$-balanced and satisfies $Hs(G)\leq\omega$, then the paratopological
group $\mathcal{C}[G]$ is $\omega$-balanced, Hausdorff, and satisfies
$Hs(\mathcal{C}[G])\leq\omega$.
###### Proof.
Let $G$ be a projectively $\Psi$ paratopological group. By Lemma 3.2, $G$ is
topologically isomorphic to a subgroup of a Tychonoff product
$Y=\prod_{\alpha\in A}Y_{\alpha}$ of Hausdorff paratopological groups of
countable pseudocharacter. We claim that the paratopological group $Y$ is
$\mathcal{C}$-complete. Consider any homomorphism $f\colon Z\to Y$ in the
class $\mathcal{C}$. By the choice of $\mathcal{C}$, $Z$ is a pseudocompact
topological group. For every $\alpha\in A$, denote by $\pi_{\alpha}\colon Y\to
Y_{\alpha}$ the natural projection and consider the continuous homomorphism
$f_{\alpha}=\pi_{\alpha}\circ f\colon Z\to Y_{\alpha}$. Let $e_{\alpha}$ be
the identity of the group $Y_{\alpha}$. Clearly
$Z_{\alpha}=f_{\alpha}^{-1}(e_{\alpha})$ is a closed invariant subgroup of $Z$
and the quotient topological group $Z/Z_{\alpha}$ admits a continuous
injective homomorphism to the group $Y_{\alpha}$ of countable pseudocharacter.
Hence $Z/Z_{\alpha}$ has countable pseudocharacter as well and we can apply
[8, Proposition 2.3.12] to conclude that $Z/Z_{\alpha}$ is a compact
metrizable group. Let $q_{\alpha}\colon Z\to Z/Z_{\alpha}$ be the quotient
homomorphism and $g_{\alpha}\colon Z/Z_{\alpha}\to Y_{\alpha}$ be a unique
continuous injective homomorphism such that $f_{\alpha}=g_{\alpha}\circ
q_{\alpha}$. Since the compact topological group $Z/Z_{\alpha}$ is Raĭkov
complete, the quotient homomorphism $q_{\alpha}\colon Z\to Z/Z_{\alpha}$
admits a continuous homomorphic extension
$\breve{q}_{\alpha}\colon\breve{Z}\to Z/Z_{\alpha}$. Then
$\breve{f}_{\alpha}=g_{\alpha}\circ\breve{q}_{\alpha}\colon\breve{Z}\to
Y_{\alpha}$ is a continuous extension of the homomorphism $f_{\alpha}$.
The diagonal product of the homomorphisms
$\breve{f}_{\alpha}\colon\breve{Z}\to Y_{\alpha}$ with $\alpha\in A$ is a
continuous homomorphism $\breve{f}\colon\breve{Z}\to Y=\prod_{\alpha\in
A}Y_{\alpha}$ extending the homomorphism $f$ and witnessing that the
paratopological group $Y$ is $\mathcal{C}$-complete. By Proposition 3.1, the
$\mathcal{C}$-completion $\mathcal{C}[X]$ of $X$ is Hausdorff.
Now assume that the paratopological group $G$ is $\omega$-balanced and
satisfies $Hs(G)\leq\omega$. By [7, Theorem 2.7], $G$ is a subgroup of a
Tychonoff product $Y=\prod_{\alpha\in A}Y_{\alpha}$ of first-countable
Hausdorff paratopological groups. Hence $G$ is projectively $\Psi$, by Lemma
3.2. By the above argument, the paratopological group $Y$ is
$\mathcal{C}$-complete and, hence, $\mathcal{C}[G]$ can be identified with a
subgroup of $Y$. By Propositions 2.1–2.3 in [7], the subgroup $\mathcal{C}[G]$
of the Tychonoff product $Y$ of first-countable Hausdorff paratopological
groups is $\omega$-balanced and satisfies $Hs(\mathcal{C}[G])\leq\omega$. ∎
A topological group $G$ is called precompact if its Raĭkov completion
$\rho{G}$ is compact. This happens if and only if for any neighborhood $U$ of
the identity in $G$, there exists a finite set $F\subseteq G$ such that
$G=UF=FU$.
###### Proposition 3.4.
Let $\mathcal{C}$ be a subclass of the class of continuous homomorphisms from
precompact topological groups to paratopological groups. A Hausdorff
paratopological group $G$ is $\mathcal{C}$-complete if and only if for any
homomorphism $h\colon X\to G$ in the class $\mathcal{C}$, the image $h[X]$ has
compact closure in $G$.
###### Proof.
_Sufficiency._ Take any homomorphism $h\colon X\to G$ in the class
$\mathcal{C}$ and assume that the image $h[X]$ has compact closure
$\overline{h[X]}$ in $G$. It follows from [1, Proposition 1.4.10] that
$H=\overline{h[X]}$ is a Hausdorff compact topological semigroup and $H$,
being a subsemigroup of a group, is cancellative. By Numakura’s theorem (see
[5] or [1, Theorem 2.5.2]), $H$ is a compact topological group. Since $H$ is
Raĭkov complete, the homomorphism $h\colon X\to H$ has a continuous
homomorphic extension $\breve{h}\colon\breve{X}\to H\subseteq G$, witnessing
that the paratopological group $G$ is $\mathcal{C}$-complete.
_Necessity._ Assume that the paratopological group $G$ is
$\mathcal{C}$-complete. Then every homomorphism $h\colon X\to G$ in the class
$\mathcal{C}$ has a continuous homomorphic extension
$\breve{h}\colon\breve{X}\to G$ to the Raĭkov completion $\breve{X}=\rho{X}$
of the topological group $X$. The precompactness of $X$ guarantees that the
topological group $\breve{X}$ is compact and so is its image
$\breve{h}[\breve{X}]$ in $G$. Since the space $G$ is Hausdorff, the compact
subspace $\breve{h}[\breve{X}]$ is closed in $G$ and the closure
$\overline{h[X]}$ of $h[X]$ in $G$, being a closed subset of the compact space
$\breve{h}[\breve{X}]$, is compact. ∎
A subset $F$ of a topological space $X$ is called _functionally closed_ (or
else a _zero-set_) if $F=f^{-1}(0)$ for some continuous function $f\colon
X\to\mathbb{R}$. A paratopological group $G$ is said to be _simply
$sm$-factorizable_ if for every functionally closed set $A$ in $G$, there
exists a continuous homomorphism $h\colon G\to H$ onto a separable metrizable
paratopological group $H$ such that $A=h^{-1}[B]$, for some closed set
$B\subseteq H$ (see [1, Definition 5.6]).
A subspace $X$ of a topological space $Y$ is called _$C$ -embedded_ in $Y$ if
each continuous real-valued function on $X$ has a continuous extension over
$Y$.
###### Proposition 3.5.
Every regular simply $sm$-factorizable paratopological group $G$ is
$C$-embedded in its $\mathcal{C}$-completion $\mathcal{C}[G]$ provided
$\mathcal{C}$ is a subclass of the class of continuous homomorphisms of
pseudocompact topological groups to paratopological groups.
###### Proof.
The space $G$ is Tychonoff, by [2, Corollary 5]. We apply [9, Theorem 4.3]
according to which the realcompactification $\upsilon{G}$ of the space $G$
admits the structure of paratopological group containing $G$ as a dense
paratopological subgroup. Since the closure of every pseudocompact subspace in
$\upsilon{G}$ is compact and every pseudocompact topological group is
precompact, the paratopological group $\upsilon{G}$ is $\mathcal{C}$-complete
according to Proposition 3.4. Then
$G\subseteq\mathcal{C}[G]\subseteq\upsilon{G}$, which implies that $G$ is
$C$-embedded in $\mathcal{C}[G]$, being $C$-embedded in $\upsilon{G}$. ∎
Now we present a necessary condition on a paratopological group $G$ for the
Hausdorffness of its $\mathcal{C}$-semicompletion $\mathcal{C}[G)$. In view of
the inclusion $\mathcal{C}[G)\subseteq\mathcal{C}[G]$, the same condition is
necessary for the Hausdorffness of $\mathcal{C}[G]$.
###### Proposition 3.6.
Let $\mathcal{C}$ be a subclass of the class of continuous homomorphisms from
precompact topological groups to paratopological groups. If the
$\mathcal{C}$-semicompletion $\mathcal{C}[G)$ of $G$ is Hausdorff, then for
any homomorphism $h\colon X\to G$ in $\mathcal{C}$, the closure of $h[X]$ in
$G$ is a precompact topological group.
###### Proof.
Let $h\colon X\to G$ be any homomorphism in the class $\mathcal{C}$. By the
definition of a $\mathcal{C}$-semicompletion, the homomorphism $h$ has a
continuous extension $\breve{h}\colon\breve{X}\to\mathcal{C}[G)$. The
precompactness of $X$ ensures that $\breve{X}=\rho{X}$ is a compact
topological group. Assume that the paratopological group $\mathcal{C}[G)$ is
Hausdorff. Then the compact subspace $\breve{h}[\breve{X}]$ of
$\mathcal{C}[G)$ is closed. It follows that $\breve{h}[\breve{X}]$ is a
compact topological group that is topologically isomorphic to a quotient group
of the compact topological group $\breve{X}$. Then the closure of $h[X]$ in
$G$ is a precompact topological group, which is equal to the intersection
$G\cap\breve{h}[\breve{X}]$ of the group $G$ with the compact topological
group $\breve{h}[\breve{X}]\subseteq\mathcal{C}[G)$. ∎
Finally we present an example of a Hausdorff paratopological group $G$ with a
non-Hausdorff $\mathcal{C}$-semicompletion, where $\mathcal{C}$ is the class
of all continuous homomorphisms from sequentially compact topological groups
to paratopological groups. As usual, a space $X$ is _sequentially compact_ if
each sequence in $X$ contains a convergent subsequence. Also, $X$ is said to
be _$\omega$ -bounded_ if the closure in $X$ of every countable set is
compact.
###### Example 3.7.
Let $\mathcal{C}$ be the class of continuous homomorphisms from sequentially
compact topological groups to paratopological groups. There exists a Hausdorff
paratopological abelian group $G$ whose $\mathcal{C}$-semicompletion
$\mathcal{C}[G)$ fails to be a $T_{1}$-space. In addition, $G$ contains a
subgroup $H$ such that $H$ is a sequentially compact $\omega$-bounded
topological group but its closure in $G$ is not a group.
###### Proof.
Let $\mathbb{T}=\\{z\in\mathbb{C}:|z|=1\\}$ be the torus group with its usual
topology and multiplication inherited from the complex plane. So the identity
of $\mathbb{T}$ is $1$. Denote by $\Sigma$ the subgroup of the Tychonoff
product $\mathbb{T}^{\omega_{1}}$ defined as follows:
$\Sigma=\\{x\in\mathbb{T}^{\omega_{1}}:|\\{\alpha\in\omega_{1}:x_{\alpha}\neq
1\\}|\leq\omega\\}.$
According to Corollaries 1.6.33 and 1.6.34 of [1], the space $\Sigma$ is
Fréchet-Urysohn and $\omega$-bounded. Hence $\Sigma$ is sequentially compact.
Take an element $c\in\mathbb{T}^{\omega_{1}}$ of infinite order such that
$\langle c\rangle\cap\Sigma=\\{e\\}$, where $e$ is the identity element of
$\mathbb{T}^{\omega_{1}}$ and $\langle c\rangle$ is the cyclic group generated
by $c$. For every open neighborhood $U$ of $e$ in $\mathbb{T}^{\omega_{1}}$,
we define a subset $O_{U}$ of $G=\Sigma\times\mathbb{Z}$ by letting
$O_{U}=\bigcup_{n\in\omega}\big{(}c^{n}U\cap\Sigma\big{)}\times\\{n\\}.$
Here $\omega$ is identified with the subset $\\{0,1,2,\ldots\\}$ of
$\mathbb{Z}$. A routine verification shows that the sets $O_{U}$ with $U$ as
above, constitute a base at the identity $(e,0)\in G$ for a Hausdorff
paratopological group topology on $G$.
It turns out that the subgroup $H=\Sigma\times\\{0\\}$ of $G$ has the required
properties. Let us show that the closure of $H$ of $G$ is the set
$\Sigma\times(-\omega)$, where $-\omega=\\{0,-1,-2,\ldots\\}$, so this closure
is not a subgroup of $G$. First, we claim that any element $g=(x,k)\in G$ with
$k>0$ is not in the closure of $H$. Indeed, let $U$ be an arbitrary open
neighborhood of $e$ in $\mathbb{T}^{\omega_{1}}$. Then the set $gO_{U}$ is an
open neighborhood of $g$ in $G$ disjoint from $H$.
Further, consider an element $g=(x,-k)\in G$, where $x\in\Sigma$ and
$k\in\omega$. If $k=0$, then $g\in\Sigma\times\\{0\\}=H$. Assume that $k>0$
and take a basic open neighborhood $O_{U}$ of the identity in $G$, where $U$
is an open neighborhood of $e$ in $\mathbb{T}^{\omega_{1}}$. Then
$g\cdot
O_{U}=g\cdot\bigcup_{n\in\omega}\big{(}(c^{n}U\cap\Sigma)\times\\{n\\}\big{)}=\bigcup_{n\in\omega}(xc^{n}U\cap\Sigma)\times\\{n-k\\}.$
Therefore, $g\cdot O_{U}\cap H=g\cdot
O_{U}\cap(\Sigma\times\\{0\\})=(xc^{k}U\cap\Sigma)\times\\{k\\}\neq\emptyset$.
Hence $g$ is in the closure of $H$. We have thus shown that the closure of $H$
in $G$ is the asymmetric subset $\Sigma\times(-\omega)$ of $G$, which implies
the second claim of the example.
The definition of the topology of the paratopological group $G$ implies that
the topology of the group coreflection $G^{\sharp}$ coincides with the product
topology on $\Sigma\times\mathbb{Z}$ and then the Raĭkov completion
$\rho{G}^{\sharp}$ can be identified with the product
$\mathbb{T}^{\omega_{1}}\times\mathbb{Z}$, where $\mathbb{Z}$ is endowed with
the discrete topology. By [3, §2.2], a neighborhood base for the topology of
the Raĭkov completion $\breve{G}$ at its identity consists of the closures
$\overline{O_{U}}$ of the basic neighborhoods $O_{U}$ in the product space
$\mathbb{T}^{\omega_{1}}\times\mathbb{Z}$. It is easy to see that for any
neighborhood $U$ of $e$ in $\mathbb{T}^{\omega_{1}}$, the closure
$\overline{O_{U}}$ of
$O_{U}=\bigcup_{n\in\omega}(c^{n}U\cap\Sigma)\times\\{n\\}$ coincides with the
set $\bigcup_{n\in\omega}(c^{n}\overline{U})\times\\{n\\}$. We see, therefore,
that each neighborhood of the identity in $\breve{G}$ contains the point
$(c,1)$, which means that the Raĭkov completion $\breve{G}$ of $G$ does not
satisfy the $T_{1}$ separation axiom.
It remains to prove that $\breve{G}$ is the $\mathcal{C}$-semicompletion of
$G$. Since the subgroup $\Sigma$ of $\mathbb{T}^{\omega_{1}}$ is sequentially
compact, the continuous homomorphism
$h\colon\Sigma\to\Sigma\times\\{0\\}\subseteq G,\quad h:x\mapsto\langle
x,0\rangle,$
extends to a continuous homomorphism
$\bar{h}\colon\mathbb{T}^{\omega_{1}}\to\mathcal{C}[G)\subseteq\breve{G}$.
Since $\mathbb{T}^{\omega_{1}}$ is a topological group, the homomorphism
$\bar{h}$ remains continuous with respect to the topology of the Raĭkov
completion $\rho{G}^{\sharp}=\mathbb{T}^{\omega_{1}}\times\mathbb{Z}$ of the
topological group $G^{\sharp}$. Now the compactness of
$\mathbb{T}^{\omega_{1}}$ implies that
$\mathbb{T}^{\omega_{1}}\times\\{0\\}=\bar{h}[\mathbb{T}^{\omega_{1}}]\subseteq\mathcal{C}[G)$
and, hence, $\mathcal{C}[G)=\breve{G}$. ∎
## References
* [1] Alexander Arhangel’skii, Mikhail Tkachenko, Topological groups and related structures, Atlantis Press, Paris; World Sci. Publ., NJ, 2008.
* [2] Taras Banakh, Alex Ravsky, Each regular paratopological group is completely regular, Proc. Amer. Math. Soc. 145:3 (2017), 1373–1382.
* [3] Taras Banakh, Alex Ravsky, On feebly compact paratopological groups, Topol. Appl. 284 (2020) 107363.
* [4] Josefa Marín, Salvador Romaguera, A bitopological view of quasi-topological groups, Indian J. Pure Appl. Math. 27:4 (1996), 393–405.
* [5] Katsumi Numakura, On bicompact semigroups, Math. J. Okayama Univ. 1 (1952), 99–108.
* [6] Manuel Sanchis, Mikhail Tkachenko, Completions of paratopological groups and bounded sets, Monatsh. Math. 183 (2017), 699–721.
* [7] Mikhail Tkachenko, Embedding paratopological groups into topological products, Topol. Appl. 156 (2009), 1298–1305.
* [8] Mikhail Tkachenko, Pseudocompact Topological groups, Section 2 in: M. Hrušák, Á. Tamariz-Mascarúa, M. Tkachenko (Eds.), Pseudocompact Topological Spaces, Developments in Mathematics, vol. 55. Springer, Cham, 2018.
* [9] Li-Hong Xie, Mikhail Tkachenko, Simply sm-factorizable (para)topological groups and their completions, Monatsh. Math. 193 (2020), 507–529.
|
# Deriving the intrinsic properties of M51 with radiative transfer models
Christopher J. Inman,1 Cristina C. Popescu,1,2 Mark Rushton 3 and David
Murphy1
1University of Central Lancashire, Jeremiah Horrocks Institute, Preston, PR1
2HE, UK
2 Max Planck Institut für Kernphysik, Saupfercheckweg 1, 69117 Heidelberg,
Germany
3The Astronomical Institute of the Romanian Academy, Str. Cutitul de Argint 5,
Bucharest, Romania E-mail<EMAIL_ADDRESS><EMAIL_ADDRESS>
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
A quantitative derivation of the intrinsic properties of galaxies related to
their fundamental building blocks, gas, dust and stars is essential for our
understanding of galaxy evolution. A fully self-consistent derivation of these
properties can be achieved with radiative transfer (RT) methods that are
constrained by panchromatic imaging observations. Here we present an axi-
symmetric RT model of the UV-optical-FIR/submm spectral and spatial energy
distribution of the face-on spiral galaxy M51. The model reproduces reasonably
well the azimuthally averaged radial profiles derived from the imaging data
available for this galaxy, from GALEX, SDSS, 2MASS, Spitzer and Herschel. We
model the galaxy with three distinct morphological components: a bulge, an
inner disc and a main disc. We derive the length parameters of the stellar
emissivity and of the dust distribution. We also derive the intrinsic global
and spatially resolved parameters of M51. We find a faint “outer disc”
bridging M51 with its companion galaxy M51b. Finally, we present and discuss
an alternative model, with dust properties that change within the galaxy.
###### keywords:
radiative transfer – galaxies: disc – galaxies: stellar content – galaxies:
structure – ISM: dust, extinction – galaxies: spiral
††pubyear: 2023††pagerange: Deriving the intrinsic properties of M51 with
radiative transfer models–C
## 1 Introduction
Dust is prevalent within the interstellar medium (ISM) of star-forming
galaxies (Trumpler 1930), and, although it constitutes only $1\%$ of the mass
of this ISM (Greenberg 1963), it plays an important role in the thermodynamic
balance of galaxies (Montier & Giard 2004; Giard et al. 2008; Popescu & Tuffs
2010; Natale et al. 2010), since it regulates the cooling and heating
mechanisms of the interstellar gas (Dwek & Werner 1981).
The interstellar dust also affects our view of galaxies. This is because
stellar photons, primarily in the ultraviolet (UV)/optical range, are
continuously absorbed and scattered by the interstellar dust. Photons that are
scattered can travel long distances within the ISM before they are scattered
again, absorbed or leave the galaxy. Due to scattering, what would be
otherwise an isotropic process related to the emission of stellar light in
galaxies, becomes highly anisotropic. As such, along any line of sight through
a galaxy, the stellar light may be diminished due to extinction, but also
amplified by stellar photons being scattered into the line of sight. Thus, for
a extended distribution of emitters and absorbers, the complex effect of
absorption and scattering processes is what we call attenuation.
The attenuation is dependent not only on the optical properties of the dust
grains (as is the case for the extinction processes), but also on the relative
distribution of stars and dust, and on the orientation of a galaxy with
respect to the observer (Byun et al. 1994; Bianchi et al. 1996; Kuchinski et
al. 1998; Ferrara et al. 1999; Baes & Dejonghe 2001; Tuffs et al. 2004;
Pierini et al. 2004; Fischera & Dopita 2004; Natale et al. 2015). Dust
attenuation affects the observed spatially integrated stellar luminosity of a
galaxy, but also its surface brightness distribution and the geometric
parameters derived from surface-brightness photometry, like scalelength of
discs, effective radii of bulges and Sérsic indexes: Byun et al. (1994); Evans
(1994); Cunow (2001); Möllenhoff et al. (2006); Gadotti et al. (2010); Pastrav
et al. (2013a, b); Savchenko et al. (2023).
Due to dust attenuation the observed UV/optical images of star forming
galaxies are far from providing a direct tracer of the constituent stellar
populations distributed throughout the galaxy. In order to derive the
intrinsic distribution of stars of all ages, and thus to understand how stars
form and evolve in galaxies, one also needs to jointly derive the distribution
of dust in galaxies (Popescu & Tuffs 2010). This can be done by considering
both the absorption and emission properties of interstellar dust. Thus, the
stellar light that is absorbed by dust is re-emitted in the infrared (IR) and
the energy balance between direct and re-radiated stellar light is an
important constraint on the modelling of transfer of radiation in galaxies. In
particular dust emission contains unique information on both the spatial and
spectral energy distribution (SED) of the energy density of the radiation
fields that heat the dust grains, and on the dust properties and dust opacity.
Accounting for both dust attenuation and dust emission is thus needed to
provide a self-consistent model of a galaxy.
Galaxy models that are based on radiative transfer (RT) methods (Kylafis &
Xilouris 2005; Steinacker et al. 2013) and constrained by panchromatic imaging
observations are the most powerful tools in quantifying the attenuation of
stellar light in galaxies and in deriving the intrinsic parameters of galaxies
(Popescu 2021). This type of modelling was initially performed on edge-on
galaxies, since in this orientation it is possible to directly constrain the
vertical distribution of stars and dust from imaging observations. The first
RT modelling of an edge-on galaxy was performed in the optical range, for NGC
891, by Kylafis & Bahcall (1987). This was the first time a RT model was
applied in conjunction with realistic distributions of stars and dust (finite
exponential discs as opposed to slabs or sandwich geometries - Disney et al.
1989 or spherical configurations - Witt et al. 1992; Witt & Gordon 1996,
2000). Fundamental work on characterising the relative distributions of stars
and dust and their statistical trends were later established in a series of
papers dedicated to the modelling of edge-on galaxies in the optical and NIR
by Xilouris et al. (1997, 1998, 1999).
The first panchromatic (UV/optical/FIR/submm) RT modelling of an edge-on
galaxy was achieved for the same edge-on spiral NGC 891 (see above for the
first RT modelling) by Popescu et al. (2000). Further work on modelling edge-
on spirals was performed by Misiriotis et al. (2001); Popescu et al. (2004);
Bianchi (2008); Baes et al. (2010); Popescu et al. (2011); MacLachlan et al.
(2011); Robitaille et al. (2012); De Geyter et al. (2015); Mosenkov et al.
(2016, 2018); Popescu et al. (2017); Natale et al. (2021).
More recently, non edge-on galaxies have also started to be modelled with RT
methods. There are two approaches to this type of modelling. The first
approach is to use non axi-symmetric RT models: De Looze et al. (2014); Viaene
et al. (2017); Williams et al. (2019); Verstocken et al. (2020); Nersesian et
al. (2020a, b). Non axi-symmetric models are, in principle, ideal to capture
the detailed geometry seen face-on. However, due to computational limitations,
they are implemented to only fit the integrated SED, with typically only four
free parameters that describe the total luminosity output for stars and the
dust mass, with the geometry being fixed. The second approach is to use axi-
symmetric RT models (Thirlwall et al. 2020; Rushton et al. 2022) and fit not
only the integrated SED, but also the geometry of stars and dust, albeit with
the assumption of axi-symmetry. So far only non interacting face-on spiral
galaxies have been modelled with axi-symmetric models (M33; Thirlwall et al.
2020, NGC 628; Rushton et al. 2022). The goal of this paper is to extend the
applicability of the model to a non-isolated spiral. We will apply the model
to M51, since, despite its interaction with a close companion, most of the
extent of the galaxy has regular spiral structure, and is not perturbed by the
interaction, thus making it suitable for the axi-symmetry approximation.
First discovered by Charles Messier in 1773, M51 is a system of interacting
galaxies consisting of M51a (=NGC 5194), the ‘Whirlpool Galaxy’ , a grand-
design spiral of Hubble type Sbc, and the smaller lenticular galaxy M51b (=NGC
5195), of Hubble type SB0. While interaction between the two galaxies is
evident from the “bridge” linking them, there is still debate of whether the
galaxies have interacted on a single fly-by collision (Toomre & Toomre 1972;
Durrell et al. 2003) or by multiple close passes (Salo & Laurikainen 2000;
Dobbs et al. 2010). Despite the interaction, as mentioned before, M51a
exhibits a good deal of azimuthal symmetry. This, in conjunction with the
close proximity ($\text{D}=8.58\,\text{Mpc}$; McQuinn et al. 2016) and the
wealth of detailed observational data, makes M51a ideal for studying the
energy balance between absorption by dust and re-emission in the far-infrared
for a non-isolated spiral galaxy that is seen not far from face-on. In the
following we will refer to M51a as M51.
In this work we will utilise the axisymmetric RT models developed in Popescu
et al. (2000), Tuffs et al. (2004) and Popescu et al. (2011), which have been
successfully used to account for the properties of star forming galaxies
derived from statistical samples (e.g. van der Giessen et al. 2022), as well
as for the detailed spatial and spectral energy distribution of individual
nearby galaxies (Popescu et al. 2000; Thirlwall et al. 2020; Rushton et al.
2022) and of the Milky Way (Popescu et al. 2017; Niederwanger et al. 2019;
Natale et al. 2021). These models describe the geometries of the stellar
emissivity and dust opacity from the UV to submm using parameterised analytic
functions. The models consider the absorption and anisotropic scattering of
stellar photons with dust grains of various sizes and chemical compositions,
with the optical constants taken from Draine & Lee (1984) and Draine & Li
(2007). The models also consider the heating of the dust grains and the re-
emission in the infrared, including explicit calculations for stochastic
heating.
Previous work on modelling M51 was done by De Looze et al. (2014) and
Nersesian et al. (2020b) using non-axisymmetric RT models. As discussed above,
these models have the inherent limitation that they cannot self-consistently
derive the geometry of emitters and absorbers. In this study we aim to model
M51, by self-consistently deriving the geometry of stars and dust through
fitting the detailed azimuthally averaged UV/optical/FIR/submm surface
brightness profiles of M51. As we will show here, our axi-symmetric model
produces a good fit to the imaging data within a radial distance of 7 kpc from
the centre of M51, within which the galaxy preserves azimuthal symmetry.
The paper is organised as follows: In Sect. 2 we present the imaging
observations used for constraining the model, and the data reduction steps
undertaken in this study. In Sect. 3 we briefly describe the model and its
geometrical components, with the fitting procedure detailed in Sect. 4. In
Sect. 5 we present the results of the fits for the surface-brightness
distributions and for the intrinsic parameters of M51, including the SFR, dust
optical depth, dust mass, the fractional contribution of the different stellar
populations in heating the dust and the attenuation curve. We also present the
results for the derived morphological components of M51. In Sect. 6 we compare
our model for M51 with previous RT modelling of this galaxy and discuss an
alternative to the standard model presented in this paper. We summarise our
main results in Sect. 7.
## 2 Observational Data
In this section we discuss the data set of M51 used in this study. To start
with, basic parameters like inclination and position angle were difficult to
determine with standard methods, and, because of this, there has been a lot of
uncertainty in the value of these parameters. Determinations of the
inclination angle vary from $15^{\circ}$ to $42^{\circ}$ (Danver 1942; Toomre
& Toomre 1972; Tully 1974; Bersier et al. 1994; Monnet et al. 1981; Considere
& Athanassoula 1982; Shetty et al. 2007; Hu et al. 2013), while determinations
for the PA vary from $-10^{\circ}$ to $36^{\circ}$ (Burbidge & Burbidge 1964;
Tully 1974; Monnet et al. 1981; Considere & Athanassoula 1982; Shetty et al.
2007; Tamburro et al. 2008; Walter et al. 2008; Hu et al. 2013). We adopted
the most recent values from Hu et al. (2013). The distance, inclination,
position angle and central coordinates used in this paper are listed in Table
1. We obtained panchromatic images from various missions, as summarised in
Table 2.
Table 1: Distance, inclination, position angle and coordinates adopted for the modelling of M51 (M51/NGC 5194), together with the relevant reference. Distance | $8.58\pm 0.1$ Mpc | McQuinn et al. (2016)
---|---|---
Inclination | $20.3^{\circ}\pm 2.8^{\circ}$ | Hu et al. (2013)
Position Angle | $12.0^{\circ}\pm 2.5^{\circ}$ | Hu et al. (2013)
Right Ascension | $13^{\text{h}}\,29^{\text{m}}\,57.11^{\text{s}}$ | Turner & Ho (1994)
Declination | $47^{\circ}\,11^{\prime}\,\,\,42.62^{\prime\prime}$ | Turner & Ho (1994)
All data was found on NASA/IPAC Extragalactic Database
(NED)111https://ned.ipac.caltech.edu, except for the GALEX data, which was
obtained from the Mikulski Archive for Space Telescopes
(MAST)222https://archive.stsci.edu/prepds/gcat/. Before analysis, the units of
the surface brightness maps were converted into [MJy sr-1].
Table 2: A summary of the observational data used in this study. The following references were used: (1) Gil de Paz et al. (2007), (2) Brown et al. (2014), (3) Bendo et al. (2012). Telescope | Filter/ | $\lambda_{0}$ [$\mu$m] | Pixel scale | $\dfrac{\epsilon_{\text{cal}}}{F_{\nu}}$[%] | $\sigma_{\text{bg}}$ $\left[\rm{kJy\,sr^{-1}}\right]$ | Band name
---|---|---|---|---|---|---
| Instrument | | [′′] | | |
GALEX1 | FUV | 0.1542 | 1.5 | 10 | 0.0019 | FUV
GALEX1 | NUV | 0.2274 | 1.5 | 10 | 0.0024 | NUV
SDSS2 | u | 0.3562 | 0.396 | 2 | 0.049 | u
SDSS2 | g | 0.4719 | 0.396 | 2 | 0.017 | g
SDSS2 | r | 0.6185 | 0.396 | 2 | 0.028 | r
SDSS2 | i | 0.7500 | 0.396 | 2 | 0.049 | i
2MASS2 | J | 1.2000 | 1.0 | 3 | 0 | J
2MASS2 | K${}_{\text{s}}$ | 2.2000 | 1.0 | 3 | 0 | K${}_{\text{s}}$
Spitzer2 | IRAC | 3.5070 | 0.75 | 3 | 0 | I1
Spitzer2 | IRAC | 4.4370 | 0.75 | 3 | 0 | I2
Spitzer2 | IRAC | 5.7390 | 0.75 | 3 | 0.27 | I3
Spitzer2 | IRAC | 7.9270 | 0.75 | 3 | 0 | I4
Spitzer2 | MIPS | 24.000 | 1.5 | 1 | 0 | MIPS24
Herschel3 | PACS | 70.000 | 1.4 | 10 | 20.0 | PACS70
Herschel3 | PACS | 160.00 | 2.85 | 20 | 0 | PACS160
Herschel3 | SPIRE | 250.00 | 6.0 | 15 | 0 | SPIRE250
Herschel3 | SPIRE | 350.00 | 8.0 | 15 | 0 | SPIRE350
Herschel3 | SPIRE | 500.00 | 12.0 | 15 | 0 | SPIRE500
### 2.1 Background removal
We performed a curve of growth analysis to determine the radius from the
centre of M51 at which the emission of the galaxy falls below the background
level. This was found to be 25 kpc. We also visually inspect each map to
ensure that the emission associated with M51 does not extend beyond this
point. The outer sampled radius (30 kpc) was chosen to ensure a large area for
measuring the background. This sampling region extended beyond the limit of
some maps, thus the full sampling area could not be used in those cases.
However, even in the most severe cases, this was a negligible portion of the
total sampling area. We then calculate the average background flux by taking
the mean surface brightness through 6 annuli of equal width contained in the
region $25\,\text{kpc}\leq R\leq 30\,\text{kpc}$. Although M51 is nearly face-
on with an inclination of 20.3∘, the interaction with M51b gives rise to some
stripped material in the outer disc, making the outer isophotes more elongated
than the inner ones. This change in isophotal shape was taken into account
when defining the background. Thus, it was found that the best fitting annuli
for the background had an inclination of 40∘ (and the same position angle as
the galaxy of 12∘).
### 2.2 Foreground star removal
Foreground stars are visible in the UV/optical/NIR wavelengths and can
influence the measured values of emission from M51. It is therefore important
to mask these stars. For this we first produce a median map where each pixel
of the median map is equal to the median of the $3\times 3$ pixels around it.
A bias map is also produced, that is principally the same as the median map,
but takes the median of the $21\times 21$ pixels. This bias map is used as an
additional check to mitigate against false positives when masking noisy data.
Pixels with values that exceed both the median and bias maps are then masked.
After this initial pass, we then manually inspect each mask for regions that
may have been missed, or unmask false positives, such as HII regions. HII
regions appear as point sources (similarly to foreground stars) and thus are
often falsely masked by this automated process. These HII regions were
identified by eye and the catalogue of Lee et al. (2011). We also masked the
companion galaxy M51b and it’s diffuse emission.
### 2.3 Convolutions
The SDSS, 2MASS and Spitzer data have a higher resolution than that of our
model, which would correspond to 50 pc at the distance of M51. We therefore
degrade the resolution of these data to match the resolution of our model
using a two dimensional Gaussian kernel. The data at all the other wavelengths
used in this study have lower resolutions, therefore, for these wavelengths,
we degrade the resolution of our model to match the corresponding ones from
the data. For this we used the kernels of Aniano et al. (2011) to convolve the
model with the PSF of the observational instrument.
### 2.4 GALEX
The Galaxy Evolution Explorer (GALEX; Martin et al. 2005) was a NASA Small
Explorer mission that performed an all-sky survey in the far-ultraviolet (FUV
$\sim 0.15\,\mu$m) and near-ultraviolet (NUV $\sim 0.22\,\mu$m). Both FUV and
NUV data are used in this study and were found on the Mikulski Archive for
Space Telescopes (MAST) 333https://galex.stsci.edu/GR6/ and made available by
Gil de Paz et al. (2007). The NUV and FUV data has a pixel scale of 1.5′′ (62
pc) per pixel and a FWHM of 4.2′′ and 5.3′′, respectively. The flux
calibration error is $\frac{\epsilon_{\text{cal}}}{F_{\nu}}$ = 10% following
Morrissey et al. (2007).
### 2.5 SDSS
The Sloan Digital Sky Survey (SDSS; York et al. 2000) has observed M51 with
two sets of CCD arrays. Each array has an integration time of 51 s (York et
al. 2000) capturing images through u, g, r, i and z filters. In this study we
use the u, g, r, and i SDSS bands, with images obtained from NED/IPAC and
processed by Brown et al. (2014). We do not use the z image as this wavelength
is not included in our model fitting. Each band has a pixel scale of 0.396′′
(16 pc) and FWHM of 1′′ (Gunn et al. 2006). Each band has a 2% rms photometric
calibration in all bands (Aihara et al. 2011). The u band required a
correction of -0.04 mag (Brown et al. 2014).
### 2.6 2MASS
The Two Micron All Sky Survey (2MASS; Jarrett et al. 2003) was performed in
the NIR bands, J, H and K${}_{\text{s}}$, and observed M51 as part of the
Large Galaxy Atlas Survey. We use the 2MASS data in the J and K${}_{\text{s}}$
bands provided on NED/IPAC, and processed by Brown et al. (2014). We do not
use the H band data as this wavelength is not included in our model. Both J
and K${}_{\text{s}}$ bands have a pixel scale of 1′′ (41 pc) and a FWHM of
2.5′′ (Skrutskie et al. 2006). The calibration errors for J and
K${}_{\text{s}}$ are 2%-3% (Jarrett et al. 2003).
### 2.7 Spitzer
#### 2.7.1 IRAC
The Spitzer Space Telescope contains the Infrared Array Camera (IRAC). IRAC is
a four channel camera that images the 3.6 $\mu$m, 4.5 $\mu$m, 5.8 $\mu$m and
8.0 $\mu$m bands simultaneously. In this study we use all four bands. The
images were found on NED/IPAC and were processed by Brown et al. (2014). Each
instrument has a pixel size of 0.75′′ (31 pc). The 3.6 $\mu$m, 4.5 $\mu$m,
5.8$\,\mu$m, and 8.0 $\mu$m data has a FWHM of 1.66′′, 1.72′′, 1.88′′, and
1.98′′, respectively (Fazio et al. 2004). The flux calibration error for the
IRAC data is 3% (Fazio et al. 2004).
#### 2.7.2 MIPS
The Multiband Imaging Photometer for Spitzer (MIPS; Rieke et al. 2004)
conducted imaging in three wavebands: 24 $\mu$m, 70 $\mu$m and 160 $\mu$m. In
this study we only use the 24 $\mu$m data, since at 70 $\mu$m and 160 $\mu$m
we use the higher quality data from Herschel. The data was processed by Brown
et al. (2014) and made available at NED/IPAC. MIPS24 has a pixel scale of
1.5′′ (62 pc) and a FWHM of 6′′ (Rieke et al. 2004). The calibration error is
1% (Rieke et al. 2004).
### 2.8 Herschel
#### 2.8.1 PACS
M51 is observed by the Photo Array Camera & Spectrometer (PACS; Poglitsch et
al. 2010) as part of the Very Nearby Galaxy Survey (VNGS) at the 70 $\mu$m and
160 $\mu$m bands. Both bands are utilised in this study and images were
obtained from NED/IPAC. The pixel scale of the 70 $\mu$m and 160 $\mu$m data
is 1.4′′ (58 pc) and 2.85′′ (118 pc), respectively. The FWHM of the 70 $\mu$m
and 160 $\mu$m data are 5.6′′ and 11.4′′, respectively (Bendo et al. 2012).
#### 2.8.2 SPIRE
The Spectral and Photometric Imaging Receiver (SPIRE; Griffin et al. 2010)
observed M51 as part of the Very Nearby Galaxy Survey (VNGS) at the 250
$\mu$m, 350 $\mu$m and 500 $\mu$m bands. All three bands were used in this
study and images were obtained from NED/IPAC and processed by Bendo et al.
(2012). The pixel scale of the 250 $\mu$m, 350 $\mu$m and 500 $\mu$m data is
6′′ (250 pc), 8′′ (333 pc), and 12′′ (500 pc), respectively. The FWHM of SPIRE
data for 250 $\mu$m, 350 $\mu$m, 500 $\mu$m are 18.2′′, 24.5′′, and 36′′,
respectively (Bendo et al. 2012). Flux calibration error is accurate to $\pm
15\%$(Griffin et al. 2010).
Examples of imaging data used in this analysis are given in the left panels of
Fig. 1. They show 15x15 kpc images of M51 in the GALEX NUV, 2MASS K band, and
the SPIRE 500 ${\mu}$m.
Figure 1: The observed (left), azimuthally averaged observed (middle) and
model (right) images of M51 of various wavebands. The emission from M51b is
masked out from the figure. We indicate on each image elliptical apertures
(green) with radial distances of 0.8 and 2.7 kpc, which correspond to the
inner radius of the inner and main disc of our model, respectively. We also
plot in white the elliptical aperture corresponding to $R_{l}=7$ kpc, the
radius within which the assumption of axi-symmetry holds. The masked
foreground stars are indicated by the white dots.
### 2.9 Photometry
The observed images were corrected for Galactic extinction using a value of
$E(B-V)=0.0308$ (Schlafly & Finkbeiner 2011) and $R_{V}=3.1$ (Fitzpatrick
1999). Azimuthally averaged surface brightness profiles and spatially
integrated flux densities $F_{\nu}$ were derived using curve of growth (CoG)
photometry. Each map was segmented into 250 annuli of a fixed width in which
the average brightness was calculated, with the corresponding radius taken as
the midpoint of the annulus. The elliptical annuli were defined by the
inclination and position angle of M51, as listed in Table 1. Examples of
azimuthally averaged images (NUV, K, and 500 ${\mu}$m) produced in this way
are given in the middle panels of Fig. 1. Comparison between azimuthally
averaged and corresponding original observed images shows that the assumption
of axi-symmetry seems to hold up to $\sim 7$ kpc, which we take to represent
the limit of applicability of our axi-symmetric model. We will refer to this
limit as $R_{l}$.
The photometric errors due to background fluctuations, calibration and
configuration noise were derived using the procedure from Thirlwall et al.
(2020).
## 3 The Model
Our model for M51 follows the axisymmetric RT models of Popescu et al. (2011),
hereafter PT11, which describe the dust opacity and stellar emissivity for
spiral galaxies from the UV to submm wavelengths using parameterised analytic
functions. The model considers the absorption and anisotropic scattering of
stellar photons with dust grains of various size and chemical composition. A
dust composition consisting of silicates, graphites and PAH molecules was
used, with a grain size distribution from Weingartner & Draine (2001), and
optical constants from Draine & Lee (1984) and Draine & Li (2007). As in our
previous work on modelling spiral galaxies, we consider a Milky Way-type dust,
since this has proven to provide consistent results for all modelled galaxies.
The parameters describing the grain size distribution and grain mixture are
taken from Table 1, line 7 of Weingartner & Draine (2001). The RT model of M51
also incorporates explicit calculations for the stochastic heating of the dust
grains.
The RT codes used for the modelling of M51 were the PT11 code and the DART-RAY
code (Natale et al. 2014, 2015; Natale et al. 2017). The PT11 code is a
modified version of the code of Kylafis & Bahcall (1987). Both the PT11 and
DART-RAY codes use ray-tracing algorithms.
Following Thirlwall et al. (2020), we adapt the model of PT11 to allow for
additional morphological components. In the case of M51 it was found that, in
addition to the bulge, an inner and a main disc component were present out to
$R_{l}$. Each morphological component comprises the generic stellar and dust
components from PT11: the stellar disc, the thin stellar disc, the dust disc
and thin dust disc. The stellar disc is made up of all the stars that are old
enough to have had time to migrate and form a thicker configuration (a few
hundred parsec scale height), while the thin stellar disc is made up by young
stars, that are still spatially closely related to the molecular layer from
which they formed (50-100 parsec scale-height). A clumpy component associated
with the star forming clouds is also incorporated in the model. All these are
described in detail in PT11. Following Thirlwall et al. (2020) we also found
the need to alter the analytic functions describing the exponential discs, to
include an inner truncation radius. Thus, we use Eq. (1) and (2) from
Thirlwall et al. (2020) to describe the stellar volume emissivity and the dust
density distribution for each disc component j:
$w_{\text{j}}(R,z)=\begin{cases}\\!0,&\text{if }R<R_{\text{tin,j}}\\\ \\\
\begin{aligned}
A_{0,\text{j}}\left[\frac{R}{R_{\text{in,}\text{j}}}\left(1-\chi_{\text{j}}\right)+\chi_{\text{j}}\right]\\\
\times\text{exp}\left(-\frac{R_{\text{in},\text{j}}}{h_{\text{j}}}\right)\text{exp}\left(-\frac{z}{z_{\text{j}}}\right),\end{aligned}&\text{if}R_{\text{tin,j}}\leq
R<R_{\text{in,j}}\\\ \\\
A_{0,\text{j}}\times\text{exp}\left(-\frac{R_{\text{in},\text{j}}}{h_{\text{j}}}\right)\text{exp}\left(-\frac{z}{z_{\text{j}}}\right),&R_{\text{in},\text{j}}\leq
R\leq R_{\text{t},\text{j}}\end{cases}$ (1)
with:
$\displaystyle\chi_{\text{j}}=\frac{w_{\text{j}}(0,z)}{w_{\text{j}}(R_{\text{in},\text{j}},z)}$
(2)
where $R$ and $z$ are the radial and vertical coordinates, $h_{\text{j}}$ and
$z_{\text{j}}$ are the scale-length and scale-height respectively,
$A_{0,\text{j}}$ is a constant scaling the amplitude, $\chi_{\text{j}}$
describes a linear slope of the radial distributions from the inner radius,
$R_{\rm{in}}$, to the centre of the galaxy. $R_{\text{tin}}$ and
$R_{\text{t}}$ are the inner and outer truncation radii.
Eqs. 1 and 2 are wavelength dependent. The spatial integration of these
formulae (Eqs. B1-B3 from Thirlwall et al. 2020) between $R_{\text{tin}}$ and
$R_{\text{t}}$ provides the intrinsic stellar luminosity $L_{\nu}$ if j
represents a stellar disc component, or the dust mass $M_{\rm d}$ if j
represents a dust disc component.
The emissivity of the bulge is described by the Sérsic distribution
$w_{\text{bulge}}(\lambda,R,z)=w(\lambda,0,0)\,\sqrt{\frac{b_{\rm
s}}{2\pi}}\,\frac{(a/b)}{R_{\rm e}}\,\eta^{(1/2n_{\rm s})-1}\exp{(-b_{\rm
s}\,\eta^{1/n_{\rm s}})}$ (3)
with:
$\eta(\lambda,R,z)=\frac{\sqrt{R^{2}+z^{2}(a/b)^{2}}}{R_{\rm e}}$ (4)
$\eta^{\text{bulge}}(\lambda,0,0)$ is the stellar emissivity at the centre of
the bulge, $R_{\text{e}}$ is the effective radius of the bulge, $a$ and $b$
are the semi-major and semi-minor axes of the bulge, respectively, and
$b_{\text{s}}$ is a constant which depends on the value of the Sérsic index
$n_{\text{s}}$ (see Eq. 8 from Natale et al. 2021). Following Thirlwall et al.
(2020), the parameters of the different components are as follows:
The stellar disc:
The stellar disc is made up of the old stellar population, defined as a
population of stars that had time to migrate from their molecular layer into a
thicker configuration (a few hundred parsec scale height). Following from
Popescu et al. (2000) and PT11, we make the approximation that the stellar
disc only emits in the optical/NIR (no UV counterpart). It is described by the
four geometrical parameters $h_{\rm{s}}^{\text{disc}}$,
$z_{\rm{s}}^{\text{disc}}$, $R_{\rm{in,s}}^{\text{disc}}$ and
$\chi_{\rm{s}}^{\text{disc}}$, and one amplitude parameter $L^{\text{disc}}$.
The free parameters are $h_{\rm{s}}^{\text{disc}}$,
$\chi_{\rm{s}}^{\text{disc}}$ and $L^{\text{disc}}(\lambda)$.
$z_{\rm{s}}^{\text{disc}}$ is fixed from the model, and
$R_{\rm{in,s}}^{\text{disc}}$ is fixed from the data. We allow
$L^{\text{disc}}$ to be a free parameter for each wavelength corresponding to
u, g, r, i, J, K, I1, I2, I3. We also allow $h_{\rm{s}}$ to be wavelength
dependent, hence we consider the scale-lengths for the g, r, i, J, K, I1, I2,
I3 bands to be free parameters. The scale-length in the u-band is taken to be
the same as in the g-band. The model of M51 includes an inner and a main
stellar disc.
The thin stellar disc
The thin stellar disc is made up of the young stellar population of stars that
did not have time to migrate from their molecular layer (50-100 pc). This disc
emits predominantly in the UV, although we assume a smooth decrease of its
emission in the optical/NIR. It is described by the four geometrical
parameters $h_{\rm{s}}^{\text{tdisc}}$, $z_{\rm{s}}^{\text{tdisc}}$,
$R_{\rm{in},s}^{\text{tdisc}}$ and $\chi_{\rm{s}}^{\text{tdisc}}$, and one
amplitude parameter $L^{\text{tdisc}}$. The free parameters are
$h_{\rm{s}}^{\text{tdisc}}$, $\chi_{\rm{s}}^{\text{tdisc}}$ and
$L^{\text{tdisc}}$. $z_{\rm{s}}^{\text{tdisc}}$ is fixed from the model and
$R_{\rm{in},s}^{\text{tdisc}}$ is fixed from the data. The scale-length of
this disc is assumed to be wavelength independent. The wavelength dependence
of $L^{\text{tdisc}}$ is fixed from the model. The model of M51 includes an
inner and a main thin stellar disc.
Keeping in line with previous modelling, we express the spectral integrated
luminosity of the young stellar disc $L^{\rm tdisc}$ in terms of a star
formation rate ${\rm SFR}$, using equations (16), (17), and (18) from PT11.
The dust disc
It is described by the geometrical parameters $h^{\rm{disc}}_{\rm{d}}$,
$z_{\rm{d}}^{\rm{disc}}$, $R_{\rm{in,d}}^{\rm{disc}}$ and
$\chi_{\rm{d}}^{\rm{disc}}$ and the amplitude parameter, the face-on optical
depth of the B-band at the inner radius
$\tau_{\rm{B}}^{\rm{f,disc}}\left(R_{\rm{in,d}}^{\rm{disc}}\right)$. The free
parameters are $h^{\rm{disc}}_{\rm{d}}$, $\chi_{\rm{d}}^{\rm{disc}}$ and
$\tau_{\rm{B}}^{\rm{f,disc}}\left(R_{\rm{in,d}}^{\rm{disc}}\right)$.
$z_{\rm{d}}^{\rm{disc}}$ is fixed from the model and
$R_{\rm{in,d}}^{\rm{disc}}$ is fixed from the data. Our model of M51 contains
an inner, and main dust disc.
The thin dust disc
The thin dust disc is described by the geometrical parameters
$h^{\rm{tdisc}}_{\rm{d}}$, $z_{\rm{d}}^{\rm{tdisc}}$,
$R_{\rm{in,d}}^{\rm{tdisc}}$ and $\chi_{\rm{d}}^{\rm{tdisc}}$ and the
amplitude parameter, the face-on optical depth of the B-band at the inner
radius $\tau_{\rm{B}}^{\rm{f,tdisc}}\left(R_{\rm{in,d}}^{\rm{tdisc}}\right)$.
Following PT11 we fix the geometric parameters of the thin dust disc to that
of the thin stellar disc. The model of M51 contains an inner, and main thin
dust disc.
The clumpy component
The clumpy component represents the dense dusty clouds associated with star
forming regions that are illuminated by the strong, UV-dominated radiation
fields produced by the photons emitted from the young stars inside the clouds.
The dust in these regions reaches thermal equilibrium with the intense
radiation fields and emits dominantly in the 24 $\mu$m to 70 $\mu$m region,
often surpassing the emission due to the diffuse component at these
wavelengths. Following PT11 we define a “clumpiness factor” $F$, which is the
amplitude parameter of the clumpy component, and represents the fraction of
the total luminosity of massive stars that is locally absorbed by the dust in
these star-forming clouds. The diffuse component is then illuminated by the
escape fraction $1-F$. The $F$ factor has a wavelength dependence that is not
determined by the optical properties of the dust grains, but by the
evolutionary stage of the birth clouds. This is because in our model the dust
cloud is very opaque, thus completely blocking any radiation from inside the
cloud. The only way photons escape is through holes in the cloud, produced by
the fragmentation in the cloud. The wavelength dependence of the escaping
radiation thus arises because lower mass and less blue stars spend a higher
proportion of their lifetime radiating when they are further away from their
birth-clouds, and because of the progressive fragmentation of the clouds. The
formalism is described in Tuffs et al. (2004). The dust emission SED of the
birth clouds is taken to follow the prescription from PT11.
## 4 Fitting the model
To directly compare the model to the data we produce azimuthally averaged
surface brightness profiles of both. The observed surface-brightness profiles
were obtained by carrying out curve of growth photometry on the maps listed in
Table 2, as described in Sect. 2. For the model profiles we carried out an
identical procedure. Examples of these profiles can be seen in Figs. 2-3.
Examples of model images are given in the right panels of Fig. 1.
From an initial inspection of the surface brightness profiles of the
observational data it was clear that a single disc described by Eqs. 1-2 would
provide an inadequate fit. Following Thirlwall et al. (2020), we segment the
galaxy into distinct morphological components to fully capture the detailed
morphology of M51. As mentioned in Sect. 3, we found that, within $R_{l}$, in
addition to the bulge, two distinct morphological components were needed to
model this galaxy: an inner and a main disc. We allow for each morphological
component to be made up of two stellar and two dust components.
We followed the same fitting procedure as in Thirlwall et al. (2020). In
brief, this involves an intelligent searching algorithm of the parameter
space, that takes into account the fact that different parameters do not
equally affect the emission at all wavelengths. This allows one to identify
pairs of parameters that predominantly shape emission at key wavelengths and
use this to fit one pair (usually made of a geometric parameter and an
amplitude parameter) of parameters at a time.
Thus, we begin modelling with some initial guess parameters. These parameters
are based on notable features of the surface brightness profiles, as well as
informed by previous modelling (Thirlwall et al. 2020). We start by fitting
the NUV profile. This is equivalent to constraining the spatial distribution
of the young stellar population residing in the thin stellar disc. Although
this stellar disc is highly attenuated in the UV, we can still derive an
estimate of the corresponding geometrical and amplitude parameters for an
initial guess of dust opacity. We thus derive an exponential scale length
$h^{\rm{tdisc}}_{\rm{s}}$ and a luminosity density $L^{\rm{tdisc}}$ in the NUV
filter, for the inner and main disc, assuming a known wavelength dependence of
the luminosity density at the other sampled wavelengths. There is also the
free parameter $\chi_{\rm s}^{\rm tdisc}$, which is fitted for the inner disc
in order to account for the fall of emissivity towards the galaxian centre.
For the main disc $\chi_{\rm s}^{\rm tdisc}$ is only used to provide a smooth
overlap between the inner and the main disc, resembling the observed averaged
profiles, instead of simply using delta functions. $\chi_{\rm s}^{\rm
m-tdisc}$ is therefore not a critical parameter in the fit. Inner radii
$R_{\rm in}$ and inner and outer truncation radii $R_{\rm tin}$ and $R_{\rm
t}$ are fixed from data and are not subject to iterations. Once we obtain a
good fit in the NUV (for the trial dust opacity), we then attempt to constrain
the dust distribution . To do this we re-run the RT calculation with the
parameters derived from fitting the NUV profile and for an initial guess for
the optical emission, and predict the dust emission at all infrared
wavelengths, including the 500 $\mu$m. We then constrain the spatial
distribution of the dust by comparing the predicted 500 $\mu$m model profiles
with the corresponding observed ones, since at this wavelength the dust
emission primarily traces dust column density rather than heating sources. At
this step small changes in the dust parameters may be needed for a best fit to
the 500 $\mu$m data. This results in a fit for the scale-length of the dust
disc $h^{\rm{disc}}_{\rm{d}}$ and a dust opacity (plus the parameter
$\chi_{\rm d}^{\rm tdisc}$). A few iterations are needed between NUV and the
500 $\mu$m, until a good fit at both NUV and the 500 $\mu$m is achieved.
Towards the end of the iteration the FUV luminosity density is also slightly
adjusted until the best fit to the FUV profile is also obtained. Once the
parameters related to the thin stellar disc and dust disc are constrained, we
run again the RT code to derive the distribution of the stellar disc and bulge
harbouring the old stellar population, by fitting the optical and NIR profiles
at each wavelength for which observational data is available. Since in a face-
on galaxy the optical and NIR emission is fairly optically thin to dust, the
final fits at these wavelengths do not alter the solution for for the dust
disc and the thin stellar disc (in the UV) obtained in previous iterations. A
clumpy component is finally fitted to account for any remaining 24-70 ${\mu}$m
observed emission not accounted for by the diffuse dust emission. Overall,
within a small number of iterations a solution is derived for the whole
spectral range. Notably, the 70-350 ${\mu}$m emission is not directly used in
the fit, but instead is predicted. To conclude, the fit mainly consists in
fitting only a pair of parameters (a scale-length and an amplitude) at any
given time, for each morphological component at a given wavelength. More
details on the fitting procedure can be found in Thirlwall et al. (2020).
Figure 2: Comparison between the azimuthally averaged surface brightness
profiles of the observations (solid black line) at selected UV/optical/NIR
wavelengths and of the corresponding model for dust attenuated stellar light
(solid red line). The blue shaded region around the observed profile
represents the corresponding errors in the averaged surface brightness, as
described in Sect. 2. The contribution from each morphological component is
plotted with dashed lines and colour-coded as follows: bulge in purple,
stellar disc in green and thin stellar disc in blue. The lower panels show the
residuals between the observations and our model, with the dashed blue lines
showing the $\pm 20\%$ residuals to guide the eye. Figure 3: Same as Fig. 2
but for selected wavelengths in the NIR/MIR/FIR/submm, where the emission from
the dust discs is plotted with the grey dotted lines and the dust emission
from the HII regions is shown by the dashed cyan lines.
The main (free) geometric parameters constrained by the procedure outlined
above are listed in Table 3. The free parameters related to stellar luminosity
densities are listed in Table 6. The free parameters related to dust opacity
are listed in Table 7. Other parameters of the model are listed in Table 8.
$h_{\text{s}}^{\text{i-disc}}$(g, r, i J, K, I1, I2, I3) | (0.74, 0.73, 0.72, 0.69, 0.66, 0.70, 0.72, 1.0)$\pm$7%
---|---
$h_{\text{s}}^{\text{m-disc}}$(g, r, i J, K, I1, I2, I3) | (3.8, 3.5, 3.7, 3.5, 3.5, 4.3, 4.9, 7.0)$\pm$13%
$h_{\text{s}}^{\text{i-tdisc}}$ | 0.6$\pm$0.15
$h_{\text{s}}^{\text{m-tdisc}}$ | 4.3$\pm$1.0
$h_{\text{d}}^{\text{i-disc}}$ | 5.0${}^{+3.5}_{-1.5}$
$h_{\text{d}}^{\text{m-disc}}$ | 6.0${}^{+3.0}_{-0.9}$
$\chi_{\text{s}}^{\text{i-tdisc}}$ | 0.1$\pm$0.02
$\chi_{\text{s}}^{\text{i-disc}}$ | 0
$\chi_{\text{d}}^{\text{i-disc}}$ | 0.45$\pm 0.05$
$R_{\text{eff}}$ | 0.35$\pm$0.1
Table 3: Main geometrical (free) parameters of the inner and main discs, and
of the bulge, derived for the standard model.
The results of the fits are presented in Figs. 2-3, showing an overall
reasonable agreement with the data. To help the comparison, the residuals $R$
between the surface brightness profiles of the observations and of the model
are also plotted on the bottom panels of Figs. 2-3. The residuals are defined
as:
$R=\frac{\text{model}-\text{observation}}{\text{observation}}$ (5)
In addition, the goodness of fit is also considered with the reduced chi-
squared $\text{chi}^{2}_{r}$ statistics444We use the notation $\text{chi}^{2}$
rather than $\chi^{2}$, as the later notation is reserved in this work for the
geometrical parameter describing the inner linear slope., which becomes more
relevant when finalising the model solution. We calculated
$\text{chi}^{2}_{r}$ for each wavelength by the following formulae:
$\displaystyle\text{chi}^{2}_{\lambda}=\sum_{n=1}^{N}\frac{\left(M_{n}-O_{n}\right)^{2}}{\epsilon^{2}_{\text{SB,}n}}$
(6)
$\displaystyle\text{chi}^{2}_{\text{r,}\lambda}=\frac{\text{chi}^{2}_{\lambda}}{N-m}\approx\frac{\text{chi}^{2}_{\lambda}}{N}$
(7)
where $N$ is the total number of annuli $n$, $m$ is the number of free
parameters fitted at wavelength ${\lambda}$, $O_{\text{n}}$ is the azimuthally
averaged flux through annuli $n$, $M_{\text{n}}$ is the model flux through
annuli $n$, $\epsilon^{2}_{\text{SB,}n}$ is the total error of the azimuthally
averaged surface brightness profile corresponding to annuli $n$. The total
number of annuli is 70 while $m$ varies between 2 in the FUV, to 5 (in the NUV
and 500 ${\mu}$m), to 8 (in the optical and NIR) for a fitted profile out to 7
kpc radius. At most other wavelengths that are predicted, we can assume zero
free parameters. Because $N$ is much larger than $m$ we neglect $m$ in Eq. 7.
Values of the $\text{chi}^{2}_{\lambda}$, for selected wavelengths, are listed
in Table 4. The total reduced chi-squared statistic over all wavelengths is:
$\text{chi}_{\rm{r}}^{2}=\frac{\sum_{l=1}^{L}\text{\text{chi}}^{2}_{\lambda,l}}{\sum_{l=1}^{L}N_{l}}$
(8)
Band | $\text{chi}^{2}_{r}$
---|---
NUV | 1.09
I2 | 0.74
PACS70 | 3.62
SPIRE500 | 1.33
Global | 2.44
Table 4: The $\text{chi}^{2}_{r}$ values at selected wavelengths, as well as
the global $\text{chi}^{2}_{r}$ (over all wavelengths), for the standard
model.
The $\text{chi}^{2}_{r}$ values from Table 4 confirm the trends seen in Figs.
2-3, proving that the axi-symmetric model provides a reasonable fit to the
data.
## 5 Results
### 5.1 The surface brightness distributions
Taking the azimuthally averaged surface brightness of the observations
produces relatively smooth exponential profiles, as seen in Figs. 2-3. Despite
the slight imperfections in the smoothness, the average profiles are still
suitable for description with analytic axi-symmetric functions. The most
significant deviations are in the UV and 24$\,\mu$m bands, where strong
asymmetries and clumpiness are present. The bumps in these profiles generally
coincide with the spiral arms as well as star forming regions. For example, a
significant bump in emission at 1.3 kpc, which can be seen most notably in the
FUV profile, is due to bright star forming regions. There is also a bump in
emission between 5 kpc and 6.5 kpc, which coincides with an overlap in the
spiral arms. Predictably, the 24 $\mu$m profiles exhibit the least smooth
curves, as at this wavelength the emission due to the diffuse dust and the
dust heated by localised SF regions is roughly equal, hence strong asymmetries
arise from these localised sources of emission. This is also seen in the IRAC
bands, where there is significant PAH emission.
Overall the model has reasonable agreement with the data with residuals
generally within 20%.
The scale-length of the observed profiles in the optical regime generally
decreases with increasing wavelength for each morphological component. For
wavelengths $\geq 3.6\,\mu$m this trend reverses for the main disc. This is
similar with trends found from previous modelling (M33; Thirlwall et al.
2020).
The observed profiles in the IRAC bands are modelled with emission from the
old stellar populations as well as dust emission, with a negligible
contribution from the young stellar population. At $3.6\,\mu$m and $4.5\,\mu$m
the emission is dominated by the old stellar population, but at $5.8\,\mu$m
the dust emission is almost equal in amplitude to that of the old stellar
population. At the $8.0\,\mu$m band the emission is dominated by the dust,
primarily due to the PAH emission.
### 5.2 The global SED of M 51
Figure 4: The global SED of the model (spatially integrated out to $R_{\rm
l}$) plotted together with the observed flux densities used to optimise our
model (shown with black squares). The error bars of the data are also shown,
although most are contained within the square symbols. The model SED is
represented by the solid black line, and covers both the dust attenuated
stellar SED in the UV/optical/NIR as well as the dust emission SED in the
NIR/MIR/submm. The contributions from each component of the model are plotted
with dashed lines as follows: thin stellar disc in dark blue, stellar disc in
green, bulge in purple, diffuse dust in grey and clumpy component in cyan. The
intrinsic stellar SED (as would be seen in the absence of dust) is plotted
with the solid brown line. We plot the relative residuals in the lower panel,
with the $\pm 20\%$ residuals indicated with the dotted blue line to guide the
eye. Figure 5: The predicted global SED of M 51a (solid black line) for the
intrinsic (dust de-attenuated stellar emission) and for the dust emission,
plotted together with the contribution from the individual morphological
components: the inner (dashed black line) and the main (dotted black line)
discs, and the bulge (dashed purple line).
The global SED of M51 was produced by spatially integrating the emission at
each wavelength out to ${\rm R}_{l}$. This was done both for observations and
for the model and the results are plotted in Fig. 4. The model SED closely
resembles the observed SED, with a maximum residual of $|R|=26.2\%$ and an
average residual of $\langle|R|\rangle=4.33\%$. The model performs worst at
$160\,\mu$m, where the emission is underestimated by our model.
We also show the contributions from the different morphological components to
the global SED in Fig. 5. It is important to note that the dust emission for a
given morphological component is not entirely powered by the stellar emission
of the same component, as for example, some fraction of the stellar photons
emitted from the inner disc will contribute to heating the dust in the main
disc. Nonetheless, we still show the separate contributions of the inner and
main discs to the dust emission, within the accepted understanding that there
is not a one-to-one spatial correspondence between the stellar and the dust
emission of these components. As expected due to its extent, the largest
contribution to the global SED is from the main disc responsible for 67.4% of
the total stellar emission, and 63.7% of the dust emission. The inner disc
contributes 30.7% to the total stellar emission and 36.3% to total dust
emission. The bulge contributes 1.89% to the total stellar emission. Here we
note that the separation between the inner disc and the bulge is rather
degenerate. We found no strong constraints for this separation. We therefore
consider the bulge versus inner disc luminosity to be subject to large
uncertainties.
The dust emission SED is dominated by the diffuse component. The HII component
has a smaller contribution, of $12.2\%$. We find a clumpiness factor of
$F=0.125$. The global dust emission SED peaks at $130\,\mu$m, which
corresponds to a dust temperature of 22 K. The hottest dust resides within the
inner disc, where the SED peaks at $102\,\mu$m, corresponding to a dust
temperature of 28 K. The main disc has the greatest contribution to the global
SED, therefore, as expected, it peaks at a similar wavelength, at around
$133\,\mu$m, corresponding to a dust temperature of 22 K.
We find that 38.3$\pm 4$% of the stellar light is absorbed and re-radiated by
the dust. This is a typical value for late-type spiral galaxies (Popescu &
Tuffs 2002). M33 was found to have a similar value of 35$\pm$3% (Thirlwall et
al. 2020). We also find that the young stellar population contributes 42.8% to
the total intrinsic stellar light, with 55.4% being due to the old stellar
population in the disc and the remaining 1.89% due to the bulge.
### 5.3 Star-formation rates in M51
We obtain a
$\text{SFR}=4.11^{+0.41}_{-0.39}\,$$\text{M}_{\odot}\text{yr}^{-1}$ which is
within the broad range of literature values. De Looze et al. (2014) obtained a
$\text{SFR}\sim 3.13\,$$\text{M}_{\odot}\text{yr}^{-1}$(555Value adjusted for
the distance of $D=8.58$ Mpc, as used in this study.) and Nersesian et al.
(2020b) found a $\text{SFR}=4.09\pm
0.35\,$$\text{M}_{\odot}\text{yr}^{-1}$(666Adjusted value.). Both these
studies use SKIRT, a Monte-Carlo RT model. Calzetti et al. (2005) obtain the
estimates $\text{SFR}(24\,\mu\text{m})\sim
3.7\,$$\text{M}_{\odot}\text{yr}^{-1}$(777Adjusted value.) and
$\text{SFR(FUV)}\sim 4.7\,$$\text{M}_{\odot}\text{yr}^{-1}$(888Adjusted
value.).
About two thirds of the star formation takes place in the main disc with
$\text{SFR}^{\text{m}}=2.73^{+0.28}_{-0.26}\,$$\text{M}_{\odot}\text{yr}^{-1}$.
This is expected due to both the size and bolometric output of the main disc.
For the inner disc we find
$\text{SFR}^{\text{i}}=1.38^{+0.15}_{-0.14}\,$$\text{M}_{\odot}\text{yr}^{-1}$.
We also calculate the surface density of star formation rate,
$\Sigma_{\text{SFR}}$, for each morphological component, by considering the
physical extent of each disc, which is shown in Tables 9-10. The global
surface density in SFR is $\Sigma_{\text{SFR}}=2.83^{+0.28}_{-0.27}\times
10^{-2}\,$$\text{M}_{\odot}\text{yr}^{-1}\text{kpc}^{-2}$.
### 5.4 Dust optical depth and dust mass
We find that the face-on optical depth in the B-band has a maximum value of
$\tau_{\text{B}}^{\text{f}}\left(R_{\text{in}}\right)=5.33\pm 0.33$ at the
inner radius of the inner disc. The main disc has a face-on optical depth at
the inner radius of $2.62\pm 0.08$. These values are listed in Table 7. We
find that M51 has a global dust mass (out to $R_{l}$) of
M${}_{\text{d}}=3.74^{+0.34}_{-0.30}\times 10^{7}\,\text{M}_{\odot}$. The dust
masses of the individual morphological components are listed in Table 11.
Using a gas mass of M${}_{\text{G}}=1.1\pm 0.03\times
10^{10}\,\text{M}_{\odot}$ (Mentuch Cooper et al. 2012) we obtain a gas-to-
dust ratio of $\text{GDR}=294\pm 4$.
### 5.5 Morphological components of M51
In modelling M51 we found three distinct morphological components out to
$R_{l}$, each requiring different geometrical parameters for the distribution
of stars and dust. Although we present the model for the entire galaxy, it is
also interesting to study the properties of these individual components. The
geometrical parameters of the three morphological components, as constrained
by our model are listed in Table 3.
#### 5.5.1 The inner disc
The inner disc extends from 0 kpc to 1.9 kpc and exists as a ring-like
structure for the young stars and dust, with an inner radius of 700 pc and 800
pc respectively. The distribution of the old stars, however, exponentially
increases towards the centre. The distribution of the dust is notably flat
with a scale length of 5 kpc, compared to a scale length of 600 pc for the
young stellar distribution. The inner disc has a
$\text{SFR}^{\text{i}}=1.38^{+0.15}_{-0.14}\,$$\text{M}_{\odot}\text{yr}^{-1}$,
contributing around 30% to the global SFR, and has the greatest surface
density of star formation rate
$\Sigma_{\text{SFR}}^{\text{i}}=1.21^{+0.28}_{-0.11}\times
10^{-1}\,$$\text{M}_{\odot}\text{yr}^{-1}\text{kpc}^{-2}$. The dust mass
within the inner disc is M${}_{\text{d}}^{\text{i}}=6.39^{+0.37}_{-0.21}\times
10^{6}\,\text{M}_{\odot}$. The temperature of the dust contained within the
inner disc is the hottest in the galaxy, at around 28 K.
#### 5.5.2 The bulge
The bulge was found to have an effective radius of 350 pc and was modelled
with a Sérsic index of 4. As mentioned in Sect. 5.2, the bulge parameters are
subject to uncertainty, in the sense that there are different configurations
in the bulge/inner disc decomposition that could result in a good fit to the
data. We based our choice on covering a small dip in the profile, that would
otherwise appear in other configurations, and as such we believe that our
choice is the most reliable result. Nonetheless, we caution the reader of the
subjectivity of the approach. The error estimates on the bulge/inner disc
parameters do not take into account the uncertainty due to the other possible
combinations.
#### 5.5.3 The main disc
The main disc resides in the region between 1.8 kpc and 6.8 kpc. The interior
of the main disc slightly overlaps with the outermost limits of the inner
disc. Both the stellar and dust distribution are relatively flat in the main
disc, with a scale length of 4.3 kpc for the young stellar disc and 6 kpc for
the dust disc. The main disc contributes most significantly to the global SFR,
with a
SFR${}^{\text{m}}=2.73^{+0.28}_{-0.26}\,$$\text{M}_{\odot}\text{yr}^{-1}$. The
main disc has a surface density SFR of
$\Sigma_{\text{SFR}}^{\text{m}}=2.00^{+0.18}_{-0.17}\times
10^{-2}\,$$\text{M}_{\odot}\text{yr}^{-1}\text{kpc}^{-2}$. The main disc has a
dust mass of M${}_{\text{d}}^{\text{m}}=3.10^{+0.26}_{-0.18}\times
10^{7}\,\text{M}_{\odot}$, with a dust temperature of 21.8 K.
### 5.6 The contribution of young and old stellar populations in heating the
dust
The fractional contribution of the different stellar populations in heating
the dust is an important quantity, since it allows a quantitative
understanding of the physical mechanisms related to the interaction between
stellar light and interstellar dust. In addition, this is key to the
understanding of key correlations, like for example the tight and universal
correlation between far-infrared and radio continuum emission of spiral
galaxies (de Jong et al. 1985; Helou et al. 1985; Wunderlich et al. 1987). The
standard interpretation of the so-called FIR-radio correlation is in terms of
young and massive star formation activity (Condon 1992). This picture assumes
that the dust emission is mainly powered by young stars which are also
responsible for the radio emission: the ionising radiation from the young
stars powers the thermal radio emission, and the remnants of the supernova
explosions which occur at the end of their lives accelerate the cosmic ray
electrons. Nonetheless both the effect of dust opacity and the contribution of
the old stellar populations play a role in shaping the slope of the
correlation (Pierini et al. 2003), and as such, knowledge of the fraction of
dust heating powered by the young and old stellar populations is a pre-
requisite for the analysis of the FIR/radio correlation. Furthermore,
knowledge of the different stellar populations in heating the dust is
equivalent to understanding the dust attenuation of the different stellar
populations, which, in turn, allows population synthesis models to be used for
deriving star formation histories in galaxies.
For the global emission of M51 we found that the fractional contribution of
the young stellar population in powering the dust emission, $F^{\rm dust}_{\rm
young}$, is $69\%$, the rest being attributed to the old stellar populations
in the disc and bulge. Previous modelling of M51 with radiative transfer
methods derived $F^{\rm dust}_{\rm young}=0.63$ (De Looze et al. 2014) and
$F^{\rm dust}_{\rm young}=0.72$ (Nersesian et al. 2020b). The latter value is
more in range with our determination.
Our value of $F^{\rm dust}_{\rm young}=0.69$ is similar to that derived with
the same models for other nearby spirals: $F^{\rm dust}_{\rm young}=0.71$ for
the Milky Way (Natale et al. 2021) and NGC 628 (Rushton et al. 2022), but
slightly lower than $F^{\rm dust}_{\rm young}=0.8$ for M33 (Thirlwall et al.
2020).
Following the formalism from Natale et al. (2015) and Popescu et al. (2000),
we use the radiative transfer calculations to not only derive $F^{\rm
dust}_{\rm young}$ and $F^{\rm dust}_{\rm old}$ for the global emission, but
also to understand the spatial variation of this quantity. In Fig. 6 we plot
$F^{\rm dust}_{\rm young}$ and $F^{\rm dust}_{\rm old}$ as a function of
radial distance. Within the inner 0.5 kpc dust emission is dominated by the
old stellar populations, mainly from the bulge. Beyond 0.5 kpc it is the young
stellar population that contributes a significant fraction of the dust
heating. At around 0.7-0.8 kpc (the inner radius of the inner disc) there is a
peak in $F^{\rm dust}_{\rm young}$, reaching above $70\%$. Within the extent
of the main disc, between 2.7 and 7 kpc, $F^{\rm dust}_{\rm young}$ is slowly
increasing with radial distance, from $60$ to $85\%$. At around 2 kpc,
corresponding to the transition region between the inner and the main disc,
$F^{\rm dust}_{\rm young}\approx F^{\rm dust}_{\rm old}$.
Figure 6: Radial profile of the fraction of dust heating $F^{\rm dust}$
powered by the young (blue) and old (red) stellar populations.
The overall trends are similar to those found for the Milky Way (Natale et al.
2021) and for NGC 628 (Rushton et al. 2022), in the sense that dust heating in
the inner regions has a significant contribution from old stellar populations,
while the rest of the dust in the disc is mainly heated by young stellar
populations, with $F^{\rm dust}_{\rm young}$ rather constant or slowly
increasing towards the outer disc. M51 has larger radial variations in $F^{\rm
dust}$ than the Milky Way or NGC 628.
### 5.7 The radial variation of $\Sigma_{\text{SFR}}$,
$\Sigma_{\text{M}_{\star}}$nd sSFR
The global SFR of M51 was found to be $4.11\,\text{M}_{\odot}\,{\rm yr}^{-1}$.
The rate of star formation is not constant throughout the extent of M51, but
has a strong radial dependence. To illustrate this we plot in the top panel of
Fig. 7 the surface density of SFR, $\Sigma_{\text{SFR}}$, versus radial
distance. One can see that there is a peak in $\Sigma_{\text{SFR}}$ at the
inner radius of the inner disc, followed by a steep decrease out to around 2
kpc. Beyond 3 kpc, within the extent of the main disc, the
$\Sigma_{\text{SFR}}$ follows a shallow decrease. Overall,
$\Sigma_{\text{SFR}}$ shows a marked difference between the inner and the main
disc. By contrast, the surface density of stellar mass,
$\Sigma_{\text{M}_{\star}}$, has a smoother radial variation, (middle panel,
Fig. 7), showing a shallow decline with increasing radial distance. The
stellar mass was calculated using the 3.6 and 4.5 ${\mu}$m flux densities and
the calibration from Eskew et al. (2012).
The specific SFR, sSFR is rather constant throughout the disc (bottom panel,
Fig. 7), except for the very central region. These trends are similar to the
trends found for M33, NGC 628 and the Milky Way.
Figure 7: Radial profiles of the SFR surface density. $\Sigma_{\text{SFR}}$
(top), stellar mass surface density, $\Sigma_{\text{M}_{\star}}$ (middle), and
specific star formation rate, sSFR (bottom).
### 5.8 The attenuation curve of M51
The attenuation curve of a galaxy depends not only on the optical constants of
the dust grains and their grain size distribution (incorporated in the dust
extinction), but also on the relative distribution of the stars and dust.
Neither the extinction nor the geometry is usually known, making predictions
for attenuation rather difficult. It is however the strength of radiative-
transfer modelling to accurately derive geometries of stars and dust, within
the framework of a given dust model.
In Fig. 8 we plot the predicted attenuation curve of M51, as derived from our
radiative transfer model. For comparison we also plot the extinction curve of
the MW from Fitzpatrick (1999). We fit the attenuation curve of our model with
a third-order polynomial plus a Drude profile, which is the functional form
presented in Salim et al. (2018). Using this form, we found that the best fit
is given by:
$\displaystyle k_{\lambda}$
$\displaystyle=-4.95+2.00\lambda^{-1}-0.29\lambda^{-2}+0.02\lambda^{-3}+D_{\lambda}+4.00$
(9) $\displaystyle D_{\lambda}$
$\displaystyle=\frac{3.00\,\lambda^{2}(0.035\,\mu\text{m})^{2}}{[\lambda^{2}-(0.2175\,\mu\text{m})^{2}]^{2}+\lambda^{2}(0.035\,\mu\text{m})^{2}}$
(10)
Figure 8: Comparison between the attenuation curve of M51 as seen at an
inclination of 20.3∘ (solid red line) with the average extinction curve of the
Milky Way (solid black line) (Fitzpatrick 1999). Both curves are normalised to
the corresponding values in the B-band.
The attenuation curve of M51 is slightly steeper than the MW extinction curve,
becoming increasingly disparate at shorter wavelengths. M51 also exhibits a
much greater 2200 Å bump compared to the MW extinction. We find that the
difference in width of the bumps to be negligible.
## 6 Discussion
Figure 9: The azimuthally averaged surface-brightness observed profiles
(black) plotted against our model (red) and the model of Nersesian et al.
(2020b) (blue), for a selection of wavebands. The lower panels indicate the
residuals for our model (red) and the model of Nersesian et al. (2020b)
(blue), with the $\pm 50$ per cent shown with the dashed blue line.
### 6.1 Comparison with the non axi-symmetric model of Nersesian et al. 2020b
As mentioned before, M51 was also modelled with non axi-symmetric RT codes by
De Looze et al. (2014) using the SKIRT code (Baes et al. 2011; Camps & Baes
2020), and more recently by Nersesian et al. (2020b), using a more up to date
version of the same formalism and code. Here we compare our model to that of
Nersesian et al. (2020b). Publicly available model images of Nersesian et al.
(2020b) were extracted from
DustPedia999https://sciences.ugent.be/skirtextdat/DustPediaData/M51/mock_images.zip
and processed in the same way as our model images, to produce azimuthally-
averaged radial profiles. We also checked that the observed fluxes used in
Nersesian et al. (2020b) were consistent with the values used in our work,
and, when small differences existed (at percent level), we adjusted for this.
In Fig. 9 we show examples of this comparison for a few selected wavelengths.
Overall our model performs better. This is to be expected, as the model of
Nersesian et al. (2020b) does not fit the geometry of the system, but only the
spatially integrated SED.
### 6.2 An outer disc
All the results presented so far were confined to the main extent of the
galaxy, out to the radial distance $R_{\rm l}=7$ kpc. Note however that
$R_{\rm l}$ is not the actual truncation radius of M51. There is very faint
and asymmetric emission beyond $R_{\rm l}$, also containing the bridge between
M51 and its companion M51b. This faint emission does not obey axi-symmetry,
therefore it is not a prime target for an axi-symmetric model. Nonetheless,
the emission is so faint that it is difficult to trace and model, and one
effective way to reveal it is through azimuthally averaging. By averaging, a
strong signal becomes apparent, in a form of an exponential decay. We model
this outer emission with another morphological component that we call “outer
disc” . While we acknowledge that an axi-symmetric model is a poor
approximation for this emission, we nevertheless attempt to at least give some
characterisation of the global properties of this emission.
The model for the outer disc morphology contains the same geometrical
components as for the main disc: a stellar disc, a thin stellar disc, a dust
disc and a thin dust disc. The analytic functions we fit are the same, and the
whole optimisation follows the same procedure as for the main body of the
galaxy.
Figure 10: Fits to the surface brightness profiles of the outer disc (7-20
kpc) at selected wavelengths, chosen to show examples of dust attenuated
stellar light in the NUV and 3.6 ${\mu}$m and dust emission in the 24 and 500
${\mu}$m. The black line is for the observed profiles and the red line is for
the corresponding model. The bottom panels show residuals between observations
and model.
Examples of fits to the outer disc are given in Fig. 10. We find the outer
disc to extend from 6.7 kpc to 20 kpc, with a scale length of 3.6 kpc and 1.45
kpc for the dust disc and young stellar disc, respectively. The main
geometrical parameters of the outer disc are listed in Table 5.
The outer disc is the morphological component with the smallest SFR, with a
SFR${}^{\text{o}}=0.57\pm 0.06\,$$\text{M}_{\odot}\text{yr}^{-1}$. The surface
density of SFR is the least dense of the components, with
$\Sigma_{\text{SFR}}^{\text{o}}=5.11\pm 0.56\times
10^{-4}\,$$\text{M}_{\odot}\text{yr}^{-1}\text{kpc}^{-2}$.
$h_{\text{s}}^{\text{o-disc}}$(g, r, i J, K, I1, I2, I3) | (2.0, 2.0, 2.0, 1.72,
---|---
| 1.77, 1.95, 1.98, 2.00)$\pm$13%
$h_{\text{s}}^{\text{o-tdisc}}$ | 1.45$\pm$0.2
$h_{\text{d}}^{\text{o-disc}}$ | 3.6${}^{+1.4}_{-1.0}$
Table 5: Main geometric parameters of the outer disc
### 6.3 A model with a spatial variation in dust properties
Inspection of the fitted profiles from Figs. 2-3 shows that the axi-symmetric
model provides a reasonable fit to the data, within $R_{l}$.
However, when deriving the spatially integrated SED (see Fig. 11, top left-
hand panel), we noticed that our predicted intrinsic (de-attenuated) stellar
SED shows a spike in the NUV, exactly at the position of the well-known “2200
Å bump” of the extinction curve. It can be seen that the inner disc
contributes most significantly to this spike, while the main disc has
negligible contribution, if any. The intrinsic stellar SED should be smooth,
with no feature at this wavelength. Because our model is in good agreement
with the data at all wavelengths, as seen in Figs. 2-3, the spike in the
intrinsic SED is indicative that the inner disc is being too heavily
attenuated in the NUV. This could be due to a spatial variation in the dust
properties, in the sense that the dust properties may be different in the
inner disc than assumed in our model. As mentioned before, our model for M51
uses the Milky Way-type dust from Weingartner & Draine (2001), described on
line 7 in Table 1 from Weingartner & Draine (2001). The extinction curve of
the Milky Way dust model along with the contributions from the different grain
types is shown in the left panel of Fig. 12, together with its prominent “2200
Å bump”. This feature is primarily due to the PAH molecules, which the model
considers to be the carbonaceous grains with sizes $a\leq 0.01\,\mu$m. The
over-attenuation of the NUV photons in the inner disc could suggest that the
extinction curve corresponding to the dust model that we use has a too strong
bump, meaning an overabundance of PAH in the inner disc, than is present in
M51. We note that it is extremely hard to constrain the shape of the UV
extinction curve of galaxies, especially in the inner region where the
attenuation is high and the geometry is complex, with just the GALEX FUV and
NUV bands. Several authors have shown that these bands alone are insufficient
to disentangle the shape of the continuum and the strength of the 2200A bump
(Hoversten et al. 2011; Hutton et al. 2015; Hagen et al. 2017; Decleir et al.
2019).
Nonetheless, within the framework of a model with varying dust properties, our
fits at $8\,\mu$m (see Fig. 11, bottom right-hand panel) provide additional
evidence that we may need a different type of dust in the inner disc. The
model profile is clearly over-predicted in the inner region. Since the
$8\,\mu$m is a strong tracer of PAH abundance, in addition to tracing the
radiation fields that are self-consistently calculated in our RT model, the
over-prediction could again be indicative of an overabundance of PAH in our
model of the inner disc.
As an alternative, we sought a dust model with a reduced PAH abundance and
chose one of the LMC dust models of Weingartner & Draine (2001), namely the
LMC average (line 1 from Table 3 of Weingartner & Draine 2001). This LMC-type
dust features the same optical constants but different grain size
distributions, particularly less PAH grains. Thus, the extinction curve
exhibits a less pronounced 2200 Å bump, as seen in the right panel of Fig. 12.
As the attenuation in the main and outer discs produced smooth intrinsic SEDs,
we chose to modify the dust properties only in the inner disc (using the LMC-
type dust). We thus fitted a new model of M51 with variable dust grain
properties. We refer to this alternative as the “hybrid model”. The new fits
produce the intended effect, reducing the NUV peak in the de-attenuated
stellar SED of the inner disc, as seen in Fig. 11. The new fits to the
$8\,\mu$m profiles (see Fig. 11, bottom right-hand panel) also provide the
right level of emission at this wavelength. We conclude that using the LMC-
type dust for the inner disc could provide an alternative to the standard
model.
The finding that PAHs could be less abundant in the inner disc than in the
rest of the galaxy raises the question of what physical processes could cause
this variation. Theoretical and observational studies suggest that strong UV
radiation fields could destroy PAH molecules (e.g., Boulanger et al. 1988;
Helou et al. 1991; Contursi et al. 2000; Siebenmorgen & Krügel 2010). Taking
into account that the inner region of M51 has the highest surface density of
SFR, and therefore intense UV radiation fields, it is plausible that PAHs are
more readily destroyed in this region, thus explaining the reduced abundance
inferred by the hybrid model. Several studies have suggested that the fraction
of PAHs becomes low in HII regions (e.g., Pety et al. 2005; Lebouteiller et
al. 2007; Thilker et al. 2007; Chastenet et al. 2023) and is a strong function
of environment, with higher fractions in less harsh environments due to star
formation (Chastenet et al. 2023). It is beyond the scope of our paper to
provide a quantitative model for PAH destruction in different ISM
environments, but we note the possibility of PAH abundance variation within
galaxies.
Figure 11: Top panels: the standard model (constant dust properties) for the
intrinsic global SED (top left) and the surface brightness profile at 8 $\mu$m
(top right). Bottoms panels: same but for the hybrid model (with spatially
varying dust properties). The left hand plots show the intrinsic global SED
(solid black line) plotted together with the contribution from the individual
morphological components: the inner (dashed black line) and main (dotted black
line) discs as well as the bulge (dashed purple line). The right hand plots
show the $8\,\mu$m profile of the observation (solid black line) compared to
the model (solid red line) along with the contributions from the dust discs
(dotted grey lines), and the HII components (dashed cyan lines). Residuals
between observed and model profiles are also shown on the bottom of each
panel.
The fits to the multiwavelength data using the hybrid model do not provide any
big differences from the standard model, with the exception of the 8 $\mu$m
profile, which is largely improved (see examples in the Appendix, in Figs.
13-14). The energy balance between direct and re-radiated stellar light seems
to be the same as in the standard model, with no major improvements in the
fit. The values of the ${\rm chi}^{2}$ are also similar. The individual values
at selected wavelengths are listed in Table 12. The values of the derived
intrinsic parameters are different, but the quality of the fit is the same.
This is perhaps not surprising, since the optical properties of the grains are
not changed. There are some other dust models available in the literature, in
particular "The Heterogeneous dust Evolution Model for Interstellar Solids"
(THEMIS) (Jones et al. 2013; Köhler et al. 2014; Ysard et al. 2015), which
consider different carbonaceous grains, with aliphatic rather than aromatic
molecules. The THEMIS model has enhanced submm efficiencies for grains, by a
factor of about 2.5 with respect to the Draine model. The THEMIS model would
produce a different energy balance and therefore different fits.
To conclude, the hybrid model with a reduced PAH abundance in the inner disc,
together with a modified grain size distribution in the central region,
provides a consistent solution to the panchromatic data of M51, and is based
on observational evidence coming from both dust extinction and dust emission.
Nonetheless, we cannot prove that these observational constraints are uniquely
met by our hybrid model.
## 7 Summary
Using the generic radiative transfer code and formalism of PT11 we derived an
axisymmetric model for M51, fitted to multiwavelength imaging data ranging
from FUV to submm. We find that, despite the interaction with the companion
galaxy M51b, M51 preserves a regular spiral structure within a radial distance
of $R_{\rm l}$=7 kpc, making it thus suitable for the assumption of axi-
symmetry within $R_{\rm l}$. The model fits to the azimuthally-averaged
surface brightness profiles were found to be in reasonable agreement with the
observations.
We find three distinct morphological components out to $R_{\rm l}$: a bulge,
an inner disc and a main disc. The bulge has an effective radius $R_{\rm
eff}=350$ pc and a Sérsic index $n=4$. The inner disc is prominent in the UV,
with a scale-length of the thin stellar disc, $h_{\rm s}^{\rm i-disc}$, of 600
pc. The distribution of dust in the inner disc is very flat, with a
scalelength of the dust disc, $h_{\rm d}^{\rm disc}$, of 5 kpc. The main disc
resides in the region between $\sim$ 1.8 kpc and $\sim$6.8 kpc. Both the
stellar and the dust distributions of this morphological component are
relatively flat, with a scale-length for the thin stellar disc, $h_{\rm
s}^{\rm t-disc}$ of 4.3 kpc and $h_{\rm d}^{\rm disc}=6$ kpc for the dust
disc.
The main intrinsic properties of M51, as derived from the model are as
follows:
* •
The global star-formation rate is
$\text{SFR}=4.11^{+0.40}_{-0.38}\,$$\text{M}_{\odot}\text{yr}^{-1}$. The inner
disc has
$\text{SFR}=1.38^{+0.15}_{-0.14}\,$$\text{M}_{\odot}\text{yr}^{-1}$and the
main disc has
$\text{SFR}=2.73^{+0.28}_{-0.26}\,$$\text{M}_{\odot}\text{yr}^{-1}$.
* •
The global surface density in SFR is
$\Sigma_{\text{SFR}}=2.83^{+0.28}_{-0.27}\times
10^{-2}\,$$\text{M}_{\odot}\text{yr}^{-1}\text{kpc}^{-2}$.
* •
The specific star formation rate, sSFR, is rather constant over much of the
extent of M51, except for the very centre.
* •
The young stars account for 69% of the dust heating. In the inner regions dust
heating has a significant contribution from old stellar populations, while the
rest of the dust in the disc is mainly heated by young stellar populations,
with their contribution rather constant or slowly increasing towards the outer
disc.
We discuss the properties of the faint and non axi-symmetric emission
extending beyond $R_{\rm l}$, also containing the bridge between M51 and its
companion M51b. We model this outer emission with another morphological
component that we call “outer disc” . The outer disc has $\text{SFR}=0.57\pm
0.06\,$$\text{M}_{\odot}\text{yr}^{-1}$and $\Sigma_{\text{SFR}}=5.11\pm
0.56\times 10^{-4}\,$$\text{M}_{\odot}\text{yr}^{-1}\text{kpc}^{-2}$.
We also show a model with varying dust properties, with the inner disc having
a reduced PAH abundance with respect to the main disc. This hybrid model was
chosen to solve the over-prediction of the standard model at 8 $\mu$m in the
inner disc, and to alleviate a residual bump in the derived intrinsic SED in
the NUV band, at the position of the 2200 Åbump.
## Acknowledgements
We would like to thank an anonymous referee for very useful and constructive
comments, that helped improve the manuscript. This work is based in part on
observations made with the NASA Galaxy Evolution Explorer. GALEX is operated
for NASA by the California Institute of Technology under NASA contract
NAS5-98034. This research has made use of the NASA/IPAC Infrared Science
Archive, which is operated by the Jet Propulsion Laboratory, California
Institute of Technology, under contract with the National Aeronautics and
Space Administration.
This work is also based on the Sloan Digital Sky Survey (SDSS) data. Funding for the SDSS IV has been provided by the Alfred P. Sloan Foundation, the U.S. Department of Energy Office of Science, and the Participating Institutions. SDSS acknowledges support and resources from the Center for High-Performance Computing at the University of Utah. The SDSS web site is www.sdss4.org. SDSS is managed by the Astrophysical Research Consortium for the Participating Institutions of the SDSS Collaboration including the Brazilian Participation Group, the Carnegie Institution for Science, Carnegie Mellon University, Center for Astrophysics | Harvard & Smithsonian (CfA), the Chilean Participation Group, the French Participation Group, Instituto de Astrofísica de Canarias, The Johns Hopkins University, Kavli Institute for the Physics and Mathematics of the Universe (IPMU) / University of Tokyo, the Korean Participation Group, Lawrence Berkeley National Laboratory, Leibniz Institut für Astrophysik Potsdam (AIP), Max-Planck-Institut für Astronomie (MPIA Heidelberg), Max-Planck-Institut für Astrophysik (MPA Garching), Max-Planck-Institut für Extraterrestrische Physik (MPE), National Astronomical Observatories of China, New Mexico State University, New York University, University of Notre Dame, Observatório Nacional / MCTI, The Ohio State University, Pennsylvania State University, Shanghai Astronomical Observatory, United Kingdom Participation Group, Universidad Nacional Autónoma de México, University of Arizona, University of Colorado Boulder, University of Oxford, University of Portsmouth, University of Utah, University of Virginia, University of Washington, University of Wisconsin, Vanderbilt University, and Yale University.
This work has also made use of data products from the Two Micron All Sky
Survey, which is a joint project of the University of Massachusetts and the
Infrared Processing and Analysis Center/California Institute of Technology,
funded by the National Aeronautics and Space Administration and the National
Science Foundation. This work is based in part on observations made with the
Spitzer Space Telescope, which is operated by the Jet Propulsion Laboratory,
California Institute of Technology under a contract with NASA. We also utilise
observations performed with the ESA Herschel Space Observatory (Pilbratt et
al. 2010), in particular to do photometry using the PACS (Poglitsch et al.
2010) and SPIRE (Griffin et al. 2010) instruments.
## Data Availability
The data underlying this article are made available at the CDS database via
http://cdsweb.u-strasbg.fr/cgi-bin/qcat?J/MNRAS/
## References
* Aihara et al. (2011) Aihara H., et al., 2011, ApJS, 193, 29
* Aniano et al. (2011) Aniano G., Draine B. T., Gordon K. D., Sandstrom K., 2011, PASP, 123, 1218–1236
* Baes & Dejonghe (2001) Baes M., Dejonghe H., 2001, MNRAS, 326, 733
* Baes et al. (2010) Baes M., et al., 2010, A&A, 518, L39
* Baes et al. (2011) Baes M., Verstappen J., De Looze I., Fritz J., Saftly W., Vidal Pérez E., Stalevski M., Valcke S., 2011, ApJS, 196, 22
* Bendo et al. (2012) Bendo G. J., et al., 2012, MNRAS, 419, 1833
* Bersier et al. (1994) Bersier D., Blecha A., Golay M., Martinet L., 1994, A&A, 286, 37
* Bianchi (2008) Bianchi S., 2008, A&A, 490, 461
* Bianchi et al. (1996) Bianchi S., Ferrara A., Giovanardi C., 1996, ApJ, 465, 127
* Boulanger et al. (1988) Boulanger F., Beichman C., Desert F. X., Helou G., Perault M., Ryter C., 1988, ApJ, 332, 328
* Brown et al. (2014) Brown M. J. I., et al., 2014, ApJS, 212, 18
* Burbidge & Burbidge (1964) Burbidge E. M., Burbidge G. R., 1964, ApJ, 140, 1445
* Byun et al. (1994) Byun Y. I., Freeman K. C., Kylafis N. D., 1994, ApJ, 432, 114
* Calzetti et al. (2005) Calzetti D., et al., 2005, ApJ, 633, 871
* Camps & Baes (2020) Camps P., Baes M., 2020, Astronomy and Computing, 31, 100381
* Chastenet et al. (2023) Chastenet J., et al., 2023, ApJ, 944, L11
* Condon (1992) Condon J. J., 1992, ARA&A, 30, 575
* Considere & Athanassoula (1982) Considere S., Athanassoula E., 1982, A&A, 111, 28
* Contursi et al. (2000) Contursi A., et al., 2000, A&A, 362, 310
* Cunow (2001) Cunow B., 2001, MNRAS, 323, 130
* Danver (1942) Danver C.-G., 1942, Annals of the Observatory of Lund, 10, 7
* De Geyter et al. (2015) De Geyter G., et al., 2015, MNRAS, 451, 1728
* De Looze et al. (2014) De Looze I., et al., 2014, A&A, 571, A69
* Decleir et al. (2019) Decleir M., et al., 2019, MNRAS, 486, 743
* Disney et al. (1989) Disney M., Davies J., Phillipps S., 1989, MNRAS, 239, 939
* Dobbs et al. (2010) Dobbs C. L., Theis C., Pringle J. E., Bate M. R., 2010, MNRAS, 403, 625
* Draine & Lee (1984) Draine B. T., Lee H. M., 1984, ApJ, 285, 89
* Draine & Li (2007) Draine B. T., Li A., 2007, ApJ, 657, 810
* Durrell et al. (2003) Durrell P. R., Mihos J. C., Feldmeier J. J., Jacoby G. H., Ciardullo R., 2003, ApJ, 582, 170
* Dwek & Werner (1981) Dwek E., Werner M. W., 1981, ApJ, 248, 138
* Eskew et al. (2012) Eskew M., Zaritsky D., Meidt S., 2012, AJ, 143, 139
* Evans (1994) Evans R., 1994, MNRAS, 266, 511
* Fazio et al. (2004) Fazio G. G., et al., 2004, ApJS, 154, 10
* Ferrara et al. (1999) Ferrara A., Bianchi S., Cimatti A., Giovanardi C., 1999, ApJS, 123, 437
* Fischera & Dopita (2004) Fischera J., Dopita M. A., 2004, ApJ, 611, 919
* Fitzpatrick (1999) Fitzpatrick E. L., 1999, PASP, 111, 63
* Gadotti et al. (2010) Gadotti D. A., Baes M., Falony S., 2010, MNRAS, 403, 2053
* Giard et al. (2008) Giard M., Montier L., Pointecouteau E., Simmat E., 2008, AAP, 490, 547
* Gil de Paz et al. (2007) Gil de Paz A., et al., 2007, ApJS, 173, 185
* Greenberg (1963) Greenberg J. M., 1963, ARA&A, 1, 267
* Griffin et al. (2010) Griffin M. J., et al., 2010, A&A, 518, L3
* Gunn et al. (2006) Gunn J. E., et al., 2006, AJ, 131, 2332
* Hagen et al. (2017) Hagen L. M. Z., Siegel M. H., Hoversten E. A., Gronwall C., Immler S., Hagen A., 2017, MNRAS, 466, 4540
* Helou et al. (1985) Helou G., Soifer B. T., Rowan-Robinson M., 1985, ApJ, 298, L7
* Helou et al. (1991) Helou G., Ryter C., Soifer B. T., 1991, ApJ, 376, 505
* Hoversten et al. (2011) Hoversten E. A., et al., 2011, AJ, 141, 205
* Hu et al. (2013) Hu T., Shao Z., Peng Q., 2013, ApJL, 762, L27
* Hutton et al. (2015) Hutton S., Ferreras I., Yershov V., 2015, MNRAS, 452, 1412
* Jarrett et al. (2003) Jarrett T. H., Chester T., Cutri R., Schneider S. E., Huchra J. P., 2003, AJ, 125, 525
* Jones et al. (2013) Jones A. P., Fanciullo L., Köhler M., Verstraete L., Guillet V., Bocchio M., Ysard N., 2013, AAP, 558, A62
* Köhler et al. (2014) Köhler M., Jones A., Ysard N., 2014, AAP, 565, L9
* Kuchinski et al. (1998) Kuchinski L. E., Terndrup D. M., Gordon K. D., Witt A. N., 1998, AJ, 115, 1438
* Kylafis & Bahcall (1987) Kylafis N. D., Bahcall J. N., 1987, ApJ, 317, 637
* Kylafis & Xilouris (2005) Kylafis N., Xilouris E., 2005, in Popescu, CC and Tuffs, RJ ed., AIP CONFERENCE PROCEEDINGS Vol. 761, Spectral Energy Distributions of Gas-Rich Galaxies: Confronting Models with Data. pp 3–16
* Lebouteiller et al. (2007) Lebouteiller V., Brandl B., Bernard-Salas J., Devost D., Houck J. R., 2007, ApJ, 665, 390
* Lee et al. (2011) Lee J. H., Hwang N., Lee M. G., 2011, The Astrophysical Journal, 735, 75
* MacLachlan et al. (2011) MacLachlan J. M., Matthews L. D., Wood K., Gallagher J. S., 2011, ApJ, 741, 6
* Martin et al. (2005) Martin D. C., et al., 2005, ApJl, 619, L1
* McQuinn et al. (2016) McQuinn K. B. W., Skillman E. D., Dolphin A. E., Berg D., Kennicutt R., 2016, ApJ, 826, 21
* Mentuch Cooper et al. (2012) Mentuch Cooper E., et al., 2012, ApJ, 755, 165
* Misiriotis et al. (2001) Misiriotis A., Popescu C. C., Tuffs R., Kylafis N. D., 2001, A&A, 372, 775
* Möllenhoff et al. (2006) Möllenhoff C., Popescu C. C., Tuffs R. J., 2006, A&A, 456, 941
* Monnet et al. (1981) Monnet G., Paturel G., Simien F., 1981, A&A, 102, 119
* Montier & Giard (2004) Montier L. A., Giard M., 2004, AAP, 417, 401
* Morrissey et al. (2007) Morrissey P., et al., 2007, ApJS, 173, 682
* Mosenkov et al. (2016) Mosenkov A. V., et al., 2016, A&A, 592, A71
* Mosenkov et al. (2018) Mosenkov A. V., et al., 2018, A&A, 616, A120
* Natale et al. (2010) Natale G., et al., 2010, ApJ, 725, 955
* Natale et al. (2014) Natale G., Popescu C. C., Tuffs R. J., Semionov D., 2014, MNRAS, 438, 3137
* Natale et al. (2015) Natale G., Popescu C. C., Tuffs R. J., Debattista V. P., Fischera J., Grootes M. W., 2015, MNRAS, 449, 243
* Natale et al. (2017) Natale G., et al., 2017, A&A, 607, A125
* Natale et al. (2021) Natale G., Popescu C. C., Rushton M., Yang R., Thirlwall J. J., Pricopi D., 2021, MNRAS,
* Nersesian et al. (2020a) Nersesian A., et al., 2020a, A&A, 637, A25
* Nersesian et al. (2020b) Nersesian A., et al., 2020b, A&A, 643, A90
* Niederwanger et al. (2019) Niederwanger F., Reimer O., Kissmann R., Strong A. W., Popescu C. C., Tuffs R., 2019, Astroparticle Physics, 107, 1
* Pastrav et al. (2013a) Pastrav B. A., Popescu C. C., Tuffs R. J., Sansom A. E., 2013a, A&A, 553, A80
* Pastrav et al. (2013b) Pastrav B. A., Popescu C. C., Tuffs R. J., Sansom A. E., 2013b, A&A, 557, A137
* Pety et al. (2005) Pety J., Teyssier D., Fossé D., Gerin M., Roueff E., Abergel A., Habart E., Cernicharo J., 2005, AAP, 435, 885
* Pierini et al. (2003) Pierini D., Popescu C. C., Tuffs R. J., Völk H. J., 2003, A&A, 409, 907
* Pierini et al. (2004) Pierini D., Gordon K. D., Witt A. N., Madsen G. J., 2004, ApJ, 617, 1022
* Pilbratt et al. (2010) Pilbratt G. L., et al., 2010, A&A, 518, L1
* Poglitsch et al. (2010) Poglitsch A., et al., 2010, A&A, 518, L2
* Popescu (2021) Popescu C. C., 2021, Star Formation Rates of Galaxies. Cambridge University Press
* Popescu & Tuffs (2002) Popescu C. C., Tuffs R. J., 2002, MNRAS, 335, L41
* Popescu & Tuffs (2010) Popescu C. C., Tuffs R. J., 2010, in Debattista V. P., Popescu C. C., eds, American Institute of Physics Conference Series Vol. 1240, Hunting for the Dark: the Hidden Side of Galaxy Formation. pp 35–46 (arXiv:1006.4606), doi:10.1063/1.3458534
* Popescu et al. (2000) Popescu C. C., Misiriotis A., Kylafis N. D., Tuffs R. J., Fischera J., 2000, A&A, 362, 138
* Popescu et al. (2004) Popescu C. C., Tuffs R. J., Kylafis N. D., Madore B. F., 2004, A&A, 414, 45
* Popescu et al. (2011) Popescu C. C., Tuffs R. J., Dopita M. A., Fischera J., Kylafis N. D., Madore B. F., 2011, A&A, 527, A109
* Popescu et al. (2017) Popescu C. C., Yang R., Tuffs R. J., Natale G., Rushton M., Aharonian F., 2017, MNRAS, 470, 2539
* Rieke et al. (2004) Rieke G. H., et al., 2004, ApJS, 154, 25
* Robitaille et al. (2012) Robitaille T. P., Churchwell E., Benjamin R. A., Whitney B. A., Wood K., Babler B. L., Meade M. R., 2012, A&A, 545, A39
* Rushton et al. (2022) Rushton M. T., Popescu C. C., Inman C., Natale G., Pricopi D., 2022, MNRAS, 514, 113
* Salim et al. (2018) Salim S., Boquien M., Lee J. C., 2018, ApJ, 859, 11
* Salo & Laurikainen (2000) Salo H., Laurikainen E., 2000, MNRAS, 319, 377
* Savchenko et al. (2023) Savchenko S. S., et al., 2023, MNRAS, 524, 4729
* Schlafly & Finkbeiner (2011) Schlafly E. F., Finkbeiner D. P., 2011, ApJ, 737, 103
* Shetty et al. (2007) Shetty R., Vogel S. N., Ostriker E. C., Teuben P. J., 2007, ApJ, 665, 1138
* Siebenmorgen & Krügel (2010) Siebenmorgen R., Krügel E., 2010, A&A, 511, A6
* Skrutskie et al. (2006) Skrutskie M. F., et al., 2006, AJ, 131, 1163
* Steinacker et al. (2013) Steinacker J., Baes M., Gordon K. D., 2013, ARA&A, 51, 63
* Tamburro et al. (2008) Tamburro D., Rix H. W., Walter F., Brinks E., de Blok W. J. G., Kennicutt R. C., Mac Low M. M., 2008, AJ, 136, 2872
* Thilker et al. (2007) Thilker D. A., et al., 2007, ApJS, 173, 572
* Thirlwall et al. (2020) Thirlwall J. J., Popescu C. C., Tuffs R. J., Natale G., Norris M., Rushton M., Grootes M., Carroll B., 2020, MNRAS, 495, 835
* Toomre & Toomre (1972) Toomre A., Toomre J., 1972, ApJ, 178, 623
* Trumpler (1930) Trumpler R. J., 1930, PASP, 42, 214
* Tuffs et al. (2004) Tuffs R. J., Popescu C. C., Völk H. J., Kylafis N. D., Dopita M. A., 2004, A&A, 419, 821
* Tully (1974) Tully R. B., 1974, ApJS, 27, 437
* Turner & Ho (1994) Turner J. L., Ho P. T. P., 1994, ApJ, 421, 122
* Verstocken et al. (2020) Verstocken S., et al., 2020, A&A, 637, A24
* Viaene et al. (2017) Viaene S., et al., 2017, A&A, 599, A64
* Walter et al. (2008) Walter F., Brinks E., de Blok W. J. G., Bigiel F., Kennicutt Robert C. J., Thornley M. D., Leroy A., 2008, AJ, 136, 2563
* Weingartner & Draine (2001) Weingartner J. C., Draine B. T., 2001, ApJ, 548, 296
* Williams et al. (2019) Williams T. G., Baes M., De Looze I., Relaño M., Smith M. W. L., Verstocken S., Viaene S., 2019, MNRAS, 487, 2753
* Witt & Gordon (1996) Witt A. N., Gordon K. D., 1996, ApJ, 463, 681
* Witt & Gordon (2000) Witt A. N., Gordon K. D., 2000, ApJ, 528, 799
* Witt et al. (1992) Witt A. N., Thronson Harley A. J., Capuano John M. J., 1992, ApJ, 393, 611
* Wunderlich et al. (1987) Wunderlich E., Klein U., Wielebinski R., 1987, A&AS, 69, 487
* Xilouris et al. (1997) Xilouris E. M., Kylafis N. D., Papamastorakis J., Paleologou E. V., Haerendel G., 1997, A&A, 325, 135
* Xilouris et al. (1998) Xilouris E. M., Alton P. B., Davies J. I., Kylafis N. D., Papamastorakis J., Trewhella M., 1998, A&A, 331, 894
* Xilouris et al. (1999) Xilouris E. M., Byun Y. I., Kylafis N. D., Paleologou E. V., Papamastorakis J., 1999, A&A, 344, 868
* York et al. (2000) York D. G., et al., 2000, AJ, 120, 1579
* Ysard et al. (2015) Ysard N., Köhler M., Jones A., Miville-Deschênes M. A., Abergel A., Fanciullo L., 2015, AAP, 577, A110
* de Jong et al. (1985) de Jong T., Klein U., Wielebinski R., Wunderlich E., 1985, A&A, 147, L6
* van der Giessen et al. (2022) van der Giessen S. A., Leslie S. K., Groves B., Hodge J. A., Popescu C. C., Sargent M. T., Schinnerer E., Tuffs R. J., 2022, A&A, 662, A26
## Appendix A Amplitude and geometrical parameters
In our model we express the amplitude parameters of the stellar discs with
spectral luminosity densities $L_{\nu}$, and the dust discs with central face-
on optical depths in the B-band, $\tau_{\text{B}}$. Table 6 lists the spectral
luminosity densities for each stellar disc and the bulge. Table 7 lists the
B-band face-on optical depth for each of the morphological components. Table 8
lists other parameters, like the $R_{\text{in}}$, $R_{\text{tin}}$ and
$R_{\text{t}}$, which are fixed from data. The table also lists the $\chi$
values for the main disc. The latter parameter serves to more cohesively join
the inner to the main disc.
Table 6: The intrinsic spectral luminosity densities of the stellar and thin stellar disc, for each morphological component. $\lambda$ | $L_{\nu}^{\text{bulge}}$ | $L_{\nu}^{\text{disc, i}}$ | $L_{\nu}^{\text{disc ,m}}$ | $L_{\nu}^{\text{tdisc, i}}$ | $L_{\nu}^{\text{tdisc, m}}$
---|---|---|---|---|---
(Å) | (W Hz-1) | (W Hz-1) | (W Hz-1) | (W Hz-1) | (W Hz-1)
1542 | - | - | - | 1.07$\times 10^{21}$ | 2.11$\times 10^{21}$
2274 | - | - | - | 3.62$\times 10^{21}$ | 4.97$\times 10^{21}$
3562 | 8.31$\times 10^{19}$ | 7.42$\times 10^{20}$ | 1.48$\times 10^{21}$ | 1.75$\times 10^{21}$ | 3.42$\times 10^{21}$
4719 | 4.16$\times 10^{20}$ | 4.36$\times 10^{21}$ | 8.20$\times 10^{21}$ | 1.87$\times 10^{21}$ | 3.39$\times 10^{21}$
6185 | 6.05$\times 10^{20}$ | 6.26$\times 10^{21}$ | 1.21$\times 10^{22}$ | 1.96$\times 10^{21}$ | 2.95$\times 10^{21}$
7500 | 6.64$\times 10^{20}$ | 6.08$\times 10^{21}$ | 1.24$\times 10^{22}$ | 1.66$\times 10^{21}$ | 2.26$\times 10^{21}$
12000 | 1.13$\times 10^{21}$ | 9.26$\times 10^{21}$ | 2.11$\times 10^{22}$ | 8.17$\times 10^{20}$ | 1.48$\times 10^{21}$
22000 | 1.02$\times 10^{21}$ | 8.94$\times 10^{21}$ | 2.12$\times 10^{22}$ | 6.40$\times 10^{20}$ | 1.16$\times 10^{21}$
35070 | 5.25$\times 10^{20}$ | 5.25$\times 10^{21}$ | 1.23$\times 10^{22}$ | 1.25$\times 10^{20}$ | 5.63$\times 10^{20}$
44370 | 3.73$\times 10^{20}$ | 2.75$\times 10^{21}$ | 6.86$\times 10^{21}$ | 1.25$\times 10^{20}$ | 5.63$\times 10^{20}$
57390 | 4.47$\times 10^{20}$ | 5.74$\times 10^{21}$ | 2.05$\times 10^{22}$ | 1.50$\times 10^{20}$ | 5.63$\times 10^{20}$
Table 7: The face-on optical depth in the B-band at the inner radius of their respective morphological component. | $\tau_{\text{B}}^{\text{f}}\left(R_{\text{in}}\right)$
---|---
Inner | 5.33$\pm 0.33$
Main | 2.62$\pm 0.08$
Fitting parameters |
---|---
$\chi_{\text{s}}^{\text{m-tdisc}}$ | -0.7$\pm 0.5$
$\chi_{\text{s}}^{\text{m-disc}}$ | 1.5$\pm 0.25$
$\chi_{\text{d}}^{\text{m-disc}}$ | 1$\pm 0.25$
Parameters fixed by data |
$R_{\text{in,s}}^{\text{(i-tdisc, m-tdisc)}}$ | (0.7, 2.7)
$R_{\text{in,s}}^{\text{(i-disc, m-disc)}}$ | (0, 2.7)
$R_{\text{in,d}}^{\text{(i-disc, m-disc)}}$ | (0.8, 2.7)
$R_{\text{tin,s}}^{\text{(i-tdisc, m-tdisc)}}$ | (0, 1.8)
$R_{\text{tin,s}}^{\text{(i-disc, m-disc)}}$ | (0, 1.8)
$R_{\text{tin,d}}^{\text{(i-disc, m-disc)}}$ | (0, 1.8)
$R_{\text{t,s}}^{\text{(i-tdisc, m-tdisc)}}$ | (1.9, 6.8)
$R_{\text{t,s}}^{\text{(i-disc, m-disc)}}$ | (1.9, 6.8)
$R_{\text{t,d}}^{\text{(i-disc, m-disc)}}$ | (1.9, 6.8)
Parameters fixed from model |
$z_{\text{s}}^{\text{(i-tdisc, m-tdisc)}}$ | (0.09, 0.09)
$z_{\text{s}}^{\text{(i-disc, m-disc)}}$ | (0.19, 0.19)
$z_{\text{d}}^{\text{(i-disc, m-disc)}}$ | (0.16, 0.16)
Table 8: Other parameters of the model. All length parameters are given in
kpc.
## Appendix B Star formation and dust masses
In this appendix we present tables with the calculated values for SFR (Table.
9), $\Sigma_{\text{SFR}}$ (Table. 10), and dust mass (Table. 11). All these
values are given for both the global emission and for each morphological
component.
Table 9: Star formation rates for the global emission and for the morphological components of M51. Values are in (M⊙ yr-1). | SFR (M⊙yr-1)
---|---
Global | 4.11${}^{+0.41}_{-0.39}$
Inner | 1.38${}^{+0.15}_{-0.14}$
Main | 2.73${}^{+0.28}_{-0.26}$
| $\Sigma_{\text{SFR}}$ (M⊙yr-1kpc-2)
---|---
Global | 2.83${}^{+0.28}_{-0.27}\times 10^{-2}$
Inner | 1.21${}^{+0.12}_{-0.11}\times 10^{-1}$
Main | 2.00${}^{+0.18}_{-0.17}\times 10^{-2}$
Table 10: Surface density of star formation for the global emission and for the morphological components of M51. Values are in (M⊙ yr-1 kpc2). | dust mass M${}_{\text{d}}$ (M⊙)
---|---
Global | 3.74${}^{+0.34}_{-0.30}\times 10^{7}$
Inner | 6.39${}^{+0.37}_{-0.21}\times 10^{6}$
Main | 3.10${}^{+0.26}_{-0.18}\times 10^{7}$
Table 11: Dust masses for the global emission and for the morphological components of M51. Values are in (M⊙). Band | $\text{chi}^{2}_{r}$
---|---
NUV | 0.97
I2 | 2.60
PACS70 | 2.01
SPIRE500 | 0.640
Global | 1.89
Table 12: The $\text{chi}^{2}_{r}$ values at selected wavelengths, as well as
the total over all wavelengths for the hybrid model.
## Appendix C Extinction curves of the Milky Way and LMC dust models
In this paper we use the Milky Way and LMC dust models given in Weingartner &
Draine (2001) in Table 1, line 7 and Table 3, line 1 respectively. Fig. 12
shows the extinction curve for each grain composition as well as the total
extinction curve for the Milky Way and LMC dust models. These plots
demonstrate that the LMC type dust exhibits a reduced 2200 Å bump compared to
the Milky Way type dust. Figs. 13-14 show the fits at various wavelengths
using the hybrid model.
Figure 12: Extinction curves for the Milky Way (left) and the LMC (right) dust
models, with the grain size distribution and optical constants from
Weingartner & Draine (2001). We also plot the contributions from the different
grain compositions: Si (dotparted), Gra (dashed) and PAH (dot-dashed) as well
as the mean extinction curve for the Milky Way (double dot-dashed)
(Fitzpatrick 1999). It can be seen that the LMC dust model features a reduced
2000 bump wh It can be seen that the LMC dust model features a reduced 2200 Å
bump when compared to the Milky Way dust.
Figure 13: Comparison between the azimuthally averaged surface brightness
profiles of the observations (solid black line) at selected UV/optical/NIR
wavelengths and of the corresponding model for dust attenuated stellar light
(solid red line) using the combined dust models of Milky Way and LMC
(Weingartner & Draine 2001). The blue shaded region around the observed
profile represents the corresponding errors in the averaged surface
brightness, as described in Sect. 2. The contribution from each morphological
component is plotted with dashed lines and colour-coded as follows: bulge in
purple, stellar disc in green and thin stellar disc in blue. The lower panels
show the residuals between the observations and our model, with the dashed
blue lines showing the $\pm 20\%$ residuals to guide the eye. Figure 14: Same
as Fig. 13 but for selected wavelengths in the NIR/MIR/FIR/submm, where the
emission from the dust discs is plotted with the grey dotted lines and the
dust emission from the HII regions is shown by the dashed cyan lines.
|
# MeerKLASS L-band deep-field intensity maps: entering the Hi dominated regime
MeerKLASS Collaboration: Matilde Barberi-Squarotti1,2,3,4, José L. Bernal5,
Philip Bull6,7, Stefano Camera1,8,9,7, Isabella P. Carucci10,11, Zhaoting
Chen12, Steven Cunnington6, Brandon N. Engelbrecht7, José Fonseca7,13,14,
Keith Grainge6, Melis O. Irfan15,7, Yichao Li16,7, Aishrila Mazumder6, Sourabh
Paul6, Alkistis Pourtsidou12,17,7, Mario G. Santos7,18, Marta Spinelli19,7,
Jingying Wang20,7, Amadeus Witzemann7,6, Laura Wolz6
1Dipartimento di Fisica, Università degli Studi di Torino, via P. Giuria 1,
10125, Torino, Italy
2Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16,
20133, Milano, Italy
3INFN – Istituto Nazionale di Fisica Nucleare, Sezione di Milano, via G.
Celoria 16, 20133, Milano, Italy
4INAF – Istituto Nazionale di Astrofisica, Osservatorio Astrofisico di Brera-
Merate, via Brera 28 20121 Milano, Italy
5 Instituto de Física de Cantabria (IFCA), CSIC-Univ. de Cantabria, Avda. de
los Castros s/n, E-39005 Santander, Spain
6Jodrell Bank Centre for Astrophysics, Department of Physics & Astronomy, The
University of Manchester, Manchester M13 9PL, UK
7Department of Physics and Astronomy, University of the Western Cape, Robert
Sobukwe Road, Cape Town 7535, South Africa
8INFN – Istituto Nazionale di Fisica Nucleare, Sezione di Torino, via P.
Giuria 1, 10125, Torino, Italy
9INAF – Istituto Nazionale di Astrofisica, Osservatorio Astrofisico di Torino,
10025 Pino Torinese, Italy
10INAF – Istituto Nazionale di Astrofisica, Osservatorio Astronomico di
Trieste, Via G.B. Tiepolo 11, 34131 Trieste, Italy
11IFPU – Institute for Fundamental Physics of the Universe, Via Beirut 2,
34151 Trieste, Italy
12Institute for Astronomy, The University of Edinburgh, Royal Observatory,
Edinburgh EH9 3HJ, UK
13Instituto de Astrofísica e Ciências do Espaço, Universidade do Porto CAUP,
4150-762 Porto, Portugal
14Departamento de Física e Astronomia, Faculdade de Ciências, Universidade do
Porto, Rua do Campo Alegre 687, PT4169-007 Porto, Portugal
15Institute of Astrophysics, University of Cambridge, Madingley Road, CB3 0HA
16Department of Physics, College of Sciences, Northeastern University, Wenhua
Road, Shenyang, 11089, China
17Higgs Centre for Theoretical Physics, School of Physics and Astronomy, The
University of Edinburgh, Edinburgh EH9 3FD, UK
18South African Radio Astronomy Observatory (SARAO), 2 Fir Street, Cape Town,
7925, South Africa
19Observatoire de la Côte d’Azur, Laboratoire Lagrange, Bd de l’Observatoire,
CS 34229, 06304 Nice cedex 4, France
20Shanghai Astronomical Observatory, Chinese Academy of Sciences, 80 Nandan
Road, Shanghai, 200030, China Corresponding
authors:steven.cunnington<EMAIL_ADDRESS>
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
We present results from MeerKAT single-dish Hi intensity maps, the final
observations to be performed in L-band in the MeerKAT Large Area Synoptic
Survey (MeerKLASS) campaign. The observations represent the deepest single-
dish Hi intensity maps to date, produced from 41 repeated scans over
$236\,\deg^{2}$, providing 62 hours of observational data for each of the 64
dishes before flagging. By introducing an iterative self-calibration process,
the estimated thermal noise of the reconstructed maps is limited to
${\sim}\,1.21$ mK ($1.2\,\times$ the theoretical noise level). This thermal
noise will be sub-dominant relative to the Hi fluctuations on large scales
($k\,{\lesssim}\,0.15\,h\,\text{Mpc}^{-1}$), which demands upgrades to power
spectrum analysis techniques, particularly for covariance estimation. In this
work, we present the improved MeerKLASS analysis pipeline, validating it on
both a suite of mock simulations and a small sample of overlapping
spectroscopic galaxies from the Galaxy And Mass Assembly (GAMA) survey.
Despite only overlapping with ${\sim}\,25\%$ of the MeerKLASS deep field, and
a conservative approach to covariance estimation, we still obtain a
${>}\,4\,\sigma$ detection of the cross-power spectrum between the intensity
maps and the 2269 galaxies at the narrow redshift range
$0.39\,{<}\,z\,{<}\,0.46$. We briefly discuss the Hi auto-power spectrum from
this data, the detection of which will be the focus of follow-up work. For the
first time with MeerKAT single-dish intensity maps, we also present evidence
of Hi emission from stacking the maps onto the positions of the GAMA galaxies.
###### keywords:
cosmology: large scale structure of Universe – cosmology: observations – radio
lines: general – methods: data analysis – methods: statistical
††pubyear: 2024††pagerange: MeerKLASS L-band deep-field intensity maps:
entering the Hi dominated regime–D
## 1 Introduction
Probing density fluctuations in large-scale cosmic structure is a cornerstone
of precision cosmology. Constructing maps that span wide areas and reach deep
redshifts provides a rich resource for testing and constraining the standard
$\Lambda$CDM cosmological model. Surveys at optical and near-infrared
wavelengths have made impressive progress in this pursuit (eBOSS
Collaboration, 2021; Heymans et al., 2021; DES Collaboration, 2022; DESI
Collaboration et al., 2024). However, extending experiments to a broader range
of wavelengths boosts the quantity of available data. Furthermore, multi-
wavelength observations provide robust checks of systematic influence due to
the alternative instrumental approach required to survey other wavelengths.
There can also be vastly different complex astrophysics for tracers at
different wavelengths. Thus alternative tracers can validate the consistency
of cosmological conclusions.
Of particular interest for surveying large-scale structure is mapping
redshifted neutral hydrogen (Hi) gas at 21cm radio wavelengths. In the post-
reionisation Universe, at redshifts $z\,{\lesssim}\,6$, the remaining Hi is
overwhelmingly contained within dark matter halos (Villaescusa-Navarro et al.,
2018), meaning the 21cm radiation spontaneously emitted from Hi acts as a
tracer of the large-scale structure. However, at the redshifts of interest for
large-scale structure cosmology ($z\,{\gtrsim}\,0.1$), the 21cm emission
becomes faint and hard to detect with the high signal-to-noise required for a
comprehensive mapping survey. This problem can be circumvented by abandoning
the requirement to resolve individual galaxies and instead integrate the
combined unresolved emission with low-aperture radio telescopes. This process
is now commonly known as Hi intensity mapping (Bharadwaj et al., 2001; Battye
et al., 2004; Wyithe et al., 2008; Chang et al., 2008).
The novel approach of intensity mapping presents some challenges. Since it
records all radiation entering the telescope receiver within its frequency
bandpass, the contribution to the signal will not come from cosmological Hi
alone; in fact, only a small percentage is. There are multiple sources of
foregrounds in the frequency ranges corresponding to redshifted 21cm emission.
These mostly originate from our Galaxy and come from diffuse synchrotron and
free-free emission, but there can also be strong extragalactic point sources
(Mozdzen et al., 2019; Lian et al., 2020; Spinelli et al., 2021; Irfan et al.,
2021; Hurley-Walker et al., 2022; Matshawule et al., 2021). These sources
dominate the cosmological Hi brightness temperature by three to five orders of
magnitude. Furthermore, instrumental effects can distort the foreground
spectra, making their removal additionally challenging (Shaw et al., 2015;
Carucci et al., 2020; Matshawule et al., 2021; Spinelli et al., 2022). In
addition to foregrounds, the telescopes will detect terrestrial, human-made
radio frequency interference (RFI) (Harper & Dickinson, 2018; Engelbrecht et
al., 2024). This can dominate even more than the foregrounds, but vary with
time and sky position so can therefore be flagged and deleted from the final
maps. Residual RFI will also average down if repeated scans are performed over
the same patch of sky, as is the case with MeerKLASS observations.
Despite foreground contamination and RFI challenges, the potential benefits
from Hi intensity mapping are too large to ignore, thus justifying the ongoing
endeavour to develop the technique into a tool for precision cosmology.
Experiments have already delivered promising results, demonstrating how
detections are possible even without purpose-built instruments (Masui et al.,
2013; Wolz et al., 2017; Anderson et al., 2018; Wolz et al., 2022). New
experiments are starting to arrive (e.g. Nan et al., 2011; Battye et al.,
2012; Chen, 2012; Newburgh et al., 2016; PUMA Collaboration et al., 2019;
Vanderlinde et al., 2019; CHIME Collaboration et al., 2022; Pal et al., 2022),
many designed specifically with Hi intensity mapping in mind, creating a
bright future ahead.
The MeerKAT telescope,111www.sarao.ac.za/science/meerkat an array of 64 dishes
in the Upper Karoo region of South Africa, has already showcased its
capability to perform Hi intensity mapping both as an interferometer, claiming
the first detection of the Hi power spectrum (Paul et al., 2023), and also in
single-dish mode, where the field-of-view can reach cosmological scales. Using
the single-dish pilot survey (Wang et al., 2021), a detection of the cross-
correlation power spectrum was achieved with overlapping WiggleZ galaxies
(Cunnington, Li et al., 2022; Drinkwater et al., 2010).
The MeerKAT Large Area Synoptic Survey (MeerKLASS) (Santos et al., 2017) will
continue adding to these observations until it is merged with the SKA
Observatory (SKAO) (SKA Cosmology SWG et al., 2020) on the same
site.222www.skao.int/en MeerKAT can observe in two frequency bands relevant
for 21cm-cosmology: L-band
($900\,\text{MHz}\,{<}\,{\nu}\,{<}\,1420\,\text{MHz}$, reaching redshifts
$z\,{<}\,0.58$) and UHF-band
($580\,\text{MHz}\,{<}\,{\nu}\,{<}\,1000\,\text{MHz}$,
$1.45\,{>}\,z\,{>}\,0.4$).
In this work, we present results from MeerKAT single-dish observations in the
MeerKLASS L-band deep field, a set of observations performed over a single
$236\,\deg^{2}$ patch with 41 repeated scans for each of the 64 dishes.
Integrating over the sky in this way will average down time-varying
contributions from thermal noise and RFI, thus boosting the signal-to-noise of
time-constant sky signals, which includes the foregrounds and the cosmological
Hi. Previous analysis (e.g. Wolz et al., 2022; Cunnington, Li et al., 2022)
exploited the fact that the foreground cleaned intensity maps were thermal
noise dominated when estimating error bars on the power spectrum measurements.
For this MeerKLASS deep-field and future MeerKLASS observations, it can no
longer be assumed that the thermal noise will be dominant, therefore demanding
more emphasis on covariance estimation. We therefore explore using the scatter
in the foreground transfer function (Cunnington et al., 2023), primarily used
to correct for signal loss from foreground cleaning, as a measure of the
covariance in measurements. The transfer function is constructed using mock
signal injection into the real data, hence its scatter should encapsulate
cosmic variance, thermal noise fluctuations, residual RFI contributions,
foreground residuals and signal loss correction uncertainties.
We validate the precision and accuracy of the analysis pipeline to ensure
estimator accuracy is within the statistical error associated with these
MeerKLASS observations. We formalise the main steps in our pipeline, such as
beam reconvolution, foreground cleaning and sampling to a Cartesian grid,
presenting validation steps on a suite of simulations which emulate the
MeerKLASS L-band deep-field observations. We also use a small sample of
overlapping spectroscopic galaxies from the Galaxy And Mass Assembly (GAMA)
(Drinkwater et al., 2010) survey to study the cross-power spectrum and its
covariance under different analysis choices.
The paper is outlined as follows: Section 2 outlines the observations for the
MeerKLASS deep field, our new calibration strategy (with an iterative self-
calibration process) and overlapping GAMA galaxies. Section 3 presents the
methods for generating our suite of mock data which emulate the observations.
Section 4 provides details of our foreground cleaning technique. Section 5
focuses on the power spectra, formalising our estimation and modelling
process, and presents the results. Section 6 shows the results from stacking
the MeerKLASS deep-field maps onto the locations of GAMA galaxies and
discusses the process and complications that come with this. We then finally
conclude in Section 7.
## 2 Observational data
### 2.1 MeerKLASS L-band deep-field
Using the same scan strategy as the MeerKLASS L-band pilot survey in Wang et
al. (2021), a series of observations were conducted using the MeerKAT
telescope between September 1st and December 29th 2021. Observations with
single dishes typically require a scanning strategy where the dishes are
rapidly moved across the sky. The MeerKAT antennas were set to scan in azimuth
at constant elevation to minimise fluctuations of ground spill and airmass. To
make sure the projected speed on the sky is $5\,\mathrm{arcmin\,s^{-1}}$, the
telescope scan speed was set to $5/\cos(el)\,\mathrm{arcmin\,s^{-1}}$ (where
$el$ is the elevation angle) along the azimuth. This speed also ensures that
the telescope pointing does not move significantly compared to the width of
the primary beam (${\sim}\,1\,$deg) during a single time dump
($2\,\mathrm{s}$).
With a time resolution of $2\,\mathrm{s}$, this gives a scan speed of no more
than $10\,\mathrm{arcmin}$ per time sample, which is well within the beam
size. Noise diodes attached to each receiver were fired for
$0.585\,\mathrm{s}$ once every $19.5\,\mathrm{s}$ during the observation to
provide a relative time-ordered data (TOD) calibration reference. The dishes
are moved back and forth with a slew of about 10 deg sky projection in each
direction, corresponding to an observing time of about $120\,\mathrm{s}$ per
stripe. The duration of each set of scans with 48 strips is about 100 minutes,
and before and after each scan we spent ${\sim}\,8$ minutes tracking a nearby
celestial point source to use as a bandpass calibrator and absolute flux
calibrator.
At fixed elevation, two scans can be performed per night, corresponding to
when the field is rising and setting respectively. The two scans will cross
each other to achieve good sky coverage in the region of overlap. There were
41 observations in total (each one referred as a block), the details of each
are listed in the Appendix in Table 2. The observations repeated scans
covering a patch of sky approximately $236\,\deg^{2}$ in the southern sky,
spanning in R.A. $(330^{\circ},360^{\circ})$ and Dec.
$(-36^{\circ},-25^{\circ})$. The 62 hours obtained for each of the 64 dishes
before flagging, provide a cumulative 3968 hours of observations,333This
number is reduced for the final useable data since some observation blocks
were entirely flagged and some dishes failed in observation. We provide
details of this later in this section. likely to be the deepest single-dish
intensity maps, in terms of observational hours per square degree, in the
MeerKLASS campaign.
### 2.2 Standard MeerKLASS calibration
Figure 1: Flowchart showing each step in the calibration pipeline. This is an
extension beyond the Wang et al. (2021) pipeline since it includes the new
self-calibration strategy.
We employ the MeerKAT single-dish calibration pipeline KATcali to calibrate
the deep-field observations and create the combined 3D data cubes. KATcali
includes several cycles of RFI flagging, calibration, and map-making. The
details of the calibration are thoroughly described in Wang et al. (2021).
Here, we provide a brief overview of the main points. We take a two-step
approach to the calibration process. The first step is to use the flux model
of a strong point source during tracking observations to calibrate the
bandpass and absolute flux, as well as the injection power of the noise diode.
The second step then uses the noise diode injections to calibrate the TOD
during the constant-elevation scanning phase.
To calibrate the scanning data, we build the model as
$\displaystyle T_{\rm model}(t,\nu)=T_{\rm diffuse}(t,\nu)+T_{\rm
el}(t,\nu)+T_{\rm diode}(t,\nu)+T_{\rm rec}(t,\nu)\,,$ (1)
where $T_{\rm el}$ is the antenna temperature model of elevation-dependent
terrestrial emission (atmosphere and ground spill), $T_{\rm diode}$ is the
noise diode contribution, $T_{\rm rec}$ is the receiver temperature, while the
diffuse celestial components $T_{\rm diffuse}$ includes Galactic emission and
the cosmic microwave background (CMB),
$\displaystyle T_{\rm diffuse}(t,\nu)=T_{\rm Gal}(t,\nu)+T_{\rm CMB}\,,$ (2)
where the CMB emission is set to be a uniform $T_{\rm CMB}\,{=}\,2.725\,$K
(Fixsen, 2009), and the Galactic emission $T_{\rm Gal}(t,\nu)$ is derived from
PySM (Thorne et al. 2017) that is based on the reprocessed Haslam 408 MHz
total intensity map (Remazeilles et al. 2015).
To compare with the raw data from the correlator, we multiply the signal model
by the gain, $g(t,\nu)$, so that
$\displaystyle\hat{T}_{\rm model}(t,\nu)=g(t,\nu)\,T_{\rm model}(t,\nu)\,,$
(3)
where the hat indicates a temperature in the (arbitrary) correlator units.
The signal and gain models are then fitted (with a Bayesian framework; see
more details in Section 3.3 of Wang et al. 2021) to the TOD for each
polarisation, frequency channel, dish, and observation scan, all of which are
treated independently. Using the gain solution, we can then obtain the
calibrated temperature,
$\displaystyle T_{\rm cal}(t,\nu)\equiv\frac{\hat{T}_{\rm
raw}(t,\nu)}{g(t,\nu)}\,,$ (4)
where ${\hat{T}_{\rm raw}}$ is the raw data collected by MeerKAT. The residual
temperature difference between the signal model and calibrated data is then
$\displaystyle T_{\rm res}(t,\nu)\equiv T_{\rm cal}(t,\nu)-T_{\rm
model}(t,\nu)\,.$ (5)
In standard calibration, we assumed that $T_{\rm res}$ is noise-dominated
during the fitting, ignoring the faint point sources in the target field. This
was a fair assumption when the noise level was relatively high (Wang et al.,
2021). In this work, we seek a better calibration approach by implementing a
self-calibration method.
Figure 2: The standard deviations of $T_{\text{res}}$ for different self-
calibration cycles (Loops 1-5) in the 971–1023 MHz region. The standard
calibration value is also shown for reference (first point).
### 2.3 Self-calibration based on the MeerKLASS sky map
The higher signal-to-noise from the MeerKLASS deep-field data may allow us to
deal more carefully with the point sources, which were ignored in the standard
calibration, to achieve a better calibration result. Additionally, we aim to
use a more accurate sky model to replace the $T_{\rm Gal}$ model that is based
on the Haslam map. The 408 MHz radio continuum all-sky map, which has been
widely used to study diffuse synchrotron radiation from our Galaxy and works
as a Galactic foreground model for cosmological research, combines data from
four different surveys that happened more than 40 years ago (Haslam et al.
1981; Haslam et al. 1982). The instrument performance in the 1980s limited the
data quality, especially for the Galactic substructures and the zero level of
the maps (Remazeilles et al., 2015). The selection of the spectral index map,
used to generate Galactic maps at MeerKAT frequencies, is also a matter of
debate (Spinelli et al., 2022).
We therefore extend the calibration process described in Section 2.2 and test
the behaviour of the results by introducing an iterative self-calibration
process with several loops. This includes
(1) replacing $T_{\rm diffuse}(t,\nu)$ with $T_{\rm sky}(t,\nu)$, obtained
from the previous Level-6 $T_{\rm sky}$ result, when the corresponding Level-6
pixel has a hit-count$\,{>}\,40$ (which ensures that $T_{\rm sky}$ values are
obtained from high signal-to-noise data);
(2) setting the time-order weight $w({t_{i}})\,{=}\,1$ for all the scan time
samples; instead of setting a higher weight for the diode (main calibrator
during the standard calibration) injection samples (see more details in
Section 3.3 of Wang et al. 2021). Thus Equation 1 becomes
$T^{\rm upd}_{\rm model}(t,\nu)=T_{\rm sky}(t,\nu)+T_{\rm el}(t,\nu)+T_{\rm
diode}(t,\nu)+T_{\rm rec}(t,\nu)\,.$ (6)
In this way, we have all components (especially point sources and Galactic
substructures) included in the input “model”. The difference between the data
and our calibrated sky data is
$T^{\rm upd}_{\rm res}(t,\nu)\equiv T_{\rm cal}(t,\nu)-T^{\rm upd}_{\rm
model}(t,\nu)\,,$ (7)
which should be noise-dominated compared to that in Equation 5. The self-
calibration process is iterative for five loops (see Figure 1), in which the
$T_{\rm sky}(t,\nu)$ always uses the Level-6 result of the previous
calibration loop. Each iteration refines the calibration, reducing the
discrepancies between the observed data and the model. We stopped the self-
calibration iteration after five loops because the standard deviation of
$T_{\rm res}$ converged to a stable value, as shown in Figure 2.
Figure 3: Number count of good (flagged time $<50\%$) channels within the
971–1075 MHz band (L-band channels 550–1050) for each dish ($x$-axis) and
observation block ($y$-axis). We only show the 27 observation blocks used, out
of the possible 41. The other blocks were entirely removed mainly due to RFI
contamination, and one failed observation run. The details of all 41 blocks
are listed in Table 2. Figure 4: MeerKLASS L-band deep-field map at
$1023\,$MHz (middle panel) in comparison with the Galactic emission (top
panel) derived from PySM. Blue circles mark the positions of known point
sources with flux ${>}\,1$ Jy at 1.4 GHz. The bottom panel shows the number of
observational recordings i.e. timestamps in each pixel at 1023 MHz, from the
surviving 27 blocks. Each timestamp is 2 seconds and we use this for noise
estimation and weighting in the power spectrum analysis. Blue scatter points
are the positions of all overlapping GAMA galaxies. Figure 5: Amplitude of
each map pixel along the line-of-sight for the previous MeerKLASS pilot survey
(top panel) and the deep field used in this work (bottom panel. We only show
the ${\sim}\,250$ channels most clean of RFI used in later cosmological
analysis. This has the frequency range
$971.2\,{<}\,\nu\,{<}\,1023.6\,\text{MHz}$ corresponding to
$0.46\,{>}\,z\,{>}\,0.39$ for redshifted 21cm emission.
### 2.4 Characteristics of final MeerKLASS data cubes
#### 2.4.1 Final data cubes
The combined calibrated temperature cubes (with a pixel size of 0.3 deg and
frequency resolution of 0.2 MHz) are derived from all available scans and
dishes. However, predominantly due to RFI, an identified repeated source of
which we discuss in the following sub-section, only 27 of the 41 observation
blocks yielded well-calibrated results. We mark them in the last column of
Table 2. In Figure 3 we present the number count of good (flagged time
$<50\%$) channels within the 971–1075 MHz band (channels 550–1050) for each
block and each dish. In particular, we found that antenna m059 has a higher
chance of surviving calibration even when all other dishes fail.444For
consistency with other MeerKAT work, we indicate MeerKAT’s 64 antennae with
this notation spanning m000-m063. This may be because m059 is the easternmost
dish of MeerKAT, located far from the array core, providing it with a more
isolated environment.
We show the final sky temperature map from the combination of the remaining 27
scans after 5 loops of self-calibration at 1023 MHz in the middle panel of
Figure 4. For comparison, the top panel shows the PySM diffuse Galactic
emission model with the same sky coverage. We also show in the bottom panel
the total number of time samples (across all dishes and observations) that
were combined in each pixel.
To demonstrate the improvement in data quality from the deep-field intensity
maps, Figure 5 plots the amplitude of each pixel along the line-of-sight as a
function of frequency. We compare this to the previous MeerKLASS L-band pilot
survey (top-panel) relative to the deep field used for this work. The maps
should be dominated by the continuum synchrotron spectra, thus their
smoothness is a rough proxy for how dominant any systematics are. There are
numerous perturbations in the pilot survey spectra that have consequences for
the effectiveness of foreground cleaning, required to access the cosmological
Hi. However, for the deep field, the spectra appear smoother and well-
calibrated. In addition to the improvements shown in Figure 2, we also find a
marginal performance improvement on the line-of-sight rms when using the self-
calibration technique relative to the previous standard calibration in the
deep-field. The average variance along the line-of-sight of each pixel
decreases to 977.2 mK2 for the self-calibration, down from 984.3 mK2 for the
standard calibration.
#### 2.4.2 RFI contaminations
From the 41 observation blocks, we had to remove 14 blocks due to excessive,
low-level RFI contaminations. In affected blocks, these contaminations were
sometimes recognisable on the map level of individual dish data as horizontal
stripes. After careful investigations, we have now identified the source of
the RFI contamination as the downlink of the mobile communication tower
(${\sim}\,54\,\text{km}$ away) that enters through the sidelobes and causes a
non-linear gain effect in the entire band. Our aggressive flagging approach in
this study, with the entire removal of 14 blocks, was to ensure high data
quality for our analysis. We are currently developing a method to correct for
this effect, which should allow us to use the currently flagged data for
future projects.
#### 2.4.3 Noise level estimation
Here we estimate the noise levels from the final maps to check if they are
consistent with theoretical expectations. We use the final data cube, $T_{\rm
sky}$, to estimate the noise level using the difference between four
neighbouring channels (the so-called ABBA technique). By adopting the ABBA
pattern, this difference statistic removes both constant and linearly varying
signals to cancel signals detected by the receiver, leaving only the noise.
The noise level is estimated by
$\displaystyle\Delta T_{i}(\nu^{\star})=\frac{1}{2}\left(T^{i}_{\rm
sky}(\nu)+T^{i}_{\rm sky}(\nu+\delta\nu)\right)$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}-\frac{1}{2}\left(T^{i}_{\rm
sky}(\nu-\delta\nu)+T^{i}_{\rm sky}(\nu+2\delta\nu)\right),$ (8)
where the index $i$ goes over all pixels in the map and ⋆ means the combined
result from the four channels
$[\nu{-}\delta\nu,\,\nu,\,\nu{+}\delta\nu,\,\nu{+}2\delta\nu]$. From the
observation data, we take a weighted rms,
$\displaystyle\Delta T_{\rm
RMS}^{2}(\nu^{\star})=\frac{N_{\text{pix}}}{N_{\text{pix}}-1}\left(\frac{\sum_{i}w_{i}\Delta
T^{2}_{i}}{\sum_{i}w_{i}}-\left[\frac{\sum_{i}w_{i}\Delta
T_{i}}{\sum_{i}{w_{i}}}\right]^{2}\right),$ (9)
where the sum is over the number of pixels in the map, $N_{\text{pix}}$, and
we have suppressed the dependence of $w_{i}$ and $\Delta T_{i}$ on
$\nu^{\star}$. For the weight, $w_{i}$, we used the theoretical expected
variance itself, $w_{i}(\nu^{\star})=1/\sigma^{2}_{{\rm th},i}(\nu^{\star})$.
The corresponding variance in each pixel would be,
$\displaystyle\sigma^{2}_{{\rm
th},i}(\nu^{\star})=\frac{1}{4}\Big{(}\sigma^{2}_{{\rm
th},i}(\nu)+\sigma^{2}_{{\rm th},i}(\nu{+}\delta\nu)+\sigma^{2}_{{\rm
th},i}(\nu{-}\delta\nu)$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}+\sigma^{2}_{{\rm
th},i}(\nu{+}2\delta\nu)\Big{)}.$ (10)
Here the noise in the initial polarised TOD is given by the radiometer
equation
$\sigma^{2}_{\rm th,pol}(t,\nu)=\frac{T^{2}_{\rm
sys}(t,\nu)}{\delta\nu\,\delta t},$ (11)
where $\delta\nu\,{=}\,0.209\,\text{MHz}$ is the frequency width of the
MeerKAT L-band channels and $\delta t\,{=}\,2$ s is the length of each time
stamp recording. For the system temperature we use the calibrated data itself,
$T_{\rm sys}\,{=}\,T_{\rm cal,HH}(t,\nu)$ or $T_{\rm cal,VV}(t,\nu)$ for
horizontal and vertical polarisation data, respectively. Thus the noise in the
initial intensity TOD is
$\sigma^{2}_{\rm th}(t,\nu)=\frac{T^{2}_{\rm cal,HH}(t,\nu)+T^{2}_{\rm
cal,VV}(t,\nu)}{4\,\delta\nu\,\delta t}.$ (12)
In order to get the final data cube, $T^{i}_{\rm sky}(\nu)$, the data goes
through several stages of averaging. We propagate the variance taking into
account this averaging in order to get to the final $\sigma^{2}_{{\rm
th},i}(\nu)$.
The theoretical rms which can be compared to $\Delta T_{\rm RMS}(\nu^{\star})$
in Equation 9, is then simply
$\displaystyle\sigma^{2}_{\rm
th}(\nu^{\star})=N_{\text{pix}}\left(\sum_{i}\sigma^{-2}_{{\rm
th},i}(\nu^{\star})\right)^{-1}.$ (13)
More details can be found in Section 5.6 of Wang et al. (2021).
Figure 6: Comparison between noise levels derived from calibrated
observational data ($\Delta T_{\rm RMS}(\nu^{\star})$) and from theory
equation ($\sigma_{\rm th}(\nu^{\star})$) in 971–1075 MHz (upper panel), a
zoom-in for above in 971–1023 MHz (middle panel), and the histogram of the
ratio $R_{\rm RMS}=\Delta T_{\rm RMS}(\nu^{\star})/\sigma_{\rm
th}(\nu^{\star})$ in 971–1023 MHz (lower panel).
In the upper panel of Figure 6, we show the noise levels (from data and
theory) in the 971–1075 MHz frequency band (channels 550–1050). Note here both
the real and theory values are related to the real $T_{\rm sys}$ and hit
counts in corresponding map pixels. The count reduction due to RFI flagging,
and unflagged low-level RFIs that are hiding in the data would cause wiggles.
Four groups of spikes are present at higher frequencies, while below 1023 MHz,
there is much less scatter compared to the previous MeerKLASS pilot survey
result. We therefore chose to restrain our cosmological analysis to the
971–1023 MHz region, for which we show the noise level in the middle panel,
finding a median of $\Delta T_{\rm RMS}\,{=}\,1.21\,$mK. In the lower panel of
Figure 6, we also show the histogram of the ratio $R_{\rm RMS}\,{=}\,\Delta
T_{\rm RMS}(\nu^{\star})/\sigma_{\rm th}(\nu^{\star})$ in the 971–1023 MHz
band. We find a median value of $R_{\rm RMS}\,{=}\,1.2$, while some channels
are noisy and $R_{\rm RMS}$ can reach $1.5$.
### 2.5 GAMA spectroscopic galaxies
Overlapping with the MeerKLASS deep field, is a catalogue of spectroscopic
galaxies recorded in the Galaxy and Mass Assembly (GAMA) survey. GAMA began
observing in February 2008 (Driver et al., 2011; Liske et al., 2015). The
survey aimed to construct catalogues of galaxies to test the cold dark matter
paradigm and probe galaxy evolution (Driver et al., 2009). For our purposes,
we exploit the fact that these galaxies will trace dark matter and should
therefore correlate with the overlapping MeerKLASS intensity maps after
foreground cleaning.
We use the fourth and final data release from the Galaxy and Mass Assembly
survey, GAMA DR4 (Driver et al., 2022). This includes galaxies from the GAMA
II survey phase which enlarged the GAMA I equatorial survey, by including
extra survey regions. One of these extra regions was the 23 hr (G23) field,
which completely overlaps in area with the MeerKLASS L-band deep field, shown
by the blue scatter points in the bottom panel of Figure 4, which represent
the GAMA G23 galaxies relative to the observational time footprint map for the
L-band deep field. The G23 field covered
$339\,\deg\,{<}\,\text{R.A.}\,{<}\,351\,\deg$ and
${-}35\,\deg\,{<}\,\text{Dec.}\,{<}\,{-}30\,\deg$. The magnitude limit of the
field was reduced from the planned 19.8 in $r$-band to 19.2 in $i$-band due to
limited observational time, but this allowed the GAMA II survey to obtain a
high $94.2\%$ redshift completion in the G23 region (Liske et al., 2015).
The GAMA DR2 only provides random catalogues that emulate the survey selection
function of the real galaxies for the G09, G12, and G15 regions, (not the G23
used in this work). Ideally, these would be used to construct a survey
completion window function, which would be used in the cross-power spectrum
estimation. In the absence of this, we assume a uniform survey completion
across the G23 area and a constant distribution with redshift, given the very
flat distribution for the redshift range used in this analysis (green region
of Figure 7). Given our signal-to-noise levels, survey masks will be a sub-
dominant effect in the cross-power, thus this is an adequate approximation to
make.
Due to our aggressive flagging strategy on the MeerKLASS L-band data, the
overlap region in redshift is restricted to $0.39\,{<}\,z\,{<}\,0.46$. As
shown by Figure 7, this catches the GAMA survey galaxy redshift distribution,
$N_{\text{g}}(z)$, near the end of its optimal range, but still provides 2269
galaxies to cross-correlate with. At this redshift range, the comoving
(volumetric) number density of the GAMA galaxies is
$\bar{n}_{\text{g}}\,{=}\,4.8\,{\times}\,10^{-4}\,h^{3}\,\text{Mpc}^{-3}$.
Figure 7: Redshift distribution of GAMA G23 galaxies in the 23hr field. The
accessible redshifted 21cm emission for the full L-band range is shown by the
light-pink shading. The region used in this work between
$0.39\,{<}\,z\,{<}\,0.46$ ($1023.6\,{<}\,\nu\,{<}\,971.2\,\text{MHz}$) is
shown by the darker green shading, in which 2269 GAMA galaxies exist.
## 3 Simulation of mock data
To accompany the calibrated observations outlined in Section 2, we produce a
suite of 500 mock simulations, used for validating the analysis pipeline,
calculating signal loss from foreground cleaning and estimating covariance in
the power spectrum measurements. To create this large number of mocks we
generate lognormal density fields (Coles & Jones, 1991) from which a map of Hi
brightness temperatures can be cut, and galaxies Poisson sampled onto for the
spectroscopic galaxy counterpart.
An important stage of the analysis pipeline is regridding the intensity
mapping voxels with celestial sky coordinates (R.A., Dec., $\nu$) onto a
regular 3D Cartesian grid with lengths $l_{\text{x}}$, $l_{\text{y}}$,
$l_{\text{z}}$ in $h^{-1}\text{Mpc}$, on which a Fast Fourier Transform (FFT)
can be applied for the power spectrum estimation (Cunnington & Wolz, 2024). We
therefore require the mocks to begin in celestial coordinates so we can
validate the regridding process. To achieve this, we begin by fitting a
Cartesian cube around the full MeerKLASS deep-field footprint, creating a grid
with dimensions
$l_{\text{x}},l_{\text{y}},l_{\text{z}}\,{=}\,774.9,335.7,252.2\,h^{-1}\text{Mpc}$
over $n_{\text{cell}}\,{=}\,512^{3}$ cells. We use a large number of cells for
this initial grid to approximately emulate a continuous sky, aiming to avoid
the simulated maps resolving the finite input grid structure. Onto this grid,
we simulate a lognormal density field. This is done by interpolating
$\boldsymbol{k}$-modes from a model power spectrum onto the FFT grid. The Hi
model power spectrum is given by
$P_{\textrm{H{i}}}(\boldsymbol{k})\equiv
P_{\textrm{H{i}}}(k,\mu)=\frac{\overline{T}_{\textrm{H{i}}}^{2}\,b_{\textrm{H{i}}}^{2}\,\left(1+f\,\mu^{2}/b_{\textrm{H{i}}}\right)^{2}\,P_{\text{m}}(k)}{1+\left(k\,\mu\,\sigma_{\text{v}}/H_{0}\right)^{2}}\,.$
(14)
The anisotropy in this model makes it convenient to parameterise
$\boldsymbol{k}$ using $\mu$, defined as the cosine of the angle between the
line-of-sight and $\boldsymbol{k}$. The Hi parameters are the mean Hi gas
temperature ($\overline{T}_{\textrm{H{i}}}$) and the Hi bias
($b_{\textrm{H{i}}}$). $f$ is the growth rate of structure and is included to
model linear redshift space distortions (RSD) which enhance clustering across
all scales. The denominator is included to approximately model the non-linear
RSD; the so-called Fingers-of-God (FoG), which suppresses clustering at small
scales. The FoG are modulated by $\sigma_{\text{v}}$, which is the velocity
dispersion of the tracer, in this case, the Hi gas. All of these quantities
have a redshift dependence, which we omit for brevity; they will also have
little variation given our narrow redshift range. Lastly, $H_{0}$ is the
Hubble parameter at $z\,{=}\,0$, known as the Hubble constant. An alternative
process for generating brightness temperatures is pursued for our stacking
analysis discussed later in the paper. More realistic lognormal line intensity
mapping mocks are available. For example, in future work, we intend to adopt
the approach in Niemeyer et al. (2023) where luminosities from the spectral
lines are painted given a luminosity function. This has been shown to
reproduce better RSD for galaxy distributions and models the whole temperature
distribution expected in the maps.
Inverse Fourier transforming the model in Equation 14 provides the Hi
correlation function $\xi_{\textrm{H{i}}}$ which is then log-transformed via
$\xi^{\prime}_{\textrm{H{i}}}\,{=}\,\ln(1\,{+}\,\xi_{\textrm{H{i}}})$. Fourier
transforming $\xi^{\prime}_{\textrm{H{i}}}$ back gives the modified lognormal
power spectrum $P^{\prime}_{\textrm{H{i}}}$. Gaussian random numbers are then
sampled into each cell of the Fourier grid $\delta_{\text{G}}(\boldsymbol{k})$
with variance $\sigma^{2}(\boldsymbol{k})\propto
P^{\prime}_{\textrm{H{i}}}(\boldsymbol{k})$. This Gaussian random field is
then inverse Fourier transformed to real space giving
$\delta_{\text{G}}(\boldsymbol{x})$ from which the lognormal Hi brightness
temperature fluctuation field is given by
$\delta
T_{\textrm{H{i}}}(\boldsymbol{x})=\exp\left[\delta_{\text{G}}(\boldsymbol{x})-\frac{\sigma^{2}_{\text{G}}}{2}\right]-1\,,$
(15)
where $\sigma_{\text{G}}^{2}$ is the measured variance of the Gaussian random
number field $\delta_{\text{G}}$. Further details on this process can be found
in the Appendix A of Beutler et al. (2011). For the galaxy over-density field
$\delta_{\text{g}}(\boldsymbol{x})$, we follow the same procedure except the
input model is for the galaxy power spectrum, given by
$P_{\text{g}}(\boldsymbol{k})\equiv
P_{\text{g}}(k,\mu)=\frac{b_{\text{g}}^{2}\,\left(1+f\,\mu^{2}/b_{\text{g}}\right)^{2}\,P_{\text{m}}(k)}{1+\left(k\,\mu\,\sigma_{\text{v}}/H_{0}\right)^{2}}\,,$
(16)
where $b_{\text{g}}$ is the bias for the galaxies. We assume that the velocity
dispersion $\sigma_{\text{v}}$ for the Hi gas and galaxies is the same for
simplicity. This will be sufficient for this work where signal-to-noise is low
at the small scales where $\sigma_{\text{v}}$ has the main effect.
The galaxy over-density $\delta_{\text{g}}$ is unit-less whereas the intensity
maps $\delta T_{\textrm{H{i}}}$ are brightness temperature fluctuations
measured in mK. We use the same seed in the random number generation to ensure
that both the $\delta T_{\textrm{H{i}}}$ and $\delta_{\text{g}}$ correlate.
The lognormal code we adopt utilises some functions from the publically
available package powerbox (Murray, 2018).555github.com/steven-murray/powerbox
### 3.1 Transforming between Cartesian grids and sky maps
As already mentioned, the galaxies and the intensity mapping voxels from
observations are in celestial coordinates at a certain frequency (or redshift)
i.e.
$(\boldsymbol{\theta},\nu)\,{\equiv}\,(\theta_{\text{R.A.}},\theta_{\text{Dec.}},\nu)$.
However, for the final power spectrum estimation, we require the fields be in
a Cartesian space with dimensions ($l_{\text{x}},l_{\text{y}},l_{\text{z}}$)
in $h^{-1}\text{Mpc}$. For this we follow the recent work in Cunnington & Wolz
(2024) which we discuss further in Section 5.
Creating lognormal mocks in celestial coordinates that emulate observations
requires the inverse of this step i.e. $\delta
T_{\textrm{H{i}}}(\boldsymbol{x})\,{\rightarrow}\,\delta
T_{\textrm{H{i}}}(\boldsymbol{\theta},\nu)$, since the lognormal fields are
generated onto Cartesian grids measured in $h^{-1}\text{Mpc}$. Transforming
mock galaxy coordinates (the Poisson sampling of which is discussed in Section
3.3) from
$(l_{\text{x}},l_{\text{y}},l_{\text{z}})\,[h^{-1}\text{Mpc}]\,{\rightarrow}\,(\theta_{\text{R.A.}},\theta_{\text{Dec.}},\nu)$
is trivial. We can use astropy666docs.astropy.org (Astropy Collaboration et
al., 2022) transformation routines on each galaxy’s coordinate. However,
transforming the tessellating field of Cartesian cells, each with an
associated brightness temperature, into a gap-less map of voxels with
identical celestial coordinates and frequency binning as the real MeerKAT
maps, requires a resampling of the field. For this we also utilise the
techniques and public code777github.com/stevecunnington/gridimp outlined
Cunnington & Wolz (2024).
This process involves a Monte-Carlo particle sampling technique where each
Cartesian cell is filled with $N_{\text{p}}$ particles, randomly distributed
within the cell. The particles are assigned the brightness temperature of the
cell they fall within. The particles are then transformed to celestial
coordinates ($\theta_{\text{R.A.}},\theta_{\text{Dec.}},z$) and added into the
map voxels associated with their celestial coordinate and the frequency
channel corresponding to their redshift. Each voxel is normalised by the
number of particles falling within it. Providing enough sampling particles per
cell are used (in this work we use $N_{\text{p}}\,{=}\,10$) and the Cartesian
grid is of sufficient resolution (hence our choice of
$n_{\text{cell}}\,{=}\,512^{3}$), it will provide an intensity map in
celestial sky coordinates, that traces the fluctuations generated by the
lognormal field.
Throughout this work, we do not require the necessary extensions of higher-
order particle assignment or field interlacing to suppress the effects of
aliasing (as trialled in Cunnington & Wolz, 2024), since these effects are
below the current statistical noise from this data. We are however, still
required to implement a correction to the power spectrum estimation from the
Nearest-Grid-Point assignment which smooths the field, which we formalise in
Section 5.
### 3.2 Hi intensity map mocks
Using the regridding process described in the previous section we are provided
with Hi intensity mapping mocks $\delta
T_{\textrm{H{i}}}({\boldsymbol{\theta},\nu})$, with identical coordinates and
dimensions as the observed MeerKAT intensity maps. From here we can add in
observational effects and simulate various systematics. The most dominant
effect comes from foreground contamination. Systematics will have a large
influence on foreground contamination, and this can be extremely complex.
Therefore, this is not something we attempt to directly simulate in this work
since we risk under-estimating the problem. Instead, we use the observational
data itself which has the relevant systematics contained within, albeit only
one realisation of them. However, adding these directly onto different
realisations of Hi mocks, still recreates stochasticity in the Hi signal’s
response to systematics. We discuss this further in Section 4.
Another important addition to the mocks will be a smoothing to the intensity
maps corresponding to the angular pixelisation. This is necessary to emulate
the radio telescope’s beam, which for single-dish intensity mapping is quite
broad giving a low angular resolution to the maps and damping small-scale
fluctuations perpendicular to the line-of-sight. In this work, we assume the
MeerKAT beam profile to be Gaussian, which is a simplification but still valid
given the signal-to-noise at small scales where the beam suppresses signal
below the noise. We will investigate implementing a more realistic MeerKAT
beam profile in future work (see e.g. Asad et al., 2021; Matshawule et al.,
2021; Spinelli et al., 2022). The assumed Gaussian beam profile has a
frequency ($\nu$) dependent standard deviation in radians given by
$\sigma_{\text{beam}}(\nu)=\frac{1}{2\sqrt{2\ln 2}}\frac{c}{\nu
D_{\text{dish}}}\,,$ (17)
where we use $D_{\text{dish}}\,{=}\,13.5\,\text{m}$ for the diameter of the
MeerKAT dishes and $c$ is the speed of light. To emulate the beam effects we
smooth the mock maps by creating a 2D Gaussian array kernel centred on the
median angular pixel $\theta_{0}$ given by
$\pazocal{B}_{\text{beam}}(\boldsymbol{\theta},\nu_{i})=\exp\left[-\frac{1}{2}\left(\frac{\boldsymbol{\theta}-\theta_{0}}{\sigma_{\text{beam}}(\nu_{i})}\right)^{2}\right]\,.$
(18)
This is normalised such that
$\sum_{\boldsymbol{\theta}}\pazocal{B}_{\text{beam}}\,{=}\,1$, so that
convolution with this beam kernel does not change the total brightness
temperatures within the maps. The Hi intensity map mocks are convolved with
Equation 18 at each frequency $\nu_{i}$ to give the beam-smoothed Hi map;
$\delta T_{\circledast\textrm{H{i}}}(\boldsymbol{\theta},\nu_{i})=\delta
T_{\textrm{H{i}}}(\boldsymbol{\theta},\nu_{i})\circledast\pazocal{B}_{\text{beam}}(\boldsymbol{\theta},\nu_{i})\,,$
(19)
where the $\circledast$ symbol represents a convolution.
To realistically emulate additional components such as foregrounds, noise and
their response to systematics, we add the real observations to each mock
iteration. This is to avoid the requirement of simulating the complex and
poorly understood systematics.
In some cases, mostly for studying the Hi auto-correlation, it will be useful
to simulate just the thermal noise originating from the thermal signature of
the receivers. We do this by generating Gaussian random noise the power of
which can then be measured with the same pipeline as the rest of the analysis,
providing a robust estimate of the thermal noise floor. We begin with an
analytical description of the thermal noise fluctuations per pixel using the
radiometer equation (as in Section 2 but now at map-level)
$\sigma_{\text{N}}(\boldsymbol{\theta},\nu)=\frac{T_{\text{sys}}(\nu)}{\sqrt{2\,\delta\nu\,t_{\text{p}}(\boldsymbol{\theta},\nu)}}\,,$
(20)
where $t_{\text{p}}$ is the summed observation time contained in each voxel,
which can be taken directly from the time stamp counts as shown in Figure 4,
with each count representing $\delta t\,{=}\,2\,$sec. This emulates the real
data by creating a non-uniform noise map, generating higher noise around the
edges where scan coverage is lower. For the system temperature, we follow
Cunnington (2022) and use a model tuned to coincide with the estimated system
temperature of the MeerKAT telescope (Wang et al., 2021)888See
skaafrica.atlassian.net/wiki/x/AYCHE, assuming
$T_{\mathrm{sys}}(\nu)=T_{\mathrm{rx}}(\nu)+T_{\mathrm{spl}}+T_{\mathrm{CMB}}+T_{\mathrm{gal}}(\nu)\,,$
(21)
where we assume the contribution from our own galaxy is
$T_{\mathrm{gal}}\,{=}\,25\text{K}\,(408\mathrm{MHz}/\nu)^{2.75}$, the CMB
temperature is $T_{\text{CMB}}\,{=}\,2.725\,\text{K}$, as in Section 2. The
spill-over temperature is assumed as $T_{\text{spl}}\,{=}\,3\,\text{K}$ and
the receiver temperature we model as
$T_{\mathrm{rx}}(\nu)=7.5\mathrm{K}+10\mathrm{K}\left(\frac{\nu}{\mathrm{GHz}}-0.75\right)^{2}\,.$
(22)
Due to the narrow frequency band we use, this gives a near-constant system
temperature with the range
$15.97\,{<}\,T_{\text{sys}}\,{<}\,16.02\,[\text{K}]$. Gaussian random numbers
can then be generated across the MeerKLASS deep-field map with the voxel
dependent rms of $\sigma_{\text{N}}(\boldsymbol{\theta},\nu)$.
### 3.3 Galaxy mock catalogues
To accompany the mock intensity maps, we also require corresponding galaxy
mocks with the same correlated clustering. Rather than transforming the
Cartesian galaxy over-density field $\delta_{\text{g}}(\boldsymbol{x})$ into
sky coordinates, it is more realistic to instead Poisson sample the field to
generate a catalogue of galaxies with the same number density as the GAMA G23
survey. This also allows additional validity checks such as testing the
stacking pipeline which cannot be done without a catalogue of galaxy
coordinates.
We generate the galaxy catalogue coordinates initially in the Cartesian domain
to exploit the high $n_{\text{cell}}\,{=}\,512^{3}$ resolution. We first
calculate the average galaxy counts in cells $\langle{n}_{\text{g}}\rangle$
for the real GAMA G23 region in the redshift range where there is overlap with
the intensity map frequency range. We then take the galaxy survey selection
window function, which for the GAMA G23 region, we have to assume is constant
between $339\,{<}\,\text{R.A.}\,{<}\,351\,\deg$ and
${-}35\,{<}\,\text{Dec.}\,{<}\,{-}30\,\deg$ (see discussion in Section 2.5).
We then transform this window function into the high resolution Cartesian
space. A galaxy number counts field $n_{\text{g}}(\boldsymbol{x})$ is then
given by
$n_{\text{g}}(\boldsymbol{x})=\langle
n_{\text{g}}\rangle\left(\delta_{\text{g}}(\boldsymbol{x})+1\right)W_{\text{g}}(\boldsymbol{x})\,,$
(23)
where we use the transformed window function $W_{\text{g}}(\boldsymbol{x})$
and the real GAMA average counts $\langle n_{\text{g}}\rangle$ to ensure the
mock galaxy counts will share the same survey selection as the real galaxy
survey, with the same number density. In this process we normalise the window
function $W_{\text{g}}$ such that its maximum value is $1$, so for the GAMA
G23 survey, this becomes a binary mask of 0 and 1. We then Poisson sample the
$n_{\text{g}}$ field to introduce some stochasticity which we expect to be
present in the real galaxy catalogues. This provides an integer value in each
Cartesian cell for the galaxy counts. The last step is to assign random
Cartesian coordinates within the cell boundaries for the
$n_{\text{g}}(\boldsymbol{x})$ galaxies within each cell. At this stage we
rely on the Cartesian field having sufficiently high resolution such that the
random assignment of a coordinate within the cell does not shift the galaxies
away from the peaks of the density fluctuations. The catalogue of galaxy
coordinates (x, y, z [$h^{-1}$Mpc]) can then be transformed into celestial
coordinates $(\theta_{\text{R.A.}},\theta_{\text{Dec.}},z)$.
In this work, we generate 500 mock intensity maps with corresponding mock GAMA
galaxy catalogues. As some early validation of these mocks, we show in Figure
8 how the power spectra for both the Hi intensity map and galaxy auto-
correlations, along with their cross-correlation, agree with the fiducial
model (black dashed line). This demonstrates that our estimator is unbiased
and processes such as the regridding are performing to within an acceptable
tolerance. The power spectrum estimation process used here on the mocks is
identical to that used on the real data and is detailed later in Section 5,
along with the modelling.
Figure 8: Averaged power spectra (blue solid line) of the 500 regridded mocks
and agreement with the fiducial model (black dashed line). Mocks include the
beam but no other systematics. Green dotted lines in the bottom panels mark
the edges of the $10\%$ residual region, while the shaded regions mark the
diagonal variance inferred from the simulations.
The mocks used in the test of Figure 8 do not include any additional
systematics beyond the telescope beam in the intensity maps. We will
investigate the inclusion of foregrounds and systematics in Section 4. The
large scatter shown by the shaded blue region, which represents the
$1\,\sigma$ range across the 500 mocks, is therefore driven by cosmic variance
due to the small volume covered by the data. We see that the errors for the
galaxy power spectrum appear larger than the Hi intensity maps. This is
because we have not introduced any thermal noise (or other observational or
systematic effects) into the intensity maps. However, the galaxy catalogues
will have shot noise contributing to the uncertainties. This nicely showcases
the potential for Hi intensity mapping if noise and systematics can be
controlled, however, this is not the purpose of this exercise so is not a
point we will labour. There are also additional caveats when considering high-
density tracers such as Hi. As shown by Foreman et al. (2024), despite the low
shot noise, stochastic noise and non-linear bias in the Hi field can limit the
potential of constraints inferred from the power spectrum.
The green dashed lines in the residual panels mark the regions where there is
${>}\,90\%$ accuracy relative to the model which we can see is achieved at all
scales for all three cases. This is also well within the statistical
uncertainty (blue-shaded regions). The largest residual comes in the cross-
correlation where discrepancies seem to increase with $k$ nearly reaching the
$-10\%$ boundary. We speculate that this may be caused by the Poisson sampling
of mock galaxies followed by random assignment of coordinates within the cell.
This extra stochasticity is not repeated in the Hi intensity map mocks and
could therefore be generating a pseudo-cross-correlation coefficient, which
would depart from $r\,{=}\,1$ at smaller scales (higher-$k$). However, this
appears to be a minor effect. Furthermore, this could be seen as a feature,
since we expect $r\,{<}\,1$ in the real observations, which will also have a
scale dependence. For these reasons, we deem this to be a suitable result for
the purposes of this work, and defer a higher precision investigation and
validation to future work.
## 4 Foreground cleaning
Foreground cleaning Hi intensity maps is a well-studied problem (Santos et
al., 2005; Ansari et al., 2012; Wolz et al., 2014; Alonso et al., 2015; Shaw
et al., 2015; Olivari et al., 2016; Carucci et al., 2020; Cunnington et al.,
2021; Spinelli et al., 2022; Soares et al., 2022; Irfan & Bull, 2021), and
generally blind cleaning techniques are the currently preferred method for
single-dish experiments. As their name suggests, blind methods do not rely on
prior knowledge of the foreground structure. This is advantageous for current
experiments because the foregrounds at these frequencies lack observational
constraints. Furthermore, blind techniques avoid placing overwhelmingly
stringent calibration requirements on the data, since the dominant foregrounds
require beyond sub-percent precision modelling for a successful subtraction.
Blind techniques instead work by exploiting the foreground dominance and their
frequency-correlated structure to assume that they can be reduced to a
distinct set of correlated modes separated from the oscillating, sub-dominant
fluctuations which will contain most of the Hi information.
We outline the formalism for our blind foreground cleaning approach in Section
4.3, but first we discuss two pre-processing steps before foreground cleaning.
These are designed to improve the foreground clean by reducing the impact from
residual RFI or systematics, plus any known frequency-dependent observational
effects.
### 4.1 Reconvolution
All radio intensity mapping experiments will have a chromatic beam whose size
evolves with frequency. This can add to the number of degrees of freedom in
the radial foreground structure, inhibiting foreground removal since more
modes need to be removed to perform a suitable clean. Smoothing the maps
differently in each frequency channel can homogenise the angular resolution,
giving them a consistent effective beam size at all frequencies. This
reconvolution should also suppress residual contamination manifesting as
localised amplitude spikes in the temperature maps, smoothing them out so
there are fewer discontinuous fluctuations which will impact the PCA and
Fourier transforms for power spectrum estimation.
The reconvolution step involves convolving the intensity maps, one frequency
channel $\nu_{i}$ at a time with the kernel
$\pazocal{K}(\boldsymbol{\theta},\nu_{i})=\exp\left[-\frac{1}{2}\frac{\left(\boldsymbol{\theta}-\theta_{0}\right)^{2}}{\gamma\sigma_{\max}^{2}-\sigma_{\text{beam}}^{2}(\nu_{i})}\right]\,,$
(24)
where $\sigma_{\text{beam}}$ is the beam size as previously defined in
Equation 17, $\sigma_{\text{max}}$ is the maximum
$\sigma_{\text{beam}}(\nu_{\text{min}})$ value at the lowest frequency, and
$\gamma$ is a scaling factor which governs how much the final effective
resolution decreases. The reconvolution kernel is normalised such that the sum
over all pixels is unity i.e.
$\Sigma_{\boldsymbol{\theta}}\pazocal{K}(\boldsymbol{\theta},\nu_{i})\,{=}\,1$.
As in previous work (Cunnington, Li et al., 2022), we use a weighted
reconvolution to mitigate the impact of edge pixels with low scan coverage.
From hereon, the weights for the intensity maps derive directly from the time
stamp count in each pixel, as shown in Figure 4. The weighted reconvolved
intensity maps are given by
$\delta T_{\circledast}(\boldsymbol{\theta},\nu_{i})=\frac{\left[\delta
T(\boldsymbol{\theta},\nu_{i})w(\boldsymbol{\theta},\nu_{i})\right]\circledast\pazocal{K}(\boldsymbol{\theta},\nu_{i})}{w(\boldsymbol{\theta},\nu_{i})\circledast\pazocal{K}(\boldsymbol{\theta},\nu_{i})}$
(25)
The weights themselves must also be reconvolved to keep them consistent and
optimal for the new $\delta T_{\circledast}$ maps. The reconvolved weights are
given by
$w_{\circledast}(\boldsymbol{\theta},\nu_{i})=\frac{\left[w(\boldsymbol{\theta},\nu_{i})\circledast\pazocal{K}(\boldsymbol{\theta},\nu_{i})\right]^{2}}{w(\boldsymbol{\theta},\nu_{i})\circledast\pazocal{K}^{2}(\boldsymbol{\theta},\nu_{i})}\,.$
(26)
In this work we use a value of $\gamma\,{=}\,1.4$ for the reconvolution kernel
in Equation 24. We find that neglecting this resmoothing step entirely
increases the influence of systematics in the results which is likely due to
residual RFI or contaminants not being smoothed and depressing their
dominance. We discuss this further in Appendix B.1 but also plan a dedicated
study into this aspect of the analysis where a more robust model of the
MeerKAT beam will also be implemented.
The resmoothed, uncleaned map is shown in Figure 9 (top panel). Careful
comparison with Figure 4 reveals how some of the foreground structure has been
smoothed, albeit subtly, which is the aim. Too much reconvolution will
suppress power in high $k_{\perp}$ modes in the final power spectrum estimate.
The reconvolution is more noticeable in the time stamp count field, which we
use as the weights $w_{\circledast}$, shown by the bottom panel of Figure 9.
Figure 9: Reconvolved MeerKLASS deep field (top-panel) using
$\gamma\,{=}\,1.4$ (see Equation 25). The bottom panel shows the reconvolved
time-stamp field which we use for weighting, thus also requires reconvolution
(using Equation 26) for consistency with the intensity maps. Both panels have
been averaged along frequency.
### 4.2 Map trimming
Shown by the time stamp count field of Figure 4 (and the resmoothed version in
Figure 9), the MeerKLASS scanning strategy of combining numerous rising and
setting scans does not generate a uniform coverage. The edges of the patch
result in relatively low recorded time stamps due to a lack of overlapping
scans in these regions.
We found that implementing an edge trim to the field improved the quality of
the foreground clean and final power spectrum estimates. This is motivated and
explored in greater detail in Carucci et al. (in prep.). The improvement from
map trimming is because pixels are removed where residual time-varying
contamination has been averaged down less, due to fewer scans.
In Figure 9, the black contour shows the cut to the patch we make, removing
everything outside the trimmed lines. This amounts to all pixels at
$334^{\circ}\,{<}\,\text{R.A.}\,{<}\,357^{\circ}$ being set to zero in both
the weights and intensity map field.
### 4.3 Blind foreground cleaning with PCA
We represent the intensity maps with matrices X which are equivalent to the
$T(\boldsymbol{\theta},\nu)$ temperature fields, discretised into arrays on
which matrix algebra can be performed.
Blind foreground cleaning assumes the matrix of the observational data cube
$\textbf{{X}}_{\mathrm{obs}}$, which contains the intensity maps with
$n_{\theta}$ pixels at each of the $n_{\nu}$ frequency channels, can be
generalised by the linear equation
$\textbf{{X}}_{\mathrm{obs}}=\textbf{{A}}\textbf{{S}}+\textbf{{R}}\,.$ (27)
A is referred to as the mixing matrix and is the set of $N_{\text{fg}}$
spectral functions estimated to project out the foregrounds. S are the
$N_{\text{fg}}$ source maps obtained by projecting A along the data;
$\textbf{{S}}=\textbf{{A}}^{\text{T}}\textbf{{X}}_{\mathrm{obs}}\,.$ (28)
The dot product AS thus has shape $[n_{\nu},n_{\theta}]$, along with the other
terms in Equation 27, and represents a subset of the observations estimated to
contain the bulk of the foreground contamination. The residual term R will
contain the remaining Hi signal and thermal noise. In reality, the situation
is more complex and there is cross-term leakage. Both Hi signal will leak into
the subtracted AS term causing signal loss, and foreground will leak into the
remaining R causing foreground residuals. The control of both these is crucial
and we discuss them further throughout the remaining sections.
Different blind foreground techniques only vary in their approach to
estimating the mixing matrix A. In this work, we adopt the straightforward
principal component analysis (PCA) approach, since it is well-tested and
understood. Implementing more sophisticated blind techniques is planned in
future work. PCA foreground cleaning begins by computing the frequency
covariance matrix of the weighted observations
$\textbf{{C}}=\frac{\left(\textbf{{w}}\circ\textbf{{X}}_{\mathrm{obs}}\right)^{\mathrm{T}}\left(\textbf{{w}}\circ\textbf{{X}}_{\mathrm{obs}}\right)}{n_{\theta}-1}\,,$
(29)
where we have used the intensity mapping weights to optimise the estimation
and $\textbf{{w}}\circ\textbf{{X}}_{\mathrm{obs}}$ represents the element-wise
product between the weights and the observational data matrix. Since the
weights are the stacked average of the counts along the line-of-sight, they
are constant in frequency by design to avoid increasing the rank of
$\textbf{{X}}_{\mathrm{obs}}$ i.e. not inject additional structure into the
frequency spectra. Weighting in this way means C is not a formally normalised
covariance estimation, however, this is not a concern for us since we are only
extracting unitless eigenvectors from the covariance and using these to
project out modes in the real observed data so sensible amplitudes are
recovered in the component separated pieces. We discuss this further, with
some examples in Appendix B.2. C is then eigen-decomposed by
$\textbf{{C}}=\textbf{{U}}\boldsymbol{\Lambda}\textbf{{U}}^{\text{T}}\,,$ (30)
where U is the matrix of eigenvectors with the corresponding eigenvalues
contained in $\boldsymbol{\Lambda}$. The mixing matrix is then formed by
taking the first $N_{\text{fg}}$ column vectors $\boldsymbol{u}_{i}$ from U
i.e.
$\textbf{{A}}\,{=}\,[\boldsymbol{u}_{1},\boldsymbol{u}_{2},...,\boldsymbol{u}_{N_{\text{fg}}}]$.
The first 9 eigenmodes are shown in the Appendix plot Figure 22. It is at this
stage a choice is required for $N_{\text{fg}}$, which determines how many
modes are removed and deemed as foregrounds. If the $N_{\text{fg}}$ is too
high, the Hi signal loss increases; on the contrary, if $N_{\text{fg}}$ is too
low, residual contamination will dominate more. The estimate of the
foregrounds is provided by
$\textbf{{X}}_{\mathrm{fg}}\,{\equiv}\,\textbf{{A}}\textbf{{S}}$ which is
subtracted to provide the residuals, revealing the final cleaned data used in
the analysis i.e. $\textbf{{X}}_{\text{clean}}\,{\equiv}\,\textbf{{R}}$ where
$\textbf{{X}}_{\text{clean}}=\textbf{{X}}_{\mathrm{obs}}-\textbf{{X}}_{\mathrm{fg}}\,.$
(31)
Figure 10 shows the resulting cleaned maps by removing
$N_{\text{fg}}\,{=}\,10$ PCA modes. We show the maps split into six frequency
sub-bands (given in the bottom-left of each panel) to illustrate the changing
intensity structures. Each panel is the frequency average of the 42 channels
in each sub-band showing a reasonably consistent amplitude of temperature
fluctuations across the sub-bands. The amplitudes in the maps are still likely
driven by non-cosmological contributions, mostly thermal noise, but they are
not far from the $0.1\,\text{mK}$ fluctuations we expect from Hi. As we will
show statistically in Section 5, over large-scale fluctuations, the
predominant contribution comes from the cosmological Hi.
Figure 10: Foreground cleaned intensity maps for the MeerKLASS L-band deep
field in frequency sub-bands (stated in bottom left). Maps have been cleaned
by removing $N_{\text{fg}}\,{=}\,10$ PCA modes.
### 4.4 Foreground cleaning tests using mocks
In order to validate our foreground removal and estimate the signal loss it
induces, we run the same PCA cleaning technique on mock intensity maps. As
discussed in Section 3.2, simulating foreground contamination that emulates
the systematics and complex observational effects is beyond current
capabilities. We therefore use the real observations themselves and inject
mock Hi intensity maps into these. We then apply the foreground removal
pipeline on the mock-injected data and estimate the signal loss from the
injected mock whose original signal is known.
We follow the formalism outlined in (Cunnington, Li et al., 2022), beginning
with a Hi mock, whose matrices we notify by M to distinguish from the real
observation X. The mock is injected into the real observations i.e.
$\textbf{{X}}_{\text{obs}}+\textbf{{M}}_{\textrm{H{i}}}$. This is then cleaned
by projecting out a subset of modes defined by the mixing matrix, as in the
previous sections, such that
$\textbf{{M}}_{\text{clean}}=\left(\textbf{{X}}_{\text{obs}}+\textbf{{M}}_{\textrm{H{i}}}\right)-\textbf{{A}}\textbf{{A}}^{\text{T}}\left(\textbf{{X}}_{\text{obs}}+\textbf{{M}}_{\textrm{H{i}}}\right)\,,$
(32)
The data in $\textbf{{M}}_{\text{clean}}$ will now contain signal loss
emulating that of the real data clean, along with residual systematics and
noise from the real data.
We show the results of mocks contaminated by the real data in Figure 11. We
show measured power spectra averaged over all 500 mocks under different
scenarios (details of power spectrum estimation are provided in the following
Section 5). Firstly, the grey dotted line shows the Hi-only mocks,
$\textbf{{M}}_{\textrm{H{i}}}$, with no foreground cleaning effects,
representing the true power. The red solid line shows the result of adding the
real observations to the mock, then PCA cleaning as in Equation 32. The power
is higher than the grey line, despite the inevitable signal loss the clean
will have caused. This is evidence that the $\textbf{{M}}_{\text{clean}}$ maps
contain residual systematics and thermal noise, which are additively biasing
the power999There will also be a small contribution from the true Hi signal in
the observed data.. We can demonstrate this by considering the orange solid
line which is the cross-correlation of the cleaned mocks
$\textbf{{M}}_{\text{clean}}$ with the Hi-only mocks
$\textbf{{M}}_{\textrm{H{i}}}$, a test only possible with simulations. Since
the residual systematics and noise will be uncorrelated with the Hi-only
mocks, the distortion in power comes purely from signal loss, hence why it is
lower than the Hi-only power at all scales.
To isolate the contribution from unknown systematics, we need to account for
thermal noise and signal loss from foreground cleaning. The thermal noise
contribution can be approximated using Equation 20 to generate 500 noise maps
and add these to the Hi-only mocks. The black-dashed line shows this and
demonstrates the boost in power relative to the grey-dotted line, more so at
small scales where, due to the beam, the Hi signal is low relative to the
scale-independent noise101010The noise is not completely scale-independent
because it has gone through the same regridding and reconvolution processes as
the data which causes some subtle damping.. The last result presented in
Figure 11 is the reconstructed power (i.e. correcting for foreground cleaning
signal loss) shown by the blue solid line. This utilises the signal-loss
evaluated from the cross-power (orange line) to reconstruct the
$\textbf{{M}}_{\text{clean}}$ auto-power. This is discussed more in the
following sub-section. The discrepancy between the blue-solid and black-dashed
line must be caused by systematics, since signal-loss and thermal noise have
now been accounted for. We find a similar excess power with the real
observations, which we discuss in the following Section 5. The results in
Figure 11 are to validate the realism of the mock pipeline, and also to
motivate the signal loss reconstruction process, which we formalise in the
next sub-section.
Figure 11: Power spectra for the Hi intensity map mocks under different
scenarios, each averaged over 500 mock realisations. Including: Hi-only mock
(grey dotted line). Hi plus thermal noise realisations (black dashed). The Hi
mock with added foreground and systematics from the real MeerKLASS data, then
cleaned as in Equation 32 (red solid). The cross-correlation between Hi-only
and cleaned mock (orange solid). Lastly, the blue solid line is the auto-
correlation of the cleaned mock but with signal loss reconstructed.
### 4.5 Signal loss correction with transfer functions
We can utilise the method of mock signal injection from the previous Section
4.4 for estimating the signal loss in the power spectrum of the real data.
This is now a well-understood process (Switzer et al., 2015; Cunnington et
al., 2023) and used in previous single-dish intensity mapping analyses (Masui
et al., 2013; Anderson et al., 2018; Wolz et al., 2022; Cunnington, Li et al.,
2022).
We begin with the $\textbf{{M}}_{\text{clean}}$ term in Equation 32, which as
demonstrated, will contain signal loss that should emulate the real data. In
some cases, the cleaned real data $\textbf{{X}}_{\text{clean}}$ can be
subtracted from $\textbf{{M}}_{\text{clean}}$. This accelerates the
convergence of the transfer function calculation by removing the additional
noise caused by the real data but has no bearing on the signal loss
calculation. Leaving this piece in the calculation of
$\textbf{{M}}_{\text{clean}}$ is however crucial when using the transfer
function for estimating errors in the power spectrum. We also find that due to
the large number of mocks used in this work, an equivalent accuracy of the
transfer function is obtained whether subtracting
$\textbf{{X}}_{\text{clean}}$ or not. For simplicity, we therefore only use
the case as defined by Equation 32, without the $\textbf{{X}}_{\text{clean}}$
subtraction.
The transfer function is defined by the power spectrum, signified by the
operator $\pazocal{P()}$, of the cleaned mock $\textbf{{M}}_{\text{clean}}$
with the foreground-free Hi-only mock $\textbf{{M}}_{\textrm{H{i}}}$, and the
ratio of this with the power spectrum of the Hi-only mock
$\pazocal{T}_{\textrm{H{i}}}(\boldsymbol{k})=\left\langle\frac{\pazocal{P}\left(\textbf{{M}}_{\text{clean}},\textbf{{M}}_{\textrm{H{i}}}\right)}{\pazocal{P}\left(\textbf{{M}}_{\textrm{H{i}}},\textbf{{M}}_{\textrm{H{i}}}\right)}\right\rangle_{N_{\text{mock}}}\,.$
(33)
The angled brackets indicate an averaging over a large number of mocks such
that a converged result is obtained. In this work, we use all mocks in our
suite, thus $N_{\text{mock}}\,{=}\,500$. The operator $\pazocal{P()}$ can
represent any power spectrum estimator i.e. we use this same recipe for
estimating signal loss in the spherically averaged power and also in the
cylindrical ($k_{\perp}$, $k_{\parallel}$) power spectrum, which we later
utilise in some analysis.
In this work, we will mostly focus on the MeerKLASS deep field’s cross-
correlation with the overlapping GAMA galaxies. The signal loss from PCA
cleaning in the cross-power spectrum should be identical to that in the auto-
power. This was established empirically in Cunnington et al. (2023) which
showed how removing a subset of PCA modes entirely but leaving their
contribution in the other half of the cross-power product (e.g. in the galaxy
maps which have no cleaning, thus no signal loss) generates no more power than
the auto-Hi. We present further reasoning for why this is the case in Appendix
B.3. To summarise, we believe it is explained by the PCA modes being
orthogonal, thus maximally uncorrelated. Therefore, the additional remaining
modes in the uncleaned field are linearly uncorrelated with what survives in
the cleaned Hi, producing no extra power.
Despite the approximately equal signal loss between auto and cross-
correlations, we still define a separate transfer function for the cross-power
spectrum purely for covariance estimation purposes whereby the only difference
is we switch in the GAMA galaxy mocks $\textbf{{M}}_{\text{g}}$ instead which
will emulate the signal strength and shot-noise of the GAMA galaxies, thus
providing a more robust error estimate in the cross-power.
$\pazocal{T}_{\textrm{H{i}},\text{g}}(\boldsymbol{k})=\left\langle\frac{\pazocal{P}\left(\textbf{{M}}_{\text{clean}},\textbf{{M}}_{\text{g}}\right)}{\pazocal{P}\left(\textbf{{M}}_{\textrm{H{i}}},\textbf{{M}}_{\text{g}}\right)}\right\rangle_{N_{\text{mock}}}\,.$
(34)
Any measured power spectra should be divided by $\pazocal{T}(\boldsymbol{k})$,
thus the transfer function represents a reconstruction of signal at the power
spectrum level.
## 5 Power spectrum estimation
In this section we present power spectra estimations from the foreground
cleaned MeerKLASS L-band deep field. We first outline the pipeline for
performing these measurements then present the results. By default, we use
maps that have been foreground cleaned by removing $N_{\text{fg}}\,{=}\,10$
PCA modes, as outlined in the previous section. However, we also examine the
results from varying $N_{\text{fg}}$.
### 5.1 Transforming to a Cartesian grid
The first task to perform in order to measure a power spectrum is to transform
the foreground cleaned intensity maps onto a Cartesian grid with comoving
$h^{-1}\text{Mpc}$ coordinates. This allows for a Fast Fourier transform to be
calculated on the Cartesian array, from which the power at different Fourier
bandpasses $\boldsymbol{k}$, can be averaged and presented.
We discussed the transformation between sky maps and Cartesian grids already
in Section 3.1 when we performed the inverse routine of sampling Cartesian
cells from lognormal mocks into celestial sky map voxels that emulate the
MeerKLASS intensity maps. We again utilise the recent work and code in
Cunnington & Wolz (2024) and first define a grid of size
$l_{\text{x}},l_{\text{y}},l_{\text{z}}\,{=}\,632.2,341.5,231.5\,h^{-1}\text{Mpc}$
with $n_{\text{x}},n_{\text{y}},n_{\text{z}}\,{=}\,88,48,168$ cells along each
dimension. The choice of cells is a factor of $1.5$ smaller than the number of
map voxels in R.A., Dec., and $\nu$. Making the Cartesian grid slightly lower
resolution than the map and using 10 sampling particles per map voxel ensures
each Cartesian cell is well-sampled, avoiding empty cells.
The foreground cleaned maps $\delta
T_{\textrm{H{i}}}(\boldsymbol{\theta},\nu)$ are transformed to this Cartesian
grid, giving $\delta T_{\textrm{H{i}}}(\boldsymbol{x})$. The intensity mapping
weights also follow an identical transformation process
$w_{\textrm{H{i}}}(\boldsymbol{\theta},\nu)\,{\rightarrow}\,w_{\textrm{H{i}}}(\boldsymbol{x})$.
The catalogue of GAMA galaxies can be directly gridded by transforming their
celestial coordinates into Cartesian, then assigning them onto the Cartesian
grid $n_{\text{g}}(\boldsymbol{x})$. For all sampling particles and galaxies
we use a straightforward nearest grid point (NGP) assignment. Cunnington &
Wolz (2024) presented extensions to this method for avoiding aliased power and
obtaining sub-percent accuracy. However, given the current signal-to-noise of
our data at small scales, aliasing effects will be negligible, thus higher-
order interpolations and interlacing are overkill and serve only to increase
compute demand. However, we correct for the dominant effect which is the
field’s smoothing by the (NGP) particle assignment. This requires deconvolving
the field, achieved by dividing through by the NGP mass assignment window
function in Fourier space (Jing, 2005)
$\tilde{\delta
T}_{\textrm{H{i}}}(\boldsymbol{k})\rightarrow\frac{\tilde{\delta
T}_{\textrm{H{i}}}(\boldsymbol{k})}{W_{\text{ngp}}(\boldsymbol{k})}\,,$ (35)
where the window function $W_{\text{ngp}}$ accounts for the sampling particles
discretisation into cells of size $H$. These are separable functions along
each Cartesian dimension, thus modelled by a Fourier transform of a top-hat
function, which becomes a sinc function;
$\operatorname{sinc}(x)\,{=}\,\sin(x)/x$ and therefore
$\tilde{W}_{\mathrm{ngp}}(\boldsymbol{k})=\operatorname{sinc}\left(\frac{k_{\mathrm{x}}H_{\mathrm{x}}}{2}\right)\operatorname{sinc}\left(\frac{k_{\mathrm{y}}H_{\mathrm{y}}}{2}\right)\operatorname{sinc}\left(\frac{k_{\mathrm{z}}H_{\mathrm{z}}}{2}\right)\,.$
(36)
Previous work (e.g. Switzer et al., 2013; Cunnington, Li et al., 2022)
employed a tapering along the radial direction of the regridded field with the
primary purpose of reducing edge effects in the Fourier transform. Our
modelling approach should however account for these effects since the model is
also convolved with the window function (Blake, 2019). We still found results
were worse when omitting a radial tapering though, which we suspect is instead
caused by edge-effects from the foreground clean (Alonso et al., 2015) which
are suppressed by the tapering. We experimented with a less aggressive taper
by using a Tukey window, also known as a cosine window (this is also explored
in Carucci et al. in prep). This has a shape parameter $\alpha$ which is
simply a top-hat with no apodisation when $\alpha\,{=}\,0$ and equivalent to a
Hann window when $\alpha\,{=}\,1$. We show these window functions in Figure
12. By opting for less aggressive Tukey windows we still effectively suppress
the extreme edges where foreground cleaning effects may be present but avoid
unnecessary tapering on the central data. We will explore this further with
future data in MeerKAT’s UHF band where the larger useable frequency ranges
may render this less of an issue. In this work we found optimal results when
using $\alpha\,{=}\,0.8$ for the Tukey window which is applied to the
intensity map weights $w_{\textrm{H{i}}}(\boldsymbol{x})$. We do not apply
this to the galaxy data as we found this increased the correlations between
$k$-bins in the power spectrum.
Figure 12: Tukey window functions with differing shape parameters $\alpha$.
For comparison, we also show the Blackman window used in previous work.
### 5.2 Power spectrum formalism
With the fields assigned to a Cartesian space and appropriately tapered, the
power spectra can then be measured. The Fourier transformed field of the Hi
temperature fluctuation maps $\delta T_{\textrm{H{i}}}$ is given by
$\tilde{F}_{\textrm{H{i}}}(\boldsymbol{k})=\sum_{\boldsymbol{x}}\delta
T_{\textrm{H{i}}}(\boldsymbol{x})w_{\textrm{H{i}}}(\boldsymbol{x})\exp(i\boldsymbol{k}{\cdot}\boldsymbol{x})\,,$
(37)
where the weights, now denoted by $w_{\textrm{H{i}}}$ to distinguish them from
the galaxy weights, still derive from the field of time-stamp counts, which
are regridded following the same process as outlined in the previous sub-
section. The Fourier transformed field of the the galaxy count field
$n_{\text{g}}$ is given by
$\tilde{F}_{\text{g}}(\boldsymbol{k})=\sum_{\boldsymbol{x}}n_{\text{g}}(\boldsymbol{x})w_{\text{g}}(\boldsymbol{x})\exp(i\boldsymbol{k}{\cdot}\boldsymbol{x})-N_{\text{g}}\tilde{W}_{\text{g}}(\boldsymbol{k})\,,$
(38)
where $N_{\text{g}}\,{=}\,\sum n_{\text{g}}$ is the total number of galaxies
in the optical maps and $\tilde{W}_{\text{g}}$ is the weighted Fourier
transform of the selection function
$\tilde{W}_{\text{g}}(\boldsymbol{k})=\sum_{\boldsymbol{x}}W_{\text{g}}(\boldsymbol{x})w_{\text{g}}(\boldsymbol{x})\exp(i\boldsymbol{k}{\cdot}\boldsymbol{x})\,.$
(39)
The galaxy window function $W_{\text{g}}$ is normally constructed by stacking
mock galaxies that carefully follow the survey selection characteristics of
the experiment to account for its incompleteness. As discussed previously,
there are no GAMA G23 mocks available for this purpose so we therefore
construct a simple window function which is 1 within the survey boundaries and
0 everywhere else. Some inclusion of the redshift evolution of $N(z)$ could
have been included, but as Figure 7 showed, the variation over our narrow
redshift range is minimal. The window function is normalised so that
$\sum_{\boldsymbol{x}}W_{\text{g}}(\boldsymbol{x})\,{=}\,1$. Since we lack
this proper galaxy window function we do not implement the optimal weighting
as per Feldman et al. (1994), but instead use uniform weighting, so equivalent
of assuming
$w_{\text{g}}(\boldsymbol{x})\,{\equiv}\,W_{\text{g}}(\boldsymbol{x})$.
The cross-power spectrum estimator is then given by
$\hat{P}_{\textrm{H{i}},\text{g}}(\boldsymbol{k})=\frac{V_{\text{cell}}}{\sum_{\boldsymbol{x}}w_{\textrm{H{i}}}(\boldsymbol{x})w_{\text{g}}(\boldsymbol{x})W_{\text{g}}(\boldsymbol{x})}\operatorname{Re}\left\\{\tilde{F}_{\textrm{H{i}}}(\boldsymbol{k})\tilde{F}^{*}_{\text{g}}(\boldsymbol{k})\right\\}\frac{1}{N_{\text{g}}}\,,$
(40)
where
$V_{\text{cell}}\,{=}\,l_{\text{x}}l_{\text{y}}l_{\text{z}}/(n_{\text{x}}n_{\text{y}}n_{\text{z}})$.
The hat notation $\left(\hat{P}\right)$ indicates that this is the estimated
power from the data, to distinguish from the modelled power, which we discuss
shortly. The Hi auto power spectrum estimator is more straightforwardly given
by
$\hat{P}_{\textrm{H{i}}}(\boldsymbol{k})=\frac{V_{\text{cell}}}{\sum_{\boldsymbol{x}}w_{\textrm{H{i}}}^{2}(\boldsymbol{x})}\lvert\tilde{F}_{\textrm{H{i}}}(\boldsymbol{k})\rvert^{2}\,,$
(41)
These power spectra are either spherically averaged into bandpowers
$\lvert\boldsymbol{k}\rvert\,{\equiv}\,k$ to provide 1D power spectrum results
or alternatively cylindrically averaged into 2D bandpowers defined by modes
perpendicular ($k_{\perp}$) and parallel ($k_{\parallel}$) to the radial
direction z, which we assume is the line-of-sight in a plane-parallel
approximation, valid for the current MeerKLASS sky sizes. They are finally
divided by their relevant foreground transfer functions (Equation 33 and
Equation 34) to reconstruct signal loss for each $k$-bin.
#### 5.2.1 Modelling the power spectrum
Here we outline the model of the power spectrum estimations for which we
mostly follow the work in Blake (2019), with some extensions. We model the Hi
power spectrum using the equation already introduced in Equation 14 when
discussing mock simulations. We repeat the model here for convenience;
$P_{\textrm{H{i}}}(\boldsymbol{k})\equiv
P_{\textrm{H{i}}}(k,\mu)=\frac{\overline{T}_{\textrm{H{i}}}^{2}\left(b_{\textrm{H{i}}}+f\mu^{2}\right)^{2}P_{\text{m}}(k)}{1+\left(k\mu\sigma_{\text{v}}/H_{0}\right)^{2}}\,.$
(42)
Additional modelling of observational and mapping effects is also necessary
since these will distort the power. To model the MeerKAT beam and the
additional smoothing from the reconvolution process we assume the field has
undergone a constant smoothing by a 2D Gaussian kernel whose standard
deviation is given by
$R_{\text{beam}}\,{=}\,\sqrt{\gamma}\,R^{\max}_{\text{beam}}$ in comoving
units. $\gamma\,{=}\,1.4$ accounts for the reconvolution to a common effective
resolution and $R^{\max}_{\text{beam}}$ is the Gaussian approximation of the
MeerKAT beam size at the deepest redshift (and lowest frequency
$\nu_{\text{min}}$)
$R^{\max}_{\text{beam}}=\chi(z_{\max})\,\sigma_{\text{beam}}(\nu_{\min})\,,$
(43)
where $\sigma_{\text{beam}}$ is given as before from Equation 17 and $\chi(z)$
is the comoving distance to redshift $z$, which in this case, is the maximum
redshift for our frequency range. The perpendicular modes will then be damped
according to
$\tilde{\pazocal{B}}_{\text{beam}}(\boldsymbol{k})=\exp\left[\frac{-k^{2}\left(1-\mu^{2}\right)R_{\text{beam
}}^{2}}{2}\right]\,.$ (44)
For this work, we find $R_{\text{beam}}\,{=}\,14.1\,h^{-1}\text{Mpc}$.
Similar to the beam, there will also be distortions to the power caused by the
finite resolution of the intensity maps. We approximate the pixel size in
comoving units by $s_{\text{pix}}\,{=}\,\chi(z_{\text{eff}})\Delta\theta$
where $\Delta\theta\,{=}\,0.3\,\deg$ is the pixel size used for the MeerKLASS
L-band mapping. The frequency channel depth is approximated by taking the mean
across all spacings between redshifts corresponding to channel boundaries i.e.
$s_{\text{chan}}\,{=}\,\Sigma_{i}\left[\chi(z_{i})-\chi(z_{i{+}1})\right]/N_{\text{chan}}$.
These quantities are used in damping functions to model the suppressed
resolution from mapping into pixels and frequency channels, the former being
the far more dominant effect for the wide-pixel MeerKLASS intensity maps.
These functions are given as
$\tilde{\pazocal{B}}_{\text{pix}}(\boldsymbol{k})=\operatorname{sinc}\left(\frac{k\sqrt{1-\mu^{2}}\,s_{\text{pix}}}{2}\right)\,,$
(45)
and
$\tilde{\pazocal{B}}_{\text{chan
}}(\boldsymbol{k})=\operatorname{sinc}\left(\frac{k\,\mu\,s_{\text{chan
}}}{2}\right)\,,$ (46)
The model and damping functions are discretised onto the same 3D grid of
$k$-modes as the FFT for the real data, then convolved with the survey window
function. The model for the observed Hi auto-power spectrum therefore becomes
$P_{\textrm{H{i}}}^{\text{obs}}(\boldsymbol{k})=\frac{W^{2}_{\textrm{H{i}}}(\boldsymbol{k})\circledast\left[P_{\textrm{H{i}}}(\boldsymbol{k})\,\pazocal{B}_{\text{beam}}^{2}(\boldsymbol{k})\,\pazocal{B}_{\text{pix}}^{2}(\boldsymbol{k})\,\pazocal{B}_{\text{chan}}^{2}(\boldsymbol{k})\right]}{\sum_{\boldsymbol{x}}w^{2}_{\textrm{H{i}}}(\boldsymbol{x})W^{2}_{\textrm{H{i}}}(\boldsymbol{x})}\,,$
(47)
where we have normalised by the sum of the weights and window function,
$W_{\textrm{H{i}}}$, to obtain an amplitude that matches the weighted and
windowed estimator. For the intensity mapping window function, we assume a
simple binary function; 1 where we have scan coverage and 0 everywhere else.
There is scope for improving this approximation and further work is needed to
examine how intensity mapping surveys are impacted by any selection effects.
A similar approach is followed for modelling the cross-power with GAMA. The
cross-correlation power spectrum is given by
$P_{\textrm{H{i}},\mathrm{g}}(\boldsymbol{k})=\frac{\overline{T}_{\textrm{H{i}}}\left[rb_{\textrm{H{i}}}b_{\mathrm{g}}+b_{\textrm{H{i}}}f\mu^{2}+b_{\mathrm{g}}f\mu^{2}+f^{2}\mu^{4}\right]P_{\mathrm{m}}(k)}{1+\left(k\mu\sigma_{\text{v}}/H_{0}\right)^{2}}\,.$
(48)
where we introduce $r$, the correlation coefficient between the Hi
fluctuations and the GAMA galaxy over-densities. This accounts for the
different astrophysics in each tracer and non-linear clustering, which can
mean the two fields are not exact multiples of each other and there is instead
some stochasticity which will suppress their cross-power (see e.g.
Moradinezhad Dizgah et al., 2022; Sato-Polito et al., 2023; Obuljen et al.,
2023, for applications in the context of line-intensity mapping). We assume
$r$ to be constant, which is sufficient for the current signal-to-noise, but
in reality it will exhibit some scale dependence. The correlation coefficient
only enters on the single term in Equation 48 because it only influences the
biased tracer fields, not their unbiased velocity fields which are modulated
by $f$. The observational cross power is then similarly convolved with the
survey window functions but only requires single factors of the damping
functions $\pazocal{B}$;
$\displaystyle P_{\textrm{H{i}}\text{,g}}^{\text{obs}}$
$\displaystyle(\boldsymbol{k})=$ (49)
$\displaystyle\frac{W_{\textrm{H{i}}}(\boldsymbol{k})W^{*}_{\text{g}}(\boldsymbol{k})\circledast\left[P_{\textrm{H{i}}\text{,g}}(\boldsymbol{k})\,\pazocal{B}_{\text{beam}}(\boldsymbol{k})\,\pazocal{B}_{\text{pix}}(\boldsymbol{k})\,\pazocal{B}_{\text{chan}}(\boldsymbol{k})\right]}{\sum_{\boldsymbol{x}}w_{\textrm{H{i}}}(\boldsymbol{x})w_{\text{g}}(\boldsymbol{x})W_{\textrm{H{i}}}(\boldsymbol{x})W_{\text{g}}(\boldsymbol{x})}\,.$
We do not include modelling of uncertainties in the GAMA galaxy positioning,
such as redshift estimation. The GAMA radial kernel equivalent of Equation 46
will likely have a larger impact due to spectroscopic redshift estimation
errors. However, this will still be a sub-dominant contribution to the overall
modelling relative to the beam and intensity map pixelisation, thus it is
reasonable to neglect observational effects from the GAMA galaxies.
We perform a simple least-squares fit for an overall amplitude to our cross-
power, equivalent to assuming fixed values for all parameters except for a
fitted $\overline{T}_{\textrm{H{i}}}$. This is deliberately simplistic to keep
the model and fitting in a low-dimensional form. The small overlap sample in
the cross-correlation means the statistical noise from such measurements is
not yet suited to high-dimensional modelling. We are not therefore extending
much from previous work (e.g. Wolz et al., 2022; Cunnington, Li et al., 2022)
in this regard. Furthermore, the auto-Hi still has evidence of additive bias,
as we will discuss, and therefore can only be trusted for upper limits in
parameter inference.
For the power amplitude fitting, we assume the parameter values outlined by
Table 1. The Hi bias comes from interpolated results from hydrodynamical
simulations (Villaescusa-Navarro et al., 2018), which can be approximated by
the polynomial
$b_{\textrm{H{i}}}(z)=0.842+0.693z-0.046z^{2}\,.$ (50)
The galaxy bias is obtained by fitting the amplitude of the GAMA auto-power
spectrum. The growth rate is obtained from
$f\,{=}\,\Omega_{\text{m}}(z_{\text{eff}})^{0.545}$ for our Planck 18
cosmology. The correlation coefficient $r$ and velocity dispersion
$\sigma_{\text{v}}$ are poorly constrained nuisance parameters tuned by eye to
provide an adequate fit. We discuss model and parameter fitting further in
Section C where we present results from a simultaneous MCMC fit on several
parameters, demonstrating the degeneracies between them and discuss plans for
breaking such degeneracies, but this is not a central focus of this work.
Parameter | $b_{\textrm{H{i}}}$ | $b_{\text{g}}$ | $r$ | $f$ | $\sigma_{\text{v}}$
---|---|---|---|---|---
Fiducial value | 1.13 | 1.90 | 0.90 | 0.74 | 400 km s-1
Table 1: Fiducial parameters assumed in model fitting.
### 5.3 Error estimation with the transfer function
In previous analysis (Cunnington, Li et al., 2022), an analytical approach was
primarily used for estimating errors on the power spectrum measurements.
However, this approach makes various assumptions, mainly that thermal noise
dominates the error budget and that Gaussian statistics are valid. Since we
are now reaching regions in $k$-space where thermal noise no longer dominates
with MeerKLASS, we are required to relax these assumptions. Furthermore, the
assumption of Gaussianity is arguably not solid, given the additive
systematics we know to be still be present within the data.
A more robust approach of estimating the errors is to use the transfer
function scatter across the different mock realisations
$\pazocal{T}_{i}(\boldsymbol{k})$. This approach has been used in previous
analysis (e.g. Anderson et al., 2018) and has also been validated in
simulations (Cunnington et al., 2023). For it to perform effectively, the
mocks in the transfer function calculation must emulate the real data
effectively which we have strived to do so, as outlined in Section 3.
The transfer function for one single realisation is defined by
$\pazocal{T}_{i}^{\textrm{H{i}},\text{g}}(\boldsymbol{k})=\frac{\pazocal{P}\left(\textbf{{M}}_{\text{clean}},\textbf{{M}}_{\text{g}}\right)}{\pazocal{P}\left(\textbf{{M}}_{\textrm{H{i}}},\textbf{{M}}_{\text{g}}\right)}\,,$
(51)
i.e. equivalent to Equation 34 but without the averaging over
$N_{\text{mock}}$. To reconstruct signal in a measured power spectrum, we take
the ensemble average of Equation 34 i.e.
$\pazocal{T}(\boldsymbol{k})\,{=}\,\langle\pazocal{T}_{i}(\boldsymbol{k})\rangle_{N_{\text{mock}}}$,
and divide the estimated power by this;
$\hat{P}_{\text{rec}}(\boldsymbol{k})\,{=}\,\hat{P}(\boldsymbol{k})/\pazocal{T}(\boldsymbol{k})$.
Note we are dropping the subscript Hi,g for brevity and also because the
formalism we present here is identically applicable to the Hi-auto. The
assumption made when using the transfer function for error estimation is that
the scatter in the reconstructed power comes from the transfer function
distribution
$\pazocal{T}(\boldsymbol{k})\,{+}\,\delta\pazocal{T}_{i}(\boldsymbol{k})$
contributed by each $i$th mock. We can isolate the scatter component from the
array of calculated transfer functions by simply subtracting the ensemble mean
for each $i$th mock
$\delta\pazocal{T}_{i}(\boldsymbol{k})=\pazocal{T}_{i}(\boldsymbol{k})-\pazocal{T}(\boldsymbol{k})\,.$
(52)
Assuming any errors add in quadrature, error propagation suggests
$\frac{\delta\hat{P}^{\text{rec}}_{i}(\boldsymbol{k})}{\hat{P}_{\text{rec}}(\boldsymbol{k})}=\sqrt{\left(\frac{\delta\hat{P}_{i}(\boldsymbol{k})}{\hat{P}(\boldsymbol{k})}\right)^{2}+\left(\frac{\delta\pazocal{T}_{i}(\boldsymbol{k})}{\pazocal{T}(\boldsymbol{k})}\right)^{2}}\,,$
(53)
however, we assume all quantifiable error lies in the transfer function
distribution, hence $\delta\hat{P}_{i}\,{=}\,0$ and so
$\delta\hat{P}^{\text{rec}}_{i}(\boldsymbol{k})=\hat{P}_{\text{rec}}(\boldsymbol{k})\frac{\delta\pazocal{T}_{i}(\boldsymbol{k})}{\pazocal{T}(\boldsymbol{k})}\,.$
(54)
Our setup thus provides $N_{\text{mock}}$ realisations of
$\delta\hat{P}_{i}^{\text{rec}}$ and the covariance can be computed from this
for any $k$-bin. We reiterate that this formalism is valid whichever way the
$\boldsymbol{k}$-modes are binned e.g. for a 2D cylindrical averaged or 1D
spherically averaged power. In this work, our error bars in any 1D power
spectrum estimate represent the limits of the 68th percentile regions of the
distribution of $\delta\hat{P}^{\text{rec}}_{i}(k)$. If the uncertainties are
Gaussian symmetric, this is equivalent to $1\,\sigma$ error bars, however, as
we will later show, this is not perfectly the case for all $k$-modes.
### 5.4 Spherical averaging $k$-cuts
Before presenting spherically averaged power spectra from the MeerKLASS data,
we first present results from the cross-correlation with GAMA galaxies in
cylindrical-($k_{\perp},k_{\parallel})$ space. This is shown in Figure 13 and
motivates a discussion on where in $k$-space we have the most signal-to-noise
so we can make optimal spherically averaged $\hat{P}(k)$ measurements. The
dashed gray lines on all panels represent the $|\boldsymbol{k}|$-bin
boundaries we use for the spherically averaged power spectra, restricting this
to $k\,{<}\,0.3\,h\,\text{Mpc}^{-1}$ to avoid highly non-linear cosmology and
increasingly poor signal-to-noise. The orange solid lines represent boundaries
for potential $k$-cuts which we will discuss.
Figure 13: Cylindrical ($k_{\perp},k_{\parallel}$) analysis of power, signal-
loss and signal-to-noise. The first panel from the left shows the cross power
between the MeerKLASS deep field and GAMA galaxies. The second panel is the
transfer function used to reconstruct the power in
$\hat{P}_{\textrm{H{i}},\text{g}}$. The third panel shows the signal-to-noise
by dividing the cross power by the error estimation discussed in the text. The
final fourth panel counts the number of discrete $\boldsymbol{k}$ wavenumber
in the Fourier grid, demonstrating the concentration of available modes. The
gray contours mark the boundaries of the $k$-bins used in spherically averaged
power $P(k)$. The orange lines indicate regions used for $k$-cuts discussed in
the text.
The first panel of Figure 13 shows the cross-power between GAMA G23 galaxies
and the MeerKLASS deep field, cleaned by removing $N_{\text{fg}}\,{=}\,10$ PCA
modes. The $k_{\perp}k_{\parallel}$ scaling is for demonstrative purposes so
the high power at low-$k$ does not saturate the plot. At
$k\,{\gtrsim}\,0.15\,h\,\text{Mpc}^{-1}$ the large MeerKAT beam begins to have
an impact and will damp Hi fluctuations, which explains why the cross-power
signal tends to zero at high-$k_{\perp}$. Except for the extreme low-$k$
regions, the rest of the $k$-space reveals a reasonably consistent positive
cross-correlation signal, providing strong evidence of consistent clustering
between the MeerKLASS maps and GAMA galaxies. As discussed in the previous
Section 4.5, the signal loss in this power has been corrected with the
foreground transfer function. The second panel shows the corresponding
transfer function for the $N_{\text{fg}}\,{=}\,10$ clean. Low values of
$\pazocal{T}_{\textrm{H{i}},\text{g}}$ signify high signal loss. This is
evidence that nearly all signal is lost for the lowest-$k_{\parallel}$ modes
which is expected since these modes should be highly degenerate with the
removed long radial modes of the continuum foregrounds. The high signal loss
at low-$k_{\perp}$ is interesting and not seen in previous MeerKLASS work
(Cunnington, Li et al., 2022) with a similar method, albeit with less robust
mocks. This is likely caused by the small GAMA footprint (see Figure 4) which
means the large-ranging modes at these low-$k$ are not adequately resolved and
deliver little signal which is easily coupled to foreground modes and stripped
away in the clean. This can be explored further in follow-up work with wider
surveys where we expect this not to be a problem.
Due to the noisy power at low-$k$ and damped power at high-$k_{\perp}$ in
Figure 13, we can choose to cut these regions of $k$-space from our
spherically averaged power to obtain a more robust result. The orange solid
lines represent chosen boundaries, where modes beyond these boundaries are
excluded in the averaging of each $k$-bin (shown by the dashed gray lines).
Providing the modelling of any power spectra also incorporates the $k$-cuts,
then it will not bias any results. The $k$-cuts approach is further motivated
by considering the signal-to-noise of the cross-power which we show in the
third panel of Figure 13. Here the errors
$\sigma(\hat{P}_{\textrm{H{i}},\text{g}})$ are computed using the transfer
function by taking the standard deviation from the
$\delta\hat{P}_{i}^{\text{rec}}(k_{\perp},k_{\parallel})$ distribution (see
previous Section 5.3). This result provides a strikingly clear window where
signal-to-noise is optimal; the orange to red regions in the third panel. The
orange $k$-cut boundary lines have been chosen to target this region. One
limitation of this approach is that it excludes the regions of $k$-space with
the most modes, as shown by the fourth and final panel. We will present
comparisons between the spherically averaged $P(k)$ with and without these
$k$-cuts in the upcoming Section 5.6.
### 5.5 Hi auto-correlation power spectrum
We next examine results for the auto-correlation Hi power spectrum from the
MeerKLASS L-band deep field. Previous single-dish intensity mapping results
have been thermal noise-dominated across all scales. We are now crossing an
important threshold with MeerKLASS where the Hi fluctuations should be
dominating the noise at large scales. We demonstrate this in Figure 14, where
we show the estimated Hi power spectrum from the MeerKLASS deep field (blue
data points) and overlay an estimate of the thermal noise (red solid line). As
outlined in Section 3.2, this noise is modelled by generating Gaussian random
fields with the expected rms given by Equation 20 which utilises the
observational hit map. Despite being the Hi auto-power, we still restrict the
measurement to the GAMA-field footprint, for consistency with the cross-
correlation results. We tested using the full field but found no improvement
in the auto-power. We leave exploration of the full-field to follow-up work
which focuses on the Hi-auto power spectrum. The results in Figure 14 include
the scale $k$-cuts discussed in the previous section.
Figure 14 also shows the expected Hi power (grey dotted line) from Equation
47. For this auto-Hi model, we do not fit the amplitude to avoid additive bias
from residual systematics over inflating the model estimate. Instead we use
the scale-independent $\overline{T}_{\textrm{H{i}}}$ fit from the upcoming
cross-power measurement. This model shows how the cosmological Hi power
surpasses the thermal noise on large scales
($k\,{\lesssim}\,0.15\,h\,\text{Mpc}^{-1}$). The thermal noise will be present
even in the absence of clustering, thus the total expected power can be
reasonably approximated as a simple addition of the Hi power, plus the thermal
noise (shown by the black dashed line).
Figure 14: The Hi auto-power spectrum for the MeerKLASS L-band deep-field
foreground cleaned by removing $N_{\text{fg}}\,{=}\,10$ PCA modes. The red
solid line is the noise estimate from measuring the average power of 10
Gaussian random fields with rms given by $\sigma_{\text{N}}$ (Equation 20).
The grey dotted line is the observational Hi power spectrum model (Equation
47) and the black dashed line is the combination of this Hi model plus the
thermal noise estimate.
The small discrepancy between the deep-field power estimate and the model
represents the lowest upper limit from a Hi auto-correlation on these large
scales. We find the mean residual factor between model and data to be just
$2.36$ across all $k$. Whilst the model and data are close in agreement,
inferred parameter values for the bias and mean Hi temperature from fitting
the amplitude of the auto power yields inconsistent values with the GAMA
cross-correlation (discussed next). This is therefore evidence of low-level
residual systematics that are additively biasing the power. In this
evaluation, we inherently assume that the transfer function reconstructs any
signal loss in the Gaussian thermal noise. If signal loss in the Gaussian
noise is underestimated, the residual between the model and data may be higher
than we report here. We therefore intend to robustly test this assumption when
pursuing a more dedicated study of the Hi auto-correlation.
In follow-up work, we will focus on improving this upper limit from the Hi
auto power which can be done in two ways: Firstly, by implementing some
corrections to remove low-level systematic contributions such as the residual
RFI structures discussed in Section 2.4.2. Secondly, we can split the
observations into sub-sets i.e. combinations of different maps from different
scan times and/or dishes. This is useful for isolating dish- and time-
dependent systematics and removing their biasing contributions to the power.
Since each scan and dish is observing the same patch of sky, we would ideally
expect all maps to agree, thus any differences will represent thermal noise,
or systematics such as residual RFI, varying polarisation leakage etc. We do
not pursue these ideas in this paper since they warrant their own detailed
investigation.
### 5.6 Cross-correlation power with GAMA
We now focus on the cross-correlation with the GAMA galaxies for the remainder
of the paper. In Figure 15 we present the cross-correlation power spectrum
between the GAMA galaxies and the MeerKLASS deep field. All points are
positive so none are lost to the log scaling, indicating a strong, positive
correlation. Generally, a good agreement with the model is obtained. We
performed a null test on the result by substituting the GAMA mock galaxies
(from Section 3.3) in place of the real galaxies. The mocks should not
correlate with the intensity maps, and indeed, the resulting cross-power
converges to zero once averaged over a sufficient number of mocks. The fitted
amplitude of the model is equivalent to an inferred parameter of
$\overline{T}_{\textrm{H{i}}}\,{=}\,0.166\,\text{mK}$, with all other values
in Table 1 assumed. We discuss this further in Appendix C, but do not focus on
parameter inference in detail due to the large statistical noise owing to the
small overlap footprint between MeerKLASS and GAMA.
Figure 15: The cross-correlation power spectrum between GAMA galaxies and the
MeerKLASS L-band deep field, foreground cleaned by removing
$N_{\text{fg}}\,{=}\,10$ PCA modes. The black dashed line represents a model
given by Equation 49 with a fitted scale-independent amplitude.
The error bars for the cross-power in Figure 15 are obtained using the scatter
from the foreground transfer function as discussed in Section 5.3. The error
size appears to be over-predicted based on the scatter of the data points,
which have a reasonably tight agreement with the model relative to what the
error bars reflect. This is supported by examining the reduced $\chi^{2}$
which we find to be
$\chi^{2}_{\text{dof}}\,{\equiv}\,\chi^{2}{/}N_{\text{dof}}\,{=}\,0.42$. A
$\chi^{2}_{\text{dof}}\,{<}\,1$ is conventionally indicative of overinflated
errors. Using the reduced $\chi^{2}$ however is not a reliable test where
there are non-linear contributions, as is likely the case for these
observations that are still prone to residual systematics. In these cases, the
number of degrees of freedom cannot be robustly determined (Andrae et al.,
2010). For our case we simply assume $N_{\text{dof}}\,{=}\,13$, for our 14
$k$-bins minus a single fitted degree of freedom, representing the amplitude
fit. Furthermore, noise within the data causes high uncertainty in $\chi^{2}$
estimates, which despite the relatively low noise levels we are now reaching,
may still be a concern. Alternatives to $\chi^{2}$ have been proposed, and
specifically for 21cm applications (e.g. Tauscher et al., 2018), but these
have not been pursued in this work. We only use the $\chi^{2}$ as an
approximate guide, mainly for relative comparisons between different cases.
For the estimation of $\chi^{2}$, we include the full contribution from off-
diagonals in the $k$-bin covariance C, which can be trivially estimated for
our error estimation method, simply the covariance over
$\hat{P}_{i}^{\text{rec}}(k)\,{=}\,\hat{P}_{\text{rec}}(k)\,{+}\,\delta\hat{P}^{\text{rec}}_{i}(k)$
(see Section 5.3). The $\chi^{2}$ is then estimated from
$\chi^{2}=(\hat{\boldsymbol{p}}-\boldsymbol{p}_{\rm
mod})^{\text{T}}\,\textbf{{C}}^{-1}\,(\hat{\boldsymbol{p}}-\boldsymbol{p}_{\rm
mod})\,,$ (55)
where $\hat{\boldsymbol{p}}$ is the data vector and
$\boldsymbol{p}_{\text{mod}}$ the corresponding model. We present the
normalised covariance matrix (or correlation matrix) in Figure 16, showing
evidence that the $k$-bins are slightly correlated, especially at low-$k$.
This could be contributing to the large error bars, hence why we include the
full covariance in the $\chi^{2}$ estimation.
Figure 16: $k$-bin correlation matrix for the cross-power between the
MeerKLASS deep field and GAMA galaxies, shown in Figure 15.
To investigate the errors from the cross-power further, we plot the full
distributions for the reconstructed power spectra $\hat{P}_{i}^{\text{rec}}$
from each $i$th mock in the foreground transfer function computation. Figure
17 shows this for each $k$-bin, including an additional low-$k$ bin which we
otherwise exclude due to the large error on these scales. We centre the
distributions on the model and normalise by it, to aid comparison with the
case of no $k$-cuts (red results), which includes the areas of $k$-space with
low signal-to-noise (discussed in Section 5.4). The plot shows how the
distributions have some non-Gaussian features, which could also impact error
estimation, along with the correlated $k$-bins shown by the covariance. We
find the average kurtosis across all $k$-bins for the $k$-cut case to be
$-0.476$, indicating the distribution is flatter than a normal distribution
with more contribution in the tails. This could also widen the central
percentiles and thus increase the error bar size.
Figure 17: Estimated power distribution for each $k$-bin from the cross-
correlation between MeerKLASS intensity maps and GAMA. This is computed using
the power distribution in the 500 mocks used in the foreground transfer
function. $k$-bin value is given in the top-left corner of each panel. The
power distributions are centred on the fitted model at each $k$ so a zero
median in the distribution represents perfect accuracy. Shaded regions show
the 68th percentiles for the distributions which are used as the limits for
the $1\,\sigma$ error bars in the power spectra plots.
Figure 17 shows a mild improvement in performance for the case with $k$-cuts.
The distributions around the model are less broad at low-$k$, which is
expected since this was one of the regions where signal-to-noise was poorest
(see third panel of Figure 13). In terms of the Gaussianity of the
distributions, we find a better kurtosis score for the case with $k$-cuts,
suggesting the profiles are more Gaussian. In contrast, we find the results
without $k$-cuts to be less skewed, although this is marginal and in general,
the skewness of the distributions is not concerning. Any improvement coming
from the $k$-cuts is small and we also find little difference in detection
significance, as we will explore shortly. As the volume of data grows beyond
this small overlap sample, the $k$-cut results may prove more optimal since we
will no longer be cutting modes from an already sample variance-limited data
set. This is something that future data sets with MeerKLASS will reveal.
#### 5.6.1 Varying the foreground clean
So far we have only considered the chosen case of removing
$N_{\text{fg}}\,{=}\,10$ PCA modes for the foreground clean. Figure 18 shows
how the $\chi^{2}$ (which we continue to use as a rough goodness-of-fit
metric) varies with $N_{\text{fg}}$. As seen in previous work (Cunnington, Li
et al., 2022), too low an $N_{\text{fg}}$ does not sufficiently clean the
foregrounds, causing a poor fit. Going too high with $N_{\text{fg}}$ also
diminishes the fit. Ideally, we would expect high $N_{\text{fg}}$ to cause
more signal loss which should be reconstructed by the transfer function,
giving unbiased results and thus not diminishing the fit. A higher uncertainty
could be expected with greater signal loss, but this should be accounted for
by the increasing error bars i.e. the $\chi^{2}$ should stabilise and not
start to rise again at high $N_{\text{fg}}$. On further scrutiny, we found
that it is a single data point at $k\,{\sim}\,0.2\,h\,\text{Mpc}^{-1}$ with
very little power distorting the fits at higher $N_{\text{fg}}$. This is
evidence of low-level residual systematics that become more influential as
signal is stripped away in the more aggressive foreground clean. Therefore,
with future results where systematics are more controlled, this is expected to
be less of a problem. For the case with no $k$-cuts (thin dashed lines), there
is a plateau in $\chi^{2}$ at higher $N_{\text{fg}}$, as we argued should be
the expected situation. This is because the no $k$-cut power spectra were less
impacted by the $k\,{\sim}\,0.2\,h\,\text{Mpc}^{-1}$ anomaly at high
$N_{\text{fg}}$ which is potentially suggesting that the $k$-cuts are overly
aggressive for this scale. With larger overlapping fields, the more accessible
$k$-bins will naturally mitigate this discrepency. The no $k$-cut results in
Figure 18 are therefore encouraging that a model and data fit should reach a
plateau for high $N_{\text{fg}}$ as expected with the current analysis
pipeline.
Figure 18: Variation of cross-correlation results with changing
$N_{\text{fg}}$ i.e. the number of PCA modes removed in the foreground clean.
The left (black) axis shows the $\chi^{2}$, the lower values generally
represent a better fit between cross-power data and model. The right (red)
axis shows the cross-correlation power spectrum detection significance. Solid
and thin-dashed lines represent the case with and without $k$-cuts
respectively.
The right hand side (red) $y$-axis in Figure 18, also shows the detection
significance defined by
$\sqrt{\Delta\chi^{2}}\,{\equiv}\,\sqrt{\chi^{2}-\chi_{\text{null}}^{2}}$,
where $\chi^{2}_{\text{null}}$ is the result’s $\chi^{2}$ with a null model
with zero power. The cross-correlation with GAMA is returning a modest
${>}\,4\sigma$ detection for the most optimal choices of foreground clean,
lower than previous MeerKLASS detections (Cunnington, Li et al., 2022). This
is however still impressive, given that galaxies only overlap with
approximately a quarter of the full MeerKLASS deep field. The detection
significance is also restricted by the more robust approach to error
estimation, which as we have already discussed, provides conservative
estimates potentially inflated by residual systematics and correlated
$k$-bins. To demonstrate this point, we artificially construct a scenario
where a $\chi^{2}_{\text{dof}}\,{\sim}\,1$ is obtained for the
$N_{\text{fg}}\,{=}\,10$ case. This is achieved by assuming a perfectly
diagonal covariance and decreasing the diagonal errors by $40\%$. This boosts
the detection significance to $9.8\,\sigma$, verifying that it is the
conservative approach to errors lowering the detection significance, rather
than the data-model residual. This is of course only a toy demonstration but
it validates the argument that if large errors, which are suspected to be
driven by systematics and $k$-bin correlations, can be contained, then
detection robustness should improve. This is expected in future work where we
will pursue dedicated studies into systematic mitigation. Some mode coupling
is always likely, driven at least by non-linear clustering. However, we are
confident that improvements can be made in covariance. MeerKLASS observations
ongoing now are covering wider areas with more overlap with Stage-IV optical
galaxy surveys. These observations are also using MeerKAT’s UHF-band, which
will allow a deeper usable redshift range, extending the
$0.39\,{<}\,z\,{<}\,0.46$ used in this work to approximately
$0.4\,{<}\,z\,{<}\,1.4$. This will naturally improve the large statistical
noise due to the small volume overlap in this work.
## 6 Stacking
In this section, we explore the possibility of a stacking detection of the Hi
signal in the MeerKLASS L-band deep-field intensity map. The technique of
spectral line stacking (Zwaan et al., 2001) can be used to detect the Hi
content of galaxies below noise limit. In combination with spectroscopic
galaxy surveys, spectral line stacking probes the average emission line
profile of selected galaxy samples. Complimentary to the intensity mapping
technique, the stacking measurement of the emission line profile probes the Hi
mass function of galaxy samples (e.g. Bera et al. 2022), the relation of Hi
content with its host environment (e.g. Guo et al. 2020; Bera et al. 2023; Dev
et al. 2023; Rhee et al. 2023), and the baryonic Tully-Fisher relation (e.g.
Meyer et al. 2016). Interferometric observations of MeerKAT using the MeerKAT
International GigaHertz Tiered Extragalactic Exploration survey (MIGHTEE;
Jarvis et al. 2016; Maddox et al. 2021) have been used to produce Hi
measurements with regard to the scaling relation (Sinigaglia et al., 2022)
with galaxy properties and the dependency on the large-scale structure
environment (Sinigaglia et al., 2024). See also Ponomareva et al. (2021);
Tudorache et al. (2022); Pan et al. (2023); Ponomareva et al. (2023) for
direct detections of Hi galaxies in the MIGHTEE survey.
For intensity mapping experiments, stacking analysis has been performed on the
Hi intensity maps using the Parkes telescope (Anderson et al., 2018; Tramonte
et al., 2019; Tramonte & Ma, 2020) and the CHIME telescope (CHIME
Collaboration et al., 2023), as well as on CO intensity maps from the CO
Mapping Array Project (Dunne et al., 2024) and Ly$\alpha$ emission of [O
III]-selected galaxies from the Hobby-Eberly Telescope Dark Energy Experiment
(Lujan Niemeyer et al., 2022a, b).
MeerKLASS, on the other hand, faces unique challenges. The $\,{\sim}\,1\,$deg
instrument power beam introduces heavy contamination from nearby sources. For
the GAMA galaxy sample used, the number density of the sources is
$\,{\sim}\,35\,{\rm deg^{-2}}$ when integrated along the frequency sub-band.
Therefore, when stacking each source, the stacking procedure will inevitably
include emissions from more than one GAMA galaxy. The velocity width of the Hi
sources further scatter the 21 cm emission into nearby channels. The double-
counting of sources from the beam creates a plateau of excess signal as shown
in Figure 26, and can be resolved by PCA cleaning, which we discuss in
Appendix D. However, the residual after PCA cleaning also incorporates
systematic effects. The chromatic structure of the systematic effects produces
oscillations in the stacked profile, with their amplitudes comparable to the
signal. For this work, we are only interested in using stacking as part of the
validation test. As discussed later, we obtain a detection of a stacked signal
in both angular and spectral space, which can be validated by performing a
null test with random galaxy positions. The existence of a stacked signal
further demonstrates that we have reached the depth needed for the possible
detection of the Hi power spectrum using the MeerKLASS L-band intensity maps.
We leave a more detailed study into refining the stacking technique for
probing Hi science to follow-up work.
### 6.1 The stacking procedure of the intensity maps
We use the PCA-cleaned map described in Section 4 and the positions of the
GAMA galaxies to perform a 3D stacking in
$(\delta\alpha,\delta\phi,\delta\nu)$ space so that
$I(\delta\alpha,\delta\phi,\delta\nu)=\frac{\sum_{i}I(\alpha_{i}{+}\delta\alpha,\,\phi_{i}{+}\delta\phi,\,\nu_{i}{+}\delta\nu)w_{i}}{N_{\rm
g}\sum_{i}w_{i}},$ (56)
where $(\alpha_{i},\phi_{i})$ are the right ascension and declination of the
map pixel in which the $i^{\rm th}$ galaxy resides, $I$ is the flux density,
$N_{\rm g}$ is the number of galaxies, and
$w_{i}\,{=}\,w_{\textrm{H{i}}}(\alpha_{i}{+}\delta\alpha,\phi_{i}{+}\delta\phi,\nu_{i}{+}\delta\nu)$
is the reconvolved weight of each pixel defined in Section 4.1. We choose the
angular and the spectral bins to be consistent with the map resolution with
$\Delta\alpha\,{=}\,\Delta\phi\,{=}\,0.3\,$deg. The frequency resolution
corresponds to an effective velocity resolution of $\Delta v\,{=}\,62.83\,$km
s-1 at the centre frequency of the sub-band $997.37\,$MHz. Note that no
Doppler correction has been performed on the data. For our analysis, the
intensity maps are coherently averaged in frequencies for PCA cleaning and
therefore can not be averaged in velocity channels.
In order to calculate the statistical noise for the stacked cube, we perform
the same stacking routine on random galaxy positions, similar to the shuffling
null test in Section 5.6. The positions are randomly sampled uniformly within
the region of the GAMA galaxy sample. The redshifts are randomly sampled
following the redshift distribution of the survey. The random process is
repeated for 100 realisations to calculate the mean and the standard deviation
of the reference cube. The reference cube is used for performing the null
test, i.e. checking that the stacked emission signal in the reference cube is
consistent with zero.
After obtaining the stacked cubes, we create the angular and spectral stacking
by summing over the cubes along the radial and transverse directions. To
control the systematics generated by the foreground cleaning, we limit the
voxels that go into the summation. For angular stacking, only voxels that are
within 150 km s-1 of the central channel are considered, corresponding to the
typical width of the HI emission line profile. For spectral stacking, only the
voxels whose centre positions are within the $1\,$deg area of the centre voxel
in each channel are included, corresponding to the size of the primary beam.
The same summation is performed on the random realisation to calculate the
mean and the standard deviation of the reference image and spectrum. We note
that there is signal loss due to an insufficient number of voxels included.
However, as the random realisations follow the same procedure, it does not
affect the validation of the detection which is our focus in this work.
Figure 19: Left panel: The angular stacking of the GAMA galaxies on the Hi
intensity map. The dashed circle denotes the FWHM of the beam. Central panel:
The reference null detection image from 100 realisations of random galaxy
positions. Right panel: The standard deviation of the reference image
multiplied by 3 to visualise the 99.7% confidence interval.
### 6.2 The detection of stacked Hi signal
In Figure 19, we show the results for the angular stacking of the GAMA
galaxies on the Hi intensity map. The stacked map shows a clear detection at
the centre of the image, with the centre pixel achieving a detection with a
statistical significance of $8.35\,\sigma$. The stacked Hi emission follows
the structure of the beam. The extended structure of emission outside the beam
FWHM is a result of the relatively large thermal noise fluctuation and
contamination from nearby sources in the stacking procedure as discussed
above. We note that, as shown in Figure 4, the lower part of the GAMA galaxy
sample region sits near the edge of the observed intensity maps. As a result,
the lower half of the stacked image has a higher level of statistical
fluctuations as shown in the right panel of Figure 19. Furthermore, the
reference image is consistent with a null detection, demonstrating that the
stacked signal does not arise from systematic effects in the map.
Figure 20: The stacked spectrum of the GAMA galaxies on the Hi intensity map.
The blue solid line marks the measured stacked spectrum. The orange dashed
line denotes the reference spectrum from 100 random realisations, and the
shaded region denotes the standard deviation of the realisations. The black
dotted line denotes the $5\,\sigma$ limit of the reference spectrum rms and is
smoothed with a Gaussian kernel of 300 km/s for better illustration.
In Figure 20, we show the stacked spectrum along the line-of-sight. The peak
of the stacked spectrum shows a detection with a significance of
$6.30\,\sigma$. The detection is validated by the reference spectrum which is
consistent with a null detection. However, as seen in the reference spectrum,
there is a systematic noise component with fluctuations on large frequency
intervals. The systematics can be most visibly seen at the peaks and troughs
of the stacked signal at $v\,{\sim}\,2500,\,4500,\,$ and $7000\,$km/s. We have
identified the source of this fluctuation as the residual systematics in the
map after PCA cleaning, which is demonstrated in Appendix D. As stated
previously, we are only interested in validating the detection against the
null test and leave the suppressing of the systematic effects for future work.
The stacking results show a clear detection of Hi signal in the positions of
the GAMA galaxies. The detection of an emission line signal is statistically
significant against the null test. At the same time, the residual systematics
in the data, while being sub-dominant comparing to the Hi signal, induce
structures in the stacked spectrum. We intend to follow up on the findings in
this work and develop analysis techniques to extract the underlying emission
line signal in the future, as well as using the stacked signal to understand
the residual systematics.
## 7 Conclusions
In this work we have presented L-band observations from the MeerKLASS
intensity mapping campaign. The survey covered a single patch of approximately
$200\,\deg^{2}$, repeatedly scanned 41 times (27 scans remained after
aggressive RFI flagging), which we have referred to as the MeerKLASS L-band
deep field. The observations used in our analysis came from only a small
section of the full L-band ($971.2\,{<}\,\nu\,{<}\,1023.6$ at redshift
$0.39\,{<}\,z\,{<}\,0.46$) to avoid RFI which we find is more present in this
band relative to the lower frequency UHF-band ($580\,{<}\,\nu\,{<}\,1000$)
which all future MeerKLASS surveys will be conducted in. We detailed the
calibration strategy used on these observations which included a new approach
of self-calibration (Section 2.3), exploiting the data’s higher signal-to-
noise to devise a less model-dependent calibration. The improvement in data
quality relative to previous MeerKLASS pilot surveys is evident (see e.g.
Figure 5) and encouraging for future MeerKLASS surveys gathering ever-larger
data sets.
We demonstrated the enhanced signal-to-noise from the calibrated deep-field
intensity maps by presenting the auto-correlation power spectrum (Figure 14)
which revealed the lowest upper-limit from a Hi power spectrum at these
cosmological scales $k\,{<}\,0.3\,h\,\text{Mpc}^{-1}$. This remains a factor
of ${\sim}\,2$ above where we would expect the Hi power to lie. We will pursue
this limit further in follow-up work where we will explore a systematic
removal strategy to reduce the quantity of flagged data plus the cross-
correlation of divided data sets to mitigate the additive bias from time-
dependent residual systematics.
We also presented results from the cross-correlation power spectrum between
the MeerKLASS deep-field and 2269 overlapping GAMA galaxies with spectroscopic
redshifts. Despite the small field of overlap (${\sim}25\%$ of the deep field
patch), we still obtained a clear ${>}\,4\,\sigma$ detection of the cross-
power spectrum (Figure 15). Furthermore, we presented a detection of Hi
emission from stacking the MeerKLASS maps onto the positions of the GAMA
galaxies. This yielded a $8.35\,\sigma$ and $6.30\,\sigma$ detection from
angular and spectral stacking respectively, the first stacking detections with
MeerKAT in single-dish mode (see Section 6).
We implemented several novel strategies into the power spectrum analysis
pipeline, including a selective procedure to target the region of $k$-space
with the highest signal-to-noise. The third panel of Figure 13 demonstrated
the clear window of Fourier modes where power dominates over the error. By
implementing some minima and maxima cuts on $k_{\perp}$ and $k_{\parallel}$ we
were able to exclude regions of $k$-space dominated by error. This provides a
more robust result since error-dominated modes will not be included in final
spherical averaging. Comparison of the error profiles for each $k$-bin (Figure
17) demonstrated a mild improvement in terms of Gaussianity when using the
$k$-cuts, however, the sacrifice of many modes contained at high-$k$ could be
restricting this improvement. We plan to investigate this strategy in future
work where larger data sets will reveal the optimal approach.
We formalised our approach to covariance estimation for the power spectra
measurements which utlises the foreground transfer function distribution from
each mock realisation. This distribution will be determined by all foreseeable
sources of error; residual systematics, thermal noise fluctuations, galaxy
shot-noise, cosmic variance and any uncertainty in signal-loss correction.
This should therefore be a robust method for error estimation. However, we
found the errors in the spherically averaged cross-power spectrum appear to be
large relative to their scatter about the model. This was supported by a
$\chi^{2}$ analysis. Further inspection revealed that the $k$-bins are mildly
correlated and the error profiles for each $k$-bin are not perfectly Gaussian,
both of which could be inflating the errors. We hypothesise that this should
be naturally mitigated with future MeerKLASS analysis where systematics can be
more reliably controlled.
This work also presented validation tests of our improved analysis pipeline.
|
# Bredon homological stability for configuration spaces of $G$-manifolds
Eva Belmont Case Western Reserve University<EMAIL_ADDRESS>, J.D.
Quigley University of Virginia<EMAIL_ADDRESS>and Chase Vogeli
Cornell University<EMAIL_ADDRESS>
###### Abstract.
McDuff and Segal proved that unordered configuration spaces of open manifolds
satisfy homological stability: there is a stabilization map
$\sigma:C_{n}(M)\to C_{n+1}(M)$ which is an isomorphism on
$H_{d}(-;\mathbb{Z})$ for $n\gg d$. For a finite group $G$ and an open
$G$-manifold $M$, under some hypotheses we define a family of equivariant
stabilization maps $\sigma_{G/H}:C_{n}(M)\to C_{n+|G/H|}(M)$ for $H\leq G$. In
general, these do not induce stability for Bredon homology, the equivariant
analogue of singular homology. Instead, we show that each $\sigma_{G/H}$
induces isomorphisms on the ordinary homology of the fixed points of
$C_{n}(M)$, and if the group is Dedekind (e.g. abelian), we obtain the
following Bredon homological stability statement: $H^{G}_{d}(\bigsqcup_{n\geq
0}C_{n}(M))$ is finitely generated over $\mathbb{Z}[\sigma_{G/H}:H\leq G]$.
This reduces to the classical statement when $G=e$.
###### Contents
1. 1 Introduction
2. 2 Equivariant homotopy theory
3. 3 Configuration spaces and stabilization maps
4. 4 Equivariant homological stability
5. 5 Generalizations and variants
## 1\. Introduction
Let $X$ be a topological space and let $C_{k}(X)$ denote the configuration
space of $k$ unordered points in $X$. If $X$ is the interior of a manifold
with nonempty boundary, then there are stabilization maps
(1.0.1) $\sigma:C_{k}(X)\to C_{k+1}(X)$
defined by “adding a point near the boundary" (cf. Section 3.1). In the
1970’s, McDuff [McD75] and Segal [Seg79] observed that these stabilization
maps induce isomorphisms in a range of integral homology groups:
###### Theorem 1.0.2 (McDuff–Segal, Stability for Configuration Spaces
(Strong Form)).
Assume $M$ is the interior of a manifold with nonempty boundary. The
stabilization map (1.0.1) induces an isomorphism
$\sigma_{*}:H_{d}(C_{k}(M))\to H_{d}(C_{k+1}(M))$
for $d\leq k/2$.
### 1.1. Statement of main results
In this paper, we investigate an equivariant analogue of the McDuff–Segal
theorem where the manifold $M$ is replaced by a $G$-manifold (where $G$ is a
finite abelian group) and singular homology is replaced by Bredon homology.
The Bredon homology of a $G$-space $X$ with coefficients in the constant
Mackey functor $\underline{\mathbb{Z}}$, written
$H^{G}_{*}(X;\underline{\mathbb{Z}})$, is an equivariant analogue of singular
homology with integer coefficients. We recall this definition of Bredon
homology, as well as its representation by a $G$-spectrum, in Section 2.1.
If $X$ is a $G$-space, then the space of unordered configurations of $k$
points in $X$
$C_{k}(X):=\\{(x_{1},\ldots,x_{k})\in X^{k}:x_{i}\neq x_{j}\text{ for all
}i\neq j\\}/\Sigma_{k}$
is also a $G$-space with $G$ acting diagonally. If $X$ is a $G$-manifold which
is the interior of a $G$-manifold with nonempty boundary containing a fixed
point, then we can define $G$-equivariant stabilization maps
(1.1.1) $\sigma_{G/G}:C_{k}(X)\to C_{k+1}(X)$
by “adding a fixed point near the boundary” (see Section 3.3). In this
situation, we say $X$ is _$G$ -stabilizable_ (Definition 3.3.1). We can then
ask if these maps induce isomorphisms in Bredon homology for $k$ sufficiently
large. Surprisingly, this is not the case:
###### Theorem A (Failure of Strong Equivariant Form of McDuff–Segal,
Proposition 4.2.1).
Let $G=C_{p}$ and let $M$ be a $G$-manifold which is $G$-stabilizable. The map
(1.1.2) $(\sigma_{G/G})_{*}:H_{0}^{G}(C_{n}(M);\underline{\mathbb{Z}})\to
H_{0}^{G}(C_{n+1}(M);\underline{\mathbb{Z}})$
is not surjective for any $n\geq 1$.
Although the strong form of the McDuff–Segal Theorem fails equivariantly,
there is a weaker form which we can extend to the equivariant setting. Let
$C(X):=\bigsqcup_{k\geq 0}C_{k}(X)$
denote the space of all unordered configurations of finitely many points in
$X$ equipped with the disjoint union topology. The nonequivariant
stabilization maps (1.0.1) give rise to a self-map
(1.1.3) $\sigma:C(X)\to C(X)$
which allows us to state the following corollary of the McDuff–Segal Theorem:
###### Theorem 1.1.4 (McDuff–Segal, Nonequivariant Stability (Finite
Generation Form)).
Assume $M$ is an open, connected manifold with $\dim(M)\geq 2$. Moreover,
suppose that $H_{d}(M)$ is finitely generated as an abelian group for all
$d\geq 0$. Then the homology group $H_{d}(C(M))$ is finitely generated as a
$\mathbb{Z}[\sigma_{*}]$-module for all $d\geq 0$.
###### Remark 1.1.5.
The assumption that $H_{d}(M)$ is finitely generated ensures that
$H_{d}(C_{k}(M))$ is finitely generated for all $k\geq 0$ (Proposition 4.1.5),
so the finite generation form of nonequivariant stability is equivalent (cf.
Proposition 4.1.2) to saying that the map (1.0.1) is an isomorphism for all
$k\gg 0$.
We note that the philosophy of encoding stability as a finite generation
condition is not new. It has become increasingly prevalent in the literature
on homological stability (e.g., [ADCK20, Theorem 1.1]) and representation
stability (e.g. [CEF15, Theorem 6.2.1]).
The obvious finite generation form of equivariant stability fails since the
equivariant stabilization map (1.1.2) does not induce isomorphisms for all
$k\gg 0$. In other words, $H_{d}^{G}(C(M);\underline{\mathbb{Z}})$ is _not_
finitely generated as a $\mathbb{Z}[(\sigma_{G/G})_{*}]$-module.
However, there is a naturally occurring polynomial ring over which
$H_{d}^{G}(C(M);\underline{\mathbb{Z}})$ is a finitely generated module. If
$X$ is a $G$-manifold which is the interior of a $G$-manifold with boundary
containing an orbit of the form $G/H$, then we can define a new equivariant
stabilization map
(1.1.6) $\sigma_{G/H}:C_{k}(M)\to C_{k+|G/H|}(M)$
by “adding an _orbit_ of type $G/H$ near the boundary.” In this situation, we
say $X$ is _$H$ -stabilizable_ (Definition 3.3.1).
We adopt the philosophy that since orbits are equivariant analogues of points,
the equivariant analogue of stabilization maps should be the family of
stabilization maps $\sigma_{G/H}$. As such, we replace the ring
$\mathbb{Z}[\sigma_{*}]$ by the ring
$P_{G}=\mathbb{Z}[(\sigma_{G/H})_{*}:(H)\leq G]$
of all equivariant stabilization maps. By considering all possible equivariant
stabilization maps, we obtain our stability result.
###### Theorem B (Bredon Homological Stability, Theorem 4.5.5).
Let $G$ be an abelian group and $M$ a $G$-manifold that is the interior of a
compact $G$-manifold with boundary. Assume that for all $H\leq G$, $M$ is
$H$-stabilizable and $M^{H}$ is connected. Then
$H_{d}^{H}(C(M);\underline{\mathbb{Z}})$ is finitely generated over $P_{G}$
for all $H\leq G$.
For example, these hypotheses are satisfied in the case where $M$ is a sum of
copies of the regular representation (see Example 4.5.6).
###### Remark 1.1.7.
See Lemma 4.4.2 for the main reason for the abelian hypothesis, and Remark
4.4.3 for a mild relaxation of this hypothesis. Group-theoretic hypotheses are
also used in Lemma 4.5.1 and 4.5.3.
###### Remark 1.1.8.
We discuss two natural extensions of B in Section 5: in Theorem 5.1.3, we show
that stability holds for $RO(G)$-graded Bredon homology, and in Theorem 5.2.5,
we show that stability holds for Mackey functor-valued Bredon homology. The
$RO(G)$-graded Bredon homology of _ordered_ configuration spaces of
$G$-manifolds will also be studied in forthcoming work of Dugger and Hazel
[DH].
### 1.2. Motivation
Before discussing the proofs of A and B, we discuss some motivating ideas and
directions for future work.
#### 1.2.1. Equivariant loop spaces and Dyer–Lashof operations
Nonequivariantly, May’s recognition principle [May72] implies that $n$-fold
loop spaces are equivalent to algebras over the $E_{n}$-operad
$\mathcal{E}_{n}$ whose $k$-th space is $C_{k}(\mathbb{R}^{n})$. Using Cohen’s
computation of $H_{*}(C_{k}(\mathbb{R}^{n}))$ in [CLM76, Ch. III], one can
then define Dyer–Lashof operations on the mod $p$ homology of any $n$-fold
loop space.
In [GM17], Guillou and May proved an equivariant analogue of the recognition
principle: if $V$ is a real orthogonal $G$-representation, then $V$-fold loop
spaces are equivalent to algebras over the $E_{V}$-operad $\mathcal{E}_{V}$
whose $k$-th space is $C_{k}(V)$. They remark that in contrast with the
nonequivariant situation, very little is known about $H_{d}^{G}(C_{k}(V))$. B
gives the first step toward a more systematic analysis of these Bredon
homology groups which could shed light on equivariant Dyer–Lashof operations
(cf. [Wil19] for $G=C_{2}$). We verify that the hypotheses of B are satisfied
in the case of sums of regular representations in Example 4.5.6.
#### 1.2.2. Equivariant diffeomorphism groups of smooth $G$-manifolds
Configuration spaces of ordinary manifolds appear in the study of
diffeomorphism groups of smooth manifolds via the “decoupling theorems" of
Bödigheimer–Tillmann [BT01] and Bonatto [Bon22]. It would be interesting to
see if the configuration spaces of $G$-manifolds arise analogously in the
equivariant diffeomorphism groups of smooth $G$-manifolds.
### 1.3. Overview of key ideas
We now summarize the key ideas going into the proofs of A and B.
#### 1.3.1. Equivariant stabilization maps
A map between $G$-spaces induces a map on Bredon homology only if it is
$G$-equivariant, so we must be careful about how we add points. One natural
option is to add a _fixed_ point near infinity (provided our space has such a
fixed point), which yields the equivariant stabilization map (1.1.1) mentioned
above with underlying map (1.0.1). However, A shows that this map does not
produce isomorphisms in Bredon homology after iteration.
One of our key observations is that under suitable hypotheses on the manifold
$M$, there is a stabilization map (1.1.2) for each subgroup $H\leq G$. B says
that the Bredon homology stabilizes when all of these equivariant
stabilization maps are taken into account simultaneously.
###### Remark 1.3.1.
The stabilization map corresponding to $H\leq G$ inserts an orbit of the form
$G/H$, i.e., inserts $|G/H|$ points at once. Other stabilization maps which
add multiple points at once have appeared in the context of higher order
representation stability [MW19].
#### 1.3.2. Formulating stability in terms of finite generation
While none of the orbit stabilization maps (1.1.2) yield Bredon homological
stability for a fixed $H$ (A), the collection of maps $\\{\sigma_{G/H}:(H)\leq
G\\}$ yield a form of stability (B). To make this precise, we translate
homological stability from a statement about maps in homology to a statement
about finite generation over a polynomial ring (Proposition 4.1.2).
#### 1.3.3. Reduction from Bredon homology to singular homology of fixed
points
The analogue of cellular homology in the context of $G$-spaces is _Bredon
homology_ , $H^{G}_{d}(-)$ (cf. [May96, XIII.4.1], Section 2.1). The Bredon
homology of a $G$-space is not simply the homology of its fixed points, but in
Section 2 we use equivariant stable homotopy theory to relate these notions.
The main lemma enabling this comparison is Lemma 2.4.2, together with
statements in Section 2.5 relating the geometric localization of Bredon
homology to singular homology.
#### 1.3.4. Expressing fixed points in terms of nonequivariant configuration
spaces
We prove that the fixed points $C_{n}(M)^{G}$ satisfy homological stability by
reducing to 1.0.2. To do so, we express $C_{n}(M)^{G}$ in terms of
configuration spaces of other manifolds using the notion of _unordered
$S$-configurations_. We note that we are not the first to consider equivariant
configuration spaces (cf. [RS00]), but we use a new convention here.
Let $X$ be a $G$-space and let $S$ be a finite $G$-set. We define the _space
of unordered $S$-configurations in $X$_ (Definition 3.2.3) by
$C_{S}^{G}(X):=\operatorname{Emb}^{G}(S,X)/\operatorname{Aut}(S).$
When $M$ is $H$-stabilizable, the equivariant stabilization map (1.1.6)
restricts to a stabilization map
$\sigma_{G/H}:C_{S}^{G}(M)\to C_{S+G/H}^{G}(M),$
where $S+G/H$ denotes the disjoint union of $S$ and $G/H$ as $G$-sets.
###### Example 1.3.2.
If $S$ is a trivial $G$-set of cardinality $k$, then we have that
$C_{S}^{G}(X)=C_{k}(X^{G})$
is the space of unordered configurations of $k$ points in the $G$-fixed points
of $X$. If $X$ is a $G$-stabilizable $G$-manifold, the stabilization map
$\sigma_{G/G}:C_{S}^{G}(X)\to C_{S+[G/G]}^{G}(X)$
is the classical stabilization map
$\sigma:C_{k}(X^{G})\to C_{k+1}(X^{G}).$
Passing through the $S$-configuration spaces $C_{S}^{G}(X)$ allows us to
identify the fixed points $C_{n}(X)^{G}$ in terms of ordinary configuration
spaces:
###### Theorem C (Proposition 3.2.10 and Corollary 3.2.9).
1. (1)
For a $G$-space $X$ and $n\geq 0$, there is a disjoint union decomposition
$C_{n}(X)^{G}\cong\coprod_{|S|=n}C^{G}_{S}(X),$
where the disjoint union runs over isomorphism classes of finite $G$-sets $S$
with cardinality $n$.
2. (2)
Let $S$ be the finite $G$-set
$S=\bigsqcup_{(H)}k_{(H)}[G/H],$
where $k_{(H)}\geq 0$ for each conjugacy class of subgroups $(H)$. Then
$C_{S}^{G}(X)\cong\prod_{(H)}C_{k_{(H)}}(X_{(H)}/G),$
where $X_{(H)}$ is the subspace of $X$ consisting of points with stabilizer
conjugate to $(H)$ (Definition 3.2.4).
We apply part (1) of C to prove A; the essential point is that when $G$ is
nontrivial, the number of summands appearing in $C_{n}(X)^{G}$ increases as
$n$ increases, so the stabilization map cannot be surjective on homology. We
apply parts (1) and (2) to prove D.
#### 1.3.5. Stability for $S$-configurations and Bredon homology
Since they can be expressed in terms of ordinary configuration spaces, the
spaces of $S$-configurations often satisfy an analogue of the McDuff–Segal
theorem:
###### Theorem D (Theorem 4.3.1).
Let $M$ be a $G$-manifold and $S$ a finite $G$-set containing $k$ copies of
the orbit $[G/H]$ for a conjugacy class of subgroups $(H)$. If $M$ is
$H$-stabilizable and $M_{(H)}/G$ is connected, then the $(H)$-stabilization
map of (3.3.3) induces an isomorphism
$(\sigma_{G/H})_{*}:H_{d}(C^{G}_{S}(M))\to H_{d}(C^{G}_{S+[G/H]}(M))$
in integral homology in degrees $d\leq k/2$.
As explained above, B follows from stability for the fixed points
$C_{n}(M)^{G}$. C allows us to express these fixed points in terms of
$S$-configurations, and D states that the equivariant stabilization map
$\sigma_{G/H}$ from (1.1.6) induces an isomorphism in the homology of
$C_{S}^{G}(M)$ when $S$ contains sufficiently many copies of $[G/H]$. In
particular, once $S$ contains sufficiently many copies of $[G/K]$ for _any_
conjugacy class of subgroups $(K)$, there is some equivariant stabilization
map which induces an isomorphism on homology. Roughly speaking, B then follows
from the fact that there are finitely many finite $G$-sets $S$ for which none
of the equivariant stabilization maps induce isomorphisms in homology.
### 1.4. Outline
In Section 2, we recall the equivariant homology theory (Bredon homology) we
work with throughout the paper. We also recall the isotropy separation
sequence and prove some key results (Lemma 2.4.2, Proposition 2.5.1) needed to
reduce from the equivariant to nonequivariant setting.
In Section 3, we recall configuration spaces, stabilization maps, and their
equivariant analogues. We then introduce the more refined spaces of
$S$-configurations and use them to prove C which identifies the fixed points
of the configuration spaces of $G$-manifolds with (sums of products of)
configuration spaces in nonequivariant manifolds.
In Section 4, we prove A, B, and D, using the results of the previous section
to reduce to classical homological stability results.
In Section 5, we extend our results to the $RO(G)$-graded and Mackey functor-
valued settings.
### 1.5. Conventions
Throughout, $G$ denotes a finite group. If $M$ is an abelian group, then
$\underline{M}$ denotes the constant Mackey functor on $M$. In this work,
“manifold” means “smooth manifold.”
### 1.6. Acknowledgments
The authors thank Dan Dugger, Jeremy Hahn, Ben Knudsen, Irakli Patchkoria,
Oscar Randal-Williams, Bridget Schreiner, Inna Zakharevich, and Mingcong Zeng
for helpful discussions. The first author was partially supported by NSF grant
DMS-2204357. The second author was partially supported by NSF grants
DMS-2039316 amd DMS-2314082, as well as an AMS-Simons Travel Grant. The second
author also thanks the Max Planck Institute in Bonn for providing a wonderful
working environment and financial support during the beginning of this
project.
## 2\. Equivariant homotopy theory
In this section, we discuss the ideas from equivariant homotopy theory which
we will need in the sequel. Section 2.1 contains a rapid recollection of
Bredon homology with coefficients in a Mackey functor, which is the central
equivariant homology theory we use throughout this work. Section 2.2 and
Section 2.3 recall two standard constructions from equivariant stable homotopy
theory: geometric fixed points and proper homotopy orbits. These auxiliary
constructions appear in the isotropy separation sequence, which we recall in
Section 2.4. Our main technical result, Lemma 2.4.2, applies the isotropy
separation sequence to pass from Bredon homology to a simpler nonequivariant
homology theory. This simpler theory is identified as a sum of singular
homology theories in Section 2.5.
### 2.1. Bredon homology
Bredon homology associates to a $G$-space a sequence of Mackey functors (see
[May96, §IX.4]). As in nonequivariant homotopy theory, Bredon homology with
coefficients in a Mackey functor $\underline{M}$ is representable in the
equivariant stable homotopy category. By [May96, Theorem XIII.4.1], there is
an Eilenberg–MacLane $G$-spectrum $H\underline{M}$ with the property that its
homotopy Mackey functors satisfy
$\underline{\pi}_{0}(H\underline{M})\cong\underline{M}$ and
$\underline{\pi}_{i}(H\underline{M})=0$ for $i>0$. If $X$ is a $G$-space, then
we have
$\underline{H}_{*}^{(-)}(X;\underline{M})\cong\underline{\pi}_{*}(\Sigma^{\infty}_{G}X\wedge
H\underline{M}),$
where $\Sigma^{\infty}_{G}X$ is the _suspension $G$-spectrum of $X$_, and
$-\wedge-$ is the smash product of $G$-spectra.
The Bredon homology of a $G$-CW complex can equivalently be defined using an
equivariant analogue of the cellular chain complex [May96, §I.4]. A
_coefficient system_ is a contravariant functor
$\mathcal{O}_{G}\to\mathcal{A}b$, where $\mathcal{O}_{G}$ is the _orbit
category_ of $G$. If $X$ is a $G$-CW complex, we may define a coefficient
system by
$\underline{C}_{n}(X)(G/H):=\underline{H}_{n}(X^{n},X^{n-1};\mathbb{Z})(G/H):=H_{n}((X^{n})^{H},(X^{n-1})^{H};\mathbb{Z}).$
The connecting homomorphisms for the triples
$((X^{n})^{H},(X^{n-1})^{H},(X^{n-2})^{H})$ define a map of coefficient
systems
$d:\underline{C}_{n}(X)\to\underline{C}_{n-1}(X).$
If $\underline{M}$ is a _Mackey functor_ , we may define cellular chains with
coefficients in $\underline{M}$ by
$C_{n}^{G}(X;\underline{M}):=\underline{C}_{n}(X)\otimes\underline{M}:=\bigoplus_{G/H\in\mathcal{O}_{G}}\left(\underline{C}_{n}(X)(G/H)\otimes\underline{M}(G/H)\right)/\sim,$
where $f^{*}c\otimes m\sim c\otimes f_{*}m$ for a map $f:G/H\to
G/K\in\mathcal{O}_{G}$ and elements $c\in\underline{C}_{n}(X)(G/H)$ and
$m\in\underline{M}(G/H)$. This is a chain complex of abelian groups, with
differential given by $\partial=d\otimes 1$. By forgetting structure, we may
analogously define
$C_{n}^{H}(X;\underline{M}):=\underline{C}_{n}(i^{*}_{H}X)\otimes
i^{*}_{H}\underline{M}$
for each subgroup $H<G$; these assemble into a chain complex of coefficient
systems whose homology
$\underline{H}_{*}^{(-)}(X;\underline{M}):=H_{*}(C_{n}^{(-)}(X;\underline{M}))$
is the _Bredon homology of $X$ with coefficients in $\underline{M}$_.
For much of the paper, we will be concerned with $\mathbb{Z}$-graded Bredon
homology _groups_. By this, we mean the $G/G$-levels
$\underline{H}_{*}^{G}(X;\underline{M})$ of the Bredon homology Mackey
functors constructed here. We return to the level of generality of Bredon
homology Mackey functors in Section 5. It is also possible to extend Bredon
homology with coefficients in a Mackey functor from a $\mathbb{Z}$-graded
collection of groups to a $RO(G)$-graded collection of groups, where $RO(G)$
denotes the representation ring of $G$ [May96, §IX.5]. We review this
extension in Section 5 as well.
### 2.2. Geometric fixed points
As outlined in the previous section, the Bredon homology groups of a $G$-space
$X$ are the (categorical) fixed points of the $G$-spectrum
$\Sigma^{\infty}_{G}X\wedge H\underline{M}$. The _geometric fixed points_
construction is related to categorical fixed points (in a way that will be
made precise in Section 2.4), but has properties that make it more amenable to
computation. We refer the reader to [HHR16, Sec. 2.5] for proofs of the
results in this section.
###### Definition 2.2.1.
Let $G$ be a finite group and let $X$ be a $G$-spectrum. Let $\mathcal{P}$ be
the family of proper subgroups of $G$, and $E\mathcal{P}$ be the unique
$G$-homotopy type such that
$E\mathcal{P}^{H}\simeq\begin{cases}\emptyset&H=G,\\\ *&H<G.\end{cases}$
Consider the collapse map $E\mathcal{P}_{+}\to S^{0}$ and denote by
$\widetilde{E\mathcal{P}}$ its cofiber. The _geometric fixed points_ of $X$
are given by
$\Phi^{G}X=(\widetilde{E\mathcal{P}}\wedge X)^{G}.$
The key properties of the geometric fixed points functor are that it commutes
with taking suspension spectra and is strong symmetric monoidal. These
properties are the content of the following propositions.
###### Proposition 2.2.2.
For any $G$-space $X$, there is an equivalence of spectra
$\Phi^{G}\Sigma^{\infty}_{G}X\simeq\Sigma^{\infty}(X^{G}).$
###### Proposition 2.2.3.
For any two $G$-spectra $X$ and $Y$, there is an equivalence of spectra
$\Phi^{G}(X\wedge Y)\simeq\Phi^{G}X\wedge\Phi^{G}Y.$
### 2.3. Proper homotopy orbits
In order to relate geometric fixed points to categorical fixed points, we will
need the following auxiliary construction.
###### Definition 2.3.1.
The _proper homotopy orbits_ of a $G$-spectrum $X$ are given by
$X_{h\mathcal{P}}=(E\mathcal{P}_{+}\wedge X)^{G}.$
### 2.4. Isotropy separation sequence
In this section, we prove Lemma 2.4.2, a central tool in our analysis of
Bredon homology. It is based on the following cofiber sequence, which allows
us to understand categorical fixed points, provided we can understand
geometric fixed points and proper homotopy orbits.
###### Lemma 2.4.1 (Isotropy Separation Sequence [HHR16, Section 2.5.2]).
The collapse map of Definition 2.2.1 induces a cofiber sequence of spectra
$X_{h\mathcal{P}}\to X^{G}\to\Phi^{G}X.$
###### Lemma 2.4.2.
Let $G$ be a finite group, let $d\geq 0$, and let $X$ be a $G$-spectrum. Let
$\mathcal{C}$ be a Serre class. Assume
$\pi_{q}\Phi^{K}(H\underline{\mathbb{Z}}\wedge X)\in\mathcal{C}$ for all
$K\leq G$, $q\leq d+1$. Then
$H_{q}^{K}(X;\underline{\mathbb{Z}})\in\mathcal{C}$ for all $K\leq G$, $q\leq
d$.
The proof relies on the following homological fact, the proof of which we
learned from Piotr Pstrągowski.
###### Lemma 2.4.3 (Four Lemma mod Serre).
Suppose we have a commutative diagram
${A}$${B}$${C}$${D}$${A^{\prime}}$${B^{\prime}}$${C^{\prime}}$${D^{\prime}}$$\scriptstyle{f}$$\scriptstyle{\ell}$$\scriptstyle{g}$$\scriptstyle{m}$$\scriptstyle{h}$$\scriptstyle{n}$$\scriptstyle{p}$$\scriptstyle{r}$$\scriptstyle{s}$$\scriptstyle{t}$
in an abelian category $\mathcal{A}$. Let $\mathcal{C}\subseteq\mathcal{A}$ be
a Serre subcategory. Suppose the rows are exact, $\ell$ is an epimorphism mod
$\mathcal{C}$ and $p$ is a monomorphism mod $\mathcal{C}$. If $n$ is an
epimorphism mod $\mathcal{C}$ then so is $m$. If $m$ is a monomorphism mod
$\mathcal{C}$ then so is $n$.
###### Proof.
We just prove the first statement, as the second is analogous. As
$\mathcal{C}$ is a Serre subcategory, there exists an abelian category
$\mathcal{A}/\mathcal{C}$ and an exact functor
$F:\mathcal{A}\to\mathcal{A}/\mathcal{C}$ which is essentially surjective with
kernel $\mathcal{C}$ [Sta18, Lemma 02MS]. Applying $F$ yields a commutative
diagram
${FA}$${FB}$${FC}$${FD}$${FA^{\prime}}$${FB^{\prime}}$${FC^{\prime}}$${FD^{\prime}}$$\scriptstyle{Ff}$$\scriptstyle{F\ell}$$\scriptstyle{Fg}$$\scriptstyle{Fm}$$\scriptstyle{Fh}$$\scriptstyle{Fn}$$\scriptstyle{Fp}$$\scriptstyle{Fr}$$\scriptstyle{Fs}$$\scriptstyle{Ft}$
in the abelian category $\mathcal{C}/\mathcal{A}$ in which the rows are exact,
$F\ell$ and $Fn$ are epimorphisms, and $Fp$ is a monomorphism. The classical
Four Lemma implies that $Fm$ is an epimorphism, so
$F\operatorname{coker}(m)=\operatorname{coker}(Fm)=0$, i.e.,
$\operatorname{coker}(m)$ lies in $\mathcal{C}$, i.e., $m$ is an epimorphism
mod $\mathcal{C}$. ∎
###### Lemma 2.4.4.
Let $\mathcal{C}\subseteq\mathcal{A}$ be a Serre category. Let
$\sigma:E^{\prime}_{r}\to E^{\prime\prime}_{r}$ be a map of spectral sequences
that is an isomorphism modulo $\mathcal{C}$ on the $E_{r}$ page for some $r$.
Then $E^{\prime}_{r+1}\overset{\sigma}{\to}E^{\prime\prime}_{r+1}$ is an
isomorphism mod $\mathcal{C}$.
###### Proof.
This is a straightforward consequence of Lemma 2.4.3. First apply Lemma 2.4.3
to the diagram
$\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{Z^{\prime}_{r}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{E^{\prime}_{r}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{d_{r}}$$\scriptstyle{\sigma}$$\textstyle{E^{\prime}_{r}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\sigma}$$\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{Z^{\prime\prime}_{r}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{E^{\prime\prime}_{r}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{d_{r}}$$\textstyle{E^{\prime\prime}_{r}}$
(where $Z^{\prime}_{r}$ denotes the submodule of cycles) to show that
$Z^{\prime}_{r}\to Z^{\prime\prime}_{r}$ is an isomorphism mod $\mathcal{C}$.
Similar arguments involving the exact sequences $0\to Z^{\prime}_{r}\to
E^{\prime}_{r}\to B^{\prime}_{r}\to 0$ and $0\to B^{\prime}_{r}\to
Z^{\prime}_{r}\to Z^{\prime}_{r}/B^{\prime}_{r}=E^{\prime}_{r+1}\to 0$ show
that the induced maps $Z^{\prime}_{r}\to Z^{\prime\prime}_{r}$ and
$E^{\prime}_{r+1}\to E^{\prime\prime}_{r+1}$, respectively, are isomorphisms
mod $\mathcal{C}$. ∎
###### Proof of Lemma 2.4.2.
By induction up the subgroup lattice. For the base case,
$H^{e}_{q}(X;\underline{\mathbb{Z}})=\pi_{q}((H\underline{\mathbb{Z}}\wedge
X)^{e})=\pi_{q}(\Phi^{e}(H\underline{\mathbb{Z}}\wedge X))\in\mathcal{C}.$
Now let $\\{e\\}<K\leq G$ and let $\mathcal{P}$ denote the family of proper
subgroups of $K$. The isotropy separation sequence
$(H\underline{\mathbb{Z}}\wedge
X)_{h\mathcal{P}}\to(H\underline{\mathbb{Z}}\wedge
X)^{K}\to\Phi^{K}(H\underline{\mathbb{Z}}\wedge X)$
gives rise to a long exact sequence
$\displaystyle\cdots\to\pi_{q+1}(\Phi^{K}(H\underline{\mathbb{Z}}\wedge
X))\to\pi_{q}((H\underline{\mathbb{Z}}\wedge X)_{h\mathcal{P}})\to
H^{K}_{q}(X;\underline{\mathbb{Z}})$
$\displaystyle\to\pi_{q}(\Phi^{K}(H\underline{\mathbb{Z}}\wedge
X))\to\pi_{q-1}((H\underline{\mathbb{Z}}\wedge X)_{h\mathcal{P}})\to\cdots.$
To show that the middle term is in $\mathcal{C}$, we claim it suffices to show
that all the other terms are in $\mathcal{C}$: apply Lemma 2.4.3 to the map
from the above exact sequence to the zero exact sequence.
The assumption says that the first and fourth terms are in $\mathcal{C}$, so
we focus on showing the second term is in $\mathcal{C}$, which would also show
the fifth term is in $\mathcal{C}$ because $q\leq d$ is arbitrary. There is a
cell decomposition of $E\mathcal{P}_{+}$ by a finite collection of cells
$\\{{K/H_{i}}_{+}\wedge S^{n_{i}}\\}_{i\in\mathcal{I}}$
with every $H_{i}<K$ (cf. [HHR16, Section 2.5.2]). Filtering
$E\mathcal{P}_{+}\wedge X$ by the cells of $E\mathcal{P}_{+}$ gives rise to a
spectral sequence
$E_{1}^{i,p-n_{i}}\cong H_{p}^{K}({K/H_{i}}_{+}\wedge S^{n_{i}}\wedge
X)\Rightarrow H^{K}_{p-n_{i}}(E\mathcal{P}_{+}\wedge X).$
We have
$\displaystyle H^{K}_{p}({K/H}_{+}\wedge S^{n}\wedge X)$
$\displaystyle\cong[S,H\underline{\mathbb{Z}}\wedge K/H\wedge
X]^{K}_{p-n}\cong[{K/H}_{+},H\underline{\mathbb{Z}}\wedge X]^{K}_{p-n}$
$\displaystyle\cong[S,H\underline{\mathbb{Z}}\wedge X]^{H}_{p-n}\cong
H^{H}_{p-n}(X;\underline{\mathbb{Z}})$
using self-duality of $K/H_{+}$. In particular, since $H_{i}<K$, we have
$E_{1}^{i,q}\in\mathcal{C}$ by the induction hypothesis. By Lemma 2.4.4,
$H^{K}_{q}(E\mathcal{P}_{+}\wedge X)$ is also in $\mathcal{C}$. ∎
### 2.5. Geometric fixed points of Eilenberg–MacLane $G$-spectra
In light of the previous section, one piece of understanding the Bredon
homology $(H\underline{M}\wedge\Sigma^{\infty}_{G}X)^{G}$ of a $G$-space $X$
is understanding a geometric fixed points term of the form
$\Phi^{G}(H\underline{M}\wedge\Sigma^{\infty}_{G}X)\simeq\Phi^{G}H\underline{M}\wedge\Phi^{G}\Sigma^{\infty}_{G}X\simeq\Phi^{G}H\underline{M}\wedge\Sigma^{\infty}X^{G},$
which is the $\Phi^{G}H\underline{M}$-homology of the space $X^{G}$. In many
cases of interest (see Proposition 2.5.1), this splits into (nonequivariant)
Eilenberg-Maclane spaces, allowing us to reduce to studying ordinary homology
of fixed point spaces.
Recall that the category of Mackey functors is symmetric monoidal with unit
the Burnside Mackey functor $\underline{A}$ and monoidal product the _box
product_. A Green functor is a monoid in the category of Mackey functors
[Lew].
###### Proposition 2.5.1.
Let $\underline{M}$ be a Green functor. For all $H\leq G$, the spectrum
$\Phi^{H}H\underline{M}$ is a connective generalized Eilenberg–MacLane
spectrum, i.e., it splits as a wedge of nonnegative suspensions of
Eilenberg–MacLane spectra. Moreover, if $\underline{M}(G/K)$ is finitely
generated as an abelian group for all $K\leq H$, then
$\pi_{k}\Phi^{H}H\underline{M}$ is a finitely generated abelian group for each
$k\in\mathbb{Z}$.
###### Proof.
The lax monoidal transformation $(-)^{H}\to\Phi^{H}(-)$ induces a map of
commutative ring spectra (cf. [HHR16, §B.10.5])
$(H\underline{M})^{H}\to\Phi^{H}H\underline{M}.$
Since $H\underline{M}$ is Eilenberg–MacLane, $(H\underline{M})^{H}\simeq
H\underline{M}(G/H)$, and since $\underline{M}$ is a Green functor,
$\underline{M}(G/H)$ is a commutative ring (see, e.g., [Shu10, §2.2.5]).
Therefore we have a map of commutative ring spectra
$H\mathbb{Z}\to
H\underline{M}(G/H)\simeq(H\underline{M})^{H}\to\Phi^{H}H\underline{M}$
which determines an $H\mathbb{Z}$-module structure on
$\Phi^{H}H\underline{M}$. The claim that $\Phi^{H}H\underline{M}$ is
generalized Eilenberg–MacLane then follows from the classical fact (usually
attributed to Adams) that any $H\mathbb{Z}$-module is generalized
Eilenberg–MacLane.
That $\Phi^{H}H\underline{M}$ is connective follows from the isotropy
separation sequence: we have
$\Phi^{H}H\underline{M}\simeq\operatorname{cofib}((H\underline{M})_{h\mathcal{P}}\to(H\underline{M})^{H})$
and both $(H\underline{M})_{h\mathcal{P}}$ and $(H\underline{M})^{H}$ are
connective, so $\Phi^{H}H\underline{M}$ is connective.
Finite generation also follows from the isotropy separation sequence: the
finite generation hypothesis on $\underline{M}(G/K)$, along with the fact that
$E\mathcal{P}$ admits an $H$-CW structure with finitely cells in each
dimension, implies that each homotopy group of
$(H\underline{M})_{h\mathcal{P}}$ and $(H\underline{M})^{H}\simeq
H(\underline{M}(G/H))$ is finitely generated, so the same is true for
$\Phi^{H}H\underline{M}$. ∎
###### Example 2.5.2.
The Burnside Mackey functor $\underline{A}$ is the Mackey functor given by
$\underline{A}(G/K)=A(K),$
where $A(K)$ denotes the Grothendieck ring of finite $K$-sets. The
restrictions and transfers in $\underline{A}$ are induced by restriction and
induction of $K$-sets, respectively.
The constant Mackey functor $\underline{\mathbb{Z}}$ has
$\underline{\mathbb{Z}}(G/K)=\mathbb{Z}$ in every level. The restrictions in
$\underline{\mathbb{Z}}$ are the identity map $\mathbb{Z}\to\mathbb{Z}$ and
the transfer $\underline{\mathbb{Z}}(G/K)\to\underline{\mathbb{Z}}(G/L)$ is
given by multiplication by the index $[L:K]$.
Both $\underline{A}$ and $\underline{\mathbb{Z}}$ are Green functors, so
$\Phi^{H}H\underline{A}$ and $\Phi^{H}H\underline{\mathbb{Z}}$ are both
connective generalized Eilenberg–MacLane spectra. Moreover,
$\underline{A}(G/K)=A(K)$ and $\underline{\mathbb{Z}}(G/K)=\mathbb{Z}$ are
finitely generated abelian groups for all $K\leq H$, so
$\pi_{k}\Phi^{H}H\underline{A}$ and $\pi_{k}\Phi^{H}H\underline{\mathbb{Z}}$
are both finitely generated abelian groups for all $k\geq 0$.
###### Remark 2.5.3.
It is possible to describe $\Phi^{G}H\underline{M}$ explicitly in many cases
of interest using previous computations of $\pi_{*}\Phi^{G}H\underline{M}$
(e.g., [HHR16, Kri20]), but this will not be necessary for our applications.
We conclude by recording the following fact which can be used to reduce to
$p$-Sylow subgroups for coefficients in a constant Mackey functor. We will not
need it later, but it can be used to simplify some arguments in the sequel if
one works with constant Mackey functor coefficients.
###### Proposition 2.5.4.
Let $\underline{R}$ be a constant Green functor and let $\underline{M}$ be an
$\underline{R}$-algebra. If $G$ is not a $p$-group, then
$\Phi^{G}H\underline{M}\simeq*.$
###### Proof.
The $G$-spectrum $H\underline{M}$ is an $H\underline{R}$-algebra, and since
geometric fixed points are strong symmetric monoidal, the spectrum
$\Phi^{G}H\underline{M}$ is a $\Phi^{G}H\underline{R}$-algebra. But [Kri20,
Prop. 11] implies that when $G$ is not a $p$-group,
$\Phi^{G}H\underline{R}\simeq*$, from which we conclude
$\Phi^{G}H\underline{M}\simeq*$. ∎
###### Example 2.5.5.
The constant Green functor $\underline{\mathbb{Z}}$ is an algebra over itself,
so $\Phi^{G}H\underline{\mathbb{Z}}\simeq*$ if $G$ is not a $p$-group. On the
other hand, one can show that $\pi_{0}\Phi^{G}H\underline{A}\cong\mathbb{Z}$
for any finite group $G$, so $\Phi^{G}H\underline{A}\not\simeq*$ for any $G$.
###### Remark 2.5.6.
For simplicity, we will work with $\underline{\mathbb{Z}}$ coefficients in the
sequel. If $\underline{M}$ is a constant Mackey functor which is levelwise
finitely generated, then it is straightforward to deduce $\underline{M}$
coefficient analogues of all of the $\underline{\mathbb{Z}}$ coefficient
statements below using the presentation of $\underline{M}$ as a
$\underline{\mathbb{Z}}$-module (cf. deducing homology isomorphisms with $M$
coefficients from homology isomorphisms with $\mathbb{Z}$ coefficients).
For certain groups, e.g., any cyclic $p$-group $C_{p^{n}}$, it is also
straightforward to pass from $\underline{\mathbb{Z}}$ coefficients to
$\underline{A}$ coefficients using induction up the subgroup lattice, the
short exact sequence of Mackey functors
$0\to\underline{I}\to\underline{A}\to\underline{\mathbb{Z}}\to 0,$
and the fact that $H_{*}^{C_{p^{n}}}(X;\underline{I})\cong
H_{*}^{C_{p^{n-1}}}(X^{C_{p}};\underline{I}^{\prime})$, where
$\underline{I}^{\prime}$ is the $C_{p^{n-1}}$ Mackey functor with
$\underline{I}^{\prime}(C_{p^{n-1}}/C_{p^{m}})=\underline{I}(C_{p^{n}}/C_{p^{m+1}})$.
Since every Mackey functor is an $\underline{A}$-module, one can deduce
analogues of our results with coefficients in any Mackey functor which is
finitely generated for cyclic $p$-groups.
We leave the extension to other coefficients and more complicated groups to
the interested reader.
## 3\. Configuration spaces and stabilization maps
As in the previous section, $G$ is any finite group. In this section, we
discuss configuration spaces and stabilization maps. We recall the
nonequivariant theory in Section 3.1 and discuss the equivariant analogues in
Section 3.2 and Section 3.3. In Section 3.2, we define the more nuanced notion
of $S$-configuration spaces (Definition 3.2.2 and Definition 3.2.3) and their
stabilization maps. Our main results are Proposition 3.2.10, which expresses
the $G$-fixed points of configuration spaces in terms of $S$-configurations,
and Corollary 3.2.9 which expresses each $S$-configuration space in terms of
products of ordinary configuration spaces. Combining these results yields C.
### 3.1. Nonequivariant configuration spaces and stabilization maps
###### Definition 3.1.1.
Let $M$ be a space. The _configuration space of $n$ ordered points in $M$_ is
$\operatorname{Conf}_{n}(M):=\operatorname{Emb}(\\{1,\ldots,n\\},M)=\\{(x_{1},\dots,x_{n})\in
M^{\times n}\hskip 2.0pt:\hskip 2.0ptx_{i}\neq x_{j}\text{ for all }i\neq
j\\}$
topologized as a subspace of $M^{\times n}$.
The _configuration space of $n$ unordered points in $M$_ is the quotient space
$C_{n}(M):=\operatorname{Conf}_{n}(M)/\Sigma_{n},$
where the symmetric group $\Sigma_{n}$ acts by permuting points.
When $M$ is an open, connected manifold with $\dim M\geq 2$, then $M$ is the
interior of a manifold with boundary, and there there are well-defined
_stabilization maps_ $C_{n}(M)\to C_{n+1}(M)$ (cf. [Pal18, Secs. 2.1-2.2]).
Since we will describe the equivariant analogue of these maps in some detail
in Section 3.3, we review the non-equivariant construction.
###### Definition 3.1.2.
We say that an $n$-dimensional manifold $M$ is _stabilizable_ if $M$ is
homeomorphic to the interior of a $n$-dimensional manifold with nonempty
boundary.
###### Construction 3.1.3.
When $M$ is stabilizable, we can define a stabilization map on unordered
configurations as follows. Let $M$ be stabilizable with $M\cong W^{\circ}$,
where $W$ is an $n$-dimensional manifold with nonempty boundary. Let
$p\in\partial W$. By the definition of manifold with boundary, there exists an
open neighborhood $U\subset W$ containing $p$ together with a diffeomorphism
$\phi:U\xrightarrow{\cong}\mathbb{R}^{n}_{+}=\\{x=(x_{1},\ldots,x_{n})\in\mathbb{R}^{n}:x_{1}\geq
0\\}$
sending $p$ to $0$. Let $b:\mathbb{R}^{n}_{+}\to\mathbb{R}^{n}_{+}$ be a
smooth bump function which sends $0$ to $(1,0,\ldots,0)$. Let $e:W\to W$ be
the self-embedding defined by
$e(x)=\begin{cases}x\quad&\text{ if }x\notin U,\\\ (\phi^{-1}\circ
b\circ\phi)(x)\quad&\text{ if }x\in U.\end{cases}$
Then by construction, $e(M)\subseteq M$ and $e(p)\notin e(M)$, so we may
define
(3.1.4) $\sigma:C_{n}(M)\to C_{n+1}(M)$
by
$\sigma(m_{1},\ldots,m_{n})=(e(m_{1}),\ldots,e(m_{n}),e(p)).$
### 3.2. Equivariant configuration spaces
In this section, we study equivariant configuration spaces. Our main result
(Corollary 3.2.9) will allow us to relate these spaces back to nonequivariant
configuration spaces.
###### Definition 3.2.1.
Let $X$ be a $G$-space and let $n\geq 1$. The configuration space of $n$
ordered points in $X$ is the $G$-space of nonequivariant embeddings
$\operatorname{Conf}_{n}(X)=\operatorname{Emb}(\\{1,\dots,n\\},X).$
The configuration space of $n$ unordered points in $X$ is the quotient
$G$-space
$C_{n}(X)=\operatorname{Conf}_{n}(X)/\Sigma_{n}.$
Let $C(X)=\bigsqcup_{n\geq 1}C_{n}(X)$ equipped with the disjoint union
topology. In $\operatorname{Conf}_{n}(X)$ and $C_{n}(X)$, the $G$-action is
induced by the action on $X$.
These are the same definitions as in the nonequivariant case, but now they are
$G$-spaces. In Proposition 3.2.10, we will show that their fixed points
$C_{n}(X)^{G}$ split into pieces $C_{S}^{G}(X)$ which are described below.
###### Definition 3.2.2.
Let $X$ be a $G$-space and let $S$ be a finite $G$-set of cardinality $n$. The
_space of ordered $S$-configurations in $X$_, denoted
$\operatorname{Conf}^{G}_{S}(X)$, is the space of $G$-equivariant embeddings
$S\to X$,
$\operatorname{Conf}_{S}^{G}(X):=\operatorname{Emb}^{G}(S,X).$
This is a $G$-equivariant analogue of the ordinary notion of an ordered
configuration space. Indeed, if $G=e$ is the trivial group, we have
$\operatorname{Conf}^{e}_{S}(X)\cong\operatorname{Conf}_{n}(X^{e}).$
Precomposing a $G$-equivariant embedding $S\to X$ by an automorphism of $S$
yields another such embedding. In this way, $\operatorname{Conf}^{G}_{S}(X)$
admits an action by the automorphism group $\operatorname{Aut}(S)$.
###### Definition 3.2.3.
The _space of unordered $S$-configurations in $X$_ is the quotient
$C^{G}_{S}(X):=\operatorname{Conf}^{G}_{S}(X)/\operatorname{Aut}(S)$
of $\operatorname{Conf}^{G}_{S}(X)$ by the $\operatorname{Aut}(S)$-action.
Similarly, this is a $G$-equivariant analogue of the ordinary notion of
unordered configuration space. If $G=e$ is the trivial group, then
$\operatorname{Aut}(S)\cong\Sigma_{n}$, and we have
$C^{e}_{S}(X)=\operatorname{Conf}^{e}_{S}(X)/\operatorname{Aut}(S)\cong\operatorname{Conf}_{n}(X^{e})/\Sigma_{n}=C_{n}(X^{e}).$
We can express $C^{G}_{S}(X)$ in terms of nonequivariant configuration spaces
of a related space.
###### Definition 3.2.4 ([tD87, Section I.5]).
For a $G$-space $X$ and subgroup $H\leq G$, we denote by $X_{(H)}$ the
subspace
$X_{(H)}=\\{x\in X:(G_{x})=(H)\\}\subseteq X$
consisting of points with stabilizer conjugate to $H$.
Since the stabilizers of points in a fixed orbit are conjugate, the $G$-action
on $X$ restricts to a $G$-action on $X_{(H)}$. The preimage of any point under
the projection $X_{(H)}\to X_{(H)}/G$ is a full orbit isomorphic to $G/H$.
###### Proposition 3.2.5.
Given a subgroup $H\subseteq G$ and $k\geq 0$, the space of unordered
$k[G/H]$-configurations in $X$ satisfies
$C^{G}_{k[G/H]}(X)\cong C_{k}\left(X_{(H)}/G\right).$
###### Proof.
A point in $C^{G}_{k[G/H]}(X)$ is a subset $S\subseteq X$ which is isomorphic
as a $G$-set to $k[G/H]$. Each point in $S$ has stabilizer conjugate to $H$,
so $S$ is furthermore a subset of $X_{(H)}$. The set $S$ consists of $k$
orbits of type $G/H$, so its image under the projection $X_{(H)}\to X_{(H)}/G$
is a size $k$ subset of $X_{(H)}/G$. This defines a map
$C^{G}_{k[G/H]}(X)\to C_{k}\left(X_{(H)}/G\right),$
which is a homeomorphism. Indeed, the preimage of any $k$ element subset of
$X_{(H)}/G$ is a subset $S\subseteq X_{(H)}\subseteq X$ of the above form,
which defines an inverse map. ∎
The space $X_{(H)}/G$ can also be expressed in terms of $H$-fixed points of
$X$ and their $W_{G}H$-action, as we show in the following lemma.
###### Lemma 3.2.6.
For any $G$-space $X$,
$X_{(H)}/G\cong\Big{(}X^{H}\setminus\bigcup_{K>H}X^{K}\Big{)}/W_{G}H.$
###### Proof.
There is a $G$-equivariant homeomorphism
(3.2.7) $\varphi:X_{(H)}\to
G\times_{N_{G}H}\Big{(}X^{H}\setminus\bigcup_{K>H}X^{K}\Big{)}$
and hence on orbits we have
$\displaystyle X_{(H)}/G$
$\displaystyle\cong\Big{(}X^{H}\setminus\bigcup_{K>H}X^{K}\big{)}/N_{G}H\cong\Big{(}X^{H}\setminus\bigcup_{K>H}X^{K}\big{)}/W_{G}H$
since $H\leq N_{G}H$ acts trivially on $X^{H}$. ∎
In Corollary 3.2.9 we generalize Proposition 3.2.5 to a formula for
$C^{G}_{S}(X)$ for an arbitrary $G$-set $S$. It follows from the next lemma,
whose proof is omitted.
###### Lemma 3.2.8.
Let $S$ and $T$ be finite $G$-sets such that no orbit of $S$ is isomorphic to
an orbit of $T$. Then,
$C^{G}_{S+T}(X)\cong C^{G}_{S}(X)\times C^{G}_{T}(X).$
Combining Proposition 3.2.5 and Lemma 3.2.8, we obtain our desired
decomposition formula.
###### Corollary 3.2.9.
Let $S$ be the finite $G$-set
$S=\coprod_{(H)}k_{(H)}[G/H],$
where $k_{(H)}\geq 0$ for each conjugacy class of subgroups $(H)$. Then,
$C^{G}_{S}(X)\cong\prod_{(H)}C_{k_{(H)}}(X_{(H)}/G).$
Recall that $C_{n}(X)$ denotes the $G$-space of nonequivariant embeddings of
$n$ points. The next proposition is the main result of this section, which
describes its fixed points in terms of the spaces $C_{S}^{G}(X)$ we have been
studying above, in the case that $X$ is a manifold.
###### Proposition 3.2.10.
For a $G$-manifold $M$ and $n\geq 0$, there is a disjoint union decomposition
$C_{n}(M)^{G}\cong\coprod_{|S|=n}C^{G}_{S}(M),$
where the disjoint union runs over isomorphism classes of finite $G$-sets $S$
with cardinality $n$.
###### Proof.
Let $S=\coprod_{i\in I}G/H_{i}$ be a $G$-set of cardinality $n$. Since
$C_{S}^{G}(M)$ and $C_{n}(M)^{G}$ are given compatible quotient topologies, it
suffices to show that $\operatorname{Conf}_{S}^{G}(M)$ is open in
$\operatorname{Conf}_{n}(M)^{G}$. Let $x\in\operatorname{Conf}_{S}^{G}(M)$ be
an ordered configuration consisting of orbits $\mathcal{O}_{i}=\\{g\cdot
x_{i}\\}_{g\in G/H_{i}}$ for $i\in I$. By the equivariant slice theorem (see
e.g. [Aud91, Theorem I.2.1.1]), there are open neighborhoods $U_{i,g}\ni
g\cdot x_{i}$ and diffeomorphisms $\phi_{i}:\bigsqcup_{g}U_{i,g}\cong
G\times_{H_{i}}T_{x_{i}}M$ with $\phi_{i}(g\cdot x_{i})=(g,0)$. Then
$U=\prod_{i,g}U_{i,g}\subseteq M^{\times n}$ is an open neighborhood of $x$.
An element of $\operatorname{Conf}_{n}^{G}(M)\cap U$ is an $n$-tuple
$(y_{i,g})$ where $y_{i,g}\in U_{i,g}$, and we may assume that the $U_{i,g}$’s
are disjoint in $M$. It suffices to show that $y_{i,g}$ has isotropy $H_{i}$.
Since each $U_{i,g}$ contains a single $y_{i,g}$ and $G$ acts on
$\\{y_{i,g}\\}$, an element $g^{\prime}\in G$ fixes $y_{i,g}$ if and only if
it preserves $U_{i,g}\cong\\{g\\}\times T_{x_{i}}M\subseteq
G\times_{H_{i}}T_{x_{i}}M$, and this occurs if and only if $g^{\prime}\in
H_{i}$. ∎
###### Remark 3.2.11.
Though Proposition 3.2.10 uses the theory of (finite-dimensional) manifolds,
we will investigate an analogous statement for $\rho^{\infty}$ in future work
and identify the pieces:
(3.2.12) $C^{G}_{G/H}(\rho^{\infty})\simeq B(W_{G}H).$
These spaces are interesting because Guillou and May [GM17, Lem. 1.2] identify
$C_{n}(\rho^{\infty})$ with the $n$-th space in the $G$-equivariant little
$\rho^{\infty}$-disks operad, which is well-known to be $G$-equivalent to
$B_{G}\Sigma_{n}$ (cf. [CW91, Pg. 489]). In particular, the splitting here
recovers the Lashof–May splitting [May96, Theorem VII.2.7] of
$(B_{G}\Sigma_{n})^{G}$.
### 3.3. Equivariant stabilization maps
Our goal in this subsection is to define an equivariant version of
Construction 3.1.3.
###### Definition 3.3.1.
Let $H\leq G$ be a subgroup. A $G$-manifold $M$ is _$H$ -stabilizable_ if it
is equivariantly homeomorphic to the interior of a $G$-manifold $W$ with
boundary containing a point of isotropy group $H$.
###### Construction 3.3.2.
Suppose $M$ is an $H$-stabilizable $G$-manifold of dimension $n$, and write
$M\cong W^{\circ}$ where $W$ is a $G$-manifold with boundary containing a
point $p\in\partial W$ with isotropy group $H$. By the equivariant slice
theorem for manifolds with boundary [Kan07, Theorem 3.6], there exists an open
subset $U\subseteq W$ containing the orbit $\\{g\cdot p\hskip 2.0pt:\hskip
2.0ptg\in G/H\\}$ together with a $G$-equivariant diffeomorphism
$\phi:U\to\mathbb{R}_{\geq 0}\times(G\times_{H}V)$
sending $g\cdot p$ to $(0,g,0)$, where $\mathbb{R}_{\geq 0}$ has the trivial
action and $V\cong T_{p}(\partial M)$ is the $G$-representation determined by
the action of $G$ on the tangent space of $\partial M$ at $p$. Let
$b:\mathbb{R}_{\geq 0}\times(G\times_{H}V)\to\mathbb{R}_{\geq
0}\times(G\times_{H}V)$ be a smooth, $G$-equivariant bump function which is
the identity in the $G\times_{H}V$ component and sends $(0,g,0)$ to $(1,g,0)$.
Let $e:W\to W$ be the $G$-equivariant self-embedding defined by
$e(x)=\begin{cases}x\quad&\text{ if }x\notin U,\\\ (\phi^{-1}\circ
b\circ\phi)(x)\quad&\text{ if }x\in U.\end{cases}$
Let $g_{1},\ldots,g_{r}$ range over the elements of $G/H$. We define
(3.3.3) $\sigma_{G/H}:C_{n}(M)\to C_{n+|G/H|}(M)$
by
$\sigma(m_{1},\ldots,m_{n})=(e(m_{1}),\ldots,e(m_{n}),e(p),g_{1}e(p),\ldots,g_{r}e(p)).$
###### Remark 3.3.4.
Observe that in 3.3.2, the bump function $\phi$, and hence $\sigma_{G/H}$,
preserves isotropy type. Moreover, since $\sigma_{G/H}$ adds a single $G/H$
orbit, it descends to maps
$\displaystyle C_{n}(M/G)$ $\displaystyle\to C_{n+1}(M/G)$ $\displaystyle
C_{n}(M)^{G}$ $\displaystyle\to C_{n+|G/H|}(M)^{G}$ $\displaystyle
C_{S}^{G}(M)$ $\displaystyle\to C_{S+[G/H]}^{G}(M).$
Putting together Corollary 3.2.9 and Proposition 3.2.10, we expect equivariant
homological stability of $M$ should be related to nonequivariant homological
stability of $M_{(H)}/G$. First (Lemma 3.3.5) we check that $M_{(H)}/G$
satisfies the hypotheses of nonequivariant homological stability except for
connectedness, which is an additional assumption and not needed for the
definition of stabilization maps. Then (Lemma 3.3.6) we check that the
stabilization maps agree. In the next section (Theorem 4.3.1) we will put
these together to prove a form of equivariant homological stability.
###### Lemma 3.3.5.
If $M$ is an $H$-stabilizable smooth $G$-manifold for $H\leq G$, then
$M_{(H)}/G$ is an open smooth manifold.
###### Proof.
Let $W=W_{H}G$. To see that $M_{(H)}/G$ is a manifold, note that
$M_{(H)}/G=M_{(H)}/W$ since $M_{(H)}$ consists entirely of $G/H$-orbit types.
Thus $W=\operatorname{Aut}(G/H)$ acts freely on the manifold $M_{(H)}$, and
its acts properly because $W$ is finite. Thus $M_{(H)}/W$ is a (smooth)
manifold.
For noncompactness, first we show noncompactness of $M_{(H)}$. The
stabilizability assumption says that there is a neighborhood of the manifold
with boundary $W$ that is diffeomorphic to $\mathbb{R}_{\geq
0}\times(G\times_{H}V)$, where $V$ is the tangent space in $\partial W$ of a
point $p\in\partial W$ of isotropy $H$ (see Construction 3.3.2). The subspace
of isotropy $H$ in this neighborhood is $r_{\geq 0}\times(G\times_{H}V^{H})$,
and taking away the boundary, we see that $M_{(H)}$ is missing a limit point,
namely $\\{0\\}\times(G\times_{H}\\{0\\})=p$. Now passing to orbits
$M_{(H)}/G$, we have a neighborhood $\mathbb{R}_{>0}\times V^{H}$ with limit
point $(0,0)$ that is also not in $M_{(H)}/G$. ∎
###### Lemma 3.3.6.
Suppose $M$ is $H$-stabilizable. Then the equivariant stabilization map
(3.3.3) induces a map
$\bar{\sigma}_{G/H}:C_{n}(M_{(H)}/G)\to C_{n+1}(M_{(H)}/G)$
that agrees with the nonequivariant stabilization map (3.1.4).
Note that Lemma 3.3.5 is used to guarantee the map (3.1.4) exists.
###### Proof.
Since $\sigma_{G/H}:C_{n}(M)\to C_{n+|G/H|}(M)$ preserves isotropy type and
adds a $G/H$ orbit, it induces a map
$\bar{\sigma}_{G/H}^{e}:C_{n}(M_{(H)}/G)\to C_{n+1}(M_{(H)}/G)$, and the
equivariant bump function $b$ induces a function $\bar{b}$ on $M_{(H)}/G$.
First note from Lemma 3.3.5 that $M_{(H)}/G$ is a manifold. To check
$\bar{\sigma}_{G/H}$ agrees with the classical stabilization map, we must
check that $\bar{b}$ is a bump function. It is clear that it has compact
support. To check it is smooth, first note that the restriction of the smooth
function $b$ to $M_{(H)}$ is smooth. To check it is smooth after taking
orbits, recall that every orbit in $M_{(H)}$ has a neighborhood diffeomorphic
to $G\times_{H}\mathbb{R}^{n}$, and taking orbits simply collapses the $|G/H|$
Euclidean components to one Euclidean component. ∎
## 4\. Equivariant homological stability
In this section, we prove our main theorems, A and B. Our proof is by
reduction to the classical stability result of McDuff–Segal (Theorem 4.1.1),
which we recall in Section 4.1. We also discuss a weaker notion of homological
stability (Corollary 4.1.3) in terms of finite generation; the relationship
between finite generation stability and sequential stability is made precise
in Proposition 4.1.2. Using nonequivariant homological stability, we prove D
(Theorem 4.3.1) that the equivariant stabilization maps eventually induce
isomorphisms in the homology of $S$-configuration spaces in Section 4.3. We
then use our decomposition of the $G$-fixed points of configuration spaces
into $S$-configuration spaces, along with our Bredon-to-geometric-fixed-point
reduction result (Lemma 2.4.2), to prove Bredon homological stability in
Section 4.4.
### 4.1. Nonequivariant homological stability
###### Theorem 4.1.1 (McDuff–Segal, as stated in [Pal18, Thm. 1.2] with
$X=*$).
Assume $M$ is an open, connected manifold with $\dim(M)\geq 2$. The
stabilization map $\sigma:C_{n}(M)\to C_{n+1}(M)$ induces an isomorphism on
integral homology in degrees $*\leq\frac{n}{2}.$
Our goal for the rest of this subsection is to relate homological stability
with respect to iterated applications of $\sigma$ with finite generation over
$\mathbb{Z}[\sigma_{*}]$. Since this does not depend on the fact that the
groups that stabilize are homology groups, we take an abstract approach,
presenting the translation for arbitrary sequences of modules that stabilize.
Fix a commutative noetherian ring $R$ and let
$A_{0}\xrightarrow{\sigma}A_{1}\xrightarrow{\sigma}A_{2}\xrightarrow{\sigma}\cdots$
be a sequence of $R$-modules $A_{n}$ for $n\geq 0$ connected by maps
$\sigma_{n}:A_{n}\to A_{n+1}$. Write
$A=\bigoplus_{n\geq 0}A_{n}.$
The collection of maps $\sigma_{n}$ assemble into an endomorphism $\sigma:A\to
A$, and this map endows $A$ with the structure of a $R[\sigma]$-module. The
next proposition relates finiteness conditions on $A$ and $\sigma$ to a
stability condition on the sequence.
###### Proposition 4.1.2.
With $R$, $A$ and $\sigma$ as above, the following are equivalent:
1. (1)
$A_{n}$ is a finitely generated $R$-module for each $n$, and the sequence
stabilizes: that is, there exists $N\geq 0$ such that for all $n\geq N$,
$\sigma:A_{n}\to A_{n+1}$ is an isomorphism; and
2. (2)
$A$ is finitely generated as an $R[\sigma]$-module.
###### Proof.
To see (1) $\Rightarrow$ (2), let $f_{n}:R^{r_{n}}\to A_{n}$ be a surjection.
Define a map
$f:R[\sigma]^{r_{0}}\oplus\cdots\oplus R[\sigma]^{r_{N}}\to A$
which maps the generators of $R[\sigma]^{r_{n}}$ to $A_{n}\subseteq A$
according to $f_{n}$ and extends $R[\sigma]$-linearly. To see that $f$ is
surjective, note that by construction it surjects onto
$\bigoplus_{n=0}^{N}A_{n}$, and it surjects onto
$\bigoplus_{n=N}^{\infty}A_{n}\cong A_{N}\otimes\mathbb{Z}[\sigma]$ because it
is a $\sigma$-linear map that surjects onto $A_{N}$.
To see (2) $\Rightarrow$ (1), let $f:R[\sigma]^{r}\to A$ be a surjection from
a free $R[\sigma]$-module of finite rank $r$. Since $f(\sigma
R[\sigma]^{r})\subseteq\sigma A$, $f$ descends to a surjective map on
quotients
$\bar{f}:R^{r}\cong R[\sigma]^{r}/\sigma R[\sigma]^{r}\to A/\sigma
A=\operatorname{coker}\sigma$
which exhibits $\operatorname{coker}\sigma$ as a finitely generated
$R$-module. Note that
$\operatorname{coker}\sigma\cong\bigoplus_{n\geq
0}(\operatorname{coker}\sigma_{n}).$
That $\operatorname{coker}\sigma$ is finitely generated thus implies that
$(\operatorname{coker}\sigma)_{n}=0$ for sufficiently large $n$. It follows
that the sequence is eventually surjective, that is, there exists a $M$ such
that $\sigma_{n}$ is surjective for $n\geq M$. If we let
$K_{i}=\operatorname{ker}(A_{M}\to A_{i}),$
for $i\geq M$, we obtain an increasing chain $K_{M}\subseteq
K_{M+1}\subseteq\cdots$ of submodules of $A_{M}$. As a finitely generated
module over the noetherian ring $R$, $A_{M}$ is noetherian. Therefore, there
must exist an $N\geq 0$ at which point this increasing chain stabilizes, that
is $K_{N}=K_{N+1}=\cdots$. It follows that for $n\geq N$, $\sigma_{n}$ is an
isomorphism.
Now we show that $A_{n}$ is finitely generated over $R$ by induction on $n$.
Since $\operatorname{coker}(\sigma)|_{A_{0}}=A_{0}$, we must have that $A_{0}$
is finitely generated. Now assume $n\geq 1$ and assume inductively that
$A_{n-1}$ is finitely generated. Consider the exact sequence
$A_{n-1}\overset{\sigma}{\to}A_{n}\overset{\sigma}{\to}\operatorname{coker}(\sigma)|_{A_{n}}\to
0.$
Since $A_{n-1}$ and $\operatorname{coker}(\sigma)|_{A_{n}}$ are finitely
generated by assumption, we have that $A_{n}$ is also finitely generated. ∎
###### Corollary 4.1.3.
Let $M$ be a stabilizable open, connected manifold with $\dim(M)\geq 2$. If
$H_{d}(C_{k}(M))$ is finitely generated as an abelian group for all $d\geq 0$,
then $H_{*}(\bigsqcup_{k\geq 0}C_{k}(M))$ is finitely generated as a
$\mathbb{Z}[\sigma_{*}]$-module.
We note that the finite generation of $H_{d}(C_{k}(M))$ in Corollary 4.1.3
follows from the finite generation of $H_{d}(M)$. First we need a lemma.
###### Lemma 4.1.4.
Suppose $X$ has a free $G$-action, and $H_{d}(X;\mathbb{Z})$ is a finitely
generated abelian group for all $d$. Then $H_{d}(X/G;\mathbb{Z})$ is finitely
generated.
###### Proof.
There is a fibration $X\to EG\times_{G}X\to BG$, and since the action of $G$
on $X$ is free, we have $EG\times_{G}X\simeq X/G$. There is a Serre spectral
sequence with local coefficients
$E_{2}^{*,*}=H_{*}(BG;H_{*}(X;\mathbb{Z}))\Rightarrow H_{*}(X/G).$
The $E_{2}$-page is the cohomology of the complex
$C_{*}(EG;\mathbb{Z})\otimes_{\mathbb{Z}[G]}H_{*}(X;\mathbb{Z})$ which is
finitely generated in each degree by assumption. Thus the $E_{2}$-term is
finitely generated in each degree, and so is the target. ∎
###### Proposition 4.1.5.
If $H_{d}(M)$ is finitely generated for all $d\geq 0$, then so is
$H_{d}(C_{n}(M))$ for all $d,n\geq 0$.
###### Proof.
By Lemma 4.1.4, it suffices to show that
$H_{d}(\operatorname{Conf}_{n}(M);\mathbb{Z})$ is finitely generated. Since
$\operatorname{Conf}_{n}(M)$ is a manifold, its homology is Poincaré dual to
cohomology with compact support
$H^{d}_{c}(\operatorname{Conf}_{n}(M);\mathcal{O})$ with coefficients in the
orientation sheaf $\mathcal{O}$ (see [Bre97, Example 2.9]). To study this, we
apply [Pet20, Theorem 5.18], which gives a spectral sequence that in our case
specializes to
$E_{1}^{p,q}=\bigoplus_{\begin{subarray}{c}T\in J_{U_{0}}\\\
|T|=n-p\end{subarray}}\bigoplus_{i+j=p+q}\widetilde{H}^{i}(J(T);H^{j}_{c}(\overline{M(T)};\mathcal{O}))\implies
H_{c}^{p+q}(\operatorname{Conf}_{n}(M);\mathcal{O})$
where $J_{U_{0}}$ is a subset of the set of partitions of $\\{1,\dots,n\\}$,
$J(T)$ is a finite CW complex, and $\overline{M(T)}$ is a closed submanifold
of
$\\{(x_{1},..,x_{n})\in M^{n}\hskip 2.0pt:\hskip
2.0ptx_{i}=x_{j}\iff\text{$x_{i}=x_{j}$ if $i$ and $j$ are in the same block
of $T$}\\},$
which is itself a closed submanifold of $M^{n}$. Closed manifolds have
cohomology that is finitely generated in each degree. Since the $E_{1}$ page
of the spectral sequence is finitely generated for each $(p,q)$ and $p,q\geq
0$ we have that the target is finitely generated in each degree. ∎
### 4.2. Instability for adding a trivial orbit
If $M$ is $G$-stabilizable, then we may add a fixed point to obtain a
$G$-equivariant stabilization map
$\sigma_{G/G}:C_{n}(M)\to C_{n+1}(M)$
whose underlying map
$\sigma_{G/G}^{e}=\sigma:C_{n}(M)\to C_{n+1}(M)$
is the classical stabilization map. In this section, we show that the Bredon
homology of $C(M)$ cannot be stable with respect to this stabilization map
alone. For simplicity of exposition, we consider the degree 0 case where
$G=C_{p}$.
###### Proposition 4.2.1.
For $G=C_{p}$ and $M$ an open $G$\- and $e$-stabilizable manifold, the
$\mathbb{Z}[(\sigma_{G/G})_{*}]$-module
$H_{0}^{G}(C(M);\underline{\mathbb{Z}})$ is not finitely generated.
###### Proof.
The cellular chain complex $C^{G}_{*}(C(M))$ computing Bredon homology (cf.
Section 2.1) is given by
$\displaystyle C^{G}_{*}(C(M))$
$\displaystyle=\Big{(}\sum_{G/H}C_{*}(C(M)^{H})\otimes\underline{\mathbb{Z}}(G/H)\Big{)}/\sim$
$\displaystyle=\Big{(}C_{*}(C(M)^{e})\otimes\underline{\mathbb{Z}}(C_{p}/e)\oplus
C_{*}(C(M)^{G})\otimes\underline{\mathbb{Z}}(C_{p}/C_{p})\Big{)}/\sim,$
In the case $G=C_{p}$, the orbit category is generated by the projection
$r:C_{p}/e\to C_{p}/C_{p}$ and automorphism $\gamma:C_{p}/e\to C_{p}/e$ which
corresponds to the action of the generator of $C_{p}$. Therefore, in
$C^{G}_{*}(C(M))$, we have the relations:
* •
$\gamma\cdot x\sim x$ for $x\in C_{*}(C(M)^{e})$ and $\gamma\in C_{p}$ the
generator of $C_{p}$, and
* •
$r^{*}x\sim px$ for $x\in C_{*}(C(M)^{G})$, where $r^{*}:C(M)^{G}\to C(M)^{e}$
is the inclusion of fixed points.
From Proposition 3.2.10, we have that the components of $C(M)^{G}$ are indexed
by finite $G$-sets, so we have
$H_{0}(C(M)^{C_{p}})\cong\mathbb{Z}[\sigma]\\{x_{0},x_{1},x_{2},\ldots\\},$
where $x_{i}$ represents the component corresponding to the $G$-set $i\cdot
C_{p}/e$. The components of $C(M)$ are indexed by natural numbers, so we have
$H_{0}(C(M)^{e})\cong\mathbb{Z}[\sigma]\\{y_{0}\\}.$
We conclude that $H^{G}_{0}(C(M))$ is given by
$H^{G}_{0}(C(M))\cong\frac{\mathbb{Z}[\sigma]\\{x_{0},x_{1},x_{2},\cdots,y_{0}\\}}{px_{i}=\sigma^{i}y_{0}},$
which is not finitely generated as a $\mathbb{Z}[\sigma]$-module. ∎
### 4.3. Homological stability for $S$-configurations
As above, let $M$ be a $G$-manifold. In this section, we fix a conjugacy class
of subgroups $(H)$ and study stabilization maps for $[G/H]$ orbits. We will
combine Corollary 3.2.9 with classical homological stability applied to the
nonequivariant space $M_{(H)}/G$, to obtain homological stability for
$C_{S}^{G}(M)$.
###### Theorem 4.3.1.
Let $M$ be a $G$-manifold and $S$ a finite $G$-set containing $k$ copies of
the orbit $[G/H]$ for a conjugacy class of subgroups $(H)$. If $M$ is
$H$-stabilizable and $M_{(H)}/G$ is connected, then the $(H)$-stabilization
map of (3.3.3) induces an isomorphism
$(\sigma_{G/H})_{*}:H_{d}(C^{G}_{S}(M))\to H_{d}(C^{G}_{S+[G/H]}(M))$
in integral homology in degrees $d\leq k/2$.
###### Proof.
Decompose the finite $G$-set $S$ as
$S\cong k[G/H]\sqcup S^{\prime},$
where $S^{\prime}$ is a finite $G$-set with no orbits isomorphic to $[G/H]$.
Corollary 3.2.9 allows us to decompose the space of unordered
$S$-configurations in $M$ as
$C^{G}_{S}(M)\cong C^{G}_{k[G/H]}(M)\times C^{G}_{S^{\prime}}\cong
C_{k}(M_{(H)}/G)\times C^{G}_{S^{\prime}}(M).$
By Corollary 3.2.9, we have
$\displaystyle H_{*}(C^{G}_{S}(M))$ $\displaystyle\cong
H_{*}(C_{k}(M_{(H)}/G)\times C^{G}_{S^{\prime}}(M)),$ $\displaystyle
H_{*}(C^{G}_{S+[G/H]}(M))$ $\displaystyle\cong H_{*}(C_{k+1}(M_{(H)}/G)\times
C^{G}_{S^{\prime}}(M)).$
By naturality, we obtain a map of Künneth short exact sequences
${\displaystyle\bigoplus_{i+j=d-1}\operatorname{Tor}_{1}\left(H_{i}(C_{k}(M_{(H)}/G)),H_{j}(C^{G}_{S^{\prime}}(M))\right)}$${\displaystyle\bigoplus_{i+j=d-1}\operatorname{Tor}_{1}\left(H_{i}(C_{k+1}(M_{(H)}/G)),H_{j}(C^{G}_{S^{\prime}}(M))\right)}$${H_{d}(C^{G}_{S}(M))}$${H_{d}(C^{G}_{S+[G/H]}(M))}$${\displaystyle\bigoplus_{i+j=d}H_{i}(C_{k}(M_{(H)}/G))\otimes
H_{j}(C^{G}_{S^{\prime}}(M))}$${\displaystyle\bigoplus_{i+j=d}H_{i}(C_{k+1}(M_{(H)}/G))\otimes
H_{j}(C^{G}_{S^{\prime}}(M))}$$\scriptstyle{(\sigma_{G/H})_{*}}$
By Lemma 3.3.6, the top and bottom horizontal arrows in the above diagram are
given by the classical stabilization maps $\sigma:C_{k}(M_{(H)}/G)\to
C_{k+1}(M_{(H)}/G)$ for the manifold $M_{(H)}/G$, and so Theorem 4.1.1 implies
that these maps are isomorphisms for $d\leq k/2$. (Note that $M_{(H)}/G$ is
not empty because we have assumed that there are $k$ points of isotropy type
$(H)$.) It follows from the five lemma that $(\sigma_{G/H})_{*}$ is an
isomorphism in the range $d\leq k/2$. ∎
### 4.4. Bredon homological stability
We now return to the situation of ordinary configuration spaces.
For $K\leq G$, let
$P_{K}=\mathbb{Z}[\sigma_{K/H}:H\subseteq K].$
Clearly, $P_{K}$ acts on $K$-Bredon cohomology $H^{K}_{*}(C(M))$, but we can
also consider the action of $P_{G}$ on $H^{K}_{*}(C(M))$ via the restriction
map $\operatorname{res}^{G}_{K}:P_{G}\to P_{K}$, which is defined as follows.
Given $H,K\leq G$, the double coset formula gives an isomorphism of $K$-sets
$G/H\ \ \cong\bigsqcup_{KgH\in K\backslash G/H}K/(K\cap gHg^{-1}),$
and so
(4.4.1) $\operatorname{res}^{G}_{K}(\sigma_{G/H})\ \ =\prod_{KgH\in
K\backslash G/H}\sigma_{K/(K\cap gHg^{-1})}.$
###### Lemma 4.4.2.
Suppose $G$ is a Dedekind group (i.e., every subgroup is normal). Then for all
$K\leq G$, $P_{K}$ is finitely generated as a $P_{G}$-module, where the action
is via the restriction map $\operatorname{res}^{G}_{K}:P_{G}\to P_{K}$.
###### Proof.
As $P_{K}$ is a finitely generated algebra, it suffices to show that $P_{K}$
is integral over $P_{G}$. We will show that there is a power of every
generator $\sigma_{K/H}$ (for $K\leq H$) in the image of
$\operatorname{res}^{G}_{K}$, which implies the same for every monomial.
Combining the Dedekind assumption with (4.4.1) shows that
$\operatorname{res}^{G}_{K}(\sigma_{G/H})=\sigma_{K/H}^{[G:K]}$. ∎
###### Remark 4.4.3.
The “Dedekind” hypothesis is needed to ensure that
$\operatorname{res}^{G}_{K}(\sigma_{G/H})=\sigma_{K/H}^{[G:K]}$ in the last
line of the previous proof. It is known [Ded97] that a finite group is
Dedekind if and only if it is abelian or isomorphic to $Q_{8}\times A$ where
$A$ is a finite abelian group whose 2-component has the form
$(\mathbb{Z}/2)^{r}$ for $r\geq 0$.
We remark that the conclusion of Lemma 4.4.2 holds for a more general class of
nonabelian groups than those mentioned in the previous paragraph, but an
explicit characterization is not known. For example, it can be checked that it
holds for dihedral groups of order $2n$ for $n$ squarefree.
###### Lemma 4.4.4.
Assume $M$ is $H$-stabilizable and $M_{(H)}/G$ is connected for every $H\leq
G$. For every $d$ there exists $N_{d}$ such that
$\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}}\wedge C(M)))\cong
P_{G}\cdot\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}}\wedge C_{\leq N_{d}}(M))).$
Moreover, if $H_{i}(M_{(H)}/G)$ is a finitely generated abelian group for all
$i\leq d$ and $(H)\leq G$, then
$\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}}\wedge C_{\leq N_{d}}(M)))$ is a
finitely generated abelian group.
###### Proof.
Using the splitting in Corollary 3.2.9 we have
$\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}}\wedge
C(M)))\cong\pi_{d}(\Phi^{G}H\underline{\mathbb{Z}}\wedge
C(M)^{G})\cong\bigoplus_{G\text{-sets
}S}\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}})\wedge C_{S}^{G}(M)).$
By Proposition 2.5.1, there is a finite collection of nonnegative integers
$n_{i}$ and finitely generated abelian groups $A_{i}$ such that
$\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}})\wedge
C_{S}^{G}(M))\cong\bigoplus_{i}H_{d-n_{i}}(C_{S}^{G}(M);A_{i}).$
For fixed $i$ and $H$, Theorem 4.3.1 says that there exists $N_{d,i,H}$ such
that
$H_{d-n_{i}}(C_{S}^{G}(M);A_{i})=\sigma_{G/H}\cdot
H_{d-n_{i}}(C_{S-[G/H]}^{G}(M);A_{i})$
whenever $S$ contains at least $N_{d,i,H}$ copies of $G/H$. For fixed $i$,
there are finitely many $G$-sets $S$ such that
$H_{d-n_{i}}(C_{S}^{G}(M);A_{i})$ is not in the image of $\sigma_{G/H}$ for
some $H$. Let $N_{d,i}$ be the maximum cardinality of the exceptional
$G$-sets, and let $N_{d}=\max_{i}\\{N_{d,i}\\}$. Then for every $S$ and
$n_{i}$, we have
$\displaystyle H_{d-n_{i}}(C_{S}^{G}(M);A_{i})$ $\displaystyle\subseteq
P_{G}\cdot\bigoplus_{|S^{\prime}|\leq
N_{d}}H_{d-n_{i}}(C_{S^{\prime}}^{G}(M);A_{i}).$
Summing over $i$, we obtain $\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}}\wedge
C^{G}_{S}(M)))\subseteq
P_{G}\cdot\pi_{d}(\Phi^{G}(H\underline{\mathbb{Z}}\wedge C_{\leq N_{d}}(M)))$.
This shows the first statement. For the second statement, combine Corollary
3.2.9 and Proposition 4.1.5, which holds for homology with
$A_{i}$-coefficients because $A_{i}$ is a finitely generated abelian group. ∎
###### Proposition 4.4.5.
Let $G$ be a Dedekind group. For $K\leq G$, let $i_{K}^{*}$ denote restriction
from $G$-spaces to $K$-spaces. Assume that for all $H\leq K\leq G$, $M$ is
$H$-stabilizable, $(i_{K}^{*}M)_{(H)}/K$ is connected, and
$H_{i}(i_{K}^{*}(M)_{(H)}/K)$ is a finitely generated abelian group for all
$i\leq d+1$. Then $H_{d}^{K}(C(M);\underline{\mathbb{Z}})$ is finitely
generated over $P_{G}$ for all $K\leq G$.
###### Proof.
We will apply Lemma 2.4.2 with the Serre class of finitely generated
$P_{G}$-modules. It suffices to check the hypothesis, namely that
$\pi_{i}(\Phi^{K}(H\underline{\mathbb{Z}}\wedge C(M)))$ is finitely generated
over $P_{G}$ for all $K\leq G$ and $i\leq d+1$. We will apply Lemma 4.4.4 to
$i^{*}_{K}(M)$. Checking the hypotheses, first note that if $M$ is
$H$-stabilizable as a $G$-manifold, then $i^{*}_{K}(M)$ is $H$-stabilizable.
Thus $\pi_{i}(\Phi^{K}(H\underline{\mathbb{Z}}\wedge
C(i^{*}_{K}(M))))=\pi_{i}(\Phi^{K}(H\underline{\mathbb{Z}}\wedge C(M)))$ is
finitely generated over $P_{K}$. Now use Lemma 4.4.2 to get finite generation
over $P_{G}$. ∎
### 4.5. Simplifying the hypotheses of Proposition 4.4.5
We will show (Lemma 4.5.3) that the finite generation hypothesis in
Proposition 4.4.5 follows from the assumption that $M$ the interior of a
compact $G$-manifold $W$. Such finite type hypotheses are common in the
literature on this subject; see for example [RW13] for the non-equivariant
version of our result. We also translate the connectedness hypothesis of
Proposition 4.4.5 into a condition which involves $M^{H}$ instead of $M_{(H)}$
(see Lemma 4.5.1). The results are combined in Theorem 4.5.5.
###### Lemma 4.5.1.
Suppose $G$ is Dedekind or a $p$-group, and let $W=W_{G}H$ denote the Weyl
group. If $M^{H}/W$ is connected then $M_{(H)}/G$ (if nonempty) is connected.
Moreover,
$M_{(H)}/G\cong\Big{(}M^{H}\backslash\bigcup_{e<L\leq W}(M^{H})^{L}\Big{)}/W.$
###### Proof.
Applying Lemma 3.2.6 to the $G$-manifold $M$ and the $W$-manifold $M^{H}$,
respectively, we have
$\displaystyle M_{(H)}/G$
$\displaystyle\cong\Big{(}M^{H}\backslash\bigcup_{H<K\leq
G}M^{K}\Big{)}\big{/}W$ $\displaystyle(M^{H})_{(e)}/W$
$\displaystyle\cong\Big{(}M^{H}\backslash\bigcup_{e<L\leq
W}(M^{H})^{L}\Big{)}\big{/}W\cong\Big{(}M^{H}\backslash\bigcup_{H<K\leq
N_{G}(H)}M^{K}\Big{)}\Big{/}W.$
Because of the group-theoretic conditions, $H<K\leq G$ implies $H<N_{K}(H)$,
and hence $M_{(H)}/G\cong(M^{H})_{(e)}/W$.
If $M_{(H)}$ is nonempty, then $(e)$ is the principal orbit type of $M^{H}$ as
a $W$-manifold. Then [tD87, Theorem I.5.14] applied to $M^{H}$ says that
$(M^{H})_{(e)}/W\cong M_{(H)}/G$ is connected. ∎
The next lemma is used to prove Lemma 4.5.3.
###### Lemma 4.5.2.
Suppose $M$ is $G$-manifold that is the interior of a compact $G$-manifold
with boundary. Let $H\leq G$. If $A$ is a $\pi_{1}(M^{H})$-module that is
finitely generated as an abelian group, then the homology with local
coefficients $H_{q}(M^{H};A)$ is a finitely generated abelian group for every
$q$.
###### Proof.
Let $Z$ be a compact $G$-manifold whose interior is $M$. First, we observe
that $M^{H}$ is the interior of $Z^{H}$; this can be checked on Euclidean
neighborhoods. Since $Z^{H}$ is a compact manifold with boundary, it has a
collar neighborhood, and this can be used to produce a homotopy equivalence
$M^{H}\simeq Z^{H}$. Since $\pi_{1}(M^{H})=\pi_{1}(Z^{H})$, we may view $A$ as
a sheaf of abelian groups over $Z^{H}$, and consider an open cover
$\mathscr{U}$ that trivializes $A$. There is a refinement
$\mathscr{U}^{\prime}$ that is a good cover (see [BT82, Corollary 5.2]) and
since $Z^{H}$ is compact, we may take $\mathscr{U}^{\prime}$ to be finite.
Since this is a good cover, homology with local coefficients
$H_{q}(Z^{H};A)=H_{q}(M^{H};A)$ agrees with Čech cohomology
$\check{H}_{q}(\mathscr{U}^{\prime};A)$. The Čech complex
$\check{C}_{q}(\mathscr{U}^{\prime};A)=\bigoplus_{(i_{0},\dots,i_{q})}A(U_{i_{0}}\cap\dots\cap
U_{i_{q}})=\bigoplus_{(i_{0},\dots,i_{q})}A$
is a finitely generated abelian group in each degree, and hence so is its
homology. ∎
###### Lemma 4.5.3.
Assume $G$ is abelian, and $M$ is the interior of a compact $G$-manifold. Then
$H_{d}(M_{(H)}/G;\mathbb{Z})$ is finitely generated for all
$d\in\mathbb{Z},H\leq G$.
###### Proof.
Let $W=W_{G}H$ be the Weyl group. Lemma 4.5.1 says that $M_{(H)}/G\cong X/W$
where
(4.5.4) $X=M^{H}\backslash\bigcup_{e<L\leq W}(M^{H})^{L}.$
By Lemma 4.1.4, it suffices to show that $H_{*}(X;\mathbb{Z})$ is finitely
generated in each degree.
If $q:G\to W=G/H$ is the quotient, then $(M^{H})^{L}=M^{q^{-1}(L)}$. Abusing
notation, we write $M^{L}=(M^{H})^{L}$; also let $U_{L}=M^{H}\backslash
M^{L}$. Since $M^{L}$ is a submanifold, there is a normal bundle $N(M^{L})$
which is open, and $N(M^{L})\cup U_{L}\cong M^{H}$. First we will show
$H_{d}(U_{L})$ is finite using a Mayer–Vietoris sequence.
The intersection term $N(M^{L})\cap U_{L}$ is homotopy equivalent to the
sphere bundle $S(N(M^{L}))$. There is a Serre spectral sequence associated to
the fibration $S^{n}\to S(N(M^{L}))\to M^{L}$,
$E_{2}^{*,*}=H_{*}(M^{L};H_{*}(S^{n};\mathbb{Z}))\Rightarrow
H_{*}(S(N(M^{L}));\mathbb{Z})$
where $\pi_{1}(M^{L})$ may act nontrivially on $H_{*}(S^{n};\mathbb{Z})$. By
Lemma 4.5.2, the $E_{2}$ term is finitely generated in each degree, and hence
so is the target.
Now consider the Mayer–Vietoris sequence
$\dots\to H_{d+1}(M^{H})\to H_{d}(S(N(M^{L})))\to H_{d}(N(M^{L}))\oplus
H_{d}(U_{L})\to\dots.$
The first term is finitely generated by Lemma 4.5.2 and above we showed the
second is as well. Hence the third term is finitely generated; in particular
so is $H_{d}(U_{L})$.
Since $G$ is abelian, write $W=\prod_{i}\mathbb{Z}/p_{i}^{r_{i}}$. In the
union (4.5.4), it suffices to restrict $L$ to products of elementary abelian
subgroups $\mathbb{Z}/p_{i}\cong\langle
p_{i}^{r_{i}-1}\rangle\leq\mathbb{Z}/p_{i}^{r_{i}}$, since any nontrivial
subgroup $H\leq\mathbb{Z}/p_{i}^{r_{i}}$ contains this $\mathbb{Z}/p_{i}$. If
$L\times L^{\prime}\leq W$, we have
$(M^{H})^{L}\cap(M^{H})^{L^{\prime}}=(M^{H})^{L\times L^{\prime}}$. More
generally, given $L=\prod_{i}L_{i}$ and $L^{\prime}=\prod_{i}L^{\prime}_{i}$,
note that $LL^{\prime}\leq W$ is a product of all $L_{i}$ and $L^{\prime}_{i}$
factors, with duplicates removed. Thus we have
$U_{L}\cup
U_{L^{\prime}}=M^{H}\backslash((M^{H})^{L}\cap(M^{H})^{L^{\prime}})=M^{H}\backslash(M^{H})^{LL^{\prime}}.$
We will show that
$X=\bigcap_{\begin{subarray}{c}L=L_{1}\times\dots\times L_{n}\\\ L_{i}\text{
elem.ab.}\end{subarray}}U_{L}$
has finitely generated homology, by induction on the number of sets
intersected. The base case was shown above. Suppose any intersection of $n$
such sets has finitely generated homology, and consider
$U_{L_{1}}\cap\dots\cap U_{L_{n}}\cap U_{L}$. There is a Mayer-Vietoris
sequence
$\dots\to H_{d+1}((\cap_{i=1}^{n}U_{L_{i}})\cup U_{L})\to
H_{d}((\cap_{i=1}^{n}U_{L_{i}})\cap U_{L})\to
H_{d}(\cap_{i=1}^{n}U_{L_{i}})\oplus H_{d}(U_{L})\to\dots.$
The inductive hypothesis implies that the third term is finitely generated, so
in order to show the second term is finitely generated, it suffices to show
the first term is finitely generated. We have
$\Big{(}\bigcap_{i=1}^{n}U_{L_{i}}\Big{)}\cup
U_{L}=\bigcap_{i=1}^{n}(U_{L_{i}}\cup U_{L})=\bigcap_{i=1}^{n}U_{L_{i}L}$
and the inductive hypothesis shows this has finitely generated homology. ∎
###### Theorem 4.5.5.
Let $G$ be an abelian group and let $M$ be a $G$-manifold that is the interior
of a compact $G$-manifold with boundary. Assume that for all $H\leq G$, $M$ is
$H$-stabilizable and $M^{H}$ is connected. Then
$H_{d}^{H}(C(M);\underline{\mathbb{Z}})$ is finitely generated over $P_{G}$
for all $H\leq G$.
###### Proof.
Combine Proposition 4.4.5, Lemma 4.5.1, and Lemma 4.5.3. ∎
###### Example 4.5.6.
If $G$ is abelian and $V=n\rho_{G}$ is a sum of regular representations, the
hypotheses of Theorem 4.5.5 are satisfied. One could also check the hypotheses
of Proposition 4.4.5 directly without using the results of this subsection: by
Lemma 3.2.6 the space $V_{(H)}$ arises from the complement of hyperplanes in
$n\rho_{G}$. This complement is Alexander dual to a union of spheres, one for
each hyperplane.
## 5\. Generalizations and variants
Throughout this section, $G$ denotes a finite group in which every subgroup is
normal. We have established stability for _$\mathbb{Z}$ -graded_ Bredon
homology _groups_. In this section, we discuss natural extensions of our
results to
1. (1)
$RO(G)$-graded Bredon homology groups, and
2. (2)
$\mathbb{Z}$-graded Bredon homology Mackey functors.
Here, $RO(G)$ is the (group completion) of the ring of isomorphism classes of
real orthogonal representations of $G$.
### 5.1. $RO(G)$-graded Bredon homological stability
Let $V\in RO(G)$. We write $-V$ for the additive inverse of $V$ in $RO(G)$ and
write $S^{V}$ for the representation sphere associated to $V$.
###### Definition 5.1.1.
Let $V\in RO(G)$. The $V$-th Bredon homology group of a $G$-space $X$ with
coefficients in the Mackey functor $\underline{M}$ is the group
$\pi_{0}^{G}\left(S^{-V}\wedge H\underline{M}\wedge X\right)\cong
H_{0}^{G}(S^{-V}\wedge X;\underline{M}).$
The following is well-known:
###### Lemma 5.1.2.
If $V\in RO(G)$, then we have
$(S^{V})^{e}=\Phi^{e}(S^{V})=S^{|V|},\quad\Phi^{G}(S^{V})=S^{|V^{G}|}.$
Recall that $P_{G}=\mathbb{Z}[(\sigma_{G/H})_{*}:(H)\leq G]$.
###### Theorem 5.1.3.
Let $V\in RO(G)$ and let $M$ be a $G$-manifold satisfying the hypotheses of
Theorem 4.5.5. Then the $P_{G}$-module
$H_{V}^{G}(C(M);\underline{\mathbb{Z}})$ is finitely generated.
###### Proof.
Applying Lemma 2.4.2 with $X=S^{-V}\wedge C(M)$ and $\mathcal{C}$ the Serre
class of finitely generated $P_{G}$-modules, it suffices to show that
$\pi_{q}\Phi^{K}(H\underline{\mathbb{Z}}\wedge S^{-V}\wedge C(M))$ is in
$\mathcal{C}$ for all $K\leq G$ and all $q\in\mathbb{Z}$. Lemma 5.1.2 implies
that
$\pi_{q}\Phi^{K}\left(H\underline{\mathbb{Z}}\wedge S^{-V}\wedge
C(M)\right)\cong\pi_{q+|V^{K}|}\Phi^{K}\left(H\underline{\mathbb{Z}}\wedge
C(M)\right),$
and this is finitely generated by combining the same ingredients as the proof
of Proposition 4.4.5 and Theorem 4.5.5. ∎
### 5.2. Mackey functor stability
If $L\leq G$ is a subgroup, we may view any $G$-manifold $M$ as an
$L$-manifold by restriction. If the $G$-manifold $M$ satisfies the hypotheses
of B, then its underlying $L$-manifold clearly also satisfies the conditions
of B (with $L$ in place of $G$) and we obtain the following:
###### Corollary 5.2.1.
Suppose $M$ satisfies the hypotheses of Theorem 4.5.5. Then
$H_{d}^{L}(C(M);\underline{\mathbb{Z}})$ is finitely generated over $P_{L}$
for all integers $d\geq 0$.
Recall that Bredon homology is actually a Mackey functor, i.e., an additive
functor from the Burnside category of $G$ to abelian groups. The previous
corollary implies that the value of the Mackey functor
$H_{d}^{(-)}(C(M);\underline{\mathbb{Z}})$
stabilizes when evaluated on any object in the Burnside category (i.e., any
finite $G$-set). We can encode this levelwise stability, and more, with the
following notion:
###### Definition 5.2.2.
A _module_ over a Green functor $\underline{R}$ is a Mackey functor
$\underline{M}$ equipped with an action map
$\underline{R}\boxtimes\underline{M}\to\underline{M}$ satisfying the usual
associativity and unitality conditions.
An $\underline{R}$-module $\underline{M}$ is _finitely generated_ if there is
a surjective $\underline{R}$-linear map
$\underline{R}\\{x_{T}\\}\to\underline{M}$, where $\underline{R}\\{x_{T}\\}$
is the free $\underline{R}$-module on a generator $x_{T}$ with $T$ a finite
$G$-set.
According to [Shu10, Lem. 2.17], a map
$\underline{R}\boxtimes\underline{M}\to\underline{M}$ determines, and is
determined by, its _Dress pairing_ : a collection of maps
$\underline{R}(S)\otimes\underline{M}(S)\to\underline{M}(S)$ for $S$ in the
Burnside category of $G$ satisfying certain compatibility axioms. Using Dress
pairings, we can relate levelwise finite generation to Mackey finite
generation:
###### Lemma 5.2.3.
An $\underline{R}$-module $\underline{M}$ is finitely generated if and only if
$\underline{M}(S)$ is finitely generated as a $\underline{R}(S)$-module for
each finite $G$-set $S$.
###### Proof.
Surjectivity for Mackey functors is checked levelwise. ∎
To express stability in terms of Mackey functors, we need a Green functor to
encode the action of the equivariant stabilization maps for each subgroup of
$G$. This is accomplished through the following:
###### Definition 5.2.4.
For each finite group $G$, let $\underline{P}_{G}$ be the constant Green
functor on $P_{G}=\mathbb{Z}[(\sigma_{G/H})_{*}:(H)\leq G]$. (_A priori_ ,
$\underline{P}_{G}$ is only a Mackey functor. It is a Green functor since it
is the fixed point functor of a commutative ring with trivial action.) We
define an action of $\underline{P}_{G}$ on
$H_{*}^{(-)}(C(M);\underline{\mathbb{Z}})$ via Dress pairing: the map
$\underline{P}_{G}(G/H)\otimes H_{*}^{H}(C(M);\underline{\mathbb{Z}})\cong
P_{G}\otimes H^{H}_{*}(C(M);\underline{\mathbb{Z}})\to
H_{*}^{H}(C(M);\underline{\mathbb{Z}})$
comes from the restriction map $\operatorname{res}^{G}_{H}:P_{G}\to P_{H}$ as
defined in Section 4.4.
In particular, if $H_{*}^{H}(C(M);\underline{\mathbb{Z}})$ is finitely
generated over $\underline{P}_{H}(H/H)$ for every subgroup $H\leq G$, and if
$\underline{P}_{H}(H/H)$ is finitely generated as a
$\underline{P}_{G}(G/H)$-module via the restriction map (e.g., if every
subgroup of $G$ is normal), then $H_{*}^{H}(C(M);\underline{\mathbb{Z}})$ is
finitely generated as a $\underline{P}_{G}(G/H)$-module. Combined with Lemma
5.2.3, this proves:
###### Theorem 5.2.5.
Let $M$ be as in Theorem 4.5.5. Then
$H_{d}^{(-)}(C(M);\underline{\mathbb{Z}})$ is finitely generated over
$\underline{P}_{G}$ for all integers $d\geq 0$.
## References
* [ADCK20] Byung Hee An, Gabriel Drummond-Cole, and Ben Knudsen. Edge stabilization in the homology of graph braid groups. Geom. Topol., 24(1):421–469, 2020.
* [Aud91] Michèle Audin. The topology of torus actions on symplectic manifolds, volume 93 of Progress in Mathematics. Birkhäuser Verlag, Basel, 1991. Translated from the French by the author.
* [Bon22] Luciana Basualdo Bonatto. Decoupling decorations on moduli spaces of manifolds. Mathematical Proceedings of the Cambridge Philosophical Society, pages 1–36, 2022.
* [Bre97] Glen E. Bredon. Sheaf theory, volume 170 of Graduate Texts in Mathematics. Springer-Verlag, New York, second edition, 1997.
* [BT82] Raoul Bott and Loring W. Tu. Differential Forms in Algebraic Topology. Springer New York, 1982.
* [BT01] Carl-Friedrich Bödigheimer and Ulrike Tillmann. Stripping and splitting decorated mapping class groups. In Cohomological methods in homotopy theory (Bellaterra, 1998), volume 196 of Progr. Math., pages 47–57. Birkhäuser, Basel, 2001.
* [CEF15] Thomas Church, Jordan S. Ellenberg, and Benson Farb. FI-modules and stability for representations of symmetric groups. Duke Math. J., 164(9):1833–1910, 2015.
* [CLM76] Frederick R. Cohen, Thomas J. Lada, and J. Peter May. The homology of iterated loop spaces. Lecture Notes in Mathematics, Vol. 533. Springer-Verlag, Berlin-New York, 1976.
* [CW91] Steven R. Costenoble and Stefan Waner. Fixed set systems of equivariant infinite loop spaces. Trans. Am. Math. Soc., 326(2):485–505, 1991.
* [Ded97] R. Dedekind. Ueber Gruppen, deren sämtliche Teiler Normalteiler sind. Math. Ann., 48:548–561, 1897.
* [DH] Daniel Dugger and Christy Hazel. The ${R}{O}({G})$-graded Bredon cohomology of Euclidean configuration spaces. To appear.
* [GM17] Bertrand J. Guillou and J. Peter May. Equivariant iterated loop space theory and permutative $G$-categories. Algebr. Geom. Topol., 17(6):3259–3339, 2017.
* [HHR16] Michael A Hill, Michael J Hopkins, and Douglas C Ravenel. On the nonexistence of elements of Kervaire invariant one. Annals of Mathematics, 184(1):1–262, 2016.
* [Kan07] Marja Kankaanrinta. Equivariant collaring, tubular neighbourhood and gluing theorems for proper Lie group actions. Algebr. Geom. Topol., 7:1–27, 2007.
* [Kri20] Sophie Kriz. Notes on equivariant homology with constant coefficients. Pacific J. Math., 309(2):381–399, 2020.
* [Lew] L. Gaunce Lewis, Jr. The theory of Green functors. Unpublished manuscript.
* [May72] J. P. May. The geometry of iterated loop spaces. Lecture Notes in Mathematics, Vol. 271. Springer-Verlag, Berlin-New York, 1972.
* [May96] J. P. May. Equivariant homotopy and cohomology theory, volume 91 of CBMS Regional Conference Series in Mathematics. Published for the Conference Board of the Mathematical Sciences, Washington, DC; by the American Mathematical Society, Providence, RI, 1996. With contributions by M. Cole, G. Comezaña, S. Costenoble, A. D. Elmendorf, J. P. C. Greenlees, L. G. Lewis, Jr., R. J. Piacenza, G. Triantafillou, and S. Waner.
* [McD75] Dusa McDuff. Configuration spaces of positive and negative particles. Topology, 14:91–107, 1975.
* [MW19] Jeremy Miller and Jennifer Wilson. Higher-order representation stability and ordered configuration spaces of manifolds. Geom. Topol., 23(5):2519–2591, 2019.
* [Pal18] Martin Palmer. Twisted homological stability for configuration spaces. Homology Homotopy Appl., 20(2):145–178, 2018.
* [Pet20] Dan Petersen. Cohomology of generalized configuration spaces. Compos. Math., 156(2):251–298, 2020.
* [RS00] Colin Rourke and Brian Sanderson. Equivariant configuration spaces. J. Lond. Math. Soc., II. Ser., 62(2):544–552, 2000.
* [RW13] Oscar Randal-Williams. Homological stability for unordered configuration spaces. Q. J. Math., 64(1):303–326, 2013.
* [Seg79] Graeme Segal. The topology of spaces of rational functions. Acta Math., 143(1-2):39–72, 1979.
* [Shu10] Megan Elizabeth Shulman. Equivariant local coefficients and the ${R}{O}({G})$-graded cohomology of classifying spaces. PhD thesis, The University of Chicago, 2010.
* [Sta18] The Stacks Project Authors. Stacks Project. https://stacks.math.columbia.edu, 2018.
* [tD87] Tammo tom Dieck. Transformation groups, volume 8 of De Gruyter Studies in Mathematics. Walter de Gruyter & Co., Berlin, 1987.
* [Wil19] Dylan Wilson. ${C}_{2}$-equivariant Homology Operations: Results and Formulas. arXiv preprint arXiv:1905.00058, 2019.
|
# Busemann process and semi-infinite geodesics in Brownian last-passage
percolation
Timo Seppäläinen Timo Seppäläinen, University of Wisconsin-Madison,
Mathematics Department, Van Vleck Hall, 480 Lincoln Dr., Madison WI
53706-1388, USA<EMAIL_ADDRESS>and Evan Sorensen Evan Sorensen,
University of Wisconsin-Madison, Mathematics Department, Van Vleck Hall, 480
Lincoln Dr., Madison WI 53706-1388, USA<EMAIL_ADDRESS>
###### Abstract.
We prove the existence of semi-infinite geodesics for Brownian last-passage
percolation (BLPP). Specifically, on a single event of probability one, there
exist semi-infinite geodesics started from every space-time point and
traveling in every asymptotic direction. Properties of these geodesics include
uniqueness for a fixed initial point and direction, non-uniqueness for fixed
direction but random initial points, and coalescence of all geodesics
traveling in a common, fixed direction. Along the way, we prove that for fixed
northeast and southwest directions, there almost surely exist no bi-infinite
geodesics in the given directions. The semi-infinite geodesics are constructed
from Busemann functions. Our starting point is a result of Alberts, Rassoul-
Agha and Simper that established Busemann functions for fixed points and
directions. Out of this, we construct the global process of Busemann functions
simultaneously for all initial points and directions, and then the family of
semi-infinite Busemann geodesics. The uncountable space of the semi-discrete
setting requires extra consideration and leads to new phenomena, compared to
discrete models.
###### Key words and phrases:
Brownian motion, Busemann function, semi-infinite geodesic, coalescence, last-
passage percolation, midpoint problem
###### 2020 Mathematics Subject Classification:
60K30,60K35,60K37
###### Contents
1. 1 Introduction
2. 2 Definitions and previous results
3. 3 Main results
4. 4 Construction and properties of the semi-infinite geodesics
5. 5 Connections to other models
6. 6 Construction and proofs for the Busemann process
7. 7 Proofs of the results for semi-infinite geodesics
8. A Deterministic facts about continuous functions
9. B Probabilistic results
10. C The Brownian queue and stationary last-passage process
11. D Time reversal
## 1\. Introduction
### 1.1. Brownian last-passage percolation
Brownian last-passage percolation (BLPP) dates back to the 1991 work of Glynn
and Whitt [GW91], where the model appeared as a large-scale limit of multiple
queues in series under heavy traffic conditions. Harrison and Williams [Har85,
HW90, HW92] also studied what is known as the Brownian queue and developed a
stability result for that model. More specifically, if the arrivals process is
given by increments of Brownian motion and the service process is given by
increments of an independent Brownian motion with drift, the departures
process is also given by increments of Brownian motion. The connection between
BLPP and the Brownian queue is expounded on by O’Connell and Yor [OY01], who
also introduced the positive temperature version of this model, known as the
Brownian polymer. We discuss the connection to queuing theory in Section 5.3
and Appendix C.
### 1.2. BLPP in the Kardar-Parisi-Zhang universality class
In the early 2000s, Gravner, Tracy, and Widom [GTW01] and Baryshnikov [Bar01]
discovered that the Brownian last-passage value $L_{(1,0),(n,1)}(\mathbf{B})$
(to be defined in (2.2) below) has the same distribution as the largest
eigenvalue of an $n\times n$ GUE random matrix. Soon after, O’Connell and Yor
[OY02] provided an alternate proof relying on the queuing interpretation of
the model. Since then, Brownian last-passage percolation and the Brownian
polymer have been widely studied as gateways to properties of the KPZ
universality class. As a semi-discrete model with one continuous and one
discrete parameter, BLPP serves as an intermediary between discrete models
such as LPP on the planar integer lattice and continuum models such as the
stochastic heat equation and the KPZ equation. In [CH14], Corwin and Hammond
constructed the Airy line ensemble and proved that certain statistics of
Brownian last-passage percolation converge in distribution to the Airy line
ensemble. Dauvergne, Nica, and Virág ([DNV19], Corollary 6.4) provided an
alternate method to prove this fact, which couples Brownian last-passage
percolation with geometric random walks.
Recently, Dauvergne, Ortmann, and Virág [DOV18] constructed the directed
landscape, a central object in the KPZ universality class, as the scaling
limit of Brownian last-passage percolation. Another central object is the KPZ
fixed point, constructed by Matetski, Quastel, and Remenik [MQR21], as the
limit of the totally asymmetric simple exclusion process. In [DOV18], a
variational duality is described between the directed landscape and KPZ fixed
point, which was rigorously proven in [NQR20]. Even more recently, convergence
to the KPZ fixed point was shown for a larger class of models, including the
height function of the KPZ equation. This was done independently by Quastel
and Sarkar [QS20] and Virág [Vir20].
### 1.3. Semi-infinite geodesics in discrete models
The study of infinite geodesics in planar random growth models has gone
through a number of stages over the last 30 years, beginning with the work of
Licea and Newman [LN96, New95] on first-passage percolation with i.i.d. edge
weights. Under a global curvature assumption on the limit shape, for
continuously distributed edge weights, they proved existence of a
deterministic full-Lebesgue measure set of directions in which there is a
unique semi-infinite geodesic out of every lattice point. They also showed
that, for each direction in this set, the semi-infinite geodesics in that
direction all coalesce.
A separate strand of work consists of long-term efforts to prove the
nonexistence of bi-infinite geodesics. In first-passage percolation, Licea and
Newman [LN96] showed that there are no bi-infinite geodesics in fixed
northeast and southwest directions. Howard and Newman [HN01] later proved
similar results for Euclidean last-passage percolation. Around this time, Wehr
and Woo [WW98] proved that, under a first moment assumption on the edge
weights, there are no bi-infinite geodesics for first-passage percolation that
lie entirely in the upper-half plane.
In 2016, Damron and Hanson [DH17] strengthened the result of Licea and Newman
by proving that, if the weights have continuous distribution and the boundary
of the limit shape is differentiable, for each fixed direction, there are no
bi-infinite geodesics with one end having that direction. The conjectured
nonexistence was finally resolved in exponential last-passage percolation
(LPP), known also as the exponential corner growth model (CGM). The proofs
came in two independent works: first by Basu, Hoffman, and Sly [BHS19] and
shortly thereafter by Balázs, Busani, and the first author [BBS20]. The latter
proof is in spirit aligned with the development in the present paper, as it
rested on understanding the joint distribution of the Busemann functions from
[FS20].
Another focus of research has been the coalescence structure of geodesics.
Ferrari and Pimentel [FP05] imported the Licea–Newman approach [LN96, New95]
to exponential LPP on the lattice. Later Pimentel [Pim16] developed a
probabilistic duality between the coalescence time of two semi-infinite
geodesics and their last exit times from the initial boundary, again in
exponential LPP. A key idea was the equality in distribution of the tree of
directed semi-infinite geodesics and the dual tree of southwest-directed
geodesics. From this, he obtained a lower bound on tail probabilities of the
coalescence time for two geodesics starting from $(-\lfloor
k^{2/3}\rfloor,\lfloor k^{2/3}\rfloor)$ and $(\lfloor k^{2/3}\rfloor,-\lfloor
k^{2/3}\rfloor)$. Coalescence bounds have seen significant recent improvement
in [BSS19] and [SS20]. A study of the coalescence structure of finite
geodesics in BLPP was undertaken by Hammond in four papers [Ham16, Ham19a,
Ham19b, Ham20]. The existence of semi-infinite geodesics in the Airy line
ensemble, for a countable dense set of directions and initial points, was
proven recently by Sarkar and Virág in [SV21].
Hoffman was the first to use ideas of Busemann functions to study geodesics in
first-passage percolation. In [Hof08] he showed the existence of disjoint
semi-infinite geodesics. Later, Damron and Hanson [DH14] constructed
generalized Busemann functions from weak subsequential limits of first-passage
times. This allowed them to develop results for semi-infinite geodesics under
weaker assumptions than the global curvature. Under the assumption that the
limit shape is strictly convex and differentiable, they proved that every
semi-infinite geodesic has an asymptotic direction and that, in every
direction, there exists a semi-infinite geodesic out of every lattice point.
On the side of discrete last-passage percolation with general i.i.d. weights,
Georgiou, Rassoul-Agha, and the first author [GRAS17b, GRAS17a] used the
stationary LPP process to prove existence of Busemann functions under mild
moment conditions. The Busemann functions were then used to construct semi-
infinite geodesics. They also showed that if the shape function is strictly
concave, every semi-infinite geodesic has an asymptotic direction. Further in
this direction, in [Sep20] the first author introduced a new proof of the
coalescence of semi-infinite geodesics that utilizes the stationary LPP
process.
Our work is related to this last approach. We use Busemann functions to
construct and study semi-infinite geodesics in BLPP. Existence of Busemann
functions in BLPP, for fixed initial points and directions, was recently
established by Alberts, Rassoul-Agha and Simper [ARAS20], along with Busemann
functions and infinite polymer measures for the semi-discrete Brownian
polymer. Their BLPP result is the starting point of our study.
### 1.4. New techniques and phenomena in the semi-discrete model
The present paper develops the global setting of Busemann functions and semi-
infinite geodesics in the semi-discrete BLPP, with a view to future study of
their finer properties. The novelty lies in going beyond the discrete set-up.
In discrete last-passage percolation, one can prove a probability-one
statement about semi-infinite geodesics out of a fixed initial point, and then
that statement extends to all initial points by a simple union bound. This is
not the case in BLPP. To overcome this difficulty, we need new methods of
proof. Additionally, the continuum of points gives rise to new results
regarding non-uniqueness of semi-infinite geodesics, as seen in Item (3)
below.
Specific items proved in this paper include the following, and are recorded as
Theorems 3.1 and 3.5.
1. (1)
The Busemann functions of BLPP from [ARAS20] are extended to a global process
on a single event of probability one, for all the uncountably many initial
points and directions.
2. (2)
Once the Busemann process is in place, we use it to construct semi-infinite
geodesics for BLPP. With probability one, there exists a family of semi-
infinite geodesics, starting from each initial point and in each asymptotic
direction.
3. (3)
With probability one, all semi-infinite geodesics, whether constructed by the
Busemann functions or not, have an asymptotic direction. For a fixed initial
point and direction, there is almost surely a unique semi-infinite geodesic
starting from the given point and traveling asymptotically in the given
direction. We also show that for a fixed direction, there is a countably
infinite set of initial points whose geodesic in that direction is not unique.
This non-uniqueness into a fixed direction is a new phenomenon that is not
present in exponential last-passage percolation on the lattice.
4. (4)
For each fixed direction, we prove that all semi-infinite geodesics, traveling
in that common direction, coalesce.
5. (5)
For fixed northeast and southwest directions, we prove the almost sure
nonexistence of bi-infinite geodesics in those directions.
To construct an infinite up-right path on the lattice $\mathbb{Z}^{2}$, one
chooses at each step whether to move upward or to the right. In the BLPP
setting, one chooses a real-valued increment to the right and then takes a
unit-size upward step. Section 4.1 explains informally this construction. The
locations of the upward steps of a Busemann geodesic are determined by a
variational problem for Brownian motion with drift. This formulation is
significant in at least two ways. First, this step is where non-uniqueness of
geodesics can arise. Understanding it involves properties of Brownian paths.
Furthermore, this variational construction can be potentially and profitably
adapted to other models. This includes both continuum models of the KPZ class,
such as the directed landscape, and lattice LPP. We address the latter in
Section 5.2.
The proof of the coalescence in Item (4) above requires technical novelties.
The underlying idea from [Sep20] is to construct a dual environment from the
original environment and a Busemann function. If two geodesics in a given
direction do not coalesce, there exists a bi-infinite geodesic in the dual
environment with given northwest and southeast directions. Then, it is proven
that there are almost surely no bi-infinite geodesics in fixed directions.
A key point is a non-intersection property between the original semi-infinite
geodesics and southwest-travelling semi-infinite geodesics in the dual
environment. This is in Theorem 7.12, which is the analogue of Lemma 4.4 in
[Sep20]. The dual environment in BLPP is constructed through dual queuing
mappings, first presented in [OY01] and further studied in [SV10, ARAS20]. See
Equation (3.9) for the precise definition. While the general approach is not
new, our contribution comes first from Theorem D.1 that shows that joint
queuing mappings can be inverted by reverse-time queuing mappings. Then, we
use this theorem and variational formulas to construct southwest-travelling
semi-infinite geodesics in the dual environment. Lastly, the non-intersection
of Theorem 7.12 is proved by comparing the maximizers of the variational
formulas.
### 1.5. Future work
The present paper lays the foundation for future work on Busemann functions
and semi-infinite geodesics, which most immediately will cover the following.
1. (1)
To study the full family of semi-infinite geodesics in BLPP, we derive the
joint distribution of Busemann functions across all directions. The analogous
result was achieved in [FS20] for the exponential CGM, which led to important
advances in the global structure of geodesic trees [JRS21], coalescence
[SS20], convergence [BBS21, BF18], and bi-infinite geodesics [BBS20]. We
likewise study the geometry of geodesics in BLPP and give a full
characterization of exceptional directions where geodesics do not coalesce, in
terms of the Busemann process.
2. (2)
We define the competition interface for each initial point of BLPP. This has a
straightforward definition in discrete models, but requires more care in the
semi-discrete setting. Specifically, we show that, on each level $m$ of the
semi-discrete space $\mathbb{Z}\times\mathbb{R}$, outside a random set of
Hausdorff dimension $\frac{1}{2}$, all points $(m,t)$ have a degenerate
competition interface. For each of the points in this set of Hausdorff
dimension $\frac{1}{2}$, there exists a random direction $\theta^{\star}$ such
that there are two semi-infinite geodesics starting at $(m,t)$ in direction
$\theta^{\star}$ that only share the common initial point.
As can be seen from Item (2), there are many rich questions to explore related
to non-uniqueness of semi-infinite geodesics in Brownian last-passage
percolation. Other questions related to Hausdorff dimension have recently been
explored in the KPZ fixed point and the directed landscape in [BGH21, BGH19,
GH21, CHHM21]. Our longer-term goal is to extend the approach of this paper to
the continuum models of the KPZ class.
### 1.6. Organization of the paper
Section 2 defines the Brownian last-passage percolation model and related
objects. We also state results from other papers, including the existence of
Busemann functions from [ARAS20]. Section 3 contains the main results on the
existence and properties of the global Busemann process and semi-infinite
geodesics. Section 4 states more technical theorems related to the specific
construction of semi-infinite geodesics from the Busemann functions. In
Section 5, we discuss connections to infinite polymer measures in the
O’Connell-Yor polymer, semi-infinite geodesics in discrete last-passage
percolation, and to queuing theory. The proofs of this paper are contained in
Sections 6 and 7. Section 6 constructs the global Busemann process and derives
its properties. All results for semi-infinite geodesics are proved in Section
7. The Appendices collect some standard material, background from the
literature, and some technical theorems.
## 2\. Definitions and previous results
### 2.1. Preliminaries
The following notation and conventions are used throughout the paper.
1. (i)
For a function $f:\mathbb{R}\rightarrow\mathbb{R}$ , we write
$f(s,t)=f(t)-f(s)$ and $\widetilde{f}(t)=-f(-t).$
2. (ii)
$\mathbb{Z}$, $\mathbb{Q}$ and $\mathbb{R}$ are restricted by subscripts, as
in for example $\mathbb{Z}_{>0}=\\{1,2,3,\dotsc\\}$.
3. (iii)
Whenever $m\leq n\in\mathbb{Z}$ and $s\leq t\in\mathbb{R}$, we say that
$(m,s)\leq(n,t)$.
4. (iv)
Let $X\sim\mathcal{N}(\mu,\sigma^{2})$ indicate that the random variable $X$
has normal distribution with mean $\mu$ and variance $\sigma^{2}$. For
$\alpha>0$, let $X\sim\operatorname{Exp}(\alpha)$ indicate that $X$ has
exponential distribution with rate $\alpha$, or equivalently, mean
$\alpha^{-1}$.
5. (v)
Equality in distribution between random variables and processes is denoted by
$\overset{d}{=}$.
6. (vi)
A two-sided Brownian motion is a continuous random process
$\\{B(t):t\in\mathbb{R}\\}$ such that $B(0)=0$ almost surely and such that
$\\{B(t):t\geq 0\\}$ and $\\{B(-t):t\geq 0\\}$ are two independent standard
Brownian motions on $[0,\infty)$.
7. (vii)
For $\lambda\in\mathbb{R}$, $\\{Y(t):t\in\mathbb{R}\\}$ is a two-sided
Brownian motion with drift $\lambda$ if the process $\\{Y(t)-\lambda
t:t\in\mathbb{R}\\}$ is a two-sided Brownian motion.
8. (viii)
The square ${\scaleobj{0.9}{\boxempty}}$ as a superscript represents a sign
$+$ or $-$.
### 2.2. Geodesics in Brownian last-passage percolation
The Brownian last-passage process is defined as follows. On a probability
space $(\Omega,\mathcal{F},\mathbb{P})$, let
$\mathbf{B}=\\{B_{r}\\}_{r\in\mathbb{Z}}$ be a field of independent, two-sided
Brownian motions. For $(m,s)\leq(n,t)$, define the set
$\Pi_{(m,s),(n,t)}:=\\{\mathbf{s}_{m,n}=(s_{m-1},s_{m},\ldots,s_{n})\in\mathbb{R}^{n-m+2}:s=s_{m-1}\leq
s_{m}\leq\cdots\leq s_{n}=t\\}.$
Denote the energy of a sequence $\mathbf{s}_{m,n}\in\Pi_{(m,s),(n,t)}$ by
(2.1) $\mathcal{E}(\mathbf{s}_{m,n})=\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r}).$
Now, for $\mathbf{x}=(m,s)\leq(n,t)=\mathbf{y}$, define the Brownian last-
passage time as
(2.2)
$L_{\mathbf{x},\mathbf{y}}(\mathbf{B})=\sup\\{\mathcal{E}(\mathbf{s}_{m,n}):\mathbf{s}_{m,n}\in\Pi_{\mathbf{x},\mathbf{y}}\\}.$
Whenever the specific field of Brownian motions used is either clear from or
not important in the context, we write $L_{(m,s),(n,t)}$.
$s$$s_{0}$$s_{1}$$s_{2}$$s_{3}$$t$$0$$1$$2$$3$$4$ Figure 2.1. The Brownian
increments $B_{r}(s_{r-1},s_{r})$ for $r=0,\dotsc,4$ in (2.1) that make up the
energy of the path depicted in Figure 2.2.
The elements of $\Pi_{\mathbf{x},\mathbf{y}}$ are in bijection with paths in
$\mathbb{R}^{2}$ between $\mathbf{x}$ and $\mathbf{y}$, that move to the right
in real-valued increments, and move upwards in integer increments. For a given
$\mathbf{s}_{m,n}\in\Pi_{(m,s),(n,t)}$, the path consists of the following
points:
(2.3)
$\bigcup_{r=m}^{n}\bigl{\\{}(r,u):u\in[s_{r-1},s_{r}]\bigr{\\}}\cup\bigcup_{r=m}^{n-1}\bigl{\\{}(v,s_{r}):v\in[r,r+1]\bigr{\\}}.$
This set consists of horizontal and vertical line segments, such that the
vertical segments occur at the points $s_{r}$. Because of this bijection, we
sometimes say $\Gamma\in\Pi_{\mathbf{x},\mathbf{y}}$ for such an up-right
path. For $(m,t)\in\mathbb{Z}\times\mathbb{R}$, we graphically represent the
$t$-coordinate as the horizontal coordinate and the $m$-coordinate as the
vertical coordinate in the plane. Since $\Pi_{\mathbf{x},\mathbf{y}}$ is a
compact set and Brownian motion is continuous, on a single event of
probability one, for all
$(m,t)=\mathbf{x}\leq\mathbf{y}=(n,t)\in\mathbb{Z}\times\mathbb{R}$, there
exists a sequence $\mathbf{s}_{m,n}\in\Pi_{\mathbf{x},\mathbf{y}}$ such that
$\mathcal{E}(\mathbf{s}_{m,n})=L_{\mathbf{x},\mathbf{y}}$. The associated path
is called a geodesic between the points.
$(0,s)$$(4,t)$$s_{0}$$s_{1}$$s_{2}$$s_{3}$$0$$1$$2$$3$$4$ Figure 2.2. Example
of a planar path from $(0,s)$ to $(4,t)$, represented by the sequence
$(s=s_{-1},s_{0},s_{1},s_{2},s_{3},s_{4}=t)\in\Pi_{(0,s),(4,t)}$.
To an infinite sequence, $s=s_{m-1}\leq s_{m}\leq s_{m+1}\leq\cdots$ we
similarly associate a semi-infinite path. It is possible that $s_{r}=\infty$
for some $r\geq m$, in which case the last segment of the path is the ray
$[s_{r-1},\infty)\times\\{r\\}$, where $r$ is the first index with
$s_{r}=\infty$. The infinite path has direction $\theta\in[0,\infty]$ or is
$\theta$-directed if
$\lim_{n\rightarrow\infty}\frac{s_{n}}{n}\qquad\text{exists and equals
}\theta.$
We call an up-right semi-infinite path a semi-infinite geodesic if, for any
two points $\mathbf{x}\leq\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$ that lie
along the path, the portion of the path between the two points is a geodesic
between the two points. Similarly, a bi-infinite geodesic is a bi-infinite
path that forms a geodesic between any two of its points. Two up-right, semi-
infinite paths coalesce if there exists
$\mathbf{z}\in\mathbb{Z}\times\mathbb{R}$ such that the two paths agree above
and to the right of $\mathbf{z}$, as shown in Figure 2.3. Alternatively, if
the paths are defined by sequences of jump times $s^{1}=s_{m_{1}-1}^{1}\leq
s_{m_{1}}^{1}\cdots$ and $s^{2}=s_{m_{2}-1}^{2}\leq s_{m_{2}}^{2}\leq\cdots$,
then the two paths coalesce if and only if there exists $N\in\mathbb{Z}$ such
that $s_{r}^{1}=s_{r}^{2}$ for all $r\geq N$.
$\mathbf{x}$$\mathbf{y}$$\mathbf{z}$ Figure 2.3. Coalescence of semi-infinite
paths
The following lemma, due to Hammond [Ham19b], establishes uniqueness of
geodesics for a fixed initial and terminal point.
###### Lemma 2.1 ([Ham19b], Theorem B.1).
Fix endpoints $\mathbf{x}\leq\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$. Then,
there is almost surely a unique path whose energy achieves
$L_{\mathbf{x},\mathbf{y}}(\mathbf{B})$.
However, it is also true that for each fixed initial point
$\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$, with probability one, there exist
points $\mathbf{y}\geq\mathbf{x}$, such that the geodesic between $\mathbf{x}$
and $\mathbf{y}$ is not unique. Hence, the following lemma is important for
our understanding. It is a deterministic statement which holds for last-
passage percolation across any field of continuous functions.
###### Lemma 2.2 ([DOV18], Lemma 3.5).
Between any two points $(m,s)\leq(n,t)\in\mathbb{Z}\times\mathbb{R}$, there is
a rightmost and a leftmost Brownian last-passage geodesic. That is, there
exist
$\mathbf{s}_{m,n}^{L},\mathbf{s}_{m,n}^{R}\in\mathbf{\Pi}_{(m,s),(n,t)}$, that
are maximal for $\mathcal{E}(\mathbf{s}_{m,n})$, such that, for any other
maximal sequence $\mathbf{s}_{m,n}$, $s_{r}^{L}\leq s_{r}\leq s_{r}^{R}$ holds
for $m\leq r\leq n$.
### 2.3. Busemann functions
In the present paper, semi-infinite geodesics are constructed from Busemann
functions. Busemann functions are defined to be the asymptotic difference of
last-passage times from two different starting points to a common terminal
point that is traveling to $\infty$ in a given direction. See Figure 2.4. The
direction is indexed by a parameter $\theta>0$.
$\mathbf{y}$$\mathbf{x}$$(n,n\theta)$$n$$n\theta$ Figure 2.4. Two geodesics
with a common terminal point $(n,n\theta)$.
Existence of these Busemann functions was proven in [ARAS20], both in the
positive temperature and zero temperature cases. We state the zero-temperature
result:
###### Theorem 2.3 ([ARAS20], Theorem 4.2).
Fix $\theta>0$ and $\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$. Then,
there exists a random variable $\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})$
and an event $\Omega_{\mathbf{x},\mathbf{y}}^{(\theta)}$ of probability one,
on which
(2.4)
$\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})=\lim_{n\rightarrow\infty}\bigl{[}L_{\mathbf{x},(n,t_{n})}-L_{\mathbf{y},(n,t_{n})}\bigr{]}$
holds for any sequence $\\{t_{n}\\}\subseteq\mathbb{R}$ satisfying
${t_{n}}/{n}\to\theta$. Further, if
(2.5) $\displaystyle
v_{m}^{\theta}(t):=\mathcal{B}^{\theta}((m-1,t),(m,t)),\text{ and}$ (2.6)
$\displaystyle h_{m}^{\theta}(t):=\mathcal{B}^{\theta}((m,0),(m,t)),$
then
$v_{m}^{\theta}(t)\sim\operatorname{Exp}\bigl{(}\frac{1}{\sqrt{\theta}}\bigr{)}$
and
$h_{m}^{\theta}(s,t)\sim\mathcal{N}\bigl{(}\frac{t-s}{\sqrt{\theta}},|t-s|\bigr{)}$
for all $s,t\in\mathbb{R}$ and $m\in\mathbb{Z}$.
## 3\. Main results
### 3.1. Semi-infinite geodesics
The following result summarizes the contributions of this paper related to
semi-infinite geodesics in BLPP.
###### Theorem 3.1.
The following hold
1. (i)
With probability one, for every initial point
$\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$ and every direction $\theta>0$,
there exists a $\theta-$directed semi-infinite geodesic starting from
$\mathbf{x}$.
2. (ii)
(Uniqueness for fixed initial points and directions) For each fixed
$\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$ and $\theta>0$, there exists an
event, $\Omega_{\mathbf{x}}^{(\theta)}$, of probability one, on which there is
exactly one $\theta$-directed semi-infinite geodesic starting from
$\mathbf{x}$.
3. (iii)
(Non-uniqueness for fixed direction and random initial points) For each
$\theta>0$, there exists an event $\widetilde{\Omega}^{(\theta)}$, of
probability one, on which the set
$\\{\mathbf{x}\in\mathbb{Z}\times\mathbb{R}:\text{the }\theta\text{-directed
semi-infinite geodesic starting from }\mathbf{x}\text{ is not unique}\\}$
is countably infinite. For every $(m,t)$ in this set, at most one of the
$\theta$-directed semi-infinite geodesics passes through the point
$(m,t+\varepsilon)$ for some $\varepsilon>0$. All the others pass through
$(m+1,t)$.
4. (iv)
With probability one, every semi-infinite geodesic is $\theta$-directed for
some $\theta\in[0,\infty]$. That is, for any infinite sequence $t=t_{m-1}\leq
t_{m}\leq t_{m+1}\leq\cdots$ defining a semi-infinite geodesic starting from
some point $(m,t)\in\mathbb{Z}\times\mathbb{R}$, the limit
(3.1) $\lim_{n\rightarrow\infty}\frac{t_{n}}{n}\ \text{ exists in
}[0,\infty].$
5. (v)
With probability one, if for any $(m,t)\in\mathbb{Z}\times\mathbb{R}$ and any
such sequence, the limit (3.1) equals $0$, then $t_{r}=t$ for all $r\geq m$.
Similarly, if this limit is $\infty$, then $t_{m}=t_{m+1}=\cdots=\infty$. That
is, the only semi-infinite geodesics that are asymptotically vertical or
horizontal are trivial (i.e. straight lines).
6. (vi)
(Non-existence of bi-infinite geodesics for fixed directions) Fix
$\theta,\eta>0$. Then, there exists an event, $\Omega^{(\theta,\eta)}$, of
probability one, on which there are no bi-infinite geodesics defined by jump
times $\cdots\leq\tau_{-1}\leq\tau_{0}\leq\tau_{1}\leq\cdots$ such that
(3.2)
$\lim_{n\rightarrow\infty}\frac{\tau_{n}}{n}=\theta\qquad\text{and}\qquad\lim_{n\rightarrow\infty}\frac{\tau_{-n}}{n}=-\eta.$
7. (vii)
(Coalescence of geodesics in a fixed direction) For each $\theta>0$, there
exists an event $\widehat{\Omega}^{(\theta)}$, of probability one, on which
all $\theta$-directed semi-infinite geodesics coalesce.
###### Remark 3.2.
As discussed in the introduction, the non-uniqueness stated in Part (iii) is a
new phenomenon that arises from the semi-discrete nature of the model. We
refer the reader to Section 4.3 for further discussion on non-uniqueness.
Theorem 3.1(vi) is proven by first solving the “midpoint problem.” This
problem first appeared in the context of first-passage percolation in a paper
of Benjamini, Kalai, and Schramm [BKS03]. In that context the problem asks
whether
$\lim_{n\rightarrow\infty}\mathbb{P}\Bigl{(}\Bigl{\lfloor}\frac{n}{2}\Bigr{\rfloor}e_{1}\text{
lies on some geodesic between }0\text{ and }ne_{1}\Bigr{)}=0.$
In 2016, Damron and Hanson [DH17] proved that this is true, under the
assumption that the boundary of the limit shape is differentiable. Later
Ahlberg and Hoffman [AH16] proved this result without the assumption of
differentiability. The following formulation of the midpoint problem more
closely matches that for exponential last-passage percolation in the arXiv
version of [Sep20], Theorem 4.12.
###### Lemma 3.3 (Midpoint problem).
Let $\theta,\eta>0$ and $(m,t)\in\mathbb{Z}\times\mathbb{R}$. Then, the
following subset of $\Omega$ is contained in an event of probability zero:
$\displaystyle\Big{\\{}\text{there exists a sequence
}\\{t_{n}\\}_{n\in\mathbb{Z}}\text{ satisfying
}\lim_{n\rightarrow\infty}\frac{t_{n}}{n}=\theta\text{ and
}\lim_{n\rightarrow\infty}\frac{t_{-n}}{-n}=\eta\text{ and }$
$\displaystyle\qquad\text{such that, for each }n\in\mathbb{Z}_{>0},\text{ some
geodesic between }(-n,t_{-n})\text{ and }(n,t_{n})\text{ passes through
}(m,t)\Big{\\}}$
### 3.2. Existence and properties of the Busemann process
To prove Theorem 3.1, we extend the individual Busemann functions of Theorem
2.3 to a global Busemann process. The following transformations are used to
understand the structure of this process. For functions
$Z,B:\mathbb{R}\rightarrow\mathbb{R}$ satisfying $Z(0)=B(0)=0$ and
$\limsup_{s\rightarrow\infty}(B(s)-Z(s))=-\infty$, define
(3.3) $\displaystyle Q(Z,B)(t)$ $\displaystyle=\sup_{t\leq
s<\infty}\\{B(t,s)-Z(t,s)\\},$ (3.4) $\displaystyle D(Z,B)(t)$
$\displaystyle=Z(t)+Q(Z,B)(0)-Q(Z,B)(t),$ (3.5) $\displaystyle R(Z,B)(t)$
$\displaystyle=B(t)+Q(Z,B)(t)-Q(Z,B)(0).$
Reverse-time analogues of these transformations are defined for continuous
functions $Y,C:\mathbb{R}\rightarrow\mathbb{R}$ satisfying $Y(0)=C(0)=0$ and
$\limsup_{s\rightarrow-\infty}(Y(s)-C(s))=-\infty$:
(3.6) $\displaystyle\overleftarrow{Q}(Y,C)(t)$
$\displaystyle=\sup_{-\infty<s\leq t}\\{C(s,t)-Y(s,t)\\},$ (3.7)
$\displaystyle\overleftarrow{D}(Y,C)(t)$
$\displaystyle=Y(t)+\overleftarrow{Q}(Y,C)(t)-\overleftarrow{Q}(Y,C)(0),$
(3.8) $\displaystyle\overleftarrow{R}(Y,C)(t)$
$\displaystyle=C(t)+\overleftarrow{Q}(Y,C)(0)-\overleftarrow{Q}(Y,C)(t).$
These transformations originate from the Brownian queue, first studied by
Glynn and Whitt [GW91], and further expounded on by Harrison and Williams
[Har85, HW90, HW92] and O’Connell and Yor [OY01]. See Section 5.3 and Appendix
C for more about the queuing interpretation. The following lemma is a
straightforward exercise. We state it for completeness, as we will refer to it
later in the paper.
###### Lemma 3.4.
Let $B,Z:\mathbb{R}\rightarrow\mathbb{R}$ be continuous functions satisfying
$\limsup_{s\rightarrow\infty}(B(s)-Z(s))=-\infty$. Then, $Q(Z,B)$, $D(Z,B)$,
and $R(Z,B)$ are continuous. Similarly, if
$Y,C:\mathbb{R}\rightarrow\mathbb{R}$ are continuous functions satisfying
$\limsup_{s\rightarrow-\infty}(Y(s)-C(s))=-\infty$, then
$\overleftarrow{Q}(Y,C),\overleftarrow{D}(Y,C)$, and $\overleftarrow{R}(Y,C)$
are continuous.
The next theorem summarizes the existence and properties of the Busemann
process. This process has discontinuities in the direction parameter $\theta$.
Instead of a single cadlag process, it is useful to retain both a left- and a
right-continuous version indicated by $\theta-$ and $\theta+$ because this
distinction captures spatial limits.
###### Theorem 3.5.
There exists a process,
$\\{\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{x},\mathbf{y}):\theta>0,\,{\scaleobj{0.9}{\boxempty}}\in\\{+,-\\},\,\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}\\},$
and for $\theta>0$, there exist events
$\Omega^{(\theta)}\subseteq\Omega_{1}\subseteq\Omega$, each of probability
one, such that the following hold. Here,
$v_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}},h_{m}^{\theta{\scaleobj{0.8}{\boxempty}}}$
are defined as in (2.5) and (2.6), placing ${\scaleobj{0.9}{\boxempty}}$ in
the appropriate superscripts.
1. (i)
(Additivity) On $\Omega_{1}$, whenever
$\mathbf{x},\mathbf{y},\mathbf{z}\in(\mathbb{Z}\times\mathbb{R}),\theta>0$,
and ${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{x},\mathbf{y})+\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{y},\mathbf{z})=\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{x},\mathbf{z}).$
2. (ii)
(Monotonicity) On $\Omega_{1}$, whenever $0<\gamma<\theta<\infty$,
$m\in\mathbb{Z}$, and $s<t\in\mathbb{R}$,
$0\leq v_{m}^{\gamma-}(s)\leq v_{m}^{\gamma+}(s)\leq v_{m}^{\theta-}(s)\leq
v_{m}^{\theta+}(s),\text{ and }$ $B_{m}(s,t)\leq h_{m}^{\theta+}(s,t)\leq
h_{m}^{\theta-}(s,t)\leq h_{m}^{\gamma+}(s,t)\leq h_{m}^{\gamma-}(s,t).$
3. (iii)
(Convergence) On $\Omega_{1}$, for every $m\in\mathbb{Z}$, $\theta>0$ and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
1. (a)
As $\gamma\nearrow\theta$, $h_{m}^{\gamma{\scaleobj{0.8}{\boxempty}}}$ and
$v_{m}^{\gamma{\scaleobj{0.8}{\boxempty}}}$ converge uniformly, on compact
subsets of $\mathbb{R}$, to $h_{m}^{\theta-}$ and $v_{m}^{\theta-}$,
respectively.
2. (b)
As $\delta\searrow\theta$, $h_{m}^{\delta{\scaleobj{0.8}{\boxempty}}}$ and
$v_{m}^{\delta{\scaleobj{0.8}{\boxempty}}}$ converge uniformly, on compact
subsets of $\mathbb{R}$, to $h_{m}^{\theta+}$ and $v_{m}^{\theta+}$,
respectively.
3. (c)
As $\gamma\rightarrow\infty$, $h_{m}^{\gamma{\scaleobj{0.8}{\boxempty}}}$
converges uniformly, on compact subsets of $\mathbb{R}$, to $B_{m}$.
4. (d)
As $\delta\searrow 0$, $v_{m}^{\delta{\scaleobj{0.8}{\boxempty}}}$ converges
uniformly, on compact subsets of $\mathbb{R}$, to $0$.
4. (iv)
(Continuity) On $\Omega_{1}$, for any $r,m\in\mathbb{Z}$, $\theta>0$, and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$(s,t)\mapsto\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}((m,s),(r,t))$ is
a continuous function $\mathbb{R}^{2}\rightarrow\mathbb{R}$.
5. (v)
(Limits) On $\Omega_{1}$, for each $\theta>0$ and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$\lim_{s\rightarrow\pm\infty}\bigl{[}B_{m}(s)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\bigr{]}=\mp\infty.$
6. (vi)
(Queuing relationships between Busemann functions) On $\Omega_{1}$, for all
$m\in\mathbb{Z},\theta>0$, and signs
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$v_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}=Q(h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}},B_{m})\qquad\text{and}\qquad
h_{m}^{\theta{\scaleobj{0.8}{\boxempty}}}=D(h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}},B_{m}).$
7. (vii)
(Independence) For any $m\in\mathbb{Z}$,
$\\{h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}:\theta>0,{\scaleobj{0.9}{\boxempty}}\in\\{+,-\\},r>m\\}\text{
is independent of }\\{B_{r}:r\leq m\\}$
8. (viii)
(Equality for fixed directions) Fix $\theta>0$. Then, on the event
$\Omega^{(\theta)}$, for all
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$ and all sequences
$\\{t_{n}\\}$ with $t_{n}/n\rightarrow\theta$,
$\mathcal{B}^{\theta-}(\mathbf{x},\mathbf{y})=\lim_{n\rightarrow\infty}\bigl{[}L_{\mathbf{x},(n,t_{n})}-L_{\mathbf{y},(n,t_{n})}\bigr{]}=\mathcal{B}^{\theta+}(\mathbf{x},\mathbf{y}).$
9. (ix)
(Shift invariance) For each $\mathbf{z}\in\mathbb{Z}\times\mathbb{R}$,
$\\{\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{x},\mathbf{y}):\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R},\theta>0,{\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}\\}\overset{d}{=}\\{\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{x}+\mathbf{z},\mathbf{y}+\mathbf{z}):\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R},\theta>0,{\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}\\}.$
###### Remark 3.6.
On account of Part (viii), when working on the event $\Omega^{(\theta)}$ we
write $\mathcal{B}^{\theta}=\mathcal{B}^{\theta-}=\mathcal{B}^{\theta+}$. The
fact that the limits exist for all initial points
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$ and fixed $\theta>0$ on a
single event of probability one gives a generalization of Theorem 2.3. As will
be seen in Section 6, the key is Lemma C.5, which generalizes a proof in
[ARAS20].
We introduce a dual field of Brownian motions used later in the proof of
coalescence of semi-infinite geodesics. Let $\theta>0$. Recall the definition
(3.5) of the mapping $R$, and on the event $\Omega^{(\theta)}$, set
(3.9) $X_{m}^{\theta}=R(h_{m}^{\theta},B_{m-1}).$
Denote the field of these random functions by
$\mathbf{X}^{\theta}:=\\{X_{m}^{\theta}\\}_{m\in\mathbb{Z}}$. The following
theorem describes the distribution of the Busemann functions for a fixed
direction $\theta$.
###### Theorem 3.7.
Fix $\theta>0$.
1. (i)
The process $t\mapsto h_{m}^{\theta}(t)$ is a two-sided Brownian motion with
drift $\frac{1}{\sqrt{\theta}}$.
2. (ii)
$\mathbf{X}^{\theta}$ is a field of independent two-sided Brownian motions.
For each $n\in\mathbb{Z}$, $\\{X_{m}^{\theta}\\}_{m>n}$ is independent of
$\\{h_{m}^{\theta}\\}_{m\leq n}$.
3. (iii)
The process $t\mapsto v_{m}^{\theta}(t)$ is a stationary and reversible strong
Markov process. For each $t\in\mathbb{R}$,
$v_{m}^{\theta}(t)\sim\operatorname{Exp}\bigl{(}\frac{1}{\sqrt{\theta}}\bigr{)}$.
4. (iv)
(Burke property) Fix $m<n\in\mathbb{Z}$ and $-\infty<t_{n}\leq
t_{n-1}\leq\cdots\leq t_{m+1}<\infty$. Then, the following random variables
and processes are mutually independent:
$\\{B_{m}(u,t_{m+1}):u\leq
t_{m+1}\\},\;\;v_{m+1}^{\theta}(t_{m+1}),\;\;\\{h_{m}^{\theta}(t_{m+1},u):u\geq
t_{m+1}\\},\;\\\\[10.00002pt] \\{B_{r}(u,t_{r+1}):u\leq
t_{r+1}\\},\;v_{r+1}^{\theta}(t_{r+1}),\;\\{h_{r}^{\theta}(t_{r+1},u):t_{r+1}\leq
u\leq t_{r}\\},\\\\[10.00002pt] \text{ and }\\{X_{r}^{\theta}(t_{r},u):u\geq
t_{r}\\},\;\text{ for }\;m+1\leq r\leq
n-1,\;\;\\{h_{n}^{\theta}(u,t_{n}):u\leq t_{n}\\},\;\text{ and
}\;\\{X_{n}^{\theta}(t_{n},u):u\geq t_{n}\\}.$
As a special case, $\\{v_{m}^{\theta}(t)\\}_{m\in\mathbb{Z}}$ is an i.i.d
sequence for each $t\in\mathbb{R}$. Refer to Figure 3.1 for clarity.
###### Remark 3.8.
Using the representation of Equation (3.12) below, many formulas for the
process $t\mapsto v_{m}^{\theta}(t)$ are well known. See [BS02], page 129 and
[SN01], specifically equations (4) and (5), for more on this process,
including the transition density. However, we caution that when comparing
formulas, in the setting of [BS02, SN01], the process is
$t\mapsto\sup_{-\infty<s\leq t}\\{B(s,t)-\mu(t-s)\\},$
where $B$ is a two-sided Brownian motion and $\mu>0$. In our setting, there is
a factor of $\sqrt{2}$ multiplied to $B$.
$m$$m+1$$m+2$$n-3$$n-2$$n-1$$n$$B_{m}$$v_{m+1}^{\theta}$$B_{m+1}$$v^{\theta}_{m+2}$$h_{m+1}^{\theta}$$X_{m+1}^{\theta}$$h_{m}^{\theta}$$B_{m+2}$$h_{m+2}^{\theta}$$v^{\theta}_{m+3}$$X^{\theta}_{m+2}$$X^{\theta}_{m+3}$$v_{n-2}^{\theta}$$B_{n-3}$$\ddots$$v^{\theta}_{n-1}$$h^{\theta}_{n-2}$$B_{n-2}$$X^{\theta}_{n-2}$$v_{n}^{\theta}$$B_{n-1}$$h^{\theta}_{n-1}$$X^{\theta}_{n-1}$$h_{n}^{\theta}$$X_{n}^{\theta}$$t_{n}$$t_{n-1}$$t_{n-2}$$\cdots$$t_{m+3}$$t_{m+2}$$t_{m+1}$
Figure 3.1. Independence structure for Busemann functions. Each process
$h^{\theta}_{r}$ is associated to the (purple/dotted) segment on level $r$,
processes $B_{r}$ and $X^{\theta}_{r}$ cover the remaining portions of
horizontal level $r$, and the process $v_{r}^{\theta}$ is associated to the
(red/vertical) edge from level $r-1$ to $r$ at time point $t_{r}$.
The construction of the Busemann process and the proof of Theorem 3.5 can be
found in Section 6. We prove Theorem 3.7 here, assuming Theorem 3.5 and with
the help of the results of the appendix.
###### Proof of Theorem 3.7.
Part (i): As will be seen from the construction in Section 6, $t\mapsto
h_{m}^{\theta}(t)$ has the proper finite-dimensional distributions and is
continuous by Theorem 3.5(iv).
Part (ii): Fix integers $p<n<m$. By Theorem 3.5(vii),
$(B_{p},\ldots,B_{n},\dots,B_{m-2},B_{m-1},h_{m}^{\theta})\quad\text{are
independent. }$
Theorem 3.5(vi) and definition (3.9) give
$h_{m-1}^{\theta}=D(h_{m}^{\theta},B_{m-1})$ and
$X_{m}^{\theta}=R(h_{m}^{\theta},B_{m-1})$, so by Theorem C.2,
$h_{m-1}^{\theta}$ and $X_{m}^{\theta}$ are independent and $X_{m}^{\theta}$
is a two-sided Brownian motion. In particular, now
$(B_{p},\ldots,B_{n},\dots,B_{m-2},h_{m-1}^{\theta},X_{m}^{\theta})\quad\text{are
independent. }$
Continue inductively by applying the transformation $(D,R)$ to successive
pairs $(B_{j-1},h_{j}^{\theta})$ for $j=m-1,m-2,\dotsc,n+1$ after which
$(B_{p},\ldots,B_{n-1},h_{n}^{\theta},X_{n+1}^{\theta},\dotsc,X_{m}^{\theta})\quad\text{are
independent. }$
To conclude, note that $(h_{p}^{\theta},\ldots,h_{n}^{\theta})$ is a function
of $(B_{p},\ldots,B_{n-1},h_{n}^{\theta})$ through iteration of
$h_{k}^{\theta}=D(h_{k+1}^{\theta},B_{k})$.
Part (iii): By Theorem 3.5(vi) and (3.9), on the event $\Omega^{(\theta)}$ we
have these relations $\forall m\in\mathbb{Z}$:
(3.10) $h_{m-1}^{\theta}=D(h_{m}^{\theta},B_{m-1}),\qquad
X_{m}^{\theta}=R(h_{m}^{\theta},B_{m-1}),\qquad\text{and}\qquad
v_{m}^{\theta}=Q(h_{m}^{\theta},B_{m-1}).$
The fact that $v_{m}^{\theta}(t)$ is exponential with rate
$\frac{1}{\sqrt{\theta}}$ then follows from Lemma B.1. Theorem D.1 allows us
to reverse these mappings, so $\forall m\in\mathbb{Z}$:
(3.11) $\displaystyle
h_{m}^{\theta}=\overleftarrow{D}(h_{m-1}^{\theta},X_{m}^{\theta}),\qquad
B_{m-1}=\overleftarrow{R}(h_{m-1}^{\theta},X_{m}^{\theta}),\qquad\text{and}\qquad
v_{m}^{\theta}=\overleftarrow{Q}(h_{m-1}^{\theta},X_{m}^{\theta}).$
Then, for $t\in\mathbb{R}$,
(3.12) $v_{m}^{\theta}(t)=\sup_{-\infty<u\leq
t}\\{X_{m}^{\theta}(u,t)-h_{m-1}^{\theta}(u,t)\\}.$
By Parts (i) and (ii), $t\mapsto X_{m}^{\theta}(t)-h_{m-1}^{\theta}(t)$ is
equal in distribution to a two-sided Brownian motion with negative drift,
multiplied by a factor of $\sqrt{2}$. Represented this way, $t\mapsto
v_{m}^{\theta}(t)$ is known as a stationary, reflected Brownian motion with
drift. Stationarity follows from the stationarity of increments. The fact that
$X$ is a reversible strong Markov process is proven in [Har85], pg. 81 (see
also pg. 49-50 in [Har85] and Equations (4) and (5) in [SN01] for a more
directly applicable statement).
Part (iv): By Part (ii) and (3.11), for any initial level $m$, the process
$\bigl{\\{}h_{r+m}^{\theta},v_{r+m+1}^{\theta},X_{r+m+1}^{\theta},B_{r+m}\bigr{\\}}_{r\geq
0}$ has the same distribution as
$\bigl{\\{}Y_{r}^{\frac{1}{\sqrt{\theta}}},q_{r+1}^{\frac{1}{\sqrt{\theta}}},B_{r+1},W_{r}^{\lambda}\bigr{\\}}_{r\geq
0}$ as defined in (C.1). Note that $B_{r+m}$ now plays the role of the
$W_{r}^{\lambda}$, as stated in the definition. Therefore, the independence
structure of Theorem C.3 holds. ∎
## 4\. Construction and properties of the semi-infinite geodesics
### 4.1. Heuristic for construction of semi-infinite geodesics
The next task is the construction of semi-infinite geodesics from each initial
point and in each asymptotic direction. For each given point
$(m,t)\in\mathbb{Z}\times\mathbb{R}$ and direction parameter $\theta>0$, we
want to find a semi-infinite geodesic, defined by jump times
$t=\tau_{m-1}\leq\tau_{m}\leq\cdots$ that satisfies
$\lim_{n\rightarrow\infty}\frac{\tau_{n}}{n}=\theta.$
We argue heuristically to motivate the useful construction. Start by finding a
maximal path for $L_{(m,t),(n,n\theta)}$ for a large value of $n$. Note that
$L_{(m,t),(n,n\theta)}=\max_{s\in[t,n\theta]}\bigl{(}B_{m}(t,s)+L_{(m+1,s),(n,n\theta)}\bigr{)},$
and the maximizer $s=\tau_{m}$ is the location where the geodesic jumps from
level $m$ to $m+1$. For all $t\leq s\leq n\theta$,
$B_{m}(t,s)+L_{(m+1,s),(n,n\theta)}\leq
B_{m}(t,\tau_{m})+L_{(m+1,\tau_{m}),(n,n\theta)}.$
Rearranging yields
$B_{m}(\tau_{m},s)\leq
L_{(m+1,\tau_{m}),(n,n\theta)}-L_{(m+1,s),(n,n\theta)}.$
As $n$ changes, so could $\tau_{m}$, but for the sake of heuristic we hold
$\tau_{m}$ constant. Take limits as $n\rightarrow\infty$ and rearrange again
to get, for some sign ${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$B_{m}(s)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\leq
B_{m}(\tau_{m})-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{m}).$
### 4.2. Busemann geodesics
The discussion of the previous section motivates this rigorous definition.
###### Definition 4.1.
On the event $\Omega_{1}$, for all $(m,t)\in\mathbb{Z}$, $\theta>0$ and
${\scaleobj{0.8}{\boxempty}}\in\\{+,-\\}$, let
$\mathbf{T}^{\theta{\scaleobj{0.8}{\boxempty}}}_{(m,t)}$ denote the set of
sequences
$t=\tau_{m-1}\leq\tau_{m}\leq\tau_{m+1}\leq\cdots$
that satisfy
(4.1)
$B_{r}(\tau_{r})-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})=\sup_{s\in[\tau_{r-1},\infty)}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}\qquad\text{for
each $r\geq m$.}$
Theorem 3.5(iv)–(v) imply that such sequences exist. At each level $r$, there
exist leftmost and rightmost maximizers. Let
$t=\tau_{(m,t),m-1}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\tau_{(m,t),m}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\tau_{(m,t),m+1}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\cdots\qquad\text{and}\qquad
t=\tau_{(m,t),m-1}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\tau_{(m,t),m}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\tau_{(m,t),m+1}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\cdots$
denote the leftmost and rightmost sequences in
$\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$. Furthermore, define
$\mathbf{T}_{(m,t)}^{\theta}:=\mathbf{T}_{(m,t)}^{\theta+}\cup\mathbf{T}_{(m,t)}^{\theta-}.$
###### Remark 4.2.
Since every non-decreasing sequence in $\mathbf{T}_{(m,t)}^{\theta}$ defines a
semi-infinite up-right path, $\mathbf{T}_{(m,t)}^{\theta}$ will be used to
denote the set of up-right paths constructed in this way. Theorem 4.3(i) below
shows that all sequences in $\mathbf{T}_{(m,t)}^{\theta}$ are semi-infinite
geodesics starting from $(m,t)$. By Theorem 3.5(viii), on the event
$\Omega^{(\theta)}\subseteq\Omega_{1}$,
$\mathbf{T}_{\mathbf{x}}^{\theta-}=\mathbf{T}_{\mathbf{x}}^{\theta+}=\mathbf{T}_{\mathbf{x}}^{\theta}$
for all $\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$. This does not imply that
$\mathbf{T}_{\mathbf{x}}^{\theta}$ contains only one element, so the leftmost
and rightmost distinction is still necessary in general for a fixed direction
$\theta$. See Theorem 4.7 below. However, by Theorem 3.1(ii), for fixed
$\theta>0$ and fixed $(m,t)\in\mathbb{Z}\times\mathbb{R}$,
$\mathbf{T}_{(m,t)}^{\theta}$ almost surely contains a single element. In this
case, Theorem B.3 gives the distribution of the first jump time $\tau_{m}$.
$(m,t)$$\tau_{m}^{\theta}$$\tau_{m+1}^{\theta}$$\tau_{m+2}^{\theta}$$\cdots$$m$$m+1$$m+2$$\vdots$
Figure 4.1. Example of an element of $\mathbf{T}_{(m,t)}^{\theta}$
Two elements $\\{\tau_{r}\\}_{r\geq m-1}$ and $\\{\tau^{\prime}_{r}\\}_{r\geq
m-1}$ of $\mathbf{T}_{\mathbf{x}}^{\theta{\scaleobj{0.8}{\boxempty}}}$ are
distinct if $\tau_{r}\neq\tau^{\prime}_{r}$ for at least one index $r$.
Uniqueness of the $\theta{\scaleobj{0.8}{\boxempty}}$ Busemann geodesic from
$\mathbf{x}$ means that
$\mathbf{T}_{\mathbf{x}}^{\theta{\scaleobj{0.8}{\boxempty}}}$ contains exactly
one sequence.
The following theorems collect the properties of the Busemann semi-infinite
geodesics, to be proved in Section 7.
###### Theorem 4.3.
There exists an event, $\Omega_{2}$, of full probability, on which the
following hold.
1. (i)
(Existence) For all $\mathbf{x}\in\mathbb{Z}\times\mathbb{R},\theta>0$, and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, every element of
$\mathbf{T}_{\mathbf{x}}^{\theta{\scaleobj{0.8}{\boxempty}}}$ defines a semi-
infinite geodesic starting from $\mathbf{x}$. More specifically, for any two
points $\mathbf{y}\leq\mathbf{z}$ along a path in
$\mathbf{T}_{\mathbf{x}}^{\theta}$, the energy of this path between
$\mathbf{y}$ and $\mathbf{z}$ is
$\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}(\mathbf{y},\mathbf{z})$, and
this energy is maximal over all paths between $\mathbf{y}$ and $\mathbf{z}$.
2. (ii)
(Leftmost and rightmost finite geodesics along paths) If, for some $\theta>0$,
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, and
$\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$, the points
$\mathbf{y}\leq\mathbf{z}\in\mathbb{Z}\times\mathbb{R}$ both lie on the
leftmost semi-infinite geodesic in
$\mathbf{T}_{\mathbf{x}}^{\theta{\scaleobj{0.8}{\boxempty}}}$, then the
portion of this geodesic between $\mathbf{y}$ and $\mathbf{z}$ coincides with
the leftmost finite geodesic between these two points. Similarly, the
rightmost semi-infinite geodesic is the rightmost geodesic between any two of
its points.
3. (iii)
(Monotonicity) The following inequalities hold.
1. (a)
For all $0<\gamma<\theta$, all $(m,t)\in\mathbb{Z}\times\mathbb{R}$, and
$r\geq m$,
$t\leq\tau_{(m,t),r}^{\gamma-,L}\leq\tau_{(m,t),r}^{\gamma+,L}\leq\tau_{(m,t),r}^{\theta-,L}\leq\tau_{(m,t),r}^{\theta+,L}\qquad\text{and}\qquad
t\leq\tau_{(m,t),r}^{\gamma-,R}\leq\tau_{(m,t),r}^{\gamma+,R}\leq\tau_{(m,t),r}^{\theta-,R}\leq\tau_{(m,t),r}^{\theta+,R}.$
2. (b)
For all $\theta>0$, $m\leq r\in\mathbb{Z}$, $s<t\in\mathbb{R}$, and
${\scaleobj{0.8}{\boxempty}}\in\\{+,-\\}$,
$\tau_{(m,s),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}\qquad\text{and}\qquad\tau_{(m,s),r}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},R}.$
3. (c)
For $\theta>0$, on the $\theta$-dependent full-probability event
$\widetilde{\Omega}^{(\theta)}$ of Theorem 3.1(iii), for all pairs of initial
points $(m,s)$ and $(m,t)$ in $\mathbb{Z}\times\mathbb{R}$ that satisfy $s<t$,
we have
$\tau_{(m,s),r}^{\theta,R}\leq\tau_{(m,t),r}^{\theta,L}\quad\text{ for all
$r\geq m$.}$
4. (iv)
(Convergence) The following limits hold.
1. (a)
For all $(m,t)\in\mathbb{Z}\times\mathbb{R}$, $r\geq m$, $\theta>0$, and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$\lim_{\gamma\nearrow\theta}\tau_{(m,t),r}^{\gamma{\scaleobj{0.8}{\boxempty}},L}=\tau_{(m,t),r}^{\theta-,L}\qquad\text{and}\qquad\lim_{\delta\searrow\theta}\tau_{(m,t),r}^{\delta{\scaleobj{0.8}{\boxempty}},R}=\tau_{(m,t),r}^{\theta+,R}.$
2. (b)
For all $(m,t)\in\mathbb{Z}\times\mathbb{R}$, $r\geq m$,
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, and $S\in\\{L,R\\}$,
$\lim_{\theta\searrow
0}\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},S}=t\qquad\text{and}\qquad\lim_{\theta\rightarrow\infty}\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},S}=\infty.$
3. (c)
For all $(m,t)\in\mathbb{Z}\times\mathbb{R}$, $r\geq m$, $\theta>0$, and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$\lim_{s\nearrow
t}\tau_{(m,s),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}=\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}\qquad\text{and}\qquad\lim_{u\searrow
t}\tau_{(m,u),r}^{\theta{\scaleobj{0.8}{\boxempty}},R}=\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},R},$
5. (v)
(Directness) For all $\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$, $\theta>0$,
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, and all $\\{\tau_{r}\\}_{r\geq
m}\in\mathbf{T}_{\mathbf{x}}^{\theta{\scaleobj{0.8}{\boxempty}}}$,
$\lim_{n\rightarrow\infty}\frac{\tau_{n}}{n}=\theta.$
###### Remark 4.4 (A look ahead).
In future work, we will use the joint distribution of Busemann functions to
build on these results and strengthen parts of Theorem 4.3. Specifically, Part
(iv)(iv)(a) can be made stronger in the following way: There exists an event
of full probability on which, for all $\theta>0,m\leq
r\in\mathbb{Z},s<t\in\mathbb{R},S\in\\{L,R\\}$ and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, there exists $\varepsilon>0$ such
that
$\tau_{(m,s),r}^{\gamma{\scaleobj{0.8}{\boxempty}},S}=\tau_{(m,s),r}^{\theta-,S}\text{
for all
}\theta-\varepsilon<\gamma<\theta,\qquad\text{and}\qquad\tau_{(m,s),r}^{\delta{\scaleobj{0.8}{\boxempty}},S}=\tau_{(m,s),r}^{\theta+,S}\text{
for all }\theta<\delta<\theta+\varepsilon.$
This is used to strengthen Part (iii)(iii)(c) to show that, on this event, for
each $m\in\mathbb{Z}$, $m\leq r\in\mathbb{Z},s<t\in\mathbb{R},\theta>0$, and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$,
$\tau_{(m,s),r}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}.$
Part (iii)(iii)(a) cannot be strengthened to compare
$\tau_{(m,t),r}^{\gamma{\scaleobj{0.8}{\boxempty}},R}$ and
$\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}$ for general
$\gamma<\theta$. Specifically, there exists $\gamma<\theta$ such that
$\tau_{(m,t),r}^{\gamma-,R}=\tau_{(m,t),r}^{\gamma+,R}>\tau_{(m,t),r}^{\theta-,L}=\tau_{(m,t),r}^{\theta+,L}.$
The following theorem shows that the Busemann geodesics give control over all
semi-infinite geodesics.
###### Theorem 4.5.
The following hold on the full probability event $\Omega_{2}$.
1. (i)
(Control of finite geodesics) Let
$\theta>0,(m,t)\in\mathbb{Z}\times\mathbb{R}$, and let $\\{t_{n}\\}$ be any
sequence that has direction $\theta$. For all $n$ sufficiently large so that
$n\geq m$ and $t_{n}\geq t$, let $t=t_{n,m-1}\leq t_{n,m}\leq\cdots\leq
t_{n,n}=t_{n}$ be any sequence that defines a (finite) geodesic between
$(m,t)$ and $(n,t_{n})$. Then, for each $r\geq m$,
$\tau_{(m,s),r}^{\theta-,L}\leq\liminf_{n\rightarrow\infty}t_{n,r}\leq\limsup_{n\rightarrow\infty}t_{n,r}\leq\tau_{(m,s),r}^{\theta+,R}.$
2. (ii)
(Control of semi-infinite geodesics) If, for some $\theta>0$ and
$(m,t)\in\mathbb{Z}\times\mathbb{R}$, any other geodesic (constructed from the
Busemann functions or not) is defined by the sequence $t=t_{m-1}\leq
t_{m}\leq\cdots$, starts at $(m,t)$, and has direction $\theta$, then for all
$r\geq m$,
$\tau_{(m,t),r}^{\theta-,L}\leq t_{r}\leq\tau_{(m,t),r}^{\theta+,R}.$
3. (iii)
(Convergence of finite geodesics) Assume that
$\mathbf{T}_{\mathbf{(}m,t)}^{\theta}$ contains a single element
$\\{\tau_{r}\\}_{r\geq m-1}$. If $\\{t_{n}\\}$ is a $\theta$-directed sequence
and, for each $n$, the sequence $t=t_{n,m-1}\leq t_{n,m}\leq\cdots\leq
t_{n,n}$ defines a finite geodesic between $(m,t)$ and $(n,t_{n})$, then
$\lim_{n\rightarrow\infty}t_{n,r}=\tau_{r}\qquad\text{for all }r\geq m.$
###### Remark 4.6.
In Part (iii), the assumption of uniqueness holds, for example, on the event
$\Omega_{(m,t)}^{(\theta)}$ of Theorem 3.1(ii). However, this assumption does
not extend to all $(m,t)\in\mathbb{Z}\times\mathbb{R}$ and $\theta>0$
simultaneously with probability one, as discussed in the following section.
### 4.3. Non-uniqueness of semi-infinite geodesics
There are two types of non-uniqueness of semi-infinite geodesics from an
initial point $\mathbf{x}$ into an asymptotic direction $\theta$.
(i) The first type, described in Theorem 3.1(iii) and in the next Theorem 4.7,
is caused by the continuum time variable and does not appear in the lattice
corner growth model. It is captured by the $L/R$ distinction. For each fixed
direction $\theta>0$ and level $m\in\mathbb{Z}$, this happens with probability
one at infinitely many locations. To illustrate, let $s^{\star}\geq 0$ be the
maximizer below:
$B_{m}(s^{\star})-h_{m+1}^{\theta}(s^{\star})=\sup_{0\leq
s<\infty}\\{B_{m}(s)-h_{m+1}^{\theta}(s)\\}.$
By Theorem 3.5(vii) and Lemma 7.1, with probability one, the maximizer
$s^{\star}$ is unique. By Theorem B.3, $s^{\star}>0$ with probability one. By
Theorem 3.5(v),
$t^{\star}:=\sup\\{t<0:B_{m}(t)-h_{m+1}^{\theta}(t)=B_{m}(s^{\star})-h_{m+1}^{\theta}(s^{\star})\\}\qquad\text{exists
in }\mathbb{R}_{<0}.$
Then, both $t^{\star}$ and $s^{\star}$ are maximizers of
$B_{m}(s)-h_{m+1}^{\theta}(s)$ on $[t^{\star},\infty)$. This gives at least
two distinct sequences in the set $\mathbf{T}_{(m,t^{\star})}^{\theta}$, with
$\tau_{(m,t^{\star}),m}^{\theta,L}=t^{\star}$ and
$\tau_{(m,t^{\star}),m}^{\theta,R}=s^{\star}$.
This presents a new type of non-uniqneness that is not present in discrete
last-passage percolation with exponential weights. However, when $\theta$ is
fixed, $\theta$-directed geodesics can disagree only for a finite amount of
time, because Theorem 3.1(vii) forces them to eventually come back together.
(ii) The second type of non-uniqueness of semi-infinite geodesics is captured
by the $\theta\pm$ distinction. Hence, it happens with probability zero at a
fixed $\theta$ and thereby requires investigation of the full Busemann process
and the full collection of all semi-infinite geodesics. In contrast to the
first type of non-uniqueness, this bears some similarity to the behavior
present in discrete last-passage percolation shown in [JRS21].
In future work, we show that there exists a random countable set of directions
$\theta$ such that, out of every initial point, there are two
$\theta$-directed geodesics. These geodesics may initially stay together for a
while, but eventually they separate for good and never come back together.
Furthermore, there is a distinguished subset of initial points at which
geodesics with the same direction split immediately. This set will be shown to
have almost surely Hausdorff dimension $\frac{1}{2}$.
The following theorem clarifies the non-uniqueness described by (i) above. Fix
$\theta>0$. On a full probability event where the $\theta\pm$ distinction is
not present, define the following sets:
$\displaystyle\operatorname{NU}_{0}^{\theta}$
$\displaystyle=\\{(m,t)\in\mathbb{Z}\times\mathbb{R}:\tau_{(m,t),r}^{\theta,L}<\tau_{(m,t),r}^{\theta,R}\text{
for some }r\geq m\\},\qquad\text{and}$
$\displaystyle\operatorname{NU}_{1}^{\theta}$
$\displaystyle=\\{(m,t)\in\operatorname{NU}_{0}^{\theta}:\tau_{(m,t),m}^{\theta,L}<\tau_{(m,t),m}^{\theta,R}\\}.$
Since $\theta>0$ is fixed, by Theorem 4.5(ii),
$\operatorname{NU}_{0}^{\theta}$ is almost surely the set of points
$\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$ such that the $\theta$-directed
semi-infinite geodesic from $\mathbf{x}$ is not unique. Its subset
$\operatorname{NU}_{1}^{\theta}$ is the set of initial points from which two
$\theta$-directed geodesics separate on the first level.
###### Theorem 4.7.
There exists a full probability event $\widetilde{\Omega}^{(\theta)}$ on which
the following hold.
1. (i)
The sets $\operatorname{NU}_{0}^{\theta}$ and $\operatorname{NU}_{1}^{\theta}$
are countably infinite and can be written as
$\displaystyle\operatorname{NU}_{0}^{\theta}$
$\displaystyle=\\{(m,t)\in\mathbb{Z}\times\mathbb{R}:t=\tau_{(m,t),r}^{\theta,L}<\tau_{(m,t),r}^{\theta,R}\text{
for some }r\geq m\\},\qquad\text{and}$
$\displaystyle\operatorname{NU}_{1}^{\theta}$
$\displaystyle=\\{(m,t)\in\operatorname{NU}_{0}^{\theta}:t=\tau_{(m,t),m}^{\theta,L}<\tau_{(m,t),m}^{\theta,R}\\}.$
For each $(m,t)\in\mathbb{Z}\times\mathbb{R}$ and $\theta>0$, at most one
geodesic in $\mathbf{T}_{(m,t)}^{\theta}$ passes horizontally through
$(m,t+\varepsilon)$ for some $\varepsilon>0$.
2. (ii)
The set $\operatorname{NU}_{1}^{\theta}$ is neither discrete nor dense in
$\mathbb{Z}\times\mathbb{R}$. More specifically, for each point
$(m,t)\in\operatorname{NU}_{1}^{\theta}$ and every $\varepsilon>0$, there
exists $s\in(t-\varepsilon,t)$ such that
$(m,s)\in\operatorname{NU}_{1}^{\theta}$. For each
$(m,t)\in\operatorname{NU}_{1}^{\theta}$, there exists $\delta>0$ such that,
for all $s\in(t,t+\delta)$, $(m,s)\notin\operatorname{NU}_{0}^{\theta}$.
###### Remark 4.8.
Part (i) states that, on $\widetilde{\Omega}^{(\theta)}$, if there exist
multiple $\theta$-directed geodesics out of $(m,t)$, then these geodesics
separate one by one from the upward vertical ray at $(m,t)$. The set
$\operatorname{NU}_{1}^{\theta}$ is the subset of
$\operatorname{NU}_{0}^{\theta}$ such that two geodesics separate immediately
at the initial point. See Figures 4.2 and 4.3.
$(r,\tau_{(m,t),r}^{\theta,R})$$(r,t)=(r,\tau_{(m,t),r}^{\theta,L})$$(m,t)$
Figure 4.2. In this figure,
$(m,t)\in\operatorname{NU}_{0}^{\theta}\setminus\operatorname{NU}_{1}^{\theta}$.
The two geodesics split on the vertical line containing the initial point, but
they must come back together to coalesce.
$(m,t)=(m,\tau_{(m,t),m}^{\theta,L})$$(m,\tau_{(m,t),m}^{\theta,R})$ Figure
4.3. In this figure, $(m,t)\in\operatorname{NU}_{1}^{\theta}$. The two
geodesics split immediately from the initial point, but later coalesce.
### 4.4. Dual geodesics and coalescence
To prove the coalescence of Theorem 3.1(vii), we use the dual field
$\mathbf{X}^{\theta}$ of independent Brownian motions from (3.9) and their
southwest-directed semi-infinite geodesics. We use this to construct the BLPP
analogue of Pimentel’s dual tree [Pim16] and then adapt the argument of
[Sep20].
Since Brownian motion is symmetric, in distribution, about the origin, there
exist $\theta$-directed dual southwest semi-infinite geodesics for the
environment $\mathbf{X}^{\theta}$. These are constructed in a very similar
manner as the northeast geodesics in Definition 4.1. Specifically, for
$(m,t)\in\mathbb{Z}\times\mathbb{R}$, let $\mathbf{T}_{(m,t)}^{\theta,\star}$
be the set of sequences $t=\tau_{m}^{\star}\geq\tau_{m-1}^{\star}\geq\cdots$
satisfying
(4.2)
$h_{r-1}^{\theta}(\tau_{r-1}^{\star})-X_{r}^{\theta}(\tau_{r-1}^{\star})=\sup_{-\infty<s\leq\tau_{r}^{\star}}\\{h_{r-1}^{\theta}(s)-X_{r}^{\theta}(s)\\}\qquad\text{for
each }r\leq m.$
Define the leftmost and rightmost sequences similarly, by
$\tau_{(m,t),r}^{\theta,L\star}$ and $\tau_{(m,t),r}^{\theta,R\star}$. These
sequences define southwest semi-infinite paths, similar as for the northeast
paths. We graphically represent southwest paths on the plane, where the
continuous coordinate is not changed, but the discrete coordinate is shifted
down by $\frac{1}{2}$. That is, $\text{for }m\in\mathbb{Z},\text{ denote
}m^{\star}=m-\frac{1}{2}$, and for
$\mathbf{x}=(m,t)\in\mathbb{Z}\times\mathbb{R},\text{ denote
}\mathbf{x}^{\star}=\bigl{(}m-\frac{1}{2},t\bigr{)}$. Then, for
$\\{\tau_{r}^{\star}\\}_{r\leq m}\in\mathbf{T}_{(m,t)}^{\theta,\star}$, the
southwest path consists of horizontal and vertical line segments, where
$\tau_{r}^{\star}$ denotes the position of the vertical segment connecting
levels $(r+1)^{\star}$ and $r^{\star}$. Specifically, the path consists of the
points
$\bigcup_{r=-\infty}^{m}\\{(r^{\star},u):u\in[\tau_{r-1}^{\star},\tau_{r}^{\star}]\\}\cup\bigcup_{r=-\infty}^{m-1}\\{(v,\tau_{r}^{\star}):v\in[(r-1)^{\star},r^{\star}]\\}.$
Figure 4.4 shows the regular axes and the dual axes together, with a southwest
dual geodesic traveling on this dual plane. Each element of
$\mathbf{T}_{\mathbf{(}m,t)}^{\theta\star}$ is a southwest semi-infinite
geodesic for the dual environment $\mathbf{X}^{\theta}$. This fact is recorded
in Theorem 7.11.
$(m-1)^{\star}$$m-1$$m^{\star}$$m$$\vdots$$t$$\tau_{(m,t),m-1}^{\star}$$\tau_{(m,t),m-2}^{\star}$$\cdots$
Figure 4.4. The original discrete levels (solid) and the dual levels (dashed).
Dual levels are labeled with a $\star$.
Since $\mathbf{X}^{\theta}$ is an environment of i.i.d. Brownian motions, the
following theorem allows us to conclude Part (vii) of Theorem 3.1 from Part
(vi). Full details of this connection are found in the proofs. Refer to Figure
4.5 for clarity.
$\mathbf{x}$$\mathbf{y}$ Figure 4.5. The outcome of Theorem 4.9: two disjoint
semi-infinite northeast paths (red/thick) and the dual bi-infinite path
(blue/thin)
###### Theorem 4.9.
Fix $\theta>0$. With probability one, if for any
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$, the rightmost semi-
infinite geodesics in $\mathbf{T}_{\mathbf{x}}^{\theta}$ and
$\mathbf{T}_{\mathbf{y}}^{\theta}$ are disjoint (i.e. the paths share no
points), then there exists a bi-infinite, upright path defined by jump times
$\cdots\leq\tau_{-1}^{\star}\leq\tau_{0}^{\star}\leq\tau_{1}^{\star}\leq\cdots$
(where $\tau_{i}^{\star}$ denotes the jump time from level $i$ to level $i+1$)
that satisfies the following:
1. (i)
For any point $\mathbf{x}^{\star}$ along the path, the portion of that path to
the south and west of $\mathbf{x}^{\star}$ is the leftmost semi-infinite
geodesic in the set $\mathbf{T}_{\mathbf{x}}^{\theta,\star}$. Specifically,
when shifted back up by $\frac{1}{2}$, the path is a bi-infinite geodesic for
the environment $\mathbf{X}^{\theta}$.
2. (ii)
The sequence $\\{\tau_{n}^{\star}\\}_{n\in\mathbb{Z}}$ satisfies
$\lim_{n\rightarrow\infty}\frac{\tau_{n}^{\star}}{n}=\theta=\lim_{n\rightarrow\infty}\frac{\tau_{-n}^{\star}}{-n}.$
The analogous result holds if we assume that two leftmost $\theta$-directed
semi-infinite geodesics are disjoint. In this case, the portion of the path to
the south and west of each of its points is the rightmost dual semi-infinite
geodesic.
## 5\. Connections to other models
### 5.1. Connection to infinite-length polymer measures for the Brownian
polymer
Recalling the definitions at the beginning of Section 2.2, for
$(m,s)\leq(n,t)\in\mathbb{Z}\times\mathbb{R}$, the point-to-point partition
function of the Brownian polymer with unit temperature is
$Z_{(m,s),(n,t)}(\mathbf{B})=\int
e^{\mathcal{E}(\mathbf{s}_{m,n})}\mathbf{1}\\{\mathbf{s}_{m,n-1}\in\Pi_{(m,s),(n,t)}\\}d\mathbf{s}_{m,n-1}.$
The associated quenched polymer measure on $\Pi_{(m,s),(n,t)}$ is
$Q^{\mathbf{B}}(\tau_{m}\in ds_{m},\ldots,\tau_{n-1}\in
ds_{n-1})=\frac{1}{Z_{(m,s),(n,t)}(\mathbf{B})}\exp\Bigl{(}\sum_{k=m}^{n}B_{k}(s_{k-1},s_{k})\Bigr{)}1\\{\mathbf{s}_{m,n-1}\in\Pi_{(m,s),(n,t)}\\}d\mathbf{s}_{m,n-1}.$
Brownian last-passage percolation is the zero-temperature analogue of the
Brownian or O’Connell-Yor polymer. This is made precise by the limit
$\lim_{\beta\rightarrow\infty}\frac{1}{\beta}\log
Z_{(m,s),(n,t)}(\beta\mathbf{B})=L_{(m,s),(n,t)}(\mathbf{B}),$
which, by convergence of $L^{p}$ norms, holds in a deterministic sense as long
as $\mathbf{B}$ is a field of continuous functions
$\mathbb{R}\rightarrow\mathbb{R}$. The parameter $\beta$ is used to denote
inverse temperature.
Alberts, Rassoul-Agha, and Simper [ARAS20] showed the existence of Busemann
functions and infinite-length limits of the quenched measures when the right
endpoint $(n,t_{n})$ satisfies $t_{n}/n\rightarrow\theta$ for some fixed
$\theta>0$. The Busemann functions are defined as
$\widehat{\mathcal{B}}^{\theta}(\mathbf{x},\mathbf{y}):=\lim_{n\rightarrow\infty}\log\frac{Z_{\mathbf{x},(n,t_{n})}(\mathbf{B})}{Z_{\mathbf{y},(n,t_{n})}(\mathbf{B})}.$
Similarly as in (2.6), define
$\widehat{h}_{m}^{\theta}(t)=\widehat{\mathcal{B}}^{\theta}((m,0),(m,t))$.
Using the “Proof of Theorem 2.5, assuming Theorem 3.1” on page 1937 of
[ARAS20] and an analogous construction of the Busemann functions for all
initial points as in Section 6 of the present paper, it can be shown that, the
process $t\mapsto\widehat{h}_{m}^{\theta}(t)$ is a two-sided Brownian motion
with drift, independent of $\\{B_{r}\\}_{r\leq m}$. By Theorems 3.5(vii) and
3.7(i), the same is true for the zero temperature Busemann process $t\mapsto
h_{m}^{\theta}(t)$. An infinite, up-right path under the quenched infinite-
length polymer measure is a continuous-time Markov chain, starting from a
point $(m,t)$ and defined by jump times $t=\tau_{m-1}\leq\tau_{m}\leq\cdots$.
By Equation (2.3) in [ARAS20], the quenched conditional distribution of
$\tau_{r}$ given $\tau_{r-1}$ is
$\displaystyle Q^{\mathbf{B}}_{(m,s),(n,t_{n})}(\tau_{r}\in
ds_{r}|\tau_{r-1}=s_{r-1})$
$\displaystyle=e^{B_{r}(s_{r-1},s_{r})}\frac{Z_{(r+1,s_{r}),(n,t_{n})}(\mathbf{B})}{Z_{(r,s_{r-1}),(n,t_{n})}(\mathbf{B})}\mathbf{1}\\{s_{r-1}\leq
s_{r}\\}\,ds_{r}.$
The proof of existence of infinite length measures requires a rigorous
tightness argument, but to motivate the connection to BLPP, we formally take
limits as $n\rightarrow\infty$ to yield the conditional measure
$\displaystyle\;\exp\bigl{(}B_{r}(s_{r-1},s_{r})+\widehat{\mathcal{B}}^{\theta}((r+1,s_{r}),(r,s_{r-1}))\bigr{)}\mathbf{1}\\{s_{r-1}\leq
s_{r}\\}\,ds_{r}$
$\displaystyle=\exp\bigl{(}B_{r}(s_{r-1},s_{r})-\widehat{h}_{r+1}(s_{r})+\widehat{\mathcal{B}}^{\theta}((r+1,0),(r,s_{r-1})\bigr{)}\mathbf{1}\\{s_{r-1}\leq
s_{r}\\}\,ds_{r}.$
Now, note that the term $\widehat{\mathcal{B}}^{\theta}((r+1,0),(r,s_{r-1})$
does not depend on $s_{r}$. We can then think of the connection between BLPP
and the O’Connell-Yor polymer in the following sense: In the positive
temperature case, given the environment $\mathbf{B}$, the transition density
from $\tau_{r-1}$ to $\tau_{r}$ is given by
$C\exp\bigl{(}B_{r}(s)-\widehat{h}_{r+1}^{\theta}(s)\bigr{)}\mathbf{1}\\{\tau_{r-1}\leq
s\\}\,ds,$
where $C$ is a normalizing constant depending on $r$ and $\tau_{r-1}$. On the
other hand, by Definition 4.1, in the zero-temperature case, given the
environment $\mathbf{B}$ and the previous jump $\tau_{r-1}$, the jump
$\tau_{r}$ is chosen in a deterministic fashion by maximizing
$B_{r}(s)-h_{r+1}^{\theta}(s)$ over $s\in[\tau_{r-1},\infty)$.
### 5.2. Connection to semi-infinite geodesics in discrete last-passage
percolation
The corner growth model, or discrete last-passage percolation, is defined as
follows. Let $\\{Y_{\mathbf{x}}\\}_{\mathbf{x}\in\mathbb{Z}^{2}}$ be a
collection of nonnegative i.i.d random variables, each associated to a vertex
on the integer lattice. For
$\mathbf{x}\leq\mathbf{y}\in\mathbb{Z}\times\mathbb{Z}$, define the last-
passage time as
$G_{\mathbf{x},\mathbf{y}}=\sup_{\mathbf{x}_{\centerdot}\in\Pi_{\mathbf{x},\mathbf{y}}}\sum_{k=0}^{|\mathbf{y}-\mathbf{x}|_{1}}Y_{\mathbf{x}_{k}},$
where $\Pi_{x,y}$ is the set of up-right paths
$\\{\mathbf{x}_{k}\\}_{k=0}^{n}$ that satisfy
$\mathbf{x}_{0}=\mathbf{x},\mathbf{x}_{n}=\mathbf{y}$, and
$\mathbf{x}_{k}-\mathbf{x}_{k-1}\in\\{\mathbf{e}_{1},\mathbf{e}_{2}\\}$. Under
the assumption that $Y_{0}$ has finite second moment, Theorem 3.1 and
Corollary 3.1 of [GW91] introduced BLPP as a universal scaling limit of the
corner growth model, where one variable is scaled, and the other is held
constant. That is, if $Y_{0}$ is normalized to have unit mean and variance,
$\\{L_{(m,s),(n,t)}:(m,s)\leq(n,t)\in\mathbb{Z}\times\mathbb{R}\\}$
is the functional limit, as $k\rightarrow\infty$, of the properly interpolated
version of the process
$\bigl{\\{}\frac{1}{\sqrt{k}}G_{(m,\lfloor sk\rfloor),(n,\lfloor
tk\rfloor)}-(t-s)k:(m,s)\leq(n,t)\in\mathbb{Z}\times\mathbb{R}\bigr{\\}}.$
The most tractable case of discrete last-passage percolation is the case where
$Y_{0}$ has the exponential distribution with rate $1$. In this case, Busemann
functions exist and are indexed by a direction vector $\mathbf{u}$. They are
defined by
$U^{\mathbf{u}}(\mathbf{x},\mathbf{y}):=\lim_{n\rightarrow\infty}G_{(\mathbf{x},\mathbf{z}_{n})}-G_{(\mathbf{y},\mathbf{z}_{n})},$
where $\mathbf{z}_{n}$ satisfies $\mathbf{z}_{n}/n\rightarrow\mathbf{u}$ for a
fixed direction $\mathbf{u}$. For a given $\omega\in\Omega$,
$\mathbf{x}\in\mathbb{Z}^{2}$, and direction $\mathbf{u}$, a semi-infinite
geodesic $\mathbf{\gamma}$ is defined by the sequence
$\\{\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}\\}_{k\in\mathbb{Z}_{\geq 0}}$.
At each step, a choice is made to move upward or to the right. First, set
$\mathbf{\gamma}^{\mathbf{u},\mathbf{x}}_{0}=\mathbf{x}$, and for $k\geq 0$,
(5.1)
$\mathbf{\gamma}_{k+1}^{\mathbf{u},\mathbf{x}}=\begin{cases}\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}+\mathbf{e}_{1},&\text{if}\quad
U^{\mathbf{u}}\bigl{(}\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}},\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}+\mathbf{e}_{1}\bigr{)}\leq
U^{\mathbf{u}}\bigl{(}\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}},\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}+\mathbf{e}_{2}\bigr{)},\\\\[10.00002pt]
\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}+\mathbf{e}_{2},&\text{if}\quad
U^{\mathbf{u}}\bigl{(}\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}},\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}+\mathbf{e}_{2}\bigr{)}<U^{\mathbf{u}}\bigl{(}\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}},\mathbf{\gamma}_{k}^{\mathbf{u},\mathbf{x}}+\mathbf{e}_{1}\bigr{)}.\end{cases}$
In the case of exponential weights, this sequence is a semi-infinite geodesic
with direction $\mathbf{u}$. This construction is inherently discrete, so it
does not extend directly to the case of BLPP. However, this construction is
equivalent to taking a sequence of maximizers of the appropriate function,
analogous to Definition 4.1. The following is a discrete analogue of Theorem
3.5(vi) that holds in the case of exponential weights.
(5.2) $U^{\mathbf{u}}((m,r),(m,r+1))=\max_{\begin{subarray}{c}m\leq
k<\infty\\\
k\in\mathbb{Z}\end{subarray}}\Big{\\{}\sum_{i=m}^{k}Y_{(i,r)}+U^{\mathbf{u}}((k,r+1),(m,r+1))\Big{\\}}.$
Now, suppose that, starting at the point $(m,r)$, the semi-infinite geodesic
constructed in (5.1) makes an $\mathbf{e}_{2}$ step from $(k,r)$ to $(k,r+1)$
for some $k\geq m$. Then, we show that $k$ is maximal for (5.2). Indeed, by
(5.1), if the path $\mathbf{\gamma}$ makes an $\mathbf{e}_{2}$ step from
$(k,r)$ to $(k,r+1)$, then
$\displaystyle U^{\mathbf{u}}((i,r),(i+1,r))\leq
U^{\mathbf{u}}((i,r),(i,r+1))\qquad\text{ for }\qquad m\leq i\leq
k-1,\qquad\text{and }$ $\displaystyle
U^{\mathbf{u}}((k,r),(k+1,r))>U^{\mathbf{u}}((k,r),(k,r+1)).$
Using the identity $Y_{\mathbf{x}}=U^{\mathbf{u}}(\mathbf{x,x+e_{1}})\wedge
U^{\mathbf{u}}(\mathbf{x,x+e_{2}})$ and additivity of the Busemann functions,
$\displaystyle\quad\sum_{i=m}^{k}Y_{(i,r)}+U^{\mathbf{u}}((k,r+1),(m,r+1))$
$\displaystyle=\sum_{i=m}^{k-1}U^{\mathbf{u}}((i,r),(i+1,r))+U^{\mathbf{u}}((k,r),(k,r+1))+U^{\mathbf{u}}((k,r+1),(m,r+1))$
$\displaystyle=U^{\mathbf{u}}((m,r),(m,r+1)),$
so $k$ is indeed maximal in for the right hand-side of (5.2). The inductive
step follows in the same manner. Hence, we see that the construction of semi-
infinite geodesics in Definition 4.1 is a continuous analogue of the procedure
for the discrete case when viewed from the perspective of maximizers.
### 5.3. Connection to queuing theory
Fix $\theta>0$, and consider the almost surely unique $\theta$-directed semi-
infinite geodesic (Theorem 3.1(ii)) starting from $(0,0)$ and defined by the
sequence of jump times $\\{\tau_{r}\\}_{r\geq-1}$ with $\tau_{-1}=0$. For each
$r\geq 0$, associate a queuing station as follows. For $s<t$, let
$\frac{1}{\sqrt{\theta}}(t-s)-X_{r+1}^{\theta}(s,t)$ denote the service
available in the interval $(s,t]$ and let
$\frac{1}{\sqrt{\theta}}(t-s)-h_{r}^{\theta}(s,t)$ denote the arrivals to the
queue in the interval $(s,t]$. Here, the parameter $\frac{1}{\sqrt{\theta}}$
dictates the rate of service. See Appendix C for more details about the
queuing setup. By (3.11),
$h_{r+1}^{\theta}=\overleftarrow{D}(h_{r}^{\theta},X_{r+1})$, so as in the
proof of Theorem C.2, $\frac{1}{\sqrt{\theta}}(t-s)-h_{r+1}^{\theta}(s,t)$
gives the departures from the $r$th station in the interval $(s,t]$, and the
departures process from the $r$th station becomes the arrivals process for the
$r+1$st station. In other words, once a customer is served at the $r$th
station, they move into the queue at the $r+1$st station. By Equations (3.10)
and (3.11), for all $r\in\mathbb{Z}$ and $t\in\mathbb{R}$,
$v_{r+1}^{\theta}(t)=\sup_{t\leq
u<\infty}\\{B_{r}(t,u)-h_{r+1}^{\theta}(t,u)\\}=\sup_{-\infty<u\leq
t}\\{X_{r+1}^{\theta}(t,u)-h_{r}^{\theta}(t,u)\\}=\overleftarrow{Q}(h_{r}^{\theta},X_{r+1}^{\theta})(t).$
In queuing terms, $v_{r+1}^{\theta}(t)$ gives the length of the $r$th queue at
time $t$. As $\tau_{r}$ is the maximizer of $B_{r}(u)-h_{r+1}^{\theta}(u)$
over $u\in[\tau_{r-1},\infty)$, $\tau_{r}$ is the first time $u\geq\tau_{r-1}$
such that $v_{r+1}^{\theta}(u)=0$ (see Lemma 7.4). In queuing terms,
$\tau_{r}$ is the first time greater than or equal to $\tau_{r-1}$ at which
the queue is empty. Thus, the semi-infinite geodesic represents the movement
of a customer through the infinite series of queuing stations: the customer
starts at station $0$ at time $0$ and is served at the $r$th station at a time
no later than $\tau_{r}$.
## 6\. Construction and proofs for the Busemann process
The remainder of the paper is devoted to the proofs of the theorems. This
section constructs the global Busemann process, with the proof of Theorem 3.5
as the ultimate result. Section 7 proves all results about the semi-infinite
geodesics, culminating in the proof of Theorem 3.1.
### 6.1. Construction of Busemann functions for a fixed direction
###### Definition 6.1 (Definition of Busemann functions for fixed $\theta$
and a countable dense set of points).
For a fixed $\theta>0$, and
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{Q}$, let
$\Omega_{0}^{(\theta)}=\bigcap_{\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{Q}}\Omega_{\mathbf{x},\mathbf{y}}^{(\theta)},$
where the $\Omega_{\mathbf{x},\mathbf{y}}^{(\theta)}$ are the events of
Theorem 2.3. Then, $\mathbb{P}(\Omega_{0}^{(\theta)})=1$. On this event, for
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{Q}$, define
$\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})$ by (2.4). By definition in terms
of limits, it is clear that the Busemann functions are additive, that is, for
$\omega\in\Omega_{0}^{(\theta)}$ and
$\mathbf{x},\mathbf{y},\mathbf{z}\in\mathbb{Z}\times\mathbb{Q}$,
(6.1)
$\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})+\mathcal{B}^{\theta}(\mathbf{y},\mathbf{z})=\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y}).$
Hence, the entire collection of Busemann functions is constructed from the
collection of horizontal and vertical Busemann functions
$\\{h_{m}^{\theta},v_{m}^{\theta}\\}_{m\in\mathbb{Z}}$ (Equations (2.6) and
(2.5)).
Set $Y(t)=-B_{0}(t)+\frac{1}{\sqrt{\theta}}t$. By Lemma C.5, as elements of
the space of functions $\mathbb{Q}\rightarrow\mathbb{R}$, equipped with the
standard product $\sigma-$algebra,
$\bigl{\\{}h_{0}^{\theta}(t):t\in\mathbb{Q}\bigr{\\}}\overset{d}{=}\bigl{\\{}Y(t):t\in\mathbb{Q}\bigr{\\}}\qquad\text{and}\qquad\bigl{\\{}v_{1}^{\theta}(t):t\in\mathbb{Q}\bigr{\\}}\overset{d}{=}\bigl{\\{}\overleftarrow{Q}(Y,B_{1})(t):t\in\mathbb{Q}\bigr{\\}}.$
For $\omega\in\Omega_{0}^{(\theta)}$, $m\in\mathbb{Z}$, and $t\in\mathbb{Q}$,
$h_{m}^{\theta}(t)=\lim_{n\rightarrow\infty}L_{(m,0),(n,n\theta)}(\mathbf{B})-L_{(m,t),(n,n\theta)}(\mathbf{B})=\lim_{n\rightarrow\infty}L_{(m,0),(n+m,(n+m)\theta)}(\mathbf{B})-L_{(m,t),(n+m,(n+m)\theta)}(\mathbf{B}).$
Since the environment $\mathbf{B}=\\{B_{m}\\}_{m\in\mathbb{Z}}$ is a field of
i.i.d. two-sided Brownian motions, the right-hand side implies that
$h_{m}^{\theta}(t)$ has the same distribution for all $m$. By this same
reasoning, for each $m\in\mathbb{Z}$,
$\\{h_{m}^{\theta}(t),v_{m+1}^{\theta}(t):t\in\mathbb{Q}\\}\overset{d}{=}\\{h_{0}^{\theta}(t),v_{1}^{\theta}(t):t\in\mathbb{Q}\\}.$
Since Brownian motion satisfies $\lim_{s\rightarrow\infty}\frac{B(s)}{s}=0$
almost surely, for all $m\in\mathbb{Z}$, the following limits also hold almost
surely:
$\lim_{\mathbb{Q}\ni
s\rightarrow\pm\infty}\bigl{[}B_{m}(s)-h_{m+1}^{\theta}(s)\bigr{]}=\mp\infty.$
Set
$\displaystyle\Omega^{(\theta)}=$
$\displaystyle\bigcap_{m,r\in\mathbb{Z}}\bigl{\\{}(s,t)\mapsto\mathcal{B}^{\theta}((m,s),(r,t))\text{
is uniformly continuous on all bounded subsets of
}\mathbb{Q}\times\mathbb{Q}\bigr{\\}}$ (6.2)
$\displaystyle\qquad\qquad\cap\bigcap_{m\in\mathbb{Z}}\\{\lim_{\mathbb{Q}\ni
s\rightarrow\pm\infty}\bigl{[}B_{m}(s)-h_{m+1}^{\theta}(s)\bigr{]}=\mp\infty\\}.$
By almost sure continuity of the functions $Y$ and
$\overleftarrow{Q}(Y,B_{1})$ (Lemma 3.4), $\mathbb{P}(\Omega^{(\theta)})=1$.
###### Definition 6.2 (Definition of Busemann functions for fixed $\theta$,
arbitrary points).
On the event $\Omega^{(\theta)}$, for arbitrary
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$, define
$\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})$ such that, for each
$m,r\in\mathbb{Z}$,
$(s,t)\mapsto\mathcal{B}^{\theta}((m,s),(r,t))$
is the unique continuous extension of this function from
$\mathbb{Q}\times\mathbb{Q}$ to $\mathbb{R}^{2}$.
The following lemma states properties for a fixed $\theta$, as a precursor to
the more general Theorems 3.5 and Theorem 3.7.
###### Lemma 6.3.
Let $\theta>0$. Then, the following hold.
1. (i)
On the event $\Omega^{(\theta)}$, for all
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$ and all sequences
$\\{t_{n}\\}$ satisfying $t_{n}/n\rightarrow\theta$,
$\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})=\lim_{n\rightarrow\infty}\bigl{[}L_{\mathbf{x},(n,t_{n})}-L_{\mathbf{y},(n,t_{n})}\bigr{]}.$
2. (ii)
On the event $\Omega^{(\theta)}$, whenever
$\mathbf{x},\mathbf{y},\mathbf{z}\in\mathbb{Z}\times\mathbb{R}$,
$\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})+\mathcal{B}^{\theta}(\mathbf{y},\mathbf{z})=\mathcal{B}^{\theta}(\mathbf{x},\mathbf{z}).$
3. (iii)
For $0<\gamma<\theta<\infty$, on the event
$\Omega^{(\theta)}\cap\Omega^{(\gamma)}$, for all $m\in\mathbb{Z}$ and
$s<t\in\mathbb{R}$,
$0\leq v_{m}^{\gamma}(s)\leq v_{m}^{\theta}(s),\text{ and }B_{m}(s,t)\leq
h_{m}^{\theta}(s,t)\leq h_{m}^{\gamma}(s,t).$
4. (iv)
On the event $\Omega^{(\theta)}$, for each $m\in\mathbb{Z}$,
$\lim_{s\rightarrow\pm\infty}B_{m}(s)-h_{m+1}^{\theta}(s)=\mp\infty.$
5. (v)
For every $m\in\mathbb{Z}$, the process $t\mapsto h_{m}^{\theta}(t)$ is a two-
sided Brownian motion with drift $\frac{1}{\sqrt{\theta}}$. For each
$m\in\mathbb{Z}$ and $t\in\mathbb{R}$,
$v_{m}^{\theta}(t)\sim\operatorname{Exp}\bigl{(}\frac{1}{\sqrt{\theta}}\bigr{)}$.
6. (vi)
For any $m\in\mathbb{Z}$, $\\{h_{r}^{\theta}\\}_{r>m}$ is independent of
$\\{B_{r}\\}_{r\leq m}$.
7. (vii)
There exists a full-probability event
$\Omega_{1}^{(\theta)}\subseteq\Omega^{(\theta)}$ on which the following hold
for all $m\in\mathbb{Z}$ and $t\in\mathbb{R}:$
$v_{m+1}^{\theta}(t)=Q(h_{m+1}^{\theta},B_{m})(t)\qquad\text{and}\qquad
h_{m}^{\theta}(t)=D(h_{m+1}^{\theta},B_{m})(t).$
###### Proof.
Part (i): On $\Omega^{(\theta)}$, let
$(m,s),(r,t)\in\mathbb{Z}\times\mathbb{R}$, and let $\\{t_{n}\\}$ be any
sequence with $t_{n}/n\rightarrow\theta$. Note that, whenever
$t_{1}<t_{2},m\in\mathbb{Z}$ and $\mathbf{y}\geq(m,t_{2})$,
$L_{(m,t_{1}),\mathbf{y}}\geq B_{m}(t_{1},t_{2})+L_{(m,t_{2}),\mathbf{y}}.$
Then, let $q_{1},q_{2}\in\mathbb{Q}$ be such that $q_{1}<s$ and $q_{2}>t$.
Then,
$L_{(m,s),(n,t_{n})}-L_{(r,t),(n,t_{n})}\leq
L_{(m,q_{1}),(n,t_{n})}-B_{m}(q_{1},s)-L_{(r,q_{2}),(n,t_{n})}-B_{r}(t,q_{2}),$
and therefore,
$\limsup_{n\rightarrow\infty}\bigl{[}L_{(m,s),(n,t_{n})}-L_{(r,t),(n,t_{n})}\bigr{]}\leq\mathcal{B}^{\theta}((m,q_{1}),(r,q_{2}))-B_{m}(q_{1},s)-B_{r}(t,q_{2}).$
Taking $q_{1}\nearrow s$ and $q_{2}\searrow t$ and using the continuity of
Brownian motion and the Busemann functions,
$\limsup_{n\rightarrow\infty}\bigl{[}L_{(m,s),(n,t_{n})}-L_{(r,t),(n,t_{n})}\bigr{]}\leq\mathcal{B}^{\theta}((m,s),(r,t)).$
A similar procedure gives the appropriate lower bound.
Part (ii): This follows from the additivity of the process on the countable
dense set (Equation (6.1)) and the construction of Definition 6.2 as the
unique continuous extension.
Part (iii): By Lemma A.4, for $s,t\in\mathbb{R}$ and all $n$ such that $n\geq
m$ and $n\gamma\geq s\vee t$,
$\displaystyle 0$ $\displaystyle\leq
L_{(m,s),(n,n\gamma)}-L_{(m+1,s),(n,n\gamma)}\leq
L_{(m,s),(n,n\theta)}-L_{(m+1,s),(n,n\theta)}$ $\displaystyle\text{and}\qquad
B_{m}(s,t)$ $\displaystyle\leq L_{(m,s),(n,n\theta)}-L_{(m,t),(n,n\theta)}\leq
L_{(m,s),(n,n\gamma)}-L_{(m,t),(n,n\gamma)}.$
The proof is complete by Part (i), taking limits as $n\rightarrow\infty$.
Part (iv): This follows from the definition of $\Omega^{(\theta)}$ and
continuity.
Part (v): Observe that $t\mapsto h_{m}^{\theta}(t)$ is a continuous process
with the correct finite-dimensional distributions by Lemma C.5, taking limits
when necessary. Part (i) and Theorem 2.3 guarantee that
$v_{m}^{\theta}(t)\sim\operatorname{Exp}\bigl{(}\frac{1}{\sqrt{\theta}}\bigr{)}$
for all $t\in\mathbb{R}$.
Part (vi): By Part (i),
$\\{h_{r}^{\theta}(t):r>m,t\in\mathbb{R}\\}=\bigl{\\{}\lim_{n\rightarrow\infty}\bigl{[}L_{(r,0),(n,n\theta)}-L_{(r,t),(n,n\theta)}\bigr{]}:r>m,t\in\mathbb{R}\bigr{\\}}.$
The right-hand side is a function of $\\{B_{r}\\}_{r>m}$, which is independent
of $\\{B_{r}\\}_{r\leq m}$.
Part (vii): We start with the first statement. We need to show that
(6.3) $v_{m+1}^{\theta}(t)=\sup_{t\leq
s<\infty}\\{B_{m}(t,s)-h_{m+1}^{\theta}(t,s)\\}.$
By Part (i),
(6.4) $\displaystyle\sup_{t\leq
s<\infty}\\{B_{m}(t,s)-h_{m+1}^{\theta}(t,s)\\}=\sup_{t\leq
s<\infty}\\{B_{m}(t,s)+\lim_{n\rightarrow\infty}\bigl{[}L_{(m+1,s),(n,n\theta)}-L_{(m+1,t),(n,n\theta)}\bigr{]}\\}.$
On the other hand,
(6.5)
$v_{m+1}^{\theta}(t)=\lim_{n\rightarrow\infty}\bigl{[}L_{(m,t),(n,n\theta)}-L_{(m+1,t),(n,n\theta)}\bigr{]}.$
For each $s\geq t$ and all $n$ sufficiently large so that $n\theta\geq s$ and
$n\geq m+1$,
$B_{m}(t,s)+L_{(m+1,s),(n,n\theta)}-L_{(m+1,t),(n,n\theta)}\leq
L_{(m,t),(n,n\theta)}-L_{(m+1,t),(n,n\theta)}.$
Taking limits as $n\rightarrow\infty$ and comparing (6.4) and (6.5)
establishes
$v_{m+1}^{\theta}(t)\geq\sup_{t\leq
s<\infty}\\{B_{m}(t,s)-h_{m+1}^{\theta}(t,s)\\}.$
Next, by Part (v),
$v_{m+1}^{\theta}(t)\sim\operatorname{Exp}(\frac{1}{\sqrt{\theta}})$. By Parts
(v) and (vi) and Lemma B.1,
$\displaystyle\sup_{t\leq
s<\infty}\\{B_{m}(t,s)-h_{m+1}^{\theta}(t,s)\\}\overset{d}{=}\sup_{0\leq
s<\infty}\bigl{\\{}\sqrt{2}B(s)-\frac{1}{\sqrt{\theta}}s\bigr{\\}}\overset{d}{=}v_{m+1}^{\theta}(t),$
where $B$ is a standard Brownian motion. Thus, since
$v_{m+1}^{\theta}(t)\geq\sup_{t\leq
s<\infty}\\{B_{m}(t,s)-h_{m+1}^{\theta}(t,s)\\}$, equality holds with
probability one for each fixed $t\in\mathbb{R}$. Let
$\Omega_{1}^{(\theta)}\subseteq\Omega^{(\theta)}$, be the event of full
probability on which (6.3) holds for all $t\in\mathbb{Q}$. Continuity of both
sides of (6.3) (Definition 6.2 and Lemma 3.4) extend the result to all
$t\in\mathbb{R}$ on $\Omega_{1}^{(\theta)}$. The equality for $h_{m}^{\theta}$
then follows by the definitions and additivity of the Busemann functions, as
shown below.
$\displaystyle D(h_{m+1}^{\theta},B_{m})(t)$
$\displaystyle=h_{m+1}^{\theta}(t)+Q(h_{m+1}^{\theta},B_{m})(0)-Q(h_{m+1}^{\theta},B_{m})(t)$
$\displaystyle=h_{m+1}^{\theta}(t)+v_{m+1}^{\theta}(0)-v_{m+1}^{\theta}(t)$
$\displaystyle=\;\mathcal{B}^{\theta}((m+1,0),(m+1,t))+\mathcal{B}^{\theta}((m,0),(m+1,0))-\mathcal{B}^{\theta}((m,t),(m+1,t))$
$\displaystyle=\mathcal{B}^{\theta}((m,0),(m,t))=h_{m}^{\theta}(t).\qed$
### 6.2. Global construction of the Busemann process
We now have the proper framework to define the global Busemann process of
Theorem 3.5. Fix a countable dense subset $D$ of $(0,\infty)$. Let
$\Omega_{0}=\bigcap_{\theta\in D}\Omega_{1}^{(\theta)},$
and on the event $\Omega_{0}$, define $\mathcal{B}^{\theta}$ for all
$\theta\in D$, as in Definition 6.2. Then, the conclusions of Lemma 6.3 hold
for all $\theta\in D$.
###### Lemma 6.4.
On $\Omega_{0}$, for each $\theta\in D,m\in\mathbb{Z}$, and $t\in\mathbb{R}$,
let $\tau_{(m,t)}^{\theta}$ denote the rightmost maximizer of
$B_{m}(s)-h_{m+1}^{\theta}(s)$ over $s\in[t,\infty)$. Such a maximizer exists
by continuity and Lemma 6.3(iv). Let $\Omega_{1}$ be the subset of
$\Omega_{0}$ on which, for each $\theta\in D,m,N\in\mathbb{Z}$, and
$t\in\mathbb{R}$, the following limits hold:
1. (i)
$\lim_{D\ni\gamma\rightarrow\theta}h_{m}^{\gamma}(N)=h_{m}^{\theta}(N)$
2. (ii)
$\lim_{D\ni\gamma\rightarrow\infty}h_{m}^{\gamma}(N)=B_{m}(N)$.
3. (iii)
$\lim_{D\ni\delta\searrow 0}\tau_{(m,t)}^{\delta}=t$.
4. (iv)
$\lim_{D\ni\gamma\rightarrow\infty}\tau_{(m,t)}^{\gamma}=\infty$.
Then, $\mathbb{P}(\Omega_{1})=1$.
###### Proof.
The almost sure uniqueness of maximizers follows from Lemma 6.3(vi) and the
$n=m$ case of Lemma 7.1. By Lemma 6.3(iii),
$h_{m}^{\gamma}(N)=h_{m}^{\gamma}(0,N)$ is monotone as
$D\ni\gamma\nearrow\theta$, and for $N>0$,
(6.6) $\lim_{D\ni\gamma\nearrow\theta}h_{m}^{\gamma}(N)\geq
h_{m}^{\theta}(N),$
For negative $N$, the inequality flips. By Theorem 2.3,
$h_{m}^{\gamma}(N)\sim\mathcal{N}\bigl{(}\frac{N}{\sqrt{\gamma}},|N|\bigr{)}$.
As $\gamma\nearrow\theta$, this converges in distribution to
$\mathcal{N}\bigl{(}\frac{N}{\sqrt{\theta}},|N|\bigr{)}$. Then, by (6.6),
$\lim_{D\ni\gamma\nearrow\theta}h_{m}^{\gamma}(N)=h_{m}^{\theta}(N)\text{ with
probability one}$. An analogous argument proves the almost sure convergence
for limits from the right and the convergence to $B_{m}(N)$.
Next, we show that, on $\Omega_{0}$, Parts (iii) and (iv) hold if and only if
they hold for all $t\in\mathbb{Q}$. Assume the statement holds for all
$t\in\mathbb{Q}$. By Lemmas 6.3(iii) and A.1, $\tau_{(m,t)}^{\delta}$ is
monotone as $\delta\searrow 0$, and so the limit exists. By definition of
$\tau_{(m,t)}^{\theta}$, we have the inequality
$\tau_{(m,s)}^{\theta}\leq\tau_{(m,t)}^{\theta}$ whenever $s<t$. Then, for any
$t\in\mathbb{R}$ and any $q_{1},q_{2}\in\mathbb{Q}$ with $q_{1}<t<q_{2}$,
$q_{1}=\lim_{D\ni\delta\searrow
0}\tau_{(m,q_{1})}^{\delta}\leq\lim_{D\ni\delta\searrow
0}\tau_{(m,t)}^{\delta}\leq\lim_{D\ni\delta\searrow
0}\tau_{(m,q_{2})}^{\delta}=q_{2}.$
Taking limits as $q_{1}\nearrow t$ and $q_{2}\searrow t$ shows the statement
for all $t\in\mathbb{R}$. The same argument can be applied to the limits as
$\gamma\rightarrow\infty$.
Lastly, we show that the limit in (iii) holds with probability one for fixed
$t\in\mathbb{R}$. By Lemma 6.3, Parts (v) and (vi),
$\bigl{\\{}B_{m}(s)-h_{m+1}^{\delta}(s):s\in\mathbb{R}\bigr{\\}}\overset{d}{=}\bigl{\\{}\sqrt{2}B(s)-\frac{s}{\sqrt{\delta}}:s\in\mathbb{R}\bigr{\\}},$
where $B$ is a standard, two-sided Brownian motion. Then, by Theorem B.3, for
$(m,t)\in\mathbb{Z}\times\mathbb{R}$ and $s\geq 0$,
$\mathbb{P}(\tau_{(m,t)}^{\delta}>s+t)=\left(2+\frac{t}{\delta}\right)\Phi\left(-\sqrt{\frac{t}{2\delta}}\right)-\sqrt{\frac{t}{\pi\delta}}\,e^{-\frac{t}{4\delta}}.$
Taking limits as $\delta\searrow 0$, it follows that $\tau_{(m,t)}^{\theta}$
converges weakly to the constant $t$. Since the limit exists almost surely by
monotonicity, the desired conclusion follows. A similar argument applies to
show $\lim_{\gamma\rightarrow\infty}\tau_{(m,t)}^{\gamma}=\infty$ with
probability one. ∎
On $\Omega_{1}$, we extend the definition of $\mathcal{B}^{\theta}$ to all
$\theta>0$. We proceed similarly as shown for the exponential corner growth
model in [Sep18]. For BLPP, there is an additional step that must be taken to
guarantee the convergence and continuity of Theorem 3.5, Parts (iii) and (iv).
###### Lemma 6.5.
On $\Omega_{1}$, for each $\theta\in D$, $m\in\mathbb{Z}$, and
$t\in\mathbb{R}$,
$\lim_{D\ni\gamma\rightarrow\theta}h_{m}^{\gamma}(t)=h_{m}^{\theta}(t)\qquad\text{and}\qquad\lim_{D\ni\gamma\rightarrow\theta}v_{m}^{\gamma}(t)=v_{m}^{\theta}(t).$
For each $m\in\mathbb{Z}$, the convergence is uniform in $t$ on compact
subsets of $\mathbb{R}$. Additionally, for each $m\in\mathbb{Z}$,
$\lim_{D\ni\gamma\rightarrow\infty}h_{m}^{\gamma}(t)=B_{m}(t),\qquad\text{and}\qquad\lim_{D\ni\delta\searrow
0}v_{m}^{\theta}(t)=0,$
uniformly in $t$ on compact subsets of $\mathbb{R}$.
###### Proof.
We first prove the statements for the $h_{m}$. We show that
$\lim_{D\ni\gamma\nearrow\theta}h_{m}^{\gamma}(t)$
exists and equals $h_{m}^{\theta}(t)$, uniformly in $t$ on compact subsets of
$\mathbb{R}$. The limits from the right (as well as the case of $B_{m}$ in
place of $h_{m}^{\theta}$) follow by analogous arguments.
By rearranging the inequality of Lemma 6.3(iii), for $0<\gamma<\theta<\infty$
and $a<b$ and any $t\in[a,b]$,
(6.7) $h_{m}^{\gamma}(a)-h_{m}^{\theta}(a)\leq
h_{m}^{\gamma}(t)-h_{m}^{\theta}(t)\leq h_{m}^{\gamma}(b)-h_{m}^{\theta}(b),$
and the inequality still holds if we replace $h_{m}^{\theta}$ with $B_{m}$.
Thus, on the event $\Omega_{1}$, $h_{m}^{\gamma}$ converges to
$h_{m}^{\theta}$, uniformly in $t$ on compact subsets of $\mathbb{R}$. Now, we
prove the convergence statements for $v_{m}$. By Lemma 6.3(vii), on
$\Omega_{0}$, for all $t\in\mathbb{R}$, and all $\theta\in D$.
(6.8) $v_{m}^{\theta}(t)=Q(h_{m}^{\theta},B_{m-1})(t)=\sup_{t\leq
s<\infty}\\{B_{m-1}(t,s)-h_{m}^{\theta}(t,s)\\}.$
Let $\omega\in\Omega_{1}$, and $t\in\mathbb{R}$. By the monotonicity of Lemma
6.3(iii) and Lemma A.1, since $\tau_{(m,t)}^{\theta+1}$ is a maximizer of
$B_{m-1}(s)-h_{m}^{\theta+1}(s)$ over $s\in[t,\infty)$, for all $\gamma\in D$
with $\gamma<\theta+1$,
$v_{m}^{\gamma}(t)=\sup_{t\leq
s<\infty}\\{B_{m-1}(t,s)-h_{m}^{\gamma}(t,s)\\}=\sup_{t\leq
s\leq\tau_{(m,t)}^{\theta+1}}\\{B_{m-1}(t,s)-h_{m}^{\gamma}(t,s)\\}.$
Since $h_{m}^{\gamma}$ converges uniformly on compact sets to
$h_{m}^{\theta}$, $v_{m}^{\gamma}(t)$ converges pointwise to
$v_{m}^{\theta}(t)$. By Lemma 6.3(iii), the convergence from both right and
left is monotone, so by the continuity of Definition 6.2 and Dini’s Theorem,
the convergence is uniform.
Lastly, for limits as $\delta\searrow 0$, we again apply Lemmas 6.3(iii) and
A.1 so that, for $\delta\in D$ with $\delta\leq 1$,
$\displaystyle v_{m}^{\delta}(t)$ $\displaystyle=\sup_{t\leq
s<\infty}\\{B_{m-1}(t,s)-h_{m}^{\delta}(t,s)\\}=\sup_{t\leq
s\leq\tau_{(m,t)}^{\delta}}\\{B_{m-1}(t,s)-h_{m}^{\delta}(t,s)\\}$
$\displaystyle\leq\sup_{t\leq
s\leq\tau_{(m,t)}^{\delta}}\\{B_{m-1}(t,s)-h_{m}^{1}(t,s)\\},$
and the right-hand side converges to $0$ as $\delta\searrow 0$ by the
continuity of $B_{m-1}-h_{m}^{1}$ and the convergence of
$\tau_{(m,t)}^{\delta}$ to $t$ given in the definition of $\Omega_{1}$. Dini’s
Theorem again strengthens the pointwise convergence to uniform convergence. ∎
By the additivity of Lemma 6.3(ii), for arbitrary
$(m,s),(r,t)\in\mathbb{Z}\times\mathbb{R}$ with $m\leq r$,
$\mathcal{B}^{\theta}((m,s),(r,t))=h_{m}^{\theta}(s,t)+\sum_{k=m}^{r-1}v_{k+1}(t).$
For $m>r$,
$\mathcal{B}^{\theta}((m,s),(r,t))=-\mathcal{B}^{\theta}((r,t),(m,s))$, so
$\mathcal{B}^{\theta}((m,s),(r,t))$ is still a sum of horizontal and vertical
increments. Then, by Lemma 6.5, on the event $\Omega_{1}$, for all $\theta\in
D$ and $\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$,
(6.9)
$\lim_{D\ni\gamma\rightarrow\theta}\mathcal{B}^{\gamma}(\mathbf{x},\mathbf{y})=\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y}).$
###### Definition 6.6.
On the event $\Omega_{1}$, for an arbitrary $\theta>0$ and
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$, define the following.
$\displaystyle\mathcal{B}^{\theta-}(\mathbf{x},\mathbf{y})=\lim_{D\ni\gamma\nearrow\theta}\mathcal{B}^{\gamma}(\mathbf{x},\mathbf{y}),\qquad\text{and}\qquad\mathcal{B}^{\theta+}(\mathbf{x},\mathbf{y})=\lim_{D\ni\delta\searrow\theta}\mathcal{B}^{\delta}(\mathbf{x},\mathbf{y}).$
###### Remark 6.7.
By additivity of Lemma 6.3(ii) and the monotonicity of Lemma 6.3(iii), these
limits exist for all $\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$ and
$\theta>0$. By (6.9), on $\Omega_{1}$, for all $\theta\in D$ and
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$,
$\mathcal{B}^{\theta+}(\mathbf{x},\mathbf{y})=\mathcal{B}^{\theta-}(\mathbf{x},\mathbf{y})=\mathcal{B}^{\theta}(\mathbf{x},\mathbf{y})$.
###### Proof of Theorem 3.5.
Parts (i)–(ii): These follow by taking limits in Parts (ii) and (iii) of Lemma
6.3.
Parts (iii)–(iv): By Part (ii), it suffices to take limits along the countable
dense set $D$. Then, Lemma 6.5 establishes Parts (iii)(c) and (iii)(d). The
monotonicity of Part (ii) can be rearranged, just as in Equation (6.7), which
strengthens the pointwise convergence of the
$h_{m}^{\gamma{\scaleobj{0.8}{\boxempty}}}$ to uniform convergence on compact
sets. Thus, since $h_{m}^{\gamma}$ is continuous for $\gamma\in D$,
$h_{m}^{\theta{\scaleobj{0.8}{\boxempty}}}$ is also continuous for
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$. Now, for the convergence of the
$v_{m}^{\theta{\scaleobj{0.8}{\boxempty}}}$, recall that on the event
$\Omega_{1}$, for all $t\in\mathbb{R}$ and $\gamma\in D$,
$v_{m}^{\gamma}(t)=Q(h_{m}^{\gamma},B_{m-1})(t)=\sup_{t\leq
s<\infty}\bigl{\\{}B_{m-1}(t,s)-h_{m}^{\gamma}(t,s)\bigr{\\}}.$
By Part (ii) and Lemma A.1, this supremum may be restricted to a common
compact set for all $\gamma<\theta$. Then, since $h_{m}^{\gamma}$ converges
uniformly to $h_{m}^{\theta-}$ as $D\ni\gamma\nearrow\theta$,
$v_{m}^{\gamma}(t)$ converges pointwise to $Q(h_{m}^{\theta-},B_{m-1})(t)$ as
$\gamma\nearrow\theta$. But, by definition,
$v_{m}^{\theta-}(t)=\lim_{D\ni\gamma\nearrow\theta}v_{m}^{\gamma}(t)$, so
$v_{m}^{\theta-}(t)=Q(h_{m}^{\theta-},B_{m-1})(t)$, which is continuous by
continuity of $h_{m}^{\theta-}$ and $B_{m-1}$. Since the convergence is
monotone, Dini’s Theorem implies the convergence is uniform on compact sets.
The proof for $\delta\searrow\theta$ is analogous. Part (iv) follows by
continuity of the Busemann functions from Definition 6.2, the uniform
convergence of Part (iii), and the additivity in Part (i).
Part (v): This follows from Lemma 6.3(iv) and the monotonicity of Part (ii).
Part (vi): The equality for $v_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}$ was
shown in the proof of Part (iii). The equality for $h_{m}^{\theta}$ follows by
the additivity of Part (i) and the same argument as in the proof of Lemma
6.3(vii).
Part (vii): This follows because
$\\{h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}:\theta>0,{\scaleobj{0.9}{\boxempty}}\in\\{+,-\\},r>m\\}$
is a function of $\\{B_{r}\\}_{r>m}$.
Part (viii): Fix $\theta>0$. We first show that, for each $\theta>0$ and
$\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{R}$, there exists an event
$\Omega_{\mathbf{x},\mathbf{y}}^{(\theta)}$ on which
(6.10)
$\mathcal{B}^{\theta-}(\mathbf{x},\mathbf{y})=\lim_{n\rightarrow\infty}\bigl{[}L_{\mathbf{x},(n,t_{n})}-L_{\mathbf{y},(n,t_{n})}\bigr{]}=\mathcal{B}^{\theta+}(\mathbf{x},\mathbf{y}),$
for all sequences $\\{t_{n}\\}$ with
$\lim_{n\rightarrow\infty}\frac{t_{n}}{n}=\theta$. By Theorem 2.3, for each
$\mathbf{x},\mathbf{y}$, and $\theta>0$, there exists an event of full
probability on which the limit in (6.10) exists and is independent of the
choice of sequence. Then, by additivity, it is sufficient to show that, for
each fixed $\theta>0,m\in\mathbb{Z}$, and $t\in\mathbb{R}$, with probability
one,
$\displaystyle
h_{m}^{\theta-}(t)=\lim_{n\rightarrow\infty}\bigl{[}L_{(m,0),(n,n\theta)}-L_{(m,t),(n,n\theta)}\bigr{]}=h_{m}^{\theta+}(t),\text{
and}$ $\displaystyle
v_{m}^{\theta-}(t)=\lim_{n\rightarrow\infty}\bigl{[}L_{(m-1,t),(n,n\theta)}-L_{(m,t),(n,n\theta)}\bigr{]}=v_{m}^{\theta+}(t).$
We show that
(6.11)
$h_{m}^{\theta-}(t)=\lim_{n\rightarrow\infty}\left[L_{(m,0),(n,n\theta)}-L_{(m,t),(n,n\theta)}\right]\qquad\text{a.s.},$
and the other statements follow by analogous arguments. By Theorem 2.3, the
right-hand side of (6.11) has distribution
$\mathcal{N}(\frac{t}{\sqrt{\theta}},|t|)$. By definition,
$h_{m}^{\theta-}(t)$ is the limit, as $\gamma\nearrow\theta$, of
$h_{m}^{\gamma}(t)$, and
$h_{m}^{\gamma}(t)\sim\mathcal{N}(\frac{t}{\sqrt{\gamma}},|t|)$, which
converges weakly to $\mathcal{N}(\frac{t}{\sqrt{\theta}},|t|)$. Using Theorem
6.3(i) and Lemma A.4, for each fixed $t>0$ and all $D\ni\gamma<\theta$,
$h_{m}^{\gamma}(t)=\lim_{n\rightarrow\infty}\left[L_{(m,0),(n,n\gamma)}-L_{(m,t),(n,n\gamma)}\right]\geq\lim_{n\rightarrow\infty}\left[L_{(m,0),(n,n\theta)}-L_{(m,t),(n,n\theta)}\right]\qquad\text{a.s.}$
The inequality flips for $t<0$. Hence, both sides in (6.11) have the same
distribution, while one dominates the other, so they are equal with
probability one.
Next, set
$\Omega^{(\theta)}=\bigcap_{\mathbf{x},\mathbf{y}\in\mathbb{Z}\times\mathbb{Q}}\Omega_{\mathbf{x},\mathbf{y}}^{(\theta)}.$
Then, using the continuity of Part (iv), the desired conclusion follows by the
same reasoning as in the proof of Lemma 6.3(i).
Part (ix) This follows from the construction of the Busemann functions as
limits of BLPP times and the shift invariance of BLPP. ∎
## 7\. Proofs of the results for semi-infinite geodesics
Throughout this section, $\Omega_{1}$ is the event defined in Lemma 6.4 and
referenced in Theorem 3.5.
### 7.1. Key lemmas
For two fields of Brownian motions, $\mathbf{B}$ and $\overline{\mathbf{B}}$,
$\lambda>0$, and an initial point $(m,s)\in\mathbb{Z}\times\mathbb{R}$, define
the point-to-line last passage time:
(7.1)
$\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}})=\sup\Bigl{\\{}\sum_{r=m}^{n}B_{r}(s_{k-1},s_{k})-\overline{B}_{n+1}(s_{n})-\lambda
s_{n}:\mathbf{s}_{m,n}\in\Pi_{(m,s),n}\Bigr{\\}},$
where $\Pi_{(m,s),n}$ denotes the set of sequences $s=s_{m-1}\leq
s_{m}\leq\cdots\leq s_{n}$. The following is due to Hammond [Ham19b]
###### Lemma 7.1 ([Ham19b], Lemma B.2).
Let $\mathbf{B}$ be a field of independent, two-sided Brownian motions and
$\overline{\mathbf{B}}$ be any other field of Brownian motions such that, for
each $n$, $\overline{B}_{n+1}$ is independent of $\\{B_{m}\\}_{m\leq n}$. Fix
an initial point $(m,s)$ and let $n\geq m$. Then, with probability one, the
quantity in (7.1) is finite, and there is a unique sequence
$\mathbf{s}_{m,n}\in\Pi_{(m,s),n}$ that is maximal for (7.1).
###### Remark 7.2.
Lemma 7.1 is stated slightly differently in [Ham19b]. In that paper, the
function $s\mapsto-\overline{B}_{n+1}(s)-\lambda s$ is replaced by an
arbitrary deterministic and measurable function
$h:\mathbb{R}\rightarrow\mathbb{R}\cup\\{-\infty\\}$ satisfying $h(t)>-\infty$
for some $t>s$ and $\limsup_{t\rightarrow\infty}h(t)/t<0$. The assumption that
$\\{\overline{B}_{n+1}\\}_{n+1}$ is independent of $\\{B_{m}\\}_{m\leq n}$
allows us to condition on $\overline{B}_{n+1}$ and obtain the desired result.
###### Lemma 7.3.
Let $\omega\in\Omega_{1}$, $(m,t)\in\mathbb{Z}\times\mathbb{R}$, $\theta>0$,
and ${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$. Then, the following hold.
1. (i)
Let $\\{\tau_{r}\\}_{r=m-1}^{\infty}$ be any sequence in
$\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$. Then, for each
$n\geq m$, the jump times $t=\tau_{m-1}\leq\tau_{m}\leq\cdots\leq\tau_{n}$ are
a maximizing sequence for
(7.2)
$\overline{L}_{(m,t),n}^{\lambda}(\mathbf{B},\overline{\mathbf{B}}):=\sup\Biggl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-h_{n+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n}):\mathbf{s}_{m,n}\in\Pi_{(m,t),n}\Biggr{\\}}.$
Here, $\overline{L}_{(m,t),n}^{\lambda}(\mathbf{B},\overline{\mathbf{B}})$ is
as in (7.1), where $\lambda=\frac{1}{\sqrt{\theta}}$,
$\mathbf{B}=\\{B_{r}\\}_{r\in\mathbb{Z}}$, and
$\overline{\mathbf{B}}=\\{h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(t)-\frac{t}{\sqrt{\theta}}:t\in\mathbb{R}\\}_{r\in\mathbb{Z}}$.
2. (ii)
Conversely, for each $n\geq m$, whenever $t=t_{m-1}\leq t_{m}\leq\cdots\leq
t_{n}$ is a maximizing sequence for (7.2), there exists
$\\{\tau_{r}\\}_{r=m-1}^{\infty}\in\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$
such that $t_{r}=\tau_{r}$ for $m\leq r\leq n$.
3. (iii)
For each $n\geq m$, the sequences
$t=\tau_{(m,t),m-1}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\cdots\leq\tau_{(m,t),n}^{\theta{\scaleobj{0.8}{\boxempty}},L}$
and
$t=\tau_{(m,t),m-1}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\cdots\leq\tau_{(m,t),n}^{\theta{\scaleobj{0.8}{\boxempty}},R}$
are, respectively, the leftmost and rightmost maximizing sequences for (7.2).
###### Proof.
Part (i): Recall by (4.1), that $\tau_{m}\geq t$ is a maximizer of
$B_{m}(t,s)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\qquad\text{ over
}\qquad s\in[t,\infty).$
Hence, the statement holds for $n=m$. Now, assume that the statement holds for
some $n\geq m$. Then, $t=\tau_{m-1}\leq\cdots\leq\tau_{n}$ satisfies
(7.3)
$\sum_{r=m}^{n}B_{r}(\tau_{r-1},\tau_{r})-h_{n+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{n})=\sup\Biggl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-h_{n+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n}):\mathbf{s}_{m,n}\in\Pi_{(m,t),n}\Biggr{\\}}.$
Using Theorem 3.5(vi) and rearranging the terms in the definition (3.4) of the
operator $D$,
(7.4)
$h_{n+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n})=B_{n+1}(s_{n})+\sup_{0\leq
s<\infty}\\{B_{n+1}(s)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}-\sup_{s_{n}\leq
s_{n+1}<\infty}\\{B_{n+1}(s_{n+1})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n+1})\\},$
for $s_{n}\in\mathbb{R}$. Specifically, since $\tau_{n+1}$ is a maximizer of
$B_{n+1}(s)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)$ over
$s\in[\tau_{n},\infty)$,
(7.5)
$h_{n+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{n})=-B_{n+1}(\tau_{n},\tau_{n+1})+h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{n+1})+\sup_{0\leq
s<\infty}\\{B_{n+1}(s)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}$
Substituting (7.4) and (7.5) into (7.3) and discarding the term $\sup_{0\leq
s<\infty}\\{B_{n+1}(s)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}$ on
both sides
$\displaystyle\qquad\sum_{r=m}^{n+1}B_{r}(\tau_{r-1},\tau_{r})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{n+1})$
$\displaystyle=$
$\displaystyle\sup\Biggl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-B_{n+1}(s_{n})+\sup_{s_{n}\leq
s_{n+1}<\infty}\\{B_{n+1}(s_{n+1})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n+1})\\}:\mathbf{s}_{m,n}\in\Pi_{(m,t),n}\Biggr{\\}}$
$\displaystyle=$
$\displaystyle\sup\Biggl{\\{}\sum_{r=m}^{n+1}B_{r}(s_{r-1},s_{r})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n+1}):\mathbf{s}_{m,n+1}\in\Pi_{(m,t),n+1}\Biggr{\\}}.$
Part (ii): We prove this part by induction. First, note that in the case
$n=m$, maximizers of
$B_{n}(t,s_{m})-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{m})$ over
$s_{m}\in[t,\infty)$ are precisely those that are the first jump times of the
Busemann semi-infinite geodesics. Now, assume that the statement holds for
$n$. We show that if $t=t_{m-1}\leq\ldots\leq t_{n+1}$ is a maximizing
sequence for (7.3) (with $n$ replaced by $n+1$), then $t_{n+1}$ is a maximizer
of $B_{n+1}(u)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)$ over
$u\in[t_{n},\infty)$, and $t=t_{m-1}\leq\cdots\leq t_{n}$ is a maximizing
sequence for (7.3).
With this procedure mapped out, observe that, as in the proof of Part (i),
$\displaystyle\sup\Biggl{\\{}\sum_{r=m}^{n+1}B_{r}(s_{r-1},s_{r})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n+1}):\mathbf{s}_{m,n+1}\in\Pi_{(m,t),n+1}\Biggr{\\}}$
$\displaystyle=$
$\displaystyle\sup\Biggl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-B_{n+1}(s_{n})+\sup_{s_{n}\leq
s_{n+1}<\infty}\\{B_{n+1}(s_{n+1})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n+1})\\}:\mathbf{s}_{m,n}\in\Pi_{(m,t),n}\Biggr{\\}}$
Hence, any maximizing sequence $t=t_{m-1}\leq\cdots\leq t_{n+1}$ must satisfy
$B_{n+1}(t_{n+1})-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(t_{n+1})=\sup_{t_{n}\leq
u<\infty}\\{B_{n+1}(u)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\\}.$
Furthermore, $t=t_{m-1}\leq\cdots\leq t_{n}$ is a maximizing sequence for
$\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-B_{n+1}(s_{n})+\sup_{s_{n}\leq
u<\infty}\\{B_{n+1}(u)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\\}$
over all sequences $\mathbf{s}_{m,n}\in\Pi_{(m,t),n}$ Subtracting off a
constant, $t=t_{m-1}\leq\cdots\leq t_{n}$ is also a maximizing sequence for
$\displaystyle\qquad\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-\Big{(}B_{n+1}(s_{n})+\sup_{0\leq
u<\infty}\\{B_{n+1}(u)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\\}-\sup_{s_{n}\leq
u<\infty}\\{B_{n+1}(u)-h_{n+2}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\\}\Big{)}$
$\displaystyle=\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-h_{n+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n}),$
where the last line comes by (7.4). This completes the inductive step.
Part (iii): This follows by Parts (i) and (ii) since
$t=\tau_{(m,t),m-1}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\tau_{(m,t),m}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\cdots$
and
$t=\tau_{(m,t),m-1}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\tau_{(m,t),m}^{\theta{\scaleobj{0.8}{\boxempty}},R}\leq\cdots$
are respectively the leftmost and rightmost sequences in
$\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$. ∎
###### Lemma 7.4.
Let $\omega\in\Omega_{1}$, $(m,t)\in\mathbb{Z}\times\mathbb{R},\theta>0$, and
${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, and $\\{\tau_{r}\\}_{r\geq
m-1}\in\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$. Then, for all
$r\geq m$,
$v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})=0,\qquad\text{and}\qquad
h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(u,v)=B_{r}(u,v)\text{ for all
}u,v\in[\tau_{r-1},\tau_{r}].$
Furthermore, the following identities hold for $r\geq m$.
(7.6)
$\displaystyle\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}=\inf\bigl{\\{}u\geq\tau^{\theta{\scaleobj{0.8}{\boxempty}},L}_{(m,t),r-1}:v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)=0\bigr{\\}}\text{
and }$ (7.7)
$\displaystyle\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r}=\sup\bigl{\\{}u\geq\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1}:h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1},u)=B_{r}(\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1},u)\bigr{\\}}$
More specifically, if
$u\geq\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1}$, then
$h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1},u)=B_{r}(\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1},u)$
if and only if $u\leq\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r}$.
###### Remark 7.5.
The first part of Lemma 7.4 says that, along any semi-infinite path in
$\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$, the Busemann process
agrees with the energy of the semi-infinite geodesic. That is, at every
vertical jump from $(r,\tau_{r})$ to $(r+1,\tau_{r})$, the vertical Busemann
function $v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}$ equals zero. This is as
it should because, according to (2.1), a vertical step of a path does not
collect any energy from the environment. Along each horizontal step from
$(r,\tau_{r-1})$ to $(r,\tau_{r})$, the increment of the horizontal Busemann
function $h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}$ agrees with the increment
of the Brownian motion $B_{r}$. Equations (7.6) and (7.7) are more subtle. Let
$s_{r-1}$ be the time when a $\theta{\scaleobj{0.8}{\boxempty}}$ geodesic
jumps from level $r-1$ to level $r$. Equation (7.6) says that the leftmost
$\theta{\scaleobj{0.8}{\boxempty}}$ geodesic jumps from level $r$ to $r+1$ at
the first time $u\geq s_{r-1}$ such that
$v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)=0$. Equation (7.7) says that
the rightmost geodesic jumps from level $r$ to $r+1$ at the last time $u\geq
s_{r-1}$ such that
$B_{r}(s_{r-1},u)=h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{r-1},u)$. At
all subsequent times $v>u$ the equality is lost, and we have
$B_{r}(s_{r-1},v)<h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{r-1},v)$.
###### Proof of Lemma 7.4.
By Theorem 3.5(vi) and the definition of $\tau_{r}$ as a maximizer,
$\displaystyle 0$ $\displaystyle\leq
v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})=\sup_{\tau_{r}\leq
s<\infty}\\{B_{r}(\tau_{r},s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r},s)\\}$
$\displaystyle=\sup_{\tau_{r}\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}-[B_{r}(\tau_{r})-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})]$
$\displaystyle\leq\sup_{\tau_{r-1}\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}-[B_{r}(\tau_{r})-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})]=0.$
To establish (7.6), assume, by way of contradiction, that for some
$\tau_{(m,t),r-1}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq
u<\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}$,
$0=v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)=\sup_{u\leq
s<\infty}\\{B_{r}(u,s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u,s)\\}.$
Then,
$B_{m}(u)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\geq
B_{m}(s)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)$
for all $s\geq u$, and specifically for
$s=\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}$. This contradicts the
definition of $\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}$ as the
leftmost maximizer of
$B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)$ over
$s\in[\tau_{(m,t),r-1}^{\theta{\scaleobj{0.8}{\boxempty}},L},\infty)$.
By Theorem 3.5(vi),
(7.8) $h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(u,v)=B_{r}(u,v)+\sup_{u\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}-\sup_{v\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\},$
so, since $\tau_{r}$ maximizes
$B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)$ on
$[\tau_{r-1},\infty)$, the two supremum terms above are both equal to
$B_{r}(\tau_{r})-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})$
whenever $u,v\in[\tau_{r-1},\tau_{r}]$. Therefore,
$h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(u,v)=B_{r}(u,v)$. Now we establish
(7.7). If, for some $r\geq m$, (7.7) fails, then by (7.8), for some
$u>\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},R}$,
$\sup_{\tau^{\theta{\scaleobj{0.8}{\boxempty}},R}_{(m,t),r-1}\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}=\sup_{u\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)\\}.$
Thus, all maximizers of
$B_{r}(s)-h_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s)$ over
$s\in[u,\infty)$ are also maximizers over the larger set
$[\tau_{(m,t),r-1}^{\theta{\scaleobj{0.8}{\boxempty}},R},\infty)$. But this is
a contradiction because $\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},R}$
is the rightmost maximizer over
$s\in[\tau_{(m,t),r-1}^{\theta{\scaleobj{0.8}{\boxempty}},R},\infty)$. ∎
Recall, by Remark 4.2, that on the event
$\Omega^{(\theta)}\subseteq\Omega_{1}$ of Theorem 3.5,
$\mathbf{T}_{\mathbf{x}}^{\theta+}=\mathbf{T}_{\mathbf{x}}^{\theta-}=\mathbf{T}_{\mathbf{x}}^{\theta}$
for all $\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$.
###### Lemma 7.6.
Fix $\theta>0$ and $\mathbf{x}=(m,t)\in\mathbb{Z}\times\mathbb{R}$. Then there
exists an event
$\widetilde{\Omega}_{\mathbf{x}}^{(\theta)}\subseteq\Omega^{(\theta)}$, of
probability one, on which the set $\mathbf{T}_{\mathbf{x}}^{\theta}$ contains
exactly one sequence $\\{\tau_{r}\\}_{r\geq m-1}$. This sequence satisfies
$\lim_{n\rightarrow\infty}\frac{\tau_{n}}{n}=\theta.$
To prove this lemma, we need some machinery from [ARAS20]. For two fields of
Brownian motions $\mathbf{B}$ and $\overline{\mathbf{B}}$ and a subset
$A\subset\mathbb{R}$, define
(7.9)
$\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}};s_{n}\in
A)=\sup\Bigl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-\overline{B}_{n+1}(s_{n})-\lambda
s_{n}:\mathbf{s}_{m,n}\in\Pi_{(m,s),n},\,s_{n}\in A\Bigr{\\}}.$
The only difference between this definition and (7.1) is the restriction on
the $s_{n}$.
###### Lemma 7.7 ([ARAS20], page 1949).
Let $\mathbf{B}$ be a field of independent, two-sided Brownian motions and
$\overline{\mathbf{B}}$ an arbitrary field of Brownian motions. Fix
$s\in\mathbb{R}$, $0\leq S\leq T\leq\infty$, and $m\in\mathbb{Z}$. Then, with
probability one,
$\lim_{n\rightarrow\infty}n^{-1}\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}};s+nS\leq
s_{n}\leq s+nT)=\sup_{S\leq t\leq T}\bigl{\\{}2\sqrt{t}-\lambda t\bigr{\\}}.$
###### Remark 7.8.
The appearance of the term $2\sqrt{t}$ in Lemma 7.7 comes from the shape
theorem for BLPP. Namely, the following limit holds with probability one.
$\lim_{n\rightarrow\infty}n^{-1}L_{(0,0),(n,nt)}=2\sqrt{t}.$
This almost sure convergence was first proved in [HMO02] (see also [MO07] for
an alternative proof). If $\overline{\mathbf{B}}_{n+1}$ is independent of
$\mathbf{B}_{m},\ldots,\mathbf{B}_{n}$, then the quantity
$\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}})$ is
distributed as the sum of $n-m+1$ independent exponential random variables
with rate $\lambda$ and has the interpretation as the sum of vertical
increments in the increment-stationary BLPP model. (see Appendix C of the
present paper and Section 4 of [OY01]). Then, in the case $S=0$ and
$T=\infty$, Lemma 7.7 degenerates to an application of the law of large
numbers, namely
$\lim_{n\rightarrow\infty}n^{-1}\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}})=\lambda^{-1}.$
###### Lemma 7.9 ([ARAS20], Lemma 4.7).
Let $\mathbf{B}$ and $\overline{\mathbf{B}}$ satisfy the same conditions of
Lemma 7.7. Fix $\lambda>0$, $s\in\mathbb{R}$, and $m\in\mathbb{Z}$. If
$0<\theta<\lambda^{-2}$, then there exist a nonrandom
$\varepsilon=\varepsilon(\lambda,\theta)>0$ such that, with probability one,
for all sufficiently large $n$,
$\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}};s_{n}\leq
s+n\theta)+n\varepsilon<\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}}).$
Similarly, if $\theta>\lambda^{-2}$, there exists a nonrandom
$\varepsilon=\varepsilon(\lambda,\theta)>0$ such that, with probability one,
for all sufficiently large $n$,
$\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}};s_{n}\geq
s+n\theta)+n\varepsilon<\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}})$
Lemma 7.9 is slightly stronger than the result stated in [ARAS20], so we
include a proof.
###### Proof.
We prove the first statement, and the second follows analogously. The unique
maximum of $2\sqrt{t}-\lambda t$ for $t\in[0,\infty)$ is achieved at
$t=\lambda^{-2}$. Then, by Lemma 7.7 and the assumption $\theta<\lambda^{-2}$,
$\lim_{n\rightarrow\infty}n^{-1}\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}};s_{n}\leq
s+n\theta)=\sup_{0\leq t\leq\theta}\bigl{\\{}2\sqrt{t}-\lambda
t\bigr{\\}}<\sup_{0\leq t<\infty}\bigl{\\{}2\sqrt{t}-\lambda
t\bigr{\\}}=\lim_{n\rightarrow\infty}n^{-1}\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}}).$
Hence, for sufficiently large $n$,
$\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}};s_{n}\leq
s+n\theta)+n\varepsilon<\overline{L}^{\lambda}_{(m,s),n}(\mathbf{B},\overline{\mathbf{B}}),$
where
$\varepsilon=\frac{1}{2}\Bigl{(}\;{\sup_{0\leq
t<\infty}\bigl{\\{}2\sqrt{t}-\lambda t\bigr{\\}}-\sup_{0\leq
t\leq\theta}\bigl{\\{}2\sqrt{t}-\lambda t\bigr{\\}}}\Bigr{)}.\qed$
###### Proof of Lemma 7.6.
By Lemma 7.3(i), for a sequence $\\{\tau_{r}\\}_{r\geq
m-1}\in\mathbf{T}_{(m,t)}^{\theta}$, the jump times
$t=\tau_{m-1}\leq\cdots\leq\tau_{n}$ are a maximizing sequence for the point-
to-line last passage time (7.2). By Theorem 3.5(vii), $h_{n+1}^{\theta}$ and
$\\{B_{r}\\}_{r\leq n}$ are independent for each $n$. This allows us to apply
the almost sure uniqueness of point-to-line last-passage maximizers in Lemma
7.1. Let $\widetilde{\Omega}_{\mathbf{x}}^{(\theta)}$ be the event on which
these maximizers are unique and on which $\tau_{n}/n\to\theta$ for the (now)
almost surely unique sequence $\\{\tau_{r}\\}_{r\geq
m-1}\in\mathbf{T}_{(m,t)}^{\theta}$. It remains to show that
$\mathbb{P}\bigl{(}\lim_{n\rightarrow\infty}\frac{\tau_{n}}{n}=\theta\bigr{)}=1.$
For $\gamma<\theta=\lambda^{-2}$, Lemma 7.9 guarantees that, with probability
one, for all sufficiently large $n$,
$\overline{L}^{\lambda}_{(m,t),n}(\mathbf{B},\overline{\mathbf{B}};s_{n}\leq
s+n\gamma)<\overline{L}^{\lambda}_{(m,t),n}(\mathbf{B},\overline{\mathbf{B}}).$
Therefore, by Lemma 7.3(i), $\tau_{n}>s+n\gamma$ for all sufficiently large
$n$. Thus, for $\gamma<\theta$,
$\mathbb{P}\bigl{(}\liminf_{n\rightarrow\infty}\frac{\tau_{n}}{n}\geq\gamma\bigr{)}=1,\qquad\text{so
by taking
}\gamma\nearrow\theta,\qquad\mathbb{P}\bigl{(}\liminf_{n\rightarrow\infty}\frac{\tau_{n}}{n}\geq\theta\bigr{)}=1.$
A symmetric argument using the second statement of Lemma 7.9 shows that
$\mathbb{P}\bigl{(}\limsup_{n\rightarrow\infty}\frac{\tau_{n}}{n}\leq\theta\bigr{)}=1.\qed$
### 7.2. Proofs of Theorems 4.3–4.7:
We now begin to prove the theorems of Section 4.2. First, we define the event
$\Omega_{2}$ used in the theorems. Let
$\widetilde{\Omega}_{\mathbf{x}}^{(\gamma)}$ be the events of Lemma 7.6. For
the countable dense set $D\subseteq(0,\infty)$ of Section 6, set
(7.10) $\Omega_{2}:=\bigcap_{\gamma\in
D,\mathbf{x}\in\mathbb{Z}\times\mathbb{Q}}\widetilde{\Omega}_{\mathbf{x}}^{(\gamma)}$
Then, $\mathbb{P}(\Omega_{2})=1$ and
$\Omega_{2}\subseteq\bigcap_{\theta\in\mathbb{Q}_{>0}}\Omega^{(\theta)}\subseteq\Omega_{1}$.
###### Proof of Theorem 4.3.
Part (i): On the event $\Omega_{2}$, let $(m,t)\in\mathbb{Z}\times\mathbb{R}$,
$\theta>0$ and ${\scaleobj{0.9}{\boxempty}}\in\\{+,-\\}$, and let
$t=\tau_{m-1}\leq\tau_{m}\leq\cdots$ be a sequence in
$\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$. Let $\Gamma$ be the
associated path. By Lemma 7.4,
(7.11)
$v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})=0,\qquad\text{and}\qquad
h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(s,t)=B_{r}(s,t)\text{ for
}\tau_{r-1}\leq s\leq t\leq\tau_{r}.$
We take $\mathbf{x}=\mathbf{y}=(m,t)$ and $\mathbf{z}=(n,\tau_{n})$. The case
for general $\mathbf{y}\leq\mathbf{z}$ along the path $\Gamma$ follows by the
same argument. By (7.11) and additivity of Busemann functions (Theorem
3.5(i)), the energy of path $\Gamma$ between $\mathbf{x}$ and $\mathbf{y}$ is
given by
$\displaystyle\sum_{r=m}^{n}\bigl{(}B_{r}(\tau_{r-1},\tau_{r})\bigr{)}=\sum_{r=m}^{n-1}\bigl{(}h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r-1},\tau_{r})+v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{r})\bigr{)}+h_{n}^{\theta{\scaleobj{0.8}{\boxempty}}}(\tau_{n-1},\tau_{n})=\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}((m,t),(n,\tau_{n})).$
Let jump times $t=s_{m-1}\leq s_{m}\leq s_{m+1}\leq\cdots\leq s_{n}=\tau_{n}$
define any other path between $(m,t)$ and $(n,\tau_{n})$. Then, by Theorem
3.5, Parts (i) and (ii), the energy of this path is
(7.12)
$\displaystyle\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})\leq\sum_{r=m}^{n-1}\bigl{(}h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{r-1},s_{r})+v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{r})\bigr{)}+h_{n}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{n-1},\tau_{n})=\mathcal{B}^{\theta{\scaleobj{0.8}{\boxempty}}}((m,t),(n,\tau_{n})).$
Part (ii): By Theorem 3.5(ii), equality holds in Equation (7.12) only if
$B_{r}(s_{r-1},s_{r})=h_{r}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{r-1},s_{r})$
for $m\leq r\leq n$ and $v_{r+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(s_{r})=0$
for $m\leq r\leq n-1$. Then, the statement follows by the Equations (7.6) and
(7.7) of Lemma 7.4.
Part (iii)(iii)(a): The key is Theorem 3.5(ii). We show that, for $r\geq m$,
$\tau_{r}^{-}:=\tau_{(m,t),r}^{\theta-,L}\leq\tau_{(m,t),r}^{\theta+,L}=:\tau_{r}^{+},$
and all other inequalities of the statement follow by the same procedure. By
definition, $\tau_{m-1}^{-}=\tau_{m-1}^{+}=t$. Inductively, assume that
$\tau_{r}^{-}\leq\tau_{r}^{+}$ for some $r\geq m$. We use the notation
$L\arg\sup$ to denote leftmost maximizer. Then,
$\displaystyle\tau_{r+1}^{-}$ $\displaystyle=L\arg\sup_{\tau_{r^{-}}\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta-,L}(s)\\}\leq L\arg\sup_{\tau_{r}^{+}\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta-,L}(s)\\}$ $\displaystyle\leq
L\arg\sup_{\tau_{r}^{+}\leq
s<\infty}\\{B_{r}(s)-h_{r+1}^{\theta+,L}(s)\\}=\tau_{r+1}^{+}.$
The first inequality above holds because $\tau_{r}^{-}\leq\tau_{r}^{+}$. The
second inequality is an application of Lemma A.1, using the fact that
$h_{m}^{\theta-}(s,t)\leq h_{m}^{\theta+}(s,t)$ for all $s\leq t$ (Theorem
3.5(ii)).
Part (iii)(iii)(b): Let $s<t$. We show that
$\tau_{s,r}:=\tau_{(m,s),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}\leq\tau_{(m,t),r}^{\theta{\scaleobj{0.8}{\boxempty}},L}=:\tau_{t,r},$
and the statement with ‘$L$’ replaced by ‘$R$’ has an analogous proof. Again,
the base case of $r=m-1$ follows by definition. Assume the inequality holds
for some $r\geq m-1$. Then,
$\tau_{s,r+1}=L\arg\sup_{\tau_{s,r}\leq
u<\infty}\\{B_{m}(u)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\\}\leq
L\arg\sup_{\tau_{t,r}\leq
u<\infty}\\{B_{m}(u)-h_{m+1}^{\theta{\scaleobj{0.8}{\boxempty}}}(u)\\}=\tau_{t,r+1}.$
Part (iii)(iii)(c): The proof of this item is postponed until the very end of
Section 7. This item is not used in any subsequent proofs.
Part (iv)(iv)(a): The monotonicity of Part (iii)(iii)(a) ensures that the
limits exist and that
(7.13)
$\tau_{r}:=\lim_{\gamma\nearrow\theta}\tau_{(m,t),r}^{\gamma{\scaleobj{0.8}{\boxempty}},L}\leq\tau_{(m,t),r}^{\theta-,L}.$
We prove equality in the above expression, and the other statements follow
analogously. By Lemma 7.3(i), for any $n\geq m$,
$t=\tau_{(m,t),m-1}^{\gamma{\scaleobj{0.8}{\boxempty}},L}\leq\ldots\leq\tau_{(m,t),n}^{\gamma{\scaleobj{0.8}{\boxempty}},L}$
is a maximizing sequence for
$\overline{L}_{(m,t),n}^{\frac{1}{\sqrt{\gamma}}}(\mathbf{B},\overline{\mathbf{B}}):=\sup\Bigl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-h_{n+1}^{\gamma{\scaleobj{0.8}{\boxempty}}}(s_{n}):\mathbf{s}_{m,n}\in\Pi_{(m,t),n}\Bigr{\\}}.$
Using the monotonicity of Part (iii)(iv)(a) again, for all $\gamma\leq\theta$,
the supremum may be restricted to the compact subset of $\Pi_{(m,t),n}$ such
that $s_{n}\leq\tau_{(m,t),n}^{(\theta+1)-,L}$. By Theorem 3.5(iii)(iii)(a),
as $\gamma\nearrow\theta$, $h_{n+1}^{\gamma{\scaleobj{0.8}{\boxempty}}}(s)$
converges to $h_{n+1}^{\theta-}(s)$, uniformly over
$s\in[t,\tau_{(m,t),n}^{(\theta+1)-,L}]$. Then, by Lemma A.2,
$t=\tau_{m-1}\leq\ldots\leq\tau_{n}$ is a maximizing sequence for
$\overline{L}_{(m,t),n}^{\frac{1}{\sqrt{\theta}}}(\mathbf{B},\overline{\mathbf{B}}):=\sup\Bigl{\\{}\sum_{r=m}^{n}B_{r}(s_{r-1},s_{r})-h_{n+1}^{\theta-}(s_{n}):\mathbf{s}_{m,n}\in\Pi_{(m,t),n}\Bigr{\\}}.$
Part (iii) of Lemma 7.3 then implies the inequality (7.13) must be an
equality. The statement for limits as $\delta\searrow\theta$ follows by the
same reasoning.
Part (iv)(iv)(b): The proof of this item is postponed until the very end of
Section 7. This item is not used in any subsequent proofs.
Part (iv)(iv)(c): This follows the same proof as that of Part (iv)(iv)(a),
replacing the use of Part (iii)(iii)(a) with Part (iii)(iii)(b) and replacing
the use of Lemma A.2 with Lemma A.3.
Part (v): Let $\omega\in\Omega_{2}$, $\theta>0$,
$(m,t)\in\mathbb{Z}\times\mathbb{Q}$, $(\tau_{r})_{r\geq
m-1}\in\mathbf{T}_{(m,t)}^{\theta{\scaleobj{0.8}{\boxempty}}}$,
$\varepsilon>0$, and let $\gamma,\delta\in D$ be such that
$\theta-\varepsilon<\gamma<\theta<\delta<\theta+\varepsilon$. Part
(iii)(iii)(a) and Lemma 7.6 imply that
$\limsup_{n\rightarrow\infty}\frac{\tau_{n}}{n}\leq\limsup_{n\rightarrow\infty}\frac{\tau_{(m,t),n}^{\theta+,R}}{n}\leq\limsup_{n\rightarrow\infty}\frac{\tau_{(m,t),n}^{\delta+,R}}{n}=\delta<\theta+\varepsilon.$
By a similar argument, on $\Omega_{2}$,
$\liminf_{n\rightarrow\infty}\frac{\tau_{n}}{n}\geq\theta-\varepsilon.$
An analogous application of Part (iii)(iii)(b) extends Theorem 4.3(v) to all
$\mathbf{x}\in\mathbb{Z}\times\mathbb{R}$, on the event $\Omega_{2}$. ∎
###### Proof of Theorem 4.5.
Part (i): Let $\omega\in\Omega_{2}$,
$\theta>0,(m,t)\in\mathbb{Z}\times\mathbb{R}$, and $\\{t_{n,r}\\}_{n\geq
m,m-1\leq r\leq n}$ satisfy the given assumptions. Let
$0<\gamma<\theta<\delta$ be arbitrary. Then, by assumption that
$t_{n}/n\rightarrow\theta$ and Theorem 4.3(v), there exists $N\in\mathbb{Z}$
such that for all $n\geq N$,
(7.14) $\tau_{(m,t),n}^{\gamma-,L}<t_{n}<\tau_{(m,t),n}^{\delta+,R}.$
By Theorem 4.3(ii), the sequence
$t=\tau_{(m,t),m-1}^{\gamma-,L}\leq\tau_{(m,t),m}^{\gamma-,L}\leq\cdots\leq\tau_{(m,t),n}^{\gamma-,L}$
defines the leftmost geodesic between $(m,t)$ and
$(n,\tau_{(m,t),n}^{\gamma-,L})$, and
$t=\tau_{(m,t),m-1}^{\delta+,R}\leq\tau_{(m,t),m}^{\delta+,R}\leq\cdots\tau_{(m,t),n}^{\delta+,R}$
defines the rightmost geodesic between $(m,t)$ and
$(n,\tau_{(m,t),n}^{\delta+,R})$. Let $t=t_{m-1}^{L}\leq t_{m}^{L}\leq\cdots
|
# Federated Multi-Agent DRL for Radio Resource Management in Industrial 6G
in-X subnetworks ††thanks: This work is partly supported by the HORIZON-JU-
SNS-2022-STREAM-B-01-02 CENTRIC project (grant agreement No. 101096379) and
the HORIZON-JU-SNS-2022-STREAM-B-01-03 6G-SHINE project (grant agreement No.
101095738).
Bjarke Madsen and Ramoni Adeogun
Department of Electronic Systems, Aalborg University, Denmark
<EMAIL_ADDRESS>
###### Abstract
Recently, 6G in-X subnetworks have been proposed as low-power short-range
radio cells to support localized extreme wireless connectivity inside entities
such as industrial robots, vehicles, and the human body. Deployment of in-X
subnetworks within these entities may result in rapid changes in interference
levels and thus, varying link quality. This paper investigates distributed
dynamic channel allocation to mitigate inter-subnetwork interference in dense
in-factory deployments of 6G in-X subnetworks. This paper introduces two new
techniques, Federated Multi-Agent Double Deep Q-Network (F-MADDQN) and
Federated Multi-Agent Deep Proximal Policy Optimization (F-MADPPO), for
channel allocation in 6G in-X subnetworks. These techniques are based on a
client-to-server horizontal federated reinforcement learning framework. The
methods require sharing only local model weights with a centralized gNB for
federated aggregation thereby preserving local data privacy and security.
Simulations were conducted using a practical indoor factory environment
proposed by 5G-ACIA and 3GPP models for in-factory environments. The results
showed that the proposed methods achieved slightly better performance than
baseline schemes with significantly reduced signaling overhead compared to the
baseline solutions. The schemes also showed better robustness and
generalization ability to changes in deployment densities and propagation
parameters.
## I Introduction
Wireless communication has evolved significantly, with advancements in 5G
technology enabling faster and more reliable communication. However, the need
for more efficient and reliable networks has led to the development of 6G
technology, which promises ultra-reliable communication, higher data rates,
and lower latency. Wireless communication has limitless applications,
including virtual and augmented reality in entertainment, education,
healthcare, and manufacturing, personalized healthcare solutions, and real-
time environmental monitoring. To achieve these benefits, new communication
technologies must be developed, such as 6G short-range low-power in-X
subnetworks [1], which can support demanding requirements inside entities like
robots, production modules, vehicles, and human-body scenarios. Ensuring data
privacy and security is crucial for advanced wireless communication networks,
especially for sensitive personal or environmental data. Techniques for
resource allocation must intelligently guarantee privacy and security while
adapting to changing wireless environments.
Related work: Interference management solutions for wireless systems have been
proposed in the literature, see e.g., [2]. There has also been a significant
amount of work targetting 6G in-X subnetworks within the last few years. These
solutions can be categorized into heuristics, machine learning (ML) methods,
and reinforcement learning (RL) techniques. Heuristic methods generate near-
optimal solutions based on simple rules or algorithms, providing good results
with low computational complexity. In [3], three heuristic-based algorithms
are presented: $\epsilon$-greedy channel selection, nearest neighbor conflict
avoidance, and minimum Signal to Interference plus Noise Ratio (SINR)
guarantee. Centralized Graph Coloring (CGC) is presented as a benchmark
algorithm, but it is impractical due to its high signaling and computation
overhead. A distributed interference-aware dynamic channel selection is
presented in [4]. In [5], channel selection is approached based on centralized
selective graph constructions. ML-based solutions may offer a higher quality
of service compared to heuristics by leveraging historical data and adapting
to evolving network conditions.
In [6], a Deep Neural Network (DNN) was successfully trained in offline
simulations using CGC with mobile subnetworks. Subsequently, this DNN was
deployed for real-time distributed channel selection. A novel solution for
centralized power control is presented in [7], where decisions are based on
positioning information using Graph Neural Network (GNN).
In [8], joint allocation of channel and transmit power based on a distributed
multi-objective optimization problem is addressed, and an approach of
Q-learning for multiple agents based on limited sensing information is
proposed. The GA-Net framework presented in [9] is a distributed framework
Multi-Agent Reinforcement Learning (MARL) resource management, based on GNN.
Contributions: This paper investigates the potential of federated MARL
techniques for efficient, privacy-preserving channel allocation in 6G in-X
subnetworks. The authors present their first study on federated MARL for radio
resource management (RRM) in short-range low-power 6G in-X subnetworks,
presenting the first contribution on FRL for RRM in 6G in-X subnetworks. They
propose two novel federated MARL solutions - F-MADDQN and F-MADPPO - for
dynamic channel allocation in 6G in-X subnetworks. These techniques enable
privacy-preserving collaborative training of a Double Deep Q-Network (DDQN)
policy and proximal Policy Optimization (PPO) agent for dynamic channel
selection by multiple subnetworks in the presence of an umbrella network. The
training phase requires only offline connectivity to the umbrella network. The
paper also presents performance evaluation in industrial factory environment
simulations based on a realistic factory floor plan from 5G-ACIA and 3GPP
channel models for in-factory environments.
## II System Model and Problem Formulation
System Model: We consider downlink transmissions in a wireless network
comprising $N$ independent and mobile subnetworks, each responsible for
serving one or more devices, such as sensors and actuators. Each subnetwork
contains an Access Point (AP) which is responsible for managing transmissions
with its associated devices. We denote the collection of subnetworks as
$\mathcal{N}={1,\cdots,N}$, and the set of devices in the $n^{\text{th}}$
subnetwork as $\mathcal{M}_{n}={1,\cdots,M_{n}}$. It is assumed that each
subnetwork’s AP is equipped with a local Resource Manager (RM), which utilizes
data acquired from wireless environment sensing or received from its devices.
Within each mobile subnetwork, wireless transmission occurs over one of $K$
shared orthogonal channels. Given the limited availability of resources, the
number of bands is typically much less than the number of co-existing
subnetworks. We assume that transmissions within each subnetwork are
orthogonal, hence there is no intra-subnetwork interference. In practical
systems, it may be impossible to make transmissions completely orthogonal due
to limitations on the bandwidth of the channel allocated to the subnetwork
leading to intra-subnetwork interference. Consideration of the effects of such
interference is however left for future work.
The received SINR, denoted $\gamma_{n,m}$, on a link between the
$n^{\text{th}}$ AP and $m^{\text{th}}$ device can be expressed
$\gamma_{n,m}(t)=\dfrac{g_{n,m}(t)}{\sum_{i\in\mathcal{I}_{k}}g_{n,i}(t)+\sigma^{2}},$
(1)
where $g_{n,m}(t)$ is the received power on the link between the $n$th AP and
device $m$, $\mathcal{I}_{k}$ denotes the APs or devices transmitting on
channel $k$ and $\sigma^{2}$ denotes the noise power and is defined as
$\sigma^{2}=10^{(-174+\text{NF}+10\log_{10}(B_{k}))/10}$, where NF denotes the
noise figure and $B_{k}$ is the bandwidth of the channel. We consider
practical 6G in-X subnetworks featuring short packets (in the order of tens of
bytes) transmission, making the infinite block length assumption and the
Shannon approximation used in most of the existing studies unrealistic. To
compute the capacity, we instead use the finite block-length approximation as
[10]
$r_{n,m}=B_{k}\left[\log_{2}\left(1+\gamma_{n,m}(t)\right)-\sqrt{\frac{V_{n,m}(t)}{l}}Q^{-1}(\epsilon)\log
e\right],$
where $Q$ is the complementary Gaussian cumulative distribution function based
on the code-word decoding error probability $\epsilon$, $\log e\approx 0.434$
is a constant constraint of the loss, and $V$ is the channel dispersion
defined as [10].
$V_{n,m}(t)=1-\dfrac{1}{\left(1+\gamma_{n,m}(t)\right)^{2}}.$ (2)
When the length of the code-word block $l$ tends toward infinity, the achieved
rate approaches the classical Shannon’s approximation.
Problem Formulation: We focus on a resource allocation problem that involves
the distributed selection of channels to maximize the achieved sum rate while
ensuring that devices within each subnetwork achieve a specified minimum rate.
This optimization problem can be mathematically expressed as
${\text{P}}:\left\\{{{{\max}_{\left\\{{{{\mathbf{c}}^{t}}}\right\\}}}\sum\limits_{m=1}^{M}{{r_{n,m}}}\left({{{\mathbf{c}}^{t}}}\right)}\right\\}_{n=1}^{N}{\text{
st: }}\>{r_{n,m}}\geq{r_{{\text{target }}}},\>\forall n,m$ (3)
Here, $\mathcal{K}=\left[1,\cdots,K\right];k\in\\{1,2,\cdots,K\\}$ represents
the vector of channel indices selected by all subnetworks at time t.
$r_{\text{target}}$ denotes the target minimum rate, assumed to be the same
for all subnetworks. The problem described in (3) involves the joint
optimization of $N$ conflicting non-convex objective functions, making it a
challenging problem to solve. In this paper, federated MARL methods are
proposed as a solution to this problem.
Figure 1: Illustration of the training procedures for (a) multi-agent DDQN and
(b) multi-agent PPO and the federated learning concept (c).
## III MARL FOR RRM IN SUBNETWORKS
### III-A MARL Algorithms
Resource selection in scenarios with multiple in-X subnetworks is cast as a
decentralized partially observable Markov decision process (DEC-POMDP). A DEC-
POMDP is formally represented as
$\left(N,\mathcal{S},\\{\mathcal{A}^{n}\\}_{n=1}^{N},\mathcal{P},\\{\mathcal{R}^{n}\\}_{n=1}^{N}\right)$.
The set of all possible states for all agents is defined as the state space
$\mathcal{S}=\mathcal{S}_{1}\times\cdots\times\mathcal{S}_{N}$, and the joint
action space containing all possible actions for the $n^{\text{th}}$ agent
$\mathcal{A}^{n}$ is denoted
$\boldsymbol{\mathcal{A}}=\mathcal{A}^{1}\times\cdots\times\mathcal{A}^{N}$.
The reward signal for the $n^{\text{th}}$ agent is denoted
$\mathcal{R}^{n}:\mathcal{S}\times\boldsymbol{\mathcal{A}}\times\mathcal{S}$
and the transition probability from state $s\in\mathcal{S}$ to state
$s^{\prime}\in\mathcal{S}$ by joint action
$\boldsymbol{a}\in\boldsymbol{\mathcal{A}}$ is denoted
$\mathcal{P}=\mathcal{S}\times\boldsymbol{\mathcal{A}}$.
#### III-A1 MADDQN
Multi-Agent Double Deep Q Network (MADDQN) [11, 12] extends the Double Deep Q
Network (DDQN) algorithm to enable decentralized learning among multiple
agents. MADDQN allows each agent to have its own local Q-network and learn
from its individual experiences. To train the Q-network, each agent interacts
with the environment and collects experiences by executing actions according
to its current policy. The experiences, including the state, action, reward,
and next state, are stored in either a shared replay buffer, $\mathcal{B}_{c}$
in case of centralized learning or individual reply buffers,
$\left\\{\mathcal{B}_{n}\right\\}_{n=1}^{N}$ in distributed learning.
To stabilize the learning process, MADDQN utilizes target networks. These
networks are copies of the local Q-networks and their weights are periodically
updated by polyak averaging, which involves updating the target network
weights with a fraction of the local network weights.
The Q-networks are optimized using mini-batches of experiences sampled from
the replay buffer with the goal of learning decentralized policies that
maximize the expected cumulative rewards for each agent. For each agent, the
Q-network is trained to minimize the mean squared Bellman error, i.e., the
squared difference between the estimated Q-value and the target Q-value. The
loss function is defined as [12]
$\mathcal{L}(\theta)=\frac{1}{N}\sum_{i=1}^{N}(y_{i}-Q(s_{i},a_{i};\theta_{i}))^{2}$
(4)
where $N$ is the mini-batch size, $s_{i}$ is the state, $a_{i}$ is the action,
and $Q(s_{i},a_{i};\theta_{i})$ is the estimated Q-value from the local
Q-network. The variable, $y_{i}$ is the target Q-value which is calculated by
using the target network to select the best action for the next state and then
evaluating it with the local Q-network. The target Q-value is typically
defined as
$y_{i}=r_{i}+\gamma
Q^{\prime}\left(s^{\prime},\max_{a}Q(s^{\prime},a^{\prime};\theta_{i});\theta_{i}^{\prime}\right)$
(5)
where $r_{i}$ denotes the reward received by agent $i$, $s^{\prime}$ is the
next state, $a^{\prime}$ is the action selected by the target Q-network, and
$Q^{\prime}$ is the target Q-network.
During training or execution, each agent selects actions using its local
Q-network based on the observed state using the well-known $\epsilon$-greedy
strategy. The actions can be either random with probability, $\epsilon$, to
encourage exploration, or greedy with probability, $1-\epsilon$, to exploit
the learned policies.
#### III-A2 MAPPO
Multi-agent Proximal Policy Optimization (MAPPO) is a state-of-the-art
algorithm designed to address challenges in multi-agent environments. It is an
extension of the Proximal Policy Optimization algorithm, originally proposed
for single-agent reinforcement learning. PPO uses an actor-critic
architecture, where each agent has its own policy and value function, allowing
them to interact and learn decentralizedly. The core idea is to update the
policy while ensuring deviation from the previous policy remains within a
certain range. This is achieved by maximizing a surrogate objective function,
which approximates the expected improvement of the policy using the current
state-action distribution and the ratio between new and old policies. This
balances exploration and exploitation, enabling stable learning in multi-agent
environments.
In MAPPO, each agent interacts with the environment by executing actions
according to its policy. Trajectories of state-action pairs are collected and
stored in a replay buffer for efficient sampling. The collected trajectories
are used to estimate the value function for each agent, typically using a
separate critic network to predict cumulative rewards from each state. The
surrogate objective is computed based on the collected trajectories and the
current policy, quantifying the expected improvement and determining the
policy update direction. Denoting the ratio of the new policy,
$\pi_{\theta^{\prime}}^{\prime}$ and the old policy, $\pi(\theta)$ as
$r(\theta)$, the surrogate objective is defined as [13]
$\mathcal{L}(\theta)=\mathbb{E}\left[\min\left(r(\theta^{\prime})A,\operatorname{clip}\left(r(\theta^{\prime},1-\eta,1+\eta\right)A\right)\right],$
(6)
where $A$ denotes the advantage function which provides a measure of the
advantage of taking a specific action under the current policy compared to the
estimated value function. The parameter $\eta$ is a hyperparameter that
controls the range of the policy deviation.
The policy is updated by optimizing the surrogate objective function using
stochastic gradient descent (SGD) or a similar optimization technique. The
objective is to maximize the surrogate objective with respect to the policy
parameters, $\theta$. This is typically done by taking multiple gradient steps
on the objective function, considering the trajectories collected from the
replay buffer.
The update of the parameters, $\theta$ of the policy can be represented by
$\displaystyle\theta^{\prime}$
$\displaystyle=\arg\max_{\theta^{\prime}}\left\\{\frac{1}{N}\sum_{m=1}^{N}\min\left(r(\theta^{\prime})A_{m},C(\theta^{\prime})A_{m}\right)\right\\},$
(7) $\displaystyle C(\theta^{\prime})$
$\displaystyle=\operatorname{clip}\left(r(\theta^{\prime},1-\eta,1+\eta\right),$
(8)
where $N$ denotes the number of trajectories and $A_{m}$ is the advantage of
the $m^{\text{th}}$ state-action pair.
The value function is updated to improve its estimation accuracy. This is
typically done by minimizing the Mean-Squared Error (MSE) between the
predicted value function and the actual cumulative rewards. The value function
update is performed using SGD or another optimization algorithm, and the
target is to minimize the loss function:
$\mathcal{L}(\theta_{v})=\frac{1}{N}\sum_{i=1}^{N}(V(s_{i};\theta_{v})-G_{i})^{2},$
(9)
where $\theta_{v}$ represents the value function parameters, $s_{i}$ is the
$i^{\text{th}}$ state, $V_{i}$ is the predicted value function for state
$s_{i}$, and $G_{i}$ is the actual cumulative reward obtained from the
$i^{\text{th}}$ trajectory.
The steps above are repeated iteratively to improve the policy and value
function over multiple iterations. The process continues until a desired level
of performance or convergence is achieved.
### III-B Proposed Methods
We propose two algorithms combining the MARL techniques, i.e., MADDQN and
MAPPO described above with federated learning [14, 15] to solve the resource
optimization problem in (3). The methods are referred to as Federated MADDQN
(F-MADDQN) and Federated MAPPO (F-MAPPO). The algorithms require the
definition of the environment, action space, state space, and reward function.
The considered environment is the wireless network with $N$ in-X subnetworks.
The other components of the proposed solutions are described below.
#### III-B1 Action Space
As stated earlier, we assume that the available frequency band is divided into
a set of $K$ equally sized channels. This leads to a $K$-dimensional action
space for each subnetwork. Denoting the $k^{\text{th}}$ channel as $b_{k}$,
the action space for the $n^{\text{th}}$ subnetwork, $\mathcal{A}^{n}$ is
defined as
$\mathcal{A}^{n}=\left\\{b_{1},\ldots,b_{K}\right\\}\quad\forall
n\in\mathcal{N}$ (10)
#### III-B2 Observation Space
The local observation of subnetwork $n$ at time $t$ is defined as
$\mathbf{S}_{n}(t)=\left[f(\mathbf{s}_{n}^{1}(t)),\cdots,f(\mathbf{s}_{n}^{K}(t))\right]^{T},$
(11)
where
$\mathbf{s}_{n}^{k}(t)=[\mathrm{SIR}_{n1}^{k}(t),\cdots,\mathrm{SIR}_{nM}^{k}(t)]\in\mathbb{R}^{M\times
1}$ denotes the vector of SIRs measured at all devices in the $n^{\text{th}}$
subnetwork on channel $k$ and $f$ is the Observation Reduction Function (ORF)
which is applied to the measurements to obtain the state to be used as input
to an RL agent. The choice of ORF affects the input to the RL model and hence
plays a significant role in the learning process. Considering the minimum rate
constraint in (3), it appears reasonable to optimize for the worst case, i.e.,
allocate channels based on the minimum SIR over each channel. However, the
combinatorial nature of the optimization problem makes it difficult to
conclude whether this is the optimal choice. We therefore propose to use two
categories of ORF defined as
$f(\mathbf{x}):\begin{cases}\mathbf{x}\mapsto\mathbf{x}\in\mathbb{R}^{M\times
1}&\text{Case I:Full state}\\\ \mathbf{x}\mapsto
f_{A}(\mathbf{x})\in\mathbb{R}^{1}&\text{Case II:Reduced state}\end{cases},$
(12)
where $f_{A}$ is the mean, max, or median aggregation functions.
#### III-B3 Reward Signal
To guide the learning of the agents toward achieving this objective in (3), we
define the reward function for the $n$ subnetwork as
$R_{n}=\lambda_{1}\sum_{m=1}^{M}r_{n,m}-\lambda_{2}\sum_{m=1}^{M}\mathds{1}(r_{n,m})(r_{\mathrm{min}}-r_{n,m})$
(13)
where $\lambda_{i};i=1,2$ are scaling factors that are selected to create a
balance between rate maximization and satisfaction of the minimum rate
constraint. The function, $\mathds{1}(r)$, is a binary indicator function with
a value equal to unity if and only if $r\geq r_{\mathrm{min}}$.
#### III-B4 Policy
The main component of any reinforcement learning solution is the policy that
defines how state measurements from the environment are mapped into actions.
Representation of this state-action mapping and the associated optimization
framework determines the policy. In this paper, we propose two methods viz:
F-MADDQN and F-MAPPO. As stated earlier F-MADDQN and F-MAPPO combines
federated learning with MADDQN and MAPPO, respectively. In both algorithms,
the policy is modeled using a DNN.
#### III-B5 Training
In both F-MADDQN and F-MAPPO, the agents are trained using federated learning
(FL) [16] as shown in Figure 1c. Consider $N$ agents
$\left\\{\mathcal{F}_{i}\right\\}_{i=1}^{N}$, each storing their respective
data-sets in experience replay buffer, $\mathcal{B}_{i}$ from which a set of
model parameters $\mathbf{\theta}_{i}$ is learned. Denoting the loss function
for the $i^{\text{th}}$ agent as $\mathcal{L}_{i}(\mathbf{\theta}_{i})$. the
common global model loss can be defined as $\mathcal{L}_{g}(\mathbf{\theta})$
[16].
$\mathcal{L}_{g}(\mathbf{\theta})=\sum_{i=1}^{N}\eta_{i}\mathcal{L}_{i}(\mathbf{\theta}),$
(14)
where $|\cdot|$ denotes the size of the set and $\eta_{i}>0$ is the relative
impact of each agent. Typically, the term $\eta$ is constrained to
$\sum_{i=1}^{N}\eta_{i}=1$. To allocate equal priority to all agents, we use
$\eta_{i}=1/N;\forall i\in\mathcal{N}$. The goal is to find the optimal
parameters $w^{*}$, which minimizes the global loss function [16].
$\mathbf{\theta}^{*}=\arg\min_{\mathbf{\theta}}\mathcal{L}_{g}(\mathbf{\theta})$
(15)
One solution for (15) would be a gradient-descent approach, known as the
federated averaging algorithm. Each agent uses its local data to perform a
number of steps in gradient descent on current model parameters
$\bar{\mathbf{\theta}}(t)$. This gradient descent step is defined as
$\mathbf{\theta}_{i}(t+1)=\bar{\mathbf{\theta}}(t)-\gamma\nabla\mathcal{L}_{i}(\bar{\mathbf{\theta}}_{i}(t))\quad\forall
i\in\mathcal{N},$ (16)
where $\gamma>0$ is the learning rate, and $\nabla f(\mathbf{\theta})$ for any
scalar expression $f(\mathbf{\theta})$ denotes the vector of partial
derivatives with respect to the components of the parameters
$\mathbf{\theta}$. During federated training, the global model is updated as
$\mathbf{\theta}_{g}(t+1)=\sum_{i=1}^{N}\dfrac{1}{N}\mathbf{\theta}_{i}(t+1)$
(17)
Once updated, the global weight is sent back to the agents at specified
intervals referred to as _aggregation interval_ , $\tau_{\mathrm{agg}}$.
TABLE I: Simulation parameters.
Parameter | | Value
---|---|---
Total factory area | $\mathcal{R}$ | $180~{}\mathrm{m}\times 80~{}\mathrm{m}$
Clutter type table | | Sparse
Number of subnetworks | $N$ | 20
Timestep | $t$ | 0.005 s
Number of episode | | 2000
Number of steps per episode | $T$ | 200
Subnetwork separation distance | $d_{\text{min}}$ | 1 m
Subnetwork radius | $d_{r}$ | 1 m
Subnetwork velocity | | 3 m/s
Transmit power | $p_{n}(t)$ | -10 dBm
Number of frequency channels | $K$ | 4
Carrier frequency | $f_{c}$ | 6 GHz
Bandwidth per subnetwork | $BW$ | 10 MHz
Noise figure | NF | 10 dB
Shadowing decorrelation distance | $d_{\delta}$ | 10 m
Max action switch delay | $\tau_{\text{max}}$ | 10
Figure 2: Illustration of the (a) Deep Q-Network (DQN) architecture for MADDQN
and (b) Deep PPO architecture for MAPPO which are used for the simulations.
## IV PERFORMANCE EVALUATION
### IV-A Simulation Settings
We consider an indoor industrial factory scenario inspired by existing
production facilities of manufacturing companies as identified by the industry
initiative, 5G alliance for connected industries, and automation [17]. Such a
layout is implemented as a $180~{}\mathrm{m}\times 80~{}\mathrm{m}$ hall
containing several separate areas for production, assembly, storage, and human
work zones. Multiple in-robot subnetworks can be deployed with the task of
transporting materials or tools around the facility. The alleys separating
laboring areas are 5 m wide taking up $\sim 1600$ m2 of the factory area and
are outlined as two-lane roads in a right-handed traffic setting. In the
deployment, the in-robot subnetworks are separated with a minimum distance of
$d_{\text{min}}=1$ m and move with a speed of 3 m/s. The in-robot subnetworks
are modeled by a circular coverage area with a radius of $d_{r}=1$ m, making
it possible for robots to pass each other in the alleys. Collisions are
avoided by prioritizing robots with the shortest distances to a common
intersection, slowing down any other robot that is within minimum separation
distance.
Except where otherwise stated, we consider a total bandwidth of
$100~{}\mathrm{MHz}$ which is partitioned into $K=4$ channels. Transmission
within each subnetwork is then performed over a single channel at each time
instant. The channel gain is calculated as
$g_{n,m}=h_{n,m}\times\sqrt{10^{(PL_{n,m}+X^{\mathrm{SD}}_{n,m})/10}}$, where
$h_{n,m}$ denotes the small scale channel gain which is modeled as a
temporally correlated Rayleigh random variable, $PL_{n,m}$ denotes the path
loss and $X^{\mathrm{SD}}_{n,m}$ is the shadow fading. We simulate the
pathloss using the 3GPP Indoor Factory (InF) channel model for the Dense-
clutter Low-antenna (InF-DL) and the Sparse-clutter Low-antenna (InF-SL)
scenarios. The shadow fading $X^{\mathrm{SD}}_{n,m}$ is generated using the
spatially correlated shadowing model used in [18]. Other simulation parameters
are given in Table I. We consider different values of the aggregation
intervals, i.e., $T^{\text{Agg}}=[128,256,512,1024]$ for F-MAPPO and F-MADDQN.
The minimum required data rate is set as, $r_{\rm{min}}=11~{}\mathrm{bps/Hz}$.
### IV-B Benchmarks
We benchmark the performance of the proposed F-MADDQN and F-MAPPO algorithms
with the following methods.
1\. Centralized Graph Coloring (CGC): The CGC algorithm utilizes improper
coloring to assign colors equivalent to channels for all subnetworks [19]. At
each timestep $t$, the pair-wise interference power relationships among
subnetworks, $\boldsymbol{I}(t)\in\mathbb{R}^{N\times N}$, are collected, and
mapped to a mutual coupling graph $G_{t}$. Each vertex corresponds to a
subnetwork, and edges are created by connecting each vertex to its $K-1$
nearest neighbors, where the weights of edges are equivalent to the
interference power between the connected subnetworks.
2\. Greedy channel selection: At each switching instant, each subnetwork
selects the channel with the highest measured SINR from the previous time
step.
3\. Random channel selection: At the beginning of every episode, a random
channel is allocated to each subnetwork.
In addition, the federated learning solutions are also compared with previous
solutions based on centralized as well as distributed training. We denote the
centralized (distributed) MADDQN and MAPPO as C-MADDQN (D-MADDQN) and C-MAPPO
(D-MAPPO), respectively.
(a) MADDQN.
(b) MAPPO.
Figure 3: Averaged reward versus number of episodes. Figure 4: CDF of achieved
rate. Figure 5: Sensitivity evaluation. Figure 6: Rate versus number of
subnetworks.
### IV-C Training and Convergence
We consider the network architecture in Figure 2a and Figure 2b for F-MADDQN
and F-MAPPO, respectively. The networks are trained using the federated
learning procedure illustrated in Figure 1c. Figure 3a shows the averaged
reward signal achieved at each training episode by the MADDQN-based methods.
At convergence approximately after 1500 episodes, the MARL methods approach
the distributed greedy selection baseline. At convergence, the F-MADDQN
methods with greater aggregation intervals achieve a marginally better
performance than the centralized and distributed baselines, indicating a
potential advantage of the federated learning framework. The figure also shows
that too frequent aggregation of DQN weights due to a low value of aggregation
interval, $\tau_{\mathrm{agg}}$ can result in a comparatively slower
convergence rate. For example, while F-MADDQN with $\tau_{\mathrm{agg}}\geq
256$ converged at $\approx 1500$ episodes, the algorithm failed to converge
after $2000$ episodes with $\tau_{\mathrm{agg}}=128$.
We evaluate the convergence performance of F-MAPPO with the in Figure 3b.At
convergence, achieved approximately after $\approx 750$ episodes, the F-MAPPO
achieves similar performance to the benchmark for all values of the
aggregation interval, $\tau_{\mathrm{agg}}$. Compared to the methods based on
MADDQN, the MAPPO-based algorithms give a reduction of about $50\%$ in the
number of episodes required for convergence. A plausible explanation for this
is the effect of the decay rate of the $\epsilon$-greedy parameter of MADDQN
and the iterative training over multiple epochs at each training step of
MAPPO.
### IV-D Performance Comparison
The trained models are deployed in each in-robot subnetwork for distributed
channel allocation, and the performance is compared with random, greedy, CGC,
and centralized and distributed training frameworks. To evaluate the trained
models, the aggregation interval is set to $T^{\text{Agg}}=512$. Figure 4
shows the CDF of the achieved rate per device. The proposed F-MADDQN performs
marginally better than D-MADDQN and C-MADDQN below the $30^{\text{th}}$
percentile. Furthermore, the performance of F-MADDQN is also better than that
of the distributed greedy scheme. F-MAPPO similarly achieves marginally better
performance than the distributed baseline and approaches the centralized
baseline below the $30^{\text{th}}$ percentile. As expected, the centralized
graph coloring baseline shows significant performance superiority to all other
schemes which are based on distributed execution. This difference is expected
since CGC exploits the global information about the scenario, unlike the
MADDQN, MAPPO, and greedy schemes which rely solely on local measurements of
only the aggregate interference power. It should, however, be noted that the
CGC method has significantly higher sensing, signalling, and computational
complexity. Such high complexity makes CGC impractical for dense deployments
of subnetworks.
### IV-E Sensitivity and Robustness Evaluation
We evaluate the ability of the proposed methods to generalize to changes in
the deployment density and the wireless environment parameters.
#### IV-E1 Robustness to changes in deployment density
The models trained in the factory with $N=20$ subnetworks are deployed in
scenarios with different numbers of subnetworks. We vary the number of
subnetworks in the test scenarios between $N=10$ and $N=50$ and evaluate the
average per-device rate. As shown in Figure 6, when the number of subnetworks
increases, an overall decrease in performance with similar proportions across
all methods is observed. While the methods based on MADDQN appear to be robust
to the changes in deployment density for all values of $N$, there seems to be
a degradation in the performance of the methods based on MAPPO. This shows
that the proposed F-MADDQN is more robust to changes in deployment density
than the F-MAPPO.
#### IV-E2 Sensitivity to changes in environment parameters
To study the ability of the proposed methods to generalize to environments
with different parameters than those used for the training, we train the
models in a factory environment with sparse clutter with an average clutter
element size of $d_{\text{clutter}}=10$ m and a density of
$r_{\text{clutter}}=20\%$. The models are then introduced to environments with
different types of clutter. We consider sparse and dense clutter cases with
varying clutter density as defined in the 3GPP model for in-factory
environment [20]. Figure 5 shows the minimum, maximum, and average values of
the achieved rate per device for the different cases using the proposed
schemes and the baseline algorithms. The figure shows that the RL-based
methods can maintain their performance in comparison to the centralized and
distributed baselines. Furthermore, similar to the observation from the
sensitivity test with varying numbers of subnetworks, the MADDQN-based
solutions appear to be slightly more robust to changes in the environment and
maintain performance close to the greedy selection baseline.
## V CONCLUSION
We proposed two federated reinforcement learning-based schemes named Federated
Multi-Agent Double DQN (F-MADDQN) and Federated Multi-Agent Deep Proximal
Policy Optimization (F-MADPPO) for dynamic channel allocation in 6G in-X
subnetworks which overcome the inherent issues of convergence and high
signalling overhead with privacy challenges associated with distributed and
centralized multi-agent reinforcement learning techniques. Our performance
evaluation results using realistic in-factory models defined by 5G-ACIA and
3GPP have shown that the proposed schemes can achieve similar performance to
the best-performing baselines and are robust to changes in the deployment
density as well as wireless environment conditions.
## References
* [1] R. Adeogun, G. Berardinelli, P. E. Mogensen, I. Rodriguez, and M. Razzaghpour, “Towards 6g in-x subnetworks with sub-millisecond communication cycles and extreme reliability,” _IEEE Access_ , vol. 8, pp. 110 172–110 188, 2020.
* [2] H. Sun, X. Chen, Q. Shi, M. Hong, X. Fu, and N. D. Sidiropoulos, “Learning to optimize: Training deep neural networks for interference management,” _IEEE Trans on Signal Proc._ , vol. 66, no. 20, pp. 5438–5453, 2018.
* [3] R. Adeogun, G. Berardinelli, I. Rodriguez, and P. Mogensen, “Distributed dynamic channel allocation in 6g in-x subnetworks for industrial automation,” in _2020 IEEE Globecom Workshops (GC Wkshps_ , 2020, pp. 1–6.
* [4] R. Adeogun, G. Berardinelli, and P. E. Mogensen, “Enhanced interference management for 6g in-x subnetworks,” _IEEE Access_ , vol. 10, pp. 45 784–45 798, 2022.
* [5] G. Berardinelli and R. Adeogun, “Spectrum assignment for industrial radio cells based on selective subgraph constructions,” in _2021 IEEE 94th Vehicular Technology Conference (VTC2021-Fall)_ , 2021, pp. 01–05.
* [6] R. O. Adeogun, G. Berardinelli, and P. E. Mogensen, “Learning to dynamically allocate radio resources in mobile 6g in-x subnetworks,” 2021.
* [7] D. Abode, R. Adeogun, and G. Berardinelli, “Power control for 6g industrial wireless subnetworks: A graph neural network approach,” in _2023 IEEE WCNC_. IEEE, 2023, pp. 1–6.
* [8] R. Adeogun and G. Berardinelli, “Multi-agent dynamic resource allocation in 6g in-x subnetworks with limited sensing information,” _Sensors_ , vol. 22, no. 13, p. 5062, 2022.
* [9] X. Du, T. Wang, Q. Feng, C. Ye, T. Tao, L. Wang, Y. Shi, and M. Chen, “Multi-agent reinforcement learning for dynamic resource management in 6g in-x subnetworks,” _IEEE transactions on wireless communications_ , vol. 22, no. 3, pp. 1–1, 2023.
* [10] B. Lu, H. Zhang, T. Xue, S. Guo, and H. Gai, “Deep reinforcement learning-based power allocation for ultra reliable low latency communications in vehicular networks,” in _2021 IEEE/CIC International Conference on Communications in China (ICCC)_ , 2021, pp. 1149–1154.
* [11] W. Du and S. Ding, “A survey on multi-agent deep reinforcement learning: from the perspective of challenges and applications,” _Artificial Intelligence Review_ , vol. 54, pp. 3215–3238, 2021.
* [12] A. Feriani and E. Hossain, “Single and multi-agent deep reinforcement learning for ai-enabled wireless networks: A tutorial,” _IEEE Comm. Surveys & Tutorials_, vol. 23, no. 2, pp. 1226–1252, 2021.
* [13] J. G. Kuba, R. Chen, M. Wen, Y. Wen, F. Sun, J. Wang, and Y. Yang, “Trust region policy optimisation in multi-agent reinforcement learning,” _arXiv preprint arXiv:2109.11251_ , 2021.
* [14] L. Collins, H. Hassani, A. Mokhtari, and S. Shakkottai, “Fedavg with fine tuning: Local updates lead to representation learning,” 2022.
* [15] A. B. Mansour, G. Carenini, A. Duplessis, and D. Naccache, “Federated learning aggregation: New robust algorithms with guarantees,” 2022.
* [16] J. Qi, Q. Zhou, L. Lei, and K. Zheng, “Federated reinforcement learning: Techniques, applications, and open challenges,” _arXiv preprint arXiv:2108.11887_ , 2021.
* [17] “5G-ACIA - 5G Alliance for Connected Industries and Automation — 5g-acia.org,” http://www.5g-acia.org/, [Accessed 16-May-2023].
* [18] R. Adeogun and G. Berardinelli, “Multi-agent dynamic resource allocation in 6g in-x subnetworks with limited sensing information,” _Sensors_ , vol. 22, no. 13, p. 5062, 2022.
* [19] R. O. Adeogun, G. Berardinelli, and P. E. Mogensen, “Learning to dynamically allocate radio resources in mobile 6g in-x subnetworks,” 2021.
* [20] 3GPP, “Study on channel model for frequencies from 0.5 to 100 GHz,” 3rd Generation Partnership Project (3GPP), Technical Report (TR) 38.901, 01 2017, version 16.1.0 Release 16.
|
# Spontaneous symmetry breaking or non minimal coupling?
M. Novello and A.E.S. Hartmann Centro de Estudos Avançados de Cosmologia/CBPF
Rua Dr. Xavier Sigaud 150, Urca 22290-180 Rio de Janeiro, RJ-Brazil
###### Abstract
We show that the combined minimal and non minimal interaction with the
gravitational field may produce the generation of a cosmological constant
without self-interaction of the scalar field. In the same vein we analyze the
existence of states of a scalar field that by a combined interaction of
minimal and non minimal coupling with the gravitational field can exhibit an
unexpected property, to wit, they are acted on by the gravitational field but
do not generate gravitational field. In other words, states that seems to
violate the action-reaction principle. We present explicit examples of this
situation in the framework of a spatially isotropic and homogeneous universe.
## I Introduction
The distribution of the energy of a scalar field dependent only on time in a
space time endowed with a spatially isotropic and homogeneous metric may be
represented as a perfect fluid even in the case of a non minimal coupling to
gravity. This allows a very unexpected result to appear, that is the
possibility of the scalar field to be acted upon by the gravitational field
that does not drive the gravitational field, as a sort of violation of the
action-reaction principle,in the realm of General Relativity theory. We will
show here how this is possible.
From the quantum theory of fields the mechanism of spontaneous symmetry
breaking set in evidence a sort of dynamical origin of a cosmological
constant. In a sense, this opened the way to allow the proposed scenario of
inflation guth , starobinski , novellosalim . The idea is very simple. Start
with a scalar field the dynamics of which, besides its kinematical term is
provided by a potential $V(\varphi)$ that is a polynomial. A typical case,
transformed in a paradigm, is $V=c_{1}\,\varphi^{2}+c_{2}\,\varphi^{4}$ for a
convenient choice of constants $c_{1}$ and $c_{2}.$ The density of energy is
associated to the constant $\varphi_{0}$ that extremizes $V.$ A further
restriction on the constants makes this extremum a stable one for the case
$\varphi_{0}.$ Then this constant field shows a non trivial energy
distribution having the appearance of a cosmological constant.
In the present paper we show an alternative of associating the energy
distribution of the scalar field (without self-interaction) to a cosmological
constant related to a combined interaction of gravity to the scalar field.
## II Dynamics: Minimal coupling
We start by making a short resume of properties that will be used in our text.
In the framework of standard cosmology the geometry of the universe is given
by a spatially isotropic and homogeneous metric with Euclidean section
$ds^{2}=dt^{2}-a^{2}(t)\,(dx^{2}+dy^{2}+dz^{2})$ (1)
In this case
$R_{00}=\dot{\theta}+\frac{1}{3}\,\theta^{2},$
$R_{ij}=\frac{1}{3}\,(\dot{\theta}+\,\theta^{2})\,g_{ij}$
$R=2\,\dot{\theta}+\frac{4}{3}\,\,\theta^{2}$
where we have defined the expansion factor $\theta=3\,\dot{a}/a$ and a dot
means time derivative.
Let us first examine the simplest case of the Lagrangian of a massless scalar
field
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi-\frac{\Lambda}{\kappa}$
(2)
The equation of motion of the scalar field and the metric are given by:
$\Box\,\varphi=0$ (3)
$R_{\mu\nu}-\frac{1}{2}\,R\,g_{\mu\nu}=-\,\kappa\,T_{\mu\nu}^{0}-\Lambda\,g_{\mu\nu}$
(4)
where (note that we use indiscriminately
$\varphi_{\mu}\equiv\varphi_{,\,\mu}$)
$T_{\mu\nu}^{0}=\varphi_{,\,\mu}\,\varphi_{,\,\nu}-\frac{1}{2}\,\varphi_{\alpha}\,\varphi^{\alpha}\,g_{\mu\nu}.$
(5)
The unique non trivial equations reduces to this case to the conservation of
energy and the evolution of the expansion factor, that is
$\dot{\rho}+(\rho+p)\,\theta=0.$ (6)
$\rho+\Lambda=\frac{1}{3}\,\theta^{2}$ (7)
where, in our case
$\rho=p=\frac{1}{2}\,\dot{\varphi}^{2}.$
It then follows that the density of energy
$\rho=\rho_{0}\,a^{-6}.$
Thus, for very small values of $a(t)$ the cosmological constant can be
completely neglected. In fact, comparison of the energy of the scalar field to
others forms of matter, like the energy of an electromagnetic fluid for
instance, where the density of the photons is provided by
$\rho_{\gamma}=constant\,a^{-4},$ shows that such scalar field dominates the
cosmical scenario in the primordial era where the scalar function $a(t)$ is
very small.
In this case the solution of these equations are immediate
$\varphi=\left(\frac{2}{3}\right)^{1/2}\,\ln t$ $a(t)=a_{0}\,t^{1/3}.$
Let us now introduce a non minimal coupling and examine its consequences. We
shall see that in certain cases, the combined coupling minimal plus non
minimal can supress totally the effect of the scalar field on the geometry.
## III Non minimal coupling: suppressing the energy
Let us set the Lagrangian
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi+B(\varphi)\,R-\frac{1}{\kappa}\,\Lambda$
(8)
where $B(\varphi)$ is for the time being an arbitrary function of the scalar
field. Then, it yields the equation for the metric as
$R_{\mu\nu}-\frac{1}{2}\,R\,g_{\mu\nu}=-\,T_{\mu\nu}^{0}-\,Z_{\mu\nu}-\Lambda\,g_{\mu\nu}$
(9)
where $T_{\mu\nu}^{0}$ is given by (5) and
$Z_{\mu\nu}=2B\,(R_{\mu\nu}-\frac{1}{2}\,R\,g_{\mu\nu})+2\\!\mathrel{\raisebox{1.49994pt}{\reflectbox{\rotatebox[origin={c}]{180.0}{$\triangle$}}}}\\!\\!_{\mu}\\!\mathrel{\raisebox{1.49994pt}{\reflectbox{\rotatebox[origin={c}]{180.0}{$\triangle$}}}}\\!\\!_{\nu}B-2\Box
B\,g_{\mu\nu}$ (10)
In the case the scalar field depends only on the cosmological time we obtain
for the density of energy the forms
$p_{0}=\rho_{0}=\frac{1}{2}\,\dot{\varphi}^{2}$
and for the non minimal part
$\rho_{Z}=-\,\frac{2}{3}\,B\,\theta^{2}-2\,\dot{B}\,\theta.$
$p_{Z}=2\,\ddot{B}+\frac{4}{3}\,\dot{B}\,\theta+\,\frac{2}{3}\,B\,\theta^{2}+\frac{4}{3}\,B\,\dot{\theta}$
We are looking for solution of the scalar field such that its total energy
$\rho=\rho_{0}+\rho_{Z}$ and pressure $p=p_{0}+p_{Z}$ vanishes, although its
field satisfies the equation of motion of the curved space-time. The
responsible for the curvature is the cosmological constant that generates a
constant expansion factor $\theta_{0}.$ Note that this situation can be
generalized for other curved space-times.
We start by solving the equation of motion for $\varphi.$ The equation is
$\Box\,\varphi-R\,B^{{}^{\prime}}=0$ (11)
where $B^{{}^{\prime}}$ represents derivative of $B$ with respect to
$\varphi.$ Let us choose, just to present an specific case the non minimal
term as given by
$B=b\,\varphi^{2}$
and the equation of motion (in the case the field depends only on time)
reduces to
$\ddot{\varphi}+\theta_{0}\,\dot{\varphi}-b\,\frac{8}{3}\,\theta^{2}_{0}\,\varphi=0.$
(12)
We set a solution of the form
$\varphi=\varphi_{0}\,e^{m\,t}$
The equation of motion is satisfied if $m$ and $b$ are related by
$m^{2}+\theta_{0}\,m-\frac{8}{3}\,b\,\theta_{0}^{2}=0.$ (13)
Using this proposal into the two independent conditions
$\rho_{0}+\rho_{Z}=0$ $p_{0}+p_{Z}=0$
yields, respectively, the equations:
$m^{2}-\frac{4}{3}b\,\theta_{0}^{2}-8\,b\,m\,\theta_{0}=0.$ (14)
$m^{2}+16\,b\,m^{2}+\frac{4}{3}b\,\theta_{0}^{2}+\frac{16}{3}\,b\,m\,\theta_{0}=0.$
(15)
The solution of these equations gives the value
$m=-\,\sqrt{3\,\Lambda}/3=-\,1/3\,\theta_{0},$ and the value of the constant
$b=-\,1/12.$ This means that the non minimal term is nothing but a conformal
coupling.
We can then state the Lemma: _The combined coupling of a minimal and conformal
coupling of a massless scalar field to gravity can annihilate the effect of
the scalar field on gravity in an universe of constant expansion._
### III.1 The case of a variable expansion factor
The suppressed process of the energy by non minimal gravitational interaction
can be generalized for the case the expansion factor is not a constant. We set
the Lagrangian
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi+b\,\varphi^{2}\,R+V(\varphi)+L_{m}$
(16)
and consider the matter term as generating an expanding universe with
expansion factor $\theta=\theta_{0}/t.$
The equation of motion of the scalar field is
$\Box\varphi-2\,b\,\varphi\,R-\frac{\delta V}{\delta\varphi}=0$
Let us consider the case in which $V=\lambda\,\varphi^{4}.$ Then, a solution
is given by
$\varphi=\frac{\varphi_{0}}{t}.$
In the case of a conformal interaction $b=-1/12$ it follows that
$\lambda\,\varphi_{0}^{2}=\frac{1}{2}\,(1-\frac{\theta_{0}}{3})^{2}.$
It is then immediate to show that for the density and pressure of the scalar
field we obtain
$\rho(\varphi)=p(\varphi)=0.$
That is, the non minimal coupling term annihilates the free part of the
energy-momentum tensor of the scalar field.
### III.2 Second example
Let us now consider the case that there is an electromagnetic field and set
the Lagrangian as given by
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi+B(\varphi)\,R-\frac{1}{4}\,F_{\mu\nu}\,F^{\mu\nu}$
(17)
We are interested in analyze the effects of the non minimal coupling in the
standard spatially isotropic and homogeneous metric. To be consistent with the
symmetries of this choice of the metric, an averaging procedure must be
performed if electromagnetic fields are to be taken as the source of the
gravitational field, according to the standard procedure tolmanbook . As a
consequence, the components of the electric $E_{i}$ and magnetic $H_{i}$
fields must satisfy the following relations:
$\displaystyle<{E}_{i}>=0,\qquad<{H}_{i}>$ $\displaystyle=$ $\displaystyle
0,\qquad<{E}_{i}\,{H}_{j}>=0,$ (18) $\displaystyle<{E}_{i}\,{E}_{j}>$
$\displaystyle=$ $\displaystyle-\,\frac{1}{3}{E}^{2}\,g_{ij},$ (19)
$\displaystyle<{H}_{i}\,{H}_{j}>$ $\displaystyle=$
$\displaystyle-\,\frac{1}{3}{H}^{2}\,g_{ij}.$ (20)
where $E$ and $H$ depends only on time.
Using the above average values it follows that the Maxwell energy-momentum
tensor $T_{\mu\nu}^{m}$ reduces to a perfect fluid configuration with energy
density $\rho_{\gamma}$ and pressure $p_{\gamma}$ given by
$<T_{\mu\nu}^{\gamma}>=(\rho_{\gamma}+p_{\gamma})\,v_{\mu}\,v_{\nu}-p_{\gamma}\,g_{\mu\nu},$
(21)
where
$\rho_{\gamma}=3p_{\gamma}=\frac{1}{2}\,(E^{2}+{H}^{2}).$ (22)
From the conservation of the energy-momentum tensor (note that there is no
direct interaction between the scalar and the electromagnetic fields) it
follows that the density of energy goes like $a^{-4}$. Consequently we obtain
the standard value for the scale factor once, as we shall see, there is no
contribution coming from the scalar field to change the curvature of space-
time
$a(t)=a_{0}\,\sqrt{t}$
The equation of motion for the scalar field reduces to
$\Box\,\varphi=0$
that is
$\varphi=\frac{\varphi_{0}}{\sqrt{t}}$
Let us set
$B=b\,\varphi^{2}$
Then we obtain for the annihilation of the density of energy of the scalar
field the two conditions
$\rho_{0}+\rho_{Z}=\frac{1}{2}\,\dot{\varphi}^{2}-\,\frac{2}{3}\,B\,\theta^{2}-2\,\dot{B}\,\theta=0.$
$p_{0}+p_{Z}=\frac{1}{2}\,\dot{\varphi}^{2}+2\,\ddot{B}+\frac{4}{3}\,\dot{B}\,\theta+\,\frac{2}{3}\,B\,\theta^{2}+\frac{4}{3}\,B\,\dot{\theta}=0.$
These equations are satisfied only in the case $b=-1/12$ that is in the case
the scalar field is coupled conformal to gravity.
We can then state the Lemma: _The combined coupling of a minimal and conformal
coupling of a massless scalar field to gravity can annihilate the effect of
the scalar field on gravity in the presence of a cosmological scenario driven
by a gas of photons._
## IV Curvature induced cosmological constant
Let us now follow another path and show how it is possible that a combination
of the minimal and a specific non minimal interaction of a scalar field to
gravity can generate a cosmological constant. We start by a very short review
of the original framework of inflationary universe linde .
In order to explore the consequences of a primordial phase in which the
universe has a constant expansion factor, $\theta=\theta_{0}$ in our notation,
it was suggested to make appeal to the spontaneous symmetry breaking mechanism
that allows a simple configuration of a constant scalar field controlled by
the minimum of a given potential $V(\varphi)$ drive by the Lagrangian
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi+V(\varphi)$
(23)
where we set $\hbar=c=1.$ In a spatially homogeneous and isotropic geometry
the scalar field must depend only on the cosmical time. Then its energy
density and pressure are given by
$\rho=\frac{1}{2}\,\dot{\varphi}^{2}-V$
$p=\frac{1}{2}\,\dot{\varphi}^{2}+V.$
The equation of motion takes the form
$\Box\varphi-\frac{\delta V}{\delta\varphi}=0$
For constant solution, in the extremum of the potential, one obtains trivially
the condition that turns the distribution of the energy of the field into a
cosmological constant. Thus the true origin of the cosmological constant in
this framework rests on the self-interaction contained in the potential
$V(\varphi).$
Let us now show that there is another possibility that does not need such
self-interaction but instead use the non minimal coupling of the scalar field
to gravity to generate a cosmological constant.
The Lagrangian for a massive scalar field coupled non minimally to gravity is
given by
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi+b\,\varphi^{2}\,R-\frac{1}{2}\,\mu^{2}\,\varphi^{2}$
(24)
In the framework of a spatially isotropic and homogeneous space time with
Euclidean section driven by the cosmological constant (where the expansion
factor is constant $\theta_{0}$ ) the minimal coupling part of the energy-
momentum of the field takes the form
$\rho_{0}=\frac{1}{2}\,\dot{\varphi}^{2}+\frac{1}{2}\,\mu^{2}\,\varphi^{2}$
$p_{0}=\frac{1}{2}\,\dot{\varphi}^{2}-\frac{1}{2}\,\mu^{2}\,\varphi^{2}$
and for the non minimal coupling it yields
$\rho_{Z}=-\,\frac{2}{3}\,B\,\theta^{2}-2\,\dot{B}\,\theta.$
$p_{Z}=2\,\ddot{B}+\frac{4}{3}\,\dot{B}\,\theta+\,\frac{2}{3}\,B\,\theta^{2}+\frac{4}{3}\,B\,\dot{\theta}$
in which $B=b\,\varphi^{2}.$
The case of a constant field $\varphi=\varphi_{0}$ produces an expanding
universe (Euclidean section) with a constant expansion $\theta=\theta_{0}.$
The equation of motion of the scalar field combined with the equation of the
metric $\kappa\,\rho=\theta^{2}/3c^{2}$ implies that
$\rho_{0}+\rho_{Z}+p_{0}+p_{Z}=0$
$\theta_{0}^{2}=\frac{3}{8}\,\frac{\mu^{2}}{b}$
$\varphi_{0}^{2}=\frac{4}{3}\,\left(\frac{\theta_{0}}{\mu}\right)^{2}.$
In other words, a cosmological constant appears thanks to the non minimal
coupling given by
$\Lambda=\frac{1}{4}\,\mu^{2}\,\varphi^{2}_{0}=\frac{1}{3}\,\theta_{0}^{2}.$
This solution is stable. This can be easily shown by making the perturbation
$\varphi_{0}\rightarrow\varphi_{0}+\delta\varphi$ and
$\theta_{0}\rightarrow\theta_{0}+\delta\theta.$ The time dependance is given
by
$\delta\varphi=\delta\varphi_{i}\,\exp{(-\theta_{0}\,t)}.$
Note that in the present case the non minimal coupling cannot be a conformal
one com .
### IV.1 Spontaneous symmetry breaking or non minimal coupling?
Let us analyze now a generalization of both cases examined in the previous
sections. We set for the Lagrangian
$L=\frac{1}{2\kappa}\,R+\frac{1}{2}\,\partial_{\mu}\varphi\,\partial^{\mu}\varphi+V(\varphi)\left(1-\frac{2\,b}{\mu^{2}}\,R\right)$
(25)
where $b$ is a dimensionless constant and for comparision to the standard
framework we set for the potential the form
$V(\varphi)=-\,\frac{\mu^{2}}{2}\,\varphi^{2}+\lambda\,\varphi^{4}.$
We limit the analysis to the cosmological scenario of spatially homogeneous
and isotropic metric with Euclidean section for constant expansion where the
scalar of curvature reduces to
$R=\frac{4}{3}\,\theta^{2}.$
The equation of motion is given by
$\Box\varphi-\left(1-\frac{2\,b}{\mu^{2}}\,R\right)\,\frac{\delta
V}{\delta\varphi}=0$
In the case of a constant solution $\varphi=\varphi_{0}$ the energy-momentum
of the field reduces to the density and pressure given by
$\rho=-V+\frac{4}{3}\,\frac{b\,\theta^{2}}{\mu^{2}}\,V$
$p=+V-\frac{4}{3}\,\frac{b\,\theta^{2}}{\mu^{2}}\,V$
that is, the distribution of energy is nothing but an effective cosmological
constant, that is, $\rho+p=0$
In this case, the dynamics of $\varphi$ admits two class of solutions:
$\frac{\delta V}{\delta\varphi}=0$ (26)
or
$1-\frac{2\,b}{\mu^{2}}\,R=0$ (27)
The first case led to the break of symmetry scenario of standard inflation.
The second one represents the combined effect of minimal and non minimal
coupling of $\varphi$ field to gravity where the expansion factor is given by
$\theta_{0}^{2}=\frac{3}{8}\,\frac{\mu^{2}}{b}.$
In both cases the energy distribution of the scalar field is identical to a
cosmological constant.
In the gravitational case the equation
$\rho=\frac{\theta^{2}}{3}$
yields c1
$\varphi_{0}^{2}=\frac{\mu^{2}}{4\,\lambda}\,\left(1+\sqrt{1-\frac{32}{3}\,\frac{\lambda\,\theta^{2}_{0}}{\mu^{4}}}\right).$
(28)
## V Final comments
We have presented some cases in which the effect of non minimal coupling of a
scalar field to the curvature of space-time allows the existence of states
such that the net effect of the free part of the energy added to the
gravitational interaction annihilate each other. This means that although the
scalar field is embedded in a curved space its contribution to the metric
vanishes. A similar interaction allows the generation of a cosmological
constant without self interaction of the scalar field. In a forth coming work
we examine these properties for other fields, in particular for the
electromagnetic field novellohartmann , hartmannenovello .
### V.1 Acknowledgements
We would like to acknowledge the financial support from brazilian agencies
Finep, Faperj and CNPq.
## References
* (1) A. Guth, Inflationary Universe in Phys Rev D23 (1981) 347.
* (2) A. Starobinsky, A new type of isotropic cosmological models without singularity in Phys Lett 91 B (1980) 99
* (3) M Novello and J M Salim in Phys Rev D (1979)
* (4) M Novello and S E P Bergliaffa: Bouncing Cosmologies in Physics Reports vol 463, no 4, july 2008.
* (5) R. Tolman _Relativity, Thermodynamics, and Cosmology_ , Oxford, Clarendon (1934).
* (6) A. D. Linde: The inflationary universe in Rep. Progr. Phys. vol 47, pp 925-986, 1984.
* (7) M. Novello and A.Hartmann in Is the elctromagnetic field responsible for the cosmic acceleration in late times? (to be submited)
* (8) A.Hartmann and M.Novello in A new bifurcation in the universe, to be submitted for publication.
* (9) We note that the expression of the constant value of the field reads $\varphi_{0}^{2}=4\theta_{0}^{2}/3\mu^{2}L_{Pl}^{2}.$ In the paper we are using units such that the Planck lenght is equal to $1.$
* (10) Compare with the case of SSB in which the condition to extremize the potential yields the same result for the particular choice of constant $\lambda$ given by $\lambda=3\,b\,\mu^{2}/4.$
|
# Global Optimization for Trajectory Design via Invariant Manifolds in the
Earth-Moon Circular Restricted Three-Body Problem
Flavio Tagliaferri Advanced Concepts Team, European Space Agency, European
Space Research and Technology Centre (ESTEC), Keplerlaan 1, 2201 AZ Noordwijk,
The Netherlands Sapienza University of Rome, Via Eudossiana 18, 00184, Rome,
Italy Emmanuel Blazquez Advanced Concepts Team, European Space Agency,
European Space Research and Technology Centre (ESTEC), Keplerlaan 1, 2201 AZ
Noordwijk, The Netherlands Giacomo Acciarini Advanced Concepts Team,
European Space Agency, European Space Research and Technology Centre (ESTEC),
Keplerlaan 1, 2201 AZ Noordwijk, The Netherlands Surrey Space Centre,
University of Surrey, GU2 7XH, Guildford, United Kingdom Dario Izzo Advanced
Concepts Team, European Space Agency, European Space Research and Technology
Centre (ESTEC), Keplerlaan 1, 2201 AZ Noordwijk, The Netherlands
_Abstract_ \- This study addresses optimal impulsive trajectory design within
the Circular Restricted Three-Body Problem (CR3BP), presenting a global
optimization-based approach to identify minimum $\Delta V$ transfers between
periodic orbits, including heteroclinic connections. By combining a Monotonic
Basin Hopping (MBH) algorithm with a sequential quadratic solver in a parallel
optimization framework, a wide range of minimum $\Delta V$ transfers are
efficiently found. To validate this approach, known connections from the
literature are reproduced. Consequently, three-dimensional periodic orbits are
explored and a systematic search for minimum propellant trajectories is
conducted within a selected interval of Jacobi constants and a maximum time of
flight. Analysis of the results reveals the presence of very low $\Delta V$
solutions and showcases the algorithm’s effectiveness across various mission
scenarios.
## I. Introduction
Nowadays, as multiple space agencies and private companies are becoming more
interested in developing missions far beyond the Earth orbit, trajectory
design in multi-body environments becomes a fundamental asset. In particular,
the Moon is considered the cornerstone of the near future space exploration
[1], with several missions aiming to bring scientific discovery, technology
advancement, and acquire the essential skills needed to thrive in a different
planetary environment, specifically in preparation for upcoming human missions
to other planets, e.g. Mars. In this context, the Earth-Moon Circular
Restricted Three-Body Problem (CR3BP) represents a key mathematical framework
for preliminary mission analysis studies, offering families of periodic orbits
about the lunar libration points with peculiar operational characteristics,
and providing a potential low-energy transport network to move between these
orbits, by means of the invariant manifolds [2]. In the CR3BP, the process of
optimal trajectory design can be challenging and traditionally involves the
use of Poincaré maps (PM), a tool from dynamical systems theory, which allows
the mission designer to visualize the solution space and select promising
candidate trajectories. In [3] Koon et al. exploited this tool to compute
propellant-free heteroclinic transfers, between $L_{1}$ and $L_{2}$ periodic
orbits in the planar system. However, when the full 6-dimensional CR3BP is
considered, the use of PMs becomes more demanding due to the higher
dimensionality of the system which does not enable a qualitative analysis of
the maps [4]. Several optimization strategies were developed to tackle this
problem. For example, in [5] Haapala and Howell implemented a combination of
interactive and automated search strategies to locate maneuver-free and low
$\Delta V$ transfers between vertical Lyapunov, halo and axial orbits with
different Jacobi constant in the Earth-Moon systems, and consequently
established a catalog of these transfers. In [6] Zhaoyu Li, together with the
other authors, proposed a global search strategy based on comparing the
orbital state of the unstable and stable manifolds, incorporated with low-
thrust techniques, to seek a suitable matching point for maneuver application.
The obtained velocity increments were then refined to achieve more fuel-
optimal transfers using Sequential Quadratic Programming techniques, also
deriving the gradient of the constraint to enhance the algorithm convergence
accuracy and rapidity. Recently, in [7] D. B. Henry and D.J. Scheeres proposed
a robust multiparameter continuation scheme to retrieve the families of
transfers between both periodic and quasi-periodic orbits, using the PMs to
identify the starting solution.
In this work, we propose a black-box approach based on combining the use of a
global optimization algorithm, the Monotonic Basin Hopping (MBH), with the
sequential quadratic solver SNOPT [8] in a parallel optimization framework to
perform trajectory design in the CR3BP. More specifically, we focus on
impulsive minimum propellant transfers via invariant manifolds, between halo
and vertical Lyapunov orbits at the same value of Jacobi constant ($C$), about
the $L_{1}$ and $L_{2}$ libration points of the Earth-Moon system. To validate
the approach, we reproduce trajectories from the literature, both in the
planar and spatial scenarios. We consequently perform a systematic search of
minimum propellant trajectories in a region of Jacobi constants and accounting
for a maximum time of flight. We finally report on the best trajectories found
with low propellant consumption, for each transfer type. The proposed approach
can be used to retrieve initial low $\Delta V$ trajectories without employing
preliminary orbit cartography.
## II. Theoretical background
In this section, we begin by describing the dynamical system considered.
Subsequently, we delve into the method for retrieving the selected periodic
orbits, followed by a discussion on computing stable and unstable manifolds.
### A. The Circular restricted three-body problem
The circular restricted three-body problem describes the motion of an object
(satellite) with a negligible mass under the gravitational influence of two
other masses $m_{1}$ and $m_{2}$. The larger of these two, $m_{1}$ (Earth) is
called the primary, while $m_{2}$ (Moon) is called the secondary. The primary
and secondary are assumed to orbit their center of mass in a circular orbit.
This dynamical system is usually implemented in a non-dimensional coordinate
system with distance unit (DU) equal to the distance between $m_{1}$ and
$m_{2}$ (i.e., $DU=384,400$ km for the Earth-Moon system). The time unit (TU)
is such that the mean angular motion of the primary and secondary bodies about
the system’s center of mass is unitary ($TU=375,699.79375$ s for the Earth-
Moon system). The origin lies at the center of mass of $m_{1}$ and $m_{2}$,
and the $\hat{x}$ axis points from the primary to the secondary. Aligning
parallel to the angular momentum vector of the system is the $\hat{z}$ axis,
while the $\hat{y}$ axis completes the right-handed system. This reference
frame, known as Synodic reference frame, undergoes a constant rotation around
the $\hat{z}$ axis, sharing the same angular velocity as the orbital motion of
the primary and secondary, as depicted in Fig.1. Thus, the coordinates of both
the primary and secondary remain fixed. The dynamical system can be fully
described by (1):
$\left\\{\begin{array}[]{@{}l@{}@{}}\ddot{x}-2\dot{y}=\pdv{U}{x}\\\
\ddot{y}+2\dot{x}=\pdv{U}{y}\\\ \ddot{z}=\pdv{U}{z}\\\ \end{array}\right.\ $
(1)
Where:
$U(x,y,z)=\frac{(x^{2}+y^{2})}{2}+\frac{1-\mu}{r_{1}}+\frac{\mu}{r_{2}}$ (2)
represents the potential function, $\mu=\frac{m_{2}}{m_{1}+m_{2}}$ is the mass
ratio parameter, which relates the masses of the primary and secondary
($\mu$=0.01215058 for the Earth-Moon system), $r_{1}$ the distance of the
object to the primary and $r_{2}$ its distance to the secondary. The CR3BP has
five equilibrium points, known as Lagrange points. Three of these points are
known as the collinear points and lie on the $\hat{x}$ axis. The two collinear
points closest to the moon, $L_{1}$ and $L_{2}$ are used for this study. This
system admits a first integral of motion, called the Jacobi constant, defined
through (3):
$C=2U-({\dot{x}}^{2}+{\dot{y}}^{2}+{\dot{z}}^{2})\ $ (3)
This parameter is used in designing heteroclinic connections between different
periodic orbits since it is required that the departure and arrival orbits
must have the same Jacobi constant. Finally, by constraining the object to
move in the xy plane, a simplified framework is retrieved, called Planar
Circular Restricted Three-Body Problem (PCR3BP).
Fig. 1: Synodic reference frame and location of the libration points, $L_{1}$
and $L_{2}$
### B. Periodic orbits
The computation of periodic orbits is performed using a predictor-corrector
method. Initially, the state is numerically integrated, which results in a
deviation from periodicity. This initial state is then corrected based on
information obtained from the monodromy matrix, defined as the state
transition matrix evaluated over one complete period T of the orbit. To
achieve this, routines from SEMpy, a Python library specialized in non-
Keplerian astrodynamics [9], were utilized and adapted to retrieve the desired
values of Jacobi constants for halo, planar, and vertical Lyapunov orbits.
### C. Invariant manifolds
Invariant Manifolds are higher-dimensional dynamical structures that govern
the asymptotic flow of motion towards and away from unstable orbits. Those can
be computed by exploiting the properties of the monodromy matrix. For each
point $\boldsymbol{X_{i}}=[x,\ y,\ z,\ \dot{x},\dot{y},\dot{z}]$ on the
periodic orbit, the initial state of the stable and unstable manifolds, can be
computed using (4) and (5):
$\boldsymbol{X_{i}^{S}}=\boldsymbol{X_{i}}\pm\epsilon\frac{\Phi_{i}{,_{0}}\nu_{0}^{S}}{\left\lVert\Phi_{i}{,_{0}}\nu_{0}^{S}\right\rVert.}$
(4)
$\boldsymbol{X_{i}^{U}}=\boldsymbol{X_{i}}\pm\epsilon\frac{\Phi_{i}{,_{0}}\nu_{0}^{U}}{\left\lVert\Phi_{i}{,_{0}}\nu_{0}^{U}\right\rVert.}$
(5)
Where $\Phi_{i}{,_{0}}$ is the state transition matrix computed from a
reference state $\boldsymbol{X_{0}}$ up to $\boldsymbol{X_{i}}$, while
$\nu_{0}^{S}$ and $\nu_{0}^{U}$ represent the stable ad unstable eigenvectors
of the monodromy matrix evaluated at $\boldsymbol{X_{0}}$, corresponding to
the smallest and largest eigenvalue, respectively. The value of the step
parameter $\epsilon$ is applied to the whole state in the positive and
negative directions, i.e. interior and exterior w.r.t. CR3BP’s center of mass
[10]. In this study, its value is taken constant and equal to 50km for both
manifolds, which is a common choice for the Earth-Moon system [11]. Finally,
the manifolds can be retrieved by numerically integrating the state backward
and forward in time for the stable and unstable manifolds, respectively.
Intersections of the unstable and stable manifolds result in low-energy
transfers.
## III. Trajectory design
In this section, the proposed approach is explained in detail. First, the
chosen problem formulation is addressed, then the specific optimization
architecture used is illustrated.
### A. Problem formulation
In the proposed approach, the search for low $\Delta V$ trajectories between
periodic orbits is formulated as a single-objective constrained optimization
problem through (6), (7) and (8):
minimize $\displaystyle J=\lVert\Delta V(\textbf{x})\rVert$ (6) subject to
$\displaystyle\Delta x=0,\Delta y=0,\Delta z=0$ (7) $\displaystyle
l_{B}\leq\textbf{x}\leq u_{B}$ (8)
where the Euclidean norm of the difference in velocity and the components of
the difference in position, i.e. $\Delta x$, $\Delta y$ and $\Delta z$, both
computed at the patching point of the unstable and stable manifolds branches,
represent the objective function $J$ to be minimized and the three equality
constraints to be fulfilled. The decision vector x is composed of four
optimization variables: $[\theta_{U},\tau_{U},\theta_{S,}\tau_{S}]$. In
particular $(\theta_{U},\theta_{S})$, called the departure and arrival angles,
are defined as fractions of the orbit’s periods $T_{d_{\text{ep}}}$ and
$T_{a_{\text{rr}}}$ through (9) and (10), respectively:
$\theta_{U}=\frac{t_{d{\text{ep}}}}{T_{d{\text{ep}}}}$ (9)
$\theta_{S}=\frac{t_{a{\text{rr}}}}{T_{a{\text{rr}}}}$ (10)
where $t_{d{\text{ep}}}$ and $t_{a{\text{rr}}}$ are the coasting times taken
from the reference starting points $\boldsymbol{X_{0}{{}^{U}}}$ and
$\boldsymbol{X_{0}{{}^{S}}}$ along the departure and arrival orbits. They are
used to reach the states $\boldsymbol{X_{i}{{}^{U}}}$ and
$\boldsymbol{X_{i}{{}^{S}}}$ from which the unstable and stable manifolds
branches are computed. The coasting times on the unstable and stable manifold
branches are represented by $(\tau_{U},\tau_{S})$. We will refer the sum of
$\tau_{U}$ and $\tau_{S}$ as the time of flight ($TOF$) of a given trajectory.
The optimization variables are bounded between $u_{L}=[0,0,0,0]$ and
$u_{B}=[1,8,1,8]$. An example of a transfer scenario is depicted in Fig.2.
Fig. 2: Example of a simplified mission scenario, with all the optimization
variables represented
### B. Optimization architecture
To solve the presented optimization problem a nested optimization architecture
is employed, where a Monotonic Basin Hopping (MBH) algorithm uses sequential
quadratic programming (and in particular SNOPT) as its local optimizer. MBH is
a stochastic global optimization algorithm [12] that works by continuously
applying a perturbation vector to the problem’s optimization variables in a
method called hopping. Following each hop, a local optimization is conducted
on the new configuration. If this step yields an improved solution, that
solution updates the set of optimization variables for the next perturbation.
This iterative process is repeated until a maximum number of hops is reached.
Here we adopted a modified version of the standard MBH. In particular, the
initial decision vector is obtained as the best individual among a population
of 200, generated from a random seed. A tolerance of $10^{-}6$ is set for the
three equality constraints. The architecture is then integrated in a parallel
optimization framework, where a number of 100 different computational unit,
called islands, are used to solve one optimization problem. The best solution
is then extracted among the best individual of each island. Note that in
principle, multiple trajectories can be retrieved with the solution of a
single optimization problem. This could allow the mission designer to make
further trade offs based on geometry and time of flight. The resulted
optimization architecture as well as the parallel framework was created using
the open-source Python library pyGMO [13]. Moreover, to increase the speed and
efficiency of the optimization process, the open-source Python library Heyoka
[14] was employed for fast integration of ODE systems.
## IV. Validation
In this section we validate the proposed approach by reproducing established
trajectories from the literature. In particular, the algorithm is first tested
in the PCR3BP, where heteroclinic connections between planar Lyapunov orbits
are sought. Then the CR3BP framework is considered, where heteroclinic
connections between periodic orbits become challenging to find. For this
reason, a minimum fuel trajectory retrieved through a state-of-the-art method
in [7] is set as a reference solution for the validation of the algorithm.
Furthermore, each trajectory is characterized by the Euclidean norm of the
velocity and position differences at the patching point, i.e $\Delta V$ and
$\Delta R$, along with the $TOF$. The specific values of the key parameters
used in the two validation cases can be found in Tab. 1.
Tab. 1: Key parameters used in the validation cases
$Case$ | $\mu$ | $\epsilon$ | $TOF[TU]$
---|---|---|---
PCR3BP | 0.012150585 | $1\times 10^{-6}$ | -
CR3BP | 0.01215 | $1\times 10^{-5}$ | $[0,6]$
### A. Planar trajectories
The first task we tackle is the computation of heteroclinic connections
between planar Lyapunov orbits, belonging to three different values of the
Jacobi constant coming from [15]. The detailed numerical results are reported
in Tab. 2., while the trajectories are represented in Fig.3., where the
unstable and stable branches are depicted in red and green, respectively. As
one can see, in all the three cases, the results successfully led to the
reconstruction of trajectories with the same qualitative geometry as the ones
in the reference paper, up to a precision of about $10^{-4}$ m/s in velocity
and of hundreds of meters in position at the patching point.
Fig. 3: Representation of the obtained numerical heteroclinic connections
between planar Lyapunov orbits for cases (a), (b) and (c) in Tab. 2. Tab. 2:
Numerical results of the validation in the PCR3BP
$Case$ | $C$ | $\Delta V$ $[m/s]$ | $\Delta R$ $[m]$ | $TOF[dd]$
---|---|---|---|---
(a) | 3.130459 | $4.0\times 10^{-5}$ | 25.53 | 48.075
(b) | 3.097474 | $1.3\times 10^{-4}$ | 491.80 | 52.787
(c) | 3.025554 | $4.3\times 10^{-4}$ | 360.08 | 75.079
### B. Spatial trajectories
As a successive test, spatial trajectories are considered. Since finding
heteroclinic connections between three-dimensional periodic orbits is very
challenging due to the low dimension of their manifolds relative to the phase
space, we perform the validation in the CR3BP case by considering the example
presented in [7]. This trajectory connects with minimum propellant a $L_{1}$
vertical Lyapunov and a $L_{2}$ southern halo orbits at a value of $C$ equal
to 3.1328. The results led to the computation of a comparable trajectory in
terms of geometry, propellant consumption and time of flight, as showed in
Tab. 3. Then, the best minimum fuel solution found by our approach is also
retrieved. As one can see from Tab. 3., the best minimum fuel solution found
by the presented algorithm requires about 45% less fuel and about 13% longer
$TOF$. The corresponding trajectories are depicted in Fig.4. and in Fig.5.
Tab. 3: Numerical results of the validation in the CR3BP
$Case$ | $C$ | $\Delta V$ [m/s] | $\Delta R$ [m] | $TOF[dd]$
---|---|---|---|---
Reference [7] | 3.1328 | 280.47 | - | 39.570
Reproduced | 3.1328 | 280.70 | 14.38 | 39.564
Best min-fuel | 3.1328 | 149.10 | 64.04 | 44.789
Fig. 4: Representation of the trajectory from the reproduced case in Tab. 3.
from [7]
(a) Three-dimensional trajectory
(b) Projections in the xy and xz planes
Fig. 5: Representation of the trajectory from the best min-fuel case in Tab.
3., retrieved with the proposed algorithm
## V. Applications
In this section, we apply the algorithm to three different transfer types:
halo to halo, vertical Lyapunov to vertical Lyapunov and halo to vertical
Lyapunov. Our specific goal is to identify the global optimal trajectories
that the method can produce, for each transfer type. To achieve this, we
systematically explore minimum fuel trajectories between halo and vertical
Lyapunov orbits. This investigation covers a Jacobi constant range of
$[3.057,3.152]$ and a maximum time of flight of approximately $70$ days.
Notably, both the interior and exterior manifolds are considered in this
study. The upper bound of the Jacobi constant corresponds to the value of the
bifurcation of the $L_{2}$ halo family from the Lyapunov family [5]. A step of
$6\times 10^{-4}$ is used to determine the orbit’s Jacobi constant and it
comes from a trade-off between the number of trajectories obtained and the
dimension of successive orbits, i.e. the z-amplitudes. Each mission scenario
is identified by:
1. 1.
The departure and arrival orbits characterized by a value of $C$ and their
respective libration points;
2. 2.
The Family of each orbit;
3. 3.
The types of manifolds branches used.
Based on these characteristics, a compact notation is introduced in Tab. 4. to
identify a particular mission scenario.
Tab. 4: Description of the notation adopted to identify a mission scenario $Notation$ | $Description$
---|---
NH1 | $L_{1}$ northern halo
SH1 | $L_{1}$ southern halo
V1 | $L_{1}$ vertical Lyapunov
NH2 | $L_{2}$ northern halo
SH2 | $L_{2}$ southern halo
V2 | $L_{2}$ vertical Lyapunov
+ | interior manifold
- | exterior manifold
outward | $L_{1}$ to $L_{2}$
return | $L_{2}$ to $L_{1}$
For example, with $[NH1-,V2+]$ we describe a trajectory obtained departing
from a $L_{1}$ northern halo orbit moving on the exterior unstable manifold
branch and arriving to a $L_{2}$ vertical Lyapunov orbit using the interior
stable manifold branch. Trajectories which either exhibit a $\Delta R\geq 1$
km or impact on the Moon surface are discarded and excluded from the analysis.
In the following subsections, we report the global optimal trajectories
discovered with this algorithm, for each transfer type.
### A. Halo to halo trajectories
The best solutions in terms of velocity increment for outward and return
trajectories are reported in Tab. 5. Observing the results, we found several
trajectories with remarkably low fuel requirements. Among them, the
trajectories with the lowest $\Delta V$ are then represented in Fig.6. and
Fig.7. Furthermore, the algorithm exhibits a convergence toward outcomes that
align with those achievable through the application of CR3BP’s symmetries
[16]. For example, the successive application of the image of the xy-plane and
backward image of the xz-plane to the $[NH1+,SH2-]$ would produce the
$[NH2-,SH1+]$ trajectory, while the use of only the backward image of the xz-
plane would result in the $[SH2-,NH1+]$, with the same $C$, $\Delta V$ and
$TOF$ values. Consistent findings emerged from solving the optimization
problems corresponding to these three mission scenarios. This observed
behavior validates the obtained results and could offer an opportunity to
streamline the investigation of mission scenarios when employing this
numerical method.
Tab. 5: Numerical results for the optimal halo to halo trajectories, in
outward and return mission scenarios
Mission scenario | $C$ | $\boldsymbol{\Delta V}$ [m/s] | $\boldsymbol{\Delta R}$ [m] | $TOF$ [dd]
---|---|---|---|---
$[NH1-,NH2-]$ | 3.132397 | 3.000 | 110.76 | 50.976
$[NH1+,SH2-]$ | 3.080797 | 2.980 | 9.02 | 64.410
$[SH1-,SH2-]$ | 3.131797 | 2.684 | 144.10 | 51.077
$[SH1+,NH2-]$ | 3.071197 | 4.991 | 0.06 | 61.276
$[NH2+,NH1-]$ | 3.129397 | 3.903 | 446.72 | 49.83
$[SH2-,NH1+]$ | 3.080797 | 2.980 | 44.49 | 64.410
$[SH2+,SH1+]$ | 3.132397 | 2.999 | 414.86 | 50.976
$[NH2-,SH1+]$ | 3.080797 | 5.029 | 31.70 | 64.405
Fig. 6: Representation of the global optimum solution for outward
trajectories, corresponding to the mission scenario $[SH1-,SH2-]$ in Tab. 5.
Fig. 7: Representation of the global optimum solution for return trajectories,
corresponding to the mission scenario $[SH2-,NH1+]$ in Tab. 5.
### B. Vertical Lyapunov to vertical Lyapunov trajectories
The second application we addressed involves computing trajectories connecting
vertical Lyapunov orbits. The best solutions in terms of velocity increment
for the outward and return scenarios are reported in Tab. 6. Among them, the
trajectories with the lowest $\Delta V$ are then represented in Fig.8. and
Fig.9. In particular, the mission scenario $[V2+,V1+]$ shows a $\Delta V$ of
4.128 m/s which is, to the best of our knowledge and based of the literature
we read, a value remarkably low for this type of transfer. As for halo to halo
trajectories, the algorithm demonstrates adherence to the symmetry properties
of the CR3BP.
Tab. 6: Numerical results for the optimal vertical to vertical Lyapunov
trajectories, in outward mission scenarios
Mission scenario | $C$ | $\boldsymbol{\Delta V}$ [m/s] | $\boldsymbol{\Delta R}$ [m] | $TOF$ [dd]
---|---|---|---|---
$[V1+,V2+]$ | 3.109600 | 39.629 | 95.48 | 41.583
$[V1-,V2+]$ | 3.086799 | 13.068 | 35.23 | 23.253
$[V1+,V2-]$ | 3.082600 | 9.150 | 1.16 | 38.803
$[V1-,V2-]$ | 3.089200 | 6.961 | 430.75 | 49.406
$[V2+,V1+]$ | 3.084400 | 4.128 | 0.50 | 63.482
$[V2-,V1+]$ | 3.104800 | 61.037 | 3.38 | 26.467
$[V2+,V1-]$ | 3.089200 | 6.962 | 2.27 | 49.406
$[V2-,V1-]$ | 3.086799 | 13.0776 | 25.45 | 23.251
Fig. 8: Representation of the global optimum solution for outward
trajectories, corresponding to the mission scenario $[V1-,V2-]$ in Tab. 6.
Fig. 9: Representation of the global optimum solution for return trajectories,
corresponding to the mission scenario $[V2+,V1+]$ in Tab. 6.
### C. Halo to vertical Lyapunov trajectories
The last application concerns transfers associated with halo orbits and
vertical Lyapunov orbits. The best solutions in terms of velocity increment
for the outward and return scenarios are reported in Tab. 7. Among them, the
trajectories with the lowest $\Delta V$ are then represented in Fig.10. and
Fig.11. Higher values of the $\Delta V$ are noticed compared to previous
transfer types. Also in this scenario, the algorithm follows symmetry
properties, as one can see in Tab. 7.
Tab. 7: Numerical results for the optimal halo to vertical Lyapunov
trajectories, in outward mission scenarios
Mission scenario | $C$ | $\boldsymbol{\Delta V}$ [m/s] | $\boldsymbol{\Delta R}$ [m] | TOF [dd]
---|---|---|---|---
$[NH1+,V2-]$ | 3.084997 | 29.338 | 6.37 | 58.522
$[NH1-,V2-]$ | 3.138997 | 37.159 | 269.83 | 63.321
$[SH1+,V2-]$ | 3.084397 | 29.283 | 0.11 | 58.630
$[SH1-,V2-]$ | 3.108397 | 32.678 | 187.17 | 63.670
$[V2+,NH1+]$ | 3.083197 | 31.950 | 13.89 | 58.866
$[V2+,NH1-]$ | 3.086197 | 30.731 | 38.40 | 58.317
$[V2+,SH1+]$ | 3.106597 | 20.939 | 49.32 | 64.209
$[V2+,SH1-]$ | 3.086197 | 30.731 | 0.14 | 58.317
Fig. 10: Representation of the global optimum solution for outward
trajectories, corresponding to the mission scenario $[SH1+,V2-]$ in Tab. 7.
Fig. 11: Representation of the global optimum solution for return
trajectories, corresponding to the mission scenario $[V2+,SH1+]$ in Tab. 7.
## VI. Conclusions
In this study, we introduce a general black-box methodology for trajectory
design within the framework of the CR3BP. Our validation demonstrates that the
proposed approach yields low-cost trajectories that are comparable to those
found in the literature. Leveraging a direct problem formulation and an
optimized computational architecture, we efficiently compute a large number of
trajectories across various scenarios in an automated manner. The numerical
results reveal very low $\Delta V$ solutions, but also alternative
trajectories that could serve as initial approximations for existing
methodologies. Remarkably, despite its reliance on a numerical approach, the
algorithm demonstrates convergence toward the symmetries of the CR3BP,
achieving a certain precision. Furthermore, our proposed approach suggests an
exploration of a broader search space compared to traditional methods that
rely on Poincaré maps. Indeed, the algorithm we employed operates without the
need for defining a surface of section or imposing problem-dependent
constraints to reduce system dimensionality. These unique characteristics
could facilitate the uncovering of new fuel-optimal solutions, as evidenced by
the numerical results for trajectories from vertical Lyapunov to southern halo
orbits. Our future research will delve into a more detailed study and analysis
of this property, along with an investigation into the convergence limits of
the algorithm.
## References
* [1] European Space Agency. “ESA Strategy for Science at the Moon.” https://exploration.esa.int/web/moon/-/61371-esa-strategy-for-science-at-the-moon, 2019.
* [2] Szebehely, V. Theory of Orbits: The Restricted Problem of Three Bodies. Academic Press, New York, 1967.
* [3] Koon, W. S., Lo, M. W., Marsden, J. E., and Ross, S. D. “Heteroclinic Connections between Periodic Orbits and Resonance Transitions in Celestial Mechanics.” Chaos: An Interdisciplinary Journal of Nonlinear Science, Vol. 10, No. 2, pp. 427–469, 2000.
* [4] Haapala, A. F. and Howell, K. C. “Representations of Higher-Dimensional Poincaré Maps with Applications to Spacecraft Trajectory Design.” Acta Astronautica, Vol. 96, pp. 23–41, March 2014.
* [5] Haapala, A. and Howell, K. “A framework for construction of transfers linking periodic libration point orbits in the Earth-Moon spatial circular restricted three-body problem.” Int. J. Bifurcat. Chaos, Vol. 26, No. 5, 2016.
* [6] Li, Z., Zeng, H., Xu, R., Peng, K., and Huang, Z. “Hierarchical global search for fuel-optimal impulsive transfers between lunar libration orbits.” Adv. Space Res., Vol. 67, No. 6, pp. 1773–1786, 2021.
* [7] Henry, D. B. and Scheeres, D. J. “Quasi-periodic orbit transfer design via whisker intersection sets.” J Guid Control Dyn, pp. 1–16, 2023.
* [8] Gill, P. E., Murray, W., and Saunders, M. A. “SNOPT: An SQP Algorithm for Large-Scale Constrained Optimization.” SIAM Review, Vol. 47, No. 1, pp. 99–131, 2005.
* [9] Blazquez, E. “SEMPy: a python open-source toolbox for non-Keplerian astrodynamics.” “International Conference on Astrodynamics Tools and Techniques (ICATT),” Virtual, 2021.
* [10] Blazquez, E. Rendezvous Optimization and GNC Design for Proximity Operations on cis-lunar Near Rectilinear Halo Orbits. Ph.D. thesis, 2021.
* [11] Gómez, G., Jorba, A., Masdemont, J., and Simó, C. “Study refinement of semi-analytical halo orbit theory.” Final Report, ESOC Contract No: 8625/89/D/MD (SC), 1991.
* [12] Eason, G., Noble, B., and Sneddon, I. N. “On certain Wales, D. J. & Doye, J. P. K. Global optimization by basin-hopping and the lowest energy structures of Lennard–Jones clusters containing up to 110 atoms.” J. Phys. Chem. A, Vol. 101, pp. 5111–5116, 1997.
* [13] Biscani, F. and Izzo, D. “A parallel global multiobjective framework for optimization: pagmo.” Journal of Open Source Software, Vol. 5, No. 53, p. 2338, 2020.
* [14] Biscani, F. and Izzo, D. “Revisiting high-order Taylor methods for astrodynamics and celestial mechanics.” Monthly Notices of the Royal Astronomical Society, Vol. 504, No. 2, pp. 2614–2628, 2021.
* [15] Barrabés, E., Mondelo, J. M., and Ollé, M. “Numerical continuation of families of heteroclinic connections between periodic orbits in a Hamiltonian system.” Nonlinearity, Vol. 26, No. 10, p. 2747, 2013.
* [16] Miele, A. “Revisit of the Theorem of Image Trajectories in the Earth-Moon Space.” Optimization Theory and Applications, Vol. 147, No. 3, pp. 483–490, 2010.
|
$\displaystyle=F\Big{[}-h\,dt^{2}+\frac{dr^{2}}{h}+r^{2}d\theta^{2}\Big{]}+\frac{r^{2}\sin^{2}\theta}{F}\bigl{(}d\phi-\Omega\,dt\bigr{)}^{2}\,,$
(C.6a) $\displaystyle A$
$\displaystyle=A_{0}\,dt+\frac{A_{3}}{F}\bigl{(}d\phi-\Omega\,dt\bigr{)}\,,$
(C.6b)
where
$\displaystyle h$
$\displaystyle=1-\frac{2M}{r}+\frac{Q^{2}}{r^{2}}+\frac{H^{2}}{r^{2}}\,,$
(C.7a) $\displaystyle F$
$\displaystyle=1+\jmath^{2}\Bigl{[}r^{2}\sin^{2}\theta+\bigl{(}Q^{2}+H^{2}\bigr{)}\cos^{2}\theta\Bigr{]}^{2}\,,$
(C.7b) $\displaystyle\Omega$ $\displaystyle=-4\,\jmath\,h\,r\cos\theta\,,$
(C.7c) $\displaystyle A_{0}$ $\displaystyle=-\frac{Q}{r}-\jmath
Hr\Bigl{[}1+h\cos^{2}\theta\Bigr{]}\,,$ (C.7d) $\displaystyle A_{3}$
$\displaystyle=\cos\theta\Bigl{[}H+Q\sqrt{F-1}\Bigr{]}\,.$ (C.7e)
From the components of the electromagnetic potential, we can see that, in a
certain sense, the swirling parameter $\jmath$ generates a magnetic charge
from an electric one, and vice-versa.
## Appendix D Embedding of the Event Horizon in $\mathbb{E}^{3}$
An event horizon can be embedded in the Euclidian three-dimensional space
$\mathbb{E}^{3}$ as a surface of revolution, by comparing the induced metric
on the horizon with that of the flat three-dimensional space in cylindrical
coordinates.
Indeed, the induced metric of an event horizon of radius $r_{H}$ is obtained
by setting $r=r_{H}$ and $t=constant$ in the complete spacetime metric, which
yields
$ds^{2}_{H}=f_{1}d\theta^{2}+f_{2}d\phi^{2}\,,$ (D.1)
where
$\displaystyle f_{1}$ $\displaystyle=g_{\theta\theta}(\theta,r=r_{H})\,,$
(D.2a) $\displaystyle f_{2}$ $\displaystyle=g_{\phi\phi}(\theta,r=r_{H})\,.$
(D.2b)
On the other hand, the flat three-dimensional metric in cylindrical
coordinates is given by
$ds_{3}^{2}=dz^{2}+dR^{2}+R^{2}d\phi^{2}\,.$ (D.3)
Therefore, considering $z$ and $R$ as functions of $\theta$, and comparing the
flat three-dimensional metric in cylindrical coordinates (D.3) with that
induced on the event horizon (D.1), it follows that
$R=f_{2}\,,$ (D.4)
while $z$ is determined by the following differential equation:
$\biggl{[}\frac{dR}{d\theta}\biggr{]}^{2}+\biggl{[}\frac{dz}{d\theta}\biggr{]}^{2}=f_{1}\,.$
(D.5)
In conclusion, the embedding of an event horizon in $\mathbb{E}^{3}$ is
determined once the functions $R=R\bigl{(}\theta\bigr{)}$ and
$z=z\bigl{(}\theta\bigr{)}$ are obtained.
## Appendix E Majorana Representation of the Gamma Matrices
The (real) Majorana representation of the gamma matrices
$\gamma_{a}=\bigl{(}\gamma_{0},\gamma_{1},\gamma_{2},\gamma_{3}\bigr{)}$ can
be obtained by applying the unitary transformation
$U=\frac{1}{\sqrt{2}}\begin{pmatrix}\mathds{1}_{2}&\sigma_{2}\\\
\sigma_{2}&-\mathds{1}_{2}\end{pmatrix}\,,$ (E.1)
to the gamma matrices $\gamma^{\prime}_{a}$ in the standard Dirac
representation:
$\gamma_{a}=U^{-1}\gamma^{\prime}_{a}U\,,$ (E.2)
which yields
$\displaystyle\gamma_{0}$ $\displaystyle=\begin{pmatrix}0&-i\sigma_{2}\\\
-i\sigma_{2}&0\end{pmatrix}\,,$ (E.3) $\displaystyle\gamma_{1}$
$\displaystyle=\begin{pmatrix}-\sigma_{3}&0\\\ 0&-\sigma_{3}\end{pmatrix}\,,$
(E.4) $\displaystyle\gamma_{2}$
$\displaystyle=\begin{pmatrix}0&-i\sigma_{2}\\\ i\sigma_{2}&0\end{pmatrix}\,,$
(E.5) $\displaystyle\gamma_{3}$ $\displaystyle=\begin{pmatrix}\sigma_{1}&0\\\
0&\sigma_{1}\end{pmatrix}\,,$ (E.6) $\displaystyle\gamma_{5}$
$\displaystyle=\gamma_{0}\gamma_{1}\gamma_{2}\gamma_{3}=\begin{pmatrix}i\sigma_{2}&0\\\
0&-i\sigma_{2}\ \end{pmatrix}\,.$ (E.7)
Where the $\sigma_{j}$ denoted the standard representation of the two-
dimensional Pauli matrices
$\sigma_{j}=\begin{pmatrix}\delta_{j3}&\delta_{j1}-i\delta_{j2}\\\
\delta_{j1}+i\delta_{j2}&-\delta_{j3}\end{pmatrix}\,.$ (E.8)
## Bibliography
* [1] A. Einstein. Zur Allgemeinen Relativitätstheorie. [On the General Theory of Relativity]. Sitzungsber. Preuss. Akad. Wiss. Berlin (Math. Phys.), 1915:778–786, (1915). [Addendum: Sitzungsber. Preuss. Akad. Wiss. Berlin (Math. Phys.), 799–801 (1915)].
* [2] A. Einstein. Die Grundlage der Allgemeinen Relativitätstheorie. [The Basis of the General Theory of Relativity.]. Annalen der Physik, 1916:769–822, (1916). doi:10.1002/andp.19163540702.
* [3] G.F.B. Riemann. Ueber die Hypothesen, welche der Geometrie zu Grunde liegen. [On the Hypotheses on which Geometry is Based]. Abhandlungen der Königlichen Gesellschaft der Wissenschaften zu Göttingen, 13:133–159, (1868).
* [4] A. Einstein. Zur Elektrodynamik bewegter Körper. [On the Electrodynamics of Moving Bodies]. Annalen der Physik, 322:891–921, (1905). doi:10.1002/andp.19053221004.
* [5] J.A. Wheeler and K. Ford. Geons, Black Holes, and Quantum Foam: A Life in Physics. W. W. Norton & Company, (1998).
* [6] C.W. Misner, K.S. Thorne, and J.A. Wheeler. Gravitation. W. H. Freeman and Company, (1973).
* [7] G.M. Clemence. The Relativity Effect in Planetary Motions. Rev. Mod. Phys., 19:361–364, (1947). doi:10.1103/RevModPhys.19.361.
* [8] F.W. Dyson, A.S. Eddington, and C. Davidson. A Determination of the Deflection of Light by the Sun’s Gravitational Field, from Observations Made at the Total Eclipse of May 29, 1919. Phil. Trans. Roy. Soc. Lond. A, 220:291–333, (1920). doi:10.1098/rsta.1920.0009.
* [9] R.V. Pound and G.A. Rebka. Gravitational Red-Shift in Nuclear Resonance. Phys. Rev. Lett., 3:439–441, (1959). doi:10.1103/PhysRevLett.3.439.
* [10] B.P. Abbott et al. Observation of Gravitational Waves from a Binary Black Hole Merger. Phys. Rev. Lett., 116:061102, (2016). arXiv:1602.03837, doi:10.1103/PhysRevLett.116.061102.
* [11] K. Schwarzschild. "Über das Gravitationsfeld eines Massenpunktes nach der Einsteinschen Theorie" [On the Gravitational Field of a Mass Point According to Einstein’s Theory]. Sitzungsber. Preuss. Akad. Wiss. Berlin (Math. Phys.), 7:189–196, (1916). arXiv:physics/9905030.
* [12] E.T. Newman and A.I. Janis. Note on the Kerr Spinning-Particle Metric. J. Math. Phys., 6:915–917, (1965). doi:10.1063/1.1704350.
* [13] E.T. Newman et al. Metric of a Rotating, Charged Mass. J. Math. Phys., 6:918–919, (1965). doi:10.1063/1.1704351.
* [14] M. Astorino, R. Martelli, and A. Viganò. Black Holes in a Swirling Universe. Phys. Rev. D, 106:064014, (2022). arXiv:2205.13548, doi:10.1103/PhysRevD.106.064014.
* [15] M.A. Melvin. Pure Magnetic and Electric Geons. Phys. Lett., 8:65–70, (1964). doi:10.1016/0031-9163(64)90801-7.
* [16] S.M. Carroll. Spacetime and Geometry: an Introduction to General Relativity. Addison Wesley, (2004). doi:10.1017/9781108770385.
* [17] R.M. Wald. General Relativity. Chicago Press, (1984). doi:10.7208/chicago/9780226870373.001.0001.
* [18] F.J. Ernst. New Formulation of the Axially Symmetric Gravitational Field Problem. Phys. Rev., 167:1175–1179, (1968). doi:10.1103/PhysRev.167.1175.
* [19] F.J. Ernst. New Formulation of the Axially Symmetric Gravitational Field Problem. II. Phys. Rev., 168:1415–1417, (1968). [Erratum: F.J. Ernst, Phys. Rev., 172(5):1850, (1968)]. doi:10.1103/PhysRev.168.1415.
* [20] A. Viganó. Black Holes and Solution Generating Techniques. Ph.D. Thesis, Milan U., (2022). arXiv:2211.00436.
* [21] R. Martelli. The Action of the Axisymmetric and Stationary Symmetry Group of General Relativity on a Static Black Hole. Bachelor Thesis, Milan U., (2022). URL: https://pcteserver.mi.infn.it/~molinari/TESI/Tesi_Riccardo_Martelli.pdf.
* [22] G. Dall’Agata and M. Zagermann. Supergravity: From First Principles to Modern Applications. Springer Berlin, (2021). doi:10.1007/978-3-662-63980-1.
* [23] L.J. Romans. Supersymmetric, Cold and Lukewarm Black Holes in Cosmological Einstein-Maxwell Theory. Nucl. Phys. B, 383:395–415, (1992). arXiv:hep-th/9203018, doi:10.1016/0550-3213(92)90684-4.
* [24] D. Klemm and M. Nozawa. Supersymmetry of the C-metric and the General Plebanski-Demianski Solution. JHEP, 05:123, (2013). arXiv:1303.3119, doi:10.1007/JHEP05(2013)123.
* [25] M.M. Caldarelli and D. Klemm. Supersymmetry of Anti-de Sitter Black Holes. Nucl. Phys. B, 545:434–460, (1999). arXiv:hep-th/9808097, doi:10.1016/S0550-3213(98)00846-3.
* [26] N. Alonso-Alberca, P. Meessen, and T. Ortin. Supersymmetry of Topological Kerr-Newman-Taub-NUT-adS Spacetimes. Class. Quant. Grav., 17:2783–2798, (2000). arXiv:hep-th/0003071, doi:10.1088/0264-9381/17/14/312.
* [27] D. Hilbert. Die Grundlagen der Physik. 1. [Foundations of Physics. 1]. Gott. Nachr., 27:395–407, (1915).
* [28] B. Carter. Killing Horizons and Orthogonally Transitive Groups in Space-time. J. Math. Phys., 10:70–81, (1969). doi:10.1063/1.1664763.
* [29] H. Weyl. Zur Gravitationstheorie [On the Theory of Gravity]. Annalen der Physik, 359:117–145, (1917). doi:10.1002/andp.19173591804.
* [30] T. Lewis. Some Special Solutions of the Equations of Axially Symmetric Gravitational Fields. Proc. Roy. Soc. Lond. A, 136:179–192, (1932). doi:10.1098/rspa.1932.0073.
* [31] A. Papapetrou. Eine Rotationssymmetrische Lösung in der Allgemeinen Relativitätstheorie [A Rotationally Symmetric Solution in General relativity]. Annalen der Physik, 447:309–315, (1953). doi:10.1002/andp.19534470412.
* [32] A. Papapetrou. Champs Gravitationnels Stationnaires à Symétrie axiale [Stationary Gravitational Fields with Axial Symmetry]. Ann. Inst. Henri Poincaré, 4:83–105, (1966).
* [33] M. Heusler. Black Hole Uniqueness Theorems. Cambridge Lecture Notes in Physics. Cambridge University Press, (1996). doi:10.1017/CBO9780511661396.
* [34] S. Chandrasekhar. The Mathematical Theory of Black Holes. Oxford University Press, (1983).
* [35] O. Aharony, M. Fabinger, G.T. Horowitz, and E. Silverstein. Clean Time-Dependent String Backgrounds from Bubble Baths. JHEP, 07:007, (2002). arXiv:hep-th/0204158, doi:10.1088/1126-6708/2002/07/007.
* [36] S.W. Hawking. Black Hole Explosions ? Nature, 248:30–31, (1974). doi:10.1038/248030a0.
* [37] J.D. Bekenstein. Black Holes and the Second Law. Lett. Nuovo Cim., 4:737–740, (1972). doi:10.1007/BF02757029.
* [38] S.W. Hawking. Particle Creation by Black Holes. Commun. Math. Phys., 43:199–220, (1975). [Erratum: Hawking, S.W. , Commun. Math. Phys. 46:206 (1976)]. doi:10.1007/BF02345020.
* [39] J.M. Bardeen, B. Carter, and S.W. Hawking. The Four Laws of Black Hole Mechanics. Commun. Math. Phys., 31:161–170, (1973). doi:10.1007/BF01645742.
* [40] J. Ehlers. Konstruktionen und Charakterisierung von Losungen der Einsteinschen Gravitationsfeldgleichungen [Constructions and Characterization of Solutions to Einstein’s gravitational Field Equations]. Ph.D. Thesis, Hamburg University, (1957).
* [41] B.K. Harrison. New Solutions of the Einstein-Maxwell Equations from Old. Journal of Mathematical Physics, 9:1744–1752, (1968). doi:10.1063/1.1664508.
* [42] H. Stephani, D. Kramer, et al. Exact Solutions of Einstein’s Field Equations. Cambridge Monographs on Mathematical Physics. Cambridge Univ. Press, (2003). doi:10.1017/CBO9780511535185.
* [43] G. Neugebauer and D. Kramer. Eine Methode zur Konstruktion stationärer Einstein-Maxwell-Felder [A Method for Constructing Stationary Einstein-Maxwell fields]. Annalen der Physik, 479:62–71, (1969). doi:10.1002/andp.19694790108.
* [44] W. Kinnersley. Generation of Stationary Einstein-Maxwell Fields. J. Math. Phys., 14:651–653, (1973). doi:10.1063/1.1666373.
* [45] M. Astorino. Charging Axisymmetric Space-times with Cosmological Constant. JHEP, 06:086, (2012). arXiv:1205.6998, doi:10.1007/JHEP06(2012)086.
* [46] M.S. Illy. Accelerated Reissner–Nordström Black Hole in a Swirling, Magnetic Universe. Bachelor Thesis, Milan U., (2023). arXiv:2312.14995.
* [47] J. Barrientos, A. Cisterna, et al. Mixing "Magnetic” and "Electric” Ehlers–Harrison Transformations: The Electromagnetic Swirling Spacetime and Novel Type I Backgrounds. (2024). arXiv:2312.17347.
* [48] F.J. Ernst. Black Holes in a Magnetic Universe. J. Math. Phys., 17:54–56, (1976). doi:10.1063/1.522781.
* [49] A.H. Taub. Empty Space-times Admitting a Three Parameter Group of Motions. Annals Math., 53:472–490, (1951). doi:10.2307/1969567.
* [50] E. Newman, L. Tamburino, and T. Unti. Empty Space Generalization of the Schwarzschild Metric. J. Math. Phys., 4:915, (1963). doi:10.1063/1.1704018.
* [51] H. Reissner. Über die Eigengravitation des Elektrischen Feldes nach der Einsteinschen Theorie [About the Self-Gravity of the Electric Field According to Einstein’s theory]. Annalen der Physik, 355:106–120, (1916). doi:10.1002/andp.19163550905.
* [52] G. Nordström. On the Energy of the Gravitation Field in Einstein’s Theory. Koninklijke Nederlandse Akademie van Wetenschappen Proceedings Series B Physical Sciences, 20:1238–1245, (1918).
* [53] C. Reina and A. Treves. NUT-like Generalization of Axisymmetric Gravitational Fields. Journal of Mathematical Physics, 16:834–835, (1975). doi:10.1063/1.522614.
* [54] G.W. Gibbons and D.L. Wiltshire. Space-Time as a Membrane in Higher Dimensions. Nucl. Phys. B, 287:717–742, (1987). arXiv:hep-th/0109093, doi:10.1016/0550-3213(87)90125-8.
* [55] G.W. Gibbons and C.A.R. Herdeiro. The Melvin Universe in Born-Infeld Theory and Other Theories of Nonlinear Electrodynamics. Class. Quant. Grav., 18:1677–1690, (2001). arXiv:hep-th/0101229, doi:10.1088/0264-9381/18/9/305.
* [56] R. Emparan and M. Gutperle. From p-branes to Fluxbranes and Back. JHEP, 12:023, (2001). arXiv:hep-th/0111177, doi:10.1088/1126-6708/2001/12/023.
* [57] M. Astorino. Removal of Conical singularities From Rotating C-metrics and Dual CFT Entropy. JHEP, 10:074, (2022). arXiv:2207.14305, doi:10.1007/JHEP10(2022)074.
* [58] G.W. Gibbons, A.H. Mujtaba, and C.N. Pope. Ergoregions in Magnetised Black Hole Spacetimes. Class. Quant. Grav., 30:125008, (2013). arXiv:1301.3927, doi:10.1088/0264-9381/30/12/125008.
* [59] P.A.M. Dirac. Quantised Singularities in the Electromagnetic Fiels. Proc. Roy. Soc. Lond. A, 133:60–72, (1931). doi:10.1098/rspa.1931.0130.
* [60] P.A.M. Dirac. The Theory of Magnetic Poles. Phys. Rev., 74:817–830, (1948). doi:10.1103/PhysRev.74.817.
* [61] K. Gödel. An Example of a New Type of Cosmological Solutions of Einstein’s Field Equations of Gravitation. Rev. Mod. Phys., 21:447–450, (1949). doi:10.1103/RevModPhys.21.447.
* [62] C.W. Misner. The Flatter Regions of Newman, Unti and Tamburino’s Generalized Schwarzschild Space. J. Math. Phys., 4:924–938, (1963). doi:10.1063/1.1704019.
* [63] D. Astefanesei, R.B. Mann, and E. Radu. Nut Charged Space-times and Closed Timelike Curves on the Boundary. JHEP, 01:049, (2005). arXiv:hep-th/0407110, doi:10.1088/1126-6708/2005/01/049.
* [64] M. Alcubierre. The Warp Drive: Hyperfast Travel Within General Relativity. Class. Quant. Grav., 11:L73–L77, (1994). arXiv:gr-qc/0009013, doi:10.1088/0264-9381/11/5/001.
* [65] M. Astorino. Magnetised Kerr/CFT Correspondence. Phys. Lett. B, 751:96–106, (2015). arXiv:1508.01583, doi:10.1016/j.physletb.2015.10.017.
* [66] H.M. Siahaan. Magnetized Kerr-Taub-NUT Spacetime and Kerr/CFT Correspondence. Phys. Lett. B, 820:136568, (2021). arXiv:2102.04345, doi:10.1016/j.physletb.2021.136568.
* [67] G. Compère. The Kerr/CFT Correspondence and its Extensions. Living Rev. Rel., 15:11, (2012). arXiv:1203.3561, doi:10.1007/s41114-017-0003-2.
* [68] D.Z. Freedman and A.K. Das. Gauge Internal Symmetry in Extended Supergravity. Nucl. Phys. B, 120:221–230, (1977). doi:10.1016/0550-3213(77)90041-4.
* [69] E.S. Fradkin and M.A. Vasiliev. Unpublished. Lebedev Institute, Preprint 197, (1976).
* [70] P. van Nieuwenhuizen and N.P. Warner. Integrability conditions for Killing spinors. Commun. Math. Phys., 93:277–284, (1984). doi:10.1007/BF01223747.
|
[a]Atreya Acharyya
# Identifying multiwavelength counterparts to astrophysical neutrino events
Marcos Santander
###### Abstract
High-energy neutrinos originating in astrophysical sources should be
accompanied by gamma-rays at production. Depending on the properties of the
emission environment and the distance of the source to the Earth, these gamma-
rays may be observed directly, or through the detection of lower energy
photons that result from interactions with the intervening radiation fields.
In this work, we present an automated tool that aims at using data from the
_Fermi_ -Large Area Telescope to identify multiwavelength counterparts to
astrophysical neutrino events. The main goal of this tool is to enable prompt
follow-up observations with ground-based and space-based observatories in
order to help pinpoint the neutrino source.
## 1 Introduction
The extragalactic gamma-ray sky is dominated by blazars [1], a subclass of
radio-loud active galactic nuclei (AGN) powered by a central supermassive
black hole (SMBH), with relativistic jets pointed close to our line-of-sight.
The spectral energy distribution (SED) of a typical blazar comprises two
distinct peaks. While the first peak, occurring in the radio to the X-ray
regime, has been attributed to synchrotron emission from electrons and
positrons within the jet, the physical mechanisms responsible for the second
peak, produced in the X-ray to gamma-ray regime, is still a matter of debate.
Towards understanding this, the _Fermi_ -Large Area Telescope (LAT, [2]) is a
pair conversion telescope capable of detecting gamma-ray photons in the energy
range between 20 MeV to above 500 GeV. Primarily operating in survey mode, it
scans the entire sky every three hours.
Leptonic models [3, 4] attribute the high-energy peak of the SED to the
inverse Compton (IC) scattering of electrons off a source of seed photons,
either the same photons emitted through synchrotron emission (synchrotron
self-Compton [SSC] model) or photon populations external to the jet (external
Inverse Compton [EIC] model). On the other hand, lepto-hadronic models [5, 6]
suggest that the second peak may be a result of either proton synchrotron or
the decay of high-energy mesons produced in cosmic-ray interactions. A
hadronic component of the gamma-ray emission [7] would potentially make AGN
prime candidate sources of astrophysical neutrinos [8, 9].
Since 2016, IceCube has been broadcasting automatic real-time alerts for
potential astrophysical neutrino candidate events in order to allow for prompt
follow-up observations with ground-based and space-based observatories. The
alerts are issued by the Astrophysical Multimessenger Observatory Network
(AMON111https://www.amon.psu.edu (accessed on 06/06/2023), [10]) and
circulated using the Gamma-ray Coordinates Network
(GCN)222https://gcn.gsfc.nasa.gov (accessed on 06/06/2023), an open-source
platform created by NASA to receive and transmit alerts about astronomical
transient phenomena. One of the main goals of these alerts is to allow for
follow-up observations with other observatories, in the hope of observing
multimessenger events.
In this work, we present an automated tool that aims at using _Fermi_ -LAT
data to identify multiwavelength counterparts to astrophysical neutrino events
with ground-based and space-based observatories in order to help pinpoint the
neutrino source. More specifically, we want to know what is in the region of
interest (RoI) around the neutrino alert, whether the RoI is observable from a
ground-based observatory and whether anything interesting, for example a
gamma-ray flare, is occurring at that particular location. In Section 2, we
introduce the main components of the analysis and processing pipeline. In
Section 3, we consider a typical example of a neutrino alert, IC230506A-gold-
rev1, and discuss the primary outputs from the analysis tool. We summarise our
conclusions and discuss some plans for future work in Section 4.
## 2 Methodology
Figure 1: A flowchart of the automated pipeline illustrating how each IceCube
GCN alert is processed and analyzed before the results are mirrored on to the
web server (https://multimessenger.ua.edu/fermi/, accessed on 06/06/2023).
The automated pipeline, illustrated in Fig. 1, listens for incoming
AMON_ICECUBE_GOLD and AMON_ICECUBE_BRONZE event alerts coming in through the
GCN. These are individual single-neutrino events in the IceCube detector with
a neutrino energy in the sub-PeV to 1 PeV energy regime and are shown in Fig.
2 for the time interval between June 19, 2019 and June 6, 2023.
More specifically, the analysis pipeline extracts information from the GCN
alert related to the type of event (GOLD or BRONZE), the sky coordinates of
the alert along with the associated confidence interval, the date and time of
the alert in UTC, and also the corresponding _Run_ID_ and _Event_ID_. The main
components of the tool include performing an automatic Fermi-LAT analysis of
the neutrino RoI, calculating the visibilities for common follow-up
instruments and collecting multiwavelength archival data for known sources in
the RoI. The entire process is then repeated for multiple revisions of a
particular alert.
The tool then runs an automatic analysis of the Fermi-LAT photons detected
from within the RoI of the neutrino alert over a 30 day interval prior to each
individual event. Each analysis uses the Fermi Science Tools version
$11-05-03$ 333http://fermi.gsfc.nasa.gov/ssc/data/analysis/software (accessed
on 06/06/2023), FERMIPY version 1.0.1 444http://fermipy.readthedocs.io
(accessed on 06/06/2023) [11] in conjunction with the PASS 8 instrument
response functions [12]. The contributions from the isotropic and Galactic
diffuse backgrounds are modeled using the most recent templates for isotropic
and Galactic diffuse emission, iso_P8R3_SOURCE_V2_v1.txt and gll_iem_v07.fits
respectively. Sources in the 4FGL-DR3 catalog [13] within a radius of
$20^{\circ}$ from the best-fit location of each alert are included in the
model with their spectral parameters fixed to their catalog values. The
gtfindsrc routine is also applied to search for any additional point sources
not accounted for in the model. Any source found to have a test statistic (TS,
[14]) $\geq$ 9 (roughly corresponding to a significance of $\sim$ 3$\sigma$)
is permanently added to the model at the position of its highest TS value.
Moreover, the normalization factor for both the isotropic and Galactic diffuse
emission templates are left free to vary, along with the spectral
normalization of all modeled sources within the RoI. Finally, the spectral
shape parameters of all modeled sources within 3∘ of the alert are left free
to vary while those of the remaining sources are fixed to the values reported
in the 4FGL-DR3 catalog. A binned likelihood analysis is then performed in
order to obtain the spectral parameters best describing the model during the
period of observation, using a spatial binning of 0.1∘ pixel-1 and two energy
bins per decade. The results, including HTML, FITS files, and images of plots,
are mirrored on to the web server (https://multimessenger.ua.edu/fermi/,
accessed on 06/06/2023).
Figure 2: A skymap, in celestial coordinates, showing AMON_ICECUBE_GOLD events
(in yellow) and AMON_ICECUBE_ BRONZE events (in red) between June 19, 2019 and
June 6, 2023, from the web server (https://multimessenger.ua.edu/fermi/,
accessed on 06/06/2023).
## 3 Discussion
Figure 3: Left: Top panel: A skymap of the neutrino alert, also containing all
sources in the confidence region from the 4FGL-DR3 and 3FHL catalog. Bottom
panel: These sources, listed separately, with links to the source entry in
_Simbad_ , as well as multiwavelength archival data repositories. Right: The
diagnostic plots produced in this _Fermi_ -LAT data analysis. These include a
skymap of the residual significance and the excess photon counts obtained in
the analysis over the RoI centred on the location of the alert. The colour
scales correspond to the excess significance of each pixel in the RoI in
Gaussian $\sigma$ and the number of excess photons at each pixel in the RoI
respectively.
Figure 4: Left: Top panel: The results obtained from the _Fermi_ -LAT
analysis, including the TS, flux and spectrum for both known 4FGL sources, as
well as potential new sources. Middle panel: A skymap and lightcurve of the
GeV photons emitted from within the RoI. Bottom panel: An example of a _Fermi_
-LAT spectrum for the source 4FGL J0325.7+2225 in red. The data are binned
into two energy bins per decade, with individual bins having a TS $<$ 4
(roughly corresponding to a significance of $\sim$ 2$\sigma$) considered as
upper limits. Also shown for comparison is the 4FGL-DR3 best-fit spectrum.
Right: Top panel: An interactive skymap containing markers for nearby sources
in the 4FGL, 3FHL, and 2WHSP catalog. Bottom panel: A visibility plot of the
RoI with VERITAS.
In this Section, we consider a typical example of a neutrino alert,
IC230506A-gold-
rev1555https://gcn.gsfc.nasa.gov/notices_amon_g_b/137910_29871391.amon
(accessed on 06/06/2023), received on 2023-05-06 at 15:53:45 UTC. This was a
revised AMON_ICECUBE_GOLD type alert, with _Run_ID_ = 137910, _Event_ID_ =
29871391, and the associated sky coordinates are RA = $50.19^{\circ}$, Dec =
$21.06^{\circ}$, with a corresponding error radius of $3.12^{\circ}$. The
plots produced by the automated tool are shown in Fig. 3 and Fig. 4.
The output results include a skymap of the neutrino alert, also containing all
sources in the confidence region from the 12-year _Fermi_ -LAT catalog (4FGL-
DR3, [13]), as well as the third _Fermi_ -LAT catalog of sources significantly
detected in the 10 GeV – 2 TeV energy range (3FHL, [15]). These sources are
also listed separately in a Table containing their 4FGL and 3FHL names, sky
coordinates, distance from the neutrino alert and links to the source entry in
_Simbad_ [16], as well as multiwavelength archival data repositories including
the _Fermi_ -LAT Light Curve Repository [17] and the _Fermi_ All-sky
Variability Analysis (FAVA, [18]) Light Curve Repository.
Also shown in Fig. 3, are the set of diagnostic plots produced in this _Fermi_
-LAT data analysis. These are obtained for each individual alert and
subsequent revision and include a skymap showing the modelled distribution of
the gamma-ray photons obtained after removing all sources found to have a TS <
9 followed by applying the gtfindsrc routine over the RoI, as well as a skymap
of the residual significance and the excess photon counts obtained in the
analysis over the RoI centred on the location of the alert.
Fig. 4 shows more results obtained from the _Fermi_ -LAT analysis, including
both known 4FGL sources having TS $\geq 9$, as well as new sources obtained
from the gtfindsrc routine. This includes the TS obtained for each source, the
measured flux values in the energy range 100 MeV – 300 GeV along with the
corresponding 4FGL flux values for known sources. There is also a link to
corresponding plots, such as the one shown in Fig. 4, of the _Fermi_ -LAT
spectrum for each source, with the 4FGL-DR3 best-fit spectrum [13] also shown
for comparison for known sources. Furthermore, we also obtain a skymap and
lightcurve of the GeV photons, emitted from within the RoI and having a $\geq$
99 $\%$ probability of originating from each individual source, over the one
month observation period investigated.
Moreover, the tool also produces an interactive skymap containing markers for
nearby sources in three catalogs, namely the 4FGL, 3FHL, and 2WHSP [19]. Other
catalogs, for example the 2RXS [20] and BZCat [21] plus a dynamic HiPS search
in _Simbad_ , can also be enabled. An additional _Simbad_ search can be made
to show a selection of objects detected in radio, X-rays, gamma-rays and
galaxies near the best fit position of the alert. The background sky can also
be changed to optical (Mellinger survey, [22]) using the radio buttons below
the image. Finally, we also produce visibility plots of the RoI for the four
ground-based instruments, namely MAGIC, H.E.S.S., VERITAS and HAWC, over the
time interval just after the alert, in order to help enable prompt follow-up
observations.
## 4 Conclusions and Future Work
In this work, we introduce an automated tool that aims at using _Fermi_ -LAT
data to identify multiwavelength counterparts to astrophysical neutrino events
and enable prompt follow-up observations with ground-based and space-based
observatories in order to help pinpoint the neutrino source. After discussing
the main components of the the analysis and processing pipeline, we walk-
through the primary outputs from the analysis tool for a typical example of a
neutrino alert, IC230506A-gold-rev1.
It should be noted that this is an early version of the automated tool with
plenty of scope for even further improvement. This includes, for example,
adding support for 1-year and full-mission _Fermi_ -LAT analysis alongside the
month long time period currently analyzed, gathering more multiwavelength data
and resources to enable SED construction, and finally improving compatibility
with the GCN Kafka Client setup 666https://gcn.nasa.gov/docs/client (accessed
on 06/06/2023).
## Acknowledgments
A.A. and M.S. acknowledge support through NASA grants 80NSSC22K1515,
80NSSC22K0950, 80NSSC20K1587, 80NSSC20K1494 and NSF grant PHY-1914579.
## References
* [1] C.D. Dermer and B. Giebels, _Active galactic nuclei at gamma-ray energies_ , _C R Phys._ 17 (2016) 594 [1602.06592].
* [2] W.B. Atwood, A.A. Abdo, M. Ackermann, W. Althouse, B. Anderson, M. Axelsson et al., _The Large Area Telescope on the Fermi Gamma-Ray Space Telescope Mission_ , _The Astrophysical Journal_ 697 (2009) 1071 [0902.1089].
* [3] R. Blandford and A. Levinson, _Pair Cascades in Extragalactic Jets. I. Gamma Rays_ , _The Astrophysical Journal_ 441 (1995) 79.
* [4] M. Georganopoulos, F.A. Aharonian and J.G. Kirk, _External Compton emission from relativistic jets in Galactic black hole candidates and ultraluminous X-ray sources_ , _The Astronomy and Astrophysics Journal_ 388 (2002) L25 [astro-ph/0110379].
* [5] G.E. Romero, D.F. Torres, M.M. Kaufman Bernadó and I.F. Mirabel, _Hadronic gamma-ray emission from windy microquasars_ , _The Astronomy and Astrophysics Journal_ 410 (2003) L1 [astro-ph/0309123].
* [6] M. Böttcher, A. Reimer, K. Sweeney and A. Prakash, _Leptonic and Hadronic Modeling of Fermi-detected Blazars_ , _The Astrophysical Journal_ 768 (2013) 54 [1304.0605].
* [7] K. Mannheim, _The proton blazar._ , _The Astronomy and Astrophysics Journal_ 269 (1993) 67 [astro-ph/9302006].
* [8] K. Mannheim, _High-energy neutrinos from extragalactic jets_ , _Astroparticle Physics_ 3 (1995) 295.
* [9] F. Halzen and E. Zas, _Neutrino Fluxes from Active Galaxies: A Model-Independent Estimate_ , _The Astrophysical Journal_ 488 (1997) 669 [astro-ph/9702193].
* [10] M.W.E. Smith, D.B. Fox, D.F. Cowen, P. Mészáros, G. Tešić, J. Fixelle et al., _The Astrophysical Multimessenger Observatory Network (AMON)_ , _Astroparticle Physics_ 45 (2013) 56 [1211.5602].
* [11] M. Wood, R. Caputo, E. Charles, M. Di Mauro, J. Magill, J.S. Perkins et al., _Fermipy: An open-source Python package for analysis of Fermi-LAT Data_ , in _35th International Cosmic Ray Conference (ICRC2017)_ , vol. 301 of _International Cosmic Ray Conference_ , p. 824, Jan., 2017 [1707.09551].
* [12] W. Atwood, A. Albert, L. Baldini, M. Tinivella, J. Bregeon, M. Pesce-Rollins et al., _Pass 8: Toward the Full Realization of the Fermi-LAT Scientific Potential_ , _arXiv e-prints_ (2013) arXiv:1303.3514 [1303.3514].
* [13] Fermi-LAT collaboration, :, S. Abdollahi, F. Acero, L. Baldini, J. Ballet et al., _Incremental Fermi Large Area Telescope Fourth Source Catalog_ , _arXiv e-prints_ (2022) arXiv:2201.11184 [2201.11184].
* [14] J.R. Mattox, D.L. Bertsch, J. Chiang, B.L. Dingus, S.W. Digel, J.A. Esposito et al., _The Likelihood Analysis of EGRET Data_ , _The Astrophysical Journal_ 461 (1996) 396.
* [15] M. Ajello, W.B. Atwood, L. Baldini, J. Ballet, G. Barbiellini, D. Bastieri et al., _3FHL: The Third Catalog of Hard Fermi-LAT Sources_ , _Astrophysical Journal Supplement Series_ 232 (2017) 18 [1702.00664].
* [16] M. Wenger, F. Ochsenbein, D. Egret, P. Dubois, F. Bonnarel, S. Borde et al., _The SIMBAD astronomical database. The CDS reference database for astronomical objects_ , _American Association of Pharmaceutical Scientists_ 143 (2000) 9 [astro-ph/0002110].
* [17] S. Abdollahi, M. Ajello, L. Baldini, J. Ballet, D. Bastieri, J. Becerra Gonzalez et al., _The Fermi-LAT Lightcurve Repository_ , _Astrophysical Journal Supplement Series_ 265 (2023) 31 [2301.01607].
* [18] S. Abdollahi, M. Ackermann, M. Ajello, A. Albert, L. Baldini, J. Ballet et al., _The Second Catalog of Flaring Gamma-Ray Sources from the Fermi All-sky Variability Analysis_ , _The Astrophysical Journal_ 846 (2017) 34 [1612.03165].
* [19] Y.L. Chang, B. Arsioli, P. Giommi and P. Padovani, _2WHSP: A multi-frequency selected catalogue of high energy and very high energy $\gamma$-ray blazars and blazar candidates_, _The Astronomy and Astrophysics Journal_ 598 (2017) A17 [1609.05808].
* [20] T. Boller, M.J. Freyberg, J. Trümper, F. Haberl, W. Voges and K. Nandra, _Second ROSAT all-sky survey (2RXS) source catalogue_ , _The Astronomy and Astrophysics Journal_ 588 (2016) A103 [1609.09244].
* [21] E. Massaro, P. Giommi, C. Leto, P. Marchegiani, A. Maselli, M. Perri et al., _Roma-BZCAT: a multifrequency catalogue of blazars_ , _The Astronomy and Astrophysics Journal_ 495 (2009) 691 [0810.2206].
* [22] A. Mellinger, _A Color All-Sky Panorama Image of the Milky Way_ , _Publications of the Astronomical Society of the Pacific_ 121 (2009) 1180 [0908.4360].
|
# Interplay of nonlocality and incompatibility breaking qubit channels
Swati Kumari<EMAIL_ADDRESS>Department of Physics and Center for
Quantum Frontiers of Research & Technology (QFort), National Cheng Kung
University, Tainan 701, Taiwan Javid Naikoo<EMAIL_ADDRESS>Centre
for Quantum Optical Technologies, Centre of New Technologies, University of
Warsaw, Banacha 2c, 02-097 Warsaw, Poland Sibasish Ghosh
<EMAIL_ADDRESS>Optics and Quantum Information Group, The Institute of
Mathematical Sciences, HNBI, CIT Campus, Taramani, Chennai 600113, India A.
K. Pan<EMAIL_ADDRESS>Department of physics, Indian Institute of
Technology Hyderabad Kandi,Telengana, India
###### Abstract
Incompatibility and nonlocality are not only of foundational interest but also
act as important resources for quantum information theory. In the Clauser-
Horne-Shimony-Holt (CHSH) scenario, the incompatibility of a pair of
observables is known to be equivalent to Bell nonlocality. Here, we
investigate these notions in the context of qubit channels. The Bell-CHSH
inequality has a greater perspective—compared to any genuine tripartite
nonlocality scenario—while determining the interplay between nonlocality
breaking qubit channels and incompatibility breaking qubit channels. In the
Bell-CHSH scenario, we prove that if the conjugate of a channel is
incompatibility breaking, then the channel is itself nonlocality breaking and
vice versa. However, this equivalence is not straightforwardly generalized to
multipartite systems, due to the absence of an equivalence relation between
incompatibility and nonlocality in the multipartite scenario. We investigate
this relation in the tripartite scenario by considering some well-known states
like Greenberger-Horne-Zeilinger and $W$ states and using the notion of Mermin
and Svetlichny nonlocality. By subjecting the parties in question to unital
qubit channels, we identify the range of state and channel parameters for
which incompatibility coexists with nonlocality. Further, we identify the set
of unital qubit channels that is Mermin or Svetlichny nonlocality breaking
irrespective of the input state.
## I Introduction
Nonlocality is one of the most profound notions in quantum mechanics Bell
(1964) and is often discussed in conjunction with incompatibility of
observables. Recent developments in quantum information theory have found
nonlocality a useful phenomenon underpinning many advantages afforded by
various quantum information processing tasks Brunner _et al._ (2014).
Nonlocality can also be considered as a potential quantum resource for
information processing, such as in developing quantum protocols to reduce the
amount of communication needed in certain computational tasks Brunner _et
al._ (2014) and providing secure quantum communications Scarani and Gisin
(2001); Ekert (1991). Incompatibility, like nonlocality, is not merely of
theoretical interest but of practical utility, for example, in order to
explore the advantage of entanglement shared by two parties in a cryptography
task, each party needs to carry out measurements that are incompatible, in the
sense that these cannot be carried out simultaneously by a single measurement
device. Incompatibility should not be confused with noncommutativity or the
related concept of uncertainty principle. The notion of incompatibility is
best understood in terms of joint measurability Heinosaari _et al._ (2016). A
collection of quantum measurements is jointly measurable, if it can be
simulated by a single common quantum measurement device. If such a single
common device cannot be constructed by a given set of quantum measurements, it
then enables the set to be used as a quantum resource. This was first noted in
Wolf _et al._ (2009) in the context of Clauser-Horne-Shimony-Holt (CHSH)
inequalities and later in the Einstein-Podolsky-Rosen steering, which is more
explicit, when incompatibility appears as a quantum resource. Incompatibility
is necessary and sufficient for the violation of the steering inequalities
Quintino _et al._ (2014); Uola _et al._ (2015). The relation between
incompatibility and contextuality has also been studied in references Kochen
and Specker (1975); Budroni _et al._ (2022). Further, a set of observables
that is pairwise incompatible, but not triplewise can violate the Liang-
Spekkens-Wiseman noncontextuality inequality Liang _et al._ (2011). Recently,
the connection between steerability and measurement incompatibility was
studied in Ku _et al._ (2022a) in the context of the so-called steerability
equivalent observables. Thus, both nonlocality and incompatibility can be
considered as quantum resources whose understanding is of utmost importance in
view of emerging quantum technologies.
The interplay of nonlocality and incompatibility has been a subject matter of
various studies. It is well known that any incompatible local measurements,
performed by the constituent parties of a system, lead to the violation of
Bell inequality provided they share a pure entangled state Bell (1964);
Brunner _et al._ (2014). Absence of either of them (i.e., entanglement or
incompatibility) will not allow the system to exhibit nonlocality. It is
important to mention here that the notion of quantum nonlocality without
entanglement has been proposed in Bennett _et al._ (1999) which is different
from Bell nonlocality Bell (1964) and amounts to the inability of
discriminating a set of product states by local operations and classical
communication, while mutual orthogonality of the states assures their perfect
global discrimination.
Further, for any pair of dichotomic incompatible observables, there always
exists an entangled state which enables the violation of a Bell inequality
Wolf _et al._ (2009). The relationship of incompatibility and nonlocality is
sensitive to the dimension of the system; for example, increasing the
dimension beyond $2$, the incompatible observables do not necessarily lead to
the violation of Bell-type inequalities, implying that the measurement
incompatibility can not guarantee nonlocality in general Bene and Vértesi
(2018); Hirsch _et al._ (2018). Here, we probe the interplay between
incompatibility and nonlocality in the tripartite case by using the well-known
Mermin and Svetlichny inequalities Svetlichny (1987). The Svetlichny
inequality, unlike the Mermin inequality, is a genuine measure of nonlocality
that assumes nonlocal correlations between two parties which are locally
related to a third party and is known to provide a suitable measure to detect
tripartite nonlocality for $W$ and Greenberger-Horne-Zeilinger (GHZ) classes
of states Ajoy and Rungta (2010). We refer the interested reader to Bancal
_et al._ (2013); Brunner _et al._ (2014) for various facets of the
multipartite nonlocality.
The extent to which a system can exhibit nonlocal correlations is also
sensitive to its interaction with the ambient environment. Such interaction is
usually accompanied with a depletion of various quantum features like
coherence, entanglement and nonlocality. The reduced dynamics of the system in
such cases is given by completely positive and trace preserving maps, also
known as quantum channels (QCs). On the other hand, the action of conjugate
channels on projective measurements turns them into unsharp positive operator-
valued measures (POVMs) which may be biased, in general. In light of the above
discussion, a study of the open system effects on the interplay of nonlocality
and incompatibility naturally leads to the notions of nonlocality breaking and
incompatibility breaking quantum channels Pal and Ghosh (2015); Heinosaari
_et al._ (2016). A nonlocality breaking channel (NBC) can be defined as a
channel which when applied to a system (or part of it) leads to a state which
is local Pal and Ghosh (2015), while the incompatibility breaking channel
(IBC) is the one that turns incompatible observables into compatible ones
Heinosaari _et al._ (2015); Heinosaari and Miyadera (2017). An IBC that
renders any set of $n$ incompatible observables compatible would be denoted by
$n$-IBC. The notion of the NBC has been introduced in a similar spirit of
well-studied entanglement breaking channels Horodecki _et al._ (2003). Every
entanglement breaking channel is nonlocality breaking but the converse is not
true. As an example, the qubit depolarizing channel
$\mathcal{E}\big{(}\rho\big{)}:=p\big{(}\rho\big{)}+(1-p)\mathbf{I}/2$ is CHSH
nonlocality breaking for all $\frac{1}{3}\leq p\leq\frac{1}{2}$, but not
entanglement breaking Pal and Ghosh (2015). Hence, based on this
classification, nonlocality and entanglement emerge as different resources.
The equivalence of the steerability breaking channels and the incompatibility
breaking channels was reported in Kiukas _et al._ (2017) and CHSH nonlocality
breaking channels were shown to be a strict subset of the steerability
breaking channels Ku _et al._ (2022b). The connection between Bell
nonlocality and incompatibility of two observable is well understood, however,
the question of the equivalence between NBC Pal and Ghosh (2015) and IBC
Heinosaari and Miyadera (2017) is rarely discussed. This motivates us to
explore the relation between CHSH nonlocality breaking channels (CHSH-NBC) and
$2$-IBC, such that the action of one may be replaced by the other. The
tripartite nonlocality has much richer and complex structure and less is known
about its synergy with incompatibility as compared to its bipartite
counterpart. Mermin inequality assumes local-realistic correlations among all
the three qubits; hence a violation would be a signature of the tripartite
nonlocality shared among the qubits. However, biseparable states were shown to
also violate the Mermin inequality Collins _et al._ (2002); Scarani and Gisin
(2001). This motivated Svetlichny to introduce the notion of genuine
tripartite nonlocality Svetlichny (1987) and provide a set of inequalities
sufficient to witness it. We make use of these notions of _absolute_ and
_genuine_ nonlocality to figure out the ranges of state and channels
parameters in which NBC and $2$-IBC coexist.
The paper is organized as follows. In Sec. II, we revisit some basic notions
and definitions used in this paper. Section III is devoted to results and
their discussion where we prove an equivalence between NBCs and $2$-IBCs in
the CHSH scenario. This is followed by an analysis of these notions in the
tripartite scenario, where we identify the state and channel parameters in
which NBCs and 2-IBCs co-exist. A conclusion is given in Sec. IV.
## II Preliminaries
In this section, we discuss the notion of incompatibility in the context of
observables and quantum channels and look at specific cases of bipartite and
tripartite scenarios.
### II.1 Incompatibility
#### II.1.1 Incompatibility of observables
A finite collection of observables $A_{1},\dots,A_{n}$ associated with the
respective outcome spaces $\Omega_{A_{1}},\dots,\Omega_{A_{n}}$, is said to be
compatible (or jointly measurable) if there exists a joint observable $G$,
defined over the product outcome space
$\Omega_{A_{1}}\times\cdots\times\Omega_{A_{n}}$, such that for all
$X_{1}\subset\Omega_{A_{1}},\dots,X_{n}\subset\Omega_{A_{n}}$, the following
marginal relations hold Busch (1986):
$\displaystyle\sum_{a_{i},i=1,\dots,n;i\neq
k}G(a_{1},\dots,a_{n})=A_{k}(a_{k})$ (1)
where $a_{k}$ is the outcome associated to observable $A_{k}$ and the
summation is carried out over all outcomes $a_{i}$ except for $i=k$. The
notion of the incompatibility of observables can be illustrated by a simple
example of Pauli matrices $\sigma_{x}$ and $\sigma_{z}$ which are noncommuting
and can not be measured jointly. However, consider the unsharp observables
$S_{x}(\pm)=\frac{1}{2}(\mathbf{I}\pm\frac{1}{\sqrt{2}}\sigma_{x})$ and
$S_{z}(\pm)=\frac{1}{2}(\mathbf{I}\pm\frac{1}{\sqrt{2}}\sigma_{z})$, with
$\mathbf{I}$ being the $2\times 2$ identity matrix. The joint observable
$G(i,j)=\frac{1}{4}(\mathbf{I}+\frac{i}{\sqrt{2}}\sigma_{x}+\frac{j}{\sqrt{2}}\sigma_{z})$
with $i,j=\pm 1$ jointly determines the probabilities of generalized
measurements $S_{x}$ and $S_{z}$, since the later can be obtained as marginals
$S_{x}(\pm)=\sum_{j}G(\pm,j)$ and $S_{z}(\pm)=\sum_{i}G(i,\pm)$.
#### II.1.2 Incompatibility breaking quantum channel
A QC, in the Schrödinger picture, is a completely positive trace preserving
map
$\mathcal{E}:\mathcal{L}({\mathcal{H^{A}}})\rightarrow{\mathcal{L}}({\mathcal{H^{B}}})$,
where ${\mathcal{L}}({\mathcal{H}^{i}})$ is the set of bounded linear
operators on Hilbert space $\mathcal{H}^{i}$ ($i=A,B$). One may write the
operator sum representation Nielsen and Chuang (2010)
$\displaystyle\rho^{\prime}=\mathcal{E}\big{(}\rho\big{)}=\sum_{j=1}^{n}K_{j}\rho
K_{j}^{\dagger}$ (2)
where $K_{i}$ are known as Kraus operators satisfying the completeness
relation $\sum_{j}K^{\dagger}_{j}K_{j}=\mathbf{I}$. The QCs which map the
identity operator to itself, i.e., $\mathcal{E}(\mathbf{I})=\mathbf{I}$, are
known as unital QCs. A quantum channel $\mathcal{E}$ in the Schrödinger
picture acting on quantum state $\rho$ can be thought of as conjugate channel
$\mathcal{E}^{*}$ acting on observable $A$ through the following duality
relation
$\Tr\left[\mathcal{E}(\rho)A\right]=\Tr\left[\rho~{}\mathcal{E}^{*}(A)\right].$
(3)
###### Definition 1.
A quantum channel $\mathcal{E}$ is said to be incompatibility breaking if the
outputs $\mathcal{E}^{*}(A_{1}),\dots,\mathcal{E}^{*}(A_{n})$ are compatible
for any choice of input observables $A_{1},\dots,A_{n}$ $(n\geq 2)$.
If a channel $\mathcal{E}$ breaks the incompatibility of every class of $n$
observables, it is said to be $n-$ IBC. For example, a channel would be
$2-{\rm IBC}$ if it breaks the incompatibility of a pair of observables. As an
example, the white noise mixing channel $\mathcal{W}_{\eta}$ is described as
$\rho^{\prime}=\mathcal{W}_{\eta}[\rho]=\eta\rho+(1-\eta)\frac{\mathbf{I}}{d}.$
(4)
Here, $\eta$ is the channel parameter, $d$ is the dimension of the underlying
Hilbert space. This channel is $n$-incompatibility breaking for all
$0\leq\eta\leq\frac{n+d}{n(d+1)}$. With $d=2$, $\mathcal{W}_{\eta}$ is $2-{\rm
IBC}$ and $3-{\rm IBC}$ for $\eta\leq 0.66$ and $0.55$, respectively
Heinosaari and Miyadera (2017).
#### II.1.3 Incompatibility of generalized spin-observables
Consider a spin observable $A=\hat{a}\cdot\vec{\sigma}$ with projectors
$P_{\pm}(\vec{a})=\frac{1}{2}(\mathbf{I}\pm A)$, with $|\vec{a}|=1$. In
presence of a noise channel $\mathcal{E}$, these projectors are mapped to the
noise induced POVM $P_{\pm}(\vec{\alpha})$, by the transformation
$A\xrightarrow[]{\mathcal{E}^{*}}A^{(x,\eta)}=x\mathbf{I}+\eta\hat{a}\cdot\vec{\sigma}.$
(5)
Here, $x$ and $\eta$ characterize the bias and sharpness such that un-biased
projective measurements correspond to $x=0$ and $\eta=1$. Two biased and
unsharp observables $A^{(x,\eta)}$ and $B^{(y,\xi)}$ are jointly measurable if
Yu _et al._ (2010); Heinosaari (2020)
$\left(1-S(x,\eta)^{2}-S(y,\xi)^{2}\right)\left(1-\frac{\eta^{2}}{S(x,\eta)^{2}}-\frac{\xi^{2}}{S(y,\xi)^{2}}\right)\leq\left(\vec{\eta}\cdot\vec{\xi}-xy\right),$
(6)
where $\vec{\eta}=\eta\hat{a}$ and $\vec{\xi}=\xi\hat{b}$ are the unsharpness
parameters, with
$S(p,q)=\frac{1}{2}\left(\sqrt{\left(1+p\right)^{2}-q^{2}}+\sqrt{\left(1-p\right)^{2}-q^{2}}\right).$
(7)
For unbiased observables $A^{(0,\eta)}$ and $B^{(0,\xi)}$, the necessary and
sufficient condition for compatibility simplifies to the following
$\displaystyle|\vec{\eta}+\vec{\xi}|+|\vec{\eta}-\vec{\xi}|\leq 2.$ (8)
When both the observables are subjected to identical noise channel i.e.,
$\xi=\eta$ and $\vec{\eta}$ and $\vec{\xi}$ are perpendicular, then the
pairwise joint measurability condition Eq. (8) becomes
$\eta\leq\frac{1}{\sqrt{2}}.$ (9)
This provides the condition for the incompatibility breaking of two
observables and the corresponding (unital) channel is called an
incompatibility breaking channel ($2-{\rm IBC}$).
### II.2 Nonlocality
#### II.2.1 Bipartite nonlocality: CHSH inequality
Consider the scenario of two spatially separated qubits with observables
$\hat{A}_{i}=\sum_{k=1}^{3}{\hat{a_{ik}}}\cdot{\hat{\sigma_{k}}}$ and
$\hat{B}_{j}=\sum_{l=1}^{3}{\hat{b_{jl}}}\cdot{\hat{\sigma_{l}}}$ acting on
each qubit respectively, where $\hat{a_{ik}}$ and $\hat{b_{jl}}$ are unit
vectors in $\mathbb{R}^{3}$; $i,j=1,2$; and the $\hat{\sigma}_{i}$’s are spin
projection operators. The Bell operator associated with the CHSH inequality
has the form
$\hat{\mathcal{B}}=\hat{A}_{1}\otimes(\hat{B}_{1}+\hat{B}_{2})+\hat{A}_{2}\otimes(\hat{B}_{1}-\hat{B}_{2}),$
(10)
Then the Bell-CHSH inequality, for any state $\rho$ is
$\Tr\left[\rho~{}\hat{\mathcal{B}}(\hat{A}_{1},~{}\hat{A}_{2},~{}\hat{B}_{1},~{}\hat{B}_{2})\right]\leq
2.$ (11)
where the observables on Alice’s and Bob’s side are pairwise compatible. The
violation of the above inequality (11) is sufficient to justify the
nonlocality of the quantum state. Since incompatible observables acting on
entangled particles enable nonlocality, thus incompatibility is necessary for
violation of (11).
The essence of Eq. (3) is that the effect of noise as decoherence of the
nonlocal resource $\rho$, can be interpreted as distortion of Alice’s local
measurement resource (incompatibility). This fact was exploited by Pal and
Ghosh Pal and Ghosh (2015) to introduce the notion of CHSH-NBC defined below.
###### Definition 2.
Any qubit channel
$\mathcal{E}:\mathcal{E}({\mathcal{H}}^{i})\rightarrow{\mathcal{E}}({\mathcal{H}}^{i})$
is said to be ${\rm IBC}$ if applying on one side of (arbitrary) bipartite
state $\rho_{AB}$, it produces a state
$\rho^{\prime}_{AB}=(\mathbf{I}\otimes\mathcal{E})(\rho_{AB})$ which satisfies
the Bell-CHSH inequality (11).
This means that for any choice of POVMs $\\{\pi_{a|x}^{A}\\}$ and
$\\{\pi_{b|y}^{B}\\}$ on subsystems $A$ and $B$, respectively, there exist
conditional distributions $P(a|x,\lambda)$ and $P(b|y,\lambda)$ and shared
variable $\lambda$, such that
$\displaystyle\Tr\left[(\pi_{a|x}^{A}\otimes\pi_{b|y}^{B})\sigma^{AB}\right]=\int
d\lambda~{}p(\lambda)~{}P(a|x,\lambda)~{}P(b|y,\lambda).$ (12)
The violation of the Bell inequalities in the measurement statistics
$P(ab|xy,\lambda)=\Tr[(\pi_{a|x}^{A}\otimes\pi_{b|y}^{B})\sigma^{AB}]$
indicates that $\sigma^{AB}$ is not a local state. A unital channel is
particularly important as it breaks the nonlocality for any state when it does
so for maximally entangled states Pal and Ghosh (2015).
#### II.2.2 Tripartite nonlocality: Svelitchny inequality
In a tripartite Bell scenario, with Alice, Bob, and Charlie performing
measurements $A$, $B$ and $C$ having outcome $a$, $b$ and $c$ respectively, if
the joint correlations can be written as
$\displaystyle P(abc|ABC)$
$\displaystyle=\sum_{l}c_{l}P_{l}(a|A)P_{l}(b|B)P_{l}(c|C)$ (13)
with $0\leq c_{l}\leq 1$ and $\sum_{l}c_{l}=1$, then they are local Mermin
(1990); Brunner _et al._ (2014). Svetlichny Svetlichny (1987) proposed the
hybrid local nonlocal form of probability correlations
$\displaystyle P(abc|ABC)$
$\displaystyle=\sum_{l}c_{l}P_{l}(ab|AB)P_{l}(c|C)+\sum_{m}c_{l}P_{m}(ac|AC)P_{m}(b|B)$
$\displaystyle+\sum_{n}c_{l}P_{n}(bc|C)P_{n}(a|A),$ (14)
with $0\leq c_{l},c_{m},c_{n}\leq 1$ and
$\sum_{l}c_{l}+\sum_{m}c_{m}+\sum_{n}c_{n}=1$. The quantum version involves
triplets of particles subjected to independent dichotomic measurements with
operators $\hat{A}_{1}$ and $\hat{A}_{2}$ for Alice, $\hat{B}_{1}$ and
$\hat{B}_{2}$ for Bob, and $\hat{C}_{1}$ and $\hat{C}_{2}$ for Charlie, with
each measurement resulting in outcome $\pm 1$. One defines the Mermin operator
Mermin (1990) as
$\hat{M}=\hat{A}_{1}\otimes\hat{B}_{1}\otimes\hat{C}_{2}+\hat{A}_{1}\otimes\hat{B}_{2}\otimes\hat{C}_{1}+\hat{A}_{2}\otimes\hat{B}_{1}\otimes\hat{C}_{1}-\hat{A}_{2}\otimes\hat{B}_{2}\otimes\hat{C}_{2}.$
(15)
Similarly, the Svetlichny operator is defined as Svetlichny (1987); Cereceda
(2002)
$\displaystyle\hat{S}$ $\displaystyle=$
$\displaystyle\hat{A}_{1}\otimes\big{[}(\hat{B}_{1}+\hat{B}_{2})\otimes\hat{C}_{1}+(\hat{B}_{1}-\hat{B}_{2})\otimes\hat{C}_{2}\big{]}$
(16) $\displaystyle+$
$\displaystyle\hat{A}_{2}\otimes\big{[}(\hat{B}_{1}-\hat{B}_{2})\otimes\hat{C}_{1}-(\hat{B}_{1}+\hat{B}_{2})\otimes\hat{C}_{2}\big{]}.$
The respective average values of these operators are classically upper bounded
in the form of the Mermin and Svetlichny inequalities written as
$|\langle\hat{M}\rangle|\leq 2,\qquad|\langle\hat{S}\rangle|\leq 4.$ (17)
The maximum quantum bounds are known to be $2\sqrt{2}$ and $4\sqrt{2}$ for
$|\langle\hat{M}\rangle|$ and $|\langle\hat{S}\rangle|$, respectively, and can
be attained by, for example, the GHZ state Mitchell _et al._ (2004). Several
approaches have been adopted to find the maximum quantum value of the
Svetlichny operator Ghose _et al._ (2009); Su _et al._ (2018). Recently,
Siddiqui and Sazim (2019); Li _et al._ (2017) have analytically found the
tight upper bound of the Mermin and Svetlichny operator as given below.
###### Definition 3.
For any three-qubit quantum state $\rho$, the maximum quantum value of the
Mermin and the Svetlichny operator is bounded as Li _et al._ (2017); Siddiqui
and Sazim (2019)
$\displaystyle\max|\langle\hat{M}\rangle_{\rho}|\leq
2\sqrt{2}\lambda_{1},\qquad\max|\langle\hat{S}\rangle_{\rho}|\leq
4\lambda_{1},$ (18)
where $\langle\hat{M}\rangle_{\rho}=\Tr[\hat{M}\rho]$,
$\langle\hat{S}\rangle_{\rho}=\Tr[\hat{S}\rho]$ and $\lambda_{1}$ is the
maximum singular value of the matrix $M=(M_{j,ik})$, with
$M=(M_{ijk}=\Tr[\rho(\sigma_{i}\otimes\sigma_{j}\otimes\sigma_{k})]$,
$i,j,k=1,2,3$.
We refer the reader to Siddiqui and Sazim (2019); Li _et al._ (2017) for the
class of states that saturate the above inequalities. We make use of above
mentioned bounds to study the nonlocality breaking property of (unital)
channels acting on one (or more) party in the tripartite scenario.
## III Results and discussion
In what follows, we will make use of the fact that every qubit channel
$\mathcal{E}$ can be represented in the Pauli basis
$\\{\sigma_{0},\sigma_{1},\sigma_{2},\sigma_{3}\\}$, where
$\sigma_{0}=\mathbf{I}$, by a unique $4\times 4$ matrix
$\mathbb{M}_{\mathcal{E}}=[1,\bm{0};\textbf{t},\textbf{T}]$ Beth Ruskai _et
al._ (2002); Naikoo _et al._ (2021). Here $\textbf{T}={\rm
diag.}[\eta_{1},\eta_{2},\eta_{3}]$ is a real diagonal matrix and
$\bm{0}=(0~{}0~{}0)$ and $\textbf{t}=(t_{1}~{}t_{2}~{}t_{2})^{T}$ are row and
column vectors, respectively. For $\mathcal{E}$ to be unital i.e.,
$\mathcal{E}(\mathbf{I})=\mathbf{I}$, we must have
$\textbf{t}=(0~{}0~{}0)^{T}$. The conjugate map $\mathcal{E}^{\dagger}$ is
characterized by
$\mathbb{M}^{\dagger}_{\mathcal{E}}=[1,\textbf{t}^{T};\bm{0}^{T},\textbf{T}]$,
such that the action on a state
$\rho=\frac{1}{2}(\mathbf{I}+\textbf{w}\cdot\bm{\sigma})$ is given by
$\displaystyle\mathbb{M}_{\mathcal{E}}:\mathbf{I}\rightarrow\mathbf{I}+\textbf{t}\cdot\bm{\sigma},\quad\sigma_{j}\rightarrow\eta_{j}\sigma_{j},$
(19)
$\displaystyle\mathbb{M}^{\dagger}_{\mathcal{E}}:\mathbf{I}\rightarrow\mathbf{I},\quad\sigma_{j}\rightarrow
t_{j}\mathbf{I}+\eta_{j}\sigma_{j}.$ (20)
### III.1 Equivalance of CHSH nonlocality breaking and incompatibility
breaking channels
Our first result establishes an equivalence of the CHSH nonlocality breaking
channel acting on one party, with its dual being an incompatibility breaking
channel – in the context of $2-{\rm IBC}$s. The result can be summarized by
the following two theorems:
###### Theorem 1.
If the conjugate of a qubit channel ${\rm\mathcal{E}}$ is $2-{\rm IBC}$ , then
the channel itself is CHSH-NBC.
###### Proof.
Consider the Bell-CHSH inequality given in (11), such that $[A_{1},A_{2}]\neq
0$ and $[B_{1},B_{2}]\neq 0$, i.e., the operators $A_{1}$ and $A_{2}$ and
$B_{1}$ and $B_{2}$ are incompatible in conjunction. Let
$\mathcal{E}^{\dagger}$ be the conjugate channel that is $2-{\rm IBC}$. Then
the action of this channel on Alice’s side makes $A_{1}$ and $A_{2}$
compatible, i.e., $[A_{1},A_{2}]=0$. Therefore, the Bell-CHSH inequality is
not violated Wolf _et al._ (2009), and we have
$\Tr\left[\rho\mathcal{B}\left(\mathcal{E}^{\dagger}[A_{1}],\mathcal{E}^{\dagger}[A_{2}],B_{1},B_{2}\right)\right]\leq
2.$ (21)
Alternatively,
$\displaystyle\Tr\left[\rho\mathcal{E}^{\dagger}[A_{1}]\otimes
B_{1}\right]+\Tr\left[\rho\mathcal{E}^{\dagger}[A_{2}]\otimes B_{1}\right]$
$\displaystyle+\Tr\left[\rho\mathcal{E}^{\dagger}[A_{1}]\otimes
B_{2}\right]-\Tr\left[\rho\mathcal{E}^{\dagger}[A_{2}]\otimes B_{2}\right]\leq
2.$ (22)
This can be viewed in the Schrödinger picture as
$\displaystyle\Tr\left[(\mathcal{E}\otimes\mathbf{I})[\rho]A_{1}\otimes
B_{1}\right]+\Tr\left[(\mathcal{E}\otimes\mathbf{I})[\rho]A_{2}\otimes
B_{1}\right]$
$\displaystyle+\Tr\left[(\mathcal{E}\otimes\mathbf{I})[\rho]A_{1}\otimes
B_{2}\right]-\Tr\left[(\mathcal{E}\otimes\mathbf{I})[\rho]A_{2}\otimes
B_{2}\right]\leq 2,$ (23)
which tells us that the CHSH inequality is satisfied even when operators
$A_{1}$ and $A_{2}$ and $B_{1}$ and $B_{2}$ are incompatible in conjunction.
Therefore, the action of $\mathcal{E}$ (to be precise of
$\mathcal{E}\otimes\mathbf{I}$) on state $\rho$ is solely responsible for
nonviolation of the CHSH inequality. We conclude that $\mathcal{E}$ is CHSH-
NBC. ∎
###### Theorem 2.
If a qubit channel $\mathcal{E}$ is CHSH-NBC, then its conjugate is $2-{\rm
IBC}$.
###### Proof.
Here we start with incompatible operators associated with the respective
subsystems, $[A_{1},A_{2}]\neq 0$ and $[B_{1},B_{2}]\neq 0$ and assume that
the channel $\mathcal{E}$ acting on Alice’s side does not allow for the
violation of the CHSH inequality, that is
$\Tr\left[\left(\mathcal{E}\otimes\mathbf{I}\right)[\rho]\mathcal{B}\left(A_{1},A_{2},B_{1},B_{2}\right)\right]\leq
2.$ (24)
In other word, looking from the measurement point of view, in the Heisenberg
picture, we have
$\displaystyle\Tr\left[\rho\mathcal{E}^{\dagger}[A_{1}]\otimes
B_{1}\right]+\Tr\left[\rho\mathcal{E}^{\dagger}[A_{1}]\otimes B_{2}\right]$
$\displaystyle+\Tr\left[\rho\mathcal{E}^{\dagger}[A_{2}]\otimes
B_{1}\right]-\Tr[\rho\mathcal{E}^{\dagger}[A_{2}]\otimes B_{2}]\leq 2$ (25)
The above inequality holds for arbitrary state $\rho$ which can even be an
entangled state. Thus the nonviolation of the CHSH inequality is coming from
the action of $\mathcal{E}^{\dagger}$ on the operators $A_{1}$ and $A_{2}$,
making them compatible,
$\left[\mathcal{E}^{\dagger}[A_{1}],\mathcal{E}^{\dagger}[A_{2}]\right]=0$. We
conclude that $(\mathcal{E}^{\dagger}\otimes\mathbf{I})$ is incompatibility
breaking. ∎
In order to verify the above results, let us consider the CHSH inequality
(11), such that the local observables are subject to some biased noise
characterized by $(x_{a},\eta_{a})$ and $(x_{b},\eta_{b})$. As a result, the
observables are modified as
$A_{k}=\mathbf{I}x_{a}+\eta_{a}\hat{a}_{k}\cdot\vec{\sigma}_{k}$,
$B_{k}=\mathbf{I}x_{b}+\eta_{b}\hat{b}_{k}\cdot\vec{\sigma}_{k}$, and $k=1,2$.
Averaging with respect to the singlet-state, the CHSH inequality is satisfied
if
$\displaystyle
2x_{a}x_{b}+\eta_{a}\eta_{b}\left|\cos\theta_{11}+\cos\theta_{12}+\cos\theta_{21}-\cos\theta_{22}\right|\leq
2.$ (26)
The modulus term has a maximum value of $2\sqrt{2}$ for
$\theta_{ab}=\theta_{a^{\prime}b}=\theta_{ab^{\prime}}=\theta$ and
$\theta_{a^{\prime}b^{\prime}}=3\theta$ with $\theta=\pi/4$. The above
inequality becomes Kar and Roy (1995)
$\eta_{a}\eta_{b}\leq\frac{1-x_{a}x_{b}}{\sqrt{2}}.$ (27)
For the unbiased noise on either side, i.e., $x_{a}=0$ or $x_{b}=0$, the
incompatibility breaking condition reduces to
$\eta_{a}\eta_{b}\leq\frac{1}{\sqrt{2}}.$ (28)
Further, if this unbiased noise is acting only on one side (for example, on
$A$) of the bipartite system, then $\eta_{b}=1$, and we have
$\eta_{a}\leq\frac{1}{\sqrt{2}}.$ (29)
Having seen the impact of noise on the violation of the CHSH inequality, we
now look at the incompatibility breaking condition for observables $A_{1}$ and
$A_{2}$ subjected to the same noise. As a result we have
$A_{1}^{(x_{a},\eta_{a})}$ and $A_{2}^{(x_{a},\eta_{a})}$ as biased and
unsharp observables. Using inequality (6), the incompatibility condition for
these observables reads
$\eta_{a}\leq\frac{1-x_{a}^{2}}{\sqrt{2}}.$ (30)
This coincides with the NB condition (29) only if $x_{a}=0$. Since the action
of the nonunital channel on a projector results in a biased observable Naikoo
_et al._ (2021), one finds that the conditions for nonlocality breaking and
incompatibility breaking agree as long as the dynamics is unital in the
Schrödinger picture.
### III.2 Nonlocality and incompatibility breaking channels in tripartite
scenario
In the tripartite scenario, the aforesaid relation between nonlocality and
incompatibility does not hold in general. In this section, we therefore first
obtain the NB condition for some well known tripartite states and then
identify the range of channel parameter where NB agrees with the $2$-IBC
condition. Let us first introduce the definition of the absolute and genuine
tripartite NBC as
###### Definition 4.
For any three-qubit state $\rho_{ABC}$, a given qubit channel $\mathcal{E}$ is
said to be absolute or genuine nonlocality breaking if acting on any qubit, it
gives a state (for example) $\rho^{\prime}_{ABC}=(\mathcal{E}\otimes I\otimes
I)(\rho_{ABC})$, which satisfies Mermin inequality
$\langle\hat{M}\rangle_{\rho^{\prime}_{ABC}}\leq 2$ or Svetlichny inequality
$\langle\hat{S}\rangle_{\rho^{\prime}_{ABC}}\leq 4$.
In terms of the largest singular value in Definition 3, the Mermin nonlocality
breaking condition (M-NBC) and Svetlichny nonlocality breaking condition
(S-NBC) are, respectively, given as
$\eta_{M}\leq\frac{1}{\sqrt{2}\lambda_{max}},\qquad\eta_{S}\leq\frac{1}{\lambda_{max}}.$
(31)
We now consider some well-known tripartite quantum states in which one party
is subjected to a noisy evolution, and use Def. 3 to obtain the conditions on
the noise parameter for breaking the Mermin and Svetlichny nonlocality.
Example 1. Our first example is the generalized GHZ state Dür _et al._
(2000); Greenberger _et al._ (2007)
$|\psi_{GHZ}\rangle=\alpha|000\rangle+\beta|111\rangle.$ (32)
The entanglement in this state is completely destroyed when any of the three
qubits is traced out, i.e.,
$\Tr_{k}\left[|\psi_{GHZ}\rangle\langle\psi_{GHZ}|\right]=\mathbf{I}/2$, with
$k=A,B,C$. Let us assume a unital noise channel $\mathcal{E}$ acting on one
qubit (for example, the first one) according to Eq. (19). This can be achieved
by expressing $|\psi_{GHZ}\rangle$ in the Pauli basis and invoking
$\sigma_{i}\rightarrow\eta\sigma_{i}$ ($i=x,y,z$, $0\leq\eta\leq 1)$ at the
first qubit
$\displaystyle\left(\mathcal{E}\otimes\mathbf{I}\otimes\mathbf{I}\right)\left[|\psi_{GHZ}\rangle\langle\psi_{GHZ}|\right]$
$\displaystyle=\frac{1}{8}\Big{[}\mathbf{I}\otimes\mathbf{I}\otimes\mathbf{I}+\mathbf{I}\otimes\sigma_{3}\otimes\sigma_{3}+\eta(\sigma_{3}\otimes\mathbf{I}\otimes\sigma_{3})$
$\displaystyle+(\alpha^{2}-\beta^{2})(\eta\sigma_{3}\otimes\mathbf{I}\otimes\mathbf{I}+\mathbf{I}\otimes\sigma_{3}\otimes\mathbf{I}$
$\displaystyle+\mathbf{I}\otimes\mathbf{I}\otimes\sigma_{3}+\eta\sigma_{3}\otimes\sigma_{3}\otimes\sigma_{3})+2\alpha\beta\eta(\sigma_{1}\otimes\sigma_{1}\otimes\sigma_{1}$
$\displaystyle-\sigma_{1}\otimes\sigma_{2}\otimes\sigma_{2}-\sigma_{2}\otimes\sigma_{1}\otimes\sigma_{2}-\sigma_{2}\otimes\sigma_{2}\otimes\sigma_{1})\Big{]}$
(33)
The matrix $M_{j,ik}$ in Def. 3 turns out to be
$\displaystyle
M_{j,ik}=\small\begin{pmatrix}2\eta\alpha\beta&0&0&0&-2\eta\alpha\beta&0&0&0&0\\\
0&-2\eta\alpha\beta&0&-2\eta\alpha\beta&0&0&0&0&0\\\
0&0&0&0&0&0&0&0&\eta(\alpha^{2}-\beta^{2})\end{pmatrix}.$
This has singular values $2\sqrt{2}\eta\alpha\beta$,
$2\sqrt{2}\eta\alpha\beta$, and $\eta\left|\alpha^{2}-\beta^{2}\right|$. Thus
the condition for the channel to be Mermin and Svetlichny NB is given by,
$\displaystyle\eta_{M}\leq\frac{1}{4\alpha\beta},\qquad\eta_{S}\leq\frac{1}{2\sqrt{2}\alpha\beta}.$
(35)
Example 2. Next, we consider the well-known generalized W-state given by
$|\psi\rangle=\alpha|100\rangle+\beta|010\rangle+\gamma|001\rangle,$ (36)
with $\alpha,\beta$, and $\gamma$ real and
$\alpha^{2}+\beta^{2}+\gamma^{2}=1$. This state is special in the sense that
if one qubit is lost, the state of the remaining two qubits is still
entangled, unlike the GHZ state. The matrix $M_{j,ik}$ in Def.3 corresponding
to unital noise acting on the first qubit of the state is given by
$\displaystyle M_{j,ik}=\begin{pmatrix}0&0&\eta\omega&0&0&0&\eta\omega&0&0\\\
0&0&0&0&0&\eta\omega&0&\eta\omega&0\\\
\eta\omega^{\prime}&0&0&0&\eta\omega^{\prime}&0&0&0&\eta(2\alpha\gamma-\beta^{2})\end{pmatrix},$
with $\omega=\alpha\beta+\beta\gamma$,
$\omega^{\prime}=(\alpha^{2}+\gamma^{2})$. The largest singular value turns
out to be $\lambda=\eta\sqrt{1+8\beta^{2}\gamma^{2}}$. Now
$\beta^{2}+\gamma^{2}=1-\alpha^{2}=k$ (for example), so that
$\beta^{2}\gamma^{2}=\beta^{2}(k-\beta^{2})$. This quantity attains the
maximum $k^{2}/4$ at $\beta^{2}=k/2$, leading to
$\lambda_{max}=\eta\sqrt{1+2k^{2}}$. The Mermin and Svetlichny nonlocality
breaking conditions then read
$\eta_{M}\leq\frac{1}{\sqrt{2}\sqrt{1+2k^{2}}},\qquad\eta_{S}\leq\frac{1}{\sqrt{1+2k^{2}}}.$
(37)
The singular values attain a maximum at $k=2/3$, i.e.,
$\alpha=\beta=\gamma=1/\sqrt{3}$. Under this condition and for $\eta_{S}$
approximately in the range $[0.707-0.727]$, the channel is S-NBC but not
$2-{\rm IBC}$. Note that in the bipartite case, a unital channel is CHSH-NBC
for all $\eta\leq 1/\sqrt{2}$ Pal and Ghosh (2015). Thus in the range
$\eta\approx 0.707-0.727$, it is S-NBC but not CHSH-NBC.
Example 3. We next consider the three-qubit partially entangled set of maximal
slice (MS) states Carteret and Sudbery (2000); Liu _et al._ (2016)
$|\psi_{MS}\rangle=\frac{1}{\sqrt{2}}\bigg{[}|000\rangle+|11\bigg{(}\alpha|0\rangle+\beta|1\rangle\bigg{)}\bigg{]},$
(38)
where $\alpha$ and $\beta$ are real with $\alpha^{2}+\beta^{2}=1$. The matrix
$M_{j,ik}$ corresponding to Eq. (38) after the action of unital noise on the
first qubit is given by
$\displaystyle M_{j,ik}$
$\displaystyle=\begin{pmatrix}\eta\beta&0&\eta\alpha&0&-\eta\beta&0&0&0&0\\\
0&-\eta\beta&0&-\eta\beta&0&-\eta\alpha&0&0&0\\\
0&0&0&0&0&0&\eta\alpha\beta&0&\eta(\frac{1+\alpha^{2}-\beta^{2}}{2})\end{pmatrix}$
with three singular values
$\lambda_{1}=\eta\sqrt{\frac{1}{4}\left(\alpha^{2}-\beta^{2}+1\right)^{2}+(\alpha\beta)^{2}}$
and two equal singular values
$\lambda_{2}=\lambda_{3}=\eta\sqrt{(\alpha^{2}+2\beta^{2})}$ respectively,
leading to the condition for Mermin and Svetlichny nonlocality breaking as
$\displaystyle\eta_{M}\leq\frac{1}{\sqrt{2(\alpha^{2}+2\beta^{2})}},\qquad\eta_{S}\leq\frac{1}{\sqrt{(\alpha^{2}+2\beta^{2})}}.$
(39)
Figure 1: (Color online): The region below the dashed (red) and solid (black)
curve in (a), (b), (c), and (d) corresponds to M-NBC and S-NBC given by Ineqs.
(35), (37), (39), and (41), respectively, plotted against the (dimensionless)
state coefficients. The pairwise incompatibility breaking condition (9)
pertains to all points below the horizontal dashed line.
Example 4. Consider the quantum state Augusiak _et al._ (2015)
$\sigma_{GHZ}=p|GHZ\rangle\langle GHZ|+(1-p)I_{2}\otimes\tilde{I}.$ (40)
Here, $|GHZ\rangle\langle=(|000\rangle+|111\rangle)/\sqrt{2}$,
$\tilde{I}=diag.(1,0,0,1)$, is diagonal matrix, and $0\leq p\leq 1$. In fact,
$\tilde{I}=|\Phi^{+}\rangle\langle\Phi^{+}|+|\Phi^{-}\rangle\langle\Phi^{-}|$,
where $\Phi^{\pm}$ are Bell states, is a separable state, and it tells us that
$\sigma_{GHZ}$ has with probability $p$ the GHZ state and with probability
$(1-p)$ the first qubit is left in the mixed state and the second and third
qubit are in separable state $\tilde{I}$. The corresponding $M_{j,ik}$ matrix
with a unital noise acting on one qubit is given by
$\displaystyle M_{j,ik}=\begin{pmatrix}p\eta&0&0&0&-p\eta&0&0&0&0\\\
0&-p\eta&0&-p\eta&0&0&0&0&0\\\ 0&0&0&0&0&0&0&0&0\end{pmatrix},$
with singular values $\lambda_{1}=\lambda_{2}=\sqrt{2}\eta p$. Thus the
condition for the channel to be Mermin and Svetlichny NBC becomes
$\displaystyle\eta_{M}\leq\frac{1}{2p},\qquad\eta_{S}\leq\frac{1}{\sqrt{2}p}.$
(41)
The conditions for Mermin and Svetlichny nonlocality breaking channel given by
Ineqs. (35), (36), (39), and (41) obtained by the application of a unital
quantum channel to one party of a tripartite system, are depicted in Fig. 1.
All the points below the solid (black) and dashed (red) curve correspond to
the nonlocality breaking channel, while as the points below the horizontal
dashed line, $\eta=1/\sqrt{2}$, pertain to pairwise incompatibility breaking.
In all the four examples, Fig. 1 (a)-(d), the minimum value of $\eta_{S}$ for
which SI is violated is $1/\sqrt{2}$, suggesting that _genuine_ nonlocal
correlations can not be established if at least one pair of observable is
compatible. The converse is not true, since there exist regions (above the
horizontal dashed line and below the solid (black) curve) of Svetlichny
nonlocality breaking even when the channel is not $2-{\rm IBC}$. Thus, these
examples illustrate that corresponding to $2$-IBC the conjugate channels are
definitely S-NLB; however, the conjugate of S-NB channels may not necessarily
be a $2$-IBC. However, in the context of Mermin nonlocality, even the first
statement does not hold, that is, existence of a $2$-IBC does not necessarily
guarantee a conjugate channel that is M-NB. Also, the minimum $\eta_{M}$ (that
is maximum noise) for which a channel is M-NB is always less by a factor of
$1/\sqrt{2}$ than the minimum noise below which that channel is S-NB. It is
worth pointing here to Fig. 1 (c), which illustrates that if a (unital)
channel breaks the Svetlichny nonlocality for the GHZ state (which in fact
violates the SI maximally) then it also does so for the mixture (40) for
$1/\sqrt{2}<p<1$. This is unlike the bipartite scenario where a unital channel
that breaks the CHSH nonlocality for the maximally entangled states is
guaranteed to do so for all other states Pal and Ghosh (2015). Summarizing, it
is clear from the above examples that the existence of M-NBC or S-NBC does not
guarantee the existence of a conjugate $2$-IBC, unlike the CHSH scenario. In
particular, with the (unital) noise acting on one party of the $W$-state,
$|\psi\rangle=(|100\rangle+|010\rangle+|001\rangle$)/$\sqrt{3}$, there exists
a range of the channel parameter $\eta\in(1/\sqrt{2},3/\sqrt{17})$, where the
channel is S-NBC but not CHSH NBC as depicted in the Fig. 2. Thus the Bell-
CHSH inequality seems to be more suitable for a study of the incompatibility
of observables than the multi-partite Bell-type inequalities.
Note that instead of one party, if two or all the three parties are subjected
to noise, the conditions (35), (37), (39), and (41), become
$\eta_{M}\leq\left(\frac{1}{\sqrt{2}\lambda_{max}}\right)^{1/n},\qquad\eta_{S}\leq\left(\frac{1}{\lambda_{max}}\right)^{1/n},$
(42)
where $n$ corresponds to the number of qubits subjected to noise. Since
$1/\lambda_{max}<(1/\lambda_{max})^{1/2}<(1/\lambda_{max})^{1/3}$ (with
$\lambda_{max}>1$), the solid (black) and dashed (red) curves in Fig. 1
(a)-(d) are shifted up, thereby decreasing the region of nonlocality with
increase in $n$.
General three qubit state subjected to general unital noise: The (unital)
noise acting on a single party considered in the above analysis involving the
tripartite system assumes identical effects on $\sigma_{x}$, $\sigma_{y}$, and
$\sigma_{z}$ corresponding to that party, in the sense that
$\sigma_{k}\rightarrow\eta\sigma_{k}$ for all $k=x,y,z$. A more general
transformation would take the particular party’s
$\sigma_{k}\rightarrow\eta_{k}\sigma_{k}$, with
$\sqrt{\eta_{x}^{2}+\eta_{y}^{2}+\eta_{z}^{2}}=\eta$, and $0\leq\eta\leq 1$,
such that
$\Phi\left(\frac{\mathbf{I}+\vec{r}\cdot\vec{\sigma}}{2}\right)=\frac{\mathbf{I}+(\textbf{T}\vec{r})\cdot\vec{\sigma}}{2},$
(43)
where $\textbf{T}={\rm diag.}[\eta_{x},\eta_{y},\eta_{z}]$, is a real diagonal
matrix. The map $\Phi$ is completely positive for King and Ruskai (2001)
$\left|\eta_{x}\pm\eta_{y}\right|\leq\left|1\pm\eta_{z}\right|,$ (44)
which is a set of four inequalities and defines a tetrahedron in
$\eta_{x}-\eta_{y}-\eta_{z}$ space. Under such a transformation, the singular
values for the GHZ in (32) and $W$ state in (36) are, respectively, given by
$\displaystyle\left(2\sqrt{2}\alpha\beta\eta_{x},2\sqrt{2}\alpha\beta\eta_{y},\left|\alpha^{2}-\beta^{2}\right|\eta_{z}\right),$
(45) $\displaystyle{\rm and}~{}~{}$
$\displaystyle\left(2\alpha\sqrt{\beta^{2}+\gamma^{2}}\eta_{x},2\alpha\sqrt{\beta^{2}+\gamma^{2}}\eta_{y},\sqrt{1+8\beta^{2}\gamma^{2}}\eta_{z}\right).$
(46)
with the three singular values depending linearly on the respective noise
parameters. Depending on which singular value is the largest, one can draw
similar conclusions about nonlocality and incompatibility breaking properties
of the noise channel as in the case with uniform noise action. However, for a
general three-qubit input state, the dependence of the singular values on
noise parameters tuns out to be complicated leading to different conclusions
regarding the nonlocality breaking property of such a channel. Let us consider
the general situation where one would like to make a statement about the
limiting noise beyond which no Mermin or Svetlichny nonlocal correlations can
be established _irrespective_ of the state chosen. In this direction, we make
use of the canonical five term decomposition of the three-qubit state Acín
_et al._ (2001):
$\ket{\psi}=\lambda_{0}\ket{000}+\lambda_{1}e^{i\phi}\ket{100}+\lambda_{2}\ket{101}+\lambda_{3}\ket{110}+\lambda_{4}\ket{111},$
(47)
with $\lambda_{i}$ and $\phi$ real parameters and
$\sum_{i}\left|\lambda_{i}\right|^{2}=1$ and $0<\phi<\pi$. In the Pauli basis,
one may denote the density matrix corresponding to $\ket{\psi}$ as
$|\psi\rangle\langle\psi|\left[\sigma_{i}\otimes\sigma_{j}\otimes\sigma_{k}\right]$
with Pauli matrices $\sigma_{i}$ $i=x,y,z$. If a unital noise acts on one
party (for example, the first) of such a state, we would have
$\displaystyle|\psi\rangle\langle\psi|\left[\sigma_{i}\otimes\sigma_{j}\otimes\sigma_{k}\right]\rightarrow|\psi\rangle\langle\psi|\left[\eta_{i}\sigma_{i}\otimes\sigma_{j}\otimes\sigma_{k}\right].$
(48)
Figure 2: (Color online): Depicting the range of channel parameter
$\eta\in(\frac{1}{\sqrt{2}},\frac{3}{\sqrt{17}})$ in which it is S-NBC (for
the $W$ state) but not CHSH-NBC.
Figure 3: (Color online) Singular values corresponding to a sample of $5\times
10^{6}$ randomly generated states of the form $(\ref{eq:SVAcin})$ plotted with
respect to the dimensionless parameter
$\eta=\sqrt{\eta_{x}^{2}+\eta_{y}^{2}+\eta_{z}^{2}}$. In terms of these
singular values, the conditions for establishing the Mermin and Svetlichny
nonlocal correlations are given in Def. 3 which demands that the largest
singular value be greater than $1/\sqrt{2}$ and $1$, respectively. The
corresponding bound on the unsharpness parameters $\eta_{M}$ and $\eta_{S}$
trivially follow and are summarized in the text.
We can now check the nonlocality breaking properties of such a general noise
channel using Def. 3 based on singular values of matrix $M$. For the above
state, the singular values calculated according to Def. 3, are plotted in Fig.
3 with respect to $\eta=\sqrt{\eta_{x}^{2}+\eta_{y}^{2}+\eta_{z}^{2}}$. The
simulation makes use of a sample of $5\times 10^{6}$ randomly generated states
and the corresponding noise parameters $\eta_{i}$ subjected to
$0\leq\sqrt{\eta_{x}^{2}+\eta_{y}^{2}+\eta_{z}^{2}}\leq 1$ and also satisfying
the completely positive condition (44). According to the condition (18), the
Mermin and Svetlichny nonlocal correlations are established only if the
largest singular value is greater than $\frac{1}{\sqrt{2}}$ for the former and
$1$ for the later. In Fig. 3 the three singular values
$\lambda_{1},\lambda_{2}$, and $\lambda_{3}$ are depicted with respect to
parameter $\eta$. One finds that below a minimum $\eta_{M}$ and $\eta_{S}$,
the singular values do not exceed $\frac{1}{\sqrt{2}}$ and $1$, respectively.
These values turn out to be $(\eta_{M},\eta_{S})=(0.090,0.128)$ for
$\lambda_{1}$, $(\eta_{M},\eta_{S})=(0.182,0.259)$ for $\lambda_{2}$, and
$(\eta_{M},\eta_{S})=(0.300,0.409)$ for $\lambda_{3}$, and one may conclude
that no Mermin (Svetlichny) nonlocal correlations are supported by the unital
noise channel if the noise parameter is below $0.090~{}(0.128)$, irrespective
of the input state.
## IV Conclusion
This paper is devoted to a study of the interplay between nonlocality breaking
and incompatibility breaking power of noisy quantum qubit channels. The action
of quantum channels on projective measurements transforms them into noisy
POVMs, characterized in particular by unsharpness parameters. As a
consequence, noise tends to increase the compatibility of observers that are
otherwise incompatible. In fact, pairwise incompatibility breaking is assured
if the channel parameter is less than or equal to $1/\sqrt{2}$. To be
specific, we consider bipartite and tripartite scenarios, with CHSH
nonlocality in the former and Mermin and Svetlichny nonlocality in the later
case. The degree of incompatibility breaking directly depends on the
unsharpness parameters. Here, we showed that in the Bell-CHSH scenario, if the
conjugate of a channel is incompatibility breaking then the channel is itself
nonlocality breaking and the converse is also true. In the tripartite
scenario, however, such an equivalence between nonlocality breaking and
incompatibility breaking does not exist. We then consider various examples of
three-qubit states and identify the state parameters for which the equivalence
of nonlocality breaking corroborates with the pairwise incompatibility. In
particular, it is illustrated that the conjugate of incompatibility breaking
channels is nonlocality breaking, however, the nonlocality breaking channels
do not guarantee the existence of conjugate channels that are incompatibility
breaking. This may be viewed as a useful feature of the Bell-CHSH inequality
when it comes to the study incompatibility of observables. Further, from
randomly generated three-qubit states subjected to general unital channels, we
conclude that no Mermin (Svetlichny) nonlocal correlations are supported for
$\eta_{M}<0.090$ ($\eta_{S}<0.128$), $\eta_{M/S}$ being the channel
unsharpness parameter.
The channel activation of nonlocality in the CHSH scenario has been studied in
Zhang _et al._ (2020). A future extension of this paper could be the study of
activation of Mermin and Svetlichny nonlocality under more general noise
models and with general three-qubit input state. This also invites for a
detailed analysis on the hierarchy of nonlocality-breaking, steerability-
breaking, and entanglement-breaking quantum channels in the tripartite
scenario.
## Acknowledgment
Authors thank Guruprasad Kar for the initial motivation of the work. SK
acknowledges Yeong-Cherng Liang for the fruitful comments and the support from
the National Science and Technology Council, Taiwan (Grants
no.109-2112-M-006-010-MY3,110-2811-M-006-511,111-2811-M-006-013). JN’s work
was supported by the Foundation for Polish Science within the “Quantum Optical
Technologies” project carried out within the International Research Agendas
programme cofinanced by the European Union under the European Regional
Development Fund. AKP acknowledges the support from the research grant
DST/ICPS/QuEST/2019/4.
## References
* Bell (1964) J. S. Bell, Physics Physique Fizika 1, 195 (1964).
* Brunner _et al._ (2014) N. Brunner, D. Cavalcanti, S. Pironio, V. Scarani, and S. Wehner, Rev. Mod. Phys. 86, 419 (2014).
* Scarani and Gisin (2001) V. Scarani and N. Gisin, Phys. Rev. Lett. 87, 117901 (2001).
* Ekert (1991) A. K. Ekert, Phys. Rev. Lett. 67, 661 (1991).
* Heinosaari _et al._ (2016) T. Heinosaari, T. Miyadera, and M. Ziman, Journal of Physics A: Mathematical and Theoretical 49, 123001 (2016).
* Wolf _et al._ (2009) M. M. Wolf, D. Perez-Garcia, and C. Fernandez, Phys. Rev. Lett. 103, 230402 (2009).
* Quintino _et al._ (2014) M. T. Quintino, T. Vértesi, and N. Brunner, Phys. Rev. Lett. 113, 160402 (2014).
* Uola _et al._ (2015) R. Uola, C. Budroni, O. Gühne, and J.-P. Pellonpää, Phys. Rev. Lett. 115, 230402 (2015).
* Kochen and Specker (1975) S. Kochen and E. P. Specker, “The problem of hidden variables in quantum mechanics,” in _The Logico-Algebraic Approach to Quantum Mechanics: Volume I: Historical Evolution_, edited by C. A. Hooker (Springer Netherlands, Dordrecht, 1975) pp. 293–328.
* Budroni _et al._ (2022) C. Budroni, A. Cabello, O. Gühne, M. Kleinmann, and J.-A. Larsson, Rev. Mod. Phys. 94, 045007 (2022).
* Liang _et al._ (2011) Y.-C. Liang, R. W. Spekkens, and H. M. Wiseman, Physics Reports 506, 1 (2011).
* Ku _et al._ (2022a) H.-Y. Ku, C.-Y. Hsieh, S.-L. Chen, Y.-N. Chen, and C. Budroni, Nature Communications 13, 4973 (2022a).
* Bennett _et al._ (1999) C. H. Bennett, D. P. DiVincenzo, C. A. Fuchs, T. Mor, E. Rains, P. W. Shor, J. A. Smolin, and W. K. Wootters, Phys. Rev. A 59, 1070 (1999).
* Bene and Vértesi (2018) E. Bene and T. Vértesi, New Journal of Physics 20, 013021 (2018).
* Hirsch _et al._ (2018) F. Hirsch, M. T. Quintino, and N. Brunner, Phys. Rev. A 97, 012129 (2018).
* Svetlichny (1987) G. Svetlichny, Phys. Rev. D 35, 3066 (1987).
* Ajoy and Rungta (2010) A. Ajoy and P. Rungta, Phys. Rev. A 81, 052334 (2010).
* Bancal _et al._ (2013) J.-D. Bancal, J. Barrett, N. Gisin, and S. Pironio, Phys. Rev. A 88, 014102 (2013).
* Pal and Ghosh (2015) R. Pal and S. Ghosh, Journal of Physics A: Mathematical and Theoretical 48, 155302 (2015).
* Heinosaari _et al._ (2015) T. Heinosaari, J. Kiukas, D. Reitzner, and J. Schultz, 48, 435301 (2015).
* Heinosaari and Miyadera (2017) T. Heinosaari and T. Miyadera, Journal of Physics A: Mathematical and Theoretical 50, 135302 (2017).
* Horodecki _et al._ (2003) M. Horodecki, P. W. Shor, and M. B. Ruskai, Reviews in Mathematical Physics 15, 629–641 (2003).
* Kiukas _et al._ (2017) J. Kiukas, C. Budroni, R. Uola, and J.-P. Pellonpää, Phys. Rev. A 96, 042331 (2017).
* Ku _et al._ (2022b) H.-Y. Ku, J. Kadlec, A. Cernoch, M. T. Quintino, W. Zhou, K. Lemr, N. Lambert, A. Miranowicz, S.-L. Chen, F. Nori, and Y.-N. Chen, PRX Quantum 3, 020338 (2022b).
* Collins _et al._ (2002) D. Collins, N. Gisin, S. Popescu, D. Roberts, and V. Scarani, Phys. Rev. Lett. 88, 170405 (2002).
* Busch (1986) P. Busch, Phys. Rev. D 33, 2253 (1986).
* Nielsen and Chuang (2010) M. A. Nielsen and I. L. Chuang, _Quantum Computation and Quantum Information: 10th Anniversary Edition_ (Cambridge University Press, 2010).
* Yu _et al._ (2010) S. Yu, N.-l. Liu, L. Li, and C. H. Oh, Phys. Rev. A 81, 062116 (2010).
* Heinosaari (2020) T. Heinosaari, Journal of Physics: Conference Series 1638, 012002 (2020).
* Mermin (1990) N. D. Mermin, Phys. Rev. Lett. 65, 1838 (1990).
* Cereceda (2002) J. L. Cereceda, Phys. Rev. A 66, 024102 (2002).
* Mitchell _et al._ (2004) P. Mitchell, S. Popescu, and D. Roberts, Phys. Rev. A 70, 060101 (2004).
* Ghose _et al._ (2009) S. Ghose, N. Sinclair, S. Debnath, P. Rungta, and R. Stock, Phys. Rev. Lett. 102, 250404 (2009).
* Su _et al._ (2018) Z. Su, L. Li, and J. Ling, Quantum Information Processing 17, 1 (2018).
* Siddiqui and Sazim (2019) M. A. Siddiqui and S. Sazim, Quantum Information Processing 18 (2019), 10.1007/s11128-019-2246-1.
* Li _et al._ (2017) M. Li, S. Shen, N. Jing, S.-M. Fei, and X. Li-Jost, Phys. Rev. A 96, 042323 (2017).
* Beth Ruskai _et al._ (2002) M. Beth Ruskai, S. Szarek, and E. Werner, Linear Algebra and its Applications 347, 159 (2002).
* Naikoo _et al._ (2021) J. Naikoo, S. Banerjee, A. K. Pan, and S. Ghosh, Phys. Rev. A 104, 042608 (2021).
* Kar and Roy (1995) G. Kar and S. Roy, Physics Letters A 199, 12 (1995).
* Dür _et al._ (2000) W. Dür, G. Vidal, and J. I. Cirac, Phys. Rev. A 62, 062314 (2000).
* Greenberger _et al._ (2007) D. M. Greenberger, M. A. Horne, and A. Zeilinger, “Going beyond bell’s theorem,” (2007), arXiv:0712.0921 [quant-ph] .
* Carteret and Sudbery (2000) H. A. Carteret and A. Sudbery, Journal of Physics A: Mathematical and General 33, 4981 (2000).
* Liu _et al._ (2016) J. Liu, Z.-w. Mo, and S.-q. Sun, International Journal of Theoretical Physics 55, 2182 (2016).
* Augusiak _et al._ (2015) R. Augusiak, M. Demianowicz, J. Tura, and A. Acín, Phys. Rev. Lett. 115, 030404 (2015).
* King and Ruskai (2001) C. King and M. Ruskai, IEEE Transactions on Information Theory 47, 192 (2001).
* Acín _et al._ (2001) A. Acín, A. Andrianov, E. Jané, and R. Tarrach, Journal of Physics A: Mathematical and General 34, 6725 (2001).
* Zhang _et al._ (2020) Y. Zhang, R. A. Bravo, V. O. Lorenz, and E. Chitambar, New Journal of Physics 22, 043003 (2020).
|
# On the relevance of lift force modelling in turbulent wall flows with small
inertial particles
Wei Gao1,2<EMAIL_ADDRESS>Pengyu Shi3,4 Matteo Parsani1,2 Pedro
Costa5 1Mechanical Engineering, Physical Science and Engineering Division,
King Abdullah University of Science and Technology, Thuwal 23955-6900, Saudi
Arabia 2Applied Mathematics and Computational Science, Computer Electrical and
Mathematical Science and Engineering Division, Extreme Computing Research
Center, King Abdullah University of Science and Technology, Thuwal 23955-6900,
Saudi Arabia 3Institut de Mécanique des Fluides de Toulouse (IMFT), Université
de Toulouse, CNRS, Toulouse, France 4Helmholtz-Zentrum Dresden – Rossendorf,
Institute of Fluid Dynamics, 01328 Dresden, Germany 5Process & Energy
Department, TU Delft, Leeghwaterstraat 39, 2628 CB, The Netherlands
###### Abstract
In particle-laden turbulent wall flows, lift forces can influence the near-
wall turbulence. This has been recently observed in particle-resolved
simulations, which, however, are too expensive to be used in upscaled models.
Instead, point-particle simulations have been the method of choice to simulate
the dynamics of these flows during the last decades. While this approach is
simpler, cheaper, and physically sound for small inertial particles in
turbulence, some issues remain. In the present work, we address challenges
associated with lift force modelling in turbulent wall flows and the impact of
lift forces in the near-wall flow. We performed direct numerical simulations
(DNS) of small inertial point particles in turbulent channel flow for fixed
Stokes number and mass loading while varying the particle size. Our results
show that the particle dynamics in the buffer region, causing the apparent
particle-to-fluid slip velocity to vanish, raise major challenges for
accurately modelling lift forces. While our results confirm that lift forces
have little influence on particle dynamics for sufficiently small particle
sizes, for inner-scaled diameters of order one and beyond, lift forces become
quite important near the wall. The different particle dynamics under lift
forces result in the modulation of streamwise momentum transport in the near-
wall region. We analyze this lift-induced turbulence modulation for different
lift force models, and the results indicate that realistic models are critical
for particle-modeled simulations to correctly predict turbulence modulation by
particles in the near-wall region.
###### keywords:
particle-laden wall turbulence, lift force, turbulence modulation
MSC Codes (Optional) Please enter your MSC Codes here
## 1 Introduction
Wall-bounded turbulent flows laden with inertial particles abound in
environmental and industrial contexts, such as the transport of particulate
matter in the atmosphere, sediment transport in rivers, the separation of fine
particles within industrial cyclones, and fluidized bed reactors. The chaotic
and multiscale nature of the fluid turbulence coupled with the particle
dynamics results in fascinating phenomena, which, however, is still
challenging to understand and model (Balachandar & Eaton, 2010; Brandt &
Coletti, 2022).
In the past decades, many studies have been devoted to the dynamics of
turbulent wall transport of small inertial particles, driven by their
prevalent nature and rich physics. These systems have been traditionally
classified by how the dispersed phase influences the overall flow behaviour
(Elghobashi, 1994). Precisely, one-way coupling (1WC) corresponds to very
small particle loadings, so small that their influence on the dynamics of the
suspending fluid turbulence is negligible; two-way coupling (2WC) denotes flow
regimes where mass loading is high enough such that the particles modify flow
observably, but particle-particle interactions are negligible; finally regimes
where both particle–particle and particle–fluid interactions influence the
flow dynamics are grouped into the four-way coupling category.
In a first-principles, fully-resolved direct numerical simulation (DNS) of a
particle-laden flow, the flow around each particle needs to be explicitly
resolved (Balachandar & Eaton, 2010; Maxey, 2017) (so-called particle-resolved
DNS, PR-DNS). While this approach is free from modelling assumptions for the
dispersed phase dynamics, it is computationally expensive due to the explicit
imposition of no-slip and no-penetration boundary conditions at the surface of
many particles moving in a turbulent medium. This is incredibly challenging
when particles are tiny and there is a scale separation between the particle
size and the smallest (Kolmogorov) turbulence scale due to the need for
extreme resolution requirements. Fortunately, in this case, one may be able to
resort to the so-called point-particle approximation (PP-DNS), where
interphase coupling is considered to be localized to a point.
Indeed, the point-particle approximation has been the method of choice for
simulating the particle dynamics in turbulent wall flows. In these cases, it
is assumed that the local properties of an undisturbed flow at the particle
position drive the dispersed-phase dynamics (Maxey & Riley, 1983; Gatignol,
1983). In the case of highly inertial particles (i.e., large particle-to-fluid
density ratios), the so-called Maxey-Riley-Gatignol equations simplify to a
drag force term (Arcen et al., 2006), which nevertheless yields highly non-
trivial particle dynamics, even in isotropic turbulence (Toschi & Bodenschatz,
2009). In wall-bounded turbulent flows, the inhomogeneous turbulence results
in even richer dynamics, with the particle distribution driven by the
interplay between small-scale clustering, turbophoresis, and the interaction
between the particles and near-wall turbulence structures (Reeks, 1983;
Soldati & Marchioli, 2009; Sardina et al., 2012), resulting in very
inhomogeneous particle concentrations peaking at the wall, with strong
preferential sampling of low-speed regions, as reproduced in a plethora of
numerical studies, such as Fessler et al. (1994); Uijttewaal & Oliemans
(1996); Marchioli et al. (2003); Kuerten (2006); Marchioli et al. (2008);
Soldati & Marchioli (2009); Bernardini et al. (2013); Jie et al. (2022).
When two-way coupling effects are important, a point-particle DNS must
describe the back-reaction of the dispersed phase in the flow. This is a major
challenge, as the point-particle dynamics are driven by a local _undisturbed_
fluid velocity, while the local flow field is being disturbed by the particles
(see, e.g. Gualtieri et al., 2015). The classical approach, known as the
particle-in-cell method, was developed by Crowe et al. (1977) and is widely
used – even in the present study – but it requires sufficiently high number of
particles per grid cell and cannot reproduce simple benchmarks of a
sedimenting sphere in a quiescent medium. Indeed, approaches for a consistent
treatment are being actively investigated; see Gualtieri et al. (2015);
Horwitz & Mani (2016); Ireland & Desjardins (2017); Horwitz & Mani (2020);
Horwitz et al. (2022); Horwitz (2023). Investigations of particle-laden
turbulent flows in the two-way-coupling regime are found in e.g. Vreman et al.
(2009); Zhao et al. (2010); Kuerten (2016); Richter & Sullivan (2014);
Capecelatro et al. (2018); Wang & Richter (2019); Battista et al. (2023).
While employing the point-particle approximation to small inertial particles
is physically sound, validating the fidelity of the approximation in one- and
two-way coupling conditions remains a challenge. Experimental data are
available (Eaton & Fessler, 1994; Kaftori et al., 1995). Still there are few
parameter-matched numerical studies due to limitations in terms of Reynolds
number and well-controlled experiments, and only in recent years, efforts in
this direction have started to appear (Wang et al., 2019). Fortunately, PR-DNS
of small inertial particles in turbulence has become possible, thanks to the
continuous growth in available computer power and development of efficient
numerical methods, with the first direct comparisons between point-particle
models and particle-resolved simulations starting to appear for forced
homogeneous isotropic turbulence and decaying HIT with moving particles
(Schneiders et al., 2017b, a; Mehrabadi et al., 2018; Fröhlich et al., 2018),
and turbulent channel flow (Horne & Mahesh, 2019; Costa et al., 2020a).
These results from PR-DNS of particles in wall-bounded turbulence confirm that
a sole drag force may not suffice to accurately reproduce the particle
dynamics even for relatively small particles with a large density ratio under
one-way coupling conditions. Where the shear rate is high near the wall,
_lift_ forces are also important. This has already been suggested in early
works using PP-DNS, since the work by McLaughlin (1989) (see also Botto et
al., 2005). In this regard, while Arcen et al. (2006) reported that the lift
force has a negligible impact on the dispersed phase statistics, Marchioli et
al. (2007) and Shin et al. (2022) found that the inclusion of the lift force
in PP-DNS can lead to weakened near-wall particle accumulation in upward and
horizontal channels, respectively; McLaughlin (1989) showed that the inclusion
of the lift force resulted in a higher deposition rate; Wang et al. (1997)
found that neglecting the lift force resulted in a slight reduction in the
deposition rate.
While these findings are not necessarily contradictory, as there are some
variations in the governing parameters in the different studies and lift force
models, many questions remain elusive: _(1) which form of lift force model is
appropriate for reproducing with high fidelity the dynamics of small inertial
particles in a turbulent wall flow; (2) under which conditions do lift forces
matter in the particle dynamics; and (3) what are the consequences of choosing
an inaccurate lift force model in dispersed phase dynamics and near-wall
turbulence modulation._
In the recent direct comparison between PR-DNS and PP-DNS for small particles
in turbulent channel flow by Costa et al. (2020a) (see also Costa et al.,
2020b), it was shown that lift models are vital for reproducing the near-wall
particle dynamics, at least for inertial particles with a size of the order of
one viscous wall unit. Surprisingly, a modified Saffman lift model predicted
the particle dynamics perfectly, with the Saffman force scaled by a normalized
shear rate, while conventional lift models yield poorer predictions (Costa et
al., 2020b). In addition to reducing the near-wall concentration peak, it was
also seen that lift force causes a large increase in correlated
streamwise–wall-normal particle velocity fluctuations. This quantity has
dramatic drag-increasing effects for sufficiently high mass loading, as it
modulates the streamwise momentum budget; see Costa et al. (2021). This direct
link between lift force and drag increase makes their accurate modelling
crucial.
In the present work, we address the three questions presented above by
performing DNS of turbulent channel flow laden with small inertial particles,
using the point-particle approximation, with three different lift models from
the classical Saffman lift model to the one that perfectly predicts previous
PR-DNS data. We consider different inner-scaled particle sizes $D^{+}\sim
1-0.1$ in one and two-way coupling conditions for a Stokes number that is
known to feature strong wall accumulation and preferential concentration in
low-speed regions. Our analysis shows that currently available lift models are
bound to fail near the wall for small inertial particles due to a vanishing
particle-to-fluid slip velocity. We then use two-way coupling simulations to
illustrate how different lift force models can result in _qualitatively_
different turbulence modulation (i.e., turbulence attenuation vs. turbulence
enhancement). Near-wall accumulation is still significantly reduced for the
smallest particle size ($D^{+}=0.1$). Still, the lift-induced increase of
correlated velocity fluctuations near the wall is negligible, and thus, lift
force has little consequences in turbulence modulation.
This paper is organized as follows. Next, in §2, the numerical method, lift
force models, and computational setups are described. Then, in §3, the effects
of lift force models on the particle dynamics and near-wall accumulation are
investigated using one-way coupling DNS, based on which we try to explore the
reason for the failure of conventional lift force models. Following this, we
qualitatively demonstrate the effect of lift force models on turbulence
modulation and momentum transfer with two-way coupling DNS. Finally,
conclusions are drawn in §4.
## 2 Methodology
### 2.1 Governing equations and numerical method
The fluid phase is governed by the incompressible Navier–Stokes equations,
$\begin{gathered}\nabla\cdot\mathbf{u}=0,\\\
\rho\left(\partial_{t}\mathbf{u}+(\mathbf{u}\cdot\nabla)\mathbf{u}\right)=-\nabla
p+\mu\nabla^{2}\mathbf{u}+\mathbf{f},\end{gathered}$ (1)
where $\mathbf{u}$ denotes the fluid velocity vector, $p$ is the fluid
pressure, $\mu$ is the dynamic viscosity, $\rho$ is the fluid density, and
$\mathbf{f}$ is the particle feedback force to the fluid phase in the case of
two-way coupling point-particle simulations, here computed by using a standard
particle-in-cell (PIC) approach, which spreads the particle force to the
nearest Eulerian grid points with a linear kernel (Boivin et al., 1998; Lee &
Lee, 2019; Zhang et al., 2023).
We consider $x$, $y$, and $z$ as the streamwise, wall-normal, and spanwise
directions, respectively. No-slip and no-penetration boundary conditions are
specified at the domain walls, while periodicity is applied along the
streamwise and spanwise directions. These equations are discretized in space
using a pseudospectral approach along $x$ and $z$ and second-order finite
differences along $y$. Wray’s low-storage third-order Runge–Kutta scheme is
employed for time marching (see Wray, 1990).
Spherical particles with density $\rho_{p}\gg\rho$ and particle diameter $D$
in the absence of gravity are tracked with a standard Lagrangian point-
particle method, with their dynamics governed by
$m_{p}\dot{\mathbf{U}}_{p}=\mathbf{F}_{d}+\mathbf{F}_{l}\mathrm{,}\;\;\dot{\mathbf{X}}_{p}=\mathbf{U}_{p}\mathrm{,}$
(2)
with the Schiller-Naumann drag force
$\mathbf{F}_{d}=-3\pi\mu
D\mathbf{U}_{s}\left(1+0.15\Rey_{p}^{0.687}\right)\mathrm{,}$ (3)
where $\mathbf{U}_{p}$, $\mathbf{X}_{p}$, $\mathbf{F}_{d}$ and
$\mathbf{F}_{l}$ are the particle velocity, position, drag and lift forces,
$m_{p}$ is the particle mass,
$\mathbf{U}_{s}=\mathbf{U}_{p}-\mathbf{u}|_{\mathbf{x}=\mathbf{X}_{p}}$ is the
local slip velocity evaluated at the particle location, and $\Rey_{p}$
$=\left|\mathbf{U}_{s}\right|D/{\nu}$ is the particle Reynolds number. Since
the particle-to-fluid density ratio is high, other dynamic effects such as
added mass, fluid acceleration, and the Basset history force are negligible.
While in most practical scenarios of particle-laden wall transport, gravity is
important at high density ratios, we neglect it in the present work. We do
this to isolate the interplay between more intricate wall accumulation
mechanisms (e.g., turbophoresis) and lift forces, in a flow which could
otherwise feature significant settling effects. Finally, for the sake of
simplicity, a perfectly elastic hard-sphere rebound is employed for particle-
wall interactions.
#### Lift force models
We consider standard shear-induced lift force models where lift force
$\mathbf{F}_{l}$ acts perpendicular to the local shear sampled by the particle
and points along $\boldsymbol{\omega}\times\mathbf{U}_{s}$, where
$\boldsymbol{\omega}=\left(\nabla\times\mathbf{u}\right)|_{\mathbf{x}=\mathbf{X}_{p}}$
is the undisturbed flow vorticity evaluated at the particle position. For
convenience, let us express the lift force in terms of the dimensionless lift
coefficient, $C_{L}$, defined as
$C_{L}=\frac{\mathbf{F}_{l}}{\frac{1}{8}\pi\rho
D^{2}|\mathbf{U}_{s}|^{2}}\cdot\frac{\boldsymbol{\omega}\times\mathbf{U}_{s}}{|\boldsymbol{\omega}\times\mathbf{U}_{s}|}.$
(4)
Assuming the Oseen length $\ell_{u}=\nu/{\left|\mathbf{U}_{s}\right|}$ is much
larger than the Saffman length
$\ell_{\omega}=\sqrt{\nu/{\left|\boldsymbol{\omega}\right|}}$, i.e.
$\varepsilon\gg 1$ with
$\varepsilon=\sqrt{\left|\boldsymbol{\omega}\right|\nu}/\left|\mathbf{U}_{s}\right|$,
an explicit lift solution can be derived:
$C_{L}=\frac{18}{\pi^{2}}\varepsilon J(\varepsilon),$ (5)
with
$J(\varepsilon\gg 1)=J^{\infty}=2.255.$ (6)
Hereafter, the lift expression (5) together with the $J$ function by Eq. (6)
will be referred to as the Saffman model, which is expected to be valid in the
double limits $\Rey_{p}\to 0$ and $\Rey_{\omega}\to 0$ provided that
$\varepsilon\gg 1$, where
$\Rey_{\omega}=\left|\boldsymbol{\omega}\right|D^{2}/\nu$ is the shear
Reynolds number.
Still in the double limits $\Rey_{p}\to 0$ and $\Rey_{\omega}\to 0$, the $J$
function in Eq. (5) at finite $\varepsilon$ turns out to be a volume integral
in Fourier space. Its value cannot be put in closed form but was estimated
numerically in Asmolov (1989), McLaughlin (1991), and more recently, Shi &
Rzehak (2019). Based on these numerical data, various empirical correlations
of $J(\varepsilon)$ were proposed (see Shi & Rzehak (2019) for a comprehensive
review), with the most commonly used being the one proposed by Mei (1992),
i.e.,
$J(\varepsilon)=0.3J^{\infty}\left(1+\tanh\left[\frac{5}{2}\left(\log_{10}\varepsilon+0.191\right)\right]\right)\left(\frac{2}{3}+\tanh(6\varepsilon-1.92)\right).$
(7)
Hereafter, the lift expression (5) together with the $J$ function by Eq. (7)
will be referred to as the Mei model, which is expected to be valid
irrespective of $\varepsilon$, provided that $\Rey_{p}$ and $\Rey_{\omega}$
are small.
The flow is assumed to be unbounded for the two lift models introduced above.
In particle-laden channel flows, near-wall accumulation of particles is often
observed, as discussed above. Typically, the peak in the particle
concentration appears within the viscous sublayer, i.e., for $y^{+}\leq 5$,
wherein the wall effect on the lift is crucial if $D^{+}\geq 1$ (Balachandar &
Eaton, 2010; Shi et al., 2021). In this context, the explicit lift solution
can be derived if the wall lies in the “inner region” of the flow disturbance
in the low-$\Rey_{p}$ limit, specifically, if the separation between the
particle and the wall, $\ell$, is much smaller than the inertial lengths
($\ell\ll\min{(\ell_{u},\ell_{\omega})}$). The corresponding lift solutions
(see Shi & Rzehak (2020) for a comprehensive review) take the general form
$C_{L}=A+BSr+CSr^{2},$ (8)
where $Sr=\left|\boldsymbol{\omega}\right|D/\left|\mathbf{U}_{s}\right|$ is
the normalized shear rate, and $A$, $B$, and $C$ are pre-factors that are
independent of $Sr$. Hereafter, the terms proportional to $Sr$ and $Sr^{2}$
will be referred to as the linear and quadratic inner contributions
$C_{L,\text{in}}^{\text{linear}}$ and $C_{L,\text{in}}^{\text{quad}}$,
respectively.
Situations where the wall is in the “outer region” of the flow disturbance
($\ell\gg\max{(\ell_{u},\ell_{\omega})}$), while still in the limit
$\Rey_{p}\to 0$, were considered in Asmolov (1990), McLaughlin (1993) and
Takemura et al. (2009). Assuming that $Sr=\Rey_{\omega}^{1/2}\varepsilon\leq
O(1)$ (i.e. the same condition for the shear rate as considered in Saffman
(1965)), the outer-region lift approaches the Saffman solution (5), namely
$C_{L,\text{out}}^{\text{linear}}\propto\varepsilon$. Together with the
scaling in the inner-region that $C_{L,\text{in}}^{\text{linear}}\propto Sr$,
it appears that the inner-region lift contribution transitions into the outer-
region by a pre-factor $\varepsilon/Sr$. If one assumes that the quadratic
contribution follows the same transition, it may be speculated that
$C_{L,\text{in}}^{\text{quad}}$ scales as $\varepsilon Sr$ in the outer
region, i.e. $C_{L,\text{out}}^{\text{quad}}\propto\varepsilon Sr$. For the
particle-laden channel flow,
$Sr\approx\frac{(u_{\tau}^{2}/\nu)D}{u_{s}}=\left(\frac{u_{s}}{u_{\tau}}\right)^{-1}D^{+},$
(9)
where $u_{s}$ is the streamwise mean slip velocity, namely
$u_{s}=\langle\mathbf{U}_{s}\cdot\mathbf{e}_{x}\rangle$ with $\mathbf{e}_{x}$
the unit vector along the streamwise direction; $u_{\tau}$ is the conventional
wall friction velocity. As will be demonstrated in figure 5 in section 3.1,
$u_{s}$ might change its sign in the buffer layer, leading to extremely large
values of $Sr$ in the near-wall region.
The analysis above implies that there is also a quadratic lift contribution
$C_{L}^{\text{quad}}\propto\varepsilon Sr$, which might dominate the lift
generation in the inner wall region. This is consistent with the form of the
lift force proposed in Costa et al. (2020a, b), where:
$C_{L}=\frac{18}{\pi^{2}}\varepsilon SrJ^{\infty},$ (10)
with $J^{\infty}=2.255$ according to Saffman (1965). Despite its simplified
form, this correlation aligns well with their PR-DNS results. Hereafter, the
lift model by Eq. (10) will be referred to as the CBP model.
Finally, it should be noted that in practice, to avoid singularities in the
numerical calculation, we implemented a slight variant of the lift coefficient
as described in (4), where the quotient on the right-hand side to define the
unit vector is modified to
$\boldsymbol{\omega}\times\mathbf{U}_{s}/|\boldsymbol{\omega}||\mathbf{U}_{s}|$.
In practice, the results are not very sensitive to this choice, as
$\boldsymbol{\omega}\times\mathbf{U}_{s}$ is nearly aligned with the wall-
normal direction.
### 2.2 Validation
Figure 1: $(a)$ Normalized particle volume fraction $\phi/\Phi$ and outer-
scaled second-order moments of particle velocity: $(b)$ streamwise velocity
r.m.s., $(c)$ wall-normal velocity r.m.s., $(d)$ Reynolds shear stress
profile. The points (colour online) denote the corresponding lift-included 1WC
PP-DNS from Costa et al. (2020b). Here $U_{b}$ is the flow bulk velocity.
The present code has demonstrated successful applications in wall-bounded
particle-laden flows, including open channel flow (Wang & Richter, 2019; Gao
et al., 2023) and planar Couette flow (Richter & Sullivan, 2014; Richter,
2015) loaded with inertial particles, but the lift force is neglected. Hence,
for completeness, we present the validation of the one-way-coupling PP-DNS
code, including the lift force models. Specifically, we reproduced in figure 1
the PP-DNS of particle-laden turbulent channel flow in the one-way coupling
regime using these three lift models, as reported in Costa et al. (2020a, b)
in the one-way coupling regime. This corresponds to case CL1 in Table 1. The
agreement is excellent, which validates our implementation.
### 2.3 Computational parameters
We perform point-particle DNS of channel flow at $\Rey_{\tau}=180$ in a
computational domain $(L_{x},L_{y},L_{z})=(6h,2h,3h)$, where $h$ is the half
channel height. The domain is discretized on
$(N_{x},N_{y},N_{z})=(160,320,160)$ grid points; the grid is slightly
stretched to refine the near-wall resolution, corresponding to a grid spacing
of $(\Delta_{x}^{+},\Delta_{y}^{+},\Delta_{z}^{+})=(6.75,0.9,3.375)$, with the
conventional ‘$+$’ superscript denoting viscous wall scaling. Note that the
grid spacing near the wall is comparable to that of the largest particle
sizes. All simulations are carried out at a constant time-marching step fixed
at $\Delta t^{+}=0.1$ (normalized by viscous unit $\nu/u^{2}_{\tau}$), which
corresponds to $\text{CFL}\approx 0.4$. The DNS results confirm this time step
to be sufficiently small that particles and fluid elements could not pass
through a grid cell per time step (Zheng et al., 2021). The total simulation
time is $T\approx 450h/u_{\tau}$ (about $80\,000$ viscous time scales
$\nu/u_{\tau}^{2}$), which is long enough to guarantee converged statistics
(Sardina et al., 2012).
Case | $D^{+}$ | $\rho_{p}/\rho$ | $N_{p}$
---|---|---|---
CPF | particle-free case
CL1 | 3 | 100 | $5\times 10^{4}$
CM1 | 1 | 900 | $1.5\times 10^{5}$
CS1 | 0.1 | 90000 | $1.5\times 10^{6}$
Table 1: Physical parameters used in the present point-particle DNS campaign.
$N_{p}$ denotes the total number of particles, and $D^{+}$ is inner-scaled
particle diameter. The particle density ratio was adjusted to keep the same
value of the Stokes number $St^{+}=\tau_{p}/(\nu/u_{\tau}^{2})=50$ and bulk
solid mass fraction $\Psi_{m}=0.337$ for all cases.
We consider three different setups for varying particle diameter, while
$St^{+}=\tau_{p}/(\nu/u_{\tau}^{2})=50$ is fixed by varying the particle
density. These parameters are described in table 1. This target particle
Stokes number, was chosen since it is known to feature highly inhomogeneous
particle distributions in wall turbulence; see Sardina et al. (2012).
Moreover, it should be noted that a value of $D^{+}=3$ corresponds to about
one Kolmogorov length scale in the channel bulk, an order of magnitude which
is often investigated in the literature using PP-DNS (see, e.g., Bernardini,
2014; Motoori et al., 2022; Zhang et al., 2022). We performed one- and two-way
coupling simulations at a fixed mass fraction for different particle sizes and
lift force models. Naturally, in the one-way coupled simulations, the mass
fraction is not a governing parameter, and the high number of particles
ensures statistical convergence of the results. In the two-way coupling PP-
DNS, the bulk mass fraction is fixed to about $\Psi_{m}\approx 30\%$, to
ensure significant turbulence modulation and highlight the effects of lift
force model choice in the flow statistics.
## 3 Results and discussion
### 3.1 Particle dynamics
We start by describing the dynamics of the dispersed phase under the effect of
lift force models using a simple, one-way coupling approach. This allows us to
measure observables sampled by the particle that dictate the validity region
of the lift models (e.g., $Sr$, $\varepsilon$, and $\Rey_{p}$). As we will
see, current lift models are bound to fail in the buffer layer.
Figure 2 shows the profiles of the normalized mean particle concentration for
all particle sizes. When the lift force is not taken into account, the
preferential accumulation of particles in the wall region is significant,
which has been widely observed (Picano et al., 2009; Sardina et al., 2012; Lee
& Lee, 2015; Gao et al., 2023; Gualtieri et al., 2023) and termed as
turbophoresis (Reeks, 1983; Johnson et al., 2020). However, the particles show
less wall accumulation when subjected to lift forces, and this effect is more
pronounced the larger the particle size. The Saffman model yields a much
smaller peak, while the other two models feature smaller deviations, with the
highest peak corresponding to the Mei model. When the particles are very small
($D^{+}=0.1$), the differences in the lift force models become almost
indistinguishable, with all the models robustly showing the same reduction of
wall accumulation. Yet, the difference in accumulation in the absence of lift
is clear.
Figure 2: Local particle mass fraction $\phi$, normalized by the bulk value
$\Phi$ as a function of the wall-normal distance (1WC). $(a)$ CL1, $(b)$ CM1,
$(c)$ CS1. The inset shows the normalized particle volume fraction, versus the
inner-scaled wall distance.
Figure 3: Inner-scaled instantaneous streamwise velocity fluctuation and the
corresponding instantaneous snapshots of particle locations in $x$-$z$
($y^{+}=12$, left column) and $y$-$z$ ($x^{+}=540$, right column) planes (Case
CL1, 1WC). The wall-parallel plane shows all particles with inner-scaled wall
normal position $Y_{p}^{+}\leq 12$. $(a)$ Lift-free, $(b)$ Saffman model,
$(c)$ Mei model, $(d)$ CBP model.
To qualitatively describe the lift force effect on the spatial localization of
wall accumulation, figure 3 shows instantaneous snapshots of streamwise
velocity fluctuations in a wall-parallel plane within the buffer layer
($y^{+}=12$) and in a plane of constant streamwise location ($x^{+}=540$),
along with the corresponding particle positions for the case with largest
particles CL1. Expectedly, the particles show strong spatial localization and
inhomogeneous distribution, with larger local density occurring in elongated
clusters (Sardina et al., 2012). Effects of the lift force model on the gross
characteristics of the particle distribution are readily discernible. One
notable manifestation is the difference in particle numbers in the wall-
parallel ($x$-$z$) plane. Consistent with figure 2, the lift-free model yields
the highest number of particles, while the Saffman model yields the lowest.
Conversely, Mei and CBP models, fall somewhere between these two. This
indicates the weakening accumulation of particles near the wall by the lift
force, consistently with observations by Marchioli et al. (2007) and Shin et
al. (2022). The other manifestation concerns small-scale particle clustering.
At first glance, lift-free, Mei and CBP simulations show a more significant
tendency to over-sample low-speed regions than the Saffman one. This tendency
is checked in figure 4, showing the probability density functions (p.d.f.s) of
streamwise fluid velocity fluctuations sampled by the particles within the
$y^{+}\leq 12$ region. For larger particle sizes, particles under the Saffman
lift force sample a less focused range of streamwise velocity fluctuations
than other models for larger particle sizes. In particular, for $D^{+}=3$ and
$1$, the CBP and Mei cases show about the same tendency to sample low-speed
regions, differently than the Saffman model and comparable to the lift-free
case for $D^{+}=3$. Preferential accumulation gets more focused with
decreasing particle size, with the lift-free case producing the strongest
preferential accumulation for $D^{+}=1$ and $0.1$.
Figure 4: Probability density functions (p.d.f.s) of inner-scaled streamwise
fluid velocity fluctuations conditioned at the particle location within the
$y^{+}\leq 12$ region (1WC). $(a)$ CL1, $(b)$ CM1, $(c)$ CS1.
Figure 5: Inner-scaled streamwise mean particle velocity and slip velocity
profiles (1WC). $(a,d)$ CL1, $(b,e)$ CM1, $(c,f)$ CS1. The horizontal dashed
lines denote $\langle u_{s}\rangle=0$.
Figure 6: Second-order moments of mean particle (solid lines) and fluid
(dashed lines) velocity (1WC). $(a,b,c)$ Inner-scaled streamwise and $(d,e,f)$
wall-normal velocity variances, and $(g,h,i)$ Reynolds shear stress profiles.
The left $(a,d,g)$, middle $(b,e,h)$ and right $(c,f,i)$ columns denote Case
CL1, CM1 and CS1, respectively.
Figure 5 shows the inner-scaled profiles of streamwise mean particle velocity
$\langle u_{p}\rangle$ and local slip velocity $\langle u_{s}\rangle$.
Compared with the lift-free particles with $D^{+}\geq 1$, lift forces increase
$\langle u_{p}\rangle$ in the viscous sublayer and buffer layer, while the
differences between cases become negligible beyond $y^{+}\sim 30$. Once more,
particles under the Saffman lift force show the highest variation with respect
to the lift-free case, yielding the highest $\langle u_{p}\rangle$; particles
under the Mei lift force yield the lowest, and the ones under the CBP lift
force fall in between. Consistently, for the smallest particle sizes, the
overall modification of $\langle u_{p}\rangle$ by lift forces is quite small.
This tendency of lift forces becoming less important with decreasing particle
size is expected – as we will discuss later in more detail (§3.3), the
relative importance of lift to drag forces close to the wall roughly scales
with the inner-scaled particle diameter $D^{+}$.
At the wall, while the fluid velocity must vanish, particles can feature a
mean apparent slip where they flow faster than the fluid due to their inertia
(Zhao et al., 2012). This is clearly shown in the particle mean slip (figure
5$d,e,f$). At larger wall distances, the tendency of particles to over-sample
slower-than-average fluid velocity (Kiger & Pan, 2002; Baker & Coletti, 2021)
is also clearly reflected in the mean particle slip velocity. A similar trend
can be observed in the corresponding streamwise mean slip velocity profiles,
however, the difference in $\langle u_{s}\rangle$ is larger than $\langle
u_{p}\rangle$, and this is attributed to the negative apparent slip velocity
($\langle u|_{\mathbf{x}=\mathbf{X}_{p}}\rangle-\langle u\rangle<0$), which
reflects preferential sampling of slower-than-average fluid. Interestingly,
lift forces increase the apparent slip velocity near the wall. This can be
understood from the sign of the lift force in this region, which points
towards the wall due to the positive slip velocity. Particles sampling higher-
momentum regions will be driven towards the wall by the lift force, and their
inertia results in a higher mean slip.
The sign of $\langle u_{s}\rangle$ changes in roughly the same location of the
buffer layer, irrespective of the particle size and lift force model. This
mechanism seems relatively robust and has been observed in numerous other
studies with inertial particles at different Stokes number, and different flow
Reynolds number (see, e.g., Mortensen et al., 2008; Zhao et al., 2012; Wang &
Richter, 2020; Gao et al., 2023). It will be shown that this change in sign
makes lift force modelling in wall turbulence extremely challenging (see
discussions in §3.3).
Figure 6 shows the $x-y$ components inner-scaled velocity covariances for the
particle and fluid phases, respectively. The spanwise velocity variances are
hardly affected by lift forces, and thus not shown. Expectedly, away from the
wall where the mean shear is low, the particle statistics closely follow those
of the lift-free cases. Near the wall, instead, the lift force enhances the
streamwise and wall-normal particle velocity variances ($\langle u^{\prime
2}_{p}\rangle$ and $\langle v^{\prime 2}_{p}\rangle$) and Reynolds stress
($\langle u^{\prime}_{p}v^{\prime}_{p}\rangle$), even for the smallest
particle size ($D^{+}=0.1$). Enhancement of wall-normal velocity fluctuations
near the wall has indeed been reported in previous numerical and experimental
studies (Fong et al., 2019; Costa et al., 2020a). Indeed, as also shown in
Costa et al. (2020a), this enhancement of velocity fluctuations near the wall
cannot be reproduced when lift forces are not considered in the particle
dynamics. It seems that lift force is very effective at correlating streamwise
and wall-normal velocity fluctuations near the wall – under high shear rate,
small variations in streamwise velocity naturally translate to changes in
vertical acceleration through the lift forces, which in turn induce wall-
normal velocity fluctuations. Streamwise velocity fluctuations will also be
naturally amplified, since streamwise and wall-normal velocity fluctuations
are correlated through the mean shear. As we will see in §3.4, this
amplification of near-wall velocity fluctuations by lift forces has major
consequences on turbulent modulation and drag changes in the flow at high
particle mass loading.
### 3.2 Mechanism for near-wall accumulation
To understand the role of the lift force in altering the near-wall particle
distributions, it is essential to examine three primary mechanisms responsible
for wall-normal inertial particle transport. The first is the lift-induced
migration, which is the direct consequence of the inertial lift force. The
second is turbophoresis (Reeks, 1983), causing particle migration from regions
of higher to lower turbulence intensity. In wall-bounded flows, turbophoresis
results in strong particle accumulation in the viscous sublayer, where
turbulent fluctuations vanish. Note however that in most cases the peak in the
particle concentration does not appear at the wall but at a distance of $O(D)$
from the wall (Marchioli & Soldati, 2002; Costa et al., 2020a), owing to the
wall-particle collisions as well as hydrodynamic wall-particle interactions
(Goldman et al., 1967; Vasseur & Cox, 1977; Zeng et al., 2005). Finally, there
is a biased sampling effect, due to the tendency of particles to sample
ejection regions in the buffer layer featuring negative streamwise velocity
and positive (i.e. repelling) wall-normal velocity fluctuations (Marchioli &
Soldati, 2002; Sardina et al., 2012). As a result, biased sampling usually
provides a net slow particle drift away from the wall. The interaction among
the three mechanisms is sketched in figure 7 and will be elaborated in the
following.
Figure 7: Illustration of the interplay between different mechanisms
responsible for the wall-normal particle transport.
Lift-induced migration is directed along
$\boldsymbol{\omega}\times\mathbf{U}_{s}$, which in the case of a turbulent
wall flow is predominantly in the wall-normal direction. According to figure
5($d,e,f$), particles lead the liquid flow for $y^{+}\lesssim 10$ and lag
behind the flow further away from the wall. Hence, the lift-induced migration
is _towards_ the wall in the viscous sublayer and acts cooperatively with the
turbophoresis effect. The way these two mechanisms interact in the viscous
sublayer may seem surprising at first glance, as figure 2 clearly indicates a
suppression of the near-wall particle accumulation by lift force. The
reasoning behind this suppression may be understood by noting the following
two issues. First, the lift force is towards the channel center within the
buffer layer and beyond $y^{+}\approx 10$. The resulting outward migration,
together with the strong turbulence intensity in the buffer layer, tends to
entrain particles in this region outwards, leading to a net decrease in the
concentration in the region $y^{+}\lesssim 10$ as revealed from figure 2.
Second, the lift force influences the wall-normal particle transport by
turbophoresis. This phenomenon is linked to the particle wall-normal velocity
variance $\langle v^{\prime 2}_{p}\rangle$ (Reeks, 1983; Johnson et al.,
2020). Indeed, in the limit of vanishing particle Reynolds number, the
particle wall-normal momentum balance at steady state yields a turbophoresis
pseudo-force proportional to $\mathrm{d}\langle v^{\prime
2}_{p}\rangle/\mathrm{d}y$, which creates the migration of particles down
gradients in particle wall-normal velocity variance (Sikovsky, 2014; Johnson
et al., 2020). Recall figure 6($d,e,f$), where the variation of $\langle
v^{\prime 2}_{p}\rangle$ with the wall distance is depicted. Clearly, the lift
force tends to increase the “apparent inertia” of the particles, enhancing the
deviation between the particle velocity variance and that of the fluid within
the viscous sublayer. This leads to an attenuation in the corresponding
turbophoresis pseudo-force. In particular, for the two cases where $D^{+}\geq
1$ (figure 6$d,e$), the prediction with the Saffman lift model yields
$\mathrm{d}\langle v^{\prime 2}_{p}\rangle/\mathrm{d}y<0$ in the viscous
sublayer, indicating that the turbophoresis pseudo-force is outwards, driving
the particles away from the wall. Consequently, the lift force competes with
the turbophoresis pseudo-force in these two cases, leading to a highly
flattened near-wall peak in the particle fraction as seen in figure 2($a,b$).
In contrast, they cooperate in predictions using the Mei and the CBP models,
as figure 2($a$) shows. There, the lift force compensates the attenuation in
turbophoresis, making the negative slopes in $\phi(y^{+})$ from results using
Mei and the CBP models closely follow that in the lift-free case (highlighted
in the figure insets). Note, however, that the magnitude of the Mei lift (CBP
lift) is proportional to $Re_{\omega}^{1/2}$ ($Re_{\omega}$), which approaches
$D^{+}$ (${D^{+}}^{2}$) in the viscous sublayer. Consequently, for the two
cases where $D^{+}\leq 1$ the lift-induced migration is weak and incapable of
compensating the still-pronounced attenuation in turbophoresis revealed from
figure 6$(e,f)$. This is why in these two cases the slopes in $\phi(y^{+})$
from results employing the Mei and CBP models are smaller in magnitude than
that in the lift-free case. Still, the difference in accumulation with respect
to the lift-free case for smaller sizes is clear, due to near-wall particles
residing for long times near the wall while being subjected to small but
persistent lift forces.
Figure 8: The inner-scaled lift-induced acceleration (i.e., normalized by
$u^{3}_{\utau}/\nu$) profiles along the wall-normal direction (1WC). $(a)$
CL1, $(b)$ CM1, $(c)$ CS1. The horizontal dashed lines denote $\langle
F_{l,y}\rangle=0$.
Given the discussion above, it is worth examining in more detail the link
between the chosen lift model and the corresponding modulation in the
turbophoresis pseudo-force, particularly whether to which extent this
modulation relates to the lift force magnitude. Combing the turbophoresis
pseudo-force (figure 6$d,e,f$) and lift force as shown in figure 8, a
monotonic increase in the suppression of the turbophoresis pseudo-force with
increasing lift is observed at $D^{+}\leq 1$, while no such trend is evident
at $D^{+}=3$. In the latter case, the suppression of the turbophoresis pseudo-
force is most pronounced in the prediction using the Saffman model, even
though the magnitude of the corresponding lift force is only half that of the
prediction using the CBP model for $y^{+}\lesssim 5$. Hence, the differences
in peak concentration are not directly linked to the lift force magnitudes,
but rather to the modulation in the turbophoresis pseudo-force induced by
these models. In the present work, the turbophoresis pseudo-force is always
suppressed by the presence of lift, leading to a decrease in the peak
concentration. Hence, the precise relationship between the extent of this
suppression and the lift force magnitude remains unclear, making it
challenging to directly connect the chosen lift models with the changes in the
predicted concentration profiles. Finally, it would be worth investigating
these dynamics at lower Stokes numbers, as both turbophoresis and preferential
sweeping will be less pronounced and the nature of the interaction between
lift and turbulence may drastically change. Given the increased issues with
lift force modelling at lower Stokes number, discussed in the next section,
addressing this question should be the object of a future, dedicated study.
To gain insights into biased sampling of sweeps/ejection events under lift
forces, we carried out a quadrant analysis for the fluid flow experienced by
the particles. In the ($u^{\prime},v^{\prime}$) plane, with positive
$v^{\prime}$ directed away from the wall, the flow experienced by the particle
is categorized into four types of events: first quadrant events (Q1),
characterized by outward motion of high-speed fluid, with $u^{\prime}>0$ and
$v^{\prime}>0$; second quadrant events (Q2), characterized by outward motion
of low-speed fluid, with $u^{\prime}<0$ and $v^{\prime}>0$, which are usually
called ejections; third quadrant events (Q3), characterized by inward motion
of low-speed fluid, with $u^{\prime}<0$ and $v^{\prime}<0$; and finally,
fourth quadrant events (Q4), which represent motions of high-speed fluid
toward the wall, with $u^{\prime}>0$ and $v^{\prime}<0$, and are usually
called sweeps.
Figure 9: Joint p.d.f.s of the streamwise and wall-normal fluid fluctuating
velocities seen by ascending ($v_{p}>0$) and descending ($v_{p}<0$) particles
within the $y^{+}\leq 12$ region (1WC). $(a)$ CL1, $(b)$ CM1, $(c)$ CS1. The
numbers in panels denote the proportion of ascending particles.
Figure 9 shows the joint p.d.f.s in the ($u^{\prime},v^{\prime}$) plane of the
fluid fluctuation seen by ascending ($v_{p}>0$) and descending ($v_{p}<0$)
particles within the $y^{+}\leq 12$ region. The first observation is that a
larger proportion of particles tends to move outwards (ascending),
irrespective of the presence of lift force. The proportion of ascending
particles without lift force decreases as the particle size decreases, whereas
particles subject to lift force exhibit an inverse trend. Second, the lift-
free ascending particles are highly likely to sample ejections (Q2), while the
lift-free descending ones are prone to sample sweeps (Q4) for particles of
$D^{+}\geq 1$. This agrees with experimental observations in channel flow
laden with finite-size particles (Kiger & Pan, 2002; Yu et al., 2016; Baker &
Coletti, 2021). However, the particles under the effect of lift forces tend
not to over-sample sweeps, irrespective of the wall-normal particle velocity
sign, which indicates that the wall-normal velocity of the particles is less
correlated to the fluid one. This also confirms that inertial particles under
the effect of lift forces are not prone to slowly drive toward the wall in
low-speed regions, because the competing wall-repelling effect of lift forces
seems to dominate their dynamics. Close inspection of the high-intensity
region of the joint p.d.f.s on the $u^{\prime}$-axis, we note that particles
under the Saffman lift force tend to sample a wider range of negative
$u^{\prime}$ regions, compared with particles under the Mei and CBP lift
forces. This aligns with the p.d.f.s of $u^{\prime}$ as shown in figure 4.
Third, irrespective of the presence of the lift force or wall-normal particle
velocity sign, the smallest particles ($D^{+}=0.1$) over-sample ejection (Q2)
and inward motions of low-speed fluid regions (Q3). Sweep events, bringing
high-momentum fluid towards the wall are only weakly experienced by particles.
Overall, the percentage of particles sampling Q2 events only slightly
surpasses that of Q3, suggesting that biased sampling may play a minor role in
driving high inertial particles ($St^{+}=50$) to depart from the inner wall
region. This observation aligns with numerical results from Marchioli &
Soldati (2002), where biased sampling is found to be more efficient for
transferring particles with smaller inertia.
### 3.3 Emerging challenges in lift force modelling
The aforementioned results allow us to address several issues concerning the
effects of lift force in turbulent wall flows, which may have been previously
overlooked in PP-DNS simulations. Ideally, such discussions would be based on
results from PR-DNS simulations. However, in this instance, we will
extrapolate from PP-DNS results, in conjunction with the CBP model, as an
analogue to PR-DNS outcomes. This approach is justified by findings from Costa
et al. (2020a, b), which demonstrate that PP-DNS simulations are capable of
satisfactorily reproducing the corresponding PR-DNS results for the most
challenging case considered in this study, with $D^{+}=3$.
Let us first address the circumstances under which the effects of lift force
could be considered negligible for large inertial particles, and thereby
disregarded in PP-DNS simulations. Given that biased sampling does not play a
key role in the wall-normal transport of inertial particles, a criterion based
on the ratio of lift to the turbophoresis pseudo-force seems sufficient to
describe the relative significance of the lift force. In this context, the
dimensionless particle size $D^{+}$ seems a promising candidate, owing to the
following two reasons. Firstly, ${D^{+}}^{2}$ approaches the shear Reynolds
number $Re_{\omega}$ in the viscous sublayer; the latter directly measures the
magnitude of the shear lift irrespective of the choice of lift models
(Saffman, 1965; McLaughlin, 1991; Mei, 1992; Shi & Rzehak, 2019; Costa et al.,
2020b). Secondly, the turbophoresis effects seem to decay with increasing
$D^{+}$. This is clearly revealed by comparing the slopes of $\langle
v^{\prime 2}_{p}\rangle(y^{+})$ revealed from figure 6$(d,e,f)$ which
correspond to results with increasing $D^{+}$. Glueing these two aspects
together, it seems that the force ratio of lift to turbophoresis roughly
scales as ${D^{+}}^{k}$ with $k>0$. Although the exact value of $k$ is
infeasible to be determined, the results summarized in §3.1 made it clear that
by no means should the lift force be neglected for $D^{+}\geq 1$. On the other
hand, it is reasonable for the lift force to be neglected for $D^{+}\leq 0.1$,
except for the underprediction of near-wall accumulation discussed above
(recall figure 2).
Figure 10: The p.d.f.s of $u_{s}/(\omega_{z}D)$ along the wall-normal
direction (1WC DNS with the CBP model). $(a)$ CL1, $(b)$ CM1, $(c)$ CS1. The
vertical dashed lines denote $u_{s}/(\omega_{z}D)=\pm 1$.
We should now address a major issue in commonly applied lift force models for
PP-DNS simulations of turbulent wall flows. As outlined in section 2.1, any of
the shear-lift models originating from the pioneering work of Saffman (1965)
are rigorously applicable only in the double limits $\Rey_{p}\to 0$ and
$\Rey_{\omega}\to 0$. In the viscous sublayer, $\Rey_{p}\approx D^{+}(\langle
u_{s}\rangle/u_{\tau})\approx D^{+}$ for $D^{+}\geq 1$ according to figure 5,
whereas $\Rey_{\omega}\approx{D^{+}}^{2}$. Consequently, none of the two
limits are satisfied for $D^{+}\geq 1$. In addition to the double limits
above, there is a constraint concerning the dimensionless shear rate $Sr$.
This is made clear by noting that $Sr=\Rey_{\omega}^{1/2}\varepsilon$ and
hence $Sr\leq O(1)$ even in the strong shear limit, $\varepsilon\geq 1$,
considered in Saffman (1965). This is why in the wall-bounded situation the
“outer-region” lift solutions (Asmolov, 1990; McLaughlin, 1993; Takemura et
al., 2009) following the methodology of Saffman (1965) match with the
corresponding “inner-region” solutions (Cherukat & McLaughlin, 1994; Magnaudet
et al., 2003; Shi & Rzehak, 2020) only for $Sr=O(1)$. For particle-laden
channel flows with small inertial particles (Mortensen et al., 2008; Zhao et
al., 2012; Wang & Richter, 2020; Gao et al., 2023), this constraint is more
seriously violated than that for the double limits above, as the slip velocity
reverses at $y^{+}\approx O(10)$ (see, e.g., figure 5$(d,e,f)$ in the present
work), making $Sr\to\infty$ for particles within the buffer layer. This issue
is highlighted in figure 10, where we show the p.d.f.s of the inverse of $Sr$
(horizontal axis) as a function of the wall-normal distance $y^{+}$ (vertical
axis). Apparently, a large portion of particles experiences a strong shear
where $Sr\geq 1$, in particular for the two cases where $D^{+}\geq 1$.
Finally, it is worth noting that the probability of a particle experiencing
very high values of $Sr$ near the wall will increase with decreasing values of
$St^{+}$ as, for the same local shear rate, particles will experience an ever
smaller slip velocity. This exacerbates the violation of the $Sr=O(1)$
constraint, meaning that current lift force models are even less suitable for
turbulent channel flow laden with particles at lower Stokes numbers.
Additionally, the discussion above focuses on the lift force arising from the
shear in the ambient flow. In wall-bounded flows, as considered in this work,
the lift force may deviate significantly from its unbounded counterpart if the
particle is close to the wall. In brief, the presence of the wall leads to a
repulsive transverse force in the absence of the ambient shear (Vasseur & Cox,
1977; Zeng et al., 2005) and, in the presence of an ambient shear, a
suppression of the shear-induced lift (McLaughlin, 1993; Magnaudet et al.,
2003; Takemura et al., 2009). Although analytical solutions for the combined
lift force are achievable in some asymptotic limits (see, e.g., Wang et al.
(1997) and, more recently, Shi & Rzehak (2020) for an overview of these
solutions), there is so far no satisfactory way to glue these solutions
together to achieve a general treatment of the wall effect (Ekanayake et al.,
2020, 2021). Here, instead of trying to make any further progress in achieving
such a general solution, it is more feasible to discuss under which condition
may this intricate wall effect be neglected. Within the $O(D)$-depth layer
close to the wall, wall-particle collisions govern the wall-normal position of
the peak particle concentration. Hence, the hydrodynamic wall-particle
interaction may play a significant role only beyond $y^{+}\approx D^{+}$. On
the other hand, it is known that (McLaughlin, 1993) the amount of reduction of
the shear-induced lift due to the wall decreases as
$(\ell/\ell_{\omega})^{-5/3}$ and reduces to only $20\%$ of its magnitude in
the limit $\ell/\ell_{\omega}\to 0$ at $\ell/\ell_{\omega}\approx 3$. Hence,
the wall effect may be considered negligible if the position of the peak
particle concentration, which is of $O(D)$ from the wall, is larger than about
$3\ell_{\omega}$, which yields roughly $D^{+}\geq 3$. This finding, together
with the previous one where (shear-induced) lift force is non-negligible for
$D^{+}\geq 1$, indicates that such a wall effect can be disregarded except in
the overlap regime where $1\leq D^{+}\leq 3$.
In addition to the ambient shear, particle rotation is known to produce a lift
force as well (Rubinow & Keller, 1961). This lift contribution is often
neglected in prior PP-DNS simulations and not considered in the present work.
Nevertheless, it might be worthy to discuss in the following its relevant
importance. Assuming that the particle inertia is small enough for the
particle to always stay in the torque-free state, particle rotation leads to a
lift contribution of $C_{L}^{\Omega}=0.5Sr$ in the double limits considered by
Saffman (1965). The ratio between this spin-induced lift contribution and that
induced by ambient shear is proportional to $Re_{\omega}^{1/2}$, indicating
that $C_{L}^{\Omega}$ is merely a component of the second-order lift
contribution from the ambient shear. This is confirmed from the recent work by
Candelier et al. (2023), where second-order lift contributions (with respect
to $Re_{\omega}^{1/2}$) were obtained using matched asymptotic expansions. In
addition to the spin-induced contribution $C_{L}^{\Omega}=0.5Sr$, they also
obtained a second-order contribution from the ambient shear
$C_{L}^{{}^{\prime}(2)}=-0.505Sr$, which counterbalances the spin-induced
contribution. In other words, the effect of particle rotation on the lift is
negligible in the torque-free state. This conclusion is valid irrespective of
the presence of the wall (Cherukat & McLaughlin, 1994; Magnaudet et al., 2003;
Shi & Rzehak, 2020). However, large-inertia particles with a large density
ratio as considered in this work might not always stay torque-free.
Specifically, for particles with density ratio $\rho_{p}/\rho\geq 100$, the
relaxation time in response to torque is comparable with that to slip (see,
e.g., Bagchi & Balachandar (2002)). This finding, together with the pronounced
slip in the viscous sublayer revealed from figure 5$(a,b)$, indicates that
particles of $D^{+}\geq 1$ may experience a strong angular acceleration and
possibly a non-negligible lift contribution from forced rotation.
Investigating the effects of rotation by accounting for the particles’
conservation of angular momentum would therefore be interesting but is left
outside the scope of the present work due to the higher complexity associated
with drag and lift closures for angular momentum.
Finally, it is worth noting that while the CBP lift force model used in Costa
et al. (2020a, b) appears to provide reasonable predictions in some cases, it
is not universally valid. Specifically, it does not align with any of the
inner-region solutions (Shi & Rzehak, 2020), which are applicable for
particles extremely close to the wall (e.g., within the viscous sublayer), nor
does it approach the unbounded solution (Saffman, 1965; McLaughlin, 1991) for
particles located within the outer layer of the channel flow. The agreement
between the PP-DNS simulations utilizing this lift force model and PR-DNS
suggests that the model offers a satisfactory blending of results in the two
limits for the specific cases explored in Costa et al. (2020a) (see table 1).
A preliminary attempt to establish a more general lift force model can be
found in the work of Ekanayake et al. (2020, 2021), where a lift model
satisfying the above conditions was proposed in the limit of small shear
Reynolds numbers, i.e. $\Rey_{\omega}\ll 1$. However, the applicability of
this model is limited to $D^{+}\ll 1$ as, in the viscous layer,
$\Rey_{\omega}\approx(u_{\tau}^{2}/\nu)D^{2}/\nu=(D^{+})^{2}$. While this
provides some insights, its practicality may be limited since the lift force
no longer plays a pivotal role for $D^{+}<1$, as demonstrated here. Hence,
further work extending the work of Ekanayake et al. (2020, 2021) to finite
$\Rey_{\omega}$ (hence $D^{+}$) would be significant.
### 3.4 Lift-induced turbulence modulation
Finally, in this section, we use two-way coupling point-particle DNS to
illustrate that lift force models can have tremendous consequences in basic
integral quantities such as the overall drag. To achieve this, we use the same
parameters shown in table 1, which fixes the total mass fraction and particle
Stokes number while varying the particle diameter. We should reiterate that
high-fidelity two-way coupling point-particle DNS is highly challenging and is
currently being actively investigated. Still, the simple particle-in-cell 2WC
DNS employed here suffices to illustrate the impact of lift force model choice
on important turbulence metrics. In what follows, we illustrate the effects of
lift force on the overall drag and, analyze the streamwise momentum fluxes
that contribute to it.
Figure 11 shows the friction coefficient for the 2WC DNS, defined as
$C_{f}=2u^{2}_{\tau}/U^{2}_{b}$. While all small inertial particles have a
drag-increasing effect, the difference between lift force models is highly
amplified as the particle size increases. At $D^{+}=3$, in particular,
particles under the Saffman lift force show a drag reduction compared to the
other particle sizes, almost reaching an overall drag decrease. Instead, the
other lift models show a monotonic drag increase.
Figure 11: The wall friction coefficient $C_{f}$ versus the particle diameter
$D^{+}$ (2WC). The horizontal dashed line denotes the mean $C_{f}$ of the
particle-free case.
Figure 12: Budget of streamwise momentum $(a,c,e)$ and contribution of each
stress contribution $\tau_{i}$ to the mean wall friction coefficient
$C_{f}=2\tau_{w}/(\rho U_{b}^{2})$ $(b,d,f)$,
$1/(h^{2})\int_{0}^{h}6(h-y)\tau_{i}\mathrm{d}y/(\rho U_{b}^{2})$ normalized
by that of the particle-free case. $(a,b)$ CL1, $(c,d)$ CM1, $(e,f)$ CS1.
Different line colours in $(a,c,e)$ denote different cases, which match the
legend in $(c)$.
To better understand and quantify this difference, we follow the analysis in
Costa et al. (2021) and investigate the stress budget in the two-way coupling
limit of vanishing volume fraction, while keeping the mass fraction finite:
$\tau=\rho
u_{\tau}^{2}\left(1-\frac{y}{h}\right)\approx\underbrace{\left(\mu\frac{\mathrm{d}\left\langle
u\right\rangle}{\mathrm{d}y}-\rho\left\langle
u^{\prime}v^{\prime}\right\rangle\right)}_{\tau_{\nu}+\tau_{T,f}}-\underbrace{\left\langle\psi\right\rangle\left\langle
u_{p}^{\prime}v_{p}^{\prime}\right\rangle}_{\tau_{T,p}},$ (11)
where $\left<\psi\right>=\rho_{p}\left<\phi\right>$ is the mean local mass
fraction. The first two terms in the right-hand-side of this equation are the
viscous stress, fluid Reynolds stress, and particle Reynolds stress (hereafter
$\tau_{\nu}$, $\tau_{T,f}$, and $\tau_{T,p}$). This equation boils down to the
stress budget in the single-phase limit ($\tau_{\nu}$ and $\tau_{T,f}$),
corrected by an additional turbulent particle momentum flux, $\tau_{T,p}$ that
involves the local mass fraction and the particle counterpart of the Reynolds
shear stress. This term relates to the transfer of streamwise momentum by
correlated particle velocity fluctuations, and shows the importance of high
particle mass fraction near the wall when combined with correlated
streamwise–wall-normal particle velocity fluctuations. We have seen in §3.1
that, while the former decreases in the presence of lift force, the latter may
be increased by it.
The profiles of stress terms and the corresponding contributions of each
momentum transfer mechanism to the overall friction coefficient, normalized by
the overall friction coefficient of the particle-free case, are shown in
figure 12. On the left panels, the normalized shear stress profiles are shown,
while on the right we show the so-called FIK identity employed to a particle
suspension (Fukagata et al., 2002; Yu et al., 2021). In the case of a
turbulent plane channel flow, this identity decomposes the mean friction
coefficient ($C_{f}$) into a laminar and turbulent contribution. Here, the
latter can be further decomposed into a contribution due to the fluid phase
and one due to the particles. Each contribution is given by the following
weighted integral
$C_{f,i}=1/h^{2}\int_{0}^{h}6(h-y)\tau_{i}\,\mathrm{d}y/(\rho U_{b}^{2})$,
with $\tau_{i}$ being one the stress mechanisms described above. Note that the
residual of this stress budget is virtually zero, which validates the
approximation in equation (11).
As anticipated from figure 11, all 2WC DNS show increased drag compared to the
unladen flow. Yet, remarkably, different lift models show qualitatively
different turbulence modulation for the largest particle size ($D^{+}=3$).
Particles under the Saffman lift force have a major net turbulence attenuation
effect, which results in an almost drag-reducing flow with respect to the
unladen case. Conversely, particles under the Mei and CBP lift forces have a
net drag increase, mostly caused by an increase in the two-way coupling stress
term $\psi\left<u^{\prime}_{p}v_{p}^{\prime}\right>$. This results from the
amplification of particle Reynolds shear stresses
$\left<u^{\prime}_{p}v^{\prime}_{p}\right>$ discussed in section 3.1, which
compensate for the decrease in local mass fraction $\psi$. Naturally, as the
particle size decreases, the differences between lift models become less
pronounced but still noticeable at all particle sizes, with particles under
the Saffman lift force consistently showing the highest fluid turbulence
attenuation. The smallest particle size $D^{+}=0.1$, the two turbulent
stresses (fluid and particle) terms still show noticeable but in practice
negligible differences (recall figure 11).
Hence, it appears that the most important dynamical effects of lift forces in
two-way coupling conditions can be seen as finite-size effects that directly
modify single-point moments of particle velocity for $D^{+}\gtrsim 1$ (recall
figure 6). However, the lift force still affects near-wall turbulence
modulation when particles are relatively small, as they strongly affect the
mean particle concentration (recall figure 2), directly linked to turbulence
modulation via the two-way coupling stress term.
## 4 Conclusions
We have investigated the importance of lift forces in the dynamics of
particle-laden turbulent channel flow. Three relevant shear-induced lift force
models are analyzed using PP-DNS for varying particle size and fixed Stokes
number. Our results confirm that shear-induced lift forces strongly influence
particle dynamics. We have analyzed the main mechanisms for near-wall particle
accumulation under lift forces: lift-induced migration, turbophoresis, and
biased sampling. Very close to the wall, where the particle-to-fluid slip
velocity is positive and particle fluctuations are vanishing, lift forces and
turbophoresis cooperate in driving particles toward the wall. Particle-wall
interactions and biased sampling balance this effect. Further away from the
wall, lift forces and turbophoresis compete, with the former having a wall-
repelling effect. Concerning biased sampling, we observe that, for larger
particle sizes ($D^{+}\geq 1$), the lift-free ascending particles prefer to
sample ejections, while the descending ones tend to sample sweeps. Conversely,
particles under lift forces tend to sample ejection regions while experiencing
a marginal occurrence of sweep events, regardless of the employed lift model
or wall-normal particle velocity sign. This is again a consequence of the
competing wall-repelling effect of lift force. For the smallest particles
($D^{+}=0.1$), all cases exhibit over-sampling in ejection and inward motions
of low-speed fluid regions (i.e., Q3 events). Overall, the sampling percentage
of the ejections events only slightly surpasses that of Q3, suggesting that
biased sampling plays a minor direct role in near-wall particle transport.
However, the biased sampling of low-speed regions exposes particles for long
times to persistent lift forces, which results in significantly reduced
particle accumulation, even for small particle sizes.
These different dynamics have major consequences for turbulence modulation due
to a modification of the near-wall values of (1) the local particle
concentration, and (2) the correlated streamwise-wall-normal particle velocity
fluctuations. While the latter is only modified for sufficiently large
particle size and therefore can be seen as a finite-size effect, the former is
still non-negligible for relatively small particle sizes, and can still result
in small but visible changes in overall drag. Indeed, correlated particle
velocity fluctuations show little sensitivity to lift forces for sufficiently
small values of inner-scaled particle diameter ($D^{+}\lesssim 0.1$).
Conversely, for small particle sizes, the modification of near-wall particle
accumulation due to lift showed little sensitivity to the flavour of shear-
induced lift model. Hence, to accurately account for lift-induced turbulence
modulation with $D^{+}=O(1)$ particle sizes, reliable shear-induced lift
models are needed that correctly reproduce the particle turbulent momentum
flux. This could be seen as a necessary correction for finite-sized effects,
in the same spirit as the Faxén correction in the Maxey-Riley-Gatignol
equation.
While such correction is important for reliable particle-modeled simulations
of turbulent wall flows, our analysis has shown that current lift force models
are bound to lose their validity in the near-wall region for small inertial
particles. The reason is that the well-known tendency of particles to over-
sample low-speed regions in the buffer layer, along with their tendency to
flow faster than the fluid in the viscous sublayer, results in a region near
the wall of vanishing particle-to-fluid slip velocity. Naturally, this issue
will be even more pronounced at lower Stokes numbers, where the range of very
low slip velocities in regions of high shear is wider. Unfortunately, current
shear-induced lift force models are only available for non-dimensional shear
rates $Sr$ of order unity, which robustly tends to infinity within the buffer
layer. Hence, the condition of $Sr\leq 1$ required by these models is
seriously violated in wall turbulence with small inertial particles.
[Acknowledgements]The Cray XC40 Shaheen II at KAUST was used for all
simulations reported. PC thanks F. Picano from Padova University for
insightful discussions.
[Funding]This research was partially supported by the KAUST Office of
Sponsored Research (OSR) under award no. OSR-2019-CCF-3666 and under baseline
research funds of MP.
[Declaration of interests]The authors report no conflict of interest.
[Author ORCIDs] Wei Gao, https://orcid.org/0000-0001-7313-0058; Pengyu Shi,
https://orcid.org/0000-0001-6402-4720; Matteo Parsani,
https://orcid.org/0000-0001-7300-1280; Pedro Costa,
https://orcid.org/0000-0001-7010-1040.
## References
* Arcen et al. (2006) Arcen, B., Tanière, A. & Oesterlé, B. 2006 On the influence of near-wall forces in particle-laden channel flows. International Journal of Multiphase Flow 32, 1326–1339.
* Asmolov (1989) Asmolov, E. S. 1989 Lift force exerted on a spherical particle in a laminar boundary layer. Fluid Dynamics 24 (5), 710–714.
* Asmolov (1990) Asmolov, E. S. 1990 Dynamics of a spherical particle in a laminar boundary layer. Fluid Dynamics 25 (6), 886–890.
* Bagchi & Balachandar (2002) Bagchi, P. & Balachandar, S. 2002 Effect of free rotation on the motion of a solid sphere in linear shear flow at moderate Re. Physics of Fluids 14 (8), 2719–2737.
* Baker & Coletti (2021) Baker, L. J. & Coletti, F. 2021 Particle–fluid–wall interaction of inertial spherical particles in a turbulent boundary layer. Journal of Fluid Mechanics 908, A39.
* Balachandar & Eaton (2010) Balachandar, S. & Eaton, J. K. 2010 Turbulent dispersed multiphase flow. Annual Review of Fluid Mechanics 42, 111–133.
* Battista et al. (2023) Battista, F., Gualtieri, P., Mollicone, J.-P., Salvadore, F. & Casciola, C. M. 2023 Drag increase and turbulence augmentation in two-way coupled particle-laden wall-bounded flows. Physics of Fluids 35, 045133.
* Bernardini (2014) Bernardini, M. 2014 Reynolds number scaling of inertial particle statistics in turbulent channel flows. Journal of Fluid Mechanics 758, R1.
* Bernardini et al. (2013) Bernardini, M., Pirozzoli, S. & Orlandi, P. 2013 The effect of large-scale turbulent structures on particle dispersion in wall-bounded flows. International Journal of Multiphase Flow 51, 55–64.
* Boivin et al. (1998) Boivin, M., Simonin, O. & Squires, K. D. 1998 Direct numerical simulation of turbulence modulation by particles in isotropic turbulence. Journal of Fluid Mechanics 375, 235–263.
* Botto et al. (2005) Botto, L., Narayanan, C., Fulgosi, M. & Lakehal, D. 2005 Effect of near-wall turbulence enhancement on the mechanisms of particle deposition. International Journal of Multiphase Flow 31, 940–956.
* Brandt & Coletti (2022) Brandt, L. & Coletti, F. 2022 Particle-laden turbulence: Progress and Perspectives. Annual Review of Fluid Mechanics 54, 159–189.
* Candelier et al. (2023) Candelier, F., Mehaddi, R., Mehlig, B. & Magnaudet, J. 2023 Second-order inertial forces and torques on a sphere in a viscous steady linear flow. Journal of Fluid Mechanics 954, A25.
* Capecelatro et al. (2018) Capecelatro, J., Desjardins, O. & Fox, R. O. 2018 On the transition between turbulence regimes in particle-laden channel flows. Journal of Fluid Mechanics 845, 499–519.
* Cherukat & McLaughlin (1994) Cherukat, P. & McLaughlin, J. B. 1994 The inertial lift on a rigid sphere in a linear shear flow field near a flat wall. Journal of Fluid Mechanics 263, 1–18.
* Costa et al. (2020a) Costa, P., Brandt, L. & Picano, F. 2020a Interface-resolved simulations of small inertial particles in turbulent channel flow. Journal of Fluid Mechanics 883, A54.
* Costa et al. (2020b) Costa, P., Brandt, L. & Picano, F. 2020b Interface-resolved simulations of small inertial particles in turbulent channel flow – CORRIGENDUM. Journal of Fluid Mechanics 891, E2.
* Costa et al. (2021) Costa, P., Brandt, L. & Picano, F. 2021 Near-wall turbulence modulation by small inertial particles. Journal of Fluid Mechanics 922, A9.
* Crowe et al. (1977) Crowe, C. T., Sharma, M. P. & Stock, D. E. 1977 The particle-source-in cell (PSI-CELL) model for gas-droplet flows. Journal of Fluids Engineering 99 (2), 325–332.
* Eaton & Fessler (1994) Eaton, J. K. & Fessler, J. R. 1994 Preferential concentration of particles by turbulence. International Journal of Multiphase Flow 20, 169–209.
* Ekanayake et al. (2021) Ekanayake, N. I. K., Berry, J. D. & Harvie, D. J. E. 2021 Lift and drag forces acting on a particle moving in the presence of slip and shear near a wall. Journal of Fluid Mechanics 915, A103.
* Ekanayake et al. (2020) Ekanayake, N. I. K., Berry, J. D., Stickland, A. D., Dunstan, D. E., Muir, I. L., Dower, S. K. & Harvie, D. J. E. 2020 Lift and drag forces acting on a particle moving with zero slip in a linear shear flow near a wall. Journal of Fluid Mechanics 904, A6.
* Elghobashi (1994) Elghobashi, S. 1994 On predicting particle-laden turbulent flows. Applied Scientific Research 52, 309–329.
* Fessler et al. (1994) Fessler, J. R., Kulick, J. D. & Eaton, J. K. 1994 Preferential concentration of heavy particles in a turbulent channel flow. Physics of Fluids 6 (11), 3742–3749.
* Fong et al. (2019) Fong, K. O., Amili, O. & Coletti, F. 2019 Velocity and spatial distribution of inertial particles in a turbulent channel flow. Journal of Fluid Mechanics 872, 367–406.
* Fröhlich et al. (2018) Fröhlich, K., Schneiders, L., Meinke, M. & Schröder, W. 2018 Validation of Lagrangian two-way coupled point-particle models in large-eddy simulations. Flow, Turbulence and Combustion 101 (2), 317–341.
* Fukagata et al. (2002) Fukagata, K., Iwamoto, K. & Kasagi, N. 2002 Contribution of Reynolds stress distribution to the skin friction in wall-bounded flows. Physics of Fluids 14 (11), L73–L76.
* Gao et al. (2023) Gao, W., Samtaney, R. & Richter, D. H. 2023 Direct numerical simulation of particle-laden flow in an open channel at ${Re}_{\tau}=5186$. Journal of Fluid Mechanics 957, A3.
* Gatignol (1983) Gatignol, R. 1983 The Faxén formulas for a rigid particle in an unsteady non-uniform Stokes-flow. J. Méc. Théor. Appl. 2 (2), 143–160.
* Goldman et al. (1967) Goldman, A. J., Cox, R. G. & Brenner, H. 1967 Slow viscous motion of a sphere parallel to a plane wall—I Motion through a quiescent fluid. Chemical Engineering Science 22 (4), 637–651.
* Gualtieri et al. (2023) Gualtieri, P., Battista, F., Salvadore, F. & Casciola, C. M. 2023 Effect of Stokes number and particle-to-fluid density ratio on turbulence modification in channel flows. Journal of Fluid Mechanics 974, A26.
* Gualtieri et al. (2015) Gualtieri, P., Picano, F., Sardina, G. & Casciola, C. M. 2015 Exact regularized point particle method for multiphase flows in the two-way coupling regime. Journal of Fluid Mechanics 773, 520–561.
* Horne & Mahesh (2019) Horne, W. J. & Mahesh, K. 2019 A massively-parallel, unstructured overset method to simulate moving bodies in turbulent flows. Journal of Computational Physics 397, 108790.
* Horwitz (2023) Horwitz, J. A. K. 2023 Improved force models for Euler–Lagrange computations. In Modeling Approaches and Computational Methods for Particle-Laden Turbulent Flows, pp. 265–298. Elsevier.
* Horwitz et al. (2022) Horwitz, J. A. K., Iaccarino, G., Eaton, J. K. & Mani, A. 2022 The discrete Green’s function paradigm for two-way coupled Euler–Lagrange simulation. Journal of Fluid Mechanics 931, A3.
* Horwitz & Mani (2016) Horwitz, J. A. K. & Mani, A. 2016 Accurate calculation of Stokes drag for point–particle tracking in two-way coupled flows. Journal of Computational Physics 318, 85–109.
* Horwitz & Mani (2020) Horwitz, J. A. K. & Mani, A. 2020 Two-way coupled particle-turbulence interaction: Effect of numerics and resolution on fluid and particle statistics. Physical Review Fluids 5, 104302\.
* Ireland & Desjardins (2017) Ireland, P. J. & Desjardins, O. 2017 Improving particle drag predictions in Euler–Lagrange simulations with two-way coupling. Journal of Computational Physics 338, 405–430.
* Jie et al. (2022) Jie, Y., Cui, Z., Xu, C. & Zhao, L. 2022 On the existence and formation of multi-scale particle streaks in turbulent channel flows. Journal of Fluid Mechanics 935, A18.
* Johnson et al. (2020) Johnson, P. L., Bassenne, M. & Moin, P. 2020 Turbophoresis of small inertial particles: theoretical considerations and application to wall-modelled large-eddy simulations. Journal of Fluid Mechanics 883, A27.
* Kaftori et al. (1995) Kaftori, D., Hetsroni, G. & Banerjee, S. 1995 Particle behavior in the turbulent boundary layer. I. Motion, deposition, and entrainment. Physics of Fluids 7 (5), 1095–1106.
* Kiger & Pan (2002) Kiger, K. T. & Pan, C. 2002 Suspension and turbulence modification effects of solid particulates on a horizontal turbulent channel flow. Journal of Turbulence 3, N19.
* Kuerten (2006) Kuerten, J. G. M. 2006 Subgrid modeling in particle-laden channel flow. Physics of Fluids 18 (2), 025108.
* Kuerten (2016) Kuerten, J. G. M. 2016 Point-particle DNS and LES of particle-laden turbulent flow - a state-of-the-art review. Flow, Turbulence and Combustion 97, 689–713.
* Lee & Lee (2015) Lee, J. & Lee, C. 2015 Modification of particle-laden near-wall turbulence: Effect of Stokes number. Physics of Fluids 27, 023303.
* Lee & Lee (2019) Lee, J. & Lee, C. 2019 The effect of wall-normal gravity on particle-laden near-wall turbulence. Journal of Fluid Mechanics 873, 475–507.
* Magnaudet et al. (2003) Magnaudet, J., Takagi, S. & Legendre, D. 2003 Drag, deformation and lateral migration of a buoyant drop moving near a wall. Journal of Fluid Mechanics 476, 115–157.
* Marchioli et al. (2003) Marchioli, C., Giusti, A., Salvetti, M. V. & Soldati, A. 2003 Direct numerical simulation of particle wall transfer and deposition in upward turbulent pipe flow. International Journal of Multiphase Flow 29 (6), 1017–1038.
* Marchioli et al. (2007) Marchioli, C., Picciotto, M. & Soldati, A. 2007 Influence of gravity and lift on particle velocity statistics and transfer rates in turbulent vertical channel flow. International Journal of Multiphase Flow 33, 227–251.
* Marchioli & Soldati (2002) Marchioli, C. & Soldati, A. 2002 Mechanisms for particle transfer and segregation in a turbulent boundary layer. Journal of Fluid Mechanics 468, 283–315.
* Marchioli et al. (2008) Marchioli, C., Soldati, A., Kuerten, J. G. M., Arcen, B., Taniere, A., Goldensoph, G., Squires, K. D., Cargnelutti, M. F. & Portela, L. M. 2008 Statistics of particle dispersion in direct numerical simulations of wall-bounded turbulence: Results of an international collaborative benchmark test. International Journal of Multiphase Flow 34 (9), 879–893.
* Maxey (2017) Maxey, M. 2017 Simulation methods for particulate flows and concentrated suspensions. Annual Review of Fluid Mechanics 49, 171–193.
* Maxey & Riley (1983) Maxey, M. R. & Riley, J. J. 1983 Equation of motion for a small rigid sphere in a nonuniform flow. Physics of Fluids 26 (4), 883–889.
* McLaughlin (1989) McLaughlin, J. B. 1989 Aerosol particle deposition in numerically simulated channel flow. Physics of Fluids A: Fluid Dynamics 1 (7), 1211–1224.
* McLaughlin (1991) McLaughlin, J. B. 1991 Inertial migration of a small sphere in linear shear flows. Journal of Fluid Mechanics 224, 261–274.
* McLaughlin (1993) McLaughlin, J. B. 1993 The lift on a small sphere in wall-bounded linear shear flows. Journal of Fluid Mechanics 246, 249–265.
* Mehrabadi et al. (2018) Mehrabadi, M., Horwitz, J. A. K., Subramaniam, S. & Mani, A. 2018 A direct comparison of particle-resolved and point-particle methods in decaying turbulence. Journal of Fluid Mechanics 850, 336–369.
* Mei (1992) Mei, R. 1992 An approximate expression for the shear lift force on a spherical particle at finite Reynolds number. International Journal of Multiphase Flow 18 (1), 145–147.
* Mortensen et al. (2008) Mortensen, P. H., Andersson, H. I., Gillissen, J. J. J. & Boersma, B. J. 2008 Dynamics of prolate ellipsoidal particles in a turbulent channel flow. Physics of Fluids 20, 093302\.
* Motoori et al. (2022) Motoori, Y., Wong, C. & Goto, S. 2022 Role of the hierarchy of coherent structures in the transport of heavy small particles in turbulent channel flow. Journal of Fluid Mechanics 942, A3.
* Picano et al. (2009) Picano, F., Sardina, G. & Casciola, C. M. 2009 Spatial development of particle-laden turbulent pipe flow. Physics of Fluids 21, 093305.
* Reeks (1983) Reeks, M. W. 1983 The transport of discrete particles in inhomogeneous turbulence. Journal of Aerosol Science 14 (6), 729–739.
* Richter (2015) Richter, D. H. 2015 Turbulence modification by inertial particles and its influence on the spectral energy budget in planar Couette flow. Physics of Fluids 27, 063304.
* Richter & Sullivan (2014) Richter, D. H. & Sullivan, P. P. 2014 Modification of near-wall coherent structures by inertial particles. Physics of Fluids 26, 103304.
* Rubinow & Keller (1961) Rubinow, S. I. & Keller, J. B. 1961 The transverse force on a spinning sphere moving in a viscous fluid. Journal of Fluid Mechanics 11 (3), 447–459.
* Saffman (1965) Saffman, P. G. 1965 The lift on a small sphere in a slow shear flow. Journal of Fluid Mechanics 22 (2), 385–400.
* Sardina et al. (2012) Sardina, G., Schlatter, P., Brandt, L., Picano, F. & Casciola, C. M. 2012 Wall accumulation and spatial localization in particle-laden wall flows. Journal of Fluid Mechanics 699, 50–78.
* Schneiders et al. (2017a) Schneiders, L., Meinke, M. & Schröder, W. 2017a Direct particle–fluid simulation of Kolmogorov-length-scale size particles in decaying isotropic turbulence. Journal of Fluid Mechanics 819, 188–227.
* Schneiders et al. (2017b) Schneiders, L., Meinke, M. & Schröder, W. 2017b On the accuracy of Lagrangian point-mass models for heavy non-spherical particles in isotropic turbulence. Fuel 201, 2–14.
* Shi & Rzehak (2019) Shi, P. & Rzehak, R. 2019 Lift forces on solid spherical particles in unbounded flows. Chemical Engineering Science 208, 115145.
* Shi & Rzehak (2020) Shi, P. & Rzehak, R. 2020 Lift forces on solid spherical particles in wall-bounded flows. Chemical Engineering Science 211, 115264.
* Shi et al. (2021) Shi, P., Rzehak, R., Lucas, D. & Magnaudet, J. 2021 Drag and lift forces on a rigid sphere immersed in a wall-bounded linear shear flow. Physical Review Fluids 6, 104309\.
* Shin et al. (2022) Shin, H. H., Portela, L. M., Schaerer, C. E. & Mangiavacchi, N. 2022 Dynamics of suspended sediment transport: A Direct Numerical Simulation study. International Journal of Multiphase Flow 155, 104165.
* Sikovsky (2014) Sikovsky, D. P. 2014 Singularity of inertial particle concentration in the viscous sublayer of wall-bounded turbulent flows. Flow, Turbulence and Combustion 92, 41–64.
* Soldati & Marchioli (2009) Soldati, A. & Marchioli, C. 2009 Physics and modelling of turbulent particle deposition and entrainment: Review of a systematic study. International Journal of Multiphase Flow 35, 827–839.
* Takemura et al. (2009) Takemura, F., Magnaudet, J. & Dimitrakopoulos, P. 2009 Migration and deformation of bubbles rising in a wall-bounded shear flow at finite Reynolds number. Journal of Fluid Mechanics 634, 463–486.
* Toschi & Bodenschatz (2009) Toschi, F. & Bodenschatz, E. 2009 Lagrangian properties of particles in turbulence. Annual Review of Fluid Mechanics 41, 375–404.
* Uijttewaal & Oliemans (1996) Uijttewaal, W. S. J. & Oliemans, R. V. A. 1996 Particle dispersion and deposition in direct numerical and large eddy simulations of vertical pipe flows. Physics of Fluids 8 (10), 2590–2604.
* Vasseur & Cox (1977) Vasseur, P. & Cox, R. G. 1977 The lateral migration of spherical particles sedimenting in a stagnant bounded fluid. Journal of Fluid Mechanics 80 (3), 561–591.
* Vreman et al. (2009) Vreman, B., Geurts, B. J., Deen, N. G., Kuipers, J. A. M. & Kuerten, J. G. M. 2009 Two-and four-way coupled Euler–Lagrangian large-eddy simulation of turbulent particle-laden channel flow. Flow, Turbulence and Combustion 82 (1), 47–71.
* Wang et al. (2019) Wang, G., Fong, K. O., Coletti, F., Capecelatro, J. & Richter, D. H. 2019 Inertial particle velocity and distribution in vertical turbulent channel flow: A numerical and experimental comparison. International Journal of Multiphase Flow 120, 103105.
* Wang & Richter (2019) Wang, G. & Richter, D. H. 2019 Two mechanisms of modulation of very-large-scale motions by inertial particles in open channel flow. Journal of Fluid Mechanics 868, 538–559.
* Wang & Richter (2020) Wang, G. & Richter, D. H. 2020 Multiscale interaction of inertial particles with turbulent motions in open channel flow. Physical Review Fluids 5, 044307.
* Wang et al. (1997) Wang, Q., Squires, K. D., Chen, M. & McLaughlin, J. B. 1997 On the role of the lift force in turbulence simulations of particle deposition. International Journal of Multiphase Flow 23 (4), 749–763.
* Wray (1990) Wray, A. A. 1990 Minimal-storage time advancement schemes for spectral methods. In Tech. Rep.. NASA Ames Research Center.
* Yu et al. (2016) Yu, W., Vinkovic, I. & Buffat, M. 2016 Finite-size particles in turbulent channel flow: quadrant analysis and acceleration statistics. Journal of Turbulence 17 (11), 1048–1071.
* Yu et al. (2021) Yu, Z., Xia, Y., Guo, Y. & Lin, J. 2021 Modulation of turbulence intensity by heavy finite-size particles in upward channel flow. Journal of Fluid Mechanics 913, A3.
* Zeng et al. (2005) Zeng, L., Balachandar, S. & Fischer, P. 2005 Wall-induced forces on a rigid sphere at finite Reynolds number. Journal of Fluid Mechanics 536, 1–25.
* Zhang et al. (2023) Zhang, H., Cui, Y. & Zheng, X. 2023 How electrostatic forces affect particle behaviour in turbulent channel flows. Journal of Fluid Mechanics 967, A8.
* Zhang et al. (2022) Zhang, X., Nathan, G. J., Tian, Z. F. & Chin, R. C. 2022 The dominant mechanisms for each regime of secondary flows in horizontal particle-laden pipe flows. Journal of Fluid Mechanics 949, A10.
* Zhao et al. (2010) Zhao, L., Andersson, H. I. & Gillissen, J. J. J. 2010 Turbulence modulation and drag reduction by spherical particles. Physics of Fluids 22, 081702.
* Zhao et al. (2012) Zhao, L., Marchioli, C. & Andersson, H. I. 2012 Stokes number effects on particle slip velocity in wall-bounded turbulence and implications for dispersion models. Physics of Fluids 24, 021705.
* Zheng et al. (2021) Zheng, X., Feng, S. & Wang, P. 2021 Modulation of turbulence by saltating particles on erodible bed surface. Journal of Fluid Mechanics 918, A16.
|
AP and the percentage overlap between the training data of two languages. For
example, monolingual models trained on Zul and Nbl with $14.14\%$ and
$13.24\%$ overlap gives an AP of $62.9\%$ and $61.5\%$. respectively. On the
other hand, a model trained on Ssw and evaluated on Xho with only $6.84\%$
overlap gives an AP of $45.2\%$. Although all three languages belong to the
same language family as Xho, the two with a higher percentage overlap
outperform the one with less overlap. Moreover, we suspect that the model
trained on Sot outperforms the model trained on Afr (from the same language
family as Eng) when evaluating on Eng because of the high number of Eng words
present in the training set of Sot, showing a $9.69\%$ overlap.
This begs the question of whether the increased performance of training on a
language related to the target language (Figure 5.5) is related to the
phonetic properties shared among them, or because a model is exposed to
training examples from the target language. We investigate this uncertainty by
reproducing the experiment yielding the results in Figure 5.5, this time
removing the overlapping word segments from the training and validation sets.
The results without overlapping data are displayed in Figure 5.6. A
monolingual model is trained on data on each language where all words that
overlap with any other language is removed, then applied to every other
language where words from the training language are removed from the
evaluation set.
First, we observe a performance decrease when training and evaluating on the
same language (shown along the diagonal), especially for the Bantu languages.
Secondly, for most languages, the AP considerably decreased when training a
model on one language and applying it to another from the same language
family. Importantly, although the performance decreased, the best results are
still achieved by training on related languages, except for the one model
where training on Ven outperforms a model trained on Nbl by $0.6\%$ when
evaluated on Xho. Finally, for most languages, the AP decreased on Eng after
the Eng words were removed from the training data of all the other languages.
For example, the AP of Sot dropped from $57.2\%$ to $26.7\%$, with Afr (same
language family as Eng) now achieving the best AP ($11.1\%$ higher than second
best). This experiment indicates the potential of using related languages in
multilingual AWE modelling which we explore in the following section.
#### 5.5.2 Multilingual evaluation
The cross-lingual experiment above was in large part our inspiration for the
subsequent analysis of multilingual models. Concretely, we hypothesise that
even better performance can be achieved by training on multiple languages from
the same family as the target zero-resource language and that this would be
superior to multilingual models trained on unrelated languages. Focusing on
two evaluation languages, Zul and Sot from distinct language families, we
investigate this hypothesis by training multilingual models with different
language combinations, as shown in Table 5.2. Here we use the original data
without disregarding any particular words since it better simulates a real-
world scenario. Firstly, we see the best result on a language when all the
training languages come from the same family as the target. Secondly, we see
how the performance gradually decrease as the number of training languages
related to the target drop. Furthermore, notice the performance boost from
including even just one training language related to the target compared to
not including any. For instance, on Sot we see an increase of more than 12%
absolute when adding just one related language (from 52.5% and 51.9% to
64.8%).
To further demonstrate the benefit of using training languages from the same
family, we train a multilingual model for each evaluation language on all its
related languages using a 10% subset of the original data. For both Zul and
Sot, the subset models outperform the models where no related languages are
used. E.g. on Zul, the Xho+Ssw+Nbl subset model outperforms the full
Tsn+Nso+Eng model (no related languages) by more than 20% in AP. Moreover,
this subset model (58.6% in AP) even outperforms the Xho+Nso+Eng model (55.7%)
where all the training data from one related language are included and almost
matches the AP when using two related languages (Xho+Ssw+Eng, 60.9%).
These comparisons do more than just show the benefit of training on related
languages: they also show that it is beneficial to train on a diverse set of
related languages.
Table 5.2: AP (%) on test data for multilingual models trained on different combinations of well-resourced languages. Models are applied to two zero-resource languages from different language families, Nguni and Sotho-Tswana. For each training language, 100k word pairs were extracted. Multilingual model | Zul∗ | Sot†
---|---|---
Nguni: | |
Xho∗ \+ Ssw∗ \+ Nbl∗ | 68.6 | —
Xho∗ \+ Ssw∗ \+ Eng‡ | 60.9 | —
Xho∗ \+ Nso† \+ Eng‡ | 55.7 | —
Tsn∗ \+ Nso† \+ Eng‡ | 37.5 | —
Xho∗ \+ Ssw∗ \+ Nbl∗ ($10$% subset) | 58.6 | —
Sotho-Tswana: | |
Nso† \+ Tsn† | — | 76.7
Nso† \+ Eng‡ | — | 64.8
Xho∗ \+ Ssw∗ | — | 51.9
Xho∗ \+ Eng‡ | — | 52.5
Nso† \+ Tsn† ($10$% subset) | — | 58.4
#### 5.5.3 Adding more languages
In the above experiments, we controlled for the amount of data per language
and saw that training on languages from the same family improves multilingual
transfer. But this raises a question: will adding additional unrelated
languages harm performance? This question is of practical relevance in cases
where related data is either unavailable or the system developer is unsure for
which language AWEs will be extracted. To answer this, we systematically train
two sequences of multilingual models on all six well-resourced languages,
evaluating each target language as a new training language is added.
Same-different results for all five evaluation languages are shown in Figure
5.7 and QbE results for Zul, Sot and Afr are shown in Figure 5.8. Notice how
the trends in the same-different and QbE results track each other closely for
corresponding evaluation languages. In the first sequence of multilingual
models (green), we start by adding the three Nguni languages (Xho, Ssw, Nbl),
followed by the two Sotho-Tswana languages (Nso, Tsn), and lastly the Germanic
training language (Eng). The second sequence (orange) does not follow a
systematic procedure.
Figure 5.7: Same-different results from two sequences of multilingual models,
trained by adding one language at a time. For each training language, 100k
positive word pairs are used, which is indicated on the $x$-axis.
On Zul, the green sequence, which starts with a related language (Xho),
initially achieves a higher score compared to the orange sequence in both
Figures 5.7 and 5.8. Then, the score gradually increase by adding more related
languages (Ssw, Nbl). Thereafter, adding additional unrelated languages (Nso,
Tsn, Eng) show no performance increase. In fact, AP decreases slightly after
adding the two Sotho-Tswana languages (Nso, Tsn), but not significantly. The
orange sequence starts low on Zul until the first related language (Ssw) is
added, causing a sudden increase. Adding the Germanic language (Eng) has
little effect. Adding the last two related languages (Nbl, Xho) again causes
the score to increase. A similar trend follows for Sot and Afr, where adding
related languages causes a noticeable performance increase, especially when
adding the first related language; after this, performance seems to plateau
when adding more unrelated languages. (Afr is the one exception, with a drop
when adding the last language in the orange sequence). On Tso, which does not
have any languages from the same family in the training set, AP gradually
increases in both sequences without any sudden jumps. Although Ven is not in
the same family as Nso and Tsn at the lowest level of the tree in Figure 5.2,
it belongs to the same family (Sotho-Tswana) at a higher level. This explains
why it closely tracks the Sot results.
Figure 5.8: QbE results on Zul, Sot and Afr, using the same sequences of
multilingual models as in Figure 5.7. DTW result are shown for baseline
performance.
Summarising these results, we see that adding unrelated languages generally
does not decrease scores, but also does not provide a big benefit (except if
it is one of the earlier languages in the training sequence, where data is
still limited). In contrast, it seems that training on languages from the same
family is again beneficial; this is especially the case for the first related
language, irrespective of where it is added in the sequence.
### 5.6 Chapter summary
We investigated the effect of training language choice when applying a
multilingual AWE model to a zero-resource language. Using word discrimination
and QbE search tasks on languages spoken in South Africa, we showed that
training a multilingual model on languages related to the target is
beneficial. We observed gains in absolute scores, but also in data efficiency:
you can achieve similar performance with much less data when training on
multiple languages from the same family as the target. We showed that even
including just one related language already gives a large gain. From a
practical perspective, these results indicate that one should prioritise
collecting data from related languages (even in modest quantities) rather than
collecting more extensive datasets from diverse unrelated families, when
building multilingual acoustic word embedding models for a zero-resource
language. In the following chapter, we present a framework for producing
semantically organised AWEs using multilingual models developed in this
chapter.
## Chapter 6 Hate Speech Detection Using Multilingual Acoustic Word Embeddings
Leading up to this chapter, we mainly aimed at improving the quality of AWEs
for low-resource languages: we developed a new AWE model (Chapter 3), explored
multilingual adaptation techniques (Chapter 4), and considered language choice
in multilingual AWE modelling (Chapter 5). In Chapter 5 we briefly measured
AWE performance in a downstream query-by-example (QbE) speech search task.
However, this was performed with limited development data in a controlled
environment. In this chapter, we evaluate the practicality of using
multilingual AWEs for real-world applications, beyond the experimental
environments considered up to now.
The content of this chapter originates from a collaborative project with
VoxCroft, a company specialising in developing speech applications in low-
resource settings. We address a specific case study involving the development
of keyword spotting (KWS) systems to monitor radio broadcast audio for low-
resource languages. Here, we confront the challenge of lacking transcribed
data in the target language and dealing with radio recordings that differs
significantly from experimental audio data. In this challenging environment,
we develop an ASR-free approach, utilising a pre-trained multilingual AWE
model in a KWS system. Our main objective is to compare the performance of
this AWE-based KWS system to the conventional ASR-based KWS systems, which
mostly rely on transcribed data from the target language. We are particularly
concerned about the trade-off between accuracy and feasibility for rapid
deployment. To measure performance in this in-the-wild environment, we perform
our final experiments on real Swahili radio broadcast audio, scraped from
local radio stations in Kenya. Moreover, our KWS systems are specifically
developed to detect instances of hate speech in radio broadcasts. Therefore,
to search the audio, we use a set of keywords carefully selected by domain
experts that are potential markers of hate speech.
The content of this chapter was presented at Interspeech:
C. Jacobs, N. C. Rakotonirina, E. A. Chimoto, B. A. Bassett, and H. Kamper,
“Towards hate speech detection in low-resource languages: Comparing ASR to
acoustic word embeddings on Wolof and Swahili,” in Proceedings of Interspeech,
2023.
Contributions per author:
C. Jacobs developed the AWE-based KWS systems (Section 6.4.2).
N. C. Rakotonirina performed the ASR experiments (Section 6.4.1).
E. A. Chimoto evaluated radio recordings flagged by our KWS systems (Section
6.5.3.2).
B. A. Bassett and H. Kamper supervised the project and helped with editing the
paper.
### 6.1 Background
The rise of the internet and communication platforms are increasing the spread
of content, some of which include the use of derogatory or discriminatory
language targeting individuals or groups based on their identity or
characteristics. This form of expression is commonly referred to as hate
speech. Given the overwhelming volume of content being generated on online
platforms, several automatic monitoring systems have been explored to curb the
spread of hate speech (usually in the form of written text), especially for
resource-rich languages.
However, in many developing countries where there is a lack of technological
advancement, radio remains the primary medium for communicating news and
information to the public [111]. These radio stations play a vital role as
platforms for public discourse, facilitating freedom of expression [112].
Therefore, it is essential to establish appropriate mechanisms for monitoring
these radio stations to ensure responsible broadcasting practices.
A striking example of the consequences of hate speech over radio is the 1994
genocide in Rwanda where propaganda was disseminated through local radio
stations targeting individuals from an ethnic minority. The dehumanising
language used on air actively fuelled violence, resulting in devastating
consequences. Similarly, in the 2007-2008 post-election violence in Kenya,
vernacular radio stations played a significant role in exacerbating the
conflict [113].
In response to this ongoing issue, the United Nations has taken proactive
measures to address hate speech on radio [59]. By actively monitoring hate
speech claims, authorities can promptly respond and mitigate instances,
preventing tension escalation and the dissemination of harmful ideologies. The
efforts of this chapter are in line with the objective of combating hate
speech on radio broadcasts, specifically for low-resource languages.
Hate speech detection can be supported by using a keyword spotting (KWS)
system that can effectively search an audio corpus for a predetermined set of
keywords or phrases that are indicative of hate speech. Such a KWS system has
the potential ability to quickly scan through large amounts of audio content
and flag parts of the audio that have been identified as potential markers of
hate speech. It is important to note that a KWS system solely relying on
keyword or phrase identification may not be sufficient for comprehensive hate
speech detection. For instance, given the keyword “blood”, a KWS system may
retrieve utterances “we must spill their blood and eradicate them from our
society” and “she donated blood to help save lives during the blood drive.”
The former contains hate speech whereas the latter highlights a positive
action. Therefore, the context in which these potential hate speech markers
occur should be carefully considered and analysed by domain experts.
### 6.2 Related work
The conventional KWS approach involves transcribing the target audio with an
automatic speech recognition (ASR) system and then searching for keywords in
the output text [60, 61]. For low-resource languages, the absence of large
quantities of transcribed speech in the target language can make it difficult
to train high-quality ASR models. Several studies have looked into this
problem, proposing methods to train ASR models on limited amounts of labelled
data [114, 115, 55, 57, 116].
Menon et al. [56] were the first to evaluate and report results of a KWS
system monitoring radio broadcasts in low-resource languages spoken in Uganda.
Their system was tailored to identify specific keywords, contributing to the
progress of development and relief initiatives in rural areas. Their system
utilises an ASR model trained on around nine hours of transcribed data from
the target language. However, relying on this amount of data poses a challenge
as it will require a significant amount of time to collect and transcribe the
audio.
Saeb et al. [55] extended the work of Menon et al. [56] and implemented an
ASR-based KWS system that combines multilingual acoustic modelling and semi-
supervised training relying on much less transcribed data from the target
language. Here, they only use twelve minutes of annotated data from the target
language to train an ASR model. However, they use several hours of
untranscribed audio from the target language, coupled with a self-supervised
learning strategy to improve their system.
Rapidly deploying ASR KWS systems for low-resource languages remains
challenging due to the time-consuming process of collecting even a few minutes
of transcribed data, which can take several weeks. Moreover, multiple hours of
untranscribed speech are not necessarily available for all target languages
for self-supervised training.
When we want to rapidly develop and deploy a KWS system for hate speech
detection in a new setting, it might be better to opt for an ASR-free
approach. In a truly zero-resource setting, ASR-free KWS systems are
favourable, where KWS systems are developed without the need for transcribed
data. The most popular methodology extends QbE (Section 2.2.3), which uses a
spoken instance of a keyword to search through an audio corpus, without the
need for transcribing the audio into text. To use QbE for KWS, we need a small
number of spoken templates to serve as queries for the keywords of interest;
this is the approach followed in [57, 62] and the methodology we also use for
our AWE-based KWS system in this chapter.
DTW (Section 2.2.2) is typically used to match the speech features of a spoken
query to a search utterance [9, 48, 49, 50]. However, DTW has drawbacks in
terms of its high time complexity and limitations, such as poor speaker
independence, often requiring a large template set.
AWEs have been proposed as an alternative approach to finding matches in QbE
where search segments and query segments are jointly mapped to the same fixed
vector space [87, 6, 52, 53]. This allows for faster comparisons between query
and search segments in linear time complexity. Various neural networks have
been considered to obtain these AWEs as described in Section 2.3.
As demonstrated in Chapter 3 and Chapter 5, one approach to quickly obtain
robust AWE for a new language is to train a multilingual AWE model on labelled
data from multiple well-resourced languages, and then apply it to the target
low-resource language. Although these multilingual AWE models have proven
successful in controlled experiments, there has been limited work
investigating the effectiveness of these systems beyond the experimental
environment.
In this chapter, for the first time, we develop and evaluate a KWS system
utilising multilingual AWEs in a real-world scenario. For this experiment, we
apply the system to real radio broadcast audio, specifically targeting the
detection of instances of hate speech. The radio data was collected from
stations in Kenya, a Sub-Saharan country in Africa, where the predominant
language is Swahili. Additionally, we conducted developmental experiments
where we included Wolof, another low-resource language spoken mainly in
Senegal, Sub-Saharan Africa.
We compare the AWE-based KWS system to ASR-based KWS systems. We train our ASR
models, leveraging advancements in self-supervised speech representation
learning (SSRL) (Section 2.2.1). These models enable fine-tuning of ASR models
using limited transcribed data from the target language. Our main research
question is to compare the AWE-based KWS system with ASR-based KWS systems,
focusing on the trade-off between accuracy and rapid deployment.
Before we describe the different KWS systems, we first briefly give the
overall methodology that we follow towards our end goal of hate speech
detection on Swahili radio broadcasts.
### 6.3 Overview of methodology
We start by developing Wolof and Swahili KWS systems in a controlled
environment where we use parallel audio-text data from experimental datasets
to train and test our models. These datasets [117, 118] contain high-quality,
noise-free read speech. We then test our models on held-out test data from
these corpora. We refer to these experiments as in-domain because the test
data comes from the same domain as the training data. In this controlled
setup, we perform KWS using a set of keywords drawn from development data.
These are not necessarily hate speech keywords. The benefit of this controlled
setting is that we have transcriptions for the test data, and can therefore
quantify performance exactly.
We then apply the models to the out-of-domain Swahili radio broadcasts. These
recordings were obtained from three radio stations in Kenya. They consist of a
mix of read and spontaneous speech, and also include non-speech audio like
music and advertisements. Additionally, there may be instances of different
accents and languages present within the recordings. This is the type of real-
world audio that we would like to monitor for hate speech using KWS. As our
keywords in this context, we utilise actual hate speech keywords. These were
labelled as inflammatory words by native Swahili experts familiar with the
media environment. We should emphasise that, when we detect occurrences of
these words, they might not necessarily correspond to hate speech. For
example, the English sentence “do not kill your neighbour” is not hate speech
despite containing the word “kill”. In a deployed system the utterances
flagged by our KWS approach would be passed on to human moderators for further
review and hate speech assessment. In this work we also use a human moderator,
but only to mark whether a detected keyword actually occurred in an utterance.
We do this for our in-the-wild test since we do not have transcriptions for
the radio broadcasts.
In the following section, we provide a detailed overview of the KWS systems
that we developed. This will be followed by the implementation details of the
KWS systems (Section 6.5).
### 6.4 Low-resource keyword spotting
The ASR-based KWS systems of this chapter were developed by a colleague.
Therefore, in Section 6.4.1, we provide a high-level overview of this
approach. The AWE-based KWS system we developed as an ASR-free alternative, is
discussed in detail in the follow-up section.
#### 6.4.1 Keyword spotting with ASR
Figure 6.1: An illustration of ASR-based keyword spotting.
In an ASR-based KWS system, the target corpus is first transcribed to text.
Occurrences of a keyword in the corpus can be efficiently identified by simply
matching the keyword tokens with the output transcriptions. An example of this
ASR-based KWS is illustrated in Figure 6.1. In this example, when using the
keyword “hair” to search the output transcriptions, the KWS system will
correctly retrieve the first utterance (a). In the third utterance (c), “air”
was incorrectly transcribed as “hair” and will therefore be an incorrect
retrieval (although the keyword matches the transcription). In the second
utterance (b), the word “hair” in the audio was incorrectly transcribed as
“air” and will not be retrieved by the KWS system. Also, note that in this
example the ASR model correctly transcribed all the other words and would
therefore have a low word error rate (WER), but WER does not always correspond
to KWS performance. In Section 6.5.3 we give a more detailed explanation of
how we evaluate KWS performance.
Typically, to train an ASR model we need audio with transcriptions from the
target language. In a low-resource setting, a lack of large quantities of
transcribed speech data from the target languages prohibits the rapid
development of ASR-based KWS systems.
For our ASR systems, we build on recent advancements in ASR models leveraging
SSRL (Section 2.2.1). These SSRL models are trained in a self-supervised
fashion using large amounts of unlabelled audio recordings to produce high-
quality latent frame-level speech representations. It has been shown that
these pre-trained SSRL models can be fine-tuned in an end-to-end fashion for
ASR by mapping the latent frame representations to target text tokens. Fine-
tuning SSRL models for ASR requires using much less training data than
training an ASR model from scratch.
In our experiments, we use the pre-trained Wav2Vec2.0 XLSR model [76, 119].
This SSRL model is a large-scale cross-lingual speech model pre-trained on
half a million hours of unlabelled speech data in 128 languages (including
Swahili but not Wolof). The speech representation learned during cross-lingual
pre-training improves performance for low-resource languages by leveraging
data from the high-resource languages on which it is pre-trained. Using the
pre-trained model as a starting point, the model is fine-tuned on labelled
data from the language of interest.
Previous work has shown that this fine-tuning approach leads to competitive
ASR performance even with limited amounts of labelled data [119]. In our
experiments (Table 6.2 specifically) we show that the sample efficiency of the
XLSR model also extends to KWS when training on just five minutes of labelled
data.
#### 6.4.2 Keyword spotting with multilingual acoustic word embeddings
Figure 6.2: Search utterances ($u_{1},u_{2},\ldots,u_{N}$) are segmented and
mapped to a fixed-dimensional space, along with a query segment $q$, to
produce embeddings $\bigl{\\{}\mathbf{z}_{n}^{(i)}\bigr{\\}}$ and
$\mathbf{z}_{q}$, respectively. The subscript $n$ represents a unique
utterance within the search collection, and superscript $i$ represents a
segment extracted from utterance $\mathbf{z}_{n}$. Subsequently, search
utterances are ranked based on the shortest distance between a search segment
$\mathbf{z}_{n}^{(i)}$ and the query vector $\mathbf{z}_{q}$.
For a severely under-resourced language, there might not even be limited
amounts of labelled data available for ASR fine-tuning. In this scenario,
collecting only a few spoken instances of a keyword of interest would allow
for KWS through QbE search (Section 2.2.3). QbE is the task of retrieving
utterances from a search corpus using a spoken instance of a keyword, instead
of a written keyword.
In the previous chapter, our results (on development data) showed improved
performance in QbE search by using multilingual AWEs instead of DTW. We use a
similar setup to the QbE experiments described in Section 5.3 for experiments
in this chapter. The workings of a QbE system utilising AWEs are displayed in
Figure 6.2. First, the search corpus is segmented into word-like segments.
Search utterances are segmented by splitting into overlapping segments of some
minimum to some maximum duration (see Figure 5.3). Concretely, utterances from
the search corpus ($u_{1},u_{2},\ldots,u_{N}$) are segmented to produce a set
of variable-length speech segments
$S=\bigl{\\{}s_{n}^{(i)}\bigr{\\}}_{n=1}^{N}$, where $i$ is the $i$th search
segment extracted from utterance $n$.
A multilingual AWE model (trained on labelled data from multiple well-
resourced languages but not the target language) is then applied to the
segmented search corpus $S$ and query segment $q$, mapping them to the same
fixed-dimensional vector space. In this experiment, we use the CAE-RNN network
(Section 2.1.3) and train a multilingual CAE-RNN AWE model. This model is
trained on word pairs using labelled data from multiple languages (but does
not include the target language). This multilingual AWE model can be
transferred and applied to any unseen language to embed word segments of
arbitrary length. Utilising such a universal, language-independent model would
enable fast deployment of practical QbE-based KWS systems.
Applying the AWE model to search segments $S$ and query segment $q$ produces
search embeddings $\big{\\{}\mathbf{z}_{n}^{(i)}\bigr{\\}}$ and query
embedding $\mathbf{z}_{q}$, respectively. The utterances in the search corpus
are ranked based on comparing the query-segment distances
$d(\mathbf{z}_{q},\mathbf{z}_{n}^{(i)})$. For each query, the lowest distance
for each utterance is used as the final utterance score. We can then predict
that all query-utterance pairs with a distance below some threshold are
positive matches (this requires a threshold to be tuned on development data).
With these predictions, we can calculate common KWS metrics and compare them
to those achieved by the ASR-based KWS systems.
### 6.5 Experimental setup
In this section, we present the implementation details of both the ASR-based
(Section 6.4.1) and AWE-based (Section 6.4.2) KWS systems. It is important to
note that the overview provided in Section 6.3 serves as a reminder for the
reader. By the end of this section, the difference between the developmental
experiments conducted in a controlled environment and our in-the-wild KWS
testing, particularly concerning data and evaluation methods, should be clear.
#### 6.5.1 Data
Here we emphasise the difference between the training and evaluation data we
use during development and the data we use for our in-the-wild testing.
##### 6.5.1.1 Development
For our controlled experiments, we train and evaluate our models on Swahili
and Wolof data, collected from the Common Voice [118] and ALFFA [117]
datasets, respectively. Common Voice is an open-source data collection
initiative by Mozilla Foundation where volunteers can contribute by recording
their voices while reading prompts collected from a variety of sources. These
recordings are then verified for quality and accuracy. For our experiments we
use the Swahili data from the Common Voice Corpus 8.0 dataset which gives
train, test, and developing sets totalling 150 hours. In order to match the
sampling rate used during the training of the XLSR model, the original
recordings, which are sampled at 48 kHz, are downsampled to 16 kHz.
The current data repository of Common Voice does not include Wolof. We
therefore use a separate dataset from the ALFFA project, which contains Wolof
recordings and their corresponding transcriptions. Wolof is primarily spoken
in Senegal, another country in Sub-Saharan Africa. However, the official
language in Senegal is French, which serves as the lingua franca in all
governmental institutions. Due to the prevalence of French as the official
language, there are limited written documents available for Wolof to construct
parallel training data. For the Wolof language, we have designated training,
testing, and development sets, consisting only of 16 hours, 2 hours, and 2
hours, respectively.
For both datasets, we use the default train, development, and test splits. In
our controlled experiments, we evaluate in-domain KWS performance on the
Swahili and Wolof test splits using available transcriptions. The test split
of both languages serves as the search collection and is not used during model
development. The Swahili search collection contains 8 941 utterances and has a
duration of approximately 14 hours, while the Wolof set contains 2 000
utterances and has a duration of approximately 2 hours.
##### 6.5.1.2 In-the-wild
We perform our in-the-wild test on recorded radio broadcasts (collected by
VoxCroft) from three different radio stations in Kenya: Ghetto Radio, Radio
Salaam, and Radio Taifa. We do not have any transcriptions, speaker
information or content information on these audio clips. The nature of these
radio recordings has distinct characteristics that are different from the
prompted read speech we use during development. These broadcasts can contain a
variety of elements, including music in the background, fast-paced speech by
radio hosts, advertisements, background music, and dialogues involving
multiple speakers.
Our systems are only applied to these out-of-domain recordings and are nowhere
used during the development of our KWS systems. Before applying our systems to
this data we first apply a preprocessing step where we segment the recordings
using an unsupervised diarisation system that also attempts to remove music.
Apart from this preprocessing step, we simply apply our systems to this out-
of-domain data without further changes or calibration. After this
preprocessing step, 19 716 utterances are obtained, which is our out-of-domain
search collection. The duration of all utterances is between three and 30
seconds. Given that we do not have transcriptions of these recordings, we
cannot evaluate system performance in the same way we do in the controlled
experiments (where we have audio with corresponding transcriptions). As we
explain below (Section 6.5.3.2), we use a domain expert to evaluate our
system’s performance on this set.
#### 6.5.2 Models
We now provide the implementation details of the ASR model and the
multilingual AWE model for ASR-based (Section 6.4.1) and AWE-based KWS
(Section 6.4.2), respectively.
##### 6.5.2.1 ASR model
To train our ASR models, we fine-tune the pre-trained XLSR model by adding a
projection head to the final transformer layer. Subsequently, the model is
trained using transcribed data from the target language by mapping the frame-
level representations to character tokens. The model is optimised using the
connectionist temporal classification (CTC) loss [120]. The CTC loss is
effective for aligning input sequences and producing output sequences that may
not be temporally aligned, as is the case with speech and text. After the
fine-tuning step, the model produces a sequence of characters given an
arbitrary audio recording, that is then decoded to give the predicted
transcription.
For the ASR-based KWS systems, we control the amount of training data and
train three ASR models for Wolof and Swahili. Specifically, we fine-tune the
XLSR model [119] for each language on 30 hours, one hour, and five minutes of
labelled training data. These models are used for our development experiments
before applying them to the radio broadcast recordings.
##### 6.5.2.2 Multilingual acoustic word embedding model
For the AWE-based KWS system (Section 6.4.2), we train a multilingual AWE
model on five different Common Voice languages: Abkhazian (ab), Czech (cs),
Basque (eu), Swedish (sv), Tamil (ta). Word boundaries are obtained using the
Montreal forced-aligner [121]. We pool the data from all languages and extract
300k positive word pairs which we use to train a multilingual CAE-RNN AWE
model similar to the CAR-RNN model of Chapter 3. We use another well-resourced
language, Bulgarian, for validation during training using the same-different
evaluation (Section 2.5). The encoder and decoder each consist of three
unidirectional RNNs with 400-dimensional hidden vectors and an embedding size
of 100 dimensions. As inputs to the encoder, we use the XLSR model to extract
speech features (Section 2.2.1) instead of MFCCs used for AWE modelling in
previous chapters. We use the output of the $12$th transformer layer,
producing input features with 1024 dimensions. In the work of [122], they show
that the lower layers of the XLSR transformer model capture more
discriminative phonetic properties than the final layers when simply using the
model to extract frame-level speech representations. Preliminary experiments
showed using these features outperforms MFCCs in AWE modelling. We provide a
short analysis in Section 6.7.1 comparing multilingual AWEs trained on MFCC
and XLSR. All models are optimised using Adam [103] with a learning rate of
$0.001$.
#### 6.5.3 Evaluation
We outline our KWS evaluation methods for the ASR-based and AWE-based KWS
systems for the two scenarios: controlled in-domain experiments with available
transcriptions and out-of-domain in-the-wild tests where transcriptions are
unavailable for evaluation.
##### 6.5.3.1 Controlled experiments
To measure in-domain KWS performance, we apply the ASR-based and AWE-based KWS
systems to the same search corpora. Each utterance in the search corpus is
assigned a label of 1 if the keyword is detected by the respective KWS system,
or 0 if the keyword is not present. We evaluate our KWS systems by using the
predicted labels and compare them against the true labels obtained from
available transcriptions. Following this approach, we can calculate precision
and recall scores.
Precision measures the accuracy of a KWS system in predicting whether an
utterance contains a keyword. It is calculated as the number of utterances
correctly predicted as keyword-containing utterances (# hits) divided by the
total number of utterances predicted to contain a keyword (# keywords
detected):
$P=\frac{\text{\\# hits}}{\text{\\# keywords detected}}$ (6.1)
Recall quantifies the KWS system’s ability to retrieve keyword-containing
utterances from the search corpus. It is calculated as the number of correctly
predicted keyword-containing utterances (# hits) divided by the total number
of keyword-containing utterances in the search corpus (# keywords):
$R=\frac{\text{\\# hits}}{\text{\\# keywords}}$ (6.2)
In addition to precision and recall, we report the $F_{1}$-score to quantify
the trade-off between these two metrics. The $F_{1}$-score represents the
balanced combination of precision and recall, calculated as the harmonic mean
of the two metrics:
$F_{1}=2\times\frac{P\times R}{P+R}$ (6.3)
To evaluate our ASR KWS systems, we simply check whether a keyword occurs in
the predicted transcript after applying it to the search corpus and label each
utterance accordingly (1 or 0) to calculate the metrics above.
The AWE-based KWS system does not output transcribed text but a query-
utterance score for each query segment (bottom of Section 6.4.2). First, sub-
segments for utterances in the search collection are obtained by extracting
windows ranging from 20 to 35 frames with a 5-frame overlap (we experimented
with different window and overlap sizes on development data where this
configuration worked best). Then, an AWE model is applied to all the sub-
segments and query segment. Finally, the similarity between a query and search
segment is calculated using cosine distance.
We use a set of ten templates of 36 unique keywords with a minimum character
length of five for Swahili, and 15 unique keywords with a minimum character
length of four, for Wolof. These keywords were randomly sampled from a larger
set of words that appear at least ten times in both the development and test
sets. These keywords do not represent hate speech keywords. Query templates
are drawn from development data, embedded, and averaged to obtain a single AWE
embedding representing a keyword. Preliminary experiments showed that
averaging all templates from the same keyword improves KWS performance
compared to using a single template for searching. The threshold value in the
AWE-based KWS is tuned for the highest $F_{1}$-score across all keywords on
the controlled test data (more in Section 6.7.2).
Table 6.1: Swahili hate speech keywords used for in-the-wild KWS on Swahili radio broadcasts, with their English translations. Swahili | English | Swahili | English
---|---|---|---
vita | war | wezi | thieves
damu | blood | majimbo | states
hama | move | wakora | conmen
kabila | tribe | panga | machete
utapeli | fraud | takataka | garbage
mende | cockroaches | mjinga | stupid
kitendawili | riddle | fala | stupid
##### 6.5.3.2 In-the-wild
For the in-the-wild KWS, we apply the systems directly to the out-of-domain
Swahili radio broadcasts. We use a set of keywords labelled as inflammatory by
expert analysts from VoxCroft who are familiar with the media environment for
the purpose of hate speech detection. These keywords with their English
translations are given in Table 6.1. For the AWE-based KWS system, ten query
templates per keyword are extracted from the in-domain Swahili data. We
evaluate KWS performance by asking each approach to give 100 utterances out of
the search collection that are most likely to contain any of the hate speech
keywords. In the absence of per-word confidence scores for the ASR models,
here we simply take 100 random utterances that were predicted to contain a
keyword (for most models, the total number of utterances was around 150).
For the AWE-based KWS, we use the 100 highest-ranked utterances. Because we do
not have transcriptions for this data, we provide a native Swahili speaker
with untranscribed recordings of the utterances that were predicted to contain
a keyword. They then mark whether the keyword was indeed present. Note that
due to the absence of transcriptions for the radio recordings, we are unable
to calculate recall or the $F_{1}$-score. Therefore, we can only measure
performance by calculating precision, based on 100 retrieved utterances per
system (Table 6.4).
### 6.6 Experiments
We first look at the results from our controlled test on in-domain Swahili and
Wolof data. We then look at the results on the out-of-domain data, which is of
particular interest to us as it would indicate the effectiveness of both these
systems for real-world application.
#### 6.6.1 KWS results in a controlled environment
Table 6.2: ASR and AWE KWS results (%) on in-domain test data. For the ASR systems, the XLSR model is fine-tuned on each target language, controlling the amount of training data. For the AWE KWS system, a supervised multilingual AWE is trained on multiple well-resourced languages and applied to the two low-resource target languages. Model | Swahili | Wolof
---|---|---
| Prec. | Rec. | $F_{1}$ | Prec. | Rec. | $F_{1}$
ASR: | | | | | |
XLS-R (30-h) | 97.6 | 98.6 | 98.0 | 93.1 | 87.1 | 90.0
XLS-R (1-h) | 96.4 | 76.1 | 85.1 | 93.1 | 74.3 | 82.7
XLS-R (5-min) | 95.4 | 42.6 | 58.9 | 88.0 | 45.6 | 60.0
Multilingual AWE: | | | | | |
CAE-RNN (ab+cs+eu+sv+ta) | 57.1 | 56.1 | 56.6 | 44.0 | 58.4 | 50.2
The first three lines in Table 6.2 report the ASR KWS results on in-domain
Swahili and Wolof data. The results show that fine-tuning XLSR on only five
minutes of labelled data achieves decent KWS performance, with $F_{1}$-scores
of 59% and 60% for Swahili and Wolof, respectively. The multilingual AWE KWS
system using roughly one minute of template data, performs worse on most
metrics. The only metric on which the AWE system is better is recall, where it
achieves higher scores than the 5-minute ASR system on both languages. In
practice, the AWE system could therefore be a better option if recall is
important.
We briefly consider how ASR performance is affected by the amount of training
data. For reference, WERs for the 30-hour, 1-hour, and 5-minute models are
respectively 9%, 36%, and 62% on Swahili. For Wolof, the WERs are 27%, 44%,
and 68%. As expected, more training data gives better WER and KWS scores. But
it is noteworthy that with just five minutes of training data, we can already
spot keywords with high precision (95% and 88% for Swahili and Wolof,
respectively). This is especially useful for our use case, where we want to
rapidly develop KWS applications in severely low-resourced settings. The other
noteworthy finding from the ASR results is that scores for Swahili are not
notability higher than Wolof, although the former is one of XLSR’s pretraining
languages.
Table 6.3: Supervised AWE KWS results (%) on in-domain Swahili test data. The supervised monolingual CAE-RNN AWE model is trained using labelled Swahili data. It is applied with two segmentation configurations: using ground-truth word boundaries (true segm.), and using a variable-length window which is swept across the search collection (random segm.). Here we also report the standard QbE metrics $P@10$ and $P@N$ (Section 5.4.3.2). Model | Prec. | Rec. | $F_{1}$ | $P@10$ | $P@N$
---|---|---|---|---|---
Supervised monolingual: | | | | |
CAE-RNN (true segm.) | 95.5 | 90.8 | 93.1 | 98.6 | 94.3
CAE-RNN (random segm.) | 79.2 | 76.1 | 77.6 | 92.2 | 90.9
Supervised multilingual: | | | | |
CAE-RNN ab+cs+eu+sv+ta (random segm.) | 57.1 | 56.1 | 56.6 | 87.8 | 64.7
We now turn to the AWE-based approach. Specifically, we ask what the upper
bound on performance would be if we had more training data or a more idealised
search setting. Therefore, for a moment, we assume we have labelled data
available and perform top-line experiments on the in-domain Swahili data,
shown in Table 6.3. A supervised Swahili AWE model is trained on 30 hours of
labelled data and applied to the search collection that is segmented using
true word boundaries. Results are shown in the first row of Table 6.3, serving
as the top-line performance for AWE KWS, with a high $F_{1}$-score of 93%.
Compared to the 30-hour ASR system (Table 6.2), the idealised AWE system comes
closer to the $F_{1}$ of 98%. The second row in Table 6.3 shows KWS
performance using a supervised Swahili AWE model but applied to search
segments extracted by sliding a variable-length window across the search
collection—the way we apply the AWE approach in practice and in Table 6.2.
Segmentation without true word boundaries incurs a significant penalty,
resulting in an $F_{1}$-score drop to 77%. It is clear that the sliding window
approach has a large effect on downstream performance, so future work should
consider more sophisticated unsupervised word segmentation techniques.
#### 6.6.2 KWS results in the wild
We now turn to our main research question: comparing ASR to multilingual AWE-
based KWS on real-life, out-of-domain audio in a low-resource setting. As
mentioned in Section 6.5.3.2, systems are applied to out-of-domain Swahili
radio broadcasts, after which, for each system, the top 100 utterances
predicted to contain a hate speech keyword (Table 6.1) are manually reviewed.
The results are given in Table 6.4.
The table reports precision: the proportion of retrieved top-100 utterances
that correctly contain a hate speech keyword. Surprisingly, we see that the
multilingual AWE KWS system achieves a precision of 45%—better than the
5-minute and even the 1-hour ASR system in this in-the-wild test. This is in
contrast to the in-domain KWS results, where the 1-hour ASR model outperformed
the AWE KWS system (Table 6.2). Further investigation is required to
understand exactly why the relative performance of the ASR and AWE systems are
affected differently when applied to out-of-domain data. However, it is worth
noting that several studies have shown that ASR system performance can drop
dramatically when it is applied to data outside of its training domain [123,
124, 125]. One example where this can be seen in this case is how many of the
ASR retrievals contain music (which we asked the human annotator to mark).
The in-the-wild search collection has been diarised automatically, which
included a step to remove music segments. But this preprocessing step is not
perfect: the search collection still ends up with some music (which neither
the ASR nor AWE systems have seen in training). Table 6.4 shows that, out of
the top-100 utterances for the 30-hour ASR system, 30% contained music. This
decreases for the 1-hour (29%) and 5-minute (17%) ASR systems. But the AWE KWS
system only retrieves two utterances containing music. And of these two, one
utterance actually did contain a hate speech keyword in the music lyrics
(there are also examples of such correct matches in the ASR music retrievals).
Nevertheless, this shows that the AWE approach seems to be more robust to
domain mismatch compared to training an ASR system on one hour or five minutes
of labelled data. We therefore conclude that implementing an AWE-based KWS
system as an ASR-free alternative for rapid deployment in an unseen language
proves to be fruitful.
Table 6.4: In-the-wild KWS results (%) on Swahili radio broadcasts. For each system, we report the percentage of utterances retrieved containing a keyword (precision) and the percentage of utterances retrieved containing music. Model | Precision | Music
---|---|---
ASR: | |
XLS-R (30-h) | 52 | 30
XLS-R (1-h) | 42 | 29
XLS-R (5-min) | 36 | 17
Multilingual AWE: | |
CAE-RNN ab+cs+eu+sv+ta | 45 | 2
### 6.7 Further analysis
We briefly examine the transition from MFCCs to XLSR features as input to our
multilingual AWE model. We also investigate the potential of selecting a more
optimal threshold value for the AWE-based KWS system (Section 6.4.2).
(a) MFCC
(b) XLSR
Figure 6.3: UMAP [126] visualisation of the ten most frequent two-character
word segments in Bulgarian development data. These plots show the difference
between AWEs produced by a multilingual AWE model trained on (a) MFCC features
and (b) XLSR features.
#### 6.7.1 XLSR vs MFCCs
For our experiments in this chapter we substituted MFCCs for XLSR speech
features as input to our AWE models. Here we briefly look at the difference in
embedding quality produced between a multilingual AWE model trained on XLSR
features compared to MFCCs. We use the same training setup in Section 6.5.2.2
and train a multilingual AWE using MFCCs. For evaluation, we use the same
developmental language, Bulgarian, and perform the same-different word task.
On the same set of evaluation pairs, we obtain an AP of 29% and 58% for the
MFCC and XLSR multilingual AWE models, respectively. With further
investigation, we found that a model trained on XLSR features is much better
at learning higher-quality embeddings for shorter word segments. Figure 6.3
shows the difference between the quality of embeddings produced for two
character words when using (a) MFCCs and (b) XLSR as input features. Overall,
we see word segments from the same class to be much better clustered together
and separated from word segments from different classes. Using XLSR features
as input clearly improves the quality of AWEs, leading to improved performance
in the AWE-based KWS approach.
#### 6.7.2 Threshold
As mentioned in Section 6.4.2, the threshold value selected for the AWE-based
KWS system was tuned on the $F_{1}$-score by varying the threshold value over
all keywords. For interest, we tune the threshold value per keyword and plot a
precision-recall curve for each of the 36 keywords, shown in Figure 6.4. We
also plot the precision-recall curve (in black) using the same threshold value
for all keywords. From the figure we see that most per-keyword scores are
higher than the average (blue lines above black line). When we average the
$F_{1}$-scores using the optimal threshold value per keyword, we obtain an
$F_{1}$ score of 69% compared to 57% (bottom of Table 6.2) when using the same
threshold for all keywords. This raises the question of the selected threshold
value and whether simply treating keywords as equal might be a suboptimal
approach. Practically, however, a per-keyword threshold can not be calculated
given that for a new arbitrary keyword in a new unseen language we do not have
development data to tune this.
But there is a possibility that a threshold can be selected based on other
properties of a new keyword. These factors include the length of the speech
segment, the number of phones present, and the relative distances between
other unknown word segments in the acoustic space. For example, we may find
that, in general, the threshold value for shorter word segments is lower than
for longer speech segments. Future work should investigate the correlation
between the threshold value of keywords and the potential factors mentioned
above.
Figure 6.4: Precision-recall curves from in-domain AWE-based KWS on in-domain
Swahili test data using the same 36 keywords as in Section 6.6.1.
### 6.8 Chapter summary
This chapter addressed the problem of keyword spotting (KWS) for the purpose
of hate speech detection in low-resource languages. We compared two KWS
systems for data from Swahili and Wolof: a fine-tuned automatic speech
recognition (ASR) model using different amounts of training data, and an ASR-
free KWS system that utilises multilingual AWEs. The results show that fine-
tuning a pre-trained multilingual ASR model using even a small amount of
labelled data can outperform an AWE-based KWS system in controlled
environments where training and test data come from the same domain. However,
the AWE-based KWS system is more robust on out-of-domain radio broadcast data
and achieves comparable results to an ASR model fine-tuned on 30 hours of
labelled data. In the end, the merits of ASR vs AWE KWS will come down to the
practical setting: it will depend on whether labelled training data can be
collected from the target domain or not, and whether precision or recall is
more important.
## Chapter 7 Leveraging Multilingual Transfer for Unsupervised Semantic
Acoustic Word Embeddings
In the previous chapters, we focussed on producing vector representations for
spoken word segments where different acoustic realisations from the same word
type should have similar embeddings. This can clearly be seen in the way we
evaluated the embeddings: we use a word discrimination task (Section 2.5), a
QbE speech search task (Section 5.3), and a keyword spotting task (Section
6.4.2), all of which require exact matching of spoken word segments. The
performance of both these tasks relies on the discriminative ability between
embeddings from different word types. Specifically, we aimed at producing
these embeddings in low-resource settings, where annotated data in the target
language is not available; here a multilingual transfer approach demonstrated
superior performance.
In this chapter, we shift our focus to a different type of representation
learning: semantic acoustic word embeddings. In a semantic-acoustic space,
embeddings should not only be similar for different realisations of the same
word type. Additionally, embeddings from distinct word types should also be
close to each other if the respective words share meaning. These embeddings,
mapping to spoken words, should reflect the same semantic properties as word
embeddings trained on text data (Figure 1.1(a)). In other words, proximity in
this new acoustic-semantic space should reflect word-type similarity as well
as semantic relatedness between embeddings from different word types.
Again, we consider a low-resource scenario where word labels in the target
language are unavailable. In this setting, there have been limited efforts to
address the challenging problem of producing semantic AWEs. As a contribution,
we propose a novel semantic AWE modelling approach: leveraging a pre-trained
multilingual AWE model to assist semantic AWE modelling. This marks the first
instance of using a multilingual transfer strategy in this context.
Furthermore, existing research on semantic AWEs lacks comprehensive
evaluation. We therefore perform our experiments in a constrained setup and
evaluate our embeddings in an intrinsic word similarity task (measuring
semantic relatedness between words), and an extrinsic semantic QbE search. In
the latter, a system should retrieve utterances containing different
realisations of the query as well as utterances containing word segments that
are semantically related to the query.
Parts of this chapter are presented in the journal paper:
C. Jacobs and H. Kamper, “Leveraging multilingual transfer for unsupervised
semantic acoustic word embeddings,” IEEE Signal Processing Letters, 2023.
### 7.1 Background
We remind the reader of the properties and differences between AWEs and
textual word embeddings described in Section 1.2.1. In the zero-resource
setting, unsupervised representation learning of whole word speech segments
primarily focuses on producing embeddings that allow discrimination between
word segments of different word types. In this context, we train our AWE
models using isolated word pairs. For example, the CAE-RNN (Section 3.2.1) is
trained to reconstruct a different instance of the same word type as the
input, and the ContrastiveRNN (Section 3.3) is explicitly trained to minimise
the distance between instances of the same word type and maximise the distance
between all instances of different word types. Consequently, different
instances of the same word type should have similar embeddings and will be
different from instances from different word types. These representations are
learned without considering the context in which they appear within a spoken
corpus. From this point onwards, we refer to these representations, which only
encapsulate the acoustic properties linked to word type, as phonetic AWEs. An
example vector space containing these phonetic AWEs is shown in Figure 7.1(a).
Here, different spoken instances of the same word type, for example “bank”,
end up close to each other. Different spoken instances of the word “tank” will
also end up close to each other. Since “bank” and “tank” are phonetically
similar, instances of both word types might also end up close to each other in
vector space. These are the type of embeddings generated in all our AWE
experiments from previous chapters.
In contrast, in the text domain, word embedding models map written words to
fixed vectors, capturing the contextual meaning of words within a text corpus.
These models learn from the co-occurrence information in large unlabelled text
corpora. For example, the Skip-gram model of Mikolov et al. [127] employs a
neural network model that, given a target word, aims to predict the
neighbouring words in that sentence (more in Section 7.2). As a result, words
that are related in meaning end up having similar embeddings. Moreover, these
textual word embedding (TWE) models assign a unique embedding to each word
type within a fixed vocabulary. Figure 7.1(b) illustrates an example semantic
space, where the written words “finance”, “bank”, and “currency” are mapped
nearby, and “tank” and “infantry”, reflecting their shared contextual usage.
Instead of written words, our goal in this chapter is now to map spoken word
segments to vector representations that preserve this contextual relationship
among words described by Mikolov et al. [127] (Figure 7.1(b)). A unique aspect
of our problem lies in the continuous nature of speech, where each spoken word
segment will be different from another. Consequently, each spoken segment will
have a unique vector representation, which is different to TWEs where a single
vector is assigned to each word type.
For instance, we would now like to see spoken instances of “finance”,
“currency”, and “bank” mapped close to each other and, “tank” and “infantry”,
to be in close proximity in the resulting acoustic-semantic space, as
illustrated in Figure 7.2.
(a) Acoustic word embeddings
(b) Textual word embeddings
Figure 7.1: A visualisation of the fundamental difference between (a) AWEs
and (b) TWEs. (a) Preserves word-type similarity among different acoustic
realisations of five different word types. (b) Preserves semantic relatedness
among the same word types mapping to written words. Figure 7.2: A
visualisation of a semantic acoustic word embedding space.
As part of our approach to create semantic AWEs, we will extend the textual
word embedding approach called skip-gram. This approach has also been used as
the basis for other existing semantic AWE approaches. We therefore present the
skip-gram approach in detail below, before going into further details on
previous work on semantic AWEs.
### 7.2 Skip-gram
The Skip-gram model of Mikolov et al. [127] is trained on large amounts of
text corpora. The objective of this model is to learn word representations
that are useful for predicting neighbouring words in a sentence. Formally,
given a sequence of training words, $w_{1},w_{2},w_{3},\ldots,w_{T}$, the
model aims to maximise the average log probability [28],
$\frac{1}{T}\sum_{t=1}^{T}\sum_{-c\leq j\leq c,j\neq
0}\text{log}\,p(w_{t+j}|w_{t}),$ (7.1)
where $c$ is the number of words before and after the target centre word
$w_{t}$. Consider an example of multiple pairs of target and context words as
training samples for a single sentence in a text corpus, shown in Table 7.1.
Table 7.1: Training samples extracted from a sentence using a context size of two. Example sentence: “two boys are kicking a ball to each other in the park”. Context size = 2 | Target word | Context words
---|---|---
[two boys are] | two | boys, are
[two boys are kicking] | boys | two, are, kicking
[two boys are kicking a] | are | two, boys, kicking, a
[boys are kicking a ball] | kicking | boys, are, a, ball
… | … | …
[each other in the park] | in | each, other, the, park
[other in the park] | the | other, in, park
[in the park] | park | in, the
For the training sample with input (the target word) “kicking” the model will
learn to predict the words “boys”, “are”, “a”, “ball”.
Figure 7.3 shows the Skip-gram model architecture. The model consists of two
fully connected feedforward layers with weight parameters
$\mathbf{W}_{1}\in\mathbb{R}^{(V\times D)}$ between input and hidden layer,
and $\mathbf{W}_{2}\in\mathbb{R}^{(D\times V)}$ between hidden layer and
output, where $V$ is the vocabulary size (number of unique words in text
corpus) and $D$ is the dimension of the hidden layer.
Figure 7.3: The Skip-gram model architecture.
For each training sample, the target and context words are encoded into one-
hot vectors where the non-zero index corresponds to a word’s index in the
vocabulary set. Given a one-hot vector as input, the model produces a latent
embedding of size $D$. The embedding is then given as input to the second
layer to produce an output vector of size $V$. A softmax function is applied
to the output vector to normalise the output to a probability distribution
over $V$ classes, where the value at each index shows the prediction of the
model given the input word. The model then updates weight parameters
$\mathbf{W}_{1}$ and $\mathbf{W}_{2}$ to maximise the objective function in
Equation 7.1. Note that the one-hot input vector selects the $i$-th row from
weight parameter $\mathbf{W}_{1}$ to produce the latent embedding. The one-hot
vector essentially acts as a lookup table for each word in the vocabulary.
Therefore, after training, $\mathbf{W}_{1}$ represents the learned word
embeddings where the $i$th row of size $D$ represents the $i$th word in the
vocabulary set.
Keep in mind that the meaning of the words itself is not important, but rather
the context in which they appear with other words, such that the words can be
replaced with any unique character sequence (i.e. label) as long as the
sequence in which they appear are retained. This allows a model to be trained
on a text corpus with unknown word labels. In one of our speech models, we
make use of this property by inferring unknown word labels for spoken word
segments.
The training strategy of the Skip-gram model, which involves predicting
surrounding words from a target centre word, served as an inspiration to the
speech community as a means to obtain semantic embeddings from unlabelled
spoken word segments. In the following section, we explore the existing body
of work addressing this problem.
### 7.3 Related work
Learning semantic AWEs from speech is extremely difficult due to channel
variability, noise, and speaker-specific information that are not present in
written text. Moreover, the presence of inherent phonetic interference poses a
significant obstacle to semantic AWE modelling. For example, distinguishing
between spoken instances of “steal” and “steel” is challenging, even with the
context provided, since from a model’s perspective they would appear
identical. Some studies, therefore, consider the use of another modality as a
grounding signal, for example, images [128, 129, 91] or text labels [130] as a
weak form of supervision.
Only a handful of studies have looked at learning semantic AWEs from
unlabelled speech alone. Chung and Glass [63] were the first to present and
gain widespread recognition for producing semantic AWEs without word labels.
They introduced a model named Speech2Vec, borrowing the training methodology
of the Skip-gram model, by training on pairs of context words. Speech2Vec uses
an RNN decoder-encoder structure to accommodate variable-length speech
segments as input and target output (similar to the CAE-RNN), while the Skip-
gram model uses a two-layer fully connected feedforward network with one-hot
encoded vectors as input and output.
Speech2Vec is evaluated as follows: After training, all word segments from the
training data are embedded using the encoder RNN. Each word type is then
represented by a single embedding by averaging all instances of that type. The
averaged embeddings are then used in multiple intrinsic word similarity tasks
measuring semantic relatedness. With this setup, Speech2Vec achieved
surprisingly good results, even outperforming textual word embeddings (trained
on the corresponding transcripts) in some tasks. This certainly raised our
suspicion, as well as others [131], concerning the validity of the embeddings
produced by this model. In Section 7.4.1 we describe the model in more detail
and address some of the concerns with Speech2Vec. Nevertheless, Speech2Vec has
not been evaluated for single sample performance nor word segments that were
not included in the model’s training data. The quality of embeddings produced
for unseen segments is essential to determine the usefulness of embeddings for
downstream applications, such as semantic QbE.
Chen et al. [64] proposed a two-step framework towards semantic AWE modelling.
Speech segments are first given to a speaker disentanglement network (trained
using the unlabelled word segments), which outputs two embeddings, separating
speaker and phonetic information. The idea is that the phonetic embedding
should only contain information related to word type (the same property of our
multilingual AWEs). To produce semantic embeddings, they construct context
training pairs using the phonetic embeddings and train a network optimising a
contrastive loss similar to the ContrastiveRNN (Equation 3.6). Their results
are not compared to Speech2Vec and also use averaged embeddings from the
training data for evaluation.
Speech2Vec [63] and the work of Chen et al. [64] only consider the unlabelled
speech from the target language during model development. In this chapter, we
propose strategies for semantic AWE modelling, utilising a pre-trained
multilingual AWE model. We evaluate our models on an intrinsic word similarity
task, measuring the quality of embeddings for single word segments not used
during training. To demonstrate the practical applicability of our embeddings,
we also apply them to a downstream semantic QbE task.
In both cases [63, 64], the problem is simplified by assuming that we know
where words start and end in the target language (but the word classes are
still unknown). However, in a true zero-resource setting, word boundaries will
be unknown. In the experiments presented in this chapter, we also assume word
boundaries are known. This problem needs to be addressed in the future to
consider complete unsupervised semantic AWE modelling (Section 8.2).
### 7.4 Semantic AWEs trained from scratch
Until now, our AWE models mapped speech segments from the same word type
nearby in vector space (Figure 7.1(a)). In our notation, a speech segment
$X=\left(\mathbf{x}_{1},\mathbf{x}_{2},\ldots,\mathbf{x}_{T}\right)$ is
projected to a fixed vector $\mathbf{z}$. From this point onwards we add the
subscript p to the AWEs from previous chapters. In other words,
$\mathbf{z}_{\text{p}}$ indicates that the embedding preserves phonetic
information related to word type only.
The goal of this chapter is to produce semantic embeddings
$\mathbf{z}_{\text{s}}$. These embeddings should not only reflect phonetic
similarity but also capture word meaning (Figure 7.2). Speech2Vec [63] and the
work of Chen et al. [64] achieve this by only training on context pairs
extracted from unlabelled speech data (from the target language). We use the
training strategies presented in these two methods as a baseline for our
multilingual enhanced semantic AWE models (Section 7.5).
#### 7.4.1 Speech2Vec
Speech2Vec is a variant of the CAE-RNN AWE model (Section 3.2.1). The CAE-RNN
is an encoder-decoder model that takes a speech segment as input and aims to
reconstruct a different instance of the same word type. Speech2Vec uses the
same encoder-decoder framework, but instead of training on pairs of instances
from the same word type (as for phonetic AWE modelling), it uses context word
pairs $(X_{\text{trg}},X_{\text{ctx}})$. $X_{\text{trg}}$ is a target centre
word segment while $X_{\text{ctx}}$ is a context word appearing somewhere in a
window around the centre. These context pairs are constructed similarly to
those used for training the Skip-gram model on text data (see Table 7.1).
However, in this case, $X_{\text{trg}}$ and $X_{\text{ctx}}$ represent spoken
words. Importantly, these context pairs are constructed without word labels by
only considering the relative position of words within an utterance (requiring
word boundaries). Figure 7.4 shows the Speech2Vec model for a single training
pair. Speech2Vec optimises the reconstruction loss of Equation 3.1, where
$X^{\prime}$ in that equation is replaced with $X_{\text{ctx}}$. Ideally the
resulting embeddings $\mathbf{z}_{\text{s}}$ should therefore be similar for
words that co-occur in the speech data. Although this approach seems
theoretically sound, concerns have been raised regarding the performance of
this model. We address this now.
Figure 7.4: The Speech2Vec model architecture. This is similar to the CAE-RNN
model architecture (Figure 3.1) but adapted for context word pairs
$(X_{\text{trg}},X_{\text{ctx}})$ instead of positive pairs $(X,X^{\prime})$.
Dashed lines in the audio indicate true word boundaries.
Criticism. The authors of Speech2Vec [63] did not make their code available,
despite multiple requests from the public. They did however release a set of
embeddings produced by the
model.111https://github.com/iamyuanchung/speech2vec-pretrained-vectors With
numerous attempts to replicate the model’s embeddings exactly as described in
the paper, and reaching out to the authors without receiving a response, we
have not been able to achieve results close to theirs. We are also not aware
of any credible publications that successfully replicated Speech2Vec to obtain
similar quality semantic embeddings on the same evaluation tasks. Other failed
attempts to replicate the results of Speech2Vec include the works of
Jang222https://github.com/yjang43/Speech2Vec,
Wang333https://github.com/ZhanpengWang96/pytorch-speech2vec and
EarthSpecies444https://github.com/earthspecies/audio-embeddings/issues/6.
Chen and Cao [131] recently performed an in-depth analysis questioning the
authenticity of the original Speech2Vec model. On a homophone inspection
analysis, they show that Speech2Vec is immune to phonetic interference. For
example, the model will be able to perfectly discriminate between spoken
instances of “steal” and “steel” despite being phonetically identical, which
should inevitably affect model performance. By inspecting the vocabulary of
the released embeddings, they found it exactly matches the vocabulary of a
word embedding model trained on the corresponding transcripts. They show this
should not be true given that some information (words) are lost when applying
a forced-aligner to segment the audio. Based on these findings, there is
compelling evidence that the embeddings released by Speech2Vec have been
influenced by word labels in some way.
Moreover, the CAE-RNN AWE was originally designed to learn embeddings that are
invariant to speaker identity, channel characteristics, etc. (Section 3.2.1).
In the Speech2Vec setup, training pairs are from the same speaker and
utterance (the way context pairs are constructed), so that the model is not
explicitly able to learn embeddings invariant to those properties. Despite the
scepticism of Speech2Vec we include the model as a baseline in our
experiments. We use our own version of the model and perform experiments in a
different setup (but still refer to it as Speech2Vec).
#### 7.4.2 Semantic ContrastiveRNN
Similar to how the CAE-RNN AWE model is adapted for semantic AWE modelling
(Speech2Vec), we modify the ContrastiveRNN AWE model (Section 3.3) to learn
semantic AWEs. We substitute $X_{a}$, $X_{p}$ and $X_{n_{1}},\ldots,X_{n_{K}}$
in Equation 3.6, with $X_{\text{trg}}$ (target centre), $X_{\text{ctx}}$
(context), and out-of-context words. Now, by minimising the contrastive loss
of Equation 3.6, the model should optimise relative distances between
embeddings such that it reflect semantic similarity. This is similar to the
contrastive optimisation for semantic AWEs by Chen et al. [64], without
speaker disentanglement (requiring speaker labels).
For both Speech2Vec and semantic ContrastiveRNN, a semantic AWE model is
trained from scratch, therefore requiring the models to learn phonetic and
semantic similarity simultaneously. We now look at new semantic AWE modelling
strategies that we propose, utilising a pre-trained multilingual AWE model.
### 7.5 Our approach: Semantic AWEs using multilingual transfer
This section introduces a novel semantic AWE modelling approach. We propose
incorporating phonetic word knowledge through a pre-trained multilingual AWE
model during semantic modelling. Incorporating phonetic details may assist the
model to prioritise learning word semantics by filtering out unrelated
characteristics, for example, speaker information. Specifically, we propose
three strategies utilising a pre-trained multilingual AWE model in various
ways.
#### 7.5.1 ContrastiveRNN with multilingual initialisation
Instead of training semantic models from scratch (Section 7.4) we warm-start
them using the learned weights of a pre-trained multilingual AWE model. In our
experiments, we use the learned weights of a multilingual AWE model’s encoder
to initialise the encoder RNN of the ContrastiveRNN and subsequently train a
semantic ContrastiveRNN (Section 7.4.2). This learned parameter initialisation
method is similar to the multilingual adaptation strategy explored in Chapter
4: a pre-trained multilingual AWE model is adapted to a target low-resource
language using discovered word pairs. Here, we adapt the multilingual weights
to produce semantic AWEs by training on context pairs from the target
language.
#### 7.5.2 Projecting multilingual AWEs
Figure 7.5: An illustration of the Project AWE semantic strategy, as detailed
in Section 7.5.2.
An alternative approach would be to first apply the multilingual AWE model to
the segmented speech segments $\\{X^{(n)}\\}^{N}_{n=1}$ where $N$ is the
number of word tokens in the training corpus. Then, train a projection
network, mapping the resulting phonetic AWEs to a new semantic AWE space.
Figure 7.5 illustrates this mapping process. Applying the multilingual model
to the unlabelled speech segments $\\{X^{(n)}\\}$ produces a set of phonetic
AWEs $\\{\mathbf{z}^{(n)}_{\text{p}}\\}$ (upper part in figure). In the
phonetic space, we note that word segments are mapped to unknown word classes
(five in this simplified example). This implies $\mathbf{z}_{\text{p}}^{(35)}$
(blue) and $\mathbf{z}_{\text{p}}^{(71)}$ most likely belong to the same
unknown word class, as well as, $\mathbf{z}_{\text{p}}^{(18)}$ (green) and
$\mathbf{z}_{\text{p}}^{(87)}$. Groups containing
$\mathbf{z}_{\text{p}}^{(35)}$ and $\mathbf{z}_{\text{p}}^{(18)}$ are in close
proximity, indicating they might also share phonetic similarities.
In the semantic space, we would like to see embeddings of segment $X^{(35)}$
and $X^{(36)}$ to be positioned close to each other (assuming the unknown
classes associated with these segments frequently co-occur). We do this by
training a projection network, mapping the phonetic AWEs to a new set of
semantic embeddings $\\{\mathbf{z}^{(n)}_{\text{s}}\\}$. We train this network
using context pairs and the contrastive loss of Equation 3.6, optimising the
distances between the output embeddings $\mathbf{z}_{\text{s}}$. The lower
part of Figure 7.5 illustrates the transformation of the phonetic space into a
new semantic space, through optimising the contrastive loss. We now observe
neighbouring word segments $X^{(35)}$ and $X^{(36)}$ end up having similar
semantic representations $\mathbf{z}^{(35)}_{\text{s}}$ and
$\mathbf{z}^{(36)}_{\text{s}}$.
This approach is similar to the semantic AWE method presented by Chen et al.
[64]. Here they used data from the target language to train a network
separating speaker and phonetic information (which requires speaker labels).
Instead, we use a pre-trained multilingual AWE model designed for generating
phonetic representations for unlabelled word segments.
#### 7.5.3 Cluster+Skip-gram
Figure 7.6: Our Cluster+Skip-gram semantic AWE approach. Speech segments
$X^{(n)}$ are represented by soft pseudo-word label vectors $\mathbf{v}^{(n)}$
which are then used to train a Skip-gram model.
This new approach is based on the Skip-gram model (Section 7.2). The Skip-gram
model uses a fixed dictionary of discrete word class labels to create input
and output vectors for training on text data. In the low-resource setting we
do not have word labels for speech segments.
Instead, we use the phonetic similarities in the original AWE space to derive
a soft pseudo-word label for each speech segment. This process is illustrated
in Figure 7.6. First, a pre-trained multilingual AWE model is applied to the
segmented speech corpus $\\{X^{(n)}\\}$, producing a set of phonetic AWEs
$\\{\mathbf{z}_{\text{p}}^{(n)}\\}$. The colourless embedding space represents
the unsupervised setting where embeddings have no associated word types or
labels. We then apply $K$-means clustering to the phonetic embedding space,
producing a set of centroids $\\{\mathbf{c}_{k}\\}^{K}_{k=1}$, with $K$
representing the predetermined number of clusters for partitioning the vector
space.
Figure 7.7: Given an acoustic embedding $\mathbf{z}_{\text{p}}^{(n)}$
(assigned to cluster 2), instead of using a one-hot encoded input
$\mathbf{x}=\\{0,1,0,\cdots,0\\}$, the value at each index is set using
Equation 7.2, for example, $\mathbf{v}=\\{0.03,0.45,0.01,\cdots,0.43\\}$. Note
that the values at index $2$ and $K$ are the highest given that
$\mathbf{z}_{\text{p}}^{(n)}$ is the closest to cluster means $\mathbf{c}_{2}$
and $\mathbf{c}_{K}$. Black triangles indicate cluster means.
Ideally, these clusters should resemble distinct word classes. For example,
all spoken instances of “apple” should be assigned to cluster one (purple) and
all instances of “water” should be assigned to cluster two (green). Note, with
a perfect clustering, where all embeddings from an unknown word label are
exclusively assigned to a unique cluster, training a Skip-gram model using the
cluster labels would be equivalent to training on the corresponding text
labels (see bottom of Section 7.2). However, due to clustering inaccuracies,
not all embeddings belonging to the same unknown word class will be assigned
to the same cluster. The main challenge arises from embeddings of the same
word class being assigned to multiple clusters; a consequence of the
uncertainty regarding vocabulary size and word class frequency range. For
example, embeddings assigned to cluster $2$ (green) and cluster $K$ (blue) in
Figure 7.7 might represent the same word class given that there is no clear
separation between them. We experimented with cluster labels as training
labels for Skip-gram modelling, but this gave very poor performance on
development data.
Rather than assigning a hard label to each segment based on its associated
cluster number, we calculate a soft vector label of an AWE belonging to each
cluster:
$v_{k}^{(n)}=\frac{\exp\left(-\text{sim}(\textbf{z}_{\text{p}}^{(n)},\textbf{c}_{k})/\sigma^{2}\right)}{\sum_{j=1}^{K}\exp\left(-\text{sim}(\textbf{z}_{\text{p}}^{(n)},\textbf{c}_{j})/\sigma^{2}\right)}$
(7.2)
where $\text{sim}(\cdot)$ denotes cosine similarity and $\sigma$ is a
hyperparameter controlling the influence of distant centroids. This non-linear
scaling approach assigns more weight to embeddings closer to the target
embedding $\mathbf{z}^{(n)}_{\text{p}}$. Figure 7.7 extends the soft labelling
calculation illustration in the bottom right in Figure 7.6. Now, each word
segment $X^{(n)}$ is associated with a new unique vector
$\mathbf{v}^{(n)}\in\mathbb{R}^{K}$, where the value at index $k$ indicates
the likelihood of the segment being contained in cluster $k$. Even if segments
from the same unknown word class are assigned to multiple clusters, their
relative position in the embedding space is captured in $\mathbf{v}$.
Lastly, we use these soft vectors as inputs and targets to train a Skip-gram
model, illustrated in Figure 7.8. Compared to the Skip-gram model trained on
text data (Figure 7.3), there is a key difference: using these soft
representations as input, compared to a one-hot vector. This means that
$\textbf{W}_{1}$ do not act as a look-up table, instead, each row contributes
to the hidden embedding with corresponding weight (visually illustrated with
shading in Figure 7.8). In other words, the hidden layer contains the weighted
sum of all rows in $\textbf{W}_{1}$, which we take as the semantic embedding
$\textbf{z}_{\text{s}}$. The same applies to $\textbf{W}_{2}$ when using soft
targets.
Figure 7.8: The skip-gram model using soft inputs and targets.
During inference, a unique embedding $\textbf{z}_{\text{s}}$ is produced for a
word segment by first obtaining a phonetic AWE using the multilingual model,
then calculating representation $\mathbf{v}$ using the trained cluster means,
and finally applying it to the learned weight matrix $\mathbf{W}_{1}$.
### 7.6 Semantic query-by-example
In Chapter 5 and Chapter 6, we performed QbE experiments using AWEs to
calculate the phonetic similarity between speech segments. This involved
searching an unlabelled audio corpus with a query word segment, aiming to
retrieve utterances containing instances of the query. We refer to this as
exact QbE, where the query only retrieves utterances if it contains a word
sharing the same class label. In this chapter, apart from intrinsic AWE
evaluations, we consider the task of semantic QbE as a downstream application
for semantic AWEs. In semantic QbE the goal is to retrieve utterances from an
audio collection that not only match segments containing the exact query
segment but also utterances which contain word segments that are semantically
similar to the query. For example, using a query segment, “football”, the
system should retrieve utterances containing exact query matches for instance,
“the football player is injured”, as well as utterances containing word
segments semantically related to the query for instance, “two kids play soccer
in the field”, and “the player was offside”.
We perform our semantic QbE experiments using a similar setup as in [132].
They performed experiments to produce semantic embeddings on the same data
corpus we use for our semantic AWE experiments. However, during training, they
use images as a weak form of supervision to produce semantic embeddings. To
the best of our knowledge, we are the first to perform semantic QbE using
semantic AWEs trained on unlabelled speech segments.
### 7.7 Setup
In this section we outline the data corpus we use for our semantic AWE
modelling. We provide implementation details for the two semantic models,
Speech2Vec and ContrastiveRNN, trained from scratch (Section 7.4).
Subsequently, we present the specifics of the three semantic AWE models
leveraging a pre-trained multilingual AWE model we introduced in Section 7.5.
Finally, we discuss the evaluation process for our semantic AWEs.
#### 7.7.1 Data
Figure 7.9: Examples of images with corresponding spoken captions from the
FACC.
For this experiment, we use the Flickr8k Audio Captions Corpus (FACC) [128].
The corpus contains 40k spoken captions in English describing the content of a
Flickr image [133]. These images depict scenes for which humans have given
verbal descriptions. Example images with corresponding spoken captions are
shown in Figure 7.9. Note, the text in Figure 7.9 is only for the purpose of
illustration and not used during training (the corresponding speech are used).
The corpus is divided into train, development, and test splits containing
$30\,000$, $5\,000$, $5\,000$ utterances respectively.
Although the amount of training data in the Flickr8k corpus is relatively low,
the content of the data suits our semantic experiments well: images come from
a fairly narrow domain, with five different spoken captions describing the
content of each image. Therefore, semantic concepts reoccur in different
utterances. This dataset has also been used for acoustic semantic modelling by
others [128, 91, 132].
We train and evaluate our models using two sets of speech features. Firstly,
we employ 13-dimensional MFCCs. Secondly, as in Chapter 6, we conduct
experiments with self-supervised speech features: we use the 12th transformer
layer of the multilingual XLSR model [119] to get 1024-dimensional features.
Utterances are normalised per speaker and segmented using true word boundaries
from forced alignments [121]. We use these word segments from the training
data to construct context word pairs. For all the semantic models, we use a
context window of three words before and after a centre word: this yields
roughly two million context pairs in our training data.
#### 7.7.2 Semantic models trained from scratch
The encoder and decoder of our Speech2Vec (Section 7.4.1) implementation each
consists of three unidirectional RNNs with 400-dimensional hidden vectors and
an embedding size of 100. The semantic ContrastiveRNN (Section 7.4.2) uses the
same encoder structure. We train both these models using the context pairs
extracted from the training set. For the ContrastiveRNN we randomly sample
word segments outside the context window as negatives; for each positive, we
sample 20 negatives.
#### 7.7.3 Semantic models using multilingual transfer
All the multilingual transfer strategies presented in Section 7.5 use a pre-
trained multilingual AWE model. The multilingual model is trained in the same
way multilingual models were trained in previous chapters, with embedding
dimension set to 100. We train a multilingual CAE-RNN AWE model on labelled
data from four different well-resourced languages included in Common Voice
[118]: Italian, Dutch, Russian, and Czech. Note, the entire speech corpus is
in English and therefore not included as a training language. We pool the data
from all four languages and extract 300k training pairs.
Multilingual initialisation. (Section 7.5.1) For the multilingual weight
initialisation method, we use the trained parameters of the multilingual CAE-
RNN encoder to initialise the ContrastiveRNN’s encoder. Furthermore, we freeze
the weights of the first two encoder layers to preserve the phonetic structure
of input words during further training. This initialisation approach is
similar to how we initialised AWE models for adaptation in Chapter 4: a
multilingual AWE model is adapted using discovered word pairs from the target
language. Here, we adapt for semantic AWEs by training on context pairs from
the target corpus.
Project AWE. (Section 7.5.2) For our projection network, we use a feed-forward
network of two linear layers with an inner dimension of 1024 and input and
output dimensions of 100. The input dimension is set to match the embedding
dimension produced by the multilingual AWE model. The inner and output
dimensions are chosen based on experimenting on development data. Again, we
train on context pairs and sample 20 negatives for each positive. The network
optimises by minimising the contrastive loss of Equation 3.6.
Cluster+Skip-gram. (Section 7.5.3) For $K$-means clustering, we use $K=5\,000$
clusters and set $\sigma=0.01$ in Equation 7.2. The number of unique word
types in the target corpus is $6\,307$ where $2\,150$ appears once. We
experimented with different values for $K$ ranging from $1\,000$ to $10\,000$
which produced similar results. We initialise our cluster means with a random
selection of acoustic embeddings, perform 50 permutations, and keep the best
means. We reimplement the Skip-gram model [127] described in Section 7.2,
setting the latent embedding dimension to $D=100$. The model is trained using
the soft-vectors as input and optimises a negative log-likelihood with class
probabilities set to the context target word’s soft-vector. The model is
optimised using Adam [103], with a learning rate of $0.001$.
#### 7.7.4 Evaluation
We evaluate the quality of resulting semantic embeddings in an intrinsic word
similarity task. To measure the applicability for real-life applications, we
also apply the semantic AWEs in a downstream semantic QbE task for extrinsic
evaluation.
##### 7.7.4.1 Intrinsic semantic AWE evaluation: Word similarity
Table 7.2: Word-pair similarity ratings extracted from the WordSim353 [134] evaluation set. Similarity scores indicate word relatedness based on human judgements. money | bank | 8.50
---|---|---
physics | proton | 8.12
tennis | racket | 7.56
journey | car | 5.85
smart | student | 4.62
problem | airport | 2.38
professor | cucumber | 0.31
In the text domain, word embeddings are evaluated for semantics by measuring
the similarity between word pairs through vector distance calculations. These
vector distances are then compared to standard reference sets containing
similarity ratings assigned by humans. As an illustration, Table 7.2 shows
some reference word-pair similarities from the WordSim353 [134] evaluation
set. For instance, in a resulting semantic space, the distance between the
embedding of “tennis” and “racket” should be smaller than the distance between
“journey” and “car”, but greater than the distance between “physics” and
“proton.” Furthermore, “money” and “bank” should have the smallest distance
than all the shown word-pair distances. Spearman’s $\rho$ is used to quantify
the similarity between the set of word-pair distances
$(\text{WP}_{\text{embed}})$ and the set of word-pair human judgement ratings
$(\text{WP}_{\text{human}})$, calculated as follows:
$\rho=\frac{\text{cov}(R(\text{WP}_{\text{embed}}),R(\text{WP}_{\text{human}}))}{\sigma_{R(\text{WP}_{\text{embed}})}\sigma_{R(\text{WP}_{\text{human}})}}$
(7.3)
with $R$ converting the scores to ranks, and $\text{cov}(\cdot\,,\cdot)$ and
$\sigma_{\cdot}$ denotes, covariance and standard deviation, respectively.
[.8] For the sentence below, select all words that could be used to search for
the described scene.
Sentence: a skateboarder in a light green shirt.
$\square$ dogs $\square$ beach $\square$ children $\square$ white
$\square$ swimming $\square$ wearing $\square$ skateboard $\square$ None of
the above
Figure 7.10: An example annotation job presented to a human for manual
semantic labelling, with keyword options. In this case, “skateboard” was
selected by all five annotators and “wearing” by four. Reproduced from [132].
In our controlled experiment, our vocabulary is limited, causing many words
from the standard evaluation sets typically used to assess word embeddings
trained on large amounts of text to be absent from our training and evaluation
data. To address this, we compile our own list of word pairs that frequently
appear in our dataset. Instead of using human judgement, we assign similarity
scores to those pairs based on the distance between textual word embeddings
generated by an off-the-shelf Skip-gram model trained on the corresponding
transcribed utterances. Given that our goal is to obtain semantic embeddings
similar to Word2Vec, we believe this approach is reasonable.
To obtain a single semantic embedding for each word class, we calculate the
average of all embeddings from the same class and report $\rho_{\text{avg}}$.
Given that we are particularly interested in obtaining semantic embeddings for
individual word segments, single-sample performance is also measured by
randomly selecting one instance of each spoken word. This is repeated ten
times and averaged to get a single score $\rho_{\text{single}}$.
##### 7.7.4.2 Semantic QbE
Table 7.3: The number of annotators (out of five) that selected the keyword for that utterance is shown. The majority vote indicates whether the keyword is deemed semantically relevant to an utterance, according to the labelling. Adapted from [132]. Keyword | Example utterance | Count 0 | Majority vote
---|---|---|---
jumps | biker jumps off of ramp | 5 / 5 0 | Y
ocean | man falling off a blue surfboard in the ocean | 5 / 5 0 | Y
race | a red and white race car racing on a dirt racetrack | 5 / 5 0 | Y
snowy | a skier catches air over the snow | 5 / 5 0 | Y
bike | a dirt biker rides through some trees | 4 / 5 0 | Y
children | a group of young boys playing soccer | 4 / 5 0 | Y
young | a little girl in a swing laughs | 4 / 5 0 | Y
field | two white dogs running in the grass together | 3 / 5 0 | Y
carrying | small dog running in the grass with a toy in its mouth | 2 / 5 | Y
face | a man in a gray shirt climbs a large rock wall | 2 / 5 | N
large | a group of people on a zig path through the mountains | 1 / 5 | N
sitting | a baby eats and has food on his face | 1 / 5 | N
hair | two women and a man smile for the camera | 0 / 5 | N
We use the same setup as [91] to evaluate downstream semantic QbE performance.
Semantic labels for 1 000 test utterances from FACC were collected from human
annotators, using a set 67 keyword classes [132]. Specifically, each of the 1
000 utterances was labelled by five annotators, indicating whether a
particular keyword is semantically relevant to that utterance (regardless of
whether the word instance appears verbatim in the utterance). An example
annotation job is shown in Figure 7.10.
We use the majority decision to assign a hard label for whether a keyword is
relevant to an utterance as in [132]. Table 7.3 shows annotations for some of
the utterances. For instance, a correct retrieval using the keyword “jump”
would be “bike jumps of ramp”, while an incorrect retrieval using the keyword
“face” would be “a man in a gray shirt climbs a large rock wall”. Using these
hard labels, we calculate semantic $P@10$, $P@N$, EER (metric details in
Section 5.3).
Given the ambiguity among annotators when determining if a keyword is
semantically related to an utterance, we also use the counts as soft labels
and calculate Spearman’s $\rho$ (Equation 7.3). Here, $\rho$ measures the
correlation between a system’s ranking and the actual number of annotators who
deemed a query keyword relevant to an utterance.
To simplify the QbE task, we still assume that ground truth word boundaries
are known: a query AWE is therefore compared to AWEs for the word segments in
an unlabelled search utterance.
### 7.8 Experimental results
We first compare embedding quality through measuring word similarity between
the semantic AWE models trained from scratch (Section 7.4) and the semantic
AWE models we proposed leveraging multilingual AWEs (Section 7.5). We then
apply one of our proposed semantic AWE models in a semantic QbE search for
extrinsic evaluation. This is followed by a qualitative analysis of our
semantic AWEs.
#### 7.8.1 Intrinsic evaluation: Word similarity
Table 7.4: Spearman’s $\rho$ (%) for semantic word similarity of AWEs on development data. Embeddings are either averaged over word segments of the same class, or a single segment is sampled per class. AWE models take either MFCC or XLSR features as input. | MFCC | XLSR
---|---|---
Model | $\rho_{\text{single}}$ | $\rho_{\text{avg}}$ | $\rho_{\text{single}}$ | $\rho_{\text{avg}}$
Trained from scratch: (Section 7.4) | | |
Speech2Vec | 01.0 | 05.3 | 07.2 | 21.3
ContrastiveRNN | 00.7 | 05.1 | 04.6 | 25.9
Using multilingual transfer: (Section 7.5) | | |
ContrastiveRNN multilingual init. | 00.6 | 05.5 | 06.1 | 24.5
Project AWE | 03.6 | 18.2 | 18.3 | 33.6
Cluster+Skip-gram | 18.0 | 35.6 | 35.9 | 41.7
Table 7.4 presents the intrinsic scores of embeddings from the semantic AWE
models, trained either from scratch (top section) or using multilingual
transfer (bottom). The benefit of multilingual transfer is evident in the
scores of the projection and Cluster+Skip-gram approaches, with the latter
outperforming all other models regardless of the input features used or
whether single or averaged embeddings are evaluated. The single-sample
performance $\rho_{\text{single}}$ is particularly significant as it shows
that individual representations can be compared accurately—a useful property
for downstream applications such as semantic QbE (Section 7.6).
The ContrastiveRNN is the one exception that does not show a clear gain from
initialising with multilingual weights compared to training from scratch. As a
sanity check, we evaluate the phonetic multilingual AWEs before semantic
training, in other words, embeddings produced by the foundation multilingual
AWE model (top of Section 7.7.3) used for transfer in the bottom section. We
obtain results for $\rho_{\text{single}}=\text{0.59\%}$ and
$\rho_{\text{avg}}=-\text{0.13\%}$. As expected, this indicates that phonetic
multilingual AWEs do not capture semantic information.
Again, we see the benefit of using self-supervised speech representations as
input to AWEs instead of conventional features, as also in previous work [135,
122]. We therefore use XSLR features for all further experiments. We apply the
semantic AWE model achieving the best results on the word similarity task,
Cluster+Skip-gram, in our semantic QbE search.
#### 7.8.2 Extrinsic evaluation: Semantic QbE
Table 7.5 compares the Cluster+Skip-gram (semantic) and multilingual AWE
(phonetic) models when used in a downstream QbE system. We evaluate both exact
and semantic QbE, where the latter gets awarded for retrieving exact query
matches as well as utterances labelled as semantically related to the search
query. To situate results, we use a random baseline model that assigns a
random relevance score to each utterance. (The relatively high scores of the
random approach are due to the narrow domain of the evaluation data, Section
7.7.1.)
Table 7.5: Exact and semantic QbE results (%) on test data. | Exact QbE | Semantic QbE
---|---|---
Model | $P@10$ | $P@N$ | EER | $P@10$ | $P@N$ | EER | $\rho$
Baselines: | | | | | | |
Random | 05.0 | 05.0 | 50.0 | 09.1 | 09.1 | 50.0 | 05.7
Multilingual AWE | 90.7 | 82.8 | 04.7 | 92.8 | 59.4 | 24.1 | 17.0
Semantic AWE model: | | | | | | |
Cluster+Skip-gram | 85.8 | 71.3 | 09.3 | 88.2 | 52.1 | 21.6 | 28.2
Table 7.6: Semantic QbE results (%) on test data, where any instance of a query is masked from the search utterances. Model | $P@10$ | $P@N$ | EER | Spearman’s $\rho$
---|---|---|---|---
Baselines: | | | |
Random | 09.2 | 09.0 | 50.0 | 05.4
Multilingual AWE | 21.5 | 15.6 | 46.1 | 06.4
Semantic AWE model: | | | |
Cluster+Skip-gram | 29.9 | 23.1 | 32.3 | 22.9
Looking at the EER and Spearman’s $\rho$ for semantic QbE, we see that the
Cluster+Skip-gram model achieves the highest score, outperforming the purely
phonetic AWEs from the multilingual AWE model. The reason why the phonetic
multilingual AWE model outperforms the semantic model in $P@10$ and $P@N$ is
due to its proficiency in detecting exact matches (which are also correct
semantic matches).
In this chapter, we are mainly interested in observing information captured by
our embeddings related to word meaning. However, in the semantic QbE task, a
system also gets rewarded for retrieving utterances containing exact instances
of the query. To get a better sense of the ability of a model to retrieve non-
verbatim semantic matches, we mask out all exact occurrences of the query
before we segment the search collection. The system will now only be able to
retrieve utterances based on word meaning. The results are shown in Table 7.6.
Now we see a clear benefit in using the Cluster+Skip-gram model, with the
phonetic multilingual AWE model becoming close to a random search given that
there are no exact matches.
Our core goal is semantic QbE, but it is worth briefly touching on exact QbE
performance of the Cluster+Skip-gram model in Table 7.5. In exact QbE, the
system should only retrieve utterances containing instances of the query word
type. The Cluster+Skip-gram model learns to predict frequently co-occurring
words, without explicit training for preserving word-type information.
Although only trained for semantics, the Cluster+Skip-gram still achieves
reasonable exact retrieval performance, with only a drop of between 5% and 10%
in scores compared to the multilingual AWE model. It is therefore clear that
this semantic model is able to retain phonetic properties for word-type while
also capturing semantic information related to context.
#### 7.8.3 Qualitative analysis
Figure 7.11: PCA projection of semantic AWEs (averaged), produced by our Cluster+Skip-gram model on development data. We highlight the five nearest neighbours for “several” (pink), “bike” (green), and “orange” (blue). Table 7.7: The top ten closest embeddings for keywords, derived from text (left) and speech segments (right), respectively. Speech embeddings were produced using the Cluster+Skip-gram model and averaged per word-type before nearest neighbour calculation. american | | swimming | | boys | | football
---|---|---|---|---|---|---
flags | african | | pool | swimmers | | children | girls | | opponent | baseball
spiderman | waving | | swim | swims | | kids | children | | soccer | hockey
tin | older | | water | swim | | men | kids | | hockey | soccer
camouflage | costume | | wet | into | | girls | women | | player | uniform
muzzles | elderly | | diving | zooms | | boy | teens | | players | miami
participating | flags | | headfirst | swings | | players | men | | volleyball | player
elderly | asian | | wading | pool | | people | are | | hitting | approaches
tiny | leather | | paddling | scratching | | dolphins | traveller | | uniform | apportioned
instrument | establishment | | swims | scratches | | gateway | cans | | field | saxophone
asian | instruments | | patiently | splashes | | girl | adults | | swim | players
Figure 7.11 visualises the semantic embedding space produced by the
Cluster+Skip-gram model. It is clear that the acoustic realisations of
semantically related words end up in similar areas in the space. For example,
the model learned that spoken instances of “orange”, “red”, “blue”, “yellow”,
and “green” should be close to each other. For additional visualisation, we
direct the reader back to Figure 1.1 and Figure 1.2. Figure 1.1(a) displays a
selected number of TWEs produced by a Skip-gram model trained on the
transcribed speech we use for acoustic semantic modelling in this chapter
(Section 7.7.1). Figure 1.1(b) displays the AWE produced by the multilingual
AWE model we use in our semantic AWE approaches (Section 7.5). Finally, Figure
1.2 displays the embeddings (without averaging) produced by our Cluster+Skip-
gram model. From these visualisations, it is evident that the semantic AWEs
satisfactory capture of word-type and semantic information.
The intrinsic results in Table 7.4 and qualitative analysis in Figure 1.2 show
that our semantic embeddings, to some extent, resemble the semantic
information captured in the corresponding text embeddings. To get a better
understanding of the difference between the embeddings produced from speech
and the corresponding transcripts, we look at the variation in the nearest
embeddings of some words. Table 7.7 displays chosen keywords, comparing
semantic AWEs (on the right) with textual word embeddings (on the left).
Firstly, we notice that across all keywords some words are shared (highlighted
in bold). Secondly, we observe a strong phonetic influence on the semantic
embeddings trained on speech. They are more likely to have words higher ranked
that are phonetically similar to the keyword. For example, “american” and
“football”, have “african” and “baseball”, as their closest neighbour,
respectively. In this instance, we could say they are semantically similar
from our own judgement, but they do not appear among the nearest words for the
corresponding text embeddings. Looking at the keyword “swimming” we can see
where this phonetic disturbance has a negative impact. For“swimming”, the word
“swims” appear in the top neighbours for both speech and text. Because “swims”
are phonetically similar to “zooms” and “swings”, the model fails to separate
them in vector space. As a result, “swimming” also ends up close to “zooms”
and “swings”. In this instance, we would not regard them as semantically
relevant to each other.
### 7.9 Chapter summary
In this chapter, we addressed a new form of acoustic representation learning,
where word embeddings should capture the meaning of words, rather than
phonetic information. To achieve this, we presented several semantic AWE
modelling strategies. We specifically promoted transferring knowledge from a
pre-trained multilingual AWE model trained for word-class discrimination.
Through intrinsic and extrinsic evaluations, we demonstrated the effectiveness
of our strategies in learning semantic representations from unlabelled speech
data. Our best semantic AWE approach involves a soft clustering on the
original multilingual AWEs, serving as input to a Skip-gram model.
The main shortcoming of our work (as also in others [63, 64]), is that the
word segmentation is assumed to be known. This was reasonable given our goal
of comparing different semantic AWE approaches on a sensible benchmark, but
future work should look into incorporating unsupervised word segmentation
methods [136, 137, 138] in order to do fully unsupervised semantic AWE
modelling.
## Chapter 8 Summary and Conclusions
Developing robust speech systems for zero-resource languages—where no
transcribed speech resources are available for model training—remains a
challenge. Most of these systems, at their core, require speech segments of
different duration to be compared. Previous studies showed the benefit of AWEs
over alignment methods for this task. In a zero-resource setting, the two
existing training strategies are unsupervised monolingual learning (Section
2.4.1) and supervised multilingual transfer (Section 2.4.2). This thesis
mainly contributes to the development of AWEs following the multilingual
transfer strategy. We presented multiple approaches toward improving the
intrinsic quality of AWEs as well as the implementation of AWEs in a
downstream speech retrieval task. Furthermore, we extended the multilingual
transfer strategy to semantic representation learning, demonstrating the
versatility of multilingual AWEs in upstream tasks. We first summarise our
main findings. We then provide recommendations on how the work can be extended
and applied in future research.
### 8.1 Main findings
In Chapter 3, we presented the ContrastiveRNN AWE model. This model optimises
a distance metric function not considered for AWE modelling in previous work.
We compared the ContrastiveRNN to two existing AWE embedding models, CAE-RNN
[21] and SiameseRNN [32]. We re-implemented both existing models and compared
our model in the same experimental setup as Kamper et al. [23]. On a word-
discrimination task measuring the intrinsic quality of the AWEs, the
ContrastiveRNN significantly outperformed the CAE-RNN and SiameseRNN following
the unsupervised monolingual training strategy, reaching up to 17.8% absolute
increase in average precision on one of the evaluation languages. In the
supervised multilingual setting, we observed marginal performance increases
from the ContrastiveRNN compared to the CAE-RNN and SiameseRNN.
In Chapter 4, we proposed a new training strategy that extends the
multilingual transfer strategy. After training a multilingual model, instead
of only applying it to an unseen target language, we first fine-tune the
multilingual model using discovered word-like pairs obtained from applying an
unsupervised term discovery system (UTD) to unlabelled speech data in the
target language. For each of the three multilingual models, we adapt a model
to every zero-resource language before applying it. Here we observe that the
ContrastiveRNN generally adapts better to a target language compared to the
CAE-RNN and SiameseRNN.
In Chapter 5, we investigated the impact a particular choice of training
languages have when training an AWE model for a specific target language. We
specifically want to see the impact of including training languages that
belong to the same language family as the target language. We performed
multiple experiments on data from languages in South Africa where we train
multilingual AWEs with different language combinations while controlling for
the amount of data. We show that including related languages over unrelated
languages are beneficial, evaluated on an intrinsic word discrimination task.
We also evaluated the multilingual AWEs in a downstream zero-resource speech
task, query-by-example, supporting the results of the intrinsic evaluation
task. To summarise the findings in this chapter, including training data from
related languages, even in small quantities, are beneficial over using
unrelated languages. For instance, a multilingual model trained on
approximately 6 hours of related languages, outperform a model trained on
approximately 60 hours of unrelated languages. The findings in this chapter
would be of great interest to practitioners who want to develop speech
applications in a zero-resource setting.
In Chapter 6, we developed keyword spotting systems to monitor hate speech in
radio broadcast audio in a collaborative effort with a company, VoxCroft. Our
goal was to compare existing ASR-based KWS systems (requiring labelled
training data from the target language) to a new multilingual AWE-based KWS
system. We first developed our systems in a controlled environment where we
used parallel audio-text data from experimental datasets to train and test our
models. In this in-domain environment, an ASR model trained on only five
minutes of transcribed audio outperforms our AWE-based KWS system. We then
applied the models, without further calibrations, to the real radio broadcast
audio scraped from radio stations in Kenya. In this in-the-wild test, our AWE-
based KWS system proves to be more robust against domain mismatch and performs
similar to an ASR-KWS system using 30 hours of labelled target audio. Our
findings in this chapter suggest multilingual AWEs can be used as an
alternative to existing ASR systems for rapidly deploying KWS systems for
languages with limited transcribed audio.
In Chapter 7, we shifted our attention to a new form of representation
learning of spoken word segment: AWEs that reflect word-type information as
well as word meaning. We specifically proposed using a pre-trained
multilingual AWE model to assist in this task compared to existing approaches
that only use unlabelled data from the target language [63, 64]. One of our
models, Cluster+Skip-gram, show great improvement over previous models in an
intrinsic word similarity task, measuring semantic relatedness. For the first
time, we applied these semantic AWEs in a downstream semantic speech retrieval
task. Our results show that semantic information is indeed being captured in
the embeddings produced by our new models, indicating promise for future
downstream applications.
### 8.2 Future work
Throughout this thesis, we identified multiple opportunities to expand on in
future research.
* •
All our AWE models up to Chapter 5 took hand-engineered MFCCs as input
features. In Chapter 6 and Chapter 7 we incorporated self-supervised speech
features as an alternative to MFCCs for input to our AWE models. These learned
speech features showed improved performance in a variety of speech tasks [76,
77, 78, 79]. We also observed the benefits in AWE modelling using these
features: improved word discrimination (Section 6.7.1) and improved semantic
representations (Table 7.4). However, we did not systematically investigate
this but merely replaced features without considering alterations to our
models or training conditions. We hypothesise further improvement by
optimising the training setup to accommodate these speech features.
* •
We notice a large performance gap between the unsupervised monolingual models
trained on discovered word pairs (top of Table 3.1) and the supervised
monolingual models trained on true word segments (top of Table 4.2). A more
accurate unsupervised segmentation of unlabelled speech data, delivering
higher quality discovered pairs, could reduce the gap between the unsupervised
and supervised training strategies. Kamper et al. [93] presented a full-
coverage segmentation and clustering system—unlabelled speech data are
segmented and clustered into unknown word types. Their system employs
primitive acoustic word embeddings. By replacing the AWEs with higher quality
AWEs, for instance, embeddings produced by a multilingual AWE model, better
segmentation can be expected. Applying this segmentation system to unlabelled
speech data in a zero-resource language could potentially yield higher quality
discovered word pairs compared to the UTD system of Jansen and Van Durme [10]
used in this thesis.
* •
Up until now, the choice of training languages in a multilingual transfer
setting have not been considered, specifically for AWEs. We showed that a more
careful selection of training languages can improve performance when
developing a multilingual AWE model for a specific target language by
including training languages that belong to the same language family as the
target. We showed that the performance increase is not simply due to the
overlap of words in related languages. However, a more in-depth analysis of
the phonetic and syntactic properties common to languages from the same
language family could give more insight assisting the development of
multilingual transfer learning in general.
* •
In Chapter 6, we identified the need for a more optimal threshold value to
improve KWS performance. We hypothesise that a per-keyword threshold could be
determined based on factors including the length of the speech segment, the
number of phones present, and the relative distances between other unknown
word segments in the acoustic space. In the work of Matusevych et al. [104],
they analyse an AWE space and discover that the embeddings encode absolute
duration and speaker information. Additionally, they observe that the distance
between word embeddings increases with phonetic dissimilarity. With such
knowledge we could explore potential correlations between these factors stored
in AWEs and the distance between instances in the acoustic space.
* •
In Chapter 7, we introduced a novel approach for learning speech
representations that reflect semantic meaning instead of acoustic similarity.
Although our approach exhibits improved performance compared to existing
methods, certain limitations remain: Similar to others [63, 64], we used true
word boundaries for segmenting unlabelled audio. Future work should consider
unsupervised segmentation methods [136, 137, 138], or explore an end-to-end
semantic AWE approach that eliminates the need for explicit word boundaries.
Additionally, we performed our experiments within a constrained setup
involving a limited vocabulary and a smaller number of words when compared to
the millions or even billions used for training textual word embeddings.
Future work should therefore investigate the scalability of the semantic
approaches we have introduced.
## References
## References
* Besacier et al. [2014] L. Besacier, E. Barnard, A. Karpov, and T. Schultz, “Automatic speech recognition for under-resourced languages: A survey,” _Speech Commun._ , vol. 56, pp. 85–100, 2014.
* Eberhard et al. [2021] Eberhard, M. David, G. F. Simons, and C. D. Fenning, “Ethnologue: Languages of the world,” 2021. [Online]. Available: https://www.ethnologue.com
* Versteegh et al. [2016] M. Versteegh, X. Anguera, A. Jansen, and E. Dupoux, “The Zero Resource speech challenge 2015: Proposed approaches and results,” in _Proc. SLTU_ , 2016.
* Jansen et al. [2013] A. Jansen, E. Dupoux, S. Goldwater, M. Johnson, S. Khudanpur, K. Church, N. Feldman, H. Hermansky, F. Metze, R. Rose, M. Seltzer, P. Clark, I. McGraw, B. Varadarajan, E. Bennett, B. Borschinger, J. Chiu, E. Dunbar, A. Fourtassi, D. Harwath, C.-y. Lee, K. Levin, A. Norouzian, V. Peddinti, R. Richardson, T. Schatz, and S. Thomas, “A summary of the 2012 JHU CLSP workshop on zero resource speech technologies and models of early language acquisition,” in _Proc. ICASSP_ , 2013.
* Räsänen [2012] O. Räsänen, “Computational modeling of phonetic and lexical learning in early language acquisition: Existing models and future directions,” _Speech Commun._ , vol. 54, no. 9, pp. 975–997, 2012.
* Levin et al. [2015] K. Levin, A. Jansen, and B. Van Durme, “Segmental acoustic indexing for zero resource keyword search,” in _Proc. ICASSP_ , 2015.
* Huang et al. [2018] S.-F. Huang, Y.-C. Chen, H.-y. Lee, and L.-s. Lee, “Improved audio embeddings by adjacency-based clustering with applications in spoken term detection,” _arXiv preprint arXiv:1811.02775_ , 2018.
* Yuan et al. [2018] Y. Yuan, C.-C. Leung, L. Xie, H. Chen, B. Ma, and H. Li, “Learning acoustic word embeddings with temporal context for query-by-example speech search,” in _Proc. Interspeech_ , 2018.
* Park and Glass [2008] A. S. Park and J. R. Glass, “Unsupervised pattern discovery in speech,” _IEEE Trans. Audio, Speech, Language Process._ , vol. 16, no. 1, pp. 186–197, 2008.
* Jansen and Van Durme [2011] A. Jansen and B. Van Durme, “Efficient spoken term discovery using randomized algorithms,” in _Proc. ASRU_ , 2011.
* Ondel et al. [2019] L. Ondel, H. K. Vydana, L. Burget, and J. Černocký, “Bayesian subspace hidden markov model for acoustic unit discovery,” in _Proc. Interspeech_ , 2019.
* Räsänen and Blandón [2020] O. Räsänen and M. A. C. Blandón, “Unsupervised discovery of recurring speech patterns using probabilistic adaptive metrics,” _arXiv preprint arXiv:2008.00731_ , 2020.
* Kamper et al. [2017a] H. Kamper, K. Livescu, and S. Goldwater, “An embedded segmental K-means model for unsupervised segmentation and clustering of speech,” in _Proc. ASRU_ , 2017.
* Seshadri and Räsänen [2019] S. Seshadri and O. Räsänen, “SylNet: An adaptable end-to-end syllable count estimator for speech,” _IEEE Signal Process. Letters_ , vol. 26, no. 9, pp. 1359–1363, 2019.
* Kreuk et al. [2020] F. Kreuk, J. Keshet, and Y. Adi, “Self-supervised contrastive learning for unsupervised phoneme segmentation,” in _Proc. Interspeech_ , 2020.
* Rabiner et al. [1978] L. Rabiner, A. Rosenberg, and S. Levinson, “Considerations in dynamic time warping algorithms for discrete word recognition,” _IEEE Trans. Acoust., Speech, Signal Process._ , vol. 26, no. 6, pp. 575–582, 1978.
* Levin et al. [2013] K. Levin, K. Henry, A. Jansen, and K. Livescu, “Fixed-dimensional acoustic embeddings of variable-length segments in low-resource settings,” in _Proc. ASRU_ , 2013.
* Holzenberger et al. [2018] N. Holzenberger, M. Du, J. Karadayi, R. Riad, and E. Dupoux, “Learning word embeddings: Unsupervised methods for fixed-size representations of variable-length speech segments,” in _Proc. Interspeech_ , 2018.
* Chung et al. [2016] Y.-A. Chung, C.-C. Wu, C.-H. Shen, H.-Y. Lee, and L.-S. Lee, “Unsupervised learning of audio segment representations using sequence-to-sequence autoencoder,” in _Proc. Interspeech_ , 2016.
* Kamper et al. [2016a] H. Kamper, W. Wang, and K. Livescu, “Deep convolutional acoustic word embeddings using word-pair side information,” in _Proc. ICASSP_ , 2016\.
* Kamper [2019] H. Kamper, “Truly unsupervised acoustic word embeddings using weak top-down constraints in encoder-decoder models,” in _Proc. ICASSP_ , 2019.
* Kamper et al. [2020] H. Kamper, Y. Matusevych, and S. Goldwater, “Multilingual acoustic word embedding models for processing zero-resource languages,” in _Proc. ICASSP_ , 2020.
* Kamper et al. [2021] ——, “Improved acoustic word embeddings for zero-resource languages using multilingual transfer,” _IEEE Trans. Audio, Speech, Language Process._ , vol. 29, pp. 1107–1118, 2021.
* Hu et al. [2021] Y. Hu, S. Settle, and K. Livescu, “Acoustic span embeddings for multilingual query-by-example search,” in _Proc. SLT_ , 2021.
* Hu et al. [2020] ——, “Multilingual jointly trained acoustic and written word embeddings,” in _Proc. Interspeech_ , 2020.
* Ruder [2019] S. Ruder, “Neural transfer learning for natural language processing,” PhD diss., NUI Galway, 2019.
* Rumelhart et al. [1986] D. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning representations by back-propagating errors,” _Nature_ , vol. 323, pp. 533–536, 1986.
* Mikolov et al. [2013a] T. Mikolov, I. Sutskever, K. Chen, G. Corrado, and J. Dean, “Distributed representations of words and phrases and their compositionality,” in _Proc. NeurIPS_ , 2013.
* Pennington et al. [2014] J. Pennington, R. Socher, and C. Manning, “Glove: Global vectors for word representation,” in _Proc. EMNLP_ , 2014.
* Jacobs et al. [2021] C. Jacobs, Y. Matusevych, and H. Kamper, “Acoustic word embeddings for zero-resource languages using self-supervised contrastive learning and multilingual adaptation,” in _Proc. SLT_ , 2021.
* Chen et al. [2015] G. Chen, C. Parada, and T. N. Sainath, “Query-by-example keyword spotting using long short-term memory networks,” in _Proc. ICASSP_ , 2015.
* Settle and Livescu [2016] S. Settle and K. Livescu, “Discriminative acoustic word embeddings: Recurrent neural network-based approaches,” in _Proc. SLT_ , 2016.
* Doersch and Zisserman [2017] C. Doersch and A. Zisserman, “Multi-task self-supervised visual learning,” in _Proc. ICCV_ , 2017.
* Asano et al. [2020] Y. M. Asano, C. Rupprecht, and A. Vedaldi, “A critical analysis of self-supervision, or what we can learn from a single image,” in _Proc. ICLR_ , 2020.
* Doersch et al. [2015] C. Doersch, A. Gupta, and A. A. Efros, “Unsupervised visual representation learning by context prediction,” in _Proc. ICCV_ , 2015.
* Noroozi and Favaro [2016] M. Noroozi and P. Favaro, “Unsupervised learning of visual representations by solving jigsaw puzzles,” in _Proc. ECCV_ , 2016.
* Gidaris et al. [2018] S. Gidaris, P. Singh, and N. Komodakis, “Unsupervised representation learning by predicting image rotations,” in _Proc. ICLR_ , 2018.
* Pascual et al. [2019] S. Pascual, M. Ravanelli, J. Serrà, A. Bonafonte, and Y. Bengio, “Learning problem-agnostic speech representations from multiple self-supervised tasks,” 2019.
* Synnaeve et al. [2020] G. Synnaeve, Q. Xu, J. Kahn, T. Likhomanenko, E. Grave, V. Pratap, A. Sriram, V. Liptchinsky, and R. Collobert, “End-to-end ASR: From supervised to semi-supervised learning with modern architectures,” in _Proc. ICML_ , 2020\.
* Baevski et al. [2020a] A. Baevski, S. Schneider, and M. Auli, “vq-wav2vec: Self-supervised learning of discrete speech representations,” in _Proc. ICLR_ , 2020.
* Baevski et al. [2020b] A. Baevski, M. Auli, and A. Mohamed, “Effectiveness of self-supervised pre-training for speech recognition,” in _Proc. ICASSP_ , 2020.
* Wang et al. [2020] W. Wang, Q. Tang, and K. Livescu, “Unsupervised pre-training of bidirectional speech encoders via masked reconstruction,” in _Proc. ICASSP_ , 2020.
* Chen et al. [2020] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework for contrastive learning of visual representations,” in _Proc. ICML_ , 2020\.
* Sohn [2016] K. Sohn, “Improved deep metric learning with multi-class N-pair loss objective,” in _Proc. NeurIPS_ , 2016.
* van der Merwe [2020] R. van der Merwe, “Triplet entropy loss: Improving the generalisation of short speech language identification systems,” _arXiv preprint arXiv:2012.03775_ , 2020.
* Yi et al. [2019] J. Yi, J. Tao, Z. Wen, and Y. Bai, “Language-adversarial transfer learning for low-resource speech recognition,” _IEEE/ACM Trans. Audio, Speech, Language Process._ , vol. 27, no. 3, pp. 621–630, 2019.
* van der Westhuizen et al. [2021] E. van der Westhuizen, T. Padhi, and T. Niesler, “Multilingual training set selection for ASR in under-resourced Malian languages,” in _Proc. SPECOM_ , 2021.
* Hazen et al. [2009] T. J. Hazen, W. Shen, and C. White, “Query-by-example spoken term detection using phonetic posteriorgram templates,” in _Proc. ASRU_ , 2009.
* Zhang and Glass [2009] Y. Zhang and J. R. Glass, “Unsupervised spoken keyword spotting via segmental DTW on Gaussian posteriorgrams,” in _Proc. ASRU_ , 2009.
* Jansen and Durme [2012] A. Jansen and B. V. Durme, “Indexing raw acoustic features for scalable zero resource search,” in _Proc. Interspeech_ , 2012.
* Anastasopoulos et al. [2016] A. Anastasopoulos, D. Chiang, and L. Duong, “An unsupervised probability model for speech-to-translation alignment of low-resource languages,” in _Proc. EMNLP_ , 2016.
* Settle et al. [2017] S. Settle, K. Levin, H. Kamper, and K. Livescu, “Query-by-example search with discriminative neural acoustic word embeddings,” in _Proc. Interspeech_ , 2017.
* Yuan et al. [2019] Y. Yuan, C.-C. Leung, L. Xie, H. Chen, and B. Ma, “Query-by-example speech search using recurrent neural acoustic word embeddings with temporal context,” _IEEE Access_ , vol. 7, pp. 67 656–67 665, 2019.
* Ram et al. [2019] D. Ram, L. Miculicich, and H. Bourlard, “Neural network based end-to-end query by example spoken term detection,” _IEEE Trans. Audio, Speech, Lang. Process._ , vol. 28, no. 1, pp. 1416–1427, 2019.
* Saeb et al. [2017] A. Saeb, R. Menon, H. Cameron, W. Kibira, J. Quinn, and T. Niesler, “Very low resource radio browsing for agile developmental and humanitarian monitoring,” in _Proc. Interspeech_ , 2017.
* Menon et al. [2017] R. Menon, A. Saeb, H. Cameron, W. Kibira, J. Quinn, and T. Niesler, “Radio-browsing for developmental monitoring in Uganda,” in _Proc. ICASSP_ , 2017.
* Menon et al. [2018] R. Menon, H. Kamper, J. Quinn, and T. Niesler, “Fast ASR-free and almost zero-resource keyword spotting using DTW and CNNs for humanitarian monitoring,” in _Proc. Interspeech_ , 2018.
* Menon et al. [2019] R. Menon, H. Kamper, E. v. d. Westhuizen, J. Quinn, and T. Niesler, “Feature exploration for almost zero-resource ASR-free keyword spotting using a multilingual bottleneck extractor and correspondence autoencoders,” in _Proc. Interspeech_ , 2019.
* [59] “United Nations strategy and plan of action on hate speech.” [Online]. Available: https://www.un.org/en/genocideprevention/documents/advising-and-mobilizing/Action_plan_on_hate_speech_EN.pdf
* Larson and Jones [2012] M. Larson and G. Jones, “Spoken content retrieval: A survey of techniques and technologies,” _Found. Trends Inform. Retrieval_ , pp. 235–422, 2012\.
|
# Lithium in red giants: the roles of the He-core flash and the luminosity
bump
Deepak1,2 and David L. Lambert3
1Indian Institute of Astrophysics, Bangalore - 560034, India
2Pondicherry University, R. V. Nagara, Kalapet, Puducherry - 605014, India
3W.J. McDonald Observatory and Department of Astronomy, The University of
Texas at Austin, Austin, TX 78712, USA E-mail<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>
(Accepted 9th July 2021. Received 9th July 2021; in original form 31st March
2021)
###### Abstract
Lithium abundances for red giants in the GALAH DR3 survey are studied. The
rare examples of Li-enriched stars with abundances A(Li) $\geq 1.5$ are
confirmed to be He-core burning stars belonging to or evolved from the red
clump with similar masses and metallicity: $M\simeq 1.1\pm 0.2{\rm M_{\odot}}$
and [Fe/H] $\simeq-0.3\pm 0.3$. Li enrichment over the Li abundance present in
a star’s predecessor at the tip of the red giant branch likely occurs in all
these red clump stars. Examination of the elemental abundances (C to Eu) in
the GALAH catalogue shows no anomalous abundances in red clump giants and, in
particular, no dependence on the Li abundance, which ranges over at least five
dex. Lithium synthesis is attributed to the He-core flash occurring in stars
at the tip of the red giant branch. Models from the Modules for Experiments in
Stellar Astrophysics (MESA) match the observed evolution of these stars along
the red giant branch and to the red clump but only at the low effective
temperature end of the observed spread of red clump giants. Run of Li
abundance on the red giant branch is fairly well reproduced by MESA models. A
speculation is presented that the series of He-core flashes not only leads to
7Li synthesis from a star’s internal reservoir of 3He but also may lead to
internal restructuring leading to the observed effective temperature spread of
red clump stars at about a constant luminosity. Giants exhibiting marked Li
enrichments are not found at other evolutionary phases and, in particular, not
directly associated with the luminosity bump on the red giant branch for which
the Li abundance increase does not exceed 0.3 dex.
###### keywords:
Surveys – Hertzsprung-Russell and colour-magnitude diagrams – Stars: evolution
– Stars: abundances – Nuclear reactions, nucleosynthesis, abundances – Stars:
individual: Li-rich giants
††pubyear: 2021††pagerange: Lithium in red giants: the roles of the He-core
flash and the luminosity bump–A
## 1 Introduction
Lithium’s presence in low mass red giants has occasioned considerable interest
across the evolutionary phases from subgiant, the red giant branch, the red
clump He-core burning giant to the asymptotic giant branch and then to the
post-asymptotic giant branch followed by ejection of the envelope with the
stellar residual present as a white dwarf. Lithium present at low abundance in
a giant’s main sequence progenitor is depleted in the giant because the deep
convective envelope drastically dilutes the Li abundance. This major episode
of dilution occurs as the giant evolves to the base of the first red giant
branch (RGB). This stage is known as the first dredge-up (FDU). Continuing
depletion of Li occurs in the lowest mass RGB giants beyond the completion of
the FDU. Given that the effect of the FDU in low mass giants is to lower the
envelope’s Li abundance by approximately a factor of 100 (more precise
estimates are provided below), a red giant post-FDU of about solar metallicity
is expected to have a Li abundance of A(Li) $\sim 1.5$.111Elemental abundances
are given on the traditional scale: A(X) = $\log$[N(X)/N(H)]+ 12 where N(X) is
the number density of element X. A Li abundance of about this value represents
the anticipated upper limit for a giant’s Li abundance because Li may have
experienced depletion, dilution or diffusion in a main sequence star’s outer
envelope well prior to the FDU and the giant’s convective envelope may post-
FDU destroy additional Li. Thus, a Li-rich giant has been widely known as one
with an abundance A(Li)$\geq 1.5$.
Discovery of Li-rich K giants was initiated by Wallerstein & Sneden (1982)’s
finding of a strong 6707 Å Li i line in the post-FDU K III giant HD 112127
with the abundance A(Li) $\simeq 3.2$. Discoveries of additional Li-rich K
giants across a range of Li abundances trickled out after 1982, but, recently,
discoveries have been provided at an accelerated pace thanks in large part to
surveys such as the Gaia-ESO, GALAH and LAMOST surveys (Casey et al., 2016;
Deepak & Reddy, 2019; Gao et al., 2019; Martell et al., 2021). Li-rich giants
are rare, representing a per cent or so of K giants in the Galactic disk
(Deepak & Reddy, 2019). At present, the most Li-rich giant known may be TYC
429-2097-1 with A(Li) = 4.5 (Yan et al., 2018).
The discovery of HD 112127 and others led to speculations about Li production
in these red giants. Almost all proposed explanations involve Li production
tied to a giant’s interior layer of 3He. This He isotope at low abundance is
present in a star from its birth with additional amounts produced in low mass
stars by initial steps of the pp-chain in main sequence H-burning. Li
production from 3He is straightforward for a nuclear astrophysicist: the
reaction 3He($\alpha,\gamma$)7Be is followed by electron capture
7Be(e-,$\nu_{e}$)7Li. For the stellar astrophysicist, this link requires
coupling the hot reaction site where 7Li is created but also destroyed by
protons to the cool stellar atmosphere where the Li may survive and is
observed. This necessary coupling acknowledges too that the intermediary 7Be
has a half-life of just 53 days. The giant’s stellar convective envelope is
required to bring the fresh (7Be and) 7Li to the atmosphere with, of course,
also convection returning 7Li to the interior for destruction and
replenishment. This coupling of nuclear and stellar astrophysics first
recognized by Cameron (1955) is now widely known as the Cameron-Fowler
mechanism.
Cameron & Fowler (1971) suggested this mechanism involving He-burning shell
flashes to account for the rare carbon stars like WZ Cas (McKellar, 1940) and
WX Cyg (Sanford, 1950) with very strong 6707 Å lines. In modern parlance,
these carbon stars are examples of luminous AGB stars experiencing helium
shell burning episodes and the third dredge-up. Lithium, carbon and
$s$-process enrichment are predicted by contemporary modelling and observed in
luminous AGB giants (Smith & Lambert, 1989). However, for K giants like HD
112127, which are not luminous AGB stars, two evolutionary phases have
received the most attention with respect to Li enrichment: giants at the so-
called luminosity bump (LB) on the RGB and more luminous low mass giants at
the RGB’s tip, which experience a He-core flash before evolving to the red
clump (horizontal branch) as a He-core burning giant.
Suspicions about the association of the He-core flash with Li synthesis were
aired in the first survey for Li-rich giants by Brown et al. (1989) and raised
subsequently by others. For example, Deepak & Reddy (2019) found that most of
the Li-rich giants in the GALAH DR2 survey belong to the red clump phase.
These suspicions were wonderfully raised to a near-certainty by Kumar et al.
(2020) whose analysis of Li abundances of giants in the GALAH DR2 survey led
to the bold suggestion that only He-core burning giants were Li-rich and,
thus, that Li synthesis occurs at or immediately following the He-core flash.
They asserted that Li enrichment of He-core burning giants is ‘ubiquitous’ and
confined to these giants, that is – all low mass giants experiencing a He-core
flash synthesize Li at a level from very mild to severe to contaminate the RC
giant which follows the flash. Subsequent studies of asteroseismology of He-
core burning giants with A(Li) determinations suggest that Li abundances
following the He-core flash may decline with age (Deepak & Lambert, 2021;
Singh et al., 2021). The common assumption is that the Li results from the
Cameron-Fowler mechanism. Kumar et al. (2020) did not detail the mechanism’s
application to the He-core flash. A recent theoretical exploration (Schwab,
2020) is discussed below.
Adaptation of the 3He reservoir to Li production has also been connected to
RGB giants experiencing the LB. This ‘bump’ on the RGB occurs at the point
when the H-burning shell moves outward across an H-discontinuity. The
evolution of the star reverses course, and it moves slowly and slightly down
the RGB before reversing again and continuing to evolve up the RGB. This
manoeuvre introduces a clumping of stars on the RGB. At the time of this LB,
operation of the Cameron-Fowler mechanism occurs through a thermohaline
instability expected on theoretical grounds (Eggleton et al., 2008; Lagarde et
al., 2011) and suggested as a possible way to account for Li enriched giants
(e.g., Martell et al., 2021). But definitive observational identification of
Li enriched LB giants remains open.222Thermohaline instability at the LB has
been suggested also as a way to change a giant’s surface 12C/13C ratio.
In this paper, we concentrate on the exploitation of the recent GALAH survey
of stellar abundances, with the focus here on the Li abundances of low-mass
red giants. Our primary goal is to set Li production limits by the He-core
flash as a function of stellar mass and initial metallicity. Our secondary
goal is to search for Li enrichment by LB giants. Our observational inquiry
based on the latest data release DR3 from the GALAH survey is supported by
predictions of Li abundances from calculations performed using the Modules for
Experiments in Stellar Astrophysics (MESA; Paxton et al., 2011, 2013, 2015,
2018, 2019) suite of computer programmes.
## 2 Data sample
Recent observational advances in understanding Li-rich giants have largely
come from combining spectroscopic abundance surveys with astrometric data from
the Gaia satellite (Gaia Collaboration et al., 2016). Li (and other) elemental
abundances from the GALAH DR2 survey have been used by Deepak & Reddy (2019);
Deepak et al. (2020); Kumar et al. (2020) with Martell et al. (2021) using the
DR3 results. The LAMOST survey was considered by Deepak et al. (2020) for C, N
and O (but not for Li) for stars in common with the GALAH DR2. LAMOST’s Li
abundances were explored by Singh et al. (2019) for stars in the Kepler field
of view and in detail by Yan et al. (2021).
Here, we consider the third data release (DR3) of the GALAH survey. Our
analysis is devoted to low mass giants with a metallicity [Fe/H]
representative of the Galactic disk, say $-1\lesssim$ [Fe/H] $\lesssim+0.5.$
GALAH DR3 (and DR2) provides too few metal-poor halo giants to probe the
matter of Li-richness among the halo population. As a secondary interest, we
make a foray into possible abundance anomalies introduced into RC giants by
the He-core flash, which converts a giant at the RGB tip to a He-core burning
giant on the RC.
The Third Data Release of the Galactic Archaeology with the HERMES survey
provides stellar parameters and elemental abundances derived from 678,423
spectra for 588,571 unique stars observed with the HERMES spectrograph at the
Anglo-Australian Telescope between November 2013 and February 2019 (see Buder
et al., 2021, for more detail on the data release). It is also accompanied by
several value-added-catalogues (VAC) providing various other intrinsic,
kinematic and dynamical properties of the observed stars (hereafter, GALAH$+$
DR3).333The GALAH$+$ DR3 data is available at
https://cloud.datacentral.org.au/teamdata/GALAH/public/GALAH_DR3/
GALAH_DR3_main_allspec_v1.fits is one of the two main catalogues of GALAH$+$
DR3 and includes stellar parameters and overall abundances of about 30
chemical species derived for each of the observed spectra. This catalogue also
includes abundances derived for each individual line for all the elements. We
joined this catalogue with the VAC GALAH_DR3_VAC_ages_v1.fits which provides
other stellar parameters like age, mass, distances, luminosity, etc.,
estimated using Bayesian Stellar Parameter Estimation code (BSTEP, Sharma et
al., 2018). The two catalogues are joined using the recommended method of
using the internal ID for each GALAH observation provided in the sobject_id
column. To eliminate multiple entries of a star in the catalogue and to select
the entry with the highest signal-to-noise, we restrict our sample to stars
with flag_repeat = 0. We further restrict our sample to stars with reliable
stellar parameters (flag_sp = 0) and metallicity (flag_fe_h = 0) which lead to
abandoning of stars with $T_{\rm eff}<4100$ K. We also discarded all the stars
with uncertainty in temperature ($\sigma_{T\rm_{eff}}$) and luminosity
($\sigma_{\log(L/{\rm L_{\odot}})}$) of more than 200 K and 0.2, respectively.
This resulted in a sample of about 425 thousand stars. Sample stars are shown
in Figure 1 in the form of the HR diagram. Note that the sample is not mass
limited, and the large scatter in the figure can be attributed in part to the
presence of relatively massive stars (with masses 2$-$5 M⊙) in the sample. For
each sample star, uncertainties in luminosity and temperature are also shown
in the background, however, due to crowding, they are visible only for the
star on the boundaries. From this sample, we select giant stars with
temperature from 4100 to 5600 K and luminosity $\log(L/{\rm L_{\odot}})$ from
0.3 to 3.4 dex. This selection of 159,841 stars is shown within the dotted
green box in Figure 1 (hereafter, the RBS, which stands for the red-giant
branch sample).
Except for Martell et al. (2021), previous discussions of Li abundances in
giants based on the GALAH survey drew on the DR2 catalogue. Of interest to the
present study is the fact that our sample of RBS giants from the DR3 catalogue
is approximately doubled over our previous sample from DR2 (Deepak et al.,
2020), abundance quality flagging is also improved, including the provision of
upper limits and abundances are now provided for 31 rather than 21 elements
previously. In the case of the 335 Li-rich giants found by Deepak & Reddy
(2019) in the DR2 catalogue, Li abundances of common stars in DR2 and DR3 show
only a slight systematic change: the least Li-rich are slightly less Li-rich,
and the super Li-rich giants are slightly more Li-rich. Our assessment is that
the differences between the data releases have not affected principal
conclusions about Li in giants in the RBS collection studied in previous
works. The increase in the RBS sample offers the chance to search for new
properties of Li-rich giants and to explore already covered properties in more
detail.
Figure 1: Sample stars in a HR diagram also showing the selection of red
giants (the RBS) within the green box. For all the stars, uncertainty in
luminosity and temperature are also plotted in the background as red error
bars.
### 2.1 Elemental abundances
The GALAH$+$ DR3 survey provides the elemental abundance ratio [X/Fe]444For
two elements, $X$ and $Y$, with number densities $N_{X}$ and $N_{Y}$, the
elemental abundance ratio [$X/Y$] $\equiv$ $\log$($N_{X}/N_{Y}$)star $-$
$\log$($N_{X}/N_{Y}$)Sun. for up to 30 chemical elements per star with X = Li,
C, O, Na, Mg, Al, Si, K, Ca, Sc, Ti, V, Cr, Mn, Co, Ni, Cu, Zn, Rb, Sr, Y, Zr,
Mo, Ru, Ba, La, Ce, Nd, Sm, and Eu. Abundances for most of the elements were
considered in detail by Deepak et al. (2020) using the DR2 catalogue. The
global analysis of [X/Fe] will not be reexamined here. Previously, the only
detectable difference between [X/Fe] for Li-rich, Li super-rich and normal
giants was that Li super-rich giants were slightly underabundant in [C/Fe] but
see Martell et al. (2021).
For Li, the DR3 catalogue provides the non-LTE correction for the Li abundance
provided by the 6707 Å resonance doublet. We convert the [Li/Fe] using the
non-LTE Li and LTE Fe abundances into A(Li) using A(Li) = [Li/Fe] + [Fe/H] +
A(Li)⊙, where A(Li)⊙ is the solar Li abundance for which we adopt A(Li)⊙ =
1.05 $\pm$ 0.10 dex, which is the value used for the GALAH data (see Buder et
al., 2021, for more details). The DR3 catalogue also provides uncertainties
($\rm\sigma_{[X/Fe]}$ available in corresponding e_X_fe column) and a quality
flag (flag_X_fe) for each of abundance entry. flag_X_fe can have any value
from 0, 1 and 32, where flag 0 means provided abundance is reliable, 1 means
provided abundance is an upper limit, and 32 means no reliable measurement
reported. In this study, we will use flag_X_fe = 0 for elements except Li. In
the case of Li, flag_Li_fe = 0 will be used to identify a Li-enriched star,
while for the Li-normal stars, we will use all three flags to avoid discarding
stars with a relatively lower Li-abundance from the sample. Throughout the
paper, we will specify if we are considering a sample of stars for which DR3
notes a detection of Li or considering the larger sample combining stars with
a detection or an upper limit of the Li 6707 Å line.
## 3 Lithium-rich giants: overview
Figure 2: HR diagram and A(Li) as a function of luminosity for the RBS giants
with masses $M/{\rm M_{\odot}}$ = 1.10 $\pm$ 0.02 and metallicity [Fe/H] =
0.00 $\pm$ 0.10 dex. The left and middle panels show giants with both Li
detections and upper limits with light yellow colour indicating a higher
density of stars. In the right-most panel, giants with Li detections are shown
in red, while giants with Li upper limits are in blue. Figure 3: The HR
diagram for the RBS giants. Colour in the top and bottom panels represents the
linear density and maximum A(Li) abundances, respectively. Note that the
density visualization scale in the top panel is linear, but in Figure 1 scale
is logarithmic. Figure 4: HR diagrams for the RBS giants for Li abundances in
different abundance intervals. All giants shown here have a Li detection
(i.e., flag_Li_fe=0). Figure 5: Li abundance maps along with metallicity and
mass information in colour for the RBS giants in the HR diagram. All the
giants shown here have Li detections (i.e., flag_Li_fe=0). Figure 6: Top row:
Metallicity and mass histograms for giants plotted in Figure 5. Histograms are
shown for different Li abundance intervals, as shown in the keys in the two
panels. Bottom row: Metallicity and mass histograms for entire RBS sample with
$\log(L/{\rm L_{\odot}})>$ 1.5.
Our principal endeavour is to use the GALAH DR3 RBS sample (Figure 1) to (i)
explore closely the Li enrichment attributable to the He-core flash in low
mass red clump giants and (ii) probe more thoroughly than before the Li
enrichment predicted from thermohaline mixing at the luminosity bump on the
RGB. Li-enrichment by the He-core flash was empirically proven by Kumar et al.
(2020) from analysis of the GALAH DR2 Li abundances for giants but theoretical
predictions of Li-enrichment are lacking apart from the recent exploration by
Schwab (2020). In contrast to this observational–theoretical imbalance
regarding Li enrichment in red clump giants, theoretical modelling of Li-
enrichment at the luminosity bump (Eggleton et al., 2008; Lagarde et al.,
2011) may be more advanced than observational efforts.
To provide an intriguing glimpse into the relative roles of the two
evolutionary phases of possible Li enrichment, Figure 2 shows in the left
panel the HR diagram for the narrow range of masses $M=1.10\pm 0.02\ {\rm
M}_{\odot}$ and metallicities [Fe/H]$=0.00\pm 0.10$ shown below to be rich in
Li enriched giants. The middle and right-most panels show the relation between
A(Li) and luminosity. The left and middle panels show giants with both Li
detections and upper limits, while in the right-most panel, giants with Li
detections are shown in red while giants with Li upper limits are in blue.
Gross features of the three panels are readily recognizable. The HR diagram
begins with subgiants at $\log(L/{\rm L_{\odot}})\sim 0.5$ and leads into the
RGB with its tip at $\log(L/{\rm L_{\odot}})\simeq 3.5$. There is a steady
decrease of the Li abundance A(Li) up the RGB. The LB occurs on the RGB at a
luminosity just below that of the RC. At the LB, there is a noticeable
increase in the counts. Superimposed on the RGB in the HR diagram is the RC at
$\log(L/{\rm L_{\odot}})\simeq 1.7$ and this marks a large spread in lithium
abundance up to A(Li) $\sim 4$. Apart from a few outliers at luminosities
above and below the RC’s luminosity, this selection does not indicate Li
enrichment associated with a density increase on the RGB arising from the LB.
If this selection by mass and metallicity from the RBS is representative, the
outstanding message is that Li enrichment among RC giants, as Kumar et al.
(2020) noted, is a striking feature. But Figure 2 also suggests that a search
for Li enrichment associated with the LB may require subtle procedures because
at the mass and metallicity ranges explored in Figure 2 (and at neighbouring
mass and metallicities) Li enrichment from the LB appears not to be a major
process.
In our effort to advance the story of Li-enrichment in giants at the red clump
post-He-core flash stage and in other giants at the luminosity bump on the
RGB, we combine Li abundances from the GALAH DR3 survey with stellar evolution
calculations for low mass stars from the MESA suite of computer programmes.
Our focus on possibilities of Li-enrichment in RC and LB low mass Population I
giants does not mean that Li enrichment cannot occur in giants other than at
the RC or LB phases, as in mass transfer to a star from a Li-rich AGB
companion (again, Li synthesis occurs via 3He) or in Population II giants or
by other mechanisms such as engulfment of planetary bodies (Alexander, 1967).
### 3.1 He-core burning giants - the red clump
In Figure 1, the prominent feature superimposed on the red giant branch (RGB)
at $\log(L/{\rm L_{\odot}})\simeq 1.7$ is the red clump (RC) comprised of
giants burning He in their core. As noted in the Introduction, Kumar et al.
(2020)’s vital contribution was to show that observations demand that all RC
giants in the GALAH survey were enriched in Li to differing degrees relative
to stars at the tip of the RGB, the immediate predecessors of RC giants, that
is Li enrichment in RC giants is ‘ubiquitous’, i.e., even RC giants with the
lowest Li abundance, A(Li) $\simeq-0.5$, are Li enriched. Lithium production
was attributed to the He-core flash occurring at the RGB tip as the star
ignites its He core. Giants with masses less than about 2.25 M⊙ at solar
metallicity [Fe/H] = 0 experience a He-core flash at the tip of the RGB. The
upper mass limit for a He-core flash decreases slightly for metal-poor stars
(Karakas, 2010). The existence of the flash has stimulated theoretical
speculation on the possibility of internal mixing and associated
nucleosynthesis including production of Li from 3He and enrichment of the
surface of the RC giant with Li. Consideration of lithium abundances among RC
stars is the principal concern of our study. (Extension of the RC to hotter
temperatures, as in globular clusters, marks the horizontal branch (HB), but
here we use RC to refer to the entire temperature range.)
Lithium abundances among the RC giants span several dex (Kumar et al., 2020).
A simple map of Li abundance over the HR diagram is provided in Figure 3. The
top panel shows the number density of giants across the RBS. Presence of the
RC is obvious from about 4600 K to, perhaps, 5400 K with a peak ratio of
counts in the mid-RC strip to the counts on either side of about 7. The Li map
in the lower panel is colour-coded according to the Li abundance and makes
clear that the most Li-rich giants are on or maybe evolved from the RC.
More detailed maps of Li abundances are provided in the eight panels of Figure
4 showing HR diagrams for giants with A(Li) abundances in the intervals
$-0.5$-0.0, 0.0-0.5, 0.5-1.0, 1.0-1.5, 1.5-2.0, 2.0-2.5, 2.5-3.0 and $>3.0$.
The full RBS includes stars of mass from about 0.8 M⊙ to 2.5 M⊙. The low Li
abundance panels in Figure 4 clearly show the results of the first dredge-up
(FDU) affecting subgiants and early residents on the RGB, most notably in the
A(Li) = 1.0-1.5 and 1.5-2.0 panels at the luminosity $\log(L/{\rm
L_{\odot}})\sim 0.5$. The limit A(Li) $\geq$ 1.5 with representation across
four panels corresponds approximately to the Li abundance expected following
the first dredge-up, a value which has often been adopted in discussions of
Li-rich giants as the lower boundary for a Li-rich giant. (Note that lithium
post-FDU is further depleted in the lower mass giants – see below.) Li
abundances in these four panels include giants in which clearly Li has been
enriched. Inspection of each of these panels confirms that Li-enriched stars
appear tightly confined to the RC and also to a few stars more luminous and
generally cooler than the RC, i.e., to stars which may have exhausted He-
burning in their core and evolved off the RC. For each of these four samples,
there is a clear gap between the RC giants and the lower luminosity RGB
giants, i.e., all giants which are Li-enriched are RC giants. There is
potential confusion at luminosities above the RC between post-RC (early AGB)
giants and RGB stars up to the tip of the RGB, at $\log(L/{\rm
L_{\odot}})\approx 3.4$. (There may also be a possibility of confusion
provided by Li-rich giants at the luminosity bump on the RGB, an issue
discussed in Section 5). In the panel covering the lowest Li abundances, A(Li)
= $-0.5$ to $-0.0$, no stars are seen on the RC, and the majority are likely
RGB giants with a luminosity below that of the RGB tip. Since further
evolution to the RGB tip may result in even lower Li abundances, RC giants
emerging from the RGB tip without experiencing Li enrichment will have maximum
Li abundances A(Li) $\lesssim 0.0$. But there are very few RC giants with this
low Li abundance and, therefore, Li synthesis at some level from the He-core
flash may occur in all giants transitioning from the RGB to the RC, as
suggested by Kumar et al. (2020).
Consideration of the panels in Figure 4 provides additional information.
First, we explore the metallicity and mass dependence of Li-rich giants (see,
for example, Deepak et al., 2020). The metallicities (colour-coded) show that
Li-rich RC giants are predominantly of near-solar metallicities (Figure 5).
This is not a new result; see, for example, Figures 3 and 4 from Deepak et al.
(2020) and Figure 5 from Martell et al. (2021) but extends earlier results by
showing that the metallicity histograms are only weakly dependent on the Li
abundance. Histograms of the metallicity for RC and (presumed) evolved RC
stars in Figure 6 well illustrate that Li-rich giants are certainly
representatives of the Galactic thin disk and its spread in metallicity.
Another family of histograms (Figure 6, top-right panel) is provided for the
masses of the Li-rich RC giants. As Figure 6 shows, the Li-rich giants have a
narrow distribution peaking about 1.1 M⊙. The histograms extend earlier
results by showing that the mass distribution is independent of Li abundance.
There is perhaps a slight increase in mass from the cool to the warm end of
the RC. Distribution of Li abundances across the RBS by mass shows Li rich RC
giants from 0.8 M⊙ to 2.3 M⊙ with, of course, the most prominent sampling
around 1.1 M⊙ to 1.2 M⊙.
The histograms in the bottom panels of Figure 6 refers to the full sample
covered by the RBS at luminosities $\log(L/{\rm L_{\odot}})>1.5$. Note the
large increase in counts between this sample and the histograms compiled from
selected Li abundance intervals in the upper part of the figure. More
importantly, note the overall similarities in the two sets of histograms. To
first order, this suggests that the frequency of Li enriched RC giants is
approximately independent of mass and metallicity. The two sets of histograms
appear most similar for intermediate Li abundances inviting the idea that
levels of Li enrichment may vary with mass and metallicity.
Close scrutiny of Figure 4 suggests several insights into Li production at the
He-core flash. Most obviously, the RC’s location in the HR diagram is
independent of the Li abundance. For example, the mean luminosity of the RC’s
core at the mean effective temperature of 4775$\pm 100$ K is $\log(L/{\rm
L_{\odot}})$ = 1.73 to within $\pm 0.05$ across the bottom four panels in
Figure 4. The mild increase in luminosity with increasing effective
temperature is independent of Li abundance. Similarly, the RC’s width at a
given effective temperature appears to be constant across these panels with
the exception of a mild increase at the lowest Li abundances, which may be
related to the increase in the sample size at the lowest Li abundances. The
range of effective temperatures spanned by the RC in the panels is independent
of Li abundance above A(Li)$\geq 1.5$. Also, the proportion of giants at
luminosities above the RC is similar for the four panels. These
characteristics about the RC and the Li abundance show that the gross
properties of RC stars appear independent of the He-core flash’s ability to
synthesize Li.
However, stellar evolutionary calculations fail not only to account for Li
enhancement in RC giants but also do not account for their considerable spread
in effective temperature at approximately constant luminosity. Such
calculations predict a 1 M⊙ of about solar metallicity to have approximately
the observed luminosity at around 4500K. These are two failures of standard
calculations, but one supposes that while their resolutions may be related,
they cannot be tightly coupled because Li abundances are independent of a
star’s effective temperature.
It seems remarkable that the vast majority of Li enriched giants have very
similar characteristics, i.e., they are or were He-core burning giants of
about 1 M⊙ and near-solar metallicity. Figure 5 shows that among apparent
exceptions are subgiants ($\log(L/{\rm L_{\odot}})\sim 0.7$ with $T_{\rm
eff}\sim 5000-5600$ K) whose numbers decline with increasing A(Li). These are
in all likelihood normal stars of about 1 M⊙ and near-solar metallicity at
different stages of an incomplete FDU. Figure 5 also shows a few higher mass
stars with a luminosity spanning that of the RC. These too are stars observed
before completion of their FDU. Two Li enriched giants at $\sim$4800 K and at
a luminosity well below that of the RC in A(Li) $\geq 3.0$ panel in Figure 4
arouse our curiosity. Both of these giants are of solar mass. This pair “Gaia
DR2 5507464674226166144” and “Gaia DR2 3340856632069339776” with A(Li) = 3.09
$\pm$ 0.11 and 3.51 $\pm$ 0.11, respectively, deserve detailed study.
In essence, setting aside few stars, our selection of Li enriched stars from
the RBS is dominated by stars of about a solar mass and a solar metallicity.
Figure 7: HR diagram (odd-numbered rows) and A(Li) as a function of luminosity
(even-numbered rows) for the RBS stars selected with mass and metallicity
ranges as mentioned in panels’ titles. In luminosity-A(Li) panels, giants with
Li detections are shown in red, while giants with Li upper limits are shown in
blue. Figure 8: continued Figure 9: HR diagrams and A(Li) as a function of
luminosity for solar mass and solar metallicity models with different
$\alpha\rm_{ML}$ and $\alpha\rm_{th}$. Top and bottom panels show a model‘s
sensitivity to $\alpha\rm_{ML}$ and $\alpha\rm_{th}$, respectively. Dashed
lines show pre-main-sequence evolution, while solid lines show the evolution
from the zero-age main sequence (ZAMS). On each track, the LB’s position is
marked by the green coloured blob. Figure 10: Temporal evolution of Li for
solar mass and solar metallicity models with different $\alpha\rm_{ML}$ and
$\alpha\rm_{th}$. Time is shown relative to the onset of LB. Model’s
sensitivity to both $\alpha\rm_{ML}$ and $\alpha\rm_{th}$ is shown. The top
panel covers evolution from about sub-giant phase to the CHeB phase, while the
bottom panel covers evolution from near RGB on-wards. Evolution up to the RGB
tip is shown as solid lines, while dashed lines show the evolution in the CHeB
phase. On each track, the LB’s position is marked by the green coloured blob.
Timestep in each case varies from about 0.1 to 0.2 million years (Myr) during
the LB, and further decreases as the star ascend the RGB. Figure 11: Top
panel: Predicted evolution of A(Li) as a function of luminosity for solar
metallicity (Z = 0.0143) giants. Masses of 0.8 M⊙ to 2.0 M⊙ are shown, as
labelled. Bottom panel: Predicted evolution of A(Li) as a function of
luminosity for $1{\rm M_{\odot}}$ mass models with metallicities Z = 0.0045,
0.0080, 0.0143, and 0.0254, which are equivalent to [Fe/H] = -0.50, -0.25,
0.00, and 0.25 dex, respectively. For all models, we use $\alpha\rm_{ML}$ =
1.85, $\alpha\rm_{th}$ = 100 and initial surface rotation $\Omega_{\it
int}$/$\Omega_{\it crit}$ = 0. For each of the case, evolution is shown from
the ZAMS to the RGB tip. For all the masses, LBs are marked by green coloured
blobs. The abrupt change in A(Li) occurring at about $\log(L/{\rm
L_{\odot}})\sim 1.5$ is attributable to the onset of thermohaline mixing at
the LB. Figure 12: Observed and predicted HR diagram (top panel) and A(Li) as
a function of luminosity (middle and bottom panels) for RBS giants of masses
1.10 $\pm$ 0.02 M⊙ and metallicity 0.0 $\pm$ 0.1 dex. Predictions are shown
for best fit model with $\alpha_{\rm ML}$ = 1.85 along with $\alpha_{\rm th}$
= 50, 100, 150 and 200. The bottom panel shows luminosity-A(Li) trends after
adjustment to Li abundance level. In A(Li)-luminosity panels, giants with Li
detections are shown in red, while giants with Li upper limits are shown in
blue. Figure 13: Observed and predicted HR diagram (top panel and middle
panel) and A(Li) as a function of luminosity (bottom panel) for RBS giants of
masses 1.10 $\pm$ 0.02 M⊙ and metallicity 0.0 $\pm$ 0.1 dex. Predictions are
shown for best fit model with $\alpha_{\rm ML}$ = 1.85 and $\alpha_{\rm th}$ =
100 along with initial surface rotation $\Omega_{\rm int}/\Omega_{\rm crit}$ =
0.00, 0.05, and 0.10, which are equivalent to initial surface rotation of
about 0, 15, and 30 kms-1, respectively. In A(Li)-luminosity panels, giants
with Li detections are shown in red, while giants with Li upper limits are
shown in blue. Figure 14: Observed and predicted HR diagram (top panel) and
A(Li) as a function of luminosity (middle and bottom panel) for RBS giants of
masses 1.10 $\pm$ 0.02 M⊙ and metallicity -0.50 $\pm$ 0.10 dex. Prediction are
shown for best fit models with $\alpha_{\rm ML}$ = 1.80 along with
$\alpha_{\rm th}$ = 50, 100, 150 and 200. In A(Li)-luminosity panels, giants
with Li detections are shown in red, while giants with Li upper limits are
shown in blue. Figure 15: Observed and predicted HR diagram (top panel) and
A(Li) as a function of luminosity (bottom panel) for RBS giants of masses 1.50
$\pm$ 0.02 M⊙ and metallicity -0.25 $\pm$ 0.10 dex. Prediction are shown for
best fit model with $\alpha_{\rm ML}$ = 1.95 and $\alpha_{\rm th}$ = 100, 200,
300 and 400. The bottom panel shows luminosity-A(Li) trends after adjustment
to Li abundance level. In A(Li)-luminosity panels, giants with Li detections
are shown in red, while giants with Li upper limits are shown in blue.
## 4 MESA modelling for low-mass stars
To further our understanding of Li in giants, we use MESA models to predict
the Li abundance from the main sequence along the RGB to the RC for stars in
the mass and metallicity ranges covered by our RBS. Lithium dilution by the
FDU and lithium evolution to the RGB’s tip is modelled by MESA. The potential
episode of Li enrichment of the atmosphere resulting from the He-core flash is
not included. Also, AGB stars experiencing the third dredge-up and possible Li
enrichment are neither included in selection of giants from our RBS nor in our
use of the MESA modules.
Our initial interest in the MESA predictions for Li abundances focussed on
predicted Li abundances for red giants up to the RGB tip and their comparison
with the observations over the mass and metallicity ranges adequately sampled
by the RBS (Figure 7). Such a comparison is also a resource with which to
search for non-standard episodes of Li enrichment (or destruction) such as may
occur on the RGB or by engulfment of planetary material (Alexander, 1967).
### 4.1 Predicted Li abundances to the RGB tip
In pursuit of the predicted Li abundances in giants, we began with the MESA
model for a solar mass and solar metallicity model assembled by Schwab
(2020).555Schwab’s MESA model is publicly available on Zenodo at
https://doi.org/10.5281/zenodo.3960434. Based on our needs, we made some
changes to Schwab’s inlist. The prescription now includes phenomena like
rotation, semi-convection, etc., along with additional resolution
controls.666The source files for our solar mass and metallicity MESA model
along with the outputs for all the cases studied in this work are publicly
available on Zenodo at https://doi.org/10.5281/zenodo.5086893. For our
calculations, we use MESA version 15140. Convection in MESA is treated using
the mixing length theory (MLT) of convection, and for the current exploration
we use its default formalism from Cox & Giuli (1968). For our models, we use
the MESA’s available recipe for thermohaline mixing adapted from Kippenhahn et
al. (1980). We also use convective premixing (Paxton et al., 2019) which is
specified by the condition use_Ledoux_criterion = .true. and do_conv_premix =
.true. in the code. In MESA models, the strength of the thermohaline mixing is
set by a dimensionless efficiency parameter $\alpha\rm_{th}$ (Paxton et al.,
2013). Thermohaline mixing is initiated when the inversion of the mean
molecular weight creates the LB and the region is formally stable against
convection. Mixing in the interior at this time may result in Li synthesis.
This fresh Li will appear at the giant’s surface if the Li enters the deep
convective envelope. For cases where there is no overlap between the region of
Li synthesis and the convective envelope, the surface Li abundance will be
unaltered at its post-FDU value. Subsequent overlap with the convective
envelope is predicted to generally reduce the surface Li abundance. Another
parameter in MESA modelling is the mixing length ($\alpha\rm_{ML}$) for which
we generally adopt a value of about 1.85 times the pressure scale height.
Semi-convective mixing is also considered with efficiency $\alpha_{\rm sc}$ =
0.1 (Paxton et al., 2013). Our models use the standard MESA pp_and_cno_extras
nuclear network, which includes 25 species and the reactions covering the pp-
chains and CNO-cycles. Our adopted MESA nuclear reaction rates are the same as
used by Schwab (2020). GALAH’s abundances are scaled to the solar composition
from Asplund et al. (2009). We also use the same abundance mixture and related
opacity tables (a09, a09_co and lowT_fa05_a09p with condition
use_Type2_opacities = .true.) from MESA’s kap module. Mass loss is not
considered. A recipe for rotation is included in our models. However, in
general, most of our models are non-rotating, i.e., the initial surface
rotation is set at $\Omega_{\rm int}$/$\Omega_{\rm crit}$ = 0.0, and the
detailed effect of rotation is discussed only for a few cases (see Section
4.1.1). Also, unlike Schwab (2020), our models do not include the ad hoc
flash-induced diffusive mixing in the He core at the RGB tip. (However, if
required, it can be easily turned on in our final prescription also.)
Selection of timestep is very crucial in the evolution of models as it helps
to decide the next timestep and assist in the solution for the new structure
and composition (see Paxton et al., 2011; Lattanzio et al., 2015, for more
details). In MESA it is primarily controlled using parameter
varcontrol_target. In our initial test runs, we used the default minimum value
of varcontrol_target = 10-4 for evolution of solar metallicity models of
masses 0.8 to 2.2 M⊙ in 0.1 M⊙ interval. Thermohaline mixing was present in
all the models except for the 1.8 M⊙ case. We also noticed that adoption of a
slightly higher value (about three times than the default value) of
varcontrol_target resulted in no Li reduction (i.e., missing thermohaline
mixing) in the post-LB phase for masses higher than about 1.2 M⊙ until the
model has already evolved near to the tip of RGB. As a fix to this issue, we
forced smaller timesteps by adopting a lower value of the parameter
time_delta_coeff = 0.75 than its default value of unity. Adoption of this
lower timestep resulted in no significant effect on the Li trends for the
cases where Li reduction near the LB was possible with a larger timestep.
Models are begun in the pre-main sequence phase at the point where the core
first becomes radiative. Approach to zero-age main sequence (ZAMS) is shown by
dotted lines in Figures 9 and 11. Initial Li abundances dependent on
metallicity are consistent with the Li abundances in local stars (Lambert &
Reddy, 2004; Deepak & Reddy, 2020; Randich et al., 2020): A(Li) = 3.57, 3.30,
3.04 and 2.78 for [Fe/H] = $+0.25,0.00,-0.25$ and $-0.50$, respectively. Li
depletion dependent on $\alpha_{\rm ML}$ occurs on the approach to the ZAMS:
for the preferred value $\alpha_{\rm ML}\simeq 1.85$ (see below) depletion is
$\sim 0.3$ and $\sim 0.15$ dex for solar metallicity stars of masses 1.0 and
1.1 M⊙, respectively (Table 1). Evolution is followed along the RGB and
through the He-core flash at the culmination of the RGB and to He-burning in
the RC CHeB giant. We evolve all our models until the central mass fraction of
4He drops below 10-6 to conclude the CHeB phase.
We first study the sensitivity of the model predictions to $\alpha\rm_{ML}$
and $\alpha\rm_{th}$. We evolved 1 M⊙ and solar metallicity models with
$\alpha\rm_{ML}$ = 1.00, 1.85, and 2.70, while keeping $\alpha\rm_{th}$ = 100.
For $\alpha\rm_{th}$, we evolved 1 M⊙ and solar metallicity models with
$\alpha\rm_{th}$ = 50, 100, and 200, while keeping $\alpha\rm_{ML}$ fixed to
1.85. The HR diagram and A(Li) trends as a function of luminosity for all
these cases are shown in Figure 9 from the ZAMS through the RGB tip to the end
of core-He burning. The top panels show the effect of $\alpha\rm_{ML}$, and
the bottom panels illustrate sensitivity to $\alpha\rm_{th}$. An increase in
$\alpha\rm_{ML}$ leads to the expected evolutionary track shifts to higher
temperatures. Higher $\alpha\rm_{ML}$ also leads to increased depletion of Li
during the FDU. The temporal evolution of Li in our model stars is shown in
Figure 10 which suggests that adoption of higher $\alpha\rm_{ML}$ leads to
increased as well as prolonged depletion of Li as the star ascends the RGB. A
closer look at Figure 9 and 10 also suggests that for a given set of input
parameters, models with lower $\alpha_{\rm ML}$ experience slightly delayed Li
reduction at the LB in comparison to models with higher $\alpha_{\rm ML}$.
Solar mass and metallicity models with $\alpha_{\rm ML}$ = 1.00, 1.85, and
2.70 starts experiencing post-LB Li reduction after about 29, 17, and 9 Myrs,
respectively, from the onset of LB. However, for all three cases, the actual
timestep during LB’s evolution ranges from about 0.1 to 0.2 million years
(Myr) and further decreases as the star evolve towards the tip of RGB. These
timesteps are very small compared to the duration of LB and delays in Li
reduction. The delay in the post-LB Li reduction is sensitive to the timestep
and decreases with a decrease in the timestep. A test run with smaller
timestep by adopting time_delta_coeff = 0.50 and 0.25 (which resulted in a
timestep of about 0.1 and 0.05 Myr, respectively) for a solar mass model with
$\alpha_{\rm ML}$ = 1.00 resulted in delays of about 8 and 6 Myr,
respectively, along with almost tripling the computation time. It may be of
interest to note that the duration of the LB depends on the assumed value of
$\alpha_{\rm ML}$ and is longer for smaller values of $\alpha_{\rm ML}$.
Similar is the case for overall age of the stars i.e. stars with lower
$\alpha_{\rm ML}$ evolves more slowly (Figure 10). The evolutionary tracks are
insensitive to $\alpha_{\rm th}$, but Li depletion commencing at the LB is, as
anticipated, dependent on $\alpha_{\rm th}$.
To obtain model predictions for various masses, we evolve solar metallicity
models stars of masses 0.8 to 1.6 M⊙ in steps of 0.1 M⊙ and 1.6 to 2.0 in
steps of 0.2 M⊙ from the ZAMS to the RGB tip. The role of metallicity in the
evolution of A(Li) during the RGB was explored by 1 M⊙ models with
metallicities Z = 0.0045, 0.0080, 0.0143, and 0.0254, which are equivalent to
[Fe/H] = -0.50, -0.25, 0.00 and 0.25 dex, respectively. For all models, we
used $\alpha\rm_{ML}$ = 1.85, $\alpha\rm_{th}$ = 100 and initial surface
rotation $\Omega_{\rm int}$/$\Omega_{\rm crit}$ = 0. Changes in the Li
abundance are shown in the top panel of Figure 11 for the series of solar
metallicity models. Metallicity dependence of the Li abundance evolution is
shown in the bottom panel of Figure 11 for 1 M⊙ models. The initial
temperature, luminosity and $\it A$(Li) along with their values at ZAMS, at
the onset of LB and at RGB tip for models in Figure 11 are tabulated in Table
1.
Conditions promoting thermohaline mixing begin at a luminosity around the LB.
Surface Li abundance is affected by Li destruction and formation in the
interior and by the connection between the interior and the outer convective
envelope. Prescriptions for surface Li on the RGB predict in general an
appreciable loss of Li beginning around the LB’s luminosity, see examples in
Figures 9 and 11. A decline of Li on the ascent of the RGB results when the
convective envelope taps the region of Li synthesis. The predicted Li
abundance is necessarily sensitive to the detailed modelling of thermohaline
mixing, as well as stellar mass and composition. Lattanzio et al. (2015)
discuss Li abundance predictions for a 1.25 M⊙ solar metallicity model given
by five different codes. Across these codes, reduction of the Li abundance
post-LB varies considerably for comparable mixing coefficients. An increase in
A(Li) at luminosities just below the RGB tip is predicted under some
conditions for some codes. (Late (small) increases in Li abundance are seen in
Figure 9 for the 0.8 M⊙ model with [Fe/H] = 0 and the 1.0 M⊙ models with
[Fe/H] = 0.) Such late small increases of Li will not be detectable in DR3 and
similar surveys of K giants. In Figure 9 evolution of A(Li) is continued
beyond the He-core flash to the termination of He core burning. It is at the
He-core flash followed by a slight Li decrease as a RC giant. (Li synthesis at
the He-core flash is not included in our models.)
A goal of our study is to confront the MESA models with the HR diagrams and
luminosity-A(Li) trends displayed in Figure 7 where in all panels, we
distinguish between Li abundance measurements and upper limits. Within the
constraints of MESA models, it is apparent that the value of $\alpha_{\rm ML}$
may be determined independently of $\alpha_{\rm th}$ from an observed HR
diagram. Then, determination of the best fit to the luminosity-A(Li) trend for
a given mass-metallicity combination may proceed by adjustment of $\alpha_{\rm
th}$. But, perhaps, the most striking observation is that the luminosity-A(Li)
trend is very similar across all (M⊙,[Fe/H]) pairings well sampled by the RBS,
but the MESA predictions in Figure 9 and 11 show a diverse set of
luminosity-A(Li) trends. Ahead of detailed comparisons, it is apparent that
the fits of MESA Li predictions to observations in Figure 7 will range in
quality. Notably lacking among the observed trends in Figure 7 are examples
matching predictions where the Li abundance is effectively constant over the
RGB from completion of the FDU through the LB’s luminosity to the RGB’s
tip.777The mass and metallicity trends shown in Figure 11 are qualitatively
similar to those provided by Charbonnel & Lagarde (2010) from the STAREVOL-
GENEVA Code.
Investigation of mixing on the RGB is well worthy of study because
observations of Li on the RGB end at a luminosity of $\log(L/{\rm
L_{\odot}})\sim 2.2$ well short of the RGB tip at $\log(L/{\rm L_{\odot}})\sim
3.4$. Thus, inference of the Li abundance at the RGB’s tip involves a
substantial extrapolation from the known Li abundances. It is the Li abundance
at the tip which sets the low bound to the Li abundance among RC clump giants,
i.e., the RC giants which did not experience Li addition as a result of the
He-core flash.
#### 4.1.1 Constraining thermohaline mixing from Li predictions for red
giants
To investigate thermohaline mixing in red giants from the MESA models, we
start with the sample of giants with mass $M/{\rm M_{\odot}}$ = 1.10 $\pm$
0.02 and [Fe/H] = 0.00 $\pm$ 0.10. As discussed above, the evolutionary track
is sensitive to $\alpha_{\rm ML}$ and insensitive to $\alpha_{\rm th}$. A
model with $\alpha_{\rm ML}=1.85\pm 0.05$ provides a good match to the
observations (Figure 12). Li abundances for $\alpha_{\rm th}=50,100,150$ and
200 were predicted. Predictions and observations are compared in the bottom
panels of Figure 12.
Predicted A(Li) throughout the observed luminosity range is higher than the
observed value. This offset, apparently independent of luminosity, may be
negated by subtracting 0.6 dex from observed Li abundances, as achieved in the
bottom panel of Figure 12 showing a fair fit between predicted and observed Li
abundances. The short segment of the RGB above the LB with observed Li
abundances (after application of the offset) is fitted with $\alpha_{\rm
th}\sim 100\pm 50$. When predictions of Li abundances are extrapolated to the
RGB’s tip, a Li abundance A(Li) $\simeq-1.0\pm 0.3$ is obtained. Setting aside
the question of the $-0.6$ offset in Li abundance, the final match of
prediction to observation might seem acceptable. Two points deserve attention.
First, the decline of observed Li abundances on the RGB does not clearly show
the abrupt change predicted to occur at the LB’s luminosity. However, as
starkly shown by the bottom panel in the figure, the RGB at luminosities below
the LB is poorly represented by giants with Li detections and, thus, the true
slope may be ill determined.
Origins of the above Li abundance offset may be imagined. (Also, the presence
of Li upper limits among similar giants with Li abundances shows processes
actively depleting Li are common among these stars, as giants or in earlier
evolutionary phases.) Li in a main-sequence star survives in a very thin skin
at the surface. Should depletion occur via the skin’s base (prior to the FDU),
the Li abundance post-FDU will be reduced from its predicted value. Unless the
Li-containing skin is convective, the depletion of Li will not be seen in the
main sequence star yet revealed in the post-FDU giant. Similarly, mass loss
from the surface of the main-sequence star may not affect its Li surface
abundance but will be revealed by the FDU. There are examples of main-sequence
stars with reduced Li abundances. The Sun is a notable example with an
abundance of about 2.3 dex less than its initial abundance. Surely, the most
extraordinary examples belong to the so-called Li-dip discovered by Boesgaard
& Tripicco (1986) in the Hyades at a mass of $M\sim 1.4{\rm M_{\odot}}$. This
dip shifts to lower masses with decreasing metallicity (Balachandran, 1990).
Chen et al. (2001) map this decline: at [Fe/H] = 0.0 and -0.50, the mass
corresponding to the dip is $M\simeq$ 1.40 M⊙ and 1.26 M⊙, respectively,
suggesting that the Li-dip is not the primary origin for the Li offset
required in Figure 12. Severe, even moderate, reductions of Li abundance in
main sequence stars result in non-detection of Li in giants such that
reductions cannot be quantified. It is curious that when well-populated with
Li abundance measurements, the A(Li) - luminosity relations for the RGB in
Figure 7 have a smooth outer boundary suggesting uniform initial Li abundances
and effectively identical treatment of Li from the main sequence to the post-
FDU giant. The other boundary defined by Li abundance limits from non-
detection of the Li i 6707 Å line may be ragged and arises from stars which
experienced Li loss prior to the RGB.888 It may be worth remarking that Li
with its survival limited to a main-sequence star’s outer skin differs greatly
in its response to the FDU from the C, N and O elemental and isotopic
abundances traditionally identified as measures of the FDU.
To understand the effect of rotation, we evolved a 1.1 M⊙ solar metallicity
model with initial surface rotation $\Omega_{\rm int}$/$\Omega_{\rm crit}$ =
0.00, 0.05, and 0.10 which are equivalent to initial surface rotation of about
0, 15, and 30 kms-1, respectively. A comparison of predictions with the
observational data is shown in Figure 13. Predicted HRDs for both rotating and
non-rotating model are similar except for slight disagreements at the RC and
the RGB’s tip. (HRDs of the rotating and non-rotating models also disagree at
the main-sequence, but not shown here.) Rotating models show slightly lower
luminosity and slightly higher temperature for the RC star compared to non-
rotating models. This behaviour of rotating models at the RC is in contrast to
observations and suggests that giants post-He-Core-flash have small or no
rotation. The RGB’s tip for rotating models lies at a slightly lower
luminosity compared to non-rotating models (see the bottom panel of Figure
13). The figure suggests that rotating models experience smaller Li depletion
during the FDU compared to non-rotating models. However, MESA’s implementation
of rotation-induced mixing is but one realization of this complex physical
process - see Charbonnel & Lagarde (2010) for an alternative prescription and
different Li depletions.
Other samples in Figure 7 from the RBS were studied to compare predicted with
observed HR diagrams and the luminosity-A(Li) trends. Predictions comparable
to observations can be produced with small acceptable changes to $\alpha_{\rm
ML}$ and $\alpha_{\rm th}$. As a test of a fit to a metal-poor selection, we
consider the sample for ($M$,[Fe/H]) = ($1.10\pm 0.02{\rm M_{\odot}},-0.50\pm
0.10$) - see Figure 14. The HR diagram is well fit with a slightly lower
$\alpha_{\rm ML}=1.80\pm 0.05$. For this sample, the predicted luminosity of
the LB falls slightly above the predicted/observed luminosity of the RC. The
LB is not obvious from the observed HR diagram because it is merged with the
RC. The middle panel of Figure 14 shows a mildly satisfactory match between
observed and predicted A(Li)-luminosity trends. An additional offset of about
$-0.3$ to the predicted A(Li) is required to get a better fit. The predicted
Li-luminosity trend with a sharp break in slope at the LB’s luminosity seems a
poor match to the observed trend, as set by giants with Li abundance
detections. In contrast to Figure 12, Li detections span the RGB from its base
to the limiting luminosity. The short segment of the RGB above the LB with
observed Li abundances (after application of the offset) is fitted with a
slightly higher $\alpha_{th}$ $\sim$ 150 $\pm$ 50\. Extrapolation of predicted
Li trends to the RGB tip gives a Li abundance A(Li) $\simeq$ -0.8 $\pm$ 0.3 at
the RGB tip.
As a representative sample of higher mass giants, we consider the ($M$,[Fe/H])
= ($1.50\pm 0.02{\rm M_{\odot}},-0.25\pm 0.10$) selection - see Figure 15. The
rather sparse HRD is again well fit with $\alpha_{\rm ML}$ = 1.95 $\pm$ 0.05.
The LB is not prominent in the observed HR diagram. The predicted luminosity
of the LB falls slightly above the predicted/observed luminosity of the RC.
Due to the small sample size, especially very few giants with a Li detection,
the observed A(Li) trend is not very well defined by giants more luminous than
the LB. An offset of $-0.2$ to the Li abundance level is required to match the
luminosity-(Li) trend for giants less luminous than the RC. The predicted
trend Li abundance with luminosity again changes slope at the LB’s luminosity
and may not match the smoother observed trend. The segment of giants more
luminous than the RC suggests $\alpha_{\rm th}\simeq$ 300 $\pm$ 100, a value
almost three times higher than required for 1.1 M⊙ stars. Extrapolation of
predicted Li trends to the RGB’s tip gives a Li abundance A(Li) $\simeq$
$-0.1\pm 0.3$.
Final results from detailed modeling for the above three (M, [Fe/H]) cases are
summarized in Table 2.
Figure 16: Left panel: HR diagram showing tracks for RGB phase for solar
metallicity (Z = 0.0143) models of masses from 0.8 to 2.2 in steps of 0.2 M⊙.
Right panel: RGB tracks for solar mass models of metallicity Z = 0.0045,
0.0080, 0.0143, and 0.0254, which is equivalent to [Fe/H] = -0.50, -0.25, 0.00
and 0.25, respectively. Shown tracks are for models with $\alpha_{\rm ML}$ =
1.85 and $\alpha_{\rm th}$ = 100. The luminous end for each of the track marks
the position of RGB tip. Also, in each of the case, green blob mark the
position of LB.
### 4.2 MESA predictions for the red clump
Our MESA models include the He-core flash but not the possibility of
accompanying Li synthesis. In low-mass stars, there is not a single flash, but
a series of He-core flashes of declining severity as He-core burning is
established (Thomas, 1967; Despain, 1981; Bildsten et al., 2012; Deepak &
Lambert, 2021) and the star’s luminosity drops from its value at the RGB tip
to the value of the RC.999Iben (2013) provides a detailed discussion of a
solar metallicity 1 M⊙ model through its He-core flashes to the third dredge-
up on the AGB. This decline occurs so quickly that very few stars are expected
to be observed in this approach to the RC. During the RC phase, the luminosity
and effective temperature are little changed for about 100 Myr for a giant of
about 1 M⊙. As He-core burning transitions to He-shell burning with continuing
H-shell burning, luminosity increases and effective temperature decreases in
the transitions from the RC to the AGB. RC evolution according to MESA models
for 0.8, 1.0 and 1.2 M⊙ stars of solar metallicity is shown in Figure 17.
Initial lithium abundances for these RC models are assumed to be the low
values achieved at the tip of the RGB. A small additional decrease in Li
abundance is predicted on the RC and in the initial ascent of the AGB (Figure
10 and 17). An increase of the Li abundance resulting from the third dredge-up
is not expected until the AGB model reaches a luminosity exceeding that of the
RGB tip and an effective temperature of 3000 K or cooler.
Notably, observations and MESA predictions disagree sharply in two regards.
First, the luminosity and effective temperature of the post He-core flash RC
models are in good agreement with observations for the coolest examples in
Figure 5 but cannot account for the RC’s extension to higher temperatures
(Figure 3 and 4) at nearly a constant luminosity and nearly the same mass. For
example, the 1 M⊙ model in Figure 17 at 10 Myr has the luminosity $\log(L/{\rm
L_{\odot}})$ = 1.70 and $T\rm_{eff}$ = 4518 K which compares favourably with
the values for the coolest RC stars in Figure 5. Second, these standard MESA
models necessarily have the very low A(Li) of their RGB progenitors and cannot
account for the highly variable Li enrichment of all - ‘ubiquitous’ enrichment
(Kumar et al., 2020) – RC stars across their temperature range. One may
speculate that these two disagreements between observations and models may
have a single explanation or at least related explanations.
Lithium synthesis at the He core flash was recently investigated by Schwab
(2020) using a 1 M⊙ solar metallicity MESA model. He demonstrated that the
high Li abundance in RC stars may be explained by an event that occurs at the
time of the first and strongest He-core flash. Three versions of Li synthesis
and mixing events shown in Schwab’s Figure 4 give final Li abundances from
about 3.7 to 1.0. The mixing introduced by Schwab is diffusive and defined by
a mixing coefficient. Another parameter is the duration of the mixing event.
Li may also be destroyed in a subsequent weaker He-core flash. Schwab remarks
that excitation of internal gravity waves by turbulent convection during the
core flash may provide the basis for the mixing, which he describes by an ad
hoc mixing coefficient. These exploratory calculations do suggest that the He-
core flash and attendant Li synthesis may account for varying degrees of Li
enrichment seen in RC giants (Figure 4).
## 5 Lithium enrichment at the luminosity bump?
Lithium synthesis at the He-core flash results in an abundance increase
between a star at the RGB tip and the RC. In a parallel endeavour, one may
search for Li abundance increases in stars at or just following the LB. Theory
and observation, as discussed above, anticipate that Li abundance changes
resulting from the LB are expected to be revealed subsequently as the star
ascends the RGB. In general, the Li abundance post-LB is predicted to be
unchanged or to decline steadily up the RGB. Nonetheless, the history of Li
among giants suggests that one may usefully see if exceptional Li-rich giants
are to be found at about the luminosity of the LB. MESA predictions for the
LB’s location in the HR diagram are shown in Figure 16 for solar composition
models with $\alpha_{\rm ML}$ = 1.8. This figure shows that the LB’s
luminosity increases with stellar mass and crosses the RC’s luminosity at
$M\simeq 1.6{\rm M_{\odot}}$. The LB occurs at somewhat higher luminosities
and higher temperatures for metal-poor stars (Figure 16). The similarity in
the luminosity of RC and LB for low mass giants potentially complicates the
identity of Li enriched stars in a luminosity-A(Li) diagram.
Attribution of Li enrichment directly to the LB seems possible for those
(mass, metallicity) pairings providing a LB luminosity operationally
distinguishable by more than 0.2 dex from the RC’s luminosity. At [Fe/H] = 0,
this condition corresponds to $M\leq 1.2{\rm M}_{\odot}$ for a LB luminosity
less than that of the RC or $M\leq 1.0{\rm M}_{\odot}$ for a LB luminosity
greater than that of the RC. At [Fe/H] $=-0.25$ these limits are reduced by
about $0.2{\rm M_{\odot}}$. Figure 7 shows that only the search among the
samples with a luminosity less than that of the RC is practical for the DR3
survey; far too few giants with $M\geq 1.6{\rm M}_{\odot}$ are included in
DR3.
One suitable sample is shown in Figure 2. The distinct locations of the RC and
LB are well shown. Surely, the outstanding feature in the luminosity- A(Li)
diagram is the wide range of Li abundances at the RC’s luminosity. At the LB’s
luminosity, there are no giants with an obvious Li enrichment; most of the Li
abundances are upper limits. Several panels in Figure 7 confirm this result
for other (mass,metallicity) combinations.
Figure 7 shows that for all (mass, metallicity) combinations, the Li abundance
post-LB declines at a similar rate as the giant evolves beyond the LB. If
significant Li enrichment were to occur at the LB, it may vary even between
stars with identical properties, as it does among RC giants. Then, the width
of the A(Li) – $\log(L/{\rm L_{\odot}})$ trend on the RGB above the LB would
be increased relative to the width below the LB. In principle, this test of Li
enrichment at the LB seems applicable to all (mass, metallicity) combinations
provided that the RGB is well populated. Unfortunately, the presence of many
upper limits to the Li abundances impairs the test. In addition, at
luminosities above the RC and LB, Li enriched RGB stars are difficult to
isolate from evolved Li enriched RC giants.
Our conclusion that Li-rich giants do not occur on the RGB at the LB differs
from remarks made by Martell et al. (2021) from their analysis of Li-rich
giants in the GALAH DR3 survey. They conclude that lithium enrichment is not
limited to RC giants but is also present among an appreciable number of RGB
giants: “red clump stars are 2.5 times as likely to be lithium-rich as red
giant branch stars.” This disagreement with our analysis may arise from
Martell et al.’s use of Kiel diagrams ($T_{\rm eff},\log g)$ where giants are
assigned to either the RGB or the RC using the Bayesian probability estimates
provided in column is_redclump_bstep. Giants with is_redclump_bstep $\geq$ 0.5
and WISE $W_{\rm 2}$ absolute magnitude in the range $|W_{\rm 2}$ $+$ 1.63$|$
$\leq$ 0.80 are classified as RC giants, while giants with is_redclump_bstep
$<$ 0.5 or absolute magnitude outside the range $|W_{\rm 2}$ $+$ 1.63$|$
$\leq$ 0.80 are classified as RGB giants. This results in a classification of
all the luminous giants with $W_{\rm 2}$ $<$ -2.43 or $log(L/{\rm L_{\odot}})$
$\gtrapprox$ 2.1 as RGB giants, which in actuality are expected to be a
mixture of RGB, early CHeB, and post-RC giants. The use of Bayesian estimates,
which vary with luminosity, effective temperature and metallicity, appear to
have introduced some intermixing of the RGB and RC populations. We prefer to
assign evolutionary phases using a giant’s location in the HR diagram
($\log(L/{\rm L_{\odot}}),T_{\rm eff}$) with respect to the appropriate
evolutionary track. As noted above, there are regions in the HR diagram where
ambiguities may arise depending on the accuracy of the parameters used to
place a giant on the HR diagram, for example, at the confluence of the RC and
the RGB.
In summary, Li enrichment at the LB is not detected in the giants well
represented by DR3. An upper limit to the Li abundance increase of about 0.3
dex is suggested by visual inspection of Figure 7. This may provide a novel
challenge to the theoretical modelling of thermohaline mixing at the LB.
Figure 17: Variation of luminosity, Teff and A(Li) as a function of time (in
million years) during the CHeB phase for solar-metallicity model stars of
masses 0.8, 1.0 and 1.2 M⊙ with $\alpha_{\rm ML}$ = 1.85 and $\alpha_{\rm th}$
= 100. For the time axis, a base-10 log scale is used to highlight the fact
that CHeB giants spend almost their entire life in a relatively stable state
except for the brief He-flashing region, which lasts for about 1-2 Myr from
the He-core flash. In the CHeB phase, giants also experience additional Li
depletion. This Li depletion starts approximately about 5 Myr after the He-
core flash for a solar mass and solar metallicity star.
## 6 Compositions of red clump giants
The chemical composition of Li enriched giants has been studied by Deepak et
al. (2020) to a great extent based on the GALAH DR2 and then by Martell et al.
(2021) based on data from the GALAH DR3. The conclusion is clear: compositions
of Li-rich giants are not distinguishable from other giants. Here, attention
is returned specifically to the RC giants. All of these stars have been
affected by the He-core flash, which led to alterations of the Li abundance by
an amount varying from small to substantial. In examination of the [X/Fe]
versus [Fe/H] plots, the emphasis here is on elements affected by He-burning
and $s$-processing. The earlier conclusion about compositions of Li enriched
and normal giants is confirmed.
Unfortunately, elements provided by the GALAH survey are not well suited to
probing surface contamination by products of He burning mixed to the surface
during the short-lived episode of He-core flashes or the longer period of He-
core burning. Just 229 giants in the RC luminosity strip (with $\log(L/\rm
L_{\odot})$ of 1.6-1.9) have a determined C abundance. Nitrogen, which might
be affected via associated H-burning CN-cycling, is not provided in the GALAH
survey. Oxygen, which is well represented in this survey, shows the [O/Fe]
versus [Fe/H] relation expected for thin-disk giants with no alterations
dependent on Li abundance. This suggests that, although internal He-burning
may have altered surface Li abundances, oxygen which is possibly a direct
product of He-burning has not been altered.
Coupling of $s$-processing with He-burning is a theoretical possibility. GALAH
provides 11 elements from Rb to Sm as at least partial monitors of an
$s$-process and Eu as a measure of the $r$-process. For four elements - Rb,
Sr, Mo and Ru - abundances are sparsely reported. Among the RC giants, there
is no evidence that abundances of these 11 elements differ according to their
Li abundances or differ at all from the [X/Fe] versus [Fe/H] trends expected
for the thin disk.101010The [X/Fe] versus [Fe/H] for X = Sr, Mo and Ru show an
odd bifurcation. A majority of the measurements provide [X/Fe] $\simeq 0$, as
found for other $s$-process dominated elements, but a minority of giants
suggest a second trend of increasing [X/Fe] with decreasing [Fe/H] attaining
[X/Fe] $\sim+1$ at [Fe/H] $\sim-0.4$. Is this exceptional behaviour a
reflection of systematic errors, perhaps linked to misidentification of lines?
In brief, elements examined across the RBS sample of RC giants show no
abundance anomalies except for Li, but this conclusion can not yet be extended
to principal products of He-burning (C and N and their isotopes) for lack of
observations.
## 7 Concluding remarks
Since their discovery in 1982 by Wallerstein & Sneden, Li-rich giants have
occasioned considerable attention by both observers and theoreticians
searching for the evolutionary phases at which a stellar atmosphere’s Li
abundance may be increased by up to several dex. For red giants considered as
single stars, the search for internal lithium production has been directed at
two evolutionary phases in low mass giants (say, $M\lesssim 2{\rm
M_{\odot}}$): the luminosity bump on the RGB and the He-core flash. It is now
clear that an overwhelming majority of lithium-rich red giants are He-core
burning giants, members of the so-called red clump, featuring lithium
synthesized by the Cameron-Fowler mechanism during (presumably) the He-core
flashes occurring in the RC star’s predecessor at the RGB’s tip. Detectable Li
synthesis is not detectable at the luminosity bump. The preeminence of the
idea that the majority of lithium-rich giants are He-core burners ends a long
debate about the relative roles of the red clump and the luminosity bump. (The
role of binary stars in creating Li-rich giants has been examined. Casey et
al. (2019) proposed that Li was synthesized in a RGB giant when Li synthesis
was triggered by perturbations arising from a companion star. The observation
that the binary frequency for Li-rich giants resembles that for Li normal
giants Jorissen et al. (2020) is a challenge to the universal viability of
this binary hypothesis.)
The conclusion about the dominance of the He-core flash in creating Li-rich
giants was strengthened by Kumar et al.’s (2020) use of the GALAH DR2 survey.
Kumar et al. noted that the population of the Li-rich population was at the
luminosity of the red clump with Li abundances A(Li) spanning from about
$-0.5$ to $+3.7$. Red clump residents have evolved from the RGB and most
likely from a star at the RGB’s luminosity tip. Thus, the lithium abundance at
the RGB’s tip provides the baseline for a clump giant. Kumar et al. inferred
by extrapolation of the GALAH abundances for the RGB that the abundance at the
tip was A(Li) $\sim-0.9$. Thus, this baseline and the Li abundance range
$-0.5$ to $+3.7$ for RC giants led Kumar et al. to a conclusion that all RC
residents were enriched in lithium: across the range of Li abundances, the
enrichment is slight to dramatic. Kumar et al. referred to lithium enrichment
as ‘ubiquitous’ among red clump giants. Our probing of red clump giants in the
GALAH DR3 endorses Kumar et al.’s conclusions about the ubiquity of lithium
enrichment among RC giants.
MESA models do not fully account for the run of Li abundance along the RGB.
Across the sample examined here, the maximum predicted Li abundances of RGB
giants are about 0.2 - 0.6 dex greater than observed, which suggests a loss of
Li within the main sequence star prior to the RGB. Loss of Li is also
suggested by the absence of detectable Li in many RGB giants. All mass-
metallicity samples of RGB giants exhibit a smooth decrease of Li abundance
with increasing luminosity, but the MESA predictions provide for a steeper
decline of Li abundance at luminosities above that of the LB. A possible
resolution of this discrepancy is to invoke missing ‘physics’ in the models,
such as rotation-induced mixing.
Our analysis shows that the lithium enriched population peaks at a mass
$M\simeq$ 1.1 M⊙ and metallicity [Fe/H] $\simeq-0.2$. These values reflect the
mass and metallicity distributions of the GALAH population as a whole. The
lithium richest and lithium poorest stars in a mass-metallicity combination
may appear at the same effective temperature and at the same luminosity. The
ubiquity of Li enrichment among RC giants demands redefinition and restatement
of the frequency of Li-rich giants among a population, previously stated to be
at the few per cent level. If ubiquity is the correct description for RC
giants, all such giants with mass $M\simeq$ 1.1 M⊙ and metallicity [Fe/H]
$\simeq-0.2$ have experienced the addition of fresh Li to their atmosphere.
Future investigations exploring red giants beyond the masses and metallicities
well covered by the RBS are awaited with interest.
MESA models fail to account for observations of RC giants. (Li synthesis was
not included in our models.) There is an observed spread in effective
temperature (and Li abundance) for giants of a given mass and metallicity. But
models predict He-core burning to occur at essentially a single effective
temperature before the star evolves to the AGB phase at lower temperatures and
higher luminosity. This observed spread in effective temperature and an
uncorrelated spread in Li abundance suggests that the sequence of He-core
flashes initiating the RC phase at the RGB’s tip may result in internal
adjustments not considered by MESA (and similar models). The primary search
for the origin for the effective temperature spread among red clump giants
must be led by theoreticians. This issue, however, may be explored by
examining compositions of red clump giants across their full spread in
effective temperature. No abundance anomalies are shown by the GALAH
compositions, but key nuclides are very incompletely reported (as C) or not
reported at all (as N).
The luminosity bump has been offered as an opportunity to enhance the surface
lithium abundance. Guided by the MESA predictions on the bump’s location, we
searched for a direct association with Li enriched giants. Across the mass and
metallicity combinations providing HRDs with well-populated RGBs, we found no
evidence for the presence of Li enriched giants at the LB: an upper limit of
0.3 dex to the increase in Li abundance is set. This provides an interesting
limit with which to test calculations of lithium synthesis occurring at the
luminosity bump. Such synthesis does directly affect the Li abundance
development on the RGB above the LB, where without obvious exception, the Li
abundances are less than at pre-LB RGB values.
The saga of the origins of lithium enrichment in low mass red giants has now
transitioned to the challenge of providing a clear picture of Li synthesis at
the He-core flashes in luminous RGB giants. Clarification of the
identification of Li enriched giants is a notable achievement, but questions
concerning the quantitative accounting for giants’ lithium abundances remain,
most especially difficult questions about the He-core flashes and their effect
on the structure of red clump giants which may in many cases depart markedly
from the clump giants described by standard evolution calculations. Although
direct observations of Li abundance at the LB have not been found,
observations of Li and light nuclides (C, N and O) affected by thermohaline
and other forms of mixing along the full RGB deserve pursuit before the
understanding of RGB giants may be deemed complete. Undoubtedly, as the
history of lithium in observational astronomy reminds us, other puzzles about
the lithium abundances among giants as single and binary stars will occur
unexpectedly.
## Acknowledgments
We are indebted to the anonymous reviewer and to John Lattanzio for helpful
correspondence and, most especially, to the GALAH team and their publication
of the data release DR3. Without the catalogues associated with DR3, our
analysis would have been impossible. A similar debt is due to Bill Paxton and
his colleagues for developing the MESA suite of programmes. This work has also
made use of NASA’s Astrophysics Data System. The data for the GALAH survey are
acquired through the Australian Astronomical Observatory, under programs:
A/2013B/13 (The GALAH pilot survey); A/2014A/25, A/2015A/19, A2017A/18 (The
GALAH survey phase 1); A2018A/18 (Open clusters with HERMES); A2019A/1
(Hierarchical star formation in Ori OB1); A2019A/15 (The GALAH survey phase
2); A/2015B/19, A/2016A/22, A/2016B/10, A/2017B/16, A/2018B/15 (The HERMES-
TESS program); and A/2015A/3, A/2015B/1, A/2015B/19, A/2016A/22, A/2016B/12,
A/2017A/14 (The HERMES K2-follow-up program). We acknowledge the traditional
owners of the land on which the AAT stands, the Gamilaraay people, and pay our
respects to elders past and present. This paper includes data that has been
provided by AAO Data Central (https://datacentral.org.au/). We also
acknowledge using the Delphinus and Fornax server at IIA, Bengaluru, to evolve
our model stars.
## ORCID iDs
Deepak: https://orcid.org/0000-0003-2048-9870
David L. Lambert: https://orcid.org/0000-0003-1814-3379
## Data Availability
The observational data underlying this article are publicly available in the
Data Central at
https://cloud.datacentral.org.au/teamdata/GALAH/public/GALAH_DR3/. The source
files for our solar mass and metallicity MESA model along with the outputs for
all the cases studied in this work are publicly available on Zenodo at
https://doi.org/10.5281/zenodo.5086893.
## References
* Alexander (1967) Alexander J. B., 1967, The Observatory, 87, 238
* Asplund et al. (2009) Asplund M., Grevesse N., Sauval A. J., Scott P., 2009, ARA&A, 47, 481
* Balachandran (1990) Balachandran S., 1990, ApJ, 354, 310
* Bildsten et al. (2012) Bildsten L., Paxton B., Moore K., Macias P. J., 2012, ApJ, 744, L6
* Boesgaard & Tripicco (1986) Boesgaard A. M., Tripicco M. J., 1986, ApJ, 302, L49
* Brown et al. (1989) Brown J. A., Sneden C., Lambert D. L., Dutchover E. J., 1989, ApJS, 71, 293
* Buder et al. (2021) Buder S., et al., 2021, MNRAS, 506, 150
* Cameron (1955) Cameron A. G. W., 1955, ApJ, 121, 144
* Cameron & Fowler (1971) Cameron A. G. W., Fowler W. A., 1971, ApJ, 164, 111
* Casey et al. (2016) Casey A. R., et al., 2016, MNRAS, 461, 3336
* Casey et al. (2019) Casey A. R., et al., 2019, ApJ, 880, 125
* Charbonnel & Lagarde (2010) Charbonnel C., Lagarde N., 2010, A&A, 522, A10
* Chen et al. (2001) Chen Y. Q., Nissen P. E., Benoni T., Zhao G., 2001, A&A, 371, 943
* Cox & Giuli (1968) Cox J. P., Giuli R. T., 1968, Principles of stellar structure. Gordon & Breach, New York
* Deepak & Lambert (2021) Deepak ., Lambert D. L., 2021, MNRAS, 505, 642
* Deepak & Reddy (2019) Deepak ., Reddy B. E., 2019, MNRAS, 484, 2000
* Deepak & Reddy (2020) Deepak ., Reddy B. E., 2020, Mem. Soc. Astron. Italiana, 91, 134
* Deepak et al. (2020) Deepak ., Lambert D. L., Reddy B. E., 2020, MNRAS, 494, 1348
* Despain (1981) Despain K. H., 1981, ApJ, 251, 639
* Eggleton et al. (2008) Eggleton P. P., Dearborn D. S. P., Lattanzio J. C., 2008, ApJ, 677, 581
* Gaia Collaboration et al. (2016) Gaia Collaboration et al., 2016, A&A, 595, A1
* Gao et al. (2019) Gao Q., Shi J.-R., Yan H.-L., Yan T.-S., Xiang M.-S., Zhou Y.-T., Li C.-Q., Zhao G., 2019, ApJS, 245, 33
* Iben (2013) Iben I. J., 2013, Stellar Evolution Physics, Volume 2: Advanced Evolution of Single Stars. Cambridge University Press
* Jorissen et al. (2020) Jorissen A., Van Winckel H., Siess L., Escorza A., Pourbaix D., Van Eck S., 2020, A&A, 639, A7
* Karakas (2010) Karakas A. I., 2010, MNRAS, 403, 1413
* Kippenhahn et al. (1980) Kippenhahn R., Ruschenplatt G., Thomas H. C., 1980, A&A, 91, 175
* Kumar et al. (2020) Kumar Y. B., Reddy B. E., Campbell S. W., Maben S., Zhao G., Ting Y.-S., 2020, Nature Astronomy, 4, 1059
* Lagarde et al. (2011) Lagarde N., Charbonnel C., Decressin T., Hagelberg J., 2011, A&A, 536, A28
* Lambert & Reddy (2004) Lambert D. L., Reddy B. E., 2004, MNRAS, 349, 757
* Lattanzio et al. (2015) Lattanzio J. C., Siess L., Church R. P., Angelou G., Stancliffe R. J., Doherty C. L., Stephen T., Campbell S. W., 2015, MNRAS, 446, 2673
* Martell et al. (2021) Martell S. L., et al., 2021, MNRAS, 505, 5340
* McKellar (1940) McKellar A., 1940, PASP, 52, 407
* Paxton et al. (2011) Paxton B., Bildsten L., Dotter A., Herwig F., Lesaffre P., Timmes F., 2011, ApJS, 192, 3
* Paxton et al. (2013) Paxton B., et al., 2013, ApJS, 208, 4
* Paxton et al. (2015) Paxton B., et al., 2015, ApJS, 220, 15
* Paxton et al. (2018) Paxton B., et al., 2018, ApJS, 234, 34
* Paxton et al. (2019) Paxton B., et al., 2019, ApJS, 243, 10
* Randich et al. (2020) Randich S., et al., 2020, A&A, 640, L1
* Sanford (1950) Sanford R. F., 1950, ApJ, 111, 262
* Schwab (2020) Schwab J., 2020, ApJ, 901, L18
* Sharma et al. (2018) Sharma S., et al., 2018, MNRAS, 473, 2004
* Singh et al. (2019) Singh R., Reddy B. E., Bharat Kumar Y., Antia H. M., 2019, ApJ, 878, L21
* Singh et al. (2021) Singh R., Reddy B. E., Campbell S. W., Kumar Y. B., Vrard M., 2021, ApJ, 913, L4
* Smith & Lambert (1989) Smith V. V., Lambert D. L., 1989, ApJ, 345, L75
* Thomas (1967) Thomas H. C., 1967, Z. Astrophys., 67, 420
* Wallerstein & Sneden (1982) Wallerstein G., Sneden C., 1982, ApJ, 255, 577
* Yan et al. (2018) Yan H.-L., et al., 2018, Nature Astronomy, 2, 790
* Yan et al. (2021) Yan H.-L., et al., 2021, Nature Astronomy, 5, 86
## Appendix A Tables summarizing results from the MESA models
Table 1: The initial temperature, luminosity and $\it A$(Li) along with their values at ZAMS, at the onset of LB and at the tip of the RGB for the models shown in Figure 11 and 16. $M$/M⊙ | Z | Temperature (K) at | $\log$ (L/L⊙) at | A(Li) at
---|---|---|---|---
| | PMS | ZAMS | LB | RGB-tip | PMS | ZAMS | LB | RGB-tip | PMS | ZAMS | LB | RGB-tip
0.8 | 0.0143 | 4082 | 4894 | 4422 | 3028 | -0.19 | -0.56 | 1.32 | 3.40 | 3.30 | 2.07 | -3.03 | -4.67
0.9 | 0.0143 | 4243 | 5323 | 4430 | 3087 | -0.01 | -0.33 | 1.42 | 3.40 | 3.30 | 2.77 | -0.26 | -2.90
1.0 | 0.0045 | 4681 | 6165 | 4600 | 3454 | 0.36 | 0.07 | 1.76 | 3.36 | 2.78 | 2.77 | 1.28 | -0.50
1.0 | 0.0080 | 4533 | 5936 | 4537 | 3307 | 0.26 | -0.01 | 1.62 | 3.38 | 3.04 | 2.99 | 1.43 | -0.49
1.0 | 0.0143 | 4380 | 5678 | 4438 | 3141 | 0.15 | -0.11 | 1.51 | 3.40 | 3.30 | 3.02 | 0.68 | -1.35
1.0 | 0.0254 | 4262 | 5403 | 4351 | 2940 | 0.06 | -0.21 | 1.37 | 3.40 | 3.57 | 1.87 | -4.92 | -4.49
1.1 | 0.0143 | 4491 | 5966 | 4455 | 3189 | 0.29 | 0.09 | 1.57 | 3.40 | 3.30 | 3.15 | 1.13 | -0.45
1.2 | 0.0143 | 4580 | 6231 | 4469 | 3230 | 0.41 | 0.27 | 1.62 | 3.39 | 3.30 | 3.22 | 1.34 | 0.12
1.3 | 0.0143 | 4651 | 6510 | 4474 | 3267 | 0.53 | 0.43 | 1.68 | 3.39 | 3.30 | 3.26 | 1.44 | 0.50
1.4 | 0.0143 | 4708 | 6863 | 4477 | 3300 | 0.63 | 0.58 | 1.73 | 3.39 | 3.30 | 3.28 | 1.49 | 0.78
1.5 | 0.0143 | 4754 | 7305 | 4486 | 3331 | 0.73 | 0.71 | 1.78 | 3.39 | 3.30 | 3.29 | 1.51 | 0.98
1.6 | 0.0143 | 4789 | 7768 | 4493 | 3361 | 0.82 | 0.83 | 1.82 | 3.39 | 3.30 | 3.29 | 1.53 | 1.13
1.8 | 0.0143 | 4838 | 8628 | 4462 | 3433 | 0.99 | 1.05 | 1.97 | 3.36 | 3.30 | 3.30 | 1.53 | 1.32
2.0 | 0.0143 | 4866 | 9388 | 4494 | 3570 | 1.14 | 1.23 | 2.01 | 3.24 | 3.30 | 3.30 | 1.51 | 1.40
2.2 | 0.0143 | 4878 | 10091 | 4373 | 3886 | 1.28 | 1.39 | 2.26 | 2.89 | 3.30 | 3.30 | 1.50 | 1.46
Table 2: Input parameters for the best fit models along with lithium abundances at the start of the different evolutionary phases for the three (M, [Fe/H]) cases studied in details in Section 4.1.1. Mass | [Fe/H] | Model parameters | $\delta A$(Li) | A(Li) after correction for $\delta A$(Li) at
---|---|---|---|---
(M⊙) | (dex) | Z | $\alpha_{\rm ML}$ | $\alpha_{\rm th}$ | (dex) | PMS | ZAMS | LB | RGB-tip
1.1 | -0.50 | 0.0045 | 1.80 $\pm$ 0.05 | 150 $\pm$ 50 | -0.3 | 2.48 | 2.48 | 0.92 | -0.77 $\pm$ 0.3
1.1 | 0.00 | 0.0143 | 1.85 $\pm$ 0.05 | 100 $\pm$ 50 | -0.6 | 2.70 | 2.55 | 0.53 | -1.05 $\pm$ 0.3
1.5 | -0.25 | 0.0080 | 1.95 $\pm$ 0.05 | 300 $\pm$ 100 | -0.2 | 2.84 | 2.83 | 1.10 | -0.08 $\pm$ 0.3
|
# Does the Heisenberg uncertainty principle apply along the time dimension?
John Ashmead Visiting Scholar, University of Pennsylvania
<EMAIL_ADDRESS>
###### Abstract
Does the Heisenberg uncertainty principle (HUP) apply along the time dimension
in the same way it applies along the three space dimensions? Relativity says
it should; current practice says no. With recent advances in measurement at
the attosecond scale it is now possible to decide this question
experimentally.
The most direct test is to measure the time-of-arrival of a quantum particle:
if the HUP applies in time, then the dispersion in the time-of-arrival will be
measurably increased.
We develop an appropriate metric of time-of-arrival in the standard case;
extend this to include the case where there is uncertainty in time; then
compare. There is – as expected – increased uncertainty in the time-of-arrival
if the HUP applies along the time axis. The results are fully constrained by
Lorentz covariance, therefore uniquely defined, therefore falsifiable.
So we have an experimental question on our hands. Any definite resolution
would have significant implications with respect to the role of time in
quantum mechanics and relativity. A positive result would also have
significant practical applications in the areas of quantum communication,
attosecond physics (e.g. protein folding), and quantum computing.
## 1 Introduction
> “You can have as much junk in the guess as you like, provided that the
> consequences can be compared with experiment.” – Richard P. Feynman [1]
##### Heisenberg uncertainty principle in time
The Heisenberg uncertainty principle in space:
$\Delta x\Delta p\geq 1$ (1)
is a foundational principle of quantum mechanics. From special relativity and
the requirement of Lorentz covariance we expect that this should be extended
to time:
$\Delta t\Delta E\geq 1$ (2)
Both Bohr and Einstein regarded the uncertainty principle in time as essential
to the integrity of quantum mechanics.
At the sixth Conference of Solvay in 1930, Einstein devised the celebrated
Clock-in-a-Box experiment to refute quantum mechanics by showing the HUP in
time could be broken. Consider a box full of photons with a small fast door
controlled by a clock. Weigh the box. Open the door for a time $\Delta t$,
long enough for one photon to escape. Weigh the box again. From the
equivalence of mass and energy, the change in the weight of the box gives you
the _exact_ energy of the photon. If $\Delta E\to 0$, then you can make
$\Delta t\Delta E<1$.
Bohr famously refuted the refutation by looking at the specifics of the
experiment. If you weigh the box by adding and removing small weights till the
box’s weight is again balanced, the process will take time. During this
process the clock will move up and down in the gravitational field, and its
time rate will be affected by the gravitational redshift. Bohr showed that
resulting uncertainty is precisely what is required to restore the HUP in time
[2, 3]. The irony of employing Einstein’s own General Relativity (via the use
of the gravitational redshift) to refute Einstein’s argument was presumably
lost on neither Einstein nor Bohr.
In later work this symmetry between time and space was in general lost. This
is clear in the Schrödinger equation:
$\imath\frac{\partial}{\partial\tau}\psi_{\tau}\left(\vec{x}\right)=H\psi_{\tau}\left(\vec{x}\right)$
(3)
Here the wave function is indexed by time: if we know the wave function at
time $\tau$ we can use this equation to compute the wave function at time
$\tau+\epsilon$. The wave function has in general non-zero dispersion in
space, but is always treated as having zero dispersion in time. In quantum
mechanics “time is a parameter not an operator” (Hilgevoord [4, 5]). Or as
Busch [6] puts it “… different types of time energy uncertainty can indeed be
deduced in specific contexts, but … there is no unique universal relation that
could stand on equal footing with the position-momentum uncertainty relation.”
See also Pauli, Dirac, and Muga [7, 8, 9, 10].
So, we have a contradiction at the heart of quantum mechanics. Symmetry
requires that the uncertainty principle apply in time; experience to date says
this is not necessary. Both are strong arguments; neither is decisive on its
own. The goal of this work is to reduce this to an experimental question; to
show we can address the question with current technology.
##### Order of magnitude estimate
The relevant time scale is the time for a photon to cross an atom. This is of
order attoseconds. This is small enough to explain why such an effect has not
already been seen by chance. Therefore the argument from experience is not
decisive.
But recent advances in ultra-fast time measurements have been extraordinary;
we can now do measurements at even the sub-attosecond time scale [11].
Therefore we should now be able to measure the effects of uncertainty in time,
if they are present.
Further the principle of covariance strongly constrains the effects. In the
same way that the inhabitants of Abbott’s Flatland [12] can infer the
properties of a Sphere from the Sphere’s projection on Flatland, we should be
able to predict the effects of uncertainty in time from the known
uncertainties in space.
##### Operational meaning of uncertainty in time
In an earlier work [13] we used the path integral approach to do
this111Detailed references are provided in the earlier work. For here we note
that the initial ideas come from the work of Stueckelberg and Feynman [14, 15,
16, 17, 18, 19] as further developed by Horwitz, Fanchi, Piron, Land, Collins,
and others [20, 21, 22, 23, 24, 25, 26, 27].. We generalized the usual paths
in the three space dimensions to extend in time as well. No other change was
required. The results were manifestly covariant by construction, uniquely
defined, and consistent with existing results in the appropriate long time
limit.
While there are presumably other routes to the same end, the requirement of
Lorentz covariance is a strong one and implies – at a minimum – that the first
order corrections of any such route will be the same. We can therefore present
a well-defined and falsifiable target to the experimentalists.
##### Time-of-arrival measurements
The most obvious line of attack is to use a time-of-arrival detector. We emit
a particle at a known time, with known average velocity and dispersion in
velocity. We measure the dispersion in time at a detector. The necessary
uncertainty in space will be associated with uncertainty in time-of-arrival.
For instance if the wave function has an uncertainty in position of $\Delta x$
and an average velocity of $v$, there will be an uncertainty of time-of-
arrival of order $\frac{\Delta x}{v}$. We refer to this as the extrinsic
uncertainty.
If there is an intrinsic uncertainty in time associated with the particle we
expect to see an additional uncertainty in time-of-arrival from that. If the
intrinsic uncertainty in time is $\Delta t$, then we would expect a total
uncertainty in time-of-arrival of order $\Delta t+\frac{\Delta x}{v}$. It is
the difference between the two predictions that is the experimental target.
For brevity we will refer to standard quantum mechanics, without uncertainty
in time, as SQM. And we will refer to quantum mechanics with intrinsic
uncertainty in time as TQM.
##### Time-of-arrival in SQM
We clearly need a solid measure of the extrinsic uncertainty to serve as a
reference point. We look at a number of available measures but are forced to
conclude none of the established metrics are entirely satisfactory: they
impose arbitrary conditions only valid in classical mechanics, have free
parameters, or are physically unrealistic. We argue that the difficulties are
of a fundamental nature. In quantum mechanics there cannot really be a crisp
boundary between the wave function detected and the equipment doing the
detection: we must – as a Bohr would surely insist – consider both
simultaneously as part of a single system. However once this requirement is
recognized and accepted, we can work out the rules for a specific setup and
get reasonably solid predictions.
##### Comparison of time-of-arrival in SQM and TQM
With this preliminary question dealt with, the actual comparison of the
results with and without intrinsic uncertainty in time is straightforward. We
first look at a generic particle detector setup in SQM, then look at the
parallel results in TQM. By writing the latter as the direct product of a time
part and a space part, we get a clean comparison. As expected, if there is
intrinsic uncertainty in time then the uncertainty in time-of-arrival is
increased and by a specific amount.
In the non-relativistic limit, the relative increase in the uncertainty in
time-of-arrival is small. But by running the initial wave function through a
single slit in time, as a rapidly opening and closing camera shutter, we can
make the relative increase arbitrarily great.
_We therefore have falsifiability._
##### Implications
Since the predictions come directly from the principle of Lorentz covariance
and the basic rules for quantum mechanics, any definite result – negative or
positive – will have implications for foundational questions in relativity and
quantum mechanics.
If there is in fact intrinsic uncertainty in time, there will be practical
implications for quantum communication, attosecond physics (e.g. protein
folding), and quantum computing. This will open up new ways to carry
information, additional kinds of interaction between quantum particles (forces
of anticipation and regret, resulting from the extension in time), and a
deeper understanding of the measurement problem.
## 2 Time-of-arrival measurements in SQM
There does not appear to be a well-defined, generally accepted, and
appropriate metric for the time-of-arrival in the case of SQM.
The Kijowski metric [28] is perhaps the most popular metric for time-of-
arrival. It has a long history, back to 1974, has seen a great deal of use,
and has some experimental confirmation. However it is based on a deeply
embedded classical assumption which is unjustifiable from a quantum mechanical
viewpoint and which produces inconsistent results in some edge cases. We will
nevertheless find it useful as a starting point.
A conceptually sounder approach is supplied by Marchewka and Schuss, who use
Feynman path integrals to compute the time-of-arrival [29, 30, 31, 32].
Unfortunately their approach has free parameters, which rules it out for use
here.
The difficulties with the Marchewka and Schuss approach are partly a function
of an ad hoc treatment of the boundary between incoming wave function and the
plane of the detector. We are able to deal with this boundary in a systematic
way by using a discrete approach along the space coordinate, then taking the
continuum limit.
Unfortunately while this discrete approach solves the problem of how to handle
the discontinuity at the edge of the detector it does not address the more
fundamental problem: of assuming a crisp boundary between the quantum
mechanical and classical parts of the problem in the first place.
We address this problem by arguing that while the assumption of a crisp
boundary between the quantum mechanical and classical parts of a problem is
not in general justified it is also not in general necessary. If we work
bottom-up, starting from a quantum mechanical perspective, we can use various
ad hoc but reasonable heuristics to determine which parts may be described
with acceptable accuracy using classical mechanics, and where only a fully
quantum mechanical approach will suffice.
### 2.1 Kijowski time-of-arrival operator
> “A common theme is that classical mechanics, deterministic or stochastic, is
> always a fundamental reference on which all of these approaches [to the
> time-of-arrival operator] are based.” – Muga and Leavens [33]
#### 2.1.1 Definition
The most popular time-of-arrival operator seems to be one developed by
Kijowski in 1974 [28]. It and variations on it have seen much use since then
[34, 35, 33, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45]. We used Kijowski as a
starting point in our previous paper: it gives reasonable results in simple
cases. Unfortunately its conceptual problems rule it out for use here.
The main problem is that classical requirements play an essential role in
Kijowski’s original derivation. In particular, Kijowski assumed that for the
part of the wave function to the left of the detector222Throughout we assume
that the particle is going left to right along the $x$ axis; that the detector
is at $x=0$; that the particle starts well to the left of the detector; and
that we do not need to consider the $y$ and $z$ axes. These assumptions are
conventional in the time-of-arrival literature. only components with $p>0$
will contribute; while for the part of the wave function to the right of the
detector only components with $p<0$ will contribute. This condition (which we
will refer to as the classical condition) is imposed on the basis of the
corresponding requirement in classical mechanics, as noted by Muga and Leavens
above.
Figure 1: The classical condition for time-of-arrival measurements
Focusing on the case of a single space dimension, with the detector at $x=0$,
the Kijowski metric is333Equation 122 in [33]. :
$\rho_{d}\left(\tau\right)={\left|{\int\limits_{0}^{\infty}dp\sqrt{\frac{p}{{2\pi
m}}}{e^{-\imath\frac{{{p^{2}}\tau}}{{2m}}}}{\varphi^{\left({left}\right)}}\left(p\right)}\right|^{2}}+{\left|{\int\limits_{-\infty}^{0}dp\sqrt{\frac{{-p}}{{2\pi
m}}}{e^{-\imath\frac{{{p^{2}}\tau}}{{2m}}}}{\varphi^{\left({right}\right)}}\left(p\right)}\right|^{2}}$
(4)
However while the restrictions on momentum may make sense
classically444Amusingly enough, this condition is sometimes violated even in
classical waves: Rayleigh surface waves in shallow water (e.g. tsunamis
approaching land) show retrograde motion of parts of the wave [46]., they are
more troubling in quantum mechanics. It is as if we were thinking of the wave
function as composed of a myriad of classical particles each of which travels
like a tiny billiard ball. This would appear to be a hidden variable
interpretation of quantum mechanics. This was shown to be inconsistent with
quantum mechanics by Bell in 1965 [47] experimentally confirmed [48, 49, 50,
51, 52].
However, Kijowski’s metric has produced reasonable results, see for instance
[53]. Therefore if we see problems with it, we need to explain the successes
as well.
#### 2.1.2 Application to Gaussian test functions
We will use Gaussian test functions (GTFs) to probe this metric. We define
GTFs as normalized Gaussian functions that are solutions to the free
Schrödinger equation. We will look at the results of applying Kijowski’s
metric to two different cases, which we refer to as the “bullet” and “wave”
variations, following Feynman’s terminology in his discussion of the double-
slit experiment [54]. By “bullet” we mean a wave function which is narrowly
focused around the central axis of motion; by “wave” we mean a wave function
which is widely spread around the central axis of motion.
##### Bullets
For the bullet version we will take a GTF which starts at time $\tau=0$,
centered at start position $x_{0}=-d$, with average momentum $p_{0}$, and
dispersion $\sigma_{p}^{2}\ll p_{0}$. These conditions are met in the Muga,
Baute, Damborenea, and Egusquiza [53] reference; they give:
$\delta
v=0.098{{cm}\mathord{\left/{\vphantom{{cm}s}}\right.\kern-1.2pt}s},{v_{0}}=10{{cm}\mathord{\left/{\vphantom{{cm}s}}\right.\kern-1.2pt}s}$
(5)
This is typical of high-energy particles, where the free particle wave
functions will have essentially no negative momentum component. The starting
wave function is:
${{\bar{\varphi}}_{0}}\left(x\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}{e^{\imath{p_{0}}x-\frac{1}{{2\sigma_{x}^{2}}}{{\left({x+d}\right)}^{2}}}}$
(6)
This is taken to satisfy the free Schrödinger equation:
$\imath\frac{\partial}{{\partial\tau}}{{\bar{\varphi}}_{\tau}}\left(x\right)=-\frac{1}{{2m}}\frac{{\partial^{2}}}{{\partial{x^{2}}}}{{\bar{\varphi}}_{\tau}}\left(x\right)$
(7)
so is given as a function of time by:
$\bar{\varphi}_{\tau}\left(x\right)=\sqrt[4]{\frac{1}{\pi\sigma_{x}^{2}}}\sqrt{\frac{1}{f_{\tau}^{\left(x\right)}}}e^{\imath
p_{0}x-\frac{1}{2\sigma_{x}^{2}f_{\tau}^{\left(x\right)}}\left(x+d-\frac{p_{0}}{m}\tau\right)^{2}-\imath\frac{p_{0}^{2}}{2m}\tau}$
(8)
with dispersion factor $f$ defined by:
$f_{\tau}^{\left(x\right)}\equiv 1+\imath\frac{\tau}{m\sigma_{x}^{2}}$ (9)
The momentum transform is:
${{\hat{\bar{\varphi}}}_{\tau}}\left(p\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{p}^{2}}}}}{e^{\imath
pd-\frac{{{\left({p-{p_{0}}}\right)}^{2}}}{{2\sigma_{p}^{2}}}-\imath\frac{{p^{2}}}{{2m}}\tau}}$
(10)
with $\sigma_{p}\equiv\frac{1}{\sigma_{x}}$. We will use this wave function as
a starting point throughout this work, referring to it as the “bullet wave
function”.
In Kijowski’s notation, there is no right hand wave function since the
particle started on the left. Further since the wave function is by
construction narrow in momentum space we can extend the lower limit in the
integral on the left from zero to negative infinity giving:
${\rho_{d}}\left(\tau\right)={\left|{\int\limits_{-\infty}^{\infty}dp\sqrt{\frac{p}{{2\pi
m}}}\sqrt[4]{{\frac{1}{{\pi\sigma_{p}^{2}}}}}{e^{\imath
pd-\frac{{{\left({p-{p_{0}}}\right)}^{2}}}{{2\sigma_{p}^{2}}}-\imath\frac{{p^{2}}}{{2m}}\tau}}}\right|^{2}}$
(11)
Since the wave function is narrowly focused, we can replace the square root in
the integral by the average:
$\sqrt{\frac{{{p_{0}}+\delta p}}{m}}\approx\sqrt{\frac{{{\text{p}}_{0}}}{m}}$
(12)
which we can then pull outside of the integral:
${\rho_{d}}\left(\tau\right)\approx\frac{{p_{0}}}{m}{\left|{\int\limits_{-\infty}^{\infty}{\frac{{dp}}{{\sqrt{2\pi}}}\sqrt[4]{{\frac{1}{{\pi\sigma_{p}^{2}}}}}{e^{\imath
pd-\frac{{{\left({p-{p_{0}}}\right)}^{2}}}{{2\sigma_{p}^{2}}}-\imath\frac{{p^{2}}}{{2m}}\tau}}}}\right|^{2}}$
(13)
We recognize the integral as simply the inverse Fourier transform of the
momentum space form:
${{\bar{\varphi}}_{\tau}}\left(d\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}\sqrt{\frac{1}{{f_{\tau}^{\left(x\right)}}}}{e^{\imath{p_{0}}d-\frac{1}{{2\sigma_{x}^{2}f_{\tau}^{\left(x\right)}}}{{\left({d-\frac{{p_{0}}}{m}\tau}\right)}^{2}}-\imath\frac{{p_{0}^{2}}}{{2m}}\tau}}$
(14)
so:
${\rho_{d}}\left(\tau\right)\approx\frac{{p_{0}}}{m}{\left|{{{\bar{\varphi}}_{\tau}}\left(d\right)}\right|^{2}}=\frac{{p_{0}}}{m}{{\bar{\rho}}_{\tau}}\left(d\right)$
(15)
with the probability density being:
${{\bar{\rho}}_{\tau}}\left(d\right)=\frac{{p_{0}}}{m}\sqrt{\frac{1}{{\pi\sigma_{x}^{2}\left({1+\frac{{\tau^{2}}}{{{m^{2}}\sigma_{x}^{4}}}}\right)}}}\exp\left({-\frac{{{\left({{\bar{x}}_{\tau}}\right)}^{2}}}{{\sigma_{x}^{2}\left({1+\frac{{\tau^{2}}}{{{m^{2}}\sigma_{x}^{4}}}}\right)}}}\right)$
(16)
and with the average location in space:
$\bar{x}_{\tau}\equiv-d+{v_{0}}\tau$ (17)
We have the probability density as a function of $x$; we need it as a function
of $\tau$. We therefore make the variable transformation:
$d\tau={v_{0}}dx=\frac{{p_{0}}}{m}dx$ (18)
And rewrite $\tau$ as the average value of $\tau$ plus an offset:
$\tau=\bar{\tau}+\delta\tau,\bar{\tau}\equiv\frac{d}{{v_{0}}}$ (19)
We rewrite the numerator in the density function in terms of $\delta\tau$:
$\bar{\rho}_{\tau}\left(d\right)\approx\sqrt{\frac{1}{\pi\sigma_{x}^{2}\left|f_{\tau}^{\left(x\right)}\right|^{2}}}e^{-\frac{\left(v_{0}\delta\tau\right)^{2}}{\sigma_{x}^{2}\left|f_{\tau}^{\left(x\right)}\right|^{2}}}$
(20)
We may estimate the uncertainty in time by expanding around the average time.
Since the numerator is already only of second order in $\delta\tau$ we need
only keep the zeroth order in $\delta\tau$ in the denominator:
$\sigma_{x}^{2}\left|f_{\tau}^{\left(x\right)}\right|^{2}=\sigma_{x}^{2}+\frac{\tau^{2}}{m^{2}\sigma_{x}^{2}}\approx\frac{\tau^{2}}{m^{2}\sigma_{x}^{2}}=\frac{\left(\bar{\tau}+\delta\tau\right)^{2}}{m^{2}\sigma_{x}^{2}}\approx\frac{\bar{\tau}^{2}}{m^{2}\sigma_{x}^{2}}$
(21)
giving:
$\bar{\rho}_{\delta\tau}\approx\sqrt{\frac{v_{0}^{2}m^{2}\sigma_{x}^{2}}{\pi\bar{\tau}^{2}}}e^{-\frac{v_{0}^{2}m^{2}\sigma_{x}^{2}}{\bar{\tau}^{2}}\left(\delta\tau\right)^{2}}$
(22)
We define an effective dispersion in time:
$\bar{\sigma}_{\tau}\equiv\frac{1}{mv_{0}\sigma_{x}}\bar{\tau}$ (23)
So we have the probability of detection as roughly:
$\bar{\rho}_{\delta\tau}\approx\sqrt{\frac{1}{\pi\bar{\sigma}_{\tau}^{2}}}e^{-\frac{\left(\delta\tau\right)^{2}}{\bar{\sigma}_{\tau}^{2}}}$
(24)
This is normalized to one, centered on $\tau=\bar{\tau}$, and with
uncertainty:
$\Delta\tau=\frac{1}{\sqrt{2}}\bar{\sigma}_{\tau}$ (25)
It is interesting that the classical condition has played no essential role in
this derivation. In spite of that, the result is about what we would expect
from a classical particle with probability distribution $\bar{\rho}$.
Therefore the implication of the experimental results in Muga and Leavens is
that they are confirmation of the reasonableness of the classical
approximation in general – rather than of Kijowski’s classical condition in
particular.
##### Waves
Now consider a more wave-like Gaussian test function: slow and wide. Take
$0<{p_{0}}\ll{\sigma_{p}}$ in equation 10, in fact make $p_{0}$ only very
slightly greater than 0.
Our GTF is still moving to the right but very slowly. If we wait long enough –
and if there were no detector in the way – it will eventually get completely
past the detector. Since there is a detector, the GTF will encounter this
instead. We will first assume a perfect detector with no reflection.
Use of the Kijowski distribution requires we drop the $p<0$ half of the wave
function because we started on the left:
${\varphi^{\left({wave}\right)}}\left(p\right)\Rightarrow\sqrt[4]{{\frac{1}{{\pi\sigma_{p}^{2}}}}}{e^{\imath
pd-\frac{{\left({p-{p_{0}}}\right)^{2}}}{{2\sigma_{p}^{2}}}}}\theta\left(p\right)$
(26)
We can no longer justify extending the lower bound on the left hand integral
from zero to negative infinity. We have instead:
${\rho_{d}}\left(\tau\right)={\left|{\int\limits_{0}^{\infty}dp\sqrt{\frac{p}{{2\pi
m}}}\sqrt[4]{{\frac{1}{{\pi\sigma_{p}^{2}}}}}{e^{\imath
pd-\frac{{{\left({p-{p_{0}}}\right)}^{2}}}{{2\sigma_{p}^{2}}}-\imath\frac{{p^{2}}}{{2m}}\tau}}}\right|^{2}}$
(27)
This means we have dropped nearly half of the wave function and therefore
expect a detection rate of only about 25%.
The integral over the momentum gives a sum over Bessel functions of fractional
order; the subsequent integrals over time are distinctly non-trivial. However
in the limiting case $d\to 0,p_{0}\to 0$ (the wave function starts at the
detector) the distribution simplifies drastically giving:
$\rho_{0}\left(\tau\right)={\left|{\frac{{\sqrt[4]{m}\sqrt{\sigma_{p}^{2}}\Gamma\left({\frac{3}{4}}\right)}}{{{{(2\pi)}^{3/4}}{{(m+\imath\sigma_{p}^{2}\tau)}^{3/4}}}}}\right|^{2}}$
(28)
And this in turn is simple enough to let us compute the norm analytically:
$\int\limits_{0}^{\infty}{d\tau\rho_{0}\left(\tau\right)}=\frac{1}{4}$ (29)
This is exactly as expected: if we throw out half the wave function, the
subsequent norms are the square of the wave function or one quarter.
This result is general. The GTF obeys the Schrödinger equation, which is norm
preserving. If we throw out fraction $F$ of the wave function, the resulting
wave function has norm ${\left({1-F}\right)^{2}}$. Assuming perfect detection,
the Kijowski distribution will under-count the actual detections by this
factor.
It is arguable that the problem is the assumption of perfect detection.
However if we drill down to the actual mechanism of detection, this will
involve some sort of interaction of the incoming particles with various atoms,
typically involving the calculation of matrix elements of the form (e.g.
Sakurai [55]):
$\left\langle{\varphi_{\tau}^{\left({n^{\prime}l^{\prime}m^{\prime}}\right)}\left({\vec{x}}\right)}\right|\sqrt{\frac{{n_{\vec{k},\alpha}}}{{2\omega
V}}}{{\vec{\varepsilon}}^{\left(\alpha\right)}}\exp\left({\imath\vec{k}\cdot\vec{x}-\imath\omega\tau}\right)\left|{\varphi_{\tau}^{\left({nlm}\right)}\left({\vec{x}}\right)}\right\rangle$
(30)
It is difficult to make sense of the classical condition in this context.
Would we have to apply it to the interaction with each atom in turn? And how
would the atoms know whether they are supposed to be detecting particles
coming from the left or the right? And therefore how would the unfortunate
atoms know whether they should drop the upper left and bottom right quadrants
on the diagram or vice versa?
#### 2.1.3 Requirements for a time-of-arrival metric
The classical condition was introduced without justification in terms of
quantum mechanics itself. As there are no cases where the rules of quantum
mechanics have failed to produce correct results, this is not acceptable.
Further it is unnecessary. The probability current, discussed below
(subsubsection 3.1.1), provides a conceptually sound and purely quantum
mechanical approach to the problem of detection.
We therefore have to dispense entirely with the classical condition. As this
is the foundation of the Kijowski class of metrics – per Muga and Leavens
above – we have to dispense with the class as well.
We have however learned a bit about our requirements: we are looking for a
time-of-arrival metric which satisfies the requirements:
1. 1.
Comes out of a completely quantum mechanical analysis, with no ad hoc
classical requirements imposed on it.
2. 2.
Includes an account of the interaction with the detector. As Bohr pointed out
during the EPR debate [56, 57], an analysis of a quantum mechanical detection
must include the specifics of the apparatus. There is no well-defined value of
a quantum parameter in the abstract, but only in the context of a specific
measurement.
3. 3.
Conserves probability.
4. 4.
And of course, is as simple as possible given the first three requirements.
### 2.2 Marchewka and Schuss path integral approach
> “According to our assumptions, trajectories that propagate into the
> absorbing boundary never return into the interval [a, b] so that the Feynman
> integral over these trajectories is supported outside the interval. On the
> other hand, the Feynman integral over the bounded trajectories in the
> interval is supported inside the interval. Thus the two integrals are
> orthogonal and give rise to no interference.” – Marchewka and Schuss [31]
We turn to an approach from Marchewka and Schuss [29, 30, 31, 32]. They use a
path integral approach plus the assumption of an absorbing boundary. To
compute the wave function, they use a recursive approach: computing the wave
function at each step, subtracting off the absorption at each step, and so to
the end.
Marchewka and Schuss start with the wave function defined over a general
interval from $a\to b$. We can simplify their treatment slightly by taking
$a=-\infty,b=0$. They break the time evolution up into steps of size
$\epsilon$. Given the wave function at step $n$, they compute the wave
function at step $n+1$ as:
${\psi_{n+1}}\left({x_{n+1}}\right)=\int\limits_{-\infty}^{0}{d{x_{n}}{K_{\varepsilon}}\left({{x_{n+1}};{x_{n}}}\right)}{\psi_{n}}\left({x_{n}}\right)$
(31)
The kernel can be quite general.
Marchewka and Schuss then introduce the assumption that the probability to be
absorbed at each endpoint is proportional to the value of the probability to
be at the end point times a characteristic length $\lambda$555They refer to
$\lambda$ as a “fudge parameter”. The value of $\lambda$ has to found by
running some reference experiments; after this is done the values are
determined for “nearby” experiments.:
${P_{n}}=\lambda{\left|{{\psi_{n}}\left(0\right)}\right|^{2}}$ (32)
Using the kernel they compute the probability of absorption as:
${P_{n}}=\frac{\varepsilon}{{2\pi{m}}}\left({\lambda\left|{\left.{{\partial_{x}}{\psi_{n}}\left(x\right)}\right|_{x=0}^{2}}\right|}\right)$
(33)
They then shrink the wave function at the next step by an appropriate factor,
the survival probability:
${\psi_{n+1}}\left(x\right)\to\sqrt{1-{P_{n}}}{\psi_{n+1}}\left(x\right)$ (34)
This guarantees the total probability is constant, where the total probability
is the probability to be detected plus the probability to survive. This
overcomes the most obvious problem with the Kijowski approach.
At the same time it represents an explicit loss of the phase information.
There is nothing to keep the “characteristic length” from being complex;
taking it as real is a choice. Therefore we are seeing the loss of information
normally associated with the act of measurement. So we have effectively broken
the total measurements at the end points down into a series of time dependent
“mini-measurements” or “mini-collapses” at each time step.
Computing the product of all these infinitesimal absorptions they get the
final corrected wave function as:
$\psi_{\tau}^{\left({corrected}\right)}\left(x\right)=\exp\left({-\frac{1}{{2\pi{m}}}\int\limits_{0}^{\tau}{d\tau^{\prime}\left({{\lambda}\left|{{\left.{{\partial_{x}}{\psi_{\tau}}\left(0\right)}\right|}^{2}}\right|}\right)}}\right){\psi_{\tau}}\left(x\right)$
(35)
They go on to apply this approach to a variety of cases, as GTFs going from
left to right.
The Marchewka and Schuss approach is a significant advance over the Kijowski
metric: it works at purely quantum mechanical level, it includes the detector
in the analysis, and it guarantees by construction overall conservation of
probability.
But at the same time the introduction of the characteristic length is a major
problem for falsifiability: how do we know whether the same length should be
used for SQM and for TQM?
And the trick for estimating the loss of probability at each step implicitly
assumes no interference in time. This is not a problem for SQM but a point we
need to keep in mind when including TQM.
We therefor further refine our requirements for a metric: we need to handle
any discontinuity at the detector in a well-defined way.
### 2.3 Discrete approach
Figure 2: Reflection principle
We can handle the discontinuity at the boundary in a well-defined way by
taking a discrete approach, by setting up a grid in $\tau,x$. We model quantum
mechanics as a random walk on the Feynman checkerboard [58]. Feynman used
steps taken at light speed; ours will travel a bit more slowly than that.
This lets us satisfy the previous requirements. And it lets us take advantage
of the extensive literature on random walks (e.g. [59, 60, 61, 62]).
We will start with a random walk model for detection, take the continuous
limit as the grid spacing goes to zero, Wick rotate in time, and then apply
the results to a GTF. This will give us a well-defined answer for the
detection rate for a GTF.
#### 2.3.1 Random Walk
We start with a grid, index by $n\in\left[{0,1,\ldots,\infty}\right]$ in the
time direction and by $m\in[-\infty,\ldots,-1,0,1,\ldots,\infty]$ in the space
direction. We start with a simple coin tossing picture. We will take the
probability to go left or right as equal, both $\frac{1}{2}$. $P_{nm}$ is the
probability to be at position $m$ at step $n.$ It is given recursively by:
${P_{n+1,m}}={P_{n,m-1}}/2+{P_{n,m+1}}/2$ (36)
The number of paths from $\left(0,0\right)\to\left(n,m\right)$ is given by the
binomial coefficient $\left({\begin{array}[]{c}n\\\
{\frac{{n+m}}{2}}\end{array}}\right)$. We will take this as zero if $n$ and
$m$ do not have the same parity, if they are not both even or both odd. If
$n>\left|{m}\right|$ then $P_{nm}=0$.
The total probability to get from $\left(0,0\right)\to\left(n,m\right)$ is
given by the number of paths times the probability $\frac{{1}}{2^{n}}$ for
each one. In our example:
${P_{nm}}=\left({\begin{array}[]{c}n\\\
{\frac{{n+m}}{2}}\end{array}}\right)\frac{1}{{2^{n}}}$ (37)
If we wish to start at a position other than zero, say $m^{\prime}$, we
replace $m$ with $m-m^{\prime}$.
##### First arrival
We can model detection as the first arrival of the path at the position of the
detector. We start the particle at step 0 at position $m=-d$ and position the
detector at $m=0$. We take the probability of the first arrival at step $n$
when starting at position $-d$ as ${F_{n}}$.
We define the function $G_{nm}$ as the probability to arrive at position $m$
at time $n$, _without_ having been previously detected. We can get this
directly from the raw probabilities using the reflection principle (e.g. [59])
as follows:
The number of paths that start at $\left(0,-d\right)$ and get to
$\left(n,m\right)$ is given by: $\left({\begin{array}[]{c}n\\\
{\frac{{n+{m+d}}}{2}}\end{array}}\right)$. Now consider all paths that get to
$m$ at step $n$ having first touched the detector at some step $f$. By
symmetry, the number of paths from $-d\to f$ is the same as the number of
paths from $d\to f$. The number of these paths is
$\left({\begin{array}[]{c}n\\\
{\frac{{n+\left|{m-d}\right|}}{2}}\end{array}}\right)$.
Therefore we have the number of paths that reach $m$ without first touching
the line $m=0$ as ([63]):
${G_{nm}}=\left({\left({\begin{array}[]{c}n\\\
{\frac{{n+m+d}}{2}}\end{array}}\right)-\left({\begin{array}[]{c}n\\\
{\frac{{n+m-d}}{2}}\end{array}}\right)}\right)\frac{1}{{2^{n}}}$ (38)
And of course, if $\left(n,m\right)$ do not have the same parity, $G_{nm}=0$.
To get to the detector at step $n$, the particle must have been one step to
the left at the previous step. It will then have a probability of
$\frac{{1}}{2}$ to get the detector. So we have:
${F_{n}}=\frac{{1}}{2}{G_{n-1,-1}}$ (39)
giving:
$F_{n}=\frac{d}{n}\left({\begin{array}[]{c}n\\\
{\frac{{n+d}}{2}}\end{array}}\right)\frac{{1}}{2^{n}}=\frac{d}{n}{P_{nd}}$
(40)
If the particle is already at position $0$ at step $0$ we declare that it has
arrived at time zero so we have:
$F_{0}={\delta_{d0}}$ (41)
#### 2.3.2 Diffusion equation
The passage to the continuum limit is well understood ([63, 62]). Define:
$\begin{gathered}\tau=n\Delta\tau\hfill\\\ x=m\Delta x\hfill\end{gathered}$
(42)
We get the diffusion equation if we take the limit so that both $\Delta x$ and
$\Delta\tau$ become infinitesimal while we keep fixed the ratio:
${D_{0}}=\mathop{\lim}\limits_{\Delta\tau,\Delta x\to 0}\frac{{{\left({\Delta
x}\right)}^{2}}}{{2\Delta\tau}}$ (43)
This gives the diffusion equation for the probability $P$:
$\frac{{\partial
P}}{{\partial\tau}}={D_{0}}\frac{{{\partial^{2}}P}}{{\partial{x^{2}}}}$ (44)
To further simplify we take $D_{0}=\frac{1}{2}$:
$\frac{{\partial{P_{\tau}}}}{{\partial\tau}}=\frac{1}{2}\frac{{\partial^{2}}}{{\partial{x^{2}}}}{P_{\tau}}$
(45)
If we start with the probability distribution
${P_{0}}\left(x-x^{\prime}\right)=\delta\left({x-x^{\prime}}\right)$ the
probability distribution as a function of time is then:
${P_{\tau}}\left({x;x^{\prime}}\right)=\frac{1}{{\sqrt{2\pi\tau}}}\exp\left({-\frac{{{\left({x-x^{\prime}}\right)}^{2}}}{{2\tau}}}\right)$
(46)
with the probability of a first arrival at $\tau$ being given by:
${F_{\tau}}\left({x;x^{\prime}}\right)=\frac{{\left|{x-x^{\prime}}\right|}}{{\sqrt{2\pi{\tau^{3}}}}}\exp\left({-\frac{{{\left({x-x^{\prime}}\right)}^{2}}}{{2\tau}}}\right)$
(47)
To include the mass, scale by $\tau\to\tau^{\prime}\equiv m\tau$ giving:
$\frac{{\partial{P_{\tau}}}}{{\partial\tau^{\prime}}}=\frac{1}{{2m}}\frac{{\partial^{2}}}{{\partial{x^{2}}}}{P_{\tau^{\prime}}}$
(48)
and:
${P_{\tau}}\left({x;x^{\prime}}\right)=\frac{{\sqrt{m}}}{{\sqrt{2\pi\tau}}}\exp\left({-m\frac{{{\left({x-x^{\prime}}\right)}^{2}}}{{2\tau}}}\right)$
(49)
To get the detection rate we have to handle the scaling of the time a bit
differently, since it is a probability distribution in time. The detection
rate only makes sense when integrated over $\tau$, so the substitution
implies:
$\int{d\tau{F_{\tau}}\to\int{\frac{{d\tau^{\prime}}}{m}{F_{\tau^{\prime}}}}}\to\int{d\tau^{\prime}{D_{\tau^{\prime}}}}\Rightarrow{D_{\tau}}=\frac{1}{m}{F_{\tau}}$
(50)
So we have:
${D_{\tau}}\left({x-x^{\prime}}\right)=\frac{{\left|{x-x^{\prime}}\right|}}{\tau}\frac{{\sqrt{m}}}{{\sqrt{2\pi\tau}}}\exp\left({-m\frac{{{\left({x-x^{\prime}}\right)}^{2}}}{{2\tau}}}\right)$
(51)
As a double check on this, we derive the detection rate directly from the
diffusion equation using, again, the method of images (see appendix A).
#### 2.3.3 Wick rotation
To go from the diffusion equation to the Schrödinger equation:
$\imath\frac{\partial}{{\partial\tau}}\psi=-\frac{1}{{2{m}}}\frac{{\partial^{2}}}{{\partial{x^{2}}}}\psi$
(52)
we use Wick rotation:
$\tau\to-\imath\tau,P\to\psi$ (53)
so we have the Wick-rotated kernel $K$:
${K_{\tau}}\left({x;x^{\prime}}\right)=\sqrt{\frac{m}{{2\pi\imath\tau}}}\exp\left({\imath
m\frac{{{\left({x-x^{\prime}}\right)}^{2}}}{{2\tau}}}\right)$ (54)
and for the kernel for first arrival $F$:
${F_{\tau}}\left({x;x^{\prime}}\right)=\frac{{\left|{x-x^{\prime}}\right|}}{\tau}{K_{\tau}}\left({x-x^{\prime}}\right)$
(55)
As a second double check, we derive $F$ directly using Laplace transforms (see
appendix B).
#### 2.3.4 Application to a Gaussian test function
We start with the “bullet” GTF from our analysis of the Kijowski metric
(equation 6):
${\varphi_{0}}\left({x}\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}{e^{\imath{p_{0}}x-\frac{1}{{2\sigma_{x}^{2}}}{{\left({x+d}\right)}^{2}}}}$
(56)
with:
$\left|{d}\right|\gg{\sigma_{x}}$ (57)
The wave function at the detector is then:
${\varphi_{\tau}}\left(0\right)=\int\limits_{-\infty}^{0}{dx^{\prime}{F_{\tau}}\left({0;x^{\prime}}\right)}{\varphi_{0}}\left({x^{\prime}}\right)=\int\limits_{-\infty}^{0}{dx^{\prime}\frac{{\left|{0-x^{\prime}}\right|}}{\tau}{K_{\tau}}\left({0;x^{\prime}}\right)}{\varphi_{0}}\left({x^{\prime}}\right)$
(58)
We can solve this explicitly in terms of error functions. However since
$\varphi_{0}$ is strongly centered on $-d$ we can use the same trick as we did
in the analysis of the Kijowski metric. We extend the upper limit of
integration to $\infty$:
${\varphi_{\tau}}\left(0\right)\approx-\frac{1}{\tau}\int\limits_{-\infty}^{\infty}{dx^{\prime}x^{\prime}{K_{\tau}}\left({0;x^{\prime}}\right)}{\varphi_{0}}\left({x^{\prime}}\right)$
(59)
We see by inspection we can pull down the $x^{\prime}$ by taking the
derivative of $\varphi_{0}$ with respect to $p_{0}$. This gives:
${\varphi_{\tau}}\left(0\right)\approx\imath\frac{1}{{\tau}}\frac{\partial}{{\partial{p_{0}}}}\int\limits_{-\infty}^{\infty}{dx^{\prime}{K_{\tau}}\left({0;x^{\prime}}\right)}{\varphi_{0}}\left({x^{\prime}}\right)$
(60)
We recognize the integral as the integral that propagates the free wave
function from $0\to\tau$:
${\varphi_{\tau}}\left(0\right)\approx\imath\frac{1}{{\tau}}\frac{\partial}{{\partial{p_{0}}}}\varphi_{\tau}^{\left({free}\right)}\left(0\right)$
(61)
We have:
$\varphi_{\tau}^{\left({free}\right)}\left(0\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}\sqrt{\frac{1}{{f_{\tau}^{\left(x\right)}}}}{e^{-\frac{1}{{2\sigma_{x}^{2}f_{\tau}^{\left(x\right)}}}{{\left({d-\frac{{p_{0}}}{m}\tau}\right)}^{2}}-\imath\frac{{p_{0}^{2}}}{{2m}}\tau}}$
(62)
so the wave function for a first arrival at $\tau$ is:
${\varphi_{\tau}}\left(0\right)\approx\left({\imath\frac{1}{{m}}\frac{{d-{v_{0}}\tau}}{{\sigma_{x}^{2}f_{\tau}^{\left(x\right)}}}+v_{0}}\right)\varphi_{\tau}^{\left({free}\right)}\left(0\right)$
(63)
#### 2.3.5 Metrics
We can compute the probability density to a sufficient order by expanding
around the average time of arrival:
$\bar{\tau}\equiv\frac{d}{{v_{0}}},\delta\tau\equiv\tau-\bar{\tau}$ (64)
The first term in the expression for the $\varphi_{\tau}$ scales as
$1/{\bar{\tau}}$; the second as a constant so at large $\tau$ we have:
${\rho_{\tau}}\left(0\right)\approx
v_{0}^{2}\rho_{\tau}^{\left({free}\right)}\left(0\right)$ (65)
The result is the same as with Kijowski, except for the overall factor of
$v_{0}^{2}$. Since the uncertainty is produced by normalizing this, the
uncertainty in time is the same as with the bullet calculation for Kijowski.
But the factor of $v_{0}^{2}$ is troubling, particularly when we recall we are
using natural units so that at non-relativistic speeds this is hard to
distinguish from zero.
The immediate problem is our somewhat thin-skinned approximation. The random
walk can only penetrate to position $m=0$ in the grid before it is absorbed.
And since the probability density is computed as the square, the result is
doubly small.
Given the taking of the limit, it is perhaps more surprising that we get a
finite detection rate at all. A realistic approximation would have the paths
penetrating some finite distance into the detector, with the absorption going
perhaps with some characteristic length – as with the Marchewka and Schuss
approach.
In return for a _more_ detailed examination of the boundary we have gotten a
_less_ physically realistic estimate. But the real problem with all three of
the approaches we have looked at is their assumption of a crisp boundary
between quantum and classical realms.
### 2.4 Implications for time-of-arrival measurements
> “What is it that we human beings ultimately depend on? We depend on our
> words. We are suspended in language. Our task is to communicate experience
> and ideas to others. We must strive continually to extend the scope of our
> description, but in such a way that our messages do not thereby lose their
> objective or unambiguous character.” – Niels Bohr as cited in [64]
In reality there can be no crisp boundary between quantum and classical
realms. Consider the lifespan of an atom built on classical principles, as in
say Bohr’s planetary model of the atom. A classical electron in orbit around
the nucleus undergoes centripetal acceleration from the nucleus; therefore
emits Larmor radiation; therefore decays in towards the nucleus. The estimated
lifespan is of order $10^{-11}$ seconds [65].
What keeps this from happening in the case of a quantum atom is the
uncertainty principle: as the electron spirals into the nucleus its
uncertainty in position is reduced so its uncertainty in momentum is
increased. A larger $\Delta p$ means a larger kinetic energy. In fact the
ground state energy can be estimated as the minimum total energy – kinetic
plus potential – consistent with the uncertainty principle.
Since everything material is built of atoms and since there are no classical
atoms there is no classical realm. And therefore it is incorrect to speak of a
transition to a classical end state. No such state exists.
However classical mechanics does work very well in practice. So what we have
is a problem in description. Which parts of our experimental setup – particle,
source, detector, observers – can be described to acceptable accuracy using
classical language; where must we use a fully quantum description?
This approach places us, with Bohr [66], firmly on the epistemological side of
the measurement debate. One possible starting point is the program of
decoherence [67, 68, 69, 70, 71, 72, 73, 74]. The key question for decoherence
is not when does a system go from quantum to classical but rather when does a
classical description become possible. Another is Quantum Trajectory Theory
(QTT) [75, 76, 77, 78]. QTT, as its name suggests, is a particularly good fit
with the path integral approach, the formalism we are primarily focused on
here.
For the moment we will explore a more informal approach, taking advantage of
the conceptual simplicity of the path integral approach. We can describe the
evolution of a quantum particle in terms of the various paths it takes. In the
case of the problem at hand, we visualize the paths as those that enter the
detector or do not, that are reflected or are not, that leave and do not
return (backflow) or return again and so on. Ultimately however each of the
paths will escape permanently or else enter an entropy sink of some kind, a
structure that is sufficiently macroscopic that it is no longer possible to
analyze the process in terms of the individual paths. Its effects will live
on, perhaps in a photo-electric cascade or a change of temperature. But the
specific path will be lost like a spy in a crowd.
The entropy sink is the one essential part needed for a measurement to take
place. In a quantum mechanical system information, per the no-cloning theorem,
is neither created nor destroyed. But the heart of a measurement is the
reduction of the complexity of the system under examination to a few numbers.
Information is necessarily lost in doing this. In general it is fairly clear
where this happens; typically in crossing the boundary from the microscopic to
the macroscopic scale.
Figure 3: Global character of the measurement problem
Typical paths may cross the boundary multiple times, ultimately either being
absorbed and registered (measured), lost in some way (detector inefficiency),
or reflected (backflow). There will be some time delay associated with all of
these cases666A typical representation could be done in terms of a Laplace
transform $\mathcal{D}\left(s\right)$ of the detection amplitude plus a
Laplace transform $\mathcal{R}\left(s\right)$ of the reflection amplitude. The
Marchewka and Schuss fudge parameter may be thought of as a useful first
approximation of $\mathcal{D}$..
This is not too far from conventional practice. A macroscopic detector
functions as the entropy sink, its efficiency is usually known, and
reflections are possible but minimized. In many cases the response time
associated with the process of detection may not require much attention – we
may only care that there was a detection. Our case is a bit trickier than
that. We have to care about the time delay associated with a detection,
especially if the time delay introduces some uncertainty in time in its own
right.
This picture is enough to show that there can not be, in general, a local
time-of-arrival operator. Suppose there is such. Consider the paths leaving it
on the right, continuing to an entropic sink of some kind. If we arrange a
quantum mirror of some kind (perhaps just a regular mirror), then some of the
paths will return to their source. The detection rate will be correspondingly
reduced. But this can only be predicted using knowledge of the global
situation, of the interposed mirror. An operator local to the boundary cannot
know about the mirror and therefore cannot give a correct prediction of the
time-of-arrival distribution. Therefore the time-of-arrival cannot in general
be given correctly by a purely local operator.
Since the fundamental laws are those of quantum mechanics, the analysis must
be carried out at a quantum mechanical level – except for those parts where we
can show the classical approximation suffices, usually the end states. In
fact, there is no guarantee that even the end states can be adequately
described in classical terms; while there are no cases – so far – of a
perdurable and macroscopic quantum system777But see the macroscopic wave
functions in [79, 80]., there is nothing to rule that out.
## 3 Effects of dispersion in time on time-of-arrival measurements
> “Clearly,” the Time Traveller proceeded, “any real body must have extension
> in four directions: it must have Length, Breadth, Thickness, and — Duration.
> But through a natural infirmity of the flesh, which I will explain to you in
> a moment, we incline to overlook this fact. There are really four
> dimensions, three which we call the three planes of Space, and a fourth,
> Time. There is, however, a tendency to draw an unreal distinction between
> the former three dimensions and the latter, because it happens that our
> consciousness moves intermittently in one direction along the latter from
> the beginning to the end of our lives.” – H. G. Wells [81]
We now return to our original problem, first estimating the dispersion in
time-of-arrival without dispersion in time, then with.
### 3.1 Time-of-arrival measurements without dispersion in time
Our efforts to correctly define a time of arrival operator have led to the
conclusion that the time-of–arrival – like measurement in general – is a
system level property; it is not something that can be correctly described by
a local operator.
We have therefore to a certain extent painted ourselves into a corner. We
would like to compute the results for the general case, but we have shown
there are only specific cases. We will deal with this by trading precision for
generality.
We will first assume a perfect detector, then justify the assumption.
#### 3.1.1 Probability current in SQM
We will first compute the detection rate by using the probability current per
Marchewka and Schuss [32]. We start with the Schrödinger equation:
$\imath\frac{\partial}{{\partial\tau}}{\psi_{\tau}}\left(x\right)=-\frac{1}{{2m}}\frac{{\partial^{2}}}{{\partial{x^{2}}}}{\psi_{\tau}}\left(x\right)$
(66)
We will assume we have in hand a wave function that includes the interaction
with the detector. We define the probability of detection at clock time $\tau$
as $D_{\tau}$.
The total probability of the wave function to be found to the left of the
detector at time $\tau$ is the integral of the probability density from
$-\infty\to 0$. The total probability to have been detected up to time $\tau$
is the integral of $D_{\tau}$ from $0\to\tau$. Therefore by conservation of
probability we have:
$1=\int\limits_{-\infty}^{0}{dx}\psi_{\tau}^{*}\left(x\right){\psi_{\tau}}\left(x\right)+\int\limits_{0}^{\tau}{d\tau^{\prime}{D_{\tau^{\prime}}}}$
(67)
We take the derivative with respect to $\tau$ to get:
$D_{\tau}=-\int\limits_{-\infty}^{0}{dx\frac{{\partial\psi_{\tau}^{*}\left(x\right)}}{{\partial\tau}}{\psi_{\tau}}\left(x\right)+\psi_{\tau}^{*}\left(x\right)\frac{{\partial{\psi_{\tau}}\left(x\right)}}{{\partial\tau}}}$
(68)
We use the Schrödinger equation:
${D_{\tau}}=\frac{\imath}{{2m}}\int\limits_{-\infty}^{0}{dx\frac{{{\partial^{2}}\psi_{\tau}^{*}\left(x\right)}}{{\partial{x^{2}}}}{\psi_{\tau}}\left(x\right)-\psi_{\tau}^{*}\left(x\right)\frac{{{\partial^{2}}{\psi_{\tau}}\left(x\right)}}{{\partial{x^{2}}}}}$
(69)
And integrate by parts to get the equation for the detection rate:
${D_{\tau}}={\left.{\frac{\imath}{{2m}}\left({\frac{{\partial\psi_{\tau}^{*}\left(x\right)}}{{\partial
x}}{\psi_{\tau}}\left(x\right)-\psi_{\tau}^{*}\left(x\right)\frac{{\partial{\psi_{\tau}}\left(x\right)}}{{\partial
x}}}\right)}\right|_{x=0}}$ (70)
We recognize the expression on the right as the probability current:
${J_{\tau}}\left(x\right)=-\frac{\imath}{{2}}\left({\psi_{\tau}^{*}\left(x\right)\frac{{\partial{\psi_{\tau}}\left(x\right)}}{{\partial
x}}-{\psi_{\tau}}\left(x\right)\frac{{\partial\psi_{\tau}^{*}\left(x\right)}}{{\partial
x}}}\right)$ (71)
or:
${J_{\tau}}=\psi_{\tau}^{*}\frac{p}{{2m}}{\psi_{\tau}}-{\psi_{\tau}}\frac{p}{{2m}}\psi_{\tau}^{*}$
(72)
so:
$D_{\tau}=J_{\tau}$ (73)
Note that $D_{\tau}$ can, in the general case, be negative. For instance, if
the detector includes – per the previous section – a mirror of some kind there
may be backflow, probability current going from right to left.
Ironically enough, this is in fact the first metric that Kijowski considered
in his paper, only to reject it because it violated the classical condition 2.
#### 3.1.2 Black box detector
Given the wave function we can compute the detection rate using the
probability current. The real problem is to compute the wave function in the
first place, given that this must in principle include the interaction of the
particle with the detector. To treat the general case, we would like a general
detector, one where the details of the interaction do not matter. We are
looking for something which is a perfect black.
We can use Einstein’s clock-in-a-box, but this time run it in reverse: have
the detector be a box that has a small window open for a fraction of a second
then check how many particles are in the box afterward by weighing the box.
The interior walls of the box provide the necessary entropy sink. We will
refer to this as a black box detector 888A “perfect black” is apparently not
quite as theoretical a concept as it sounds: the commercial material
Vantablack can absorb 99.965% of incoming light [82]. And Vantablack even uses
a similar mechanism: incoming light is trapped in vertically aligned nanotube
arrays..
In fact, this is a reasonable model of the elementary treatment of detection.
When, for instance, we compute the rate of scattering of particles into a bit
of solid angle, we assume that the particles will be absorbed with some
efficiency but we do not generally worry about subtleties as time delay till a
detection is registered or backflow.
#### 3.1.3 Application to a Gaussian test function
We will take a GTF at the plane as above:
${\varphi_{\tau}}\left(x\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}\sqrt{\frac{1}{{f_{\tau}^{\left(x\right)}}}}{e^{\imath{p_{0}}x-\frac{1}{{2\sigma_{x}^{2}f_{\tau}^{\left(x\right)}}}{{\left({x+d-\frac{{p_{0}}}{m}{\tau}}\right)}^{2}}-\imath\frac{{p_{0}^{2}}}{{2m}}{\tau}}}$
(74)
An elementary application of the probability current gives the detection rate
as the velocity times the probability density:
${D_{\tau}}=\frac{{p_{0}}}{m}{\rho_{\tau}}\left(0\right)$ (75)
This is the same as we had for the Kijowski bullet (equation 16). And it is,
as noted there, also the classical rate for detection of a probability
distribution impacting a wall.
Our use of the Einstein box in reverse is intended to provide a lowest common
denominator for quantum measurements; it is not surprising that the lowest
common denominator for quantum measurements should be the corresponding
classical distribution.
We have the uncertainty in time-of-arrival as before (equation 25):
$\Delta\tau=\frac{1}{{\sqrt{2}}}{{\bar{\sigma}}_{\tau}}=\frac{1}{{\sqrt{2}}}\frac{1}{{m{v_{0}}{\sigma_{x}}}}\bar{\tau}$
(76)
Since we have $\bar{p}=m{v_{0}}$ and – for a minimum uncertainty wave function
– ${\sigma_{p}}=\frac{1}{{\sigma_{x}}}$ we can also write:
$\frac{{\sigma_{\tau}}}{{\bar{\tau}}}=\frac{{{\hat{\sigma}}_{p}}}{{\bar{p}}}$
(77)
so the dispersion in momentum and in clock time are closely related, as one
might expect.
While the results are the same as with the Kijowski metric (for bullet wave
functions), the advantage is that we now have a clearer sense of what is
actually going on and therefore what corrections we might have to make in
practice. These may include (but will probably not be limited to):
1. 1.
Depletion, e.g. inefficiencies of the detector,
2. 2.
Backflow, e.g. a reflection coefficient,
3. 3.
Delay in the detection (not normally important, but significant here),
4. 4.
Self-interference of the wave function with itself,
5. 5.
Edge effects – one can imagine the incoming part of the wave function skimming
along the edge of the detector for instance,
6. 6.
And of course the increasingly complex effects that result from increasingly
complex models of the wave function-detector interaction.
### 3.2 Time-of-arrival measurements with dispersion in time
#### 3.2.1 Quantum mechanics with uncertainty in time
Figure 4: Paths in three and in four dimensions
Path integrals provide a natural way to extend quantum mechanics to include
time in a fully covariant way. In [13] we did this by first extending the
usual three-dimensional paths to four dimensions. The rest of the treatment
followed from this single assumption together with the twin requirements of
covariance and of consistency with known results. But as one might expect,
there are a fair number of questions to be addressed along the way – making
the full paper both rather long and rather technical. To make this treatment
self-contained we summarize the key points here.
##### Feynman path integrals in four dimensions
We start with clock time, defined operationally as what the laboratory clock
measures. This is the parameter $\tau$ in the Schrödinger equation, Klein-
Gordon equation, path integral sums, and so on.
The normal three dimensional paths are parameterized by clock time; we
generalize them to four dimensional paths, introducing the coordinate time $t$
to do so:
${\bar{\pi}_{\tau}}\left({x,y,z}\right)\to{\pi_{\tau}}\left({t,x,y,z}\right)$
(78)
It can be helpful to think of coordinate time as another space dimension. We
will review the relationship between clock time and coordinate time at the
end, once the formalism has been laid out. (This is a case where it is helpful
to let the formalism precede the intuition.)
In path integrals we get the amplitude to go from one point to another by
summing over all intervening paths, weighing each path by the corresponding
action. The path integral with four dimensional paths is formally identical to
the path integral with three dimensions, except that the paths take all
possible routes in four rather than three dimensions. We define the kernel as:
${K_{\tau}}\left({x^{\prime\prime};x^{\prime}}\right)=\mathop{\lim}\limits_{N\to\infty}\int\mathcal{D}\pi{e^{\imath\int\limits_{0}^{\tau}{d\tau^{\prime}\mathcal{L}\left[{\pi,\dot{\pi}}\right]}}}$
(79)
We use a Lagrangian which works for both the three and four dimensional cases
(see [83]):
$\mathcal{L}\left({{x^{\mu}},{{\dot{x}}^{\mu}}}\right)=-\frac{1}{2}m{{\dot{x}}^{\mu}}{{\dot{x}}_{\mu}}-q{{\dot{x}}^{\mu}}{A_{\mu}}\left(x\right)-\frac{m}{2}$
(80)
As usual with path integrals, to actually do the integrals we discretize clock
time $\tau=\epsilon n$ and replace the integral over clock time with a sum
over steps:
$\int\limits_{0}^{\tau}{d\tau^{\prime}\mathcal{L}}\to\sum\limits_{j=0}^{n}{\mathcal{L}_{j}}$
with:
$\mathcal{L}_{j}\equiv-m\frac{\left(x_{j}-x_{j-1}\right)^{2}}{2\varepsilon^{2}}-q\frac{x_{j}-x_{j-1}}{\epsilon}\frac{A\left(x_{j}\right)+A\left(x_{j-1}\right)}{2}-\frac{m}{2}$
(81)
And measure:
$\mathcal{D}\pi\equiv\left(-\imath\frac{m^{2}}{4\pi^{2}\varepsilon^{2}}\right)^{N}\prod\limits_{n=1}^{N-1}d^{4}x_{n}$
(82)
This gives us the ability to compute the amplitude to get from one point to
another in a way that includes paths that vary in time as well as space.
##### Schrödinger equation in four dimensions
Usually we derive the path integral expressions from the Schrödinger equation.
But here we derive the four dimensional Schrödinger equation from the short
time limit of the path integral kernel, running the usual derivation in
reverse and with one extra dimension. We get:
$\imath\frac{\partial\psi_{\tau}}{\partial\tau}=-\frac{1}{{2m}}\left({\left({{p_{\mu}}-q{A_{\mu}}}\right)\left({{p^{\mu}}-q{A^{\mu}}}\right)-m^{2}}\right){\psi_{\tau}}$
(83)
Note that here $p$ is an operator, $A$ is an external field, and $m$ is the
constant operator999In the extension to QED, $A$ becomes an operator as well..
This equation goes back to Stueckelberg [14, 15] with further development by
Feynman, Fanchi, Land, and Horwitz [16, 23, 24, 27].
We need only the free Schrödinger equation here. With the vector potential $A$
set to zero we have:
$\imath\frac{\partial}{{\partial\tau}}{\psi_{\tau}}=-\frac{{{E^{2}}-{\overrightarrow{p}^{2}}-m^{2}}}{{2m}}{\psi_{\tau}}$
(84)
Or spelled out:
$\imath\frac{\partial}{{\partial\tau}}{\psi_{\tau}}\left({t,\vec{x}}\right)=\frac{1}{{2m}}\left({\frac{{\partial^{2}}}{{\partial{t^{2}}}}-{\nabla^{2}}-{m^{2}}}\right){\psi_{\tau}}\left({t,\vec{x}}\right)$
(85)
If the left hand side were zero, the right hand side would be the Klein-Gordon
equation (with $t\to\tau$). Over longer times – femtoseconds or more – the
left side will in general average to zero, giving the Klein-Gordon equation as
the long time limit. More formally we expect that if we average over times
$\Delta\tau$ of femtoseconds or greater, we will have:
$\int\limits_{\tau}^{\tau+\Delta\tau}{d\tau^{\prime}\int{{d^{4}}x}}\psi_{\tau}^{*}\left(x\right)\frac{{{E^{2}}-{{\vec{p}}^{2}}-{m^{2}}}}{{2m}}{\psi_{\tau}}\left(x\right)\approx
0$ (86)
But at short times – attoseconds – we should see effects associated with
uncertainty in time. This is in general how we get from a fully four
dimensional approach to the usual three dimensional approach; the fluctuations
in coordinate time tend to average out over femtosecond and longer time
scales. But at shorter times – attoseconds – we will see the effects of
uncertainty in time.
It is much the same way that in SQM quantum effects in space average out over
longer time and distance scales to give classical mechanics. TQM is to SQM –
with respect to time – as SQM is to classical mechanics with respect to the
three space dimensions.
##### Wave functions in coordinate time
We need an estimate of the wave function at source, its evolution in clock
time, and the rules for detection – the birth, life, and death of the wave
function if you will.
###### Initial wave function
We start with the free Klein-Gordon equation and use the method of separation
of variables to break the wave function out into a space and a time part:
$\varphi_{0}\left(t,x\right)=\tilde{\varphi}_{0}\left(t\right)\bar{\varphi}_{0}\left(x\right)$
(87)
Or in energy/momentum space:
$\hat{\varphi}_{0}\left(E,p\right)=\hat{\tilde{\varphi}}_{0}\left(E\right)\hat{\bar{\varphi}}_{0}\left(p\right)$
(88)
We assume we already have the space part by standard methods. For instance, if
this is a GTF in momentum space it will look like:
${{\hat{\bar{\varphi}}}_{\tau}}\left(p\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{p}^{2}}}}}{e^{-\imath
p{x_{0}}-\frac{{{\left({p-{p_{0}}}\right)}^{2}}}{{2\sigma_{p}^{2}}}-\imath\frac{{p^{2}}}{{2m}}\tau}}$
(89)
This plus the Klein-Gordon equation give us constraints on the time part:
$\begin{array}[]{l}\left\langle 1\right\rangle=1\\\ \bar{E}\equiv\left\langle
E\right\rangle=\sqrt{m^{2}+\left\langle\vec{p}\right\rangle^{2}}\\\
\left\langle E^{2}\right\rangle=\left\langle
m^{2}+\vec{p}^{2}\right\rangle=m^{2}+\left\langle\vec{p}^{2}\right\rangle\end{array}$
(90)
To get a robust estimate of the shape of the time part we assume it is the
maximum entropy solution that satisfies the constraints. We used the method of
Lagrange multipliers to find this getting:
$\hat{\tilde{\varphi}}_{0}\left(E\right)=\sqrt[4]{\frac{1}{\pi\sigma_{E}^{2}}}e^{\imath\left(E-E_{0}\right)t_{0}-\frac{\left(E-E_{0}\right)^{2}}{2\sigma_{E}^{2}}}$
(91)
with values:
$\sigma_{E}^{2}=\sigma_{p}^{2}$ (92)
and:
$E_{0}=\sqrt{m^{2}+\bar{p}^{2}}$ (93)
To get the starting wave function in time we take the inverse Fourier
transform:
$\tilde{\varphi}_{0}\left(t\right)=\sqrt[4]{\frac{1}{\pi\sigma_{t}^{2}}}e^{-\imath
E_{0}\left(t-t_{0}\right)-\frac{t^{2}}{2\sigma_{t}^{2}}}$ (94)
$\sigma_{t}^{2}=\frac{1}{\sigma_{E}^{2}}$ (95)
The value of the $\sigma_{t}$ will normally be of order attoseconds. We set
$t_{0}=0$ as the overall phase is already supplied by the space/momentum part.
Since estimates from maximum entropy tend to be robust we expect this approach
will give estimates which are order-of-magnitude correct.
###### Propagation of a wave function
The TQM form of the Klein-Gordon equation is formally identical to the non-
relativistic Schrödinger equation with the additional coordinate time term. In
momentum space one can use this insight to write the kernel out by inspection:
${{\hat{K}}_{\tau}}\left({p^{\prime\prime};p^{\prime}}\right)={\delta^{\left(4\right)}}\left({p^{\prime\prime}-p^{\prime}}\right)\exp\left({\imath\frac{{{{p^{\prime}}^{2}}-{m^{2}}}}{{2m}}\tau}\right)\theta\left(\tau\right)$
(96)
The coordinate space form is:
${K_{\tau}}\left({x^{\prime\prime};x^{\prime}}\right)=-\imath\frac{{m^{2}}}{{4{\pi^{2}}{\tau^{2}}}}{e^{-\imath
m\frac{{{\left({x^{\prime\prime}-x^{\prime}}\right)}^{2}}}{{2\tau}}-\imath\frac{m}{2}\tau}}\theta\left(\tau\right)$
(97)
In both cases we have the product of a coordinate time part by the familiar
space part. Spelling this out in coordinate space101010We are using an over-
tilde and over-bar to distinguish between the coordinate time and the space
parts, when this is useful.:
$\begin{gathered}{K_{\tau}}\left({x^{\prime\prime};x^{\prime}}\right)={{\tilde{K}}_{\tau}}\left({{t^{\prime\prime}};{t^{\prime}}}\right){{\bar{K}}_{\tau}}\left({{{\vec{x}}^{\prime\prime}};{{\vec{x}}^{\prime}}}\right)\exp\left({-\imath\frac{m}{2}\tau}\right)\theta\left(\tau\right)\hfill\\\
{{\tilde{K}}_{\tau}}\left({{t^{\prime\prime}};{t^{\prime}}}\right)=\sqrt{\frac{{\imath
m}}{{2\pi\tau}}}\exp\left({-\imath
m\frac{{{\left({{t^{\prime\prime}}-{t^{\prime}}}\right)}^{2}}}{{2\tau}}}\right)\hfill\\\
{{\bar{K}}_{\tau}}\left({{{\vec{x}}^{\prime\prime}};{{\vec{x}}^{\prime}}}\right)={\sqrt{-\frac{\imath
m}{{2\pi\tau}}}^{3}}\exp\left({\imath
m\frac{{{\left({\vec{x}^{\prime\prime}-\vec{x}^{\prime}}\right)}^{2}}}{{2\tau}}}\right)\hfill\end{gathered}$
(98)
where the space part kernel is merely the familiar non-relativistic kernel
(e.g. Merzbacher [84]). The additional factor of
$\exp\left({-\imath\frac{m}{2}\tau}\right)$ only contributes an overall phase
which plays no role in single particle calculations.
If the initial wave function can be written as the product of a coordinate
time and a three-space part, then the propagation of the three-space part is
the same as in non-relativistic quantum mechanics. In general, if the free
wave function starts as a direct product in time and space it stays that way.
Of particular interest here is the behavior of a GTF in time. If at clock time
zero it is given by:
$\tilde{\varphi}_{0}\left(t\right)\equiv\sqrt[4]{\frac{1}{\pi\sigma_{t}^{2}}}e^{-\imath
E_{0}\left(t-t_{0}\right)-\frac{\left(t-t_{0}\right)^{2}}{2\sigma_{t}^{2}}}$
(99)
then as a function of clock time it is:
$\tilde{\varphi}_{\tau}\left(t\right)=\sqrt[4]{\frac{1}{\pi\sigma_{t}^{2}}}\sqrt{\frac{1}{f_{\tau}^{\left(t\right)}}}e^{-\imath
E_{0}t-\frac{1}{2\sigma_{t}^{2}f_{\tau}^{\left(t\right)}}\left(t-t_{0}-\frac{E_{0}}{m}\tau\right)^{2}+\imath\frac{E_{0}^{2}}{2m}\tau}$
(100)
with dispersion factor $f_{\tau}^{\left(t\right)}\equiv
1-\imath\frac{\tau}{m\sigma_{t}^{2}}$ and with expectation, probability
density, and uncertainty:
$\begin{array}[]{l}\left\langle{t_{\tau}}\right\rangle={t_{0}}+\frac{E}{m}\tau={t_{0}}+\gamma\tau\\\
{{\tilde{\rho}}_{\tau}}\left(t\right)=\sqrt{\frac{1}{{\pi\sigma_{t}^{2}}\left({1+\frac{{\tau^{2}}}{{{m^{2}}\sigma_{t}^{4}}}}\right)}}\exp\left({-\frac{{{\left({t-\left\langle{t_{\tau}}\right\rangle}\right)}^{2}}}{{\sigma_{t}^{2}\left({1+\frac{{\tau^{2}}}{{{m^{2}}\sigma_{t}^{4}}}}\right)}}}\right)\\\
{\left({\Delta t}\right)^{2}}\equiv\left\langle
t^{2}\right\rangle-{\left\langle
t\right\rangle^{2}}=\frac{{\sigma_{t}^{2}}}{2}\left|{1+\frac{{\tau^{2}}}{{{m^{2}}\sigma_{t}^{4}}}}\right|\end{array}$
(101)
The behavior of the time part is given by replacing $x\to t$ and taking the
complex conjugate. As noted, coordinate time behaves like a fourth space
dimension, albeit one that enters with the sign of $\imath$ flipped.
###### Detection of a wave function
To complete the birth, life, death cycle we look at the detection of a wave
function. This is obviously the place where things can get very tricky.
However the general rule that coordinate time acts as a fourth space
coordinate eliminates much of the ambiguity. If we were doing a measurement
along the $y$ dimension and we registered a click in a detector located at
$y_{detector}\pm\frac{\Delta y}{2}$ we would take this as meaning that we had
measured the particle as being at position $y_{detector}$ with uncertainty
$\Delta y$.
Because of our requirement of the strictest possible correspondence between
coordinate time and the space dimensions, the same rule necessarily applies in
coordinate time. If we have a clock in the detector with time resolution
$\Delta\tau$, then a click in the detector implies a measurement of the
particle at coordinate time $t=\tau$ with uncertainty $\Delta\tau.$
So the detector does an implicit conversion from coordinate time to clock
time, because we have assumed we know the position of the detector in clock
time. How do we justify this assumption?
The justification is that we take the clock time of the detector as being
itself the average of the coordinate time of the detector:
${\tau_{Detector}}\equiv\left\langle{t_{Detector}}\right\rangle$ (102)
The detector will in general be made up of a large number of individual
particles, each with a four-position in coordinate time and the three space
dimensions. While in TQM, the individual particles may be a bit in the future
or past, the average will behave as a classical variable, specifically as the
clock time. This drops out of the path integral formalism, which supports an
Ehrenfest’s theorem in time (again in [13]). If we take the coordinate time of
a macroscopic object as the sum of the coordinate times of its component
particles, the principle is macroscopically stronger. The motion in time of a
macroscopic object will no more display quantum fluctuations in time than its
motion in space displays quantum fluctuations in the $x,y,z$ directions.
This explains the differences in the properties of the clock time and
coordinate time. The clock time is something we think of as only going
forward; its corresponding energy operator is bounded below, usually by zero.
But if it is really an expectation of the coordinate time of a macroscopic
object, then it is not that the clock time cannot go backward, it is that it
is statistically extremely unlikely to do so. The principle is the same as the
argument that while the gas molecules in a room could suddenly pile up in one
half of a room and leave the other half a vacuum, they are statistically
extremely unlikely to do so.
Further the conjugate variable for coordinate time
${E^{\left({op}\right)}}\equiv\imath\frac{\partial}{{\partial t}}$ is not
subject to Pauli’s theorem [85]. The values of $E$ – the momentum conjugate to
the coordinate time $t$ – can go negative. If we look at a non-relativistic
GTF:
${{\hat{\tilde{\varphi}}}_{0}}\left(E\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{E}^{2}}}}}{e^{\imath
E{t_{0}}-\frac{{{\left({E-{E_{0}}}\right)}^{2}}}{{2\sigma_{E}^{2}}}}},{\sigma_{E}}=\frac{1}{{\sigma_{t}}},{E_{0}}\approx
m+\frac{{{\vec{p}}^{2}}}{{2m}}$ (103)
the value of $m$ for an electron is about 500,000 eV, while for $\sigma_{t}$
of order attoseconds the $\sigma_{E}$ will be of order 6000 eV. Therefore the
negative energy part is about $500000/6000$ or $83$ standard deviations away
from the average. And therefore the likelihood of detecting a negative energy
part of the wave function is zero to an excellent approximation. But not
_exactly_ zero.
So, clock time is the classical variable corresponding to the fully quantum
mechanical coordinate time. They are two perspectives on the same universe.
#### 3.2.2 Probability current in TQM
We now compute the probability current in TQM, working in parallel to the
Marchewka and Schuss derivation for SQM covered above (and with similar
caveats about its applicability).
By conservation of probability the sum of the detections as of clock time
$\tau$ plus the total probability remaining at $\tau$ must equal one:
$1=\int\limits_{0}^{\tau}{d\tau^{\prime}\int\limits_{-\infty}^{\infty}{dt{D_{\tau^{\prime}}}\left(t\right)}}+\int\limits_{-\infty}^{0}{dx}\int\limits_{-\infty}^{\infty}{dt}\psi_{\tau}^{*}\left({t,x}\right){\psi_{\tau}}\left({t,x}\right)$
(104)
Again we take the derivative with respect to clock time to get:
$\int\limits_{-\infty}^{\infty}{dt{D_{\tau}}\left(t\right)}=-\int\limits_{-\infty}^{0}{dx}\int\limits_{-\infty}^{\infty}{dt}\left({\dot{\psi}_{\tau}^{*}\left({t,x}\right){\psi_{\tau}}\left({t,x}\right)+\psi_{\tau}^{*}\left({t,x}\right){{\dot{\psi}}_{\tau}}\left({t,x}\right)}\right)$
(105)
From the Schrödinger equation:
$\begin{gathered}{{\dot{\psi}}_{\tau}}=-\imath\frac{1}{{2m}}\left({\frac{{\partial^{2}}}{{\partial{t^{2}}}}-\frac{{\partial^{2}}}{{\partial{x^{2}}}}}\right){\psi_{\tau}}\hfill\\\
\dot{\psi}_{\tau}^{*}=\imath\frac{1}{{2m}}\left({\frac{{\partial^{2}}}{{\partial{t^{2}}}}-\frac{{\partial^{2}}}{{\partial{x^{2}}}}}\right)\psi_{\tau}^{*}\hfill\end{gathered}$
(106)
We use integration by parts in coordinate time to show the terms in the second
derivative of coordinate time cancel:
$\imath\frac{1}{{2m}}\left({\frac{{\partial^{2}}}{{\partial{t^{2}}}}\psi_{\tau}^{*}}\right){\psi_{\tau}}-\imath\frac{1}{{2m}}\psi_{\tau}^{*}\frac{{\partial^{2}}}{{\partial{t^{2}}}}{\psi_{\tau}}\to-\imath\frac{1}{{2m}}\left({\frac{\partial}{{\partial
t}}\psi_{\tau}^{*}}\right)\frac{\partial}{{\partial
t}}{\psi_{\tau}}+\imath\frac{1}{{2m}}\left({\frac{\partial}{{\partial
t}}\psi_{\tau}^{*}}\right)\frac{\partial}{{\partial t}}{\psi_{\tau}}=0$ (107)
leaving:
$\mathop{\smallint}\limits_{-\infty}^{\infty}dt{D_{\tau}}\left(t\right)=\frac{\imath}{{2m}}\mathop{\smallint}\limits_{-\infty}^{\infty}dt\mathop{\smallint}\limits_{-\infty}^{0}dx\left({\left({\frac{{{\partial^{2}}\psi_{\tau}^{*}\left({t,x}\right)}}{{\partial{x^{2}}}}}\right){\psi_{\tau}}\left({t,x}\right)-\psi_{\tau}^{*}\left({t,x}\right)\left({\frac{{{\partial^{2}}{\psi_{\tau}}\left({t,x}\right)}}{{\partial{x^{2}}}}}\right)}\right)$
(108)
We have the probability current in the $x$ direction at each instant in
coordinate time for fixed clock time:
$j_{\tau}\left({t,x}\right)\equiv-\frac{\imath}{{2m}}\frac{{\partial\psi_{\tau}^{*}\left({t,x}\right)}}{{\partial
x}}{\psi_{\tau}}\left({t,x}\right)+\frac{\imath}{{2m}}\psi_{\tau}^{*}\left({t,x}\right)\frac{{\partial{\psi_{\tau}}\left({t,x}\right)}}{{\partial
x}}$ (109)
$j_{\tau}\left({t,x}\right)=\left({\frac{{p_{x}}}{{2m}}\psi_{\tau}^{*}\left({t,x}\right)}\right){\psi_{\tau}}\left({t,x}\right)+\psi_{\tau}^{*}\left({t,x}\right)\frac{{p_{x}}}{{2m}}{\psi_{\tau}}\left({t,x}\right)$
(110)
What we are after is the full detection rate at a specific coordinate time:
$D\left(t\right)=\int{d\tau{D_{\tau}}\left(t\right)}$ (111)
If we accept that the equality in equation 108 applies at each instant in
coordinate time (detailed balance) we have:
${D_{\tau}}\left(t\right)={\left.{{j_{\tau}}\left({t,x}\right)}\right|_{x=0}}$
(112)
giving:
$D\left(t\right)={\left.{\int{d\tau{j_{\tau}}\left({t,x}\right)}}\right|_{x=0}}$
(113)
Recall that the duration in clock time of a path represents the length of the
path – in the discrete case the number of steps. To get the sum of all paths
ending at a specific coordinate time, we need to sum over all possible path
lengths. That is this integral.
As in the SQM we are assuming that once the paths enter the detector they do
not return. And as with SQM, this means we have at best a reasonable first
approximation. However as the SQM approximation works well in practice and as
we are interested only in first corrections to SQM, this should be enough to
achieve falsifiability.
#### 3.2.3 Application to a Gaussian test function
If we construct the wave function as the direct product of time and space
parts:
${\varphi_{\tau}}\left({t,x}\right)={{\tilde{\varphi}}_{\tau}}\left(t\right){{\bar{\varphi}}_{\tau}}\left(x\right)$
(114)
then the expression for the probability current simplifies:
${j_{\tau}}\left({t,x}\right)=\left({\left({\frac{{p_{x}}}{{2m}}\bar{\psi}_{\tau}^{*}\left(x\right)}\right){{\bar{\psi}}_{\tau}}\left(x\right)+\bar{\psi}_{\tau}^{*}\left(x\right)\frac{{p_{x}}}{{2m}}{{\bar{\psi}}_{\tau}}\left(x\right)}\right)\tilde{\varphi}_{\tau}^{*}\left(t\right){{\tilde{\varphi}}_{\tau}}\left(t\right)$
(115)
or:
${D_{\tau}}\left(t\right)={{\bar{D}}_{\tau}}{{\tilde{\rho}}_{\tau}}\left(t\right)$
(116)
Using the time wave function from equation 100 we have for the non-
relativistic probability density in time:
$\tilde{\rho}_{\tau}\left(t\right)=\sqrt{\frac{1}{\pi\sigma_{t}^{2}\left|f_{\tau}^{\left(t\right)}\right|^{2}}}e^{-\frac{1}{\sigma_{t}^{2}\left|f_{\tau}^{\left(t\right)}\right|^{2}}\left(t-\tau\right)^{2}}$
(117)
To get the probability for detection at coordinate time $t$ we convolute clock
time with coordinate time. To solve we look at the limit in long time (as with
equation 21):
$\sigma_{t}^{2}\left|f_{\tau}^{\left(t\right)}\right|^{2}=\sigma_{t}^{2}+\frac{\tau^{2}}{m^{2}\sigma_{t}^{2}}=\sigma_{t}^{2}+\frac{\left(\bar{\tau}+\delta\tau\right)^{2}}{m^{2}\sigma_{t}^{2}}\approx\frac{\left(\bar{\tau}+\delta\tau\right)^{2}}{m^{2}\sigma_{t}^{2}}\approx\frac{\bar{\tau}^{2}}{m^{2}\sigma_{t}^{2}}$
(118)
This gives the probability density:
${{\tilde{\rho}}_{\tau}}\left(t\right)\approx\sqrt{\frac{1}{{\pi\tilde{\sigma}_{\tau}^{2}}}}{e^{-\frac{1}{{\tilde{\sigma}_{\tau}^{2}}}{{\left({t-\tau}\right)}^{2}}}},{{\tilde{\sigma}}_{\tau}}\equiv\frac{{\bar{\tau}}}{{m{\sigma_{t}}}}$
(119)
The convolution over $\tau$ is trivial giving:
${{\rho_{\bar{\tau}}}\left(t\right)=\sqrt{\frac{1}{{\pi\sigma_{\tau}^{2}}}}{e^{-\frac{{{\left({t-\bar{\tau}}\right)}^{2}}}{{\sigma_{\tau}^{2}}}}}}$
(120)
with the total dispersion in clock time being the sum of the dispersions in
coordinate time and in space:
${\sigma_{\tau}^{2}\equiv\tilde{\sigma}_{\tau}^{2}+\bar{\sigma}_{\tau}^{2}}$
(121)
This is reasonably simple. The uncertainty is:
$\Delta\tau=\frac{1}{{\sqrt{2}}}\sqrt{\tilde{\sigma}_{\tau}^{2}+\bar{\sigma}_{\tau}^{2}}$
(122)
We collect the definitions for the two dispersions:
$\begin{array}[]{c}\bar{\sigma}_{\tau}^{2}=\frac{{{\bar{\tau}}^{2}}}{{{m^{2}}{v^{2}}\sigma_{x}^{2}}}\hfill\\\
\tilde{\sigma_{t}}^{2}\approx\frac{{{\bar{\tau}}^{2}}}{{{m^{2}}\sigma_{t}^{2}}}\hfill\end{array}$
(123)
As noted, we expect the particle wave functions to have initial dispersions in
energy/time comparable to their dispersions in momentum/space:
$\sigma_{t}\sim\sigma_{x}$ (124)
In the non-relativistic case, $v\ll 1$ the total uncertainty will be dominated
by the space part. While we have looked specifically at wave functions
composed of a direct product of GTFs in time and space, the underlying
arguments are general. Therefore we expect that in the case of non-
relativistic particles, the uncertainty in time-of-arrival will show evidence
of dispersion in coordinate time, but will be dominated by the dispersion in
space, because this enters with a $\frac{1}{v}$ factor and in the non-
relativistic case $v$ is by definition small.
## 4 Applications
##### Single slit in time
At non-relativistic velocities, we expect that the uncertainty in time-of-
arrival will be dominated by the space part. To increase the effect of
uncertainty in time we can run the wave function through a single slit in
time, i.e. a camera shutter which opens and closes very rapidly. In SQM, the
wave function will merely be clipped and the uncertainty in time at the
detector correspondingly reduced. But in TQM the reduction in $\Delta t$ will
increase $\Delta E$. The increase in uncertainty in energy will cause the wave
function to be more spread out in time, making the uncertainty in time at the
detector arbitrarily great. The case is analogous to the corresponding single
slit in space, with $\Delta E,\Delta t$ substituting for $\Delta p,\Delta x$.
In our previous paper we examined this case, but our analysis took as its
starting point the Kijowski metric and associated approaches. In particular,
we shifted back and forth between $q$ and $p$ in phase space in a way that we
now recognize as suspect.
This obscured an important subtlety in the case of non-relativistic particles.
Consider a non-relativistic particle going through a time gate. If the gate is
open for only a short time, the particle must have come from an area close to
the time gate. Therefore its position in space is also fairly certain. If the
particle is traveling with non-relativistic velocity $v\ll 1$ then $\Delta
x\sim v\Delta t\Rightarrow\Delta x\ll\Delta t$. This in turn drives up the
uncertainty in momentum, $\Delta p\sim{1\mathord{\left/{\vphantom{1{\Delta
x}}}\right.\kern-1.2pt}{\Delta x}}$, leading to increased uncertainty in time
at the detector, even for SQM.
To avoid this problem, we replace the gate with a time-dependent source. We
let the particles propagate for a short time past the source, then compare the
expected uncertainties without and with dispersion in time. The basic
conclusion of the previous work is unchanged: we can make the difference
between the two cases arbitrarily large. We work out the details in appendix
C.
##### Double slit in time
The double slit experiment is often cited as _the_ key experiment for quantum
mechanics (mostly famously by Feynman [54]). And Lindner’s “Attosecond Double-
Slit Experiment” [86] provided a key inspiration for the current work, in that
it showed we could now explore interference effects at the necessary
attosecond scale. Building on this, Palacios et al [87] have proposed an
experiment using spin-correlated pairs of electrons which takes advantage of
the electrons being indistinguishable to look for interference in the time-
energy domain111111Unfortunately as far as we know this experiment has not yet
been performed..
Both experiments use the standard quantum theory to do the analysis of
interference in the time/energy dimension. Horwitz [25, 88, 27] notes that as
time is a parameter in standard quantum mechanics it is difficult to fully
justify this analysis. He provides an alternative analysis in terms of the
Stueckelberg equation (our eq 83) which is not subject to this objection. He
gets the same spacing for the fringes as Lindner found.
Unfortunately for this investigation since the fringe spacing is the same
using the Lindner’s analysis or the one here the fringe pattern does not let
us distinguish between the two approaches; it contributes nothing to
falsifiability.
We do expect, from the analysis of the single slit experiment, that the
individual fringes will be smoothed out from the additional dispersion in
time: the brights less bright, the darks less dark. However the contribution
of this effect to falsifiability is already covered by the analysis of the
single slit experiment.
##### Quantum electrodynamics
While the above is sufficient to establish that we can detect the effects of
uncertainty in time, it is clear that the strongest and most interesting
effects will be seen at high energies and relativistic speeds, where the
effects of quantum field theory (QFT) will need to be taken into account.
In the previous paper, we looked at the QFT approach to systems of spinless,
massive particles. We were able to extend TQM to QFT in a straightforward and
unambiguous way. We also showed that we recover the usual Feynman rules in the
appropriate limit. And that we will see various interesting effects from the
additional uncertainty in time.
One obvious concern was that the additional dimension in the loop integrals
might make the theory still more divergent than QFT already is, perhaps even
unrenormalizable. But in the spin 0 massive case we do not see the usual
ultraviolet (UV) divergences: the combination of uncertainty in time,
entanglement in time, and the use of finite initial wave functions (rather
than plane waves or $\delta$ functions) suppresses the UV divergences without
the need for regularization.
The extension to spin 1/2 particles and to photons appears to be
straightforward. We expect to see the various effects of time-as-observable
for spin 1/2 particles and photons: interference in time, entanglement in
time, correlations in time, and so on. We will also look at the implications
for spin correlations, symmetry/anti-symmetry under exchange of identical
particles, and the like.
One caveat: if we analyze the photon propagator using the familiar trick of
letting the photon mass $\mu\to 0$[89] we can see from inspection of the
$\frac{1}{2\mu}$ factor in the propagator in momentum space:
${K_{\tau}}\left({k;k^{\prime}}\right)=\exp\left({\imath\frac{{{k^{2}}-{\mu^{2}}}}{{2\mu}}\tau}\right){\delta^{4}}\left({k-k^{\prime}}\right)$
(125)
that excursions off-shell will be severely penalized. This in turn will limit
the size of the effects of TQM in experiments such as the single slit in time.
Such experiments are therefore better run with massive particles, the more
massive the better in fact.
In this paper we have estimated the initial wave functions in time using
dimensional and maximum entropy arguments. Once TQM has been extended to
include QED, we should be able to combine the standard QED results with these
to serve as the zeroth term in a perturbation expansion, then compute first
and higher corrections using the standard methods of time-dependent
perturbation theory.
This should open up a wide variety of interesting experiments.
##### General Relativity
As noted in the introduction, TQM is a part of the relativistic dynamics
program so can draw on the extensive relativistic dynamics literature. In
particular we can take advantage of Horwitz’s extension of relativistic
dynamics to General Relativity for single particles, the classical many body
problem, and the quantum many body problem [90, 91, 92].
There would appear to be an interesting complementarity between the work of
Horwitz and this work. First, TQM supplies an explicit source for the
invariant monotonic parameter $\tau$ so that we do not need to introduce this
as an additional assumption. Second, TQM avoids the UV divergences which have
created significant difficulties for work to extend QFT to General Relativity.
Therefore an appropriate combination of the two approaches might provide
useful insights into the extension of QFT to General Relativity.
## 5 Discussion
> “It is not surprising that our language should be incapable of describing
> the processes occurring within the atoms, for, as has been remarked, it was
> invented to describe the experience of daily life, and these consist only of
> processes involving exceedingly large numbers of atoms. Furthermore, it is
> very difficult to modify our language so that it will be able to describe
> these atomic processes, for words can only describe things of which we can
> form mental pictures, and this ability, too, is a result of daily
> experience. Fortunately, mathematics is not subject to this limitation, and
> it has been possible to invent a mathematical scheme — the quantum theory —
> which seems entirely adequate for the treatment of atomic processes…” –
> Werner Heisenberg [93]
###### Falsifiability
The Heisenberg uncertainty principle (HUP) in time follows directly from
quantum mechanics and relativity. This was clear to both Einstein and Bohr in
1930. However quantum mechanics since then has not in general included it.
Given that the relevant attosecond time scales have been too small to work
with – until recently – this is reasonable enough.
However it is now possible to look at the time at the relevant scales; it is
therefore necessary to do so, if we are understand either time or quantum
mechanics. The specific predictions made here are based on only covariance and
the requirement of consistency with existing results, they are by construction
without free parameters. They should therefore provide at a minimum a
reasonable order-of-magnitude target.
The time-of-arrival experiment we have focused on here provides merely the
most obvious line of attack. Essentially any experiment looking at time-
dependent phenomena is a potential candidate. For instance in Auletta [94]
there are about three hundred foundational experiments; most can be converted
into a test of uncertainty in time by replacing an $x$ with a $t.$ Examples
include the single slit in time, the double slit in time, and so on. One may
also look at variations intended to multiply normally small effects, as
experiments using resonance or diffraction effects.
###### Implications of negative result striking
From symmetry, one may argue that a positive result is the more likely, but
even a negative result would be interesting.
For instance, assume we have established there is no HUP in time in one frame.
Consider the same experiment from another frame moving at high velocity
relative to the first. Consider say a Gaussian test function (GTF) in $x$ in
the first; make a Lorentz boost of this into the second frame. The Lorentz
boost will transform $x\to\gamma\left(x-v\tau\right)$. This will turn the
uncertainty in space into a mixture of uncertainty in space and in time. We
then look for uncertainty in time in the boosted frame.
If the HUP in time is also rejected in the second, how do we maintain the
principle of covariance? If the HUP in time is present in the second frame,
then we can defined a preferred frame as the one in which HUP in time is
maximally falsified. Such a preferred frame is anathema to general relativity.
Therefore exploring the precise character of the negative result – uniform
across frames, more in some frames than others, and so on – would itself
represent an interesting research program.
###### Positive result would have significant practical applications
If on the other hand, the wave function extends in time, this would not only
be interesting from the point of view of fundamental physics, it would open up
a variety of practical applications. For instance, there would be an
additional channel for quantum communication systems to use. Memristors and
other time-dependent circuit elements would show interesting effects. In
attosecond chemistry and biochemistry we would expect to see forces of
anticipation and regret; if the wave function extends in time, it will cause
interactions to start earlier and end later than would otherwise be the case.
The mysteries of protein folding could be attacked from a fresh perspective
and perhaps unexpected temporal subtleties found.
The applications in quantum computing are particularly interesting. Quantum
computers will need to compensate for the effects of decoherence along the
time dimension. But they should also be able to take advantage of additional
computing opportunities along the time dimension. And if we have a deeper
understanding of the relationship between time and measurement, we may find
opportunities to “cut across the diagonal” in the design of quantum computers.
I thank my long time friend Jonathan Smith for invaluable encouragement,
guidance, and practical assistance.
I thank Ferne Cohen Welch for extraordinary moral and practical support.
I thank Martin Land, L. P. Horwitz, and the other organizers and participants
of the International Association for Relativistic Dynamics (IARD) 2020
Conference for encouragement, useful discussions, and hosting a talk on this
paper at the IARD 2020 conference.
I thank the reviewer who drew my attention to Horwitz’s [90].
I thank Steven Libby for several useful conversations.
I thank Larry Sorensen for many helpful references. I thank Ashley Fidler for
helpful references to the attosecond physics literature.
And I thank Avi Marchewka for an interesting and instructive conversation
about various approaches to time-of-arrival measurements.
And I note none of the above are in any way responsible for any errors of
commission or omission in this work.
.
## Appendix A Direct computation of the detection rate in diffusion
We use conservation of probability and the method of images to make a direct
computation of the detection rate in the case of diffusion. The approach is
similar to the one used by Marchewka and Schuss to compute the probability
current (subsection 3.1.1), although the context is classical rather than
quantum.
Define ${G_{\tau}}\left(x\right)$ as the probability to get from $d\to x$
without touching the detector at $0$. We get $G$ from the method of images by
the same logic as in the discrete case:
${G_{\tau}}\left(x\right)={P_{\tau}}\left({x;-d}\right)-{P_{\tau}}\left({x;d}\right)$
(126)
$G$ obeys the diffusion equation:
$\frac{{\partial{G_{\tau}}}}{{\partial\tau}}=\frac{1}{{2m}}\frac{{\partial^{2}}}{{\partial{x^{2}}}}{G_{\tau}}$
(127)
Take $D_{\tau}$ as the rate of detection at $\tau$. From conservation of
probability we have:
$1=\int\limits_{-\infty}^{0}{dx{G_{\tau}}\left(x\right)}+\int\limits_{0}^{\tau}{d\tau^{\prime}{D_{\tau^{\prime}}}}$
(128)
Take the derivative with respect to $\tau$ and apply the diffusion equation:
${D_{\tau}}=-\int\limits_{-\infty}^{0}{dx\frac{{\partial{G_{\tau}}}}{{\partial\tau}}\left(x\right)}=-\int\limits_{-\infty}^{0}{dx\frac{1}{{2m}}\frac{{\partial^{2}}}{{\partial{x^{2}}}}G}$
(129)
Since the term on the right is the integral of a derivative we have:
${D_{\tau}}=-\frac{1}{{2m}}{\left.{\frac{\partial}{{\partial
x}}G}\right|_{x=0}}$ (130)
By using the explicit form for $P$ (equation 49) we get:
${D_{\tau}}=\frac{d}{\tau}{P_{\tau}}\left(d\right)$ (131)
as above (equation 51).
## Appendix B Alternate derivation of the detection amplitude
Figure 5: Computation of amplitude for first arrival by using the Laplace
transform
We compute $F$ by using the Laplace transform.
To get to any specific point $x$ we have to get to the point the first time,
then return to it zero or more times. Therefore we can write the full kernel
$K$ as the convolution of the kernel $F$ to arrive for the first time and the
kernel $U$ to return zero or more times:
${K_{\tau}}\left({x;x^{\prime}}\right)=\int\limits_{0}^{\tau}{d\tau^{\prime}{U_{\tau^{\prime}}}\left(x\right){F_{\tau^{\prime}}}\left({x;x^{\prime}}\right)}$
(132)
Since the kernel is invariant under space translation:
$\begin{gathered}{K_{\tau}}\left({x;x^{\prime}}\right)={K_{\tau}}\left({x-x^{\prime}}\right)\hfill\\\
{F_{\tau^{\prime}}}\left({x;x^{\prime}}\right)={F_{\tau}}\left({x-x^{\prime}}\right)\hfill\end{gathered}$
(133)
and:
${U_{\tau}}\left(x\right)={U_{\tau}}\left(0\right)={K_{\tau}}\left(0\right)$
(134)
we can simplify the convolution to:
${K_{\tau}}\left({x-x^{\prime}}\right)=\int\limits_{0}^{\tau}{d\tau^{\prime}{U_{\tau^{\prime}}}F_{\tau^{\prime}}\left({x-x^{\prime}}\right)}$
(135)
The Laplace transform of $K$ is the product of the Laplace transform of $U$
and $F$:
$\mathcal{L}\left[K\right]=\mathcal{L}\left[U\right]\mathcal{L}\left[F\right]$
(136)
The free kernel is:
${K_{\tau}}\left(x\right)=\sqrt{\frac{m}{{2\pi\imath\tau}}}\exp\left({\imath
m\frac{{x^{2}}}{{2\tau}}}\right)$ (137)
with Laplace transform:
$\mathcal{L}\left[K\right]=-\left({-{1^{3/4}}}\right)\frac{{e^{\left({-1+\imath}\right)\sqrt{ms}\left|x\right|}}}{{\sqrt{2}\sqrt{s}}}\sqrt{m}$
(138)
To get the Laplace transform of $U$ take the $x=0$ case:
$\mathcal{L}\left[U\right]=-\frac{{\left({-{1^{3/4}}}\right)}}{{\sqrt{2}\sqrt{s}}}\sqrt{m}$
(139)
Therefore we have the Laplace transform for $F$:
$\mathcal{L}\left[F\right]={e^{\left({-1+\imath}\right)\sqrt{ms}\left|x\right|}}$
(140)
and we get $F_{\tau}$ from the inverse Laplace transform:
$\begin{gathered}\hfill\\\
{F_{\tau}}=\frac{\left|x\right|}{\tau}{K_{\tau}}\left(x\right)\hfill\end{gathered}$
(141)
## Appendix C Single slit in time
We compare the results of the single slit in time in SQM and TQM. We model the
single slit in time as a particle source located at $x=-d$, emitting particles
with momentum $p_{0}$ in the $x$-direction, velocity
${v_{0}}=\frac{{p_{0}}}{m}$.
In the SQM case, we assume that the wave function is emitted with probability
$\bar{G}_{\tau}$:
$\bar{G}\left(\tau\right)\equiv\frac{1}{{\sqrt{2\pi{W^{2}}}}}\exp\left({-\frac{{\tau^{2}}}{{2{W^{2}}}}}\right)$
(142)
This is normalized to one, with uncertainty in time $\Delta\tau=W$.
To extend to TQM we will replace this probability with an amplitude:
${{\tilde{\varphi}}_{0}}\left(t_{0}\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{t}^{2}}}}}\exp\left({-\imath
E_{0}t_{0}-\frac{{t_{0}^{2}}}{{2\sigma_{t}^{2}}}}\right)$ (143)
This has probability distribution and uncertainty:
$\begin{gathered}{{\tilde{\rho}}_{0}}\left({t_{0}}\right)=\sqrt{\frac{1}{{\pi\sigma_{t}^{2}}}}\exp\left({-\frac{{t_{0}^{2}}}{{\sigma_{t}^{2}}}}\right)\\\
\Delta t=\frac{{\sigma_{t}}}{{\sqrt{2}}}\end{gathered}$ (144)
We will take ${\sigma_{t}}=\sqrt{2}W$ so that the uncertainty from the gate is
equal to the uncertainty from the initial distribution in coordinate time, to
make the comparison as fair as possible.
The detector will be positioned at $x=0$. The average time of arrival is
$\bar{t}=\bar{\tau}=\frac{d}{{v_{0}}}$. We are interested in the uncertainty
in time at the detector as defined above for SQM and TQM.
### C.1 Single slit in time in SQM
Figure 6: Single slit in time in SQM
In SQM the wave function extends in only the three space dimensions. If we
break time up into slices, there is no interference across slices. At each
time slice, there is an amplitude for paths from that slice to get to the
detector. But if we look at the individual paths emitted during that time
slice, some get to the detector at one time, some at an earlier or a later
time. At the detector all paths that arrive at the same clock time interfere,
constructively or destructively, as may happen. This is the picture used by
Lindner et al in their analysis. The peaks in the incoming electric wavelet
ejected electrons at different times, but when the electrons ejected from
different times arrived at the detector at the same time they interfered.
To make this specific, we will use a simple model.
For convenience, we will assume that the source has an overall time dependency
of $\exp\left({-\imath\frac{{p_{0}^{2}}}{{2m}}{\tau_{G}}}\right)$ so that our
initial wave function is:
${{\bar{\varphi}}_{G}}\left({x_{G}}\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}{e^{\imath{p_{0}}{x_{G}}-\frac{1}{{2\sigma_{x}^{2}}}x_{G}^{2}-\frac{{p_{0}^{2}}}{{2m}}{\tau_{G}}}}$
(145)
This is normalized to one at the start:
$1=\int{d{x_{G}}\varphi_{G}^{*}\left({x_{G}}\right){\varphi_{G}}\left({x_{G}}\right)}$
(146)
So the particle will have a total probability of being emitted of one:
$1=\int\limits_{-\infty}^{\infty}{d\tau{{\bar{G}}_{\tau}}\int{d{x_{G}}\varphi_{G}^{*}\left({x_{G}}\right){\varphi_{G}}\left({x_{G}}\right)}}$
(147)
The amplitude at the detector from a single moment at the gate will be given
by:
${{\bar{\varphi}}_{DG}}\left(x\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}\sqrt{\frac{1}{{f_{DG}^{\left(x\right)}}}}{e^{\imath{p_{0}}x-\frac{1}{{2\sigma_{x}^{2}f_{DG}^{\left(x\right)}}}{{\left({x+d-{v_{0}}{\tau_{DG}}}\right)}^{2}}-\imath\frac{{p_{0}^{2}}}{{2m}}{\tau_{D}}}}$
(148)
with ancillary definitions:
$\begin{gathered}{\tau_{DG}}\equiv{\tau_{D}}-{\tau_{G}}\hfill\\\
f_{DG}^{\left(x\right)}=1+\imath\frac{{\tau_{DG}}}{{m\sigma_{x}^{2}}}\hfill\end{gathered}$
(149)
We have:
$x+d-{v_{0}}\left({{\tau_{D}}-{\tau_{G}}}\right)=-{v_{0}}\left({\delta\tau-{\tau_{G}}}\right)$
(150)
Both $\delta\tau$ and $\tau_{G}$ are expected small. We can therefore justify
taking:
$f_{DG}^{\left(x\right)}=1+\imath\frac{{\tau_{DG}}}{{m\sigma_{x}^{2}}}\approx
f_{\bar{\tau}}^{\left(x\right)}=1+\imath\frac{{\bar{\tau}}}{{m\sigma_{x}^{2}}}$
(151)
Giving:
${{\bar{\varphi}}_{DG}}\left(x\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}\sqrt{\frac{1}{{f_{\vec{\tau}}^{\left(x\right)}}}}{e^{\imath{p_{0}}x-\frac{{v_{0}^{2}}}{{2\sigma_{x}^{2}f_{\vec{\tau}}^{\left(x\right)}}}{{\left({\delta\tau-{\tau_{G}}}\right)}^{2}}-\imath\frac{{p_{0}^{2}}}{{2m}}{\tau_{D}}}}$
(152)
To get the full wave function at the detector we need to take the convolution
of this with the gate function:
${{\bar{\psi}}_{D}}\left(x\right)=\int\limits_{-\infty}^{\infty}{d{\tau_{G}}\bar{G}\left({\tau_{G}}\right){{\bar{\varphi}}_{DG}}\left(x\right)}$
(153)
giving:
${{\bar{\psi}}_{D}}\left(x\right)=\frac{{\sigma_{x}}}{{v_{0}}}\sqrt[4]{{\frac{1}{{\pi\sigma_{x}^{2}}}}}\frac{1}{{\sqrt{2\pi\left({\frac{{\sigma_{x}^{2}}}{{v_{0}^{2}}}+{W^{2}}+\imath\frac{{\bar{\tau}}}{{mv_{0}^{2}}}}\right)}}}\exp\left({\imath{p_{0}}x-\frac{{{\left({\delta\tau}\right)}^{2}}}{{2\left({\frac{{\sigma_{x}^{2}}}{{v_{0}^{2}}}+{W^{2}}+\imath\frac{{\bar{\tau}}}{{mv_{0}^{2}}}}\right)}}-\imath\frac{{p_{0}^{2}}}{{2m}}{\tau_{D}}}\right)$
(154)
We can see that the effect of the gate is to increase the effective dispersion
in space:
$\frac{{\sigma_{x}^{2}}}{{v_{0}^{2}}}\to\frac{{\sigma_{x}^{2}}}{{v_{0}^{2}}}+{W^{2}}\Rightarrow\sigma_{x}^{2}\to\Sigma_{x}^{2}\equiv\sigma_{x}^{2}+v_{0}^{2}{W^{2}}$
(155)
The gate effectively adds an uncertainty of $vW$ to the original uncertainty
in space. As $vW$ is about the distance a particle would cross while the gate
is open, this is reasonable.
We get then as the associated uncertainty in time (eq: 76):
$\Delta\tau=\frac{1}{{\sqrt{2}}}{{\bar{\Sigma}}_{\tau}}=\frac{1}{{\sqrt{2}}}\frac{1}{{m{v_{0}}{\Sigma_{x}}}}\bar{\tau}$
The longer the gate stays open the greater the resulting uncertainty in time
at the detector. The shorter the gate stays open the less the uncertainty in
time at the detector, with the minimum being the uncertainty for a free GTF
released at the time and location of the gate.
### C.2 Single slit in time in TQM
Figure 7: Single slit in time in TQM
In TQM the wave function extends in all four dimensions. There is interference
across time so it is no longer legitimate to break time up into separate
slices (except for purposes of analysis of course). We will take the source as
centered at a specific moment in coordinate time, $t=0$.
We can write the four dimensional wave function as the product of the time
(eq: 143) and space parts:
${\varphi_{0}}\left({{t_{0}},{{\vec{x}}_{0}}}\right)={\tilde{\varphi}_{0}}\left({t_{0}}\right){\bar{\varphi}_{0}}\left({{\vec{x}}_{0}}\right)$
(156)
The space part is as above for clock time $\tau=0$.
The time part at clock time $\tau$ is:
>
> ${{\tilde{\varphi}}_{\tau}}\left(t\right)=\sqrt[4]{{\frac{1}{{\pi\sigma_{t}^{2}}}}}\sqrt{\frac{1}{{f_{\tau}^{\left(t\right)}}}}{e^{-\imath{E_{0}}t-\frac{1}{{2\sigma_{t}^{2}f_{\tau}^{\left(t\right)}}}{{\left({t-\frac{{E_{0}}}{m}\tau}\right)}^{2}}}}$
> (157)
For a non-relativistic particle $\frac{E}{m}\approx 1$. The treatment in
subsection 3.2.3 applies giving the uncertainty in time at the detector as:
$\Delta\tau=\frac{1}{{\sqrt{2}}}\sqrt{\tilde{\sigma}_{\tau}^{2}+\bar{\sigma}_{\tau}^{2}}$
(158)
with contributions from the space and time parts of:
$\begin{gathered}\bar{\sigma}_{\tau}^{2}=\frac{{{\bar{\tau}}^{2}}}{{{m^{2}}{v^{2}}\sigma_{x}^{2}}}\hfill\\\
\tilde{\sigma}_{\tau}^{2}=\frac{{{\bar{\tau}}^{2}}}{{{m^{2}}\sigma_{t}^{2}}}=\frac{{{\bar{\tau}}^{2}}}{{2{m^{2}}{W^{2}}}}\hfill\end{gathered}$
(159)
So we have the scaled uncertainty in time as:
$\frac{{\Delta\tau}}{{\bar{\tau}}}=\frac{1}{{\sqrt{2}}}\frac{1}{m}\sqrt{\frac{1}{{{v^{2}}\sigma_{x}^{2}}}+\frac{1}{{2{W^{2}}}}}$
(160)
We can see that when $W\sim v\sigma_{x}$ the effects of dispersion in time
will be comparable to those from dispersion in space. And as $W\to 0$, the
uncertainty in time at the detector will be dominated by the width of the gate
in time, going as $\frac{1}{W}$.
So the intrinsic uncertainty in space of a GTF creates a corresponding minimum
uncertainty in time at the detector given by $\bar{\sigma}_{\tau}$. In SQM the
effects of the gate drop to zero as $W\to 0$ while in TQM they go to infinity.
In SQM the wave function is clipped in time; in TQM it is diffracted.
We therefore have the unambiguous signal – even in the non-relativistic case –
that we need to achieve practical falsifiability.
## References
* [1] Feynman R P 1965 The Character of Physical Law (Modern Library) ISBN 0 679 60127 9
* [2] Schilpp P A and Bohr N 1949 Discussion with Einstein on Epistemological Problems in Atomic Physics (Evanston: The Library of Living Philosophers) pp 200–241
* [3] Pais A 1982 Subtle is the Lord: The Science and Life of Albert Einstein (Oxford University Press)
* [4] Hilgevoord J 1996 American Journal of Physics 64 1451–1456
* [5] Hilgevoord J 1998 American Journal of Physics 66 396–402
* [6] Busch P 2001 The Time Energy Uncertainty Relation (Berlin: Springer-Verlag) pp 69–98 Lecture Notes in Physics
* [7] Pauli W 1980 General Principles of Quantum Mechanics (Springer-Verlag) URL http://www.springer.com/us/book/9783540098423
* [8] Dirac P A M 1958 General Principles of Quantum Mechanics 4th ed International series of monographs on physics (Oxford, Clarendon Press)
* [9] Muga J G, Sala Mayato R and Egusquiza I L 2002 Time in Quantum Mechanics (Berlin; New York: Springer) ISBN 3540432949
* [10] Muga J G, Sala Mayato R and Egusquiza I L 2008 Time in Quantum Mechanics - Vol 2 (Berlin; New York: Springer-Verlag) ISBN 3540734724 9783540734727 0075-8450 ;
* [11] Ossiander M, Siegrist F, Shirvanyan V, Pazourek R, Sommer A, Latka T, Guggenmos A, Nagele S, Feist J, Burgdörfer J, Kienberger R and Schultze M 2016 Nature Physics 13 280
* [12] Abbott E A 1884 Flatland: A Romance of Many Dimensions (Seeley and Co. of London)
* [13] Ashmead J 2019 Journal of Physics: Conference Series 1239 012015 URL https://iopscience.iop.org/article/10.1088/1742-6596/1239/1/012015
* [14] Stueckelberg E C G 1941 Helv. Phys. Acta. 14 51
* [15] Stueckelberg E C G 1941 Helv. Phys. Acta. 14 322–323
* [16] Feynman R P 1948 Rev. Mod. Phys. 20(2) 367–387
* [17] Feynman R P 1949 Phys Rev 76 749–759
* [18] Feynman R P 1949 Phys Rev 76 769–789
* [19] Feynman R P 1950 Physical Review 80 440–457
* [20] Horwitz L P and Piron C 1973 Helvetica Physica Acta 46
* [21] Fanchi J R and Collins R E 1978 Found Phys 8 851–877
* [22] Fanchi J R 1993 Found. Phys. 23
* [23] Fanchi J R 1993 Parameterized Relativistic Quantum Theory (Fundamental Theories of Physics vol 56) (Kluwer Academic Publishers)
* [24] Land M C and Horwitz L P 1996 ArXiv e-prints (Preprint hep-th/9601021v1) URL http://lanl.arxiv.org/abs/hep-th/9601021v1
* [25] Horwitz L P 2006 Physics Letters A 355 1–6 (Preprint quant-ph/0507044) URL http://arxiv.org/abs/quant-ph/0507044
* [26] Fanchi J R 2011 Found Phys 41 4–32
* [27] Horwitz L P 2015 Relativistic Quantum Mechanics Fundamental Theories of Physics (Springer Dordrecht Heidelberg New York London)
* [28] Kijowski J 1974 Reports on Mathematical Physics 6 361–386
* [29] Marchewka A and Schuss Z 1998 Phys.Lett. A240 177–184 (Preprint quant-ph/9708034) URL https://arxiv.org/abs/quant-ph/9708034
* [30] Marchewka A and Schuss Z 1999 Phys. Lett. A 65 042112 (Preprint quant-ph/9906078) URL https://arxiv.org/abs/quant-ph/9906078
* [31] Marchewka A and Schuss Z 1999 ArXiv e-prints (Preprint quant-ph/9906003) URL https://arxiv.org/abs/quant-ph/9906003
* [32] Marchewka A and Schuss Z 2000 Phys. Rev. A 61 052107 (Preprint quant-ph/9903076) URL https://arxiv.org/abs/quant-ph/9903076
* [33] Muga J G and Leavens C R 2000 Physics Reports 338 353–438
* [34] Baute A D, Mayato R S, Palao J P, Muga J G and Egusquiza I L 2000 Phys.Rev.A 61 022118 (Preprint quant-ph/9904055) URL https://arxiv.org/abs/quant-ph/9904055
* [35] Baute A D, Egusquiza I L, Muga J G and Sala-Mayato R 2000 Phys. Rev. A 61 052111 (Preprint quant-ph/9911088) URL https://arxiv.org/abs/quant-ph/9911088
* [36] Baute A D, Egusquiza I L and Muga J G 2001 Phys. Rev. A 64 012501 (Preprint quant-ph/0102005) URL https://arxiv.org/abs/quant-ph/0102005
* [37] Ruggenthaler M, Gruebl G and Kreidl S 2005 Journal of Physics A: Mathematical and General 38 (Preprint quant-ph/0504185) URL https://arxiv.org/abs/quant-ph/0504185
* [38] Anastopoulos C and Savvidou N 2006 J. Math. Phys. 47 122106 (Preprint quant-ph/0509020) URL https://arxiv.org/abs/quant-ph/0509020
* [39] Yearsley J M 2010 ArXiv e-prints (Preprint 1012.2575v1) URL https://arxiv.org/abs/1012.2575v1
* [40] Kiukas J, Ruschhaupt A, Schmidt P O and Werner R F 2012 Journal of Physics A: Mathematical and Theoretical 45 (Preprint 1109.5087) URL https://arxiv.org/abs/1109.5087
* [41] Yearsley J M, Downs D A, Halliwell J J and Hashagen A K 2011 Phys. Rev. A 84 022109 (Preprint 1106.4767) URL https://arxiv.org/abs/1106.4767
* [42] Halliwell J J, Evaeus J, London J and Malik Y 2015 Physics Letters A 379 2445 (Preprint 1504.02509) URL https://arxiv.org/abs/1504.02509
* [43] Kijowski J, Waluk P and Senger K 2015 ArXiv e-prints (Preprint 1512.02867) URL https://arxiv.org/abs/1512.02867
* [44] Das S 2018 ArXiv e-prints (Preprint 1802.01863) URL https://arxiv.org/abs/1802.01863
* [45] Yearsley J M 2011 ArXiv e-prints (Preprint 1110.5790) URL https://arxiv.org/abs/1110.5790
* [46] Rayleigh L 1885 Proceedings of the London Mathematical Society s1-17 4–11 ISSN 0024-6115 (Preprint https://academic.oup.com/plms/article-pdf/s1-17/1/4/4368144/s1-17-1-4.pdf)
* [47] Bell J S 1964 Physics 1 195–200
* [48] Clauser J F, Horne M A, Shimony A and Holt R A 1969 Phys. Rev. Lett. 23(15) 880–884 URL htt
* [49] Aspect A, Grangier P and Roger G 1981 Phys. Rev. Lett. 47(7) 460–463
* [50] Aspect A, Grangier P and Roger G 1982 Phys. Rev. Lett. 49(2) 91–94
* [51] Aspect A, Dalibard J and Roger G 1982 Phys. Rev. Lett. 49(25) 1804–1807
* [52] Hensen B, Bernien H, Dréau A E, Reiserer A, Kalb N, Blok M S, Ruitenberg J, Vermeulen R F L, Schouten R N, Abellán C, Amaya W, Pruneri V, Mitchell M W, Markham M, Twitchen D J, Elkouss D, Wehner S, Taminiau T H and Hanson R 2015 Nature 526 682–686
* [53] Muga J G, Baute A D, Damborenea J A and Egusquiza I L 2000 ArXiv e-prints (Preprint quant-ph/0009111) URL https://arxiv.org/abs/quant-ph/0009111
* [54] Feynman R P, Leighton R B and Sands M 1965 The Feynman Lectures on Physics vol Volumes I, II, and III (Addison-Wesley)
* [55] Sakurai J J 1967 Advanced Quantum Mechanics (Addison-Wesley) ISBN 0 201 06710 2
* [56] Einstein A, Podolsky B and Rosen N 1935 Phys. Rev. 47 777–780
* [57] Bohr N 1935 Physical Review 48 696–702
* [58] Feynman R P, Hibbs A R and Styer D F 2010 Quantum Mechanics and Path Integrals (Mineola, N.Y.: Dover Publications) ISBN 9780486477220 0486477223 URL http://www.loc.gov/catdir/enhancements/fy1006/2010004550-d.html
* [59] Feller W 1968 An Introduction to Probability Theory and Its Applications vol I (John Wiley and Sons, Inc.)
* [60] Biggs N L 1989 Discrete Mathematics (Oxford Science Publications)
* [61] Graham R L, Knuth D E and Patashnik O 1994 Concrete Mathematics: A Foundation for Computer Science (Addison-Wesley Professional) ISBN 0201558025
* [62] Ibe O C 2013 Elements of Random Walk and Diffusion Processes (Hoboken, N.J.: John Wiley & Sons, Inc.) ISBN 1-118-61793-2; 1-118-61805-X; 1-118-62985-X
* [63] Grimmett G R and Stirzaker D R 2001 Probability and Random Processes 3rd ed (Oxford University Press)
* [64] Petersen A 1963 The Bulletin of the Atomic Scientists 19
* [65] Olsen J D and McDonald K T 2017 URL reference (Preprint https://www.physics.princeton.edu/ mcdonald/examples/orbitdecay.pdf) URL https://www.physics.princeton.edu/ mcdonald/examples/orbitdecay.pdf
* [66] Camilleri K and Schlosshauer M 2015 Stud. Hist. Phil. Mod. Phys. 49, 73-83 (2015) (Preprint 1502.06547v1) URL http://arxiv.org/abs/1502.06547v1
* [67] Zurek W H 1991 Physics Today 44 33–44 URL https://arxiv.org/abs/quant-ph/0306072
* [68] Omnès R 1994 The Interpretation of Quantum Mechanics (Princeton: Princeton University Press) ISBN 0 691 03669 1
* [69] Giulini D, Joos E, Kiefer C, Kupsch J, Stamatescu I O and Zeh H D 1996 Decoherence and the Appearance of a Classical World in Quantum Theory (Berlin: Springer-Verlag)
* [70] Zeh H D 1996 ArXiv e-prints (Preprint quant-ph/9610014) URL https://arxiv.org/abs/quant-ph/9610014
* [71] Zeh H D 2000 Lect.Notes Phys 538 19–42 URL https://arxiv.org/abs/quant-ph/9905004
* [72] Kim Y S and Noz M E 2003 Optics and Spectroscopy 94 (Preprint quant-ph/0206146) URL https://arxiv.org/pdf/quant-ph/0206146.pdf
* [73] Joos E 2003 Decoherence and the Appearance of a Classical World in Quantum Theory 2nd ed (Berlin: Springer) ISBN 3540003908 (acid-free paper) URL http://www.loc.gov/catdir/toc/fy038/2003042813.html
* [74] Schlosshauer M A 2007 Decoherence and the Quantum-to-Classical Transition (Berlin: Springer) ISBN 9783540357735 (hbk.) URL http://www.loc.gov/catdir/enhancements/fy0814/2007930038-b.html
* [75] Carmichael H 1991 An Open Systems Approach to Quantum Optics: Lectures Presented at the Université Libre de Bruxelles, October 28 to November 4, 1991 (Springer)
* [76] Dalibard J, Castin Y and Mølmer K 1992 Phys. Rev. Lett. 68(5) 580–583
* [77] Dum R, Zoller P and Ritsch H 1992 Phys. Rev. A 45(7) 4879–4887
* [78] Hegerfeldt G C and Wilser T S 1991 Classical and Quantum Systems: Foundations and Symmetries ed HD Doebner; W Scherer; F Schroeck J (World Scientific)
* [79] Kovachy T, Asenbaum P, Overstreet C, Donnelly C A, Dickerson S M, Sugarbaker A, Hogan J M and Kasevich M A 2015 Nature 528 530–533 URL https://doi.org/10.1038/nature16155
* [80] Asenbaum P, Overstreet C, Kovachy T, Brown D D, Hogan J M and Kasevich M A 2017 Phys. Rev. Lett. 118(18) 183602 URL https://link.aps.org/doi/10.1103/PhysRevLett.118.183602
* [81] Wells H G 1935 The Time Machine (Everyman) URL http://www.fourmilab.ch/etexts/www/wells/timemach/html/
* [82] Jackson J J, Puretzky A A, More K L, Rouleau C M, Eres G and Geohegan D B 2010 ACS Nano 4 7573–7581 URL https://doi.org/10.1021/nn102029y
* [83] Goldstein H 1980 Classical Mechanics Second Edition (Reading, MA: Addison-Wesley)
* [84] Merzbacher E 1998 Quantum Mechanics (New York: John Wiley and Sons, Inc.)
* [85] Pashby T 2014 Time and the Foundations of Quantum Mechanics Ph.D. thesis Dietrich School of Arts and Sciences, University of Pittsburgh URL http://d-scholarship.pitt.edu/21183/1/Pashby2014TimeFoundationsQM.pdf
* [86] Lindner F, Schätzel M G, Walther H, Baltuška A, Goulielmakis E, Krausz F, Milošević D B, Bauer D, Becker W and Paulus G G 2005 Physical Review Letters 95 040401 URL http://link.aps.org/doi/10.1103/PhysRevLett.95.040401 http://lanl.arxiv.org/abs/quant-ph/0503165
* [87] Palacios A, Rescigno T N and McCurdy C W 2009 Phys. Rev. Lett. 103(25) 253001 URL https://link.aps.org/doi/10.1103/PhysRevLett.103.253001
* [88] Horwitz L P 2007 Foundations of Physics 37 734–746 URL https://doi.org/10.1007/s10701-007-9127-7
* [89] Zee A 2010 Quantum Field Theory in a Nutshell (Princeton, N.J.: Princeton University Press) ISBN 9780691140346 (hardcover alk. paper)
* [90] Horwitz L P 2019 The European Physical Journal Plus 134 313 (Preprint 1810.09248) URL https://arxiv.org/abs/1810.09248
* [91] Horwitz L P 2020 The European Physical Journal Plus 135 479 URL https://doi.org/10.1140/epjp/s13360-020-00446-0
* [92] Horwitz L P 2021 The European Physical Journal Plus 136 32 URL https://doi.org/10.1140/epjp/s13360-020-00967-8
* [93] Heisenberg W 1930 The Physical Principles of the Quantum Theory (Chicago: University of Chicago Press)
* [94] Auletta G 2000 Foundations and Interpretation of Quantum Mechanics: In the Light of a Critical-Historical Analysis of the Problems and of a Synthesis of the Results (Singapore: World Scientific Publishing Co. Pte. Ltd.)
|
# Helioseismic Properties of Dynamo Waves in the Variation of Solar
Differential Rotation
Krishnendu Mandal New Jersey Institute of Technology, Newark, NJ 07102, USA
Alexander G. Kosovichev New Jersey Institute of Technology, Newark, NJ 07102,
USA NASA Ames Research Center, Moffett Field, CA 94035, USA Valery V. Pipin
Institute of Solar-Terrestrial Physics, Russian Academy of Sciences, Irkutsk,
664033, Russia
(Received –; Revised –; Accepted –)
###### Abstract
Solar differential rotation exhibits a prominent feature: its cyclic
variations over the solar cycle, referred to as zonal flows or torsional
oscillations, are observed throughout the convection zone. Given the challenge
of measuring magnetic fields in subsurface layers, understanding deep
torsional oscillations becomes pivotal in deciphering the underlying solar
dynamo mechanism. In this study, we address the critical question of
identifying specific signatures within helioseismic frequency-splitting data
associated with the torsional oscillations. To achieve this, a comprehensive
forward modeling approach is employed to simulate the helioseismic data for a
dynamo model that, to some extent, reproduces solar-cycle variations of
magnetic fields and flows. We provide a comprehensive derivation of the
forward modeling process utilizing generalized spherical harmonics, as it
involves intricate algebraic computations. All estimated frequency-splitting
coefficients from the model display an $11$-year periodicity. Using the
simulated splitting coefficients and realistic noise, we show that it is
possible to identify the dynamo wave signal present in the solar zonal flow
from the tachocline to the solar surface. By analyzing observed data, we find
similar dynamo wave patterns in the observational data from MDI, HMI, and
GONG. This validates the earlier detection of dynamo waves and holds potential
implications for the solar dynamo theory models.
Sun: waves – Sun: oscillations – solar interior – solar convection zone –
solar differential rotation
## 1 Introduction
Parker (1955) showed that dynamo action involves cyclic transformations of the
poloidal and toroidal magnetic fields of the Sun. This scenario suggests that
the magnetic field of bipolar magnetic regions is formed from the large-scale
toroidal magnetic field that is generated from the axisymmetric poloidal
magnetic field, twisted by differential rotation deep in the convection zone.
As the solar dynamo revolves around comprehending the cyclic evolution of
toroidal and poloidal magnetic fields, there is a consensus on the generation
of the toroidal field. However, the generation process of the poloidal field
remains less understood. Parker (1955) suggested that the poloidal magnetic
field is generated from the toroidal field by electromotive force excited by
cyclonic convection which is known as “$\alpha-$effect”. Several alternative
mechanisms for the poloidal field generation can be found in the literature
(Babcock, 1961; Choudhuri & Dikpati, 1999; Charbonneau & Barlet, 2011). An
alternative scenario, known as the flux transport model proposed by Babcock
(1961); Leighton (1969), posits that the cyclical evolution of solar activity
is influenced by meridional circulation. This circulation is responsible for
transporting the magnetic field of decaying sunspots from low latitudes toward
the polar region, as outlined by Dikpati et al. (2009). The transported
magnetic field subsequently undergoes reconnection with the polar field of the
preceding solar cycle and migrates to the base of the convection zone. In the
tachocline, this magnetic field is carried equatorward by the meridional flow
at the base of the convection zone. Eventually, it surfaces due to buoyancy
instability, resulting in the formation of observable sunspots. This scenario
gains validation through the observed migration of sunspots, originating from
mid-latitudes at the onset of the solar cycle and progressively migrating
towards the equator by its culmination. Observational evidence further
indicates that the polar magnetic field attains its peak during sunspot
minima, with its magnitude correlating with the subsequent cycle’s sunspot
number (Schatten et al., 1978). This information enables forecasting the
magnitude of the upcoming solar cycle (Choudhuri et al., 2007). Using this
technique, the forecast window spans $5$ to $6$ years, half the duration of
the $11$-year solar cycle. This supports the propositions by Parker (1955) and
Babcock (1961) that the toroidal magnetic field is generated through the
stretching of the poloidal magnetic field in polar regions. Whether this
conveyor belt scenario, which depends on the structure of the meridional flow,
is responsible for flux transport dynamo was under scrutiny. Helioseismology
has found that it can either be single cell (Rajaguru & Antia, 2015; Mandal et
al., 2018; Gizon et al., 2020) or double cell meridional flow (Zhao et al.,
2013; Chen & Zhao, 2017). However, all the studies to detect the deep
structure of the meridional flow suffer from centre-to-limb systematics in the
time-distance helioseismology measurements (Zhao et al., 2012). The
contribution from the systematic and random errors is comparable in magnitude
to the signal itself (Stejko et al., 2022). Aside from this difficulty, both
the surface flux transport and dynamo-waves can reproduce the basic observed
features of magnetic field evolution observed on the solar surface, including
the correlation of the polar magnetic field during the magnetic cycle minimum
with the subsequent sunspot cycle magnitude. The surface observations are not
sufficient to distinguish between two dynamo scenarios. Currently, we can not
measure sub-surface magnetic fields. Thus, the information about the dynamo
process comes from the measurement of large-scale subsurface flows such as
differential rotation and meridional flow. It is expected that variations of
these flows over the scale of $11$ years are due to the solar-cycle evolution
of subsurface magnetic fields. It is to be noted that these variations are not
only due to magnetic torque. Other inertial forces can also contribute to this
variation. Nevertheless observed variations in subsurface flows provide an
important clue about the mechanism of solar dynamo.
The torsional oscillations were discovered by Howard & Labonte (1980). After
subtracting the mean rotation, the data revealed alternating zones of fast and
slow zonal flow bands, which originated at mid-latitudes and subsequently
migrated to the equator just like the magnetic butterfly diagram. It was then
suggested that these zone flows (called torsional oscillations because of
their cyclic variations) are due to the action of the magnetic field of active
regions. The observed zonal flow pattern is long-living and coherent over
essentially the whole solar circumference. Therefore, it can not be of
convective origin but may be induced by the dynamo-generated magnetic fields
and associated processes. Numerical 3D simulation by Guerrero et al. (2016)
suggested that torsional oscillation in the model is due to the magnetic
torque at the base of the convection zone. However, modulation of the
turbulent heat transport by the magnetic field in the convection zone results
in variations in the zonal and meridional flows, producing the ‘extended’
22-year cycle of the torsional oscillations (Pipin & Kosovichev, 2019).
Helioseismology can probe the subsurface flow profile of solar zonal flow
(Howe et al., 2018; Basu & Antia, 2019). Kosovichev & Schou (1997) found that
torsional oscillations are extended beneath the solar surface. Later Vorontsov
et al. (2002) found that torsional oscillations are present in the entire
convection zone with the phase propagating poleward and equatorward from mid-
latitude at all depths throughout the convective envelope. Pipin & Kosovichev
(2019) suggested that torsional oscillation can be helpful for advanced
prediction of solar cycles, $1-2$ sunspot cycles ahead. Kosovichev & Pipin
(2019) argued that the amplitude of zonal acceleration in the high-latitude
region of the base of the convection zone during solar maxima may give
information about the strength of following solar cycle maxima. If confirmed,
this relationship would give the solar activity forecast a full $11$-year
cycle ahead. Pipin & Kosovichev (2020) argued that the sign of the correlation
between the flow amplitude in the deep convection zone and the strength of the
subsequent solar cycle depends on the properties of long-term variations of
the dynamo cycles.
Recognizing the significance of conclusions drawn from the aforementioned
studies, we systematically revisit the problem. We conduct synthetic tests to
validate the detection of the dynamo wave patterns within the solar zonal
flow. Utilizing the variations of the zonal flow velocity from the dynamo
model proposed by Pipin & Kosovichev (2020), we calculate associated
coefficients of rotational splitting of frequencies of solar oscillations
corresponding to the p-modes observed by the MDI. HMI, and GONG instruments.
The flow variations depicted in Figure 1, feature two distinct branches of
faster-than-average rotational components akin to the observed solar zonal
flows. Our aim is to determine the precision with which we can reconstruct the
zonal flows in the solar convection zones, from the tachocline to the solar
surface by using helioseismic data. Through these tests, we demonstrate the
high accuracy achievable in capturing the zonal flow across this entire depth
range if the observational data are averaged over two-year periods. We then
implement this methodology for measuring the zonal flow across the convection
zone using the current helioseismic data from MDI, HMI, and GONG.
## 2 Data analysis and methods
### 2.1 Rotational Frequency Splitting Data
We utilize a comprehensive set of global helioseismology data products
obtained from various instruments, including space-based ones like the
Helioseismic Magnetic Imager (HMI) (HMI; Schou et al., 2012) on board the
Solar Dynamic Observatory (SDO) and Michelson Doppler Imager (MDI) (MDI;
Scherrer et al., 1995) on board the Solar and Heliospheric Observatory (SoHO),
as well as ground-based instrument Global Oscillation Network Group (GONG)
(GONG; Hill et al., 1996). Each acoustic mode is characterized by a triplet of
quantum numbers, denoted as $(n,\ell,m)$, where $n$ signifies the radial
order, $\ell$ represents the spherical harmonic degree, and $m$ designates the
azimuthal order. In a spherically symmetric model, the mode frequency
$\nu_{n\ell m}$ remains independent of the azimuthal order, $m$. However, the
presence of axisymmetric perturbations, such as differential rotation, breaks
this degeneracy in $m$ which can be expressed as
$\nu_{n\ell
m}=\nu_{n\ell}+\sum_{s}^{s_{\text{max}}}a_{s}(n,\ell)\mathcal{P}_{s}^{\ell}(m),$
(1)
where $\nu_{n\ell}$ is the central frequency of a frequency multiple of radial
order $n$ and angular degree $\ell$; $a_{s}$ are the rotational splitting
coefficients (called the a-coefficients). The odd-order coefficients,
$a_{1},a_{3},a_{5}$, etc., capture information about the solar rotation.
$\mathcal{P}_{s}^{\ell}$ are polynomials of degree $s$ (Ritzwoller & Lavely,
1991), and these polynomials have orthogonal properties. The MDI data
encompass the period from May $1996$ to February $2011$, while the SDO/HMI
data used in this study cover the period from April $2010$ to May $2023$. We
collect solar frequency data and their corresponding splitting information
from distinct, non-overlapping $72$-day time series. The time series data for
the MDI and HMI instruments are sourced from the SDO Joint Science Operations
Center (JSOC) website111JSOC: http://jsoc.stanford.edu/. To investigate the
temporal evolution of solar zonal flow across the entire duration of our
analysis, we utilize a consolidated dataset comprising both MDI and HMI
instruments. In the overlapping time frame of both instruments, we rely on the
data exclusively from HMI.
GONG’s data span from May $1995$ to January $2023$. The data are categorized
by GONG’s “months”, each spanning a duration of 36 days. Solar oscillation
frequencies and splitting coefficients are derived through the analysis of a
108-day period, which is equal to observation of 3 GONG’s months. There is an
overlap of $72-$ days between different data sets.
Figure 1: The zonal flow velocities, derived from the dynamo model of Pipin &
Kosovichev (2020), are depicted across various depths. Notably, these
velocities exhibit both equatorward and poleward migrating branches stemming
from the mid-latitude area. Employing this zonal flow velocity, forward
modeling is conducted to estimate splitting coefficients. The right panel
illustrates the temporal evolution of splitting coefficients measured in nHz.
### 2.2 Forward Modelling Approach
The solar rotational velocity can be expressed as the following
$v(r,\theta)=\sum_{s=1,3,5,\cdots}w_{s}(r)\partial_{\theta}Y_{s,0}(\theta,\phi)$
(2)
where $Y_{s,0}(\theta,\phi)$ is a spherical harmonic of angular degree $s$ and
zero azimuthal order, $w_{s}(r)$ are the radial coefficients in the velocity
expansion, $r$ is the radius from the solar center, $\theta$ is the co-
latitude, and $\phi$ is the azimuthal angle. Rotation leads to the splitting
of p and f-mode frequencies with respect to the azimuthal order, $m$, of the
oscillation eigenfunctions. This phenomenon can be expressed through the
following integral equation:
$a_{s}(n,\ell)=\int_{0}^{R}K_{s}(n,\ell;r)w_{s}(r)r^{2}dr,$ (3)
Here, $K_{s}(n,\ell)$ represents the sensitivity kernel for the mode
$(n,\ell),$ determining the extent to which this specific mode is influenced
by rotation, characterized by a spatial scale $s$. The precise expression of
this function can be found in the appendix (see Appendix A). By leveraging
vector spherical harmonics, we have formulated the kernel expression. A
similar approach was previously employed by Ritzwoller & Lavely (1991). The
beauty of using the generalized spherical harmonics lies in their capability
to enable expressing derivatives with respect to $\theta$ and $\phi$ using
vector spherical harmonics itself. Consequently, the entire derivation
transforms into an elegant algebraic exercise. The use of generalized
spherical harmonics has recently become prevalent in studies focusing on
normal mode coupling (see e.g. Hanasoge et al., 2017). For an in-depth
derivation, please refer to Section A. We present the entire derivation due to
the availability of the splitting a-coefficients using various normalizations.
This often necessitates converting these coefficients into a preferred form
for ease of use. Equation 3 allows us to estimate the a-coefficients provided
we have a known rotation profile, denoted as $v(r,\theta)$, which is
decomposed into a spherical harmonic basis to derive $w_{s}$ using Equation 2.
Figure 2: The splitting coefficients’ variation over time is depicted here.
Left panel: we compare average a-coefficients measured in nHz for all the
modes with the turning point radius of the p-modes between $0.9R_{\odot}$ and
$1R_{\odot}$ by using data from three instruments, SOHO/MDI (blue marker),
SDO/HMI (red marker), and GONG (black marker). Right panel: we limit our
analysis to a maximum harmonic degree of $120$ for MDI. This results in
consistency across all three instruments.
### 2.3 Inversion Method
In the preceding section, we discussed the application of forward modeling
using a predefined rotation or zonal flow profile. These artificially
generated a-coefficients serve as the basis for inversion, enabling an
assessment of our ability to accurately replicate the zonal flow. To enhance
the realism of the inversion process, random noise is introduced based on
observed data. Employing a regularized least squares formula, we aim to
minimize the misfit function, denoted as $\chi$, as the following
$\chi^{2}=\sum_{n,\ell}\frac{1}{\sigma_{n,\ell}^{2}}\left(a_{s}(n,\ell)-\int
K_{n,\ell,s}(r)w_{s}(r)r^{2}dr\right)^{2}+\lambda\int\left(\frac{\partial^{2}w_{s}}{\partial
r^{2}}\right)^{2}r^{2}dr,$ (4)
We expect our solution for $w_{s}(r)$ to manifest as a linear combination of
splitting coefficients $a_{s}$.
$\displaystyle w_{s}(r_{0})=\sum_{i\in M}c_{i}(r_{0})a_{s}(i)$
$\displaystyle=\sum_{i\in M}c_{i}(r_{0})\int K_{i,s}(r)w_{s}(r)r^{2}dr$
$\displaystyle=\int\mathcal{K}(r;r_{0})w_{s}(r)r^{2}dr.$ (5)
Equation 5 reveals that our inverted $w_{s}$ at depth $r_{0}$ represents a
convolution of the true $w_{s}(r)$ profile, with $\mathcal{K}$ representing
the averaging kernel, $M$ represents the set of measurements for all
accessible acoustic modes. In Section D, we derive the expression for the
averaging kernel $\mathcal{K}(r;r_{0})$ and demonstrate its behavior at
various depths. Figure 9 illustrates that the averaging kernel tends to be
narrower near the surface and widens with greater depths. This observation
underscores our higher resolution capability close to the surface compared to
the base of the convection zone. To attain a smooth solution for $w_{s}(r)$,
we employ second-order derivative smoothing. The function, $w_{s}$, is
expanded on the basis of B-spline functions with $50$ equally spaced knots
along the acoustic depth, ranging from the surface to the center. Additional
details regarding our inversion method are available in Appendix C. To compute
error bars, we utilize the Monte Carlo method. This involves perturbing the
measured splitting coefficients in accordance with observed noise, generating
a sample set of $100$ realizations. Subsequently, we conduct inversions with
each sample to derive profiles for $w_{s}(r)$. The variation among these
different profiles of $w_{s}(r)$ allows us to calculate the standard
deviation, which in turn quantifies the error in $w_{s}$. We then use Equation
2 to estimate the error in velocity, $v(r,\theta)$. Our inversion technique
follows a $1.5$D approach. Other inversion methods are also available; for
instance, refer to Schou & Brown (1994); Schou et al. (1998).
## 3 Results
### 3.1 Forward Modeling of Helioseismic Effects of Torsional Oscillations
In their research, Pipin & Kosovichev (2020) delved into the connection
between the amplitude of the “extended” mode of migrating zonal flows and
solar magnetic cycles using a nonlinear mean-field solar dynamo model. Their
findings unveiled a notable correlation between torsional oscillation
parameters, such as zonal acceleration, and the subsequent cycle magnitudes,
albeit with a time lag ranging from $11$ to $20$ years. This discovery holds
great significance, suggesting that helioseismic observations of torsional
oscillations can offer valuable insights for more advanced predictions of
solar cycles, potentially forecasting solar activity $1$ to $2$ sunspot cycles
in advance. Given the substantial importance of this finding, we employ zonal
flow data from their simulation and engage in a forward modeling approach.
This allows us to gain insight into the specific signatures we should be
seeking within observed a-coefficients. This result is depicted in Figure 1.
We consider modes with turning-point radii between 0.95-1R⊙. We calculate the
average values of the a-coefficients of those mode sets. We only compare the
variations of the a-coefficients with time. Upon examining Figure 1, it
becomes evident that periodicity is present among all the a-coefficients
across all harmonic degrees. To determine if similar observations can be made
in the actual observed data, we conduct identical tests. We focused on the
modes with the turning-point radii ranging from 0.9-1R⊙. The results for the
observed a-coefficients from all three instruments are shown in Figure 2.
Similar to the findings of Basu & Antia (2019), we see periodicity in all the
a-coefficients in the observed data except for $a_{3}$, which does not exhibit
periodicity during the available time period of data. It is possible that it
reflects a longer trend. While the splitting coefficients from GONG and HMI
exhibit consistency (refer to the left panel of Figure 2), those from MDI do
not align. Following the approach of Basu & Antia (2019); Antia et al. (2008),
we limit our analysis to a maximum harmonic degree of $120$ for MDI, which
results in consistency across all three instruments (see right panel of Figure
2). To unveil the signature of dynamo waves from the tachocline to the solar
surface, we must address the presence of outliers in our observed data. Given
the relatively small magnitude of the signal, we can either eliminate outliers
through data cleaning or extend the duration of the time series to enhance the
signal-to-noise ratio.
### 3.2 Inversion for synthetic tests
We use noise from the $72$-day observational data to perturb the measurement.
We use a model from Pipin & Kosovichev (2020) to generate artificial
a-coefficients and perform inversion of these coefficients. In Figure 3, we
present a comparison between our inversion results and the original profile.
We showcase results for both scenarios: without noise and with added noise.
It’s evident that without noise, the inversion performs exceptionally well.
Even with added noise, the inversion effectively captures all the features of
the dynamo wave pattern from the tachocline to the surface. We specifically
show two harmonic degrees, $s=1$ and $s=3$.
Figure 3: Zonal flow derived from the dynamo model undergoes spherical
harmonic decomposition, as described by Equation 2. The left and right panels
illustrate our inversion outcomes for the $a$-coefficients of degrees $s=1$
and $s=3$, respectively. The top panel displays the original profile of
$w_{s}$, utilized in forward modeling to derive the $a$-coefficients from
Equation 3. These $a$-coefficients serve as inputs for the inversion process,
showcased in the middle panel. Following the initial calculation of the
$a$-coefficients as previously mentioned, they are subsequently subjected to
perturbation using Gaussian random noise proportional to the observed values.
The inversion is then conducted based on this perturbed data, with the
resulting analysis depicted in the bottom panel.
In Figure 4 within the dynamo model, we observe a band originating from the
tachocline and migrating toward the surface. This migration takes longer near
low-latitude regions ($<30^{\circ}$) but considerably less time at higher
latitudes, around $60^{\circ}$. This identical profile has been decomposed
into spherical harmonics to derive $w_{s}(r)$ using Equation 2, facilitating
forward modeling via Equation 3. Having demonstrated in Figure 3 that we can
accurately recover the original profile using our inversion methods, we
combine components from all harmonic degrees to derive the velocity using
Equation 2. We compare our inverted profile against the original profile in
Figure 4.
Figure 4: Zonal flow from the dynamo model is plotted against radius over time
across various latitudes, indicated in each panel’s title. The left panel
illustrates the inverted profile, while the right panel displays the original
profile.
### 3.3 Accuracy near the tachocline
Kosovichev & Pipin (2019) discussed the potential for predicting the strength
of future solar cycles by measuring torsional oscillation strength at the base
of the convection zone, particularly at latitude $60^{\circ}$. To undertake a
similar analysis, validating our accuracy near the base of the convection zone
is imperative. To accomplish this, we assess the accuracy through synthetic
tests. In Figure 5, we compare our inverted values with the original ones at
depth $0.75R_{\odot}$ and latitude $60^{\circ}$. We consider two cases. In one
instance, we apply noise akin to a $72$-day measurement, while in another, we
simulate noise comparable to a $2$-year measurement. We observe consistent
accuracy in both scenarios near the surface layers, depth $0.98R_{\odot}$.
However, there is a notable shift in the results as we approach the base of
the convection zone. This variation aligns with expectations due to reduced
sensitivity in deeper layers. P-modes spend less time traversing these depths,
primarily because of the increasing sound speed from the surface towards the
center of the Sun. This leads to an increase in the inversion error bars from
the surface towards the center. As the error bars in splitting coefficients
decrease with the square root of the total time, the inversion error bars near
the tachocline tend to be smaller for longer time series measurements. This
characteristic aids in interpreting variations in torsional oscillations
around the tachocline. In Figure 5, we illustrate this trend. JSOC typically
provides the $a$-coefficient data for the time series spanning 72 days.
However, Korzennik (2023) has presented the splitting coefficients across
various time series, including longer periods such as $1$ or $2$ years. This
varied dataset might significantly reduce uncertainties near the base of the
convection zone, aiding in understanding how torsional oscillation strength
evolves over time. This understanding could be pivotal in forecasting the
traits of forthcoming solar cycles (Kosovichev & Pipin, 2019; Pipin &
Kosovichev, 2020). However, it is important to note that long-term averaged
data might blend various dynamical changes occurring at the tachocline.
Figure 5: We present our synthetic inversion results across varying noise
levels. The left panels depict outcomes using noise akin to a $72$-day time
series, while the right panel reflects results obtained from a $2$-year time
series. All panels depict results corresponding to a latitude of $60^{\circ}$.
Evidently, employing the $2$-year time series yields superior outcomes,
particularly near the tachocline. However, both cases demonstrate equally
commendable results near the surface at $0.98R_{\odot}$. To display the error
bar at a depth of $0.98R_{\odot}$, we scaled it by a factor of $6$.
### 3.4 Inversion with observed data
After showcasing our ability to retrieve dynamo waves across the entire
convection zone through the $72$-day measurement, we proceeded to conduct
inversions using both MDI and HMI data sets, and during the overlap period, we
used the HMI data. Kosovichev & Pipin (2019) demonstrated the presence of
dynamo waves in solar zonal flow by utilizing a-coefficients from the JSOC
pipeline. To validate their analysis, we cross-checked using additional
datasets from the NSO/GONG pipeline. Confirming the presence of the dynamo
wave pattern in these alternative datasets would substantiate Kosovichev &
Pipin (2019) earlier detection. The results from the analysis of datasets
collected by all three instruments are depicted in Figure 6 and 7. We employ a
Gaussian filter with a one-year width in time to smooth the solution. In
Figure 7, two branches of zonal flow are visible, both originating near mid-
latitudes. One migrates towards the equator, while the other moves poleward.
Zonal flow patterns persist across three solar cycles, encompassing cycles
$23$, $24$, and ongoing cycle $25$. From Figure 6, the radial migration of the
high-latitude branch is fast compared to the low-latitude branch, which takes
a longer time to migrate from the tachocline to the surface.
Figure 6: Zonal flow velocity is plotted as a function of time and radius for
four latitudes, indicated in the title of each panel. The MDI and HMI results
are shown in the left panel, and the GONG results are shown in the right
panel.
Figure 7: Zonal flow velocities plotted as a function of time and latitude for
four depths mentioned in the title of each panel. The left panel illustrates
the results derived from the combined MDI and HMI datasets, while the right
panel showcases the findings specifically from the GONG datasets. Figure 8:
Evolution of solar zonal flow in the solar convection zone from the analysis
of MDI and HMI combined data sets.
## 4 Discussion and conclusions
The detection of dynamo waves holds significant importance in discerning the
operational dynamo within the solar interior—whether it aligns with Babcock-
Leighton or Parker’s dynamo. This study aims to ascertain the feasibility of
detecting dynamo-like wave patterns within the solar zonal flow. However, this
detection poses a challenge due to its subtle nature compared to the
prevailing mean rotation, which has a significantly higher magnitude.
Understanding the dynamo-wave-like structure within the zonal flows assumes
crucial importance as it could elucidate the type of dynamo active in the Sun.
This comprehension, in turn, might aid in predicting the characteristics of
subsequent solar cycles, thereby contributing significantly to space weather
prediction. Through forward modeling, we observe that variations in $a_{s}$
exhibit an $11$-year periodicity, paralleling similar characteristics observed
in the a-coefficients depicted in Figure 2, barring the potential for longer
periodicity in $a_{3}$ coefficients. This forward modeling, along with the
inversions, is necessary to decipher any subtle structure like dynamo waves
present in the solar zonal flow measurements. After developing a forward
modeling approach, we perform the inversion of the synthetic data sets. We
show in Figure 4 that we can quite accurately recover the dynamo waves-like
pattern in the zonal flow inferences from the tachocline to the surface.
After this testing, we used the observed data from three instruments MDI, HMI,
and GONG, and found a similar dynamo-wave-like pattern in the inferred solar
zonal flows. The similarity of the zonal flows deduced from the observations
and the dynamo models allows us to link the observed variations to the dynamo
waves, confirming the interpretation of Kosovichev & Pipin (2019). We find
that at high latitudes, the radial migration of the dynamo waves from the
tachocline to the surface is very fast, whereas it is of the order of several
years in the low-latitude areas. It exhibits a distinct behavior within the
near-surface layers, where the zonal flows migrate from the surface to the
bottom of the near-surface shear layer in the opposite direction compared to
the rest of the convection zone, as illustrated in Figure 6. This phenomenon
mirrors a similar structure previously observed by Vorontsov et al. (2002)
(refer to Figure $3$ in their study). Its origin has to be studied further. It
can be connected with the magnetic field feedback on the on the angular
momentum fluxes and the heat-transport in the near-surface layers. In the
presented dynamo model, a similar feature can be noticed in the time-radius
diagram at 45∘ latitude in Figure 4, but it is much less pronounced.
We can speculate a plausible dynamo scenario, previously discussed by
Kosovichev & Pipin (2019), in which the generation of the poloidal magnetic
field occurs due to helical turbulence in this high-latitude zone around
$60^{\circ}$, near the base of the convection zone. The field swiftly migrates
to the surface within a time frame of $1$ to $2$ years. However, in low
latitudes, this migration process takes longer, reaching the surface after
being twisted by differential rotation, transforming from the poloidal to the
toroidal component. This aspect resembles Parker’s dynamo theory (Parker,
1955), and is reproduced in our test dynamo model (Pipin & Kosovichev, 2019).
Due to buoyancy instability, these magnetic fields emerge at the surface,
forming sunspots. As sunspots decay, the remnant magnetic fields are
transported by meridional flows to the polar regions, where they eventually
descend and likely contribute to the seed field for the subsequent solar
dynamo cycle. This facet of the dynamo process shares similarities with the
models proposed by Dikpati et al. (2009) and Pipin et al. (2023). Our results
show clear signs of the magnetic cycle effects on the near-surface dynamics of
the torsional oscillations.
Emerging dynamo models propose that the inflow toward active regions could
influence the characteristics of the solar dynamo (Teweldebirhan et al.,
2023). This prompts an investigation into the potential implications of these
inflows of the zonal flows and the corresponding observable in the frequency-
splitting data. Both forward modeling and inversion techniques are essential
to decipher the specific signature that should manifest in the frequency-
splitting data, providing valuable insights into the solar dynamo’s
characteristics.
In summary, in this work, we developed a methodology for forward modeling of
the effects of migrating zonal flows on the rotational frequency splitting
data obtained from global helioseismology measurements and inversion for
inferring the flows in the solar convection zone. Our inversion is based on
the regularized least-squares RLS method. We find that using our approach, we
can detect the migrating dynamo-like wave patterns in the solar zonal flows
using only 72-day measurements of p-modes of solar oscillations. However, the
error bars are notably large near the tachocline, and our study shows that
helioseismic analysis of the 2-year time series of solar oscillations should
be able to measure the zonal flows in the tachocline. We will address this
task in future work.
##
Acknowledgments: This work was partially supported by the NASA grants
80NSSC20K1320, 80NSSC20K0602, and 80NSSC22M0162. VP thanks the financial
support of the Ministry of Science and Higher Education (Subsidy
No.075-GZ/C3569/278).
## Appendix A Theory
### A.1 Properties of Generalized Spherical Harmonics
We use the following relations of generalized spherical harmonics
$\displaystyle Y_{\ell m}(\theta,\phi)$
$\displaystyle=\gamma_{\ell}Y_{\ell}^{0m}(\theta,\phi),$ (A1)
$\displaystyle\frac{\partial Y_{\ell}^{Nm}}{\partial\theta}$
$\displaystyle=\frac{1}{\sqrt{2}}\left(\Omega_{N}^{\ell}Y_{\ell}^{N-1,m}-\Omega_{N+1}^{\ell}Y_{\ell}^{N+1,m}\right),$
(A2) $\displaystyle-\frac{m}{\sin\theta}Y_{\ell}^{0m}$
$\displaystyle=\frac{\Omega_{0}^{\ell}}{\sqrt{2}}\left(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m}\right)$
(A3) $\displaystyle\frac{N\cos\theta-m}{\sin\theta}Y_{\ell}^{Nm}$
$\displaystyle=\frac{1}{\sqrt{2}}\left(\Omega_{N}^{\ell}Y_{\ell}^{N-1,m}+\Omega_{N+1}^{\ell}Y_{\ell}^{N+1,m}\right)$
(A4)
where $\gamma_{\ell}=\sqrt{(2\ell+1)/(4\pi)}$ and
$\Omega_{N}^{\ell}=\sqrt{\frac{1}{2}(\ell+N)(\ell-N+1)}$.
### A.2 Properties of Wigner-3j symbol
We use the following symmetry properties of Wigner’s 3j symbol
$\begin{pmatrix}J_{1}&J_{2}&J_{3}\\\
m_{1}&m_{2}&m_{3}\end{pmatrix}=\begin{pmatrix}J_{2}&J_{3}&J_{1}\\\
m_{2}&m_{3}&m_{1}\end{pmatrix}=\begin{pmatrix}J_{3}&J_{1}&J_{2}\\\
m_{3}&m_{1}&m_{2}\end{pmatrix}$ (A6)
For odd permutations we have
$\begin{pmatrix}J_{1}&J_{2}&J_{3}\\\
m_{1}&m_{2}&m_{3}\end{pmatrix}=(-1)^{J_{1}+J_{2}+J_{3}}\begin{pmatrix}J_{2}&J_{1}&J_{3}\\\
m_{2}&m_{1}&m_{3}\end{pmatrix}$ (A7)
The following recurrence relation will be necessary for our future analysis.
$\displaystyle\left[(j_{2}-m_{2})(j_{2}+m_{2}+1)(j_{3}+m_{3})(j_{3}-m_{3}+1)\right]^{1/2}\begin{pmatrix}j_{1}&j_{2}&j_{3}\\\
m_{1}&m_{2}+1&m_{3}-1\end{pmatrix}$ (A8)
$\displaystyle+\left[(j_{2}+m_{2})(j_{2}-m_{2}+1)(j_{3}-m_{3})(j_{3}+m_{3}+1)\right]^{1/2}\begin{pmatrix}j_{1}&j_{2}&j_{3}\\\
m_{1}&m_{2}-1&m_{3}+1\end{pmatrix}$
$\displaystyle=\left(j_{1}(j_{1}+1)-j_{2}(j_{2}+1)-j_{3}(j_{3}+1)-2m_{2}m_{3}\right)\begin{pmatrix}j_{1}&j_{2}&j_{3}\\\
m_{1}&m_{2}&m_{3}\end{pmatrix}$ (A9)
The integral of three generalized spherical harmonics obeys the following
relations
$\displaystyle\int_{0}^{2\pi}d\phi\int_{0}^{\pi}d\theta\sin\theta\left(Y_{\ell^{\prime}}^{N^{\prime}m^{\prime}}\right)^{*}Y_{\ell^{\prime\prime}}^{N^{\prime\prime}m^{\prime\prime}}Y_{\ell}^{Nm}=4\pi(-1)^{(N^{\prime}-m^{\prime})}\begin{pmatrix}\ell^{\prime}&\ell^{\prime\prime}&\ell\\\
-N^{\prime}&N^{\prime\prime}&N\end{pmatrix}\begin{pmatrix}\ell^{\prime}&\ell^{\prime\prime}&\ell\\\
-m^{\prime}&m^{\prime\prime}&m\end{pmatrix}$ (A10)
It is to be noted that Wigner-3j symbols in Equation A6 are non-zero when the
following relation is satisfied
$m1+m2+m3=0$
Here we define another term
$B^{(N)\pm}_{\ell^{\prime}\ell^{\prime\prime}\ell}=\frac{(-1)^{N}}{2}(1\pm(-1)^{\ell+\ell^{\prime}+\ell^{\prime\prime}})\left[\frac{(\ell^{\prime}+N)!(\ell+N)!}{(\ell-N)!(\ell-N)!}\right]\begin{pmatrix}\ell^{\prime}&\ell^{\prime\prime}&\ell\\\
-N&0&N\end{pmatrix}$ (A11)
One relation that shall be used to calculate the kernel is the following which
can be derived from Equation A9
$\displaystyle(\Omega^{\ell}_{0})^{2}B^{(1)-}_{s\ell\ell}+\Omega_{0}^{s}(\Omega^{\ell}_{0})^{2}\Omega^{\ell}_{2}(1-(-1)^{2\ell+s}\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}=\frac{1}{2}[2\ell(\ell+1)-s(s+1)]B^{(1)-}_{\ell s\ell},$
(A12)
### A.3 Kernel calculations
Rotation frequency splitting is calculated using the perturbation theory based
on the variational principle:
$\delta\omega_{k}=-i\frac{\int_{V}\rho_{0}\boldsymbol{\xi}_{k}^{*}\cdot(\mathbf{v}_{0}\cdot\nabla)\boldsymbol{\xi}_{k}dV}{\int_{V}\rho_{0}|\delta\mathbf{r}|^{2}dV},$
(A13)
$\boldsymbol{\xi}_{k}$ is the displacement eigenfunction of a mode, and $k$
signifies the combination of the radial order, angular degree, and azimuthal
number $(n,\ell,m)$ of a mode. The eigenfunction of a spherically symmetric
Sun can be written as
$\boldsymbol{\xi}_{k}=U(r)Y_{\ell m}\hat{r}+V(r)\boldsymbol{\nabla}_{h}Y_{\ell
m},$ (A14)
where $U(r)$ and $V(r)$ are the radial and horizontal components of the
displacement vector. Reformulating Equation A14 using generalized spherical
harmonics and presenting them as components within a vector representation,
$\displaystyle\xi_{k,r}$ $\displaystyle=\gamma_{\ell}U(r)Y_{\ell}^{0,m}$
$\displaystyle\xi_{k,\theta}$
$\displaystyle=\gamma_{\ell}V(r)\partial_{\theta}Y_{\ell}^{0,m}=\frac{\gamma_{\ell}}{\sqrt{2}}V(r)\Omega^{\ell}_{0}(Y_{\ell}^{-1,m}-Y_{\ell}^{1,m})$
$\displaystyle\xi_{k,\phi}$
$\displaystyle=\gamma_{\ell}V(r)\partial_{\phi}Y_{\ell}^{0,m}=-i\frac{\gamma_{\ell}}{\sqrt{2}}V(r)\Omega^{\ell}_{0}(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m})$
(A15)
The velocity of solar rotation can be written as
$\displaystyle\mathbf{v}_{0}(r,\theta)$
$\displaystyle=-\sum_{s=1,3,5}w_{s}^{0}(r)\partial_{\theta}Y_{s0}\hat{\phi}$
$\displaystyle\mathbf{v}_{0}(r,\theta)$
$\displaystyle=-\sum_{s=1,3,5}\gamma_{s}w_{s}^{0}(r)\partial_{\theta}Y_{s}^{0,0}\hat{\phi}=-\sum_{s=1,3,5}\frac{\gamma_{s}}{\sqrt{2}}w_{s}^{0}(r)\Omega_{0}^{s}(Y_{s}^{-1,0}-Y_{s}^{1,0})\hat{\phi}$
(A17)
We use Equation A2 within Equation A17 to denote the derivative with respect
to $\theta$. In order to calculate Equation A13, we need to define a tensor
called $T$.
$T=\nabla\boldsymbol{\xi}_{k}$ (A18)
The components of tensor $T$, which are required to calculate the sensitivity
kernels are shown below. Temporarily, we omit the subscript $k$ from the
eigenfunction $\boldsymbol{\xi}_{k}$ for the sake of brevity in notation and
use the $r$, $\theta$, and $\phi$ components of the eigenfunction
$\boldsymbol{\xi}_{k}$ from Equation A14.
$\displaystyle T_{\phi r}$
$\displaystyle=\frac{1}{r\sin\theta}(\partial_{\phi}\boldsymbol{\xi}).\hat{r}=\frac{1}{r\sin\theta}(\partial_{\phi}\xi_{r}-\xi_{\phi}\sin\theta)$
$\displaystyle=\gamma_{\ell}U\frac{im}{r\sin\theta}Y_{\ell}^{0m}-(-i)\frac{\gamma_{\ell}}{\sqrt{2}}V\Omega_{0}^{\ell}r^{-1}\left(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m}\right)$
$\displaystyle=-ir^{-1}\gamma_{\ell}\frac{\Omega_{0}^{\ell}}{\sqrt{2}}(U-V)\left(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m}\right),$
(A19) $\displaystyle T_{\phi\theta}$
$\displaystyle=\frac{1}{r\sin\theta}(\partial_{\phi}\xi_{\theta}-\cos\theta\xi_{\phi})$
$\displaystyle=r^{-1}\left(\frac{\gamma_{\ell}}{\sin\theta\sqrt{2}}V\Omega_{0}^{\ell}\partial_{\phi}(Y_{\ell}^{-1,m}-Y_{\ell}^{1,m})+\cot\theta
i\frac{\gamma_{\ell}}{\sqrt{2}}V\Omega_{0}^{\ell}(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m})\right)$
$\displaystyle=ir^{-1}\frac{\gamma_{\ell}\Omega_{0}^{\ell}}{\sqrt{2}}V\left(\frac{m}{\sin\theta}(Y_{\ell}^{-1,m}-Y_{\ell}^{1,m})+\cot\theta(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m})\right)$
$\displaystyle=ir^{-1}\frac{\gamma_{\ell}}{2}V\Omega_{0}^{\ell}\Omega_{2}^{\ell}\left(Y_{\ell}^{2m}-Y_{\ell}^{-2m}\right),$
(A20) $\displaystyle T_{\phi\phi}$
$\displaystyle=\frac{1}{r\sin\theta}\left[\xi_{r}(\partial_{\phi}\hat{r})\cdot\hat{\phi}+\xi_{\theta}(\partial_{\phi}\hat{\theta})\cdot\hat{\phi}+\partial_{\phi}\xi_{\phi}\right]$
$\displaystyle=\frac{1}{r\sin\theta}\left[\xi_{r}\sin\theta+\xi_{\theta}\cos\theta+\partial_{\phi}\xi_{\phi}\right]$
$\displaystyle=\frac{1}{r}\left[\gamma_{\ell}U(r)Y_{\ell}^{0m}+\cot\theta\frac{\gamma_{\ell}}{\sqrt{2}}V(r)\Omega_{0}^{\ell}(Y_{\ell}^{-1,m}-Y_{\ell}^{1,m})+\frac{m}{\sin\theta}\frac{\gamma_{\ell}}{\sqrt{2}}V(r)\Omega_{0}^{\ell}(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m})\right]$
$\displaystyle=r^{-1}\left[\gamma_{\ell}U(r)Y_{\ell}^{0m}-\frac{\gamma_{\ell}}{2}V\Omega_{0}^{\ell}\left[\Omega_{2}^{\ell}\left(Y_{\ell}^{2m}+Y_{\ell}^{-2m}\right)+2\Omega_{0}^{\ell}Y_{\ell}^{0m}\right]\right]$
$\displaystyle=r^{-1}\gamma_{\ell}\left[(U-\Omega_{0}^{\ell}\Omega_{0}^{\ell}V)Y_{\ell}^{0m}-V\Omega_{0}^{\ell}\Omega_{2}^{\ell}\frac{Y_{\ell}^{2m}+Y_{\ell}^{-2m}}{2}\right],,$
(A21)
The numerator in Equation A13 can be calculated using the tensor components of
$T$ as following
$i\int\rho d\mathbf{r}\left[v_{\phi}(\xi_{k,r}^{*}T_{\phi
r}+\xi^{*}_{k,\theta}T_{\phi\theta}+\xi^{*}_{k,\phi}T_{\phi\phi})\right]=T1+T2+T3,$
(A22)
where
$\displaystyle T1=i\int\rho d\mathbf{r}v_{\phi}(\xi_{k,r}^{*}T_{\phi r}$
$\displaystyle T2=i\int\rho
d\mathbf{r}v_{\phi}\xi^{*}_{k,\theta}T_{\phi\theta}$ $\displaystyle
T3=i\int\rho d\mathbf{r}v_{\phi}\xi^{*}_{k,\phi}T_{\phi\phi}$ (A23)
We individually analyze each term in the above equation $T1$, $T2$, and $T3$
before combining them in the final step.
### A.4 Analyzing $T1$ term
Substituting $v_{\phi}$, $\xi_{k,r}$ and $T_{\phi r}$ from Equations A17,
LABEL:eq:xi_sph and A19 into Equation A23 for $T1$ term we get
$\displaystyle i\int\rho
r^{2}dr\left(-\frac{\gamma_{s}}{\sqrt{2}}w_{s}^{0}\Omega_{0}^{s}(Y_{s}^{-1,0}-Y_{s}^{1,0})\right)(-ir^{-1})\gamma_{\ell}\frac{\Omega_{0}^{\ell}}{\sqrt{2}}(U-V)\left(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m}\right)\gamma_{\ell}UY_{\ell}^{0m*}$
$\displaystyle=-4\pi\frac{\gamma_{\ell}^{2}\gamma_{s}\Omega_{0}^{\ell}\Omega_{0}^{s}}{2}\int\rho
w_{s}^{0}d\mathbf{r}(U^{2}-UV)Y_{\ell}^{0m*}(Y_{s}^{-1,0}-Y_{s}^{1,0})\left(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m}\right)$
$\displaystyle=-4\pi\frac{\gamma_{\ell}^{2}\gamma_{s}\Omega_{0}^{\ell}\Omega_{0}^{s}}{2}(-1)^{m}\begin{pmatrix}\ell&s&\ell\\\
-m&0&m\end{pmatrix}\left(1-(-1)^{2\ell+s}\right)\begin{pmatrix}\ell&s&\ell\\\
0&-1&1\end{pmatrix}\int\rho w_{s}^{0}r^{2}dr(U^{2}-UV)$
$\displaystyle=-4\pi\frac{\gamma_{\ell}^{2}\gamma_{s}}{2}(-1)^{m}\begin{pmatrix}\ell&s&\ell\\\
-m&0&m\end{pmatrix}\int\rho
w_{s}^{0}d\mathbf{r}(U^{2}-UV)B_{s\ell\ell}^{(1)-}$
In our case $m^{\prime}=m$ and selection rules of the 3j symbol dictates
$\displaystyle-N^{\prime}+N^{\prime\prime}+N$ $\displaystyle=0$
$\displaystyle-m^{\prime}+m^{\prime\prime}+m$
$\displaystyle=0=m^{\prime\prime}$
### A.5 Analyzing $T2$ term
Substituting $v_{\phi}$, $\xi_{k,\theta}$ and $T_{\phi\theta}$ from Equations
A17, LABEL:eq:xi_sph and A20 into Equation A23 for $T2$ term we get
$\displaystyle i\int\rho
r^{2}dr\left(-\frac{\gamma_{s}}{\sqrt{2}}w_{s}^{0}\Omega_{0}^{s}(Y_{s}^{-1,0}-Y_{s}^{1,0})\right)\left(ir^{-1}\frac{\gamma_{\ell}}{2}V\Omega_{0}^{\ell}\Omega_{2}^{\ell}\left(Y_{\ell}^{2m}-Y_{\ell}^{-2m}\right)\right)\left(\frac{\gamma_{\ell}}{\sqrt{2}}V(r)\Omega_{0}^{\ell}(Y_{\ell}^{-1,m}-Y_{\ell}^{1,m})\right)^{*}$
$\displaystyle=4\pi\frac{\gamma_{s}\gamma_{\ell}^{2}}{4}\Omega_{0}^{s}\left(\Omega_{0}^{\ell}\right)^{2}\Omega_{2}^{\ell}\int\rho
rV^{2}w_{s}^{0}dr\begin{pmatrix}\ell&s&\ell\\\
-m&0&m\end{pmatrix}\left((-1)^{(-1-m)}\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}-(-1)^{1-m}\begin{pmatrix}\ell&s&\ell\\\
-1&-1&2\end{pmatrix}\right)$
$\displaystyle=-4\pi\frac{\gamma_{s}\gamma_{\ell}^{2}}{4}\Omega_{0}^{s}\left(\Omega_{0}^{\ell}\right)^{2}\Omega_{2}^{\ell}(-1)^{m}\begin{pmatrix}\ell&s&\ell\\\
-m&0&m\end{pmatrix}\int\rho
rV^{2}w_{s}^{0}dr(1-(-1)^{2\ell+s})\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}$
The term that we need to simplify is
$\displaystyle(-1)^{(-1-m)}\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}-(-1)^{1-m}\begin{pmatrix}\ell&s&\ell\\\
-1&-1&2\end{pmatrix}$
$\displaystyle=-(-1)^{m}(1-(-1)^{2\ell+s})\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}$
where we have used the following identity
$\begin{pmatrix}J_{1}&J_{2}&J_{3}\\\
m_{1}&m_{2}&m_{3}\end{pmatrix}=(-1)^{J_{1}+J_{2}+J_{3}}\begin{pmatrix}J_{1}&J_{2}&J_{3}\\\
-m_{1}&-m_{2}&-m_{3}\end{pmatrix}$ (A24)
### A.6 Analyzing $T3$ term
Substituting $v_{\phi}$, $\xi_{k,\phi}$ and $T_{\phi\phi}$ from Equations A17,
LABEL:eq:xi_sph and A21 into Equation A23 for $T3$ term we get
$\displaystyle T3$ $\displaystyle=i\int\rho
d\mathbf{r}\left(-\frac{\gamma_{s}}{\sqrt{2}}w_{s}^{0}\Omega_{0}^{s}(Y_{s}^{-1,0}-Y_{s}^{1,0})\right)$
$\displaystyle\left\\{r^{-1}\gamma_{\ell}\left[(U-\Omega_{0}^{\ell}\Omega_{0}^{\ell}V)Y_{\ell}^{0m}-V\Omega_{0}^{\ell}\Omega_{2}^{\ell}\frac{Y_{\ell}^{2m}+Y_{\ell}^{-2m}}{2}\right]\right\\}$
$\displaystyle\left(-i\frac{\gamma_{\ell}}{\sqrt{2}}V(r)\Omega_{0}^{\ell}(Y_{\ell}^{-1,m}+Y_{\ell}^{1,m})\right)^{*}$
We can simplify this equation as following
$\displaystyle
T3=4\pi\frac{\gamma_{s}\gamma_{\ell}^{2}}{2}\Omega_{0}^{s}\Omega_{0}^{\ell}(-1)^{m}\int\rho
r^{-1}w_{s}^{0}r^{2}dr\begin{pmatrix}\ell&s&\ell\\\ -m&0&m\end{pmatrix}$
$\displaystyle\left(-(1-(-1)^{2\ell+s})\begin{pmatrix}s&\ell&\ell\\\
-1&0&1\end{pmatrix}\left(UV-\Omega_{0}^{\ell}\Omega_{0}^{\ell}V^{2}\right)-\frac{V^{2}\Omega_{0}^{\ell}\Omega_{2}^{\ell}}{2}[1-(-1)^{2\ell+s}]\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}\right)$ (A25)
After using the formula for $B^{(1)-}_{s\ell\ell}$ above Equation can be
simplified further
$\displaystyle T3=4\pi\frac{\gamma_{s}\gamma_{\ell}^{2}}{2}(-1)^{m}\int\rho
r^{-1}w_{s}^{0}d\mathbf{r}\begin{pmatrix}\ell&s&\ell\\\ -m&0&m\end{pmatrix}$
$\displaystyle\left(\left(UV-\Omega_{0}^{\ell}\Omega_{0}^{\ell}V^{2}\right)B^{(1)-}_{s\ell\ell}-\frac{V^{2}\Omega_{0}^{s}(\Omega_{0}^{\ell})^{2}\Omega_{2}^{\ell}}{2}[1-(-1)^{2\ell+s}]\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}\right)$ (A26)
### A.7 Summing over all the terms
After summing over $T1$, $T2$ and $T3$, coefficient associated with the term
$V^{2}$ is following
$-\Omega_{0}^{s}(\Omega^{\ell}_{0})^{2}\Omega^{\ell}_{2}(1-(-1)^{2\ell+s})\begin{pmatrix}\ell&s&\ell\\\
1&1&-2\end{pmatrix}-(\Omega^{\ell}_{0})^{2}B^{(1)-}_{s\ell\ell}=-\frac{1}{2}[2\ell(\ell+1)-s(s+1)]B^{(1)-}_{\ell
s\ell}$ (A27)
where we have used the identity Equation A12. We get the equation for the
kernel as follows
$M_{n\ell}=4\pi\frac{\gamma_{s}\gamma_{\ell}^{2}}{2}(-1)^{m}\int\rho
r^{-1}w_{s}^{0}d\mathbf{r}\begin{pmatrix}\ell&s&\ell\\\
-m&0&m\end{pmatrix}\left[-U^{2}+2UV-\frac{V^{2}}{2}(2\ell(\ell+1)-s(s+1))\right]B^{(1)-}_{\ell
s\ell},$ (A28)
where the form for $B^{(1)-}_{\ell s\ell}$ from Equation A11 is the following
$B^{(1)-}_{\ell
s\ell}=\frac{-1}{2}(1-(-1)^{2\ell+s})\ell(\ell+1)\begin{pmatrix}\ell&s&\ell\\\
-1&0&1\end{pmatrix}$ (A29)
We can rearrange all the terms in the Equation A28 and write it as follows
$M_{n\ell}=\sqrt{\frac{2s+1}{4\pi}}(2\ell+1)L^{2}H_{s}^{1}H_{s}^{m}F_{s}^{2}\int\rho
r^{-1}w_{s}^{0}d\mathbf{r}\left[-U^{2}+2UV-\frac{V^{2}}{2}(2\ell(\ell+1)-s(s+1))\right]$
(A30)
where we have used a similar notation as used by Ritzwoller & Lavely (1991)
(Equation $32-35$ of their paper).
## Appendix B Asymptotic relation for the $a$-coefficients
Matrix element due to frequency splitting is following (when $s<<l$)
$\displaystyle M_{n\ell}$ $\displaystyle=-i\int_{0}^{R_{\odot}}\rho
dV\bm{\xi}_{n\ell m}\cdot(\mathbf{u}\cdot\bm{\nabla}\bm{\xi}_{e\ell m}$
$\displaystyle=(-1)^{m+1}\left[\ell(\ell+1)(2\ell+1)\right]\sqrt{\frac{2s+1}{4\pi}}\sum_{s=1,2}\begin{pmatrix}s&\ell&\ell\\\
0&1&-1\end{pmatrix}\begin{pmatrix}s&\ell&\ell\\\
0&m&-m\end{pmatrix}\int_{0}^{R}\rho
r^{2}dr\frac{w_{s}(r)}{r}\left[U^{2}-2UV+\frac{2\ell(\ell+1)-s(s+1)}{2}V^{2}\right]$
$\displaystyle=(-1)^{m+1}\left[\ell(\ell+1)(2\ell+1)\right]\sqrt{\frac{2s+1}{4\pi}}\sum_{s=1,2}(-1)^{\frac{s+2m+1}{2}}\frac{(s!!)^{2}}{2\ell^{2}s!}P_{s}(m/\ell)\langle\frac{w_{s}(r)}{r}\rangle_{n\ell}$
$\displaystyle\approx\sum_{s}\left((-1)^{(s-1)/2}\sqrt{\frac{2s+1}{4\pi}}\frac{(s!!)^{2}}{s!}\langle\frac{w_{s}(r)}{r}\rangle_{n\ell}\right)\ell
P_{s}(m/\ell)$ $\displaystyle=\sum_{s}a_{s}\ell P_{s}(m/\ell)$
where the expression for $a_{s}$ is the following
$a_{s}(n,\ell)=(-1)^{(s-1)/2}\sqrt{\frac{2s+1}{4\pi}}\frac{(s!!)^{2}}{s!}\langle\frac{w_{s}(r)}{r}\rangle_{n\ell}$
(B1)
where we have used the following simplified relation (Equation $A4$ from
Vorontsov (2011) )
$\begin{pmatrix}s&\ell&\ell\\\
0&1&-1\end{pmatrix}\begin{pmatrix}s&\ell&\ell\\\
0&m&-m\end{pmatrix}=(-1)^{(s+2m+1)/2}\frac{(s!!)^{2}}{2\ell^{2}s!}P_{s}(m/\ell)$
(B2)
### B.1 Full expression of kernel
We can obtain relations for the $a$-coefficients as following from Equation
A30
$\displaystyle
a_{s}=4\ell(\ell+1)(2\ell+1)\frac{2s+1}{4\pi}\frac{\\{(2\ell-1)!\\}^{2}}{(2\ell-s)!(2\ell+s+1)!}\mathcal{P}_{s}^{\ell}(1)\langle\frac{w_{s}(r)}{r}\rangle$
$\displaystyle=\int_{0}^{R}K_{s}(n,\ell;r)w_{s}(r)rdr$ (B3)
where the expression for $K_{s}(n,\ell)$ is following
$K_{s}(n,\ell;r)=4\ell(\ell+1)(2\ell+1)\frac{2s+1}{4\pi}\frac{\\{(2\ell-1)!\\}^{2}}{(2\ell-s)!(2\ell+s+1)!}\mathcal{P}_{s}^{\ell}(1)\rho\left[U^{2}-2UV+\frac{2\ell(\ell+1)-s(s+1)}{2}V^{2}\right]$
(B4)
## Appendix C Inversion method
The frequency splitting is expressed in terms of the $a-$coefficients.
$\omega_{n\ell
m}/(2\pi)=\nu_{n\ell}+\sum_{s=1,3,5,\dots,s_{\text{max}}}a_{n,\ell,s}\mathcal{P}_{s}^{\ell}(m)$
(C1)
where splitting coefficients are related as follows
$a_{s}(n,\ell)=\int_{0}^{R_{\odot}}K_{n,\ell,s}(r)w_{s}(r)r^{2}dr,$ (C2)
where $w_{s}$ are used to express velocity profile
$v(r,\theta)=\sum_{s}w_{s}(r)\partial_{\theta}Y_{s0}(\theta,\phi)$ (C3)
We minimize the following equation to obtain $w_{s}(r)$
$\chi=\sum_{n,\ell}\frac{1}{\sigma_{n,\ell}^{2}}\left(a_{n,\ell,s}-\int
K_{n,\ell,s}(r)w_{s}(r)r^{2}dr\right)^{2}+\lambda\int\left(\frac{\partial^{2}w_{s}}{\partial
r^{2}}\right)^{2}r^{2}dr,$ (C4)
We express $w_{s}$ in terms of B-spline in radius as following
$w_{s}(r)=\sum_{i=1}^{N}b_{i}B_{i}(r),$ (C5)
$N$ corresponds to the total number of unknowns. If we substitute Equation C5
into Equation C4, we get
$\displaystyle\chi=\sum_{n,\ell}\frac{1}{\sigma_{n,\ell}^{2}}\left(a_{s}(n,\ell)-\sum_{i}\int
K_{n,\ell,s}(r)b_{i}B_{i}(r)r^{2}dr\right)^{2}+\lambda\int\left(\frac{\partial^{2}\sum_{i}b_{i}B_{i}(r)}{\partial
r^{2}}\right)^{2}r^{2}dr$ (C6)
We minimize the above equation with respect to unknown coefficients $b_{j}$ to
obtain
$\displaystyle\frac{\partial\chi}{\partial
b_{j}}=-\sum_{n,\ell}\frac{1}{\sigma_{n,\ell}^{2}}\left(a_{s}(n,\ell)-\sum_{i}\int
K_{n,\ell,s}(r)b_{i}B_{i}(r)r^{2}dr)\right)\int
K_{n,\ell,s}(r)B_{j}(r)r^{2}dr$
$\displaystyle+\lambda\int\sum_{i}b_{i}\frac{\partial^{2}B_{i}(r)}{\partial
r^{2}}r^{2}dr\int\frac{\partial^{2}B_{j}(r)}{\partial r^{2}}r^{2}dr$ (C7)
We define matrix $A$
$\displaystyle A_{M,i}$ $\displaystyle=\int K_{M}B_{i}(r)r^{2}dr$
$\displaystyle D_{ij}$ $\displaystyle=\int\frac{\partial^{2}B_{i}(r)}{\partial
r^{2}}r^{2}dr\int\frac{\partial^{2}B_{j}(r)}{\partial r^{2}}r^{2}dr$
where $M$ is a multi-index variable for $(n,\ell)$ and assume there are total
$N_{0}$ such measurements. We get
$\sum_{M\in
N_{0}}\frac{1}{\sigma_{M}^{2}}A_{Mi}A_{Mj}b_{i}+\lambda\sum_{i}D_{ij}b_{i}=\sum_{M\in
N_{0}}\frac{1}{\sigma_{M}^{2}}a_{M}A_{Mj}$ (C8)
Which we can rewrite as
$\sum_{i=1}^{N}\left[\sum_{M\in
N_{0}}(A^{T})_{jM}\Lambda_{MM}^{-1}A_{Mi}+\lambda
D_{ij}\right]b_{i}=\sum_{M\in N_{0}}A_{jM}^{T}\Lambda_{MM}^{-1}a_{M},$ (C9)
where $\Lambda$ is a diagonal matrix
$\Lambda^{-1}=\begin{pmatrix}1/\sigma_{1}^{2}&0&0&0&\dots\\\
0&1/\sigma_{2}^{2}&0&0&\dots\\\ 0&0&1/\sigma_{3}^{2}&0&\dots\\\
0&0&0&1/\sigma_{4}^{2}&\dots\\\ \dots&\dots&\dots&\dots&\dots\end{pmatrix}$
(C10)
$\Lambda^{-1}$ is a $N_{0}\times N_{0}$ matrix. The above equation can be
written in a matrix form as
$\displaystyle\left[(A^{T}\Lambda^{-1}A)_{ji}+\lambda D_{ji}\right]b_{i}$
$\displaystyle=(A^{T}\Lambda^{-1})_{jM}a_{M}$
$\displaystyle\left[A^{T}\Lambda^{-1}A+\lambda D\right]b$
$\displaystyle=A^{T}\Lambda^{-1}a$ (C11)
where matrix $b=(b_{1},b_{2},\dots,b_{N})$. Solving these equations provides
us with the values of $b$ which are used to reconstruct $w_{s}(r)$ from
equation C5.
## Appendix D Averaging kernels
The splitting coefficients, $a_{s}$ are related to the velocity coefficients,
$w_{s}$:
$a_{s}(n,\ell)=\int_{0}^{R_{\odot}}K_{s}(n,\ell;r)w_{s}(r)r^{2}dr,$ (D1)
We look for a linear combination of $a_{s}(n,\ell)$ such that we can get an
estimate of $w_{s}(r)$ at a target depth $r_{0}$ as following
$\bar{w}_{s}(r)=\sum_{M\in N_{0}}C_{M}a_{s}(M),$ (D2)
$C_{M}$ are the coefficients that need to be determined. Therefore, formula
for the averaging kernels is
$\mathcal{K}(r,r_{0})=\sum_{M}C_{M}K_{s}(M;r),$ (D3)
From Section C we know
$w_{s}(r_{0})=\sum_{i=1}^{N}b_{i}B_{i}(r_{0})$ (D4)
where $b_{i}$ is derived by performing the matrix operations outlined in
Equation C11.
$b=(A^{T}\Lambda^{-1}A+\lambda D)^{-1}(A^{T}\Lambda^{-1})a$ (D5)
Since $b$ and $w_{s}$ are related through $w_{s}(r)=\sum_{i}b_{i}B_{i}(r)$, we
can find the equation for coefficients $c_{M}$ using the following equations
$C_{M}=\sum_{j}B_{j}(r_{0})((A^{T}\Lambda^{-1}A+\lambda
D)^{-1}(A^{T}\Lambda^{-1}))_{jM}$ (D6)
or in matrix notation
$C=\mathbf{B}((A^{T}\Lambda^{-1}A+\lambda D)^{-1}(A^{T}\Lambda^{-1})),$ (D7)
where $B=[B_{1}(r_{0}),B_{2}(r_{0}),\cdots]$ represents a matrix created by
applying B-spline basis functions evaluated at $r_{0}$. After determining the
coefficients $C$, we substitute them into Equation D3 to derive the averaging
kernel at the target depth $r_{0}$. In Figure 9, we display averaging kernel
at two different depths $0.75R_{\odot}$ and $0.9R_{\odot}$.
Figure 9: The averaging kernels for two distinct depths, $r=0.75R_{\odot}$
(indicated by a black dashed line) and $r=0.9R_{\odot}$ (shown as a black
solid line). The target depths are highlighted by vertical dashed grey lines
and are explicitly labeled in the legend of the plot. It is evident that the
averaging kernel tends to have a narrower width closer to the surface, while
it becomes broader in deeper layers.
### D.1 Averaging Kernels in 2D
Solar rotation, $\Omega(r,\theta)$ can be expressed as a toroidal function
$\Omega(r,\theta)=\sum_{s}w_{s}(r)\Psi_{s}(\theta)$ (D8)
where $w_{s}$ is obtained from Equation D2 as a linear combination of
splitting coefficients $a_{s}$ as following
$w_{s}(r)=\sum_{n,\ell}C_{s}(r;n,\ell)a_{s}(n,\ell)$ (D9)
We can find a relationship between the $a-$coefficients and frequency
splitting by minimizing the following misfit function (Pijpers, 1997)
$\chi^{2}=\sum_{m=-\ell}^{\ell}\left(\nu_{n\ell
m}-\sum_{s=0}^{s_{\text{max}}}a_{s}(n,\ell)\mathcal{P}_{s}^{\ell}(m)\right)^{2}$
(D10)
with respect to $a_{s}$. This yields
$a_{s}(n,\ell)=\frac{\sum_{-\ell}^{\ell}\nu_{n\ell
m}\mathcal{P}_{s}^{\ell}(m)}{\sum_{m=-\ell}^{\ell}\mathcal{P}_{s}^{\ell}(m)^{2}}=\sum_{m=-\ell}^{\ell}A_{s}(n,\ell,m)\nu_{n\ell
m}$ (D11)
where $A_{s}(n,\ell,m$ is defined as
$A_{s}(n,\ell,m)=\frac{\mathcal{P}_{s}^{\ell}(m)}{\sum_{m^{\prime}=-\ell}^{\ell}\left(\mathcal{P}_{s}^{\ell}(m^{\prime})\right)^{2}}$
(D12)
Substituting Equation D11 and D9 into Equation D8 we get
$\displaystyle\Omega(r,\theta)=\sum_{s}\sum_{n,\ell}C_{s}(r;n,\ell)\left(\sum_{m=-\ell}^{\ell}A_{s}(n,\ell)\nu_{n\ell
m}\right)\psi(\theta)$ $\displaystyle=\sum_{n\ell
m}C_{s}(r;n,\ell)A_{s}(n,\ell)\psi(\theta)$ $\displaystyle=\sum_{n\ell
m}C_{n\ell m}^{2D}\nu_{n\ell m},$ (D13)
where
$C_{n\ell m}^{2D}=\sum_{s}c_{s}(r;n,\ell)A_{s}(n,\ell)\psi(\theta),$ (D14)
Therefore, we can write the equation for the averaging kernels as the
following
$\mathcal{K}(r,\theta,r_{0},\theta_{0})=\sum_{n\ell m}C_{n\ell
m}^{2D}(r_{0},\theta_{0})K_{n\ell m}(r,\theta),$ (D15)
where $K_{n\ell m}(r,\theta)$ is the $2-$D kernel defined as
$K_{n\ell m}(r,\theta)=\frac{R_{n\ell m}}{I_{n\ell m}}$ (D16)
where $R_{n\ell m}$ is defined as the following
$\displaystyle R_{n\ell m}$ $\displaystyle=m\rho r\sin\theta\Bigg{\\{}\lvert
U\rvert^{2}P_{\ell}^{m}(\cos\theta)^{2}+\lvert
V(r)\rvert^{2}\Bigg{[}\left(\frac{dP_{\ell}^{m}}{d\theta}\right)^{2}$
$\displaystyle\quad+\frac{m^{2}}{\sin^{2}\theta}P_{\ell}^{m}(\cos\theta)^{2}\Bigg{]}-P_{\ell}^{m}(\cos\theta)^{2}[U^{*}V+UV^{*}]$
$\displaystyle\quad-2P_{\ell}^{m}(\cos\theta)\frac{dP_{\ell}^{m}}{d\theta}\frac{\cos\theta}{\sin\theta}\lvert
V\rvert^{2}\Bigg{\\}}$ (D17)
where $I_{n\ell m}$ is
$I_{n\ell
m}=\frac{2}{2\ell+1}\frac{(\ell+|m|)!}{(\ell-|m|)!}\int_{0}^{R}\left[U^{2}+\ell(\ell+1)V^{2}\right]\rho
r^{2}dr$ (D18)
In Figure 10, we present the averaging kernels for various depths and
latitudes. The figure clearly illustrates a distinctive localization in both
latitude and radius.
Figure 10: Left panel: averaging kernel plotted at radius $0.8R_{\odot}$ and
latitude $20^{\circ}$. Right panel: averaging kernel plotted at radius
$0.75R_{\odot}$ and latitude $40^{\circ}$.
## References
* Antia et al. (2008) Antia, H. M., Basu, S., & Chitre, S. M. 2008, ApJ, 681, 680, doi: 10.1086/588523
* Babcock (1961) Babcock, H. W. 1961, ApJ, 133, 572, doi: 10.1086/147060
* Basu & Antia (2019) Basu, S., & Antia, H. M. 2019, ApJ, 883, 93, doi: 10.3847/1538-4357/ab3b57
* Charbonneau & Barlet (2011) Charbonneau, P., & Barlet, G. 2011, Journal of Atmospheric and Solar-Terrestrial Physics, 73, 198, doi: 10.1016/j.jastp.2009.12.020
* Chen & Zhao (2017) Chen, R., & Zhao, J. 2017, ApJ, 849, 144, doi: 10.3847/1538-4357/aa8eec
* Choudhuri et al. (2007) Choudhuri, A. R., Chatterjee, P., & Jiang, J. 2007, Phys. Rev. Lett., 98, 131103, doi: 10.1103/PhysRevLett.98.131103
* Choudhuri & Dikpati (1999) Choudhuri, A. R., & Dikpati, M. 1999, Sol. Phys., 184, 61, doi: 10.1023/A:1005092601436
* Dikpati et al. (2009) Dikpati, M., Gilman, P. A., Cally, P. S., & Miesch, M. S. 2009, ApJ, 692, 1421, doi: 10.1088/0004-637X/692/2/1421
* Gizon et al. (2020) Gizon, L., Cameron, R. H., Pourabdian, M., et al. 2020, Science, 368, 1469, doi: 10.1126/science.aaz7119
* Guerrero et al. (2016) Guerrero, G., Smolarkiewicz, P. K., de Gouveia Dal Pino, E. M., Kosovichev, A. G., & Mansour, N. N. 2016, ApJ, 828, L3, doi: 10.3847/2041-8205/828/1/L3
* Hanasoge et al. (2017) Hanasoge, S. M., Woodard, M., Antia, H. M., Gizon, L., & Sreenivasan, K. R. 2017, MNRAS, 470, 1404, doi: 10.1093/mnras/stx1298
* Hill et al. (1996) Hill, F., Stark, P. B., Stebbins, R. T., et al. 1996, Science, 272, 1292, doi: 10.1126/science.272.5266.1292
* Howard & Labonte (1980) Howard, R., & Labonte, B. J. 1980, ApJ, 239, L33, doi: 10.1086/183286
* Howe et al. (2018) Howe, R., Hill, F., Komm, R., et al. 2018, ApJ, 862, L5, doi: 10.3847/2041-8213/aad1ed
* Korzennik (2023) Korzennik, S. G. 2023, Frontiers in Astronomy and Space Sciences, 9, 1031313, doi: 10.3389/fspas.2022.1031313
* Kosovichev & Pipin (2019) Kosovichev, A. G., & Pipin, V. V. 2019, ApJ, 871, L20, doi: 10.3847/2041-8213/aafe82
* Kosovichev & Schou (1997) Kosovichev, A. G., & Schou, J. 1997, ApJ, 482, L207, doi: 10.1086/310708
* Leighton (1969) Leighton, R. B. 1969, ApJ, 156, 1, doi: 10.1086/149943
* Mandal et al. (2018) Mandal, K., Hanasoge, S. M., Rajaguru, S. P., & Antia, H. M. 2018, ApJ, 863, 39, doi: 10.3847/1538-4357/aacea2
* Parker (1955) Parker, E. N. 1955, ApJ, 122, 293, doi: 10.1086/146087
* Pijpers (1997) Pijpers, F. P. 1997, A&A, 326, 1235, doi: 10.48550/arXiv.astro-ph/9705234
* Pipin & Kosovichev (2019) Pipin, V. V., & Kosovichev, A. G. 2019, ApJ, 887, 215, doi: 10.3847/1538-4357/ab5952
* Pipin & Kosovichev (2020) —. 2020, ApJ, 900, 26, doi: 10.3847/1538-4357/aba4ad
* Pipin et al. (2023) Pipin, V. V., Kosovichev, A. G., & Tomin, V. E. 2023, The Astrophysical Journal, 949, 7, doi: 10.3847/1538-4357/acaf69
* Rajaguru & Antia (2015) Rajaguru, S. P., & Antia, H. M. 2015, ApJ, 813, 114, doi: 10.1088/0004-637X/813/2/114
* Ritzwoller & Lavely (1991) Ritzwoller, M. H., & Lavely, E. M. 1991, ApJ, 369, 557, doi: 10.1086/169785
* Schatten et al. (1978) Schatten, K. H., Scherrer, P. H., Svalgaard, L., & Wilcox, J. M. 1978, Geophys. Res. Lett., 5, 411, doi: 10.1029/GL005i005p00411
* Scherrer et al. (1995) Scherrer, P. H., Bogart, R. S., Bush, R. I., et al. 1995, Sol. Phys., 162, 129, doi: 10.1007/BF00733429
* Schou & Brown (1994) Schou, J., & Brown, T. M. 1994, A&AS, 107, 541
* Schou et al. (1998) Schou, J., Antia, H. M., Basu, S., et al. 1998, ApJ, 505, 390, doi: 10.1086/306146
* Schou et al. (2012) Schou, J., Scherrer, P. H., Bush, R. I., et al. 2012, Sol. Phys., 275, 229, doi: 10.1007/s11207-011-9842-2
* Stejko et al. (2022) Stejko, A. M., Kosovichev, A. G., Featherstone, N. A., et al. 2022, The Astrophysical Journal, 934, 161, doi: 10.3847/1538-4357/ac7a44
* Teweldebirhan et al. (2023) Teweldebirhan, K., Miesch, M., & Gibson, S. 2023, arXiv e-prints, arXiv:2310.00738, doi: 10.48550/arXiv.2310.00738
* Vorontsov (2011) Vorontsov, S. V. 2011, Monthly Notices of the Royal Astronomical Society, 418, 1146, doi: 10.1111/j.1365-2966.2011.19564.x
* Vorontsov et al. (2002) Vorontsov, S. V., Christensen-Dalsgaard, J., Schou, J., Strakhov, V. N., & Thompson, M. J. 2002, Science, 296, 101, doi: 10.1126/science.1069190
* Zhao et al. (2013) Zhao, J., Bogart, R. S., Kosovichev, A. G., Duvall, T. L., J., & Hartlep, T. 2013, ApJ, 774, L29, doi: 10.1088/2041-8205/774/2/L29
* Zhao et al. (2012) Zhao, J., Nagashima, K., Bogart, R. S., Kosovichev, A. G., & Duvall, T. L., J. 2012, ApJ, 749, L5, doi: 10.1088/2041-8205/749/1/L5
|
We present a microlocal analysis of two novel Radon transforms of interest in Compton Scattering Tomography (CST),
which map compactly supported
$L^2$ functions to their integrals over seven-dimensional sets of apple and lemon
surfaces. Specifically, we show that the apple and lemon transforms
are elliptic Fourier Integral Operators (FIO), which satisfy the
Bolker condition. After an analysis of the full seven-dimensional
case, we focus our attention on $n$-D subsets of apple and lemon surfaces
with fixed central axis, where $n<7$. Such subsets of surface integrals have
applications in airport baggage and security screening. When the data dimensionality is restricted, the
apple transform is shown to violate the Bolker condition, and there
are artifacts which occur on apple-cylinder intersections. The lemon transform is shown to satisfy the Bolker
condition, when the support of the function is restricted to the strip $\sparen{0<z<1}$.
[1]
L. Borg, J. Frikel, J. S. Jørgensen, and E. T. Quinto.
Analyzing reconstruction artifacts from arbitrary incomplete X-ray
CT data.
SIAM Journal on Imaging Sciences, 11(4):2786–2814, 2018.
[2]
J. Cebeiro, C. Tarpau, M. A. Morvidone, D. Rubio, and M. K. Nguyen.
On a three-dimensional compton scattering tomography system with
fixed source.
Inverse Problems, 37(5):054001, 2021.
[3]
J. J. Duistermaat.
Fourier integral operators, volume 130 of Progress in
Birkhäuser, Inc., Boston, MA, 1996.
[4]
J. J. Duistermaat and L. Hormander.
Fourier integral operators, volume 2.
Springer, 1996.
[5]
V. Guillemin and S. Sternberg.
Geometric Asymptotics.
American Mathematical Society, Providence, RI, 1977.
[6]
L. Hörmander.
The analysis of linear partial differential operators. I.
Classics in Mathematics. Springer-Verlag, Berlin, 2003.
Distribution theory and Fourier analysis, Reprint of the second
(1990) edition [Springer, Berlin].
[7]
L. Hörmander.
The analysis of linear partial differential operators. III.
Classics in Mathematics. Springer, Berlin, 2007.
Pseudo-differential operators, Reprint of the 1994 edition.
[8]
L. Hörmander.
The analysis of linear partial differential operators. IV.
Classics in Mathematics. Springer-Verlag, Berlin, 2009.
Fourier integral operators, Reprint of the 1994 edition.
[9]
C.-Y. Jung and S. Moon.
Inversion formulas for cone transforms arising in application of
Compton cameras.
Inverse Problems, 31(1):015006, 2015.
[10]
A. I. Katsevich.
Local tomography for the limited-angle problem.
J. Math. Anal. Appl., 213(1):160–182, 1997.
[11]
V. P. Krishnan and E. T. Quinto.
Microlocal analysis in tomography.
Handbook of mathematical methods in imaging, pages 1–50, 2014.
[12]
P. Kuchment and F. Terzioglu.
Three-dimensional image reconstruction from Compton camera data.
SIAM Journal on Imaging Sciences, 9(4):1708–1725, 2016.
[13]
S. Moon and M. Haltmeier.
Analytic inversion of a conical radon transform arising in
application of Compton cameras on the cylinder.
SIAM Journal on imaging sciences, 10(2):535–557, 2017.
[14]
M. K. Nguyen, T. T. Truong, and P. Grangeat.
Radon transforms on a class of cones with fixed axis direction.
Journal of Physics A: Mathematical and General, 38(37):8003,
[15]
E. T. Quinto.
The dependence of the generalized Radon transform on defining
Trans. Amer. Math. Soc., 257:331–346, 1980.
[16]
G. Rigaud.
3d compton scattering imaging with multiple scattering: Analysis by
fio and contour reconstruction.
Inverse Problems, 2021.
[17]
G. Rigaud and B. Hahn.
Reconstruction algorithm for 3d compton scattering imaging with
incomplete data.
Inverse Problems in Science and Engineering, 29(7):967–989,
[18]
G. Rigaud and B. N. Hahn.
3D Compton scattering imaging and contour reconstruction for a
class of Radon transforms.
Inverse Problems, 34(7):075004, 2018.
[19]
W. Rudin.
Functional analysis.
McGraw-Hill Book Co., New York, 1973.
McGraw-Hill Series in Higher Mathematics.
[20]
Sylvester's determinant theorem.
Accessed 11/30/2021.
[21]
J. J. Sylvester.
On the relation between the minor determinants of linearly
equivalent quadratic functions.
Philosophical Magazine, 1:295–305, 1851.
[22]
T. T. Truong, M. K. Nguyen, and H. Zaidi.
The mathematical foundations of 3D Compton scatter emission
International journal of biomedical imaging, 2007, 2007.
[23]
J. Webber and E. Miller.
Compton scattering tomography in translational geometries.
Technical report, Tufts University, 2019.
[24]
J. W. Webber and S. Holman.
Microlocal analysis of a spindle transform.
Inverse Problems & Imaging, 13(2):231–261, 2019.
[25]
J. W. Webber and W. R. Lionheart.
Three dimensional Compton scattering tomography.
Inverse Problems, 34(8):084001, 2018.
[26]
J. W. Webber and E. T. Quinto.
Microlocal analysis of generalized radon transforms from scattering
SIAM Journal on Imaging Sciences, 14(3):976–1003, 2021.
[27]
J. W. Webber, E. T. Quinto, and E. L. Miller.
A joint reconstruction and lambda tomography regularization technique
for energy-resolved x-ray imaging.
Inverse Problems, 36(7):074002, 2020.
|
# Realizing degree sequences as $Z_{3}$-connected graphs
Fan Yang1 , Xiangwen Li2 , Hong -Jian Lai3
1Department of Mathematics and Physics, Jiangsu University of Science and
Technology,
Zhenjiang 212003, China
2Department of Mathematics, Huazhong Normal University, Wuhan 430079, China
3Department of Mathematics, West Virginia University, Morgantown, WV 26506,
USA Supported by the Natural Science Foundation of China (11326215)Supported
by the Natural Science Foundation of China (11171129) and by Doctoral Fund of
Ministry of Education of China (20130144110001)Corresponding author:
<EMAIL_ADDRESS>
###### Abstract
An integer-valued sequence $\pi=(d_{1},\ldots,d_{n})$ is graphic if there is a
simple graph $G$ with degree sequence of $\pi$. We say the $\pi$ has a
realization $G$. Let $Z_{3}$ be a cyclic group of order three. A graph $G$ is
$Z_{3}$-connected if for every mapping $b:V(G)\to Z_{3}$ such that $\sum_{v\in
V(G)}b(v)=0$, there is an orientation of $G$ and a mapping $f:E(G)\to
Z_{3}-\\{0\\}$ such that for each vertex $v\in V(G)$, the sum of the values of
$f$ on all the edges leaving from $v$ minus the sum of the values of $f$ on
the all edges coming to $v$ is equal to $b(v)$. If an integer-valued sequence
$\pi$ has a realization $G$ which is $Z_{3}$-connected, then $\pi$ has a
$Z_{3}$-connected realization $G$. Let $\pi=(d_{1},\ldots,d_{n})$ be a graphic
sequence with $d_{1}\geq\ldots\geq d_{n}\geq 3$. We prove in this paper that
if $d_{1}\geq n-3$, then either $\pi$ has a $Z_{3}$-connected realization
unless the sequence is $(n-3,3^{n-1})$ or is $(k,3^{k})$ or $(k^{2},3^{k-1})$
where $k=n-1$ and $n$ is even; if $d_{n-5}\geq 4$, then either $\pi$ has a
$Z_{3}$-connected realization unless the sequence is $(5^{2},3^{4})$ or
$(5,3^{5})$.
## 1 Introduction
Graphs here are finite, and may have multiple edges without loops. We follow
the notation and terminology in [2] except otherwise stated.
For a given orientation of a graph $G$, if an edge $e\in E(G)$ is directed
from a vertex $u$ to a vertex $v$, then $u$ is the tail of $e$ and $v$ is the
head of $e$. For a vertex $v\in V(G)$, let $E^{+}(v)$ and $E^{-}(v)$ denote
the sets of all edges having tail $v$ or head $v$, respectively. A graph $G$
is $k$-flowable if all the edges of $G$ can be oriented and assigned nonzero
numbers with absolute value less than $k$ so that for every vertex $v\in
V(G)$, the sum of the values on all the edges in $E^{+}(v)$ equals that of the
values of all the edges in $E^{-}(v)$. If $G$ is $k$-flowable we also say that
$G$ admits a nowhere-zero $k$-flow.
Let $A$ be an abelian group with identity 0, and let $A^{*}=A-\\{0\\}$. Given
an orientation and a mapping $f:E(G)\rightarrow A$, the boundary of $f$ is a
function $\partial f:V(G)\rightarrow A$ defined by, for each vertex $v\in
V(G)$,
$\partial f(v)=\sum\limits_{e\in E^{+}(v)}f(e)-\sum\limits_{e\in
E^{-}(v)}f(e),$
where “$\sum$” refers to the addition in $A$.
A mapping $b:V(G)\rightarrow A$ is a zero-sum function if $\sum\limits_{v\in
V(G)}b(v)=0$. A graph $G$ is $A$-connected if for every zero-sum function
$b:V(G)\rightarrow A$, there exists an orientation of $G$ and a mapping
$f:E(G)\rightarrow A^{*}$ such that $\partial f(v)=b(v)$ for each $v\in V(G)$.
The concept of $k$-flowability was first introduced by Tutte [19], and this
theory provides an interesting way to investigate the coloring of planar
graphs in the sense that Tutte [19] proved a classical theorem: a planar graph
is $k$-colorable if and only if it is $k$-flowable. Jaeger et al. [10]
successfully generalized nowhere-zero flow problems to group connectivity. The
purpose of study in group connectivity is to characterize contractible
configurations for integer flow problems. Let $Z_{3}$ be a cyclic group of
order three. Obviously, if $G$ is $Z_{3}$-connected, then $G$ is $3$-flowable.
An integer-valued sequence $\pi=(d_{1},\ldots,d_{n})$ is graphic if there is a
simple graph $G$ with degree sequence $\pi$. We say $\pi$ has a realization
$G$, and we also say $G$ is a realization of $\pi$. If an integer-valued
sequence $\pi$ has a realization $G$ which is $A$-connected, then we say that
$G$ is a $A$-connected realization of $\pi$ for an abelian group $A$. In
particular, if $A=Z_{3}$, then $G$ is a $Z_{3}$-connected realization of
$\pi$, and we also say that $\pi$ has a $Z_{3}$-connected realization $G$. In
this paper, we write every degree sequence $(d_{1},\ldots,d_{n})$ is in
nonincrease order. For simplicity, we use exponents to denote degree
multiplicities, for example, we write $(6,5,4^{4},3)$ for $(6,5,4,4,4,4,3)$.
The problem of realizing degree sequences by graphs that have nowhere-zero
flows or are $A$-connected, where $A$ is an abelian group, has been studied.
Luo et al. [15] proved that every bipartite graphic sequence with least
element at least 2 has a 4-flowable realization. As a corollary, they
confirmed the simultaneous edge-coloring conjecture of Cameron [3]. Fan et al.
[7] proved that every degree sequence with least element at least 2 has a
realization which contains a spanning eulerian subgraph; such graphs is
4-flowable. Let $A$ be an abelian group with $|A|=4$. For a nonincreasing
$n$-element graphic sequence $\pi$ with least element at least 2 and sum at
least $3n-3$, Luo et al. [17] proved that $\pi$ has a realization that is
$A$-connected. Yin and Guo [20] determined the smallest degree sum that yields
graphic sequences with a $Z_{3}$-connected realization. For the literature for
this topic, the readers can see a survey [13]. In particular, Luo et al.[16]
completely answered the question of Archdeacon [1]: Characterize all graphic
sequences $\pi$ realizable by a 3-flowable graph. The natural group
connectivity version of Archdeacon’s problem is as follows.
###### Problem 1.1
Characterize all graphic sequences $\pi$ realizable by a $Z_{3}$-connected
graph.
On this problem, Luo et al. [16] obtained the next two results.
###### Theorem 1.2
Every nonincreasing graphic sequence $(d_{1},\ldots,d_{n})$ with $d_{1}=n-1$
and $d_{n}\geq 3$ has a $Z_{3}$-connected realization unless $n$ is even and
the sequence is $(k,3^{k})$ or $(k^{2},3^{k-1})$, where $k=n-1$.
###### Theorem 1.3
Every nonincreasing graphic sequence $(d_{1},\ldots,d_{n})$ with $d_{n}\geq 3$
and $d_{n-3}\geq 4$ has a $Z_{3}$-connected realization.
Motivated by Problem 1.1 and the results above, we present the following two
theorems in this paper. These results extend the results of [16] by extending
the characterizations to a large set of sequences.
###### Theorem 1.4
A nonincreasing graphic sequence $(d_{1},\ldots,d_{n})$ with $d_{1}\geq n-3$
and $d_{n}\geq 3$ has a $Z_{3}$-connected realization unless the sequence is
$(n-3,3^{n-1})$ for any $n$ or is $(k,3^{k})$ or $(k^{2},3^{k-1})$, where
$k=n-1$ and $n$ is even.
###### Theorem 1.5
A nonincreasing graphic sequence $(d_{1},\ldots,d_{n})$ with $d_{n}\geq 3$ and
$d_{n-5}\geq 4$ has a $Z_{3}$-connected realization unless the sequence is
$(5^{2},3^{4})$ or $(5,3^{5})$.
We end this section with some notation and terminology. A graph is trivial if
$E(G)=\emptyset$ and nontrivial otherwise. A $k$-vertex denotes a vertex of
degree $k$. Let $P_{n}$ denote the path on $n$ vertices and we call $P_{n}$ a
$n$-path. An $n$-cycle is a cycle on $n$ vertices. The wheel $W_{k}$ is the
graph obtained from a $k$-cycle by adding a new vertex, the center of the
wheel, and joining it to every vertex of the $k$-cycle. A wheel $W_{k}$ is an
odd (even) wheel if $k$ is odd (even). For simplicity, we say $W_{1}$ is a
triangle. For a graph $G$ and $X\subseteq V(G)$, denote by $G[X]$ the subgraph
of $G$ induced by $X$. For two vertex-disjoint subsets $V_{1},V_{2}$ of
$V(G)$, denote by $e(V_{1},V_{2})$ the number of edges with one endpoint in
$V_{1}$ and the other endpoint in $V_{2}$.
We organize this paper as follows. In Section 2, we state some results and
establish some lemmas that will be used in the following proofs. We will deal
with some special degree sequences, each of which has a $Z_{3}$-connected
realization in Section 3. In Sections 4 and 5, we will give the proofs of
Theorems 1.4 and 1.5.
## 2 Lemmas
Let $\pi=(d_{1},\ldots,d_{n})$ be a graphic sequence with $d_{1}\geq\ldots\geq
d_{n}$. Throughout this paper, we use $\bar{\pi}$ to represent the sequence
$(d_{1}-1,\ldots,d_{d_{n}}-1,d_{d_{n}+1},\ldots,d_{n-1})$, which is called the
_residual sequence_ obtained from $\pi$ by deleting $d_{n}$. The following
well-known result is due to Hakimi [8, 9] and Kleitman and Wang [11].
###### Theorem 2.1
A graphic sequence has even sum. Furthermore, a sequence $\pi$ is graphic if
and only if $\bar{\pi}$ is graphic.
Some results in [4, 5, 6, 10, 12] on group connectivity are summarized as
follows.
###### Lemma 2.2
Let $A$ be an abelian group with $|A|\geq 3$. The following results are known:
(1) $K_{1}$ is $A$-connected;
(2) $K_{n}$ and $K_{n}^{-}$ are $A$-connected if $n\geq 5$;
(3) An $n$-cycle is $A$-connected if and only if $|A|\geq n+1$;
(4) $K_{m,n}$ is $A$-connected if $m\geq n\geq 4$; neither $K_{2,t}$ $(t\geq
2)$ nor $K_{3,s}$ $(s\geq 3)$ is $Z_{3}$-connected;
(5) Each even wheel is $Z_{3}$-connected and each odd wheel is not;
(6) Let $H\subseteq G$ and $H$ be $A$-connected. $G$ is $A$-connected if and
only if $G/H$ is $A$-connected;
(7) If $G$ is not $A$-connected, then any spanning subgraph of $G$ is not
$A$-connected.
(8) Let $v$ be not a vertex of $G$. If $G$ is $A$-connected and $e(v,G)\geq
2$, then $G\cup\\{v\\}$ is $A$-connected.
Let $G$ be a graph having an induced path with three vertices $v,u,w$ in
order. Let $G_{[uv,uw]}$ be the graph by deleting $uv$ and $uw$ and adding a
new edge $vw$. The following lemma was first proved by Lai in [12] and
reformulated by Chen et al. in [4].
###### Lemma 2.3
Let $G$ be a graph with $u\in V(G)$, $uv,uw\in E(G)$ and $d(u)\geq 4$, and let
$A$ be an abelian group with $|A|\geq 3$. If $G_{[uv,uw]}$ is $A$-connected,
then so is $G$.
A graph $G$ is triangularly connected if for every edge $e,f\in E$ there
exists a sequence of cycles $C_{1},C_{2},\ldots,C_{k}$ such that $e\in
E(C_{1})$, $f\in E(C_{k})$, and $|E(C_{i})|\leq 3$ for $1\leq i\leq k$, and
$|E(C_{j})\cap E(C_{j+1})|\neq\emptyset$ for $1\leq j\leq k-1$.
###### Lemma 2.4
([5]) A triangularly connected graph $G$ is $Z_{3}$-connected if $G$ has
minimum degree at least 4 or has a nontrivial $Z_{3}$-connected subgraph.
An orientation $D$ of $G$ is a modular 3-orientation if
$|E^{+}(v)|-|E^{-}(v)|\equiv 0$ (mod 3) for every vertex $v\in V(G)$.
Steinberg and Younger [18] established the following relationship.
###### Lemma 2.5
A graph $G$ is 3-flowable if and only if $G$ admits a modular 3-orientation.
Let $v$ be a 3-vertex in a graph $G$, and let $N(v)=\\{v_{1},v_{2},v_{3}\\}$.
Denote by $G_{(v,v_{1})}$ the graph obtained from $G$ by deleting vertex $v$
and adding a new edge $v_{2}v_{3}$. The following lemma is due to Luo et al.
[14].
###### Lemma 2.6
Let $A$ be an abelian group with $|A|\geq 3$, and let $b:V(G)\mapsto A$ be a
zero-sum function with $b(v)\neq 0$. If $G_{(v,v_{1})}$ is $Z_{3}$-connected,
then there exists an orientation $D$ of $G$ and a nowhere-zero mapping
$f^{\prime}:E(G)\mapsto A$ such that $\partial f^{\prime}=b$ under the
orientation $D$ of $G$.
For any odd integer $k$, Luo et al. [16] proved that no realization of the
graphic sequence $(k,3^{k})$ and $(k^{2},3^{k-1})$ is 3-flowable. This yields
the following lemma.
###### Lemma 2.7
If $k$ is odd, then neither $(k,3^{k})$ nor $(k^{2},3^{k-1})$ has a
$Z_{3}$-connected realization.
Next we provide $Z_{3}$-connected realizations for some degree sequences.
$v_{1}$$v_{5}$$v_{3}$$v_{4}$$v_{6}$$v_{2}$(a) $\pi=(4^{2},3^{4})$
$v_{2}$$v_{1}$$v_{3}$$v_{4}$$v_{5}$$v_{6}$$v_{7}$(b) $\pi=(5,4,3^{5})$
$v_{5}$(120,25,130,25) (126,18,132,19) (130,25,132,19) (126,18,140,15)
(130,25,140,15) (120,25,126,18)
$v_{6}$$v_{7}$$v_{8}$$v_{1}$$v_{4}$$v_{3}$$v_{2}$(c) $\pi=(6,4,3^{6})$
$v_{5}$(120,25,130,25) (126,18,132,19) (130,25,132,19) (126,18,140,15)
(130,25,140,15) (120,25,126,18) (132,19,140,15)
$v_{6}$$v_{7}$$v_{8}$$v_{1}$$v_{4}$$v_{3}$$v_{2}$(d) $\pi=(5^{2},3^{6})$Fig.
1: Realizations of four degree sequences
###### Lemma 2.8
Each of the graphs in Fig. 1 is $Z_{3}$-connected.
Proof. If $G$ is the graph (a) in Fig. 1, then $G$ is $Z_{3}$-connected by
Lemma 2.2 of [14]. Thus, we may assume that $G$ is one of the graphs (b), (c)
and (d) shown in Fig 1.
We first prove here that the graph (b) in Fig. 1 is $Z_{3}$-connected. Assume
that $G$ is the graph (b) shown in Fig. 1. Let $b:V(G)\to Z_{3}$ be a zero-sum
function. If $b(v_{3})\neq 0$, then $G_{(v_{3},v_{4})}$ contains a 2-cycle
$(v_{1},v_{2})$. Contracting this 2-cycle and repeatedly contracting all
2-cycles generated in the process, we obtain $K_{1}$. By parts (1) and (6) of
Lemma 2.2, $G_{(v_{3},v_{4})}$ is $Z_{3}$-connected. It follows by Lemma 2.6
that there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial
f=b$. Thus, we may assume that $b(v_{3})=0$. Similarly, we may assume that
$b(v_{4})=b(v_{5})=b(v_{6})=b(v_{7})=0$. This means that for such $b$, there
are only three possibilities to be considered:
$(b(v_{1}),b(v_{2}))\in\\{(0,0),(1,2),(2,1)\\}$.
If $(b(v_{1}),b(v_{2}))=(0,0)$, we show that $G$ is 3-flowable. Note that each
vertex of $v_{3},v_{4},v_{5},v_{6}$ and $v_{7}$ is of degree 3. The edges of
$G$ are oriented as follows: $|E^{+}(v_{3})|=3$, $|E^{-}(v_{4})|=3$,
$|E^{+}(v_{5})|=3$, $|E^{-}(v_{6})|=3$, $|E^{+}(v_{7})|=3$, and $v_{2}v_{1}$
is oriented from $v_{2}$ to $v_{1}$. It is easy to verify that
$|E^{+}(v)|-|E^{-}(v)|=0$ (mod 3) for each vertex $v\in V(G)$. By Lemma 2.5,
$G$ is 3-flowable. Thus, there is an $f:E(G)\to Z_{3}^{*}$ such that $\partial
f(v)=b(v)$ for each $v\in V(G)$.
If $(b(v_{1}),b(v_{2})=(1,2)$, note that $b(v)=0$ for each $v\in
V(G)-\\{v_{1},v_{2}\\}$. The edges of $G$ are oriented as follows:
$|E^{-}(v_{3})|=3$, $|E^{+}(v_{4})|=3$, $|E^{-}(v_{5})|=3$,
$|E^{+}(v_{6})|=3$, $|E^{-}(v_{7})|=3$ and edge $v_{2}v_{1}$ is oriented from
$v_{2}$ to $v_{1}$. If $(b(v_{1}),b(v_{2})=(2,1)$, then the edges of $G$ are
oriented as follows: $|E^{+}(v_{3})|=3$, $|E^{-}(v_{4})|=3$,
$|E^{+}(v_{5})|=3$, $|E^{-}(v_{6})|=3$, $|E^{+}(v_{7})|=3$ and edge
$v_{1}v_{2}$ is oriented from $v_{1}$ to $v_{2}$. In each case, for each $e\in
E(G)$, define $f(e)=1$. It is easy to see that for
$v\in\\{v_{3},v_{4},v_{5},v_{6},v_{7}\\}$, $\partial f(v)=0=b(v)$, $\partial
f(v_{1})=b(v_{1})$ and $\partial f(v_{2})=b(v_{2})$.
Thus, for any zero-sum function $b$, there exists an orientation of $G$ and a
nowhere-zero mapping $f:E(G)\to Z_{3}$ such that $\partial f=b$. Therefore,
$G$ is $Z_{3}$-connected.
Next, assume that $G$ is the graph (c) shown in Fig. 1. Let $b:V(G)\to Z_{3}$
be a zero-sum function. If $b(v_{3})\neq 0$, then $G_{(v_{3},v_{4})}$ contains
an even wheel $W_{4}$ induced by $\\{v_{1},v_{2},v_{6},v_{7},v_{8}\\}$ with
the center $v_{1}$. Contracting this $W_{4}$ and recursively contracting all
2-cycles generated in the process, we get $K_{1}$. By parts (1), (5) and (6)
of Lemma 2.2, $G_{(v_{3},v_{4})}$ is $Z_{3}$-connected. By Lemma 2.6, there
exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial f=b$. Then we
may assume $b(v_{3})=0$. Similarly, $G_{(v_{5},v_{4})}$, $G_{(v_{6},v_{7})}$
and $G_{(v_{8},v_{7})}$ are all $Z_{3}$-connected. Thus, we may assume
$b(v_{5})=b(v_{6})=b(v_{8})=0$.
If $b(v_{4})\neq 0$, then $G_{(v_{4},v_{3})}$ contains a 2-cycle
$(v_{1},v_{5})$. Contracting this 2-cycle, we get the graph containing an even
wheel $W_{4}$ induced by $\\{v_{1},v_{2},v_{6},v_{7},v_{8}\\}$ with the center
at $v_{1}$. Contracting this wheel $W_{4}$ and contracting the 2-cycle
generated in the process, we get $K_{1}$. By parts (1), (3), (5) and (6) of
Lemma 2.2, $G_{(v_{4},v_{3})}$ is $Z_{3}$-connected. Lemma 2.6 shows that
there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial f=b$.
Thus, we may assume $b(v_{4})=0$. Similarly, $G_{(v_{7},v_{8})}$ is
$Z_{3}$-connected. Therefore, we may assume $b(v_{7})=0$.
Thus, we are left to consider the case that $b(v_{i})=0$ for $3\leq i\leq 8$.
This implies that for such function $b$, we need to consider three cases:
$b(v_{1})=b(v_{2})=0$; $b(v_{1})=1$ and $b(v_{2})=2$; $b(v_{1})=2$ and
$b(v_{2})=1$.
In the case that $b(v_{1})=b(v_{2})=0$, we have $b(v)=0$ for each $v\in V(G)$.
The edges of $G$ are oriented as follows: $|E^{+}(v_{3})|=3$,
$|E^{-}(v_{4})|=3$, $|E^{+}(v_{5})|=3$, $|E^{-}(v_{6})|=3$, $|E^{+}(v_{7})|=3$
and $|E^{-}(v_{8})|=3$. It is easy to verify that for each vertex $v\in V(G)$,
$|E^{+}(v)|-|E^{-}(v)|=0$ (mod 3). By Lemma 2.5, $G$ is 3-flowable. Thus,
there is an $f:E(G)\to Z_{3}^{*}$ such that $\partial f(v)=b(v)$ for each
$v\in V(G)$.
In the case that $b(v_{1})=1$ and $b(v_{2})=2$, we have $b(v)=0$ for each
$v\in V(G)-\\{v_{1},v_{2}\\}$. The edges of $G$ are oriented as follows:
$|E^{+}(v_{3})|=3$, $|E^{-}(v_{4})|=3$, $|E^{+}(v_{5})|=3$,
$|E^{+}(v_{6})|=3$, $|E^{-}(v_{7})|=3$, $|E^{+}(v_{8})|=3$. For each $e\in
E(G)$, define $f(e)=1$. It is easy to verify that $\partial f(v)=b(v)$ for
each $v\in V(G)$.
In the case that $b(v_{1})=2$ and $b(v_{2})=1$, we have $b(v)=0$ for each
$v\in V(G)-\\{v_{1},v_{2}\\}$. The edges of $G$ are oriented as follows:
$|E^{-}(v_{3})|=3$, $|E^{+}(v_{4})|=3$, $|E^{-}(v_{5})|=3$,
$|E^{-}(v_{6})|=3$, $|E^{+}(v_{7})|=3$, $|E^{-}(v_{8})|=3$. For each $e\in
E(G)$, define $f(e)=1$. It is easy to verify that $\partial f(v)=b(v)$ for
each $v\in V(G)$.
Thus, for every zero-sum function $b$, there exists an orientation of $G$ and
a mapping $f:E(G)\mapsto Z_{3}\setminus\\{0\\}$ such that $\partial f=b$.
Therefore, $G$ is $Z_{3}$-connected.
Finally, assume that $G$ is the graph (d) shown in Fig. 1. Let $b:V(G)\to
Z_{3}$ be a zero-sum function. As in the proof of the case when $G$ is the
graph (c) in Fig. 1, we may assume that that $b(v_{i})=0$ for $3\leq i\leq 8$.
This implies that for such $b$, we need to consider only three cases:
$b(v_{1})=b(v_{2})=0$; $b(v_{1})=1$ and $b(v_{2})=2$; $b(v_{1})=2$ and
$b(v_{2})=1$.
In the case that $b(v_{1})=b(v_{2})=0$, we have $b(v)=0$ for each $v\in V(G)$.
Assume that the edges of $G$ are oriented as follows: $|E^{+}(v_{3})|=3$,
$|E^{-}(v_{4})|=3$, $|E^{+}(v_{5})|=3$, $|E^{+}(v_{6})|=3$,
$|E^{-}(v_{7})|=3$, $|E^{+}(v_{8})|=3$. It is easy to verify that for each
vertex $v\in V(G)$, $|E^{+}(v)|-|E^{-}(v)|=0$ (mod 3). By Lemma 2.5, $G$ is
3-flowable. Thus, there is an $f:E(G)\to Z_{3}\setminus\\{0\\}$ such that
$\partial f(v)=b(v)$ for each $v\in V(G)$.
In the case that $b(v_{1})=1$ and $b(v_{2})=2$, we have $b(v)=0$ for each
$v\in V(G)-\\{v_{1},v_{2}\\}$. Assume that the edges of $G$ are oriented as
follows: $|E^{+}(v_{3})|=3$, $|E^{-}(v_{4})|=3$, $|E^{+}(v_{5})|=3$,
$|E^{-}(v_{6})|=3$, $|E^{+}(v_{7})|=3$, $|E^{-}(v_{8})|=3$. For each $e\in
E(G)$, define $f(e)=1$. It is easy to verify that $\partial f(v)=b(v)$ for
each $v\in V(G)$.
In the case that $b(v_{1})=2$ and $b(v_{2})=1$, we have $b(v)=0$ for each
$v\in V(G)-\\{v_{1},v_{2}\\}$. Assume that the edges of $G$ are oriented as
follows: $|E^{-}(v_{3})|=3$, $|E^{+}(v_{4})|=3$, $|E^{-}(v_{5})|=3$,
$|E^{+}(v_{6})|=3$, $|E^{-}(v_{7})|=3$, $|E^{+}(v_{8})|=3$. For each $e\in
E(G)$, define $f(e)=1$. It is easy to verify that $\partial f(v)=b(v)$ for
each $v\in V(G)$.
Thus, for every zero-sum function $b$, there exists an orientation of $G$ and
a mapping $f:E(G)\to Z_{3}\setminus\\{0\\}$ such that $\partial f=b$.
Therefore, $G$ is $Z_{3}$-connected.
###### Lemma 2.9
Each graph in Fig.2 is $Z_{3}$-connected.
Proof. We first prove here that the graph (a) in Fig. 2 is $Z_{3}$-connected.
Denote by $G$ the graph (a) in Fig. 2. We claim that $G$ is 3-flowable. Assume
that the edges of the graph are oriented as follows: $|E^{+}(v_{4})|=3$,
$|E^{+}(v_{5})|=0$, $|E^{+}(v_{6})|=3$, $|E^{+}(v_{7})|=0$ and $v_{2}v_{3}$
from $v_{2}$ to $v_{3}$. Define $f(e)=1$ for all $e\in E(G)$. It is easy to
verify that $\partial f(v)=0$ for each $v\in V(G)$. By Lemma 2.5, the graph
(a) is 3-flowable.
(10,35,25,28) (40,35,25,28) (30,25,25,28) (20,25,25,28) (20,25,40,15)
$v_{2}$$v_{1}$$v_{4}$$v_{5}$$v_{3}$$v_{6}$$v_{7}$(a) $\pi=(4^{3},3^{4})$
(75,20,70,15) $v_{2}$$v_{1}$$v_{7}$$v_{8}$$v_{6}$$v_{5}$$v_{3}$$v_{4}$(b)
$\pi=(5,4^{2},3^{5})$
$v_{3}$(125,30,110,15) (125,30,140,15) (125,30,110,35) (125,30,140,35)
(125,20,110,15) (125,20,140,15) (125,20,110,28) (125,20,140,28)
$v_{2}$$v_{7}$$v_{8}$$v_{1}$$v_{5}$$v_{6}$$v_{4}$(c) $\pi=(4^{4},3^{4})$Fig.
2: Realizations of three degree sequences
Let $b:V(G)\to Z_{3}$ be a zero-sum function. If $b(v_{4})\neq 0$, then
$G_{(v_{4},v_{2})}$ contains a 2-cycle $(v_{1},v_{5})$. Contracting the
2-cycle, we obtain an even wheel $W_{4}$ induced by
$\\{v_{1},v_{2},v_{3},v_{6},v_{7}\\}$ with the center at $v_{3}$. By parts
(3), (5) and (6) of Lemma 2.2, $G_{(v_{4},v_{2})}$ is $Z_{3}$-connected. By
Lemma 2.6, there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with
$\partial f=b$. Thus, we assume $b(v_{4})=0$. By symmetry, we may assume
$b(v_{5})=0$. If $b(v_{6})\neq 0$, then $G_{(v_{6},v_{3})}$ is a graph
isomorphic to Fig 1 (a) which is $Z_{3}$-connected by Lemma 2.8. By Lemma 2.6,
there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial f=b$. We
thus assume $b(v_{6})=0$. By symmetry, we assume $b(v_{7})=0$.
So far, we may assume $b(v_{4})=b(v_{5})=b(v_{6})=b(v_{7})=0$. We claim that
$b(v_{2})\not=0$. If $b(v_{2})=0$, then denote by $G(v_{2})$ the graph
obtained from $G$ by deleting $v_{2}$ and adding edges $v_{3}v_{7}$ and
$v_{4}v_{6}$. Contracting all 2-cycles, we finally get an even wheel $W_{4}$
with the center at $v_{1}$. By Lemma 2.2, $G(v_{2})$ is $Z_{3}$-connected.
Thus, there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial
f=b$. By symmetry, we assume that $b(v_{3})\not=0$. Thus, we are left to
discuss three cases
$(b(v_{1}),b(v_{2}),b(v_{3}))\in\\{(1,1,1),(0,1,2),(2,2,2)\\}$.
If $(b(v_{1}),b(v_{2}),b(v_{3}))=(1,1,1)$, then we orient the edges of $G$ as
follows: $|E^{+}(v_{4})|=3$, $|E^{+}(v_{5})|=0$, $|E^{+}(v_{6})|=3$,
$|E^{+}(v_{7})|=3$, and $v_{2}v_{3}$ from $v_{2}$ to $v_{3}$; if
$(b(v_{1}),b(v_{2}),b(v_{3}))=(0,1,2)$, then we orient edges of $G$ as
follows: $|E^{+}(v_{4})|=3$, $|E^{+}(v_{5})|=0$, $|E^{+}(v_{6})|=3$,
$|E^{+}(v_{7})|=0$, and $v_{3}v_{2}$ from $v_{3}$ to $v_{2}$; if
$(b(v_{1}),b(v_{2}),b(v_{3}))=(2,2,2)$, then we orient edges of $G$ as
follows: $|E^{+}(v_{4})|=3$, $|E^{+}(v_{5})|=0$, $|E^{+}(v_{6})|=0$,
$|E^{+}(v_{7})|=0$, and $v_{2}v_{3}$ from $v_{2}$ to $v_{3}$. In each case,
for each $e\in E(G)$ define $f(e)=1$. It is easy to verify that $\partial
f(v)=b(v)$ for each $v\in V(G)$.
In each case, there exists an orientation of $G$ and a nowhere-zero mapping
$f:E(G)\to Z_{3}$ such that $\partial f=b$. Therefore, $G$ is
$Z_{3}$-connected.
Next we consider $G$ is the graph $(b)$ in Fig. 2. Let $b:V(G)\to Z_{3}$ be a
zero-sum function. If $b(v_{4})\neq 0$, then $G_{(v_{4},v_{3})}$ is isomorphic
the graph shown in Fig 1 (b). By Lemma 2.8, $G_{(v_{4},v_{3})}$ is
$Z_{3}$-connected. By Lemma 2.6, there exists a nowhere-zero mapping
$f:E(G)\to Z_{3}$ with $\partial f=b$. Thus, we may assume $b(v_{4})=0$.
Similarly, if $b(v_{5})\not=0$, we can prove that $G_{(v_{5},v_{3})}$ is
$Z_{3}$-connected. Thus we also assume $b(v_{5})=0$.
If $b(v_{6})\neq 0$, then $G_{(v_{6},v_{5})}$ contains a 2-cycle
$(v_{1},v_{8})$. Contracting 2-cycles obtained will result in a graph $K_{1}$.
By parts (1), (3) and (6) of Lemma 2.2, $G_{(v_{6},v_{5})}$ is
$Z_{3}$-connected. By Lemma 2.6, there exists a nowhere-zero mapping
$f:E(G)\to Z_{3}$ with $\partial f=b$. Thus, we may assume $b(v_{6})=0$.
Similarly, if $b(v_{7})\not=0$, we can prove that $G_{(v_{7},v_{8})}$,
$G_{(v_{8},v_{6})}$ is $Z_{3}$-connected. Thus, we may assume
$b(v_{7})=b(v_{8})=0$.
From now on we may assume $b(v_{4})=b(v_{5})=b(v_{6})=b(v_{7})=b(v_{8})=0$. We
claim that $b(v_{2})\not=0$. If $b(v_{2})=0$, then denote by $G(v_{2})$ the
graph obtained from $G$ by deleting $v_{2}$ and adding edges $v_{1}v_{7}$ and
$v_{3}v_{4}$. By contracting all 2-cycles, finally we get $K_{1}$. By Lemma
2.2, $G(v_{2})$ is $Z_{3}$-connected. Thus, for $b(v_{2})=0$, there exists a
nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial f=b$. Similarly, we may
assume that $b(v_{3})\not=0$. Thus, it remains for us to discuss four cases
$(b(v_{1}),b(v_{2}),b(v_{3}))\in\\{(1,1,1),(0,1,2),(2,2,2),(0,2,1)\\}$.
If $(b(v_{1}),b(v_{2}),b(v_{3}))=(1,1,1)$, then we orient the edges of $G$ as
follows: $|E^{+}(v_{7})|=3$, $|E^{+}(v_{8})|=0$, $|E^{+}(v_{6})|=3$,
$|E^{+}(v_{5})|=0$, $|E^{+}(v_{4})|=3$ and $v_{1}v_{2}$ from $v_{1}$ to
$v_{2}$, $v_{2}v_{3}$ from $v_{2}$ to $v_{3}$, $v_{1}v_{3}$ from $v_{1}$ to
$v_{3}$; if $(b(v_{1}),b(v_{2}),b(v_{3}))=(0,1,2)$, then we orient the edges
of $G$ as follows: $|E^{+}(v_{7})|=3$, $|E^{+}(v_{8})|=0$, $|E^{+}(v_{6})|=3$,
$|E^{+}(v_{5})|=0$, $|E^{+}(v_{4})|=3$ and $v_{2}v_{1}$ from $v_{2}$ to
$v_{1}$, $v_{3}v_{2}$ from $v_{3}$ to $v_{2}$, $v_{3}v_{1}$ from $v_{3}$ to
$v_{1}$; if $(b(v_{1}),b(v_{2}),b(v_{3}))=(2,2,2)$, then we orient the edges
of $G$ as follows: $|E^{+}(v_{7})|=0$, $|E^{+}(v_{8})|=3$, $|E^{+}(v_{6})|=0$,
$|E^{+}(v_{5})|=3$, $|E^{+}(v_{4})|=0$ and $v_{2}v_{1}$ from $v_{2}$ to
$v_{1}$, $v_{3}v_{2}$ from $v_{3}$ to $v_{2}$, $v_{3}v_{1}$ from $v_{3}$ to
$v_{1}$; if $(b(v_{1}),b(v_{2}),b(v_{3}))=(0,2,1)$, then we orient the edges
of $G$ as follows: $|E^{+}(v_{7})|=0$, $|E^{+}(v_{8})|=3$, $|E^{+}(v_{6})|=0$,
$|E^{+}(v_{5})|=3$, $|E^{+}(v_{4})|=0$ and $v_{1}v_{2}$ from $v_{1}$ to
$v_{2}$, $v_{2}v_{3}$ from $v_{2}$ to $v_{3}$, $v_{1}v_{3}$ from $v_{1}$ to
$v_{3}$. In each case, for each $e\in E(G)$ define $f(e)=1$ in such
orientation. Clearly, we can verify that $\partial f(v)=b(v)$ for each $v\in
V(G)$. Thus, $G$ is $Z_{3}$-connected.
We are left to consider the case that $G$ is the graph $(c)$ in Fig. 2. Let
$b:V(G)\to Z_{3}$ be a zero-sum function. If $b(v_{5})\neq 0$, then
$G_{(v_{5},v_{6})}$ contains an even wheel $W_{4}$ with the center at $v_{3}$.
By parts (5) and (8) of Lemma 2.2, $G_{(v_{5},v_{6})}$ is $Z_{3}$-connected.
By Lemma 2.6, there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with
$\partial f=b$. Thus, we assume that $b(v_{5})=0$. By symmetry, we may assume
that $b(v_{6})=0$.
If $b(v_{7})\neq 0$, then $G_{(v_{7},v_{5})}$ contains a 2-cycle
$(v_{2},v_{3})$. Contracting this 2-cycle and repeatedly contracting all
2-cycles generated in the process, we finally get $K_{1}$. By parts (1) and
(6) of Lemma 2.2, $G_{(v_{7},v_{5})}$ is $Z_{3}$-connected. By Lemma 2.6,
there exists a nowhere-zero mapping $f:E(G)\to Z_{3}$ with $\partial f=b$.
Thus, we may assume $b(v_{7})=0$. By symmetry, we may assume $b(v_{8})=0$.
If $b(v_{1})=0$, then denote by $G(v_{1})$ the graph obtained from $G$ by
deleting $v_{1}$ and adding edges $v_{5}v_{6}$ and $v_{3}v_{4}$. Contracting
all 2-cycles in the process, we finally get $K_{1}$. By Lemma 2.2, $G(v_{1})$
is $Z_{3}$-connected. Thus, for $b(v_{1})=0$, there exists a nowhere-zero
mapping $f:E(G)\to Z_{3}$ with $\partial f=b$. Thus, we may assume that
$b(v_{1})\not=0$. Similarly, we also assume that $b(v_{2})\not=0$,
$b(v_{3})\not=0$ and $b(v_{4})\not=0$. Thus, we only need to discuss four
cases
$(b(v_{1}),b(v_{2}),b(v_{3}),b(v_{4}))\in\\{(2,2,1,1),(2,1,2,1),(1,2,1,2),(1,1,2,2)\\}$
and $b(v_{5})=b(v_{6})=b(v_{7})=b(v_{8})=0$.
If $(b(v_{1}),b(v_{2}),b(v_{3}),b(v_{4}))=(2,2,1,1)$, then we orient the edges
of $G$ as follows: $|E^{+}(v_{7})|=0$, $|E^{+}(v_{5})|=3$, $|E^{+}(v_{6})|=0$,
$|E^{+}(v_{8})|=3$, and $v_{1}v_{3}$ from $v_{1}$ to $v_{3}$, $v_{1}v_{4}$
from $v_{1}$ to $v_{4}$, $v_{2}v_{3}$ from $v_{2}$ to $v_{3}$, $v_{2}v_{4}$
from $v_{2}$ to $v_{4}$, $v_{4}v_{3}$ from $v_{4}$ to $v_{3}$; if
$(b(v_{1}),b(v_{2}),b(v_{3}),b(v_{4}))=(2,1,2,1)$, then we orient the edges of
$G$ as follows: $|E^{+}(v_{7})|=0$, $|E^{+}(v_{5})|=3$, $|E^{+}(v_{6})|=0$,
$|E^{+}(v_{8})|=3$, and $v_{1}v_{3}$ from $v_{1}$ to $v_{3}$, $v_{1}v_{4}$
from $v_{1}$ to $v_{4}$, $v_{3}v_{2}$ from $v_{3}$ to $v_{2}$, $v_{4}v_{2}$
from $v_{4}$ to $v_{2}$, $v_{3}v_{4}$ from $v_{3}$ to $v_{4}$; if
$(b(v_{1}),b(v_{2}),b(v_{3}),b(v_{4}))=(1,2,1,2)$, then we orient the edges of
$G$ as follows: $|E^{+}(v_{7})|=3$, $|E^{+}(v_{5})|=0$, $|E^{+}(v_{6})|=3$,
$|E^{+}(v_{8})|=0$, and $v_{3}v_{1}$ from $v_{3}$ to $v_{1}$, $v_{4}v_{1}$
from $v_{4}$ to $v_{1}$, $v_{2}v_{3}$ from $v_{2}$ to $v_{3}$, $v_{2}v_{4}$
from $v_{2}$ to $v_{4}$, $v_{4}v_{3}$ from $v_{4}$ to $v_{3}$; if
$(b(v_{1}),b(v_{2}),b(v_{3}),b(v_{4}))=(1,1,2,2)$, then we orient the edges of
$G$ as follows: $|E^{+}(v_{7})|=0$, $|E^{+}(v_{5})|=3$, $|E^{+}(v_{6})|=0$,
$|E^{+}(v_{8})|=3$, and $v_{3}v_{1}$ from $v_{3}$ to $v_{1}$, $v_{4}v_{1}$
from $v_{4}$ to $v_{1}$, $v_{3}v_{2}$ from $v_{3}$ to $v_{2}$, $v_{4}v_{2}$
from $v_{4}$ to $v_{2}$, $v_{4}v_{3}$ from $v_{4}$ to $v_{3}$. In each case,
for each $e\in E(G)$ define $f(e)=1$ in such orientation. Clearly, we can
verify that $\partial f(v)=b(v)$ for each $v\in V(G)$. Thus, $G$ is
$Z_{3}$-connected.
## 3 Some special cases
Throughout this section, all sequences are graphic sequences. We provide
$Z_{3}$-connected realizations for some graphic sequences.
###### Lemma 3.1
Suppose that one of the following holds,
(i) $n\geq 6$ and $\pi=(n-2,4,3^{n-2})$;
(ii) $n\geq 5$ and $\pi=(4^{n-4},3^{4})$;
(iii) $n\geq 7$ and $\pi=(5,4^{n-6},3^{5})$.
Then $\pi$ has a $Z_{3}$-connected realization.
Proof. (i) If $n=6$, then by Lemma 2.8, $\pi$ has a $Z_{3}$-connected
realization $G$ in Fig. 1 (a). Thus, we assume that $n\geq 7$.
If $n=7,8$, then by Lemma 2.8, $\pi$ has a $Z_{3}$-connected realization $G$
in Fig. 1 (b) (c). Thus, we assume that $n\geq 9$.
Assume that $n$ is odd. Let $W_{n-5}$ be an even wheel with the center at
$v_{1}$ and $K_{4}^{-}$ on vertex set $\\{u_{1},u_{2},u_{3},u_{4}\\}$ with
$d_{K_{4}^{-}}(u_{1})=d_{K_{4}^{-}}(u_{3})=2$. Denote by $G$ the graph
obtained from $W_{n-5}$ and $K_{4}^{-}$ by adding edges $u_{i}v_{1}$ for each
$i\in\\{1,2,3\\}$. Obviously, the graph $G$ has a degree sequence
$(n-2,4,3^{n-2})$. By part (5) of Lemma 2.2, $W_{n-5}$ is $Z_{3}$-connected.
The graph $G/W_{n-5}$ is an even wheel $W_{4}$. By part (5) and (6) of Lemma
2.2, $G$ is $Z_{3}$-connected. This means that $\pi$ has a $Z_{3}$-connected
realization.
Assume that $n$ is even. Let $G_{0}$ be the graph in Fig. 1 (a) and $W_{n-6}$
be an even wheel with the center at $u_{1}$. Denote by $G$ the graph obtained
from $W_{n-6}$ and $G_{0}$ by identifying $u_{1}$ and $v_{1}$. Clearly, $G$
has a degree sequence $(n-2,4,3^{n-2})$. Since $n\geq 10$ is even, $W_{n-6}$
is $Z_{3}$-connected by (5) of Lemma 2.2. By Lemma 2.8, $G_{0}$ is
$Z_{3}$-connected. This shows that $G$ is $Z_{3}$-connected.
(ii) If $n=5$, then an even wheel $W_{4}$ is a $Z_{3}$-connected realization
of $\pi$; if $n=6$, then by Lemma 2.8, $\pi$ has a $Z_{3}$-connected
realization $G$ in Fig. 1 (a); if $n=7$, then by Lemma 2.9, $\pi$ has a
$Z_{3}$-connected realization shown in Fig. 2 (a); if $n=8$, then by Lemma
2.9, the graph (c) in Fig. 2 is $Z_{3}$-connected realization of $\pi$. If
$n=9$, then let $G_{1}$ be an even wheel $W_{4}$ induced by
$\\{u_{0},u_{1},u_{2},u_{3},u_{4}\\}$ with the center at $u_{0}$ and $G_{2}$
be a $K_{4}^{-}$ induced by $\\{v_{1},v_{2},v_{3},v_{4}\\}$ with
$d_{G_{2}}(v_{1})=d_{G_{2}}(v_{3})=3$. We construct a graph $G$ from $W_{4}$
and $K_{4}^{-}$ by adding three edges $u_{1}v_{2}$, $u_{2}v_{4}$ and
$u_{3}v_{1}$. Then $G$ is a $Z_{3}$-connected realization of $(4^{5},3^{4})$.
Thus, we assume that $n\geq 10$.
Assume that $n=2k$, where $k\geq 5$. By induction of hypothesis, let $G_{i}$
be a $Z_{3}$-connected realization of the degree sequence $(4^{k-4},3^{4})$
for $i\in\\{1,2\\}$. Assume that $n=2k+1$, where $k\geq 5$. By induction
hypothesis, let $G_{1}$ be a $Z_{3}$-connected realization of the degree
sequence $(4^{k-4},3^{4})$ and $G_{2}$ be a $Z_{3}$-connected realization of
the degree sequence $(4^{k-3},3^{4})$. In each case, we construct a graph $G$
from $G_{1}$ and $G_{2}$ by connecting a pair of 3-vertices of $G_{1}$ to a
pair of 3-vertices of $G_{2}$ one by one. It is easy to verify that $G$ is a
$Z_{3}$-connected realization of the degree sequence $(4^{n-4},3^{4})$.
(iii) If $n=7$, then by Lemma 2.8, the graph (b) in Fig. 1 is a
$Z_{3}$-connected realization of $\pi$; if $n=8$, then by Lemma 2.9, the graph
(b) in Fig. 2 is a $Z_{3}$-connected realization of $\pi$. If $n=9$, then
$\pi=(5,4^{3},3^{5})$. Let $G_{1}$ be an even wheel $W_{4}$ induced by
$\\{u_{0},u_{1},u_{2},u_{3},u_{4}\\}$ with the center at $u_{0}$ and $G_{2}$
be a $K_{4}^{-}$ induced by $\\{v_{1},v_{2},v_{3},v_{4}\\}$ with
$d_{G_{2}}(v_{1})=d_{G_{2}}(v_{3})=3$. We construct a graph $G$ from $W_{4}$
and $K_{4}^{-}$ by adding three edges $u_{0}v_{2}$, $u_{1}v_{1}$,
$u_{2}v_{4}$. We conclude that $G$ is a $Z_{3}$-connected realization of
degree sequence $(5,4^{3},3^{5})$. Thus, $n\geq 10$.
Assume that $n=2k$, where $k\geq 5$. By (ii), let $G_{1}$ and $G_{2}$ be
$Z_{3}$-connected realizations of degree sequence $(4^{k-4},3^{4})$. Assume
that $n=2k+1$, where $k\geq 5$. By (ii), let $G_{1}$ be a $Z_{3}$-connected
realization of degree sequence $(4^{k-4},3^{4})$ and $G_{2}$ be a
$Z_{3}$-connected realization of degree sequence $(4^{k-3},3^{4})$. In each
case, choose one 4-vertex $u_{1}$ and one 3-vertex $u_{2}$ of $G_{1}$; choose
two 3-vertices $v_{1},v_{2}$ of $G_{2}$. We construct a graph $G$ from $G_{1}$
and $G_{2}$ by adding $u_{1}v_{1}$ and $u_{2}v_{2}$. Thus, $G$ is a
$Z_{3}$-connected realization of degree sequence $(5,4^{n-6},3^{5})$.
###### Lemma 3.2
If $\pi=(n-3,3^{n-1})$, then $\pi$ has not a $Z_{3}$-connected realization.
Proof. Suppose otherwise that $G$ has a $Z_{3}$-connected realization of
degree sequence $(n-3,3^{n-1})$. Let
$V(G)=\\{u,u_{1},\ldots,u_{n-3},x_{1},x_{2}\\}$,
$N_{G}(u)=\\{u_{1},\ldots,u_{n-3}\\}$ ($N$ for short), and
$X=\\{x_{1},x_{2}\\}$. We now consider the following two cases.
Case 1. $x_{1}x_{2}\in E(G)$.
Since $G$ is $Z_{3}$-connected, $G$ is 3-flowable. By Lemma 2.5 and symmetry,
we may assume that $|E^{+}(x_{1})|=3$ and $|E^{-}(x_{2})|=3$. Since
$d(u_{i})=3$ for $i\in\\{1,\ldots,n-3\\}$, by Lemma 2.5, either
$|E^{+}(u_{i})|=3$ or $|E^{-}(u_{i})|=3$. This implies that there exists no
vertex $u_{i}$ in $N$ such that $u_{i}x_{1},u_{i}x_{2}\in E(G)$. Thus, $G[N]$
is the union of two paths $P_{1}$ and $P_{2}$. We relabel the vertices of $N$
such that $P_{1}=u_{1}\ldots u_{k}$ and $P_{2}=u_{k+1}\ldots u_{n-3}$.
Suppose first that $x_{1}u_{1},x_{1}u_{k}\in E(G)$ and
$x_{2}u_{k+1},x_{2}u_{n-3}\in E(G)$. Since $G$ is 3-flowable, by Lemma 2.5,
$P_{i}$ contains odd number of vertices for each $i\in\\{1,2\\}$. Define
$b(u)=b(x_{1})=b(x_{2})=1$ and $b(u_{i})=0$ for each $i\in\\{1,\ldots,n-3\\}$.
It is easy to verify that there exists no $f:E(G)\to Z_{3}^{*}$ such that
$\partial f(v)=b(v)$ for each $v\in V(G)$, contrary to that $G$ is
$Z_{3}$-connected.
Next, suppose that $x_{1}u_{1},x_{1}u_{k+1}\in E(G)$ and
$x_{2}u_{k},x_{2}u_{n-3}\in E(G)$. Since $G$ is 3-flowable, by Lemma 2.5
$P_{i}$ contains even number of vertices for each $i\in\\{1,2\\}$. Define
$b(u)=1$, $b(x_{2})=2$ and $b(u_{i})=b(x_{1})=0$ for each
$i\in\\{1,\ldots,n-3\\}$. It is easy to verify that there exists no $f:E(G)\to
Z_{3}^{*}$ such that $\partial f(v)=b(v)$ for each $v\in V(G)$, contrary to
that $G$ is $Z_{3}$-connected.
Case 2. $x_{1}x_{2}\notin E(G)$.
Since $d(x_{i})=3$ for each $i=1,2$, $0\leq|N(x_{1})\cap N(x_{2})|\leq 3$.
Assume first that $|N(x_{1})\cap N(x_{2})|=3$. We assume, without loss of
generality, that $u_{1},u_{2},u_{3}\in N(x_{1})\cap N(x_{2})$. The subgraph
induced by $\\{u,x_{1},x_{2},u_{1},u_{2},u_{3}\\}$ is $K_{3,3}$ which is not
$Z_{3}$-connected by part (4) of Lemma 2.2. By part (6) of Lemma 2.2, $G$ is
not $Z_{3}$-connected, a contradiction.
Assume that $|N(x_{1})\cap N(x_{2})|=2$. We assume, without loss of
generality, that $u_{1},u_{2}\in N(x_{1})\cap N(x_{2})$. Since $G$ is
3-flowable, the graph $H$ induced by $N\setminus\\{u_{1},u_{2}\\}$ consists of
even cycles and a path of length even. This means that $n$ is even. If $n=6$,
then this case cannot occur. Thus $n\geq 8$. Define $b(x_{1})=1$, $b(x_{2})=2$
and $b(u_{i})=b(u)=0$ for each $i\in\\{1,\ldots,n-3\\}$. It is easy to verify
that there exists no $f:E(G)\to Z_{3}^{*}$ such that $\partial f(v)=b(v)$ for
each $v\in V(G)$, contrary to that $G$ is $Z_{3}$-connected.
Next, assume that $|N(x_{1})\cap N(x_{2})|=1$. We assume, without loss of
generality, that $u_{1}\in N(x_{1})\cap N(x_{2})$. The graph induced by
$N\setminus\\{u_{1}\\}$ consists of even cycles and two paths $P_{1}$ and
$P_{2}$. Since $G$ is 3-flowable, $P_{i}$ contains odd vertices for each
$i\in\\{1,2\\}$. Then $n$ is even. If $n=6,8$, then this case cannot occur.
Thus, we assume that $n\geq 10$. Define $b(x_{1})=1$, $b(x_{2})=2$ and
$b(u_{i})=b(u)=0$ for each $i\in\\{1,\ldots,n-3\\}$. In this case, there
exists no $f:E(G)\to Z_{3}^{*}$ such that $\partial f(v)=b(v)$ for each $v\in
V(G)$, contrary to that $G$ is $Z_{3}$-connected.
Finally, assume that $|N(x_{1})\cap N(x_{2})|=0$. Then the graph induced by
the vertices of $N$ consists of three paths $P_{1}$, $P_{2}$ and $P_{3}$,
together with even cycles. We relabel the vertices of $N$ such that
$P_{1}=u_{1}\ldots u_{s}$, $P_{2}=u_{s+1}\ldots u_{t}$ and
$P_{3}=u_{t+1}\ldots u_{n-3}$. By symmetry, we consider two cases: $x_{1}$ is
adjacent to both the end vertices of some $P_{i}$; $x_{1}$ is adjacent to one
of each $P_{j}$ for $j\in\\{1,2,3\\}$.
In the former case, we may assume that $u_{1}x_{1},u_{s}x_{1}\in E(G)$ and
$x_{2}u_{s+1},x_{2}u_{t}$. Since $G$ is 3-flowable, by Lemma 2.5, both
$|V(P_{1})|$ and $|V(P_{2})|$ are odd. If $|V(P_{3})|$ is odd, then define
$b(x_{1})=1$, $b(x_{2})=2$ and $b(u_{i})=b(u)=0$ for each
$i\in\\{1,\ldots,n-3\\}$. If $|V(P_{3})|$ is even, then define $b(x_{1})=1$,
$b(x_{2})=1$, $b(u)=1$ and $b(u_{i})=0$ for each $i\in\\{1,\ldots,n-3\\}$. In
either case, there exists no $f:E(G)\to Z_{3}^{*}$ such that $\partial
f(v)=b(v)$ for each $v\in V(G)$, contrary to that $G$ is $Z_{3}$-connected.
In the latter case,
$x_{1}u_{1},x_{1}u_{s+1},x_{1}u_{t+1},x_{2}u_{s},x_{2}u_{t},x_{2}u_{n-3}\in
E(G)$. It follows that $|V(P_{1})|,|V(P_{2})|$ and $|V(P_{3})|$ have the same
parity. If each of $|V(P_{i})|$ for $i\in\\{1,2,3\\}$ is even, then define
$b(x_{1})=1$, $b(x_{2})=1$, $b(u)=1$ and $b(u_{i})=0$ for each
$i\in\\{1,\ldots,n-3\\}$. If each of $|V(P_{i})|$ for $i\in\\{1,2,3\\}$ is
odd, then define define $b(x_{1})=1$, $b(x_{2})=2$, $b(u)=b(u_{i})=0$ for each
$i\in\\{1,\ldots,n-3\\}$. In either case, there exists no $f:E(G)\to
Z_{3}^{*}$ such that $\partial f(v)=b(v)$ for each $v\in V(G)$, contrary to
that $G$ is $Z_{3}$-connected.
## 4 Proof of Theorem 1.4
In order to prove Theorem 1.4, we establish the following lemma.
###### Lemma 4.1
Suppose that $\pi=(d_{1},\ldots,d_{n})$ is a nonincreasing graphic sequence
with $d_{n}\geq 3$. If $d_{1}=n-2$, then $\pi$ has a $Z_{3}$-connected
realization.
Proof. Suppose, to the contrary, that $\pi=(d_{1},\ldots,d_{n})$ has no
$Z_{3}$-connected realization with $n$ minimized, where $d_{1}=n-2$. By
Theorem 1.3, we may assume that $d_{n-3}\leq 3$. In order to prove our lemma,
we need the following claim.
Claim 1. Each of the following holds.
(i) $d_{n-3}=d_{n-2}=d_{n-1}=d_{n}=3$;
(ii) $n\geq 6$.
Proof of Claim 1. (i) follows since $d_{n}\geq 3$.
(ii) Since $d_{n}=3$, $n\geq 4$. If $n=4$, then $d_{1}=3=n-1$, contrary to
that $d_{1}=n-2$. If $n=5$, then $\pi=(3^{5})$ is not graphic by Theorem 2.1.
This proves Claim 1.
If $n=6$, then $d_{1}=4$ and $d_{3}=d_{4}=d_{5}=d_{6}=3$. By Theorem 2.1,
$d_{2}=4$. By Lemma 2.8, $\pi=(4^{2},3^{4})$ has a $Z_{3}$-realization, a
contradiction. Thus, we may assume that $n\geq 7$.
Claim 2. $d_{3}=3$.
Proof of Claim 2. Suppose otherwise that $d_{3}\geq 4$ and $G$ is a
counterexample with $|V(G)|=n$ minimized. Then $d_{2}\geq d_{3}\geq 4$. Hence,
$\bar{\pi}=(n-3,d_{2}-1,d_{3}-1,d_{4},\ldots,d_{n-1})$=$(\bar{d_{1}},\ldots,\bar{d}_{n-1})$
with $\bar{d_{1}}\geq\ldots\geq\bar{d}_{n-1}$. This implies that
$\bar{d}_{n-1}\geq 3$ and $\bar{d_{1}}=(n-1)-2$ or $\bar{d_{1}}=d_{4}$. In the
former case, since $\bar{d_{1}}=n-3=(n-1)-2$, by the minimality of $n$,
$\bar{\pi}$ has a $Z_{3}$-connected realization $\bar{G}$. In the latter case,
$\bar{d_{1}}\neq n-3$ and hence $\bar{d_{1}}=d_{4}>n-3$. Since $d_{1}=n-2\geq
d_{4}$, $d_{4}=n-2$. This means that $\bar{d_{1}}=d_{4}=(n-1)-1$. By Theorem
1.2, either $\bar{\pi}$ has a $Z_{3}$-connected realization $\bar{G}$ or
$\bar{\pi}=(k,3^{k})$, $(k^{2},3^{k-1})$, where $k$ is odd. If
$\bar{\pi}=(k,3^{k})$, then $d_{1}=k+1=n-2$ and $d_{2}=d_{3}=4$. On the other
hand, $n=k+1+1=k+2$. This contradiction proves that $\pi\not=(k,3^{k})$.
Similarly, $\bar{\pi}\not=(k^{2},3^{k-1})$. If $\bar{\pi}$ has a
$Z_{3}$-connected realization $\bar{G}$, then $\pi$ has a realization $G$ of
$\pi$ from $\bar{G}$ by adding a new vertex $v$ and three edges joining $v$ to
the corresponding vertices of $\bar{G}$. By part (8) of Lemma 2.2, $G$ is
$Z_{3}$-connected, a contradiction. Thus $d_{3}\leq 3$. Clearly, $d_{3}\geq
3$. Then $d_{3}=3$. This proves Claim 2.
By Claims 1 and 2, $\pi=(n-2,d_{2},3^{n-2})$. Since $\pi$ is graphic, $d_{2}$
is even whenever $n$ is even or odd. Moreover, $d_{2}\geq 4$. Recall that
$n\geq 7$. In this case, $\pi=(n-2,4,3^{n-2})$. By (i) of Lemma 3.1, $\pi$ has
a $Z_{3}$-connected realization $G$, a contradiction. Thus, we may assume that
$d_{2}\geq 6$. Since $n-2=d_{1}\geq d_{2}\geq 6$, $n\geq d_{2}+2\geq 8$.
Consider the case that $n$ is even. Denote by $W_{n-d_{2}+2}$ an even wheel
with the center at $v_{1}$ and by $S$ a vertex set such that $|S|=d_{2}-4$ and
$V(W_{n-d_{2}+2})\cap S=\emptyset$. Note that $|S|$ is even. We construct a
graph $G$ from $W_{n-d_{2}+2}$ and $S$ as follows: First, pick two vertices
$s_{1},s_{2}$ of $S$ and add $(d_{2}-6)/2$ edges such that the subgraph
induced by $S\setminus\\{s_{1},s_{2}\\}$ is a perfect matching. Second, let
$v_{1}$ connect to each vertex of $S$. Third, pick a vertex $v_{2}$ in
$W_{n-d_{2}+2}$ and let $v_{2}$ join to each vertex of $S$. Finally, add one
new vertex $x$ adjacent to $v_{2}$, $s_{1}$ and $s_{2}$.
We claim that $G$ has a degree sequence $(n-2,d_{2},3^{n-2})$. Since
$d_{W_{n-d_{2}+2}}(v_{1})=n-d_{2}+2\geq 4$, $d(v_{1})=n-d_{2}+2+d_{2}-4=n-2$,
$d(v_{2})=3+d_{2}-4+1=d_{2}$, each vertex of $V(G)\setminus\\{v_{1},v_{2}\\}$
is a 3-vertex. Since $W_{n-d_{2}+2}$ is an even wheel, by part (5) of Lemma
2.2, this wheel is $Z_{3}$-connected. By part (8) of Lemma 2.2, $G$ is
$Z_{3}$-connected, a contradiction.
Consider the case that $n$ is odd. Denote by $W_{n-d_{2}+1}$ an even wheel
with the center at $v_{1}$ and by $S$ a vertex set with $|S|=d_{2}-3$ and
$V(W_{n-d_{2}+1})\cap S=\emptyset$. We construct a graph $G$ from
$W_{n-d_{2}+1}$ and $S$ as follows: First, let $v_{1}$ connect to each vertex
of $S$. Second, pick one vertex $v_{2}$ in $W_{n-d_{2}+1}$ and let $v_{2}$
join to each vertex of $S$. Third, add one vertex $x$ adjacent to three
vertices of $S$. Finally, add $(d_{2}-6)/2$ edges in $S$ so that the subgraph
induced by vertices of $S$, each of which is not adjacent to $x$, is a perfect
matching. We claim that $G$ is a realization of degree sequence
$(n-2,d_{2},3^{n-2})$. Since $d_{W_{n-d_{2}+1}}(v_{1})=n-d_{2}+1\geq 4$,
$d(v_{1})=n-d_{2}+1+d_{2}-3=n-2$. Note that $d(v_{2})=3+d_{2}-3=d_{2}$, each
vertex of $V(G)\setminus\\{v_{1},v_{2}\\}$ is a 3-vertex. Similarly, it can be
verified that $G$ is a $Z_{3}$-connected realization of $\pi$, a
contradiction.
Proof of Theorem 1.4. Assume that $\pi=(d_{1},\ldots,d_{n})$ is a
nonincreasing graphic sequence with $d_{1}\geq n-3$. If $\pi$ is one of
$(n-3,3^{n-1})$, $(k,3^{k})$ and $(k^{2},3^{k-1})$, then by Lemmas 2.7 and
3.2, $\pi$ has no $Z_{3}$-connected realization.
Conversely, assume that
$\pi\notin\\{(n-3,3^{n-1}),(k,3^{k}),(k^{2},3^{k-1})\\}$. Since $d_{1}\geq
n-3$ and $d_{n}\geq 3$, $n\geq 6$. In the case that $n=6$, by Theorem 1.2,
$d_{1}=3,4$. If $d_{1}=3$, then $\pi=(3^{6})$. Since $n=6$,
$(3^{6})=(n-3,3^{n-1})$, contrary to our assumption. If $d_{1}=4$, then by
(ii) of Lemma 3.1 $\pi$ has a $Z_{3}$-connected realization. In the case that
$n=7$, by Theorems 1.2 and 2.1, $4\leq d_{1}\leq 5$. If $d_{1}=5$, then any
realization of $\pi$ contains the graph (b) of fig. 1. By Lemma 2.8, $\pi$ has
a $Z_{3}$-connected realization. Assume that $d_{1}=4$. Since $n=7$,
$(4,3^{6})=(n-3,3^{n-1})$. Thus, by our assumption, $\pi\not=(4,3^{6})$. In
this case, any realization of $\pi$ contains the graph (a) in Fig. 2. By Lemma
2.9, $\pi$ has a $Z_{3}$-connected realization. Thus, assume that $n\geq 8$.
By Theorem 1.2 and by Lemmas 3.2 and 4.1, we are left to prove that if
$d_{1}=n-3,d_{n}\geq 3$ and $d_{2}\neq 3$, then $\pi$ has a $Z_{3}$-connected
realization. Suppose otherwise that $\pi=(d_{1},\ldots,d_{n})$ satisfying
$d_{1}=n-3$, $d_{2}\neq 3$, $d_{n}\geq 3$. (1)
Subject to (1),
$\pi$ has no $Z_{3}$-realization with $n$ minimized. (2)
We establish the following claim first.
Claim 1. (i) $d_{n-3}=d_{n-2}=d_{n-1}=d_{n}=3$.
(ii) $3\leq d_{3}\leq 4$.
_Proof of Claim 1._ By Theorem 1.3, $d_{n-3}\leq 3$. (i) follows since
$d_{n}\geq 3$.
(ii) Suppose otherwise that subject to (1) and (2), $\pi$ satisfies $d_{3}\geq
5$. Since $d_{2}\geq d_{3}$, $d_{2}\geq 5$. Define
$\bar{\pi}=(n-4,d_{2}-1,d_{3}-1,d_{4},\ldots,d_{n-1})$=$(\bar{d_{1}},\ldots,\bar{d}_{n-1})$
with $\bar{d_{1}}\geq\ldots\geq\bar{d}_{n-1}$. Since $d_{3}\geq 5$,
$d_{2}-1\geq d_{3}-1\geq 4$. This means that $\bar{d_{1}}\geq\bar{d_{2}}\geq
4$, and $\bar{d}_{n-1}\geq 3$. If $d_{1}>d_{4}$, then $\bar{d_{1}}=(n-1)-3$.
In this case, by the minimality of $n$, $\bar{\pi}$ has a $Z_{3}$-connected
realization $\bar{G}$. If $d_{1}=d_{4}$, then $\bar{d_{1}}=d_{4}$. It follows
that $d_{4}>n-4$. This implies that $d_{1}=d_{2}=d_{3}=d_{4}=n-3$. Thus,
$\bar{d_{1}}=d_{4}=n-3=(n-1)-2$. By Lemma 4.1, $\bar{\pi}$ has a
$Z_{3}$-connected realization $\bar{G}$. In either case, $\pi$ has a
realization $G$ obtained from $\bar{G}$ by adding a new vertex $v$ and three
edges joining $v$ to the corresponding vertices of $\bar{G}$. By part (8) of
Lemma 2.2, $G$ is $Z_{3}$-connected, a contradiction. Thus $d_{3}\leq 4$.
Since $d_{3}\geq 3$, $3\leq d_{3}\leq 4$. This proves Claim 1.
By Claim 1, we may assume that $\pi=(n-3,d_{2},d_{3},\ldots,d_{n-4},3^{4})$
with $d_{3}\in\\{3,4\\}$. We consider the following two cases.
Case 1. $d_{3}=3$.
In this case, $\pi=(n-3,d_{2},3^{n-2})$. Since $\pi$ is graphic, $d_{2}$ is
odd. Since $d_{2}\not=3$, $d_{2}\geq 5$. We first assume that $d_{2}=5$. In
this case, $\pi=(n-3,5,3^{n-2})$. If $n=8$, by Lemma 2.8, the graph (d) in
Fig. 1 is a $Z_{3}$-connected realization of $\pi=(5^{2},3^{6})$. Thus, assume
that $n\geq 9$.
Assume that $n$ is odd. Denote by $W_{n-5}$ an even wheel with the center at
$v_{1}$ and by $S$ a vertex set with $|S|=2$. We construct graph $G$ from
$W_{n-5}$ and $S$ as follows: First, connect $v_{1}$ to each vertex of $S$.
Second, choose one vertex $v_{2}$ in $W_{n-5}$ and add two vertices
$x_{1},x_{2}$ such that $x_{i}$ is adjacent to $v_{2}$ and each vertex of $S$
for each $i\in\\{1,2\\}$.
Since $d(v_{1})=n-5+2=n-3$, $d(v_{2})=3+2=5$ and each vertex of
$V(G)\setminus\\{v_{1},v_{2}\\}$ is a 3-vertex, this means that $G$ is a
realization of degree sequence $(n-3,5,3^{n-2})$. By part (5) of Lemma 2.2,
$W_{n-5}$ is $Z_{3}$-connected. Note that $G/W_{n-5}$ is an even wheel $W_{4}$
which is also $Z_{3}$-connected by Lemma 2.2. It follows by part (6) of Lemma
2.2 that $G$ is $Z_{3}$-connected, a contradiction.
Thus we may assume that $n$ is even. Denote by $W_{n-6}$ an even wheel with
the center at $v_{1}$ and let $S=\\{s_{1},s_{2},s_{3}\\}$ be a vertex set. We
construct graph $G$ from $W_{n-6}$ and $S$ as follows: First, connect $v_{1}$
to each vertex of $S$. Second, choose one vertex $v_{2}$ in $W_{n-6}$ and and
let $v_{2}$ be adjacent to $s_{1}$. Finally, add two vertices $x_{1},x_{2}$
such that $x_{1}$ is adjacent to $v_{2}$ and $s_{2}$, $s_{3}$; $x_{2}$ is
adjacent to each vertex of $S$.
It is easy to verify that $d(v_{1})=n-6+3=n-3$, $d(v_{2})=3+2=5$ and each
vertex of $V(G)\setminus\\{v_{1},v_{2}\\}$ is a 3-vertex. This means that $G$
is a realization of degree sequence $(n-3,5,3^{n-2})$. By (5) of Lemma 2.2,
$W_{n-6}$ is $Z_{3}$-connected. By part (8) of Lemma 2.2,
$W_{n-6}\cup\\{s_{1}\\}$ is $Z_{3}$-connected. Note that
$G/\\{W_{n-6}\cup\\{s_{1}\\}\\}$ is an even wheel $W_{4}$ which is
$Z_{3}$-connected. It follows by (6) of Lemma 2.2 that $G$ is
$Z_{3}$-connected, a contradiction.
From now on, we assume that $d_{2}\geq 7$. In this case, $n\geq d_{2}+3\geq
10$. Consider the case that $n$ is even. Denote by $W_{n-d_{2}+1}$ an even
wheel with the center at $v_{1}$ and by $S$ a vertex set with $|S|=d_{2}-4$.
We construct graph $G$ from $W_{n-d_{2}+1}$ and $S$ as follows: First, connect
$v_{1}$ to each vertex of $S$. Second, pick one vertex $s$ of $S$ and let
$S_{1}=S\setminus\\{s\\}$, pick one vertex $v_{2}$ in $W_{n-d_{2}+1}$ and
connect $v_{2}$ to each vertex of $S_{1}$. Third, pick two vertices
$s_{1},s_{2}$ of $S_{1}$, and add $(d_{2}-7)/2$ edges such that the induced
subgraph by $S_{1}\setminus\\{s_{1},s_{2}\\}$ is a perfect matching. Finally,
we add two vertices $x_{1}$ and $x_{2}$ such that $x_{i}$ is adjacent to
$v_{2}$, $x_{i}$ is adjacent to $s_{i}$ for $i=1,2$ and $s$ is adjacent to
each of $x_{1}$ and $x_{2}$.
Since $d(v_{1})=n-d_{2}+1+d_{2}-4=n-3$, $d(v_{2})=3+d_{2}-5+2=d_{2}$ and each
vertex of $V(G)-\\{v_{1},v_{2}\\}$ is a 3-vertex, this implies s that $G$ is a
realization of degree sequence $(n-3,d_{2},3^{n-2})$. By (5) of Lemma 2.2,
$W_{n-d_{2}+1}$ is $Z_{3}$-connected. Contracting this even wheel
$W_{n-d_{2}+1}$ and contracting all 2-cycles generated in the process, we get
$K_{1}$. By (8) of Lemma 2.2, $G$ is $Z_{3}$-connected, a contradiction.
Consider the case that $n$ is odd. Denote by $W_{n-d_{2}}$ an even wheel with
the center at $v_{1}$ and by $S$ a vertex set with $|S|=d_{2}-3$. We construct
a graph from $W_{n-d_{2}}$ and $S$ as follows. First, let $v_{1}$ be adjacent
to each vertex of $S$. Second, pick two vertices $s_{3}$ and $s_{4}$ of $S$,
define $S_{1}=S\setminus\\{s_{3},s_{4}\\}$ and pick one vertex $v_{2}$ in $W$
so that $v_{2}$ is adjacent to each vertex of $S_{1}$. Third, add two new
vertex $x_{1},x_{2}$ such that $x_{i}$ is adjacent to each of $v_{2}$, $s_{3}$
and $s_{4}$. Finally, add $(d_{2}-5)/2$ edges in $S_{1}$ so that the subgraph
induced by vertices of $S_{1}\setminus\\{s_{3},s_{4}\\}$ is a perfect
matching.
Since $d(v_{1})=n-d_{2}+d_{2}-3=n-3$, $d(v_{2})=3+d_{2}-5+2=d_{2}$, and each
vertex of $V(G)\setminus\\{v_{1},v_{2}\\}$ is a 3-vertex, $G$ is a realization
of degree sequence $(n-3,d_{2},3^{n-2})$. Similarly, by parts (5) and (8) of
Lemma 2.2, $G$ is $Z_{3}$-connected, a contradiction.
Case 2. $d_{3}=4$.
In this case, $d_{2}\geq 4$ and
$\pi=(n-3,d_{2},4,d_{4},\ldots,d_{n-4},3^{4})$. Define
$\bar{\pi}=(n-4,d_{2}-1,3,d_{4},\ldots,d_{n-4},\\\
3^{3})=(\bar{d}_{1},\ldots,\bar{d}_{n-1})$ with
$\bar{d}_{1}\geq\ldots\geq\bar{d}_{n-1}$. If $d_{1}=d_{4}$, then $n-3=4$ and
hence $n=7$, contrary to assumption that $n\geq 8$. Thus, $d_{1}>d_{4}$. In
this case, $\bar{d}_{1}=n-4$.
Claim 2. $d_{2}=4$.
Proof of Claim 2. Suppose otherwise that $d_{2}\geq 5$. Then $\bar{d_{2}}\geq
4$ and $\bar{\pi}$ satisfies (1). By the minimality of $n$, $\bar{\pi}$ has a
$Z_{3}$-connected realization $\bar{G}$. Thus, we conclude that $G$ is a
$Z_{3}$-connected realization of $\pi$ obtained from $\bar{G}$ by adding a new
vertex $v$ and three edges joining $v$ to the corresponding vertices of
$\bar{G}$. This contradiction proves Claim 2.
By Claim 2, $d_{2}=4$. Assume that $i\in\\{3,\ldots,n-4\\}$ such that
$d_{i}=4$ and $d_{i+1}=3$. Thus $\pi=(n-3,4^{i-1},3^{n-i})$.
Claim 3. $i=3$.
Proof of Claim 3. If $i$ is even, then $n-i$ is odd (even) when $n$ is odd
(even). No matter whether $n$ is odd or even, there are odd vertices of odd
degree, a contradiction. Thus, $i$ is odd. If $i\geq 5$, then
$\bar{\pi}=(n-4,4^{i-3},3^{n-i+1})$ satisfies (1). Recall that $n\geq 8$, by
the minimality of $n$, $\bar{\pi}$ has a $Z_{3}$-connected realization
$\bar{G}$. In this case, we can obtain a realization $G$ of $\pi$ from
$\bar{G}$ by adding a new vertex $v$ and three edges joining $v$ to the
corresponding vertices of $\bar{G}$. By (8) of Lemma 2.2, $G$ is
$Z_{3}$-connected, a contradiction. This proves Claim 3.
By Claim 3, $i=3$. This leads to that $\pi=(n-3,4^{2},3^{n-3})$. Recall that
$n\geq 8$. If $n=8$, then by Lemma 2.9, $\pi$ has a $Z_{3}$-connected
realization. Thus, we may assume that $n\geq 9$.
In the case that $n$ is odd, denote by $W_{n-5}$ the even wheel with the
center at $v_{1}$ and by $S$ a vertex set with $|S|=2$. We construct a graph
$G$ from $W_{n-5}$ and $S$ as follows. First, let $v_{1}$ be adjacent to each
vertex of $S$. Second, pick two vertices $v_{2},v_{3}$ in $W_{n-5}$ and add
two vertices $x_{1},x_{2}$ such that $x_{i}$ is adjacent to $v_{i+1}$ and each
vertex of $S$ for each $i\in\\{1,2\\}$.
It is easy to verify that $d(v_{1})=n-5+2=n-3$, $d(v_{i})=3+1=4$ for each
$i\in\\{2,3\\}$, and each vertex of $V(G)-\\{v_{1},v_{2},v_{3}\\}$ is a
3-vertex. Obviously, $G$ has a degree sequence $(n-3,4^{2},3^{n-3})$. By (5)
of Lemma 2.2, $W_{n-5}$ is $Z_{3}$-connected. $G/W_{n-5}$ is an even wheel
$W_{4}$ which is $Z_{3}$-connected. By (6) of Lemma 2.2, $G$ is
$Z_{3}$-connected, a contradiction.
In the case that $n$ is even, denote by $W_{n-6}$ the even wheel with the
center at $v_{1}$ and let $S=\\{s_{1},s_{2},s_{3}\\}$. We construct a graph
$G$ from $W_{n-6}$ and $S$ as follows. First, let $v_{1}$ be adjacent to each
vertex of $S$. Second, pick two vertices $v_{2},v_{3}$ in $W_{n-6}$ so that
$v_{2}$ is adjacent to $s_{1}$. Finally, we add two vertices $x_{1},x_{2}$
such that $x_{1}$ is adjacent to $v_{3}$ and $s_{2}$, $s_{3}$ and such that
$x_{2}$ is adjacent to each vertex of $S$.
It is easy to verify that $d(v_{1})=n-6+3=n-3$, $d(v_{i})=3+1=4$ for each
$i\in\\{2,3\\}$, and each vertex of $V(G)-\\{v_{1},v_{2},v_{3}\\}$ is a
3-vertex. Obviously, $G$ has a degree sequence $(n-3,4^{2},3^{n-3})$. By (5)
Lemma 2.2, $W_{n-6}$ is $Z_{3}$-connected. By (8) of Lemma 2.2,
$W_{n-6}\cup\\{s_{1}\\}$ is $Z_{3}$-connected.
$G/\\{W_{n-6}\cup\\{s_{1}\\}\\}$ is an even wheel $W_{4}$ which is
$Z_{3}$-connected. By (6) of Lemma 2.2, $G$ is $Z_{3}$-connected a
contradiction. $\blacksquare$
## 5 Proof of Theorem 1.5
We first establish the following lemma which is used in the proof of Theorem
1.5.
###### Lemma 5.1
Let $\pi=(d_{1},\ldots,d_{n})$ be a nonincreasing graphic sequence. If
$d_{n}\geq 3$ and $d_{n-4}\geq 4$, then either $\pi$ has a $Z_{3}$-connected
realization or $\pi=(5^{2},3^{4})$.
Proof. Since $d_{n-4}\geq 4$, $n\geq 5$. If $n=5$, then by Theorem 1.3,
$\pi=(4,3^{4})$. In this case, an even wheel $W_{4}$ is a $Z_{3}$-connected
realization of $\pi$. If $n=6$, then by Theorem 1.3, $\pi=(4^{2},3^{4})$ or
$(5^{2},3^{4})$. If $\pi=(4^{2},3^{4})$, then by Lemma 2.8, the graph (a)
shown in Fig.1 is a $Z_{3}$-connected realization of $\pi$. If $n=7$, then by
Theorem 1.3, $\pi=(5^{2},4,3^{4})$. Let $G$ be the graph (b) shown in Fig. 1
which has degree sequence $\pi=(5,4,3^{5})$. Denote by $G^{\prime}$ the graph
obtained from $G$ by adding an edge joining a vertex of degree 3 to a vertex
of degree 4. By Lemma 2.8, $G$ is a $Z_{3}$-connected realization of
$(5,4,3^{4})$ and so $G^{\prime}$ is a $Z_{3}$-connected realization of
$(5^{2},4,3^{4})$. Thus, assume that $n\geq 8$.
By Theorems 1.3 and 1.4, it is sufficient to prove that if $d_{n-3}=3$,
$d_{n-4}\geq 4$ and $d_{1}\leq n-4$, then $\pi$ has a $Z_{3}$-connected
realization. In this case, $\pi=(d_{1},\ldots,d_{n-4},3^{4})$. Suppose, to the
contrary, that $\pi=(d_{1},\ldots,d_{n})$ satisfies
$d_{n-3}=3$, $d_{n-4}\geq 4$ and $d_{1}\leq n-4$. (3)
Subject to (3),
$\pi$ has no $Z_{3}$-connected realization with $n$ minimized. (4)
Assume that $d_{n-4}\geq 5$. Define
$\bar{\pi}=(d_{1}-1,d_{2}-1,d_{3}-1,d_{4},\ldots,d_{n-4},3^{3})=(\bar{d_{1}},\ldots,\bar{d}_{n-1})$.
Since $d_{1},d_{2},d_{3}\geq 5$ and $n\geq 8$, $\bar{d}_{n-5}\geq 4$. This
implies that $\bar{\pi}$ satisfies (3), by the minimality of $n$, $\bar{\pi}$
has a $Z_{3}$-connected realization $\bar{G}$. Thus, we construct a
realization $G$ of $\pi$ from $\bar{G}$ by adding a new vertex $v$ and three
edges joining $v$ to the corresponding vertices of $\bar{G}$. It follows by
(8) of Lemma 2.2 that $G$ is $Z_{3}$-connected, a contradiction. Thus, we may
assume that $d_{n-4}=4$.
On the other hand, if $d_{1}=4$, then $\pi=(4^{n-4},3^{4})$. By Lemma 3.1,
$\pi$ has a $Z_{3}$-connected realization, a contradiction. Thus, assume
$d_{1}\geq 5$. Since $d_{n-4}=4$ and $n\geq 8$, $d_{2}\geq 5$ or $d_{2}=4$.
In the former case, $\bar{d}_{n-5}\geq 4$. In this case, $\bar{\pi}$ satisfies
(3), by the minimality of $n$, $\bar{\pi}$ has a $Z_{3}$-connected realization
$\bar{G}$. Thus, we can construct a realization $G$ of $\pi$ from $\bar{G}$ by
adding a new vertex $v$ and three edges joining $v$ to the corresponding
vertices of $\bar{G}$. By (8) of Lemma 2.2, $G$ is $Z_{3}$-connected, a
contradiction.
In the latter case, $\pi=(d_{1},4^{n-5},3^{4})$. Since $\pi$ is graphic,
$d_{1}$ is even. Since $d_{1}\leq n-4$, $n-d_{1}-1\geq 3$. In the case that
$n-d_{1}-1=3$, we have $d_{1}=n-4$ and $n\geq 10$ is even. Denote by $W_{n-4}$
an even wheel with the center at $v_{1}$. We construct a graph $G$ from
$W_{n-4}$ as follows. First, choose five vertices
$v_{2},v_{3},v_{4},v_{5},v_{6}$ of $W_{n-4}$. Second, add three vertices
$x_{1},x_{2},x_{3}$ and edges $x_{1}x_{2}$, $x_{2}x_{3}$. Finally, add edges
$v_{2}x_{1},v_{3}x_{1},v_{4}x_{2},v_{5}x_{3},v_{6}x_{3}$. In this case, for
each $i\in\\{1,2,3\\}$, $x_{i}$ is a 3-vertex, and for each
$j\in\\{2,\ldots,6\\}$ $v_{j}$ is a 4-vertex.
It is easy to see that $G$ is a $Z_{3}$-connected realization of degree
sequence $(n-4,4^{5},3^{n-6})$. Define
$S=V(W_{n-4})\setminus\\{v_{2},\ldots,v_{6}\\}=\\{v_{1},v_{7},\ldots,v_{n-3}\\}$.
Then $|S|=n-4-5=n-9\geq 1$. If $n=10$, then $G$ is a realization of
$(6,4^{5},3^{4})$. If $n\geq 12$, then define $G^{\prime}$ from $G$ by adding
$v_{j}v_{n-j+4}$ for $7\leq j\leq\frac{n}{2}+1$, that is, adding $(n-10)/2$
edges in $S$. Obviously, $G^{\prime}$ has a degree sequence
$(n-4,4^{n-5},3^{4})$. We conclude that $G^{\prime}$ is a $Z_{3}$-connected
realization of $\pi$.
In the case that $n-d_{1}-1=4$, we have $d_{1}=n-5$ and $n\geq 11$ is odd.
Denote by $W_{n-5}$ an even wheel with the center at $v_{1}$. We construct a
graph $G$ from $W_{n-5}$ as follows. First, choose six vertices
$v_{2},v_{3},v_{4},v_{5},v_{6},v_{7}$ of $W_{n-5}$. Second, add four vertices
$x_{1},x_{2},x_{3},x_{4}$ and edges $x_{1}x_{2}$, $x_{2}x_{3}$, $x_{3}x_{4}$.
Third, add edges
$x_{1}v_{2},x_{1}v_{3},x_{2}v_{4},x_{3}v_{5},x_{4}v_{6},x_{4}v_{7}$. In this
case, for each $i\in\\{1,2,3,4\\}$, $x_{i}$ is a 3-vertex, and for each
$j\in\\{2,3,\ldots,7\\}$ $v_{j}$ is a 4-vertex.
It is easy to see that $G$ is a $Z_{3}$-connected realization of degree
sequence $(n-5,4^{6},3^{n-7})$. Define
$S=V(W_{n-5})\setminus\\{v_{1},v_{2},\ldots,v_{7}\\}=\\{v_{8},\ldots,v_{n-4}\\}$.
Then $|S|=n-5-6=n-11\geq 0$. If $n=11$, then $G$ is a realization of
$(6,4^{6},3^{4})$. If $n\geq 13$, then define $G^{\prime}$ from $G$ by adding
edges $v_{j}v_{n-j+4}$ for $8\leq j\leq\frac{n+3}{2}$, that is, adding
$(n-11)/2$ edges in $S$. Obviously, $G^{\prime}$ is a $Z_{3}$-connected
realization of degree sequence $(n-5,4^{n-5},3^{4})$, a contradiction.
In the case that $n-d_{1}-1\geq 5$, denote by $W_{d_{1}}$ an even wheel with
the center at $v_{0}$. Let $V(W_{d_{1}})=\\{v_{0},v_{1},\ldots,v_{d_{1}}\\}$.
Let $C:u_{1}\ldots u_{n-d_{1}-1}u_{1}$ be a cycle of length $n-d_{1}-1$ and
and define a graph $H$ obtained from $C$ adding edges $u_{i}u_{i+2}$ for each
$i\in\\{1,\ldots,n-d_{1}-1\\}$, where the subscripts are taken modular
$n-d_{1}$. Clearly, $H$ is a 4-regular and is triangularly connected. Define
$H^{\prime}=H-\\{u_{2}u_{n-d_{1}-1}\\}$. Now we prove $H^{\prime}$ is
$Z_{3}$-connected. Clearly, $H^{\prime}_{[u_{1}u_{2},u_{1}u_{3}]}$ is
triangularly connected and contains a 2-circuit $u_{2}u_{3}u_{2}$. By Lemma
2.4 (a) and Lemma 2.2 (3), $H^{\prime}_{[u_{1}u_{2},u_{1}u_{3}]}$ is
$Z_{3}$-connected, and hence $H^{\prime}$ by Lemma 2.3. We construct a graph
$G$ from $W_{d_{1}}$ and $H^{\prime}$ as follows. If $d_{1}\geq 8$, then we
add two edges $v_{1}u_{n-d_{1}-1}$, $v_{2}u_{2}$ and add edges
$v_{j}v_{d_{1}-j+3}$ for $3\leq j\leq\frac{d_{1}}{2}-1$, that is, add
$(d_{1}-6)/2$ edges between vertices
$\\{v_{3},\ldots,v_{d_{1}-4}\\}\setminus\\{v_{\frac{d_{1}}{2}},v_{\frac{d_{1}}{2}+1},v_{\frac{d_{1}}{2}+2},v_{\frac{d_{1}}{2}+3}\\}$
such that $d(v_{i})=4$ for each vertex of
$\\{v_{3},\ldots,v_{d_{1}-4}\\}\setminus\\{v_{\frac{d_{1}}{2}},v_{\frac{d_{1}}{2}+1},v_{\frac{d_{1}}{2}+2},v_{\frac{d_{1}}{2}+3}\\}$
and the new graph is simple. If $d_{1}=6$, then we add two edges
$v_{1}u_{n-d_{1}-1}$, $v_{2}u_{2}$. In either case, $G$ is a $Z_{3}$-connected
realization of has a degree sequence $(d_{1},4^{n-5},3^{4})$, a contradiction
Proof of Theorem 1.5. If $\pi=(5^{2},3^{4})$ or $(5,3^{5})$, then by Lemma
2.7, $\pi$ has no $Z_{3}$-connected realization. Thus, assume that
$\pi\not=(5^{2},3^{4}),(5,3^{5})$.
Since $d_{n-5}\geq 4$, $n\geq 6$. If $n=6$, then by Theorem 1.3,
$\pi=(d_{1},d_{2},3^{4})$. By Lemma 5.1, $\pi=(d_{1},3^{5})$. By our
assumption that $d_{n-5}\geq 4$, $\pi=(5,3^{5})$, a contradiction. If $n=7$,
then by Theorem 1.3 and Lemma 5.1, $\pi=(d_{1},d_{2},3^{5})$. By our
assumption that $d_{n-5}\geq 4$, $\pi=(5,4,3^{5})$ or $(6,5,3^{5})$. In the
former, the graph (b) in Fig. 1 is a $Z_{3}$-connected realization of $\pi$.
In the latter case, Theorem 1.4 shows that $\pi$ has a $Z_{3}$-connected
realization. Thus, we may assume that $n\geq 8$.
By Theorems 1.3 and 1.4, and Lemma 5.1, it is sufficient to prove that if
$d_{n-4}=3$, $d_{n-5}\geq 4$ and $d_{1}\leq n-4$, then $\pi$ has a
$Z_{3}$-connected realization. Then $\pi=(d_{1},\ldots,d_{n-5},3^{5})$.
Suppose to the contrary that $\pi$ satisfies
$d_{n-4}=3$, $d_{n-5}\geq 4$ and $d_{1}\leq n-4$. (5)
Subject to (5),
$\pi$ has no $Z_{3}$-connected realization with $n$ minimized. (6)
We claim that $d_{3}=4$. Suppose otherwise that $d_{3}\geq 5$. Define
$\bar{\pi}=(d_{1}-1,d_{2}-1,d_{3}-1,d_{4},\ldots,d_{n-5},3^{4})=(\bar{d}_{1},\ldots,\bar{d}_{n-1})$.
Since $d_{1},d_{2},d_{3}\geq 5$, $\bar{d}_{n-6}\geq 4$. Thus $\bar{\pi}$
satisfies (5). By the minimality of $n$, $\bar{\pi}$ has a $Z_{3}$-connected
realization $\bar{G}$. Denote by $G$ the graph obtained from $\bar{G}$ by
adding a new vertex $v$ and three edges joining $v$ to the corresponding
vertices. It follows by (8) of Lemma 2.2 that $G$ is a $Z_{3}$-connected
realization of $\pi$, a contradiction. Thus, $d_{3}=4$ and
$\pi=(d_{1},d_{2},4^{n-7},3^{5})$.
We claim that $d_{2}=4$. Suppose otherwise that $d_{2}\geq 5$. In this case,
$\bar{\pi}=(d_{1}-1,d_{2}-1,4^{n-8},3^{6})=(\bar{d}_{1},\ldots,\bar{d}_{n-1})$.
Since $d_{2}\geq 5$ and $d_{3}=4$, $n\geq 3+5=8$. This implies that
$\bar{d}_{n-6}\geq 4$. Thus, $\bar{\pi}$ satisfies (5). By the minimality of
$n$, $\bar{\pi}$ has a $Z_{3}$-connected realization $\bar{G}$. Denote by $G$
the graph obtained from $\bar{G}$ by adding a new vertex $v$ and three edges
joining $v$ to the corresponding vertices of $\bar{G}$. It follows from (8) of
Lemma 2.2 that $G$ is a $Z_{3}$-connected realization of $\pi$, a
contradiction. Thus, $d_{2}=4$ and $\pi=(d_{1},4^{n-6},3^{5})$.
Since $\pi$ is graphic, $d_{1}$ is odd and $d_{1}\geq 5$. If $d_{1}=5$, then
by (iii) of Lemma 3.1, $\pi$ has a $Z_{3}$-connected realization.
We are left to the case that $d_{1}\geq 7$. Since $d_{1}\leq n-4$, $n\geq
d_{1}+4\geq 11$. By (iii) of Lemma 3.1, Let $G^{\prime}$ be a
$Z_{3}$-connected realization of degree sequence $(5,4^{n-6},3^{5})$. By the
construction of $G^{\prime}$ in (iii) of Lemma 3.1, $G^{\prime}$ has at least
$|E(G^{\prime})|-5-14=2n-21\geq 2(d_{1}+4)-21=2d_{1}-13\geq(d_{1}-5)/2$ edges
not incident with the any vertex of $N_{G^{\prime}}(u)\cup\\{u\\}$, where $u$
is a 5-vertex in $G^{\prime}$ since $G^{\prime}$ contains a pair of adjacent
neighbors of $u$. Choose $(d_{1}-5)/2$ such edges, say $u_{i}v_{i}$ for each
$i\in\\{1,\ldots,(d_{1}-5)/2)\\}$. Denote by the graph $G$ from $G^{\prime}$
by deleting edges $u_{i}v_{i}$ and adding edge $uu_{i}$, $uv_{i}$ for each
$i\in\\{1,\ldots,(d_{1}-5)/2\\}$. It follows by Lemma 2.3 that $G$ is a
$Z_{3}$-connected realization of degree sequence $(d_{1},4^{n-6},3^{5})$, a
contradiction. We complete our proof.
## References
* [1] D. Archdeacon, Realizing degree sequences with graphs having 3-flows, http://www.cems. uvm.edu/ archdeac/problems/seqflow. html.
* [2] J. A. Bondy, U. S. R. Murty, Graph Theory with Applications, North-Holland, New York, 1976.
* [3] P. J. Cameron, Problems from the 16th British Combinatorial Conference, Discrete Math., 197/198 (1999) 799-812.
* [4] J. Chen, E. Eschen, H. -J. Lai, Group connectivity of certain graphs, Ars Combin., 89 (2008) 141–158.
* [5] M. DeVos, R. Xu, G. Yu, Nowhere-zero $Z_{3}$-flows through $Z_{3}$-connectivity, Discrete Math., 306 (2006) 26–30.
* [6] G. Fan, H. -J. Lai, R. Xu, C. -Q. Zhang, C. Zhou, Nowhere-zero 3-flows in triangularly connected graphs, J. Combin. Theory, Ser. B 98 (2008) 1325–1336.
* [7] S. Fan, H. -J. Lai, Y. Shao, T. Zhang, J. Zhou, Degree sequence and supereulerian graphs, Discrete Math., 308 (2008) 6626–6631.
* [8] S. L. Hakimi, On realizability of a set of integers as degrees of the vertices of a linear graph. I, SIAM J. Appl. Math., 10 (1962) 496–506.
* [9] S. L. Hakimi, On realizability of a set of integers as degrees of the vertices of a linear graph. H. Uniqueness, SIAM J. Appl. Math., 11 (1963) 135–147.
* [10] F. Jaeger, N. Linial, C. Payan, N. Tarsi, Group connectivity of graphs–a nonhomogeneous analogue of nowhere zero flow properties, J. Combin. Theory, Ser. B 56 (1992) 165–182.
* [11] D. J. Kleitman and D. L. Wang, Algorithm for constructing graphs and digraphs with given valences and factors, Discrete Math., 6 (1973) 79–88.
* [12] H. -J. Lai, Group connectivity of 3-edge-connected chordal graphs, Graphs Combin., 16 (2000) 165–176.
* [13] H.-J. Lai, X. Li, Y. Shao, M. Zhan, Group connectivity and group colorings of graphs-a survey, Acta Math. Sinica, 27 (2011) 405–434.
* [14] R. Luo, R. Xu, J. Yin, G. Yu, Ore-condition and $Z_{3}$-connectivity, European J. of Combin., 29 (2008) 1587–1595.
* [15] R. Luo, W. Zang, and C. -Q Zhang, Nowhere-zero 4-flows, simultaneous edge-colorings, and critical partial latin squares, Combinatorica, 24 (2004) 641–657.
* [16] R. Luo, R. Xu, W. Zang, and C. -Q Zhang, Realizing degree sequences with graphs having nowhere-zero 3-flows, SIAM J. Discrete Math., 22 (2008) 500–519.
* [17] R. Luo, R. Xu, X. Yu, An extremal problem on group connectivity of graphs, European J. Combin., 33 (2012) 1078–1085.
* [18] R. Steinberg and D. H. Younger, Grötzch’s theorem for the projectice plane, Ars Combin. 28 (1989) 15–31.
* [19] W. T. Tutte, A contribution on the theory of chromatic polynomial, Canad. J. Math., 6 (1954), 80–91.
* [20] J. Yin, G. Guo, The smallest degree sum that yields graphic sequences with a $Z_{3}$-connected realization, European J. Combin., 34 (2013) 806–811.
|
# On Vacuum Fluctuations in Quantum Gravity and Interferometer Arm
Fluctuations
Kathryn M. Zurek Walter Burke Institute for Theoretical Physics
California Institute of Technology, Pasadena, CA USA
###### Abstract
We propose a simple model of spacetime vacuum fluctuations motivated by
AdS/CFT, where the vacuum is described by a thermal density matrix,
$\rho=\frac{e^{-K}}{\mbox{Tr}(e^{-K})}$ with $K$ the modular Hamiltonian. In
AdS/CFT, both the expectation value of $K$ and its fluctuations $\langle\Delta
K^{2}\rangle$ have been calculated; both obey an area law identical to the
Bekenstein-Hawking area law of black hole mechanics: $\langle
K\rangle=\langle\Delta K^{2}\rangle=\frac{A}{4G_{N}}$, where $A$ is the area
of an (extremal) entangling surface. It has also been shown that $\Delta K$
gravitates in AdS, and hence generates metric fluctuations. These theoretical
results are intriguing, but it is not known how to precisely extend such ideas
about holographic quantum gravity to ordinary flat space. We take the approach
of considering whether experimental signatures in metric fluctuations could
determine properties of the vacuum of quantum gravity in flat space. In
particular, we propose a theoretical model motived by the AdS/CFT calculations
that reproduces the most important features of modular Hamiltonian
fluctuations; the model consists of a high occupation number bosonic degree of
freedom. We show that if this theory couples through ordinary gravitational
couplings to the mirrors in an interferometer with strain sensitivity similar
to what will be available for gravitational waves, vacuum fluctuations could
be observable.
Overview. One of the most important open problems in fundamental physics is
bridging the theoretical and observational chasm between quantum mechanics and
gravity, to understand quantum gravity. The divide can be understood at a
basic level by considering the length and time scales of gravity, which are
derived from the Newton constant, $l_{p}=\sqrt{8\pi G\hbar/c^{3}}\simeq
10^{-34}\mbox{ m},~{}t_{p}=l_{p}/c\simeq 10^{-43}\mbox{ s}$.111For the
remainer of this letter we will take $\hbar=c=1$. From the point of view of
quantum mechanics and Effective Field Theory (EFT), these numbers represent
the time and length scales of quantum fluctuations. They are far out of reach
of observational capabilities, and neatly encapsulate why the quantum effects
of gravity have never been probed.
On the other hand, theoretical progress, mostly in the context of AdS/CFT and
the black hole information paradox, suggests that non-locality and
entanglement plays an important role in the quantum theory of gravity (see
Ref. Van Raamsdonk (2016) for a pedagogical review and references therein).
The simplest way to see why quantum gravity may be highly non-local, and hence
why the naïve expectations of EFT may break down, is through holography.
Quantum Field Theory (QFT) dictates that the number of degrees of freedom in a
spacetime volume scales with the volume (in units of $l_{p}$). Holography, on
the other hand, says that the degrees-of-freedom of a spacetime volume scales
with the area of the surface bounding that volume. As a result, this implies
that QFT grossly over-counts the number of degrees-of-freedom when gravity is
involved, suggesting that in any EFT description of spacetime, there should be
long range correlations between the degrees-of-freedom. It is such long range
correlations that we seek to model in this letter, where we call each
spacetime degree-of-freedom a pixel.
Our ability to quantify spacetime fluctuations from pixels relies on
theoretical progress made recently in connecting the holographic notion of
spacetime to entanglement. Not surprisingly, the connection between geometry,
quantum information (QI) and holography has been formulated most clearly and
rigorously in the context of AdS/CFT Ryu and Takayanagi (2006a, b), though
there are reasons to think that the connection appears in generic spacetimes
Srednicki (1993); Callan and Wilczek (1994); Jacobson (1995, 2016). In the
context of AdS/CFT, the geometry-QI-holography connection is encapsulated most
prominently in the relation
$\displaystyle\beta\langle K\rangle=S_{\rm ent}=\frac{A(\Sigma)}{4G}.$ (1)
Let us discuss each part of this equality. $K$ is a measure of the total
energy in a spacetime volume called the modular Hamiltonian of the boundary
CFT ($K\equiv\int T_{\mu\nu}\xi^{\mu}dV^{\nu}$ with $T_{\mu\nu}$ the CFT
stress tensor, $\xi^{\mu}$ the conformal Killing vector on the boundary and
$dV^{\nu}$ an infinitesimal volume element); since energy directly sources the
metric through the Einstein equation, it represents geometry. The entanglement
entropy $S_{\rm ent}$ quantifies mixing of the vacuum state $\rho_{\beta}$
across an entangling boundary and represents the QI part of the
correspondence: $S_{\rm ent}=-\mbox{Tr}\rho_{\beta}\log\rho_{\beta}$, where
the entangling surface $\Sigma$ has area $A(\Sigma)$ and an associated inverse
temperature $\beta=2\pi R$, where $R$ is the AdS curvature.222We have
reintroduced the temperature; it can be reabsorbed in the definition of $K$ as
in the abstract, but we will find it convenient to leave it explicit. It was
remarkably shown Ryu and Takayanagi (2006b) that there is a special surface in
AdS, called the RT surface, where the entanglement entropy of the vacuum state
has precisely the form of the Bekenstein-Hawking area law for the entropy of
black hole horizons; this represents the holographic part of the connection.
Further, by calculating fluctuations of the modular Hamiltonian (and of the
entanglement entropy), one can compute fluctuations in the bulk (AdS)
geometry. The fluctuations of the modular Hamiltonian have been calculated in
the AdS bulk Verlinde and Zurek (2020), with the result
$\displaystyle\beta^{2}\langle\Delta K^{2}\rangle=\frac{A(\Sigma)}{4G}.$ (2)
With suitable identifications, this result agrees with an equivalent boundary
calculation Perlmutter (2014), and with the “capacity of entanglement”
Nakaguchi and Nishioka (2016); De Boer _et al._ (2019). This implies that,
when restricted to a finite part of the spacetime (defined by a causal diamond
with bifurcate horizon $\Sigma$), the vacuum has energy fluctuations $\Delta
K\neq 0$.
Eqs. (1), (2) together encourage an interpretation of spacetime as bits of
information, with the number ${\cal N}$ of degrees-of-freedom in a given
volume bounded by a surface of area $A$ given by the entanglement entropy
$\displaystyle S_{\rm ent}={\cal N}=\frac{A}{4G}.$ (3)
This result can be interpreted in light of the fact that the vacuum state of
any QFT, restricted to a causal diamond, is given by a thermal density matrix
Casini _et al._ (2011)
$\displaystyle\rho_{\beta}=\frac{e^{-\beta K}}{Z_{\beta}}.$ (4)
For example, in a high-temperature system with Maxwell-Boltzmann statistics,
the root-mean-square fluctuations of these ${\cal N}$ degrees-of-freedom is
given by
$\displaystyle\frac{\sqrt{\langle\Delta K^{2}\rangle}}{\langle
K\rangle}=\frac{1}{\sqrt{\cal N}},$ (5)
in agreement with Eqs. (1), (2). It was shown in Ref. Verlinde and Zurek
(2020) that the fluctuations in the modular Hamiltonian, $\Delta K$, gravitate
and hence, for certain observers, behave like a mass, sourcing metric
fluctuations.
It is not precisely known whether the vacuum state of quantum gravity in
ordinary flat space, when restricted to a causal diamond, can be described by
Eq. (4), and whether the results derived for AdS/CFT apply to the Universe we
observe. There are reasons to think that entropy, entanglement and their
connection to geometry are very generic concepts that apply to any spacetime,
including ours. Such ideas underly currents dating more than twenty years on
the entropic and holographic nature of spacetime Jacobson (1995); Verlinde
(2011); Banks (2011). For example, it was shown in Ref. Jacobson (2016) that
taking fixed volume variations of the first law of entanglement gives rise to
the Einstein Equations.
In this letter we approach quantum gravity in flat space in terms of
observational signatures derived from a model motivated by the known AdS/CFT
results. If vacuum fluctuations could be observed in an experiment consistent
with an entropic or thermal nature of the vacuum state, this would be a leap
forward in our understanding of quantum gravity. To this end, we propose a
simple model for the degrees-of-freedom in the density matrix, which we call
pixellons because the excitations are associated with each holographic pixel
of a volume bounded by an entangling surface of area $A$. More specifically,
we consider whether fluctuations in the spacetime degrees-of-freedom of the
density matrix Eq. (4) could be observable as fluctuations in the arm length
of an interferometer. The two arms of an interferometer mark out a
(spherically symmetric) volume of spacetime with the beamsplitter at the
center of the volume and the mirrors on the surface with area $A$; the
interferometer measures the geometric fluctuations in this volume.
The motivation for our pixellon Ansatz is as follows. Based on the discussion
of Eqs. (1)-(3), we interpret the spacetime volume bounded by a surface of
area $A$ as having ${\cal N}$ bits with total energy $\langle K\rangle$. The
energy per bit is then $\omega\sim\beta^{-1}\sim 1/L$. However, it was shown
in Ref. Verlinde and Zurek (2020) that $\langle K\rangle$ itself does not
gravitate; rather the fluctuations $\Delta K$ gravitate. From this point of
view, $\langle K\rangle$ should be treated as a chemical potential counting
the background degrees-of-freedom, and the energy per excitation is
$\beta\omega=\Delta K/K=1/\sqrt{\cal N}\ll 1$. We assume the low-energy
excitations will be bosonic; this is appropriate because, as we will see
below, they are associated with a gravitational potential. Because the energy
of these bosonic degrees-of-freedom is so low, $\beta\omega\ll 1$, they will
form a high-occupation-number bosonic state. We further show that such low-
energy bosonic excitations of the vacuum, when gravitationally coupled to test
masses, may give rise to observably large fluctuations of the mirror positions
in an interferometer.
The detailed outline of our proposal is as follows. In the next section we
introduce the bosonic excitations as vacuum fluctuations, and we suggest that
these bosonic fluctuations be associated with a scalar gravitational
potential. In the following section, we propose that these scalar degrees-of-
freedom have a Bose-Einstein density-of-states and a high occupation number
due to their low energy. Then, we gravitationally couple the pixellons to a
test mass whose position fluctuates due to the boson fluctuations. We utilize
the Feynman-Vernon influence functional–a path integral realization of the
fluctuation-dissipation theorem–to compute the size of those fluctuations. We
will conclude that vacuum fluctuations from a thermal density matrix can give
rise to interferometer mirror position fluctuations that appear as noise with
a peculiar angular correlation, a smoking gun signature. Note that while the
experimental system of interest for measuring metric fluctuations, an
interferometer, and the motivation derived from the discussion surrounding
Eqs. (1)-(5), is the same as Ref. Verlinde and Zurek (2020, 2019), the models
operationally share no overlap. We will nevertheless find a very similar
effect, perhaps suggesting dual languages to describe the vacuum state of
quantum gravity.
Pixellons and Vacuum Fluctuations. As outlined in the introduction,
excitations of the degrees-of-freedom, associated with the entanglement
entropy of a finite volume of space, we refer to as pixellons. We expect that
pixellons are complicated non-linear states of all the degrees-of-freedom
available in the complete theory. Our goal is to describe a consistent low-
energy theory based on the quantum information theoretic picture presented in
the introduction.
Our starting point is energy fluctuations given by Eq. (5), which we associate
with pixellons. The energy fluctuations in a volume of radius $L$ source a
potential
$\displaystyle 2\phi=\sqrt{\frac{\alpha}{\cal N}}=\frac{\bar{l}_{p}}{4\pi L},$
(6)
where $l_{p}^{2}\equiv 8\pi G_{N}$ and
$\bar{l}_{p}^{2}\equiv\frac{\alpha}{2}l_{p}^{2}$. Here we have allowed for an
${\cal O}(1)$ number $\alpha$ in the relation $\beta^{2}\langle\Delta
K^{2}\rangle=\alpha\beta\langle K\rangle$ ($\alpha=1$ in AdS/CFT) as a
generalization of Eq. (5). This gravitational potential should be interpreted
as the root-mean-square (RMS) fluctuations in the potential about the vacuum
state, for which $\langle\phi\rangle=0$. Note that this potential is in
agreement with the results of Ref. Verlinde and Zurek (2019), where $\alpha=2$
(in four dimensions) was derived by computing mass fluctuations
thermodynamically and then extracting the gravitational potential from a
topological black hole foliation of 4-d Minkowski space. Next we consider how
to represent this gravitational potential as a bosonic degree-of-freedom.
Thermal Representation of Vacuum. The effects of interest occur because of the
finite volume, and hence finite ${\cal N}$, under observation; in the limit
${\cal N}\rightarrow\infty$, fluctuations vanish. We divide our space into two
regions, $A$ (which is interrogated by an experiment), and $\tilde{A}$
(outside of the experimental apparatus). We assume that the vacuum state of
the quantized Einstein-Hilbert action, restricted to the subspace $A$, has a
density matrix given by Eq. (4), such that $\rho_{A}=\rho_{\beta}$. This
assumption is supported by Jacobson’s derivation Jacobson (2016) of the
Einstein Equation from the entanglement entropy of a diamond utilizing the
density matrix in Eq. (4).
Furthermore, we are seeking to describe the excitations associated with the
fluctuating scalar potential in Eq. (6). Accordingly, we assume that there is
an algebra of $N=\sqrt{\cal N}$ bosonic degrees-of-freedom to describe the
pixellon excitations, with thermal density matrix given by
$\displaystyle\mbox{Tr}(\rho_{\beta}a_{i}^{\dagger}a_{i})=\frac{1}{e^{\beta(\epsilon_{i}-\mu)}-1},$
(7)
where the bosonic creation and annihilation operators satisfy the usual
commutation relations
$\displaystyle[a_{i},a_{i^{\prime}}^{\dagger}]=\delta_{i,i^{\prime}},$ (8)
with $i,~{}i^{\prime}$ running from 1 to $N$. The Hilbert space of these
operators has been cut-off by the finite size of the causal diamond. If
$\epsilon_{i}-\mu$ represents the energy per degree-of-freedom of excitations
of the vacuum state, we have
$\displaystyle\beta(\epsilon_{i}-\mu)=\beta\frac{K-\langle K\rangle}{{\cal
N}}.$ (9)
Since $\beta^{2}\langle\Delta K^{2}\rangle=\alpha\cal N$, and $\beta\sim L$,
we immediately learn that the amplitude of a mass fluctuation is Planckian,
$\sqrt{\langle\Delta K^{2}\rangle}\sim 1/l_{p}$. This is a large mass
fluctuation, but it is divided over ${\cal N}$ pixels such that the typical
energy per degree-of-freedom, relative to the momentum $p\sim 1/\beta$, is
$\displaystyle\beta(\epsilon_{\rm pix}-\mu)\sim\frac{1}{\sqrt{\cal N}}\ll 1.$
(10)
To fully describe the distribution, we need the momentum dependence of
$\beta(\epsilon_{\rm pix}-\mu)$. To obtain this, we employ an Ansatz motivated
by the observation that we produce the scaling in Eq. (10) if we postulate
$\displaystyle\epsilon_{\rm pix}-\mu\equiv\frac{p^{2}}{2m_{\rm pix}},$ (11)
with $m_{\rm pix}\sim 1/l_{p}$. We thus have a pixellon density of states
given by
$\displaystyle\rho_{\rm pix}(p)=\frac{1}{e^{\beta\frac{p^{2}}{2m_{\rm
pix}}}-1}.$ (12)
Because of the extremely low energy of the bosonic degrees-of-freedom, the
occupation number of the excitations is high:
$\displaystyle\rho_{\rm pix}\approx\frac{2m_{\rm pix}}{\beta
p^{2}}\sim\sqrt{\cal N}.$ (13)
This high occupation number will amplify the effective gravitational coupling
of a massive object to the pixellons. Note that while Eq. (12) is the density
of states for the collective pixellon modes, one should not identify the
dispersion of an individual pixellon as $\omega_{p}=p^{2}/2m_{\rm pix}$; the
individual pixellon dispersion will be read off from the action Eq. (18)
below. We also note in passing that, while the context is different, our
hypothesis shares features in common with Ref. Dvali and Gomez (2013) for
black holes, a connection that might be interesting to explore in the future.
Length from Vacuum Fluctuations. For our purposes, the most important question
is whether such excitations are observable. The Ansatz of a pixellon
excitations, $\phi$, with density of states given by Eq. (12), coupled
gravitationally to the mirror system of an interferometer, is ideally suited
to the Feynman-Vernon (FV) influence functional Feynman and Vernon (1963).
Refs. Parikh _et al._ (2020a, b) have utilized the FV set-up to compute the
impact of graviton fluctuations coupled to an interferometer mirror; they find
that the mirror position fluctuations are Planckian (i.e. unobservably small)
for the vacuum state. In this section we utilize similar methods, but focus on
the pixellon vacuum fluctuations, and find the effects are large, amplified by
the high occupation number of the pixellon.
We will consider a single arm of a simple Michelson interferometer as the test
system, with the far mirror of mass $m_{M}$ at coordinate ${\bf M}$, with
equilibrium position a distance $L$ from a beam splitter at position ${\bf
B}$. Our starting point is the mode decomposition $h_{ij}(t,{\bf
x})=\int\frac{d^{3}k}{(2\pi)^{3}}\phi(t,{\bf k})e^{i{\bf k}\cdot{\bf
x}}\epsilon_{ij}^{s},$ where $\epsilon_{ij}^{s}$ is the graviton polarization
vector. We choose TT gauge where
$\epsilon_{ij}^{s}\epsilon^{ij}_{t}=2\delta^{s}_{t}$, and the metric
fluctuations are transverse to the interferometer arm. $\phi(t,{\bf k})$ will
be identified with the pixellon momentum modes. The action for the mirror and
metric is
$\displaystyle S$ $\displaystyle=$ $\displaystyle\int
d^{4}x\left[\frac{\partial_{\mu}h_{ij}\partial^{\mu}h^{ij}}{64\pi G}\right.$
(14) $\displaystyle+\left.\delta^{3}({\bf x}-{\bf
M})\frac{m_{M}(\delta_{ij}+h_{ij}(t,{\bf
x}))}{2}\dot{M}^{i}\dot{M}^{j}\right].$
The first term is identified as the usual Einstein-Hilbert action, and the
second as the mirror kinetic term. We have the freedom to fix the origin of
the coordinate system at the beam splitter, with displacements relative to the
beam splitter given by $M^{i}-B^{i}=X^{i}-\frac{h^{i}_{k}X^{k}}{2}$. We also
take the interferometer arm of interest aligned along the $x$ axis such that
$X^{x}\equiv X$. The action Eq. (14) then becomes
$\displaystyle S=$ $\displaystyle\int
dt\int\frac{d^{3}k}{(2\pi)^{3}}\frac{m_{\phi}}{2}\left(\dot{\tilde{\phi}}(t,{\bf
k})^{2}-k^{2}\tilde{\phi}(t,{\bf k})^{2}\right)$ $\displaystyle+\int
dt\frac{m_{M}}{2}\left(\dot{X}^{2}+g\int\frac{d^{3}k}{(2\pi)^{3}}\tilde{\phi}(t,{\bf
k})\frac{d}{dt}(\dot{X}X)\right),$
where $m_{\phi}=1/16\pi G^{2}$, $\tilde{\phi}=\sqrt{G}\phi$, and
$g=1/\sqrt{G}$. In this form, the action has been explicitly written as a non-
relativistic quantum mechanical system for which the FV action is ideally
suited.
Our goal is to compute the fluctuations in the mirror position $X$ driven by
the gravitational $\phi-X$ coupling, with the FV influence functional. The FV
influence functional is a path integral formulation for the transition
probability $P_{nm}$ of some test system $Q$ from a state $\psi_{n}$ at time
$t=\tau$ to $\psi_{m}$ at time $t=T$, when coupled to an interaction system
characterized by an influence functional ${\cal F}(Q,Q^{\prime})$:
$\displaystyle P_{nm}$ $\displaystyle=$
$\displaystyle\int\psi_{m}^{*}(Q_{T})\psi_{m}(Q^{\prime}_{T})e^{i(S_{0}(Q)-S_{0}(Q^{\prime}))}{\cal
F}(Q,Q^{\prime})$ (16) $\displaystyle\times$
$\displaystyle\psi_{n}^{*}(Q_{\tau})\psi_{n}(Q^{\prime}_{\tau}){\cal D}Q{\cal
D}Q^{\prime}dQ_{T}dQ^{\prime}_{T}dQ_{\tau}dQ^{\prime}_{\tau}.$
Here $S_{0}(Q)$ is the action in the absence of any coupling to an external
system, and the influences of the external system are contained in ${\cal
F}(Q,Q^{\prime})$, which is characterized by the influence phase
$\Phi(Q,Q^{\prime})$:
$\displaystyle{\cal F}(Q,Q^{\prime})=e^{i\Phi(Q,Q^{\prime})}.$ (17)
The influence phase is most simply derived for an oscillator $\tilde{\phi}$
coupled to the test system $Q$ through a linear interaction of the type (see
Eq. (4.6) of Ref. Feynman and Vernon (1963)),
$\displaystyle
S_{k}=\int_{\tau}^{T}dt\left[\frac{m_{\phi}}{2}(\dot{\tilde{\phi}}^{2}-k^{2}\tilde{\phi}^{2})+\frac{gm_{M}}{2}\tilde{\phi}Q\right]+S_{0}^{Q},$
(18)
where $S_{0}^{Q}$ is the uncoupled action of the test system, and we have left
a subscript $k$ explicit to denote that we have chosen a single $k$ mode. For
this action FV showed the influence functional is
$\displaystyle i\Phi_{k}(Q,Q^{\prime})$ $\displaystyle=$
$\displaystyle-\frac{g^{2}m_{M}^{2}}{8m_{\phi}k}\int_{\tau}^{T}dt\int_{\tau}^{t}ds(Q_{t}-Q^{\prime}_{t})$
(19)
$\displaystyle\times\left(Q_{s}e^{-ik(t-s)}-Q_{s}^{\prime}e^{ik(t-s)}\right).$
For a distribution of oscillators $\rho(k)$ of momentum $k$ acting on the test
system, we have
$\displaystyle\Phi(Q,Q^{\prime})=\int\frac{d^{3}k}{(2\pi)^{3}}\rho(k)\Phi_{k}(Q,Q^{\prime}).$
(20)
Now we decompose the influence functional into real and imaginary parts,
according to
$\displaystyle f(\Delta t)$ $\displaystyle\equiv$
$\displaystyle\frac{g^{2}m_{M}^{2}}{4m_{\phi}}\int\frac{d^{3}k}{(2\pi)^{3}}\frac{\rho(k)}{k}e^{ik\Delta
t}$ (21) $\displaystyle\equiv$ $\displaystyle A_{FV}(\Delta t)+iB_{FV}(\Delta
t).$
The real and imaginary parts of the influence phase are fixed relative to each
other by the oscillator density of states $\rho(k)$, which is the path
integral realization of the Fluctuation-Dissipation Theorem. The influence
phase can be written in terms of this function as
$\displaystyle i\Phi(Q,Q^{\prime})$ $\displaystyle=$
$\displaystyle-\frac{1}{2}\int_{\tau}^{T}dt\int_{\tau}^{t}ds(Q_{t}-Q^{\prime}_{t})$
(22) $\displaystyle\times\left(Q_{s}f^{*}(t-s)-Q_{s}^{\prime}f(t-s)\right).$
One can see that $A_{FV}(t-s)$ represents the variance in $Q(t)$, i.e. the
amplitude of the noise in the variable $Q$. This is most clearly seen by
rewriting
$\displaystyle\mbox{exp}\left[\mbox{Re}\left[i\Phi(Q,Q^{\prime})\right]\right]$
$\displaystyle=$
$\displaystyle\mbox{exp}\left[-\frac{1}{2}\int_{0}^{T}dt\int_{0}^{t}ds~{}A_{FV}(t-s)(Q_{t}-Q^{\prime}_{t})(Q_{s}-Q^{\prime}_{s})\right]$
$\displaystyle=$ $\displaystyle\int{\cal
D}F_{Q}\mbox{exp}\left[-\frac{1}{2}\int_{0}^{T}dt\int_{0}^{t}ds~{}A_{FV}^{-1}(t-s)F_{Q}(t)F_{Q}(s)\right.$
$\displaystyle\left.+i\int_{0}^{T}dtF_{Q}(t)(Q_{t}-Q^{\prime}_{t})\right],$
from which one reads
$\displaystyle\langle F_{Q}^{2}\rangle=A_{FV}(\Delta t).$ (24)
The only step that remains is to relate this noise to the observable,
fluctuations in the position of the mirror, $X(t)$. The influence phase has
allowed us to effectively integrate out the pixellon oscillators $\phi$ and
write their effect only in terms of the degree-of-freedom $F_{Q}$ which
describes the distribution of the mirror displacements according to
$Q\equiv\frac{d}{dt}(\dot{X}X)$. The noise $F_{Q}$ accordingly becomes a term
in the effective action $S_{\rm eff}$ replacing the pixellon:
$\displaystyle S_{\rm eff}=\int
dt\left[\frac{m_{M}}{2}\dot{X}^{2}+F_{Q}(t)\frac{d}{dt}(\dot{X}X)\right].$
(25)
The equation of motion is thus
$\displaystyle m_{M}\ddot{X}(t)=\ddot{F}_{Q}(t)X(t),$ (26)
where we have integrated the right-hand-side by parts as necessary to move all
time derivatives onto $F_{Q}$. This result agrees with Ref. Parikh _et al._
(2020b) in the relevant limit (up to conventions), though we have arrived at
it more compactly by utilizing the FV results directly. Identifying the time-
dependent change in $X(t)$ with the mirror displacement $\delta L(t)$ observed
in an interferometer, we have
$\displaystyle\frac{\delta L(t)}{L}=\frac{F_{Q}(t)}{m_{M}},$ (27)
or equivalently
$\displaystyle\langle\frac{\delta L^{2}(\Delta t)}{L^{2}}\rangle$
$\displaystyle=$ $\displaystyle\frac{A_{FV}(\Delta t)}{m_{M}^{2}}$
$\displaystyle=$ $\displaystyle 4\pi
G\int\frac{d^{3}k}{(2\pi)^{3}}\frac{\rho(k)}{k}\cos(k\Delta t)dk.$
Observational Consequences of the Pixellon. We now identify the pixellon with
the degrees-of-freedom having density of states $\rho(k)$ in Eq. (20). Taking
$p=k$ in Eq. (13), we have $\rho_{\rm pix}(k)\approx\frac{2m_{\rm pix}}{\beta
k^{2}}$ such that
$\displaystyle\langle\frac{\delta L^{2}(\Delta
t)}{L^{2}}\rangle\approx\frac{4\pi Gm_{\rm pix}}{\pi^{2}\beta}\int_{k_{\rm
min}}^{k_{\rm max}}\frac{dk}{k}\cos(k\Delta t),$ (29)
where ${k_{\rm min}},~{}{k_{\rm max}}$ are the maximum and minimum frequencies
detectable by the apparatus measuring the metric fluctuations. We previously
argued that $m_{\rm pix}$ is near the Planck scale; ultimately it should be
experimentally determined. Taking, for example, $2m_{\rm pix}/\beta=a/l_{p}L$,
and in the limit $\Delta t\rightarrow 0$, we recover
$\displaystyle\langle\frac{\delta
L^{2}}{L^{2}}\rangle\approx\frac{al_{p}}{4\pi^{2}L}\log\left(\frac{k_{\rm
max}}{k_{\rm min}}\right).$ (30)
Now $k$ is the momentum conjugate to the spacetime position $x$. Thus we can
identify $~{}k_{\rm min}$ with the overall apparatus size $k_{\rm min}\sim
1/L$, and the spatial separation between the two arms of the interferometer
$k_{\rm max}\sim 1/|{\bf r}_{1}-{\bf r}_{2}|$. Note that this is remarkably
similar to the result obtained in Ref. Verlinde and Zurek (2019) despite the
distinct Ansatz and calculational techniques.
The amplitude of the effect is within reach. For example, the interferometer
of Ref. Chou _et al._ (2017) was able to constrain a power spectral density
$S_{h}(\omega)\equiv\int_{-\infty}^{\infty}d\Delta t\langle\frac{\delta
L^{2}(\Delta t)}{L^{2}}\rangle e^{-i\omega\Delta t}\lesssim
0.15~{}l_{p}/\sqrt{8\pi}(\operatorname{sinc}^{2}(\omega
L/2)-2\operatorname{sinc}^{2}(\omega L))$. While a precise phenomenological
analysis comparing our model to the data is not the purpose of this present
work, we can obtain an estimate by integrating Eq. (29) times
$e^{-i\omega\Delta t}$ over $-L$ to $L$, taking $\omega\rightarrow 0$. With
this naive estimate, we find that the data is roughly consistent with
$a\lesssim 2$. This motivates future work to make precise phenomenological
predictions and to search for the unique signatures of this model. For
example, Ref. Verlinde and Zurek (2019) went one step further and used the
spherical symmetry of the entangling surface to argue that length fluctuations
correlated between two interferometer arms should obey the Green function of
the 2-d Laplacian on the sphere. While we have not repeated this calculation
here, the general form of Eq. (30) suggests similar behavior holds for this
model. The observation of these angular correlations would be a spectacular
confirmation of the idea of low momentum vacuum fluctuations in quantum
gravity, and could be tested with a suitable modification of a planned table
top interferometer experiment Vermeulen _et al._ (2020). We anticipate a
vibrant experimental program along these directions.
Acknowledgments. I especially thank Tom Banks, Cindy Keeler, Michele Papucci
and Julio Parra-Martinez for detailed comments on the draft. I thank many
others for discussion and criticism, including Dan Carney, Yanbei Chen,
Clifford Cheung, Patrick Draper, David B. Kaplan, Dongjun Li, Juan Maldacena,
Maulik Parikh, and Erik Verlinde.
## References
* Van Raamsdonk (2016) M. Van Raamsdonk, in _Theoretical Advanced Study Institute in Elementary Particle Physics: New Frontiers in Fields and Strings_ (2016) arXiv:1609.00026 [hep-th] .
* Ryu and Takayanagi (2006a) S. Ryu and T. Takayanagi, JHEP 08, 045 (2006a), arXiv:hep-th/0605073 .
* Ryu and Takayanagi (2006b) S. Ryu and T. Takayanagi, Phys. Rev. Lett. 96, 181602 (2006b), arXiv:hep-th/0603001 .
* Srednicki (1993) M. Srednicki, Phys. Rev. Lett. 71, 666 (1993), arXiv:hep-th/9303048 .
* Callan and Wilczek (1994) C. G. Callan, Jr. and F. Wilczek, Phys. Lett. B 333, 55 (1994), arXiv:hep-th/9401072 .
* Jacobson (1995) T. Jacobson, Phys. Rev. Lett. 75, 1260 (1995), arXiv:gr-qc/9504004 .
* Jacobson (2016) T. Jacobson, Phys. Rev. Lett. 116, 201101 (2016), arXiv:1505.04753 [gr-qc] .
* Verlinde and Zurek (2020) E. Verlinde and K. M. Zurek, JHEP 04, 209 (2020), arXiv:1911.02018 [hep-th] .
* Perlmutter (2014) E. Perlmutter, JHEP 03, 117 (2014), arXiv:1308.1083 [hep-th] .
* Nakaguchi and Nishioka (2016) Y. Nakaguchi and T. Nishioka, JHEP 12, 129 (2016), arXiv:1606.08443 [hep-th] .
* De Boer _et al._ (2019) J. De Boer, J. Järvelä, and E. Keski-Vakkuri, Phys. Rev. D 99, 066012 (2019), arXiv:1807.07357 [hep-th] .
* Casini _et al._ (2011) H. Casini, M. Huerta, and R. C. Myers, JHEP 05, 036 (2011), arXiv:1102.0440 [hep-th] .
* Verlinde (2011) E. P. Verlinde, JHEP 04, 029 (2011), arXiv:1001.0785 [hep-th] .
* Banks (2011) T. Banks, (2011), arXiv:1109.2435 [hep-th] .
* Verlinde and Zurek (2019) E. P. Verlinde and K. M. Zurek, (2019), arXiv:1902.08207 [gr-qc] .
* Dvali and Gomez (2013) G. Dvali and C. Gomez, Fortsch. Phys. 61, 742 (2013), arXiv:1112.3359 [hep-th] .
* Feynman and Vernon (1963) R. Feynman and J. Vernon, F.L., Annals Phys. 24, 118 (1963).
* Parikh _et al._ (2020a) M. Parikh, F. Wilczek, and G. Zahariade, (2020a), arXiv:2005.07211 [hep-th] .
* Parikh _et al._ (2020b) M. Parikh, F. Wilczek, and G. Zahariade, (2020b), arXiv:2010.08208 [hep-th] .
* Chou _et al._ (2017) A. Chou _et al._ (Holometer), Class. Quant. Grav. 34, 165005 (2017), arXiv:1703.08503 [gr-qc] .
* Vermeulen _et al._ (2020) S. Vermeulen, L. Aiello, A. Ejlli, W. Griffiths, A. James, K. Dooley, and H. Grote, (2020), arXiv:2008.04957 [gr-qc] .
|
# The Prime Geodesic Theorem in Arithmetic Progressions
Dimitrios Chatzakos Department of Mathematics, University of Patras, 26 110,
Patras, Greece https://dchatzakos.math.upatras.gr<EMAIL_ADDRESS>, Gergely Harcos Alfréd Rényi Institute of Mathematics, POB 127, Budapest
H-1364, Hungary https://users.renyi.hu/ gharcos/<EMAIL_ADDRESS>and Ikuya
Kaneko The Division of Physics, Mathematics and Astronomy, California
Institute of Technology, 1200 E. California Blvd., Pasadena, CA 91125, USA
https://sites.google.com/view/ikuyakaneko/<EMAIL_ADDRESS>
###### Abstract.
We address the prime geodesic theorem in arithmetic progressions, and resolve
conjectures of Golovchanskiĭ–Smotrov (1999). In particular, we prove that the
traces of closed geodesics on the modular surface do not equidistribute in the
reduced residue classes of a given modulus.
###### Key words and phrases:
prime geodesic theorem, arithmetic progressions, Kuznetsov–Bykovskiĭ formula
###### 2020 Mathematics Subject Classification:
11F72 (primary); 11M36 (secondary)
The first author acknowledges the financial support from the ELKE of the
University of Patras (MEDIKOS Program no. 82043). The second author was
supported by the Rényi Intézet Lendület Automorphic Research Group and NKFIH
(National Research, Development and Innovation Office) grant K 143876. The
third author acknowledges the support of the Masason Foundation.
## 1\. Introduction
### 1.1. Historical prelude
The prime geodesic theorem asks for an asymptotic evaluation of the counting
function of oriented primitive closed geodesics on hyperbolic manifolds. If
the underlying group is a cofinite Fuchsian group
$\Gamma\subset\mathrm{PSL}_{2}(\mathbb{R})$, then this problem has received
distinguished attention among number theorists; see for instance [Hej76a,
Hej76b, Hej83, Hub61a, Hub61b, Kuz78, Sar80, Sel89, Ven90] and references
therein for classical triumphs.
A closed geodesic $P$ on $\Gamma\backslash\mathbb{H}$ corresponds bijectively
to a hyperbolic conjugacy class in $\Gamma$ (cf. [Hub59]). If $\mathrm{N}(P)$
denotes the norm of this conjugacy class, then the hyperbolic length of $P$
equals $\log\mathrm{N}(P)$. As usual, we set
$\Lambda_{\Gamma}(P)=\log\mathrm{N}(P_{0})$, where $P_{0}$ is the primitive
closed geodesic underlying $P$, and we introduce the Chebyshev-like counting
function
$\Psi_{\Gamma}(x)\coloneqq\sum_{\mathrm{N}(P)\leq x}\Lambda_{\Gamma}(P).$
Because the prime geodesic theorem is reminiscent of the prime number theorem,
the norms are sometimes called pseudoprimes. Selberg [Sel89] established an
asymptotic formula of the shape
$\Psi_{\Gamma}(x)=\sum_{\frac{1}{2}<s_{j}\leq
1}\frac{x^{s_{j}}}{s_{j}}+\mathcal{E}_{\Gamma}(x),$ (1.1)
where the main term emerges from the small eigenvalues
$\lambda_{j}=s_{j}(1-s_{j})<\frac{1}{4}$ of the Laplacian on
$\Gamma\backslash\mathbb{H}$ for the upper half-plane $\mathbb{H}$, and
$\mathcal{E}_{\Gamma}(x)$ is an error term. It is known that
$\mathcal{E}_{\Gamma}(x)\ll_{\Gamma,\varepsilon}x^{\frac{3}{4}+\varepsilon}$;
see the explicit formulæ in [Iwa84b, KK22]. This barrier is often termed the
trivial bound. Given an analogue of the Riemann Hypothesis for Selberg zeta
functions apart from a finite number of the exceptional zeros, one should
expect the best possible estimate to be
$\mathcal{E}_{\Gamma}(x)\ll_{\Gamma,\varepsilon}x^{\frac{1}{2}+\varepsilon}$.
This remains unresolved due to the abundance of Laplace eigenvalues.
If $\Gamma$ is arithmetic, then an improvement over the barrier was achieved
by Iwaniec [Iwa84b], who proved that
$\mathcal{E}_{\Gamma}(x)\ll_{\varepsilon}x^{\frac{35}{48}+\varepsilon}$ for
the full modular group $\Gamma=\mathrm{PSL}_{2}(\mathbb{Z})$. Iwaniec [Iwa84a]
stated that the stronger exponent $\frac{2}{3}+\varepsilon$ follows from the
Generalised Lindelöf Hypothesis for quadratic Dirichlet $L$-functions.
Furthermore, under the Generalised Lindelöf Hypothesis for Rankin–Selberg
$L$-functions, the same exponent follows from the Weil bound for Kloosterman
sums. Luo–Sarnak [LS95] then strengthened the machinery of Iwaniec to obtain
the exponent $\frac{7}{10}+\varepsilon$; see also [Koy98, LRS95]. As a further
refinement, Cai [Cai02] derived the exponent $\frac{71}{102}+\varepsilon$. The
crucial gist in all these works is to estimate nontrivially a certain spectral
exponential sum via the Kuznetsov formula. On the other hand, the subsequent
work of Soundararajan–Young [SY13] demonstrated that
$\mathcal{E}_{\Gamma}(x)\ll_{\varepsilon}x^{\frac{2}{3}+\frac{\vartheta}{6}+\varepsilon},$
where $\vartheta$ is a subconvex exponent for quadratic Dirichlet
$L$-functions. The current record $\vartheta=\frac{1}{6}$ by Conrey–Iwaniec
[CI00] implies the best known exponent $\frac{25}{36}+\varepsilon$. The proof
utilises the Kuznetsov–Bykovskiĭ formula (see [Byk94, Kuz78, SY13]), leaving
the theory of Selberg zeta functions aside. For recent progress on the prime
geodesic theorem and its generalisations, we direct the reader to [BBCL22,
BBHM19, BCC+19, BF19, BF20, BFR22, CCL22, CG18, CWZ22, DM23, Kan20, Kan22,
Kan23, Koy01, PR17].
### 1.2. Statement of main results
By a classical theorem of Dirichlet, the primes equidistribute in the reduced
residue classes of a given modulus. As we shall see, the prime geodesic
analogue of this phenomenon breaks down, and the corresponding non-uniform
distribution can be determined explicitly.
We fix $\Gamma=\mathrm{PSL}_{2}(\mathbb{Z})$. By definition, an element
$P\in\Gamma$ is said to be hyperbolic if as a Möbius transformation
$Pz=\frac{az+b}{cz+d},$
it possesses two distinct real fixed points. By conjugation, any hyperbolic
element $P\in\Gamma$ may be expressed as $P=\sigma^{-1}\widetilde{P}\sigma$,
where $\sigma\in\mathrm{PSL}_{2}(\mathbb{R})$ and
$\widetilde{P}=\begin{pmatrix}\lambda&0\\\ 0&\lambda^{-1}\end{pmatrix}$ with
$\lambda>1$. Here $\widetilde{P}$ acts as multiplication by $\lambda^{2}$,
namely $\widetilde{P}z=\lambda^{2}z$. The factor $\lambda^{2}$ is called the
norm of $P$, which depends only on the
$\mathrm{PSL}_{2}(\mathbb{R})$-conjugacy class of $P$. We note that the
positive trace
$t=\mathrm{tr}(P)=\mathrm{tr}(\widetilde{P})=\lambda+\lambda^{-1}$
is an integer exceeding $2$, and hence the norm takes the form
$\mathrm{N}(P)=\left(\frac{t+\sqrt{t^{2}-4}}{2}\right)^{2}=t^{2}-2+O(t^{-2}).$
Given a prime $p\geq 2$, we define the Chebyshev-like counting function in
arithmetic progressions modulo $p$ by
$\Psi_{\Gamma}(x;p,a)\coloneqq\sum_{\begin{subarray}{c}\mathrm{N}(P)\leq x\\\
\mathrm{tr}(P)\equiv
a\,(\mathrm{mod}{\,p})\end{subarray}}\Lambda_{\Gamma}(P).$
The following result shows that the main term in an asymptotic for
$\Psi_{\Gamma}(x;p,a)$ depends on the residue class $a\,(\mathrm{mod}{\,p})$
unlike for primes in arithmetic progressions.
###### Theorem 1.1.
Let $\Gamma=\mathrm{PSL}_{2}(\mathbb{Z})$, and let $p\geq 3$ be a prime. Then
we have that
$\Psi_{\Gamma}(x;p,a)=\begin{dcases}\frac{1}{p-1}\cdot
x+O_{\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon})&\text{if
$\left(\frac{a^{2}-4}{p}\right)=1$,}\\\ \frac{1}{p+1}\cdot
x+O_{\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon})&\text{if
$\left(\frac{a^{2}-4}{p}\right)=-1$,}\\\ \frac{p}{p^{2}-1}\cdot
x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon})&\text{if
$\left(\frac{a^{2}-4}{p}\right)=0$,}\end{dcases}$ (1.2)
where $\vartheta$ is a subconvex exponent for quadratic Dirichlet
$L$-functions.
###### Remark 1.
The implied constant in the error term is independent of $p$ when
$a\not\equiv\pm 2\,(\mathrm{mod}{\,p})$.
###### Remark 2.
When $p=3$, the first case of (1.2) is void, while the second case is covered
with a stronger error term by [GS99, Theorem $1$].
###### Remark 3.
Apart from the size of the error term, Theorem 1.1 resolves [GS99, Conjecture
$2$] for level $N=1$. In fact, we expect that our method works for
$\Gamma=\Gamma_{0}(N)$ when $(N,p)=1$, but we restrict to
$\Gamma=\mathrm{PSL}_{2}(\mathbb{Z})$ for simplicity. Furthermore, we expect
that the error term can be improved significantly by a more careful analysis
(e.g. by combining the Kuznetsov–Bykovskiĭ formula with an adèlic trace
formula), but we solely focused on determining the main term according to the
sign of the Legendre symbol. We leave such pursuits for future work.
The method of Golovchanskiĭ–Smotrov [GS99, Theorem $1$] is different from
ours, and they delve into properties of the norms and traces, expressing
$\Psi_{\Gamma}(x;p,a)$ as a linear combination of
$\Psi_{\Gamma_{0}(2^{k})}(x)$ for some $k\geq 0$, for which an asymptotic
formula is already known as in (1.1). For example, they derived a general
linear combination of the shape
$3\Psi_{\Gamma_{0}(N)}(x)-3\Psi_{\Gamma_{0}(2N)}(x)+\Psi_{\Gamma_{0}(4N)}(x)=3\Psi_{\Gamma_{0}(N)}(x;2,1),$
from which it follows that111We emphasise that the case of $p=2$ is not
contained in Theorem 1.1.
$\Psi_{\Gamma_{0}(N)}(x;2,1)=\frac{1}{3}\cdot
x+\mathcal{E}_{\Gamma_{0}(N)}(x),\qquad\Psi_{\Gamma_{0}(N)}(x;2,0)=\frac{2}{3}\cdot
x+\mathcal{E}_{\Gamma_{0}(N)}(x).$
The level structure that they developed is delicate, and it appears that their
idea only works for some specific values of $p$ and $a$. Hence, some new ideas
are needed to prove Theorem 1.1.
An elementary counting argument implies the following result.
###### Corollary 1.2.
Let $\Gamma=\mathrm{PSL}_{2}(\mathbb{Z})$, and let $p\geq 3$ be a prime. Then
we have that
$\displaystyle\sum_{\begin{subarray}{c}a\,(\mathrm{mod}{\,p})\\\
\left(\frac{a^{2}-4}{p}\right)=1\end{subarray}}\Psi_{\Gamma}(x;p,a)$
$\displaystyle=\frac{p-3}{2(p-1)}\cdot
x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}),$
$\displaystyle\sum_{\begin{subarray}{c}a\,(\mathrm{mod}{\,p})\\\
\left(\frac{a^{2}-4}{p}\right)=-1\end{subarray}}\Psi_{\Gamma}(x;p,a)$
$\displaystyle=\frac{p-1}{2(p+1)}\cdot
x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}),$
$\displaystyle\sum_{\begin{subarray}{c}a\,(\mathrm{mod}{\,p})\\\
\left(\frac{a^{2}-4}{p}\right)=0\end{subarray}}\Psi_{\Gamma}(x;p,a)$
$\displaystyle=\frac{2p}{p^{2}-1}\cdot
x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}),$
where $\vartheta$ is a subconvex exponent for quadratic Dirichlet
$L$-functions.
###### Remark 4.
Apart from the size of the error term, Corollary 1.2 resolves [GS99,
Conjecture $1$] in the case of full level $N=1$, and again the method should
work for $\Gamma=\Gamma_{0}(N)$ when $(N,p)=1$.
### Acknowledgements
The authors thank Mikhail Nikolaevich Smotrov for sending us the preprint
[GS99].
## 2\. Key propositions
This section prepares for the proof of Theorem 1.1. Throughout, we follow
[SY13, Section 2] closely.
Let $\Gamma=\mathrm{PSL}_{2}(\mathbb{Z})$ as before. Sarnak [Sar82,
Proposition 1.4] showed that the primitive hyperbolic conjugacy classes in
$\Gamma$ correspond bijectively to the $\Gamma$-equivalence classes of
primitive indefinite binary quadratic forms. We recall this correspondence
briefly. For a given primitive quadratic form $ax^{2}+bxy+cy^{2}$ of
discriminant $d>0$, the automorphs are the elements
$P(t,u)=\begin{pmatrix}\dfrac{t-bu}{2}&-cu\\\
au&\dfrac{t+bu}{2}\end{pmatrix}\in\Gamma,$
with $t^{2}-du^{2}=4$ being a solution of the Pell equation. For $u$ nonzero,
$P(t,u)$ is hyperbolic with norm $(t+u\sqrt{d})^{2}/4$ and trace $t$. Because
$P(-t,-u)=P(t,u)$ holds in $\Gamma$, we shall restrict to $t>0$ without loss
of generality. This is in harmony with our convention in Section 1.2 that
$\mathrm{tr}(P)>2$ for a hyperbolic element $P\in\Gamma$. If $(t_{d},u_{d})$
denotes the fundamental solution of the Pell equation, then $P(t_{d},u_{d})$
is a primitive hyperbolic matrix of norm $\varepsilon_{d}^{2}$ and trace
$t_{d}$. Moreover, every automorph $P(t,u)$ with $u>0$ (resp. $u<0$) is a
unique positive (resp. negative) integral power of $P(t_{d},u_{d})$. Sarnak’s
bijection sends the quadratic form $ax^{2}+bxy+cy^{2}$ to the conjugacy class
of $P(t_{d},u_{d})$ in $\Gamma$. Thus, for a given discriminant $d>0$, there
are $h(d)$ primitive hyperbolic conjugacy classes in $\Gamma$, each of norm
$\varepsilon_{d}^{2}$ and trace $t_{d}$.
Now every hyperbolic conjugacy class $\\{P\\}$ can be written uniquely as
$\\{P_{0}^{n}\\}$ for $n\geq 1$ and a primitive hyperbolic conjugacy class
$\\{P_{0}\\}$ (cf. [Hub59]). Combining this with Sarnak’s bijection described
above, we obtain
$\Psi_{\Gamma}(x;p,a)=2\sum_{\begin{subarray}{c}3\leq t\leq X\\\ t\equiv
a\,(\mathrm{mod}{\,p})\end{subarray}}\sum_{t^{2}-du^{2}=4}h(d)\log\varepsilon_{d},$
where $X$ abbreviates $\sqrt{x}+\frac{1}{\sqrt{x}}$, and $d>0$ (resp. $u>0$)
runs through discriminants (resp. integers). The class number formula
$h(d)\log\varepsilon_{d}=\sqrt{d}L(1,\chi_{d})$, where $\chi_{d}$ is the not
necessarily primitive quadratic Dirichlet character associated to the
discriminant $d$, allows us to write
$\Psi_{\Gamma}(x;p,a)=2\sum_{\begin{subarray}{c}3\leq t\leq X\\\ t\equiv
a\,(\mathrm{mod}{\,p})\end{subarray}}\sum_{t^{2}-du^{2}=4}\sqrt{d}L(1,\chi_{d}).$
(2.1)
For an arbitrary discriminant $\delta>0$, we define Zagier’s $L$-series by
(cf. [SY13, (6) & (3)])
$L(s,\delta)\coloneqq\sum_{du^{2}=\delta}L(s,\chi_{d})u^{1-2s}=\sum_{q=1}^{\infty}\frac{\lambda_{q}(\delta)}{q^{s}},$
where $d>0$ (resp. $u>0$) runs through discriminants (resp. integers). This
series admits a transparent Euler product expansion (to be discussed below),
while it simplifies (2.1) as
$\Psi_{\Gamma}(x;p,a)=2\sum_{\begin{subarray}{c}3\leq t\leq X\\\ t\equiv
a\,(\mathrm{mod}{\,p})\end{subarray}}\sqrt{t^{2}-4}L(1,t^{2}-4).$ (2.2)
If $\delta=Dl^{2}$, where $D>0$ is a fundamental discriminant and $l>0$ is an
integer, then we obtain the following Euler product expansion of Zagier’s
$L$-series (cf. [SY13, (2)]):
$\displaystyle L(s,\delta)$ $\displaystyle=\sum_{u\mid
l}L(s,\chi_{Dl^{2}/u^{2}})u^{1-2s}$ $\displaystyle=L(s,\chi_{D})\sum_{u\mid
l}u^{1-2s}\prod_{\bm{p}\mid\frac{l}{u}}(1-\chi_{D}(\bm{p}))$
$\displaystyle=\prod_{\bm{p}}\left(\sum_{0\leq
m<v_{\bm{p}}(l)}\bm{p}^{m(1-2s)}+\frac{\bm{p}^{v_{\bm{p}}(l)(1-2s)}}{1-\chi_{D}(\bm{p})\bm{p}^{-s}}\right).$
(2.3)
In particular, for fixed $\delta$, the arithmetic function
$q\mapsto\lambda_{q}(\delta)$ is multiplicative. The idea of the proof of
Theorem 1.1 is to group together certain values of $t$ in (2.2) such that the
corresponding Zagier $L$-series $L(s,t^{2}-4)$ has a constant Euler factor at
$\bm{p}=p$. Thus we are led to consider $L(s,\delta)$ with its Euler factor at
$\bm{p}=p$ removed:
$L^{p}(s,\delta)\coloneqq\sum_{\begin{subarray}{c}q\geq 1\\\
(q,p)=1\end{subarray}}\frac{\lambda_{q}(\delta)}{q^{s}}.$
###### Proposition 2.1.
Let $p\geq 3$ be a prime, and let $n\geq 1$ be an integer. Let
$r\,(\mathrm{mod}{\,p^{n}})$ be an arbitrary residue class. If $(q,p)=1$ and
$b$ denotes the squarefree part of $q$, then
$\sum_{\begin{subarray}{c}3\leq t\leq X\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}\lambda_{q}(t^{2}-4)=\frac{X}{p^{n}}\cdot\frac{\mu(b)}{b}+O_{\varepsilon}(q^{\frac{1}{2}+\varepsilon}).$
###### Proof.
It follows from [SY13, Lemma 2.3] that
$\lambda_{q}(t^{2}-4)=\sum_{q_{1}^{2}q_{2}=q}\frac{1}{q_{2}}\sum_{k\,(\mathrm{mod}{\,q_{2}})}e\left(\frac{kt}{q_{2}}\right)S(k^{2},1;q_{2}).$
This leads to
$\sum_{\begin{subarray}{c}3\leq t\leq X\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}\lambda_{q}(t^{2}-4)=\sum_{q_{1}^{2}q_{2}=q}\frac{1}{q_{2}}\sum_{k\,(\mathrm{mod}{\,q_{2}})}S(k^{2},1;q_{2})\sum_{\begin{subarray}{c}3\leq
t\leq X\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}e\left(\frac{kt}{q_{2}}\right).$
When $k\equiv 0\,(\mathrm{mod}{\,q_{2}})$, the inner sum over $t$ is
$\frac{X}{p^{n}}+O(1)$, and $S(0,1;q_{2})=\mu(q_{2})$, yielding the expected
main term. When $k\not\equiv 0\,(\mathrm{mod}{\,q_{2}})$, we apply the Weil
bound for Kloosterman sums and the estimate
$\sum_{\begin{subarray}{c}3\leq t\leq X\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}e\left(\frac{kt}{q_{2}}\right)\ll\left\lVert\frac{kp^{n}}{q_{2}}\right\rVert^{-1}\leq
q_{2},$
where $\left\lVert\cdot\right\rVert$ is the distance to the nearest integer.
The proof is complete. ∎
Guided by Proposition 2.1 and (2.2), we consider the sum
$\Psi_{\Gamma}^{\star}(x;p^{n},r)\coloneqq 2\sum_{\begin{subarray}{c}3\leq
t\leq X\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}\sqrt{t^{2}-4}L^{p}(1,t^{2}-4).$
(2.4)
We shall deduce Theorem 1.1 from the following analogue of [SY13, Theorem
3.2]:
###### Proposition 2.2.
Let $p\geq 3$ be a prime, and let $n\geq 1$ be an integer. Let
$r\,(\mathrm{mod}{\,p^{n}})$ be an arbitrary residue class. Then
$\Psi_{\Gamma}^{\star}(x+u;p^{n},r)-\Psi_{\Gamma}^{\star}(x;p^{n},r)=\frac{u}{p^{n}}+O_{\varepsilon}(u^{\frac{1}{2}}x^{\frac{1}{4}+\frac{\vartheta}{2}+\varepsilon}),\qquad\sqrt{x}\leq
u\leq x.$ (2.5)
###### Proof.
Let $\sqrt{x}\leq u\leq x$, and set
$X\coloneqq\sqrt{x}+\frac{1}{\sqrt{x}}\qquad\text{and}\qquad
X^{\prime}\coloneqq\sqrt{x+u}+\frac{1}{\sqrt{x+u}}.$
From the definition (2.4), it is clear that
$\displaystyle\Psi_{\Gamma}^{\star}(x+u;p^{n},r)-\Psi_{\Gamma}^{\star}(x;p^{n},r)$
$\displaystyle=2\sum_{\begin{subarray}{c}X<t\leq X^{\prime}\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}\sqrt{t^{2}-4}L^{p}(1,t^{2}-4)$
$\displaystyle=\left(2+O\left(x^{-1}\right)\right)\sum_{\begin{subarray}{c}X<t\leq
X^{\prime}\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}tL^{p}(1,t^{2}-4),$
because $\sqrt{t^{2}-4}=t(1+O(t^{-2}))$. We shall approximate
$L^{p}(1,t^{2}-4)$ in terms of a suitable Dirichlet series. Let $V\geq 1$ be a
parameter to be chosen later, and let
$\delta=t^{2}-4=Dl^{2},$
where $D>0$ is a fundamental discriminant and $l>0$ is an integer. Consider
$\Lambda_{V}^{p}(\delta)\coloneqq\sum_{\begin{subarray}{c}q\geq 1\\\
(q,p)=1\end{subarray}}\frac{\lambda_{q}(\delta)}{q}e^{-\frac{q}{V}}.$
Shifting the contour yields the expression
$\Lambda_{V}^{p}(\delta)=\int_{(1)}L^{p}(1+s,\delta)V^{s}\Gamma(s)\frac{ds}{2\pi
i}=L^{p}(1,\delta)+\int_{(-\frac{1}{2})}L^{p}(1+s,\delta)V^{s}\Gamma(s)\frac{ds}{2\pi
i}.$
On the right-hand side, for some $A>0$,
$L^{p}(1+s,\delta)\ll|L(1+s,\delta)|\ll_{\varepsilon}|L(1+s,\chi_{D})|l^{\varepsilon}\ll_{\varepsilon}\delta^{\vartheta+\varepsilon}|s|^{A},\qquad\mathrm{Re}s=-\frac{1}{2},$
while $\Gamma(s)$ decays exponentially. It follows that
$L^{p}(1,\delta)=\Lambda_{V}^{p}(\delta)+O_{\varepsilon}(\delta^{\vartheta+\varepsilon}V^{-\frac{1}{2}}),$
and hence
$\Psi_{\Gamma}^{\star}(x+u;p^{n},r)-\Psi_{\Gamma}^{\star}(x;p^{n},r)=(2+O(x^{-1}))\sum_{\begin{subarray}{c}X<t\leq
X^{\prime}\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}t\Lambda_{V}^{p}(t^{2}-4)+O_{\varepsilon}(ux^{\vartheta+\varepsilon}V^{-\frac{1}{2}}).$
If $(q,p)=1$ and $q=bc^{2}$ with $b$ squarefree, then Proposition 2.1 along
with partial summation leads to
$2\sum_{\begin{subarray}{c}X<t\leq X^{\prime}\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}t\lambda_{q}(t^{2}-4)=\frac{u}{p^{n}}\cdot\frac{\mu(b)}{b}+O_{\varepsilon}(Xq^{\frac{1}{2}+\varepsilon}).$
It thus follows that
$2\sum_{\begin{subarray}{c}X<t\leq X^{\prime}\\\ t\equiv
r\,(\mathrm{mod}{\,p^{n}})\end{subarray}}t\Lambda_{V}^{p}(t^{2}-4)=\frac{u}{p^{n}}\sum_{\begin{subarray}{c}b,c\geq
1\\\
(bc,p)=1\end{subarray}}\frac{\mu(b)}{b^{2}c^{2}}e^{-\frac{bc^{2}}{V}}+O_{\varepsilon}(XV^{\frac{1}{2}+\varepsilon}).$
A standard contour shift argument gives
$\sum_{\begin{subarray}{c}b,c\geq 1\\\
(bc,p)=1\end{subarray}}\frac{\mu(b)}{b^{2}c^{2}}e^{-\frac{bc^{2}}{V}}=\int_{(1)}V^{s}\Gamma(s)\frac{\zeta^{p}(2+2s)}{\zeta^{p}(2+s)}\frac{ds}{2\pi
i}=1+O(V^{-\frac{1}{2}}),$
where $\zeta^{p}(s)$ is the Riemann zeta function with the Euler factor at $p$
removed. As a result,
$\Psi_{\Gamma}^{\star}(x+u;p^{n},r)-\Psi_{\Gamma}^{\star}(x;p^{n},r)=\frac{u}{p^{n}}+O_{\varepsilon}(x^{\frac{1}{2}}V^{\frac{1}{2}+\varepsilon}+ux^{\vartheta+\varepsilon}V^{-\frac{1}{2}}).$
Setting $V=ux^{-\frac{1}{2}+\vartheta}$ yields (2.5). ∎
###### Corollary 2.3.
Let $p\geq 3$ be a prime, and let $n\geq 1$ be an integer. Let
$r\,(\mathrm{mod}{\,p^{n}})$ be an arbitrary residue class. Then
$\Psi_{\Gamma}^{\star}(x;p^{n},r)=\frac{x}{p^{n}}+O_{\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}).$
(2.6)
###### Proof.
Setting $u=x$ in (2.5) yields
$\Psi_{\Gamma}^{\star}(2x;p^{n},r)-\Psi_{\Gamma}^{\star}(x;p^{n},r)=\frac{x}{p^{n}}+O_{\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}).$
Now (2.6) follows readily by a dyadic subdivision. ∎
## 3\. Proof of Theorem 1.1
We shall approximate the $t$-sum in (2.2). As before, we write
$\delta=t^{2}-4=Dl^{2},$
where $D>0$ is a fundamental discriminant and $l>0$ is an integer.
If $a\not\equiv\pm 2\,(\mathrm{mod}{\,p})$, then for every $t$ participating
in (2.2), we have that $p\nmid l$ and
$\chi_{D}(p)=\left(\frac{D}{p}\right)=\left(\frac{Dl^{2}}{p}\right)=\left(\frac{t^{2}-4}{p}\right)=\left(\frac{a^{2}-4}{p}\right).$
By (2.3), the corresponding Zagier $L$-series $L(s,t^{2}-4)$ factorises as
$L(s,t^{2}-4)=\left(1-\left(\frac{a^{2}-4}{p}\right)p^{-s}\right)^{-1}L^{p}(s,t^{2}-4),$
hence by (2.2) and (2.4), it also follows that
$\Psi_{\Gamma}(x;p,a)=\left(1-\left(\frac{a^{2}-4}{p}\right)p^{-1}\right)^{-1}\Psi_{\Gamma}^{\star}(x;p,a).$
Applying (2.6), we obtain the first two cases of (1.2).
If $a\equiv\pm 2\,(\mathrm{mod}{\,p})$, then we shall assume (without loss of
generality) that $a=\pm 2$. We subdivide the $t$-sum in (2.2) according to the
exponent of $p$ in the positive integer $t-a$:
$\Psi_{\Gamma}(x;p,a)=\sum_{k=1}^{\infty}\Psi_{\Gamma}(x;p,a;k),$
where
$\Psi_{\Gamma}(x;p,a;k)\coloneqq 2\sum_{\begin{subarray}{c}3\leq t\leq X\\\
v_{p}(t-a)=k\end{subarray}}\sqrt{t^{2}-4}L(1,t^{2}-4).$
We shall approximate these pieces individually. Note that
$\Psi_{\Gamma}(x;p,a;k)=0$ for $p^{k}>t-a$. Moreover, the condition
$v_{p}(t-a)=k$ constrains $t$ to $p-1$ residue classes modulo $p^{k+1}$, and
it also yields $v_{p}(t^{2}-4)=k$.
If $k=2n-1$ is odd, then $p\mid D$ and $v_{p}(l)=n-1$, hence by (2.3),
$L(s,t^{2}-4)=\frac{1-p^{n(1-2s)}}{1-p^{1-2s}}L^{p}(s,t^{2}-4).$
Using also (2.4) and (2.6), we obtain
$\displaystyle\Psi_{\Gamma}(x;p,a;2n-1)$
$\displaystyle=\frac{p-1}{p^{2n}}\cdot\frac{1-p^{-n}}{1-p^{-1}}\cdot
x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon})$
$\displaystyle=(p^{1-2n}-p^{1-3n})x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}).$
(3.1)
It is important that the implied constant in the error term is independent of
$n$.
If $k=2n$ is even, then $p\nmid D$ and $v_{p}(l)=n$, hence by (2.3),
$L(s,t^{2}-4)=\left(\frac{1-p^{n(1-2s)}}{1-p^{1-2s}}+\frac{p^{n(1-2s)}}{1-\chi_{D}(p)p^{-s}}\right)L^{p}(s,t^{2}-4).$
We can understand $\chi_{D}(p)$ by writing $t=a+p^{2n}r$. Indeed, then
$t^{2}-4=2ap^{2n}r+p^{4n}r^{2}$, hence
$\chi_{D}(p)=\left(\frac{D}{p}\right)=\left(\frac{Dl^{2}p^{-2n}}{p}\right)=\left(\frac{2ar}{p}\right).$
This means that among the $p-1$ choices for $t\,(\mathrm{mod}{\,p^{2n+1}})$,
half the time $\chi_{D}(p)$ equals $1$ and half the time it equals $-1$. Using
also (2.4) and (2.6), we obtain
$\displaystyle\Psi_{\Gamma}(x;p,a;2n)$
$\displaystyle=\frac{p-1}{p^{2n+1}}\left(\frac{1-p^{-n}}{1-p^{-1}}+\frac{(1/2)p^{-n}}{1-p^{-1}}+\frac{(1/2)p^{-n}}{1+p^{-1}}\right)x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon})$
$\displaystyle=\left(p^{-2n}-\frac{p^{-3n}}{p+1}\right)x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}).$
(3.2)
Again, the implied constant in the error term is independent of $n$.
Summing up the pieces $\Psi_{\Gamma}(x;p,a;2n-1)$ and
$\Psi_{\Gamma}(x;p,a;2n)$ for $1\leq n\leq\log(X+2)$, and inserting the
approximations (3.1)–(3.2), we deduce the asymptotic formula
$\Psi_{\Gamma}(x;p,\pm
2)=c_{p}x+O_{p,\varepsilon}(x^{\frac{3}{4}+\frac{\vartheta}{2}+\varepsilon}),$
where
$c_{p}\coloneqq\sum_{n=1}^{\infty}\left(p^{1-2n}-p^{1-3n}+p^{-2n}-\frac{p^{-3n}}{p+1}\right)=\frac{p}{p^{2}-1}.$
This is the third case of (1.2), and the proof of Theorem 1.1 is complete.
## 4\. Proof of Corollary 1.2
There are $\frac{p-3}{2}$ residues $a\,(\mathrm{mod}{\,p})$ such that
$a^{2}-4$ is a nonzero quadratic residue. Indeed, this occurs if and only if
$a^{2}-4\equiv b^{2}\,(\mathrm{mod}{\,p})$ for some $b\not\equiv
0\,(\mathrm{mod}{\,p})$, which can be written as $(a+b)(a-b)\equiv
4\,(\mathrm{mod}{\,p})$. Making the change of variables $a+b\equiv
2x\,(\mathrm{mod}{\,p})$ and $a-b\equiv 2x^{-1}\,(\mathrm{mod}{\,p})$, we
obtain $a\equiv x+x^{-1}\,(\mathrm{mod}{\,p})$ and $b\equiv
x-x^{-1}\,(\mathrm{mod}{\,p})$ with the condition
$x\not\equiv-1,0,1\,(\mathrm{mod}{\,p})$. This restriction leaves
$\frac{p-3}{2}$ different ways to choose $a\,(\mathrm{mod}{\,p})$. Since there
are two solutions to $(\frac{a^{2}-4}{p})=0$, we conclude that there are
$\frac{p-1}{2}$ residues $a\,(\mathrm{mod}{\,p})$ such that $a^{2}-4$ is a
nonzero quadratic nonresidue. Now Corollary 1.2 is immediate from Theorem 1.1.
## References
* [BBCL22] A. Balog, A. Biró, G. Cherubini, and N. Laaksonen, _Bykovskii-type theorem for the Picard manifold_ , Int. Math. Res. Not. IMRN 2022 (2022), no. 3, 1893–1921.
* [BBHM19] A. Balog, A. Biró, G. Harcos, and P. Maga, _The prime geodesic theorem in square mean_ , J. Number Theory 198 (2019), 239–249, Errata: https://users.renyi.hu/~gharcos/primegeodesic_errata.txt.
* [BCC+19] O. Balkanova, D. Chatzakos, G. Cherubini, D. Frolenkov, and N. Laaksonen, _Prime geodesic theorem in the 3-dimensional hyperbolic space_ , Trans. Amer. Math. Soc. 372 (2019), no. 8, 5355–5374.
* [BF19] O. Balkanova and D. Frolenkov, _Sums of Kloosterman sums in the prime geodesic theorem_ , Quart. J. Math. 70 (2019), no. 2, 649–674.
* [BF20] by same author, _Prime geodesic theorem for the Picard manifold_ , Adv. Math. 375 (2020), Paper No. 107377, 42 pages.
* [BFR22] O. Balkanova, D. Frolenkov, and M. S. Risager, _Prime geodesics and averages of the Zagier $L$-series_, Math. Proc. Cambridge Philos. Soc. 172 (2022), no. 3, 705–728.
* [Byk94] V. A. Bykovskiĭ, _Density theorems and the mean value of arithmetic functions in short intervals (Russian)_ , Zap. Nauchn. Sem. S.-Peterburg. Otdel. Mat. Inst. Steklov. (POMI) 212 (1994), no. Anal. Teor. Chisel i Teor. Funktsiĭ. 12, 56–70, 196, translation in J. Math. Sci. (N.Y.) 83 (1997), no. 6, 720–730.
* [Cai02] Y. Cai, _Prime geodesic theorem_ , J. Théor. Nombres Bordeaux 14 (2002), no. 2, 59–72.
* [CCL22] D. Chatzakos, G. Cherubini, and N. Laaksonen, _Second moment of the prime geodesic theorem for $\mathrm{PSL}(2,\mathbb{Z}[i])$_, Math. Z. 300 (2022), no. 1, 791–806.
* [CG18] G. Cherubini and J. Guerreiro, _Mean square in the prime geodesic theorem_ , Algebra & Number Theory 12 (2018), no. 3, 571–597.
* [CI00] J. B. Conrey and H. Iwaniec, _The cubic moment of central values of automorphic $L$-functions_, Ann. of Math. (2) 151 (2000), no. 3, 1175–1216.
* [CWZ22] G. Cherubini, H. Wu, and G. Zábrádi, _On Kuznetsov–Bykovskii’s formula of counting prime geodesics_ , Math. Z. 300 (2022), no. 1, 881–928.
* [DM23] L. Dever and D. Milićević, _Ambient prime geodesic theorems on hyperbolic $3$-manifolds_, Int. Math. Res. Not. IMRN 2023 (2023), no. 1, 588–635.
* [GS99] V. V. Golovchanskiĭ and M. N. Smotrov, _The distribution of the number classes of the primitive hyperbolic elements of the group $\Gamma_{0}(N)$ in arithmetic progressions (Russian)_, Preprint FEB RAS Khabarovsk Division of the Institute for Applied Mathematics (1999), no. 15, 15 pages, https://drive.google.com/file/d/1MxrJbbhLTGKzh4-1lCWa4Up_EDGeaqHc/view.
* [Hej76a] D. Hejhal, _The Selberg trace formula and the Riemann zeta function_ , Duke Math. J. 43 (1976), no. 3, 441–482.
* [Hej76b] by same author, _The Selberg trace formula for $\mathrm{PSL}(2,\mathbb{R})$ I_, Lecture Notes in Mathematics, vol. 548, Springer-Verlag, Berlin Heidelberg, 1976.
* [Hej83] by same author, _The Selberg trace formula for $\mathrm{PSL}(2,\mathbb{R})$ II_, Lecture Notes in Mathematics, vol. 1001, Springer-Verlag, Berlin Heidelberg, 1983.
* [Hub59] H. Huber, _Zur analytischen Theorie hyperbolischer Raumformen und Bewegungsgruppen_ , Math. Ann. 138 (1959), no. 1, 1–26.
* [Hub61a] by same author, _Zur analytischen Theorie hyperbolischer Raumformen und Bewegungsgruppen. II_ , Math. Ann. 142 (1961), no. 4, 385–398.
* [Hub61b] by same author, _Zur analytischen Theorie hyperbolischer Raumformen und Bewegungsgruppen. II. Nachtrag zu Math. Ann. 142, 385–398 (1961)_, Math. Ann. 143 (1961), no. 5, 463–464.
* [Iwa84a] H. Iwaniec, _Non-holomorphic modular forms and their applications_ , Modular forms (Durham, 1983) (R. A. Rankin, ed.), Ellis Horwood Ser. Math. Appl., Statist. Oper. Res., Horwood, Chichester, 1984, pp. 157–196.
* [Iwa84b] by same author, _Prime geodesic theorem_ , J. Reine Angew. Math. 349 (1984), 136–159.
* [Kan20] I. Kaneko, _The second moment for counting prime geodesics_ , Proc. Japan Acad. Ser. A Math. Sci. 96 (2020), no. 1, 7–12.
* [Kan22] by same author, _The prime geodesic theorem for $\mathrm{PSL}_{2}(\mathbb{Z}[i])$ and spectral exponential sums_, Algebra Number Theory 16 (2022), no. 8, 1845–1887.
* [Kan23] by same author, _Spectral exponential sums on hyperbolic surfaces_ , Ramanujan J. 60 (2023), no. 3, 837–846.
* [KK22] I. Kaneko and S. Koyama, _Euler products of Selberg zeta functions in the critical strip_ , Ramanujan J. 59 (2022), no. 2, 437–458.
* [Koy98] S. Koyama, _Prime geodesic theorem for arithmetic compact surfaces_ , Int. Math. Res. Not. IMRN 1998 (1998), no. 8, 383–388.
* [Koy01] by same author, _Prime geodesic theorem for the Picard manifold under the mean-Lindelöf hypothesis_ , Forum Math. 13 (2001), no. 6, 781–793.
* [Kuz78] N. V. Kuznetsov, _The arithmetic form of Selberg’s trace formula and the distribution of norms of the primitive hyperbolic classes of the modular group_ , Preprint (Khabarovsk), 1978.
* [LRS95] W. Luo, Z. Rudnick, and P. Sarnak, _On Selberg’s eigenvalue conjecture_ , Geom. Funct. Anal. 5 (1995), no. 2, 387–401.
* [LS95] W. Luo and P. Sarnak, _Quantum ergodicity of eigenfunctions on $\mathrm{PSL}_{2}(\mathbf{Z})\backslash\mathbf{H}^{2}$_, Publ. Math. Inst. Hautes Études Sci. 81 (1995), 207–237.
* [PR17] Y. N. Petridis and M. S. Risager, _Local average in hyperbolic lattice point counting, with an appendix by Niko Laaksonen_ , Math. Z. 285 (2017), no. 3–4, 1319–1344.
* [Sar80] P. Sarnak, _Prime geodesic theorems_ , Ph.D. thesis, Stanford University, August 1980.
* [Sar82] by same author, _Class numbers of indefinite binary quadratic forms_ , J. Number Theory 15 (1982), no. 2, 229–247.
* [Sel89] A. Selberg, _Collected papers_ , Springer Collected Works in Mathematics, vol. 1, Springer-Verlag Berlin Heidelberg, 1989.
* [SY13] K. Soundararajan and M. P. Young, _The prime geodesic theorem_ , J. Reine Angew. Math. 676 (2013), 105–120, Errata: https://www.math.tamu.edu/%7Emyoung/PrimeGeodesicCorrections.pdf.
* [Ven90] A. B. Venkov, _Spectral theory of automorphic functions and its applications_ , Mathematics and its Applications (Soviet Series), vol. 51, Kluwer Academic Publishers Group, 1990, Translated from the Russian by N. B. Lebedinskaya.
|
††thanks<EMAIL_ADDRESS>
# Differentiable Predictions for Large Scale Structure with SHAMNet
Andrew P. Hearin1,⋆ Nesar Ramachandra2,1 Matthew R. Becker1 Joseph DeRose3
1High Energy Physics Division, Argonne National Laboratory, 9700 South Cass
Avenue, Lemont, IL 60439, USA 2Computational Science Division, Argonne
National Laboratory, 9700 South Cass Avenue, Lemont, IL 60439, USA 3Lawrence
Berkeley National Laboratory, 1 Cyclotron Road, Berkeley, CA, 94720 USA
###### Abstract
In simulation-based models of the galaxy–halo connection, theoretical
predictions for galaxy clustering and lensing are typically made based on
Monte Carlo realizations of a mock universe. In this paper, we use Subhalo
Abundance Matching (SHAM) as a toy model to introduce an alternative to
stochastic predictions based on mock population, demonstrating how to make
simulation-based predictions for clustering and lensing that are both exact
and differentiable with respect to the parameters of the model. Conventional
implementations of SHAM are based on iterative algorithms such as Richardson-
Lucy deconvolution; here we use the JAX library for automatic differentiation
to train SHAMNet, a neural network that accurately approximates the stellar-
to-halo mass relation (SMHM) defined by abundance matching. In our approach to
making differentiable predictions of large scale structure, we map
parameterized PDFs onto each simulated halo, and calculate gradients of
summary statistics of the galaxy distribution by using autodiff to propagate
the gradients of the SMHM through the statistical estimators used to measure
one- and two-point functions. Our techniques are quite general, and we
conclude with an overview of how they can be applied in tandem with more
complex, higher-dimensional models, creating the capability to make
differentiable predictions for the multi-wavelength universe of galaxies.
###### keywords:
Cosmology: large-scale structure of Universe; methods: N-body simulations
## 1 Introduction
Numerical simulations of cosmological structure formation can be viewed as
prediction engines for the density field, and for the abundance and spatial
distribution of dark matter halos. Of course, neither dark matter nor
gravitationally self-bound halos are directly observed in the sky, and so
additional modeling is required in order to transform a cosmological
simulation into a prediction that is commensurable with astronomical
measurements of galaxies. Contemporary simulations have by now achieved
percent-level precision in the ability to characterize the density field, halo
abundance and halo clustering in the nonlinear regime, whereas models
connecting observed galaxies to the fundamental quantities in simulations lag
far behind this level of accuracy (see Wechsler & Tinker, 2018, for a recent
review). Improving theoretical techniques for transforming cosmological
simulations into predictions for the galaxy density field is thus a critical
component of the precision-cosmology program.
In many of the structure formation models that are used for cosmological
inference, predictions for the galaxy distribution are made by modeling
galaxies as biased tracers of the underlying density field of matter. There
are a wide range of techniques that can be used for this purpose, including
methods based on Lagrangian perturbation theory (LPT, Bernardeau et al., 2002;
Desjacques et al., 2018), effective field theory (Carrasco et al., 2012), the
Zel’dovich approximation (White, 2014), and hybrid techniques that blend LPT
with simulations (Kokron et al., 2021; DeRose et al., 2021b). This approach to
generating cosmological predictions has now been used by numerous galaxy
surveys to derive constraints on the fundamental parameters of cosmology
(e.g., DES Collaboration et al., 2021; Joudaki et al., 2018).
An alternative approach is to model the connection between dark matter halos
and the galaxies residing within them. The conventional “halo occupation
model” approach proceeds with an initial step in which a fitting function
and/or a machine learning algorithm is calibrated to capture the cosmology-
dependence of various summary statistics of dark matter halos, such as the
halo mass function (e.g., Jenkins et al., 2001; McClintock et al., 2019a;
Bocquet et al., 2020) and halo bias (e.g, Tinker et al., 2010; McClintock et
al., 2019b); once the approximations for these quantities are specified,
models for the galaxy–halo connection determine the prediction for the
clustering and lensing of galaxies, enabling the derivation of constraints on
cosmological parameters (Cacciato et al., 2013; Reddick et al., 2014; Miyatake
et al., 2021).
As an alternative to traditional implementations of halo occupation modeling,
it has become increasingly common to directly populate simulated halos with a
Monte Carlo realization of the galaxy population; predictions for summary
statistics of large scale structure are then made directly from the synthetic
galaxy distribution using the same point estimators used to analyze
observational data. In the conventional approach to halo occupation modeling,
phenomena such as halo exclusion (García et al., 2021), satellite anisotropy
(Sgró et al., 2013), and galaxy assembly bias (Zentner et al., 2014)
constitute a major technical challenge, particularly in the one-to-two-halo
regime (van den Bosch et al., 2013); one of the key advantages of the mock-
population approach is that the numerical treatment of such effects is exact,
with a precision limited only by the resolution and finite size of the
simulation.
One of the principal sources of motivation for using mock-population
techniques lies in the cosmological information content of the nonlinear
regime. The potential to substantially improve cosmological constraints by
incorporating smaller-scale information has been known for many years (Zentner
et al., 2013; Reid et al., 2014; Krause & Eifler, 2017). Recent studies of the
clustering of galaxies in the Baryon Oscillation Spectroscopic Survey (BOSS)
have confirmed the long-forecasted constraining power of the nonlinear regime.
In an analysis of the redshift-space clustering of the LOWZ galaxy sample
(Lange et al., 2021), the authors derived better-than 5% constraints on the
cosmological growth of structure, $f\sigma_{8},$ a full factor of two stronger
than any previous BOSS analysis that restricted attention to larger-scale
measurements. Comparable gains in cosmological constraining power from the
nonlinear regime of BOSS LOWZ galaxies were also found in Wibking et al.
(2020). In closely related work analyzing the clustering of the Luminous Red
Galaxy sample in eBOSS (Chapman et al., 2021), it was found that constraints
on $f\sigma_{8}$ are improved by 70% when including information from nonlinear
scales.
Even when simulation-based predictions are made based on simple empirical
models implemented in specialized libraries such as Halotools (Hearin et al.,
2017), Corrfunc (Sinha & Garrison, 2020), or AbacusHOD (Yuan et al., 2021),
the computational demands of conducting Bayesian inference with survey-scale
simulations are considerable. This challenge is tailor-made for a set of
methods generally referred to as “emulation”, which is a machine learning
technique that is specifically designed for situations in which the behavior
of some parametric function, $\mathcal{P}(\theta),$ is calculable, but
expensive to evaluate. When confronted with such a situation, a natural way to
proceed is to pre-compute $\mathcal{P}(\theta)$ for a finite collection of
$\theta_{\rm i}$ that spans the domain of interest, and to use
$\mathcal{P}(\theta_{\rm i})$ to train a machine learning algorithm to serve
as an “emulator”, or surrogate function, $\mathcal{F}(\theta),$ that
approximates $\mathcal{P}(\theta).$ Once the emulator has been trained, one
then proceeds to carry out performance-critical analyses such as MCMCs using
$\mathcal{F}(\theta)$ rather than $\mathcal{P}(\theta),$ since the surrogate
function is typically inexpensive to evaluate. Since the introduction of
emulation methods to computational cosmology (Heitmann et al., 2006), these
techniques have become a widely-used tool throughout the field (e.g., Kwan et
al., 2015; Nishimichi et al., 2019; Euclid Collaboration et al., 2020).
The conventional approach to emulation is based on Gaussian Process (GP)
regression. Briefly, in GP regression, it is assumed that the values
$\mathcal{P}(\theta_{\rm i})$ are drawn from a multi-dimensional Gaussian
distribution, $\mathcal{N}(\mathcal{P}(\theta_{\rm i}),\Sigma(\theta_{\rm
i})),$ where $\Sigma(\theta_{\rm i})$ is a correlation matrix with hyper-
parameters that are optimized during training (see Rasmussen & Williams, 2006,
for a contemporary review). For example, in Lawrence et al. (2017), the
abscissa $\theta_{\rm i}$ are cosmological parameters, and
$\mathcal{P}(\theta_{\rm i})$ are values of the matter power spectrum; as
another example, in Zhai et al. (2019), the abscissa reside in the joint space
of cosmology and parameters of the Halo Occupation Distribution (HOD, Berlind
& Weinberg, 2002; Zheng et al., 2005), and the ordinates are values of the
redshift-space galaxy correlation function.
Although the GP approach to emulation has thus far been quite successful, this
technique has limitations that create a significant impediment to applying it
in tandem with the realistically complex models that will be required by near-
future cosmological datasets. In Zhai et al. (2019), the authors found that
the errors of their emulator were comparable to present-day measurement
uncertainties on their predicted data vector, resulting in a nearly $50\%$
degradation of the cosmological constraints due to emulator noise. In
principle, these inaccuracies could be remedied by increasing the number of
training points $\theta_{\rm i};$ in practice, however, the size of the
training data cannot be too large or the hyper-parameter optimization becomes
computationally intractable due to the need to invert $\Sigma(\theta_{\rm
i}).$ We point out that the effort in Zhai et al. (2019) is one of the most
ambitious GP applications in cosmology to date in the sense that the emulated
parameter space includes parameters that jointly encode variations in
cosmology as well as the galaxy–halo connection. And yet, the HOD model
emulated in Zhai et al. (2019) is essentially the simplest, lowest-dimensional
model that can plausibly be used to interpret galaxy clustering data with
present-day levels of uncertainty. Thus these shortcomings will only become
more severe as observational cosmology progresses further into the 2020s: not
only will the precision of cosmological measurements improve dramatically, but
perhaps even more importantly, the dimension of the emulated model will need
to increase substantially in order for the theoretical predictions to match
the quality and richness of the data.
In the present work, we introduce a new theoretical framework for the
galaxy–halo connection that is designed to address these issues. Our formalism
differs from the typical simulation-based methodology outlined above in
several respects. First, our simulation-based predictions for large scale
structure observables are not based on stochastic Monte Carlo realizations;
instead, we analytically propagate probability distributions from the
galaxy–halo connection through to the corresponding summary statistics, and so
the only source of stochasticity in our predictions derives from the finite
size and resolution of the simulation. Second, instead of relying upon
classical machine learning techniques such as Gaussian Process emulation, we
instead rely on Artificial Intelligence (AI) algorithms for our surrogate
functions. Since the neural networks implemented in contemporary deep learning
libraries are routinely used in industry applications to approximate the
behavior of million-parameter systems, our use of AI-based surrogate functions
improves upon the problems associated with emulator accuracy and complexity
outlined above. As a result of these techniques, our predictions for large
scale structure observables are fully differentiable, end-to-end, enabling us
to use gradient-based optimization and inference algorithms that exhibit much
better scaling to the problem sizes that will characterize cosmological
modeling in the 2020s.
In the present paper, for our model of the galaxy–halo connection, we use
subhalo abundance matching (SHAM, Kravtsov et al., 2004; Conroy et al., 2006)
as a toy model to demonstrate our framework; in §2, we provide a self-
contained overview of SHAM. We train a neural network, SHAMNet, that serves as
a surrogate function approximating the SHAM mapping between stellar mass and
halo mass; we describe our implementation of SHAMNet in §3, relegating
technical details to the appendices. Although any suitably formulated neural
network is naturally differentiable, this is not the case for typical
pipelines that make downstream predictions for the $n-$point functions of
large scale structure; in §4, we describe our framework for constructing
differentiable point estimators of the stellar mass function, galaxy-galaxy
lensing, and galaxy clustering. We discuss our results in the broader context
of related efforts in the literature in §5, and we conclude by summarizing our
primary findings in §6.
Throughout the paper, values of halo mass and distance are quoted assuming
$h=1.$ For example, when writing $M_{\rm peak}=10^{12}M_{\odot},$ we suppress
the $M_{\odot}/h$ notation and write the units as $M_{\odot}.$ Values of
stellar mass are quoted assuming $h=0.7.$
## 2 Abundance Matching Primer
The $M_{\star}-M_{\rm halo}$ relation in traditional abundance matching is the
unique, non-parametric mapping defined by equating the abundance of galaxies
to the abundance of halos at the same redshift,
$\displaystyle\Phi_{\rm g}(>M_{\star}|z)=\Phi_{\rm h}(>M_{\rm halo}|z),$ (1)
where $\Phi_{\rm g}(>M_{\star})$ is the cumulative number density of galaxies,
and $\Phi_{\rm h}(>M_{\rm halo})$ is the cumulative number density of
subhalos. Contemporary forms of SHAM generalize Eq.1 to define non-parametric
mappings between some subhalo property, $x_{\rm h},$ and some observed galaxy
property, $y_{\rm g}.$ Application of the defining abundance matching equation
guarantees that no matter the choice for $x_{\rm h}$ and $y_{\rm g},$ the
observed and predicted number density of galaxies will be in exact
correspondence, by construction. We restrict the present investigation to
$x_{\rm h}=M_{\rm peak}$ and $y_{\rm g}=M_{\star},$ and refer the reader to §5
for discussion of how our methodology could be extended to generalized
galaxy/halo properties.
Even in the early literature on abundance matching (e.g., Tasitsiomi et al.,
2004), the importance of stochasticity in the $M_{\star}-M_{\rm halo}$
relation was recognized to play an important role in the predictions of the
model. Stochasticity in abundance matching is typically treated by assuming a
model for the probability density $P(M_{\star}|M_{\rm peak})$ such as a log-
normal, and then equating the number density of galaxies to the convolution of
the subhalo mass function against the assumed PDF :
$\displaystyle\phi_{\rm g}(M_{\star})=\int_{0}^{\infty}dM_{\rm peak}\phi_{\rm
h}(M_{\rm peak})P(M_{\star}|M_{\rm peak}),$ (2)
where $\phi_{\rm g}(M_{\star})$ is the differential number density galaxies as
a function of stellar mass, i.e. the stellar mass function (SMF), and
$\phi_{\rm h}(M_{\rm peak})$ is the differential number density of subhalos as
a function of mass, i.e. the subhalo mass function (SHMF).
In general, the $M_{\star}-M_{\rm peak}$ relation defined by Equation 2 has no
closed-form analytic solution, and so one must rely on approximations and
numerical techniques to determine the scaling relation $\langle
M_{\star}|M_{\rm peak}\rangle_{\rm med}$ that gives rise to the observed SMF
when applied to the SHMF of a subhalo population. In §2.1, we discuss the
conventional, non-parametric approach to determining the $M_{\star}-M_{\rm
peak}$ relation predicted by SHAM, and in §2.2 we describe the commonly-used
alternative approach based on a parameterized approximation to SHAM, in both
cases highlighting the close relationship between the SMF, the median relation
$\langle M_{\star}|M_{\rm peak}\rangle_{\rm med},$ and scatter in stellar mass
at fixed halo mass.
### 2.1 Non-parametric SHAM with scatter
In non-parametric approaches to SHAM, the starting point is typically a
catalog of simulated subhalos that defines the SHMF, and a volume-limited
galaxy sample that defines the SMF. A particularly simple way to measure the
cumulative abundance of galaxies or halos is simply to divide each object’s
rank-order by the volume containing the sample, although one may instead use a
fitting function approximation to one or both abundance functions. However
$\phi_{\rm g}$ and $\phi_{\rm h}$ are characterized, in non-parametric SHAM
the SMF and SHMF are held fixed, and one numerically solves for the quantity
$P(M_{\star}|M_{\rm peak})$ that is constrained by Eq. 2. Thus when generating
a synthetic galaxy population with non-parametric SHAM, the stellar mass
function of the mock is guaranteed to match the observed SMF, by construction,
with an accuracy limited only by the convergence and robustness of the
computational technique used to numerically solve Eq. 2. In §2.1.1 we describe
two commonly used approaches to obtaining such a solution, and in §2.1.2 we
highlight the basic features of the stellar-to-halo mass relation derived from
the solution.
#### 2.1.1 Numerical Methods
The most widely used numerical approach to solving Eq. 2 is based on
Richardson-Lucy deconvolution (Richardson, 1972; Lucy, 1974), an iterative
algorithm originally developed to recover a true underlying image that has
been blurred by a point-spread function (PSF). In the context of abundance
matching, we can think of the “noisy image” as $\phi_{\rm g}(M_{\star}),$
i.e., the SMF that we measure in our observed galaxy sample; the PSF blurring
the image is typically assumed to be a log-normal distribution in stellar mass
at fixed halo mass; the RL deconvolution algorithm determines the “true
image”, $\phi^{\rm true}_{\rm g}(M_{\star}),$ which in this case is the SMF
that one would measure in the absence of any scatter in the $M_{\star}-M_{\rm
halo}$ relation. Once $\phi^{\rm true}_{\rm g}(M_{\star})$ is determined, then
Eq. 1 is used to define $\langle M_{\star}|M_{\rm peak}\rangle_{\rm med};$
together with the assumed level of log-normal scatter, this scaling relation
then defines the quantity $P(M_{\star}|M_{\rm peak})$ that is used to map
stellar mass onto a simulated subhalo population. In the conventional Monte
Carlo based implementation, a synthetic galaxy population is generated by
randomly drawing a value of $M_{\star}$ from $P(M_{\star}|M_{\rm peak})$ for
each subhalo; by construction, the resulting SMF of the synthetic population
will agree with the SMF of the observed galaxy sample, $\phi_{\rm
g}(M_{\star}),$ again with accuracy limited only by the RL deconvoluation-
based estimation of $\phi^{\rm true}_{\rm g}(M_{\star})$ (see Behroozi et al.,
2010, Section 3, for further details).
An alternative approach presented in Kravtsov et al. (2018) is instead based
on the assumption that the $M_{\star}-M_{\rm peak}$ relation can be locally
approximated as a power law. Under this assumption, for any particular value
of $M_{\rm peak}$ and level of scatter, it is straightforward to solve for the
power-law normalization and index that produces the observed $\phi_{\rm
g}(M_{\star})$ from the simulated $\phi_{\rm h}(M_{\rm peak}).$ By repeating
this exercise at a finite set of control points in $M_{\rm peak}$ that densely
spans the relevant range, one can use the results of the computation as an
interpolation table that defines the $\langle M_{\star}|M_{\rm
peak}\rangle_{\rm med}$ scaling relation (see Kravtsov et al., 2018, Appendix
A, for further details).
#### 2.1.2 Stellar-to-Halo Mass Relation of Non-Parameteric SHAM
Figure 1 gives a simple demonstration of the role of scatter in non-parametric
abundance matching. Each curve shows the abundance matching prediction for
$\langle M_{\star}|M_{\rm peak}\rangle_{\rm med}$ using the same stellar mass
function, and the same subhalo mass function, but with different levels of
scatter as indicated in the legend. For our SMF, $\phi_{\rm fid},$ we use a
Schechter function with best-fit parameters taken from Panter et al. (2007),
and for the subhalo mass function we use the fitting function presented in
Appendix A. We calculate $\langle M_{\star}|M_{\rm peak}\rangle_{\rm med}$
using a publicly available python
wrapper111https://bitbucket.org/yymao/abundancematching/src of the RL
deconvolution implementation originally developed in Behroozi et al. (2010).
All curves in Figure 1 present the same qualitative shape of a double power-
law, with stronger levels of scatter producing shallower power-law relations
at the high-mass end. We can understand this characteristic flattening in
terms of Eddington bias (Eddington, 1913): stochasticity in the
$M_{\star}-M_{\rm peak}$ relation will naturally result in proportionally more
up-scatter above a threshold relative to down-scatter below the threshold,
simply because low-mass halos are more abundant than high-mass halos. Thus it
is sensible that we see stronger effects of scatter at high mass, where the
slope of the mass function is rapidly steepening. We return to this issue in
the subsequent section on parametric approximations to abundance matching.
Figure 1: Role of scatter in non-parametric SHAM. Each curve shows the
$M_{\star}-M_{\rm halo}$ relation of a SHAM model that predicts the same
stellar mass function, $\phi_{\rm fid}(M_{\star}),$ but with different levels
of scatter in $M_{\star}$ at fixed halo mass, as indicated in the legend. For
each model, the median relation $\langle M_{\star}|M_{\rm peak}\rangle_{\rm
med}$ is plotted on the vertical axis as a function of halo mass, $M_{\rm
peak}.$ Figure 2: Role of scatter in parameterized abundance matching. Each
curve shows the stellar mass function predicted by a parameterized abundance
matching model with the same $\langle M_{\star}|M_{\rm peak}\rangle_{\rm
med},$ but with different levels of scatter in $M_{\star}$ at fixed halo mass.
For each model, the number density of galaxies, $\phi(M_{\star}),$ is plotted
on the vertical axis as a function of stellar mass.
### 2.2 Parametric Approximations to SHAM with scatter
In studies of the galaxy-halo connection connecting stellar mass to halo mass,
numerous models have directly parameterized the scaling relation $\langle
M_{\star}|M_{\rm peak}\rangle_{\rm med},$ rather than using numerical methods
of solution to Eq. 2 (e.g., Moster et al., 2010, 2013; Behroozi et al., 2013a;
Rodríguez-Puebla et al., 2015). In such models, the assumed functional forms
for $\langle M_{\star}|M_{\rm peak}\rangle_{\rm med}$ all have the same
general double power-law shape shown in Figure 1; the free parameters of these
models regulate the low- and high-mass slopes, as well as the normalization
and shape of the transition between the two regimes.
It has been shown explicitly that the $\langle M_{\star}|M_{\rm
peak}\rangle_{\rm med}$ predicted via RL deconvolution can indeed be
accurately approximated with the functional forms used in Behroozi et al.
(2010) and related studies, and so it is common for such parametric models to
be informally referred to as “abundance matching.” However, we point out a
clear distinction in the way these two models utilize information from
observations: in parameterized models of $\langle M_{\star}|M_{\rm
peak}\rangle_{\rm med},$ there is no guarantee that the observed stellar mass
function will be correctly predicted, as is the case with non-parametric
abundance matching. Instead, in parameterized models, observations of the SMF
are treated as constraining data that can be used to derive confidence
intervals on the parameters controlling the stellar-to-halo mass relation.
Thus parametric vs. non-parametric SHAM models use information from
observations in quite a different manner when evaluating the likelihood of a
proposed point in parameter space.
Figure 2 illustrates the role of scatter in predictions for the stellar mass
function made by parameterized abundance matching. The axes show the number
density of galaxies as a function of stellar mass, $\phi(M_{\star}).$ Each
curve shows results for models with different levels of scatter about the same
$M_{\star}-M_{\rm halo}$ relation. Again we see the proportionally larger
influence of Eddington bias at higher mass, due to up-scatter from lower-mass
halos that outnumber higher-mass halos.
## 3 SHAMNet
In this section, we describe a new approach to abundance matching based on
SHAMNet: a neural network approximation to the numerical solution to Eq. 2. We
begin in §3.1 with a high-level overview of how SHAMNet is defined. In §3.2 we
describe the key ingredients we use in our implementation, and in §3.3 we
describe how we have used these ingredients to train SHAMNet to approximate
the abundance matching prediction for the SMHM scaling relation, $\langle
M_{\star}|M_{\rm peak}\rangle_{\rm med}.$ We give a detailed account of the
architecture and training of SHAMNet in the appendices; our source code is
publicly available at https://github.com/ArgonneCPAC/shamnet, and is available
for installation with pip.
### 3.1 SHAMNet Overview
As discussed in §2, the SHAM technique defines the probability distribution
$P(M_{\star}|M_{\rm peak})$ such that when this PDF is convolved against the
subhalo mass function, $\phi_{\rm h}(M_{\rm peak}),$ the galaxy stellar mass
function, $\phi_{\rm g}(M_{\star}),$ is recovered. Under the assumption of
log-normal scatter, this PDF is fully described by its first and second
moments, $\langle M_{\star}|M_{\rm peak}\rangle_{\rm med}$ and $\sigma(M_{\rm
peak}),$ respectively. Thus the goal of any SHAM implementation is to accept
ingredients for $\phi_{\rm g}(M_{\star}),$ $\phi_{\rm h}(M_{\rm peak}),$ and
$\sigma(M_{\rm peak})$ as inputs, and to return the scaling relation $\langle
M_{\star}|M_{\rm peak}\rangle_{\rm med}$ that provides a solution to Eq. 2.
SHAMNet is simply a neural network that provides a mapping from these
ingredients to the desired stellar-to-halo mass relation.
In building SHAMNet, we assume that the observed SMF can be characterized with
sufficient precision using some parametrized functional form, $\phi_{\rm
g}(M_{\star}|\theta_{1});$ we similarly assume that the subhalo mass function
can be parametrically described as $\phi_{\rm h}(M_{\rm peak}|\theta_{2});$
finally, we assume log-normal scatter in stellar mass at fixed halo mass,
allowing the level of scatter, $\sigma,$ to be a parametrized function of halo
mass, $\sigma(M_{\rm peak}|\theta_{3}).$ For notational convenience, we will
use the generic variable $\theta$ to refer to the collection of these
parameters, so that $\theta$ fully specifies $\phi_{\rm g}(M_{\star}),$
$\phi_{\rm h}(M_{\rm peak}),$ and $\sigma(M_{\rm peak}).$ SHAMNet is defined
to be a neural network that accepts $\theta$ and $M_{\rm peak}$ as input, and
returns $M_{\star}$ as output; the network parameters of SHAMNet are trained
so that Eq. 2 is satisfied by the mapping.
### 3.2 SHAMNet ingredients
Our formulation of SHAMNet requires a parametric description of $\phi_{\rm
g}(M_{\star}),$ $\phi_{\rm h}(M_{\rm peak}),$ and $\sigma(M_{\rm peak}).$ In
this section, we describe our models for each of these ingredients in turn.
Our parameterization of $\phi_{\rm g},$ the differential number density of
galaxies, is based on the Schechter function, $\phi_{\rm S},$ defined as
(3) $\displaystyle\phi_{\rm S}(x|x_{\ast},\phi_{\ast},\alpha)={\rm
ln}(10)\phi_{\ast}10^{(x-x_{\ast})\cdot(\alpha+1)}\exp(-10^{x-x_{\ast}}),$
where $x=\log_{10}M_{\star}$ in units of $M_{\odot}.$ Our galaxy SMF is thus
characterized by $x_{\ast},\phi_{\ast},$ and $\alpha,$ and we define SHAMNet
with a fixed value $x_{\ast}=10.85,$ closely mimicking the SMF in the low-
redshift universe (Li & White, 2009).
We parameterize the subhalo mass function in terms of the cumulative number
density of subhalos, $\Phi_{\rm h}(>M_{\rm peak}|z),$ using the fitting
function described in Appendix A. Briefly, $\Phi_{\rm h}(>M_{\rm peak}|z)$
behaves like a power-law at low mass, with a normalization $A_{\rm h},$ and an
index $\beta_{\rm h},$ with an exponential cutoff the high-mass end
characterized by a cutoff mass, $x_{\rm h},$ and cutoff speed, $k_{\rm h}.$ In
defining SHAMNet, we hold the subhalo mass function parameters fixed to the
values supplied in Appendix A that have been tuned to closely match the SHMF
in the Bolshoi-Planck simulation (BPL, Klypin et al., 2011).
We model the scatter in $M_{\star}$ at fixed $M_{\rm peak}$ as a log-normal
distribution with a width that is allowed to vary as a function of halo mass,
$\sigma(M_{\rm peak}).$ For the mass-dependence of the scatter, we use the
sigmoid function given in 21, with the parameters $y^{\sigma}_{\rm lo}$ and
$y^{\sigma}_{\rm hi}$ specifying the level of scatter at the low- and high-
mass end, respectively, and with $k^{\sigma}=1$ and $x^{\sigma}_{0}=12$ held
fixed:
$\displaystyle\sigma(x|y_{\rm lo}^{\sigma},y_{\rm
hi}^{\sigma})=\mathcal{S}(x,12,1,y_{\rm lo}^{\sigma},y_{\rm hi}^{\sigma}),$
(4)
where $x=\log_{10}M_{\rm peak}.$
### 3.3 SHAMNet Definition
With the ingredients defined in §3.2, the parameters $\theta$ collectively
describe our model for the SMF, $\phi_{\rm g}(M_{\star}|\theta),$ the subhalo
mass function, $\phi_{\rm h}(M_{\rm peak}|\theta),$ and the level of log-
normal scatter, $\sigma(M_{\rm peak}|\theta).$ For any particular value of
$\theta,$ we seek to identify the stellar-to-halo mass relation, $\langle
M_{\star}|M_{\rm peak};\theta\rangle_{\rm med},$ that provides the self-
consistent relationship between these quantities (as defined by Eq. 2). We
define SHAMNet, $\mathcal{F}_{\rm SHAM},$ to be a neural network that has been
trained to supply the appropriate scaling relation:
$\displaystyle\mathcal{F}_{\rm SHAM}(M_{\rm peak},\theta|\psi)\equiv\langle
M_{\star}|M_{\rm peak};\theta\rangle_{\rm med},$ (5)
where the variables $\psi$ are the weights and biases of the neural network.
Thus for any particular values of $\psi,$ the SHAMNet function
$\mathcal{F}_{\rm SHAM}$ accepts $M_{\rm peak}$ and $\theta$ as input, and
returns $M_{\star}$ as output.
The objective of training SHAMNet is to optimize the weights and biases $\psi$
such that the resulting mapping in Eq. 5 supplies a stellar-to-halo mass
relation that satisfies Eq. 2 to the required precision. Once trained, the
parameters $\psi$ are thereafter held fixed, and $\mathcal{F}_{\rm
SHAM}(M_{\rm peak},\theta)=M_{\star}$ defines the stellar-to-halo mass
relation for any input combination of SMF, subhalo mass function, and scatter.
As shown in Figure 3, accurately training SHAMNet does not present a
significant challenge for a simply-connected network with only a few short
layers. On the vertical axis, we show the logarithmic difference between
target SMF defined by $\theta,$ and the SMF that results from the SHAMNet
prediction for $\langle M_{\star}|M_{\rm peak};\theta\rangle_{\rm med}.$ The
level of success shown in Fig. 3 should not be surprising, since neural
networks are routinely called upon to approximate far more complex functions
than abundance matching. We refer the reader to Appendix D for detailed
information on how we trained SHAMNet.
Figure 3: Accuracy of SHAMNet. The vertical axis shows the logarithmic
difference between the target stellar mass function and the SMF that results
from the SHAMNet prediction for the stellar-to-halo-mass relation. The red
band shows the variance in the residual error amongst 100 randomly selected
points in the SMF parameter space; the blue curves show the particular
residuals for 10 of these points.
Beyond its accuracy and convenience, the function $\mathcal{F}_{\rm
SHAM}(M_{\rm peak},\theta)$ is an analytically differentiable function of the
parameters $\theta,$ because the behavior of the function is simply the
composition of a chain of $C^{\infty}$ functions. As described in Appendix D,
our implementation of SHAMNet in the JAX library makes it straightforward and
efficient to compute these gradients via automatic differentiation. In §4
below, we describe how these SHAMNet gradients can be propagated to permit
calculation of derivatives of large scale structure observables such as two-
point clustering and lensing.
## 4 Differentiable Predictions for Large Scale Structure
In §3, we described how SHAMNet, $\mathcal{F}_{\rm SHAM},$ serves as a
surrogate function that approximates the stellar-to-halo mass relation of
abundance matching, $\mathcal{F}_{\rm SHAM}(M_{\rm peak}|\theta)\equiv\langle
M_{\star}|M_{\rm peak};\theta\rangle_{\rm med}.$ We remind the reader that the
parameters $\theta$ control the behavior of the galaxy stellar mass function,
$\phi_{\rm g}(M_{\star}|\theta),$ the subhalo mass function, $\phi_{\rm
h}(M_{\rm peak}|\theta),$ and halo mass-dependent scatter, $\sigma(M_{\rm
peak}|\theta).$ For the remainder of the paper, our principal focus will be on
leveraging the differentiability of $\mathcal{F}_{\rm SHAM}(M_{\rm
peak}|\theta)$ with respect to the parameters $\theta.$ Physically, the
gradients $\partial\mathcal{F}_{\rm SHAM}/\partial\theta$ encode how the
stellar-to-halo mass relation changes in response to changes in $\phi_{\rm
g}(M_{\star}),$ $\phi_{\rm h}(M_{\rm peak}),$ and $\sigma(M_{\rm peak}).$
In conventional SHAM implementations, the scaling relation $\langle
M_{\star}|M_{\rm peak};\theta\rangle_{\rm med}$ is computed numerically via an
iterative algorithm such as RL deconvolution, and so one must rely on finite
differencing methods to estimate the gradients of the scaling relation with
respect to the parameters $\theta.$ By contrast, our JAX implementation of
$\mathcal{F}_{\rm SHAM}$ allows us to calculate $\partial\langle
M_{\star}|M_{\rm peak};\theta\rangle_{\rm med}/\partial\theta$ to machine
precision with high efficiency. In this section, we describe a set of
techniques that will allow us to leverage the availability of these gradients
to directly differentiate simulation-based forward-modeling predictions for
large scale structure observables.
### 4.1 Stellar Mass Function
Simulation-based predictions for large scale structure observables are
typically based on Monte Carlo realizations of a model applied to a (sub)halo
catalog. Using such Monte Carlo methods, the SHAM prediction for the stellar
mass function proceeds in two steps:
1. 1.
Assign a value $M_{\star}$ to every subhalo in the catalog by randomly drawing
from a log-normal PDF centered at $\langle M_{\star}|M_{\rm peak}\rangle_{\rm
med}$ with scatter $\sigma.$
2. 2.
For the $i^{\rm th}$ bin of the SMF, $\phi_{\rm g}(M_{\star}^{\rm i}),$ sum
the number of synthetic galaxies with $M_{\star}^{\rm
i}<=M_{\star}<M_{\star}^{\rm i+1},$ and divide by the normalization factor
appropriate for the population volume and bin size, $V_{\rm i}.$
In order to calculate gradients of the predicted SMF with respect to model
parameters $\theta,$ the above procedure is repeated for parameter values
$\theta$ that have been perturbed about some fiducial value, and the gradient
is estimated via finite-differencing methods, e.g., $\Delta\phi_{\rm
g}(M_{\star})/\Delta\theta.$
Here we consider an alternative approach to predicting $\phi_{\rm
g}(M_{\star})$ that does not rely on a Monte Carlo realization. For the
$\alpha^{\rm th}$ subhalo in the catalog with halo mass $M_{\rm
peak}^{\alpha},$ we can analytically calculate the probability that the galaxy
mapped onto the subhalo will fall within the ${\rm i^{th}}$ bin:
$\displaystyle w_{\rm i}^{\alpha}$ $\displaystyle=$
$\displaystyle\int_{M_{\star}^{\rm i}}^{M_{\star}^{\rm i+1}}{\rm
dM_{\star}}P(M_{\star}|M_{\rm peak}^{\alpha}).$ (6)
Once all the weights $w_{\rm i}^{\alpha}$ have been calculated for each
subhalo in the catalog, then predicting the SMF is simply a matter of summing
the weights in each bin and normalizing the result:
$\displaystyle\phi_{\rm g}(M_{\star}^{\rm i})$ $\displaystyle=$
$\displaystyle\sum_{\alpha}w_{\rm i}^{\alpha}/V_{\rm i}.$ (7)
Figure 4: Gradients of simulation-based SMF predictions. In the top panel we
show three different SMFs, with different values of $\alpha$ as indicated in
the legend. In the bottom panel we show the logarithmic derivative of the SMF
with respect to $\alpha,$ calculating the result with both finite-differencing
methods and automatic differentiation.
Operationally, the only difference between these two approaches is that we
compute the integrals in Eq. 6 analytically rather than via Monte Carlo
integration. While Monte Carlo-based predictions will be statistically
commensurable with the SMF computed via Eq. 7, the analytical approach can be
more computationally efficient because numerical convergence may require a
large number of realizations, particularly at the high-mass end due to the
rarity of massive galaxies. Beyond this computational advantage, from
Equations 6-7 we can readily see that the gradient $\partial\phi_{\rm
g}(M_{\star})/\partial\theta$ can be calculated exactly, since it is simply a
sum of $C^{\infty}$ functions:
$\displaystyle\frac{\partial}{\partial\theta}\phi_{\rm g}(M_{\star}^{\rm
i})=\frac{1}{V_{\rm i}}\sum_{\alpha}\frac{\partial w_{\rm
i}^{\alpha}}{\partial\theta}$ (8)
In Figure 4, we show how the stellar mass function responds to changes with
respect to $\alpha,$ the parameter controlling the power-law slope of the
Schechter function defined in Eq. 3. In the top panel we show three different
SMFs, with different values of $\alpha$ as indicated in the legend. To compute
$\phi_{\rm g}(M_{\star}),$ we used Eq. 7 to calculate the weight attached to
every subhalo of the BPL simulation at $z=0,$ using 50 logarithmically-spaced
bins spanning $10^{9}M_{\odot}<M_{\star}<10^{11.5}M_{\odot}.$ In the bottom
panel of Figure 4, we show the logarithmic derivative of the SMF with respect
to $\alpha.$ For the curve labeled “finite differencing”, we calculated the
gradient numerically by repeatedly calculating the result of Eq. 7 for low and
high values of $\alpha.$ For the curve labeled “autodiff”, we use JAX to
calculate $\partial w_{\alpha}/\partial\theta$ for every subhalo, and then we
use numba to propagate these derivatives through to the computation of the SMF
according to Eq. 8.
### 4.2 Galaxy Lensing
In this section, we adapt the methods described in §4.1 to make differentiable
predictions for galaxy lensing, $\Delta\Sigma(R),$ the excess surface mass
density at a projected distance $R$ from the center of a stacked sample of
galaxies, defined as
$\displaystyle\Delta\Sigma(R)\equiv\bar{\Sigma}(<R)-\Sigma(R).$ (9)
In Eq. 9, $\Sigma(R)$ is the surface mass density projected along the line-of-
sight to the stack, and $\bar{\Sigma}(<R)$ is the value of $\Sigma$ averaged
over the cylinder interior to $R.$
Using the Halotools implementation of the technique presented in Lange et al.
(2019) and reviewed in Appendix F, the profiles $\Delta\Sigma(R)$ can be
computed in advance on a per-object basis for every subhalo in a simulated
snapshot. Once tabulated, the lensing signal produced by a stack of galaxies
can be computed as the average signal produced by each of the subhalos in the
sample,
$\displaystyle\langle\Delta\Sigma(R)\rangle=\langle\Delta\Sigma^{\alpha}(R)\rangle_{\alpha\in\
{\rm sample}},$ (10)
where $\Delta\Sigma^{\alpha}$ is the lensing of the $\alpha^{\rm th}$ subhalo
in the catalog.
Traditional SHAM predictions for galaxy lensing are straightforward to compute
using standard Monte Carlo methods. First, a value of $M_{\star}^{\alpha}$ is
mapped onto every subhalo in the snapshot according to a random draw from the
appropriate log-normal. The SHAM prediction for the lensing signal produced by
a stack of galaxies is then given simply by Eq. 10, taking the average over
only those galaxies with randomly drawn values of $M_{\star}^{\alpha}$ that
satisfy the desired selection criteria, e.g., $M_{\star}^{\rm
lo}<M_{\star}^{\alpha}<M_{\star}^{\rm hi}.$
Figure 5: Gradients of simulation-based $\Delta\Sigma$ predictions. In the top
panel, we show $\Delta\Sigma$ for galaxies with $M_{\star}>10^{11}M_{\odot}$
residing in subhalos in the BPL simulation, with different curves
corresponding to values of the parameters controlling mass-dependent scatter,
as indicated in the legend. In the bottom panel we show the logarithmic
derivative of $\Delta\Sigma,$ calculating the result with both finite-
differencing methods and automatic differentiation.
As a differentiable alternative to Monte Carlo-based predictions, we can
instead compute the left-hand side of Eq. 10 as a weighted average, where the
weight attached to each subhalo, $w_{\alpha},$ is the probability that the
galaxy passes the selection function, calculated according to Eq. 6:
$\displaystyle\langle\Delta\Sigma(R)\rangle=\frac{\sum_{\alpha}w_{\alpha}\cdot\Delta\Sigma^{\alpha}(R)}{\sum_{\alpha}w_{\alpha}},$
(11)
where the summations in Eq. 11 are performed over all subhalos with non-zero
probability of residing in the stellar mass bin. In the limit of an infinite
simulated volume, Eq. 10 converges to Eq. 11, but the latter method has the
advantage of being differentiable with respect to parameters $\theta$ of the
galaxy–halo connection:
$\displaystyle\frac{\partial}{\partial\theta}\langle\Delta\Sigma(R)\rangle=\sum_{\alpha}\frac{\partial}{\partial\theta}\left(\frac{w_{\alpha}}{\sum_{\alpha}w_{\alpha}}\right)\Delta\Sigma^{\alpha}(R).$
(12)
To compute the left-hand side of Eq. 12, we use JAX to calculate the quantity
$\frac{\partial}{\partial\theta}(w_{\alpha}/\sum_{\alpha}w_{\alpha})$ for each
simulated subhalo, and we then we use numba to perform the weighted-sum on the
right-hand-side of Eq. 12 using the precomputed lensing profile of each
individual subhalo.
In the top panel of Figure 5, we show a few examples of computations of
$\Delta\Sigma$ for galaxies with $M_{\star}>10^{11}M_{\odot}$ based on the BPL
simulation, with different curves corresponding to different values of the
parameters controlling mass-dependent scatter. For the fiducial model shown
with the black curve, we use $(\sigma_{\rm lo},\sigma_{\rm hi})=(0.4,0.25);$
results based on independent perturbations to $\sigma_{\rm lo}$ and
$\sigma_{\rm hi}$ are color-coded as indicated in the legend. In the bottom
panel of Figure 5, we show gradients of $\Delta\Sigma$ with respect to
$\sigma_{\rm lo}$ and $\sigma_{\rm hi},$ again comparing the results of the
computation based on autodiff vs. finite-differencing methods. Due to the
power-law shape of the subhalo mass function, larger scatter in the SMHM at
any mass corresponds to a lower amplitude of $\Delta\Sigma,$ since increasing
the scatter results in a higher proportion of low-mass, weakly-clustered
subhalos that up-scatter into the sample. The effect on $\Delta\Sigma$ for
this relatively massive galaxy sample is more pronounced for $\sigma_{\rm
hi};$ this is sensible, since at high mass the slope of the mass function is
falling off rapidly, and the halo bias function is rapidly steepening.
### 4.3 Galaxy Clustering
The computation of differentiable two-point functions proceeds in much the
same way as that for galaxy lensing. Each object again has a probability
$w_{\alpha}$ of being in the sample. The product of the weights,
$w_{\alpha}\cdot w_{\beta}$, is interpreted as the probability that a given
pair of objects would be counted when accumulating the number of pairs at a
given separation for, e.g., computing a two-point function. Formally, we are
assuming that when conditioned on the values of the weights themselves, the
presence of any given object in the sample is independent of all of the
others. Finally, there exists some ambiguity in how exactly to define an
unclustered or “random” sample given a set of objects with weights. We define
the random sample as a set of points with a random spatial distribution, each
with a weight that has been randomly assigned from the original sample. We
show below that this definition properly reproduces zero clustering signal for
samples where points are distributed randomly, and only kept with a
probability equal to the weight.
With these assumptions, we use the following estimator for the two-point
correlation function
$\displaystyle\hat{\xi}_{i}=\frac{DD_{i}}{RR_{i}}-1$ (13)
where
$\displaystyle
DD_{i}=\sum_{\alpha}\sum_{\beta\neq\alpha}w_{\alpha}w_{\beta}{\cal
S}_{i}(x_{\alpha},x_{\beta})$ (14)
and
$\displaystyle
RR_{i}=\left(\sum_{\alpha}w_{\alpha}\right)^{2}\left(1-\frac{1}{N_{\rm
eff}}\right)\frac{V_{i}}{V}\ .$ (15)
Here ${\cal S}_{i}(x_{\alpha},x_{\beta})$ is the bin selection function, which
is unity if points $\alpha$ and $\beta$ are separated such that their
separation falls in bin $i$, and zero otherwise; $N$ is the total number of
points in the sample, $V_{i}$ is the volume of the bin, and $V$ is the total
volume of the domain; finally, the quantity $N_{\rm eff}$ is defined as:
$\displaystyle N_{\rm
eff}\equiv\frac{\left(\sum_{\alpha}w_{\alpha}\right)^{2}}{\sum_{\alpha}w_{\alpha}^{2}},$
(16)
so that $N_{\rm eff}$ is interpreted as the effective sample size.
To better understand the role of the denominator, $RR_{i},$ we can rewrite the
estimator as
$\displaystyle
RR_{i}=\left(\sum_{\alpha}w_{\alpha}\left(\sum_{\beta}w_{\beta}-w_{\alpha}\right)\right)\times\frac{V_{i}}{V}$
(17)
In this form, we can see that this estimator directly computes the expected
pair counts, excluding self-pairs, if we have randomly assigned the weights to
the random sample. This estimator is equivalent to the
$\mathcal{O}(\frac{1}{N})$ corrections for finite samples described in the
Corrfunc documentation (Sinha & Garrison, 2020), but with weights that
represent membership probabilities. Finally, the factor $V_{i}/V$ is the
fraction of the total volume occupied by the $i^{\rm th}$ bin.
Figure 6: Gradients of simulation-based $w_{\rm p}$ predictions. Same as Fig.
5, but for $w_{\rm p}({\rm r_{p}}).$ In the top panel, different curves show
$w_{\rm p}({\rm r_{p}})$ for different values of the parameters controlling
mass-dependent scatter, as indicated in the legend. In the bottom panel we
show the logarithmic derivative of $w_{\rm p}({\rm r_{p}}),$ calculating the
result with both finite-differencing methods and automatic differentiation.
From this definition of the two-point clustering estimator, we can use the
chain rule to compute the derivative,
$\displaystyle\frac{\partial\hat{\xi}_{i}}{\partial\theta}=\frac{1}{RR_{i}}\frac{\partial
DD_{i}}{\partial\theta}-\frac{DD_{i}}{RR_{i}^{2}}\frac{\partial
RR_{i}}{\partial\theta}$ (18)
where
(19) $\displaystyle\frac{\partial
DD_{i}}{\partial\theta}=\sum_{\alpha}\sum_{\beta\neq\alpha}\left[\frac{\partial
w_{\alpha}}{\partial\theta}w_{\beta}+w_{\alpha}\frac{\partial
w_{\beta}}{\partial\theta}\right]{\cal S}_{i}(x_{\alpha},x_{\beta}),$
and
$\displaystyle\frac{\partial RR_{i}}{\partial\theta}$ $\displaystyle=$
$\displaystyle\left(\sum_{\alpha}\left[\frac{\partial
w_{\alpha}}{\partial\theta}\left(\sum_{\beta}w_{\beta}-w_{\alpha}\right)\right.\right.$
$\displaystyle\ \ \ \ \ \ \ \ \ \ \
\left.\left.+w_{\alpha}\left(\sum_{\beta}\frac{\partial
w_{\beta}}{\partial\theta}-\frac{\partial
w_{\alpha}}{\partial\theta}\right)\right]\right)\times\frac{V_{i}}{V}$
We assume here that the parameters $\theta$ exclusively impact the value of
the weights mapped onto each object. Should the object positions also depend
on $\theta$, additional terms proportional to the spatial derivative of the
bin selection function ${\cal S}_{i}(x_{\alpha},x_{\beta})$ would arise. We
leave the treatment of such contributions to the derivative to future work
exploring models that parameterize changes to galaxy position.
We note that the expressions derived in this section can be quite useful in
cases where one has existing weighted pair-counting code to efficiently
compute two-point functions. In this case, an autodiff library such as JAX can
be used to calculate the per-object weights, $w_{\alpha},$ and the gradients
$\partial w_{\alpha}/\partial\theta;$ subsequently, the pair-counting code can
proceed with its usual computation, using the equations derived in this
section to transform the results into a computation of a two-point function
and its exact derivatives. In carrying out the calculations in this section,
we use used the corrfunc code for this purpose.
In the top panel of Figure 6, we show $w_{\rm p}({\rm r_{p}})$ for galaxies
with $M_{\star}>10^{11}M_{\odot}$ based on the BPL simulation, with the same
fiducial scatter model and perturbations used in Fig. 5. In the bottom panel
of Figure 6, we show gradients of $w_{\rm p}({\rm r_{p}})$ with respect to
$\sigma_{\rm lo}$ and $\sigma_{\rm hi}.$ We see the same trends in Figure 6 as
we saw in Fig. 5: increasing scatter in the SMHM decreases clustering
strength, with more pronounced effects from scatter at high mass.
## 5 Discussion & Future Work
We have presented a new approach to making simulation-based predictions of the
galaxy–halo connection. In the conventional methodology, synthetic galaxy
populations are generated from a stochastic Monte Carlo realization of some
probability distribution defined by the underlying model. For example, in most
abundance matching studies, stellar masses are drawn from a realization of a
log-normal PDF; the synthetic galaxy at the center of each subhalo is then
assigned to a unique bin of stellar mass (or otherwise discarded from the
sample), and predictions for clustering and lensing are made from the
resulting bins of point data (as in, e.g., Reddick et al., 2013; Hearin et
al., 2013). By contrast, the approach taken here is not based on Monte Carlo
realizations. Instead, every subhalo makes a PDF-weighted contribution to each
stellar mass bin, and the statistical estimators used to calculate summary
statistics are defined in terms of the weighted point data.
Previous implementations of the galaxy–halo connection have also opted in
favor of weighted summary statistics over hard-edged bins filled with noisy
Monte Carlo realizations. For example, a variation on this technique was used
in Reid et al. (2014) in their treatment of fiber collisions; the
UniverseMachine source code used PDF-weighting methods in the computation of
its DR1 summary statistics (Behroozi et al., 2019); these same techniques form
the basis of the TabCorr source code222https://github.com/johannesulf/TabCorr,
as well as a variety of previous works on the galaxy–halo connection that
utilize pre-computation methods (e.g., Neistein & Khochfar, 2012; Zheng & Guo,
2016).
Although implementations of the galaxy–halo connection utilizing pre-
computation methods also enjoy the same reduction of stochasticity as the
approach presented here, our framework has two distinct advantages that make
it more extensible to the problem sizes of cosmological modeling in the 2020s.
First, conventional pre-computation methods are implemented by tabulating a
set of results over a grid of (sub)halo properties, and then linearly
interpolating from the resulting lookup table. While this methodology has
proven to be quite effective in deriving constraints on cosmological
parameters (e.g., Lange et al., 2021), extending these techniques to models
that depend upon more than one or two (sub)halo properties would be quite
challenging, since implementations based on lookup-table interpolation have
memory demands that rise sharply with the dimension of the table. Our use of
AI resolves this issue of excessive memory demands, in essence by storing the
lookup-table information in the weights and biases of a neural network.
Second, our predictions for large scale structure are differentiable. Although
we achieve this property in part due to our use of PDF-weighted summary
statistics, as discussed above this is not the novel feature of our
methodology. As described in §4, our predictions for the stellar mass
function, galaxy–galaxy lensing, and galaxy clustering are differentiable as a
result of the analytical propagation of the gradients of each synthetic
galaxy’s weight to the point estimators of the summary statistics. The
availability of gradient information is the central feature that enables
modern Bayesian inference techniques such as Hamiltonian Monte Carlo to derive
converged posteriors for models of hundreds of parameters (Hoffman & Gelman,
2014), and the extensibility of our framework to physical models of higher
dimension is one of the principal motivations underlying this paper.
Our paper focuses on the most widely-used 1- and 2-point functions in
cosmology, but our framework could naturally extend to higher-order summary
statistics. In some cases, virtually no adaptation of the computations
detailed in §4 would be required. For example, both the under-density
probability function (UPF) and counts-in-cells (CIC) summary statistics have
been shown to be sensitive probes of galaxy–halo information that is not
contained in galaxy clustering or lensing (Tinker et al., 2008; Wang et al.,
2019). Both of these summary statistics are defined in terms of number counts
of galaxies residing within some enclosing volume. If the UPF and CIC
estimators are instead defined in terms of the total sum of galaxy weights
residing in the enclosing volume, the predictions for these statistics become
differentiable using the same techniques in §4. Recent work has shown that the
information content of $n-$point functions of arbitrarily high order is
formally contained in summary statistics based on $k$ nearest-neighbor ($k$NN)
computations (Banerjee & Abel, 2021a, b); this approach to extracting
information from the density field is an attractive alternative to using
higher-order $n-$point functions directly, since the computation of $k$NN-
based summary statistics scales like $\mathcal{O}(N\log N);$ moreover, it has
also recently been shown that $k$NN-based measurements are sensitive probes of
the galaxy–halo connection (Behroozi et al., 2021). In the standard
computation of these summary statistics, the estimator is defined in terms of
the spatial distance to the $N^{\rm th}-$nearest member of a galaxy sample
defined by some hard-edged bin or threshold; in the corresponding
differentiable version, one would instead use the smallest distance for which
the sum of galaxy weights exceeds some chosen value. We relegate the extension
of our framework to these and other higher-order summary statistics to future
work.
Our work is closely related to Horowitz et al. (2021), who have implemented
DiffHOD, a differentiable form of the Halo Occupation Distribution (HOD).
Whereas the HOD populates host dark matter halos with variable numbers of
satellites, SHAM populates each simulated subhalo with a single synthetic
galaxy. Nonetheless, the theoretical roots of these two empirical models are
tightly connected (Kravtsov et al., 2004), and DiffHOD and SHAMNet are part of
a growing trend of differentiable formulations of galaxy–halo connection
models. In these two examples, the differentiability of the model predictions
is achieved through a different methodology. In SHAMNet, the derivatives of
the galaxy–halo connection model are propagated through analytical PDFs of the
galaxy residing in each halo, so that SHAMNet predictions have no
stochasticity; by contrast, DiffHOD uses a differentiable form of stochastic
sampling based on a Gumbel-Softmax distribution, coupled with an annealing
technique (Jang et al., 2016).
The GalaxyNet model (Moster et al., 2020) is also a differentiable form of the
galaxy–halo connection, and so the techniques presented here could naturally
be used to transform GalaxyNet predictions for clustering and lensing into
differentiable computations. Whereas DiffHOD and SHAMNet use AI-based methods
in a manner that mimics the scaling relations of simple empirical models,
GalaxyNet builds a highly nonlinear galaxy–halo connection with its neural
network, and so is a far more expansive application of AI. Our two-phase
training of SHAMNet is an adaption of the same technique used to train
GalaxyNet. As detailed in Appendix D, we trained SHAMNet by first carrying out
an initialization phase in which the weights and biases of the network were
tuned to reproduce an existing model; training then proceeded with a
subsequent phase in which the cost function was defined directly in terms of
target summary statistics of the galaxy distribution. The initialization phase
of SHAMNet was based on a flexible parametric form of the stellar-to-halo-mass
relation (see Appendix C), whereas GalaxyNet was initialized based on the
EMERGE model of star formation history (Moster et al., 2018).
Our primary interest in the differentiable techniques introduced here is their
application to more physically complex and higher-dimensional models than
abundance matching. In particular, in closely related work (Hearin et al.,
2021b), we have introduced diffmah, a differentiable model for the assembly of
dark matter halos; the diffmah model is the basis of a new approach to the
galaxy–halo connection, diffstar, in which parameterized star formation
histories (SFH) are statistically mapped onto the merger trees of dark matter
halos (Alarcon et al., 2021). Formulating the galaxy–halo connection in terms
of parameterized SFHs allows us to additionally forward model galaxy SEDs in a
differentiable fashion through the use of DSPS, a JAX-based implementation of
stellar population synthesis (Hearin et al., 2021a). Our focus on the benefits
of gradient information in likelihood analyses anticipates the expansion of
the dimension of the parameter space required by the complexity of these
models.
In the present work, we have used SHAM primarily as a toy model to demonstrate
how gradients of parameters of a galaxy–halo connection model propagate
through to derivatives of predictions for the summary statistics of large
scale structure. However, analyzing the galaxy distribution with SHAM is an
active area of research unto itself, and SHAMNet offers some improvements upon
conventional deconvolution-based abundance matching that may be useful for
such purposes. For example, the RL deconvolution algorithm outlined in §2 has
notoriously finicky convergence properties at the high-mass end, whereas
SHAMNet provides a robust solution to Eq. 2 even at very high mass (see Figure
3). More importantly, SHAMNet naturally incorporates mass-dependent scatter in
the stellar-to-halo mass relation, whereas all previous deconvolution-based
SHAM constraints have been derived under the assumption of constant scatter,
at least in part due to the technical challenge of incorporating this feature
into the RL-deconvolution algorithm.
With only modest extensions to SHAMNet, differentiable versions of
contemporary analyses could be applied to observational data. The present work
is based on $M_{\rm peak}-$based abundance matching, and so in order to
support abundance matching based on alternative mass proxies such as $V_{\rm
peak}$ (Reddick et al., 2013), $V_{\rm relax}$ (Chaves-Montero et al., 2016),
or more generalized proxies (Lehmann et al., 2017), it would be necessary to
retrain SHAMNet based on an alternative parameterization of the subhalo
abundance function. Although in §4 we focused on the gradients of the mass-
dependent scatter, we trained SHAMNet to additionally capture the dependence
of $\langle M_{\star}|M_{\rm peak}\rangle_{\rm med}$ upon parameters
specifying the SMF, and so our approach could also be used in abundance
matching formulations such as Saito et al. (2016) that incorporate uncertainty
in the SMF. In order to incorporate cosmology-dependence into abundance
matching, as in Contreras et al. (2021), an additional model providing a
mapping from cosmological parameters to the parameterized subhalo abundance
function would need to be developed. If an analysis required additional
marginalization over uncertainty in orphan abundance, as in DeRose et al.
(2021a), it would be necessary to develop an additional ingredient for how the
parameters of the orphan prescription change the parameters of the subhalo
abundance function.
With comparably modest adaptations, our framework could also be used to
conduct differentiable analyses of conditional abundance matching (CAM, Hearin
et al., 2014). While SHAM supplies a mapping from e.g., halo mass to stellar
mass, the CAM framework supplies a mapping from some secondary subhalo
property (such as a halo formation time proxy, as in Hearin & Watson, 2013;
Masaki et al., 2013), to some secondary galaxy property (such as specific star
formation rate, as in Watson et al., 2015). In order to make CAM
differentiable, rather than parameterizing the SMF and SHMF, one would instead
need to calibrate parameterized models for the conditional abundance of the
secondary halo and galaxy properties, and then use, e.g., the
conditional_abunmatch function in halotools to generate training data for
CAMNet. Although the version of SHAMNet that we trained in the present work
does not directly support these features, we have made our source code
publicly available to facilitate these and other adaptations in future work.
## 6 Summary
1. 1.
We have devised a new, differentiable framework for making simulation-based
predictions for large scale structure. Our approach is not based on stochastic
Monte Carlo realizations, but is instead formulated in terms of parameterized
PDFs that are mapped onto each simulated halo. We then use automatic
differentiation to propagate gradients of the galaxy–halo connection through
the point estimators used to measure summary statistics of the galaxy
distribution.
2. 2.
We have used a neural network, SHAMNet, to approximate the stellar-to-halo
mass relationship (SMHM) of abundance matching. Our JAX-based implementation
of SHAMNet is available for installation with pip. In addition to being
differentiable, SHAMNet captures variable levels of $M_{\rm halo}$-dependent
scatter in the SMHM, which is challenging to incorporate into conventional
deconvolution-based implementations of abundance matching.
3. 3.
Our source code is publicly available at
https://github.com/ArgonneCPAC/shamnet, and provides a set of recipes that can
be adapted to formulate other differentiable, AI-accelerated models of the
galaxy–halo connection.
## 7 Acknowledgements
We thank Peter Behroozi for useful discussions. APH thanks José Feliciano
again for Feliz Navidad.
We thank the developers of NumPy (Van Der Walt et al., 2011), SciPy (Jones et
al., 2001-2016), Jupyter (Ragan-Kelley et al., 2014), IPython (Pérez &
Granger, 2007), scikit-learn (Pedregosa et al., 2011), JAX (Bradbury et al.,
2018), numba (Lam et al., 2015), conda-forge (conda-forge community, 2015),
and Matplotlib (Hunter, 2007) for their extremely useful free software. While
writing this paper we made extensive use of the Astrophysics Data Service
(ADS) and arXiv preprint repository. The Bolshoi simulations used in this work
have been performed within the Bolshoi project of the University of California
High-Performance AstroComputing Center (UC-HiPACC) and were run at the NASA
Ames Research Center.
Work done at Argonne National Laboratory was supported under the DOE contract
DE-AC02-06CH11357. JD is supported by the Chamberlain Fellowship at Lawrence
Berkeley National Laboratory. We gratefully acknowledge use of the Bebop
cluster in the Laboratory Computing Resource Center at Argonne National
Laboratory.
## References
* Alarcon et al. (2021) Alarcon, A., Hearin, A., Becker, M., & Chaves-Montero, J. 2021, in prep
* Banerjee & Abel (2021a) Banerjee, A., & Abel, T. 2021a, MNRAS, 500, 5479, doi: 10.1093/mnras/staa3604
* Banerjee & Abel (2021b) —. 2021b, MNRAS, 504, 2911, doi: 10.1093/mnras/stab961
* Behroozi et al. (2021) Behroozi, P., Hearin, A., & Moster, B. P. 2021, arXiv:2101.05280, arXiv:2101.05280. https://arxiv.org/abs/2101.05280
* Behroozi et al. (2019) Behroozi, P., Wechsler, R. H., Hearin, A. P., & Conroy, C. 2019, MNRAS, 488, 3143, doi: 10.1093/mnras/stz1182
* Behroozi et al. (2010) Behroozi, P. S., Conroy, C., & Wechsler, R. H. 2010, ApJ, 717, 379, doi: 10.1088/0004-637X/717/1/379
* Behroozi et al. (2013a) Behroozi, P. S., Wechsler, R. H., & Conroy, C. 2013a, ApJ, 770, 57, doi: 10.1088/0004-637X/770/1/57
* Behroozi et al. (2013b) Behroozi, P. S., Wechsler, R. H., & Wu, H.-Y. 2013b, ApJ, 762, 109, doi: 10.1088/0004-637X/762/2/109
* Behroozi et al. (2013c) Behroozi, P. S., Wechsler, R. H., Wu, H.-Y., et al. 2013c, ApJ, 763, 18, doi: 10.1088/0004-637X/763/1/18
* Berlind & Weinberg (2002) Berlind, A. A., & Weinberg, D. H. 2002, ApJ, 575, 587, doi: 10.1086/341469
* Bernardeau et al. (2002) Bernardeau, F., Colombi, S., Gaztañaga, E., & Scoccimarro, R. 2002, Phys. Rep., 367, 1, doi: 10.1016/S0370-1573(02)00135-7
* Bocquet et al. (2020) Bocquet, S., Heitmann, K., Habib, S., et al. 2020, ApJ, 901, 5, doi: 10.3847/1538-4357/abac5c
* Bradbury et al. (2018) Bradbury, J., Frostig, R., Hawkins, P., et al. 2018, JAX: composable transformations of Python+NumPy programs, 0.2.10. http://github.com/google/jax
* Cacciato et al. (2013) Cacciato, M., van den Bosch, F. C., More, S., Mo, H., & Yang, X. 2013, MNRAS, 430, 767, doi: 10.1093/mnras/sts525
* Carrasco et al. (2012) Carrasco, J. J. M., Hertzberg, M. P., & Senatore, L. 2012, Journal of High Energy Physics, 2012, 82, doi: 10.1007/JHEP09(2012)082
* Chapman et al. (2021) Chapman, M. J., Mohammad, F. G., Zhai, Z., et al. 2021, arXiv:2106.14961, arXiv:2106.14961. https://arxiv.org/abs/2106.14961
* Chaves-Montero et al. (2016) Chaves-Montero, J., Angulo, R. E., Schaye, J., et al. 2016, MNRAS, 460, 3100, doi: 10.1093/mnras/stw1225
* conda-forge community (2015) conda-forge community. 2015, The conda-forge Project: Community-based Software Distribution Built on the conda Package Format and Ecosystem, Zenodo, doi: 10.5281/zenodo.4774216
* Conroy et al. (2006) Conroy, C., Wechsler, R. H., & Kravtsov, A. V. 2006, ApJ, 647, 201, doi: 10.1086/503602
* Contreras et al. (2021) Contreras, S., Chaves-Montero, J., Zennaro, M., & Angulo, R. E. 2021, arXiv e-prints, arXiv:2105.05854. https://arxiv.org/abs/2105.05854
* DeRose et al. (2021a) DeRose, J., Becker, M. R., & Wechsler, R. H. 2021a, arXiv e-prints, arXiv:2105.12104. https://arxiv.org/abs/2105.12104
* DeRose et al. (2021b) DeRose, J., Chen, S.-F., White, M., & Kokron, N. 2021b, arXiv:2112.05889, arXiv:2112.05889. https://arxiv.org/abs/2112.05889
* DES Collaboration et al. (2021) DES Collaboration, Abbott, T. M. C., Aguena, M., et al. 2021, arXiv e-prints, arXiv:2105.13549. https://arxiv.org/abs/2105.13549
* Desjacques et al. (2018) Desjacques, V., Jeong, D., & Schmidt, F. 2018, Phys. Rep., 733, 1, doi: 10.1016/j.physrep.2017.12.002
* Eddington (1913) Eddington, A. S. 1913, MNRAS, 73, 359, doi: 10.1093/mnras/73.5.359
* Euclid Collaboration et al. (2020) Euclid Collaboration, Knabenhans, M., Stadel, J., et al. 2020, arXiv:2010.11288, arXiv:2010.11288. https://arxiv.org/abs/2010.11288
* Foreman-Mackey et al. (2019) Foreman-Mackey, D., Farr, W., Sinha, M., et al. 2019, The Journal of Open Source Software, 4, 1864, doi: 10.21105/joss.01864
* García et al. (2021) García, R., Rozo, E., Becker, M. R., & More, S. 2021, MNRAS, doi: 10.1093/mnras/stab1317
* Hearin et al. (2021a) Hearin, A. P., Chaves-Montero, J., Alarcon, A., Becker, M. R., & Benson, A. 2021a, arXiv:2112.06830, arXiv:2112.06830. https://arxiv.org/abs/2112.06830
* Hearin et al. (2021b) Hearin, A. P., Chaves-Montero, J., Becker, M. R., & Alarcon, A. 2021b, arXiv e-prints, arXiv:2105.05859. https://arxiv.org/abs/2105.05859
* Hearin & Watson (2013) Hearin, A. P., & Watson, D. F. 2013, MNRAS, 435, 1313, doi: 10.1093/mnras/stt1374
* Hearin et al. (2014) Hearin, A. P., Watson, D. F., Becker, M. R., et al. 2014, MNRAS, 444, 729, doi: 10.1093/mnras/stu1443
* Hearin et al. (2013) Hearin, A. P., Zentner, A. R., Berlind, A. A., & Newman, J. A. 2013, MNRAS, 433, 659, doi: 10.1093/mnras/stt755
* Hearin et al. (2017) Hearin, A. P., Campbell, D., Tollerud, E., et al. 2017, AJ, 154, 190, doi: 10.3847/1538-3881/aa859f
* Heitmann et al. (2006) Heitmann, K., Higdon, D., Nakhleh, C., & Habib, S. 2006, ApJ, 646, L1, doi: 10.1086/506448
* Hoffman & Gelman (2014) Hoffman, M., & Gelman, A. 2014, Journal of Machine Learning Research, 15, 1593\. http://jmlr.org/papers/v15/hoffman14a.html
* Horowitz et al. (2021) Horowitz, B., Hahn, C., Lanusse, F., Modi, C., & Ferraro, S. 2021, in prep
* Hunter (2007) Hunter, J. D. 2007, Computing In Science & Engineering, 9, 90, doi: 10.1109/MCSE.2007.55
* Iman et al. (1981) Iman, R., Helton, J., & Campbell, J. 1981, Journal of Quality Technology, 13, 174, doi: 10.1080/00224065.1981.11978748
* Jang et al. (2016) Jang, E., Gu, S., & Poole, B. 2016, arXiv:1611.01144, arXiv:1611.01144. https://arxiv.org/abs/1611.01144
* Jenkins et al. (2001) Jenkins, A., Frenk, C. S., White, S. D. M., et al. 2001, MNRAS, 321, 372, doi: 10.1046/j.1365-8711.2001.04029.x
* Jones et al. (2001-2016) Jones, E., Oliphant, T., Peterson, P., et al. 2001-2016, http://www.scipy.org
* Joudaki et al. (2018) Joudaki, S., Blake, C., Johnson, A., et al. 2018, MNRAS, 474, 4894, doi: 10.1093/mnras/stx2820
* Kingma & Ba (2014) Kingma, D. P., & Ba, J. 2014, arXiv:1412.6980, arXiv:1412.6980. https://arxiv.org/abs/1412.6980
* Klambauer et al. (2017) Klambauer, G., Unterthiner, T., Mayr, A., & Hochreiter, S. 2017, arXiv:1706.02515, arXiv:1706.02515. https://arxiv.org/abs/1706.02515
* Klypin et al. (2011) Klypin, A. A., Trujillo-Gomez, S., & Primack, J. 2011, ApJ, 740, 102, doi: 10.1088/0004-637X/740/2/102
* Kokron et al. (2021) Kokron, N., DeRose, J., Chen, S.-F., White, M., & Wechsler, R. H. 2021, MNRAS, 505, 1422, doi: 10.1093/mnras/stab1358
* Krause & Eifler (2017) Krause, E., & Eifler, T. 2017, MNRAS, 470, 2100, doi: 10.1093/mnras/stx1261
* Kravtsov et al. (2004) Kravtsov, A. V., Berlind, A. A., Wechsler, R. H., et al. 2004, ApJ, 609, 35, doi: 10.1086/420959
* Kravtsov et al. (1997) Kravtsov, A. V., Klypin, A. A., & Khokhlov, A. M. 1997, ApJS, 111, 73, doi: 10.1086/313015
* Kravtsov et al. (2018) Kravtsov, A. V., Vikhlinin, A. A., & Meshcheryakov, A. V. 2018, Astronomy Letters, 44, 8, doi: 10.1134/S1063773717120015
* Kwan et al. (2015) Kwan, J., Heitmann, K., Habib, S., et al. 2015, ApJ, 810, 35, doi: 10.1088/0004-637X/810/1/35
* Lam et al. (2015) Lam, S. K., Pitrou, A., & Seibert, S. 2015, in Proceedings of the Second Workshop on the LLVM Compiler Infrastructure in HPC, 1–6
* Lange et al. (2021) Lange, J. U., Hearin, A. P., Leauthaud, A., et al. 2021, arXiv:2101.12261, arXiv:2101.12261. https://arxiv.org/abs/2101.12261
* Lange et al. (2019) Lange, J. U., Yang, X., Guo, H., Luo, W., & van den Bosch, F. C. 2019, MNRAS, 488, 5771, doi: 10.1093/mnras/stz2124
* Lawrence et al. (2017) Lawrence, E., Heitmann, K., Kwan, J., et al. 2017, ApJ, 847, 50, doi: 10.3847/1538-4357/aa86a9
* Lehmann et al. (2017) Lehmann, B. V., Mao, Y.-Y., Becker, M. R., Skillman, S. W., & Wechsler, R. H. 2017, ApJ, 834, 37, doi: 10.3847/1538-4357/834/1/37
* Li & White (2009) Li, C., & White, S. D. M. 2009, MNRAS, 398, 2177, doi: 10.1111/j.1365-2966.2009.15268.x
* Lucy (1974) Lucy, L. B. 1974, AJ, 79, 745, doi: 10.1086/111605
* Lupton et al. (1999) Lupton, R. H., Gunn, J. E., & Szalay, A. S. 1999, AJ, 118, 1406, doi: 10.1086/301004
* Masaki et al. (2013) Masaki, S., Lin, Y.-T., & Yoshida, N. 2013, MNRAS, 436, 2286, doi: 10.1093/mnras/stt1729
* McClintock et al. (2019a) McClintock, T., Rozo, E., Becker, M. R., et al. 2019a, ApJ, 872, 53, doi: 10.3847/1538-4357/aaf568
* McClintock et al. (2019b) McClintock, T., Rozo, E., Banerjee, A., et al. 2019b, arXiv:1907.13167, arXiv:1907.13167. https://arxiv.org/abs/1907.13167
* McKay et al. (1979) McKay, M., Beckman, R., & Conover, W. 1979, Technometrics, 21, 239, doi: 10.1080/00401706.1979.10489755
* Miyatake et al. (2021) Miyatake, H., Sugiyama, S., Takada, M., et al. 2021, arXiv:2111.02419, arXiv:2111.02419. https://arxiv.org/abs/2111.02419
* Moster et al. (2020) Moster, B. P., Naab, T., Lindström, M., & O’Leary, J. A. 2020, arXiv:2005.12276, arXiv:2005.12276. https://arxiv.org/abs/2005.12276
* Moster et al. (2013) Moster, B. P., Naab, T., & White, S. D. M. 2013, MNRAS, 428, 3121, doi: 10.1093/mnras/sts261
* Moster et al. (2018) —. 2018, MNRAS, 477, 1822, doi: 10.1093/mnras/sty655
* Moster et al. (2010) Moster, B. P., Somerville, R. S., Maulbetsch, C., et al. 2010, ApJ, 710, 903, doi: 10.1088/0004-637X/710/2/903
* Neistein & Khochfar (2012) Neistein, E., & Khochfar, S. 2012, arXiv:1209.0463, arXiv:1209.0463. https://arxiv.org/abs/1209.0463
* Nishimichi et al. (2019) Nishimichi, T., Takada, M., Takahashi, R., et al. 2019, ApJ, 884, 29, doi: 10.3847/1538-4357/ab3719
* Panter et al. (2007) Panter, B., Jimenez, R., Heavens, A. F., & Charlot, S. 2007, MNRAS, 378, 1550, doi: 10.1111/j.1365-2966.2007.11909.x
* Pedregosa et al. (2011) Pedregosa, F., Varoquaux, G., Gramfort, A., et al. 2011, Journal of Machine Learning Research, 12, 2825
* Pérez & Granger (2007) Pérez, F., & Granger, B. E. 2007, Computing in Science and Engineering, 9, 21, doi: 10.1109/MCSE.2007.53
* Planck Collaboration et al. (2014) Planck Collaboration, Ade, P. A. R., Aghanim, N., et al. 2014, A&A, 571, A16, doi: 10.1051/0004-6361/201321591
* Press & Schechter (1974) Press, W. H., & Schechter, P. 1974, ApJ, 187, 425, doi: 10.1086/152650
* Ragan-Kelley et al. (2014) Ragan-Kelley, M., Perez, F., Granger, B., et al. 2014, in American Geophysical Union Fall Meeting Abstracts, Vol. D7
* Rasmussen & Williams (2006) Rasmussen, C. E., & Williams, C. K. I. 2006, Gaussian Processes for Machine Learning
* Reddick et al. (2014) Reddick, R. M., Tinker, J. L., Wechsler, R. H., & Lu, Y. 2014, ApJ, 783, 118, doi: 10.1088/0004-637X/783/2/118
* Reddick et al. (2013) Reddick, R. M., Wechsler, R. H., Tinker, J. L., & Behroozi, P. S. 2013, ApJ, 771, 30, doi: 10.1088/0004-637X/771/1/30
* Reid et al. (2014) Reid, B. A., Seo, H.-J., Leauthaud, A., Tinker, J. L., & White, M. 2014, MNRAS, 444, 476, doi: 10.1093/mnras/stu1391
* Richardson (1972) Richardson, W. H. 1972, J. Opt. Soc. Am., 62, 55, doi: 10.1364/JOSA.62.000055
* Rodríguez-Puebla et al. (2015) Rodríguez-Puebla, A., Avila-Reese, V., Yang, X., et al. 2015, ApJ, 799, 130, doi: 10.1088/0004-637X/799/2/130
* Rodríguez-Puebla et al. (2016) Rodríguez-Puebla, A., Behroozi, P., Primack, J., et al. 2016, MNRAS, 462, 893, doi: 10.1093/mnras/stw1705
* Saito et al. (2016) Saito, S., Leauthaud, A., Hearin, A. P., et al. 2016, MNRAS, 460, 1457, doi: 10.1093/mnras/stw1080
* Sgró et al. (2013) Sgró, M. A., Paz, D. J., & Merchán, M. 2013, MNRAS, 433, 787, doi: 10.1093/mnras/stt773
* Sinha & Garrison (2020) Sinha, M., & Garrison, L. H. 2020, MNRAS, 491, 3022, doi: 10.1093/mnras/stz3157
* Tasitsiomi et al. (2004) Tasitsiomi, A., Kravtsov, A. V., Wechsler, R. H., & Primack, J. R. 2004, ApJ, 614, 533, doi: 10.1086/423784
* Tinker et al. (2008) Tinker, J. L., Conroy, C., Norberg, P., et al. 2008, ApJ, 686, 53, doi: 10.1086/589983
* Tinker et al. (2010) Tinker, J. L., Robertson, B. E., Kravtsov, A. V., et al. 2010, ApJ, 724, 878, doi: 10.1088/0004-637X/724/2/878
* van den Bosch et al. (2013) van den Bosch, F. C., More, S., Cacciato, M., Mo, H., & Yang, X. 2013, MNRAS, 430, 725, doi: 10.1093/mnras/sts006
* Van Der Walt et al. (2011) Van Der Walt, S., Colbert, S. C., & Varoquaux, G. 2011, ArXiv:1102.1523
* Wang et al. (2019) Wang, K., Mao, Y.-Y., Zentner, A. R., et al. 2019, MNRAS, 488, 3541, doi: 10.1093/mnras/stz1733
* Watson et al. (2015) Watson, D. F., Hearin, A. P., Berlind, A. A., et al. 2015, MNRAS, 446, 651, doi: 10.1093/mnras/stu2065
* Wechsler & Tinker (2018) Wechsler, R. H., & Tinker, J. L. 2018, ARA&A, 56, 435, doi: 10.1146/annurev-astro-081817-051756
* White (2014) White, M. 2014, MNRAS, 439, 3630, doi: 10.1093/mnras/stu209
* Wibking et al. (2020) Wibking, B. D., Weinberg, D. H., Salcedo, A. N., et al. 2020, MNRAS, 492, 2872, doi: 10.1093/mnras/stz3423
* Yuan et al. (2021) Yuan, S., Garrison, L. H., Hadzhiyska, B., Bose, S., & Eisenstein, D. J. 2021, MNRAS, doi: 10.1093/mnras/stab3355
* Zentner (2007) Zentner, A. R. 2007, International Journal of Modern Physics D, 16, 763, doi: 10.1142/S0218271807010511
* Zentner et al. (2014) Zentner, A. R., Hearin, A. P., & van den Bosch, F. C. 2014, MNRAS, 443, 3044, doi: 10.1093/mnras/stu1383
* Zentner et al. (2013) Zentner, A. R., Semboloni, E., Dodelson, S., et al. 2013, Phys. Rev. D, 87, 043509, doi: 10.1103/PhysRevD.87.043509
* Zhai et al. (2019) Zhai, Z., Tinker, J. L., Becker, M. R., et al. 2019, ApJ, 874, 95, doi: 10.3847/1538-4357/ab0d7b
* Zheng & Guo (2016) Zheng, Z., & Guo, H. 2016, MNRAS, 458, 4015, doi: 10.1093/mnras/stw523
* Zheng et al. (2005) Zheng, Z., Berlind, A. A., Weinberg, D. H., et al. 2005, ApJ, 633, 791, doi: 10.1086/466510
## Appendix A Parameterized Subhalo Mass Function
In this section, we describe the fitting function we use to approximate the
subhalo mass function in the Bolshoi-Planck simulation (BPL, Klypin et al.,
2011). The BPL simulation was run with cosmological parameters closely
matching Planck Collaboration et al. (2014), and was carried out using the ART
code (Kravtsov et al., 1997) by evolving $2048^{3}$ dark-matter particles of
mass $m_{\rm p}=1.55\times 10^{8}M_{\odot}$ on a simulation box of $250\,{\rm
Mpc}$ on a side.
For notational convenience, throughout this section we will refer to subhalo
mass using the variable $m_{\rm h}\equiv\log_{10}M_{\rm peak}.$ The basic
quantity that we model is the cumulative mass function, $\rm dN_{\rm
h}(>m_{\rm h})/\rm dV,$ where the quantity $V$ is the comoving volume. The
relationship the cumulative mass function and the differential mass function,
$\rm dn_{\rm h}(m_{\rm h})/\rm dm_{\rm h}\rm dV,$ is given by the following
equation:
$\rm dN_{\rm h}(>m_{\rm h})/\rm dV=\int_{m_{\rm h}}^{\infty}\rm dm_{\rm
h}^{\prime}\rm dn_{\rm h}/\rm dm_{\rm h}^{\prime}\rm dV.$
Figure 7: Fitting function approximation of the subhalo mass function in the
Bolshoi-Planck simulation. Values for $\rm dN_{\rm h}(>x)/\rm dV$ plotted on
the vertical axis are offset by an order of magnitude at different redshifts
for visual clarity.
We parameterize the cumulative mass function to behave like a power-law with
constant index at the low-mass end, and like a power-law with a steeply
declining power-law index at high mass. We capture this behavior by modeling
$y_{\rm h}\equiv\log_{10}N_{\rm h}$ to behave like a linear function of
$m_{\rm h}$ at low mass, and to exhibit a rapidly declining slope above some
critical value of subhalo mass, $m_{\rm c}.$ To accomplish this behavior, we
utilize a sigmoid function:
(21) $\displaystyle\mathcal{S}(x|x_{\rm c},k,y_{\rm lo},y_{\rm hi})\equiv
y_{\rm lo}+\frac{y_{\rm hi}-y_{\rm lo}}{1+\exp(-k\cdot(x-x_{\rm c}))}.$
Using the sigmoid function defined in Eq. 21, we model the cumulative number
density as follows:
(22) $\displaystyle y_{\rm h}(m_{\rm h})=y_{0}+S(m_{\rm h},m_{\rm c},k,y_{\rm
lo},y_{\rm hi})\cdot(m_{\rm h}-m_{\rm c})$
In Eq. 22, the parameter $y_{0}$ controls the normalization of the subhalo
mass function at the critical mass $m_{\rm c};$ the power-law index at the
low-mass end is controlled by $y_{\rm lo},$ and the shape of the high-mass
cutoff is characterized by both $k$ and $y_{\rm hi}.$ We note that basic
theoretical considerations lead one to expect the true halo mass function to
have an exponential cutoff at the high-mass end (see, e.g., Press & Schechter,
1974; Zentner, 2007), whereas even in the limit of infinite mass, Eq. 22 never
attains exponentially declining behavior. For our purposes, however, this
shortcoming is immaterial, since our fitting function closely mimics an
exponential decline for all halo masses relevant for the present study, and we
find that Eq. 22 has more numerically stable behavior in applications
requiring automatic differentiation at the high-mass end.
We calibrate the values of our fitting function parameters using publicly
available333https://www.peterbehroozi.com/data.html subhalo catalog identified
with Rockstar and ConsistentTrees (Behroozi et al., 2013b, c; Rodríguez-Puebla
et al., 2016). While fitting the free parameters, we hold fixed $y_{\rm
lo}=-1.4$, $k=0.9,$ and $y_{\rm hi}=-5.25.$ The remaining two parameters,
$y_{0}$ and $m_{\rm c},$ exhibit redshift dependence that is well approximated
with the sigmoid function defined by Eq. 21. That is, we model $m_{\rm c}(z)$
as:
$\displaystyle m_{\rm c}(z)$ $\displaystyle=$
$\displaystyle\mathcal{S}(z,z_{\rm c}^{m_{\rm c}},k_{\rm c}^{m_{\rm c}},y_{\rm
lo}^{m_{\rm c}},y_{\rm hi}^{m_{\rm c}}),$
and similarly for $y_{0}(z).$ For our best-fitting parameters controlling
$m_{\rm c}(z),$ we have $z_{\rm c}^{m_{\rm c}}=0.75,k_{\rm c}^{m_{\rm
c}}=1,y_{\rm lo}^{m_{\rm c}}=16.175,$ and $y_{\rm hi}^{m_{\rm c}}=12.925.$ For
$y_{0}(z),$ we have $z_{\rm c}^{y_{0}}=0.5,k_{\rm c}^{y_{0}}=1.35,y_{\rm
lo}^{y_{0}}=-8.1,$ and $y_{\rm hi}^{y_{0}}=-5.9.$ In Figure 7 we show the
fidelity with which our calibration approximates the cumulative mass function
of subhalos in BPL. Throughout this paper, we define SHAMNet according to the
$z=0$ mass function, and we supply these parameters as a convenience for
future applications.
## Appendix B Analytical SMF predictions
In §4.1, we outlined a technique to make a differentiable prediction for the
SMF based on an input subhalo catalog. In this section, we describe a
variation on this method that allows one to predict the SMF based an
analytical subhalo mass function (SHMF), such as the one supplied in Appendix
A.
In the absence of scatter between stellar mass, and halo mass, it is
straightforward to predict the SMF, $\phi_{\rm g}(M_{\star}),$ from knowledge
of the SHMF, $\phi_{\rm h}(M_{\rm peak}),$ and the stellar-to-halo mass
relation, $\langle M_{\star}|M_{\rm peak}\rangle,$
$\displaystyle\phi_{\rm g}(M_{\star})=\phi_{\rm h}(\langle M_{\star}|M_{\rm
peak}\rangle)\frac{\rm dM_{\rm peak}}{\rm dM_{\star}},$ (23)
where ${\rm dM_{\rm peak}}/{\rm dM_{\star}}$ is the Jacobian of $\langle
M_{\star}|M_{\rm peak}\rangle.$
Equation 23 is simply the differential version of Equation 1, and so the
calculation of $\phi_{\rm g}(M_{\star})$ in the presence of scatter is given
by Equation 2, repeated here for convenience:
$\displaystyle\phi_{\rm g}(M_{\star})$ $\displaystyle=$
$\displaystyle\int_{0}^{\infty}dM_{\rm peak}\phi_{\rm h}(M_{\rm
peak})P(M_{\star}|M_{\rm peak})$ $\displaystyle=$
$\displaystyle\int_{0}^{\infty}\rm dM_{\star}^{\prime}\frac{\rm dM_{\rm
peak}}{\rm dM_{\star}^{\prime}}\phi_{\rm h}(\langle M_{\star}^{\prime}|M_{\rm
peak}\rangle)P(M_{\star}^{\prime}|M_{\rm peak})$
where scatter in the stellar-to-halo mass relation is encoded by
$P(M_{\star}|M_{\rm peak}).$ To differentiably calculate the left-hand side of
Equation B, one can simply tabulate the integrand on the right-hand side using
a grid of $M_{\star}^{\prime}$ that is sufficiently broad to cover the support
of convolution, and sufficiently dense to achieve the desired precision:
$\displaystyle\phi_{\rm g}(M_{\star})=\sum_{i}\Delta M_{\star,i}^{\prime}$
$\displaystyle\frac{\rm dM_{\rm peak}}{\rm dM_{\star,i}^{\prime}}\phi_{\rm
h}(\langle M_{\star,i}^{\prime}|M_{\rm peak}\rangle)$ (25)
$\displaystyle\times P(M_{\star,i}^{\prime}|M_{\rm peak}).$
The principal difference between Equation 25 and Equation 2 in the main body
of the paper is the presence of the Jacobian, ${\rm dM_{\rm peak}}/{\rm
dM_{\star}};$ this factor does not appear in the simulation-based formulation
because its effect is accounted for by the relative abundance of simulated
halos as a function of $M_{\rm peak}.$
Equation 25 makes it plain to see how to calculate derivatives of $\phi_{\rm
g}(M_{\star}|\theta)$ with respect to the model SHAMNet parameters, $\theta$:
the gradient operator passes through the summation, and $\partial\phi_{\rm
g}(M_{\star})/\partial\theta$ can be calculated by simply summing each term.
For the analytical approximations we use throughout this paper, each of these
terms can in principle be calculated symbolically; we refer the reader to the
smf_scatter_convolution.py module in our source code for our JAX-based
computation of the gradients of Equation 25 based on automatic
differentiation.
## Appendix C Three-roll stellar-to-halo mass relation
As discussed in Appendix §D, in order to train SHAMNet we used an analytical
parameterization of the stellar-to-halo mass relation to support the
initialization of the network. In §C.1, we describe the functional form we use
for this purpose, and in §C.2 we detail our procedure for optimizing the
parameters of this model as a function of the variables $\theta$ that define
the abundance matching problem space reviewed in §3.2.
### C.1 Basic behavior of the three-roll SMHM
In this section, we describe a new functional form we developed to capture the
principal scaling relation of abundance matching, the stellar-to-halo-mass
relation (SMHM), i.e., $\langle M_{\star}|M_{\rm peak}\rangle_{\rm med}.$ In
its general shape, the SMHM relation exhibits a characteristic peak at $M_{\rm
halo}\approx 10^{12}M_{\odot},$ and declines roughly like a power-law at both
the low- and high-mass ends. We parametrically capture this behavior in terms
of the relationship between $m_{\rm h}\equiv\log_{10}M_{\rm peak},$ and
$m_{\star}\equiv\log_{10}M_{\star}:$
$\displaystyle m_{\star}(m_{\rm h})=m_{0}+\alpha(m_{\rm h})\cdot(m_{\rm
h}-m_{\rm crit}).$ (26)
From Eq. 26 we see that $M_{\star}$ scales like a power-law with $M_{\rm
halo},$ with rolling index $\alpha,$ and normalization $m_{0}$ defined by the
stellar mass at halo mass $m_{\rm crit}.$ For the function $\alpha(x),$ we use
the sigmoid function $\mathcal{S}(x)$ defined in Eq. 21 to control the
transition between the low-mass slope, $\alpha_{\rm lo}=\alpha(x\rightarrow
0),$ and the high-mass slope, $\alpha_{\rm hi}=\alpha(x\rightarrow\infty).$
That is, we have
$\alpha(m_{\rm h})=\mathcal{S}(m_{\rm h},m_{\rm crit},k_{\rm crit},\alpha_{\rm
lo},\alpha_{\rm hi}).$
We further allow the power-law indices at low- and high-mass to be functions
of halo mass, $\alpha_{\rm lo}\rightarrow\alpha_{\rm lo}(m_{\rm h}),$ and
$\alpha_{\rm hi}\rightarrow\alpha_{\rm hi}(m_{\rm h}),$ where the mass-
dependence is again controlled by a sigmoid as defined by Eq. 21. The sigmoid-
dependence to $\alpha_{\rm lo}(m_{\rm h})$ and $\alpha_{\rm hi}(m_{\rm h})$
gives the model freedom for an additional roll in the power-law index at both
low- and high-mass ends.
The above formulation gives our SMHM relation a total of eleven parameters:
$m_{0},m_{\rm crit},k_{\rm crit},$ plus four parameters for each of the two
sigmoid functions, $\alpha_{\rm lo}(m_{\rm h}),$ and $\alpha_{\rm hi}(m_{\rm
h}).$ We find that none of the three $k$ parameters provides a physically
useful degree of freedom, and so in all of our applications of this SMHM
relation, we hold these fixed to unity, giving us an eight-dimensional model
for this three-roll SMHM relation.
Figure 8: Flexible model for the stellar-to-halo-mass relation (SMHM). Using
the model described in Appendix §C, different curves show the median stellar
mass as a function of halo mass for different values of $\omega$ in our eight-
dimensional parameter space.
### C.2 Parameterized SHAM with the three-roll SMHM
In this section, we describe how we use the functional form defined in §C.1 to
identify particular solutions to the stellar-to-halo mass relation that
defines abundance matching (see Equation 2). In particular, here describe how
we identify an optimal choice of parameters, $\omega,$ such that the
functional form $m_{\star}(m_{\rm h}|\omega)$ defined in Equation 26 supplies
a solution to Equation 2 for an input $\theta$ that defines the SMF,
$\phi_{\rm g}(M_{\star}|\theta),$ the SHMF, $\phi_{\rm h}(M_{\rm
peak}|\theta),$ and log-normal scatter, $\sigma(M_{\star}|M_{\rm peak}),$
where the $\theta$-dependence of these quantities is specified in §3.
Since the goal of solving Equation 2 is to define a scaling relation that
gives rise to some desired SMF, then in order to solve for $\omega$ we must
quantify our choice for the closeness of two stellar mass functions,
$\phi_{1}(M_{\star})$ and $\phi_{2}(M_{\star}).$ Due to the exponentially
declining nature of the Schechter function, we have found that minimizing the
simple least-squares difference between two SMFs can lead to numerically
unstable results during gradient descent with automatic differentiation,
particularly when seeking high-quality fits at the very massive end. Computing
the logarithmic difference between the two mass functions improves the
numerical instability, but does not resolve it; we also find instabilities
when implementing a hard-edged clip, $\phi_{\rm g}(M_{\star})\rightarrow{\rm
max}\left\\{\phi_{\rm g}(M_{\star}),\phi_{\rm min}\right\\},$ due to the
discontinuity of the derivative at the tiny value of the clip. To resolve
these issues, we have found it beneficial to use a differentiable clipping
function, $\mathcal{R}(x;y_{\rm min}),$ defined as follows
(27) $\displaystyle\mathcal{R}(x;y_{\rm min})\equiv\frac{1}{\ln(10)}\left({\rm
sinh}^{-1}(x/2y_{\rm min})+\ln(y_{\rm min})\right),$
which largely behaves like the base-10 logarithm, but smoothly asymptotes to
$y_{\rm min}$ rather than falling below this value (note that this is the same
transformation used to define magnitudes of SDSS galaxies in Lupton et al.,
1999). Thus when minimizing the difference between two stellar mass functions,
in practice we minimize the difference between
$\mathcal{R}(\phi_{1}(M_{\star}),\phi_{\rm min})$ and
$\mathcal{R}(\phi_{2}(M_{\star}),\phi_{\rm min}),$ using $\phi_{\rm
min}=10^{-15}{\rm Mpc^{-3}M_{\odot}^{-1}}$ to protect against numerical
instabilities in the gradient evaluations.
For each choice of $\theta,$ we searched our model parameter space for an
optimum value of $\omega$ by minimizing the quantity $\mathcal{L}_{\rm MSE},$
defined as
$\displaystyle\mathcal{L}_{\rm MSE}(\theta,\omega)$ $\displaystyle\equiv$
$\displaystyle\frac{1}{N}\sum_{i}\left(y^{\rm i}(\theta)-x^{\rm
i}(\theta,\omega)\right)^{2}$ $\displaystyle y^{\rm i}(\theta)$
$\displaystyle=$ $\displaystyle\mathcal{R}(\phi_{\rm
g}(M_{\star,i}|\phi_{\ast},\alpha))$ (28) $\displaystyle x^{\rm
i}(\theta,\omega)$ $\displaystyle=$ $\displaystyle\mathcal{R}(\phi_{\rm
g}(M_{\star,i}|\theta,\omega))$
where $y^{\rm i}(\theta)$ is the differentiably-clipped logarithm of target
double-Schechter function defined only by $\\{\phi_{\ast},\alpha\\},$ and
$x^{\rm i}(\theta,\omega)$ is the theoretical prediction for the SMF, computed
by the method described in Appendix B. We calculate $\mathcal{L}_{\rm MSE}$ by
evaluating the predicted and target mass functions at a set of $N$ control
points, $M_{\star}^{\rm i},$ using $N=100$ logarithmically-spaced values
spanning the range $10^{9}M_{\odot}<M_{\star}<10^{12}M_{\odot}.$ To minimize
$\mathcal{L}_{\rm MSE}$, we use the JAX implementation of the Adam algorithm
(Kingma & Ba, 2014), which is a gradient descent technique with an adaptive
learning rate, where we also use JAX to compute the gradients. We use 2
successive burn-in cycles with a step-size parameter $s=0.05$ for $\sim 50$
updates, followed by $1000$ updates with $s=0.01.$
In this minimization calculation, the eight parameters $\omega$ defined in
§C.1 are not the actual variables we use to calculate our predicted values for
the SMF; instead we minimize $\omega^{\prime},$ defined by:
$\omega^{\prime}\equiv\mathcal{S}(\omega|\omega_{0},k_{\omega},\omega_{\rm
lo},\omega_{\rm hi}),$
where $\mathcal{S}(x)$ is the sigmoid function defined by Eq. 21. We tailor
the lower and upper bounds of each dimension of $\omega$ according to
reasonable expectations for physically plausible SMHMs, and we hold
$k_{\omega}$ fixed to a constant value of $0.1,$ and $\omega_{0}$ fixed to the
halfway point between $\omega_{\rm lo}$ and $\omega_{\rm hi}.$ We refer the
reader to the relevant section of our source code for additional details of
our implementation.
## Appendix D SHAMNet Definition and Training
In this appendix, we describe our procedure for training SHAMNet,
$\mathcal{F}_{\rm SHAM}(M_{\rm peak},\theta|\psi).$ We remind the reader that
in this notation, the parameters $\psi$ refer to the weights and biases of the
neural network described in detail in this section, and as described in §3.2,
the parameters $\theta$ control the behavior of the three ingredients needed
to define SHAM:
1. 1.
$\phi_{\rm g}(M_{\star}|\theta),$ the galaxy stellar mass function,
2. 2.
$\phi_{\rm h}(M_{\rm peak}|\theta),$ the subhalo mass function described in
§A,
3. 3.
$\sigma(M_{\rm peak}|\theta),$ the halo mass-dependent scatter.
As outlined in §2, for a particular choice of $\theta,$ SHAM is defined to be
the scaling relation, $\langle M_{\star}|M_{\rm peak};\theta\rangle_{\rm
med},$ that gives rise to the target stellar mass function, $\phi_{\rm
g}(M_{\star}|\theta),$ when applied to the subhalo population, $\phi_{\rm
h}(M_{\rm peak}|\theta),$ in the presence of (log-normal) scatter,
$\sigma(M_{\rm peak}|\theta).$ The goal of training SHAMNet is to identify the
parameters $\psi$ such that for all physically relevant values of $\theta,$
the function $\mathcal{F}_{\rm SHAM}(M_{\rm peak},\theta|\psi)=\langle
M_{\star}|M_{\rm peak};\theta\rangle_{\rm med}$ satisfies the SHAM-defined
relationship between $\phi_{\rm g},$ $\phi_{\rm h},$ and $\sigma.$ The nature
of this objective makes training $\mathcal{F}_{\rm SHAM}(M_{\rm
peak},\theta|\psi)$ different from a standard problem in neural network
regression, since there does not exist a target function whose output defines
the training data; instead, the parameters $\psi$ will be considered optimal
when the SMF that emerges from $\langle M_{\star}|M_{\rm
peak};\theta\rangle_{\rm med}$ agrees with $\phi_{\rm g}(M_{\star}|\theta)$ to
the desired precision.
We begin in §D.1 with a description of the design of the neural network we use
for $\mathcal{F}_{\rm SHAM}.$ Our training then proceeds in two phases. In the
first phase described in §D.2, we tune the network parameters $\psi$ so that
$\mathcal{F}_{\rm SHAM}(M_{\rm peak},\theta|\psi)$ closely agrees with a
parametric form for the stellar-to-halo mass relation that has been tuned in
advance to approximately solve Eq. 2. After this initial phase, we optimize
the parameters $\psi$ using the training procedure described in §D.3. We will
provide a reasonably comprehensive description of this two-phase procedure in
this appendix, but we refer the reader to
https://github.com/ArgonneCPAC/shamnet for all the quotidian details.
### D.1 Architecture
In this section, we describe the architecture of the neural network we use for
SHAMNet, which provides a mapping
$\mathcal{F}_{\rm SHAM}:\\{\theta,M_{\rm peak}\\}\rightarrow M_{\star}$
based on a simply-connected multi-layer perceptron (MLP). For the input
variables $\theta,$ we use a $2$-parameter Schechter function to capture
variations in the galaxy stellar mass function, and a 2-parameter sigmoid to
describe the halo mass dependence of scatter, adopting the forms described in
§3.2. For the subhalo mass function, we use the parametric model defined in
§A, but in this case we hold these parameters fixed to the values calibrated
to approximate the $z=0$ mass function in the Bolshoi simulation. Thus in
addition to $M_{\rm peak},$ SHAMNet accepts a 5-dimensional parameter,
$\theta,$ and returns stellar mass, $M_{\star},$ so that $\mathcal{F}_{\rm
SHAM}:\mathbb{R}^{5}\rightarrow\mathbb{R}^{1}.$
Our neural network $\mathcal{F}_{\rm SHAM}$ thus has an input layer of 5
nodes, one for each dimension of $\\{\theta,M_{\rm peak}\\},$ and an output
layer of a single node, $M_{\star}.$ Between the inputs and outputs, we use 4
hidden layers composed of $64,32,8,$ and $4$ nodes, respectively. Each node in
a layer is connected to every node in the previous layer, so that all 4 of our
hidden layers are dense. As with all simply-connected MLPs, the computations
performed by the $i^{\rm th}$ node of a layer, $y_{\rm i},$ consists of two
successive operations performed on the outputs of all the nodes of the
previous layer, $x_{\rm j}.$ The first set of operations consists of a linear
transformation,
$\displaystyle y_{\rm i}=\sum_{\rm j}W_{\rm ij}x_{\rm j}+b_{\rm i},$ (29)
and the second operation is a nonlinear transformation, $\Upsilon_{\rm act},$
that is independently applied to the each node $y_{\rm i}$ in the layer; the
result $\Upsilon_{\rm act}(y_{\rm i})$ is then fed as an input to each node in
the next layer. For the activation function $\Upsilon_{\rm act}$ in every node
of each dense hidden layer, we use a SELU function,
$\displaystyle\Upsilon_{\rm act}(z)=\lambda\begin{cases}z&\mbox{if }z>0\\\
\alpha e^{z}-\alpha&\mbox{if }z\leq 0\end{cases},$ (30)
where $\lambda=1.05$ and $\alpha=1.67$ are chosen so that the mean and
variance of the inputs are preserved between two consecutive layers (Klambauer
et al., 2017). For the activation function applied to the output layer, we use
the sigmoid function defined in Eq. 21 to enforce that the returned stellar
mass is bounded within the range
$10^{-5}M_{\odot}<M_{\star}<10^{25}M_{\odot}.$ Our MLP is implemented in the
stax subpackage of the JAX library.
### D.2 Initialization
Our goal for training SHAMNet is to identify the weights, $W_{\rm ij},$ and
biases, $b_{\rm j},$ collectively represented with the variable $\psi,$ such
that $\mathcal{F}_{\rm SHAM}$ provides the desired mapping from $M_{\rm
peak}\rightarrow M_{\star}$ for any physically relevant parameters, $\theta.$
We begin our procedure for optimizing $\psi$ by generating approximations to
the desired mapping, $\langle M_{\star}|M_{\rm peak};\theta\rangle_{\rm med},$
for a large collection of $\theta;$ the collection of $\theta$ and associated
mappings will serve as training data for the first phase of optimizing
$\mathcal{F}_{\rm SHAM}.$ Thus at the end of this phase, we will have a
collection of weights and biases, $\psi_{\rm init},$ that roughly provides the
desired scaling relation, $\langle M_{\star}|M_{\rm peak};\theta\rangle_{\rm
med},$ for each $\theta$ in the training set. The point in parameter space
$\psi_{\rm init}$ will be used as the starting point for the optimization
procedure described in §D.3, and so we refer to calculations described in the
present section as the initialization phase of our training.
As described in §3, the parameters $\theta$ define a specific combination of
$\phi_{\rm g}(M_{\star}|\theta),\phi_{\rm h}(M_{\rm peak}|\theta),$ and
$\sigma(M_{\rm peak}|\theta);$ for each such combination, we seek to determine
the abundance matching relation, $\langle M_{\star}|M_{\rm peak}\rangle_{\rm
med},$ that provides a solution to Eq. 2. In the initialization phase of our
training, we assume that the parametric form $m_{\star}(m_{\rm h}|\omega),$ as
defined by Eq. 26, is sufficiently flexible to supply an adequate stellar-to-
halo mass relation, and we train SHAMNet to reproduce this parameterized
relation. Thus in order to generate training data for the initialization phase
of our training, we need to generate a large collection of pairs,
$\\{\theta,\omega_{\theta}\\},$ where the values of $\theta$ span the
physically relevant range, and the accompanying value $\omega_{\theta}$
represents an approximate solution to Eq. 2. To generate the collection of
pairs, $\\{\theta,\omega_{\theta}\\},$ we randomly draw a value of $\theta$
from an observationally relevant range, and for each random draw, we use the
gradient descent technique described in Appendix C to find a point
$\omega_{\theta}$ that optimally solves Eq. 2.
Since $\theta=\\{\phi_{\ast},\alpha,y^{\sigma}_{\rm lo},y^{\sigma}_{\rm
hi}\\},$ then we have four dimensions for which we need to define an
observationally relevant range. The last two dimensions control the level of
$M_{\rm halo}$-dependent scatter, which we allow to vary independently within
the open interval, (0.1, 0.5). The first two variables define the variations
in the SMF, which for purposes of this paper, we wish to be reasonably close
to SDSS measurements of the low-redshift universe (e.g., Li & White, 2009). To
determine the relevant range of the SMF parameters, we define an SDSS-like
SMF, $\phi_{\rm SDSS}(M_{\star}),$ by selecting fiducial values
$\phi_{\ast}=0.005\ {\rm Mpc^{-3}M_{\odot}^{-1}},$ and $\alpha=-1.06.$ Using
emcee (Foreman-Mackey et al., 2019), we run an MCMC to determine posteriors on
the parameters $\phi_{\ast}$ and $\alpha,$ assuming a Gaussian likelihood with
a diagonal covariance matrix defined by $M_{\star}$-independent uncertainty of
$0.1$ dex on $\phi_{\rm SDSS}(M_{\star}).$ The results of this MCMC supply a
Gaussian-like distribution of points in our SMF parameter space, and we fit
this posterior distribution with a two-dimensional Gaussian distribution,
$\mathcal{N}(\mu,{\rm Cov(}\phi_{\ast},\alpha)).$ When generating training
data for SHAMNet, we sample SMF parameters by randomly draw pairs
$\\{\phi_{\ast},\alpha\\}$ based on a Latin Hypercube, where the axes of the
Latin Hypercube are aligned with the eigenvectors of ${\rm
Cov(}\phi_{\ast},\alpha),$ and span a $5\sigma$ length in each direction; any
values with $\alpha<1$ correspond to non-monotonic SMFs, and are
discarded.555We use the pyDOE2 package444https://github.com/clicumu/pyDOE2 in
our random sampling based on a Latin Hypercube (McKay et al., 1979; Iman et
al., 1981).
For each value of $\theta$ generated by this sampling method, we identify a
best-fitting stellar-to-halo mass relation, $\omega_{\theta},$ using the
optimization techniques described in §C.2; the quality of each fit is
quantified by $\mathcal{L}_{\rm MSE}(\omega,\theta),$ as defined in Eq. C.2.
For some points $\theta,$ the best-fitting value $\omega_{\theta}$ produces
only a rough recovery of the target stellar mass function, $\phi_{\rm
g}(M_{\star}|\theta).$ In the initialization phase of training SHAMNet, we
discard all pairs $\\{\theta,\omega_{\theta}\\}$ with $\mathcal{L}_{\rm
MSE}>0.05,$ which we find is about 5% of the sampling data. While this
rejection rate implies that our training set will not fairly sample the full,
observationally relevant region of parameter space as defined above, this is
rather harmless because the primary purpose of this procedure is just to
generate some collection $\\{\theta,\omega_{\theta}\\}$ used to identify
$\psi_{\rm init}.$ As described in §D.1, the design of SHAMNet has no reliance
upon the functional form of the stellar-to-halo mass relation defined in
Appendix C, $m_{\star}(m_{\rm h}|\omega)$, and so rejected points
$\omega_{\theta}$ simply represent an inflexibility of the three-roll SMHM
functional form, but this will not impact SHAMNet as we will not reject such
pairs $\\{\theta,\omega_{\theta}\\}$ after $\psi_{\rm init}$ has been
identified.
The goal of the initialization phase of our training is to identify a set of
weights and biases, $\psi_{\rm init},$ such that the behavior of our neural
network, $\mathcal{F}_{\rm SHAM},$ gives an accurate approximation of the
parametric function defined in Appendix C. To achieve this goal, we seek to
identify the point $\psi_{\rm init}$ that minimizes the following cost
function:
$\displaystyle\mathcal{L}_{\rm MSE}(\theta,\omega)$ $\displaystyle\equiv$
$\displaystyle\frac{1}{N}\sum_{i}\left(y^{\rm i}(\theta)-x^{\rm
i}(\theta,\omega)\right)^{2}$ $\displaystyle y^{\rm i}(\theta)$
$\displaystyle=$ $\displaystyle m_{\star}(m_{\rm h}|\omega_{\theta})$ (31)
$\displaystyle x^{\rm i}(\theta,\omega)$ $\displaystyle=$
$\displaystyle\mathcal{F}_{\rm SHAM}(M_{\rm peak},\theta|\phi)$
In evaluating $\mathcal{L}_{\rm MSE}(\theta,\omega),$ we used batch sizes of
$N=50$ pairs of points, $\\{\theta,\omega_{\theta}\\},$ that randomly sample
the parameter space of $\theta$ using the procedure outlined above. We used
the implementation of the Adam optimizer in JAX with a step-size parameter of
0.001 to identify $\psi_{\rm init}$ through 1500 gradient evaluations, which
we found through experimentation to be sufficient for the network to achieve
the level of accuracy required for this initializing phase of training the
network.
### D.3 SHAMNet Training
The goal of the final phase of our training is to identify a set of weights
and biases, $\psi_{\rm SHAM},$ such that the behavior of our neural network,
$\mathcal{F}_{\rm SHAM},$ supplies a stellar-to-halo mass relation that
accurately solves the defining equation of abundance matching, Eq. 2. We note
that if the parametric function $m_{\star}(m_{\rm h}|\omega)$ defined in
Appendix C were sufficiently flexible, then this goal would already be
achieved via the identification of $\psi_{\rm init}.$ However, as described in
Appendix D.2, roughly 5% of points in the observationally relevant range have
SMFs that cannot be described to high accuracy using a stellar-to-halo mass
relation defined by this functional form. Thus in the final training phase of
SHAMNet, we no longer rely on training data based on the parametric function
$m_{\star}(m_{\rm h}|\omega).$ Instead, we seek to identify the point
$\psi_{\rm SHAM}$ that directly minimizes the difference between the predicted
and target SMF, and so we minimize a cost function that directly quantifies
this difference:
$\displaystyle\mathcal{L}_{\rm MSE}(\theta)$ $\displaystyle\equiv$
$\displaystyle\frac{1}{N}\sum_{i}\left(y^{\rm i}(\theta)-x^{\rm
i}(\theta)\right)^{2}$ $\displaystyle y^{\rm i}(\theta)$ $\displaystyle=$
$\displaystyle\mathcal{R}(\phi_{\rm g}(M_{\star,i}|\phi_{\ast},\alpha))$ (32)
$\displaystyle x^{\rm i}(\theta,\phi)$ $\displaystyle=$
$\displaystyle\mathcal{R}(\phi_{\rm g}(M_{\star,i}|\theta,\phi)).$
Equation D.3 is very similar to Eq. C.2, where the latter defined the cost
function that we minimized in order to identify the parametric function
$m_{\star}(m_{\rm h}|\omega_{\theta})$ that solves the defining equation of
abundance matching for a particular value of $\theta.$ In fact, in both
equations the values of $y^{\rm i}(\theta)$ that specify the target values of
the SMF are the same, and in both cases, the predicted values of the SMF,
$x^{\rm i}(\theta),$ are derived using the same techniques detailed in
Appendix B for calculating the SMF in the presence of scatter via an
analytical expression for the stellar-to-halo mass relation. But in evaluating
the cost function defined by Eq. D.3, here our predicted values $x^{\rm
i}(\theta)$ come directly from $\mathcal{F}_{\rm SHAM}(M_{\rm
peak},\theta|\psi).$
We use the Adam optimizer with a step size parameter of 0.001 to minimize
$\mathcal{L}_{\rm MSE}(\theta),$ where each of our $10^{5}$ cost function
evaluations used a batch size of $N=50$ points that span the parameter space
of $\theta,$ again using the same sampling method detailed in §D.2. The end
result of this procedure is the identification of $\psi_{\rm SHAM},$ which
defines the behavior of the $\mathcal{F}_{\rm SHAM}$ function whose
performance is illustrated in Figure 3.
## Appendix E Triweight Kernel Convolutions
Many of the calculations in this paper involve convolutions of log-normal
distributions such as the one appearing in Eq. 2. In this section, we review
an alternative to Gaussian convolution based on the triweight kernel,
$\mathcal{K},$ defined as:
$\displaystyle\mathcal{K}(z)\equiv\begin{cases}\frac{35}{96}\left[1-(z/3)^{2}\right]^{3},&-3\sigma\leq
x\leq 3\sigma\\\ 0,&\text{otherwise}\end{cases}$ (33)
where $z\equiv(x-\mu)/\sigma.$ The comparison to a Gaussian is shown in Figure
9. The two distributions have the same first and second moments, $\mu$ and
$\sigma,$ but differ in their higher-order moments.
Figure 9: Triweight convolution kernel. The blue curve shows the probability
density function of a typical log-normal distribution in stellar mass,
$\mathcal{N}(\mu=10,\sigma=0.25);$ the orange curve shows the triweight
kernel, $\mathcal{K},$ defined in Eq. 33. The triweight function
$\mathcal{K}(x)$ vanishes at points beyond $\mu\pm 3\sigma,$ is differentiable
for all $x,$ and is highly performant on GPUs.
The function $\mathcal{K}(x)$ has several properties that make it convenient
for our purposes. As described in §4, differentiable predictions for
cosmological summary statistics involve weights, $w_{\rm halo},$ computed by
integrating some assumed PDF (typically a log-normal) across some bin(s) of
stellar mass. Once $w_{\rm halo}$ has been computed for every simulated
subhalo, various point-estimators are applied to the population of synthetic
galaxies in order to make predictions for summary statistics.
Because Gaussians are everywhere non-zero, computing weighted two-point
summary statistics using exact log-normals can be expensive due to the need to
keep track of a large number of tiny contributions. One simple technique to
mitigate this computational expense is to impose a clip at some multiple of
$\sigma.$ This gains back performance at the cost of some (controllably small)
roundoff error, but results in discontinuity in the weights at $x=\mu\pm
3\sigma.$ On the other hand, the triweight kernel is a $C^{\infty}$ function
on the real line, and points with $|x-\mu|>\pm 3\sigma$ contribute formally
zero weight, and so can be neglected from two-point calculations without
consequence. Moreover, computations of predictions based on Gaussians require
special-function evaluations that can be far slower on GPU accelerator devices
in comparison to the small number of elementary arithmetical operations
required to evaluate Eq. 33.
In computing predictions based on $\mathcal{K}$ rather than $\mathcal{N},$ we
are, in effect, proposing a slightly different probability distribution
$P(M_{\star}|M_{\rm halo})$ than the traditional log-normal. However, the
difference between models distinguished only by $\mathcal{K}$ and
$\mathcal{N}$ is likely to be observationally immaterial, as it is already
challenging to obtain tight constraints on the second moment
$P(M_{\star}|M_{\rm halo}),$ which is the same in the two distributions.
## Appendix F Simulation-Based $\Delta\Sigma$ Predictions
In this section, we derive the equations associated with our calculation of
$\Delta\Sigma(R),$ the excess surface mass density profile of a sample of
points in a cosmological simulation. We highlight that the computation
described below is exact, can be used to calculate the lensing of either
galaxies or halos simulated with or without hydrodynamics, and applies on all
spatial scales resolved by the simulation, including the deeply nonlinear
regime. Our approach also facilitates a highly convenient pre-computation of
the lensing profile on a per-object basis, so that galaxy-halo model
predictions for the lensing of a stacked sample can be calculated simply as a
masked or weighted sum over the pre-computed profiles.
We begin by considering the gravitational lensing at a projected distance $R$
from a single point mass, ${\rm m_{p}}.$ The surface density profile,
$\Sigma(R),$ satisfies:
$\displaystyle\int_{R_{\rm min}}^{R_{\rm max}}\Sigma(R)2\pi R{\rm
d}R=\begin{cases}{\rm m_{p}},&\text{if}\ R_{\rm min}<R<R_{\rm max}\\\
0,&\text{otherwise}\end{cases}$
Using this expression, we can calculate $\langle\Sigma(R)\rangle,$ the average
surface density in an annulus with inner radius $R_{\rm min}$ and outer radius
$R_{\rm max}:$
$\displaystyle\langle\Sigma(R)\rangle$ $\displaystyle\equiv$
$\displaystyle\frac{\int_{R_{\rm min}}^{R_{\rm max}}\Sigma(R)2\pi R{\rm
d}R}{\int_{R_{\rm min}}^{R_{\rm max}}2\pi R{\rm d}R}$ (34) $\displaystyle=$
$\displaystyle\begin{cases}{\rm m_{p}}/A_{\rm ann},&\text{if}\ R_{\rm
min}<R<R_{\rm max}\\\ 0,&\text{otherwise}\end{cases}$
where $A_{\rm ann}=\pi(R_{\rm max}^{2}-R_{\rm min}^{2})$ is the area of the
annulus.
We can similarly compute $\bar{\Sigma}(<R)={\rm m_{p}}/\pi R^{2}$ averaged
over the same annulus:
$\displaystyle\langle\bar{\Sigma}(<R)\rangle$ $\displaystyle\equiv$
$\displaystyle\frac{1}{A_{\rm ann}}\int_{R_{\rm min}}^{R_{\rm
max}}\bar{\Sigma}(<R)2\pi R{\rm d}R,$
which reduces to
$\displaystyle\langle\bar{\Sigma}(<R)\rangle=\frac{2{\rm m_{p}}}{A_{\rm
ann}}\ln(R_{\rm max}/a),$ (35)
where the value of $a$ depends on whether the point mass is located inside the
annulus:
$\displaystyle a=\begin{cases}R,&R>R_{\rm min}\\\ R_{\rm min},&R<R_{\rm
min}\end{cases}$ (36)
Using Eqs. 34 & 35 together with Eq. 9, we can calculate the value of
$\Delta\Sigma$ averaged over the annulus:
$\displaystyle\langle\Delta\Sigma(R)\rangle$ $\displaystyle\equiv$
$\displaystyle\langle\bar{\Sigma}(<R)-\Sigma(R)\rangle$ $\displaystyle=$
$\displaystyle\frac{{\rm m_{p}}}{A_{\rm ann}}G(R,R_{\rm min},R_{\rm max}),$
where
$\displaystyle G$ $\displaystyle=$ $\displaystyle\begin{cases}2\ln(R_{\rm
max}/R_{\rm min}),&R<R_{\rm min}\\\ 2\ln(R_{\rm max}/R)-1,&R_{\rm
min}<R<R_{\rm max}\\\ 0,&R_{\rm max}<=R.\end{cases}$
In order to calculate the average lensing profile
$\langle\Delta\Sigma(R)\rangle$ about an individual point in a N-body
simulation, Equation F generalizes in the obvious way as a sum over the point-
masses used to trace the simulated density field. For a simulated snapshot at
redshift $z_{\rm snap},$ if the cosmological distance $D(z_{\rm snap})$
satisfies $D(z_{\rm snap})\gg R_{\rm max},$ then we can use any of the
Cartesian axes of the snapshot as the line-of-sight, and the remaining two
axes to define $R.$
For a fixed choice of $R_{\rm min}$ and $R_{\rm max},$ the average lensing
profile $\langle\Delta\Sigma\rangle$ can be computed once and for all about
every subhalo in a simulated snapshot. Once tabulated, predicting the lensing
produced by a stack of subhalos can be computed simply by averaging over the
pre-computed values of $\langle\Delta\Sigma\rangle$ for each member of the
stack (or alternatively, by calculating an average that has been weighted by a
continuously-valued sample selection function, as in §4.2).
|
# Optimal Learning Rates for Regularized Least-Squares with a Fourier Capacity
Condition ††thanks: Citation: Authors. Title. Pages…. DOI:000000/11111.
Prem Talwai
Operations Research Center
MIT David Simchi-Levi
Institute for Data, Systems, and Society
MIT
###### Abstract
We derive minimax adaptive rates for a new, broad class of Tikhonov-
regularized learning problems in Hilbert scales under general source
conditions. Our analysis does not require the regression function to be
contained in the hypothesis class, and most notably does not employ the
conventional a priori assumptions on kernel eigendecay. Using the theory of
interpolation, we demonstrate that the spectrum of the Mercer operator can be
inferred in the presence of “tight” $L^{\infty}$ embeddings of suitable
Hilbert scales. Our analysis utilizes a new Fourier capacity condition, which
characterizes the optimal Lorentz range space of a modified Mercer operator in
certain parameter regimes.
## 1 Introduction
Consider the classical learning problem:
$Y=f^{*}(X)+\epsilon$ (1)
where we wish to estimate the regression function $f^{*}$ in the presence of
additive noise $\epsilon$ using a dataset
$\mathcal{D}=\\{(X_{1},Y_{1}),(X_{2},Y_{2}),\ldots(X_{n},Y_{n})\\}$ sampled
i.i.d from a distribution $P$ over $\mathcal{X}\times\mathcal{Y}$. Here, we
focus on kernel-based regularized least-squares (ridge) regression, i.e. we
wish to solve the optimization problem
$f_{\lambda,D}=\text{arg}\min_{f\in\mathcal{H}_{K}}\frac{1}{n}\sum_{i=1}^{n}(Y_{i}-f(X_{i}))^{2}+\lambda||f||^{2}_{\mathcal{K}}$
(2)
where $\mathcal{H}_{K}\subset L^{2}(P_{\mathcal{X}})$ is a reproducing kernel
Hilbert space (RKHS) with kernel $K(\cdot,\cdot)$.
The performance of a regularization scheme is typically quantified using the
expected risk:
$R(f_{D,\lambda})=||f_{D,\lambda}-f^{*}||^{2}_{L^{2}(P)}$ (3)
This problem has been studied extensively in the statistics and machine
learning literature [17, 8, 4, 36]. These works primarily focus on
establishing optimal convergence rates for $R(f_{D,\lambda})$ under light-
tailed noise (typically subgaussian or subexponential) and various source
conditions, which characterize the regularity/smoothness of the learning
target $f^{*}$. The two main approaches for this task have included the
integral operator technique (e.g. [17, 4, 8]) and empirical process technique
[41, 31]. While the latter approach enjoys easy adaptability to general convex
losses, the former technique more directly exploits the Hilbertian structure
of the hypothesis space, and aligns seamlessly with the method of real
interpolation between $\mathcal{H}_{K}$ and $L^{2}(P_{\mathcal{X}})$ (see
section 2.1).
We consider the hard-learning scenario, where $f^{*}\not\in\mathcal{H}$. This
setting has been treated in previous works [17, 24, 4, 37, 36] both for the
Hölder source condition [17, 4] where $f^{*}$ lies in some interpolation space
$[L^{2}(P_{\mathcal{X}}),\mathcal{H}_{K}]_{\theta,2}$ (for $\theta\in(0,1)$)
and the general source condition [37] where $f^{*}$ lies in a suitable Hilbert
scale (defined in section 2.1), which correspond to more general interpolation
spaces between $L^{2}(P_{\mathcal{X}})$ and $\mathcal{H}_{K}$ obtained using a
function parameter. While these works treat the problem of misspecification,
their analysis hinges on a kernel eigendecay assumption, which characterizes
the capacity of the hypothesis space $\mathcal{H}_{K}$ by the decay rate of
the eigenvalues of the kernel Mercer operator. While this spectral assumption
has been ubiquitous in the statistical analysis of kernel ridge regression
[26, 4, 8], it depends crucially on the data-generating measure $P$, which is
typically unknown in practice.
In this paper, we replace the classical kernel eigendecay condition with a new
Fourier capacity condition which characterizes the decay of the Fourier power
spectrum of the kernel ($P$-independent) as opposed to its Mercer spectrum
($P$-dependent). Indeed, our capacity assumption relates to $P$ only through
the eigenbasis of the Mercer operator, which may be shared by several
different kernels and is often independent of the specific interplay between
$\mathcal{H}_{K}$ and $P$ (see discussion in section 2.4). Fourier analytic
approaches have been applied previously to study the capacity of the
hypothesis class [51, 49, 40]. However, these works typically either assume a
uniform (Lebesgue-equivalent) data-generating measure or particular RKHS
structure (i.e a Sobolev space) in order to apply Fourier transform
techniques. In this paper, we demonstrate that Fourier techniques remain
powerful over general probability spaces, in the presence of the embedding of
a suitable Hilbert scale in $L^{\infty}(\mathcal{X})$. Indeed, we demonstrate
that if the embedding of an intermediate Hilbert scale is sufficiently “sharp”
(characterized by the optimality of its index function in “covering” the range
of a modified Mercer operator), then a Gagliardo-Nirenberg type inequality
between the scales enables us to estimate kernel eigendecay via the Bernstein
widths of the RKHS in $L^{\infty}(\mathcal{X})$. To the best of our knowledge,
this is the first work to establish minimax learning rates under general
source conditions without appealing to any direct assumptions on the
eigendecay or effective dimension of the Mercer operator.
## 2 Preliminaries
We recall our learning problem:
$f_{\lambda,D}=\text{arg}\min_{f\in\mathcal{H}_{K}}\frac{1}{n}\sum_{i=1}^{n}(Y_{i}-f(X_{i}))^{2}+\lambda||f||^{2}_{\mathcal{K}}$
where $\mathcal{H}_{K}$ be a separable reproducing kernel Hilbert space (RKHS)
on $\mathcal{X}$ (see e.g. [41] for a definition). A solution to this problem
can be computed in closed form:
$f_{\lambda,D}=(C_{D}+\lambda)^{-1}g_{D}$
where $C_{D}=\mathbb{E}_{\mathcal{D}}[k(X,\cdot)\otimes k(X,\cdot)]$ is the
empirical covariance, and $g_{D}=\sum_{i=1}^{n}y_{i}k(x_{i},\cdot)$. Before
discussing our model framework and results, we begin with an overview of some
mathematical preliminaries required for this section.
### 2.1 Hilbert Scales
Suppose the imbedding $I_{\nu}:\mathcal{H}_{K}\to L^{2}(\nu)$ of
$\mathcal{H}_{K}$ into $\mathcal{L}^{2}(\nu)$ is injective (here $\nu=P_{X}$
is the marginal on $\mathcal{X}$). Let $S_{\nu}=I^{*}_{\nu}$ be its adjoint.
Then, it can be shown that $S_{\nu}$ is an integral operator given by:
$S_{\nu}f(x)=\int_{\mathcal{X}}k(x,\cdot)f(y)d\nu(y)$ (4)
Using $S_{\nu}$ and $I_{\nu}$, we construct the following positive self-
adjoint operators on $\mathcal{H}_{K}$ and $\mathcal{L}^{2}(\nu)$,
respectively:
$\displaystyle C_{\nu}$ $\displaystyle=S_{\nu}I_{\nu}=I^{*}_{\nu}I_{\nu}$
$\displaystyle T_{\nu}$ $\displaystyle=I_{\nu}S_{\nu}=I_{\nu}I^{*}_{\nu}$
We observe that $C_{\nu}$ and $T_{\nu}$ are nuclear (see Lemma 2.2/2.3 in
[42]) Since, $T_{\nu}$ is nuclear and self-adjoint, it admits a spectral
representation:
$T_{\nu}=\sum_{j=1}^{\infty}\mu_{j}e_{j}\langle
e_{j},\cdot\rangle_{L^{2}(\nu)}$
where $\\{\mu_{j}\\}_{j=1}^{\infty}\in(0,\infty)$ are nonzero eigenvalues of
$T_{\nu}$ (ordered nonincreasingly) and $\\{e_{j}\\}_{j=1}^{\infty}\subset
L^{2}(\nu)$ form an orthonormal system of corresponding eigenfunctions. Note
that formally, the elements $e_{j}$ of $L^{2}(\nu)$ are equivalence classes
$[e_{j}]_{\nu}$ whose members only differ on a set of $\nu$-measure zero—
notationally, we consider this formalism to be understood here and simply
write $e_{j}$ to refer to elements in both $\mathcal{H}_{K}$, $L^{2}(\nu)$,
and their interpolation spaces (with the residence of $e_{j}$ understood from
context). Given a nonincreasing index function
$\phi:(0,||T_{\nu}||)\to\mathbb{R}_{+}$, we define the Hilbert scales
$\mathcal{H}^{\phi}_{K}$ as [30]:
###### Definition 2.1.
Let $\phi:(0,||T_{\nu}||)\to\mathbb{R}_{+}$ be nondecreasing, continuous, with
$\phi(0)=0$. Then, the Hilbert scale $\mathcal{H}^{\phi}_{K}$ is the
completion of the space:
$\mathcal{H}^{\phi}_{K}=\Big{\\{}f\in
L^{2}(\nu):\sum_{i=1}^{\infty}\frac{\langle
f,e_{i}\rangle_{L^{2}(\nu)}^{2}}{\phi(\mu_{i})}<\infty\Big{\\}}$
with respect to the inner product $\langle
f,g\rangle_{\phi}=\sum_{i}\frac{\langle f,e_{i}\rangle\langle
g,e_{i}\rangle}{\phi(\mu_{i})}$.
It is easy to see that
$\mathcal{H}^{\phi}_{K}\cong\text{ran}(\phi^{\frac{1}{2}}(T_{\nu}))$. The
subscript $K$ in $\mathcal{H}^{\phi}_{K}$ reflects that $\phi$ is acting on
the spectrum of $K$, when the kernel is fixed and understood from context, we
will omit this subscript and simply denote the Hilbert scale
$\mathcal{H}^{\phi}$.
We define:
$\|k^{\phi}\|^{2}_{\infty}=\sup_{x\in\mathcal{X}}\sum_{i=1}^{\infty}\phi(\mu_{i})e_{i}^{2}(x)$
where we allow $\|k^{\phi}\|_{\infty}=\infty$. If
$\|k^{\phi}\|_{\infty}<\infty$, then it is easy to show that
$\mathcal{H}^{\phi}_{K}$ is continuously embedded in $L^{\infty}(\mathcal{X})$
with norm $\|k^{\phi}\|_{\infty}$ (see e.g. Theorem 9 in [17] for the case of
Holder source conditions).
The Hilbert scale $\mathcal{H}^{\phi}_{K}$ generalizes the notion of RKHS
interpolation (power) spaces discussed in [42], which result from the specific
choice of $\phi(t)=t^{\alpha}$ for some $\alpha\in(0,1)$. Intuitively, the
Hilbert scales can be viewed as a nonlinear transformation of the infinite-
dimensional RKHS ellipsoid $B(\mathcal{H}_{K})$, realized by transforming the
axes lengths with the index function $\phi$. As we will see later, the growth
properties of $\phi$ play a fundamental role in specifying the smoothness of
the elements of $\mathcal{H}^{\phi}_{K}$.
### 2.2 Rearrangement Invariant Spaces
Let $(R,\mu)$ be a $\sigma$-finite measure space. For any positive increasing
function $\phi:(0,\infty)\to(0,\infty)$, we define the dilation function
$d_{\phi}:(0,\infty)\to(0,\infty)$ as:
$d_{\phi}(t)=\sup_{s\in(0,\infty)}\frac{\phi(st)}{\phi(s)}$ (5)
We define the extension indices:
$\displaystyle\alpha_{\phi}$ $\displaystyle\equiv\lim_{t\to 0}\frac{\log
d_{\phi}(t)}{\log t}$ $\displaystyle\beta_{\phi}$
$\displaystyle\equiv\lim_{t\to\infty}\frac{\log d_{\phi}(t)}{\log t}$
If $\phi$ has finite extension indices (i.e.
$-\infty<\alpha_{\phi}<\beta_{\phi}<\infty$), we define the Lorentz endpoint
space $\Lambda_{\phi}$ and Marcinkiewicz endpoint space $M_{\phi}$ via the
(quasi)-norms:
$||g||_{M_{\phi}}=\sup_{t\in(0,\mu(R))}\phi(t)g^{*}(t)\hskip
11.38109pt||g||_{\Lambda_{\phi}}=\int_{0}^{\mu(R)}g^{*}(t)d\phi(t)$
where $g^{*}$ denotes the decreasing rearrangement of $g$ (see e.g. [34]) for
a definition). Note, in the definition of $M_{\phi}$, we follow [35], rather
than the more standard convention in [34] (which replaces $g^{*}$ by the
maximal decreasing rearrangement), in order to allow for quasi-normed spaces
(under suitable growth conditions on $\phi$, which we demonstrate later on,
these two definitions are equivalent by Proposition 7.10.5 in [34]).
Every rearrangement invariant space $X$ with fundamental function $\phi$ lies
between $\Lambda_{\phi}$ and $M_{\phi}$, i.e. $\Lambda_{\phi}\subset X\subset
M_{\phi}$ (Propositions 7.10.6 and 7.10.15 in [34]). For $p\geq 1$, we hence
define the $p$-Lorentz spaces $\Lambda^{p}_{\phi}$ as:
$\Lambda^{p}_{\phi}=\Big{\\{}g:\int_{0}^{\mu(R)}(g^{*}(t))^{p}d\phi^{p}(t)<\infty\Big{\\}}$
(6)
A prototypical example of this scale are the “classical” Lorentz spaces:
$\Lambda_{t^{\frac{1}{p}}}(\mathbb{R}^{d})\equiv
L^{p,1}(\mathbb{R}^{d})\subset L^{p,q}(\mathbb{R}^{d})\subset
L^{p,\infty}(\mathbb{R}^{d})\equiv M_{t^{\frac{1}{p}}}$ for $q\in(1,\infty)$.
For a continuous, increasing function $\Phi$, we may further define the
Orlicz-Lorentz spaces $L^{\Phi,w}$ [19] containing measurable functions for
which the quantity:
$||f||_{\Phi,w}=\inf\Big{\\{}\lambda>0:\int_{0}^{\infty}\Phi\Big{(}\frac{f^{*}(t)}{\lambda}\Big{)}w(t)dt\leq
1\Big{\\}}$ (7)
is finite (where $w(t)$ is a decreasing weight function). Note that
$L^{\Phi,w}$ in (7) (following [19]) are sometimes known as weighted Orlicz
spaces, with Orlicz-Lorentz spaces instead being defined with respect to the
homogeneous measure $\frac{dt}{t}$ (e.g. [2]). This vernacular difference
significantly impacts the interpolation properties of the Lorentz-Orlicz space
(as elaborated below). More precisely, the latter definition enables the space
to be characterized as an “ultrasymmetric” space intermediate between Lorentz
and Marcinkiewicz endpoint spaces, while this characterization is rarely
possible with the formulation in (7) (see [2]). We use the formulation in (7)
for the added generality provided by the nonincreasing weight function $w(t)$.
When, $w=1$, we simply write $L^{\Phi}$.
It is also important to note that we have again not restricted $\Phi$ to be an
Orlicz or Young function. Hence, (7) only defines a quasi-norm. In Lemma 2 in
the appendix, we derive a fundamental function for for $L^{\Phi,w}$ (up to
equivalence) when the extension indices $\alpha_{\Phi}$ and $\beta_{\Phi}$
coincide.
Finally, we say that a function $\xi$ satisfies $\Delta_{2}$ condition if
there exists constants $D_{1},D_{2}>0$, such that:
$D_{1}\xi(\lambda)\leq\xi(2\lambda)\leq D_{2}\xi(\lambda)\hskip
5.69054pt\forall\lambda>0$ (8)
### 2.3 Approximation Widths
We review some standard widths in approximation theory that describe the
compactness of spaces in various norms. The $n^{\text{th}}$ entropy number
$\epsilon_{n}(\mathcal{X})$ of a compact set $\mathcal{X}$, is informally the
smallest radius $\epsilon$ such that there exists an $\epsilon$-covering of
$\mathcal{X}$ of at most $n$ balls. Precisely, we have:
$\epsilon_{n}(\mathcal{X})=\inf\\{\epsilon>0:\exists\hskip
1.42262pt\\{x_{i}\\}_{i=1}^{n}\subset\mathcal{X};\text{s.t.}\hskip
5.69054pt\mathcal{X}\subset\cup_{i=1}^{n}B(x_{i},\epsilon)\\}$ (9)
We also define the $n$-Bernstein width of a Banach space $X$ embedded in
another Banach space $Y$:
$b_{n}(X,Y)=\sup_{Z\subset X,\text{dim}(Z)\geq n+1}\inf_{z\in
Z}\frac{||z||_{Y}}{||z||_{X}}$ (10)
Intuitively, the Bernstein width gives the radius of the largest $Y$-ball that
may be inscribed in a unit $X$-ball in subspaces of dimension greater than
$n$. When $X$ and $Y$ are both Hilbert spaces, then $b_{n}$ coincides with the
$n+1^{\text{st}}$ singular number of the embedding $X\hookrightarrow Y$.
Moreover, if $X\hookrightarrow Y$ is compact, then $b_{n}(X,Y)\to 0$. For a
detailed discussion of the relationships between various approximation widths
and function space embeddings, we direct the reader to [9].
###### Remark (Notation).
For any two Banach spaces $\mathcal{A}$ and $\mathcal{B}$, we write
$\mathcal{A}\hookrightarrow\mathcal{B}$, if $\mathcal{A}$ is continuously
embedded in $\mathcal{B}$, and $\mathcal{A}\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}\mathcal{B}$ if this embedding is further compact. We
denote by $B(\mathcal{A})$ the unit ball in $\mathcal{A}$.
$\mathbb{E}_{\mathcal{D}}[\cdot]$ denotes the sample expectation, and
$C_{D}=\mathbb{E}_{\mathcal{D}}[k(X,\cdot)\otimes k(X,\cdot)]$ denotes the
sample covariance. We will denote by $f_{D,\lambda}$ the solution to (2) and
$f_{\lambda}=(C_{\nu}+\lambda)^{-1}S_{\nu}f^{*}$, the regularized population
solution. $\mathcal{F}$ denotes the Fourier transform, while $\mathcal{F}_{d}$
denotes the $d$-dimensional Fourier transform of a univariate function (see
e.g. Theorem 5.26 in [47]). $L^{p}(\mathcal{X},\mu)$ denotes the $L^{p}$ space
on the measure space $(\mathcal{X},\mu)$ — the domain will be omitted in the
notation if it is understood from context; likewise the measure will be
omitted if it is equivalent to Lebesgue measure. $||\cdot||_{2}$ or
$||\cdot||_{L^{2}(\nu)}$ are used to for norms in $L^{2}(\nu)$ (the latter if
we want to make the space explicit); $||\cdot||_{\phi}$ for the norm in
Hilbert scale $\mathcal{H}^{\phi}$, and $||\cdot||_{K}$ for the RKHS
$\mathcal{H}_{K}$. Typically, we will take $\nu=P_{X}=P|_{\mathcal{X}}$ where
$P$ is the data-generating measure, and $|_{\mathcal{Z}}$ denotes the
“marginal on” the space $\mathcal{Z}$. Finally, we will write $a(x)\preceq
b(x)$ ($a(x)\succeq b(x)$) if, $a(x)\leq Kb(x)$ (resp. $b(x)\leq Ka(x)$), for
some $K>0$ that depends only on problem parameters (and independent of $x$ or
the confidence level in the statement). If $a\preceq b$ and $b\preceq a$, then
we write $a\asymp b$.
### 2.4 Assumptions
Let $\phi$ and $\psi$ be two admissible index functions (i.e. $\phi$ and
$\psi$ are nondecreasing, nonnegative, and continuous, with
$\phi(0)=\psi(0)=0$)
###### Assumption 1.
$H_{K}\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}H^{\phi}\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}H^{\psi}\hookrightarrow L^{\infty}$
###### Assumption 2.
(a) $\frac{t}{\phi(t)}$ and $\frac{\phi(t)}{\psi(t)}$ are nondecreasing, and
$\frac{t}{\psi(t)}$ is concave; (b) $\phi,\psi$ satisfy the $\Delta_{2}$
condition (8) for some $D^{\phi}_{1},D^{\phi}_{2},D^{\psi}_{1},D^{\psi}_{2}>1$
###### Assumption 3.
$f^{*}\in\mathcal{H}^{\phi}$
###### Assumption 4.
$\sup\Big{\\{}\Big{\|}\sum_{i=1}^{n}c_{i}e_{i}(\cdot)\Big{\|}_{\infty}:\sum_{i=1}^{n}c^{2}_{i}=1\Big{\\}}\asymp\sqrt{s(n)}$
for some increasing functions $s$, where
$\sum_{n=1}^{\infty}\frac{1}{s(n)}<\infty$ and $s$ satisfies (8) for some
$D^{s}_{1},D^{s}_{2}>1$
###### Assumption 5.
There are constants $\sigma,L>0$ such that:
$\int_{\mathbb{R}}|y-f^{*}(x)|^{m}dP(y|x)\leq\frac{m!\sigma^{2}L^{m-2}}{2}$
for $\nu$-almost all $x\in\mathcal{X}$
###### Assumption 6*.
$k$ is radial, i.e. $k(x,y)=\kappa(||x-y||)$ for some positive definite
function $\kappa\in L^{1}(\mathbb{R})$, such that either
$d_{\kappa}(t)<\infty$ or $d_{\mathcal{F}_{d}\kappa}(t)<\infty$ for all $t>0$.
###### Assumption 7*.
$\mathcal{X}\subset\mathbb{R}^{d}$ is compact with
$\epsilon_{n}(\mathcal{X})\asymp n^{-\frac{1}{d}}$ for $n\in\mathbb{N}$.
###### Assumption 8*.
$\psi\Big{(}\frac{\mathcal{F}_{d}\kappa(t^{\frac{1}{d}})}{s(t)}\Big{)}\asymp\frac{1}{s(t)}$
as $t\to\infty$
Assumptions 1 and 2 dictates the relative growth rates of the index functions
$\phi$ and $\psi$. Indeed, the compactness of the embedding
$H^{\phi}\stackrel{{\scriptstyle c}}{{\hookrightarrow}}H^{\psi}$ ensures that
$\frac{\phi(t)}{\psi(t)}\to 0$ as $t\to 0$. Moreover, the first embedding
$H_{K}\stackrel{{\scriptstyle c}}{{\hookrightarrow}}H^{\phi}$ ensures that
both $\frac{t}{\phi(t)}\to 0$ and $\frac{t}{\psi(t)}\to 0$ as $t\to 0$.
Intuitively, this condition implies that the norms in $\mathcal{H}_{\psi}$ and
$\mathcal{H}_{\phi}$ are weaker than those in our hypothesis class
$\mathcal{H}_{K}$, allowing us to study convergence in these larger spaces of
sample estimators constructed in $\mathcal{H}_{K}$. The second part of the
assumption requiring $\frac{t}{\phi(t)}$ and $\frac{\phi(t)}{\psi(t)}$ to be
nondecreasing extends this behavior as $t\to\infty$; collectively the two
growth conditions characterize $\phi^{-1}$ and $\psi^{-1}$ as resembling
Young’s functions; a requirement commonly imposed in the study of Orlicz
spaces [1]. The requirement that $\frac{t}{\phi(t)}$ is additionally concave
enables the application of a Gagliardo-Nirenberg type interpolation inequality
[29] that relates $L^{\infty}$ norms to those in $\mathcal{H}_{K}$ and
$L^{2}(\nu)$, which will be crucial in our analysis of uniform error rates.
Finally, the condition that index functions are $\Delta_{2}$ is found quite
commonly in the literature on statistical inverse problems [30, 39], where it
has been used to demonstrate adaptivity of balancing strategies to unknown
source conditions. It should be noted that all the growth conditions in
Assumption 2 are satisfied when $\phi(t)=t^{\beta}$ and $\psi(t)=t^{\alpha}$
are power functions with $0<\alpha<\beta<1$ (the so-called Hölder source
conditions).
Assumption 3 characterizes the smoothness of the learning target $f^{*}$. Note
a distinctive feature of our analysis is that we allow $f^{*}$ to lie outside
the hypothesis class $\mathcal{H}_{K}$ (in light of Assumption 1), while still
in $L^{\infty}(\mathcal{X})$ (see Assumption 1). Assumption 5 is standard
subexponential noise condition in kernel regression [17, 43].
Assumption 4 characterizes the embedding of $L^{\infty}$ in $L^{2}(\nu)$ via
the eigenspaces of $T_{\nu}$. The “Lebesgue” function
$M_{n}(x)=\sum_{i=1}^{n}e_{i}(x)$ was first considered by Littlewood [18] for
the Fourier basis; later its $L^{p}$-norms were studied in several works [21,
6, 7], primarily in the context of finding roots for various generalized
polynomials. [44] and [20] later explored the intimate connection between such
Lebesgue-type inequalities and the greedy approximation properties of the
basis $\\{e_{i}\\}_{i=1}^{\infty}$. Indeed, it can be seen directly from the
definition (10) that assumption 4 implies a lower bound of $s(n)^{-1}$ on the
Bernstein widths $b^{2}_{n-1}(L^{\infty},L^{2}(\nu))$ of $L^{\infty}$ in
$L^{2}(\nu)$. Note, that although the orthonormality of
$\\{e_{i}\\}_{i=1}^{\infty}$ in $L^{2}(\nu)$ is $\nu$-dependent, this system
is always an orthonormal basis in $\mathcal{H}_{K}$ by separability and the
injectivity of $I_{\nu}$ (see Theorem 3.3 in [42]). Moreover, Assumption 4 is
independent of the Mercer eigenvalues $\\{\mu_{i}\\}_{i=1}^{\infty}$ and is
hence applicable to any RKHS with a common eigenbasis (e.g. the popular
Fourier basis in $L^{2}([-\pi,\pi]^{d})$). These situations arise commonly in
the study of stochastic differential equations [13, 14], where $\nu$ is a
Gaussian measure, $\\{e_{i}\\}_{i=1}^{\infty}$ is a fixed basis of
multivariate Hermite polynomials, and $\\{\mu^{-1}_{i}\\}_{i=1}^{\infty}$ are
the (possibly unknown) eigenvalues of a general operator drift coefficient.
The square of the left-hand side in Assumption 4 is known as the Christoffel
function, whose asymptotic growth (and its interaction with the measure $\mu$)
is an active area of research [28, 48, 3] with diverse applications to
quadrature/interpolation and random matrices [45] and [3]. Moreover, it is
important to note that Assumption 3 is significantly more general than the
more typical assumption of uniformly bounded bases, which has been shown to be
violated by several common kernels (see discussion in [50, 42]).
While Assumptions 6*-8* are not necessary for establishing the upper bound in
Theorem 2, they are required to demonstrate its optimality. Namely, the
radiality assumption on the kernel $k$ in Assumption 6* enables the tight
estimation of the approximation widths of $\mathcal{H}_{K}$ in $L^{\infty}$
via the compactness properties of the domain $\mathcal{X}$, the latter of
which are characterized by the entropy numbers specified in Assumption 7*. The
dilation condition can be satisfied either by the kernel or its
$d$-dimensional Fourier transform; an example of the latter being the popular
Matern kernel. This dilation condition is primarily intended for streamlining
the analysis, it could be removed at the expense of requiring Assumption 8* be
satisfied instead by a dilation of the Fourier transform
$\mathcal{F}_{d}\kappa$ (this tradeoff is highlighted in Remark Remark in the
Appendix). Similarly, the precise asymptotic behavior of the entropy numbers
of $\mathcal{X}$ is only included for simplicity; the general case would
require replacing $\mathcal{F}_{d}\kappa(t^{\frac{1}{d}})$ with
$\mathcal{F}_{d}\kappa(\epsilon_{t}(\mathcal{X})^{-1})$ in Assumption 8*.
Assumption 8* ensures that the index function $\psi$ is indeed the “optimal”
choice for characterizing the embedding of our hypothesis class
$\mathcal{H}_{K}$ in $L^{\infty}$. It is important to note that while
Assumptions 1 and 3 implicitly depend on the ambient measure $\nu$ (as the
index functions act spectrally on the Mercer operator $T_{\nu}$), the
optimality condition is independent of this measure (which is typically
unknown in practice and determined by the data-generating process) given the
orthonormal basis $\\{e_{i}\\}_{i=1}^{\infty}$ of $\mathcal{H}_{K}$ in
Assumption 4. This contrasts with the exact decay rates of the eigenvalues
$\mu_{i}(T_{\nu})$ or the effective dimension [17, 4] traditionally required
to establish minimax optimal learning rates. Indeed, as is shown in Appendix
C, the combination of Assumptions 1, 2, 4, and 8* enable us to infer that
$\mu_{i}\asymp\psi^{-1}(s(i)^{-1})$ via a comparison of
$b^{2}_{n}(\mathcal{H}_{K},L^{\infty})$ (estimated by
$\mathcal{F}_{d}\kappa(n^{\frac{1}{d}}))$) and
$b^{2}_{n}(L^{\infty},L^{2}(\nu))$ (estimated by $s(n)^{-1}$). In the
following lemma, we demonstrate Assumption 8* is satisfied when $\psi$
characterizes the “optimal” scale of range spaces for a modified Mercer
operator.
###### Lemma 1.
Suppose $s(n)=n^{\alpha}$ for some $\alpha>1$ and
$\alpha_{\psi}=\beta_{\psi}\equiv\rho$. Define the Fourier multiplier operator
$T$ by
$\mathcal{F}(Tf)=\mathcal{F}k(||\cdot||^{\frac{1}{\alpha}})\mathcal{F}f$. Let
$\Lambda_{\text{rad}}=\\{f=\tilde{f}(||x||):\tilde{f}(r)r^{\frac{1}{2}}\in
L^{1}(0,\infty),\hskip 2.84526pt\tilde{f}(r)r^{d-1}\hskip
2.84526pt\text{decreasing}\\}$ be a space of radially decreasing functions on
$\mathbb{R}^{d}$. Then, if $\rho>\frac{p}{p+1}$ and Assumption 8* holds, $T$
is bounded from $L^{p,q}(\mathbb{R}^{d})\cap\Lambda_{\text{rad}}$ to
$\Lambda^{q}_{\Psi}$ where $\Psi$ is the fundamental function of
$L^{\psi,t^{\bar{p}}}$ with $\bar{p}=\rho\Big{(}1+\frac{1}{p}\Big{)}-2$.
Moreover, $\Lambda^{q}_{\Psi}$ is the smallest $q$-Lorentz space satisfying
this conclusion.
The choice of polynomial growth $s(n)=n^{\alpha}$ is inspired by Lebesgue-type
inequalities for common greedy bases (see e.g. Chapter 2 of [44]). Note, that
in the limiting case when $\alpha=1$, $T$ is simply the Mercer operator
$T_{m}$ with respect to the Lebesgue measure $m$. Hence, intuitively Lemma 1
states that when $\alpha_{\psi}=\beta_{\psi}$ the interpolation scale
$(\Lambda_{\Psi},M_{\Psi})$ characterizes the optimal range space for the
kernel embedding.
The requirement $\alpha_{\psi}=\beta_{\psi}$ is crucial in Lemma 1 in order to
identify the couple $(\Lambda_{\Psi},M_{\Psi})$ with an Orlicz-Lorentz space.
This is unsurprising, when considered in analogy with the Banach space case
(i.e. when $\psi$ is a Young function), for which stronger results hold [25].
Indeed, there it is known that $L^{\psi}$ is only a Lorentz endpoint space
when $\psi(t)=t^{\frac{1}{r}}v(t)$ for some $r\in(1,\infty)$ and function $v$
that satisfies the Lorentz condition (Theorem 3 in [25]) which implies sub-
power growth. Hence, intuitively, Lemma 1 suggests that we may only associate
the optimal range scale for $T$ with an Orlicz-Lorentz space when the latter
space itself also “belongs” to this interpolation scale (see Remark after the
proof of Lemma 1 for further discussion).
We observe $\alpha_{\psi}=\beta_{\psi}\equiv\rho$ most notably when
$\psi(t)=t^{\rho}$ (e.g. Assumption EMB in [17]). Here, Lemma 1 states that
$T$ is bounded from $L^{p,q}(\mathbb{R}^{d})\cap\Lambda_{\text{rad}}$ to
$L^{\frac{p\rho^{*}}{\rho^{*}+p},q}(\mathbb{R}^{d})$, where $\rho^{*}<0$ is
the dual exponent of $\rho\in(0,1)$. Observe that
$\frac{p\rho^{*}}{\rho^{*}+p}>1$ as $\rho<p$ by Assumption 2.
In the general case of $\alpha_{\psi}<\beta_{\psi}$, we only obtain the
following Boas-type result (see [23]):
###### Corollary 1.1.
Suppose $s(n)=n^{\alpha}$ for some $\alpha>1$. Define the Fourier multiplier
operator $T$ as in Lemma 1. Then, if $\beta_{\psi}>\frac{p}{p+1}$ and
Assumption 8* holds, $T$ is bounded from
$L^{p,q}(\mathbb{R}^{d})\cap\Lambda_{\text{rad}}$ to $\Lambda^{q}_{\Psi}$
where $\Psi(t)=\frac{t^{1+\frac{1}{p}}}{\psi^{-1}(t)}$. Moreover,
$\Lambda^{q}_{\Psi}$ is the smallest $q$-Lorentz space satisfying this
conclusion.
#### 2.4.1 Related Work
Assumptions 1 and 3 can be viewed as a generalization of those in [17] to
Hilbert scales. Namely, these assumptions reduce to the corresponding source
and embedding conditions in [17] with the choice of $\phi(t)=t^{\beta}$ and
$\psi(t)=t^{\alpha}$ (using their notation). The additional embedding
$H^{\phi}\hookrightarrow H^{\psi}$ in Assumption 1 is not strictly necessary
for the derivation of minimax rates, but helps simplify the analysis.
Intuitively, this embedding ensures that $f^{*}$ is bounded.
We emphasize that a distinctive feature of our analysis is that we do not make
any assumption on the eigendecay of $T_{\nu}$ or its effective dimension.
Indeed, while [4, 5] established learning rates in the stronger
$\|\cdot\|_{K}$ norm for more general regularization schemes, their analysis
only considered Holder-type source conditions and hinged on a particular
asymptotic eigendecay of $T_{\nu}$ (which depends on the unknown measure
$\nu=P|_{\mathcal{X}}$). [36] and [37] consider general source conditions
similar to this paper (and for more general regularization schemes), although
they too place conditions on the decay of the effective dimension of $T_{\nu}$
(which is strongly related to its eigendecay). Recently, [26] considered
Lepski-based parameter selection strategy for general regularization schemes,
and demonstrated that known estimates on the effective dimension of $T_{\nu}$
can be effectively leveraged for adapting to misspecification.
To the best of our knowledge, this is the first work to establish minimax
learning rates under general source conditions without appealing to any direct
assumptions on the eigendecay or effective dimension of the Mercer operator
$T_{\nu}$. Instead, our approach leverages the $L^{\infty}$ embedding of the
Hilbert scale $\mathcal{H}^{\psi}$, the distance between $L^{\infty}$ and
$L^{2}(\nu)$ (Assumption 4), and a compatibility condition between $\psi$ and
the Fourier transform of the kernel (Assumption 8*) to infer spectral
information on $T_{\nu}$. In a sense, our analysis demonstrates that, via an
interpolation inequality, evaluating the tightness of the embedding
$\mathcal{H}^{\psi}\hookrightarrow L^{\infty}$ is in a sense equivalent to
evaluating the optimality of $\psi$ in characterizing the range of a Mercer-
type operator (via Lemma 1). When $\psi$ is indeed optimal, the interpolation
inequality (Lemma 1) provides the correct “change of measure”. However, this
delicate relationship hinges on the radiality of the kernel, and does not
obviously extend to more general kernel families.
### 2.5 Results
We first present our upper bound on the mean-square error in Theorem 2.
###### Theorem 2.
Suppose Assumptions 1-8* hold. Let $\tilde{s}(t)=s^{-1}(t^{-1})$. Choosing
$\lambda_{n}\asymp(\frac{\phi}{\tilde{s}\circ\psi})^{-1}(n^{-1})$, we obtain,
with probability $1-2\delta$:
$||f_{D,\lambda_{n}}-f^{*}||_{L^{2}(\nu)}\preceq\log(\delta^{-1})\sqrt{\phi\Big{(}\Big{(}\frac{\phi}{\tilde{s}\circ\psi}\Big{)}^{-1}(n^{-1})\Big{)}}$
(11)
Intuitively, the function
$\phi\Big{(}\Big{(}\frac{\phi}{\tilde{s}\circ\psi}\Big{)}^{-1}\Big{)}$
appearing in (11) captures the distance between $\mathcal{H}^{\phi}$ and the
hypothesis class $\mathcal{H}_{K}$, i.e. the “degree of misspecification”.
Indeed, since $\mu_{i}\asymp\psi^{-1}(s(i)^{-1})$ (as shown in Appendix C),
the upper rate effectively compares $\phi(\mu_{i})$ to $\mu_{i}$. For example,
when $\phi(t)=t^{\beta}$ and $\mu_{i}\asymp i^{-\frac{1}{p}}$ (for
$\beta,p\in(0,1)$), then
$\phi\Big{(}\Big{(}\frac{\phi}{\tilde{s}\circ\psi}\Big{)}^{-1}\Big{)}(n^{-1})\asymp
n^{-\frac{\beta}{p+\beta}}$. If $\frac{\beta}{p}$ is close to $1$, then
$\mathcal{H}^{\phi}$ is weakly embedded in $L^{2}(\nu)$ (i.e. “far” from
$\mathcal{H}_{K}$) and the setting is poorly specified, producing the slow
rate of $n^{-\frac{1}{4}}$. If $\beta\gg p$ is large, i.e $\mathcal{H}^{\phi}$
is close to $\mathcal{H}_{K}$, our setting is well-specified, and we approach
the fast rate of $n^{-\frac{1}{2}}$.
The proof of Theorem 2 hinges on the following bias-variance decomposition,
which illustrates that, up to a constant, both components of the error can be
controlled purely in terms of $\phi,\psi$, and $s$:
###### Lemma 3.
Suppose Assumptions 1-8* hold. Then:
$||f^{*}-f_{D,\lambda}||_{L^{2}(\nu)}\preceq||f^{*}||_{\phi}\sqrt{\phi(\lambda)}+||k_{\psi}||_{\infty}||f^{*}||_{\phi}\sqrt{\frac{\log(2\delta^{-1})}{n}\Big{(}\frac{1}{n\psi(\lambda)}+\frac{\phi(\lambda)}{\psi(\lambda)}+s^{-1}(\psi(\lambda)^{-1})\Big{)}}$
(12)
with probability $1-3\delta$.
As noted above, we observe here that the kernel only appears in (12) via the
constant term $||k_{\psi}||_{\infty}$, and hence $k$ does not influence the
asymptotic behavior of the upper bound as $\lambda\to 0$. This surprising
absence of the kernel (beyond the eigenbasis $\\{e_{i}\\}_{i=1}^{\infty}$),
can be attributed to the embedding conditions in Assumption 1, which ensure
that (via a Gagliardo-Nirenberg type inequality) uniform bounds on the bias
depend only on the ratios between norms in $\mathcal{H}^{\psi}$, and
$\mathcal{H}^{\phi}$ and $\mathcal{H}_{K}$, respectively. Moreover, due to
relative growth rates of the index functions implied by compactness, these
capacity ratios depend solely on $\phi$, $\psi$, and $s$.
We now discuss the optimality of the upper bound in Theorem 2. The derivation
of a minimax lower bound matching (11) is significantly more involved and
hinges crucially on the sharpness of the embedding
$\mathcal{H}^{\psi}\hookrightarrow L^{\infty}$, which in turn depends on the
compactness properties of the domain $\mathcal{X}\subset\mathbb{R}^{d}$
(captured by the entropy condition in Assumption 7*) and the smoothness of the
kernel $k$. Unlike the derivation of the upper bound, here the analysis does
not reduce simply to a comparison of the index functions, as we must
demonstrate the maximality of the spectral function $\psi$ in “stretching” the
kernel class $\mathcal{H}_{K}$. Informally, we need to ensure that the
enlarged (infinite-dimensional) ellipsoid $B(\psi(\mathcal{H}_{K}))$ is
“maximal” in $L^{\infty}$, otherwise our use of $\mathcal{H}^{\psi}$ as a
proxy for $L^{\infty}$ in the derivation of (11) is suboptimal (i.e.
$||\cdot||_{\mathcal{H}^{\psi}}$ may provide a very coarse approximation of
$||\cdot||_{\infty}$). We quantify this maximality via estimating the
Bernstein widths of $\mathcal{H}_{K}$ in $L^{\infty}$, which can then be
compared to the eigenvalues $\mu_{i}$ (the Bernstein widths of
$\mathcal{H}_{K}$ in $L^{2}(\nu)$; see e.g. Prop. 5 in [29]) via the Gagliardo
interpolation inequality. The former widths are independent of the measure
$\nu$ and depend only on the smoothness of $\kappa$ (captured by its Fourier
decay) and the compactness of $\mathcal{X}$. Intuitively, the embedding
$\mathcal{H}^{\psi}\hookrightarrow L^{\infty}$ is sharp (and our learning rate
in (11) is optimal) if the index function $\psi$ provides the correct “change-
of-measure” from $\nu$ to the $d$-dimensional Lebesgue measure.
###### Theorem 4.
There exists a distribution $P$ on $\mathcal{X}\times\mathbb{R}$ satisfying
Assumption 5 with $P|_{\mathcal{X}}=\nu$, $||f^{*}_{P}||_{\phi}\leq B_{\phi}$,
$||f^{*}_{P}||_{\infty}\leq B_{\infty}$, such that, for any learning algorithm
$\mathcal{D}\mapsto f_{D}$, we have with $P^{n}$ probability not less than
$1-\frac{432\tau K^{\log\tau}}{U_{d,\infty,\phi}\sigma^{2}\log 2}$:
$||f_{D}-f^{*}_{P}||_{L^{2}(\nu)}\succeq\sqrt{\tau\phi\Big{(}\Big{(}\frac{\phi}{\tilde{s}\circ\psi}\Big{)}^{-1}(n^{-1})\Big{)}}$
where $U_{d,\infty,\phi}$ does not depend on $n$ or $\tau$ and $K>1$ is an
absolute constant.
## 3 Conclusion
In this paper, we derive novel minimax learning rates for ridge regression
over Hilbert scales with general source conditions. A notable feature of our
analysis is that we do not impose the conventional conditions on eigendecay
rates, and instead present a new approach based on estimating the Bernstein
widths of the kernel class in $L^{\infty}$ to demonstrate minimax optimality.
We introduce a new Fourier capacity condition, which characterizes the
complexity of the kernel via the optimal Lorentz range space of a modified
Mercer operator. Future work involves extending our convergence analysis to
stronger interpolation norms and online regression settings.
## References
* [1] Waldo Arriagada “Matuszewska–orlicz indices of the sobolev conjugate young function” In _Partial Differential Equations in Applied Mathematics_ 3 Elsevier, 2021, pp. 100029
* [2] Sergei V Astashkin and Lech Maligranda “Ultrasymmetric Orlicz spaces” In _Journal of mathematical analysis and applications_ 347.1 Elsevier, 2008, pp. 273–285
* [3] P Nevai AT983168RF14 and Géza Freud “Orthogonal polynomials and Christoffel functions: A case study” In _J. Approx. Theory_ 48, 1986, pp. 3–167
* [4] Gilles Blanchard and Nicole Mücke “Optimal rates for regularization of statistical inverse learning problems” In _Foundations of Computational Mathematics_ 18.4 Springer, 2018, pp. 971–1013
* [5] Gilles Blanchard and Nicole Mücke “Kernel regression, minimax rates and effective dimensionality: Beyond the regular case” In _Analysis and Applications_ 18.04 World Scientific, 2020, pp. 683–696
* [6] Sergei Viktorovich Bochkarev “Multiplicative inequalities for the L 1 norm: Applications in analysis and number theory” In _Proceedings of the Steklov Institute of Mathematics_ 255.1 Springer, 2006, pp. 49–64
* [7] Peter Borwein and Richard Lockhart “The expected $L_{p}$ norm of random polynomials” In _Proceedings of the American Mathematical Society_ 129.5, 2001, pp. 1463–1472
* [8] Andrea Caponnetto and Ernesto De Vito “Optimal rates for the regularized least-squares algorithm” In _Foundations of Computational Mathematics_ 7.3 Springer, 2007, pp. 331–368
* [9] Bernd Carl and Irmtraud Stephani “Entropy, Compactness and the Approximation of Operators”, Cambridge Tracts in Mathematics Cambridge University Press, 1990 DOI: 10.1017/CBO9780511897467
* [10] M Carro, Luboš Pick, Javier Soria and Vladimir D Stepanov “On embeddings between classical Lorentz spaces” In _Mathematical Inequalities and Applications_ 4 ELEMENT, 2001, pp. 397–428
* [11] C Carton-Lebrun and HP Heinig “Weighted Fourier transform inequalities for radially decreasing functions” In _SIAM journal on mathematical analysis_ 23.3 SIAM, 1992, pp. 785–798
* [12] Marianna Chatzakou and Vishvesh Kumar “$L^{p}L^{q}$ boundedness of Fourier multipliers associated with the anharmonic Oscillator” In _arXiv preprint arXiv:2004.07801_ , 2020
* [13] Pao-Liu Chow “Infinite-dimensional Kolmogorov equations in Gauss-Sobolev spaces” In _Stochastic Analysis and Applications_ 14.3 Taylor & Francis, 1996, pp. 257–282
* [14] Pao-Liu Chow “Solutions of linear elliptic equations in Gauss-Sobolev spaces” In _Communications on Stochastic Analysis_ 6.3, 2012, pp. 8
* [15] A Debernardi “Uniform convergence of Hankel transforms” In _Journal of Mathematical Analysis and Applications_ 468.2 Elsevier, 2018, pp. 1179–1206
* [16] Alberto Debernardi “The Boas problem on Hankel transforms” In _Journal of Fourier Analysis and Applications_ 25.6 Springer, 2019, pp. 3310–3341
* [17] Simon Fischer and Ingo Steinwart “Sobolev Norm Learning Rates for Regularized Least-Squares Algorithms.” In _J. Mach. Learn. Res._ 21, 2020, pp. 205–1
* [18] Godfrey Harold Hardy and John Edensor Littlewood “A new proof of a theorem on rearrangements” In _Journal of the London Mathematical Society_ 1.3 Wiley Online Library, 1948, pp. 163–168
* [19] Anna Kamińska “Some remarks on Orlicz-Lorentz spaces” In _Mathematische Nachrichten_ 147.1 Wiley Online Library, 1990, pp. 29–38
* [20] Boris Sergeevic Kashin and Artur Artushovich Saakyan “Orthogonal series” American Mathematical Soc., 2005
* [21] Sergei Vladimirovich Konyagin “On a problem of Littlewood” In _Izvestiya Rossiiskoi Akademii Nauk. Seriya Matematicheskaya_ 45.2 Russian Academy of Sciences, Steklov Mathematical Institute of Russian …, 1981, pp. 243–265
* [22] Aigerim Kopezhanova “Summability of Fourier transforms of functions from Lorentz spaces”, 2017
* [23] Aigerim Kopezhanova, Erlan Nursultanov and Lars Erik Persson “A new generalization of Boas theorem for some Lorentz spaces $\Lambda$q ($\omega$)” Ele-Math, 2018
* [24] Junhong Lin, Alessandro Rudi, Lorenzo Rosasco and Volkan Cevher “Optimal rates for spectral algorithms with least-squares regression over Hilbert spaces” In _Applied and Computational Harmonic Analysis_ 48.3 Elsevier, 2020, pp. 868–890
* [25] GG Lorentz “Relations between function spaces” Mathematical Science Directorate, Office of Scientific Research, US Air Force, 1959
* [26] Shuai Lu, Peter Mathé and Sergei V Pereverzev “Balancing principle in supervised learning for a general regularization scheme” In _Applied and Computational Harmonic Analysis_ 48.1 Elsevier, 2020, pp. 123–148
* [27] Attila Máté, Paul Nevai and Vilmos Totik “Szego’s extremum problem on the unit circle” In _Annals of Mathematics_ JSTOR, 1991, pp. 433–453
* [28] Attila Máté and Paul G Nevai “Bernstein’s Inequality in Lp for and (C, 1) Bounds for Orthogonal Polynomials” In _Annals of Mathematics_ JSTOR, 1980, pp. 145–154
* [29] Peter Mathé and Bernd Hofmann “Direct and inverse results in variable Hilbert scales” In _Journal of approximation theory_ 154.2 Elsevier, 2008, pp. 77–89
* [30] Peter Mathé and Sergei V Pereverzev “Geometry of linear ill-posed problems in variable Hilbert scales” In _Inverse problems_ 19.3 IOP Publishing, 2003, pp. 789
* [31] Shahar Mendelson and Joseph Neeman “Regularization in kernel learning” In _The Annals of Statistics_ 38.1 Institute of Mathematical Statistics, 2010, pp. 526–565
* [32] Stephen J Montgomery-Smith “The Hardy operator and Boyd indices” In _Lecture Notes Pure Appl. Math_ 175, 1995, pp. 359–364
* [33] Alain Pajor and Nicole Tomczak-Jaegermann “Subspaces of small codimension of finite-dimensional Banach spaces” In _Proceedings of the American Mathematical Society_ 97.4, 1986, pp. 637–642
* [34] Luboš Pick, Alois Kufner, Oldřich John and Svatopluk Fucı́k “Function spaces, 1” In _Function Spaces, 1_ de Gruyter, 2012
* [35] Evgeniy Pustylnik “Ultrasymmetric spaces” In _Journal of the London Mathematical Society_ 68.1 Cambridge University Press, 2003, pp. 165–182
* [36] Abhishake Rastogi “Tikhonov regularization with oversmoothing penalty for nonlinear statistical inverse problems” In _arXiv preprint arXiv:2002.01303_ , 2020
* [37] Abhishake Rastogi and Sivananthan Sampath “Optimal rates for the regularized learning algorithms under general source condition” In _Frontiers in Applied Mathematics and Statistics_ 3 Frontiers, 2017, pp. 3
* [38] Michael Ruzhansky and Niyaz Tokmagambetov “Nonharmonic analysis of boundary value problems” In _International Mathematics Research Notices_ 2016.12 Oxford University Press, 2016, pp. 3548–3615
* [39] Thomas Schuster, Barbara Kaltenbacher, Bernd Hofmann and Kamil S Kazimierski “Regularization methods in Banach spaces” In _Regularization Methods in Banach Spaces_ de Gruyter, 2012
* [40] Steve Smale and Ding-Xuan Zhou “Shannon sampling and function reconstruction from point values” In _Bulletin of the American Mathematical Society_ 41.3, 2004, pp. 279–305
* [41] Ingo Steinwart and Andreas Christmann “Support vector machines” Springer Science & Business Media, 2008
* [42] Ingo Steinwart and Clint Scovel “Mercer’s theorem on general domains: On the interaction between measures, kernels, and RKHSs” In _Constructive Approximation_ 35.3 Springer, 2012, pp. 363–417
* [43] Prem Talwai, Ali Shameli and David Simchi-Levi “Sobolev Norm Learning Rates for Conditional Mean Embeddings” In _International Conference on Artificial Intelligence and Statistics_ , 2022, pp. 10422–10447 PMLR
* [44] Vladimir Temlyakov “Greedy approximation” Cambridge University Press, 2011
* [45] Walter Van Assche “Asymptotics for orthogonal polynomials” Springer, 2006
* [46] Roman Vershynin and Convex Bodies “Lecture notes in geometric functional analysis” Citeseer, 2009
* [47] Holger Wendland “Scattered data approximation” Cambridge university press, 2004
* [48] Yuan Xu “Asymptotics for orthogonal polynomials and Christoffel functions on a ball” In _Methods and Applications of Analysis_ 3.2 International Press of Boston, 1996, pp. 257–272
* [49] Yiming Ying and Ding-Xuan Zhou “Learnability of Gaussians with flexible variances” Microtome Publishing, 2007
* [50] Ding-Xuan Zhou “The covering number in learning theory” In _Journal of Complexity_ 18.3 Elsevier, 2002, pp. 739–767
* [51] Ding-Xuan Zhou “Capacity of reproducing kernel spaces in learning theory” In _IEEE Transactions on Information Theory_ 49.7 IEEE, 2003, pp. 1743–1752
## Appendix A Proof of Lemma 1
We first demonstrate a two-sided Boas-type result (a two-sided Hardy-
Littlewood type inequality, see e.g. [23, 22]) for the radial Fourier
transform on rearrangement invariant spaces. We expect that this result is
well-known to experts, but could not find a definite source, and hence prove
it here:
###### Lemma 1.
Suppose $f=\tilde{f}(||\cdot||)\in\mathbb{R}^{d}$ is a radial function with
$\tilde{f}(r)r^{d-1}$ decreasing on $(0,\infty)$ and
$\tilde{f}(r)r^{\frac{1}{2}}\in L^{1}(0,\infty)$. Let $w$ be a positive weight
function on $\mathbb{R}^{+}$ and define $\tilde{w}(\cdot)=w(||\cdot||^{d})$ on
$\mathbb{R}^{d}$. For $1\leq p<\infty$, define:
$\displaystyle W^{1}_{p}(t)$
$\displaystyle=\int_{0}^{t}r^{p}w^{p}(r^{-1})\frac{dr}{r}$ (13) $\displaystyle
W^{2}_{p}(t)$ $\displaystyle=\int_{0}^{t}w^{p}(r)\frac{dr}{r}\ $ (14)
Suppose $\beta_{W^{1}_{p}},\beta_{W^{2}_{p}}<p$ (see section 2.2 for a
definition). Then:
$\int_{\mathbb{R}^{d}}\Big{(}|\mathcal{F}f(x)|\tilde{w}\Big{(}\frac{x^{\prime}}{||x||}\Big{)}||x||^{d}\Big{)}^{p}\frac{dx}{||x||^{d}}\asymp\int_{\mathbb{R}^{d}}(\tilde{w}(x)|f(x)|)^{p}\frac{dx}{||x||^{d}}$
(15)
where $x=||x||x^{\prime}$.
###### Proof.
We only show the backward direction as the forward direction follows
analogously (under the assumption $\beta_{W^{2}_{p}}>\frac{1}{d^{2}}$); the
latter may also be deduced from the proof of Theorem 2.3 in [11] through the
application of a specialized Hardy inequality (Theorem 2(i) in [32]; see
below) customized to our growth condition in (13) and (14). Hence, for the
reverse direction, we have, by the Fourier inversion formula (which holds
since $\tilde{f}(r)r^{\frac{1}{2}}\in L^{1}(0,\infty)$ by assumption; see [15,
16]):
$\displaystyle f(x)$
$\displaystyle=\int_{\mathbb{R}^{d}}\mathcal{F}f(||u||)e^{2\pi x\cdot u}du$
$\displaystyle=|S_{d-2}|\int_{0}^{\infty}s^{d-1}\mathcal{F}f(s)\int_{0}^{1}\cos(2\pi
s||x||t)(1-t^{2})^{\frac{d-3}{2}}dtds$
$\displaystyle=\int_{0}^{\frac{1}{||x||}}+\int_{\frac{1}{||x||}}^{\infty}s^{d-1}\mathcal{F}f(s)\int_{0}^{1}\cos(2\pi
s||x||t)(1-t^{2})^{\frac{d-3}{2}}dtds$
where $|S_{d-2}|$ is the surface area of the $(d-1)$-sphere. Then, analogously
to the proof of Theorem 2.3 in [11] (see (2.5)-(2.7) on p. 787-788), we can
demonstrate that:
$\displaystyle\Big{|}\int_{0}^{\frac{1}{||x||}}s^{d-1}\mathcal{F}f(s)\int_{0}^{1}\cos(2\pi
s||x||t)(1-t^{2})^{\frac{d}{3}}dtds\Big{|}$ $\displaystyle\leq
C_{1}\int_{0}^{\frac{1}{||x||}}|\mathcal{F}f(s)|s^{d-1}ds$ (16)
$\displaystyle\Big{|}\int_{\frac{1}{||x||}}^{\infty}s^{d-1}\mathcal{F}f(s)\int_{0}^{1}\cos(2\pi
s||x||t)(1-t^{2})^{\frac{d}{3}}dtds\Big{|}$ $\displaystyle\leq
C_{2}\int_{0}^{\frac{1}{||x||}}|\mathcal{F}f(s)|s^{d-1}ds$ (17)
for some constants $C_{1},C_{2}>0$. Therefore, we have that:
$\displaystyle\int_{\mathbb{R}^{d}}(w(||x||^{d})|f(x)|)^{p}\frac{dx}{||x||^{d}}$
$\displaystyle=\int_{0}^{\infty}(w(r^{d})|f(r)|)^{p}\frac{dr}{r}$
$\displaystyle\leq
C\int_{0}^{\infty}\Big{(}w(r^{d})\int_{0}^{\frac{1}{r}}|\mathcal{F}f(s)|s^{d-1}ds\Big{)}^{p}\frac{dr}{r}$
$\displaystyle=C\int_{0}^{\infty}r^{pd}w^{p}(r^{-d})\Big{(}\frac{1}{r^{d}}\int_{0}^{r}|\mathcal{F}f(s)|s^{d-1}ds\Big{)}^{p}\frac{dr}{r}$
$\displaystyle\leq
C\int_{0}^{\infty}r^{pd}w^{p}(r^{-d})(|\mathcal{F}f(r)|)^{p}\frac{dr}{r}$ (18)
$\displaystyle=C\int_{\mathbb{R}^{d}}\Big{(}|\mathcal{F}f(x)|\tilde{w}\Big{(}\frac{x^{\prime}}{||x||^{d}}\Big{)}||x||^{d}\Big{)}^{p}\frac{dx}{||x||^{d}}$
where in (18) we have applied Hardy’s inequality for rearrangement-invariant
spaces , i.e. Theorem 2(i) in [32] with $p=\frac{1}{d}$ and $r=1$ (using their
notation), and $p_{X}=\frac{p}{d\beta_{W^{1}_{p}}}$ (since here the r.i. space
$X=L^{p}(w^{p}(t^{-d})dt^{pd})$ in their notation). The requirement
$\frac{p}{d\beta_{W^{1}_{p}}}>\frac{1}{d}$ in this theorem is satisfied by
assumption. ∎
###### Remark.
We emphasize that the validity of Lemma 1 hinges on the condition that
$\tilde{f}(r)r^{d-1}$ is decreasing, which factors crucially in demonstrating
(16) and (17) (see the p. 787-788 in [11] for details). Simple counterexamples
to (15) may be found when this condition is violated.
###### Lemma 2.
Let $w(t)=t^{-p}$ for some $p\in(0,1)$. Suppose
$\rho=\alpha_{\Phi}=\beta_{\Phi}$. Then the fundamental function $\phi_{p}(t)$
of $L^{\Phi,w}$ is equivalent to $\frac{t^{\frac{2-p}{\rho}}}{\Phi^{-1}(t)}$.
###### Proof.
Note, that by a standard argument, we may find a function
$\bar{\Phi}\asymp\Phi$ such that
$\frac{t\bar{\Phi}^{\prime}(t)}{\bar{\Phi}(t)}=\rho$ and hence $\Phi(t)\sim
t^{\rho}$ (see e.g. [35]). Hence, we have
$\displaystyle\int_{0}^{\infty}\Phi\Big{(}\frac{\chi_{(0,s)}(t)}{\lambda}\Big{)}w(t)dt$
$\displaystyle=\int_{0}^{s}\Phi\Big{(}\frac{1}{\lambda}\Big{)}t^{-p}dt$
$\displaystyle=s^{1-p}\Phi\Big{(}\frac{1}{\lambda}\Big{)}$
$\displaystyle\asymp\frac{s^{2-p}}{s\lambda^{\rho}}$
$\displaystyle=\frac{1}{s}\cdot\Big{(}\frac{s^{\frac{2-p}{\rho}}}{\lambda}\Big{)}^{\rho}$
$\displaystyle\asymp\frac{1}{s}\cdot\Phi\Big{(}\frac{s^{\frac{2-p}{\rho}}}{\lambda}\Big{)}$
Setting the RHS of the last line to $1$ and solving for $\lambda$ we obtain
our result. ∎
###### Proof of Lemma 1.
Let $g=T\tilde{f}$, for some
$\tilde{f}\in\Lambda_{\text{rad}}(\mathbb{R}^{d})\cap
L^{p,q}(\mathbb{R}^{d})$. We may express $\tilde{f}=f(||\cdot||)$ with
$f(r)r^{d-1}$ decreasing. Then, we have that
$\mathcal{F}g(||\cdot||)=\mathcal{F}_{d}\kappa(||\cdot||^{\frac{1}{\alpha}})\mathcal{F}_{d}f(||\cdot||)$.
This follows from the fact that the elements of $\Lambda_{\text{rad}}$ are
radial, and hence so are their Fourier transforms. Moreover, we have that the
decreasing (symmetric) rearrangement $(\mathcal{F}g)^{*}(t)$ (with respect to
the $d$-dimensional Lebesgue measure) is given by
$(\mathcal{F}g)^{*}(t)\asymp\mathcal{F}_{d}\kappa(t^{\frac{1}{d\alpha}})\mathcal{F}_{d}f(t^{\frac{1}{d}})$.
By Lemma 2, the fundamental function of $L^{\psi,t^{\frac{\rho(p+1)}{p}-2}}$
is equivalent to $\Psi(t)=\frac{t^{\frac{p+1}{p}}}{\psi^{-1}(t)}$. Observe
that $\alpha_{\Psi}=\beta_{\Psi}=1+\frac{1}{p}-\frac{1}{\rho}\in(0,1)$ by
assumption. Hence, applying Lemma 1 we have:
$\displaystyle||g||_{\Lambda^{q}_{\Psi}}$
$\displaystyle\asymp\Big{\|}\Psi(t)\cdot
g^{*}(t)\Big{\|}_{L^{q}\big{(}\frac{dt}{t}\big{)}}$
$\displaystyle\asymp\Big{\|}t\Psi\Big{(}\frac{1}{t}\Big{)}\cdot(\mathcal{F}g)^{*}(t)\Big{\|}_{L^{q}\big{(}\frac{dt}{t}\big{)}}$
(19)
$\displaystyle\asymp\Big{\|}\frac{t^{-\frac{1}{p}}}{\psi^{-1}(t^{-1})}\cdot\mathcal{F}_{d}\kappa(t^{\frac{1}{d\alpha}})\mathcal{F}_{d}f(t^{\frac{1}{d}})\Big{\|}_{L^{q}\big{(}\frac{dt}{t}\big{)}}$
$\displaystyle=\Big{\|}\frac{t^{-1}}{\psi^{-1}(t^{-1})}\cdot\mathcal{F}_{d}\kappa(t^{\frac{1}{d\alpha}})t^{1-\frac{1}{p}}\mathcal{F}_{d}f(t^{\frac{1}{d}})\Big{\|}_{L^{q}\Big{(}\frac{dt}{t}\Big{)}}$
$\displaystyle\asymp||t^{1-\frac{1}{p}}\mathcal{F}_{d}f(t^{\frac{1}{d}})||_{L^{q}\Big{(}\frac{dt}{t}\Big{)}}$
(20)
$\displaystyle\asymp||t^{\frac{1}{p}}f(t^{\frac{1}{d}})||_{L^{q}\Big{(}\frac{dt}{t}\Big{)}}$
(21) $\displaystyle\asymp||\tilde{f}||_{L^{p,q}(\mathbb{R}^{d})}$ (22)
where (19) and (21) follow from Lemma 1 after noting that
$\frac{1}{p},1-\frac{1}{p}<1\leq q$ and
$\beta_{\Psi^{1}_{q}},\beta_{\Psi^{2}_{q}}<q$ (since here
$\Psi^{1}_{q}(t)\asymp t\Psi(t^{-1})$ and $\Psi^{2}_{q}(t)\asymp\Psi(t)$ by
the assumption $\alpha_{\psi}=\beta_{\psi}$), (20) follows from observing
$\frac{t^{-1}\mathcal{F}_{d}k(t^{\frac{1}{d\alpha}})}{\psi^{-1}(t^{-1})}\stackrel{{\scriptstyle
t=s^{\alpha}}}{{\asymp}}\frac{\mathcal{F}_{d}k(s^{\frac{1}{d}})}{s^{\alpha}\psi^{-1}(s^{-\alpha})}\asymp
1$ by Assumption 8*, and (22) follows from recalling that
$\tilde{f}\in\Lambda_{\text{rad}}(\mathbb{R}^{d})\cap L^{p,q}(\mathbb{R}^{d})$
is radially decreasing. The optimality of the domain space
$L^{\psi,\frac{\rho(p+1)}{p}-2}$ follows from the asymptotic equivalence in
(20), and Theorem 3.1 in [10]. ∎
###### Remark.
In Assumption 8*, the expression $s(t)$ can be interpreted as a change-of-
basis operation, that converts between the Fourier basis and the Mercer
eigensystem (this interpretation is exact when the eigenspace consists of
complex polynomials, see e.g. Theorem 1 in [27]). It is important to note that
our framework implicitly requires the nonequivalence of $\mu$ and the Lebesgue
(uniform) measure. To see this, suppose for simplicity, $\mathcal{X}=[0,1]$,
and $f_{2n}(x)=\sqrt{2}\cos(2\pi nx),f_{2n+1}(x)=\sqrt{2}\sin(2\pi nx)$. Then,
we have that
$\sup\Big{\\{}\Big{\|}\sum_{i=1}^{n}c_{i}f_{i}(\cdot)\Big{\|}_{\infty}:\sum_{i=1}^{n}c^{2}_{i}=1\Big{\\}}^{2}\asymp
n$, i.e. $s$ is the identity function, which does not satisfy Assumption 4 as
the harmonic series is not summable; in other words, the classical Nikolskii
inequality between $L^{\infty}$ and $L^{2}$ does not suffice. Intuitively, the
exclusion of Lebesgue-equivalent probability measures here is not problematic
— our framework is designed for the setting in which the entropy numbers of
$\mathcal{H}_{K}$ in $L^{2}(\mu)$ are unknown, which is typically not the case
when $\mu$ is Lebesgue-equivalent. Change-of-basis conditions appear
prominently in the study of Fourier multiplier operators associated with
anharmonic oscillators (see [12, 38]). Moreover, the study of the asymptotic
properties of the Christoffel function (square of the left-hand side of
Assumption 4) in relation to the properties of $\mu$ has been an active area
of research [28, 48, 3] with diverse applications to quadrature/interpolation
and random matrices [45] and [3].
## Appendix B Proof of Theorem 2
The main effort in the proof of Theorem 2 involves proving Lemma 3, from which
the Theorem 2 easily follows; hence we will begin with the proof of Lemma 3.
We start with the following bound on the bias:
###### Lemma 1.
$||f^{*}-f_{\lambda}||^{2}_{L^{2}(\nu)}\leq\phi(\lambda)||f^{*}||^{2}_{\phi}$
###### Proof.
Since $f\in\mathcal{H}^{\phi}$, we have that there exists a
$\\{a_{i}\\}_{i=1}^{\infty}\in\ell^{2}$ such that:
$f^{*}=\sum_{i}a_{i}\phi^{\frac{1}{2}}(\mu_{i})e_{i}$
Hence,
$\displaystyle
f^{*}-f_{\lambda}=\sum_{i=1}^{\infty}\frac{a_{i}\lambda\phi^{\frac{1}{2}}(\mu_{i})}{\mu_{i}+\lambda}e_{i}$
(23)
Therefore, by the Cauchy-Schwartz inequality:
$\displaystyle||f^{*}-f_{\lambda}||^{2}_{L^{2}(\nu)}$
$\displaystyle\leq\sum_{i=1}^{\infty}\Big{(}\frac{a_{i}\lambda\phi^{\frac{1}{2}}(\mu_{i})}{\mu_{i}+\lambda}\Big{)}^{2}$
$\displaystyle\leq\Big{(}\sup_{i}\frac{\lambda^{2}\phi(\mu_{i})}{(\mu_{i}+\lambda)^{2}}\Big{)}\sum_{i=1}^{\infty}a^{2}_{i}$
$\displaystyle\leq\Big{(}\sup_{i}\frac{\lambda^{2}\phi(\mu_{i})}{(\mu_{i}+\lambda)^{2}}\Big{)}||f^{*}||^{2}_{\phi}$
Now, observe that for $\mu_{i}<\lambda$:
$\sup_{i:\mu_{i}<\lambda}\frac{\lambda^{2}\phi(\mu_{i})}{(\mu_{i}+\lambda)^{2}}\leq\phi(\lambda)$
as $\phi$ is nondecreasing. For $\mu_{i}\geq\lambda$, we have that:
$\sup_{i:\mu_{i}\geq\lambda}\frac{\lambda^{2}\phi(\mu_{i})}{(\mu_{i}+\lambda)^{2}}=\sup_{i:\mu_{i}\geq\lambda}\frac{\lambda^{2}\mu_{i}^{2}\phi(\mu_{i})}{\mu^{2}_{i}(\mu_{i}+\lambda)^{2}}\leq\sup_{i:\mu_{i}\geq\lambda}\frac{\lambda^{2}\mu_{i}^{2}\phi(\lambda)}{\lambda^{2}(\mu_{i}+\lambda)^{2}}\leq\phi(\lambda)$
where the second inequality follows from the fact that $\frac{\phi(t)}{t^{2}}$
is nonincreasing (from Assumption 2). Putting this together, we obtain our
result. ∎
###### Lemma 2.
$||f^{*}-f_{\lambda}||^{2}_{\infty}\leq\frac{\phi(\lambda)||k_{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}}{\psi(\lambda)}$
###### Proof.
We have from (23):
$\displaystyle||f^{*}-f_{\lambda}||^{2}_{\infty}$
$\displaystyle\leq\Big{|}\sum_{i=1}^{\infty}\frac{a_{i}\lambda\phi(\mu_{i})^{\frac{1}{2}}e_{i}(\cdot)}{\mu_{i}+\lambda}\Big{|}^{2}_{\infty}$
$\displaystyle\leq\Big{(}\sum_{i}\frac{a^{2}_{i}\lambda^{2}\phi(\mu_{i})}{(\mu_{i}+\lambda)^{2}\psi(\mu_{i})}\Big{)}\Big{|}\sum_{i=1}^{\infty}\psi(\mu)e^{2}_{i}(\cdot)\Big{|}_{\infty}$
$\displaystyle\leq\frac{\phi(\lambda)||k^{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}}{\psi(\lambda)}$
where the last line follows from the definition of $||k^{\psi}||$ and the same
logic as in the proof of Lemma 1, after noting that
$\frac{\phi(t)}{t^{2}\psi(t)}$ is decreasing for $t\geq\lambda$ and
$\frac{\phi(t)}{\psi(t)}$ is increasing for $t<\lambda$ by Assumption 2. ∎
We now derive a bound on the variance of our estimator
$f_{\lambda}-f_{D,\lambda}$:
###### Lemma 3.
$||f_{\lambda}-f_{D,\lambda}||_{L^{2}(\nu)}\preceq\log(\delta^{-1})\sqrt{\frac{1152\sigma^{2}||k_{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}}{n\psi(\lambda)}\Big{(}\phi(\lambda)+\frac{1}{n}\Big{)}+\frac{s^{-1}(\psi(\lambda)^{-1})}{n}}$
with probability $1-3\delta$
###### Proof.
We begin with a standard decomposition applied in the proof of Theorem 16 in
[17] and Theorem 7 in [43]. We first observe that:
$f_{\lambda}-f_{D,\lambda}=f_{\lambda}-(C_{D}+\lambda)^{-1}g_{D}=(C_{D}+\lambda)^{-1}((C_{D}+\lambda)f_{\lambda}-g_{D})$
(24)
where $g_{D}=\sum_{i=1}^{n}y_{i}k(x_{i},\cdot)\in\mathcal{H}_{K}$. Hence, we
have that:
$\displaystyle||f_{\lambda}-f_{D,\lambda}||_{L^{2}(\nu)}$
$\displaystyle=||C^{\frac{1}{2}}_{\nu}(f_{\lambda}-f_{D,\lambda})||_{K}$ (25)
$\displaystyle\leq||C^{\frac{1}{2}}_{\nu}(C_{\nu}+\lambda)^{-\frac{1}{2}}||||(C_{\nu}+\lambda)^{\frac{1}{2}}(C_{D}+\lambda)^{-1}(C_{\nu}+\lambda)^{\frac{1}{2}}||||(C_{\nu}+\lambda)^{-\frac{1}{2}}((C_{D}+\lambda)f_{\lambda}-g_{D})||$
(26)
where (25) follows from the fact that for any $f\in\mathcal{H}_{K}$, $\langle
I_{\nu}f,I_{\nu}f\rangle_{2}=\langle f,I^{*}_{\nu}I_{\nu}f\rangle_{K}=\langle
f,C_{\nu}f\rangle_{K}$ and (4), and (26) follows from (24) and algebraic
manipulation. It is easy to see that:
$||C^{\frac{1}{2}}_{\nu}(C_{\nu}+\lambda)^{-\frac{1}{2}}||\leq 1$
We now analyze second factor in (26). Indeed, we have that:
$(C_{\nu}+\lambda)^{\frac{1}{2}}(C_{D}+\lambda)^{-1}(C_{\nu}+\lambda)^{\frac{1}{2}}=\Big{(}I-(C_{\nu}+\lambda)^{-\frac{1}{2}}(C_{\nu}-C_{D})(C_{\nu}+\lambda)^{-\frac{1}{2}}\Big{)}^{-1}$
by the same algebraic manipulation as in the proof of Theorem 16 in [17] (for
the sake of space, we avoid repeating the argument verbatim here). Hence,
estimating the middle factor in (26) boils down to estimating the
concentration of
$(C_{\nu}+\lambda)^{-\frac{1}{2}}(C_{\nu}-C_{D})(C_{\nu}+\lambda)^{-\frac{1}{2}}$.
We first note that:
$(C_{\nu}+\lambda)^{-\frac{1}{2}}(C_{\nu}-C_{D})(C_{\nu}+\lambda)^{-\frac{1}{2}}=\mathbb{E}_{X}[h(X,\cdot)\otimes
h(X,\cdot)]-\mathbb{E}_{\mathcal{D}}[h(X,\cdot)\otimes h(X,\cdot)]$
where $h(x,\cdot)=(C_{\nu}+\lambda)^{-\frac{1}{2}}k(x,\cdot)$. Note, by Lemma
2, we have that
$||h(x,\cdot)||_{K}\leq\sqrt{\frac{||k^{\psi}||^{2}_{\infty}}{\psi(\lambda)}}$.
Then, analogously to Lemma 17 in [17], we may demonstrate that:
$(C_{\nu}+\lambda)^{-\frac{1}{2}}(C_{\nu}-C_{D})(C_{\nu}+\lambda)^{-\frac{1}{2}}\leq\frac{2}{3}$
with probability $1-2\delta$, when $n\geq
8\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}g_{\lambda}\psi(\lambda)^{-1}$,
where
$g_{\lambda}=\log(2e\mathcal{N}(\lambda)\Big{(}1+\frac{\lambda}{||C_{\nu}||}\Big{)})$,
where $\mathcal{N}(\lambda)=\text{tr}(C_{\nu}(C_{\nu}+\lambda)^{-1})$ Applying
a Neumann series expansion, we obtain (like in [17]):
$\|(C_{\nu}+\lambda)^{\frac{1}{2}}(C_{D}+\lambda)^{-1}(C_{\nu}+\lambda)^{\frac{1}{2}}\|=||\Big{(}I-(C_{\nu}+\lambda)^{-\frac{1}{2}}(C_{\nu}-C_{D})(C_{\nu}+\lambda)^{-\frac{1}{2}}\Big{)}^{-1}||\leq
3$
for $n\geq
8\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}g_{\lambda}\psi(\lambda)^{-1}$.
Now, it remains to bound the third factor in (26). We observe that we may
write:
$\displaystyle(C_{\nu}+\lambda)^{-\frac{1}{2}}((C_{D}+\lambda)f_{\lambda}-g_{D})$
$\displaystyle=(C_{\nu}+\lambda)^{-\frac{1}{2}}((C_{D}-C_{\nu})f_{\lambda}+S_{\nu}f^{*}-g_{D})$
$\displaystyle=\mathbb{E}_{\mathcal{D}}[h(X,\cdot)(f_{\lambda}(X)-Y)]-\mathbb{E}_{XY}[h(X,\cdot)(f_{\lambda}(X)-Y)]$
which follow by the definition of $f_{\lambda}$ and $h(x,\cdot)$. In order to
estimate this concentration, we compute absolute moments:
$\displaystyle\mathbb{E}_{XY}[||h(X,\cdot)(f_{\lambda}(X)-Y)||_{K}^{p}]$
$\displaystyle\leq\mathbb{E}_{XY}[||h(X,\cdot)(f_{\lambda}(X)-f^{*}(X)+f^{*}(X)-Y)||_{K}^{p}]$
$\displaystyle\leq
2^{p-1}\mathbb{E}_{XY}[||h(X,\cdot)||^{p}_{K}(|f_{\lambda}(X)-f^{*}(X)|^{p}+|f^{*}(X)-Y|^{p})]$
$\displaystyle\leq
2^{p-1}\mathbb{E}_{X}\Big{[}||h(X,\cdot)||^{p}_{K}\Big{(}\|f_{\lambda}-f^{*}\|_{\infty}^{p-2}|f_{\lambda}(X)-f^{*}(X)|^{2}+\frac{p!L^{p-2}\sigma^{2}}{2}\Big{)}\Big{]}$
(27) $\displaystyle\leq
2^{p-1}\Big{(}\phi^{\frac{p}{2}}(\lambda)\Big{(}\frac{||k_{\psi}||_{\infty}||f^{*}||_{\phi}}{\psi(\lambda)}\Big{)}^{p}+\frac{p!\sigma^{2}}{2}\Big{(}\frac{||k_{\psi}||^{2}_{\infty}L^{2}}{\psi(\lambda)}\Big{)}^{\frac{p-2}{2}}\text{tr}(C_{\nu}(C_{\nu}+\lambda)^{-1})\Big{)}$
(28) $\displaystyle\preceq
2^{p-1}\Big{(}\Big{(}\frac{||k_{\psi}||_{\infty}||f^{*}||_{\phi}L}{\sqrt{\psi(\lambda)}}\Big{)}^{p-2}\Big{[}\Big{(}\frac{||k_{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}\phi(\lambda)}{\psi(\lambda)}\Big{)}+\sigma^{2}p!\Gamma\Big{(}\frac{p}{2}\Big{)}s^{-1}(\psi(\lambda)^{-1})\Big{]}\Big{)}$
(29) $\displaystyle\leq\frac{p!L_{\lambda}^{p-2}R^{2}}{2}$
with
$L_{\lambda}=\frac{2\sigma||f^{*}||_{\phi}||k_{\psi}||_{\infty}}{\sqrt{\psi(\lambda)}}$
and
$R^{2}=\frac{4||f^{*}||^{2}_{\phi}||k_{\psi}||^{2}_{\infty}\phi(\lambda)}{\psi(\lambda)}+4\sigma^{2}s^{-1}(\psi(\lambda)^{-1})$.
Here, (27) follows from Assumption 5, (28) invokes Lemmas 2, 1 and 2, and
finally (29) follows from Lemma 3 and the fact that
$\frac{\phi(\lambda)}{\psi(\lambda)}\to 0$ as $\lambda\to 0$. Hence, applying
Theorem 26 in [17], we have that:
$||(C_{\nu}+\lambda)^{-\frac{1}{2}}((C_{D}+\lambda)f_{\lambda}-g_{D})||^{2}_{K}\preceq\frac{32\log^{2}(\delta^{-1})}{n}\Big{(}\frac{4||f^{*}||^{2}_{\phi}||k_{\psi}||^{2}_{\infty}\phi(\lambda)}{\psi(\lambda)}+4\sigma^{2}s^{-1}(\psi(\lambda)^{-1})+\frac{4\sigma^{2}||k_{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}}{n\psi(\lambda)}\Big{)}$
with probability $1-2\delta$. Putting this all together, we have that:
$||f_{\lambda}-f_{D,\lambda}||_{L^{2}(\nu)}\preceq\log(\delta^{-1})\sqrt{\frac{1152\sigma^{2}||k_{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}}{n\psi(\lambda)}\Big{(}\phi(\lambda)+\frac{1}{n}\Big{)}+\frac{s^{-1}(\psi(\lambda)^{-1})}{n}}$
(30)
with probability $1-3\delta$, when $n\geq
8\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}g_{\lambda}\psi(\lambda)^{-1}$. We
demonstrate that this latter condition is eventually met, when the
regularization program is
$\lambda_{n}\asymp(\frac{\phi}{\tilde{s}\circ\psi})^{-1}(n^{-1})$. Indeed, we
have that:
$\displaystyle\frac{8\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}g_{\lambda_{n}}}{n\psi(\lambda_{n})}$
$\displaystyle=\frac{8\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}\log\Big{(}2e\mathcal{N}(\lambda_{n})\Big{(}1+\frac{\lambda_{n}}{||C_{\nu}||}\Big{)}\Big{)}}{n\psi(\lambda_{n})}$
$\displaystyle\leq\frac{16\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}\log\Big{(}2e\mathcal{N}(\lambda_{n})\Big{)}}{n\psi(\lambda_{n})}$
$\displaystyle\leq\frac{16\log(\delta^{-1})||k^{\psi}||^{2}_{\infty}\log\Big{(}s^{-1}(\psi(\lambda_{n})^{-1})\Big{)}}{n\psi(\lambda_{n})}$
where the last line follows by Lemma 3. Hence, it is sufficient to demonstrate
that
$\frac{\log\Big{(}s^{-1}(\psi(\lambda_{n})^{-1})\Big{)}}{n\psi(\lambda_{n})}\to
0$ for the regularization program. Indeed, since
$\lambda_{n}\asymp(\frac{\phi}{\tilde{s}\circ\psi})^{-1}(n^{-1})$, we have
that:
$\displaystyle\frac{\log\Big{(}s^{-1}(\psi(\lambda_{n})^{-1})\Big{)}}{n\psi(\lambda_{n})}$
$\displaystyle\preceq\frac{\log\Big{(}s^{-1}(\psi(\lambda_{n})^{-1})\Big{)}}{n\phi(\lambda_{n})}$
(31)
$\displaystyle\asymp\frac{\log\Big{(}s^{-1}(\psi(\lambda_{n})^{-1})\Big{)}}{\tilde{s}(\psi(\lambda_{n}))}$
$\displaystyle=\frac{\log\Big{(}s^{-1}(\psi(\lambda_{n})^{-1})\Big{)}}{s^{-1}(\psi(\lambda_{n})^{-1})}\to
0$ (32)
where (31) follows from Assumption 2 and (32) follows from the definition of
$\tilde{s}$.
###### Proof of Lemma 3.
This follows immediately by combining Lemma 3 with Lemma 1. ∎
###### Proof of Theorem 2.
From Lemma 3, we have:
$\displaystyle||f^{*}-f_{D,\lambda_{n}}||_{L^{2}(\nu)}$
$\displaystyle\preceq||f^{*}||_{\phi}\sqrt{\phi(\lambda_{n})}+\log(\delta^{-1})\sqrt{\frac{\sigma^{2}||k_{\psi}||^{2}_{\infty}||f^{*}||^{2}_{\phi}}{n\psi(\lambda_{n})}\Big{(}\phi(\lambda_{n})+\frac{1}{n}\Big{)}+\frac{s^{-1}(\psi(\lambda_{n})^{-1})}{n}}$
$\displaystyle\preceq||f^{*}||_{\phi}\sqrt{\phi(\lambda_{n})}+\log(\delta^{-1})\sqrt{\frac{s^{-1}(\psi(\lambda_{n})^{-1})}{n}}$
(33)
where the last line follows from the fact that
$\frac{\phi(\lambda_{n})}{\psi(\lambda_{n})}\to 0$ by Assumption 2 and
$\phi(\lambda_{n})\asymp\frac{\tilde{s}(\psi(\lambda_{n}))}{n}=\frac{s^{-1}(\psi(\lambda_{n})^{-1})}{n}\succeq\frac{1}{n}$
as $s^{-1}(\psi(\lambda_{n})^{-1})\to\infty$ when
$\lambda_{n}\asymp(\frac{\phi}{\tilde{s}\circ\psi})^{-1}(n^{-1})$. The result
follows immediately after substituting the latter choice of regularizer into
(33) (which was indeed proposed via balancing (33)). ∎
∎
## Appendix C Proof of Theorem 4
###### Lemma 1.
$b^{2}_{n-1}(\mathcal{H}_{K},L^{\infty})\leq||k^{\psi}||^{2}_{\infty}\Big{(}\frac{t}{\psi}\Big{)}(\mu_{n})$
###### Proof.
We first note that since $\frac{t}{\psi(t)}$ is concave by Assumption 2, we
have by the Gagliardo-Nirenberg interpolation inequality (Lemma 1), that for
all $f\in\mathcal{H}_{K}$:
$\frac{||f||^{2}_{\psi}}{||f||^{2}_{K}}\leq\Big{(}\frac{t}{\psi}\Big{)}\Big{(}\frac{||f||^{2}_{L^{2}(\nu)}}{||f||^{2}_{K}}\Big{)}$
By the embedding condition $\mathcal{H}^{\psi}\hookrightarrow L^{\infty}$, we
have that:
$\frac{||f||^{2}_{\infty}}{||f||^{2}_{K}}\leq||k^{\psi}||^{2}_{\infty}\Big{(}\frac{t}{\psi}\Big{)}\Big{(}\frac{||f||^{2}_{L^{2}(\nu)}}{||f||^{2}_{K}}\Big{)}$
Hence, we have that:
$\displaystyle\sup_{dim(Z)\geq n}\inf_{f\in
Z\cap\mathcal{H}_{K}}\frac{||f||^{2}_{\infty}}{||f||^{2}_{K}}$
$\displaystyle\leq\sup_{dim(Z)\geq n}\inf_{f\in
Z\cap\mathcal{H}_{K}}||k^{\psi}||^{2}_{\infty}\Big{(}\frac{t}{\psi}\Big{)}\Big{(}\frac{||f||^{2}_{L^{2}(\nu)}}{||f||^{2}_{K}}\Big{)}$
$\displaystyle=||k^{\psi}||^{2}_{\infty}\Big{(}\frac{t}{\psi}\Big{)}\Big{(}\sup_{dim(Z)\geq
n}\inf_{f\in
Z\cap\mathcal{H}_{K}}\frac{||f||^{2}_{L^{2}(\nu)}}{||f||^{2}_{K}}\Big{)}$
where the last step follows from the fact that $\frac{t}{\psi(t)}$ is
nondecreasing and continuous. Hence, by the definition of the Bernstein width
(10), and Proposition 5 in [29], we obtain our result. ∎
###### Lemma 2.
$b_{n-1}(\mathcal{H}_{K},L^{\infty})\geq\sqrt{C_{d}\mathcal{F}_{d}\kappa(n^{\frac{1}{d}})}$
where $C_{d}>0$ is constant depending on the dimension $d$ and independent of
$n$.
###### Proof.
Let $\\{z_{i}\\}_{i=1}^{2n}\subset\mathcal{X}$ be a set of $2n$ distinct
points in $\mathcal{X}$. Consider the subspace
$Z=\text{span}\\{k(z_{i},\cdot)\\}_{i=1}^{2n}\subset\mathcal{H}_{K}$. Since,
$k$ is positive definite by assumption, we have that $\text{dim}(Z)=2n$. Since
$\mathcal{H}_{K}$ is a Hilbert space, $Z$ is isomorphic to $\mathbb{R}^{2n}$
(with the Euclidean norm). By Theorem IV.3.1 in [46] (see also Theorem 1 in
[33]), there exists a subspace $\tilde{Z}\subset Z$ of dimension $n$ and
constant $C>0$, such that:
$||z||_{K}\leq\sqrt{\frac{C}{2}}M^{*}(B(L^{\infty})\cap Z)$
for all $z\in\tilde{Z}\cap B(L^{\infty})$ (we will absorb additional constants
into $C$ throughout this proof). Here, $M^{*}(B(L^{\infty})\cap Z)$ denotes
the mean width of $B(L^{\infty})\cap Z$ given by:
$M^{*}(B(L^{\infty})\cap Z)=\int_{\partial B(Z)}\sup_{x\in B(L^{\infty})\cap
Z}\langle x,y\rangle_{K}dS_{2n}(y)$
where $S_{2n}(\cdot)$ is the canonical uniform measure over the unit sphere
$\partial B(Z)$ in $Z$. Let $\gamma_{2n}$ denote the standard Gaussian measure
on $\mathbb{R}^{2n}$, and $\mu_{2n}$ its pushforward onto $Z$ via the
canonical isomorphism. We write $g=\sum_{i=1}^{2n}a_{i}(g)k(z_{i},\cdot)$ for
$g\in Z$ (where $\\{a_{i}(g)\\}_{i=1}^{2n}\in\mathbb{R}^{2n}$). Then may
compute, via a standard change of variables, that for an absolute constant
$C_{1}>0$:
$\displaystyle M^{*}(B(L^{\infty})\cap Z)$ $\displaystyle=\int_{\partial
B(Z)}\sup_{f\in B(L^{\infty})\cap Z}\langle f,g\rangle_{K}dS_{2n}(g)$
$\displaystyle=\frac{C_{1}}{\sqrt{2n}}\int_{Z}\sup_{f\in B(L^{\infty})\cap
Z}\langle f,g\rangle_{K}d\mu_{2n}(g)$
$\displaystyle=\frac{C_{1}}{\sqrt{2n}}\int_{Z}\sup_{f\in B(L^{\infty})\cap
Z}\Big{\langle}f,\sum_{i=1}^{2n}a_{i}(g)k(z_{i},\cdot)\Big{\rangle}_{K}d\mu_{2n}(g)$
$\displaystyle=\frac{C_{1}}{\sqrt{2n}}\int_{Z}\sup_{f\in B(L^{\infty})\cap
Z}\sum_{i=1}^{2n}a_{i}(g)f(z_{i})d\mu_{2n}(g)$
$\displaystyle\leq\frac{C_{1}}{\sqrt{2n}}\int_{Z}\sum_{i=1}^{2n}|a_{i}(g)|d\mu_{2n}(g)$
since $|f(z_{i})|\leq 1$ as $f\in B(L^{\infty})$. Now, observe that
$g\in\partial B(Z)$ iff $\mathbf{a}(g)\cdot(K_{2n}\mathbf{a}(g))=1$, where
$\mathbf{a}(g)=[a_{1}(g),\ldots a_{2n}(g)]^{T}$ and
$(K_{2n})_{i,j}=k(z_{i},z_{j})$ is the kernel Gram matrix on
$\\{z_{i}\\}_{i=1}^{2n}$. Hence, performing a change of variables (and
renaming), we obtain, for a constant $C_{1}>0$:
$\displaystyle M^{*}(B(L_{\infty})\cap Z)$
$\displaystyle\leq\frac{C_{1}}{\sqrt{2n}}\int_{\mathbb{R}^{2n}}|K^{-\frac{1}{2}}_{2n}\mathbf{a}|_{1}d\gamma_{2n}(\mathbf{a})$
$\displaystyle\leq
C_{1}\int_{\mathbb{R}^{2n}}\sqrt{||K^{-1}_{2n}||_{2}||\mathbf{a}||^{2}}d\gamma_{2n}(\mathbf{a})$
$\displaystyle\leq C_{1}\sqrt{2n||K^{-1}_{2n}||_{2}}$
since $\int_{\mathbb{R}^{2n}}||\mathbf{a}||^{2}d\gamma_{2n}(a)=2n$ as
$\gamma_{2n}$ is standard normal. Hence, putting this all together, we have:
$\displaystyle||z||_{K}\leq\sqrt{C_{3}n||K^{-1}_{2n}||_{2}}$
for all $z\in\tilde{Z}\cap B(L_{\infty})$. Hence, for all $z\in\tilde{Z}$, we
have that:
$||z||_{K}\leq\sqrt{C_{3}n||K^{-1}_{2n}||_{2}}||z||_{\infty}$
where $C_{3}=CC_{1}$. By the definition of the Bernstein width, we therefore
have:
$b_{n-1}(\mathcal{H}_{K},L^{\infty})\geq\sqrt{\frac{C}{n||K^{-1}_{2n}||_{2}}}$
(34)
We now wish to upper bound $||K^{-1}_{2n}||_{2}$ — we do this by using the
following lower bound on the minimum eigenvalue of $K_{2n}$, given in Theorem
12.3 of [47]:
$\lambda_{\min}(K_{2n})\geq
C_{d}\Big{(}\frac{12.76d}{q_{z}}\Big{)}^{d}\mathcal{F}_{d}\kappa\Big{(}\frac{12.76d}{q_{z}}\Big{)}$
where $q_{z}\equiv\min_{i\neq j}||z_{i}-z_{j}||$ and $C_{d}>0$ is an absolute
constant depending only the dimension $d$. We choose a maximal $2n$-set
$\\{z_{i}\\}_{i=1}^{2n}$, so that $q_{z}$ is the packing number
$p_{n}(\mathcal{X})$ of the domain $\mathcal{X}$. From the standard
equivalence $p_{n}(\mathcal{X})\asymp\epsilon_{n}(\mathcal{X})$ and Assumption
7*, we have that:
$\lambda_{\min}(K_{2n})\geq
C_{d}n\mathcal{F}_{d}\kappa\Big{(}12.76d(2n)^{\frac{1}{d}}\Big{)}$
From the integral expression for the $d$-dimensional Fourier transform (see
e.g. Theorem 5.26 in [47]), and the dilation condition in Assumption 6*, we
therefore have that:
$\lambda_{\min}(K_{2n})\geq C_{d}n\mathcal{F}_{d}\kappa(n^{\frac{1}{d}})$
where again we recall that the overhead constant $C_{d}$ depends only on the
ambient dimension in Assumption 7*. Noting that
$\|K^{-1}_{2n}\|_{2}\leq\lambda^{-1}_{\min}(K_{2n})$ and plugging back into
(34), we have that:
$b_{n-1}(\mathcal{H}_{K},L^{\infty})\geq\sqrt{C_{d}\mathcal{F}_{d}\kappa(n^{\frac{1}{d}})}$
∎
###### Corollary 2.1.
$\mu_{n}\asymp\psi^{-1}(s(n)^{-1})$
###### Proof.
From Assumption 4, for any $\alpha\in(0,1)$, we can choose a $c\in l^{n}_{2}$
with $\sum_{i=1}^{n}c^{2}_{i}=1$ and $x^{*}\in\mathcal{X}$, such that:
$\sum_{i=1}^{n}c_{i}e_{i}(x^{*})\succeq\alpha\sqrt{s(n)}$
Hence, noting that $\\{c_{i}\\}_{i=1}^{n}$ can always be chosen such that
$c_{i}e_{i}(x^{*})\geq 0$, for all $i\in[n]$, we have that that:
$\displaystyle\alpha\sqrt{\psi(\mu_{n})s(n)}$
$\displaystyle\preceq\sum_{i=1}^{n}\sqrt{\psi(\mu_{i})}c_{i}e_{i}(x^{*})$
$\displaystyle\leq||k^{\psi}||_{\infty}$
Observing that $||k^{\psi}||_{\infty}<\infty$ since $H^{\psi}\hookrightarrow
L^{\infty}$ in Assumption 1 and that $\alpha$ was chosen arbitrarily, we
obtain the upper estimate:
$\mu_{n}\preceq\psi^{-1}(s(n)^{-1})$
For the lower estimate, we apply Lemma 1 and Lemma 2 to obtain:
$\frac{t}{\psi}\Big{(}\mu_{n}\Big{)}\succeq
b^{2}_{n-1}(\mathcal{H}_{K},L^{\infty})\succeq\mathcal{F}_{d}k(n^{\frac{1}{d}})$
(35)
Now, from Assumption 8*, we have that:
$\displaystyle\psi^{-1}(s(t)^{-1})$
$\displaystyle=\Big{(}\frac{t}{\psi}\Big{)}^{-1}\Big{(}\Big{(}\frac{t}{\psi}\Big{)}(\psi^{-1}(s(t)^{-1}))\Big{)}$
$\displaystyle\Big{(}\frac{t}{\psi}\Big{)}^{-1}\Big{(}\frac{\psi^{-1}(s(t)^{-1})}{s(t)^{-1}}\Big{)}$
$\displaystyle\asymp\Big{(}\frac{t}{\psi}\Big{)}^{-1}\Big{(}\frac{s(t)\mathcal{F}_{d}k(t^{\frac{1}{d}})}{s(t)}\Big{)}$
$\displaystyle=\Big{(}\frac{t}{\psi}\Big{)}^{-1}(\mathcal{F}_{d}k(t^{\frac{1}{d}}))$
Hence, combining this with (35), we have obtain the lower estimate:
$\mu_{n}\succeq\psi^{-1}(s(n)^{-1})$
∎
###### Remark.
Observe, that in light of the proof of Lemma 2, Corollary 2.1 would still hold
if in Assumption 8*, $\mathcal{F}_{d}\kappa(t^{\frac{1}{d}})$ was replaced by
$\mathcal{F}_{d}\kappa\Big{(}12.76d(2t)^{\frac{1}{d}}\Big{)}$. However, in
order to streamline the analysis, we have chosen the simpler assumption, as
also discussed in section 2.4.
###### Proof of Theorem 4.
We follow the general outline of the proof of Theorem 2 in [17] that was
earlier utilized in [8] and [4]. Our goal is to construct a sequence of
probability measures $\\{P_{j}\\}_{j=1}^{M_{\epsilon}}$ on
$\mathcal{X}\times\mathbb{R}$ with $P_{j}|_{\mathcal{X}}=\nu$ (i.e. the
measures share the same marginal on $\mathcal{X}$) that are hard to learn,
i.e. their regression functions $f_{j}$ satisfy the source conditions but are
sufficiently “well spread out” in $L^{2}(\nu)$ so that it is sufficiently hard
to distinguish between these functions given only the sample data. We make
these ideas more precise in the subsequent analysis. Our candidate regression
functions take the following form — for some fixed $\epsilon\in(0,1)$ and
$m\in\mathbb{N}$, we consider the element:
$f_{\omega}=2\sqrt{\frac{8\epsilon}{m}}\sum_{i=1}^{m}\omega_{i}e_{i+m}$
where $\omega\in\\{0,1\\}^{m}$ is some binary string. Since the sum is finite,
we have that $f\in\mathcal{H}_{K}\subset L^{\infty}\cap\mathcal{H}^{\phi}$. We
will demonstrate that $||f||_{\phi}\leq B_{\phi}$ and $||f||_{\infty}\leq
B_{\infty}$ for sufficiently large choices of $m$. Indeed, we have that:
$\displaystyle||f_{\omega}||^{2}_{\phi}$
$\displaystyle=\frac{32\epsilon}{m}\sum_{i=1}^{m}\frac{\omega^{2}_{i}}{\phi(\mu_{i+m})}\leq\frac{32\epsilon}{\phi(\mu_{2m})}$
$\displaystyle||f_{\omega}||^{2}_{\infty}$
$\displaystyle\leq||k^{\psi}||_{\infty}^{2}||f_{\omega}||^{2}_{\psi}=\frac{32||k^{\psi}||_{\infty}^{2}\epsilon}{m}\sum_{i=1}^{m}\frac{\omega^{2}_{i}}{\psi(\mu_{i+m})}\leq\frac{32||k^{\psi}||_{\infty}^{2}\epsilon}{\psi(\mu_{2m})}$
Now, by Corollary 2.1, we have that:
$\displaystyle||f_{\omega}||^{2}_{\phi}$
$\displaystyle\leq\frac{32\epsilon}{\phi(C_{d}\psi^{-1}(s(m)^{-1}))}$
$\displaystyle||f_{\omega}||^{2}_{\infty}$
$\displaystyle\leq\frac{64||k^{\psi}||_{\infty}^{2}\epsilon}{C_{d}s(m)^{-1}}$
where $C_{d}>0$ depends only $d$ (and the constants in the $\Delta_{2}$
condition on $\phi,\psi$, and $s$). Hence, by the $\Delta_{2}$ condition on
$\phi$ and $\psi$, we have that $||f||^{2}_{\phi}\leq B^{2}_{\phi}$ and
$||f||^{2}_{\infty}\leq B^{2}_{\infty}$, when:
$m\leq
s^{-1}\Big{(}\min\Big{\\{}\frac{C_{d}B^{2}_{\infty}}{64||k^{\psi}||_{\infty}^{2}\epsilon},\Big{(}\psi\Big{(}C_{d}\phi^{-1}\Big{(}\frac{32\epsilon}{B^{2}_{\phi}}\Big{)}\Big{)}\Big{)}^{-1}\Big{\\}}\Big{)}$
(where we have renamed the constant $C_{d}$, as we will throughout this
proof). By Assumption 2, we have that $\frac{\phi(t)}{\psi(t)}\to 0$ as $t\to
0$. Hence, for sufficiently small $t>0$, we have that
$\frac{\psi(\phi^{-1}(t))}{t}=\frac{\psi(\phi^{-1}(t))}{\phi(\phi^{-1}(t))}\geq
1$. Hence, we may choose $\epsilon_{1}\in(0,1)$, such that for
$\epsilon\in(0,\epsilon_{1}]$, we have $||f||^{2}_{\phi}\leq B^{2}_{\phi}$ and
$||f||^{2}_{\infty}\leq B^{2}_{\infty}$, when:
$m\leq
U_{d,\infty,\phi}s^{-1}\Big{(}\frac{1}{\psi(\phi^{-1}(\epsilon))}\Big{)}=U_{d,\infty,\phi}\tilde{s}(\psi(\phi^{-1}(\epsilon)))$
where we have recalled the definition of
$\tilde{s}(t)=s^{-1}\Big{(}\frac{1}{t}\Big{)}$ and $U_{d,\infty,\phi}>0$ is a
constant depending only on $d,B_{\infty},B_{\phi}$ that we have again
extracted via the $\Delta_{2}$ conditions on $s$, $\psi$ and $\phi$. We can
ensure that there is an $m\geq 1$ satisfying this bound by choosing
$\epsilon\leq\epsilon_{2}=\min\\{\epsilon_{1},\phi(\psi^{-1}(s(U_{d,\infty,\phi}^{-1})^{-1}))\\}$.
Moreover, we have for any two binary strings
$\omega^{(1)},\omega^{(2)}\in\\{0,1\\}^{m}$:
$||f_{\omega^{(1)}}-f_{\omega^{(2)}}||^{2}_{L^{2}(\nu)}=\frac{32\epsilon}{m}\sum_{i=1}^{m}(\omega^{(1)}_{i}-\omega^{(2)}_{i})^{2}\leq
32\epsilon$
Now, observe that if we choose
$\epsilon\leq\epsilon_{3}\equiv\min\\{\epsilon_{2},\phi(\psi^{-1}(s(9U_{d,\infty,\phi}^{-1})^{-1}))\\}$,
and choose $m_{\epsilon}=\lfloor
U_{d,\infty,\phi}\tilde{s}(\psi(\phi^{-1}(\epsilon)))\rfloor$ like before, we
have $m_{\epsilon}\geq 9$. Then, by the Gilbert-Varshamov bound (see e.g.
Lemma 24 in [17]), there exists a $M_{\epsilon}\geq
2^{\frac{m_{\epsilon}}{8}}$, and some binary strings
$\omega^{(1)},\omega^{(2)},\ldots,\omega^{(M_{\epsilon})}$, such that:
$\sum_{i=1}^{m_{\epsilon}}(\omega^{(k)}_{i}-\omega^{(l)}_{i})^{2}\geq\frac{m_{\epsilon}}{8}$
for all $k,l\in[M_{\epsilon}]$. Therefore, we have:
$||f_{\omega^{(k)}}-f_{\omega^{(l)}}||^{2}_{L^{2}(\nu)}\geq 4\epsilon$ (36)
for $k,l\in[M_{\epsilon}]$. We now construct measures
$\\{P_{j}\\}_{j=1}^{M_{\epsilon}}$ on $\mathcal{X}\times\mathcal{Y}$ such that
$P_{j}|_{\mathcal{X}}=\nu=P_{X}$, with
$\mathbb{E}[Y|X]=f_{\omega^{(j)}}(X)\equiv f_{j}(X)$ and
$Y-\mathbb{E}[Y|X]\sim\mathcal{N}(0,\bar{\sigma}^{2})$ for $(X,Y)\sim P_{j}$
and all $j\in[M_{\epsilon}]$ (where $\bar{\sigma}=\max\\{\sigma,L\\}$). The
measures $P_{j}$ satisfy Assumption 5 by Lemma 21 in [17]. Letting $P_{0}$ be
such that $f_{0}=0$. Then, we have that:
$\frac{1}{M_{\epsilon}}\sum_{i=1}^{M_{\epsilon}}\text{KL}(P^{n}_{i}||P^{n}_{0})=\frac{n}{2\sigma^{2}M_{\epsilon}}\sum_{i=1}^{M_{\epsilon}}||f_{i}-f_{0}||^{2}_{L^{2}(\nu)}\leq\frac{16n\epsilon}{\sigma^{2}}\equiv\alpha^{*}$
Thus, for any measurable function
$\Theta:(\mathcal{X}\times\mathbb{R})^{n}\to\\{0,1,\ldots,M_{\epsilon}\\}$, we
have by Theorem 20 in [17] that:
$\max_{j\in[M_{\epsilon}]}P_{j}(\Theta(D)\neq
j)\geq\frac{\sqrt{M_{\epsilon}}}{1+\sqrt{M_{\epsilon}}}\Big{(}1-\frac{48n\epsilon}{\sigma^{2}\log
M_{\epsilon}}-\frac{1}{2\log M_{\epsilon}}\Big{)}$ (37)
Note that $M_{\epsilon}\geq 2^{\frac{m_{\epsilon}}{8}}\geq
2^{\frac{U_{d,\infty,\phi}\tilde{s}(\psi(\phi^{-1}(\epsilon)))}{9}}$ using the
definition of $m_{\epsilon}$ and the fact that $m_{\epsilon}\geq 9$.
Substituting this into (37), we obtain:
$\max_{j\in[M_{\epsilon}]}P_{j}(\Theta(D)\neq
j)\geq\frac{\sqrt{M_{\epsilon}}}{1+\sqrt{M_{\epsilon}}}\Big{(}1-\frac{432n\epsilon}{\tilde{s}(\psi(\phi^{-1}(\epsilon)))\cdot
U_{d,\infty,\phi}\sigma^{2}\log 2}-\frac{1}{2\log M_{\epsilon}}\Big{)}$
Now, we choose
$\epsilon_{n}=\tau\phi\Big{(}(\frac{\phi}{\tilde{s}\circ\psi})^{-1}(n^{-1})\Big{)}$
where $\tau<1$ has been chosen so that $\epsilon_{n}\leq\epsilon_{3}$. Define:
$\Theta(D)\equiv\text{arg}\min_{j\in[M_{\epsilon_{n}}]}||f_{D}-f_{j}||_{L^{2}(\nu)}$
Then for any $j\neq\Theta(D)$, we have by (36):
$2\sqrt{\epsilon_{n}}\leq||f_{\Theta(D)}-f_{j}||_{L^{2}(\nu)}\leq||f_{D}-f_{\Theta(D)}||_{L^{2}(\nu)}+||f_{D}-f_{j}||_{L^{2}(\nu)}\leq
2||f_{D}-f_{j}||_{L^{2}(\nu)}$
Thus, we have:
$\displaystyle\max_{j\in[M_{\epsilon_{n}}]}P_{j}(||f_{D}-f_{j}||_{L^{2}(\nu)}\geq\sqrt{\epsilon_{n}})$
$\displaystyle\geq\max_{j\in[M_{\epsilon_{n}}]}P_{j}(\Theta(D)\neq j)$
$\displaystyle\geq\frac{\sqrt{M_{\epsilon_{n}}}}{1+\sqrt{M_{\epsilon_{n}}}}\Big{(}1-\frac{432n\epsilon_{n}}{\tilde{s}(\psi(\phi^{-1}(\epsilon_{n})))\cdot
U_{d,\infty,\phi}\sigma^{2}\log 2}-\frac{1}{2\log M_{\epsilon_{n}}}\Big{)}$
$\displaystyle\geq\frac{\sqrt{M_{\epsilon_{n}}}}{1+\sqrt{M_{\epsilon_{n}}}}\Big{(}1-C_{\tau}-\frac{1}{2\log
M_{\epsilon_{n}}}\Big{)}$
after substituting the choice of $\epsilon_{n}$ and letting
$C_{\tau}=\frac{432\tau K^{u}}{U_{d,\infty,\phi}\sigma^{2}\log 2}$ (where
$u=\log\tau$ and $K=K(\psi,s,\phi)>1$ is a constant depending only on
$D^{1}_{\psi},D^{2}_{\psi},D^{1}_{\phi},D^{2}_{\phi},D^{1}_{s},D^{2}_{s}>1$).
Observing that $M_{\epsilon_{n}}\to\infty$ as $\epsilon_{n}\to 0$ (as
$n\to\infty$), we obtain our result. ∎
## Appendix D Auxiliary Results
###### Lemma 1.
Suppose $\frac{t}{\psi(t)}$ is concave. Then, if Assumption 1 holds, we have
for all $f\in\mathcal{H}_{K}$:
$\frac{||f||^{2}_{\psi}}{||f||^{2}_{K}}\leq\frac{t}{\psi}\Big{(}\frac{||f||^{2}_{L^{2}(\nu)}}{||f||^{2}_{K}}\Big{)}$
###### Proof.
Follows immediately from Proposition 7 in [29] with $\kappa=\sqrt{t}$ and
$\phi=\sqrt{\psi(t)}$ (note here $\kappa$ and $\phi$ denote notation used in
[29] and do not coincide with the objects $\kappa$ and $\phi$ in this paper).
∎
###### Lemma 2.
$\sup_{x\in\mathcal{X}}||(C_{\nu}+\lambda)^{-\frac{1}{2}}k(x,\cdot)||_{K}\leq\sqrt{\frac{||k^{\psi}||_{\infty}^{2}}{\psi(\lambda)}}$
###### Proof.
We first note that, for any linear operator $T$ on $\mathcal{H}_{K}$, we have:
$\displaystyle\sup_{x\in\mathcal{X}}||Tk(x,\cdot)||_{K}$
$\displaystyle=\sup_{||f||_{K}=1,x\in\mathcal{X}}\langle
Tk(x,\cdot),f\rangle_{K}$
$\displaystyle=\sup_{||f||_{K}=1,x\in\mathcal{X}}\langle
k(x,\cdot),T^{*}f\rangle_{K}$
$\displaystyle=\sup_{||f||_{K}=1,x\in\mathcal{X}}(T^{*}f)(x)$
$\displaystyle=\sup_{||f||_{K}=1}||T^{*}f||_{\infty}$
$\displaystyle=||T^{*}||_{\mathcal{H}\to L^{\infty}}$
Hence, since $(C_{\nu}+\lambda)^{-\frac{1}{2}}$ is self-adjoint, we have by
the interpolation inequality (Lemma 1), that:
$||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||^{2}_{\infty}\leq\frac{||k^{\psi}||^{2}_{\infty}||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||^{2}_{2}}{\psi\Big{(}\frac{||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||^{2}_{2}}{||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||^{2}_{K}}\Big{)}}$
(38)
Taking the supremum over $f\in\mathcal{H}_{K}$, and noting that the RHS of
(38) is jointly increasing in $||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||_{2}$ and
$||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||_{K}$ by the fact that
$\frac{t}{\psi(t)}$ is concave and nondecreasing in Assumption 2, we obtain:
$||(C_{\nu}+\lambda)^{-\frac{1}{2}}||^{2}_{\mathcal{H}\to
L^{\infty}}\leq\frac{||k^{\psi}||^{2}_{\infty}}{\psi(\lambda)}$
where we have used the fact that
$\sup_{f\in\mathcal{H}_{K}}||(C_{\nu}+\lambda)^{-\frac{1}{2}}f||^{2}_{2}=\sup_{f\in\mathcal{H}_{K}}||C^{\frac{1}{2}}_{\nu}(C_{\nu}+\lambda)^{-\frac{1}{2}}f||^{2}_{K}$.
∎
###### Lemma 3.
$\mathcal{N}(\lambda)\equiv\text{tr}(C_{\nu}(C_{\nu}+\lambda)^{-1})\preceq
s^{-1}(\psi(\lambda)^{-1})$
###### Proof.
From Corollary 2.1, we obtain $\mu_{i}\asymp\psi^{-1}(s(i)^{-1})$. Therefore,
there we have:
$\displaystyle\mathcal{N}(\lambda)$
$\displaystyle\equiv\text{tr}(C_{\nu}(C_{\nu}+\lambda)^{-1})$
$\displaystyle=\sum_{i=1}^{\infty}\frac{\mu_{i}}{\mu_{i}+\lambda}$
$\displaystyle=\sum_{i=1}^{\infty}\frac{1}{1+\lambda\mu^{-1}_{i}}$
$\displaystyle\leq\sum_{i=1}^{\infty}\frac{1}{1+\frac{C_{1}\lambda}{\psi^{-1}(s(i)^{-1})}}$
$\displaystyle\leq\int_{0}^{\infty}\frac{1}{1+\frac{C_{1}\lambda}{\psi^{-1}(s(t)^{-1})}}dt$
$\displaystyle=\int_{0}^{s^{-1}(\psi(\lambda)^{-1})}\frac{1}{1+\frac{C_{1}\lambda}{\psi^{-1}(s(t)^{-1})}}dt+\int_{s^{-1}(\psi(\lambda)^{-1})}^{\infty}\frac{1}{1+\frac{C_{1}\lambda}{\psi^{-1}(s(t)^{-1})}}dt$
We observe that, by Assumption 2 we have that:
$2\psi^{-1}\Big{(}\frac{x}{D^{\psi}_{2}}\Big{)}\leq\psi^{-1}(x)\leq
2\psi^{-1}\Big{(}\frac{x}{D^{\psi}_{1}}\Big{)}$
and by Assumption 4:
$\Big{(}D^{s}_{2}s\Big{(}\frac{x}{2}\Big{)}\Big{)}^{-1}\leq
s(x)^{-1}\leq\Big{(}D^{s}_{1}s\Big{(}\frac{x}{2}\Big{)}\Big{)}^{-1}$
Putting these together, we have:
$2\psi^{-1}\Big{(}s(D^{\psi,s}_{1}x)^{-1}\Big{)}\leq\psi^{-1}(s(x)^{-1})\leq
2\psi^{-1}\Big{(}s(D^{\psi,s}_{2}x)^{-1}\Big{)}$
where $D^{\psi,s}_{1}=2^{\log_{D^{s}_{1}}D^{\psi}_{2}}$ and
$D^{\psi,s}_{2}=2^{\log_{D^{s}_{2}}D^{\psi}_{1}}$. Hence, for
$t>s^{-1}(\psi(\lambda)^{-1})$, we have:
$\frac{\lambda}{\psi^{-1}(s(t)^{-1})}=\frac{\psi^{-1}(s(s^{-1}(\psi(\lambda)^{-1}))^{-1})}{\psi^{-1}(s(t)^{-1})}=\frac{\psi^{-1}(s(t^{-1}s^{-1}(\psi(\lambda)^{-1})t)^{-1})}{\psi^{-1}(s(t)^{-1})}\geq
2^{-\log_{D^{\psi,s}_{2}}t^{-1}s^{-1}(\psi(\lambda)^{-1})}$
Observing that
$2^{-\log_{D^{\psi,s}_{2}}t^{-1}s^{-1}(\psi(\lambda)^{-1})}=\Big{(}\frac{t}{s^{-1}(\psi(\lambda)^{-1})}\Big{)}^{r}$,
where $r\equiv\log_{D^{\psi,s}_{2}}(2)$. Hence, we have:
$\displaystyle\mathcal{N}(\lambda)$
$\displaystyle\leq\int_{0}^{s^{-1}(\psi(\lambda)^{-1})}\frac{dt}{1+\frac{C_{1}\lambda}{\psi^{-1}(s(t)^{-1})}}+\int_{s^{-1}(\psi(\lambda)^{-1})}^{\infty}\frac{dt}{1+\frac{C_{1}\lambda}{\psi^{-1}(s(t)^{-1})}}$
$\displaystyle\leq
s^{-1}(\psi(\lambda)^{-1})+\int_{s^{-1}(\psi(\lambda)^{-1})}^{\infty}\frac{dt}{1+C_{1}\Big{(}\frac{t}{s^{-1}(\psi(\lambda)^{-1})}\Big{)}^{r}}$
$\displaystyle\leq
s^{-1}(\psi(\lambda)^{-1})+s^{-1}(\psi(\lambda)^{-1})\int_{1}^{\infty}\frac{ds}{1+C_{1}s^{r}}$
$\displaystyle\leq Cs^{-1}(\psi(\lambda)^{-1})$
for some $C>0$, since the second integral converges as
$r=\log_{D^{\psi,s}_{2}}(2)>1$ (as we can always choose
$D^{s}_{2}>D^{\psi}_{1}$). ∎
|
[table]capposition=top
# A Model for Attribute Based Role-Role Assignment (ARRA)
aJiwan L. Ninglekhu and bRam Krishnan Department of Electrical and Computer
Engineering
The University of Texas at San Antonio
San Antonio, Texas 78249
Email<EMAIL_ADDRESS>and<EMAIL_ADDRESS>
###### Abstract
Administrative Role Based Access Control (ARBAC) models specify how to manage
user-role assignments (URA), permission-role assignments (PRA), and role-role
assignments (RRA). Many approaches have been proposed in the literature for
URA, PRA and RRA. In this paper, we propose a model for attribute-based role-
role assignment (ARRA), a novel way to unify prior RRA approaches. We leverage
the idea that attributes of various RBAC entities such as admin users and
regular roles can be used to administer RRA in a highly-flexible manner. We
demonstrate that ARRA can express and unify prior RRA models.
###### Index Terms:
Roles, Role Hierarchy, RBAC, ARBAC, Access Control, Administration.
## I Introduction
Role-based access control (RBAC) [1, 2] model is well studied in academia[3]
and well adopted in the industry[4]. In RBAC, assignment of users and
permissions to roles allows users to execute permissions by activating roles.
Administrative RBAC (ARBAC), which involves in user-role assignment and
revocation (URA), permission-role assignment and revocation (PRA), and role-
role assignment and revocation (RRA), is a challenging task [5]. ARBAC has
been well explored [5, 6, 7, 8, 9]. All of these works explore URA and PRA.
However, only few cover RRA [5, 8]. Each of these models use few fixed set of
properties of RBAC in making assignment decisions. For example, In RRA97 [5],
an admin role is given admin authority over set of roles, called role range,
where admin role can perform role assignment.
Recently, Attribute-Based Access Control (ABAC) has gained popularity because
of its flexibility [10, 11, 12, 13, 14]. ABAC has also proven its ability to
represent different access control models [12]. However, it has rarely been
used in administration of RBAC.
Among three main aspects of ARBAC, we have explored models for attribute-based
user-role assignment (AURA) and attribute-based permission-role assignment
(ARPA) models [15].
In this paper we present our model for attribute-based role-role assignment
(ARRA), a follow-up work on AARBAC [15]. ARRA is driven by our objective to
design attribute-based model that allows us to express features from previous
RRA models and more. For example, it allows us to express features as a
combination of two prior models and in addition, allows us to add new
features. Thus, this work is motivated largely by two critical factors: (a) an
objective to build a coherent model, which can be configured to express prior
models and more (b) build a unified model that can be analyzed for various
desirable security features, which can provide a single codebase to express
prior models and beyond. The contributions of this paper are as follows:
* •
We develop an attribute-based administrative model for role-role assignment
(ARRA).
* •
We demonstrate that ARRA is capable of expressing prior approaches to RRA such
as RRA97 [5] and UARBAC’s RRA [8].
The remainder of this paper is organized as follows. In Section II, we discuss
related work. In Section III, ARRA model is presented. Section IV presents
algorithms that translate prior RRA instances into equivalent ARRA instances.
Section V concludes the paper.
## II Related work
Among many prominent works done for ARBAC [5, 6, 7, 9, 8], ARBAC97 [5] and
UARBAC [8] present role-role assignment (RRA) as part of their model.
In RRA97, a set of roles in the hierarchy called authority range is given to
an admin role for decentralized administration. A user with admin role can
perform operations like inserting an edge or creating a role, on roles in her
authority range. To confirm autonomy RRA97 requires role range to be
encapsulated [5].
UARBAC [8] redefines RBAC model with class objects. It treats each entity in
the RBAC system such as files, roles and users as objects. UARBAC’s RRA deals
with assigning/revoking role-role on the basis of admin user’s authority over
individual roles specified by access modes. Class permissions in UARBAC allows
an admin user to conduct an operation over every object of a class.
In both of these models, the policy for assigning a role to a role is based on
a fixed set of properties of the entities such as admin role, admin user and
regular roles, involved in assigning/revoking roles.
Crampton et. al [16] present models for RRA with administrative scopes. Admin
scopes are plausible approach for role hierarchy operations in RBAC. However,
admin scopes may not be intuitive to express as a role attribute. Many
literature [10, 17, 18] present benefits of integrating attributes into an
RBAC operational model. An operational model deals with making decisions on
user’s access request for objects. There are many works on operational aspects
of ABAC [19, 12, 11, 20]. In contrast, ARRA is an administrative model that
uses attributes of RBAC entities for assigning/revoking roles.
Table I: ARRA Model
o 1 | X[l] | – AU, AOP, ROLES, AR are finite sets of administrative users, administrative operations such as assign and revoke, regular roles and administrative roles, respectively.
---
– AUA $\subseteq$ AU $\times$ AR, administrative user to administrative role
assignment relation.
– RH $\subseteq$ ROLES $\times$ ROLES, a partial ordering on the set ROLES.
– ARH $\subseteq$ AR $\times$ AR, a partial ordering on the set AR.
– AATT, ARATT, and RATT are finite sets of administrative user attribute
functions, administrative role attribute functions, and regular role attribute
functions, respectively.
– For each att in AATT $\cup$ ARATT $\cup$ RATT, Scope(att) is a finite set of
atomic values from which the range
of the attribute function att is derived. – attType : AATT $\cup$ ARATT $\cup$
RATT $\rightarrow$ {set, atomic}, which specifies whether the range of a given
attribute
is atomic or set valued.
– Each attribute function maps elements in AU, AR and ROLES to atomic or set
values. $\forall aatt\in\mbox{AATT}.\
aatt:\mbox{AU}\rightarrow\left\\{\begin{array}[]{ll}\mbox{Scope}(aatt)\>\mbox{if}\>\mbox{attType}(aatt)=\mbox{atomic}\\\
2\textsuperscript{Scope({aatt})}\>\mbox{if}\>\mbox{attType}(aatt)=\mbox{set}\\\
\end{array}\right.$ $\forall aratt\in\mbox{ARATT}.\
aratt:\mbox{AR}\rightarrow\left\\{\begin{array}[]{ll}\mbox{Scope}(aratt)\>\mbox{if}\>\mbox{attType}(aratt)=\mbox{atomic}\\\
2\textsuperscript{Scope({aratt})}\>\mbox{if}\>\mbox{attType}(aratt)=\mbox{set}\\\
\end{array}\right.$ $\forall ratt\in\mbox{RATT}.\
ratt:\mbox{ROLES}\rightarrow\left\\{\begin{array}[]{ll}\mbox{Scope}(ratt)\>\mbox{if}\>\mbox{attType}(ratt)=\mbox{atomic}\\\
2\textsuperscript{Scope({ratt})}\>\mbox{if}\>\mbox{attType}(ratt)=\mbox{set}\\\
\end{array}\right.$ – is_ordered : AATT $\cup$ ARATT $\cup$ RATT $\rightarrow$
{True, False}, specifies if the scope is ordered for each of the
attributes.
– For each att $\in$ AATT $\cup$ ARATT $\cup$ RATT, if is_ordered(att) = True,
Hatt $\subseteq$ Scope(att) $\times$ Scope(att), a partially ordered attribute
hierarchy, and Hatt $\neq$ $\phi$, else, if is_ordered(att) = False, Hatt =
$\phi$
(For some att $\in$ AATT $\cup$ ARATT $\cup$ RATT for which attType(att) = set
and is_ordered(att) = True, if $\\{a,b\\}$, $\\{c,d\\}$ $\in$ 2Scope(att)
(where $a,b,c,d\in$ Scope($att$)), we infer $\\{a,b\\}$ $\geq$ $\\{c,d\\}$ if
$(a,c)$, $(a,d)$, $(b,c)$, $(b,d)$ $\in$ H${}^{\mbox{*}}_{\mbox{{att}}}$.)
ARRA model allows an administrator to perform an operation on a single role or
a set of roles at a time. The authorization rule for performing an operation
on a single role is as follows:
For each op in AOP, is_authorizedRop(au: AU, r1 : ROLES, r2 : ROLES) specifies
if the admin user au is allowed to perform the operation op (e.g. assign,
revoke, etc.) between the regular roles r1 and the role r2. Assigning a role
r1 to r2 makes r1 junior to r2. This rule is written as a logical expression
using attributes of admin user au, admin role, ar, and regular role, r.
The authorization rule for performing an operation on a set of users is as
follows:
For each op in AOP, is_authorizedRop(au: AU, $\chi$ : 2ROLES, r : ROLES)
specifies if the admin user au is allowed to perform the operation op (e.g.
assign, revoke, etc.) between the roles in the set $\chi$ and the role r. Here
$\chi$ is a set of roles that can be specified using a set-builder notation,
whose rule is written using role attributes.
## III ARRA Model
In this section, we present our approach for attribute-based role-role
assignment (ARRA). Inspired by prior RRA models we have included attributes
for admin users, admin roles and regular roles. Based on these attributes role
assignment and revocation decisions on are made.
Table II presents formal ARRA model. The entities involved in ARRA comprise of
admin users (AU), regular roles (ROLES) and their hierarchy (RH), admin roles
(AR) and their hierarchy (ARH), admin user to admin role relation (AUA) and
admin operations (AOP).
In ARRA, admin user in AU wants to perform admin operation such as assign or
revoke from AOP using attributes of entities in the model. We have developed
admin users attribute functions (AATT) and admin roles attribute functions
(ARATT). We have also introduced regular roles attribute functions (RATT)
based on the need we have observed. Attributes from other RBAC entities can
also be developed if needed. We will later see that we need attributes from
different entities in representing properties of RRA97 and UARBAC’s RRA in
ARRA.
The attribute functions or _simply_ attributes are defined as mapping from its
domain such as AU or AR to its range. Range of an attribute att can be atomic
or set valued, which is specified by function attType. It is derived from set
of scope or atomic values denoted by Scope(att). Furthermore, the scope of an
attribute can be either ordered or unordered, which is given by a function
called is_ordered(att). If it is ordered, we must specify that attribute’s
hierarchy denoted by Hatt on its scope Scope(att). Hatt is a partial ordering
on Scope(att). Note that even in the case of a set valued attribute att, the
hierarchy Hatt is specified on Scope(att) instead of 2Scope(att). We infer the
ordering between two set values given an ordering on atomic values.
H${}^{\mbox{*}}_{\mbox{{att}}}$ in Table II denotes the reflexive transitive
closure of Hatt.
In ARRA, there are two ways by which an admin user can select a set of regular roles for assignment to a target regular role. The first way allows an admin user to select a single role and a target role, and perform an admin operation like assign or revoke. The second way allows an admin user to select a set of regular roles, the target role and perform similar operation on those roles. In the latter case, the selection criteria for the set of regular roles can be expressed using a set-builder notation whose rule is based on the regular role attributes. For example, is_authorizedRassign(au, {r1 | r1 $\in$ ROLES $\wedge$ Lead $\in$ roleTitle(r1)}, r2) would specify a policy for an admin user au that selects the set of all the roles with role title Lead in order to assign those roles to a role r2. Assigning a role r1 to role r2 makes role r1 junior to r2. In other words, it adds an entry <r2, r1> in RH. It is also referred to as edge insetion.
Authorization rule is specified as a logical expression on the attributes of
admin users, admin roles, and that of regular roles considered for assignment.
## IV Mapping Prior RRA Models in ARRA
In this section, we demonstrate that ARRA can intuitively simulate the
features of prior RRA models. In particular, we have developed concrete
algorithms that can convert any instance of RRA97 and UARBAC’s RRA into an
equivalent ARRA instance. For each of these models, we first present an
example instance followed by that instance’s corresponding ARRA equivalent.
Then finally, we present algorithms for each to show a programmable mapping
procedure.
Figure 1: Role and Administrative Role Hierarchies in RRA97
### IV-A Mapping RRA97 to AARA
In this section, we present RRA97 example instance and manually convert it
into its equivalent ARRA instance. We then present a mapping algorithm for the
same, which demostrates a programmable procedure for translation of any RRA97
instance to its equivalent ARRA instance.
#### IV-A1 An RRA97 Example Instance
RRA97 [5] assumes an existing regular role hierarchy and an administrative
role hierarchy as depicted in Figure 1. An example instance for RRA97 model is
presented as follows: Sets and Functions:
* •
USERS = {u1, u2, u3, u4}
* •
ROLES = {ED, E1, PE1, QE1, PL1, E2, PE2,
* QE2, PL2, DIR}
* •
AR = {DSO, PSO1}
* •
AUA = {(u3, DSO), (u4, PSO1)}
* •
RH = {<ED, E1>, <E1, QE1>, <E1, PE1>,
* <PE1, PL1>, <QE1, PL1>, <ED, E2>,
* <E2, PE2>, <E2, QE2>, <PE2, PL2>,
* <QE2, PL2>, <PL1, DIR>, <PL2, DIR>}
* •
ARH = {<SSO, DSO>, <DSO, PSO1>,
<DSO, PSO2>} • can-modify = {(DSO, (ED, DIR)), (PSO1, (E1, PL1)), (PSO1, (E2,
PL2))} Where, (ED, DIR) = {E1, E2, PE1, PE2, QE1, QE2, PL1, PL2}, (E1, PL1)
= {PE1, QE1} and, (E2, PL2) = {PE2, QE2} USERS, ROLES, and AR are sets of
users, regular roles and admin roles, respectively. User u3 is given admin
role DSO and PSO1 is given to role u4. RH and ARH represent regular role and
admin role hierarchies as depicted in Figure 1. We have taken original can-
modify example from RRA97 paper [5] where each admin role is mapped with
authority ranges. The can-modify set in our instance depicts the same scenario
with three elements: (DSO, (ED, DIR)), (PSO1, (E1, PL1)) and (PSO1, (E2,
PL2)). It shows that role DSO is given an authority range (ED, DIR) while role
PSO1 is given two authority ranges (E1, PL1) and (E2, PL2). In RRA97, an
authority range, (x, y) = {r : ROLES $|$ x < r < y}, where x and y are end
points of the range. For every (ar, (x, y)) in can-modify, the range denoted
by (x, y) must be an encapsulated authority range. A range (x, y) is said to
be encapsulated if, $\forall$r1 $\in$ (x, y) $\wedge$ $\forall$r2 $\notin$ (x,
y). [r2 > r1 $\leftrightarrow$ r2 > y $\wedge$ r2 < r1 $\leftrightarrow$ r2 <
x] A new edge AB can be inserted between two incomparable roles A and B under
following two conditions: • ARimmediate(A) = ARimmediate(B). An immediate
authority range of a role r, ARimmediate(r) is the authority range (x,y) such
that r $\in$ (x, y) and for all authority ranges (x', y') junior to (x, y) we
have r $\notin$ (x', y'). or, • if (x, y) is an authority range such that (A =
y $\wedge$ B > x) $\vee$ (B = x $\wedge$ A < y) then insertion of AB must
preserve encapsulation of (x, y).
#### IV-A2 ARRA Instance Equivalent to RRA97 Instance
In this section, we show a manual translation of RRA97 example instance
presented in previous segment into an equivalent ARRA instance. Equivalent
ARRA instance is as follows:
Map RRA97 sets to ARRA sets
* •
AU = {u1, u2, u3, u4}
* •
AOP = {insertEdge, deleteEdge}
* •
ROLES = {ED, DIR, E1, PE1, QE1, PL1, E2,
* PE2, QE2, PL2, DIR}
* •
AUA = {(u3, DSO), (u4, PSO1)}
* •
RH = {<ED, E1>, <E1, QE1>, <E1, PE1>,
* <PE1, PL1>, <QE1, PL1>, <ED, E2>,
* <E2, PE2>, <E2, QE2>, <PE2, PL2>,
* <QE2, PL2>, <PL1, DIR>, <PL2, DIR>}
Define attributes
* •
AATT = {}, RATT = {}
* •
ARATT= {authRange}
* •
Scope(authRange) = RH+,
* attType(authRange) = set,
* is_ordered(authRange) = False, HauthRange = $\phi$
* •
authRange(DSO) = {ED, DIR},
* authRange(PSO1) = {(E1, PL1), (E2, PL2)}
Construct authorization rule for insertEdge To assign role r1 to role r2 (or
to insert an edge r1r2 and hence add <r2, r1> to RH), –
is_authorizedRinsertEdge(au : AU, r1 : ROLES, r2 : ROLES) $\equiv$
($\exists$(au, ar1) $\in$ AUA, $\exists$(s, t) $\in$ authRange(ar1). r1, r2
$\in$ $\lceil$s, t$\rceil$) $\wedge$ (($\exists$(m, n), (m', n') $\in$
$\bigcup\limits_{ar\textsubscript{2}\in\ \textrm{AR}}$ authRange(ar2). r1, r2
$\lceil$m, n$\rceil$ $\wedge$ ($\lceil$m', n'$\rceil$ $\subset$ $\lceil$m,
n$\rceil$ $\rightarrow$ r1, r2 $\notin$ (m', n'))) $\vee$ ($\exists$(x,y)
$\in$ $\bigcup\limits_{ar\textsubscript{3}\in\ \textrm{AR}}$ authRange(ar3).
((r1 = y $\wedge$ r2 > x) $\vee$ (r2 = x $\wedge$ r1 < y)) $\wedge$
($\forall$p $\in$ $\lceil$x, y$\rceil$ $\wedge$ $\forall$q $\notin$ $\lceil$x,
y$\rceil$. (<q, p> $\in$ (RH $\cup$ <r2, r1>)* $\leftrightarrow$ <q, y> $\in$
(RH $\cup$ <r2, r1>)*) $\wedge$ (<p, q> $\in$ (RH $\cup$ <r2, r1>)*
$\leftrightarrow$ <x, q> $\in$ (RH $\cup$ <r2, r1>)*))))
Construct authorization rule for deleteEdge – is_authorizedRdeleteEdge(au :
AU, r1 : ROLES, r2 : ROLES) $\equiv$ $\exists$(au, ar) $\in$ AUA, $\exists$(x,
y) $\in$ authRange(ar). r1, r2 $\in$ $\lceil$x, y$\rceil$ $\wedge$
$\forall$(u, v) $\in$ $\bigcup\limits_{ar\textsubscript{4}\in\ \textrm{AR}}$
authRange(ar4), $\exists$<r2, r1> $\in$ RH. (r1 $\neq$ u $\wedge$ r2 $\neq$ v)
As we can observe, the process of translation involves four main steps. First,
we map given sets from RRA97 to equivalent sets in ARRA. Second, required
attribute functions are defined. We have designed admin role attribute
authRange that maps authority ranges given to admin roles. Attribute authRange
is set valued and unordered. Scope of authRange is a transitive clousure of
regular role hierarchy, RH+. That is, if <a, b>, <b, c>, <c, d> $\in$ RH+,
then <a, d> is also true.
We know that authority range represents a set of roles over which an admin
role have authority to conduct operations such as assign and revoke. In ARRA,
we use symbolic representation of authority range with end points a and b as
(a, b). However, whenever we need to work with roles present in the set
represented by authority range (a, b), we denote the set of roles with
$\lceil$a, b$\rceil$. That is, $\lceil$a, b$\rceil$ = {r $|$ r $\in$ ROLES
$\wedge$ a < r < b}.
Next, we construct authorization rule for the admin operation insertEdge.
There are fundamentally three requirements that an admin user must consider to
assign role r1 to role r2. Firstly, both the roles r1 and r2 must be within
the authority range of an admin role assigned to a given admin user. Secondly,
both the roles taken for edge insertion must have same immediate authority
range, or (thirdly) if the edge is to be inserted between a role on the top
end point of an authority range and a role senior to the lower end point of
authority range, or if the edge is to be inserted between a role on the bottom
end of an authority range and a role junior to the top end of the authority
range, then that edge insertion must preserve encapsulation of an authority
range with those (top and bottom) end points. Finally, we construct an
authorization formula for operation deleteEdge. An admin user with an admin
role can delete an edge between any two roles in an authority range except
that any edge between the end points of an authority range can not be deleted.
Algorithm 1 MapRRA97
0: Input: RRA97 instance
0: Output: ARRA instance
0: /* Map basic sets and functions in AURA */
0: a. AU $\leftarrow$ USERS97
0: b. AOPA $\leftarrow$ {insertEdge, deleteEdge}
0: c. ROLESA $\leftarrow$ ROLES97 ; AUAA $\leftarrow$ AUAU
0: b. RHA $\leftarrow$ RH97
0: /* Map attribute functions in AURA */
0: a. AATTA $\leftarrow$ {} ; RATTA $\leftarrow$ {}
0: b. ARATTA $\leftarrow$ {authRange}
0: c. ScopeA(authRange) = RH+A
0: d. attTypeA(authRange) = set
0: e. is_orderedA(authRange) = False ; H${}^{\mbox{{A}}}_{\mbox{{authRange}}}$
= $\phi$
0: f. For each ar $\in$ ARA, authRange(ar) = $\phi$
0: g. For each (ar, (ri, rj)) $\in$ can-modify97,
0: authRange(ar) = authRange(ar) $\cup$ (ri, rj)
0: /* Construct assign rule in AURA */
0: a. assign_formula = $\phi$
0: b. For each (ar, (ri, rj)) $\in$ can-modify97,
0: assign_formula' = assign_formula $\vee$ ($\exists$(au, ar1) $\in$ AUA,
$\exists$(s, t) $\in$ authRange(ar1). r1, r2 $\in$ $\lceil$s, t$\rceil$)
$\wedge$ (($\exists$(m, n), (m', n') $\in$
$\bigcup\limits_{ar\textsubscript{2}\in\ \textrm{AR}}$ authRange(ar2). r1, r2
$\lceil$m, n$\rceil$ $\wedge$ ($\lceil$m', n'$\rceil$ $\subset$ $\lceil$m,
n$\rceil$ $\rightarrow$ r1, r2 $\notin$ (m', n'))) $\vee$ ($\exists$(x,y)
$\in$ $\bigcup\limits_{ar\textsubscript{3}\in\ \textrm{AR}}$ authRange(ar3).
((r1 = y $\wedge$ r2 > x) $\vee$ (r2 = x $\wedge$ r1 < y)) $\wedge$
($\forall$p $\in$ $\lceil$x, y$\rceil$ $\wedge$ $\forall$q $\notin$ $\lceil$x,
y$\rceil$. (<q, p> $\in$ (RH $\cup$ <r2, r1>)* $\leftrightarrow$ <q, y> $\in$
(RH $\cup$ <r2, r1>)*) $\wedge$ (<p, q> $\in$ (RH $\cup$ <r2, r1>)*
$\leftrightarrow$ <x, q> $\in$ (RH $\cup$ <r2, r1>)*))))
0: c. auth_assign = is_authorizedRinsertEdge(au : AUA,
0: r1 : ROLESA, r2 : ROLESA) $\equiv$ assign_formula'
0: /* Construct revoke rule for AURA */
0: a. revoke_formula = $\phi$
0: b. For each (ar1, (r1, r2)) $\in$ can-modify97,
0: revoke_formula' = revoke_formula $\vee$
0: $\exists$(au, ar) $\in$ AUA, $\exists$(x, y) $\in$ authRange(ar). r1, r2
$\in$ $\lceil$x, y$\rceil$ $\wedge$ $\forall$(u, v) $\in$
$\bigcup\limits_{ar\textsubscript{4}\in\ \textrm{AR}}$ authRange(ar4),
$\exists$<r2, r1> $\in$ RH. (r1 $\neq$ u $\wedge$ r2 $\neq$ v)
0: c. auth_revoke = is_authorizedRdeleteEdge(au : AUA,
0: r1 : ROLESA, r2 : ROLESA) $\equiv$ revoke_formula'
### IV-B MapRRA97
Algorithm 1 presents MapRRA97, a translation algorithm for any RRA97 instance
into an equivalent ARRA instance. Superscript labels 97 and A in the algorithm
represent sets and functions from RRA97 and ARRA, respectively. MapRRA97 takes
an instance of RRA97 as input. In particular, input consists of USERS97,
ROLES97, AR97, AUA97, RH97, ARH97, and can-modify97. The can-modify
instruction covers operations for inserting an edge, deleting an edge,
creating a role and deleting a role. ARRA can simulate inserting edge and
deleting an edge. However, creating and deleting roles are beyond the scope of
current ARRA model.
Output from MapRRA97 algorithm is an equivalent ARRA instance, with following
sets and functions. AUA, AOPA, ROLESA, AUAA, RHA, ARHA, AATTA, ARATTA, For
each attribute att $\in$ ARATTA, ScopeA(att), attTypeA(att), is_orderedA(att)
and HAatt, For each admin role ar $\in$ ARA and for each att $\in$ ARATTA,
att(ar), Authorization rule for assign (auth_assign), and Authorization rule
for revoke (auth_revoke).
Steps indicated in MapRRA97 correspond to the steps described in equivalent
ARRA instance for RRA97 in section IV-A2. We explain it here briefly. Step 1
maps sets from RRA97 to ARRA sets and functions. In Step 2, admin user
attributes and admin role attributes are expressed. AATT and RATT are left
empty as there is no use case for these attributes in translating RRA97. Admin
role attribute authRange captures the mapping between an authority range as
defined in RRA97, and an admin role. For each (ar, (ri, rj)) in set can-
modifySS97, authRange(ar) populates set of authority ranges for ar.
In Step 3, we construct assign_formula rule for ARRA that is equivalent to
inserting an edge in can_modify97. For each (ar, (ri, rj)) in set can-
modifySS97, logical expression that checks the assignment conditions is
constructed. Equivalent translation in ARRA is given by
is_authorizedRinsertEdge(au : AUA, rß1 : ROLESA, rß2 : ROLESA). This formula
must checks if roles rß1 and rß2 belong in an authority range of an admin
role, which is mapped to admin user au. It further checks if rß1 and rß2 have
same immediate authority range, or checks if any other assignment action
doesn’t violate encapsulation of admin role’s authority range. Similarly, In
Step 4, revoke_formula equivalent to deleting an edge from can_modify97 is
expressed. It checks if admin role mapped to admin role au has authority range
where rß1 and rß2 belong and restricts from deleting an edge between end
points of any authority range.
### IV-C Mapping UARBAC’s RRA to ARRA
In this section, we present an example instance for UARBAC’s RRA and its
equivalent ARRA instance. We then present an algorithm that outlines the
translation procedure.
#### IV-C1 An Example UARBAC’s RRA Instance
RBAC schema
* •
C = {role}
* •
OBJS(user) = USERS, OBJS(role) = ROLES
* •
AM(role) = {grant, empower, admin}
RBAC state
* •
USERS = OBJ(user) = {u1, u2, u3, u4}
* •
ROLES = OBJ(role)= {r1, r2, r3}
* •
P = [role, r1, grant], [role, r1, empower], [role, r1, admin], [role, r2,
grant], [role, r2, empower], [role, r2, admin], [role, r3, grant], [role, r3,
empower], [role, r3, admin], [role, grant], [role, empower], [role, admin]}
* •
RH = {<r2, r3>}
Administrative permissions of UARBAC’s RRA Following is the list of
administrative permissions each user has for role-role assignment:
* •
authorized_perms[u1] = {[role, r1, grant],
* [role, r2, admin], [role, r2, empower],
* [role, r3, admin]}
* •
authorized_perms[u2] = {[role, r1, admin],
* [role, r1, empower], [role, r2, empower],
* [role, r3, grant]}
* •
authorized_perms[u3] = {[role, admin]}
* •
authorized_perms[u4] = {[role, grant],
* [role, empower], [role, admin]}
Role-role assignment condition One can perform following operation to assign a
role r1 to another role r2.
* •
grantRoleToRole(r1, r2)
To perform this operation one needs one of the following two permissions:
* •
[role, r1, grant] and [role, r2, empower] or,
* •
[role, grant] and [role, r2, empower] or,
* •
[role, r1, grant] and [role, empower] or,
* •
[role, grant] and [role, empower]
Condition for revoking a role from another role To revoke a role r2 from a
role r3, admin user performs following operation.
* •
revokeRoleFromUser(r2, r3)
To conduct this operation one needs one of the following options:
* •
[role, r2, grant] and [role, r3, empower]
* •
[role, r2, admin]
* •
[role, r3, admin]
* •
[role, admin]
### IV-D ARRA instance equivalent to UARBAC’s RRA instance
This segment presents equivalent instance for UARBAC’s RRA example instance
presented in previous segment. Map UARBAC’s RRA sets to ARRA sets
* •
AU = {u1, u2, u3, u4}
* •
AOP = {assign, revoke}
* •
ROLES = {r1, r2, r3},
* •
AUA = {}
* •
RH = {<r2, r3>}
Define Attributes
* •
AATT = {grantAuth, empowerAuth, adminAuth, roleClassAuth}, ARATT = {}, RATT =
{}
* •
Scope(grantAuth) = {r1, r2, r3}
* attType(grantAuth) = set,
* is_ordered(grantAuth) = ROLES, HgrantAuth = RH
* •
grantAuth(u1) = {r1}, grantAuth(u2) = {r3},
* grantAuth(u3) = {}, grantAuth(u4) = {}
* •
Scope(empowerAuth) = {r1, r2, r3}
* attType(empowerAuth) = set,
* is_ordered(empowerAuth) = ROLES
* HempowerAuth = RH
* •
empowerAuth(u1) = {r2},
* empowerAuth(u2) = {r1, r2},
* empowerAuth(u3) = {}, empowerAuth(u4) = {}
* •
Scope(adminAuth) = {r1, r2, r3}
* attType(adminAuth) = set,
* is_ordered(adminAuth) = ROLES,
* HadminAuth = RH
* •
adminAuth(u1) = {r2, r3},
* adminAuth(u2) = {r1},
* adminAuth(u3) = {}, adminAuth(u4) = {}
* •
Scope(roleClassAuth) = AM(role)
* attType(roleClassAuth) = set,
* is_ordered(roleClassAuth) = False,
* HroleClassAuth = $\phi$
* •
roleClassAuth(u1) = {}, roleClassAuth(u2) = {},
* roleClassAuth(u3) = {admin},
* roleClassAuth(u4) = {grant, empower, admin}
Construct authorization rule for role-role assignment
For each op in AOP, authorization rule to assign/revoke role-role can be
expressed as follows: To assign any regular role r1 $\in$ ROLES to regular
role r2 $\in$ ROLES, – is_authorizedUassign(au : AU, r1 : ROLES, r2 : ROLES)
$\equiv$ (r1 $\in$ grantAuth(au) $\wedge$ r2 $\in$ empowerAuth(au)) $\vee$ (r1
$\in$ grantAuth(au) $\wedge$ empower $\in$ roleClassAuth(au)) $\vee$ (grant
$\in$ roleClassAuth(au) $\wedge$ r2 $\in$ empowerAuth(au)) $\vee$ (grant $\in$
roleClassAuth(au) $\wedge$ empower $\in$ roleClassAuth(au))
Construct authorization rule for revoking role-role
To revoke any regular role r1 $\in$ ROLES from another regular role r2 $\in$
ROLES, – is_authorizedUrevoke(au1 : AU, r1 : ROLES, r2 : ROLES) $\equiv$ (r1
$\in$ grantAuth(au) $\wedge$ r2 $\in$ empowerAuth(au)) $\vee$ r1 $\in$
adminAuth(au) $\vee$ r2 $\in$ adminAuth(au) $\vee$ admin $\in$
roleClassAuth(au)
There are four main stages in this translation process. First, we map main
sets from UARBAC’s RRA to ARRA model. Set of admin users consists of set of
users from UARBAC’s RRA, namely u1, u2, u3, and u4. There are two admin
operations, assign and revoke. Sets ROLES and RH map set of roles and role
hierarchy from UARBAC’s RRA, respectively. AUA is left empty as there is no
notion of admin roles in UARBAC.
Next, we define required attribute functions. We used admin user attributes to
express ARRA for UARBAC’s RRA. Among four admin user attributes, grantAuth,
empowerAuth and adminAuth are object level attributes. Each of these
attributes express the nature of authority an admin user has over different
roles. For example, empowerAuth over a role rx, allows admin user to assign
any object such as a user and/or another role to role rx. Class level
attribtue roleClassAuth captures the nature of access mode an admin user has
towards role class. The scope of this attribute is AM(role), which has three
different types of access modes, namely grant, empower and admin. For example,
an admin user with roleClassAuth of grant allows that admin user to grant
(assign) any role to other object such as a user and/or role. Translation of
UARBAC’s RRA doesn’t require admin role and regular role attributes, hence
left empty.
After attributes have been defined we construct the assignment formula
(assign_formula), using usual logical expression. As mentioned in the UARBAC’s
RRA instance, there are four different combination of permissions an admin
user can use to make role-role assignment. An equivalent logical expression is
captured as is_authorizedUassign(au : AU, r1 : ROLES, r2 : ROLES). Finally, we
construct a logical expression which captures four different permission
options that an admin user can use for revoking a role from another role. The
logical expression is equivalent to is_authorizedUrevoke(au1 : AU, r1 : ROLES,
r2 : ROLES).
Algorithm 2 MapRRA-UARBAC
0: Input: Instance of RRA in UARBAC
0: Output: ARRA instance
0: /* Map basic sets and functions in AURA */
0: a. AUA $\leftarrow$ USERSU ; AOPA $\leftarrow$ {assign, revoke}
0: b. ROLESA $\leftarrow$ ROLESU ; AUAA = $\phi$
0: c. RHA $\leftarrow$ RHU
0: /* Map attribute functions in AURA */
0: a. AATTA $\leftarrow$ {grantAuth, empowerAuth, adminAuth,
0: roleClassAuth}, ARATT = {}, RATT = {}
0: b. ScopeA(grantAuth) = ROLESA,
0: c. attTypeA(grantAuth) = set
0: d. is_orderedA(grantAuth) = True, HAgrantAuth = RHA
0: e. For each u in AUU, grantAuth(u) = $\phi$
0: f. For each u in USERSU and
0: for each [role, r, grant] $\in$ authorized_permsU[u],
0: grantAuth(u)' = grantAuth(u) $\cup$ (r)
0: g. ScopeA(empowerAuth) = ROLESA,
0: h. attTypeA(empowerAuth) = set
0: i. is_orderedA(empowerAuth) = True,
0: j. HAempowerAuth = RHA
0: k. For each u in AUU, empowerAuth(u) = $\phi$
0: l. For each u in USERSU and
0: for each [role, r, empower]
0: $\in$ authorized_permsU[u],
0: empowerAuth(u)' = empowerAuth(u) $\cup$ r
0: m. ScopeA(adminAuth) = ROLESA
0: n. attTypeA(adminAuth) = set
0: o. is_orderedA(adminAuth) = True,
0: p. HAadminAuth = RHA
0: q. For each u in USERSA, adminAuth(u) = $\phi$
0: r. For each u in USERSU and
0: for each [role, r, admin] $\in$ authorized_permsU[u],
0: adminAuth(u)' = adminAuth(u) $\cup$ r
0: s. ScopeA(roleClassAuth) = AMU(role)
0: t. attTypeA(roleClassAuth) = set
0: u. is_orderedA(roleClassAuth) = False
0: v. HAroleClassAuth = $\phi$
0: w. For each u in USERSA, roleClassAuth(u) = $\phi$
0: x. For each u in USERSU
0: for each [c, am] $\in$ authorized_permsU[u],
0: roleClassAuth(u)'= roleClassAuth(u) $\cup$ am
0: /* Construct assign rule in AURA */
0: a. assign_formula = (r1 $\in$ grantAuth(au) $\wedge$ r2 $\in$
empowerAuth(au)) $\vee$ (r1 $\in$ grantAuth(au) $\wedge$ empower $\in$
roleClassAuth(au)) $\vee$ (grant $\in$ roleClassAuth(au) $\wedge$ r2 $\in$
empowerAuth(au)) $\vee$ (grant $\in$ roleClassAuth(au) $\wedge$ empower $\in$
roleClassAuth(au))
Continuation of Algorithm 1 MapRRA-UARBAC
0:
0: b. auth_assign =
0: is_authorizedUassign(au : AU, r1 : ROLES, r2 : ROLES) $\equiv$
assign_formula
0: /* Construct revoke rule for AURA */
0: a. revoke_formula = r1 $\in$ grantAuth(au) $\wedge$ r2 $\in$
empowerAuth(au)) $\vee$ r1 $\in$ adminAuth(au) $\vee$ r2 $\in$ adminAuth(au)
$\vee$ admin $\in$ roleClassAuth(au)
0: b. auth_revoke =
0: is_authorizedUrevoke(au1 : AU, r1 : ROLES, r2 : ROLES) $\equiv$
revoke_formula
### IV-E MapRRA-UARBAC
We presented example instances for UARBAC’s RRA and its equivalent AARA
instance in previous segments. Algorithm IV-D presents MapRRA-UARBAC, a
procedure to map any instance of UARBAC’s RRA [8] to its equivalent ARRA
instance. It take instance of UARBAC’s RRA as its input and yields an
equivalent AARA instance as its output. Sets and functions from UARBAC and
ARRA are labeled with superscripts U and A, respectively. Input consists of
CU, USERSU, ROLESU, PU, RHU, AMU(role), For each u $\in$ USERSU,
authorized_permsU[u], and For every r1, r2 $\in$ ROLESU. Grant operation
grantRoleToRole(r1, r2) will be true if the granter has either [role, r2,
empower] and [role, r1, grant] or, [role, r2, empower] and [role, grant] or,
[role, empower] and [role, r1, grant] or, [user, empower] and [role, grant]
permissions towards roles.
For each r1, r2 $\in$ ROLESU, revokeRoleFromUser(r1, r2) is true if the
granter has either [role, r2, empower] and [role, r1, grant] or, [role, r1,
admin] or, [role, r2, admin] or, [role, admin] permission on roles.
MapRRA-UARBAC yields an ARRA instance consisting of AUA, AOPA, ROLESA, ARA,
AUAA, RHA, ARHA, AATTA, For each attribute att $\in$ AATTA, ScopeA(att),
attTypeA(att), is_orderedA(att) and HAatt, For each user au $\in$ AUA, and for
each att $\in$ AATTA, att(au), Authorization rule for assign (auth_assign),
and Authorization rule for revoke (auth_revoke).
In Step 1 primary sets from UARBAC’s RRA are mapped to AURA equivalent sets.
In Step 2, admin user attributes are defined. In UARBAC model, role-role
assignment decisions are based on the admin user’s _access modes_ such as
grant, empower and admin towards regular roles. For example, for each
permission [role, r, grant] in authorized_permsU[u], grantAuth(au) attribute
extracts set of roles towards which an admin user au has grant authority. In
other words, admin user can grant those roles to any other object such as user
and role. Similarly, empowerAuth yields set of roles to which an admin user
has empower access mode and adminAuth yields sets of roles about which admin
user has admin access mode. The meaning of each access mode is preserved by
UARBAC model. For each class level attribute [c, am] in authorized_permsU[u],
where c is a type of class and am is type of access mode, roleClassAuth(au)
specifies the nature of class level access mode such as grant or admin admin
user au has on role class.
In Step 3, an authorization formula equivalent to grantRoleToRole(r1, r2) in
UARBAC is expressed. Equivalent expression is_authorizedRassign(au : AUA, r1 :
ROLESA, r2 : ROLESA) checks if an admin user au grant authority over role r1
and empower authority over role r2. Similarly, in Step 4 revoke_formula
equivalent to revokeRoleFromRole(r1, r2) is constructed using revoke
permissions given to admin user. It checks whether an admin user au has an
authority to revoke roles r1 and r2.
### IV-F An Example Instance for ARRA with role attributes
Previous ARRA model simulation examples did not include regular role
attribute. In this section, we present a simple yet plausible example that
demonstrates a use case for role attributes in assigning role to role.
There are two admin users, Sam and Tom, and regular roles IT Director,
Development Manager, Quality Manager, Marketing Mganager, Finance Manager,
Support Engineer and System Analyst. There can be many other roles in an
organization. We have considered few roles, enough to illustrate our case.
Admin user attribute dept maps each admin user to set of departments they have
admin authority over. Although, it is likely that departments have hierarchy
in practice, we have not included hierarchy for departments for simplicity.
There are three departments in the organization, namely Operations, Account
and IT departments. Sam has admin authority over all the departments while Tom
has authority over IT department only. There is a regular role attribute dept
for mapping regular roles to their departments.
Authorization condition is that if an admin user au with admin role ar was
given authority over a department d and if roles r1 and r2 belonged to the
same department d then r1 can be assigned to r2.
Sets
– AU = {Sam, Tom}, AOP = {assign, revoke}
– ROLES = {IT Director, Devlopment Mgr., Quality
Mgr., Marketing Mgr., Finance Mgr., Support Engineer, System Analyst}
– AR = {}
– AUA = {}
– RH = {}, ARH = {}
Attributes definition
– AATT = {dept},
– Scope(dept) = {Operations, Account, IT},
attType(dept) = set, is_ordered(dept) = False,
Hdept = $\phi$
– dept(Sam) = {Operations, Account, IT},
dept(Tom) = {IT}
– ARATT = {}, RATT = {dept, level}
– Scope(dept) = {Operations, Account, IT},
attType(dept) = atomic,
is_ordered(dept) = False, Hdept = $\phi$
– dept(IT Director) = IT, dept(Dev. Manager) = IT,
dept(Marketing Mgr.) = Operations,
dept(QA Mgr.) = IT,
Authorization functions
– is_authorizedRassign(au : AU, r1 : ROLES, r2 : ROLES) $\equiv$ $\exists$d
$\in$ Scope(dept). dept(au) = dept(r1) = dept(r2)
– is_authorizedRrevoke(au : AU, r1 : ROLES, r2 : ROLES) $\equiv$
is_authorizedRassign(au : AU, r1 : ROLES, r2 : ROLES)
## V Conclusion
In this paper, we presented ARRA, a model for attribute based role-role
assignment. A design motivation behind ARRA model was to make it enough to
express prior RRA models. In particular, we took RRA97 and UARBAC’s RRA model
as foundation. For these models, we have presented MapRRA97 and MapRRA-UARBAC
algorithms, which map any instances of these models to ARRA model,
respectively.
ARRA model can not only express existing RRA models but also has a capability
to do more. For instance, prior models included in our research did not fit
the notion of regular role attributes. However, we acknowledge that it is an
essential attribute in making role assignment decisions to say the least.
Motivated by this notion, we presented a simple example scenario where we made
use of regular role attributes for role-role assignment.
## Acknowledgment
This work is partially supported by NSF grants CNS-1423481 and CNS-1553696.
## References
* [1] D. F. Ferraiolo, R. Sandhu, S. Gavrila, D. R. Kuhn, and R. Chandramouli, “Proposed NIST standard for role-based access control,” _ACM Transactions Inf. Syst. Secur._ , vol. 4, pp. 224–274, Aug. 2001.
* [2] R. S. Sandhu, E. J. Coyne, H. L. Feinstein, and C. E. Youman, “Role-based access control models,” _Computer_ , vol. 29, pp. 38–47, Feb 1996.
* [3] L. Fuchs, G. Pernul, and R. Sandhu, “Roles in information security - A survey and classification of the research area,” _Computers & Security_, vol. 30, pp. 748 – 769, 2011.
* [4] A. C. OConnor and R. J. Loomis, “2010 economic analysis of role-based access control,” _NIST, Gaithersburg, MD_ , 2010.
* [5] R. Sandhu, V. Bhamidipati, and Q. Munawer, “The ARBAC97 model for role-based administration of roles,” _ACM Transactions on Information and System Security (TISSEC)_ , vol. 2, pp. 105–135, 1999.
* [6] R. Sandhu and Q. Munawer, “The ARBAC99 model for administration of roles,” in _Proc. of the 15th Annual Computer Security Applications Conference (ACSAC’99)_. IEEE, 1999, pp. 229–238.
* [7] S. Oh and R. Sandhu, “A model for role administration using organization structure,” in _Proc. of the seventh ACM symposium on Access control models and technologies_. ACM, 2002, pp. 155–162.
* [8] N. Li and Z. Mao, “Administration in role-based access control,” in _Proc. of the 2nd ACM symposium on Information, computer and communications security_ , 2007, pp. 127–138.
* [9] P. Biswas, R. Sandhu, and R. Krishnan, “Uni-ARBAC: A unified administrative model for role-based access control,” in _International Conference on Information Security_. Springer, 2016, pp. 218–230.
* [10] D. R. Kuhn, E. J. Coyne, and T. R. Weil, “Adding attributes to role-based access control,” _Computer_ , vol. 43, pp. 79–81, 2010.
* [11] V. C. Hu, D. Ferraiolo, R. Kuhn, A. Schnitzer, K. Sandlin, R. Miller, K. Scarfone _et al._ , “Guide to attribute based access control (ABAC) definition and considerations,” _NIST special publication_ , vol. 800, no. 162, 2014.
* [12] X. Jin, R. Krishnan, and R. Sandhu, “A unified attribute-based access control model covering DAC, MAC and RBAC,” in _IFIP Annual Conference on Data and Applications Security and Privacy_. Springer, 2012, pp. 41–55.
* [13] S. Bhatt, F. Patwa, and R. Sandhu, “ABAC with group attributes and attribute hierarchies utilizing the policy machine,” in _Proc. of the 2nd ACM Workshop on Attribute-Based Access Control_. ACM, 2017, pp. 17–28.
* [14] P. Biswas, R. Sandhu, and R. Krishnan, “Attribute transformation for attribute-based access control,” in _Proc. of the 2nd ACM Workshop on Attribute-Based Access Control_. ACM, 2017, pp. 1–8.
* [15] J. Ninglekhu and R. Krishnan, “Attribute based administration of role based access control: A detail description,” _arXiv preprint arXiv:1706.03171_ , 2017.
* [16] J. Crampton, “Administrative scope and role hierarchy operations,” in _Proceedings of the seventh ACM symposium on Access control models and technologies_. ACM, 2002, pp. 145–154.
* [17] X. Jin, R. Sandhu, and R. Krishnan, “RABAC: role-centric attribute-based access control,” in _International Conference on Mathematical Methods, Models, and Architectures for Computer Network Security_. Springer, 2012, pp. 84–96.
* [18] Q. M. Rajpoot, C. D. Jensen, and R. Krishnan, “Attributes enhanced role-based access control model,” in _International Conference on Trust and Privacy in Digital Business_. Springer, 2015, pp. 3–17.
* [19] E. Yuan and J. Tong, “Attributed based access control (ABAC) for web services,” in _Proc. of the IEEE International Conference on Web Services (ICWS’05)_. IEEE, 2005.
* [20] D. Servos and S. L. Osborn, “HGABAC: Towards a formal model of hierarchical attribute-based access control,” in _International Symposium on Foundations and Practice of Security (FPS 2014)_. Springer, 2014, pp. 187–204.
|
∎
11institutetext: Devansh Mehta 22institutetext: Voicedeck Technologies
22email<EMAIL_ADDRESS>33institutetext: Kalika Bali 44institutetext:
Microsoft Research India
44email<EMAIL_ADDRESS>
# Learnings from Technological Interventions in a Low Resource Language
Enhancing Information Access in Gondi
Devansh Mehta∗∗ ** Equal Contribution Harshita Diddee∗11footnotemark: 1 Ananya
Saxena Anurag Shukla Sebastin Santy Ramaravind Kommiya Mothilal Brij Mohan
Lal Srivastava Alok Sharma Vishnu Prasad Venkanna U Kalika Bali
(Received: date / Accepted: date)
###### Abstract
The primary obstacle to developing technologies for low-resource languages is
the lack of representative, usable data. In this paper, we report the
deployment of technology-driven data collection methods for creating a corpus
of more than 60,000 translations from Hindi to Gondi, a low-resource
vulnerable language spoken by around 2.3 million tribal people in south and
central India. During this process, we help expand information access in Gondi
across 2 different dimensions (a) The creation of linguistic resources that
can be used by the community, such as a dictionary, children’s stories, Gondi
translations from multiple sources and an Interactive Voice Response (IVR)
based mass awareness platform; (b) Enabling its use in the digital domain by
developing a Hindi-Gondi machine translation model, which is compressed by
nearly 4 times to enable it’s edge deployment on low-resource edge devices and
in areas of little to no internet connectivity. We also present preliminary
evaluations of utilizing the developed machine translation model to provide
assistance to volunteers who are involved in collecting more data for the
target language. Through these interventions, we not only created a refined
and evaluated corpus of 26,240 Hindi-Gondi translations that was used for
building the translation model but also engaged nearly 850 community members
who can help take Gondi onto the internet.
###### Keywords:
Low-Resource Languages, Deployment, Applications
††journal: Language Resources and Evaluation
## 1 Introduction
In the present era of globalization and integration of technology into almost
every aspect of life, native speakers are turning to dominant languages at a
faster rate than ever before. As seen in Figure1, around 40% of all languages
in the world face the danger of extinction in the near future 31 , while 95%
have lost the capacity to ascend to the digital realm kornai2013digital . When
a language spoken in a particular community dies out, we lose a vital part of
the culture that is necessary to completely understand it. At a bare minimum,
languages need to be integrated with the Internet to give them a fighting
chance at survival.
In this work, we design, deploy and critically assess technological
interventions that aim to create a digital ecosystem of Gondi content, a
South-Central Dravidian tribal language spoken by the Gond tribe in Central
India. The case of Gondi is representative of many languages across the world
and presents a unique case study of how a language can be in danger despite
having all the ingredients of sustainability such as (1) long historical
continuity (2) a population of 3 million people speaking it and (3) widely
spoken in around 6 states of India with various dialects and forms. The
complexities arise as Gondi is a predominantly spoken language with no single
standard variety but a number of dialects, some mutually unintelligible
beine1994sociolinguistic . Our objectives with these interventions is creating
a repository of linguistic resources in Gondi that can be used for: (1)
building language technologies like machine translation or speech to text
systems that are essential for taking Gondi onto the internet; and (2)
expanding the information available to the Gond community in their language.
There needs to be groundwork and identification of the real problems that can
be solved by deployment of language technologies in minority communities
dearden2015ethical , as there are legitimate technological and ethical
concerns surrounding the use of technology in low-resource languages
joshi2019unsung ; 10.1145/3530190.3534792 . Technical systems aren’t simply
ethically neutral exchanges of information and those that work in isolation
from social contexts can be detrimental to the ecosystem of minority language
communities. At the same time, inaction on the part of language technologists
is an ethically dubious proposition in its own right, as minority communities
are currently forced to learn mainstream languages for availing the benefits
of the Internet and to access power structures enabling socio-economic
mobility. We believe it to be fairly non-controversial to maintain that anyone
should be able to surf the internet in the language that they are comfortable
in. Ethical issues only arise on the methods one uses to reach this goal.
Figure 1: UNESCO 2017 World Atlas: Demonstrating the endangered state of the
languages of the world.
Our focus when working with the Gond community is centered around devising
novel approaches for creation, evaluation and dissemination of Gondi content,
unlike well-resourced languages where the focus is more on engineering
language technologies using abundantly available data. As far as possible, we
have ensured that the technology interventions described in this study were
led by the Gond community and non-profit organizations working with them.
The first intervention is creating a 3,500 word Gondi dictionary that is
accessible to the community as an Android app. The second is 230 children’s
books that were translated by the Gond community in a 10 day workshop. The
third is crowdsourcing translations from community volunteers via an Android
app, the WhatsApp Business API and an assistive translation tool. The fourth
intervention is a phone number that community members could call to gain
awareness about a local election in their area, upon completion of which they
earn mobile credits. These interventions respectively resulted in facilitating
80-100 community members to create a 3,500 word dictionary; 20 community
members that translated 230 children’s stories from Hindi to Gondi; 219
community members that translated nearly 55,000 sentences; and 557 native
speakers of Gondi that learned about elections in their own language and who
can be called for future workshops.
This data was used to develop a Hindi-Gondi Machine Translation
Model111https://github.com/microsoft/INMT-lite, compressed for streamlined
operation on low-resource edge devices for use by the community. The model was
evaluated by community members that engaged with us through these
interventions and got an average Direct Assessment score specia-
etal-2020-findings-wmt of 63, indicating the semantic accuracy of our model’s
output. We also run preliminary experiments for evaluating the model’s ability
in assisting community members with translations. Our results show promising
efficacy with more than 50% of the suggestions being accepted.
## 2 Context
According to the 2011 census chandramouli2011census , the total population of
the Gond tribe is approximately 11.3 million. However, the total Gondi
speaking population is only around 2.7 million. That is, only about 25 percent
of the entire tribe now speaks it as a first language. UNESCO’s Atlas of the
Worlds Languages in Danger moseley2010atlas lists Gondi as belonging to the
vulnerable category. There is an added difficulty of creating resources for
Gondi due to the linguistic heterogeneity within the Gond community. As seen
in Figure2, Gondi is spread over 6 states in India. It is heavily influenced
by the dominant language of each state to the point where a Gond Adivasi from
Telangana (a Telugu speaking Southern state), finds it difficult to understand
a Gond Adivasi from Madhya Pradesh (a Central state with Hindi as the dominant
language). Recent scholarship has found the influence of these dominant
languages to be a major factor in language loss in the Gond community
boruah2020language .
Figure 2: Gondi speaking areas in India
As a predominantly oral language, the proportion of Gondi speakers is expected
to go down further as opportunities are shrinking to hear the language spoken
outside of their everyday surroundings. All India Radio, the only radio
station in India allowed to broadcast news, does not have regular Gondi news
bulletins. There is no TV station or channel catering to Gondi speakers. There
is also a severe dearth of online content in Gondi, resulting in members of
the tribe having to learn a mainstream language to enjoy the benefits of
internet connectivity. “The Wire”222Gondi Bulletin:
https://youtu.be/M3q2ycJ_U7g is one of the few Indian news outlet that
publishes a news bulletin in Gondi. However, the Gondi spoken in their
broadcast caters to the Gond Adivasis from the Central states of India and it
is difficult for Gonds from the Southern states to understand the content.
Gondi is also not included in the 8th Schedule of the Indian Constitution,
with the result that education and exams for government jobs cannot be
administered in the language. The deleterious effects of the marginalization
of Gondi on their community are manifold. Gondi is considered the lingua
franca of the local insurgents, who use their knowledge of the language and
the perceived neglect by the government to recruit candidates from the tribe
to join their civil war against the state kumar2019gondi . Further, there are
high dropout rates among monolingual children that speak Gondi as a first
language. A UNESCO study found that children whose mother tongue is not the
medium of instruction at primary school are more likely to fail in early
grades or drop out buhmann2008mother , which in turn increases the chances of
them joining the insurgency. Working with the Gond tribes on reviving their
language is thus important not just for cultural reasons, but may also serve
as an instrument for bringing peace to their society. These factors can be
mitigated provided there is political will and a standardized version of the
language that is agreed upon. While the younger, bilingual generation in the
Gond tribe attach less importance to their mother tongue, it was not difficult
to find community members worried about the extinction of their language and
it proved easy to recruit participants for the workshops we held, even without
any financial incentive.
## 3 Related Work
The first section focuses on how we address the constraints of efficiently
collecting data through traditional channels for low-resource languages. The
second section situates our study within existing language preservation and
revitalization programs and lays out our unique contributions to this prior
literature.
##### Data Collection Channels
Traditional channels of data collection can vary across 3 dimensions (a)
Channel of Data Collection (b) Quality of Collected Data and (c) Quantity of
Collected Data. There are two common channels of data collections: automated
mining and human reliant crowdsourcing. Mining includes a wide spectrum of
methods varying from crawling data off the web to utilizing content after
digitization by systems such as Optical Character Recognition. Unfortunately,
these channels assume the existence of data resources in a specific form: for
example, Automatic or Semi Automatic methods like Web mining expects the data
to be digitally readable, while OCR expects data to be digitally consumable by
an existing service, which might not be the case for many languages,
especially low-resource ones rijhwani2020ocr .
In our study, we rely on crowdsourcing large quantities of data through human
reliant channels. This approach is fraught with logistical challenges around
participant recruitment nangia2021ingredients , the cost of response
validation, and mitigating subjectivity in responses goldman-
etal-2018-strategies , as well as ethical considerations pertaining to
devising fair pay kummerfeld-2021-quantifying shmueli2021beyond and
maintaining response diversity by controlling for human bias in the inputs
provided by a crowdsourced audience liu-etal-2022-toward . Our study adds
novel insights to this prior literature by building a dataset from community
volunteers and staff at partner organizations without the provision of any
fiscal incentive, which is often a given attribute of human reliant data
crowdsourcing channels. We find that that a volunteer led workforce for data
generation can yield higher quality datasets by self-selecting for those
motivated to preserve their language.
Prior work has identified quality constraints for corpora arising out of
human-collection channels 10.1162/tacl_a_00447 caswell-etal-2020-language .
Specifically, noise derivative of code-mixing, non-linguistic characters,
short/meta-data segments, untranslated/copied content has been shown to
considerably harm the output of neural machine translation (NMT) systems
khayrallah2018impact . For high resource languages, the sheer quantity of a
corpus is usually significant enough to allow the selection of a subset of
high-quality sentences from the larger set using intelligent validation
techniques, especially since some percentage of noisy data is shown to improve
generalization robustness of NMT models. Training and evaluation corpora in
low-resource languages may not be as effective due to the paucity of data
10.1162/tacl_a_00447 . We observe this explicitly with our interventions,
since a large proportion of our data (nearly 60K data samples that we
collected) were sifted to a set of 27K samples, greatly reducing the effective
yield of our collection channels. Our work thus presents a case study on
developing alternative methods of data collection and evaluation for low-
resource languages.
##### Language Technologies for Revitalization and Preservation
Preservation programs generally attend to the need to develop data corpora in
that language abraham2020crowdsourcing , article , mus2021toward . However,
the notion that language documentation or artifact creation can independently
bring about revitalization is largely dismissed by the community zariquiey-
etal-2022-cld2 . Community-centric technical interventions that can infuse the
language technology into the community for sustained use are increasingly
coming under focus. One dimension of this focus has been towards developing
fundamental NLP tools for low-resource languages such as Part-Of-Speech
taggers finn-etal-2022-developing , parsers bowers-etal-2017-morphological ,
etc. Another dimension is the development of more end-user oriented language
technologies such as edge-deployments bettinson-bird-2017-developing
hermes2013ojibwe bird2014aikuma and interactive interfaces that can be
consumed rapidly by the community little-2017-connecting adams-etal-2021-user
. A critical point to note here is that access bottlenecks can limit the
consumption of outcomes from revitalization and preservation endeavours. For
instance, Roche highlights the political nature of language endangerment,
pointing out that low-resource language communities are usually under-
resourced in other aspects of life roche2020abandoning . Leibling et al
10.1145/3313831.3376261 discuss these constraints across the adaptation of
translation systems to edge devices across three globally distributed and
divergent demographic groups. We contribute to this identified gap by creating
a software artifact that displays NMT translations on low cost devices like
smartphones, which can be used by other speech communities and researchers
working with marginalized populations.
Due to the well-studied constraints of data collection through traditional
channels elazar-etal-2020-extraordinary millour-fort-2020-text liu-
etal-2022-toward , our work explores newer mechanisms for collecting low-
resource language data that are of wider relevance. For example, our Adivasi
Radio translation app allowed users to bulk download a corpus of sentences for
translation, which they could then complete even from areas without Internet.
Another area our work sheds new light on is reducing the cognitive burden
faced by communities in understanding an entirely novel interface for data
provision, as discussed in prior work like 10.1145/2858036.2858448
10.1145/2556288.2557155 . Our study uses low friction approaches for
collecting translations such as through the WhatsApp Business API, which is
well integrated into many communities. Lastly, our work briefly touches upon
the attempt to integrate assistive translation interfaces to provide sub-
optimal help in data collection efforts 10.1145/3290605.3300461
10.1145/3530190.3534792 .
Overall, this paper is a case study in collaborating with target communities
for creating linguistic resources to enable the development of language
technologies that can then give rise to a digital ecosystem for such
languages. In this regard, we follow the lead set by Kornai and Bhattacharyya
in building online tools for low-resource languages like spellchecks that can
be incrementally refined over time kornai2014indian . Our work distinguishes
itself by empirically studying methods of community engagement that are
cognizant of the constraints faced in low-resource environments.
## 4 Technological Interventions
The larger framework of our interventions is that language resource creation
feeds into building language technologies and enhancing access to information
in that language. We rolled out a series of interventions that created
linguistic resources useful to the community, such as children’s stories,
which in turn fed into building language technologies such as a Hindi-Gondi
translation system. These initiatives helped increase information access in
Gondi while supporting the end goal of taking Gondi onto the Internet.
### 4.1 Gondi Dictionary Development
This section discusses the need and complexity of building a common dictionary
across the various Gondi dialects.
#### 4.1.1 Motivation
Several researchers have found that mutual intelligibility of Gondi decreases
with distance, in part due to the influence of dominant state languages
creeping into the various dialects beine1994sociolinguistic ;
shapiro1981language ; tyler1969koya . Beine (1994) conducted mutual
intelligibility tests across the Gondi speaking areas and found there to be 7
mutually unintelligible dialects of the Gondi language (pp 89). He thus
recommended the creation of dialect centres to cover each Gondi speaking
region, with literacy materials separately developed for each center. These
workshops helped determine whether the community wanted separate efforts for
each dialect or a common effort towards one language.
#### 4.1.2 Intervention
Prior to the involvement of our research team, a citizen journalism platform
called CGNet Swara held 7 workshops beginning from 2014 to develop a Gondi
thesaurus containing all the different words used by native speakers from 6
states. Some words, such as water, had as many as 8 different words for it. At
the 8th workshop in 2018, which saw more than 80 people in attendance, the
thesaurus was developed into a dictionary containing 3,500 words.
The dictionary was made into an Android app, Gondi Manak
Shabdkosh333aka.ms/gondi-dictionary 444aka.ms/indian-express-gondi-dictionary,
depicted in Figure3. It allows users to enter a Hindi or Gondi word and hear
or read its equivalent translation, similar to the Ma! Iwaidja dictionary
app555ma-iwaidja-dictionary.soft112.com/ without the wheel based interface for
conjugation and sentence formation.
#### 4.1.3 Takeaways
CGNet Swara reported an overwhelming consensus by the community to remain as
one language. Gondi is primarily used within its local speech community and
participants recognized the need to develop a common vocabulary and spelling
for its emergence as a standard language. The three techniques commonly used
in standardizing languages are comparative (linguistic reconstruction to build
a mother tongue for all), archaizing (deriving a variety from older, written
texts) and the statistical (combining the different dialects having the widest
usage) haugen1966dialect , with most participants favoring the latter
approach.
CGNet’s initial aim with the dictionary app was studying whether it can allow
some basic communication and learning to take place in primary schools where a
teacher does not know Gondi while their student is a monolingual speaker.
However, our approach of finding a central dialect mediating between the
extremes resulted in a variety of Gondi promoted as everyone’s language that
became nobody’s language. For example, monolingual Gonds may only be aware of
their local word for water, but if it hasn’t been selected in the dictionary
then its use in such scenarios would be limited. In retrospect, a thesaurus
might have been more applicable for this particular use case. The dictionary
provided greater utility in pretraining our translation model. Following Wang
et al wang2022expand we augmented digitally consumable monolingual data to
improve performance of the model, as demonstrated in Table 2.
Figure 3: Gondi Manak Shabdkosh, a dictionary of 3,500 words
### 4.2 Creating children’s books in Gondi through Pratham Books
We narrate our efforts at translating children’s stories into Gondi for use in
schools and as a data source for language technologies.
#### 4.2.1 Motivation
Imparting basic, primary education in a child’s mother tongue inculcates pride
in their language and cultural identity formation among younger generations.
It also targets the most important stakeholder for ensuring continued language
use. More practically, having Gondi as a medium of instruction or a subject at
schools can help increase enrollment among monolingual Gond tribes within the
education system and decrease drop-out rates among indigenous children not
conversant in Hindi. This intervention focused on developing the necessary
linguistic resources in Gondi that could serve as educational material.
#### 4.2.2 Intervention
In 2018, our research team conducted a 10 day translation workshop with CGNet
Swara and Pratham Books, a nonprofit publisher with the motto ”Book in every
child’s hands”. 20 Gond tribals from three states of India came together and
translated 230 children’s books from Hindi to Gondi, some of which introduced
Gondi children to climate change for the first time 666Article:
aka.ms/news18-climatechange. These stories were published on the Storyweaver
platform 777storyweaver.org.in/, an initiative of Pratham Books that hosts
more than 15,000 children’s stories in various languages and dialects. To the
best of our knowledge, this is the first online repository of children’s
stories in Gondi 999Article: aka.ms/hindu-gondi-nextgen888aka.ms/storyweaver-
gondi. They were also printed out and distributed in primary schools in the
Gondi speaking districts of Chhattisgarh, with efforts now ongoing to convince
the state government to include them as part of the school curriculum across
the tribal belts of the state.
#### 4.2.3 Takeaways
Although no payment was provided to workshop participants (besides covering
food, travel and lodging), a total of around 20 volunteers translated 8000
sentences over 10 days. Having a volunteer only workforce helped in self-
selecting individuals that cared about the languages survival among future
generations. We found it most effective to create teams of 4-5 people for
translating a story, with more educated participants taking on the role of
writing the translation on paper or typing the written translation into the
computer interface. Monolingual Gond speakers would help find the right
vocabulary for translations, while bilingual speakers would help them
understand the Hindi text that needed translating. Having groups complete
translations (rather than individuals) resulted in considerable debate around
which Gondi words to use for translating a Hindi sentence. This intervention
yielded the most high quality data compared to the other efforts, as is
evident from our manual evaluation of the data referenced in Table 1.
### 4.3 Crowdsourcing Gondi translations through Adivasi Radio
Unlike the Pratham Books workshop featuring in-person group translations,
Adivasi Radio was geared towards the post pandemic world and allowed
volunteers to provide translations from the comfort of their homes.
#### 4.3.1 Motivation
At the children’s book workshop, we found that many participants wanted to
continue the translation work from home but there was no avenue for them to do
so. Taking inspiration from Aikumabird2014aikuma , we developed Adivasi Radio,
an Android application that presents users with Hindi words or sentences for
which they need to provide the Gondi translation. In addition to the
translation role, Adivasi Radio was designed as the go-to place for native
speakers to find outlets and sites publishing Gondi content.
Figure 4: Adivasi Radio app to collect translations and access Gondi content
#### 4.3.2 Intervention
The app was first deployed between March and June 2020, during the initial
months of the coronavirus induced lockdown. The interface of the app can be
seen in Figure4 while translations over time are shown in Figure5. In 4
months, 17,756 sentences were translated through the app from 164 unique
users.
6 superusers, most of whom were staff at our partner organization CGNet Swara,
accounted for the majority of the translations. Many volunteers gave up after
completing a few translations, indicating that sustaining data collection
efforts is a challenge, especially without fiscal incentives.
We uploaded the children stories produced with Pratham Books, Gondi stories on
CGnet Swara’s citizen journalism platform and The Wire’s Gondi news bulletin
into the app. A Devanagari text-to-speech system read out the written Gondi
text so that monolingual speakers could enjoy the content.
Figure 5: Number of translations collected per day.
#### 4.3.3 Takeaways
After Adivasi Radio’s deployment, we held meetings with the translators to
learn about their experience of using the application. We found that several
users resided in areas having weak internet connectivity, thus prompting us to
update the application so that they could bulk download a large corpus of
sentences that they could translate even when they were offline. Similar to
the feature used in d2014mobile ; mehta2020facilitating for enabling citizen
journalism in areas without internet, translated sentences would get sent to
our server whenever the user managed to connect to the internet. The attention
to providing accurate translations was reflected in their request for both a
’skip’ feature to avoid providing translations they were unsure of and an
option to edit their earlier submissions, as they would sometimes realize
mistakes only after completing a translation. However, we found that the skip
feature was used to avoid translating long sentences or short paras of three
or more sentences. This concern could be alleviated by having them translate
sentences forming a cohesive narrative like a Wikipedia article rather than
disjointed sentences unrelated to one another. Finally, we did not see
significant engagement with the uploaded Gondi content available on the app as
most users were bilingual and accustomed to a richer online environment in
Hindi.
Our evaluation of the quality of submitted translations reiterated the
importance of incorporating rigorous validation checks on data collected
through human-channels. Out of the nearly 18000 sentences collected through
this channel, we were only able to utilize 10105 sentences after weeding out
transliteration, skipped sentences and gibberish inputs with a high density of
special and/or numeric characters.
### 4.4 Crowdsourcing Translations through WhatsApp Business API
A growing body of research advocates integration with existing platforms for
coordinating social mobilization instead of creating custom made ones such as
an app or website lambton2020unplatformed ; saldivar2019online ;
starbird2012crowdwork . The majority of Adivasi Radio users made extensive use
of WhatsApp in their daily life, prompting us to explore how we could use this
medium to solicit translations from the community.
Figure 6: Payments were discontinued for translations received via WhatsApp
after finding users were gaming the system
#### 4.4.1 Motivation
During field testing with Adivasi Radio users, we found that some users did
not have space on their phones to install our app. In such cases, we had to
delete media files and unused apps before we could install the Adivasi Radio
application on their phone. Moreover, training had to be provided on how to
navigate the interface of our app. These factors limited our ability to reach
more members in the community without our physical presence.
We thus applied for - and received - permission to use the WhatsApp Business
API, which would enable users to provide translations over WhatsApp directly
to our server without any human intervention. We accessed the API through the
IMI Connect platform, which has a monthly charge of about USD 350 per month in
addition to variable costs per message.
#### 4.4.2 Intervention
To seed usage of the system, we needed users that had a smartphone with
regular internet connectivity and were conversant in both Hindi and Gondi.
From May 2021 to September 2021, we collected 37,173 sentences from 55
volunteers and staff members at CGNet Swara. 6 superusers accounted for nearly
half of the submitted translations. Many translations were written in the
latin script (as opposed to Devanagari), resulting in a little less than
12,000 sentences that could be fed into the translation model.
#### 4.4.3 Takeaways
To address declining interest in completing translations from home via Adivasi
Radio, we released a probe in February 2021 that for the first time paid users
for completing translations via WhatsApp. We found that payment per sentence
translated led to deteriorating quality by rewarding speed over quality.
Indeed, we saw some of the most devoted community members spend more than 10
minutes on a single sentence, even asking their neighbors for the right word,
so that they could get it right. When these members saw their peers provide
inaccurate translations and earn more than them, they withdrew their support.
We ended up discarding nearly 12,000 sentences collected via this process and
switched back to a volunteer driven model. We also observed frustration when
the sentence that translators were given timed out and they were provided a
new sentence upon logging back in, as they had researched the earlier sentence
but there was no option for them to provide it anymore.
We relaunched the WhatsApp chatbot in June 2021 without the option of earning
any money for translations. The majority of the translations came from staff
members at CGNet Swara. In the absence of strict oversight and punishments for
incorrect translations, we recommend paying a fixed salary to translators
rather than one varying by number of sentences translated or other such
quantitative metrics. It also proved effective to integrate data collection
into the day to day responsibilities of employees at partner organizations.
### 4.5 Data Quality
We requested community members to manually evaluate the data collected and
then compare the quality of data collected via each channel.
#### 4.5.1 Annotation Setup
We recruited 3 Gondi speaking annotators from the Indian states of
Maharashtra, Madhya Pradesh and Chhattisgarh to assist us with the manual
evaluations of our data and the translation model developed from it. This
selection was made bearing in mind the dialectal coverage of our dataset, as
most translations were provided by volunteers from these three states. The
annotation instructions were made visible to all annotators via the interface
we provide to them as shown in Figure7. We also discussed a set of samples and
the mechanism to acquaint them with the scoring system before the evaluation
exercise. The participating annotators performed 3 tasks where they scored a
provided sample with the Direct Assessment score described in specia-
etal-2020-findings-wmt . They were compensated using the following
system101010 All compensations are specified in Indian National Rupee \- per
ranking task, each annotator was given Rs. 5 (Rs. 3 base pay, and Rs. 2 upon
completion). Per assistive translation task, they were provided Rs. 5 and per
translation task, they were provided Rs.10 ( Rs. 7 base pay, Rs. 3 upon
completion ).
Figure 7: The scoring instructions are specified in Hindi on the annotation
screen for quick reference. A score slider is provided for limiting scores to
a valid range.
To analyze the quality of the data collected through Pratham Books, Adivasi
Radio and WhatsApp API, we requested each annotator to score 60 sentences
sampled across any two data sources i.e. 2 annotators score 30 instances per
source. These samples were taken from our training corpus, and we apply a
sequence length filter in the range of [4, 15] tokens to avoid undue weightage
to sentences coming from any one source.
#### 4.5.2 Results
We report average direct assessment scores for each data source in Table 1.
Channel | Average DA Score | Standard Deviation
---|---|---
Pratham Books | 65.4 | 16.1
Adivasi Radio | 58.6 | 19.1
WhatsApp | 36.5 | 21.9
Model(Hindi-Gondi) | 63.2 | 23.2
Table 1: Average Direct Assessment Scores with Standard Deviations for data
collected via all channels and the Hindi-Gondi model trained on the data
The average DA scores on the sample set indicate that the data collected from
Pratham Books is of the best quality, followed by Adivasi Radio and WhatsApp
respectively. This could be due to the setup in which participants were
placed: synchronous and supervised for Pratham Books and asynchronous and
unsupervised for WhatsApp or Adivasi Radio.
As is observed with the standard deviation, the inter-annotator score
agreement was the poorest for WhatsApp. The annotators report that WhatsApp
data inputs had a very high degree of code-switching, sometimes not even with
Hindi but other indic languages. For example, all annotators reported
observing Marathi tokens as a part of the provided translations.
### 4.6 Machine Translation
Using the data collected, we train a machine-translation model between Hindi-
Gondi and evaluate the efficacy of its output.
#### 4.6.1 Motivation
The development of the machine translation model aimed at developing a core-
technology that serves the larger goal of improving information access for the
community by enabling automated translation of web content from Hindi to
Gondi. It also let us evaluate the utility of the data when it is plugged into
a real-world NLP usecase. We envisage the usage of this model in other low-
resource languages like that of Bribri, Wixarica and Mundari
https://doi.org/10.48550/arxiv.2210.15184 to demonstrate the applicability of
our artifact across other data-deficient setups.
#### 4.6.2 Intervention
Using the data we had available from all channels, we finetuned MT5 xue2021mt5
for training our translation model in both the Hindi-Gondi and Gondi-Hindi
direction. The model used in this work was adapted from
https://doi.org/10.48550/arxiv.2210.15184 and we replicate the training setup
specified there. The models were evaluated using sacrebleu (v2.2.0) with the
spm tokenizer. As explored by wang2022expand , we used the Hindi-Gondi
dictionary to generate 200K noisy monolingual data to train our models, which
we further used for continual pretraining. We also generated forward
translated data using the same model to fuse with the lexicon-adapted data.
Additionally, we utilize the quantized model developed by
https://doi.org/10.48550/arxiv.2210.15184 to evaluate the performance of the
same model’s compressed version (400 MB in disk size compared to 2.28 GB), for
the purpose of commenting on the edge deployment feasibility in low-resource
environments.
The annotation setup is same as the one described in section 4.5.1. All
annotators scored 40 inferences generated from the model to analyze its
quality.
Model | BLEU | Size
---|---|---
mt5-small (Hindi-Gondi) | 14.3 | 1.2GB
mt5-small (Gondi-Hindi) | 30.9 | 1.2GB
mt5-base (Hindi-Gondi) | 15.6 | 2.2GB
mt5-small with continued pretraining | 14.9 | 1.2GB
mt5-small with fused data (With 50K forward translated) | 12.6 | 1.2GB
Quantized MT5 | 13.8 | 400 MB
Table 2: Performance of the Hindi-Gondi translation model with model
adaptations like lexicon-adaptation, continual pretraining and quantization.
#### 4.6.3 Takeaways
As shown in Table 1, the annotators report an average DA Score of 63 for the
Hindi-Gondi model which connotes that the model’s output was semantically
appropriate but contained an identifiable degree of grammatical errors and
typos. The annotators mentioned that the model’s output was visibly poor in
terms of producing appropriate punctuation. While the structural integrity of
the sentence seems to have been preserved, spelling differences and code-
switched tokens ( i.e. Hindi tokens on the target side ) were common.
Annotators also observed the occurrence of ambiguous-language tokens that they
were unable to identify as being correct in the given context, which could be
attributed to the dialectal differences that existed within the language.
The BLEU scores of the models are reported in Table 2. The significant drop in
the BLEU score in the Hindi-Gondi direction, in comparison to Gondi-Hindi,
appears to be most significantly associated with the inclusion of Hindi in the
pretraining corpus of MT5. The mC4 training corpus contains 1.21% Hindi data,
which greatly appears to aid the target side reconstruction in the Gondi-Hindi
model xue2020mt5 . Additionally, we posit that the mt5 tokenizer, which has
never seen Gondi, expectedly falters in the reconstruction of Gondi on the
target side ( as is seen in the Hindi-Gondi ) model, so we expect a high
degree of tokenization ambiguities in the Gondi output which would adversely
impact the BLEU evaluation. This is further substantiated by observation of
annotators who claimed that in some cases constructions of a Gondi token
emulated the Hindi construction of the token. We intend to do a deeper
investigation of such spurious correlations that massively-multilingual models
might develop when a dominant attribute of language, its script in this case,
matches other well-represented languages in the pretraining corpus despite
being linguistically quite divergent.
One of the primary learnings from this task was the need to integrate
linguistically-motivated validation structures in a data collection channel to
optimize the yield of the data collection process. In our specific case, the
lack of any validation structures around the dialects against which we
collected our data for led to the collection of a dataset with samples from at
least 3 dialects. This especially harmed the impact of the machine translation
system as it’s responses catered to neither dialect completely.
### 4.7 Assistive Translation
We built an interface showing suggestions from our translation model to see if
it improves the quality or speed of translation efforts.
#### 4.7.1 Motivation
We use the same Hindi-Gondi model to generate candidate words that could be
used when translating Hindi sentences to Gondi. The motivation for this task
was to understand if poor-accuracy models could be used to accelerate the data
collection process by providing assistive recommendations to the participating
translators. This task holds unique practical value because it allows us to
comment if sub-optimal machine translations (which would be an obvious
intermediate artifact in such community interventions) can help enhance the
yield of data-generation pipelines in low-resource communities.
Figure 8: Gisting Interface for providing recommendations to data-providers.
Our motivation to develop this system did not come directly from the Gond
community but through partners at an educational institution and the citizen
journalism platform CGNet Swara. They believed it could enable speedy
translation of books for primary education and media stories into Gondi.
#### 4.7.2 Intervention
We realized that developing such a language technology had to take local
constraints into account, which in this specific case entailed the lack of a
stable internet connection. We thus developed an offline variant of the model,
which was compressed enough to operate on relatively low-resource devices
without requiring an internet connection.
The interface for this task, as shown in Figure8, is a screen that shows users
the model’s output as a set of candidate tokens that they can opt for while
submitting the translation. We sample 150 sentences from our development set
and provided it for translation through the Karya app. Inferences for these
sample sentences were provided to the annotators as a panel of assistive
candidate tokens. The annotators were requested to post-edit these candidate
gists until they were correct using the provided suggestions wherever
possible. During this exercise we monitor the number of suggestions accepted
by the annotators over the total number of suggestions presented.
#### 4.7.3 Takeaways
We find that annotators use 3.66 out of an average of 5.56 options shown to
them in each sentence translation iteration. This is further substantiated by
the observations of the annotators, who reported that nearly half of the
suggestions provided by the model were accurate and useful in generating the
final output. They also commented on the interface design by mentioning that
they preferred to type a token displayed in the panel of candidate tokens
instead of clicking on the token to avoid disrupting their flow while typing
out a translation. This indicates that tracking the number of accepted tokens
via clicks would have provided us a pessimistic view of the model’s usefulness
since annotators were visually aided by tokens (which might have been
completely or partially correct) and our experimental design would not have
logged that as a successful instance of aid. Annotators also pointed out that
some stop words were useful candidates in the Bag Of Words interface as they
preferred opting for them while typing a translation (rather than post-editing
them as a part of a gist). Hence, providing even noisy assistance with the
right interface appears to be a promising direction.
The annotators overarching evaluation of the translation system deemed it to
be semantically accurate. This gives hope that we can setup a virtuous cycle
where generation of linguistic resources enables the development of community-
oriented language technologies, which in turn could help in generating more
linguistic resources that can then further improve language technologies.
Additionally, their feedback on finding the candidate translations being
visually assistive encourages a deeper investigation into if and how
relatively low-accuracy models could be leveraged to improve the experience of
data providers in low-resource communities.
### 4.8 Disseminating Gondi content via Interactive Voice Response
Collecting data and building language technologies is only one part of the
equation; the other is creating avenues for communities to hear their language
broadcast over mass media and other communication channels.
#### 4.8.1 Motivation
There is an established body of work on the role of Interactive Voice Response
(IVR) forums for reaching communities that are too poor to afford smartphones,
too remote to access the Internet or too low-literate to navigate text-driven
environments swaminathan2019learn2earn ; dipanjan2019 ; revisitCG ; designLess
; ivrFarmers ; raza2018baang ; sawaal ; raza2013job ; sangeet . As an example,
Learn2Earn swaminathan2019learn2earn is an IVR based system that was used as
an awareness campaign tool to spread farmers’ land rights in rural India, HIV
literacy and voter awareness mehta2020using . Learn2Earn awards mobile
talktime to users who call a toll-free number and answer all multiple-choice
questions on the message correctly. Starting from an initial set of just 17
users, it was successful in spreading land rights awareness to 17,000 farmers
in 45 days via additional rewards for successful referrals
swaminathan2019learn2earn .
We adapted Learn2Earn for spreading voter awareness among Gondi speakers in
Dantewada (a rural district in the state of Chhattisgarh in India), during the
time that a bypoll election was being held. Prior work has shown that larger
the contexts, identities and communicative functions associated with a
language, the more likely it is to thrive walsh2005will . Voter rights content
in local languages during election season has the potential to encourage
conversations on topics of wider contexts and functions as well as contribute
to establishing representative and effective governance.
#### 4.8.2 Intervention
The pilot obtained users through seeding activities and referrals. Figure9
shows the number of unique calls that were made to our system and the number
of users who passed the quiz over time. Further, the Figureshows that a
majority of quiz passers came to know about our system via direct seeding,
though people knew that additional credits can be earned through referrals.
Usage dropped sharply after the elections in September concluded and seeding
activities were discontinued.
Figure 9: Number of calls to the system (blue line) and the number of users
who answered all questions correctly (bar plot). The blue and orange bars
represent users onboarded via seeding and through referrals
---
Metrics | Value
Unique callers (total, during and after seeding) | (557, 480, 77)
Unique callers per day - during seeding (min, mean, median, max) | (13, 48, 49, 80)
Unique callers per day - after seeding (min, mean, median, max) | (0, 4, 3, 20)
Callers answering all questions correctly | 313
Callers answering all questions correctly in their first call | 104
Calls made by callers (min, mean, median, max) | (1, 3, 2, 64)
Table 3: Summary statistics for our Learn2Earn deployment in Gondi
Table 3 highlights some important statistics in our Learn2Earn pilot. We
spread voter awareness in Gondi to 557 speakers, of which a majority (86%)
were reached during active seeding. Nearly 60% of the users correctly answered
all questions, either in their first attempt or in successive attempts,
indicating the effectiveness of the system for content dissemination. Further,
many users called the system more than once, with one user calling 64 times
(to refer someone, users had to place another call to the system). A follow-up
survey revealed that 22 of 113 users surveyed could not vote as they did not
have a voter ID card, potentially useful information for authorities to
increase voter turnout. One disappointing finding was that only about 8.3
percent of users were women.
#### 4.8.3 Takeaways
We see three clear benefits from conducting Learn2Earn pilots in endangered or
vulnerable languages. First, since it is entirely in the spoken form, only
native speakers of an endangered or vulnerable language can comprehend the
content and earn the reward. It is also easily accessible in low-resource
environments as it requires a missed call and nothing more. Second, the phone
numbers collected are an important dataset of speakers of that language and
can be used for future translation workshops and related programs that help
their economic and linguistic development. Finally, an oral language has a
tendency to die out unless there is an opportunity for that language to be
used outside of everyday surroundings, which periodic Learn2Earn campaigns on
important issues can help achieve.
## 5 Discussion
Steven Bird alerts us to the dangers inherent in a technology driven approach
to language revival, contending that commodifying indigenous languages as data
alienates native speakers. He argues for an approach rooted in self-
determination where outside experts are only engaged with to help implement a
program or strategy that the community has already settled on
bird2020decolonising . This sentiment is echoed by Stebbins et al (2017) who
maintain that ‘the community must be the driver and driving interest of the
research and language projects”stebbins2017living . Lewis and Simons carve out
an additional role for reflective practitioners of language development, that
of ”facilitating opportunities for local language users to interact with each
other” book .
There are also concerns that experts profit off their domain expertise in an
indigenous language with little benefit accruing to the communities
themselves. For example, Bird asks ”where are the full implementations,
deployed in robust software, in active use to capture primary data, leading to
curated language products that are being mobilized in speech communities?” By
his yardstick, even the interventions described in our paper fall short as
none have translated to scale or persistent use by the community. With the
benefit of hindsight, we can now reflect on whether our initiatives have been
exploitative.
Our position is that these interventions avoided an exploitative approach by
providing tangible outputs advancing one of the five FAMED conditions laid out
by Lewis and Simons as an ethical guide to language practitioners. This
acronym expands to having Functions associated with the language that are in
use and recognized by the community; a means for community members to Acquire
proficiency in the language; a Motivation for community members to use the
language, which is frequently but not exclusively economic; a policy
Environment that is conducive to the growth of the language; and a
Differentiated sphere for use of the language in adherence to established
societal norms.
Our machine translation artifact can be built upon by the Gondi community and
other low resource communities around the world to expand the functions
associated with their language, through automated translation of digital
content. The assistive translation app can help in language acquisition and
documentation by improving the speed of translating literary material and
providing vocabulary that translators may themselves be unaware of. Our
Learn2Earn pilot provided incentive payments to those speaking Gondi, thus
increasing (at a small scale) the motivation to use and recognize the
language. Translating children’s stories into Gondi can create a fertile
policy environment that makes possible the use of tribal languages in primary
education. The dictionary development process kickstarted a differentiated
sphere of Gondi usage between tribals of different states, whereas it has
historically been used only within their own region. Our thematic approach of
collecting language data to build technologies while simultaneously creating
resources for the community that advance one of the FAMED conditions can help
avoid the pitfall of alienating local speakers by commodifying their language
as data.
Overall, our work raises interesting questions on the role of the outside
interventionist in building langugage technologies and the relationship they
should have with the community. For example, the dictionary development
workshops primarily comprised bilingual and educated Gond speakers, which may
have contributed to their desire for standardizing the language. However, we
realized that this approach would be discriminatory to monolingual Gond
speakers who would have to accept the language thrown at them by the
intellectuals within their community since the standardization efforts would
not be universally agreed upon or even known to the majority of Gond speakers.
As Bird writes, ”No community speaks with a single voice, and in building
relationships we can unwittingly align ourselves with agendas, clans and gate-
keepers.” This prompts reflection on whether the outside interventionist
should simply fill in ”white man’s paperwork” and let the community lead, or
if there is a deeper role that we need to play, one where we use our perceived
neutrality to arbitrate between the different sections of a community. The
latter role finds expression in Lewis and Simons’ contention that ”it may be
an appropriate role for an outsider to act as an intermediary between the
different speech communities.” While we do not presume to have answers to
these questions, we hope that our work helps stimulate wider debate in the
language development community on defining our relationship with the speech
communities we work with and depend upon.
## 6 Conclusion
Engineering is the primary obstacle to designing technologies for well-
resourced languages. For low-resource ones, more focus needs to be given
towards designing methods for robust data collection and evaluation upon which
the language technology can be built.
To keep community members motivated through the data collection process, our
team strived to achieve 2 simultaneous goals: the collection of data upon
which language technologies such as speech to text or machine translation
could be built, and building a literary resource that community members could
point to as an immediate, demonstrable success. For example, the Learn2Earn
pilot in Gondi not only provided the community with an opportunity to earn
money for answering a quiz in their language and referring others to it, but
it also provided a dataset of native Gondi speakers. Similarly, translating
children’s stories and creating a standardized dictionary resulted in both
data upon which a machine translation tool was built and also tangible
language resources that can be used by the community.
We made use of the data collected to develop a machine translation model and
an assistive translation interface. Community evaluators reported that the
model fared poorly in punctuation and grammar, and roughly half the suggested
words were applicable to the sentence they were translating. Our future goals
are improving the models accuracy by feeding it more data and testing its
performance with data from other low-resource languages.
In 2020, the Chhattisgarh government began allowing primary education in 10
tribal languages, including Gondi
111111https://indianexpress.com/article/governance/chhattisgarh-education-
reforms-tribal-languages-to-be-a-medium-of-education-in-pre-school-6271547/.
At the community level, our focus is on working with the government to
integrate the linguistic resources and technological artifacts we’ve created
into their apparatus. We also plan on integrating the translation tool with
CGNet Swara’s citizen journalism newsroom so that they can regularly put out
Gondi content.
## References
* (1) Abraham, B., Goel, D., Siddarth, D., Bali, K., Chopra, M., Choudhury, M., Joshi, P., Jyoti, P., Sitaram, S., Seshadri, V.: Crowdsourcing speech data for low-resource languages from low-income workers. In: Proceedings of The 12th Language Resources and Evaluation Conference, pp. 2819–2826 (2020)
* (2) Adams, O., Galliot, B., Wisniewski, G., Lambourne, N., Foley, B., Sanders-Dwyer, R., Wiles, J., Michaud, A., Guillaume, S., Besacier, L., Cox, C., Aplonova, K., Jacques, G., Hill, N.: User-friendly automatic transcription of low-resource languages: Plugging ESPnet into elpis. In: Proceedings of the 4th Workshop on the Use of Computational Methods in the Study of Endangered Languages Volume 1 (Papers), pp. 51–62. Association for Computational Linguistics, Online (2021). URL https://aclanthology.org/2021.computel-1.7
* (3) Aiton, G.: Translating fieldwork into datasets: The development of a corpus for the quantitative investigation of grammatical phenomena in eibela. Proceedings of the Workshop on Computational Methods for Endangered Languages 2 (2021). DOI 10.33011/computel.v2i.973
* (4) Alagarai Sampath, H., Rajeshuni, R., Indurkhya, B.: Cognitively inspired task design to improve user performance on crowdsourcing platforms. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI ’14, p. 3665–3674. Association for Computing Machinery, New York, NY, USA (2014). DOI 10.1145/2556288.2557155. URL https://doi.org/10.1145/2556288.2557155
* (5) Beine, D.K.: A sociolinguistic survey of the gondi-speaking communities of central india. Ph.D. thesis, Univ. San Diego (1994)
* (6) Bettinson, M., Bird, S.: Developing a suite of mobile applications for collaborative language documentation. In: Proceedings of the 2nd Workshop on the Use of Computational Methods in the Study of Endangered Languages, pp. 156–164. Association for Computational Linguistics, Honolulu (2017). DOI 10.18653/v1/W17-0121. URL https://aclanthology.org/W17-0121
* (7) Bird, S.: Decolonising speech and language technology. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 3504–3519 (2020)
* (8) Bird, S., Hanke, F.R., Adams, O., Lee, H.: Aikuma: A mobile app for collaborative language documentation. In: Proceedings of the 2014 Workshop on the Use of Computational Methods in the Study of Endangered Languages, pp. 1–5 (2014)
* (9) Boruah, D.M.: Language loss and revitalization of gondi language: An endangered language of central india. Language in India 20(9) (2020)
* (10) Bowers, D., Arppe, A., Lachler, J., Moshagen, S., Trosterud, T.: A morphological parser for odawa. In: Proceedings of the 2nd Workshop on the Use of Computational Methods in the Study of Endangered Languages, pp. 1–9. Association for Computational Linguistics, Honolulu (2017). DOI 10.18653/v1/W17-0101. URL https://aclanthology.org/W17-0101
* (11) Bühmann, D., Trudell, B.: Mother tongue matters: Local language as a key to effective learning. France: UNESCO (2008)
* (12) Caswell, I., Breiner, T., van Esch, D., Bapna, A.: Language ID in the wild: Unexpected challenges on the path to a thousand-language web text corpus. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 6588–6608. International Committee on Computational Linguistics, Barcelona, Spain (Online) (2020). DOI 10.18653/v1/2020.coling-main.579. URL https://aclanthology.org/2020.coling-main.579
* (13) Chakraborty, D., Gupta, A., Team, G.V., Seth, A.: Experiences from a Mobile-based Behaviour Change Campaign on Maternal and Child Nutrition in Rural India. In: ICTD (2019)
* (14) Chandramouli, C., General, R.: Census of india 2011. Provisional Population Totals. New Delhi: Government of India (2011)
* (15) Dearden, A., Tucker, W.D.: The ethical limits of bungee research in ictd. In: 2015 IEEE International Symposium on Technology and Society (ISTAS), pp. 1–6. IEEE (2015)
* (16) Diddee, H., Bali, K., Choudhury, M., Mukhija, N.: The six conundrums of building and deploying language technologies for social good. In: ACM SIGCAS/SIGCHI Conference on Computing and Sustainable Societies (COMPASS), COMPASS ’22, p. 12–19. Association for Computing Machinery, New York, NY, USA (2022). DOI 10.1145/3530190.3534792. URL https://doi.org/10.1145/3530190.3534792
* (17) Diddee, H., Dandapat, S., Choudhury, M., Ganu, T., Bali, K.: Too brittle to touch: Comparing the stability of quantization and distillation towards developing lightweight low-resource mt models (2022). DOI 10.48550/ARXIV.2210.15184. URL https://arxiv.org/abs/2210.15184
* (18) D’Silva, K., Marathe, M., Vashistha, A., Borriello, G., Thies, W.: A mobile application for interactive voice forums: Design and pilot deployment in rural india. In: Proceedings of the Fifth ACM Symposium on Computing for Development, pp. 121–122 (2014)
* (19) Elazar, Y., Basmov, V., Ravfogel, S., Goldberg, Y., Tsarfaty, R.: The extraordinary failure of complement coercion crowdsourcing. In: Proceedings of the First Workshop on Insights from Negative Results in NLP, pp. 106–116. Association for Computational Linguistics, Online (2020). DOI 10.18653/v1/2020.insights-1.17. URL https://aclanthology.org/2020.insights-1.17
* (20) Ethnologue: https://www.ethnologue.com/world
* (21) Finn, A., Jones, P.L., Mahelona, K., Duncan, S., Leoni, G.: Developing a part-of-speech tagger for te reo Māori. In: Proceedings of the Fifth Workshop on the Use of Computational Methods in the Study of Endangered Languages, pp. 93–98. Association for Computational Linguistics, Dublin, Ireland (2022). DOI 10.18653/v1/2022.computel-1.12. URL https://aclanthology.org/2022.computel-1.12
* (22) Goldman, J.P., Clematide, S., Avanzi, M., Tandler, R.: Strategies and challenges for crowdsourcing regional dialect perception data for Swiss German and Swiss French. In: Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). European Language Resources Association (ELRA), Miyazaki, Japan (2018). URL https://aclanthology.org/L18-1234
* (23) Haugen, E.: Dialect, language, nation 1. American anthropologist 68(4), 922–935 (1966)
* (24) Herbig, N., Pal, S., van Genabith, J., Krüger, A.: Multi-modal approaches for post-editing machine translation. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, CHI ’19, p. 1–11. Association for Computing Machinery, New York, NY, USA (2019). DOI 10.1145/3290605.3300461. URL https://doi.org/10.1145/3290605.3300461
* (25) Hermes, M., King, K.A.: Ojibwe language revitalization, multimedia technology, and family language learning. Language Learning & Technology 17(1), 125–144 (2013)
* (26) Joshi, P., Barnes, C., Santy, S., Khanuja, S., Shah, S., Srinivasan, A., Bhattamishra, S., Sitaram, S., Choudhury, M., Bali, K.: Unsung challenges of building and deploying language technologies for low resource language communities. arXiv preprint arXiv:1912.03457 (2019)
* (27) Khayrallah, H., Koehn, P.: On the impact of various types of noise on neural machine translation. arXiv preprint arXiv:1805.12282 (2018)
* (28) Kornai, A.: Digital language death. PloS one 8(10), e77,056 (2013)
* (29) Kornai, A., Bhattacharyya, P.: Indian subcontinent language vitalization (2014)
* (30) Kreutzer, J., Caswell, I., Wang, L., Wahab, A., van Esch, D., Ulzii-Orshikh, N., Tapo, A., Subramani, N., Sokolov, A., Sikasote, C., Setyawan, M., Sarin, S., Samb, S., Sagot, B., Rivera, C., Rios, A., Papadimitriou, I., Osei, S., Suarez, P.O., Orife, I., Ogueji, K., Rubungo, A.N., Nguyen, T.Q., Müller, M., Müller, A., Muhammad, S.H., Muhammad, N., Mnyakeni, A., Mirzakhalov, J., Matangira, T., Leong, C., Lawson, N., Kudugunta, S., Jernite, Y., Jenny, M., Firat, O., Dossou, B.F.P., Dlamini, S., de Silva, N., Çabuk Ballı, S., Biderman, S., Battisti, A., Baruwa, A., Bapna, A., Baljekar, P., Azime, I.A., Awokoya, A., Ataman, D., Ahia, O., Ahia, O., Agrawal, S., Adeyemi, M.: Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets. Transactions of the Association for Computational Linguistics 10, 50–72 (2022). DOI 10.1162/tacl˙a˙00447. URL https://doi.org/10.1162/tacl_a_00447
* (31) Kumar, S.R.: Gondi language - identity, politics and struggle in india. IOSR Journal 24(2), 51–57 (2019)
* (32) Kummerfeld, J.K.: Quantifying and avoiding unfair qualification labour in crowdsourcing. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 343–349. Association for Computational Linguistics, Online (2021). DOI 10.18653/v1/2021.acl-short.44. URL https://aclanthology.org/2021.acl-short.44
* (33) Lambton-Howard, D., Olivier, P., Vlachokyriakos, V., Celina, H., Kharrufa, A.: Unplatformed design: A model for appropriating social media technologies for coordinated participation. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–13 (2020)
* (34) Lewis, M., Simons, G.: Sustaining language use: Perspectives on community-based language development. (2016)
* (35) Liebling, D.J., Lahav, M., Evans, A., Donsbach, A., Holbrook, J., Smus, B., Boran, L.: Unmet needs and opportunities for mobile translation ai. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, CHI ’20, p. 1–13. Association for Computing Machinery, New York, NY, USA (2020). DOI 10.1145/3313831.3376261. URL https://doi.org/10.1145/3313831.3376261
* (36) Little, A.N.: Connecting documentation and revitalization: A new approach to language apps. In: Proceedings of the 2nd Workshop on the Use of Computational Methods in the Study of Endangered Languages, pp. 151–155. Association for Computational Linguistics, Honolulu (2017). DOI 10.18653/v1/W17-0120. URL https://aclanthology.org/W17-0120
* (37) Liu, H., Thekinen, J., Mollaoglu, S., Tang, D., Yang, J., Cheng, Y., Liu, H., Tang, J.: Toward annotator group bias in crowdsourcing. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1797–1806. Association for Computational Linguistics, Dublin, Ireland (2022). DOI 10.18653/v1/2022.acl-long.126. URL https://aclanthology.org/2022.acl-long.126
* (38) Marathe, M., O’Neill, J., Pain, P., Thies, W.: Revisiting CGNet Swara and its Impact in Rural India. In: ICTD (2015)
* (39) Mehta, D., Mothilal, R.K., Sharma, A., Thies, W., Sharma, A.: Using mobile airtime credits to incentivize learning, sharing and survey response: Experiences from the field. In: Proceedings of the 3rd ACM SIGCAS Conference on Computing and Sustainable Societies, pp. 254–264 (2020)
* (40) Mehta, D., Sharma, A., Kommiya Mothilal, R.K., Shukla, A., Prasad, V., Thies, W., U, V., Scott, C., Sharma, A.: Facilitating media distribution with monetary incentives. In: Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–7 (2020)
* (41) Millour, A., Fort, K.: Text corpora and the challenge of newly written languages. In: Proceedings of the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL), pp. 111–120. European Language Resources association, Marseille, France (2020). URL https://aclanthology.org/2020.sltu-1.15
* (42) Moitra, A., Das, V., Kumar, A., Seth, A.: Design Lessons from Creating a Mobile-based Community Media Platform in Rural India. In: ICTD (2016)
* (43) Moseley, C.: Atlas of the World’s Languages in Danger. Unesco (2010)
* (44) Mus, N., Metzger, R.: Toward a corpus of tundra nenets: stages and challenges in building a corpus. In: Proceedings of the Workshop on Computational Methods for Endangered Languages, vol. 2, pp. 4–9 (2021)
* (45) Nangia, N., Sugawara, S., Trivedi, H., Warstadt, A., Vania, C., Bowman, S.R.: What ingredients make for an effective crowdsourcing protocol for difficult nlu data collection tasks? arXiv preprint arXiv:2106.00794 (2021)
* (46) Patel, N., Chittamuru, D., Jain, A., Dave, P., Parikh, T.S.: Avaaj otalo - a field study of an interactive voice forum for small farmers in rural india. In: CHI (2010)
* (47) Raza, A.A., Saleem, B., Randhawa, S., Tariq, Z., Athar, A., Saif, U., Rosenfeld, R.: Baang: a viral speech-based social platform for under-connected populations. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, p. 643. ACM (2018)
* (48) Raza, A.A., Tariq, Z., Randhawa, S., Saleem, B., Athar, A., Saif, U., Rosenfeld, R.: Voice-based quizzes for measuring knowledge retention in under-connected populations. In: CHI (2019)
* (49) Raza, A.A., Ul Haq, F., Tariq, Z., Pervaiz, M., Razaq, S., Saif, U., Rosenfeld, R.: Job opportunities through entertainment: Virally spread speech-based services for low-literate users. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 2803–2812. ACM (2013)
* (50) Rijhwani, S., Anastasopoulos, A., Neubig, G.: Ocr post correction for endangered language texts. arXiv preprint arXiv:2011.05402 (2020)
* (51) Roche, G.: Abandoning endangered languages: Ethical loneliness, language oppression, and social justice. American Anthropologist 122(1), 164–169 (2020)
* (52) Saldivar, J., Daniel, F., Cernuzzi, L., Casati, F.: Online idea management for civic engagement: A study on the benefits of integration with social networking. ACM Transactions on Social Computing 2(1), 1–29 (2019)
* (53) Shapiro, M.C., Schiffman, H.F.: Language and society in South Asia, vol. 1. Motilal Banarsidass Publishe (1981)
* (54) Shmueli, B., Fell, J., Ray, S., Ku, L.W.: Beyond fair pay: Ethical implications of nlp crowdsourcing. arXiv preprint arXiv:2104.10097 (2021)
* (55) Specia, L., Blain, F., Fomicheva, M., Fonseca, E., Chaudhary, V., Guzmán, F., Martins, A.F.T.: Findings of the WMT 2020 shared task on quality estimation. In: Proceedings of the Fifth Conference on Machine Translation, pp. 743–764. Association for Computational Linguistics, Online (2020). URL https://aclanthology.org/2020.wmt-1.79
* (56) Starbird, K.: Crowdwork, crisis and convergence: How the connected crowd organizes information during mass disruption events. Ph.D. thesis, University of Colorado at Boulder (2012)
* (57) Stebbins, T.N., Eira, K., Couzens, V.L.: Living languages and new approaches to language revitalisation research. Routledge (2017)
* (58) Suh, H., Shahriaree, N., Hekler, E.B., Kientz, J.A.: Developing and validating the user burden scale: A tool for assessing user burden in computing systems. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, CHI ’16, p. 3988–3999. Association for Computing Machinery, New York, NY, USA (2016). DOI 10.1145/2858036.2858448. URL https://doi.org/10.1145/2858036.2858448
* (59) Swaminathan, S., Medhi Thies, I., Mehta, D., Cutrell, E., Sharma, A., Thies, W.: Learn2earn: Using mobile airtime incentives to bolster public awareness campaigns. Proceedings of the ACM on Human-Computer Interaction 3(CSCW), 49 (2019)
* (60) Tyler, S.A.: Koya: an outline grammar: Gommu dialect, vol. 54. Berkeley: University of California Press (1969)
* (61) Vashistha, A., Cutrell, E., Borriello, G., Thies, W.: Sangeet Swara: A Community-Moderated Voice Forum in Rural India. In: CHI (2015)
* (62) Walsh, M.: Will indigenous languages survive? Annu. Rev. Anthropol. 34, 293–315 (2005)
* (63) Wang, X., Ruder, S., Neubig, G.: Expanding pretrained models to thousands more languages via lexicon-based adaptation. In: ACL (2022)
* (64) Xue, L., Constant, N., Roberts, A., Kale, M., Al-Rfou, R., Siddhant, A., Barua, A., Raffel, C.: mt5: A massively multilingual pre-trained text-to-text transformer. arXiv preprint arXiv:2010.11934 (2020)
* (65) Xue, L., Constant, N., Roberts, A., Kale, M., Al-Rfou, R., Siddhant, A., Barua, A., Raffel, C.: mT5: A massively multilingual pre-trained text-to-text transformer. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 483–498. Association for Computational Linguistics, Online (2021). DOI 10.18653/v1/2021.naacl-main.41. URL https://aclanthology.org/2021.naacl-main.41
* (66) Zariquiey, R., Oncevay, A., Vera, J.: CLD2 language documentation meets natural language processing for revitalising endangered languages. In: Proceedings of the Fifth Workshop on the Use of Computational Methods in the Study of Endangered Languages, pp. 20–30. Association for Computational Linguistics, Dublin, Ireland (2022). DOI 10.18653/v1/2022.computel-1.4. URL https://aclanthology.org/2022.computel-1.4
|
# Extremal numbers of disjoint triangles in $r$-partite graphs
Junxue Zhang111Corresponding author
Center for Combinatorics and LPMC, Nankai University, Tianjin 300071, China
Email<EMAIL_ADDRESS>
###### Abstract
For two graphs $G$ and $F$, the extremal number of $F$ in $G$, denoted by
ex$(G,F)$, is the maximum number of edges in a spanning subgraph of $G$ not
containing $F$ as a subgraph. Determining ex$(K_{n},F)$ for a given graph $F$
is a classical extremal problem in graph theory. In 1962, Erdős determined
ex$(K_{n},kK_{3})$, which generalized Mantel’s Theorem. On the other hand, in
1974, Bollobás, Erdős, and Straus determined
ex$(K_{n_{1},n_{2},\dots,n_{r}},K_{t})$, which extended Turán’s Theorem to
complete multipartite graphs. In this paper, we determine
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{3})$ for $r\geq 4$ and $10k-4\leq
n_{1}+4k\leq n_{2}\leq n_{3}\leq\cdots\leq n_{r}$.
Keywords: extremal numbers; multipartite graphs; disjoint triangles
## 1 Introduction
All graphs considered in this paper are finite and simple. We follow [12] for
undefined notation and terminology. Let $G=G(V(G),E(G))$ be a graph, where
$V(G)$ is the vertex set and $E(G)$ is the edge set. Denote $e(G)=|E(G)|$. For
any subset $S\subseteq V(G)$, we use $G[S]$ to denote the subgraph of $G$
induced by $S$ and let $G-S=G[V(G)\setminus S]$. Given two disjoint vertex
sets $V_{1}$ and $V_{2}$, the join $V_{1}\vee V_{2}$ is the edge set obtained
by joining each vertex of $V_{1}$ to each vertex of $V_{2}$. Furthermore,
given two graphs $G$ and $H$, the join $G\vee H$ is the graph obtained from
the disjoint union of $G$ and $H$ by joining each vertex of $G$ to each vertex
of $H$. Let $r,t,k$ be three positive integers. For convenience, we write
$[r]=\\{1,2,3,...,r\\}$ in the context. Denote by $K_{t}$ the complete graph
on $t$ vertices and $kK_{t}$ the disjoint union of $k$ copies of $K_{t}$.
Moreover, for $r$ positive integers $n_{1},n_{2},\dots,n_{r}$, we use
$K_{n_{1},n_{2},\dots,n_{r}}$ to denote the complete $r$-partite graph with
parts of sizes $n_{1},n_{2},\dots,n_{r}$.
Let ex$(G,H)=\max\\{e(G^{\prime}):$ $|V(G^{\prime})|=|V(G)|$,
$G^{\prime}\subseteq G$, and $H\nsubseteq G^{\prime}\\}$ and call it the
extremal number of $H$ in $G$. Let $n$ and $t$ be two integers with $n\geq t$.
The classical Turán problem considers the case $G=K_{n}$, i.e., determining
the value of ex$(K_{n},H)$ for a given $H$. For instance, the well-known
Mantel’s Theorem [8] and Turán’s Theorem [11] determined ex$(K_{n},K_{3})$ and
ex$(K_{n},K_{t})$, respectively. Turán’s Theorem is widely considered to be
the first extremal theorem on graphs and initiated the study of extremal graph
theory. Let $T_{r}(n)$ be a balanced complete $r$-partite graph on $n$
vertices, that is, any two parts of sizes differ by at most one. In 1959,
Erdős and Gallai [6] determined ex$(K_{n},kK_{2})$ for any positive integers
$n$ and $k$. Later, Erdős [5] proved ex$(K_{n},kK_{3})=e(K_{k-1}\vee
T_{2}(n-k+1))$ for $n>400(k-1)^{2}$ and Moon [9] proved that for
$n>\frac{9k}{2}+4$. Moreover, Moon [9] (only when $n-k+1$ is divisible by $p$)
and Simonovits [10] showed that $K_{k-1}\vee T_{r-1}(n-k+1)$ is the unique
extremal graph containing no copy of $kK_{r}$ for sufficiently large $n$. The
extremal problem on multipartite graphs was first considered by Bollobás,
Erdős, and Straus [1]. They determined ex$(K_{n_{1},n_{2},\dots,n_{r}},K_{t})$
for $r\geq t\geq 2$. Later, Chen, Li, and Tu [2] determined
ex$(K_{n_{1},n_{2}},kK_{2})$. Recently, De Silva, Heysse and Young [4] proved
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{2})=(k-1)(n_{2}+n_{3}+\cdots+n_{r})$ for
$n_{1}\leq n_{2}\leq\dots\leq n_{r}$.
To make it easier to state the results in this paper, we define a function
$f$. Let $x_{1},x_{2},\dots,x_{r}$ be positive integers and $t\geq 2$ be an
integer. Note that $\\{x_{1},x_{2},\dots,x_{r}\\}$ is a multiset. For any
subset $P\subseteq[r]$, let $x_{P}:=\sum_{i\in P}x_{i}$. Now, we define the
desired function $f$ as follows. For $t=2$, let
$f_{2}(\\{x_{1},x_{2},\dots,x_{r}\\})=0$. For $t\geq 3$, let
$f_{t}(\\{x_{1},x_{2},\dots,x_{r}\\})=\max\limits_{\mathcal{P}}~{}\\{\sum\limits_{1\leq
i<j\leq t-1}x_{P_{i}}\cdot x_{P_{j}}\\}$, where
$\mathcal{P}=(P_{1},P_{2},\dots,P_{t-1})$ is a partition of $[r]$ into $t-1$
nonempty parts. For convenience, we simply write
$f_{t}(x_{1},x_{2},\dots,x_{r})$ for $f_{t}(\\{x_{1},x_{2},\dots,x_{r}\\})$.
Notice that $f_{r}(x_{1},x_{2},\dots,x_{r})=(\sum_{1\leq i<j\leq
r}x_{i}x_{j})-x_{1}x_{2}$ if $x_{1}\leq x_{2}\leq\dots\leq x_{r}$.
###### Theorem 1.1
[1] Let $n_{1},n_{2},\dots,n_{r},t$ be positive integers with $r\geq t\geq 2$.
Then
ex$(K_{n_{1},n_{2},\dots,n_{r}},K_{t})=f_{t}(n_{1},n_{2},\dots,n_{r})$.
By replacing the forbidden graph $K_{t}$ with $kK_{t}$, De Silva et al. [3]
considered a special case $t=r$ and raised the problem about
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})$ for $r>t$.
###### Theorem 1.2
[3] For any integers $r\geq 2$ and $1\leq k\leq n_{1}\leq n_{2}\leq\cdots\leq
n_{r}$, we have
$\displaystyle{ex}(K_{n_{1},n_{2},\dots,n_{r}},kK_{r})$
$\displaystyle=\sum_{1\leq i<j\leq r}n_{i}n_{j}-n_{1}n_{2}+(k-1)n_{2}$
$\displaystyle{=}(k-1)(n-n_{1})+{f_{r}(n_{1}-(k-1),n_{2},\dots,n_{r}),}$
where $n=\sum n_{i}$.
###### Problem 1.3
[3] Determine ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})$ for $r>t$.
Han and Zhao [7] determined ex$(K_{n_{1},n_{2},n_{3},n_{4}},kK_{3})$ for a
sufficiently large integer $n_{1}$ with $n_{1}\leq n_{2}\leq n_{3}\leq n_{4}$.
###### Theorem 1.4
[7] Let $n_{1},\dots,n_{4}$ be sufficiently large and $n_{1}\leq n_{2}\leq
n_{3}\leq n_{4}$. For any integer $k\geq 1$, we have
$\displaystyle ex(K_{n_{1},n_{2},n_{3},n_{4}},kK_{3})$
$\displaystyle=(k-1)(n-n_{1})+f_{3}(n_{1}-{(k-1)},n_{2},n_{3},n_{4})$
$\displaystyle=\left\\{\begin{array}[]{lcl}n_{4}(n_{1}+n_{2}+n_{3})+(k-1)(n_{2}+n_{3}),&&{\text{if}~{}n_{4}>n_{2}+n_{3}};\\\
(n_{1}+n_{4})(n_{2}+n_{3})+(k-1)n_{4},&&{\text{if}~{}n_{4}\leq
n_{2}+n_{3}}.\\\ \end{array}\right.$
In addition, Han and Zhao [7] proposed a conjecture on
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})$.
###### Conjecture 1.5
[7] Given three integers $k$, $r$, and $t$ with $r>t\geq 3$ and $k\geq 2$, let
$n_{1},\dots,n_{r}$ be sufficiently large. For $I\subseteq[r]$, write
$m_{I}:={\min}_{i\in I}n_{i}$. Given a partition $\mathcal{P}$ of $[r]$, let
$n_{\mathcal{P}}:={\max}_{I\in\mathcal{P}}\\{n_{I}-m_{I}\\}$. Then
$\displaystyle
ex(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})=\mathop{max}\limits_{\mathcal{P}}\left\\{(k-1)n_{\mathcal{P}}+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}}}n_{I}\cdot
n_{I^{\prime}}\right\\},$
where the maximum is taken over all partitions $\mathcal{P}$ of $[r]$ into
$t-1$ parts.
Actually, we can obtain an equivalent statement of Conjecture 1.5 as follows,
equivalence will be proved in Section 2.
###### Conjecture 1.6
Given three integers $k$, $r$, and $t$ with $r>t\geq 3$ and $k\geq 2$, let
$n_{1}\leq n_{2}\dots\leq n_{r}$ be integers with $\sum_{i=1}^{r}n_{i}=n$. If
$n_{1}$ is sufficiently large, then
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})=(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r})$.
To support Conjecture 1.6, we provide a simple construction, which gives a
lower bound for ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t}).$
###### Theorem 1.7
Let $k$ and $t$ be two integers with $k\geq 2$ and $t\geq 3$. Let
$n_{1},n_{2},\dots,n_{r}$ be $r$ integers with $n_{1}\leq\min_{i\geq
2}\\{n_{i}\\}$, $n_{1}\geq k$, $r\geq t$, and $\sum_{i=1}^{r}n_{i}=n$. Then
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})\geq(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r})$.
Furthermore, with Theorem 1.7, we confirm Conjecture 1.6 for $t=3$, $r\geq 4$,
and $n_{1}+4k\leq n_{2}$.
###### Theorem 1.8
Let $r,n_{1},n_{2},\dots,n_{r}$ be integers with $r\geq 4$, $10k-4\leq
n_{1}+4k\leq\min_{i\geq 2}\\{n_{i}\\}$ and $\sum_{i=1}^{r}n_{i}=n$. Then
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{3})=(k-1)(n-n_{1})+f_{3}(n_{1}-(k-1),n_{2},\dots,n_{r})$.
In Conjecture 1.6, it is required that “$n_{1}$ is sufficiently large”. But
how large it can be? By aforementioned results and Theorem 1.8, we conjecture
that the lower bound for $n_{1}$ is $k$.
###### Conjecture 1.9
Given three integers $k$, $r$, and $t$ with $r\geq t\geq 2$ and $k\geq 1$, let
$n_{1},n_{2},\dots,n_{r}$ be integers with $n_{1}\leq\min_{i\geq
2}\\{n_{i}\\}$ and $\sum_{i=1}^{r}n_{i}=n$. If $n_{1}\geq k$, then
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})=(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r})$.
The rest of this paper is organized as follows. In Section 2, we first prove
the equivalence of Conjectures 1.5 and 1.6. Later, some basic properties of
$f_{3}(n_{1},n_{2},\dots,n_{r})$ are provided, which will be used frequently
in the proof of Theorem 1.8. The proofs of Theorems 1.7 and 1.8 are presented
in Sections 3 and 4, respectively.
## 2 Equivalence and Properties of $f_{3}(n_{1},n_{2},\dots,n_{r})$
###### Proposition 2.1
Conjecture 1.5 and Conjecture 1.6 are equivalent.
Proof. Let $r>t\geq 3$, $k\geq 2$, $n_{1}\leq n_{2}\dots\leq n_{r}$. Our goal
is to prove
$\displaystyle\mathop{max}\limits_{\mathcal{P}}\left\\{(k-1)n_{\mathcal{P}}+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}}}n_{I}\cdot
n_{I^{\prime}}\right\\}=(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r}){.}$
(1)
First we show the $\geq$ direction of (1). Let $x_{1}=n_{1}-(k-1)$,
$x_{i}=n_{i}$ for any $i\geq 2$, and
$\mathcal{P}_{0}=(P_{1},P_{2},\dots,P_{t-1})$ be a partition of $[r]$
maximizing $f_{t}(x_{1},x_{2},\dots,x_{r})$. Assume $1\in P_{1}$. Hence
$\displaystyle(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r})$
$\displaystyle=(k-1)(n_{P_{1}}-n_{1})+(k-1)(n-n_{P_{1}})+\sum_{i\neq
j\in[t-1]}x_{P_{i}}\cdot x_{P_{j}}$
$\displaystyle\leq(k-1)n_{\mathcal{P}_{0}}+\sum_{i\neq
j\in[t-1]}n_{P_{i}}\cdot n_{P_{j}}$
$\displaystyle\leq\mathop{max}\limits_{\mathcal{P}}\left\\{(k-1)n_{\mathcal{P}}+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}}}n_{I}\cdot
n_{I^{\prime}}\right\\}.$
Now we prove the $\leq$ direction of (1). For any given partition
$\mathcal{P}=(P_{1},P_{2},\dots,P_{t-1})$ of $[r]$, let $\ell\in[t-1]$ such
that $n_{\mathcal{P}}=n_{P_{\ell}}-m_{P_{\ell}}$ and $\alpha\in P_{\ell}$ such
that $n_{\alpha}=m_{P_{\ell}}$.
Notice that
$\displaystyle(k-1)n_{\mathcal{P}}+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}}}n_{I}\cdot
n_{I^{\prime}}$
$\displaystyle=(k-1)(n_{P_{\ell}}-m_{P_{\ell}})+n_{P_{\ell}}\cdot(n-n_{P_{\ell}})+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}\setminus\\{P_{\ell}\\}}}n_{I}\cdot
n_{I^{\prime}}$
$\displaystyle=(k-1)(n-m_{P_{\ell}})+(n_{P_{\ell}}-(k-1))\cdot(n-n_{P_{\ell}})+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}\setminus\\{P_{\ell}\\}}}n_{I}\cdot
n_{I^{\prime}}$
$\displaystyle\leq(k-1)(n-n_{\alpha})+f_{t}(\\{n_{1},\cdots,n_{r}\\}\cup\\{n_{\alpha}-(k-1)\\}\setminus\\{n_{\alpha}\\}).$
Next we assume that $\mathcal{P}$ be the partition maximizing the value
$(k-1)n_{\mathcal{P}}+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}}}n_{I}\cdot
n_{I^{\prime}}$. It remains to show
$\displaystyle(k-1)(n-n_{\alpha})+f_{t}(\\{n_{1},\cdots,n_{r}\\}\cup\\{n_{\alpha}-(k-1)\\}\setminus\\{n_{\alpha}\\})$
$\displaystyle\leq(k-1)(n-n_{1})+f_{t}(\\{n_{1},\cdots,n_{r}\\}\cup\\{n_{1}-(k-1)\\}\setminus\\{n_{1}\\}).$
(2)
If $n_{\alpha}=n_{1}$, then we are done. We may assume that
$n_{\alpha}>n_{1}$. We can see $1\notin P_{\ell}$. We may assume $1\in{P}_{1}$
and $\ell\neq 1$. Now note that $m_{P_{1}}=n_{1}$ and
$m_{P_{\ell}}=n_{\alpha}$. Then $n_{P_{\ell}}-n_{\alpha}>n_{P_{1}}-n_{1}$ by
$\ell\neq 1$. Since $n_{\alpha}>n_{1}$ and
$n_{P_{\ell}}-n_{\alpha}>n_{P_{1}}-n_{1}$, we can switch $1$ and $\alpha$ in
the partition $\mathcal{P}$ and thus increases
$(k-1)n_{\mathcal{P}}+\sum_{I\neq{I^{\prime}}\in{\mathcal{P}}}n_{I}\cdot
n_{I^{\prime}}$, a contradiction. Thus the Inequation (2) holds.
Next we provided some properties of the function
$f_{3}(n_{1},n_{2},\dots,n_{r})$.
###### Proposition 2.2
Let $n_{1},n_{2},\dots,n_{r},\mu$ be positive integers with $\mu\in[r]$ and
$n_{\mu}-1\geq n_{1}+1$. Then
$f_{3}(\\{n_{1},n_{2},\dots,n_{r}\\}\setminus\\{n_{1},n_{\mu}\\}\cup\\{n_{1}+1,n_{\mu}-1\\})\leq
f_{3}(n_{1},n_{2},\dots,n_{r})+n_{\mu}-(n_{1}+1)$.
Proof. Let $x_{1}=n_{1}+1$, $x_{\mu}=n_{\mu}-1$, and $x_{j}=n_{j}$ for any
$j\in[r]\setminus\\{1,\mu\\}$. Then
$f_{3}(\\{n_{1},n_{2},\dots,n_{r}\\}\setminus\\{n_{1},n_{\mu}\\}\cup\\{n_{1}+1,n_{\mu}-1\\})=f_{3}(x_{1},x_{2},\dots,x_{r})$.
Let $\mathcal{P}=(P_{1},P_{2})$ be the partition of $[r]$ maximizing
$f_{3}(x_{1},x_{2},...,x_{r})$. Note that
$\sum_{j=1}^{r}x_{j}=\sum_{j=1}^{r}n_{j}$. If $\\{1,\mu\\}\subseteq P_{1}$ or
$\\{1,\mu\\}\subseteq P_{2}$, then $x_{P_{1}}=n_{P_{1}}$ and
$x_{P_{2}}=n_{P_{2}}$. Thus, $f_{3}(x_{1},x_{2},\dots,x_{r})=x_{P_{1}}\cdot
x_{P_{2}}=n_{P_{1}}\cdot n_{P_{2}}\leq f_{3}(n_{1},n_{2},\dots,n_{r})$. Hence,
without loss of generality, we may assume that $1\in P_{1}$ and $\mu\in
P_{2}$. Observe that $x_{P_{1}}=n_{P_{1}}+1$, $x_{P_{2}}=n_{P_{2}}-1$,
$x_{P_{1}\setminus\\{1\\}}=n_{P_{1}\setminus\\{1\\}}$, and
$x_{P_{2}\setminus\\{\mu\\}}=n_{P_{2}\setminus\\{\mu\\}}$. If
$x_{P_{1}\setminus\\{1\\}}\geq x_{P_{2}\setminus\\{\mu\\}}$, then
$n_{P_{1}\setminus\\{1\\}}\geq n_{P_{2}\setminus\\{\mu\\}}$, and we have
$\displaystyle f_{3}(x_{1},x_{2},\dots,x_{r})=$ $\displaystyle x_{P_{1}}\cdot
x_{P_{2}}$ $\displaystyle=$ $\displaystyle(n_{P_{1}}+1)\cdot(n_{P_{2}}-1)$
$\displaystyle=$ $\displaystyle n_{P_{1}}\cdot
n_{P_{2}}-n_{P_{1}}+n_{P_{2}}-1$ $\displaystyle\leq$ $\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})-(n_{P_{1}\setminus\\{1\\}}+n_{1})+(n_{P_{2}\setminus\\{\mu\\}}+n_{\mu})-1$
$\displaystyle\leq$ $\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})+n_{\mu}-(n_{1}+1).$
Now $x_{P_{1}\setminus\\{1\\}}<x_{P_{2}\setminus\\{\mu\\}}$, equivalently,
$n_{P_{1}\setminus\\{1\\}}<n_{P_{2}\setminus\\{\mu\\}}$. Recall that
$x_{1}=n_{1}+1\leq n_{\mu}-1=x_{\mu}$. If $n_{1}+1<n_{\mu}-1$, then
$|(x_{P_{1}\setminus\\{1\\}}+x_{\mu})-(x_{P_{2}\setminus\\{\mu\\}}+x_{1})|<|(x_{P_{1}\setminus\\{1\\}}+x_{1})-(x_{P_{2}\setminus\\{\mu\\}}+x_{\mu})|=|x_{P_{1}}-x_{P_{2}}|$,
which contradicts the choice of $\mathcal{P}$ that minimizes
$|x_{P_{1}}-x_{P_{2}}|$, i.e., maximizes $x_{P_{1}}\cdot x_{P_{2}}$. Thus, we
obtain $n_{1}+1=n_{\mu}-1$. This implies $x_{P_{1}}\cdot
x_{P_{2}}=(x_{P_{1}\setminus\\{1\\}}+x_{\mu})\cdot(x_{P_{2}\setminus\\{\mu\\}}+x_{1})$.
Then,
$\displaystyle f_{3}(x_{1},x_{2},\dots,x_{r})=$ $\displaystyle x_{P_{1}}\cdot
x_{P_{2}}$ $\displaystyle=$
$\displaystyle(x_{P_{1}\setminus\\{1\\}}+x_{\mu})\cdot(x_{P_{2}\setminus\\{\mu\\}}+x_{1})$
$\displaystyle=$
$\displaystyle(n_{P_{1}\setminus\\{1\\}}+n_{\mu}-1)\cdot(n_{P_{2}\setminus\\{\mu\\}}+n_{1}+1)$
$\displaystyle=$
$\displaystyle(n_{P_{1}\setminus\\{1\\}}+n_{\mu})\cdot(n_{P_{2}\setminus\\{\mu\\}}+n_{1})+(n_{P_{1}\setminus\\{1\\}}+n_{\mu})-(n_{P_{2}\setminus\\{\mu\\}}+n_{1})-1$
$\displaystyle<$ $\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})+n_{\mu}-(n_{1}+1).$
###### Proposition 2.3
Let $n_{1},n_{2},\dots,n_{r}$ be $r$ integers with $n_{1}+2\leq\min_{i\geq
2}\\{n_{i}\\}$. For two indices $i,j\in[r]$ with $i\neq j$, let
$\mathcal{P}=(P_{1},P_{2})$ be the partition attaining
$f_{3}(\\{n_{1},\dots,n_{r}\\}\setminus\\{n_{i},n_{j}\\}\cup\\{n_{i}-1,n_{j}-1\\})$.
Then
$\displaystyle
f_{3}(\\{n_{1},\dots,n_{r}\\}\setminus\\{n_{i},n_{j}\\}\cup\\{n_{i}-1,n_{j}-1\\})\leq$
$\displaystyle f_{3}(n_{1},n_{2},\dots,n_{r})-\sum_{m=1}^{r}n_{m}$
$\displaystyle+max\\{n_{1}+2,n_{i}-n_{1}+1\\}$
Proof. Let $x_{i}=n_{i}-1$, $x_{j}=n_{j}-1$, and $x_{\ell}=n_{\ell}$ for any
$\ell\in[r]\setminus\\{i,j\\}$. Then we know
$f_{3}(\\{n_{1},\dots,n_{r}\\}\setminus\\{n_{i},n_{j}\\}\cup\\{n_{i}-1,n_{j}-1\\})=f_{3}(x_{1},x_{2},\dots,x_{r})$
and the partition $\mathcal{P}$ attains $f_{3}(x_{1},x_{2},\dots,x_{r})$.
Moreover, for any $\ell\in[r]\setminus\\{1\\}$, we have $x_{\ell}\geq
n_{\ell}-1>n_{1}\geq x_{1}$ since $n_{\ell}\geq n_{1}+2$.
If $\\{i,j\\}\nsubseteq P_{\zeta}$ for any $\zeta\in[2]$, we have
$x_{P_{\zeta}}=n_{P_{\zeta}}-1$ and $x_{P_{3-\zeta}}=n_{P_{3-\zeta}}-1$. Thus
$\displaystyle f_{3}(x_{1},x_{2},\dots,x_{r})=$ $\displaystyle
x_{P_{\zeta}}\cdot x_{P_{3-\zeta}}=(n_{P_{\zeta}}-1)\cdot(n_{P_{3-\zeta}}-1)$
$\displaystyle=$ $\displaystyle n_{P_{\zeta}}\cdot
n_{P_{3-\zeta}}-n_{P_{\zeta}}-n_{P_{3-\zeta}}+1$ $\displaystyle\leq$
$\displaystyle f_{3}(n_{1},n_{2},\dots,n_{r})+1-\sum_{m=1}^{r}n_{m}.$
Assume $\\{i,j\\}\subseteq P_{\zeta}$ for some $\zeta\in[2]$. Notice that
$x_{P_{\zeta}}=n_{P_{\zeta}}-2$ and $x_{P_{3-\zeta}}=n_{P_{3-\zeta}}$. Thus,
$\displaystyle f_{3}(x_{1},x_{2},\dots,x_{r})=$ $\displaystyle
x_{P_{\zeta}}\cdot x_{P_{3-\zeta}}=(n_{P_{\zeta}}-2)\cdot n_{P_{3-\zeta}}$
$\displaystyle=$ $\displaystyle n_{P_{\zeta}}\cdot
n_{P_{3-\zeta}}-2n_{P_{3-\zeta}}$ $\displaystyle\leq$ $\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})-2n_{P_{3-\zeta}}.$ (3)
If $1\in{P_{\zeta}}$, we have $x_{P_{3-\zeta}}\geq x_{P_{\zeta}}-x_{1}$.
Otherwise $x_{P_{\zeta}}\cdot x_{P_{3-\zeta}}<x_{P_{\zeta}}\cdot
x_{P_{3-\zeta}}+x_{1}\cdot(x_{P_{\zeta}}-x_{1}-x_{P_{3-\zeta}})=(x_{P_{\zeta}}-x_{1})\cdot(x_{P_{3-\zeta}}+x_{1})$,
which contradicts to the choice of $\mathcal{P}$. With
$x_{P_{\zeta}}+x_{P_{3-\zeta}}=x_{\mathcal{P}}$, we know
$n_{P_{3-\zeta}}=x_{P_{3-\zeta}}\geq\frac{x_{\mathcal{P}}-x_{1}}{2}\geq\frac{\sum_{m=1}^{r}n_{m}-2-n_{1}}{2}.$
Thus, by the inequality (2),
$\displaystyle f_{3}(x_{1},x_{2},\dots,x_{r})\leq$ $\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})-2n_{P_{3-\zeta}}$ $\displaystyle\leq$
$\displaystyle f_{3}(n_{1},n_{2},\dots,n_{r})+n_{1}+2-\sum_{m=1}^{r}n_{m}.$
It remains to consider the case $1\in{P_{3-\zeta}}$. Recall $x_{i}>x_{1}$ and
$i\in{P_{\zeta}}$, we have $x_{P_{3-\zeta}}-x_{1}\geq x_{P_{\zeta}}-x_{i}$.
Otherwise $x_{P_{\zeta}}\cdot x_{P_{3-\zeta}}<x_{P_{\zeta}}\cdot
x_{P_{3-\zeta}}+(x_{i}-x_{1})\cdot(x_{P_{\zeta}}-x_{i}-x_{P_{3-\zeta}}+x_{1})=(x_{P_{\zeta}}-x_{i}+x_{1})\cdot(x_{P_{3-\zeta}}-x_{1}+x_{i})$,
a contradiction to the choice of $\mathcal{P}$ again. With
$x_{P_{\zeta}}+x_{P_{3-\zeta}}=x_{\mathcal{P}}$, we know
$n_{P_{3-\zeta}}=x_{P_{3-\zeta}}\geq\frac{x_{\mathcal{P}}+x_{1}-x_{i}}{2}=\frac{\sum_{m=1}^{r}n_{m}-1+n_{1}-n_{i}}{2}.$
Thus, by the inequality (2),
$\displaystyle f_{3}(x_{1},x_{2},\dots,x_{r})\leq$ $\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})-2n_{P_{3-\zeta}}$ $\displaystyle\leq$
$\displaystyle
f_{3}(n_{1},n_{2},\dots,n_{r})-n_{1}+n_{i}+1-\sum_{m=1}^{r}n_{m}.$
## 3 Proof of Theorem 1.7
We shall provide a construction to complete the proof. Let $x_{1}=n_{1}-(k-1)$
and $x_{i}=n_{i}$ for any $i\neq 1$. Then
$f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r})=f_{t}(x_{1},x_{2},\dots,x_{r})$. Let
$\mathcal{P}=(P_{1},P_{2},\ldots,P_{t-1})$ be the partition of $[r]$ that
attains $f_{t}(x_{1},x_{2},\dots,x_{r})$. Assume that $V_{0}$ is a set of
$(k-1)$ vertices and $V_{i}$ is a set of $x_{i}$ vertices for any $i\in[r]$
such that $V_{0}$, $V_{1}$, …, $V_{r}$ are pairwise disjoint. According to the
partition $\mathcal{P}$ of $[r]$, we set $V_{P_{\ell}}=\cup_{i\in
P_{\ell}}V_{i}$ for each $\ell\in[t-1]$. So
$V_{[r]}=\cup_{\ell\in[t-1]}V_{P_{\ell}}$ and $|V_{P_{\ell}}|=x_{P_{\ell}}$.
Now we construct the graph $G$ as follows. Let $V(G)=V_{0}\cup V_{[r]}$ and
$E(G)=\\{V_{0}\vee(\cup_{i=2}^{r}V_{i})\\}\cup\\{\cup_{1\leq\ell<j\leq
t-1}(V_{P_{\ell}}\vee V_{P_{j}})\\}.$
We see $|V(G)|=(k-1)+\sum_{i=1}^{r}x_{i}=\sum_{i=1}^{r}n_{i}=n$, and
$\displaystyle|E(G)|=$ $\displaystyle(k-1)(n-n_{1})+\sum_{1\leq\ell<j\leq
t-1}x_{P_{\ell}}\cdot x_{P_{j}}$ $\displaystyle=$
$\displaystyle(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r}).$
We can see $G$ is a subgraph of $K_{n_{1},n_{2},\dots,n_{r}}$ and $G-V_{0}$ is
a complete $(t-1)$-partite graph. So any copy of $K_{t}$ in $G$ must contain
at least one vertex of $V_{0}$. Since $|V_{0}|=k-1$, $G$ contains no copy of
$kK_{t}$. Thus
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{t})\geq(k-1)(n-n_{1})+f_{t}(n_{1}-(k-1),n_{2},\dots,n_{r})$.
This completes the proof of Theorem 1.7.
## 4 Proof of Theorem 1.8
Recall that, in this section, $r,n_{1},n_{2},\dots,n_{r}$ are integers with
$r\geq 4$, $10k-4\leq n_{1}+4k\leq\min_{i\geq 2}\\{n_{i}\\}$, and
$\sum_{i=1}^{r}n_{i}=n$. By Theorem 1.7, it suffices to prove
$\displaystyle
ex(K_{n_{1},n_{2},\dots,n_{r}},kK_{3})\leq(k-1)(n-n_{1})+f_{3}(n_{1}-(k-1),n_{2},\dots,n_{r}).$
(4)
By way of contradiction, we suppose that $k$ is the minimum positive integer
such that
ex$(K_{n_{1},n_{2},\dots,n_{r}},kK_{3})>(k-1)(n-n_{1})+f_{3}(n_{1}-(k-1),n_{2},\dots,n_{r})$.
Thus, there exists a graph $G$, as a spanning subgraph of
$K_{n_{1},n_{2},\dots,n_{r}}$, containing no copy of $kK_{3}$ and satisfying
$\displaystyle
e(G)=ex(K_{n_{1},n_{2},\dots,n_{r}},kK_{3})>(k-1)(n-n_{1})+f_{3}(n_{1}-(k-1),n_{2},\dots,n_{r}).$
(5)
Moreover, we use $V_{1},V_{2},\dots,V_{r}$ to denote the parts of $G$ , where
$|V_{i}|=n_{i}$ for each $i\in[r]$. It is worth noting that the inequality (4)
holds for $k=1$ by Theorem 1.1. Hence, we may assume $k\geq 2$.
###### Claim 1
$G-\\{v\\}$ contains at least one copy of $(k-1)K_{3}$ for each $v\in V(G)$.
Proof. By contradiction, suppose that there is a vertex $v$ in $V(G)$ such
that $G-\\{v\\}$ contains no copy of $(k-1)K_{3}$. Suppose that $v\in
V_{\ell}$ for some $\ell\in[r]$. Let $n_{\ell}^{\prime}=n_{\ell}-1$,
$n_{j}^{\prime}=n_{j}$ for any $j\neq\ell$, and $n^{\prime}=n-1$. Then
$10(k-1)-4\leq n^{\prime}_{1}+4(k-1)\leq\min_{i\geq 2}\\{n^{\prime}_{i}\\}$.
Moreover, $G-\\{v\\}\subseteq
K_{n^{\prime}_{1},n^{\prime}_{2},\dots,n^{\prime}_{r}}$ by the construction.
Because $G-\\{v\\}$ contains no copy of $(k-1)K_{3}$, we obtain
$e(G-\\{v\\})\leq
ex(K_{n^{\prime}_{1},n^{\prime}_{2},\dots,n^{\prime}_{r}},(k-1)K_{3})$. By the
minimality of $k$, we have
$\displaystyle e(G)=$ $\displaystyle e(G-\\{v\\})+d_{G}(v)$
$\displaystyle\leq$ $\displaystyle
ex(K_{n^{\prime}_{1},n^{\prime}_{2},\dots,n^{\prime}_{r}},(k-1)K_{3})+d_{G}(v)$
$\displaystyle\leq$
$\displaystyle(k-2)(n^{\prime}-n_{1}^{\prime})+f_{3}(n_{1}^{\prime}-(k-2),n_{2}^{\prime},n_{3}^{\prime},\dots,n_{r}^{\prime})+d_{G}(v).$
(6)
If $v\in V_{1}$, then $n_{1}^{\prime}=n_{1}-1$ and
$d_{G}(v)=|N_{G}(v)|\leq|V(G)\setminus V_{1}|=n-n_{1}$ since $G\subseteq
K_{n_{1},n_{2},\dots,n_{r}}$. By the inequality (4), we have
$\displaystyle e(G)\leq$
$\displaystyle(k-2)(n-1-n_{1}+1)+f_{3}(n_{1}-1-(k-2),n_{2},n_{3},\dots,n_{r})+n-n_{1}$
$\displaystyle=$
$\displaystyle(k-1)(n-n_{1})+f_{3}(n_{1}-(k-1),n_{2},n_{3},\dots,n_{r}),$
which contradicts the inequality (5). Hence, we may assume $v\notin V_{1}$.
Then $d_{G}(v)\leq n-n_{\ell}$, $n_{\ell}^{\prime}=n_{\ell}-1$, and
$n_{i}^{\prime}=n_{i}$ for any $i\neq\ell$. Note that $n_{\ell}-1\geq
n_{1}+4k-1\geq n_{1}-k+2$. By the inequality (4) and Proposition 2.2, we have
$\displaystyle e(G)\leq$
$\displaystyle(k-2)(n-1-n_{1})+f_{3}(\\{n_{1}-(k-2),n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\ell}\\}\cup\\{n_{\ell}-1\\})+n-n_{\ell}$
$\displaystyle\leq$
$\displaystyle(k-2)(n-1-n_{1})+(f_{3}(n_{1}-(k-1),n_{2},n_{3},\dots,n_{r})+n_{\ell}-(n_{1}-k+2))+n-n_{\ell}$
$\displaystyle{=}$
$\displaystyle(k-1)(n-n_{1})-k+2+n_{1}+f_{3}(n_{1}-(k-1),n_{2},n_{3},\dots,n_{r})-(n_{1}-k+2)$
$\displaystyle=$
$\displaystyle(k-1)(n-n_{1})+f_{3}(n_{1}-(k-1),n_{2},n_{3},\dots,n_{r}),$
which is a contradiction to the inequality (5). This completes the proof of
Claim 1.
Since $k\geq 2$, the graph $G$ contains at least one triangle by Claim 1.
Arbitrarily choose a triangle $K^{*}_{3}$ in $G$ with vertices
$u_{1},u_{2},u_{3}$. Assume that $u_{1}\in V_{\alpha}$, $u_{2}\in V_{\eta}$,
and $u_{3}\in V_{\xi}$, where $\alpha,\eta,\xi$ are three distinct integers in
$[r]$. Moreover, let $S=\\{\alpha,\eta,\xi\\}$, $n^{\prime}_{\ell}=n_{\ell}-1$
for any $\ell\in S$, $n_{j}^{\prime}=n_{j}$ for any $j\in[r]\setminus S$, and
$n^{\prime}=\sum_{i=1}^{r}n^{\prime}_{i}$. Notice that
$n^{\prime}=\sum_{i=1}^{r}n_{i}-3=n-3$ and $G-\\{u_{1},u_{2},u_{3}\\}\subseteq
K_{n_{1}^{\prime},n_{2}^{\prime},\dots,n_{r}^{\prime}}$.
###### Claim 2
$e(G)\leq(k-2)(n^{\prime}-n^{\prime}_{1})+f_{3}(n^{\prime}_{1}-(k-2),n_{2}^{\prime},\dots,n_{r}^{\prime})+\sum_{1\leq
i<j\leq 3}|N_{G}(u_{i})\cap N_{G}(u_{j})|+n-3$.
Proof. Note that $G-\\{u_{1},u_{2},u_{3}\\}$ contains no copy of $(k-1)K_{3}$
since $G$ contains no copy of $kK_{3}$. Moreover, we have $10(k-1)-4\leq
n^{\prime}_{1}+4(k-1)\leq\min_{i\geq 2}\\{n^{\prime}_{i}\\}$ . By the
minimality of $k$, we know $e(G-\\{u_{1},u_{2},u_{3}\\})\leq
ex(K_{n_{1}^{\prime},n_{2}^{\prime},\dots,n_{r}^{\prime}},(k-1)K_{3})\leq(k-2)(n^{\prime}-n^{\prime}_{1})+f_{3}(n^{\prime}_{1}-(k-2),n_{2}^{\prime},\dots,n_{r}^{\prime})$.
Thus,
$\displaystyle e(G)=$ $\displaystyle
e(G-\\{u_{1},u_{2},u_{3}\\})+\sum_{i=1}^{3}|N_{G}(u_{i})|-3$
$\displaystyle\leq$
$\displaystyle(k-2)(n^{\prime}-n^{\prime}_{1})+f_{3}(n^{\prime}_{1}-(k-2),n_{2}^{\prime},\dots,n_{r}^{\prime})+\sum_{i=1}^{3}|N_{G}(u_{i})|-3.$
(7)
By the Principle of Inclusion-Exclusion,
$\displaystyle\sum_{i=1}^{3}|N_{G}(u_{i})|=$
$\displaystyle|\bigcup_{i=1}^{3}N_{G}(u_{i})|+\sum_{1\leq i<j\leq
3}|N_{G}(u_{i}))\cap N_{G}(u_{j})|-|\bigcap_{i=1}^{3}N_{G}(u_{i})|$
$\displaystyle\leq$ $\displaystyle|V(G)|+\sum_{1\leq i<j\leq
3}|N_{G}(u_{i})\cap N_{G}(u_{j})|$ $\displaystyle=$ $\displaystyle
n+\sum_{1\leq i<j\leq 3}|N_{G}(u_{i})\cap N_{G}(u_{j})|.$ (8)
Combining inequalities (4) and (4), we see
$\displaystyle
e(G)\leq(k-2)(n^{\prime}-n^{\prime}_{1})+f_{3}(n^{\prime}_{1}-(k-2),n_{2}^{\prime},\dots,n_{r}^{\prime})+\sum_{1\leq
i<j\leq 3}|N_{G}(u_{i})\cap N_{G}(u_{j})|+n-3{.}$
We finish our proof in the following two cases.
Case 1. $|N_{G}(u_{i})\cap N_{G}(u_{j})|\leq 6(k-1)$ for every two distinct
vertices $u_{i}$ and $u_{j}$ in $\\{u_{1},u_{2},u_{3}\\}$.
Recall that $G-\\{u_{1},u_{2},u_{3}\\}\subseteq
K_{n_{1}^{\prime},n_{2}^{\prime},\dots,n_{r}^{\prime}}$. We shall consider the
two cases when $1\in\\{\alpha,\eta,\xi\\}$ and $1\notin\\{\alpha,\eta,\xi\\}$.
First, assume $1\in\\{\alpha,\eta,\xi\\}$. Without loss of generality, say
$1=\alpha$. Then $n_{1}^{\prime}=n_{1}-1$ and
$\\{n_{2}^{\prime},n_{3}^{\prime},\dots,n_{r}^{\prime}\\}=\\{n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\}$.
By Claim 2, we have
$\displaystyle e(G)\leq$ $\displaystyle(k-2)(n-3-n_{1}+1)+18(k-1)+n-3$
$\displaystyle+f_{3}(\\{n_{1}-1-k+2,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\})$
$\displaystyle=$ $\displaystyle(k-1)(n-n_{1})-2(k-2)-(n-n_{1})+18(k-1)+n-3$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\})$
$\displaystyle=$ $\displaystyle(k-1)(n-n_{1})+n_{1}+16(k-1)-1$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\}).$
Now, we assume $1\notin\\{\alpha,\eta,\xi\\}$. Then
$n_{i}^{\prime}=n_{i}-1>n_{1}$ for any $i\in\\{\alpha,\eta,\xi\\}$, and
$n_{j}^{\prime}=n_{j}$ for any $j\in[r]\setminus\\{\alpha,\eta,\xi\\}$. By
Claim 2,
$\displaystyle e(G)\leq$ $\displaystyle(k-2)(n-3-n_{1})+18(k-1)+n-3$
$\displaystyle+f_{3}(\\{n_{1}-k+2,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha},n_{\eta},n_{\xi}\\}\cup\\{n_{\alpha}-1,n_{\eta}-1,n_{\xi}-1\\})$
$\displaystyle=$ $\displaystyle(k-1)(n-n_{1})+15(k-1)+n_{1}$
$\displaystyle+f_{3}(\\{n_{1}-k+2,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha},n_{\eta},n_{\xi}\\}\cup\\{n_{\alpha}-1,n_{\eta}-1,n_{\xi}-1\\}).$
(9)
Note that $n_{\alpha}-1\geq n_{1}-k+2$. By Proposition 2.2,
$\displaystyle f_{3}$
$\displaystyle(\\{n_{1}-k+2,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha},n_{\eta},n_{\xi}\\}\cup\\{n_{\alpha}-1,n_{\eta}-1,n_{\xi}-1\\})$
$\displaystyle\leq
f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\})+n_{\alpha}-(n_{1}-k+2).$
(10)
Thus, combining inequalities (4) and (4),
$\displaystyle e(G)\leq$
$\displaystyle(k-1)(n-n_{1})+15(k-1)+n_{1}+n_{\alpha}-(n_{1}-k+2)$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\})$
$\displaystyle=$ $\displaystyle(k-1)(n-n_{1})+16(k-1)+n_{\alpha}-1$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\eta},n_{\xi}\\}\cup\\{n_{\eta}-1,n_{\xi}-1\\}).$
(11)
In both cases ($1\in\\{\alpha,\eta,\xi\\}$ and
$1\notin\\{\alpha,\eta,\xi\\}$), we have inequality (4) holds. Since
$n_{1}-k+1+2\leq\min_{i\geq 2}\\{n_{i}\\}$, by Proposition 2.3 and the
inequality (4),
$\displaystyle e(G)\leq$
$\displaystyle(k-1)(n-n_{1})+n_{\alpha}+16(k-1)-1+f_{3}(n_{1}-k+1,n_{2},n_{3},\dots,n_{r})-(n-k+1)$
$\displaystyle+{\max\\{n_{1}-(k-1)+2,n_{\eta}-n_{1}+(k-1)+1\\}}$
$\displaystyle=$
$\displaystyle(k-1)(n-n_{1})+f_{3}(n_{1}-k+1,n_{2},n_{3},\dots,n_{r})$
$\displaystyle+{17k-18-n+\max\\{n_{\alpha}+n_{1}-k+3,n_{\eta}+n_{\alpha}-n_{1}+k\\}}.$
Since $n-n_{\alpha}-n_{1}\geq(r-2)n_{2}\geq 2n_{2}\geq 20k-8$ and
$n+n_{1}-n_{\eta}-n_{\alpha}\geq(r-1)n_{1}\geq 3n_{1}\geq 18k-12$, we have
$e(G)\leq(k-1)(n-n_{1})+f_{3}(n_{1}-k+1,n_{2},n_{3},\dots,n_{r}),$
which contradicts the inequality (5).
Case 2. There exist two distinct vertices $u_{i}$ and $u_{j}$ in
$\\{u_{1},u_{2},u_{3}\\}$ such that $|N_{G}(u_{i})\cap N_{G}(u_{j})|>6(k-1)$.
Without loss of generality, we may assume $|N_{G}(u_{2})\cap N_{G}(u_{3})|\geq
6(k-1)+1$. Then we have the following claim.
###### Claim 3
For any vertex set $A\subseteq V(G)$, if $|N_{G}(u_{2})\cap N_{G}(u_{3})\cap
A|\geq 6(k-1)+1$, then there is a vertex $u_{0}$ with $u_{0}\in A\cap
N_{G}(u_{2})\cap N_{G}(u_{3})$ such that $|N_{G}(u_{0})\cap N_{G}(u_{i})|\leq
3(k-1)$ for any $i\in\\{2,3\\}$.
Proof. By Claim 1, denote by $\mathcal{T}_{i}$ a copy of $(k-1)K_{3}$ in
$G\setminus\\{u_{i}\\}$ for any $i\in\\{2,3\\}$. Since $|N_{G}(u_{2})\cap
N_{G}(u_{3})\cap A|\geq 6(k-1)+1$ and
$|\mathcal{T}_{1}|=|\mathcal{T}_{2}|=3(k-1)$, there is a vertex, say $u_{0}$,
such that $u_{0}\in(N_{G}(u_{2})\cap N_{G}(u_{3})\cap
A)\setminus(\mathcal{T}_{1}\cup\mathcal{T}_{2})$. If $|N_{G}(u_{0})\cap
N_{G}(u_{i})|\geq 3(k-1)+1$ for some $i\in\\{2,3\\}$, then there exists a
vertex $w_{i}\in(N_{G}(u_{0})\cap N_{G}(u_{i}))\setminus\mathcal{T}_{i}$.
Notice that $\\{w_{i},u_{0},u_{i}\\}\cap\mathcal{T}_{i}=\emptyset$ and
$u_{0}u_{i}w_{i}u_{0}$ is a triangle. So the disjoint union of
$u_{0}u_{i}w_{i}u_{0}$ and $\mathcal{T}_{i}$ is a copy of $kK_{3}$ in $G$, a
contradiction to the choice of $G$. Thus, $|N_{G}(u_{0})\cap N_{G}(u_{i})|\leq
3(k-1)$ for any $i\in\\{2,3\\}$.
By setting $A=V(G)$ in Claim 3, there exists a vertex $u_{0}$ with $u_{0}\in
N_{G}(u_{2})\cap N_{G}(u_{3})$ such that $|N_{G}(u_{0})\cap N_{G}(u_{i})|\leq
3(k-1)$ for any $i\in\\{2,3\\}$. Note that $u_{0}u_{2}u_{3}u_{0}$ is a
triangle in $G$. Recall $u_{2}\in V_{\eta}$ and $u_{3}\in V_{\xi}$. Assume
$u_{0}\in V_{\alpha^{\prime}}$.
According to the arbitrariness of the choice of $K^{*}_{3}$ in $G$, we know
that Claims 2 and 3 are applicable to the triangle $u_{0}u_{2}u_{3}u_{0}$ by
replacing $K^{*}_{3}$ with $u_{0}u_{2}u_{3}u_{0}$ and replacing $S$ with
$\\{\alpha^{\prime},\eta,\xi\\}$. The remaining proof is divided into two
subcases.
Subcase 1. $1\in\\{\alpha^{\prime},\eta,\xi\\}$.
Note that $n_{1}^{\prime}=n_{1}-1$ and $n^{\prime}=n-3$. Moreover, we have
$|N_{G}(u_{2})\cap N_{G}(u_{3})|\leq n-n_{\eta}-n_{\xi}$ by $u_{2}\in
V_{\eta}$ and $u_{3}\in V_{\xi}$. Let
$\\{1,s,\ell\\}=\\{\alpha^{\prime},\eta,\xi\\}$. Due to
$\\{s,\ell\\}\cap\\{\eta,\xi\\}\neq\emptyset$, without loss of generality, we
may assume $s\in\\{\eta,\xi\\}$. Since $|N_{G}(u_{0})\cap N_{G}(u_{i})|\leq
3(k-1)$ for any $i\in\\{2,3\\}$, by Claim 2, we know
$\displaystyle e(G)\leq$
$\displaystyle(k-2)(n-3-n_{1}+1)+(6(k-1)+|N_{G}(u_{2})\cap N_{G}(u_{3})|)+n-3$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{s},n_{\ell}\\}\cup\\{n_{s}-1,n_{\ell}-1\\})$
$\displaystyle\leq$
$\displaystyle(k-1)(n-n_{1})+4k-5+n+n_{1}-n_{\eta}-n_{\xi}$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{s},n_{\ell}\\}\cup\\{n_{s}-1,n_{\ell}-1\\})$
$\displaystyle\leq$
$\displaystyle{(k-1)(n-n_{1})+4k-5+n+n_{1}-n_{\eta}-n_{\xi}+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\})}$
$\displaystyle{-(n-(k-1))+\max\\{(n_{1}-k+1)+2,n_{s}-(n_{1}-k+1)+1\\}}$
$\displaystyle\leq$
$\displaystyle{(k-1)(n-n_{1})+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\})}$
$\displaystyle{+5k-6+n_{1}-n_{\eta}-n_{\xi}+\max\\{n_{1}-k+3,n_{s}-n_{1}+k+2\\}}$
$\displaystyle\leq$
$\displaystyle{(k-1)(n-n_{1})+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}),}$
where the third inequality holds by Proposition 2.3 and the last inequality
holds by $4k-3+2n_{1}\leq n_{\eta}+n_{\xi}$ and $6k-4+n_{s}\leq
n_{\eta}+n_{\xi}$.
Subcase 2. $1\notin\\{\alpha^{\prime},\eta,\xi\\}$.
By setting $A=V_{1}$ in Claim 3, if $|N(u_{2})\cap N(u_{3})\cap V_{1}|\geq
6(k-1)+1$, then there exists a vertex $u^{\prime}_{0}$ with $u^{\prime}_{0}\in
N_{G}(u_{2})\cap N_{G}(u_{3})\cap V_{1}$ such that $|N_{G}(u^{\prime}_{0})\cap
N_{G}(u_{i})|\leq 3(k-1)$ for any $i\in\\{2,3\\}$. Note that
$u^{\prime}_{0}u_{2}u_{3}u^{\prime}_{0}$ is also a triangle in $G$. This cases
reduces to Subcase 1. Hence, we may assume $|N(u_{2})\cap N(u_{3})\cap
V_{1}|\leq 6(k-1)$. Then
$\displaystyle|N_{G}(u_{2})\cap N_{G}(u_{3})|$
$\displaystyle=|N_{G}(u_{2})\cap N_{G}(u_{3})\cap(V(G)\setminus
V_{1})|+|N_{G}(u_{2})\cap N_{G}(u_{3})\cap V_{1}|$ $\displaystyle\leq
n-n_{\eta}-n_{\xi}-n_{1}+6(k-1).$
Recall that $n^{\prime}=n-3$, $n^{\prime}_{1}=n_{1}$, and $|N_{G}(u_{0})\cap
N_{G}(u_{i})|\leq 3(k-1)$ for any $i\in\\{2,3\\}$. By Claim 2,
$\displaystyle e(G)\leq$
$\displaystyle(k-2)(n-3-n_{1})+(6(k-1)+n-n_{\eta}-n_{\xi}-n_{1}+6(k-1))+n-3$
$\displaystyle+f_{3}(\\{n_{1}-k+2,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha^{\prime}},n_{\eta},n_{\xi}\\}\cup\\{n_{\alpha^{\prime}}-1,n_{\eta}-1,n_{\xi}-1\\})$
$\displaystyle=$ $\displaystyle(k-1)(n-n_{1})+9k-9+n-n_{\eta}-n_{\xi}$
$\displaystyle+f_{3}(\\{n_{1}-k+2,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha^{\prime}},n_{\eta},n_{\xi}\\}\cup\\{n_{\alpha^{\prime}}-1,n_{\eta}-1,n_{\xi}-1\\}).$
(12)
Since $n_{1}+4k\leq\min_{i\geq 2}\\{n_{i}\\}$, we know
$n_{1}-k+3\leq\min_{i\geq 2}\\{n_{i}\\}$. By Proposition 2.2 and the
inequality (4),
$\displaystyle e(G)\leq$
$\displaystyle(k-1)(n-n_{1})+9k-9+n-n_{\eta}-n_{\xi}+n_{\xi}-(n_{1}-k+2)$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha^{\prime}},n_{\eta}\\}\cup\\{n_{\alpha^{\prime}}-1,n_{\eta}-1\\})$
$\displaystyle=$ $\displaystyle(k-1)(n-n_{1})+10k-11+n-n_{\eta}-n_{1}$
$\displaystyle+f_{3}(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha^{\prime}},n_{\eta}\\}\cup\\{n_{\alpha^{\prime}}-1,n_{\eta}-1\\}).$
(13)
Proposition 2.3 implies that
$\displaystyle f_{3}$
$\displaystyle(\\{n_{1}-k+1,n_{2},n_{3},\dots,n_{r}\\}\setminus\\{n_{\alpha^{\prime}},n_{\eta}\\}\cup\\{n_{\alpha^{\prime}}-1,n_{\eta}-1\\})$
$\displaystyle\leq
f_{3}(n_{1}-k+1,n_{2},\dots,n_{r})+{\max\\{n_{1}-(k-1)+2,n_{\eta}-(n_{1}-k+1)+1\\}-n+k-1.}$
(14)
Since $n_{1}\geq 6k-4$ and $n_{\eta}\geq n_{1}+4k\geq 10k-4$, combining
inequalities (4) and (4), we have
$\displaystyle e(G)\leq$ $\displaystyle(k-1)(n-n_{1})+10k-11+n-n_{\eta}-n_{1}$
$\displaystyle+f_{3}(n_{1}-k+1,n_{2},\dots,n_{r})+{\max\\{n_{1}-k+3,n_{\eta}-n_{1}+k\\}-n+k-1.}$
$\displaystyle=$
$\displaystyle(k-1)(n-n_{1})+f_{3}(n_{1}-k+1,n_{2},\dots,n_{r})+11k-12$
$\displaystyle-\max\\{n_{\eta}+k-3,2n_{1}-k\\}$ $\displaystyle\leq$
$\displaystyle(k-1)(n-n_{1})+f_{3}(n_{1}-k+1,n_{2},\dots,n_{r}),$
which contradicts to the inequality (5). This completes the proof of the
inequality (4), and so Theorem 1.8 follows.
Acknowledgements. The author would like to thank the editor and the two
referees very much for their valuable suggestions and comments. Junxue Zhang
was partially supported by the National Natural Science Foundation of China
(Nos. 12161141006 and 12111540249), the Natural Science Foundation of Tianjin
(Nos. 20JCJQJC00090 and 20JCZDJC00840), the Fundamental Research Funds for the
Central Universities, Nankai University (No. 63223063).
## References
* [1] B. Bollobás, P. Erdős, E. G. Straus, Complete subgraphs of chromatic graphs and hypergraphs, Utilitas Math. 6 (1974), 343–347.
* [2] H. Chen, X. Li, J. Tu, Complete solution for the rainbow numbers of matchings, Discrete Math. 309 (2009), 3370–3380.
* [3] J. De Silva, K. Heysse, A. Kapilow, A. Schenfisch, M. Young, Turán numbers of vertex-disjoint cliques in $r$-partite graphs, Discrete Math. 341 (2018), 492–496.
* [4] J. De Silva, K. Heysse, M. Young, Rainbow number for matchings in $r$-partite graphs, preprint.
* [5] P. Erdős, über ein Extremalproblem in der Graphentheorie (German), Arch. Math. (Basel) 13 (1962), 122–127.
* [6] P. Erdős, T. Gallai, On maximal paths and circuits of graphs, Acta Math. Acad. Sci. Hungar. 10 (1959), 337–356.
* [7] J. Han, Y. Zhao, Turán number of disjoint triangles in 4-partite graphs, Electron. J. Combin. 29 (2022), #P2.35.
* [8] W. Mantel, Problem 28, Wiskundige Opgaven 10 (1907), 60–61.
* [9] J. W. Moon, On independent complete subgraphs in a graph, Canad. J. Math. 20 (1968), 95–102.
* [10] M. Simonovits, A method for solving extremal problems in graph theory, Theory of Graphs, Proc. Colloq. Tihany, (1966), (P. Erdős and G. Katona, Eds.), 279–319, Acad. Press, New York, 1968.
* [11] P. Turán, On an extremal problem in graph theory, (Hungarian), Mat. Fiz. Lapok 48 (1941), 436–452.
* [12] D. West, Introduction to Graph Theory, Prentice hall, Upper Saddle River, 2001.
|
# Generalizing Yee’s method: Scalable geometric higher-order
FEEC algorithms for Maxwell’s equations on an unstructured mesh
Alexander S. Glasser Hong Qin
###### Abstract
The Yee algorithm for electromagnetic simulations is widely known to have many
advantages, including the following crucial two: (i) Its calculations are
local and therefore efficiently parallelizable—enabling simulations that
capitalize on the speed and scalability of high-performance computing
architecture. (ii) Yee’s method faithfully preserves the symplectic geometry
of Maxwell’s equations, improving its accuracy in long-time numerical
simulations. Whereas previous geometric generalizations of Yee’s method have
sacrificed its scalability, in this article the Yee algorithm is generalized
to higher order and unstructured meshes in a manner that fully preserves both
its scalability and geometric naturalness. This generalization is achieved by
prioritizing the locality of the algorithm, reflecting the physical locality
of Maxwell’s equations. Specifically, we demonstrate that Yee’s method is but
a special case of a larger family of symplectic, finite element exterior
calculus (FEEC) methods that use scalable, local approximations of mass
matrices. We discuss the numerical advantages of this family of methods, which
we call _scalable FEEC_ (SFEEC) methods.
[1]Princeton Plasma Physics Laboratory, Princeton University, Princeton, New
Jersey 08543 [2]Department of Astrophysical Sciences, Princeton University,
Princeton, New Jersey 08544
## 1 Introduction
The Yee algorithm [1, 2]—alternatively, the finite difference time domain
(FDTD) method—defines electromagnetic fields on a cubic mesh. It associates to
each edge a component of the electric field $\mathbf{E}$, and to each face a
component of the magnetic field $\mathbf{B}$. This discretization reflects a
natural geometric description of Maxwell’s equations, in which one defines
${\mathbf{E}\in\Lambda^{1}(\mathbb{R}^{3})}$ as a differential 1-form on
$\mathbb{R}^{3}$ and ${\mathbf{B}\in\Lambda^{2}(\mathbb{R}^{3})}$ as a
differential 2-form on $\mathbb{R}^{3}$. In this way, Yee’s method employs a
technique adopted in many _structure-preserving algorithms_ [3], wherein
differential $k$-forms are discretized by associating them with
$k$-dimensional features of a mesh [4, 5, 6, 7].
Structure-preserving algorithms have been widely adopted in many subfields of
computational physics, including gravitational simulations [8, 9, 10, 11, 12],
geophysics [13, 14] and plasma physics [15, 16, 17, 18, 19, 20, 21, 22, 23].
Such algorithms generally derive from variational principles or Hamiltonian
systems. As a result, they preserve essential mathematical features of their
underlying physical systems, including symplectic structure, topology,
symmetries, and conservation laws. These properties contribute to the
numerical fidelity of structure-preserving algorithms, especially in long-time
numerical simulations. Despite Yee’s omission of any overt Lagrangian or
Hamiltonian formulations in his original work [1], Yee’s method (apparently
serendipitously) is one of the most historically successful examples of a
structure-preserving algorithm [24].
More recently, the methodology of structure-preserving algorithms has led to
the development of advanced algorithms for the simulation of plasmas, whose
electromagnetic fields are constructed using the formalism of finite element
exterior calculus (FEEC) [6, 7, 20, 23]. Although such methods are in
principle readily generalizable to unstructured meshes and high order finite
elements, they lack the computational efficiency and scalability of Yee’s
method. In particular, the time evolution of electromagnetic fields in these
FEEC methods requires communication between all nodes of a simulation, thereby
destroying their parallelism. As we shall discuss and address, this problem
arises because sparse finite element mass matrices generally have dense
inverses.
On the other hand, there have also been numerous efforts (e.g. [25, 26, 27,
28]) to generalize Yee’s method using scalable finite element methods, called
finite element time domain (FETD) methods. Such efforts leverage the crucial
technique of sparse approximate inverse (SPAI) mass matrices, and preserve the
scalability of Yee’s method. However, these methods’ preservation of
symplectic structure has not been established.
Motivated by the desire to overcome the scalability limitation of structure-
preserving FEEC plasma methods, and to guarantee the structure preservation of
FETD methods, in this article we develop _scalable FEEC_ (SFEEC) methods, a
family of symplectic finite element methods for electromagnetic fields that
includes Yee’s method as a special case. SFEEC methods enable the higher order
simulation of electromagnetic fields on structured and unstructured meshes in
a manner that preserves the two aforementioned crucial advantages of Yee’s
method, namely: (i) its scalability on modern architectures; and (ii) its
symplectic geometry (and the resulting conservation of electric charge and
Gauss’ law).
From a finite element point of view, the scalability of Yee’s method will be
reframed as a result of its simplified (or ‘pruned’) approximation of finite
element mass matrices. To retain their scalability, SFEEC methods employ a
comparable, if more flexible treatment of mass matrices. Relative to Yee’s
method, SFFEC methods afford a greater flexibility to improve algorithmic
accuracy without sacrificing scalability. Their use of finite elements and the
FEEC formalism further enables their viability on more general meshes.
While Yee’s algorithm has been generalized numerous times in the literature,
including via higher order and finite element schemes (e.g. [24, 25, 27, 28,
29, 30, 31, 32]), no such generalization is known to us that simultaneously
affords higher order accuracy and scalability while ensuring the geometric
structure-preservation of Yee’s method. In particular, the Yee method’s
preservation of symplectic structure is an important aspect of the method’s
stability and long-term accuracy [33]. In this work, we will review this
symplectic structure of Yee’s method and demonstrate its exact preservation in
the SFEEC family of algorithms we define. In so doing, we also offer a means
to overcome the scalability limitations of structure-preserving FEEC
algorithms (e.g. [20, 23]) in a massively parallel, high-performance computing
architecture.
The remainder of this article is organized as follows. In Section 2, the
formalism of finite element exterior calculus (FEEC) [6, 7] and its
discretization of electromagnetic fields will be reviewed. In Section 3, it
will be demonstrated that Yee’s algorithm can be interpreted as an FEEC method
with simplified mass matrices. In Section 4 we will define SFEEC methods,
which extend Yee’s method to higher order finite element schemes on a general
mesh. In Section 5, numerical results will be presented that demonstrate the
improved higher order accuracy of the resulting SFEEC methods relative to
Yee’s method, without sacrificing its scalability. Section 6 will then
summarize and conclude.
## 2 Finite Element Exterior Calculus (FEEC)
for Electromagnetic Simulations
In this section, we briefly review aspects of FEEC [6, 7] and its application
in electromagnetic simulations. We refer the reader to [20, 23] for additional
background. We let $\Lambda^{p}(\mathcal{T}_{h})$ denote a vector space of
finite element differential $p\text{-form}$s on a simplicial or cubical
complex ${\mathcal{T}_{h}\subset\mathbb{R}^{n}}$ (where the subscript $h$
denotes the maximal diameter, or edge length, of any simplex in
$\mathcal{T}_{h}$). For all ${0\leq p\leq n}$, $\Lambda^{p}(\mathcal{T}_{h})$
may be defined as the span of a finite ($N_{p}$-dimensional) basis
${\mbox{\boldmath$\Lambda$}^{p}}$, whose $i^{\text{th}}$ basis element
${\Lambda^{p}_{i}\in\Lambda^{p}(\mathcal{T}_{h})}$ is a piecewise polynomial
$p\text{-form}$ on $\mathcal{T}_{h}$. Such a basis element typically has
support localized to one or more adjacent cells in ${\mathcal{T}_{h}}$. An
arbitrary $p\text{-form}$ ${\mathbf{S}\in\Lambda^{p}(\mathcal{T}_{h})}$ can be
expressed in the $\mbox{\boldmath$\Lambda$}^{p}$ basis as
$\displaystyle\begin{split}\mathbf{S}(\mathbf{x})&=\mathbf{s}\cdot\mbox{\boldmath$\Lambda$}^{p}(\mathbf{x})=s_{i}\Lambda^{p}_{i}(\mathbf{x})\end{split}$
(1)
$\forall$ ${\mathbf{s}\in\mathbb{R}^{N_{p}}}$ and
${\mathbf{x}\in\left|\mathcal{T}_{h}\right|}$, the convex hull of
${\mathcal{T}_{h}}$. (Einstein summation convention is used for the repeated
index in Eq. (1) and hereafter.) Individual components of
${\mathbf{S}\in\Lambda^{p}(\mathcal{T}_{h})}$ will be denoted
$\displaystyle\begin{split}\mathbf{S}(\mathbf{x})_{\mu_{1}\cdots\mu_{p}}&=\mathbf{s}\cdot\mbox{\boldmath$\Lambda$}^{p}(\mathbf{x})_{\mu_{1}\cdots\mu_{p}}=s_{i}\Lambda^{p}_{i}(\mathbf{x})_{\mu_{1}\cdots\mu_{p}}\end{split}$
(2)
where Greek letters denote coordinate indices. For example, given
${\left|\mathcal{T}_{h}\right|\subset\mathbb{R}^{3}}$, the $\mu^{\text{th}}$
component of the 1-form basis element $\Lambda^{1}_{i}(\mathbf{x})$ may be
written as ${\Lambda^{1}_{i}(\mathbf{x})_{\mu}}$ $\forall$
${\mu\in\\{1,2,3\\}}$, such that
${\Lambda^{1}_{i}(\mathbf{x})=\Lambda^{1}_{i}(\mathbf{x})_{\mu}\mathrm{d}x^{\mu}}$.
Because each basis ${\mbox{\boldmath$\Lambda$}^{p}}$ is finite $\forall$
${0\leq p\leq n}$, the exterior derivative
${\mathrm{d}:\Lambda^{p}(\mathcal{T}_{h})\rightarrow\Lambda^{p+1}(\mathcal{T}_{h})}$
of a finite element $p\text{-form}$ on ${\mathcal{T}_{h}}$ can be computed in
the ${\mbox{\boldmath$\Lambda$}^{0},\dots,\mbox{\boldmath$\Lambda$}^{n}}$
bases by straightforward matrix multiplication. A choice of basis for each
${\Lambda^{p}(\mathcal{T}_{h})}$ in three dimensions
(${\mathcal{T}_{h}\subset\mathbb{R}^{3}}$) determines, for example, matrices
that represent the gradient ($\mathbb{G}$), curl ($\mathbb{C}$), and
divergence ($\mathbb{D}$)—as defined in Table 1.
$p$-Form | Abstract $\mathrm{d}$ | Matrix $\mathrm{d}$ | Defined by
---|---|---|---
$\mathbf{S}=\mathbf{s}\cdot\mbox{\boldmath$\Lambda$}^{0}$ | - | - | -
$\mathbf{A}=\mathbf{a}\cdot\mbox{\boldmath$\Lambda$}^{1}$ | $\mathbf{A}=\mathrm{d}\mathbf{S}$ | $\mathbf{a}=\mathbb{G}\mathbf{s}$ | $\mathbb{G}^{T}\mbox{\boldmath$\Lambda$}^{1}=\mathrm{d}\mbox{\boldmath$\Lambda$}^{0}$
$\mathbf{B}=\mathbf{b}\cdot\mbox{\boldmath$\Lambda$}^{2}$ | $\mathbf{B}=\mathrm{d}\mathbf{A}$ | $\mathbf{b}=\mathbb{C}\mathbf{a}$ | $\mathbb{C}^{T}\mbox{\boldmath$\Lambda$}^{2}=\mathrm{d}\mbox{\boldmath$\Lambda$}^{1}$
$\mathbf{C}=\mathbf{c}\cdot\mbox{\boldmath$\Lambda$}^{3}$ | $\mathbf{C}=\mathrm{d}\mathbf{B}$ | ${\mathbf{c}=\mathbb{D}\mathbf{b}}$ | ${\mathbb{D}^{T}\mbox{\boldmath$\Lambda$}^{3}=\mathrm{d}\mbox{\boldmath$\Lambda$}^{2}}$
Table 1: FEEC matrix implementation of $\mathrm{d}$ on
${\mathcal{T}_{h}\subset\mathbb{R}^{3}}$. The property
${\mathrm{d}\circ\mathrm{d}=0}$ implies that
${\mathbb{C}\mathbb{G}=\mathbb{0}}$ and ${\mathbb{D}\mathbb{C}=\mathbb{0}}$.
To apply FEEC in a physical setting, it will also be essential to compute
_mass matrices_ on $\mathcal{T}_{h}$ for each basis
$\mbox{\boldmath$\Lambda$}^{p}$ of finite element $p\text{-form}$s.
Specifically, the mass matrix ${\mathbb{M}_{p}\in\mathbb{R}^{N_{p}\times
N_{p}}}$ for $\mbox{\boldmath$\Lambda$}^{p}$ is defined by
$\displaystyle\begin{split}(\mathbb{M}_{p})_{ij}&=\int_{\left|\mathcal{T}_{h}\right|}\mathrm{d}\mathbf{x}\Big{(}\Lambda^{p}_{i},\Lambda^{p}_{j}\Big{)}_{p}\end{split}$
(3)
where ${(\cdot,\cdot)_{p}}$ denotes the pointwise inner product on
$p\text{-form}$s induced by the metric $g_{\mu\nu}$—specifically
$\displaystyle\begin{split}(\alpha,\beta)_{p}&=\frac{1}{p!}\alpha_{\mu_{1}\cdots\mu_{p}}\beta^{\mu_{1}\cdots\mu_{p}}\\\
&=\frac{1}{p!}\alpha_{\mu_{1}\cdots\mu_{p}}\beta_{\nu_{1}\cdots\nu_{p}}g^{\mu_{1}\nu_{1}}\cdots
g^{\mu_{p}\nu_{p}}.\end{split}$ (4)
For ${p=1,2}$ on $\mathbb{R}^{3}$, for example, Eq. (4) defines the inner
product ${(\alpha,\beta)_{p}}$ as the standard inner product
${\alpha\cdot\beta}$. After all, 1- and 2-forms each have three independent
components in $\mathbb{R}^{3}$, and the metric ${g_{\mu\nu}}$ (and its inverse
${g^{\mu\nu}}$) is given by the Kronecker
delta—${g_{\mu\nu}=\delta_{\mu\nu}}$. We note that ${(\alpha,\beta)_{p}}$ is
symmetric, such that Eq. (3) is symmetric and
${\mathbb{M}_{p}^{T}=\mathbb{M}_{p}}$. Eq. (3) also implies that
$\mathbb{M}_{p}$ is typically sparse (so long as each basis element
$\Lambda_{i}^{p}$ has only local support). Importantly, however, its inverse
matrix $\mathbb{M}_{p}^{-1}$ is typically dense.
To apply the FEEC formalism to electromagnetic simulations, we will first
discretize the magnetic vector potential as an FEEC 1-form, that is,
$\displaystyle\begin{split}\mathbf{A}=\mathbf{a}\cdot\mbox{\boldmath$\Lambda$}^{1}\end{split}$
(5)
with degrees of freedom given by the vector of coefficients
${\mathbf{a}\in\mathbb{R}^{N_{1}}}$. (We work in the temporal gauge, such that
the electric potential $\phi$ vanishes, ${\phi=0}$.) The magnetic field is
then given by a 2-form,
$\displaystyle\begin{split}\mathbf{B}=\mathbf{b}\cdot\mbox{\boldmath$\Lambda$}^{2}=\mathbb{C}\mathbf{a}\cdot\mbox{\boldmath$\Lambda$}^{2}=\mathbf{a}\cdot\mathbb{C}^{T}\mbox{\boldmath$\Lambda$}^{2}=\mathbf{a}\cdot\mathrm{d}\mbox{\boldmath$\Lambda$}^{1}=\mathrm{d}\mathbf{A}\end{split}$
(6)
with ${\mathbf{b}\in\mathbb{R}^{N_{2}}}$, in keeping with the geometry of
Maxwell’s equations and the FEEC implementation of the curl, as in Table 1.
Lastly, the electric field will be defined as a discrete 1-form
$\displaystyle\begin{split}\mathbf{E}=\mathbf{e}\cdot\mathbb{M}_{1}^{-1}\cdot\mbox{\boldmath$\Lambda$}^{1}\end{split}$
(7)
with ${\mathbf{e}\in\mathbb{R}^{N_{1}}}$. Following [23], we use a convention
wherein the coefficients $\mathbf{e}$ determine the electric field
$\mathbf{E}$ with an additional factor of $\mathbb{M}_{1}^{-1}$. We shall
regard the pair ${(\mathbf{a},\mathbf{e})\in\mathbb{R}^{2N_{1}}}$ as defining
the degrees of freedom of the electromagnetic finite element dynamical system.
To describe the dynamics of these discrete fields via Maxwell’s equations, we
first recall the canonical symplectic structure of electromagnetic fields in
the continuum, defined with Poisson bracket and Hamiltonian in SI units as
[34]
$\displaystyle\begin{split}\\{F,G\\}_{\text{EM}}&=\frac{1}{\epsilon_{0}}\int\mathrm{d}\mathbf{x}\left(\frac{\delta
F}{\delta\mathbf{E}}\cdot\frac{\delta G}{\delta\mathbf{A}}-\frac{\delta
G}{\delta\mathbf{E}}\cdot\frac{\delta F}{\delta\mathbf{A}}\right)\\\
H_{\text{EM}}&=\frac{1}{2}\int\mathrm{d}\mathbf{x}\left(\epsilon_{0}\left|\mathbf{E}\right|^{2}+\frac{1}{\mu_{0}}\left|\nabla\times\mathbf{A}\right|^{2}\right).\end{split}$
(8)
We now substitute the FEEC fields
${\mathbf{A}=\mathbf{a}\cdot\mbox{\boldmath$\Lambda$}^{1}}$ and
${\mathbf{E}=\mathbf{e}\cdot\mathbb{M}_{1}^{-1}\cdot\mbox{\boldmath$\Lambda$}^{1}}$
into Eq. (8) to derive (see [23]) the following discrete Poisson bracket
${\\{\cdot,\cdot\\}_{\Delta}}$ and discrete Hamiltonian $H_{\Delta}$
approximating this system,
$\displaystyle\begin{split}\\{F,G\\}_{\Delta}&=\frac{1}{\epsilon_{0}}\left(\frac{\partial
F}{\partial\mathbf{e}}\cdot\frac{\partial
G}{\partial\mathbf{a}}-\frac{\partial
G}{\partial\mathbf{e}}\cdot\frac{\partial F}{\partial\mathbf{a}}\right)\\\
H_{\Delta}&=\frac{1}{2}\left(\epsilon_{0}\mathbf{e}^{T}\mathbb{M}_{1}^{-1}\mathbf{e}+\frac{1}{\mu_{0}}\mathbf{a}^{T}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}\right),\end{split}$
(9)
where we have applied
${\mathrm{d}\mbox{\boldmath$\Lambda$}^{1}=\mathbb{C}^{T}\mbox{\boldmath$\Lambda$}^{2}}$
as in Table 1.
Equations of motion are now readily calculated as usual with a Poisson
bracket. We plug the FEEC coefficient vectors $\mathbf{a}$ and $\mathbf{e}$
into ${\\{\cdot,\cdot\\}_{\Delta}}$ with $H_{\Delta}$, to find
$\displaystyle\begin{split}\dot{\mathbf{a}}&=\\{\mathbf{a},H_{\Delta}\\}_{\Delta}=-\mathbb{M}_{1}^{-1}\mathbf{e}\\\
\dot{\mathbf{e}}&=\\{\mathbf{e},H_{\Delta}\\}_{\Delta}=c^{2}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}.\end{split}$
(10)
These equations are discrete forms of Maxwell’s equations,
${\dot{\mathbf{A}}=-\mathbf{E}}$ and
${\dot{\mathbf{E}}=c^{2}\nabla\times\nabla\times\mathbf{A}}$, respectively. In
Section 5, the discrete approximation of the curl-of-curl operator
$\displaystyle\begin{split}\nabla\times\nabla\times\approx\mathbb{M}_{1}^{-1}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C},\end{split}$
(11)
whose calculation (until our later simplification) incorporates the dense
matrix $\mathbb{M}_{1}^{-1}$, will be a central focus of our study.
Note, it will also be convenient to rewrite Eq. (10) in terms of the magnetic
field by substituting ${\mathbf{b}=\mathbb{C}\mathbf{a}}$ (the discrete
realization of ${\mathbf{B}=\nabla\times\mathbf{A}}$) to find
$\displaystyle\begin{split}\dot{\mathbf{b}}&=-\mathbb{C}\mathbb{M}_{1}^{-1}\mathbf{e}\\\
\dot{\mathbf{e}}&=c^{2}\mathbb{C}^{T}\mathbb{M}_{2}\mathbf{b}.\end{split}$
(12)
Because of the gauge symmetry of Maxwell’s equations, Eq. (12) is an
equivalent formulation of Eq. (10) and exactly preserves the physical degrees
of freedom of interest [22, 23].
To algorithmically evolve Eq. (10) in discrete time, we follow [17] (see also
[22]) and define a splitting algorithm that decomposes $H_{\Delta}$ of Eq. (9)
into two pieces:
$\displaystyle\begin{split}H_{\Delta}&=H_{\mathbf{a}}+H_{\mathbf{e}}\\\
\text{where}~{}~{}~{}H_{\mathbf{a}}&=\frac{1}{2\mu_{0}}\mathbf{a}^{T}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}\\\
H_{\mathbf{e}}&=\frac{\epsilon_{0}}{2}\mathbf{e}^{T}\mathbb{M}_{1}^{-1}\mathbf{e}.\end{split}$
(13)
These ‘sub-Hamiltonians’ define two different subsystems which may be evolved
separately, that is,
$\displaystyle\begin{split}H_{\mathbf{a}}&\begin{cases}\dot{\mathbf{a}}&=\\{\mathbf{a},H_{\mathbf{a}}\\}_{\Delta}=0\\\
\dot{\mathbf{e}}&=\\{\mathbf{e},H_{\mathbf{a}}\\}_{\Delta}=c^{2}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}.\end{cases}\\\
H_{\mathbf{e}}&\begin{cases}\dot{\mathbf{a}}&=\\{\mathbf{a},H_{\mathbf{e}}\\}_{\Delta}=-\mathbb{M}_{1}^{-1}\mathbf{e}\\\
\dot{\mathbf{e}}&=\\{\mathbf{e},H_{\mathbf{e}}\\}_{\Delta}=0.\end{cases}\end{split}$
(14)
We see that $\mathbf{a}$ is constant in the subsystem $H_{\mathbf{a}}$, while
$\mathbf{e}$ is constant in the subsystem $H_{\mathbf{e}}$. As a result, each
subsystem is exactly solvable over a finite time interval $\Delta t$, in
particular,
$\displaystyle\begin{split}H_{\mathbf{a}}:~{}~{}~{}~{}\mathbf{e}(t_{0}+\Delta
t)&=\mathbf{e}(t_{0})+\Delta t\cdot
c^{2}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}(t_{0})\\\
H_{\mathbf{e}}:~{}~{}~{}~{}\mathbf{a}(t_{0}+\Delta
t)&=\mathbf{a}(t_{0})-\Delta
t\cdot\mathbb{M}_{1}^{-1}\mathbf{e}(t_{0}).\end{split}$ (15)
By choosing an appropriate sequence of these subsystems’ discrete time
evolutions, a splitting method may be implemented that approximates a timestep
of the entire electromagnetic system $H_{\Delta}$.
A typical example of such a splitting scheme is called Strang splitting [35],
which evolves $\mathbf{a}$ and $\mathbf{e}$ according to the approximation
$\displaystyle\begin{split}\left[\begin{matrix}\mathbf{a}\\\
\mathbf{e}\end{matrix}\right]_{\mathrlap{t_{0}+\Delta
t}}&~{}~{}~{}=\exp\left(\Delta
tH_{\Delta}\right)\cdot\left[\begin{matrix}\mathbf{a}\\\
\mathbf{e}\end{matrix}\right]_{t_{0}}\\\ &\approx\exp\left(\frac{\Delta
t}{2}H_{\mathbf{e}}\right)\exp\left(\Delta
tH_{\mathbf{a}}\right)\exp\left(\frac{\Delta
t}{2}H_{\mathbf{e}}\right)\cdot\left[\begin{matrix}\mathbf{a}\\\
\mathbf{e}\end{matrix}\right]_{t_{0}}.\end{split}$ (16)
Here, we use the notation ${\exp(\Delta tH_{i})}$ to denote a discrete time
mapping—as appears in Eq. (15)—that evolves according to the Hamiltonian
$H_{i}$ for a time interval $\Delta t$. Since $\exp(\Delta tH_{\mathbf{a}})$
and $\exp(\Delta tH_{\mathbf{e}})$ do not commute, the second line of Eq. (16)
is only an approximation of the first. Nevertheless, such a splitting method
is an effective algorithmic implementation of our dynamical system. It is
exactly solvable, accurate to second order, and preserves the symplectic
structure of the system as defined by the Poisson bracket of Eq. (9).
Splitting methods of arbitrarily high accuracy in $\Delta t$ can also be
constructed (see, e.g. [3]).
## 3 Reinterpreting Yee’s Method via FEEC
We now rederive Yee’s method using the finite element formalism of the
previous section. We proceed in three steps: (i) we first choose generalized
Whitney forms as an FEEC basis on a cubical mesh; (ii) we then maximally
simplify the mass matrices of the corresponding electromagnetic algorithm; and
finally (iii) we choose Strang splitting as our method of time evolution.
These three steps will exactly recover Yee’s method, demonstrating its
interpretation as an implementation of the FEEC electromagnetic algorithm
defined in Eq. (15)—with simplified mass matrices.
We begin by considering a cubical lattice
${\mathcal{T}_{h}\subset\mathbb{R}^{3}}$ with lattice spacings
${\\{\Delta_{x},\Delta_{y},\Delta_{z}\\}}$. We make the simplest possible
choice for an FEEC basis on $\mathcal{T}_{h}$, namely, the _generalized
Whitney forms_ , a family of piecewise polynomial finite elements denoted
${Q_{1}^{-}\Lambda^{p}(\mathcal{T}_{h})}$ [36]). This finite element basis is
also usefully defined in [37] Example 5.2.
A generalized Whitney $p\text{-form}$ can be readily understood by its 1-to-1
correspondence with a $p$-dimensional feature of a mesh. In particular, we
define the generalized Whitney $p\text{-form}$ ${\mathcal{W}_{\sigma^{p}}\in
Q_{1}^{-}\Lambda^{p}(\mathcal{T}_{h})}$ associated to a given $p$-face
${\sigma^{p}\subset\mathcal{T}_{h}\subset\mathbb{R}^{n}}$ by requiring that
$\displaystyle\begin{split}\frac{1}{\left|\tau^{p}\right|}\int_{\tau^{p}}\mathcal{W}_{\sigma^{p}}=\begin{cases}1&\tau^{p}=\sigma^{p}\\\
0&\tau^{p}\neq\sigma^{p}.\end{cases}\end{split}$ (17)
In this way, generalized Whitney $p\text{-form}$s are _dual_ to $p$-faces of
$\mathcal{T}_{h}$ via integration. Here, ${\left|\tau^{p}\right|}$ denotes the
$p$-volume of $\tau^{p}$ (defined as 1, length, area, and volume respectively
for ${p=0,1,2,3}$). This normalization is chosen to mimic fields as they are
typically represented in Yee’s method.
A concrete example of this is represented on
${\mathcal{T}_{h}\subset\mathbb{R}^{3}}$ in Fig. 1. The 1-form
${\mathcal{W}_{x_{1}x_{2}}\in Q_{1}^{-}\Lambda^{1}(\mathcal{T}_{h})}$ and the
2-form ${\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}\in
Q_{1}^{-}\Lambda^{2}(\mathcal{T}_{h})}$ are depicted, defined respectively by
$\displaystyle\begin{split}\mathcal{W}_{x_{1}x_{2}}&=\left(1-\tfrac{y}{\Delta_{y}}\right)\left(1-\tfrac{z}{\Delta_{z}}\right)\mathrm{d}x\\\
\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}&=\left(1-\tfrac{z}{\Delta_{z}}\right)\mathrm{d}x\wedge\mathrm{d}y.\end{split}$
(18)
As in Eq. (17), $\mathcal{W}_{x_{1}x_{2}}$ is the least-order piecewise
polynomial 1-form satisfying
${\int_{x_{1}x_{2}}\mathcal{W}_{x_{1}x_{2}}=\left|x_{1}x_{2}\right|}$ and
${\int_{\sigma^{1}}\mathcal{W}_{x_{1}x_{2}}=0}$ $\forall$ ${\sigma^{1}\neq
x_{1}x_{2}}$. $\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}$ is analogously defined so
that its integral satisfies
${\int_{x_{1}x_{2}x_{4}x_{3}}\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}=\left|x_{1}x_{2}x_{4}x_{3}\right|}$
and vanishes on any other face. In later calculations, it will also be useful
to explicitly define the following Whitney 2-form associated to the face
${x_{1}x_{5}x_{6}x_{2}}$ in Fig. 1,
$\displaystyle\begin{split}\mathcal{W}_{x_{1}x_{5}x_{6}x_{2}}&=\left(1-\tfrac{y}{\Delta_{y}}\right)\mathrm{d}z\wedge\mathrm{d}x.\end{split}$
(19)
0ptOlFFL
Figure 1: The generalized Whitney 1-forms $\mathcal{W}_{x_{1}x_{2}}$ and
$\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}$ are schematically depicted, respectively,
on the left and right. $\mathcal{W}_{x_{1}x_{2}}$ evaluates to the length
$\left|x_{1}x_{2}\right|$ when integrated along the edge ${x_{1}x_{2}}$ (in
blue) and vanishes on all other edges (in orange).
$\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}$ likewise yields the area
$\left|x_{1}x_{2}x_{4}x_{3}\right|$ when integrated over the blue face
${x_{1}x_{2}x_{4}x_{3}}$, and vanishes on all other faces.
Now consider the exterior derivative $\mathrm{d}$ of these forms. The curl
matrix ${\mathbb{C}:\mathbb{R}^{N_{1}}\rightarrow\mathbb{R}^{N_{2}}}$ is a
linear operator that precisely computes $\mathrm{d}$ for 1-forms, mapping the
coefficients of a discrete 1-form—expanded in the
${Q_{1}^{-}\Lambda^{1}(\mathcal{T}_{h})}$ basis—to coefficients of a discrete
2-form—expanded in the ${Q_{1}^{-}\Lambda^{2}(\mathcal{T}_{h})}$ basis. The
1-to-1 correspondence between these generalized Whitney forms and elements of
the mesh leads to a convenient interpretation of $\mathbb{C}$ in that basis,
as follows.
We find by directly computing the exterior derivative of
$\mathcal{W}_{x_{1}x_{2}}$ using Eqs. (18-19), for example, that
$\displaystyle\begin{split}\mathrm{d}\mathcal{W}_{x_{1}x_{2}}&=\frac{1}{\Delta_{y}}\mathcal{W}_{x_{1}x_{2}x_{4}x_{3}}-\frac{1}{\Delta_{z}}\mathcal{W}_{x_{1}x_{5}x_{6}x_{2}}.\end{split}$
(20)
We observe that a Whitney 2-form $\mathcal{W}_{\sigma^{2}}$ appears on the
right hand side above if and only if its associated face $\sigma^{2}$ contains
the edge $x_{1}x_{2}$ on its boundary. Its coefficient ${\pm
1/\Delta_{x^{\mu}}}$ is determined by the $\mu^{\text{th}}$ dimension in which
the associated face extends $x_{1}x_{2}$, and its sign is set by the relative
orientation between the edge and face.
Note, this interpretation of $\mathrm{d}$—as a mapping with coefficients ${\pm
1/\Delta_{x^{\mu}}}$ from 1-forms to the 2-forms of faces which contain them
on the boundary—is applicable only to the particular (cubical) FEEC mesh and
(Whitney form) FEEC basis that we have chosen. In this context, however, our
interpretation leads us to note that the curl operator $\mathbb{C}$, (which
implements $\mathrm{d}$ as defined in Table 1), acts on on the generalized
Whitney forms of a cubical lattice as nothing more than a finite difference
operator. To see this more explicitly, let us denote by
${\mathbf{a}_{\sigma^{1}}}$ the entry of ${\mathbf{a}\in\mathbb{R}^{N_{1}}}$
corresponding to the Whitney 1-form basis element
${\mathcal{W}_{\sigma^{1}}}$. Likewise, we let ${\mathbf{b}_{\sigma^{2}}}$
denote the coefficient in
${\mathbf{b}=\mathbb{C}\mathbf{a}\in\mathbb{R}^{N_{2}}}$ corresponding to
${\mathcal{W}_{\sigma^{2}}}$. With reference again to Fig. 1, we find that
$\displaystyle\begin{split}\mathbf{b}_{x_{1}x_{2}x_{4}x_{3}}&=(\mathbb{C}\mathbf{a})_{x_{1}x_{2}x_{4}x_{3}}\\\
&=\frac{1}{\Delta_{x}}\Big{(}\mathbf{a}_{x_{2}x_{4}}-\mathbf{a}_{x_{1}x_{3}}\Big{)}-\frac{1}{\Delta_{y}}\Big{(}\mathbf{a}_{x_{3}x_{4}}-\mathbf{a}_{x_{1}x_{2}}\Big{)}.\end{split}$
(21)
Under the map $\mathbb{C}$, the finite differences of the edge coefficients
are therefore summed to derive the coefficient on a face.
The transpose of the curl operator $\mathbb{C}^{T}$, yields an analogous
result, with ${(\mathbb{C}^{T}\mathbf{b})_{\sigma^{1}}}$ computed via finite
differences between the faces adjoining edge $\sigma^{1}$. Using Fig. 2 as a
reference, for example, we find
$\displaystyle\begin{split}(\mathbb{C}^{T}\mathbf{b})_{x_{1}x_{5}}=\frac{1}{\Delta_{x}}\Big{(}\mathbf{b}&{}_{x_{1}x_{5}x_{6}x_{2}}-\mathbf{b}_{x_{10}x_{12}x_{5}x_{1}}\Big{)}\\\
&-\frac{1}{\Delta_{y}}\Big{(}\mathbf{b}_{x_{1}x_{3}x_{7}x_{5}}-\mathbf{b}_{x_{9}x_{1}x_{5}x_{11}}\Big{)}.\end{split}$
(22)
Figure 2: A depiction of the degrees of freedom involved in
$\mathbb{C}^{T}$—the transposed curl operator—for generalized Whitney forms on
a cubic mesh.
The observation that $\mathbb{C}$ acts as a finite difference operator on
cubical generalized Whitney forms marks a first step toward recovering Yee’s
method from Eq. (15).
As a second step, let us first consider what $\mathbb{M}_{1}$ and
$\mathbb{M}_{2}$ look like for our chosen Whitney form FEEC basis on a cubical
mesh. Ignoring boundary cells for simplicity, we use Eq. (3) to integrate
inner products of forms such as those appearing in Eq. (18) to find that
$\displaystyle\begin{split}(\mathbb{M}_{1})_{\sigma^{1},\tau^{1}}&=\Delta_{V}\cdot\begin{cases}4/9&\sigma^{1}=\tau^{1}\\\
1/9&\sigma^{1}\parallel\tau^{1}\text{ and
}\exists~{}\tau^{2}\supset\\{\sigma^{1},\tau^{1}\\}\\\
1/36&\sigma^{1}\parallel\tau^{1}\text{ and
}\exists~{}\tau^{3}\supset\\{\sigma^{1},\tau^{1}\\}\end{cases}\\\
(\mathbb{M}_{2})_{\sigma^{2},\tau^{2}}&=\Delta_{V}\cdot\begin{cases}2/3&\sigma^{2}=\tau^{2}\\\
1/6&\sigma^{2}\parallel\tau^{2}\text{ and
}\exists~{}\tau^{3}\supset\\{\sigma^{2},\tau^{2}\\}\end{cases}\end{split}$
(23)
where ${\Delta_{V}=\Delta_{x}\Delta_{y}\Delta_{z}}$ denotes a cell volume.
Given these matrix entries, each row (and column) of $\mathbb{M}_{1}$ and
$\mathbb{M}_{2}$ can be readily shown to sum to $\Delta_{V}$. In the interest
of recovering Yee’s method, therefore, we define the simplified mass matrices
$\displaystyle\begin{split}\mathbb{M}_{1}^{Y}&=\Delta_{V}\cdot\mathbb{1}_{N_{1}\times
N_{1}}\\\ \mathbb{M}_{2}^{Y}&=\Delta_{V}\cdot\mathbb{1}_{N_{2}\times
N_{2}}.\end{split}$ (24)
Here, $\mathbb{M}_{p}^{Y}$ signifies the ‘Yee-modified’ mass matrix for
$p\text{-form}$s, and $\mathbb{1}$ denotes the identity matrix. This process
is often referred to as ‘lumping’ the mass matrix (of Eq. (23), for example)
into diagonal form.
Importantly, using such simplified mass matrices leaves the symplectic
structure defined by ${\\{\cdot,\cdot\\}_{\Delta}}$ in Eq. (9) entirely
undisturbed. Rather, a substitution of the simplified mass matrix
${\mathbb{M}_{p}\rightarrow\mathbb{M}_{p}^{Y}}$ is properly viewed as taking
place within the discrete Hamiltonian $H_{\Delta}$ in Eq. (9). While this
coarser approximation of $H_{\text{EM}}$ reduces the accuracy of the resulting
(Yee’s) algorithm, it has no deleterious effect on its characterization as a
structure-preserving algorithm.
We now complete our program to recover Yee’s method. Let us consider again the
Strang splitting defined in Eq. (16). We can rewrite this system more
explicitly as
$\displaystyle\begin{split}\left[\begin{matrix}\mathbf{a}\\\
\mathbf{e}\end{matrix}\right]_{\mathrlap{t_{0}+\Delta
t}}~{}~{}&=~{}~{}\mathbb{H}_{\mathbf{e}}^{\Delta
t/2}\mathbb{H}_{\mathbf{a}}^{\Delta t}\mathbb{H}_{\mathbf{e}}^{\Delta
t/2}\left[\begin{matrix}\mathbf{a}\\\
\mathbf{e}\end{matrix}\right]_{t_{0}}~{}~{}~{}\text{where}\\\
\mathbb{H}_{\mathbf{e}}^{\Delta t}&=\left[\begin{matrix}\mathbb{1}&-\Delta
t\mathbb{M}_{1}^{-1}\\\
\mathbb{0}&\mathbb{1}\end{matrix}\right]~{}~{}\text{,}~{}~{}\mathbb{H}_{\mathbf{a}}^{\Delta
t}=\left[\begin{matrix}\mathbb{1}&\mathbb{0}\\\ \Delta
tc^{2}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}&\mathbb{1}\end{matrix}\right]\end{split}$
(25)
or equivalently, in terms of ${\mathbf{b}=\mathbb{C}\mathbf{a}}$:
$\displaystyle\begin{split}\left[\begin{matrix}\mathbf{b}\\\
\mathbf{e}\end{matrix}\right]_{\mathrlap{t_{0}+\Delta
t}}&~{}~{}=~{}~{}\mathcal{H}_{\mathbf{e}}^{\Delta
t/2}\mathcal{H}_{\mathbf{b}}^{\Delta t}\mathcal{H}_{\mathbf{e}}^{\Delta
t/2}\left[\begin{matrix}\mathbf{b}\\\
\mathbf{e}\end{matrix}\right]_{t_{0}}~{}~{}~{}\text{where}\\\
\mathcal{H}_{\mathbf{e}}^{\Delta t}&=\left[\begin{matrix}\mathbb{1}&-\Delta
t\mathbb{C}\mathbb{M}_{1}^{-1}\\\
\mathbb{0}&\mathbb{1}\end{matrix}\right]~{}~{}~{}\text{,}~{}~{}~{}\mathcal{H}_{\mathbf{b}}^{\Delta
t}=\left[\begin{matrix}\mathbb{1}&\mathbb{0}\\\ \Delta
tc^{2}\mathbb{C}^{T}\mathbb{M}_{2}&\mathbb{1}\end{matrix}\right].\end{split}$
(26)
Eq. (26) is iterated in a simulation, so that after ${n+1}$ timesteps its
evolution operator is ${\mathcal{H}_{\mathbf{e}}^{\Delta
t/2}\left[\mathcal{H}_{\mathbf{e}}^{\Delta t}\mathcal{H}_{\mathbf{b}}^{\Delta
t}\right]^{n}\mathcal{H}_{\mathbf{e}}^{\Delta t/2}}$. Therefore, applying the
‘finite-difference operators’ $\mathbb{C}$ and $\mathbb{C}^{T}$ from Eqs.
(21-22) and the mass matrix approximations $\mathbb{M}_{p}^{Y}$ from Eq. (24),
let us examine our splitting method at ‘mid-step’—after
${[\mathcal{H}_{\mathbf{e}}^{\Delta t}\mathcal{H}_{\mathbf{b}}^{\Delta
t}]\mathcal{H}_{\mathbf{e}}^{\Delta t/2}}$—and compare it with Yee’s method.
Since ${\mathcal{H}_{\mathbf{e}}^{\Delta t/2}}$ only evolves $\mathbf{b}$, we
regard this as an ‘initialization step’ that gives us
$\displaystyle\begin{split}\mathcal{H}_{\mathbf{e}}^{\Delta
t/2}:\left(\begin{matrix}\mathbf{b}[t_{0}]\\\
\mathbf{e}[t_{0}]\end{matrix}\right)\mapsto\left(\begin{matrix}\mathbf{b}\left[t_{1/2}\right]\\\
\mathbf{e}\left[t_{0}\right]\end{matrix}\right).\end{split}$ (27)
In particular, this half-step establishes initial field data as it appears in
Yee’s method. Next, ${\mathcal{H}_{\mathbf{b}}^{\Delta t}}$ only evolves
$\mathbf{e}$, such that
$\displaystyle\begin{split}\mathcal{H}_{\mathbf{b}}^{\Delta
t}:\left(\begin{matrix}\mathbf{b}\left[t_{1/2}\right]\\\
\mathbf{e}\left[t_{0}\right]\end{matrix}\right)\mapsto\left(\begin{matrix}\mathbf{b}\left[t_{1/2}\right]\\\
\mathbf{e}\left[t_{1}\right]\end{matrix}\right)\\\\[5.0pt] \text{ where
}\mathbf{e}\left[t_{1}\right]=\mathbf{e}\left[t_{0}\right]+\Delta
t\Delta_{V}c^{2}\mathbb{C}^{T}\mathbf{b}[t_{1/2}].\end{split}$ (28)
This timestep
${(\mathbf{e}\left[t_{1}\right]-\mathbf{e}\left[t_{0}\right])/\Delta
tc^{2}=\Delta_{V}\mathbb{C}^{T}\mathbf{b}[t_{1/2}]}$ replicates the Yee
method’s evolution of the electric field, e.g.
$\displaystyle\begin{split}&\frac{1}{c^{2}\Delta
t}\left(E_{x}^{n+1}\left[i+\tfrac{1}{2},j,k\right]-E_{x}^{n}\left[i+\tfrac{1}{2},j,k\right]\right)\\\
&=\frac{1}{\Delta
y}\left(B_{z}^{n+\frac{1}{2}}\left[i+\tfrac{1}{2},j+\tfrac{1}{2},k\right]-B_{z}^{n+\frac{1}{2}}\left[i+\tfrac{1}{2},j-\tfrac{1}{2},k\right]\right)\\\
&-\frac{1}{\Delta
z}\left(B_{y}^{n+\frac{1}{2}}\left[i+\tfrac{1}{2},j,k+\tfrac{1}{2}\right]-B_{y}^{n+\frac{1}{2}}\left[i+\tfrac{1}{2},j,k-\tfrac{1}{2}\right]\right).\end{split}$
(29)
As depicted in Fig. 2 and described in Eq. (22), the operator $\mathbb{C}^{T}$
implements precisely the finite differences of Eq. (29). The additional
constant factor $\Delta_{V}$ in Eq. (28) has no effect other than changing the
effective unit in which $\mathbf{b}$ is expressed. Finally,
${\mathcal{H}_{\mathbf{e}}^{\Delta t}}$ only evolves $\mathbf{b}$, such that
$\displaystyle\begin{split}\mathcal{H}_{\mathbf{e}}^{\Delta
t}:\left(\begin{matrix}\mathbf{b}\left[t_{1/2}\right]\\\
\mathbf{e}\left[t_{1}\right]\end{matrix}\right)\mapsto\left(\begin{matrix}\mathbf{b}\left[t_{3/2}\right]\\\
\mathbf{e}\left[t_{1}\right]\end{matrix}\right)\\\\[5.0pt] \text{ where
}\mathbf{b}\left[t_{3/2}\right]=\mathbf{b}\left[t_{1/2}\right]-(\Delta
t/\Delta_{V})\mathbb{C}\mathbf{e}[t_{1}].\end{split}$ (30)
This timestep
${(\mathbf{b}\left[t_{3/2}\right]-\mathbf{b}\left[t_{1/2}\right])\Delta_{V}/\Delta
t=\mathbb{C}\mathbf{e}[t_{1}]}$ replicates Yee’s evolution of the magnetic
field, e.g.
$\displaystyle\begin{split}\frac{1}{\Delta
t}&\left(B_{x}^{n+\frac{1}{2}}\left[i,j+\tfrac{1}{2},k+\tfrac{1}{2}\right]-B_{x}^{n-\frac{1}{2}}\left[i,j+\tfrac{1}{2},k+\tfrac{1}{2}\right]\right)\\\
&=\frac{1}{\Delta
z}\left(E_{y}^{n}\left[i,j+\tfrac{1}{2},k+1\right]-E_{y}^{n}\left[i,j+\tfrac{1}{2},k\right]\right)\\\
&-\frac{1}{\Delta
y}\left(E_{z}^{n}\left[i,j+1,k+\tfrac{1}{2}\right]-E_{z}^{n}\left[i,j,k+\tfrac{1}{2}\right]\right).\end{split}$
(31)
As described in Eq. (21), $\mathbb{C}$ implements precisely the finite
differences of Eq. (31). The factor of $\Delta_{V}$ again only impacts the
unit of $\mathbf{b}$.
Consequently, we have demonstrated that Yee’s algorithm is equivalent to the
structure-preserving FEEC method of Eq. (15), using Whitney forms on a cubical
mesh, maximally simplified (diagonal) mass matrices, and Strang splitting time
evolution.
## 4 Generalizing Yee’s Method
| Yee’s Method | SFEEC methods
---|---|---
Splitting method | Strang | any (e.g. Lie-Trotter [38], Strang)
Mesh | cubical | any (e.g. simplicial, cubical)
Finite elements | Whitney forms | any FEEC basis (e.g. ${\mathcal{P}_{r}^{-}\Lambda^{k},\mathcal{P}_{r}\Lambda^{k}}$)
$\mathbb{M}_{1}^{-1}$ and $\mathbb{M}_{2}$ | diagonal approximation | any sparse approximation
Table 2: The SFEEC generalization of Yee’s method affords considerable
flexibility in the choice of splitting scheme, mesh, and finite element basis
in its implementation of Eq. (15). SFEEC methods are symplectic and are also
required to be scalable. Therefore, they require sparse mass matrices and
their sparse approximate inversions.
The previous section demonstrates that Yee’s method is a special case of the
FEEC structure-preserving algorithm defined by Eq. (15). Specifically, Yee’s
method is seen to make several selections for the FEEC algorithm, including
1. 1.
timesteps given by a Strang splitting scheme;
2. 2.
a cubical grid;
3. 3.
the least-order FEEC finite element basis—cubical Whitney forms; and
4. 4.
mass matrices simplified into diagonal form.
The last of these—the mass matrix approximation—can be seen not so much as a
‘selection’ but as a ‘departure’ from the algorithm defined by Eq. (15). As
previously described, however, a simplified treatment of mass matrices occurs
at the level of the Hamiltonian $H_{\Delta}$ in Eq. (9), and does not affect
the Poisson bracket ${\\{\cdot,\cdot\\}_{\Delta}}$. As a result, Yee’s method
remains symplectic and—though it loses accuracy—retains all of the advantages
of a structure-preserving algorithm. Yee’s method thereby preserves the
geometric structure of Maxwell’s equations while employing a parsimonious
description of electromagnetic fields. Moreover, its sacrifice of accuracy in
mass matrices is arguably compensated by the scalability the Yee method
achieves.
Here, we propose to relax all four of the above selections made for the Yee
scheme, and in so doing we define a new class of algorithms, _scalable FEEC_
(SFEEC) methods. In particular, Table 2 enumerates the generalizations of
Yee’s method afforded by SFEEC methods.
As defined, SFEEC methods offer considerable flexibility in generalizing the
Yee scheme. They have the full flexibility of FEEC behind them, and can
therefore handle quite general meshes and high order finite elements.
Moreover, the splitting scheme used for SFEEC methods can be chosen to be of
arbitrarily high order accuracy in time.
As we shall see, however, this improvement in accuracy is constrained by the
requirement that SFEEC methods remain scalable. In particular, while mass
matrices will no longer be maximally simplified into diagonal form (as in
Yee’s approach), $\mathbb{M}_{1}^{-1}$ will require sparse approximation. Even
with this caveat, we shall demonstrate that SFEEC methods offer a significant
improvement upon the accuracy of Yee’s method, at limited additional cost in
computational effort.
Let us consider in greater detail the need for simplified mass matrices in
large scale implementations of Eq. (15). We note that essentially any choice
of FEEC finite element has only local support, so that the mass matrices
$\mathbb{M}_{1}$ and $\mathbb{M}_{2}$ are generally sparse by definition.
However, Eq. (15) requires the use of $\mathbb{M}_{1}^{-1}$, and the inverse
of a sparse matrix is generally dense. Given $\mathbb{M}_{1}^{-1}$ dense,
however, a computation that evolves subsystem ${H_{\mathbf{e}}}$ of Eq. (15)
would require every node of a simulation to pass its local $\mathbf{e}$ data
to every other node, so that $\mathbf{a}$ could be evolved. This communication
would clearly spoil the efficacy of parallelization.
On the other hand, we may leverage the spirit of Yee’s algorithm as seen
through an FEEC lens, and consider more parsimonious approximations of the
matrix $\mathbb{M}_{1}^{-1}$. We emphasize from our discussion above that we
can ‘prune’ $\mathbb{M}_{1}^{-1}$ as minimally or as maximally as desired: Any
approximation of $\mathbb{M}_{1}^{-1}$ will produce, as Yee’s method produces,
a symplectic algorithm. The choice of approximation should be guided,
therefore, strictly by the trade-off between its scalability and its accuracy
in approximating the electromagnetic Hamiltonian, $H_{\Delta}\approx
H_{\text{EM}}$.
Before proceeding to describe our numerical results in the next section, we
first describe our method to find sparse approximations of
$\mathbb{M}_{1}^{-1}$. Our general strategy (following [26],[39]) is to
identify a desirable sparsity pattern a priori, and find a matrix
${\mathbb{Q}\approx\mathbb{M}_{1}^{-1}}$ of the desired sparsity pattern that
best approximates ${\mathbb{M}_{1}^{-1}}$. We shall denote the best fit for
$\mathbb{M}_{1}^{-1}$ with sparsity pattern ${S(\mathbb{A})}$ as
$\mathbb{Q}_{S(\mathbb{A})}$, where ${S(\mathbb{A})}$ denotes the sparsity
pattern of an appropriate matrix $\mathbb{A}$. The goodness of fit may be
determined by the Froebenius norm $\left\lVert\cdot\right\rVert_{F}$, in which
case $\mathbb{Q}$ can be solved for column-by-column, since
$\displaystyle\begin{split}\left\lVert\mathbb{M}_{1}\mathbb{Q}-\mathbb{1}\right\rVert_{F}^{2}=\sum\limits_{\ell=1}^{N_{1}}\left\lVert\mathbb{M}_{1}\mathbf{q}_{\ell}-1_{\ell}\right\rVert^{2}.\end{split}$
(32)
Here, $\mathbf{q}_{\ell}$ is the $\ell^{\text{th}}$ column of ${\mathbb{Q}}$
and $1_{\ell}$ is the $\ell^{\text{th}}$ column of the identity matrix, (a
standard basis vector). Thus, $\mathbf{q}_{\ell}$ can be found $\forall$
$\ell$ independently—in parallel, if desired—as the solution to a least
squares problem, in which the only entries of $\mathbf{q}_{\ell}$ permitted to
vary are those that fit the desired sparsity pattern for the
$\ell^{\text{th}}$ column of $\mathbb{Q}$.
More explicitly, consider the problem of solving for the entries of one such
column $\mathbf{q}_{\ell}$. We may define an index
${\mathcal{I}_{\ell}\subset\\{1,\dots,N_{1}\\}}$ corresponding to the row
numbers of nonzero entries permitted in $\mathbf{q}_{\ell}$ according to the
desired sparsity pattern $S(\mathbb{Q})$. Denote the cardinality of such an
index ${\left|\mathcal{I}_{\ell}\right|\leq N_{1}}$. Then let
${\mathbb{M}_{1}(:,\mathcal{I}_{\ell})}$ denote the
${N_{1}\times\left|\mathcal{I}_{\ell}\right|}$ submatrix of $\mathbb{M}_{1}$
comprised of the subset of its columns indicated by $\mathcal{I}_{\ell}$. For
each ${1\leq\ell\leq N_{1}}$ we minimize
${\left\lVert\mathbb{M}_{1}(:,\mathcal{I}_{\ell})\mathbf{q}_{\ell}(\mathcal{I}_{\ell})-1_{\ell}\right\rVert}$,
that is,
$\displaystyle\begin{split}\operatorname*{arg\,min}\limits_{\mathbf{q}_{\ell}(\mathcal{I}_{\ell})}\sum\limits_{\begin{subarray}{c}1\leq
i\leq N_{1}\\\
j\in\mathcal{I}_{\ell}\end{subarray}}\Big{(}(\mathbb{M}_{1})_{ij}(\mathbf{q}_{\ell})_{j}-\delta_{i\ell}\Big{)}^{2}.\end{split}$
(33)
In the numerical examples we describe below, we solve Eq. (33) directly via
$\displaystyle\begin{split}\mathbf{q}_{\ell}(\mathcal{I}_{\ell})=[\mathbb{M}_{1}(:,\mathcal{I}_{\ell})^{T}\mathbb{M}_{1}(:,\mathcal{I}_{\ell})]^{-1}\mathbb{M}_{1}(:,\mathcal{I}_{\ell})^{T}1_{\ell}.\end{split}$
(34)
Most implementations of Eq. (33) will be sparse enough to be solved this way.
Denser choices of the sparsity pattern (${\left|\mathcal{I}_{\ell}\right|\gg
1}$), however, for which the inversion in Eq. (34) is more difficult, can also
be readily solved using, for example, the conjugate gradient method (see [40])
or, if desired, constrained versions thereof (see [41]).
## 5 Numerical Results
Figure 3: The figures above depict the ${L^{2}\Lambda^{1}}$ log relative error
in an FEEC approximation of
$\dot{\mathbf{E}}=c^{2}\nabla\times\nabla\times\mathbf{A}$ vs. log cell size
$h$. The $x$-axis measures the number of cells (of size $h$) per wavelength
${\lambda_{\mathbf{A}}=2\pi/k_{n}}$ of the vector potential
${\mathbf{A}=\sin(k_{n}y)\mathrm{d}x}$. The left plot uses the first order
(Whitney form) ${\mathcal{P}_{1}^{-}\Lambda^{p}(\mathcal{T}_{h})}$ FEEC basis,
while the right plot uses a more accurate second order basis,
${\mathcal{P}_{2}^{-}\Lambda^{p}(\mathcal{T}_{h})}$. For each basis, we plot
the relative errors of four approximations of the FEEC curl-of-curl operator,
${\nabla\times\nabla\times\approx\mathbb{Q}_{i}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$,
${i\in\\{1,\dots,4\\}}$, and fit their power scalings against $h$. Here,
${\mathbb{Q}_{i}}$ is an approximation of the inverse mass matrix
$\mathbb{M}_{1}^{-1}$ with $i$ indicating one of four sparsity patterns:
diagonal, $\mathbb{M}_{1}$, $(\mathbb{M}_{1})^{2}$, and dense (which recovers
the exact FEEC operator). The left plot demonstrates that the accuracy of
Yee’s method, (which effectively uses Whitney forms and diagonal mass
matrices), is meaningfully improved upon by an approximation of
$\mathbb{M}_{1}^{-1}$ that has the sparsity pattern of $\mathbb{M}_{1}$. The
right plot demonstrates that approximating $\mathbb{M}_{1}^{-1}$ by a matrix
with the sparsity pattern of ${(\mathbb{M}_{1})^{2}}$ achieves much of the
improved accuracy possible with second order finite elements. It is seen that
the power scalings extend to a finite resolution, at which point relative
error saturates.
To measure the accuracy of an approximation
${\mathbb{Q}\approx\mathbb{M}_{1}^{-1}}$ in a manner relevant to our
electromagnetic problem, we examine the curl-of-curl operator
${\nabla\times\nabla\times}$, discretized by
${\mathbb{M}_{1}^{-1}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$ in the FEEC
setting, as in Eq. (11). For simplicity, we work with simplicial finite
elements in 2-D, and generate Delaunay triangulations ${\mathcal{T}_{h}}$ of a
2-torus domain
${\left|\mathcal{T}_{h}\right|=[0,L_{x}]\times[0,L_{y}]\subset\mathbb{R}^{2}}$
with periodic boundary conditions. On $\left|\mathcal{T}_{h}\right|$, we
consider a sinusoidal vector potential ${\mathbf{A}=\sin(k_{n}y)\mathrm{d}x}$
for ${k_{n}=2\pi n/L_{y}}$ and ${n\in\mathbb{N}}$, and canonically project it
onto some choice of discrete FEEC basis [6],
${\mathbf{a}\cdot\mbox{\boldmath$\Lambda$}^{1}\in\Lambda^{1}(\mathcal{T}_{h})}$.
We measure the ${L^{2}\Lambda^{1}}$ error between
${\nabla\times\nabla\times\mathbf{A}=k_{n}^{2}\sin(k_{n}y)\mathrm{d}x}$
(exactly representing ${\dot{\mathbf{E}}/c^{2}}$ in the continuum) and its
FEEC discretization,
${\mathbb{M}_{1}^{-1}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}}$
(representing ${\mathbb{M}_{1}^{-1}\dot{\mathbf{e}}/c^{2}}$). Finally, we
investigate how the error of approximating ${\mathbb{M}_{1}^{-1}}$ with
$\mathbb{Q}$—i.e., of approximating ${\nabla\times\nabla\times\mathbf{A}}$
with ${\mathbb{Q}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}\mathbf{a}}$—varies
with the sparsity pattern ${S(\mathbb{Q})}$.
We repeat this procedure over a range of frequencies ${k_{n}=2\pi n/L_{y}}$
and mesh diameters $h$, (the latter achieved by varying the number of vertices
in $\mathbb{T}^{2}$ and generating a Delaunay triangulation $\mathcal{T}_{h}$
for each). We examine two different FEEC bases on this 2-D simplicial mesh:
${\mathcal{P}_{1}^{-}\Lambda^{p}(\mathcal{T}_{h})}$, the Whitney 1-forms
(described above for a cubical mesh), and their counterparts at the next order
of accuracy, ${\mathcal{P}_{2}^{-}\Lambda^{p}(\mathcal{T}_{h})}$—the second-
order trimmed polynomial family of finite elements. For each basis, we compare
four different sparsity patterns for our approximation of
${\mathbb{M}_{1}^{-1}}$, including: diagonal (Yee’s implicit pattern),
$\mathbb{M}_{1}$ sparsity,
${(\mathbb{M}_{1})^{2}=\mathbb{M}_{1}\cdot\mathbb{M}_{1}}$
sparsity,111${(\mathbb{M}_{1})_{ij}}$ is nonzero whenever basis elements
$\Lambda^{1}_{i}$ and $\Lambda^{1}_{j}$ of $\mbox{\boldmath$\Lambda$}^{1}$
have nontrivial ‘overlap,’ in the sense of Eq. (3).
${(\mathbb{M}_{1})^{2}_{ij}}$ is generally nonzero whenever $\Lambda^{1}_{i}$
and $\Lambda^{1}_{j}$ each has nontrivial overlap with a third basis finite
element $\Lambda^{1}_{k}$ (where ${k\in\\{i,j\\}}$ is also permitted). In our
context, therefore, ${(\mathbb{M}_{1})^{2}}$ is strictly denser than
$\mathbb{M}_{1}$. By extension, the Neumann representation of
$\mathbb{M}_{1}^{-1}$ demonstrates that ${S((\mathbb{M}_{1})^{n})\rightarrow
S(\mathbb{M}_{1}^{-1})}$ as $n$ approaches $N_{1}$ [39]. and dense (the exact
inverse). The results of this investigation are depicted in Fig. 3.
With the setup described above, the two subplots of Fig. 3 display the log
relative $L^{2}\Lambda^{1}$ errors in our approximations of
${\dot{\mathbf{E}}=c^{2}\nabla\times\nabla\times\mathbf{A}}$. (For brevity, we
use ${\dot{\mathbf{E}}}$ merely as a shorthand notation.) In particular, given
the $L^{2}\Lambda^{1}$ norm (see Eq. (4))
$\displaystyle\begin{split}||\dot{\mathbf{E}}||_{L^{2}\Lambda^{1}}=\int_{\left|\mathcal{T}_{h}\right|}(\dot{\mathbf{E}},\dot{\mathbf{E}})_{1}\mathrm{d}\mathbf{x},\end{split}$
(35)
we compute
$\displaystyle\begin{split}||\hat{\dot{\mathbf{E}}}-\dot{\mathbf{E}}||_{L^{2}\Lambda^{1}}/||\dot{\mathbf{E}}||_{L^{2}\Lambda^{1}}\end{split}$
(36)
where
${\hat{\dot{\mathbf{E}}}=c^{2}\mathbb{Q}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$
denotes the relevant finite element approximation of the continuum 1-form
${\dot{\mathbf{E}}=c^{2}k_{n}^{2}\sin(k_{n}y)\mathrm{d}x}$.
The left plot demonstrates that the FEEC curl-of-curl operator
${\mathbb{M}_{1}^{-1}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$ is well
approximated for a Whitney form basis using
${\mathbb{Q}_{S(\mathbb{M}_{1})}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$
(where $\mathbb{Q}_{S(\mathbb{M}_{1})}$ denotes the approximation of
$\mathbb{M}_{1}^{-1}$ with a sparsity pattern of $\mathbb{M}_{1}$). In
particular, the relative error of
${\mathbb{Q}_{S(\mathbb{M}_{1})}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$
scales with cell size as ${h^{0.73}}$, while the exact FEEC operator
${\mathbb{M}_{1}^{-1}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$ error scales
comparably as ${h^{0.82}}$.
Remarkably, the first plot demonstrates that a substantial improvement in
accuracy over a diagonal sparsity pattern analogous to Yee’s method is
achieved by using only a slightly less parsimonious estimate of
$\mathbb{M}_{1}^{-1}$ in the curl-of-curl FEEC operator. In particular, while
${\mathbb{Q}_{S(\mathbb{M}_{1})}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$
scales as ${h^{0.73}}$ (as noted just above), the accuracy of
${\mathbb{Q}_{S(\mathbb{1})}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$ (with
$\mathbb{Q}_{S(\mathbb{1})}$ diagonal-patterned) scales as ${h^{0.56}}$.
More important than this modest improvement in power scaling, however, is its
applicability to finer meshes. The first plot of Fig. 3 shows that the power
scaling of the diagonal mass matrix curl-of-curl approximation can only be
assumed for meshes of relatively low resolution. The relative error in the
${\mathbb{Q}_{S(\mathbb{1})}}$ discrete differential operator is seen to
saturate when the number of cells per signal wavelength climbs above 5—that is
${\lambda_{\mathbf{A}}/h\gtrsim 5}$. This is in contrast to the
${\mathbb{Q}_{S(\mathbb{M}_{1})}}$ power scaling, which continues to hold even
at the higher resolution of ${\lambda_{\mathbf{A}}/h\sim 20}$. This could be a
particular advantage in simulations of small-scale electromagnetic structure,
such as in turbulent plasmas or nanomaterials.
It must be further emphasized that the additional computational cost
introduced by this less parsimonious $\mathbb{Q}_{S(\mathbb{M}_{1})}$
approximation of $\mathbb{M}_{1}^{-1}$ will generally be fairly modest. The
pairs of nodes that communicate in a simulation with an
$\mathbb{M}_{1}$-sparsity approximation are typically the same as those that
communicate with a diagonal-sparsity approximation, such that no ‘new
channels’ of communication are introduced. Each communication between nodes
will pass more data, however—in our 2-D example, roughly 5 times as much.
Nevertheless, the amount of data passed is small to begin with, as it scales
only with the number of boundary cells of a node.
The right plot of Fig. 3 demonstrates that, to achieve the accuracy of higher
order finite elements, denser approximations of $\mathbb{M}_{1}^{-1}$ must be
used in the curl-of-curl operator. Nevertheless, greater accuracy is readily
attained. In our 2-D simplicial example, we see that much of the improved
accuracy possible with second order finite elements is achieved by using a
curl-of-curl operator
${\mathbb{Q}_{S(\mathbb{M}_{1}^{2})}\mathbb{C}^{T}\mathbb{M}_{2}\mathbb{C}}$,
which approximates $\mathbb{M}_{1}^{-1}$ with a sparsity pattern of
${(\mathbb{M}_{1})^{2}}$. Indeed, this approximate FEEC operator achieves an
error scaling with cell size of $h^{1.70}$, while the exact FEEC curl-of-curl
operator error scales in our results as $h^{1.98}$. It is worth noting,
however, that the approximation’s scaling holds only until a mesh resolution
of roughly ${\lambda_{\mathbf{A}}/h\gtrsim 5}$. Nevertheless, due to the
higher order finite element, this saturation occurs at a significantly higher
accuracy overall.
## 6 Discussion
We have demonstrated that Yee’s method can be interpreted as a structure-
preserving splitting method using an FEEC formalism on a cubical mesh with
simplified mass matrices. In so doing, we have identified Yee’s algorithm to
be a special case of a larger family of methods we call SFEEC (scalable finite
element exterior calculus) methods, summarized by Eq. (15) and Table 2. These
methods are structure-preserving, as Yee’s method is, and are therefore
accurate in long-time numerical simulations. They also respect the topological
and geometric properties of Maxwell’s equations. Crucially, SFEEC methods are
also scalable; they adopt the strategy implicit in Yee’s method of using
sparse approximations of finite element mass matrices to achieve computational
efficiency.
The classification of SFEEC methods enabled us to identify higher order
extensions of Yee’s method that preserve its scalability nevertheless (see
Fig. 3). Depending on the computing architecture employed and the desired
accuracy of the problem of interest, these alternative methods may enable
greater long-time accuracy than Yee’s method provides, with little additional
computational cost. The applicability of SFEEC methods on general meshes may
also be of particular benefit in problems with irregular geometries.
The saturation of the relative errors plotted in Fig. 3 suggest that it would
be fruitful to investigate in future work whether improved sparse
approximations of the discrete codifferential operator,
${\mathbb{M}_{1}^{-1}\mathbb{C}^{T}\mathbb{M}_{2}}$ can be found. Such efforts
might be geared to preserve higher order finite element error scaling at
arbitrarily high mesh resolution.
## 7 Acknowledgments
Thank you to Josh Burby and Tyrus Berry for helpful discussions, and to Phil
Morrison for his support. This research was further supported by the U.S.
Department of Energy (DE-AC02-09CH11466), as well as the U.S. Department of
Energy Fusion Energy Sciences Postdoctoral Research Program administered by
the Oak Ridge Institute for Science and Education (ORISE) for the DOE. ORISE
is managed by Oak Ridge Associated Universities (ORAU) under DOE contract
number DE-SC0014664. All opinions expressed in this paper are the authors’ and
do not necessarily reflect the policies and views of DOE, ORAU, or ORISE.
## References
* [1] K. Yee, “Numerical solution of initial boundary value problems involving maxwell’s equations in isotropic media,” _IEEE Transactions on Antennas and Propagation_ , vol. 14, no. 3, pp. 302–307, 1966.
* [2] A. Taflove and S. C. Hagness, _Computational electrodynamics: the finite-difference time-domain method_ , 3rd ed. Boston: Artech House, 2005.
* [3] E. Hairer, C. Lubich, and G. Wanner, _Geometric Numerical Integration_ , 2nd ed. Berlin: Springer, 2006.
* [4] H. Whitney, _Geometric Integration Theory_. Princeton, NJ: Princeton University Press, 1957.
* [5] M. Desbrun, A. N. Hirani, M. Leok, and J. E. Marsden, “Discrete exterior calculus,” _arXiv preprint math/0508341_ , 2005. [Online]. Available: https://arxiv.org/abs/math/0508341
* [6] D. N. Arnold, R. S. Falk, and R. Winther, “Finite element exterior calculus, homological techniques, and applications,” _Acta Numerica_ , vol. 15, 2006\.
* [7] ——, “Finite element exterior calculus: from Hodge theory to numerical stability,” _Bulletin of the American Mathematical Society_ , vol. 47, no. 2, pp. 281–354, 2010.
* [8] H. Kinoshita, H. Yoshida, and H. Nakai, “Symplectic integrators and their application to dynamical astronomy,” _CELESTIAL MECHANICS AND DYNAMICAL ASTRONOMY_ , vol. 50, no. 1, pp. 59–71, 1991.
* [9] B. Gladman, M. Duncan, and J. Candy, “Symplectic integrators for long-term integrations in celestial mechanics,” _Celestial Mechanics and Dynamical Astronomy_ , vol. 52, no. 3, pp. 221–240, 1991.
* [10] J. E. Chambers, E. V. Quintana, M. J. Duncan, and J. J. Lissauer, “Symplectic Integrator Algorithms for Modeling Planetary Accretion in Binary Star Systems,” _The Astronomical Journal_ , vol. 123, no. 5, pp. 2884–2894, 2002.
* [11] A. Bravetti, M. Seri, M. Vermeeren, and F. Zadra, “Numerical integration in celestial mechanics: a case for contact geometry,” _Celestial Mechanics and Dynamical Astronomy_ , vol. 132, no. 1, 2020.
* [12] E. Kur and A. S. Glasser, “Discrete gravity with local Lorentz invariance,” _Physical Review D_ , vol. 106, no. 6, p. 064001, 2022.
* [13] X. Li, W. Wang, M. Lu, M. Zhang, and Y. Li, “Structure-preserving modelling of elastic waves,” _Geophysical Journal International_ , vol. 188, no. 3, pp. 1382–1392, 2012.
* [14] S. Liu, X. Li, W. Wang, L. Xu, and B. Li, “A modified symplectic scheme for seismic wave modeling,” _Journal of Applied Geophysics_ , vol. 116, pp. 110–120, 2015.
* [15] J. Squire, H. Qin, and W. M. Tang, “Geometric integration of the Vlasov-Maxwell system with a variational particle-in-cell scheme,” _Physics of Plasmas_ , vol. 19, no. 8, p. 084501, 2012.
* [16] J. Xiao, H. Qin, J. Liu, Y. He, R. Zhang, and Y. Sun, “Explicit high-order non-canonical symplectic particle-in-cell algorithms for Vlasov-Maxwell systems,” _Physics of Plasmas_ , vol. 22, no. 11, p. 112504, 2015.
* [17] Y. He, H. Qin, Y. Sun, J. Xiao, R. Zhang, and J. Liu, “Hamiltonian time integrators for Vlasov-Maxwell equations,” _Physics of Plasmas_ , vol. 22, no. 12, p. 124503, 2015.
* [18] N. Crouseilles, L. Einkemmer, and E. Faou, “Hamiltonian splitting for the Vlasov–Maxwell equations,” _Journal of Computational Physics_ , vol. 283, pp. 224–240, feb 2015.
* [19] H. Qin, J. Liu, J. Xiao, R. Zhang, Y. He, Y. Wang, Y. Sun, J. W. Burby, L. Ellison, and Y. Zhou, “Canonical symplectic particle-in-cell method for long-term large-scale simulations of the Vlasov–Maxwell equations,” _Nuclear Fusion_ , vol. 56, no. 1, p. 014001, 2016.
* [20] M. Kraus, K. Kormann, P. J. Morrison, and E. Sonnendrücker, “GEMPIC: Geometric ElectroMagnetic Particle-In-Cell Methods,” _Journal of Plasma Physics_ , vol. 83, no. 4, 2017.
* [21] P. J. Morrison, “Structure and structure-preserving algorithms for plasma physics,” _Physics of Plasmas_ , vol. 24, no. 5, p. 055502, 2017.
* [22] A. S. Glasser and H. Qin, “The geometric theory of charge conservation in particle-in-cell simulations,” _Journal of Plasma Physics_ , vol. 86, no. 3, p. 835860303, 2020.
* [23] ——, “A gauge-compatible Hamiltonian splitting algorithm for particle-in-cell simulations using finite element exterior calculus,” _Journal of Plasma Physics_ , vol. 88, no. 2, p. 835880202, 2022.
* [24] A. Stern, “Geometric Discretization of Lagrangian Mechanics and Field Theories,” Ph.D. dissertation, California Institute of Technology, Pasadena, California, 2009.
* [25] Bo He and F. Teixeira, “Sparse and explicit FETD via approximate inverse Hodge (mass) matrix,” _IEEE Microwave and Wireless Components Letters_ , vol. 16, no. 6, pp. 348–350, Jun. 2006.
* [26] B. He and F. L. Teixeira, “Differential Forms, Galerkin Duality, and Sparse Inverse Approximations in Finite Element Solutions of Maxwell Equations,” _IEEE Transactions on Antennas and Propagation_ , vol. 55, no. 5, pp. 1359–1368, May 2007.
* [27] J. Kim and F. L. Teixeira, “Parallel and Explicit Finite-Element Time-Domain Method for Maxwell’s Equations,” _IEEE Transactions on Antennas and Propagation_ , vol. 59, no. 6, pp. 2350–2356, Jun. 2011.
* [28] F. L. Teixeira, “Differential Forms in Lattice Field Theories: An Overview,” _ISRN Mathematical Physics_ , vol. 2013, pp. 1–16, Feb. 2013\.
* [29] M. Hano, “Generalized time-domain method for solution of maxwell’s integral equations,” _AIP Conference Proceedings_ , vol. 391, no. 1, pp. 197–202, 1997.
* [30] J. Cole, “A high-accuracy realization of the Yee algorithm using non-standard finite differences,” _IEEE Transactions on Microwave Theory and Techniques_ , vol. 45, no. 6, pp. 991–996, 1997.
* [31] Z. Chen and S. Luo, “Generalization of the Finite-Difference-Based Time-Domain Methods Using the Method of Moments,” _IEEE Transactions on Antennas and Propagation_ , vol. 54, no. 9, pp. 2515–2524, 2006\.
* [32] F. L. Teixeira and S. Member, “Time-Domain Finite-Difference and Finite-Element Methods for Maxwell Equations in Complex Media,” _IEEE Transactions on Antennas and Propagation_ , vol. 56, no. 8, pp. 2150–2166, 2008.
* [33] A. Stern, Y. Tong, M. Desbrun, and J. E. Marsden, “Geometric Computational Electrodynamics with Variational Integrators and Discrete Differential Forms,” in _Geometry, Mechanics, and Dynamics: The Legacy of Jerry Marsden_ , ser. Fields Institute Communications, D. E. Chang, D. D. Holm, G. Patrick, and T. Ratiu, Eds. New York, NY: Springer New York, 2015, pp. 437–475.
* [34] J. E. Marsden and A. Weinstein, “The Hamiltonian Structure of the Maxwell-Vlasov Equations,” _Physica D_ , vol. 4, no. 3, p. 394, 1982\.
* [35] G. Strang, “On the Construction and Comparison of Difference Schemes,” _SIAM Journal on Numerical Analysis_ , vol. 5, no. 3, pp. 506–517, 1968.
* [36] D. N. Arnold, “Periodic Table of the Finite Elements.” [Online]. Available: http://www-users.math.umn.edu/~arnold/femtable/index.html
* [37] J. Lohi and L. Kettunen, “Whitney forms and their extensions,” _Journal of Computational and Applied Mathematics_ , vol. 393, p. 113520, 2021.
* [38] H. F. Trotter, “On the Product of Semi-Groups of Operators,” _Proceedings of the American Mathematical Society_ , vol. 10, no. 4, pp. 545–551, 1959.
* [39] T. Huckle, “Approximate Sparsity Patterns for the Inverse of a Matrix and Preconditioning,” Technische Universitat Munchen, Tech. Rep. TUM-I9829, 1998.
* [40] L. N. Trefethen and D. Bau, _Numerical linear algebra_. Philadelphia: Society for Industrial and Applied Mathematics, 1997.
* [41] M. Shariff, “A constrained conjugate gradient method and the solution of linear equations,” _Computers & Mathematics with Applications_, vol. 30, no. 11, pp. 25–37, Dec. 1995.
|
# A One-Shot Reparameterization Method for Reducing the Loss of Tile Pruning
on DNNs
Yanchen Li1, Qingzhong Ai2, and Fumihiko Ino1 1: Osaka University, 1-5
Yamadaoka, Suita, Osaka 565-0871, Japan, {yc-li<EMAIL_ADDRESS>2:
University of Electronic Science and Technology of China, Chengdu, China,
<EMAIL_ADDRESS>
###### Abstract
Recently, tile pruning has been widely studied to accelerate the inference of
deep neural networks (DNNs). However, we found that the loss due to tile
pruning, which can eliminate important elements together with unimportant
elements, is large on trained DNNs. In this study, we propose a one-shot
reparameterization method, called TileTrans, to reduce the loss of tile
pruning. Specifically, we repermute the rows or columns of the weight matrix
such that the model architecture can be kept unchanged after
reparameterization. This repermutation realizes the reparameterization of the
DNN model without any retraining. The proposed reparameterization method
combines important elements into the same tile; thus, preserving the important
elements after the tile pruning. Furthermore, TileTrans can be seamlessly
integrated into existing tile pruning methods because it is a pre-processing
method executed before pruning, which is orthogonal to most existing methods.
The experimental results demonstrate that our method is essential in reducing
the loss of tile pruning on DNNs. Specifically, the accuracy is improved by up
to 17% for AlexNet while 5% for ResNet-34, where both models are pre-trained
on ImageNet.
###### Index Terms:
Deep learning, tile pruning, reparameterization
## I Introduction
Deep neural networks (DNNs) have demonstrated excellent performance on a wide
variety of tasks, such as computer vision [1, 2], speech recognition [3, 4],
and robotics [5, 6]. One backbone behind the growth in the abilities of DNNs
is an increasing number of trainable parameters. For instance, a recent state-
of-the-art model, CogView [7], achieved superior performance on text-to-image
generation tasks, thanks to its almost four billion parameters. However,
training a model with many parameters is time- and resource-intensive.
Therefore, the need for accelerating the computation on DNN is also
increasing. Nowadays, several mainstream acceleration methods, such as pruning
[8, 9, 10], mixed precision [11], and parallelization [12] have recently
attracted significant attention. Among these methods, mixed precision and
parallelization focus on the acceleration in the training phase of DNNs,
whereas pruning concentrates on the inference time.
This study focuses on pruning for the increasing need to accelerate the
application of DNNs. Pruning accelerates DNNs by deleting the trivial elements
in the weight matrix. The importance of each element is evaluated by the
importance score [13] so that the elements with small importance scores will
be deleted. Furthermore, to quantify the result of pruning, we calculate the
loss of pruning based on the importance score of the deleted elements,
generally summation. As the goal of pruning, we hope to get a minor pruning
loss, which brings higher inference accuracy, with a certain sparsity of
pruning.
However, there is a tradeoff between pruning loss and acceleration. On the one
hand, pruning loss increases when pruning sparsity increases. On the other
hand, acceleration from pruning is proportional to the sparsity. Consequently,
various pruning methods [14, 15, 16, 17] managed to better balance the
tradeoff between the pruning loss and acceleration.
Among those pruning methods, tile pruning [18] achieves a better tradeoff.
Tile pruning deletes the elements in the weight matrix by tiles while other
structured pruning methods eliminate the elements by a larger shape, such as a
row. The advantages of deleting elements by tiles can be reflected in the
following two aspects. First, owing to the smaller pruning unit of tiles, the
pruning loss is relatively smaller than other structured pruning methods.
Second, the computation for pruned tiles can be skipped with a small overhead,
and thus the pruned DNNs are accelerated. However, we found that the trained
DNNs empirically do not always fit well with tile pruning. Those important
elements are usually eliminated with unimportant elements, leading to large
pruning loss.
To this end, we propose a reparameterization method, called TileTrans, based
on the DNNs model for tile pruning. Instead of changing the pruning strategy,
TileTrans reparameterizes the DNNs model before the pruning. Specifically, we
use permutation to separate elements by their importance so that there are
either important or unimportant elements in the same tile. In this way, we
preserve more important elements after tile pruning and achieve minor pruning
loss. TileTrans is user-friendly in the sense that we reparameterize the DNN
model by a trick of matrix multiplication instead of retraining. Besides, we
keep the model architecture unchanged and therefore introduce no extra
computation for inference. Additionally, our algorithm is a straightforward
heuristic algorithm, which means that the algorithm’s efficiency is
guaranteed.
We evaluated TileTrans on two classical DNN models: AlexNet [19] and ResNet-34
[20], where both models are first pretrained on the ImageNet dataset [21]. We
reparameterized the DNN models by TileTrans before tile pruning on the models.
The experimental results show that TileTrans reduces the pruning loss and
improves accuracy by 17% on AlexNet and 5% on ResNet-34, respectively.
The main contributions of our study are as follows:
* •
We proposed a reparameterization method, named TileTrans, to permute the
weight matrix such that important and unimportant elements are permuted in the
same tile before tile pruning, respectively. As a result, we preserve more
important elements after tile pruning, and thus we reduce the pruning loss.
* •
We reparameterized DNN models in a novel way based on a trick of matrix
multiplication. Compared with other reparameterization methods, our method is
user-friendly because TileTrans requires no extra retraining on DNN models.
Furthermore, We kept the architecture of the model unchanged so that no
additional computation occurs after TileTrans.
* •
We proposed a heuristic method to build an appropriate permutation for
TileTrans to avoid time-consuming searching for the optimal permutation.
Besides, we verified that this heuristic method does reduce the loss of tile
pruning experimentally.
Eventually, we deduce that TileTrans is capable of building DNN models that
are more adaptable to tile pruning.
## II Related work
Notably, TileTrans can be fused with any tile pruning method seamlessly.
Various pruning strategies [22, 18, 14] have been proposed to reduce the loss
of tile pruning. For example, Guo et al. [18] deleted the tiles by the number
of important elements in the tile. TileTrans reparameterizes DNN before the
procedure of pruning; thus, it can be used together with these pruning
strategies to improve the performance of tile pruning.
Compared with the major reparameterization methods on DNNs, TileTrans needs no
extra training. Reparameterization has been used to change the data structure
of DNN for many purposes [23, 24, 25, 26]. Specifically, reparameterization is
also used to build a model with structural sparsity. Most reparameterization
methods design a particular loss function and train the model for a large
number of iterations. For example, Wen et al. [27] introduced the Lasso
regression to the loss function and trained a model with structural sparsity;
Ding et al.[28] proposed the ResRep, which punished unimportant channels in
the weight matrices during training. These studies achieved outstanding
performance in building structural sparsity, but they required extra training
on the models. In contrast, our TileTrans only reparameterizes DNN by matrix
multiplication without any training.
Some previous works also reduced the pruning loss by transforming the weight
matrix. However, these methods changed the model architectures; thus,
increasing the cost of computation. Liu et al. [29] added a transformation to
each layer, by which they transformed the format of each weight matrix into
the frequency domain. Guo et al. [18] duplicated each weight matrix and saved
the important elements in one of them. These methods improved the accuracy of
the pruned models but called for extra computation compared with the original
architecture. In contrast, TileTrans transforms weight matrices without
modifying the architecture. In other words, we add no extra computation to the
model.
## III Preliminary
Before introducing TileTrans, we have to present the preliminary about tile
pruning because we specialize TileTrans for reducing the loss of tile pruning.
Therefore, we will introduce the optimization objective and algorithm of tile
pruning as follows.
### III-A Importance Score
The importance score is a criterion used to evaluate the importance of an
element in the weight matrix. Given an element with weight $w$, we calculate
the importance score as $\mathcal{E}(w)$, where
$\mathcal{E}:\mathbb{R}\to\mathbb{R}$ denotes the criterion of importance
score. There are many criteria in previous works, such as calculating the
Hessian matrix [30] and deriving from Taylor expansion [13].
We can quantitatively define the loss of pruning with a certain importance
score. Generally, the pruning loss is an important score summation of the
deleted elements. Given the weights of the model as $\mathbf{W}$ and targeted
sparsity $s$, weights are pruned by $\mathit{f}(\mathbf{W},s;\mathcal{E})$,
where $f$ deletes the elements in $\mathbf{W}$ to the sparsity of $s$
depending on $\mathcal{E}$. Furthermore, pruning can be described as an
optimization problem with the objective as follows:
$\displaystyle\min_{f}$
$\displaystyle\mathcal{L}(\mathbf{W},s;\mathcal{E})=\left|\sum_{w}^{w\in\mathbf{W}}\mathcal{E}(w)-\sum_{w}^{w\in\mathit{f}(\mathbf{W},s;\mathcal{E})}\mathcal{E}(w)\right|$
(1) s.t. $\displaystyle\mathbf{W}\in\mathbb{W},0\leq s\leq 1,$
where $\mathcal{L}$ denotes the loss function of pruning and $\mathbb{W}$ is
the set of all possible weights. In other words, the loss of pruning can be
minimized by deleting the elements with the smallest importance scores. In
tile pruning, there are certain restrictions on the shape of the pruning unit.
Therefore, we evaluate the average importance score of the elements in the
tiles for tile pruning and delete the most unimportant tiles.
### III-B Tile Shape
The shape of tile pruning depends on the devices or libraries. For example,
Chen et al. [17] pruned the weights in the shape of the loading unit for the
tensor core so that the computation for pruned elements can be skipped with a
small overhead. Guo et al. [18] used a shape of $128\times 1$, because CUTLASS
[31] loads $128\times 1$ elements at once from the global memory to the shared
memory on the GPU system. Given the tile shape of $a\times b$, tile pruning
sometimes is called the “block pruning” [14] when $a=b$. Note that we regard
the block pruning as tile pruning in this paper. Especially, tile pruning in
the shape of $1\times 1$ is called unstructured pruning [32]. The pruning loss
of unstructured pruning is usually lower than that of tile pruning, where
$a>1$ and $b>1$. Unstructured pruning precisely deletes the unimportant
elements while tile pruning eliminates the important elements together with
the unimportant elements in the same tile. Therefore, unstructured pruning is
usually used as the baseline for tile pruning when evaluating the accuracy of
pruned models.
(a)
(b)
(c)
Figure 1: Examples for pruning the $4\times 4$ weight matrix to the sparsity
of 0.5 with the unstructured pruning (a), the $2\times 2$ tile pruning before
reparameterization (b), and the $2\times 2$ tile pruning after
reparameterization (c), respectively. We fill the background with different
colors for the corresponding tiles in (b) and (c) for easy observation. The
numbers in cells are the importance scores for elements. The values of pruning
loss are shown at the bottom of the figures.
### III-C Algorithm of Tile Pruning
For tile pruning, we evaluate the importance of every tile and delete
unimportant tiles according to the importance score, which is calculated from
the elements in the tile. We show details in Algorithm 1. From lines 2 to 6,
we traverse all weight matrices in the model and calculate the importance
scores for every tile. The importance score of a tile is defined as the
average of the importance scores of the elements in that tile. After
evaluating for tiles, we set a threshold according to the sparsity at line 7.
Because the sparsity is the percentage of zero elements in all matrices, we
delete unimportant tiles whose importance scores are smaller than the
threshold. Finally, we retrain the model to recover the accuracy for inference
after tile pruning.
Algorithm 1 TilePruning($\mathbf{W}$, $\mathcal{E}$, $a$, $b$, $s$)
1:Weight matrices of the model $\mathbf{W}$, metrics for importance score
$\mathcal{E}$, height of tile $a$, width of tile $b$, and targeted sparsity
$s$.
2:Pruned weight matrices $\mathbf{W}$.
3:$\mathbf{V}\leftarrow\emptyset$
4:foreach $W\in\mathbf{W}$ do $\triangleright$ Traverse the weights in the
layers
5: foreach $T\in W$ do
6: $\mathbf{V}\leftarrow\mathbf{V}\cup\\{v\leavevmode\nobreak\
|\leavevmode\nobreak\ v=\sum_{i}^{a}\sum_{j}^{b}\mathcal{E}(T_{i,j})/ab\\}$
7: end for
8:end for
9:Set threshold $t$ such that $|\\{v\leavevmode\nobreak\ |\leavevmode\nobreak\
v\in\mathbf{V},v<t\\}|/|\mathbf{V}|=s$
10:Delete the tile with importance score smaller than $t$.
## IV Weights Transformation
Naturally, the distribution of the elements in the weight matrices is usually
unfriendly to tile pruning. The important and unimportant elements are mixed
in the same tile. We show the examples in Figs. 1(a) and 1(b). For
unstructured pruning in Fig. 1(a), unimportant elements are precisely deleted,
and thus the loss for unstructured pruning is the smallest. For tile pruning
in Fig. 1(b), the elements that are alive in unstructured pruning are deleted.
Consequently, the loss for tile pruning is more significant than for
unstructured pruning. To separate the important and unimportant elements, we
propose TileTrans that reparameterizes DNN before the pruning by matrix
transformation.
### IV-A Idea of Permutation
The loss for tile pruning can be reduced by matrix transformation. We
transform the weight matrices to a better format, where the important elements
are combined together in the same tiles as far as possible. For example, in
Fig. 1(c), we switch the first row of the matrix with the third row.
Accordingly, more important elements are permuted in the same tiles than in
Fig. 1(b). Finally, the loss for the $2\times 2$ tile pruning on the
reparameterized weight matrix becomes smaller than that on the original weight
matrix.
The critical point is that the transformation must keep the model’s output
unchanged. The output of DNN is changed if we transform the weight matrix. One
option to recover the output is adding extra calculation after the inference.
However, we should avoid increasing the computational amount for inference;
thus, we keep the model architecture unchanged. Without modifying the
architecture, we managed to recover the model’s output by a trick of matrix
multiplication.
### IV-B Algorithm of TileTrans
We designed a reparameterization method based on matrix multiplication. The
main components of the DNN models are the linear layer [33] and the
convolutional layer [34], which can be represented as matrix multiplication.
Furthermore, the DNN model consists of multiple layers, where the calculations
are transitive. For example, Fig. 2 shows a partial architecture of ResNet,
where the output of a layer is the input of another layer. Utilizing this
transitivity of the calculations in DNN, we can reparameterize DNN without any
training by the properties of matrix multiplication.
Figure 2: Partial architecture of ResNet. A block denotes a layer with
weights. An arrow represents a data dependence inherent in the data flow.
The reparameterization procedure is a series of operations that transform two
layers. For a single operation, we first build an invertible transformation
$R$ for the previous layer and build its inverse $R^{-1}$ for the next layer.
In this way, the output of the transformed layer can be recovered by
calculating the next layer. We take Fig. 2 as an example to show how this
trick works. Suppose that the weights of the layer $i$ is $W_{i}$, the input
is $I_{i}$, and the output is $O_{i}$. The calculations in layer $A$ and $B$
then can be given as
$\begin{split}O_{A}=I_{A}W_{A}^{T},\end{split}$ (2) $O_{B}=O_{A}W_{B}^{T}.$
(3)
After that, we transform these weights by a matrix transformation $R$ as
follows
$\begin{split}O_{A}R^{T}=I_{A}(RW_{A})^{T},\end{split}$ (4)
$O_{B}=O_{A}R^{T}(W_{B}R^{-1})^{T}.$ (5)
In other words, we transform the previous layer and use the next layer to
recover the output. Therefore, the output of the two layers is unchanged, even
though we transform their weights. Finally, we reparameterize the entire DNN
model based on this trick. However, the restraint for the reparameterization
becomes more strict when the model architecture is more complex.
Some layers in the model have to share the same matrix transformation when
there are residuals in the model. Residual [20] is a calculation that reuses
the output of the previous layer as the input of the deeper layer. For
example, in Fig. 2, the output of layer A is the input of layers B and D.
Before calculating the matrix multiplication of layer D, we first calculate
the sum of $O_{A}$ and $O_{C}$ as follows:
$O_{A}=I_{A}W_{A}^{T},$ (6) $O_{B}=O_{A}W_{B}^{T},$ (7)
$O_{C}=O_{B}W_{C}^{T},$ (8) $O_{D}=(O_{A}+O_{C})W_{D}^{T}.$ (9)
Supposing the transformation for layer $i$ as $R_{i}$, the calculations for
the transformed layers are given as
$O_{A}R_{A}^{T}=I_{A}(R_{A}W_{A})^{T},$ (10)
$O_{B}R_{B}^{T}=O_{A}R_{A}^{T}(R_{B}W_{A}R_{A}^{-1})^{T},$ (11)
$O_{C}R_{C}^{T}=O_{B}R_{B}^{T}(R_{C}W_{C}R_{B}^{-1})^{T},$ (12)
$O_{D}=(O_{A}R_{A}^{T}+O_{C}R_{C}^{T})(W_{D}R_{C}^{-1})^{T},$ (13)
where $O_{D}=(O_{A}+O_{C})W_{D}^{T}$ only if $R_{A}=R_{C}$. Thus, layers A and
C must share the same transformation. The example of Fig. 2 is simple, but the
actual model is usually more sophisticated.
There may be groups of layers that must share the same transformation. For
simplicity, we define the layers that have to share the same transformation as
the layer group. This layer group is built according to the relations of
parents and children, which can be defined as follows.
###### Definition 1 (Parents and Children).
Given a layer A, parents of A are the layers that send their output to A as
the inputs of A. Meanwhile, children of A are the layers that receive the
output of A as their inputs.
Specifically, we present a theorem as follows.
###### Theorem 1.
Given a layer A in the DNN model, the parents of A must share the same
transformation.
###### Proof.
According to Eq. (13), assuming the weight is $W$, the transformation is $R$,
and the output is $O$, if there are $n$ parents for a layer, the output of the
layer is given by
$O=(O_{1}R_{1}^{T}+O_{2}R_{2}^{T}+...+O_{n}R_{n}^{T})(WR^{-1})^{T},$ (14)
where $O_{i}$ denotes the output of the $i$-th parent, and $R_{i}$ denotes the
transformation of the $i$-th parent. Then
$O=(O_{1}+O_{2}+...+O_{n})W^{T},$ (15)
only if $R_{1}=R_{2}=...=R_{n}=R$. ∎
Algorithm 2 TileTrans($\mathbf{M}$, $\mathcal{E}$)
1:DNN model $\mathbf{M}$ and metrics for importance score $\mathcal{E}$.
2:Reparameterized DNN model $\mathbf{M}$.
3:$\mathbb{G}\leftarrow\emptyset$ $\triangleright$ Initialize the layer group
4:foreach $L\in\mathbf{M}$ do
5: $\mathbf{G}\leftarrow$ the parents of $L$
6: $\mathbb{G}\leftarrow\mathbb{G}\cup\mathbf{G}$
7:end for
8:$\hat{\mathbb{G}}\leftarrow\\{\mathbf{G}\leavevmode\nobreak\
|\leavevmode\nobreak\
\mathbf{G}\leftarrow\mathbf{G}_{1}\cup\mathbf{G}_{2}\text{ if
}\mathbf{G}_{1}\cap\mathbf{G}_{2}\neq\emptyset,\text{ and
}\mathbf{G}_{1},\mathbf{G}_{2}\in\mathbb{G}\\}$ $\triangleright$ Merge the
intersecting groups
9:foreach $\mathbf{G}\in\hat{\mathbb{G}}$ do
10: $R\leftarrow$ BuildTrans($\mathbf{G},\mathcal{E}$) $\triangleright$ Build
the transformation
11: foreach $L\in\mathbf{L}$ do
12: Transform the weights of $L$ by $R$
13: end for
14: Transform the weights in the children of the layers in $\mathbf{G}$ by
$R^{-1}$
15:end for
According to Theorem 1, we design an algorithm for TileTrans, which is shown
in Algorithm 2. From lines 2 to 4, we first find the sets of weights that must
share the same transformation. Notably, some subsets in $\mathbb{G}$ may
intersect each other. For example, given two sets of layers
$\mathbf{G}_{0}\text{ and }\mathbf{G}_{1}$, $\mathbf{G_{0}}=\\{L_{A},L_{B}\\}$
and $\mathbf{G_{1}}=\\{L_{B},L_{C}\\}$, all $L_{A}$, $L_{B}$, and $L_{C}$ must
share the same transformation for their weights. Therefore, we merge the
subsets intersecting each other at line 6. According to
$\mathbf{G}\in\mathbb{G}$, we build the appropriate transformation $R$, which
is introduced later, for the weights at line 8. To obtain the correct output,
we transform the weights in the children of the layers in $\mathbf{G}$ by
$R^{-1}$ at line 12.
The modern DNN models improve their performance by introducing nonlinear
layers, such as ReLU [35], pooling [34], and normalization [36]. For such
nonlinear layers, we restraint the transformation $R$ to only the permutation
of rows in Algorithm 2. This restraint guarantees the correctness of the model
because permutation is a linear transformation that only changes the location
of the elements. By contrast, a linear transformation that changes the scale
of elements does not have this guarantee because it fails to recover the
output of the model.
Additionally, the permutation of columns works in the TileTrans, but Algorithm
2 needs to be “reversed.” In Algorithm 2, we define the permutation of rows
$R$ for the parents at line 8 and permute the weight matrices at line 10.
After that, we use the permutation of columns $R^{-1}$ to recover the output
for the children at line 12. Therefore, TileTrans in rows is a procedure from
parents to children. Meanwhile, TileTrans in columns is a procedure from
children to parents. Thus, for the permutation of columns, we build the
transformation for children and then recover the output by the calculations in
parents.
(a)
(b)
(c)
(d)
(e)
(f)
Figure 3: Loss of tile pruning on AlexNet (a, b, and c) and ResNet-34 (d, e,
and f) that are pre-trained on the ImageNet for different sparsity. The tile
shape in Figs. (a) and (d) are $256\times 1$, that in Figs. (b) and (e) are
$16\times 16$, and that in Figs. (c) and (f) are $32\times 32$. Specially, we
define the result for unstructured tile pruning, i.e., the $1\times 1$ tile
pruning, as the baseline for tile pruning. Besides, the legend with “trans”
denotes the result for the model reparameterized by TileTrans. Meanwhile,
“diff” denotes the difference between tile pruning and unstructured pruning,
which is defined as Eq. (16). For simplicity, we normalize the pruning loss
and difference by 0-1 normalization.
### IV-C Heuristic Permutation
Ideally, if all-important elements were aggregated in the same tile, all
important elements then exist after tile pruning. That is, the result of tile
pruning becomes the same as that of unstructured pruning. TileTrans aims to
make the result of tile pruning as close as possible to that of unstructured
pruning. Because the loss for unstructured pruning is the smallest, we assume
the loss for unstructured pruning is the baseline. Based on this baseline, we
give the optimization problem for the TileTrans as follows:
$\displaystyle\min\left|\sum_{w}^{w\in\mathit{f}_{u}(\mathbf{W},s;\mathcal{E})}\mathcal{E}(w)-\sum_{w}^{w\in\mathit{f}_{t}(\mathcal{R}(\mathbf{W}),s;\mathcal{E})}\mathcal{E}(w)\right|,$
(16) $\displaystyle\textrm{s.t.}\mathbf{W}\in\mathbb{W},0\leq s\leq 1,$
where $\mathit{f}_{u}:\mathbb{W}\to\mathbb{W}$ denotes the unstructured
pruning method, $\mathit{f}_{t}:\mathbb{W}\to\mathbb{W}$ denotes the tile
pruning method, and $\mathcal{R}:\mathbb{W}\to\mathbb{W}$ denotes the
reparameterization method. According to Eq. (16), given specific sparsity $s$,
we need to find the permutation that minimizes the difference between tile
pruning and unstructured pruning methods.
However, searching for the optimal solution for Eq. (16) is time consuming.
The most intuitive method to solve Eq. (16) is to try all possible
permutations and verify the effect, which costs too much time. Given a model
with $k$ layers and each layer with weights of $n$ rows, the time complexity
to traverse all possible permutations is $O((n!)^{k})$. Therefore, we propose
a heuristic method to build a relatively good permutation without searching.
Algorithm 3 BuildTrans($\mathbf{G}$, $\mathcal{E}$)
1:Group of layers $\mathbf{G}$ and metrics for importance score $\mathcal{E}$.
2:Permutation $R$.
3:$\mathbf{W}\leftarrow$ all weights of the layers in $\mathbf{G}$
4:$\hat{W}\leftarrow$ concatenate $W\in\mathbf{W}$
5:$\hat{W}_{R}\leftarrow\hat{W}$
6:Sort the rows in $\hat{W}_{R}$ $\triangleright$ According to $\mathcal{E}$
7:Build $R$ such that $R\hat{W}=\hat{W}_{R}$
We sort the rows according to their average importance scores, which are shown
in Algorithm 3. In line 1, we concatenate the weights that share the same
transformation in the dimension of rows to evaluate them together. These
weights can be concatenated because they have the same number of rows. After
that, we sort the rows of the concatenated weights according to the average
importance score, which is calculated by the metrics $\mathcal{E}$. Finally,
we build the transformation $R$ such that $R\hat{W}=\hat{W}_{R}$, where
$\hat{W}$ denotes the concatenated weights, and $\hat{W}_{R}$ denotes the
sorted concatenated weights.
The sorting procedure is based on two assumptions:
###### Assumption 1.
The averages of the rows follow the normal distribution.
###### Assumption 2.
The variance of importance scores for elements in a row is small.
We make the assumptions depending on previous studies and our experience.
Assumption 1 is inspired by previous studies on channel pruning [37], which
prunes the elements in rows. In channel pruning, the importance score for a
row is evaluated by the average of the importance scores for the elements in
the row. After channel pruning, rows are deleted without a few important rows,
meaning that a few important rows are in the matrix, whereas the other rows
are unimportant. Similarly, there are few large values in the normal
distribution, while the others are smaller. We use the normal distribution to
approximate the distribution of the importance scores for the rows. For
Assumption 2, we make it empirically because we observed that importance
scores for the elements in a row are usually all large when the row is
important. The unimportant and important elements are naturally separated
after sorting rows depending on the assumptions. Finally, the important
elements are preserved after tile pruning because the important elements are
in the same tiles. In other words, we reduce the loss of tile pruning.
(a)
(b)
(c)
(d)
(e)
(f)
Figure 4: Accuracy of the pruned AlexNet (a, b, and c) and ResNet-34 (d, e,
and f) that are pre-trained on the ImageNet for different sparsity. The tile
shape in Figs. (a) and (d) are $256\times 1$, that in Figs. (b) and (e) are
$16\times 16$, and that in Figs. (c) and (f) are $32\times 32$. Specially, we
define the result for unstructured tile pruning, i.e., the $1\times 1$ tile
pruning, as the baseline for tile pruning. Besides, the legend with “trans”
denotes the result for the model reparameterized by TileTrans.
## V Experiment
In the experiments, we evaluated the proposed TileTrans on the following
metrics:
* •
Loss. Pruning loss is our direct optimization objective. Therefore, we first
show the pruning loss before and after the reparameterization of TileTrans. At
the same time, we calculate the difference between tile pruning and
unstructured pruning for the following two reasons. First, the difference in
the pruning loss before and after TileTrans is negligible; therefore, we
illustrate the normalized difference in figures for better observation.
Second, we designed an optimization objective for TileTrans as Eq. (16), which
is the difference between tile pruning and unstructured pruning; thus, we
calculate this difference to verify the heuristic permutation does reduce the
pruning loss.
* •
Accuracy. The accuracy of the model is the most intuitive metric that
evaluates the performance of the model. We are unaware of the performance of
the pruned models only by evaluating pruning loss. Besides, the difference
between the pruning loss with different tile shapes on the same model is
negligible, so we cannot assess how tile shape affects the performance of
TileTrans only by loss. Therefore, we calculate the accuracy of the pruned
models to evaluate TileTrans further.
We design sets of controlled experiments on AlexNet and ResNet-34 to evaluate
the performance of TileTrans. For one set, we first reparameterized the model
by TileTrans and then pruned the model; for another set, we pruned the model
directly. For simplicity, we choose the L1 norm as the criteria of importance
score. Finally, we calculated the pruning loss and model accuracy of the
pruned models.
Besides, we evaluate TileTrans on tile pruning of different tile shapes to
assess how tile shape affects the performance of TileTrans. Expressly, we set
the tile shape as $256\times 1$, $16\times 16$, or $32\times 32$. For
$256\times 1$ and $16\times 16$, the sizes are the same, while the aspect
ratios are different; for $16\times 16$ and $32\times 32$, the aspect ratios
are the same, while the sizes are different. Therefore, we know how the size
and aspect ratio of tile affect the performance of TileTrans by comparing the
result of these tile shapes.
We realized TileTran with PyTorch and retrained the models on eight GeForce
GTX 1080 GPUs. Our code is available on https://github.com/LoonLi/TileTrans.
### V-A Loss
The results of pruning loss are shown in Fig. 3. Specially, we also illustrate
the result of unstructured pruning, i.e., the $1\times 1$ pruning, in Fig. 3
as the baseline of tile pruning. Accordingly, with specific sparsity, the
difference between tile pruning and unstructured pruning is the distance
between the point of tile pruning and unstructured pruning.
TileTrans reduced most pruning loss at the sparsity of about 60; however, it
reduced little loss when the sparsity was sufficiently small or large. This is
due to the minority tiles with all unimportant or important elements, which
exist naturally in weight matrices. These tiles with all unimportant elements
are deleted when the sparsity is small. In contrast, those with all important
elements are still preserved when the sparsity is large. Consequently,
TileTrans avoids degrading the pruning loss when the sparsity is sufficiently
small or large. Meanwhile, those tiles with important and unimportant elements
are permuted by TileTrans; thus, the pruning loss is reduced when the sparsity
is moderate.
TileTrans reduced less pruning loss on ResNet-34 than on AlexNet. This is
attributed to the following reasons: more complex architecture and smaller
weight matrices. First, the architecture of ResNet-34 is more complex than
that of AlexNet because there are multiple connections of residual in
ResNet-34. According to Theorem 1, layers with residuals must share the
transformation with other layers that send the residual to them. In contrast,
AlexNet is a simple architecture in which no layer needs to share the same
transformation. Thus, searching for good transformations for ResNet-34 is more
difficult. Second, the weight matrices of ResNet-34 are small for the
permutation. Specifically, the heights of most weight matrices in ResNet-34
are less than 512. Meanwhile, there are large weight matrices with a height of
4096 in AlexNet. In other words, we have fewer choices to permute the weight
matrices in rows on ResNet-34 than on AlexNet.
Pruning loss is the most direct metric to evaluate TileTrans, but it is not
intuitive enough for understanding the performance of the models.
Specifically, we are unaware of the difference between different tile shapes
since their results are very similar in the same model. Therefore, we also
evaluate TileTrans on the metric of the model accuracy.
### V-B Accuracy
We define accuracy as the probability that models correctly infer the types of
images in the test data. The accuracy $P_{a}$ is calculated as
$P_{a}=N_{c}/N_{t}$, where $N_{t}$ is the amount of test data $N_{t}$ and
$N_{c}$ is the amount of correctly inferred images.
TileTrans improved the accuracies of both AlexNet and ResNet-34 after tile
pruning, which is a natural result because TileTrans reduced the pruning loss.
We show results of model accuracy in Fig. 4. Specifically, the accuracies of
both AlexNet and ResNet-34 were improved most under the tile shape of
$32\times 32$. In detail, the pruned AlexNet was improved by up to 17% at the
sparsity of 60%; the pruned ResNet-34 was improved by up to 5% at the sparsity
of 70%. Specially, the accuracy of the pruned AlexNet decreased after
TileTrans at the sparsity of 80% under the tile shape of $16\times 16$.
However, in this case, the pruning loss was reduced according to Fig. 3(b).
Consequently, this decrease is not caused by increased pruning loss and may be
improved by changing the criteria of importance score.
Between the tile shapes of $256\times 1$ and $16\times 16$, TileTrans improved
the accuracy more for $256\times 1$ on AlexNet while for $16\times 16$ on
ResNet-34. According to this result, we draw two conclusions about the
performance of TileTrans. On the one hand, TileTrans improve more accuracy for
strip-shaped tile when the weight matrix is large enough. On the other hand,
TileTrans improves more accuracy for square-shaped tiles when the weight
matrix is too small. ResNet-34 consists of multiple convolutional layers with
heights smaller than $512$. As a result, TileTrans reduced little pruning loss
on ResNet-34 for $256\times 1$ because the permutation cannot exchange the
elements among tiles. Meanwhile, AlexNet contains two large linear layers
whose weight matrices are $4096\times 4096$. Therefore, the important and
unimportant elements were combined in different tiles on AlexNet for
$256\times 1$. In this case, the unimportant elements are deleted more
precisely for $256\times 1$ than $16\times 16$.
Between the tile shapes of $16\times 16$ and $32\times 32$, TileTrans improved
the accuracy more for $32\times 32$ on both models. Consequently, we conclude
that TileTrans improves accuracy more for larger tile sizes. The probability
of a tile in the weight matrix filled with important elements is low when the
tile size is large. Meanwhile, TileTrans builds weight matrices with more
tiles that are filled with important elements. In other words, TileTrans
reduces pruning loss more when the tile size is large.
In summary, we conclude three rules for TileTrans as follows:
1. 1.
TileTrans reduces more loss for strip-shaped than square-shaped tiles when the
sizes of tiles are the same, and the weight matrix is large enough for
permutation.
2. 2.
TileTrans reduces more loss for square-shaped than strip-shaped tiles when the
sizes of tiles are the same, and the weight matrix is small.
3. 3.
TileTrans reduces more loss for the tiles with large size.
## VI Conclusion
In this study, we proposed a one-shot reparameterization method, called
TileTrans, to reduce the loss of tile pruning on DNNs by permuting the weight
elements. We heuristically build an appropriate permutation to avoid time-
consuming searching for the optimal permutation. Besides, TileTrans requires
no retraining and keeps the model architecture unchanged. The experiments
demonstrated that TileTrans reduced the loss of tile pruning on AlexNet and
ResNet-34.
## Acknowledgments
This research was in part supported by “Program for Leading Graduate Schools”
of the Ministry of Education, Culture, Sports, Science and Technology, Japan
and the Japan Society for the Promotion of Science KAKENHI grant number
20K21794.
## References
* [1] S. Xiong et al., “Explore Visual Concept Formation for Image Classification,” in Proc. 22nd International Conference on Machine Learning (ICML 2021), pp. 11470–11479, 2021.
* [2] Z. Liu et al., “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the 18th IEEE/CVF International Conference on Computer Vision (ICCV 2021), pp. 10012–10022, 2021.
* [3] V. Popov et al., “Grad-TTS: A Diffusion Probabilistic Model For Text-to-Speech.” ArXiv, 2021. abs/2105.06337.
* [4] D. Min et al., “Meta-StyleSpeech: Multi-Speaker Adaptive Text-to-Speech Generation.” ArXiv, 2021. abs/2106.03153.
* [5] Y. Chebotar et al., “Actionable Models: Unsupervised Offline Reinforcement Learning of Robotic Skills.” ArXiv, 2021. abs/2104.07749.
* [6] N. Sünderhauf et al., “The Limits and Potentials of Deep Learning for Robotics,” The International Journal of Robotics Research, vol. 37, no. 4-5, pp. 405–420, 2018.
* [7] M. Ding et al., “CogView: Mastering Text-to-Image Generation via Transformers.” ArXiv, 2021. abs/2105.13290.
* [8] S. Han et al., “Learning Both Weights and Connections for Efficient Neural Networks.” ArXiv, 2015. abs/1506.02626.
* [9] W. Wang et al., “Accelerate CNNs from Three Dimensions: A Comprehensive Pruning Framework,” in Proc. 22nd International Conference on Machine Learning (ICML 2021), pp. 10717–10726, 2021.
* [10] A. Jordao et al., “Discriminative Layer Pruning for Convolutional Neural Networks,” IEEE Journal of Selected Topics in Signal Processing, vol. 14, no. 4, pp. 828–837, 2020.
* [11] P. Micikevicius et al., “Mixed Precision Training.” ArXiv, 2017. abs/1710.03740.
* [12] J. Dean et al., “Large scale distributed deep networks,” Advances in Neural Information Processing Systems, vol. 25, 2012.
* [13] P. Molchanov et al., “Importance Estimation for Neural Network Pruning,” in Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), pp. 11264–11272, 2019.
* [14] H. Peng et al., “Accelerating Transformer-based Deep Learning Models on FPGAs Using Column Balanced Block Pruning,” in Proc. 22nd International Symposium on Quality Electronic Design (ISQED 2021), pp. 142–148, 2021.
* [15] A. Ren et al., “DARB: A Density-Adaptive Regular-Block Pruning for Deep Neural Networks.,” in Proc. 34th Association for the Advancement of Artificial Intelligence (AAAI 2020), pp. 5495–5502, 2020.
* [16] S. Narang et al., “Block-Sparse Recurrent Neural Networks.” ArXiv, 2017\. abs/1711.02782.
* [17] S. Chen et al., “ET: Re-thinking Self-Attention for Transformer Models on GPUs,” in Proc. 34th International Conference for High Performance Computing, Networking, Storage and Analysis (SC 2021), pp. 1–18, 2021.
* [18] C. Guo et al., “Accelerating Sparse DNN Models without Hardware-Support via Tile-Wise Sparsity,” in Proc. 33rd International Conference for High Performance Computing, Networking, Storage and Analysis (SC 2020), pp. 1–15, 2020.
* [19] A. Krizhevsky et al., “Imagenet Classification with Deep Convolutional Neural Networks,” Advances in Neural Information Processing Systems, vol. 25, pp. 1097–1105, 2012.
* [20] K. He et al., “Deep residual Learning for Image Recognition,” in Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), pp. 770–778, 2016.
* [21] J. Deng et al., “Imagenet: A Large-Scale Hierarchical Image Database,” in Proc. 24th IEEE Conference on Computer Vision and Pattern Recognition (CPVR 2009), pp. 248–255, 2009.
* [22] S. Lin et al., “Towards Optimal Structured CNN Pruning via Generative Adversarial Learning,” in Proc. 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), pp. 2790–2799, 2019.
* [23] X. Ding et al., “Acnet: Strengthening the Kernel Skeletons for Powerful CNN via Asymmetric Convolution Blocks,” in Proc. 17th IEEE/CVF International Conference on Computer Vision (ICCV 2019), pp. 1911–1920, 2019\.
* [24] X. Ding et al., “Diverse Branch Block: Building a Convolution as an Inception-like Unit,” in Proc. 34th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2021), pp. 10886–10895, 2021.
* [25] X. Ding et al., “Repmlp: Re-parameterizing convolutions into fully-connected layers for image recognition.” ArXiv, 2021. abs/2105.01883.
* [26] X. Ding et al., “Repvgg: Making Vgg-Style Convnets Great Again,” in Proc. 34th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2021), pp. 13733–13742, 2021.
* [27] W. Wen et al., “Learning Structured Sparsity in Deep Neural Networks,” Advances in Neural Information Processing Systems, vol. 29, pp. 2074–2082, 2016.
* [28] X. Ding et al., “Resrep: Lossless CNN Pruning via Decoupling Remembering and Forgetting,” in Proc. 18th IEEE/CVF International Conference on Computer Vision (ICCV 2021), pp. 4510–4520, 2021.
* [29] Z. Liu et al., “Frequency-Domain Dynamic Pruning for Convolutional Neural Networks,” in Proc. 32nd International Conference on Neural Information Processing Systems (NeurIPS 2018), pp. 1051–1061, 2018.
* [30] Y. LeCun et al., “Optimal Brain Damage,” in Proc. 3rd Advances in Neural Information Processing Systems (NeurIPS 1990), pp. 598–605, 1990.
* [31] K. Andrew et al., “CUTLASS: Fast Linear Algebra in CUDA C++.” https://developer.nvidia.com/blog/cutlass-linear-algebra-cuda/, 2017. accessed on February 1, 2022.
* [32] Z. Liu et al., “Rethinking the Value of Network Pruning.” ArXiv, 2018\. abs/1810.05270.
* [33] W. Liu et al., “A Survey of Deep Neural Network Architectures and Their Applications,” Neurocomputing, vol. 234, pp. 11–26, 2017.
* [34] S. Albawi et al., “Understanding of A Convolutional Neural Network,” in Proc. 24th International Conference on Engineering and Technology (ICET 2017), pp. 1–6, 2017.
* [35] A. F. Agarap, “Deep Learning Using Rectified Linear Units (ReLu).” ArXiv, 2018\. abs/1803.08375.
* [36] S. Ioffe et al., “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” in Proceedings of the 16th International Conference on Machine Learning (ICML 2015), pp. 448–456, 2015\.
* [37] J.-H. Luo et al., “Thinet: A Filter Level Pruning Method for Deep Neural Network Compression,” in Proc. 16th IEEE International Conference on Computer Vision (ICCV 2017), pp. 5058–5066, 2017.
|
# An example for Kuznetsov-Shinder conjecture
Tanya Kaushal Srivastava Indian Institute of Technology Gandhinagar,
Gandhinagar-382355, India<EMAIL_ADDRESS>
###### Abstract.
We give an example for the Kuznetsov-Shinder conjecture, with infinitely many
non-isomorphic but L-equivalent varieties.
## 1\. The Grothendieck ring of varieties
Let $k$ be a field of characteristic zero. One define the Grothendieck ring of
varieties over $k$ as the ring generated by isomorphism classes $[X]$ of
algebraic varieties $X$ over $k$ with relations $[X]=[Z]+[U]$ for every closed
subvariety $Z\subset X$ having open complement $U\subset X$. The product
structure of this ring is induced by the products of varieties. The ring is
denoted by $K_{0}(Var/k)$. The ring has a unit element given by the
equivalence class of a point $[Spec(k)]$. This ring was introduced by
Grothendieck in his correspondence with Serre [3]. Results of Bittner [1] and
Looijenga [16] give an alternate description of this ring, showing that the
ring is generated by isomorphism classes of smooth projective varieties
subject to the relation: $[X]+[Z]=[Y]+[E]$, whenever $Y$ is smooth projective,
$Z\subset Y$ is a smooth subvariety, $X$ is a blow-up of $Y$ along $Z$, and
$E\subset X$ is the exceptional divisor. The following lemmas are very useful
and easy consequences of the definition of Grothendieck ring of varieties.
###### Lemma 1.1.
[12, Lemma 2.1] Assume $M\rightarrow S$ is a Zariski locally trivial fibration
with fiber $F$. Then $[M]=[S][F]$ in $K_{0}(Var/k)$.
###### Lemma 1.2.
In the Grothendieck ring of varieties $K_{0}(Var/k)$, we have:
1. (1)
$[\emptyset]=0$
2. (2)
$[Spec\ k]=1$
3. (3)
$[\mathbb{P}^{n}]=1+\mathbb{L}+\ldots+\mathbb{L}^{n}$, where
$\mathbb{L}=[\mathbb{A}^{1}]$.
The above lemma is very easy to prove, see for example [18].
Poonen has shown that the ring $K_{0}(Var/k)$ is not a domain [19]. Let us
denote by $\mathbb{L}$ the equivalence class of the affine line. It has been
further shown that $\mathbb{L}$ is a zero-divisor in the ring $K_{0}(Var/k)$
[2], [17], [8].
###### Definition 1.3.
Let $X$ and $Y$ be smooth projective varieties. They are called $D$-equivalent
if their derived categories $D^{b}(X)$ and $D^{b}(Y)$ of bounded complexes of
coherent sheaves are equivalent as triangulated categories, i.e., there exists
an equivalence of categories $\phi:D^{b}(X)\rightarrow D^{b}(Y)$ which
commutes with the translations and sends any distinguished triangle to a
distinguished triangle.
###### Definition 1.4.
Two algebraic varieties $X$ and $Y$ are called $L$-equivalent if the
differences $[X]-[Y]$ vanishes in the localization
$K_{0}(Var_{\mathbb{C}})[\mathbb{L}^{-1}]$. Equivalently, for some $n>0$ we
have $([X]-[Y]).\mathbb{L}^{n}=0$.
Question(Kuznetsov-Shinder): Assume X and Y are $D$-equivalent smooth
projective varieties. What is the relation between the classes $[X]$ and $[Y]$
in the Grothendieck ring $K_{0}(Var/k)$?
A partial answer to the above question of theirs was conjectured by them.
Kuznetsov-Shinder Conjecture: If $X$ and $Y$ are smooth projective simply
connected varieties such that $D(X)$ and $D(Y)$, then there is a non-negative
integer $r\geq 0$ such that $([X]-[Y])\mathbb{L}^{r}=0$. Equivalently,
$D$-equivalence of simply connected varieties implies their $L$-equivalence.
###### Remark 1.5.
The condition of being simply connected is necessary. Efimov [5] and
independently Ito et. al [9], showed that there exist derived equivalent
Abelian varieties over $\mathbb{C}$ which are not $L$-equivalent.
For a list of known positive cases of Kusnetsov-Shinder conjecture, we refer
the reader to the survey by Ito et. al [9].
### 1.1. An example in support: Rational 3-folds
Lesieutre [14] gave an example of a rational threefold to show that a smooth
projective variety can have infinitely many Fourier-Mukai partners, i.e.,
there are infinitely many non-isomorphic smooth projective varieties (all
rational threefolds) whose derived category of coherent sheaves are all
isomorphic.
###### Theorem 1.6.
[14] Let p denote an ordered $8$-tuple of distinct points in $\mathbb{P}^{3}$,
and let $X_{\textbf{p}}$ be the blow-up of $\mathbb{P}^{3}$ at the points of
p. There is an infinite set $W$ of configurations of $8$ points in
$\mathbb{P}^{3}$ such that if p and q are distinct elements of $W$, then
$D(X_{\textbf{p}})\cong D(X_{\textbf{q}})$ but are not isomorphic.
A rational variety is simply connected (with the etale fundamental group being
trivial or if defined over $\mathbb{C}$ then with the usual topological
fundamental group) [4, Corollary 4.18]. In particular this example satisfies
the assumptions of Kuznetsov-Shinder conjecture. We can easily show that this
example provide support for the conjecture. The conjecture holds for it.
###### Proposition 1.7.
Each pair of non-isomorphic derived equivalent $X_{\textbf{p}}$ and
$X_{\textbf{q}}$ have the same class in the Grothendieck ring of varieties. In
particular, they are $\mathbb{L}$-equivalent.
###### Proof.
From [14, Lemma 1], we know that $X_{\textbf{p}}$ and $X_{\textbf{q}}$ are
isomorphic iff p and q coincide upto an automorphism of $\mathbb{P}^{3}$ and
permutation of points. From the alternate description of $K_{0}(Var/k)$ using
blowups (see [1] and [16]), we can write
(1) $[X_{\textbf{p}}]=[\mathbb{P}^{3}]-8[k]+[E],$
where $E$ is the exceptional divisor and $[k]$ is the equivalence class of a
point, which is our unit element. The exceptional divisor of the blow up is
Zariski locally isomorphic to $\\{pt\\}\times\mathbb{P}^{2}$ ([7, Theorem II
8.24]). Thus, from the relation above and lemma 1.2, we get
$\displaystyle[X_{\textbf{p}}]$
$\displaystyle=\mathbb{L}^{3}+\mathbb{L}^{2}+\mathbb{L}+1-8+\mathbb{L}^{2}+\mathbb{L}+1$
$\displaystyle=\mathbb{L}^{3}+2\mathbb{L}^{2}+2\mathbb{L}-6.$
Since this formula, does not depend on the configuration of the 8 points in
$\mathbb{P}^{3}$, we see that all the varieties $X_{\textbf{p}}$ have the same
class in the Grothendieck ring of varieties and hence are
$\mathbb{L}$-equivalent. ∎
###### Remark 1.8.
Note that the above example shows that there can be infinitely many
isomorphism classes of varieties which have the same equivalence class in
Grothendieck ring of varieties.
### 1.2. Grothendieck Ring of Varieties and Birationality
One can study rationality problems in algebraic geometry using the
Grothendieck ring of varieties. Consider the following lemma:
###### Lemma 1.9.
[6, Lemma 2.1] Let $X,X^{\prime}$ be smooth birationally equivalent varieties
of dimension $d$. Then we have an equality in the Grothendieck ring
$K_{0}(Var/k)$:
$[X^{\prime}]-[X]=\mathbb{L}.\mathcal{M}$
where $\mathcal{M}$ is a linear combination of classes of smooth projective
varieties of dimension $d-2$.
This immediately gives the following rationality condition as a corollary:
###### Corollary 1.10.
If $X$ is a rational smooth $d$-dimensional variety, then
(2) $[X]=[\mathbb{P}^{d}]+\mathbb{L}.\mathcal{M}_{X}$
where $\mathcal{M}_{X}$ is a linear combination of classes of smooth
projective varieties of dimension $d-2$.
###### Remark 1.11.
Localizing with respect to a zero divisor:
Further, there is another open conjecture of Larsen and Lunts [13].
Conjecture: If $[X]=[Y]$, then $X$ is birational to $Y$.
The above conjecture is trivially true for the example of Lesieutre as we have
constructed varieties which are birational to each other. The conjecture has
been proven true for Calabi-Yau’s by Liu and Sebag in [15, Corollary 1]. A
weaker result in this direction was already proven by Larsen and Lunts in
[13]:
###### Theorem 1.12.
([13]) The quotient-ring $K_{0}(Var/k)/\mathbb{L}$ is naturally isomorphic to
the free abelian group generated by classes of stable birational equivalence
of smooth projective connected varieties together with its natural ring
structure. In particular, if $X$ and $Y_{1},\ldots,Y_{m}$ are smooth
projective connected varieties and
$[X]=\sum_{j=1}^{m}n_{j}[Y_{j}]\mod\mathbb{L}$
for some $n_{j}\in\mathbb{Z}$, then $X$ is stably birationally equivalent to
one of the $Y_{j}$.
Recall that two smooth projective varieties $X$ and $Y$ are called stably
birationally equivalent if for some $m,n\geq 1$, $X\times\mathbb{P}^{m}$ is
birationally equivalent to $Y\times\mathbb{P}^{n}$.
###### Remark 1.13.
Generally the notion of stable birational equivalence is weaker than
birational equivalence. However, for varieties with non-negative Kodaira
dimension these two notions are the same [6, Lemma 2.6].
### 1.3. Comparison with K-equivalence
Kawamata introduced K-equivalence in [10].
###### Definition 1.14.
Let $X$ and $Y$ be smooth projective varieties. They are called $K$-equivalent
if they are birationally equivalent and if there exists a smooth projective
variety $Z$ with birational morphisms $f:Z\rightarrow X$ and $g:Z\rightarrow
Y$ such that the pull-backs of the canonical divisors are linearly equivalent:
$f^{*}K_{X}\sim g^{*}K_{Y}$ .
He further conjectured the following
Kawamata’s DK-Conjecture: [10] Let X and Y be birationally equivalent smooth
projective varieties. Then the following are equivalent.
1. (1)
There exists an equivalence of triangulated categories $D^{b}(X)\cong
D^{b}(Y)$.
2. (2)
There exists a smooth projective variety Z and birational morphisms
$f:Z\rightarrow X$ and $g:Z\rightarrow Y$ such that $f^{*}K_{X}\sim
g^{*}K_{Y}$.
In other words,for birational varieties D-equivalence is the same as
K-equivalence. For a survey of the current known examples for this conjecture
along with its more general form, see [11].
The above example of Lesieture satisfies the DK-conjecture as well. This can
be seen easily by taking $Z$ to be the blowup of $\mathbb{P}^{3}$ at all the
distinct points in the sets p and q of $X_{\textbf{p}}$ and $X_{\textbf{q}}$.
The maps $f$ and $q$ are just the respective blow downs. Note that pullback of
the canonical divisors of $X_{\textbf{p}}$ and $X_{\textbf{q}}$ are the same.
## Acknowledgements
This work was prepared for talk at 38th Annual Conference of Ramanujan
Mathematical Society, 2023 and is supported by the SERB POWER grant No.
SPG/2021/002307.
## References
* [1] Bittner F., The universal Euler characteristic for varieties of characteristic zero, Compos Math. 140(4), 1011-1032, 2004.
* [2] Borisov L., Class of the affine line is a zero divisor in the Grothendieck ring, arXiv:1412.6194.
* [3] Colmez P., Serre J.P., Grothendieck-Serre Correspondence: Bilingual Edition, American Mathematical Society, 2022.
* [4] Debarre, O, Higher-dimensional algebraic geometry, 1st Edition, Universitext, Springer, 2001.
* [5] Efimov A.I., Some remarks on L-equivalence of algebraic varieties, Sel. Math. New Series, 24:3753-3762, 2018.
* [6] Galkin, S., Shinder, E., The Fano variety of lines and rationality problem for a cubic hypersurface, arXiv: 1405.5154, 2014.
* [7] Hartshorne R., Algebraic Geometry, Graduate Text in Mathematics 52, Springer, 1977.
* [8] Ito, A., Miura, M., Okawa, S., Ueda, K., The class of the affine line is a zero divisor in the Grothendieck ring: Via $G_{2}$-Grassmannians, J. Algebraic Geom. 28 (2019), 245-250. DOI: https://doi.org/10.1090/jag/731
* [9] Ito, A., Miura, M., Okawa, S., Ueda, K., Derived equivalence and Grothendieck ring of varieties: the case of K3 surfaces of degree 12 and abelian varieties. Sel. Math. New Ser. 26, 38 (2020). https://doi.org/10.1007/s00029-020-00561-x
* [10] Kawamata, Y. D-equivalence and K-equivalence, J. Differential Geom., 61, no. 1, 147–171, 2002.
* [11] Kawamata Y., Birational geometry and derived categories, Surveys in Differential Geometry, 22(1):291–317, 2017.
* [12] Kuzentsov A., Shinder E., Grothendieck ring of varieties, D- and L-equivalence, and families of quadrics, Sel. Math. New Series, 2017.
* [13] M. Larsen, V. Lunts, Motivic measures and stable birational geometry, Moscow Mathematical Journal, Vol.3 (1), 2003, 85–95.
* [14] Lesieture J., Derived-Equivalent Rational Threefolds, International Mathematics Research Notices, Volume 2015, Issue 15, 2015, Pages 6011–6020, https://doi.org/10.1093/imrn/rnu117.
* [15] Liu Q., Sebag J., The Grothendieck ring of varieties and piecewise isomorphisms, Math Z., 265, 321-342, 2010.
* [16] Looijenga, E., Motivic measures, Seminaire Bourbaki 42, 267-297 (1999-2000).
* [17] Martin N., The class of the affine line is a zero divisor in the Grothendieck ring: an improvement, Comptes Rendus Mathematique, Volume 354, Issue 9, 936-939, 2016. https://doi.org/10.1016/j.crma.2016.05.016.
* [18] Mustaţă M., Lecture 8 in Lecture notes on Zeta-functions in Algebraic Geometry, https://dept.math.lsa.umich.edu/~mmustata/lecture8.pdf
* [19] Poonen B., The Grothendieck ring of varieties is not a domain, Math. Res. Letters, 9(4):493-498, 2002.
|
# Fluid-gravity correspondence and causal first-order relativistic viscous
hydrodynamics
Luca Ciambelli<EMAIL_ADDRESS>Luis Lehner
<EMAIL_ADDRESS>Perimeter Institute for Theoretical Physics,
31 Caroline St North, Waterloo ON N2L 2Y5, Canada
###### Abstract
The fluid-gravity correspondence is a duality between anti-de Sitter Einstein
gravity and a relativistic fluid living at the conformal boundary. We show
that one can accommodate the causal first-order viscous hydrodynamics recently
developed by Bemfica, Disconzi, Noronha, and Kovtun in this framework, by
requiring a set of natural conditions for the geometric data at the horizon.
The latter hosts an induced Carrollian fluid, whose equations of motion are
shown to be tightly tied to the ones describing the fluid at the boundary.
Functional expressions for the transport coefficients are found –with those
associated to viscosity and heat flux uniquely determined–, satisfying a set
of known causality requirements for the underlying equations of motion.
## I Introduction
In a series of works by Bemfica, Disconzi, Noronha, and Kovtun (BDNK), a
formulation for viscous, relativistic hydrodynamics has been introduced where
dissipative corrections are accounted for via first-order derivatives of the
energy density and flow velocity [1, 2, 3], see also [4, 5, 6], and where
causality of the resulting equations of motion is achieved when choosing
transport coefficients within particular bounds. Such a formulation, at first
sight, is in tension with standard results where at least second-order
corrections are required to account for viscous relativistic hydrodynamics [7,
8, 9, 10, 11]. A key observation is that such results require a strictly non-
negative entropy change, while the first-order formulation does so up to
higher order in gradients. Arguably, this is not necessarily a significant
shortcoming as such higher order terms should be subleading in the effective
field theory regime where such theory can be written. Further, another key
aspect of BDNK is formulating such theory in a more general frame than the
typical Landau or Eckart frames where causality is violated in the first order
formulation.
The Landau frame [12] was introduced requiring that the heat current vanishes,
such that the fluid velocity is an eigenvector of the energy-momentum tensor.
On the other hand, in the Eckart frame [13] the fluid velocity is aligned with
the particle number flux, such that the equations are similar to those of non-
relativistic hydrodynamics. As pointed out by BDNK, the frame discussed above
should not be chosen driven by aesthetics, but instead requiring that the
resulting hydrodynamic equations lead to well-posed problems, thus the
equations of motions should be hyperbolic and causal.
In a parallel development, the celebrated fluid-gravity correspondence [14,
15, 16, 17, 18, 19] has linked the behavior of perturbed black holes (with
asymptotically anti-de Sitter boundary conditions) to viscous relativistic
hydrodynamics in one lower dimension. This remarkable correspondence, was
fully developed to second order in gradients, but specialized in the Landau
frame by judicious choices made when solving Einstein equations for a
perturbed black brane. Under restricted assumptions on the bulk duals, the
Landau frame was abandoned in [20, 21, 22], where the heat current was
considered in the fluid-gravity correspondence.
In the current work, with the aim of shedding light on the connection between
the fluid-gravity correspondence and the BDNK first order, viscous
relativistic hydrodynamics, we first show that the fluid-gravity
correspondence is well-suited to accommodate the full first order hydrodynamic
frame spectrum. This freedom was already present in [14], but the
correspondence was fully developed in the Landau frame. Given this freedom,
are there reasonable choices –in particular, from a gravity perspective–
leading to BDNK? To answer this question, we study the properties of the bulk
projected on the horizon, which is a null hypersurface.
It is known since the original “membrane paradigm” [23, 24] that the Einstein
equations projected to the horizon are conservation equations of a fluid,
which has been recently understood to be a Carrollian fluid [25, 26, 27, 28,
29, 30, 31]. We show that the Carrollian equations of motion, at first order,
are equal to those of perfect fluid conservation for a conformal fluid111In
particular, in agreement with the boundary fluid equations of motion for the
perfect fluid.. We also observe that requiring the null vector generating the
horizon to be aligned with the fluid velocity at first order in the derivative
expansion selects exactly the BDNK form of the heat current. Similarly, the
energy density at first order is the one used by BDNK if it is proportional to
the horizon expansion. Under these assumptions we derive the values induced by
Einstein gravity of most transport coefficients and a condition on a remaining
one for a conformal equation of state. We find that the transport coefficients
than can be fully fixed this way are within the causality and stability
bounds. These observations may open the door to unraveling a deeper connection
between the horizon Carrollian fluid and relativistic conformal fluids.
The rest of the manuscript is organized as follows. In section II we review
the construction of the fluid-gravity correspondence of [14], and extrapolate
the boundary energy-momentum tensor using the holographic prescription [32,
33]. In section III we discuss the horizon geometry using the framework of
[34], and study it at zeroth and first order derivative expansion. We then
make our geometrical choices, that we show in section IV lead to BDNK for the
boundary fluid. We conclude with final remarks in section V.
## II Setup
Following the presentation in [14], we consider a boosted black brane in 4+1
dimensions with asymptotically anti-de Sitter boundary conditions. In the
stationary case –zeroth order in the gradient expansion–, the spacetime metric
is given by
$\displaystyle\mathrm{d}s^{2}=-2u_{a}\mathrm{d}x^{a}\mathrm{d}r+r^{2}(P_{ab}-f(br)u_{a}u_{b})\mathrm{d}x^{a}\mathrm{d}x^{b}\,;$
(1)
with $f(r)=1-r^{-4}$, $u^{a}u^{b}\eta_{ab}=-1$ and
$P_{ab}=\eta_{ab}+u_{a}u_{b}$ the projector orthogonal to $u^{a}$,
$u^{a}P_{ab}=0$. The vector $u^{a}$ is constant and defines the boost, while
the function $f(br)$ describes a black brane with radius $r_{H}=b^{-1}$.
Perturbed solutions, in terms of a gradient expansion (also known as
derivative expansion) can be obtained by considering $(b,u^{a})$ as slowly
varying functions of $x^{a}$222We use bulk coordinates $x^{\mu}=(r,x^{a})$,
such that $x^{a}$ are coordinates of fixed-$r$ hypersurfaces, and in
particular of the boundary., inserting into Einstein equations and solving for
non-trivial functions of $r$ (see [14]).
As opposed to the treatment in that work, we refrain from adopting the
specific choice of vanishing “zero-modes” (as also considered in [6]). In
other terms, we allow for small coordinate changes in the $x^{a}$ sector,
which we capture in terms of a scalar function $\chi(x^{a})$ and a vector
$q^{a}(x^{b})$ transverse to $u^{a}$, $u^{a}q_{a}=0$. The resulting solution,
at first order in the gradient expansion is333Using the convention
$A_{(ab)}=\frac{1}{2}(A_{ab}+A_{ba})$.
$\displaystyle\mathrm{d}s^{2}$ $\displaystyle=$
$\displaystyle-2u_{a}\mathrm{d}x^{a}\mathrm{d}r+r^{2}(P_{ab}-f(br)u_{a}u_{b})\mathrm{d}x^{a}\mathrm{d}x^{b}$
(2)
$\displaystyle-\frac{2}{b^{4}r^{2}}u_{({a}}q_{{b})}\mathrm{d}x^{a}\mathrm{d}x^{b}+\frac{\chi}{b^{4}r^{2}}u_{a}u_{b}\mathrm{d}x^{a}\mathrm{d}x^{b}$
$\displaystyle+2r^{2}bF(br)\sigma_{ab}\mathrm{d}x^{a}\mathrm{d}x^{b}+\frac{2}{3}r\theta
u_{a}u_{b}\mathrm{d}x^{a}\mathrm{d}x^{b}$
$\displaystyle-2ru_{({a}}a_{{b})}\mathrm{d}x^{a}\mathrm{d}x^{b},$
where $F(br)=\frac{1}{br}-\frac{1}{4b^{4}r^{4}}$, and we introduced the shear,
expansion, and acceleration of $u^{a}$,
$\displaystyle\sigma^{ab}=P^{c(a}\partial_{c}u^{b)}-\frac{1}{3}P^{ab}\partial_{c}u^{c}$
(3) $\displaystyle
a_{a}=u^{b}\partial_{b}u_{a}\,,\qquad\theta=\partial_{a}u^{a},$ (4)
satisfying $u^{a}\sigma_{ab}=0$, $\sigma_{a}{}^{a}=0$ ($\eta_{ab}P^{ab}=3$),
$u^{a}a_{a}=0$.
Armed with this solution, we can now obtain the stress tensor induced at the
timelike asymptotic boundary. To do so, we follow the holographic prescription
discussed in [32, 33], of which we now recap the salient ingredients. Given a
bulk metric $g_{\mu\nu}$, we introduce a hypersurface at fixed $r$ and its
projector $h_{\mu\nu}$. Using the normalized normal form
$N=N_{\mu}\mathrm{d}x^{\mu}=\frac{\mathrm{d}r}{\sqrt{g^{rr}}}$, the projector
reads
$\displaystyle h_{\mu\nu}=g_{\mu\nu}-N_{\mu}N_{\nu}\quad
h_{\mu}{}^{\nu}N_{\nu}=0.$ (5)
The extrinsic curvature (second fundamental form) is defined as
$\displaystyle K_{ab}$ $\displaystyle=$ $\displaystyle
h_{a}{}^{\mu}h_{b}{}^{\nu}\frac{1}{2}\pounds_{N}g_{\mu\nu}$ (6)
$\displaystyle=$ $\displaystyle
h_{a}{}^{\mu}h_{b}{}^{\nu}\frac{1}{2}(\nabla_{\mu}N_{\nu}+\nabla_{\nu}N_{\mu}),$
(7)
where $\nabla$ is the bulk Levi-Civita connection. The induced inverse metric
is
$\displaystyle\overline{g}^{ab}=h_{\mu}{}^{a}h_{\nu}{}^{b}g^{\mu\nu},$ (8)
which can be used to define the trace on the hypersurface
$\displaystyle K=\overline{g}^{ab}K_{ab}.$ (9)
The traceless part of the extrinsic curvature defines the boundary stress
tensor
$\displaystyle
T^{a}{}_{b}=-2\lim_{r\to\infty}r^{4}\left(K^{a}{}_{b}-\frac{K}{4}\delta^{a}_{b}\right),$
(10)
where the $r$ pre-factor comes from the holographic dictionary and ensures its
finiteness approaching the boundary.
Applying this procedure to our line element (2), the final result is
$\displaystyle
T_{ab}=3\frac{1+\chi}{b^{4}}u_{a}u_{b}+\frac{1+\chi}{b^{4}}P_{ab}-\frac{2}{b^{3}}\sigma_{ab}-\frac{8}{b^{4}}u_{(a}q_{b)}.$
(11)
This is the stress tensor of a conformal viscous fluid with fluid velocity
$u^{a}$, which accounts for heat-flux through $q^{a}$, and a correction to the
perfect fluid energy density through $\chi$. Since a generic stress tensor is
decomposed as
$\displaystyle T_{ab}={\cal E}u_{a}u_{b}+{\cal
P}P_{ab}-2\eta\sigma_{ab}+u_{a}Q_{b}+Q_{a}u_{b},$ (12)
one has,
$\displaystyle{\cal E}$ $\displaystyle=$ $\displaystyle
3\frac{1+\chi}{b^{4}},\quad{\cal P}=\frac{1+\chi}{b^{4}}\quad$
$\displaystyle\eta$ $\displaystyle=$ $\displaystyle\frac{1}{b^{3}}\,,\quad
Q_{a}=-\frac{4}{b^{4}}q_{a}.$ (13)
We note that ${\cal E}=3{\cal P}$ is the conformal equation of state implied
by the asymptotic conformal symmetry, streaming from the vanishing of the
stress tensor trace. Notice that at equilibrium the temperature is given by
$T=\frac{1}{b}$.
From here, one could straightforwardly identify conditions for $(\chi,q^{a})$
to recover BDNK. This would amount to using a particular formulation of
viscous-relativistic hydrodynamics to fix conditions on the gravitational
sector. However, our goal is to go in the opposite way, namely to consider
arguably natural choices on the gravitational sector –specifically at the
horizon– and explore what they correspond to in the hydrodynamic side.
## III Choices
To consistently deal with a degenerate metric at the null hypersurface
describing the horizon, we adopt the null Rigging formalism described in [34].
The horizon is generically located at $r=r_{H}(x)$, and thus the one-form
normal to the horizon is
$\displaystyle\underline{n}=\tilde{\alpha}\mathrm{d}(r-r_{H}(x)),$ (14)
and we adopt $\tilde{\alpha}=1$ in the following. Next, we introduce the
vector $k=\partial_{r}$ with the defining properties
$\displaystyle\underline{n}(k)=1\,,\qquad\underline{k}(k)=0.$ (15)
This vector is called the null Rigging vector. We can then define the Rigging
projector as
$\displaystyle\Pi_{\mu}{}^{\nu}=\delta_{\mu}^{\nu}-n_{\mu}k^{\nu},$ (16)
such that
$\displaystyle\Pi_{\mu}{}^{\nu}n_{\nu}=0\qquad k^{\nu}\Pi_{\nu}{}^{\mu}=0.$
(17)
The Rigging projector projects to the null hypersurface, since indeed the form
$\underline{n}$ and the vector $k$ are normal to it.
The bulk metric duals $n$ and $\underline{k}=-u_{a}\mathrm{d}x^{a}$ satisfy
$\displaystyle\Pi_{\mu}{}^{\nu}k_{\nu}=k_{\mu}\qquad\ell^{\mu}=n^{\nu}\Pi_{\nu}{}^{\mu}.$
(18)
Furthermore, the projected metric is given by
$\displaystyle q_{\mu\nu}$ $\displaystyle=$
$\displaystyle\Pi_{\mu}{}^{\rho}\Pi_{\nu}{}^{\sigma}g_{\rho\sigma}=g_{\mu\nu}-n_{\mu}k_{\nu}-k_{\mu}n_{\nu}.$
(19)
The components intrinsic to the hypersurface, $(k_{a},\ell^{a},q_{ab})$, form
the ruled Carrollian structure discussed in [31] (with the same conventions).
In particular, $\ell^{a}$ is the Carrollian vector field, $k_{a}$ is the
Ehresmann connection, and $q_{ab}$ is the degenerate Carrollian metric
satisfying
$\displaystyle\ell^{a}q_{ab}=0$ (20)
at the horizon.
The other relevant quantities for the horizon physics are the surface gravity,
expansion, Hájiček connection, and acceleration. They are defined in the bulk
by,
(1) Surface gravity:444This quantity should be called inaffinity, but for non-
expanding horizons these two concepts coincide. Here, by construction at
zeroth order and as a consequence of the equations of motion at first order,
the horizon expansion vanishes so we are in this framework.
$\displaystyle\ell^{\mu}\nabla_{\mu}\ell^{\nu}=\kappa\ell^{\nu},\qquad
k_{\nu}\ell^{\mu}\nabla_{\mu}\ell^{\nu}=\kappa\,;$ (21)
(2) Expansion:
$\displaystyle\Theta=q_{\nu}{}^{\mu}\nabla_{\mu}n^{\nu}\,;$ (22)
(3) Hájiček connection:
$\displaystyle\pi_{\mu}=q_{\mu}{}^{\nu}k_{\rho}\nabla_{\nu}n^{\rho}\,;$ (23)
(4) Acceleration
$\displaystyle\varphi_{\mu}=n^{\nu}\nabla_{[\mu}k_{\nu]}\,.$ (24)
We now proceed to compute these quantities, for clarity we do so first in the
stationary solution (zeroth order), and then in the first order perturbed
case.
### III.1 Zeroth Order
At this order, the location of the horizon, and associated normal form and
vector are:
$\displaystyle r_{H}=\frac{1}{b}\,,\qquad\underline{n}=\mathrm{d}r\,,\qquad
k=\partial_{r}\,.$ (25)
The bulk metric duals are
$\displaystyle
n=r^{2}f(br)\partial_{r}+u^{a}\partial_{a}\,,\qquad\underline{k}=-u_{a}\mathrm{d}x^{a}$
(26)
and thus the Carrollian vector is exactly given by the boundary fluid
congruence (which is constant at zeroth order)
$\displaystyle\ell^{\mu}=n^{\nu}\Pi_{\nu}{}^{\mu}=u^{a}\delta_{a}^{\mu}.$ (27)
This implies
$\displaystyle\ell^{\mu}k_{\mu}=-u^{a}u_{a}=1.$ (28)
The degenerate metric on the null surface is
$\displaystyle q_{\mu\nu}=\begin{pmatrix}0&0\\\
0&r^{2}(P_{ab}-f(br)u_{a}u_{b})\end{pmatrix}\xrightarrow{r=r_{H}}\begin{pmatrix}0&0\\\
0&\frac{P_{ab}}{b^{2}}\end{pmatrix}$ (29)
which indeed satisfies at the horizon
$\displaystyle\ell^{\mu}q_{\mu\nu}=u^{a}q_{ab}\delta^{b}_{\nu}\xrightarrow{r=r_{H}}u^{a}\frac{P_{ab}}{b^{2}}=0.$
(30)
With the above quantities, it is straightforward to obtain:
$\displaystyle\kappa=\frac{2}{b},\quad\Theta=0,\quad\pi_{\mu}=0,\quad\varphi_{\mu}=0.$
(31)
Exactly like the relativistic conformal fluid at the boundary, the Carrollian
fluid at the horizon is a perfect fluid at zeroth order. Delving into the
properties of the Carrollian fluid on the horizon and its connection to the
boundary fluid would bring us too afield from the subject of this manuscript.
We leave this exploration to a future work.
### III.2 First Order
We now perturb the stationary solution using the first order gradient
expansion. Details on how to establish the location of the perturbed horizon
are in [15] (in particular subsection 2.3), so we just report the result here.
At first order, the horizon and associated normal form are
$\displaystyle
r_{H}=\frac{1}{b}+\frac{\theta}{6}+\frac{\chi}{4b}-\frac{u^{a}\partial_{a}b}{2b},\qquad\underline{n}=\mathrm{d}r+\frac{\mathrm{d}b}{b^{2}},$
(32)
where $\theta$, $\chi$, and $\mathrm{d}b$ are first order quantities.
Following the steps described above, we gather
$\displaystyle\underline{k}=-u_{a}\mathrm{d}x^{a},$ (33)
$\displaystyle\ell^{a}=u^{a}-ba^{a}-q^{a}+P^{ab}\partial_{b}b\,,\quad\ell^{r}=-u^{a}\frac{\partial_{a}b}{b^{2}};$
(34)
where the indices of the various quantities ($a^{a}$, $q^{a}$, and $P_{ab}$)
are raised using $\eta_{ab}$, and we note that $\ell^{r}$ is non-vanishing due
to the fact that the horizon position is now a function of $x^{a}$.
With the above, through a direct, but naturally more involved calculation, one
obtains:
$\displaystyle\kappa$ $\displaystyle=$
$\displaystyle\frac{2}{b}-\frac{2\theta}{3}+\frac{\chi}{2b}+\frac{u^{a}\partial_{a}b}{b}$
(35) $\displaystyle\Theta$ $\displaystyle=$
$\displaystyle\theta-3\frac{u^{a}\partial_{a}b}{b}$ (36)
$\displaystyle\pi_{a}$ $\displaystyle=$ $\displaystyle
2\frac{q_{a}}{b}-\frac{P_{a}{}^{b}\partial_{b}b}{b}$ (37)
$\displaystyle\varphi_{a}$ $\displaystyle=$ $\displaystyle a_{a}.$ (38)
With these, we are now ready to argue for some particular choices. First, one
could demand that at first order, the component of the null vector
$\ell^{\mu}$ orthogonal to $r$ should be aligned with $u^{a}$ (just as in the
zeroth order case). This allows one to still identify the Carrollian vector
with the boundary fluid velocity, even at first order. Such a choice implies
$\displaystyle q^{a}=-ba^{a}+P^{ab}\partial_{b}b\,.$ (39)
This, as we shall discuss below, is precisely in line with the hydrodynamic
choice in BDNK.
Before such discussion, we must address the choice of $\chi$. First, note that
$r_{H}$ can be re-expressed as,
$\displaystyle r_{H}=\frac{1}{b}+\frac{1}{6}\Theta+\frac{\chi}{4b}.$ (40)
We shall now show that, to first order, $\Theta=0$ as a consequence of the
Einstein equations projected on the horizon, specifically the Raychaudhuri
equation. Thus, the choice $\chi\propto\Theta$, conveniently keeps $r_{H}=1/b$
on-shell. Note that, with this choice, $\kappa$ receives non-trivial first-
order corrections. We discuss the consequences of choosing it to remain
unchanged in appendix V.1. Since $\kappa$ depends on the generators’
parameterization, we regard keeping $r_{H}$ unchanged as the more natural
choice.
To see that $\Theta=0$ to first order, let us recall that Raychaudhuri’s and
Damour’s equations in vacuum are,
$\displaystyle(\pounds_{\ell}+\Theta)[\Theta]=\mu\Theta-\sigma_{a}{}^{b}\sigma_{b}{}^{a},$
(41) $\displaystyle
q_{a}{}^{b}\left(\pounds_{\ell}+\Theta\right)[\pi_{b}]+\Theta\varphi_{a}=(\overline{D}_{b}+\varphi_{b})(\mu
q_{a}{}^{b}-\sigma_{a}{}^{b})$ (42)
where $\pounds_{\ell}$ is the Lie-derivative along $\ell$, $\overline{D}_{a}$
the (Carrollian) covariant derivative associated to $q_{ab}$,
$\mu=\Theta/2+\kappa$ and we used the conventions of [31]. Since here we will
be interested only in the first order expression, where most terms in these
equations vanish, we refer to this reference for an explanation of all the
quantities involved in general. Notice $\kappa$ has an order 0 contribution,
so, eq. (41) implies that at first order $\Theta=0$. A similar analysis of eq.
(42) implies $q_{a}{}^{b}\partial_{b}\kappa+\varphi_{a}\kappa=0$, where
$q_{a}{}^{b}$ is the projector orthogonal to $\ell^{a}$ at the horizon, which
at zeroth order is simply $P_{a}{}^{b}$, and thus (using (38)) this equation
is equal to $a_{a}=P_{a}{}^{b}\frac{\partial_{b}b}{b}$.
These observations have several consequences. First, since to the order we
work, $\Theta=0$, the choice stated above for $\chi$ indeed implies
$r_{H}=1/b$ to this order. Further, and importantly, they indicate that at
first order Raychaudhuri’s and Damour’s equations are exactly equal to the
conservation of the boundary perfect fluid stress tensor. Indeed, one can
easily show that $\partial_{a}T^{a}{}_{b}=0$, using (11) and the relationships
(35), (36), and (38), gives exactly the Raychuadhuri’s and Damour’s equations,
once projected on $u^{a}$ and $P_{a}{}^{b}$, respectively. This is ultimately
tied to the fact that these equations all come from the bulk Einstein
equations and their particular hierarchical structure arising from the
characteristic treatment along a timelike-null foliation of the spacetime.
To summarize then, examining the resulting structure at the horizon, our
choices are:
$\displaystyle\ell^{a}=u^{a}\Leftrightarrow
q^{a}=-ba^{a}+P^{ab}\partial_{b}b\,$ (43) $\displaystyle
r_{H}=\frac{1}{b}+\left(2+3\alpha\right)\frac{\Theta}{12}\Leftrightarrow\chi=\alpha\,b\,\Theta\,,$
(44)
with $\alpha$ a proportionality function that remains to be specified. We
reported these results off-shell of the conservation laws discussed above. If
we now impose these conservation laws, we obtain $q^{a}=0$ and $\chi=0$. This
is precisely the outcome of the intrinsic hydrodynamic BDNK analysis for a
conformal relativistic fluid: the heat current and the first-order correction
to the energy that implement causality are zero on-shell of the first order
conservation law. In what follows, we discuss in detail the structure implied
by the geometrical identifications/choices on the resulting hydrodynamical
equations.
## IV Consequences
We can now examine the consequences of these choices on the thermodynamic
quantities obtained in (II). First, note that
$\displaystyle{\cal E}^{(0)}=3{\cal P}^{(0)}$ $\displaystyle\equiv$
$\displaystyle e=\frac{3}{b^{4}}$ (45) $\displaystyle{\cal E}^{(1)}=3{\cal
P}^{(1)}$ $\displaystyle=$
$\displaystyle\frac{3\alpha}{b^{3}}\left(\partial_{a}u^{a}-\frac{3}{b}u^{a}\partial_{a}b\right)$
(46) $\displaystyle Q^{a}$ $\displaystyle=$
$\displaystyle\frac{4}{b^{3}}\left(a^{a}-\frac{P^{ab}\partial_{b}b}{b}\right),$
(47)
where we introduced $\\{e,p=\frac{e}{3}\\}$ to denote the zeroth order
expressions for energy and pressure respectively.
The first order expressions can be re-expressed in terms of $e$ and $p$ as,
$\displaystyle{\cal E}^{(1)}$ $\displaystyle=$
$\displaystyle\frac{3\alpha}{b^{3}}\left(\partial_{a}u^{a}+\frac{u^{a}\partial_{a}e}{(e+p)}\right)$
(48) $\displaystyle Q^{a}$ $\displaystyle=$
$\displaystyle\frac{4}{b^{3}}\left(a^{a}+\frac{P^{ab}\partial_{b}e}{3(e+p)}\right)$
(49)
(the expressions for the pressure is trivially set by the conformal
condition). We can now compare with the expressions adopted by BDNK for the
conformal case, as this is the one that corresponds to our case [3]. Namely,
denoting with an overbar their choices,
$\displaystyle\bar{\cal E}^{(1)}$ $\displaystyle=$
$\displaystyle\left(\chi_{2}\,\partial_{a}u^{a}+\chi_{1}\,\frac{u^{a}\partial_{a}e}{(e+p)}\right)$
(50) $\displaystyle\bar{Q}^{a}$ $\displaystyle=$
$\displaystyle\lambda\left(a^{a}+\frac{P^{ab}\partial_{b}e}{3(e+p)}\right),$
(51)
with $\lambda,\chi_{i}$ transport coefficients555 Which include
$\\{\chi_{3},\chi_{4}\\}$ analogously introduced for the first-order pressure
${\cal P}^{(1)}$. that are chosen to ensure causality of the underlying
equations, together with $\eta$ defined in (12).
Remarkably, the functional form for the first order corrections are in
excellent agreement with the proposed terms in [3]. Moreover, our choices
motivated by considerations at the horizon also imply for the transport
coefficients (for $\eta$ we recall (II)),
$\displaystyle\eta=\frac{1}{b^{3}}\,,\,\lambda=\frac{4}{b^{3}}\,,$
$\displaystyle\chi_{1}=\chi_{2}=3\chi_{3}=3\chi_{4}=\frac{3\alpha}{b^{3}}\,,$
(52)
where $\\{\chi_{3},\chi_{4}\\}$ are linked to $\\{\chi_{1},\chi_{2}\\}$ by
conformality.
Not only do the transport coefficients have the temperature dependency of
$T^{3}$ as expected from kinetic theory [3], but the shear viscosity and heat
transport coefficients are uniquely determined666The value of the viscous
transport coefficient is tied to the lowest-lying quasinormal modes of the
perturbed black brane (see, e.g. [35]).. In particular, they satisfy the
criteria for causality $\lambda\geq\eta$ identified in [3]. Notice however our
expressions make the transport coefficients $\chi_{i}$ all proportional to
each other but do not completely fix them, nor provide bounds for them which
need not be surprising. Namely, conditions on $\chi_{i}$ determined by the
causality analysis of [3], effectively, come from the high frequency limit
(through the standard analysis within PDE theory). This can be seen by
examining the dispersion relations for the shear and sound modes and their
dependency on $\\{\eta,\lambda,\alpha\\}$. Their roles appear at order
$k^{2}$, $k^{4}$ and $k^{6}$ respectively. On the other hand, the fluid-
gravity correspondence is obtained in the long wavelength regime of perturbed
black holes in General Relativity –which is a causal theory–, thus it is
natural to expect that in the regime where the duality can be established,
conditions on relevant parameters on the hydrodynamic side can be obtained
implying such property.
For the unfixed parameter, we only demand $\alpha>0$, as this choice ensures
equilibrium can be reached, i.e. ${\cal E}^{(0)}+{\cal
E}^{(1)}\rightarrow{\cal E}^{(0)}$ within a timescale given by $\alpha$. Of
course, one can choose a suitable value for $\alpha$ such that the full set of
requirements for causality are satisfied (e.g. $\alpha=4/3$, so that
$\chi_{\\{1,2\\}}=3\chi_{\\{3,4\\}}=\lambda$) but there is no geometric reason
at this order we can demand to argue for a specific value.
## V Final words
In this work we examined from a gravitation angle how the BDNK first order
formulation of relativistic, viscous hydrodynamics is connected to the fluid-
gravity correspondence. Such a formulation, which in practice is simpler to
deal with than standard, second order viscous formulations [36, 37], has
received significant attention in recent years both at the theoretical level
[3, 2, 38, 4, 5] and also in incipient numerical investigations (e.g. [39, 40,
41]). The results obtained also revealed new connections between relativistic
and Carrollian hydrodynamics as well as with gravity.
Our analysis unearthed a natural way to motivate the BDNK formulation from a
gravitational perspective. Further, the expected functional dependence of
transport coefficients was obtained and, for the viscous and heat-flux
coefficients, a unique expression was found. As well, our analysis revealed a
connection between the effective Carrollian hydrodynamic description of null
surfaces and the asymptotic relativistic fluid that is identified at the
timelike infinity of perturbed black branes in AdS. Such connection implies
that, at leading order, Raychaudhuri’s and Damour’s equations encode the
conservation of a conformal perfect fluid. The analysis of higher orders and
the exploration of Carrollian hydrodynamics from this perspective is an
interesting task which we defer to future work.
In a similar vein, it would be interesting to explore the horizon physics
deeper, as results there would also hold for asymptotically flat spacetimes.
Importantly, in the latter case, there is also an interesting relation between
the structure of interior null surfaces (like the horizon), and future null
infinity. However, the relationship between the horizon membrane paradigm and
the asymptotic (e.g. ${\cal I}^{+}$) null boundary Carrollian fluid is still
largely unexplored. The latter fluid however enjoys Weyl symmetry, which makes
it special. This could also help motivate a fluid interpretation of
(particular) quasi-normal modes in asymptotically flat spacetimes. Another
avenue for exploration is to consider a potential entropy current, both for
the relativistic fluid at the boundary and the horizon Carrollian fluid. This
current could help us connect with its microscopic origin and inform standing
questions on Carrolllian hydrodynamics. Finally, a deeper understanding of
potential connections between phenomena in non-linear gravity and
hydrodynamics can motivate new avenues to identify and study non-linear
gravitational behavior (e.g. [16, 42, 43, 44, 45, 46, 47, 48, 49, 30]).
###### Acknowledgements.
We thank F. Bemfica, M. Disconzi, L. Freidel, S. Giri, R. E. Hoult, P. Kovtun,
R. Leigh, J. Noronha, M. Petropoulos, E. Poisson, F. Pretorius, M. Rangamani,
and J. Senovilla for discussions. This work was supported in part by Perimeter
Institute for Theoretical Physics. Research at Perimeter Institute is
supported by the Government of Canada through the Department of Innovation,
Science and Economic Development Canada and by the Province of Ontario through
the Ministry of Economic Development, Job Creation and Trade. L. L. thanks
financial support via the Carlo Fidani Rainer Weiss Chair at Perimeter
Institute. L. L. receives additional financial support from the Natural
Sciences and Engineering Research Council of Canada through a Discovery Grant
and CIFAR. L. L. thanks KITP - UC Santa Barbara for its hospitality during
“The Many Faces of Relativistic Fluid Dynamics” Program, where this work’s
initial stages were completed.
### V.1 Alternative choice for $\chi$
An alternative choice for $\chi$, which fixes it completely, would be to
demand that to first order $\kappa=2/b$. This would imply
$\displaystyle\chi$ $\displaystyle=$ $\displaystyle
2\left(\frac{2b}{3}\partial_{a}u^{a}-u^{a}\partial_{a}b\right)$ (53)
$\displaystyle=$
$\displaystyle\frac{2b}{3}\left(2\partial_{a}u^{a}+\frac{u^{a}\partial_{a}e}{(e+p)}\right),$
and as a consequence, $\chi_{1}=3\chi_{3}=2/b^{3}$,
$\chi_{2}=3\chi_{4}=4/b^{3}$. These values however, (complemented by
$\lambda=4/b^{3},\eta=1/b^{3}$) are not within the causality bounds of [3].
Further, on-shell dynamical solutions have an associated energy density at
first order which differs from that at zeroth order.
## References
* [1] F. S. Bemfica, M. M. Disconzi, and J. Noronha, “Causality and existence of solutions of relativistic viscous fluid dynamics with gravity,” Phys. Rev. D 98 (2018) no. 10, 104064, arXiv:1708.06255 [gr-qc].
* [2] P. Kovtun, “First-order relativistic hydrodynamics is stable,” JHEP 10 (2019) 034, arXiv:1907.08191 [hep-th].
* [3] F. S. Bemfica, M. M. Disconzi, and J. Noronha, “Nonlinear Causality of General First-Order Relativistic Viscous Hydrodynamics,” Phys. Rev. D 100 (2019) no. 10, 104020, arXiv:1907.12695 [gr-qc]. [Erratum: Phys.Rev.D 105, 069902 (2022)].
* [4] R. E. Hoult and P. Kovtun, “Stable and causal relativistic Navier-Stokes equations,” JHEP 06 (2020) 067, arXiv:2004.04102 [hep-th].
* [5] F. S. Bemfica, M. M. Disconzi, and J. Noronha, “First-Order General-Relativistic Viscous Fluid Dynamics,” Phys. Rev. X 12 (2022) no. 2, 021044, arXiv:2009.11388 [gr-qc].
* [6] R. E. Hoult and P. Kovtun, “Causal first-order hydrodynamics from kinetic theory and holography,” Phys. Rev. D 106 (2022) no. 6, 066023, arXiv:2112.14042 [hep-th].
* [7] R. Geroch and L. Lindblom, “Causal theories of dissipative relativistic fluids,” Annals of Physics 207 (1991) no. 2, 394–416.
* [8] G. S. Denicol, T. Kodama, T. Koide, and P. Mota, “Stability and Causality in relativistic dissipative hydrodynamics,” J. Phys. G 35 (2008) 115102, arXiv:0807.3120 [hep-ph].
* [9] S. Pu, T. Koide, and D. H. Rischke, “Does stability of relativistic dissipative fluid dynamics imply causality?,” Phys. Rev. D 81 (2010) 114039, arXiv:0907.3906 [hep-ph].
* [10] L. Lehner, O. A. Reula, and M. E. Rubio, “Hyperbolic theory of relativistic conformal dissipative fluids,” Phys. Rev. D 97 (2018) no. 2, 024013, arXiv:1710.08033 [gr-qc].
* [11] M. Aguilar and E. Calzetta, “Causal Relativistic Hydrodynamics of Conformal Fermi-Dirac Gases,” Phys. Rev. D 95 (2017) no. 7, 076022, arXiv:1701.01916 [hep-ph].
* [12] L. D. Landau and E. M. Lifshitz, Fluid mechanics. 1959\.
* [13] C. Eckart, “The Thermodynamics of irreversible processes. 3.. Relativistic theory of the simple fluid,” Phys. Rev. 58 (1940) 919–924.
* [14] S. Bhattacharyya, V. E. Hubeny, S. Minwalla, and M. Rangamani, “Nonlinear Fluid Dynamics from Gravity,” JHEP 02 (2008) 045, arXiv:0712.2456 [hep-th].
* [15] S. Bhattacharyya, V. E. Hubeny, R. Loganayagam, G. Mandal, S. Minwalla, T. Morita, M. Rangamani, and H. S. Reall, “Local Fluid Dynamical Entropy from Gravity,” JHEP 06 (2008) 055, arXiv:0803.2526 [hep-th].
* [16] M. Van Raamsdonk, “Black Hole Dynamics From Atmospheric Science,” JHEP 05 (2008) 106, arXiv:0802.3224 [hep-th].
* [17] M. Haack and A. Yarom, “Nonlinear viscous hydrodynamics in various dimensions using AdS/CFT,” JHEP 10 (2008) 063, arXiv:0806.4602 [hep-th].
* [18] N. Banerjee, J. Bhattacharya, S. Bhattacharyya, S. Dutta, R. Loganayagam, and P. Surowka, “Hydrodynamics from charged black branes,” JHEP 01 (2011) 094, arXiv:0809.2596 [hep-th].
* [19] M. Rangamani, “Gravity and Hydrodynamics: Lectures on the fluid-gravity correspondence,” Class. Quant. Grav. 26 (2009) 224003, arXiv:0905.4352 [hep-th].
* [20] M. M. Caldarelli, R. G. Leigh, A. C. Petkou, P. M. Petropoulos, V. Pozzoli, and K. Siampos, “Vorticity in holographic fluids,” PoS CORFU2011 (2011) 076, arXiv:1206.4351 [hep-th].
* [21] A. Mukhopadhyay, A. C. Petkou, P. M. Petropoulos, V. Pozzoli, and K. Siampos, “Holographic perfect fluidity, Cotton energy-momentum duality and transport properties,” JHEP 04 (2014) 136, arXiv:1309.2310 [hep-th].
* [22] J. Gath, A. Mukhopadhyay, A. C. Petkou, P. M. Petropoulos, and K. Siampos, “Petrov Classification and holographic reconstruction of spacetime,” JHEP 09 (2015) 005, arXiv:1506.04813 [hep-th].
* [23] T. Damour, Quelques proprietes mecaniques, electromagnetiques, thermodynamiques et quantiques des trous noir. PhD thesis, Paris U., VI-VII, 1979.
* [24] K. S. Thorne, R. H. Price, and D. A. Macdonald, eds., Black holes: the membrane paradigm. 1986\.
* [25] R. F. Penna, “Near-horizon Carroll symmetry and black hole Love numbers,” arXiv:1812.05643 [hep-th].
* [26] L. Donnay and C. Marteau, “Carrollian Physics at the Black Hole Horizon,” Class. Quant. Grav. 36 (2019) no. 16, 165002, arXiv:1903.09654 [hep-th].
* [27] L. Ciambelli, R. G. Leigh, C. Marteau, and P. M. Petropoulos, “Carroll Structures, Null Geometry and Conformal Isometries,” Phys. Rev. D 100 (2019) no. 4, 046010, arXiv:1905.02221 [hep-th].
* [28] L. Freidel and P. Jai-akson, “Carrollian hydrodynamics from symmetries,” Class. Quant. Grav. 40 (2023) no. 5, 055009, arXiv:2209.03328 [hep-th].
* [29] L. Freidel and P. Jai-akson, “Carrollian hydrodynamics and symplectic structure on stretched horizons,” arXiv:2211.06415 [gr-qc].
* [30] J. Redondo-Yuste and L. Lehner, “Non-linear black hole dynamics and Carrollian fluids,” JHEP 02 (2023) 240, arXiv:2212.06175 [gr-qc].
* [31] L. Ciambelli, L. Freidel, and R. G. Leigh, “Null Raychaudhuri: Canonical Structure and the Dressing Time,” arXiv:2309.03932 [hep-th].
* [32] M. Henningson and K. Skenderis, “The Holographic Weyl anomaly,” JHEP 07 (1998) 023, arXiv:hep-th/9806087.
* [33] V. Balasubramanian and P. Kraus, “A Stress tensor for Anti-de Sitter gravity,” Commun. Math. Phys. 208 (1999) 413–428, arXiv:hep-th/9902121.
* [34] M. Mars and J. M. M. Senovilla, “Geometry of general hypersurfaces in space-time: Junction conditions,” Class. Quant. Grav. 10 (1993) 1865–1897, arXiv:gr-qc/0201054.
* [35] E. Berti, V. Cardoso, and A. O. Starinets, “Quasinormal modes of black holes and black branes,” Class. Quant. Grav. 26 (2009) 163001, arXiv:0905.2975 [gr-qc].
* [36] I. Muller, “Zum Paradoxon der Warmeleitungstheorie,” Z. Phys. 198 (1967) 329–344.
* [37] W. Israel and J. M. Stewart, “Transient relativistic thermodynamics and kinetic theory,” Annals Phys. 118 (1979) 341–372.
* [38] F. S. Bemfica, M. M. Disconzi, C. Rodriguez, and Y. Shao, “Local well-posedness in Sobolev spaces for first-order conformal causal relativistic viscous hydrodynamics,” arXiv:1911.02504 [math.AP].
* [39] A. Pandya and F. Pretorius, “Numerical exploration of first-order relativistic hydrodynamics,” Phys. Rev. D 104 (2021) no. 2, 023015, arXiv:2104.00804 [gr-qc].
* [40] U. Danielsson, L. Lehner, and F. Pretorius, “Dynamics and observational signatures of shell-like black hole mimickers,” Phys. Rev. D 104 (2021) no. 12, 124011, arXiv:2109.09814 [gr-qc].
* [41] H. Bantilan, Y. Bea, and P. Figueras, “Evolutions in first-order viscous hydrodynamics,” JHEP 08 (2022) 298, arXiv:2201.13359 [hep-th].
* [42] C. Eling, I. Fouxon, and Y. Oz, “Gravity and a Geometrization of Turbulence: An Intriguing Correspondence,” arXiv:1004.2632 [hep-th].
* [43] H. Yang, A. Zimmerman, and L. Lehner, “Turbulent Black Holes,” Phys. Rev. Lett. 114 (2015) 081101, arXiv:1402.4859 [gr-qc].
* [44] A. Adams, P. M. Chesler, and H. Liu, “Holographic turbulence,” Phys. Rev. Lett. 112 (2014) no. 15, 151602, arXiv:1307.7267 [hep-th].
* [45] F. Carrasco, L. Lehner, R. C. Myers, O. Reula, and A. Singh, “Turbulent flows for relativistic conformal fluids in 2+1 dimensions,” Phys. Rev. D 86 (2012) 126006, arXiv:1210.6702 [hep-th].
* [46] S. R. Green, F. Carrasco, and L. Lehner, “Holographic Path to the Turbulent Side of Gravity,” Phys. Rev. X 4 (2014) no. 1, 011001, arXiv:1309.7940 [hep-th].
* [47] J. R. Westernacher-Schneider, “Fractal dimension of turbulent black holes,” Phys. Rev. D 96 (2017) no. 10, 104054, arXiv:1710.04264 [gr-qc].
* [48] S. Aretakis, Dynamics of Extremal Black Holes, vol. 33 of SpringerBriefs in Mathematical Physics. Springer International Publishing, Cham, 2018.
* [49] G. Moschidis, Two instability results in general relativity. PhD thesis, Princeton U., Math. Dept., 2018.
|
# Emergent inflation of the Efimov spectrum under three-body spin-exchange
interactions
J. van de Kraats<EMAIL_ADDRESS>Eindhoven University of Technology, P.
O. Box 513, 5600 MB Eindhoven, The Netherlands D. J. M. Ahmed-Braun
Eindhoven University of Technology, P. O. Box 513, 5600 MB Eindhoven, The
Netherlands J.-L. Li Institut für Quantenmaterie and Center for Integrated
Quantum Science and Technology IQ ST, Universität Ulm, D-89069 Ulm, Germany
S. J. J. M. F. Kokkelmans Eindhoven University of Technology, P. O. Box 513,
5600 MB Eindhoven, The Netherlands
###### Abstract
One of the most fascinating predictions of few-body quantum physics is the
Efimov effect, a universal accumulation of an infinite geometric series of
three-body bound states at a two-body scattering resonance. Ever since the
first experimental observation of such an Efimov state, the precise
characterization of their physical properties has continued to challenge few-
body theory. This is demonstrated most strongly by the lithium few-body
puzzle, a remarkable theoretical discrepancy with the observed Efimov spectrum
in 7Li. Here, we resolve this long-standing puzzle, demonstrating that the
discrepancy arises out of the presence of strong non-universal three-body
spin-exchange interactions. This conclusion is obtained from a thorough
numerical solution of the quantum mechanical three-body problem, including
precise interatomic interactions and all spin degrees of freedom for three
alkali-metal atoms. Our results show excellent agreement with the experimental
data regarding both the Efimov spectrum and the absolute rate constants of
three-body recombination, and in addition reveal a general product propensity
for such triatomic reactions in the Paschen-Back regime, stemming from
Wigner’s spin conservation rule.
There exists a general desire in physics to formulate accurate descriptions of
nature from a minimal number of adjustable parameters, thus uncovering the
presence of universal behavior. A paradigmatic example of a system where this
ideal picture is realized is the scattering of two particles at low energy.
Here, the wave function delocalizes to the point that the observable
properties of the system become insensitive to the exact microscopic detail of
the interaction, allowing for a description purely in terms of the $s$-wave
scattering length $a$ [1]. This remarkable universality carries through to
systems of more than two particles, most strikingly exemplified at the three-
body level by virtue of the Efimov effect. At a two-body scattering resonance,
where $a\rightarrow\infty$, the Efimov effect induces a universal emergence of
an infinite tower of geometrically spaced three-body bound states [2, 3]. The
resulting spectrum is fully determined by a single length scale, the three-
body parameter, typically expressed as the negative scattering length $a_{-}$
where the ground state trimer dissociates into the three-body scattering
continuum [1, 4, 5, 6]. In turn, the Efimov effect and three-body parameter
induce universal properties in few-body clusters of four or more particles,
further extending the applicability of universal theory [5].
The vast majority of experimental studies of the Efimov effect utilize
ultracold atomic gases, where the scattering length can be directly controlled
by means of a magnetic Feshbach resonance [7, 8]. Near such a resonance, the
three-body parameter can be extracted from a characteristic log-periodic
modulation of the rate of three-body recombination [9, 1, 4]. Interestingly,
although the precise value of $a_{-}$ is typically sensitive to non-universal
short-range physics, the Efimov spectrum in atomic systems possesses an
additional van der Waals universality $a_{-}\approx-9.7\ r_{\mathrm{vdW}}$
[10, 11, 12], where $r_{\mathrm{vdW}}$ gives the characteristic length scale
associated with the two-body interaction. Theoretical analyses have shown that
this universality is robust for Efimov states near broad Feshbach resonances,
and originates from a universal suppression of three-body probability in the
short range [13, 14].
For Feshbach resonances of intermediate to narrow width, both theoretical and
experimental works have demonstrated an increase of $\absolutevalue{a_{-}}$,
arising from the associated growth of the two-body effective range scale [15,
16, 17, 18, 19, 20, 21, 22]. However, a series of experiments in this regime
with the lightest bosonic alkali, 7Li, have failed to observe this behavior,
and in fact measured values for $\absolutevalue*{a_{-}}$ that, remarkably,
recede slightly below the universal van der Waals value [23, 24, 25, 26, 27].
While similar behavior can be obtained in some theoretical scenarios [28, 20],
it is generally unclear how to connect these to 7Li, and sophisticated
numerical models have so far failed to reproduce the data. The long-standing
challenge to explain this unexpected mismatch between theory and experiment is
now referred to as the lithium few-body puzzle [29, 30, 31].
Figure 1: Schematic representation of three-body recombination through two
distinct spin-exchange pathways, where the color of particles represents their
spin state and two connected particles represent a molecule. The upper pathway
contains purely two-body spin-exchange, where one of the particles conserves
its spin throughout the recombination. In the bottom pathway, all three
particles partake in spin-exchange such that no single spin is conserved. We
note that there is an additional pathway, not pictured here, in which all
particles preserve their initial spin state. This pathway is included in all
our calculations.
In this work, we investigate the connection between the anomalous value of
$\absolutevalue*{a_{-}}$ in 7Li and the presence of three-body spin-exchange
interactions. Here, spin-exchange refers to a process in which the internal
spin state of the atoms is altered by coupling of the valence electrons, and
thus necessarily occurs at short length scales. Due to the aforementioned
suppression of three-body probability in this regime, a useful distinction can
be made between two-body spin exchange, where the third particle only
spectates and its state is conserved, and three-body spin-exchange, where all
particles partake. As we illustrate in Fig. 1, this distinction can also
naturally be applied to three-body recombination. For many applications, the
contributions of three-body spin-exchange are negligible, which significantly
simplifies the three-body problem [15, 16, 32, 17, 33, 18, 19, 34, 20].
Recently however, studies have found three-body spin-exchange to contribute
significantly to three-body observables in 39K [35] and 7Li [36], both at
relatively large magnetic fields. Motivated by these findings we study the
Efimov spectrum in 7Li, using a recently developed numerical approach to the
quantum mechanical three-body problem which, together with high-performance
computing facilities, allows us to include all coupled three-body channels in
the Hamiltonian [37, 35] (see Methods).
## Inflation of the Efimov spectrum
Following the experiment of Ref. [26], we analyze two high-field Feshbach
resonances in spin-polarized ultracold 7Li gases, in the hyperfine states
$\ket*{f,m_{f}}_{\mathrm{in}}=\ket*{1,1}$ and
$\ket*{f,m_{f}}_{\mathrm{in}}=\ket*{1,0}$ respectively. Both resonances have
similar negative background scattering lengths on the order of
$r_{\mathrm{vdW}}=32.4863\ a_{0}$, and are both of intermediate to narrow
resonance width [26, 38]. In line with experiment we study the rate of three-
body loss of the trapped gas density $n$, typically expressed in terms of a
recombination rate constant $L_{3}$ as $\mathrm{d}n/\mathrm{d}t=-L_{3}\
n^{3}$. We calculate $L_{3}$ for a range of scattering lengths on the
attractive ($a<0$) side of the Feshbach resonance, and subsequently extract
the values of the three-body parameter $a_{-}$ and trimer width $\eta_{-}$ by
fitting the data to universal predictions from effective field theory [1]. To
highlight the role of three-body spin-exchange, we compare a full multichannel
spin (FMS) model with a fixed spectating spin (FSS) model [35], where the
latter is constrained to purely two-body spin exchange (i.e. the upper pathway
in Fig. 1, see Methods for more detail). To model the pairwise interactions we
use realistic singlet and triplet Born-Oppenheimer interaction potentials
[38]. It follows that the FMS model is quantum mechanically rigorous, save for
the omission of nonadditive three-body forces which are generally assumed to
be negligible in ultracold atomic gases [4].
Figure 2: Results of our three-body calculations, compared directly with the
experimental data of Ref. [26]. Table (a) shows the three-body parameter
$a_{-}$ and Efimov trimer width $\eta_{-}$ obtained using FSS and FMS models.
The associated values of the three-body recombination rate coefficient $L_{3}$
as a function of the scattering length $a$ are shown in Figs. (b) and (c), for
the two distinct incoming hyperfine states. The fits of $L_{3}$ to universal
theory giving the results in table (a) are shown as dash-dotted lines in
matching color. Insets show the two-body scattering length as a function of
magnetic field surrounding the Feshbach resonance, where the green shaded
region of field values matches the range of scattering lengths in the
enclosing figure. We note that next to Ref. [26], an additional independent
measurement of the Efimov trimer in the $\ket*{1,1}$ state obtained
$a_{-}/r_{\mathrm{vdW}}=-7.76(31)$ and $\eta_{-}=0.17$ [27].
Our results are shown in Fig. 2. In an FSS calculation, the value of
$\absolutevalue*{a_{-}}$ is significantly larger than both the universal van
der Waals value and the experimental data, suggesting a significant squeezing
of the spectrum for this narrow resonance which is in line with the majority
of multichannel three-body models in the current literature [17, 18, 22]. Our
main result is that upon including three-body spin-exchange processes, the
additional accessible states induce non-trivial multichannel physics that acts
to cancel the increase of $\absolutevalue*{a_{-}}$, and can even decrease
$\absolutevalue*{a_{-}}$ to below the universal van der Waals value thus
affecting an inflation of the spectrum. Consequently, our FMS calculations
significantly improve on the FSS results with respect to the experimental
data, and for the $\ket*{f,m_{f}}_{\mathrm{in}}=\ket*{1,1}$ state specifically
our FMS results for both $a_{-}$ and $\eta_{-}$ fall within the experimental
uncertainty of Ref. [26]. Due to the increased number of coupled channels, the
FMS calculations in the $\ket*{f,m_{f}}_{\mathrm{in}}=\ket*{1,0}$ state are
too intensive numerically to get fully converged, but the similarity in
behavior to the $\ket*{f,m_{f}}_{\mathrm{in}}=\ket*{1,1}$ state strongly
indicates that a similar match with the experiment is achievable if numerical
resources allow (see Supplementary Material).
Previous three-body studies have shown that the universal increase in
$\absolutevalue{a_{-}}$ near narrow resonances arises from a repulsive barrier
in the three-body potential scaling with the effective range, which
progressively squeezes the Efimov spectrum [39]. Recently however, the
unexpected observation of a trimer state above the atom-dimer disassociation
threshold has prompted further theoretical analysis, which indicates that the
Efimov state of 7Li may actually exist behind the universal repulsive barrier
[40]. While the universal effects of the barrier are evident in our FSS
results, the observed sensitivity of the value of $\absolutevalue{a_{-}}$ to
short-range three-body spin-exchange processes is in fact consistent with the
presence of a non-universal trimer state in the inner potential well. In this
sense, our FMS results can serve as important numerical confirmation of this
novel trimer binding mechanism. Such an identification furthermore indicates
that three-body spin-exchange couplings induce an effective attractive
interaction that can tug the trimer state into the inner potential well, thus
causing the inflation of the Efimov spectrum. In the future, it may be
interesting to analyze the exact nature of this trimer and the corresponding
potential in more detail, for which an approach similar to Ref. [22] could
prove a useful starting point.
Figure 3: Analysis of the electronic spin-propensity rule for three-body
recombination. In Figs. (a) and (b) we show the components of the incoming
two-body radial wave functions $r\psi_{\mathrm{in}}(r)$, highlighting the
electronic spin components $\ket*{{\downarrow}{\downarrow}}$ in green and
$\ket*{{\downarrow}{\uparrow}}$ in dash-dotted orange. The singlet
$\ket*{{\downarrow}{\uparrow}}$ components are resonantly enhanced by the
Feshbach level and thus dominate the wave function at short distance. In Fig.
(c) and (d) we show the averaged partial recombination rate
$\mathcal{L}_{3}(M_{S})$ to channels with definite total electronic projection
$M_{S}$, which automatically fixes the total nuclear projection $M_{I}$. The
fraction in blue originates from atom-dimer states accessible by purely two-
body spin exchange, while the red portion comes from states only accessible by
three-body spin exchange.
## Propensity rule for three-body recombination
Next to the excellent match with the three-body parameter our calculations
also show good agreement with the individual measurements of $L_{3}$. As
three-body recombination is an important and ubiquitous chemical process,
relevant far beyond the specific context of Efimov physics, this agreement
motivates us to analyze the recombination rates more closely [41].
Specifically we aim to characterize the nature of the spin-exchange pathways,
for which we decompose $L_{3}$ into partial recombination rates to all
distinct atom-dimer channels, providing a measure of the population
distribution of product states following three-body recombination.
Interestingly our calculations show that, even though the FSS approximation
fails, the actual number of strongly coupled channels remains remarkably
small, with the vast majority of product channels having near negligible
relative recombination rates (see Supplementary Material for more detail).
We will now show that this behavior results from a manifestation of Wigner’s
electronic-spin conservation rule [42, 43] for three atoms, originating from
the relatively weak coupling between the electronic and nuclear spins of the
atoms at large magnetic fields. In this Paschen-Back regime [44], the spins
independently precess around the magnetic field direction, such that single-
particle states are best described by the individual projection quantum
numbers $m_{s}$ and $m_{i}$. For simplicity let us briefly neglect the
subdominant contribution from the nuclear spin, whose coupling to the magnetic
field is relatively weak. Then both incoming single-particle states studied in
this work may be written as,
$\ket*{f,m_{f}}_{\mathrm{in}}\sim\ket*{\downarrow}+\delta\ket*{\uparrow},$ (1)
where $\ket*{{\downarrow}/{\uparrow}}$ represent the down and up electronic
spin states $m_{s}=-1/2$ and $m_{s}=1/2$ respectively, and $\delta$ scales
with the ratio of the hyperfine and Zeeman energies [45].
In the Paschen-Back regime $\delta$ is a small number, which motivates us to
expand the incoming three-body state into four distinct components scaling as
$\delta^{n}$, where $n$ equals the number of electronic spins pointing up.
Each component can be uniquely identified with a definite value of the total
electronic spin projection $M_{S}=m_{s_{1}}+m_{s_{2}}+m_{s_{3}}=-3/2+n$, which
is rigorously conserved in this basis as it is fully uncoupled from the
nuclear spin. Hence, if it is possible to identify a dominant incoming
projection $M_{S}$, then the outgoing product state distribution will show a
propensity to states that conserve this projection. To determine the dominant
incoming component we have to consider both the scaling with the small
parameter $\delta$ and the amplitude of the associated wave functions at small
nuclear separations, where recombination processes typically take place. As we
illustrate in Figs. 3(a) and 3(b), the dominant short-range components of the
incoming two-body wave functions are in the singlet two-body state
$\ket*{{\downarrow}{\uparrow}}$, correspondent with the spin character of the
resonantly coupled Feshbach level. It follows that recombination preferably
occurs through three-body states with (partial) singlet character, of which
the state $\ket*{{\downarrow}{\downarrow}{\uparrow}}$ ($n=1$) has the dominant
scaling with $\delta$. Thus, we finally deduce that three-body recombination
will show a propensity to product channels with $M_{S}=-\frac{1}{2}$.
To confirm the presence of this propensity in our numerics, we introduce the
quantity $\mathcal{L}_{3}(M_{S})$ as the average of the partial recombination
rates with respect to the projection of the associated outgoing state on a
three-body spin state with definite total electronic spin projection (see
Methods section). In this way $\mathcal{L}_{3}(M_{S})$ provides a measure for
the relative prevalence of a given value of $M_{S}$ in the product state
distribution following recombination. As shown in Figs. 3(c) and 3(d), the
electronic spin-conservation propensity is clearly present for both Feshbach
resonances we study. Note that in the argument outlined above all three
particles are treated equally, and no reference is made to a special role for
the spectating particle. Indeed, we observe that the
$\ket*{{\downarrow}{\downarrow}{\uparrow}}$ component of $\mathcal{L}_{3}$ is
split almost equally between channels in which the spectating spin is
conserved or changed, with the latter being slightly larger.
## Outlook
Our findings suggest several new avenues for future research. First the
excellent match between our 7Li results and the experimental data now provides
a new benchmark for the theoretical description of strongly interacting few-
body systems. Our method thus shows great promise for studying other systems
where measurements deviate from the current theoretical predictions [21].
Aside from these experimental concerns, there is also a more conceptual
challenge to now further characterize the physical mechanism underpinning the
formation of the non-universal Efimov trimer observed in this work, which will
require untangling the exact reshaping of the three-body potential in the
presence of three-body spin-exchange [22, 39, 40].
Our results also have interesting implications beyond the realm of Efimov
physics. The uncovered spin-propensity rule in the rate of three-body
recombination provides a remarkably simple picture of triatomic chemical
reactions in large magnetic fields, which can now aid in the understanding and
possible experimental control of state-to-state quantum chemistry in these
regimes [33, 34]. Further studies in this direction may also seek to elucidate
the more subtle role of the individual nuclear spins, which should have a
similar propensity to be conserved in the Paschen-Back regime.
###### Acknowledgements.
We thank Jose D’Incao for fruitful discussions. J.v.d.K. and S.J.J.M.F.K.
acknowledge financial support from the Dutch Ministry of Economic Affairs and
Climate Policy (EZK), as part of the Quantum Delta NL program. D.J.M.A.B.
acknowledges financial support from the Netherlands Organisation for
Scientific Research (NWO) under Grant No. 680-47-623. The results presented in
this work were obtained on the Dutch national supercomputer Snellius, with
support from TU/e HPC Lab and Surf.
## References
* Braaten and Hammer [2006] E. Braaten and H.-W. Hammer, Physics Reports 428, 259 (2006).
* Efimov [1970] V. Efimov, Phys. Lett. B 33, 563 (1970).
* Efimov [1971] V. Efimov, Sov. J. Nucl. Phys. 12, 589 (1971).
* Naidon and Endo [2017] P. Naidon and S. Endo, Rep. Prog. Phys. 80, 056001 (2017).
* Greene _et al._ [2017] C. H. Greene, P. Giannakeas, and J. Pérez-Ríos, Rev. Mod. Phys. 89, 035006 (2017).
* D’Incao [2018] J. P. D’Incao, J. Phys. B: At. Mol. Opt. Phys. 51, 043001 (2018).
* Köhler _et al._ [2006] T. Köhler, K. Góral, and P. S. Julienne, Rev. Mod. Phys. 78, 1311 (2006).
* Chin _et al._ [2010] C. Chin, R. Grimm, P. Julienne, and E. Tiesinga, Rev. Mod. Phys. 82, 1225 (2010).
* Esry _et al._ [1999] B. D. Esry, C. H. Greene, and J. P. Burke, Phys. Rev. Lett. 83, 1751 (1999).
* Kraemer _et al._ [2006] T. Kraemer, M. Mark, P. Waldburger, J. G. Danzl, C. Chin, B. Engeser, A. D. Lange, K. Pilch, A. Jaakkola, H.-C. Nägerl, and R. Grimm, Nature 440, 315 (2006).
* Berninger _et al._ [2011] M. Berninger, A. Zenesini, B. Huang, W. Harm, H.-C. Nägerl, F. Ferlaino, R. Grimm, P. S. Julienne, and J. M. Hutson, Phys. Rev. Lett. 107, 120401 (2011).
* Wild _et al._ [2012] R. J. Wild, P. Makotyn, J. M. Pino, E. A. Cornell, and D. S. Jin, Phys. Rev. Lett. 108, 145305 (2012).
* Wang _et al._ [2012] J. Wang, J. P. D’Incao, B. D. Esry, and C. H. Greene, Phys. Rev. Lett. 108, 263001 (2012).
* Naidon _et al._ [2014] P. Naidon, S. Endo, and M. Ueda, Phys. Rev. A 90, 022106 (2014).
* Petrov [2004] D. S. Petrov, Phys. Rev. Lett. 93, 143201 (2004).
* Gogolin _et al._ [2008] A. O. Gogolin, C. Mora, and R. Egger, Phys. Rev. Lett. 100, 140404 (2008).
* Schmidt _et al._ [2012] R. Schmidt, R. Rath, and W. Zwerger, Eur. Phys. J. B 85, 386 (2012).
* Langmack _et al._ [2018] C. Langmack, R. Schmidt, and W. Zwerger, Phys. Rev. A 97, 033623 (2018).
* Chapurin _et al._ [2019] R. Chapurin, X. Xie, M. J. Van de Graaff, J. S. Popowski, J. P. D’Incao, P. S. Julienne, J. Ye, and E. A. Cornell, Phys. Rev. Lett. 123, 233402 (2019).
* Secker _et al._ [2021a] T. Secker, D. J. M. Ahmed-Braun, P. M. A. Mestrom, and S. J. J. M. F. Kokkelmans, Phys. Rev. A 103, 052805 (2021a).
* Etrych _et al._ [2023] J. c. v. Etrych, G. Martirosyan, A. Cao, J. A. P. Glidden, L. H. Dogra, J. M. Hutson, Z. Hadzibabic, and C. Eigen, Phys. Rev. Res. 5, 013174 (2023).
* van de Kraats _et al._ [2023] J. van de Kraats, D. J. M. Ahmed-Braun, J.-L. Li, and S. J. J. M. F. Kokkelmans, Phys. Rev. A 107, 023301 (2023).
* Pollack _et al._ [2009] S. E. Pollack, D. Dries, and R. G. Hulet, Science 326, 1683 (2009).
* Gross _et al._ [2009] N. Gross, Z. Shotan, S. Kokkelmans, and L. Khaykovich, Phys. Rev. Lett. 103, 163202 (2009).
* Gross _et al._ [2010] N. Gross, Z. Shotan, S. Kokkelmans, and L. Khaykovich, Phys. Rev. Lett. 105, 103203 (2010).
* Gross _et al._ [2011] N. Gross, Z. Shotan, O. Machtey, S. Kokkelmans, and L. Khaykovich, C. R. Phys. 12, 4 (2011).
* Dyke _et al._ [2013] P. Dyke, S. E. Pollack, and R. G. Hulet, Phys. Rev. A 88, 023625 (2013).
* Sørensen _et al._ [2012] P. K. Sørensen, D. V. Fedorov, A. S. Jensen, and N. T. Zinner, Phys. Rev. A 86, 052516 (2012).
* Yudkin and Khaykovich [2021] Y. Yudkin and L. Khaykovich, Phys. Rev. A 103, 063303 (2021).
* D’Incao _et al._ [2022] J. D’Incao, Y. Yudkin, P. Julienne, and L. Khaykovich, Bulletin of the American Physical Society (2022).
* Yudkin _et al._ [2023a] Y. Yudkin, P. S. Julienne, and L. Khaykovich, Phys. Rev. A 107, 053303 (2023a).
* Massignan and Stoof [2008] P. Massignan and H. T. C. Stoof, Phys. Rev. A 78, 030701(R) (2008).
* Wolf _et al._ [2017] J. Wolf, M. Deiß, A. Krükow, E. Tiemann, B. P. Ruzic, Y. Wang, J. P. D’Incao, P. S. Julienne, and J. H. Denschlag, Science 358, 921 (2017).
* Haze _et al._ [2022] S. Haze, J. P. D’Incao, D. Dorer, M. Deiß, E. Tiemann, P. S. Julienne, and J. H. Denschlag, Phys. Rev. Lett. 128, 133401 (2022).
* Secker _et al._ [2021b] T. Secker, J.-L. Li, P. M. A. Mestrom, and S. J. J. M. F. Kokkelmans, Phys. Rev. A 103, 022825 (2021b).
* Li _et al._ [2022] J.-L. Li, T. Secker, P. M. A. Mestrom, and S. J. J. M. F. Kokkelmans, Phys. Rev. Research 4, 023103 (2022).
* Secker _et al._ [2021c] T. Secker, J.-L. Li, P. M. A. Mestrom, and S. J. J. M. F. Kokkelmans, Phys. Rev. A 103, 032817 (2021c).
* Julienne and Hutson [2014] P. S. Julienne and J. M. Hutson, Phys. Rev. A 89, 052715 (2014).
* Tempest and Jonsell [2023] K.-M. Tempest and S. Jonsell, Phys. Rev. A 107, 053319 (2023).
* Yudkin _et al._ [2023b] Y. Yudkin, R. Elbaz, J. P. D’Incao, P. S. Julienne, and L. Khaykovich, arXiv 2308, 06237 (2023b).
* Mirahmadi and Pérez-Ríos [2022] M. Mirahmadi and J. Pérez-Ríos, International Reviews in Physical Chemistry 41, 233 (2022).
* Wigner [1927] E. Wigner, Nachr. Akad. Wiss. Goettingen, Math.-Physik Kl. IIa, 375 (1927).
* Moore [1973] J. H. Moore, Phys. Rev. A 8, 2359 (1973).
* Paschen and Back [1921] F. Paschen and E. Back, Physica 1, 261 (1921).
* Stoof _et al._ [1988] H. T. C. Stoof, J. M. V. A. Koelman, and B. J. Verhaar, Phys. Rev. B 38, 4688 (1988).
* Lee _et al._ [2007] M. D. Lee, T. Köhler, and P. S. Julienne, Phys. Rev. A 76, 012720 (2007).
* Moerdijk _et al._ [1996] A. J. Moerdijk, H. M. J. M. Boesten, and B. J. Verhaar, Phys. Rev. A 53, 916 (1996).
* Smirne _et al._ [2007] G. Smirne, R. M. Godun, D. Cassettari, V. Boyer, C. J. Foot, T. Volz, N. Syassen, S. Dürr, G. Rempe, M. D. Lee, K. Góral, and T. Köhler, Phys. Rev. A 75, 020702 (2007).
* Alt _et al._ [1967] E. Alt, P. Grassberger, and W. Sandhas, Nuclear Physics B 2, 167 (1967).
## Methods
### Three-body recombination rate
The three-body recombination rate constant $L_{3}$ in the zero-energy limit
can be formally expressed as [46, 47, 48, 37],
$L_{3}=\sum_{{\nu},c_{3}}L_{3}^{\mathrm{part}}(\varphi_{\nu},c_{3}),$ (2)
where,
$\displaystyle L_{3}^{\mathrm{part}}(\varphi_{\nu},c_{3})=\frac{12\pi
m}{\hbar}(2\pi\hbar)^{6}$ (3) $\displaystyle\times\lim_{E\rightarrow 0}\int
d\hat{\mathbf{q}}_{3}\
q_{3}\absolutevalue*{\matrixelement*{(\mathbf{q}_{3},c_{3}),\varphi_{\nu}}{U_{\alpha
0}(E)}{\Psi_{\mathrm{in}}}}^{2},$
reminiscent of Fermi’s golden rule. Each element
$\matrixelement*{(\mathbf{q}_{3},c_{3}),\varphi_{\nu}}{U_{\alpha
0}(z)}{\Psi_{\mathrm{in}}}$ describes transition from an incoming three-body
state with energy $E$ into a molecular state $\ket*{\varphi_{\nu}}$ of a pair
$\alpha$ and a third free particle with relative momentum $\mathbf{q}_{3}$ in
spin state $\ket*{c_{3}}$.
To calculate the operator $U_{\alpha 0}(E)$ we solve the inelastic three-body
scattering problem in momentum space, using the Alt-Grassberger-Sandhas (AGS)
equations [49] (see Supplemental Material). As any two-body subsystem can
transfer energy to the third particle, the three-body problem embeds the off-
shell solution to the two-body problem, which we obtain by exact
diagonalization of the two-body Hamiltonian. For the two-body interaction $V$
we use realistic Born-Oppenheimer singlet and triplet potentials, which allow
us to accurately account for the molecular level structure underpinning the
three-body recombination [38].
### Three-body spin models
To separate the contributions from two-body and three-body spin-exchange
processes, we use two different models for the spin coupling [35]. In a full
multichannel spin (FMS) model, we make no approximations and include all
three-body channels that conserve the total spin projection
$M_{F}=m_{f_{1}}+m_{f_{2}}+m_{f_{3}}$. In the fixed spectating spin (FSS)
model, we turn off three-body spin exchange processes by replacing the two-
body interaction $V$ with the restricted form,
$V_{\mathrm{FSS}}=V\ket*{c_{3}^{\mathrm{in}}}\bra{c_{3}^{\mathrm{in}}}.$ (4)
In this way the two-body interaction is turned off if the third particle is
not in the incoming spin state. In the AGS equations this replacement is
equivalent to restricting $\ket*{c_{3}}$ to $\ket*{c_{3}^{\mathrm{in}}}$
everywhere, such that $m_{f_{12}}=m_{f_{1}}+m_{f_{2}}$ and $m_{f_{3}}$ are
separately conserved.
### Fitting of the recombination rate
For large negative scattering length, universal theory predicts a power-law
increase of the three-body recombination rate coefficient, given as
$L_{3}=3C_{-}(a)\hbar a^{4}/m$. The function $C_{-}(a)$ gives the log-periodic
oscillation of the recombination rate arising from the Efimov effect. A two-
parameter analytical prediction for $C_{-}(a)$ can be obtained in the
formalism of effective field theory, which gives [1],
$C_{-}(a)=4590\
\frac{\sinh(2\eta_{-})}{\sin^{2}\left[s_{0}\ln(a/a_{-})\right]+\sinh^{2}(\eta_{-})}.$
(5)
To perform the fit of $L_{3}$ we first compute the dimensionless quantity
$mL_{3}/(3\hbar a^{4})$, and subsequently fit $C_{-}(a)$ directly. Provided
that we compute $L_{3}$ close to an Efimov resonance, then 3 data points are
typically sufficient to obtain a converged fit.
### Partial recombination probability
To examine the presence of spin-propensity for three-body recombination we
define an augmented partial recombination rate
$\mathcal{L}_{3}(\left\\{m_{s_{j}},m_{i_{j}}\right\\})$ as,
$\displaystyle\mathcal{L}_{3}(\left\\{m_{s_{j}},m_{i_{j}}\right\\})$
$\displaystyle=$
$\displaystyle\sum_{\nu,c_{3}}\absolutevalue*{\innerproduct*{m_{s_{1}},m_{i_{1}},m_{s_{2}},m_{i_{2}}}{\varphi_{\nu}}}^{2}$
$\displaystyle\times\absolutevalue*{\innerproduct*{m_{s_{3}},m_{i_{3}}}{c_{3}}}^{2}\frac{L_{3}^{\mathrm{part}}(\varphi_{\nu},c_{3})}{L_{3}}.$
By this definition $\mathcal{L}_{3}(\left\\{m_{s_{j}},m_{i_{j}}\right\\})$
averages the spin-projection
$\absolutevalue*{\innerproduct*{m_{s_{1}},m_{i_{1}},m_{s_{2}},m_{i_{2}}}{\varphi_{\nu}}}^{2}\absolutevalue*{\innerproduct*{m_{s_{3}},m_{i_{3}}}{c_{3}}}^{2}$
with respect to the discrete probability distribution
$L_{3}^{\mathrm{part}}(\varphi_{\nu},c_{3})/L_{3}$, and hence forms a measure
of the relative importance of a state defined by a set of quantum numbers
$\left\\{m_{s_{j}},m_{i_{j}}\right\\}$ for three-body recombination,
normalised to unity if summed over all available states. To obtain
$\mathcal{L}_{3}(M_{S})$ we sum
$\mathcal{L}_{3}(\left\\{m_{s_{j}},m_{i_{j}}\right\\})$ over all spin states
with definite $M_{S}$, which then finally gives the results shown in Figs.
3(c) and 3(d).
|
# Development of a sub-milimeter position sensitive gas detector ††thanks:
Supported by NSFC (11075095) and Shandong Province Science Foundation
(ZR2010AM015)
DU Yanyan1 XU Tongye1 SHAO Ruobin1
WANG Xu1 ZHU Chengguang1;1<EMAIL_ADDRESS>1 MOE key lab on particle physics
and particle irradiation,
Shandong University, Ji’nan 250100, China
###### Abstract
A position sensitive thin gap chamber has been developed. The position
resolution was measured using the cosmic muons. This paper presents the
structure of this detector, position resolution measurement method and
results.
###### keywords:
Thin gap chamber, Position resolution
###### pacs:
0
7.77.Ka, 29.40.Gx
## 1 Introduction
TGC( Thin Gap Chamber) used in ATLAS experiment [1] shows good performance in
the fast response and time resolution, but with limited position resolution.
The improvement of the position resolution with the timing performance
retained is straightforward for its flexibility to be used in the future
experiments and radiation measurement, for example, the upgrade of the trigger
system of ATLAS experiment. The main goal of the study described in this paper
aimed to build a prototype detector based on TGC, which can have a position
resolution better than 300${\mu}m$, while keeping timing performance not
deteriorated
The TGC detector operates in saturated mode by using a highly quenching gas
mixture of carbon dioxide and n-pentane, $55\%$ :$45\%$, which has many
advantages, such as small sensitivity to mechanical deformations, small
parallax, small Landau tails and good time resolution, but a position
sensitivity of around 1$cm$, decided by the geometrical width of the readout
channel and the strength of the induced signal. To improve the position
resolution, we concentrated on the improving of the method of signal readout
by fine tuning of the structure of the detector.
The new detector, named as pTGC(precision Thin Gap Chamber) based on the ATLAS
TGC, is constructed and tested. We found the position resolution can be
improved to be less than $300{\mu}m$, which meets the requirements.
In section 2, the structure of pTGC detector is described. Section 3 is
devoted to pTGC’s position resolution measurement. Results of the measurement
is summarized in section 4.
Figure 1: The schematic structure of pTGC-I chamber. Anode wires are placed in
the middle, with copper strip etched on the inner surface of the PCB board,
perpendicular to the wire direction.
Figure 2: The schematic structure of pTGC-II chamber. Compared to pTGC-I,
additional isolation layer and graphite layer cover the etched copper strips.
## 2 Construction of pTGC
In the pTGC development, two versions of detectors are constructed and tested,
which are referred as $pTGC-I$ in the first stage and $pTGC-II$ in the second
stage, respectively.
The schematic structure of $pTGC-I$ is shown in Fig. 2, similar to the
structure of ATLAS TGC, except that the position of the strips for signal
collection are modified. 48 copper strips of $0.8mm$ wide and $0.2mm$ spaced
are etched on the inner surface of the 2 parallel PCB boards, which form a
thin spaced chamber. The wires, segmented at $1.8mm$ interval and
perpendicular to the strip direction, are sandwiched in between the two PCB
boards. The resulted size of the detector is defined by the number of wires
and strips, which are $290mm{\times}50mm$.
In the test of $pTGC-I$, the discharge happened between wires and strips
resulted in fatal damage on the frontend electronics, even though we have
designed a protection circuit to insert between detector and frontend
electronics board. This means an instability for big detector and for long
time running. Besides, the induced charge on strips spread roughly $5$ to
$6mm$, which leaves limited rooms for reducing the quantity of the channels by
enlarging the width of strips. Based on $pTGC-I$, the $pTGC-II$ is developed
to deal with these problems.
The schematic structure of $pTGC-II$ is shown in the Fig. 2. The strip width
is enlarged to $3.8mm$ ($0.2mm$ spaced), and a thin ($100{\mu}m$) insolation
layer is pasted on the strip layer. The isolation layer is then coated with a
thin ( $30{\mu}m$) graphite layer as the electric ground to form the electric
field with wires. This graphite layer acts as the protector of the frontend
electronics from discharge and can enlarge the spreading size of the induced
charge on the strip layer. We tune the resistivity of the graphite layer to be
around $100k\Omega$, considering the diffusion speed of the charge, as well.
the resulted size of $pTGC-II$ is $290mm*200mm$.
Both detector use gas mixture of carbon dioxide and n-pentane, $55\%$ :$45\%$,
as working gas, and the anode wire is set to high voltage of 2900v, which are
all the same configuration as the ATLAS TGC detector to maintain the its
features relative to the time measurement of the detector.
## 3 Position resolution measurement
With 3 layers of identical chambers placed in parallel, and 2 layers of
scintillator detectors to build a muon hodoscope(see Fig 3), the $pTGC-I$ and
$pTGC-II$ detectors are tested. The induced charge on each strip is integrated
for the the position calculation based on the charge center-of-gravity
algorithms. The measured hitting position on the 3 layers of chamber are
supposed to be aligned into a straight line concerning the penetration power
of muons. The residue of the position relative to the straight line is then
used to calculate the position resolution of the detectors.
### 3.1 Signal definition
Using oscilloscope, we first observed the induced signal in one wire group and
3 adjacent strips (limited by channels of oscilloscope), as shown in Fig. 4.
It’s apparent that the signals are great significant above the noise and the
signals on the strip are in different magnitudes as expected.
For position resolution measurement, we designed a much more complicated
DAQ(data acquisition) system based on gassiplex frontend electronics [2] to
readout and digitize the induced charge from a quantity of channels of the 3
chambers in a more complex hodoscope [3]. Once the two scintillator detector
of the hodoscope are both fired, the DAQ is triggered. The trigger signal is
sent to the detector front end electronics, which then close the gate for the
discharge of capacitance which has integrated the signal charge on. The charge
on the capacitance are then read out one by one controlled by the clock
distributed from the DAQ system. The charge are then digitized and saved into
computer.
The digitized charge, denoted by $Q_{i}$ where $i$ is the channel number,
consists of three parts: electronic pedestal, noise, and charge induced by
muon hit.
First of all, we need to figure out the pedestal and noise for each channel.
The method is to histogram the integrated charge for each channel using a soft
trigger where no real muon induced signal appear in the data. Fitting the
histogram with a gaussian function to get the pedestal and the noise, denoted
by $P_{i}$ and $\sigma_{i}$, as shown in Fig. 5, where the height of the
histogram represents the pedestal and the error bar represents the noise of
that channel.
In the analysis, if $Q_{i}>P_{i}+3\sigma_{i}$, the channel is considered to be
fired by real muon hit, and the signal charge is calculated as:
$S_{i}=Q_{i}-P_{i},$ (1)
Figure 3: The comic muon hodoscope used for the chamber testing. Plastic
scintillator detector are used for trigger. 3 identical pTGC chambers placed
in parallel in between the 2 scintillator detectors. Figure 4: The observed
signals on wires and several copper strips induced by the same incident cosmic
ray. The signal on wire are negative, and the signal on strips are positive.
Figure 5: The noise and pedestal distribution of 96 signal strips in one
chamber (The x-axis is the signal strip number, the vertical coordinate is the
pedestal value and the error bars presents the noise of that channel.)
The signal magnitude distribution of the largest signal in each cluster
(cluster definition is in next section), named as the peak signal, is shown in
Fig. 6. The distribution of the second largest signal in each cluster, named
as second peak signal, is shown in Fig. 7. The distribution of the sum of all
charge in one cluster is shown in Fig. 8. The correspondence between the
magnitude of the signal and the charge is $1fC/3.6bits$. We can then calculate
that the maximum probable charge of the largest signal in one cluster is
$69fC$, the maximum probable total charge of one cluster is $470fC$, which is
consistent with the measurement in [1]
### 3.2 Cluster definition
The induced charge by the incident muons are distributed on several adjacent
strips, which are grouped in ”cluster” in the analysis and used for the hit
position calculation. In one event, we search all the channels of one
detector, and define group of fired adjacent strips without space as a
cluster. To suppress the fake signals from noise, if the cluster contains only
one strip, the cluster is dropped. The cluster size and number of cluster per
detector per event are shown in Fig. 9 for $pTGC-I$ and Fig. 10 for $pTGC-II$.
It can be seen that in both cases one cluster contain average six strips and
almost every event contains one cluster, which is consistent with the
expected. The hit position is then calculated for each cluster by
$x=\sum_{i}(S_{i}*x_{i})/\sum_{i}(S_{i}),$ (2)
where $x_{i}$ is the center coordination of the $i-th$ strip.
Figure 6: The distribution of the largest signal in one cluster. The x-axis is
the digitized charge collected.
Figure 7: The distribution of the second largest signal in one cluster
Figure 8: The distribution of total charge induced in one cluster
Figure 9: For pTGC-I: (Left) The distribution of cluster size (quantity of
strips in one cluster). (Right) The quantity of cluster in one chamber per
triggered event.
Figure 10: For pTGC-II: (Left) The distribution of cluster size (quantity of
strips in one cluster). (Right) The quantity of cluster in one chamber per
triggered event.
### 3.3 Position resolution
As redundant design, the strips are etched on both inner surface of the PCB
boards. Signals will be induced by the same avalanche on the 2 face-to-face
strips, which corresponds to an double measurements of a single hit. To
compare the two measurements, denoted as $x_{1}$ and $x_{1}^{\prime}$, we fill
$x_{1}-x_{1}^{\prime}$ into histogram to see the broadness of the
distribution. From a simple gaussian function fit, we observed a narrow width
of around $36{\mu}m$, which means that the electronics noise effect on the
resolution is much small. This is consistent with the expectation when to
compare Fig. 5 and Fig. 6, where it shows the signal is great significant
compared to the noise.
After the three hit positions $x_{1}$, $x_{2}$, $x_{3}$ are calculated for the
3 parallel chambers, to simplify the calculation, we first use $x_{1}$ and
$x_{3}$ to calculate the expected hit position on the second layer $x_{2c}$:
$x_{2c}=x_{1}\frac{L_{23}}{L_{12}+L_{23}}+x_{3}\frac{L_{12}}{L_{12}+L_{23}},$
(3)
where $L_{12}$ and $L_{23}$ are the vertical distance between the detector 1,2
and 2,3. To assume the same position resolution $\sigma$ for the 3 identical
detectors, we know the resolution of $x_{2c}$, with the error propagation, is:
$\sigma_{2c}=\sqrt{\frac{L_{23}^{2}}{(L_{12}+L_{23})^{2}}+\frac{L_{12}^{2}}{(L_{12}+L_{23})^{2}}}\sigma{\equiv}k\sigma,$
(4)
Filling $x_{2}-x_{2c}$ into the histogram and then fit with gaussian function,
the width is $w=\sqrt{1+k^{2}}\sigma$. So we can directly calculate the
position resolution of the detector as
$\sigma=\frac{w}{\sqrt{1+k^{2}}}.$ (5)
From Fig. 11 and Fig. 12, we can obtain that the position resolution are
$359um$ for $pTGC-I$ and $233um$ for $pTGC-II$. In both of the cases, the
detector resolution has reach our design requirement. In test, we see that
$pTGC-II$ are more stable with the graphite layer protection and achieve a
better resolution even with less channels.
Figure 11: The distribution of $x_{2}-x_{2c}$ for pTGC-I. The corresponding
position resolution of the chamber is
$\sigma=\frac{w}{\sqrt{1+k^{2}}}=\frac{439{\mu}m}{1.22}=359{\mu}m$.
Figure 12: The distribution of $x_{2}-x_{2c}$ for pTGC-II. The corresponding
position resolution of the chamber is
$\sigma=\frac{w}{\sqrt{1+k^{2}}}=\frac{286{\mu}m}{1.22}=233{\mu}m$.
To look at the dependence of the position sensitivity of the detector to the
incident angle of the muon, we divide the data into groups. Each group of data
contains the events of muon with specific incident angle. To redo the analysis
above, the result is shown in Fig. 13, which shows that the position
resolution of $pTGC-II$ is insensitive to the incident angle of muons.
To check the effect of the electronic noise, we use part of the top highest
signals in one cluster to calculate the position resolution. The result is
shown in Fig. 14, which shows that the resolution are similar and the
electronic noise doesn’t affect much.
Figure 13: The position resolution variance relative to the incident angle of
the cosmic rays. The x-axis is the incident angle of cosmic rays.
Figure 14: The position resolution variance relative to the quantity of strips
in one cluster used for position calculation. The x-axis is the quantity of
strips in one cluster used for position calculation.
## 4 Summary
Two pTGC version $pTGC-I$ and $pTGC-II$, have been constructed and tested.
With the basic structure and working gas unchanged, the detector can attains
the exiting features like good time resolution and fast response, which are
essential for trigger. By revising the signal collecting structure and method,
the position resolution is improved from the level of centimeter to be less
than $300{\mu}m$, which meet the requirement of design. To be noticed that the
resolution measured is a global resolution of the detector, which include the
effect of the non-uniformity of the detector all over the sensitive area. The
3 detectors are placed in parallel with mechanical method, the relative
rotation of the 3 detectors will deteriorate the final measured resolution,
which means that the measured resolution is much conservative.
## References
* [1] Atlas Collaboration, ATLAS muon spectrometer: Technical design report, 1997. CEAN/LHCC/97-22.
* [2] Liu Minghui $etal.$ Nuclear electronics and detection technology, 2008(5)
* [3] Xu Tongye $etal.$ arXiv:1308.5751
|
# On Stability and Denominators of $F$-pure Thresholds in Families of Diagonal
Hypersurfaces
Gari Lincoln Chua Department of Mathematics and Statistics, University of New
Mexico, 1 University of New Mexico, MSC01 1115
Albuquerque, NM 87131, United States of America<EMAIL_ADDRESS>
###### Abstract.
Given a prime number $p$ and a positive integer $m$, we provide a family of
diagonal hypersurfaces $\\{f_{n}\\}_{n=1}^{\infty}$ in $m$ variables, for
which the denominator of $\operatorname{fpt}(f_{n})$ (in lowest terms) is
always $p$ and whose $F$-pure thresholds stabilize after a certain $n$. We
also provide another family of diagonal hypersurfaces
$\\{g_{n}\\}_{n=1}^{\infty}$ in $m$ variables, for which the power of $p$ in
the denominator of $\operatorname{fpt}(g_{n})$ (in lowest terms) diverges to
$\infty$ as $n\to\infty$. This behavior of the denominator of the $F$-pure
thresholds is dependent on the congruence class of $p$ modulo the smallest two
exponents of $\\{f_{n}\\}$ and $\\{g_{n}\\}$.
## 1\. Introduction
Let $R=k[x_{1},\dots,x_{n}]$ where $k$ is a perfect field of characteristic
$p>0$ and denote the maximal ideal $\mathfrak{m}=(x_{1},\dots,x_{n})$. The
$F$-pure threshold of a polynomial $f\in R$, introduced in [10], is defined to
be
$\operatorname{fpt}(f)=\sup_{e}\left\\{\frac{N}{p^{e}}:f^{N}\not\in\mathfrak{m}^{[p^{e}]}\right\\}$,
where $\mathfrak{m}^{[p^{e}]}=(x_{1}^{p^{e}},\dots,x_{n}^{p^{e}})$ is the
$e$th Frobenius power of $\mathfrak{m}$. The $F$-pure threshold
$\operatorname{fpt}(f)$ is a rational number in $(0,1]$ that is an invariant
of positive characteristic singularities, with smaller values meaning that the
hypersurface is “more singular."
The $F$-pure threshold (fpt) is closely related to the log canonical threshold
(lct), an invariant of characteristic $0$ singularities. If
$f\in\mathbb{Q}[x_{1},\dots,x_{n}]$ and $\overline{f}$ is the reduction of $f$
modulo $p$, both $\operatorname{lct}(f)$ and
$\operatorname{fpt}(\overline{f})$ are rational numbers between $0$ and $1$.
Moreover,
$\operatorname{fpt}(\overline{f})\leq\operatorname{lct}(f)\hskip 10.0pt\text{
and }\hskip
10.0pt\lim_{p\to\infty}\operatorname{fpt}(\overline{f})=\operatorname{lct}(f).$
This motivates the following question (see for example, [8]): When is
$\operatorname{fpt}(\overline{f})=\operatorname{lct}(f)$, and if
$\operatorname{fpt}(\overline{f})\neq\operatorname{lct}(f)$, does $p$ always
divide the denominator of $\operatorname{fpt}(\overline{f})$?
The answer to the previous question is complicated. Canton, Hernandez,
Schwede, and Witt in [2] have shown a family of polynomials $f$ for which
$\operatorname{fpt}(\overline{f})\neq\operatorname{lct}(f)$ but the
denominator of $\operatorname{fpt}(\overline{f})$ is not a multiple of $p$. On
the other hand, [3], [5], [7] have shown that in many classes of polynomials,
if $\operatorname{fpt}(\overline{f})\neq\operatorname{lct}(f)$, then the
denominator of $\operatorname{fpt}(\overline{f})$ must be a multiple of $p$.
In particular, in [5], the authors have shown that if $f$ is a quasi-
homogeneous polynomial with an isolated singularity at the origin, then
$\operatorname{fpt}(\overline{f})$ is either $\operatorname{lct}(f)$ or its
denominator must be a power of $p$. The same paper also gives a bound on the
power of $p$ (depending on $\overline{f}$) appearing in the denominator of
$\operatorname{fpt}(\overline{f})$ depending on the choice of grading on
$k[x_{1},\dots,x_{n}]$ that turns $\overline{f}$ homogeneous.
The aformentioned bound in the power of $p$ in the $F$-pure threshold
motivates the following question: Given a family of quasi-homogeneous
polynomials $\\{f_{n}\\}$ in characteristic $p$ for which
$\operatorname{fpt}(f_{n})\neq\operatorname{lct}(f_{n})$ for all $n$, is there
a positive integer $M$ such that the power of $p$ in the denominator of
$\operatorname{fpt}(f_{n})$ is less than or equal to $M$ for all $n$?
In this paper, we consider families of diagonal hypersurfaces; i.e.,
hypersurfaces cut out by polynomials of the form
$x_{1}^{a_{1}}+\dots+x_{m}^{a_{m}}$, where $m>2$, the exponents $a_{1}$ and
$a_{2}$ are always fixed, while the other exponents may vary across the
natural numbers. We will see that the power of $p$ in the $F$-pure thresholds
of these families is either bounded by 1, or diverges to $\infty$ depending on
whether $p\equiv-1$ or $1\mod\operatorname{lcm}(a_{1},a_{2})$. This provides
some interesting behavior of $F$-pure thresholds. Our main results are as
follows:
###### Theorem 1.1.
Let
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a_{1}}+x_{2}^{a_{2}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{n}\in
k[x_{1},\dots,x_{m}]$, where $1<a_{1}\leq a_{2}\leq\dots\leq a_{m-1}$ are
fixed integers, $n\geq a_{m-1}$, and $(a_{1},a_{2})\neq(2,2)$. If
$p\geq\frac{a_{1}a_{2}\left(a_{1}+a_{2}+\frac{1}{a_{1}}+\frac{1}{a_{2}}-4\right)}{a_{1}a_{2}-a_{1}-a_{2}}$
and $p\equiv-1\mod\operatorname{lcm}(a_{1},a_{2})$, then either
$\operatorname{fpt}(f_{n})=1$ or the denominator of
$\operatorname{fpt}(f_{n})$ (in lowest terms) is $p$. Moreover,
$\operatorname{fpt}(f_{n})=\operatorname{fpt}(x_{1}^{a_{1}}+x_{2}^{a_{2}}+x_{3}^{a_{3}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{p})$
for all $n\geq p$.
###### Theorem 1.2.
Let
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a}+x_{2}^{b}+x_{3}^{n}+x_{4}^{n}+\dots+x_{m}^{n}$,
where $1<a\leq b$ are fixed integers, integers, $n\geq b$, and
$(a,b)\neq(2,2)$. If $p\equiv 1\mod\operatorname{lcm}(a,b)$ and $p^{d}\leq
n<p^{d+1}$ for some positive integer $d$ and
$\operatorname{fpt}(f_{n})\neq\frac{1}{a}+\frac{1}{b}+\frac{m-2}{n}$, then the
power of $p$ in the denominator of $\operatorname{fpt}(f_{n})$ (in lowest
terms) is at least $d$.
###### Theorem 1.3.
Let $s$ be a positive integer between 2 and $p-1$ inclusive. For $n\geq 2$,
let
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a}+x_{2}^{b}+x_{3}^{p^{n}-s}+x_{4}^{p^{n}-s}+\dots+x_{m}^{p^{n}-s}$,
where $1<a\leq b$ are integers and $(a,b)\neq(2,2)$. Let
$r=\left\lceil\log_{s}\left(\left(1-\frac{1}{a}-\frac{1}{b}\right)(p-1)\right)\right\rceil$.
If $p\equiv 1\mod\operatorname{lcm}(a,b)$ and $s^{r}<p-1$, then the power of
$p$ in the denominator of $\operatorname{fpt}(f_{n})$ is exactly
$d\left(r+1\right)-1$.
###### Corollary 1.4.
Given a prime $p>5$ and positive integers $\ell,m$ with $m>2$, there exists a
family of diagonal hypersurfaces $\\{f_{n}\\}_{n\in\mathbb{N}}$ in $m$
variables such that the power of $p$ in the denominator of
$\operatorname{fpt}(f_{n})$ is less than or equal to $\ell$ for all
$n\in\mathbb{N}$, and
$\operatorname{fpt}(f_{n})=\operatorname{fpt}(f_{p^{\ell}})$ for all $n\geq
p^{\ell}$. There also exists a different family of diagonal hypersurfaces
$\\{g_{n}\\}_{n\in\mathbb{N}}$ such that the power of $p$ in the denominator
of $\operatorname{fpt}(g_{n})$ diverges to $\infty$ as $n\to\infty$.
Acknowledgements. The author thanks Dr. Janet Vassilev for the many insightful
discussions.
## 2\. Nonterminating Base $p$ Expansions
In this section, we review relevant results regarding nonterminating base $p$
expansions. The content of this section is lifted from [3] and [5].
###### Definition 2.1.
Let $\alpha\in\mathbb{Q}\cap[0,1]$ and $p$ be a prime number. The
nonterminating base $p$ expansion of $\alpha$ is the expression
$\alpha=\sum_{e=1}^{\infty}\alpha^{(e)}p^{-e}$, where for each $e$,
$0\leq\alpha^{(e)}\leq p-1$, and the sequence
$\\{\alpha^{(e)}\\}_{e\in\mathbb{N}}$ is not eventually zero. The
nonterminating base $p$ expansion of $\alpha$ is unique, and for each positive
integer $e$, we call $\alpha^{(e)}$ the $e$-th digit of $\alpha$ (base $p$).
###### Definition 2.2.
Let $\alpha\in\mathbb{Q}\cap[0,1]$ and $p$ be a prime number. For
$e\in\mathbb{N}$, define the $e$-th truncation of $\alpha$ (base $p$) as
$\left\langle\alpha\right\rangle_{e}=\alpha^{(1)}p^{-1}+\dots+\alpha^{(e)}p^{-e}$,
and for convenience, we may also write
$\left\langle\alpha\right\rangle_{e}=.\alpha^{(1)}:\alpha^{(2)}:\dots:\alpha^{(e)}(\text{base
}p).$
By convention, $\left\langle\alpha\right\rangle_{0}=0$ and
$\left\langle\alpha\right\rangle_{\infty}=\alpha$.
Some relevant properties of nonterminating base $p$ expansions are encoded in
the following lemma:
###### Lemma 2.3.
Fix $\alpha,\beta\in\mathbb{Q}\cap[0,1]$.
1. (1)
If $\left\langle\alpha\right\rangle_{e}\neq 0$, then the denominator of
$\left\langle\alpha\right\rangle_{e}$ is less than or equal to $p^{d}$ for
some $d\leq e$.
2. (2)
$\alpha\leq\beta$ if and only if
$\left\langle\alpha\right\rangle_{e}\leq\left\langle\beta\right\rangle_{e}$
for all $e\geq 1$; if $\alpha<\beta$, these inequalities are strict for $e\gg
0$.
3. (3)
If $(p^{d}-1)\alpha\in\mathbb{Z}$ for some positive integer $d$, then the base
$p$ expansion of $\alpha$ is periodic with period dividing $d$.
###### Proof.
(1) and (2) follow from the definition; see [5, Lemmas 2.5 and 2.6] for (3). ∎
## 3\. $F$-pure Thresholds of Diagonal and Quasi-homogeneous Polynomials
From this point onward, we set $k$ to be a perfect field of characteristic
$p>0$ (typically, $k=\mathbb{F}_{p}$), $R=k[x_{1},\dots,x_{n}]$ the polynomial
ring in $n$ variables over $k$, and $\mathfrak{m}=(x_{1},\dots,x_{n})\subseteq
R$ be the maximal ideal of $R$ generated by the variables.
###### Definition 3.1.
Let $f\in\mathfrak{m}$. For each positive integer $e$, define
$\nu_{f}(p^{e})=\sup\left\\{N\in\mathbb{Z}:f^{N}\notin\mathfrak{m}^{[p^{e}]}\right\\}$
where $\mathfrak{m}^{[p^{e}]}=(x^{p^{e}}_{1},\dots,x^{p^{e}}_{n})$ denotes the
$e$-th Frobenius power of $\mathfrak{m}$. Define the $F$-pure threshold of $f$
as
$\operatorname{fpt}(f)=\lim_{e\to\infty}\frac{\nu_{f}(p^{e})}{p^{e}}.$
###### Remark 3.2.
For any $f\in\mathfrak{m}$, the sequence
$\\{p^{-e}\nu_{f}(p^{e})\\}_{e=1}^{\infty}$ is a nondecreasing sequence of
rational numbers contained in $(0,1]$ (see [6], Lemma 1.1 and Remark 1.2 for
the proof). Therefore, the limit defining the $F$-pure threshold of $f$
exists, and the $F$-pure threshold is in $(0,1]$.
Developing formulas and algorithms for $F$-pure threshold for certain classes
of polynomials has been an active area of research. For example, see [3] for
$F$-pure thresholds of diagonal hypersurfaces, [4] for $F$-pure thresholds of
binomial hypersurfaces, [1] and [7] for $F$-pure thresholds of Calabi-Yau
hypersurfaces, and [9] for $F$-pure thresholds of generic homogeneous
polynomials. We record some relevant results.
###### Theorem 3.3 ([3] Theorem 3.4, Remark 3.5).
Let $f\in k[x_{1},\dots,x_{n}]$ be a $k^{*}$-linear combination of the
monomials $x_{1}^{d_{1}},\dots,x_{n}^{d_{n}}$. Let $L=\inf\\{e\geq
0:(d_{1}^{-1})^{(e+1)}+\dots+(d_{n}^{-1})^{(e+1)}\geq p\\}$.
* •
If $L<\infty$, then $\operatorname{fpt}(f)=\left\langle
d_{1}^{-1}\right\rangle_{L}+\dots+\left\langle
d_{n}^{-1}\right\rangle_{L}+p^{-L}$.
* •
If $L=\infty$, then $\operatorname{fpt}(f)=d_{1}^{-1}+\dots+d_{n}^{-1}$.
###### Theorem 3.4 ([5], Theorem 3.5, Corollary 3.9).
Fix an $\mathbb{N}$-grading on the polynomial ring $k[x_{1},\dots,x_{n}]$, and
let $f$ be a homogeneous polynomial under this grading. Write
$\lambda=\frac{\deg x_{1}+\dots+\deg x_{n}}{\deg f}=\frac{s}{t}$ in lowest
terms.
* •
If $\operatorname{fpt}(f)\neq\lambda$, then
$\operatorname{fpt}(f)=\left\langle\lambda\right\rangle_{L}-\frac{E}{p^{L}}$
for some pair $(L,E)\in\mathbb{N}^{2}$ satisfying $L\geq 1$ and $1\leq E\leq
n-1-\frac{\left\lceil[sp^{L}\%t]+s\right\rceil}{t}$, where $[m\%t]$ denotes
the smallest positive residue of $m$ modulo $t$.
* •
If $\operatorname{fpt}(f)\neq\lambda$ and $p\not|t$, then the power of $p$ in
the denominator of $\operatorname{fpt}(f)$ (in lowest terms) is less than or
equal to $M=2\phi(t)+\left\lceil\log_{2}(n-1)\right\rceil$, where $\phi$
denotes Euler’s totient function.
These two important theorems tell us that the $F$-pure thresholds of certain
classes of polynomials can be computed off nonterminating base $p$ expansions.
In particular, the proofs of the main results will be based on finding $L$ and
then using these two theorems.
The second part of Theorem 3.4 says that the power of $p$ appearing in the
denominator of the $F$-pure threshold of a quasi-homogeneous polynomial $f$ is
bounded by a number depending on the $\mathbb{N}$-grading that turns $f$
homogeneous. This along with [7], Theorem 4.3, motivates the premise of the
paper: if we consider a family of quasi-homogeneous polynomials where one or
more exponents vary through the family, what is the growth of the power of $p$
in the denominator of the $F$-pure thresholds?
## 4\. Main Results
We now prove the main results.
###### Theorem 4.1.
Let
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a_{1}}+x_{2}^{a_{2}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{n}\in
k[x_{1},\dots,x_{m}]$, where $1<a_{1}\leq a_{2}\leq\dots\leq a_{m-1}$ are
fixed integers, $n\geq a_{m-1}$, and $(a_{1},a_{2})\neq(2,2)$. If
$p\geq\frac{a_{1}a_{2}\left(a_{1}+a_{2}+\frac{1}{a_{1}}+\frac{1}{a_{2}}-4\right)}{a_{1}a_{2}-a_{1}-a_{2}}$
and $p\equiv-1\mod\operatorname{lcm}(a_{1},a_{2})$, then either
$\operatorname{fpt}(f_{n})=1$ or the denominator of
$\operatorname{fpt}(f_{n})$ (in lowest terms) is $p$. Moreover,
$\operatorname{fpt}(f_{n})=\operatorname{fpt}(x_{1}^{a_{1}}+x_{2}^{a_{2}}+x_{3}^{a_{3}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{p})$
for all $n\geq p$.
###### Proof.
Since $p\equiv-1\mod\operatorname{lcm}(a_{1},a_{2})$, we must have
$p\equiv-1\mod a_{1}$ and $p\equiv-1\mod a_{2}$. Therefore, the first $2$
digits of the nonterminating base $p$ expansion of $\frac{1}{a_{1}}$ satisfies
$\left(\frac{1}{a_{1}}\right)^{(1)}=\frac{p-(a_{1}-1)}{a_{1}},\hskip
10.0pt\left(\frac{1}{a_{1}}\right)^{(2)}=\left\lfloor
p\left(\frac{a_{1}-1}{a_{1}}\right)\right\rfloor\geq(a_{1}-1)\left(\frac{p-(a_{1}-1)}{a_{1}}\right),$
and similarly for $\frac{1}{a_{2}}$. Now,
$(a_{1}-1)\left(\frac{p-(a_{1}-1)}{a_{1}}\right)+(a_{2}-1)\left(\frac{p-(a_{2}-1)}{a_{2}}\right)\geq
p\iff
p\geq\frac{a_{1}a_{2}\left(a_{1}+a_{2}+\frac{1}{a_{1}}+\frac{1}{a_{2}}-4\right)}{a_{1}a_{2}-a_{1}-a_{2}}.$
Therefore, for these $p$, we have $L<2$. On the other hand,
$\left(\frac{1}{a_{1}}\right)^{(1)}+\left(\frac{1}{a_{2}}\right)^{(1)}+\left(\frac{1}{n}\right)^{(1)}\geq
p\iff
n\leq\frac{p}{(p+2)\left(1-\frac{1}{a_{1}}-\frac{1}{a_{2}}\right)}<1-\frac{a_{1}+a_{2}}{a_{1}a_{2}-a_{1}-a_{2}}.$
Since $(a_{1},a_{2})\neq(2,2)$, $a_{1}a_{2}>a_{1}+a_{2}$, and so the last
inequality implies that to “carry at the first place," we must have $n<1$.
This is impossible. Therefore, $L\geq 1$. This then proves that $L=1$, and so
$\operatorname{fpt}(f)=\left\langle\frac{1}{a_{1}}\right\rangle_{1}+\left\langle\frac{1}{a_{2}}\right\rangle_{1}+\dots+\left\langle\frac{1}{a_{m-1}}\right\rangle_{1}+\left\langle\frac{1}{n}\right\rangle_{1}+\frac{1}{p}$
by Theorem 3.3. Since
$\left(\frac{1}{a_{1}}\right)^{(1)}+\left(\frac{1}{a_{2}}\right)^{(1)}=(p+1)\left(\frac{1}{a_{1}}+\frac{1}{a_{2}}\right)-2<p$
for all $(a_{1},a_{2})\neq(2,2)$, this proves the result.
Finally, if $n\geq p$, then $\left(\frac{1}{n}\right)^{(1)}=0$, and
$\operatorname{fpt}(f)=\left\langle\frac{1}{a_{1}}\right\rangle_{1}+\left\langle\frac{1}{a_{2}}\right\rangle_{1}+\dots+\left\langle\frac{1}{a_{m-1}}\right\rangle_{1}+\frac{1}{p}=\operatorname{fpt}(x_{1}^{a_{1}}+x_{2}^{a_{2}}+x_{3}^{a_{3}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{p}).$
∎
Theorem 4.1 shows that for certain families of hypersurfaces
$\\{f_{n}\\}_{n\in\mathbb{N}}$, the $F$-pure thresholds stabilize as
$n\to\infty$ and their denominators are always either 1 or $p$. On the other
hand, by considering different primes and indexing, we can make the power of
$p$ in the $F$-pure thresholds of a family of diagonal hypersurfaces become
unbounded as $n\to\infty$. One such method to do this is in the next result.
###### Theorem 4.2.
Let
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a}+x_{2}^{b}+x_{3}^{n}+x_{4}^{n}+\dots+x_{m}^{n}$,
where $1<a\leq b$ are fixed integers, integers, $n\geq b$, and
$(a,b)\neq(2,2)$. If $p\equiv 1\mod\operatorname{lcm}(a,b)$ and $p^{d}\leq
n<p^{d+1}$ for some positive integer $d$ and
$\operatorname{fpt}(f_{n})\neq\frac{1}{a}+\frac{1}{b}+\frac{m-2}{n}$, then the
power of $p$ in the denominator of $\operatorname{fpt}(f_{n})$ (in lowest
terms) is at least $d$.
###### Proof.
For all $e\geq 1$,
$\left(\frac{1}{a}\right)^{(e)}+\left(\frac{1}{b}\right)^{(e)}=(p-1)\left(\frac{1}{a}+\frac{1}{b}\right)<p$
for all $(a,b)$. Moreover, $p^{d}\leq n<p^{d+1}$ implies that
$\left(\frac{1}{n}\right)^{(e)}=0$ for all $1\leq e\leq d$. Therefore, $L\geq
d$. If $L=\infty$, then
$\operatorname{fpt}(f)=\frac{1}{a}+\frac{1}{b}+\frac{m-2}{n}.$
Otherwise,
$\operatorname{fpt}(f)=\left\langle\frac{1}{a}\right\rangle_{L}+\left\langle\frac{1}{b}\right\rangle_{L}+(m-2)\left\langle\frac{1}{n}\right\rangle_{L}+\frac{1}{p^{L}},$
which has denominator $p^{L}$ in lowest terms. ∎
Theorem 4.2 shows a linear asymptotic growth in the power of $p$ in the
denominator of the $F$-pure thresholds. If we restrict to a subsequence, we
can make the growth of the power of $p$ faster.
###### Theorem 4.3.
Let $s$ be a positive integer between 2 and $p-1$ inclusive. For $n\geq 2$,
let
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a}+x_{2}^{b}+x_{3}^{p^{n}-s}+x_{4}^{p^{n}-s}+\dots+x_{m}^{p^{n}-s}$,
where $1<a\leq b$ are integers and $(a,b)\neq(2,2)$. Let
$r=\left\lceil\log_{s}\left(\left(1-\frac{1}{a}-\frac{1}{b}\right)(p-1)\right)\right\rceil$.
If $p\equiv 1\mod\operatorname{lcm}(a,b)$ and $s^{r}<p-1$, then the power of
$p$ in the denominator of $\operatorname{fpt}(f_{n})$ is exactly
$d\left(r+1\right)-1$.
###### Proof.
Let $t=p^{n}-s$. We compute the nonterminating base $p$ expansion of
$\frac{1}{t}=\frac{1}{p^{n}-s}$:
$\frac{1}{p^{n}-s}=\left(\frac{1}{p^{n}}\right)\left(\frac{1}{1-\frac{s}{p^{n}}}\right)=\sum_{q=0}^{\infty}\frac{s^{q}}{p^{n(q+1)}}.$
Note that if $s^{q}\leq p-1$, then
$\left(\frac{1}{p^{n}-s}\right)^{(n(q+1))}=s^{q}$ and
$\left(\frac{1}{p^{n}-s}\right)^{(e)}=0$ for $e<n(q+1)$ and $d$ not dividing
$e$. Now,
$\left(\frac{1}{a}\right)^{(d(q+1))}+\left(\frac{1}{b}\right)^{(n(q+1))}+\left(\frac{1}{t}\right)^{(n(q+1))}>p-1\iff
q>\log_{s}\left((p-1)\left(1-\frac{1}{a}-\frac{1}{b}\right)\right).$
Therefore, we may choose
$q=\left\lceil\log_{s}\left(\left(1-\frac{1}{a}-\frac{1}{b}\right)(p-1)\right)\right\rceil$
(assuming $s^{q}<p-1$), and so, $L=n\left(q+1\right)-1$. ∎
###### Remark 4.4.
We consider the special case $m=3$ and fix the family of diagonal
hypersurfaces $f_{n}(x,y,z)=x^{a}+y^{b}+z^{n}$. Then Theorems 4.1 and either
4.2 or 4.3 take the following form: For any
$p\equiv-1\mod\operatorname{lcm}(a,b)$, then the denominator of
$\operatorname{fpt}(f_{n})$ is either $1$ or $p$, and
$\operatorname{fpt}(f_{n})=\operatorname{fpt}(f_{p})$ for all $n\geq p$. For
any $p\equiv 1\mod\operatorname{lcm}(a,b)$ and $n=p^{d}-s$ for some integer
$1<s<p-1$, then for
$r=\left\lceil\log_{s}\left(\left(1-\frac{1}{a}-\frac{1}{b}\right)(p-1)\right)\right\rceil$,
if $s^{r}<p-1$, then the power of $p$ in the denominator of
$\operatorname{fpt}(f_{n})$ is exactly $d\left(r+1\right)-1$.
###### Corollary 4.5.
Given a prime $p>5$ and positive integers $\ell,m$ with $m>2$, there exists a
family of diagonal hypersurfaces $\\{f_{n}\\}_{n\in\mathbb{N}}$ in $m$
variables such that the power of $p$ in the denominator of
$\operatorname{fpt}(f_{n})$ is less than or equal to $\ell$ for all
$n\in\mathbb{N}$, and
$\operatorname{fpt}(f_{n})=\operatorname{fpt}(f_{p^{\ell}})$ for all $n\geq
p^{\ell}$. There also exists a different family of diagonal hypersurfaces
$\\{g_{n}\\}_{n\in\mathbb{N}}$ such that the power of $p$ in the denominator
of $\operatorname{fpt}(g_{n})$ diverges to $\infty$ as $n\to\infty$.
###### Proof.
The case where $\ell=1$ is done by Theorems 4.1 and either 4.2 or 4.3. For the
case $\ell>1$, if the family
$f_{n}(x_{1},\dots,x_{m})=x_{1}^{a_{1}}+x_{2}^{a_{2}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{n}$
satisfies Theorem 4.1, then consider the family
$g_{n}(x_{1},\dots,x_{m})=x_{1}^{p^{\ell}a_{1}}+x_{2}^{p^{\ell}a_{2}}+\dots+x_{m-1}^{p^{\ell}a_{m-1}}+x_{m}^{p^{\ell}n}$.
Since the nonterminating base $p$ expansion of $\frac{1}{p^{\ell}a}$ for any
positive integer $a$ satisfies
$\left(\frac{1}{a}\right)^{(e)}=\left(\frac{1}{p^{\ell}a}\right)^{(e+\ell)}$
for all positive integers $e$, we may then proceed similarly to the proof of
Theorem 4.1. ∎
We end with some computational remarks. All computations are made using the
FrobeniusThresholds package of Macaulay2.
###### Remark 4.6.
It is possible for the power of $p$ in the denominator of
$\operatorname{fpt}(f_{n})$ for some $n$ between $p^{d}$ and $p^{d+1}$ to be
higher than the prescribed power in Theorem 4.3. For example, let
$a=2,b=3,p=97,d=2,n=9216$. Then the power of $p$ in the denominator of
$\operatorname{fpt}(x^{2}+y^{3}+z^{9216})$ is $16$, but the power of $p$ in
the denominator of $\operatorname{fpt}(x^{2}+y^{3}+z^{97^{2}-s})$ for $2\leq
s\leq 96$ is at most $11$.
###### Remark 4.7.
Theorem 4.1 may inspire the conjecture that the denominator of
$\operatorname{fpt}(x_{1}^{a_{1}}+x_{2}^{a_{2}}+\dots+x_{m-1}^{a_{m-1}}+x_{m}^{n})$
for $p\equiv-c\mod\text{lcm }(a,b)$ and $c\leq\frac{p-1}{2}$ is $p^{e}$ for
some positive integer $e\leq c$. Unfortunately, this is false. For example,
let $a=3,b=5,n=699,p=71$: note that $p\equiv-4\mod 15$, but
$\operatorname{fpt}(x^{a}+y^{b}+z^{n})=\frac{964836811}{71^{5}}$.
###### Remark 4.8.
The growth of the denominator in Theorems 4.2 and 4.3 may not hold if $p\equiv
c\mod\operatorname{lcm}(a,b)$ and $c\leq\frac{p-1}{2}$. For example, let
$a=3,b=5,n=500,p=17$: note that $p\equiv 2\mod 15$, and the nonterminating
base $p$ expansions of $\frac{1}{3}$ and $\frac{1}{5}$ are
$\frac{1}{3}=.5:11:\dots$ and $\frac{1}{5}=.3:6:\dots,$ giving
$\operatorname{fpt}(x^{3}+y^{5}+z^{500})=\left<\frac{1}{3}\right>_{1}+\left<\frac{1}{5}\right>_{1}+\left<\frac{1}{500}\right>_{1}+\frac{1}{17}=\frac{9}{17}.$
###### Remark 4.9.
(Necessity of $s^{r}<p-1$) For $a=3,b=4,n=165,p=13$: By Macaulay2’s
computations,
$\operatorname{fpt}(x^{a}+y^{b}+z^{n})=\frac{480786743}{13^{8}}$. Note that
$n=p^{2}-4$, and the nonterminating base $p$ expansions of $\frac{1}{3}$ and
$\frac{1}{4}$ are $\frac{1}{3}=.\overline{4}$ and $\frac{1}{4}=.\overline{3}$.
Now, for $r=1$, $4+3+s^{r}=4+3+4=11<p$. However, for $r=2$, $s^{r}=16>p-1$, so
the “predicted" power would be $5$, which is wrong.
## References
* [1] Bhargav Bhatt and Anurag Singh, The $F$-pure threshold of a Calabi-Yau hypersurface, Math. Ann. 362 (2015), no. 1-2, 551-567.
* [2] Eric Canton, Daniel Hernandez, Karl Schwede, and Emily Witt, On the behavior of singularities at the $F$-pure threshold, Illinois J. Math 60 (2016), no. 3-4, 669-685.
* [3] Daniel Hernandez, $F$-invariants of diagonal hypersurfaces, Proc. Am. Math. Soc. 143 (2014), no. 1, 87-104.
* [4] Daniel Hernandez, $F$-pure thresholds of binomial hypersurfaces, Proc. Am. Math. Soc. 142 (2014), no. 7, 2227-2242.
* [5] Daniel Hernandez, Luis Nuñez-Betancourt, Emily Witt, and Wen-Liang Zhang, $F$-pure thresholds of homogeneous polynomials, Michigan Math. J. 65 (2016), no. 1, 57-87.
* [6] Mircea Mustata, Shunsuke Takagi, and Kei-ichi Watanabe, $F$-thresholds and Bernstein-Sato polynomials, European Congress of Mathematics (Ari Laptev, ed.), European Mathematical Society, Stockholm, 2005, pp. 341-364.
* [7] Susanne Müller, The $F$-pure threshold of quasi-homogeneous polynomials, J. Pure Appl. Algebra 222 (2017), no. 1, 75-96.
* [8] Karl Schwede, Generalized test ideals, sharp $F$-purity, and sharp test elements, Math. Res. Lett. 15 (2008), no. 6, 1251-1261.
* [9] Karen Smith and Adela Vraciu, Values of the $F$-pure threshold for homogeneous polynomials, J. London Math. Soc. 103 (2023), no. 3, 1004-1035.
* [10] Shunsuke Takagi and Kei-ichi Watanabe, On $F$-pure thresholds, J. Algebra 282 (2004), no. 52, 278-297.
|
# A Puzzle of Proportions: Two Popular Bayesian Tests Can Yield Dramatically
Different Conclusions
Fabian Dablander Department of Psychological Methods, University of Amsterdam
Karoline Huth Department of Psychological Methods, University of Amsterdam
Quentin F. Gronau Department of Psychological Methods, University of
Amsterdam Alexander Etz University of California, Irvine Eric-Jan
Wagenmakers Department of Psychological Methods, University of Amsterdam
(blueThis manuscript has been accepted for publication in Statistics in
Medicine.)
###### Abstract
Testing the equality of two proportions is a common procedure in science,
especially in medicine and public health. In these domains it is crucial to be
able to quantify evidence for the absence of a treatment effect. Bayesian
hypothesis testing by means of the Bayes factor provides one avenue to do so,
requiring the specification of prior distributions for parameters. The most
popular analysis approach views the comparison of proportions from a
contingency table perspective, assigning prior distributions directly to the
two proportions. Another, less popular approach views the problem from a
logistic regression perspective, assigning prior distributions to logit-
transformed parameters. Reanalyzing 39 null results from the New England
Journal of Medicine with both approaches, we find that they can lead to
markedly different conclusions, especially when the observed proportions are
at the extremes (i.e., very low or very high). We explain these stark
differences and provide recommendations for researchers interested in testing
the equality of two proportions and users of Bayes factors more generally.
blackThe test that assigns prior distributions to logit-transformed parameters
creates prior dependence between the two proportions and yields weaker
evidence when the observations are at the extremes. When comparing two
proportions, we argue that this test should become the new default.
Keywords: Bayesian testing, Evidence of absence, Equality of proportions, Data
reanalysis
## 1 Introduction
Researchers frequently wish to test whether two populations differ. In
medicine and public health, for example, the resulting statistical analysis
frequently concerns testing whether or not two proportions differ. Examples
include testing whether a vaccine decreases the number of infections compared
to a control [[, e.g.,]]polack2020safety, whether sexual minorities are more
prone to suicide compared to their heterosexual counterparts [[,
e.g.,]]ploderl2013suicide, or whether tightly or less-tightly controlling
hypertension leads to fewer miscarriages in pregnant women [19].
In these applications it is crucial to be able to discriminate between
evidence of absence and absence of evidence. For example, [19] conducted a
trial to investigate the effect of a tight (target diastolic blood pressure,
85 mm Hg) or a less-tight (target diastolic blood pressure, 100 mm Hg) control
of hypertension in pregnant woman on, among other outcomes, pregnancy loss.
They found no significant difference between the two conditions, with 15 out
of 493 women in the less-tight control condition and 13 out of 488 in the
tight control condition having lost their child, yielding an estimated odds
ratio of 1.14 (95% CI: [0.53, 2.45]). How confident are we that there is
indeed no difference between the two conditions rather than the data being
inconclusive? Bayesian statistics provides a principled way of quantifying
evidence via the Bayes factor [11, 18, 20], thus providing one avenue to
discriminate between evidence of absence and absence of evidence [[,
e.g.,]]keysers2020using.
The Bayes factor quantifies how well one hypothesis predicts the data compared
to another. Using the test of equality between two proportions as an example,
let $\mathcal{D}=(y_{1},y_{2},n_{1},n_{2})$ denote the combined data from the
two groups. We have:
$\displaystyle Y_{1}$ $\displaystyle\sim\text{Binomial}(n_{1},\theta_{1})$
$\displaystyle Y_{2}$
$\displaystyle\sim\text{Binomial}(n_{2},\theta_{2})\enspace,$
where the sample sizes $(n_{1},n_{2})$ are assumed fixed and under
$\mathcal{H}_{0}$ we have that $\theta\equiv\theta_{1}=\theta_{2}$ while under
$\mathcal{H}_{1}$ we have that $\theta_{1}\neq\theta_{2}$. By quantifying
relative predictive performance, the Bayes factor tells us how we should
update our prior beliefs about $\mathcal{H}_{0}$ relative to $\mathcal{H}_{1}$
after observing the data [14]:
$\underbrace{\frac{p(\mathcal{H}_{0}\mid\mathcal{D})}{p(\mathcal{H}_{1}\mid\mathcal{D})}}_{\text{Posterior
odds}}=\underbrace{\frac{p(\mathcal{D}\mid\mathcal{H}_{0})}{p(\mathcal{D}\mid\mathcal{H}_{1})}}_{\text{Bayes
factor}}\,\times\underbrace{\frac{p(\mathcal{H}_{0})}{p(\mathcal{H}_{1})}}_{\text{Prior
odds}}\enspace.$
blackA Bayes factor of, say, 15 means that the data are 15 times more likely
under one hypothesis compared to the other. While there exist verbal
guidelines that may aid in the interpretation of the Bayes factor [[, for
example, Bayes factors in the range from 1$-$3 constitute _weak_ evidence,
those in the range from 3$-$10 constitute _moderate_ evidence, and values
larger than $10$ constitute _strong_ evidence;]]jeffreys1939theory,
wasserman2000bayesian, lee2013book, the Bayes factor should be understood as a
continuous measure of evidence [20].
While the Bayes factor does not depend on the prior probability of hypotheses,
it does depend crucially on the prior over parameters in the models
instantiating $\mathcal{H}_{0}$ and $\mathcal{H}_{1}$, which becomes apparent
when expanding:
$\frac{p(\mathcal{D}\mid\mathcal{H}_{0})}{p(\mathcal{D}\mid\mathcal{H}_{1})}=\frac{\int_{\theta}p(\mathcal{D}\mid\theta,\mathcal{H}_{0})\pi_{0}(\theta\mid\mathcal{H}_{0})\,\,\mathrm{d}\theta}{\int_{\theta_{1}}\int_{\theta_{2}}p(\mathcal{D}\mid\theta_{1},\theta_{2},\mathcal{H}_{1})\pi_{1}(\theta_{1},\theta_{2}\mid\mathcal{H}_{1})\,\,\mathrm{d}\theta_{1}\,\mathrm{d}\theta_{2}}\enspace,$
where $\pi_{0}$ and $\pi_{1}$ indicate the respective prior distributions.
There exist two main Bayes factor approaches for testing the equality of two
proportions. The more popular one comes from the analysis of contingency
tables, and assigns independent beta distributions directly to
$(\theta_{1},\theta_{2})$ [[, e.g.,]]gunel1974bayes, jamil2017default. We call
this approach the “Independent Beta” (IB) approach. The second approach is
less widely used, and assigns a prior to the average log odds $\beta$ and the
log odds ratio $\psi$ [15, 3]. We call this approach the “Logit
Transformation” (LT) approach. In this paper, we show that these two
approaches can yield markedly different results. This is especially the case
when the observed proportions are at the extremes (i.e., very low or very
high), as is the case for a large number of applications including the three
examples mentioned above. Consider the study by [19] again. The IB approach
yields a Bayes factor of 12.30 in favour of $\mathcal{H}_{0}$, while the LT
approach yields a mere 1.17. In other words, under the IB approach the data
are about 12 times more likely under the hypothesis that tightly or less-
tightly controlling hypertension have the same effect on miscarriages compared
to a hypothesis assuming a difference. Under the LT approach, however, the
data are about equally likely under both hypotheses, which constitutes
equivocal evidence. The answer to the question “is observing two equally small
proportions strong or weak evidence for the null hypothesis?” depends,
therefore, crucially — and non-trivially — on the prior setup.
This paper is structured as follows. In Section 2, we outline these two ways
of testing the equality of two proportions in more detail. In Section 3, we
highlight the occasionally stark differences of the two approaches by
reanalyzing 39 statistical tests reported in the New England Journal of
Medicine and explain why these differences occur. In Section 4, we end by
reviewing the implications of the prior setup and what users of Bayes factors
should be mindful of when testing the equality of two population parameters.
blackWe argue that the LT approach should become the default when testing the
equality of two proportions because it (a) induces prior dependence between
proportions which almost always are, in fact, dependent, and (b) yields a
sensibly milder assessment of the evidence compared to the IB approach when
the observations are at the extremes.
## 2 Two Ways of Testing the Equality of Two Proportions
In this section, we outline two ways of testing the equality of two
proportions. In Section 2.1 we describe the Independent Beta approach, and in
Section 2.2 we describe the Logit Transformation approach.
### 2.1 The Independent Beta (IB) Approach
In order to nest the null hypothesis under the alternative hypothesis, we
introduce the difference parameter $\eta=\theta_{2}-\theta_{1}$ and the grand
mean $\zeta=\frac{1}{2}\left(\theta_{1}+\theta_{2}\right)$. Using this
parameterization, we have that:
$\displaystyle\theta_{1}$ $\displaystyle=\zeta-\frac{\eta}{2}$
$\displaystyle\theta_{2}$ $\displaystyle=\zeta+\frac{\eta}{2}\enspace.$
The hypotheses are then specified as:
$\displaystyle\mathcal{H}_{0}$ $\displaystyle:\eta=0$
$\displaystyle\mathcal{H}_{1}$ $\displaystyle:\eta\neq 0\enspace.$
Under this approach, we assign independent $\text{Beta}(a,a)$ priors to
$\theta_{1}$ and $\theta_{2}$. Figure 1 visualizes the joint prior
distribution (top) under $\mathcal{H}_{1}$ for $a=1$ (left) and $a=2$ (right).
Increasing values of $a$ implies that the joint prior mass is more
concentrated around
$(\theta_{1},\theta_{2})=(\nicefrac{{1}}{{2}},\nicefrac{{1}}{{2}})$. The
bottom panels visualize the conditional prior distribution of $\theta_{2}$
given that we know that $\theta_{1}=0.10$. Knowing the value of $\theta_{1}$
does not change our prior about $\theta_{2}$, which follows from the
assumption of prior independence.
Figure 1: Top: Joint prior distribution assigned to $(\theta_{1},\theta_{2})$
under $a=1$ (left) and $a=2$ (right). Bottom: Conditional prior distribution
of $\theta_{2}$ given that $\theta_{1}=0.10$.
The IB Bayes factor is available in analytic form [[, see
e.g.,]]dickey1970weighted, gunel1974bayes, jeffreys1939theory. In the
literature on contingency tables, our setup corresponds to an independent
multinomial sampling scheme where the row (or column) sums of the contingency
table (here $n_{1}$ and $n_{2}$) are fixed; for an extension to other sampling
schemes and more than two groups, see [4] and [10]. [4] suggest $a=1$ as a
default value. Note that values $a<1$ lead to an undefined prior density for
$\theta=0$ and $\theta=1$, which implies model selection inconsistency in case
$\theta$ is indeed 1 or 0. Hence these values should be avoided on principle
grounds. Because the beta prior is conjugate for the binomial likelihood, the
posterior distributions of $\theta_{1}$ and $\theta_{2}$ are again
(independent) beta distributions.
### 2.2 The Logit Transformation (LT) Approach
The test proposed by [15] and implemented by [3] does not assign a prior
directly to $(\theta_{1},\theta_{2})$, but applies a logit transformation and
assigns priors to the transformed parameters ($\beta$, $\psi$). Specifically,
we write:
$\displaystyle\text{log}\left(\frac{\theta_{1}}{1-\theta_{1}}\right)$
$\displaystyle=\beta-\frac{\psi}{2}$
$\displaystyle\text{log}\left(\frac{\theta_{2}}{1-\theta_{2}}\right)$
$\displaystyle=\beta+\frac{\psi}{2}\enspace,$
where $\beta$ is a grand mean and $\psi$ is the difference in log odds (i.e.,
the log odds ratio):
$\displaystyle\beta$
$\displaystyle=\frac{1}{2}\left(\text{log}\left(\frac{\theta_{1}}{1-\theta_{1}}\right)+\text{log}\left(\frac{\theta_{2}}{1-\theta_{2}}\right)\right)$
$\displaystyle\psi$
$\displaystyle=\text{log}\left(\frac{\theta_{2}}{1-\theta_{2}}\right)-\text{log}\left(\frac{\theta_{1}}{1-\theta_{1}}\right)\enspace.$
While this is a more involved reparameterization than in the IB approach
above, another way to formulate this setup is by writing:
$\displaystyle\theta_{1}$
$\displaystyle=\frac{e^{\beta-\frac{\psi}{2}}}{1+e^{\beta-\frac{\psi}{2}}}$
$\displaystyle\theta_{2}$
$\displaystyle=\frac{e^{\beta+\frac{\psi}{2}}}{1+e^{\beta+\frac{\psi}{2}}}\enspace,$
which readers familiar with logistic regression may recognize. Using this
setup, we test the hypotheses:
$\displaystyle\mathcal{H}_{0}$ $\displaystyle:\psi=0$
$\displaystyle\mathcal{H}_{1}$ $\displaystyle:\psi\neq 0\enspace.$
Figure 2: Top: Joint prior assigned to $(\theta_{1},\theta_{2})$ for
$\sigma_{\psi}=1$ (left) and $\sigma_{\psi}=2$ (right). Bottom: Conditional
prior distribution of $\theta_{2}$ given that $\theta_{1}=0.10$. In both cases
we assume $\sigma_{\beta}=1$.
In contrast to above, we now assign priors to $\beta$ and $\psi$ rather than
to $\theta_{1}$ and $\theta_{2}$ directly. In particular, under both
hypotheses we assume $\beta\sim\mathcal{N}(0,\sigma_{\beta})$ with
$\sigma_{\beta}=1$. Under $\mathcal{H}_{1}$ we assume
$\psi\sim\mathcal{N}(0,\sigma_{\psi})$.111While this prior specification
assigns the log odds ratio $\psi$ a Gaussian distribution, the IB prior
specification results in a non-standard distribution on $\psi$. For more
details, see Appendix A. The top left panel in Figure 2 visualizes the implied
joint prior distribution on $(\theta_{1},\theta_{2})$ under $\mathcal{H}_{1}$
for $\sigma_{\psi}=1$. The prior mass is concentrated along the diagonal,
which indicates that $\theta_{1}$ and $\theta_{2}$ are dependent. The bottom
left panel illustrates this fact: if we know that $\theta_{1}=0.10$, then the
prior on $\theta_{2}$ shifts toward this value. Setting $\sigma_{\psi}=2$
removes the prior dependency, as the right column in Figure 2 shows. For
values of $\sigma_{\psi}>2$, $\theta_{1}$ and $\theta_{2}$ become anti-
correlated and hence observing a small value of $\theta_{1}$ results in a
prior that puts more mass on large values for $\theta_{2}$. Such an inverse
relation is undesirable in almost all empirical applications, and so values
$\sigma_{\psi}>2$ are therefore to be avoided. [3] developed software to
compute the Bayes factor using this prior specification, first proposed by
[15], suggesting $\sigma_{\psi}=1$ as a default value.
### 2.3 Comparison of Priors
A direct comparison of the two prior specifications may be helpful to get
further intuition for their differences. While the IB approach does not assign
a prior distribution to $(\beta,\psi)$ or $(\eta,\zeta)$ explicitly, assigning
a prior to $\theta_{1}$ and $\theta_{2}$ induces a prior distribution on these
quantities. Conversely, the LT approach assigns a prior to $\psi$ and $\beta$
and this induces a prior on $(\theta_{1},\theta_{2})$ and $(\eta,\zeta)$. The
induced prior distributions under both approaches are non-standard (see
Appendix A), but their densities can be calculated numerically.
Figure 3: Top: Prior distributions assigned to $\eta$ (left) and $\psi$
(right) under the LT (blue, vertical lines) and IB approach (red, horizontal
lines), for two different parameter settings, respectively. Bottom: Marginal
prior distribution of $\theta_{1}$ and $\theta_{2}$ under the two approaches.
blackThe density that is filled and has the highest peak corresponds to
$\sigma_{\psi}=1$.
The top left panel in Figure 3 shows the prior distribution assigned to $\eta$
by the LT approach for $\sigma_{\psi}=1$ (shaded blue) and $\sigma_{\psi}=2$
(striped blue) and the IB approach for $a=1$ (shaded red) and $a=2$ (striped
red) under $\mathcal{H}_{1}$. Similarly, the top right panel shows the prior
distribution assigned to $\psi$ for the two approaches and prior parameter
values. The (default) IB approach assigns comparatively more mass to large
values of $\eta$ and $\psi$, which in practice means that it expects larger
differences between the sample proportions. The bottom panel shows the
marginal priors for $\theta_{1}$ and $\theta_{2}$, where we find that the LT
approach assigns comparatively less mass to extreme values. The LT approach
cannot result in a uniform distribution on the proportions under
$\mathcal{H}_{0}$ because of the Gaussian prior on $\beta$. If it instead
would assign a (standard) logistic prior to $\beta$ (which has fatter tails),
the prior on the proportions would be uniform; see a related discussion in
Appendix B. In the next section, we discuss a somewhat surprising difference
between these two tests.
## 3 Practical Implications of the Prior Setup
To see the implications of the two different prior specifications in practice,
in Section 3.1 we reanalyze 39 null results published in the New England
Journal of Medicine, previously analyzed by [6] using the IB approach. We then
explain why these difference occur in Section 3.2. The data and code to
reproduce all analyses and figures are available from
https://github.com/fdabl/Proportion-Puzzle.
### 3.1 Reanalysis of New England Journal of Medicine Studies
[6] considered all 207 articles published in the New England Journal of
Medicine in 2015. The abstract of 45 of these articles contained a claim about
the absence or non-significance of an effect for a primary outcome measure,
and 37 of those allowed for a comparison of proportions, reporting 43 null
results in total. We focus on those results that can be reanalyzed using a
test between two proportions, which results in a total of 39 tests from 32
articles. The top left panel in Figure 4 contrasts Bayes factors in favour of
$\mathcal{H}_{0}$ computed using the IB approach (rectangles) across
$a\in[1,5]$ with Bayes factors computed using the LT approach (circles) across
$\sigma_{\psi}\in[1,2]$. In virtually all cases and across specifications, the
Bayes factor in favour of $\mathcal{H}_{0}$ is higher under the IB approach,
and this difference is frequently substantial.222In contrast, the conclusions
one would draw based on posterior distributions are very similar, see Appendix
C. As the parameter $a$ is increased under the IB approach, the expected
difference between the two groups is smaller (see top left panel in Figure 3).
Therefore, the predictions under $\mathcal{H}_{1}$ become more similar to the
predictions under $\mathcal{H}_{0}$, and the Bayes factor decreases.
Conversely, as $\sigma_{\psi}$ is increased under the LT approach, the
expected difference between the two group increases, and the Bayes factor in
favour of $\mathcal{H}_{0}$ increases.
Figure 4: Top: Bayes factors using the IB (rectangles) and LT approach
(circles) in favour of $\mathcal{H}_{0}$ across studies reported in [6] (left)
or for simulated equal proportions with $n=100$ (right) for values $a\in[1,5]$
and $\sigma_{\psi}\in[1,2]$ with $\sigma_{\beta}=1$. Bottom: Joint prior
distribution of $(\theta_{1},\theta_{2})$ under the IB approach with $a=1$
(left) and under the LT approach with $\sigma_{\psi}=1$ and $\sigma_{\beta}=1$
(right). Black dots and blue rectangles indicate the maximum likelihood
estimates of the proportions in the studies analyzed by [6].
The top left panel in Figure 4 shows that, for some studies, the IB and LT
Bayes factors cannot be brought into the vicinity of each other by changing
the prior parameters in the way specified above, while for other studies the
Bayes factors do overlap substantially. The studies without overlap are those
indexed as 1-12. They are shown as blue rectangles in the bottom panels in
Figure 4, which shows the joint prior density for the IB (left) and the LT
approach (right) with the symbols indicating the maximum likelihood estimates
for the individual studies. Note that the larger the proportions in the bottom
panels, the more likely are the two Bayes factors to overlap in the top left
panel.
While [6] are reassured by the fact that their Bayesian reanalysis (using the
IB approach) yields strong evidence in favour of $\mathcal{H}_{0}$ on average
(given that all studies reported a non-significant $p$-value), using the LT
approach yields a more uncertain picture. While the median Bayes factor across
the studies under the IB approach is 12.30, the median Bayes factor under the
LT approach is only 4.79. This difference is driven by the extremes, where the
two approaches suggest substantially different conclusions. We have already
seen one example in the introduction concerning hypertension and pregnancy
loss (analysis identifier 3). Another example is given by [13], which compared
the efficacy of the 9vHPV against the qHPV vaccine for preventing, among
others, cervical, vulvar, and varginal disease and persistent human
papillomavirus related infections in women. Comparing different outcome
measures in women who were HPV-negative and HPV-positive at baseline, the IB
approach yields Bayes factors consistently around 100 while the LT approach
never shows Bayes factors larger than 10 (see analysis identifiers 1, 2, 5,
and 6). The IB approach thus strongly suggests that the two vaccines have a
similar efficacy, while the LT approach suggests that more data is needed to
reach a firm conclusion.
The initially diverging and then converging pattern of the Bayes factors can
be seen neatly in the top right panel in Figure 4, which shows the Bayes
factors under the null hypothesis for increasing values of
$\nicefrac{{y}}{{100}}\equiv\nicefrac{{y_{1}}}{{100}}=\nicefrac{{y_{2}}}{{100}}$.
Note that the pattern is symmetric, so that for
$\nicefrac{{y}}{{100}}\in[\nicefrac{{1}}{{2}},1]$ the same holds, just
mirrored. At the extremes $\nicefrac{{y}}{{n}}\in\\{0,1\\}$, the IB approach
yields a Bayes factor in favour of $\mathcal{H}_{0}$ of 50.75, while the LT
Bayes factor gives a mere 1.40. The Bayes factor decreases as
$\nicefrac{{y}}{{n}}\rightarrow\nicefrac{{1}}{{2}}$ under the IB approach, but
increases under the LT approach. Consequently, the difference between the
Bayes factors becomes less pronounced as we move to more central proportions,
which matches the observation in the empirical analysis shown in the top left
panel in Figure 4. For $\nicefrac{{y}}{{n}}=\nicefrac{{1}}{{2}}$, the IB Bayes
factor yields 5.70 and the LT Bayes factor yields 3.67.
In summary, we have found that the two approaches give a very different answer
to the question “is observing two equally small proportions strong or weak
evidence in favour of the null hypothesis?”. The IB Bayes factor suggests that
the evidence is often orders of magnitude larger than the LT Bayes factor.
### 3.2 Explaining the Difference
Why does the IB Bayes factor yield so much stronger evidence than the LT Bayes
factor for $\mathcal{H}_{0}$ when the proportions are small and of roughly
equal size? And why does the IB Bayes factor decrease as the proportions get
closer to $\nicefrac{{1}}{{2}}$ while the LT Bayes factor increases? To answer
these questions, we need to zoom in on the differences in the respective prior
specifications. The two approaches differ in two key ways. First, while the IB
approach assigns independent priors to $\theta_{1}$ and $\theta_{2}$, the LT
approach assigns dependent priors to them. Second, the LT approach employs a
logit transformation while the IB approach does not. Appendix D shows that it
is not the prior dependence that underlies the difference between the two
approaches. Instead, as we will see below, it is the logit transformation.
The fact that the IB Bayes factor is larger than the LT Bayes factor even when
the latter approach expects a larger difference in the proportions (e.g., a
larger $\eta$ or $\psi$, compare $a=2$ with $\sigma_{\psi}=2$ in Figure 3)
means that focusing only on the difference parameter is not sufficient to
explain the difference between the Bayes factors. Instead, we turn to a
sequential predictive perspective. From such a perspective, we may first use
the prior to predict the data from group one, update the prior to a posterior,
and then predict the data from group two. One can rewrite the marginal
likelihoods to make this sequential perspective apparent, see Appendix E. This
perspective shows that there is a crucial difference in the predictions that
the CT and LT approach make under $\mathcal{H}_{1}$. Under both approaches
there is a common parameter $\theta$ under $\mathcal{H}_{0}$, and the prior
assigned to $\theta$ gets updated by data from the first group, after which
predictions about the second group are made. Under the alternative hypothesis,
however, the IB approach implies that observing data from group one does not
update our beliefs about likely values of $\theta_{2}$, and so data from group
one cannot inform the subsequent prediction about group two. Under the LT
approach, such information sharing does take place. This difference is shown
in the top panels in Figure 5, which visualizes the joint prior distribution
for $\theta_{1}$ and $\eta\equiv\theta_{2}-\theta_{1}$ under the IB (left) and
LT (right) approaches. We see that under the IB approach, learning about
$\theta_{1}$ does not influence our predictions about likely differences
between the two groups. In contrast, under the LT approach we find that when
we learn that $\theta_{1}$ is either small or large, we expect small
differences between the proportions compared to when we learn that
$\theta_{1}$ is about half. This explains why the LT Bayes factor in favour of
$\mathcal{H}_{0}$ increases as we move from extreme values of
$\nicefrac{{y}}{{n}}$ towards values around $\nicefrac{{1}}{{2}}$, as shown in
the top right panel in Figure 4: at the extremes, $\mathcal{H}_{1}$ expects
smaller differences between $\theta_{2}$ and $\theta_{1}$, and hence it more
closely resembles $\mathcal{H}_{0}$, leading to a more equivocal Bayes factor.
Moving towards more central values, $\mathcal{H}_{1}$ expects larger
differences and hence $\mathcal{H}_{0}$ outpredicts it by a larger margin.
Figure 5: Top: Joint prior distribution for $(\theta_{1},\eta)$ under the IB
approach with $a=1$ (left) and under the LT approach with $\sigma_{\psi}=1$
and $\sigma_{\beta}=1$ (right). Bottom: Joint prior distribution for
$(\theta_{1},\psi)$ under the two approaches.
While prior dependence is necessary to change one’s beliefs about likely
differences between the proportions upon learning $\theta_{1}$, the fact that
the expected difference decreases in the LT approach is due to the logit
transformation. When the two rates $(\theta_{1},\theta_{2})$ are small, say
$\theta_{1}=0.05$ and $\theta_{2}=0.10$, the log odds difference $\psi$ is
large, in this case $\psi=0.75$. The prior on $\psi$ in the LT approach
renders such large log odds differences to be unlikely, which means that a
smaller rate difference is expected. Conversely, if the rates are somewhere in
the center, say $\theta_{1}=0.50$ and $\theta_{2}=0.55$, then the log odds
difference $\psi$ is small, in this case $\psi=0.20$. The prior on $\psi$ in
the LT approach views such small log odds differences as likely; in fact, much
larger ones are possible. Therefore, the LT approach expects larger rate
differences when the rates are in the center. On the log odds scale, the
expected difference does not change markedly as $\theta_{1}$ varies, as the
bottom right panel in Figure 5 shows.333The fact that there is barely any
prior mass at extreme values of $\theta_{1}$ under the IB approach is due to
the Gaussian prior on $\beta$, as discussed in Section 2.3.
In contrast to the LT approach, the IB approach does not automatically reduce
the expectations about likely differences as we learn that $\theta_{1}$ takes
on extreme values. Taking slices at particular values for $\theta$ in the top
left panel in Figure 5 would show uniform distributions ranging from
$-\theta_{1}$ to $1-\theta_{1}$. Instead of reducing the size of an expected
difference at the extremes, the IB approach actually amplifies it. This occurs
because of the boundedness of the rate parameters. In particular, if
$\theta_{1}=0$, then the only values for which, say, $|\eta|<0.10$ are
$\theta_{2}\in[0,0.10]$. Differences in absolute magnitude between the two
rates of up to $0.10$ if $\theta_{1}=0$ are therefore assigned a probability
of $0.10$. If, however, $\theta_{1}=0.50$, the values of $\theta_{2}$ for
which $|\eta|<0.10$ are $\theta_{2}\in[0.40,0.60]$. This means that
differences in absolute magnitude between the two rates of up to $0.10$ if
$\theta_{1}=0.50$ are assigned a probability of $0.20$, or twice as much as in
the case that $\theta_{1}=0$. Another way to see this is to note that the
expected difference in the case that $\theta_{1}=0$ is $\eta=0.50$, while the
expected difference in the case that $\theta_{1}=0.50$ is $\eta=0$; more
generally, the expected difference increases linearly as we move to the
extremes, as can be seen in the top left panel in Figure 5. The bottom left
panel in Figure 5 shows an associated and pronounced increase in the expected
log odds difference as $\theta_{1}$ moves towards extreme values.
## 4 Discussion
Being able to discriminate between evidence of absence and absence of evidence
is key in medicine and public health. The Bayes factor, which pits the
predictive performance of two hypotheses against each other, is a method for
doing so. Reanalyzing 39 null results published in the New England Journal of
Medicine, we have found that the strength of evidence depends crucially on the
prior specification. Comparing an approach that assigns independent beta
distributions directly to the rate parameters (IB) to one that employs a logit
transformation (LT), we have found that the former approach suggests evidence
that is orders of magnitudes larger than the latter approach when observing
small proportions of roughly equal size. Consider again the effect of tightly
or less-tightly controlling hypertension in pregnant women [19]. The IB
approach suggests a Bayes factor about ten times larger in favour of the null
hypothesis than the LT approach. Similarly, the Bayes factor in favour of the
hypothesis that two vaccines against the human papillomavirus are equally
effective [13] is up to 38 times larger under the IB approach, depending on
the outcome measure. While the original statistical analysis by the authors is
more involved (e.g., adjusting for various covariates), these differences are
striking. As we have seen, they occur because of the boundedness of the
parameter space, which leads to the expectations of large differences at the
extremes under the IB approach. In contrast, the LT approach expects smaller
differences at the extremes.
Although the two Bayes factors frequently differed markedly in size, they
always both provided evidence in favour of $\mathcal{H}_{0}$ in the cases we
studied in this paper. This need not be the case, however. For example, a
large-scale investigation into the effect of aspirin found that 26 out of
$n_{1}=11,034$ who received a placebo suffered a fatal heart attack, but only
10 out of $n_{2}=11,037$ who received aspirin did [23]. The IB Bayes factor
yields strong evidence in favour of $\mathcal{H}_{0}$ ($\text{BF}_{01}=19.78$)
while the LT Bayes factor finds moderate evidence in favour of
$\mathcal{H}_{1}$ ($\text{BF}_{10}=5.36$). This is again owed to the fact that
the IB approach expects larger differences than the LR approach; a small
effect size is more likely under $\mathcal{H}_{0}$ than under
$\mathcal{H}_{1}$ in the IB approach simply because under $\mathcal{H}_{1}$
unrealistically large effect sizes are assumed to be plausible.
Our results have a number of implications. First, researchers should think
carefully about their prior setup. This not only implies thinking about the
parameter values for the priors, but also the nature of the priors and what
they imply for the data. As we have seen, one may be fooled by assessing how
sensitive the Bayes factor is by varying only prior parameters conditional on
the model. For all studies we re-analyzed that had proportions at the
extremes, varying the prior parameters of the IB approach still resulted in
substantial evidence in favour of the null hypothesis. However, changing the
prior setup to the LT approach, we found markedly reduced evidence. If there
are sensible alternative parameterizations, it may be prudent to explore how
sensitive the results are to these. In particular, our exposition suggests
that assessing the predictions that follow from the model can help get a
better intuition of its assumptions. In our case, this made clear that the IB
approach expects larger rate differences at the boundaries of the parameter
space compared to the center, while the LT approach expects the reverse. We
were struck by the initially puzzling contrast between these two tests, given
that the task — comparing two proportions — seemed so simple. This suggests
that even more caution needs to be applied when using Bayes factors for
comparing models that are invariably more complex. In the same spirit of
sensitivity analysis, it is prudent to compare inferences based on the Bayes
factor with inferences based on the posterior distributions of the parameters.
While testing precise hypotheses based on credible intervals is arguably
inappropriate since it makes inference conditional on $\mathcal{H}_{1}$,
ignoring $\mathcal{H}_{0}$ [[, e.g.,]]berger2006bayes,
wagenmakers2020principle, alarm bells should go off when they show stark
differences to the Bayes factor results, as in our case (compare Figure 4 and
6). blackResearchers who rely only on the posterior distribution for inference
or for trial design are unaffected by our results.
The second implication of our work is that the LT approach appears better
suited for testing the equality of two proportions than the currently more
popular IB approach. First, assuming prior dependence strikes us as a more
sensible approach especially for medicine and public health. This is because
it is generally unlikely that gaining information about one group does not
influence our beliefs about the other group. For example, if a particular
treatment for cancer yields a $30\%$ success rate, surely this should inform
our expectations for the success rate of a comparison treatment. Incorporating
this dependency results in more cautious predictions, which reduces the
eagerness with which the standard IB approach suggests strong evidence in
favour of $\mathcal{H}_{0}$ (see also Appendix D). While we have focused on
testing the sharp null hypothesis $\theta_{1}=\theta_{2}$, the difference
between dependent and independent priors is important also when testing other
hypotheses. [9] discusses the difference between a dependent and independent
prior in the context of comparing $\theta_{1}<\theta_{2}$ against
$\theta_{1}>\theta_{2}$, and similarly recommends dependent priors, as
independent priors lead to a test that is not “sufficiently cautious”.
Most importantly, however, is the fact that the LT approach avoids the issues
associated with a bounded parameter space. In particular, the IB approach is
not sensitive to the fact that the implications of a rate difference
$\theta_{2}-\theta_{1}$ depends on the actual values of the rates. For
example, $\theta_{2}=0.10$ and $\theta_{1}=0.05$ and $\theta_{2}=0.55$ and
$\theta_{1}=0.50$ imply the same rate difference, but while the former
represents a 100% increase, the latter represents only a 10% increase. Yet the
IB approach assigns more prior mass to large rate differences at the extremes
of the parameter space compared to its center. In contrast, the reverse holds
for the LT approach because of the logit transformation. This strikes us as
more sensible. With the recent development of accessible software that makes
the LT approach easily available [3, 8], we hope that the use of the LT
approach for testing the equality of two proportions becomes more widespread.
To sum up, we have seen an initially puzzling divergence in evidence when
reanalyzing 39 null results from the New England Journal of New Medicine with
two different Bayes factor tests. We have explained this divergence and sided
with the approach that employs a logit transformation and assumes prior
dependence between the rates. We suggest that this approach, rather than the
one motivated from a contingency table perspective, should be routinely
applied when testing the equality of two proportions. Our journey should also
act as a cautionary tale for users of Bayes factors, stressing the importance
of assessing the predictions of one’s models, conducting thorough sensitivity
analyses, and never relying on just a single quantity. We have demonstrated
the usefulness of this in the simple case of testing the equality of two
proportions. Most applications will arguably be more complex, however,
increasing the potential for puzzles and stressing the need for a holistic
evidence assessment.
#### Author contributions.
EJW proposed the study. KH, QFG, AE, and EJW worked on the problem initially.
After an interim period, FD and EJW resumed the work and solved the puzzle. FD
and KH analyzed the data. FD wrote the manuscript. All authors read, provided
comments, and approved the submitted version of the paper. They also declare
that there were no conflicts of interest.
#### Acknowledgements.
We thank Rink Hoekstra, Rei Monden, and Don van Ravenzwaaij for sharing the
New England Journal of Medicine data with us.
## References
* [1] J.. Berger “Bayes Factors” In _Encyclopedia of Statistical Sciences, vol. 1_ Hoboken, NJ: Wiley, 2006, pp. 378–386
* [2] James M Dickey and BP Lientz “The weighted likelihood ratio, sharp hypotheses about chances, the order of a Markov chain” In _The Annals of Mathematical Statistics_ JSTOR, 1970, pp. 214–226
* [3] Quentin F Gronau, A Raj and Eric-Jan Wagenmakers “Informed Bayesian inference for the A/B test” In _Journal of Statistical Software_ , in press
* [4] Erdogan Gunel and James Dickey “Bayes factors for independence in contingency tables” In _Biometrika_ 61.3 Oxford University Press, 1974, pp. 545–557
* [5] Max Hinne, Quentin F Gronau, Don van den Bergh and Eric-Jan Wagenmakers “A conceptual introduction to Bayesian model averaging” In _Advances in Methods and Practices in Psychological Science_ 3.2 SAGE Publications Sage CA: Los Angeles, CA, 2020, pp. 200–215
* [6] Rink Hoekstra, Rei Monden, Don van Ravenzwaaij and Eric-Jan Wagenmakers “Bayesian reanalysis of null results reported in medicine: Strong yet variable evidence for the absence of treatment effects” In _PloS One_ 13.4 Public Library of Science San Francisco, CA USA, 2018, pp. e0195474
* [7] Jennifer A Hoeting, David Madigan, Adrian E Raftery and Chris T Volinsky “Bayesian model averaging: A tutorial” In _Statistical Science_ JSTOR, 1999, pp. 382–401
* [8] Tabea Hoffmann and Eric-Jan Wagenmakers “Bayesian Inference for the A/B Test: Example Applications with R and JASP” PsyArXiv, 2021 DOI: 10.31234/osf.io/z64th
* [9] JV Howard “The 2 $\times$ 2 table: A discussion from a Bayesian viewpoint” In _Statistical Science_ JSTOR, 1998, pp. 351–367
* [10] Tahira Jamil et al. “Default “Gunel and Dickey” Bayes factors for contingency tables” In _Behavior Research Methods_ 49.2 Springer, 2017, pp. 638–652
* [11] Harold Jeffreys “Some tests of significance, treated by the theory of probability” In _Proceedings of the Cambridge Philosophical Society_ 31.2, 1935, pp. 203–222
* [12] Harold Jeffreys “Theory of Probability” Oxford, UK: Oxford University Press, 1939
* [13] Elmar A Joura et al. “A 9-valent HPV vaccine against infection and intraepithelial neoplasia in women” In _New England Journal of Medicine_ 372.8 Mass Medical Soc, 2015, pp. 711–723
* [14] Robert E Kass and Adrian E Raftery “Bayes factors” In _Journal of the American Statistical Association_ 90.430 Taylor & Francis, 1995, pp. 773–795
* [15] Robert E Kass and Suresh K Vaidyanathan “Approximate Bayes factors and orthogonal parameters, with application to testing equality of two binomial proportions” In _Journal of the Royal Statistical Society: Series B (Methodological)_ 54.1 Wiley Online Library, 1992, pp. 129–144
* [16] Christian Keysers, Valeria Gazzola and Eric-Jan Wagenmakers “Using Bayes factor hypothesis testing in neuroscience to establish evidence of absence” In _Nature Neuroscience_ 23.7 Nature Publishing Group, 2020, pp. 788–799
* [17] M.. Lee and E.–J. Wagenmakers “Bayesian Cognitive Modeling: A Practical Course” Cambridge University Press, 2013
* [18] Alexander Ly, Josine Verhagen and Eric-Jan Wagenmakers “Harold Jeffreys’s default Bayes factor hypothesis tests: Explanation, extension, and application in psychology” In _Journal of Mathematical Psychology_ 72 Elsevier, 2016, pp. 19–32
* [19] Laura A Magee et al. “Less-tight versus tight control of hypertension in pregnancy” In _New England Journal of Medicine_ 372.5 Mass Medical Soc, 2015, pp. 407–417
* [20] Richard D Morey, Jan-Willem Romeijn and Jeffrey N Rouder “The philosophy of Bayes factors and the quantification of statistical evidence” In _Journal of Mathematical Psychology_ 72 Elsevier, 2016, pp. 6–18
* [21] Matthew Oladejo Ojo “A remark on the convolution of the generalized logistic random variables” In _ASSET serves A_ 1.2, 2003
* [22] Thu Pham-Gia, Noyan Turkkan and P Eng “Bayesian analysis of the difference of two proportions” In _Communications in Statistics-Theory and Methods_ 22.6 Taylor & Francis, 1993, pp. 1755–1771
* [23] Steering Committee of the Physicians’ Health Study Research Group “Final report on the aspirin component of the ongoing Physicians’ Health Study” In _New England Journal of Medicine_ 321.3 Mass Medical Soc, 1989, pp. 129–135
* [24] Martin Plöderl et al. “Suicide risk and sexual orientation: a critical review” In _Archives of Sexual Behavior_ 42.5 Springer, 2013, pp. 715–727
* [25] Fernando P Polack et al. “Safety and efficacy of the BNT162b2 mRNA Covid-19 vaccine” In _New England Journal of Medicine_ 383.27 Mass Medical Soc, 2020, pp. 2603–2615
* [26] E.–J. Wagenmakers, M.. Lee, J.. Rouder and R.. Morey “The Principle of Predictive Irrelevance or Why Intervals Should not be Used for Model Comparison Featuring a Point Null Hypothesis” In _The Theory of Statistics in Psychology – Applications, Use and Misunderstandings_ Cham: Springer, 2020, pp. 111–129
* [27] Larry Wasserman “Bayesian model selection and model averaging” In _Journal of Mathematical Psychology_ 44.1 Elsevier, 2000, pp. 92–107
## Appendix A Induced Priors
Under the LT approach, the marginal priors for $\theta_{1}$ and $\theta_{2}$
as well as $\eta$ are given in integral representation by [3]. Under the IB
approach, the density of $\eta$ is given by:
$f(\eta;a)=\begin{cases}\frac{1}{B(a,a)}\eta^{2a-1}(1-\eta)^{2a-1}\,F_{1}\left({a},{4a-2},{1-a};{2a};{1-\eta},{1-\eta^{2}}\right)&0<\eta\leq
1\\\\[4.5pt]
\frac{1}{B(a,a)}(-\eta)^{2a-1}(1+\eta)^{2a-1}\,F_{1}\left({a},{1-a},{4a-2};{2a};{1-\eta^{2}},{1+\eta}\right)&-1\leq\eta<0\end{cases}\enspace,$
see [22]. In our case, we always have that $2a>1$ and so
$f(0)=\frac{B(2a-1,2a-1)}{B(a,a)^{2}}$ [22]. We can derive the density for
$\psi$ under prior independence for the special case of $a=1$. Note that,
because $\theta\sim\text{Beta}(1,1)$, it follows that:
$\text{log}\left(\frac{\theta}{1-\theta}\right)\sim\text{Logistic}(0,1)\enspace.$
Recall that:
$\psi=\text{log}\left(\frac{\theta_{2}}{1-\theta_{2}}\right)-\text{log}\left(\frac{\theta_{1}}{1-\theta_{1}}\right)\enspace.$
[21] gives the distribution function of the sum of $n$ i.i.d. logistic random
variables. Since the logistic distribution is symmetric, we have that if
$X\sim\text{Logistic}$ then $-X\sim\text{Logistic}$. Hence we can use the
results by [21] and write the density function of $\psi$ as:
$f(\psi;a=1)=\frac{e^{\psi}(e^{\psi}(\psi-2)+\psi+2)}{(e^{\psi}-1)^{3}}\enspace.$
## Appendix B Combining Approaches through Model-averaging
If one is uncertain as to which test one should employ, an approach would be
to combine both through model-averaging [[, e.g.,]]hoeting1999bayesian,
hinne2020conceptual. Let $\mathcal{M}_{0}^{\text{IB}}$ and
$\mathcal{M}_{0}^{\text{LT}}$ denote the null models under the IB and the LT
approach, respectively, and let $\mathcal{M}_{1}^{\text{IB}}$ and
$\mathcal{M}_{1}^{\text{LT}}$ denote the respective alternative models. The
model-averaged Bayes factor is given by:
$\text{BF}^{\text{AVG}}_{01}=\frac{p(\mathcal{D}\mid\mathcal{M}_{0}^{\text{IB}})\pi(\mathcal{M}_{0}^{\text{IB}})+p(\mathcal{D}\mid\mathcal{M}_{0}^{\text{LT}})\pi(\mathcal{M}_{0}^{\text{LT}})}{p(\mathcal{D}\mid\mathcal{M}_{1}^{\text{IB}})\pi(\mathcal{M}_{1}^{\text{IB}})+p(\mathcal{D}\mid\mathcal{M}_{1}^{\text{LT}})\pi(\mathcal{M}_{1}^{\text{LT}})}\enspace,$
where $\pi()$ gives the prior probability of the particular model. One has to
be careful when model-averaging. While the nuisance parameters generally do
not matter when testing nested models [15, 18], they do here. In particular,
the IB approach assigns $\beta$ a logistic distribution, while the LT approach
assigns $\beta$ a Gaussian distribution. The logistic distribution has fatter
tails, and thus puts more mass on extreme values of $\theta$ than the Gaussian
distribution. In our case, this can result in somewhat surprising results. Let
$\mathcal{D}_{1}=(y_{1},n_{1})$ and $\mathcal{D}_{2}=(y_{2},n_{2})$ denote
data in the two groups, respectively. For data $\mathcal{D}_{1}=(0,50)$ and
$\mathcal{D}_{2}=(0,50)$, for example, the IB model that assumes a difference
in the population ($\mathcal{M}^{\text{IB}}_{1}$) outpredicts the LT null
model ($\mathcal{M}^{\text{LT}}_{0}$) by a factor of seven! Since we want to
focus on differences under $\mathcal{H}_{1}$ between the IB and LT approach,
we would need to make the models identical under $\mathcal{H}_{0}$, for
example by assigning $\beta$ a logistic prior not only under the IB approach,
but also under the LT approach.
## Appendix C Inference based on Posterior Distributions
The contrasting results discussed in Section 3 do not occur when inferences
are based on the posterior distribution of the log odds difference $\psi$ or
the rate difference $\eta$. Figure 6 shows the posterior mean and 95% credible
interval of $\psi$ (top) and $\eta$ (bottom) for the IB (rectangles) and the
LT approach (circles) with the respective default parameterizations $a=1$ and
$\sigma_{\psi}=1$ across all studies. The conclusions one would draw from
these posterior distributions are very similar (and changing the prior
parameters does not affect them). This is as expected, since the effect of the
prior generally washes out as more and more data are observed, which is not
the case for Bayes factors.
Figure 6: Shows the posterior mean and 95% credible interval of the log odds
difference $\psi$ (top) and the rate difference $\eta$ (bottom) under the IB
approach (rectangles, $a=1$) and the LT approach (circles, $\sigma_{\psi}=1$).
## Appendix D A Dependent IB Approach
If prior dependence were responsible for the stark differences between the
Bayes factors at the extremes, we would find that a IB approach which assumes
prior dependence shows a pattern similar to the LT approach. Here, we show
that this is not the case. Recall that $\eta=\theta_{2}-\theta_{1}$ and
$\zeta=\frac{1}{2}\left(\theta_{1}+\theta_{2}\right)$. We define:
$\displaystyle\theta_{1}$
$\displaystyle=\text{min}\left(\text{max}\left(\zeta-\frac{\eta}{2},0\right),1\right)$
$\displaystyle\theta_{2}$
$\displaystyle=\text{min}\left(\text{max}\left(\zeta+\frac{\eta}{2},0\right),1\right)\enspace,$
and assign truncated Gaussian priors to $(\eta,\zeta$), i.e.,
$\eta\sim\mathcal{N}(0,\sigma_{\eta})_{\mathbb{I}(-1,1)}$ and
$\zeta\sim\mathcal{N}(0,\sigma_{\zeta})_{\mathbb{I}(0,1)}$. Using
$\sigma_{\eta}=\nicefrac{{1}}{{5}}$ and $\sigma_{\zeta}=\nicefrac{{1}}{{2}}$,
the left panel in Figure 7 shows the joint prior distribution for
$(\theta_{1},\theta_{2})$ from which a strong prior dependence between
$\theta_{1}$ and $\theta_{2}$ is apparent. The right panel, however, shows the
characteristic pattern of the IB approach as discussed in the main text,
rather than the characteristic pattern of the LT approach. Since the rates a
dependent a priori, it is not the prior dependence that is responsible for the
pattern. However, employing a prior that has a stronger dependency between the
two rates naturally reduces the size of their expected difference. Increasing
$\sigma_{\eta}$ results in a lower correlation (going from about 0.77 at
$\sigma_{\eta}=\nicefrac{{1}}{{5}}$ to zero at $\sigma_{\eta}=1$) and a
reduced Bayes factor in favour of $\mathcal{H}_{0}$, as the right panel in
Figure 7 shows.
Figure 7: Left: Joint distribution on $(\theta_{1},\theta_{2})$ using
$\sigma_{\eta}=\nicefrac{{1}}{{5}}$ and $\sigma_{\zeta}=\nicefrac{{1}}{{2}}$.
Right: Bayes factors in favour of $\mathcal{H}_{0}$ for simulated proportions
with $n=100$ for values $\sigma_{\eta}=[\nicefrac{{1}}{{5}},1]$ and
$\sigma_{\zeta}=\nicefrac{{1}}{{2}}$.
## Appendix E Sequential Predictive Perspective
Under $\mathcal{H}_{0}$ we have for both the IB (i.e., independent) and the LT
(i.e., dependent) setup that (suppressing conditioning on $\mathcal{H}_{0}$):
$\displaystyle p(\mathcal{D}_{2},\mathcal{D}_{1}\mid\mathcal{H}_{0})$
$\displaystyle=\int_{\theta}p(\mathcal{D}_{2},\mathcal{D}_{1}\mid\theta)\pi_{0}(\theta)\,\,\mathrm{d}\theta$
(1)
$\displaystyle=\int_{\theta}p(\mathcal{D}_{2}\mid\mathcal{D}_{1},\theta)p(\mathcal{D}_{1}\mid\theta)\pi_{0}(\theta)\,\,\mathrm{d}\theta$
(2)
$\displaystyle=\mathcal{Z}\int_{\theta}p(\mathcal{D}_{2}\mid\theta)p(\theta\mid\mathcal{D}_{1})\,\,\mathrm{d}\theta\enspace,$
(3)
where
$\mathcal{Z}=\int_{\theta}p(\mathcal{D}_{1}\mid\theta)\pi_{0}(\theta)\,\,\mathrm{d}\theta$
is the marginal likelihood for $\mathcal{D}_{1}$,
$p(\theta\mid\mathcal{D}_{1})$ is the posterior of $\theta$ after observing
$\mathcal{D}_{1}$, and we can remove the conditioning on $\mathcal{D}_{1}$ in
(2) because all relevant information is in $\theta$. Similarly, under
$\mathcal{H}_{1}$ for LT setup we have that:
$\displaystyle p(\mathcal{D}_{2},\mathcal{D}_{1}\mid\mathcal{H}_{1})$
$\displaystyle=\int_{\theta_{2}}\int_{\theta_{1}}p(\mathcal{D}_{2},\mathcal{D}_{1}\mid\theta_{1},\theta_{2})\pi_{1}(\theta_{1},\theta_{2})\,\mathrm{d}\theta_{1}\,\mathrm{d}\theta_{2}$
(4)
$\displaystyle=\int_{\theta_{2}}\int_{\theta_{1}}p(\mathcal{D}_{2}\mid\mathcal{D}_{1},\theta_{1},\theta_{2})p(\mathcal{D}_{1}\mid\theta_{1},\theta_{2})\pi_{1}(\theta_{1},\theta_{2})\,\mathrm{d}\theta_{1}\,\mathrm{d}\theta_{2}$
(5)
$\displaystyle=\mathcal{Z}\int_{\theta_{2}}\int_{\theta_{1}}p(\mathcal{D}_{2}\mid\theta_{1},\theta_{2})p(\theta_{1},\theta_{2}\mid\mathcal{D}_{1})\,\mathrm{d}\theta_{1}\,\mathrm{d}\theta_{2}\enspace,$
(6)
where $\mathcal{Z}$ is the marginal likelihood for $\mathcal{D}_{1}$,
$p(\theta_{1},\theta_{2}\mid\mathcal{D}_{1})$ is the posterior after observing
$\mathcal{D}_{1}$, and we can remove the conditioning on $\mathcal{D}_{1}$ in
(2) because all relevant information is in $(\theta_{1},\theta_{2})$. This is
in contrast to the IB setup, where:
$\displaystyle p(\mathcal{D}_{2},\mathcal{D}_{1}\mid\mathcal{H}_{1})$
$\displaystyle=\int_{\theta_{2}}\int_{\theta_{1}}p(\mathcal{D}_{2},\mathcal{D}_{1}\mid\theta_{1},\theta_{2})\pi_{1}(\theta_{1},\theta_{2})\,\mathrm{d}\theta_{1}\,\mathrm{d}\theta_{2}$
(7)
$\displaystyle=\int_{\theta_{2}}p(\mathcal{D}_{2}\mid\theta_{2})\pi_{1}(\theta_{2})\,\mathrm{d}\theta_{2}\int_{\theta_{1}}p(\mathcal{D}_{2}\mid\theta_{1})\pi_{1}(\theta_{1})\,\mathrm{d}\theta_{1}\enspace,$
(8)
and thus no sharing of information across the two groups takes place.
|
# High-dimensional entanglement for quantum communication in the frequency
domain
Meritxell Cabrejo-Ponce<EMAIL_ADDRESS>Fraunhofer
Institute for Applied Optics and Precision Engineering, Albert-Einstein-
Strasse 7, 07745 Jena Abbe Center of PhotonicsFriedrich Schiller University
JenaAlbert-Einstein-Str. 6, 07745 Jena, Germany André Luiz Marques Muniz
Fraunhofer Institute for Applied Optics and Precision Engineering, Albert-
Einstein-Strasse 7, 07745 Jena Marcus Huber<EMAIL_ADDRESS>Atominstitut, Technische Universität Wien, Stadionallee 2, 1020 Vienna,
Austria Institute for Quantum Optics and Quantum Information (IQOQI),
Austrian Academy of Sciences, 1090 Vienna, Austria Fabian Steinlechner
<EMAIL_ADDRESS>Fraunhofer Institute for Applied Optics
and Precision Engineering, Albert-Einstein-Strasse 7, 07745 Jena Abbe Center
of Photonics, Friedrich Schiller University Jena, Albert-Einstein-Str. 6,
07745 Jena, Germany
###### Abstract
High-dimensional photonic entanglement is a promising candidate for error-
protected quantum information processing with improved capacity. Encoding
high-dimensional qudits in the carrier frequency of photons combines ease of
generation, universal single-photon gates, and compatibility with fiber
transmission for high-capacity quantum communication. Recent landmark
experiments have impressively demonstrated quantum interference of a few
frequency modes, yet the certification of massive-dimensional frequency
entanglement has remained an open challenge. This study shows how to harness
the large frequency-entanglement inherent in standard continuous-wave
spontaneous parametric down-conversion processes. It further reports a record
certification of discretized frequency entanglement, combined with a novel
approach for certification that is both highly efficient and nonlocally
implementable. This technique requires very few measurements and does not
require assumptions on the state. The work opens the possibility for utilizing
this encoding in quantum communications and in quantum information science in
general.
## I Introduction
Entanglement is a unique and powerful quantum feature with a multitude of
applications in quantum information processing. The nonlocal correlations of
the entangled states may be used in quantum communications, imaging,
metrology, and quantum processors. In the case of photons, polarization-
entangled states have been traditionally used to demonstrate a multitude of
quantum gates and quantum information protocols [1] and the principles of
rapidly developing quantum networks [2, 3]. These qubit states are easy to
manipulate with linear optics and can be distributed in fiber or free-space
links because of their low interaction with the environment. Nevertheless,
larger alphabets in quantum communication are highly pursued not only to
increase the capacity of the quantum channel. High-dimensional encoding also
provides a stronger tolerance to noise [4], an essential asset to overcome the
transmission limits of polarization qubits. On the other hand, an increase in
dimensionality can also boost the computational power of quantum computers
[5]. In this context, a recent landmark experiment [6] used high-dimensional
entanglement of propagation paths on a silicon photonic chip to realise error-
protected logical qubits and thus improve the performance of a quantum phase
estimation algorithm. Harnessing the full state space of photonic degrees of
freedom (DOF) such as transverse spatial mode, time, and frequency, will be
key to future generations of photonic quantum information processors.
Figure 1: Schematic of our frequency-entangled photon pair source and the
state analysis parts. We generate broadband frequency entanglement at telecom
wavelength via spontaneous parametric down-conversion (SPDC) in a periodically
poled lithium niobate waveguide (ppLN). Afterwards, the photon pairs are
coherently manipulated with a pulse shaper and an electro-optic modulator
(EOM) to analyze the correlations in a superposition basis. CW: continuous-
wave; DEMUX: demultiplexer.
Although spatial [7, 8, 9] and temporal modes [10, 11, 12] have been
straightforward to operate in very large dimensional spaces, the frequency DOF
has remained behind in these advances. The frequency domain at optical scales
is of particular interest because of its parallelization capabilities, in
particular its compatibility with telecom multiplexing or frequency modulation
techniques. Earlier experiments exploiting electro-optic modulation were able
to demonstrate two-dimensional frequency entanglement [13], while the use of
pulse shapers and up-conversion processes allowed the characterization of up
to four-dimensional states [14, 15]. The combination of both components was
used later to coherently control the quantum states emerging from integrated
resonators [16]. These important building blocks opened the possibility to
measure in the superposition basis at optical scales and were used to
demonstrate discretized frequency entanglement with few dimensions (up to 6
[17]). Since then, approaches have been developed to perform arbitrary
manipulations of single-frequency qubits [18] and the first steps have been
taken towards full control of single qudits, its high-dimensional version [19,
20, 21]. While Hong-Ou-Mandel interference has also been used to verify two-
dimensional frequency entanglement [22], the requirement of local measurements
limits the utility of these methods in quantum networks.
Furthermore, it was recently shown that the HOM based approach cannot be used
to verify higher dimensional entanglement [23] , at least not without imposing
additional assumptions such as energy conservation [24].
Entangled frequency states can also be generated in combination with
entanglement in other DOFs like polarization, path [25] or even temporal,
whenever the relevant time and frequency properties can be manipulated
independently, because they refer to vastly different time scales. Such
hyperentangled states can also be used to enlarge the dimensionality of the
system, to generate cluster states [26] or to perform more advanced quantum
gates [27]. Yet, all these approaches are only able to manipulate a small
frequency mode set of typically extensive underlying quantum states.
Certifying real high-dimensional entanglement is not trivial, particularly in
the frequency domain, and its immense potential is still unexploited.
In this Letter, we show that it is not always necessary to design quantum
sources with a discretized frequency space, such as those built in cavities,
and therefore continuous spectra can also provide access to massive-
dimensional and well-controlled Hilbert space. We show quantum interference
with up to 7 modes and $>98\%$ visibility and demonstrate complete state
control in a 7-mode state space. Subsequently, based on previous work to
characterize time-bin qudits [28, 29], we certify genuine high-dimensional
frequency entanglement without prior assumptions regarding, e.g., the purity
of the quantum state. We can further alleviate the harsh requirements on the
number of measurements necessary to characterize the relevant correlations,
which grows with dimensionality, by taking into account energy-conservation.
This step is not necessary for our certification method, but it allows us
exploit a novel bucket detection approach that requires very few measurement
settings, in a similar fashion to the compressed measurements used to
characterize spatial correlations [30, 31]. Finally, by recovering the
information from high-quality quantum interference of two-dimensional (2D)
subspaces, we are able to certify a minimum of 33 entangled frequency modes,
the highest dimensionality of entanglement reported in time and frequency
degrees of freedom.
## II Results and Discussion
### II.1 Frequency comb generation
In our work, we analyze the frequency content of two-photon states generated
in standard $\chi^{(2)}$ non-linear crystals: periodically poled lithium
niobate (ppLN) waveguides (see Experimental Section). They provide continuous
and broadband spontaneous parametric down-converted photons (SPDC) with high
efficiency and 60 nm of bandwidth. The SPDC process is temperature tuned to a
degeneracy wavelength of 1548 nm to cover the entire C band with high
uniformity (see inset in Fig. 1).
At this point, the frequency space of the generated photon pairs would
typically be discretized, e.g., with etalon cavities to carve the spectrum
[32, 33]. This approach is useful for the isolation of frequency modes and for
performing sideband modulation. However, a considerable contribution of the
photon spectrum is directly rejected, reducing the total throughput. For this
reason, we avoid this step and employ the maximum bandwidth per frequency
mode.
Due to the energy conservation of the SPDC process, the emitted photon pairs
are strongly anticorrelated in frequency, and the state can be described as:
$|j\rangle_{s}|j\rangle_{i}=\int\Pi(\Omega-j\Delta\omega,\Omega+j\Delta\omega)|\omega_{0}+\Omega\rangle_{s}|\omega_{0}-\Omega\rangle_{i}\,d\Omega$
(1)
where $|j\rangle_{s,i}$ is the label of the $j^{\mathrm{th}}$ frequency mode
corresponding to the signal and the idler photon. $\omega_{0}$ is the
degeneracy frequency of the SPDC process, $\Delta\omega$ is the FSR between
modes and $\Pi$ is the spectral shape of each mode. We discretize the system
in bins of 25 GHz bandwidth and the same FSR over the whole C-band, yielding
the state:
$|\psi_{d}\rangle=\sum_{j=1}^{d}\alpha_{j}|j\rangle_{s}|j\rangle_{i}$ (2)
The term $\alpha_{j}$ refers to the phase and amplitude of the mode and is
determined by the spectral characteristics of the source. For a very broad and
uniform spectrum as here, $\alpha_{j}\approx 1$. Although after propagation,
each photon pair corresponding to mode $j$ accumulate different phases due to
material dispersion.
In a quantum key distribution (QKD) scenario, the frequency-entangled photon
pairs emerging from a single optical fiber can be distributed into different
paths, e.g., via wavelength- or polarization demultiplexing. In our
experiment, we use only one device simultaneously for both photons, but in
principle, manipulation would be just as easily possible at two separate
locations. To reveal the entanglement content, we use a commercial pulse
shaper to control the phase and amplitude of each of the frequency modes [16],
and electro-optic modulation to achieve mode superpositions [13] (see
Experimental Section).
### II.2 The superposition basis
For either quantum state characterization via full state tomography (FST),
evaluation of Bell-type tests, or implementation of QKD protocols,
measurements in superposition bases are fundamental to uncover quantum
correlations and statistics of the state. The eigenvectors of these bases may
be the superposition of some or all elements of the computational basis, here
the frequency basis, with certain phases for each mode. Here, we show that
with standard levels of RF signal amplification ($P_{\mathrm{max}}=26$ dBm),
it is possible to perform a full superposition of up to 7 modes with a low
contribution of accidental coincidence detection events. To prove this, we
performed high-dimensional Bell-type tests, also known as CGLMP [34], based on
the CHSH inequality for two-photon qubits [35]. Instead of measuring quantum
correlations only for fixed phase settings, we performed a phase scan for all
contributing modes [36]. The measurement projector we use for each photon is:
$|\Psi_{proj}\rangle=\frac{1}{\sqrt{d}}\sum_{j=1}^{d}\left(e^{ij\theta_{s,i}}|j\rangle_{s,i}\right)$
(3)
where $\theta_{s,i}$ is the phase applied to the signal or the idler photon,
and we use $\theta_{s}=\theta_{i}$. The phase of interference depends on the
sum of the signal and idler phases. By scanning their phase, we obtain the
quantum interferences shown in Fig. 2 for dimensions $d=2,3,5,7$. The
visibilities are 96.7%, 97.7%, 98.1% and 98.2%, respectively, without fitting
or subtraction of accidental coincidences and much above the thresholds 70.7%,
77.5%, 84.6% and 88.3% to rule out hidden variable theories [36]. To perform
these measurements, we have selected states centered on the $6^{\mathrm{th}}$
mode. Electro-optic modulation shifted photons from the neighboring modes into
the $6^{\mathrm{th}}$ mode, and demultiplexing filters (DEMUX) postselected
the superposition state. Our constrained RF signal only allowed a limited
efficiency of the frequency modulation of photons; thus, added mode loss by
the pulse shaper provided an equal contribution. Even dimensionality can also
be evaluated by using the same parameters as here.
Figure 2: Bell-type tests in the frequency domain for dimensions $d=2,3,5,7$.
We chose the modes centered on $|6\rangle$ for the signal and the idler
photons.
### II.3 Higher dimensionality certification
While these CGLMP tests provide a (partially) device-independent certification
of entanglement and demonstrate the reliability of our devices, they do not
easily test for actual entanglement dimensionality, i.e. the dimension of
entanglement needed to reproduce the correlations. Generically, to demonstrate
even higher levels of entanglement, one would need to project distant spectral
modes into superposition states, which may be limited or physically impossible
to perform. In our large frequency space and current setup, this would imply
unreachable RF power levels for electro-optic modulation. Although similar
practical limitations are to be expected for any type of encoding, there exist
other modulation methods that can provide frequency shifts of several
nanometers [37, 38]. On the other hand, common entanglement certification
techniques, such as FST, are expensive procedures that require measurements in
at least $(d+1)^{2}$ bases for bipartite systems, an equivalent of
$d^{2}(d+1)^{2}$ single-outcome measurement settings. The obvious consequence
is that the number of total measurements increases rapidly with dimension.
Adapting the techniques developed for the time-bin domain, we show that it is
sufficient to characterize the quantum coherence of a few 2D subspaces to
certify a high dimensionality of entanglement in the frequency domain.
The certification process is structured into two main blocks: the measurement
of some elements of the density matrix $\rho$ and the posterior lower bound of
the remaining unknowns. Measurements on the computational basis, that is, the
frequency basis, can be performed with standard filters and are, in fact,
directly related to the diagonal elements of the density matrix $\langle
j,k|\rho|j,k\rangle$. Only this characterization step would usually require
$d^{2}=10404$ filter settings, which can take arbitrary long times. Here, we
propose making good use of the frequency parallelization and bucket detection
of all uncorrelated frequencies with a single and broadband filter setting
$\sum_{k\neq j}\langle j,k|\rho|j,k\rangle$ (i.e. 101 modes $\times$ 25 GHz),
while still measuring the correlated set $|j,j\rangle$ with narrowband filter
settings (25 GHz). This method reduces the high number of measurements
required to only $2d$. The results of the maximum frequency correlation are
shown in Fig. 3a. The background noise measured for uncorrelated frequencies
originates from accidental coincidence detection events due to the high number
of single counts and imperfect filters. The detected coincidence to accidental
ratio (CAR) averaged throughout the space amounts to $1.4\times 10^{3}$.
Figure 3: (a) Intensity probabilities and (b) visibilities of the two-photon
interferences for the 2D subspaces of modes $|j,j+1\rangle$, $|j,j+2\rangle$
and $|j,j+6\rangle$. (c) Comparison of measured data with the expected lower
bound with our method. Note that, from all possible values of elements that
produce a positive density matrix $\rho$, we take the worst scenario.
Outside of the diagonal of $\rho$ we find two types of elements: those close
to zero due to energy conservation, as observed from the computational basis
measurements and upper limited by the accidentals in our system, and those
non-zero elements $\langle j,j|\rho|k,k\rangle$ that indicate the coherence of
the entangled state. We now measure some of the coherence elements to which we
have access with our measurement system. They can be estimated with the mode
amplitude from the computational basis and the strength of neighboring
interference. That is, we only need to measure the quantum interference
between any two modes of the whole system. We measure the interferences for
the first, second and sixth neighboring modes corresponding to the terms
$\langle j,j|\rho|j+i,j+i\rangle$ for $i=1,2,6$. The dispersion shifts the
interference patterns proportional to the frequency distance with respect to
the center wavelength [33]. Knowledge of the exact dispersion values would
allow us to directly measure the maxima and minima of the interference
patterns. For an unknown exact value as in our case, we perform a dispersion
calibration, a full phase scan of the interference for different frequency
modes. Finally, to collect statistics, we record 60 samples of the maxima and
minima of the expected interference for each subspace and calculate the
visibility. The total number of filter settings (not considering the
dispersion calibration) are now $2(d-1)+2(d-2)+2(d-6)=6(d-3)$ and the results
of all recorded visibilities are presented in Fig. 3b. The average visibility
for the 102 modes with 1st, 2nd, and 6th neighbors are, respectively,
96.85(7)%, 97.94(5)% and 96.8(1)%.
The fact that these high-contrast interference is preserved over the whole
investigated space indicates that the quantum state at hand is very close to a
maximally entangled state. However, in potentially adversarial scenarios, such
as QKD, we do not want to make any assumption on the distributed state. Thus,
we proceed with a rigorous analysis to demonstrate entanglement. Indeed, these
strong coherences allow us to finally certify a large amount of entanglement
without further measurements. Similarly to the methods proposed for
entanglement certification in the time domain [29, 28], we lower bound the
remaining unknown elements $\langle j,j|\rho|j+i,j+i\rangle$ (see Fig. 4a) by
using the fact that the density matrix must be positive semidefinite to
represent a valid quantum state. Thus, every element, submatrix, and therefore
subdeterminant of $\rho$ must be positive or equal to zero at the very least.
The magnitude of unknown density matrix elements can thus be lower-bounded
iteratively by solving $3\times 3$ subdeterminants, composed of measured or
estimated parameters and one unknown, and we keep the largest bound extracted
from all combinations of submatrices. The resulting submatrix $\langle
j,j|\rho|k,k\rangle$ is shown in Fig. 4b. Notice that this method causes a
rather fast loss of information, many of the elements are still lower-bounded
to zero, yet it is sufficient to certify high-dimensional entanglement. To
visualize this loss, we display in Fig. 3c the measured quantities for
$\langle j,j|\rho|j+2,j+2\rangle$ and the calculated bound if we would not use
those measurements.
Figure 4: (a) Simplified representation of the density matrix, considering
only the elements $\langle j,j|\rho|k,k\rangle$. While some visibilities
$V_{j}$ can be measured, the remaining unknowns in red are lower-bounded with
our method. (b) Lower bound of the density submatrix $\langle
j,j|\rho|k,k\rangle$ and (c) minimum certified dimensionality of our system,
according to the measured data. Notice that for low space dimension up to
$d=11$, we can certify the maximum dimensionality with very few measurement
settings.
We can now compare the lower-bounded density matrix with a target state
$|\Phi\rangle$ by computing the fidelity
$F(\rho,\Phi)=\mathrm{Tr}\left(\sqrt{\sqrt{\Phi}\rho\sqrt{\Phi}}\right)^{2}$.
There exists an upper bound for the fidelity of any state of Schmidt rank
$k\leq d$ [39, 40]. Fidelities with the maximally entangled state above the
threshold $B_{k}(\Phi)=k/d$ indicate a dimensionality of at least $k+1$. We
thus choose a maximally entangled state as the target state
$|\Phi\rangle=1/\sqrt{d}\sum_{j=1}^{d}|j,j\rangle$, iteratively calculate the
fidelity for $d=2$ to $d=102$ and compare it with the threshold values for
varying Schmidt rank. The final certification is plotted in Fig. 4c, where we
also show the dependence on the amount of measurement data. Certainly,
collecting the coherence information from more distant modes helps to improve
the certification. By using all the measured data, we find at least 33
entangled modes in a space of 101 to 102. To estimate the error of the
certification, we perform Monte Carlo simulations with our data error. On
average, the statistical method yields an even better certification, with 34.8
+- 0.7 entangled modes in a space of 102. It is worth emphasizing that with
only these very few measurement settings, $2d$ on the computational basis and
$\sim 6d$ on 2D subspaces, we are able to certify 11 dimensions in a space of
11 modes. Higher visibility would directly increase the amount of certified
entanglement per number of modes, and further measurements that could fill
more elements of the density matrix would also improve the certification.
## III Conclusion
The time-frequency domain straightforwardly provides entanglement of very high
dimensionality simply owing to energy conservation in spontaneous parametric
nonlinear processes. However, the characterization of this entanglement is a
key challenge. To date, high-dimensional frequency entanglement has never been
completely certified beyond a few dimensions. In our work, we have shown
methods on how to characterize these quantum states that unquestionably
possess a huge dimensionality. Similarly as in the time domain, the limits on
the dimensionality for continuously pumped processes depend on the resolution
of our devices, here the spectral filters.
In this work, we have shown full superpositions up to $d=7$ and great
interference visibilities over the full scanned spectrum. The used subspaces
are easily exploitable with few fiber-integrated optical components and can be
implemented nonlocally in quantum communication tasks. Due to physical
measurement constrains, limited amount of information is accessible from these
large states. Nevertheless, we have finally certified the full 11-dimensional
entanglement in a subspace of 11 modes with only $\sim 8d$ measurement
settings, and at least 33-dimensional entanglement in our frequency space of
102 frequency modes.
These results hold great promise for further enhancing the information
capacity of quantum information processing. Entanglement in the frequency
domain can also be used in combination with entanglement in time, path and
polarization DOFs, leading to huge state spaces. The main building blocks and
techniques can be readily combined with established spatial and temporal mode
approaches and thus provide the near-term prospect of more efficient hybrid
time-space photonic information processors.
We hope that these results on the quantification of quantum coherence in a
massive-dimensional frequency space will motivate the community to use the
frequency DOF beyond wavelength multiplexing schemes. Moreover, the further
development of photonic technology, in particular low-loss electro-optic
modulation and wave-shaping technologies, could render frequency encoding a
viable contender for near-term photonic quantum information processing with
massive bandwidth.
## IV Experimental section
To generate the frequency-entangled state, a commercial second harmonic
generation (SHG) module was used, a 40 mm type-0 ppLN waveguide (Covesion). To
pump the nonlinear process, a standard continuous-wave telecom laser was
upconverted with a second SHG module to better align the SPDC wavelength to
ITU channels. To process the frequency entanglement, a telecom pulse shaper
(Waveshaper 16000A) was used to select specific frequency subspaces for the
signal and the idler photons and to tune their relative phases. This device
limits the operational frequency range of the source to 40 nm of the telecom
C-band only. This space was then divided into 102 frequency modes for the
signal and the idler photon, with a standard free spectral range (FSR) of 25
GHz and the same bandwidth. This leads to a state space of $102\times 102$
dimensions. Note that subspace selection was an actual frequency
discretization procedure, and it could be employed as a resource for flexible
bandwidth allocation in reconfigurable QKD networks. The photon frequencies
were then modulated with an electro-optic modulator, driven by a radio-
frequency sine with the same frequency as the FSR. This sideband modulation
technique allowed us to distribute photons into neighboring modes according to
Bessel function amplitudes [41]. Choosing a low FSR and bandwidth allowed
slower modulation signals than in earlier work, where FSR values of 200 GHz
[16] down to 50 GHz [42] were used. Lower FSR would be better suited to
increase dimensionality, but would eventually be limited by the resolution of
the available optical filters and the number of photons per fraction of
spectral bandwidth. Lastly, due to photon scattering into distant spectral
modes, postselection was necessary to measure the right superposition states.
To reduce the noise that imperfect filters may introduce, a tunable and narrow
DEMUX of 22 GHz bandwidth centered at the corresponding signal and idler
frequencies was used prior to coincidence measurement. The colorful
illustrations at the bottom of Figure 1 depict the sideband modulation and
photon scattering.
## Funding
This work was supported by the Fraunhofer Internal Programs under Grant no.
Attract 066-604178, by the joint program of the Natural Sciences and
Engineering Research Council of Canada (NSERC) and the European Comission
under Grant no. HyperSpace 101070168, and by the Federal Ministry of Education
and Research of Germany (BMBF) through the project Quantum Photonics Labs
(QPL). The presented results were partially acquired using facilities and
devices funded by the Free State of Thuringia within the application center
Quantum Engineering. M.C.P is part of the Max Planck School of Photonics
supported by BMBF, Max Planck Society, and Fraunhofer Society.
## References
* [1] P. Kok, W. J. Munro, K. Nemoto, T. C. Ralph, J. P. Dowling, G. J. Milburn, _Reviews of Modern Physics_ 2007, _79_ , 1 135.
* [2] J.-W. Pan, Z.-B. Chen, C.-Y. Lu, H. Weinfurter, A. Zeilinger, M. Żukowski, _Reviews of Modern Physics_ 2012, _84_ , 2 777.
* [3] A. Anwar, C. Perumangatt, F. Steinlechner, T. Jennewein, A. Ling, _Review of Scientific Instruments_ 2021, _92_ , 4 041101\.
* [4] S. Ecker, F. Bouchard, L. Bulla, F. Brandt, O. Kohout, F. Steinlechner, R. Fickler, M. Malik, Y. Guryanova, R. Ursin, M. Huber, _Physical Review X_ 2019, _9_ , 4 041042.
* [5] B. P. Lanyon, M. Barbieri, M. P. Almeida, T. Jennewein, T. C. Ralph, K. J. Resch, G. J. Pryde, J. L. O’Brien, A. Gilchrist, A. G. White, _Nature Physics_ 2009, _5_ , 2 134.
* [6] C. Vigliar, S. Paesani, Y. Ding, J. C. Adcock, J. Wang, S. Morley-Short, D. Bacco, L. K. Oxenløwe, M. G. Thompson, J. G. Rarity, A. Laing, _Nature Physics_ 2021, _17_ , 10 1137.
* [7] A. Mair, A. Vaziri, G. Weihs, A. Zeilinger, _Nature_ 2001, _412_ , 6844 313.
* [8] M. Krenn, M. Huber, R. Fickler, R. Lapkiewicz, S. Ramelow, A. Zeilinger, _Proceedings of the National Academy of Sciences_ 2014, _111_ , 17 6243.
* [9] N. K. Fontaine, R. Ryf, H. Chen, D. T. Neilson, K. Kim, J. Carpenter, _Nature Communications_ 2019, _10_ , 1 1865.
* [10] J. D. Franson, _Physical Review Letters_ 1989, _62_ , 19 2205.
* [11] J. Brendel, N. Gisin, W. Tittel, H. Zbinden, _Physical Review Letters_ 1999, _82_ , 12 2594.
* [12] D. Richart, Y. Fischer, H. Weinfurter, _Applied Physics B_ 2012, _106_ , 3 543.
* [13] L. Olislager, I. Mbodji, E. Woodhead, J. Cussey, L. Furfaro, P. Emplit, S. Massar, K. P. Huy, J.-M. Merolla, _New Journal of Physics_ 2012, _14_ , 4 043015.
* [14] A. Pe’er, B. Dayan, A. A. Friesem, Y. Silberberg, _Physical Review Letters_ 2005, _94_ , 7 073601.
* [15] C. Bernhard, B. Bessire, T. Feurer, A. Stefanov, _Physical Review A_ 2013, _88_ , 3 032322.
* [16] M. Kues, C. Reimer, P. Roztocki, L. R. Cortés, S. Sciara, B. Wetzel, Y. Zhang, A. Cino, S. T. Chu, B. E. Little, D. J. Moss, L. Caspani, J. Azaña, R. Morandotti, _Nature_ 2017, _546_ , 7660 622.
* [17] H.-H. Lu, K. V. Myilswamy, R. S. Bennink, S. Seshadri, M. S. Alshaykh, J. Liu, T. J. Kippenberg, D. E. Leaird, A. M. Weiner, J. M. Lukens, _Nature Communications_ 2022, _13_ , 1 4338.
* [18] H.-H. Lu, E. M. Simmerman, P. Lougovski, A. M. Weiner, J. M. Lukens, _Physical Review Letters_ 2020, _125_ , 12 120503.
* [19] H.-H. Lu, J. M. Lukens, N. A. Peters, B. P. Williams, A. M. Weiner, P. Lougovski, _Optica_ 2018, _5_ , 11 1455.
* [20] H.-H. Lu, J. M. Lukens, B. P. Williams, P. Imany, N. A. Peters, A. M. Weiner, P. Lougovski, _npj Quantum Information_ 2019, _5_ , 1 24.
* [21] C. Cui, K. P. Seshadreesan, S. Guha, L. Fan, _Physical Review Letters_ 2020, _124_ , 19 190502.
* [22] Y. Chen, S. Ecker, J. Bavaresco, T. Scheidl, L. Chen, F. Steinlechner, M. Huber, R. Ursin, _Physical Review A_ 2020, _101_ , 3 032302.
* [23] N. B. Lingaraju, H.-H. Lu, S. Seshadri, P. Imany, D. E. Leaird, J. M. Lukens, A. M. Weiner, _Optics Express_ 2019, _27_ , 26 38683.
* [24] Y. Chen, S. Ecker, L. Chen, F. Steinlechner, M. Huber, R. Ursin, _npj Quantum Information_ 2021, _7_ , 1 1.
* [25] L. Zhang, Cui, Chaohan, Yan, Jianchang, Guo, Yanan, Wang, Junxi, Fan, Linran, On-chip parallel processing of quantum frequency combs for high-dimensional hyper-entanglement generation, Technical Report arXiv:2111.12784, 2021.
* [26] C. Reimer, S. Sciara, P. Roztocki, M. Islam, L. Romero Cortés, Y. Zhang, B. Fischer, S. Loranger, R. Kashyap, A. Cino, S. T. Chu, B. E. Little, D. J. Moss, L. Caspani, W. J. Munro, J. Azaña, M. Kues, R. Morandotti, _Nature Physics_ 2019, _15_ , 2 148.
* [27] P. Imany, J. A. Jaramillo-Villegas, M. S. Alshaykh, J. M. Lukens, O. D. Odele, A. J. Moore, D. E. Leaird, M. Qi, A. M. Weiner, _npj Quantum Information_ 2019, _5_ , 1 59.
* [28] A. Martin, T. Guerreiro, A. Tiranov, S. Designolle, F. Fröwis, N. Brunner, M. Huber, N. Gisin, _PHYSICAL REVIEW LETTERS_ 2017, 5.
* [29] A. Tiranov, S. Designolle, E. Z. Cruzeiro, J. Lavoie, N. Brunner, M. Afzelius, M. Huber, N. Gisin, _PHYSICAL REVIEW A_ 2017, 6.
* [30] J. Schneeloch, G. A. Howland, _Physical Review A_ 2018, _97_ , 4 042338.
* [31] J. Schneeloch, C. C. Tison, M. L. Fanto, P. M. Alsing, G. A. Howland, _Nature Communications_ 2019, _10_ , 1 2785.
* [32] Z. Xie, T. Zhong, S. Shrestha, X. Xu, J. Liang, Y.-X. Gong, J. C. Bienfang, A. Restelli, J. H. Shapiro, F. N. C. Wong, C. Wei Wong, _Nature Photonics_ 2015, _9_ , 8 536.
* [33] P. Imany, O. D. Odele, J. A. Jaramillo-Villegas, D. E. Leaird, A. M. Weiner, _Physical Review A_ 2018, _97_ , 1 013813.
* [34] D. Collins, N. Gisin, N. Linden, S. Massar, S. Popescu, _Physical Review Letters_ 2002, _88_ , 4 040404.
* [35] J. F. Clauser, M. A. Horne, A. Shimony, R. A. Holt, _Physical Review Letters_ 1969, _23_ , 15 880.
* [36] R. T. Thew, A. Acín, H. Zbinden, N. Gisin, _Physical Review Letters_ 2004, _93_ , 1 010503.
* [37] C. Joshi, A. Farsi, S. Clemmen, S. Ramelow, A. L. Gaeta, _Nature Communications_ 2018, _9_ , 1 847.
* [38] D. Zhu, C. Chen, M. Yu, L. Shao, Y. Hu, C. J. Xin, M. Yeh, S. Ghosh, L. He, C. Reimer, N. Sinclair, F. N. C. Wong, M. Zhang, M. Lončar, _Light: Science & Applications_ 2022, _11_ , 1 327\.
* [39] R. Fickler, R. Lapkiewicz, M. Huber, M. P. Lavery, M. J. Padgett, A. Zeilinger, _Nature Communications_ 2014, _5_ , 1 4502.
* [40] J. Bavaresco, N. Herrera Valencia, C. Klöckl, M. Pivoluska, P. Erker, N. Friis, M. Malik, M. Huber, _Nature Physics_ 2018, _14_ , 10 1032.
* [41] J. Capmany, C. R. Fernández-Pousa, _JOSA B_ 2010, _27_ , 6 A119.
* [42] P. Imany, J. A. Jaramillo-Villegas, O. D. Odele, K. Han, D. E. Leaird, J. M. Lukens, P. Lougovski, M. Qi, A. M. Weiner, _Optics Express_ 2018, _26_ , 2 1825.
|
# FRETting about Requirements: Formalised Requirements for an Aircraft Engine
Controller††thanks: The authors thank Georgios Giantamidis, Stylianos
Basagiannis, and Vassilios A. Tsachouridis (UTRC, Ireland) for their help in
requirements elicitation; and Anastasia Mavridou (NASA Ames Research Center,
USA) for her help with FRET. This research was funded by the European Union’s
Horizon 2020 research and innovation programme under the VALU3S project (grant
No 876852), and by Enterprise Ireland (grant No IR20200054). The funders had
no role in study design, data collection and analysis, decision to publish, or
preparation of the manuscript.
Marie Farrell Matt Luckcuck Oisín Sheridan Rosemary Monahan
(Department of Computer Science, Maynooth University, Ireland
<EMAIL_ADDRESS>
###### Abstract
[Context & motivation] Eliciting requirements that are detailed and logical
enough to be amenable to formal verification is a difficult task. Multiple
tools exist for requirements elicitation and some of these also support
formalisation of requirements in a way that is useful for formal methods.
[Question/problem] This paper reports on our experience of using the Formal
Requirements Elicitation Tool (FRET) alongside our industrial partner. The use
case that we investigate is an aircraft engine controller. In this context, we
evaluate the use of FRET to bridge the communication gap between formal
methods experts and aerospace industry specialists. [Principal ideas/results]
We describe our journey from ambiguous, natural-language requirements to
concise, formalised FRET requirements. We include our analysis of the
formalised requirements from the perspective of patterns, translation into
other formal methods and the relationship between parent-child requirements in
this set. We also provide insight into lessons learned throughout this process
and identify future improvements to FRET. [Contribution] Previous experience
reports have been published by the FRET team, but this is the first such
report of an industrial use case that was written by researchers that have not
been involved FRET’s development.
## 1 Introduction
Formal verification uses mathematically-based techniques to guarantee that a
system obeys certain properties, which is particularly useful when developing
safety-critical systems like those used in the aerospace domain. Developing a
correct set of requirements necessitates discussion with people who have
expertise in the system under development, who may not have skills in formal
methods. In which case, it can be beneficial to the requirements elicitation
process to write the requirements in an intermediate language. Tools like
NASA’s Formal Requirements Elicitation Tool (FRET) provide a gateway for
developing formal requirements with developers who are not familiar with
formal languages [7].
In this paper, we examine how FRET can be used in an industrial case study of
an aircraft engine controller that has been supplied by our industrial
partner, United Technologies Research Center (Ireland). FRET has previously
been used to formalise the requirements for the 10 Lockheed Martin Cyber-
Physical Challenges [11]. However, to the best of our knowledge this paper
provides the first experience report on FRET’s use on an industrial case
study, by a team not involved in FRET’s development.
Our approach provides external and internal traceability. Using a tool like
FRET to develop the requirements provides a link between natural-language
requirements and formally verified artefacts. FRET also enables the user to
describe a link between requirements at different levels of abstraction, which
means that this traceability is maintained within the developing set of
requirements. We also use FRET to collect information about the rationale
behind a requirement, further improving the traceability; either back to a
natural-language requirement, or forward to a more concrete requirement. These
traceability features encourage better explainability of a requirement’s
source, and the intermediate language improves the explainability of the
requirements themselves.
The rest of the paper is laid out as follows. §2 outlines the relevant
background material pertaining to FRET and the aircraft engine controller use
case. Then, we describe our requirements elicitation process and present
detailed requirements in §3. These requirements are analysed in §4. We discuss
the lessons that were learned through this work in §5 and §6 concludes. We
also provide a detailed appendix showing the full set of requirements, test
cases, Simulink model and fretish requirements.
## 2 Background
This section provides an overview of FRET and the aircraft engine controller
use case for which we were developing requirements.
#### FRET:
is an open-source tool that enables developers to write and formalise system
requirements [7]. FRET accepts requirements written in a structured natural-
language called fretish, in which requirements take the form:
scope condition component shall timing response
The condition, component, and response fields are mandatory; scope and timing
are optional fields. This allows responses that are tied to a scope, are
triggered by conditions, relate to a system component, and may have timing
constraints.
The underlying semantics of a fretish requirement is determined by the scope,
condition, timing, and response fields. There is a template for each possible
combination of a requirement’s fields, currently FRET provides 160 such
templates [8]. The selected template is used to generate formalisations of the
associated requirement in both past- and future-time metric Linear-time
Temporal Logic (LTL). FRET displays a diagramatic semantics for each
requirement, which shows: the time interval where it should hold, and its
triggering and stopping conditions (if they exist). Both versions of the
requirements are helpful for sanity-checking what has been written in fretish.
The user must give each fretish requirement an ID, which can be used to create
a many-to-many, hierarchical link between requirements: a parent requirement
may have many child requirements, and one child may have many parents. While
this link facilitates traceability, FRET does not define this relationship
(formally or otherwise). For example, a child requirement does not inherit
definitions from its parent. We discuss possible improvements to this link in
§4.2. FRET also allows the user to enter ‘Rationale’ and ‘Comments’ for a
requirement, which further supports traceability and encourages explainability
of requirements.
FRET can automatically translate requirements into contracts for a Simulink
diagram, written in CoCoSpec, which are checked during Simulink simulations by
the CoCoSim tool, using the Kind2 model checker [4]. FRET can also generate
runtime monitors for the Copilot framework [6].
#### Aircraft Engine Controller:
Our use case is a software controller for a high-bypass civilian aircraft
turbofan engine, provided by our industrial partner on the VALU3S [2] project,
based on existing controller designs [13, 14]. It is an example of a Full
Authority Digital Engine Control (FADEC) system, which monitors and controls
everything about the engine,using input from a variety of sensors. The engine
itself contains two compressors (high-pressure and low-pressure) turning a
central spool, which drives the engine.
As described in our prior work [9], the controller’s high-level objectives are
to manage the engine thrust, regulate the compressor pressure and speeds, and
limit engine parameters to safe values. It should continue to operate, keeping
settling time, overshoot, and steady state errors within acceptable limits,
while respecting the engine’s operating limits (e.g. the spool’s speed limit),
in the presence of:
* •
sensor faults (a sensor value deviating too far from from its nominal value,
or being unavailable),
* •
perturbation of system parameters (a system parameter deviating too far from
from its nominal value), and
* •
other low-probability hazards (e.g. abrupt changes in outside air pressure).
The controller is also required to detect engine surge or stall and change
mode to prevent these hazardous situations.
Our industrial partner has supplied us with 14 English-language requirements
(Table 1) and 20 abstract test cases, which provide more detail about the
controller’s required behaviour. The naming convention for requirements is:
$<$use case id$>$_R_$<$parent requirement id$>$.$<$child requirement id$>$
For example, because this is Use Case 5 in the VALU3S project111The VALU3S
project: https://valu3s.eu/, requirement one is named UC5_R_1. Note, we use a
similar naming convention for test cases. Table 2 shows the abstract test
cases for UC5_R_1. Our industrial partner also designed the controller in
Simulink222Simulink: https://mathworks.com/products/simulink.html, shown in
the extended version of this paper333This Paper Extended Version:
https://arxiv.org/abs/2112.04251.
For our Use Case, we collaborated with scientists in the System Analysis and
Assurance division of an aerospace systems company. The hour-long requirements
elicitation meetings were held monthly, over a period of 10 months, with
additional meetings as needed. In these meetings, our collaborators reviewed
the fretish versions of their natural-language requirements, validating our
formalisation and clarifying ambiguities for us. Since our collaborators were
already familiar with other formal tools we were able to introduce them to
FRET quite quickly. However, we produced a training video for other members of
the project consortium444“Formalising Verifiable Requirements” Presentation:
https://www.youtube.com/watch?v=FQGKbYCbxPY&list=PLGtGM9euw6A66ceQbywXGjVoTKEhP-
Of7&index=9.
## 3 Our Requirements Elicitation Process Using FRET
In this section we describe our requirements elicitation process. We begin by
outlining how this fits into our larger approach to verification for the
aircraft engine controller use case. We then describe our journey from
natural-language requirements to formalised fretish requirements.
### 3.1 Requirements-Driven Methodology
As part of the three-phase verification methodology outlined in our prior work
[9], we used FRET to elicit and formalise requirements for the aircraft engine
controller. Focussing on Phase 1, this paper includes the full set of fretish
requirements and presents lessons learnt, which were not discussed in [9].
Fig. 1 shows a high-level flowchart of our methodology, with an exploded view
of the relationship between the artefacts involved in Phase 1. The methodology
takes the natural-language requirements (Table 1), test cases, and Simulink
diagram of the engine controller as input, and enables the formal verification
of the system’s design against the requirements.
Figure 1: Flowchart of our three-phase requirements-driven verification
methodology [9] (left) with an exploded view of Phase 1’s artefacts (right).
The solid lines and arrowheads show direct information flow between artefacts
(and back into Phase 2A and 2B), the dashed lines and open arrowheads show an
artefact being implemented by another.
Phase 1 of our methodology involves formalising natural-language requirements
using FRET, eliciting further detail as we progress. Phase 2 consists of two,
potentially parallel, phases. Phase 2A uses FRET’s built-in support for
generating CoCoSpec contracts that can be incorporated into a Simulink diagram
for verification with the Kind2 model-checker. In Phase 2B, the formalised
requirements drive a (manual) translation into other formal methods, as chosen
by the verifier. These tools typically require the construction of a formal
model of the system, which is verified against the translated requirements.
This step requires translation of the fretish requirements into the formalism
of the chosen verification tool. Finally, Phase 3 produces a report collecting
the verification results and other useful artefacts, such as formal models,
Simulink diagrams, various versions of the requirements, counter-examples,
proofs, etc. This supports tracing the requirements through the system’s
development lifecycle.
The following subsections describe the requirements elicitation process (Phase
1) in more detail. Figure 1’s exploded view, shows how natural-language
requirements are translated into fretish parent requirements (solid arrow),
and the test cases are translated into child requirements. Since we view the
test cases as implementations of the natural-language requirements (dashed
arrow), the child requirements are similarly viewed as implementations of
their corresponding parent requirements. The left-hand side of Fig. 1 shows
how the work in this paper fits within our development and verification
methodology; the solid arrows from the fretish parent and child requirements,
to Phases 2A and 2B, show how the output of this work is consumed by the next
phase.
ID | Description
---|---
UC5_R_1 | Under sensor faults, while tracking pilot commands, control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits)
UC5_R_2 | Under sensor faults, during regulation of nominal system operation (no change in pilot input), control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits)
UC5_R_3 | Under sensor faults, while tracking pilot commands, operating limit objectives shall be satisfied (e.g., respecting upper limit in shaft speed)
UC5_R_4 | Under sensor faults, during regulation of nominal system operation (no change in pilot input), operating limit objectives shall be satisfied (e.g., respecting upper limit in shaft speed)
UC5_R_5 | Under mechanical fatigue conditions, while tracking pilot commands, control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits)
UC5_R_6 | Under mechanical fatigue conditions, during regulation of nominal system operation (no change in pilot input), control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits)
UC5_R_7 | Under mechanical fatigue conditions, while tracking pilot commands, operating limit objectives shall be satisfied (e.g., respecting upper limit in shaft speed)
UC5_R_8 | Under mechanical fatigue conditions, during regulation of nominal system operation (no change in pilot input), operating limit objectives shall be satisfied (e.g., respecting upper limit in shaft speed)
UC5_R_9 | Under low probability hazardous events, while tracking pilot commands, control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits)
UC5_R_10 | Under low probability hazardous events, during regulation of nominal system operation (no change in pilot input), control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits)
UC5_R_11 | Under low probability hazardous events, while tracking pilot commands, operating limit objectives shall be satisfied (e.g., respecting upper limit in shaft speed)
UC5_R_12 | Under low probability hazardous events, during regulation of nominal system operation (no change in pilot input), operating limit objectives shall be satisfied (e.g., respecting upper limit in shaft speed)
UC5_R_13 | While tracking pilot commands, controller operating mode shall appropriately switch between nominal and surge / stall prevention operating state
UC5_R_14 | During regulation of nominal system operation (no change in pilot input), controller operating mode shall appropriately switch between nominal and surge / stall prevention operating state
Table 1: Natural-language requirements for the aircraft engine controller as produced by the aerospace use case in the VALU3S project. These 14 requirements are mainly concerned with continued operation of the controller in the presence of sensor faults (UC5_R_1–UC5_R_4), perturbation of system parameters (UC5_R_5–UC5_R_8) and low probability hazards (UC5_R_9–UC5_R_12). There are also requirements for switching between modes if engine surge/stall is detected (UC5_R_13–UC5_R_14). Test Case ID | Requirement ID | Description
---|---|---
UC5_TC_1 | UC5_R_1 | Preconditions: Aircraft is in operating mode M and sensor S value deviates at most +/- R % from nominal value
Input conditions/steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting control objectives (settling time, overshoot, steady state error)
UC5_TC_2 | UC5_R_1 | Preconditions: Aircraft is in operating mode M and sensor S value is not available (sensor is out of order)
Input conditions/steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting control objectives (settling time, overshoot, steady state error)
Table 2: Abstract test cases corresponding to requirement UC5_R_1. Each
specifies the preconditions for the test case, the input conditions/steps and
the expected results.
### 3.2 Speaking FRETISH: Parent Requirements
ID | FRETISH
---|---
UC5_R_1 | if ((sensorfaults) & (trackingPilotCommands)) Controller shall satisfy (controlObjectives)
UC5_R_2 | if ((sensorfaults) & (!trackingPilotCommands)) Controller shall satisfy (controlObjectives)
UC5_R_3 | if ((sensorfaults) & (trackingPilotCommands)) Controller shall satisfy (operatingLimitObjectives)
UC5_R_4 | if ((sensorfaults) & (!trackingPilotCommands)) Controller shall satisfy (operatingLimitObjectives)
UC5_R_5 | if ((mechanicalFatigue) & (trackingPilotCommands)) Controller shall satisfy (controlObjectives)
UC5_R_6 | if ((mechanicalFatigue) & (!trackingPilotCommands)) Controller shall satisfy (controlObjectives)
UC5_R_7 | if ((mechanicalFatigue) & (trackingPilotCommands)) Controller shall satisfy (operatingLimitObjectives)
UC5_R_8 | if ((mechanicalFatigue) & (!trackingPilotCommands)) Controller shall satisfy (operatingLimitObjectives)
UC5_R_9 | if ((lowProbabilityHazardousEvents) & (trackingPilotCommands)) Controller shall satisfy (controlObjectives)
UC5_R_10 | if ((lowProbabilityHazardousEvents) & (!trackingPilotCommands)) Controller shall satisfy (controlObjectives)
UC5_R_11 | if ((lowProbabilityHazardousEvents) & (trackingPilotCommands)) Controller shall satisfy (operatingLimitObjectives)
UC5_R_12 | if ((lowProbabilityHazardousEvents) & (!trackingPilotCommands)) Controller shall satisfy (operatingLimitObjectives)
UC5_R_13 | if (trackingPilotCommands) Controller shall satisfy (changeMode(nominal)) $|$ (changeMode(surgeStallPrevention))
UC5_R_14 | if (!trackingPilotCommands) Controller shall satisfy (changeMode(nominal)) $|$ (changeMode(surgeStallPrevention))
Table 3: fretish parent requirements corresponding to the natural-language
requirements outlined in Table 1. The correspondance is clear to see and we
have used booleans to represent the ambiguous terms from the natural-language
requirements.
The inputs to our requirements elicitation process were the Simulink diagram,
14 natural-language requirements (Table 1), and 20 abstract test cases that
were supplied by our industrial partner. We elicited further information about
the requirements through regular team discussions with our industrial partner.
We started by translating the natural-language requirements into fretish,
producing the set of 14 fretish requirements in Table 3. The correspondence
between the fretish requirements and their natural-language counterparts is
clear. For example, requirement UC5_R_1 states that:
* Under sensor faults, while tracking pilot commands, control objectives shall be satisfied (e.g., settling time, overshoot, and steady state error will be within predefined, acceptable limits).
This became the corresponding fretish requirement:
* if ((sensorfaults) & (trackingPilotCommands)) Controller shall
satisfy (controlObjectives)
Producing this initial set of requirements enabled us to identify the
ambiguous parts of the natural-language requirements. For example, the phrase
“sensor faults” simply becomes a boolean in our fretish requirements,
highlighting that we need to elicit more details. We captured these additional
details as child requirements, as described in §3.3
### 3.3 Adding Detail: Child Requirements
Once the fretish parent requirements (Table 3) were complete, we added more
detail to make the requirements set more concrete. We paid particular
attention to ambiguous phrases translated from the natural-language
requirements. These extra details were drawn from the abstract test cases and
from detailed discussions with our industrial collaborators, who clarified
specific ambiguities.
We captured the extra details in 28 child requirements. As mentioned in §2, a
child requirement does not inherit definitions from its parent(s). However, we
use this hierarchical link to group the detail in the child requirements under
a common parent, which enables the detailed child requirements to be traced
back to the more abstract parent requirements.
For example, UC5_R_1 was distilled into three requirements (UC5_R_1.1,
UC5_R_1.2 and UC5_R_1.3), shown in Table 4. These three child requirements
each have the same condition and component, but differ in their responses.
Each child requirement specifies one of the “control objectives” (settling
time, overshoot and steady state error) mentioned in the natural-language
version of UC5_R_1. During elicitation discussions, it was revealed that these
were actually the only control objectives that were of concern for this use
case. Here, using FRET encouraged us to question exactly what the phrase
“control objectives” meant.
Each of these requirements includes the condition when (diff(r(i),y(i)) $>$ E)
and the timing constraint until (diff(r(i),y(i)) $<$ e), which were initially
overlooked in the natural-language requirements but revealed during
elicitation discussions with our industrial partner. The response must hold
when the difference between the reference sensor value, $r(i)$, and the
observed sensor value, $y(i)$, falls between specific bounds ($E$ and $e$).
This important detail was missing from the parent requirement but was
uncovered during our requirements elicitation.
The “Preconditions” of test cases UC5_TC_1 and UC5_TC_2 (Table 2) showed us
that the phrase “Under sensor faults” meant a period where a sensor value
deviates by ±R% from its nominal value or returns a null value. To represent
this, the child requirements use the function sensorValue(S) where S is a
parameter representing each of the 4 sensors for the engine controller. These
requirements are thus applied to all of the sensors in the model.
In UC5_TC_1 and UC5_TC_2, the “Input conditions/steps” refer to the aircraft
thrust and a change in the pilot’s input. We encoded this as the condition and
response pair (pilotInput $=>$ setThrust = V2) & (observedThrust = V1) and
satisfy (observedThrust = V2), where V1 and V2 are thrust variables and => is
logical implication. During elicitation discussions we found that this pair
corresponds to the condition, trackingPilotCommands. This was a particularly
important clarification because trackingPilotCommands models the phrase “while
tracking pilot commands”, which the natural-language requirements use
extensively. This underlines that it is possible for an ambiguous statement to
have a very precise meaning that was lost while drafting the requirements.
The thrust variables V1 and V2 in our fretish requirements correspond to
variables V1, V2, A1, and A2 in the test cases. During elicitation
discussions, we found that V1 and V2 alone were sufficient to capture the
requirement. V1 and A1 are used interchangeably as the initial thrust value,
which we label V1. Similarly, V2 and A2 refer to the updated thrust value,
which we label V2 for consistency. This is another ambiguity that our
translation from natural-language to fretish helped to clarify.
Our industrial partner checked the child requirements to ensure that there
were no errors or omissions. The intuitive meaning of fretish constructs
simplified this check, and features like the requirements’ diagramatic
semantics provided quick feedback when we edited the requirements during
elicitation discussions. The act of formalising the requirements helped us to
identify ambiguities in the requirements, prompting elicitation of further
detail from our industrial partner.
ID | Parent | FRETISH
---|---|---
UC5_R_1.1 | UC5_R_1 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null) & (pilotInput $=>$ setThrust = V2) & (observedThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V2)
UC5_R_1.2 | UC5_R_1 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (pilotInput $=>$ setThrust = V2) & (observedThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V2)
UC5_R_1.3 | UC5_R_1 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (pilotInput $=>$ setThrust = V2)& (observedThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateErrorMax) & (observedThrust = V2)
Table 4: We have three distinct child requirements for UC5_R_1 that capture
the correct behaviour with respect to each of settling time, overshoot and
steady state error.
## 4 An Analysis of Elicited Requirements
This section provides an analysis of the fretish requirements that we produced
for the aircraft engine controller use case. We note that the requirements
only refer to one component, the Controller, but this could be decomposed to
refer to specific blocks in the use case Simulink design.
### 4.1 Requirement Templates
Each of the 14 fretish parent requirements (Table 3) uses the same pattern:
condition component shall response. As described in §2, FRET maps each
requirement into a semantic template so that it can generate the associated
LTL specification. Our parent requirements all correspond to the template
[null, regular, eventually], which specifies the scope-option, condition-
option and timing-option, respectively (if the timing-option is omitted, then
eventually is the default). Specific details about templates in FRET are given
in [8].
We introduced until clauses into all of the 28 child requirements, although
with different timing constraints. The introduction of the until clauses was
identified through a combination of the information in the test cases and from
extensive discussions with our industrial parter. However, the specific timing
constraints required in-depth discussion with our industrial partner to
identify. Most of the child requirements correspond to the template [null,
regular, until]. However, some child requirements differed slightly as
outlined below.
ID | Parent | FRETISH
---|---|---
UC5_R_13.1 | UC5_R_13 | in nominal mode when (diff(setNL, observedNL) $>$ NLmax) if (pilotInput $=>$ surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $<$ NLmin) satisfy (changeMode(surgeStallPrevention))
UC5_R_13.2 | UC5_R_13 | in surgeStallPrevention mode when (diff(setNL, observedNL) $<$ NLmax) if (pilotInput $=>$ !surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $>$ NLmin) satisfy (changeMode(nominal))
UC5_R_14.1 | UC5_R_14 | in nominal mode when (diff(setNL, observedNL) $>$ NLmax) if (!pilotInput $=>$ surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $<$ NLmin) satisfy (changeMode(surgeStallPrevention))
UC5_R_14.2 | UC5_R_14 | in surgeStallPrevention mode when (diff(setNL, observedNL) $<$ NLmax) if (!pilotInput $=>$ !surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $>$ NLmin) satisfy (changeMode(nominal))
Table 5: Child requirements corresponding to UC5_R_13 and UC5_R_14. These
differ from the previous requirements because we use the scope field to assert
which mode of operation the controller is in.
UC5_R_13 and UC5_R_14 generated a lot of discussion, because they differ so
much from the other requirements; here, the system changes between modes of
operation, so we use the scope clause. This produced the child requirements
shown in Table 5. The when and until clauses differ from the other
requirements, because here the mode change is triggered by comparing the set
value of the low-pressure compressor’s spool speed (setNL) to the value
produced by the sensor (observedNL). It is necessary to differentiate between
the cause of the difference, i.e. whether it was directly caused by pilotInput
or by external factors (!pilotInput). In either case the system must change
modes, but our industrial partner felt that it was important that the
requirements distinguish the difference.
Figure 2: fretish and semantics diagram generated for UC5_R_14.2. Here, ‘M’
indicates the mode, ‘TC’ the triggering condition, and ‘SC’ the stopping
condition.
Fig. 2 contains the semantics diagram produced by FRET for UC5_R_14.2. The
semantic template that was used is [in, regular, until]. In a recent study
using FRET to formalise the 10 Lockheed Martin Cyber Physical Challenge
problems, the most commonly used semantic template was [null, null, always]
[11]. Of these 10 problems, the autopilot system is the closest to our case
study, and it was the only requirement set to use the in scope-option. The
timing-option in their requirements was different to ours; but we use until,
which was introduced into fretish after that study.
We used all of the fields available in fretish in our use case, although a lot
of our individual requirements used a subset of them. We only used scope in
the four child requirements of UC5_R_13 and UC5_R_14. FRET provides many ways
of specifying modes, but we only used in mode for this; there are many ways to
specify a condition, but we only used when and if. There are also multiple
ways to specify timing, but in this case study we only used until clauses.
Despite until providing timing constraints, we did not use explicit times
(e.g. ticks/timesteps/seconds) in our requirements. This is because the
natural-language requirements (Table 1) do not mention timing, and our
introduction of timing constraints came from elicitation discussions. However,
time points are implicit in some of the child requirements, e.g. comparing
r(i) and y(i) in the child requirements of UC5_R_1(Table 4), or the T1 and T2
variables in UC5_R_11.1 (Table 6). The timing clause was not intentionally
avoided, but we felt that the implicit time constraints better suited the
requirements and was closer to the description agreed with our industrial
partner.
ID | Parent | FRETISH
---|---|---
UC5_R_11.1 | UC5_R_11 | when (diff(r(i),y(i)) $>$ E) if (outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) &(abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) &(observedThrust = V1) &(pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V2)
Table 6: Child requirement of UC5_R_11 which has timing implicit through the
use of the timestamp variables T1 and T2.
### 4.2 Parent-Child Relationship in our Use Case
As previously mentioned, FRET allows a requirement to be related to another as
a ‘parent’ or a ‘child’, but this relationship is not well defined, formally
or otherwise. The parent-child relationship in FRET could be viewed as formal
refinement [1]: a concept supported by a variety of formal methods that enable
formal specifications to be gradually made more concrete, while proving that
they still obey a more abstract version of the specification. Similar
approaches exist in the literature on refactoring goal-driven requirements [5,
15].
If viewed through the lens of refinement, we would need to introduce
abstraction invariants to relate the abstract and concrete specifications.
These invariants facilitate the proof that the concrete specification does not
permit any behaviours that the abstract specification forbids.
Here, we investigate whether FRET’s parent-child relationship can be expressed
as formal refinement. In particular, it is possible to formalise the following
abstraction invariant in relation to sensorfaults:
sensorfaults $\iff$ (sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S)
$<$ nominalValue - R) $|$ (sensorValue(S) = null)
Intuitively this means that the boolean sensorfaults (from the parent
requirement) corresponds to the condition on the right of the ‘$\iff$’ (from
the child requirement). This kind of refinement is referred to as data
refinement.
Similarly, the abstraction invariant between trackingPilotCommands and the
condition and response pair (pilotInput $=>$ setThrust = V2) & (observedThrust
= V1) and satisfy (observedThrust = V2) could be specified as:
trackingPilotCommands $\iff$ pilotInput
The remainder of the condition-response pair above is then treated as
superposition refinement, which adds detail during refinement. This approach
is used because of the update of the observedThrust variable which is
difficult to express in an abstraction invariant because it provides a
behavioural update rather than a simple match between booleans. The additional
when and until clauses in the child requirement are also superposition
refinements.
The parent-child relationship in FRET appears to us to be formal refinement,
at least for our set of requirements. In which case UC5_R_1 is refined by its
three child requirements (UC5_R_1.1, UC5_R_1.2, UC5_R_1.3). We will examine
this further in future work, where we will seek to translate these
requirements into a formalism that supports refinement, and then examine
whether the appropriate proof obligations can be discharged by theorem
provers.
### 4.3 Translatable Requirements
As mentioned in §1, our aim is to formally verify the aircraft engine
controller system described in §2. It is often difficult to identify what
properties a system should obey, for example what does it mean for the system
to operate ‘correctly’. Identifying the properties to verify often causes
difficulties for non-domain experts. FRET helped to guide conversations with
the domain experts to facilitate the formalisation of the requirements.
FRET currently supports translation to the CoCoSim [4] and Copilot [12]
verification tools. We are particularly interested in using CoCoSim since it
works directly on Simulink diagrams. Thus, we have started to generate CoCoSim
contracts for these requirements automatically using FRET [10]. This is
described in [9] and corresponds to Phase 2A of the methodology outlined in
Fig. 1.
As described in §4.1, we didn’t rely heavily on timing constraints that
specified specific time steps, rather we used until constraints that could
potentially be translated into boolean flags in other formalisms. As such, we
believe that the vast majority of the requirements that we formalised in FRET
could be used by other formal methods. For example, we may need to model the
aircraft engine controller in an alternative formalism if some of these
properties fail to verify using CoCoSim due to the state space explosion. This
approach has been taken, manually, in previous work [3].
## 5 Lessons Learnt and Future Improvements
This section summarises some of the lessons that we learnt from this case
study.
Communication Barrier: We found that FRET and fretish provided a useful
conduit for conversation with our industrial partner. Formalising natural-
language requirements is often time-consuming because of contradictions and
ambiguities. fretish provides a stepping-stone between readable natural-
language requirements and their fully-formal counterparts, and helped us to
step-wise reduce ambiguity. This process produced requirements that are easier
to read than if they had been fully-formal, but which FRET can still
automatically formalise.
We used FRET during elicitation discussions to explain and update our
requirements, alongside our industrial partner. The diagramatic semantics gave
a useful visualisation of the requirements, helping both us and our industrial
partner to sanity-check updates. FRET also enabled our documentation of
information for each natural-language requirement, recording the reasoning for
any changes, alongside each fretish requirement, thus facilitating
requirements explainability.
Parent-Child Relationship: While not a formal relationship, the link between
parent and child requirements enabled us to gradually make the requirements
more concrete, by adding details and removing ambiguities. For example, the
term sensorfaults in UC5_R_1 was replaced with (sensorValue(S) $>$
nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$
(sensorValue(S) = null) in its child requirements (Table 4). Documenting these
links, via the ‘Parent ID’ and ‘Rationale’ fields in FRET, provides a
structuring mechanism that enables traceability within the requirement set.
However, a more concrete definition of this link would be beneficial. We have
suggested a definition using formal refinement, but an object-oriented
inheritance relationship could also provide structure here.
Limitations of FRETISH: While a useful language, we note some limitations of
fretish. Logical quantifiers ($\forall,\exists$) would be a welcome addition
to fretish. For example, in UC5_R_1.1, we used sensorValue(S), where the
parameter S indicates that this condition applies to all sensors. This is
slight abuse of notation, it would have been more accurate to use a $\forall$
quantifier.
We also suggest that highlighting assignments to variables (which hold after
the requirement is triggered) would be beneficial. For example, in UC5_R_1.1
we use the observedThrust variable in both the condition and the response. We
expect that observedThrust has been updated by the response but this is not
obvious, and may have implications when translating to other verification
tools.
An Industrial Perspective: Our industrial partner had not used FRET before, so
we asked them about their experience with it. They felt that the fretish
requirements were ‘much more clear’ than the natural-language requirements,
and that using a ‘controlled-natural language with precise semantics is always
better than natural-language’. When asked if FRET was difficult to use or
understand they said that FRET was ‘very easy to use; interface is intuitive;
formal language is adequately documented inside the tool (along with usage
examples)’. Overall, they found that FRET was useful ‘because it forces you to
think about the actual meaning behind the natural-language requirements’.
Having installed FRET, our industrial partner found some usability
improvements that could be made. Some were problems with the GUI, which have a
low impact but happen very frequently. Other issues related to fretish; for
example, they would like to be able to add user-defined templates and
patterns, such as specifying timing within the condition component. Finally,
to aid interoperability they ‘would like to be able to export to a format
where the formalised requirements are machine readable (e.g. parse tree)’.
Impact: Formalising the requirements made them more detailed and less
ambiguous; crucially much of the detail came from elicitation discussions, not
from existing documents. FRET captures the links between requirements, and
explanations of their intent (which was often more detailed than what already
existed). These two things mean that the FRET requirements are a valuable
development artefact. We are currently pursuing Phase 2 of our methodology
(Fig. 1), in which we will assess the impact of the fretish requirements on
verification.
We believe that FRET can scale to larger requirements sets, with the parent-
child relationship providing a grouping function. However, for large sets of
requirements it might be necessary to modularise or refactor the requirements
so that they are easier to maintain. We are currently examining how fretish
requirements can be refactored for our use case.
## 6 Conclusions and Future Work
This paper provides an experience report of requirements elicitation and
formalisation of an aircraft engine controller in FRET. Our industrial partner
provided a set of natural-language requirements, test cases, and a Simulink
diagram. In close collaboration with our industrial partner, we clarified
ambiguous text in the requirements and test cases. This was essential, as we
had originally misunderstood some of the text. This iterative process produced
a set of detailed fretish requirements that we, and our industrial partner,
are confident correspond to the intent of the natural-language requirements.
The fretish requirements are now ready for use in formal verification
activities.
During this work we identified improvements that could be made to FRET, which
we plan to investigate in future work. First, our fretish requirements contain
quite a lot of repetition, so if changes were needed we often had to make the
change manually in several places. This was very time-consuming, so we propose
adding automatic requirement refactoring. Second, we plan to investigate how
to introduce globally-declared variable types. This would improve the
readability of requirements; clarifying what operations are valid for a
particular variable, while encapsulating definitions that might change in the
future. This could be made optional, to retain the ability to write very
abstract initial requirements. Finally, we would like to improve the
interoperability of FRET with other formal verification tools. For example,
adding a translator to the input language of a theorem prover to avoid the
state-explosion faced by model checkers (like Kind2, which is used to verify
CoCoSpec contracts); or outputting the requirement to a parse tree, as
suggested by our industrial partner.
## References
* [1] R.-J. Back and J. von Wright. Refinement Calculus: A Systematic Introduction. Springer, 1998.
* [2] R. Barbosa, S. Basagiannis, G. Giantamidis, H. Becker, E. Ferrari, J. Jahic, A. Kanak, M. L. Esnaola, V. Orani, D. Pereira, et al. The VALU3S ECSEL Project: Verification and Validation of Automated Systems Safety and Security. In Euromicro Conference on Digital System Design, pages 352–359. IEEE, 2020.
* [3] H. Bourbouh, M. Farrell, A. Mavridou, I. Sljivo, G. Brat, L. A. Dennis, and M. Fisher. Integrating Formal Verification and Assurance: An Inspection Rover Case Study. In NASA Formal Methods Symposium, pages 53–71. Springer, 2021.
* [4] H. Bourbouh, P.-L. Garoche, T. Loquen, É. Noulard, and C. Pagetti. CoCoSim, a code generation framework for control/command applications An overview of CoCoSim for multi-periodic discrete Simulink models. In European Congress on Embedded Real Time Software and Systems, 2020.
* [5] R. Darimont and A. Van Lamsweerde. Formal Refinement Patterns for Goal-Driven Requirements Elaboration. ACM SIGSOFT Software Engineering Notes, 21(6):179–190, 1996.
* [6] A. Dutle, C. Muñoz, E. Conrad, A. Goodloe, I. Perez, S. Balachandran, D. Giannakopoulou, A. Mavridou, T. Pressburger, et al. From Requirements to Autonomous Flight: An Overview of the Monitoring ICAROUS Project. In Workshop on Formal Methods for Autonomous Systems, pages 23–30. EPTCS, 2020.
* [7] D. Giannakopoulou, A. Mavridou, J. Rhein, T. Pressburger, J. Schumann, and N. Shi. Formal Requirements Elicitation with FRET. In International Conference on Requirements Engineering: Foundation for Software Quality, 2020.
* [8] D. Giannakopoulou, T. Pressburger, A. Mavridou, and J. Schumann. Automated formalization of structured natural language requirements. Information and Software Technology, 137, 2021.
* [9] M. Luckcuck, M. Farrell, O. Sheridan, and R. Monahan. A Methodology for Developing a Verifiable Aircraft Engine Controller from Formal Requirements. In IEEE Aerospace Conference, 2022.
* [10] A. Mavridou, H. Bourbouh, P. L. Garoche, D. Giannakopoulou, T. Pessburger, and J. Schumann. Bridging the Gap Between Requirements and Simulink Model Analysis. In International Conference on Requirements Engineering: Foundation for Software Quality, 2020.
* [11] A. Mavridou, H. Bourbouh, D. Giannakopoulou, T. Pressburger, M. Hejase, P.-L. Garoche, and J. Schumann. The Ten Lockheed Martin Cyber-Physical Challenges: Formalized, Analyzed, and Explained. In International Requirements Engineering Conference, pages 300–310. IEEE, 2020.
* [12] I. Perez, F. Dedden, and A. Goodloe. Copilot 3. Technical report, NASA/TM-2020-220587, National Aeronautics and Space Administration, 2020.
* [13] I. Postlethwaite, R. Samar, B. Choi, and D. Gu. A Digital Multimode H$\infty$ Controller for the Spey Turbofan Engine. In European Control Conference, 1995.
* [14] R. Samar and I. Postlethwaite. Design and Implementation of a Digital Multimode H$\infty$ Controller for the Spey Turbofan Engine. Journal of Dynamic Systems, Measurement, and Control, 132(1):011010, 2010.
* [15] P. Zave and M. Jackson. Four Dark Corners of Requirements Engineering. ACM Transactions on Software Engineering and Methodology (TOSEM), 6(1):1–30, 1997.
## Appendix Appendix A Simulink Diagram
This appendix contains a high-level view of the Simulink diagram of the
aircraft engine controller that we described in Sect. 2.
Figure 3: High-level view of the Simulink diagram that our industrial partner
constructed to model the controller for this use case.
## Appendix Appendix B Test Cases
This appendix contains all of the test cases for the aircraft engine
controller, provided by our industry partners. There are 20 test cases in
total and these are shown in Tables 7 – 10.
Test Case ID | Req. ID | Description
---|---|---
UC5_TC_1 | UC5_R_1 | Preconditions: Aircraft is in operating mode M and sensor S value deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting control objectives (settling time, overshoot, steady state error)
UC5_TC_2 | UC5_R_1 | Preconditions: Aircraft is in operating mode M and sensor S value is not available (sensor is out of order)
Input conditions / steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting control objectives (settling time, overshoot, steady state error)
UC5_TC_3 | UC5_R_2 | Preconditions: Aircraft is in operating mode M and sensor S value deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and
perturbations in nonpilot input cause it to change to V2
Expected results: Observed aircraft thrust returns to value V1, respecting
control objectives (settling time, overshoot, steady state error)
UC5_TC_4 | UC5_R_2 | Preconditions: Aircraft is in operating mode M and sensor S value is not available (sensor is out of order)
Input conditions / steps: Observed aircraft thrust is at value V1 and
perturbations in nonpilot input cause it to change to V2
Expected results: Observed aircraft thrust returns to value V1, respecting
control objectives (settling time, overshoot, steady state error)
UC5_TC_5 | UC5_R_3 | Preconditions: Aircraft is in operating mode M and sensor S value deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting operating limit objectives (e.g. upper limit in shaft speed)
UC5_TC_6 | UC5_R_3 | Preconditions: Aircraft is in operating mode M and sensor S value is not available (sensor is out of order)
Input conditions / steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting operating limit objectives (e.g. upper limit in shaft speed)
Table 7: The test cases for the aircraft engine controller, provided by our industrial partner corresponding to requirements UC5_R_1, UC5_R_2 and UC5_R_3. Test Case ID | Req. ID | Description
---|---|---
UC5_TC_7 | UC5_R_4 | Preconditions: Aircraft is in operating mode M and sensor S value deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and
perturbations in nonpilot input cause it to change to V2
Expected results: Observed aircraft thrust returns to value V1, respecting
operating limit objectives (e.g. upper limit in shaft speed)
UC5_TC_8 | UC5_R_4 | Preconditions: Aircraft is in operating mode M and sensor S value is not available (sensor is out of order)
Input conditions / steps: Observed aircraft thrust is at value V1 and
perturbations in nonpilot input cause it to change to V2
Expected results: Observed aircraft thrust returns to value V1, respecting
operating limit objectives (e.g. upper limit in shaft speed)
UC5_TC_9 | UC5_R_5 | Preconditions: Aircraft is in operating mode M and system parameter P deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting control objectives (settling time, overshoot, steady state error)
UC5_TC_10 | UC5_R_6 | Preconditions: Aircraft is in operating mode M and system parameter P deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and
perturbations in nonpilot input cause it to change to V2
Expected results: Observed aircraft thrust returns to value V1, respecting
control objectives (settling time, overshoot, steady state error)
UC5_TC_11 | UC5_R_7 | Preconditions: Aircraft is in operating mode M and system parameter P deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and pilot
input changes from A1 to A2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting operating limit objectives (e.g. upper limit in shaft speed)
Table 8: The test cases for the aircraft engine controller, provided by our industrial partner corresponding to requirements UC5_R_4, UC5_R_5, UC5_R_6 and UC5_R_7. Test Case ID | Req. ID | Description
---|---|---
UC5_TC_12 | UC5_R_8 | Preconditions: Aircraft is in operating mode M and system parameter P deviates at most +/ R % from nominal value
Input conditions / steps: Observed aircraft thrust is at value V1 and
perturbations in nonpilot input cause it to change to V2
Expected results: Observed aircraft thrust returns to value V1, respecting
operating limit objectives (e.g. upper limit in shaft speed)
UC5_TC_13 | UC5_R_9 | Preconditions: Aircraft is in operating mode M
Input conditions / steps: Observed aircraft thrust is at value V1, pilot input
changes from A1 to A2, and outside air pressure abruptly changes from P1 to P2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting control objectives (settling time, overshoot, steady state error)
UC5_TC_14 | UC5_R_10 | Preconditions: Aircraft is in operating mode M
Input conditions / steps: Observed aircraft thrust is at value V1, small
perturbations in nonpilot input cause it to change to V2, and outside air
pressure abruptly changes from P1 to P2
Expected results: Observed aircraft thrust returns to value V1, respecting
control objectives (settling time, overshoot, steady state error)
UC5_TC_15 | UC5_R_11 | Preconditions: Aircraft is in operating mode M
Input conditions / steps: Observed aircraft thrust is at value V1, pilot input
changes from A1 to A2, and outside air pressure abruptly changes from P1 to P2
Expected results: Observed aircraft thrust changes and settles to value V2,
respecting operating limit objectives (e.g. upper limit in shaft speed)
UC5_TC_16 | UC5_R_12 | Preconditions: Aircraft is in operating mode M
Input conditions / steps: Observed aircraft thrust is at value V1, small
perturbations in nonpilot input cause it to change to V2, and outside air
pressure abruptly changes from P1 to P2
Expected results: Observed aircraft thrust returns to value V1, respecting
operating limit objectives (e.g. upper limit in shaft speed)
Table 9: The test cases for the aircraft engine controller, provided by our industrial partner corresponding to requirements UC5_R_8, UC5_R_9, UC5_R_10, UC5_R_11 and UC5_R_12. Test Case ID | Req. ID | Description
---|---|---
UC5_TC_17 | UC5_R_13 | Preconditions: Aircraft is in nominal operating mode
Input conditions / steps: Pilot input changes from A1 to A2, causing surge /
stall avoidance indicator signal to be set
Expected results: Aircraft switches to surge / stall prevention operating mode
UC5_TC_18 | UC5_R_13 | Preconditions: Aircraft is in surge / stall prevention operating mode
Input conditions / steps: Pilot input changes from A1 to A2, causing surge /
stall avoidance indicator signal to be cleared
Expected results: Aircraft switches to nominal operating mode
UC5_TC_19 | UC5_R_14 | Preconditions: Aircraft is in nominal operating mode
Input conditions / steps: Perturbations in nonpilot input cause surge / stall
avoidance indicator signal to be set
Expected results: Aircraft switches to surge / stall prevention operating mode
UC5_TC_20 | UC5_R_14 | Preconditions: Aircraft is in surge / stall prevention operating mode
Input conditions / steps: Perturbations in nonpilot input cause surge / stall
avoidance indicator signal to be cleared
Expected results: Aircraft switches to nominal operating mode
Table 10: The test cases for the aircraft engine controller, provided by our
industrial partner corresponding to requirements UC5_R_13, and UC5_R_14.
## Appendix Appendix C Child Requirements
This appendix contains the complete set of fretish child requirements for the
aircraft engine controller use case. These requirements are contained in
Tables 11 –15.
ID | Parent | FRETISH
---|---|---
UC5_R_1.1 | UC5_R_1 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null) & (pilotInput $=>$ setThrust = V2) & (observedThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V2)
UC5_R_1.2 | UC5_R_1 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (pilotInput $=>$ setThrust = V2) & (observedThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V2)
UC5_R_1.3 | UC5_R_1 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (pilotInput $=>$ setThrust = V2)& (observedThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateErrorMax) & (observedThrust = V2)
UC5_R_2.1 | UC5_R_2 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (!pilotInput $=>$ setThrust = V1)& (observedThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V1)
UC5_R_2.2 | UC5_R_2 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (!pilotInput $=>$ setThrust = V1)& (observedThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V1)
UC5_R_2.3 | UC5_R_2 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (!pilotInput $=>$ setThrust = V1)& (observedThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateError Max) & (observedThrust = V1)
Table 11: Child requirements for UC5_R_1 and UC5_R_2. ID | Parent | FRETISH
---|---|---
UC5_R_3.1 | UC5_R_3 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null) & (observedThrust = V1))& (pilotInput $=>$ setThrust = V2) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V2)
UC5_R_4.1 | UC5_R_4 | when (diff(r(i),y(i)) $>$ E) if((sensorValue(S) $>$ nominalValue + R) $|$ (sensorValue(S) $<$ nominalValue - R) $|$ (sensorValue(S) = null)& (!pilotInput $=>$ setThrust = V1) & (observedThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V1)
UC5_R_5.1 | UC5_R_5 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V2)
UC5_R_5.2 | UC5_R_5 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V2)
UC5_R_5.3 | UC5_R_5 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateErrorMax) & (observedThrust = V2)
Table 12: Child requirements for UC5_R_3, UC5_R_4 and UC5_R_5. ID | Parent | FRETISH
---|---|---
UC5_R_6.1 | UC5_R_6 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V1)
UC5_R_6.2 | UC5_R_6 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V1)
UC5_R_6.3 | UC5_R_6 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateErrorMax) & (observedThrust = V1)
UC5_R_7.1 | UC5_R_7 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V2)
UC5_R_8.1 | UC5_R_8 | when (diff(r(i),y(i)) $>$ E) if((systemParameter(P) $>$ nominalValue + R) $|$ (systemParameter(P) $<$ nominalValue - R) $|$ (systemParameter(P) = null)& (!pilotInput $=>$ setThrust = V1)& (observedThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V1)
Table 13: Child requirements for UC5_R_6, UC5_R_7 and UC5_R_8. ID | Parent | FRETISH
---|---|---
UC5_R_9.1 | UC5_R_9 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V2)
UC5_R_9.2 | UC5_R_9 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V2)
UC5_R_9.3 | UC5_R_9 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateErrorMax) & (observedThrust = V2)
UC5_R_10.1 | UC5_R_10 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (settlingTime $>=$ 0) & (settlingTime $<=$ settlingTimeMax) & (observedThrust = V1)
UC5_R_10.2 | UC5_R_10 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (overshoot $>=$ 0) & (overshoot $<=$ overshootMax) & (observedThrust = V1)
UC5_R_10.3 | UC5_R_10 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (steadyStateError $>=$ 0) & (steadyStateError $<=$ steadyStateErrorMax) & (observedThrust = V1)
Table 14: Child requirements for UC5_R_9 and UC5_R_10. ID | Parent | FRETISH
---|---|---
UC5_R_11.1 | UC5_R_11 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V1)& (pilotInput $=>$ setThrust = V2)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V2)
UC5_R_12.1 | UC5_R_12 | when (diff(r(i),y(i)) $>$ E) if ((outsideAirPressure(T1) != outsideAirPressure(T2) & (diff(t2,t1) $<$ threshold) & (abs(outsideAirPressure (T1) - outsideAirPressure(T2)) $>$ pressureThreshold) & (observedThrust = V2)& (!pilotInput $=>$ setThrust = V1)) Controller shall until (diff(r(i),y(i)) $<$ e) satisfy (shaftSpeed $>=$ operatingLowerBound) & (shaftSpeed $<=$ operatingUpperBound) & (observedThrust = V1)
UC5_R_13.1 | UC5_R_13 | in nominal mode when (diff(setNL, observedNL) $>$ NLmax) if (pilotInput $=>$ surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $<$ NLmin) satisfy (changeMode(surgeStallPrevention))
UC5_R_13.2 | UC5_R_13 | in surgeStallPrevention mode when (diff(setNL, observedNL) $<$ NLmax) if (pilotInput $=>$ !surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $>$ NLmin) satisfy (changeMode(nominal))
UC5_R_14.1 | UC5_R_14 | in nominal mode when (diff(setNL, observedNL) $>$ NLmax) if (!pilotInput $=>$ surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $<$ NLmin) satisfy (changeMode(surgeStallPrevention))
UC5_R_14.2 | UC5_R_14 | in surgeStallPrevention mode when (diff(setNL, observedNL) $<$ NLmax) if (!pilotInput $=>$ !surgeStallAvoidance) Controller shall until (diff(setNL, observedNL) $>$ NLmin) satisfy (changeMode(nominal))
Table 15: Child requirements for UC5_R_11, UC5_R_12, UC5_R_13 and UC5_R_14.
|
# A portrait of the rotation of Ultra-Cool Dwarfs revealed by TESS
D. O. Fontinele Departamento de Física Teórica e Experimental, Universidade
Federal do Rio Grande do Norte, Campus Universitário, Natal, RN, 59072-970,
Brazil P. D. S. de Lima Departamento de Física Teórica e Experimental,
Universidade Federal do Rio Grande do Norte, Campus Universitário, Natal, RN,
59072-970, Brazil School of Physics, Trinity College Dublin, Dublin 2,
Ireland Y. S. Messias Departamento de Física Teórica e Experimental,
Universidade Federal do Rio Grande do Norte, Campus Universitário, Natal, RN,
59072-970, Brazil Université de Montréal, Département de Physique, IREX,
Montréal, QC H3C 3J7, Canada R. L. Gomes Departamento de Física Teórica e
Experimental, Universidade Federal do Rio Grande do Norte, Campus
Universitário, Natal, RN, 59072-970, Brazil Université de Montréal,
Département de Physique, IREX, Montréal, QC H3C 3J7, Canada C. E. Ferreira
Lopes Instituto de Astronomía y Ciencias Planetarias, Universidad de Atacama,
Copayapu 485, Copiapó, Chile Millennium Institute of Astrophysics, Nuncio
Monseñor Sotero Sanz 100, Of. 104, 7500000 Providencia, Santiago, Chile B. L.
Canto Martins Departamento de Física Teórica e Experimental, Universidade
Federal do Rio Grande do Norte, Campus Universitário, Natal, RN, 59072-970,
Brazil INAF - Osservatorio Astrofisico di Arcetri, Largo Enrico Fermi 5,
50125 Firenze, Italy I. C. Leão Departamento de Física Teórica e
Experimental, Universidade Federal do Rio Grande do Norte, Campus
Universitário, Natal, RN, 59072-970, Brazil J. M. DE Araújo Departamento de
Física Teórica e Experimental, Universidade Federal do Rio Grande do Norte,
Campus Universitário, Natal, RN, 59072-970, Brazil E. Janot Pacheco Instituto
de Astronomia, Geofísica e Ciências Atmosféricas, Universidade de São Paulo,
05509-090, São Paulo, SP, Brazil Observatoire de la Côte d’Azur, Laboratoire
Lagrange, Bâtiment Fizeau, Parc Valrose, 06108 Nice, France J. R. De Medeiros
Departamento de Física Teórica e Experimental, Universidade Federal do Rio
Grande do Norte, Campus Universitário, Natal, RN, 59072-970, Brazil
(Received December, 2023; Revised March, 2024; Accepted May)
###### Abstract
This study presents the results of a search for rotation signature in 250 Gaia
DR3 Ultra-Cool Dwarfs (UCDs) with TESS light curves. We identified 71 targets
with unambiguous periodicities, of which 61 present rotation signatures and a
single source behavior, with periods between 0.133 and 5.81 days. Five UCDs
show double-dip features, namely variations with two periods, one
approximately double or half the other. The remaining ten UCDs with
unambiguous variability present a likely non-single behavior. We also found 20
UCDs showing complex behavior in their light curves, with noticeable
fluctuations and irregular structure, with a few exhibiting apparent changes
in their temporal structure. The remaining 159 targets show noisy light curves
corresponding to low-amplitude signals, whose temporal variation cannot be
easily identified. The distribution of the UCDs with rotation signature in the
CMD diagram points to a lack of rotating objects within about
$11.5<M_{G}<12.5$ and $G-G_{RP}<1.5$ separating them into two regimes, one
mainly composed of less massive late-M stars with $P_{rot}\geq 1.0$ d, and
another mainly composed of more massive early-M stars with $P_{rot}<1.0$,d. It
is important to emphasize that by separating stars into age intervals, one
observes that UCDs with $P_{rot}\geq 1.0$ d tend to be located in regions of
younger objects, and, in contrast, those with $P_{rot}<1.0$ d are mainly
concentrated in regions of older objects. Whether these trends of stars
contrasting the sample separation is physical or produced by observational
biases is a question to be verified in future studies.
Low mass stars; Stellar rotation; Brown dwarfs; Stellar photometry; Space
observatories; Stellar astronomy
††journal: ApJ
## 1 Introduction
Ultra-Cool Dwarfs (UCDs) are defined as objects with spectral types M7 or
later, including L, T, and Y types, of stellar and substellar nature. These
objects are characterized by intense and broad potassium lines in the optical,
and water, methane, and ammonia absorption bands in the near-IR (e.g.,
Kirkpatrick et al., 1999; Burgasser et al., 2002; Burningham et al., 2008).
Among their most interesting characteristics is the formation of atmospheric
cloud-like structures, composed of iron and silicates for L dwarfs, chlorides
and sulfides for T and early-type Y dwarfs, and water clouds in the coolest Y
dwarfs (e.g., Saumon et al., 2000; Zahnle & Marley, 2014; Suárez & Metchev,
2023). Another fundamental characteristic of the UCDs concerns their well-
established variability, characterized by brightness changes caused by large-
scale atmospheric structures, such as spots or longitudinal bands, which can
provide fundamental information on the physical processes operating at their
atmospheres (e.g., Artigau et al., 2009; Radigan et al., 2014; Buenzli et al.,
2014; Metchev et al., 2015).
To date, the literature reports a growing list of studies on the photometric
variability of UCDs, based on observations from ground-based telescopes (e.g.,
Tinney & Tolley, 1999; Koen et al., 2005; Harding et al., 2013; Radigan et
al., 2014; Miles-Páez et al., 2017), and from space (e.g., Buenzli et al.,
2014; Metchev et al., 2015; Cushing et al., 2016; Tannock et al., 2021; Miles-
Páez et al., 2023; Vos et al., 2022; Petrucci et al., 2024). These studies
revealed essential characteristics on the behavior of the periodicities,
including a clustering of shortest rotation periods near 1 hr (Tannock et al.,
2021) and a lower envelope of rotation periodicities that runs from 2 h at
spectral type M8 down to 1 h at spectral type T (Miles-Páez et al., 2023).
Nevertheless, the astrophysical root cause for the variability in UCDs still
needs to be better understood. On the stellar side of the hydrogen-fusing
limit, starspots are the usual explanation for the observed variability (e.g.,
Rockenfeller et al., 2006; Irwin et al., 2011; Donati et al., 2023). For the
UCDs, the variability can have multiple origins, such as magnetic processes
(e.g., Hooten & Hall, 1990), condensated particles that form atmospheric
cloud-like structures (Tsuji et al., 1996), or even brightness changes due to
an atmosphere out of chemical equilibrium (Tremblin et al., 2015). Pineda et
al. (2017) pointed to evidence for a complex interplay between the starspots
driving the variability of UCDs and robust large-scale magnetospheric current
systems.
The NASA TESS space mission (Ricker et al., 2015), launched in 2018, is
producing differential photometric light curves (LCs) for hundreds of
thousands of stars. Primarily dedicated to the search for terrestrial planets
transiting nearby bright stars, the large number of observed targets combined
with the high quality of the acquired data is opening new horizons for the
study of a variety of astrophysical phenomena, including stellar rotation and
activity (e.g., Canto Martins et al., 2020; Barraza et al., 2022; Doyle et
al., 2020). Thanks to its high-quality photometry data and continuous
monitoring, TESS could also represent an optimal laboratory in the search for
rotation periodicities in UCDs.
Aiming to enlarge the current sample of Ultra-Cool Dwarfs with measured
rotation period, this study is dedicated to the search for rotation signatures
in sources identified as Ultra-Cool Dwarfs in the Gaia DR3 (Sarro et al.,
2023), observed by the TESS mission. This paper is part of a collaborative
effort to determine rotation periodicities for different families of stars
located throughout the HR Diagram, with LCs collected by the referred space
mission. The paper is organized as follows. Section 2 presents the stellar
sample and observational data set used in this study and the procedure used to
analyze the LCs. Section 3 provides the main results. A summary is presented
in Section 4.
## 2 The Ultra-Cool Dwarf Sample and Observational Data Analysis
The TESS primary mission plan is to survey almost the entire sky by monitoring
26 segments (or sectors) of $90^{\circ}\times 24^{\circ}$, each one with a
duration of 27 days at a time, with four on-board 10.5 cm telescopes in a red
(600–1000 nm) band-pass. A given sector is, therefore, revisited about every
two years. However, overlaps between sectors can increase the coverage, and
targets in the Continuous Viewing Zone are observed with almost unbroken
coverage for a year. The mission provides photometric time series at different
cadences, namely 2 and 30 minutes in Cycles 1 and 2, and 20 s, 2 minutes, and
10 minutes in Cycle 3, with a time baseline from 27 to 351 days, depending on
sector overlaps. While 2-minute cadence data, also known as Target Pixel (TP)
files, are available for a subset of targets, all CCDs, called full-frame
images (FFIs), are read out every 30 minutes (10 minutes in the extended
mission). Ten subsets of TESS targets were observed for multiple sectors, with
approximately 1–2 percent of targets located in the continuous viewing zone
during the primary mission (Barclay et al., 2018). Such an aspect makes these
targets invaluable for extracting rotation periodicities and analyzing the
persistence of stellar cycles (Ferreira Lopes et al., 2015).
For this study, searching for rotation signatures in UCDs and determining
their periodicities, we have used the sample of 7630 Gaia DR3 UCDs, given by
Sarro et al. (2023), as a starting point. For the definition of this sample,
those authors used the Gaia DR3 set of UCD candidates and complemented the
Gaia spectrophotometry with additional photometry to characterize the
candidates’ global properties. This procedure includes the determination of
the distances and their locus in the Gaia color-absolute magnitude diagram
(CAMD). More specifically, they present the typical (median) RP spectra in
spectral-type bins and compare them to UCD standards, ground-based high-
resolution spectra, and CAMDs, including Gaia and external photometry. Then,
that sample was compared with the TESS Input Catalogue, from which 7360
objects were identified out of the total sample of 7630. Of this latter list,
we found 250 targets with TESS 2-minute cadence LCs available on October 30th,
2023. On this date, the TESS data basis also reported observations of 674
targets in a 30-min cadence, 124 in a 10-min cadence, and one target in a
20-seg cadence; all these additional data are not suitable for our purpose.
The 2-min cadence data were downloaded from the FFI-TP-LC-DV Bulk Downloads
Page of the Mikulski Archive for Space
Telescopes111https://archive.stsci.edu/tess/bulk/downloads.html (STScI, 2018)
using the cURL scripts available for retrieving Pre-search Data Conditioned
Simple Aperture Photometry (PDCSAP) reduced light curves. Jenkins et al.
(2016) described the TESS science processing operations center (SPOC) pipeline
that produces the 2-min cadence light curves. The distribution of the spectral
types for the list of 250 UCDs with TESS LCs is displayed in Figure 1, which
ranges from M5 to L3. The referred spectral types were computed following the
calibration by Pecaut & Mamajek (2013). These targets are located in the
following star-forming regions, open clusters, or moving groups, as listed in
Table 6: Taurus (TAU), Carina-Near (CARN), Tucana-Horologium Association
(THA), $\epsilon$ Chamaeleontis Association (EPSC), AB Doradus moving group
(ABDMG), $\beta$ Pic Moving Group (BPMG), Upper Centaurus Lupus (UCL), Hyades
(HYA), TW Hya Association (TWA), Lower Centaurus Crux (LCC), Pleiades cluster
(PLE), Columba (COL), Argus (ARG), Coma Berenices(CBER), and Carina
Association (Car). The determination of membership to the referred star
clusters and moving groups were derived by Sarro et al. (2023), using the
BANYAN $\Sigma$ software tool (Gagné et al., 2018). Those authors have used
the Gaia DR3 UCD candidates, namely their celestial positions and proper
motions and parallaxes, as input to BANYAN $\Sigma$.
When required, additional processing on these LCs was performed manually to
avoid possible distortions in the signature of periodicities. Such processing
was explicitly the removal of outliers and correction of instrumental trends,
following the recipe by De Medeiros et al. (2013), Paz-Chinchón et al. (2015),
and Canto Martins et al. (2020). In short, flux measurements exceeding
3$\times$ the standard deviation of the light curve compared to the average
neighbor data were considered outliers and then removed from the time series.
The correction of instrumental trends consisted of detrending the light curve
of each sector from a polynomial fit of third order. In the analysis of the
post-processed LCs for the 250 UCDs with TESS observations, we have used the
same procedure applied by Canto Martins et al. (2020), which is based on a
manifold interactive platform built in the Interactive Data Language
(IDL222https://www.l3harrisgeospatial.com/Software-Technology/IDL). Readers
are referred to Canto Martins et al. (2020) and Barraza et al. (2022) for a
complete description of the referred manifold package. In short, our analysis
consisted of computing Fast Fourier Transforms (FFT) (e.g., Zhan et al.,
2019), Lomb-Scargle periodograms (e.g., Scargle, 1982; Horne & Baliunas, 1986;
Press & Rybicki, 1989), as well as wavelet maps and global spectra (Grossmann
& Morlet, 1984), of each light curve. Based on the known range of the rotation
periods of UCDs (e.g., Bailer-Jones, 2004; Tannock et al., 2021), we searched
for the main periodicities, namely peaks, computed from those methods. The
errors on the corresponding peak periods were estimated using Eq. (2) of Lamm
et al. (2004). The Lomb-Scargle method was also used to consider the false
alarm probability (FAP) of the periodicities, based on Eq. (22) of Horne &
Baliunas (1986), given by
$F=1-\left\\{1-exp[-(N_{0}/2)(1+xi^{-1})-1]\right\\}^{N_{i}}$. The formula
gives the periodogram power threshold, $F$, corresponding to a given FAP,
$\xi$, where $N_{0}$ is the number of observations of a light curve, $\xi$ is
the FAP, and $N_{i}$ is the number of independent frequencies, namely
$N_{i}=-6.362+1.193N_{0}+0.00098N_{0}^{2}$, according to Eq. (13) of Horne &
Baliunas (1986). In the present work, only peaks with FAP less than 1%,
corresponding to significance levels greater than 99%, were taken into
consideration. Additionally, wavelet maps were used to examine the persistence
of the main signals within the time span of TESS observations of a given
target in different sectors, and phase diagrams of the possible periodicities
were computed to inspect the global morphology of each signal.
After obtaining the FFT, Lomb–Scargle, wavelet analyses, and phase diagrams
results, we visually inspected each light curve to identify the present
modulation features, adopting the procedure applied by Canto Martins et al.
(2020).
Specifically, we separated the targets into three groups: (i) those with a
typical rotational modulation in the LC, namely a signature characterized by
semi-regular flux variability that used to be multi-sinusoidal; (ii) those
with an ambiguous, dubious, or complex variability profile, namely those
targets with LCs displaying an asymmetric shape in the flux variation, and
(iii) targets with noisy LCs. From the sample of 250 UCDs with bonafide TESS
LCs, listed in Table A of the Online Material, we have identified 54 UCDs
presenting unambiguous variability, with variability amplitudes ranging from
0.6 to 40.9 percent, for which we have computed the associated periodicities.
This list of 54 UCDs is presented in Table 1, which also contains the referred
periodicities. Figure 2 displays three examples of LCs with typical rotation
signatures identified in our sample of UCDs, with the corresponding phase
folded LC, FFT, Lomb–Scargle periodograms, and wavelet maps. Figures following
the same design as Figure 2 for all the UCDs with unambiguous variability are
provided in the figure set given in the Online Material. Among the UCDs with
evident variability, we identified 20 targets presenting a complex temporal
structure in their LCs. The referred LCs are also provided in the figure set
given in the Online Material.
### 2.1 A search for periodicities in Gaia DR3 UCDs with supposedly noisy LCs
Our manifold analysis has initially identified 176 Gaia DR3 UCDs with LCs
presenting supposedly noisy signatures, an aspect assumed particularly for LCs
with a low-amplitude signal. Indeed, as underlined by different studies (e.g.,
Canto Martins et al., 2020; Gilliland et al., 2011), a noisy behavior is a
complex combination of different factors, including, for example, instrumental
noise contributions and data reduction processes. Some UCDs classified as
noisy LCs may have been classified this way because of data reduction issues.
Considering that different studies point to a clustering of shortest rotation
periods of UCDs near 1 to 2 hr (e.g., Miles-Páez et al., 2023; Tannock et al.,
2021), we revisited the sample of 176 Gaia DR3 targets with LCs previously
classified as noisy in the search for short periods, specifically for periods
shorter than 1 d. First, we reanalyzed each LC using the Wavelet method to
search for persistent temporal structures on time scales shorter than 1 d, all
along the considered LC time span. Such a procedure revealed 17 targets with
potential low amplitude variability ranging from 0.22 to 21 percent, a
signature also clearly visible in the related phase folded light curve. For
these targets, it was possible to compute Wavelet and LS periodicities, listed
in Table 2, with the respective variability amplitude. Table 2 lists Sectors
where it was possible to extract the period variability and Sectors with
negligible or absent modulation. To reinforce the presence of variability with
potential physical meaning, not reflecting instrumental systematics, we have
analyzed the LCs of these 17 UCDs using the W Transform (Wang, 2021, 2022),
which computes a time-frequency as the Wavelet Transform, but with a
significant difference. The W Transform uses a nonstationary window function
to concentrate the spectral energy in the region of a dominant frequency and
to remove the singularity problem related to the zero frequency in the time-
frequency spectrum. Indeed, whereas the standard Wavelet transform shifts the
energy of a signal toward a higher frequency, the W transform centralizes the
referred energy around the dominant frequency. The periodicity values computed
from the W Transfom are also listed in Table 2, which reinforces those
obtained from Wavelet and LS methods. Figures following the same design as
Figure 2 for these 17 UCDs with likely rotation signatures and low amplitude
variability are provided in the figure set given in the Online Material.
### 2.2 Cleaning the sample of UCDs with modulation
We have checked all the stars with unambiguous periodicities to verify if the
variability signals obtained from their LCs were associated with the
considered UCD. Indeed, given the TESS large plate scale of 21′′ pixel-1,
analysis of signals obtained from the LCs of this mission should be made with
caution due to potential contamination by blended LCs, which can induce an
interpretation of the observed variability to a wrong source (Mullally et al.,
2022; Higgins & Bell, 2023; Pedersen & Bell, 2023). To assess the potential
impact of contamination by nearby sources, we analyzed the crowding metric
(CROWD) calculated by the SPOC pipeline (Caldwell et al., 2020), which
indicates the fraction of the light in the TESS aperture that comes from the
considered target given the positions and amplitudes of stars in the TESS
Input Catalog. A CROWD value near 1.0 indicates a potentially isolated star,
while lower values indicate a potential crowding from neighbors. This analysis
found 42 UCDs with a CROWD parameter larger than 0.7 within the total sample
of 71 UCDs with modulation listed in Tables 1 and 2. These tables also give
the CROWD parameters, with the listed values corresponding to an average for
stars observed in more than one sector. Hence, the variability amplitude of
UCDs with a CROWD parameter smaller than 0.7 should be taken cautiously since
these objects could be significantly affected by the flux contamination of
nearby targets.
## 3 The Gaia DR3 UCDs with TESS rotation periods
Among 250 DR3 Ultra-Cool Dwarfs with TESS bonafide available 2-min cadence
LCs, we have identified 71 targets presenting unambiguous periodicities,
composing two sub-samples. The first one is composed of 54 UCDs with high
amplitude modulation, showing persistent temporal structures on time scales
ranging from 0.133 to 5.81 days, and the second one is composed of 17 UCDs
with modulation signatures despite the low amplitude of their LCs, also with
persistent temporal structures on time scales shorter than one day. However,
an important aspect should be underlined: a closer look into the whole of the
LCs of eight of these 17 UCDs with low amplitude, namely TIC 448590406, TIC
299007548, TIC 356632694, TIC 371080717, TIC 275133493, TIC 201688405, TIC
302396416, and TIC 342967337, reveals that, in some epochs (TESS Sectors),
they show a quasi-sinusoidal modulation, while in other epochs, there is a
negligible or undetectable sign of variability. Tables 1 and 2 give the
periodicities for the referred samples of 54 and 17 UCDs, respectively,
including associated error in the period ($eP_{rot}$), effective time span
($t_{SPAN}$) of each LC (the total time span subtracted by the duration of
eventual gaps), the effective number of cycles of the rotational modulation
(defined as $N_{Cycle}=t_{SPAN}/P_{rot}$), and the TESS observation sectors.
We have also identified 20 DR3 UCDs with dubious or ambiguous variability
signatures in their LCs. The Dubious variability corresponds to targets
showing potential rotation signature but whose period could not be
disentangled among two or more possibilities (from periodogram peaks and
wavelet maps) and stars with $N_{Cycle}<$ 3\. Ambiguous variability
corresponds to objects showing visually large-amplitude variations with a very
irregular or complex structure in the LCs, noticeable fluctuations, or an
insufficient time span to correctly identify the nature of their variability
signature (see Section 2.5 of Canto Martins et al. (2020)). The referred UCDs
are listed in Table 2. A few of these targets, TIC 125882881, TIC 367089503,
TIC 58285782, and TIC 58229208, present apparent changes in their temporal
structure in the same TESS sector.
Tables 1 and 2 show a relatively high number of UCDs with short-period
measurements, 40 percent out of the 71 objects with computed periodicities. We
carefully analyzed if some of those short periods could be associated with
binarity. Then, we visually inspected the short-period LCs, searching for
eclipsing binary signatures. However, no sign of binarity was found in the
LCs. In this sense, we have evaluated the Gaia RUWE parameter (Gaia
Renormalized Unit Weight Error) for the entire list of 71 UCDs, which may be
an indicator of multiplicity (Lindegren et al., 2021). A RUWE value of around
1.0 is expected for a single-star astrometric solution. In contrast, a value
significantly greater than 1.0, typically greater than 1.4, could indicate a
non-single source or problematic for the astrometric solution (e.g., Kervella
et al., 2022). Ten targets with RUWE greater than 1.4 were identified, six
from Table 1 (TIC 17518894, TIC 268325250, TIC 56002511, TIC 58229181, TIC
56624850, and TIC 902237947), and four from Table 2 (TIC 356632694, TIC
281668854, TIC 388682292, and TIC 448590406). The star TIC 58229181 is listed
in the literature as an EB, from TESS observations (Prša et al., 2022),
whereas TIC 56624850 is classified as a circumbinary protoplanetary disk
system (Czekala et al., 2019). In addition, among the UCDs with dubious or
ambiguous variability signatures in their LCs, 18 present a RUWE greater than
1.4.
Then, for the present study, namely, the search for rotation signatures in
sources identified as UCDs in the Gaia DR3, observed by the TESS mission, we
consider as potential bonafide rotation signatures only those UCDs with RUWE
less than 1.4. This criterion amounts to 48 and 13 objects from the lists
given in Tables 1 and 2, respectively. The distribution of the computed
rotation periods for the corresponding sample of 61 UCDs with RUWE less than
1.4, namely those expected to be likely single-rotating objects, is displayed
in Figure 3. As it arises, the periodicities for these UCDs range from 0.133 d
to 5.81 d for the sample listed in Table 1 and from 0.116 d to 0.71 d for the
UCDs listed in Table 2, the latter with low LC amplitudes.
Five of the UCDs listed in Table 1 have two period values (TIC 150094175, TIC
20305594, TIC 359892714, TIC 58638214, and TIC 96680763), four of which
present a second period at half the main period (TIC 20305594, TIC 359892714,
TIC 58638214, and TIC96680763). These cases can be interpreted as double-dip
signatures (Basri, 2018; Basri & Nguyen, 2018), typically produced by two
groups of heterogeneities on the stellar surface, such as spots, clouds, or
large wave bands (e.g., Apai et al., 2017), with phases shifted by about 90°.
Those phase shifts may evolve along time, producing beating effects in the
light curves or a variation of the peak amplitudes in the periodograms that
could be caused by differential rotation (e.g., Apai et al., 2017; Basri,
2018; Basri & Nguyen, 2018). In wavelet maps, a distinctive pattern often
includes two dominant features over time, where one feature’s period is either
double or half of the other. Frequently, the longer period corresponds to the
stellar rotation period, while the shorter one results from the combination of
two semi-sinusoids linked to the double-dip signature (e.g., Canto Martins et
al., 2020).
A no less relevant result concerns the high number of 159 targets showing
noisy behavior in their LCs, corresponding to 64 percent of the original
sample of 250 targets. Although those targets present typically low-amplitude
signals whose physical periodicities cannot be easily identified, they can
hide essential information. Typically, a noisy signature is a complex
combination of instrumental noise contributions (related, for instance, with
Poisson statistics and readout noise) plus a relevant contribution of
intrinsic stellar noise, the Galactic position, light from neighboring stars,
and sky background contamination (e.g., Gilliland et al., 2011). When the LCs
considered in this work present a low-amplitude signal, we assumed them to be,
in principle, a noisy signature. The noisy behavior could reflect low activity
or long periodicities for some stars, particularly for targets with short
observational time spans. Table B in the Online Material gives the list of
UCDs supposedly presenting noisy LCs.
Miles-Páez et al. (2023) have analyzed the TESS LCs of a sample of 23 known
rapidly rotating UCDs, typically objects with a $V\sin i>30$ km/s to explore
their photometric variability. These authors used a Lomb-Scargle periodogram
in their study, applying a first-order correction by removing a 24-hour median
filter from the data and then a Gaussian Process regression. From such an
analysis, they claimed that significant periodicities in 18 targets are
compatible with rotation signature, with amplitudes ranging from 0.17 to 0.77
percent. For comparative purposes, even though only two UCDs from Miles-Páez
et al. (2023) are present in our Gaia DR3 UCDs whole list, we have then
applied our manifold procedure to the 21 targets of their sample with 2-min
cadence LCs available. The result of this analysis is presented in Table 4,
from where one observes that, for most cases, our variability periods are in
good agreement with those computed by Miles-Páez et al. (2023), except for
three targets (TIC 288506050, TIC 441000085, TIC 311188315), for which our
analysis points instead for a noisy behavior in their LCs. Further, it is
worth underlining that among the 21 UCDs listed in Table 4, ten have a RUWE
greater than 1.4. Recently, Petrucci et al. (2024) have claimed for rotation
periodicities for a sample of 87 UCDs from TESS LCs, based on a Lomb-Scargle
periodogram. Among these objects, eighty percent show $P_{rot}<1$ d. Four
objects from the referred sample (TIC 298907057, TIC 902237947, TIC 201688405,
and TIC 371080717) are in common with our study, showing a good agreement in
the corresponding $P_{rot}$ values.
The location of the Gaia DR3 UCDs with rotation periodicities and RUWE shorter
than 1.4, listed in Tables 1, 2, and 4, is shown in the color-magnitude
diagram (CMD) displayed in Figure 4, from where one observes an apparent trend
for a transition in the distribution of UCD rotation periods. In the left
panel UCDs with $P_{rot}\geq 1.0$ d tend to be located mainly in the regions
above $11.5<M_{G}<12.5$, whereas those with $P_{rot}<1.0$ d tend to populate
the region below this range. In addition, the CMD diagram in the referred
panel also appears to point to a dependence on the distribution of $P_{rot}$
with stellar mass, separated into two different regimes. The upper region,
populated mainly by UCDs with $P_{rot}\geq 1.0$ d, is composed of lower mass M
stars, whereas the lower region, mainly populated by UCDs with $P_{rot}<1.0$
d, is composed of larger mass M stars (Reylé, 2018). Thanks to the nature of
the present UCDs sample, which consists of objects located in stellar
associations with measured ages (Kenyon & Hartmann, 1995; Zuckerman et al.,
2006, 2011; Murphy et al., 2013; Bell et al., 2015; Pecaut & Mamajek, 2016),
the right panel in Figure 4 displays the same CMD. In this panel, stars are
separated into age intervals based on the values given in Tables 1 and 2, to
tentatively define regions where younger and older stars are expected to
dominate. The scenario emerging from the right panel in Figure 4 indicates an
interesting trend: UCDs with $P_{rot}\geq 1.0$ d tend to be located mainly in
the regions of younger objects, whereas those with $P_{rot}<1.0$ d tend to
populate the regions occupied by older objects.
Nevertheless, we should be cautious with the referred trend observed in the
CMD diagram, which could be affected by observational biases. The
incompleteness of the analyzed UCD sample and the instrumental sensitivity in
the observed wavelength band may contribute to the portrait observed in the
rotation period distribution (e.g., Canto Martins et al., 2023; Leão et al.,
2015). For instance, the UCD sample may be incomplete to periods longer than
24 hours because TESS is mainly sensitive to L0 dwarfs and, on the other side,
the lower period limits are also uncertain because most UCD campaigns have
been limited to $\sim$8–24-h cadences (Miles-Páez et al., 2023). Besides, the
region with an apparent lack of stars with rotation period measurements could
be populated by low-amplitude rotators below the TESS sensitivity. Despite
these potential biases, it is worth underlining that, based on the analysis of
the projected rotation velocity $V\sin i$ of UCDs, Mohanty & Basri (2003)
showed that this parameter tends to increase from mid-M to L types and that L
UCDs are probably rapidly rotating in general, with a lack of slowly rotating
in this latter spectral region.
## 4 Summary and Conclusions
Based on a manifold procedure, which considers the analysis of TESS LCs taking
into account Lomb-Scargle periodograms, Fast Fourier Transform (FFT), and
Wavelet Transform, we have analyzed a total of 250 Gaia DR3 UCDs presenting
public LCs, with short-cadence TESS observations in sectors 5 to 63. We
identified 48 UCDs with rotational modulation, presenting an unambiguous
rotation period. Among these UCDs, five show two periods, with four presenting
a second period that is approximately double or half of the rotation period.
This aspect may reveal the presence of differential rotation in the referred
targets. We have also identified 13 UCDs with typical modulation signatures,
presenting periodicities shorter than 1 d despite a low-amplitude behavior in
their LCs. Ten UCDs with unambiguous periodicities show a potential non-single
behavior, pointing to a dubious origin of the referred periodicities. The
present study also revealed 20 UCDs with ambiguous or dubious variability
signatures in their LCs, a few of them with LCs presenting apparent changes in
the temporal structure. The TESS light curves for the remaining 159 objects
show a noisy pattern with very low-amplitude signals.
Considering the whole sample of UCDs with rotation periods, this study shows
that 47.5 percent have a period shorter than 1 d. Another relevant result
arises from the CMD diagram, which seems to show a lack of rotating UCDs
within about $11.5<M_{G}<12.5$ and $G-G_{RP}<1.5$ that separates them into two
different regimes. One is mainly composed of less massive late-M stars with
$P_{rot}\geq 1.0$ d, and another regime is mainly composed of more massive
early-M stars with $P_{rot}<1.0$ d. Also, by separating the stars into age
intervals, we found that UCDs with $P_{rot}\geq 1.0$ d tend to be in regions
of younger objects, while those with $P_{rot}<1.0$ d are concentrated in
regions of older objects. Whether the apparent lack of stars contrasting the
sample separation is physical or produced by observational biases is a
question to be verified in future studies based on a more extensive and
homogeneous sample, as well as better statistics. If, on one side, the
signatures of rotation in UCDs are associated with multiple origins, including
magnetic processes, condensate particles that form atmospheric cloud-like
structures, or even brightness changes due to an atmosphere out of chemical
equilibrium, the objects with ambiguous variability or a noisy pattern appear
to reflect also the presence of different factors, such as low activity phases
or polar phenomena. In this context, the large sample of bonafide rapidly
rotating UCDs revealed in the present work represents a solid laboratory to
identify the cause of photometric variability in ultra UCDs through
spectroscopic analysis, in particular, to identify traces that can
differentiate between magnetically-induced star spots and inhomogeneous
atmospheric dust clouds. A first step in this direction should be the study of
their spectroscopic line profiles. If these targets are rapidly rotating,
their line profiles will be significantly Doppler-broadened. The large sample
of noisy UCDs also represents a relevant material for spectroscopic studies.
If such behavior reflects a low activity level and a slow rotation, these
objects will not show much line broadening. Identifying the root cause of
photometric variability in UCDs is also mandatory to understand the structural
stability of rapidly rotating Ultra Cool Dwarfs.
## 5 ACKNOWLEDGMENTS
Research activities of the Observational Astronomy Board at the Federal
University of Rio Grande do Norte (NAOS) are supported by continuous grants
from the Brazilian funding agency CNPq. This study was financed in part by the
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior $-$ Brasil (CAPES)
$-$ Finance Code 001. YSM and DOF acknowledge CAPES graduate fellowships, and
RLG acknowledges a CNPq PDE fellowship. PDSL and YSM acknowledge
UFRN/CAPES/Print fellowship, respectively, process 88887.838338/2023-00 and
88887.717782/2022-00. JMA acknowledges the CNPq Brazilian research agency for
funding (grant 311589/2021-9). BLCM, ICL, and JRM acknowledge CNPq research
fellowships. This work includes data collected by the TESS mission. The NASA
Explorer Program provides funding for the TESS mission. All the TESS data used
in this paper can be found in MAST:http://dx.doi.org/10.17909/fwdt-2x66
(catalog 10.17909/fwdt-2x66). We made use of data from the European Space
Agency (ESA) mission Gaia (https: //www.cosmos.esa.int/gaia), processed by the
Gaia Data Processing and Analysis Consortium (DPAC,
https://www.cosmos.esa.int/web/gaia/dpac/consortium). Funding for the DPAC has
been provided by national institutions, particularly the institutions
participating in the Gaia Multilateral Agreement. We want to thank the Referee
for carefully reading the manuscript and a number of suggestions that largely
improved the paper.
## References
* Apai et al. (2017) Apai, D., Karalidi, T., Marley, M. S., et al. 2017, Science, 357, 683. doi:10.1126/science.aam9848
* Artigau et al. (2009) Artigau, É., Bouchard, S., Doyon, R., et al. 2009, ApJ, 701, 1534. doi:10.1088/0004-637X/701/2/1534
* Bailer-Jones (2004) Bailer-Jones, C. A. L. 2004, A&A, 419, 703. doi:10.1051/0004-6361:20040965
* Barclay et al. (2018) Barclay, T., Pepper, J., & Quintana, E. V. 2018, ApJS, 239, 2. doi:10.3847/1538-4365/aae3e9
* Barraza et al. (2022) Barraza, L. F., Gomes, R. L., Messias, Y. S., et al. 2022, ApJ, 924, 117. doi:10.3847/1538-4357/ac3335
* Basri (2018) Basri, G. 2018, ApJ, 865, 142. doi:10.3847/1538-4357/aade45
* Basri & Nguyen (2018) Basri, G. & Nguyen, H. T. 2018, ApJ, 863, 190. doi:10.3847/1538-4357/aad3b6
* Bell et al. (2015) Bell, C. P. M., Mamajek, E. E., & Naylor, T. 2015, MNRAS, 454, 593. doi:10.1093/mnras/stv1981
* Buenzli et al. (2014) Buenzli, E., Apai, D., Radigan, J., et al. 2014, ApJ, 782, 77. doi:10.1088/0004-637X/782/2/77
* Burgasser et al. (2002) Burgasser, A. J., Kirkpatrick, J. D., Brown, M. E., et al. 2002, ApJ, 564, 421. doi:10.1086/324033
* Burningham et al. (2008) Burningham, B., Pinfield, D. J., Leggett, S. K., et al. 2008, MNRAS, 391, 320. doi:10.1111/j.1365-2966.2008.13885.
* Caldwell et al. (2020) Caldwell, D. A., Tenenbaum, P., Twicken, J. D., et al. 2020, Research Notes of the American Astronomical Society, 4, 201. doi:10.3847/2515-5172/abc9b3
* Canto Martins et al. (2020) Canto Martins, B. L., Gomes, R. L., Messias, Y. S., et al. 2020, ApJS, 250, 20. doi:10.3847/1538-4365/aba73f
* Canto Martins et al. (2023) Canto Martins, B. L., Messias, Y. S., Arruda Gonçalves, M. I., et al. 2023, Nature Astronomy, 7, 900. doi:10.1038/s41550-023-01976-0
* Cushing et al. (2016) Cushing, M. C., Hardegree-Ullman, K. K., Trucks, J. L., et al. 2016, ApJ, 823, 152. doi:10.3847/0004-637X/823/2/152
* Czekala et al. (2019) Czekala, I., Chiang, E., Andrews, S. M., et al. 2019, ApJ, 883, 22. doi:10.3847/1538-4357/ab287b
* De Medeiros et al. (2013) De Medeiros, J. R., Ferreira Lopes, C. E., Leão, I. C., et al. 2013, A&A, 555, A63. doi:10.1051/0004-6361/201219415
* Donati et al. (2023) Donati, J.-F., Lehmann, L. T., Cristofari, P. I., et al. 2023, MNRAS, 525, 2015. doi:10.1093/mnras/stad2301
* Doyle et al. (2020) Doyle, L., Ramsay, G., & Doyle, J. G. 2020, MNRAS, 494, 3596. doi:10.1093/mnras/staa923
* Ferreira Lopes et al. (2015) Ferreira Lopes, C. E., Leão, I. C., de Freitas, D. B., et al. 2015, A&A, 583, A134. doi:10.1051/0004-6361/201424900
* Gagné et al. (2018) Gagné, J., Mamajek, E. E., Malo, L., et al. 2018, ApJ, 856, 23. doi:10.3847/1538-4357/aaae09
* Gilliland et al. (2011) Gilliland, R. L., Chaplin, W. J., Dunham, E. W., et al. 2011, ApJS, 197, 6. doi:10.1088/0067-0049/197/1/6
* Grossmann & Morlet (1984) Grossmann, A. & Morlet, J. 1984, SIAM Journal on Mathematical Analysis, 15, 723. doi:10.1137/0515056
* Harding et al. (2013) Harding, L. K., Hallinan, G., Boyle, R. P., et al. 2013, ApJ, 779, 101. doi:10.1088/0004-637X/779/2/101
* Higgins & Bell (2023) Higgins, M. E. & Bell, K. J. 2023, AJ, 165, 141. doi:10.3847/1538-3881/acb20c
* Hooten & Hall (1990) Hooten, J. T. & Hall, D. S. 1990, ApJS, 74, 225. doi:10.1086/191497
* Horne & Baliunas (1986) Horne, J. H. & Baliunas, S. L. 1986, ApJ, 302, 757. doi:10.1086/164037
* Irwin et al. (2011) Irwin, J., Berta, Z. K., Burke, C. J., et al. 2011, ApJ, 727, 56. doi:10.1088/0004-637X/727/1/56
* Jenkins et al. (2016) Jenkins, J. M., Twicken, J. D., McCauliff, S., et al. 2016, Proc. SPIE, 9913, 99133E. doi:10.1117/12.2233418
* Kenyon & Hartmann (1995) Kenyon, S. J. & Hartmann, L. 1995, ApJS, 101, 117. doi:10.1086/192235
* Kervella et al. (2022) Kervella, P., Arenou, F., & Thévenin, F. 2022, A&A, 657, A7. doi:10.1051/0004-6361/202142146
* Kirkpatrick et al. (1999) Kirkpatrick, J. D., Reid, I. N., Liebert, J., et al. 1999, ApJ, 519, 802. doi:10.1086/307414
* Koen et al. (2005) Koen, C., Tanabé, T., Tamura, M., et al. 2005, MNRAS, 362, 727. doi:10.1111/j.1365-2966.2005.09280.x
* Lamm et al. (2004) Lamm, M. H., Bailer-Jones, C. A. L., Mundt, R., et al. 2004, A&A, 417, 557. doi:10.1051/0004-6361:20035588
* Leão et al. (2015) Leão, I. C., Pasquini, L., Ferreira Lopes, C. E., et al. 2015, A&A, 582, A85. doi:10.1051/0004-6361/201526085
* Lindegren et al. (2021) Lindegren, L., Klioner, S. A., Hernández, J., et al. 2021, A&A, 649, A2. doi:10.1051/0004-6361/202039709
* Metchev et al. (2015) Metchev, S. A., Heinze, A., Apai, D., et al. 2015, ApJ, 799, 154. doi:10.1088/0004-637X/799/2/154
* Miles-Páez et al. (2017) Miles-Páez, P. A., Pallé, E., & Zapatero Osorio, M. R. 2017, MNRAS, 472, 2297. doi:10.1093/mnras/stx2191
* Miles-Páez et al. (2023) Miles-Páez, P. A., Metchev, S. A., & George, B. 2023, MNRAS, 521, 952. doi:10.1093/mnras/stad273
* Mohanty & Basri (2003) Mohanty, S. & Basri, G. 2003, ApJ, 583, 451. doi:10.1086/345097
* Mullally et al. (2022) Mullally, S. E., Sloan, G. C., Hermes, J. J., et al. 2022, AJ, 163, 136. doi:10.3847/1538-3881/ac4bce
* Murphy et al. (2013) Murphy, S. J., Lawson, W. A., & Bessell, M. S. 2013, MNRAS, 435, 1325. doi:10.1093/mnras/stt1375
* Paz-Chinchón et al. (2015) Paz-Chinchón, F., Leão, I. C., Bravo, J. P., et al. 2015, ApJ, 803, 69
* Pecaut & Mamajek (2013) Pecaut, M. J. & Mamajek, E. E. 2013, ApJS, 208, 9. doi:10.1088/0067-0049/208/1/9
* Pecaut & Mamajek (2016) Pecaut, M. J. & Mamajek, E. E. 2016, MNRAS, 461, 794. doi:10.1093/mnras/stw1300
* Pedersen & Bell (2023) Pedersen, M. G. & Bell, K. J. 2023, AJ, 165, 239. doi:10.3847/1538-3881/accc31
* Petrucci et al. (2024) Petrucci, R. P., Gómez Maqueo Chew, Y., Jofré, E., et al. 2024, MNRAS, 527, 8290. doi:10.1093/mnras/stad3720
* Pineda et al. (2017) Pineda, J. S., Hallinan, G., & Kao, M. M. 2017, ApJ, 846, 75. doi:10.3847/1538-4357/aa8596
* Press & Rybicki (1989) Press, W. H. & Rybicki, G. B. 1989, ApJ, 338, 277. doi:10.1086/167197
* Prša et al. (2022) Prša, A., Kochoska, A., Conroy, K. E., et al. 2022, ApJS, 258, 16. doi:10.3847/1538-4365/ac324a
* Radigan et al. (2014) Radigan, J., Lafrenière, D., Jayawardhana, R., et al. 2014, ApJ, 793, 75. doi:10.1088/0004-637X/793/2/75
* Reylé (2018) Reylé, C. 2018, A&A, 619, L8. doi:10.1051/0004-6361/201834082
* Ricker et al. (2015) Ricker, G. R., Winn, J. N., Vanderspek, R., et al. 2015, Journal of Astronomical Telescopes, Instruments, and Systems, 1, 014003. doi:10.1117/1.JATIS.1.1.014003
* Rockenfeller et al. (2006) Rockenfeller, B., Bailer-Jones, C. A. L., & Mundt, R. 2006, A&A, 448, 1111. doi:10.1051/0004-6361:20054150
* Sarro et al. (2023) Sarro, L. M., Berihuete, A., Smart, R. L., et al. 2023, A&A, 669, A139. doi:10.1051/0004-6361/202244507
* Saumon et al. (2000) Saumon, D., Geballe, T. R., Leggett, S. K., et al. 2000, ApJ, 541, 374. doi:10.1086/309410
* Scargle (1982) Scargle, J. D. 1982, ApJ, 263, 835. doi:10.1086/160554
* STScI (2018) STScI 2018, TESS Input Catalog and Candidate Target List, STScI/MAST, doi:10.17909/FWDT-2X66
* Suárez & Metchev (2023) Suárez, G. & Metchev, S. 2023, MNRAS, 523, 4739. doi:10.1093/mnras/stad1711
* Tannock et al. (2021) Tannock, M. E., Metchev, S., Heinze, A., et al. 2021, AJ, 161, 224. doi:10.3847/1538-3881/abeb67
* Tinney & Tolley (1999) Tinney, C. G. & Tolley, A. J. 1999, MNRAS, 304, 119. doi:10.1046/j.1365-8711.1999.02297.x
* Tremblin et al. (2015) Tremblin, P., Amundsen, D. S., Mourier, P., et al. 2015, ApJ, 804, L17. doi:10.1088/2041-8205/804/1/L17
* Tsuji et al. (1996) Tsuji, T., Ohnaka, K., Aoki, W., et al. 1996, A&A, 308, L29
* Vos et al. (2022) Vos, J. M., Faherty, J. K., Gagné, J., et al. 2022, ApJ, 924, 68. doi:10.3847/1538-4357/ac4502
* Wang (2021) Wang, Y. 2021, Geophysics, 86, V31. doi:10.1190/geo2020-0316.1
* Wang (2022) Wang, Y. 2022, Geophysics, Time-Frequency Analysis of Seismic Signals: , Wiley. ISBN 978-1-119-89234-2. John Wiley & Sons, Ltd, 2022. doi:10.1002/9781119892373
* Zahnle & Marley (2014) Zahnle, K. J. & Marley, M. S. 2014, ApJ, 797, 41. doi:10.1088/0004-637X/797/1/41
* Zhan et al. (2019) Zhan, Z., Günther, M. N., Rappaport, S., et al. 2019, ApJ, 876, 127. doi:10.3847/1538-4357/ab158c
* Zuckerman et al. (2006) Zuckerman, B., Bessell, M. S., Song, I., et al. 2006, ApJ, 649, L115. doi:10.1086/508060
* Zuckerman et al. (2011) Zuckerman, B., Rhee, J. H., Song, I., et al. 2011, ApJ, 732, 61. doi:10.1088/0004-637X/732/2/61
Figure 1: Distribution of spectral types for the present working sample of 250
UCs with bonafide TESS LCs. Red color stands for UCDs with unambiguous
rotation and blue represents the entire UCD sample.
Figure 2: Examples of diagnostic plots displaying FFT and Lomb–Scargle
periodograms, LCs, LCs phase folded, and wavelet maps for three Gaia DR3
Ultra-Cool Dwarfs with typical TESS rotation signatures. Persistent
variability of rotation period 0.329, 0.519 and 0.472 days, respectively, for
TIC 170849193 (top panels), TIC 360329359 (middle panels), and TIC 359892714
(bottom panels) are observed in their wavelet maps and confirmed by FFT and
Lomb–Scargle peaks labeled A. Figures for the entire sample, following the
same design, are provided in the figure set given in the Online Material
Table 1: Gaia DR3 Ultra-Cool Dwarfs with TESS unambiguous periodicities
TIC ID | GAIA ID | Prot | eProt | tSPAN | NCycle | Sectors | CROWD | RUWE | Age
---|---|---|---|---|---|---|---|---|---
| | (days) | (days) | (days) | | | | | (Myr)
5630425 | 3493736924979792768 | 1.499 | 0.018 | 21.000 | 14.000 | 10 | 0.494 | 1.104 | 24aafootnotemark:
17308621 | 48193316928788352 | 1.332 | 0.020 | 43.981 | 33.019 | 43,44 | 0.589 | 1.357 | 2bbfootnotemark:
17518894 | 3314586142482134144 | 2.065 | 0.049 | 43.916 | 21.267 | 43,44 | 0.988 | 1.491 | 2bbfootnotemark:
20305594 | 1388684770725399296 | 0.281 | 0.001 | 73.210 | 260.534 | 23,24,50,51 | 0.994 | 1.183 | 149aafootnotemark:
45193132 | 5314960632252986880 | 2.486 | 0.025 | 123.355 | 49.620 | 8-10,35-37 | 0.093 | 1.01 | 149aafootnotemark:
49178529 | 5401822669314874240 | 0.967 | 0.022 | 21.503 | 22.237 | 36 | 0.436 | 1.085 | 10aafootnotemark:
56002511 | 3185321611286286976 | 0.377 | 0.003 | 23.785 | 63.089 | 5 | 0.956 | 2.317 | 149aafootnotemark:
56624850 | 163178353176600448 | 3.155 | 0.113 | 44.177 | 14.002 | 43,44 | 0.782 | 1.551 | 149aafootnotemark:
56658216 | 164802984685384320 | 2.606 | 0.077 | 44.140 | 16.938 | 43,44 | 0.823 | 1.013 | 2bbfootnotemark:
56658273 | 164800235906367232 | 5.81 | 0.382 | 44.160 | 7.601 | 43,44 | 0.894 | 0.888 | 2bbfootnotemark:
58229181 | 164513022853468160 | 2.11 | 0.050 | 44.140 | 20.919 | 43,44 | 0.978 | 1.54 | 2bbfootnotemark:
58285779 | 164513400810646912 | 1.669 | 0.063 | 22.000 | 13.182 | 44 | 0.183 | 1.205 | 2bbfootnotemark:
58436495 | 150501362066641664 | 0.58 | 0.004 | 43.885 | 75.664 | 43,44 | 0.270 | 0.979 | 2bbfootnotemark:
58538719 | 152516079683687680 | 2.447 | 0.068 | 44.046 | 18.000 | 43,44 | 0.872 | 1.036 | 2bbfootnotemark:
58601455 | 152108882425024128 | 4.66 | 0.251 | 43.211 | 9.273 | 44 | 0.568 | 0.976 | 2bbfootnotemark:
58638214 | 152917298349085824 | 0.752 | 0.004 | 67.494 | 89.753 | 19,43,44 | 0.726 | 1.372 | 2bbfootnotemark:
78772203 | 5424690587034982144 | 0.133 | 0.000 | 88.644 | 666.499 | 35,36,62,63 | 0.173 | 1.08 | 40ccfootnotemark:
96680763 | 156915878541979008 | 0.302 | 0.001 | 44.205 | 146.423 | 43,44 | 0.107 | 0.928 | 2bbfootnotemark:
101256057 | 216677698470254976 | 1.691 | 0.023 | 61.735 | 36.493 | 42,43,44 | 0.017 | 0.87 | $-$
102076870 | 3459372646830687104 | 0.982 | 0.011 | 41.944 | 42.713 | 10,37 | 0.412 | 1.068 | 10aafootnotemark:
118520413 | 145238687096970496 | 1.498 | 0.026 | 43.212 | 28.846 | 43,44 | 0.894 | 0.981 | 2bbfootnotemark:
125882806 | 148420639387738112 | 1.027 | 0.012 | 43.514 | 42.342 | 43,44 | 0.980 | 0.944 | 2bbfootnotemark:
125977598 | 147441558642852736 | 4.418 | 0.223 | 43.794 | 9.913 | 43,44 | 0.970 | 1.001 | 2bbfootnotemark:
150005829 | 148354733113981696 | 3.138 | 0.113 | 43.534 | 13.873 | 43,44 | 0.895 | 1.285 | 2bbfootnotemark:
150058659 | 148400229703257856 | 3.438 | 0.136 | 43.566 | 12.672 | 43,44 | 0.980 | 1.021 | 2bbfootnotemark:
150094175 | 157644373715415424 | 0.726 | 0.006 | 43.727 | 60.230 | 43,44 | 0.674 | 1.109 | 2bbfootnotemark:
150122702 | 148196510814073728 | 3.782 | 0.164 | 43.681 | 11.550 | 43,44 | 0.894 | 0.969 | 2bbfootnotemark:
167808464 | 5286310069347838080 | 0.279 | 0.000 | 145.412 | 583.983 | 7-13 | 0.223 | 0.99 | 45aafootnotemark:
170849193 | 4877824564574114304 | 0.329 | 0.002 | 23.768 | 72.243 | 5 | 0.845 | 1.079 | 45aafootnotemark:
232064183 | 6406967509044866560 | 0.718 | 0.006 | 41.292 | 57.510 | 27,28 | 0.064 | 1.062 | 24aafootnotemark:
268017136 | 151262876946558976 | 2.78 | 0.088 | 43.860 | 15.777 | 43,44 | 0.767 | 0.995 | 2bbfootnotemark:
268018471 | 152466120624336896 | 2.037 | 0.047 | 43.914 | 21.558 | 43,44 | 0.879 | 1.08 | 2bbfootnotemark:
268144659 | 152643240779301632 | 1.926 | 0.042 | 43.84 | 22.762 | 43,44 | 0.287 | 1.033 | 2bbfootnotemark:
268148996 | 151327159721125888 | 1.389 | 0.022 | 43.38 | 31.231 | 43,44 | 0.885 | 1.133 | 2bbfootnotemark:
268217504 | 149369139966814976 | 3.918 | 0.176 | 43.553 | 11.116 | 43,44 | 0.966 | 1.095 | 2bbfootnotemark:
268218180 | 146366442430208640 | 2.382 | 0.065 | 43.528 | 18.274 | 43,44 | 0.976 | 1.128 | 2bbfootnotemark:
268324394 | 146277553787186048 | 3.379 | 0.132 | 43.310 | 12.817 | 43,44 | 0.555 | 1.049 | 2bbfootnotemark:
268324578 | 147799209159857280 | 2.248 | 0.058 | 43.346 | 19.282 | 43,44 | 0.142 | 1.213 | 2bbfootnotemark:
268325250 | 151028990206478080 | 2.386 | 0.066 | 43.273 | 18.136 | 43,44 | 0.723 | 1.5 | 2bbfootnotemark:
268397898 | 147801339463632000 | 1.004 | 0.012 | 43.282 | 43.110 | 43,44 | 0.987 | 1.353 | 2bbfootnotemark:
268399069 | 151130591952773632 | 1.866 | 0.040 | 43.456 | 23.288 | 43,44 | 0.952 | 1.249 | 2bbfootnotemark:
298907057 | 3200303384927512960 | 0.484 | 0.005 | 23.879 | 49.337 | 5,32 | 0.825 | 1.099 | 200ddfootnotemark:
359892714 | 5841324542416019712 | 0.228 | 0.000 | 97.523 | 427.733 | 11,12,38,39 | 0.669 | 1.069 | 3.7 eefootnotemark:
360329359 | 5919792529769594240 | 0.502 | 0.005 | 26.857 | 53.500 | 39 | 0.142 | 0.974 | 149aafootnotemark:
367079252 | 2278720295036404736 | 2.769 | 0.029 | 132.993 | 48.029 | 17-19,24-26 | 0.906 | 1.026 | $-$
379097589 | 3724596914697213056 | 0.237 | 0.001 | 33.159 | 139.910 | 23,50 | 0.606 | 1.146 | 200ddfootnotemark:
397287296 | 3314309890186259712 | 2.655 | 0.080 | 43.785 | 16.492 | 43,44 | 0.826 | 1.125 | 2bbfootnotemark:
401838575 | 4954453580066220800 | 0.315 | 0.001 | 43.106 | 136.844 | 2,3 | 0.978 | 1.178 | 45aafootnotemark:
427770178 | 147614422487144960 | 1.479 | 0.025 | 43.355 | 29.314 | 43,44 | 0.726 | 1.198 | 2bbfootnotemark:
440859813 | 6224387727748521344 | 2.04 | 0.047 | 44.361 | 21.746 | 11,38 | 0.480 | 1.256 | 149aafootnotemark:
454227159 | 5201185574182015744 | 2.042 | 0.020 | 51.000 | 25.07 | 38,39 | 0.395 | 1.145 | 2bbfootnotemark:
454291779 | 5201129563515179520 | 3.23 | 0.075 | 70.000 | 21.672 | 11,38,39 | 0.848 | 1.024 | $-$
456944264 | 3314299238667410176 | 2.22 | 0.056 | 43.834 | 19.745 | 43,44 | 0.898 | 1.048 | 2bbfootnotemark:
902237947 | 3562157781229213312 | 0.241 | 0.001 | 21.439 | 88.884 | 36 | 0.210 | 3.117 | 200ddfootnotemark:
Targets with 2nd period | |
20305594 | 1388684770725399296 | 0.563 | 0.002 | 73.210 | 130.035 | 23,24,50,51 | 0.994 | 1.183 | 149aafootnotemark:
58638214 | 152917298349085824 | 0.367 | 0.001 | 67.494 | 183.758 | 19,43,44 | 0.726 | 1.372 | 2bbfootnotemark:
96680763 | 156915878541979008 | 0.117 | 0.000 | 44.205 | 377.821 | 43,44 | 0.107 | 0.928 | 2bbfootnotemark:
150094175 | 157644373715415424 | 0.966 | 0.011 | 43.727 | 45.266 | 43,44 | 0.674 | 1.109 | 2bbfootnotemark:
359892714 | 5841324542416019712 | 0.458 | 0.001 | 97.523 | 212.932 | 11,12,38,39 | 0.669 | 1.069 | 3.7eefootnotemark:
Note. — Ages are from (a)(Bell et al., 2015), (b)(Kenyon & Hartmann, 1995),
(c)(Zuckerman et al., 2011), (d)(Zuckerman et al., 2006), (e)(Murphy et al.,
2013)
Table 2: Gaia DR3 Ultra-Cool Dwarfs with low-amplitude and TESS unambiguous
periodicities.
TIC ID | GAIA ID | Prot (LS) | Amplitude | Prot (WT) | TESS sectors | Noisy | CROWD | RUWE | Age
---|---|---|---|---|---|---|---|---|---
| | (days) | (%) | (days) | with modulation | TESS sectors | | | (Myr)
26126812 | 68012529415816832 | 0.229 | 0.50 | 0.260 | 42,43,44 | | 0.974 | 1.217 | 24aafootnotemark:
50085560 | 2613754712222934656 | 0.427 | 0.80 | 0.444 | 42 | | 0.978 | 0.958 | 149aafootnotemark:
52256020 | 4691426694779173376 | 0.265 | 0.50 | 0.254 | 1,2,27-29 | | 0.971 | 1.292 | 45aafootnotemark:
58432630 | 152284735566828032 | 0.642 | 0.85 | 0.690 | 43,44 | | 0.511 | 0.945 | 2bbfootnotemark:
201688405 | 6473160651658069120 | 0.692 | 0.39 | 0.731 | 27 | 13 | 0.850 | 0.983 | 149aafootnotemark:
266012525 | 2556491429388743296 | 0.313 | 1.76 | 0.313 | 43 | | 0.971 | 1.079 | 45aafootnotemark:
275133493 | 3852362605385919616 | 0.438 | 0.73 | 0.446 | 8,35 | 45 | 0.808 | 1.042 | 149aafootnotemark:
281668854 | 4920330653311504256 | 0.116 | 0.56 | 0.118 | 2,28,29 | | 0.877 | 9.984 | 45aafootnotemark:
299007548 | 3230008650057256960 | 0.529 | 0.40 | 0.568 | 32 | 5 | 0.809 | 1.140 | 24aafootnotemark:
302396416 | 1113083282052337792 | 0.458 | 0.42 | 0.464 | 19,20,26,40,47 | 60,53 | 0.832 | 1.020 | 149aafootnotemark:
342967337 | 5316349173696052736 | 0.705 | 2.62 | 0.697 | 8 | 9,10 | 0.126 | 1.056 | 45aafootnotemark:
356632694 | 1640572714166158976 | 0.240 | 0.27 | 0.260 | 14,16,21,49 | 15,22-24,41,48,50 | 0.705 | 1.426 | 149 aafootnotemark:
371080717 | 898275195732381440 | 0.714 | 0.35 | 0.714 | 20 | 60 | 0.311 | 1.057 | 40ccfootnotemark:
388682292 | 5514929155583865216 | 0.533 | 9.80 | 0.543 | 8 | | 0.099 | 2.109 | 149aafootnotemark:
405514195 | 6059992460039778560 | 0.237 | 2.90 | 0.240 | 11 | | 0.160 | 1.047 | 15ddfootnotemark:
448590406 | 3478519134297202560 | 0.404 | 0.16 | 0.431 | 36 | 10 | 0.769 | 2.201 | 10aafootnotemark:
959690391 | 5856405272135505024 | 0.221 | 21.69 | 0.223 | 38 | | 0.005 | 1.144 | 149aafootnotemark:
Note. — Ages are from (a)(Bell et al., 2015), (b)(Kenyon & Hartmann, 1995),
(c)(Zuckerman et al., 2011), (d)(Pecaut & Mamajek, 2016)
Table 3: Gaia DR3 Ultra-Cool Dwarfs with TESS ambiguous or dubious variability. TIC ID | GAIA ID | Ambiguous or Dubious | Noisy | RUWE
---|---|---|---|---
| | TESS sectors | TESS sectors |
26390890 | 216704503361774080 | 42,43,44 | | 1.057
58108762 | 164474986623118592 | 43,44 | | 0.880
58175278 | 164783811951433856 | 43,44 | | 3.177
58229208 | 164702070133970944 | 19,43,44 | | 1.216
58285782 | 164502062096975744 | 19,43,44 | | 1.408
58544613 | 149629483705467008 | 43,44 | | 1.152
58601266 | 152029992465874560 | 43,44 | | 1.024
118710667 | 146874275068113664 | 43,44 | | 0.958
125882881 | 148449845165337600 | 43,44 | | 1.080
167670065 | 6118581861234228352 | 11 | 38 | 0.974
260304085 | 5496160698257650048 | 4 | 5,7,8,10-13 | 0.991
346721695 | 216575791780899200 | 42,44 | 43 | 1.004
348844681 | 5479241065435684352 | 13 | 5,7,8,12 | 0.987
367089503 | 2279554205887675392 | 17,19,24,25 | | 1.149
385557215 | 71504509626338048 | 42 | 43,44 | 0.909
387813466 | 2270043769607851264 | 16, 17,18,24, 25 | | 1.135
389053345 | 3311785239689504512 | 44 | 43 | 1.020
397367343 | 3480771277705948928 | 10 | | 1.041
436574130 | 3406128761895775872 | 43,44 | | 0.964
467753980 | 169163647804297216 | 44 | 43 | 1.010
Table 4: Gaia DR3 Ultra-Cool Dwarfs with previous TESS periodicities in the literature. TIC ID | Prot(h) | Amplitude(%) | TESS sectors | TESS sectors | Prot(h) | Amplitude(%) | TESS sectors | CROWD | RUWE
---|---|---|---|---|---|---|---|---|---
| this work | with modulation-this work | Noisy-this work | Miles-Páez et al. (2022) | |
8259901 | 3.072 $\pm$ 0.007 | 0.93 | 21 | — | 2.5267 $\pm$ 0.0002 | 0.44$\pm$0.01 | 21 | 0.240 | 2.785
8259902 | 3.072 $\pm$ 0.005 | 0.26 | 21,48 | — | 3.0889 $\pm$ 0.0002 | 0.44$\pm$0.01 | 21 | 0.829 | 1.498
16227031 | — | — | — | 44,45,46 | — | — | 44 | 0.663 | 1.212
17308656 | — | — | — | — | 4.9199$\pm$0.0035 | 0.58$\pm$0.08 | 43,44 | — | 2.794
27858644 | 2.832 $\pm$ 0.002 | 0.17 | 26,40,54 | 53 | 2.84140 $\pm$ 0.00039 | 0.34$\pm$0.03 | 26,40 | 0.703 | 1.096
32422219 | 2.112 $\pm$ 0.002 | 0.19 | 3,30 | — | 2.15012$\pm$0.00024 | 0.53$\pm$0.06 | 3,30 | 0.902 | 2.698
34014829 | 3.859 $\pm$ 0.007 | 0.24 | 2,29 | — | 3.8632 $\pm$ 0.0016 | 0.44$\pm$0.03 | 2,29 | 0.892 | 5.417
143029977 | 3.840 $\pm$ 0.014 | 0.1 | 31 | 4 | 3.8356 $\pm$0.0010 | 0.17$\pm$0.04 | 4,31 | 0.888 | 0.999
193974787 | 3.072 $\pm$ 0.009 | 0.24 | 16 | 23,24,49,51 | 3.0942$\pm$0.0010 | 0.48$\pm$0.08 | 16,23,24 | 0.651 | 0.998
198107795 | 11.292 $\pm$ 0.039 | 0.52 | 16,23,49,50 | — | 11.4334 $\pm$ 0.0058 | 0.59$\pm$0.03 | 16,23 | 0.845 | 11.011
202408306 | 15.336 $\pm$ 0.153 | 0.76 | 23,50 | — | 15.34$\pm$0.03 | 0.66$\pm$0.06 | 23 | 0.993 | 0.946
219095664 | 3.888 $\pm$ 0.001 | 0.24 | 14-21,24 | 22,23,40,41, | 3.89442 $\pm$ 0.00041 | 0.18$\pm$0.04 | 14-25,40,41 | 0.852 | 4.316
| | | 25,55,57,58 | 47-52,54,56,59 | | | | |
229653723 | — | — | — | — | — | $<$0.34 | 22 | — | 1.231
239097694 | — | — | — | 19,59 | — | — | 19 | 0.267 | 0.993
286963145 | — | — | — | 23 | — | $<$0.28 | 23 | 0.962 | 1.095
288506050 | — | — | — | 16,22,23 | 2.4221 $\pm$ 0.0003 | 0.70$\pm$0.05 | 16,22,23 | 0.986 | 1.178
307956653 | 3.614 $\pm$ 0.001 | 0.4 | 16-19,21,22,25,26,40 | 14,15,24,51 | 3.62703$\pm$0.00031 | 0.49$\pm$0.1 | 14-26,40,41 | 0.853 | 1.150
| | | 41,47-50,53,55-58,60 | 52,54,59 | | | | |
311188315 | — | — | — | 24,51 | 1.95976$\pm$0.00043 | 0.43$\pm$0.08 | 24 | 0.372 | 1.661
438883548 | — | — | — | — | 3.992$\pm$0.017 | 0.42$\pm$0.07 | 11 | — | 1.445
441000085 | — | — | — | 11,38 | 3.9614 $\pm$ 0.0009 | 0.77$\pm$0.06 | 11,38 | 0.831 | 1.185
902237947 | 6.000 $\pm$ 0.036 | 3.93 | 36 | — | 5.984$\pm$0.001 | 0.6$\pm$0.02 | 9 | 0.210 | 3.117
Figure 3: The distribution of rotation periods from the present analysis for
the UCDs with rotation signature and single source behavior. The dashed black
histogram stands for the 48 UCDs with unambiguous rotation periods, and the
red color for the 13 UCDs withunambiguous rotation and low amplitude
variability in the LCs.
Figure 4: The Gaia CMD for the sample of UCDs composing the present study,
with black circles illustrating the whole sample of 7630 Gaia DR3 UCDs, given
by Sarro et al. (2023). In the left panel circles in blue are for the UCDs
with $P_{rot}\geq 1.0$ d, red circles are for the UCDs with P${}_{rot}<1$ d,
and red open diamonds stand for the UCDs listed by Miles-Páez et al. (2023).
In the right panel, colors follow the color bar age intervals, which gives the
age in Myr.
|
# Towards Dynamic Urban Bike Usage Prediction for Station Network
Reconfiguration
Xi Yang and Suining He The University of ConnecticutStorrsConnecticut
xi.yang<EMAIL_ADDRESS>
###### Abstract.
Bike sharing has become one of the major choices of transportation for
residents in metropolitan cities worldwide. A station-based bike sharing
system is usually operated in the way that a user picks up a bike from one
station, and drops it off at another. Bike stations are, however, not static,
as the bike stations are often reconfigured to accommodate changing demands or
city urbanization over time. One of the key operations is to evaluate
candidate locations and install new stations to expand the bike sharing
station network. Conventional practices have been studied to predict existing
station usage, while evaluating new stations is highly challenging due to the
lack of the historical bike usage.
To fill this gap, in this work we propose a novel and efficient bike station-
level prediction algorithm called AtCoR, which can predict the bike usage at
both existing and new stations (candidate locations during reconfiguration).
In order to address the lack of historical data issues, virtual historical
usage of new stations is generated according to their correlations with the
surrounding existing stations, for AtCoR model initialization. We have
designed novel station-centered heatmaps which characterize for each target
station centered at the heatmap the trend that riders travel between it and
the station’s neighboring regions, enabling the model to capture the learnable
features of the bike station network. The captured features are further
applied to the prediction of bike usage for new stations. Our extensive
experiment study on more than 23 million trips from three major bike sharing
systems in US, including New York City, Chicago and Los Angeles, shows that
AtCoR outperforms baselines and state-of-art models in prediction of both
existing and future stations.
bike sharing, usage prediction, station-level, new stations, pick-ups and
drop-offs, attention, spatio-temporal
††copyright: none††conference: UrbComp ’20: The 9th SIGKDD International
Workshop on Urban Computing; August 24, 2020; San Diego, CA
## 1\. Introduction
Thanks to mobile networking and location-based services, bike sharing has
become one of the major transportation modalities for urban residents
worldwide due to its convenience and efficiency. As a representative product
of the sharing economy, it is often hailed as the excellent helper to solve
the “last mile” problem in citizen transportation. Given the social and
business importance, the bike sharing market is estimated to hit US$5 billion
by 2025111https://www.globenewswire.com/news-
release/2019/12/03/1955257/0/en/Bike-Sharing-Market-is-Predicted-to-
Hit-5-Billion-by-2025-P-S-Intelligence.html.
A station-based bike sharing system (each station is equipped with multiple
docks for bike parking) is usually operated in the way that a user or rider
picks up a bike from one station (pick-ups) and drops it off at another (drop-
offs). All the resultant bike trips (usage that consists of pairs of pick-ups
and drop-offs) connect different parts of the city, forming the bike sharing
station network. The bike stations are, however, not static, as the bike
sharing operators often reconfigure the stations to accommodate changing
demands or city urbanization over time (He and Shin, 2018; Liu et al., 2015),
where a key operation is to evaluate candidate locations and then install new
stations to expand the bike sharing station network (Zhang et al., 2016; Liu
et al., 2017) as shown in Figure 1. It is essential for the operators to know
the potential bike usage of a future station at a certain candidate location
beforehand, which helps predicting the station profitability as well as its
positive/negative effect upon the local mobility and traffic networks,
enhancing the bike sharing service quality to the local community. This
further benefits the future bike sharing network operations, including station
re-balancing (O’Mahony and Shmoys, 2015; Liu et al., 2016; Hulot et al., 2018;
Wang et al., 2018; Pan et al., 2019a) and bike route pre-planning (Zhang and
Philip, 2016).
Despite the business and social importance, predicting bike usage at future
stations is extremely challenging due to the lack of historical bike usage
data of those stations. Conventional practices include user survey,
crowdsourced feedbacks and public hearing, investigating the local demands,
which is often costly and time-consuming. Operating regulations are also
provided for bike sharing operators as guidance on candidate station
locations, which is based primarily on geo-information such as distance to
road intersections/public transit stations and station network
density222https://www.transformative-mobility.org/assets/publications/The-
Bikeshare-Planning-Guide-ITDP-Datei.pdf. However, such guidance is often too
general to take into account uniqueness of different cities, causing
inefficiency in station reconfiguration as well as reduced service quality and
profit loss. While most of the existing studies focus on predicting the bike
usage at the existing stations, only a few of them have comprehensively
explored how to forecast new stations for the station reconfiguration
purposes.
Figure 1. Illustration of bike sharing system and new station usage
prediction.
To fill this gap, in this work we propose a novel bike station prediction
algorithm called AtCoR, based on Attention Convolutional Recurrent Neural
Network, for predicting station-based bike usage of future stations. To tackle
the lack of historical data issues for future/new stations to be deployed,
virtual historical usage of new stations is generated according to their
correlation with the surrounding existing stations. To predict the usage of
new bike stations for reconfiguration decisions, we have designed novel
station-centered heatmaps which characterize for each target station centered
at the heatmap the trend that riders travel between it and the neighboring
regions, so that the model is able to capture the learnable common patterns of
the bike station network through a convolutional neural network (CNN). A Long
Short-Term Memory (LSTM) neural network with temporal attention mask leverages
the common patterns with integration of historical data and external factors,
such as weekends, holidays, and weather, to predict the bike usage for
existing stations. The captured bike usage features are then used to predict
the pick-ups and drop-offs for new stations along with the virtual historical
usage. An overview of information flow of this work is illustrated in Figure
2, summarizing the above process.
Our main technical contributions are as follows:
* 1)
Comprehensive bike data analysis for station prediction designs: We have
conducted comprehensive and detailed real-world data analysis on how the
weather conditions, regional bike usage and surrounding POIs impact the bike
usage in the metropolitan area like New York City (NYC), Chicago and Los
Angeles (LA), and visualized them to validate our design insights for station
reconfiguration prediction. These features serve as the shared patterns
leveraged for the bike usage prediction, including predicting for the new
stations.
* 2)
Dynamic urban bike usage prediction for reconfigured station networks: We have
designed a novel scheme called AtCoR to predict the station-level bike usage
for new/future stations (candidate locations for reconfiguration) as well as
existing/fixed ones. We propose a novel design of station-centered feature
heatmap representation as an input of AtCoR, which calculates the differences
between the features at the location of each station and those at the
surrounding areas. The heatmaps account for the trend that riders travel from
the center station to the neighborhood. The inclusion of the heatmaps
significantly improves the accuracy of usage predictions. Heatmaps are then
fed to AtCoR which consists of a deep Convolutional Neural Network (CNN) and a
Long Short-Term Memory (LSTM) network with integration of temporal attention
mechanism.
* 3)
Extensive experimental studies with real-world datasets: We have conducted
extensive experimental studies upon 23,955,989 trips in total, across three
major bike sharing systems in the United States: 20,551,697 from Citi Bike in
New York City (NYC), 3,113,950 from Divvy in Chicago, and 290,342 from Metro
Bike in Los Angeles (LA). Our experimental studies, upon both the existing and
new stations in the bike station network reconfiguration, have validated that
AtCoR outperforms the other baseline models in multi-station predictions,
often by more than 20% error reduction.
Figure 2. System overview and information flow of AtCoR.
While our studies focus on station-based bike sharing, the model and algorithm
derived can be further extended to other transportation platform with
deployment and expansion operations, including ride/car sharing (Chen et al.,
2017; Jauhri et al., 2017; He and Shin, 2019a) and scooter sharing (He and
Shin, 2020a).
The rest of the paper is organized as follows. We first conduct a brief survey
on related studies in Section 2. Then we analyze the datasets, and find out
the features and their corresponding representations for model input Section
3. Then the model is further presented in Section 4. Afterwards, we present
our extensive experimental studies to evaluate the performance of AtCoR for
both existing and new stations in Section 5. We will discuss the deployment in
Section 6, and conclude in Section 7.
## 2\. Related Work
Traffic flow prediction has been studied recently due to the advances of
intelligent transportation and smart cities. Gong et al. proposed a potential
passenger flow predictor to help the decision on the places to construct new
metro station (Gong et al., 2019). Tang et al. tackled the dense and
incomplete trajectories for citywide traffic volume inference (Tang et al.,
2019). For dockless bike sharing system, the potential bike distribution and
the detection of parking hotspots in a new city has been made by Liu et al.
(Liu et al., 2018a, b). Different from above works on dockless bike sharing,
we focus on station-based deployment due to its wider social acceptance, and
our studies provide highly comprehensive studies of predicting usage for
new/future stations before the bike sharing station reconfiguration (He and
Shin, 2018).
Based on bike usage prediction granularity, there are three categories of
prediction models in current works in the bike sharing systems: city-level,
cluster-level, and station-level (Li et al., 2019). In city-level prediction,
the aim is to predict bike usage for a whole city, while at the cluster level
the goal is to predict bike usage for clusters of bike stations. The station
cluster is generated by clustering algorithms such as the Bipartite Station
Clustering (Li et al., 2015), the Community Detection and the Agglomerative
Hierarchical Clustering method (Zhou, 2015), the K-means clustering and the
Latent Dirichlet Allocation (LDA). Chen et al. further considered clusters as
dynamic rather than static (Chen et al., 2016). While city-level and cluster-
level predictions save the computational cost by simplifying the problems,
station-level (Hulot et al., 2018; Chai et al., 2018; Chen et al., 2020; He
and Shin, 2020b) prediction still benefits the bike sharing system management
the most, including fine-grained station rebalancing, but yet is challenging.
Recent research efforts have also been made upon traffic prediction beyond the
bike sharing systems. Deep learning approaches have been studied for traffic
flow prediction (Lv et al., 2014; Huang et al., 2014; Yu et al., 2017;
Fadlullah et al., 2017; He and Shin, 2020a, 2019b). Yao et al. proposed a CNN-
LSTM based transfer learning method to predict traffic flows across different
cities (Yao et al., 2019). Pan et al. used a deep meta learning model to
predict urban traffic (Pan et al., 2019b). Ma et al. constructed spatial-
temporal graphs of urban traffic which was learned by a deep convolutional
neural network (Ma et al., 2017). In this work, we study a novel approach
based on deep learning designs and data-driven studies to handle the new
stations usage prediction problems. Specifically, we propose a novel approach
AtCoR which consists of a CNN component for modeling spatial characteristics
and a LSTM with integration of a temporal attention mechanism for capturing
temporal characteristics .
## 3\. Data Analysis & Feature Designs
Because no historical usage data of new stations is available, future mobility
patterns might only be predicted through accessible spatial and temporal
characteristics of the stations. In this work, we propose using station
specific regional usage, points of interest (POIs), weather conditions and
holidays as input features as they are all available once the locations and
launching time of new stations are provided by system operators.
In this section, we first present the preliminary concepts defined for the
feature designs of bike sharing stations in Section 3.1. Then in Section 3.2
we discuss our representation design of the features we choose, namely
station-centered heatmaps, which significantly improve the model performance.
We further cluster station networks based on such heatmaps to save
computational cost of training.
Table 1 summarizes all the symbols as well as their definitions in this work.
Table 1. List of symbols and definitions Symbols | Definitions
---|---
$n$ | New stations
$f$ | Existing stations
$\mathbf{F}$ | Set of existing stations
$j,k$ | Index of the stations
$T,S$ | Ranges of timestamps
$t,\tau$ | Indices of timestamps
$\mathcal{T}$ | Number of timestamps
$g_{lat},g_{lon}$ | Height and width of each grid
$\mathcal{G}_{lat},\mathcal{G}_{lon},\mathcal{P}$ | Dimensions of station-centered heatmaps
$\mathbf{H}$ | Station-centered heatmaps
$e$ | Entries of station-centered heatmaps
$L$ | Station usage
$h,w,c$ | Parameters of the convolutional layers
$\mathbf{x}$ | Input of LSTM
$\mathbf{h}$ | LSTM hidden state
$\mathbf{c}$ | LSTM cell state
$d$ | Number of hidden units of LSTM
$\mathbf{v},\mathbf{W},\mathbf{U},\mathbf{b}$ | Trainable parameters
$\lambda,\gamma$ | Attention scores between a pair of states
$\mathbf{d}$ | Decoder’s input
$\mathbf{ex}$ | External features
$\omega$ | | Similarity scores between a new station and
---
its surrounding existing peers
### 3.1. Preliminary for Station Feature Studies
#### 3.1.1. Overview of Datasets and New/Existing Stations
We conduct our data analytics upon three datasets: Citi Bike of
NYC333https://www.citibikenyc.com/ of 2019 (20,551,697 trips), Divvy Bike of
Chicago444https://data.cityofchicago.org/Transportation/Divvy-Trips/fg6s-gzvg
of the first three quarters of 2019 (3,113,950 trips), and Metro Bike of Los
Angeles555https://bikeshare.metro.net/ of 2019 (3,113,950 trips).
In this study, the new stations established in a set of certain time intervals
$[T,T^{\prime}]$ are defined as the ones that have no historical bike usage
data in the past 30 days, i.e. from $(T/24-30)$ days to $T$ considering $T$ on
hourly basis, while the active existing stations (or existing stations in
short) in a certain time interval $[S,S^{\prime}]$ are defined as the ones
that have historical usage data every day in this time interval. Note that
active existing stations focus on the existing stations which are populous
with everyday bike usage.
For example, there are 454 active existing stations at the New York City (NYC)
in 2019 in total, and from April 11, 2019 to July 19, 2019 there are 631
existing stations, while in June 2019 there are 8 new stations established,
compared to totally 1,047 unique bike station coordinates in the whole year
including those stations that are newly installed, removed, and relocated. The
monthly number of active existing stations and new stations are shown in
Figure 3 for the three datasets as demonstration.
Figure 3. The number of active existing stations and new stations per month in
NYC, Chicago and LA.
We select the following spatial and temporal features in our modeling:
regional usage in areas where bike stations are located, points of interest,
weather conditions and holidays and weekends. Since there is no historical
bike usage at the future locations of the new bike stations, it is difficult
to directly characterize the mobility patterns for the target locations.
However, the mobility patterns are strongly related to the spatial and
temporal features of those locations. Thus, in this work AtCoR incorporates
the correlations with the above features for the new station prediction.
#### 3.1.2. Regional Usage
To characterize the urban bike usage in a computationally efficient manner, we
discretize the neighborhood city map around each station into grids, each of
which is an $g_{lat}$ m $\times$ $g_{lon}$ m rectangular region. The total
bike pick-ups/drop-offs within the grid can represent the bike usage
popularity of this specific region. Bikes rented at the popular regions tend
to be returned at surrounding regions. Therefore, the station-centered
regional usage distribution is chosen an essential input feature.
#### 3.1.3. Points of Interest (POIs)
Another key insight of feature selection is that the differences in the POI
distributions around each station can steer the bike riders’ travels with the
corresponding preferences or purposes. Therefore, the POI distributions are
used as another important features. Following the manners of defining regional
station usage, the POI distributions are defined as the total numbers of POIs
within a $g_{lat}$ m $\times$ $g_{lon}$ m grid for each POI category.
Figure 4 illustrates some examples of the different distributions of different
categories of POIs in a 500m $\times$ 500m grid region. POIs around active
existing stations from April 11, 2019 to July 19, 2019 are shown here with the
numbers of POIs normalized by the min-max method.
Figure 4. POIs distributions around bike stations in NYC, Chicago and LA.
Specifically, the POIs data of NYC are obtained through NYC Open
Data666https://data.cityofnewyork.us/City-Government/Points-Of-
Interest/rxuy-2muj which contain 13 major categories, such as residential,
education facility, cultural facility, recreational facility, social services,
transportation facility, commercial, government facility (non public safety),
religious institution, health services, public safety, water and others. The
POIs of Chicago are obtained through OpenStreetMap Overpass
API777https://wiki.openstreetmap.org/wiki/Overpass_API where POIs are
categorized by the OSM tag of amenity including sustenance, education,
transportation, financial, healthcare, entertainment, arts $\&$ culture, and
others. The POIs of LA are obtained through LAC Open
Data888https://data.lacounty.gov/ which include communications,
transportation, private industry, health and mental health, social services,
postal, arts and recreation, community groups, municipal services, public
safety, education, government, emergency response, physical features and
environment.
#### 3.1.4. Weather Conditions, Holidays and Weekends
The station usage is highly correlated with the weather conditions. We have
collected and analyzed the weather condition data from open source weather
data API999https://api.weather.com. We have analyzed the correlations between
weather conditions and the station usage. Analysis of one-year bike usage
reveals that both high and low temperature decreases the bike usage. The
effect of daily temperature, precipitation and wind speed on daily overall
usage of the city is illustrated in Figures 5 and 6. Clearly, precipitation,
including rain, snow and large wind speed, significantly reduces the bike
usage. Given above, the weather conditions including temperature,
precipitation and wind speed are chosen as the input features. Besides weather
conditions, the bike usage has different patterns on federal holidays and
weekends from that on workdays. Specifically, we set the indicator as 1 if a
time interval belongs to holiday/weekend periods, or 0 for weekdays otherwise.
As an example, for the time interval of [0:00 a.m., 1:00 a.m.], 2019-01-01,
the external vector including temperature, wind speed, precipitation and
holiday/weekends, is given by [47 °F, 1.5 mph, 0.08 in, 1].
Figure 5. Temperature effect on daily overall bike usage. Figure 6. The effect
of precipitation and wind speed on city’s daily overall bike usage.
### 3.2. Feature Representations & Designs
Given the spatial and temporal features presented above for each station, we
present a representation design to integrate them as the model inputs, which
will be discussed in details as followed.
#### 3.2.1. Station-centered Heatmaps
We construct a $\mathcal{G}_{lat}\times\mathcal{G}_{lon}\times\mathcal{P}$
heatmap centered at the stations studied where each grid is a $g_{lat}$ m
$\times$ $g_{lon}$ m area on city map with $\mathcal{P}$ channels including
the regional usage and POIs distribution of the grid area. The first two
channels of the heatmap are the regional pick-ups and drop-offs, respectively,
with each of the following channel as the POIs distributions of one POIs
category. Every entry of the heatmaps is the regional usage or the POIs amount
for the corresponding grid location. Figure 7 shows a $11\times 11$ regional
pick-ups as the first channel of the station-centered heatmap for a candidate
station in NYC.
After construction of the heatmaps centered at a particular station, the
heatmaps are normalized by subtracting each one of the
$\mathcal{G}_{lat}\times\mathcal{G}_{lon}$ grid features by the features of
the center grid. This way, the normalized heatmaps represent the riders’
motivations or mobility trends departing from this station to the
neighborhoods, characterizing the spatial-temporal features near a station.
(a)
(b)
Figure 7. Regional pick-ups for a candidate station (red cycle) in Manhattan,
NYC on (a) Sunday 01-20-2019 from 8:00 a.m. to 9:00 a.m.; (b) Monday
01-21-2019 from 8:00 a.m. to 9:00 a.m.
#### 3.2.2. Station Clustering
To enhance the learning efficiency upon large-scale bike station network, we
design a clustering scheme for the bike stations and train the AtCoR for
certain cluster. We cluster existing stations and newly established ones all
together. This way, we can save computational cost by reducing the amount of
stations being trained, while the predictions upon new stations can be enabled
by the patterns learned from other stations.
We adopt the K-means clustering algorithm to find out the clusters from the
normalized station-centered heatmaps constructed as in Section 3.2. We first
calculate the sum of all the $\mathcal{G}_{lat}\times\mathcal{G}_{lon}$
entries at each channel, and the mean of heatmaps over the timeframe
$[T,T^{\prime}]$, generating vectors of length $\mathcal{P}$. The POI metric
scores between the heatmaps of stations $j$ and $k$ for the clustering are
calculated from the Euclidean distance between the generated vector:
(1)
$Sim_{j,k}=\left\|\sum_{e=1}^{\mathcal{G}_{lat}\times\mathcal{G}_{lon}}\overline{\mathbf{H}}_{j}^{{\textit{e}}}-\sum_{e=1}^{\mathcal{G}_{lat}\times\mathcal{G}_{lon}}\overline{\mathbf{H}}_{k}^{{\textit{e}}}\right\|.$
The combination of both existing and new stations in each city are then
clustered based on the above metric scores. The results of clustering existing
and new stations for NYC and Chicago based on $11\times 11\times\mathcal{P}$
heatmaps are shown below on Figure 8.
Figure 8. The cluster results of existing and new stations. The clusters
represented by dots contain new stations, while the cross ones have only
existing stations.
## 4\. AtCoR: Usage Prediction for Station Network Reconfiguration
We further discuss the details of AtCoR model in this section. We first define
our usage problem, followed by a description of model structure in Section
4.1. Our model is trained on the datasets of active existing stations, and the
details will be discussed in Section 4.2.
### 4.1. Problem & Model Definitions
#### 4.1.1. Problem Definition
The problem in this study can be defined as: given the input station-centered
heatmaps, $\mathbf{H_{n,\tau^{\prime}}}$, and the external features
$\mathbf{ex_{\tau^{\prime}}}$, including the weather conditions and
weekend/holidays information, for each station $n$ at the period of time
$\tau^{\prime}\in[1,\mathcal{T}]$, either an existing or new one after
reconfiguration, predict future usage (pick-ups and drop-offs), $L_{n,\tau}$,
where $\tau=\mathcal{T}+1$ is the target time interval for an existing station
or the one when the new station is established after reconfiguration.
#### 4.1.2. Model Overview
Our model consists of three major modules.
1. 1)
Spatial feature learning: A deep-channel CNN learns the station specific
heatmaps which describe the driving force of usage for each particular station
based on its spatial characteristics.
2. 2)
Temporal feature learning: The output of CNN combined with historical usage
data is then fed as input of a LSTM which learns the temporal pattern of
station’s usage.
3. 3)
Feature attention characterization: A temporal attention mechanism is applied
to further capture and differentiate the correlations between features across
different timestamps.
An overview of AtCoR is further illustrated in Figure 9. The details of each
module will be presented as follows.
Figure 9. Illustration of model designs in AtCoR.
#### 4.1.3. Convolutional Neural Network (CNN)
As described before, the station-specific or station-centered heatmaps as the
representation of spatial features, i.e. the POIs distribution and regional
usage popularity, indicate the motivation of people heading to places around
and hence correlate with the bike usage. Differences in this driving force may
result in different directions of bike usage and are represented by the
designed heatmaps. However, the mechanism behind this driving force of bike
usage can be too complicated to formulate, and simple vector concatenation is
not enough. Therefore, we propose a deep-channel CNN model to learn the
correlation between station specific spatial features in an area and station
bike usage. Note that different from previous studies (Ma et al., 2017), we
focus on learning station-centered heatmaps to further identify the useful
correlations between the center station’s usage and the neighborhood.
The CNN has $C$ convolutional layers, each of which is followed by a max
pooling layers. The size of the convolutional layers are $h\times w\times c$,
all with relu activation function. A fully connected layer is implemented as
the last layer to generate the output of CNN of size 1.
#### 4.1.4. Long Short-Term Memory (LSTM)
The output of the CNN is concatenated with historical usage data as the input
of LSTM. The last hidden state of the LSTM cell is connected to a fully
connected layer to generate predictions. The equations of an LSTM cell at
timestamp $t$ are listed below:
(2) $\displaystyle\mathbf{i}_{t}$
$\displaystyle=\sigma\left(\mathbf{W}^{i}\left[\mathbf{h}_{t-1},\mathbf{x}_{t}\right]+\mathbf{b}^{i}\right),$
$\displaystyle\mathbf{f}_{t}$
$\displaystyle=\sigma\left(\mathbf{W}^{f}\left[\mathbf{h}_{t-1},\mathbf{x}_{t}\right]+\mathbf{b}^{f}\right),$
$\displaystyle\mathbf{o}_{t}$
$\displaystyle=\sigma\left(\mathbf{W}^{o}\left[\mathbf{h}_{t-1},\mathbf{x}_{t}\right]+\mathbf{b}^{o}\right),$
$\displaystyle\mathbf{\tilde{c}}_{t}$
$\displaystyle=\text{tanh}\left(\mathbf{W}^{i}\left[\mathbf{h}_{t-1},\mathbf{x}_{t}\right]+\mathbf{b}^{i}\right),$
$\displaystyle\mathbf{c}_{t}$
$\displaystyle=\left(\mathbf{f}_{t}*\mathbf{c}_{t-1}+\mathbf{i}_{t}*\mathbf{\tilde{c}}_{t}\right),$
$\displaystyle\mathbf{h}_{t}$
$\displaystyle=\text{tanh}\left(\mathbf{c}_{t}\right)*\mathbf{o}_{t},$
where $\mathbf{x}_{t}\in\mathbb{R}^{l}$ is the input of this timestamp of size
$l$, which is the concatenation of historical usage and the CNN output,
$\mathbf{c}_{t}$ and $\mathbf{h}_{t}$ are the cell state and hidden state of
LSTM cell at time $t$, $\mathbf{i}_{t}$, $\mathbf{f}_{t}$, $\mathbf{o}_{t}$
and $\mathbf{\tilde{c}}_{t}$ are intermediate variables of LSTM cell, and
$\mathbf{W}^{q}\in\mathbb{R}^{d\times(d+l)}$ and
$\mathbf{b}^{q}\in\mathbb{R}^{d}$ where $q=\mathbf{i},\mathbf{f},\mathbf{o}$
are trainable variables, and $d$ is the number of hidden units of LSTM cell.
#### 4.1.5. Temporal Attention
The temporal attention mechanism captures correlation between features across
different timestamps. The intuition behind this is that there is a strong
temporal dependency of the usage of a station at time $t$, for example, on the
usage at the same time one day ago. The temporal attention scores between the
current decoder LSTM timestep $t$ and one of the previous encoder LSTM hidden
states, $\lambda_{t,t^{\prime}}$, are calculated as Eq. (3) by a concatenation
manner:
(3)
$\lambda_{t,t^{\prime}}=\mathbf{v}_{a}^{\intercal}\texttt{tanh}\left(\mathbf{W}_{a}[\mathbf{h}_{t-1};\mathbf{c}_{t-1}]+\mathbf{U}_{a}\mathbf{h}_{t^{\prime}}+\mathbf{b}_{a}\right),$
where $\mathbf{v}_{a},\mathbf{b}_{a}\in\mathbb{R}^{d}$,
$\mathbf{W}_{a}\in\mathbb{R}^{d\times 2d}$ and
$\mathbf{U}_{a}\in\mathbb{R}^{d\times d}$ are learnable parameters, $d$ is the
number of hidden units of LSTM cell, $\mathbf{h}_{t}$ and $\mathbf{c}_{t}$ are
hidden state and cell state of decoder at timestamp $t$, and
$\mathbf{h}_{t^{\prime}}$ is hidden state of encoder at timestamp
$t^{\prime}$, which is in range of $[1,\mathcal{T}]$. The attention weight,
denoted as $\gamma_{t,t^{\prime}}$, is then a softmax function of
$\lambda_{t,t^{\prime}}$ as in Eq. (4):
(4)
$\gamma_{t,t^{\prime}}=\frac{\texttt{exp}\left(\lambda_{t,t^{\prime}}\right)}{\sum_{t=1}^{\mathcal{T}}\texttt{exp}\left(\lambda_{t,t^{\prime}}\right)}.$
The weighted sum of the encoder LSTM hidden states, $\mathbf{d}_{t}$, shown in
Eq. (5) concatenated with external features $\mathbf{ex_{t}}$,
$[\mathbf{d}_{t}:\mathbf{ex_{t}}]$, serves as the input of decoder:
(5)
$\mathbf{d}_{t}=\sum_{t^{\prime}=1}^{\mathcal{T}}\gamma_{t,t^{\prime}}\mathbf{h}_{t^{\prime}},$
where $\mathcal{T}$ is the number of encoder timestamps.
### 4.2. Model Training & Prediction for New Stations
As conventional practices, we handle the existing station prediction based on
the historical data available. Given absence of historical data, the
predictions of the usage for new stations are based on the model trained on
the active existing stations. To address the initialization problem of the
model, during the training process, we generate the virtual trip data based on
the weighted average of those from multiple existing stations. This is
achieved by randomly picking a batch from the samples of all the existing
stations at each training epoch. This way, the model is able to learn their
shared bike usage patterns across all the existing stations, and such
knowledge learned can be used to predict the usage of new stations,
considering that the new stations share correlated spatial and temporal usage
features of existing stations.
The model is trained upon the existing stations within each cluster. To
address the model initialization without historical data, for the new stations
we design an efficient mechanism to generate the virtual historical usage from
those of surrounding existing stations. We note that the two adjacent stations
have similar bike usage and mobility patterns because of their similar
spatial-temporal characteristics. Therefore, we generate the virtual bike
usage by the distances between stations. Specifically, we compute the
geographic similarity score between a new station, $n$, and an existing peer
in its neighborhood, $f$, based on their mutual geographic distance in km:
(6) $Sim_{n,f}=\frac{1}{\text{distance}\left(n,f\right)}.$
Then the similarity scores across all the existing stations is normalized to
find the weights for the known bike usage assigned upon each existing station:
(7) $\omega_{n,f}=\frac{Sim_{n,f}^{2}}{\sum_{f=1}^{\mathbf{F}}Sim_{n,f}^{2}}.$
The virtual bike usage for the new station at time $\tau$ based on the
historical usage of those existing stations, $L_{f,\tau}$ is finally
calculated by:
(8) $L_{n,\tau}=\sum_{f\in\mathbf{F}}^{\mathbf{F}}\omega_{n,f}L_{f,\tau},$
where $\mathbf{F}$ is the set of existing stations.
## 5\. Experimental Studies
In this section, we present the experimental results based on the datasets of
Citi (NYC), Divvy (Chicago) and Metro (LA). The experimental settings are
first introduced in Section 5.1, followed by experimental results on
existing/new stations in Section 5.2.
### 5.1. Experimental Settings
We compare AtCoR with the following baselines and state-of-art models on the
given datasets:
1. (1)
ARIMA: Auto Regressive Integrated Moving Average for time-series forecasting.
The size of sliding window is set to be 24.
2. (2)
RNN: Simple Recurrent Neural Network for time-series predictions (Pan et al.,
2019c). The length of input sequence is 24.
3. (3)
LSTM: Long Short-Term Memory neural network (Lin et al., 1996) predicts future
usage with historical data of last 24 hour.
4. (4)
GRU: Gated Recurrent Units as another recurrent neural network for time-series
predictions (Cho et al., 2014).
5. (5)
GCN: Graph Convolutional Neural Networks with Data-driven Graph Filter (Chai
et al., 2018) predicts station-level usage taking into account of station
correlations.
We are leveraging the features for previous 24 hours to predict the station-
level bike usage for the next timestamp. For the evaluation of the existing
stations for NYC and Chicago datasets, our model as well as comparison schemes
are trained based on the hourly usage data from April 11, 2019 to July 19,
2019 (2,400 hours duration), and are tested on the hourly usage data from July
20, 2019 to August 18, 2019 (720 hours duration). For LA datasets, the models
are trained based on the 4-hour bike usage data in June 2019 (720 hours
duration), and are tested on the data for the following 30 days (720 hours
duration).
For the evaluation of the new stations, we find out all the deployed stations
from May to August 2019 for NYC and Chicago, and those from June to December
2019 for LA. Our model as well as the baselines are tested on the usage data
within 4 weeks (672 hours duration) for NYC and Chicago and 2 weeks for LA
from the first appearance of the hourly usage of the stations. The stations
with sufficient bike usage are of our interest here, and we discard those with
little usage (say, less than 10 pick-ups/drop-offs per day). For new stations
in LA, due to the large sparsity in the bike usage data, we consider the
appearance of consecutive bike usage as the first usage and the start of our
evaluations. The numbers of existing/new stations studied in our experimental
evaluations are further identified and listed as follows. The numbers of
existing and new stations studied, (existing, new), for each city are as
follows: (631, 15) for NYC, (229, 13) for Chicago, (48, 7) for LA. The
existing and new stations are clustered together in the manner described in
Section 3.2.2. We train AtCoR and baselines models on all the clusters of the
three bike sharing systems. Then we evaluate models regarding their
predictions of new stations.
Figure 10. Distributions of bike trip distance (meters) for NYC, Chicago and
LA.
Figure 10 shows that a majority of bikes rented at one station in NYC, Chicago
and LA, are returned at stations beyond 500m away, demonstrating that for a
specific station there is a small number ($<20\%$) of bike transitions between
the stations within a grid of 500m $\times$ 500m and the bikes rented at this
region are mostly returned to regions somewhere else. In addition, a majority
of bike trips are within the range of 500m and 5,000m as shown in Figure 10.
Therefore, the size of station-centered heatmaps discussed in Section 3.2.1 is
set to be $11\times 11\times\mathcal{P}$, where $\mathcal{P}$ is 2 (regional
pick-ups and drop-offs) plus the total number of POIs categories (13 for NYC,
7 for Chicago and 15 for LA), and each grid of the heatmaps is a 500m $\times$
500m area on the city map. This way, the heatmaps cover most of the areas
where the riders can reach by the bikes rented from the center station.
Other model parameters of AtCoR are set as followed. The CNN component has 3
layers convolutional layers with sizes $h\times w\times c$; we set $3\times
3\times 256$, $3\times 3\times 128$ and $2\times 2\times 64$, respectively,
and relu activation function for each of them. Each of the first two
convolutional layers is followed by a maxpooling layer. The final
convolutional layer is connected with a fully connected layer which converts
the CNN’s output into the one of size 1. The number of layers of LSTM is
1,024. The dropout rate for the LSTM is 0.5. The learning rate is 0.001. The
batch size is set to be 128. Total number of training epochs is 5,000. We have
implemented AtCoR and other schemes in Python 3.7 and Tensorflow 2.1, and the
models are trained and evaluated upon a desktop server with Intel i5-8700K,
16GB RAM, Nvidia GTX 1060/1050Ti and Windows 10.
We use mean square error (MSE) as the training metric, and we evaluate model
performance as well as the results based on both MSE and mean abosolute error
(MAE):
(9) $\displaystyle
MSE=\frac{1}{M}\sum_{i}^{M}\left(y_{i}-\hat{y}_{i}\right)^{2},\quad
MAE=\frac{1}{M}\sum_{i}^{M}\left|y_{i}-\hat{y}_{i}\right|,$
where M, $y_{i}$ and $\hat{y}_{i}$ are the total number of predictions made by
the evaluated model, the ground-truth of the bike usage and the predicted bike
usage, respectively.
### 5.2. Experimental Results
#### 5.2.1. Usage Prediction for Existing Stations
First, we test our model on all active existing stations for all three
datasets. The accuracy of bike usage predictions across the existing stations
is shown in Table 2. It is shown that our model achieves overall better
accuracy compared with other baselines for the systems in NYC, Chicago and LA.
It is mainly because the station-centered heatmaps incorporate the trend that
the bike riders travel between the stations. Focusing upon each station’s
neighborhood, a heatmap characterizes the mobility patterns, which are learned
and captured by AtCoR, and thus the bike usage at the new stations with
similar neighborhood patterns can be further predicted. For Metro in LA, GCN
achieves comparable or slightly better performance like AtCoR, likely due to
the sparsity of bike station network in LA. Despite this, AtCoR has shown high
accuracy and robustness with complicated mobility patterns.
(a) A station at NYC (40.7659∘, -73.9763∘).
(b) A station at Chicago (41.8782∘, -87.6319∘).
(c) A station at LA (34.085∘,-118.259∘).
Figure 11. Hourly bike usage prediction for the existing stations in the three
datasets.
Figure 11 illustrates the bike usage (pick-ups and drop-offs) predictions for
the three existing stations in NYC, Chicago and LA. We can observe that the
predictions of AtCoR are highly accurate, and close to the ground-truth
measurements. Such accuracy can enable advanced bike station operational
applications like station demand and supply rebalancing.
Table 2. Comparison between AtCoR and other baseline and state-of-art models
for existing stations of the three datasets.
Schemes | ARIMA | RNN | GCN | LSTM | GRU | AtCoR
---|---|---|---|---|---|---
Citi | Pick-up | MAE | 2.682 | 2.345 | 2.240 | 2.164 | 2.166 | 1.771
MSE | 30.028 | 18.711 | 12.226 | 16.080 | 16.143 | 10.599
Drop-off | MAE | 2.564 | 2.265 | 2.084 | 2.100 | 2.112 | 1.730
MSE | 24.762 | 16.514 | 11.318 | 14.210 | 14.245 | 9.447
Divvy | Pick-up | MAE | 1.964 | 1.832 | 2.300 | 1.437 | 1.418 | 1.263
MSE | 23.693 | 14.866 | 6.382 | 11.293 | 10.732 | 7.687
Drop-off | MAE | 1.927 | 1.633 | 1.067 | 1.247 | 1.404 | 1.247
MSE | 21.075 | 13.584 | 6.382 | 10.937 | 10.508 | 7.980
Metro | Pick-up | MAE | 2.126 | 3.258 | 1.276 | 1.878 | 1.936 | 1.462
MSE | 8.578 | 16.071 | 4.895 | 7.471 | 7.26 | 4.879
Drop-off | MAE | 2.154 | 1.478 | 1.480 | 1.251 | 1.005 | 1.435
MSE | 8.752 | 4.535 | 3.965 | 3.393 | 2.245 | 4.610
Table 3. Performance comparison for predictions of new stations.
Schemes | RNN | LSTM | GRU | AtCoR
---|---|---|---|---
Citi | Pick-up | MAE | 2.922 | 2.760 | 2.736 | 2.495
MSE | 20.961 | 18.652 | 18.461 | 15.883
Drop-off | MAE | 2.774 | 2.728 | 2.738 | 2.568
MSE | 17.833 | 18.930 | 19.016 | 15.835
Divvy | Pick-up | MAE | 2.092 | 1.814 | 1.825 | 1.627
MSE | 23.475 | 21.089 | 21.025 | 13.282
Drop-off | MAE | 1.812 | 1.589 | 1.639 | 1.442
MSE | 14.126 | 11.491 | 12.317 | 8.223
Metro | Pick-up | MAE | 1.552 | 1.249 | 1.246 | 1.098
MSE | 3.286 | 3.425 | 3.135 | 2.439
Drop-off | MAE | 1.588 | 1.583 | 1.216 | 1.043
MSE | 3.683 | 3.662 | 2.463 | 2.795
Table 4. Comparison of performance between AtCoR w/ and w/o virtual generated
data.
Schemes | Citi | Divvy | Metro
---|---|---|---
MAE | MSE | MAE | MSE | MAE | MSE
Pick-up | AtCoR w/ virtual | 2.384 | 15.018 | 1.638 | 11.983 | 1.305 | 4.639
AtCoR w/o virtual | 2.387 | 16.661 | 1.685 | 14.528 | 1.395 | 5.014
Drop-off | AtCoR w/ virtual | 2.666 | 15.892 | 1.589 | 9.320 | 1.561 | 5.797
AtCoR w/o virtual | 2.547 | 16.510 | 1.522 | 10.300 | 1.727 | 7.929
#### 5.2.2. Bike Usage Prediction for New Stations
The predictions of bike usage for new stations can be done using the knowledge
learned from the existing stations. Given the trained models from all the
existing stations, RNN, LSTM and GRU can be directly applied to predict each
individual new stations, and therefore we focus on comparing AtCoR with the
three approaches here. We leverage the knowledge learned from the existing
stations for the predictions of new stations given the bike station network
reconfiguration. MAE and MSE of new stations’ usage predictions, pick-ups and
drop-offs, by AtCoR as well as three baselines, RNN, LSTM and GRU, are shown
in Table 3. As is shown, AtCoR outperforms those three baseline models in
predicting the usage of new stations for all three bike sharing systems.
We illustrate the performance with and without the virtual bike usage data
generated. Virtual bike usage one day ahead of the first bike usage of new
stations are generated as in Section 4.2 as the starting points. The virtual
usage provides an initial inference for the models regarding how the mobility
pattern of a new station possibly look like, enhancing the model accuracy as
is shown in Table 4. Here we only compare the MAE and MSE for predicting the
next 24 hours after the first use of the stations given the data availability.
## 6\. Discussion
Incorporating Other Information: Despite the features selected in this work as
described in Section 3.1, other features may also be correlated with bike
usage, such as events and demographic distributions (Chen et al., 2016).
However, such information is not considered due to the limit of our current
resources. Our prediction accuracy could be further increased with the
inclusion of those features in our model. Nevertheless, the generic design of
station-centered heatmaps as the feature representations in this study allows
the easy integration of other information.
Sparsity of Usage Data: Though we focus on active existing stations in this
study, a lot of stations are not so active that their bike usage is low,
especially in LA . For experimental study of LA bike usage, we consider 4-hour
time interval as one timestamp to lower the influence of data sparsity.
Predicting the bike usage with sparsity is challenging, yet important for
system management. Further study on how to deal with sparse historical data is
necessary.
Similarity Scores for Virtual Data Generation: As mentioned in Section 5.2.2,
virtual historical usage is essential for the first few predictions, and their
accuracy depends on the quality of the initial inference in addition to the
model performance. In this work, we compute the virtual historical usage based
on the mutual geometric distances between stations, which has been shown to
benefit predicting the new stations. However, further enhanced inference can
be achieved by inclusion of other information such as crowd’s awareness of new
stations, where a comprehensive mechanism is needed for future study.
## 7\. Conclusion
In this work, we propose a novel bike station prediction algorithm called
AtCoR for predicting station-based bike usage of future stations given bike
station network reconfiguration. We design novel station-centered heatmaps
which characterize for each target station centered at the heatmap the trend
that the riders travel between it and the neighboring regions, making the
common patterns of the bike station network learnable. AtCoR further leverages
such knowledge learned to predict the usage for new stations with the aid of
virtual historical usage of new stations generated according to their
correlation to the surrounding existing stations. Extensive experiment study
on the bike sharing systems of NYC, Chicago and LA shows AtCoR is capable in
predicting usage at new/future stations as well as existing stations, and
outperforms the baseline and state-of-art models in our experimental
evaluations.
## References
* (1)
* Chai et al. (2018) Di Chai, Leye Wang, and Qiang Yang. 2018. Bike flow prediction with multi-graph convolutional networks. In _Proc. ACM SIGSPATIAL_. 397–400.
* Chen et al. (2016) Longbiao Chen, Daqing Zhang, Leye Wang, Dingqi Yang, Xiaojuan Ma, Shijian Li, Zhaohui Wu, Gang Pan, Thi-Mai-Trang Nguyen, and Jérémie Jakubowicz. 2016\. Dynamic cluster-based over-demand prediction in bike sharing systems. In _Proc. ACM UbiComp_. 841–852.
* Chen et al. (2017) Min Hao Chen, Abhinav Jauhri, and John Paul Shen. 2017\. Data driven analysis of the potentials of dynamic ride pooling. In _Proc. ACM SIGSPATIAL_. 7–12.
* Chen et al. (2020) Po-Chuan Chen, He-Yen Hsieh, Kuan-Wu Su, Xanno Kharis Sigalingging, Yan-Ru Chen, and Jenq-Shiou Leu. 2020. Predicting station level demand in a bike-sharing system using recurrent neural networks. _IET Intelligent Transport Systems_ (2020).
* Cho et al. (2014) Kyunghyun Cho, Bart Van Merriënboer, Dzmitry Bahdanau, and Yoshua Bengio. 2014. On the properties of neural machine translation: Encoder-decoder approaches. _arXiv preprint arXiv:1409.1259_ (2014).
* Fadlullah et al. (2017) Zubair Md Fadlullah, Fengxiao Tang, Bomin Mao, Nei Kato, Osamu Akashi, Takeru Inoue, and Kimihiro Mizutani. 2017\. State-of-the-art deep learning: Evolving machine intelligence toward tomorrow’s intelligent network traffic control systems. _IEEE COMST_ 19, 4 (2017), 2432–2455.
* Gong et al. (2019) Yongshun Gong, Zhibin Li, Jian Zhang, Wei Liu, and Jinfeng Yi. 2019. Potential Passenger Flow Prediction: A Novel Study for Urban Transportation Development. _arXiv preprint arXiv:1912.03440_ (2019).
* He and Shin (2018) Suining He and Kang G Shin. 2018. (Re) Configuring Bike Station Network via Crowdsourced Information Fusion and Joint Optimization. In _Proc. ACM MobiHoc_. 1–10.
* He and Shin (2019a) Suining He and Kang G. Shin. 2019a. Spatio-Temporal Adaptive Pricing for Balancing Mobility-on-Demand Networks. _ACM Trans. Intell. Syst. Technol._ 10, 4, Article 39 (July 2019), 28 pages.
* He and Shin (2019b) Suining He and Kang G Shin. 2019b. Spatio-temporal capsule-based reinforcement learning for mobility-on-demand network coordination. In _Proc. WWW_. 2806–2813.
* He and Shin (2020a) Suining He and Kang G Shin. 2020a. Dynamic Flow Distribution Prediction for Urban Dockless E-Scooter Sharing Reconfiguration. In _Proc. WWW_. 133–143.
* He and Shin (2020b) Suining He and Kang G Shin. 2020b. Towards Fine-grained Flow Forecasting: A Graph Attention Approach for Bike Sharing Systems. In _Proc. WWW_. 88–98.
* Huang et al. (2014) Wenhao Huang, Guojie Song, Haikun Hong, and Kunqing Xie. 2014\. Deep architecture for traffic flow prediction: deep belief networks with multitask learning. _IEEE T-ITS_ 15, 5 (2014), 2191–2201.
* Hulot et al. (2018) Pierre Hulot, Daniel Aloise, and Sanjay Dominik Jena. 2018\. Towards station-level demand prediction for effective rebalancing in bike-sharing systems. In _Proc. ACM KDD_. 378–386.
* Jauhri et al. (2017) Abhinav Jauhri, Brian Foo, Jerome Berclaz, Chih Chi Hu, Radek Grzeszczuk, Vasu Parameswaran, and John Paul Shen. 2017. Space-time graph modeling of ride requests based on real-world data. In _Proc. AAAI_.
* Li et al. (2015) Yexin Li, Yu Zheng, Huichu Zhang, and Lei Chen. 2015\. Traffic prediction in a bike-sharing system. In _Proc. ACM SIGSPATIAL_. 33.
* Li et al. (2019) Youru Li, Zhenfeng Zhu, Deqiang Kong, Meixiang Xu, and Yao Zhao. 2019. Learning Heterogeneous Spatial-Temporal Representation for Bike-Sharing Demand Prediction. In _Proc. AAAI_ , Vol. 33. 1004–1011.
* Lin et al. (1996) Tsungnan Lin, Bill G Horne, Peter Tino, and C Lee Giles. 1996\. Learning long-term dependencies in NARX recurrent neural networks. _IEEE TNNLS_ 7, 6 (1996), 1329–1338.
* Liu et al. (2015) Junming Liu, Qiao Li, Meng Qu, Weiwei Chen, Jingyuan Yang, Hui Xiong, Hao Zhong, and Yanjie Fu. 2015\. Station site optimization in bike sharing systems. In _Proc. IEEE ICDM_. IEEE, 883–888.
* Liu et al. (2016) Junming Liu, Leilei Sun, Weiwei Chen, and Hui Xiong. 2016\. Rebalancing bike sharing systems: A multi-source data smart optimization. In _Proc. ACM KDD_. 1005–1014.
* Liu et al. (2017) Junming Liu, Leilei Sun, Qiao Li, Jingci Ming, Yanchi Liu, and Hui Xiong. 2017\. Functional zone based hierarchical demand prediction for bike system expansion. In _Proc. ACM SIGKDD_. 957–966.
* Liu et al. (2018a) Zhaoyang Liu, Yanyan Shen, and Yanmin Zhu. 2018a. Inferring dockless shared bike distribution in new cities. In _Proc. ACM WDSM_. 378–386.
* Liu et al. (2018b) Zhaoyang Liu, Yanyan Shen, and Yanmin Zhu. 2018b. Where Will Dockless Shared Bikes be Stacked? —Parking Hotspots Detection in a New City. In _Proc. ACM KDD_. 566–575.
* Lv et al. (2014) Yisheng Lv, Yanjie Duan, Wenwen Kang, Zhengxi Li, and Fei-Yue Wang. 2014. Traffic flow prediction with big data: a deep learning approach. _IEEE T-ITS_ 16, 2 (2014), 865–873.
* Ma et al. (2017) Xiaolei Ma, Zhuang Dai, Zhengbing He, Jihui Ma, Yong Wang, and Yunpeng Wang. 2017\. Learning traffic as images: a deep convolutional neural network for large-scale transportation network speed prediction. _Sensors_ 17, 4 (2017), 818.
* O’Mahony and Shmoys (2015) Eoin O’Mahony and David B Shmoys. 2015. Data analysis and optimization for (Citi)bike sharing. In _Proc. AAAI_.
* Pan et al. (2019a) Ling Pan, Qingpeng Cai, Zhixuan Fang, Pingzhong Tang, and Longbo Huang. 2019a. A deep reinforcement learning framework for rebalancing dockless bike sharing systems. In _Proc. AAAI_ , Vol. 33. 1393–1400.
* Pan et al. (2019c) Yan Pan, Ray Chen Zheng, Jiaxi Zhang, and Xin Yao. 2019c. Predicting bike sharing demand using recurrent neural networks. _Procedia Computer Science_ 147 (2019), 562–566.
* Pan et al. (2019b) Zheyi Pan, Yuxuan Liang, Weifeng Wang, Yong Yu, Yu Zheng, and Junbo Zhang. 2019b. Urban traffic prediction from spatio-temporal data using deep meta learning. In _Proc. ACM KDD_. 1720–1730.
* Tang et al. (2019) Xianfeng Tang, Boqing Gong, Yanwei Yu, Huaxiu Yao, Yandong Li, Haiyong Xie, and Xiaoyu Wang. 2019. Joint modeling of dense and incomplete trajectories for citywide traffic volume inference. In _Proc. WWW_. 1806–1817.
* Wang et al. (2018) Shuai Wang, Tian He, Desheng Zhang, Yuanchao Shu, Yunhuai Liu, Yu Gu, Cong Liu, Haengju Lee, and Sang H Son. 2018\. BRAVO: Improving the rebalancing operation in bike sharing with rebalancing range prediction. _Proc. ACM IMWUT_ 2, 1 (2018), 1–22.
* Yao et al. (2019) Huaxiu Yao, Yiding Liu, Ying Wei, Xianfeng Tang, and Zhenhui Li. 2019. Learning from multiple cities: A meta-learning approach for spatial-temporal prediction. In _Proc. WWW_. 2181–2191.
* Yu et al. (2017) Bing Yu, Haoteng Yin, and Zhanxing Zhu. 2017. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. _arXiv preprint arXiv:1709.04875_ (2017).
* Zhang et al. (2016) Jiawei Zhang, Xiao Pan, Moyin Li, and Philip S Yu. 2016\. Bicycle-sharing systems expansion: station re-deployment through crowd planning. In _Proc. ACM SIGSPATIAL_. 1–10.
* Zhang and Philip (2016) Jiawei Zhang and S Yu Philip. 2016. Trip route planning for bicycle-sharing systems. In _Proc. IEEE CIC_. IEEE, 381–390.
* Zhou (2015) Xiaolu Zhou. 2015\. Understanding spatiotemporal patterns of biking behavior by analyzing massive bike sharing data in Chicago. _PloS One_ 10, 10 (2015), e0137922.
|
# Ion-driven destabilization of a toroidal electron plasma - A 3D3VPIC
Simulation
S. Khamaru Institute for Plasma Research, Bhat, Gandhinagar 382428, India
<EMAIL_ADDRESS>Homi Bhabha National Institute, Training School
Complex, Anushaktinagar, Mumbai 400094, India R. Ganesh Institute for Plasma
Research, Bhat, Gandhinagar 382428, India Homi Bhabha National Institute,
Training School Complex, Anushaktinagar, Mumbai 400094, India M. Sengupta
Lawrence Livermore National Laboratory, Livermore, California, 94551, USA
###### Abstract
Ion resonance instability of toroidal electron plasmas in a tight aspect ratio
axisymmetric toroidal device is reported for ${Ar}^{+}$ ions of different
initial density values using a high fidelity 3D3V PIC solver. Stability of a
recently discovered quiescent quasi-steady state (QQS) of a toroidal electron
plasma obtained from “seed” solution as a result of entropy extremization at
zero inertia, is addressed to the presence of a small ion population. An ion
fraction ($f$) and corresponding number of secondary electrons are preloaded
into the system after the electron plasma attains a QQS state. Driven by the
ions, the electron plasma exhibits destabilized “center of charge motion” ($m$
$=$ $1$) along with increased poloidal mode coupling ($m$ $=$ $1$ to $9$) with
dominant $m$ $=$ $2$ mode. The growth in wall probe current is algebraic in
nature and increases for $f$ $\geq$ $0.005$, showing saturation at later time.
Higher values of ion temperatures than the electron temperatures indicate a
resonant energy transfer from electron plasma to ions via ion-resonance and
concomitant ion heating. The volume averaged temperature value of the electron
plasma rises with simulation time, attaining a quasi-steady nature near the
end of the simulation time. As can be expected from conservation of adiabatic
invariants, the flux tube averaged electron temperatures along parallel and
perpendicular directions are found to scale as $1/R^{2}$ and $1/R$
respectively, where $R$ is the major radial variable, though the plasma is
nearly collision-less.
††preprint: AIP/123-QED
## I Introduction
Instability induced confinement issues in charged particle traps are
ubiquitous and effective mitigation of such issues requires knowledge of the
underlying fundamental processes, which in turn often result in further
technical improvement of the device. Besides providing useful insight into
several fundamental physical processes, these particle trapping devices become
very useful in various applications, as discussed next.
External electric field and homogeneous magnetic field aided linear traps such
as Penning-MalmbergMalmberg and deGrassie (1975) (PM) trap, Paul trapPaul
(1990) and some modified PM trapsHurst _et al._ (2014); Yamada and Himura
(2016); Lane and Orperformedz (2019) are used extensively to trap pure
electron plasmaDavidson and Felice (1998), pure ion plasmaO’Neil and Dubin
(1998); Landa (2019) and unconventional charge species such as positron
plasmaGreaves and Surko (2000), anti-protron plasma.Ahmadi and et al. (2016);
Ahmadi _et al._ (2018) In straight cylindrical PM trap, radial and axial
confinement are provided to the pure electron (or ion) plasma, with the aid of
an axial homogeneous magnetic field and electrostatic end plugs, respectively.
The result achieved in such an arrangement is so robust that the electron
plasma plasma can be confined for a long time (over several days) in near
absolute equilibrium stateO’Neil and Driscoll (1979); Malmberg and Driscoll
(1980); Driscoll, Malmberg, and Fine (1988); Dubin and O’Neil (1999), allowing
numerous experiments to be performed. For example, pure electron plasmas share
fundamental properties with inviscid 2D Euler fluids at low density limit
($m\rightarrow 0$) and hence are often investigated experimentallyFine _et
al._ (1995) to understand complex nonlinear dynamics of such fluids. These
findings often corroborate with the corresponding theoreticalBriggs,
Daugherty, and Levy (1970) and numericalGanesh and Pahari (2006); Rome _et
al._ (2000) studies. Extensive numerical studiesSengupta and Ganesh (2014)
using two dimensional Particle-in-Cell (PIC)Birdsall and Langdon (2004)
simulation also provide essential insight into two dimensional electron plasma
dynamics in the cylindrical cross-section of the PM trap. Experimental
applications of pure electron plasmas are found in high-precision measurement
techniquesNiemann _et al._ (2019); Schuh _et al._ (2019), mass
spectrometryGutiérrez _et al._ (2019); Peurrung, Kouzes, and Barlow (1996)
etc. Positron, antiprotons are confined in separate PM traps with modified
electric fields and axial magnetic field, to synthesize antihydrogen used in
anti-matter investigationsAhmadi and et al. (2016); Ahmadi _et al._ (2018);
Fajans and Surko (2020). Linear Paul trapsPaul (1990); Landa (2019) are mainly
motivated to trap low density ions using time varying RF fields and axial
magnetic fields, whereas stabilization of the ions is performed with laser
coolingMorigi _et al._ (1999). Such stable ions are used in quantum
information processingMorigi and Walther (2001), mass spectrometryDouglas,
Frank, and Mao (2005) etc. Stable pair plasmas such as positron-electron,
found in astrophysical environment, are studied experimentally using combined
Penning-Paul trapGreaves and Surko (2002), inhomogeneous magnetic field in
magnetic mirror configurationHigaki _et al._ (2017) and recently reported
numerical experimentHicks, Bowman, and Godden (2019) with RF multipole
electric field. In a recent experiment, stable electron-ion coexistence is
achieved in a PM likeNakajima, Himura, and Sanpei (2021) trap under certain
conditions.
In contrast to the above discussed devices, toroidal trapsJanes _et al._
(1966); Daugherty, Eninger, and Janes (1969); Clark _et al._ (1976); Zaveri
_et al._ (1992); Khirwadkar _et al._ (1993); Lachhvani _et al._ (2017);
Stoneking _et al._ (2004) of different aspect ratios ($a/b$, $a$ is the major
radius, $b$ is the minor radius) apply an inhomogeneous external magnetic
field with inherent curvature to confine an electron plasma. The electron
plasma goes through rotational cross-field $\bf{E}\times\bf{B}$ drift where
$\bf{E}$ field is the self electric field of the plasma and $\bf{B}$ is the
external toroidal field. Because of the strong rotational $\bf{E}\times\bf{B}$
drift, electron plasma avoids the vertical loss of particles due to
${\bf{\nabla}}B$ and curvature drift. Additionally, the electron plasma
undergoes toroidal $m=1$ DiocotronDavidson (1990) rotation or center of charge
dynamics, which is similar to the Diocotron motion of cylindrical electron
plasma but due to magnetic field curvature are strongly coupled to elliptical
(i.e. $m=2$) mode and often to other higher order poloidal modes (i.e.
$m=3,4,5$ etc)Khamaru, Sengupta, and Ganesh (2019) depending on the aspect
ratio of the device and the amplitude of the $m=1$ mode. Apart from toroidal
magnetic field, electron plasma is also confined using magnetic surfaces in
stellaratorMarksteiner _et al._ (2008) and dipole magnetic fieldSaitoh _et
al._ (2010) based devices.
Earliest experiments of trapped electron plasmas in toroidal geometry were
performed by Janes et al. ($a/b$ $\sim$ 5)Janes _et al._ (1966), Daugherty et
al. ($a/b$ $\sim$ 4.6)Daugherty, Eninger, and Janes (1969), where the
axisymmetric toroidal device was proposed to be used as heavy ion plasma
accelerator (HIPAC) producing highly stripped heavy ions. The device was
intended to utilize the potential well of the electron plasma to trap and
accelerate heavy ions. Subsequent experiments were performed to confine
electron plasma in axisymmetric toroidal geometry of different aspect ratios
($a/b$ $\sim$ 6.2Clark _et al._ (1976) and $\sim$ 1.5Zaveri _et al._ (1992);
Khirwadkar _et al._ (1993)), followed by a number of experiments in partial
toroidal geometry: small aspect ratio device SMARTEX-C ($a/b$ $\sim$
1.6)Lachhvani _et al._ (2017), large aspect ratio device LNT II ($a/b$ $\sim$
13.7)Stoneking _et al._ (2004).
Such an electron plasma in inhomogeneous magnetic field is useful to
understand nonlinear couplingKhamaru, Sengupta, and Ganesh (2019); Khamaru,
Ganesh, and Sengupta (2021a, b) between different toroidal Diocotron modes,
collisional/non-collisional transport mechanismsCrooks and O’Neil (1996) of
electron plasma energy and also crucial to initiate an ongoing construction of
electron-positron plasma experiment using dipole magnetic fieldStoneking _et
al._ (2020); Singer _et al._ (2021). These systems can often serve as test-
beds to understand plasma dynamics in curved magnetic field which are often
found in astrophysical phenomenaBrambilla _et al._ (2018).
As indicated before, near-absolute equilibria of the electron plasma in PM
traps with uniform magnetic fields allows various experiments to be performed
in the device. Such near-absolute equilibrium in toroidal trap was recently
addressed and a quiescent quasi-steady (QQS)Khamaru, Ganesh, and Sengupta
(2021a, b) toroidal electron plasma has been found in a numerical experiment
with similar parametric space of that of an existing small aspect ratio
torusLachhvani _et al._ (2017). This finding also suggests the possibility of
realizing such a quiescent pure ion plasma in toroidal equilibrium, which
might provide excellent testbed/new directions for quantum information
processingLanda (2019); Morigi and Walther (2001) in future.
Coming back to electron plasmas, trapped pure electron plasmas are prone to
loss of electrons due to various instabilities associated with the electron
plasma dynamics. These instabilities are mainly resistive wall
instabilityWhite, Malmberg, and Driscoll (1982), electrostatic instability
induced by electron-neutral collisionsDavidson and Chao (1996) and ion
resonance instabilityLevy, Daugherty, and Buneman (1969); Davidson and Uhm
(1977); Fajans (1993); Peurrung, Notte, and Fajans (1993). Most of these
instabilities are subdued in devices with confining walls made of low
resistivity materials and better vacuum conditions (even at low working
pressure $\sim$ $10^{-9}$ torrLachhvani _et al._ (2016) or better). Amongst
these instabilities, ion resonance instability in cylindrical traps as well as
in toroidal traps, is more frequent and predominant over the rest. Even in the
low background pressure $\sim$ $10^{-9}$ torr, residual gases (mainly $N_{2}$
and $H_{2}$) remain in the system. Through the process of electron-impact
ionization of these background natural gas, ions and secondary electrons are
generated. It is believed that the ions are trapped in the device by the
potential well of the electron plasma leading to ion resonance instability,
inciting the destabilization of the electron plasma by the onset of an $m=1$
like toroidal Diocotron instability. To illuminate the underlying processes,
various analytical models of this instability that were constructed in the
past should be addressed first.
Initial analytical model of the cylindrical electron plasma in the presence of
few “trapped” ions was proposed by Levy et al.Levy, Daugherty, and Buneman
(1969) suggesting an exponential growth (i.e. mode amplitude is $\propto$
$exp(\gamma t)$, with $\gamma>0$) of the $m=1$ Diocotron mode of the electron
plasma. It was shown that the exact resonance condition arises if the
oscillation frequency of the ion trapped in the potential well of the electron
plasma matches the Diocotron frequency of the off centered electron plasma. In
another simplified linear model developed by Davidson and UhmDavidson and Uhm
(1977) for infinitely long cylindrical electron plasma under weak
neutralization by “trapped” ions, collisionless rotating two stream
instability between electron and ion plasma approach was adopted. The results
of this model are found to be in good agreement with those proposed by Levy el
al. showing the $m=1$ Diocotron mode of the electron plasma as fastest growing
mode among other exponentially growing modes. Later, for a 3D straight
cylindrical trap (i.e. PM trap) with electrostatic end plugs, ion resonance
instability of the cylindrical electron plasma in the presence of axially
drifting ions or “transient” ions (due to rapid loss of ions in the axial
direction as compared to an infinitely long cylinder) was investigated
analytically by FajansFajans (1993). This theory suggested algebraic growth of
the Diocotron instability, which was also reported to be confirmed in the
experiment performed by Peurrung et al.Peurrung, Notte, and Fajans (1993)
under similar experimental conditions. Earlier, simulation based studies using
2D Particle-in-Cell(PIC) code PEC2PICSengupta and Ganesh (2015, 2016, 2017),
have been performed addressing the ion resonance instability of cylindrically
confined partially neutralized electron plasma, under different initial
conditions. Through collisional and collisionless processes, the growth rate
of different Diocotron modes were addressed where the results were found to be
in good agreement with the Davidson and Uhm model. It was also shown that the
non-ionizing collisions do not initiate instabilities in contrary to then
existing theorySengupta and Ganesh (2016).
As of now, numerous experiments of cylindrical electron plasmasPeurrung,
Notte, and Fajans (1993); Eckhouse, Fisher, and Rostoker (1981); Bettega _et
al._ (2006); Kabantsev and Driscoll (2007) have exhibited ion resonance
instability under different configurations. These experiments show
exponentialEckhouse, Fisher, and Rostoker (1981); Kabantsev and Driscoll
(2007) or algebraic growthPeurrung, Notte, and Fajans (1993); Bettega _et
al._ (2006) depending on the nature of the ion entrapment. A toroidal electron
plasma, is also greatly affected by the presence of ions in the
systemLachhvani _et al._ (2016); Stoneking _et al._ (2002); Marksteiner _et
al._ (2008). Extensive investigation of the dependency of the growth on ion
species, magnetic field strength and pressure has been performed in
stellaratorMarksteiner _et al._ (2008) trap under magnetic surface trapping.
In SMARTEX-CLachhvani _et al._ (2016) trap, it is reported that the improved
vacuum condition of the device resulted in the damping of the growth of the
instability.
Recently, collisionless dynamics of toroidal electron plasma has been
addressed in tight aspect ratio device (in a similar parametric space as
SMARTEX-C) using the 3D3V Particle-in-Cell code PEC3PICKhamaru, Sengupta, and
Ganesh (2019), for both axisymmetricKhamaru, Sengupta, and Ganesh (2019);
Khamaru, Ganesh, and Sengupta (2021a, b) and partialSengupta, Khamaru, and
Ganesh (2021) or nonaxisymmetric configurations with electrostatic end-plugs
along the toroidal direction. Of late, the existence of a quiescent quasi-
steady state (QQS)Khamaru, Ganesh, and Sengupta (2021a, b) has been
discovered, for the toroidal electron plasma in axisymmetric configuration
using a maximum-entropy based zero-inertia solution as an initial condition to
the PEC3PIC, which then solves the full inertia-based equation of motion of
electrons. This slowly evolving state is shown to exhibit nearly free of
center of charge motion of the naturally shaped toroidal electron plasma with
finite parallel and perpendicular temperatures. We believe that the discovery
of such a quiescent quasi-steady state of the toroidal electron plasma may
become crucial to the future of quantum information processingLanda (2019);
Morigi and Walther (2001) and quantum computing. As a subsequent step,
collisionless ion-driven destabilization of the toroidal electron in QQS state
has been investigated in this paper. In an ongoing study, impact ionization
and various collisional mechanisms between electron-neutral, ion-neutral
collisions have been considered and shall be reported in future.
In the present study, we have addressed the effect of ion resonance
instability on the toroidal electron plasma in QQS state, via preloaded
$Ar^{+}$ ion plasma for different fractional neutralizationDavidson (1990)
values $f$ $=$ $n_{i}/n_{e}$ i.e. the ratio of ion density to electron
density, in an axisymmetric tight aspect ratio device using PEC3PICKhamaru,
Sengupta, and Ganesh (2019); Khamaru, Ganesh, and Sengupta (2021a, b);
Sengupta, Khamaru, and Ganesh (2021). It is shown that the electron plasma is
destabilized and the wall probe current shows algebraic growth. It is found
that the growth rate increases with increasing values of $f$. Along with
“center of charge motion” ($m$ $=$ $1$) of the electron plasma, mode coupling
between toroidal Diocotron modes ($m$ $=$ $1$ to $9$) has been observed with
$m$ $=$ $2$ as the dominant mode. The growth of the wall probe current later
saturates as ion (and electron) losses increase with simulation time. It is
shown that the calculated value of ion plasma rotational frequency of the
$Ar^{+}$ ion plasma in cylindrical approximation is at $\sim 31\%$ deviation
from the toroidal $m=1$ DiocotronDavidson (1990) frequency of the electron
plasma. Destabilization of the electron plasma and algebraic growth of the
wall probe current is brought out unambiguously. It has been observed that the
ion resonance instability of the electron plasma is possible for a certain
range of the ion plasma rotational frequency (via adding $N^{+}$ and $H^{+}$
ions into the system and observing the electron plasma dynamics, though the
results are not shown here). Collision-less heating of the ions has been
observed resulting in high ion temperature values, which is interpreted as due
to a resonant transfer of energy of the electron plasma to ions via toroidal
ion-resonance instability.
In Sec. II, the toroidal device configuration is given and the initial
conditions including the initialization steps of electron/ion plasma loading
in the toroidal device are explained. Sec. III various diagnostics implemented
for this study are shown. In Sec. IV the conclusion of the present study is
described.
## II Initial condition and simulation parameters
As discussed previously, 3D3V particle-in-cellBirdsall and Langdon (2004) code
PEC3PICKhamaru, Sengupta, and Ganesh (2019); Khamaru, Ganesh, and Sengupta
(2021a) has been used in the present study, addressing the dynamics of
toroidal electron plasma in tight aspect ratio axisymmetric toroidal device
(in a similar parametric space as SMARTEX-CLachhvani _et al._ (2017)).
Correct loading of the electron and ion plasma in this device is crucial for
the desired experiments and needs careful study along with corresponding
simulation parameters used in the study. In the next Subsec. II.1, the device
parameters and initialization process of the plasmas are explained in details.
### II.1 Device parameters and initialization of the electron and ion plasma
Figure 1: A schematic diagram of the axisymmetric toroidal vessel with
conducting boundary walls. The radius of the inner wall boundary is $R_{1}$
$=$ $0.05~{}m$, outer wall is $R_{2}$ $=$ $0.22~{}m$ (aspect ratio $\sim$ 1.6)
and the height is $L_{z}$ $=$ $0.32~{}m$. The vessel is closed by upper and
lower boundary walls. Wall probes located near the inner boundary wall,
separated by $\theta$ $=$ $\pi/2$ from each other are shown at $L_{z}$ $=$
$0.16~{}m$ mid-plane. The device sizes described are identical to SMARTEX-
CLachhvani _et al._ (2017) device. However unlike the experimental device,
here electrostatic endplugs in the toroidal direction are absent, though the
particle dynamics is fully 3D3V.
The schematic diagram of the axisymmetric toroidal vessel used in this study
is presented in Fig. 1 in cylindrical geometry. Inner wall radius of the torus
is $R_{1}$ $=$ $0.05~{}m$ and the outer wall radius is $R_{2}$ $=$ $0.22~{}m$,
making the device aspect ratio $\sim$ $1.6$. Height of the torus is $L_{z}$
$=$ $0.32~{}m$. The vessel is closed by upper and lower boundary walls. The
boundary walls of the torus are conducting. Wall probes are located near the
inner boundary wall, separated by $\theta$ $=$ $\pi/2$ locations at $L_{z}$
$=$ $0.16~{}m$ mid-plane are shown. The device size is identical to that of
SMARTEX-CLachhvani _et al._ (2017).
At background pressure of $10^{-9}$ to $10^{-7}$ torrMarksteiner _et al._
(2008); Lachhvani _et al._ (2016), ions and secondary electrons are generated
by the energetic primary electrons due to ionization of the background neutral
atoms. In typical tight aspect ratio toroidal experimentsLachhvani _et al._
(2016) with pressure $10^{-8}$ torr, the ion build up is upto $\sim$ $2.5\%$
within $5\times 10^{-4}$ sec time span. To investigate the effects of low
density ion population on the primary electron plasma in our study, the
primary electron plasma, $Ar^{+}$ ion plasma along with secondary electron
plasma are loaded in the system in two consecutive steps, described next in
detail.
Figure 2: Initial density distribution of the primary electrons in the 3D
toroidal vessel is shown at $\tau$ $=$ $0.00$, where $\tau$ $=$
$t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal Diocotron
periodKhamaru, Sengupta, and Ganesh (2019) ($\sim$ $1.8\times 10^{-6}s$) for
QQS stateKhamaru, Ganesh, and Sengupta (2021a, b). Only $3/4th$ of the
toroidal direction is shown to present the poloidal density contours along
with the toroidal density contours. The initial density of the electron plasma
has the radial variation of two orders of magnitude, viz $\sim 7.1\times
10^{12}m^{-3}-2.485\times 10^{14}m^{-3}$ i.e. $f_{b}\sim 0.002-0.07$, where
$f_{b}\sim 0.07$ corresponds to the mode (the locally maximum value of the
density) of the total density distribution at $z$ $=$ $L_{z}/2$ plane
($0.16~{}m$) of the torus. Few electrons are present outside the contour
region making the density negligible in that area. In the absence of ions,
this QQS state is shown to suffer very little radial transportKhamaru, Ganesh,
and Sengupta (2021a, b). Figure 3: (a)The density distributions of primary
electron plasma at $\tau$ $=$ $11.11$, where $\tau$ $=$ $t/\overline{t}_{D}$
and $\overline{t}_{D}$ is toroidal Diocotron period for the QQS state. The
density variation of the electron plasma is in the range $\sim 6.83\times
10^{12}m^{-3}-2.39\times 10^{14}m^{-3}$ i.e. $f_{b}\sim 0.0019-0.067$. Here
$f_{b}\sim 0.067$ corresponds to the mode of the total electron density
distribution at $z$ $=$ $L_{z}/2$ plane of the torus. (b) The density
distribution of $Ar^{+}$ ion plasma at $\tau$ $=$ $11.11$ for fractional
neutralization factor $f$ $=$ $0.007$ which corresponds to the ratio of ion
density to the electron density i.e. $n_{i}/n_{e}$. The density of the ion
plasma is in the range $\sim 4.78\times 10^{10}m^{-3}-1.67\times
10^{12}m^{-3}$. Here $1.67\times 10^{12}m^{-3}$ corresponds to the mode of the
total ion density distribution at $z$ $=$ $L_{z}/2$ plane of the torus.
As described in Ref.45Khamaru, Ganesh, and Sengupta (2021a), in the first
step, zero-inertia maximum entropy density distribution functionKhamaru,
Ganesh, and Sengupta (2021a) is used as the “seed” particle density profile of
the primary electron plasma in the system at the beginning of the simulation
which has been previously shown to result in a novel quiescent quasi-steady
state (QQS)Khamaru, Ganesh, and Sengupta (2021a) of the electron plasma. Thus
the initial spatial density distribution of the electron plasma, following
this initial “seed” solution, is shown to evolve into an axisymmetric
($\partial/\partial\theta=0$) nonuniform toroidal plasma plasma which now is
accurate to all orders in ${\rho_{{Le}}}/L$, where ${\rho_{{Le}}}$ is the
average electron Larmor radius and $L$ is a typical magnetic field gradient
length scale. Corresponding density distribution of the electrons in the 3D
toroidal vessel is shown in the Fig. 2 at $\tau$ $=$ $0.00$, where $\tau$ $=$
$t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal Diocotron
periodKhamaru, Sengupta, and Ganesh (2019) $\sim$ $1.8\times 10^{-6}s$ for the
QQSKhamaru, Ganesh, and Sengupta (2021a, b) state of the electron plasma. The
method to obtain $\overline{t}_{D}$ was elaborated in a previous studyKhamaru,
Sengupta, and Ganesh (2019). The quantity toroidal Diocotron period is
explained later in Sec. III.
The electrons are cold loaded with zero velocities. This method of initial
loading is in general different than the conventional electron injection
technique used in typical toroidal pure electron plasma experimentsLachhvani
_et al._ (2017); Stoneking _et al._ (2004), where a negatively biased
injector grid filament emit the electrons in toroidal direction, filling the
toroidal vessel via another positive biased grid placed “nearby”, followed by
axial confinement of electrons by negative end plugs. This procedure results
in a toroidal or parallel velocity such that toroidal or parallel bounce time
is much smaller than the typical $\bf{E}\times\bf{B}$ time scale. The parallel
kinetic energy of electrons often ionize the weak background neutrals, giving
rise to ions. In the present study, zero-inertia maximum entropy distribution
results in nonuniform density distribution of the electrons featuring
contoured structures with peaked centers in toroidal and poloidal planes, as
shown in the Fig. 2. Though the electrons fill up the whole torus, only
$3/4th$ of the whole torus is shown in this figure (Fig. 2) to present the
poloidal density contours along with the toroidal density contours. This
configuration of the density distribution will be used throughout this paper.
The initial density of the electron plasma is in the range $\sim 7.1\times
10^{12}m^{-3}-2.485\times 10^{14}m^{-3}$ i.e. $f_{b}\sim 0.002-0.07$. Here
$f_{b}$ is the “cylindrical” Brillouin fractionDavidson (1990) ratio,
$f_{b}=2{\omega_{p}}^{2}/{\omega_{c}}^{2}$, ($\omega_{p}$ is the plasma
frequency and $\omega_{c}$ is the cyclotron frequency for magnetic field
$B_{0}$ ($0.03T$) at inner wall radius). (Inferred range of $f_{b}$ is in the
range $f_{b}\sim 10^{-3}-10^{-4}$ in the experimentsLachhvani _et al._
(2016), which is typically an order of magnitude lower than our simulation
parameters.) As there are very few electrons in the outer regions of the
plasma, the density contours are seen to sharply fall to zero in this region.
The electron plasma attains QQS state shortly after the initialization and is
left to evolve upto $\tau$ $=$ $11.11$.
In the second step, the $Ar^{+}$ ions are introduced in the system at $\tau$
$=$ $11.11$ (which corresponds to $t$ $=$ $2\times 10^{-5}$ s) according to
same spatial distribution of QQS as the primary electrons have evolved into.
Though at $\tau$ $=$ $11.11$ the electrons acquire certain finite width of
velocity distribution function, ions are loaded cold at this instant. In
cylindrical experimentsPeurrung, Notte, and Fajans (1993); Bettega _et al._
(2006); Kabantsev and Driscoll (2007) in PM traps with ion resonance
instability, the values of the fractional neutralization factorDavidson (1990)
$f$, which corresponds to the fraction of ion population to the primary
electron population i.e. $n_{i}/n_{e}$, is reported as typically $\sim$
$10^{-5}$ to $10^{-4}$. For partial toroidal experimentsMarksteiner _et al._
(2008); Lachhvani _et al._ (2016), $f$ $\sim$ $10^{-2}$ to $10^{-1}$. In our
axisymmetric toroidal simulation, $f$ values are chosen between $0.001$ and
$0.01$, which is in the range of experimental estimates. The 3D density
distributions of primary electron plasma and $Ar^{+}$ ion plasma at $\tau$ $=$
$11.11$, is shown in the Fig. 3 (a) and (b) respectively for $f$ $=$ $0.007$.
In Fig. 3 (a), the density of the electron plasma is in the range $\sim
6.83\times 10^{12}m^{-3}-2.39\times 10^{14}m^{-3}$ i.e. $f_{b}\sim
0.0019-0.067$. Fig. 3 (b), the density of the ion plasma is in the range $\sim
4.78\times 10^{10}m^{-3}-1.67\times 10^{12}m^{-3}$. The ion densities used are
two orders of magnitude smaller than the electron density at any given spatial
location.
Figure 4: Inner wall probe current for fractional neutralization factor $f$
(i.e. $n_{i}/n_{e}$) ranging from $0.001$ to $0.01$ for $Ar^{+}$ upto $\tau$
$=$ $34$, where $\tau$ $=$ $t/\overline{t}_{D}$ and $\overline{t}_{D}$ is
toroidal Diocotron period for QQS state. The time of introduction of ions is
at $\tau$ $=$ $11.11$, as indicated by black dotted line. The inset plot shows
onset of the growth of wall probe current amplitude.
Along with the ions, cold (zero velocity) secondary electrons are also loaded
at $\tau$ $=$ $11.11$ with density value $f$ times the density of the primary
electrons i.e. the density of secondary electron plasma is same as the ion
plasma loaded. This method mimics the ionization process in the experiments
and conserves the total charge of the system at $\tau$ $=$ $11.11$. Total
energy of the system is not effected by addition of the secondary electron
population/ion population because of the low density $\sim$ $10^{-3}$ to
$10^{-2}$. After the loading processes, this nonneutral plasma consisting of
the primary and secondary electrons along with ions is left to evolve and the
dynamics is monitored through the simulation time period upto $34$
$\overline{t}_{D}$. In the next Subsec. II.2, the simulation parameters are
described.
### II.2 Simulation Parameters
In this study, 3D3V OPEN-MP parallelized particle-in-cellBirdsall and Langdon
(2004) (PIC) code PEC3PICKhamaru, Sengupta, and Ganesh (2019); Khamaru,
Ganesh, and Sengupta (2021a) is used. More details on the code and the
algorithm can be found in the recent QQS studyKhamaru, Ganesh, and Sengupta
(2021a) of electron plasmas where same PIC code has been used. Here initial
super particle number for the primary electron plasma is 2560640 in a grid
size of $192\times 192\times 192$ in Cartesian coordinates. At $\tau$ $=$
$11.11$, super particle number for the primary electron plasma becomes 2464401
due to initial loss of the electrons. Also initial super particle numbers for
the ion plasma and secondary electron plasma are 2464401 each at $\tau$ $=$
$11.11$. The density values of the ion plasma and secondary electron plasma
are made $f$ times lesser than the primary electron plasma using the
fractional neutralization factor $f$ $=$ $0.007$ in the code. At $\tau$ $=$
$11.11$, most of the electron and ion particles are confined in toroidal tube
with high density value which has lesser volume than the whole torus, as shown
in the Fig. 3. Thus the effective number of particles (electrons and ions) per
3D cell of this toroidal tube is $\sim 10$ and the number of particles per 3D
cell of the whole toroidal simulation device is $\sim 3.4$. Numerical
conservation tests for the current PEC3PIC solver has been performed in a
previous studyKhamaru, Sengupta, and Ganesh (2019). $Ar^{+}$ ion used in the
study have mass $72820.77$ $m_{e}$ where $m_{e}$ is the electron mass.
The toroidal magnetic field used in this device is spatially inhomogeneous. At
inner wall radius $R_{1}$, the magnitude of the magnetic field is highest
($B_{0}=0.03T$) and gradually falls with cylindrical radial coordinate $R$ as
$B={B_{0}R_{1}}/R$. At $R=R_{1}$ the estimated electron cyclotron time period
$T_{ce}=1.19\times 10^{-9}s$ and the ion cyclotron time period
$T_{ci}=8.68\times 10^{-5}s$. As is obvious, the cyclotron periods are
smallest at the largest magnetic field in the system, which is at the inner
wall at $R=R_{1}$. The simulation time step chosen is $10^{-10}s$ which is
about 10 times smaller than the smallest cyclotron time in the system.
The parameters used here are typical of recent experiments Lachhvani _et al._
(2016), though any collisional processes are not included in the present
study. As neutrals are not present in our simulation, collisional processes
with neutrals are exempted as well. Ion-ion collision are also exempted
because of their mass and extremely low density. Other possible processes are
the electron-electron and electron-ion collisions. Here we have estimated the
respective collisional parameters for electron-electron collision and
electron-ion collision (for $Ar^{+}$), $f=0.007$ case as follows: Large angle
electron-electron Coulomb collisions time scaleHuba (2013) ${\tau}_{col}^{e}$
$\simeq$ $\frac{{{T_{e}}^{3/2}}}{2.91\times 10^{-6}~{}n_{e}ln\Gamma}$ sec
$\simeq$ $0.099$ sec i.e. $\overline{\tau}_{col}^{e}$ $\simeq$
${{\tau}_{col}^{e}}/{\overline{t}_{D}}$ $\simeq$ $5.5$ $\times$ $10^{4}$.
Electron-ion Coulomb collisions time scale ${\tau}_{col}^{i}$ $\simeq$
$\frac{{{T_{i}}^{3/2}}}{4.8\times 10^{-8}~{}n_{i}\mu Z_{a}^{4}ln\Gamma}$ sec
$\simeq$ $141.44$ sec i.e. $\overline{\tau}_{col}^{i}$ $\simeq$ $7.86$
$\times$ $10^{7}$. (In the formulae for collision times, temperature is in eV
units, whereas all the other quantities are in CGS units). Consequently,
${\overline{t}_{D}}\sim 1.8\times 10^{-6}s$ is the toroidal Diocotron time
period, $ln\Gamma$ $\sim$ 10, $T_{max}^{e}$ $\sim$ $80.0~{}eV$, $n_{e}$ $\sim$
$2.485\times 10^{8}{cm}^{-3}$, $n_{i}$ $\sim$ $1.74\times 10^{6}{cm}^{-3}$,
$T_{max}^{i}$ $\sim$ $280.0~{}eV$, atomic number $Z_{a}$ $=$ $1$, ratio of ion
mass to proton mass $\mu$ $=$ $\frac{m_{i}}{m_{p}}$ $=$ 39.66. In the present
study, total simulation time is $\tau$ $=$ $t/\overline{t}_{D}$ $=$ $34$,
where $t$ is time. Both electron-electron Coulomb collisions time scale and
electron-ion collisions time scale are larger than the present simulation time
scale. Thus the plasma is considered as collision-less plasma.
Figure 5: Time evolution of primary electron plasma and ion plasma with
respective density values for $f$ $=$ $0.007$ at different simulation time
periods $\tau$ $=$ $16.67$, $21.11$, $23.33$ and $25.56$, where $\tau$ $=$
$t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal Diocotron period for
QQS state. The density contours of electron plasma along with the spatial
position of the plasma is shown in (a)-(d). These figures show displaced
electron plasma with “center of charge motion” ($m$ $=$ $1$ mode) in the
vertical and horizontal direction, with distinguishable compressed and
expanded forms of the plasma along with periodic peak density value evolution.
The time evolution of the ion plasma and density is shown in (e)-(h). (e) at
$\tau$ $=$ $16.67$ shows that the initially cold loaded ion plasma starts to
gain energy from the system and initial shape of the plasma (Fig. 3 (b) and
(d)) at $\tau$ $=$ $11.11$) is distorted. In (f)-(h), at $\tau$ $=$ $21.11$,
$23.33$, $25.56$ respectively, the ion plasma spreads out into the torus
representing rapid loss form the torus boundaries. This loss of ions is also
reflected in decrement of the peak poloidal and toroidal density values of the
ion plasma through the simulation time. Figure 6: (a)-(d) shows electrostatic
potential surface plots of the system for $f$ $=$ $0.007$ at different
simulation time periods $\tau$ $=$ $16.67$, $21.11$, $23.33$ and $25.56$,
where $\tau$ $=$ $t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal
Diocotron period for QQS state. The potential surface plots display closed
axisymmetric toroidal potential well of the system, with spatial location of
the well minima around the same spatial location of the electron plasma
density maxima (Fig. 5 (a)-(d)).
Wall probes similar to experimental devices are placed at four different
azimuthal locations (Fig. 1) (separated by $\theta$ $=$ $\pi/2$) near the
inner wall at $z$ $=$ $L_{z}/2$ plane. Wall probe current obtained from the
inner wall probe is used as a diagnostic method to analyse the electron plasma
dynamics in the vessel. This “wall probe current” depends mainly on the charge
induced by center of charge motion of the electron plasma and expressed as
$I_{wall}=\frac{d(CV)}{dt}$. Here $V(t)$ and $C(t)$ are the instantaneous
probe voltage and capacitance respectively. Thus the charge induced on the
probe is $C(t)V(t)$ at that instance. In this work, a constant capacitance
value is chosen to be $C=0.08pF$ so that the obtained order of magnitude of
probe current is similar to typical experimental values Ganesh and Pahari
(2006).
In the next Sec. III, the effect of the ion resonance instability on electron
dynamics is investigated via different diagnostics methods: (i) Growth in wall
probe current and dependency on $f$, electron/ion plasma spatial evolution in
the torus and corresponding density evolution, spectrogram analysis of the
wall probe current and electrostatic potential of the system, cylindrically
approximated ion plasma rotational frequency and corresponding toroidal $m=1$
Diocotron frequency of the electron plasma (ii) electron/ion particle loss in
the system (iii) energy of the electron/ion plasma (iv) growth rate of the
wall probe current at different periods of the simulation, and (v) temperature
analysis of the electron and ion plasmas.
## III Diagnostics
As indicated earlier, to investigate the electron and ion plasma dynamics, we
have implemented several diagnostics suitable for this study. The diagnostics
results covers the simulation time period $\tau$ $=$ $0.00$ to $\tau$ $=$
$34$.
### III.1 Growth in wall probe current and dependency on $f$
To study the electron dynamics in the presence of ions in the system, wall
probe signal (wall probe voltage in Volts) is obtained as a function of time
from one of the inner wall probe and the corresponding wall probe current is
shown in Fig. 4. Wall probe current from three other probes at three remaining
toroidal locations provide identical information which suggests absence of any
toroidal mode in the system (not shown here). In Fig. 4, wall probe current
growth is observed for $f$ values ranging from $0.001$ to $0.01$ upto $\tau$
$=$ $34$. As discussed earlier, at $\tau$ $=$ $0.00$, a cold electron plasma
density distribution obtained by extremizing inertia-less entropy is
initializedKhamaru, Sengupta, and Ganesh (2019); Khamaru, Ganesh, and Sengupta
(2021a), which evolves self consistently into a quiescent quasi-steady (QQS)
state, as reflected in the wall probe data shown in Fig. 4. As discussed
earlier, ions are then introduced “by hand” or “pre-loaded” at $\tau$ $=$
$11.11$ (along with equal number of secondary electrons, as it would happen in
an ionizing process), soon after which, the amplitudes of the probe currents
start to grow with time and the growth is more for higher values of $f$ for
any instant of time. This is clearly demonstrated in the inset plot displayed
at the onset of the growth. Also an initial phase difference between the probe
currents for different values of $f$ can be observed. The growth of the wall
probe current is algebraic in nature for all values of $f$ and saturates at
later time of the simulation, mainly for higher $f$ values. More analysis on
the growth rate is performed in Subsec. III.5.
To aid the above analysis, spatial positions of the electron plasma inside the
torus along with density contours are shown in Fig. 5 (a)-(d), at simulation
times $\tau$ $=$ $16.67$, $21.11$, $23.33$, $25.56$ respectively with $f$ $=$
$0.007$ case. Before $\tau$ $=$ $11.11$, the electron plasma has attained a
quiescent state with negligible displacement, shown in previous study of QQS
stateKhamaru, Ganesh, and Sengupta (2021a, b). After ions were introduced at
$\tau$ $=$ $11.11$, the “center of charge motion” ($m$ $=$ $1$) of the
electron plasma in the poloidal plane gains amplitude as time progresses. The
electron plasma is displaced along the vertical and horizontal directions,
with a distinguishable compression-expansion cycle due to toroidicity. The
peak density values at the toroidal/poloidal plane evolve periodically along
with the “center of charge motion” of the plasma. In these figures, the
poloidal shape of the plasma shows elliptic nature representing $m$ $=$ $2$
mode. Referring to Fig. 4, the probe currents attain “double peak” nature
which actually corresponds to the “center of charge motion” ($m$ $=$ $1$) of
the electron plasma in the poloidal plane and onset of strong compression-
expansion cycle. Due to this center of charge motion, electron plasma makes a
finite amplitude elliptical trajectory in the poloidal plane, which we dub
here as nonlinear toroidal Diocotron motion and effectively similar to the
cylindrical $m$ $=$ $1$ Diocotron motion of the electron plasma with a
remarkable difference that at such small aspect ratios as studies here, the
modes are coupled in a poloidal plane due to strong toroidicity or $1/R$
dependence of the toroidal magnetic field. Among these double peaks, the
higher amplitude peak corresponds to the electron plasma position near the
inner wall of the torus with most compressed form of the plasma and the lower
amplitude peak appears while the plasma is away from the inner wall in most
expanded formKhamaru, Sengupta, and Ganesh (2019). This can be seen from the
Fig. 4, (i) the growth rate of the amplitudes of the peaks increases with time
for every $f$ values (for higher $f$ values growth with time is more) and (ii)
at any instant of time, the amplitude values of the peaks depend on $f$ values
i.e. with high value of $f$ the peak amplitude is higher. This brings out
clearly the fact that with higher ion density the “center of charge” motion
becomes stronger, which means an extended elliptical trajectory of the
electron plasma from the QQS state, in the poloidal plane of the torus. This
displacement of the electron plasma increases rapidly with time for higher
values of the ion density. These signatures point out clearly that the physics
of destabilization of QQS state is strongly connected to the presence of ions
and its amplitude to the initial fraction of ions $f$ present in the system.
Following the above said arguments, we believe that the wall probe growth or
amplitude increase is indeed due to a toroidal ion resonance instability.
Figure 7: Spectrogram of the wall probe current for $f$ $=$ $0.0$ (QQS state),
$0.001$, $0.005$, $0.01$ upto $\tau$ $=$ $34$, where $\tau$ $=$
$t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal Diocotron period for
QQS state. White dotted line represents the introduction time of the ion
plasma at $\tau$ $=$ $11.11$. Left y axis presents the frequency of the data
where as in the right y axis the frequency is normalized by the nonlinear
toroidal Diocotron frequency ($\sim$ $0.58$ MHz) of the QQS state. For all
values of $f$, without the ion plasma before $\tau$ $=$ $11.11$, $m$ $=$ $1$
and $m$ $=$ $2$ Diocotron modes are present in the system. For $f$ $=$
$0.001$, $0.005$ and $0.01$ cases higher modes (upto $m$ $=$ $9$ for $f$ $=$
$0.01$) start to appear with time and $m$ $=$ $2$ mode becomes dominant. The
temporal evolution of the spectrogram plots for $f$ $=$ $0.001$, $0.005$ and
$0.01$ cases reveal the ion density dependency of the dominant modes ($m$ $=$
$1$ and $m$ $=$ $2$) and dynamical chirping infused coupling of low poloidal
modes ($m$ $=$ $1$ to $m$ $=$ $9$). For $f$ $=$ $0.01$ case, a high chirping
region between $\tau$ $\sim$ $15$ and $16$ is noted.
The time evolution of the ion plasma and density is shown in Fig. 5. At $\tau$
$=$ $16.67$ (see Fig. 5 (e)) the initially cold loaded ion plasma starts to
gain energy from the system and initial shape of the plasma (Fig. 3 (b) at
$\tau$ $=$ $11.11$) is distorted. In Fig. (f)-(h), at $\tau$ $=$ $21.11$,
$23.33$, $25.56$ respectively, the plasma spreads out into the torus
representing rapid loss of ions from the torus boundaries. Initial loss of the
ions is higher at the top boundary wall and at later simulation times from
other boundary walls also. This loss of ions is also reflected in decrement of
the peak poloidal and toroidal density values of the ion plasma through the
simulation time. Also rapid ion loss is discussed in the next Subsec. III.3.
The cold loaded ion plasma gains enough kinetic energy from the electron
plasma as a result of the ion resonance instability. As our simulation is
collisionless, we believe that the underlying exchange of energy between ions
and electrons is a collisionless resonant process. The thermal energy gain of
the ions results in increase of ion Larmour radii (for example, at
perpendicular ion temperature $50~{}eV$, average Larmour radius of an ion is
${\rho_{{Li}}}$ $\sim$ $0.214~{}m$ near the inner wall location and
${\rho_{{Li}}}$ $\sim$ $1.073~{}m$ near outer wall location). Also various
inertia driven drifts due to the toroidal inhomogeneous magnetic fields
($\bm{\nabla}B$ and curvature drift) aid the loss of ions. Thus the ions are
rapidly lost from the system. The energy transfer process is investigated
further in the Subsec. III.4.
Corresponding to the presented spatial locations and density plots of the
electron and ion plasma, electrostatic potential surface plots of the system
are shown in Fig. 6 (a)-(d). The potential surface plots display closed
axisymmetric toroidal potential well of the system, with spatial location of
the well minima around the same spatial location of the electron plasma
density maxima (Fig. 5 (a)-(d)). This potential well has parabolic nature in
the poloidal plane and the ions are initially trapped in the potential well.
Though the location of the maxima of the ion plasma density is near to that of
the electron plasma/potential minima at $\tau$ $=$ $16.67$, after that the ion
plasma density maxima do not exhibit the same. This observation suggests the
decoupling of the ion plasma from the electron plasma at later simulation time
period after $\tau$ $=$ $16.67$. More analysis on this is presented in Subsec.
III.4.
To further understand the dynamics of the dynamic toroidal Diocotron modes of
the system and possible mode coupling associated with the electron and ion
plasma dynamics, spectrogram analysis of the wall probe current has been
performed for $f$ $=$ $0.0$ (QQS state), $0.001$, $0.005$, $0.01$ upto $\tau$
$=$ $34$ and shown in Fig. 7. Spectrogram frequency is normalized by the
nonlinear toroidal Diocotron frequency ($\sim$ $0.58$ MHz) of the QQS state
and shown in the right y axis. For all values of $f$ considered, without the
ion plasma before $\tau$ $=$ $11.11$, $m$ $=$ $1$ and $m$ $=$ $2$ Diocotron
modes are the only dominant ones that are present in the system. For $f$ $=$
$0.001$, $0.005$ and $0.01$ cases higher modes (upto $m$ $=$ $9$ for $f$ $=$
$0.01$) start to appear with time with introduction of ions at $\tau$ $=$
$11.11$ and $m$ $=$ $2$ mode becomes dominant. The temporal evolution of the
spectrogram plots for $f$ $=$ $0.001$, $0.005$ and $0.01$ cases reveal the ion
density dependency of the dominant modes ($m$ $=$ $1$ and $m$ $=$ $2$) and
dynamical chirping infused coupling of low poloidal modes ($m$ $=$ $1$ to $m$
$=$ $9$). Similar feature of the poloidal modes with $m$ $=$ $2$ as dominant
mode, was also seen in our previous studyKhamaru, Sengupta, and Ganesh (2019)
with an initial loading of the electron plasma while the electron plasma had
stronger “center of charge” motion than that of the QQS state. Also for $f$
$=$ $0.01$ case, a high chirping region at very low power has been observed
between $\tau$ $\sim$ $15$ and $16$.
### III.2 Ion resonance instability and ion plasma rotational frequency
In straight cylinder geometryDavidson (1990), ion plasma rotational frequency
in a partially neutralized electron plasma is given by
$\omega_{ri}^{-}=-\frac{\omega_{ci}}{2}[1-(1+\frac{2\omega_{pe}^{2}}{\omega_{ce}^{2}}(1-f)\frac{m_{i}}{m_{e}})^{1/2}]$
(1)
where $\omega_{ci}$ and $\omega_{ce}$ are the cyclotron frequency of ion and
electron respectively at uniform $B$ field, $\omega_{pe}$ is the electron
plasma frequency, $m_{i}$ and $m_{e}$ are the ion mass and electron mass
respectively. The condition for ion resonance instabilityDavidson (1990) of
the electron plasma in cylindrical geometry is that the ion plasma rotational
frequency should be nearly equal to the cylindrical $m=1$ Diocotron frequency
of the electron plasma. In the present study with radially varying toroidal
magnetic field, the an approximated ion plasma rotational frequency has been
considered to compare with the toroidal $m=1$ Diocotron rotation of the
electron plasma, as discussed in several ion resonance instability
studiesMarksteiner _et al._ (2008); Lachhvani _et al._ (2016). With the
present parameters for $Ar^{+}$ ions, the ion plasma rotational frequency
(magnetic field at the minor axis location of the torus $B=0.011T$) $\sim$ 0.4
MHz under cylindrical approximation where “measured” toroidal Diocotron
frequency of the electron plasma $\sim$ 0.58 MHz in our system. To the lowest
order in inverse aspect ratio, thus the ion rotational frequency is at $\sim
31\%$ deviation from toroidal Diocotron frequency of the electron plasma. As
the electron plasma and ion plasma dynamics are in toroidal geometry with
inhomogeneous magnetic field, the presented deviation should be acceptable, if
a permissible range of the deviation can be obtained. To estimate the
permissible range of the deviation, $N^{+}$ ($m_{i}=25532.65m_{e}$) and
$H^{+}$ ($m_{i}=1836m_{e}$) ions are also introduced in the device replacing
$Ar^{+}$ ions to observe the electron plasma dynamics. For $N^{+}$ and $H^{+}$
ions, corresponding ion plasma rotational frequencies are $\sim$ 0.67 MHz
($\sim 9\%$ deviation) and $\sim$ 2.4 MHz ($\sim 182\%$ deviation)
respectively. Algebraic growth of the wall probe current similar to ${Ar}^{+}$
ions has been observed for $N^{+}$ ions where as for $H^{+}$ ions, no growth
is observed (the results are not shown here). These findings indicate the
dependency of the ion resonance instability on the ion mass. Thus we can
conclude that the permissible value of the ion plasma rotational frequencies
to initiate an ion resonance instability of the electron plasma, for the
current simulation parameters, lies between a certain range. Further analysis
in the present study mainly focused on the ${Ar}^{+}$ ions and detailed
investigation of the ion plasma rotational frequency range can be performed in
future with different ion species.
To investigate the ion resonance induced instability in this system in more
detail, loss of electron and ion particles and temporal evolution of the
ion/electron plasma’s energy with simulation time, are described next.
### III.3 Loss of electron and ion
Figure 8: Ratio of the instantaneous electron particle number ($N_{E}(t)$) to
the initial value of the electron particle number ($N_{E0}$) i.e.
${N_{E}(t)}/N_{E0}$ is shown for different values of $f$ upto $\tau$ $=$ $34$,
where $\tau$ $=$ $t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal
Diocotron period for QQS state. At $\tau$ $=$ $11.11$ (indicated by black
dotted line) introduction of ions results into a dip in ${N_{E}(t)}/N_{E0}$
values which is more for higher values of $f$. The inset plot shows the
effective electron loss, around $\tau$ $=$ $11.11$. Figure 9: Ratio of the
instantaneous ion particle number (${N_{I}(t)}$) to the ion particle number
($N_{I}(\tau=11.11)$) at $\tau=11.11$, i.e. ${N_{I}(t)}/N_{I}(\tau=11.11)$ is
shown for different values of $f$ upto $\tau$ $=$ $34$, where $\tau$ $=$
$t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal Diocotron period for
QQS state. After $\tau$ $=$ $11.11$ (indicated by black dotted line),
${N_{I}(t)}/N_{I}(\tau=11.11)$ show finite value (ion particles are introduced
in the system at $\tau$ $=$ $11.11$) and its value decreases with simulation
time. The decrement is more for higher values of $f$. The inset plot shows the
starting of ion loss, around $\tau$ $=$ $12.5$.
Though electron plasma in a quiescent quasi-steady state has better stability
properties, inertia effects i.e. various drifts due to the toroidal
inhomogeneous magnetic fields: $\bm{\nabla}B$ and curvature drift (and
$\overline{\rho}_{Le}/{L}$ to all ordersKhamaru, Ganesh, and Sengupta (2021a,
b)) are present in the system leading to some electron losses ($\sim$
$2\%$)Khamaru, Ganesh, and Sengupta (2021a) in the system. As ions are
introduced in the system, destabilization of electron plasma due to toroidal
ion resonance instability, initiates more rapid loss of electron particles
from the system. Such loss results in confinement issues of electron plasma in
the experimental studies. For the present study, primary electron particle
loss is shown in Fig. 8 where ratio of the instantaneous electron particle
number $N_{E}(t)$ to the initial value of the electron particle number
$N_{E0}$ i.e. ${N_{E}(t)}/N_{E0}$ for different values of $f$ upto $\tau$ $=$
$34$ is shown. Here $\tau$ $=$ $t/\overline{t}_{D}$ and $\overline{t}_{D}$ is
toroidal Diocotron period for QQS state. The time of introduction of ions is
at $\tau$ $=$ $11.11$, as indicated by black dotted line. Before $\tau$ $=$
$11.11$, the electron plasma evolves into QQS state and the the plasma losses
$\sim$ $3.5\%$ of initial total number of electrons. At $\tau$ $=$ $11.11$
introduction of ions results into a dip in the quantity ${N_{E}(t)}/N_{E0}$
and the loss rate increases with increasing value of $f$ or ion density. At
$\tau$ $=$ $34$, the quantity ${N_{E}(t)}/N_{E0}$ losses $\sim$ $5\%$ of the
initial value for $f$ $=$ $0.001$ case, and $\sim$ $8\%$ for $f$ $=$ $0.01$
case. The inset plot around $\tau$ $=$ $11.11$ shows the the effective
electron loss, around the introduction period of ions in the system.
Figure 10: Energy variations of the system upto $\tau$ $=$ $34$, where $\tau$
$=$ $t/\overline{t}_{D}$ and $\overline{t}_{D}$ is toroidal Diocotron period
for QQS state. The initial value of the total energy ${TE}_{0}$
($3.503\times{10}^{-5}$ J) of the system at $\tau$ $=$ $0.00$ is subtracted
from the total energy ($TE$), kinetic energy ($KE$) and potential energy
($PE$) of the system and shown for the values of $f$ ranging from $0.001$ to
$0.01$ in (a), (b) and (c) respectively. The time of introduction of ions is
at $\tau$ $=$ $11.11$, as indicated by black dotted line in all three plots.
The inset plot in (a) around $\tau$ $=$ $11.11$ shows that total energy of the
system starts to decrease before $\sim$ one toroidal Diocotron time period
after $\tau$ $=$ $11.11$. The inset plots in (b) and (c) shows the initial
values of ${KE}$ $-$ ${TE}_{0}$ and ${PE}$ $-$ ${TE}_{0}$ respectively.
In Fig. 9, ratio of the instantaneous ion particle number (${N_{I}(t)}$) to
the ion particle number ($N_{I}(\tau=11.11)$) at $\tau=11.11$, i.e.
${N_{I}(t)}/N_{I}(\tau=11.11)$ is shown for different values of $f$ upto
$\tau$ $=$ $34$, where $\tau$ $=$ $t/\overline{t}_{D}$ and $\overline{t}_{D}$
is toroidal Diocotron period for QQS state. The time of introduction of ions
is at $\tau$ $=$ $11.11$, as indicated by black dotted line. Upto $\tau$ $=$
$11.11$, ${N_{I}(t)}/N_{I}(\tau=11.11)$ shows zero value as the system
contained no ions at all, whereas after $\tau$ $=$ $11.11$,
${N_{I}(t)}/N_{I}(\tau=11.11)$ has finite value (ion particles are introduced
in the system at $\tau$ $=$ $11.11$). At $\tau$ $=$ $34$, ion plasma losses
$\sim$ $32\%$ of its initial value for $f$ $=$ $0.001$ case and $\sim$ $80\%$
for $f$ $=$ $0.01$ case. The inset plot around $\tau$ $=$ $11.11$ shows the
introduction period of ions and the starting point of the loss of ions from
the system. After $\tau$ $=$ $11.11$, the loss of ion particles starts after
$\sim$ one toroidal Diocotron cycle ($\sim$ $\tau$ $=$ $12.5$) and the
starting time varies with different values of $f$. In the case of electron
loss, the effective electron loss (after $\tau$ $=$ $11.11$) starts just
before $\sim$ one toroidal Diocotron cycle $\sim$ $\tau$ $=$ $12.5$). This
observation indicates that though the loss of ion at later stage of the
simulation is rapid and significantly high when compared to the electron loss
within the same simulation time period, initially the ion particles were
trapped and gained energy while the system endured electron particle loss.
Also, in spite of the ion loss at later stage of the simulation, the electrons
plasma remains destabilized throughout that simulation time period showing
saturation of the amplitude of the wall probe current (Fig. 4). This suggests,
the initial presence of the ion plasma at $\tau$ $=$ $11.11$ started the
process of electron plasma destabilization which became independent of the
presence of ion plasma in the system at a later time. It is also to be noted
that, as the ion loss rate is more than the electron loss rate along with the
simulation time, the dynamic $f$ values become lower than the initial values.
To understand the plasma dynamics in detail, temporal evolution of the kinetic
and potential energy of the total system as well as for electron/ion plasma
are described.
### III.4 Energy variation with simulation time
In Fig. 10 energy evolution of the system are shown upto $\tau$ $=$ $34$. The
initial value of the total energy ${TE}_{0}$ ($3.503\times{10}^{-5}$ J) of the
system at $\tau$ $=$ $0.00$ is subtracted from the total energy ($TE$),
kinetic energy ($KE$) and potential energy ($PE$) of the system and shown for
the values of $f$ ranging from $0.001$ to $0.01$ in Fig. 10 (a), (b) and (c)
respectively. The time of introduction of ions is at $\tau$ $=$ $11.11$, as
indicated by black dotted line in all three plots. Fig. 10 (a) shows ${TE}$
$-$ ${TE}_{0}$ $\sim$ ${10}^{-7}$, indicating that the total energy of the
system is conserved for all $f$ values (for lowest $f$ $=$ $0.001$, the
increment is $<$ $1\%$). The inset plot in Fig. 10 (a) around $\tau$ $=$
$11.11$ shows that total energy (${TE}$ $-$ ${TE}_{0}$) of the system starts
to decrease before one toroidal Diocotron time period after $\tau$ $=$
$11.11$. The inset plots in Fig. 10 (b) and (c) shows the initial values of
${KE}$ $-$ ${TE}_{0}$ and ${PE}$ $-$ ${TE}_{0}$ respectively. From $\tau$ $=$
$0.00$ to $\tau$ $=$ $11.11$, while the system remains in QQS state without
ion plasma, kinetic energy of the system/the electron plasma increases, along
with an increase oscillation amplitude, at the cost of electrostatic potential
energy of the system (Fig. 10 (c): $\tau$ $=$ $0.00$ to $\tau$ $=$ $11.11$).
After the introduction of cold ion plasma in the system at $\tau$ $=$ $11.11$,
average kinetic energy of the system (electrons and ions) rises with time
while the oscillation amplitude starts to grow with time for all values of
$f$. For higher values of $f$, the growth in the oscillation amplitude are
higher, where as average kinetic energy of the system reaches to close values
for all $f$ at $\tau$ $=$ $34$. In the same time interval, average potential
energy of the system decreases for all values of $f$ (Fig. 10 (c)), along with
growth of the oscillation amplitudes similar as the kinetic energy. However,
the average potential energy values drops with higher $f$ values at $\tau$ $=$
$34$, mainly because of its high transfer to the kinetic energy of the system
(electrons and ions) with higher $f$ values. Also loss of particles from the
system with time (Fig. 8 and 9) aids to small order decrement of the potential
energy of the system. Clearly, rise in the oscillation amplitude of kinetic
energy as well as potential energy with higher values of $f$ needs to be
addressed separately. In the ion resonance instability process, energy is
transferred from electron plasma to ion plasma. Thus the energy profile
variations of the electron and ion plasma are investigated separately and
explained in the next paragraph.
Figure 11: (a) and (b) represent volume averaged parallel and perpendicular
velocity distribution functions of electron plasma, $\langle
f_{e}(v_{\parallel})\rangle$ and $\langle f_{e}(v_{\bot})\rangle$
respectively, at different simulation time periods $\tau$ $=$ $12.22$,
$15.56$, $18.89$, $20.00$ and $34.00$ along with corresponding parallel and
perpendicular temperatures $\langle T_{\parallel}^{e}\rangle$ and $\langle
T_{\bot}^{e}\rangle$, for the case of $f$ $=$ $0.007$. In (c) and (d), volume
averaged parallel and perpendicular velocity distribution functions of ion
plasma, $\langle f_{i}(v_{\parallel})\rangle$ and $\langle
f_{i}(v_{\bot})\rangle$, at same simulation time periods, are shown
respectively along with corresponding parallel and perpendicular temperatures,
$\langle T_{\parallel}^{i}\rangle$ and $\langle T_{\bot}^{i}\rangle$, for same
$f$ value. In (a) and (b), the temperatures increase from ${\langle
T_{\parallel}^{e}\rangle}_{12.22\tau}$ $=$ $10~{}eV$ to ${\langle
T_{\parallel}^{e}\rangle}_{34\tau}$ $=$ $16.5~{}eV$ and from ${\langle
T_{\bot}^{e}\rangle}_{12.22\tau}$ $=$ $15~{}eV$ to ${\langle
T_{\bot}^{e}\rangle}_{34\tau}$ $=$ $21.03~{}eV$ respectively. In (c), parallel
distribution function of the ion plasma at $\tau$ $=$ $12.22$, results into
${\langle T_{\bot}^{i}\rangle}_{12.22\tau}$ $=$ $0.0004~{}eV$. The parallel
temperature attains ${\langle T_{\bot}^{i}\rangle}_{34\tau}$ $=$
$0.0546~{}eV$. In (d), for $\tau$ $=$ $12.22$, $15.56$ and $18.89$, the
distribution functions do not form near-Maxwellian shape, though
$v_{\perp}^{i}$ $\sim$ $10^{5}$ m/s, $10^{2}$ order $>$ $v_{\parallel}^{i}$.
For $\tau$ $=$ $20.00$ and $34.00$, the distribution function starts to form
near-Maxwellian shape resulting ${\langle T_{\bot}^{i}\rangle}_{20\tau}$ $=$
$218.46~{}eV$ and ${\langle T_{\bot}^{i}\rangle}_{34\tau}$ $=$ $280.36~{}eV$.
Figure 12: Normalized kinetic energy and potential energy variations of the
electron plasma and ion plasma are shown upto $\tau$ $=$ $34$ for $f$ $=$
$0.001$ to $0.01$. The energy values are normalized by initial total energy
${TE}_{0}$ ($3.503\times{10}^{-5}$ J) of the system at $\tau$ $=$ $0.00$. The
time of introduction of ions is at $\tau$ $=$ $11.11$, as indicated by black
dotted line in all plots. (a) and (b) represents the variation of kinetic
energy ${KE}_{e}$/${TE}_{0}$ and potential energy ${PE}_{e}$/${TE}_{0}$ of the
electron plasma respectively. The inset plot in (a) around $\tau$ $=$ $11.11$
shows that oscillation amplitude in the kinetic energy of the electron plasma
starts to increase at $\sim$ $\tau$ $=$ $11.11$. The inset plots in (b),
around $\tau$ $=$ $11.11$ shows $\sim$ ${10}^{-3}$ to ${10}^{-2}$ order
increase in the amplitude of the ${PE}_{e}$/${TE}_{0}$ values at $\tau$ $=$
$11.11$ because of the addition of secondary electrons. (c) and (d) shows
kinetic energy ${KE}_{i}$/${TE}_{0}$ and potential energy
${PE}_{i}$/${TE}_{0}$ of the ion plasma respectively from $\tau$ $=$ $11.11$
to $34$. ${KE}_{i}$/${TE}_{0}$ linearly increases after $\tau$ $=$ $11.11$ for
all $f$ values. The linear fits yield: growth rates ${\alpha}_{KE}^{0.001}$
$=$ $0.00087$ and ${\alpha}_{KE}^{0.01}$ $=$ $0.01052$. After the time period
$\tau$ $=$ $11.11$ to $\tau$ $\sim$ $19.17$, the fall of the quantity
${KE}_{i}$/${TE}_{0}$ is not followed by rise of the order $>$ ${10}^{-3}$ for
any values of $f$. In the period form $\tau$ $=$ $11.11$ to $\tau$ $\sim$
$19.17$ the ion plasma gains energy from the potential of the electron plasma.
In (d), the potential energy of the ion plasma, increases after $\tau$ $=$
$11.11$ (at the cost of the potential energy of the electron plasma) with
similar fluctuating trend as the kinetic energy. Around $\tau$ $\sim$ $19.17$,
the fluctuation reduces and the potential energy of the plasma slowly
decreases with simulation time.
In the above discussion, it is found that the kinetic energy of the system
(i.e. the electrons and ions) increases with simulation time. This suggests
that the average temperature also increases. To verify, volume averaged
parallel and perpendicular velocity distribution functions of the ion plasma
as well as electron plasma are shown in Fig. 11 to check if the electron or
ion plasma reach a certain equilibrium with a Maxwellian-like distribution
function. Here parallel velocity, $v_{\parallel}$ is defined as the 1D
velocity vector with velocity component along toroidal direction ($\theta$).
Perpendicular velocity, $v_{\bot}$ is defined as a 2D vector having a minor-
radial component and a poloidal component, or equivalently through a basis
transformation, a major-radial component, $v_{R}$ together with a cylindrical-
axial component, $v_{Z}$. Choosing the later basis for convenience, the
perpendicular speed is $|v_{\perp}|=\sqrt{{v_{R}}^{2}+{v_{Z}}^{2}}$. To
calculate corresponding volume averaged parallel and perpendicular
temperatures, volume averaged distribution functions, $\langle
f(v_{\parallel})\rangle$ and $\langle f(v_{\bot})\rangle$, which are near-
Maxwellian, are shown and fitted with respective perfect Maxwellian
distribution function (1D/2D) for a certain temperature value. Then $\langle
T_{\parallel}\rangle$ and $\langle T_{\bot}\rangle$ are calculated from the
respective fitted temperature values. Using the expressions given, for case of
$f$ $=$ $0.007$, volume averaged parallel and perpendicular velocity
distribution functions of electron plasma, $\langle
f_{e}(v_{\parallel})\rangle$ and $\langle f_{e}(v_{\bot})\rangle$, at
different simulation time periods (at $\tau$ $=$ $12.22$, $15.56$, $18.89$,
$20.00$ and $34.00$) are shown in Fig. 11 (a) and (b) respectively along with
corresponding parallel and perpendicular temperatures, $\langle
T_{\parallel}^{e}\rangle$ and $\langle T_{\bot}^{e}\rangle$. In Fig. 11 (c)
and (d), volume averaged parallel and perpendicular velocity distribution
functions of ion plasma, $\langle f_{i}(v_{\parallel})\rangle$ and $\langle
f_{i}(v_{\bot})\rangle$, at same simulation time periods, are shown
respectively along with corresponding parallel and perpendicular temperatures,
$\langle T_{\parallel}^{i}\rangle$ and $\langle T_{\bot}^{i}\rangle$, for same
$f$ value. In Fig. 11 (a) and (b), the widths of the electron distribution
functions spread with simulation time and the temperatures also increase as
${\langle T_{\parallel}^{e}\rangle}_{12.22\tau}$ $=$ $10~{}eV$ to ${\langle
T_{\parallel}^{e}\rangle}_{34\tau}$ $=$ $16.5~{}eV$ (Fig. 11 (a)) and
${\langle T_{\bot}^{e}\rangle}_{12.22\tau}$ $=$ $15~{}eV$ to ${\langle
T_{\bot}^{e}\rangle}_{34\tau}$ $=$ $21.03~{}eV$ (Fig. 11 (b)). The rise in
temperatures are also reflecting the rise in the kinetic energy of the
electron plasma. However, the ion distribution functions are more informative
about the energy transfer process and need to be discussed. In Fig. 11 (c),
parallel distribution function of the ion plasma at $\tau$ $=$ $12.22$, with
narrow width, corresponds to ${\langle T_{\parallel}^{i}\rangle}_{12.22\tau}$
$=$ $0.0004~{}eV$. The parallel temperature does not increase much and attains
${\langle T_{\parallel}^{i}\rangle}_{34\tau}$ $=$ $0.0546~{}eV$. This result
suggests that cold loaded ion plasma at $\tau$ $=$ $11.11$ does not acquire
enough parallel energy ($v_{\perp}^{i}$ $\sim$ $10^{3}$ m/s) obstructing the
ion particle motion in the toroidal plane. In Fig. 11 (d), for $\tau$ $=$
$12.22$, $15.56$ and $18.89$, the distribution functions do not form near-
Maxwellian shape, thus it is inaccurate to attribute any temperature
information to the plasma in these time periods. However, the spread in these
distribution functions ($v_{\perp}^{i}$ $\sim$ $10^{5}$ m/s, $10^{2}$ order
higher than $v_{\parallel}^{i}$) suggest the ion particle motion mainly in the
poloidal plane. For $\tau$ $=$ $20.00$ and $34.00$, the distribution function
starts to form near-Maxwellian shape which suggests that the ion plasma
approaches equilibrium state. By fitting the curves we obtain ${\langle
T_{\bot}^{i}\rangle}_{20\tau}$ $=$ $218.46~{}eV$ and ${\langle
T_{\bot}^{i}\rangle}_{34\tau}$ $=$ $280.36~{}eV$. As seen in this analysis,
the ion plasma forms near-Maxwellian distribution function in the region
between $\tau$ $=$ $18.89$ to $\tau$ $=$ $20.00$, indicating that the ion
plasma attains equilibrium after $\tau$ $=$ $18.89$. The temperature values
${\langle T_{\bot}^{i}\rangle}_{20\tau}$ $=$ $218.46~{}eV$ and ${\langle
T_{\bot}^{i}\rangle}_{34\tau}$ $=$ $280.36~{}eV$ are higher than that of the
corresponding electron temperatures. Such high temperature of ion plasma
indicates anomalous heating of ions via toroidal ion resonance instability
process and to the best of our knowledge has never been reported prior to this
work.
To get a better idea of the energy transfer processes, kinetic and potential
energies of the ion and electron plasmas have been obtained. Normalized
kinetic energy and potential energy variations of the electron plasma and ion
plasma with simulation time are shown in Fig. 12 upto $\tau$ $=$ $34$ for $f$
$=$ $0.001$ to $0.01$. The energy values are normalized by initial total
energy ${TE}_{0}$ of the system at $\tau$ $=$ $0.00$. The time of introduction
of ions is at $\tau$ $=$ $11.11$, as indicated by black dotted line in all
plots. Fig. 12 (a) and (b) represents the variation of kinetic energy
${KE}_{e}$/${TE}_{0}$ and potential energy ${PE}_{e}$/${TE}_{0}$ of the
electron plasma respectively, upto the simulation time $\tau$ $=$ $34$. The
inset plot in Fig. 12 (a) around $\tau$ $=$ $11.11$ shows that oscillation
amplitude in the kinetic energy of the electron plasma starts to increase at
$\sim$ $\tau$ $=$ $11.11$. In Fig. 12 (b), the inset plots around $\tau$ $=$
$11.11$ shows $\sim$ ${10}^{-3}$ to ${10}^{-2}$ order increase in the
amplitude of the ${PE}_{e}$/${TE}_{0}$ values at $\tau$ $=$ $11.11$. This is
because of the addition of secondary electrons of density $\sim$ ${10}^{-3}$
to ${10}^{-2}$ times the density of the primary electrons. After $\tau$ $=$
$11.11$, the kinetic energy and potential energy of the electron plasma show
similar trend as seen in the total kinetic energy and potential energy
variation of the system (Fig. 10 (b) and (c)) through the simulation time with
the values of $f$. Following initial drop after $\tau$ $=$ $11.11$, potential
energy of the electron plasma is drained out as the kinetic energy of the
plasma for all values of $f$. The drop is more for higher $f$ values as seen
in Fig. 10 (c). Loss of electron particles from the system with time (Fig. 8)
also aids to small order decrement of the potential energy of the plasma. The
orders of the values of ${KE}_{e}$/${TE}_{0}$ are that of the total kinetic
energy of the system throughout the simulation, as the ions and secondary
electrons are loaded cold contributing zero kinetic energies to the system.
Similarly, the orders of the values of ${PE}_{e}$/${TE}_{0}$ are that of the
total potential energy of the system as the ions and secondary electrons are
loaded with lower density than that of the primary electron plasma ($f$ $=$
$0.001$ to $0.01$). In Fig. 12 (a) and (b), $\tau$ $=$ $19.17$ is marked by
vertical dotted line in magenta and explained in next paragraph.
To understand the energy variation of ion plasma through the simulation time,
normalized kinetic energy ${KE}_{i}$/${TE}_{0}$ and potential energy
${PE}_{i}$/${TE}_{0}$ of the ion plasma are shown in Fig. 12 (c) and (d)
respectively for the values of $f$ ranging from $0.001$ to $0.01$ from
simulation time $\tau$ $=$ $11.11$ to $\tau$ $=$ $34$. As the ion plasma was
cold loaded at $\tau$ $=$ $11.11$, kinetic energy is zero at $\tau$ $=$
$11.11$ in Fig. 12 (c). The parameter ${PE}_{i}$/${TE}_{0}$ $\sim$ ${10}^{-3}$
to ${10}^{-2}$ times the total energy of the system at $\tau$ $=$ $11.11$ in
Fig. 12 (d). Demonstrating typical outcome of the ion resonance instability
process, the ion particles gain energy (kinetic/potential) from the potential
energy of the electron plasma. As our simulation is collisionless, we believe
that there exists an underlying collisionless process, resulting in
collisionless exchange of energy between ions and electrons. In Fig. 12 (c),
the parameter ${KE}_{i}$/${TE}_{0}$ linearly increases after $\tau$ $=$
$11.11$ for all $f$ values. To check the growth rate with respect to $f$
values, the parameter ${KE}_{i}$/${TE}_{0}$ is linearly fitted upto $\tau$ $=$
$11.8$ for $f$ $=$ $0.001$ and $0.01$ cases and shown in the inset plot. The
linear fits yield slope of ${KE}_{i}$/${TE}_{0}$ growth as
${\alpha}_{KE}^{0.001}$ $=$ $0.00087$ for $f$ $=$ $0.001$ and
${\alpha}_{KE}^{0.01}$ $=$ $0.01052$ for $f$ $=$ $0.01$, suggesting $\sim$
${10}$ time higher growth rate for one order of magnitude increase in $f$
value ($\frac{{\alpha}_{KE}^{0.01}}{{\alpha}_{KE}^{0.001}}$ $\sim$ ${12.09}$).
After $\tau$ $=$ $11.11$, the quantity ${KE}_{i}$/${TE}_{0}$ displays
fluctuating rise/fall profile with occasional gain/loss of ion kinetic energy,
for all values of $f$. We have identified a region, $\tau$ $=$ $11.11$ to
$\tau$ $\sim$ $19.17$ ($\tau$ $=$ $19.17$ marked by vertical dotted line in
magenta), after when the fall of the quantity ${KE}_{i}$/${TE}_{0}$ is not
followed by rise of the order $>$ ${10}^{-3}$ for any values of $f$. After
$\tau$ $\sim$ $19.17$, ${KE}_{i}$/${TE}_{0}$ slowly decreases with simulation
time with small amount of fluctuations (though for $f$ $=$ $0.001$ to $0.005$,
the gradual fall of the kinetic energy starts at later simulation time). In
the period form $\tau$ $=$ $11.11$ to $\tau$ $\sim$ $19.17$, the ion plasma
gains energy from the potential of the electron plasma. After $\tau$ $\sim$
$19.17$ the ion plasma starts to equilibrate. This transition time is within
the region between $\tau$ $=$ $18.89$ to $\tau$ $=$ $20.00$, as seen in the
velocity distribution function analysis of Fig. 11 (d). Further loss of
kinetic energy after $\tau$ $\sim$ $19.17$ can also be related to the loss of
ion particles from the system though the particle loss starts sooner than
$\tau$ $\sim$ $19.17$ (Fig. 9). The potential energy of the ion plasma, shown
in Fig. 12 (d), also increases after $\tau$ $=$ $11.11$ (at the cost of the
potential energy of the electron plasma) with similar fluctuating trend as the
kinetic energy. Around $\tau$ $\sim$ $19.17$, the fluctuation reduces and the
potential energy of the plasma slowly decreases with simulation time, though
the transition of the fluctuation amount around $\tau$ $\sim$ $19.17$ is not
strikingly distinguishable as the kinetic energy case. The region around
$\tau$ $\sim$ $19.17$ is also identified in Fig. 12 (a) and (b) (vertical
dotted line in magenta) for the electron plasma, but there is no noticeable
transition of kinetic/potential energy of the electron plasma around that time
period. The reason might be the low order magnitude of the ion plasma energies
compared to the electron plasma energies.
### III.5 Growth rate of the wall probe current at different periods of the
simulation
Figure 13: (a) Growth rate analysis of the inner wall probe current for $f$
$=$ $0.001$ and $0.01$ upto $\tau$ $=$ $34$. $\tau$ $=$ $t/\overline{t}_{D}$
and $\overline{t}_{D}$ is toroidal Diocotron period for QQS state. The time of
introduction of ions is at $\tau$ $=$ $11.11$, as indicated by black dotted
line. $\tau$ $=$ $19.17$ is indicated by vertical dotted line in magenta.
Algebraic growth rate of the wall probe current is calculated by linear
fitting of the maximum peak among the double peaks of the probe current, in
two regions: (1) $\tau$ $=$ $11.11$ to $19.17$ and (2) $\tau$ $=$ $19.17$ to
$34$, for both of $f$ $=$ $0.001$ and $0.01$. The fits yield
${\gamma}_{1}^{0.001}$ $=$ $0.0004$ $mA/\overline{t}_{D}$,
${\gamma}_{1}^{0.01}$ $=$ $0.008$ $mA/\overline{t}_{D}$ and
${\gamma}_{1}^{0.001}$ $=$ $0.0008$ $mA/\overline{t}_{D}$,
${\gamma}_{1}^{0.01}$ $=$ $0.0037$ $mA/\overline{t}_{D}$. (b) Growth rates for
the two regions (1) $\tau$ $=$ $11.11$ to $19.17$ and (2) $\tau$ $=$ $19.17$
to $34$ are shown as a function of $f$, as ${\gamma}_{1}$ and ${\gamma}_{2}$
respectively. Linear fit of ${\gamma}_{1}$ and ${\gamma}_{2}$ reveals growth
rate as ${\beta}_{1}$ $=$ $0.821$ $mA/\overline{t}_{D}$ and ${\beta}_{1}$ $=$
$0.263$ $mA/\overline{t}_{D}$.
Fig. 13 (a) shows growth rate analysis of the inner wall probe current for $f$
$=$ $0.001$ and $0.01$ upto $\tau$ $=$ $34$. Wall probe currents for extrema
values of $f$ are chosen for comparison purpose i.e. to reveal the dependency
of growth on lower and higher $f$ values. The time of introduction of ions is
at $\tau$ $=$ $11.11$, as indicated by black dotted line. $\tau$ $=$ $19.17$,
simulation time close to the transition time of the ion plasma to equilibrium
state, is indicated by vertical dotted line in magenta. Algebraic growth rate
of the wall probe current is calculated by linear fitting of the maximum peak
among the double peaks of the probe current, in two regions: (1) $\tau$ $=$
$11.11$ to $19.17$ and (2) $\tau$ $=$ $19.17$ to $34$, for both of $f$ $=$
$0.001$ and $0.01$. In the initial ($\tau$ $=$ $11.11$ to $19.17$) time
period, the linear fit reveals growth rates ${\gamma}_{1}^{0.001}$ $=$
$0.0004$ $mA/\overline{t}_{D}$ and ${\gamma}_{1}^{0.01}$ $=$ $0.008$
$mA/\overline{t}_{D}$. For later time period ($\tau$ $=$ $19.17$ to $34$),
growth rates are ${\gamma}_{1}^{0.001}$ $=$ $0.0008$ $mA/\overline{t}_{D}$ and
${\gamma}_{1}^{0.01}$ $=$ $0.0037$ $mA/\overline{t}_{D}$. As expected, the
growth rate is high in the initial region and attains lower value in the later
region, indicating saturation of the wall probe current for $f$ $=$ $0.01$
case. This result echos the same explanation of energy transfer given in the
ion plasma energy analysis of Subsec. III.4. For $f$ $=$ $0.001$ case though,
the growth rate is higher at later part (lower values than $f$ $=$ $0.01$
case). The discrepancy suggests that for lower $f$ values, the ion plasma
gains energy for longer time (low energy gain than the high $f$ value) and
equilibrates at a later time than $\tau$ $=$ $19.17$. This feature is also
seen in Fig. 12 (a), where for $f$ $=$ $0.001$ to $0.005$, the fall of the
kinetic energy started at later simulation time ($\sim$ $\tau$ $=$ $25$). To
get an idea about the variation of the growth rate for every $f$ values in the
present study, growth rate for $f$ $=$ $0.001$ to $0.01$ are shown as a
function of $f$ in Fig. 13 (b), for the two regions (1) $\tau$ $=$ $11.11$ to
$19.17$ and (2) $\tau$ $=$ $19.17$ to $34$ as ${\gamma}_{1}$ and
${\gamma}_{2}$ respectively. Linear fit of ${\gamma}_{1}$ and ${\gamma}_{2}$
values with respect to $f$ values shows that, ${\gamma}_{1}$ (${\beta}_{1}$
$=$ $0.821$ $mA/\overline{t}_{D}$) has higher slope value than ${\gamma}_{2}$
(${\beta}_{1}$ $=$ $0.263$ $mA/\overline{t}_{D}$) i.e. the average growth rate
of the wall probe current is higher at the initial time period $\tau$ $=$
$11.11$ to $19.17$. Interestingly, ${\gamma}_{1}$ and ${\gamma}_{2}$ values
decreases with low values of $f$ and approaches to a common value ($\sim$ $f$
$=$ $0.0015$).
As discussed in the Sec. I (Introduction), algebraic growthFajans (1993);
Peurrung, Notte, and Fajans (1993); Bettega _et al._ (2006) is seen in the
ion resonance instability with transient nature of the ion in the cylindrical
trap. In tight aspect ratio toroidal trap experimentLachhvani _et al._ (2016)
with toroidal end plugs, the ions are produced and lost along toroidal
direction and thus are effectively transient. Though the growth rate is high
but slowerLachhvani _et al._ (2016) than theoretically proposed rates by Levy
et al.Levy, Daugherty, and Buneman (1969). Here growth of the wall probe
current has been found out to be algebraic in nature. In our case there is no
continues source of ions but ion loss occurs throughout the simulation, which
is a unique situation never studied before. But it is evident that the nature
of the instability is similar to that of the transient ions.
In the next Subsec. III.6, the global temperature profiles of electron and ion
plasma are described as function time. Also the radial temperature profiles of
the electron plasma and their evolution is addressed.
### III.6 Temperature of the electron and ion plasma
To understand the temperature evolution of electron and ion plasma, volume
averaged parallel and perpendicular temperatures are evaluated through out the
simulation time and explained next. Average temperatures of electron flux
tubes at different radial locations are also found out and corresponding
radial variations are also discussed.
#### III.6.1 Global temperature of the electron plasma
Figure 14: Volume averaged parallel and perpendicular temperatures of the
electron plasma, $\langle T_{\parallel}^{e}\rangle$ and $\langle
T_{\bot}^{e}\rangle$ are shown for case of $f$ $=$ $0.007$ upto $\tau$ $=$
$34$. $\langle T_{\parallel}^{e}\rangle$ $\sim$ $13.8$ eV and $\langle
T_{\bot}^{e}\rangle$ $\sim$ $8.8$ eV at $\tau$ $=$ $11.11$ (indicated by black
dotted line), when the ion plasma was introduced. At $\tau$ $=$ $34$, $\langle
T_{\parallel}^{e}\rangle$ $\sim$ $21$ eV and $\langle T_{\bot}^{e}\rangle$
$\sim$ $17$ eV. Though the rise in the temperature values was almost steady
initially, after $\tau$ $\sim$ $25$ the fluctuations in $\langle
T_{\parallel}^{e}\rangle$ values were $\sim$ $5$ev and for $\langle
T_{\bot}^{e}\rangle$ $\sim$ $2.5$ev. Volume averaged parallel and
perpendicular distribution functions, $\langle f_{e}(v_{\parallel})\rangle$
and $\langle f_{e}(v_{\bot})\rangle$, at $\tau$ $=$ $11.11$, are shown in the
inset plot.
The volume averaged parallel and perpendicular temperatures of the electron
plasma, $\langle T_{\parallel}^{e}\rangle$ and $\langle T_{\bot}^{e}\rangle$
are shown in Fig. 14 for case of $f$ $=$ $0.007$ upto $\tau$ $=$ $34$. As the
electrons were loaded cold, the temperature values start from zero $\tau$ $=$
$0.00$ and reaches to finite temperature values ($\langle
T_{\parallel}^{e}\rangle$ $\sim$ $13.8$ eV and $\langle T_{\bot}^{e}\rangle$
$\sim$ $8.8$ eV), at $\tau$ $=$ $11.11$, at the instant the ion plasma was
introduced. Volume averaged parallel and perpendicular distribution functions,
$\langle f_{e}(v_{\parallel})\rangle$ and $\langle f_{e}(v_{\bot})\rangle$, at
$\tau$ $=$ $11.11$, are shown in the inset plot. After $\tau$ $=$ $11.11$, the
temperature values rise further and attains values as $\langle
T_{\parallel}^{e}\rangle$ $\sim$ $21$ eV and $\langle T_{\bot}^{e}\rangle$
$\sim$ $17$ eV. Though the rise was almost steady initially, after $\tau$
$\sim$ $25$ the fluctuations in $\langle T_{\parallel}^{e}\rangle$ values were
$\sim$ $5$ev and for $\langle T_{\bot}^{e}\rangle$ $\sim$ $2.5$ev. The orders
of these global temperature values are in the same range ($\simeq 10-20~{}eV$)
of the values for pure electron plasma, if the ion plasma was absent from the
system, though the oscillating features at later simulation time represents
the high amplitude oscillation of the electron plasma after $\tau$ $\sim$ $25$
seen in Fig. 12 (a). Volume averaged parallel and perpendicular temperatures
of the ion plasma are also calculated and discussed next.
#### III.6.2 Global temperature of the ion plasma
Figure 15: Volume averaged parallel and perpendicular temperatures of the ion
plasma, $\langle T_{\parallel}^{i}\rangle$ and $\langle T_{\bot}^{i}\rangle$
are shown for case of $f$ $=$ $0.007$ upto $\tau$ $=$ $34$. As seen in Fig.
11, the perpendicular distribution functions of ion plasma form near-
Maxwellian shape around $\tau$ $=$ $19.17$ (indicated by vertical dotted line
in magenta) with $\langle T_{\bot}^{i}\rangle$ $\sim$ $211$ eV at $\tau$ $=$
$19.17$. $\langle T_{\bot}^{i}\rangle$ $\sim$ $280$ eV at $\tau$ $=$ $34$. As
the ions were loaded cold, the $\langle T_{\parallel}^{i}\rangle$ values start
from zero at $\tau$ $=$ $11.11$ and reaches to low order finite temperature
values ($\langle T_{\parallel}^{i}\rangle$ $\sim$ $0.055$ eV ) at $\tau$ $=$
$34$. Steep rise in $\langle T_{\parallel}^{i}\rangle$ values is seen after
$\tau$ $=$ $19.17$. Volume averaged parallel and perpendicular distribution
functions, $\langle f_{i}(v_{\parallel})\rangle$ and $\langle
f_{i}(v_{\bot})\rangle$, at $\tau$ $=$ $34$, are shown in the inset plot.
As seen in Fig. 11, volume averaged temperature evolution of the ion plasma
also helps to clarify the equilibration process of the ion plasma. Thus, the
volume averaged parallel and perpendicular temperatures of the ion plasma,
$\langle T_{\parallel}^{i}\rangle$ and $\langle T_{\bot}^{i}\rangle$ are shown
for case of $f$ $=$ $0.007$ upto $\tau$ $=$ $34$. $\tau$ $=$ $19.17$ is
indicated by vertical dotted line in magenta. As seen in Fig. 11, the
perpendicular distribution functions of ion plasma form near-Maxwellian shape
around $\tau$ $=$ $19.17$ with $\langle T_{\bot}^{i}\rangle$ $\sim$ $211$ eV
at $\tau$ $=$ $19.17$. After $\tau$ $=$ $19.17$, $\langle T_{\bot}^{i}\rangle$
values increases and attains $\sim$ $280$ eV at $\tau$ $=$ $34$. As the ions
were loaded cold, the $\langle T_{\parallel}^{i}\rangle$ values start from
zero at $\tau$ $=$ $11.11$ and reaches to low order finite temperature values
($\langle T_{\parallel}^{i}\rangle$ $\sim$ $0.055$ eV) at $\tau$ $=$ $34$.
However, steep rise in $\langle T_{\parallel}^{i}\rangle$ values is seen after
$\tau$ $=$ $19.17$ suggesting energy gain of the ion plasma, though very small
in the parallel direction. Volume averaged parallel and perpendicular
distribution functions, $\langle f_{i}(v_{\parallel})\rangle$ and $\langle
f_{i}(v_{\bot})\rangle$, at $\tau$ $=$ $34$, are shown in the inset plot.
#### III.6.3 Radial variation of the electron plasma temperature
Figure 16: Averaged temperatures ${\bar{T}}_{\parallel}^{e}(R,t)$ and
${\bar{T}}_{\bot}^{e}(R,t)$ are shown as functions of $R$, at $\tau$ $=$
$16.67$ and $\tau$ $=$ $25.56$ for $f$ $=$ $0.007$ case. Z-averaged quantities
are expressed as ${\bar{T}}_{\parallel}^{e}(R,t)(Z_{avg})$,
${\bar{T}}_{\bot}^{e}(R,t)(Z_{avg})$ and flux tube averaged quantities are
expressed as ${\bar{T}}_{\parallel}^{e}(R,t)(Fluxtube)$,
${\bar{T}}_{\bot}^{e}(R,t)(Fluxtube)$ in all of these Figures. The magenta
colored vertical lines represents $1/e$ times value of the peak electron
density. For high accuracy the analysis is performed over different sized bins
along $R$ and the respective range of the temperatures are shown by the error
bars in these Figures. In (a) and (b),
${\bar{T}}_{\parallel}^{e}(R,t)(Z_{avg})$ and
${\bar{T}}_{\parallel}^{e}(R,t)(Fluxtube)$ values are higher at the inboard
side and gradually fall with $R$. The temperature values are fitted with
$\alpha/R^{2}$ and $\alpha_{2}/R^{2}$ ($\alpha$ and $\alpha_{2}$ are an
arbitrary constants) showing $1/R^{2}$ scaling with $R$. In (c) and (d),
${\bar{T}}_{\bot}^{e}(R,t)(Z_{avg})$ and ${\bar{T}}_{\bot}^{e}(R,t)(Fluxtube)$
values are also higher at the inboard side but showing non-monotonic variation
with $R$. The average temperature values are fitted with $\beta/R$ and
$\beta_{2}/R$ ($\beta$ and $\beta_{2}$ are arbitrary constants) showing $1/R$
scaling with $R$.
The electron plasma performs a center of charge motion in the poloidal plane,
the toroidal electron flux tubes passes across the $1/R$ varying toroidal
magnetic field resulting in periodic compression–expansion cycle of the tubes
(which is also known as magnetic pumping). Under this cross-field dynamics,
mechanisms like electron-electron large angle collisions, results into
asymmetric exchange of energies between parallel and perpendicular directions.
As a result, the parallel and perpendicular temperatures of the electron flux
tubes undergo radial variations which, in general, scales inversely with $R$.
As our simulation is collisionless, we believe that there exists an underlying
collisionless process, which introduces asymmetry in the compression–expansion
cycle (i.e. MP), resulting in collisionless exchange of energy between
parallel and perpendicular directions. More details of this mechanism remain
to be explored in future. Extensive analysis of the radial variation of
parallel and perpendicular temperature of pure toroidal electron plasma had
been performed in our previous work with QQS state. It was found out that the
parallel temperature of toroidal electron plasma flux tubes scales as
$1/R{{}^{2}}$ and the perpendicular temperature scales as $1/R$, which is
consistent with the angular momentum conservation and mean magnetic moment
conservation principles. In the present study, we have observed similar
scaling laws despite the presence of the ion plasmas in the system. As
performed in the previous study, parallel and perpendicular temperatures are
averaged in two different ways: 1. In toroidal and vertical directions
(Z-averaged). 2. In toroidal direction only, around mid-Z plane (averaged over
toroidal flux tubes). In the first method, particles from different vertical
coordinates contribute in the calculations. In the second method, particles
within different toroidal flux tubes along with radial coordinates, around
mid-z plane have been considered. In the second method, the particles within a
flux tube at certain R may be expected to follow similar dynamics to each
other. In Fig. 16 (a)-(d), these averaged temperatures
${\bar{T}}_{\parallel}^{e}(R,t)$ (Fig. 16 (a),(b)) and
${\bar{T}}_{\bot}^{e}(R,t)$ (Fig. 16 (c),(d)) are shown as functions of $R$,
at two different time periods $\tau$ $=$ $16.67$ and $\tau$ $=$ $25.56$ for
$f$ $=$ $0.007$ case. Z-averaged quantities are expressed as
${\bar{T}}_{\parallel}^{e}(R,t)(Z_{avg})$,
${\bar{T}}_{\bot}^{e}(R,t)(Z_{avg})$ and flux tube averaged quantities are
expressed as ${\bar{T}}_{\parallel}^{e}(R,t)(Fluxtube)$,
${\bar{T}}_{\bot}^{e}(R,t)(Fluxtube)$ in all of these Figures. The magenta
colored vertical lines represents $1/e$ times value of the peak electron
density. For high accuracy the analysis is performed over different sized bins
along $R$ and the respective range of the temperatures are shown by the error
bars in these figures. In Fig. 16 (a) and (b),
${\bar{T}}_{\parallel}^{e}(R,t)(Z_{avg})$ and
${\bar{T}}_{\parallel}^{e}(R,t)(Fluxtube)$ values are higher at the inboard
side and gradually fall with $R$. The temperature values are fitted with
$\alpha/R^{2}$ and $\alpha_{2}/R^{2}$ ($\alpha$ and $\alpha_{2}$ are an
arbitrary constants) showing $1/R^{2}$ scaling with $R$. In Fig. 16 (c) and
(d), ${\bar{T}}_{\bot}^{e}(R,t)(Z_{avg})$ and
${\bar{T}}_{\bot}^{e}(R,t)(Fluxtube)$ values are also higher at the inboard
side but showing non-monotonic variation with $R$. The average temperature
values are fitted with $\beta/R$ and $\beta_{2}/R$ ($\beta$ and $\beta_{2}$
are arbitrary constants) showing $1/R$ scaling with $R$. In these plots, the
scaling laws are preserved at later simulation times representing the
conservation of angular momentum and mean magnetic moment, thus suggesting the
conserved quantities of the system are not affected by the introduction of ion
plasma in the system. The reason for this weak non-monotonicity in the
perpendicular temperature profile is unknown and also seen in the analysis of
pure electron plasma in QQSKhamaru, Ganesh, and Sengupta (2021b) state.
## IV Conclusions and Discussions
In this paper, toroidal electron plasmas in a tight aspect ratio axisymmetric
toroidal device to the presence of small fraction of ions, has been
investigated with a three dimensional collisionless Particle-in-Cell
simulation study with $Ar^{+}$ ions as the ion species. The electron plasma is
loaded with a “seed” solution obtained by zero-inertia entropy maximization
and initialized in the PIC code to evolve to attain a quiescent quasi-steady
state (QQS). After a QQS state is attained, $Ar^{+}$ ions and secondary
electrons are introduced cold into the system. $Ar^{+}$ ions of different
number density values i.e. $f$ values (0.001 to 0.01) are used and the
dependency of the instability on the ion density is addressed via different
diagnostics methods.
We have demonstrated that the electron plasma dynamics and the ion plasma
dynamics are affected by presence of small number of ions, for mainly $f$
values $>$ 0.001 i.e. for such ion population in the system results in growth
of the wall probe current after the ions are introduced. Thus the results are
mainly interpreted on the basis of high $f$ value datum ($f$ $\geq$ 0.005).
The growth of the wall probe current is algebraic in nature and saturates at
later simulation time, when a substantial amount of ions are lost. In our
preloaded ion case there is no continuous source of ions but ions are being
lost with simulation time. This is a unique situation not belonging to
strictly trapped ions or transient ions in the system. Spatial evolution of
the electron plasma and corresponding spectrogram analysis demonstrate the
destabilized “center of charge motion” ($m$ $=$ $1$) of the electron plasma in
poloidal plane along with chirped coupling between toroidal Diocotron modes
($m$ $=$ $1$ to $9$ for $f$ $=$ $0.01$) with $m$ $=$ $2$ as dominant mode.
Investigation of the evolution of the three dimensional potential well of the
system, reveals interesting features. Based on the above, we have argued that
the ion fraction destabilizes the electron plasma in QQS state via a toroidal
ion resonance instability mechanism.
The variations in the ion and electron plasma energies show collisionless
transfer of electron potential energy to the kinetic energy of the ion plasma.
Though ions start to exhibit a dynamic nature in the poloidal direction,
toroidal motion of the ions are negligible. This is understandable as the
$E_{\parallel}=-\nabla_{\parallel}\phi\sim 0$, there is no parallel
acceleration or energizing dynamics for ions in the parallel direction. After
initial energy gain, the ion plasmas start to equilibrate and attains near-
Maxwellian distribution in the poloidal plane. We have identified a certain
time period after which the ion energy gain is trivial. Correlating the wall
probe growth in the same time period, it has been found out that the wall
probe current also starts to saturate after that time period. Also a linear
scaling of the growth rate of the instability with respect to the ion density
($f$ values) has been obtained for the above mentioned two time windows.
In this study, volume averaged parallel and perpendicular temperatures are
defined and used to calculate the temperatures of the ion and electron
plasmas. The volume averaged perpendicular temperature of the ion plasma is
obtained when ions start to equilibrate, shows high temperature value,
indicating ion heating due to gain of energy via the toroidal ion-resonance
instability. However, the exact process of equilibration of ions resulting
Maxwellian distributions is not clear. The volume averaged temperature values
of the electron plasma rises with simulation time though attaining a
fluctuating nature near the end of the simulation time.
For two different simulation time periods, the averaged temperatures
${\bar{T}}_{\parallel}^{e}$ and ${\bar{T}}_{\bot}^{e}$ show $1/R^{2}$ and
$1/R$ scaling with $R$ respectively, as can be expected from conservation of
angular momentum and conservation of $\mu$.
We have demonstrated toroidal ion resonance instability of toroidal electron
plasma and the underlying processes through extensive analysis and showed that
under controlled ion population, the growth of the instability can be
addressed via a scaling law of the growth rate with the ion density value. In
experimental conditions, ions are produced continuously via electron-impact
ionization and other collisional processes play a crucial role in the energy
exchange processes. An investigation considering the impact ionization,
electron-neutral and ion-neutral collision operations is required. In an
ongoing study we have implemented these collisional processes and the results
will be reported soon.
###### Acknowledgements.
All numerical results reported in this study were obtained using the ANTYA
cluster at IPR. The authors would like to thank the staff of the Computer
Center of IPR for facilities of the ANTYA HPC. The authors would like to thank
Dr. Nirmal Kumar Bisai for careful reading of the manuscript and for his
comments.
## V Data availability
The data that support the findings of this study are available from the
corresponding author upon reasonable request.
## References
* Malmberg and deGrassie (1975) J. H. Malmberg and J. S. deGrassie, “Properties of nonneutral plasma,” Phys. Rev. Lett. 35, 577–580 (1975).
* Paul (1990) W. Paul, “Electromagnetic traps for charged and neutral particles (nobel lecture),” Angewandte Chemie International Edition in English 29, 739–748 (1990), https://onlinelibrary.wiley.com/doi/pdf/10.1002/anie.199007391 .
* Yamada and Himura (2016) S. Yamada and H. Himura, “Note: Consecutive capture of images of ions and electrons using high-voltage vacuum relay,” Review of Scientific Instruments 87, 036109 (2016), https://aip.scitation.org/doi/pdf/10.1063/1.4944861 .
* Hurst _et al._ (2014) N. C. Hurst, J. R. Danielson, C. J. Baker, and C. M. Surko, “Electron plasma orbits from competing Diocotron drifts,” Phys. Rev. Lett. 113, 025004 (2014).
* Lane and Orperformedz (2019) R. A. Lane and C. A. Orperformedz, “Electrostatic equilibria of non-neutral plasmas confined in a penning trap with axially varying magnetic field,” Physics of Plasmas 26, 052511 (2019), https://doi.org/10.1063/1.5092136 .
* Davidson and Felice (1998) R. C. Davidson and G. M. Felice, “Influence of profile shape on the Diocotron instability in a non-neutral plasma column,” Physics of Plasmas 5, 3497–3511 (1998), https://doi.org/10.1063/1.873067 .
* O’Neil and Dubin (1998) T. M. O’Neil and D. H. E. Dubin, “Thermal equilibria and thermodynamics of trapped plasmas with a single sign of charge,” Physics of Plasmas 5, 2163–2193 (1998), https://doi.org/10.1063/1.872925 .
* Landa (2019) H. Landa, “Tuning nonthermal distributions to thermal ones in time-dependent paul traps,” Phys. Rev. A 100, 013413 (2019).
* Greaves and Surko (2000) R. G. Greaves and C. M. Surko, “Inward transport and compression of a positron plasma by a rotating electric field,” Phys. Rev. Lett. 85, 1883–1886 (2000).
* Ahmadi and et al. (2016) M. Ahmadi, M. Baquero-Ruiz, W. Bertsche, E. Butler, A. Capra, C. Carruth, C. L. Cesar, M. Charlton, A. E. Charman, S. Eriksson et al., Nature 529, 373–376 (2016).
* Ahmadi _et al._ (2018) M. Ahmadi, B. X. R. Alves, C. J. Baker, W. Bertsche, A. Capra, C. Carruth, C. L. Cesar, M. Charlton, S. Cohen, R. Collister et al., (ALPHA Collaboration), “Enhanced control and reproducibility of non-neutral plasmas,” Phys. Rev. Lett. 120, 025001 (2018).
* O’Neil and Driscoll (1979) T. M. O’Neil and C. F. Driscoll, “Transport to thermal equilibrium of a pure electron plasma,” The Physics of Fluids 22, 266–277 (1979), https://aip.scitation.org/doi/pdf/10.1063/1.862577 .
* Malmberg and Driscoll (1980) J. H. Malmberg and C. F. Driscoll, “Long-time containment of a pure electron plasma,” Phys. Rev. Lett. 44, 654–657 (1980).
* Driscoll, Malmberg, and Fine (1988) C. F. Driscoll, J. H. Malmberg, and K. S. Fine, “Observation of transport to thermal equilibrium in pure electron plasmas,” Phys. Rev. Lett. 60, 1290–1293 (1988).
* Dubin and O’Neil (1999) D. H. E. Dubin and T. M. O’Neil, Rev. Mod. Phys. 71, 87–172 (1999).
* Fine _et al._ (1995) K. S. Fine, A. C. Cass, W. G. Flynn, and C. F. Driscoll, Phys. Rev. Lett. 75, 3277–3280 (1995).
* Briggs, Daugherty, and Levy (1970) R. J. Briggs, J. D. Daugherty, and R. H. Levy, “Role of landau damping in crossed‐field electron beams and inviscid shear flow,” The Physics of Fluids 13, 421–432 (1970), https://aip.scitation.org/doi/pdf/10.1063/1.1692936 .
* Ganesh and Pahari (2006) R. Ganesh and S. Pahari, in _Frontiers in Turbulence and Coherent Structures_, Vol. 6 (World Scientific, 2006) p. 471.
* Rome _et al._ (2000) M. Rome, M. Brunetti, F. Califano, F. Pegoraro, and R. Pozzoli, “Motion of extended vortices in an inhomogeneous pure electron plasma,” Physics of Plasmas 7, 2856–2865 (2000), https://doi.org/10.1063/1.874135 .
* Sengupta and Ganesh (2014) M. Sengupta and R. Ganesh, “Inertia driven radial breathing and nonlinear relaxation in cylindrically confined pure electron plasma,” Physics of Plasmas 21, 022116 (2014), https://doi.org/10.1063/1.4866022 .
* Birdsall and Langdon (2004) C. Birdsall and A. Langdon, _Plasma Physics via Computer Simulation_, Series in Plasma Physics and Fluid Dynamics (Taylor & Francis, 2004).
* Niemann _et al._ (2019) M. Niemann, T. Meiners, J. Mielke, M. J. Borchert, J. M. Cornejo, S. Ulmer, and C. Ospelkaus, Measurement Science and Technology 31, 035003 (2019).
* Schuh _et al._ (2019) M. Schuh, F. Heiße, T. Eronen, J. Ketter, F. Köhler-Langes, S. Rau, T. Segal, W. Quint, S. Sturm, and K. Blaum, “Image charge shift in high-precision penning traps,” Phys. Rev. A 100, 023411 (2019).
* Gutiérrez _et al._ (2019) M. J. Gutiérrez, J. Berrocal, F. Domínguez, I. Arrazola, M. Block, E. Solano, and D. Rodríguez, “Dynamics of an unbalanced two-ion crystal in a penning trap for application in optical mass spectrometry,” Phys. Rev. A 100, 063415 (2019).
* Peurrung, Kouzes, and Barlow (1996) A. Peurrung, R. Kouzes, and S. Barlow, “The non-neutral plasma: an introduction to physics with relevance to cyclotron resonance mass spectrometry,” International Journal of Mass Spectrometry and Ion Processes 157-158, 39 – 83 (1996).
* Fajans and Surko (2020) J. Fajans and C. M. Surko, “Plasma and trap-based techniques for science with antimatter,” Physics of Plasmas 27, 030601 (2020), https://doi.org/10.1063/1.5131273 .
* Morigi _et al._ (1999) G. Morigi, J. Eschner, J. I. Cirac, and P. Zoller, “Laser cooling of two trapped ions: Sideband cooling beyond the lamb-dicke limit,” Phys. Rev. A 59, 3797–3808 (1999).
* Morigi and Walther (2001) G. Morigi and H. Walther, “Two-species coulomb chains for quantum information,” The European Physical Journal D - Atomic, Molecular, Optical and Plasma Physics 13 (2001), 10.1007/s100530170275.
* Douglas, Frank, and Mao (2005) D. J. Douglas, A. J. Frank, and D. Mao, “Linear ion traps in mass spectrometry,” Mass Spectrometry Reviews 24, 1–29 (2005) .
* Greaves and Surko (2002) R. G. Greaves and C. M. Surko, “Practical limits on positron accumulation and the creation of electron-positron plasmas,” AIP Conference Proceedings 606, 10–23 (2002), https://aip.scitation.org/doi/pdf/10.1063/1.1454263 .
* Higaki _et al._ (2017) H. Higaki, C. Kaga, K. Fukushima, H. Okamoto, Y. Nagata, Y. Kanai, and Y. Yamazaki, “Simultaneous confinement of low-energy electrons and positrons in a compact magnetic mirror trap,” New Journal of Physics 19, 023016 (2017).
* Hicks, Bowman, and Godden (2019) N. K. Hicks, A. Bowman, and K. Godden, “Particle-in-cell simulation of quasi-neutral plasma trapping by rf multipole electric fields,” Physics 1, 392–401 (2019).
* Nakajima, Himura, and Sanpei (2021) Y. Nakajima, H. Himura, and A. Sanpei, “Counter differential rigid-rotation equilibrium of electrically non-neutral two-fluid plasma with finite pressure,” Journal of Plasma Physics 87, 905870415 (2021).
* Janes _et al._ (1966) G. S. Janes, R. H. Levy, H. A. Bethe, and B. T. Feld, Phys. Rev. 145, 925–952 (1966).
* Daugherty, Eninger, and Janes (1969) J. D. Daugherty, J. E. Eninger, and G. S. Janes, “Experiments on the injection and containment of electron plasmas in a toroidal apparatus,” The Physics of Fluids 12, 2677–2693 (1969), https://aip.scitation.org/doi/pdf/10.1063/1.1692411 .
* Clark _et al._ (1976) W. Clark, P. Korn, A. Mondelli, and N. Rostoker, Phys. Rev. Lett. 37, 592–595 (1976).
* Zaveri _et al._ (1992) P. Zaveri, P. I. John, K. Avinash, and P. K. Kaw, “Low-aspect-ratio toroidal equilibria of electron plasmas,” Phys. Rev. Lett. 68, 3295–3298 (1992).
* Khirwadkar _et al._ (1993) S. S. Khirwadkar, P. I. John, K. Avinash, A. K. Agarwal, and P. K. Kaw, “Steady state formation of a toroidal electron plasma,” Phys. Rev. Lett. 71, 4334–4337 (1993).
* Lachhvani _et al._ (2017) L. Lachhvani, S. Pahari, S. Sengupta, Y. G. Yeole, M. Bajpai, and P. K. Chattopadhyay, Physics of Plasmas 24, 102132 (2017), https://doi.org/10.1063/1.5009013 .
* Stoneking _et al._ (2004) M. R. Stoneking, M. A. Growdon, M. L. Milne, and R. T. Peterson, Phys. Rev. Lett. 92, 095003 (2004).
* Davidson (1990) R. C. Davidson, _An introduction to the physics of nonneutral plasmas._ (IEEE, 1990).
* Marksteiner _et al._ (2008) Q. R. Marksteiner, T. S. Pedersen, J. W. Berkery, M. S. Hahn, J. M. Mendez, B. Durand de Gevigney, and H. Himura, Phys. Rev. Lett. 100, 065002 (2008).
* Saitoh _et al._ (2010) H. Saitoh, Z. Yoshida, J. Morikawa, Y. Yano, H. Hayashi, T. Mizushima, Y. Kawai, M. Kobayashi, and H. Mikami, “Confinement of electron plasma by levitating dipole magnet,” Physics of Plasmas 17, 112111 (2010), https://doi.org/10.1063/1.3514207 .
* Khamaru, Sengupta, and Ganesh (2019) S. Khamaru, M. Sengupta, and R. Ganesh, “Dynamics of a toroidal pure electron plasma using 3d pic simulations,” Physics of Plasmas 26, 112106 (2019), https://doi.org/10.1063/1.5111747 .
* Khamaru, Ganesh, and Sengupta (2021a) S. Khamaru, R. Ganesh, and M. Sengupta, “A novel quiescent quasi-steady state of a toroidal electron plasma,” Physics of Plasmas 28, 042101 (2021a), https://doi.org/10.1063/5.0032880 .
* Khamaru, Ganesh, and Sengupta (2021b) S. Khamaru, R. Ganesh, and M. Sengupta, “Erratum: “a novel quiescent quasi-steady state of a toroidal electron plasma” [phys. plasmas 28, 042101 (2021)],” Physics of Plasmas 28, 119901 (2021b), https://doi.org/10.1063/5.0074384 .
* Crooks and O’Neil (1996) S. M. Crooks and T. M. O’Neil, “Transport in a toroidally confined pure electron plasma,” Physics of Plasmas 3, 2533–2537 (1996), https://doi.org/10.1063/1.871971 .
* Brambilla _et al._ (2018) G. Brambilla, C. Kalapotharakos, A. N. Timokhin, A. K. Harding, and D. Kazanas, “Electron–positron pair flow and current composition in the pulsar magnetosphere,” The Astrophysical Journal 858, 81 (2018).
* Stoneking _et al._ (2020) M. R. Stoneking, T. S. Pedersen, P. Helander, H. Chen, U. Hergenhahn, E. V. Stenson, G. Fiksel, J. von der Linden, H. Saitoh, C. M. Surko, and et al., “A new frontier in laboratory physics: magnetized electron–positron plasmas,” Journal of Plasma Physics 86, 155860601 (2020).
* Singer _et al._ (2021) M. Singer, M. R. Stoneking, E. V. Stenson, S. Nißl, A. Deller, A. Card, J. Horn-Stanja, T. Sunn Pedersen, H. Saitoh, and C. Hugenschmidt, “Injection of positrons into a dense electron plasma in a magnetic dipole trap,” Physics of Plasmas 28, 062506 (2021), https://doi.org/10.1063/5.0050881 .
* White, Malmberg, and Driscoll (1982) W. D. White, J. H. Malmberg, and C. F. Driscoll, “Resistive-wall destabilization of Diocotron waves,” Phys. Rev. Lett. 49, 1822–1826 (1982).
* Davidson and Chao (1996) R. C. Davidson and E. H. Chao, “l=1 electrostatic instability induced by electron‐neutral collisions in a nonneutral electron plasma interacting with background neutral gas,” Physics of Plasmas 3, 3279–3287 (1996), https://doi.org/10.1063/1.871610 .
* Levy, Daugherty, and Buneman (1969) R. H. Levy, J. D. Daugherty, and O. Buneman, “Ion resonance instability in grossly nonneutral plasmas,” The Physics of Fluids 12, 2616–2629 (1969), https://aip.scitation.org/doi/pdf/10.1063/1.1692404 .
* Davidson and Uhm (1977) R. C. Davidson and H. Uhm, “Influence of strong self‐electric fields on the ion resonance instability in a nonneutral plasma column,” The Physics of Fluids 20, 1938–1946 (1977), https://aip.scitation.org/doi/pdf/10.1063/1.861813 .
* Fajans (1993) J. Fajans, “Transient ion resonance instability,” Physics of Fluids B: Plasma Physics 5, 3127–3135 (1993), https://doi.org/10.1063/1.860649 .
* Peurrung, Notte, and Fajans (1993) A. J. Peurrung, J. Notte, and J. Fajans, “Observation of the ion resonance instability,” Phys. Rev. Lett. 70, 295–298 (1993).
* Lachhvani _et al._ (2016) L. Lachhvani, S. Pahari, R. Goswami, M. Bajpai, Y. Yeole, and P. K. Chattopadhyay, Physics of Plasmas 23, 062109 (2016), https://doi.org/10.1063/1.4953440 .
* Eckhouse, Fisher, and Rostoker (1981) S. Eckhouse, A. Fisher, and N. Rostoker, “Possible observation of the ion‐resonance instability in a non‐neutral electron plasma,” Applied Physics Letters 38, 318–320 (1981), https://doi.org/10.1063/1.92354 .
* Bettega _et al._ (2006) G. Bettega, F. Cavaliere, M. Cavenago, F. De Luca, A. Illiberi, R. Pozzoli, and M. Romé, “Active control of the ion resonance instability by ion removing fields,” Physics of Plasmas 13, 112102 (2006), https://doi.org/10.1063/1.2363175 .
* Kabantsev and Driscoll (2007) A. Kabantsev and C. Driscoll, “Ion-induced instability of Diocotron modes in electron plasmas modelling curvature-driven flute modes,” Fusion Science and Technology 51, 96–99 (2007), https://doi.org/10.13182/FST07-A1324 .
* Stoneking _et al._ (2002) M. R. Stoneking, P. W. Fontana, R. L. Sampson, and D. J. Thuecks, “Electron plasmas in a “partial” torus,” Physics of Plasmas 9, 766–771 (2002), https://doi.org/10.1063/1.1445425 .
* Marksteiner _et al._ (2008) Q. R. Marksteiner, T. S. Pedersen, J. W. Berkery, M. S. Hahn, J. M. Mendez, B. Durand de Gevigney, and H. Himura, “Observations of an ion-driven instability in non-neutral plasmas confined on magnetic surfaces,” Phys. Rev. Lett. 100, 065002 (2008).
* Sengupta and Ganesh (2015) M. Sengupta and R. Ganesh, “Linear and nonlinear evolution of the ion resonance instability in cylindrical traps: A numerical study,” Physics of Plasmas 22, 072112 (2015), https://doi.org/10.1063/1.4927126 .
* Sengupta and Ganesh (2016) M. Sengupta and R. Ganesh, “Influence of electron-neutral elastic collisions on the instability of an ion-contaminated cylindrical electron plasma: 2d3v pic-with-mcc simulations,” Physics of Plasmas 23, 102111 (2016), https://doi.org/10.1063/1.4964913 .
* Sengupta and Ganesh (2017) M. Sengupta and R. Ganesh, “Destabilization of a cylindrically confined electron plasma by impact ionization of background neutrals: 2d3v pic simulation with monte-carlo-collisions,” Physics of Plasmas 24, 032105 (2017), https://doi.org/10.1063/1.4978473 .
* Khamaru, Sengupta, and Ganesh (2019) S. Khamaru, M. Sengupta, and R. Ganesh, “Dynamics of a toroidal pure electron plasma using 3d pic simulations,” Physics of Plasmas 26, 112106 (2019), https://doi.org/10.1063/1.5111747 .
* Sengupta, Khamaru, and Ganesh (2021) M. Sengupta, S. Khamaru, and R. Ganesh, “Self-organization of pure electron plasma in a partially toroidal magnetic-electrostatic trap: A 3d particle-in-cell simulation,” Journal of Applied Physics 130, 133305 (2021), https://doi.org/10.1063/5.0055828 .
* Huba (2013) J. D. Huba, _Plasma Physics_ (Naval Research Laboratory, Washington, DC, 2013) p. 28.
|
# Radiative stability of tiny cosmological constant from the Swampland and
quantized compactification
Cao H. Nam<EMAIL_ADDRESS>Phenikaa Institute for Advanced
Study and Faculty of Fundamental Sciences, Phenikaa University, Yen Nghia, Ha
Dong, Hanoi 12116, Vietnam
###### Abstract
We address a quantization mechanism that can allow us to understand why the
cosmological constant is not large under the quantum corrections from studying
the circle compactification solution of the Standard Model coupled to Einstein
gravity which is subject to the constraint of the Swampland conjectures. A
novel result in the present work compared to the previous investigations in
the literature is that the radius of the compactified dimension and the 4D
cosmological constant $\Lambda_{4}$ must in fact be quantized. The
quantization rule of the cosmological constant is given by $\Lambda_{4}\propto
n^{2}$ with $n=1,2,3,4,\ldots$, which means that the values of $\Lambda_{4}$
are not arbitrary but only its specific values are allowed. In general, the
quantum corrections as well as other effects would break this quantization
rule. Hence, it could prevent the quantum fluctuations from generating zero-
point energy contributions to the cosmological constant.
The current cosmological and astrophysical observations Planck2018 have
implied a positive cosmological constant of the value around $\sim
10^{-120}M^{4}_{\text{Pl}}\simeq\left(2.4\times 10^{-3}\text{eV}\right)^{4}$
Tanabashi2018 (responsible for the late accelerating expansion of the
universe) which is a very small number compared to the electroweak scale
around $\sim 10^{2}$ GeV and the Planck energy scale $M_{\text{Pl}}\sim
10^{19}$ GeV. There is no problem at all with respect to having such a tiny
value for the cosmological constant at the classical level. However, the
problem arises when the quantum corrections (and other corrections such as
non-perturbative effects coming from the QCD instantons) for the cosmological
constant are included, which would make the value of the cosmological constant
much bigger, and hence its small value is not stable to the large quantum
corrections Zeldovich6768 ; Birrell1994 . For instance, with the cutoff at the
Planck energy scale, the quantum fluctuations would additionally contribute to
the cosmological constant with a quantity around $\sim M^{4}_{\text{Pl}}$. In
order to yield the correct physical result, a cancellation between the bare
cosmological constant and the quantum corrections must be fine-tuned to an
unimaginable precision which is around $\sim 120$ orders of magnitude in
order. Therefore, the cosmological constant problem has been regarded as one
of the most important problems in modern physics Weinberg1989 ; Bousso2008 .
The traditional approach to understanding why the cosmological constant is not
large under quantum corrections is based on a certain symmetry (either
continuous or discrete), which provides a mechanism to cancel the quantum
contributions precisely. The most promising symmetry is perhaps supersymmetry
which if unbroken would lead to a very precise cancellation between boson and
fermions contributing to the vacuum energy since their contributions are
opposite signs. Unfortunately, the superpartners of the Standard Model (SM)
particles have not been observed so far since supersymmetry must be
spontaneously broken down at least at the TeV scale. It was also suggested
that the one-loop contribution to the cosmological constant might vanish in
certain string models with broken supersymmetry Kachru1998 ; Kachru1999 .
However, it was argued that the higher-order quantum corrections would spoil
this vanishing Harvey1999 ; Iengo2000 . In addition, these string models need
to have the supersymmetric non-Abelian sector Blumenhagen ; Angelantonj1999 ;
Angelantonj2004 . Another promising symmetry to solve the cosmological
constant problem is scale invariance or conformal symmetry which forbids the
quantities or the terms in Lagrangian with the length scale dimension and
hence the cosmological constant has to be zero. However, the observed world
simply is not invariant under the scale transformation. For other ideas which
have been proposed in the literature, see Nobbenhuis2006 ; Burgess2013 for
reviews. Although there have been attempts which have been made to solve the
cosmological constant problem there is indeed no known solution to this
problem at present.
The Swampland program aims Vafa2005 ; Ooguri2007 ; Palti2019 to seek the
universal features of quantum gravity and the consistency criteria which
distinguish effective field theories UV-completed consistent with quantum
gravity from those which cannot or are in the Swampland. In particular, this
program leads to interesting implications for inflationary cosmology
(corresponding to the early accelerating expansion of the universe) which is
constrained by the Swampland criterion for de Sitter Obied2018 ; Garg2019 and
trans-Planckian censorship conjecture Bedroya2020 ; Brandenberger2021 . In
this paper, based on the Swampland program we will point to a novel and simple
mechanism which could forbid quantum fluctuations (as well as other effects)
to generate zero-point energy contributions to the cosmological constant.
Although this mechanism is far from a solution to the cosmological constant
problem, we hope that it could give a step toward a complete solution.
If a theory satisfies the Swampland constraints and is completed into quantum
gravity in UV, then a lower-dimensional theory that is obtained from its
compactification to lower dimensions must do as well. On the contrary, if the
theory upon compactification is inconsistent with quantum gravity then its
parent higher dimensional theory itself is pathological. Because the SM
coupled to General Relativity (GR) is a good effective low energy theory and
thus cannot be in the Swampland, its compactification solution down to lower
dimensions Arkani-Hamed2007 ; Arnold2010 ; Fornal2011 must be consistent with
quantum gravity. In this sense, the Swampland conjectures impose the
constraints not only on the SM coupled to GR itself but also on its
compactification solution. This has motivated the intensive investigations in
the literature with the significant implications between the cosmological
constant and the neutrino masses Martin-Lozano2017 ; Valenzuela2017 ;
Hamada2017 ; Gonzalo2018a ; Gonzalo2018b arising when applying the non-SUSY
AdS conjecture Ooguri2017 (motivated by the weak gravity conjecture Arkani-
Hamed2007b ) and the AdS distance conjecture Lust2019 (realized in string
theory and also supported by the evidence of bottom-up physics Rudelius2021 ;
Gonzalo2021 ; Nam2022 ).
In the present work, in this direction, we study the compactification of the
SM coupled to GR on a circle $S^{1}$. However, a new point here is that we
consider the compactified coordinate dependence of the 3D tensor component of
the 4D metric in a natural way which is usually ignored in the literature.
Interestingly, from investigating the wavefunction profile of the 3D tensor
component along the compactified dimension we find a novel aspect that the
radius of the compactified dimension and the 4D cosmological constant are in
fact quantized. Here, we obtain a quantization rule of the 4D cosmological
constant $\Lambda_{4}$ as $\Lambda_{4}\propto n^{2}$ with $n=1,2,3,4,\ldots$.
In this way, the values of the 4D cosmological constant are not arbitrary but
only its specific values are allowed. This quantization or this discrete
spectrum could prevent the quantum fluctuations from generating zero-point
energy contributions to the 4D cosmological constant because in general, the
additional contribution from the quantum fluctuations would not respect the
quantization rule. This is in analogy to that the hydrogen atom would be
unable to absorb the energy which is not equal to the energy difference
between two initial and final states due to the quantization of its energy
levels.
The action of the SM coupled to GR in the presence of the observed
cosmological constant $\Lambda_{4}$ is given as
$\displaystyle S$ $\displaystyle=$ $\displaystyle\int
d^{4}x\sqrt{-g}\left[\frac{M^{2}_{\text{Pl}}}{2}\left(\mathcal{R}^{(4)}-2\Lambda_{4}\right)+\mathcal{L}_{SM}\right],$
(1)
where $\mathcal{R}^{(4)}$ and $\mathcal{L}_{SM}$ refer to the scalar curvature
of 4D spacetime and Lagrangian of the SM, respectively. We consider the above
system compactified on a circle $S^{1}$ where one of three spatial dimensions
of 4D spacetime is periodic with the period of $2\pi$, which means that the
points $x^{3}$ and $x^{3}+2\pi$ are identified. The 4D metric is generally
parametrized in terms of the 3D fields as follows
$\displaystyle
ds^{2}_{4}=g_{ij}dx^{i}dx^{i}+R^{2}\left[dx^{3}+g_{{}_{A}}A_{i}dx^{i}\right]^{2},$
(2)
where $g_{ij}$ is the 3D metric component, $A_{i}$ the graviphoton, $R$ is the
radion field which determines the radius of $S^{1}$, $g_{{}_{A}}$ is a
coupling constant, and $i,j=0,1,2$.111We use the metric sign as $(-,+,+,+)$.
Using this ansatz of the 4D metric, we can expand explicitly the 4D scalar
curvature $\mathcal{R}^{(4)}$ given in the action (1) as
$\displaystyle\mathcal{R}^{(4)}$ $\displaystyle=$
$\displaystyle\mathcal{R}+\frac{1}{4R^{2}}\left(\partial_{3}g^{ij}\partial_{3}g_{ij}+g^{ij}g^{kl}\partial_{3}g_{ij}\partial_{3}g_{kl}\right)$
(3) $\displaystyle-\frac{g^{2}_{{}_{A}}R^{2}}{4}F_{ij}F^{ij},$
where $\mathcal{R}\equiv
g^{ij}\left(\hat{\partial}_{k}\hat{\Gamma}^{k}_{ji}-\hat{\partial}_{j}\hat{\Gamma}^{k}_{ki}+\hat{\Gamma}^{k}_{ji}\hat{\Gamma}^{l}_{lk}-\hat{\Gamma}^{k}_{li}\hat{\Gamma}^{l}_{jk}\right)$
with
$\hat{\Gamma}^{k}_{ij}\equiv\frac{g^{kl}}{2}(\hat{\partial}_{i}g_{jl}+\hat{\partial}_{j}g_{il}-\hat{\partial}_{l}g_{ij})$,
$\hat{\partial}_{i}\equiv\partial_{i}-g_{{}_{A}}A_{i}\partial_{3}$, and
$F_{ij}=\partial_{i}A_{j}-\partial_{j}A_{i}$. It is important to emphasize
here that the previous investigations about the compactification of the SM
coupled with GR on $S^{1}$ Arkani-Hamed2007 ; Arnold2010 ; Fornal2011 ;
Martin-Lozano2017 ; Valenzuela2017 ; Hamada2017 ; Gonzalo2018a ; Gonzalo2018b
have not considered the compactified dimension dependence of the 3D metric
$g_{ij}$ and hence the second term in the right-hand side of Eq. (3) was
absent. But, in the present work, we will point out that the presence of this
term would lead to the quantization of the 4D cosmological constant which
could provide a mechanism to prevent the vacuum fluctuations as well as other
effects contributing additionally to the cosmological constant.
Let us find the wavefunction profile of the 3D metric $g_{ij}$. In order to do
this, first we obtain the equation of motion for $g_{ij}$ in the vacuum
$R=\text{constant}$, $A_{i}=0$, and the vanishing matter fields by varying the
action (1) in the 3D tensor component of the 4D metric as follows
$\displaystyle\bar{\mathcal{R}}_{ij}-\frac{1}{2}g_{ij}\bar{\mathcal{R}}+\Lambda_{4}g_{ij}+\frac{1}{4R^{2}}\left[g_{ik}g_{jl}\partial^{2}_{3}g^{kl}\right.$
$\displaystyle\left.-\partial^{2}_{3}g_{ij}+g^{kl}\partial_{3}g_{kl}\partial_{3}g_{ij}+2\partial_{3}\left(g_{ij}g^{kl}\partial_{3}g_{kl}\right)\right.$
$\displaystyle\left.-\frac{1}{2}g_{ij}\left\\{\partial_{3}g^{kl}\partial_{3}g_{kl}-\left(g^{kl}\partial_{3}g_{kl}\right)^{2}\right\\}\right]=0,$
(4)
where
$\bar{\mathcal{R}}_{ij}\equiv(\partial_{k}\bar{\Gamma}^{k}_{ji}-\partial_{j}\bar{\Gamma}^{k}_{ki}+\bar{\Gamma}^{l}_{ji}\bar{\Gamma}^{k}_{kl}-\bar{\Gamma}^{l}_{ki}\bar{\Gamma}^{k}_{jl})$
with
$\bar{\Gamma}^{k}_{ij}\equiv\frac{g^{kl}}{2}(\partial_{i}g_{lj}+\partial_{j}g_{ki}-\partial_{l}g_{ij})$
and $\bar{\mathcal{R}}\equiv g^{ij}\bar{\mathcal{R}}_{ij}$. One can solve Eq.
(4) by separating the variables as
$g_{ij}(x^{i},x^{3})=\chi(x^{3})g^{(3)}_{ij}(x^{i})$ where
$g^{(3)}_{ij}(x^{i})$ is identified as the metric of the 3D effective theory
and $\chi(x^{3})$ is its wavefunction profile along the compactified
dimension. Then, substituting this variable separation into Eq. (4), we obtain
the 3D Einstein field equations determining the 3D effective geometry of
spacetime and the equation for the wavefunction profile of the 3D metric as
follows
$\displaystyle\mathcal{R}^{(3)}_{ij}-\frac{1}{2}g^{(3)}_{ij}\mathcal{R}^{(3)}+\Lambda_{3}g^{(3)}_{ij}$
$\displaystyle=$ $\displaystyle 0,$ (5)
$\displaystyle\chi^{\prime\prime}+\frac{11}{4}\frac{\chi^{\prime
2}}{\chi}+\kappa^{2}\chi$ $\displaystyle=$
$\displaystyle\frac{\Lambda_{3}}{R^{-2}},$ (6)
where $\mathcal{R}^{(3)}_{ij}$ ($\mathcal{R}^{(3)}$) is the Ricci (scalar)
curvature of 3D effective spacetime, $\kappa\equiv\sqrt{\Lambda_{4}}/R^{-1}$,
and $\Lambda_{3}$ is a constant which characterizes the dynamics of the 3D
metric $g_{ij}$ along the compactified dimension and from Eq. (5) we see that
it plays the role of a 3D cosmological constant.
It is difficult to find an analytical solution for $\chi(x^{3})$ in the
general case. However, in the situation of $\Lambda_{3}/R^{-2}\ll 1$ (which
can be seen from Table. 1 in Ref. Nam2023 ), we can perturbatively solve Eq.
(6) in the order of $\Lambda_{3}/R^{-2}$. The solution of $\chi(\theta)$ at
the leading order is easily found as follows
$\displaystyle\chi(x^{3})=\left[\frac{1}{2}\left\\{1+\cos\left(\sqrt{15}\kappa
x^{3}\right)\right\\}\right]^{2/15}.$ (7)
Due to the topology of $S^{1}$, the wavefunction profile $\chi(x^{3})$ must be
periodic with the period of $2\pi$, i.e. $\chi(x^{3})=\chi(x^{3}+2\pi)$, which
implies $\sqrt{15}\kappa=n$ with $n=1,2,3,\ldots.$ This thus leads to the
quantization of the radius $R$ of the compactified dimension and 4D
cosmological constant $\Lambda_{4}$ as
$\displaystyle\Lambda_{4}R^{2}=\frac{n^{2}}{15}.$ (8)
Equation (8) means that the radius of the compactified dimension and the 4D
cosmological constant are not arbitrary but must obtain the discrete values
according to the quantization relation (8). On the other hand, the radius of
the compactified dimension and the 4D cosmological constant have no continuous
spectrum but a discrete one. This is a novel point of the circle
compactification solution of the SM coupled to GR which is pointed to the
first time.
The quantization relation (8) results from the nontrivial dynamics of the 3D
metric along the compact direction in the presence of the cosmological
constant. From Eq. (6) that describes the dynamic propagation of the 3D metric
along the compact direction, we see that the 4D cosmological constant
$\Lambda_{4}$ plays the role of a force proportional linearly to the profile
$\chi$ because its presence curves the spacetime geometry, whereas the inverse
radius $R^{-1}$ of the compactified dimension plays the role of a mass for the
3D metric because its propagation is confined along the compact direction.
(Note that, the term $11\chi^{\prime 2}/(4\chi)$ in Eq. (6) comes from the
nonlinear property of the gravitational field which means that gravity is
itself a source creating the curvature of the spacetime geometry.) As a
result, it leads to the oscillation behavior of the 3D metric along the
compactified dimension with the corresponding period depending on the ratio of
$\sqrt{\Lambda_{4}}$ to $R^{-1}$. In addition, the dynamic propagation of the
3D metric is constrained by the periodic property of the compactified
dimension. Hence, both of these aspects result in the quantization of the
radius $R$ of the compactified dimension and 4D cosmological constant
$\Lambda_{4}$.
In fact, the size of the compactified dimension can be physically determined
by the 3D effective potential of the radion field which is generated by the
dynamics of $3$D tensor component along the compactified dimension and the
one-loop quantum corrections with the Casimir energy density calculated in
Ref. Arkani-Hamed2007 (see Appendix D for detailed computation). Taking into
account the stabilization mechanism and from Eq. (8), we can find a general
quantization rule of the 4D cosmological constant in terms of the mass of the
light particles in the spectrum and the parameter $\Lambda_{3}$ as follows
$\displaystyle\Lambda_{4}=f(m_{i},n_{i},\Lambda_{3})n^{2}\propto n^{2},$ (9)
where $f(m_{i},n_{i},\Lambda_{3})$ is a function of $m_{i}$ (the mass of the
$i$th light particle), $n_{i}$ (the number of degrees of freedom corresponding
to the $i$th light particle), and $\Lambda_{3}$. Note that, the function
$f(m_{i},n_{i},\Lambda_{3})$ can be completely determined from the minimum of
the radion potential. The quantization rule (9) implies that the ratio of the
4D cosmological constant to a specific combination of the 3D cosmological
constant and the masses of the particles contributing to the radion potential
must lead to 1, 4, 9, 16, 25, $\ldots$. In general, under the quantum
corrections this ratio would not lead to 1, 4, 9, 16, 25, $\ldots$ which are
only the allowed values. Hence, we expect that the quantization rule (9) can
provide a mechanism to prevent the quantum fluctuations from generating
additional contributions to the 4D cosmological constant.
It is important to emphasize that Eq. (8) can be realized as a quantum gravity
sign that arises at the low-energy regime. Hence, one expects that the
quantization rule for the radius of the compactified dimension and the
cosmological constant could provide a step toward solving not only the
cosmological constant problem but also other important problems (which require
a complete theory of quantum gravity) such as a microscopic description of the
black hole entropy. By considering the compactification of $\text{dS}_{5}$ on
a circle $S^{1}$ to obtain 4D effective field theory, we can find a
quantization relation for the size of the extra dimension as
$R^{-1}\propto\sqrt{\Lambda_{5}}/n$ where $\Lambda_{5}$ is the bulk
cosmological constant and $n=1,2,3,\ldots$, in analogy to what we have done
for the $S^{1}$ compactification of the SM. Each value of the positive integer
$n$ corresponding to a possible value of the size of the extra dimension
implies a possible configuration of 4D effective field theory. Hence, we can
calculate the black hole entropy by summing over all relevant configurations.
We leave an exploration of this proposal to future work.
## Appendix A: The expansion of $\mathcal{R}^{(4)}$ in terms of 3D components
Because Eq. (3) is the main technical starting point, in this appendix we
present detailed computations of how to obtain it. It is easy to do the
calculations in the frame
$\\{\hat{\partial}_{i},\hat{\partial}_{3}\\}\equiv\\{\hat{\partial}_{\mu}\\}$
where $\hat{\partial}_{i}\equiv\partial_{i}-g_{{}_{A}}A_{i}\partial_{3}$ and
$\hat{\partial}_{3}\equiv\partial_{3}$ which transform as a 3D vector and a 1D
vector, respectively. The corresponding coframe is given by
$\\{dx^{i},dx^{3}+g_{{}_{A}}A_{i}dx^{i}\\}$ which dual to
$\\{\hat{\partial}_{i},\hat{\partial}_{3}\\}$, respectively. The components of
4D metric and its dual are given in these bases as
$\displaystyle\hat{g}_{\mu\nu}$ $\displaystyle=$
$\displaystyle\textrm{diag}\left(g_{ij},R^{2}\right),$
$\displaystyle\hat{g}^{\mu\nu}$ $\displaystyle=$
$\displaystyle\textrm{diag}\left(g^{ij},R^{-2}\right).$ (10)
Then, we can determine the Christoffel connection
$\hat{\Gamma}^{\rho}_{\mu\nu}$ and the Riemann curvature tensor
$\hat{\mathcal{R}}^{\lambda}_{\mu\rho\nu}$ in these bases as follows
$\displaystyle\hat{\Gamma}^{\rho}_{\mu\nu}$ $\displaystyle=$
$\displaystyle\frac{\hat{g}^{\rho\lambda}}{2}\left(\hat{\partial}_{\mu}\hat{g}_{\nu\lambda}+\hat{\partial}_{\nu}\hat{g}_{\mu\lambda}-\hat{\partial}_{\lambda}\hat{g}_{\mu\nu}\right)$
$\displaystyle+\frac{\hat{g}^{\rho\lambda}}{2}\left(C^{\sigma}_{\lambda\mu}\hat{g}_{\nu\sigma}+C^{\sigma}_{\lambda\nu}\hat{g}_{\mu\sigma}\right)+\frac{C^{\rho}_{\mu\nu}}{2},$
$\displaystyle\hat{\mathcal{R}}^{\lambda}_{\mu\rho\nu}$ $\displaystyle=$
$\displaystyle\hat{\partial}_{\rho}\hat{\Gamma}^{\lambda}_{\nu\mu}-\hat{\partial}_{\nu}\hat{\Gamma}^{\lambda}_{\rho\mu}+\hat{\Gamma}^{\sigma}_{\nu\mu}\hat{\Gamma}^{\lambda}_{\rho\sigma}-\hat{\Gamma}^{\sigma}_{\rho\mu}\hat{\Gamma}^{\lambda}_{\nu\sigma}$
(11) $\displaystyle-C^{\sigma}_{\rho\nu}\hat{\Gamma}^{\lambda}_{\sigma\mu},$
where $C^{\rho}_{\mu\nu}$ defines the commutation of two frame fields as
$\left[\hat{\partial}_{\mu},\hat{\partial}_{\nu}\right]=C^{\rho}_{\mu\nu}\hat{\partial}_{\rho}.$
(12)
The 4D scalar curvature $\mathcal{R}^{(4)}$ reads
$\displaystyle\mathcal{R}^{(4)}=\hat{g}^{\mu\nu}\hat{\mathcal{R}}^{\rho}_{\mu\rho\nu}=g^{ij}\hat{\mathcal{R}}^{\rho}_{i\rho
j}+g^{33}\hat{\mathcal{R}}^{\rho}_{3\rho 3}.$ (13)
The first and second terms are explicitly expanded as follows
$\displaystyle g^{ij}\hat{\mathcal{R}}^{\rho}_{i\rho j}$ $\displaystyle=$
$\displaystyle
g^{ij}\left(\hat{\partial}_{\rho}\hat{\Gamma}^{\rho}_{ji}-\hat{\partial}_{j}\hat{\Gamma}^{\rho}_{\rho
i}+\hat{\Gamma}^{\rho}_{ji}\hat{\Gamma}^{\lambda}_{\lambda\rho}-\hat{\Gamma}^{\rho}_{\lambda
i}\hat{\Gamma}^{\lambda}_{j\rho}\right.$ (14)
$\displaystyle\left.-C^{\rho}_{\lambda j}\hat{\Gamma}^{\lambda}_{\rho
i}\right)$ $\displaystyle=$ $\displaystyle
g^{ij}\left[\left(\hat{\partial}_{k}\hat{\Gamma}^{k}_{ji}-\hat{\partial}_{j}\hat{\Gamma}^{k}_{ki}+\hat{\Gamma}^{k}_{ji}\hat{\Gamma}^{l}_{lk}-\hat{\Gamma}^{k}_{li}\hat{\Gamma}^{l}_{jk}\right)\right.$
$\displaystyle+\left.\left(\hat{\partial}_{3}\hat{\Gamma}^{3}_{ji}+\hat{\Gamma}^{3}_{ji}\hat{\Gamma}^{k}_{k3}+\hat{\Gamma}^{k}_{ji}\hat{\Gamma}^{3}_{3k}+\hat{\Gamma}^{3}_{ji}\hat{\Gamma}^{3}_{33}\right)\right.$
$\displaystyle-\left.\left(\hat{\partial}_{j}\hat{\Gamma}^{3}_{3i}+\hat{\Gamma}^{k}_{3i}\hat{\Gamma}^{3}_{jk}+\hat{\Gamma}^{3}_{ki}\hat{\Gamma}^{k}_{j3}+\hat{\Gamma}^{3}_{3i}\hat{\Gamma}^{3}_{j3}\right)\right.$
$\displaystyle-\left.\left(C^{3}_{kj}\hat{\Gamma}^{k}_{3i}+C^{3}_{3j}\hat{\Gamma}^{3}_{3i}\right)\right],$
$\displaystyle g^{33}\hat{\mathcal{R}}^{\rho}_{3\rho 3}$ $\displaystyle=$
$\displaystyle
g^{33}\left(\hat{\partial}_{\rho}\hat{\Gamma}^{\rho}_{33}-\hat{\partial}_{3}\hat{\Gamma}^{\rho}_{\rho
3}+\hat{\Gamma}^{\rho}_{33}\hat{\Gamma}^{\lambda}_{\lambda\rho}-\hat{\Gamma}^{\rho}_{\lambda
3}\hat{\Gamma}^{\lambda}_{3\rho}\right.$ (15)
$\displaystyle\left.-C^{\rho}_{\lambda 3}\hat{\Gamma}^{\lambda}_{\rho
3}\right)$ $\displaystyle=$ $\displaystyle
g^{33}\left(\hat{\partial}_{i}\hat{\Gamma}^{i}_{33}-\hat{\partial}_{3}\hat{\Gamma}^{i}_{i3}+\hat{\Gamma}^{i}_{33}\hat{\Gamma}^{j}_{ji}+\hat{\Gamma}^{3}_{33}\hat{\Gamma}^{i}_{i3}\right.$
$\displaystyle\left.-\hat{\Gamma}^{j}_{i3}\hat{\Gamma}^{i}_{3j}-\hat{\Gamma}^{3}_{i3}\hat{\Gamma}^{i}_{33}-C^{3}_{i3}\hat{\Gamma}^{i}_{33}\right).$
In (14) and (15), we find the following combinations
$\displaystyle
g^{ij}\left(\hat{\partial}_{3}\hat{\Gamma}^{3}_{ji}+\hat{\Gamma}^{3}_{ji}\hat{\Gamma}^{k}_{k3}+\hat{\Gamma}^{3}_{ji}\hat{\Gamma}^{3}_{33}\right)\equiv\nabla_{\mu}Y^{\mu}_{1}$
$\displaystyle+\frac{g^{33}}{2}\hat{\partial}_{3}g^{ij}\hat{\partial}_{3}g_{ij},$
(16) $\displaystyle
g^{ij}\left(\hat{\partial}_{j}\hat{\Gamma}^{3}_{3i}-\hat{\Gamma}^{k}_{ji}\hat{\Gamma}^{3}_{3k}\right)-g^{33}\hat{\Gamma}^{i}_{33}\hat{\Gamma}^{3}_{3i}$
$\displaystyle\equiv$ $\displaystyle\nabla_{\mu}Y^{\mu}_{2},$ (17)
$\displaystyle
g^{33}\left(\hat{\partial}_{i}\hat{\Gamma}^{i}_{33}+\hat{\Gamma}^{i}_{33}\hat{\Gamma}^{j}_{ji}+\hat{\Gamma}^{i}_{33}\hat{\Gamma}^{3}_{3i}\right)\equiv\nabla_{\mu}Y^{\mu}_{3}$
$\displaystyle+\frac{g^{ij}}{2}\hat{\partial}_{i}g^{33}\hat{\partial}_{j}g_{33}+g_{{}_{A}}g^{ij}\hat{\partial}_{3}A_{i}g^{33}\hat{\partial}_{j}g_{33},$
(18) $\displaystyle
g^{33}\left(\hat{\partial}_{3}\hat{\Gamma}^{i}_{i3}-\hat{\Gamma}^{3}_{33}\hat{\Gamma}^{i}_{i3}\right)-g^{ij}\hat{\Gamma}^{3}_{ij}\hat{\Gamma}^{k}_{k3}$
$\displaystyle\equiv$ $\displaystyle\nabla_{\mu}Y^{\mu}_{4},$ (19)
where
$\displaystyle Y^{\mu}_{1}$ $\displaystyle\equiv$
$\displaystyle\left(0,g^{ij}\hat{\Gamma}^{3}_{ji}\right),$ $\displaystyle
Y^{\mu}_{2}$ $\displaystyle\equiv$
$\displaystyle\left(g^{ij}\hat{\Gamma}^{3}_{3j},0\right),$ $\displaystyle
Y^{\mu}_{2}$ $\displaystyle\equiv$
$\displaystyle\left(g^{33}\hat{\Gamma}^{i}_{33},0\right),$ $\displaystyle
Y^{\mu}_{4}$ $\displaystyle\equiv$
$\displaystyle\left(0,g^{33}\hat{\Gamma}^{i}_{i3}\right).$ (20)
The terms $\nabla_{\mu}Y^{\mu}_{a}$ with $a=1,2,3,4$ are divergences
corresponding to the boundary terms and as a result, they vanish at infinity.
Then, $\mathcal{R}^{(4)}$ is a sum of the remaining terms as
$\displaystyle\mathcal{R}^{(4)}$ $\displaystyle=$
$\displaystyle\mathcal{R}-g^{ij}\left(\hat{\Gamma}^{k}_{3i}\hat{\Gamma}^{3}_{jk}+\hat{\Gamma}^{3}_{ki}\hat{\Gamma}^{k}_{j3}+\hat{\Gamma}^{3}_{3i}\hat{\Gamma}^{3}_{j3}+\hat{\Gamma}^{3}_{ij}\hat{\Gamma}^{k}_{k3}\right.$
(21)
$\displaystyle\left.+C^{3}_{kj}\hat{\Gamma}^{k}_{3i}+C^{3}_{3j}\hat{\Gamma}^{3}_{3i}\right)-g^{33}\left(\hat{\Gamma}^{j}_{i3}\hat{\Gamma}^{i}_{3j}+\hat{\Gamma}^{3}_{i3}\hat{\Gamma}^{i}_{33}\right.$
$\displaystyle\left.+2\hat{\Gamma}^{3}_{3i}\hat{\Gamma}^{i}_{33}+C^{3}_{i3}\hat{\Gamma}^{i}_{33}\right)+\frac{g^{33}}{2}\hat{\partial}_{3}g^{ij}\hat{\partial}_{3}g_{ij}$
$\displaystyle+\frac{g^{ij}}{2}\hat{\partial}_{i}g^{33}\hat{\partial}_{j}g_{33}+g_{{}_{A}}g^{ij}\hat{\partial}_{3}A_{i}g^{33}\hat{\partial}_{j}g_{33}$
$\displaystyle=$
$\displaystyle\mathcal{R}+\frac{1}{4R^{2}}\left(\partial_{3}g^{ij}\partial_{3}g_{ij}+g^{ij}g^{kl}\partial_{3}g_{ij}\partial_{3}g_{kl}\right)$
$\displaystyle-\frac{g^{2}_{{}_{A}}R^{2}}{4}F_{ij}F^{ij},$
where $\mathcal{R}\equiv
g^{ij}\left(\hat{\partial}_{k}\hat{\Gamma}^{k}_{ji}-\hat{\partial}_{j}\hat{\Gamma}^{k}_{ki}+\hat{\Gamma}^{k}_{ji}\hat{\Gamma}^{l}_{lk}-\hat{\Gamma}^{k}_{li}\hat{\Gamma}^{l}_{jk}\right)\equiv
g^{ij}R_{ij}$ and $F_{ij}=\partial_{i}A_{j}-\partial_{j}A_{i}$. Note that, in
the calculation of the scalar curvature $\mathcal{R}^{(4)}$, we have
considered the components $g_{ij}$, $A_{i}$, and $R$ to be the general
functions of $(x^{i},x^{3})$. However, we observe that first the terms
relating to $\hat{\partial}_{3}A_{i}$ automatically cancel together and hence
they do not contribute to $\mathcal{R}^{(4)}$. This is because the spacetime
curvature which is caused by the non-triviality of the $U(1)$ principal bundle
is measured by the curvature of the connection $A_{i}$, i.e. $F_{ij}$. Second,
the first-order derivative terms of $R$ associated with its kinetic terms do
not appear in $\mathcal{R}^{(4)}$ because the curvature of the $S^{1}$ fiber
is zero. But, the kinetic terms of $R$ would appear in Einstein frame derived
by rescaling the 3D metric as $g_{ij}\rightarrow\Omega^{-2}g_{ij}$ with
$\Omega=R/r$ where $r$ is introduced to keep the rescaled 3D metric
dimensionless and would be fixed equal to the vacuum expectation value of the
radion field.
## Appendix B: Equations of motion for the 3D components of the bulk metric
First, let us obtain equations of motion for the 3D tensor component $g_{ij}$
from the following variation
$\displaystyle\delta_{g_{3}}S_{\text{EH}}$ $\displaystyle=$
$\displaystyle\delta_{g_{3}}\int
d^{4}x\sqrt{-g}\frac{M^{2}_{\text{Pl}}}{2}\left(\mathcal{R}^{(4)}-2\Lambda_{4}\right)$
(22) $\displaystyle=$ $\displaystyle\frac{M^{2}_{\text{Pl}}}{2}\int
d^{4}x\left[\left(\mathcal{R}^{(4)}-2\Lambda_{4}\right)\delta_{g_{3}}\sqrt{-g}\right.$
$\displaystyle\left.+\sqrt{-g}\delta_{g_{3}}\mathcal{R}^{(4)}\right]=0,$
where $\delta_{g_{3}}$ refers to the variation in terms of $g_{ij}$ and the
expansion of $\mathcal{R}^{(4)}$ is given in Eq. (21). We find
$\delta_{g_{3}}\sqrt{-g}=-\frac{1}{2}\sqrt{-g}g_{ij}\delta g^{ij}$ and
$\displaystyle\delta_{g_{3}}\mathcal{R}^{(4)}$ $\displaystyle=$
$\displaystyle\delta g^{ij}R_{ij}+g^{ij}\delta_{g_{3}}R_{ij}$ (23)
$\displaystyle+\frac{1}{4}\delta_{g_{3}}\left(\partial_{3}g^{ij}\partial^{3}g_{ij}+g^{ij}g^{kl}\partial_{3}g_{ij}\partial^{3}g_{kl}\right)$
$\displaystyle-\frac{g^{2}_{{}_{A}}R^{2}}{4}\delta_{g_{3}}\left(F_{ij}F^{ij}\right).$
The second variation term in (23) is calculated as
$\displaystyle g^{ij}\delta_{g_{3}}R_{ij}$ $\displaystyle=$
$\displaystyle\nabla_{k}\left(g_{ij}\nabla^{k}\delta g^{ij}-\nabla_{l}\delta
g^{kl}\right)$ (24) $\displaystyle=$
$\displaystyle\nabla_{\mu}X^{\mu}_{1}-\hat{\Gamma}^{3}_{3k}\left(g_{ij}\nabla^{k}\delta
g^{ij}-\nabla_{l}\delta g^{kl}\right)$ $\displaystyle=$
$\displaystyle\nabla_{\mu}X^{\mu}_{1}+\nabla_{\mu}X^{\mu}_{2}+\nabla_{\mu}X^{\mu}_{3}+\hat{\Gamma}^{3}_{3k}\hat{\Gamma}^{3}_{3l}g^{kl}g_{ij}\delta
g^{ij}$
$\displaystyle+\nabla_{k}\left(g^{kl}\hat{\Gamma}^{3}_{3l}g_{ij}\right)\delta
g^{ij}-\hat{\Gamma}^{3}_{3i}\hat{\Gamma}^{3}_{3j}\delta g^{ij}$
$\displaystyle-\nabla_{i}\left(\hat{\Gamma}^{3}_{3j}\right)\delta g^{ij},$
where
$\displaystyle X^{\mu}_{1}$ $\displaystyle=$
$\displaystyle\left(g_{ij}\nabla^{k}\delta g^{ij}-\nabla_{l}\delta
g^{kl},0\right),$ $\displaystyle X^{\mu}_{2}$ $\displaystyle=$
$\displaystyle\left(-g^{kl}\hat{\Gamma}^{3}_{3l}g_{ij}\delta g^{ij},0\right),$
(25) $\displaystyle X^{\mu}_{3}$ $\displaystyle=$
$\displaystyle\left(\hat{\Gamma}^{3}_{3l}\delta g^{lk},0\right).$
The variation of the third term in (23) reads
$\displaystyle\delta_{g_{3}}\left(\partial_{3}g^{ij}\partial^{3}g_{ij}+g^{ij}g^{kl}\partial_{3}g_{ij}\partial^{3}g_{kl}\right)=\nabla_{\mu}X^{\mu}_{4}+\nabla_{\mu}X^{\mu}_{5}$
$\displaystyle+g_{ik}g_{jl}\partial_{3}\partial^{3}g^{kl}-\partial_{3}\partial^{3}g_{ij}+g^{kl}\partial_{3}g_{kl}\partial^{3}g_{ij}\delta
g^{ij}$
$\displaystyle+2\partial_{3}\left(g_{ij}g^{kl}\partial^{3}g_{kl}\right)+\left(g^{kl}\partial_{3}g_{kl}\right)^{2}g_{ij}\delta
g^{ij}$
$\displaystyle-2\hat{\Gamma}^{3}_{33}\left(\partial^{3}g_{ij}+g_{ij}g_{kl}\partial^{3}g^{kl}\right)\delta
g^{ij},$ (26)
where
$\displaystyle X^{\mu}_{4}$ $\displaystyle=$
$\displaystyle\left(0,2\partial^{3}g_{ij}\delta g^{ij}\right),$ $\displaystyle
X^{\mu}_{5}$ $\displaystyle=$
$\displaystyle\left(0,2g_{ij}g_{kl}\partial^{3}g^{kl}\delta g^{ij}\right).$
(27)
The last variation term is
$\displaystyle\delta_{g_{3}}\left(F_{ij}F^{ij}\right)=2{F_{i}}^{k}F_{jk}.$
(28)
Finally, we find the equations of motion for $g_{ij}$ as follows
$\displaystyle\mathcal{R}_{ij}-\frac{1}{2}g_{ij}\mathcal{R}+\Lambda_{4}g_{ij}+\frac{1}{4R^{2}}\left[g_{ik}g_{jl}\partial^{2}_{3}g^{kl}\right.$
$\displaystyle\left.-\partial^{2}_{3}g_{ij}+g^{kl}\partial_{3}g_{kl}\partial_{3}g_{ij}+2\partial_{3}\left(g_{ij}g^{kl}\partial_{3}g_{kl}\right)\right.$
$\displaystyle\left.-\frac{1}{2}g_{ij}\left\\{\partial_{3}g^{kl}\partial_{3}g_{kl}-\left(g^{kl}\partial_{3}g_{kl}\right)^{2}\right\\}\right]$
$\displaystyle-\frac{g^{2}_{{}_{A}}R^{2}}{2}\left({F_{i}}^{k}F_{jk}-\frac{1}{4}F_{kl}F^{kl}g_{ij}\right)+f_{1}\left(\hat{\Gamma}^{3}_{3k},\hat{\Gamma}^{3}_{33}\right)$
$\displaystyle=$ $\displaystyle 0,$
where $f_{1}\left(\hat{\Gamma}^{k}_{k3},\hat{\Gamma}^{3}_{33}\right)$ refers
to the terms which are proportional to $\hat{\Gamma}^{3}_{3k}$,
$(\hat{\Gamma}^{3}_{3k})^{2}$, $\nabla_{k}(\hat{\Gamma}^{3}_{3l})$, and
$\hat{\Gamma}^{3}_{33}$ which are zero with $\hat{\partial}_{i}R=0$,
$\hat{\partial}_{3}R=0$, and $\partial_{3}A_{i}=0$.
Equations of motion for the 3D vector component $A_{i}$ are derived from the
following variation
$\displaystyle\delta_{A}S_{\text{EH}}$ $\displaystyle=$
$\displaystyle\frac{M^{2}_{\text{Pl}}}{2}\int
d^{4}x\sqrt{-g}\left[\delta_{A}\mathcal{R}-\frac{g^{2}_{{}_{A}}R^{2}}{4}\delta_{A}\left(F_{ij}F^{ij}\right)\right]$
(30) $\displaystyle=$ $\displaystyle 0$
This variation leads to the equations of motion for $A_{i}$ as
$\nabla_{i}F^{ij}+R^{-2}f_{2}\left(A_{j},\hat{\partial}_{j}g_{kl},\hat{\partial}_{3}g_{kl}\right)=0.$
(31)
Note that, the second term in Eq. (31) is negligible because it is strongly
suppressed by $R^{-2}$ which is in the order of $\Lambda_{4}$ ($\approx
5.06\times 10^{-84}$ GeV2 Tanabashi2018 ) as indicated in Appendix D. Thus,
Eq. (31) can be approximated as $\nabla_{i}F^{ij}=0$.
Because the radion field $R$ would get an effective potential (generated from
the cosmological constant term and one-loop quantum corrections) by which the
size of the compactified dimension is physically fixed, we write the equation
of motion for the radion field $R$ including this radion potential,
corresponding to the variation of the 3D action (given in Appendix D) as
follows
$\displaystyle\square R$ $\displaystyle=$
$\displaystyle\frac{1}{R}(\partial_{i}R)^{2}+\frac{R^{2}}{2M_{3}}\frac{\partial
V}{\partial R}+f_{3}(A_{i},F_{ij}),$ (32)
where $f_{3}(A_{i},F_{ij})$ is a functional of $A_{i}$ and $F_{ij}$ which
vanishes for $A_{i}=0$.
## Appendix C: A derivation of Eqs. (5) and (6)
In order to solve Eq. (4), we use the method of the variable separation as
$g_{ij}(x^{i},x^{3})=\chi(x^{3})g^{(3)}_{ij}(x^{i})$ where
$g^{(3)}_{ij}(x^{i})$ is the metric of the 3D effective theory whose
wavefunction profile along the compactified dimension is $\chi(x^{3})$. First,
by substituting $g_{ij}(x^{i},x^{3})=\chi(x^{3})g^{(3)}_{ij}(x^{i})$ into Eq.
(4), we find
$\displaystyle\mathcal{R}^{(3)}_{ij}-\frac{1}{2}g^{(3)}_{ij}\mathcal{R}^{(3)}+\Lambda_{4}\chi
g^{(3)}_{ij}+\frac{1}{4R^{2}}\left[\left(\frac{1}{\chi}\right)^{{}^{\prime\prime}}\chi^{2}-\chi{{}^{\prime\prime}}\right.$
$\displaystyle\left.+\frac{3\chi^{\prime
2}}{\chi}+6\chi^{\prime\prime}-\frac{3\chi}{2}\left\\{\left(\frac{1}{\chi}\right)^{{}^{\prime}}\chi^{\prime}-3\left(\frac{\chi^{\prime}}{\chi}\right)^{2}\right\\}\right]g^{(3)}_{ij},$
(33)
where
$\mathcal{R}^{(3)}_{ij}\equiv(\partial_{k}\Gamma^{k}_{ji}-\partial_{j}\Gamma^{k}_{ki}+\Gamma^{l}_{ji}\Gamma^{k}_{kl}-\Gamma^{l}_{ki}\Gamma^{k}_{jl})$
with $\Gamma^{k}_{ij}\equiv
g^{(3)kl}(\partial_{i}g^{(3)}_{lj}+\partial_{j}g^{(3)}_{ki}-\partial_{l}g^{(3)}_{ij})/2$
and $\mathcal{R}^{(3)}\equiv g^{(3)ij}\mathcal{R}^{(3)}_{ij}$, and ′ denotes
the derivative with respect to the coordinate $x^{3}$. After simplification,
we obtain
$\displaystyle\mathcal{R}^{(3)}_{ij}-\frac{1}{2}g^{(3)}_{ij}\mathcal{R}^{(3)}+\left[\frac{1}{R^{2}}\left(\chi^{\prime\prime}+\frac{11}{4}\frac{\chi^{\prime
2}}{\chi}\right)+\Lambda_{4}\chi\right]g^{(3)}_{ij}=0.$
This equation can be written in the following way
$\displaystyle\frac{\mathcal{R}^{(3)}}{6}=\frac{1}{R^{2}}\left(\chi^{\prime\prime}+\frac{11}{4}\frac{\chi^{\prime
2}}{\chi}\right)+\Lambda_{4}\chi.$ (35)
It is clear that the left-handed and right-handed sides of Eq. (35) depend on
the variables $x^{i}$ and $x^{3}$, respectively. This implies that these sides
are equal to a constant denoted by $\Lambda_{3}$, which leads to
$\displaystyle\mathcal{R}^{(3)}=6\Lambda_{3},$ (36)
$\displaystyle\chi^{\prime\prime}+\frac{11}{4}\frac{\chi^{\prime
2}}{\chi}+\kappa^{2}\chi$ $\displaystyle=$
$\displaystyle\frac{\Lambda_{3}}{R^{-2}},$ (37)
$\kappa\equiv\sqrt{\Lambda_{4}}/R^{-1}$. Note that, Eq. (36) is equivalent to
Eq. (5) in the paper.
## Appendix D: The radion potential
The dimensional reduction of the 4D Einstein gravity on $S^{1}$ given in the
Einstein frame reads
$\displaystyle S_{3D}$ $\displaystyle\supset$ $\displaystyle\int
d^{3}x\sqrt{-g_{3}}\left[\frac{M_{3}}{2}\left\\{\mathcal{R}^{(3)}-2\left(\frac{\partial_{i}R}{R}\right)^{2}\right\\}\right.$
(38)
$\displaystyle\left.-M_{3}\left(\frac{r}{R}\right)^{2}\Lambda_{3}\right],$
where $g_{3}\equiv\text{det}\left[g^{(3)}_{ij}\right]$ and the 3D Planck
energy scale is identified as
$\displaystyle M_{3}$ $\displaystyle\equiv$ $\displaystyle
rM^{2}_{\text{Pl}}\int^{\pi}_{-\pi}dx^{3}\chi^{1/2}$ (39) $\displaystyle=$
$\displaystyle\frac{2\sqrt{\pi}\Gamma(17/30)}{\Gamma(16/15)}rM^{2}_{\text{Pl}}.$
The last term in the action (38) is the tree-level potential of the radion
field generated by the dynamics of $3$D tensor component along the
compactified dimension. In addition, the one-loop quantum corrections would
contribute to the radion potential as
$\displaystyle V_{\text{1L}}(R)$ $\displaystyle=$
$\displaystyle\sum_{i}(-1)^{s_{i}}n_{i}R\left(\frac{r}{R}\right)^{3}\rho_{i}(R)\int^{\pi}_{-\pi}dx^{3}\chi^{\frac{3}{2}}.\
\ $ (40)
Here $s_{i}$ is equal to $0(1)$ for the fermions(bosons), $n_{i}$ is the
number of degrees of freedom corresponding to the $i$th particle, and the
Casimir energy density with respect to the $i$th particle is given by Arkani-
Hamed2007
$\displaystyle\rho_{i}(R)=\sum^{\infty}_{n=1}\frac{2m^{4}_{i}}{(2\pi)^{2}}\frac{K_{2}(2\pi
nm_{i}R)}{(2\pi nm_{i}R)^{2}},$ (41)
where $m_{i}$ and $K_{2}(z)$ are the mass of the $i$th particle and the
modified Bessel function, respectively. It should be noted here that due to
the function $K_{2}(z)$ suppressed for $z\ll 1$ the particles with their mass
which is much larger than $R^{-1}$ do not contribute significantly to the one-
loop term of the radion potential and hence we can ignore their contribution.
On the other hand, only the light degrees of freedom contribute significantly
to $V_{\text{1L}}(R)$.
The radion potential $V(R)$ thus is a sum of the tree and loop level
contributions, which is expanded in terms of $m_{i}R$ for $m_{i}R\ll 1$ as
$\displaystyle\frac{V(R)}{2\sqrt{\pi}r^{3}}$ $\displaystyle\simeq$
$\displaystyle\left[\frac{\Gamma(17/30)}{\Gamma(16/15)}\frac{M^{2}_{P}\Lambda_{3}}{R^{2}}+\frac{1}{16\pi^{2}}\frac{\Gamma(7/10)}{\Gamma(6/5)}\frac{1}{R^{6}}\right.$
$\displaystyle\left.\times\sum_{i}(-1)^{s_{i}}n_{i}\left\\{\frac{1}{90}-\frac{(m_{i}R)^{2}}{6}+\frac{(m_{i}R)^{4}}{48}\right\\}\right].$
The behavior of the radion potential is depicted in Fig. 1. Here, the bosonic
contributions come from graviton and photon, whereas the fermionic
contributions come from the lightest neutrino and an additional massless Dirac
fermion.
---
Figure 1: The scaled radion potential $\bar{V}(R)\equiv
V(R)/(2\sqrt{\pi}r^{3}\Lambda^{3}_{4})$ versus the scaled $S^{1}$ radius
$R\sqrt{\Lambda_{4}}$ for various values of $m_{\nu}/\sqrt{\Lambda_{4}}$ with
$M^{2}_{\text{Pl}}\Lambda_{3}/\Lambda^{2}_{4}=0.1$. The red and orange curves
correspond to the AdS vacuum and no vacuum, respectively, whereas the blue and
purple curves lead to the dS vacuum.
## References
* (1) N. Aghanim _et al._ (Planck Collaboration), Astron. Astrophys 641, A6 (2020).
* (2) M. Tanabashi _et al._ (Particle Data Group), Phys. Rev. D 98, 030001 (2018).
* (3) Y. B. Zeldovich, Sov. Phys. Usp. 11, 381 (1968) [JETP Lett. 6, 316 (1967)].
* (4) N. D. Birrell and P. C. W. Davies, _Quantum Fields in Curved Space_ (Cambridge University Press, Cambridge, England, 1994).
* (5) S. Weinberg, Rev. Mod. Phys. 61, 1 (1989).
* (6) R. Bousso, Gen. Relativ. Gravit. 40, 607 (2008).
* (7) S. Kachru and E. Silverstein, Phys. Rev. Lett. 80, 4855 (1998).
* (8) S. Kachru, J. Kumar, and E. Silverstein, Phys. Rev. D 59, 106004 (1999).
* (9) J. A. Harvey, Phys. Rev. D 59, 026002 (1998).
* (10) R. Iengo and C. Zhu, JHEP 04 (2000) 028.
* (11) R. Blumenhagen and L. Gorlich, Nucl. Phys. B 551, 601 (1999).
* (12) C. Angelantonj, I. Antoniadis, and K. Forger, Nucl. Phys. B 555, 116 (1999).
* (13) C. Angelantonj and I. Antoniadis, Nucl. Phys. B 676, 129 (2004).
* (14) S. Nobbenhuis, arXiv:gr-qc/0609011.
* (15) C. P. Burgess, _Post-Planck Cosmology: Lecture Notes of the Les Houches Summer School_ (Oxford Academic, New York, 2015).
* (16) C. Vafa, arXiv: hep-th/0509212.
* (17) H. Ooguri and C. Vafa, Nucl. Phys. B 766, 21 (2007).
* (18) E. Palti, Fortschr. Phys. 67, 1900037 (2019).
* (19) G. Obied, H. Ooguri, L. Spodyneiko, and C. Vafa, arXiv:1806.08362.
* (20) S. K. Garg and C. Krishnan, JHEP 11 (2019) 075.
* (21) A. Bedroya, R. Brandenberger, M. Loverde, and C. Vafa, Phys. Rev. D 101, 103502 (2020).
* (22) R. Brandenberger, LHEP 2021 (2021) 198.
* (23) N. Arkani-Hamed, S. Dubovsky, A. Nicolis, and G. Villadoro, JHEP 0706 (2007) 078.
* (24) J. M. Arnold, B. Fornal, and M. B. Wise, JHEP 12 (2010) 083.
* (25) B. Fornal and M. B. Wise, JHEP 07 (2011) 086.
* (26) L. E. Ibáñez, V. Martin-Lozano, and I. Valenzuela, JHEP 11 (2017) 066.
* (27) L. E. Ibáñez, V. Martin-Lozano, and I. Valenzuela, arXiv:1707.05811.
* (28) Y. Hamada and G. Shiu, JHEP 11 (2017) 043.
* (29) E. Gonzalo, A. Herráez, and L. E. Ibáñez, JHEP 06 (2018) 051.
* (30) E. Gonzalo and L. E. Ibáñez, Phys. Lett. B 786, 272 (2018).
* (31) H. Ooguri and C. Vafa, Adv. Theor. Math. Phys. 21 (2017) 1787.
* (32) N. Arkani-Hamed, L. Motl, A. Nicolis, and C. Vafa, JHEP 06 060 (2007)
* (33) D. Lüst, E. Palti, and C. Vafa, Phys. Lett. B 797, 134867 (2019).
* (34) T. Rudelius, JHEP 08 (2021) 041.
* (35) E. Gonzalo, L. E. Ibáñez, and I. Valenzuela, Phys. Lett. B 822, 136691 (2021).
* (36) C. H. Nam, Phys. Rev. D 107, L041901 (2023).
* (37) C. H. Nam, _A prediction for the mass of the lightest neutrino from the Swampland and quantized compactification_ , in preparation.
|
# Constraints on Spin-Spin-Velocity-Dependent Interaction
Wei Ji Helmholtz-Institut, GSI Helmholtzzentrum für Schwerionenforschung,
Mainz 55128, Germany State Key Laboratory of Low-Dimensional Quantum Physics,
Department of Physics, Tsinghua University, Beijing 100084, China Johannes
Gutenberg-Universität Mainz, Mainz 55128, Germany Weipeng Li State Key
Laboratory of Low-Dimensional Quantum Physics, Department of Physics, Tsinghua
University, Beijing 100084, China Pavel Fadeev Helmholtz-Institut, GSI
Helmholtzzentrum für Schwerionenforschung, Mainz 55128, Germany Johannes
Gutenberg-Universität Mainz, Mainz 55128, Germany Filip Ficek Institute of
Theoretical Physics, Jagiellonian University, Łojasiewicza 11, 30-348 Kraków,
Poland Jianan Qin Helmholtz-Institut, GSI Helmholtzzentrum für
Schwerionenforschung, Mainz 55128, Germany Key Laboratory of Geophysical
Exploration Equipment, Ministry of Education of China, Jilin University,
Changchun 130012, China Kai Wei<EMAIL_ADDRESS>School of Instrumentation
Science and Opto-electronics Engineering, Beihang University, Beijing, 100191,
China Hangzhou Innovation Institute, Beihang University, Hangzhou, 310051,
China Hangzhou Extremely Weak Magnetic Field Major Science and Technology
Infrastructure Research Institute, Hangzhou, 310051, China Yong-Chun Liu
<EMAIL_ADDRESS>State Key Laboratory of Low-Dimensional Quantum Physics,
Department of Physics, Tsinghua University, Beijing 100084, China Frontier
Science Center for Quantum Information, Beijing, China Dmitry Budker
Helmholtz-Institut, GSI Helmholtzzentrum für Schwerionenforschung, Mainz
55128, Germany Johannes Gutenberg-Universität Mainz, Mainz 55128, Germany
Department of Physics, University of California, Berkeley, CA 94720-7300, USA
###### Abstract
The existence of exotic spin-dependent forces may shine light on new physics
beyond the Standard Model. We utilize two iron shielded SmCo5 electron-spin
sources and two optically pumped magnetometers to search for exotic long-range
spin-spin-velocity-dependent force. The orientations of spin sources and
magnetometers are optimized such that the exotic force is enhanced and common-
mode noise is effectively subtracted. We set direct limit on proton-electron
interaction in the force range from 1 cm to 1 km. Our experiment represents
more than ten orders of magnitude improvement than previous works.
The nature of dark matter is one of the most profound mysteries in modern
physics. Many new light bosons introduced by theories beyond the Standard
Model are proposed to be dark matter candidates, such as spin-0 bosons
including axions and axion like particles (ALPs) axion-PhysRevLett.38.1440 ;
preskill1983cosmology ; graham2015experimental , spin-1 bosons including dark
photons darkPhoton.2010 ; Dark-Photon-AN2015331 , and Z′ bosons
(okada2020dark, ; croon2021supernova, ). Furthermore, the new bosons may
mediate new types of long-range fundamental forces (moody1984new, ;
dobrescu2006spin, ; Fadeev2019, ; fayet1986fifth, ).
If we consider the spin, relative position and velocity of two fermions, the
exotic interaction between them can be classified to 16 terms dobrescu2006spin
; Fadeev2019 , and then generally classified into static terms and velocity-
dependent terms. A conventional velocity-dependent force in classical physics
is the Lorentz force of a moving charged particle.
Many experimental methods have been used to search for exotic forces,
including experiments with torsional resonators heckel2008preferred ;
hammond2007new ; Loong-Nature2003 ; terrano2015short , nuclear magnetic
resonance petukhov2010polarized ; yan2015searching ; chu2013laboratory ;
arvanitaki2014resonantly , magnetometers based on hot atoms and nitrogen-
vacancy center in diamond Romalis-PRL2009 ; Ji2018 ; jiao2021experimental ;
hunter2014using ; su2021search ; almasi2020new ; kim2018experimental , and
other high-sensitivity technologies tullney2013constraints ;
serebrov2010search ; ficek2017constraints ; ren2021search ;
stadnik2018improved . Most of these efforts focus on static interactions,
while the velocity-dependent interactions have also been gaining attention in
recent years jiao2021experimental ; su2021search ; hunter2014using ;
wei2022new ; kim2018experimental .
In this experiment, we focus on one term of Spin-Spin-Velocity-Dependent
Interaction (SSVDI) proposed by dobrescu2006spin :
$\displaystyle V=$ $\displaystyle\frac{f\hbar}{4\pi
c}\left[(\hat{\bm{\sigma}}_{1}\cdot\bm{v})(\hat{\bm{\sigma}}_{2}\cdot\bm{v})\right]\
\frac{e^{-r/\lambda}}{r},$ (1)
where $f$ is a dimensionless coupling coefficient, $\hat{\bm{\sigma}}_{1}$,
$\hat{\bm{\sigma}}_{2}$ are the respective Pauli spin-matrix vectors of the
two fermions, $r$ and $\mathbf{v}$ are the relative position and velocity
between two fermions. (This potential is called $V_{8}$ in Ref.
dobrescu2006spin .) To search for this force, a spin polarized test object is
required as the spin source, and an ultra-sensitive magnetometer is required
as the sensor.
In this experiment, the spin sources are two iron shielded SmCo5 magnets
(ISSCs) that have high net electron spin and small magnetic leakage
ji2017searching . The sensor is a pair of optically pumped magnetometers (OPM)
that operate in the spin-exchange-relaxation-free (SERF) mode allred2002high ;
shah2007subpicotesla , which use spin polarized Rb as the sensing atoms. By
designing the setup, the experiment is sensitive to the exotic force, while
common-mode noise is reduced. Our experiment sets new limits on exotic SSVDI
for electron-proton coupling.
Figure 1 shows a schematic of the experimental setup. Each of the two spin
sources ISSC1,2 contains a 40.00 mm diameter cylindrical SmCo5 magnet enclosed
in three layers of pure iron. The magnetization of the SmCo5 magnets is about
1 T. The magnetic field of the magnet is shielded by the iron layers, and the
magnetic leakage outside the iron layers is smaller than 10 $\mu$T. However,
the net spin of the ISSCs is not canceled, which is mostly due to the fact
that the orbital magnetic moment and spin magnetic moments of the 4$f$ rare
earth metal (Sm) and 3$d$ metals (Co and Fe) are differently oriented, and
thus the total magnetic moments are canceled but the net orbital magnetic
moment and spin magnetic moment are not ji2017searching ; heckel2008preferred
. The net electron spin for each ISSC is 1.75 $(21)\times 10^{24}$
ji2017searching . The ISSCs are connected with titanium-alloy supports and are
driven with a motor to rotate clockwise (CW) and counterclockwise (CCW). The
motor frequency is controlled with a direct current (DC) power supply.
Figure 1: The experimental setup (not to scale). Two QuSpin OPMs noted as
OPM1 and OPM2 are enclosed in a five-layer magnetic shield. Their sensitive
axis orientations are antiparallel along the $\hat{x}$-axis. Two spin sources
noted as ISSC1,2 are put in the other, four-layer shield. The spin source is
driven with a motor to rotate clockwise or counterclockwise. The blue arrows
show the direction of net spin in OPMs and ISSCs.
The OPMs are QuSpin Vector Zero-Field Magnetometers (QZFM Gen-2) Quspin that
work in the SERF regime. They are placed in the center of a five-layer
$\mu$-metal magnetic shield. The arrows along the $\hat{z}$-axis demonstrate
the direction of the circularly polarized laser beam that passes through the
87Rb vapor cell. A narrow linewidth Rb Hannle resonance in near-zero field is
used to detect the magnetic field dupont1969detection . Because the
orientation of two OPMs along $\hat{x}$ are anti-parallel, their responses to
the magnetic field have opposite signs. If there is a magnetic field $B_{0}$
applied, the responses of the OPMs are $S_{1}=B_{0}+N_{C}+N_{1}$ and
$S_{2}=-B_{0}+N_{C}+N_{2}$, respectively, where $N_{C}$ is the common-mode
noise and $N_{1}$ and $N_{2}$ are other noises. Subtracting the readings of
the two sensors can diminish the common noise and yields a signal of
$S_{\mathrm{sub}}=(S_{1}-S_{2})/2=B_{0}+(N_{1}-N_{2})/2$.
To test the validity of the subtraction procedure, an 8 Hz and 1.5 pT uniform
magnetic field is applied along $\hat{x}$ with a set of Helmholtz coils. The
spectrum of the OPM signals and the subtraction result are shown in Fig. 2
(a). By taking the difference, the uniform magnetic field is unaffected, the
common-mode (for example, electrical or gradient) noise is reduced by as high
as a factor of five, and the 8 Hz target signal is successfully extracted. The
noise level around 8 Hz is about 13 fT/$\sqrt{\textrm{Hz}}$.
Figure 2: Up: A typical spectrum of two OPMs and the subtraction result. A
uniform AC magnetic field of 8 Hz is applied along the $\hat{x}$-axis. The
dashed-blue line and the red-dot-dashed line are the spectrum of the OPMs on
the left side and right side, respectively. The yellow-solid line is their
difference. Down: The OPMs’ response to the pseudomagnetic field along the
$\hat{x}$-axis. The blue-dot and red-dot line are the pseudomagnetic field
sensed by OPM1 and OPM2 respectively, the black-dash line is the subtraction
result. The subtraction result agrees well with the result from OPM1.
The SSVDIs will manifest as a pseudomagnetic fields that could be sensed by
the Rb atoms like the Zeeman effect. The potential can be expressed as
$V^{n}\zeta^{n}+V^{p}\zeta^{p}+V^{e}\zeta^{e}=-\bm{\mu}\cdot\textbf{B}$, where
$\bm{\mu}$ is the magnetic moment of the Rb atom, B is the pseudomagnetic
field from the exotic interaction, $\zeta^{n,p,e}$ are the neutron, proton and
electron’s fraction of spin polarization in 87Rb atoms, which could be
obtained by the the Russel-Saunders LS-coupling and the Schmidt model of
nuclear physics.
In this experiment, we search for the coupling between the proton, neutron and
electron spins in the Rb atoms and the electron spins in ISSCs. The
pseudomagnetic field sensed by the OPM can be obtained by integrating the
exotic interaction from the electron spins over the ISSCs:
$\displaystyle\begin{split}\mathbf{B}^{p,e,n}=&\frac{f\zeta^{p,e,n}\hbar}{4\pi\mu
c}\iiint\rho(\mathbf{r})(\hat{\bm{\sigma}}_{2}\cdot\mathbf{v})\frac{\mathbf{v}}{r}e^{-r/\lambda}\rm{d}\mathbf{r},\end{split}$
(2)
where $\textbf{B}^{p,n,e}$ are the fractions of B that couple to proton,
neutron and electron respectively,
$\mathbf{v}(\mathbf{r})=\mathbf{\omega}\times\mathbf{r}$ and
$\rho(\mathbf{r})$ are the velocity and spin density at location $\mathbf{r}$,
$\mathbf{\omega}$ is the angular velocity of the ISSCs. The proton and
electron fractions of polarization in 87Rb are $\zeta^{p}=0.29$ and
$\zeta^{e}=0.13$ respectively, and neutron polarization $\zeta^{n}$ is assumed
to be zero under the basic nuclear shell model. The calculation of the
fraction of spin polarization is explained in the supplemental document.
The experimental parameters and a benchmark coupling coefficients $f^{0}=1$
are put in the simulation to obtain $\textbf{B}^{p}$. The benchmark parameter
$f^{0}$ is set to $1$ for convenience; a different $f^{0}$ does not affect the
final result. The orientation of the OPMs and ISSC sources are optimized by
simulating different configurations, such that the OPMs can sense the maximum
pseudomagnetic field. The best configuration is shown in Figure 1 and Table 1.
The distance between the spin sources and the OPMs is much larger than the
distance between two OPMs, such that two OPMs experience almost the same
pseudomagnetic field. Thus the signal subtraction procedure works well for
this pseudomagnetic field. The simulated responses of the two OPMs and their
subtraction result are shown in Fig. 2 (b).
Table 1: Experimental parameters and the error budget of $f^{\rm{ep}}$. The origin of coordinates is at the midpoint between the centers of the two OPMs. The contributions to the error budget are evaluated for $\lambda=20$ m. The final systematic error is derived from the uncertainties of the parameters listed. Parameter | Value | $\Delta f^{\rm{ep}}(\times 10^{-{22}})$
---|---|---
ISSC net spin ($\times 10^{24}$) | $1.75(21)$ | 0.084
Position of ISSCs x(m) | $0.000(2)$ | $0.001$
Position of ISSCs y(m) | $-0.477(2)$ | $0.001$
Position of ISSCs z(m) | $0.000(2)$ | $0.001$
Distance between ISSCs(m) | $0.251(1)$ | $0.044$
Distance between OPM cells(m) | $0.017(1)$ | $0.004$
Rotation frequency CW (Hz) | $4.11(1)$ |
Rotation frequency CCW (Hz) | $4.09(1)$ |
phase uncertainty ($\deg$) | $\pm 2.8$ | $\pm 1.190$
Final $f^{\rm{exp}}(\times 10^{-22}$) | $-0.7$ | $\pm 10.1\ (stat.)$
$(\lambda=20\,m)$ | | $\pm 1.2(syst.)$
The ISSC spin sources are driven with a DC motor. The positions of the spin
sources are monitored with a photoelectronic encoder placed on the rotation
axle. The signals of the encoder and the OPMs are taken simultaneously and
recorded with a data-acquisition (DAQ) device. The motor is tuned to rotate
CCW and CW alternatively for every two hours. The DC motor works in a good
stability with frequency of $4.09(1)$ Hz and $4.11(1)$ Hz for CW and CCW
rotations.
The two OPMs signals are subtracted and then transformed to frequency domain
by fast Fourier transformations (FFT). The 50 Hz power line interference and
its 100 Hz and 200 Hz harmonics are removed in the frequency domain. The data
were then transformed back to the time domain with inverse FFT.
The signals are then cut to one-period-long segments based on the encoder
signal of the spin source rotation. The DC components in each period are
removed. The data are noted as $\mathbf{S}^{\rm{exp}}_{i}(t_{j})$, where $i$
represents the $i$-th period, and $t_{j}$ is the time of the $j$-th point in
this period.
The coupling coefficient $f^{\rm{ep}}$ can be obtained by a similarity
comparison method between the experimental data and simulation results:
$f^{\rm{ep}}_{i}=k_{i}\,\sqrt{\frac{\sum_{j}\left[\mathbf{S}^{\rm{exp}}_{i}(t_{j})\right]^{2}}{\sum_{j}\left[\mathbf{S}^{\rm{sim}}(t_{j})\right]^{2}}}\,,$
(3)
where $k_{i}$ is the similarity score to weigh the similarity between
$\mathbf{S}^{\rm{exp}}_{i}$ and $\mathbf{S}^{\rm{sim}}(t)$ krasichkov2015shape
, which is defined as
$k_{i}\equiv\frac{\sum_{j}\mathbf{S}^{\rm{sim}}(t_{j})\cdot{\mathbf{S}}_{i}^{\rm{exp}}(t_{j})}{\sqrt{\sum_{j}\left[\mathbf{S}^{\rm{sim}}(t_{j})\right]^{2}}\sqrt{\sum_{j}\left[{\mathbf{S}_{i}}^{\rm{exp}}(t_{j})\right]^{2}}}.$
(4)
The expectation values and standard error for the CW and CCW rotation are
$\langle f^{\rm{ep}}\rangle^{+}$, $\langle f^{\rm{ep}}\rangle^{-}$, and
$\sigma_{f}^{+}$, $\sigma_{f}^{-}$ respectivly, the final coupling coefficient
can be obtained by
$\langle f^{\rm{ep}}\rangle=\frac{\langle
f^{\rm{ep}}\rangle^{+}/{\sigma^{+}}^{2}+\langle
f^{\rm{ep}}\rangle^{-}/{\sigma^{-}}^{2}}{1/{\sigma^{+}}^{2}+1/{\sigma^{-}}^{2}}.$
(5)
Some systematic bias could be removed by averaging over CW an CCW. The
distributions of the $f^{\rm{ep}+}$ and $f^{\rm{ep}-}$ are shown in Fig. 3.
Figure 3: Statistical results of the $f^{\rm{ep}}$. Each data point represents
an average of about one 2.7-hour long data set. The distribution of
$f^{\rm{ep}}$ for one data set is shown in the insert. The result is well
fitted with a Gaussian distribution (red line) with $\bar{\chi}^{2}=1.18$.
Figure 4: Limits on the SSVDI coupling coefficients between electron and
proton. The black solid line is our constraints. The “Hunter2014 e-p” is from
Ref. hunter2014using that uses geo-electrons and atomic magnetometer; the
“Fadeev 2022 e-p” is from Refs. Fadeev2022Pseudovector that compare the
experimental and theoretical results of hydrogen spectroscopy; the “Chu 2020
e-p” chu2020search propose to use 3He as sensor and dysprosium iron garnet as
spin source, the line is based on their sensitivity at $3\times 10^{-17}$ T
and is rescaled using the fraction of spin polarisation
$\zeta_{p}^{{}^{3}\text{He}}=-0.027$ vasilakis2009limits .
The parameters of the experiment and their corresponding uncertainties on
$\Delta f^{\rm{exp}}$ for range $\lambda=20$ m are shown in Table 1. The
$f^{\rm{ep}}$ is determined to be $f^{\rm{ep}}=0.7\pm 10.1_{stat.}\pm
1.2_{syst.}(\times 10^{-22})$. No evidence of the SSVDI is observed. New
constraints on the $f$ between electron-proton is set to be $|f^{\rm{ep}}|\leq
2.0\times 10^{-21}$ by the 95% confidence level. The values for other
$\lambda$s are obtained with the same procedure, and the final limits are
shown in Fig. 4.
If the mediator of the SSVDI is a spin-1 boson such as Z′, which is a dark
matter candidate and may resolve other discrepancies such as that in the
anomalous magnetic moment of the muon arushi2021solving ; okada2020dark , the
coupling coefficient can be rewritten as $f^{\rm{ep}}=-g_{A}^{e}g_{A}^{p}/2$
dobrescu2006spin ; Fadeev2019 . For $\lambda=20$ m, $|g_{A}^{e}g_{A}^{p}|\leq
4.0\times 10^{-21}$, where to set the limit on one of these coupling-constant
products, we assume that the other one is zero. Note that the velocity-
independent term provides significantly tighter limit on $g_{A}g_{A}$
coefficients Fadeev2019 ; dobrescu2006spin , however, the SSVDI provides a
unique way to explore the velocity-dependent interactions.
A comparison between our results (black and dashed-red lines) and the
literature is shown in Fig. 4. With the same hydrogen-spectrum analysis used
in Ref. Fadeev2022Pseudovector we obtained a bound on the SSVDF of
$f^{ep}<2.0\times 10^{-11}$ for the range larger than 1 cm (the green line
“Fadeev 2022 e-p” in Fig. 4). Results on the couplings between other fermions,
such as the coupling between electron-electron ficek2017constraints ;
hunter2014using ; Ji2018 neutron-proton kimball2010constraints and electron-
antiproton ficek2018constraints are not plotted on Fig. 4.
The major advance of our experiment is that the ISSC spin sources have much
larger numbers of spins compared to those in precision-spectroscopy
experiments yielding data for the analyses in ficek2017constraints ;
ficek2018constraints and spin-exchange approaches kimball2010constraints ,
which are most sensitive to forces with ranges on the atomic to microscopic
scale. The other advantage is that the OPMs typically have energy resolution
on the order of $10^{-18}$ eV kimball2015nuclear , significantly better than
for the spectroscopy used in Refs.ficek2017constraints ; ficek2018constraints
. On the other hand, spectroscopy experiments have an advantage over
macroscopic once in the short range, because of the exponential decay of the
exotic force. Our search covers the range of parameters inaccessible for the
geoelectron experiment hunter2014using . Using the same method and data, we
also set limits for the electron-proton coupling on the V6+7, V15 and V16
terms of SSVDF dobrescu2006spin ; Fadeev2019 . The results are shown in the
supplemental document.
A major concern in this experiment was magnetic leakage from the ISSCs. With
the iron shielding, at a distance of 10 cm away from the ISSC’s surface, its
residual magnetic field was measured to be less than 10 $\mu$T. The shielding
factors for the magnetic shielding of the ISSCs and OPMs were measured to be
both greater than $10^{6}$. Considering all the decay and shielding factors,
we conservatively expect the magnetic leakage from the ISSCs to the position
of the OPMs to be smaller than 0.1 aT, which was insignificant with regards to
the error budget.
The stability of the OPM is monitored throughout the experiment. The DC drift
of the OPM is less than 2 pT within two hours. A servo motor has a better
frequency precision, however, commercial servo motor’s control systems have
electromagnetic coupling with the magnetometer Ji2018 . A DC motor is chosen
to diminish this coupling. The experiment can further be improved if a larger
size ISSC could be used. The vapor cell can also be replaced with a
magnetometer that uses a levitated ferromagnetic sphere and has orders of
magnitude better potential magnetic sensitivity vinante2021surpassing .
In summary, we utilized a pair of OPMs that can reduce the common noise and
have ultrahigh sensitivity to search for exotic spin-dependent physics.
Together with the high electron spin density iron-shielded SmCo5 spin source,
the new experiment sets new limits on SSVDI, with more than 10 orders of
magnitude improvement for the electron-proton coupling.
###### Acknowledgements.
We thank Dr. Derek F. Jackson Kimball and Dr. Changbo Fu for valuable
discussions. This work is supported by Key-Area Research and Development
Program of Guangdong Province (Grant No. 2019B030330001), the National Natural
Science Foundation of China (NSFC) (Grant Nos. 12275145, 92050110, 91736106,
11674390, and 91836302), and the National Key R&D Program of China (Grants No.
2018YFA0306504), the DFG Project ID 390831469: EXC 2118 (PRISMA+ Cluster of
Excellence), by the German Federal Ministry of Education and Research (BMBF)
within the Quantumtechnologien program (Grant No. 13N15064). and by the
QuantERA project LEMAQUME (DFG Project Number 500314265).
## References
* (1) R.D.Peccei & Quinn, H. Cp conservation in the presence of pseudoparticles. _Phys. Rev. Lett._ 38, 1440 (1977).
* (2) Preskill, J., Wise, M. B. & Wilczek, F. Cosmology of the invisible axion. _Phys. Lett. B_ 120, 127–132 (1983).
* (3) Graham, P. W., Irastorza, I. G., Lamoreaux, S. K., Lindner, A. & van Bibber, K. A. Experimental searches for the axion and axion like particles. _Annual Review of Nuclear and Particle Science_ 65, 485–514 (2015).
* (4) Jaeckel, J. & Ringwald, A. _Phys. Rev. Lett._ 40, 223 (1978).
* (5) An, H., Pospelov, M., Pradler, J. & Ritz, A. Direct detection constraints on dark photon dark matter. _Phys. Lett. B_ 747, 331 – 338 (2015).
* (6) Okada, N., Okada, S., Raut, D. & Shafi, Q. Dark matter z’ and xenon1t excess from u (1) x extended standard model. _Phys. Lett. B_ 810, 135785 (2020).
* (7) Croon, D., Elor, G., Leane, R. K. & McDermott, S. D. Supernova muons: new constraints on z’ bosons, axions and alps. _J. High Energy Phys._ 2021, 1–28 (2021).
* (8) Moody, J. & Wilczek, F. New macroscopic forces? _Phys. Rev. D_ 30, 130 (1984).
* (9) Dobrescu, B. A. & Mocioiu, I. Spin-dependent macroscopic forces from new particle exchange. _J. High Energy Phys._ 2006, 005 (2006).
* (10) Fadeev, P. _et al._ Revisiting spin-dependent forces mediated by new bosons : Potentials in the coordinate-space representation for macroscopic- and atomic-scale experiments. _Phys. Rev. A_ 022113, 1–7 (2019).
* (11) Fayet, P. The fifth interaction in grand-unified theories: a new force acting mostly on neutrons and particle spins. _Phys. Lett. B_ 172, 363–368 (1986).
* (12) Heckel, B. R. _et al._ Preferred-frame and c p-violation tests with polarized electrons. _Phys. Rev. D_ 78, 092006 (2008).
* (13) Hammond, G. D., Speake, C. C., Trenkel, C. & Patón, A. P. New constraints on short-range forces coupling mass to intrinsic spin. _Phys. Rev. Lett._ 98, 081101 (2007).
* (14) Long, J. _et al._ Upper limits to submillimetre-range forces from extra space-time dimensions. _NATURE_ 421, 922–925 (2003).
* (15) Terrano, W., Adelberger, E., Lee, J. & Heckel, B. Short-range, spin-dependent interactions of electrons: A probe for exotic pseudo-goldstone bosons. _Phys. Rev. Lett._ 115, 201801 (2015).
* (16) Petukhov, A., Pignol, G., Jullien, D. & Andersen, K. Polarized he 3 as a probe for short-range spin-dependent interactions. _Phys. Rev. Lett._ 105, 170401 (2010).
* (17) Yan, H. _et al._ Searching for new spin-and velocity-dependent interactions by spin relaxation of polarized he 3 gas. _Phys. Rev. Lett._ 115, 182001 (2015).
* (18) Chu, P.-H. _et al._ Laboratory search for spin-dependent short-range force from axionlike particles using optically polarized he 3 gas. _Phys. Rev. D_ 87, 011105 (2013).
* (19) Arvanitaki, A. & Geraci, A. A. Resonantly detecting axion-mediated forces with nuclear magnetic resonance. _Phys. Rev. Lett._ 113, 161801 (2014).
* (20) Vasilakis, G., Brown, J. M., Kornack, T. W. & Romalis, M. V. Limits on new long range nuclear spin-dependent forces set with a $\mathbf{K}\mathrm{\text{-}}^{3}\mathrm{He}$ comagnetometer. _Phys. Rev. Lett._ 103, 261801 (2009).
* (21) Ji, W. _et al._ New Experimental Limits on Exotic Spin-Spin-Velocity-Dependent Interactions by Using SmCo 5 Spin Sources. _Phys. Rev. Lett._ 121, 261803 (2018).
* (22) Jiao, M., Guo, M., Rong, X., Cai, Y.-F. & Du, J. Experimental constraint on an exotic parity-odd spin-and velocity-dependent interaction with a single electron spin quantum sensor. _Phys. Rev. Lett._ 127, 010501 (2021).
* (23) Hunter, L. & Ang, D. Using geoelectrons to search for velocity-dependent spin-spin interactions. _Phys. Rev. Lett._ 112, 091803 (2014).
* (24) Su, H. _et al._ Search for exotic spin-dependent interactions with a spin-based amplifier. _Science advances_ 7, eabi9535 (2021).
* (25) Almasi, A., Lee, J., Winarto, H., Smiciklas, M. & Romalis, M. V. New limits on anomalous spin-spin interactions. _Phys. Rev. Lett._ 125, 201802 (2020).
* (26) Kim, Y. J., Chu, P.-H. & Savukov, I. Experimental constraint on an exotic spin-and velocity-dependent interaction in the sub-mev range of axion mass with a spin-exchange relaxation-free magnetometer. _Phys. Rev. Lett._ 121, 091802 (2018).
* (27) Tullney, K. _et al._ Constraints on spin-dependent short-range interaction between nucleons. _Phys. Rev. Lett._ 111, 100801 (2013).
* (28) Serebrov, A. P. _et al._ Search for macroscopic cp violating forces using a neutron edm spectrometer. _JETP letters_ 91, 6–10 (2010).
* (29) Ficek, F. _et al._ Constraints on exotic spin-dependent interactions between electrons from helium fine-structure spectroscopy. _Phys. Rev. A_ 95, 032505 (2017).
* (30) Ren, X. _et al._ Search for an exotic parity-odd spin-and velocity-dependent interaction using a magnetic force microscope. _Phys. Rev. D_ 104, 032008 (2021).
* (31) Stadnik, Y., Dzuba, V. & Flambaum, V. Improved limits on axionlike-particle-mediated p, t-violating interactions between electrons and nucleons from electric dipole moments of atoms and molecules. _Phys. Rev. Lett._ 120, 013202 (2018).
* (32) Wei, K. _et al._ New constraints on exotic spin-velocity-dependent interactions. _arXiv preprint arXiv:2203.07050_ (2022).
* (33) Ji, W., Fu, C. & Gao, H. Searching for new spin-dependent interactions with smco 5 spin sources and a spin-exchange-relaxation-free comagnetometer. _Phys. Rev. D_ 95, 075014 (2017).
* (34) Allred, J., Lyman, R., Kornack, T. & Romalis, M. V. High-sensitivity atomic magnetometer unaffected by spin-exchange relaxation. _Phys. Rev. Lett._ 89, 130801 (2002).
* (35) Shah, V., Knappe, S., Schwindt, P. D. & Kitching, J. Subpicotesla atomic magnetometry with a microfabricated vapour cell. _Nature Photonics_ 1, 649–652 (2007).
* (36) GZFM Gen-2, QuSpin. http://quspin.com/products-qzfm. Accessed: 2021-03-30.
* (37) Dupont-Roc, J., Haroche, S. & Cohen-Tannoudji, C. Detection of very weak magnetic fields (10- 9gauss) by 87rb zero-field level crossing resonances. _Phys. Lett. A_ 28, 638–639 (1969).
* (38) Krasichkov, A. S., Grigoriev, E. B., Bogachev, M. I. & Nifontov, E. M. Shape anomaly detection under strong measurement noise: An analytical approach to adaptive thresholding. _Phys. Rev. E_ 92, 042927 (2015).
* (39) Fadeev, P., Ficek, F., Kozlov, M. G., Budker, D. & Flambaum, V. V. Pseudovector and pseudoscalar spin-dependent interactions in atoms. _Phys. Rev. A_ 105, 022812 (2022).
* (40) Chu, P.-H., Kim, Y. & Savukov, I. Search for exotic spin-dependent interactions using polarized helium. _arXiv preprint arXiv:2002.02495_ (2020).
* (41) Vasilakis, G., Brown, J., Kornack, T. & Romalis, M. Limits on new long range nuclear spin-dependent forces set with a k- he 3 comagnetometer. _Phys. Rev. Lett._ 103, 261801 (2009).
* (42) Arushi, B. & Rupert, C. Solving the electron and muon g-2 anomalies in z’ models. _Eur. Phys. J. C, Particles and Fields._ 81 (2021).
* (43) Kimball, D. J., Boyd, A. & Budker, D. Constraints on anomalous spin-spin interactions from spin-exchange collisions. _Phys. Rev. A_ 82, 062714 (2010).
* (44) Ficek, F. _et al._ Constraints on exotic spin-dependent interactions between matter and antimatter from antiprotonic helium spectroscopy. _Phys. Rev. Lett._ 120, 183002 (2018).
* (45) Kimball, D. J. Nuclear spin content and constraints on exotic spin-dependent couplings. _New Journal of Physics_ 17, 073008 (2015).
* (46) Vinante, A. _et al._ Surpassing the energy resolution limit with ferromagnetic torque sensors. _Phys. Rev. Lett._ 127, 070801 (2021).
|
# Non-relativistic torque and Edelstein effect in noncollinear magnets
Rafael González-Hernández1, Philipp Ritzinger2, Karel Výborný2, Jakub
Železný2, Aurélien Manchon2 1Grupó de Investigación en Física Aplicada,
Departamento de Física, Universidad del Norte, Barranquilla, Colombia; 2
Institute of Physics, Czech Academy of Sciences, Cukrovarnická 10, 162 00
Praha 6 Czech Republic; 3Aix-Marseille Univ, CNRS, CINaM, Marseille, France
###### Abstract
The Edelstein effect is the origin of the spin-orbit torque: a current-induced
torque that is used for the electrical control of ferromagnetic and
antiferromagnetic materials. This effect originates from the relativistic
spin-orbit coupling, which necessitates utilizing materials with heavy
elements. Here we show that in magnetic materials with non-collinear magnetic
order, the Edelstein effect and consequently also a current-induced torque can
exist even in the absence of the spin-orbit coupling. Using group symmetry
analysis, model calculations, and realistic simulations on selected compounds,
we identify large classes of non-collinear magnet candidates and demonstrate
that the current-driven torque is of similar magnitude as the celebrated spin-
orbit torque in conventional transition metal structures. We also show that
this torque can exist in an insulating material, which could allow for highly
efficient electrical control of magnetic order.
## I Introduction
In materials and heterostructures with spin-orbit coupling, the
interconnection between the spin and momentum degrees of freedom of the
electronic Bloch states underscores a rich landscape of microscopic ”spin-
orbitronics” phenomena, such as anomalous Hall effect Nagaosa _et al._ (2010)
and anisotropic magnetoresistance Ritzinger and Vyborny (2022), spin Hall
effect Sinova _et al._ (2015), Dzyaloshinskii-Moriya interaction or spin-
orbit torques Soumyanarayanan _et al._ (2016); Manchon _et al._ (2019). To
maximize these effects, materials displaying reasonably large spin-orbit
coupling are necessary, which implies using metals with large atomic numbers
Z, such as Pt, W, Bi, etc. Some of these elements are however scarce,
expensive, and environmentally unfriendly. In addition, arbitrarily large
spin-orbit coupling does not necessarily lead to arbitrarily large spin-
orbitronics phenomena Sunko _et al._ (2017); Železný _et al._ (2021) because
of the competition with crystal field and exchange.
Contrary to a common conception though, spin-orbit coupling is not a mandatory
ingredient to obtain spin-momentum locking. In fact, as noticed by Pekar and
Rashba in the mid-sixties Pekar and Rashba (1965), electronics states in
materials with a spatially inhomogeneous magnetization display a spin texture
in momentum space that share similarities with the one obtained through spin-
orbit coupling. In other words, non-collinear magnetism mimics spin-orbit
coupling to some extent and can support a number of phenomena that are well
known in spin-orbit coupled materials such as electric-dipole spin resonance
Pekar and Rashba (1965); Ramazashvili (2009), topological Hall effect Shindou
and Nagaosa (2001); Martin and Batista (2008); Ndiaye _et al._ (2019), spin
Hall effect Zhang _et al._ (2018a), and magnetic spin Hall effect Zelezný
_et al._ (2017); Kimata _et al._ (2019), the latter being specific to
magnetic materials. It is therefore natural to wonder whether another hallmark
of spin-orbit coupled materials, the Edelstein effect F.T. Vasko (1979);
Edelstein (1990) (also called the Rashba-Edelstein effect, inverse spin-
galvanic effect, or the magneto-electric effect), and its associated spin-
orbit torque can also be achieved in spin-orbit free non-collinear magnets.
The Edelstein effect refers to the generation of non-equilibrium spin density
by an applied electric field in non-centrosymmetric semiconducting or metallic
materials and heterostructures with spin-orbit coupling. The magnitude of the
nonequilibrium spin density is governed by the competition between the spin-
orbit coupling energy and the crystal field energy associated with inversion
symmetry breaking. In magnetic materials, the spin-momentum locking is
governed by the magnetic exchange between local and itinerant electrons,
rather than by the atomic spin-orbit coupling, suggesting that a large
Edelstein effect can be obtained in non-centrosymmetric magnetic materials. A
possible advantage of such a mechanism is that it does not require the
presence of heavy elements and it could exist even in materials with
negligible spin-orbit coupling such as organic magnetic materials. In
addition, since the magnitude of the Edelstein effect is directly related to
the magnetic configuration of the material, it should be highly tunable using
an external magnetic field.
A first step towards the realization of this non-relativistic Edelstein effect
was established recently in a model non-collinear coplanar kagomé
antiferromagnet with in-plane broken inversion symmetry Hayami _et al._
(2020). In this particular case, the current-driven spin density is polarized
out of plane. What makes this Edelstein effect particularly appealing is that
the nonequilibrium spin density can exert a torque on the local magnetization
that is responsible for generating this spin density. This feature suggests
that current-driven dynamics, high-frequency excitation, and reversal, might
exhibit properties fundamentally different from those usually observed in
spin-transfer and spin-orbit torque devices. We have recently noticed such a
torque in a work studying spin-transfer torque in magnetic junctions composed
of centrosymmetric non-collinear antiferromagnets Ghosh _et al._ (2022).
Here, we study the non-relativistic Edelstein effect in non-collinear systems
in detail, focusing on systems where the spin-orbit torque is normally
studied, i.e. bulk non-centrosymmetric magnets and heterostructures with
current flowing parallel to the interface.
We demonstrate that wide classes of antiferromagnets lacking a center of
inversion can support the current-driven Edelstein effect in the absence of
spin-orbit coupling, and therefore also current-driven spin torques. We
establish general symmetry principles to discover new materials, propose
selected promising candidates, demonstrate and quantify the effect in specific
materials, and extend the idea to the case of magnetic multilayers. We also
show the existence of non-relativistic antisymmetric spin-textures in
reciprocal space, which are in some cases directly related to the Edelstein
effect, and discuss their general symmetry properties. We implemented an
algorithm for determining the symmetry of the non-relativistic Edelstein
effect as well as other non-relativistic phenomena and we released it within
an open source code. Remarkably, we show that the non-relativistic Edelstein
effect can also be present in insulating materials. This could allow for
controlling the magnetic order by a voltage in the absence of any Ohmic
conduction, resulting in a much higher efficiency than the conventional
current-induced torques.
## II General principles
### II.1 Conditions for an antisymmetric spin texture
In non-magnetic materials lacking inversion symmetry, the relativistic
Edelstein effect is associated with an antisymmetric spin texture in the
reciprocal space Bihlmayer _et al._ (2022). These spin textures arise from
the spin-momentum locking imposed by the spin-orbit coupling and are
characterized by a spin direction that varies in momentum space. In the
absence of spin-orbit coupling and in the presence of non-collinear magnetism,
one expects non-relativistic analogs of the antisymmetric spin textures.
Therefore, before addressing the non-relativistic Edelstein effect and its
associated torque, we first consider the conditions of the emergence of such
antisymmetric spin textures. Recently, spin textures in the absence of spin-
orbit coupling have been studied in non-collinear Zelezný _et al._ (2017);
Yuan _et al._ (2021); Hayami _et al._ (2020) as well as in collinear
magnetic materials Hayami _et al._ (2019); Yuan _et al._ (2020, 2021);
Šmejkal _et al._ (2022); Egorov _et al._ (2021). In collinear systems,
however, the direction of spin is fixed and only the magnitude and sign of the
spin-splitting varies in momentum space. In addition, most of the non-
relativistic spin-textures studied so far (with the exception of Ref. Hayami
_et al._ , 2020) are typically symmetric in momentum $\mathbf{k}$,
$\mathbf{S}_{n\mathbf{k}}=\mathbf{S}_{n-\mathbf{k}}$, $n$ being the band
index, which forbids the realization of the non-relativistic Edelstein effect
at the level of the magnetic unit cell.
In the absence of relativistic spin-orbit coupling, the spin and orbital
degrees of freedom are decoupled, which also means that the spin is not
coupled to the lattice. In such a case the symmetry of magnetic systems is
described by the so-called spin space groups Brinkman and Elliott (1966);
Litvin and Opechowski (1974). In addition to crystallographic symmetry
operations that form the magnetic space groups, which describe the
relativistic symmetry of magnetic systems, the spin space groups contain also
pure spin rotations. Elements of the spin space groups can be written in the
form $\\{R_{s}||R|{\bm{\tau}}\\}$, where $R_{s}$ denotes the spin-rotation,
$R$ is a crystallographic point group operation, i.e. a proper or improper
rotation, and ${\bm{\tau}}$ is a translation. We denote symmetry operations
that contain time-reversal as $\\{R_{s}||R|{\bm{\tau}}\\}^{\prime}$.
In a 3D periodic system, the rules for the existence of spin-splitting are
simple to determine. For an arbitrary $\mathbf{k}$-point (that is, a
$\mathbf{k}$-point lying away from any high-symmetry lines or planes), the
only symmetry operations that can keep the spin invariant are the combined
space-inversion and time-reversal (the so-called $P\cal{T}$ symmetry), a pure
spin rotation, translation, or any combination of these symmetry operations.
In a $P\cal{T}$ symmetric system, the bands with opposite spin must be
degenerate, which is known as Kramers degeneracy and holds even in the
presence of spin-orbit coupling. If a pure spin rotation is present, the spin
of all states must lie along the spin-rotation axis. If more than one spin
rotation with different spin axes is present, this cannot be satisfied without
a spin degeneracy. This can also be seen from the fact that spin rotations
along different axes do not commute. Since translation does not change spin,
the same conclusions apply to symmetry operations that contain translation.
Thus spin-splitting exists in all systems, except those that have a $P\cal{T}$
symmetry or two spin rotation axes. In ferromagnetic systems, spin splitting
can exist anywhere in the Brillouin zone since no symmetry operations
connecting states with opposite spin exist. In antiferromagnetic materials,
there are specific lines or planes where the bands with opposite spin must be
degenerate, see Ref. Krempaský _et al._ (2023). This has been studied
systematically for collinear antiferromagnets Hayami _et al._ (2019); Yuan
_et al._ (2020, 2021); Šmejkal _et al._ (2020); Egorov _et al._ (2021). Note
that the spin-split collinear antiferromagnets have sometimes been referred to
as “altermagnets” Šmejkal _et al._ (2022).
In a collinear magnetic system, any spin rotation along the magnetic axis is a
symmetry. Thus if there exists another spin-rotation around a perpendicular
axis, the bands must be degenerate. Such a spin rotation must contain
translation (otherwise the system could not be magnetic) and can only be a
$180^{\circ}$ rotation, which in a collinear system has the same effect on the
magnetic order as the time-reversal. The existence of such a symmetry thus
implies that the system is invariant under a $\cal{T}{\bm{\tau}}$ (a combined
time-reversal and translation) symmetry. Collinear antiferromagnetic systems
can thus be separated into three types. In systems with $P\cal{T}$ symmetry,
bands are spin degenerate even with spin-orbit coupling. In systems with
$\cal{T}{\bm{\tau}}$ but broken $P\cal{T}$ symmetry, spin-splitting occurs
only when the spin-orbit coupling is present. Finally, in systems with broken
$P\cal{T}$ and $\cal{T}{\bm{\tau}}$ symmetries, a non-relativistic spin-
splitting can be present. In non-collinear magnets, this does not hold since
time reversal does not have the same effect as a $180^{\circ}$ spin rotation,
and spin rotations with different angles can also occur.
The existence of antisymmetric spin textures is governed by symmetries that
transform $\mathbf{k}\rightarrow-\mathbf{k}$. This involves in particular the
inversion symmetry, which implies
$\mathbf{S}_{n\mathbf{k}}=\mathbf{S}_{n-\mathbf{k}}$. In systems with
inversion symmetry, any spin texture thus must be symmetric. In a coplanar
system a combined spin-rotation and time-reversal operation
$\\{R_{s}(\hat{\mathbf{n}}_{\perp},180^{\circ})||E||E\\}^{\prime}$ is a
symmetry. Here $R_{s}(\hat{\mathbf{n}}_{\perp},180^{\circ})$ denotes a spin
rotation by $180^{\circ}$ around the direction perpendicular to the magnetic
plane $\hat{\mathbf{n}}_{\perp}$. As a consequence, it must hold that
$\mathbf{S}^{||}_{n\mathbf{k}}=\mathbf{S}^{||}_{n-\mathbf{k}}$ and
$\mathbf{S}^{\perp}_{n\mathbf{k}}=-\mathbf{S}^{\perp}_{n-\mathbf{k}}$, where
$\mathbf{S}^{||}$ and $\mathbf{S}^{\perp}$ denote the components of spin
parallel and perpendicular to the plane, respectively. In a coplanar system,
the only antisymmetric component is thus perpendicular to the magnetic plane,
as in the case studied by Hayami et al. Hayami _et al._ (2020). We note that
even when all magnetic moments lie within a plane, the electron spins can
contain an out-of-plane component. In a collinear magnetic system, this is not
possible, however, since in this case spin is a good quantum number and all
spins must lie along a single axis. There, any non-relativistic spin splitting
must thus be symmetric in momentum.
### II.2 Conditions for nonequilibrium spin densities and torques
Let us now turn our attention towards the non-relativistic Edelstein effect.
The nonequilibrium properties of materials obtained via the Kubo formula are
often parsed into so-called Fermi surface and Fermi sea contributions Železný
_et al._ (2017); Bonbien and Manchon (2020), the former being even under
$\cal{T}$ and the latter being odd. In the context of the Edelstein effect,
the $\cal{T}$-even Fermi surface contribution is related to the antisymmetric
spin texture in momentum space F.T. Vasko (1979); Edelstein (1990), whereas
the $\cal{T}$-odd Fermi sea contribution is related to the Berry curvature in
mixed spin-momentum space in the weak scattering limit Kurebayashi _et al._
(2014); Li _et al._ (2015); Hanke _et al._ (2017). As a consequence, in
spin-orbit coupled non-centrosymmetric magnetic heterostructures, the Fermi
surface contribution produces the so-called field-like torque whereas the
Fermi sea contribution is responsible for the antidamping-like torque Manchon
_et al._ (2019). Notice that the $\cal{T}$-odd Fermi sea contribution can also
be nonzero in $P\cal{T}$ symmetric antiferromagnets with Kramers degeneracy.
Furthermore, for manipulating the magnetic order in more complex magnetic
systems, especially in antiferromagnets, the nonequilibrium spin density one
should be concerned with is the local one, projected on the magnetic
sublattices, rather than the global one, at the level of the magnetic unit
cell Železný _et al._ (2014, 2017). The $\cal{T}$-even component of the local
Edelstein effect can be understood as originating from the antisymmetric spin
texture obtained upon projecting on the local atom. Such a “hidden” texture
can again exist even in systems with Kramers degeneracy Zhang _et al._
(2014). Consequently, the symmetry conditions that allow for the existence of
an Edelstein effect and a torque on the magnetic order are distinct from those
for the existence of anti-symmetric spin textures.
The symmetry of the non-relativistic global and local Edelstein effects and
the resulting torque can be determined in a similar fashion as for the
relativistic one, just replacing the magnetic space groups with spin groups.
The key symmetry that needs to be broken for the existence of the Edelstein
effect is the inversion symmetry. This holds regardless of the presence of
spin-orbit coupling. For the global Edelstein effect, the global inversion
symmetry must be broken, whereas for the local Edelstein effect, it has to be
broken locally (see, e.g., Zhang _et al._ (2014)). This means that for the
presence of the Edelstein effect on a given magnetic site, there must be no
inversion symmetry operation that would leave this site invariant.
As already mentioned, in magnets the Edelstein effect can be nonzero even
without spin-orbit coupling, similar to the spin-Hall effect, for example,
Zhang _et al._ (2018b). This applies even to collinear magnets; however, in
such a case the induced spin density must be oriented along the magnetic order
and does not lead to a torque (although it could play a role for example in
the presence of magnons). Consequently, we focus here on non-collinear
magnetic systems. In presence of a pure spin-rotation
$\\{R_{s}(\hat{\mathbf{n}},\theta)||E||{\bm{\tau}}\\}\\}$, where ${\bm{\tau}}$
could also be zero, the global Edelstein effect must obey
$\mathbf{S}||\mathbf{n}$. In presence of spin-rotation coupled with time-
reversal $\\{R_{s}(\hat{\mathbf{n}},\theta)||E||{\bm{\tau}}\\}\\}^{\prime}$ it
obeys $\mathbf{S}^{\text{even}}||\mathbf{n}$ and
$\mathbf{S}^{\text{odd}}\perp\mathbf{n}$. The same holds for the local
Edelstein effect as long as the site is invariant under ${\bm{\tau}}$.
Consequently in coplanar systems, $\mathbf{S}^{\text{even}}$ must be oriented
perpendicular to the magnetic plane and $\mathbf{S}^{\text{odd}}$ must lie
within the plane for both the global and the local Edelstein effects.
To determine the full symmetry of the non-relativistic Edelstein effect, it is
necessary to consider all the symmetry operations of the spin group. We have
implemented an algorithm for determining all spin group symmetry operations of
a given magnetic system within the freely available open-source _Symmetr_ code
Zelezný (2017). The process of determining the non-relativistic symmetry is
described in detail in Supplementary Materials. We have utilized this program
to explore the symmetry of non-collinear materials from the MAGNDATA database
of magnetic materials. We have analyzed the symmetry of 484 non-collinear
magnetic materials and have found that the global Edelstein effect is allowed
in 160 of these materials, whereas the local Edelstein effect on a magnetic
sublattice is allowed in 355 compounds. The full list is given in the
Supplementary Materials Sup . As also described in the Supplementary
Materials, the Symmetr code allows one to directly obtain the non-relativistic
symmetry of the Edelstein effect (as well as other phenomena) for materials
from the MAGNDATA.
Among the noticeable materials whose crystal structure admits both a global
and local (sublattice) torque, we identified ferroelectric antiferromagnets
such as orthorhombic DyFeO3, hexagonal HoMnO3, YbMnO3, and LuFeO3, as well as
metallic antiferromagnets such as $\alpha$-Mn, Tb3Ge5 and Tb5Ge4.
Interestingly, the centro-symmetric metallic antiferromagnets Mn5Si3, Mn3(Sn,
Ge, As), and Mn3CuN do not display a global torque but do admit a local torque
on the individual magnetic sublattices. These torques are expected to induce
magnetic excitations and potentially magnetic order reversal. In the
following, we explicitly compute the global and local Edelstein effects in
both LuFeO3 and Mn3Sn as an illustration of both cases.
## III Non-relativistic Edelstein effect in non-collinear antiferromagnets
To calculate the Edelstein effect and torque we use the Kubo formula within
the constant relaxation time approximation. We only consider an Edelstein
effect linear in an electric field: $\delta S_{i}=\chi_{ij}E_{j}$, where
$\delta S_{i}$ is the induced spin, $E_{j}$ is the electric field, and
$\chi_{ij}$ is a response tensor. The $\cal{T}$-even and $\cal{T}$-odd
components are computed using the Kubo formula derived in Refs. Freimuth _et
al._ (2014); Li _et al._ (2015) and replacing the torque operator with the
spin operator,
$\displaystyle\chi^{\rm even}_{ij}$ $\displaystyle=$
$\displaystyle-\frac{e\hbar}{\pi}\sum_{\textbf{k},m,n}\frac{\text{Re}[\Bra{\psi_{\textbf{k}n}}\hat{S}_{i}\Ket{\psi_{\textbf{k}m}}\Bra{\psi_{\textbf{k}m}}\hat{v}_{j}\Ket{\psi_{\textbf{k}n}}]\Gamma^{2}}{(\left(\varepsilon_{F}-\varepsilon_{\mathbf{k}n})^{2}+\Gamma^{2})((\varepsilon_{F}-\varepsilon_{\mathbf{k}m}\right)^{2}+\Gamma^{2})},$
$\displaystyle\chi^{\rm odd}_{ij}$ $\displaystyle=$
$\displaystyle{2e\hbar}\sum_{\mathbf{k},n\neq m}^{\begin{subarray}{c}\\\ n\
\text{occ.}\\\ m\
\text{unocc.}\end{subarray}}\text{Im}[\Bra{\psi_{n\mathbf{k}}}\hat{S}_{i}\Ket{\psi_{m\mathbf{k}}}\Bra{\psi_{m\mathbf{k}}}\hat{v}_{j}\Ket{\psi_{n\mathbf{k}}}]$
(2)
$\displaystyle\times\frac{\Gamma^{2}-(\varepsilon_{\mathbf{k}n}-\varepsilon_{\mathbf{k}m})^{2}}{[(\varepsilon_{\mathbf{k}n}-\varepsilon_{\mathbf{k}m})^{2}+\Gamma^{2}]^{2}}.$
Here $\psi_{\textbf{k}n}$ is the Bloch function of band $n$, k is the Bloch
wave vector, $\varepsilon_{\mathbf{k}n}$ is the band energy, $\varepsilon_{F}$
is the Fermi energy, $\hat{v}_{j}$ is the velocity operator, $e>0$ is the
elementary charge, $\hat{S}_{i}$ is the spin operator, and $\Gamma$ is a
parameter that describes the strength of disorder, which is related to the
relaxation time $\tau=\hbar/2\Gamma$. To calculate the local Edelstein effect
on a given sublattice, a local spin operator is used instead.
In the limit $\Gamma\rightarrow 0$, Eq. (1) goes to the semiclassical
Boltzmann constant relaxation formula, which scales as $1/\Gamma$ whereas Eq.
(2) goes to the so-called intrinsic formula, which is $\Gamma$ independent and
can be understood in terms of Berry curvature in mixed spin-momentum space
Hanke _et al._ (2017).
Figure 1: The 3Q triangular antiferromagnet: (a) Sketch of the triangular
lattice with 3Q non-coplanar configuration of the magnetic moments. The green
atoms break the planar inversion symmetry. (b) Top view of the triangular
lattice. The hopping parameters of the black and red bounds are $t$ and
$t^{\prime}$, respectively. (c, f) Band structure for $t^{\prime}=t$ and
$t^{\prime}=t/2$, respectively. (e) In-plane spin texture in momentum space at
energy $\varepsilon=-5$ eV, corresponding to the red dashed line in panel (f).
In the absence of inversion symmetry breaking, $t^{\prime}=t$, the degenerate
bands display compensating spin texture in momentum space. When
$t^{\prime}\neq t$, the band degeneracy is lifted and the spin textures do not
longer compensate. Notice that a momentum-asymmetric out-of-plane spin texture
is also present (not shown). (d, g) $\cal{T}$-even (d) $\cal{T}$-odd (g)
contributions for $t^{\prime}=t/2$ and $\varepsilon=-4eV$ when rotating the
electric field direction in the (x,y) plane. We set $t=1$ eV, $\Gamma=0.1$ eV
and the exchange is $J=-2$ eV.
### III.1 A non-coplanar 3Q antiferromagnet
An example of a non-relativistic Edelstein effect in a non-collinear coplanar
antiferromagnet was recently given by Hayami et al. Hayami _et al._ (2020).
In this case, the coplanarity of the magnetic texture imposes the current-
driven spin density to be oriented perpendicular to the magnetic plane. Here,
we adopt a triangular antiferromagnet with a 3Q spin texture, as displayed in
Fig. 1(a). This magnetic texture can be stabilized in the presence of 4-spin
interaction Kurz _et al._ (2001); Kato _et al._ (2010) and hosts quantum
anomalous Hall effect Shindou and Nagaosa (2001); Martin and Batista (2008);
Ndiaye _et al._ (2019). The 3Q texture is also commonly observed in three-
dimensional materials such as $\gamma$-FeMnEndoh and Ishikawa (1971) and
pyrochlores Gardner _et al._ (2010). We use a simple tight-binding model with
a 3Q spin texture to illustrate the physical properties of such systems. The
model is given by
$\displaystyle
H=-\sum_{<ab>\alpha}t_{ab}c^{\dagger}_{a\alpha}c_{b\alpha}+J\sum_{a\alpha,\beta}({\bm{\sigma}}\cdot\mathbf{m}_{a})_{\alpha\beta}c^{\dagger}_{a\alpha}c_{a\beta}.$
(3)
Here, $c^{\dagger}$ and $c$ denote the creation and annihilation operators
respectively; $a,b$ denote the site index and $\alpha,\beta$ the spin index.
The first term is the nearest neighbor hopping term, with $t_{ab}$
representing the hopping magnitude. The second term represents the coupling of
the conduction electrons to the on-site magnetic moments. Here
$\mathbf{m}_{i}$ is the magnetic moment direction, $J$ is the exchange
parameter and ${\bm{\sigma}}$ is the vector of Pauli matrices. We only
consider nearest-neighbor hopping. To break the inversion symmetry we use two
different hopping magnitudes as shown in Fig. 1(b). This could be understood
as due to the presence of another atom illustrated in Fig. 1(a).
The band structures of the 3Q antiferromagnet are given in Figs. 1(c) and (f),
without and with inversion symmetry breaking. In the absence of inversion
symmetry breaking, the band structure is doubly degenerate. Breaking the
inversion symmetry lifts the band degeneracy [Fig. 1(f)] and results in a
spin-texture, shown in Fig. 1(e). This spin texture contains both symmetric
and anti-symmetric components, the latter giving rise to the current-driven
Edelstein effect and its torque. When the inversion symmetry is broken, we
observe a finite Edelstein effect as shown in Fig. 1(d) and (g). Several
features are worth noticing. First, because the magnetic texture of the 3Q
antiferromagnet spans the 3D space, the current-driven spin density possesses
all three components, $S_{x}$, $S_{y}$ and $S_{z}$, which strikingly contrasts
with the result of Ref. Hayami _et al._ (2020). Second, both $\cal{T}$-even
and $\cal{T}$-odd components contribute with similar magnitude. Finally, for
the set of parameters adopted in this calculation, i.e., the exchange and
hopping energies are of comparable magnitude ($\Delta$ =2t=4t’= 2 eV), we
obtain a nonequilibrium spin density of about 10-11 $\hbar$m/V. For the sake
of comparison, in a two-dimensional Rashba gas, the nonequilibrium spin
density is Edelstein (1990) $S^{R}_{\rm surf}/eE=(\alpha_{\rm
R}/\hbar^{2})(m_{0}/\pi\Gamma)$. Taking $\Gamma=0.1$ eV, $m_{0}$ being the
free electron mass, $\alpha_{\rm R}=10^{-9}$ eV$\cdot$m as the typical Rashba
strength expected in transition metal heterostructures Manchon _et al._
(2015), and $(3\mskip 3.0mu\textrm{\AA})^{2}$ as a unit cell area, the
Edelstein effect yields $\chi_{S}\sim 3.6\times 10^{-11}\mskip
3.0mu\hbar\text{m/V}$, which is in the same range as our calculations for the
two-dimensional 3Q system reported in Fig. 1.
### III.2 A centrosymmetric antiferromagnet: ${\rm Mn_{3}Sn}$
In antiferromagnets, and in general in more complex magnetic systems, the
magnetic dynamics is not determined by the global Edelstein effect, but rather
by the local Edelstein effect on each magnetic sublattice. Consequently, in
antiferromagnets, it is the local rather than the global inversion symmetry
breaking that is necessary for the existence of the Edelstein effect and the
current-induced torque Železný _et al._ (2017). An example of an
antiferromagnet with a global inversion symmetry and a local inversion
symmetry breaking is the well-known non-collinear antiferromagnet Mn3Sn
Tomiyoshi (1982); Nakatsuji _et al._ (2015). In this material, the global
Edelstein effect vanishes but the local Edelstein effect is allowed on each
sublattice, even in the absence of spin-orbit coupling.
The crystal and magnetic structure of Mn3Sn is given in Fig. 2(c). Mn3Sn has
six magnetic sublattices, which are composed of three pairs of sites with
equivalent moments connected by inversion symmetry. The inversion partners are
denoted by ′ in Fig. 2(b). Due to the inversion symmetry, the Edelstein effect
on the two inversion-connected sites must be opposite. The local Edelstein
effect tends to drive the system into a state where the magnetic moments of
the inversion-connected sites are not parallel and thus it acts against the
exchange. As such, it is unlikely to reverse the magnetic order but it can
excite different magnon modes. Leaving the rigorous analysis of the magnetic
dynamics to future studies, we emphasize that the global inversion symmetry
can be broken, for example, by an interface. Then the two sites with the same
moments are no longer connected by inversion symmetry and consequently can
experience the Edelstein effect of the same sign, enabling the electric
manipulation of the magnetic order.
We evaluate the Edelstein effect in Mn3Sn using ab-initio calculations with
and without spin-orbit coupling (see the Methods section for details of the
calculation setup). The result of the calculation for $\Gamma=0.01\mskip
3.0mu\text{eV}$ is shown in Fig. 2(a). We find substantial $\cal{T}$-even and
$\cal{T}$-odd Edelstein effect on all sublattices. Including the spin-orbit
coupling does not change the results substantially, similar to previous
calculations of the spin Hall effect in this material Zhang _et al._ (2018a).
Our calculations agree well with the symmetry analysis shown in the
Supplementary Materials. Notice that, again, the magnitude of the current-
driven spin density is rather large, corresponding to a Rashba strength of
$10^{-9}-10^{-10}$ eV$\cdot$m.
We note that current-induced switching of Mn3Sn has been experimentally
observed in Mn3Sn/non-magnetic metal heterostructures Tsai _et al._ (2020);
Krishnaswamy _et al._ (2022); Pal _et al._ (2022); Xie _et al._ (2022). The
switching has been attributed to the spin-Hall effect from the non-magnetic
metal layer and to spin-transfer torque and inter-grain spin-transfer torque,
however, it is possible that the non-relativistic Edelstein effect also
contributes.
Figure 2: (a),(b) Calculation of the $\cal{T}$-even (a) and $\cal{T}$-odd (b)
local Edelstein effect in Mn3Sn with (dashed lines) and without (solid lines)
spin-orbit coupling. The individual lines correspond to different tensor
components, see Eq. (LABEL:eq:Kubo_even)-(2). (c) The crystal structure and
magnetic structure of Mn3Sn.
### III.3 A non-centrosymmetric antiferromagnet: ${\rm LuFeO_{3}}$
Figure 3: (a) The crystal and magnetic structure of the hexagonal LuFeO3. (b)
LuFeO3 band structure without spin-orbit coupling. The color denotes the
$S_{x}$ projection. $X$ points represent opposite $k_{x}$ coordenates, $Z$
points represent opposite $k_{z}$ coordenates and $P$ points represent
opposite $k_{x},k_{y},k_{z}$ coordenates in the Brillouin zone. Asymmetric
-$odd$ in k- and symmetric -$even$\- spin splitting is labeled in the
correspondent $k$-path. (c) The spin texture of LuFeO3 at the Fermi surface
for Fermi level 0.45 eV below the top of the valence band. We plot the spin-
texture for two planes corresponding to $k_{z}$ = 0.25 Å-1 and $k_{z}$ = -0.25
Å-1. The center of the figure lies at the $\Gamma$ point. The arrows show the
spin and we use the color to highlight the $z$-component of the spin since it
would be hard to distinguish otherwise.
As an example of a real non-collinear antiferromagnet that can exhibit a
global non-relativistic Edelstein effect and torques, we consider the
hexagonal LuFeO3, a multiferroic with antiferromagnetic order. In bulk, LuFeO3
is typically orthorhombic. However, the hexagonal phase has been stabilized in
thin layers Disseler _et al._ (2015) and can also be stabilized in the bulk
Suresh _et al._ (2018). It has a non-collinear coplanar antiferromagnetic
structure with magnetic space group (MSG) #185.201 (P$6_{3}$c’m’) as presented
in Figure 3(a) Disseler _et al._ (2015); Suresh _et al._ (2018). The
inversion symmetry is broken in this material by the crystal structure, which
suggests the possibility of non-relativistic spin-torques. The system has a
small net moment $\sim 0.02\mskip 3.0mu\mu_{\textrm{B}}$ along the $z$
direction (weak ferromagnetism). This moment is of relativistic origin, and
thus in the absence of spin-orbit coupling the magnetic structure is perfectly
compensated. Apart from the magnetic order, hexagonal LuFeO3 also exhibits a
ferroelectric order that is present below $\sim 1000\mskip 3.0mu\textrm{K}$
and the material has attracted large attention for its multiferroic properties
and the possibility of magneto-electric coupling Das _et al._ (2014);
Disseler _et al._ (2015); Suresh _et al._ (2018).
The non-relativistic electronic structure is shown in Fig. 3(b). The material
is insulating; here we only show the valence bands, which are relevant to our
calculations. As can be also seen in Fig. 3(b) the bands are spin-split and
thus there is also a non-relativistic spin-texture, shown in Fig. 3(c) for two
cuts through the Brillouin zone. Due to the coplanarity of the magnetic order,
the spin-texture is symmetric for the $S_{x}$ and $S_{y}$ components and anti-
symmetric for the $S_{z}$ component. The $S_{z}$ component is non-zero but
very small.
For the calculation of the Edelstein effect, we move the Fermi level into the
valence band to simulate doping. Our symmetry analysis shown in the
Supplementary Materials shows that the Edelstein effect is allowed in LuFeO3
even with no spin-orbit coupling. Results of the calculation for
$\Gamma=0.01\mskip 3.0mu\text{eV}$ with and without spin-orbit coupling are
given in Fig. 4. We calculate both the global Edelstein effect and the local
one for all Fe sublattices. For brevity though, we only show here the result
for the global effect and for one sublattice. The full results are shown in
the Supplementary Materials.
Our calculations reveal a large non-relativistic global and local Edelstein
effect in good agreement with the symmetry analysis. Without spin-orbit
coupling for the global effect only the $\cal{T}$-odd component is allowed
[Fig. 4(b)]. With spin-orbit coupling even the global $\cal{T}$-even component
appears [Fig. 4(a)]. We find that the effect of spin-orbit coupling is quite
small for the $\cal{T}$-odd component, but fairly large for the $\cal{T}$-even
component. For the local effect [Fig. 4(c,d)], both $\cal{T}$-even and
$\cal{T}$-odd components are allowed.
An important remark is in order. The $\cal{T}$-even component has to vanish
within the gap since it is a Fermi surface property [see Eq.
(LABEL:eq:Kubo_even)]. The global $\cal{T}$-odd component vanishes within the
gap as well [Fig. 4(b)]. However, we find that the local $\cal{T}$-odd
components on the Fe atoms are non-zero within the gap, as shown in Fig. 4(d)
and in the Supplementary Materials. Only the $xz$ and $yz$ components are non-
zero within the gap, reaching a constant value. Such a result is intriguing as
within the gap there is no Ohmic conduction and thus heat dissipation is
absent. This could consequently allow for electric field control of magnetic
order in the absence of Ohmic dissipation. The existence of spin-orbit torque
in an insulator was previously studied in topological materials Hanke _et
al._ (2017); Tang and Cheng (2023). Our results are similar except that in the
case of LuFeO3 the origin of the torque is non-relativistic, due to the
coexistence of the non-collinear magnetic order with inversion symmetry
breaking. We point out that the torque is not quantized, contrary to the
quantized magnetoelectric effect in topological insulators Qi _et al._
(2008), and therefore unlikely to be of topological origin. We are also not
aware of any topological properties of LuFeO3. The $\cal{T}$-odd torque is
governed by the Berry curvature in mixed spin-momentum space and involves
electrically-driven interband transitions, resulting in a finite (but not
quantized) value in the gap.
We note that in metals the torques induced by an electric field are
accompanied by an electric current and thus often referred to as “current-
induced” torques. However, even in metals, the torques are in fact due to the
electric field, rather than to the current flow, although the torque cannot
exist without Ohmic conduction. In non-centrosymmetric insulating magnets
though, Ohmic conduction is suppressed while the electrically-driven torque
remains sizable as demonstrated in LuFeO3. This opens promising perspectives
for the dissipation-free electrical control of magnetization.
Figure 4: Edelstein effect calculation in LuFeO3 with (dashed lines) and
without spin-orbit coupling (solid lines). Here zero energy corresponds to the
top of the valence band. (a) $\cal{T}$-even component of the total Edelstein
effect. (b) $\cal{T}$-odd component of the total Edelstein effect. (c)
$\cal{T}$-even component of the local Edelstein effect. (d) $\cal{T}$-odd
component of the local Edelstein effect.
### III.4 Non-centrosymmetric heterostructures
The examples we have discussed so far all have inversion symmetry (globally or
locally) broken in the bulk of their crystal structure. Such a constraint
however severely restricts the Edelstein effect to the materials listed in the
Supplemental Materials. For this reason, we propose to exploit the broken
inversion symmetry taking place at the interface between the non-collinear
antiferromagnet and an adjacent metal. Such heterostructures are commonly
utilized for spin-orbit torque, where the ferro- or antiferromagnet is
typically interfaced with a heavy element metal such as platinum Manchon _et
al._ (2019). This simple but instrumental configuration allows for observing
the spin-orbit torque in a wide variety of systems and enables interfacial
engineering of the spin-orbit torque properties.
The same concept can be applied to the non-relativistic Edelstein effect. When
a non-collinear magnetic material with inversion symmetry is interfaced with a
different material, the broken inversion symmetry can result in a non-
relativistic Edelstein effect, which in turn generates a torque on the
magnetic order. We illustrate this concept using the example of the well-known
non-collinear antiferromagnet Mn3Ir whose crystal and magnetic structures are
displayed in Fig. 5(a). In this material, each magnetic site is an inversion
center and thus no Edelstein effect is allowed in the bulk. To break the
inversion symmetry we consider a thin layer of Mn3Ir interfaced with a thin
layer of a non-magnetic material. When Mn3Ir is grown along the [001]
direction, the non-relativistic Edelstein effect in such a heterostructure is
only allowed for an electric field along the [001] direction. In such a case
no electric current can flow, however. Thus we instead consider Mn3Ir grown
along the [111] direction. For this orientation, the symmetry is lowered and
the Edelstein effect is allowed for an electric field oriented along the
interface. We consider a structure composed of 12 atomic layers of the Mn3Ir,
as shown in Fig. 5(b). The individual atomic layers are shown in Fig. 5(c).
We utilize a simple tight-binding model that is not meant to give quantitative
predictions but rather to confirm that the effect can exist and illustrate its
basic properties. The model is analogous to the one we have used for the 3Q
antiferromagnet. It is composed of s-electrons on each site with nearest
neighbor hopping and exchange coupling to the atomic magnetic moments. Similar
models have been utilized to demonstrate other properties of the Mn3Ir and
similar antiferromagnets such as the non-relativistic spin currents Zhang _et
al._ (2018a) or the anomalous Hall effect Chen _et al._ (2014). We do not
include any spin-orbit coupling in the model. The Hamiltonian is in Eq. (3),
where we consider nearest neighbor hopping $t=1\ \text{eV}$ and magnetic
exchange $J=1.7\ \text{eV}$ on Mn atoms (gray and gold atoms representing non-
magnetic layer and Ir atoms, respectively, are not endowed with magnetic
moment).
Figure 5: (a) Mn3Ir bulk structure. (b) Side view of the Mn3Ir [111] bilayer.
Here we do not show the magnetic moments. (c) Top-down view of a single atomic
layer in the Mn3Ir [111] bilayer. Here $A$, $B$ and $C$ denote the three
different sublattices. Note that we have constructed the system in such a way
that there are 6 Mn atoms in each atomic layer, however, the two atoms for
each sublattice are connected by translation and thus there are only three
non-equivalent sublattices. (d) Edelstein effect calculation in the Mn3Ir
[111] for each sublattice as a function of the atomic layer. The top row plots
show the $\cal{T}$-even component and the bottom plots show the $\cal{T}$-odd
component.
The calculated Edelstein effect is shown in Fig. 5(c) for each sublattice and
atomic layer. For this calculation we have used $\Gamma=0.01\mskip
3.0mu\text{eV}$ and $\varepsilon_{F}=0\mskip 3.0mu\text{eV}$. Our calculations
are fully in agreement with the symmetry analysis, shown in the Supplementary
Materials. Both $\cal{T}$-even and $\cal{T}$-odd components are present. We
find the largest effect close to the interfaces, although the current-driven
spin density remains sizable in the center of the Mn3Ir layer. A large effect
is found both at the interface with the non-magnetic layer and at the top
surface, which illustrates that the presence of another layer is in principle
not necessary.
In Figure 5, we only give the result of the calculation for the tensor
components that correspond to an in-plane electric field. Interestingly, the
out-of-plane components do not vanish even though no current can flow in the
out-of-plane direction, similarly to the case of the LuFeO3. In this case,
however, since the system is metallic, the out-of-plane electric field is
screened and thus the effect is hard to observe in practice.
## IV Discussion and conclusion
The torque induced by the non-relativistic Edelstein effect shares important
similarities with the conventional spin-orbit torque. Both torques are
electrically driven self-induced torques that necessitate inversion symmetry
breaking. The key difference though is that the torque due to the non-
relativistic Edelstein effect does not originate from spin-orbit coupling, but
rather from the non-collinear magnetic order. As a consequence, the
microscopic origin of these two torques is quite distinct. In the non-
relativistic limit, spin is conserved and the torque is directly associated
with a spin current: the torque corresponds to spin sources Ghosh _et al._
(2022) and can be understood as a local transfer of spin angular momentum
within the magnetic unit cell.
In the present work, we have computed the non-relativistic Edelstein effect in
four different systems, all displaying inversion symmetry breaking, either in
the magnetic unit cell or locally. It is quite remarkable that all the
examples discussed here display a sizable electrically induced spin density,
in spite of the absence of spin-orbit coupling. For the sake of comparison, in
our previous calculations of the relativistic Edelstein effect in a collinear
antiferromagnet Mn2Au, we found a magnitude of $\chi_{S}\sim 4.3\times
10^{-11}\mskip 3.0mu\hbar\text{m/V}$ for $\Gamma=0.01\mskip 3.0mu\text{eV}$
Železný _et al._ (2017), corresponding to a Rashba strength of $10^{-9}$
eV$\cdot$m, similarly to the magnitude reported in our realistic simulations
on LuFeO3 and Mn3Sn. Further systematic studies as necessary to determine the
conditions for a maximal non-relativistic Edelstein effect.
A central feature of the non-relativistic nature of the torque is its
dependence on the magnetic order. In most magnetic systems, the magnetic
exchange is much larger than any other magnetic interactions or torques acting
on the system. Hence, during the dynamics of the magnetic order, the angles
between the individual magnetic moments stay approximately unchanged.
Therefore, the dynamics of the magnetic order are described by an overall
rotation of all magnetic moments and a small canting. In the non-relativistic
limit (ignoring the small canting) the rotated states are connected by a spin
rotation and, consequently, the corresponding torques must also be transformed
by this spin rotation. Specifically, any torque acting on magnetic moment
$\mathbf{M}_{i}$ can be written as
$\mathbf{T}_{i}=\mathbf{M}_{i}\times\mathbf{B}_{i}$, where $\mathbf{B}_{i}$ is
an effective magnetic field. When the magnetic moments are rotated by rotation
$R$ then the torque reads $\mathbf{T}_{i}=R\mathbf{M}_{i}\times
R\mathbf{B}_{i}$. This is quite distinct from the conventional spin-orbit
torque for which the two most important terms are the field-like torque, in
which $\mathbf{B}_{i}$ is independent of $\mathbf{M}$, and the anti-damping
torque, in which $\mathbf{B}_{i}\sim\mathbf{M}_{i}\times\mathbf{p}$,
$\mathbf{p}$ being some constant direction Manchon _et al._ (2019).
Because of the dependence of the non-relativistic torque on the magnetic
order, it may be difficult to realize reversible switching since there can be
no magnetic configuration for which the effective field $\mathbf{B}_{i}$
vanishes. This might not be such a limitation in practice, however, since some
spin-orbit coupling is always present, which may enable deterministic
switching even in cases where the non-relativistic torque is dominant.
Furthermore, deterministic switching could be achieved by using field-assisted
switching or precise pulse timing. In the presence of antiferromagnetic domain
walls, the non-relativistic torque could provide an additional source of spin
current and therefore enhance or quench the domain wall mobility, depending on
the wall configuration. In fact, the very dependence of the torque on the
magnetic ordering makes the interplay between the flowing electrons and the
magnetic order particularly rich and, as such, the non-relativistic torque is
well adapted to excite magnetic modes and self-oscillations which, in
antiferromagnets, are particularly appealing for THz applications.
## V acknowledgments
We acknowledge the Grant Agency of the Czech Republic Grant No. 22-21974S. A.
M. acknowledges support from the Excellence Initiative of Aix-Marseille
Université - A*Midex, a French ”Investissements d’Avenir” program.
## VI Methods
The DFT calculations use the VASP code Kresse and Furthmüller (1996) and we
use the Wannier90 code Pizzi _et al._ (2020) to construct the Wannier
Hamiltonian that serves as an input to the linear response calculations. For
LuFeO3 we use 9x9x3 k-point mesh and $520\mskip 3.0mu\text{eV}$ cutoff energy.
For the Wannierization we use the $s$ and $d$ orbitals for Fe, $s$ and $p$
orbitals for O and $\varepsilon_{F}\pm 3\mskip 3.0mu\text{eV}$ frozen energy
window. For Mn3Sn we use 11x11x11 k-point mesh and set the cutoff energy to
$520\mskip 3.0mu\text{eV}$. For the Wannierization we use the $s$,$p$, and $d$
orbitals for the Mn atoms, $s$ and $p$ orbitals for the Sn atoms and we set
the frozen energy window to $\varepsilon_{F}\pm 2\mskip 3.0mu\text{eV}$.
For the linear response calculations we use the Linres code Železný . This
code uses the Wannier or tight-binding Hamiltonian as an input. This
Hamiltonian is then Fourier transformed to a dense mesh in the reciprocal
space, which is used for evaluating the Kubo formulas as described in Ref.
Zelezný _et al._ (2017).
## References
* Nagaosa _et al._ (2010) N. Nagaosa, J. Sinova, S. Onoda, A. H. MacDonald, and N. P. Ong, Review of Modern Physics 82, 1539 (2010).
* Ritzinger and Vyborny (2022) P. Ritzinger and K. Vyborny, arXiv:2212.03700v2 (2022).
* Sinova _et al._ (2015) J. Sinova, S. O. Valenzuela, J. Wunderlich, C. H. Back, and T. Jungwirth, Review of Modern Physics 87, 1213 (2015).
* Soumyanarayanan _et al._ (2016) A. Soumyanarayanan, N. Reyren, A. Fert, and C. Panagopoulos, Nature 539, 509 (2016), arXiv:1611.09521 .
* Manchon _et al._ (2019) A. Manchon, J. Zelezný, M. Miron, T. Jungwirth, J. Sinova, A. Thiaville, K. Garello, and P. Gambardella, Review of Modern Physics 91, 035004 (2019).
* Sunko _et al._ (2017) V. Sunko, H. Rosner, P. Kushwaha, S. Khim, F. Mazzola, L. Bawden, O. J. Clark, J. M. Riley, D. Kasinathan, M. W. Haverkort, T. K. Kim, M. Hoesch, J. Fujii, I. Vobornik, A. P. Mackenzie, and P. D. C. King, Nature 549, 492 (2017), arXiv:1708.03887 .
* Železný _et al._ (2021) J. Železný, Z. Fang, K. Olejník, J. Patchett, F. Gerhard, C. Gould, L. W. Molenkamp, C. Gomez-Olivella, J. Zemen, T. Tichý, T. Jungwirth, and C. Ciccarelli, Physical Review B 104, 054429 (2021), arXiv:2102.12838 .
* Pekar and Rashba (1965) S. Pekar and E. I. Rashba, Soviet Physics JETP 20, 1295 (1965).
* Ramazashvili (2009) R. Ramazashvili, Physical Review B 79, 184432 (2009), arXiv:arXiv:0809.1893v3 .
* Shindou and Nagaosa (2001) R. Shindou and N. Nagaosa, Physical Review Letters 87, 116801 (2001).
* Martin and Batista (2008) I. Martin and C. D. Batista, Physical Review Letters 101, 156402 (2008), arXiv:0807.3552 .
* Ndiaye _et al._ (2019) P. B. Ndiaye, A. Abbout, V. M. L. D. P. Goli, and A. Manchon, Physical Review B 100, 144440 (2019).
* Zhang _et al._ (2018a) Y. Zhang, J. Zelezny, Y. Sun, J. V. D. Brink, and B. Yan, New Journal of Physics 20, 073028 (2018a).
* Zelezný _et al._ (2017) J. Zelezný, Y. Zhang, C. Felser, and B. Yan, Physical Review Letters 119, 187204 (2017).
* Kimata _et al._ (2019) M. Kimata, H. Chen, K. Kondou, S. Sugimoto, P. K. Muduli, M. Ikhlas, Y. Omori, T. Tomita, A. H. Macdonald, S. Nakatsuji, and Y. Otani, Nature 565, 627 (2019).
* F.T. Vasko (1979) F.T. Vasko, Pis’ma Zh. Eksp. Teor. Fiz 30, 574 (1979).
* Edelstein (1990) V. M. Edelstein, Solid State Communications 73, 233 (1990).
* Hayami _et al._ (2020) S. Hayami, Y. Yanagi, and H. Kusunose, Physical Review B 101, 220403 (2020), arXiv:2001.05630 .
* Ghosh _et al._ (2022) S. Ghosh, A. Manchon, and J. Železný, Physical Review Letters 128, 097702 (2022).
* Bihlmayer _et al._ (2022) G. Bihlmayer, P. Noël, D. V. Vyalikh, E. V. Chulkov, and A. Manchon, Nature Reviews Physics (2022), 10.1038/s42254-022-00490-y.
* Yuan _et al._ (2021) L. D. Yuan, Z. Wang, J. W. Luo, and A. Zunger, Physical Review Materials 5, 014409 (2021), arXiv:2008.08532 .
* Hayami _et al._ (2019) S. Hayami, Y. Yanagi, and H. Kusunose, Journal of the Physical Sociecty of Japan 88, 123702 (2019).
* Yuan _et al._ (2020) L.-d. Yuan, Z. Wang, J.-w. Luo, E. I. Rashba, and A. Zunger, Physical Review B 102, 014422 (2020).
* Šmejkal _et al._ (2022) L. Šmejkal, J. Sinova, and T. Jungwirth, Phys. Rev. X 12, 040501 (2022).
* Egorov _et al._ (2021) S. A. Egorov, D. B. Litvin, and R. A. Evarestov, The Journal of Physical Chemistry C 125, 16147 (2021).
* Brinkman and Elliott (1966) W. F. Brinkman and R. J. Elliott, Proceedings of the Royal Society of London Series A 294, 343 (1966).
* Litvin and Opechowski (1974) D. B. Litvin and W. Opechowski, Physica 76, 538 (1974).
* Krempaský _et al._ (2023) J. Krempaský, L. Šmejkal, S. W. D’Souza, M. Hajlaoui, G. Springholz, K. Uhlířová, F. Alarab, P. C. Constantinou, V. Strokov, D. Usanov, W. R. Pudelko, R. González-Hernández, A. B. Hellenes, Z. Jansa, H. Reichlová, Z. Šobáň, R. D. G. Betancourt, P. Wadley, J. Sinova, D. Kriegner, J. Minár, J. H. Dil, and T. Jungwirth, (2023).
* Šmejkal _et al._ (2020) L. Šmejkal, R. González-hernández, T. Jungwirth, and J. Sinova, Science Advances 6, eaaz8809 (2020).
* Železný _et al._ (2017) J. Železný, H. Gao, A. Manchon, F. Freimuth, Y. Mokrousov, J. Zemen, J. Mašek, J. Sinova, and T. Jungwirth, Physical Review B 95, 014403 (2017).
* Bonbien and Manchon (2020) V. Bonbien and A. Manchon, Physical Review B 102, 085113 (2020).
* Kurebayashi _et al._ (2014) H. Kurebayashi, J. Sinova, D. Fang, A. Irvine, T. D. Skinner, J. Wunderlich, V. Novák, R. P. Campion, B. L. Gallagher, E. K. Vehstedt, L. P. Zârbo, K. Výborný, A. J. Ferguson, and T. Jungwirth, Nature Nanotechnology 9, 211 (2014).
* Li _et al._ (2015) H. Li, H. Gao, L. P. Zârbo, K. Výborný, X. Wang, I. Garate, F. Dogan, A. Cejchan, J. Sinova, T. Jungwirth, and A. Manchon, Physical Review B 91, 134402 (2015).
* Hanke _et al._ (2017) J.-P. Hanke, F. Freimuth, C. Niu, S. Blügel, and Y. Mokrousov, Nat Commun 8, 1479 (2017), arXiv:1701.08050 .
* Železný _et al._ (2014) J. Železný, H. Gao, K. Výborný, J. Zemen, J. Mašek, A. Manchon, J. Wunderlich, J. Sinova, and T. Jungwirth, Physical Review Letters 113, 157201 (2014).
* Zhang _et al._ (2014) X. Zhang, Q. Liu, J.-w. Luo, A. J. Freeman, and A. Zunger, Nature Physics 10, 387 (2014).
* Zhang _et al._ (2018b) Y. Zhang, S. Okamoto, and D. Xiao, Physical Review B 98, 035424 (2018b).
* Zelezný (2017) J. Zelezný, “symcode,” (2017).
* (39) “Supplemental Material,” .
* Freimuth _et al._ (2014) F. Freimuth, S. Blügel, and Y. Mokrousov, Physical Review B 90, 174423 (2014).
* Kurz _et al._ (2001) P. Kurz, G. Bihlmayer, K. Hirai, and S. Blügel, Physical Review Letters 86, 1106 (2001).
* Kato _et al._ (2010) Y. Kato, I. Martin, and C. D. Batista, Physical Review Letters 105, 266405 (2010).
* Endoh and Ishikawa (1971) Y. Endoh and Y. Ishikawa, Journal of the Physical Society of Japan 30, 1614 (1971).
* Gardner _et al._ (2010) J. S. Gardner, M. J. Gingras, and J. E. Greedan, Reviews of Modern Physics 82, 53 (2010), arXiv:0906.3661 .
* Manchon _et al._ (2015) A. Manchon, H. C. Koo, J. Nitta, S. M. Frolov, and R. A. Duine, Nature Materials 14, 871 (2015).
* Tomiyoshi (1982) S. Tomiyoshi, Journal of the Physical Society of Japan 51, 803 (1982).
* Nakatsuji _et al._ (2015) S. Nakatsuji, N. Kiyohara, and T. Higo, Nature 527, 212 (2015).
* Tsai _et al._ (2020) H. Tsai, T. Higo, K. Kondou, T. Nomoto, A. Sakai, A. Kobayashi, T. Nakano, K. Yakushiji, R. Arita, S. Miwa, Y. C. Otani, and S. Nakatsuji, Nature 580, 608 (2020).
* Krishnaswamy _et al._ (2022) G. K. Krishnaswamy, G. Sala, B. Jacot, C.-H. Lambert, R. Schlitz, M. D. Rossell, P. Nöel, and P. Gambardella, Physical Review Applied 18, 024064 (2022).
* Pal _et al._ (2022) B. Pal, B. K. Hazra, B. Göbel, J.-C. Jeon, A. K. Pandeya, A. Chakraborty, O. Busch, A. K. Srivastava, H. Deniz, J. M. Taylor, H. Meyerheim, I. Mertig, S.-H. Yang, and S. S. P. Parkin, Science Advances 8, eabo5930 (2022).
* Xie _et al._ (2022) H. Xie, X. Chen, Q. Zhang, Z. Mu, X. Zhang, B. Yan, and Y. Wu, Nature Communications 13, 5744 (2022).
* Disseler _et al._ (2015) S. M. Disseler, J. A. Borchers, C. M. Brooks, J. A. Mundy, J. A. Moyer, D. A. Hillsberry, E. L. Thies, D. A. Tenne, J. Heron, M. E. Holtz, J. D. Clarkson, G. M. Stiehl, P. Schiffer, D. A. Muller, D. G. Schlom, and W. D. Ratcliff, Phys. Rev. Lett. 114, 217602 (2015).
* Suresh _et al._ (2018) P. Suresh, K. Vijaya Laxmi, A. K. Bera, S. M. Yusuf, B. L. Chittari, J. Jung, and P. S. Anil Kumar, Phys. Rev. B 97, 184419 (2018).
* Das _et al._ (2014) H. Das, A. L. Wysocki, Y. Geng, W. Wu, and C. J. Fennie, Nature Communications 5, 2998 (2014).
* Tang and Cheng (2023) J. Tang and R. Cheng, arXiv preprint arXiv:2303.06181 (2023).
* Qi _et al._ (2008) X.-L. Qi, T. L. Hughes, and S.-C. Zhang, Physical Review B 78, 195424 (2008).
* Chen _et al._ (2014) H. Chen, Q. Niu, and A. H. Macdonald, Physical Review Letters 112, 017205 (2014).
* Kresse and Furthmüller (1996) G. Kresse and J. Furthmüller, Phys. Rev. B 54, 11169 (1996).
* Pizzi _et al._ (2020) G. Pizzi, V. Vitale, R. Arita, S. Blügel, F. Freimuth, G. Géranton, M. Gibertini, D. Gresch, C. Johnson, T. Koretsune, J. Ibanez-Azpiroz, H. Lee, J. M. Lihm, D. Marchand, A. Marrazzo, Y. Mokrousov, J. I. Mustafa, Y. Nohara, Y. Nomura, L. Paulatto, S. Poncé, T. Ponweiser, J. Qiao, F. Thöle, S. S. Tsirkin, M. Wierzbowska, N. Marzari, D. Vanderbilt, I. Souza, A. A. Mostofi, and J. R. Yates, Journal of Physics Condensed Matter 32, 165902 (2020), arXiv:1907.09788 .
* (60) J. Železný, https://bitbucket.org/zeleznyj/wannier-linear-response.
|
# On the Ehrhart Theory of Generalized Symmetric Edge Polytopes
Robert Davis Department of Mathematics, Colgate University, Hamilton, NY, USA
<EMAIL_ADDRESS>, Akihiro Higashitani Department of Pure and Applied
Mathematics, Graduate School of Information Science and Technology, Osaka
University, Osaka, Japan<EMAIL_ADDRESS>and Hidefumi Ohsugi
Department of Mathematical Sciences, School of Science, Kwansei Gakuin
University, Sanda, Hyogo 669-1337, Japan<EMAIL_ADDRESS>
###### Abstract.
The symmetric edge polytope (SEP) of a (finite, undirected) graph is a
centrally symmetric lattice polytope whose vertices are defined by the edges
of the graph. SEPs have been studied extensively in the past twenty years.
Recently, Tóthmérész and, independently, D’Alí, Juhnke-Kubitzke, and Koch
generalized the definition of an SEP to regular matroids, as these are the
matroids that can be characterized by totally unimodular matrices. Generalized
SEPs are known to have symmetric Ehrhart $h^{*}$-polynomials, and Ohsugi and
Tsuchiya conjectured that (ordinary) SEPs have nonnegative $\gamma$-vectors.
In this article, we use combinatorial and Gröbner basis techniques to extend
additional known properties of SEPs to generalized SEPs. Along the way, we
show that generalized SEPs are not necessarily $\gamma$-nonnegative by
providing explicit examples. We prove these polytopes to be “nearly”
$\gamma$-nonnegative in the sense that, by deleting exactly two elements from
the matroid, one obtains SEPs for graphs that are $\gamma$-nonnegative. This
provides further evidence that Ohsugi and Tsuchiya’s conjecture holds in the
ordinary case.
## 1\. Introduction
A _lattice polytope_ $P$ in $\mathbb{R}^{n}$ is the convex hull of finitely
many points in $\mathbb{Z}^{n}$. That is, $P$ is a lattice polytope if and
only if there are some $v_{1},\dots,v_{k}\in\mathbb{Z}^{n}$ such that
$P=\operatorname{conv}\\{v_{1},\dots,v_{k}\\}=\left\\{\sum_{i=1}^{k}\lambda_{i}v_{i}\mid\lambda_{1},\dots,\lambda_{k}\geq
0,\,\sum_{j=1}^{k}\lambda_{j}=1\right\\}.$
The main objects of study in this paper will be lattice polytopes arising from
(finite, undirected) graphs and matroids. First, given an undirected graph
$G=([n],E)$ where $[n]=\\{1,\dots,n\\}$, the _symmetric edge polytope_ (or
_SEP_) associated to $G$ is
$\Sigma(G)=\operatorname{conv}\\{\pm(e_{i}-e_{j})\in\mathbb{R}^{n}\mid ij\in
E\\}.$
This polytope was introduced in [11] and has been the object of much study for
its algebraic and combinatorial properties [6, 9, 12, 13], as well as its
applications to problems arising from engineering [3, 4, 5].
SEPs have recently been generalized to regular matroids in two ways. Suppose
$A$ is a totally unimodular matrix representing a matroid $\mathcal{M}$.
Tóthmérész [17], working within the context of oriented matroids, defined the
_root polytope_ of $\mathcal{M}$ to be the convex hull of the columns of $A$.
Independently, D’Alì, Juhnke-Kubitzke, and Koch [7] defined the _generalized
symmetric edge polytope_ (or _generalized SEP_) associated to $\mathcal{M}$ as
$\Sigma(\mathcal{M})=\operatorname{conv}\\{\pm u\mid u\text{ is a column of
}A\\}.$
Thus, $\Sigma(\mathcal{M})$ may be considered the root polytope of the matroid
represented by the matrix $[A\mid-A]$, since this is a totally unimodular
matrix. Because we will not be working with oriented matroids, and because are
focused on the polytopes $\Sigma(\mathcal{M})$ in this work, we will follow
the terminology and notation in [7].
Although the definition of $\Sigma(\mathcal{M})$ depends on choice of $A$, the
polytope is well-defined up to unimodular equivalence ([17, Proposition 3.2]
and [7, Theorem 3.2]). In [7], the authors proved that many properties held by
ordinary SEPs also hold for generalized SEPs. Further, this more general
setting allowed the authors to show that two SEPs are unimodularly equivalent
if and only if the underlying matroids are isomorphic [7, Theorem 4.6]. In
addition to these aspects of SEPs, much attention has been given to their
Ehrhart theory, which we briefly introduce here.
The function $m\mapsto|mP\cap\mathbb{Z}^{n}|$, defined on the positive
integers, agrees with a polynomial $L_{P}(m)$ of degree $\dim(P)$, called the
_Ehrhart polynomial_ of $P$. Consequently, the _Ehrhart series_ of $P$, which
is the (ordinary) generating function for the sequence $\\{L_{P}(m)\\}_{m\geq
0}$, may be written as the rational function
$E(P;t)=\sum_{m\geq
0}L_{P}(m)t^{m}=\frac{h_{0}^{*}+h_{1}^{*}t+\cdots+h_{d}^{*}t^{d}}{(1-t)^{\dim(P)+1}}$
for some $d\leq\dim(P)$. We denote the numerator by $h^{*}(P;t)$ and refer to
it as the _$h^{*}$ -polynomial_ of $P$. The _$h^{*}$ -vector_ of $P$,
$h^{*}(P)$, is the list of coefficients $h^{*}(P)=(h^{*}_{0},\dots,h^{*}_{d})$
of $h^{*}(P;t)$.
The $h^{*}$-vectors of polytopes have been the focus of much study in recent
decades, as it lies within the intersection of polyhedral geometry,
commutative algebra, algebraic geometry, combinatorics, and much more (see,
e.g., [15, 16]). For example, when the $h^{*}$-vector is _symmetric_ , meaning
$h^{*}_{i}=h^{*}_{d-i}$ for each $i=0,\dots,d$, then the associated semigroup
algebra
$K[P]=K[x^{v}z^{m}\mid v\in mP\cap\mathbb{Z}^{n}]\subseteq
K[x_{1}^{\pm},\dots,x_{d}^{\pm},z],$
over a field $K$, is Gorenstein. In fact, the converse holds as well.
Determining when $K[P]$ is Gorenstein can be detected entirely using
polyhedral geometry, so, from this perspective, there is a complete
characterization of when $h^{*}(P)$ is symmetric.
On the other hand, a more mysterious property held by some $h^{*}$-vectors is
unimodality. We call $h^{*}(P)$ _unimodal_ if there is some index $0\leq r\leq
d$ for which $h^{*}_{0}\leq\cdots\leq h^{*}_{r}$ and $h^{*}_{r}\geq\cdots\geq
h^{*}_{d}$. Unlike symmetry, there is no complete characterization of when an
$h^{*}$-vector is unimodal. There are various sets of sufficient conditions
that ensure a polytope has a unimodal $h^{*}$-vector and various sets of
necessary conditions [2], but no set of conditions capturing both
simultaneously.
When an $h^{*}$-vector is already known to be symmetric, new avenues for
proving $h^{*}$-unimodality appear. For instance, in this setting, the
$h^{*}$-polynomial may be expressed
$\sum_{i=0}^{d}h^{*}_{i}t^{i}=\sum_{i=0}^{\lfloor
d/2\rfloor}\gamma_{i}t^{i}(1+t)^{d-2i}$
for certain choices of $\gamma_{i}\in\mathbb{Q}$. The polynomial on the right
is the _$\gamma$ -polynomial_ of $P$ and is often denoted $\gamma(P;t)$, and
the vector $\gamma(P)=(\gamma_{0},\dots,\gamma_{\lfloor d/2\rfloor})$ is the
_$\gamma$ -vector_ of $P$. Note that if $\gamma_{i}\geq 0$ for each $i$, then
$h^{*}_{i}\geq 0$ for each $i$ as well. This is not necessary, though: the
$h^{*}$-vector of $(1,1,1)$, realized by the lattice triangle with vertices
$(1,0)$, $(0,1)$, and $(-1,-1)$, is unimodal but has a $\gamma$-vector of
$(1,-1)$. Nevertheless, $\gamma$-vectors remain the topic of intense study,
due in part to wide-open conjectures surrounding their nonnegativity. The most
famous of these is _Gal’s Conjecture_ : that the $h$-vector of a flag homology
sphere is $\gamma$-nonnegative. To an Ehrhart theorist, the conjecture states
that the $h^{*}$-polynomial of a lattice polytope with a regular, unimodular,
flag triangulation is $\gamma$-nonnegative.
In this article, we use combinatorial and Gröbner basis techniques to extend
additional known properties of SEPs to generalized SEPs. Along the way, we
show that generalized SEPs are not $\gamma$-nonnegative by providing an
infinite class of explicit examples (Theorem 4.2). Computational evidence
strongly suggests that these examples are “nearly” $\gamma$-nonnegative in the
sense that, by deleting exactly two elements from the matroid, we obtain SEPs
for graphs that appear to be $\gamma$-nonnegative. We present and prove a
formula for their $h^{*}$-polynomials (Theorem 4.3), deduce the normalized
volumes of these SEPs (Corollary 4.6), and compute their $\gamma$-polynomials
(Theorem 4.4). This provides further evidence that Ohsugi and Tsuchiya’s
conjecture holds in the ordinary case.
A brief structure of the article is as follows. In Section 2 we discuss the
necessary background for matroids and the relationships among Ehrhart theory,
triangulations of polytopes, and toric ideals. Section 3 discusses how various
important results for ordinary SEPs extend to generalized SEPs. In Section 4
we present explicit examples of generalized SEPs whose $h^{*}$-polynomials are
not $\gamma$-nonnegative (Theorem 4.2) and examine the consequences of making
minor variations to them. We also state two of the main results of this
article: Theorem 4.3 and Theorem 4.4. In Section 5 we both prepare for and
present the full proof of Theorem 4.3, which follows a combinatorial argument.
Finally, in Section 6, we use a generating function argument to prove Theorem
4.4.
## 2\. Background
To keep this article as self-contained as possible, this section provides the
necessary background on matroids, Ehrhart theory, and generalized SEPs. For a
comprehensive reference on the first two topics, we refer the reader to [14]
for matroids and [1] for Ehrhart theory. Throughout this article we use the
convention of writing $A\cup x$ and $A-x$ for the union or deletion,
respectively, of $A$ by a one-element set $\\{x\\}$.
### 2.1. Matroids
A _matroid_ $\mathcal{M}$ consists of a pair $(E,\mathcal{I})$ where $E$ is
finite and $\mathcal{I}\subseteq 2^{E}$ satisfies the following three
properties:
1. (i)
$\emptyset\in\mathcal{I}$;
2. (ii)
if $I\in\mathcal{I}$ and $J\subseteq I$, then $J\in\mathcal{I}$; and
3. (iii)
if $I,J\in\mathcal{I}$ and $|I|<|J|$, then there is some $x\in J-I$ such that
$I\cup x\in\mathcal{I}$.
The set $E$ is called the _ground set_ of $\mathcal{M}$ and the elements of
$\mathcal{I}$ are called the _independent sets_ of $\mathcal{M}$.
Matroids are useful for unifying various notions of “independence” in
mathematics. For example, if $G$ is a (finite) graph, then there is a
corresponding matroid $M(G)$ whose ground set is $E(G)$, the edge set of $G$,
and whose independent sets are the sets of edges that form acyclic subgraphs
of $G$. This matroid is called the _cycle matroid_ of $G$, and any matroid
that is isomorphic to $M(G)$ for some $G$ is a _graphic matroid_.
Another fundamental example of a matroid uses the columns of a $k\times n$
matrix $A$ as the ground set. Here, the independent sets are the columns of
$A$ that form linearly independent sets. This matroid, denoted $M(A)$, is the
_vector matroid_ of $A$, and any matroid isomorphic to $M(A)$ for some matrix
$A$ is called a _representable matroid_. More precisely, for a field $K$, the
matrix $A$ is called _$K$ -representable_ if $M(A)$ is isomorphic to a vector
matroid whose ground set consists of vectors in $K^{k}$. Hence, to say a
matroid is representable is to say it is $K$-representable for some field $K$.
On the other hand, a matroid is _regular_ if it is $K$-representable over
every field $K$.
The _rank_ of a matroid $\mathcal{M}=(E,\mathcal{I})$ is defined as
$\operatorname{rank}(\mathcal{M})=\max_{I\in\mathcal{I}}|I|.$
When $\mathcal{M}$ is graphic, $\operatorname{rank}(\mathcal{M})$ is the size
of a maximal spanning forest for any $G$ satisfying $\mathcal{M}=M(G)$. When
$\mathcal{M}$ is representable,
$\operatorname{rank}(\mathcal{M})=\operatorname{rank}(A)$ for any matrix $A$
satisfying $\mathcal{M}=M(A)$. With these definitions in hand, we can now
describe some of the many benefits of working with regular matroids.
###### Theorem 2.1 ([7, Definition/Theorem 2.3]).
A positive-rank matroid $\mathcal{M}$ is regular if and only if any of the
following hold:
1. (i)
$\mathcal{M}$ is representable over every field.
2. (ii)
$\mathcal{M}=M(A)$ for some matrix $A$ in which every minor is $\pm 1$ or $0$.
3. (iii)
$\mathcal{M}=M(A)$ for some full-rank matrix $A$ in which every maximal minor
is $\pm 1$ or $0$.
Note that a matrix satisfying the minor condition in (ii) is called _totally
unimodular_ and a matrix satisfying the minor condition in (iii) is called
_weakly unimodular_. Borrowing more terminology from linear algebra, an
element of $\mathcal{I}$ of cardinality $\operatorname{rank}(\mathcal{M})$ is
called a _basis_ of $\mathcal{M}$. The set
$\mathcal{B}=\mathcal{B}(\mathcal{M})$ of bases, together with the ground set
set $E$, can be used to define a matroid without explicit mention of
independent sets; when this is done, we usually write
$\mathcal{M}=(E,\mathcal{B})$.
Now, from a matroid $\mathcal{M}=(E,\mathcal{B})$ we may produce its _dual_
matroid $\mathcal{M}^{*}=(E,\mathcal{B}^{*})$ by setting
$\mathcal{B}^{*}=\\{B-E\subseteq E\mid B\in\mathcal{B}\\}.$
Many properties of $\mathcal{M}^{*}$ can be easily deduced from $\mathcal{M}$.
For example, $\mathcal{M}^{*}$ is representable if and only if $\mathcal{M}$
is representable. This does not hold for graphic matroids, though: if
$\mathcal{M}$ is graphic, then $\mathcal{M}^{*}$ is graphic if and only if
$\mathcal{M}=M(G)$ for some planar graph $G$.
There are additional notions in graph theory that have corresponding matroidal
analogues. For example, a _circuit_ in a matroid is a minimal dependent set;
this is the analogue of a cycle in a graph. Letting
$\mathcal{C}=\mathcal{C}(\mathcal{M})$ denote the set of circuits of a
matroid, we define the matroid to be _bipartite_ if each of the circuits has
even size. This further allows one to call a matroid _Eulerian_ if it is the
dual of a bipartite matroid. These definitions reflect the fact that a
connected, planar graph is Eulerian if and only if its dual is bipartite.
A _loop_ of $\mathcal{M}$ is a single element of $E$ that forms a circuit;
equivalently, a loop is an element that is in no basis of $\mathcal{M}$. On
the other hand, a _coloop_ of $\mathcal{M}$, called by some authors an
_isthmus_ , is a single element of $E$ that is in every basis of
$\mathcal{M}$. Now, given a matroid $\mathcal{M}=(E,\mathcal{I})$ and an
element $e\in E$ that is not a coloop, the _deletion_ of $e$ from
$\mathcal{M}$ is the matroid $\mathcal{M}-e$ having ground set $E-e$ and
independent sets $\\{I\in\mathcal{I}\mid e\notin I\\}$. The _contraction_ of
$\mathcal{M}$ by $e\in E$, where $e$ is not a loop, is denoted $\mathcal{M}/e$
and has ground set $E-e$ and independent sets $\\{I-e\mid e\in
I\in\mathcal{I}\\}$. Deletion and contraction are dual operations: if $e$ is
neither a loop nor a coloop of $\mathcal{M}$, then
$(\mathcal{M}-e)^{*}=\mathcal{M}^{*}/e$. Many binary operations on graphs
extend to matroids as well: the _direct sum_ , or _1-sum_ , of matroids
$\mathcal{M}_{1}$ and $\mathcal{M}_{2}$ having disjoint ground sets $E_{1}$
and $E_{2}$ is $\mathcal{M}_{1}\oplus\mathcal{M}_{2}=(E_{1}\cup
E_{2},\mathcal{I}_{3})$, where $\mathcal{I}_{3}$ consists of independent sets
of the form $I=I_{1}\cup I_{2}$ where $I_{1}\in\mathcal{I}_{1}$ and
$I_{2}\in\mathcal{I}_{2}$.
To define the next operation we assume that $E_{1}\cap E_{2}=\\{p\\}$ and call
this element the _basepoint_. The _parallel connection_ , denoted
$P(\mathcal{M}_{1},\mathcal{M}_{2})$, has ground set $E_{1}\cup E_{2}$ and is
most efficiently defined through its circuits: if $p$ is neither a loop nor a
coloop of $\mathcal{M}_{1}$, then
$\mathcal{C}(P(\mathcal{M}_{1},\mathcal{M}_{2}))=\,\mathcal{C}(\mathcal{M}_{1})\cup\mathcal{C}(\mathcal{M}_{2})\cup\\{(C_{1}\cup
C_{2})-p\mid p\in C_{i}\in\mathcal{C}(\mathcal{M}_{i})\text{ for each }i\\}.$
If $p$ is a loop of $\mathcal{M}_{1}$, then we set
$P(\mathcal{M}_{1},\mathcal{M}_{2})=\mathcal{M}_{1}\oplus(\mathcal{M}_{2}/p).$
Similarly, if $p$ is a coloop of $\mathcal{M}_{1}$, then we set
$P(\mathcal{M}_{1},\mathcal{M}_{2})=(\mathcal{M}_{1}-p)\oplus\mathcal{M}_{2}.$
Of course, if $\mathcal{M}_{1}$ and $\mathcal{M}_{2}$ have disjoint ground
sets, then a point of each ground set may be selected and relabeled $p$ to
fall into these definition, although the resulting parallel connection may
then depend on which points were selected. An important property of this
operation is that it preserves regularity: if $\mathcal{M}_{1}$ and
$\mathcal{M}_{2}$ are both regular matroids, then so is
$P(\mathcal{M}_{1},\mathcal{M}_{2})$ [14, Corollary 7.1.25].
### 2.2. Dissecting Tree Sets and Toric Algebra
It is frequently more tractable to gather the desired information about a
polytope by decomposing it into subpolytopes. A _dissection_ , or
_subdivision_ , of a polytope $P$ is a collection $\mathcal{S}$ of full-
dimensional subpolytopes of $P$ whose union is $P$ and for any
$S_{1},S_{2}\in\mathcal{S}$, the intersection $S_{1}\cap S_{2}$ is a face of
both. A _triangulation_ of $P$ is a dissection in which every
$S\in\mathcal{S}$ is a simplex. There are many ways to create triangulations
of a polytope; we will be concerned with one that has been recently introduced
in the study of ordinary SEPs in particular.
Given an oriented subgraph $H$ of a graph $G$ on $[n]$, let
$Q_{H}=\operatorname{conv}\\{e_{i}-e_{j}\mid ij\in H\\}$. Thus, if $T$ is a
spanning tree of $G$, then $Q_{T}$ is a simplex contained in $\Sigma(G)$. In
fact, $Q_{T}$ is a full-dimensional simplex contained in the boundary of
$\Sigma(G)$. Further still, $Q_{T}$ is a _unimodular_ simplex: it has
normalized volume $1$. It follows that the simplex
$\operatorname{conv}\\{Q_{T}\cup\\{0\\}\\}$ is a full-dimensional, unimodular
simplex contained in $\Sigma(G)$. It is desirable, then, to find a collection
$\mathcal{T}$ of oriented spanning trees of $G$ so that the set $\\{Q_{T}\mid
T\in\mathcal{T}\\}$ forms a triangulation of the boundary of
$\operatorname{\Sigma}(G)$. Such a set $\mathcal{T}$ is called a _dissecting
tree set_ for $G$.
Of course, if $\mathcal{T}$ is a dissecting tree set for $G$, then the
normalized volume of $\Sigma(G)$ is exactly $|\mathcal{T}|$. However, there is
more information about $\Sigma(G)$ to be extracted from $\mathcal{T}$. To
describe this information, consider a fixed vertex $v$ in an oriented tree
$T$. For each edge $e\in T$ we say that _$e$ points away from $v$ in $T$_ if
$v$ is contained in the same component as the tail of $e$ in $T-e$. Otherwise,
we say _$e$ points towards $v$ in $T$_.
###### Theorem 2.2 ([10, Theorem 1.2]).
Let $G$ be a connected graph, let $\mathcal{T}$ be a dissecting tree set for
$G$, and let $v$ be any vertex of $G$. Then the $i^{th}$ entry in
$h^{*}(\Sigma(G))$ is
$|\\{T\in\mathcal{T}\mid T\text{ has exactly }i\text{ edges pointing away from
}v\\}|.$
In order to determine when we have a dissecting tree set, we turn to Gröbner
basis techniques. Here, we will introduce the necessary algebraic background
to prove our main results. We largely follow the framework given in [9]. For
more details about monomial orders and their relationships to triangulations
of polytopes, see [16, Chapter 8].
Recall that, given a lattice polytope $P\subseteq\mathbb{R}^{n}$ and a field
$K$, we may construct a semigroup algebra
$K[P]=K[x^{v}z^{m}\mid v\in mP\cap\mathbb{Z}^{n}]$
where $x^{v}=x_{1}^{v_{1}}\cdots x_{n}^{v_{n}}$. The _toric ideal_ of $P$ is
defined as the kernel of the map
$\pi_{P}:K[t_{v}\mid v\in P\cap\mathbb{Z}^{n}]\to K[P]$
defined by setting $\pi_{P}(t_{v})=x^{v}z$; this ideal is denoted $I_{P}$.
Consider a monomial order $\prec$ on $K[t_{v}\mid v\in P\cap\mathbb{Z}^{n}]$
and any ideal $I$ of this algebra. For each polynomial $f\in I$, its _initial_
term with respect to $\prec$, denoted $\operatorname{in}_{\prec}(f)$, is the
term of $f$ that is greatest with respect to $\prec$. The _initial ideal_ of
$I$ with respect to $\prec$ is
$\operatorname{in}_{\prec}(I)=(\operatorname{in}_{\prec}(f)\mid f\in I).$
A _Gröbner basis_ of $I$, which we denote by $\mathscr{G}$, is a finite
generating set of $I$ such that
$\operatorname{in}_{\prec}(I)=(\operatorname{in}_{\prec}(g)\ |\
g\in\mathscr{G})$. We call $\mathscr{G}$ _reduced_ if each element has a
leading coefficient of $1$ and if for any $g_{1},g_{2}\in\mathscr{G}$,
$\operatorname{in}_{\prec}(g_{1})$ does not divide any term of $g_{2}$. There
are many Gröbner bases of $I$ but there is exactly one reduced Gröbner basis
of $I$.
Now suppose $P\subseteq\mathbb{R}^{n}$ is an $n$-dimensional lattice polytope
and $P\cap\mathbb{Z}^{n}=\\{l_{1},\ldots,l_{s}\\}$. Choose a weight vector
$w=(w_{1},\ldots,w_{s})\in\mathbb{R}^{s}$ such that the polytope
$P_{w}:=\operatorname{conv}\\{(l_{1},w_{1}),\ldots,(l_{s},w_{s})\\}\subseteq\mathbb{R}^{n+1}$
is $(n+1)$-dimensional, i.e., $P_{w}$ does not lie in an affine hyperplane of
$\mathbb{R}^{n+1}$. Some facets of $P_{w}$ will have outward-pointing normal
vectors with a negative last coordinate. By projecting these facets back to
$\mathbb{R}^{n}$, we obtain the facets of a polyhedral subdivision
$\Delta_{w}(P)$ of $P$. Any triangulation that can be obtained in this way for
an appropriate choice of $w$ is called a _regular_ triangulation. Importantly,
regular unimodular triangulations are in one-to-one correspondence with
reduced Gröbner bases whose initial terms are all squarefree [16, Corollary
8.9]. It is well-known that any monomial order obtained as a graded reverse
lexicographic (grevlex) order can be represented by a weight vector.
In the case of $P=\Sigma(\mathcal{M})$ for a regular matroid
$\mathcal{M}=(E,\mathcal{I})$, we can identify $K[t_{v}\mid v\in
P\cap\mathbb{Z}^{n}]$ with
(2.1) $K\left[\\{z\\}\cup\bigcup_{e\in E}\\{x_{e},y_{e}\\}\right].$
Here, we let $x_{e}$ correspond to the column of a fixed matrix $M$ realizing
$\mathcal{M}$ that corresponds to $e$, and we will let $y_{e}$ correspond to
the negative of this column. The variable $z$, then, corresponds to the
origin. So, for each $e\in E$, we can treat $x_{e}$ and $y_{e}$ as encoding an
orientation of $e$. For ease of notation, if we are considering $e\in E$ with
a particular orientation, then we let $p_{e}$ denote the corresponding
variable and $q_{e}$ the variable corresponding to the opposite orientation.
###### Lemma 2.3 (cf. [9, Proposition 3.8]).
Let $\mathcal{M}$ be a regular matroid on the ground set $E$, let
$z<x_{e_{1}}<y_{e_{1}}<\cdots<x_{e_{n}}<y_{e_{n}}$ be an order on the lattice
points of $\Sigma(\mathcal{M})$, and let $\prec$ be the grevlex order on (2.1)
induced from this ordering on the variables. The collection of all binomials
of the following types forms a Gröbner basis of $I_{\Sigma(\mathcal{M})}$ with
respect to grevlex order:
1. (i)
For every $2k$-element circuit $C$ of $\mathcal{M}$ and choice of orientation
for each $e\in C$, and any $k$-element subset $I\subseteq C$ not containing
the weakest element among those of $C$,
$\prod_{e\in I}p_{e}-\prod_{e\in C-I}q_{e}.$
2. (ii)
For every $(2k+1)$-element circuit $C$ of $\mathcal{M}$ and choice of
orientation for each $e\in C$, and any $(k+1)$-element subset $I\subseteq C$,
$\prod_{e\in I}p_{e}-z\prod_{e\in C-I}q_{e}.$
3. (iii)
For any $e\in E$, $x_{e}y_{e}-z^{2}$.
In each case, the binomial is written so that the initial term has positive
sign.
###### Proof.
This proof follows the same approach as the proof to [9, Proposition 3.8].
First, recognize we need to show that for any binomial $m_{1}-m_{2}$ in
$I_{\Sigma(\mathcal{M})}$, one of $m_{1}$ or $m_{2}$ is divisible by the
leading monomial of one of the binomials described above.
Fix a totally unimodular matrix $A$ representing $\mathcal{M}$, and label its
columns $a_{1},\dots,a_{n}$. Both monomials $m_{1}$ and $m_{2}$ correspond to
a matrix so that $a_{i}$ is a column of the matrix if and only if $x_{a_{i}}$
is present in the monomial, and $-a_{i}$ is a column of the matrix if and only
if $y_{a_{i}}$ is present in the monomial. Let $A_{1}$ (resp. $A_{2}$) be the
matrix corresponding to $m_{1}$ (resp. $m_{2}$). Of course, we may assume that
neither matrix of $A_{1}$ and $A_{2}$ has both $a_{i}$ and $-a_{i}$ for any
$i$, since then the monomial would be divisible by $x_{a_{i}}y_{a_{i}}$.
Consider $-A_{2}$. Since $m_{1}-m_{2}$ is contained in the toric ideal of
$I_{\Sigma(\mathcal{M})}$, there is some minimal linearly dependent set $C$ of
columns in the matrix $[A_{1}\mid-A_{2}]$. Denote by $C_{1}$ the set of
vectors in $C$ that are columns in $A_{1}$, and let $c_{1}=|C_{1}|$.
Similarly, denote by $C_{2}$ the set of vectors in $C$ that are columns in
$-A_{2}$, and let $c_{2}=|C_{2}|$. We may assume $c_{1}\geq c_{2}$.
First, suppose that $c_{1}+c_{2}=2k$ for some $k>1$. If $c_{1}>c_{2}$,
consider the product of the $k$ largest variables, with respect to $\prec$,
among those corresponding to vectors in $C_{1}$. The leading term of the
corresponding binomial in (i) then divides $m_{1}$. If $c_{1}=c_{2}$, we may
assume without loss of generality that the smallest variable corresponding to
a vector in $C$ belongs to $A_{2}$ and proceed as before. If
$c_{1}+c_{2}=2k+1$, then since $a>b$, the leading term of the binomial in (ii)
corresponding to $(k+1)$ vectors in $A_{1}$ divides $m_{1}$. ∎
## 3\. Extensions of known properties of SEPs
Before providing the extensions of known properties of ordinary SEPs to
generalized SEPs, we recall the formula of the dimensions of them.
###### Proposition 3.1 (cf. [7, Theorem 4.3]).
Let $\mathcal{M}$ be a regular matroid on the ground set $E$. Then
$\dim\Sigma(\mathcal{M})=\operatorname{rank}(\mathcal{M})$. In particular,
given a connected graph $G$ with $n$ vertices, we have $\dim\Sigma(G)=n-1$.
Moreover, we have
$\dim\Sigma(\mathcal{M}^{*})=|E|-\operatorname{rank}(\mathcal{M})$.
First, we extend [13, Proposition 4.4], using the same overall strategy used
in its proof.
###### Proposition 3.2 (cf. [13, Proposition 4.4]).
Let $\mathcal{M}$ be a bipartite regular matroid on $E$. For any $e\in E$,
$h^{*}(\operatorname{\Sigma}(\mathcal{M});t)=(1+t)h^{*}(\operatorname{\Sigma}(\mathcal{M}/e);t).$
###### Proof.
Let $A$ be a full-rank, totally unimodular matrix of rank $r$ representing
$\mathcal{M}$, and denote its columns by $a_{1},\dots,a_{n}$ so that
$e=a_{1}$. Also, let $A_{e}$ be the matrix representing $\mathcal{M}/e$. Since
$\mathcal{M}$ is bipartite, the Gröbner basis $\mathscr{G}$ from Lemma 2.3
consists of the binomials of the form (i) and (iii). Let $A^{\prime}$ be the
rank-$(r+1)$ matrix
$A^{\prime}=\left[\begin{array}[]{@{}c|c@{}}A_{e}&\begin{matrix}0\\\ \vdots\\\
0\end{matrix}\\\
\hline\cr\begin{matrix}0&\cdots&0\end{matrix}&1\end{array}\right]$
and let $\mathcal{M}^{\prime}$ be the matroid represented by $A^{\prime}$. The
Gröbner basis of $\Sigma(\mathcal{M}^{\prime})$ consists of all binomials $b$
satisfying one of the following:
$b=\prod_{a_{i}\in I}x_{a_{i}}-\prod_{a_{i}\in C-I}y_{a_{i}},$
where $C$ is a set of columns of $A$ corresponding to a circuit in
$\mathcal{M}$ of size $2k$ and $a_{1}\notin C$, and $I$ is a $k$-subset of $C$
such that $a_{j}\notin I$ where $j=\min\\{i\mid a_{i}\in C\\}$;
$b=\prod_{a_{i}\in I}x_{a_{i}}-z\prod_{a_{i}\in C-I}y_{a_{i}},$
where $C\cup a_{1}$ corresponds to a circuit in $\mathcal{M}$ of size $2k+2$
and $I$ is a $(k+1)$-subset of $C$;
$b=x_{a_{i}}y_{a_{i}}-z^{2}$
for $1\leq i\leq k$. Thus, the initial terms in the Gröbner bases for
$I_{\Sigma(\mathcal{M})}$ and $I_{\Sigma(\mathcal{M}/e)}$ are the same.
Following the proof strategy of [8, Theorem 3.1], it follows that
$h^{*}(\Sigma(\mathcal{M});t)=h^{*}(\Sigma(\mathcal{M}^{\prime});t)=h^{*}(\Sigma(e_{r});t)h^{*}(\Sigma(\mathcal{M});t)=(1+t)h^{*}(\Sigma(\mathcal{M});t),$
where $e_{r}$ is a standard basis vector, as claimed. ∎
Next we present an extension of [6, Theorem 44]. Its proof is entirely
analogous to the proof of [6, Theorem 44], in the same way that our proof of
Lemma 2.3 adapts the proof [9, Proposition 3.8], and our proof of Proposition
3.2 adapts the proof of [13, Proposition 4.4]. So, due to the length of the
proof and the similarity in how we adapt it from [6, Theorem 44], we omit its
details.
###### Theorem 3.3 (cf. [6, Theorem 44]).
Let $\mathcal{M}$ be a regular matroid and let $\mathcal{M}^{\prime}$ be a
bipartite regular matroid. Then
$h^{*}(\operatorname{\Sigma}(P(\mathcal{M}_{1},\mathcal{M}_{2}));t)=\frac{h^{*}(\operatorname{\Sigma}(\mathcal{M}_{1});t)h^{*}(\operatorname{\Sigma}(\mathcal{M}_{2});t)}{1-t}.$
## 4\. A counterexample of $\gamma$-nonnegativity for generalized SEPs
In [13], the authors conjectured that SEPs are $\gamma$-nonnegative. It is
natural to want to extend this conjecture to generalized SEPs, but the
conjecture in this setting is false. Throughout the remainder of this section,
we will be considering the matroids $M^{*}(K_{3,n})$ for $n\geq 3$.
###### Proposition 4.1.
Let $\mathcal{M}=M^{*}(K_{3,n})$ with $n\geq 3$ and let $E$ be the ground set
of $\mathcal{M}$.
1. (i)
$\mathcal{M}$ can be represented by $[I_{2(n-1)}\mid B_{n}]$, where
$B_{n}=\begin{bmatrix}1&1&0&-1&0&0&\cdots&0\\\ 1&0&1&-1&0&0&\cdots&0\\\
1&1&0&0&-1&0&\cdots&0\\\ 1&0&1&0&-1&0&\cdots&0\\\ 1&1&0&0&0&-1&\cdots&0\\\
1&0&1&0&0&-1&\cdots&0\\\
\vdots&\vdots&\vdots&\vdots&\vdots&\vdots&\ddots&\vdots\\\
1&1&0&0&0&0&\cdots&-1\\\ 1&0&1&0&0&0&\cdots&-1\\\ \end{bmatrix}.$
2. (ii)
$\mathcal{M}-e$ is not graphic for any $e\in E$ if $n\geq 4$.
###### Proof.
1. (i)
It is known that if a matroid $\mathcal{M}$ is represented by a matrix
$A=[I\mid D]$, then $\mathcal{M}^{*}$ is represented by $[I\mid-D^{T}]$, where
$\cdot^{T}$ denotes the transpose ([14, Theorem 2.2.8]). Hence, we prove that
the matrix $[I\mid D]$ represents $M(K_{3,n})$ where $D=-B_{n}^{T}$.
Let $V(K_{3,n})=\\{u_{1},u_{2},u_{3}\\}\cup\\{v_{1},\ldots,v_{n}\\}$ and
$E(K_{3,n})=\\{u_{i}v_{j}\mid 1\leq i\leq 3,1\leq j\leq n\\}$. Fix a spanning
tree consisting of the following edges:
$e_{1}=u_{1}v_{1},\quad e_{2}=v_{1}u_{2},\quad e_{3}=v_{1}u_{3},\quad
e_{4}=u_{1}v_{2},\,\ldots\,,e_{n+2}=u_{1}v_{n}.$
This corresponds to the identity matrix in the representing matrix $[I\mid
D^{\prime}]$ of $M(K_{3,n})$. Our goal is to show that $D^{\prime}=-D^{T}$. By
assigning the orientation of each of these edges as described (i.e., from the
first vertex to the second vertex), for the remaining edges of $K_{3,n}$, we
see that $u_{2}v_{j}$ (resp. $u_{3}v_{j}$) with $2\leq j\leq n$ form minimal
dependent sets with the opposite of $e_{2}$, the opposite of $e_{1}$, and
$e_{j+2}$. This dependent set corresponds to the $(2j-1)$-th (resp. $2j$-th)
column of $-D^{T}$, as claimed.
2. (ii)
It is known that given a matroid $\mathcal{M}$ and an element $e$ of its
ground set, we have $\mathcal{M}^{*}-e=(\mathcal{M}/e)^{*}$ (see [14, 3.1.1]).
Hence, it is enough to show that $K_{3,n}/e$ never becomes planar for any edge
of $K_{3,n}$ This is easy to see: the edge set of $K_{3,n}/u_{1}v_{1}$ is
$\\{u_{i}v_{j}\mid i=2,3,\,1\leq j\leq n\\}\cup\\{v_{1}v_{k}\mid 2\leq k\leq
n\\}$, which contains $K_{3,n-1}$ as a subgraph.
∎
###### Theorem 4.2.
In every dimension at least $10$ there is a generalized SEP that is not
$\gamma$-nonnegative.
###### Proof.
Consider the matroid $\mathcal{M}=M^{*}(K_{3,6})$. Since $K_{3,6}$ has $18$
edges, we know that $\dim\Sigma(M^{*}(K_{3,6}))=18-(9-1)=10$. From its matrix
description, it is possible to show that
$h^{*}(\Sigma(\mathcal{M}))=(1,26,297,1908,6264,9108,6264,1908,297,26,1),$
hence $\gamma(\Sigma(\mathcal{M}))=(1,16,124,596,914,-148)$.
Now let $\mathcal{M}_{k}=\mathcal{M}\oplus x_{1}\oplus\cdots\oplus x_{k}$
where the $x_{i}$ are pairwise distinct elements, none of which is already in
$\mathcal{M}$. The generalized SEP $\Sigma(\mathcal{M}_{k})$ is the $k$-fold
direct sum of $\Sigma(\mathcal{M})$ with copies of the interval $[-1,1]$.
Consequently, $\dim(\mathcal{M}_{k})=10+k$,
$h^{*}(\Sigma(\mathcal{M}_{k});t)=h^{*}(\mathcal{M};t)(1+t)^{k},$
from which it follows that
$\gamma(\Sigma(\mathcal{M}_{k}))=\gamma(\Sigma(\mathcal{M}))$. ∎
Computational evidence suggests that the $\gamma$-vectors for
$\Sigma(M^{*}(K_{3,n}))$ fail to be nonnegative for all even $n\geq 6$,
although we do not aim to prove so here. Instead, we examine how these
matroids are situated between ordinary SEPs and generalized SEPs: by deleting
exactly two elements from $M^{*}(K_{3,n})$, the resulting matroid is now
graphic. In fact, by invoking Whitney’s $2$-isomorphism theorem [14, Theorem
5.3.1], it is a brief exercise to show that there is, up to isomorphism, a
unique graph $\Gamma(n)$ such that
$M^{*}(K_{3,n})-\\{e,e^{\prime}\\}=M(\Gamma(n))$. We will be more precise
about what this graph looks like in Section 5. For now, we simply state our
main results regarding $\Sigma(\Gamma(n))$.
###### Theorem 4.3.
For each $n\geq 1$,
(4.1)
$\begin{split}h^{*}(\Sigma(\Gamma(n+1));t)&=\sum_{k=0}^{n}\binom{n}{k}(2t)^{n-k}\sum_{(p,q)\in
S_{k}}\binom{2\lfloor\frac{k}{2}\rfloor+1}{p+q+1}f_{p,q,k}(t)\\\
&=\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2t)^{n-2\ell-1}\left(2t\binom{n}{2\ell}+(1+t)^{2}\binom{n}{2\ell+1}\right)\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell}(t),\end{split}$
where
$\displaystyle S_{k}$ $\displaystyle=\\{(x,y)\in\mathbb{Z}:x\geq 0,y\geq
0,x+y\leq k\\},$ $\displaystyle f_{p,q,2\ell}(t)$ $\displaystyle=t^{2\ell-
p-q}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}\binom{2\ell-
p-q}{\ell-q-i+j}t^{2(i+j)},\text{ and }$ $\displaystyle f_{p,q,2\ell+1}(t)$
$\displaystyle=(1+t)^{2}f_{p,q,2\ell}(t).$
For example, the following are the computations of $h^{*}(\Gamma(n+1);t)$ for
some small values of $n$:
$\displaystyle h^{*}(\Sigma(\Gamma(3));t)$ $\displaystyle=$
$\displaystyle\binom{2}{0}(2t)^{2}\binom{1}{1}f_{0,0,0}(t)+\binom{2}{1}(2t)^{1}(1+t)^{2}\left(\binom{1}{1}f_{0,0,0}(t)+\binom{1}{2}(f_{1,0,0}(t)+f_{0,1,0}(t))\right)$
$\displaystyle+$
$\displaystyle\binom{2}{2}(2t)^{0}\left(\binom{3}{1}f_{0,0,2}(t)+\binom{3}{2}(f_{1,0,2}(t)+f_{0,1,2}(t))+\binom{3}{3}(f_{1,1,2}(t)+f_{2,0,2}(t)+f_{0,2,2}(t))\right),$
$\displaystyle h^{*}(\Sigma(\Gamma(4));t)$ $\displaystyle=$
$\displaystyle\binom{3}{0}(2t)^{3}\binom{1}{1}f_{0,0,0}(t)+\binom{3}{1}(2t)^{2}(1+t)^{2}\left(\binom{1}{1}f_{0,0,0}(t)+\binom{1}{2}(f_{1,0,0}(t)+f_{0,1,0}(t))\right)$
$\displaystyle+$
$\displaystyle\binom{3}{2}(2t)^{1}\left(\binom{3}{1}f_{0,0,2}(t)+\binom{3}{2}(f_{1,0,2}(t)+f_{0,1,2}(t))+\binom{3}{3}(f_{1,1,2}(t)+f_{2,0,2}(t)+f_{0,2,2}(t))\right)$
$\displaystyle+$
$\displaystyle\binom{3}{3}(2t)^{0}(1+t)^{2}\left(\binom{3}{1}f_{0,0,2}(t)+\binom{3}{2}(f_{1,0,2}(t)+f_{0,1,2}(t))+\binom{3}{3}(f_{1,1,2}(t)+f_{2,0,2}(t)+f_{0,2,2}(t))\right).$
Our second main result regarding $\Gamma(n)$ is the following. For
readability, we defer its proof to Section 6.
###### Theorem 4.4.
For each $n\geq 1$,
$\displaystyle\gamma(\Gamma(n+1);t)$ $\displaystyle=\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2t)^{n-2\ell-1}\left(2\binom{n}{2\ell}t+\binom{n}{2\ell+1}\right)\sum_{a=0}^{\ell}\binom{2a}{a}t^{a}$
$\displaystyle=\sum_{m=0}^{n}\binom{n}{m}(2t)^{n-m}\sum_{a=0}^{\lfloor
m/2\rfloor}\binom{2a}{a}t^{a}.$
In particular, $\Sigma(\Gamma(n+1))$ is $\gamma$-nonnegative for all $n$.
As a corollary of Theorem 4.3, we can obtain the normalized volume of
$\Sigma(\Gamma(n+1))$. For its proof, as well as the proof of Theorem 4.3, we
collect some well-known formulas on the summation of binomial coefficients.
###### Lemma 4.5.
Let $a,b,c,d\in\mathbb{Z}_{\geq 0}$. Then the following hold:
$\displaystyle\text{(i)}\;\sum_{i=0}^{c}\binom{a}{i}\binom{b}{c-i}=\binom{a+b}{c},\;\;\text{(ii)}\;\sum_{i=0}^{c}\binom{a+i}{b}\binom{c-i}{d}=\binom{a+c+1}{b+d+1},\;\;\text{(iii)}\;\binom{a}{b}\binom{b}{c}=\binom{a-c}{b-c}\binom{a}{c}.$
###### Corollary 4.6.
The normalized volume of $\Sigma(\Gamma(n+1))$ is
$2^{n}\sum_{k=0}^{n}(k+1)\binom{n}{k}\binom{k}{\lfloor\frac{k}{2}\rfloor}.$
###### Proof.
We may set $t=1$ in (4.1). For $f_{p,q,2\ell}(1)$, we see that
$f_{p,q,2\ell}(1)=\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}\binom{2\ell-
p-q}{\ell-
q-i+j}=\sum_{j=0}^{q}\binom{q}{j}\binom{2\ell-q}{\ell-q+j}=\sum_{j=0}^{q}\binom{q}{j}\binom{2\ell-q}{\ell-j}=\binom{2\ell}{\ell}$
by using Lemma 4.5 (i) twice. Hence,
$\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}\binom{2\ell}{\ell}=\binom{2\ell}{\ell}\sum_{m=0}^{2\ell}(m+1)\binom{2\ell+1}{m+1}=\binom{2\ell}{\ell}\sum_{m=0}^{2\ell}(2\ell+1)\binom{2\ell}{m}=2^{2\ell}(2\ell+1)\binom{2\ell}{\ell}.$
Therefore,
$\displaystyle h^{*}(\Sigma(\Gamma(n+1));1)$
$\displaystyle=\sum_{\ell=0}^{\lfloor
n/2\rfloor}2^{n-2\ell}\left(\binom{n}{2\ell}+2\binom{n}{2\ell+1}\right)\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}\binom{2\ell}{\ell}$
$\displaystyle=\sum_{\ell=0}^{\lfloor
n/2\rfloor}2^{n-2\ell}\left(\binom{n}{2\ell}+2\binom{n}{2\ell+1}\right)2^{2\ell}(2\ell+1)\binom{2\ell}{\ell}$
$\displaystyle=2^{n}\left(\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2\ell+1)\binom{n}{2\ell}\binom{2\ell}{\ell}+2\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2\ell+1)\binom{n}{2\ell+1}\binom{2\ell}{\ell}\right)$
$\displaystyle=2^{n}\left(\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2\ell+1)\binom{n}{2\ell}\binom{2\ell}{\ell}+\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2\ell+2)\binom{n}{2\ell+1}\binom{2\ell+1}{\ell}\right)$
$\displaystyle=2^{n}\sum_{k=0}^{n}(k+1)\binom{n}{k}\binom{k}{\lfloor\frac{k}{2}\rfloor}.$
∎
## 5\. Proof of Theorem 4.3
In this section, we put all of the necessary pieces together for the proof of
Theorem 4.3. We will produce a dissecting tree set by identifying which
oriented spanning trees correspond to monomials that are not divisible by an
initial monomial of one of the types in Lemma 2.3. Before this, we collect
several identities needed in the proof of Theorem 4.3.
###### Lemma 5.1.
1. (i)
For all $n,p,k,\ell\geq 0$ such that $p+q\leq 2\ell\leq n$,
$\binom{n-(p+q)}{2\ell-(p+q)}\sum_{i=0}^{n}\binom{n-i}{p}\binom{i-1}{q-1}+\binom{n-1-(p+q)}{2\ell-1-(p+q)}\sum_{i=0}^{n}\binom{n-i}{p}\binom{i-1}{q}=\binom{n}{2\ell}\binom{2\ell+1}{p+q+1}.$
2. (ii)
For all $n,p,k,\ell\geq 0$ such that $p+q\leq 2\ell\leq n-1$,
$\binom{n-1-(p+q)}{2\ell-(p+q)}\sum_{i=0}^{n}\binom{n-i}{p}\binom{i-1}{q}=\binom{n}{2\ell+1}\binom{2\ell+1}{p+q+1}.$
###### Proof.
These straightforwardly follow by applying Lemma 4.5 (i) and (ii). ∎
Now, note that $M^{*}(K_{3,n+1})-e$ is never graphic for $n\geq 3$ (see
Proposition 4.1). However, by an appropriate choice $e,e^{\prime}$ from
$M^{*}(K_{3,n+1})$, we see that the matroid
$M^{*}(K_{3,n+1})-\\{e,e^{\prime}\\}$ may be represented by the matrix
$[I_{2n}\mid B^{\prime}_{n+1}]$ where $B^{\prime}_{n}$ is the matrix $B_{n}$
with the second and third columns removed. This matrix also represents the
graph $\Gamma(n+1)$ on the ground set
$\\{u_{1},v_{1},\dots,u_{n},v_{n},u_{n+1}\\}$ with edges $u_{i}u_{i+1}$,
$u_{i}v_{i}$, $v_{i}u_{i+1}$ for each $i=1,\dots,n$, as well as
$u_{1}u_{n+1}$. See Figure 1 for examples.
Figure 1. The graphs $\Gamma(3)$, left, and $\Gamma(4)$, right.
###### Remark 5.2.
We can obtain $\Gamma(n)$ as the graph representing
$M^{*}(K_{3,n})-\\{e,e^{\prime}\\}$ in a theoretical way. Although $K_{3,n}/e$
is not planar for any edge $e$ of $K_{3,n}$ as described in the proof of
Proposition 4.1 (ii), we can get a planar graph $\tilde{G}$ by the contraction
with a new edge of $K_{3,n}/e$. By taking the dual graph of $\tilde{G}$ (in
the sense of planar graphs), a planar graph $\Gamma(n)$ appears.
In what follows, we concentrate on the graph $\Gamma(n+1)$. Let
$\widetilde{e}=u_{1}u_{n+1}$. To describe these spanning trees, it will be
helpful to refer to a specific planar embedding of $\Gamma(n+1)$. We may view
$\Gamma(n+1)$ as the $(2n+1)$-cycle $u_{1}v_{1}u_{2}v_{2}\dots u_{n+1}u_{1}$,
together with the chords $u_{i}u_{i+1}$ for each $i=1,\dots,n$. In particular,
throughout the remainder of the article, we will use “chord” to refer to any
of the $n$ edges of the form $u_{i}u_{i+1}$.
Referring to a specific planar embedding will allow us to construct the
$h^{*}$-polynomial by applying Theorem 2.2. To do this, we need to identify a
dissecting tree set for $\Gamma(n+1)$. Let $\mathcal{G}_{n+1}$ be the reduced
Gröbner basis of $I_{\Sigma(\Gamma(n+1))}$ with respect to any grevlex order
in which the variable $z$ corresponding to the origin is weakest and the
variables $p_{\widetilde{e}}$ and $q_{\widetilde{e}}$ corresponding to
$\widetilde{e}$ are the second- and third-weakest. Using the framework
developed in Section 2, we can identify which oriented spanning trees
contribute to our dissecting tree set. To help us describe them, we introduce
some new terminology.
Observe that if $T$ is a spanning tree of $\Gamma(n+1)$, then for each of the
triangles with vertices $u_{i},v_{i},u_{i+1}$, there are exactly three
possibilities of edges appearing in $T$: $T$ contains either
* •
both $u_{i}v_{i}$ and $v_{i}u_{i+1}$, which we call a _chordless pair_ ,
* •
both $u_{i}v_{i}$ and $u_{i}u_{i+1}$ or both $u_{i}u_{i+1}$ and
$u_{i+1}v_{i}$, each of which we call a _chorded pair_ , or
* •
only $u_{i}v_{i}$ or only $v_{i}u_{i+1}$, each of which we call an _unpaired
edge_.
For example, the spanning tree of $\Gamma(5)$ in Figure 2 contains one
chordless pair of edges, two chorded pairs of edges, and one unpaired edge.
Figure 2. A spanning tree of $\Gamma(5)$. The edges $u_{1}v_{1}$ and
$v_{1}u_{2}$ form a chordless pair. The edges $u_{2}v_{2}$ and $u_{2}u_{3}$
form a chorded pair of type $\alpha$, while the edges $u_{4}v_{4}$ and
$u_{4}u_{5}$ form a chorded pair of type $\beta$. The edge $v_{3}u_{4}$ is an
unpaired edge.
###### Proposition 5.3.
An oriented spanning tree $T$ of $\Gamma(n+1)$ is part of the dissecting tree
set if and only if, for every cycle $C$ in $\Gamma(n+1)$, either $T$ contains
no more than $\lfloor\frac{|C|-1}{2}\rfloor$ edges of $C$ that are oriented in
the same direction, or the following conditions holds:
* When $|C|$ is even, both $\widetilde{e}\in T$ and $T$ contains exactly $|C|/2$ edges of $C$ that are oriented in the same direction.
###### Proof.
This follows from Lemma 2.3. ∎
###### Corollary 5.4.
Suppose $T$ is an oriented spanning tree in the dissecting tree set. Then the
edges in the chordless pair share the same tail or share the same head.
Moreover, their orientations may be reversed to obtain another oriented
spanning tree in the dissecting tree set.
###### Proof.
If the edges in the chordless pair, say, $u_{i}v_{i}$ and $v_{i}u_{i+1}$,
share neither the same tail nor the same head, then the $3$-cycle
$u_{i},v_{i},u_{i+1}$ of $\Gamma(n+1)$ contains two edges that are oriented in
the same direction. ∎
For a spanning tree $T$ of $\Gamma(n+1)$, let $c(T)$ denote the number of
chords $u_{i}u_{i+1}$ of $T$ and let
$\epsilon(T)=\begin{cases}1&\text{ if }\widetilde{e}\in T\\\ 0&\text{
otherwise.}\end{cases}$
###### Lemma 5.5.
Let $T$ be a member of our dissecting tree set.
1. (i)
If $\epsilon(T)=1$, then $T$ contains a total of $n-1$ chordless pairs and
chorded pairs of edges as well as exactly one unpaired edge.
2. (ii)
If $\epsilon(T)=0$, then $T$ contains a total of $n$ chordless and chorded
pairs of edges. Moreover, $c(T)$ must be always even in this case.
###### Proof.
Note that each spanning tree of $\Gamma(n+1)$ has $2n$ edges. Hence, the
statements on the numbers of chordless/chorded pairs and unpaired edge follow.
Regarding the evenness of $c(T)$, on the contrary, suppose that $c(T)=2a-1$
for some $a\in\mathbb{Z}_{>0}$. Then there are at least $a$ edges that have
the same direction. On the other hand, since there are $(n-2a+1)$ chordless
pairs that consist of $2(n-2a+1)$ edges, Corollary 5.4 claims that the exactly
half of these edges should have the same oriantation. Hence, we can find a
cycle $C$ consisting of those chords of chorded pairs, the edges of chordless
pairs and $\widetilde{e}$, that have length $2(n-2a+1)+(2a-1)+1=2(n-a+1)$,
such that there are at least $(n-a+1)$ edges in $C$ in the same direction, a
contradiction to Proposition 5.3. ∎
At this point we will take time here to closely examine some particular cases.
We do this with the goal of helping motivate and illustrate the key aspects of
the full proof, which becomes somewhat technical.
$c(T)+\epsilon(T)=0$: Let $T$ be a member of our dissecting tree set with
$c(T)+\epsilon(T)=0$. Then the (undirected) edges of $T$ consist entirely of
chordless pairs. When orienting this tree, each pair of edges $u_{i}v_{i}$,
$v_{i}u_{i+1}$ has two possibilities: either both edges point toward $v_{i}$
or both edges point away from $v_{i}$ (see Corollary 5.4). In either case,
there are $n$ edges that will point towards $u_{1}$. So, by Theorem 2.2 with
$v=u_{1}$, each such oriented tree corresponds to a summand of $t^{n}$ in
$h^{*}(\Sigma(\Gamma(n+1));t)$. There are $2^{n}$ valid orientations of this
tree, resulting in a summand of $(2t)^{n}$ in the $h^{*}$-polynomial after
over all of these trees.
$c(T)+\epsilon(T)=1$: Let $T$ be a member of our dissecting tree with
$c(T)+\epsilon(T)=1$. Then $\epsilon(T)=1$ and $c(T)=0$ (see Lemma 5.5 (ii)).
We can establish orientations on the chordless pairs as in the previous case,
although there are now only $n-1$ of these pairs. Any cycle containing one
edge of a chordless pair must also contain the other. By Corollary 5.4, these
two edges must have opposite orientations. For any cycle containing
$\widetilde{e}$, the orientations of $\widetilde{e}$ and the unpaired edge
should have the opposite direction; otherwise we can find at least $(n+1)$
edges with the same direction in $(2n+1)$-cycle of $\Gamma(n+1)$ (see
Corollary 5.4 and Lemma 5.5).
For $T$ to contribute to $h^{*}(\Sigma(\Gamma(n+1)))$, the orientation of
$\widetilde{e}$ and the unpaired edge must be opposite. To help describe which
summand of the $h^{*}$-polynomial corresponds to $T$, we will refer to two
portions of it relative to $u_{1}$. Given a spanning tree $T$ of $\Gamma(n+1)$
that contains $\widetilde{e}$, the graph $T-\widetilde{e}$ consists of two
components. Call the component containing $u_{1}$ the _left side_ of $T$ and
denote it by $L(T)$. The _right side_ of the tree, which we denote $R(T)$, is
the subgraph of $T$ consisting of $\widetilde{e}$ and all edges not in $L(T)$.
This definition makes it clear that $T$ can correspond to any of the following
summands of the $h^{*}$-polynomial, depending on its orientation and the
location of the unpaired edge, $e$:
* •
If $\widetilde{e}$ and $e$ are pointing away each other and $e\in L(T)$, then
$T$ corresponds to the summand $(2t)^{n-1}$.
* •
If $\widetilde{e}$ and $e$ are pointing away each other and $e\in R(T)$, then
$T$ corresponds to the summand $t(2t)^{n-1}$.
* •
If $\widetilde{e}$ and $e$ are pointing towards each other and $e\in L(T)$,
then $T$ corresponds to the summand $t^{2}(2t)^{n-1}$.
* •
If $\widetilde{e}$ and $e$ are pointing towards each other and $e\in R(T)$,
then $T$ corresponds to the summand $t(2t)^{n-1}$.
Thus, by considering all orientations on the edges of a spanning tree
satisfying $c(T)+\epsilon(T)=1$, the tree will correspond to a summand of
$(1+t)^{2}(2t)^{n-1}$. Note that $(1+t)^{2}=f_{0,0,1}(t)$.
$c(T)+\epsilon(T)=2$: Because of one additional subtlety that arises when
$c(T)+\epsilon(T)=2$, we will discuss this case as well. When this happens,
there are two subcases: the first being where $c(T)=2$ and $\epsilon(T)=0$ and
the second being where $c(T)=1$ and $\epsilon(T)=1$. In the first case, we can
repeat the same argument as when $c(T)+\epsilon(T)=0$, but we now have only
$n-2$ possible chordless pairs. As before, the orientations on the chordless
pairs result in a factor of $(2t)^{n-2}$ in the contribution $T$ makes to the
$h^{*}$-polynomial. Such a factor occurs $\binom{n}{2}$ times, since this is
the number of chords $u_{i}u_{i+1}$ that can appear as parts of chorded pairs.
Now, for each chorded pair, the orientation on the edge $u_{i}u_{i+1}$
dictates the orientation of the other edge $u_{j}u_{j+1}$ of the pair: they
have the opposite direction. However, whether the chorded pair consists of
$u_{i}v_{i}$ or $v_{i}u_{i+1}$ will make a difference in the summand of the
$h^{*}$-polynomial corresponding to $T$. We say that a chorded pair has type
$\alpha$ if the non-chord edge is incident to the vertex that is closer to
$u_{1}$ within $T$; otherwise, we say that a chorded pair has type $\beta$.
See Figure 2 for an example of each.
When summing the monomials of the $h^{*}$-polynomial corresponding to $T$, if
a chorded pair has type $\alpha$, then one of the orientations of its edges
results in a factor of $t^{2}$, while the other results in a factor of $1$. On
the other hand, if a chorded pair has type $\beta$, then each of the
orientations of its edges results in a factor of $t$. So, if we consider all
spanning trees $T$ of $\Gamma(n+1)$ satisfying $c(T)=2$ and $\epsilon(T)=0$,
then we know that there are $p$ chorded pairs of type $\alpha$ in $L(T)$ and
$q$ chorded pairs of type $\alpha$ in $R(T)$ for some nonnegative $p,q$
satisfying $p+q\leq 2$. This leaves $2-(p+q)$ chords to be part of a chorded
pair of type $\beta$. This is where the set $S_{k}$ in (4.1) will come from:
the pair $(p,q)$ will indicate the presence of $p$ chorded pairs of type
$\alpha$ in $L(T^{\prime})$ and $q$ of them in $R(T^{\prime})$, where the
notation $T^{\prime}$ is explained below.
Recognize that in any oriented spanning tree $T$ of $\Gamma(n+1)$, there is at
most one unpaired edge. Consequently, if an unpaired edge exists, then
$\widetilde{e}\in T$ and the unpaired edge must have endpoints $u_{i}$ and
$v_{j}$ for some $i$ and $j=i\pm 1$. Thus we may, without affecting the
monomial to which $T$ corresponds, modify $T$ by deleting the vertex $v_{j}$,
adding a new vertex $v_{n+1}$, and adding a new edge as follows:
* •
If $\widetilde{e}$ and $u_{i}v_{j}$ are pointing away each other and
$u_{i}v_{j}\in L(T)$, then add the directed edge $u_{1}v_{n+1}$.
* •
If $\widetilde{e}$ and $u_{i}v_{j}$ are pointing away each other and
$u_{i}v_{j}\in R(T)$, then add the directed edge $u_{n+1}v_{n+1}$.
* •
If $\widetilde{e}$ and $u_{i}v_{j}$ are pointing towards each other and
$u_{i}v_{j}\in L(T)$, then add the directed edge $v_{n+1}u_{1}$.
* •
If $\widetilde{e}$ and $u_{i}v_{j}$ are pointing towards each other and
$u_{i}v_{j}\in R(T)$, then add the directed edge $v_{n+1}u_{n+1}$.
See Figure 3 for illustrations of each of these cases. We call $T^{\prime}$ a
_modified (oriented) spanning tree_ of $\Gamma(n+1)$. Modified spanning trees
allow us to consider $\widetilde{e}$ and $u_{1}v_{n+1}$ or $u_{n+1}v_{n+1}$ as
a chorded pair, so that all edges in $T^{\prime}$ are part of a chorded or
chordless pair. Note that, in each case, the number of edges pointing towards
or away from $u_{1}$ in each of $L(T)$ and $R(T)$ is preserved by this
operation. We are now prepared to present the proof of Theorem 4.3.
Figure 3. Examples of modified spanning trees in $\Gamma(4)$. The original
spanning trees and their corresponding modified trees are written above and
below each other.
###### Proof of Theorem 4.3.
To begin, we first show the second equation of (4.1) holds. Consider a summand
$\binom{n}{k}(2t)^{n-k}\sum_{(p,q)\in
S_{k}}\binom{2\lfloor\frac{k}{2}\rfloor+1}{p+q+1}f_{p,q,k}(t).$
If $k=2\ell$ for some $\ell$, then adding this summand and the summand for
$2\ell+1$ becomes
$\binom{n}{2\ell}(2t)^{n-2\ell}\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell}(t)+\binom{n}{2\ell+1}(2t)^{n-2\ell-1}\sum_{(p,q)\in
S_{2\ell+1}}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell+1}(t).$
Note that if $p+q=2\ell+1$, then the binomial coefficient
$\binom{2\ell+1}{p+q+1}$ becomes $0$. So, there is no harm in replacing
$S_{2\ell+1}$ with $S_{2\ell}$. Further, since
$f_{p,q,2\ell+1}(t)=(1+t)^{2}f_{p,q,2\ell}(t)$, the above expression may be
written
$\binom{n}{2\ell}(2t)^{n-2\ell}\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell}(t)+\binom{n}{2\ell+1}(2t)^{n-2\ell-1}\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}(1+t)^{2}f_{p,q,2\ell}(t).$
Factoring appropriately results in the expression
$(2t)^{n-2\ell-1}\left(2t\binom{n}{2\ell}+(1+t)^{2}\binom{n}{2\ell+1}\right)\sum_{(p,q)\in
S_{2\ell}}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell}(t).$
Thus, ranging over all $\ell=0,1,2,\dots,\lfloor n/2\rfloor$ yields the second
equation of (4.1).
Now, fix an arbitrary $\ell$. We will treat the summand corresponding to
$\ell$ in two pieces:
(5.1) $\sum_{(p,q)\in
S_{2\ell}}(2t)^{n-2\ell}\binom{n}{2\ell}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell}(t)$
and
(5.2) $\sum_{(p,q)\in
S_{2\ell}}(2t)^{n-2\ell-1}(1+t)^{2}\binom{n}{2\ell+1}\binom{2\ell+1}{p+q+1}f_{p,q,2\ell}(t)$
since these will more closely resemble how to identify the oriented spanning
trees in our dissecting tree set, and their contributions to the
$h^{*}$-polynomial.
For (5.1), we interpret $2\ell$ as the number of chords in the modified tree
$T^{\prime}$. Then there is exactly one $i$ such that all of
$u_{n-i}u_{n-i+1}$, $u_{n-i}v_{n-i}$ and $v_{n-i}u_{n-i+1}$ are missing in
$T^{\prime}$. Namely, we treat both cases, $c(T)=2\ell$ and $\epsilon(T)=0$,
and $c(T)=2\ell-1$ and $\epsilon(T)=1$, simultaneously. Passing to the
modified spanning trees, we can say that $\widetilde{e}$ is part of a chorded
pair, and is of type $\alpha$ or $\beta$. Suppose first that $\widetilde{e}$
is part of a pair of type $\alpha$ in $T^{\prime}$. Then $p$ of the chords
$u_{1}u_{2},\dots,u_{n-i-1}u_{n-i}$ may be chosen to be parts of chorded pairs
of type $\alpha$, and these will all be part of $L(T^{\prime})$. We may then
choose $q-1$ of the chords $u_{n-i+1}u_{n-i+2},\dots,u_{n}u_{n+1}$ to be part
of chorded pairs of type $\alpha$, to ensure that $R(T^{\prime})$ contains $q$
of them. Of the remaining chords of the form $u_{j}u_{j+1}$ (excluding
$u_{n-i}u_{n-i+1}$), there are $n-(p+q)$ of them, and we must choose
$2\ell-(p+q)$ of them to be the chorded pairs of type $\beta$. All together we
have constructed
$\sum_{i=0}^{n}\binom{n-i}{p}\binom{i-1}{q-1}\binom{n-(p+q)}{2\ell-(p+q)}$
trees. On the other hand, suppose $\widetilde{e}$ is part of a chorded pair of
type $\beta$. Following an analogous argument, we obtain
$\sum_{i=0}^{n}\binom{n-i}{p}\binom{i-1}{q}\binom{n-1-(p+q)}{2\ell-1-(p+q)}$
spanning trees.
Now we may apply Lemma 5.1 (i) to obtain
$\binom{n}{2\ell}\binom{2\ell+1}{p+q+1}$
spanning trees that contribute to the $h^{*}$-polynomial when there are
$2\ell$ chords in the modified tree $T^{\prime}$.
For each of the edges $u_{i}u_{i+1}$ that are part of a chorded pair, the
structure of $\Gamma(n+1)$ and Proposition 5.3 imply that exactly half of them
must be pointing towards $u_{1}$ and exactly half of them must be pointing
away from $u_{1}$. The contribution to the $h^{*}$-polynomial depends on
whether those chorded pairs are of type $\alpha$ or type $\beta$.
For each of these trees, we can determine a valid orientation by considering
which orientations satisfy Proposition 5.3. For each spanning tree $T$ with
$p$ chorded pairs of type $\alpha$ in $L(T)$, we choose $i$ of them to point
toward $u_{1}$. Similarly, we choose $j$ of them from $R(T)$ to point toward
$u_{1}$. There are $2\ell-p-q$ chorded pairs remaining, all of type $\beta$;
we must orient these so that exactly $\ell$ of the chords $u_{r}u_{r+1}$ are
pointing in the same direction. By selecting which pairs are oriented so that
the chord is pointing counterclockwise in our planar embedding of
$\Gamma(n+1)$, this must be done in $\binom{2\ell-p-q}{\ell-(q-j)-i}$ ways,
since $i+(q-j)$ of the edges have already been given counterclockwise
orientations.
For each of these trees, there will be $(n-2\ell)$ edges oriented in
$T^{\prime}$ toward $u_{1}$ that come from the chordless pairs, and this is
true for either possible orientation of the pairs. From the paragraph above,
there will be $2(i+j)$ edges pointing toward $u_{1}$ that come from chorded
pairs of type $\alpha$, and $2\ell-p-q$ edges pointing toward $u_{1}$ that
come from chorded pairs of type $\beta$. Summing over all $i$ and $j$ gives us
a summand of
$(2t)^{n-2\ell}f_{p,q,2\ell}(t)$
in the computation of the $h^{*}$-polynomial. Further, because we are
considering the case of modified spanning trees, we have shown that there are
$\binom{n}{2\ell}\binom{2\ell+1}{p+q+1}$
of them. By multiplying the above two expressions and summing over all
$(p,q)\in S_{2\ell}$, we obtain (5.1).
Now, suppose $c(T)+\epsilon(T)=2\ell+1$. By Lemma 5.5, $\epsilon(T)=1$. We can
apply an argument analogous to the one above, using Lemma 5.1 (ii), to show
that there are
$\binom{n}{2\ell+1}\binom{2\ell+1}{p+q+1}$
oriented spanning trees with $p$ and $q$ choreded pairs of type $\alpha$ in
$L(T)$ and $R(T)$ in $T$, respectively. Note that the argument holds for each
possibility of whether $\widetilde{e}$ is of type $\alpha$ or $\beta$. So, we
must multiply each of these by the sum of the $h^{*}$-polynomial contributions
from these chorded pairs, which is $1+2t+t^{2}=(1+t)^{2}$. This establishes
(5.2), which completes the proof.
∎
## 6\. Proof of Theorem 4.4
To prove Theorem 4.4, we will use a generating function argument. We will make
regular use of the following three well-known identities:
(6.1) $\sum_{n\geq 0}\binom{n+k}{n}x^{n}=\frac{1}{(1-x)^{k+1}},$ (6.2)
$\sum_{k\geq 0}\binom{2k}{k}x^{k}=\frac{1}{\sqrt{1-4x}},\\\ $ (6.3)
$\sum_{n\geq
0}\binom{2n+k}{n}x^{n}=\frac{1}{\sqrt{1-4x}}\left(\frac{1-\sqrt{1-4x}}{2x}\right)^{k}.$
The first lemma we will need is the following.
###### Lemma 6.1.
For all $0\leq k\leq\ell$ we have
$\sum_{a=0}^{\ell}(-1)^{k-a}4^{\ell-a}\binom{2a}{a}\binom{\ell-k}{\ell-a}=\frac{\binom{\ell}{k}\binom{2\ell}{\ell}}{\binom{2\ell}{2k}}.$
###### Proof.
Using (6.1), (6.2) we compute the generating function for the left side of the
identity:
$\displaystyle\sum_{k\geq 0}\sum_{\ell\geq
0}\sum_{a=0}^{\ell}(-1)^{k-a}4^{\ell-a}\binom{2a}{a}\binom{\ell-k}{\ell-a}x^{k}y^{\ell}$
$\displaystyle=\sum_{k\geq 0}x^{k}\sum_{a\geq 0}\sum_{\ell\geq
a}(-1)^{k-a}4^{\ell-a}\binom{2a}{a}\binom{\ell-k}{\ell-a}y^{\ell}$
$\displaystyle=\sum_{k\geq 0}x^{k}\sum_{a\geq
0}(-1)^{k-a}\binom{2a}{a}\sum_{\ell\geq
0}4^{\ell}\binom{\ell+a-k}{\ell}y^{\ell+a}$ $\displaystyle=\sum_{k\geq
0}x^{k}\sum_{a\geq 0}(-1)^{k-a}\binom{2a}{a}\frac{y^{a}}{(1-4y)^{a-k+1}}$
$\displaystyle=\frac{1}{1-4y}\sum_{k\geq
0}\left(-x(1-4y)\right)^{k}\sum_{a\geq
0}\binom{2a}{a}\left(\frac{-y}{1-4y}\right)^{a}$
$\displaystyle=\frac{1}{(1-4y)(1-(-x(1-4y)))\sqrt{1-4\frac{-y}{1-4y}}}$
$\displaystyle=\frac{1}{\sqrt{1-4y}\ (1+x-4xy)}.$
Applying the Taylor expansion to this rational function, we find
$\displaystyle\frac{1}{\sqrt{1-4y}\ (1+x-4xy)}$ $\displaystyle=$
$\displaystyle\sum_{k\geq 0}(-1)^{k}k!(1-4y)^{k-\frac{1}{2}}\
\frac{x^{k}}{k!}$ $\displaystyle=$ $\displaystyle\sum_{k\geq 0}\sum_{\ell\geq
0}(-1)^{k}(-4)^{\ell}\left(k-\frac{1}{2}\right)\cdots\left(k-\frac{1}{2}-\ell+1\right)x^{k}\frac{y^{\ell}}{\ell!}$
$\displaystyle=$ $\displaystyle\sum_{k\geq 0}\sum_{\ell\geq
0}(-1)^{k}2^{\ell}(2\ell-2k-1)\cdots(-2k+1)x^{k}\frac{y^{\ell}}{\ell!}$
$\displaystyle=$ $\displaystyle\sum_{k\geq 0}\sum_{\ell\geq
0}\frac{(2\ell-2k)!(2k)!}{(\ell-k)!k!}x^{k}\frac{y^{\ell}}{\ell!}$
$\displaystyle=$ $\displaystyle\sum_{k\geq 0}\sum_{\ell\geq
0}\frac{\binom{\ell}{k}\binom{2\ell}{\ell}}{\binom{2\ell}{2k}}x^{k}y^{\ell}.$
∎
###### Lemma 6.2.
For all $k\geq 0$,
$\sum_{\ell\geq
0}\frac{2\ell+1}{2k+1}\binom{\ell}{k}\binom{2\ell}{\ell}x^{\ell}=\binom{2k}{k}\frac{x^{k}}{(1-4x)^{k+\frac{3}{2}}}.$
###### Proof.
By Taylor expansion, we have
$\displaystyle\binom{2k}{k}\frac{x^{k}}{(1-4x)^{k+\frac{3}{2}}}$
$\displaystyle=$ $\displaystyle\binom{2k}{k}x^{k}\sum_{m\geq
0}4^{m}\left(k+\frac{3}{2}\right)\left(k+\frac{5}{2}\right)\cdots\left(k+\frac{2m+1}{2}\right)\frac{x^{m}}{m!}$
$\displaystyle=$ $\displaystyle\binom{2k}{k}x^{k}\sum_{m\geq
0}2^{m}\left(2k+3\right)\left(2k+5\right)\cdots(2k+2m+1)\frac{x^{m}}{m!}$
$\displaystyle=$ $\displaystyle\binom{2k}{k}x^{k}\sum_{m\geq
0}2^{m}\frac{(2k+2m+1)!!}{(2k+1)!!}\frac{x^{m}}{m!}$ $\displaystyle=$
$\displaystyle x^{k}\sum_{m\geq
0}2^{m}\frac{(2k)!}{k!k!}\frac{(2k+2m+2)!}{2^{k+m+1}(k+m+1)!}\frac{2^{k+1}(k+1)!}{(2k+2)!}\frac{x^{m}}{m!}$
$\displaystyle=$ $\displaystyle x^{k}\sum_{m\geq
0}\frac{(2k+2m+2)!}{2(2k+1)(k+m+1)!k!m!}\ \ x^{m}$ $\displaystyle=$
$\displaystyle\sum_{\ell\geq
0}\frac{(2\ell+2)!}{2(2k+1)(\ell+1)!k!(\ell-k)!}x^{\ell}$ $\displaystyle=$
$\displaystyle\sum_{\ell\geq
0}\frac{2\ell+1}{2k+1}\binom{\ell}{k}\binom{2\ell}{\ell}x^{\ell}.$
∎
The second lemma we will need is the following.
###### Lemma 6.3.
For $\ell\geq 0$,
$\sum_{p,q\geq
0}\binom{2\ell+1}{p+q+1}x^{p}y^{q}=\frac{(x+1)^{2\ell+1}-(y+1)^{2\ell+1}}{x-y}.$
###### Proof.
We again use a generating function argument, beginning with the left side of
the desired identity multiplied by $x-y$.
$\displaystyle(x-y)\sum_{p,q\geq 0}\binom{2\ell+1}{p+q+1}x^{p}y^{q}$
$\displaystyle=$ $\displaystyle\sum_{p,q\geq
0}\binom{2\ell+1}{p+q+1}x^{p+1}y^{q}-\sum_{p,q\geq
0}\binom{2\ell+1}{p+q+1}x^{p}y^{q+1}$ $\displaystyle=$
$\displaystyle\sum_{p\geq 1}\sum_{q\geq
0}\binom{2\ell+1}{p+q}x^{p}y^{q}-\sum_{p\geq 0}\sum_{q\geq
1}\binom{2\ell+1}{p+q}x^{p}y^{q}$ $\displaystyle=$ $\displaystyle\sum_{p\geq
1}\binom{2\ell+1}{p}x^{p}-\sum_{q\geq 1}\binom{2\ell+1}{q}y^{q}$
$\displaystyle=$ $\displaystyle\left(1+\sum_{p\geq
1}\binom{2\ell+1}{p}x^{p}\right)-\left(1+\sum_{q\geq
1}\binom{2\ell+1}{q}y^{q}\right)$ $\displaystyle=$
$\displaystyle(x+1)^{2\ell+1}-(y+1)^{2\ell+1}.$
∎
###### Theorem 6.4.
For all $\ell\geq 0$ we have
$\sum_{p+q\leq 2\ell}\binom{2\ell+1}{p+q+1}t^{2\ell-
p-q}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}\binom{2\ell-
p-q}{\ell-
q-i+j}t^{2i+2j}=\sum_{a=0}^{\ell}\binom{2a}{a}t^{a}(t+1)^{4\ell-2a}.$
###### Proof.
For notational convenience, set
$F_{\ell}=\sum_{a=0}^{\ell}\binom{2a}{a}t^{a}(t+1)^{4\ell-2a}$
and
$G_{\ell,m}=\sum_{p+q\leq 2\ell}\binom{2\ell+1}{p+q+1}t^{2\ell-
p-q}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}\binom{2m-p-q}{m-q-i+j}t^{2i+2j}.$
We compute the generating function for $F_{\ell}$ as follows:
$\displaystyle\sum_{\ell\geq 0}F_{\ell}\lambda^{\ell}$ $\displaystyle=$
$\displaystyle\sum_{\ell\geq
0}\sum_{a=0}^{\ell}\binom{2a}{a}t^{a}(t+1)^{4\ell-2a}\lambda^{\ell}$
$\displaystyle=$ $\displaystyle\sum_{a\geq 0}\binom{2a}{a}t^{a}\sum_{\ell\geq
a}(t+1)^{4\ell-2a}\lambda^{\ell}$ $\displaystyle=$ $\displaystyle\sum_{a\geq
0}\binom{2a}{a}t^{a}\frac{(t+1)^{2a}\lambda^{a}}{1-(t+1)^{4}\lambda}$
$\displaystyle=$ $\displaystyle\frac{1}{1-(t+1)^{4}\lambda}\sum_{a\geq
0}\binom{2a}{a}t^{a}(t+1)^{2a}\lambda^{a}$ $\displaystyle=$
$\displaystyle\frac{1}{(1-(t+1)^{4}\lambda)\sqrt{1-4t(t+1)^{2}\lambda}}.$
Hence it is enough to show that
$\sum_{\ell\geq
0}G_{\ell,\ell}\lambda^{\ell}=\frac{1}{(1-(t+1)^{4}\lambda)\sqrt{1-4t(t+1)^{2}\lambda}}.$
Using Lemma 6.3, we have
$\displaystyle\sum_{m\geq 0}G_{\ell,m}\lambda^{m}$ $\displaystyle=\sum_{m\geq
0}\sum_{p,q\geq 0}\binom{2\ell+1}{p+q+1}t^{2\ell-
p-q}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}\binom{2m-p-q}{m-q-i+j}t^{2i+2j}\lambda^{m}$
$\displaystyle=\sum_{p,q\geq
0}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}t^{2\ell-
p-q+2i+2j}\binom{2\ell+1}{p+q+1}\sum_{m\geq
0}\binom{2m-p-q}{m-q-i+j}\lambda^{m}$ $\displaystyle=\sum_{p,q\geq
0}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}t^{2\ell-
p-q+2i+2j}\binom{2\ell+1}{p+q+1}\lambda^{q+i-j}\sum_{m\geq
0}\binom{2m-p+q+2i-2j}{m}\lambda^{m}$ $\displaystyle=\sum_{p,q\geq
0}\sum_{i=0}^{p}\sum_{j=0}^{q}\binom{p}{i}\binom{q}{j}t^{2\ell-
p-q+2i+2j}\binom{2\ell+1}{p+q+1}\frac{\lambda^{q+i-j}}{\sqrt{1-4\lambda}}\ \
L^{-p+q+2i-2j}$
$\displaystyle=\frac{t^{2\ell}}{\sqrt{1-4\lambda}}\sum_{p,q\geq
0}\binom{2\ell+1}{p+q+1}\left(\lambda
Lt^{+}\frac{1}{tL}\right)^{p}\left(\frac{\lambda L}{t}+\frac{t}{L}\right)^{q}$
$\displaystyle=\frac{t^{2\ell}}{\sqrt{1-4\lambda}}\frac{(\lambda
Lt+\frac{1}{tL}+1)^{2\ell+1}-(\frac{\lambda
L}{t}+\frac{t}{L}+1)^{2\ell+1}}{(\lambda Lt+\frac{1}{tL})-(\frac{\lambda
L}{t}+\frac{t}{L})},$
where $L=\frac{1-\sqrt{1-4\lambda}}{2\lambda}$. It follows that
$\frac{1}{L}=\frac{1+\sqrt{1-4\lambda}}{2}$, that
$\displaystyle\lambda Lt+\frac{1}{tL}+1$
$\displaystyle=\frac{1-\sqrt{1-4\lambda}}{2}t+\frac{1+\sqrt{1-4\lambda}}{2t}+1$
$\displaystyle=\frac{1}{2t}\left(\sqrt{1-4\lambda}(1-t^{2})+(t+1)^{2}\right)$
$\displaystyle=\frac{t+1}{2t}\left(\sqrt{1-4\lambda}(1-t)+(t+1)\right),$
and that
$\displaystyle\frac{\lambda L}{t}+\frac{t}{L}+1$
$\displaystyle=\frac{1-\sqrt{1-4\lambda}}{2t}+\frac{1+\sqrt{1-4\lambda}}{2}t+1$
$\displaystyle=\frac{1}{2t}\left(\sqrt{1-4\lambda}(t^{2}-1)+(t+1)^{2}\right)$
$\displaystyle=\frac{t+1}{2t}\left(\sqrt{1-4\lambda}(t-1)+(t+1)\right).$
Hence,
$\left(\lambda Lt+\frac{1}{tL}\right)-\left(\frac{\lambda
L}{t}+\frac{t}{L}\right)=\left(\lambda
Lt+\frac{1}{tL}+1\right)-\left(\frac{\lambda
L}{t}+\frac{t}{L}+1\right)=\frac{1-t^{2}}{t}\sqrt{1-4\lambda}.$
Thus
$\displaystyle\sum_{m\geq 0}G_{\ell,m}\lambda^{m}$
$\displaystyle=\frac{(t+1)^{2\ell+1}}{2^{2\ell+1}(1-t^{2})(1-4\lambda)}\left(\left(\sqrt{1-4\lambda}(1-t)+(t+1)\right)^{2\ell+1}+\left(\sqrt{1-4\lambda}(1-t)-(t+1)\right)^{2\ell+1}\right)$
$\displaystyle=\frac{(t+1)^{2\ell+1}}{2^{2\ell}(1-t^{2})(1-4\lambda)}\sum_{k\geq
0}\binom{2\ell+1}{2k}(\sqrt{1-4\lambda}(1-t))^{2\ell+1-2k}(t+1)^{2k}$
$\displaystyle=\frac{(t+1)^{2\ell}(1-t)^{2\ell}}{2^{2\ell}\sqrt{1-4\lambda}}\sum_{k\geq
0}\binom{2\ell+1}{2k}(1-4\lambda)^{\ell-k}\left(\frac{t+1}{1-t}\right)^{2k}$
$\displaystyle=4^{-\ell}(t+1)^{2\ell}(1-t)^{2\ell}\sum_{a\geq
0}\binom{2a}{a}\lambda^{a}\sum_{k\geq
0}\binom{2\ell+1}{2k}(1-4\lambda)^{\ell-k}\left(\frac{t+1}{1-t}\right)^{2k}$
$\displaystyle=4^{-\ell}(t+1)^{2\ell}(1-t)^{2\ell}\sum_{a\geq
0}\binom{2a}{a}\lambda^{a}\sum_{k\geq
0}\binom{2\ell+1}{2k}\sum_{i=0}^{\ell-k}\binom{\ell-k}{i}(-4\lambda)^{i}\left(\frac{t+1}{1-t}\right)^{2k}.$
Since $G_{\ell,\ell}$ is the coefficient of $\lambda^{\ell}$ in $\sum_{m\geq
0}G_{\ell,m}\lambda^{m}$, we have
$G_{\ell,\ell}=4^{-\ell}(t+1)^{2\ell}(1-t)^{2\ell}\sum_{a\geq 0}\sum_{k\geq
0}\binom{2a}{a}\binom{2\ell+1}{2k}\binom{\ell-k}{\ell-a}(-4)^{\ell-a}\left(\frac{t+1}{1-t}\right)^{2k}.$
Additionally, from Lemma 6.1,
$\displaystyle G_{\ell,\ell}$
$\displaystyle=4^{-\ell}(t+1)^{2\ell}(1-t)^{2\ell}\sum_{k=0}^{\ell}(-1)^{\ell-k}\frac{2\ell+1}{2\ell-2k+1}\binom{\ell}{k}\binom{2\ell}{\ell}\left(\frac{t+1}{1-t}\right)^{2k}$
$\displaystyle=4^{-\ell}(t+1)^{2\ell}(1-t)^{2\ell}\sum_{k=0}^{\ell}(-1)^{k}\frac{2\ell+1}{2k+1}\binom{\ell}{k}\binom{2\ell}{\ell}\left(\frac{t+1}{1-t}\right)^{2\ell-2k}.$
Moreover, from Lemma 6.2,
$\displaystyle\sum_{\ell\geq 0}G_{\ell,\ell}\lambda^{\ell}$
$\displaystyle=\sum_{\ell\geq
0}4^{-\ell}(t+1)^{2\ell}(1-t)^{2\ell}\lambda^{\ell}\sum_{k=0}^{\ell}(-1)^{k}\frac{2\ell+1}{2k+1}\binom{\ell}{k}\binom{2\ell}{\ell}\left(\frac{t+1}{1-t}\right)^{2\ell-2k}$
$\displaystyle=\sum_{k\geq
0}(-1)^{k}\left(\frac{t+1}{1-t}\right)^{-2k}\sum_{\ell\geq
0}\frac{2\ell+1}{2k-1}\binom{\ell}{k}\binom{2\ell}{\ell}\left(\frac{\lambda}{4}(t+1)^{4}\right)^{\ell}$
$\displaystyle=\sum_{k\geq
0}(-1)^{k}\left(\frac{t+1}{1-t}\right)^{-2k}\binom{2k}{k}\frac{\left(\frac{\lambda}{4}(t+1)^{4}\right)^{k}}{\left(1-(t+1)^{4}\lambda\right)^{k+\frac{3}{2}}}.$
$\displaystyle=\frac{1}{\left(1-(t+1)^{4}\lambda\right)^{\frac{3}{2}}}\sum_{k\geq
0}\binom{2k}{k}\left(-\left(\frac{1-t}{t+1}\right)^{2}\frac{\frac{\lambda}{4}(t+1)^{4}}{1-(t+1)^{4}\lambda}\right)^{k}$
$\displaystyle=\frac{1}{\left(1-(t+1)^{4}\lambda\right)^{\frac{3}{2}}}\sum_{k\geq
0}\binom{2k}{k}\left(-\frac{\lambda(t+1)^{2}(1-t)^{2}}{4(1-(t+1)^{4}\lambda)}\right)^{k}$
$\displaystyle=\frac{1}{\left(1-(t+1)^{4}\lambda\right)^{\frac{3}{2}}\sqrt{1+\frac{\lambda(t+1)^{2}(1-t)^{2}}{1-(t+1)^{4}\lambda}}}$
$\displaystyle=\frac{1}{(1-(t+1)^{4}\lambda)\sqrt{1-4t(t+1)^{2}\lambda}},$
as desired. ∎
###### Proof of Theorem 4.4.
The equations
$\displaystyle\gamma(\Sigma(\Gamma(n+1);t))=$
$\displaystyle\sum_{\ell=0}^{\lfloor
n/2\rfloor}(2t)^{n-2\ell-1}\left(2\binom{n}{2\ell}t+\binom{n}{2\ell+1}\right)\sum_{a=0}^{\ell}\binom{2a}{a}t^{a}$
$\displaystyle=\sum_{m=0}^{n}\binom{n}{m}(2t)^{n-m}\sum_{a=0}^{\lfloor
m/2\rfloor}\binom{2a}{a}t^{a}$
follow from substituting the equation from Theorem 6.4 into equation (4.1). ∎
## References
* [1] Matthias Beck and Sinai Robins. Computing the continuous discretely. Undergraduate Texts in Mathematics. Springer, New York, 2007. Integer-point enumeration in polyhedra.
* [2] Benjamin Braun. Unimodality problems in Ehrhart theory. In Recent trends in combinatorics, volume 159 of IMA Vol. Math. Appl., pages 687–711. Springer, 2016.
* [3] Tianran Chen. Directed acyclic decomposition of Kuramoto equations. Chaos, 29(9):093101, 12, 2019.
* [4] Tianran Chen, Robert Davis, and Evgeniia Korchevskaia. Facets and facet subgraphs of symmetric edge polytopes. Discrete Appl. Math., 328:139–153, 2023.
* [5] Tianran Chen, Robert Davis, and Dhagash Mehta. Counting equilibria of the Kuramoto model using birationally invariant intersection index. SIAM J. Appl. Algebra Geom., 2(4):489–507, 2018.
* [6] Alessio D’Alì, Emanuele Delucchi, and Mateusz Michałek. Many faces of symmetric edge polytopes. Electron. J. Combin., 29(3):Paper No. 3.24, 42, 2022.
* [7] Alessio D’Alì, Martina Juhnke-Kubitzke, and Melissa Koch. On a generalization of symmetric edge polytopes to regular matroids, 2023\. arXiv:2307.04933.
* [8] Takayuki Hibi and Akiyoshi Tsuchiya. Reflexive polytopes arising from perfect graphs. Journal of Combinatorial Theory, Series A, 157:233–246, 2018.
* [9] Akihiro Higashitani, Katharina Jochemko, and Mateusz Michałek. Arithmetic aspects of symmetric edge polytopes. Mathematika, 65(3):763–784, 2019.
* [10] Tamás Kálmán and Lilla Tóthmérész. $h^{*}$-vectors of graph polytopes using activities of dissecting spanning trees, 2023. arXiv:2203.17127.
* [11] Tetsushi Matsui, Akihiro Higashitani, Yuuki Nagazawa, Hidefumi Ohsugi, and Takayuki Hibi. Roots of Ehrhart polynomials arising from graphs. J. Algebr. Comb., 34(4):721–749, 2011.
* [12] Hidefumi Ohsugi and Takayuki Hibi. Centrally symmetric configurations of integer matrices. Nagoya Math. J., 216:153–170, 2014.
* [13] Hidefumi Ohsugi and Akiyoshi Tsuchiya. The $h^{*}$-polynomials of locally anti-blocking lattice polytopes and their $\gamma$-positivity. Discrete Comput. Geom., 66(2):701–722, 2021.
* [14] James G. Oxley. Matroid theory. Oxford Science Publications. The Clarendon Press, Oxford University Press, New York, 1992.
* [15] Richard P. Stanley. Combinatorics and commutative algebra, volume 41 of Progress in Mathematics. Birkhäuser Boston, Inc., Boston, MA, second edition, 1996.
* [16] Bernd Sturmfels. Gröbner bases and convex polytopes, volume 8 of University Lecture Series. American Mathematical Society, Providence, RI, 1996.
* [17] Lilla Tóthmérész. A geometric proof for the root-independence of the greedoid polynomial of eulerian branching greedoids, 2022. arXiv:2204.12419.
|
# Segmentation of Breast Microcalcifications: a Multi-Scale Approach
Chrysostomos Marasinou, Bo Li, Jeremy Paige, Akinyinka Omigbodun, Noor
Nakhaei,
Anne Hoyt, William Hsu, Senior Member, IEEE This work was supported in part
by the National Science Foundation under Grant No. 1722516 and the Department
of Radiological Sciences through the generous support of the Iris Cantor
Foundation.C. Marasinou, L. Bo, J. Paige, A. Omigbodun, A. Hoyt, and W. Hsu
are with the David Geffen School of Medicine at UCLA, Department of
Radiological Sciences, Los Angeles, CA 90024, United States of America (email:
[email protected])N. Nakhaei is with the University of California, Los
Angeles, Department of Computer Science, Los Angeles, CA 90024, United States
of America
###### Abstract
Accurate characterization of microcalcifications (MCs) in 2D full-field
digital screening mammography is a necessary step towards reducing diagnostic
uncertainty associated with the callback of women with suspicious MCs.
Quantitative analysis of MCs has the potential to better identify MCs that
have a higher likelihood of corresponding to invasive cancer. However,
automated identification and segmentation of MCs remains a challenging task
with high false positive rates. We present Hessian Difference of Gaussians
Regression (HDoGReg), a two stage multi-scale approach to MC segmentation.
Candidate high optical density objects are first delineated using blob
detection and Hessian analysis. A regression convolutional network, trained to
output a function with higher response near MCs, chooses the objects which
constitute actual MCs. The method is trained and validated on 435 mammograms
from two separate datasets. HDoGReg achieved a mean intersection over the
union of 0.670$\pm$0.121 per image, intersection over the union per MC object
of 0.607$\pm$0.250 and true positive rate of 0.744 at 0.4 false positive
detections per $cm^{2}$. The results of HDoGReg perform better when compared
to state-of-the-art MC segmentation and detection methods.
###### Index Terms:
breast cancer, full-field digital mammography, microcalcifications,
segmentation
## I Introduction
Breast cancer is the most prevalent cancer among women. More than two million
new cases and over 620,000 deaths were estimated for 2018 worldwide [1].
Studies have shown that early detection using mammography reduces breast
cancer mortality [2]. In many countries, screening programs have been
established with sensitivity levels for detecting cancers ranging between
80-95% [3, 4]. However, screening also results in false positive outcomes,
including the identification of clinically insignificant cancers, raising
concerns about overdetection.
Microcalcifications (MCs), which are small calcium deposits, are common
mammographic findings where they typically appear as bright structures. Nearly
50% of the biopsied MCs are associated with ductal carcinoma in situ (DCIS)
[5], a largely benign form of cancer but a nonobligate precursor to invasive
cancer [6, 7]. MCs are reported by radiologists using a set of qualitative
descriptors related to shape (morphology) and distribution, as defined by the
American College of Radiology Breast Imaging Reporting and Data System (ACR
BI-RADS). Descriptors correspond to varying levels of suspicion for cancer.
For example, amorphous MCs are assigned a moderate suspicion level (i.e., BI-
RADS 4B) with a positive predictive value (PPV) of 21% [8]. However,
interpretation of MCs varies by radiologist due to subtle differences in their
size, shape, texture, and inhomogeneity of the background tissue [9]. Hence,
determining whether a group of calcifications is malignant is challenging, and
the current PPV of assessing suspicious MCs on 2D mammography is in the range
of 20-41% [10].
Many computerized methods have been developed to aid radiologists in detecting
MCs [11, 12, 13, 14, 15, 9]. These methods, generally categorized as computer-
aided detection (CADe) systems, automatically mark groups of suspicious MCs in
mammograms. Current CADe systems achieve high sensitivity but at the cost of a
large number of false positive marks per mammogram, increasing the
interpretation time. Furthermore, most of the automated methods identify
regions with suspicious MCs but do not delineate the exact shape of individual
MCs. Studies have shown that using shape and intensity features from segmented
MCs can improve malignancy classification [16, 17, 18, 19, 20]. Precise
segmentation also allows for more accurate quantitative characterization of
shape and distribution of MCs and texture analysis of the surrounding breast
parenchyma that could be used to classify cancerous regions better.
In this paper, a quantitative approach for characterizing MCs based on their
shape is demonstrated. Given a 2D full-field digital mammogram, we initially
identify bright salient structures using the difference of Gaussians’ (DoG)
blob detection algorithm. Hessian analysis is then applied to segment these
structures. Next, dense regression is applied to segment regions containing
structures that are likely to be MCs. Dense regression has been used for
similar tasks such as cell & nuclei detection [21, 22], retinal optical disc
and fovea detection [23], and focal vascular lesion localization on brain MRI
[24]. The idea is that human experts’ reference annotations are mapped to a
smooth proximity function that reaches its maximum value when corresponding to
the annotated points. Dense regression models are then trained to map the
input mammogram to the proximity function. The proximity function method is
advantageous when the manual annotations are incomplete (e.g., impractical to
annotate every possible MC) or when objects are annotated by a single pixel
rather than their actual boundaries (e.g., many MCs are tiny and time-
consuming to delineate the boundary). To perform dense regression, a
pretrained, fully convolutional network with pretrained weights is utilized.
The outputs of the dense regression model and the blob segmentation algorithm
are combined to generate the final MC segmentation.
The novelty of our work is summarized as follows:
* •
Proximity functions represent individual MCs, providing a way to accommodate
uncertainty in boundaries as part of classifier training.
* •
A dense regression model and a novel blob segmentation algorithm are applied
to generate MCs’ accurate segmentation while achieving fewer false positives
than other state-of-the-art algorithms.
### I-A Prior Work
Detection of MCs generally occurs in two steps: (i) detection of individual
MCs, (ii) identification of clusters by grouping individual MCs following a
set of spatial clustering rules. For detecting individual MCs, various
approaches have been investigated, ranging from traditional image processing
techniques to deep learning methods. [11] used a DoG filter for spot detection
and shape reconstruction by morphological filtering. [12] used 2D wavelet
transform to suppress the low frequency background and enhance small signals.
[13] applied a multiscale spot detector based on Laplacian filters and size
and contrast estimation based on filter response. [14] used support vector
machines with the input being local image windows to classify locations as MCs
or non-MCs. [15] extracted local features using a predefined filter bank and
applied supervised learning with a boosting classifier. Most of these methods
focused on detecting the point locations of MCs, disregarding the task of
extracting their shape. More recently, [9] developed a context-sensitive deep
convolutional neural network to classify MCs’ presence at locations. In their
approach, candidate locations were generated using DoG blob detection,
filtering out non-MCs using a convolutional neural network. In contrast, our
study performs MC segmentation (i.e., each MC is characterized by its shape
and spatial location).
Several studies have examined MC segmentation. Techniques include wavelet
transform to isolate high frequency components [25, 26], gray-level
morphological operations [11, 27, 28, 29, 30], fuzzy logic [31], and machine
learning for binary pixel classification [17]. Although segmentation of MCs is
performed in these studies, all but one [30]) evaluated the performance of
their algorithm as a detection task, not a segmentation task. Reference [30]
reported an intersection over the union of 70.8% between the segmented MCs and
the radiologist annotations on a set of 200 regions. We have reimplemented
their approach, comparing its performance to our HDoGReg approach using the
same dataset.
### I-B Paper Organization
The rest of the paper is organized as follows. In section II, we present the
data utilized for training and validating our model. In section III, we
formulate the problem and describe the methodology of Hessian Difference of
Gaussians Regression (HDoGReg) and the corresponding evaluation metrics. The
experiments for validating our models are described in section IV, where the
comparison methods are given. In section V, we present and discuss the
results. Finally, conclusions are drawn in section VI.
## II Data
Two distinct full-field digital mammography datasets were used in this work:
The publicly available INbreast dataset and cases seen at our institution
(UCLA).
INbreast: For model training and internal validation, we utilized a public
dataset called INbreast [32], a collection of 2D mammograms generated using a
Siemens MammoNovation system. A total of 115 screening cases with 410 images
were collected at a 0.070 mm per pixel resolution and 14-bit greyscale. The
dataset included detailed annotations provided by two experts for several
types of lesions (i.e., masses, MCs, asymmetries, and distortions). Fifty-six
cases had pathology-confirmed diagnoses, out of which 45 were cancerous (DCIS
and invasive). We used 294 images with annotations of individual MCs. MCs were
annotated in two ways: (1) small MCs were annotated by a single pixel to
denote their location and (2) larger MCs were annotated using pixel-wise
contours. It should be noted that the guideline of what was considered small
versus larger MC was not reported [32].
UCLA: As an additional external test set, we utilized data collected
retrospectively from patients who had a mammogram performed by our
institution, following an institutional review board-approved protocol. The
dataset consisted of 79 diagnostic cases with 141 full-field digital images
where MCs were present. All images were acquired using Hologic Selenia full-
field digital mammography devices at a 0.070 mm per pixel resolution and
12-bit greyscale. After collecting the data, suspicious MCs were annotated by
a breast fellowship-trained, board-certified radiologist. An open-source
medical image viewer, Horos, was utilized to generate the annotations.
Individual MCs were annotated by single pixels indicating their locations. To
assess the annotation task’s interreader reliability, a sample of 5 cases was
annotated by a second board-certified radiologist. The index of specific
agreement and the kappa statistic were determined. The two radiologists’
agreement was moderate, with an index of specific agreement of 0.664
(0.606-0.729, 95% confidence intervals), see supplementary
material111Supplementary material will be made available when published.
## III Methods
The overall approach to MC segmentation, HDoGReg, is illustrated in Figure 1.
Figure 1: HDoGReg: Approach for segmenting individual MCs. While the
segmentation is performed on entire 2D full-field mammograms, for
visualization purposes, a small patch is shown. In the upper branch, blob
segmentation is performed to segment bright blob-like and tubular structures.
In the lower branch, a regression convolutional neural network gives a
continuous function with higher response close to MCs. A threshold is then
applied to segment regions where MCs are likely present. The two branches’
output is combined based on an overlap criterion, resulting in the final
segmentation mask.
### III-A Blob segmentation
The first stage in HDoGReg is the segmentation of granular structures that are
candidate MCs. To generate candidate MC segments, we developed Hessian DoG for
blob segmentation. This module’s objective is to obtain an accurate
segmentation of bright salient structures that are candidate MC objects, as
shown in Figure 1.
Scale-space theory is a framework formulated to represent signals at multiple
scales. The Gaussian scale-space representation of an image $I(x,y)$ is
defined as [33]:
$\displaystyle L(x,y;\sigma)=G(x,y;\sigma)\ast I(x,y),$ (1)
where $\ast$ is the convolution and $G(x,y,\sigma)$ the two-dimensional
Gaussian function
$\displaystyle G(x,y;\sigma)$
$\displaystyle=\frac{1}{2\pi\sigma^{2}}e^{-(x^{2}+y^{2})/(2\sigma^{2})}$ (2)
In the DoG method, blobs with associated scale levels are detected from scale-
space maxima of the scale-normalized DoG function. The normalized DoG function
is defined as:
$\displaystyle DoG(x,y;\sigma)$
$\displaystyle=\frac{\sigma}{\Delta\sigma}\left(L(x,y;\sigma+\Delta\sigma)-L(x,y;\sigma)\right)$
(3)
where $\Delta\sigma$ is the difference between two scales. To construct the
DoG scale-space representation, a sequence of scales is considered
$\sigma_{n}=k^{n}\sigma_{\textrm{min}}$ where $k$ is a constant multiplicative
factor and $n=[0,1,\cdots,n_{\textrm{max}}]$. The DoG representations (3) are
computed for all adjacent scales (i.e.,
$\Delta\sigma=\sigma_{n+1}-\sigma_{n}$) forming a 3-dimensional
representation:
$\displaystyle DoG(x,y,n)$
$\displaystyle=\frac{\sigma_{n}}{\sigma_{n+1}-\sigma_{n}}\left(L(x,y;\sigma_{n+1})-L(x,y;\sigma_{n})\right)$
(4)
with $x,y$ the two spatial dimensions and $n=[0,1,\cdots,n_{\max-1}]$ a scale
dimension. Local-maxima in the 3-dimensional representation are computed
giving a blob set $(x^{(i)},y^{(i)},\sigma^{(i)})$ where $i$ identifies each
blob. The number of blob detections is controlled by a threshold,
$T_{\textrm{DoG}}$, that is applied as a lower bound on the stacked
representation before obtaining the local maxima. Moreover, in the case of
overlapping blobs, the smaller blob is eliminated if the overlapping fraction
is greater than the threshold $O_{\textrm{DoG}}$.
We extended this method using Hessian analysis to achieve blob segmentation.
The geometrical structure of a blob-like object can be described by the
eigenvalues of the Hessian [34]. In particular, a bright blob-like structure
corresponds to two negative and large eigenvalues, whereas a bright tubular
structure corresponds to one large negative eigenvalue and a small eigenvalue
of an arbitrary sign. These structures correspond to the target MC candidates.
The Hessian of the DoG representation at scale $\sigma$ is given by
$\displaystyle\textrm{HDoG}(x,y;\sigma)=\begin{pmatrix}\frac{\partial^{2}DoG(x,y;\sigma)}{\partial
x^{2}}&\frac{\partial^{2}DoG(x,y;\sigma)}{\partial x\partial y}\\\
\frac{\partial^{2}DoG(x,y;\sigma)}{\partial x\partial
y}&\frac{\partial^{2}DoG(x,y;\sigma)}{\partial y^{2}}\end{pmatrix}$ (5)
HDoG is computed across all scales in the sequence $\sigma_{n}$. At each scale
the following constraints are imposed:
$\displaystyle\textrm{tr}(H)<0\quad\land\quad\left(\det(H)<0\quad\lor\quad\frac{|\det(H)|}{\textrm{tr}(H)^{2}}\leq
h_{\textrm{thr}}\right)$ (6)
where $h_{\textrm{thr}}$ is a tunable parameter. The constraints ensure that
the Hessian is either negative definite or has a small positive eigenvalue. In
this way, only bright salient blob-like and tubular structures are segmented.
The constraint generates a binary mask at each scale. Iterating over the blob
set found in the DoG algorithm $(x^{(i)},y^{(i)},\sigma^{(i)})$, the
corresponding objects are found in the Hessian masks. More specifically, for
the Hessian mask at scale $\sigma^{(i)}$, the object spanning the location
$(x^{(i)},y^{(i)})$ is found. The final blob segmentation result is comprised
of all detected objects merged into a single binary mask.
Our HDoG algorithm works on a predetermined scale range
$[\sigma_{\min},\sigma_{\max}]$, which was fine-tuned on the validation set to
achieve a high true positive rate (>90%). The segmented objects from this
phase were considered candidate MC objects and filtered in the subsequent step
to reduce the number of false positives.
### III-B Regression Convolutional Neural Network
Figure 2: FPN with Inception-v4 encoder used for regression. In the encoding
branch the image is processed with the Inception-v4 classification network.
Skip connections (dashed lines) are inserted after layers where the output was
reduced in spatial size by factors of 4, 8, 16 and 32 respectively. The skip
connections feed FPN blocks where they are processed by convolutions. The
outputs are up-sampled independently by factors of 1, 2, 4 and 8 respectively.
Their outputs are added and inserted in a spatial dropout layer activated only
during training for regularization purposes. After dropout, convolutions are
followed by an up-sampling by a factor of 4 to match the original image size
and the sigmoid activation function.
In the second stage, we identified regions where MCs are most likely to be
present in the mammogram. The task here was to segment the MCs’ area and use
the segmentation to choose relevant MC objects from the previous stage. We
performed this task using a fully convolutional neural network, previously
used in image segmentation, as a regression model. The model’s output is a
smooth proximity map reaching a maximum value at the predicted MC locations.
Segmentation approaches commonly perform binary classification on each pixel.
However, applying this method requires precise pixel-wise annotations for all
objects of interest appearing in our images. As previously described,
generating precise annotations of MCs is challenging, and the reference
annotations only provide the point locations of MCs, not their exact
boundaries.
The task of MC region segmentation is analogous to that of cell & nuclei
detection in microscopy images. The two tasks share the following
characteristics: (1) they are highly imbalanced ( i.e., the positive class
[MCs or cells/nuclei] captures a small region compared to the background
within an image, and it often consists of many small structures), (2) the
background is highly inhomogeneous, (3) individual objects (MCs or
cells/nuclei) exhibit large variation in sizes, shapes, and textures, (4)
boundaries of the structures are often blurry, (5) the resolution of both
types of images is large, and (6) the annotations are usually a mixture of
individual points or exact boundaries.
Inspired by this analogy, we adapted methods previously used in cell & nuclei
detection [21, 22]. In [21], the authors proposed regression as a method for
detecting cell centers. The human-annotated binary masks containing cell
centers’ locations were transformed into a continuous function flat on the
background with localized peaks at each cell’s center. These functions were
then used to train a Random Forest Regression algorithm on a set of image
patches. The cell centers were identified with a local maxima in the model’s
output. In [22], the authors showed that the same technique could be applied
using a deep learning model. Their regression model was a fully convolutional
neural network with a large receptive field capable of encoding high-
resolution information.
Our MC segmentation model is formulated as follows: Given a mask generated
from reference annotations $M(x,y)\in\\{0,1\\}$, the MC locations are given by
$\\{(x_{i},y_{i})\\}$ where $M(x_{i},y_{i})=1$. The proximity function is then
defined as:
$\displaystyle P(x,y)$ $\displaystyle=\max_{i}g(x,y,x_{i},y_{i})$ (7)
$\displaystyle g(x,y,x_{i},y_{i})$
$\displaystyle=\begin{cases}(e^{\alpha(1-r/\xi)}-1)/(e^{\alpha}-1),&r\leq\xi\\\
0,&r>\xi\end{cases}$ (8) $\displaystyle r$
$\displaystyle=\sqrt{(x-x_{i})^{2}+(y-y_{i})^{2}}$ (9)
where $\alpha,\xi$ are tunable parameters. The function maps MC locations on
an exponentially curved surface, expanding to a distance $\xi$ with decay rate
$\alpha$ before it vanishes. An example of the transformation is illustrated
in Figure 3. This transformed mask compensated for the fact that we had mixed
quality annotations (i.e., point-like and exact) and forced the model to learn
information from both the precise locations of MCs and the surrounding
background.
(a)
(b)
(c)
Figure 3: (a) A mammographic image patch which includes MCs; (b) The
corresponding annotation mask; (c) The corresponding proximity function map
with parameters $\xi=10$ and $\alpha=1$.
We constructed a model which predicts the proximity function $P(x,y)$ given
the image $I(x,y)$. A feature pyramid network (FPN) [35] was used with
Inception-v4 [36] as the backbone. The FPN architecture was introduced for
applications such as region proposal, object detection, and instance
segmentation. It adopts a pyramidal shape structure similar to many
segmentation networks, such as the U-net [37], with an encoder that produces
semantic features at different scales and a decoder that combines the encoder
features by upsampling them.
The FPN structure is a suitable architecture because it allows features from
all scales to contribute to the final prediction independently. An
illustration of the network is presented in Figure 2. The network consists of
an encoding and a decoding branch. In the encoding branch, the Inception-v4
architecture was adopted with weights pretrained on ImageNet [38]. Features
were extracted at four different scales (down-sampled compared to the original
image by factors of 4, 8, 16, and 32). The features were then transferred to
the decoding branch via skip connections. They were upsampled by factors of 1,
2, 4, and 8, respectively, to match their spatial sizes. The resulting
features were aggregated using addition and further upsampled to match the
image size. The number of output channels was set to 1 and passed through a
sigmoid function to generate a value between 0 and 1. This value was
thresholded to achieve the final segmentation.
The model was trained using a soft Dice loss function, which was introduced as
an optimization objective in biomedical segmentation applications [39, 40].
The formulation in [40] was used:
$\displaystyle
L_{\textrm{DICE}}(\hat{P},P)=1-\frac{2\sum\limits_{x,y}P(x,y)\hat{P}(x,y)+\epsilon}{\sum\limits_{x,y}(P(x,y)+\hat{P}(x,y))+\epsilon}$
(10)
with $\epsilon$ set to 1 where $\epsilon$ was introduced for numerical
stability and $P$ and $\hat{P}$ correspond to the target and predicted
proximity map, respectively. A segmentation binary mask was generated by
applying a cut-off on the resulting proximity mask, $\hat{P}(x,y)\geq p_{thr}$
where $p_{thr}\in[0,1]$.
### III-C Combining approach
HDoG resulted in the segmentation of bright objects, whereas the regression
CNN outputted a mask of MC regions. These outputs were combined to achieve the
segmentation of individual MCs. The objects identified with HDoG were
candidate MC objects, and their shape had to remain intact. Our approach chose
a subset of these candidates in their entirety. We retained the HDoG objects
that have an overlap $\leq o_{\textrm{thr}}$ with the CNN region mask, where
$o_{\textrm{thr}}$ is a tunable percentage parameter.
### III-D Evaluation metrics
The performance of HDoGReg was assessed using Intersection over the Union
(IoU). We defined IoU per object as the averaged IoU between each reference
annotation object and the object that has the most overlap with in the
prediction mask222MC objects annotated by a single pixel were disregarded in
computing IoU per MC object. IoU is not well-defined in such cases.. To
evaluate the image-wise segmentation, the mean IoU between the background and
the positive MC class per image was computed. The IoU per MC object was
measured to examine the performance of segmenting individual MCs.
HDoGReg was also evaluated as a MC detection task using Free-Response
Operating Characteristic (FROC) analysis, similar to prior work [14, 9]. In
FROC analysis, the true positive detection rate was contrasted with false
positive detection per image. The analysis required the definition of
localization rules to determine true positives. We defined a detected object
as a true positive if its distance from a ground truth object was at most 5
pixels (0.35 mm)333Centroids of individual objects were used in computing
their distance. or if it demonstrated an IoU value of at least 0.3 with a
ground truth object.
## IV Experiments
(a)
(b)
(c)
(d)
Figure 4: Five 256x256 patches extracted from different diagnostic mammograms
showing the results of HDoGReg and a comparison method. From left to right:
(a) unannotated images, (b) reference annotations, (c) results using HDoGReg,
and (d) results using [30]. The first three rows are from INbreast data and
the last two are from UCLA data. For better visualization the patches were
normalized. Note the inherent difference in the appearance of the mammograms
between INbreast and UCLA data.
### IV-A Training, Validation, and Test Sets
We partitioned the INbreast dataset into a training set with 51 cases (173
images), a validation set with 17 cases (56 images), and a test set with 18
cases (65 images). The UCLA external test dataset was partitioned into a
training set used for fine-tuning the model with 112 images and a held-out
test set with 29 images. We used the INbreast validation set to fine-tune
HDoGReg, and the INbreast test set to assess the performance. Cases were kept
independent (i.e., all images from an individual case were included within the
same subset) to avoid potential bias.
### IV-B Blob Segmentation Optimization
With HDoG blob segmentation, our goal was to achieve a maximized IoU per
object and a true positive rate (TPR) higher than 90%. We extracted
representative patches from the INbreast validation set and optimized HDoG
hyperparameters on them to satisfy the above criteria.
Through experimentation on the validation set, we set $\sigma_{\min}=1.18$,
$\sigma_{\max}=3.1$, overlapping fraction $O_{\textrm{DoG}}=1$, DoG threshold
$T_{\textrm{DoG}}=0.006$ and Hessian threshold $h_{\textrm{ thr}}=1.4$. On the
validation set, our algorithm achieved an IoU per object 0.648 and a TPR of
0.903 at 136.5 false positives per unit area ($cm^{2}$).
### IV-C Regression Model Training and Tuning
To train the regression model, we extracted patches from the images and
corresponding masks. We applied the sliding window approach with patch size
512 pixels and stride 480 to permit overlapping patches. Only patches with
annotated MCs present were considered. From INbreast cases, a total of 1045
patches were extracted from the training set and 329 from the validation set.
From UCLA cases, a total of 252 patches from the training set were extracted.
The mask patches were transformed using the proximity function map (7). We set
$\xi=\\{6,8,10,12\\}$ for the characteristic distance and
$\alpha=\\{-1,-2,10^{-4},1,2\\}$ for the decay rate. The proximity function
(7) is not well defined when $\alpha=0$.
A regression neural network was trained using the image patches of the
INbreast training set as input and the corresponding proximity function maps
as output. Data augmentation was performed to enrich the training set by
randomly applying horizontal flipping, magnification, spatial translations in
both directions, cropping, contrast enhancement, brightness adjustment, and
gamma correction (details are given in Supplementary Materials). The resulting
patches had size 320x320 pixels. The soft Dice loss was used to compute the
error between target and predicted proximity functions (10). The model was
trained for 40 epochs using the adaptive moment estimation (Adam) optimization
method [41] with batch size 8, learning rate $10^{-4}$, $\beta_{1}=0.9$,
$\beta_{2}=0.999$ and $\epsilon=10^{-8}$. At the end of each epoch, the model
was evaluated on the image patches of the INbreast validation set, with the
average IoU per patch as the metric. The model achieving the highest IoU over
all epochs was kept. The Inception-v4 weights were initialized with weights
pretrained on ImageNet. The rest of the model weights were initialized
randomly following He initialization [42]. The configuration $\xi=10$,
$\alpha=1$ achieved the highest performance for HDoGReg on the validation set.
We updated our HDoGReg model to work with data from our institution by
updating the model with an additional 40 epochs using a combination of patches
from the INbreast and UCLA datasets.
### IV-D Combining Approach Optimization
The combination approach keeps the HDoG candidate objects based on how much
they overlap with the regression model mask. Optimizing the value of
$o_{\textrm{thr}}=\\{0.2,0.3,0.4,0.5,0.6\\}$, $o_{\textrm{thr}}=0.3$ achieved
the highest performance on the validation set.
### IV-E Comparison
We compared HDoGReg against two state-of-the-art methods. For the MC detection
task, we compared our approach to Wang and Yang [9], which used two
subnetworks, focusing on the local features and the other on features
extracted from the background tissue around the location. They reported a
detection performance of 80% true positive fraction (TPF) at a false positive
rate of 1.03 FPs/cm2. We implemented their context-sensitive deep neural
network, which classifies a location as MC or non-MC, training our
implementation on the INbreast dataset. We implemented DoG based on their
reported parameters adjusting the scales to the resolution of our dataset.
We compared our MC segmentation results with the approach presented in [30].
In [30], MC segmentation was performed using morphological operations. In the
first step, morphological operators were applied to the original image to
detect the MCs’ locations. Specifically, a morphological pyramid was generated
using the closing-opening filter. Differences in the pyramid representations
of the original image were obtained and combined using the extended maximum of
the original image and morphological reconstruction. In the second step, the
MC shapes were extracted using watershed segmentation, where the output of the
first step was utilized as a marker. We report the mean IoU per image and the
IoU per object for this approach.
## V Results
### V-A Regression Model Selection
The entire segmentation pipeline was evaluated using the INbreast validation
set. The FROC analysis is presented in Table I, and the mean IoU per image and
IoU per object are summarized in Table II. For the FROC analysis, a total of
100 bootstrap samples were used to find the partial area under the curve
(pAUC) in each experiment. The pAUC was computed for the range between 0 and 1
FPs per unit area, and the 95% confidence interval is reported. For the
computation of the segmentation metrics, a threshold on the predicted
proximity function was applied. To determine the optimal threshold for each
experiment, we referred to the corresponding FROC curve and found the point
closest to TPR 1 and false positives per unit area 0. All configurations
performed similarly in terms of the FROC analysis and segmentation metrics. We
chose the model with the highest mean value of the FROC pAUC with $\xi=10$ and
$\alpha=1$.
TABLE I: Individual MC FROC pAUC values for validating different regression models. The highest pAUC is bolded. $\alpha/\xi$ | 6 | 8 | 10 | 12
---|---|---|---|---
-2 | 0.804$\pm$0.048 | 0.812$\pm$0.049 | 0.793$\pm$0.040 | 0.773$\pm$0.045
-1 | 0.783$\pm$0.058 | 0.808$\pm$0.047 | 0.790$\pm$0.047 | 0.794$\pm$0.045
10-4 | 0.813$\pm$0.054 | 0.783$\pm$0.056 | 0.790$\pm$0.044 | 0.784$\pm$0.048
1 | 0.799$\pm$0.054 | 0.776$\pm$0.056 | 0.819$\pm$0.046 | 0.790$\pm$0.053
2 | 0.775$\pm$0.056 | 0.791$\pm$0.055 | 0.802$\pm$0.049 | 0.789$\pm$0.053
TABLE II: Segmentation results of different regression models on validation
set. The highest IoUs are bolded. Mean IoU per image
---
$\alpha/\xi$ | 6 | 8 | 10 | 12
-2 | 0.593$\pm$0.109 | 0.590$\pm$0.102 | 0.576$\pm$0.088 | 0.560$\pm$0.066
-1 | 0.590$\pm$0.105 | 0.579$\pm$0.094 | 0.572$\pm$0.089 | 0.566$\pm$0.077
10-4 | 0.596$\pm$0.115 | 0.584$\pm$0.098 | 0.586$\pm$0.097 | 0.584$\pm$0.093
1 | 0.619$\pm$0.125 | 0.601$\pm$0.109 | 0.583$\pm$0.105 | 0.587$\pm$0.098
2 | 0.610$\pm$0.123 | 0.588$\pm$0.108 | 0.592$\pm$0.105 | 0.580$\pm$0.094
IoU per Object
$\alpha/\xi$ | 6 | 8 | 10 | 12
-2 | 0.645$\pm$0.207 | 0.645$\pm$0.206 | 0.648$\pm$0.200 | 0.626$\pm$0.228
-1 | 0.647$\pm$0.203 | 0.648$\pm$0.201 | 0.643$\pm$0.208 | 0.640$\pm$0.212
10-4 | 0.648$\pm$0.201 | 0.644$\pm$0.207 | 0.641$\pm$0.214 | 0.644$\pm$0.207
1 | 0.649$\pm$0.200 | 0.646$\pm$0.206 | 0.647$\pm$0.203 | 0.643$\pm$0.208
2 | 0.650$\pm$0.197 | 0.648$\pm$0.202 | 0.649$\pm$0.200 | 0.643$\pm$0.208
### V-B Individual Microcalcifications
Table III reports the segmentation results of HDoGReg on the INbreast
validation and test sets. For comparison, the segmentation results of the
morphological method of Ciecholewski et al. (see IV-E), are also presented.
Using the paired Wilcoxon signed-rank test, we showed that HDoGReg achieved
superior performance in both mIoU per image and IoU per object for both
subsets with $p<0.01$. Figure 4 presents a sampling of model outputs.
TABLE III: Segmentation Results of Final Model on Validation and Test Sets
HDoGReg
---
Metric/dataset | Validation | Test
mean IoU per image | 0.583$\pm$0.105 | 0.670$\pm$0.121
IoU per object | 0.647$\pm$0.203 | 0.607$\pm$0.250
Ciecholewski [30]
Metric/dataset | Validation | Test
mean IoU per image | 0.517$\pm$0.037 | 0.524$\pm$0.034
IoU per object | 0.408$\pm$0.286 | 0.363$\pm$0.278
Figure 5: Individual MC FROC analysis for final HDoGReg model and comparison
with baseline model. Figure 6: Individual MC FROC analysis on UCLA data.
Figure 5 presents the individual MC FROC analysis between HDoGReg and [9]. In
this plot, the true positive detection rate on the y-axis was plotted against
the false positive counts per unit area (1 $cm^{2}$). The FROC analysis was
performed on the INbreast validation and test sets. HDoGReg achieved FROC pAUC
0.819$\pm$0.046 with a TPR of 0.852 at 0.4 false positives per unit area on
the validation set. On the test set, the FROC pAUC was 0.697$\pm$0.078 with a
TPR of 0.744 at 0.4 false positives per unit area. In comparison, our
implementation of the approach in [9] achieved FROC pAUC 0.703$\pm$0.057 and
0.581$\pm$0.072 in the validation and test sets.
Figure 6 shows the detection performance of the final model on the UCLA data.
We compared the performance of the model trained solely on INbreast data and
the model fine-tuned on data from UCLA. The performance of the two models was
comparable since the original model achieved 0.313$\pm$0.109 FROC pAUC and the
fine-tuned model achieved 0.420$\pm$0.107. However, for the range 0.2 to 0.6
FPs per unit area, the fine-tuned model outperformed the original based on
TPR.
### V-C Microcalcification Clustering Analysis
MCs can be clinically significant when they appear in a group or a
distribution that would suggest deposits in a duct. In this section, we
analyzed HDoGReg segmentation masks to obtain MC clusters and assessed the
relationship between the extracted clusters and five predefined groups based
on radiologist-interpreted mammographic distributions. To cluster individual
MCs into groups, we applied a method called Ordering Points To Identify the
Clustering Structure (OPTICS) [43]. OPTICS operates on density-based
considerations to identify clusters among a given set of points in a multi-
dimensional space. OPTICS defines a notion of neighborhood density of points,
based on a threshold for the minimum number of adjacent points within a region
of a fixed radius. To perform the clustering, we mapped all MC objects in our
segmentation mask to their centroid locations and used them as input to
OPTICS. Upon clustering, a total of 24 expert-defined features were extracted
describing each cluster’s size, shape, and density of MCs (details are given
in the supplementary material). The features for all clusters were
standardized and fed into K-means clustering with k=5, corresponding to the
number of predefined groups.
The correspondence between the clustering approach versus radiologist
descriptors was evaluated as follows: (1) A board-certified fellowship-trained
breast radiologist provided labels for mammographic distribution at the image
level. Some (n=13, 10.7%) images were given multiple labels based on the
dominant and secondary distribution types that appeared in the image. (2) Each
detected cluster was assigned into one of five groups using the k-means
algorithm. (3) To associate each point (defined by the features of each
cluster) with a distribution label, we duplicated the clusters appearing in
images with multiple distribution labels and associated each copy with one
distribution label. Thus, we ended up with a list of distributions and their
associated K-means group. (4) We then computed the homogeneity score h [44],
which is a measure describing the purity of all clusters. If all clusters
contain points of a single class, h has the value of 1.0. OPTICS was fine-
tuned using a random grid search of 100 experiments. We hypothesized that the
MC clustering with the highest homogeneity score is the one closest to
capturing the actual mammographic distributions. Our clustering approach
achieved a homogeneity score $h=0.074$ on the INbreast dataset. Using the
extracted quantitative features, we achieved a moderate level of
correspondence to predefined radiologist-provided labels of distribution.
## VI Discussion
We present an approach that combines the difference of Gaussians with Hessian
analysis and dense regression to achieve precise MC segmentation in full-field
2D digital mammograms. To our knowledge, this is one of the first works
applying a fully convolutional architecture for MC segmentation, which permits
concurrent prediction on multiple adjacent locations. The method was trained
and validated on 435 mammograms from two separate datasets. The results showed
that HDoGReg outperformed comparable approaches that have been recently
published. In terms of the FROC analysis using the INbreast dataset, our
method achieved a TPR of 0.744 at 0.4 false positives per unit area in
comparison with 0.35 TPR at the same false positive counts as what was
presented in [9]. On the segmentation task, HDoGReg achieved 0.670 mean IoU
per image and 0.607 IoU per object compared to 0.524 mean IoU per image and
0.363 IoU per object for the morphological approach presented in [30]. The
addition of UCLA data, even when coarsely annotated by a human reader,
improved the performance of HDoGReg. The ability to utilize a mixture of
annotations, including exact segmentation of larger calcifications or
individual points representing the centroid of smaller calcifications, is a
strength of our approach.
While HDoGReg achieved a lower number of false positives compared to other
approaches, the overall number of false positives per image is still high. The
majority of false positives occur near larger calcifications and correspond to
more irregular shapes compared to actual MCs. The irregular detection can be
attributed to the regression model which was designed to segment regions
containing calcifications. In the case of larger calcifications, the segmented
regions span larger areas, increasing the likelihood of retaining false-
positive objects. Additional filtering based on size and shape criteria in
areas where large calcifications are identified could lead to a substantial
false positive reduction. Given that annotating every possible MC is
impractical, our algorithm likely identified MCs that were not annotated by
human readers, contributing to the false positive count. HDoGReg also
undersegments or oversegments in certain scenarios. Undersegmentation occurs
most often in large objects due to: (1) interior regions of objects having
lower intensities that are omitted and (2) incorrect delineation of boundaries
due to subtle contrast differences between the MC and surrounding tissue.
Nevertheless, large calcifications are typically considered benign and not
clinically significant. Their undersegmentation will have little effect on
quantitative features that may predict invasive cancers. Oversegmentation
occurs primarily when bright objects identified with HDoG are close together
and erroneously combined into a single object when only part of it corresponds
to an actual MC.
Several limitations of our approach exist. Labeling all MCs in full-field
mammograms is a time-consuming process and prone to human error and inter-
annotator variability. Hence, our work is limited by the dataset size and by
variations in how MCs are annotated, ranging from point-like annotations to
detailed contours. Our use of a proximity function to reflect the uncertainty
associated with MC annotations allows our approach to be robust to training
data variations. Moreover, the inherent differences in mammograms acquired
with equipment manufactured by different vendors present another challenge.
The UCLA dataset was obtained using equipment manufactured by Hologic while
the public dataset INbreast were obtained using Siemens equipment. The
brightness and contrast levels of the images varied substantially between
manufacturers. Given that the INbreast dataset had four times as many cases as
the UCLA dataset, our model was fine-tuned with a limited number of training
patches from UCLA. Ongoing work includes annotating additional UCLA cases that
would allow us to further fine-tune the model and experiment with different
training strategies to improve the generalizability of our approach.
Currently, a model with an encoder initialized with pretrained weights was
trained first on INbreast data and fine-tuned on mixed UCLA and INbreast data.
Provided additional UCLA annotated data, an approach where mixed data are
utilized from the beginning of training can be examined. In addition, the
significance of using pretrained weights could be explored. Additionally,
magnification views acquired as part of diagnostic mammograms typically
improve the conspicuity of confirmed suspicious or benign regions. One
direction of future research is to improve the conspicuity of findings in
magnification views, potentially yielding more accurate segmentations.
In summary, we presented HDoGReg, a new quantitative approach for MC
segmentation, based on blob segmentation and dense regression. We showed that
HDoGReg achieves better performance in comparison with state-of-the-art MC
segmentation and detection methods. Additionally, we explored how segmented
MCs can be clustered in an unsupervised manner, resulting in five groups that
are loosely related to qualitative descriptors of MC distributions provided by
radiologists. Our results suggest that HDoGReg could be used in large-scale,
multi-center studies to define reliable quantitative descriptions of MCs
associated with malignancy. Our work serves as the basis for improved
quantitative characterization of MCs. A number of shape, intensity, and
texture features can be extracted from individually segmented MCs and used to
yield quantitative descriptors of MC morphology and distribution. We expect
that quantitative descriptors of morphology and distribution may differ from
the subjective labels assigned by radiologists. Further studies are needed to
evaluate the PPV of our quantitative features and their ability to identify
MCs that are associated with invasive cancers.
## Acknowledgment
We gratefully acknowledge the support of the NVIDIA Corporation. Also, we
would like to thank the Breast Research Group at University of Porto for
developing and sharing the INbreast dataset.
## References
* [1] F. Bray, J. Ferlay, I. Soerjomataram, R. L. Siegel, L. A. Torre, and A. Jemal, “Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries,” _CA: A Cancer Journal for Clinicians_ , vol. 68, no. 6, pp. 394–424, 2018.
* [2] M. G. Marmot, D. G. Altman, D. A. Cameron, J. A. Dewar, S. G. Thompson, M. Wilcox, and T. I. U. P. on Breast Cancer Screening, “The benefits and harms of breast cancer screening: an independent review,” _British Journal of Cancer_ , vol. 108, no. 11, pp. 2205–2240, 2013.
* [3] K. Kemp Jacobsen, E. S. O’Meara, D. Key, D. S.M. Buist, K. Kerlikowske, I. Vejborg, B. L. Sprague, E. Lynge, and M. von Euler-Chelpin, “Comparing sensitivity and specificity of screening mammography in the united states and denmark,” _International Journal of Cancer_ , vol. 137, no. 9, pp. 2198–2207, 2015.
* [4] M. v. Euler-Chelpin, M. Lillholm, G. Napolitano, I. Vejborg, M. Nielsen, and E. Lynge, “Screening mammography: benefit of double reading by breast density,” _Breast Cancer Research and Treatment_ , vol. 171, no. 3, pp. 767–776, Oct 2018.
* [5] G. Farshid, T. Sullivan, P. Downey, P. G. Gill, and S. Pieterse, “Independent predictors of breast malignancy in screen-detected microcalcifications: Biopsy results in 2545 cases,” _British Journal of Cancer_ , vol. 105, no. 11, pp. 1669–1675, 2011.
* [6] S. Hofvind, B. F. Iversen, L. Eriksen, B. M. Styr, K. Kjellevold, and K. D. Kurz, “Mammographic morphology and distribution of calcifications in ductal carcinoma in situ diagnosed in organized screening,” _Acta Radiologica_ , vol. 52, no. 5, pp. 481–487, jun 2011.
* [7] R. F. Cox, A. Hernandez-Santana, S. Ramdass, G. McMahon, J. H. Harmey, and M. P. Morgan, “Microcalcifications in breast cancer: Novel insights into the molecular mechanism and functional consequence of mammary mineralisation,” _British Journal of Cancer_ , vol. 106, no. 3, pp. 525–537, 2012.
* [8] E. A. Sickles, D. C. J., B. L. W. _et al._ , “ACR BI-RADS® Mammography. In: ACR BI-RADS® Atlas, Breast Imaging Reporting and Data System,” 2013.
* [9] J. Wang and Y. Yang, “A context-sensitive deep learning approach for microcalcification detection in mammograms,” _Pattern Recognition_ , vol. 78, pp. 12 – 22, 2018.
* [10] L. Wilkinson, V. Thomas, and N. Sharma, “Microcalcification on mammography: Approaches to interpretation and biopsy,” _British Journal of Radiology_ , vol. 90, no. 1069, 2017.
* [11] J. Dengler, S. Behrens, and J. F. Desaga, “Segmentation of Microcalcifications in Mammograms,” _IEEE Transactions on Medical Imaging_ , vol. 12, no. 4, pp. 634–642, 1993.
* [12] H. Yoshida, K. Doi, and R. M. Nishikawa, “Automated detection of clustered microcalcifications in digital mammograms using wavelet processing techniques,” _Medical Imaging 1994: Image Processing_ , vol. 2167, no. May 1994, pp. 868–886, 1994.
* [13] T. Netsch and H. O. Peitgen, “Scale-space signatures for the detection of clustered microcalcifications in digital mammograms,” _IEEE Transactions on Medical Imaging_ , vol. 18, no. 9, pp. 774–786, 1999.
* [14] I. El-Naqa, Y. Yang, M. N. Wernick, N. P. Galatsanos, and R. M. Nishikawa, “A support vector machine approach for detection of microcalcifications,” _IEEE Transactions on Medical Imaging_ , vol. 21, no. 12, pp. 1552–1563, 2002\.
* [15] A. Oliver, A. Torrent, X. Lladó, M. Tortajada, L. Tortajada, M. Sentís, J. Freixenet, and R. Zwiggelaar, “Automatic microcalcification and cluster detection for digital and digitised mammograms,” _Knowledge-Based Systems_ , vol. 28, pp. 68–75, 2012.
* [16] H. Strange, Z. Chen, E. R. Denton, and R. Zwiggelaar, “Modelling mammographic microcalcification clusters using persistent mereotopology,” _Pattern Recognition Letters_ , vol. 47, pp. 157–163, 2014.
* [17] A. Bria, N. Karssemeijer, and F. Tortorella, “Learning from unbalanced data: A cascade-based approach for detecting clustered microcalcifications,” _Medical Image Analysis_ , vol. 18, no. 2, pp. 241–252, 2014.
* [18] Z. Chen, H. Strange, A. Oliver, E. R. Denton, C. Boggis, and R. Zwiggelaar, “Topological Modeling and Classification of Mammographic Microcalcification Clusters,” _IEEE Transactions on Biomedical Engineering_ , vol. 62, no. 4, pp. 1203–1214, 2015.
* [19] N. Alam, E. R. Denton, and R. Zwiggelaar, “Classification of microcalcification clusters in digital mammograms using a stack generalization based classifier,” _Journal of Imaging_ , vol. 5, no. 9, 2019\.
* [20] H. Cai, Q. Huang, W. Rong, Y. Song, J. Li, J. Wang, J. Chen, and L. Li, “Breast Microcalcification Diagnosis Using Deep Convolutional Neural Network from Digital Mammograms,” _Computational and Mathematical Methods in Medicine_ , vol. 2019, pp. 1–10, 2019.
* [21] P. Kainz, M. Urschler, S. Schulter, P. Wohlhart, and V. Lepetit, “You should use regression to detect cells,” in _Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015_ , N. Navab, J. Hornegger, W. M. Wells, and A. F. Frangi, Eds. Cham: Springer International Publishing, 2015, pp. 276–283.
* [22] Y. Xie, F. Xing, X. Shi, X. Kong, H. Su, and L. Yang, “Efficient and robust cell detection: A structured regression approach,” _Medical Image Analysis_ , vol. 44, pp. 245 – 254, 2018.
* [23] M. I. Meyer, A. Galdran, A. M. Mendonça, and A. Campilho, “A pixel-wise distance regression approach for joint retinal optical disc and fovea detection,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 11071 LNCS, pp. 39–47, 2018.
* [24] K. M. H. van Wijnen, F. Dubost, P. Yilmaz, M. A. Ikram, W. J. Niessen, H. Adams, M. W. Vernooij, and M. de Bruijne, “Automated lesion detection by regressing intensity-based distance with a neural network,” in _Medical Image Computing and Computer Assisted Intervention – MICCAI 2019_ , D. Shen, T. Liu, T. M. Peters, L. H. Staib, C. Essert, S. Zhou, P.-T. Yap, and A. Khan, Eds. Cham: Springer International Publishing, 2019, pp. 234–242.
* [25] R. Strickland and Hee Il Hahn, “Wavelet transforms for detecting microcalcifications in mammograms,” _IEEE Transactions on Medical Imaging_ , vol. 15, no. 2, pp. 218–229, 2002.
* [26] E. Regentova, L. Zhang, J. Zheng, and G. Veni, “Microcalcification detection based on wavelet domain hidden Markov tree model: Study for inclusion to computer aided diagnostic prompting system,” _Medical Physics_ , vol. 34, no. 6, pp. 2206–2219, 2007.
* [27] D. Betal, N. Roberts, and G. H. Whitehouse, “Segmentation and numerical analysis of microcalcifications on mammograms using mathematical morphology,” _British Journal of Radiology_ , vol. 70, no. SEPT., pp. 903–917, sep 1997.
* [28] S. Halkiotis and J. Mantas, “Automatic detection of clustered microcalcifications in digital mammograms,” _Studies in Health Technology and Informatics_ , vol. 90, pp. 24–29, 2002.
* [29] S. Xu, H. Liu, and E. Song, “Marker-controlled watershed for lesion segmentation in mammograms,” _Journal of Digital Imaging_ , vol. 24, no. 5, pp. 754–763, 2011.
* [30] M. Ciecholewski, “Microcalcification Segmentation from Mammograms: A Morphological Approach,” _Journal of Digital Imaging_ , vol. 30, no. 2, pp. 172–184, apr 2017.
* [31] Heng-Da Cheng, Yui Man Lui, and R. Freimanis, “A novel approach to microcalcification detection using fuzzy logic technique,” _IEEE Transactions on Medical Imaging_ , vol. 17, no. 3, pp. 442–450, 2002.
* [32] I. C. Moreira, I. Amaral, I. Domingues, A. Cardoso, M. J. Cardoso, and J. S. Cardoso, “Inbreast: Toward a full-field digital mammographic database,” _Academic Radiology_ , vol. 19, no. 2, pp. 236 – 248, 2012.
* [33] T. Lindeberg, “Feature Detection with Automatic Scale Selection,” _International Journal of Computer Vision_ , vol. 30, no. 2, pp. 79–116, 1998\.
* [34] A. F. Frangi, W. J. Niessen, K. L. Vincken, and M. A. Viergever, “Multiscale vessel enhancement filtering,” in _Medical Image Computing and Computer-Assisted Intervention — MICCAI’98_ , W. M. Wells, A. Colchester, and S. Delp, Eds. Berlin, Heidelberg: Springer Berlin Heidelberg, 1998, pp. 130–137.
* [35] T. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie, “Feature pyramid networks for object detection,” in _2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , 2017, pp. 936–944.
* [36] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4, inception-ResNet and the impact of residual connections on learning,” _31st AAAI Conference on Artificial Intelligence, AAAI 2017_ , pp. 4278–4284, 2017.
* [37] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 9351, pp. 234–241, 2015.
* [38] P. Yakubovskiy, “Segmentation models pytorch,” https://github.com/qubvel/segmentation_models.pytorch (accessed 1 November 2020), 2020.
* [39] F. Milletari, N. Navab, and S. A. Ahmadi, “V-Net: Fully convolutional neural networks for volumetric medical image segmentation,” _Proceedings - 2016 4th International Conference on 3D Vision, 3DV 2016_ , pp. 565–571, 2016\.
* [40] M. Drozdzal, E. Vorontsov, G. Chartrand, S. Kadoury, and C. Pal, “The importance of skip connections in biomedical image segmentation,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 10008 LNCS, pp. 179–187, 2016.
* [41] D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,” _3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings_ , pp. 1–15, 2015.
* [42] K. He, X. Zhang, S. Ren, and J. Sun, “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification,” in _Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV)_ , ser. ICCV ’15. USA: IEEE Computer Society, 2015, p. 1026–1034.
* [43] M. Ankerst, M. M. Breunig, H.-p. Kriegel, and J. Sander, “OPTICS: Ordering Points To Identify the Clustering Structure,” _ACM SIGMOD Record_ , vol. 28, no. 2, pp. 49–60, 1999.
* [44] A. Rosenberg and J. Hirschberg, “V-measure: A conditional entropy-based external cluster evaluation measure,” in _Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL)_. Prague, Czech Republic: Association for Computational Linguistics, Jun. 2007, pp. 410–420.
|
11institutetext: Department of Physics, Chungbuk National University,
Cheongju 28644, Republic of Korea
11email<EMAIL_ADDRESS>22institutetext: Max Planck Institute
for Astronomy, Königstuhl 17, D-69117 Heidelberg, Germany 33institutetext:
Department of Astronomy, The Ohio State University, 140 W. 18th Ave.,
Columbus, OH 43210, USA 44institutetext: Astronomical Observatory, University
of Warsaw, Al. Ujazdowskie 4, 00-478 Warszawa, Poland 55institutetext:
Institute of Information and Mathematical Sciences, Massey University, Private
Bag 102-904, North Shore Mail Centre, Auckland, New Zealand 66institutetext:
Dipartimento di Fisica ‘E.R. Caianiello’, Universit a di Salerno, Via Giovanni
Paolo 132, Fisciano I-84084, Italy 77institutetext: Korea Astronomy and Space
Science Institute, Daejon 34055, Republic of Korea 88institutetext: University
of Canterbury, Department of Physics and Astronomy, Private Bag 4800,
Christchurch 8020, New Zealand 99institutetext: Department of Particle Physics
and Astrophysics, Weizmann Institute of Science, Rehovot 76100, Israel
1010institutetext: Center for Astrophysics $|$ Harvard & Smithsonian 60 Garden
St., Cambridge, MA 02138, USA 1111institutetext: Department of Astronomy and
Tsinghua Centre for Astrophysics, Tsinghua University, Beijing 100084, China
1212institutetext: School of Space Research, Kyung Hee University, Yongin,
Kyeonggi 17104, Republic of Korea 1313institutetext: Korea University of
Science and Technology, 217 Gajeong-ro, Yuseong-gu, Daejeon, 34113, Republic
of Korea 1414institutetext: Department of Physics, University of Warwick,
Gibbet Hill Road, Coventry, CV4 7AL, UK 1515institutetext: Institute for
Space-Earth Environmental Research, Nagoya University, Nagoya 464-8601, Japan
1616institutetext: Astrophysics Science Division, NASA/Goddard Space Flight
Center, Greenbelt, MD20771, USA 1717institutetext: Laboratory for Exoplanets
and Stellar Astrophysics, NASA / Goddard Space Flight Center, Greenbelt, MD
20771, USA 1818institutetext: Department of Astronomy, University of Maryland,
College Park, MD 20742, USA 1919institutetext: Department of Earth and
Planetary Science, Graduate School of Science, The University of Tokyo, 7-3-1
Hongo, Bunkyo-ku, Tokyo 2020institutetext: Instituto de Astrofísica de
Canarias, Vía Láctea s/n, E-38205 La Laguna, Tenerife, Spain
2121institutetext: Department of Earth and Space Science, Graduate School of
Science, Osaka University, 1-1 Machikaneyama, Toyonaka, Osaka 560-0043, Japan
2222institutetext: Department of Astronomy, Graduate School of Science, The
University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-0033, Japan
2323institutetext: Code 667, NASA Goddard Space Flight Center, Greenbelt, MD
20771, USA 2424institutetext: Zentrum für Astronomie der Universität
Heidelberg, Astronomisches Rechen-Institut, Mönchhofstr. 12-14, 69120
Heidelberg, Germany 2525institutetext: Department of Physics, University of
Auckland, Private Bag 92019, Auckland, New Zealand 2626institutetext:
Department of Physics, The Catholic University of America, Washington, DC
20064, USA 2727institutetext: Institute of Space and Astronautical Science,
Japan Aerospace Exploration Agency, 3-1-1 Yoshinodai, Chuo, Sagamihara,
Kanagawa, 252-5210, Japan 2828institutetext: University of Canterbury Mt. John
Observatory, P.O. Box 56, Lake Tekapo 8770, New Zealand
# Four sub-Jovian-mass planets detected by high-cadence microlensing surveys
Cheongho Han 0101 Doeon Kim 0101 Andrew Gould 02020303 Andrzej Udalski 0404
Ian A. Bond 0505 Valerio Bozza 0606 Youn Kil Jung
(Leading authors)
Michael D. Albrow 07070808 Sun-Ju Chung 0707 Kyu-Ha Hwang 0707 Yoon-Hyun Ryu
0707 In-Gu Shin 0707 Yossi Shvartzvald 0909 Jennifer C. Yee 1010 Weicheng Zang
1111 Sang-Mok Cha 07071212 Dong-Jin Kim 0707 Seung-Lee Kim 0707 Chung-Uk Lee
0707 Dong-Joo Lee 0707 Yongseok Lee 07071313 Byeong-Gon Park 07071313 Richard
W. Pogge
(The KMTNet Collaboration) 0303
Przemek Mróz 0404 Michał K. Szymański 0404 Jan Skowron 0404 Radosław Poleski
0404 Igor Soszyński 0404 Paweł Pietrukowicz 0404 Szymon Kozłowski 0404
Krzysztof Ulaczyk 1414 Krzysztof A. Rybicki 0404 Patryk Iwanek
(The OGLE Collaboration) 0404
Fumio Abe 1515 Richard K. Barry 1515 David P. Bennett 17171818 Aparna
Bhattacharya 17171818 Hirosane Fujii 1515 Akihiko Fukui 19192020 Yuki Hirao
161617172121 Yoshitaka Itow 1515 Rintaro Kirikawa 2121 Naoki Koshimoto 2121
Iona Kondo 2121 Yutaka Matsubara 1515 Sho Matsumoto 2121 Shota Miyazaki 2121
Yasushi Muraki 2121 Greg Olmschenk 2323 Arisa Okamura 2121 Clément Ranc 2424
Nicholas J. Rattenbury 2525 Yuki Satoh 2121 Stela Ishitani Silva 17172626
Takahiro Sumi 2121 Daisuke Suzuki 2727 Taiga Toda 2121 Paul J. Tristram 2828
Aikaterini Vandorou 2323 Hibiki Yama
(The MOA Collaboration)
2121
(Received ; accepted)
###### Abstract
Aims. With the aim of finding short-term planetary signals, we investigated
the data collected from the high-cadence microlensing surveys.
Methods. From this investigation, we found four planetary systems with low
planet-to-host mass ratios, including OGLE-2017-BLG-1691L, KMT-2021-BLG-0320L,
KMT-2021-BLG-1303L, and KMT-2021-BLG-1554L. Despite the short durations,
ranging from a few hours to a couple of days, the planetary signals were
clearly detected by the combined data of the lensing surveys. It is found that
three of the planetary systems have mass ratios of the order of $10^{-4}$ and
the other has a mass ratio slightly greater than $10^{-3}$.
Results. The estimated masses indicate that all discovered planets have sub-
Jovian masses. The planet masses of KMT-2021-BLG-0320Lb, KMT-2021-BLG-1303Lb,
and KMT-2021-BLG-1554Lb correspond to $\sim 0.10$, $\sim 0.38$, and $\sim
0.12$ times of the mass of the Jupiter, and the mass of OGLE-2017-BLG-1691Lb
corresponds to that of the Uranus. The estimated mass of the planet host
KMT-2021-BLG-1554L, $M_{\rm host}\sim 0.08\leavevmode\nobreak\ M_{\odot}$,
corresponds to the boundary between a star and a brown dwarf. Besides this
system, the host stars of the other planetary systems are low-mass stars with
masses in the range of $\sim[0.3$–$0.6]\leavevmode\nobreak\ M_{\odot}$. The
discoveries of the planets well demonstrate the capability of the current
high-cadence microlensing surveys in detecting low-mass planets.
###### Key Words.:
gravitational microlensing – planets and satellites: detection
## 1 Introduction
During the 2010s, microlensing entered an era of high-cadence surveys with the
instrumental upgrade of previously established experiments and the
participation of a new experiment. The new era started when the Microlensing
Observations in Astrophysics survey (MOA: Bond et al., 2001), which had
previously carried out a lensing survey using a 0.61 m telescope and a camera
with a 1.3 deg2 field of view (FOV) during the early phase, launched its
second phase experiment with the employment of a 1.8 m telescope equipped with
a wide-field camera yielding a 2.2 deg2 FOV. Since its first operation in
1992, the Optical Gravitational Lensing Experiment (OGLE) has been upgraded
multiple times, and it is in its fourth phase (OGLE-IV: Udalski et al., 2015)
with the employment of a 1.3 m telescope and a camera with a 1.4 deg2 FOV. The
Korea Microlensing Telescope Network (KMTNet: Kim et al., 2016), which was
launched in 2016, is being carried out with the use of three globally
distributed 1.6 m telescopes, each of which is mounted with a wide-field
detector providing a 4 deg2 FOV. With the use of wide-field cameras mounted on
multiple telescopes, the current lensing surveys can monitor lensing events
with a dramatically enhanced observational cadence.
Table 1: Coordinates, fields, alert dates, and baseline magnitudes
Event | (RA, decl)J2000 | $(l,b)$ | Field | Alert date | $I_{\rm base}$ (mag)
---|---|---|---|---|---
OGLE-2017-BLG-1691 | (17:34:22.52, -29:17:05.39) | $(-1^{\circ}\hskip-2.0pt.601,1^{\circ}\hskip-2.0pt.895)$ | OGLE (BLG654.31) | 2017 Sep 6 | 19.90
(KMT-2017-BLG-0752) | | | KMT (BLG14) | postseason |
KMT-2021-BLG-0320 | (17:57:33.15, -30:30:14.18) | $(2^{\circ}\hskip-2.0pt.095,-4^{\circ}\hskip-2.0pt.299)$ | KMT (BLG01, BLG41) | 2021 Apr 9 | 20.08
KMT-2021-BLG-1303 | (18:07:27.33, -29:16:53.69) | $(-0^{\circ}\hskip-2.0pt.027,-3^{\circ}\hskip-2.0pt.028)$ | KMT (BLG04) | 2021 Jun 14 | 19.67
(MOA-2021-BLG-182) | | | MOA (gb13) | 2021 Jun 17 |
KMT-2021-BLG-1554 | (17:51:12.82, -31:51:51.70) | $(-1^{\circ}\hskip-2.0pt.888,-2^{\circ}\hskip-2.0pt.543)$ | KMT (BLG01, BLG41) | 2021 Jul 1 | 22.39
111 ${\rm HJD}^{\prime}={\rm HJD}-2450000$.
The greatly increased observational cadence of the lensing surveys has brought
out important changes not only in the observational strategy but also in the
outcome of planetary microlensing searches. First, being able to resolve
short-lasting planetary signals from survey observations, the high-cadence
surveys can function without the survey+followup experiment mode, in which
low-cadence surveys mainly detect and alert lensing events and followup groups
densely observe the alerted events. Second, the number of microlensing planets
has substantially increased with the operation of the high-cadence surveys.
Among the total 135 microlensing planets with measured mass ratios222The
sample includes 119 planets listed in the NASA Exoplanet Archive
(https://exoplanetarchive.ipac.caltech.edu/index.html) plus 16 planets that
are not included in the archive: KMT-2020-BLG-0414Lb (Zang et al., 2021a),
OGLE-2019-BLG-1053Lb (Zang et al., 2021b), KMT-2019-BLG-0253Lb,
KMT-2019-BLG-0953Lb, OGLE-2018-BLG-0977Lb, OGLE-2018-BLG-0506Lb,
OGLE-2018-BLG-0516Lb, OGLE-2019-BLG-1492Lb (Hwang et al., 2022),
KMT-2017-BLG-2509Lb, OGLE-2017-BLG-1099Lb, OGLE-2019-BLG-0299Lb (Han et al.,
2021), OGLE-2016-BLG-1093Lb (Shin et al., 2022), KMT-2018-BLG-1988Lb (Han et
al., 2022b), KMT-2021-BLG-0912Lb (Han et al., 2022a), OGLE-2019-BLG-0468Lb,
and OGLE-2019-BLG-0468Lc (Han et al., 2022c)., 80 (59%) planets were found
with the use of the data from the KMTNet survey. This can be seen in Figure 1,
in which we present the histogram of microlensing planets as a function of the
planet-to-host mass ratio $q$. The histogram also shows that the high-cadence
surveys contribute to the detections of planets with very low mass ratios,
especially those with $q<10^{-4}$.
Figure 1: Histogram of published microlensing planets as a function of the
planet-to-host mass ratio. The histogram of planets detected with the data
from KMTNet survey is separately marked in brown. The solid dots with error
bars indicate the four planets reported in this work.
In this paper, we report four planetary systems with low planet-to-host mass
ratios detected from the microlensing surveys. Despite the short durations,
ranging from a few hours to a couple of days, the planetary signals were
clearly detected by the combined data of the lensing surveys. It is found that
all of the discovered planets have sub-Jovian masses lying in the range of
[0.05–0.38] $M_{\rm J}$, illustrating the importance of the high-cadence
surveys in detecting low-mass planets.
We present the analysis of the planetary lensing events according to the
following organization. In Sect. 2, we describe the observations of the
individual lensing events, instrument used for observations, and the process
of the data reduction. In Sect. 3, we explain the procedure of analysis that
is applied to each of the individual events. In the subsequent subsections, we
describe the detailed features of the planetary signals and present the
results of the analyses conducted for the individual lensing events. In Sect.
4, we specify the types of the source stars and estimate the angular Einstein
radii of the lens systems. In Sect. 5, we estimate the physical parameters of
the planetary systems by conducting Bayesian analyses. We summarize the
results of the analyses and conclude in Sect. 6.
## 2 Observations and data
The four lensing events for which we present analyses are (1)
OGLE-2017-BLG-1691 (KMT-2017-BLG-0752), (2) KMT-2021-BLG-0320, (3)
KMT-2021-BLG-1303 (MOA-2021-BLG-182), and (4) KMT-2021-BLG-1554. The source
stars of all events lie toward the Galactic bulge field. In Table 1, we list
the equatorial and galactic coordinates, observation fields of the surveys,
alert dates, and baseline magnitudes of the individual events. The baseline
magnitude $I_{\rm base}$ is approximately scaled to the OGLE-III photometry
system.
The events were detected from the combined observations of the lensing surveys
conducted by the KMTNet, OGLE, and MOA groups. The KMTNet survey utilizes
three identical 1.6 m telescopes that are distributed in three countries of
the Southern Hemisphere for the continuous coverage of lensing events. The
sites of the individual telescopes are the Siding Spring Observatory in
Australia (KMTA), the Cerro Tololo InterAmerican Observatory in Chile (KMTC),
and the South African Astronomical Observatory in South Africa (KMTS).
Observations by the OGLE survey were conducted using the 1.3 m telescope
located at Las Campanas Observatory in Chile. The MOA survey uses the 1.8 m
telescope at the Mt. John Observatory in New Zealand. Observations were done
mainly in the $I$ band for the KMTNet and OGLE surveys and in the customized
MOA-$R$ band for the MOA survey. A subset of images were acquired in the $V$
band for the purpose of estimating source colors. The detailed procedure of
the source color estimation will be discussed in Sect. 4.
Table 2: Error bar readjustment factors Event | Data set | $k$ | $\sigma_{\rm min}$ (mag)
---|---|---|---
OGLE-2017-BLG-1691 | OGLE | 1.010 | 0.010
| KMTA | 0.866 | 0.040
| KMTC | 1.170 | 0.010
| KMTS | 1.131 | 0.020
KMT-2021-BLG-0320 | KMTA (BLG01) | 1.127 | 0.040
| KMTA (BLG41) | 1.389 | 0.025
| KMTC (BLG01) | 1.091 | 0.020
| KMTC (BLG41) | 1.175 | 0.020
| KMTS (BLG01) | 1.356 | 0.010
| KMTS (BLG41) | 1.319 | 0.010
KMT-2021-BLG-1303 | KMTA | 1.304 | 0.020
| KMTC | 1.237 | 0.010
| KMTS | 1.364 | 0.010
| MOA | 1.034 | 0.020
KMT-2021-BLG-1554 | KMTA (BLG01) | 1.371 | 0.020
| KMTA (BLG41) | 1.446 | 0.020
| KMTC (BLG01) | 1.418 | 0.020
| KMTC (BLG41) | 1.367 | 0.020
| KMTS (BLG41) | 1.364 | 0.020
Reduction and photometry of the data were done using the pipelines of the
individual survey groups: Albrow et al. (2009) for the KMTNet survey, Woźniak
(2000) for the OGLE survey, and Bond et al. (2001) for the MOA survey. All
these pipelines apply the difference image analysis algorithm (Tomaney &
Crotts, 1996; Alard & Lupton, 1998) developed for the optimal photometry of
stars lying in very dense star fields. For a subset of the KMTC data set, we
carried out additional photometry utilizing the pyDIA code (Albrow, 2017) for
the specification of the source colors. Following the routine described in Yee
et al. (2012), we rescale the error bars of data by $\sigma=k(\sigma_{\rm
min}^{2}+\sigma_{0}^{2})^{1/2}$, where $\sigma_{0}$ represents the error
estimated from the photometry pipeline, $\sigma_{\rm min}$ is a factor used to
make the data consistent with the scatter of data, and the factor $k$ is used
to make $\chi^{2}$ per degree of freedom for each data set become unity. In
Table 2, we list the factors $k$ and $\sigma_{\rm min}$ of the individual data
sets.
## 3 Analyses
The light curves of the analyzed events share a common characteristic that
short-lived anomalies appear on the otherwise smooth and symmetric form of a
single-lens single-source (1L1S) event. Such anomalies in lensing light curves
can be produced by two channels, in which the first channel is a perturbation
induced by a low-mass companion, such as a planet, to the lens (Gould & Loeb,
1992), and the second channel is an anomaly caused by a faint companion to the
source (Gaudi, 1998). Hereafter we denote events with a binary lens and a
binary source as 2L1S and 1L2S events, respectively. We examine the origins of
the anomalies by modeling the light curve of the individual events under these
2L1S and 1L2S interpretations.
Lensing light curves are described by the combination of various lensing
parameters. For a 1L1S event, the light curve is described by three parameters
of $(t_{0},u_{0},t_{\rm E})$, which denote the time of the closest approach of
the source to the lens, the separation (scaled to the angular Einstein radius
$\theta_{\rm E}$) between the source and lens at that time, and the time scale
of the event, respectively. The event time scale is defined as the time
required for a source to transit $\theta_{\rm E}$.
In addition to these basic parameters, describing the light curve of a 2L1S
event requires one to include extra parameters of $(s,q,\alpha)$, which
represent the separation (scaled to $\theta_{\rm E}$) and mass ratio between
the binary lens components, and the angle between the binary-lens axis and the
direction of the source motion (source trajectory angle), respectively.
Because planet-induced anomalies are usually generated by the crossings over
or close approach of the source to the planet-induced caustics, an additional
parameter of the normalized source radius $\rho$, which is defined as the
ratio of the angular radius of the source $\theta_{*}$ to $\theta_{\rm E}$, is
needed to describe the deformation of the anomaly by finite-source effects
(Bennett & Rhie, 1996). In computing finite-source magnifications, we consider
the limb-darkening variation of the source.
One also needs extra parameters to describe the lensing light curve of a 1L2S
event. These parameters are $(t_{0,2},u_{0,2},q_{F})$, which represent the
closest approach time and separation between the lens and the second source,
and the flux ratio between the binary source stars, respectively. In the 1L2S
model, we designate the lensing parameters related to the primary source,
$S_{1}$, as $(t_{0,1},u_{0,1})$ to distinguish them from those related to the
source companion, $S_{2}$. In order to consider finite-source effects
occurring when the lens passes over either of the source stars, we add two
additional parameters of the normalized source radii $\rho_{1}$ and $\rho_{2}$
for the lens transits over $S_{1}$ and $S_{2}$, respectively (Dominik, 2019).
In modeling 1L1S and 1L2S light curves, for which the lensing magnification is
smooth with the variation of the lensing parameters, we search for the
solution of the lensing parameters via a downhill approach by minimizing
$\chi^{2}$ using the Markow Chain Monte Carlo (MCMC) algorithm. In modeling
2L1S light curves, for which the lensing magnification variation is
discontinuous due to the formation of caustics and there may exist multiple
local solutions caused by various types of degeneracy, we model the light
curves in two steps. In the first step, we conduct grid searches for the
binary parameters $s$ and $q$, construct a $\chi^{2}$ map on the $s$–$q$
parameter plane, and then identify local solutions appearing on the $\chi^{2}$
map. In the second step, we polish the individual local solutions using the
downhill approach. Below we present the details of the modeling conducted for
the individual lensing events.
Figure 2: Lensing light curve of OGLE-2017-BLG-1691. The bottom panel shows
the whole view, and the top panel displays the enlargement of the region
around the anomaly (the region enclosed by a box in the bottom panel). In the
top panel, curves of the three tested models, 2L1S (outer), 1L2S, and 1L1S
models, are drawn over the data points, and the residuals from the individual
models are presented in the three middle panels. The curve drawn in the bottom
panel is the best-fit model (outer 2L1S model). Colors of data points are set
to match those of the telescopes used for observations marked in the legend.
The curves drawn in the 1L2S and 1L1S residual panels represent the
differences from the 2L1S model.
### 3.1 OGLE-2017-BLG-1691 (KMT-2017-BLG-0752)
The lensing event OGLE-2017-BLG-1691 occurred during the 2017 bulge season. It
was first found by the OGLE survey, and later confirmed by the KMTNet survey
from the post-season analyses of the data collected during the season. The
KMTNet group designated the event as KMT-2017-BLG-0752. Hereafter, we use the
nomenclatures of events by the ID references of the surveys who first found
the events in accordance with the convention of the microlensing community.
The baseline magnitude of the event was $I_{\rm base}=19.900\pm 0.004$.
The lensing light curve of OGLE-2017-BLG-1691 is shown in Figure 2, in which
the bottom panel shows the whole view and the top panel displays the
enlargement of the region around the anomaly. The event was alerted by the
OGLE survey on 2017 September 6, ${\rm HJD}^{\prime}\equiv{\rm
HJD}-2450000\sim 8002.5$, which approximately corresponds to the time near the
peak. An anomaly occurred about one day after the peak, but it was not noticed
during the progress of the event because it was covered by just a single OGLE
data point, and the KMTNet data were not released during the lensing
magnification. The existence of the anomaly was identified five years after
the event, shortly after the re-reduction of all 2017 KMT light curves, from a
project of reinvestigating the previous KMTNet data conducted to find
unnoticed planetary signals (Han et al., 2022b). The top panel of Figure 2
shows that the anomaly, which lasted for $\sim 5$ hours centered at ${\rm
HJD}^{\prime}\sim 8003.6$, was additionally covered by three KMTC data points,
and this confirmed that the single anomalous OGLE point was real.
Figure 3: Lens system configurations of OGLE-2017-BLG-1691: upper panel
(inner solution) and lower panel (outer solution). The inset in each panel
shows the whole view of the lens system, and the main panel presents the zoom-
in view of the caustics. The red cuspy figure is the caustic and the line with
an arrow is the source trajectory. The orange circle on the trajectory is
drawn to represent the source scale with respect to the caustic. The blue
solid dots marked by $M_{1}$ and $M_{2}$ indicate the positions of the host
and planet, respectively. The dashed circle with unit radius in the inset
represents the Einstein ring. Table 3: Lensing parameters of
OGLE-2017-BLG-1691
Parameter | 2L1S (Inner) | 2L1S (Outer) | 1L2S
---|---|---|---
$\chi^{2}$ | 866.9 | 866.5 | 880.4
$t_{0}$ (HJD′) | $8002.540\pm 0.010$ | $8002.531\pm 0.010$ | $8002.499\pm 0.012$
$u_{0}$ ($10^{-2}$) | $4.83\pm 0.27$ | $4.95\pm 0.26$ | $6.55\pm 0.49$
$t_{\rm E}$ (days) | $19.42\pm 0.83$ | $18.94\pm 0.82$ | $18.77\pm 0.83$
$s$ | $1.058\pm 0.011$ | $1.003\pm 0.014$ | –
$q$ ($10^{-4}$) | $0.708\pm 0.23$ | $0.97\pm 0.34$ | –
$\alpha$ (rad) | $3.873\pm 0.012$ | $3.865\pm 0.010$ | –
$\rho$ ($10^{-3}$) | $3.40\pm 0.47$ | $3.54\pm 0.52$ | –
$t_{0,2}$ (HJD′) | – | – | $8003.587\pm 0.015$
$u_{0,2}$ ($10^{-2}$) | – | – | $0.029\pm 0.124$
$\rho_{2}$ ($10^{-3}$) | – | – | $3.39\pm 0.93$
$q_{F}$ ($10^{-3}$) | – | – | $6.11\pm 1.85$
333 ${\rm HJD}^{\prime}={\rm HJD}-2450000$.
We model the light curve of the event under the 1L1S, 2L1S and 1L2S
interpretations. The residuals of the individual tested models are compared in
Figure 2. It is found that the 2L1S model best describes the anomaly, being
favored over the 1L1S and 1L2S models by $\Delta\chi^{2}=456.1$ and 13.9,
respectively. From the comparison of the residuals, it is found that the 2L1S
model is confirmed not only by the 4 data points (3 KMTC plus 1 OGLE points)
with positive deviations near the peak of the bump but also by the 3
additional points (2 KMTS and 1 KMTC points) before the major bump with
slightly negative deviations.
We find two degenerate 2L1S solutions with binary parameters of
$(s,q)\sim(1.058,0.71\times 10^{-4})$ and $\sim(1.003,0.97\times 10^{-4})$,
which we designate as “inner” and “outer” solutions, respectively, for the
reason to be mentioned below. We list the full lensing parameters of the two
2L1S solutions in Table 3, together with the parameters of the 1L2S model. The
degeneracy is very severe and the outer solution is favored only by
$\Delta\chi^{2}=0.4$. The lensing parameters of the solutions indicate that
the anomaly was produced by a very small mass-ratio planetary companion to the
lens lying close to the Einstein ring of the primary regardless of the
solutions. From the fact that both solutions have separations greater than
unity and do not follow the relation of $s_{\rm inner}\times s_{\rm
outer}\simeq 1$, the degeneracy is different from the “close-wide” degeneracy
(Griest & Safizadeh, 1998; Dominik, 1999; An, 2005), which arises due to the
similarity between the central caustics of planetary lens systems with planet
separations $s$ and $1/s$. Instead, the planet separations of the two
degenerate solutions follow the relation
$\sqrt{s_{\rm in}\times s_{\rm out}}=s^{\dagger};\qquad
s^{\dagger}={\sqrt{u_{\rm anom}^{2}+4}+u_{\rm anom}\over 2},$ (1)
where $u_{\rm anom}^{2}=\tau_{\rm anom}^{2}+u_{0}^{2}$, $\tau_{\rm
anom}=(t_{\rm anom}-t_{0})/t_{\rm E}$, and $t_{\rm anom}$ indicates the time
of the planetary anomaly (Hwang et al., 2022; Zhang et al., 2022; Ryu et al.,
2022). With the values $t_{0}\simeq 8002.5$, $u_{0}\simeq 4.85\times 10^{-2}$,
$t_{\rm E}\simeq 19$ day and $t_{\rm anom}\simeq 8003.6$, one finds that
$s^{\dagger}\simeq 1.04$, which matches well $\sqrt{s_{\rm in}\times s_{\rm
out}}\simeq 1.03$. This indicates that the similarity between the light curves
of the two solutions is caused by the degeneracy identified by Yee et al.
(2021), who first mentioned the continuous transition between the “close-wide”
and “inner-outer” (Gaudi & Gould, 1997) degeneracies. Hereafter, we refer to
this degeneracy as “offset degeneracy” following Zhang et al. (2022).
Figure 3 shows the lens systems configurations, in which the source trajectory
(line with an arrow) with respect to the caustic and positions of the lens
components (blue solid dots marked by $M_{1}$ and $M_{2}$) is presented: inner
solution in the upper panel and outer solution in the lower panel. The source
passed the inner side (with respect to $M_{1}$) of the caustic according to
the inner solution, while the source passes the outer side according to the
outer solution. For both solutions, the source crossed the cusp of the
caustic, and thus the anomaly is affected by finite-source effects during the
caustic crossing, allowing us to precisely measure the normalized source
radius. For each solution, the source size relative to the caustic is
represented by an orange circle marked on the source trajectory. As will be
discussed in Sect. 4, the measurement of $\rho$ is important to estimate the
lensing observable of the angular Einstein radius $\theta_{\rm E}$, which can
be used to constrain the physical lens parameters. However, the microlens
parallax vector, $\pi_{\rm E}=(\pi_{\rm rel}/\theta_{\rm
E})/(\mbox{\boldmath$\mu$}/\mu)$, which is another observable constraining the
physical lens parameters, cannot be securely measured because the event time
scale, $t_{\rm E}\sim 19$ days, is not long enough to produce detectable
deviations induced by the orbital motion of Earth around the Sun (Gould,
1992). Here $\mu$ represents the vector of the relative lens-source proper
motion.
### 3.2 KMT-2021-BLG-0320
The lensing event KMT-2021-BLG-0320 was found on 2021 April 9 (${\rm
HJD}^{\prime}\sim 9313$), when the event had not yet reached its peak, with
the employment of the KMTNet AlertFinder system (Kim et al., 2018), which
began full operation since the 2019 season. Two days after the detection, the
event reached its peak with a magnification of $A_{\rm max}\sim 170$, and then
gradually declined until it reached its baseline of $I_{\rm base}=20.08$. The
source of the event lies in the two KMTNet prime fields of BLG01 and BLG41,
toward which observations were conducted with a 30 min cadence for each field,
and thus with a 15 min combined cadence. The areas covered by the two fields
overlap except for about 15% of the area of each field filling the gaps
between the chips of the camera. Because the data were taken from two fields
of three telescopes, there are 6 data sets: KMTA (BLG01), KMTA (BLG41), KMTC
(BLG01), KMT (BLG41), KMTS (BLG01), and KMTS (BLG41).
Figure 4: Lensing light curve of KMT-2021-BLG-0320. Notations are same as those in Fig. 2. Table 4: Lensing parameters of KMT-2021-BLG-0320 Parameter | Close | Wide
---|---|---
$\chi^{2}$ | 4075.5 | 4075.5
$t_{0}$ (HJD′) | $9315.615\pm 0.001$ | $9315.615\pm 0.001$
$u_{0}$ ($10^{-3}$) | $5.86\pm 0.21$ | $5.93\pm 0.21$
$t_{\rm E}$ (days) | $13.12\pm 0.39$ | $12.97\pm 0.40$
$s$ | $0.771\pm 0.014$ | $1.274\pm 0.024$
$q$ ($10^{-4}$) | $3.02\pm 0.42$ | $2.89\pm 0.44$
$\alpha$ (rad) | $0.604\pm 0.014$ | $0.602\pm 0.014$
$\rho$ ($10^{-3}$) | – | –
In Figure 4, we present the light curve constructed from the combination of
the 6 KMTNet data sets. Although it would be difficult to notice the anomaly
from a glimpse, we inspected the light curve because the event reached a very
high magnification at the peak, near which the light curve is susceptible to
perturbations induced by a planet (Griest & Safizadeh, 1998). From this
inspection, it was found that the light curve exhibited an anomaly that lasted
for about 4 hours with a negative deviation with respect to a 1L1S model. See
the enlarged view around the peak region of the light curve presented in the
top panel of Figure 4. It is known that a planetary companion to a lens can
induce anomalies with both positive and negative deviations, while a faint
companion to a source can induce anomalies with only positive deviations
(Gaudi, 1998). We, therefore, modeled the light curve under the 2L1S
interpretation.
It is found that the anomaly is well described by a 2L1S model, in which the
mass ratio between the binary lens components is very low. We found two sets
of solutions with $(s,q)_{\rm close}\sim(0.77,3.0\times 10^{-4})$ and
$(s,q)_{\rm wide}\sim(1.27,2.9\times 10^{-4})$. We designate the individual
solutions as “close” and “wide” solutions, because the former solution has
$s<1.0$ and the latter has $s>1.0$. In Table 4, we list the full lensing
parameters for the two sets of solutions. The fits of the two solutions are
nearly identical with $\chi^{2}$ values that are equal to the first digit
after the decimal point, indicating that the degeneracy between the two
solutions is very severe. The fact that the binary separations of the two
solutions follow the relation of $s_{\rm close}\times s_{\rm wide}\simeq 1.0$
indicates that the similarity between the solution stems from the close-wide
degeneracy. It is known that the relation between the planet separations of
the solutions under the offset degeneracy in Equation (1) applies to more
general cases, including resonant case, than the $s_{\rm close}\times s_{\rm
wide}\simeq 1.0$ relation of the close-wide degeneracy, and we confirm this.
In addition to the relation between $s_{\rm in}$ and $s_{\rm out}$, Hwang et
al. (2022) provided analytic formulas for the heuristic estimation of the
source trajectory angle and the mass ratio;
$\alpha=\tan^{-1}\left({u_{0}\over\tau_{\rm anom}}\right);\qquad
q=\left({\Delta t_{\rm anom}\over 4t_{\rm
E}}\right)\left({s\over|u_{0}|}\right)\left|\sin^{3}\alpha\right|,$ (2)
where $\tau_{\rm anom}$ denotes the duration of the planet-induced anomaly. We
confirm that the heuristically estimated values of $\alpha$ and $q$ match well
those estimated from the modeling. It was found that the normalized source
radius could not be constrained, not even the upper limit, and thus the line
for $\rho$ in Table 4 is left blank.
Figure 5: Lens system configurations of KMT-2021-BLG-0320. The upper and
lower panels show the configurations of the close and wide solutions,
respectively.
Figure 5 shows the lensing configuration of KMT-2021-BLG-0320 for the close
(upper panel) and wide (lower panel) solutions. It shows that the source
passed the back-end region of the tiny central caustic induced by a planet,
and this generated the negative deviation of the observed anomaly. The central
caustics of the close and wide solutions are very similar, resulting in nearly
identical deviations. Because the source passed well outside the caustic,
finite-source effects could not be detected, and thus we do not mark an orange
circle representing the source size. Microlens-parallax effects could not be
securely detected because the event time scale, $t_{\rm E}\sim 13$ days, is
much shorter than the orbital period of Earth, that is, 1 year.
Figure 6: Lensing light curve of KMT-2021-BLG-1303. Notations are same as
those in Fig. 2.
### 3.3 KMT-2021-BLG-1303 (MOA-2021-BLG-182)
The event KMT-2021-BLG-1303 was first found on 2021 June 14 (${\rm
HJD}^{\prime}\sim 9379$) by the KMTNet survey, and later identified by the MOA
survey on 2021 June 17 (${\rm HJD}^{\prime}\sim 9382$). The MOA survey
designated the event as MOA-2021-BLG-182. A day after the first discovery, the
event displayed an anomaly that lasted for about 2 days. After the anomaly,
the event reached peak at ${\rm HJD}^{\prime}\sim 9385.1$ with a magnification
of $A_{\rm max}\sim 46$, and then gradually declined to the baseline of
$I_{\rm base}=19.67$.
The light curve of KMT-2021-BLG-1303 constructed with the combined KMTNet and
MOA data is shown in Figure 6. Compared to the previous two events, the light
curve of this event displays a very obvious anomaly with a maximum deviation
of $\Delta I\sim 1.4$ mag from a 1L1S model. The rapid brightening at ${\rm
HJD}^{\prime}\sim 9381.0$ indicates that the event experienced a caustic
crossing. Because caustics form due to the multiplicity of lens components, we
rule out the 1L2S origin of the anomaly and test only the 2L1S interpretation.
Table 5: Lensing parameters of KMT-2021-BLG-1303 Parameter | Value
---|---
$\chi^{2}$ | 1654.9
$t_{0}$ (HJD′) | $9385.091\pm 0.003$
$u_{0}$ ($10^{-2}$) | $2.17\pm 0.06$
$t_{\rm E}$ (days) | $25.21\pm 0.68$
$s$ | $1.029\pm 0.001$
$q$ ($10^{-4}$) | $6.42\pm 0.45$
$\alpha$ (rad) | $6.138\pm 0.001$
$\rho$ ($10^{-3}$) | $0.95\pm 0.08$
We find that the anomaly is well explained by a 2L1S model with a planetary
companion lying close to the Einstein ring of the host. We find a unique
solution without any degeneracy, and the binary parameters of the solution are
$(s,q)\sim(1.029,6.4\times 10^{-4})$. Here again the planet-to-host mass ratio
is of the order of $10^{-4}$ like the two previous events. We double checked
the uniqueness of the solution by thoroughly inspecting the region around the
$s$ and $q$ parameters predicted by the relations in Equations (1) and (2),
and confirmed that there is only a single solution. The full lensing
parameters of the event are listed in Table 5.
Figure 7: Lens system configuration of KMT-2021-BLG-1303. Notations are same as those in Fig. 3. Table 6: Lensing parameters of KMT-2021-BLG-1554 Parameter | 2L1S (Close) | 2L1S (Wide) | 1L2S
---|---|---|---
$\chi^{2}$ | 3486.0 | 3485.7 | 3496.1
$t_{0}$ (HJD′) | $9394.737\pm 0.010$ | $9394.742\pm 0.010$ | $9394.719\pm 0.019$
$u_{0}$ | $0.051\pm 0.009$ | $0.051\pm 0.011$ | $0.003\pm 0.054$
$t_{\rm E}$ (days) | $5.04\pm 0.78$ | $5.15\pm 0.72$ | $6.55\pm 1.11$
$s$ | $0.888\pm 0.039$ | $1.189\pm 0.051$ | –
$q$ ($10^{-4}$) | $14.14\pm 4.48$ | $14.94\pm 5.41$ | –
$\alpha$ (rad) | $4.705\pm 0.038$ | $4.727\pm 0.037$ | –
$\rho$ ($10^{-3}$) | $6.98\pm 1.18$ | $7.00\pm 1.28$ | $98.8\pm 24.2$
$t_{0,2}$ (HJD′) | – | – | $9394.740\pm 0.002$
$u_{0,2}$ | – | – | $0.002\pm 0.002$
$\rho_{2}$ ($10^{-3}$) | – | – | –
$q_{F}$ | – | – | $0.048\pm 0.008$
The lensing configuration of the event is presented in Figure 7, which shows
that the planet lies very close to the Einstein ring and induces a single six-
sided resonant caustic near the primary of the lens. The source entered the
caustic at ${\rm HJD}^{\prime}\sim 9381.0$, and this produced the sharp rise
of the light curve at the corresponding time. According to the model, the
source exited the caustic at ${\rm HJD}^{\prime}\sim 9384.2$, but the caustic-
crossing feature at this epoch is not obvious in the lensing light curve due
to the combination of the weak caustic and the poor coverage of the region.
The pattern of the anomaly between the caustic entrance and exit deviates from
a typical “U”-shape pattern because the source passed along the fold of the
caustic. The light curve during the caustic entrance of the source was well
resolved by the KMTA data, and thus the normalized source radius,
$\rho=(0.95\pm 0.08)\times 10^{-3}$, is tightly constrained. However, it was
difficult to constrain the microlens parallax because the time scale, $t_{\rm
E}\sim 25$ days, is not long enough and the photometric precision of the faint
source with $I\sim 21$ is not high enough to detect subtle deviations induced
by microlens-parallax effects.
Figure 8: Lensing light curve of KMT-2021-BLG-1554. Notations are same as
those in Fig. 2.
### 3.4 KMT-2021-BLG-1554
This short event with a time scale of $t_{\rm E}\sim 5$ days and a faint
baseline magnitude of $I_{\rm base}=22.39$ was detected on 2021 Jul 1 (${\rm
HJD}^{\prime}\sim 9396$), about two days after the event peaked at $t_{0}\sim
9394.7$, by the KMTNet survey. The source lies in the prime fields of BLG01
and BLG41, and thus there are 6 data sets from the three KMTNet telescopes.
Among these data sets, the data set of the BLG41 field acquired by the KMTS
telescope was not used in the analysis due to its poor photometric quality
caused by bad seeing during the lensing magnification. Although the other data
set from KMTS and those from KMTA were included in the analysis, the results
are mainly derived from the KMTC data sets, which cover the peak region of the
light curve.
The light curve constructed with the available KMTNet data sets is shown in
Figure 8. In the peak region, it exhibits an anomaly that lasted for about 3
hours from a 1L1S model with a positive deviation of $\Delta I\sim 0.5$ mag.
The anomaly appears both in the BLG01 (two data points) and BLG41 (3 points)
data sets obtained from KMTC observations, confirming that the signal is real.
Because a positive anomaly can be produced by a binary companion to either a
lens or a source, we test both 2L1S and 1L2S models.
Table 7: De-reddened colors, magnitudes, and spectral types of source stars Event | $(V-I)_{0}$ | $I_{0}$ | Spectral type
---|---|---|---
OGLE-2017-BLG-1691 | $0.714\pm 0.065$ | $17.397\pm 0.039$ | G1 turnoff or subgiant
KMT-2021-BLG-0320 | $0.700\pm 0.040$ | $18.843\pm 0.002$ | G1V
KMT-2021-BLG-1303 | $0.849\pm 0.015$ | $19.809\pm 0.003$ | G9V
KMT-2021-BLG-1554 | $0.873\pm 0.061$ | $18.649\pm 0.008$ | K1V
Figure 9: Lens system configuration of KMT-2021-BLG-1554. Notations are same as those in Fig. 5. Table 8: Angular source radii, Einstein radii, and relative lens-source proper motions Event | $\theta_{*}$ ($\mu$as) | $\theta_{\rm E}$ (mas) | $\mu$ (mas yr-1)
---|---|---|---
OGLE-2017-BLG-1691 | $1.043\pm 0.100$ | $0.30\pm 0.05$ | $5.68\pm 1.01$
KMT-2021-BLG-0320 | $0.528\pm 0.042$ | – | –
KMT-2021-BLG-1303 | $0.401\pm 0.029$ | $0.42\pm 0.05$ | $6.13\pm 0.70$
KMT-2021-BLG-1554 | $0.703\pm 0.065$ | $0.10\pm 0.02$ | $7.30\pm 1.50$
Detailed analysis of the light curve indicates that the anomaly was produced
by a planetary companion to the lens. We find two sets of solutions with
planet parameters of $(s,q)_{\rm close}\sim(0.89,1.4\times 10^{-3})$ and
$(s,q)_{\rm wide}\sim(1.12,1.5\times 10^{-3})$ resulting from the close-wide
degeneracy. The full lensing parameters of both solutions are provided in
Table 6. It is found that the degeneracy between the two 2L1S solution is
severe with $\Delta\chi^{2}=0.3$, and the 2L1S solutions are favored over the
model under the 1L2S interpretation by $\Delta\chi^{2}\sim 11$. The lensing
parameters of the 1L2S model are listed in Table 6. We compare the residuals
of the 2L1S and 1L2S models in the two middle panels of Figure 8. It turns out
that the source crossed a caustic, and thus we are able to measure the
normalized source radius, $\rho=(6.98\pm 1.18)\times 10^{-3}$.
The lensing configurations of the event according to the close and wide
solutions are shown in the upper and lower panels of Figure 9, respectively.
The figure shows that the anomaly was produced by the crossing of the source
over the sharp tip of the central caustic induced by a planet lying near the
Einstein ring of the planet host. The gap between the caustic entrance and
exit is much smaller than the source size, and thus the individual caustic-
crossing features do not show up in the anomaly and instead the anomaly
appears as a single bump. The incidence angle of the source on the binary axis
is nearly $90^{\circ}$, and thus the features on the rising and falling sides
of the anomaly are symmetric.
## 4 Source stars and angular Einstein radii
Among the four analyzed planetary events, the anomalies of the three events
(KMT-2017-BLG-0752, KMT-2021-BLG-1303, and KMT-2021-BLG-1554) were affected by
finite-source effects, and thus the normalized source radii can be measured.
In this section, we estimate the angular Einstein radii for these events.
Estimating $\theta_{\rm E}$ from a measured $\rho$ value requires one to
specify the source type, from which the angular source radius $\theta_{*}$ is
deduced and the Einstein radius is determined by
$\theta_{\rm E}={\theta_{*}\over\rho}.$ (3)
Although the normalized source radius and thus $\theta_{\rm E}$ cannot be
measured for the event KMT-2021-BLG-0320, we specify the source type for the
sake of completeness.
The specification of the source type is done by estimating the color and
brightness of the source. We measure the $I$ and $V$-band magnitudes of the
source from the regression of the data processed using the pyDIA photometry
code (Albrow, 2017) with the variation of the lensing magnification. We then
place the source on the instrumental color-magnitude diagram (CMD) of stars
lying around the source constructed using the same photometry code. Following
the method of Yoo et al. (2004), the instrumental source color and magnitude,
$(V-I,I)$, are then calibrated using the centroid of red giant clump (RGC),
for which its reddening and extinction-corrected (de-reddened) color and
magnitude, $(V-I,I)_{\rm RGC,0}$, are known, on the CMD as a reference.
Figure 10 shows the positions of the source stars (marked by blue dots) with
respect to those of the RGC centroids (red dots) on the CMDs of the individual
events. For OGLE-2017-BLG-1691, the source color could not be securely
measured not only because the $V$-band data sparsely covered the light curve
during the lensing magnification but also because the photometry quality of
these data is low, although the $I$-band magnitude was relatively well
measured. In this case, we estimated the source color as the median color of
stars with the same $I$-band magnitude offset from the RGC centroid in the CMD
constructed from the images of Baade’s window taken from the observations
using the Hubble Space Telescope (Holtzman et al., 1998). From the offsets in
color and magnitudes between the source and RGC centroid, $\Delta(V-I,I)$,
together with the known de-reddened color and magnitude of the RGC centroid,
$(V-I,I)_{\rm RGC,0}$ (Bensby et al., 2013; Nataf et al., 2013), we estimate
the calibrated source color and magnitude as $(V-I,I)_{0}=(V-I,I)_{\rm
RGC,0}+\Delta(V-I,I)$. In Table 7, we list the source colors and magnitudes of
the individual events estimated from this procedure along with the spectral
types of the source stars. It is found that the source of OGLE-2017-BLG-1691
is a turnoff star or a subgiant of a G spectral type, while the source stars
of the other events are main-sequence stars with spectral types ranging from G
to K.
Figure 10: Source locations with respect to the centroids of the red giant
clump (RGC) on the instrumental color-magnitude diagrams of stars lying around
the source stars of the events OGLE-2017-BLG-1691, KMT-2021-BLG-0320,
KMT-2021-BLG-1303, and KMT-2021-BLG-1554. For OGLE-2017-BLG-1691, the CMD is
constructed by combining the two CMDs from KMTC (grey dots) and HST (brown
dots) observations. For each event, the locations of the source and RGC
centroid are marked by blue and red solid dots, respectively.
For each event, the angular source radius and Einstein radius were estimated
from the measured color and magnitude of the source. For this, the measured
$V-I$ color was converted into $V-K$ color using the color-color relation of
Bessell & Brett (1988), $\theta_{*}$ value was interpolated from the
($V-K$)–$\theta_{*}$ relation of Kervella et al. (2004), and then the angular
Einstein radius was estimated using the relation in Equation (3). With the
measured Einstein radius together with the event time scale, the value of the
relative lens-source proper motion was assessed as
$\mu={\theta_{\rm E}\over t_{\rm E}}.$ (4)
In Table 8, we list the estimated values of $\theta_{*}$, $\theta_{\rm E}$,
and $\mu$ for the individual events. The $\theta_{\rm E}$ and $\mu$ values for
the event KMT-2021-BLG-0320 are left blank because finite-source effects in
the lensing light curve were not detected and subsequently the values of
$\rho$, $\theta_{\rm E}$, and $\mu$ could not be measured. We note that the
angular Einstein radius of KMT-2021-BLG-1554, $\theta_{\rm E}\sim 0.10$ mas,
is substantially smaller than those of the other events, and this together
with its short time scale, $t_{\rm E}\sim 5$ days, suggests that the mass of
the lens would be very low. KMT-2021-BLG-1554 is the seventh shortest
microlensing event with a bound planet. See Table 3 of Ryu et al. (2021).
Table 9: Physical lens parameters Event | $M_{\rm host}$ ($M_{\odot}$) | $M_{\rm planet}$ ($M_{\rm J}$) | $D_{\rm L}$ (kpc) | $a_{\perp}$ (AU)
---|---|---|---|---
OGLE-2017-BLG-1691 | $0.45^{+0.36}_{-0.25}$ | $0.046^{+0.037}_{-0.025}$ | $7.29^{+1.01}_{-1.33}$ | $2.41^{+0.34}_{-0.44}$ (inner)
| – | – | – | $2.54^{+0.36}_{-0.46}$ (outer)
KMT-2021-BLG-0320 | $0.32^{+0.39}_{-0.21}$ | $0.10^{+0.13}_{-0.07}$ | $6.95^{+1.07}_{-1.33}$ | $1.54^{+0.24}_{-0.30}$ (close)
| – | – | – | $2.55^{+0.39}_{-0.49}$ (wide)
KMT-2021-BLG-1303 | $0.57^{+0.32}_{-0.29}$ | $0.38^{+0.22}_{-0.20}$ | $6.28^{+0.95}_{-1.43}$ | $2.89^{+0.44}_{-0.66}$
KMT-2021-BLG-1554 | $0.08^{+0.13}_{-0.04}$ | $0.12^{+0.20}_{-0.07}$ | $7.68^{+1.04}_{-1.10}$ | $0.72^{+0.13}_{-0.14}$ (close)
| – | – | – | $0.96^{+0.15}_{-0.16}$ (wide)
## 5 Physical parameters
In this section, we estimate the physical parameters of the planetary systems.
For the unique constraint of the physical lens parameters, one must
simultaneously measure the extra observables of $\theta_{\rm E}$ and $\pi_{\rm
E}$, from which the mass $M$ and distance to the lens, $D_{\rm L}$, are
determined as
$M={\theta_{\rm E}\over\kappa\pi_{\rm E}};\qquad D_{\rm L}={{\rm
AU}\over\pi_{\rm E}\theta_{\rm E}+\pi_{\rm S}}.$ (5)
Here $\kappa=4G/(c^{2}{\rm AU})$, $\pi_{\rm S}={\rm AU}/D_{\rm S}$, and
$D_{\rm S}$ denotes the distance to the source (Gould, 1992, 2000). The values
of the microlens parallax could not be measured for any of the events, and
thus the physical parameters cannot be unambiguously determined from the
relation in Equation (5). However, one can still constrain $M$ and $D_{\rm L}$
using the other measured observables of $t_{\rm E}$ and $\theta_{\rm E}$,
which are related to the physical lens parameters as
$t_{\rm E}={\theta_{\rm E}\over\mu};\qquad\theta_{\rm E}=(\kappa M\pi_{\rm
rel})^{1/2},$ (6)
where $\pi_{\rm rel}={\rm AU}(D_{\rm L}^{-1}-D_{\rm S}^{-1})$ denotes the
relative parallax between the lens and source. The event time scales were well
measured for all events, and the Einstein radii were assessed for three of the
four events. In order to estimate the physical lens parameters with the
constraints provided by the measured observables of the individual events, we
conduct Bayesian analyses using a Galactic model.
Figure 11: Bayesian posteriors of the host masses of the planetary systems.
For each distribution, the solid vertical line indicates the median value, and
the dotted lines represent 1$\sigma$ range of the distribution. The blue and
red curves represent the distributions contributed by disk and bulge
populations of lenses, respectively, and the black curve is the sum from the
two lens populations.
In the Bayesian analysis, we started by producing many lensing events from a
Monte Carlo simulation conducted with the use of a Galactic model, which
defines the matter-density and kinetic distributions, and mass function of
Galactic objects. We adopted the Galactic model constructed by Jung et al.
(2021). In this model, the density distributions of disk and bulge objects are
described by the Robin et al. (2003) and Han & Gould (2003) models,
respectively. The kinematic distribution of disk objects is based on the
modified version of the Han & Gould (1995) model, in which the original
version is modified to reconcile the changed density model, that is, the Robin
et al. (2003) model. The bulge kinematic distribution is modeled based on
proper motions of stars in the Gaia catalog (Gaia Collaboration, 2016, 2018).
For the details of the density and kinematic distributions, see Jung et al.
(2021). In the Galactic model, the mass function is constructed with the
adoption of the initial mass function and the present-day mass function of
Chabrier (2003) for the bulge and disk lens populations, respectively.
Figure 12: Bayesian posteriors of the distances to the planetary systems.
Notations are same as those in Fig. 11.
In the second step, we computed the time scales and Einstein radii of the
simulated events using the relations in Equation (6), and then constructed
posterior distributions of $M$ and $D_{\rm L}$ for the simulated events with
the $t_{\rm E}$ and $\theta_{\rm E}$ values that are consistent with the
observables of the individual lensing events. For the three events
OGLE-2017-BLG-1691, KMT-2021-BLG-1303, and KMT-2021-BLG-1554, we use the
constraints of both $t_{\rm E}$ and $\theta_{\rm E}$, and for
KMT-2021-BLG-0320, we apply the constraint of only $t_{\rm E}$ because
$\theta_{\rm E}$ is not measured for the event.
Figures 11 and 12 show the Bayesian posteriors of $M$ and $D_{\rm L}$ of the
individual events, respectively. In Table 9, we summarize the estimated values
of the host and planet masses, $M_{\rm host}$ and $M_{\rm planet}$, distance,
and projected host-planet separation, $a_{\perp}=sD_{\rm L}\theta_{\rm E}$,
for all of which medians are presented as representative values and the
uncertainties are estimated as 16% and 84% of the Bayesian posterior
distributions. The median values and uncertainty ranges of the individual
events are marked by solid and dotted vertical lines in the Bayesian
posteriors presented in Figures 11 and 12, respectively. For the events with
degenerate solutions, OGLE-2017-BLG-1691, KMT-2021-BLG-0320 and
KMT-2021-BLG-1554, we present the planetary separations $a_{\perp}$
corresponding to both the close and wide (or inner and outer) solutions. We
note that the uncertainties of the estimated physical parameters are big
because of the intrinsically weak Bayesian constraint. For example, from the
comparison of the posterior distributions of OGLE-2017-BLG-1691 and
KMT-2021-BLG-0320, one finds that the difference between the two posterior
distributions is not significant, although the mean lens mass for the former
event is slightly bigger than that of the latter event due to the longer time
scale and the uncertainty is smaller due to the additional constrain of
$\theta_{\rm E}$.
The estimated masses indicate that all of the discovered planets have sub-
Jovian masses. The planet masses KMT-2021-BLG-0320L, KMT-2021-BLG-1303L, and
KMT-2021-BLG-1554L correspond to $\sim 0.10$, $\sim 0.38$, and $\sim 0.12$
times of the mass of the Jupiter, and the planet mass of OGLE-2017-BLG-1691L
corresponds to that of Uranus. To be noted among the planet hosts is that the
estimated mass of the planetary system KMT-2021-BLG-1554L, $M_{\rm host}\sim
0.08\leavevmode\nobreak\ M_{\odot}$, corresponds to the boundary between a
star and a brown dwarf (BD). Together with the previously detected planetary
systems with very low-mass hosts444MOA-2011-BLG-262 (Bennett et al., 2014),
OGLE-2012-BLG-0358 (Han et al., 2013), MOA-2015-BLG-337 (Miyazaki et al.,
2018), OGLE-2015-BLG-1771 (Zhang et al., 2020), KMT-2016-BLG-1820 (Jung et
al., 2018a), KMT-2016-BLG-2605 (Ryu et al., 2021), OGLE-2017-BLG-1522 (Jung et
al., 2018b), OGLE-2018-BLG-0677 (Herrera-Martín et al., 2020),
KMT-2018-BLG-0748 (Han et al., 2020b), and KMT-2019-BLG-1339L (Han et al.,
2020a), the discovery of the system demonstrates the microlensing capability
of detecting planetary systems with very faint or substellar hosts. Besides
KMT-2021-BLG-1554L, the host stars of the other planetary systems are low-mass
stars with masses lying in the range of $\sim[0.3$–$0.6]\leavevmode\nobreak\
M_{\odot}$. Under the approximation that the snow line distance scales with
the host mass as $a_{\rm sl}\sim 2.7(M/M_{\odot})$ AU, all discovered planets
lie beyond the snow lines of the hosts regardless of the solutions,
demonstrating the high microlensing sensitivity to cold planets. The planetary
systems lie in the distance range of $\sim[6.3$–$7.6]$ kpc, demonstrating the
usefulness of the microlensing method in detecting remote planets.
## 6 Summary and conclusion
We presented the analyses of four planetary microlensing events
OGLE-2017-BLG-1691, KMT-2021-BLG-0320, KMT-2021-BLG-1303, and
KMT-2021-BLG-1554. The events share a common characteristic that the planetary
signals appeared as anomalies with very short durations, ranging from a few
hours to a couple of days, and they were clearly detected solely by the
combined data of the high-cadence lensing surveys without additional data from
followup observations.
From the detailed analyses of the events, it was found that the signals were
generated by planets with low planet-to-host mass ratios: three of the
planetary systems with mass ratios of the order of $10^{-4}$ and the other
with a mass ratio slightly greater than $10^{-3}$. In the histogram of
microlensing planets presented in Figure 1, we mark the positions of the four
planets discovered in this work. The estimated masses indicated that all
discovered planets have sub-Jovian masses, in which the planet masses of
KMT-2021-BLG-0320Lb, KMT-2021-BLG-1303Lb, and KMT-2021-BLG-1554Lb correspond
to $\sim 0.10$, $\sim 0.38$, and $\sim 0.12$ times of the mass of the Jupiter,
and the mass of OGLE-2017-BLG-1691Lb corresponds to that of the Uranus. It was
found that the host of the planetary system KMT-2021-BLG-1554L has a mass at
around the boundary between a star and a brown dwarf. Besides this system, it
was found that the host stars of the other planetary systems are low-mass
stars with masses in the range of $\sim[0.3$–$0.6]\leavevmode\nobreak\
M_{\odot}$. The discoveries of the planets well demonstrate the capability of
the current high-cadence microlensing surveys in detecting low-mass planets.
###### Acknowledgements.
Work by C.H. was supported by the grants of National Research Foundation of
Korea (2020R1A4A2002885 and 2019R1A2C2085965). This research has made use of
the KMTNet system operated by the Korea Astronomy and Space Science Institute
(KASI) and the data were obtained at three host sites of CTIO in Chile, SAAO
in South Africa, and SSO in Australia. The OGLE project has received funding
from the National Science Centre, Poland, grant MAESTRO 2014/14/A/ST9/00121 to
AU. he MOA project is supported by JSPS KAKENHI grant Nos. JSPS24253004,
JSPS26247023, JSPS23340064, JSPS15H00781, JP16H06287, JP17H02871, and
JP19KK0082.
## References
* Alard & Lupton (1998) Alard, C., & Lupton, R. H. 1998, ApJ, 503, 325
* Albrow (2017) Albrow, M. 2017, MichaelDAlbrow/pyDIA: Initial Release on Github,Versionv1.0.0, Zenodo, doi:10.5281/zenodo.268049
* Albrow et al. (2009) Albrow, M., Horne, K., Bramich, D. M., et al. 2009, MNRAS, 397, 2099
* An (2005) An, J. H. 2005, MNRAS, 356, 1409
* Bennett et al. (2014) Bennett, D. P., Batista, V., Bond, I. A., et al. 2014, ApJ, 785, 155
* Bennett & Rhie (1996) Bennett, D. P., & Rhie, S. H. 1996, ApJ, 472, 660
* Bensby et al. (2013) Bensby, T., Yee, J. C., Feltzing, S., et al. 2013, A&A, 549, A147
* Bessell & Brett (1988) Bessell, M. S., & Brett, J. M. 1988, PASP, 100, 1134
* Bond et al. (2001) Bond, I. A., Abe, F., Dodd, R. J., et al. 2001, MNRAS, 327, 868
* Chabrier (2003) Chabrier, G. 2003, PASP, 115, 763
* Dominik (1999) Dominik, M. 1999, A&A, 349, 108
* Dominik (2019) Dominik, M., Bachelet, E., Bozza, V., et al. 2019, MNRAS, 484, 5608
* Gaia Collaboration (2016) Gaia Collaboration, Prusti, T., de Bruijne, J. H. J., et al. 2016, A&A, 595, A1
* Gaia Collaboration (2018) Gaia Collaboration, Brown, A. G. A., Vallenari, A., et al. 2018, A&A, 616, A1
* Gaudi (1998) Gaudi, B. S. 1998, ApJ, 506, 533
* Gaudi & Gould (1997) Gaudi, B. S., & Gould, A. 1997, ApJ, 486, 85
* Gould (1992) Gould, A. 1992, ApJ, 392, 442
* Gould & Loeb (1992) Gould, A., & Loeb, A. 1992, ApJ, 396, 104
* Gould (2000) Gould, A. 2000, ApJ, 542, 785
* Griest & Safizadeh (1998) Griest, K., & Safizadeh, N. 1998, ApJ, 500, 37
* Han et al. (2022a) Han, C. Bond, I. A., Yee, J. C., et al. 2022a, A&A, 658, A62
* Han & Gould (1995) Han, C., & Gould, A. 1995, ApJ, 447, 53
* Han & Gould (2003) Han, C., & Gould, A. 2003, ApJ, 592, 172
* Han et al. (2022b) Han, C., Gould, A., Albrow, M. D., et al. 2022b, A&A, 658, A94
* Han et al. (2013) Han, C., Jung, Y. K., Udalski, A., et al. 2013, ApJ, 778, 38
* Han et al. (2020a) Han, C., Kim, D., Udalski, A., et al. 2020a, AJ, 160, 64
* Han et al. (2020b) Han, C., Shin, I.-G., Jung, Y. K., et al. 2020b, A&A, 641, A105
* Han et al. (2021) Han, C., Udalski, A., Kim, D., et al. 2021, A&A, 655, A21
* Han et al. (2022c) Han, C., Udalski, A., Lee, C.-U., et al. 2022c, A&A, 658, A93
* Herrera-Martín et al. (2020) Herrera-Martín, A., Albrow, A., Udalski, A., et al. 2020, AJ, 159, 134
* Holtzman et al. (1998) Holtzman, J. A., Watson, A. M., Baum, W. A., Grillmair, C. J., Groth, E. J., Light, R. M., Lynds, R., & O’Neil, E. J.,Jr. 1998, AJ, 115, 1946
* Hwang et al. (2022) Hwang, K.-H., Zang, W., Gould, A., et al. 2022, AJ, 163, 43
* Jung et al. (2021) Jung, Y. K., Han, C., Udalski, A., et al. 2021, AJ, 161, 293
* Jung et al. (2018a) Jung, Y. K., Hwang, K.-H., Ryu, Y.-H., et al. 2018a, AJ, 156, 208
* Jung et al. (2018b) Jung, Y. K., Udalski, A., Gould, A., et al. 2018b, AJ, 155, 219
* Kervella et al. (2004) Kervella, P., Thévenin, F., Di Folco, E., & Ségransan, D. 2004, A&A, 426, 29
* Kim et al. (2018) Kim, D.-J., Kim, H.-W., Hwang, K.-H., et al. 2018, AJ, 155, 76
* Kim et al. (2016) Kim, S.-L., Lee, C.-U., Park, B.-G., et al. 2016, JKAS, 49, 37
* Miyazaki et al. (2018) Miyazaki, S., Sumi, T., Bennett, D. P., et al. 2018, AJ, 156, 136
* Nataf et al. (2013) Nataf, D. M., Gould, A., Fouqué, P., et al. 2013, ApJ, 769, 88
* Robin et al. (2003) Robin, A. C., Reylé, C., Derriére, S., & Picaud, S. 2003, A&A, 409, 523
* Ryu et al. (2021) Ryu, Y. H., Hwang, K.-H., Gould, A., et al. 2021, AJ, 162, 96.
* Ryu et al. (2022) Ryu, Y. H., Jung Y. K., Yang, H., et al. 2022, in preparation
* Shin et al. (2022) Shin, I.-G., Yee, J. C., Hwang K.-H., et al. 2022, AJ, in press (arXiv:2201.04312)
* Tomaney & Crotts (1996) Tomaney, A. B., & Crotts, A. P. S. 1996, AJ, 112, 2872
* Udalski et al. (2015) Udalski, A., Szymański, M. K., & Szymański, G. 2015, Acta Astron., 65, 1
* Woźniak (2000) Woźniak, P. R. 2000, Acta Astron., 50, 42
* Yee et al. (2012) Yee, J. C., Shvartzvald, Y., Gal-Yam, A., et al. 2012, ApJ, 755, 102
* Yee et al. (2021) Yee, J. C., Zang, W, Udalski, A., et al. 2021, AJ, 162, 180
* Yoo et al. (2004) Yoo, J., DePoy, D. L., Gal-Yam, A., et al. 2004, ApJ, 603, 139
* Zang et al. (2021a) Zang, W., Han, C., Kondo, I., et al. 2021, Rearcxh in Astron and Astroph., 2021a, 21, 239
* Zang et al. (2021b) Zang, W., Hwang, K.-H., Udalski, A., et al. 2021b, AJ, 162, 163
* Zhang et al. (2022) Zhang, K., Gaudi, B. S., & Bloom, J. S. 2022, Nature Astronomy, submitted, arXiv:2111.13696
* Zhang et al. (2020) Zhang, X., Zang, W., Udalski, A., et al. 2020, AJ, 159, 116
|
# Appraisal Theories for Emotion Classification in Text
Jan Hofmann1, Enrica Troiano1, Kai Sassenberg2,3 Roman Klinger1
1Institut für Maschinelle Sprachverarbeitung, University of Stuttgart, Germany
2Leibniz-Institut für Wissensmedien, Tübingen, Germany
3University of Tübingen, Germany
<EMAIL_ADDRESS>
<EMAIL_ADDRESS>
###### Abstract
Automatic emotion categorization has been predominantly formulated as text
classification in which textual units are assigned to an emotion from a
predefined inventory, for instance following the fundamental emotion classes
proposed by Paul Ekman (fear, joy, anger, disgust, sadness, surprise) or
Robert Plutchik (adding trust, anticipation). This approach ignores existing
psychological theories to some degree, which provide explanations regarding
the perception of events. For instance, the description that somebody
discovers a snake is associated with fear, based on the appraisal as being an
unpleasant and non-controllable situation. This emotion reconstruction is even
possible without having access to explicit reports of a subjective feeling
(for instance expressing this with the words “I am afraid.”). Automatic
classification approaches therefore need to learn properties of events as
latent variables (for instance that the uncertainty and the mental or physical
effort associated with the encounter of a snake leads to fear). With this
paper, we propose to make such interpretations of events explicit, following
theories of cognitive appraisal of events, and show their potential for
emotion classification when being encoded in classification models. Our
results show that high quality appraisal dimension assignments in event
descriptions lead to an improvement in the classification of discrete emotion
categories. We make our corpus of appraisal-annotated emotion-associated event
descriptions publicly available.
00footnotetext: This work is licensed under a Creative Commons Attribution
4.0 International License. License details:
http://creativecommons.org/licenses/by/4.0/.
## 1 Introduction
The task of emotion analysis is commonly formulated as classification or
regression in which textual units (documents, paragraphs, sentences, words)
are mapped to a predefined reference system, for instance the sets of
fundamental emotions fear, anger, joy, surprise, disgust, and sadness proposed
by ?), or by ?), which includes also trust and anticipation. Machine learning-
based models need to figure out which words point to a particular emotion
experienced by a reader, by the author of a text, or a character in it.
Depending on the resource which has been annotated, the description of an
emotion experience can vary. On Twitter, for instance, other than direct
reports of an emotion state (“I feel depressed”), hashtags are used as emotion
labels to enrich the description of events and stances (“I just got my exam
result #sad”). In news articles, emotional events are sometimes explicitly
mentioned (“couple infuriate officials” [Bostan et al., 2020]) and other times
require world knowledge (“Tom Cruise and Katie Holmes set wedding date”,
labeled as surprise [Strapparava and Mihalcea, 2007]). In literature, a
sequence of events which forms the narrative leads to an emotion in the
reader. In this paper, we focus on those texts which communicate emotions
without an explicit emotion word, but rather describe events for which an
emotion association is evident.
Such textual examples became popular in natural language processing research
with the use of the data generated in the ISEAR project [Scherer and Wallbott,
1997]. The project led to a dataset of descriptions of events triggering
specific affective states, which was originally collected to study event
interpretations with a psychological focus. In text analysis, to infer the
emotion felt by the writers of those reports, an event interpretation needs to
be accomplished. For instance, in the text “When a car is overtaking another
and I am forced to drive off the road”, the model needs to associate the event
with fear.
To date, nearly all computational approaches that associate text with emotions
are agnostic to the way in which emotions are communicated, they do “not know”
how to interpret events, but, presumably, they purely learn word associations
instead of actual event interpretations. One might argue that approaches
predicting fine-grained dimensions of affect, namely arousal and valence,
actually tackle this problem [Buechel and Hahn, 2017, Preoţiuc-Pietro et al.,
2016]. However, these typically do not infer downstream emotion categories.
Further, particularly regarding events, psychological theories offer more
detailed information. As an example, the emotion component model [Scherer,
2005] advocates that cognitive appraisal dimensions underly discrete emotion
classes [Smith and Ellsworth, 1985]. These appraisal dimensions111The examples
follow the results by ?), an excerpt is shown in Table 1. evaluate (1) how
pleasant an event is (pleasantness, likely to be associated with joy, but
unlikely to appear with disgust), (2) how much effort an event can be expected
to cause (anticipated effort, likely to be high when anger or fear is
experienced), (3) how certain the experiencer is in a specific situation
(certainty, low, e.g., in the context of hope or surprise), (4) how much
attention is devoted to the event (attention, likely to be low, e.g., in the
case of boredom or disgust), (5) how much responsibility the experiencer of
the emotion holds for what has happened (self-other responsibility/control,
high for feeling challenged or pride), and (6) how much the experiencer has
control over the situation (situational control, low in the case of anger).
Despite their richness, cognitive theories of appraisal and their empirical
results have not been exploited for emotion prediction in text yet. We fill
this gap with this paper and analyze the relation between appraisal dimensions
and emotion categories in a text classification setting. We post-annotate an
English emotion corpus of self-reports of emotion events [Troiano et al.,
2019], which already contains annotations related to the emotions of anger,
disgust, fear, guilt, joy, sadness, and shame, and add the appraisal
dimensions by ?) mentioned above. Further, we analyze if an automatic
prediction of these dimensions from text is possible with standard neural
methods, and if these predictions contribute to emotion classifications. Our
main contributions are: (1) the first event-centered text corpus annotated
with appraisal dimensions; (2) the evaluation of how well text classification
models can recognize these appraisal dimensions; (3) we show emotion
classification benefits from the information of appraisal dimensions, when
high quality predictions of these are available. Further, (4), we replicate
the study by ?) from a CL/NLP perspective, based on textual event
descriptions.
## 2 Background on Emotion Psychology and Analysis
### 2.1 Emotion and Affect Theories
As a component of humans’ life, emotions have been thoroughly studied in the
field of psychology, where they are generally deemed responses to salient
events. The debates surrounding their definition, however, has never come to a
consensus, producing a varied literature on the topic. This has a clear
implication for computational emotion analyses, for they must choose and
follow one of the available psychological theories in order to motivate the
emotion phenomenon that they research in language.
Some of such theories focus on the evolutionary function of emotions, and
accordingly, on their link to actions [Izard, 1971, Tooby and Cosmides, 2008].
The core idea is that emotions help humans accomplish every-day life tasks and
communicate socially relevant information by triggering specific physiological
symptoms. In particular, there are patterns of behaviour (e.g., smiling) that
reflect discrete emotion terms (e.g., joy), which suggests that emotional
states can be grouped based on a few natural language categories. One of the
most popular sources for a set of fundamental emotions is the theory by ?).
Ekman studied the relation between emotions and both culture and facial
expressions: he claimed that the set of fundamental emotions, namely, anger,
disgust, fear, joy, sadness, and surprise can be distinguished by facial
muscular movements across cultures (which is partially doubted these days,
?)). As an addition to this model, ?) makes explicit the assumption that
different fundamental emotions can occur together, for instance trust and joy,
which is the case when love is experienced. Such emotion mixtures, as well as
an opposition between anger and fear, joy and sadness, surprise and
anticipation, trust and disgust, has been included in this model. In natural
language processing, mostly a set of four to eight fundamental emotions is
used, where anger, fear, joy, and sadness are shared by most approaches (an
exception with 24 emotion classes is ?)).
A diametrically opposite view is held by the constructivist tradition
[Averill, 1980, Oatley, 1993, Barrett and Russell, 2015], in which actions and
physiological changes are the building blocks that construct emotions, rather
than their direct effect [Feldman Barrett, 2006]. Feeling an emotion means
categorizing the fluctuations of an affect system along some components. For
instance, the affect components _valence_ (degree of polarity), _arousal_
(degree of excitement), and _dominance_ (degree of control over a situation)
[Posner et al., 2005] are used as dimensions to describe affect experiences in
a 3-dimensional space, which can then be mapped to discrete emotion
categories.
### 2.2 Theories of Cognitive Appraisal
Extensions to this model in terms of underlying components are the works of
?), ?) and ?), who qualified emotions as component processes that arise to
face salient circumstances: an emotion is an “episode of interrelated,
synchronized changes in the states of all or most of the five organismic
subsystems in response to the evaluation of an external or internal stimulus-
event as relevant to major concerns of the organism” [Scherer et al., 2001].
According to this view, there is an appraisal, that is, an information
processing component, which enables people to determine the significance of a
situation with respect to their needs and values. In the context of this
appraisal (e.g., judging a snake as dangerous), the resources of four other
components are mobilitated to deal with the situation. These are, next to the
_cognitive component (appraisal)_ , a _neurophysiological component (bodily
symptoms)_ , a _motivational component (action tendencies)_ , a _motor
expression (facial and vocal expression)_ , and a _subjective feeling
component (emotional experience)_ [Scherer, 2005, Table 1].
While the notions of subjective experience and bodily symptoms were common to
other emotion theories, appraisal represents a novelty that fills in some
shortcomings of basic models. First, it explains how emotions are elicited.
The origin of emotions is to be seen in the stimulus as appraised rather than
in the stimulus as such. Second, appraisals provide a structured account for
the differences among emotions. For instance, anger and fear are experienced
when the evaluation of a negative event attributes it to external factors,
whereas guilt and shame are felt if the causes of such event are identified in
the self, as stable and uncontrollable personality traits, like in the case of
shame (e.g., “I’m dumb”), or unstable and controllable behaviours for guilt
(e.g., “I did not observe the speed limit”) [Tracy and Robins, 2006].
We argue in this paper that this makes appraisals particularly useful for
natural language processing, because they both provide a framework for
research and represent a way of enriching existing data. As a matter of fact,
few dimensions are sufficient to explain emotions based on cognitive
appraisal. ?) explain 15 emotions by leveraging pleasantness (polarity), self-
other responsibility/control (for initiating the situation), certainty (about
what is going on), attention (whether the emotion stimulus is worth
attending), anticipated effort (the amount of physical or mental activation
before the stimulus), and situational control (the ability to cope with the
situation). Compared to the valence-arousal-dominance model, where it is left
unclear if the polarity dimension refers to a quality of the emotion stimulus
or a quality of the feeling [Scherer, 2005], all these dimensions are
unambiguously event-directed. In this paper, we focus on modelling the
cognitive components described by ?). We show their main findings in Table 1,
limited to the emotions that are available in the corpus we use.
Emotion | Unpleasant | Responsibility | Uncertainty | Attention | Effort | Control
---|---|---|---|---|---|---
Happiness | $-$1.46 | 0.09 | $-$0.46 | 0.15 | $-$0.33 | $-$0.21
Sadness | 0.87 | $-$0.36 | 0.00 | $-$0.21 | $-$0.14 | 1.15
Anger | 0.85 | $-$0.94 | $-$0.29 | 0.12 | 0.53 | $-$0.96
Fear | 0.44 | $-$0.17 | 0.73 | 0.03 | 0.63 | 0.59
Disgust | 0.38 | $-$0.50 | $-$0.39 | $-$0.96 | 0.06 | $-$0.19
Shame | 0.73 | 1.31 | 0.21 | $-$0.11 | 0.07 | $-$0.07
Guilt | 0.60 | 1.31 | $-$0.15 | $-$0.36 | 0.00 | $-$0.29
Table 1: The locations of emotions along appraisal dimensions (according to a
principle component analysis) as published by ?), Table 6, filtered to those
emotions which are available in the text corpus we use.
### 2.3 Automatic Emotion Classification
Previous work on emotion analysis in natural language processing focuses
either on resource creation or on emotion classification for a specific task
and domain. On the side of resource creation, the early and influential work
of ?) is a dictionary of words being associated with different psychologically
relevant categories, including a subset of emotions. Later, ?) made WordNet
Affect available to target word classes and differences regarding their
emotional connotation, ?) released the NRC dictionary with more than 14,000
words for a set of discrete emotion classes, and a valence-arousal-dominance
dictionary was provided by ?). ?) have developed a methodological framework to
adapt existing affect lexicons to specific use cases. Other than dictionaries,
emotion analysis relies on labeled corpora. Some of them include information
relative to valence and arousal [Buechel and Hahn, 2017, Preoţiuc-Pietro et
al., 2016], but the majority of resources use discrete emotion classes, for
instance to label fairy tales [Alm et al., 2005], blogs [Aman and Szpakowicz,
2007], tweets [Mohammad et al., 2017, Schuff et al., 2017, Mohammad, 2012,
Mohammad and Bravo-Marquez, 2017, Klinger et al., 2018], Facebook posts
[Preoţiuc-Pietro et al., 2016], news headlines [Strapparava and Mihalcea,
2007], dialogues [Li et al., 2017], literary texts [Kim et al., 2017], or self
reports on emotion events [Scherer and Wallbott, 1997, Troiano et al., 2019].
We point the reader to the survey by ?) for a more an overview.
Most automatic methods that assign labels to text rely on machine learning
[Alm et al., 2005, Aman and Szpakowicz, 2007, Schuff et al., 2017, i.a.].
Recent shared tasks showed an increase in transfer learning from generic
representations [Klinger et al., 2018, Mohammad et al., 2018, Mohammad and
Bravo-Marquez, 2017]. ?) proposed to use emoji representations for
pretraining, and ?) performed pretraining on existing emotion corpora followed
by fine-tuning for a specific domain for which only little training data was
available.
We are only aware of one preliminary study which considered appraisal
dimensions to improve text-based emotion prediction, namely ?). In their
study, subjects labeled 200 stories with 38 appraisal features (which remain
unmentioned), to evaluate if a text-based representation adds on top of an
fMRI-based classification. Apart from this study, all previous machine
learning-based approaches used models to predict emotions or affect values
directly from text, without any access to appraisal dimensions. Only a couple
of works incorporated cognitive components, for instance those coming from the
OCC model (named after the authors Ortony, Clore and Collins’ initials), which
sees every appraisal as an evaluation of the pleasantness of events, objects,
or actions with respect to one’s goals, tastes or behavioural and moral
standards [Clore and Ortony, 2013]. Based on the OCC model, ?) devised a rule-
based approach to interpret text. They did not explicitly formulate their
model following appraisal theories, but they moved towards a cognitively-
motivated interpretation of events and interpersonal descriptions. Others have
adopted patterns of appraisal to predict the emotions triggered by actions, as
described in a text. Specifically, ?) and ?) have created EmotiNet, a
knowledge base of action chains that includes information about the elements
on which the appraisal is performed within an affective situation, namely, the
agent, the action and the object involved in a chain. We share their
motivation to delve into event representations based on the descriptions of
their experiencers. Unlike their work, ours explicitly encodes appraisal
dimensions and uses the classification into these categories for emotion
prediction.
We are not aware of any textual corpus annotated with appraisals according to
emotion theories, or associated computational models. However, some research
exists which studies these theories with added value for computational models.
?), e.g., explicitly encode realizations of appraisals of events and compare
those between production and perception of facial expressions. ?) model
emotions in video sequences computationally, based on appraisals of events. ?)
inform intelligent agents about emotions based on formalizations of cognitive
appraisal. Further, in ?), the chapter authored by ?) argues for an integrated
computational approach of emotions which also embeds appraisal theories.
## 3 Corpus
The main objective of this study is to understand the relation between
appraisal dimensions and emotion categories. Therefore, we build appraisal
annotations on top of enISEAR, an existing corpus of 1001 English event
descriptions which are already labeled with the discrete categories of anger,
disgust, fear, guilt, joy, sadness, and shame [Troiano et al., 2019]. Each
instance has been generated by a crowdworker on the platform FigureEight by
completing the sentence “I feel [emotion name], when …”. This corpus has an
advantage over the original ISEAR resource because it has a German counterpart
which can be used in further studies; moreover, its emotion labels have been
intersubjectively validated. The corpus, annotations, and our implementations
are available at http://www.ims.uni-stuttgart.de/data/appraisalemotion.
### 3.1 Annotation
One presumable challenge in the post-annotation of events regarding the
appraisal dimensions is that our annotators do not have access to the private
state of the event experiencers. However, under the assumption that events are
perceived similarly in subjective feeling and evaluated comparably based on
cognitive appraisal, we assume that this is not a major flaw in the design of
the study. An alternative would have been to perform the text generation task
as ?) did, but asking the authors of event descriptions for their appraisal in
addition. We opted against such procedure as it would have meant to reproduce
an existing study in addition to our research goal.
For the post-labeling of enISEAR, we aimed at formulating unambiguous and
intuitive descriptions of appraisal dimensions, which would be faithful to
those in ?). As opposed to the subjects of their study, however, our
annotators had to judge events that they did not personally experience. For
this reason, we simplified our annotation guidelines in two respects. First,
we opted for a binary setting, while ?) used continuous scales to rate
discrete emotion categories on the appraisal dimensions. Second, we split
control into _Control_ and _Circumstance_ (i.e., _self_ and situational
control), in line with the discussion of this variable by ?, p. 824f.), while
retaining the category of responsibility. This was motivated by the
observation of a low inter-annotator agreement in preliminary annotation
rounds and a series of discussions that revealed the difficulty for annotators
to separate the concepts of responsibility and self control. Then, the
annotators were instructed to read an event description, without having access
to the emotion label, and to answer the following questions:
Most probably, at the time when the event happened, the writer…
* •
…wanted to devote further attention to the event. (Attention)
* •
…was certain about what was happening. (Certainty)
* •
…had to expend mental or physical effort to deal with the situation. (Effort)
* •
…found that the event was pleasant. (Pleasantness)
* •
…was responsible for the situation. (Responsibility)
* •
…found that he/she was in control of the situation. (Control)
* •
…found that the event could not have been changed or influenced by anyone.
(Circumstance)
Each event description from enISEAR was judged by three annotators between the
age of 26 and 29. One of them is a female Ph.D. student of computational
linguistics, the others are male graduate students of software engineering.
Two of the annotators are co-authors of this paper. The judges familiarised
themselves with their task through four training iterations. At every
iteration, we hand-picked 15–20 samples from the ISEAR dataset [Scherer and
Wallbott, 1997], such that instances used for training would not be seen
during the actual annotation, but had a comparable structure. Dissimilarities
in the annotation were discussed in face-to-face meetings and the annotation
guideline was refined.
The agreement improved from $\kappa$=$0.62$ to 0.67 in the four iterations. In
one of them, we experimented with giving access to the emotion label, which
lead to a large improvement in agreement ($\kappa$=$0.83$). Nevertheless, we
decided to continue without this information, in order to evaluate the
annotator’s performance in a similar setting as we evaluate the automatic
model – to predict appraisal for emotion classification. We show the pairwise
inter-annotator scores of the final set in Table 2. The agreement scores
between the different annotator pairs are comparable.
| Cohen’s $\kappa$
---|---
| between annotators | annotator–majority
Appraisal Dimension | A1/A2 | A1/A3 | A2/A3 | avg. | A1 | A2 | A3 | avg.
Attentional Activity | .28 | .24 | .41 | .31 | .50 | .76 | .66 | .64
Certainty | .41 | .23 | .29 | .31 | .62 | .77 | .46 | .62
Anticipated Effort | .38 | .33 | .26 | .32 | .69 | .67 | .62 | .66
Pleasantness | .89 | .88 | .90 | .89 | .93 | .96 | .94 | .94
Responsibility | .68 | .57 | .63 | .63 | .80 | .88 | .76 | .81
Control | .65 | .56 | .52 | .58 | .84 | .81 | .70 | .78
Circumstance | .52 | .32 | .28 | .37 | .80 | .69 | .49 | .66
Average | .59 | .48 | .52 | .53 | .77 | .82 | .70 | .76
Table 2: Cohen’s $\kappa$ between all annotator pairs and between each
annotator and the majority vote.
These scores tell us that rating appraisal dimensions for given events is
challenging, and its difficulty varies depending on the categories. Given the
comparably low agreement obtained for a subset of dimensions, we opt for a
“crowd-sourcing”-like aggregation by taking the majority vote to form the
final annotation, included in Table 2, on the right side of the table. We
observe that the agreement between majority vote and each annotator is
constantly above $\kappa$=.62, which is an acceptable agreement
(avg.$\kappa$=.76).
| Appraisal Dimension
---|---
Emotion | Attention | Certainty | Effort | Pleasant | Respons. | Control | Circum.
Anger | 129 | .90 | 119 | .83 | 60 | .42 | 0 | .00 | 9 | .06 | 1 | .01 | 5 | .03
Disgust | 67 | .47 | 134 | .94 | 40 | .28 | 2 | .01 | 14 | .10 | 11 | .08 | 24 | .17
Fear | 129 | .90 | 13 | .09 | 121 | .85 | 4 | .03 | 43 | .30 | 18 | .13 | 66 | .46
Guilt | 55 | .38 | 132 | .92 | 36 | .25 | 0 | .00 | 133 | .93 | 88 | .62 | 11 | .08
Joy | 139 | .97 | 140 | .98 | 4 | .03 | 141 | .99 | 65 | .45 | 41 | .29 | 25 | .17
Sadness | 122 | .85 | 112 | .78 | 88 | .62 | 1 | .01 | 7 | .05 | 2 | .01 | 97 | .68
Shame | 32 | .22 | 111 | .78 | 51 | .36 | 1 | .01 | 106 | .74 | 67 | .47 | 12 | .08
Total | 673 | | 761 | | 400 | | 149 | | 377 | | 228 | | 240 |
Table 3: Instance counts and ratios across emotions and appraisal annotations.
### 3.2 Analysis
In Table 3 are the cooccurrence counts across emotion and appraisal dimension
pairs, as well as the relative counts normalized by emotion (enISEAR provides
143 descriptions per emotion). The most frequently annotated class is
certainty, followed by attention. Appraisal dimensions are differently
distributed across emotions: anger and fear require attention, guilt and shame
do not; disgust and anger show the highest association with certainty, in
opposition to fear. _Responsibility_ and control play the biggest role in
guilt and shame, while joy, non-surprisingly, strongly relates to
pleasantness. _Fear_ has a clear link with anticipated effort and, together
with sadness, it is characterized by the inability to control the
circumstance.
These numbers are particularly interesting in comparison with the findings of
?), who report the average scores along the appraisal dimensions (based on a
principle component analysis) for each emotion222We report the subset of
emotions that overlap with ours. Also note that their “Control” corresponds to
our “Circum.”.. Results are consistent in most cases. For instance, joy (or
happiness in Table 1) stands out as highly pleasant and barely related to
anticipated effort. Self responsibility is lowest in anger, an emotion that
arises when blame is externalized, and mostly present in shame and guilt,
which derive from blaming the self [Tracy and Robins, 2006]. These two are
also the emotions that annotators associated with control more than others.
Attention is prominent for events that elicited anger and which were under the
control of others, as suggested by the low situational control. The highest
situational control, on the contrary, appears with the data points labeled as
fear, also characterized by a strong feeling of uncertainty and anticipated
effort. There are also dissimilarities between the two tables, like the level
of attention, reaching the lowest score for disgust in their study and not in
ours. They also find that situational control is a stronger indicator for
shame than for guilt, while anticipated effort is more marked in our sadness-
related events than theirs.
These differences may partly be data-specific, partly due to the type of
metrics shown in the tables. Most importantly, they can be traced to the
annotation setup: while their subjects recalled and appraised personal events,
our annotators evaluated the descriptions of events that are foreign to them.
It should be noted that for a reader/annotator it is challenging to
impersonate in the writer: although some events have a shared understanding
(e.g., “I passed the exam” is most likely appraised as pleasantness), others
are tied to one’s personal background, values and preferences. This may
represent a source of disaccord both between the tables, among the annotators,
and with the emotion labels themselves (e.g., “I felt … when my mom offered me
curry” has a pleasant gold label, while the original author meant it as a
negative emotion, namely disgust).
## 4 Experiments
We now move to our evaluation if automatic methods to recognize emotions can
benefit from being informed about appraisal dimensions explicitly. We first
describe our models and how we address our research questions and then turn to
the results.
### 4.1 Model Configuration
Figure 1 illustrates the four different tasks addressed by our models. Task
T$\rightarrow$E is the prediction of _emotions from text_ , namely the
standard setting in nearly all previous work of emotion analysis. We use a
convolutional neural network (CNN) inspired by ?), with pretrained GloVe
(Glove840B) as a 300-dimensional embedding layer [Pennington et al.,
2014]333https://nlp.stanford.edu/projects/glove/ with convolution filter sizes
of 2, 3, and 4 with a ReLu activation function [Nair and Hinton, 2010],
followed by a max pooling layer of length 2 and a dropout of 0.5 followed by
another dense layer.
As another model to predict emotions, we use a pipeline based on two steps,
one to detect the appraisal from text, and the second to assign the
appropriate emotion to the appraisal. We refer to Task T$\rightarrow$A as the
step of identifying _appraisal from text_. We use the model configuration of
Task T$\rightarrow$E, except for the sigmoid activation function and binary
cross entropy loss instead of softmax and cross-entropy loss. As a second
step, Task A$\rightarrow$E predicts _emotion from appraisal_. The features are
seven boolean variables corresponding to the appraisal dimensions. We use a
neural network with two hidden layers with ReLU activation, followed by a
dropout of 0.5.444We do not perform any further hyperparameter search. We also
compared all model configurations to MaxEnt models to ensure that we do not
suffer from overfitting due to a large number of parameters, in comparison to
the limited training data. In all settings, the neural models were superior.
We therefore limit our explanations to those.
A disadvantage of the pipeline setting could be that the emotion prediction
needs to handle propagated errors from the first step, and that the first step
cannot benefit from what the second model learns. Therefore, we compare the
pipeline setting (T$\rightarrow$A, A$\rightarrow$E) with a multi-task learning
setting (T$\rightarrow$A/E). The model is similar to Task T$\rightarrow$E. The
convolutional layer is shared by the tasks of predicting _emotions from text_
and predicting _appraisal from text_ and we use two output layers, one for
emotion predictions with softmax activation and one for appraisal predictions
with sigmoid activation.
Figure 1: Tasks investigated in experiments on appraisal-driven emotion
analysis.
| T$\rightarrow$A
---|---
Appraisal | P | R | $\textrm{F}_{1}$
Attention | 81 | 84 | 82
Certainty | 84 | 86 | 85
Effort | 68 | 68 | 68
Pleasantness | 79 | 63 | 70
Responsibility | 74 | 68 | 71
Control | 63 | 49 | 55
Circumstance | 65 | 58 | 61
Macro avg. | 73 | 68 | 70
Micro avg. | 77 | 74 | 75
Table 4: Classifier performance on predicting appraisal dimensions.
| | | | | | | | | | | | | Oracle Ensembles
---|---|---|---|---|---|---|---|---|---|---|---|---|---
| T$\rightarrow$E | T$\rightarrow$A,A$\rightarrow$E | A$\rightarrow$E (Gold) | T$\rightarrow$A/E | T5pt 0pt 0pt -1pt$\rightarrow$A5pt 0pt 0pt -1pt$\rightarrow$E + T5pt 0pt 0pt -1pt$\rightarrow$E | T5pt 0pt 0pt -1pt$\rightarrow$A/E + T5pt 0pt 0pt -1pt$\rightarrow$E
Emotion | P | R | $\textrm{F}_{1}$ | P | R | $\textrm{F}_{1}$ | P | R | $\textrm{F}_{1}$ | P | R | $\textrm{F}_{1}$ | P | R | $\textrm{F}_{1}$ | P | R | $\textrm{F}_{1}$
Anger | 51 | 52 | 52 | 34 | 62 | 44 | 55 | 71 | 62 | 51 | 52 | 52 | 66 | 81 | 73 | 59 | 59 | 59
Disgust | 65 | 63 | 64 | 59 | 34 | 43 | 53 | 48 | 51 | 64 | 64 | 64 | 78 | 68 | 73 | 69 | 66 | 67
Fear | 69 | 71 | 70 | 55 | 55 | 55 | 79 | 78 | 78 | 70 | 68 | 69 | 76 | 77 | 77 | 73 | 75 | 74
Guilt | 47 | 42 | 44 | 38 | 50 | 43 | 57 | 70 | 63 | 45 | 42 | 44 | 60 | 63 | 62 | 58 | 54 | 56
Joy | 74 | 80 | 77 | 77 | 69 | 72 | 94 | 98 | 96 | 77 | 77 | 77 | 79 | 80 | 80 | 79 | 85 | 82
Sadness | 69 | 67 | 68 | 58 | 40 | 47 | 69 | 63 | 66 | 68 | 68 | 68 | 74 | 70 | 72 | 73 | 71 | 72
Shame | 44 | 45 | 45 | 36 | 24 | 29 | 56 | 35 | 43 | 43 | 43 | 43 | 58 | 51 | 54 | 51 | 52 | 52
Macro avg. | 60 | 60 | 60 | 51 | 48 | 48 | 66 | 66 | 65 | 60 | 59 | 59 | 70 | 70 | 70 | 66 | 66 | 66
Micro avg. | | | 60 | | | 48 | | | 66 | | | 59 | | | 70 | | | 66
Table 5: Comparison of the Text-to-Emotion baseline (T$\rightarrow$E) with the
performance of first prediction appraisal followed by emotion analysis
(T$\rightarrow$A,A$\rightarrow$E) and the multi-task setting
(T$\rightarrow$A/E). The oracle consists of a combination of two models and is
informed which model is more likely to make the correct prediction.
### 4.2 Results
We perform each experiment in a repeated 10$\times$10-fold cross-validation
setting and report average results. All partitions of training and test sets
are the same across all experiments.
Experiment 1 (Appraisal Prediction, T$\rightarrow$A) aims at understanding how
well appraisal dimensions can be predicted from text. Emotion classification
is an established task, and one might have some intuition on the expected
performance with a given data set, but the prediction of appraisal dimensions
has never been performed before. Hence, we report precision, recall, and
$\textrm{F}_{1}$ for each appraisal component considered in Table 4. The
prediction of certainty works best (85%$\textrm{F}_{1}$) followed by attention
(82%$\textrm{F}_{1}$). The lowest performance is seen for control
(55%$\textrm{F}_{1}$) and circumstance (61%$\textrm{F}_{1}$). These results
are only partially in line with the inter-annotator agreement scores. We
obtain a .75 micro average $\textrm{F}_{1}$ score.
| Appraisal |
---|---|---
| Emotion (G/P) | A | Ce | Ex | Px | Rx | Co | Ci | Text
Appr+Emo corr. | Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when my neighbour started to throw rubbish in my garden for no reason.
Disgust | 0 | 1 | 0 | 0 | 0 | 0 | 0 | to watch someone eat insects on television.
Fear | 1 | 0 | 1 | 0 | 0 | 0 | 1 | when our kitten escaped in the late evening and we thought he was lost.
Guilt | 0 | 1 | 0 | 0 | 1 | 1 | 0 | when I took something without paying.
Joy | 1 | 1 | 0 | 1 | 1 | 0 | 0 | when I found a rare item I had wanted for a long time.
Sadness | 1 | 1 | 1 | 0 | 0 | 0 | 1 | when my dog died. He was ill for a while. Still miss him.
Shame | 0 | 1 | 0 | 0 | 1 | 0 | 0 | when I remember an embarrassing social faux pas from my teenage years.
Emo incorr. | Anger/Fear | 1 | 0 | 1 | 0 | 0 | 0 | 0 | when someone drove into my car causing damage and fear to myself – then drove off before exchanging insurance details.
Disgust/Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I saw a bird being mistreated when on holiday.
Fear/Sadness | 1 | 1 | 1 | 0 | 0 | 0 | 1 | a huge spider just plopped on down on the sofa besides me, staring me out.
Guilt/Disgust | 0 | 1 | 0 | 0 | 0 | 0 | 0 | when I watched a documentary that showed footage of farms of pigs and chickens and as a meat eater I felt awful guilt at how they are treated.
Sadness/Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I saw a group of homeless people and it was cold outside.
Shame/Guilt | 0 | 1 | 0 | 0 | 1 | 1 | 0 | because I did something silly.
Ap+Emo incorr. | Anger/Shame | 0 | 1 | 0 | 0 | 1 | 0 | 0 | I feel … because I can’t stand when people lie.
Disgust/Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I saw a medical operation on a TV show.
Fear/Guilt | 1 | 0 | 0 | 0 | 1 | 0 | 0 | when I was on a flight as I am … of flying.
Guilt/Shame | 0 | 1 | 1 | 0 | 1 | 1 | 0 | when I lost my sister’s necklace that I had borrowed.
Joy/Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I saw bees coming back to my garden after few years of absence.
Sadness/Guilt | 1 | 1 | 0 | 0 | 1 | 0 | 0 | when I watched some of the sad cases of children in need.
Shame/Guilt | 0 | 1 | 0 | 0 | 1 | 1 | 0 | when I forgot a hairdressers appointment.
Table 6: Examples for the prediction of the pipeline setting (T$\rightarrow$A,
A$\rightarrow$E) . A: Attention, Ce: Certainty, E: Effort, P: Pleasantness, R:
Responsibility, Co: Control, Ci: Circumstance. First emotion mention is gold,
second is prediction. Appraisal shown is prediction with errors shown in bold.
Experiment 2 (Appraisal Interpretation, A$\rightarrow$E) aims at understanding
how well emotions can be predicted from appraisals. We compare the baseline
text-to-emotion setting (T$\rightarrow$E) to the pipeline setting that first
predicts the appraisal and then, from those, the emotion. In the pipeline
setting we train the second step (A$\rightarrow$E) on the gold appraisal
annotations, not on the predictions.555We also tested if training on the
prediction leads to better results, but it constantly underperformed. We
compare this setting to the performance of the appraisal-to-emotion model
(A$\rightarrow$E), when applied on gold appraisal annotations. This serves as
an upper bound which can be reached with the best-performing appraisal
prediction model.
We first turn to the results of the model which predicts the emotion based on
annotated (gold) appraisal dimensions (A$\rightarrow$E (gold)). Here, we
observe a clear improvement in contrast to the emotion classification which
has access to the text (T$\rightarrow$E). _Anger_ increases from .52 to .62,
disgust decreases from .64 to .51, fear increases from .70 to .78, guilt from
.44 to .63, joy from .77 to .96, sadness decreases from .68 to .66 and shame
decreased from .45 to .43. On micro average, the performance increases from
.60 to .66% $\textrm{F}_{1}$. These results are an upper-bound for the
performance that can be achieved with the pipeline model, under the assumption
of having access to perfect appraisal predictions.
When moving to the real-world setting of first predicting the appraisal
dimensions and then, based on those, predicting the emotion, the performance
scores drop from 66 to 48% $\textrm{F}_{1}$. This is an indicator that the
performance of our current appraisal prediction, though comparably reasonable
with 75% $\textrm{F}_{1}$, is not yet sufficient to support emotion
predictions, at least partially. The clear improvement in emotion prediction
based on perfect appraisal annotations and the performance drop in the real-
world setting for emotion prediction suggest that annotating more data with
appraisal dimensions is necessary to further develop our approach.
Finally, Experiment 3 (Multi-Task Learning, T$\rightarrow$A/E and Oracle
Ensembles) evaluates if the model which learns appraisals and emotions jointly
performs better than the pipeline model (both being models applicable in a
real-world scenario). We show these results also in Table 5 and see that the
multi-task learning model cannot improve over the text-only setting.
A remaining question is on the complementarity of the pipeline and the
multitask model to the T$\rightarrow$E model. To look into this, we define two
oracle ensembles (T$\rightarrow$A, A$\rightarrow$E and T$\rightarrow$E as well
as T$\rightarrow$A/E and T$\rightarrow$E), in which an oracle predicts which
of the two approaches will obtain the correct result. In this experiment, we
therefore accept a prediction as true positive if one of the two parts of the
ensemble is correct. These results are shown in Table 5 in the columns “Oracle
Ensemble”. We see a clear improvement over the isolated text-based prediction
for both oracle ensembles, while the pipeline model shows a higher
contribution in addition to the T$\rightarrow$E model (70% $\textrm{F}_{1}$ in
contrast to 66% $\textrm{F}_{1}$).
### 4.3 Discussion and Analysis
We have seen in the experiments and results that the approach of predicting
emotions based on appraisal shows a clear potential for a better performance.
Though we have not been able to reach a substantial improvement in a real-
world setting, in which appraisal dimensions are first predicted as a basis
for the emotion prediction or in the multi-task setting, we observe that text-
and appraisal-based models behave differently. Table 6 shows examples for the
prediction in the real-world setting (T$\rightarrow$A, A$\rightarrow$E). In
the top block, one example is shown for each emotion in which both appraisal
and emotion are correctly predicted. This does not only include cases in which
clear emotion indicators exist. The second block reports instances in which
the appraisal is correct, but the emotion prediction is not. Here, the first
sentence (“when someone drove into my car…”) is an example in which a flip of
certainty would have changed the emotion. Similarly, the change of attention
in “when I saw a group…” would have lead to the correct emotion prediction.
These are therefore untypical cases of appraisal assignment. The last block
shows examples where the wrong appraisal prediction leads to wrong emotion
assignment.
It is further interesting to look into those cases which are wrongly predicted
from text, but correctly predicted based on the gold appraisal annotations. We
show examples for such cases in Table 7. Several of these cases are examples
in which a word seems to indicate a particular emotion, which is actually not
relevant to infer the emotion in the first place (e.g., “animal”, “vomiting”,
“kids”, “high school”). Often, shame is wrongly predicted when the event is
about the self. This is particularly problematic if the actual word pointing
to an emotion appears to be non-typical (e.g., “crossword”, “anaesthetic”).
Gold Emotion | A$\rightarrow$E | T$\rightarrow$E | Text
---|---|---|---
Anger | Anger | Fear | because I was overlooked at work.
Anger | Anger | Disgust | when I saw someone mistreating an animal.
Anger | Anger | Fear | when someone overtook my car on a blind bend and nearly caused an accident.
Disgust | Disgust | Shame | because I ate a sausage that was horrible.
Disgust | Disgust | Fear | when I was on a ferry in a storm and lots of people were vomiting.
Disgust | Disgust | Shame | because the milk I put in my coffee had lumps in it.
Fear | Fear | Shame | because I had to have a general anaesthetic for an operation.
Fear | Fear | Sadness | when my 2 year old broke her leg, and we felt helpless to assist her.
Fear | Fear | Anger | because we were driving fast in the rain in order to get somewhere before it shut, and the driver was going over the speed limit.
Guilt | Guilt | Shame | when I took something without paying.
Guilt | Guilt | Joy | for denying to offer my kids what they demanded of me.
Guilt | Guilt | Anger | when I had not done a job for a friend that I had promised to do.
Joy | Joy | Sadness | when witnessing the joy on my children’s face on Christmas morning.
Joy | Joy | Shame | when I managed to complete a cryptic crossword.
Joy | Joy | Disgust | when I found a twenty pound note on the ground outside.
Sadness | Sadness | Fear | when it was raining this morning as I been planning to go on a camping trip.
Sadness | Sadness | Joy | I feel … when I see the Christmas decorations come down, and know they won’t be up again for another year.
Sadness | Sadness | Shame | when my friend’s eye was watering after an injection into it and I could do nothing to help.
Shame | Shame | Joy | when I failed my ninth year at high school.
Shame | Shame | Guilt | when I had too much to drink in a pub, fell over and had to go to hospital.
Shame | Shame | Anger | when my mom caught me lying.
Table 7: Examples in which the appraisal model (on gold appraisal annotation)
predicts the correct emotion and the baseline system does not.
## 5 Conclusions and Future Work
We investigated the hypothesis that informing an emotion classification model
about the cognitive appraisal regarding a situation is beneficial for the
model performance. We were able to show that emotion classification performs
better than text-based classification, under the assumption that perfect
appraisal predictions are possible and shows complementary correct
predictions. Yet, neither in a multi-task learning nor a pipeline, in which
the appraisal was predicted as a basis, could we show an improvement in
emotion classification. This provides evidence that, though our appraisal
predictor is of reasonable performance, the model suffers from error
propagation. This is still an encouraging result, suggesting that future work
should further investigate the combination of appraisal information with
emotion prediction, particularly in the light of our oracle ensemble that
indicated a clear improvement.
This first study on the topic raises a couple of research questions: Would
there be other neural architectures which are better suited for including the
appraisal information? Will more annotated data improve the prediction quality
sufficiently? Finally, it should be analyzed if giving the annotators access
to the emotion label when making the appraisal annotation could have changed
the results.
## Acknowledgements
This work was supported by Leibniz WissenschaftsCampus Tübingen “Cognitive
Interfaces” and Deutsche Forschungsgemeinschaft (project SEAT, KL 2869/1). We
thank Laura Ana Maria Oberländer for inspiration and fruitful discussions and
Valentino Sabbatino for his annotation work. Further, we thank the three
anonymous reviewers for their constructive and helpful criticism.
## References
* [Abdul-Mageed and Ungar, 2017] Muhammad Abdul-Mageed and Lyle Ungar. 2017\. EmoNet: Fine-grained emotion detection with gated recurrent neural networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 718–728, Vancouver, Canada, July. Association for Computational Linguistics.
* [Alm et al., 2005] Cecilia Ovesdotter Alm, Dan Roth, and Richard Sproat. 2005\. Emotions from text: Machine learning for text-based emotion prediction. In Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing, pages 579–586, Vancouver, British Columbia, Canada. Association for Computational Linguistics.
* [Aman and Szpakowicz, 2007] Saima Aman and Stan Szpakowicz. 2007\. Identifying expressions of emotion in text. In International Conference on Text, Speech and Dialogue, pages 196–205. Springer.
* [Averill, 1980] James R Averill. 1980\. A constructivist view of emotion. In Theories of emotion, pages 305–339. Elsevier.
* [Balahur et al., 2011] Alexandra Balahur, Jesús M. Hermida, Andrés Montoyo, and Rafael Muñoz. 2011\. Emotinet: A knowledge base for emotion detection in text built on the appraisal theories. In Rafael Muñoz, Andrés Montoyo, and Elisabeth Métais, editors, Natural Language Processing and Information Systems, pages 27–39, Berlin, Heidelberg. Springer Berlin Heidelberg.
* [Balahur et al., 2012] Alexandra Balahur, Jesus M. Hermida, and Andrew Montoyo. 2012\. Building and exploiting emotinet, a knowledge base for emotion detection based on the appraisal theory model. IEEE Transactions on Affective Computing, 3(1):88–101.
* [Barrett and Russell, 2015] Lisa Feldman Barrett and James A Russell. 2015\. The psychological construction of emotion. Guilford Publications, New York.
* [Bostan and Klinger, 2018] Laura-Ana-Maria Bostan and Roman Klinger. 2018\. An analysis of annotated corpora for emotion classification in text. In Proceedings of the 27th International Conference on Computational Linguistics, pages 2104–2119, Santa Fe, New Mexico, USA. Association for Computational Linguistics.
* [Bostan et al., 2020] Laura Ana Maria Bostan, Evgeny Kim, and Roman Klinger. 2020\. GoodNewsEveryone: A corpus of news headlines annotated with emotions, semantic roles, and reader perception. In Nicoletta Calzolari, Khalid Choukri, Thierry Declerck, Hrafn Loftsson, Bente Maegaard, Joseph Mariani, Asuncion Moreno, Jan Odijk, and Stelios Piperidis, editors, Proceedings of the 12th International Conference on Language Resources and Evaluation (LREC’20), Marseille, France. European Language Resources Association (ELRA).
* [Broekens et al., 2008] Joost Broekens, Doug DeGroot, and Walter A. Kosters. 2008\. Formal models of appraisal: Theory, specification, and computational model. Cognitive Systems Research, 9(3):173 – 197.
* [Buechel and Hahn, 2017] Sven Buechel and Udo Hahn. 2017\. EmoBank: Studying the impact of annotation perspective and representation format on dimensional emotion analysis. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pages 578–585, Valencia, Spain. Association for Computational Linguistics.
* [Buechel et al., 2016] Sven Buechel, Johannes Hellrich, and Udo Hahn. 2016\. Feelings from the Past—Adapting affective lexicons for historical emotion analysis. In Proceedings of the Workshop on Language Technology Resources and Tools for Digital Humanities (LT4DH), pages 54–61, Osaka, Japan, December. The COLING 2016 Organizing Committee.
* [Campero et al., 2017] Andres Campero, Bjarke Felbo, Joshua B. Tenenbaum, and Rebecca Saxe. 2017\. A first step in combining cognitive event features and natural language representations to predict emotions. In Conference on Cognitive Computational Neuroscience.
* [Cevher et al., 2019] Deniz Cevher, Sebastian Zepf, and Roman Klinger. 2019\. Towards multimodal emotion recognition in german speech events in cars using transfer learning. In Proceedings of the 15th Conference on Natural Language Processing (KONVENS 2019): Long Papers, pages 79–90, Erlangen, Germany. German Society for Computational Linguistics & Language Technology.
* [Clore and Ortony, 2013] Gerald. L. Clore and Andrew Ortony. 2013\. Psychological Construction in the OCC Model of Emotion. Emotion Review, 5(4):335–343.
* [Ekman, 1992] Paul Ekman. 1992\. An argument for basic emotions. Cognition & emotion, 6(3-4):169–200.
* [Ekman, 1999] Paul Ekman. 1999\. Basic emotions. In Tim Dalgleish and Mick J. Power, editors, Handbook of Cognition and Emotion. John Wiley & Sons, Sussex, UK.
* [Felbo et al., 2017] Bjarke Felbo, Alan Mislove, Anders Søgaard, Iyad Rahwan, and Sune Lehmann. 2017\. Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 1615–1625, Copenhagen, Denmark. Association for Computational Linguistics.
* [Feldman Barrett, 2006] Lisa Feldman Barrett. 2006\. Solving the emotion paradox: categorization and the experience of emotion. Personality and Social Psychology Review, 10(1):20–46.
* [Gendron et al., 2014] Maria Gendron, Debi Roberson, Jacoba Marietta van der Vyver, and Lisa Feldman Barrett. 2014\. Perceptions of emotion from facial expressions are not culturally universal: evidence from a remote culture. Emotion, 14(2):251–262.
* [Izard, 1971] Carroll E. Izard. 1971\. The face of emotion. Appleton-Century-Crofts, New York.
* [Kim et al., 2017] Evgeny Kim, Sebastian Padó, and Roman Klinger. 2017\. Investigating the relationship between literary genres and emotional plot development. In Proceedings of the Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature, pages 17–26, Vancouver, Canada. Association for Computational Linguistics.
* [Kim, 2014] Yoon Kim. 2014\. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1746–1751, Doha, Qatar. Association for Computational Linguistics.
* [Klinger et al., 2018] Roman Klinger, Orphée De Clercq, Saif Mohammad, and Alexandra Balahur. 2018\. IEST: WASSA-2018 implicit emotions shared task. In Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 31–42, Brussels, Belgium. Association for Computational Linguistics.
* [Li et al., 2017] Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. 2017\. DailyDialog: A manually labelled multi-turn dialogue dataset. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 986–995, Taipei, Taiwan. Asian Federation of Natural Language Processing.
* [Marsella and Gratch, 2009] Stacy C. Marsella and Jonathan Gratch. 2009\. EMA: A process model of appraisal dynamics. Cognitive Systems Research, 10(1):70 – 90. Modeling the Cognitive Antecedents and Consequences of Emotion.
* [Marsella et al., 2010] Stacy C. Marsella, Jonathan Gratch, and Paolo Petta. 2010\. Computational models of emotion. In Klaus R. Scherer, Tanja Bänziger, and Etienne Roesch, editors, A Blueprint for Affective Computing. Oxford University Press.
* [Mohammad and Bravo-Marquez, 2017] Saif Mohammad and Felipe Bravo-Marquez. 2017\. WASSA-2017 shared task on emotion intensity. In Proceedings of the 8th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 34–49, Copenhagen, Denmark. Association for Computational Linguistics.
* [Mohammad and Turney, 2012] Saif M. Mohammad and Peter D. Turney. 2012\. Crowdsourcing a word-emotion association lexicon. Computational Intelligence, 29(3).
* [Mohammad et al., 2017] Saif M. Mohammad, Parinaz Sobhani, and Svetlana Kiritchenko. 2017\. Stance and sentiment in tweets. ACM Trans. Internet Technol., 17(3).
* [Mohammad et al., 2018] Saif Mohammad, Felipe Bravo-Marquez, Mohammad Salameh, and Svetlana Kiritchenko. 2018\. SemEval-2018 task 1: Affect in tweets. In Proceedings of The 12th International Workshop on Semantic Evaluation, pages 1–17, New Orleans, Louisiana. Association for Computational Linguistics.
* [Mohammad, 2012] Saif Mohammad. 2012\. #emotional tweets. In *SEM 2012: The First Joint Conference on Lexical and Computational Semantics – Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012), pages 246–255, Montréal, Canada. Association for Computational Linguistics.
* [Mohammad, 2018] Saif M. Mohammad. 2018\. Obtaining reliable human ratings of valence, arousal, and dominance for 20,000 english words. In Proceedings of The Annual Conference of the Association for Computational Linguistics (ACL), Melbourne, Australia.
* [Nair and Hinton, 2010] Vinod Nair and Geoffrey E. Hinton. 2010\. Rectified linear units improve restricted Boltzmann machines. In Johannes Fürnkranz and Thorsten Joachims, editors, Proceedings of the 27th International Conference on Machine Learning (ICML-10), pages 807–814, Haifa, Israel. Omnipress.
* [Oatley and Johnson-Laird, 1987] Keith Oatley and Philip N. Johnson-Laird. 1987\. Towards a cognitive theory of emotions. Cognition and emotion, 1(1):29–50.
* [Oatley, 1993] Keith Oatley. 1993\. Social construction in emotions. In M. Lewis and J. M. Haviland, editors, Handbook of emotions, pages 341–352. Guilford Press, New York.
* [Pennebaker et al., 2001] James W Pennebaker, Martha E Francis, and Roger J Booth. 2001\. Linguistic inquiry and word count: Liwc 2001. Mahway: Lawrence Erlbaum Associates, 71:2001.
* [Pennington et al., 2014] Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014\. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1532–1543, Doha, Qatar. Association for Computational Linguistics.
* [Plutchik, 2001] Robert Plutchik. 2001\. The nature of emotions human emotions have deep evolutionary roots, a fact that may explain their complexity and provide tools for clinical practice. American Scientist, 89(4):344–350.
* [Posner et al., 2005] Jonathan Posner, James A. Russell, and Bradley S. Peterson. 2005\. The circumplex model of affect: an integrative approach to affective neuroscience, cognitive development, and psychopathology. Development and Psychopathology, 17(3):715–734.
* [Preoţiuc-Pietro et al., 2016] Daniel Preoţiuc-Pietro, H. Andrew Schwartz, Gregory Park, Johannes Eichstaedt, Margaret Kern, Lyle Ungar, and Elisabeth Shulman. 2016\. Modelling valence and arousal in Facebook posts. In Proceedings of the 7th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 9–15, San Diego, California. Association for Computational Linguistics.
* [Scherer and Wallbott, 1997] Klaus R. Scherer and Harald G. Wallbott. 1997\. The ISEAR questionnaire and codebook. Geneva Emotion Research Group. https://www.unige.ch/cisa/research/materials-and-online-research/research-material/.
* [Scherer et al., 2001] Klaus R. Scherer, Angela Schorr, and Tom Johnstone. 2001\. Appraisal processes in emotion: Theory, methods, research. Oxford University Press.
* [Scherer et al., 2010] Klaus R. Scherer, Tanja Bänziger, and Etienne Roesch. 2010\. A Blueprint for Affective Computing. Oxford University Press.
* [Scherer et al., 2019] Klaus R. Scherer, Heiner Ellgring, Anja Dieckmann, Matthias Unfried, and Marcello Mortillaro. 2019\. Dynamic facial expression of emotion and observer inference. Frontiers in Psychology, 10:508.
* [Scherer, 1982] Klaus R. Scherer. 1982\. Emotion as a process: Function, origin and regulation. Social Science Information, 48:555–70.
* [Scherer, 2005] Klaus R. Scherer. 2005\. What are emotions? And how can they be measured? Social Science Information, 44(4):695–729.
* [Schuff et al., 2017] Hendrik Schuff, Jeremy Barnes, Julian Mohme, Sebastian Padó, and Roman Klinger. 2017\. Annotation, modelling and analysis of fine-grained emotions on a stance and sentiment detection corpus. In Proceedings of the 8th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 13–23, Copenhagen, Denmark. Association for Computational Linguistics.
* [Shaikh et al., 2009] Mostafa Al Masum Shaikh, Helmut Prendinger, and Mitsuru Ishizuka. 2009\. A linguistic interpretation of the occ emotion model for affect sensing from text. Affective Information Processing, pages 45–73.
* [Smith and Ellsworth, 1985] Craig. A. Smith and Phoebe. C. Ellsworth. 1985\. Patterns of cognitive appraisal in emotion. Journal of Personality and Social Psychology, 48(4):813–838.
* [Strapparava and Mihalcea, 2007] Carlo Strapparava and Rada Mihalcea. 2007\. SemEval-2007 task 14: Affective text. In Proceedings of the Fourth International Workshop on Semantic Evaluations (SemEval-2007), pages 70–74, Prague, Czech Republic, June. Association for Computational Linguistics.
* [Strapparava and Valitutti, 2004] Carlo Strapparava and Alessandro Valitutti. 2004\. WordNet affect: an affective extension of WordNet. In Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC’04), Lisbon, Portugal. European Language Resources Association (ELRA).
* [Tooby and Cosmides, 2008] John Tooby and Leda Cosmides. 2008\. The evolutionary psychology of the emotions and their relationship to internal regulatory variables. In J.M. Haviland-Jones M. Lewis and L. F. Barrett, editors, Handbook of emotions, pages 114–137. Guilford Press, New York.
* [Tracy and Robins, 2006] Jessica L. Tracy and Richard W. Robins. 2006\. Appraisal antecedents of shame and guilt: Support for a theoretical model. Personality and social psychology bulletin, 32(10):1339–1351.
* [Troiano et al., 2019] Enrica Troiano, Sebastian Padó, and Roman Klinger. 2019\. Crowdsourcing and validating event-focused emotion corpora for German and English. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4005–4011, Florence, Italy. Association for Computational Linguistics.
## Appendix A Appraisal Prediction Performance of the Multitask Model
| T$\rightarrow$A/E
---|---
Appraisal | P | R | $\textrm{F}_{1}$
Attention | 82 | 78 | 80
Certainty | 85 | 76 | 80
Effort | 65 | 62 | 64
Pleasantness | 76 | 59 | 67
Responsibility | 71 | 69 | 70
Control | 60 | 42 | 49
Circumstance | 67 | 48 | 56
Macro avg. | 72 | 62 | 67
Micro avg. | 76 | 68 | 71
Table 8: Multi-task learning model performance on predicting appraisal
dimensions.
We focused in the discussion of the results in Section 4.2 on the goal to
improve the performance of the emotion prediction with appraisal information.
Though we evaluated the appraisal prediction model as an intermediate step in
the pipeline to emotion prediction, we did only evaluate the emotion
performance in the multitask learning setting due to our focus on emotion
categorization.
However, it would be meaningful to also experiment with a
T$\rightarrow$E,E$\rightarrow$A model when the goal would be to predict the
appraisal dimensions with or without knowledge of an emotion category. We
leave such evaluation for future work. However, our multitask model
T$\rightarrow$A/E also produces appraisal predictions in the context of
emotion predictions. We therefore provide the appraisal evaluation of the
T$\rightarrow$A/E model in Table 8. We see that the appraisal prediction does
not show any improvement over the T$\rightarrow$A model with the respective
results in Table 4.
## Appendix B Examples from the Multitask Model
We show examples from the T$\rightarrow$A/E model in Table 9, similarly to the
results from T$\rightarrow$A,A$\rightarrow$E examples in Table 6. Note that,
given the bidirectional interdependencies between emotion and appraisal taking
place in the multitask learning, in contrast to the unidirectional information
flow from appraisal to emotion in the pipeline, these results are more
challenging to interpret.
| Appraisal |
---|---|---
| Emotion (G/P) | A | Ce | Ex | Px | Rx | Co | Ci | Text
Appr+Emo corr. | Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I heard someone gossip about a friend of mine.
Disgust | 0 | 1 | 0 | 0 | 0 | 0 | 1 | because the drain smelled.
Fear | 1 | 0 | 1 | 0 | 0 | 0 | 0 | when I thought my son could be in trouble when mountain climbing.
Guilt | 0 | 1 | 0 | 0 | 1 | 1 | 0 | because I drank way too much, knowing that I had things to do the next day.
Joy | 1 | 1 | 0 | 1 | 0 | 0 | 0 | because my son gave me a big hug when I got home from work.
Sadness | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I read about orangutans and palm oil
Shame | 0 | 1 | 0 | 0 | 1 | 1 | 0 | because I walked past a beggar and didn’t bother to stop and give her a pound.
Emo incorr. | Anger/Shame | 1 | 0 | 0 | 0 | 0 | 0 | 0 | because the parents dropping their kids off to school were driving and parking dangerously.
Anger/Disgust | 1 | 0 | 1 | 0 | 0 | 0 | 0 | someone stabbed a man on a train.
Disgust/Sadness | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I read that hunters had killed one of the world famous lions.
Disgust/Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I watched videos of how some people treat animals.
Guilt/Shame | 0 | 1 | 0 | 0 | 1 | 1 | 0 | because I cheated on a previous girlfriend.
Sadness/Anger | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when listening to the national news.
Shame/Guilt | 0 | 1 | 0 | 0 | 1 | 1 | 0 | when my daughter was rude to my wife.
Ap+Emo incorr. | Anger/Disgust | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when a drunk man insulted my intelligence.
Disgust/Fear | 1 | 1 | 0 | 0 | 1 | 0 | 0 | towards my friend because he stole my wallet.
Fear/Shame | 0 | 1 | 0 | 0 | 1 | 0 | 0 | because a fight broke out in a bar where I was drinking
Guilt/Disgust | 0 | 1 | 0 | 0 | 0 | 0 | 0 | because I accidentally killed my pet fish by decorating his tank with untreated sand.
Joy/Sadness | 1 | 1 | 0 | 0 | 0 | 0 | 0 | when I rescued a dog at the local shelter and he became my best friend.
Sadness/Joy | 1 | 1 | 0 | 0 | 0 | 0 | 1 | when I was on holiday and it was the last day and had to come home.
Shame/Guilt | 0 | 1 | 0 | 0 | 1 | 0 | 1 | when I ate my whole Easter egg in one go.
Table 9: Examples for the prediction of the multi-task setting
(T$\rightarrow$A/E) . A: Attention, Ce: Certainty, E: Effort, P: Pleasantness,
R: Responsibility, Co: Control, Ci: Circumstance. First emotion mention is
gold, second is prediction. Appraisal shown is prediction with errors shown in
bold.
|
# MoRS: An Approximate Fault Modelling Framework for Reduced-Voltage SRAMs
İsmail Emir Yüksel3, Behzad Salami2, Oğuz Ergin3, Osman S. Ünsal2, Adrián
Cristal Kestelman24 3 TOBB University of Economics and Technology (TOBB ETÜ)
2 Barcelona Supercomputing Center (BSC) 4 Universitat Politècnica de
Catalunya (UPC)
###### Abstract
On-chip memory (usually based on Static RAMs-SRAMs) are crucial components for
various computing devices including heterogeneous devices, e.g, GPUs, FPGAs,
ASICs to achieve high performance. Modern workloads such as Deep Neural
Networks (DNNs) running on these heterogeneous fabrics are highly dependent on
the on-chip memory architecture for efficient acceleration. Hence, improving
the energy-efficiency of such memories directly leads to an efficient system.
One of the common methods to save energy is undervolting i.e., supply voltage
underscaling below the nominal level. Such systems can be safely undervolted
without incurring faults down to a certain voltage limit. This safe range is
also called voltage guardband. However, reducing voltage below the guardband
level without decreasing frequency causes timing-based faults.
In this paper, we propose MoRS, a framework that generates the first
approximate undervolting fault model using real faults extracted from
experimental undervolting studies on SRAMs to build the model. We inject the
faults generated by MoRS into the on-chip memory of the DNN accelerator to
evaluate the resilience of the system under the test. MoRS has the advantage
of simplicity without any need for high-time overhead experiments while being
accurate enough in comparison to a fully randomly-generated fault injection
approach. We evaluate our experiment in popular DNN workloads by mapping
weights to SRAMs and measure the accuracy difference between the output of the
MoRS and the real data. Our results show that the maximum difference between
real fault data and the output fault model of MoRS is 6.21%, whereas the
maximum difference between real data and random fault injection model is
23.2%. In terms of average proximity to the real data, the output of MoRS
outperforms the random fault injection approach by 3.21x.
###### Index Terms:
Modeling, Fault-injection, Neural Networks, Undervolting, SRAM.
## I Introduction
SRAMs are traditionally the building block of different components in
different computing systems such as branch predictor (in CPUs), register files
(in GPUs), on-chip buffer memory (in hardware accelerators like FPGAs and
ASICs), thanks to the low-latency access time of such memories. However, the
power consumption of SRAMs significantly contributes to the total system
power. For instance, prior works on GPUs[1] show that register file in GPUs
consumes 15-20% of the total power. Another work on modern out-of-order
cores[2] estimates that the SRAM-dominated front-end of the microprocessor
consumes up to 33% of the total power of the CPU. Other work on DNN
accelerators on FPGAs [3] shows that on-chip SRAMs consume 27% of the total
power.
Since the total power consumption of any underlying hardware is directly
related to its supply voltage, voltage underscaling is an effective solution
to save power [4, 5]. This technique is widely used in various devices such as
CPUs[6, 7, 8, 9, 10], GPUs[11] FPGAs[3, 12, 13, 14, 15], and ASICs[16, 17, 18]
as well as DRAMs[19, 20, 21], HBMs[22], SRAMs[23, 24], and Flash Disks[25, 26,
27, 28]. However, while performing voltage underscaling, reliability issues
can arise due to the increased circuit delay at reduced voltage levels. In
most commercial devices, there is a timing-fault free guardband between
nominal voltage and minimum safe voltage, i.e., $V_{min}$. Below this voltage
guardband, faults occur as a consequence of circuit delays.
Although further aggressive voltage underscaling below $V_{min}$ can achieve
more reduction in power consumption, it compromises system reliability through
undervolting faults. This particular region between $V_{min}$ and the lowest
operational voltage level, i.e., $V_{crash}$, still operational but with
faults is called the critical area[3]. Several prior fault mitigation
techniques were proposed[3, 16, 17, 29, 30]. However, these techniques need
either high-effort engineering[3, 30] or totally random fault injection
campaigns[16, 17]. These approaches are either impractical or not accurate.
Our solution offers the advantages of both random fault injection and
empirical experiments.
In this study, we propose MoRS, a framework that generates the first
approximate voltage underscaled model for SRAMs. MoRS consists of three steps:
1) Experiment, 2) Behavior Extraction, and 3) Model Generation. In the
Experiment step, MoRS uses publicly available undervolted fault map data[31]
of SRAMs based on our prior work[3]. In the Behavior Extraction step, we
extract the characteristic fault behavior features of undervolted SRAM blocks.
We establish and confirm that undervolting based faults do not occur randomly.
These faults are correlated with each other in space. We examine the faults of
row-based and column-based approaches and see the distance between
consecutively faulty bitcells, the number of each bit-fault in both rows and
columns, and the total number of faulty rows and columns per SRAM block are
not uniformly distributed. These fine-grained features show characteristic
behaviors. In this step, we extract characteristic features and categorize
them into two profiles: coarse-grained and fine-grained profiling. The reason
behind these categorizing is that random fault injection studies use only
coarse-grained features, the number of bit faults, and the number of faulty
SRAM blocks. The output of MoRS that we term Mixed Model uses both fine-
grained and coarse-grained features and applies probabilistic custom modeling
algorithm to achieve an approximate model as the last step.
In recent years, the power consumption of SRAMs on DNNs is increasing
drastically[16, 3, 17]. We evaluate MoRS for DNN accelerators where on-chip
SRAMs play an important role. MoRS generates fault models that are not limited
to a certain domain and can be potentially used in many domains such as branch
prediction units, register files, caches, and any other memories based on
SRAM, unlike the other prior studies. To evaluate MoRS, we generate a baseline
model by applying a uniform random distribution function to coarse-grained
features. This baseline, which we term as Random Model, is a standard fault
injection scheme and used in prior studies[32, 17, 16, 33]. Besides, we
process empirical data to see the difference in accuracy between each
artificial model (Random Model and Mixed Model) and real data.
In our evaluation methodology, we map weights to SRAM blocks. When we map
weights to SRAM blocks, there can be different mapping options such as MSB
mapping, LSB mapping, the first half of bits MSB the other half of bits LSB
mapping, and the first half of bits LSB the other half of bits MSB mapping.
Another method to save energy on DNNs is quantization. Our quantization method
is reducing the precision of weights from 32-bit to 16-bit, 8-bit, 4-bit, and
1-bit, respectively. While performing undervolting, some unwanted bits can be
flipped, and then the corresponding value becomes infinity or NaN. To avoid
these values masking techniques are used[34]. We mask infinity or NaN value to
1 or 0. We examine the behavior of different weight mappings, quantizations,
and masking options. We evaluate our experiments on trained LeNeT-5[35] and
cuda-convnet[36] DNNs in the classification stage. We examine the
classification accuracy for each voltage level. Our experiments show that
generated artificial model has similar behavior with the real data on
different DNN benchmarks with an accuracy of 96.4%. We also see that Random
Model is not close enough with real data with the difference in accuracy up to
23%. We find that on average our Mixed Model has 3.6% difference with real
data and on average 3x up to 7x closer than the baseline random model.
Our contributions are as follows:
* •
We propose a framework, MoRS that generates an artificial model that can
realistically emulate real undervolting fault data with a difference in the
accuracy of 3.6% on average. To the best of our knowledge, this study provides
the first reasonably accurate model for SRAM blocks under low-voltage
conditions.
* •
We evaluate our models and real data on state-of-the-art DNNs to see how
different weight mappings, quantization, and value maskings affect the
accuracy of inference. We find the similar observation with prior works[37,
38] that if we continue to reduce the precision of weights, DNNs become more
resilient to undervolting. At the lowest reduced voltage level 8-bit LeNeT
accuracy is 14% while 4-bit LeNeT accuracy is 60%. Even in this unlikely
situation, our Mixed Model shows similar behavior to real data.
The remainder of this paper is structured as follows. In Section II, we
introduce the most important concepts used in this paper. In Section III, we
propose our approximate fault modeling framework, MoRS. Section IV describes
the methodology that we perform our experiments. Section V detail the
experimental results. Related works are introduced in Section VI. Finally,
Section VII concludes this paper.
## II Background
### II-A Undervolting SRAM based on-chip Memories
CMOS is the dominant circuit technology for current computing devices. The
power consumption of CMOS-based SRAMs is the sum of two parts: the first is
dynamic power, also called active power, and the other one is leakage power,
also called static power. Earlier studies[39, 40, 41] show that the power
consumption of SRAMs is dominated by dynamic power.
Dynamic power i.e., $P_{dyn}$, is dissipated when there is switching activity
at some nodes in transistors. Leakage power i.e., $P_{leak}$, is dissipated by
the leakage currents flowing even when the device is not active.
Mathematically, these power equations are given by,
$P_{dyn}\propto f\times V_{dd}^{2}$ (1) $P_{leak}=k_{design}\times n\times
I_{leak}\times V_{dd}$ (2)
Here $V_{dd}$ shows the supply voltage, $f$ shows the operating frequency in
Equation 1. Further, $n$ denotes the number of transistors, $k_{design}$ is
the design-dependent parameter, and $I_{leak}$ shows the leakage current a
technology-dependent parameter in Equation 2. From Equation 1, dynamic power
consumption can be reduced by adjusting the supply voltage and operating
frequency. Likewise, from Equation 2, leakage power consumption can be reduced
by underscaling the supply voltage and reducing the total number of
transistors. Total power consumption can be reduced by underscaling the supply
voltage.
Voltage underscaling is a widely used technique for energy-efficient
applications. From Equation 1, dynamic power consumption is reduced
quadratically by underscaling the supply voltage. This technique can achieve
nominal voltage level i.e., $V_{nom}$ performance due to the operating
frequency not being changed. However, further aggressive undervolting below
the minimum safe voltage i.e., $V_{min}$ may cause reliability issues as the
result of timing faults. Between $V_{nom}$ and $V_{min}$ is called voltage
guardband. This guardband is to provide an assurance of correct functionality
even in the worst environmental case. Prior works achieve by applying
undervolting that power consumption reduces; 39% on FPGA on-chip memories [3],
20% in GPUs [42], and 16% in DRAMs [19] without any timing-related errors.
Also, recent studies show that undervolting internal components of FPGAs [12]
lead to around 3x power-efficiency and underscaling supply voltage of HBMs
[22] achieve a total of 2.3x power savings. In Minerva[16], lowering SRAM
voltages achieves a total of 2.7x power savings.
### II-B Fault Injection
The fault injection mechanism is a way to examine the behavior of systems
under different circumstances. Prior works on fault injection study for
reduced-voltage SRAMs are applied on branch prediction units [43, 32], caches
[44, 45], and FPGA on-chip memories [33]. There are several approaches for the
fault injection with having a trade-off on the engineering effort and
accuracy. i) The first one is applying random faults to random locations
without any information used from empirical experiments. ii) Another one is
directly using the empirical data as a fault map. iii) The last one is the
approximate modeling that is based on real data from empirical experiments and
close enough to empirical data. Table I gives a summary of comparison of these
three fault injection techniques in terms of effort and how close they are to
real data. The accuracy percentage is from the results of this study.
TABLE I: Comparison of fault injection techniques in terms of engineering
effort and accuracy of representing real data
Fault Injection Method | Engineering Effort | Accuracy (Min)
---|---|---
Random Fault Injection | Low | 77%
MoRS | Low | 94%
Empirical Data | High | 100%
### II-C Deep Neural Networks
Deep Neural Networks (DNNs) are widely used as an effective solution for
object recognition, classification and segmentation, and many other areas.
MNIST[46] and CIFAR-10 [47] datasets are widely used by the ML community to
showcase the latest technology advancements in DNNs. DNNs perform in two
phases: training and prediction (inference). The first process is training
that weights of DNNs are fitted to data. The training phase is mainly
performed in high-performance computing platforms like CPUs or GPUs. The
second phase is the inference phase, DNN models, trained in the first phase,
are used to extract unknown output from input data. Usually, the training
phase performs once whereas the prediction phase performs repeatedly. The
inference phase consists of mainly three elements: input, weights, and output.
Figure 1: A high level abstraction of one layer of DNNs.
Figure 1 shows a high-level abstraction of one inference layer. Each inference
layer of the network consists of weights, an input matrix, and an output
matrix. Weights of these layers are stored in on-chip memories. Due to their
large size and computational intensity, DNNs consume power; while they are
fault-tolerant to some extend. Some studies improve the power-efficiency and
reduce the total power consumption of DNNs by applying voltage underscaling
[3, 12], architectural techniques[48, 49, 50, 51, 52, 53, 54, 55], and
hardware-level techniques[56, 57, 58].
Quantization is an architectural optimization technique that reduces the
precision of data types. This technique can improve performance and energy-
efficiency. Recent studies[59, 16, 60, 61, 62] show that quantization to some
extent does not significantly affect the accuracy of DNN models. In our study,
we reduce weights precisions to 16-bit half-precision floating point,
8-bit($Q_{4.4}$), 4-bit($Q_{2.2}$) in fixed-point format, and 1-bit binary
values.
## III MoRS Framework
We propose MoRS, a framework to generate approximate artificial reduced-
voltage SRAM fault models. This mechanism is the first framework based on real
fault maps. MoRS stands between fully hardware and fully software fault
injection techniques. This framework generates an approximate model that is
close enough to real data compared to the fully software fault injection
mechanism. Also, MoRS does not require high-effort engineering in comparison
to a fully hardware approach that uses real data from empirical experiments.
As shown in Figure 2 MoRS consists of three steps: ❶ Experiment, ❷ Behavior
Extraction, and ❸ Model Generation. We first explain the first step existing
experiment that provides real fault maps from real SRAM blocks in Section 3.1.
In Section 3.2 we extract the behavior of fault maps as fine-grained and
coarse-grained profiles by using the output of the first step. In fine-grained
profiling extract the row and column behaviors of SRAM blocks in terms of
physical distance, the number of each bit-faults in rows and columns, and the
number of faulty rows and columns per block. Coarse-grained profiling is a
shallow approach generally used in prior fault injection studies. In coarse-
grained profiling, we extract only the total number of bit-faults and faulty
SRAM blocks. In Section 3.3 we generate an artificial model with outputs of
the second steps.
At the endpoint, we provide Mixed Model which is an approximate model to real
data. It is generated by applying probabilistic modeling on both fine-grained
and coarse-grained fault profiles. Figure 2 provides an overview of the three
steps of the MoRS.
Figure 2: Overview of the MoRS
### III-A Experiment
Voltage underscaling, i.e., undervolting is a widely used technique to save
energy. We study undervolting for modern SRAM blocks. Each SRAM block is a
matrix of bitcells formed of rows and columns. Also, each SRAM block is 16
Kbits with 1024 rows and 16 columns.
Board Name | VC707 | ZC702 | KC705
---|---|---|---
Technology Node | 28nm | 28nm | 28nm
Nominal Voltage | 1V | 1V | 1V
Operating Temperature | $50^{\circ}C$ | $50^{\circ}C$ | $50^{\circ}C$
Minimum Voltage Level | 0.54V | 0.53V | 0.54V
Number of SRAM Block | 2060 | 280 | 890
TABLE II: The summary of deployed FPGA boards
We perform this empirical study on SRAM blocks available in the off-the-shelf
Xilinx FPGAs, i.e., VC707, ZC702, and two identical samples of KC705 (referred
as KC705-A and KC705-B). Table II shows the detail of deployed FPGAs. This
experiment consists of two parts. The first part is the monitor SRAM blocks to
see undervolting faults. The second part is adjusting the supply voltage of
SRAM blocks by using a power management unit of FPGAs. FPGAs have a power
management bus (PMBUS) a voltage regulator that monitors and adjusts voltage
rails on FPGAs. This bus uses PMBUS standard and has an I2C protocol to
execute commands. The supply voltage of SRAM blocks of FPGAs is named
$V_{CCBRAM}$. By reducing $V_{CCBRAM}$ through PMBUS, only the power
consumption of SRAM blocks reduces. Since this situation does not affect any
other logic parts (DSPs, LUTs, etc..) the effect of undervolting on SRAM
blocks is seen clearly. The setup of this methodology is shown in Figure 3.
Figure 3: The methodology of empirical experiment
The method of this experiment follows an algorithm that as a first step writes
data to SRAMs after in the second step analyzes the faults in terms of rate
and location, in the third step reduces the supply voltage by 10$mV$, finally
repeat these steps until FPGA crashes.
When undervolting is applied below the voltage guardband i.e., $V_{min}$, the
fault rate exponentially increases. Voltage can be reduced until the voltage
that FPGA stops operating i.e., $V_{crash}$. Between $V_{min}$ and $V_{crash}$
faults occur due to timing failures while the power consumption significantly
reduces. SRAM-based real fault maps significantly vary even for different
FPGAs of the same platform as the result of the process variation. Besides,
the study shows that the pattern of faults is mostly permanent. Also, we
observe that the fault rate and location for different run are mostly the
same. Most importantly, undervolting faults are not uniformly distributed over
different SRAMs.
We use the publicly available data[31] of the prior work [3] as the output of
the first stage (Experimental Data) to generate artificial fault maps. We use
2000 SRAM blocks for extracting behavior and 950 SRAM blocks for testing our
framework, 2060 SRAM blocks from VC707 and 890 SRAM blocks from KC705. We make
the methodology of MoRS reliable by using different data set to create and
also to test our artificial model.
At $V_{crash}$ the fault rate, up to 0.06% and 0.005% per 1 Mbits for VC707
and KC705-B, respectively. It should be noted that faults appear in between
$V_{min}$ = 0.6V, $V_{crash}$ = 0.54V for VC707 and $V_{min}$ = 0.59V,
$V_{crash}$ = 0.53V for KC705-B. As the prior work[3] mentioned VC707 has the
most bit-faults among the other three boards. At the lowest voltage level,
VC707 has 10.2% faulty SRAMs with 23706 bit-faults. We refer [3] to more
detailed information about this prior work.
TABLE III: Features and which model they are used in
Profiling Type | Features | Models
---|---|---
Name | Percentage | Mixed-Model | Random-Model
Coarse-grained | Total number of bit-faults | $P_{F}$ | ✓ | ✓
Total number of faulty SRAM blocks | $P_{S}$ | ✓ | ✓
Fine-grained | Total number of faulty rows per SRAM block | $P_{SR_{0..1024}}$ | ✓ | ✗
Total number of faulty columns per SRAM block | $P_{SC_{0..16}}$ | ✓ | ✗
Number of each bit-fault in rows | $P_{FR_{0..15}}$ | ✓ | ✗
Number of each bit-fault in columns | $P_{FC_{0..1023}}$ | ✓ | ✗
Distance between consecutively faulty bitcells in rows | $P_{FDR_{1..15}}$ | ✓ | ✗
Distance between consecutively faulty bitcells in columns | $P_{FDC_{1..1023}}$ | ✓ | ✗
### III-B Behavior Extraction
As we mentioned in Section 3.1 voltage underscaling faults have patterns and
these fault patterns i.e., fault maps, are mostly the same. There are features
that affect the probability of bit-faults. In this step, we profile the
behavior of undervolting-related bit-faults to extract such important
features. We perform the profiling in two steps: coarse-grained profiling and
fine-grained profiling. A summary of all features, profiling types, and which
model they are used in can be found in Table III.
Coarse-grained profiling consists of two features. The first one is the
percentage of bit-faults in all SRAMs bitcell i.e., $P_{F}$. The second one is
the percentage of faulty SRAMs in all SRAM blocks i.e., $P_{S}$. For VC707, at
$V_{crash}$, total bit-faults are 0.07% and the percentage of faulty SRAMs is
10.2%.
Fine-grained profiling comprises two parts: row-based, column-based. Both row-
based and column-based have three features. The first one is the percentage of
faulty rows i.e., $P_{SR_{0..1024}}$ and faulty columns $P_{SC_{0..16}}$ in
faulty SRAM blocks. The second one is the percentage of each bit-faults in
rows i.e., $P_{FR_{0..16}}$, or columns i.e., $P_{FC_{0..1024}}$. The last one
is the percentage of physical distance between consecutively faulty bitcells
i.e., bitcell-distances, in the same row i.e., $P_{FDR_{1..15}}$, or column
i.e., $P_{FDC_{1..1023}}$. We discover that these two features for row and
column are not randomly or uniformly distributed.
At $V_{crash}$,
* •
The percentage of each bit-faults
* –
In terms of row concentrates at 2-bit faults ($P_{FR_{2}}$) and no faults
($P_{FR_{0}}$).
* –
In terms of column it concentrates no faults to 10-bit faults
($P_{FC_{0..10}}$).
* •
The percentage of bitcell-distances between consecutively faulty bitcells
* –
In terms of row concentrates at 8-bit distance ($P_{FDR_{8}}$).
* –
For rows that there is no such bitcell-distance is more than 8-bitcells
($P_{FDR_{9..15}}$).
* –
For the columns, it is concentrated in even numbers in decreasing
order$P_{FDC_{0..2..1024}}$, 2-bitcells distance ($P_{FDC_{2}}$) has the
highest percentage while 1022-bitcells distance ($P_{FDC_{1022}}$) has the
lowest in even numbers.
* –
Also, for columns, the bitcell-distances in odd numbers
($P_{FDC_{1..3..1023}}$) are stuck between 0 and 4.
Figure 4: Physical features on a faulty SRAM block as an example
As an illustration, we show the fault behavior of an $8*8$ SRAM block in
Figure 4. In this example, the column-based distance between cell A and cell B
is 6-bitcells. For cell C and cell D, the row-based distance is 5-bitcells.
When we examine this example in Figure 4 in terms of bit-fault for both row-
based and column-based, a column that contains cell C has 2-bit faults. Also,
cell A’s row has 4-bit faults. One of the coarse-grained profile features is
the percentage of total bit-faults ($P_{F}$). It can be calculated by dividing
the total number of faulty cells by all cells. Hence, for Figure 4’s SRAM
block $P_{F}$ is 28.125%.
As we extract the behavior of this example in Figure 4, we performed this
process for experimental data. When the profiling step is done, we start
generating models by using these coarse-grained and fine-grained profiles with
probabilistic modeling and uniform random distribution.
Algorithm 1 Generate Mixed Model
0: $n\leftarrow\\#ofSRAMblocks$
$bitfaults\leftarrow n\times 1024\times 16\times P_{F}$
$faultyS\leftarrow n\times P_{S}$
while $faultyS>0\ and\ bitfaults>0$ do
$block\leftarrow random$
$faultyBlock\leftarrow SRAMblocks[block]$
$faultyrows\leftarrow selectfrom(F_{SR_{0..16}})$
while $faultyrows>0$ do
$row\leftarrow random$
$column\leftarrow random$
$bitfaultsinrow\leftarrow selectfrom(F_{FR_{0..16}})$
while $bitfaultsinrow>0$ do
$distance\leftarrow selectfrom(F_{FDR_{1..15}})$
$column\leftarrow column+distance$
$faultyBlock[row][column]\leftarrow fault$
$bitfaultsinrow\leftarrow bitfaultsinrow-1$
end while
$faultyrows\leftarrow faultyrows-1$
end while
$ArtificialCF\leftarrow ColumnFeatures(faultyBlock)$
$RealCF\leftarrow ColumnFeatures(RealData)$
if $Similarity(ArtificialCF,RealCF)>80\%$ then
$SRAMblocks[block]\leftarrow faultyBlock$
$faultyS\leftarrow faultyS-1$
$bitfaults\leftarrow bitfaults-\\#offaults(faultyBlock)$
end if
end while
### III-C Model Generation
Probabilistic modeling [34] and uniform random distribution [34, 17, 63, 43]
are widely used in many modeling studies to generate fault maps and to inject
faults. For an approximate model i.e., Mixed Model, we use both fine-grained
and coarse-grained features with custom probabilistic modeling function. In
addition to these, we need the number of SRAM blocks that will generate.
To generate Mixed Model we follow the method shown in Algorithm 1. The input
of Algorithm 1 is the number of SRAM blocks and the output is faulty SRAM
data, also called fault map. First, in Algorithm 1, faulty SRAM blocks are
determined. After that, by using $P_{F}$ value, the number of faulty cells are
calculated. First, we randomly select faulty SRAMs in all SRAM blocks. Then,
we inject faults corresponding cells in faulty SRAMs. This injection algorithm
uses fine-grained features. This process is performed in two stages: row-based
fault injection and column-based control mechanism. The row-based fault
injection uses row-based features. We determine how many rows will be faulty
by using the probability of the number of faulty rows per SRAM block i.e.,
$F_{SR_{0..1024}}$ derivating from $P_{SR_{0..1024}}$. Then according to the
probability of the number of each bit-faults in rows i.e., $F_{FR_{0..16}}$
derivating from $P_{FR_{0..16}}$ we inject bit-faults to the corresponding
row. To inject more than one fault in a row we use the probability of physical
distance between consecutively faulty bitcells in rows $F_{FDR_{1..15}}$
derivating from $P_{FDR_{1..15}}$. In the second step of this algorithm,
first, we extract column-based features of each artificial fault map.
Figure 5: The running time overhead and accuracy difference between real data
and the generated artificial model for different similarity threshold levels
After extraction, we compare these features with fine-grained column-based
features extracted from the second step. If an artificial faulty SRAM has a
lower than 80% of similarity with experimental real data, we perform these
steps again until the similarity is 80% or higher. We select the 80%
similarity threshold as a good trade-off between the run-time (to generate the
fault models) and accuracy (of the generated fault models). As shown in Figure
5, if we increase the threshold level, the run time of MoRS increases
drastically. However, increasing the threshold is not achieve significant
accuracy compared to the optimal threshold. Below this threshold, the
approximate model converges to the Random Model that do not have an acceptable
accuracy.
## IV Experimental Methodology
MoRS is a general framework that generates approximate fault maps for
undervolted SRAM blocks. In this study, we test MoRS on state-of-the-art Deep
Neural Networks. Our experiments are based on injecting faults into weights of
trained DNNs. To evaluate how precise MoRS is we use Caffe[64]. Also, we
perform different quantizations (precisions), bit-mappings, and value masking
to diversify our experiments. The summary of these options is in Table IV.
TABLE IV: Different options for evaluation
Options | Name
---|---
Precision | 32-bit single-precision floating point
16-bit half-precision floating point
8-bit fixed point (Q4.4)
4-bit fixed point (Q2.2)
| 1-bit (Binary)
Bit-Mapping | MSB
LSB
First half MSB and other half LSB
First half LSB and other half MSB
Value Masking | Infinity or NaN to 1
Infinity or NaN to 0
The experiment is performed for each voltage level between $V_{min}$ and
$V_{crash}$, precision, mapping, and masking option. To compare artificial
models with real data we also process this methodology for real data.
Artificial Models are Random Model and Mixed Model. Random Model is a naive
random baseline used in prior works to inject faults. Mixed Model is an
approximate model, the output of the MoRS. Real Data is the experimental
data[31] extracted from the prior empirical study[3]. To evaluate real data,
we select the required amount of SRAM blocks randomly from real data. Instead
of Artificial Models, we process real data to evaluate in Figure 6.
Figure 6: Overall Methodology
The methodology consists of four parts as shown in Figure 6: ❶ MoRS &
Experiment, ❷ Bit-Precision & Bit-Mapping Unit, 3 ❸ Fault Injection Unit
(Weights to SRAM Mapping) and ❹ DNN Inference. In the first step, we acquire
artificial models from MoRS that are explained in Section 3. In the second
step, the Fault Injection Unit changes healthy weights into updated weights by
performing mapping and precision options. In the third step, with the outputs
of the first and second steps, we generate faulty weights. In the last step,
we obtain the final accuracy percentage.
1\. MoRS & Experiment. In this step, we choose which SRAM data is sent to
Fault Injection Unit. To evaluate MoRS, we generate a baseline model, i.e.,
Random Model. Random Model is generated by uniform random distribution with
coarse-grained features. We process every step in MoRS described in Section
II. However, we apply a directly random distribution function to only coarse-
grained features instead of applying custom probabilistic modeling and
algorithm to coarse-grained and fine-grained features.
In Random Model, first, with a given number of SRAM blocks we determine which
and how many blocks are faulty by using $P_{S}$ value. Then, using $P_{F}$
value we calculate how many cells are going to be faulty. When it is
calculated, we randomly inject faults in cells of randomly selected faulty
SRAM blocks. Because of the uniform random distribution, every cell in faulty
blocks has the same probability. The differences between Mixed Model and
Random Model are summarized in Table III.
In Figure 6, we call Mixed Model and Random Model Artificial Models. To
understand how accurate our approximate model we evaluate empirical data also
called Real Data. Since evaluated networks utilize 850 SRAM blocks at maximum,
we randomly choose the required amount of SRAM blocks.
2\. Bit-Precision & Bit-Mapping Unit. Quantization and undervolting are both
effective techniques to improve the energy efficiency of DNNs. However, they
may lead to accuracy loss with aggressive exploitation. MoRS enables us to
explore their correlation to find an optimal operating point.
In this step, we change weights according to precision and mapping options.
Caffe’s weights are 32-bit single-precision floating points. Since each row of
SRAM block has 16-bit, we store those weights in two rows when precision is
not reduced.
We use four fixed point precisions: 16-bit half-precision floating point,
8-bit (Q.4.4), 4-bit (Q2.2), and binary. To enable fixed-precision options we
use a prior study[65] an adapted version of the original Caffe with limited
numerical precision of weights. For 16-bit half-precision floating point, we
use NVCaffe [66], NVIDIA-maintained Caffe that supports 16-bit half-precision
floating point train and inference. When precision is reduced to $X-bit$, we
store $16/X$ weights in one row consecutively. Therefore, by reducing
precision, the usage of SRAM blocks and power consumption decrease with a cost
of accuracy loss.
In addition to precision options, we change the mapping of weights to SRAM
blocks. There are four mapping options: MSB, LSB, the first half of bit MSB,
and another half of bits LSB, the first half of bits LSB, and another half of
bit MSB. MSB means the most significant bit of weights maps to the first cell
of a row whereas LSB means the least significant bit of weights maps to the
first cell of a row.
3\. Fault Injection Unit (Weights to SRAM Mapping). After precision and
mapping options we obtain updated weights. In this step, we use artificial
models to inject faults in updated weights. Each cell of artificial models
contains either faulty or healthy information. If a bit of weight is mapped in
the faulty cell, we flip its value. Else, the value is not changed. When this
bit-flip operation performs, sometimes the value of weights could be infinity
or NaN. To prevent this situation we mask these values to either one or zero.
The masking operation is only performed for 32-bit single-precision and 16-bit
half-precision floating point. Because fixed point does not have any mantissa
or exponent parts to converge NaN or infinite value. In our study, the largest
fixed point representation is $Q_{4.4}$ and its maximum value is 15.
4\. DNN Inference. After injecting faults to the weights, we use Caffe
Framework to measure the accuracy of neural networks. Our most accurate
baseline is Real Data. We train our model based on part of real data and we
test it using another of that data. To diversify we have four different bit-
mapping, three different precision, and two different masking options for each
voltage level.
(a) $Infinity-NaN\ value\leftarrow 0$ $Mapping\leftarrow MSB$
(b) $Infinity-NaN\ value\leftarrow 1$ $Mapping\leftarrow MSB$
(c) $Infinity-NaN\ value\leftarrow 0$ $Mapping\leftarrow LSB$
(d) $Infinity-NaN\ value\leftarrow 1$ $Mapping\leftarrow LSB$
(e) $Infinity-NaN\ value\leftarrow 0$ $Mapping\leftarrow MSB\mid LSB$
(f) $Infinity-NaN\ value\leftarrow 1$ $Mapping\leftarrow MSB\mid LSB$
(g) $Infinity-NaN\ value\leftarrow 0$ $Mapping\leftarrow LSB\mid MSB$
(h) $Infinity-NaN\ value\leftarrow 1$ $Mapping\leftarrow LSB\mid MSB$
Figure 7: Voltage and resilience behavior of artificial models and
experimental (real) data on LeNeT-5[35] network for each bit-mapping and value
masking option when precision is not reduced (32-bit single-precision floating
point.)
## V Experimental Results
As we mentioned in the previous section we use Caffe[64], a deep learning
framework, and test the output of MoRS and random fault injection model on two
different neural network models: LeNeT-5[35] with MNIST dataset[46] and cuda-
convnet[36] with CIFAR-10 dataset[47]. We perform different bit-mappings and
value maskings for each neural network architecture. In addition to these
tests, we also perform reduced precision tests on LeNeT-5. Details of each
evaluated benchmark are summarized in Table V.
TABLE V: Details of evaluated neural networks
NN Model | LeNeT-5[35] | cuda-convnet[36]
---|---|---
Dataset Name | MNIST[46] | CIFAR-10[47]
# of Weights | 430500 | 89440
# of SRAM Blocks Utilized | 850 | 180
Inference Accuracy (%) | 99.05% | 79.59%
### V-A Overall Resilience
Figure 7 shows the accuracy of LeNeT-5[35] on the MNIST dataset[46] with
different bit-mapping and value masking options. We observe that in all
options Mixed Model is more precise than Random Model. Also, if the
application becomes less resilient, Mixed Model is closer to the real data and
the gap between the baseline and Mixed Model is increasing in terms of how
close they are to the real data.
We observe that that if we mask infinity and NaN value to 0, the LeNeT-5
network is more resilient than masking to 1. We think that the cause of this
situation is in the MNIST dataset hand-written digits are represented by one
and the rest of the background is represented by zero. Therefore, ones more
impact than zeros when it comes to classification.
We see that MSB and $MSB\mid LSB$ mapping cause more faults and are less
resilient than LSB and $LSB\mid MSB$ mapping. We find that the cause of this
is undervolting-based faults generally occur in first cells. Since MSB means
the most significant bit, when bit-flips happen it affects the corresponding
value more than others. Figure 7b and Figure 7f show more characteristic
behavior and are less resilient than others since both have MSB mappings and
masking to 1 option.
In Figure,7 we perform 300 iterations for each option (value masking, bit
mapping). Then we average these options for LeNeT-5 and cuda-convnet[36]
network architectures. In addition to these, we perform these experiments at
different precision levels for LeNeT-5 and for each precision level, we
average all options.
Figure 8 shows the accuracy of cuda-convnet on the CIFAR-10 dataset[47]
without reducing precision, i.e, the precision of weights is 32-bit floating
points. As we evaluate the LeNeT-5[35] network in Figure 9a, we average every
bit-mapping and value masking option for cuda-convnet [36] network.
We find that the cause of LeNeT-5 works better is cuda-convnet on CIFAR-10 is
rather smaller in terms of the number of weights and SRAM block utilization.
However, even in this network, Mixed Model is 1.47x more accurate than Random
Model in terms of accuracy difference with Real Data on average. While the
supply voltage of SRAM blocks is reducing, we observe more characteristic and
distinguishing behavior from real data. At $V_{crash}$ level, the accuracy
difference between Mixed Model and Real Data is only 6% whereas between Random
Model and Real Data is 10%.
### V-B Quantization
Figure 9 shows the accuracy of LeNeT-5 on the MNIST dataset at different
precision levels. We see that in each precision level our the Mixed Model has
more similar behavior than Random Model has to the real data. Figure 9a is the
average of all options demonstrated in Figure 7.
Figure 8: Average of all options on voltage and resilience behaviour of
artificial models and experimental (real) data on cuda-convnet network without
reducing precision.
(a) 32-bit single-precision floating point
(b) 16-bit half-precision floating point
(c) 8-bit fixed point (Q4.4)
(d) 4-bit fixed point (Q2.2)
(e) 1-bit (Binary)
Figure 9: Average of all options on voltage and resilience behaviour of
artificial models and experimental (real) data on LeNeT-5 network with
different precisions.
Figure 9a shows the resilience of both artificial models and real data under
nominal precision level (32-bit single-precision floating point) in terms of
accuracy. Mixed Model is 3.74x closer than Random Model to the real data on
average. At $V_{crash}$, the accuracy difference between Mixed Model and Real
Data is 2.46%, on the other hand, the difference between Random Model and Real
Data is 12.47%.
Figure 9b shows the accuracy of LeNeT for weights with the precision reduced
to 16-bit half-precision floating point. The behavior of all models are
similar to the 32-bit architecture, since 16-bit precision is mostly cover all
values in 32-bit [67, 68] without significant affect on accuracy. The highest
accuracy difference between real data and Mixed Model is 4.47% at 0.55V
whereas the highest accuracy difference between real data and Random Model is
9.05% at 0.57V. The average accuracy difference of all voltage level between
real data and Mixed Model is 2.25% whereas between real data and Random Model
is 4.84%, which is 2.15x worse than Mixed Model in terms of similarity of the
behavior of real data.
Figure 9c shows the accuracy of LeNeT with 8-bit precision for weights. We see
that compared to 32-bit, the network are less resilient to undervolting
faults. On average, Mixed Model is 7x better than Random Model in terms of
approximation to real data. At $560mV$, the accuracy gap between Mixed Model
and Real Data is 4% while the difference between Random Model and Real Data is
23%.
Figure 9d shows the accuracy of both artificial models and real data when the
precision of weights reduces to 4-bit. We see that reducing the precision of
weights 8-bit to 4-bit network become more resilient and more fault-tolerant
to errors based on voltage underscaling. Although, at $V_{nom}$, 8-bit LeNeT
is more accurate than 4-bit LeNeT, at $V_{crash}$, 8-bit accuracy is 14%
whereas 4-bit accuracy is 62%. Even in this unexpected situation, artificial
models of the MoRS framework have similar behavior to real data. On average,
Mixed Model is 2x closer to real data than Random Model. At $550mV$, the
difference in accuracy between Mixed Model and real data is around 1.5%, while
the difference in accuracy between Random Model and real data is 3%.
Figure 9e shows the accuracy of LeNeT when precision reduces to 1-bit. In
1-bit tests, we map weights to three different value sets. First one is
{-1,1}, second is {-1,0} and the last one is {0,1}. However, we have not
observed much difference between value sets. As we see in 4-bit, 1-bit LeNeT
network becomes more resilient to faults. The difference between $V_{nom}$
accuracy and $V_{crash}$ accuracy is 0.79%. Both artificial models have the
same behavior and do not have significant difference in accuracy. Mixed Model
has 0.03% difference in accuracy, whereas Random Model is 0.02%. Because of
these negligible statistics, we do not add 1-bit LeNeT to Figure 10.
To point out the resilience of reduced precision networks, we examine the
accuracy drop between $V_{nom}$ and $V_{crash}$. The drop is 68.4%, 12.9% and
0.79% for 8-bit, 4-bit and 1-bit LeNet, respectively.
LeNeT-5 Precision | SRAM Utilization | Accuracy
---|---|---
32-bit FP | 850 | 99.05%
16-bit Half FP | 425 | 97.03%
8-bit (Q4.4) | 213 | 82.79%
4-bit (Q2.2) | 107 | 74.75%
1-bit (Binary) | 27 | 49.59%
TABLE VI: LeNeT-5 SRAM Block Utilization and Inference Accuracy(at $V_{nom}$)
under different precisions
Table VI shows the SRAM utilization and accuracy in various precision levels
of weights. Inference accuracy represents the accuracy at nominal voltage
level ($V_{nom}$). 32-bit FP denotes the 32-bit single-precision floating
point, 16-bit Half FP stands for 16-bit half-precision floating points.
### V-C Comparison of Artificial Models
Figure 10 shows the accuracy gap between Real Data and both two artificial
models. We see that on the average of all benchmarks, Mixed Model is 3.21x
closer than Random Model on average. For most of the benchmarks, the maximum
difference in accuracy between Mixed Model and Real Model is under 5%.
However, for the Random Model, the maximum difference in accuracy is 23.2%.
We conclude that the proposed model not only has the same behavior with real
data against undervolting effects but also can imitate real data with a
tolerable difference in terms of accuracy. Most importantly, if the system is
not resilient to faults, randomized fault injection does not show the behavior
of real data. To be precise in how undervolting affects systems, it has to
profile the real data in fine-grained. Coarse-grained features are
insufficient to model real SRAM behavior when undervolting is performed.
Figure 10: The accuracy difference between Real Data and each artificial
model for all benchmarks.
## VI Related Work
To the best of our knowledge, this study provides the first approximate fault
modeling framework and injection in voltage underscaled SRAMs. In this
section, we discuss related work on fault injection and modeling on
undervolted systems and the resilience of DNNs.
Resilience of DNNs. DNNs are inherently reliable to faults. However, in harsh
environments, process variations, voltage underscaling can cause significant
accuracy loss. Reagen et al.[16] propose Minerva a fault mitigation mechanism
to mitigate low-voltage SRAM faults effects in DNN accelerators. Salami et
al.[3] study undervolting SRAM-based FPGA on-chip memories and present
intelligently-constrained BRAM placement to mitigate undervolting faults in
the NN classification layer. Torres-Huitzil et al.[29] present a comprehensive
review on fault and error tolerance in neural networks and mention mitigation
and fault injection techniques. They also mention that more realist/novel
fault models need to be developed to understand the effects of faults on
neural networks deeply as an open challenge. Deng et al.[30] present a
retraining-based mitigation technique for neural networks to become more
resilient.
Fault injection. Fault injection is a widely used technique in resilience
studies. Also, fault injection is used as bit-flip-based adversarial weight
attacks also called bit-flip attacks(BFA)[69, 70, 71] and fault-injection
attacks[72, 73]. Many studies focus on the reliability and resilience of
systems on soft-errors, noise[74, 75, 63, 76], and voltage underscaling by
injecting faults. Koppula et al.[34] propose a framework, EDEN, that proposes
combining training on profiled DRAM faults in terms of timing violations and
voltage underscaling with mitigation strategies and mappings. EDEN provides
four different error models since uniform random distribution does not cover
whole DRAMs. Chatzidimitrou et al.[43, 32] inject faults randomly to branch
prediction units to examine the effects of voltage underscaling.
Chandramoorthy et al.[17] study the undervolting faults in different layers of
networks by injecting faults to SRAMs randomly and do not take account of the
patterns or spatial distribution of bit errors. Stutz et al.[77] propose
random bit error training assumed voltage underscaled SRAMs faults distribute
randomly. Salami et al. [33] study the resilience of RTL NN Accelerators and
fault characterization and mitigation. To characterize and mitigate they
assume that each bitcell of SRAMs has the same probability. Yang et al. [23]
study energy-efficient CNNs by performing voltage scaling on SRAMs. To study
the effect of bit errors they hypothesize that errors in SRAM are roughly
uniformly distributed. The prior work[78] on Near-Threshold Voltage FinFET
SRAMs presents a fault model for SRAMs based on uniform random distribution.
Givaki et al.[79] study the resilience of DNNs under reduced voltage SRAM-
based FPGA on-chip memories by using directly the experimental data to examine
the training phase of DNNs.
All of these undervolted on-chip fault injection studies perform injection
randomly and do not take into account fine-grained profiling such as spatial
distances between cells, row-based and column-based approaches. Randomly
injecting faults approach can cause misleading to understand how the system
works under the low-voltage domain. As we mentioned in Section 4, if the
system does not have much resilience to voltage underscaling, the Mixed Model
of MoRS is 7x closer than the randomly injected model to the real data.
## VII Conclusion
In this paper, we propose MoRS, a framework that generates the first
approximate fault injection model i.e., artificial fault maps. The advantage
of the proposed framework is to inject errors into various systems including
heterogeneous computing devices. We evaluated the accuracy of the proposed
framework for state-of-the-art DNN applications. To evaluate our proposed
model, we measure the difference in accuracy between artificial error models
and real data. We show that compared to random model-based error injections,
the proposed model can provide 3.21x on average closer than to the real data.
## Acknowledgement
This work is partially funded by Open Transprecision Computing (OPRECOM)
project, Summer of Code 2020.
## References
* [1] H. Jeon, G. S. Ravi, N. S. Kim, and M. Annavaram, “Gpu register file virtualization,” in _Proceedings of the 48th International Symposium on Microarchitecture_ , ser. MICRO-48. New York, NY, USA: Association for Computing Machinery, 2015, p. 420–432. [Online]. Available: https://doi.org/10.1145/2830772.2830784
* [2] J. Haj-Yihia, A. Yasin, Y. B. Asher, and A. Mendelson, “Fine-grain power breakdown of modern out-of-order cores and its implications on skylake-based systems,” _ACM Trans. Archit. Code Optim._ , vol. 13, no. 4, Dec. 2016. [Online]. Available: https://doi.org/10.1145/3018112
* [3] B. Salami, O. S. Unsal, and A. C. Kestelman, “Comprehensive evaluation of supply voltage underscaling in fpga on-chip memories,” in _2018 51st Annual IEEE/ACM International Symposium on Microarchitecture (MICRO)_. IEEE, 2018, pp. 724–736.
* [4] G. Papadimitriou, A. Chatzidimitriou, D. Gizopoulos, V. J. Reddi, J. Leng, B. Salami, O. S. Unsal, and A. C. Kestelman, “Exceeding conservative limits: A consolidated analysis on modern hardware margins,” _IEEE Transactions on Device and Materials Reliability_ , vol. 20, no. 2, pp. 341–350, 2020.
* [5] D. Gizopoulos, G. Papadimitriou, A. Chatzidimitriou, V. J. Reddi, B. Salami, O. S. Unsal, A. C. Kestelman, and J. Leng, “Modern hardware margins: Cpus, gpus, fpgas recent system-level studies,” in _2019 IEEE 25th International Symposium on On-Line Testing and Robust System Design (IOLTS)_. IEEE, 2019, pp. 129–134.
* [6] A. Bacha and R. Teodorescu, “Using ecc feedback to guide voltage speculation in low-voltage processors,” in _2014 47th Annual IEEE/ACM International Symposium on Microarchitecture_ , 2014, pp. 306–318.
* [7] G. Papadimitriou, A. Chatzidimitriou, and D. Gizopoulos, “Adaptive voltage/frequency scaling and core allocation for balanced energy and performance on multicore cpus,” in _2019 IEEE International Symposium on High Performance Computer Architecture (HPCA)_ , 2019, pp. 133–146.
* [8] K. Parasyris, P. Koutsovasilis, V. Vassiliadis, C. D. Antonopoulos, N. Bellas, and S. Lalis, “A framework for evaluating software on reduced margins hardware,” in _2018 48th Annual IEEE/IFIP International Conference on Dependable Systems and Networks (DSN)_ , 2018, pp. 330–337.
* [9] G. Papadimitriou, M. Kaliorakis, A. Chatzidimitriou, D. Gizopoulos, P. Lawthers, and S. Das, “Harnessing voltage margins for energy efficiency in multicore cpus,” in _2017 50th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO)_ , 2017, pp. 503–516.
* [10] A. Bacha and R. Teodorescu, “Dynamic reduction of voltage margins by leveraging on-chip ecc in itanium ii processors,” _SIGARCH Comput. Archit. News_ , vol. 41, no. 3, p. 297–307, Jun. 2013. [Online]. Available: https://doi.org/10.1145/2508148.2485948
* [11] A. Zou, J. Leng, X. He, Y. Zu, C. D. Gill, V. J. Reddi, and X. Zhang, “Voltage-stacked gpus: A control theory driven cross-layer solution for practical voltage stacking in gpus,” in _Proceedings of the 51st Annual IEEE/ACM International Symposium on Microarchitecture_ , ser. MICRO-51. IEEE Press, 2018, p. 390–402. [Online]. Available: https://doi.org/10.1109/MICRO.2018.00039
* [12] B. Salami, E. B. Onural, I. E. Yuksel, F. Koc, O. Ergin, A. C. Kestelman, O. S. Unsal, H. Sarbazi-Azad, and O. Mutlu, “An experimental study of reduced-voltage operation in modern fpgas for neural network acceleration,” 2020\.
* [13] B. Salami, O. S. Unsal, and A. C. Kestelman, “Evaluating built-in ecc of fpga on-chip memories for the mitigation of undervolting faults,” in _2019 27th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP)_. IEEE, 2019, pp. 242–246.
* [14] B. Salami, O. Unsal, and A. Cristal, “Fault characterization through fpga undervolting,” in _2018 28th International Conference on Field Programmable Logic and Applications (FPL)_. IEEE, 2018, pp. 85–853.
* [15] B. Salami, “Aggressive undervolting of fpgas: power & reliability trade-offs,” 2018.
* [16] B. Reagen, P. Whatmough, R. Adolf, S. Rama, H. Lee, S. K. Lee, J. M. Hernández-Lobato, G. Wei, and D. Brooks, “Minerva: Enabling low-power, highly-accurate deep neural network accelerators,” in _2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA)_ , 2016, pp. 267–278.
* [17] N. Chandramoorthy, K. Swaminathan, M. Cochet, A. Paidimarri, S. Eldridge, R. V. Joshi, M. M. Ziegler, A. Buyuktosunoglu, and P. Bose, “Resilient low voltage accelerators for high energy efficiency,” in _2019 IEEE International Symposium on High Performance Computer Architecture (HPCA)_ , 2019, pp. 147–158.
* [18] J. Zhang, K. Rangineni, Z. Ghodsi, and S. Garg, “Thundervolt: enabling aggressive voltage underscaling and timing error resilience for energy efficient deep learning accelerators,” in _Proceedings of the 55th Annual Design Automation Conference_ , 2018, pp. 1–6.
* [19] K. K. Chang, A. G. Yağlıkçı, S. Ghose, A. Agrawal, N. Chatterjee, A. Kashyap, D. Lee, M. O’Connor, H. Hassan, and O. Mutlu, “Understanding reduced-voltage operation in modern dram devices: Experimental characterization, analysis, and mechanisms,” _Proc. ACM Meas. Anal. Comput. Syst._ , vol. 1, no. 1, Jun. 2017. [Online]. Available: https://doi.org/10.1145/3084447
* [20] H. David, C. Fallin, E. Gorbatov, U. R. Hanebutte, and O. Mutlu, “Memory power management via dynamic voltage/frequency scaling,” in _Proceedings of the 8th ACM International Conference on Autonomic Computing_ , ser. ICAC ’11. New York, NY, USA: Association for Computing Machinery, 2011, p. 31–40. [Online]. Available: https://doi.org/10.1145/1998582.1998590
* [21] Q. Deng, D. Meisner, L. Ramos, T. F. Wenisch, and R. Bianchini, “Memscale: Active low-power modes for main memory,” in _Proceedings of the Sixteenth International Conference on Architectural Support for Programming Languages and Operating Systems_ , ser. ASPLOS XVI. New York, NY, USA: Association for Computing Machinery, 2011, p. 225–238. [Online]. Available: https://doi.org/10.1145/1950365.1950392
* [22] S. S. N. Larimi, B. Salami, O. S. Unsal, A. C. Kestelman, H. Sarbazi-Azad, and O. Mutlu, “Understanding power consumption and reliability of high-bandwidth memory with voltage underscaling,” 2020.
* [23] L. Yang and B. Murmann, “Sram voltage scaling for energy-efficient convolutional neural networks,” in _2017 18th International Symposium on Quality Electronic Design (ISQED)_ , 2017, pp. 7–12.
* [24] L. Yang and B. Murmann, “Approximate sram for energy-efficient, privacy-preserving convolutional neural networks,” _2017 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)_ , pp. 689–694, 2017.
* [25] Y. Cai, G. Yalcin, O. Mutlu, E. F. Haratsch, A. Crista, O. S. Unsal, and K. Mai, “Error analysis and retention-aware error management for nand flash memory.” _Intel Technology Journal_ , vol. 17, no. 1, 2013.
* [26] Y. Cai, S. Ghose, E. F. Haratsch, Y. Luo, and O. Mutlu, “Error characterization, mitigation, and recovery in flash-memory-based solid-state drives,” _Proceedings of the IEEE_ , vol. 105, no. 9, pp. 1666–1704, 2017\.
* [27] Y. Cai, E. F. Haratsch, O. Mutlu, and K. Mai, “Error patterns in mlc nand flash memory: Measurement, characterization, and analysis,” in _2012 Design, Automation Test in Europe Conference Exhibition (DATE)_ , 2012, pp. 521–526.
* [28] Y. Cai, Y. Luo, S. Ghose, and O. Mutlu, “Read disturb errors in mlc nand flash memory: Characterization, mitigation, and recovery,” in _2015 45th Annual IEEE/IFIP International Conference on Dependable Systems and Networks_ , 2015, pp. 438–449.
* [29] C. Torres-Huitzil and B. Girau, “Fault and error tolerance in neural networks: A review,” _IEEE Access_ , vol. 5, pp. 17 322–17 341, 2017.
* [30] J. Deng, Y. Fang, Z. Du, Y. Wang, H. Li, O. Temam, P. Ienne, D. Novo, X. Li, Y. Chen, and C. Wu, “Retraining-based timing error mitigation for hardware neural networks,” in _2015 Design, Automation Test in Europe Conference Exhibition (DATE)_ , 2015, pp. 593–596.
* [31] B. Salami., “FPGA BRAMs Undervolting Study,” https://github.com/behzadsalami/FPGA-BRAMs-Undervoltig-Study, 2018.
* [32] A. Chatzidimitriou, G. Papadimitriou, D. Gizopoulos, S. Ganapathy, and J. Kalamatianos, “Analysis and characterization of ultra low power branch predictors,” in _2018 IEEE 36th International Conference on Computer Design (ICCD)_ , 2018, pp. 144–147.
* [33] B. Salami, O. S. Unsal, and A. C. Kestelman, “On the resilience of rtl nn accelerators: Fault characterization and mitigation,” in _2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)_ , 2018, pp. 322–329.
* [34] S. Koppula, L. Orosa, A. G. Yağlıkçı, R. Azizi, T. Shahroodi, K. Kanellopoulos, and O. Mutlu, “Eden: Enabling energy-efficient, high-performance deep neural network inference using approximate dram,” in _Proceedings of the 52nd Annual IEEE/ACM International Symposium on Microarchitecture_ , ser. MICRO ’52. New York, NY, USA: Association for Computing Machinery, 2019, p. 166–181. [Online]. Available: https://doi.org/10.1145/3352460.3358280
* [35] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” _Proceedings of the IEEE_ , vol. 86, no. 11, pp. 2278–2324, 1998.
* [36] A. Krizhevsky., “cuda-convnet,” https://github.com/akrizhevsky/cuda-convnet2, 2014.
* [37] M. Sabbagh, C. Gongye, Y. Fei, and Y. Wang, “Evaluating fault resiliency of compressed deep neural networks,” in _2019 IEEE International Conference on Embedded Software and Systems (ICESS)_ , 2019, pp. 1–7.
* [38] G. Li, S. K. S. Hari, M. Sullivan, T. Tsai, K. Pattabiraman, J. Emer, and S. W. Keckler, “Understanding error propagation in deep learning neural network (dnn) accelerators and applications,” in _Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis_ , ser. SC ’17. New York, NY, USA: Association for Computing Machinery, 2017. [Online]. Available: https://doi.org/10.1145/3126908.3126964
* [39] P. Upadhyay, S. Ghosh, R. Kar, D. Mandal, and S. P. Ghoshal, “Low static and dynamic power mtcmos based 12t sram cell for high speed memory system,” in _2014 11th International Joint Conference on Computer Science and Software Engineering (JCSSE)_ , 2014, pp. 212–217.
* [40] T. Azam, B. Cheng, and D. R. S. Cumming, “Variability resilient low-power 7t-sram design for nano-scaled technologies,” in _2010 11th International Symposium on Quality Electronic Design (ISQED)_ , 2010, pp. 9–14.
* [41] G. Chen, D. Sylvester, D. Blaauw, and T. Mudge, “Yield-driven near-threshold sram design,” _IEEE Transactions on Very Large Scale Integration (VLSI) Systems_ , vol. 18, no. 11, pp. 1590–1598, 2010.
* [42] J. Leng, A. Buyuktosunoglu, R. Bertran, P. Bose, and V. J. Reddi, “Safe limits on voltage reduction efficiency in gpus: A direct measurement approach,” in _2015 48th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO)_ , 2015, pp. 294–307.
* [43] A. Chatzidimitriou, G. Panadimitriou, D. Gizopoulos, S. Ganapathy, and J. Kalamatianos, “Assessing the effects of low voltage in branch prediction units,” in _2019 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS)_ , 2019, pp. 127–136.
* [44] L. Rivière, Z. Najm, P. Rauzy, J. Danger, J. Bringer, and L. Sauvage, “High precision fault injections on the instruction cache of armv7-m architectures,” in _2015 IEEE International Symposium on Hardware Oriented Security and Trust (HOST)_ , 2015, pp. 62–67.
* [45] M. Kaliorakis, S. Tselonis, A. Chatzidimitriou, N. Foutris, and D. Gizopoulos, “Differential fault injection on microarchitectural simulators,” in _2015 IEEE International Symposium on Workload Characterization_ , 2015, pp. 172–182.
* [46] Y. Lecun and C. Cortes, “The MNIST database of handwritten digits.” http://yann.lecun.com/exdb/mnist/, 1999.
* [47] A. Krizhevsky, V. Nair, and G. Hinton, “he CIFAR-10 dataset,” https://www.cs.toronto.edu/ kriz/cifar.html.
* [48] A. Zhou, A. Yao, Y. Guo, L. Xu, and Y. Chen, “Incremental network quantization: Towards lossless cnns with low-precision weights,” 2017.
* [49] S. Han, X. Liu, H. Mao, J. Pu, A. Pedram, M. A. Horowitz, and W. J. Dally, “Eie: Efficient inference engine on compressed deep neural network,” 2016.
* [50] Z. Zhu, H. Sun, Y. Lin, G. Dai, L. Xia, S. Han, Y. Wang, and H. Yang, “A configurable multi-precision cnn computing framework based on single bit rram,” in _Proceedings of the 56th Annual Design Automation Conference 2019_ , ser. DAC ’19. New York, NY, USA: Association for Computing Machinery, 2019. [Online]. Available: https://doi.org/10.1145/3316781.3317739
* [51] P. Molchanov, S. Tyree, T. Karras, T. Aila, and J. Kautz, “Pruning convolutional neural networks for resource efficient inference,” 2017.
* [52] R. Yazdani, M. Riera, J. Arnau, and A. González, “The dark side of dnn pruning,” in _2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA)_ , 2018, pp. 790–801.
* [53] S. Han, J. Pool, J. Tran, and W. J. Dally, “Learning both weights and connections for efficient neural networks,” 2015.
* [54] Y. Shen, M. Ferdman, and P. Milder, “Escher: A cnn accelerator with flexible buffering to minimize off-chip transfer,” in _2017 IEEE 25th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM)_ , 2017, pp. 93–100.
* [55] C. Deng, S. Liao, Y. Xie, K. K. Parhi, X. Qian, and B. Yuan, “Permdnn: Efficient compressed dnn architecture with permuted diagonal matrices,” 2020\.
* [56] Y. Shen, M. Ferdman, and P. Milder, “Maximizing cnn accelerator efficiency through resource partitioning,” 2018.
* [57] C. Zhang, P. Li, G. Sun, Y. Guan, B. Xiao, and J. Cong, “Optimizing fpga-based accelerator design for deep convolutional neural networks,” in _Proceedings of the 2015 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays_ , ser. FPGA ’15. New York, NY, USA: Association for Computing Machinery, 2015, p. 161–170. [Online]. Available: https://doi.org/10.1145/2684746.2689060
* [58] M. Riera, J. Arnau, and A. Gonzalez, “Computation reuse in dnns by exploiting input similarity,” in _2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA)_ , 2018, pp. 57–68.
* [59] M. Courbariaux and Y. Bengio, “Binarynet: Training deep neural networks with weights and activations constrained to +1 or -1,” _CoRR_ , vol. abs/1602.02830, 2016. [Online]. Available: http://arxiv.org/abs/1602.02830
* [60] B. Jacob, S. Kligys, B. Chen, M. Zhu, M. Tang, A. G. Howard, H. Adam, and D. Kalenichenko, “Quantization and training of neural networks for efficient integer-arithmetic-only inference,” _CoRR_ , vol. abs/1712.05877, 2017. [Online]. Available: http://arxiv.org/abs/1712.05877
* [61] J. Wu, C. Leng, Y. Wang, Q. Hu, and J. Cheng, “Quantized convolutional neural networks for mobile devices,” _CoRR_ , vol. abs/1512.06473, 2015. [Online]. Available: http://arxiv.org/abs/1512.06473
* [62] K. Ueyoshi, K. Ando, K. Hirose, S. Takamaeda-Yamazaki, J. Kadomoto, T. Miyata, M. Hamada, T. Kuroda, and M. Motomura, “Quest: A 7.49tops multi-purpose log-quantized dnn inference engine stacked on 96mb 3d sram using inductive-coupling technology in 40nm cmos,” in _2018 IEEE International Solid - State Circuits Conference - (ISSCC)_ , 2018, pp. 216–218.
* [63] B. Reagen, U. Gupta, L. Pentecost, P. Whatmough, S. K. Lee, N. Mulholland, D. Brooks, and G. Wei, “Ares: A framework for quantifying the resilience of deep neural networks,” in _2018 55th ACM/ESDA/IEEE Design Automation Conference (DAC)_ , 2018, pp. 1–6.
* [64] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell, “Caffe: Convolutional architecture for fast feature embedding,” in _Proceedings of the 22nd ACM International Conference on Multimedia_ , ser. MM ’14. New York, NY, USA: Association for Computing Machinery, 2014, p. 675–678. [Online]. Available: https://doi.org/10.1145/2647868.2654889
* [65] M. Milde, D. Neil, A. Aimar, T. Delbruck, and G. Indiveri, “Adaption: Toolbox and benchmark for training convolutional neural networks with reduced numerical precision weights and activation,” 11 2017.
* [66] NVIDIA., “NVIDIA-caffe extension,” https://github.com/NVIDIA/caffe, 2017\.
* [67] P. Micikevicius, S. Narang, J. Alben, G. Diamos, E. Elsen, D. Garcia, B. Ginsburg, M. Houston, O. Kuchaiev, G. Venkatesh _et al._ , “Mixed precision training,” _arXiv preprint arXiv:1710.03740_ , 2017.
* [68] S. Markidis, S. W. Der Chien, E. Laure, I. B. Peng, and J. S. Vetter, “Nvidia tensor core programmability, performance & precision,” in _2018 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)_. IEEE, 2018, pp. 522–531.
* [69] A. S. Rakin, Z. He, J. Li, F. Yao, C. Chakrabarti, and D. Fan, “T-bfa: Targeted bit-flip adversarial weight attack,” 2021.
* [70] Z. He, A. S. Rakin, J. Li, C. Chakrabarti, and D. Fan, “Defending and harnessing the bit-flip based adversarial weight attack,” in _2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)_ , 2020, pp. 14 083–14 091.
* [71] A. S. Rakin, Z. He, and D. Fan, “Bit-flip attack: Crushing neural network with progressive bit search,” 2019.
* [72] Y. Liu, L. Wei, B. Luo, and Q. Xu, “Fault injection attack on deep neural network,” in _2017 IEEE/ACM International Conference on Computer-Aided Design (ICCAD)_ , 2017, pp. 131–138.
* [73] M. S. Kelly, K. Mayes, and J. F. Walker, “Characterising a cpu fault attack model via run-time data analysis,” in _2017 IEEE International Symposium on Hardware Oriented Security and Trust (HOST)_ , 2017, pp. 79–84.
* [74] G. Li, S. K. S. Hari, M. Sullivan, T. Tsai, K. Pattabiraman, J. Emer, and S. W. Keckler, “Understanding error propagation in deep learning neural network (dnn) accelerators and applications,” in _Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis_ , ser. SC ’17. New York, NY, USA: Association for Computing Machinery, 2017. [Online]. Available: https://doi.org/10.1145/3126908.3126964
* [75] M. A. Neggaz, I. Alouani, P. R. Lorenzo, and S. Niar, “A reliability study on cnns for critical embedded systems,” in _2018 IEEE 36th International Conference on Computer Design (ICCD)_ , 2018, pp. 476–479.
* [76] A. H. Salavati and A. Karbasi, “Multi-level error-resilient neural networks,” in _2012 IEEE International Symposium on Information Theory Proceedings_ , 2012, pp. 1064–1068.
* [77] D. Stutz, N. Chandramoorthy, M. Hein, and B. Schiele, “Bit error robustness for energy-efficient dnn accelerators,” 2020.
* [78] S. Ganapathy, J. Kalamatianos, K. Kasprak, and S. Raasch, “On characterizing near-threshold sram failures in finfet technology,” in _2017 54th ACM/EDAC/IEEE Design Automation Conference (DAC)_ , 2017, pp. 1–6.
* [79] K. Givaki, B. Salami, R. Hojabr, S. R. Tayaranian, A. Khonsari, D. Rahmati, S. Gorgin, A. Cristal, and O. S. Unsal, “On the resilience of deep learning for reduced-voltage fpgas,” in _2020 28th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP)_. IEEE, 2020, pp. 110–117.
| İsmail Emir Yüksel is an MSc student and researcher in the Computer
Engineering Department of TOBB University of Economics and Technology (TOBB
ETÜ). He received his BSc in Electrical Electronics Engineering from TOBB
University of Economics and Technology in 2019. His research interests are
energy-efficient heterogeneous computing and low-power & fault-resilient
hardware accelerators.
---|---
| Behzad Salami is a post-doctoral researcher in the Computer Science (CS)
department of Barcelona Supercomputing Center (BSC) and an affiliated research
member of SAFARI Research Group at ETH Zurich. He received his Ph.D. with
honors in Computer Architecture from Universitat Politécnica de Catalunya
(UPC) in 2018. Also, he obtained MSc and BSc degrees in Computer Engineering
from Amirkabir University of Technology (AUT) and Iran University of Science
and Technology (IUST), respectively. He has received mutiple awards and grants
for his research. His research interests are heterogeneous systems, low-power
& fault-resilient hardware accelerators, and near-data processing systems.
Contact him at<EMAIL_ADDRESS>
---|---
| Oğuz Ergin is a professor in the department of computer engineering in TOBB
University of Economics and Technology. He received his BS in electrical and
electronics engineering from Middle East Technical University, MS, and Ph.D.
in computer science from the State University of New York at Binghamton. He
was a senior research scientist in Intel Barcelona Research Center prior to
joining TOBB ETÜ. He is currently leading a research group in TOBB ETÜ working
on energy-efficient, reliable, and high-performance computer architectures.
---|---
| Osman Sabri Ünsal is co-manager of the Parallel Paradigms for Computer
Architecture research group at Barcelona Supercomputing Center (BSC). He got
his B.S., M.S., and Ph.D. in Computer Engineering from Istanbul Technical
University, Brown University, and the University of Massachusetts, Amherst
respectively. His current research interests are in computer architecture,
fault-tolerance, energy-efficiency, and heterogeneous computing. He is
currently leading LEGaTO EU H2020 research project on heterogeneous energy-
efficiency computing.
---|---
| Adrián Cristal Kestelman received the Licenciatura degree in Computer
Science from the Faculty of Exact and Natural Sciences, Universidad de Buenos
Aires, Buenos Aires, Argentina, in 1995, and the Ph.D. degree in Computer
Science from the Universitat Politécnica de Catalunya (UPC), Barcelona, Spain.
Since 2006, he is a co-manager of the Computer Architecture for Parallel
Paradigms Research Group at Barcelona Supercomputing Center (BSC). His current
research interests include the areas of microarchitecture, multicore, and
heterogeneous architectures, and programming models for multicore
architectures. Currently, he is leading the architecture development of the
vector processor unit in the European Processor Initiative.
---|---
|
# Stochastic modulational instability in the nonlinear Schrödinger equation
with colored random dispersion
Andrea Armaroli Univ. Lille, CNRS, UMR 8523-PhLAM-Physique des Lasers Atomes
et Molécules, F-59000 Lille, France Guillaume Dujardin Univ. Lille, Inria,
CNRS, UMR 8524 - Laboratoire Paul Painlevé, F-59000 Lille, France Alexandre
Kudlinski Univ. Lille, CNRS, UMR 8523-PhLAM-Physique des Lasers Atomes et
Molécules, F-59000 Lille, France Arnaud Mussot Univ. Lille, CNRS, UMR
8523-PhLAM-Physique des Lasers Atomes et Molécules, F-59000 Lille, France
Stefano Trillo Department of Engineering, University of Ferrara, I-44122
Ferrara, Italy Stephan De Bièvre Univ. Lille, CNRS, Inria, UMR 8524 -
Laboratoire Paul Painlevé, F-59000 Lille, France Matteo Conforti Univ.
Lille, CNRS, UMR 8523-PhLAM-Physique des Lasers Atomes et Molécules, F-59000
Lille, France
###### Abstract
We study modulational instability (MI) in optical fibers with random group-
velocity dispersion (GVD). We consider Gaussian and dichotomous colored
stochastic processes. We resort to different analytical methods (namely, the
cumulant expansion and the functional approach) and assess their reliability
in estimating the MI gain of stochastic origin. If the power spectral density
(PSD) of the GVD fluctuations is centered at null wavenumber, we obtain low-
frequency MI sidelobes which converge to those given by a white noise
perturbation when the correlation length tends to $0$. If instead the
stochastic processes are modulated in space, one or more MI sidelobe pairs
corresponding to the well-known parametric resonance (PR) condition can be
found. A transition from small and broad sidelobes to peaks nearly
indistinguishable from PR-MI is predicted, in the limit of large perturbation
amplitudes and correlation lengths of the random process. We find that the
cumulant expansion provides good analytical estimates for small PSD values and
small correlation lengths, when the MI gain is very small. The functional
approach is rigorous only for the dichotomous processes, but allows us to
model a wider range of parameters and to predict the existence of MI sidelobes
comparable to those observed in homogeneous fibers of anomalous GVD.
## I Introduction
A physical system exhibiting an interplay of weak nonlinearity and group
velocity dispersion (GVD) is subject to modulational instability (MI), i.e.,
the destabilization of a homogeneous state (plane or continuous waves), via
the exponential growth of small harmonic perturbations on a uniform background
[1]. After pioneering works in fluid mechanics [2, 3], MI was discovered in
electromagnetic waves [4] as well as in plasmas [5]; in the 80s the phenomenon
was observed in nonlinear fiber optics [6]. In uniform fibers, MI arises for
anomalous (negative) GVD, but it may also appear for normal GVD if
polarization [7], higher-order modes [8] or higher-order dispersion are
considered [9]. A different kind of MI related to a parametric resonance (PR)
mechanism emerges when the dispersion or the nonlinearity of the fiber are
periodically modulated [10, 11, 12, 13].
The impact of a random variation of GVD on MI was also the subject of a
considerable research effort. The particular case in which the GVD is
perturbed by a Gaussian white noise, i.e., a process exhibiting a vanishing
correlation length or equivalently a flat power spectral density (PSD), is
exactly solvable[14, 15, 16, 17, 18]. When the unperturbed fiber has an
anomalous GVD, the conventional MI gain profile is deformed due to the random
perturbation. In addition, MI sidebands of stochastic origin appear in the
case of normal GVD. A white noise, however, implies arbitrarily large
variations of GVD on arbitrarily small scales: an idealization that does not
always provide a relevant modeling of the randomness that may occur in
physical fibers. A non-conclusive theoretical study of parametric
amplification in the case of a GVD perturbed by a Gaussian process with a
finite correlation length was proposed in [19], and a numerical study can be
found in [20].
We aim at studying the MI problem in a class of random-GVD fibers that is both
experimentally accessible and theoretically tractable. In [21], we studied the
case of a GVD perturbed by randomly located sharp and large kicks. Two
different families of random processes were chosen to generate their mutual
spacing and amplitude. Different MI sidebands were predicted, including
multibump ones around zero detuning and others localized around PR
frequencies.
Here, we consider random fluctuations extended in space as in Ref. [14], but
focus on colored processes exhibiting an exponentially decaying
autocorrelation function. A lowpass (LP) and a bandpass (BP, modulated) case
are considered. We resort to cumulant expansion [22, 23, 24, 25, 26, 27, 28],
functional methods [29, 30, 31], and numerical simulations. The LP processes
converge to the white noise results for vanishing correlation lengths. The MI
sidelobes are located in the same detuning range and have amplitudes of the
same order of magnitude of those calculated from white noise. The BP processes
yield PR-like MI sidebands that are also comparable (in their maximal values)
to the white noise MI gain. For this reason, the white noise turns out to be a
reference for many crucial properties of the MI gain in a more generic
setting, eventhough there is little hope of an experimental realization of it.
Since the values of frequency detuning span a large range, the validity of the
different analytical methods has to be questioned, like in other physical
settings and systems [32, 33]. We thus have to rely on different approaches
and comparatively assess their soundness in describing the different features
of the MI sidelobes (position, height, width).
The rest of the paper is organized as follows. After presenting the model
equations (Sec. II), we study in Sec. III the lowpass random fluctuation by
means of the two different approaches mentioned above. We then analyze in Sec.
IV bandpass random fluctuations, for which we find convenient to introduce an
averaging approach to the stochastic equations. In each case we present a
thorough comparison of analytical approaches to numerical results. To close
the paper we report our conclusions and perspectives.
## II Model description
The propagation of optical pulses in a nonlinear optical fiber can be modeled
by the universal nonlinear Schrödinger equation (NLSE),
$i\partial_{z}U-\frac{1}{2}\beta_{2}(z)\partial_{tt}U+\gamma|U|^{2}U=0.$ (1)
Here $U(z,t)$ is the envelope of the optical pulse field in units of
$\sqrt{W}$, function of the propagation distance $z$ and time $t$ in the frame
of reference propagating at the group velocity of the mode propagating in the
fiber, $\beta_{2}$ is the GVD and $\gamma$ is the nonlinear coefficient [34].
We suppose that $\beta_{2}$ fluctuates randomly in $z$, while $\gamma$ stays
constant.
Let $\beta_{2}(z)=\beta_{2}^{0}+\delta\\!\beta(z)$, where $\beta_{2}^{0}>0$
(normal average GVD) and $\delta\\!\beta(z)$ is a stochastic process of zero
mean, that we specify below.
We observe that, for $\delta\\!\beta=0$, Eq. (1) has a continuous wave
($t$-independent) solution $U_{0}(z)=\sqrt{P}\exp\left(i\gamma Pz\right)$.
In order to study the stability of this continuous wave solution, we insert in
Eq. (1) the perturbed solution
${U}(z,t)=\left[\sqrt{P}+\check{x}_{1}(z,t)+i\check{x}_{2}(z,t)\right]\exp(i\gamma
Pz)$, linearize and Fourier-transform the resulting equation with respect to
$t$ ($\omega$ is used as the associated angular frequency detuning from the
carrier $U_{0}$), to obtain
$\frac{\mathrm{d}x}{\mathrm{d}z}=\begin{bmatrix}0&-g(z)\\\
h(z)&0\end{bmatrix}x,$ (2)
with $x\equiv(x_{1},x_{2})^{\mathrm{T}}$ (functions of $\omega$ and $z$),
$g(z)=g_{0}+\delta\\!g(z)$ and $h(z)=h_{0}+\delta\\!g(z)$, with
$g_{0}\equiv\beta_{2}^{0}\frac{\omega^{2}}{2}$, $h_{0}\equiv g_{0}+2\gamma P$,
and $\delta\\!g(z)\equiv{\delta\\!\beta(z)}\frac{\omega^{2}}{2}$. Eq. (2) is a
system of stochastic differential equations (SDEs) for each value $\omega$. In
the following sections we will discuss how to reduce it to a system of
ordinary differential equations (ODEs) for the first and second moments of the
probability density function of $x$ in order to estimate the MI gain.
We recall that we consider only normal average GVD, thus
$k^{2}\equiv g_{0}h_{0}>0$ (3)
and no conventional MI appears for $\delta\\!\beta(z)=0$. The MI sidebands we
predict below are therefore completely ascribed to random fluctuations. The
case of anomalous average GVD will be the subject of future work.
We focus here on two families of random processes, where the stochastic
process $\delta\\!\beta$ is characterized by two parameters $N_{0}>0$ and
$B>0$. The first family, for which we denote $\delta\\!\beta\equiv\chi$, is
characterized by an autocorrelation function of the form
$R_{\chi}(\zeta\equiv
z-z^{\prime})\equiv\langle\chi(z)\chi(z^{\prime})\rangle=\frac{N_{0}B}{4}\exp(-B|\zeta|).$
(4)
We recall that the variance of the process is
$\sigma_{\chi}^{2}=R_{\chi}(0)=\frac{N_{0}B}{4}$. For $B\to\infty$,
$R_{\chi}\to\frac{N_{0}}{2}\delta(\zeta)$, i.e., the white noise
autocorrelation function. Different stochastic processes exhibit this same
autocorrelation function. Here, we consider the Gaussian (often denoted in the
physics literature as Ornstein-Uhlenbeck [28]) and the dichotomous processes,
which find important applications in physics and allow us to obtain workable
approximations [33].
Both are stationary in $z$. By virtue of the Wiener-Khinchin theorem, the PSD
of $\chi$ coincides with the Fourier transform of $R_{\chi}$
$S_{\chi}(\kappa)=\int\limits_{-\infty}^{\infty}{\mathrm{d}\,\zeta}R_{\chi}(\zeta)\exp(i\kappa\zeta)=\frac{N_{0}}{2}\frac{B^{2}}{B^{2}+\kappa^{2}}.$
(5)
The Gaussian process is numerically generated in the $\kappa$-domain by
filtering an approximated white noise of PSD $N_{0}/2$ by means of a lowpass
filter of transfer function
$H_{\mathrm{LP}}(\kappa)=\frac{B}{B+i\kappa}.$ (6)
The dichotomous process is more conveniently obtained directly in the $z$
domain by switching the amplitude of the fluctuation between
$\pm\sigma_{\chi}$ with an exponentially distributed _waiting_ (i.e., between
switching points) length with mean $2/B$. Both are pertinent to fiber optics.
Indeed, the Gaussian process corresponds to a continuous variation of
dispersion as can be obtained by varying the fiber radius during the drawing
process. The dichotomous process corresponds to splicing together fibers with
different GVD $\beta_{2}=\beta_{2}^{0}(1\pm\sigma_{\chi})$ and random lengths.
We refer to the processes belonging to this class, with autocorrelation
function as in Eq. (4) and PSD as in Eq. (6) as LP processes.
The second family of stochastic processes, for which we denote
$\delta\\!\beta\equiv\xi$, is obtained as the modulated version of $\chi$ with
central wavenumber $\kappa_{0}=\frac{2\pi}{\Lambda_{0}}>0$ ($\Lambda_{0}$ is
the associated spatial period), that is written in phase-quadrature
representation
$\xi(z)=\psi_{1}(z)\cos\kappa_{0}z+\psi_{2}(z)\sin\kappa_{0}z.$ (7)
Here $\psi_{1,2}$ two stationary (in $z$) and independent random processes
with zero mean and autocorrelation functions
$R_{\psi_{i}}(\zeta)=\frac{N_{0}B}{2}\exp(-B|\zeta|),$ (8)
for $i=1,2$; moreover,
$\langle\psi_{i}\psi_{j}\rangle=\frac{N_{0}B}{2}\delta_{ij}$. In analogy to
$\chi$, we consider either two Gaussian or two dichotomous processes for
$\psi_{1,2}$. They are generated according to their distribution as is done
for $\chi$. The process $\xi$ thus exhibits an autocorrelation function of the
form
$R_{\xi}(\zeta)=\frac{N_{0}B}{2}\cos\kappa_{0}\zeta\exp(-B|\zeta|).$ (9)
The variance of the process is $\sigma^{2}_{\xi}=\frac{N_{0}B}{2}$, like for
$\psi_{1,2}$.
The PSD of $\xi$ reads
$S_{\xi}(\kappa)=\frac{N_{0}}{2}\left[\frac{B^{2}}{B^{2}+(\kappa-\kappa_{0})^{2}}+\frac{B^{2}}{B^{2}+(\kappa+\kappa_{0})^{2}}\right].$
(10)
We note that, for $B\ll\kappa_{0}$, $S_{\xi}$ is centered approximately around
the wavenumber $\pm\kappa_{0}$, with
$S_{\xi}(\pm\kappa_{0})\approx\frac{N_{0}}{2}=S_{\chi}(0)$ and bandwidth $B$
(in wavenumber units). We refer to this family with autocorrelation function
as in Eq. (9) and PSD as in Eq. (10) as BP processes.
For both LP and BP families, we employ the definition of correlation length
[35]
$\zeta^{\mathrm{c}}\equiv\frac{1}{R(0)}\int\limits_{0}^{\infty}{\mathrm{d}\zeta|R(\zeta)}|$,
which gives $\zeta_{\chi}^{\mathrm{c}}=1/B$ for $\chi$ and
$\zeta^{\mathrm{c}}_{\xi}\approx 2/(\pi B)$ for $\xi$, if $B\ll\kappa_{0}$.
In the next two sections we will study the effect $\chi$ and $\xi$,
respectively, on the MI predicted by Eq. (2).
## III Lowpass random dispersion
First we consider processes with the autocorrelation function given in Eq.
(4). We will discuss both the first and the second moment equations associated
to Eq. (2).
### III.1 Cumulant expansion (first moments)
The cumulant expansion yields a series development for the ODEs associated to
a SDE [28]. It is similar to the Dyson series of scattering theory [36] and
provides a solid base for more ad-hoc schemes [22]. See [26] for a systematic
derivation of terms to arbitrary order.
Let us rewrite Eq. (2) in the standard form
$\dot{{x}}=\left[{A_{1}}+\alpha\eta(z){C_{1}}\right]{x}$ with
${A_{1}}=\begin{bmatrix}0&-g_{0}\\\
h_{0}&0\end{bmatrix},\;{C_{1}}=\begin{bmatrix}0&-1\\\ 1&0\end{bmatrix},$
$\alpha=\frac{\omega^{2}}{4}\sqrt{N_{0}B}$ and $\eta=2\chi/\sqrt{N_{0}B}$ a
random process with unit variance and zero mean. The expansion is performed in
the formal parameter $\alpha$. To second order (the first-order term is
obviously 0), we write the ODE for the first moment
$\langle{x}\rangle=(\langle x_{1}\rangle,\langle x_{2}\rangle)^{\mathrm{T}}$
as
$\frac{\mathrm{d}}{\mathrm{d}z}\langle{x}\rangle=\left[\mathrm{A_{1}}+\alpha^{2}K_{2}^{1}\right]\langle{x}\rangle,$
(11)
with
$K_{2}^{1}=\int\limits_{0}^{\infty}\\!\mathrm{d}\zeta\,C_{1}e^{A_{1}\zeta}C_{1}e^{-A_{1}\zeta}R_{\eta}(\zeta).$
(12)
Other terms can be added in the expansion of Eq. (11): their contribution to
the solution rapidly decreases if
$\varepsilon\equiv\alpha\zeta^{\mathrm{c}},$ (13)
is small, i.e. $\frac{\omega^{2}}{4}\sqrt{\frac{N_{0}}{B}}\ll 1$. For large
detuning or small filter bandwidth $B$, the approximation may be invalid. A
fixed initial condition $x(0)$ can be considered and the solutions of Eq. (11)
do not keep memory of it for $z\gg\zeta_{\mathrm{c}}$. The long term dynamics
being our main focus, we set the limit of integration to infinity in Eq. (12),
see [25].
By tedious but straightforward algebra, we obtain
$\frac{\mathrm{d}}{\mathrm{d}z}\langle{x}\rangle=\begin{bmatrix}-\frac{\omega^{4}}{4}\frac{g_{0}}{2k^{2}}\left[(g_{0}+h_{0})c_{1}-(h_{0}-g_{0})c_{2}\right]&-g_{0}+\frac{\omega^{4}}{8k}(h_{0}-g_{0})c_{3}\\\
h_{0}+\frac{\omega^{4}}{8k}(h_{0}-g_{0})c_{3}&-\frac{\omega^{4}}{4}\frac{h_{0}}{2k^{2}}\left[(g_{0}+h_{0})c_{1}+(h_{0}-g_{0})c_{2}\right]\end{bmatrix}\langle{x}\rangle,$
(14)
with
$\displaystyle c_{1}$
$\displaystyle\equiv\int\limits_{0}^{\infty}\mathrm{d}\zeta
R_{\chi}(\zeta)=\frac{1}{2}S_{\chi}(0)=\frac{N_{0}}{4},$ (15) $\displaystyle
c_{2}$ $\displaystyle\equiv\int\limits_{0}^{\infty}\mathrm{d}\zeta
R_{\chi}(\zeta)\cos
2k\zeta=\frac{1}{2}S_{\chi}(2k)=\frac{N_{0}B^{2}}{4}\frac{1}{B^{2}+4k^{2}}$
$\displaystyle c_{3}$
$\displaystyle\equiv\int\limits_{0}^{\infty}\mathrm{d}\zeta
R_{\chi}(\zeta)\sin(2k\zeta)=\frac{N_{0}Bk}{2}\frac{1}{B^{2}+4k^{2}}.$
Since $c_{i}>0$, $i=1,2,3$, for all $\omega$, and $c_{1}>c_{2}$, it is easy to
verify that the eigenvalues of the matrix in Eq. (14) have both a negative
real part, so that the system Eq. (14) does not predict any MI gain. This
finding is analogous to the conventional harmonic oscillator with random
frequency [28, 33], for which the first moment undergoes only damping. For
this reason it is necessary to resort to the equations for the second moments.
We recall that, for white noise, the cumulant expansion at second order is
exact and in that limit, $B\to\infty$, $c_{1}=c_{2}$ and $c_{3}=0$, so the
eigenvalues of Eq. (14) reduce to Eq. (23) of Ref. [14].
### III.2 Cumulant expansion (second moments)
We now consider second moments. First we let $X_{1}=x_{1}^{2}$,
$X_{2}=x_{2}^{2}$, and $X_{3}=x_{1}x_{2}$ and derive from Eq. (2) a system for
their evolution, which reads
$\frac{\mathrm{d}}{\mathrm{d}z}{X}=\begin{bmatrix}0&0&-2g(z)\\\ 0&0&2h(z)\\\
h(z)&-g(z)&0\end{bmatrix}{X}.$ (16)
In order to perform the cumulant expansion, we write Eq. (16) in standard form
by letting
${A_{2}}=\begin{bmatrix}0&0&-2g_{0}\\\ 0&0&2h_{0}\\\
h_{0}&-g_{0}&0\end{bmatrix},\;{C_{2}}=\begin{bmatrix}0&0&-2\\\ 0&0&2\\\
1&-1&0\end{bmatrix},$
the other quantities $\alpha$ and $\eta$ being the same as in the previous
subsection.
Up to second order, the cumulant expansion reads
$\begin{gathered}\frac{\mathrm{d}}{\mathrm{d}z}\langle{X}\rangle=\left[\mathrm{A_{2}}+\alpha^{2}K_{2}^{2}\right]\langle{X}\rangle,\\\
\text{with
}K_{2}^{2}=\int\limits_{0}^{\infty}\\!\mathrm{d}\zeta\,C_{2}e^{A_{2}\zeta}C_{2}e^{-A_{2}\zeta}R_{\eta}(\zeta),\end{gathered}$
(17)
which gives
$\frac{\mathrm{d}}{\mathrm{d}z}\langle{X}\rangle=\begin{bmatrix}-\frac{\omega^{4}}{4g_{0}}\left[(g_{0}+h_{0})c_{1}+(g_{0}-h_{0})c_{2}\right]&\frac{\omega^{4}}{4h_{0}}\left[(g_{0}+h_{0})c_{1}-(g_{0}-h_{0})c_{2}\right]&-2g_{0}\\\
\frac{\omega^{4}}{4g_{0}}\left[(g_{0}+h_{0})c_{1}+(g_{0}-h_{0})c_{2}\right]&-\frac{\omega^{4}}{4h_{0}}\left[(g_{0}+h_{0})c_{1}-(g_{0}-h_{0})c_{2}\right]&2h_{0}\\\
h_{0}+\frac{\omega^{4}}{4k}(h_{0}-g_{0})c_{3}&-g_{0}+\frac{\omega^{4}}{4k}(h_{0}-g_{0})c_{3}&-\frac{\omega^{4}}{4k^{2}}\left[(g_{0}+h_{0})^{2}c_{1}-(g_{0}-h_{0})^{2}c_{2}\right]\end{bmatrix}\langle{X}\rangle,$
(18)
with the $c_{i}$s defined as above. The validity condition of the expansion is
the same as in the previous sub-section.
As in Ref. [21], the MI of stochastic origin is related to the growth rate of
the second moment. The eigenvalues of the matrix in Eq. (18) can be written
analytically. Their form is rather complicated: in general we have two complex
conjugate eigenvalues ($\lambda_{\pm}$) with negative real part and one
positive real eigenvalue, $\lambda_{0}$. The MI gain is thus defined as
$G_{2}(\omega)\equiv\frac{\lambda_{0}}{2}$. Since $G_{2}(\omega)$ is small for
small $N_{0}$, we proceed like in Ref. [14] to derive the following
approximation
$G_{2}(\omega)\approx\frac{4(\gamma
P)^{2}k\omega^{4}\left[8c_{2}k^{3}+c_{3}\omega^{4}\left(c_{1}j_{0}^{2}-4c_{2}(\gamma
P)^{2}\right)\right]}{64k^{6}-32c_{3}(\gamma
P)^{2}k^{3}\omega^{4}-\omega^{8}\left(c_{1}j_{0}^{2}-4c_{2}(\gamma
P)^{2}\right)},$ (19)
with $j_{0}\equiv g_{0}+h_{0}$. Nevertheless, this expression is still very
cumbersome and, below, we rely only on the numerically computed eigenvalue
$\lambda_{0}$.
We note that the cumulant expansion could be extended to fourth order (the
process being Gaussian, the third-order terms vanish), but the resulting terms
are very involved and do not clarify the behavior of gain at large $\omega$,
where the method breaks down (see below).
### III.3 Functional approach
An alternative approach follows [37, 33] and generalizes the use of Furustu-
Novikov-Shapiro-Loginov formulas, on which the treatment of white noise in
Ref. [14] is based.
Let us consider one of the second moments $X_{i}$, $i=1,2,3$. They are
functionals of $\delta\\!g$. Since $\delta\\!g\propto\chi$, its
autocorrelation function has the form of Eq. (4). According to Ref. [31], we
have
$\langle\delta\\!g\frac{\mathrm{d}X_{i}}{\mathrm{d}z}\rangle=\left(\frac{\mathrm{d}}{\mathrm{d}z}+B\right)\langle\delta\\!gX_{i}\rangle.$
(20)
Two steps are needed to write an averaged system: (i) average directly Eq.
(16); (ii) multiply each row by $\delta\\!g$ and average. We introduce three
new variables, $X_{3+i}\equiv\delta\\!gX_{i}$ and, in order to truncate the
system, we assume
$\langle\delta\\!g^{2}X_{i}\rangle=\sigma_{\delta\\!g}^{2}\langle
X_{i}\rangle$, where $\sigma_{\delta\\!g}^{2}\equiv
N_{0}B\frac{\omega^{4}}{16}$ is the variance of the process. This last
assumption is rigorously valid only for a dichotomous process [33] and in
general an infinite hierarchy of equations is obtained for a Gaussian one, see
[31].
We obtain a 6th-order system of ODEs
$\frac{\mathrm{d}}{\mathrm{d}z}\langle{X}\rangle=\begin{bmatrix}0&0&-2g_{0}&0&0&-2\\\
0&0&2h_{0}&0&0&2\\\ h_{0}&-g_{0}&0&1&-1&0\\\
0&0&-2\sigma_{\delta\\!g}^{2}&-B&0&-2g_{0}\\\
0&0&2\sigma_{\delta\\!g}^{2}&0&-B&2h_{0}\\\
\sigma_{\delta\\!g}^{2}&-\sigma_{\delta\\!g}^{2}&0&h_{0}&-g_{0}&-B\\\
\end{bmatrix}\langle{X}\rangle.$ (21)
The matrix in Eq. (21) has six eigenvalues. We observse numerically that one
is real and negative, one ($\lambda_{0}$) is real and positive, the last four
are two pairs of complex conjugate values with negative real part. The MI gain
is defined as above, $G_{2}(\omega)\equiv\frac{\lambda_{0}}{2}$. By numerical
inspection, we notice that the system in Eq. (21) generally gives different
eigenvalues with respect to Eq. (18). We stress that, in contrast to Sec.
III.2, the accuracy of the functional approach does not require any condition
on $\varepsilon$. However, in the white noise limit of $B\to\infty$, the
system reduces to three independent variables [33] and we obtain the Eqs.
(26-30) of [14], which as expected coincide with the cumulant expansion
results.
In the next subsection, we will show what are the limits of validity of the
two approximations and which fits better to numerical results.
### III.4 Results
Figure 1: MI gain as a function of detuning $\omega$ for a LP random
dispersion. Comparison of numerical values for a Gaussian (blue solid lines)
and dichotomous process (green dotted lines) obtained from Eq. (2) with the
estimates provided by Eq. (18) (red dashed lines), by Eq. (21) (dash-dotted
purple lines). We include also the gain corresponding to the white process
(yellow dotted line with pluses). In panel (a) $N_{0}=0.005$ and $B=4\pi$, so
that $\varepsilon<0.5$ over the considered $\omega$ range, (b) $N_{0}=0.4$ and
$B=4\pi$, so that $\varepsilon\approx 1$ around the maximum gain. The inset in
panel (a) shows a single realization of the two processes.
In order to validate our theoretical analysis, we resort to solving Eq. (2) by
generating a large number, $N$, of realizations of $\chi$ and studying the
moments of the resulting sample of solutions. The solution of Eq. (2) is known
exactly for piecewise constant $\beta_{2}(z)$ in terms of transfer matrices
[34]. For the Gaussian process, we choose the length $L$ of the domain
sufficiently large to ensure that the sampling rate in the spatial frequency,
i.e., $\Delta\\!\kappa=\frac{2\pi}{L}$, fairly represents the PSD of the
process, namely $B/\Delta\\!\kappa\ll 1$. An array of identically distributed
Gaussian random variables of zero mean and variance
$\sigma^{2}=\frac{N_{0}}{2\Delta\\!z}$, with $\Delta\\!z$ the sampling
distance in $z$ (over which the dispersion is assumed constant), is generated
numerically. This array is transformed to the $\kappa$ domain by means of FFT,
then multiplied by the Lorentzian filter of Eq. (6) and transformed back to
space domain. A realization of the dichotomous process is instead obtained in
the $z$ domain, by randomly generating the switching distances (where $\chi$
changes sign) from a randomly generated exponential distribution of mean
$2/B$. Between two switching points the GVD is assumed constant.
We solve Eq. (2) for a given initial condition
$(x_{1}(0),x_{2}(0))^{\mathrm{T}}=(1,0)$, corresponding to balanced sidebands,
over each random sequence of constant dispersion segments, to obtain a set of
output vectors $(x_{1}(L),x_{2}(L))^{\mathrm{T}}$ from which we estimate the
MI gain. Let $P_{\mathrm{out}}\equiv x_{1}^{2}(L)+x_{2}^{2}(L)$. We then
compute the mean gain, defined as [21, 20]
$G\equiv\frac{1}{2L}\log\langle\frac{P_{\mathrm{out}}}{P_{\mathrm{in}}}\rangle.$
(22)
In all our results we take $\gamma P=\beta_{2}^{0}=1$. This is equivalent to
putting Eq. (1) in the standard adimensional form.
In Fig. 1 we consider two examples of MI gain curves. We notice that the
sidelobes exhibit a single maximum $G_{\mathrm{max}}$ at detuning
$\omega_{\mathrm{max}}$.
We consider two different values of $N_{0}$. For small values, the gain is
generally small, see Fig. 1(a), where $N_{0}=0.005$. We choose a long domain
$L=5000$ to prevent finite-size effects that may appear at small $\omega$, and
$\Delta\\!z=0.01$ for the Gaussian case. Satisfactory statistics are obtained
for $N=2000$.
In this case $\varepsilon\approx 0.5$ at $\omega=10$ (see the definition of
$\varepsilon$ in (13)), the cumulant expansion is thus expected to be valid.
The functional approach gives very similar results, apart from some deviations
in the large $\omega$ tails. The numerical data show that the two processes
provide the same trend, that matches almost perfectly with the analytical
estimates (the functional approach proves more accurate in the tails, as
expected). The behavior near $\omega=0$ is due to the above-mentioned finite-
size effects. We notice that the LP process provide a much narrower and
smaller gain than the white noise (dotted yellow line with pluses). In this
regime, the correlations of the stochastic process suppress the MI gain.
Then we study the case of an intermediate value $N_{0}=0.4$, where
$\varepsilon>1$. We choose $L=50$, $\Delta\\!z=0.01$, and $N=1000$. We show in
Fig. 1(b) how the numerical results compare to the analytical estimates. The
functional approach proves very accurate for the dichotomous process over the
full range of $\omega$. The cumulant expansion models the numerical data quite
well, except in the large $\omega$ tails. We notice that the white noise gain
is smaller than the LP gain lobes, particularly for the dichotomous case.
Thus, for large $N_{0}$, correlations can improve the MI gain.
Figure 2: Detuning at the gain maximum as a function of $B$. Numerical results
of Gaussian (blue crosses) and dichotomous (green pluses) process are compared
to Eq. (18) (red dashed line), Eq. (21) (purple dash-dotted line). We include
$\omega_{\mathrm{max}}$ for the white noise, too, as reference (yellow dotted
line). The dashed vertical lines highlight the value of $B$ used in Fig. 1.
(a) $N_{0}=0.005$, (b) $N_{0}=0.4$.
Figure 3: Maximum gain as a function of $B$. Same convention as in Fig. 2.
In Figs. 2-3 we show the effect of variations of $B$ on
$\omega_{\mathrm{max}}$ and $G_{\mathrm{max}}$, respectively; $L$,
$\Delta\\!z$, and $N$ are chosen to guarantee a satisfactory statistical
sample for each point.
For $N_{0}=0.005$ [Figs. 2(a) and 3(a)] the position as well as the value of
the maximum gain are, for both stochastic processes, in the same range of
values and are, within the residual oscillation margins, well approximated by
both methods. Both $\omega_{\mathrm{max}}$ and $G_{\mathrm{max}}$ are monotone
increasing functions of $B$. As $B\to+\infty$, they converge from below to the
corresponding values for a white noise process. The limiting values are not
shown, because they are well beyond the axis scales.
The behavior for $N_{0}=0.4$ is very different [Figs. 2(b) and 3(b)]. First,
for the dichotomous case, the numerical points are well approximated by the
eigenvalues of Eq. (21). The maximum MI gain exhibits a sharp peak, see Fig.
3(b). We notice in Fig. 2(b) that $\omega_{\mathrm{max}}$ diverges for this
value of $B$: this means that the gain lobe never decays.
To understand this phenomenon, we study the eigenvalues of Eq. (21) at
$\omega\to\infty$ and we notice that, for $\sigma_{\delta\\!g}=g_{0}$, i.e.,
$B=4\left(\beta_{2}^{0}\right)^{2}/N_{0}$, $G_{2}(\omega)$ converges, for
small $N_{0}$, to
$G_{\mathrm{max}}\approx\frac{1}{4}\left[-B+\sqrt{\frac{B^{2}+B\sqrt{B^{2}+64(\gamma
P)^{2}}}{2}}\right]$. Notice also that for this value $\sigma^{2}_{\chi}=1$,
i.e., in dimensional units the GVD switches from $0$ to $2\beta_{2}^{0}$. In
the present case, $B=10$ and $G_{\mathrm{max}}\approx 0.17$, as observed in
Fig. 3(b). For $N_{0}>1$, the reasoning is still valid: the value of $B$ for
which $G_{2}(\omega)$ does not decay at $\omega\to\infty$ is well predicted,
while its value is generally larger than the approximated $G_{\mathrm{max}}$
above. From a physical point of view, this case is very pathological because
higher-order dispersion effects should be included in Eq. (1).
For the Gaussian process, $\omega_{\mathrm{max}}$ and $G_{\mathrm{max}}$ stay
close to the white noise limit (yellow dotted line) even for small $B$ and are
satisfactorily described by the cumulant expansion for large $B$. As expected,
the cumulant expansion is not accurate for small $B$ and is very different
from the estimate of the functional approach.
We conclude that the cumulant expansion provides an accurate approximation for
$G_{\mathrm{max}}$ and $\omega_{\mathrm{max}}$ for the Gaussian process. We
recall that, on the contrary, this is not the case for the behavior of the MI
gain at large $\omega$, see Fig. 1(b).
In conclusion, the LP processes yield a very small gain for $\varepsilon\ll
1$, always less than white noise limit. For $\varepsilon\sim 1$,
$G_{\mathrm{max}}$ can yield an MI gain larger than the white noise gain, but
is about one order of magnitude smaller than the conventional MI gain in
homogeneous fibers with anomalous GVD.
## IV Narrowband random dispersion
Here we consider the modulated process $\xi$ with the autocorrelation function
in Eq. (9), for which a new degree of freedom, $\kappa_{0}$, is included. We
start directly from Eq. (16).
### IV.1 Direct cumulant expansion
The cumulant expansion gives the same result as Eq. (18) with
$\displaystyle c_{1}$
$\displaystyle=\frac{N_{0}}{2}\frac{B^{2}}{B^{2}+\kappa_{0}^{2}},$ (23)
$\displaystyle c_{2}$
$\displaystyle=\frac{N_{0}B^{2}}{4}\left[\frac{1}{B^{2}+(2k-\kappa_{0})^{2}}+\frac{1}{B^{2}+(2k+\kappa_{0})^{2}}\right],$
$\displaystyle c_{3}$
$\displaystyle=N_{0}B\kappa\frac{B^{2}-\kappa_{0}^{2}+4k^{2}}{\left[B^{2}+(2k-\kappa_{0})^{2}\right]\left[B^{2}+(2k+\kappa_{0})^{2}\right]}.$
We observe that, while $c_{1}$ is constant, $c_{2}$ and $c_{3}$ exhibit a
resonant lineshape behavior close to $2k=\kappa_{0}$, i.e., the 1st PR
condition: the former exhibits a maximum at
$4k^{2}\approx\kappa_{0}^{2}-\frac{B^{4}}{4\kappa_{0}^{2}}$, while the latter
crosses zero at $4k^{2}=\kappa_{0}^{2}-B^{2}$. We recall that the $m$-th order
PR condition is in general $2k=m\kappa_{0}$, which gives
$\omega_{\mathrm{PR},m}^{2}=\sqrt{4\left(\frac{\gamma
P}{\beta_{2}^{0}}\right)^{2}+\left(\frac{m\kappa_{0}}{\beta_{2}^{0}}\right)^{2}}-\frac{2\gamma
P}{\beta_{2}^{0}}.$ (24)
For $B\ll\kappa_{0}$, we can set $c_{1}\approx 0$. Then around the first
parametric resonance detuning, tedious but straightforward calculations show
that the instability gain is approximated by
$G_{2}(\omega)\approx\frac{1}{2}\frac{(2\gamma
P)^{2}\omega^{4}}{\kappa_{0}^{2}}c_{2},$ (25)
that attains its maximum approximately at $\omega_{\mathrm{PR,1}}$,
$\displaystyle G_{2}^{\mathrm{max}}\approx G_{2}(\omega_{\mathrm{PR,1}})$
$\displaystyle\approx\frac{N_{0}(\gamma
P)^{2}}{2}\frac{\omega_{\mathrm{PR,1}}^{4}}{\kappa_{0}^{2}}$ (26)
$\displaystyle=\frac{N_{0}(\gamma
P)^{2}}{\left(2\beta_{2}^{0}\right)^{2}}\frac{\omega_{\mathrm{PR,1}}^{2}}{\omega_{\mathrm{PR,1}}^{2}+\frac{4\gamma
P}{\beta_{2}^{0}}}.$
We remark also that Eq. (25) is composed of a Lorentzian factor depending on
the process PSD ($c_{2}$) and a factor independent of the process. Moreover,
it turns out that the maximum MI gain, Eq. (26), coincides, for small $\omega$
and a given $N_{0}$ with the white noise MI gain evaluated at
$\omega=\omega_{{\rm PR},1}$, as hinted at in Ref. [32].
Notice that in Eq. (26), the MI gain is proportional to
$\frac{N_{0}\omega_{\mathrm{PR,1}}^{4}}{\kappa_{0}^{2}}$, while the PR-MI
scales like $\theta\frac{\omega_{\mathrm{PR,1}}^{2}}{\kappa_{0}}$ [12], with
$\theta$ the (constant) amplitude of the periodic variation. As in the low-
pass case, according to Eq. (26) the maximal MI gain depends mainly on
$N_{0}$. For $B\to 0$, the analogy to PR would suggest, instead, a dependence
on the amplitude of the fluctuations in real space, i.e., $N_{0}B$, as
discussed also in Ref. [32]. Below, we clarify this ostensible inconsistency.
### IV.2 Near-resonance reduction
In analogy to Ref. [32], a significant simplification of Eq. (2) can be
obtained by the conventional averaging method used for PR [38].
In order to average Eq. (2), we let
$\displaystyle x_{1}(z)$
$\displaystyle=y_{1}(z)\cos\left(\frac{\kappa_{0}}{2}z\right)+y_{2}(z)\sin\left(\frac{\kappa_{0}}{2}z\right)$
(27) $\displaystyle x_{2}(z)$
$\displaystyle=\frac{\kappa_{0}}{2g_{0}}\left[y_{1}(z)\sin\left(\frac{\kappa_{0}}{2}z\right)-y_{2}(z)\cos\left(\frac{\kappa_{0}}{2}z\right)\right],$
where $x_{2}$ is written assuming $g=g_{0}$ constant, in the spirit of the
variation of constants.
By averaging out oscillating terms, noticing that
$\frac{\delta\\!g}{g_{0}}=\xi$, and employing the phase-quadrature
representation of $\xi$, see Eq. (7), we obtain
$\displaystyle\kappa_{0}\dot{y}_{1}$
$\displaystyle=\Delta^{2}y_{2}+\Gamma\psi_{1}y_{2}-\Gamma\psi_{2}y_{1}$ (28)
$\displaystyle\kappa_{0}\dot{y}_{2}$
$\displaystyle=-\Delta^{2}y_{1}+\Gamma\psi_{1}y_{1}+\Gamma\psi_{2}y_{2},$
with $\Gamma\equiv\frac{1}{2}\left(\frac{\kappa_{0}^{2}}{4}-g_{0}^{2}\right)$
and $\Delta^{2}\equiv k^{2}-\frac{\kappa_{0}^{2}}{4}$ quantifies the detuning
from the PR condition; close to resonance,
$\Delta^{2}=(k+\kappa_{0}/2)(k-\kappa_{0}/2)\approx\kappa_{0}(k-\kappa_{0}/2)$.
Notice that the PR-MI gain of the first PR tongue is obtained from Eq. (28) by
putting $\psi_{1}=\psi_{2}=\theta$. It reads
$G_{\mathrm{PR,1}}=\frac{\sqrt{2(\Gamma\theta)^{2}-\Delta^{2}}}{\kappa_{0}}.$
(29)
Starting from Eq. (28), in this section we only study the evolution of second
moments. We introduce $Y_{1}\equiv y_{1}^{2}$, $Y_{2}\equiv y_{2}^{2}$, and
$Y_{3}\equiv y_{1}y_{2}$. It is easy to verify that
${Y}\equiv(Y_{1},Y_{2},Y_{3})^{\mathrm{T}}$ obeys
$\kappa_{0}\frac{\mathrm{d}}{\mathrm{d}z}{Y}=\begin{bmatrix}-2\Gamma\psi_{2}&0&2\Delta^{2}+2\Gamma\psi_{1}\\\
0&2\Gamma\psi_{2}&-2\Delta^{2}+2\Gamma\psi_{1}\\\
-\Delta^{2}+\Gamma\psi_{1}&\Delta^{2}+\Gamma\psi_{1}&0\end{bmatrix}{Y},$ (30)
where the parameters are the same as those used throughout this section.
Two random processes appear in Eq. (30) and both the cumulant expansion and
functional approach need generalizing accordingly. While we present the former
in App. A, because it provides very similar results to the direct cumulant
expansion of the previous subsection, the latter is reported below.
### IV.3 Functional approach near resonance
Following [39], we generalize the functional approach. We define
$Y_{3+i}\equiv\psi_{1}Y_{i}$, $Y_{6+i}\equiv\psi_{2}Y_{i}$, and
$Y_{9+i}\equiv\psi_{1}\psi_{2}Y_{i}$, $i=1,2,3$. We perform four different
averaging steps: (i) average directly Eq. (30), (ii) multiply each row of Eq.
(30) by $\psi_{1}$ and average, (iii) multiply by $\psi_{2}$ and average; (iv)
multiply by $\psi_{1}\psi_{2}$ and average. We employ the formula of
differentiation in Eq. (20) and its generalization
$\begin{gathered}\langle\psi_{1}\psi_{2}\frac{\mathrm{d}Y_{i}}{\mathrm{d}t}\rangle=\left(\frac{\mathrm{d}}{\mathrm{d}t}+2B\right)\langle\psi_{1}\psi_{2}Y_{i}\rangle.\end{gathered}\
$ (31)
If we assume as above that we can factor the variance out if the same process
occurs twice in an average, we obtain
$\kappa_{0}\frac{\mathrm{d}\langle{Y}\rangle}{\mathrm{d}z}=\left[\begin{array}[]{c|c|c|c}A_{2}&\Gamma
C_{2}^{\prime}&\Gamma C_{2}^{\prime\prime}&\mathbf{0}\\\
\hline\cr\sigma^{2}_{\xi}\Gamma
C_{2}^{\prime}&A_{2}-B\mathbf{I}&\mathbf{0}&\Gamma C_{2}^{\prime\prime}\\\
\hline\cr\sigma^{2}_{\xi}\Gamma
C_{2}^{\prime\prime}&\mathbf{0}&A_{2}-B\mathbf{I}&\Gamma C_{2}^{\prime}\\\
\hline\cr\mathbf{0}&\sigma^{2}_{\xi}\Gamma
C_{2}^{\prime\prime}&\sigma^{2}_{\xi}\Gamma
C_{2}^{\prime}&A_{2}-2B\mathbf{I}\end{array}\right]\langle{Y}\rangle,$ (32)
with
$\begin{gathered}{A_{2}}=\begin{bmatrix}0&0&2\Delta^{2}\\\ 0&0&-2\Delta^{2}\\\
-\Delta^{2}&\Delta^{2}&0\end{bmatrix},\\\
{C_{2}^{\prime}}=\begin{bmatrix}0&0&2\\\ 0&0&2\\\
1&1&0\end{bmatrix},\;{C_{2}^{\prime\prime}}=\begin{bmatrix}-2&0&0\\\ 0&2&0\\\
0&0&0\end{bmatrix},\end{gathered}$ (33)
and denoting $\mathbf{0}$ and $\mathbf{I}$ the null and identiy matrix,
respectively.
We compute numerically the eigenvalues of the matrix in Eq. (32) and look for
the dominant one, $\lambda^{*}$. The maximum MI gain, for $\Delta^{2}=0$, is
found analytically as
$(G_{2}^{\mathrm{PR}})^{\mathrm{max}}=\frac{1}{4\kappa_{0}}\left[\sqrt{B^{2}+4N_{0}B\omega_{\mathrm{PR,1}}^{4}}-B\right].$
(34)
We remark that the dependence on $\sigma_{\xi}$ is different from Eq. (26).
For small $B$, the gain is no longer proportional to
$\frac{\omega_{\mathrm{PR,1}}^{4}N_{0}}{\kappa_{0}^{2}}$, but to
$\frac{\omega_{\mathrm{PR,1}}^{2}\sqrt{N_{0}B}}{\kappa_{0}}$, as we would
formally obtain in the conventional periodic dispersion case, once we replace
the amplitude of parametric oscillation $\theta$ with $\sqrt{2}\sigma_{\xi}$.
Thus a very different result is found when compared to the cumulant expansions
above. We now assess which approximation works better by comparing them to
numerical solutions of Eq. (2).
### IV.4 Results
Figure 4: MI gain as a function of detuning $\omega$ for a BP random
dispersion, $\kappa_{0}=2\pi$. Comparison of numerical values for a Gaussian
(blue solid lines) and dichotomous process (green dotted lines) obtained from
Eq. (2) with the estimates provided by Eqs. (18) and (23) (red dashed lines),
and by Eq. (32) (dash-dotted purple lines). In panel (a) $N_{0}=0.005$ and
$B=\pi/4$, so that $\varepsilon<0.1$ around the PR resonance, (b) $N_{0}=3.2$
and $B=\pi/32$, so that $\varepsilon\gg 1$ around the $\omega_{\mathrm{max}}$.
In panel (b), we include the PR-MI gain of Eq. (29) with
$\theta=\sqrt{2}\sigma_{\xi}=\sqrt{\pi/5}$ (yellow dotted line with circles).
The insets show the numerical results for a Gaussian process on a larger
$\omega$ range.
In order to generate a realization of the process $\xi$ with Gaussian
distribution (resp. dichotomous), we employ the same approach in the spectral
(resp. spatial) domain as above, generate two independent LP processes and
modulate them according to Eq. (7). The numerical domain is obviously
discretized in both cases: for the Gaussian process a lower limit on $L$ is
required as in Sec. III.4, the dichotomous process requires a short
$\Delta\\!z$ to avoid spurious correlations. We do not forget to correctly
sample the period $\Lambda_{0}$ of the process spatial oscillations.
While in the previous section we were interested in the limit $B\to\infty$ to
contrast the LP processes to white noise, here we aim at understanding the
opposite limit, $B\to 0$, with fixed and finite $N_{0}B$. We expect that the
gain associated to the stochastic fluctuations converges to the one of the
periodically modulated fiber in this limit.
First, we consider a small perturbation, $N_{0}=0.005$, with a intermediate
bandwidth $B=\pi/4$ around $\kappa_{0}=2\pi$. We take $L=500$,
$\Delta\\!z=0.01$; a consistent statistical sample is collected for $N=2000$.
In Fig. 4(a) we compare the numerical mean gain $G(\omega)$ with the
analytical estimates obtained above. For such a small perturbation only a
single MI peak can be observed (see the inset for a larger detuning range). We
notice that the MI gain is centered about $\omega_{\mathrm{PR,1}}$ with gain
$G_{\mathrm{max}}$ and width $\Delta\\!\omega$. The Gausssian and dichotomous
processes give two almost identical results. The cumulant expansion provides a
very good approximation, because $\varepsilon\approx 0.1$. The near-resonance
functional approach proves instead very imprecise, as far as both
$G_{\mathrm{max}}$ and $\Delta\\!\omega$ are concerned. We explain this as
follows: it is easy to verify that the period of the modulation is comparable
to correlation length, $\Lambda_{0}\approx\zeta_{c}$. Thus, we apply
successively two distinct averaging procedures (near-resonant expansion and
functional approach) upon two perturbations occurring at the same scale as if
they were independent. This is a sure recipe for failure.
Moreover, we are not showing here the result of Eq. (37) that coincide with
the direct cumulant expansion around $\omega_{\mathrm{PR,1}}$, but is skewed
towards $\omega\to 0$, contrary to the numerical results.
As a second case, we consider a large $N_{0}=3.2$ and a small bandwidth
$B=\pi/32$. All the other parameters are the same as the previous. In Fig.
4(b) we observe, as above, that the MI lobe occurs around
$\omega_{\mathrm{PR,1}}$ and the Gaussian and dichotomous processes provide
two very close results. For such a large perturbation, several MI sidelobes
can be observed, in analogy to PR-MI, see the inset of Fig. 4(b). Now
$\varepsilon\approx 6.3\gg 1$; we thus expect that the cumulant expansion
fails to correctly describe the numerical results: indeed it overestimates the
peak MI while it underestimates its width (red dashed line). It captures
approximately the behavior of the MI gain in the tails (both left-hand and
right-hand sides). The functional method provides instead a good approximation
(purple dash-dotted line). For these parameters, Eq. (32) provides a result
very close to the conventional PR gain (yellow dotted line with circles) apart
from the tails. We can thus state that for $\zeta_{c}\gg\Lambda_{0}$, the
proposed functional approach gives a good estimate of the numerically
estimated mean gain, because the two independent approximations are performed
in the correct order on the two different scales, i.e., for $B\ll\kappa_{0}$,
the BP process can be considered a small perturbation to the PR-MI effect.
Figure 5: Maximum gain $G_{\mathrm{max}}$ as a function of $B$ at constant
$N_{0}B$. Same convention as in Fig. 2, dichotomous process excluded. We take
a constant $\kappa_{0}=2\pi$ and (a) $N_{0}B=0.0039$, (b)
$N_{0}B=\pi/10=0.314$. Notice the ordinate axis is in logarithmic scale in
panel (a). The dashed vertical lines highlight the values of $B$ used in Fig.
4(a)[(b)], respectively.
We show in Fig. 5 the effect of variations of $B$ on $G_{\mathrm{max}}$. We
keep $\kappa_{0}$ and $N_{0}B$ constant; $L$, $\Delta\\!z$, and $N$ are chosen
to guarantee a significant statistical sample for each value. Only the
Gaussian process is considered.
For small $N_{0}B=0.0039$ the MI gain follows very well the cumulant
expansion: $G_{\mathrm{max}}$ grows for $B\to 0$ (notice the logarithmic
scale). The functional approach always overestimate it. For larger
$N_{0}B=\pi/10$ we still observe a decreasing trend of $G_{\mathrm{max}}$. It
is apparent that the cumulant expansion is valid only for large $B>\pi/2$ and
completely loses its validity below $B=\pi/4$. The functional approach works
well in the PR limit, i.e., for $B<\pi/16$. Indeed, $\zeta_{c}$ depends
inversely on $B$ and the approximation is expected to become more faithful.
The residual discrepancies may depend on a systematic error of the averaging
procedure or on numerical inaccuracies. In brief, we show the transition from
the stochastic regime, where the MI gain scales with $N_{0}$, to the
parametric regime where it depends on $\sigma_{\xi}$, i.e. the fluctuation
amplitude.
Figure 6: MI gain curves $G(\omega)$ and their maxima $G_{\mathrm{max}}$
obtained for different values of
$\kappa_{0}\in\\{\pi,2\pi,4\pi,8\pi,16\pi,32\pi,64\pi\\}$. Comparisons of
numerical results [blue crosses and blue solid lines] with the peak of Eq.
(26) (red dashed line), of Eq. (34) [purple dash-dotted line, not shown in
panel (a), out of scale]. We include also $G_{2}(\omega)$ for the white noise
limit at the corresponding (yellow dotted line). The dashed vertical lines
highlight the value of $\kappa_{0}$ used in Figs. 4 and 5. (a) $N_{0}=0.005$,
$B=\pi/4$; (b) $N_{0}=3.2$, $B=\pi/32$.
Finally, we report in Fig. 6 different MI gain curves obtained from numerical
simultations of the Gaussian process by varying
$\kappa_{0}\in\\{\pi,2\pi,4\pi,8\pi,16\pi,32\pi,64\pi\\}$. The simulation
parameters are chosen for each value to ensure that the sample is statically
significant.
In Fig. 6(a), we choose $N_{0}=0.005$ and $B=\pi/4$ and show that the MI lobes
have a Lorentzian shape dominated by a simple envelope, as predicted by Eq.
(25). We explicitly show $G_{2}^{\mathrm{max}}$ obtained from Eq. (26) for the
values of $\omega_{\mathrm{PR,1}}$ corresponding to a continuous set of
$\kappa_{0}$ (red dashed line) and compare it to the values of
$G_{\mathrm{max}}$ (blue crosses). They match very well over the full range of
$\omega$. The result of the functional approach, Eq. (34), is not shown here,
because it always overestimates the MI gain. We also include the white noise
gain for the given $N_{0}$ (yellow dotted line). For such a small PSD value,
it exhibits a monotone growth up to $\omega\approx 9$, compare to Fig. 1(a).
The expression in Eq. (26) coincides with it up to this point and tends to
$N_{0}/2$ for $\kappa_{0},\omega_{\mathrm{PR,1}}\to\infty$. This strengthens
the idea that the MI gain of stochastic origin is ruled predominantly by
$N_{0}$, at least for small values.
If, instead, we choose $N_{0}=3.2$ and $B=\pi/32$, we observe, see Fig. 6(b),
that the MI lobes are sharper than a Lorentzian. Their maxima (blue crosses)
are well approximated by the functional approximation (purple dash-dotted
line), while the cumulant expansion (red dashed line) always overestimates the
MI gain by more than 3 times. For such a small $B$, Eq. (34) effectively
coincides with Eq. (29) (not shown) as expected. As in the previous case, the
gain monotonously increases with $\kappa_{0}$, consistently with the PR nature
of the phenomenon, see Eq. (29). If we consider the white noise, we see that
the MI gain exhibits a narrow lobe, that attains a maximum for $\omega\approx
1.25$ and coincides with (26) only for small detuning, as expected.
As $\kappa_{0}$ is increased, the modulation period of the process becomes
smaller, while $\zeta_{c}$ is kept constant; thus, the near-resonant
approximation performs better for large $\kappa_{0}$ and the discrepancies
observed 6(b) are probably due to a systematic error or residual numerical
inaccuracies.
## V Conclusions
We discuss the modulational instability in nonlinear optical fibers in which
the group-velocity dispersion is randomly modulated. In contrast to the
exactly solvable case of white noise or of random kicks, we consider the case
of stochastic processes with exponentially decaying autocorrelation function.
This is equivalent to a Lorentzian-shaped power spectral-density, i.e., the
process is colored. Two families of colored processes are studied: low-pass
and band-pass. For each family, we consider Gaussian and dichotomous
processes.
The distinction between the two stochastic processes turns out to be important
in the LP and marginal in the BP case.
While for very small perturbation the LP yields very small MI gain, for larger
power spectral densities it can yield MI sidelobes larger than the white
noise, sitting in the same range of detuning and exhibiting larger gain in the
tails. For small bandwidth, the gain disappears, while in the opposite limit
it converges to the white noise limit. The variance demanded for obtaining a
measurable MI gain is large, though, and the dichotomous process looks more
promising in view on an experimental characterization of the phenomenon,
because values of bandwidth (correlation length) exist where the gain is quite
larger than the white noise limit.
As far as a BP process is concerned, if the perturbation is large enough, we
may observe several MI sidelobes sitting around PR detunings. We focus on the
dominant first peak: it converges to the PR sidelobe for small bandwidth and
is generally broader and smaller for small correlation lengths.
We compare our numerical results to different analytical approximations, based
on the cumulant expansion (as formalized by van Kampen) or the functional
(Furutsu-Novikov-Loginov-Shapiro) formulas.
While the former is reliable only for small perturbations and small detuning
and provides some qualitative estimates elsewhere, the latter provides good
results for the dichotomous processes, for which the closure of the system is
rigorously obtained.
For both families of correlation functions, the functional method emerges as
more reliable and allows us to describe the transition from parametric to
stochastic resonances in the BP case. Notwithstanding, the cumulant expansion
provides good estimates of the tails of resonant peaks even for relatively
small bandwidth values (large correlation length), beyond the expected
validity range of the approximation.
Our results pave the way for tailoring MI gain sidebands in optical fibers by
means of stochastic GVD fluctuations and suggest the regimes to achieve that.
Such fluctuations can be implemented by continuous or discrete variations of
fiber specifications.
## Acknowledgments
The work was supported in part the by the French government through the
Programme Investissement d’Avenir with the Labex CEMPI (Grant
ANR-11-LABX-0007-01) and the I-SITE ULNE (Grant ANR-16-IDEX-0004 ULNE,
projects VERIFICO) managed by the Agence Nationale de la Recherche. The work
was also supported by the Nord-Pas-de-Calais Regional Council and the European
Regional Development Fund through the Contrat de Projets État-Région (CPER)
and IRCICA.
## Appendix A Cumulant expansion near resonance
In this section we will apply the cumulant expansion to Eq. (30).
The cumulant expansion to second order is built by decomposing Eq. (30) as
$\kappa_{0}\dot{Y}=\left[A_{2}+\Gamma\psi_{1}C_{2}^{\prime}+\Gamma\psi_{2}C_{2}^{\prime\prime}\right]Y$,
with $A_{2}$, $C_{2}$, and $C_{2}^{\prime}$ as in Eq. (33).
As $\psi_{1,2}$ are mutually independent, the expansion is obtained by
computing $K_{2}$ separately for $C_{2}^{\prime}$ and $C_{2}^{\prime\prime}$,
according to Eq. (17), and adding them up.
We obtain
$\kappa_{0}\frac{\mathrm{d}\langle{Y}\rangle}{\mathrm{d}z}=\begin{bmatrix}6d_{2}\Gamma^{2}&2d_{2}\beta^{2}&2\Delta^{2}-4\Gamma^{2}d_{3}\\\
2d_{2}\beta^{2}&6d_{2}\beta^{2}&-2\Delta^{2}+4\Gamma^{2}d_{3}\\\
2\Gamma^{2}d_{3}-\Delta^{2}&-2\Gamma^{2}d_{3}+\Delta^{2}&4\Gamma^{2}d_{2}\end{bmatrix}\langle{Y}\rangle,$
(35)
with
$\displaystyle d_{2}$
$\displaystyle\equiv\frac{1}{\kappa_{0}}\int\limits_{0}^{\infty}\mathrm{d}\zeta
R_{\psi_{1}}(\zeta)\cos\frac{2\Delta^{2}}{\kappa_{0}}\zeta=\frac{N_{0}B^{2}}{2\kappa_{0}}\frac{1}{B^{2}+\frac{4\Delta^{4}}{\kappa_{0}^{2}}}$
(36) $\displaystyle d_{3}$
$\displaystyle\equiv\frac{1}{\kappa_{0}}\int\limits_{0}^{\infty}\mathrm{d}\zeta
R_{\psi_{1}}(\zeta)\sin\frac{2\Delta^{2}}{\kappa_{0}}\zeta=\frac{N_{0}B}{2\kappa_{0}}\frac{\frac{2\Delta^{2}}{\kappa_{0}}}{B^{2}+\frac{4\Delta^{4}}{\kappa_{0}^{2}}}.$
We note that there is here no counterpart to $c_{1}$ and near PR
$\kappa_{0}d_{2}\approx 2c_{2}$.
The dominant eigenvalue of the matrix in Eq. (35) is exactly
$\lambda^{*}=8d_{2}\Gamma^{2}$, so the instability gain is
$G_{2}^{\mathrm{PR}}=\frac{8d_{2}\Gamma^{2}}{\kappa_{0}},$ (37)
which attains a maximum
$(G_{2}^{\mathrm{PR}})^{\mathrm{max}}\approx\frac{N_{0}\omega_{\mathrm{PR,1}}^{4}}{2\kappa_{0}^{2}}$,
identical to what found above in Eq. (26).
## References
* Zakharov and Ostrovsky [2009] V. Zakharov and L. Ostrovsky, Modulation instability: The beginning, Physica D 238, 540 (2009).
* Benjamin and Feir [1967] T. B. Benjamin and J. E. Feir, The disintegration of wave trains on deep water Part 1. Theory, Journal of Fluid Mechanics 27, 417 (1967).
* Zakharov [1968] V. E. Zakharov, Stability of periodic waves of finite amplitude on the surface of a deep fluid, J. Appl. Mech. Tech. Phys. 9, 86 (1968).
* Bespalov and Talanov [1966] V. Bespalov and V. Talanov, Filamentary Structure of Light Beams in Nonlinear Liquids, Soviet Journal of Experimental and Theoretical Physics Letters 3, 307 (1966).
* H. Ichikawa _et al._ [1973] Y. H. Ichikawa, T. Suzuki, and T. Taniuti, Modulation Instability of Electron Plasma Wave, J. Phys. Soc. Jpn 34, 1089 (1973).
* Tai _et al._ [1986] K. Tai, A. Hasegawa, and A. Tomita, Observation of modulational instability in optical fibers, Phys. Rev. Lett. 56, 135 (1986).
* Berkhoer and Zakharov [1970] A. L. Berkhoer and V. E. Zakharov, Self Excitation of Waves with Different Polarizations in Nonlinear Media, Soviet Journal of Experimental and Theoretical Physics 31, 486 (1970).
* Stolen _et al._ [1974] R. H. Stolen, J. E. Bjorkholm, and A. Ashkin, Phase‐matched three‐wave mixing in silica fiber optical waveguides, Appl. Phys. Lett. 24, 308 (1974).
* Cavalcanti _et al._ [1991] S. B. Cavalcanti, J. C. Cressoni, H. R. da Cruz, and A. S. Gouveia-Neto, Modulation instability in the region of minimum group-velocity dispersion of single-mode optical fibers via an extended nonlinear Schrödinger equation, Phys. Rev. A 43, 6162 (1991).
* Smith and Doran [1996] N. J. Smith and N. Doran, Modulational instabilities in fibers with periodic dispersion management., Opt. Lett. 21, 570 (1996).
* Droques _et al._ [2012] M. Droques, A. Kudlinski, G. Bouwmans, G. Martinelli, and A. Mussot, Experimental demonstration of modulation instability in an optical fiber with a periodic dispersion landscape., Opt. Lett. 37, 4832 (2012).
* Armaroli and Biancalana [2012] A. Armaroli and F. Biancalana, Tunable modulational instability sidebands via parametric resonance in periodically tapered optical fibers, Opt. Expr. 20, 25096 (2012).
* Mussot _et al._ [2018] A. Mussot, M. Conforti, S. Trillo, F. Copie, and A. Kudlinski, Modulation instability in dispersion oscillating fibers, Adv. Opt. Phot. 10, 1 (2018).
* Abdullaev _et al._ [1996] F. K. Abdullaev, S. A. Darmanyan, A. Kobyakov, and F. Lederer, Modulational instability in optical fibers with variable dispersion, Phys. Lett. A 220, 213 (1996).
* Abdullaev _et al._ [1997] F. K. Abdullaev, S. A. Darmanyan, S. Bischoff, and M. P. Sørensen, Modulational instability of electromagnetic waves in media with varying nonlinearity, J. Opt. Soc. Am. B 14, 27 (1997).
* Abdullaev and Garnier [1999] F. K. Abdullaev and J. Garnier, Modulational instability of electromagnetic waves in birefringent fibers with periodic and random dispersion, Physical Review E 60, 1042 (1999).
* Garnier and Abdullaev [2000] J. Garnier and F. K. Abdullaev, Modulational instability induced by randomly varying coefficients for the nonlinear Schrödinger equation, Phys. D 145, 65 (2000).
* Chertkov _et al._ [2001] M. Chertkov, I. Gabitov, and J. Moeser, Pulse confinement in optical fibers with random dispersion, Proceedings of the National Academy of Sciences 98, 14208 (2001).
* Karlsson [1998] M. Karlsson, Four-wave mixing in fibers with randomly varying zero-dispersion wavelength, J. Opt. Soc. Am. B 15, 2269 (1998).
* Farahmand and de Sterke [2004] M. Farahmand and M. de Sterke, Parametric amplification in presence of dispersion fluctuations, Opt. Expr. 12, 136 (2004).
* Dujardin _et al._ [2021] G. Dujardin, A. Armaroli, S. R. Nodari, A. Mussot, A. Kudlinski, S. Trillo, M. Conforti, and S. De Bievre, Modulational instability in optical fibers with randomly kicked normal dispersion, Phys. Rev. A 103, 1 (2021).
* Bourret [1962] R. C. Bourret, Propagation of Randomly Perturbed Fields, Canadian Journal of Physics 40, 782 (1962).
* Van Kampen [1976] N. G. Van Kampen, Stochastic differential equations, Phys. Rep. 24, 171 (1976).
* Van Kampen [1974a] N. G. Van Kampen, A cumulant expansion for stochastic linear differential equations. I, Physica 74, 215 (1974a).
* Van Kampen [1974b] N. G. Van Kampen, A cumulant expansion for stochastic linear differential equations. II, Physica 74, 239 (1974b).
* Terwiel [1974] R. H. Terwiel, Projection operator method applied to stochastic linear differential equations, Physica 74, 248 (1974).
* Fox [1986] R. F. Fox, Uniform convergence to an effective Fokker-Planck equation for weakly colored noise, Phys. Rev. A 34, 4525 (1986).
* Kampen [2007] V. N. Kampen, _Stochastic Processes in Physics and Chemistry_ (Elsevier, 2007).
* Furutsu [1963] K. Furutsu, On the statistical theory of electromagnetic waves in a fluctuating medium (I), J. Res. Nat. Bur. Stand., Sect. D 67D, 303 (1963).
* Novikov [1965] E. A. Novikov, Functionals and the Random-force Method in Turbulence Theory, Sov. Phys.—JETP 20, 1290 (1965).
* Shapiro and Loginov [1978] V. E. Shapiro and V. M. Loginov, ”Formulae of differentiation” and their use for solving stochastic equations, Physica A 91, 563 (1978).
* Zhang _et al._ [1992] W. Zhang, J. Casademunt, and J. Viñals, Study of the parametric oscillator driven by narrow-band noise to model the response of a fluid surface to time-dependent accelerations, Phys. Fluids A 5, 3147 (1992).
* Gitterman [2005a] M. Gitterman, _Noisy Oscillator, The: Random Mass, Frequency, Damping_ (World Scientific Publishing Co. Pte. Ltd., 2005).
* Agrawal [2012] G. P. Agrawal, _Nonlinear Fiber Optics_ , fifth edit ed. (Academic Press, Oxford, 2012) p. 648.
* Stratonovich [1963] R. L. Stratonovich, _Topics in the theory of random noise (Vol. 1)_ , 1st ed. (Gordon and Breach Science Publishers, New York, 1963).
* Dubkov and Muzychuk [1977] A. A. Dubkov and O. V. Muzychuk, Analysis of higher approximations of Dyson’s equation for the mean value of the Green function, Radiophys. Quant. Electr. 20, 623 (1977).
* Gitterman [2005b] M. Gitterman, Classical harmonic oscillator with multiplicative noise, Physica A 352, 309 (2005b).
* Landau [1976] L. D. Landau, _Mechanics Course of theoretical physics (Volume 1)_ , 3rd ed. (1976) p. 170.
* Burov and Gitterman [2016] S. Burov and M. Gitterman, Noisy oscillator: Random mass and random damping, Phys. Rev. E 94, 1 (2016).
|
20cm(1cm,1cm) Accepted by ACM Transactions on Software Engineering and
Methodology (TOSEM)
# Hierarchical Distribution-Aware Testing of Deep Learning
Wei Huang Purple Mountain LaboratoriesChina,University of LiverpoolU.K.
<EMAIL_ADDRESS>, Xingyu Zhao WMG, University of WarwickU.K.
<EMAIL_ADDRESS>, Alec Banks Defence Science and Technology
LaboratoryU.K<EMAIL_ADDRESS>, Victoria Cox Defence Science and
Technology LaboratoryU.K<EMAIL_ADDRESS>and Xiaowei Huang University of
LiverpoolU.K<EMAIL_ADDRESS>
(2023)
###### Abstract.
With its growing use in safety/security-critical applications, Deep Learning
(DL) has raised increasing concerns regarding its dependability. In
particular, DL has a notorious problem of lacking robustness. Input added with
adversarial perturbations, i.e. Adversarial Examples (AEs) are easily mis-
predicted by the DL model. Despite recent efforts made in detecting AEs via
state-of-the-art attack and testing methods, they are normally input
distribution agnostic and/or disregard the perceptual quality of adversarial
perturbations. Consequently, the detected AEs are irrelevant inputs in the
application context or unrealistic that can be easily noticed by humans. This
may lead to a limited effect on improving the DL model’s dependability, as the
testing budget is likely to be wasted on detecting AEs that are encountered
very rarely in its real-life operations.
In this paper, we propose a new robustness testing approach for detecting AEs
that considers both the feature level distribution and the pixel level
distribution, capturing the perceptual quality of adversarial perturbations.
The two considerations are encoded by a novel hierarchical mechanism. First,
we select test seeds based on the density of feature level distribution and
the vulnerability of adversarial robustness. The vulnerability of test seeds
are indicated by the auxiliary information, that are highly correlated with
local robustness. Given a test seed, we then develop a novel genetic algorithm
based local test case generation method, in which two fitness functions work
alternatively to control the perceptual quality of detected AEs. Finally,
extensive experiments confirm that our holistic approach considering
hierarchical distributions is superior to the state-of-the-arts that either
disregard any input distribution or only consider a single (non-hierarchical)
distribution, in terms of not only detecting imperceptible AEs but also
improving the overall robustness of the DL model under testing.
Deep learning robustness, adversarial examples detection, natural
perturbations, distribution-aware testing, robustness growth, safe AI
††copyright: acmcopyright††journalyear: 2023††doi: XXXXXXX.XXXXXXX††journal:
TOSEM††journalvolume: 0††journalnumber: 0††article: X††publicationmonth:
9††ccs: Software and its engineering Software testing and debugging††ccs:
Software and its engineering Software reliability††ccs: Computing
methodologies Machine learning
## 1\. Introduction
Deep Learning (DL) is being explored to provide transformational capabilities
to many industrial sectors including automotive, healthcare and finance. The
reality that DL is not as dependable as required now becomes a major
impediment. For instance, key industrial foresight reviews identified that the
biggest obstacle to gaining benefits of DL is its dependability (Lane et al.,
2016). There is an urgent need to develop methods to enable the dependable use
of DL, for which great efforts have been made in recent years in the field of
DL Verification and Validation (V&V) (Huang et al., 2020; Zhang et al., 2022).
DL robustness is arguably the property in the limelight. Informally,
robustness requires that the decision of the DL model is invariant against
small perturbations on inputs. That is, all inputs in a small input region
(e.g., a norm ball defined in some $L_{p}$-norm distance) should share the
same prediction label by the DL model. Inside that region, if an input is
predicted differently to the given label, then this input is normally called
an Adversarial Example (AE). Most V&V methods designed for DL robustness are
essentially about detecting AEs, e.g., adversarial attack based methods
(Goodfellow et al., 2015; Madry et al., 2018) and coverage-guided testing (Du
et al., 2019; Xie et al., 2022; Ma et al., 2018; Pei et al., 2017; Huang et
al., 2022; Sun et al., 2019).
As recently noticed by the software engineering community, emerging studies on
systematically evaluating AEs detected by aforementioned state-of-the-arts
have two major drawbacks: (i) they do not take the input data distribution
into consideration, therefore it is hard to judge whether the identified AEs
are meaningful to the DL application (Berend et al., 2020; Dola et al., 2021);
(ii) most detected AEs are of poor perceptual quality that are too
unnatural/unrealistic (Harel-Canada et al., 2020) to be seen in real-life
operations. That said, not all AEs are equal nor significantly contribute to
the robustness improvement, given limited resources. A wise strategy is to
detect those AEs that are both being “distribution-aware” and with
natural/realistic pixel-level perturbations, which motivates this work.
Prior to this work, a few notable attempts at distribution-aware testing for
DL have been made. Broadly speaking, the field has developed two types of
approaches: Out-Of-Distribution (OOD) detector based (Dola et al., 2021;
Berend, 2021) and feature-only based (Toledo et al., 2021; Byun et al., 2020).
The former can only detect anomalies/outliers, rather than being “fully-aware”
of the distribution. While the latter is indeed generating new test cases
according to the learnt distribution (in a latent space), it ignores the
pixel-level information due to the compression nature of generative models
used (Zhong et al., 2020). To this end, our approach is advancing in this
direction with the following novelties and contributions:
a) We provide a “divide and conquer” solution—Hierarchical Distribution-Aware
(HDA) testing—by decomposing the input distribution into two levels (named as
global and local) capturing how the feature-wise and pixel-wise information
are distributed, respectively. At the global level, isolated problems of
estimating the feature distribution and selecting best test seeds can be
solved by dedicated techniques. At the local level where features are fixed,
the clear objective is to precisely generate test cases considering perceptual
quality111While determining perceptual quality typically involves subjective
assessments from human observers, objective metrics such as peak signal-to-
noise ratio (PSNR) and structural similarity index (SSIM) can also be used as
measures of perceptual quality (Wang et al., 2004). Throughout the paper,
perceptual quality is defined as the metrics in Sec. 2.4 that are commonly
used in computer vision.. Our extensive experiments show that such
hierarchical consideration is more effective to detect high-quality AEs than
state-of-the-art that either disregards any data distribution or only
considers a single (non-hierarchical) distribution. Consequently, we also show
the DL model under testing exhibits higher robustness after “fixing” the high-
quality AEs detected.
b) At the global level, we propose novel methods to select test seeds based on
the approximated feature distribution of the training data and predictive
robustness indicators, so that the norm balls of the selected seeds are both
from the high-density area of the distribution and relatively unrobust (thus
more cost-effective to detect AEs in later stages). Notably, state-of-the-art
DL testing methods normally select test seeds randomly from the training
dataset without any principled rules. Thus, from a software engineering
perspective, our test seed selection is more practically useful in the given
application context.
c) Given a carefully selected test seed, we propose a novel two-step Genetic
Algorithm (GA) to generate test cases locally (i.e. within a norm ball) to
control the perceptual quality of detected AEs. At this local level, the
perceptual quality distribution of data-points inside a norm ball requires
pixel-level information that cannot be sufficiently obtained from the training
data alone. Thus, we innovatively use common perceptual metrics that quantify
image quality as an approximation of such local distribution. Our experiments
confirm that the proposed GA is not only effective after being integrated into
HDA (as a holistic testing framework), but also outperforms other pixel level
AE detectors in terms of perception quality when applied separately.
d) We investigate black-box (to the DL model under testing) methods for the
main tasks at both levels. Thus, to the best of our knowledge, our HDA
approach provides an end-to-end, black-box solution, which is the first of its
kind and more versatile in software engineering practice.
e) A publicly accessible tool of our HDA testing framework with all source
code, datasets, DL models and experimental results.
## 2\. Preliminaries and Related Work
In this section, we first introduce preliminaries and related work on DL
robustness, together with formal definitions of concepts adopted in our HDA
approach. Then existing works on distribution-aware testing are discussed.
Since our HDA testing also considers the naturalness of detected AEs, some
common perception quality metrics are introduced. In summary, we present Fig.
1 to show the stark contrast of our proposed HDA testing (the green route) to
other related works (the red and amber routes).
Figure 1. Comparison between our proposed Hierarchical Distribution-Aware
(HDA) testing and related works.
### 2.1. DL Robustness and Adversarial Examples
We denote the prediction output of DL model as the vector $f(x)$ with size
equal to the total number of labels. The predicted label
$\hat{f}(x)\\!=\\!\operatorname*{arg\,max}_{i}f_{i}(x)$ where $f_{i}(x)$ is
the $i^{th}$ attribute of vector $f(x)$.
DL robustness requires that the decision of the DL model $\hat{f}(x)$ is
invariant against small perturbations on input $x$. That is, all inputs in an
input region $\eta$ have the same prediction label, where $\eta$ is usually a
small norm ball (defined with some $L_{p}$-norm distance222$p=0,1,2$ and
$\infty$. $L_{\infty}$ norm is more commonly used.) around an input $x$. If an
input $x^{\prime}$ inside $\eta$ is predicted differently to $x$ by the DL
model, then $x^{\prime}$ is called an Adversarial Example (AE).
DL robustness V&V can be based on formal methods (Huang et al., 2017; Ruan et
al., 2018) or statistical approaches (Weng et al., 2018; Webb et al., 2019),
and normally aims at detecting AEs. In general, we may classify two types of
methods (the two branches in the red route of Fig. 1) depends on how the test
cases are generated: (i) Adversarial attack based methods normally optimise
the DL prediction loss to find AEs, which include white-box attack methods
like Fast Gradient Sign Method (FGSM) (Goodfellow et al., 2015) and Projected
Gradient Descent (PGD) (Madry et al., 2018), as well as black-box attacks
(Alzantot et al., 2019; Wu et al., 2021) using GA with gradient-free
optimisation. (ii) Coverage-guided testing optimises the certain coverage
metrics on the DL model’s internal structure, which is inspired by the
coverage testing for traditional software. Several popular test metrics, like
neuron coverage (Pei et al., 2017; Ma et al., 2018), modified
condition/decision coverage (Sun et al., 2019) for CNNs and temporal coverage
(Huang et al., 2022; Du et al., 2019) for RNNs are proposed. While it is
argued that coverage metrics are not strongly correlated with DL robustness
(Yan et al., 2020; Harel-Canada et al., 2020), they are seen as providing
insights into the internal behaviours of DL models and hence may guide test
selection to find more diverse AEs (Huang et al., 2022).
Without loss of generality, we reuse the formal definition of DL robustness in
(Webb et al., 2019; Weng et al., 2019) in this work:
###### Definition 0 (Local Robustness).
The local robustness of the DL model $f(x)$, w.r.t. a local region $\eta$ and
a target label $y$, is:
(1) $\mathcal{R}_{l}(\eta,y):=\int_{x\in\eta}I(x)p_{l}(x\mid
x\in\eta)\mathop{}\\!\mathrm{d}{x}$
where $p_{l}(x\mid x\in\eta)$ is the local distribution of region $\eta$ which
is precisely the “input model” used by both (Webb et al., 2019; Weng et al.,
2019). $I(x)$ is an indicator function, and $I(x)=1$ when $\hat{f}(x)=y$,
$I(x)=0$ otherwise.
To detect as many AEs as possible, normally the first question is—which local
region shall we search for those AEs? I.e. how to select test seeds? To be
cost-effective, we want to explore unrobust regions, rather than regions where
AEs are relatively rare. This requires the local robustness of a region to be
known a priori, which may imply a paradox (cf. Remark 4 later). In this
regard, we can only predict the local robustness of some regions before doing
the actual testing in those regions. We define:
###### Definition 0 (Local Robustness Indicator).
Auxiliary information that strongly correlated with $\mathcal{R}_{l}(\eta,y)$
(thus can be leveraged in its prediction) is named as a local robustness
indicator.
We later seek for such indicators (and empirically show their correlation with
the local robustness), which forms one of the two key factors considered in
selecting test seeds in our method.
Given a test seed, we search for AEs in a local region $\eta$ (around the test
seed) that produce different label from the test seed. This involves the
question on what size of $\eta$ should be, for which we later utilise the
property of:
###### Remark 1 ($r$-separation).
For real-world image datasets, any data-points with different ground truth
labels are at least distance $2r$ apart in the input (pixel) space, where $r$
is estimated case by case and depending on the dataset.
The $r$-separation property was first observed by (Yang et al., 2020):
intuitively it says, there is a minimum distance between two real-world
objects of different labels.
Finally, not all AEs are equal in terms of the “strength of being adversarial”
(stronger AEs may lead to greater robustness improvement in, e.g., adversarial
training (Wang et al., 2019)), for which we define:
###### Definition 0 (Prediction Loss).
Given a test seed $x$ with label $y$, the prediction loss of an input
$x^{\prime}$, which is obtained by adding perturbations to $x$, is defined as:
(2) $\mathcal{J}(f(x^{\prime}),y)=\max_{i\neq
y}(f_{i}(x^{\prime})-f_{y}(x^{\prime}))$
where $f_{i}(x^{\prime})$ returns the probability of label $i$ after input
$x^{\prime}$ being processed by the DL model $f$.
Note, $\mathcal{J}\geq 0$ implies $\operatorname*{arg\,max}_{i}f_{i}(x)\neq y$
and thus $x^{\prime}$ is an AE of $x$.
Next, to measure the DL models’ overall robustness across the whole input
domain, we introduce a notion of global robustness. Being different to some
existing definitions where robustness of local regions are treated equally
when calculating global robustness over several regions (Wang et al., 2021a;
Wang et al., 2021c), ours is essentially a “weighted sum” of the robustness of
local regions where each weight is the probability of the associated region on
the input data distribution. Defining global robustness in such a
“distribution-aware” manner aligns with our motivation—as revealed later by
empirically estimated global robustness, our HDA appears to be more effective
in supporting the growth of the overall robustness after “fixing” those
distribution-aware AEs.
###### Definition 0 (Global Robustness).
The global robustness of the DL model $f(x)$ is defined as:
(3)
$\mathcal{R}_{g}:=\sum_{\eta\in\mathcal{X}}p_{g}(x\in\eta)\mathcal{R}_{l}(\eta,y)$
where $p_{g}(x\mid x\in\eta)$ is the global distribution of region $\eta$
(i.e., a pooled probability of all inputs in the region $\eta_{z}$) and
$\mathcal{R}_{l}(\eta,y)$ is the local robustness of region $\eta$ to the
label $y$.
The estimation of $\mathcal{R}_{g}$, unfortunately, is very expensive that
requires to compute the local robustness $\mathcal{R}_{l}$ of a large number
of regions over the input domain $\mathcal{X}$. Thus, from a practical
standpoint, we adopt an empirical definition of the global robustness in our
later experiments, which has been commonly used for DL robustness evaluation
in the adversarial training (Wang et al., 2021b; Madry et al., 2018; Wang et
al., 2019; Zhang et al., 2019).
###### Definition 0 (Empirical Global Robustness).
Given a DL model $f$ and a validation dataset $D_{v}$, we define the empirical
global robustness as $\hat{\mathcal{R}_{g}}:(f,D_{v},T)\rightarrow[0,1]$ where
T denotes a given type of AE detection method and $\hat{\mathcal{R}_{g}}$ is
the weighted accuracy on AEs obtained by conducting T on $\langle
f,D_{v}\rangle$.
To be “distribution-aware”, the synthesis of $D_{v}$ should conform to the
global distribution. $D_{v}$ can be sampled from the train/test data according
to global distribution. The norm ball around each input data $x$ in $D_{v}$
represents a region $\eta$. Each region $\eta$ is explicitly assigned a weight
to indicate the density on global distribution. For each region $\eta$, we
calculate the prediction accuracy on AEs, detect by T according to local
distribution, to approximate the local robustness $\mathcal{R}_{l}$.
Consequently, the set of AEs for $D_{v}$ may represent the input distribution
and the weighted accuracy on these AEs approximate the global robustness.
### 2.2. Distribution-Aware Testing for DL
There are increasing amount of DL testing works developed towards being
distribution-aware (as summarised in the amber route of Fig. 1). Deep
generative models, such as Variational Auto-Encoders (VAE) and Generative
Adversarial Networks (GAN), are applied to approximate the training data
distribution, since the inputs (like images) to Deep Neural Network (DNN) are
usually in a high dimensional space. Previous works heavily rely on OOD
detection (Dola et al., 2021; Berend, 2021) or synthesising new test cases
directly from latent spaces (Byun and Rayadurgam, 2020; Toledo et al., 2021;
Dunn et al., 2020; Kang et al., 2020; Dunn et al., 2021). The former does not
comprehensively consider the whole distribution, rather flags outliers, thus a
more pertinent name of it should be out-of-distribution-aware (OODA) testing.
While for both types of methods, another problem arises that the distribution
encoded by generative models only contain the feature-wise information and
easily filter out the pixel-wise perturbations (Zhong et al., 2020). The
images added with pixel-wise perturbations should still fall into the local
region $\eta$, which is decided by $r$-separation property (ref. Remark 1).
Although, Dunn et al. (Dunn et al., 2020) propose to generate the fine-grained
perturbations by perturbing the output of last layers of GAN’s generator, the
fine-grained perturbations cannot be guaranteed to follow $r$-separation
property and thus may change the class of the image. Consequently, directly
searching and generating test cases from the latent space of generative models
may only perturb features, thus called Feature-Only Distribution-Aware (FODA)
in this paper (while also named as semantic AEs in some literature (Hosseini
and Poovendran, 2018; Zhao et al., 2018)). As an example in later experiment
results, i.e. Table 8(b), FODA will produce AEs, the perturbations of which
exceed the $r$-separation limit. Our approach, the green route in Fig. 1,
differs from aforementioned works by 1) considering both the global (feature
level) distribution in latent spaces and the local (pixel level) perceptual
quality distribution in the input space; 2) leveraging the robustness
indicator $\mathcal{R}$ to select test seeds which is more error-prone and
thus easier for detecting AEs; 3) proposing novel GA based test case
generation to detect AEs with high perceptual quality. Therefore, when HDA is
experimentally compared to the existing works, the selected test seeds of HDA
have higher probability density and lower robustness. The detected AEs by HDA
also have higher perceptual quality (in terms of those metrics encoded by GA)
than others.
### 2.3. Test Input Prioritisation and Generation
When testing DL based systems, in order to save computation and reduce the
cost on labelling data, test input prioritisation strategies are adopted.
(Weiss and Tonella, 2022) experimentally confirms that DeepGini outperforms
various types of surprise and neuron coverage metrics in terms of the
capability to detect misclassifications. (Attaoui et al., 2023) propose the
unsafe set selection algorithm, leveraging the density-based clustering of
error-inducing images. Both test selection methods aim at detecting
misclassified images for retraining and improving generalisation performance
of DNNs.
There are also some test input generation methods, the main idea of which is
to promote diversity of test cases in order to cover more faults of DL
systems. (Riccio and Tonella, 2020) develops a search-based tool to generate
frontier inputs for DL systems. DEEPMETIS (Riccio et al., 2021) augment the
test set for DL by increasing the mutation scores.
The above test input generation/prioritisation methods have different goals
from our HDA testing. They focus on detecting or generating more misclassified
test cases within certain test budget, while our HDA testing targets at
generating on-distribution AEs to improve operational robustness of DL
systems. Feature distribution learnt from VAE as well as the predicted
robustness indicator facilitate the generation of those on-distribution AEs.
### 2.4. Perceptual Quality of Images
Locally, data-points (around the selected seed) sharing the same feature
information may exhibit perceptual difference from the selected seed. To
capture such distribution, some perceptual quality metric can be utilised to
compare the perceptual difference between the original image $x$ and perturbed
image $x^{\prime}$. Some common metrics for perceptual quality include:
* •
Mean Squared Error (MSE): measures the mean squared difference between the
original image $x$ and perturbed image $x^{\prime}$,
(4) $MSE=\frac{1}{n}\sum_{i=1}^{n}(x_{i}-x_{i}^{\prime})^{2}$
where $n$ is the image size, and $x_{i}$ is the value of image pixel $i$.
* •
Peak Signal-to-Noise Ratio (PSNR) (Gonzalez and Woods, 1992): measures the
quality of a signal’s representation after being subject to noise during
transmission or processing. PSNR is expressed in decibels (dB) and is
calculated by comparing the maximum possible power of the original signal to
the power of the noise that affects its fidelity,
(5) $PSNR=20*log_{10}\frac{MAX}{\sqrt{MSE}}$
where $MAX$ is the maximum possible pixel value of the image. When the pixels
are represented using 8 bits per sample, this is 255.
* •
Structural Similarity Index Measure (SSIM) (Wang et al., 2004): measures the
similarity between two images based on the idea that the human visual system
is highly sensitive to changes in structural information, such as edges,
textures, and patterns, rather than just changes in pixel values. It works by
comparing the luminance (l), contrast (c), and structural information (s) of
two images, and producing a score between 0 (completely dissimilar) and 1
(identical) that indicates their similarity,
(6) $\begin{split}SSIM(x,x^{\prime})=&l(x,x^{\prime})^{\alpha}\cdot
c(x,x^{\prime})^{\beta}\cdot s(x,x^{\prime})^{\gamma}\\\
l(x,x^{\prime})=\frac{2\mu_{x}\mu_{x^{\prime}}+c_{1}}{\mu_{x}^{2}+\mu_{x^{\prime}}^{2}+c_{1}},\;c(x,x^{\prime})&=\frac{2\sigma_{x}\sigma_{x^{\prime}}+c_{2}}{\sigma_{x}^{2}+\sigma_{x^{\prime}}^{2}+c_{2}},\;s(x,x^{\prime})=\frac{\sigma_{xx^{\prime}}+c_{3}}{\sigma_{x}\sigma_{x^{\prime}}+c_{3}}\end{split}$
where $\mu_{x}$ and $\mu_{x^{\prime}}$ are the mean values of $x$ and
$x^{\prime}$, respectively. $\sigma_{x}$ and $\sigma_{x^{\prime}}$ are the
standard deviations of $x$ and $x^{\prime}$, respectively.
$\sigma_{xx^{\prime}}$ is the covariance between $x$ and $x^{\prime}$.
$c_{1}$, $c_{2}$, and $c_{3}$ are constants that prevent division by zero. The
constants $\alpha$, $\beta$, and $\gamma$ are typically set to 1, 1, and 1,
respectively, although different values may be used depending on the
application. Unlike other similarity measures, such as MSE and PSNR, SSIM is
able to account for perceptual differences in image quality, and is often
considered a more accurate measure of image quality.
* •
Fréchet Inception Distance (FID) (Heusel et al., 2017): compares the
distribution between a set of original images and a set of perturbed images.
It is calculated by first using a pre-trained Inception-v3 neural network
(Szegedy et al., 2016) to extract features from both sets of images. Then, the
mean and covariance of these features are calculated for each set, and the
distance between these mean and covariance matrices is calculated using the
Fréchet distance,
(7)
$FID=\|\mu-\mu^{\prime}\|_{2}^{2}+Tr(\Sigma+\Sigma^{\prime}-2(\Sigma^{1/2}\cdot\Sigma^{\prime}\cdot\Sigma^{1/2})^{1/2})$
where $\mu$ and $\mu^{\prime}$ are the mean feature vectors of the original
and perturbed image sets, respectively. $\Sigma$ and $\Sigma^{\prime}$ are the
covariance matrices of the original and perturbed image sets, respectively.
$Tr()$ denotes the trace of a matrix. Lower FID scores indicate that the
perturbed images are closer in terms of perception to the original images,
while higher FID scores indicate greater differences between the perturbed and
original images.
Notably, all these metrics are current standards for assessing the quality of
images, as widely used in the experiments of aforementioned related works.
## 3\. The Proposed Method
Figure 2. An example of Hierarchical Distribution Aware Testing
We first present an overview of our HDA testing, cf. the green route in Fig.
1, and then dive into details of how we implement each stage by referring to
an illustrative example in Fig. 2.
### 3.1. Overview of HDA Testing
The core of HDA testing is the hierarchical structure of two distributions. We
formally define the following two levels of distributions:
###### Definition 0 (Global Distribution).
The global distribution captures how feature level information is distributed
in some (low-dimensional) latent space after data compression.
###### Definition 0 (Local Distribution).
Given a data-point sampled from the latent space, we consider its norm ball in
the input pixel space. The local distribution is a conditional distribution
capturing the perceptual quality of all data-points within the norm ball.
Due to the sparsity of data over the high dimensional input space, it is hard
to estimate the input distribution. Therefore, we turn to estimate the feature
level (global) distribution in latent space, which is the representation of
compressed data, in which data points with similar features are closer to each
other (Liu et al., 2019). DNNs, e.g. encoder of VAEs, map any data points in
the high dimensional input space to the low dimensional latent space. It
infers that the input space can be partitioned into well-defined regions,
where each region corresponds to a particular data point in the latent space.
latent space is an abstract, multidimensional space where each dimension
represents a feature or characteristic of the data (Liu et al., 2019). The
regions in the input space are determined based on the values of these
features and how they relate to each other. In other words, we can identify
distinct patterns or clusters of data in the input space, which share the same
set of features, and maps them to corresponding point in the latent space. By
fitting a global distribution in the latent space, we actually model the
distribution of distinct regions over the input space. The local distribution
is defined as a conditional distribution within each region. Thus, we propose
the following remark.
###### Remark 2 (Decompose one distribution into two levels).
Given the definitions of global and local distributions, denoted as $p_{g}$
and $p_{l}$ respectively, we may decompose a single distribution over the
entire input domain $\mathcal{X}$ as:
(8) $p(x)=\int
p_{l}(x|x\in\eta_{z})p_{g}(x\in\eta_{z})\mathop{}\\!\mathrm{d}{z}$
where variable $z$ represents a set of features while $\eta_{z}$ represents a
region in the input space that “maps” to the $z$ point in the latent space.
Intuitively, compared to modelling a single distribution, our hierarchical
structure of distributions is superior in that the global distribution guides
for which regions of the input space to test, while the local distribution can
be leveraged to precisely control the quality of test cases. Given the
definition of distribution, the goal of HDA testing can be formalised as
below:
###### Remark 3 (Goal of HDA testing).
The goal of HDA testing is to detect AE $x^{\prime}$, which is around the
local region of test seed $x$, such that $x^{\prime}$ lies within the high
probability density region of distribution $p(x^{\prime})$, and $x$ and
$x^{\prime}$ are classified differently.
To detect such on distribution AEs, HDA testing has the following process,
which can divided into three stages, depicted as the green route in Fig. 1:
##### Stage 1: Explicitly Approximate the Global Distribution
We first extract feature-level information from the given dataset by using
data compression techniques—the encoder of VAEs in our case, and then
explicitly approximate the global distribution in the latent-feature space,
using Kernel Density Estimator (KDE).
##### Stage 2: Select Test Seeds Based on the Global Distribution and Local
Robustness Indicators
Given the limited testing budget, we want to test in those local input regions
that are both more error-prone and representative of the input distribution.
Thus, when selecting test seeds, we consider two factors—the local robustness
indicators (cf. Definition 2) and the global distribution. For the former, we
propose several auxiliary information with empirical studies showing their
correlation with the local robustness, while the latter has already been
quantified in the first stage via KDE.
##### Stage 3: Generate Test Cases Around Test Seeds Considering the Local
Distribution and Prediction Loss of AEs
When searching for AEs locally around a test seed given by the 2nd stage, we
develop a two-step GA in which the objective function is defined as a fusion
of the prediction loss (cf. Definition 3) and the local distribution (modelled
by common perceptual quality metrics). Such fusion of two fitness functions
allows the trade-off between the “strength of being adversarial” and the
perceptual quality of the detected AEs. The optimisation is subject to the
constraint of only exploring in a norm ball whose central point is the test
seed and with a radius smaller than the $r$-separation distance (cf. Remark
1).
While there are some alternatives may also suffice for the purpose of each
stage, our chosen technical solutions are the most effective and popular in
consideration of the distribution awareness.
### 3.2. Approximation of the Global Distribution
Given the training dataset $\mathcal{D}$, the task of approximating the input
distribution is equivalent to estimating a Probability Density Function (PDF)
over the input domain $\mathcal{X}$ given $\mathcal{D}$. Despite this is a
common problem with many established solutions, it is hard to accurately
approximate the distribution due to the relatively sparse data of
$\mathcal{D}$, compared to the high dimensionality of the input domain
$\mathcal{X}$. So the practical solution is to do dimensionality reduction and
then estimate the global distribution, which indeed is the first step of all
existing methods of distribution-aware DL testing.
Specifically, we choose VAE-Encoder+KDE 333We only use the encoder of VAEs for
feature extraction, rather than generate new data from the decoder, which is
different to other methods mentioned in Section 2.2. for their simplicity and
popularity. To effectively train the VAE model, we use a combination of a
reconstruction loss and a KL divergence loss to optimise the model. The
reconstruction loss measures the difference between the input data and the
output of the decoder, while the KL divergence loss measures the difference
between the learned latent distribution and the prior distribution. During
training, the reconstruction loss and KL divergence loss are calculated and
minimised simultaneously. Regularisation techniques such as dropout and weight
decay can be used to prevent overfitting.
Assume $\mathcal{D}$ contains $n$ samples and each $x_{i}\in\mathcal{D}$ is
encoded by VAE-Encoder as a Gaussian distribution $z_{i}$ in the latent space,
we can estimate the PDF of $z$ (denoted as $Pr(z)$) based on the encoded
$\mathcal{D}$. The $Pr(z)$ conforms to the mixture of Gaussian distributions,
i.e., $z\sim\mathcal{N}(\mu_{z_{i}},\sigma_{z_{i}})$. Notably, this mixture of
Gaussian distributions nicely aligns with the gist of adaptive KDE (Lokerse et
al., 1995), which uses the following estimator:
(9) $p_{g}(x\in\eta_{z})\propto
Pr(z)\simeq\frac{1}{n}\sum_{i=1}^{n}K_{h_{i}}(z-\mu_{z_{i}})$
That is, when choosing a Gaussian kernel for $K$ in Eqn. (9) and adaptively
setting the bandwidth parameter $h_{i}=\sigma_{z_{i}}$ (i.e., the standard
deviation of the Gaussian distribution representing the compressed sample
$z_{i}$), the VAE-Encoder and KDE are combined “seamlessly”. Finally, our
global distribution $p_{g}(x\in\eta_{z})$ (a pooled probability of all inputs
in the region $\eta_{z}$ that corresponds to a point $z$ in the latent space)
is proportional to the approximated distribution of $z$ with the PDF $Pr(z)$.
Running Example: The left diagram in Fig. 2 depicts the global distribution
learnt by KDE, after projected to a two-dimensional space for visualisation.
The peaks444Most training data lie in this region or gather around the region.
are evaluated with highest probability density over the latent space by KDE.
### 3.3. Test Seeds Selection
Selecting test seeds is actually about choosing which norm balls (around the
test seeds) to test for AEs. To be cost-effective, we want to test those with
higher global probabilities and lower local robustness at the same time. For
the latter requirement, there is potentially a paradox:
###### Remark 4 (A Paradox of Selecting Unrobust Norm Balls).
To be informative on which norm balls to test for AEs, we need to estimate the
local robustness of candidate norm balls (by invoking robustness estimators to
quantify $\mathcal{R}_{l}(\eta,y)$, e.g., (Webb et al., 2019)). However, local
robustness evaluation itself is usually about sampling for AEs (then fed into
statistical estimators) that consumes the testing resources.
To this end, instead of directly evaluating the local robustness of a norm
ball, we can only indirectly predict it (i.e., without testing/searching for
AEs) via auxiliary information that we call local robustness indicators (cf.
Definition 2). In doing so, we save all the testing budget for the later stage
when generating local test cases.
Given a test seed $x$ with label $y$, we propose two robustness indicators
(both relate to the vulnerability of the test seed to adversarial attacks)—the
prediction gradient based score (denoted as ${S}_{\textit{grad}}$) and the
score based on separation distance of the output-layer activation (denoted as
$S_{\textit{sep}}$):
(10)
$\begin{split}&S_{\textit{grad}}=||\nabla_{x}\mathcal{J}(f(x),y)||_{\infty}\\\
&S_{\textit{sep}}=\min_{\hat{x}}||f(x)-f(\hat{x})||_{\infty}\;\quad\textrm{s.t.}\;y\neq\hat{y}\end{split}$
These allow prediction of a whole norm ball’s local robustness by the limited
information of its central point (the test seed). The gradient of a DNN’s
prediction with respect to the input is a white-box metric, that is widely
used in adversarial attacks, such as FGSM (Goodfellow et al., 2015) and PGD
(Madry et al., 2018) attacks. A greater gradient calculated at a test seed
implies that AEs are more likely to be found around it. The activation
separation distance is regarded as a black-box metric and refers to the
minimum $L_{\infty}$ norm between the output activations of the test seed and
any other data with different labels. Intuitively, a smaller separation
distance implies a greater vulnerability of the seed to adversarial attacks.
We later show empirically that indeed these two indicators are highly
correlated with the local robustness.
After quantifying the two required factors, we combine them in a way that was
inspired by (Zhao et al., 2021a). In (Zhao et al., 2021a), the DL reliability
metric is formalised as a weighted sum of local robustness where the weights
are operational probabilities of local regions. To align with that reliability
metric, we do the following steps to select test seeds:
(i) For each data-point $x_{i}$ in the test set, we calculate its global
probability (i.e., $Pr(z_{i})$ where $z_{i}$ is its compressed point in the
VAE latent space) and one of the local robustness indicators (either white-box
or black-box, depending on the available information).
(ii) Normalise both quantities to the same scale.
(iii) Rank all data-points by the product of their global probability and
local robustness indicator.
(iv) Finally we select top-$k$ data-points as our test seeds, and $k$ depends
on the testing budget.
Running Example: In the middle diagram of Fig. 2, we add in the local
robustness indicator results of the training data which are represented by a
scale of colours—darker means lower predicted local robustness while lighter
means higher predicated local robustness. By our method, test seeds selected
are both from the highest peak (high probability density area of the global
distribution) and relatively darker ones (lower predicated local robustness).
### 3.4. Local Test Cases Generation
Not all AEs are equal in terms of the “strength of being adversarial”, and
stronger AEs are associated with higher prediction loss (cf. Definition 3).
Detecting AEs with higher prediction loss may benefit more when considering
the future “debugging” step, e.g., by adversarial retraining (Wang et al.,
2019). Thus, at this stage, we want to search for AEs that exhibit a high
degree of adversarial behaviour while also conform to the local distribution.
That is, the local test case generation can be formulated as the following
optimisation given a seed $(x,y)$:
(11) $\displaystyle\max_{x^{\prime}}$
$\displaystyle\mathcal{J}(f(x^{\prime}),y)+\alpha\cdot
p_{l}(x^{\prime}|x^{\prime}\in\eta_{z_{x}})$ s.t.
$\displaystyle||x-x^{\prime}||_{\infty}\leq r$
where $\mathcal{J}$ is the prediction loss,
$p_{l}(x^{\prime}|x^{\prime}\in\eta_{z_{x}})$ is the local distribution (note,
$z_{x}$ represents the latent features of test seed $x$), $r$ is the
$r$-separation distance, and $\alpha$ is a coefficient to balance the two
terms. As what follows, we note two points on Eqn. (11): why we need the
constraint and how we quantify the local distribution.
The constraint in Eqn. (11) determines the right locality of local
robustness—the “neighbours” that should have the same ground truth label $y$
as the test seed. We notice the $r$-separation property of real-world image
datasets (cf. Remark 1) provides a sound basis to the question. Thus, it is
formalised as a constraint that the optimiser can only search in a norm ball
with a radius smaller than $r$, to guarantee the detected AEs are indeed
“adversarial” to label $y$.
While the feature level information is captured by the global distribution
over a latent space, we only consider how the pixel level information is
locally distributed in terms of perceptual quality. Three common quantitative
metrics—MSE, PSNR and SSIM introduced in Section 2.4—are investigated. We
note, those three metrics by no means are the true local distribution
representing perceptual quality, rather quantifiable indicators from different
aspects. Thus, in the optimisation problem of Eqn. (11), replacing the local
distribution term with them would suffice our purpose. So, we redefine the
optimisation problem as:
(12)
$\displaystyle\max_{x^{\prime}}\mathcal{J}(f(x^{\prime}),y)\\!+\\!\alpha\\!\cdot\\!\mathcal{L}(x,x^{\prime}),\quad\textrm{s.t.}\,||x\\!-\\!x^{\prime}||_{\infty}\\!\leq\\!r$
where $\mathcal{L}(x,x^{\prime})$ represents those perceptual quality metrics
correlated with the local distribution of the seed $x$. Certainly,
implementing $\mathcal{L}(x,x^{\prime})$ requires some prepossessing, e.g.,
normalisation and negation, depending on which metric is adopted.
Considering that the second term of the objective function in Eqn. (12) may
not be differentiable and/or the DL model’s parameters are not always
accessible, Non-dominated Sorting Genetic Algorithm II (NSGA-II) (Deb et al.,
2002) may be adopted here to solve the multi-objective optimisation. NSGA-II
is designed to search for a diverse set of solutions along the Pareto-optimal
front, rather than just a single solution. This means that NSGA-II produces a
set of solutions that span the entire Pareto-optimal front, providing
decision-makers with a range of options to choose from. Therefore, NSGA-II is
usually computationally intensive and requires a huge amount of queries of DL
model’s prediction. Since we are only concern about test cases which have high
perceptual quality and conditioned on AEs (prediction loss greater than 0), it
is not necessary to produce a range of options for the trades-off between
perceptual quality and prediction loss, which may waste a lot of computational
resource on generating high perceptual quality but no adversarial examples.
For this reason, we propose to scalarize the vector of objectives into one
objective by averaging the objectives with weight vector and reduce the weight
dependency with alternation mechanism. That is, we develop a GA with two
fitness functions to effectively and efficiently detect AEs, as shown in
Algorithm 1.
Algorithm 1 Two-Step GA Based Local Test Cases Generation
1:Test seed $(x,y)$, neural network function $f(x)$, local perceptual quality
metric $\mathcal{L}(x,x^{\prime})$, population size $N$, maximum iterations
$T$, norm ball radius $r$, weight parameter $\alpha$, number of generated test
cases $m$.
2:A set of $m$ test cases $\mathcal{T}$
3:$F_{1}=\mathcal{J}(f(x^{\prime}),y)$,
$F_{2}=\mathcal{J}(f(x^{\prime}),y)+\alpha\cdot\mathcal{L}(x,x^{\prime})$
4:for $i=1,...,N$ do
5: $\mathcal{T}[i]=x+\mathit{uniform}(-r,+r)$
6:end for
7:while $t<T$ or $\max(\mathit{fit\\_list_{2}})$ does not converge do
8: $\mathit{fit\\_list_{1}}=\mathit{cal\\_fitness}(F_{1},\mathcal{T})$
9: $\mathit{fit\\_list_{2}}=\mathit{cal\\_fitness}(F_{2},\mathcal{T})$
10: if $\mathit{majority}(\mathit{fit\\_list_{1}}<0)$ then
11: $\mathit{parents}=\mathit{selection}(\mathit{fit\\_list_{1}},\mathcal{T})$
12: else
13: $\mathit{parents}=\mathit{selection}(\mathit{fit\\_list_{2}},\mathcal{T})$
14: end if
15: $\mathcal{T}=\mathit{crossover}(\mathit{parents},N)$
16: $\mathcal{T}=\mathit{mutation}(\mathcal{T})\cup\mathit{parents}$
17: $t=t+1$
18:end while
19:$\mathit{fit\\_list_{2}}=\mathit{cal\\_fitness}(F_{2},\mathcal{T})$
20:$idx=\operatorname*{arg\,max}(fit\\_list_{2})[:m]$
21:$\mathcal{T}=\mathcal{T}[idx]$
22:return test set $\mathcal{T}$
Algorithm 1 presents the process of generating a set of $m$ test cases
$\mathcal{T}$ from a given seed $x$ with label $y$ (denoted as $(x,y)$). At
line 1, we define two fitness functions (the reason behind it will be
discussed later). GA based test case generation consists of 4 steps:
initialisation, selection, crossover, and mutation, the last three of which
are repeated until the convergence of fitness values or the maximum iterations
are reached.
##### Initialisation
The initialisation of population is crucial to the quick convergence.
Diversity of initial population could promise approximate global optimal
(Konak et al., 2006). We initialise the population by adding uniform noise in
range $(-r,+r)$ to the test seed, at line 2-4.
##### Selection
The fitness function is defined to select fitted individuals as parents for
the latter operations. We use the fitness proportionate selection (Lipowski
and Lipowska, 2012) for operator $selection()$.
(13)
$p_{i}=\frac{\mathcal{F}_{i}}{\sum_{i=1}^{n}\mathcal{F}_{i}},\;\mathcal{F}_{i}\in
fit\\_list$
The fitness value is used to associate a probability of selection $p_{i}$ for
each individuals to maintaining good diversity of population and avoid
premature convergence. The fitness function is the objective function to be
optimised. At line 6-12, we calculate two defined fitness function values on
the population and select individuals based on one of the fitness values
according to some judgement (the reason behind it will be discussed later).
Figure 3. Illustration of crossover and mutation in Two-Step GA Based Local
Test Cases Generation
##### Crossover
At line 13, the crossover operator will combine a pair of parents from last
step to generate a pair of children, which share many of the characteristics
from the parents. The half elements of parents are randomly exchanged as an
example shown in Fig. 3.
##### Mutation
At line 14, some elements of children are randomly altered to add variance in
the evolution. It should be noticed that the mutated samples should still fall
into the local region $\eta$ around test seed $x$. Finally, the children and
parents will be the individuals for the next generation.
##### Termination
At line 5, the termination condition of GA is either maximum number of
iterations is reached or the highest ranking of fitness reaches a plateau such
that successive iterations no longer produce better results.
As seen above, the main difference between our approach and common GA test
case generation is that we use two fitness functions, which work alternatively
to guide the selection of parents. The reason why we propose two fitness
functions is because, we notice that there is a trade-off between the two
objectives $\mathcal{J}$ and $\mathcal{L}$ in the optimisation. Prediction
loss $\mathcal{J}$ is related to the adversarial strength, while $\mathcal{L}$
indicates the local distribution. Intuitively, generating the test cases with
high local probability tends to add small amount of perturbations to the seed,
while a greater perturbation is more likely to induce high prediction loss. To
avoid the competition between the two terms that may finally leads to a
failure of detecting AEs, we define two fitness functions to precisely control
the preference at different stages:
(14) $\displaystyle F_{1}=\mathcal{J}(f(x^{\prime}),y),\quad
F_{2}=\mathcal{J}(f(x^{\prime}),y)+\alpha\cdot\mathcal{L}(x,x^{\prime})$
At early stage, $F_{1}$ is optimised to quickly direct the generation of AEs,
with $F_{1}>0$ meaning the detection of an AE. When most individuals in the
population are AEs, i.e., $\mathit{majority}(\mathit{fit\\_list_{1}}\geq 0)$,
the optimisation moves to the second stage, in which $F_{1}$ is replaced by
$F_{2}$ to optimise the local distribution indicator as well as the prediction
loss. It is possible555Especially when a large $\alpha$ is used, i.e., with
preference on detecting AEs with high local probability than with high
adversarial strength, cf. the aforementioned trade-off. that the prediction
loss of most individuals again becomes negative, then the optimisation will go
back to the first stage. With such a mechanism of alternatively using two
fitness functions in the optimisation, the proportion of AEs in the population
is effectively prevented from decreasing.
Algorithm 1 describes the process for generating $m$ local test cases given a
single test seed. Suppose $n$ test seeds are selected earlier and in total $M$
local test cases are affordable, we can allocate, for each test seed $x_{i}$,
the number of local test cases $m_{i}$, according to the $n$ (re-normalised)
global probabilities, which emphasises more on the role of distribution in our
detected set of AEs.
Running Example: The right diagram in Fig. 2 plots the local distribution
using MSE as its indicator, and visualises the detected AEs by different
testing methods. Unsurprisingly, all AEs detected by our proposed HDA testing
are located at the high density regions (and very close to the central test
seed), given it considers the perceptual quality metric as one of the
optimisation objectives in the two-step GA based test case generation. In
contrast, other methods (PGD and coverage-guided testing) are less effective.
## 4\. Evaluation
We evaluate the proposed HDA testing method by performing extensive
experiments to address the following research questions (RQs):
##### RQ1 (Effectiveness): How effective are the methods adopted in the three
main stages of HDA?
Namely, we conduct experiments to i) examine the accuracy of combining VAE-
Encoder+KDE to approximate the global distribution; ii) check the correlation
significance of the two proposed local robustness indicators with the local
robustness; iii) investigate the effectiveness of our two-step GA for local
test cases generation.
##### RQ2 (AE Quality): How is the quality of AEs detected by HDA?
Comparing to conventional adversarial attack and coverage-guided testing
methods and more recent distribution-aware testing methods, such as OODA and
FODA, we introduce a comprehensive set of metrics to evaluate the quality of
AEs detected by HDA and others.
##### RQ3 (Sensitivity): How sensitive is HDA to the DL models under testing?
We carry out experiments to assess the capability of HDA applied on DL models
(adversarially trained) with different levels of robustness.
##### RQ4 (Robustness Growth): How useful is HDA to support robustness growth
of the DL model under testing?
We examine the global robustness of DL models after “fixing” the AEs detected
by various testing methods.
### 4.1. Experiment Setup
We consider five popular benchmark datasets and five diverse model
architectures for evaluation. In order to obtain statistical results, we train
10 models for each benchmark dataset, initialising their weights according to
the normal distribution. We use Adam optimiser with learning rate $10^{-2}$
and weight decay $10^{-5}$, and train 100 epochs. Details of the datasets and
average accuracy (Mean $\pm$ SD) of trained DL models under testing are listed
in Table 1. The norm ball radius $r$ is calculated based on the $r$-separation
distance (cf. Remark 1) of each dataset. When comparing different AE detection
approaches on 10 models, the evaluation results are dependent on individual
model and the difference approximately follow the normal distribution by
visualisation and normality test666Some statistical tests for AE Prop. and %
of Valid AEs will output “NaN” due to the identical evaluation results between
different approaches.. Therefore, we perform paired two-sample T-test and
discuss the statistical significance in the experiments. The null hypothesis
is that different AE detection approaches have identical average values for
evaluation metrics. The calculated $p-value>0.05$ indicates that the null
hypothesis is true, otherwise it is false.
In RQ1, we validate the accuracy and effectiveness of HDA on detecting high
quality AEs from high probability density region and decide the hyper-
parameter settings for the following experiments. That is, for RQ2-RQ4, we use
the activation separation distance score ${S}_{\textit{sep}}$ as local
robustness indicator and MSE as perceptual quality metric for images. In RQ2
we compare the quality of AEs detected by HDA and others. In RQ3, we add the
comparison on DL models, enhanced by PGD-based adversarial training, for
sensitivity analysis. Table 1 also records the accuracy of these adversarially
trained models. Adversarial training trades the generalisation accuracy for
the robustness as expected (thus a noticeable decrement of the training and
testing accuracy) (Zhang et al., 2019). In RQ4, we firstly sample 10000 data
points from the global distribution as validation set and detect AEs around
them by different methods. Then, we fine-tune the normally trained models with
training dataset augmented by these AEs. 10 epochs are taken along with ‘slow
start, fast decay’ learning rate schedule (Jeddi et al., 2021) to reduce the
computational cost while improving the accuracy-drop and robustness. To
empirically estimate the global robustness on validation set, we find another
set of AEs according to local distribution, different from the fine-tuning
data. These AEs, crafted from validation datasets, are miss-classified by
normally trained models. Thus, empirical global robustness of normally trained
models is set to 0 as the baseline.
When training VAE models, the loss function is a combination of reconstruction
loss and KL divergence loss. Two losses are weighted equally and
simultaneously optimised. We also use Adam optimiser with learning rate
$10^{-2}$ and weight decay $10^{-5}$, and train 100 epochs to obtain VAE
models. To avoid the posterior collapse that VAE often converges to a
degenerated local optimum, we add Batch Normalisation (Zhu et al., 2020)
before the output of encoder. The latent dimensions and the reconstruction
loss are listed in Table 3.
For readers’ convenience, all the metrics used in RQ2, RQ3 and RQ4 for
comparisons are listed in Table 2. The metrics are introduced to
comprehensively evaluate the quality of detected AEs and the DL models from
different aspects. When comparing HDA testing with PGD attack, coverage guided
testing, OODA and FODA, we have the following settings for the tools.
Specifically, we use PGD attack with 10 steps for gradient ascent, and step
size 2/255; HDA testing with population size $N=1000$, maximum iterations
$T=500$, and weight parameter $\alpha=1$; neuron coverage (Sun et al., 2018)
with Gaussian noise $mean=0,std=1$ to generate perturbed data, and 100
iterations to increase the coverage rate by fuzzing; OODA testing (Dola et
al., 2021) with 10 steps for gradient ascent, step size 2/255, default
hyperparameter for balancing two goals and default reconstruction probability
threshold used in the released code 777https://github.com/swa112003/
DistributionAwareDNNTesting.; FODA testing (Byun et al., 2020) with the same
latent space encoded by VAE used in HDA, random sampling to search for AEs in
the latent space. To achieve the fair comparison, we set the same perturbation
radius $r$ for PGD attack, HDA testing and coverage guided testing, since they
have the restrictions for the validity of test cases. In addition, we set the
same number of steps and step size for PGD attack and OODA testing, and
utilise the same latent space across HDA and FODA.
Table 1. Details of the datasets and DL models under testing.
Dataset | Image Size | $r$ | DL Model | Normal Training | Adversarial Training
---|---|---|---|---|---
Avg. Train Acc. | Avg. Test Acc. | Avg. Train Acc. | Avg. Test Acc.
MNIST | $1\times 32\times 32$ | 0.1 | LeNet5 | $99.88\%\pm 0.06\%$ | $98.73\%\pm 0.12\%$ | $99.77\%\pm 0.01\%$ | $98.84\%\pm 0.01\%$
Fashion-MNIST | $1\times 32\times 32$ | 0.08 | AlexNet | $95.12\%\pm 0.85\%$ | $90.70\%\pm 0.25\%$ | $86.23\%\pm 0.05\%$ | $85.11\%\pm 0.05\%$
SVHN | $3\times 32\times 32$ | 0.03 | VGG11 | $96.12\%\pm 0.45\%$ | $94.86\%\pm 0.21\%$ | $89.89\%\pm 0.69\%$ | $90.32\%\pm 0.81\%$
CIFAR-10 | $3\times 32\times 32$ | 0.03 | ResNet20 | $97.81\%\pm 0.58\%$ | $88.28\%\pm 0.63\%$ | $79.48\%\pm 0.72\%$ | $76.42\%\pm 1.01\%$
CelebA | $3\times 64\times 64$ | 0.05 | MobileNetV1 | $94.19\%\pm 1.63\%$ | $90.79\%\pm 0.49\%$ | $77.74\%\pm 0.22\%$ | $79.72\%\pm 0.22\%$
Table 2. Evaluation metrics for the quality of detected AEs and DL models
Metrics | Meanings
---|---
AE Prop. | Proportion of test seeds from which AEs can be detected over the total number of test seeds
Pred. Loss | Adversarial strength of AEs as formally defined by Definition 3
$p_{g}$ | Normalised global probability density of test-seeds/AEs
$\mathcal{R}_{l}$ | Local robustness to the correct classification label, as formally defined by Definition 1
$\hat{\mathcal{R}_{g}}$ | Empirical global robustness of DL models over input domain as defined in Definition 5
FID | Distribution difference between original images (test seeds) and perturbed images (AEs)
$\epsilon$ | Average perturbation distance between test seeds and AEs
% of Valid AEs | Percentage of “in-distribution” AEs in all detected AEs
All experiments were run on a machine of Ubuntu 18.04.5 LTS x86_64 with Nvidia
A100 GPU and 40G RAM. The source code, DL models, datasets and all experiment
results are publicly available at https://github.com/havelhuang/HDA-Testing.
### 4.2. Evaluation Results and Discussions
#### 4.2.1. RQ1
There are 3 sets of experiments in RQ1 to examine the accuracy of technical
solutions in our tool-chain, corresponding to the 3 main stages respectively.
First, to approximate the global distribution, we essentially proceed in two
steps—dimensionality reduction and PDF fitting, for which we adopt the VAE-
Encoder+KDE solution. Notably, the VAE trained in this step is for data-
compression only (not for generating new data). To reflect the effectiveness
of both aforementioned steps, we (i) compare VAE-Encoder with the Principal
Component Analysis (PCA), and (ii) measure the FID between the training
dataset and a set of random samples drawn from the fitted global distribution
by KDE.
PCA is a common approach for dimensionality reduction. We use scikit-learn
(Pedregosa et al., 2011) to implement the PCA with ’auto’ solver for applying
Singular Value Decomposition (SVD). That is, if the input data is larger than
500x500 and the number of components to extract (latent dimensions) is lower
than 80% of the smallest dimension of the data, then the more efficient
‘randomised’ method (Pedregosa et al., 2011) is enabled. Otherwise the exact
full SVD is computed and optionally truncated afterwards. To achieve the fair
comparison, the latent dimensions of PCA and VAE-Encoder are set to be the
same. We compare the performance of VAE-Encoder and PCA from the following two
perspectives. The quality of latent representation can be measured by the
clustering and reconstruction accuracy.
To learn the feature level (global) distribution from latent data, we require
that latent representations should group together data points that share
similar semantic features. To evaluate this clustering ability, we apply
K-means clustering to the latent data, which partitions the data points into
clusters based on their similarity. Then, we calculate the Completeness Score
(CS), Homogeneity Score (HS) and V-measure Score (VS) (Rosenberg and
Hirschberg, 2007). These scores provide a measure of how well the resulting
clusters group together data points that share similar semantic features.
Specifically, the CS measures how well the clustering captures all data points
that belong to the same true class in a single cluster, while the HS measures
how well the clustering captures all data points within a cluster that belong
to the same true class. The VS is a harmonic mean of the CS and HS, providing
an overall measure of the quality of the clustering.
In addition to ensuring that latent representations group together similar
data points, we also require that the latent representations can be decoded to
reconstruct the original images with minimal information loss. The
reconstruction loss is calculated based on the MSE. As is shown in Table 3,
VAE-Encoder achieves higher CS, HS, VS scores and less reconstruction loss
than PCA. In other words, the latent representations encoded by VAE-Encoder is
better in terms of capturing feature information than that of PCA.
Table 3. Quality of Latent Representation in PCA & VAE-Encoder
Dataset | Latent Dimensions | PCA | VAE-Encoder
---|---|---|---
Clustering | Recon. Loss | Clustering | Recon. Loss
CS | HS | VS | CS | HS | VS
MNIST | 8 | 0.505 | 0.508 | 0.507 | 44.09 | 0.564 | 0.566 | 0.565 | 27.13
F.-MNIST | 4 | 0.497 | 0.520 | 0.508 | 55.56 | 0.586 | 0.601 | 0.594 | 23.72
SVHN | 4 | 0.007 | 0.007 | 0.007 | 65.75 | 0.013 | 0.011 | 0.015 | 62.38
CIFAR-10 | 8 | 0.084 | 0.085 | 0.085 | 188.22 | 0.105 | 0.105 | 0.105 | 168.44
CelebA | 32 | 0.112 | 0.092 | 0.101 | 764.94 | 0.185 | 0.150 | 0.166 | 590.54
Dataset | Global Dist. | Uni. Dist.
---|---|---
MNIST | 0.395 | 13.745
Fashion-MNIST | 0.936 | 90.235
SVHN | 0.961 | 141.134
CIFAR-10 | 0.285 | 12.053
CelebA | 0.231 | 8.907
[table]..
Figure 4. Samples drawn from the approximated global distribution by KDE and a
uniform distribution over the latent feature space (Figure); and FID to the
ground truth based on 1000 samples (Table).
To evaluate the accuracy of using KDE to fit the global distribution, we
calculate the FID between a new dataset (with 1000 samples) based on the
fitted global distribution by KDE and the training dataset. The new dataset is
sampled from the fitted global distribution over latent space and decoded by
VAE decoder into images. The FID scores are shown in Table 4. As a baseline,
we also present the results of using a uniform distribution over the latent
space. As expected, we observe that all FID scores based on approximated
distributions are significantly smaller (better). We further decode the newly
generated images for visualisation in Fig. 4, from which we can see that
generated images by KDE keep high fidelity while the uniformly sampled images
are more difficult to be recognised.
Answer to RQ1 on HDA stage 1: The combination of VAE-Encoder+KDE may
accurately approximate the global distribution, since the new sampled data
from approximated distribution keep high fidelity.
Move on to stage 2, we study the correlations between a norm ball’s local
robustness and its two indicators proposed earlier—the prediction gradient
based score and the score based on separation distance of output-layer
activation (cf. Eq. 10).
Figure 5. Scatter plots of the local robustness evaluation vs. its two
indicators, based on 1000 random norm balls. Table 4. Pearson correlation
coefficients (in absolute values) between the local robustness & its two
indicators.
Dataset | $S_{\textit{grad}}$ | $S_{\textit{sep}}$ | $T(S_{\textit{grad}},S_{\textit{sep}})$
---|---|---|---
$t$ | $p-value$
MNIST | $\mathbf{0.631\pm 0.025}$ | $0.564\pm 0.019$ | $5.488$ | $5.813\times 10^{-4}$
Fashion-MNIST | $0.717\pm 0.109$ | $\mathbf{0.789\pm 0.056}$ | $-1.793$ | $0.107$
SVHN | $\mathbf{0.747\pm 0.039}$ | $0.745\pm 0.030$ | $0.150$ | $0.884$
CIFAR-10 | $0.603\pm 0.038$ | $\mathbf{0.668\pm 0.065}$ | $-5.466$ | $3.975\times 10^{-4}$
CelebA | $0.639\pm 0.073$ | $\mathbf{0.728\pm 0.077}$ | $-3.872$ | $0.004$
We invoke the tool (Webb et al., 2019) for estimating the local robustness
$\mathcal{R}_{l}$ defined in Definition 1. Based on 1000 randomly selected
data-points from the test set as the central point of 1000 norm balls, we
calculate the local robustness of each norm ball888Radius $r$ is usually small
by definition (cf. Remark 1), yielding very small $log{(1-\mathcal{R}_{l})}$.
as well as the two proposed indicators. Then, we do the scatter plots (in log-
log scale999There are dots collapsed on the vertical line of $log{(1-R)}=-70$,
due to a limitation of the estimator (Webb et al., 2019)—it terminates with
the specified threshold when the estimation is lower than that value. Note,
the correlation calculated with such noise is not undermining our conclusion,
rather the real correlation would be even higher.), as shown in Fig. 5.
Apparently, for all 5 datasets, the indicator based on activation separation
distance is negatively correlated (1st row), while the gradient based
indicator is positively correlated with the estimated local robustness (2nd
row). We further quantify the correlation by calculating the Pearson
correlation coefficients, as recorded in Table 4. There is a rule of thumb
that Pearson correlation coefficients greater than 0.6 indicate strong
correlations (Akoglu, 2018). We observe, both indicators are highly correlated
with the local robustness, while the separation distance based indicator is
slightly better. The statistical test shows that the separation distance based
indicator is significantly better than gradient based indicator in CIFAR10 and
CelebA datasets. Therefore, for the latter experiment in RQ2-RQ4, we choose
separation distance based indicator $S_{\textit{sep}}$ to guide the selection
of test seeds when comparing HDA testing with other AE detection methods.
Answer to RQ1 on HDA stage 2: The two proposed local robustness indicators are
significantly correlated with the local robustness.
Figure 6. The prediction loss (red) and the three quantified local
distribution indicators (blue) of the best fitted test case during the
iterations of our two-step GA based local test case generation.
For the local test case generation in stage 3, by configuring the parameter
$\alpha$ in our two-step GA, we may do trade-off between the “strength of
being adversarial” (measured by prediction loss $\mathcal{J}$) and the local
distribution (measured by a specific perceptual quality metric $\mathcal{L}$,
they are MSE, PSNR and SSIM), so that the quality of detected AEs can be
optimised.
In Fig. 6, we visualise the changes of the two fitness values as the
iterations of the GA. As shown in the first plot, only the prediction loss
$\mathcal{J}$ is taken as the fitness function (i.e., $\alpha=0$) during the
whole iteration process. The GA can effectively find AEs with maximised
adversarial strength, as observed by the convergence of the prediction loss of
the best-fitted test case in the population after hundreds of iterations. From
the second to the last plot, the fitness function consists not only of
prediction loss $\mathcal{J}$, but also of a perceptual quality metric
$\mathcal{L}$, representing the local distribution information (i.e.,
$\alpha>0$). Intuitively, a smaller MSE or greater PSNR and SSIM implies
higher local probability density.
Thanks to the two-step setting of the fitness functions, the prediction loss
$\mathcal{J}$ of best-fitted test case goes over 0 quickly in less than 200
iterations, which means it detects a first AE in the population. The
$\mathcal{J}$ of the best fitted test case is always quite close to the rest
in the population, thus we may confidently claim that many AEs are efficiently
detected by the population not long after the first AE was detected. Then, the
optimisation goes to the second stage, in which the quantified local
distribution indicator $\mathcal{L}$ is pursued. The $\mathcal{J}$ and
$\mathcal{L}$ finally converge and achieve a balance between them. If we
configure the coefficient $\alpha$, the balance point will change
correspondingly. A greater $\alpha$ (e.g., $\alpha=1.1$ in the plots) detects
less perceptible AEs (i.e., with higher local probability density), and the
price paid is that the detected AEs are with weaker adversarial strength
(i.e., with smaller but still positive prediction loss).
Figure 7. Comparison between regular GA and two-step GA. Figure 8. AEs
detected by our two-step GA (last 3 columns) & other methods
We further investigate the advantages of our 2-step GA over the regular GA
(using $F_{2}$ as the objective function). In Fig. 7, as $\alpha$ increases,
the proportion of AEs in the population exhibits a sharp drop to 0 when using
the regular GA. In contrast, the two-step GA prevents such decreasing of the
AE proportion while preserving it at a high-level of 0.6, even when $\alpha$
is quite large. Moreover, larger $\alpha$ represents the situations when the
AEs are less perceptible in terms of perceptual quality metrics—as shown by
the blue curves101010The blue dashed line stops earlier as there is no AEs in
the population when $\alpha$ is big., the imperceptibility ( measured by SSIM
in this case) is only sufficiently high when $\alpha$ is big enough. Thus,
compared to the regular GA, we may claim our novel 2-step GA is more robust
(in detecting AEs) to the choices of $\alpha$ and more suitable in our
framework for detecting AEs with high perceptual quality.
Table 5. Perceptual quality measured by FID between a set of original images
and a set of AEs detected by two-step GA with different fitness functions.
Dataset | $J$ | $J+MSE$ | $J+PSNR$ | $J+SSIM$
---|---|---|---|---
MNIST | $1.185\pm 0.157$ | $\mathbf{0.667\pm 0.202}$ | $0.668\pm 0.203$ | $0.875\pm 0.184$
Fashion-MNIST | $2.736\pm 0.680$ | $0.138\pm 0.106$ | $\mathbf{0.131\pm 0.100}$ | $0.679\pm 0.182$
SVHN | $120.172\pm 7.529$ | $105.399\pm 9.009$ | $\mathbf{104.147\pm 9.934}$ | $111.139\pm 6.754$
CIFAR-10 | $96.449\pm 5.906$ | $\mathbf{65.727\pm 5.672}$ | $67.426\pm 7.398$ | $76.293\pm 6.493$
CelebA | $85.658\pm 3.334$ | $\mathbf{65.981\pm 4.976}$ | $66.599\pm 3.312$ | $71.878\pm 3.153$
(a) Results
Dataset | $T(J+MSE,J)$ | $T(J+MSE,J+PSNR)$ | $T(J+MSE,J+SSIM)$
---|---|---|---
$t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$
MNIST | $-13.631$ | $2.581\times 10^{-7}$ | $-0.912$ | $0.385$ | $-10.279$ | $2.844\times 10^{-6}$
Fashion-MNIST | $-13.362$ | $3.065\times 10^{-7}$ | $2.188$ | $0.056$ | $-11.957$ | $7.934\times 10^{-7}$
SVHN | $-9.214$ | $7.036\times 10^{-6}$ | $1.271$ | $0.235$ | $-4.443$ | $0.002$
CIFAR-10 | $-17.32$ | $3.215\times 10^{-8}$ | $-1.363$ | $0.206$ | $-8.125$ | $1.954\times 10^{-5}$
CelebA | $-23.491$ | $2.187\times 10^{-9}$ | $-0.511$ | $0.621$ | $-3.931$ | $0.003$
(b) Statistical Test
Fig. 8 displays some selected AEs from the five datasets. Same as the PGD
attack and the coverage-guided testing, if we only use the prediction loss
$\mathcal{J}$ as the objective function in the GA, the perturbations added to
the images can be easily recognised. In stark contrast, AEs generated by our
two-step GA (with the 3 perceptual quality metrics in the last 3 columns) are
of high quality and less distinguishable from the original images (first
column). We further calculate the FID to quantify the perceptual quality of
detected AEs in Table 5(b). Since the comparison with PGD attack, coverage
guided testing, FODA, and OODA are given in the subsequent experiments, i.e.
Table 7(b) and Table 8(b), we focus on comparing the performance of 3
perceptual quality metrics here. Results show that MSE is significantly better
than J and SSIM to guide the generation of high fidelity AEs for the
experiment dataset. While MSE has similar performance with PSNR. Therefore, we
decide to utilise the J+MSE in the following experiments for the comparison
with the state-of-the-art.
Answer to RQ1 on HDA stage 3: Two-step GA based local test case generation can
effectively detect AEs with high perception quality (in terms of those metrics
encoded by GA).
#### 4.2.2. RQ2
We compare HDA with the state-of-the-art AE detection methods in two sets of
experiments. In the first set, we focus on comparing with the adversarial
attack and coverage-guided testing (i.e., the typical PGD attack and neuron
coverage metric for brevity, while the conclusion can be generalised to other
attacks and coverage metrics, since they all lack the consideration of
distribution when detecting AEs). Then in the second set of experiments, we
show the advantages of our HDA testing over other distribution-aware testing
methods.
In fact, both PGD attack and coverage-guided testing do not contribute to test
seeds selection. They simply use randomly sampled data from the test set as
test seeds, by default. We also notice that a large amount of test seeds
prioritisation metrics are proposed, the typical ones among which are
Distance-Based Surprise Adequacy (DSA) (Weiss and Tonella, 2022), DeepGini
(Weiss and Tonella, 2022). Thus, we compare the randomly selected test seeds,
DSA guided test seeds, DeepGini guided test seeds with our “global
distribution probability111111Refer to Section 3.2 for the calculation. The
value of probability density is further normalised by training dataset for a
better presentation. plus local robustness indicated” test seeds, shown as
“‘$p_{g}+\mathcal{R}_{l}$” in Table 6(b). Specifically, for each test seed, we
calculate two metrics—the local robustness $\mathcal{R}_{l}$ of its norm ball
and its corresponding global probability $p_{g}$. We invoke the estimator of
(Webb et al., 2019) to calculate the former ($log{(1-\mathcal{R}_{l})}$, to be
exact). To reduce the sampling noise, we repeat the test seed selection 100
times and present the averaged results in Table 6(b).
Table 6. Comparison between randomly selected test seeds, DSA guided test
seeds, DeepGini guided test seeds and our “$p_{g}+\mathcal{R}_{l}$ indicated”
test seeds (based on 100 test seeds).
Dataset | Random Test Seeds | DSA Test Seeds | DeepGini Test Seeds | $p_{g}+\mathcal{R}_{l}$ Test Seeds
---|---|---|---|---
$log{(1-\mathcal{R}_{l})}$ | $p_{g}$ | $log{(1-\mathcal{R}_{l})}$ | $p_{g}$ | $log{(1-\mathcal{R}_{l})}$ | $p_{g}$ | $log{(1-\mathcal{R}_{l})}$ | $p_{g}$
MNIST | $-64.67\pm 1.27$ | $0.0034\pm 0.0005$ | $-64.59\pm 1.15$ | $0.0016\pm 0.0003$ | $-65.16\pm 1.31$ | $0.0018\pm 0.0002$ | $\mathbf{-22.12\pm 3.01}$ | $\mathbf{0.0293\pm 0.0072}$
Fashion-MNIST | $-14.98\pm 4.50$ | $0.0033\pm 0.0009$ | $-15.74\pm 4.82$ | $0.0016\pm 0.0003$ | $-15.29\pm 4.89$ | $0.0017\pm 0.0003$ | $\mathbf{-1.24\pm 0.48}$ | $\mathbf{0.0226\pm 0.0088}$
SVHN | $-53.34\pm 3.98$ | $0.0032\pm 0.0002$ | $-53.44\pm 2.10$ | $0.0013\pm 0.0002$ | $-53.62\pm 2.32$ | $0.0008\pm 0.0001$ | $\mathbf{-8.27\pm 2.43}$ | $\mathbf{0.0132\pm 0.0021}$
CIFAR-10 | $-13.77\pm 1.24$ | $0.0033\pm 0.0009$ | $-14.25\pm 2.26$ | $0.0018\pm 0.0003$ | $-14.82\pm 0.86$ | $0.0018\pm 0.0003$ | $\mathbf{-2.49\pm 0.88}$ | $\mathbf{0.0397\pm 0.0121}$
CelebA | $-17.68\pm 5.00$ | $0.0034\pm 0.0002$ | $-17.27\pm 5.44$ | $0.0010\pm 0.0001$ | $-17.28\pm 5.37$ | $0.0005\pm 0.0001$ | $\mathbf{-1.59\pm 0.54}$ | $\mathbf{0.0118\pm 0.0005}$
(a) Results
Dataset | Metric | $T(p_{g}+\mathcal{R}_{l},\text{Random})$ | $T(p_{g}+\mathcal{R}_{l},\text{DSA})$ | $T(p_{g}+\mathcal{R}_{l},\text{DeepGini})$
---|---|---|---|---
$t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$
MNIST | $log{(1-\mathcal{R}_{l})}$ | $41.186$ | $2.890\times 10^{-19}$ | $41.680$ | $2.337\times 10^{-19}$ | $41.461$ | $2.568\times 10^{-19}$
$p_{g}$ | $11.348$ | $1.234\times 10^{-9}$ | $12.155$ | $4.101\times{-10}$ | $12.073$ | $4.574\times 10^{-10}$
Fashion-MNIST | $log{(1-\mathcal{R}_{l})}$ | $9.601$ | $1.665\times 10^{-8}$ | $9.466$ | $2.063\times 10^{-8}$ | $9.042$ | $4.106\times 10^{-8}$
$p_{g}$ | $6.899$ | $1.885\times 10^{-6}$ | $7.542$ | $5.616\times 10^{-7}$ | $7.506$ | $6.001\times 10^{-7}$
SVHN | $log{(1-\mathcal{R}_{l})}$ | $30.564$ | $5.769\times 10^{-17}$ | $44.475$ | $7.345\times 10^{-20}$ | $42.686$ | $1.528\times 10^{-19}$
$p_{g}$ | $14.991$ | $1.303\times 10^{-11}$ | $17.839$ | $6.867\times 10^{-13}$ | $18.651$ | $3.203\times 10^{-13}$
CIFAR-10 | $log{(1-\mathcal{R}_{l})}$ | $23.459$ | $6.033\times 10^{-15}$ | $15.334$ | $8.916\times 10^{-12}$ | $31.688$ | $3.045\times 10^{-17}$
$p_{g}$ | $9.487$ | $1.997\times 10^{-8}$ | $9.902$ | $1.039\times 10^{-8}$ | $9.902$ | $1.039\times 10^{-8}$
CelebA | $log{(1-\mathcal{R}_{l})}$ | $10.117$ | $7.461\times 10^{-9}$ | $9.071$ | $3.922\times 10^{-8}$ | $9.193$ | $3.207\times 10^{-8}$
$p_{g}$ | $49.326$ | $1.156\times 10^{-20}$ | $66.979$ | $4.830\times 10^{-23}$ | $70.079$ | $2.145\times 10^{-23}$
(b) Statistical Test
From Table 6(b), we observe: (i) test seeds selected by our method have much
higher global probability density, meaning their norm balls are much more
representative of the data distribution; (ii) the norm balls of our test seeds
have worse local robustness, meaning it is more cost-effective to detect AEs
in them. Both metrics of HDA are significantly different from those of other
test seeds selection methods, as indicated by large t-scores and small
p-values. This is unsurprising because we have explicitly considered the
distribution and local robustness information in the test seed selection.
(iii) DSA and DeepGini guided test seeds are even worse than random test
seeds, since they target at prioritising the test seeds which are prone to be
misclassified. These misclassified test seeds are usually out of distribution.
Finally, the overall evaluation on the generated test cases and the detected
AEs by them are shown in Table 7(b). The results are presented in two
dimensions—3 types of testing methods versus 2 ways of test seeds selection,
yielding 6 combinations (although by default, PGD attack and coverage-guided
methods are using random seeds, while our method is using the
“$p_{g}+\mathcal{R}_{l}$” seeds). For each combination, we study 4 metrics
(cf. Table 2 for meanings behind them): (i) the AE proportion; (ii) the
average prediction loss; (iii) the FID121212To show how close the perturbed
test cases are to the test seeds in the latent space, we use the last
convolutional layer of InceptionV3 to extract the latent representations of
colour images for FID. InceptionV3 is a well-trained CNN and commonly used to
show FID that captures the perturbation levels, e.g., in (Heusel et al.,
2017). While InceptionV3 is used for colour images, VAE is used for grey-scale
datasets MNIST and Fashion-MNIST. of the test set quantifying the image
quality; and (iv) the computational time (and an additional coverage rate for
coverage-guided testing). We note the observations on these 4 metrics in the
following paragraphs.
Table 7. Evaluation of the generated test cases and detected AEs by PGD
Attack, coverage-guided testing and the proposed HDA testing (all results are
averaged over 100 seeds). HDA can detect more high perception quality AEs by
$p_{g}$+$\mathcal{R}_{l}$ seeds selection, which is evidenced by high AE Prop.
and low FID.
| AE Detection
---
Method
Test Seeds | Metric | MNIST | F.-MNIST | SVHN | CIFAR-10 | CelebA
PGD Attack | Random Seeds | AE Prop. | $\mathbf{0.422\pm 0.046}$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{0.951\pm 0.031}$ | $\mathbf{1.000\pm 0.000}$ | $0.989\pm 0.035$
Pred. Loss | $\mathbf{6.362\pm 1.089}$ | $\mathbf{29.200\pm 15.754}$ | $\mathbf{6.911\pm 0.571}$ | $\mathbf{46.099\pm 4.037}$ | $\mathbf{46.502\pm 48.154}$
FID | $0.705\pm 0.047$ | $2.938\pm 0.249$ | $106.581\pm 1.275$ | $96.149\pm 2.265$ | $88.401\pm 1.459$
Time(s) | $\mathbf{0.038\pm 0.070}$ | $\mathbf{0.035\pm 0.013}$ | $\mathbf{9.632\pm 0.274}$ | $\mathbf{9.294\pm 0.532}$ | $\mathbf{8.998\pm 0.646}$
$p_{g}$+$\mathcal{R}_{l}$ Seeds | AE Prop. | $\mathbf{0.785\pm 0.103}$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{0.989\pm 0.006}$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{1.000\pm 0.000}$
Pred. Loss | $\mathbf{10.184\pm 1.135}$ | $\mathbf{36.139\pm 18.205}$ | $\mathbf{9.385\pm 0.763}$ | $\mathbf{44.816\pm 3.221}$ | $\mathbf{52.136\pm 58.991}$
FID | $0.696\pm 0.051$ | $1.351\pm 0.208$ | $103.269\pm 1.153$ | $98.592\pm 1.187$ | $83.709\pm 1.095$
Time(s) | $\mathbf{0.031\pm 0.024}$ | $\mathbf{0.035\pm 0.014}$ | $\mathbf{9.541\pm 0.499}$ | $\mathbf{9.739\pm 0.763}$ | $\mathbf{9.473\pm 0.529}$
Cov. Guided Testing | Random Seeds | Cov. Rate | $0.980\pm 0.002$ | $0.946\pm 0.023$ | $0.966\pm 0.002$ | $0.979\pm 0.004$ | $0.985\pm 0.002$
AE Prop. | $0.0002\pm 0.0006$ | $0.153\pm 0.021$ | $0.014\pm 0.005$ | $0.150\pm 0.016$ | $0.080\pm 0.033$
Pred. Loss | $0.019\pm 0.001$ | $2.404\pm 1.512$ | $0.230\pm 0.045$ | $3.732\pm 1.098$ | $2.841\pm 4.311$
FID | $0.802\pm 0.026$ | $3.771\pm 0.357$ | $101.034\pm 1.014$ | $87.899\pm 1.807$ | $85.918\pm 1.109$
Time(s) | $248.902\pm 11.088$ | $1189.87\pm 223.93$ | $4416.95\pm 144.93$ | $6678.35\pm 690.49$ | $972.17\pm 83.57$
$p_{g}$+$\mathcal{R}_{l}$ Seeds | Cov. Rate | $0.977\pm 0.003$ | $0.915\pm 0.025$ | $0.921\pm 0.008$ | $0.978\pm 0.004$ | $0.978\pm 0.002$
AE Prop. | $0.030\pm 0.017$ | $0.485\pm 0.068$ | $0.138\pm 0.030$ | $0.449\pm 0.049$ | $0.286\pm 0.056$
Pred. Loss | $1.184\pm 0.632$ | $2.867\pm 1.244$ | $0.198\pm 0.039$ | $3.607\pm 0.459$ | $2.137\pm 2.283$
FID | $0.804\pm 0.043$ | $1.816\pm 0.285$ | $94.043\pm 1.786$ | $90.491\pm 1.071$ | $81.243\pm 1.162$
Time(s) | $81.33\pm 7.09$ | $216.59\pm 152.62$ | $3073.97\pm 598.92$ | $1822.64\pm 102.24$ | $2065.70\pm 32.68$
HDA | Random Seeds | AE Prop. | $0.208\pm 0.042$ | $0.999\pm 0.003$ | $0.843\pm 0.046$ | $1.000\pm 0.000$ | $\mathbf{1.000\pm 0.000}$
Pred. Loss | $1.076\pm 0.341$ | $5.214\pm 3.325$ | $2.784\pm 0.198$ | $32.251\pm 2.287$ | $16.020\pm 15.306$
FID | $\mathbf{0.094\pm 0.009}$ | $\mathbf{0.119\pm 0.043}$ | $\mathbf{91.431\pm 1.118}$ | $\mathbf{62.587\pm 2.144}$ | $\mathbf{59.162\pm 1.469}$
Time(s) | $74.211\pm 1.71$ | $136.652\pm 22.71$ | $153.956\pm 62.92$ | $351.992\pm 74.62$ | $191.313\pm 68.35$
$p_{g}$+$\mathcal{R}_{l}$ Seeds | AE Prop. | $0.676\pm 0.127$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{0.989\pm 0.006}$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{1.000\pm 0.000}$
Pred. Loss | $1.591\pm 0.306$ | $10.020\pm 5.961$ | $3.925\pm 0.433$ | $33.361\pm 1.873$ | $20.702\pm 21.941$
FID | $\mathbf{0.042\pm 0.011}$ | $\mathbf{0.016\pm 0.004}$ | $\mathbf{91.102\pm 1.342}$ | $\mathbf{65.829\pm 2.154}$ | $\mathbf{55.478\pm 1.399}$
Time(s) | $204.002\pm 85.96$ | $61.793\pm 1.69$ | $237.882\pm 98.26$ | $708.832\pm 269.81$ | $305.499\pm 82.47$
(a) Results
Dataset | Metric | Random Seeds | $p_{g}$+$\mathcal{R}_{l}$ Seeds
---|---|---|---
$T(\text{HDA},\text{PGD})$ | $T(\text{HDA},\text{Cov.})$ | $T(\text{HDA},\text{PGD})$ | $T(\text{HDA},\text{Cov.})$
$t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$
MNIST | AE Prop. | $-10.864$ | $2.458\times 10^{-9}$ | $15.644$ | $6.365\times 10^{-12}$ | $-2.108$ | $0.049$ | $15.943$ | $4.627\times 10^{-12}$
FID | $-40.376$ | $4.118\times 10^{-19}$ | $-81.374$ | $1.469\times 10^{-24}$ | $-39.640$ | $5.715\times 10^{-19}$ | $-54.290$ | $2.080\times 10^{-21}$
Fashion-MNIST | AE Prop. | $-1.054$ | $0.306$ | $126.114$ | $5.595\times 10^{-28}$ | $nan$ | $nan$ | $23.949$ | $4.204\times 10^{-15}$
FID | $-35.279$ | $4.541\times 10^{-18}$ | $-32.117$ | $2.400\times 10^{-17}$ | $-20.293$ | $7.495\times 10^{-14}$ | $-19.970$ | $9.882\times 10^{-14}$
SVHN | AE Prop. | $-6.157$ | $8.186\times 10^{-6}$ | $56.656$ | $9.695\times 10^{-22}$ | $nan$ | $nan$ | $87.961$ | $3.629\times 10^{-25}$
FID | $-28.252$ | $2.312\times 10^{-16}$ | $-20.119$ | $8.690\times 10^{-14}$ | $-21.746$ | $2.259\times 10^{-14}$ | $-4.163$ | $5.843\times 10^{-4}$
CIFAR-10 | AE Prop. | $nan$ | $nan$ | $167.996$ | $3.221\times 10^{-30}$ | $nan$ | $nan$ | $35.559$ | $3.945\times 10^{-18}$
FID | $-34.030$ | $8.609\times 10^{-18}$ | $-28.547$ | $1.925\times 10^{-16}$ | $-42.126$ | $1.933\times 10^{-19}$ | $-32.419$ | $2.033\times 10^{-17}$
CelebA | AE Prop. | $0.994$ | $0.333$ | $88.160$ | $3.485\times 10^{-25}$ | $nan$ | $nan$ | $40.319$ | $4.223\times 10^{-19}$
FID | $-44.658$ | $6.825\times 10^{-20}$ | $-45.968$ | $4.073\times 10^{-20}$ | $-50.251$ | $8.295\times 10^{-21}$ | $-44.801$ | $6.449\times 10^{-20}$
(b) Statistical Test
Regarding the AE proportion in the set of generated test cases, the default
setting of our proposed approach ($p_{g}$+$\mathcal{R}_{l}$ Seeds with HDA)
has comparable performance with PGD attack, as indicated by identical results
and $nan$ output by statistical test. Both our novel $p_{g}$+$\mathcal{R}_{l}$
test seed selection and two-step GA local test case generation methods
contribute to the comparable performance. This is evident from the decreased
AE proportion when using random seeds in our method, but still the result is
relatively higher than most combinations. PGD attack, as a white-box approach
using the gradient information, is definitely quite efficient in detecting
AEs, especially when paired with our new test seed selection method. On the
other hand, coverage-guided testing is comparatively less effective in
detecting AEs (even with high coverage rate), but our test seed selection
method can improve it.
As per the results of prediction loss, PGD attack, as a gradient based attack
method, unsurprisingly finds the AEs with the largest prediction loss. With
better test seeds considering local robustness information by our method, the
prediction loss of PGD attack can be even higher. Both coverage-guided testing
and our HDA testing can detect AEs with relatively lower prediction loss,
meaning the AEs are with “weaker adversarial strength”. The reason for the low
prediction loss of AEs detected by our approach is that two-step GA makes the
trade-off and sacrifices it for higher local probabilities (i.e., perceptual
quality). This can be seen through the significantly smaller FID of test set
generated by HDA, compared with PGD attack and coverage-guided testing. PGD
attack has relatively high FID scores, as well as coverage-guided testing.
On the computational overheads, we observe PGD is the most efficient, given it
is by nature a white-box approach using the gradient information. While, our
approach is an end-to-end black-box approach (if without using the gradient
based indicator when selecting test seeds) requiring less information and
being more generic, at the price of being relatively less efficient. That
said, the computational time of our approach is still acceptable and better
than coverage-guided testing.
Answer to RQ2 on comparing with adversarial attack and coverage-guided
testing: HDA can select more significant test seeds, which are from high
probability density region and lack of robustness. HDA can generate higher
perception quality AEs, which are measured with smaller FID values.
Table 8. Evaluation of AEs detected by OODA, FODA and our HDA testing methods
(based on 100 test seeds).
Dataset | | AE Detection
---
Method
$p_{g}$ | | % of
---
Valid AEs
$\epsilon$ | FID
MNIST | OODA | $0.0018\pm 0.0002$ | $22.5\pm 8.3$ | $0.917\pm 0.031$ | $3.463\pm 0.729$
FODA | $0.0032\pm 0.0006$ | $98.5\pm 0.4$ | $0.558\pm 0.009$ | $0.158\pm 0.017$
HDA | $\mathbf{0.0292\pm 0.0071}$ | $\mathbf{99.3\pm 0.5}$ | $\mathbf{0.076\pm 0.003}$ | $\mathbf{0.042\pm 0.011}$
SVHN | OODA | $0.0026\pm 0.0004$ | $11.3\pm 1.2$ | $0.811\pm 0.009$ | $125.126\pm 2.838$
FODA | $0.0034\pm 0.0002$ | $100\pm 0$ | $0.252\pm 0.012$ | $111.694\pm 1.461$
HDA | $\mathbf{0.0132\pm 0.0021}$ | $\mathbf{100\pm 0}$ | $\mathbf{0.029\pm 0.001}$ | $\mathbf{91.102\pm 1.342}$
(a) Results
Dataset | Metric | $T(\text{HDA},\text{OODA})$ | $T(\text{HDA},\text{FODA})$
---|---|---|---
$t$ | $p-value$ | $t$ | $p-value$
MNIST | $p_{g}$ | $12.199$ | $3.871\times 10^{-10}$ | $11.539$ | $9.456\times 10^{-10}$
% of Valid AEs | $29.208$ | $1.286\times 10^{-16}$ | $3.951$ | $9.368\times 10^{-4}$
$\epsilon$ | $-85.391$ | $6.183\times 10^{-25}$ | $-160.667$ | $7.186\times 10^{-30}$
FID | $-14.838$ | $1.546\times 10^{-11}$ | $-18.116$ | $5.275\times 10^{-13}$
SVHN | $p_{g}$ | $15.680$ | $6.124\times 10^{-12}$ | $14.691$ | $1.826\times 10^{-11}$
% of Valid AEs | $233.745$ | $8.456\times 10^{-33}$ | $nan$ | $nan$
$\epsilon$ | $-273.086$ | $5.146\times 10^{-34}$ | $-58.563$ | $5.359\times 10^{-22}$
FID | $-34.273$ | $7.588\times 10^{-18}$ | $-32.825$ | $1.632\times 10^{-17}$
(b) Statistical Test
Next, we try to answer the difference between HDA testing and other
distribution-aware testing as summarised earlier (the amber route of Fig. 1).
We not only study the common evaluation metrics in earlier RQs, but also the
input validation method in (Dola et al., 2021), which flags the validity of
AEs according to a user-defined reconstruction probability threshold. Overall,
HDA is significantly better than OODA and FODA in four evaluation metrics, as
observed from the results of statistical test.
As shown in Table 8(b), HDA can select test seeds from much higher density
region on the global distribution and find more valid AEs than OODA. The
reason behind this is that OODA aims at detecting outliers—only AEs with lower
reconstruction probabilities (from the test seed) than the given threshold
will be marked as invalid test cases. While, HDA explicitly explores the high
density meanwhile error-prone regions by combining the global distribution and
local robustness indicators. In other words, HDA performs priority ordering
(according to the global distribution and local robustness) and then selects
the best, while OODA rules out the worst. As expected, FODA performs similarly
poorly as OODA in terms of $p_{g}$, since both use randomly selected seeds.
However, FODA has high proportion of valid AEs since the test cases are
directly sampled from the distribution in latent space.
Regarding the perceptual quality of detected AEs, HDA can always find AEs with
small pixel-level perturbations ($\epsilon$) in consideration of the
$r$-separation constraint, and with small FID thanks to the use of perceptual
quality metrics (MSE in this case) as objective functions. While OODA only
utilises the reconstruction probability (from VAE) to choose AEs, and FODA
directly samples test cases from VAE without any restrictions (thus may suffer
from the oracle problem, cf. Remark 13 later). Due to the compression nature
of generative models—they are good at extracting feature level information but
ignore pixel level information (Zhong et al., 2020), AEs detected by OODA and
FODA are all distant to the original test seeds, yielding large $\epsilon$ and
FID scores. Notably, the average distance $\epsilon$ between test seeds and
AEs detected by OODA and FODA are much (7$\sim$28 times) greater than the
$r$-separation constraints (cf. Table 1), leading to the potential oracle
issues of those AEs, for which we have the following remark:
###### Remark 5 (Oracle Issues of AEs Detected by OODA and FODA).
AEs detected by OODA and FODA are normally distant to the test seeds with a
perturbation distance even greater than the $r$-separation constraint.
Consequently, there is the risk that the perturbed image may not share the
same ground truth label of the test seed, and thus hard to determine the
ground truth label of the “AE”131313In quotes, because the perturbed image
could be a “benign example” with a correct predicted label (but different to
the test seed)..
Figure 9. Example AEs detected by different distribution-aware testing
methods. AEs detected by our HDA are indistinguishable from the original
images, while AEs detected by FODA and OODA are of low perceptual quality and
subject to the oracle issues noted by Remark 13.
To visualise the difference between AEs detected by HDA, FODA and OODA, we
present 4 examples in Fig. 9. We may observe the AEs detected by HDA are
almost indistinguishable from the original images. Moreover, the AEs by FODA
is a set of concrete evidence for Remark 13—it is actually quite hard to tell
what is the ground truth label of some perturbed image (e.g., the bottom left
one), while others appear to have a different label of the seed (e.g., the
bottom right one should be with a label “1” instead of “7”).
Answer to RQ2 on comparing with other distribution-aware testing: Compared to
OODA and FODA, the proposed HDA testing can detect more valid AEs, free of
oracle issues, with higher global probabilities and perception quality.
#### 4.2.3. RQ3
In earlier RQs, we have varied the datasets and model architectures to check
the effectiveness of HDA. In this RQ3, we are concern about HDA’s sensitivity
to DL models with different levels of robustness. Adversarial training may
greatly improve the robustness of DL models and is widely used as the defence
to adversarial attack. To this end, we apply HDA on both normally and
adversarially trained models (by (Madry et al., 2018) to be exact), and then
compare with three most representative adversarial attack methods—the most
classic FGSM, the most popular PGD, and the most advanced AutoAttack (Croce
and Hein, 2020). Experimental results are presented in Table 9(b).
Table 9. Evaluation of AEs generated by FGSM, PGD, AutoAttack and HDA on
normally and adversarially trained DL models (all results are averaged over
100 test seeds).
Model | | AE Detection
---
Method
Eval. Metric | MNIST | F.-MNIST | SVHN | CIFAR-10 | CelebA
Normally Trained | FGSM | $p_{g}$ | $0.0045\pm 0.0005$ | $0.0034\pm 0.0008$ | $0.0031\pm 0.0002$ | $0.0028\pm 0.0004$ | $0.0034\pm 0.0002$
AE Prop. | $0.413\pm 0.093$ | $0.797\pm 0.039$ | $0.727\pm 0.069$ | $0.893\pm 0.016$ | $0.989\pm 0.035$
FID | $1.022\pm 0.053$ | $4.134\pm 0.322$ | $114.256\pm 2.931$ | $104.877\pm 1.832$ | $88.401\pm 1.459$
PGD | $p_{g}$ | $0.0040\pm 0.0005$ | $0.0034\pm 0.0008$ | $0.0032\pm 0.0002$ | $0.0028\pm 0.0004$ | $0.0034\pm 0.0002$
AE Prop. | $0.422\pm 0.046$ | $1.000\pm 0.000$ | $0.984\pm 0.011$ | $1.000\pm 0.000$ | $0.989\pm 0.035$
FID | $0.705\pm 0.047$ | $2.938\pm 0.248$ | $107.370\pm 1.206$ | $101.157\pm 1.774$ | $88.401\pm 1.459$
AutoAttack | $p_{g}$ | $0.0042\pm 0.0004$ | $0.0034\pm 0.0008$ | $0.0032\pm 0.0002$ | $0.0028\pm 0.0004$ | $0.0035\pm 0.0002$
AE Prop. | $\mathbf{0.787\pm 0.075}$ | $1.000\pm 0.000$ | $0.984\pm 0.011$ | $1.000\pm 0.000$ | $1.000\pm 0.000$
FID | $0.795\pm 0.099$ | $4.962\pm 0.359$ | $106.780\pm 1.297$ | $101.173\pm 1.543$ | $90.385\pm 1.305$
HDA | $p_{g}$ | $\mathbf{0.0292\pm 0.0071}$ | $\mathbf{0.0224\pm 0.0087}$ | $\mathbf{0.0132\pm 0.0021}$ | $\mathbf{0.0406\pm 0.012}$ | $\mathbf{0.0124\pm 0.001}$
AE Prop. | $0.676\pm 0.127$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{0.989\pm 0.006}$ | $\mathbf{1.000\pm 0.000}$ | $\mathbf{1.000\pm 0.000}$
FID | $\mathbf{0.042\pm 0.011}$ | $\mathbf{0.016\pm 0.004}$ | $\mathbf{91.102\pm 1.342}$ | $\mathbf{65.829\pm 2.154}$ | $\mathbf{55.478\pm 1.399}$
Adversarially Trained | FGSM | $p_{g}$ | $0.0042\pm 0.0005$ | $0.0033\pm 0.0004$ | $0.0036\pm 0.0004$ | $0.0028\pm 0.0005$ | $0.0031\pm 0.0003$
AE Prop. | $0.033\pm 0.021$ | $0.208\pm 0.037$ | $0.323\pm 0.061$ | $0.455\pm 0.039$ | $0.404\pm 0.045$
FID | $1.064\pm 0.055$ | $4.898\pm 0.503$ | $135.730\pm 4.321$ | $113.106\pm 3.989$ | $108.070\pm 4.368$
PGD | $p_{g}$ | $0.0038\pm 0.0005$ | $0.0032\pm 0.0004$ | $0.0034\pm 0.0004$ | $0.0028\pm 0.0005$ | $0.0032\pm 0.0003$
AE Prop. | $0.014\pm 0.017$ | $0.201\pm 0.039$ | $0.507\pm 0.049$ | $0.455\pm 0.039$ | $0.372\pm 0.036$
FID | $0.762\pm 0.040$ | $3.871\pm 0.407$ | $124.388\pm 2.613$ | $113.106\pm 3.989$ | $111.931\pm 2.539$
AutoAttack | $p_{g}$ | $0.0031\pm 0.0004$ | $0.0031\pm 0.0004$ | $0.0034\pm 0.0004$ | $0.0028\pm 0.0005$ | $0.0031\pm 0.0003$
AE Prop. | $0.049\pm 0.021$ | $0.268\pm 0.038$ | $0.505\pm 0.049$ | $0.419\pm 0.039$ | $0.405\pm 0.045$
FID | $0.035\pm 0.009$ | $0.128\pm 0.046$ | $125.382\pm 2.315$ | $117.257\pm 3.750$ | $107.742\pm 4.169$
HDA | $p_{g}$ | $\mathbf{0.0214\pm 0.0075}$ | $\mathbf{0.0118\pm 0.0032}$ | $\mathbf{0.0154\pm 0.0021}$ | $\mathbf{0.0261\pm 0.0112}$ | $\mathbf{0.0103\pm 0.0007}$
AE Prop. | $\mathbf{0.368\pm 0.071}$ | $\mathbf{0.903\pm 0.021}$ | $\mathbf{0.960\pm 0.029}$ | $\mathbf{0.865\pm 0.051}$ | $\mathbf{0.968\pm 0.012}$
FID | $\mathbf{0.029\pm 0.003}$ | $\mathbf{0.058\pm 0.010}$ | $\mathbf{92.389\pm 1.152}$ | $\mathbf{62.804\pm 2.398}$ | $\mathbf{58.558\pm 0.895}$
(a) Results
Model | Dataset | Metric | $T(\text{HDA},\text{FGSM})$ | $T(\text{HDA},\text{PGD})$ | $T(\text{HDA},\text{AutoAttack})$
---|---|---|---|---|---
$t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$
Normally Trained | MNIST | $p_{g}$ | $10.97$ | $2.10\times 10^{-9}$ | $11.20$ | $1.53\times 10^{-9}$ | $11.12$ | $1.71\times 10^{-9}$
AE Prop. | $5.28$ | $5.05\times 10^{-5}$ | $5.95$ | $1.26\times 10^{-5}$ | $-2.38$ | $0.03$
FID | $-57.25$ | $8.04\times 10^{-22}$ | $-43.43$ | $1.12\times 10^{-19}$ | $-23.91$ | $4.34\times 10^{-15}$
F.-MNIST | $p_{g}$ | $6.88$ | $1.97\times 10^{-6}$ | $6.88$ | $1.97\times 10^{-6}$ | $6.88$ | $1.97\times 10^{-6}$
AE Prop. | $16.46$ | $2.70\times 10^{-12}$ | $nan$ | $nan$ | $nan$ | $nan$
FID | $-40.44$ | $4.01\times 10^{-19}$ | $-37.25$ | $1.72\times 10^{-18}$ | $-43.56$ | $1.06\times 10^{-19}$
SVHN | $p_{g}$ | $15.14$ | $1.10\times 10^{-11}$ | $14.99$ | $1.30\times 10^{-11}$ | $14.99$ | $1.30\times 10^{-11}$
AE Prop. | $11.96$ | $5.31\times 10^{-10}$ | $1.26$ | $0.22$ | $1.26$ | $0.22$
FID | $-22.71$ | $1.05\times 10^{-14}$ | $-28.51$ | $1.97\times 10^{-16}$ | $-26.56$ | $6.84\times 10^{-16}$
CIFAR-10 | $p_{g}$ | $9.96$ | $9.56\times 10^{-9}$ | $9.96$ | $9.56\times 10^{-9}$ | $9.96$ | $9.56\times 10^{-9}$
AE Prop. | $21.15$ | $3.67\times 10^{-14}$ | $nan$ | $nan$ | $nan$ | $nan$
FID | $-43.67$ | $1.02\times 10^{-19}$ | $-40.03$ | $4.79\times 10^{-19}$ | $-42.18$ | $1.89\times 10^{-19}$
CelebA | $p_{g}$ | $27.91$ | $2.87\times 10^{-16}$ | $27.91$ | $2.87\times 10^{-16}$ | $27.59$ | $3.49\times 10^{-16}$
AE Prop. | $0.99$ | $0.33$ | $0.99$ | $0.33$ | $nan$ | $nan$
FID | $-51.51$ | $5.34\times 10^{-21}$ | $-51.51$ | $5.34\times 10^{-21}$ | $-57.69$ | $6.99\times 10^{-22}$
Adversarially Trained | MNIST | $p_{g}$ | $7.24$ | $9.92\times 10^{-7}$ | $7.40$ | $7.24\times 10^{-7}$ | $7.71$ | $4.17\times 10^{-7}$
AE Prop. | $14.31$ | $2.83\times 10^{-11}$ | $15.33$ | $8.9\times 10^{-12}$ | $13.62$ | $6.37\times 10^{-11}$
FID | $-59.42$ | $4.12\times 10^{-22}$ | $-57.79$ | $6.81\times 10^{-22}$ | $-2.00$ | $0.06$
F.-MNIST | $p_{g}$ | $8.33$ | $1.36\times 10^{-7}$ | $8.43$ | $1.15\times 10^{-7}$ | $8.53$ | $9.70\times 10^{-8}$
AE Prop. | $51.66$ | $5.06\times 10^{-21}$ | $50.12$ | $8.70\times 10^{-21}$ | $46.25$ | $3.65\times 10^{-20}$
FID | $-30.42$ | $6.26\times 10^{-17}$ | $-29.62$ | $1.06\times 10^{-16}$ | $-4.70$ | $1.78\times 10^{-4}$
SVHN | $p_{g}$ | $17.46$ | $9.95\times 10^{-13}$ | $17.75$ | $7.47\times 10^{-13}$ | $17.75$ | $7.47\times 10^{-13}$
AE Prop. | $29.82$ | $8.90\times 10^{-17}$ | $25.16$ | $1.77\times 10^{-15}$ | $25.27$ | $1.64\times 10^{-15}$
FID | $-30.65$ | $5.50\times 10^{-17}$ | $-35.43$ | $4.19\times 10^{-18}$ | $-40.35$ | $4.17\times 10^{-19}$
CIFAR-10 | $p_{g}$ | $6.57$ | $3.57\times 10^{-6}$ | $6.57$ | $3.57\times 10^{-6}$ | $6.57$ | $3.57\times 10^{-6}$
AE Prop. | $20.19$ | $8.15\times 10^{-14}$ | $20.19$ | $8.15\times 10^{-14}$ | $21.97$ | $1.89\times 10^{-14}$
FID | $-34.18$ | $7.98\times 10^{-18}$ | $-34.18$ | $7.98\times 10^{-18}$ | $-38.69$ | $8.82\times 10^{-19}$
CelebA | $p_{g}$ | $29.90$ | $8.52\times 10^{-17}$ | $29.48$ | $1.09\times 10^{-16}$ | $29.90$ | $8.52\times 10^{-17}$
AE Prop. | $38.30$ | $1.05\times 10^{-18}$ | $49.67$ | $1.02\times 10^{-20}$ | $38.23$ | $1.09\times 10^{-18}$
FID | $-35.12$ | $4.93\times 10^{-18}$ | $-62.69$ | $1.58\times 10^{-22}$ | $-36.48$ | $2.51\times 10^{-18}$
(b) Statistical Test
As expected, after the adversarial training by (Madry et al., 2018), the
robustness of all five DL models are greatly improved. This can be observed
from the metric of AE Prop.: For all four methods, the proportion of AEs
detected in the set of test case is sharply decreased for adversarially
trained models. Nevertheless, the AE detection performance of HDA is less
affected by the adversarial training. HDA testing can maintain significantly
higher proportion of AEs in the test set for adversarially trained models,
compared with the state of art adversarial attack. This is evident from the
fact that there is an insignificant difference in AE Prop. between HDA and
other adversarial attacks for normally trained models, while the difference is
significant for adversarially trained models.
In terms of the probability density $p_{g}$ and perception quality measured by
FID on generated test cases, HDA significantly outperforms others for both
normally and adversarially trained models, as observed from the large t-scores
and p-values $<<0.05$ . This is unsurprising, since the rationales behind the
three adversarial attacks disregard the consideration of input data
distribution and perception quality.
Finally, we find that the measured $p_{g}$ on test cases detected by HDA
changes due to the variations in local robustness before and after the
adversarial training, yet it remained much higher than all other attack
methods.
Answer to RQ3: HDA is shown to be capable and superior to common adversarial
attacks when applied on DL models with different levels of robustness.
#### 4.2.4. RQ4
The ultimate goal of developing HDA testing is to improve the global
robustness of DL models. To this end, we refer to a validation set of 10000
test seeds. We fine-tune (Jeddi et al., 2021) the DL models with AEs detected
for validation set from different methods. Then, we calculate the train
accuracy, test accuracy and empirical global robustness before and after the
adversarial fine-tuning. Empirical global robustness is measured on a new set
of on-distribution AEs for validation set, different from the fine-tuning
data. Results are presented in Table 10(b).
Table 10. Evaluation of DL models’ train accuracy, test accuracy, and
empirical global robustness (based on 10000 on-distribution AEs) after
adversarial fine-tuning, using different number (N) of test cases.
AE Detection Method | Metric | MNIST | SVHN
---|---|---|---
N = 500 | N = 5000 | N = 50000 | N = 500 | N = 5000 | N = 50000
PGD Attack | Train Acc. | $98.26\%\pm 0.05\%$ | $98.10\%\pm 0.07\%$ | $97.70\%\pm 0.05\%$ | $94.85\%\pm 0.43\%$ | $92.76\%\pm 0.39\%$ | $94.02\%\pm 0.40\%$
Test Acc. | $97.64\%\pm 0.11\%$ | $97.05\%\pm 0.09\%$ | $96.90\%\pm 0.08\%$ | $93.32\%\pm 0.21\%$ | $81.43\%\pm 0.31\%$ | $63.81\%\pm 0.28\%$
$\mathcal{R}_{g}$ | $46.27\%\pm 0.02\%$ | $84.09\%\pm 0.05\%$ | $90.52\%\pm 0.05\%$ | $46.43\%\pm 0.11\%$ | $72.88\%\pm 0.10\%$ | $70.09\%\pm 0.12\%$
Cov. Guided Testing | Train Acc. | $99.89\%\pm 0.05\%$ | $99.61\%\pm 0.07\%$ | $99.12\%\pm 0.05\%$ | $95.73\%\pm 0.44\%$ | $93.65\%\pm 0.42\%$ | $95.76\%\pm 0.43\%$
Test Acc. | $98.93\%\pm 0.08\%$ | $98.77\%\pm 0.08\%$ | $98.41\%\pm 0.07\%$ | $94.11\%\pm 0.19\%$ | $85.66\%\pm 0.19\%$ | $76.43\%\pm 0.19\%$
$\mathcal{R}_{g}$ | $36.71\%\pm 0.04\%$ | $48.91\%\pm 0.04\%$ | $71.12\%\pm 0.05\%$ | $16.21\%\pm 0.11\%$ | $38.59\%\pm 0.16\%$ | $56.58\%\pm 0.12\%$
OODA | Train Acc. | $98.45\%\pm 0.06\%$ | $98.01\%\pm 0.06\%$ | $97.66\%\pm 0.07\%$ | $94.12\%\pm 0.38\%$ | $92.11\%\pm 0.39\%$ | $93.21\%\pm 0.38\%$
Test Acc. | $98.12\%\pm 0.10\%$ | $97.87\%\pm 0.09\%$ | $97.12\%\pm 0.08\%$ | $94.75\%\pm 0.19\%$ | $80.23\%\pm 0.18\%$ | $72.19\%\pm 0.19\%$
$\mathcal{R}_{g}$ | $40.21\%\pm 0.05\%$ | $45.69\%\pm 0.04\%$ | $51.12\%\pm 0.05\%$ | $11.21\%\pm 0.11\%$ | $16.23\%\pm 0.12\%$ | $18.21\%\pm 0.11\%$
FODA | Train Acc. | $98.44\%\pm 0.05\%$ | $98.18\%\pm 0.05\%$ | $97.32\%\pm 0.06\%$ | $94.66\%\pm 0.42\%$ | $92.75\%\pm 0.41\%$ | $94.11\%\pm 0.41\%$
Test Acc. | $97.87\%\pm 0.10\%$ | $97.43\%\pm 0.10\%$ | $97.11\%\pm 0.11\%$ | $92.13\%\pm 0.22\%$ | $82.32\%\pm 0.21\%$ | $78.19\%\pm 0.22\%$
$\mathcal{R}_{g}$ | $37.71\%\pm 0.04\%$ | $47.26\%\pm 0.05\%$ | $55.37\%\pm 0.05\%$ | $12.21\%\pm 0.12\%$ | $20.56\%\pm 0.12\%$ | $23.98\%\pm 0.11\%$
HDA | Train Acc. | $99.56\%\pm 0.07\%$ | $99.10\%\pm 0.07\%$ | $98.89\%\pm 0.06\%$ | $95.00\%\pm 0.33\%$ | $92.83\%\pm 0.39\%$ | $94.40\%\pm 0.40\%$
Test Acc. | $98.52\%\pm 0.09\%$ | $98.42\%\pm 0.08\%$ | $98.21\%\pm 0.08\%$ | $93.86\%\pm 0.24\%$ | $88.67\%\pm 0.25\%$ | $80.60\%\pm 0.22\%$
$\mathcal{R}_{g}$ | $\mathbf{89.67\%\pm 0.03\%}$ | $\mathbf{96.71\%\pm 0.06\%}$ | $\mathbf{99.12\%\pm 0.05\%}$ | $\mathbf{51.15\%\pm 0.09\%}$ | $\mathbf{86.88\%\pm 0.09\%}$ | $\mathbf{91.26\%\pm 0.05\%}$
(a) Results
Dataset | No. of Test Cases | Metric | T(HDA, PGD) | T(HDA, Cov.) | T(HDA, OODA) | T(HDA, FODA)
---|---|---|---|---|---|---
$t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$ | $t$ | $p-value$
MNIST | N=500 | Train Acc. | $47.79$ | $2.04\times 10^{-20}$ | $-12.13$ | $4.24\times 10^{-10}$ | $38.07$ | $1.17\times 10^{-18}$ | $41.17$ | $2.91\times 10^{-19}$
Test Acc. | $19.58$ | $1.39\times 10^{-13}$ | $-10.77$ | $2.83\times 10^{-9}$ | $9.40$ | $2.29\times 10^{-8}$ | $15.28$ | $9.47\times 10^{-12}$
$\mathcal{R}_{g}$ | $3806.43$ | $1.31\times 10^{-54}$ | $3349.48$ | $1.31\times 10^{-53}$ | $2682.35$ | $7.12\times 10^{-52}$ | $3286.24$ | $1.84\times 10^{-53}$
N=5000 | Train Acc. | $31.94$ | $2.64\times 10^{-17}$ | $-16.29$ | $3.21\times 10^{-12}$ | $37.39$ | $1.62\times 10^{-18}$ | $33.82$ | $9.61\times 10^{-18}$
Test Acc. | $35.98$ | $3.20\times 10^{-18}$ | $-9.78$ | $1.25\times 10^{-8}$ | $14.44$ | $2.42\times 10^{-11}$ | $24.45$ | $2.94\times 10^{-15}$
$\mathcal{R}_{g}$ | $510.97$ | $6.52\times 10^{-39}$ | $2096.17$ | $6.03\times 10^{-50}$ | $2237.38$ | $1.86\times 10^{-50}$ | $2002.17$ | $1.38\times 10^{-49}$
N=50000 | Train Acc. | $48.18$ | $1.76\times 10^{-20}$ | $-9.31$ | $2.64\times 10^{-8}$ | $42.19$ | $1.88\times 10^{-19}$ | $58.51$ | $5.45\times 10^{-22}$
Test Acc. | $36.62$ | $2.35\times 10^{-18}$ | $-5.95$ | $1.25\times 10^{-5}$ | $30.47$ | $6.10\times 10^{-17}$ | $25.57$ | $1.33\times 10^{-15}$
$\mathcal{R}_{g}$ | $384.60$ | $1.08\times 10^{-36}$ | $1252.19$ | $6.42\times 10^{-46}$ | $2146.63$ | $3.93\times 10^{-50}$ | $1956.56$ | $2.08\times 10^{-49}$
SVHN | N=500 | Train Acc. | $0.88$ | $0.39$ | $-4.19$ | $5.42\times 10^{-4}$ | $5.53$ | $2.99\times 10^{-5}$ | $2.01$ | $0.06$
Test Acc. | $5.35$ | $4.34\times 10^{-5}$ | $-2.58$ | $0.02$ | $-9.19$ | $3.20\times 10^{-8}$ | $16.80$ | $1.90\times 10^{-12}$
$\mathcal{R}_{g}$ | $105.02$ | $1.50\times 10^{-26}$ | $777.40$ | $3.42\times 10^{-42}$ | $888.65$ | $3.08\times 10^{-43}$ | $820.93$ | $1.28\times 10^{-42}$
N=5000 | Train Acc. | $0.40$ | $0.69$ | $-4.52$ | $2.63\times 10^{-4}$ | $4.13$ | $6.31\times 10^{-4}$ | $0.45$ | $0.66$
Test Acc. | $57.49$ | $7.46\times 10^{-22}$ | $30.31$ | $6.67\times 10^{-17}$ | $86.64$ | $4.77\times 10^{-25}$ | $61.50$ | $2.23\times 10^{-22}$
$\mathcal{R}_{g}$ | $329.07$ | $1.79\times 10^{-35}$ | $831.84$ | $1.01\times 10^{-42}$ | $1489.43$ | $2.83\times 10^{-47}$ | $1398.15$ | $8.83\times 10^{-47}$
N=50000 | Train Acc. | $2.12$ | $0.05$ | $-7.32$ | $8.43\times 10^{-7}$ | $6.82$ | $2.19\times 10^{-6}$ | $1.60$ | $0.13$
Test Acc. | $149.10$ | $2.75\times 10^{-29}$ | $45.36$ | $5.16\times 10^{-20}$ | $91.49$ | $1.79\times 10^{-25}$ | $24.50$ | $1.83\times 10^{-15}$
$\mathcal{R}_{g}$ | $514.96$ | $5.67\times 10^{-39}$ | $843.60$ | $7.86\times 10^{-43}$ | $1911.81$ | $3.16\times 10^{-49}$ | $1760.80$ | $1.39\times 10^{-48}$
(b) Statistical Test
We first observe that adversarial fine-tuning is effective to improve the DL
models’ empirical global robustness, measured by the prediction accuracy on
AEs, detected from normally trained models ($\mathcal{R}_{g}$), while
compromising the train/test accuracy as expected (in contrast to normal
training in Table 1). In most cases, DL models enhanced by HDA testing suffer
from the least drop of generalisation, compared with PGD attack, OODA and
FODA. This can been from significant difference of Test Acc. between HDA and
others. The reason behind this is that HDA testing targets at AEs from high
density regions on distributions, usually with small prediction loss, shown in
Fig. 6. Thus, eliminating AEs detected by HDA testing requires relatively
minor adjustment to DL’s models, the generalisation of which can be easily
tampered during the fine-tuning with new samples.
In terms of empirical global robustness, HDA testing detects AEs around test
seeds from the high global distribution region, which are more significant to
the global robustness improvement. When comparing HDA with other methods, the
p-values of $\mathcal{R}_{g}$ is way smaller than 0.05, indicating HDA
contributes more to the global robustness improvement than others. It also can
be seen that with 5000 test cases generated by utilising 1000 test seeds, the
HDA testing can improve empirical global robustness to nearly or over $90\%$,
very closed to the fine-tuning with 50000 test cases from 10000 test seeds.
This means the distribution-based test seeds selection is more efficient than
random test seeds selection. Moreover, even fine-tuning with 50000 test cases,
leveraging all the test seeds in the validation set, HDA is still better than
others, due to the consideration of local distributions (approximated by
perceptual quality metrics). We notice that PGD-based adversarial fine-tuning
minimises the maximum prediction loss within the local region, which is also
effective to eliminate the high perceptual quality AEs, but sacrificing more
train/test accuracy. DL models fined-tuned with HDA testing achieve the best
balance between the generalisation and global robustness.
Answer to RQ4: Compared with adversarial attack and coverage-guide testing,
HDA contributes more to the growth of global robustness, while mitigating the
drop of train/test accuracy during adversarial fine-tuning.
## 5\. Threats to Validity
### 5.1. Internal Validity
Threats may arise due to bias in establishing cause-effect relationships,
simplifications and assumptions made in our experiments. In what follows, we
list the main threats of each research question and discuss how we mitigate
them.
#### 5.1.1. Threats from HDA Techniques
In RQ1, both the performance of the VAE-Encoder and KDE are threats. For the
former, it is mitigated by using four established quality metrics (in Table 3)
on evaluating dimensionality reduction techniques and compared to the common
PCA method. It is known that KDE performs poorly with high-dimensional data
and works well when the data dimension is modest (Scott, 1991; Liu et al.,
2007). The data dimensions in our experiments are relatively low given the
datasets have been compressed by VAE-Encoder, which mitigates the second
threat. When studying the local robustness indicators, quantifying both the
indicators and the local robustness may subject to errors, for which we reduce
them by carefully inspecting the correctness of the script on calculating the
indicators and invoking a reliable local robustness estimator (Webb et al.,
2019) with fine-tuned hyper-parameters. For using two-step GA to generate
local test cases, a threat arises by the calculation of norm ball radius,
which has been mitigated by $r$-separation distance presented in the paper
(Yang et al., 2020). Also, the threat related to estimating the local
distribution is mitigated by quantifying its three indicators (MSE, PSNR and
SSIM) that are typically used in representing image-quality by human-
perception.
#### 5.1.2. Threats from AEs’ Quality Measurement
A threat for RQ1, RQ2 and RQ3 (when examining how effective our method models
the global distribution and local distribution respectively) is the use of FID
as a metric, quantifying how “similar” two image datasets are. Given FID is
currently the standard metric for this purpose, this threat is sufficiently
mitigated now and can be further mitigated with new metrics in future. RQ2
includes the method of validating AEs developed in (Dola et al., 2021), which
utilises generative models and OOD techniques to flag valid AEs with
reconstruction probabilities greater than a threshold. The determination of
this threshold is critical, thus poses a threat to RQ2. To mitigate it, we use
same settings across all the experiments for fair comparisons.
#### 5.1.3. Threats from Adversarial Training and Fine-Tuning
In RQ3 and RQ4, the first threat rises from the fact that adversarial training
and adversarial fine-tuning will sacrifice the DL model’s generalisation for
robustness. Since the training process is data-driven and of black-box nature,
it is hard to know how the predication of a single data-point will be
affected, while it is meaningless to study the robustness of an incorrectly
predicted seed. To mitigate this threat when we compare the robustness before
and after adversarial training/fine-tuning, we select enough number of seeds
and check the prediction of each selected seed (filtering out incorrect ones
if necessary) to make sure test seeds are always predicted correctly. For the
global robustness computation in RQ4, we refer to a validation dataset, where
a threat may arise if the empirical result based on the validation dataset
cannot represent the global robustness. To mitigate it, we synthesise the
validation set with enough data—10000 inputs sampled from global distribution.
We further attack the validation dataset to find an AE per seed according to
the local distribution. Thus, DL models’ prediction accuracy on this dataset
empirically represents the global robustness as defined. For the
training/fine-tuning to be effective, we need a sufficient number of AEs to
augment the training dataset. A threat may arise due to a small proportion of
AEs in the augmented training dataset (the DL model will be dominated by the
original training data during the training/fine-tuning). To mitigate such a
threat, we generate a large proportion of AEs in our experiments.
### 5.2. External Validity
Threats might challenge the generalisability of our findings, e.g. the number
of models and datasets considered for experimentation; thus we mitigate these
threats as follows. All our experiments are conducted on 5 popular benchmark
datasets, covering 5 typical types of DL models, cf. Table 1. Experimental
results on the effectiveness of each stage in our framework are all based on
averaging a large number of samples, reducing the random noise in the
experiments. In two-step GA based test case generation, a wide range of the
$\alpha$ parameter has been studied showing converging trends. Finally, we
enable replication by making all experimental results publicly
available/reproducible on our project website to further mitigate the threat.
## 6\. Conclusion & Future Work
In this paper, we propose a HDA testing approach for detecting AEs that
considers both the data distribution (thus with higher operational impact
assuming the training data statistically representing the future inputs) and
perceptual quality (thus looks natural and realistic to humans). The key
novelty lies in the hierarchical consideration of two levels of distributions.
To the best of our knowledge, it is the first DL testing approach that
explicitly and collectively models both (i) the feature-level information when
selecting test seeds and (ii) pixel-level information when generating local
test cases. To this end, we have developed a tool chain that provides
technical solutions for each stage of our HDA testing. Our experiments not
only show the effectiveness of each testing stage, but also the overall
advantages of HDA testing over state-of-the-arts. From a software
engineering’s perspective, HDA is cost-effective (by focusing on practically
meaningful AEs), flexible (with end-to-end, black-box technical solutions) and
may effectively contribute to the robustness growth of the DL software under
testing.
The purpose of detecting AEs is to fix them. Although existing DL
retraining/repairing techniques (e.g. (Jeddi et al., 2021) used in RQ4 and
(Wang et al., 2019; Yu et al., 2022)) may satisfy the purpose to some extent,
bespoke “debugging” methods with more emphasise on the feature-distribution
and perceptual quality can be integrated into our framework in a more
efficient way. To this end, our important future work is to close the loop of
“detect-fix-assess” as depicted in (Zhao et al., 2021b) and then organise all
generated evidence as safety cases (Zhao et al., 2020; Dong et al., 2023).
Finally, same as other distribution-aware testing methods, we assume the input
data distribution is same as the training data distribution. To relax this
assumption, we plan to take distribution-shift into consideration in future
versions of HDA. Distribution-aware testing for systematically detecting
explanation AEs (Huang et al., 2023) will also be explored.
###### Acknowledgements.
This work is supported by the U.K. DSTL (through the project of Safety
Argument for Learning-enabled Autonomous Underwater Vehicles) and the U.K.
EPSRC (through End-to-End Conceptual Guarding of Neural Architectures
[EP/T026995/1]). Xingyu Zhao and Alec Banks’ contribution to the work is
partially supported through Fellowships at the Assuring Autonomy International
Programme. This project has received funding from the European Union’s Horizon
2020 research and innovation programme under grant agreement No 956123. We
thank all three anonymous reviewers whose comments helped improve the paper.
This document is an overview of U.K. MOD (part) sponsored research and is
released for informational purposes only. The contents of this document should
not be interpreted as representing the views of the U.K. MOD, nor should it be
assumed that they reflect any current or future U.K. MOD policy. The
information contained in this document cannot supersede any statutory or
contractual requirements or liabilities and is offered without prejudice or
commitment.
## References
* (1)
* Akoglu (2018) Haldun Akoglu. 2018\. User’s guide to correlation coefficients. _Turkish journal of emergency medicine_ 18, 3 (2018), 91–93.
* Alzantot et al. (2019) Moustafa Alzantot, Yash Sharma, Supriyo Chakraborty, Huan Zhang, Cho-Jui Hsieh, and Mani B Srivastava. 2019\. Genattack: Practical black-box attacks with gradient-free optimization. In _Proceedings of the Genetic and Evolutionary Computation Conference_. 1111–1119.
* Attaoui et al. (2023) Mohammed Attaoui, Hazem Fahmy, Fabrizio Pastore, and Lionel Briand. 2023. Black-Box Safety Analysis and Retraining of DNNs Based on Feature Extraction and Clustering. _ACM Trans. Softw. Eng. Methodol._ 32, 3, Article 79 (2023), 40 pages.
* Berend (2021) David Berend. 2021\. Distribution Awareness for AI System Testing. In _43rd IEEE/ACM Int. Conf. on Software Engineering: Companion Proceedings, ICSE Companion 2021, Madrid, Spain, May 25-28, 2021_. IEEE, 96–98.
* Berend et al. (2020) David Berend, Xiaofei Xie, Lei Ma, Lingjun Zhou, Yang Liu, Chi Xu, and Jianjun Zhao. 2020. Cats Are Not Fish: Deep Learning Testing Calls for out-of-Distribution Awareness. In _Proc. of the 35th IEEE/ACM Int. Conference on Automated Software Engineering_ _(ASE’20)_. ACM, New York, NY, USA, 1041–1052. https://doi.org/10.1145/3324884.3416609
* Byun and Rayadurgam (2020) Taejoon Byun and Sanjai Rayadurgam. 2020. Manifold-based Test Generation for Image Classifiers. In _ICSE ’20: 42nd Int. Conference on Software Engineering, Workshops_. ACM, 221\. https://doi.org/10.1145/3387940.3391460
* Byun et al. (2020) Taejoon Byun, Abhishek Vijayakumar, Sanjai Rayadurgam, and Darren Cofer. 2020. Manifold-based Test Generation for Image Classifiers. In _Int. Conf. On Artificial Intelligence Testing (AITest)_. IEEE, Oxford, UK, 15–22.
* Croce and Hein (2020) Francesco Croce and Matthias Hein. 2020. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In _Proc. of the 37th Int. Conf. on Machine Learning (ICML’20)_ , Vol. 119. PMLR, 2206–2216.
* Deb et al. (2002) Kalyanmoy Deb, Amrit Pratap, Sameer Agarwal, and TAMT Meyarivan. 2002. A fast and elitist multiobjective genetic algorithm: NSGA-II. _IEEE transactions on evolutionary computation_ 6, 2 (2002), 182–197.
* Dola et al. (2021) Swaroopa Dola, Matthew B. Dwyer, and Mary Lou Soffa. 2021\. Distribution-Aware Testing of Neural Networks Using Generative Models. In _IEEE/ACM 43rd Int. Conference on Software Engineering_ _(ICSE’21)_. IEEE, Madrid, Spain, 226–237.
* Dong et al. (2023) Yi Dong, Wei Huang, Vibhav Bharti, Victoria Cox, Alec Banks, Sen Wang, Xingyu Zhao, Sven Schewe, and Xiaowei Huang. 2023. Reliability Assessment and Safety Arguments for Machine Learning Components in System Assurance. _ACM Trans. Embed. Comput. Syst._ 22, 3 (2023).
* Du et al. (2019) Xiaoning Du, Xiaofei Xie, Yi Li, Lei Ma, Yang Liu, and Jianjun Zhao. 2019. Deepstellar: Model-based quantitative analysis of stateful deep learning systems. In _Proc. of the 27th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering_. 477–487.
* Dunn et al. (2020) Isaac Dunn, Laura Hanu, Hadrien Pouget, Daniel Kroening, and Tom Melham. 2020. Evaluating robustness to context-sensitive feature perturbations of different granularities. _arXiv preprint arXiv:2001.11055_ (2020).
* Dunn et al. (2021) Isaac Dunn, Hadrien Pouget, Daniel Kroening, and Tom Melham. 2021. Exposing Previously Undetectable Faults in Deep Neural Networks. In _ACM SIGSOFT Int. Symposium on Software Testing and Analysis (ISSTA’21)_. in press.
* Gonzalez and Woods (1992) Rafael C. Gonzalez and Richard E. Woods. 1992. _Digital image processing_. 793 pages.
* Goodfellow et al. (2015) Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy. 2015. Explaining and Harnessing Adversarial Examples. In _3rd Int. Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, Conference Track Proceedings_.
* Harel-Canada et al. (2020) Fabrice Harel-Canada, Lingxiao Wang, Muhammad Ali Gulzar, Quanquan Gu, and Miryung Kim. 2020\. Is Neuron Coverage a Meaningful Measure for Testing Deep Neural Networks?. In _Proc. of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering_. ACM, 851–862.
* Heusel et al. (2017) Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. 2017\. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In _Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017_. 6626–6637.
* Hosseini and Poovendran (2018) Hossein Hosseini and Radha Poovendran. 2018. Semantic Adversarial Examples. In _2018 IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2018_. 1614–1619.
* Huang et al. (2022) Wei Huang, Youcheng Sun, Xingyu Zhao, James Sharp, Wenjie Ruan, Jie Meng, and Xiaowei Huang. 2022. Coverage-Guided Testing for Recurrent Neural Networks. _IEEE Transactions on Reliability_ 71, 3 (2022), 1191–1206.
* Huang et al. (2023) Wei Huang, Xingyu Zhao, Gaojie Jin, and Xiaowei Huang. 2023\. SAFARI: Versatile and Efficient Evaluations for Robustness of Interpretability. In _IEEE/CVF Int. Conf. on Computer Vision (ICCV’23)_.
* Huang et al. (2020) Xiaowei Huang, Daniel Kroening, Wenjie Ruan, and et al. 2020\. A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability. _Computer Science Review_ 37 (2020), 100270.
* Huang et al. (2017) Xiaowei Huang, Marta Kwiatkowska, Sen Wang, and Min Wu. 2017\. Safety verification of deep neural networks. In _Computer Aided Verification_ _(LNCS, Vol. 10426)_. Springer International Publishing, Cham, 3–29.
* Jeddi et al. (2021) Ahmadreza Jeddi, Mohammad Javad Shafiee, and Alexander Wong. 2021. A simple fine-tuning is all you need: Towards robust deep learning via adversarial fine-tuning. In _Workshop on Adversarial Machine Learning in Real-World Computer Vision Systems and Online Challenges (AML-CV) @ CVPR’21_. 1–5.
* Kang et al. (2020) Sungmin Kang, Robert Feldt, and Shin Yoo. 2020. Sinvad: Search-based image space navigation for dnn image classifier test input generation. In _Proc. of the IEEE/ACM 42nd Int. Conf. on Software Engineering Workshops_. 521–528.
* Konak et al. (2006) Abdullah Konak, David W Coit, and Alice E Smith. 2006\. Multi-objective optimization using genetic algorithms: A tutorial. _Reliability engineering & system safety_ 91, 9 (2006), 992–1007.
* Lane et al. (2016) David Lane, David Bisset, Rob Buckingham, Geoff Pegman, and Tony Prescott. 2016. _New foresight review on robotics and autonomous systems_. Technical Report No. 2016.1. LRF. 65 pages.
* Lipowski and Lipowska (2012) Adam Lipowski and Dorota Lipowska. 2012. Roulette-wheel selection via stochastic acceptance. _Physica A: Statistical Mechanics and its Applications_ 391, 6 (2012), 2193–2196.
* Liu et al. (2007) Han Liu, John Lafferty, and Larry Wasserman. 2007. Sparse nonparametric density estimation in high dimensions using the rodeo. In _Artificial Intelligence and Statistics_. PMLR, 283–290.
* Liu et al. (2019) Yang Liu, Eunice Jun, Qisheng Li, and Jeffrey Heer. 2019\. Latent space cartography: Visual analysis of vector space embeddings. In _Computer graphics forum_ , Vol. 38. Wiley Online Library, 67–78.
* Lokerse et al. (1995) S. H. Lokerse, L. P. J. Veelenturf, and J. G. Beltman. 1995\. Density Estimation Using SOFM and Adaptive Kernels. In _Neural Networks: Artificial Intelligence and Industrial Applications - Proceedings of the Third Annual SNN Symposium on Neural Networks, Nijmegen, The Netherlands, September 14-15, 1995_. Springer, 203–206.
* Ma et al. (2018) Lei Ma, Felix Juefei-Xu, Fuyuan Zhang, Jiyuan Sun, Minhui Xue, Bo Li, Chunyang Chen, Ting Su, Li Li, Yang Liu, et al. 2018\. Deepgauge: Multi-granularity testing criteria for deep learning systems. In _Proce. of the 33rd ACM/IEEE Int. Conference on Automated Software Engineering (ASE’18)_. 120–131.
* Madry et al. (2018) Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2018\. Towards Deep Learning Models Resistant to Adversarial Attacks. In _6th Int. Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings_. OpenReview.net.
* Pedregosa et al. (2011) F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine Learning in Python. _Journal of Machine Learning Research_ 12 (2011), 2825–2830.
* Pei et al. (2017) Kexin Pei, Yinzhi Cao, Junfeng Yang, and Suman Jana. 2017\. DeepXplore: Automated Whitebox Testing of Deep Learning Systems. In _Proceedings of the 26th Symposium on Operating Systems Principles_. ACM, 1–18.
* Riccio et al. (2021) Vincenzo Riccio, Nargiz Humbatova, Gunel Jahangirova, and Paolo Tonella. 2021. Deepmetis: Augmenting a deep learning test set to increase its mutation score. In _36th IEEE/ACM Int. Conf. on Automated Software Engineering_. 355–367.
* Riccio and Tonella (2020) Vincenzo Riccio and Paolo Tonella. 2020. Model-based exploration of the frontier of behaviours for deep learning system testing. In _Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering_. 876–888.
* Rosenberg and Hirschberg (2007) Andrew Rosenberg and Julia Hirschberg. 2007. V-Measure: A Conditional Entropy-Based External Cluster Evaluation Measure. In _EMNLP-CoNLL 2007, Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, June 28-30, 2007, Prague, Czech Republic_. ACL, 410–420.
* Ruan et al. (2018) Wenjie Ruan, Xiaowei Huang, and Marta Kwiatkowska. 2018\. Reachability Analysis of Deep Neural Networks with Provable Guarantees. In _Proceedings of the Twenty-Seventh Int. Joint Conference on Artificial Intelligence, IJCAI-18_. Int. Joint Conferences on Artificial Intelligence Organization, 2651–2659.
|
# Visibility Aware Human-Object Interaction Tracking from Single RGB Camera
Xianghui Xie Bharat Lal Bhatnagar Gerard Pons-Moll
University of Tübingen, Tübingen AI Center, Germany
Max Planck Institute for Informatics, Saarland Informatics Campus, Germany
{xxie<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Capturing the interactions between humans and their environment in 3D is
important for many applications in robotics, graphics, and vision. Recent
works to reconstruct the 3D human and object from a single RGB image do not
have consistent relative translation across frames because they assume a fixed
depth. Moreover, their performance drops significantly when the object is
occluded. In this work, we propose a novel method to track the 3D human,
object, contacts between them, and their relative translation across frames
from a single RGB camera, while being robust to heavy occlusions. Our method
is built on two key insights. First, we condition our neural field
reconstructions for human and object on per-frame SMPL model estimates
obtained by pre-fitting SMPL to a video sequence. This improves neural
reconstruction accuracy and produces coherent relative translation across
frames. Second, human and object motion from visible frames provides valuable
information to infer the occluded object. We propose a novel transformer-based
neural network that explicitly uses object visibility and human motion to
leverage neighbouring frames to make predictions for the occluded frames.
Building on these insights, our method is able to track both human and object
robustly even under occlusions. Experiments on two datasets show that our
method significantly improves over the state-of-the-art methods. Our code and
pretrained models are available at: https://virtualhumans.mpi-
inf.mpg.de/VisTracker.
## 1 Introduction
Figure 1: From a monocular RGB video, our method tracks the human, object and
contacts between them even under occlusions.
Perceiving and understanding human as well as their interaction with the
surroundings has lots of applications in robotics, gaming, animation and
virtual reality etc. Accurate interaction capture is however very hard. Early
works employ high-end systems such as dense camera arrays [37, 8, 17] that
allow accurate capture but are expensive to deploy. Recent works [6, 36, 34]
reduce the requirement to multi-view RGBD cameras but it is still complicated
to setup the full capture system hence is not friendly for consumer-level
usage. This calls for methods that can capture human-object interaction from a
single RGB camera, which is more convenient and user-friendly.
However, reasoning about the 3D human and object from monocular RGB images is
very challenging. The lack of depth information makes the predictions
susceptible to depth-scale ambiguity, leading to temporally incoherent
tracking. Furthermore, the object or human can get heavily occluded, making
inference very hard. Prior work PHOSA [91] relies on hand-crafted heuristics
to reduce the ambiguity but such heuristic-based method is neither very
accurate nor scalable. More recently, CHORE [80] combines neural field
reconstructions with model based fitting obtaining promising results. However,
CHORE, assumes humans are at a fixed depth from the camera and predicts scale
alone, thereby losing the important _relative translation_ across frames.
Another limitation of CHORE is that it is not robust under occlusions as
little information is available from single-frame when the object is barely
visible. Hence CHORE often fails in these cases, see Fig. 3.
In this work, we propose the first method that can track both human and object
accurately from monocular RGB videos. Our approach combines neural field
predictions and model fitting, which has been consistently shown to be more
effective than directly regressing pose [4, 6, 80, 5]. In contrast to existing
neural field based reconstruction methods [61, 80], we can do tracking
including inference of relative translation. Instead of assuming a fixed
depth, we condition the neural field reconstructions (for object and human) on
per frame SMPL estimates (SMPL-T) including translation in camera space
obtained by pre-fitting SMPL to the video sequence. This results in coherent
translation and improved neural reconstruction. In addition, we argue that
during human-object interaction, the object motion is highly correlated with
the human motion, which provides us valuable information to recover the object
pose even when it is occluded (see Fig. 1 column 3-4). To this end, we propose
a novel transformer based network that leverages the human motion and object
motion from nearby visible frames to predict the object pose under heavy
occlusions.
We evaluate our method on the BEHAVE [6] and InterCap dataset [35].
Experiments show that our method can robustly track human, object and
realistic contacts between them even under heavy occlusions and significantly
outperforms the currrent state of the art method, CHORE [80]. We further
ablate the proposed _SMPL-T conditioning_ and _human and visibility aware_
object pose prediction network and demonstrate that they are key for accurate
human-object interaction tracking.
In summary, our key contributions include:
* •
We propose the first method that can jointly track full-body human interacting
with a movable object from a monocular RGB camera.
* •
We propose _SMPL-T conditioned interaction fields_ , predicted by a neural
network that allows consistent 4D tracking of human and object.
* •
We introduce a novel _human and visibility aware_ object pose prediction
network along with an object visibility prediction network that can recover
object poses even under heavy occlusions.
* •
Our code and pretrained models are publicly available to foster future
research in this direction.
Figure 2: Given an input RGB sequence of a human interacting with an object
and their corresponding human-object masks (A), we aim to reconstruct and
track the 3D human, object and the contacts between them (E). Our first key
idea is a SMPL-T conditioned interaction field network (SIF-Net, details in
Sec. 3.3) that predicts neural fields conditioned on estimated SMPL meshes in
camera space (col. B, SMPL-T, details in Sec. 3.2). SMPL-T conditioning
provides us temporally consistent relative translation, which is important for
coherent 4D tracking. Our second key insight is to predict object pose under
occlusions (D) leveraging human motion and object visibility information
(HVOP-Net, details in Sec. 3.4). This prediction provides robust object
tracking for frames with heavy occlusion. We then jointly optimize human and
object (details in Sec. 3.5) to satisfy image observations and contact
constraints.
## 2 Related Work
In this section, we first review recent works that deal with human or object
pose estimation and tracking separately. We then discuss recent progresses
that model human and object interactions and works that deal with occlusions
explicitly.
Human or object pose estimation and tracking. After the introduction of SMPL
[49] body model, tremendous progress has been made in human mesh recovery
(HMR) from single images [55, 7, 52, 3, 2, 43, 18] or videos [89, 56, 57, 40,
20]. We refer readers to a recent review of HMR methods in [70]. On the other
hand, deep learning method has also significantly improved object 6D pose
estimation from single RGB images [73, 50, 25, 45, 22, 53, 33]. However,
object pose tracking has received less attention and most works focus on RGBD
inputs [76, 65, 94, 21, 75]. Two works explore the camera localization ideas
from SLAM communities and can track object from RGB videos [48, 68].
Nevertheless, they heavily rely on visual evidence and the performance is
unknown under heavy occlusions. They also do not track human-object
interactions.
Human-object interaction. Modelling human object interaction is an emerging
research topic in recent years. Hand-object interaction is studied with works
modelling hand-object interaction from RGB [38, 19, 30, 24, 85], RGBD [9, 11,
28] or 3D inputs [69, 10, 95, 54]. There are also works that model human
interacting with a static scene [29, 34, 63, 77, 87, 27, 86] or deformable
surface [46]. More recently, the release of BEHAVE[6] and InterCap[35]
datasets allows bench-marking of full-body interacting with a movable object.
However, human-object interaction capture usually deploys multi-view RGB [67]
or RGBD[6, 34, 35, 93, 28, 9, 28, 20] cameras. Only a few works [91, 80, 74]
reconstruct dynamic human and object from monocular RGB input and our
experiments show that they are not suitable for tracking.
Pose estimation under occlusion. Most existing methods assume occlusion-free
input images hence are not robust under occlusions. Only a few methods address
human pose estimation under partial occlusions [41, 92, 58, 59, 26, 42] or
long term occlusions [89]. For object pose estimation, pixel-wise voting [53]
and self-occlusion [22] are explored for more robust prediction under
occlusions. More recently, TP-AE [94] predicts object occlusion ratio to guide
pose estimation but relies on depth input. Although being impressive on
separate human or object pose estimation, these methods do not reason about
human-object interaction. Our method is the first one that takes both human
and object visibility into account for interaction tracking.
## 3 Method
We present a novel method for jointly tracking the human, the object and the
contacts between them, in 3D, from a monocular RGB video. The first main
challenge in monocular tracking is the estimation of human and object
translations in camera space due to the depth-scale ambiguity problem.
Existing method, CHORE [80], reconstructs human and object at a fixed depth to
the camera, leading to inconsistent 3D translation across frames. Our key idea
is to fit a SMPL model with single shape parameters to a video sequence to
obtain consistent relative translation across frames. We call the estimated
SMPL as SMPL-T and describe it in more details in Sec. 3.2. Based on the
estimated SMPL-T, we then jointly model the 3D human, object and the
interactions using our proposed _SMPL-T conditioned Interaction Fields_
network (SIF-Net, Sec. 3.3)
Object tracking from a single-frame only, is difficult when the object is
barely visible. Hence we introduce a _Human and Visibility aware Object Pose
Network_ (HVOP-Net) that leverages human and object motion from visible frames
to recover the object under occlusion (Sec. 3.4). We then use the SIF-Net and
HVOP-Net outputs to optimize SMPL model and object pose parameters of this
sequence to satisfy neural prediction and image observations (Sec. 3.5). An
overview of our approach can be found in Fig. 2.
### 3.1 Preliminaries
In this work, we focus on a single human interacting with an object, which is
a common setting in other hand-object interaction [85, 69, 95] and full body-
object interaction works [80, 6, 35]. We represent human using the SMPL [49]
body model $H(\boldsymbol{\theta},\boldsymbol{\beta})$ that parameterises the
3D human mesh using pose $\boldsymbol{\theta}$ (including global translation)
and shape $\boldsymbol{\beta}$ parameters. The object is represented by a
known mesh template and we estimate the rotation $\mathbf{R}^{o}\in SO(3)$ and
translation $\mathbf{t}^{o}\in\mathbb{R}^{3}$ parameters. Given a video
sequence $\\{\mathbf{I}_{1},...,\mathbf{I}_{T}\\}$ where
$\mathbf{I}_{i}\in\mathbb{R}^{H\times W\times 5}$ (RGB, human and object
masks), our goal is to estimate the SMPL shape $\boldsymbol{\beta}$, a
sequence of SMPL pose
$\mathbf{\Theta}=\\{\boldsymbol{\theta}_{1},...,\boldsymbol{\theta}_{T}\\}$,
object rotation
$\mathbf{\mathcal{R}}^{o}=\\{\mathbf{R}_{1}^{o},...,\mathbf{R}_{T}^{o}\\}$ and
translation
$\mathbf{\mathcal{T}}^{o}=\\{\mathbf{t}^{o}_{1},...,\mathbf{t}^{o}_{T}\\}$
parameters that satisfy 2D image observation and realistic interaction
constraints between the human and the object.
### 3.2 SMPL-T: Temporally consistent SMPL meshes in camera space
Our first step is to obtain SMPL meshes in camera space that have consistent
translation in a video sequence. We leverage 2D body keypoint predictions from
openpose[12] and natural motion smoothness for this. Specifically, we use
FrankMocap [60] to initialise SMPL-T pose
$\mathbf{\Theta}=\\{\boldsymbol{\theta}_{1},...,\boldsymbol{\theta}_{T}\\}$
and shape
$\mathbf{\mathcal{B}}=\\{\boldsymbol{\beta}_{1},...,\boldsymbol{\beta}_{T}\\}$
parameters for a sequence of images. Note here the original SMPL meshes from
FrankMocap are centered at origin. We average the SMPL-T shape parameters over
the sequence as the shape for the person and optimize the SMPL-T global
translation and body poses to minimize 2D reprojection and temporal smoothness
error:
$E(\mathbf{\Theta})=\lambda_{\text{J2D}}L_{\text{J2D}}+\lambda_{\text{reg}}L_{\text{reg}}+\lambda_{a}L_{\text{accel}}+\lambda_{\text{pi}}L_{\text{pi}}$
(1)
where $L_{\text{J2D}}$ is the sum of body keypoint reprojection losses[7] over
all frames and $L_{\text{reg}}$ is a regularization on body poses using priors
learned from data[49, 52, 71]. $L_{\text{accel}}$ is a temporal smoothness
term that penalizes large accelerations over SMPL-T vertices $H_{i}$:
$L_{\text{accel}}=\sum_{i=0}^{T-2}||H_{i}-2H_{i+1}+H_{i+2}||_{2}^{2}$.
$L_{\text{pi}}$, an L2 loss between optimized and initial body pose, prevents
the pose from deviating too much from initialization. $\lambda_{*}$ denotes
the loss weights detailed in Supp.
Note that this optimization does not guarantee that we get the absolute
translation in the world coordinates, it just ensures that our predictions
will be consistent with the SMPL model over a sequence, i.e. we will be off by
one rigid transformation for the entire sequence.
### 3.3 SIF-Net: SMPL-T conditioned interaction field
Our SMPL-T provides translation about the human but does not reason about the
object and the interaction between them. Existing method CHORE [80] can
jointly reason human and object but their humans are predicted at fixed depth.
Our key idea is to leverage our SMPL-T meshes to jointly reason human, object
and interaction while having consistent relative translation. We model this
using a single neural network which we call SIF-Net. The input to SIF-Net
consists of RGB image, human and object masks, and our estimated SMPL-T. With
features from SMPL-T and input images, it then predicts interaction field
which consists of human and object distance fields, SMPL part correspondence
field, object pose and visibility field.
SIF-Net feature encoding. Existing neural implicit methods [62, 80, 61] rely
mainly on features from input image, a main reason that limits their human
prediction at fixed depth. Instead, we extract features from both our
estimated SMPL-T meshes and input image, providing more distinct features for
the query points along the same ray. Inspired by EG3D [13], we use the
triplane representation for SMPL-T feature learning due to its efficiency.
Specifically, we use orthographic camera $\pi^{o}(\cdot)$, to render the
SMPL-T mesh silhouette from right, back and top-down views and obtain three
images $\mathbf{S}^{r}_{i},\mathbf{S}^{b}_{i},\mathbf{S}^{t}_{i}$
respectively, where $\mathbf{S}^{i}\in\mathbb{R}^{H\times W}$, see
supplementary for more visualization. Note here the triplane origin is placed
at our SMPL-T mesh center (Fig. 2 B). We then train an image encoder
$f^{\text{tri}}(\cdot)$ that extracts a pixel aligned feature grid
$\mathbf{D}^{j}_{i}\in\mathbb{R}^{H_{c}\times W_{c}\times C}$ from each
rendered view $\mathbf{S}^{j}_{i}$, where $j\in\\{r,b,t\\}$ and
$H_{c},W_{c},C$ are the feature grid dimensions. To extract features for a
query point $\mathbf{p}\in\mathbb{R}^{3}$, we project $\mathbf{p}$ into the
three planes using the same orthographic projection
$\pi_{\mathbf{p}}^{o}=\pi^{o}(\mathbf{p})$ and extract local features
$\mathbf{D}_{i}^{\mathbf{p}}=(\mathbf{D}^{r}_{i}(\pi_{\mathbf{p}}^{o}),\mathbf{D}^{b}_{i}(\pi_{\mathbf{p}}^{o}),\mathbf{D}^{t}_{i}(\pi_{\mathbf{p}}^{o}))$
using bilinear interpolation.
In addition to the SMPL-T features, SIF-Net also extracts information from
input images. More specifically, we train an image encoder
$f^{\text{enc}}(\cdot)$ to extract feature grid
$\mathbf{Z}_{i}\in\mathbb{R}^{H_{f}\times W_{f}\times C_{f}}$ from input image
$\mathbf{I}_{i}\in\mathbb{R}^{H\times W\times 5}$, here $H_{f},W_{f},C_{f}$
and $H,W$ are feature grid and input image dimensions respectively. Given
query point $\mathbf{p}\in\mathbb{R}^{3}$, we project it to 2D image using
full perspective projection $\pi_{\mathbf{p}}=\pi(\mathbf{p})$ and extract
pixel-aligned features
$\mathbf{Z}_{i}^{\mathbf{p}}=\mathbf{Z}_{i}(\pi_{\mathbf{p}})$. The input
image feature is concatenated with the SMPL-T feature to form an input and
translation aware point feature:
$\mathbf{F}_{i}^{\mathbf{p}}=(\mathbf{Z}_{i}^{\mathbf{p}},\mathbf{D}_{i}^{\mathbf{p}})$.
SIF-Net predictions. From the point feature $\mathbf{F}^{\mathbf{p}}_{i}$
discussed above, we predict our interaction fields that jointly model human,
object and their contacts, similar to CHORE [80]. Specifically, we predict the
unsigned distances to human and object surfaces using
$f^{u}:\mathbf{F}_{i}^{\mathbf{p}}\mapsto\mathbb{R}^{2}_{\geq 0}$. This allows
fitting the SMPL mesh and object template by minimizing the predicted
distances at mesh vertices. We can also infer contacts, as the points having
small distances to both human and object surfaces. For more robust SMPL
fitting [4] and modelling which body part the object point is in contact with,
we predict SMPL part correspondence using
$f^{p}:\mathbf{F}_{i}^{\mathbf{p}}\mapsto\\{1,2,...,K\\}$ where $K$ is the
number of SMPL parts. For more accurate object fitting, we additionally
predict object rotation with
$f^{R}:\mathbf{F}_{i}^{\mathbf{p}}\mapsto\mathbb{R}^{3\times 3}$ and
translation with $f^{c}:\mathbf{F}_{i}^{\mathbf{p}}\mapsto\mathbb{R}^{3}$. The
predicted $3\times 3$ matrix is projected to SO(3) using symmetric
orthogonalization[44]. At test time, we first use $f^{u}$ to find points on
the object surface[16] and take the average rotation and translation
predictions of these points as the object pose. To handle occlusions, we also
predict the object visibility field using
$f^{\text{vis}}:\mathbf{F}_{i}^{\mathbf{p}}\mapsto[0,1]$. The visibility is
useful to recover the object pose under occlusion, see more details in Sec.
3.4.
Why use triplane to encode SMPL-T? A direct alternative to triplane based
SMPL-T encoding is to find the closest point in SMPL-T mesh and concatenate
that coordinate to the point features. But such a method is slow (computing
point to surface distance) and does not allow flexible learning of local and
global features. Another choice is to voxelize the SMPL-T mesh and extract
point local features using IF-Nets [15] but such a method is still expensive.
Therefore we chose the more efficient triplane representation to encode our
estimated SMPL-T meshes.
Implementation. Our SMPL-T feature extractor $f^{\text{tri}}$ is shared for
three views and trained end to end with image encoder $f^{\text{enc}}$ and
other neural field predictors. At training time, we input the renderings from
ground truth SMPL meshes and train the network to predict GT labels. At test
time, we obtain the SMPL-T meshes using Eq. 1. In order to have smoother
SMPL-T feature in a sequence, we use SmoothNet [90] to smooth the optimized
SMPL parameters. We evaluate this component and provide more implementation
details in supplementary.
### 3.4 HVOP-Net: Human and Visibility aware Object Pose under occlusions
Our SIF-Net recovers translation and more accurate object pose. However, the
object pose prediction under very heavy occlusions remains challenging because
no image evidence from single frame is available for accurate prediction, see
Fig. 5. Our key idea is to use the SMPL-T and object pose from other visible
frames to predict the object of occluded frames. To this end, we first predict
object visibility scores in each image, which are then leveraged together with
the human evidence from neighbouring frames to predict the object poses of the
occluded frames.
Object visibility score. Our visibility score denotes how much the object is
visible in the input image. We train a visibility decoder
$f^{\text{vis}}(\cdot)$, a prediction head of SIF-Net, that takes a point
feature $\mathbf{F}_{i}^{\mathbf{p}}$ as input and predicts visibility score
$v_{i}\in[0,1]$ for frame $i$. At test time, we first use the neural object
distance predictor $f^{u}$ to find object surface points [16] and then take
the average visibility predictions of these points as the object visibility
score for this image.
Object pose prediction under heavy occlusion. Our goal now is to predict
accurate object pose for heavily occluded frames. We consider frames whose
visibility score $v_{i}$ is smaller than $\delta=0.5$ as the occluded frames.
Inspired by works from motion infill [39, 23] and synthesis [88, 93], we
design our HVOP-Net that leverages transformer [72] and explicitly takes the
human motion and object visibility into account to recover object pose under
heavy occlusions.
More specifically, we first use a transformer $f^{s}(\cdot)$ to aggregate
temporal information of the SMPL-T poses:
$f^{s}:\mathbb{R}^{T\times|\theta_{i}|}\mapsto\mathbb{R}^{T\times D_{hs}}$,
where $|\theta_{i}|$ is the SMPL pose dimension and $D_{hs}$ is the hidden
feature dimension. Similarly, we use a transformer $f^{o}(\cdot)$ to aggregate
temporal information of the object poses: $f^{o}:\mathbb{R}^{T\times
D_{o}}\mapsto\mathbb{R}^{T\times D_{ho}}$. Note here the SMPL-T transformer
$f^{s}$ attends to all frames while the object transformer $f^{o}$ only
attends to frames where object is visible ($v_{i}\geq\delta$). We then
concatenate the SMPL-T and object features and use another transformer
$f^{\text{comb}}$ to aggregate both human and object information and predict
the object poses:
$f^{\text{comb}}:\mathbb{R}^{T\times(D_{hs}+D_{ho})}\mapsto\mathbb{R}^{T\times
D_{o}}$. The joint transformer $f^{\text{comb}}$ attends to all frames. Our
experiments show that our HVOP-Net is important to accurately predict object
pose under heavy occlusions, see Fig. 5.
Implementation. The visibility decoder $f^{\text{vis}}$ is trained end to end
with other SIF-Net components using L2 loss. The GT visibility score is
computed as the number of visible object pixels (from object mask) divided by
total number of object pixels (from GT object rendering). To train our HVOP-
Net, we randomly zero out the object pose for a small clip of the input
sequence and provide ground truth SMPL and object poses for other frames as
input. We train our network to accept input sequence of a fixed length but at
test time, the sequence can have various length and object occlusion can last
for a longer time. To this end, we use an auto-regressive algorithm to recover
the object pose of a full video, similar to [89]. At test time, we use SIF-Net
object pose predictions and zero out highly occluded frames based on the
predicted visibility scores. We empirically find that having smooth object
pose as input is helpful for more accurate prediction. Hence we use SmoothNet
[90] to smooth SIF-Net object pose predictions before inputing them to HVOP-
Net. The evaluation of this component and more training details are described
in our supplementary.
### 3.5 Visibility aware joint optimization
To obtain SMPL and object meshes that align with input images and satisfy
contact constraints, we leverage our network predictions from Sec. 3.3 and
Sec. 3.4 to formulate a robust joint optimization objective. Our goal is to
obtain an optimal set of parameters
$\mathbf{\Phi}=\\{\mathbf{\Theta},\boldsymbol{\beta},\mathbf{\mathcal{R}}^{o},\mathbf{\mathcal{T}}^{o}\\}$
for SMPL pose, shape, object rotation and translation respectively. We
initialize the SMPL parameters from our estimated SMPL-T (Eq. 1) and object
parameters from our HVOP-Net predictions (Sec. 3.4). Inspired by CHORE[80],
our energy function consists of human data $E_{data}^{h}$, object data
$E_{data}^{o}$, contact data $E_{data}^{c}$ and SMPL pose prior term
$E_{reg}$:
$E_{(}\mathbf{\Phi})=E_{data}^{h}+E_{data}^{o}+E_{data}^{c}+E_{reg}.$ (2)
here $E_{reg}$ is a body pose and shape prior loss[49]. We explain other loss
data terms next.
Human data term. $E_{data}^{h}$ minimizes the discrepancy between SIF-Net
prediction and SMPL meshes as well as a temporal smoothness error:
$E_{data}^{h}(\mathbf{\Theta},\boldsymbol{\beta})=\sum_{i=1}^{T}L_{\text{neural}}^{h}(\boldsymbol{\theta}_{i},\boldsymbol{\beta})+\lambda_{\text{ah}}L_{\text{accel}}(\mathbf{\Theta})$,
where $L_{\text{accel}}$ is the same used in Eq. 1. $L_{\text{neural}}^{h}$
pushes the SMPL vertices to the zero-level set of the human distance field
represented by neural predictor $f^{u,h}_{i}$ and forces correct SMPL part
locations predicted by $f^{p}_{i}$:
$E^{h}_{\text{neural}}(\mathbf{\Theta},\boldsymbol{\beta})=\sum_{i=1}^{T}(\sum_{\mathbf{p}\in
H(\boldsymbol{\theta}_{i},\boldsymbol{\beta})}(\lambda_{h}\min(f^{u,h}_{i}(\mathbf{F}^{\mathbf{p}}_{i}),\delta_{h})+\\\
\lambda_{p}L_{p}(l_{\mathbf{p}},f^{p}_{i}(\mathbf{F}^{\mathbf{p}}_{i}))))$ (3)
here $l_{\mathbf{p}}$ is the predefined SMPL part label [4] of SMPL vertex
$\mathbf{p}$ and $L_{p}$ is the categorical cross entropy loss function.
$\delta_{h}$ is a small clamping value.
Object data term. We transform the object template vertices
$\mathbf{O}\in\mathbb{R}^{3\times N}$ using object pose parameters of frame
$i$ by:
$\mathbf{O}^{\prime}_{i}=\mathbf{R}^{o}_{i}\mathbf{O}+\mathbf{t}^{o}_{i}$.
Intuitively, the object vertices should lie on the zero-level set of the
object distance field represented by $f^{u,o}_{i}$ and the rendered silhouette
should match the 2D object mask $\mathbf{M}^{o}_{i}$. Hence we formulate the
loss as:
$E^{o}_{data}(\mathbf{\mathcal{R}}^{o},\mathbf{\mathcal{T}}^{o})=\sum_{i=1}^{T}v_{i}(\sum_{\mathbf{p}\in\mathbf{O}^{\prime}_{i}}\lambda_{o}\min(f^{u,o}_{i}(\mathbf{F}^{\mathbf{p}}_{i}),\delta_{o})\\\
+\lambda_{\text{occ}}L_{\text{occ-
sil}}(\mathbf{O}_{i}^{\prime},\mathbf{M}^{o}_{i}))+\lambda_{\text{ao}}L_{\text{ao}}$
(4)
where $v_{i}$ is the predicted object visibility score described in Sec. 3.4.
This down-weights the loss values of network predictions for frames where
object is occluded and allows more temporal regularization. $L_{\text{occ-
sil}}$ is an occlusion-aware silhouette loss[91] and $L_{\text{ao}}$ is a
temporal smoothness loss applied to object vertices $\mathbf{O}^{\prime}_{i}$,
similar to $L_{\text{accel}}$ in Eq. 1.
Contact data term. The contact data term [80] minimizes the distance between
human and object points that are predicted to be in contact:
$E_{\text{data}}^{c}(\mathbf{\mathcal{R}}^{o},\mathbf{\mathcal{T}}^{o})=\lambda_{c}\sum_{i=1}^{T}(\sum_{j=1}^{K}d(H_{j}^{c}(\boldsymbol{\theta}_{i},\boldsymbol{\beta}),\mathbf{O}_{ij}^{c}))$
(5)
here $d(\cdot,\cdot)$ is chamfer distance. We consider human points on the
$j^{\text{th}}$ body part of SMPL mesh $H_{i}$ of frame $i$ (denoted as
$H_{ij}$) are in contact when their predicted distance to the object is
smaller than a threshold:
$H^{c}_{j}(\boldsymbol{\theta}_{i},\boldsymbol{\beta})=\\{\mathbf{p}|\mathbf{p}\in
H_{ij}\,\text{and}\,f^{u,o}_{i}(\mathbf{F}^{\mathbf{p}}_{i})\leq\epsilon\\}$.
Similarly, we find contact points on object meshes with
$\mathbf{O}_{ij}^{c}=\\{\mathbf{p}|\mathbf{p}\in\mathbf{O}^{\prime}_{i}\,\text{and}\,f^{u,h}_{i}(\mathbf{F}^{\mathbf{p}}_{i})\leq\epsilon\,\text{and}\,f^{p}_{i}(\mathbf{F}^{\mathbf{p}}_{i})=j\\}$.
Please see Supp. for more details about loss weights $\lambda_{*}$.
## 4 Experiments
Figure 3: Comparison with PHOSA [91] and CHORE[80] on BEHAVE[6] (row 1-2) and
InterCap [35] (row 3-4). PHOSA’s object pose optimization often gets stuck in
local minima due to heavy occlusions. CHORE also fails to predict accurate
object pose as it does not fully explore the human, temporal and visibility
information while our method can robustly track human and object in these
challenging cases.
In this section, we first compare our method against existing approaches on
tracking human and object and then evaluate the key components of our methods.
Our experiments show that our method clearly outperforms existing joint human
object reconstruction method and our novel _SMPL-T conditioned interaction
fields_ (SIF-Net) as well as _human and visibility aware_ object pose
prediction (HVOP-Net) works better than existing state of the art methods.
Baselines. (1) Joint human and object tracking. We compare against PHOSA [91]
and CHORE [80] in the joint human-object reconstruction task. (2) Object pose
prediction. Our HVOP-Net leverages nearby (un-occluded) frames to predict the
object pose of occluded frames. We compare this with a simple baseline that
linearly interpolates the object pose between visible frames to recover
occluded poses. We also find similarity between our task and motion
smoothing/infilling. Hence we compare our HVOP-Net with SoTA smoothing [90]
and infill method [39].
Datasets. We conduct experiments on the BEHAVE [6] and InterCap [35] dataset.
(1) BEHAVE [6] captures 7 subjects interacting with 20 different objects in
natural environments and contains SMPL and object registrations annotated at
1fps. We use the extended BEHAVE dataset, which registers SMPL and object for
BEHAVE sequences at 30 fps. We follow the official split [80] with $217$
sequences for training and $82$ for testing. (2) InterCap [35] is a similar
dataset that captures 10 subjects interacting with 10 different objects. The
dataset comes with pseudo ground truth SMPL and object registrations at 30fps.
We train our model on sequences from subject 01-08 (173 sequences) and test on
sequences from subject 09-10 (38 sequences).
We compare with CHORE [80] on the full test set of both datasets. PHOSA [91]
optimizes hundreds of random object pose initialization per image hence the
inference speed is very slow (2min/image), which makes it infeasible to run on
full video sequences. Hence we compare with PHOSA on key frames only, denoted
as $\text{BEHAVE}^{*}$(3.9k images) and $\text{InterCap}^{*}$(1.1k images)
respectively. Due to the large number of frames in the full BEHAVE test set
(127k frames), we conduct other ablation experiments in a sub test set (42k
frames) of BEHAVE.
Evaluation metrics. (1) Joint human-object tracking. We evaluate the
performance of SMPL and object reconstruction using Chamfer distance between
predicted SMPL and object meshes, and the ground truth. CHORE [80] uses
Procrustes alignment on combined SMPL and object meshes for _each frame_
before computing errors. However, this does not reflect the real accuracy in
terms of the relative translation between nearby frames in a video. Inspired
by the world space errors proposed by SPEC[42] and GLAMR[89], we propose to
perform joint Procrustes alignment in a sliding window, as also used in SLAM
evaluations [31, 66]. More specifically, we combine all SMPL and object
vertices within a sliding window and compute a single optimal Procrustes
alignment to the ground truth vertices. This alignment is then applied to all
SMPL and object vertices within this window and Chamfer distance of SMPL and
object meshes are computed respectively. We report both the errors using per-
frame alignment (w=1) and alignment with a sliding window of 10s (w=10).
(2) Object only evaluation. For experiments evaluating object pose only, we
evaluate the rotation accuracy using rotation angle [32]. The object
translation error is computed as the distance between reconstructed and GT
translation. We report all errors in centimetre in our experiments.
### 4.1 Evaluation of tracking results
Dateset | Methods | Align w=1 | Align w=10
---|---|---|---
SMPL $\downarrow$ | Obj. $\downarrow$ | SMPL $\downarrow$ | Obj. $\downarrow$
BEHAVE | CHORE | 5.55 | 10.02 | 18.33 | 20.32
Ours | 5.25 | 8.04 | 7.81 | 8.49
$\text{BEHAVE}^{*}$ | PHOSA | 12.86 | 26.90 | 27.01 | 59.08
CHORE | 5.54 | 10.12 | 21.28 | 22.39
Ours | 5.24 | 7.89 | 8.24 | 8.49
InterCap | CHORE | 7.12 | 12.59 | 16.11 | 21.05
Ours | 6.76 | 10.32 | 9.35 | 11.38
$\text{InterCap}^{*}$ | PHOSA | 11.20 | 20.57 | 24.16 | 43.06
CHORE | 7.01 | 12.81 | 16.10 | 21.08
Ours | 6.78 | 10.34 | 9.35 | 11.54
Table 1: Human and object tracking results on BEHAVE [6] and InterCap [35]
datasets (unit: cm). * denotes key frames only. w is the temporal window size
used for Procrustes alignment where w=1 means per-frame Procrustes and w=10
means alignment over a sliding window of 10s. We can see our method clearly
outperforms baseline PHOSA[91] and CHORE[80] in all metrics.
We compare our human and object tracking results against baseline PHOSA [91]
and CHORE [80] and report the errors in Tab. 1. Note that the comparison with
PHOSA and CHORE is not strictly fair as they do not use temporal information.
Nevertheless they are our closest baselines and we show that our method
outperforms them using per-frame alignment and is significantly better under a
more relevant evaluation metric (align w=10) for video tracking. We also show
qualitative comparisons in Fig. 3. It can be seen that both PHOSA and CHORE
are not able to accurately capture the object under heavy occlusions while our
method is more robust under these challenging occlusion cases.
### 4.2 Importance of SMPL-T conditioning
We propose to condition our interaction field predictions on SMPL-T meshes, we
evaluate this for the joint tracking task in Tab. 2. Our SMPL-T conditioning
allows us to obtain consistent relative translation instead of predicting the
human at fixed depth. This significantly reduces the error when evaluating
using alignment of temporal sliding window, see Tab. 2 and qualitative
examples in Supp.
Note that the SIF-Net prediction relies on the estimated SMPL-T discussed in
Sec. 3.2. We also evaluate how the noisy SMPL-T predictions can affect the
joint tracking performance. We input GT SMPL to our SIF-Net and HVOP-Net to
predict the object pose and perform joint optimization. The object errors
(Chamfer distance in cm) are: 17.29 (w/o SMPL-T cond.), 8.23 (w/ SMPL-T cond.,
ours), 6.50 (GT SMPL). Our method is robust and is close to the ideal lower
bound with GT SMPL.
Method | w/o SMPL-T | w/o HVOP-Net | w/o joint opt. | Ours
---|---|---|---|---
SMPL $\downarrow$ | 14.40 | 8.05 | 8.20 | 8.03
Obj. $\downarrow$ | 17.29 | 9.36 | 16.02 | 8.23
Table 2: Ablation studies. We report the joint tracking error (cm) after an
alignment window of 10s. It can be seen that our proposed SMPL-T conditioning,
HVOP-Net and joint optimization are important to achieve the best results.
### 4.3 Importance of HVOP-Net
Figure 4: Object rotation(left) and translation(right) error vs. occlusion
(1-fully occluded) for variants of our method. Our full model with HVOP-Net
predicts more robust rotation in occlusions. Figure 5: Importance of our HVOP-
Net object pose prediction. We can see that our HVOP-Net corrects erroneous
object pose under heavy occlusions. Better viewed after zoom in.
We propose a _human and visibility aware_ object pose prediction network
(HVOP-Net) to reason about the object under heavy occlusions. Without this
component, the object error is much higher (Tab. 2 column 3), which suggests
the importance of our HVOP-Net.
For the object pose prediction task, there are other similar alternatives to
HVOP-Net: 1). Linear interpolation (SLERP), infill the object pose of
invisible frames using spherical linear interpolation between two visible
frames. 2). SmoothNet [90], a fully connected neural network trained to smooth
the object motion. 3). CMIB [39], a SoTA method for human motion infilling. To
evaluate the effectiveness of our HVOP-Net, we replace HVOP-Net with these
methods and run the full tracking pipeline respectively. The SMPL errors are
similar (deviate $<$ 0.1cm) as HVOP-Net only affects objects. We separate
object error into rotation (angle distance) and translation, and further
analyse the error under varying occlusion for CHORE, our method (SMPL-T +
HVOP-Net) and its variants (SMPL-T + SmoothNet/SLERP/CMIB) in Fig. 4.
All methods except CHORE predict similar translation (due to SMPL-T cond.) but
our HVOP-Net obtains clearly more robust rotation under heavy occlusions.
SmoothNet smooths the object motion but cannot correct errors from long-term
occlusions. SLERP and CMIB [39] are able to correct some pose errors but do
not take the human context into account hence cannot handle heavy occlusions
very well. Our method leverages the human motion and object pose from visible
frames hence achieves the best result. We show one example where our HVOP-Net
corrects the erroneous raw prediction in Fig. 5. Please see our supplementary
for more examples.
### 4.4 Generalization
Figure 6: Results in NTU-RGBD dataset [47]. It can be seen that our method
generalizes well and wormks better than CHORE[80].
To verify the generalization ability of our method, we apply our model trained
on BEHAVE to the NTU-RGBD [47] dataset. We leverage Detectron [79],
interactive [64] and video [14] segmentation to obtain the input human and
object masks. Two example comparisons with CHORE [80] are shown in Fig. 6. We
can see our method generalizes well to NTU-RGBD and works better than CHORE.
Please see Supp. for evaluation details and more comparisons.
## 5 Limitations
Although our method works robustly under heavy occlusions, there are still
some limitations. For instance, we assume known object templates for tracking.
An interesting direction is to build such a template from videos as
demonstrated by recent works [82, 83, 84, 78]. Our method may fail under
challenging cases such as significant object pose change under heavy
occlusions. It can also make noisy pose predictions when the object is
symmetric or the pose is uncommon. Example failure cases are shown in Supp.
## 6 Conclusion
We present a novel method for tracking human, object and realistic contacts
between them from monocular RGB cameras. Our first contribution is a SMPL-T
conditioned neural field network that allows consistent and more accurate 3D
reconstruction in a video sequence. Our second contribution is a human motion
and object visibility aware network that can recover the object pose under
heavy occlusions. Our experiments show that our method significantly
outperforms state of the art methods in two datasets. Our extensive ablations
show that our SMPL-T conditioning and HVOP-Net are important for accurate
tracking. We also show that our method generalizes well to another dataset it
is not trained on. Our code and model are released to promote future research
in this direction.
Acknowledgements. We thank RVH group members [1] for their helpful
discussions. We also thank reviewers for their feedback which improves the
manuscript. This work is funded by the Deutsche Forschungsgemeinschaft (DFG,
German Research Foundation) - 409792180 (Emmy Noether Programme, project: Real
Virtual Humans), and German Federal Ministry of Education and Research (BMBF):
Tübingen AI Center, FKZ: 01IS18039A. Gerard Pons-Moll is a member of the
Machine Learning Cluster of Excellence, EXC number 2064/1 – Project number
390727645.
In this supplementary, we first list all the implementation details of our
method and then show more ablation study results as well as comparison with
CHORE[80] on NTU-RGBD[47] dataset. We end with discussions of failure cases
and future works.
## Appendix A Implementation details
### A.1 Obtaining SMPL-T meshes
To obtain the image-aligned SMPL meshes that have consistent translation
(SMPL-T) we keep the SMPL shape parameters and optimize the body pose and
global translation values. The loss weights for this optimization are:
$\lambda_{\text{J2D}}=0.09,\lambda_{\text{reg}}=1.0\times
10^{-5},\lambda_{a}=25,\lambda_{\text{pi}}=900$. We optimize the parameters
until convergence with a maximum iteration of 1000.
### A.2 SIF-Net: SMPL-T conditioned interaction field
A visualization of our SMPL-T triplane rendering and query point projection
can be found in Fig. 7. We discuss our network architecture and training
details next.
#### Network architecture.
We use the stacked hourglass network [51] for both RGB image encoder
$f^{\text{enc}}$ and SMPL rendering encoder $f^{\text{tri}}$. We use 3 stacks
for $f^{\text{tri}}$ and the output feature dimension is
$d^{\text{tri}}_{\text{o}}=64$. Hence $f^{\text{tri}}:\mathbb{R}^{H\times
W}\mapsto\mathbb{R}^{H/4\times W/4\times 64}$ where $H=W=512$. We also use 3
stacks for $f^{\text{enc}}$ but the feature dimension is
$d^{\text{enc}}_{\text{o}}=256$. Hence $f^{\text{enc}}:\mathbb{R}^{H\times
W\times 5}\mapsto\mathbb{R}^{H/4\times W/4\times 256}$. We also concatenate
the image features extracted from the first convolution layer and query point
coordinate to the features. Thus the total feature dimension to our decoders
is: $d=(d^{\text{tri}}_{1}+d^{\text{tri}}_{\text{o}})\times
3+d^{\text{enc}}_{1}+d^{\text{enc}}_{\text{o}}+3=611$, here
$d^{\text{tri}}_{1}=32,d^{\text{enc}}_{1}=64$. All decoders consist of three
FC layers with ReLU activation and one output FC layer with hidden dimension
of 128 for the intermediate features. The visibility decoder $f^{v}$
additionally has a sigmoid output activation layer. The output shape is
$2,14,9,3,1$ for $f^{u},f^{p},f^{R},f^{c},f^{v}$ respectively.
Figure 7: Visualization of our SMPL-T triplane feature extraction and
rendering. The triplane origin is placed at the SMPL-T body center and we
render the mesh from three views using orthographic projection: righ-left (A),
back-front (B) and top-down (C). The query point $p$ is projected into the
three planes using same projection for rendering and we extract pixel aligned
features $\mathbf{F}^{A},\mathbf{F}^{B},\mathbf{F}^{C}$ from the feature
planes respectively. Note that we render the SMPL-T with color here for
visualization, the actual input to our network are silhouette images only.
#### Training.
All feature encoders and decoders are trained end to end with the loss:
$L=\lambda_{u}(L_{u_{h}}+L_{u_{o}})+\lambda_{p}L_{p}+\lambda_{R}L_{R}+\lambda_{c}L_{c}+\lambda_{v}L_{v}$.
Here $L_{u_{i}}$ is the $L_{1}$ distance between ground truth and predicted
unsigned distance to human or object surface [80]. $L_{p}$ is a standard
categorical cross entropy loss for SMPL part correspondence prediction.
$L_{R},L_{c},L_{v}$ are mean square losses between ground truth and predicted
values for rotation matrix, translation vector and visibility score
respectively. The loss weights are:
$\lambda_{u}=1.0,\lambda_{R}=0.006,\lambda_{c}=500,\lambda_{c}=\lambda_{v}=1000$.
The model is trained for 18 epochs and it takes 25h to converge on a machine
with 4 RTX8000 GPUs each with 48GB memory. The training batch size is 8.
### A.3 HVOP-Net: object pose under occlusion
We use three transformers $f^{s},f^{o},f^{\text{comb}}$ to aggregate features
from SMPL-T, object pose and joint human object information respectively. We
use the 6D vector [96] to represent the ration matrix of SMPL-T and object
pose parameters. Hence the SMPL-T pose dimension is $24\times 6+3=147$, where
3 denotes the global translation. We predict the object rotation only thus the
object data dimension is 6. The SMPL-T transformer $f^{s}$ consists of an MLP:
$\mathbb{R}^{T\times 147}\mapsto\mathbb{R}^{T\times 128}$ and two layers of
multi-head self-attention (MHSA) module [72] with 4 heads. Similarly, the
object transformer $f^{o}$ consists of an MLP: $\mathbb{R}^{T\times
6}\mapsto\mathbb{R}^{T\times 32}$ and two layers of MHSA module with 2 heads.
The joint transformer $f^{\text{comb}}$ consists of 4 layers of MHSA module
with 1 head only. GeLU activation is used in all MHSA modules. We finally
predict the object pose using two MLP layers with an intermediate feature
dimension of 32 and LeakyReLU activation.
The model is trained to minimize the $L_{1}$ losses of pose value and
accelerations:
$L=\lambda_{\text{pose}}L_{\text{pose}}+\lambda_{\text{accel}}L_{\text{accel}}$,
where $\lambda_{\text{pose}}=1.0,\lambda_{\text{accel}}=0.1$. It is trained on
a server with 2 RTX8000 GPUs, each GPU has 48GB memory capacity. It takes
around 7h to converge (64 epochs).
### A.4 SmoothNet for SMPL-T and object
We use SmoothNet[90] to smooth our SMPL-T and SIF-Net object pose predictions.
We use exactly the same model and training strategy proposed by the original
paper. The input to the SMPL-T SmoothNet is our estimated SMPL-T pose and
translation (relative to the first frame). The input to the object SmoothNet
is the object rotation (6D vector). Following the standard practice of
SmoothNet[90], we train both models on the predictions from the BEHAVE [6]
training set. Note that we do not fine-tune them on InterCap[35] dataset. We
evaluate this component in Sec. B.3.
### A.5 Visibility aware joint optimization
The objective function defined in Eq. 2 is highly non-convex thus we solve
this optimization problem in two stages. We first optimize the SMPL pose and
shape parameters using human data term only. We then optimize the object
parameters using the object and contact data terms. The loss weights are set
to: $\lambda_{\text{reg}}=2.5\times
10^{-4},\lambda_{\text{ah}}=10^{4},\lambda_{h}=10^{4},\lambda_{p}=t\times
10^{-4},\lambda_{o}=900,\lambda_{\text{occ}}=9\times
10^{-4},\lambda_{\text{ao}}=225,\lambda_{c}=900$, where $\lambda_{c}$ is the
loss weight for the contact data term defined in Eq. 5.
## Appendix B Additional ablation results
### B.1 Further evaluation of SMPL-T conditioning
We show some example images from one sequence in Fig. 8 to evaluate the
importance of our SMPL-T conditioning. It can be seen that without this
conditioning, the human is reconstructed at fixed depth, leading to
inconsistent relative translation across time. Our method predicts more
coherent relative human translation and more accurate object pose.
To further evaluate SMPL-T conditioning, we compute the object pose error from
the raw network predictions and compare it with the object pose of CHORE which
is also the raw prediction from the network. The pose error is computed as
Chamfer distance (CD) and vertex to vertex (v2v) error after centring the
prediction and GT mesh at origin. We also report the translation error
(transl.) as the distance between predicted and GT translation. The results
are shown in Tab. 3. We can clearly see that our SMPL feature improves both
the raw object pose prediction and distance fields (results after optimization
are also improved).
Method | Raw prediction | After opt. w=10
---|---|---
CD$\downarrow$ | v2v$\downarrow$ | transl.$\downarrow$ | SMPL$\downarrow$ | obj.$\downarrow$
w/o SMPL-T | 5.56 | 16.10 | 14.28 | 14.40 | 17.29
Ours | 3.98 | 12.34 | 9.53 | 8.03 | 8.23
Table 3: Importance of SMPL-T conditioning (errors in cm). We can see that our
SMPL-T feature improves both the raw object pose prediction and distance
fields (after opt.). Without our SMPL-T conditioning, the reconstructed
translation is not consistent across frames, leading to large errors after
alignment of temporal window of 10s (w=10). Figure 8: Evaluating SMPL-T
conditioning for neural field prediction. We can see that without conditioning
on SMPL-T meshes, the object pose prediction is worse and human is
reconstructed at fixed depth, leading to inconsistent relative location across
frames. Our method recovers the relative translation more faithfully and
obtain better object pose predictions.
### B.2 Comparing different pose prediction methods
We show some example comparisons of different object pose prediction methods
under heavy occlusions in Fig. 10. We compare our method against: 1). Raw
prediction from our SIF-Net. 2). Linearly interpolate the occluded poses from
visible frames (SLERP). 3). CMIB [39], a transformer based model trained to
infill the object motion using visible frames. Note here the evaluation is
based on the final tracking results and we report the object errors only as
the difference of SMPL error is very small. Similar to Sec. B.1, the object
errors are computed as Chamfer distance, v2v error and translation error.
It can be seen that the raw pose prediction is noisy due to occlusion. SLERP
and CMIB corrects some pose errors but is not robust as they do not leverage
the human information. Our method is more accurate as it takes the human
context and object pose into account.
### B.3 Evaluating SmoothNet
Method | Chamfer | v2v | Acceleration
---|---|---|---
w/o SmoothNet | 8.71 | 9.84 | 1.38
w/ SmoothNet | 8.01 | 9.12 | 1.18
Table 4: Ablate SMPL SmoothNet (errors in cm). We can see that SmoothNet [90] improves the overall smoothness and slightly reduces the pose errors. Method | Chamfer | v2v | Translation
---|---|---|---
a. Raw prediction | 5.03 | 10.39 | 10.01
b. Raw + SmoothNet | 4.22 | 8.60 | 10.16
c. Raw + our pose pred. | 4.09 | 8.02 | 10.20
d. Our full model | 3.62 | 7.20 | 9.96
Table 5: Ablate SmoothNet for object pose prediction (errors in cm). We can
see our pose prediction (c) is better than SmoothNet [90] (b). Combing both we
obtain the best result (d).
SmoothNet [90] is used to smooth the SMPL-T parameters after 2D keypoint based
optimization. We evaluate this step by computing the SMPL errors, shown in
Tab. 4. We can see that SmoothNet reduces the SMPL error slightly.
We also use SmoothNet to smooth the object pose before sending it to our human
and visibility aware object pose prediction network. SmoothNet cannot correct
errors under long-term occlusions. However, it provides smoother object motion
for visible frames which can benefit our pose prediction network. We evaluate
this using object pose errors and report the results in Tab. 5. It can be seen
that our method (Tab. 5c) works better than SmoothNet (Tab. 5b) on raw
predictions. Nevertheless, with smoothed pose after SmoothNet, our method
achieves the best result (Tab. 5 d).
### B.4 Runtime cost
SMPL-T pre-fitting and joint optimization can be run in batches hence the
average runtime per frame is not long: SMPL-T pre-fitting: 6.38s, SIF-Net
object pose prediction: 0.89s, HVOP-Net: 1.3ms, joint optimization: 9.26s,
total: 16.53s. Compared to CHORE ($\sim$12s/frame) [80], the additional cost
is mainly from the SMPL-T pre-fitting. Yet, SMPL-T conditioning allows faster
convergence of joint optimization and much better reconstruction. Since we use
efficient 2D encoder instead of 3D encoder, it takes only 1.05GB GPU memory to
load the SIF-Net model. This allows us to do joint optimization with batch
size up to 128 on a GPU with 48GB memory.
## Appendix C Generalization to NTU-RGBD dataset
#### Obtaining input masks.
Unlike BEHAVE and InterCap where the human and object masks are provided by
the dataset, there are no masks in NTU-RGBD. To this end, we run DetectronV2
[79] to obtain the human masks. We manually segment the object in the first
frame using interactive segmentation [64] ($<$1min/image) and then use video
segmentation [14] to propagate the masks. The overhead of 1min/video manual
label is small.
We show more results from our method on NTU-RGBD dataset[47] and compare
against CHORE[80] in Fig. 11. It can be seen that CHORE may predict some
reasonable object pose but it fails quite often to capture the fine-grained
contacts between the human and object. Our method obtains more coherent
reconstruction for different subjects, human-backpack interactions, camera
view points and backgrounds. Please see our project website for comparison in
full sequences.
## Appendix D Limitations and future works
Figure 9: Failure cases analysis. We show three typical failure cases of our
method: A. The occluded object pose (T=156) changes significantly between two
visible frames (T=0 and T=319) and it is difficult to accurate track the
contact changes. B. The object pose is not commonly seen during interaction
and it is difficult to predict for this rare pose. C. The object is symmetric.
The joint optimization satisfies the object mask and contacts but is not
semantically correct.
Although our method works robustly under heavy occlusions, there are still
some limitations. Firstly, we assume known object templates for tracking, an
interesting direction is to build such a template from videos as demonstrated
by recent works [82, 83, 84, 78]. Secondly, it would be interesting to model
multi-person or even multi-object interactions which is a more realistic
setting in real-life applications. In addition, the backpack can also deform
non-rigidly which is not modelled in our method. Further works can incorporate
the surface deformation [46] or object articulation [81] into the human object
interaction. We leave these for future works.
We identify three typical failure cases of our method, some examples are shown
in Fig. 9. The first typical failure case comes from heavy occlusion when the
object undergoes significant changes (object pose and contact locations)
between two visible frames. In this case, it is very difficult to track the
pose and contact changes accurately (Fig. 9 A). Second typical failure is due
to the difficulty of pose prediction itself even the object is fully visible.
In this case the object pose is uncommon and the network failed to predict it
correctly (Fig. 9 B). Another failure is caused by symmetric objects. Our
optimization minimizes the 2D mask loss and contact constraints but the
network is confused by the symmetry and the initial pose prediction is not
semantically correct (Fig. 9 C). In addition, the training data for these
objects is very limited (only $1/3$ of other objects). More training data or
explicitly reasoning about the symmetry [94] can be helpful.
Figure 10: Comparing different object pose prediction method under heavy
occlusions. Raw prediction is from our SIF-Net output, SLERP denotes linear
interpolation and CMIB is from [39]. We can see SLERP and CMIB can correct
some errors (row 5) but they do not take the human motion into account hence
often fail in more challenging cases. Our method is more robust as it
leverages information from both human motion and object pose from visible
frames. Figure 11: Comparing our method with CHORE[80] on NTU-RGBD [47]
dataset. It can be seen that CHORE does not capture the realistic contacts
between the person and the backpack. Our method recovers the 3D human, the
object and contacts more faithfully in different interaction types, camera
view points and backgrounds.
## References
* [1] http://virtualhumans.mpi-inf.mpg.de/people.html.
* [2] Thiemo Alldieck, Marcus Magnor, Weipeng Xu, Christian Theobalt, and Gerard Pons-Moll. Detailed human avatars from monocular video. In International Conf. on 3D Vision, sep 2018.
* [3] Thiemo Alldieck, Marcus Magnor, Weipeng Xu, Christian Theobalt, and Gerard Pons-Moll. Video based reconstruction of 3D people models. In IEEE Conf. on Computer Vision and Pattern Recognition, 2018\.
* [4] Bharat Lal Bhatnagar, Cristian Sminchisescu, Christian Theobalt, and Gerard Pons-Moll. Combining implicit function learning and parametric models for 3d human reconstruction. In European Conference on Computer Vision (ECCV). Springer, August 2020.
* [5] Bharat Lal Bhatnagar, Cristian Sminchisescu, Christian Theobalt, and Gerard Pons-Moll. Loopreg: Self-supervised learning of implicit surface correspondences, pose and shape for 3d human mesh registration. In Advances in Neural Information Processing Systems (NeurIPS), December 2020.
* [6] Bharat Lal Bhatnagar, Xianghui Xie, Ilya Petrov, Cristian Sminchisescu, Christian Theobalt, and Gerard Pons-Moll. Behave: Dataset and method for tracking human object interactions. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2022.
* [7] Federica Bogo, Angjoo Kanazawa, Christoph Lassner, Peter Gehler, Javier Romero, and Michael J Black. Keep it SMPL: Automatic estimation of 3D human pose and shape from a single image. In European Conf. on Computer Vision. Springer International Publishing, 2016.
* [8] Derek Bradley, Tiberiu Popa, Alla Sheffer, Wolfgang Heidrich, and Tamy Boubekeur. Markerless garment capture. ACM Trans. Graph., 27(3):1–9, aug 2008.
* [9] Samarth Brahmbhatt, Cusuh Ham, Charles C. Kemp, and James Hays. ContactDB: Analyzing and predicting grasp contact via thermal imaging. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
* [10] Samarth Brahmbhatt, Ankur Handa, James Hays, and Dieter Fox. Contactgrasp: Functional multi-finger grasp synthesis from contact. In IROS, 04 2019.
* [11] Samarth Brahmbhatt, Chengcheng Tang, Christopher D. Twigg, Charles C. Kemp, and James Hays. ContactPose: A dataset of grasps with object contact and hand pose. In The European Conference on Computer Vision (ECCV), August 2020\.
* [12] Z. Cao, G. Hidalgo Martinez, T. Simon, S. Wei, and Y. A. Sheikh. Openpose: Realtime multi-person 2d pose estimation using part affinity fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019\.
* [13] Eric R. Chan, Connor Z. Lin, Matthew A. Chan, Koki Nagano, Boxiao Pan, Shalini De Mello, Orazio Gallo, Leonidas Guibas, Jonathan Tremblay, Sameh Khamis, Tero Karras, and Gordon Wetzstein. Efficient geometry-aware 3D generative adversarial networks. In CVPR, 2022.
* [14] Ho Kei Cheng, Yu-Wing Tai, and Chi-Keung Tang. Modular interactive video object segmentation: Interaction-to-mask, propagation and difference-aware fusion. In CVPR, 2021.
* [15] Julian Chibane, Thiemo Alldieck, and Gerard Pons-Moll. Implicit functions in feature space for 3d shape reconstruction and completion. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, jun 2020.
* [16] Julian Chibane, Aymen Mir, and Gerard Pons-Moll. Neural unsigned distance fields for implicit function learning. In Neural Information Processing Systems (NeurIPS). , December 2020.
* [17] Alvaro Collet, Ming Chuang, Pat Sweeney, Don Gillett, Dennis Evseev, David Calabrese, Hugues Hoppe, Adam Kirk, and Steve Sullivan. High-quality streamable free-viewpoint video. ACM Trans. Graph., 34(4), jul 2015.
* [18] Enric Corona, Gerard Pons-Moll, Guillem Alenya, and Francesc Moreno-Noguer. Learned vertex descent: A new direction for 3d human model fitting. In European Conference on Computer Vision (ECCV). Springer, October 2022.
* [19] Enric Corona, Albert Pumarola, Guillem Alenya, Francesc Moreno-Noguer, and Gregory Rogez. Ganhand: Predicting human grasp affordances in multi-object scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
* [20] Yudi Dai, YiTai Lin, XiPing Lin, Chenglu Wen, Lan Xu, Hongwei Yi, Siqi Shen, Yuexin Ma, and Cheng Wang. Sloper4d: A scene-aware dataset for global 4d human pose estimation in urban environments. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2023.
* [21] Xinke Deng, Arsalan Mousavian, Yu Xiang, Fei Xia, Timothy Bretl, and Dieter Fox. PoseRBPF: A Rao–Blackwellized Particle Filter for 6-D Object Pose Tracking. IEEE Transactions on Robotics, 37(5):1328–1342, Oct. 2021. Conference Name: IEEE Transactions on Robotics.
* [22] Yan Di, Fabian Manhardt, Gu Wang, Xiangyang Ji, Nassir Navab, and Federico Tombari. SO-Pose: Exploiting Self-Occlusion for Direct 6D Pose Estimation. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pages 12376–12385, Montreal, QC, Canada, Oct. 2021. IEEE.
* [23] Yinglin Duan, Yue Lin, Zhengxia Zou, Yi Yuan, Zhehui Qian, and Bohan Zhang. A unified framework for real time motion completion. Proceedings of the AAAI Conference on Artificial Intelligence, 36(4):4459–4467, Jun. 2022.
* [24] Kiana Ehsani, Shubham Tulsiani, Saurabh Gupta, Ali Farhadi, and Abhinav Gupta. Use the force, luke! learning to predict physical forces by simulating effects. In CVPR, 2020.
* [25] Zhaoxin Fan, Yazhi Zhu, Yulin He, Qi Sun, Hongyan Liu, and Jun He. Deep Learning on Monocular Object Pose Detection and Tracking: A Comprehensive Overview, Apr. 2022. arXiv:2105.14291 [cs].
* [26] Mihai Fieraru, Mihai Zanfir, Elisabeta Oneata, Alin-Ionut Popa, Vlad Olaru, and Cristian Sminchisescu. Three-dimensional reconstruction of human interactions. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7212–7221, 2020.
* [27] Vladimir Guzov, Aymen Mir, Torsten Sattler, and Gerard Pons-Moll. Human poseitioning system (hps): 3d human pose estimation and self-localization in large scenes from body-mounted sensors. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, jun 2021.
* [28] Shreyas Hampali, Mahdi Rad, Markus Oberweger, and Vincent Lepetit. Honnotate: A method for 3d annotation of hand and object poses. In CVPR, 2020.
* [29] Mohamed Hassan, Vasileios Choutas, Dimitrios Tzionas, and Michael J. Black. Resolving 3d human pose ambiguities with 3d scene constraints. In International Conference on Computer Vision, 2019.
* [30] Yana Hasson, Gül Varol, Dimitrios Tzionas, Igor Kalevatykh, Michael J. Black, Ivan Laptev, and Cordelia Schmid. Learning joint reconstruction of hands and manipulated objects. In CVPR, 2019.
* [31] Sachini Herath, Hang Yan, and Yasutaka Furukawa. Ronin: Robust neural inertial navigation in the wild: Benchmark, evaluations, new methods. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 3146–3152, 2020.
* [32] Tomáš Hodaň, Jiří Matas, and Štěpán Obdržálek. On evaluation of 6d object pose estimation. In Gang Hua and Hervé Jégou, editors, Computer Vision – ECCV 2016 Workshops, pages 606–619, Cham, 2016. Springer International Publishing.
* [33] Yinlin Hu, Pascal Fua, Wei Wang, and Mathieu Salzmann. Single-stage 6d object pose estimation. In CVPR, 2020.
* [34] Chun-Hao P. Huang, Hongwei Yi, Markus Höschle, Matvey Safroshkin, Tsvetelina Alexiadis, Senya Polikovsky, Daniel Scharstein, and Michael J. Black. Capturing and inferring dense full-body human-scene contact. In IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), pages 13274–13285, June 2022.
* [35] Yinghao Huang, Omid Taheri, Michael J. Black, and Dimitrios Tzionas. InterCap: Joint markerless 3D tracking of humans and objects in interaction. In German Conference on Pattern Recognition (GCPR), volume 13485 of Lecture Notes in Computer Science, pages 281–299. Springer, 2022\.
* [36] Yuheng Jiang, Suyi Jiang, Guoxing Sun, Zhuo Su, Kaiwen Guo, Minye Wu, Jingyi Yu, and Lan Xu. Neuralhofusion: Neural volumetric rendering under human-object interactions. arXiv preprint arXiv:2202.12825, 2022.
* [37] Hanbyul Joo, Tomas Simon, and Yaser Sheikh. Total capture: A 3d deformation model for tracking faces, hands, and bodies. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8320–8329, 2018.
* [38] Korrawe Karunratanakul, Jinlong Yang, Yan Zhang, Michael Black, Krikamol Muandet, and Siyu Tang. Grasping field: Learning implicit representations for human grasps. In 8th International Conference on 3D Vision, pages 333–344. IEEE, Nov. 2020.
* [39] Jihoon Kim, Taehyun Byun, Seungyoun Shin, Jungdam Won, and Sungjoon Choi. Conditional motion in-betweening. Pattern Recognition, page 108894, 2022.
* [40] Muhammed Kocabas, Nikos Athanasiou, and Michael J. Black. Vibe: Video inference for human body pose and shape estimation. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
* [41] Muhammed Kocabas, Chun-Hao P. Huang, Otmar Hilliges, and Michael J. Black. PARE: Part attention regressor for 3D human body estimation. In Proceedings International Conference on Computer Vision (ICCV), pages 11127–11137. IEEE, Oct. 2021.
* [42] Muhammed Kocabas, Chun-Hao P. Huang, Joachim Tesch, Lea Müller, Otmar Hilliges, and Michael J. Black. SPEC: Seeing people in the wild with an estimated camera. In Proceedings International Conference on Computer Vision (ICCV), pages 11035–11045. IEEE, Oct. 2021.
* [43] Nikos Kolotouros, Georgios Pavlakos, Michael J Black, and Kostas Daniilidis. Learning to reconstruct 3d human pose and shape via model-fitting in the loop. In ICCV, 2019.
* [44] Jake Levinson, Carlos Esteves, Kefan Chen, Noah Snavely, Angjoo Kanazawa, Afshin Rostamizadeh, and Ameesh Makadia. An analysis of svd for deep rotation estimation. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 22554–22565. Curran Associates, Inc., 2020.
* [45] Yi Li, Gu Wang, Xiangyang Ji, Yu Xiang, and Dieter Fox. Deepim: Deep iterative matching for 6d pose estimation. In European Conference on Computer Vision (ECCV), 2018.
* [46] Zhi Li, Soshi Shimada, Bernt Schiele, Christian Theobalt, and Vladislav Golyanik. Mocapdeform: Monocular 3d human motion capture in deformable scenes. In International Conference on 3D Vision (3DV), 2022.
* [47] Jun Liu, Amir Shahroudy, Mauricio Perez, Gang Wang, Ling-Yu Duan, and Alex C. Kot. Ntu rgb+d 120: A large-scale benchmark for 3d human activity understanding. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019\.
* [48] Yuan Liu, Yilin Wen, Sida Peng, Cheng Lin, Xiaoxiao Long, Taku Komura, and Wenping Wang. Gen6d: Generalizable model-free 6-dof object pose estimation from rgb images. In ECCV, 2022.
* [49] Matthew Loper, Naureen Mahmood, Javier Romero, Gerard Pons-Moll, and Michael J. Black. SMPL: A skinned multi-person linear model. In ACM Transactions on Graphics. ACM, Oct. 2015.
* [50] Mateusz Majcher and Bogdan Kwolek. Shape enhanced keypoints learning with geometric prior for 6d object pose tracking. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 2985–2991, 2022.
* [51] Alejandro Newell, Kaiyu Yang, and Jia Deng. Stacked hourglass networks for human pose estimation. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision – ECCV 2016, pages 483–499, Cham, 2016. Springer International Publishing.
* [52] Georgios Pavlakos, Vasileios Choutas, Nima Ghorbani, Timo Bolkart, Ahmed A. A. Osman, Dimitrios Tzionas, and Michael J. Black. Expressive body capture: 3d hands, face, and body from a single image. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019.
* [53] Sida Peng, Yuan Liu, Qixing Huang, Xiaowei Zhou, and Hujun Bao. PVNet: Pixel-Wise Voting Network for 6DoF Pose Estimation. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4556–4565, Long Beach, CA, USA, June 2019. IEEE.
* [54] Ilya A Petrov, Riccardo Marin, Julian Chibane, and Gerard Pons-Moll. Object pop-up: Can we infer 3d objects and their poses from human interactions alone? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
* [55] Gerard Pons-Moll and Bodo Rosenhahn. Model-Based Pose Estimation, chapter 9, pages 139–170. Springer, 2011.
* [56] Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, and Jitendra Malik. Tracking people with 3d representations. Advances in Neural Information Processing Systems, 34:23703–23713, 2021.
* [57] Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, and Jitendra Malik. Tracking people by predicting 3d appearance, location and pose. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2740–2749, 2022.
* [58] Davis Rempe, Tolga Birdal, Aaron Hertzmann, Jimei Yang, Srinath Sridhar, and Leonidas J. Guibas. Humor: 3d human motion model for robust pose estimation. In International Conference on Computer Vision (ICCV), 2021.
* [59] Chris Rockwell and David F. Fouhey. Full-body awareness from partial observations. In ECCV, 2020.
* [60] Yu Rong, Takaaki Shiratori, and Hanbyul Joo. Frankmocap: A monocular 3d whole-body pose estimation system via regression and integration. In IEEE International Conference on Computer Vision Workshops, 2021\.
* [61] Shunsuke Saito, , Zeng Huang, Ryota Natsume, Shigeo Morishima, Angjoo Kanazawa, and Hao Li. Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In IEEE International Conference on Computer Vision (ICCV). IEEE, oct 2019.
* [62] Shunsuke Saito, Tomas Simon, Jason Saragih, and Hanbyul Joo. Pifuhd: Multi-level pixel-aligned implicit function for high-resolution 3d human digitization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2020.
* [63] Soshi Shimada, Vladislav Golyanik, Zhi Li, Patrick Pérez, Weipeng Xu, and Christian Theobalt. Hulc: 3d human motion capture with pose manifold sampling and dense contact guidance. In European Conference on Computer Vision (ECCV), pages 516–533, 2022.
* [64] Konstantin Sofiiuk, Ilia Petrov, Olga Barinova, and Anton Konushin. f-brs: Rethinking backpropagating refinement for interactive segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8623–8632, 2020.
* [65] Manuel Stoiber, Martin Sundermeyer, and Rudolph Triebel. Iterative Corresponding Geometry: Fusing Region and Depth for Highly Efficient 3D Tracking of Textureless Objects. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 6845–6855, New Orleans, LA, USA, June 2022. IEEE.
* [66] Jürgen Sturm, Stéphane Magnenat, Nikolas Engelhard, François Pomerleau, Francis Colas, Daniel Cremers, Roland Siegwart, and Wolfram Burgard. Towards a benchmark for RGB-D SLAM evaluation. In RGB-D Workshop on Advanced Reasoning with Depth Cameras at Robotics: Science and Systems Conf.(RSS), Los Angeles, United States, 2011.
* [67] Guoxing Sun, Xin Chen, Yizhang Chen, Anqi Pang, Pei Lin, Yuheng Jiang, Lan Xu, Jingya Wang, and Jingyi Yu. Neural free-viewpoint performance rendering under complex human-object interactions. In Proceedings of the 29th ACM International Conference on Multimedia, 2021.
* [68] Jiaming Sun, Zihao Wang, Siyu Zhang, Xingyi He, Hongcheng Zhao, Guofeng Zhang, and Xiaowei Zhou. OnePose: One-shot object pose estimation without CAD models. CVPR, 2022.
* [69] Omid Taheri, Nima Ghorbani, Michael J. Black, and Dimitrios Tzionas. GRAB: A dataset of whole-body human grasping of objects. In European Conference on Computer Vision (ECCV), 2020.
* [70] Yating Tian, Hongwen Zhang, Yebin Liu, and Limin Wang. Recovering 3d human mesh from monocular images: A survey. arXiv preprint arXiv:2203.01923, 2022.
* [71] Garvita Tiwari, Dimitrije Antic, Jan Eric Lenssen, Nikolaos Sarafianos, Tony Tung, and Gerard Pons-Moll. Pose-ndf: Modeling human pose manifolds with neural distance fields. In European Conference on Computer Vision (ECCV). Springer, October 2022.
* [72] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
* [73] Gu Wang, Fabian Manhardt, Federico Tombari, and Xiangyang Ji. GDR-Net: Geometry-guided direct regression network for monocular 6d object pose estimation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 16611–16621, June 2021.
* [74] Xi Wang, Gen Li, Yen-Ling Kuo, Muhammed Kocabas, Emre Aksan, and Otmar Hilliges. Reconstructing action-conditioned human-object interactions using commonsense knowledge priors. In International Conference on 3D Vision (3DV), 2022.
* [75] Bowen Wen and Kostas Bekris. BundleTrack: 6D Pose Tracking for Novel Objects without Instance or Category-Level 3D Models. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 8067–8074, Prague, Czech Republic, Sept. 2021. IEEE Press.
* [76] Bowen Wen, Chaitanya Mitash, Baozhang Ren, and Kostas E. Bekris. se(3)-tracknet: Data-driven 6d pose tracking by calibrating image residuals in synthetic domains. 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Oct 2020.
* [77] Zhenzhen Weng and Serena Yeung. Holistic 3d human and scene mesh estimation from single view images. arXiv preprint arXiv:2012.01591, 2020.
* [78] Shangzhe Wu, Tomas Jakab, Christian Rupprecht, and Andrea Vedaldi. DOVE: Learning deformable 3d objects by watching videos. arXiv preprint arXiv:2107.10844, 2021.
* [79] Yuxin Wu, Alexander Kirillov, Francisco Massa, Wan-Yen Lo, and Ross Girshick. Detectron2. https://github.com/facebookresearch/detectron2, 2019.
* [80] Xianghui Xie, Bharat Lal Bhatnagar, and Gerard Pons-Moll. Chore: Contact, human and object reconstruction from a single rgb image. In European Conference on Computer Vision (ECCV). Springer, October 2022.
* [81] Xiang Xu, Hanbyul Joo, Greg Mori, and Manolis Savva. D3d-hoi: Dynamic 3d human-object interactions from videos. arXiv preprint arXiv:2108.08420, 2021.
* [82] Gengshan Yang, Deqing Sun, Varun Jampani, Daniel Vlasic, Forrester Cole, Huiwen Chang, Deva Ramanan, William T Freeman, and Ce Liu. Lasr: Learning articulated shape reconstruction from a monocular video. In CVPR, 2021.
* [83] Gengshan Yang, Deqing Sun, Varun Jampani, Daniel Vlasic, Forrester Cole, Ce Liu, and Deva Ramanan. Viser: Video-specific surface embeddings for articulated 3d shape reconstruction. In NeurIPS, 2021.
* [84] Gengshan Yang, Minh Vo, Natalia Neverova, Deva Ramanan, Andrea Vedaldi, and Hanbyul Joo. Banmo: Building animatable 3d neural models from many casual videos. In CVPR, 2022.
* [85] Lixin Yang, Xinyu Zhan, Kailin Li, Wenqiang Xu, Jiefeng Li, and Cewu Lu. CPF: Learning a contact potential field to model the hand-object interaction. In ICCV, 2021.
* [86] Hongwei Yi, Chun-Hao P. Huang, Shashank Tripathi, Lea Hering, Justus Thies, and Michael J. Black. MIME: Human-aware 3D scene generation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2023.
* [87] Hongwei Yi, Chun-Hao P. Huang, Dimitrios Tzionas, Muhammed Kocabas, Mohamed Hassan, Siyu Tang, Justus Thies, and Michael J. Black. Human-aware object placement for visual environment reconstruction. In IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), pages 3959–3970, June 2022.
* [88] Hongwei Yi, Hualin Liang, Yifei Liu, Qiong Cao, Yandong Wen, Timo Bolkart, Dacheng Tao, and Michael J Black. Generating holistic 3d human motion from speech. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2023.
* [89] Ye Yuan, Umar Iqbal, Pavlo Molchanov, Kris Kitani, and Jan Kautz. Glamr: Global occlusion-aware human mesh recovery with dynamic cameras. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
* [90] Ailing Zeng, Lei Yang, Xuan Ju, Jiefeng Li, Jianyi Wang, and Qiang Xu. Smoothnet: A plug-and-play network for refining human poses in videos. In European Conference on Computer Vision. Springer, 2022.
* [91] Jason Y. Zhang, Sam Pepose, Hanbyul Joo, Deva Ramanan, Jitendra Malik, and Angjoo Kanazawa. Perceiving 3d human-object spatial arrangements from a single image in the wild. In European Conference on Computer Vision (ECCV), 2020.
* [92] Tianshu Zhang, Buzhen Huang, and Yangang Wang. Object-occluded human shape and pose estimation from a single color image. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7374–7383, 2020.
* [93] Xiaohan Zhang, Bharat Lal Bhatnagar, Sebastian Starke, Vladimir Guzov, and Gerard Pons-Moll. Couch: Towards controllable human-chair interactions. In European Conference on Computer Vision (ECCV). Springer, October 2022.
* [94] Linfang Zheng, Aleš Leonardis, Tze Ho Elden Tse, Nora Horanyi, Hua Chen, Wei Zhang, and Hyung Jin Chang. TP-AE: Temporally Primed 6D Object Pose Tracking with Auto-Encoders. In 2022 International Conference on Robotics and Automation (ICRA), pages 10616–10623, Philadelphia, PA, USA, May 2022. IEEE Press.
* [95] Keyang Zhou, Bharat Lal Bhatnagar, Jan Eric Lenssen, and Gerard Pons-Moll. Toch: Spatio-temporal object correspondence to hand for motion refinement. In European Conference on Computer Vision (ECCV). Springer, October 2022.
* [96] Yi Zhou, Connelly Barnes, Lu Jingwan, Yang Jimei, and Li Hao. On the continuity of rotation representations in neural networks. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
|
# Gradient estimates of nonlinear equation on complete noncompact metric
measure space with compact boundary
Xiang-Zhi Cao School of information engineering, Nanjing Xiaozhuang
University, Nanjing 211171, ChinaEmail<EMAIL_ADDRESS>
|
11institutetext: Delft University of Technology, Delft, The Netherlands
11email: {L.Lyu, N.Roy<EMAIL_ADDRESS>22institutetext: Radboud
University, Nijmegen, The Netherlands
22email<EMAIL_ADDRESS>
# Is Interpretable Machine Learning Effective at Feature Selection for Neural
Learning-to-Rank?
Lijun Lyu 11 0000-0002-7268-4902 Nirmal Roy 11 0000-0003-0860-5269 Harrie
Oosterhuis 22 0000-0002-0458-9233 Avishek Anand 11 0000-0002-0163-0739
###### Abstract
Neural ranking models have become increasingly popular for real-world search
and recommendation systems in recent years. Unlike their tree-based
counterparts, neural models are much less interpretable. That is, it is very
difficult to understand their inner workings and answer questions like how do
they make their ranking decisions? or what document features do they find
important? This is particularly disadvantageous since interpretability is
highly important for real-world systems. In this work, we explore feature
selection for neural learning-to-rank (LTR). In particular, we investigate six
widely-used methods from the field of interpretable machine learning (ML) and
introduce our own modification, to select the input features that are most
important to the ranking behavior. To understand whether these methods are
useful for practitioners, we further study whether they contribute to
efficiency enhancement. Our experimental results reveal a large feature
redundancy in several LTR benchmarks: the local selection method TabNet can
achieve optimal ranking performance with less than 10 features; the global
methods, particularly our G-L2x, require slightly more selected features, but
exhibit higher potential in improving efficiency. We hope that our analysis of
these feature selection methods will bring the fields of interpretable ML and
LTR closer together.
## 1 Introduction
Learning-to-rank (LTR) is at the core of many information retrieval (IR) and
recommendation tasks [38]. The defining characteristic of LTR, and what
differentiates it from other machine learning (ML) areas, is that LTR methods
aim to predict the optimal ordering of items. This means that LTR methods are
not trying to estimate the exact relevance of an item, but instead predict
relative relevance differences, i.e., whether it is more or less relevant than
other items. Traditionally, the most widely adopted and prevalent LTR methods
were based on Gradient Boosted Decision Trees (GBDT) [12, 65, 29]. However, in
recent years, neural LTR methods have become increasingly popular [21, 46,
47]. Recently, [51] have shown that neural models can provide ranking
performance that is comparable, and sometimes better, than that of state-of-
the-art GBDT LTR models on established LTR benchmark datasets [49, 14, 17]. It
thus seems likely that the prevalence of neural LTR models will only continue
to grow in the foreseeable future.
Besides the quality of the results that ranking systems return, there is an
increasing interest in building trustworthy systems through interpretability,
e.g., by understanding which features contribute the most to ranking results.
Additionally, the speed at which results are provided is also highly important
[5, 7, 3]. Users expect ranking systems to be highly responsive and previous
work indicates that even half-second increases in latency can contribute to a
negative user experience [7]. A large part of ranking latency stems from the
retrieval and computation of input features for the ranking model.
Consequently, _feature selection_ for ranking systems has been an important
topic in the LTR field [22, 23, 59, 52, 45, 68, 48]. These methods reduce the
number of features used, thereby helping users understand and greatly reduce
latency and infrastructure costs, while maintaining ranking quality as much as
possible. In line with the history of the LTR field, existing work on feature
selection has predominantly focused on GBDT and support-vector-machine (SVM)
ranking models [26, 21], but has overlooked neural ranking models. To the best
of our knowledge, only two existing works have looked at feature selection for
neural LTR [48, 52]. This scarcity is in stark contrast with the importance of
feature selection and the increasing prevalence of neural models in LTR.
Outside of the LTR field, feature selection for neural models has received
much more attention, for the sake of efficiency [35, 36], and also to better
understand the model behaviours [72, 4]. Those methods mainly come from the
_interpretable_ ML field [18, 44], where the idea is that the so-called
_concrete_ feature selection can give insights into what input information a
ML model uses to make decisions. This tactic has already been successfully
applied to natural language processing [71], computer vision [6], and tabular
data [69, 4]. Accordingly, there is a potential for these methods to also
enable embedded feature selection for neural LTR models, where the selection
and prediction are optimized simultaneously. However, the effectiveness of
these interpretable ML methods for LTR tasks is currently unexplored, and
thus, it remains unclear whether their application can translate into useful
insights for LTR practitioners.
The goal of this work is to investigate whether six prevalent feature
selection methods – each representing one of the main branches of
interpretable ML field – can be applied effectively to neural LTR. In
addition, we also propose a novel method with minor modifications. Our aim is
to bridge the gap between the two fields by translating the important concepts
of the interpretable ML field to the LTR setting, and by demonstrating how
interpretable ML methods can be adapted for the LTR task. Moreover, our
experiments consider whether these methods can bring efficiency into the
practical application by reducing irrelevant input features for neural ranking
models.
Our results reveal a large feature redundancy in LTR benchmark datasets, but
this redundancy can be understood differently for interpretability and for
efficiency: For understanding the model, feature selection can vary per
document and less than 10 features are required to approximate optimal ranking
behavior. In contrast, for practical efficiency purposes, the selection should
be static, and then 30% of features are needed. We conclude that – when
adapted for the LTR task – not all, but a few interpretable ML methods lead to
effective and practical feature selection for neural LTR.
To the best of our knowledge, this is the first work that extensively studies
embedded feature selection for neural LTR. We hope our contributions bring
more attention to the potential of interpretable ML for IR field. To stimulate
future work and enable reproducibility, we have made our implementation
publicly available at:
https://github.com/GarfieldLyu/NeuralFeatureSelectionLTR (MIT license).
## 2 Related Work
Learning-to-Rank (LTR). Traditional LTR algorithms mainly rely on ML models,
such as SVMs and decision trees to learn the correlation between numerical
input features and human-annotated relevance labels [37, 15, 20, 70, 67, 65,
31, 27]. Neural approaches [11, 10, 13, 66, 54, 61] have also been proposed,
but did not show significant improvements over traditional non-neural models.
Inspired by the transformer architecture [63], recent works have also adapted
self-attention [46, 47, 51] and produced the neural LTR methods that
outperform LambdaMART [65], albeit with a relatively small difference. It
shows that neural rankers can provide competitive performance, consequently,
the interest and effort towards neural models for LTR are expected to increase
considerably in the near future.
Efficiency is crucial in real-world systems since users expect them to be
highly responsive [5, 7, 3]. Aside from model execution, the latency of
ranking system is largely due to feature construction, as it happens _on-the-
fly_ for incoming queries. Thus, efficiency is often reached by reducing
(expensive) features. Previous works [21, 64] apply a cascading setup to
reduce the usage of expensive features. Another growing trend in LTR is to
design interpretable models. Existing methods rely on specific architecture
design, such as general additive model (GAM) [73] or a feature-interaction
constrained and depth-reduced tree model [39].
Feature Selection for LTR. Feature selection can achieve both efficiency and
interpretability [4, 35, 36, 72]. By selecting a subset of input features, the
input complexity of models is reduced while maintaining competitive
performance. This helps with (1) efficiency as it avoids unnecessary
construction of features [35, 36], and (2) interpretability as fewer input
features are involved in prediction [72, 4].
Existing feature selection methods in LTR are classified commonly as filter,
wrapper and embedded methods [22, 23, 48]. Filter and wrapper methods are
applied to given static ranking models which are not updated in any way;
filter methods are model-agnostic [22] while wrapper methods are designed for
a particular type of model [23]. In this work we will focus on the third
category, embedded methods, where feature selection is performed
simultaneously with model optimization. Most embedded methods are limited to
particular model designs such as SVMs [32, 33, 34] or decision trees [45, 68,
40]. To the best of our knowledge, only two methods are designed for neural
LTR [52, 48]: one applies group regularization methods [52] to reduce both
input and other model parameters; the other [48] uses the gradients of a
static ranking model to infer feature importance, and thus it belongs to the
_filter_ category. We do not investigate these two methods further, as the
focus of this work is on _embedded_ input feature selection methods.
Interpretable Machine Learning. The earliest work in interpretable ML
attempted to explain a trained model in _post-hoc_ manner, mainly relying on
input perturbations [53, 41], gradients [58, 60] and so on [57]. In parallel,
more recent works advocated intrinsically interpretable models, that are
categorized as _interpretable-by-design_ methods [56, 2, 1]. For neural
networks, explaining the decision path is challenging due to the large set of
parameters. Therefore, the more prevalent choice for intrinsic interpretable
neural models is to shift the transparency to the input features. Namely, the
final prediction comes from a subset selection of input elements, e.g., words
or pixels and the rest irrelevant features are masked out [71, 16, 36].
Importantly, this selection decision can be learned jointly with the
predictive accuracy of a model. Thereby, we limit our research focus in
intrinsic interpretable ML models.
Due to the discrete nature of selection, many approaches such as L2X [16],
Concrete AutoEncoders (CAE) [6], Instance-wise Feature grouping (IFG) [43]
apply Gumbel-Softmax sampling [24] to enable backpropagation through the
feature selection process. Alternatively, regularization is also a commonly-
used feature selection approach in traditional ML algorithms [62, 31], and is
applicable to neural models, i.e., with INVASE [69] or LassoNet [35].
Moreover, TabNet [4] applies both regularization and the sparsemax activation
function [42] to realize sparse selection. These approaches have been
successfully applied in language, vision and tabular domains, and suggested
that the resulting feature selections substantially improved the user
understanding of models and datasets [28, 55].
Despite their success in other domains, we find that the above-mentioned
feature selection methods for neural models (L2X, CAE, IFG, INVASE, LassoNet
and TabNet) have not been studied in the LTR setting. In response, we hope to
bridge this gap between the interpretable ML and the LTR field by adapting and
applying these methods to neural ranking models.
## 3 Background
### 3.1 Learning-to-Rank (LTR)
The LTR task can be formulated as optimizing a scoring function $f$ that given
item features $x$ predicts an item score $f(x)\in\mathbb{R}$, so that ordering
items according to their scores corresponds to the optimal ranking [38].
Generally, there are relevance labels $y$ available for each item, often these
are labels provided by experts where $y\in\\{0,1,2,3,4\\}$ [14, 50, 49]. Given
a training set $\mathcal{D}_{q}=\\{(x_{i},y_{i})\\}^{N_{q}}_{i=1}$ for a
single query $q$, optimization is done by minimizing a LTR loss, for instance,
the _softmax cross entropy loss_ [9, 13]:
$\mathcal{L}(f\mid\mathcal{D}_{q})=-\frac{1}{|\mathcal{D}_{q}|}\underset{(x,y)\in\mathcal{D}_{q}}{\sum}\sum_{i=1}^{N_{q}}y_{i}\log\sigma(x_{i}\mid
f,\mathcal{D}_{q}),$ (1)
where $\sigma$ is the softmax activation function:
$\sigma(x\mid
f,\mathcal{D}_{q})=\frac{\exp(f(x))}{\sum_{x^{\prime}\in\mathcal{D}_{q}}\exp(f(x^{\prime}))}.$
(2)
The resulting $f$ is then commonly evaluated with a ranking metric, for
instance, the widely-used normalized discounted cumulative gain metric (NDCG)
[25].
### 3.2 Properties of Feature Selection Methods
Table 1: Properties of feature selection methods from the interpretable ML field as discussed in Section 3. Method | Global | Local | Sampling | Regularization | Fixed-Budget | Composable
---|---|---|---|---|---|---
L2x [16] | | ✓ | ✓ | | ✓ | ✓
Invase [69] | | ✓ | ✓ | ✓ | | ✓
Cae [6] | ✓ | | ✓ | | ✓ | ✓
Ifg [43] | | ✓ | ✓ | | | ✓
LassoNet [35] | ✓ | | | ✓ | | ✓
TabNet [4] | | ✓ | | ✓ | |
G-L2x (ours) | ✓ | | ✓ | | ✓ | ✓
As discussed before, feature selection is used in the interpretable ML field
to better understand which input features ML models use to make their
predictions. Furthermore, feature selection is also important to LTR for
increasing the efficiency of ranking systems. However, selecting a subset of
input features without compromising the model performance is an NP-hard
problem, since the number of possible subsets grows exponentially with the
number of available features [23, 48]. As a solution, the interpretable ML
field has proposed several methods that approach feature selection as an
optimization problem. We will now lay out several important properties that
can be used to categorize these methods, which will be elaborated in next
section.
Global vs. local. Global methods select a single subset of features for the
entire dataset, whereas local methods can vary their selection over different
items.
Composable vs. non-composable. Non-composable methods are designed for a
specific model architecture, and therefore, they can only perform feature
selection for those models. Conversely, composable methods are not constrained
to specific architectures, and thus, they work for any (differentiable) model.
Fixed-budget vs. budget-agnostic. Fixed-budget methods work with a pre-defined
selection budget, i.e., what number of features should be selected in total,
or a cost per feature and a maximum total cost for the selection. Their
counterparts are budget-agnostic methods that do not use an explicit budget,
consequently, one has to carefully fine-tune their hyper-parameters to achieve
a desired performance-sparsity trade-off.
Sampling-based vs. regularization-based. As their names imply, sampling-based
methods optimize a sampling procedure to perform the feature selection,
whereas regularization-based methods use an added regularization loss to
stimulate sparsity in feature selection. While these groups apply very
different approaches, whether one is significantly more useful for LTR
purposes than the other remains unknown.
## 4 Feature Selection from Interpretable ML for LTR
Figure 1: Methods overview, as described in Section 4.
In this section, we present a brief technical overview of our selection of six
interpretable ML methods and their adaption to neural LTR models, and propose
our G-L2x method based on a minor modification. While any ranking loss can be
chosen, we use a listwise softmax cross entropy (Eq. 1) with all methods, for
the sake of simplicity. Therefore, the training of each query is conducted
after generating the output of all documents associated with the query. Table
1 highlights the properties of all methods and Figure 1 provides a visual
overview to accompany this section.
### 4.1 Sampling-based Feature Selection
Sampling-based approaches use a two-stage architecture consisting of a
_selector_ that generates a sparse selection over the input features; and a
_ranker_ that only takes selected features as its input, in the form of a
masked vector $\hat{x}$.
The training of a ranker follows conventional LTR, i.e., Eq 1 with $x_{i}$
replaced by $\hat{x_{i}}$. But the optimization of a selector ($\zeta$) is not
as straightforward; Usually, $\zeta$ constructs a probability distribution
$\mathbf{p}=[p_{1},p_{2},\cdots p_{d}]$, indicating a selection probability
per feature. However, the ranker uses a concrete selection $m\in\\{0,1\\}^{d}$
from the probability distribution, and this concrete operation does not allow
optimization of the selector via backpropagation. The common solution is to
generate a differentiable approximation $\tilde{m}$, by _concrete relaxation_
or the Gumbel-Softmax trick [24]. Namely, the selection of $p_{i}$ can be
approximated with the differentiable $c_{i}$ as:
$c_{i}=\frac{exp\\{(\log p_{i}+g_{i})/\tau\\}}{\sum_{j=1}^{d}exp\\{(\log
p_{j}+g_{i})/\tau\\}},$ (3)
where $g$ is the Gumbel noise and $\tau\in\mathbb{R}^{>0}$ is the temperature
parameter. Now, the selector $\zeta$ can be optimized with stochastic gradient
descent by using $\tilde{m}$. The following four sampling-based methods apply
this overall procedure, but differ in how they generate $\mathbf{p}$ and
$\tilde{m}$.
Learning to explain (L2x). L2x [16] is a local selection method since its
neural selector generates a probability distribution $\mathbf{p}$ for each
individual input instance. To generate $\tilde{m}$, L2x repeats the sampling
procedure $k$ times (Eq. 3), and subsequently, uses the maximum $c_{i}$ out of
the $k$ repeats for the $i_{th}$ element in $\tilde{m}$. The intention behind
this maximization step is to make the top-$k$ important features more likely
to have high probability scores (ideally close to 1).
Global learning to explain (G-L2x, ours). As a counterpart, we propose a
global method G-L2xbased on L2x. Our change is straightforward, where L2x
generates a different distribution $\mathbf{p}$ for each item, we apply the
same $\mathbf{p}$ to all items. In other words, G-L2x includes a global
selector layer $\zeta$ ($\zeta\in\mathbb{R}^{d}$) to simulate $\mathbf{p}$,
and sampling is conducted in the same way as L2x on the selector weights.
Thereby, G-L2x will select the same features for all items in the dataset.
Concrete autoencoder (Cae). Cae [6] is a global method where the selector is
the encoder part of an auto-encoder model [30]. Specifically, the selector
compresses the input into a smaller representation $\hat{x}$, by linearly
combining selected features, i.e. $x^{\top}\tilde{m}$, where
$\tilde{m}\in\mathbb{R}^{k\times d}$ can be viewed as approximated k-hot
concrete selection, sampled from the selector weights
($\zeta\in\mathbb{R}^{k\times d}$). Therefore, Cae might result in repetitive
selection, and the input dimension to the predictor is reduced to $k$.
Instance-wise Feature Grouping (Ifg). Ifg [43] applies a similar approach as
L2x, but clusters features into groups and then selects $k$ feature groups for
prediction. Ifg first assigns a group for each feature via Gumbel-sampling,
and then makes a feature selection by Gumbel-sampling $k$ out of the resulting
groups. This grouping decision is also guided by how rich the selected
features are to recover the original input. Therefore, apart from the ranking
objective, Ifg jointly optimizes an additional input restoring objective as
well (similar to auto-encoders [30]). Ifg is agnostic to the number of
selected features and the group sizes, it can produce oversized groups and
very large selections.
### 4.2 Regularization-based Feature Selection
Instead of the budget-explicit feature selection, regularization-based methods
induce sparsity through implicit constraints enforced by regularization terms
in the training objective. We propose modifications to three existing methods
to make them applicable to the LTR setting.
Invase. We consider Invase [69] to be a hybrid approach involving both
sampling and regularization. Built on the same structure as L2x, the selector
of Invase generates a boolean/hard mask $m$ (instead of the approximation
$\tilde{m}$) via Bernoulli sampling to train the predictor. Since this
disables backpropagation, Invase uses a customized loss function that does not
need the gradients from the predictor to train the selector. The idea is to
apply another individual baseline predictor model that takes the full-feature
input, simultaneously with the predictor that takes the masked input. The loss
difference between the two predictors is used as a scale to train the
selector. Meanwhile, the L1 regularization is applied to the selector output
to enforce selection sparsity. Ultimately, Invase will push the selector to
output a small set of selections which leads to the most similar predictions
as using all features.
LassoNet. As the name suggests, LassoNet [35] adapts a traditional Lasso (L1
regularization) [62] on the first layer of a neural model to eliminate
unnecessary features. The challenge with neural models is, all weights
corresponding to a particular feature entry in the layer have to be zero in
order to mask out the feature. Towards this, LassoNet adds a residual layer
with one weight per input feature to the original model to perform as the
traditional Lasso. Then, after every optimization step, LassoNet develops a
proximal optimization algorithm to adjust the weights of the first layer, so
that all absolute elements of each row are smaller than the respective weight
of residual layer corresponding to a specific feature. Thereby, LassoNet
performs global selection and the sparsity scale is adjusted by the L1
regularization on the residual layer weights.
TabNet. Unlike the previous methods, TabNet [4] is non-composable and tied to
a specific tree-style neural architecture. It imitates a step-wise selection
process before it outputs the final prediction based only on the selected
features. Each step has the same neural component/block but with its own
parameters, thus the model complexity and selection budget grow linearly with
the number of steps. At each step, the full input is transformed by a feature
transformer block first, and then an attentive transformer block conducts
feature selection by sparsemax activation [42], as the weights in the
resulting distribution corresponding to unselected features are zeros. The
final prediction is aggregated from all steps to simulate ensemble models. A
final mask $m$ is a union of selections from all steps, and the entropy of the
selection probabilities is used as the sparsity regularization.
## 5 Experimental Setup
Since feature selection can be applied in various manners and situations, we
structure our experiments around three scenarios:
* •
Scenario 1: Simultaneously train and select. Both the ranking model and the
feature selection are learned once and jointly. The methods are evaluated by
the performance-sparsity trade-off. It is the standard setup for evaluating
embedded feature selection methods in the interpretable ML field [52, 43].
* •
Scenario 2: Train then select with an enforced budget. Practitioners generally
set hard limits to the computational costs a system may incur and the
efficiency of the system can be greatly enhanced if it only requires a much
smaller amount of features to reach competitive performance. Following the
previous scenario, we evaluate the trained model with test instances where
only a fixed amount of features (which the method deems important and selects
frequently during training) are presented and the rest are masked out. The
resulting ranking performance and the costs of computing the required features
indicate how practical the method is in efficiency improvements.
Datasets and preprocessing. We choose three public benchmark datasets: MQ2008
(46 features) [50], Web30k (136 features) [49] and Yahoo (699 features) [14],
to cover varying numbers of available features. We apply a $\log_{1p}$
transformation to the features of Web30k, as suggested in [51]. Yahoo contains
cost labels for each feature, for Web30k we use cost estimates suggested by
previous work [21].111MQ2008 is omitted from cost analysis since no associated
cost information is available. All reported results are evaluated on the held-
out test set partitions of the datasets.
Models. We use a standard feed-forward neural network with batch
normalization, three fully-connected layers and tanh activation as the ranking
model, denoted as Dnn. According to the findings in [51], this simple model
performs closely to the most effective transformer-based models, but requires
much less resources to run. The selector models of L2x and Invase have the
same architecture, and as the only exception, TabNet is applied with its own
unalterable model (see Section 4).
Implementation. Our experimental implementation is done in PyTorch Lightning
[19]. For TabNet and LassoNet existing implementations were
used.222https://github.com/dreamquark-ai/tabnet; https://github.com/lasso-
net/lassonet We created our own implementations for the rest of the methods.
Table 2: Results of ranking performance and feature sparsity for methods
applied in Scenario 1. For comparison, we also include Gbdt [29, 8] and Dnn
baselines without feature selection as upper bound. #F denotes the number of
selected features. Reported results are averaged over $5$ random seeds (std in
parentheses). Bold font indicates the highest performing selection method; the
$\star$ and underlines denote scores that are _not_ significantly outperformed
by Gbdt and the bold-score method, respectively ($p>0.05$, paired t-tests
using Bonferonni’s correction).
Listwise | MQ2008 NDCG@k (%) | Web30k NDCG@k (%) | Yahoo NDCG@k (%)
---|---|---|---
loss | @1 | @10 | #F | @1 | @10 | #F | @1 | @10 | #F
Without feature selection.
Gbdt | 69.3 (2.5) | 80.8 (1.7) | 46 | 50.4 (0.1) | 52 (0.1) | 136 | 72.2 (0.1) | 79.2 (0.1) | 699
Dnn | 66.2⋆ (2) | 80.2⋆ (0.6) | 46 | 46.1 (0.6) | 47.7 (0.2) | 136 | 69.4 (0.3) | 76.9 (0.1) | 699
Fixed-budget feature selection using the Dnn ranking model. |
Cae | 63.0 (1.1) | 78.7 (0.5) | 4 | 32.9 (2.9) | 36.6 (2.2) | 13 | 59.2 (0.2) | 69.5 (0.1) | 6
G-L2x | 63.8⋆ (1.3) | 79.1 (0.4) | 4 | 41.1 (0.9) | 44.4 (0.3) | 13 | 65.4 (0.1) | 74.0 (0.0) | 6
L2x | 63.0 (2.1) | 78.7 (0.7) | 4 | 34.5 (2.4) | 39.7 (1.9) | 13 | 61.9 (1.1) | 73.2 (0.3) | 6
Budget-agnostic feature selection using the Dnn ranking model. |
Invase | 62.5 (2.2) | 77.5 (2.1) | 5 (2) | 15.1 (0.0) | 22.1 (0.0) | 0 | 38.7 (0.0) | 57.8 (0.0) | 0
Ifg | 66.4⋆ (0.9) | 80.4⋆ (0.5) | 20 (2) | 32.5 (5.3) | 37.5 (5.3) | 72 (30) | 69.6 (0.2) | 77.1 (0.2) | 58 (3)
LassoNet | 64.7⋆ (2.2) | 79.3 (1.2) | 6 (3) | 39.4 (0.8) | 42.1 (0.3) | 8 (2) | 63.1 (2.3) | 72.4 (1.5) | 12 (4)
Budget-agnostic feature selection using a method-specific ranking model. |
TabNet | 64.7⋆ (2.7) | 78.2 (1.2) | 7 (3) | 47.0 (0.4) | 49.2 (0.1) | 8 (1) | 70.2 (0.4) | 77.7 (0.1) | 6 (1)
| MQ2008 | Web30k | Yahoo
---|---|---|---
NDCG@10 | | |
Figure 2: Results of three fixed-budget methods applied to scenario 1. The
x-axis indicates the pre-specified percentile of selected features ($k$). The
shaded area shows the standard deviation over 5 random seeds.
## 6 Results
We report the findings in this section, aiming to answer two questions: (1)
how effective are investigated methods in the ranking setup? and (2) can those
methods improve efficiency? Each question corresponds to one of the scenarios
described in Section 5.
Simultaneous Optimization and Selection. We begin by investigating the
effectiveness of the feature selection methods when applied to Scenario 1,
where feature selection and model optimization are performed simultaneously.
The results for this scenario are displayed in Table 2 and Figure 2.
Table 2 shows the ranking performance and the respective feature sparsity of
all feature selection methods and two baselines without any selection as the
upper-bound reference. For fixed-budget methods, the budgets were set to 10%
of the total number of features for MQ2008 and Web30k, and 1% for Yahoo (the
results with varying budgets are displayed in Figure 2). Since the sparsity of
budget-agnostic methods is more difficult to control, we performed an
extensive grid search and used the hyper-parameters that produced the highest
ranking performance with a comparable feature sparsity as the other methods.
The results in Table 2 show that not all feature selection methods are equally
effective, and their performance can vary greatly over datasets. For instance,
on MQ2008 all methods perform closely to the baselines, with only a fraction
of the features. However, this is not the case for bigger datasets like Web30k
and Yahoo. In particular, Invase selects no features at all due to big
uncertainty in selection (for this reason, we omit Invase from all further
comparisons). On the other hand, Ifg performs poorly in inducing sparsity,
mainly because of its input reconstruction objective, whereas the ranking
performance is not substantially better than the rest of methods.
Additionally, Cae does not seem effective either, and furthermore, increasing
the selected features does not always result in better ranking performance
(cf. Figure 2). This is most-likely because Cae samples with replacement, and
thus the same features can be selected repeatedly.
In contrast, the other two sampling-based methods L2x and G-L2x are designed
to avoid the repetitive selection issue. Overall, the global selection G-L2x
outperforms the local counterpart L2x, possibly because global selection
generalizes better to unseen data. Another global method LassoNet is also
inferior to G-L2x, mainly due to the difficulties in sparsity weight tuning
and manually adjusting weights in the input layer.
Lastly, TabNet shows the best performance-sparsity balance across all
datasets, and even outperforms the Dnn baseline. Although, the comparison
between TabNet and Dnn is not completely fair, as they optimize different
neural architectures. It does reveal large feature redundancies in these
datasets: TabNet uses <10% of features on Web30k and 1% on Yahoo, yet still
beats the Dnn baseline with all features.
To summarize, we find that the local method TabNet is the most effective at
balancing ranking performance and sparsity. Slightly inferior but competitive
enough is the global method G-L2x, which reached $>95\%$ of baseline
performance with only 1% features on Yahoo and $>93\%$ with 10% on Web30k.
Feature Cost | Ranking Performance (NDCG@10)
---|---
Web30k | Yahoo | Web30k | Yahoo
| | |
Figure 3: Scenario 2. Feature cost (left two) and ranking performance (right
two) under incomplete input. The x-axis indicates how many percentages of
features are present in the input, to test the trained ranking model. Note
this differs from specifying $k$ during training for fixed-budget methods in
scenario 1.
Feature Selection for Trained Ranking Models. Next, we evaluate the methods in
Scenario 2, where only a specified budget (i.e., a given number of features)
of features are present in the test input. Figure 3 displays both the ranking
performance and the total feature cost for varying degrees of sparsity. The
costs represent the time it requires to retrieve the selected feature sets,
and allow us to estimate the actual efficiency improvements they provide.
Unlike previous scenario, all local methods including TabNet, are no longer
able to maintain superior performance. This is because for local methods, the
selection is made conditioned on full input information, and an incomplete
input could affect the selection and thus disrupt its prediction performance.
In contrast, global methods are immune to input changes. Therefore, Cae is
still not performing well as it did in Scenario 1; G-L2x and LassoNet provide
the best overall performance under small costs. LassoNet maintains baseline
performance with less than 40% of features on both datasets, while G-L2x
outperforms LassoNet when selected features are less than 30%. Meanwhile, it
also shows LassoNet tends to select more costly features than G-L2x.
To conclude, we find that global methods G-L2x and LassoNet perform the best
in Scenario 2, where upcoming query inputs are masked under enforced feature
budgets. Particularly, G-L2x is superior in both ranking and computing cost
when the feature budget is small. This translates to substantial efficiency
improvements in practical terms, as ranking performance is maintained by
selected features only.
## 7 Conclusion
The main goal of this work is to bring the interpretable ML and the LTR fields
closer together. To this end, we have studied whether feature selection
methods from the interpretable ML are effective for neural LTR, for both
interpretability and efficiency purposes.
Inspired by the scarcity of feature selection methods for neural ranking
models in previous work, we adapted six existing methods from the
interpretable ML for the neural LTR setting, and also proposed our own G-L2x
approach. We discussed different properties of these methods and their
relevance to the LTR task. Lastly, we performed extensive experiments to
evaluate the methods in terms of their trade-offs between ranking performance
and sparsity, in addition, their efficiency improvements through feature cost
reductions. Our results have shown that several methods from interpretable ML
are highly effective at embedded feature selection for neural LTR. In
particular, the local method TabNet can reach the upper bound with less than
10 features; the global methods, in particular G-L2x, can reduce feature
retrieval costs by more than 70%, while maintaining 96% of performance
relative to a full feature model.
We hope our investigation can bridge the gap between the LTR and interpretable
ML fields. The future work can be developing more interpretable and efficient
ranking systems, and how that interpretability could support both
practitioners and the users of ranking systems.
#### 7.0.1 Acknowledgements
This work is partially supported by German Research Foundation (DFG), under
the Project IREM with grant No. AN 996/1-1, and by the Netherlands
Organisation for Scientific Research (NWO) under grant No. VI.Veni.222.269.
## References
* [1] Abdul, A.M., von der Weth, C., Kankanhalli, M.S., Lim, B.Y.: COGAM: measuring and moderating cognitive load in machine learning model explanations. In: Bernhaupt, R., Mueller, F.F., Verweij, D., Andres, J., McGrenere, J., Cockburn, A., Avellino, I., Goguey, A., Bjøn, P., Zhao, S., Samson, B.P., Kocielnik, R. (eds.) CHI ’20: CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, April 25-30, 2020. pp. 1–14. ACM (2020). https://doi.org/10.1145/3313831.3376615, https://doi.org/10.1145/3313831.3376615
* [2] Alvarez-Melis, D., Jaakkola, T.S.: Towards robust interpretability with self-explaining neural networks. In: Bengio, S., Wallach, H.M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (eds.) Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, December 3-8, 2018, Montréal, Canada. pp. 7786–7795 (2018), https://proceedings.neurips.cc/paper/2018/hash/3e9f0fc9b2f89e043bc6233994dfcf76-Abstract.html
* [3] Arapakis, I., Bai, X., Cambazoglu, B.B.: Impact of response latency on user behavior in web search. In: Geva, S., Trotman, A., Bruza, P., Clarke, C.L.A., Järvelin, K. (eds.) The 37th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’14, Gold Coast , QLD, Australia - July 06 - 11, 2014. pp. 103–112. ACM (2014). https://doi.org/10.1145/2600428.2609627, https://doi.org/10.1145/2600428.2609627
* [4] Arik, S.Ö., Pfister, T.: Tabnet: Attentive interpretable tabular learning. In: Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021. pp. 6679–6687. AAAI Press (2021). https://doi.org/10.1609/AAAI.V35I8.16826, https://doi.org/10.1609/aaai.v35i8.16826
* [5] Bai, X., Cambazoglu, B.B.: Impact of response latency on sponsored search. Inf. Process. Manag. 56(1), 110–129 (2019). https://doi.org/10.1016/J.IPM.2018.10.005, https://doi.org/10.1016/j.ipm.2018.10.005
* [6] Balin, M.F., Abid, A., Zou, J.Y.: Concrete autoencoders: Differentiable feature selection and reconstruction. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA. Proceedings of Machine Learning Research, vol. 97, pp. 444–453. PMLR (2019), http://proceedings.mlr.press/v97/balin19a.html
* [7] Barreda-Ángeles, M., Arapakis, I., Bai, X., Cambazoglu, B.B., Pereda-Baños, A.: Unconscious physiological effects of search latency on users and their click behaviour. In: Baeza-Yates, R., Lalmas, M., Moffat, A., Ribeiro-Neto, B.A. (eds.) Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, Santiago, Chile, August 9-13, 2015. pp. 203–212. ACM (2015). https://doi.org/10.1145/2766462.2767719, https://doi.org/10.1145/2766462.2767719
* [8] Bruch, S.: An alternative cross entropy loss for learning-to-rank. In: Leskovec, J., Grobelnik, M., Najork, M., Tang, J., Zia, L. (eds.) WWW ’21: The Web Conference 2021, Virtual Event / Ljubljana, Slovenia, April 19-23, 2021. pp. 118–126. ACM / IW3C2 (2021). https://doi.org/10.1145/3442381.3449794, https://doi.org/10.1145/3442381.3449794
* [9] Bruch, S., Wang, X., Bendersky, M., Najork, M.: An analysis of the softmax cross entropy loss for learning-to-rank with binary relevance. In: Fang, Y., Zhang, Y., Allan, J., Balog, K., Carterette, B., Guo, J. (eds.) Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, ICTIR 2019, Santa Clara, CA, USA, October 2-5, 2019. pp. 75–78. ACM (2019). https://doi.org/10.1145/3341981.3344221, https://doi.org/10.1145/3341981.3344221
* [10] Burges, C.J.C., Ragno, R., Le, Q.V.: Learning to rank with nonsmooth cost functions. In: Schölkopf, B., Platt, J.C., Hofmann, T. (eds.) Advances in Neural Information Processing Systems 19, Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, British Columbia, Canada, December 4-7, 2006. pp. 193–200. MIT Press (2006), https://proceedings.neurips.cc/paper/2006/hash/af44c4c56f385c43f2529f9b1b018f6a-Abstract.html
* [11] Burges, C.J.C., Shaked, T., Renshaw, E., Lazier, A., Deeds, M., Hamilton, N., Hullender, G.N.: Learning to rank using gradient descent. In: Raedt, L.D., Wrobel, S. (eds.) Machine Learning, Proceedings of the Twenty-Second International Conference (ICML 2005), Bonn, Germany, August 7-11, 2005. ACM International Conference Proceeding Series, vol. 119, pp. 89–96. ACM (2005). https://doi.org/10.1145/1102351.1102363, https://doi.org/10.1145/1102351.1102363
* [12] Burges, C.J.: From ranknet to lambdarank to lambdamart: An overview. Learning 11(23-581), 81 (2010)
* [13] Cao, Z., Qin, T., Liu, T., Tsai, M., Li, H.: Learning to rank: from pairwise approach to listwise approach. In: Ghahramani, Z. (ed.) Machine Learning, Proceedings of the Twenty-Fourth International Conference (ICML 2007), Corvallis, Oregon, USA, June 20-24, 2007. ACM International Conference Proceeding Series, vol. 227, pp. 129–136. ACM (2007). https://doi.org/10.1145/1273496.1273513, https://doi.org/10.1145/1273496.1273513
* [14] Chapelle, O., Chang, Y.: Yahoo! learning to rank challenge overview. In: Chapelle, O., Chang, Y., Liu, T. (eds.) Proceedings of the Yahoo! Learning to Rank Challenge, held at ICML 2010, Haifa, Israel, June 25, 2010. JMLR Proceedings, vol. 14, pp. 1–24. JMLR.org (2011), http://proceedings.mlr.press/v14/chapelle11a.html
* [15] Chapelle, O., Keerthi, S.S.: Efficient algorithms for ranking with svms. Inf. Retr. 13(3), 201–215 (2010). https://doi.org/10.1007/S10791-009-9109-9, https://doi.org/10.1007/s10791-009-9109-9
* [16] Chen, J., Song, L., Wainwright, M.J., Jordan, M.I.: Learning to explain: An information-theoretic perspective on model interpretation. In: Dy, J.G., Krause, A. (eds.) Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018. Proceedings of Machine Learning Research, vol. 80, pp. 882–891. PMLR (2018), http://proceedings.mlr.press/v80/chen18j.html
* [17] Dato, D., Lucchese, C., Nardini, F.M., Orlando, S., Perego, R., Tonellotto, N., Venturini, R.: Fast ranking with additive ensembles of oblivious and non-oblivious regression trees. ACM Trans. Inf. Syst. 35(2), 15:1–15:31 (2016). https://doi.org/10.1145/2987380, https://doi.org/10.1145/2987380
* [18] Du, M., Liu, N., Hu, X.: Techniques for interpretable machine learning. Commun. ACM 63(1), 68–77 (2020). https://doi.org/10.1145/3359786, https://doi.org/10.1145/3359786
* [19] Falcon, W., et al.: Pytorch lightning. GitHub. Note: https://github. com/PyTorchLightning/pytorch-lightning 3, 6 (2019)
* [20] Freund, Y., Iyer, R.D., Schapire, R.E., Singer, Y.: An efficient boosting algorithm for combining preferences. In: Shavlik, J.W. (ed.) Proceedings of the Fifteenth International Conference on Machine Learning (ICML 1998), Madison, Wisconsin, USA, July 24-27, 1998. pp. 170–178. Morgan Kaufmann (1998)
* [21] Gallagher, L., Chen, R., Blanco, R., Culpepper, J.S.: Joint optimization of cascade ranking models. In: Culpepper, J.S., Moffat, A., Bennett, P.N., Lerman, K. (eds.) Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, WSDM 2019, Melbourne, VIC, Australia, February 11-15, 2019. pp. 15–23. ACM (2019). https://doi.org/10.1145/3289600.3290986, https://doi.org/10.1145/3289600.3290986
* [22] Geng, X., Liu, T., Qin, T., Li, H.: Feature selection for ranking. In: Kraaij, W., de Vries, A.P., Clarke, C.L.A., Fuhr, N., Kando, N. (eds.) SIGIR 2007: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, July 23-27, 2007. pp. 407–414. ACM (2007). https://doi.org/10.1145/1277741.1277811, https://doi.org/10.1145/1277741.1277811
* [23] Gigli, A., Lucchese, C., Nardini, F.M., Perego, R.: Fast feature selection for learning to rank. In: Carterette, B., Fang, H., Lalmas, M., Nie, J. (eds.) Proceedings of the 2016 ACM on International Conference on the Theory of Information Retrieval, ICTIR 2016, Newark, DE, USA, September 12- 6, 2016. pp. 167–170. ACM (2016). https://doi.org/10.1145/2970398.2970433, https://doi.org/10.1145/2970398.2970433
* [24] Jang, E., Gu, S., Poole, B.: Categorical reparametrization with gumble-softmax (2017)
* [25] Järvelin, K., Kekäläinen, J.: Cumulated gain-based evaluation of IR techniques. ACM Trans. Inf. Syst. 20(4), 422–446 (2002). https://doi.org/10.1145/582415.582418, http://doi.acm.org/10.1145/582415.582418
* [26] Joachims, T.: Optimizing search engines using clickthrough data. In: Proceedings of the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, July 23-26, 2002, Edmonton, Alberta, Canada. pp. 133–142. ACM (2002). https://doi.org/10.1145/775047.775067, https://doi.org/10.1145/775047.775067
* [27] Joachims, T.: Training linear svms in linear time. In: Eliassi-Rad, T., Ungar, L.H., Craven, M., Gunopulos, D. (eds.) Proceedings of the Twelfth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, August 20-23, 2006. pp. 217–226. ACM (2006). https://doi.org/10.1145/1150402.1150429, https://doi.org/10.1145/1150402.1150429
* [28] Kaur, H., Nori, H., Jenkins, S., Caruana, R., Wallach, H.M., Vaughan, J.W.: Interpreting interpretability: Understanding data scientists’ use of interpretability tools for machine learning. In: Bernhaupt, R., Mueller, F.F., Verweij, D., Andres, J., McGrenere, J., Cockburn, A., Avellino, I., Goguey, A., Bjøn, P., Zhao, S., Samson, B.P., Kocielnik, R. (eds.) CHI ’20: CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, April 25-30, 2020. pp. 1–14. ACM (2020). https://doi.org/10.1145/3313831.3376219, https://doi.org/10.1145/3313831.3376219
* [29] Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., Ye, Q., Liu, T.: Lightgbm: A highly efficient gradient boosting decision tree. In: Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R. (eds.) Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA. pp. 3146–3154 (2017), https://proceedings.neurips.cc/paper/2017/hash/6449f44a102fde848669bdd9eb6b76fa-Abstract.html
* [30] Kingma, D.P., Welling, M.: Auto-encoding variational bayes. In: Bengio, Y., LeCun, Y. (eds.) 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings (2014), http://arxiv.org/abs/1312.6114
* [31] Lai, H., Pan, Y., Liu, C., Lin, L., Wu, J.: Sparse learning-to-rank via an efficient primal-dual algorithm. IEEE Trans. Computers 62(6), 1221–1233 (2013). https://doi.org/10.1109/TC.2012.62, https://doi.org/10.1109/TC.2012.62
* [32] Lai, H., Pan, Y., Liu, C., Lin, L., Wu, J.: Sparse learning-to-rank via an efficient primal-dual algorithm. IEEE Trans. Computers 62(6), 1221–1233 (2013). https://doi.org/10.1109/TC.2012.62, https://doi.org/10.1109/TC.2012.62
* [33] Lai, H., Pan, Y., Tang, Y., Yu, R.: Fsmrank: Feature selection algorithm for learning to rank. IEEE Trans. Neural Networks Learn. Syst. 24(6), 940–952 (2013). https://doi.org/10.1109/TNNLS.2013.2247628, https://doi.org/10.1109/TNNLS.2013.2247628
* [34] Laporte, L., Flamary, R., Canu, S., Déjean, S., Mothe, J.: Nonconvex regularizations for feature selection in ranking with sparse SVM. IEEE Trans. Neural Networks Learn. Syst. 25(6), 1118–1130 (2014). https://doi.org/10.1109/TNNLS.2013.2286696, https://doi.org/10.1109/TNNLS.2013.2286696
* [35] Lemhadri, I., Ruan, F., Abraham, L., Tibshirani, R.: Lassonet: A neural network with feature sparsity. J. Mach. Learn. Res. 22, 127:1–127:29 (2021), http://jmlr.org/papers/v22/20-848.html
* [36] Leonhardt, J., Rudra, K., Anand, A.: Extractive explanations for interpretable text ranking. ACM Trans. Inf. Syst. (dec 2022). https://doi.org/10.1145/3576924, https://doi.org/10.1145/3576924
* [37] Li, P., Burges, C.J.C., Wu, Q.: Mcrank: Learning to rank using multiple classification and gradient boosting. In: Platt, J.C., Koller, D., Singer, Y., Roweis, S.T. (eds.) Advances in Neural Information Processing Systems 20, Proceedings of the Twenty-First Annual Conference on Neural Information Processing Systems, Vancouver, British Columbia, Canada, December 3-6, 2007. pp. 897–904. Curran Associates, Inc. (2007), https://proceedings.neurips.cc/paper/2007/hash/b86e8d03fe992d1b0e19656875ee557c-Abstract.html
* [38] Liu, T.: Learning to rank for information retrieval. Found. Trends Inf. Retr. 3(3), 225–331 (2009). https://doi.org/10.1561/1500000016, https://doi.org/10.1561/1500000016
* [39] Lucchese, C., Nardini, F.M., Orlando, S., Perego, R., Veneri, A.: ILMART: interpretable ranking with constrained lambdamart. In: SIGIR ’22: The 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022. pp. 2255–2259. ACM (2022). https://doi.org/10.1145/3477495.3531840, https://doi.org/10.1145/3477495.3531840
* [40] Lucchese, C., Nardini, F.M., Orlando, S., Perego, R., Veneri, A.: ILMART: interpretable ranking with constrained lambdamart. In: Amigó, E., Castells, P., Gonzalo, J., Carterette, B., Culpepper, J.S., Kazai, G. (eds.) SIGIR ’22: The 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022. pp. 2255–2259. ACM (2022). https://doi.org/10.1145/3477495.3531840, https://doi.org/10.1145/3477495.3531840
* [41] Lundberg, S.M., Lee, S.: A unified approach to interpreting model predictions. In: Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R. (eds.) Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA. pp. 4765–4774 (2017), https://proceedings.neurips.cc/paper/2017/hash/8a20a8621978632d76c43dfd28b67767-Abstract.html
* [42] Martins, A.F.T., Astudillo, R.F.: From softmax to sparsemax: A sparse model of attention and multi-label classification. In: Balcan, M., Weinberger, K.Q. (eds.) Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016. JMLR Workshop and Conference Proceedings, vol. 48, pp. 1614–1623. JMLR.org (2016), http://proceedings.mlr.press/v48/martins16.html
* [43] Masoomi, A., Wu, C., Zhao, T., Wang, Z., Castaldi, P.J., Dy, J.G.: Instance-wise feature grouping. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (eds.) Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual (2020), https://proceedings.neurips.cc/paper/2020/hash/9b10a919ddeb07e103dc05ff523afe38-Abstract.html
* [44] Molnar, C.: Interpretable machine learning. Lulu. com (2020)
* [45] Pan, F., Converse, T., Ahn, D., Salvetti, F., Donato, G.: Feature selection for ranking using boosted trees. In: Cheung, D.W., Song, I., Chu, W.W., Hu, X., Lin, J. (eds.) Proceedings of the 18th ACM Conference on Information and Knowledge Management, CIKM 2009, Hong Kong, China, November 2-6, 2009. pp. 2025–2028. ACM (2009). https://doi.org/10.1145/1645953.1646292, https://doi.org/10.1145/1645953.1646292
* [46] Pang, L., Xu, J., Ai, Q., Lan, Y., Cheng, X., Wen, J.: Setrank: Learning a permutation-invariant ranking model for information retrieval. In: Huang, J.X., Chang, Y., Cheng, X., Kamps, J., Murdock, V., Wen, J., Liu, Y. (eds.) Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, China, July 25-30, 2020. pp. 499–508. ACM (2020). https://doi.org/10.1145/3397271.3401104, https://doi.org/10.1145/3397271.3401104
* [47] Pobrotyn, P., Bartczak, T., Synowiec, M., Bialobrzeski, R., Bojar, J.: Context-aware learning to rank with self-attention. CoRR abs/2005.10084 (2020), https://arxiv.org/abs/2005.10084
* [48] Purpura, A., Buchner, K., Silvello, G., Susto, G.A.: Neural feature selection for learning to rank. In: Hiemstra, D., Moens, M., Mothe, J., Perego, R., Potthast, M., Sebastiani, F. (eds.) Advances in Information Retrieval - 43rd European Conference on IR Research, ECIR 2021, Virtual Event, March 28 - April 1, 2021, Proceedings, Part II. Lecture Notes in Computer Science, vol. 12657, pp. 342–349. Springer (2021). https://doi.org/10.1007/978-3-030-72240-1_34, https://doi.org/10.1007/978-3-030-72240-1_34
* [49] Qin, T., Liu, T.: Introducing LETOR 4.0 datasets. CoRR abs/1306.2597 (2013), http://arxiv.org/abs/1306.2597
* [50] Qin, T., Liu, T., Xu, J., Li, H.: LETOR: A benchmark collection for research on learning to rank for information retrieval. Inf. Retr. 13(4), 346–374 (2010). https://doi.org/10.1007/S10791-009-9123-Y, https://doi.org/10.1007/s10791-009-9123-y
* [51] Qin, Z., Yan, L., Zhuang, H., Tay, Y., Pasumarthi, R.K., Wang, X., Bendersky, M., Najork, M.: Are neural rankers still outperformed by gradient boosted decision trees? In: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net (2021), https://openreview.net/forum?id=Ut1vF_q_vC
* [52] Rahangdale, A., Raut, S.A.: Deep neural network regularization for feature selection in learning-to-rank. IEEE Access 7, 53988–54006 (2019). https://doi.org/10.1109/ACCESS.2019.2902640, https://doi.org/10.1109/ACCESS.2019.2902640
* [53] Ribeiro, M.T., Singh, S., Guestrin, C.: "why should I trust you?": Explaining the predictions of any classifier. In: Krishnapuram, B., Shah, M., Smola, A.J., Aggarwal, C.C., Shen, D., Rastogi, R. (eds.) Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016. pp. 1135–1144. ACM (2016). https://doi.org/10.1145/2939672.2939778, https://doi.org/10.1145/2939672.2939778
* [54] Rigutini, L., Papini, T., Maggini, M., Scarselli, F.: Sortnet: Learning to rank by a neural-based sorting algorithm. CoRR abs/2311.01864 (2023). https://doi.org/10.48550/ARXIV.2311.01864, https://doi.org/10.48550/arXiv.2311.01864
* [55] Rong, Y., Leemann, T., Nguyen, T., Fiedler, L., Seidel, T., Kasneci, G., Kasneci, E.: Towards human-centered explainable AI: user studies for model explanations. CoRR abs/2210.11584 (2022). https://doi.org/10.48550/arXiv.2210.11584, https://doi.org/10.48550/arXiv.2210.11584
* [56] Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1(5), 206–215 (2019). https://doi.org/10.1038/s42256-019-0048-x, https://doi.org/10.1038/s42256-019-0048-x
* [57] Shrikumar, A., Greenside, P., Kundaje, A.: Learning important features through propagating activation differences. In: Precup, D., Teh, Y.W. (eds.) Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017. Proceedings of Machine Learning Research, vol. 70, pp. 3145–3153. PMLR (2017), http://proceedings.mlr.press/v70/shrikumar17a.html
* [58] Simonyan, K., Vedaldi, A., Zisserman, A.: Deep inside convolutional networks: Visualising image classification models and saliency maps. In: Bengio, Y., LeCun, Y. (eds.) 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Workshop Track Proceedings (2014), http://arxiv.org/abs/1312.6034
* [59] Sun, Z., Qin, T., Tao, Q., Wang, J.: Robust sparse rank learning for non-smooth ranking measures. In: Allan, J., Aslam, J.A., Sanderson, M., Zhai, C., Zobel, J. (eds.) Proceedings of the 32nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2009, Boston, MA, USA, July 19-23, 2009. pp. 259–266. ACM (2009). https://doi.org/10.1145/1571941.1571987, https://doi.org/10.1145/1571941.1571987
* [60] Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: Precup, D., Teh, Y.W. (eds.) Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017. Proceedings of Machine Learning Research, vol. 70, pp. 3319–3328. PMLR (2017), http://proceedings.mlr.press/v70/sundararajan17a.html
* [61] Taylor, M.J., Guiver, J., Robertson, S., Minka, T.: Softrank: optimizing non-smooth rank metrics. In: Najork, M., Broder, A.Z., Chakrabarti, S. (eds.) Proceedings of the International Conference on Web Search and Web Data Mining, WSDM 2008, Palo Alto, California, USA, February 11-12, 2008. pp. 77–86. ACM (2008). https://doi.org/10.1145/1341531.1341544, https://doi.org/10.1145/1341531.1341544
* [62] Tibshirani, R.: Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological) 58(1), 267–288 (1996)
* [63] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. In: Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R. (eds.) Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA. pp. 5998–6008 (2017), https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html
* [64] Wang, L., Lin, J., Metzler, D.: A cascade ranking model for efficient ranked retrieval. In: Ma, W., Nie, J., Baeza-Yates, R., Chua, T., Croft, W.B. (eds.) Proceeding of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2011, Beijing, China, July 25-29, 2011. pp. 105–114. ACM (2011). https://doi.org/10.1145/2009916.2009934, https://doi.org/10.1145/2009916.2009934
* [65] Wu, Q., Burges, C.J.C., Svore, K.M., Gao, J.: Adapting boosting for information retrieval measures. Inf. Retr. 13(3), 254–270 (2010). https://doi.org/10.1007/S10791-009-9112-1, https://doi.org/10.1007/s10791-009-9112-1
* [66] Xia, F., Liu, T., Wang, J., Zhang, W., Li, H.: Listwise approach to learning to rank: theory and algorithm. In: Cohen, W.W., McCallum, A., Roweis, S.T. (eds.) Machine Learning, Proceedings of the Twenty-Fifth International Conference (ICML 2008), Helsinki, Finland, June 5-9, 2008. ACM International Conference Proceeding Series, vol. 307, pp. 1192–1199. ACM (2008). https://doi.org/10.1145/1390156.1390306, https://doi.org/10.1145/1390156.1390306
* [67] Xu, J., Li, H.: Adarank: a boosting algorithm for information retrieval. In: Kraaij, W., de Vries, A.P., Clarke, C.L.A., Fuhr, N., Kando, N. (eds.) SIGIR 2007: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, July 23-27, 2007. pp. 391–398. ACM (2007). https://doi.org/10.1145/1277741.1277809, https://doi.org/10.1145/1277741.1277809
* [68] Xu, Z.E., Huang, G., Weinberger, K.Q., Zheng, A.X.: Gradient boosted feature selection. In: Macskassy, S.A., Perlich, C., Leskovec, J., Wang, W., Ghani, R. (eds.) The 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’14, New York, NY, USA - August 24 - 27, 2014. pp. 522–531. ACM (2014). https://doi.org/10.1145/2623330.2623635, https://doi.org/10.1145/2623330.2623635
* [69] Yoon, J., Jordon, J., van der Schaar, M.: INVASE: instance-wise variable selection using neural networks. In: 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net (2019), https://openreview.net/forum?id=BJg_roAcK7
* [70] Yue, Y., Finley, T., Radlinski, F., Joachims, T.: A support vector method for optimizing average precision. In: Kraaij, W., de Vries, A.P., Clarke, C.L.A., Fuhr, N., Kando, N. (eds.) SIGIR 2007: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, July 23-27, 2007. pp. 271–278. ACM (2007). https://doi.org/10.1145/1277741.1277790, https://doi.org/10.1145/1277741.1277790
* [71] Zhang, Z., Rudra, K., Anand, A.: Explain and predict, and then predict again. In: WSDM ’21, The Fourteenth ACM International Conference on Web Search and Data Mining, Virtual Event, Israel, March 8-12, 2021. pp. 418–426. ACM (2021). https://doi.org/10.1145/3437963.3441758, https://doi.org/10.1145/3437963.3441758
* [72] Zhang, Z., Setty, V., Anand, A.: Sparcassist: A model risk assessment assistant based on sparse generated counterfactuals. In: Amigó, E., Castells, P., Gonzalo, J., Carterette, B., Culpepper, J.S., Kazai, G. (eds.) SIGIR ’22: The 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022. pp. 3219–3223. ACM (2022). https://doi.org/10.1145/3477495.3531677, https://doi.org/10.1145/3477495.3531677
* [73] Zhuang, H., Wang, X., Bendersky, M., Grushetsky, A., Wu, Y., Mitrichev, P., Sterling, E., Bell, N., Ravina, W., Qian, H.: Interpretable ranking with generalized additive models. In: WSDM ’21, The Fourteenth ACM International Conference on Web Search and Data Mining, Virtual Event, Israel, March 8-12, 2021. pp. 499–507. ACM (2021). https://doi.org/10.1145/3437963.3441796, https://doi.org/10.1145/3437963.3441796
|
# Existence of Invariant Measures for Stochastic Inviscid Multi-Layer Quasi-
Geostrophic Equations
Federico Butori Scuola Normale Superiore, Piazza dei Cavalieri 7, 56126,
Pisa, Italy federico.butori at sns.it , Francesco Grotto Università di
Pisa, Dipartimento di Matematica, 5 Largo Bruno Pontecorvo, 56127 Pisa, Italia
francesco.grotto at unipi.it , Eliseo Luongo Scuola Normale Superiore,
Piazza dei Cavalieri 7, 56126, Pisa, Italy eliseo.luongo at sns.it and
Leonardo Roveri Università di Pisa, Dipartimento di Matematica, 5 Largo Bruno
Pontecorvo, 56127 Pisa, Italia leonardo.roveri at phd.unipi.it
###### Abstract.
We consider an inviscid 3-layer quasi-geostrophic model with stochastic
forcing in a 2D bounded domain. After establishing well-posedness of such
system under natural regularity assumptions on the initial condition and the
(additive) noise, we prove the existence of an invariant measure supported on
bounded functions by means of the Krylov-Bogoliubov approach developed by
Ferrario and Bessaih (Comm. Math. Phys. 377, 2020).
###### Acknowledgements.
F. G. and L. R. were both supported by the project _Mathematical methods for
climate science_ , funded by he Ministry of University and Research (MUR) as
part of the PON 2014-2020 ”Research and Innovation” resources - Green Action -
DM MUR 1061/2022
###### 1991 Mathematics Subject Classification:
35R60, 37N10, 60H15, 76U60, 86A10
## 1\. Introduction
Multi-layer quasi-geostrophic equations are a standard model in the
theoretical study of atmospheric and oceanic phenomena. The equations are
derived from general Navier-Stokes (shallow water) models [34, Section 5.3],
and they describe the dynamics of stratified fluids characterized by an
approximate balance between the pressure gradient and the Coriolis force.
Our study is focused on an inviscid and stochastically driven 3-layer quasi-
geostrophic system:
(1.1) $\displaystyle\frac{\partial}{\partial
t}q^{i}+(u^{i}\cdot\nabla)q^{i}=f^{i},\quad
u^{i}=\operatorname{\nabla}^{\perp}\psi^{i},\quad i=1,2,3,$
in which dynamics is expressed in terms of potential vorticity $q^{i}$ of the
$i$-th layer and the associated stream function $\psi^{i}$, with $f^{i}$ being
a random forcing term. The interaction between different layers is encoded in
the linear equations linking potential vorticity and stream functions,
(1.2) $\begin{cases}q^{1}=\Delta\psi^{1}+\beta
y+\frac{c^{2}}{H_{1}g_{1}}(\psi^{2}-\psi^{1})+\frac{c^{2}}{gH_{2}}\psi^{1},\\\
q^{2}=\Delta\psi 2+\beta
y+\frac{c^{2}}{H_{2}g_{1}}(\psi^{1}-\psi^{2})+\frac{c^{2}}{H_{2}g_{2}}(\psi^{3}-\psi^{2}),\\\
q^{3}=\Delta\psi^{3}+\beta
y+\frac{c^{2}}{H_{3}g_{2}}(\psi^{2}-\psi^{3})+\frac{c}{H_{3}}\eta,\end{cases}$
the space domain is a bounded region $D\subset\mathbb{R}^{2}$ and the latter
elliptic equations for potential vorticity are solved under Dirichlet boundary
conditions. In (1.2), parameters $H_{i}>0$ describe the basic-state thickness
of single layers, $g_{i}=g\left(\rho_{i+1}-\right.$
$\left.\rho_{i}\right)/\rho_{1}$ the gravitational acceleration due to density
differences between layers, $\eta$ is the height of the bottom topography, and
finally $c$ is a variable Coriolis parameter. There is no variation in the
fluid densities $\rho_{i}$ (Boussinesq approximation). Moreover, viscosity is
absent and the fluid motion includes stochastic forcing modelling the
interaction with small scales, coherently with the setup for the observation
of geostrophic turbulence (cf. [34, Chapter 9]). The parameter $\beta>0$ rules
the $\beta$-plane approximation of Coriolis force. All the forthcoming
arguments are easily generalized to an arbitrary number of interacting layers,
and we restrict to the case of three layers for the sake of a clear
exposition.
From a mathematical standpoint, when the $f^{i}$’s are deterministic or null
the well-posedness theory of the PDE system described above is completely
analogous to that of 2D Euler’s equations, and there exists a unique solution
for potential vorticity of class $L^{\infty}$. The proof follows by the
classical argument of Yudovich [35], and we refer to [11] for details. In
order to preserve well-posedness, stochastic forcing terms must satisfy
suitable regularity assumptions. Moreover, in order to observe stationary
states of the system, friction terms must be included in the dynamics in order
to compensate the input energy (see (3.4) below for a proper formulation of
the stochastic PDE problem we consider). Both regularity of noise and the
presence of friction are physically sensible (cf. [28, p. 212]) and allow to
consider long-time evolution of the system towards equilibrium states.
The aim of the present contribution is to rigorously establish an existence
result for invariant measures, i.e. equilibrium states, supported by the well-
posedness class $L^{\infty}$ of potential vorticity. Invariant measures and
convergence to equilibrium in quasi-geostrophic and related models have been
the object of numerous works in the viscous case: we mention for instance the
recent works [10, 9] and refer to [8] for an overview. Equilibrium states for
geophysical models in low regularity regimes were considered in [19, 20]
(mirroring analogous techniques from the theory of Euler equations [1, 16, 17,
22, 21]), but to the best of our knowledge our result is the first to study
invariant measures in inviscid geophysical model within the well-posedness
regularity regime. Let us stress that in this context we can not rely on
parabolic regularity theory, and instead need to resort to the averaging
effect of stochastic forcing.
We will establish well-posedness of a stochastic 3-layer quasi-geostrophic
system by means of an inviscid limit approach, using non-physical viscous
approximations that prevent vorticity creation at the boundary, thus avoiding
a critical issue of viscous models in bounded domains (cf. [12, 18]).
Concerning invariant measures, our arguments will be based on the infinite-
dimensional Krylov-Bogoliubov approach developed in [3] for 2D Euler equations
on a periodic domain, allowing to obtain existence of invariant measures
supported by function spaces suited for the study of turbulent phenomena. In
generalizing the techniques of [3] to quasi-geostrophic systems, we are also
able to improve their results in terms of Sobolev regularity both for well-
posedness of the stochastic PDE system and for the support of invariant
measures.
The paper is organized as follows. Section 2 recalls basic notions to be used
in the remainder of the paper, and with which we can rigorously present the
main results of the paper and outline their proofs in Section 3. Sections 4
and 4.4 establish well-posedness of the stochastic PDE under consideration and
properties of solutions, proving Theorem 3.3 Finally, Section 5 proves the
main theorem of the paper, Theorem 3.6.
## 2\. Functional Analytic Setup
Assume from now on that $D\subset\mathbb{R}^{2}$ is a simply connected, smooth
bounded domain. We denote $\mathcal{D}=C^{\infty}_{c}(D)$ and by
$\mathcal{D}^{\prime}=C^{\infty}_{c}(D)^{\prime}$ its topological dual, i.e.
the space of distributions. Brackets $\left\langle\cdot,\cdot\right\rangle$
will denote duality couplings in $\mathcal{D}\times\mathcal{D}^{\prime}$, to
be understood as $L^{2}$ products whenever both arguments belong to the
latter.
We will also denote by $\left\langle\cdot,\cdot\right\rangle$ the $L^{2}$
scalar product for time-space functions, i.e. elements of spaces
$L^{p}(0,T;L^{q}(D))$, whose norm will be denoted by
$\left\|\cdot\right\|_{L^{p}_{t}L^{q}_{x}}$.
### 2.1. Functional spaces
We will rely on fractional Sobolev spaces and fractional powers of the
Dirichlet Laplacian, which have various possible definitions, often non-
equivalent. We choose the approach base on the _spectral_ fractional Laplace
operator. We refer to [7],[26, Chapter 1], [33, Chapter 3] for details on the
material covered in this paragraph.
By $W^{s,p}(D)$, $s\in\mathbb{N},\ p\in[1,+\infty]$, we denote Sobolev spaces
obtained as the closure of $C^{s}(\bar{D})$ under the usual $L^{p}$-based
Sobolev norm, and we set
$H^{s}(D)=W^{s,2}(D),\quad W^{0,p}(D)=L^{p}(D).$
Fractional Sobolev spaces $H^{s}(D)$, $0<s<1$, are the completion of
$C^{1}(\bar{D})$) under the Sobolev-Slobodeckij norm
$\displaystyle\left(\lVert
u\rVert^{2}_{L^{2}(D)}+\int_{D}\int_{D}\frac{(u(x)-u(y))^{2}}{|x-y|^{2+2s}}dxdy\right)^{1/2};$
the definition is extended to any $s=m+\sigma$, $m\in\mathbb{N}$,
$\sigma\in(0,1)$, defining the norm
$\displaystyle\lVert u\rVert_{H^{s}(D)}^{2}=\lVert
u\rVert_{H^{m}(D)}^{2}+\sum_{|\alpha|=m}\lVert
D^{\alpha}u\rVert_{H^{s}(D)}^{2}.$
Considering the closure of $\mathcal{D}$ instead of $C^{s}(\bar{D})$, we can
define as above the spaces $W_{0}^{s,p}(D)$ and $H_{0}^{s}(D)$ (again, we set
$W_{0}^{0,p}(D)=L^{p}(D)$). In particular, there exist an explicit
relationship between these different spaces: for $0\leq s\leq 1/2$ it holds
$H^{s}_{0}(D)=H^{s}(D)$, while for $s>1/2$ one has
$H^{s}_{0}(D)\hookrightarrow H^{s}(D)$, the trace operator is well defined and
$H^{s}_{0}(D)$ contains the null-trace elements.
For each $s>0$ the topological dual of $H^{s}_{0}(D)$, denoted by $H^{-s}(D)$,
is a subspace of the space $\mathcal{D}^{\prime}$ of distributions on $D$,
since $H^{s}_{0}(D)$ was obtained as the closure of $\mathcal{D}$ with respect
to a suitable norm. Moreover, embeddings $H^{s}_{0}(D)\hookrightarrow
L^{2}(D)\hookrightarrow H^{-s}(D)$ are continuous and dense.
The following proposition gathers some classical statements on the relations
between the spaces defined until now, intended in a two-dimensional setting.
###### Proposition 2.1.
The following hold:
* •
(Sobolev embeddings) given $1<q<\infty$ and $\alpha>0$,
$H^{\alpha}(D)\hookrightarrow L^{q}(D),\qquad H_{0}^{\alpha}(D)\hookrightarrow
L^{q}(D)$
whenever $\alpha-1\geq-\frac{2}{q}$. These embeddings are also compact if the
inequality is strict;
* •
(Rellich Theorem) given $\alpha>\beta\geq 0$,
$H^{\alpha}(D)\stackrel{{\scriptstyle c}}{{\hookrightarrow}}H^{\beta}(D),\quad
H_{0}^{\alpha}(D)\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}H_{0}^{\beta}(D),\quad
H^{-\beta}(D)\stackrel{{\scriptstyle c}}{{\hookrightarrow}}H^{-\alpha}(D);$
* •
(Morrey’s inequality) given $l\in\mathbb{N}$ and $\lambda\in[0,1)$, for every
$\alpha$ such that $\alpha-1>l+\lambda$
$H^{\alpha}(D)\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}C^{l,\lambda}\left(\bar{D}\right),\qquad
H_{0}^{\alpha}(D)\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}C^{l,\lambda}\left(\bar{D}\right),$
and for every $p$ such that $1-\frac{2}{p}>l+\lambda$
$W^{1,p}(D)\stackrel{{\scriptstyle
c}}{{\hookrightarrow}}C^{l,\lambda}(\bar{D});$
* •
(Interpolation) given $\theta\in[0,1]$ and $(1-\theta)s+\theta
s^{\prime}\notin(\mathbb{N}+1/2)$,
$\displaystyle\left[H^{s}_{0}(D),H^{s^{\prime}}_{0}(D)\right]_{\theta}$
$\displaystyle=H^{(1-\theta)s+\theta s^{\prime}}_{0}(D),$
$\displaystyle\left[H^{-s}(D),H^{-s^{\prime}}(D)\right]_{\theta}$
$\displaystyle=H^{-(1-\theta)s+\theta s^{\prime}}(D).$
We are also going to use a different definition of fractional Sobolev space,
based on Fourier series expansions. There exists an orthonormal basis of
$L^{2}(D)$ of eigenfunctions $\left\\{e_{n}\right\\}_{n\in\mathbb{N}}\subset
H^{1}_{0}(D)$ of the Dirichlet Laplacian $-\Delta$, i.e. $-\Delta
e_{n}=\lambda_{n}e_{n}$ for some $\lambda_{n}\in\mathbb{R}$. For $s\geq 0$,
spectral fractional Laplacian is defined as a densely defined, positive self-
adjoint operator on $L^{2}(D)$ by
$\displaystyle(-\Delta)^{s}u=\sum_{n=1}^{\infty}\lambda_{n}^{s}\hat{u}_{n}e_{n}\in
L^{2}(D),\quad\hat{u}_{n}=\left\langle u,e_{n}\right\rangle,\quad
u\in\mathsf{D}\big{(}(-\Delta)^{s}\big{)},$
$\displaystyle\mathsf{D}\big{(}(-\Delta)^{s}\big{)}=\left\\{u=\sum_{n=1}^{\infty}\hat{u}_{n}e_{n}:\hat{u}_{n}\in\mathbb{R},\sum_{n=1}^{\infty}\lambda_{n}^{2s}|\hat{u}_{n}|^{2}<\infty\right\\}:=\mathcal{H}^{s}.$
It is easy to check that $\mathcal{H}^{s}$, endowed with the norm
$\lVert{u}\rVert_{\mathcal{H}^{s}}^{2}=\left\langle(-\Delta)^{s/2}u,(-\Delta)^{s/2}u\right\rangle,$
is a Hilbert space. Moreover, $\mathcal{H}^{s}=H^{s}(D)$ for $s\in(0,1/2)$,
$\mathcal{H}^{s}=H^{s}_{0}(D)$ for $s\in(1/2,3/2)$, and
$\mathcal{H}^{s}=H^{1}_{0}(D)\cap H^{s}(D)$ for $s\in[3/2,2]$ in the sense
that they coincide as function spaces and their norms are equivalent (cf.
[7]). For completeness, in the case $s=1/2$, $\mathcal{H}^{1/2}$ coincides
with the so-called Lions-Magenes space $H^{1/2}_{00}(D)$ [26], but we are not
going to use it.
### 2.2. Weak$\star$ and bounded weak$\star$ topologies
In the last part of this work we will make use of some notions on weak
topologies that we now recall. Consider the space $L^{\infty}$ with the norm
topology $\tau_{n}$. The weak$\star$ topology $\tau_{\star}$ on $L^{\infty}$
is the coarsest topology among those making the duality pairing
$\left\langle\cdot,\psi\right\rangle:L^{\infty}\rightarrow\mathbb{R}$
continuous, for all $\psi\in L^{1}$.
A sequence of elements $\phi_{n}\in L^{\infty}$ is said to converge to some
$\phi\in L^{\infty}$ in the weak$\star$ sense,
$\phi_{n}\overset{\star}{\rightharpoonup}\phi$, if
$\left\langle\phi_{n},\psi\right\rangle\rightarrow\left\langle\phi,\psi\right\rangle$
for all $\psi\in L^{1}$. Due to the impossibility of metrizing $\tau_{\star}$,
such notion of convergence does not completely identify the topology. However,
we can circumvent this issue by only considering bounded sets as, being
$L^{1}$ separable, finite-radius balls are weakly$\star$ metrizable (cf. [5]
for an extended discussion). Since we will mostly be working with sequences,
it is then convenient to introduce the notion of bounded weak$\star$ topology
$\tau^{b}_{\star}$, that is, the finest topology on $L^{\infty}$ coinciding
with the weak$\star$ topology on all norm-bounded sets. We have the strict
inclusions
$\tau_{\star}\varsubsetneq\tau^{b}_{\star}\varsubsetneq\tau_{n}.$
As a consequence of the weak$\star$ metrizability of balls, a map
$F:L^{\infty}\rightarrow\mathbb{R}$ is sequentially $\tau_{\star}$-continuous
(i.e. $F(\phi_{n})\rightarrow F(\phi)$ whenever
$\phi_{n}\overset{\star}{\rightharpoonup}\phi$) if and only if $F$ is
$\tau^{b}_{\star}$-continuous (see [3]).
Concerning measurability, Borel sets generated by $\tau_{\star}$ and
$\tau_{\star}^{b}$ coincide, but they are a proper subset of those generated
by $\tau_{n}$, as opposed to the case where the space is separable in the
strong topology, in which $\tau_{\star}^{b}$, $\tau_{\star}$ and $\tau_{n}$
coincide, cf. [31]. We will therefore always specify with respect to which
norm measurability is intended.
## 3\. Overview of Main Results and Methodology
### 3.1. Notation for the multi-layer system
The quasi-geostrophic system (1.1), (1.2) can be written as
(3.1) $\left(\frac{\partial}{\partial
t}+u^{i}\cdot\operatorname{\nabla}\right)\bigg{(}\zeta^{i}+\sum_{j=1}^{3}\tilde{l}_{ij}\psi^{j}\bigg{)}=\tilde{f}^{i},\quad
i=1,2,3,$
where $u^{i}$ is the velocity field of the $i$-th layer,
$\zeta^{i}=\operatorname{\nabla}\times
u^{i}=\partial_{2}u^{i}_{1}-\partial_{1}u^{i}_{2}$ is a scalar function and
$\psi^{i}$ is the (scalar) stream function associated to $u^{i}$, i.e.
$\psi^{i}$ satisfies $\operatorname{\nabla}^{\perp}\psi^{i}=u^{i}$, and
$\Delta\psi^{i}=\zeta^{i}$. The vector $\tilde{f}^{i}$ accounts for external
forces, while by $\tilde{L}=(\tilde{l}_{ij})_{i,j=1,2,3}$ we denote the
layers’ interaction matrix: in the case of (1.2) it has the form
$\tilde{L}=\begin{pmatrix}-\lambda_{1}&\lambda_{1}&0\\\
\lambda_{2}&-2\lambda_{2}&\lambda_{2}\\\
0&\lambda_{3}&-\lambda_{3}\end{pmatrix},$
where $\lambda_{i}$’s are combinations of physical constants of the problem.
Setting $u=(u^{1},u^{2},u^{3})$, $\psi=(\psi^{1},\psi^{2},\psi^{3})$,
$f=(f^{1},f^{2},f^{3})$ and denoting by $\tilde{q}$ the potential vorticity
$\tilde{q}=\Delta\psi+\tilde{L}\psi$
(where $\Delta$ stands for $\operatorname{diag}(\Delta,\Delta,\Delta)$ abusing
notation), the system (3.1) can be conveniently rewritten in the form
(3.2) $\frac{\partial}{\partial
t}\tilde{q}+u\cdot\operatorname{\nabla}\tilde{q}=\tilde{f},$
in which the velocity field $u$ can be formally reconstructed from $\tilde{q}$
via $u=\operatorname{\nabla}^{\perp}(\Delta+\tilde{L})^{-1}\tilde{q}$. In the
following, it may be convenient to assume that the matrix $\tilde{L}$ is
symmetric and negative definite. This is not consistent with the physical
reality that the model aims at describing, but we can circumvent the problem
by introducing a scaling factor in the equation. Indeed, if
$D=\operatorname{diag}(h_{1},h_{2},h_{3})$ such that $h_{i}>0$ and
$h_{i}\lambda_{i}=\lambda$ for all $i=1,2,3$, then $D\tilde{L}$ is symmetric
and negative semidefinite (since for example $(1,1,1)^{\top}\in\ker(L)$). By
multiplying (3.2) by $D$, we can thus define
$q=(D\Delta+D\tilde{L})\psi,\quad
u=\operatorname{\nabla}^{\perp}(D\Delta+D\tilde{L})^{-1}\tilde{q}=\tilde{u},\quad
f=D\tilde{f},$
and obtain an equation of the same form,
$\frac{\partial}{\partial t}q+u\cdot\operatorname{\nabla}q=f,$
where now the operator $(D\Delta+D\tilde{L})$ is symmetric and negative
definite. For the sake of notation, in the following we write it as $(A+L)$,
with
$A\coloneqq D\Delta\quad\text{and}\quad L\coloneqq D\tilde{L},$
now assuming that $L$ is symmetric and negative semidefinite.
In order to properly define the PDE problem and to give meaning to the inverse
of $(A+L)$ we need to specify boundary conditions, and we will choose
Dirichlet conditions,
$\psi(t,\cdot)_{|\partial D}=0.$
As for function spaces we will be working with, dealing with a multi-layer-
problem we are interested in vector fields having some Sobolev regularity. Let
us therefore define, for $p\in[1,+\infty]$, $s\in\mathbb{N}$,
$\displaystyle\mathbf{W}^{s,p}=\\{u=(u^{1},u^{2},u^{3})^{t}:\
u^{1},u^{2},u^{3}\in
W^{s,p}\\},\quad\left\|u\right\|_{\mathbf{W}^{s,p}}=\sum_{i=1}^{3}\left\|u^{i}\right\|_{W^{s,p}}.$
and for $s\in\mathbb{R}$
$\displaystyle\mathbf{H}^{s}=\\{u=(u^{1},u^{2},u^{3})^{t}:\
u^{1},u^{2},u^{3}\in
H^{s}\\},\quad\left\|u\right\|_{\mathbf{H}^{s}}^{2}=\sum_{i=1}^{3}\left\|u^{i}\right\|_{H^{s}}^{2}.$
In an analogous way we can define $\mathbf{W}_{0}^{s,p}$ and
$\mathbf{H}_{0}^{s}$ (for $s\geq 0$). When $s=0$, we set
$\mathbf{W}^{s,p}_{0}(D)=\mathbf{W}^{s,p}(D)=\mathbf{L}^{p}(D)$.
### 3.2. Solutions of Quasi-Geostrophic Equations
We will make use of the following classic result for elliptic boundary value
problems, for which we refer to [2].
###### Lemma 3.1.
For every $q\in\mathbf{L}^{\infty}(D)$ the boundary value problem
(3.3) $\begin{cases}q=A\psi+L\psi,&x\in D,\\\ \psi(x)=0,&x\in\partial
D,\end{cases}$
has a unique weak solution $\psi\in\mathbf{W}^{2,p}(D)$. Moreover, for every
$i=1,\ldots,3$,
$\|\operatorname{\nabla}^{2}\psi^{i}\|_{\mathbf{L}^{p}}\leq
Cp\|q^{i}\|_{\mathbf{L}^{\infty}}\qquad\forall p\in[1,\infty).$
We are interested in a stochastic version of (3.2). Consider a filtered
probability space $(\Omega,\mathcal{F},\mathcal{F}_{t},\mathbb{P})$, let
$\\{\rho_{k}\\}_{k\geq 0}$ be an orthonormal basis of $\mathbf{L}^{2}$ made of
eigenvectors of the operator $(A+L)$ and define the Wiener process
$W_{t}(x)\coloneqq\sum_{k=0}^{+\infty}c_{k}\rho_{k}(x)W^{k}_{t},$
where $\\{c_{k}\\}_{k}\subset\mathbb{R}$ and $\\{W^{k}\\}_{k}$ are independent
standard $\mathbb{R}^{3}$-valued Brownian motions. Assume that
$\\{\mathcal{F}_{t}\\}_{t\geq 0}$ is the filtration generated by this family
of Brownian motions. This choice implies that $W_{t}|_{\partial D}\equiv 0$.
Notice also that one can consider $t\in\mathbb{R}$ just by taking
$W^{k}_{t}\coloneqq\tilde{W}^{k}_{-t}$ for $t<0$ where
$\\{\tilde{W}^{k}\\}_{k}$ are independent copies of $\\{W^{k}\\}_{k}$. The
space regularity for $W$ is encoded in its Fourier coefficients. From now on
we will assume that $W_{t}\in\mathbf{H}^{5/2}$ for all $t\in\mathbb{R}$, hence
$\sum_{k}c_{k}^{2}\|\rho_{k}\|^{2}_{\mathbf{H}^{5/2}}<\infty.$
The problem under consideration can now be formulated as
(3.4) $\begin{cases}dq_{t}+\left[(u_{t}\cdot\operatorname{\nabla})q_{t}+\gamma
q_{t}\right]dt=dW_{t},\qquad&\text{on }D,\\\ \operatorname{\nabla}\cdot
u_{t}=0,\,\,u_{t}=\operatorname{\nabla}^{\bot}(A+L)^{-1}q_{t},&\text{on }D,\\\
q_{t}=0,&\text{on }\partial D,\end{cases}$
where $\gamma>0$ is a damping coefficient ruling the friction term that
prevents the accumulation of energy introduced by the random forcing (we refer
again to [28] for a physical motivation). Recall that, given a filtered
probability space $(\Omega,\mathcal{F},\\{\mathcal{F}_{t}\\}_{t\geq
0},\mathbb{P})$, a collection of random variables $X_{t}(\omega),\ t\in[0,T],\
\omega\in\Omega$ taking values in a Banach space $B$ is an adapted stochastic
process if for every $t\in\mathbb{R}$ $X_{t}$ if $\mathcal{F}_{t}$-measurable.
It is said progressively measurable if for every $t\leq T$, the mapping
$[0,t]\times\Omega\ni(s,\omega)\rightarrow X_{s}(\omega)\in B$ is measurable
with respect to $\mathcal{B}([0,t])\otimes\mathcal{F}_{t}$.
###### Definition 3.2.
Given $q_{0}\in\mathbf{L}^{2}$, a stochastic process $q_{t}$ such that
$q\in C_{w}(0,T;\mathbf{L}^{2})\qquad\mathbb{P}\text{-a.s.}$
is said to be a weak solution of (3.4) if it is adapted with respect to
$\\{\mathcal{F}_{t}\\}_{t\leq 0}$ and for all
$\phi\in\mathbf{H}^{1}_{0}\cap\mathbf{H}^{2}$ and for all $t\in[0,T]$, it
holds $\mathbb{P}$-almost surely
(3.5) $\left\langle q_{t},\phi\right\rangle=\left\langle
q_{0},\phi\right\rangle+\int_{0}^{t}\left\langle
q_{s},u_{s}\cdot\operatorname{\nabla}\phi\right\rangle
ds-\gamma\int_{0}^{t}\left\langle q_{s},\phi\right\rangle
ds+\sum_{k=0}^{\infty}c_{k}\left\langle\rho_{k},\phi\right\rangle W^{k}_{t}.$
The regularity requirement on test functions is necessary to make sense of the
nonlinear term. We will be able to loosen it when assuming better
integrability of the initial datum. The following is the first main result of
the present paper:
###### Theorem 3.3 (Well-posedness).
For every $q_{0}\in\mathbf{L}^{\infty}$ there exists a pathwise unique
solution to equation (3.4) in the sense of Definition 3.2. Moreover, $q\in
C_{w}(0,T;\mathbf{L}^{\infty})$ $\mathbb{P}$-almost surely, $q_{t}$ is
progressively measurable as a process with values in
$(\mathbf{L}^{\infty},\tau_{\star})$ and the formulation (3.5) is valid for
all $\phi\in\mathbf{H}^{1}_{0}$. It also holds that the difference $q-W\in
W^{1,2}(0,T;\mathbf{H}^{-1})$ $\mathbb{P}$-almost surely.
In addition, if $q_{0}\in\mathbf{W}^{1,4}$, then $q\in
C_{w}(0,T;\mathbf{W}^{1,4})$ $\mathbb{P}$-almost surely and $q_{t}$ is
progressively measurable as a process in $\mathbf{W}^{1,4}$.
Existence of solutions will be established by means of a vanishing viscosity
argument in Proposition 4.8, while uniqueness follows from the standard
argument of Yudovich [35], cf. Proposition 4.10.
The following extends the notion of weak solution to time-dependent test
functions, and it will prove to be a convenient formulation in the forthcoming
arguments. Definitions 3.2 and 3.4 are in fact equivalent, as one can prove
along the lines of [27, Lemma 3].
###### Definition 3.4.
Given $q_{0}\in\mathbf{L}^{2}$, a stochastic process $q^{\varepsilon}_{t}$
such that
$q\in C_{w}(0,T;\mathbf{L}^{2})\qquad\mathbb{P}\text{-a.s.}$
is said to be a weak solution of (3.4) if it is adapted with respect to
$\\{\mathcal{F}_{t}\\}_{t\leq 0}$ and, for all $\phi\in
C(0,T;\mathbf{H}^{1}_{0}\cap\mathbf{H}^{2})\cap C^{1}(0,T;\mathbf{L}^{2})$ and
for all $t\in[0,T]$, $\mathbb{P}$-almost surely
(3.6) $\left\langle q_{t},\phi_{t}\right\rangle=\left\langle
q_{0},\phi_{0}\right\rangle+\int_{0}^{t}\left\langle
q_{s},\partial_{s}\phi_{s}\right\rangle ds+\int_{0}^{t}\left\langle
q_{s},u_{s}\cdot\operatorname{\nabla}\phi_{s}\right\rangle
ds-\gamma\int_{0}^{t}\left\langle q_{s},\phi_{s}\right\rangle ds\\\
+\sum_{k=0}^{\infty}c_{k}\left\langle\rho_{k},\phi_{t}\right\rangle
W_{t}^{k}-\sum_{k=0}^{\infty}\int_{0}^{t}c_{k}\left\langle\rho_{k},\partial_{s}\phi_{s}\right\rangle
W_{s}^{k}ds$
### 3.3. Properties of Solutions and Existence of Invariant Measures
Consider the solution $q_{t}$ of (3.4) obtained in Theorem 3.3. Let
$B_{b}(\mathbf{L}^{\infty},\tau_{\star})$ be the set of bounded real-valued
functions on $\mathbf{L}^{\infty}$ that are measurable with respect to the
weak$\star$ topology. For $\phi\in B_{b}(\mathbf{L}^{\infty},\tau_{\star})$ we
denote
$P_{t}\phi(\chi)\coloneqq\mathbb{E}\left[\phi(q(t,\chi))\right].$
If $q_{t}$ possesses the Markov property, the latter is the associated Markov
semigroup of operators. This is in fact the case, but the functional setting
in $\mathbf{L}^{\infty}$ requires a careful argument, to which we devote
Section 5.1. In particular, Lemma 5.1 provides sufficient conditions for the
Markov property of stochastic flows taking values in topological vector
spaces, under the assumption that a separable Banach space embedded in the
latter is preserved by the evolution. In order to show that such condition is
satisfied in our particular case, that is for the solution of Theorem 3.3, we
can rely on the stability of solutions in $\mathbf{W}^{1,4}$, to be proved in
Section 4.4.
Once established the Markov property of the stochastic process under
consideration, we can adopt the following standard definition of invariant
measure:
###### Definition 3.5.
An invariant measure for the system (3.4) is a probability measure $\mu$ on
$(\mathbf{L}^{\infty}(D),\mathcal{B}(\tau^{b}_{\star}))$ such that
$\int P_{t}fd\mu=\int fd\mu,\qquad\forall t\geq 0,\forall\phi\in
C_{b}(L^{\infty},\mathcal{B}(\tau^{b}_{\star})).$
Section 5 will be devoted to the proof of the anticipated existence theorem
for invariant measures:
###### Theorem 3.6.
Under the hypothesis of Theorem 3.3, there exist at least one invariant
measure for the problem (3.4).
The argument consists in averaging over larger and larger time intervals the
law of the stochastic process $q_{t}$, in the classical Krylov-Bogoliubov
fashion. The crucial step is thus to prove tightness of the time-averaged law
of $q_{t}$: as in [3] we rely on an idea introduced by Flandoli [15] in order
to obtain the necessary uniform bounds, to be detailed in Proposition 5.3.
###### Remark 3.7.
Our regularity assumption on the Wiener process $W_{t}$ slightly improves the
ones in [4, 3], where the authors assumed regularity in $H^{3+}$ for the
stochastic forcing (at the level of vorticity). In particular, less regularity
on the noise is required both to perform the vanishing viscosity argument,
thanks to the introduction of an additional stochastic term which vanishes in
the limit, and to ensure the stability of the solution in $W^{1,4}$ thanks to
a more careful energy estimate. As a consequence, either Theorem 3.3 and
Theorem 3.6 are improved, as they hold under weaker assumption on the space
regularity of the noise.
## 4\. Well-posedness of Stochastic Quasi-Geostrophic Systems
In order to study the existence of solutions to (3.4) it is convenient to
focus on the quantity $\eta=q-W$, and to study a vanishing viscosity limit for
the dynamics of $\eta$. Consider the Navier-Stokes system
(4.1)
$\begin{cases}dq_{t}^{\varepsilon}+(u_{t}^{\varepsilon}\cdot\operatorname{\nabla}q_{t}^{\varepsilon}+\gamma
q_{t}^{\varepsilon})dt=\varepsilon^{2}\Delta
q_{t}^{\varepsilon}dt+-\varepsilon^{2}\Delta W_{t}dt+dW_{t},\\\
u^{\varepsilon}=\operatorname{\nabla}^{\bot}(A+L)^{-1}q^{\varepsilon},\\\
q^{\varepsilon}_{t}|_{\partial D}=0,\end{cases}$
including both a viscosity term $\varepsilon^{2}\Delta q_{t}dt$ and a further
stochastic term $-\varepsilon^{2}\Delta W_{t}dt$ that will be instrumental in
dealing with the quantity $\eta=q-W$ (see Remark 4.5 below). Notice that by
assumption such additional stochastic forcing takes values in
$C(0,T;\mathbf{H}^{1/2})$.
The condition $u^{\varepsilon}\cdot n=0$ on $\partial D$ is implicit in the
very definition of $u^{\varepsilon}$, as reconstructed from the operator
$\operatorname{\nabla}^{\bot}(A+L)^{-1}$ with zero boundary condition. The
notion of solution for (4.1) is similar to the one already given for the
problem with $\varepsilon=0$, except for the fact that a second order term in
the equation allows to consider more regular solutions.
###### Definition 4.1.
Given $q_{0}\in\mathbf{L}^{2}$, a stochastic process $q^{\varepsilon}_{t}$
such that
$q^{\varepsilon}(\omega)\in C(0,T;\mathbf{L}^{2})\cap
L^{2}(0,T;\mathbf{H}^{1}_{0})\qquad a.s.$
and is progressively measurable with respect to the norm-induced topologies is
said to be a weak solution of (4.1) if, for all
$\phi\in\mathbf{H}^{1}_{0}\cap\mathbf{H}^{2}$ and every $t\in[0,T]$, it holds
$\mathbb{P}$-almost surely
$\left\langle q^{\varepsilon}_{t},\phi\right\rangle=\left\langle
q^{\varepsilon}_{0},\phi\right\rangle+\int_{0}^{t}\left\langle
q^{\varepsilon}_{s},u^{\varepsilon}_{s}\cdot\operatorname{\nabla}\phi\right\rangle
ds-\gamma\int_{0}^{t}\left\langle q^{\varepsilon}_{s},\phi\right\rangle ds\\\
+\varepsilon^{2}\int_{0}^{t}{\left\langle\operatorname{\nabla}W_{s}-\operatorname{\nabla}q^{\varepsilon}_{s},\operatorname{\nabla}\phi_{s}\right\rangle}ds+\sum_{k\geq
0}c_{k}\left\langle\rho_{k},\phi\right\rangle W^{k}_{t},$
where
$u^{\varepsilon}_{t}\coloneqq\operatorname{\nabla}^{\perp}(A+L)^{-1}q^{\varepsilon}_{t}$.
As in the case of Definitions 3.2 and 3.4, we can use an equivalent time-
dependent definition for (4.1):
###### Lemma 4.2.
Let $q^{\varepsilon}$ be solution of the system (4.1). Then for every
$\phi\in C(0,T;\mathbf{L}^{2})\cap L^{2}(0,T;\mathbf{H}_{0}^{1})\cap
H^{1}(0,T;\mathbf{H}^{-1})$
it holds
$\left\langle q^{\varepsilon}_{t},\phi_{t}\right\rangle=\left\langle
q_{0},\phi_{0}\right\rangle+\int_{0}^{t}\left\langle
q^{\varepsilon}_{s},\partial_{s}\phi_{s}\right\rangle
ds+\int_{0}^{t}\left\langle
q^{\varepsilon}_{s},u^{\varepsilon}_{s}\cdot\operatorname{\nabla}\phi_{s}\right\rangle
ds-\gamma\int_{0}^{t}\left\langle q_{s},\phi_{s}\right\rangle ds\\\
-\varepsilon^{2}\int_{0}^{t}\left\langle\operatorname{\nabla}q^{\varepsilon}_{s}-\operatorname{\nabla}W_{s},\operatorname{\nabla}\phi_{s}\right\rangle
ds\\\ +\sum_{k=0}^{\infty}c_{k}\left\langle\rho_{k},\phi_{t}\right\rangle
W_{t}^{k}-\sum_{k=0}^{\infty}\int_{0}^{t}c_{k}\left\langle\rho_{k},\partial_{s}\phi_{s}\right\rangle
W^{k}_{s}ds.$
Since we are considering additive noise, no explicit stochastic integral
appears in the definition of a solution. This allows to adopt a pathwise
approach by focusing on $\eta^{\varepsilon}\coloneqq q^{\varepsilon}-W$, which
solves
(4.2)
$\begin{cases}\partial_{t}\eta^{\varepsilon}+({u}^{\varepsilon}\cdot\operatorname{\nabla})\eta^{\varepsilon}+\gamma\eta^{\varepsilon}=\varepsilon^{2}\Delta\eta^{\varepsilon}-({u}^{\varepsilon}\cdot\operatorname{\nabla})W-\gamma
W,\\\
{{u}^{\varepsilon}=\operatorname{\nabla}^{\bot}(A+L)^{-1}(\eta^{\varepsilon}+W)},\\\
\eta^{\varepsilon}_{t}|_{\partial D}=0.\end{cases}$
The above is equivalent to the equation for $q^{\varepsilon}$, as we
understand it in the analogous weak sense. Since $W_{t}$ is regular enough, it
is easy to prove that if $q^{\varepsilon}$ is a solution of (4.1) then
$\frac{d}{dt}\eta^{\varepsilon}$ belongs to the space
$L^{2}(0,T;\mathbf{H}^{-1})$ almost surely. Its
$L^{2}(0,T;\mathbf{H}^{-1})$-norm however diverges as $\varepsilon\rightarrow
0$, unless we prove higher regularity of $\eta$ (cf. computations in Lemma
4.4).
###### Remark 4.3.
The viscous approximation we consider is non-physical: it is motivated by
mathematical convenience and, as mentioned above, it avoids the confrontation
with the (physical) creation of vorticity at the boundary $\partial D$.
Physical viscous approximations satisfying no-slip boundary conditions do not
allow to prove inviscid limit results, either in the deterministic [25] or the
stochastic framework [27, 6].
### 4.1. Galerkin approximation
The following provides an a priori bound on the solution of (4.1) in $L^{p}$
norm, independent of $p$ and $\varepsilon$.
###### Lemma 4.4.
Let $q_{0}\in\mathbf{L}^{\infty}$ and assume $q^{\varepsilon}_{t}$ to be a
weak solution of (4.1). Then $q^{\varepsilon}\in
L^{\infty}(0,T;\mathbf{L}^{\infty})$ $\mathbb{P}$-almost surely and, for all
$k\in\mathbb{N}$,
$\left\|q^{\varepsilon}\right\|_{L_{t}^{\infty}\mathbf{L}_{x}^{2k}}\leq
C_{1}\left(\gamma,T,\left\|q_{0}\right\|_{\mathbf{L}_{x}^{\infty}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right)$
and
(4.3)
$\left\|q^{\varepsilon}\right\|_{L^{2}_{t}\mathbf{H}^{1}_{x}}\leq\frac{1}{\varepsilon}C_{2}\left(T,\left\|q_{0}\right\|_{\mathbf{L}_{x}^{\infty}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right),$
where constants $C_{1},C_{2}$ are in fact random variables as they depend on
the noise sample $\omega\in\Omega$.
###### Remark 4.5.
We will actually prove something more, namely that the bound still holds with
the $\mathbf{H}^{2+\delta}$ norm of $W_{t}$ for each $\delta>0$. Since we will
assume more regularity on the noise term in later arguments, we state a sub-
optimal result.
Let us also remark that the addition of the term $-\varepsilon^{2}\Delta
W_{t}dt$ in (4.1) is essential for the latter estimates in the case $k\geq 1$.
Without that additional term and the subsequent cancellation in the dynamics
of $\eta^{\varepsilon}$, the estimate would require better regularity of the
noise $W$.
###### Proof.
We take a pathwise approach and assume that $\omega\in\Omega$ belongs to a set
of full probability such that $W(\omega,t)\in
L^{\infty}(0,T;\mathbf{H}^{5/2})$. Let
$\eta^{\varepsilon}_{t}=q^{\varepsilon}_{t}-W_{t}$ solve (4.2). Since
$\eta^{\varepsilon}\in L^{2}(0,T;\mathbf{H}^{1})$, it is possible to use
$\eta^{\varepsilon,i}|\eta^{\varepsilon,i}|^{2k-2}$, $i=1,2,3$ as test
functions. Considering the equation term by term,
$\left\langle\frac{d}{dt}\eta^{\varepsilon,i}_{t},\eta^{\varepsilon,i}|\eta^{\varepsilon,i}|^{2k-2}\right\rangle=\left\langle|\eta^{\varepsilon,i}_{t}|^{k-1}{\frac{d}{dt}\eta^{\varepsilon,i}_{t},|\eta^{\varepsilon,i}_{t}|^{k}}\right\rangle\\\
=\frac{1}{2k}\frac{d}{dt}\left\|\eta^{\varepsilon,i}\right\|_{L^{2k}}^{2k}=\left\|\eta^{\varepsilon,i}\right\|_{L^{2k}}^{2k-1}\frac{d}{dt}\left\|\eta^{\varepsilon,i}\right\|_{L^{2k}};$
also, by the incompressibility of $u^{\varepsilon}$
$\left\langle\eta^{\varepsilon,i}_{t},u^{\varepsilon,i}_{t}\cdot\operatorname{\nabla}[(\eta^{\varepsilon,i}|\eta^{\varepsilon,i}|^{2k-2})]\right\rangle=\frac{2k-1}{2k}\left\langle
u^{\varepsilon,i}_{t},\operatorname{\nabla}[(\eta^{\varepsilon,i}_{t})^{2k}]\right\rangle=0,$
while the dissipative term yields
$\left\langle\operatorname{\nabla}\eta^{\varepsilon,i}_{t},\operatorname{\nabla}[(\eta^{\varepsilon,i}|\eta^{\varepsilon,i}|^{2k-2})]\right\rangle=(2k-1)\left\|(\eta^{\varepsilon,i}_{t})^{k-1}\operatorname{\nabla}\eta^{\varepsilon,i}_{t}\right\|_{L^{2}}^{2}.$
By Hölder inequality with $p=2k$ and $p^{\prime}=\frac{2k}{2k-1}$ on the terms
involving $W$ and $|\eta^{\varepsilon,i}|^{2k-1}$, and combining the
contributions of all layers,
(4.4)
$\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}^{2k-1}\frac{d}{dt}\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}+\varepsilon^{2}(2k-1)\left\|(\eta^{\varepsilon}_{t})^{k-1}\operatorname{\nabla}\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2}}^{2}\\\
\leq-\gamma\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}^{2k}+\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}^{2k-1}\Big{(}\left\|(u^{\varepsilon}_{t}\cdot\operatorname{\nabla})W_{t}\right\|_{\mathbf{L}^{2k}}+\gamma\left\|W_{t}\right\|_{\mathbf{L}^{2k}}\Big{)}.$
Neglecting positive terms on the left-hand side, the latter simplifies to
$\frac{d}{dt}\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}\leq-\gamma\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}+\left\|u_{t}^{\varepsilon}\right\|_{\mathbf{L}^{2k}}\left\|\operatorname{\nabla}W_{t}\right\|_{\mathbf{L}^{\infty}}+\gamma\left\|W_{t}\right\|_{\mathbf{L}^{2k}}.$
In order to close the estimate, we only need to treat carefully the nonlinear
term involving $\left\|u^{\varepsilon}\right\|_{\mathbf{L}^{2k}}$, applying
the regularizing property of the elliptic problem 3.3 stated in Lemma 3.1.
Since
$u^{\varepsilon}=\operatorname{\nabla}^{\bot}(A+L)^{-1}(\eta^{\varepsilon}+W)$,
there are two cases: using Sobolev embeddings one gets, for $k=1,2$,
$\|u^{\varepsilon}\|_{\mathbf{L}^{k}}\leq
c\|u^{\varepsilon}\|_{\mathbf{H}^{1}}\leq
C(\|\eta^{\varepsilon}\|_{\mathbf{L}^{2}}+\|W_{t}\|_{\mathbf{L}^{2}}),$
while for $k>2$
$\|u^{\varepsilon}\|_{\mathbf{L}^{\infty}}\leq
c\|u^{\varepsilon}\|_{\mathbf{W}^{1,4}}\leq
C(\|\eta^{\varepsilon}\|_{\mathbf{L}^{4}}+\|W_{t}\|_{\mathbf{L}^{4}}),$
where the constants are independent of $\varepsilon$ and $k$.
We conclude by applying Grönwall’s Lemma. Consider first the case $k=1$: we
deduce a uniform bound for the $\mathbf{L}^{2}$ norm of $\eta^{\varepsilon}$
of the form
$\sup_{[0,T]}\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2}}\leq
C_{1}\left(\gamma,T,\left\|q_{0}\right\|_{\mathbf{L}_{x}^{\infty}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right).$
We then integrate (4.4) in time, getting rid of unnecessary terms, and obtain
$\displaystyle{\varepsilon}^{2}\int_{0}^{t}\left\|\operatorname{\nabla}\eta^{\varepsilon}\right\|_{\mathbf{L}^{2}}^{2}\leq\frac{1}{2}$
$\displaystyle\left\|\eta_{0}\right\|_{\mathbf{L}^{2}}^{2}+\int_{0}^{t}\left\|\eta^{\varepsilon}_{s}\right\|_{\mathbf{L}^{2}}\Big{(}\left\|(u_{s}^{\varepsilon}\cdot\operatorname{\nabla})W_{s}\right\|_{\mathbf{L}^{2}}+\gamma\left\|W_{s}\right\|_{\mathbf{L}^{2}}\Big{)}ds,$
which implies the second statement. Taking $k=2$, we use the bound we just
proved to obtain a similar one for the $\mathbf{L}^{4}$ norm, again by
Grönwall’s Lemma. Now, using the $\mathbf{L}^{4}$ estimate and knowing that
$u^{\varepsilon}$ only depends on
$\left\|\eta^{{\varepsilon}}\right\|_{L_{t}^{\infty}\mathbf{L}_{x}^{4}}$, we
can estimate all other $2k$-norms uniformly in $k$. In conclusion we get, for
all $k\geq 1$,
$\sup_{[0,T]}\left\|\eta^{\varepsilon}_{t}\right\|_{\mathbf{L}^{2k}}\leq
C_{1}\left(\gamma,T,\left\|q_{0}\right\|_{\mathbf{L}_{x}^{\infty}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right),$
with the constant $C_{1}$ independent of $k$. This implies that
$\eta^{\varepsilon}\in L^{\infty}(0,T;\mathbf{L}^{p})\cap
L^{2}(0,T;\mathbf{H}^{1})$ uniformly for all $p$, hence
$\eta^{\varepsilon}\in L^{\infty}(0,T;\mathbf{L}^{\infty})\cap
L^{2}(0,T;\mathbf{H}^{1}).\qed$
###### Remark 4.6.
The family of bounds in $L^{\infty}(0,T;\mathbf{L}^{2k})$ is stable under the
weak$\star$ convergence of $\eta^{\varepsilon}\rightharpoonup\eta$ in
$L^{\infty}(0,T;\mathbf{L}^{2k})$. Indeed, the weak$\star$ convergence implies
a uniform bound on the $L^{\infty}(0,T;\mathbf{L}^{2})$ norms, which in turn
gives a uniform bound in $L^{\infty}(0,T;\mathbf{L}^{4})$, by what was showed
in the lemma. As we have proved, this gives a uniform bound for all other
norms, meaning that, up to passing to a subsequence which converges
weakly$\star$ in all $L^{\infty}(0,T;\mathbf{L}^{2k})$, the bounds hold for
any limit function $\eta$ due to the lower-semicontinuity of the
$\mathbf{L}^{2k}$-norms.
###### Proposition 4.7.
Let $\varepsilon>0$, $q_{0}\in\mathbf{L}^{2}$ and $W\in
C([0,T];\mathbf{H}^{5/2})$ $\mathbb{P}$-almost surely Then the problem (4.2)
admits a unique solution $\eta^{\varepsilon}\coloneqq q^{\varepsilon}-W$. In
particular,
$\eta^{\varepsilon}\in C(0,T;\mathbf{\mathbf{L}}^{2})\cap
L^{2}(0,T;\mathbf{H}^{1})\cap W^{1,2}(0,T;\mathbf{H}^{-1}).$
###### Proof.
Consider a complete linearly independent system $\\{e_{k}\\}_{k\geq 0}$ in
$\mathbf{H}^{1}(\Omega)$, made of eigenfunctions of $(A+L)$, and
$\mathbf{H}_{N}=\operatorname{Span}(e_{1},\dots,e_{N})$. Let
$\eta^{\varepsilon,N}(x,t)=\sum_{k=1}^{N}c_{k}(t)\operatorname{e}_{k}(x)$,
with coefficients $c_{k}$ defined in such a way that $\eta^{\varepsilon,N}$
solves the equation on $\mathbf{H}_{N}$, meaning that for every $k=1\ldots N$
$\left\langle\partial_{t}\eta^{\varepsilon,N},e_{k}\right\rangle-\left\langle\eta^{\varepsilon,N},u^{\varepsilon,N}\operatorname{\nabla}e_{k}\right\rangle+\gamma\left\langle\eta^{\varepsilon,N},e_{k}\right\rangle\\\
=-\varepsilon^{2}\left\langle\operatorname{\nabla}\eta^{\varepsilon,N},\operatorname{\nabla}e_{k}\right\rangle-\left\langle(u^{\varepsilon,N}\cdot\operatorname{\nabla})W,e_{k}\right\rangle-\gamma\left\langle
W,e_{k}\right\rangle.$
Local well-posedness of the system above follows from classical results about
stochastic differential equations with locally Lipshitz coefficients, cf. for
example [24, 30]. As for global well-posedness and the possibility to pass to
the limit in $N$, we need some estimates on the solutions independent from $N$
and from time $t$. We give now an energy estimate: by testing
$\eta^{\varepsilon,N}$ against itself we obtain, arguing as in Lemma 4.4,
$\frac{1}{2}\frac{d}{dt}\left\|\eta^{\varepsilon,N}_{t}\right\|_{\mathbf{L}^{2}}^{2}+\varepsilon^{2}\left\|\operatorname{\nabla}\eta^{\varepsilon,N}_{t}\right\|_{\mathbf{L}^{2}}^{2}\\\
\leq-\gamma\left\|\eta^{\varepsilon,N}_{t}\right\|_{\mathbf{L}^{2}}^{2}+\Big{(}\left\|(u^{\varepsilon,N}_{t}\cdot\operatorname{\nabla})W_{t}\right\|_{\mathbf{L}^{2}}+\gamma\left\|W_{t}\right\|_{\mathbf{L}^{2}}\Big{)}\left\|\eta^{\varepsilon,N}_{t}\right\|_{\mathbf{L}^{2}},$
from which we obtain
(4.5)
$\sup_{[0,T]}\left\|\eta^{\varepsilon,N}_{t}\right\|_{\mathbf{L}^{2}}\leq
C_{1}\left(\gamma,T,\left\|q^{N}_{0}\right\|_{\mathbf{L}_{x}^{2}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right)\\\
\leq
C_{1}\left(\gamma,T,\left\|q_{0}\right\|_{\mathbf{L}_{x}^{2}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right).$
and consequently
(4.6)
$\left\|q^{\varepsilon,N}\right\|_{L^{2}_{t}\mathbf{H}^{1}_{x}}\leq\frac{1}{\varepsilon}C_{2}\left(T,\left\|q_{0}\right\|_{\mathbf{L}_{x}^{2}},\left\|W(\omega)\right\|_{L_{t}^{\infty}\mathbf{H}_{x}^{5/2}}\right).$
We can also say something about the regularity in time of
$\eta^{\varepsilon,N}$. Indeed, testing against a function
$\phi\in\mathbf{H}^{1}$ yields
$\left\langle\frac{d}{dt}\eta^{\varepsilon,N},\phi\right\rangle+\left\langle(u^{\varepsilon,N}\cdot\operatorname{\nabla})\eta^{\varepsilon,N},\phi\right\rangle+\gamma\left\langle\eta^{\varepsilon,N},\phi\right\rangle=\\\
-\varepsilon^{2}\left\langle\operatorname{\nabla}\eta^{\varepsilon,N},\operatorname{\nabla}\phi\right\rangle-\left\langle(u^{\varepsilon,N}\cdot\operatorname{\nabla})W,\phi\right\rangle-\gamma\left\langle
W,\phi\right\rangle.$
By Hölder inequality, Sobolev embedding $H^{1}\hookrightarrow L^{4}$, Lemma
3.1 we obtain
$\displaystyle\int_{0}^{T}\Big{|}\Big{\langle}\frac{d}{dt}\eta^{\varepsilon,N}_{t},\phi\Big{\rangle}\Big{|}^{2}dt$
$\displaystyle\qquad\leq\int_{0}^{T}\left(\big{\|}u_{t}^{\varepsilon,N}\big{\|}^{2}_{\mathbf{L}^{2}}\|\phi\|^{2}_{\mathbf{H}^{1}}\big{\|}\eta_{t}^{\varepsilon,N}\big{\|}^{2}_{\mathbf{L}^{4}}+\gamma\big{\|}\eta_{t}^{\varepsilon,N}\big{\|}_{\mathbf{L}^{2}}+\|\phi\|^{2}_{\mathbf{H}^{1}}\|\operatorname{\nabla}\eta_{t}^{\varepsilon,N}\|_{\mathbf{L}^{2}}\right)dt$
$\displaystyle\qquad\quad+\int_{0}^{T}\left(\|\phi\|^{2}_{\mathbf{L}^{2}}\|W_{t}\|^{2}_{\mathbf{L}^{2}}+\|u_{t}^{\varepsilon,N}\|^{2}_{\mathbf{L}^{2}}\|\phi\|_{\mathbf{H}^{1}}\|W_{t}\|^{2}_{\mathbf{L}^{4}}\right)dt$
$\displaystyle\qquad\leq
C\left(\gamma+\left\|W\right\|_{L^{\infty}_{t}\mathbf{L}_{x}^{2}}+\left\|\eta^{\varepsilon,N}\right\|_{L^{\infty}_{t}\mathbf{L}_{x}^{2}}\right)^{2}\left(\left\|W\right\|_{L^{\infty}_{t}\mathbf{L}_{x}^{4}}+\left\|\eta^{\varepsilon,N}\right\|_{L^{\infty}_{t}\mathbf{H}_{x}^{1}}\right)^{2}\left\|\phi\right\|_{H_{x}^{1}}^{2}$
$\displaystyle\qquad\quad+\varepsilon^{2}M^{2}\left\|\phi\right\|_{H_{x}^{1}}^{2},$
therefore, by Eq. 4.5 and
$\|\operatorname{\nabla}\eta^{\varepsilon,N}\|_{L^{2}_{t}\mathbf{L}^{2}_{x}}\leq
M/\varepsilon$ (following from (4.6)),
(4.7)
$\int_{0}^{T}\Big{|}\Big{\langle}\frac{d}{dt}\eta^{\varepsilon,N}_{t},\phi\Big{\rangle}\Big{|}^{2}dt\leq\frac{1}{\varepsilon}C(\gamma,\|q_{0}\|_{L}^{2},\omega)\|\phi\|_{H_{x}^{1}}^{2}.$
This shows that $\frac{d}{dt}\eta^{\varepsilon,N}$ defines an operator in
$L^{2}(0,T;\mathbf{H}^{-1})$ whose norm is independent of $N$ (but diverges
for $\varepsilon\rightarrow 0$) and, together with (4.3) and Aubin-Lions-Simon
lemma (cf. [29]), we obtain that the sequence
$\\{\eta^{\varepsilon,N}\\}_{N\geq 0}$ is precompact in
$C(0,T;\mathbf{\mathbf{L}}^{2})$. We can then extract a subsequence converging
in $C(0,T;\mathbf{\mathbf{L}}^{2})$ to some limit point $\eta^{\varepsilon}$
satisfying (4.2) and assume weak convergence in $L^{2}(0,T;\mathbf{H}^{1})$
and weak$\star$ convergence in $L^{\infty}(0,T;\mathbf{L}^{2})$ . Considering
the weak formulation for $\eta^{\varepsilon,N}$,
$\left\langle\eta_{t}^{\varepsilon,N},\phi\right\rangle-\left\langle
q_{0}^{N},\phi\right\rangle+\int_{0}^{t}\big{[}\left\langle(u_{s}^{\varepsilon,N}\cdot{}\nabla)\eta_{s}^{\varepsilon,N},\phi\right\rangle+\gamma\left\langle\eta_{s}^{\varepsilon,N},\phi\right\rangle\big{]}ds\\\
=-\int_{0}^{t}\big{[}\varepsilon^{2}\left\langle\nabla\eta_{s}^{\varepsilon,N},\nabla\phi\right\rangle-\left\langle(u_{s}^{\varepsilon,N}\cdot\nabla)W_{s},\phi\right\rangle-\gamma\left\langle
W_{s},\phi\right\rangle\big{]}ds.$
we are able to pass the equation to the limit $N\rightarrow+\infty$: weak
convergence in $L^{2}(0,T;\mathbf{H}^{1})$ yields convergence for the linear
terms on the right hand side and convergence in $C(0,T,\mathbf{L}^{2})$
assures the convergence of
$\left\langle\eta_{t}^{\varepsilon,N},\phi\right\rangle$, while the non-linear
term converges due to the additional strong convergence of $u^{\varepsilon,N}$
in $C(0,T;\mathbf{H}^{1})$. We deduce that $\eta_{t}$ solves (4.1) and that
$\eta^{\varepsilon}\in C(0,T;\mathbf{L}^{2})\cap L^{2}(0,T;\mathbf{H}^{1}).$
Finally, uniqueness follows by very similar arguments based on energy
estimates that we omit. Let us notice that uniqueness implies the possibility
of repeating existence proof without using subsequences. This gives for free
the progressive measurability of $\eta$ as a process in $\mathbf{L}^{2}$,
since any finite-dimensional pathwise approximation is measurable and
converges (in the appropriate sense) to the limit, which is a continuous
process in $\mathbf{L}^{2}$, without passing to $\omega$-dependent
subsequences. ∎
### 4.2. Vanishing viscosity
Fix $q_{0}\in\mathbf{L}^{\infty}$. The bound of
$\frac{d}{dt}\eta^{\varepsilon}$ in $L^{2}(0,T;\mathbf{H}^{-1})$ can be
improved and made independent of $\varepsilon$: it is sufficient to repeat the
computations in (4.7) estimating directly the $\mathbf{L}^{4}$ norm of
$\eta^{\varepsilon}$, thanks to Lemma 4.4, instead of using the Sobolev
embedding $\mathbf{H}^{1}\hookrightarrow\mathbf{L}^{4}$. We are then ready to
prove an existence result.
###### Proposition 4.8.
Let $\eta^{\varepsilon}$ be a solution of (4.2). For every sequence
$\varepsilon_{n}\rightarrow 0$ there exist a subsequence $\varepsilon_{k_{n}}$
and a function $\eta$ such that $\eta^{\varepsilon_{k_{n}}}\rightarrow\eta$
strongly in $C(0,T;\mathbf{H}^{-\delta})$ and weakly$\star$ in
$L^{\infty}(0,T;\mathbf{L}^{\infty})$. Moreover, $q_{t}=\eta_{t}+W_{t}$ is a
solution of
$\begin{cases}dq_{t}+([U_{t}\cdot\operatorname{\nabla}]q_{t}+\gamma
q_{t})dt=dW_{t},\\\ U=\operatorname{\nabla}^{\bot}(A+L)^{-1}q,\\\
\operatorname{\nabla}\cdot U=0\end{cases}$
in the sense of Definition 3.2 and $q_{t}\in C_{w}(0,T;\mathbf{L}^{\infty}).$
###### Proof.
By Lemma 4.4, the family $\\{\eta^{\varepsilon}\\}_{\varepsilon\geq 0}$ is
bounded in $L^{\infty}(0,T;\mathbf{L}^{2})$, uniformly in $\varepsilon$, and
since the bound in $L^{2}(0,T;\mathbf{H}^{-1})$ is now uniform in
$\varepsilon$ as well, by Aubin-Lions-Simon lemma
$\\{\eta^{\varepsilon}\\}_{\varepsilon\geq 0}$ is compactly embedded in
$C(0,T;\mathbf{H}^{-\delta})$. Therefore, for every sequence
$\varepsilon_{n}\rightarrow 0$ there exists a subsequence
$\varepsilon_{k_{n}}$ such that $\eta^{\varepsilon_{k_{n}}}$ converges
strongly to some $\eta$ in $C(0,T;\mathbf{H}^{-\delta})$, and weakly$\star$ in
$L^{\infty}(0,T;\mathbf{L}^{\infty})$. With these convergences, we are able to
pass to the limit in the weak formulation of (4.2) exactly as in the proof of
Proposition 4.7 and obtain that $\eta$ solves
$\left\langle\eta_{t},\phi\right\rangle+\int_{0}^{t}\big{[}\left\langle(u_{s}\cdot\operatorname{\nabla})\eta_{s},\phi\right\rangle+\gamma\left\langle\eta_{s},\phi\right\rangle\big{]}ds\\\
=\left\langle
q_{0},\phi\right\rangle+\int_{0}^{t}[\left\langle(u_{s}\cdot\operatorname{\nabla})W_{s},\phi\right\rangle+\gamma\left\langle
W_{s},\phi\right\rangle\big{]}ds.$
Now, since $\eta\in C(0,T;\mathbf{H}^{-\delta})\cap
L^{\infty}(0,T;\mathbf{L}^{\infty})$, by [32, Ch. 3, Lemma 1.2] it holds that
$\eta\in C_{w}(0,T;\mathbf{L}^{\infty})$. Clearly, the same also holds for
$q=\eta+W$ thanks to the regularity of $W$. ∎
###### Remark 4.9.
We have not proved measurability of the process $\omega\mapsto q_{t}(\omega)$.
This will follow, as in the case of $q^{\varepsilon}$, by the uniqueness
result we prove in the following section, so we postpone the issue. However,
by standard theory of Banach spaces the measurability of the process $q_{t}$
can be proved also without uniqueness of the solutions, arguing as in [4,
Section 3.1].
### 4.3. Uniqueness
The following establishes the uniqueness statement in Theorem 3.3.
###### Proposition 4.10.
Given $q_{0}\in L^{\infty}$ and a stochastic basis
$(\Omega,\mathcal{F},\\{\mathcal{F}_{t}\\}_{t},\mathbb{P},\\{W_{t}\\}_{t})$,
there is at most one pathwise solution $q$ to problem (3.4) in the space
$L^{\infty}(0,T;\mathbf{L}^{\infty})$.
In fact, we prove a stronger result than probabilistically strong well-
posedness: for $\mathbb{P}$-a.e. $\omega\in\Omega$, existence and uniqueness
hold for the deterministic equation with rough forcing $W(\omega)$. This
notion is often referred to as “path-by-path” uniqueness.
###### Proof.
Assume there exist two solutions $q$ and $q^{\prime}$ and let $u$,
$u^{\prime}$ be the associated velocities, $\psi$, $\psi^{\prime}$ the stream
functions. We denote their difference by
$\bar{\psi}=\psi-\psi^{\prime}.$
By Definition 3.4, for all test functions $\phi\in\mathbf{C}^{1}_{t,x}$ with
compact support it holds
$\displaystyle\left\langle(A+L)\bar{\psi}_{t},\phi(t)\right\rangle={}$
$\displaystyle\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},\partial_{s}\phi(s)\right\rangle
ds-\gamma\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},\phi(s)\right\rangle ds$
$\displaystyle+\int_{0}^{t}\left\langle(A+L)\psi^{\prime}_{s},\operatorname{\nabla}^{\perp}\bar{\psi}_{s}\cdot\operatorname{\nabla}\phi(s)\right\rangle
ds$
$\displaystyle+\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},u_{s}\cdot\operatorname{\nabla}\phi(s)\right\rangle
ds.$
If $q$ is a solution of (3.4) in $L^{\infty}(0,T;\mathbf{L}^{\infty})$, then
$\frac{d}{dt}(q-W)\in L^{2}(0,T;\mathbf{H}^{-1})$, as it is easily derived
like in (4.7). Therefore $q-q^{\prime}$ has the same regularity, and we can
integrate by parts in time to obtain
$\displaystyle\int_{0}^{t}\left\langle\partial_{s}(A+L)\bar{\psi}_{s},\phi(s)\right\rangle
ds={}$
$\displaystyle-\gamma\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},\phi(s)\right\rangle
ds$
$\displaystyle+\int_{0}^{t}\left\langle(A+L)\psi^{\prime}_{s},\operatorname{\nabla}^{\perp}\bar{\psi}_{s}\cdot\operatorname{\nabla}\phi(s)\right\rangle
ds$
$\displaystyle+\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},u_{s}\cdot\operatorname{\nabla}\phi(s)\right\rangle
ds.$
By a density argument this is extended to $\phi\in
L^{\infty}(0,T;\mathbf{H}^{1})$, hence we can choose $\phi=\bar{\psi}$ and
obtain
$\frac{1}{2}\left\|{D^{1/2}}\operatorname{\nabla}\bar{\psi}_{t}\right\|_{\mathbf{L}^{2}}^{2}=\frac{1}{2}\left\langle
L\bar{\psi}_{t},\bar{\psi}_{t}\right\rangle-\int_{0}^{t}\left\langle(A+L)\psi^{\prime}_{s},\operatorname{\nabla}^{\perp}\bar{\psi}_{s}\cdot\operatorname{\nabla}\bar{\psi}_{s}\right\rangle
ds\\\
-\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},u_{s}\cdot\operatorname{\nabla}\bar{\psi}_{s}\right\rangle
ds+\gamma\int_{0}^{t}\left\langle(A+L)\bar{\psi}_{s},\bar{\psi}_{s}\right\rangle
ds\\\ =:I+II+III+IV.$
Since we are assuming $L$ to be negative semi-definite the term $I$ is
negative, and $II$ vanishes because
$\operatorname{\nabla}^{\perp}\bar{\psi}_{s}\cdot\operatorname{\nabla}\bar{\psi}_{s}=0$.
We now need estimates on the remaining terms. As for $III$, the idea is to
integrate by parts as in [35]: the equality
$u=\operatorname{\nabla}^{\perp}\psi$ leads, up to a constant depending on
$A$, to
$\int_{0}^{t}\int_{D}\frac{\partial^{2}\psi_{s}}{\partial x_{1}\partial
x_{2}}\Big{[}\big{(}\frac{\partial\bar{\psi}_{s}}{\partial
x_{1}}\big{)}^{2}-\big{(}\frac{\partial\bar{\psi}_{s}}{\partial
x_{2}}\big{)}^{2}\Big{]}+\Big{(}\frac{\partial^{2}\psi_{s}}{\partial
x_{2}^{2}}-\frac{\partial^{2}\psi_{s}}{\partial
x_{1}^{2}}\Big{)}\frac{\partial\bar{\psi}_{s}}{\partial
x_{1}}\frac{\partial\bar{\psi}_{s}}{\partial x_{2}}dxds;$
taking absolute values we reduce ourselves to estimate the quantity
(4.8)
$\int_{0}^{t}\int_{D}|\operatorname{\nabla}^{2}\psi_{s}||\operatorname{\nabla}\bar{\psi}_{s}|^{2}dxds.$
By assumption $\psi\in L^{\infty}(0,T;B^{\infty})$, where $B^{\infty}$ is the
space of those $\psi$ solving (3.3) for some $q_{0}\in\mathbf{L}^{\infty}$ (it
is a Banach space if endowed with the norm
$\left\|\psi\right\|_{B^{\infty}}\coloneqq\left\|(A+L)\psi\right\|_{\mathbf{L}^{\infty}}$),
so we do not have bounds on the $\mathbf{L}^{\infty}_{t,x}$ norm of the first
factor, but we have a control on any $\mathbf{L}^{p}$ norm, for $p<\infty$: in
fact, we know that
$\operatorname{\nabla}\psi,\operatorname{\nabla}\psi^{\prime}\in
L^{\infty}_{t}(W^{1,p}_{0,x})^{3}\subset L^{\infty}_{t,x}$. Let
$M^{\varepsilon}\coloneqq\|\operatorname{\nabla}\bar{\psi}\|^{\varepsilon}_{L^{\infty}_{t}\mathbf{L}^{\infty}_{x}}$,
$\forall\varepsilon>0$. By Hölder inequality with conjugated exponents
$2/(2-\varepsilon)$ and $2/\varepsilon$, and since by Lemma 3.1
$\left\|\operatorname{\nabla}^{2}\psi\right\|_{\mathbf{L}^{2/\varepsilon}}\leq\frac{C}{\varepsilon}\left\|q\right\|_{\mathbf{L}^{\infty}},$
(4.8) can be bounded by
$\frac{CM^{\varepsilon}}{\varepsilon}\left\|q\right\|_{L^{\infty}_{t}\mathbf{L}^{\infty}_{x}}\int_{0}^{t}\left\|\operatorname{\nabla}\bar{\psi}_{s}\right\|_{\mathbf{L}_{x}^{2}}^{2-\varepsilon}ds.$
We are thus left with $IV$: we use again Hölder inequality with exponents $2$,
$\infty$, and $2$, then we estimate
$\left\|L\bar{\psi}_{s}\right\|_{\mathbf{L}^{2}}\leq\left\|L\right\|_{HS}\left\|\bar{\psi}_{s}\right\|_{\mathbf{L}^{2}}\leq
C\left\|\operatorname{\nabla}\bar{\psi}_{s}\right\|_{\mathbf{L}^{2}}$, where
$\left\|\cdot\right\|_{HS}$ is the Hilbert-Schmidt norm and the last
inequality follows from the Poincaré inequality (since
$\bar{\psi}\in\mathbf{W}^{1,2}_{0}$). Hence, it holds
$\int_{0}^{t}\left\langle
L\bar{\psi}_{s},u_{s}\cdot\operatorname{\nabla}\bar{\psi}_{s}\right\rangle
ds\leq
C\int_{0}^{t}\left\|\operatorname{\nabla}\bar{\psi}_{s}\right\|_{\mathbf{L}^{2}}^{2}ds$
Gathering all estimates together and setting
$z_{t}\coloneqq|\operatorname{\nabla}\bar{\psi}_{t}|_{\mathbf{L}_{x}^{2}}$ we
obtain
$\frac{1}{2}z^{2}_{t}\leq\frac{C^{\varepsilon}_{1}}{\varepsilon}\int_{0}^{t}z^{2-\varepsilon}_{s}ds+C_{2}\int_{0}^{t}z^{2}_{s}ds.$
Differentiation in time then yields
$\frac{d}{dt}z_{t}\leq\frac{C^{\varepsilon}_{1}}{\varepsilon}z^{1-\varepsilon}_{t}+C_{2}z_{t}$
from which, by Grönwall inequality,
$z_{t}\leq(C_{1}^{\varepsilon}t)^{\frac{1}{\varepsilon}}e^{tC_{2}}.$
Recalling that $C_{1}^{\varepsilon}$ is bounded in $\varepsilon$, it is
possible choose some $t_{0}\leq T$ such that $C_{1}^{\varepsilon}t<1$ whenever
$t\leq t_{0}$. Eventually, sending $\varepsilon$ to zero we obtain $z\equiv 0$
in $[0,t_{0}]$. We can then repeat the same argument starting from $t_{0}$ and
see that $z\equiv 0$ on $[0,T]$. This, together with $\bar{\psi}=0$ on
$\partial D$, implies $\bar{\psi}\equiv 0$ and consequently $q\equiv
q^{\prime}.$ ∎
### 4.4. Stability Properties of Solutions
In order to complete the proof of Theorem 3.3 we need some further stability
results, which lay the ground for the proof of the existence of an invariant
measure, Theorem 3.6. We will denote by $q(t_{0},t,\chi)$ the solution to
(3.4) at time $t$, started at time $t_{0}$ with initial datum $\chi$. Whenever
the initial time is $t_{0}=0$, it is omitted.
###### Proposition 4.11.
Let $\chi^{n}\overset{\star}{\rightharpoonup}\chi$ in $\mathbf{L}^{\infty}(D)$
and let $q^{n}_{t}\coloneqq q(t_{0},t,\chi^{n})$ be the solution of (3.4) with
initial datum $\chi^{n}$ at some fixed initial time $t_{0}$. Then
$\mathbb{P}$-almost surely
$q_{t}^{n}\overset{\star}{\rightharpoonup}q_{t}\coloneqq
q(t_{0},t,\chi)\in\mathbf{L}^{\infty}(D)$ for every $t>t_{0}$.
###### Proof.
By the a priori estimates in Lemma 4.4 and by Remark 4.6,
$\eta^{n}_{t}=q^{n}_{t}-W_{t}$ is uniformly bounded in
$L^{\infty}(0,T;\mathbf{L}^{\infty})$ and $\frac{d}{dt}\eta^{n}_{t}$ is
uniformly bounded in $L^{2}(0,T;\mathbf{H}^{-1})$, hence $\eta^{n}$ is
relatively compact in $C(0,T;\mathbf{H}^{-\delta})$. Consequently we can
extract a subsequence (still denoted as $\eta^{n}$) such that
$\displaystyle\eta^{n}\rightarrow\eta\qquad$ $\displaystyle\in
C(0,T;\mathbf{H}^{-\delta})$
$\displaystyle\eta^{n}\overset{\star}{\rightharpoonup}\eta\qquad$
$\displaystyle\in L^{\infty}(0,T;\mathbf{L}^{\infty})$
for some $\eta\in C(0,T;\mathbf{H}^{-\delta})\cap
L^{\infty}(0,T;\mathbf{L}^{\infty})$. The same holds for $q^{n}$, with limit
$q=\eta+W_{t}$. Then, for any $g\in\mathbf{L}^{1}(D)$
$\int_{t_{0}}^{T}\left\langle q_{t}^{n},g\right\rangle
dt\rightarrow\int_{t_{0}}^{T}\left\langle q_{t},g\right\rangle dt$
and, testing against $\chi^{\delta}_{t}g$ for some $\chi^{\delta}_{t}$
localized around $t$ at scale $\delta$, we get, sending $\delta\rightarrow 0$,
(4.9) $\left\langle q_{t}^{n},g\right\rangle\rightarrow\left\langle
q_{t},g\right\rangle$
$\mathbb{P}$-almost surely for almost every $t\in[t_{0},T]$. Now we are ready
to see that $q_{t}$ solves (3.4). In fact, arguing as in the proof of
Proposition 4.8 and using what we just showed for the first term, we are able
to pass to the limit the weak formulation for $\eta^{n}$ for every
$g\in\mathbf{C}_{c}^{1}(D)$,
$\left\langle\eta^{n}_{t},g\right\rangle=\left\langle\chi^{n},g\right\rangle+\int_{t_{0}}^{t}\left(\left\langle\eta^{n}_{s},u^{n}\cdot\operatorname{\nabla}g\right\rangle-\gamma\left\langle\eta^{n}_{s},g\right\rangle+\left\langle
W_{s},(u^{n}_{s})\cdot\operatorname{\nabla}g\right\rangle-\gamma\left\langle
W_{s},\eta^{n}_{s}\right\rangle\right)ds.$
from which it follows that $q_{t}$ is the unique solution of (3.4) with
initial data $\chi$, that $q_{t}\in C_{w}(0,T,\mathbf{L}^{\infty})$ and
consequently that (4.9) holds for all $t\in[0,T]$, but only if
$g\in\mathbf{C}_{c}^{1}(D)$. We can conclude by considering $q_{t}^{n}-q_{t}$
as a linear operator on $\mathbf{L}^{1}$, for every $t>t_{0}$: it converges
pointwise to zero on the dense subspace $\mathbf{C}^{1}_{c}(D)$ and, since
$\\{q_{t}^{n}\\}$ and $q_{t}$ are uniformely bounded in $\mathbf{L}^{\infty}$,
a simple density argument yields $\left\langle
q_{t}^{n}-q_{t},g\right\rangle\rightarrow 0$ for all $t\geq t_{0}$,
$g\in\mathbf{L}^{1}$. ∎
Proposition 4.11 allows to derive a weak Feller property for the semigroup
associated to the solution of our problem. Recall (cf. also Section 5) that,
for every $\phi\in B_{b}(\mathbf{L}^{\infty},\tau_{\star})$, such semigroup is
defined as
(4.10) $P_{t}\phi(\chi)\coloneqq\mathbb{E}\left[\phi(q(t,\chi))\right]$
###### Corollary 4.12.
The solution map $(t,\chi)\mapsto q(t_{0},t,\chi)$ defines a sequentially
weakly$\star$ Feller semigroup in $\mathbf{L}^{\infty}$, i.e. $P_{t}$ is
continuous as an operator
$P_{t}:SC_{b}(\mathbf{L}^{\infty},\tau_{\star})\rightarrow
SC_{b}(\mathbf{L}^{\infty},\tau_{\star})$
on the space of bounded sequentially weakly$\star$ continuous functions ($f\in
SC_{b}(\mathbf{L}^{\infty},\tau_{\star})$ if $f(\xi^{n})\rightarrow f(\xi)$
whenever $\xi\overset{\star}{\rightharpoonup}\xi$ in $\mathbf{L}^{\infty}$).
Since the weak$\star$ topology is in general non-metrizable, continuity and
sequential continuity do not coincide. Considering instead the bounded
weak$\star$ topology, continuity with respect to $\tau^{b}_{\star}$ is
equivalent to sequential continuity with respect to $\tau_{\star}$, as
recalled in Section 2. Hence, $P_{t}$ is continuous as an operator
$P_{t}:C_{b}(\mathbf{L}^{\infty},\tau^{b}_{\star})\rightarrow
C_{b}(\mathbf{L}^{\infty},\tau^{b}_{\star}).$
The forthcoming Proposition concerns regularity of solutions when the initial
datum is in $\mathbf{W}^{1,4}$, a separable proper subspace of
$\mathbf{L}^{\infty}$. Together with the dense embedding of $\mathbf{W}^{1,4}$
into $(\mathbf{L}^{\infty},\tau_{\star})$, this will be intstrumental in
establishing the Markov property of the solution semigroup.
###### Proposition 4.13.
Let $q_{0}\in\mathbf{W}^{1,4}$. Then the solution $q(t_{0},t,q_{0})$ to (3.4)
started at $q_{0}$ belongs to $C_{w}(t_{0},T;\mathbf{W}^{1,4})$
$\mathbb{P}$-almost surely and is adapted as a process in $\mathbf{W}^{1,4}$.
The proof consists in a Grönwall argument and closely resembles that of [3,
Theorem 5]. It is however possible to weaken the regularity hypothesis on
$W_{t}$ required in the cited work: in order to highlight how so, we provide a
sketch of the proof. Let us state first a preliminary result, adapting a
classic bound from elliptic PDE theory.
###### Lemma 4.14.
Given a solution $q$ of (3.4), if $u$ is the associated velocity it holds
$\left\|\operatorname{\nabla}u\right\|_{\mathbf{L}^{\infty}}\leq
C\left\|q\right\|_{\mathbf{L}^{\infty}}\big{(}1+\log_{+}\big{(}\left\|\operatorname{\nabla}q\right\|_{\mathbf{L}^{4}_{x}}\big{)}\big{)}$
for some constant $C>0$.
###### Proof.
The estimate is a consequence of the one contained in the work of A. Ferrari,
[14, Proposition 1]. By definition, each component $u^{i}$ of $u$ solves the
elliptic problem
$h_{i}\operatorname{\nabla}\times u^{i}=q^{i}-\sum_{j}L_{ij}\psi^{j}$
where both the coefficients $h_{j}$ of $D$ and the boundary conditions are as
discussed in Section 3. Set for simplicity $Q^{i}\coloneqq
q^{i}-\sum_{j}L_{ij}\psi^{j}$ By the estimate in [14], we then obtain
$h_{i}\|\operatorname{\nabla}u^{i}\|_{\mathbf{L}^{\infty}}\leq\bar{C}\left(1+\log\left(1+\frac{\left\|Q^{i}\right\|_{\mathbf{W}_{x}^{1,4}}}{\left\|Q^{i}\right\|_{\mathbf{L}^{\infty}}}\right)\right)\Big{\|}Q^{i}\Big{\|}_{\mathbf{L}^{\infty}}.$
In fact, the regularity requirements in the statement of Ferrari are stricter,
due to the fact that he works in a $3$-dimensional setting; since we only work
in $2$ dimensions, reproducing his result requires weaker assumptions due to
lower regularity loss in Sobolev embeddings.
In order to recover the estimate we need, we can work a little more on the
right-hand side. Using that $\log(x+y)\leq\log 2+\log_{+}x+\log_{+}y$, as well
as $-x\log x\leq 1/e$, we get to
$h_{i}\left\|\operatorname{\nabla}u^{i}\right\|_{\mathbf{L}^{\infty}}\leq
C\left(1+\left\|Q^{i}\right\|_{\mathbf{L}^{\infty}}\left(1+\log_{+}\left(\left\|Q^{i}\right\|_{\mathbf{W}_{x}^{1,4}}\right)\right)\right).$
We then take the maximum over $i=1,2,3$, use the triangular inequality and
again the properties of the logarithm again. This gives the estimate
$\|L_{ij}\psi^{j}\|_{\mathbf{W}_{x}^{1,4}}\leq
C\|q^{j}\|_{\mathbf{W}_{x}^{1,4}}$ (same for the $\mathbf{L}^{\infty}$ norm,
by Lemma 3.1). In the end, we get
$\|\operatorname{\nabla}u\|_{\mathbf{L}^{\infty}}\leq\bar{C}\|q\|_{\mathbf{L}^{\infty}}\left(C+\log_{+}\left(\|q\|_{\mathbf{W}^{1,4}_{x}}\right)\right).$
Finally, a few more computations and Poincaré inequality allow to substitute
$\|\operatorname{\nabla}q\|_{\mathbf{L}^{4}}$ to
$\|q\|_{\mathbf{W}^{1,4}_{x}}$, up to introducing new constants. ∎
###### Proof of Proposition 4.13 (sketch)..
By Proposition 4.11, we only need a bound on
$\left\|\operatorname{\nabla}q\right\|_{\mathbf{L}^{4}}$. We begin by taking
the gradient of (3.4), then set $\eta=q-W$: for each layer $j$ and component
$i$ we formally obtain the equation
$\frac{d}{dt}\partial_{i}\eta^{j}+\partial_{i}(u^{j}\cdot\operatorname{\nabla}(\eta^{j}+W^{j}))+\gamma\partial_{i}(\eta^{j}+W^{j})=0\qquad
j=1,\ldots,3.$
The following steps, as usual, should be performed first on the Galerkin’s
approximants of $\eta$ and then passed to the limit. We omit such passage for
brevity. By multiplying the equation by
$\partial_{i}\eta^{j}|\operatorname{\nabla}\eta^{j}|^{2}$, summing over $i$
and integrating over the spacial domain $D$,
$\frac{1}{4}\frac{d}{dt}\left\|\operatorname{\nabla}\eta^{j}\right\|_{\mathbf{L}^{2}}^{4}+\gamma\left\|\operatorname{\nabla}\eta^{j}\right\|_{\mathbf{L}^{2}}^{4}=-\gamma\sum_{i}^{3}\left\langle\partial_{i}W^{j},\partial_{i}\eta^{j}|\operatorname{\nabla}\eta^{j}|^{2}\right\rangle\\\
=-\sum_{i,k}^{3}\left\langle\partial_{i}u^{j}_{k}\partial_{k}\eta^{j},\partial_{i}\eta^{j}|\operatorname{\nabla}\eta^{j}|^{2}\right\rangle-\sum_{i,k}^{3}\left\langle\partial_{i}(u^{j}_{k}\partial_{k}W^{j}),\partial_{i}\eta^{j}|\operatorname{\nabla}\eta^{j}|^{2}\right\rangle\\\
=:I+II+III.$
The terms on the right-hand side are estimated by Hölder and Young
inequalities:
$\displaystyle|I|$
$\displaystyle\leq\frac{\gamma}{2}\left\|\operatorname{\nabla}\eta^{j}\right\|_{L^{4}}^{4}+C_{\gamma}\left\|W^{j}\right\|_{W^{1,4}},$
$\displaystyle|II|$ $\displaystyle\leq
C\left\|\operatorname{\nabla}u^{j}\right\|_{L^{\infty}}\left\|\operatorname{\nabla}\eta^{j}\right\|_{L^{4}}^{4},$
$\displaystyle|III|$
$\displaystyle\leq\sum_{i,k}^{3}\left\langle\partial_{i}u^{j}_{k}\partial_{k}W^{j}+u^{j}_{k}\partial^{2}_{ik}W^{j},\partial_{i}\eta^{j}|\operatorname{\nabla}\eta^{j}|^{2}\right\rangle$
$\displaystyle\leq\left\|\operatorname{\nabla}\eta^{j}\right\|_{L^{4}}^{4}+C\left\|\operatorname{\nabla}u^{j}\right\|_{L^{4}}^{4}\left\|\operatorname{\nabla}W^{j}\right\|_{L^{4}}^{4}+C^{\prime}\left\|u^{j}\right\|_{L^{\infty}}^{4}\left\|W^{j}\right\|_{W^{2,4}}.$
This way it becomes evident that the most regularity we need on the Brownian
motion $W^{j}$ is for it to be in $W^{2,4}$. We can therefore limit ourselves,
by Sobolev embedding, to requiring $W_{t}\in\mathbf{H}^{5/2}$. From the above
computations it emerges that there still are three terms to estimate:
$\left\|\operatorname{\nabla}u^{j}\right\|_{L^{\infty}}$,
$\left\|\operatorname{\nabla}u^{j}\right\|_{L^{4}}$ and
$\left\|u^{j}\right\|_{L^{\infty}}$. The first one is easily done via Lemma
4.14. As for the second one, either Lemma 4.14 or Lemma 3.1 and Lemma 4.4 can
be used. Finally, the third term is simply bounded by the second one.
Thanks to the estimates on $I$, $II$ and $III$ we have that, summing over $j$,
$\frac{1}{4}\frac{d}{dt}\left\|\operatorname{\nabla}\eta(t)\right\|_{\mathbf{L}^{4}}^{4}\leq\left\|\operatorname{\nabla}\eta\right\|_{\mathbf{L}^{4}}^{4}+C\left\|q\right\|_{\mathbf{L}^{\infty}}\left(C+\log_{+}\left(\|\operatorname{\nabla}q\|_{\mathbf{L}^{4}_{x}}\right)\right)\left\|\operatorname{\nabla}\eta\right\|_{\mathbf{L}^{4}}^{4}\\\
+C\left(\gamma,t_{0},T,\left\|q_{0}\right\|_{\mathbf{L}^{\infty}},\left\|W\right\|_{C\left(0,T;\mathbf{H}^{5/2}\right)}\right).$
At this point, the proof goes along exactly as in [3, Theorem 5], coming to
$\sup_{t_{0}\leq t\leq T}\|\operatorname{\nabla}\eta\|_{\mathbf{L}^{4}}\leq
C\left(\gamma,t_{0},T,\left\|q_{0}\right\|_{\mathbf{L}^{\infty}},\left\|W\right\|_{C\left(0,T;\mathbf{H}^{5/2}\right)}\right).$
Weak continuity in $\mathbf{W}^{1,4}$ is obtained as usual observing that it
has been proved $\frac{d}{dt}\eta\in L^{2}(0,T;\mathbf{H}^{-1})$ and
$\operatorname{\nabla}\eta\in L^{\infty}(0,T;\mathbf{L}^{4})$[32].
Measurability is obtained by a standard argument, see for instance [4]. ∎
###### Proof of Theorem 3.3.
We have already proved existence (Proposition 4.8) and uniqueness (Proposition
4.10). Exploiting the additional integrability of $q$ and the density of
$\mathbf{H}^{1}_{0}\cap\mathbf{H}^{2}$ into $\mathbf{H}^{1}_{0}$, we can relax
the regularity requirements for test functions in the definition of weak
solution and simply ask $\phi\in\mathbf{H}^{1}_{0}$, as
$\int_{0}^{t}\left\langle
q_{s},U_{s}\cdot\operatorname{\nabla}\phi\right\rangle
ds\leq\left\|q\right\|_{L^{\infty}_{t}\mathbf{L}^{\infty}_{x}}\left\|U\right\|_{L^{\infty}_{t}\mathbf{L}^{2}_{x}}\left\|\phi\right\|_{\mathbf{H}^{1}}$
now gives a continuous operator on $\mathbf{H}^{1}_{0}$.
As for measurability, it follows by uniqueness and weak continuity in
$\mathbf{L}^{2}$ . Indeed, we know from Proposition 4.7 that the Navier-Stokes
approximants are measurable and, thanks to the well-posedness of the limiting
problem, the whole sequence converges (in the appropriate sense) to the
solution of the inviscid problem. Thus, since there is no dependence on
$\omega\in\Omega$ in the choice of a suitable subsequence, solutions of the
limit problem are also measurable.
Finally, stability in $\mathbf{W}^{1,4}$ is the content of Proposition 4.13 we
just proved. ∎
## 5\. Invariant measure
### 5.1. Markov property
The following technical result establishes in our context a well-known general
principle, i.e. that time-homogeneous stochastic flows with independent
increments are Markov processes. In order to be able to treat non-metric
spaces, following [3] we exploit a further assumption, that is the invariance
of an embedded Banach space under the stochastic flow.
###### Lemma 5.1 (Markov Property for Stochastic Flows).
Let $\big{(}\Omega,\mathcal{F},\mathcal{F}_{t},\mathbb{P}\big{)}$ be a
filtered probability space and let $(H,\mathcal{T})$ be a topological vector
space.
Consider an $H$-valued stochastic flow $\xi_{s,t}^{\chi}(\omega)$, i.e. a
random field indexed by $0\leq s\leq t\leq T$, $\chi\in H$ such that:
* •
$\xi_{s,s}$ is the identity map $\mathbb{P}$-almost surely;
* •
$\mathbb{P}$-almost surely, for all $s\leq r\leq t$,
$\xi_{s,t}^{\chi}=\xi_{r,t}^{\xi_{s,r}^{\chi}}$;
* •
if $\chi_{n}\to\chi$ in $H$ then $\mathbb{P}$-almost surely for all
$s,t\in[0,T]$ it holds $\xi^{\chi_{n}}_{s,t}\to\xi^{\chi}_{s,t}$ in $H$.
Assume that
* •
(adaptedness) $\omega\mapsto\xi_{0,t}^{\chi}(\omega)$ is
$\mathcal{F}_{t}$-measurable for all $t\in[0,T]$;
* •
(independence of increments) for all $s\leq t$, $\xi_{s,t}^{\chi}$ is
independent of $\mathcal{F}_{s}$;
* •
(time homogeneity) for all $\chi\in H$, $\xi_{t,t+h}^{\chi}$ and
$\xi_{s,s+h}^{\chi}$ have the same law for all $s,t\in[0,T]$ and admissible
$h>0$;
* •
(continuous Banach embedding) there exists a separable Banach space $Y\subset
H$ such that the embedding is dense and continuous. Moreover, for all $\chi\in
Y$, $\mathbb{P}$-almost surely $\xi_{s,t}^{\chi}\in Y$ for all $s,t\in[0,T]$.
Then the evolution operator
$P_{t}:SC_{b}(H,\mathcal{T})\to
SC_{b}(H,\mathcal{T}),\quad(P_{t}\phi)(\chi)=\mathbb{E}[\phi(\xi_{0,t}^{\chi})].$
satisfies the Chapman-Kolmogorov equation: for all $\chi\in H$,
$P_{t+s}\phi(\chi)=P_{t}P_{s}\phi(\chi).$
###### Proof.
It is enough to prove that
$\mathbb{E}\left[\phi\left(\xi_{t+s}^{\chi}\right)|\mathcal{F}_{t}\right]=P_{s}\phi\left(\xi_{t}^{\chi}\right),$
as taking the expected value on both sides yields the thesis; in particular,
the above equality is equivalent to
$\mathbb{E}\left[\phi\left(\xi_{t+s}^{\chi}\right)Z\right]=\mathbb{E}\left[P_{s}\phi\left(\xi_{t}^{\chi}\right)Z\right]$
for every bounded $\mathcal{F}_{t}$-measurable random variable $Z$. Since
$\xi_{t+s}^{\chi}=\xi_{t,t+s}^{\xi_{t}^{\chi}}$ almost surely, it is
sufficient to prove
(5.1)
$\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\zeta}\right)Z\right]=\mathbb{E}\left[P_{s}\phi\left(\zeta\right)Z\right]$
for every $H$-valued $\mathcal{F}_{t}$-measurable random variable $\zeta$.
Let us start by choosing $\zeta$ taking values in $Y$. Since $Y$ is a
separable metric space, there exists a sequence of $Y$-valued,
$\mathcal{F}_{t}$-measurable random variables
$\\{\zeta_{n}\\}_{n}\overset{Y}{\rightarrow}\zeta$ of the form
$\zeta_{n}=\sum_{i=1}^{k_{n}}\zeta_{n}^{(i)}\mathbbm{1}_{A_{n}^{(i)}},$
with $\zeta_{n}^{(i)}\in Y$ and
$\\{A_{n}^{(1)},\dots,A_{n}^{(k_{n})}\\}\subset\mathcal{F}_{t}$ a partition of
$\Omega$. Convergence in $Y$ implies convergence in $(H,\mathcal{T})$ and, due
to the continuity of $\xi$ with respect to initial conditions and to the
sequential continuity of $\phi$ and $P_{s}\phi$, it is enough to prove
(5.2)
$\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\zeta_{n}}\right)Z\right]=\mathbb{E}\left[P_{s}\phi\left(\zeta_{n}\right)Z\right]\quad\forall
n.$
Eq. 5.1 then follows by dominated convergence (whenever $\zeta(\omega)\in Y$
$\mathbb{P}$-almost surely).
Given that
$P_{s}\phi\left(\zeta_{n}\right)=\sum_{i=1}^{k_{n}}P_{s}\phi\left(\zeta_{n}^{(i)}\right)\mathbbm{1}_{A_{n}^{(i)}}\quad\mathbb{P}-a.s.$
and that
$\phi\left(\xi_{t,t+s}^{\zeta_{n}}\right)=\sum_{i=1}^{k_{n}}\phi\left(\xi_{t,t+s}^{\zeta_{n}^{(i)}}\right)\mathbbm{1}_{A_{n}^{(i)}}$
(because
$\xi_{t,t+s}^{\zeta_{n}}=\sum_{i=1}^{k_{n}}\xi_{t,t+s}^{\zeta_{n}^{(i)}}\mathbbm{1}_{A_{n}^{(i)}}$),
we can reduce to proving that
$\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\eta}\right)\mathbbm{1}_{A}Z\right]=\mathbb{E}\left[P_{s}\phi\left(\eta\right)\mathbbm{1}_{A}Z\right]$
for all $Y$-valued, $\mathcal{F}_{t}$-measurable random variables $\eta$ and
for all $A\in\mathcal{F}_{t}$. Moreover, we can neglect the term
$\mathbbm{1}_{A}$, as it is itself a bounded $\mathcal{F}_{t}$-measurable
random variable. In sight of this, we have
$\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\eta}\right)Z\right]=\mathbb{E}\left[\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\eta}\right)Z|\mathcal{F}_{t}\right]\right]=\mathbb{E}\left[\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\eta}\right)\right]Z\right]\\\
=\mathbb{E}\left[\phi\left(\xi_{t,t+s}^{\eta}\right)\right]\mathbb{E}\left[Z\right]=\mathbb{E}\left[\phi\left(\xi_{s}^{\eta}\right)x\right]\mathbb{E}\left[Z\right]=P_{s}\phi\left(\eta\right)\mathbb{E}\left[Z\right]=\mathbb{E}\left[P_{s}\phi\left(\eta\right)Z\right],$
which proves (5.2).
We are thus left to verify that (5.1) holds for every $H$-valued,
$\mathcal{F}_{t}$-measurable random variable we exploit the dense embedding of
$Y$ into $H$. Let $\chi(\omega)\in H$ $\mathbb{P}$-almost surely and
$\\{\chi_{n}\\}_{n\in\mathbb{N}}$ be a sequence of $Y$-valued random variables
converging to $\chi$ with respect to $\mathcal{T}$. We have already proved
that for all $n$ it holds
$\mathbb{E}\left[\phi\left(\xi_{t+s}^{\chi_{n}}\right)Z\right]=\mathbb{E}\left[P_{s}\phi\left(\xi_{t}^{\chi_{n}}\right)Z\right]$
whenever $\phi\in SC_{b}(H,\mathcal{T})$ and $Z$ is a bounded,
$\mathcal{F}_{t}$-measurable random variable. By continuity of $\xi$ with
respect to $\chi_{n}$ we have that both
$\phi\left(\xi_{t+s}^{\chi_{n}}\right)\rightarrow\phi\left(\xi_{t+s}^{\chi}\right)$
and $P_{s}\phi\left(\xi_{t}^{\chi_{n}}\right)\rightarrow
P_{s}\phi\left(\xi_{t}^{\chi}\right)$, hence by dominated convergence their
expected values converge as well, yielding (5.1). ∎
###### Corollary 5.2.
Let $P_{t}$ be the solution semigroup for the problem (3.4), as defined in
(4.10). Then $P_{t}$ enjoys the Markov property, i.e. for any $s,t\geq 0$,
$\phi\in SC_{b}(\mathbf{L}^{\infty},\tau_{\star})$
$P_{t+s}\phi=P_{t}P_{s}\phi.$
###### Proof.
It is easy to verify that $P_{t}$ meets the conditions of Lemma 5.1. In
particular, properties of the stochastic flow can be derived from Theorem 3.3
and from the properties of the Wiener process we chose as stochastic forcing.
As for the continuous Banach embedding, we can use the dense embedding of the
Banach space $\mathbf{W}^{1,4}$ into $(\mathbf{L}^{\infty},\tau_{\star})$. ∎
### 5.2. Construction of the invariant measure
We are finally able to show the existence of an invariant measure for our
problem.
###### Proof of Theorem 3.6.
Let $q$ be the pseudo-vorticity solution to (3.4) as in Theorem 3.3, started
at $t_{0}=0$, and let
$m_{t}\coloneqq\operatorname{\mathcal{L}}(q_{t})$
be its law on $\mathcal{B}(\tau^{b}_{\star})$. Consider the averaged measures
$\mu_{n}\coloneqq\frac{1}{n}\int_{0}^{n}m_{t}dt.$
This integration is allowed since we have proved joint measurability of the
process $q_{t}$, with respect to the corresponding topologies, in Theorem 3.3.
The family $\\{\mu_{n}\\}_{n\in\mathbb{N}}$ is tight (see next section for the
proof). On a separable metric space one would thus use Prokhorov’s theorem to
extract a subsequence converging in the sense of measures. However, the space
$(\mathbf{L}^{\infty},\mathcal{B}(\tau_{\star}^{b}))$ being non metrizable, we
have to resort to the variant given in [23, Theorem 3]. What we have to
verify, in order to apply such variant, is the existence of a countable family
of functionals $\\{f_{i}\\}_{i\in\mathbb{N}}$ on $\mathbf{L}^{\infty}$ taking
values in $[-1,1]$, $\tau^{b}_{\star}$-continuous and point-separating in
$\mathbf{L}^{\infty}$. The condition is satisfied: since all elements of the
separable space $\mathbf{L}^{1}$ are $\tau^{b}_{\star}$-continuous (they are
$\tau_{\star}$-continuous), any countable dense subset of $\mathbf{L}^{1}$ has
the required property. Hence, there exists a subsequence $\mu_{n_{k}}$
converging to some probability measure $\mu$ in the following sense:
$\int fd\mu_{n_{k}}\rightarrow\int fd\mu\qquad\forall f\in
C_{b}((\mathbf{L}^{\infty},\tau^{b}_{\star});\mathbb{R}).$
It is then immediate to conclude that $\mu$ is an invariant measure, since for
every $\phi\in C_{b}((\mathbf{L}^{\infty},\tau^{b}_{\star});\mathbb{R})$
$\left\langle
P_{t}\phi,\mu_{n_{k}}\right\rangle=\frac{1}{n_{k}}\int_{n_{k}}^{t+n_{k}}\left\langle
m_{r},\phi\right\rangle
dr+\left\langle\phi,\mu\right\rangle-\frac{1}{n_{k}}\int_{0}^{t}\left\langle
m_{r},\phi\right\rangle dr$
and the right-hand side converges to $\left\langle\phi,\mu\right\rangle$. By
Proposition 4.11 $P_{t}\phi\in
C_{b}((\mathbf{L}^{\infty},\tau^{b}_{\star});\mathbb{R})$, thus it also holds
that $\left\langle P_{t}\phi,\mu_{n_{k}}\right\rangle\rightarrow\left\langle
P_{t}\phi,\mu\right\rangle$ and by uniqueness of the limit
$\left\langle
P_{t}\phi,\mu\right\rangle=\left\langle\phi,\mu\right\rangle.\qed$
### 5.3. Tightness of averaged measures
###### Proposition 5.3.
Let $q(t_{0},t,\chi)$ be solution of (3.4). There exists a real-valued,
$\mathbb{P}$-almost surely finite random variable $r$ such that
$\sup_{t_{0}\leq
0}\left\|q(t_{0},0,q_{t_{0}}=0)\right\|_{\mathbf{L}^{\infty}}\leq
r\qquad\mathbb{P}\text{-a.s}.$
###### Proof.
Consider, for $\lambda>0$, the stochastic convolution
$\zeta_{\lambda}(t)\coloneqq\int_{-\infty}^{t}e^{-\lambda(t-s)}dW_{s},$
where $\zeta_{\lambda}$ has one component per layer. It is known that
$\zeta_{\lambda}$ is the unique solution of the linear equation
$d\zeta_{\lambda}(t)=-\lambda\zeta_{\lambda}(t)+dW(t)$ and has the same
regularity as $W_{t}$. Set $\theta_{\lambda}(t)\coloneqq
q(t_{0},t,0)-\zeta_{\lambda}(t)$. $\theta_{\lambda}$ solves the deterministic
equation
$\frac{d}{dt}\theta_{\lambda}=-\text{BS}(\theta_{\lambda}+\zeta_{\lambda})\cdot\operatorname{\nabla}\theta_{\lambda}-\text{BS}(\theta_{\lambda}+\zeta_{\lambda})\cdot\operatorname{\nabla}\zeta_{\lambda}-\gamma\theta_{\lambda}+(\lambda-\gamma)\zeta_{\lambda}$
where BS formally stands for the operator
$\operatorname{\nabla}^{\perp}(A+L)^{-1}$, in analogy with the Biot-Savart
kernel usually found in the formulation of the standard Euler equation.
Proceeding as in Section 4, i.e. by employing a vanishing viscosity
approximation (cf. also [3]) and with computations analogous to those used to
prove Lemma 4.4, we can test $\theta_{\lambda}$ against
$|\theta_{\lambda}|^{p-2}\theta_{\lambda}$ and obtain
$\frac{d}{dt}\left\|\theta_{\lambda}\right\|_{\mathbf{L}^{p}}\leq
C\left\|\operatorname{\nabla}\zeta_{\lambda}\right\|_{\mathbf{L}^{\infty}}\left\|\zeta_{\lambda}\right\|_{\mathbf{L}^{p}}+|\lambda-\gamma|\left\|\zeta_{\lambda}\right\|_{\mathbf{L}^{p}}+\left(C\left\|\operatorname{\nabla}\zeta_{\lambda}\right\|_{\mathbf{L}^{\infty}}-\gamma\right)\left\|\theta_{\lambda}\right\|_{\mathbf{L}^{p}},$
which yields, by Grönwall inequality on $[t_{0},0]$,
$\left\|\theta_{\lambda}(0)\right\|_{\mathbf{L}^{p}}\leq\left\|\theta_{\lambda}(t_{0})\right\|_{\mathbf{L}^{p}}e^{\int_{t_{0}}^{0}(C\left\|\operatorname{\nabla}\zeta_{\lambda}(s)\right\|_{\mathbf{L}^{\infty}}-\gamma)ds}\\\
+\int_{t_{0}}^{0}\left(C\left\|\operatorname{\nabla}\zeta_{\lambda}(r)\right\|_{\mathbf{L}^{\infty}}+|\lambda-\gamma|\right)\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{L}^{p}}e^{\int_{s}^{0}(C\left\|\operatorname{\nabla}\zeta_{\lambda}(r)\right\|_{\mathbf{L}^{\infty}}-\gamma)dr}ds.$
Taking the limit $p\rightarrow\infty$ we obtain, since $H^{\alpha-1}\subset
L^{\infty}$ for $\alpha>2$ and
$\theta_{\lambda}(t_{0})=\zeta_{\lambda}(t_{0})$,
(5.3)
$\left\|\theta_{\lambda}(0)\right\|_{\mathbf{L}^{\infty}}\leq\left\|\zeta_{\lambda}(t_{0})\right\|_{\mathbf{H}^{\alpha}}e^{\int_{t_{0}}^{0}(C\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{H}^{\alpha}}-\gamma)ds}\\\
+\int_{t_{0}}^{0}C\left(\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{H}^{\alpha}}+|\lambda-\gamma|\right)\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{H}^{\alpha}}e^{\int_{s}^{0}(C\left\|\zeta_{\lambda}(r)\right\|_{\mathbf{H}^{\alpha}}-\gamma)dr}ds,$
where we allowed the constant $C$ to change due to Sobolev embedding.
A uniform bound on the right-hand side of (5.3) can now be obtained working
term-by-term as follows. First, notice that
$\mathbb{E}\left[\left\|\zeta_{\lambda}(t)\right\|_{\mathbf{H}^{\alpha}}^{2}\right]=\frac{1}{2\lambda}\mathbb{E}\left[\left\|W(1)\right\|_{\mathbf{H}^{\alpha}}^{2}\right]\qquad\forall
t\in\mathbb{R}.$
By ergodicity of $\zeta_{\lambda}$ (cf. [13]),
$\lim_{t_{0}\rightarrow-\infty}\frac{1}{-t_{0}}\int_{t_{0}}^{0}\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{H}^{\alpha}}ds=\mathbb{E}\left[\left\|\zeta_{\lambda}(0)\right\|_{\mathbf{H}^{\alpha}}\right]\qquad\mathbb{P}\text{-a.s.}$
and we can choose $\lambda$ in such a way that
$\lim_{t_{0}\rightarrow-\infty}\frac{1}{-t_{0}}\int_{t_{0}}^{0}\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{H}^{\alpha}}ds<\frac{\gamma}{2C}\qquad\mathbb{P}\text{-a.s.},$
$C$ coming from (5.3). This means that there exists a random time
$\tau(\omega)\leq 0$ $\mathbb{P}$-almost surely such that, for every
$t_{0}<\tau(\omega)$,
$\int_{t_{0}}^{0}C\left\|\zeta_{\lambda}(s)\right\|_{\mathbf{H}^{\alpha}}ds\leq\frac{1}{2}\gamma(-t_{0})\leq
0.$
Hence, for $-t_{0}$ and $\lambda$ large enough,
$\exp\left(\int_{t_{0}}^{0}(C\|\zeta_{\lambda}(s)\|_{H^{\alpha}}-\gamma)ds\right)\leq\exp\left(\frac{\gamma}{2}t_{0}\right)$.
On the other hand, by continuity of $\zeta_{\lambda}(t)$ in
$\mathbf{H}^{\alpha}$ there exists a $\mathbb{P}$-almost surely finite random
variable $r_{1}$ such that
$\sup_{\tau(\omega)\leq t_{0}\leq
0}\int_{t_{0}}^{0}\|\zeta_{\lambda}(s,\omega)\|_{\mathbf{H}^{\alpha}}ds\leq
r_{1}(\omega)\qquad\mathbb{P}\text{-a.e. }\omega\in\Omega,$
and this yields a uniform bound for the exponential term, which moreover
vanishes as $t_{0}\rightarrow-\infty$. In a similar way one can also obtain,
for $t<0$, a uniform bound on
$\left\|\zeta_{\lambda}\right\|_{\mathbf{H}^{\alpha}}$ of the form
$\left\|\zeta_{\lambda}(t,\omega)\right\|_{\mathbf{H}^{\alpha}}\leq
r_{2}(1+|t|)$ for some $\mathbb{P}$-almost surely finite random variable
$r_{2}$. We have thus proved that
$\sup_{t_{0}\leq
0}\left\|\theta_{\lambda}\left(0,\theta_{\lambda}(t_{0})=-\zeta_{\lambda}(t_{0})\right)\right\|_{\mathbf{L}^{\infty}}\leq
r_{3}(\omega)$
for some $\mathbb{P}$-almost surely finite random variable $r_{3}$. The thesis
now follows recalling that $q=\theta_{\lambda}+\zeta_{\lambda}$. ∎
Tightness of averaged measures is a direct corollary of the above proposition.
###### Corollary 5.4.
For every $\varepsilon>0$ there exist a real number $R_{\varepsilon}>0$ such
that
$\inf_{t\geq
0}\mathbb{P}\left(\left\|q(0,t,q_{0}=0)\right\|_{\mathbf{L}^{\infty}}<R_{\varepsilon}\right)>1-\varepsilon.$
In particular, the family of measures $\\{\mu_{n}\\}_{n\in\mathbb{N}}$ is
tight in $\left(\mathbf{L}^{\infty},\tau^{b}_{\star}\right).$
###### Proof.
By time homogeneity $q(t_{0},0,q_{t_{0}}=0)\sim q(0,-t_{0},q_{0}=0)$ $\forall
t_{0}<0$. Since for all random variables $r$ $\mathbb{P}$-almost surely finite
and $\varepsilon>0$ there exists $R_{\varepsilon}>0$ such that
$\mathbb{P}(r<R_{\varepsilon})>1-\varepsilon$, by Proposition 5.3 we have the
following uniform estimate:
$\mathbb{P}\left(\left\|q(0,t,q_{0}=0)\right\|_{\mathbf{L}^{\infty}}<R_{\varepsilon}\right)=\mathbb{P}\left(\left\|q(-t,0,q_{-t}=0)\right\|_{\mathbf{L}^{\infty}}<R_{\varepsilon}\right)\\\
=\mathbb{P}(r<R_{\varepsilon})>1-\varepsilon.\qed$
## References
* [1] Sergio Albeverio and Ana Bela Cruzeiro. Global flows with invariant (Gibbs) measures for Euler and Navier-Stokes two-dimensional fluids. Comm. Math. Phys., 129(3):431–444, 1990.
* [2] Luigi Ambrosio, Alessandro Carlotto, and Annalisa Massaccesi. Lectures on elliptic partial differential equations, volume 18 of Appunti. Scuola Normale Superiore di Pisa (Nuova Serie) [Lecture Notes. Scuola Normale Superiore di Pisa (New Series)]. Edizioni della Normale, Pisa, 2018.
* [3] Hakima Bessaih and Benedetta Ferrario. Invariant measures for stochastic damped 2D Euler equations. Comm. Math. Phys., 377(1):531–549, 2020.
* [4] Hakima Bessaih and Franco Flandoli. $2$-D Euler equation perturbed by noise. NoDEA Nonlinear Differential Equations Appl., 6(1):35–54, 1999\.
* [5] Haim Brezis. Function Analysis, Sobolev Spaces and Partial Differential Equations. Springer, 2010.
* [6] Federico Butori and Eliseo Luongo. Large deviations principle for the inviscid limit of fluid dynamic systems in 2d bounded domains. arXiv preprint arXiv:2305.11148, 2023.
* [7] Luis Ángel Caffarelli and Pablo Raúl Stinga. Fractional elliptic equations, Caccioppoli estimates and regularity. Annales de l’Institut Henri Poincaré C. Analyse Non Linéaire, 33(3):767–807, 2016.
* [8] Giulia Carigi. Ergodic properties and response theory for a stochastic two-layer model of geophysical fluid dynamics. PhD thesis, 06 2021.
* [9] Giulia Carigi, Jochen Bröcker, and Tobias Kuna. Exponential ergodicity for a stochastic two-layer quasi-geostrophic model. Stoch. Dyn., 23(2):Paper No. 2350011, 33, 2023.
* [10] Giulia Carigi and Eliseo Luongo. Dissipation properties of transport noise in the two-layer quasi-geostrophic model. J. Math. Fluid Mech., 25(2):Paper No. 28, 27, 2023.
* [11] Qingshan Chen. On the well-posedness of the inviscid multi-layer quasi-geostrophic equations. Discrete Contin. Dyn. Syst., 39(6):3215–3237, 2019.
* [12] Georges-Henri Cottet. Boundary conditions and deterministic vortex methods for the Navier-Stokes equations. In Mathematical aspects of vortex dynamics (Leesburg, VA, 1988), pages 128–143. SIAM, Philadelphia, PA, 1989.
* [13] Giuseppe Da Prato and Jerzy Zabczyk. Stochastic equations in infinite dimensions, volume 152 of Encyclopedia of Mathematics and its Applications. Cambridge University Press, Cambridge, second edition, 2014.
* [14] Andrew Becton Ferrari. On the blow-up of solutions of the $3$-D Euler equations in a bounded domain. Comm. Math. Phys., 155(2):277–294, 1993.
* [15] Franco Flandoli. Dissipativity and invariant measures for stochastic Navier-Stokes equations. NoDEA Nonlinear Differential Equations Appl., 1(4):403–423, 1994\.
* [16] Franco Flandoli, Francesco Grotto, and Dejun Luo. Fokker-Planck equation for dissipative 2D Euler equations with cylindrical noise. Theory Probab. Math. Statist., (102):117–143, 2020.
* [17] Francesco Grotto. Stationary solutions of damped stochastic 2-dimensional Euler’s equation. Electron. J. Probab., 25:Paper No. 69, 24, 2020.
* [18] Francesco Grotto, Eliseo Luongo, and Mario Maurelli. Uniform Approximation of 2D Navier-Stokes Equations with Vorticity Creation by Stochastic Interacting Particle Systems. arXiv e-prints, page arXiv:2212.12753, December 2022.
* [19] Francesco Grotto and Umberto Pappalettera. Equilibrium statistical mechanics of barotropic quasi-geostrophic equations. Infin. Dimens. Anal. Quantum Probab. Relat. Top., 24(1):Paper No. 2150007, 23, 2021.
* [20] Francesco Grotto and Umberto Pappalettera. Gaussian invariant measures and stationary solutions of 2D primitive equations. Discrete Contin. Dyn. Syst. Ser. B, 27(5):2683–2699, 2022.
* [21] Francesco Grotto and Giovanni Peccati. Infinitesimal invariance of completely random measures for 2D Euler equations. Theory Probab. Math. Statist., (107):15–35, 2022.
* [22] Francesco Grotto and Marco Romito. A central limit theorem for Gibbsian invariant measures of 2D Euler equations. Comm. Math. Phys., 376(3):2197–2228, 2020.
* [23] Adam Jakubowski. The almost sure skorokhod representation for subsequences in nonmetric spaces. Theory of Probability & Its Applications, 42(1):167–174, 1998\.
* [24] Ioannis Karatzas and Steven Eugene Shreve. Brownian motion and stochastic calculus, volume 113 of Graduate Texts in Mathematics. Springer-Verlag, New York, second edition, 1991.
* [25] Tosio Kato. Remarks on zero viscosity limit for nonstationary Navier-Stokes flows with boundary. In Seminar on nonlinear partial differential equations (Berkeley, Calif., 1983), volume 2 of Math. Sci. Res. Inst. Publ., pages 85–98. Springer, New York, 1984.
* [26] Jacques-Louis Lions and Enrico Magenes. Non-homogeneous boundary value problems and applications. Vol. I. Springer-Verlag, New York-Heidelberg, 1972. Translated from the French by P. Kenneth.
* [27] Eliseo Luongo. Inviscid limit for stochastic navier-stokes equations under general initial conditions. arXiv preprint arXiv:2111.14189, 2021.
* [28] Joseph Pedlosky. Geophysical fluid dynamics. Springer Science & Business Media, 2013.
* [29] Jacques Simon. Compact sets in the space $l^{p}(0,t;b)$. Annals of pure and applied mathematics, CXLVI(IV):65–96, 1987.
* [30] Anatoliy Volodymyrovych Skorokhod. Studies in the theory of random processes. Addison-Wesley Publishing Co., 1965. Translated from the Russian by Scripta Technica, Inc.
* [31] Michel Talagrand. Comparaison des boréliens pour les topologies fortes et faibles. Indiana Univ. Math. J., 21:1001–1004, 1978.
* [32] Roger Temam. Navier-Stokes equations. North-Holland publishing Company, 1979. Studies In Mathematics and Applications.
* [33] Hans Triebel. Theory of function spaces. Modern Birkhäuser Classics. Birkhäuser/Springer Basel AG, Basel, 2010. Reprint of 1983 edition [MR0730762], Also published in 1983 by Birkhäuser Verlag [MR0781540].
* [34] Geoffrey K. Vallis. Atmospheric and oceanic fluid dynamics. Cambridge University Press, 2017.
* [35] Victor Iosifovich Yudovich. Non-stationary flow of an ideal incompressible liquid. USSR Computational Mathematics and Mathematical Physics, 3(6):1407–1456, 1963.
|
# Algorithmic Improvements for Deep Reinforcement Learning applied to
Interactive Fiction
Vishal Jain,1, 3 William Fedus,1, 2 Hugo Larochelle,1, 2, 5 Doina Precup,1, 3,
4, 5 Marc G. Bellemare1, 2, 3, 5
1Mila, 2Google Brain, 3McGill University, 4DeepMind, 5CIFAR Fellow
<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
Text-based games are a natural challenge domain for deep reinforcement
learning algorithms. Their state and action spaces are combinatorially large,
their reward function is sparse, and they are partially observable: the agent
is informed of the consequences of its actions through textual feedback. In
this paper we emphasize this latter point and consider the design of a deep
reinforcement learning agent that can play from feedback alone. Our design
recognizes and takes advantage of the structural characteristics of text-based
games. We first propose a contextualisation mechanism, based on accumulated
reward, which simplifies the learning problem and mitigates partial
observability. We then study different methods that rely on the notion that
most actions are ineffectual in any given situation, following Zahavy et al.’s
idea of an admissible action. We evaluate these techniques in a series of
text-based games of increasing difficulty based on the TextWorld framework, as
well as the iconic game Zork. Empirically, we find that these techniques
improve the performance of a baseline deep reinforcement learning agent
applied to text-based games.
## 1 Introduction
In a text-based game, also called _interactive fiction_ (IF), an agent
interacts with its environment through a natural language interface. Actions
consist of short textual commands, while observations are paragraphs
describing the outcome of these actions (Figure 1). Recently, interactive
fiction has emerged as an important challenge for AI techniques (? ?), in
great part because the genre combines natural language with sequential
decision-making.
Figure 1: The introductory gameplay from Zork.
From a reinforcement learning perspective, IF domains pose a number of
challenges. First, the state space is typically combinatorial in nature, due
to the presence of objects and characters with which the player can interact.
Since any natural language sentence may be given as a valid command, the
action space is similarly combinatorial. The player observes its environment
through feedback in natural language, making this a partially observable
problem. The reward structure is usually sparse, with non-zero rewards only
received when the agent accomplishes something meaningful, such as retrieving
an important object or unlocking a new part of the domain.
We are particularly interested in bringing deep reinforcement learning
techniques to bear on this problem. In this paper, we consider how to design
an agent architecture that can learn to play text adventure games from
feedback alone. Despite the inherent challenges of the domain, we identify
three structural aspects that make progress possible:
* •
Rewards from subtasks. The optimal behaviour completes a series of subtasks
towards the eventual game end;
* •
Transition structure. Most actions have no effect in a given state;
* •
Memory as state. Remembering key past events is often sufficient to deal with
partial observability.
While these properties have been remarked on in previous work (?; ?), here we
relax some of the assumptions previously made and provide fresh tools to more
tractably solve IF domains. More generally, we believe these tools to be
useful in partially observable domains with similar structure.
Our first contribution takes advantage of the special reward structure of IF
domains. In IF, the accumulated reward within an episode correlates with the
number of completed subtasks and provides a good proxy for an agent’s
progress. Our _score contextualisation_ architecture makes use of this fact by
defining a piecewise value function composed of different deep network heads,
where each piece corresponds to a particular level of cumulative reward. This
separation allows the network to learn separate value functions for different
portions of the complete task; in particular, when the problem is linear
(i.e., there is a fixed ordering in which subtasks must be completed), our
method can be used to learn a separate value function for each subtask.
Our second contribution extends the work of ? (?) on action elimination. We
make exploration and action selection more tractable by determining which
actions are _admissible_ in the current state. Formally, we say that an action
is admissible if it leads to a change in the underlying game state. While the
set of available actions is typically large in IF domains, there are usually
few commands that are actually admissible in any particular context. Since the
state is not directly observable, we first learn an LSTM-based _auxiliary
classifier_ that predicts which actions are admissible given the agent’s
history of recent feedback. We use the predicted probability of an action
being admissible to modulate or _gate_ which actions are available to the
agent at each time step. We propose and compare three simple modulation
methods: masking, drop out, and finally consistent Q-learning (? ?). Compared
to ?’s algorithm, our techniques are simpler in spirit and can be learned from
feedback alone.
We show the effectiveness of our methods on a suite of seven IF problems of
increasing difficulty generated using the TextWorld platform (? ?). We find
that combining the score contextualisation approach to an otherwise standard
recurrent deep RL architecture leads to faster learning than when using a
single value function. Furthermore, our action gating mechanism enables the
learning agent to progress on the harder levels of our suite of problems.
## 2 Problem Setting
We represent an interactive fiction environment as a partially observable
Markov decision process (POMDP) with deterministic observations. This POMDP is
summarized by the tuple
$(\mathcal{S},\mathcal{A},P,r,\mathcal{O},\psi,\gamma)$, where $\mathcal{S}$
is the state space, $\mathcal{A}$ the action space, $P$ is the transition
function, $r:\mathcal{S}\times\mathcal{A}\to\mathbb{R}$ is the reward
function, and $\gamma\in[0,1)$ is the discount factor. The function
$\psi:\mathcal{S}\times\mathcal{A}\times\mathcal{S}\to\mathcal{O}$ describes
the observation $o=\psi(s,a,s^{\prime})$ provided to the agent when action $a$
is taken in state $s$ and leads to state $s^{\prime}$.
Throughout we will make use of standard notions from reinforcement learning (?
?) as adapted to the POMDP literature (?; ?). At time step $t$, the agent
selects an action according to a policy $\pi$ which maps a _history_
$h_{t}:=o_{1},a_{1},\dots,o_{t}$ to a distribution over actions, denoted
$\pi(\cdot\,|\,h_{t})$. This history is a sequence of observations and actions
which, from the agent’s perspective, replaces the unobserved environment state
$s_{t}$. We denote by $B(s\,|\,h_{t})$ the probability or _belief_ of being in
state $s$ after observing $h_{t}$. Finally, we will find it convenient to rely
on time indices to indicate the relationship between a history $h_{t}$ and its
successor, and denote by $h_{t+1}$ the history resulting from taking action
$a_{t}$ in $h_{t}$ and observing $o_{t+1}$ as emitted by the hidden state
$s_{t+1}$.
The action-value function $Q^{\pi}$ describes the expected discounted sum of
rewards when choosing action $a$ after observing history $h_{t}$, and
subsequently following policy $\pi$:
$Q^{\pi}(h_{t},a)=\operatorname*{{\mathbb{E}}}\big{[}\sum_{i\geq
0}\gamma^{i}r(s_{t+i},a_{t+i})\,|\,h_{t},a\big{]},$
where we assume that the action at time $t+j$ is drawn from
$\pi(\cdot\,|\,h_{t+j})$; note that the reward depends on the sequence of
hidden states $s_{t+1},s_{t+2},\dots$ implied by the belief state
$B(\cdot\,|\,h_{t})$. The action-value function satisfies the Bellman equation
over histories
$Q^{\pi}(h_{t},a)=\operatorname*{{\mathbb{E}}}_{s_{t},s_{t+1}}\big{[}r(s_{t},a)+\gamma\max_{a^{\prime}\in\mathcal{A}}Q^{\pi}(h_{t+1},a^{\prime})\big{]}.$
When the state is observed at each step ($\mathcal{O}=\mathcal{S}$), this
simplifies to the usual Bellman equation for Markov decision processes:
$Q^{\pi}(s_{t},a)=r(s_{t},a)+\gamma\operatorname*{{\mathbb{E}}}_{s_{t+1}\sim
P}\max_{a^{\prime}\in\mathcal{A}}Q^{\pi}(s_{t+1},a^{\prime}).$ (1)
In the fully observable case we will conflate $s_{t}$ and $h_{t}$.
The Q-learning algorithm (? ?) over histories maintains an approximate action-
value function $Q$ which is updated from samples $h_{t},a_{t},r_{t},o_{t+1}$
using a step-size parameter $\alpha\in[0,1)$:
$\displaystyle Q(h_{t},a_{t})$ $\displaystyle\leftarrow
Q(h_{t},a_{t})+\alpha\delta_{t}$ $\displaystyle\delta_{t}$
$\displaystyle=r_{t}+\gamma\max_{a\in\mathcal{A}}Q(h_{t+1},a)-Q(h_{t},a_{t}).$
(2)
Q-learning is used to estimate the _optimal action-value function_ attained by
a policy which maximizes $Q^{\pi}$ for all histories. In the context of our
work, we will assume that this policy exists. Storing this action-value
function in a lookup table is impractical, as there are in general an
exponential number of histories to consider. Instead, we use recurrent neural
networks approximate the Q-learning process.
### 2.1 Consistent Q-Learning
Consistent Q-learning (? ?) learns a value function which is consistent with
respect to a local form of policy stationarity. Defined for a Markov decision
process, it replaces the term $\delta_{t}$ in (2) by
$\delta^{\textsc{cql}}_{t}=r_{t}+\left\\{\begin{array}[]{cc}\gamma\max_{a\in\mathcal{A}}Q(s_{t+1},a)-Q(s_{t},a_{t})&s_{t+1}\neq
s_{t}\\\ (\gamma-1)Q(s_{t},a_{t})&s_{t+1}=s_{t}.\end{array}\right.$ (3)
Consistent Q-learning can be shown to decrease the action-value of suboptimal
actions while maintaining the action-value of the optimal action, leading to
larger _action gaps_ and a potentially easier value estimation problem.
Observe that consistent Q-learning is not immediately adaptable to the
history-based formulation, since $h_{t+1}$ and $h_{t}$ are sequences of
different lengths (and therefore not comparable). One of our contributions in
this paper is to derive a related algorithm suited to the history-based
setting.
### 2.2 Admissible Actions
We will make use of the notion of an _admissible action_ , following
terminology by ? (?).111Note that our definition technically differs from ?
(?)’s, who define an admissible action as one that is not ruled out by the
learning algorithm.
###### Definition 1.
An action $a$ is _admissible in state $s$_ if
$P(s\,|\,s,a)<1.$
That is, $a$ is admissible in $s$ if its application _may_ result in a change
in the environment state. When $P(s\,|\,s,a)=1$, we say that an action is
_inadmissible_.
We extend the notion of admissibility to histories as follows. We say that an
action $a$ is admissible given a history $h$ if it is admissible in _some_
state that is possible given $h$, or equivalently:
$\sum_{s\in\mathcal{S}}B(s\,|\,h)P(s\,|\,s,a)<1.$
We denote by $\xi(s)\subseteq\mathcal{A}$ the set of admissible actions in
state $s$. Abusing notation, we define the _admissibility function_
$\displaystyle\xi(s,a)$ $\displaystyle:=\mathbb{I}_{\left[a\in\xi(s)\right]}$
$\displaystyle\xi(h,a)$ $\displaystyle:=\Pr\\{a\in\xi(S)\\},S\sim
B(\cdot\,|\,h).$
We write $\mathcal{A}_{t}$ for the set of admissible actions given history
$h_{t}$, i.e. the actions whose admissibility in $h_{t}$ is strictly greater
than zero. In IF domains, inadmissible actions are usually dominated, and we
will deprioritize or altogether rule them out based on our estimate of
$\xi(h,a)$.
## 3 More Efficient Learning for IF Domains
Figure 2: Our IF architecture consists of three modules: a representation
generator $\Phi_{R}$ that learns an embedding for a sentence, an action scorer
$\Phi_{A}$ that chooses a network head $i$ (a feed-forward network)
conditional on score $u_{t}$, learns its Q-values and outputs
$Q(h_{t},:,u_{t})$ and finally, an auxilliary classifier $\Phi_{C}$ that
learns an approximate admissibility function $\hat{\xi}(h_{t},:)$. The
architecture is trained end-to-end.
We are interested in learning an action-value function which is close to
optimal and from which can be derived a near-optimal policy. We would also
like learning to proceed in a sample-efficient manner. In the context of IF
domains, this is hindered by both the partially observable nature of the
environment and the size of the action space. In this paper we propose two
complementary ideas that alleviate some of the issues caused by partial
observability and large action sets. The first idea contextualizes the action-
value function on a surrogate notion of progress based on total reward so far,
while the second seeks to eliminate inadmissible actions from the exploration
and learning process.
Although our ideas are broadly applicable, for concreteness we describe their
implementation in a deep reinforcement learning framework. Our agent
architecture (Figure 2) is derived from the LSTM-DRQN agent (? ?) and the work
of ? (?).
### 3.1 Score Contextualisation
In applying reinforcement learning to games, it is by now customary to
translate the player’s score differential into rewards (?; ?). Our setting is
similar to Arcade Learning Environment in the sense that the environment
provides the score. In IF, the player is awarded points for acquiring an
important object, or completing some task relevant to progressing through the
game. These awards occur in a linear, or almost linear structure, reflecting
the agent’s progression through the story, and are relatively sparse. We
emphasize that this is in contrast to the more general reinforcement learning
setting, which may provide reward for surviving, or achieving something at a
certain rate. In the video game Space Invaders, for example, the notion of
“finishing the game” is ill-defined: the player’s objective is to keep
increasing their score until they run out of lives.
We make use of the IF reward structure as follows. We call _score_ the agent’s
total (undiscounted) reward since the beginning of an episode, remarking that
the term extends beyond game-like domains. At time step $t$, the score $u_{t}$
is
$u_{t}:=\sum_{i=0}^{t-1}r_{i}.$
In IF domains, where the score reflects the agent’s progress, it is reasonable
to treat it as a state variable. We propose maintaining a separate action-
value function for each possible score. This action-value function is denoted
$Q(h_{t},a_{t},u_{t})$. We call this approach _score contextualisation_. The
use of additional context variables has by now been demonstrated in a number
of settings (? (?); ? (?); ? (?)). First, credit assignment becomes easier
since the score provides clues as to the hidden state. Second, in settings
with function approximation we expect optimization to be simpler since for
each $u$, the function $Q(\cdot,\cdot,u)$ needs only be trained on a subset of
the data, and hence can focus on features relevant to this part of the
environment.
In a deep network, we implement score contextualisation using $K$ network
heads and a map $\mathcal{J}:\mathbb{N}\to\\{1,\dots,K\\}$ such that the
$\mathcal{J}(u_{t})^{th}$ head is used when the agent has received a score of
$u_{t}$ at time $t$. This provides the flexibility to either map each score to
a separate network head, or multiple scores to one head. Taking $K=1$ uses one
monolothic network for all subtasks, and fully relies on this network to
identify state from feedback. In our experiments, we assign scores to networks
heads using a round-robin scheme with a fixed $K$. Using ? (?)’s terminology,
our architecture consists of a shared _representation generator_ $\Phi_{R}$
with $K$ independent LSTM heads, followed by a feed-forward _action scorer_
$\Phi_{A}(i)$ which outputs the action-values (Figure 2).
### 3.2 Action Gating Based on Admissibility
In this section we revisit the idea of using the admissibility function to
eliminate or more generally _gate_ actions. Consider an action $a$ which is
inadmissible in state $s$. By definition, taking this action does not affect
the state. We further assume that inadmissible actions produce a constant
level of reward, which we take to be 0 without loss of generality:
$a\text{ inadmissible in }s\implies r(s,a)=0.$
This assumption is reasonable in IF domains, and more generally holds true in
domains that exhibit subtask structure, such as the video game Montezuma’s
Revenge (? ?). We can combine knowledge of $P$ and $r$ for inadmissible
actions with Bellman’s equation (1) to deduce that for any policy $\pi$,
$a\text{ inadmissible in }s\implies
Q^{\pi}(s,a)\leq\max_{a^{\prime}\in\mathcal{A}}Q^{\pi}(s,a^{\prime})$ (4)
If we know that $a$ is inadmissible, then we do not need to learn its action-
value.
We propose learning a classifier whose purpose is to predict the admissibility
function. Given a history $h$, this classifier outputs, for each action $a$,
the probability $\hat{\xi}(h,a)$ that this action is admissible. Because of
state aliasing, this probability is in general strictly between 0 and 1;
furthermore, it may be inaccurate due to approximation error. We therefore
consider action gating schemes that are sensitive to intermediate values of
$\hat{\xi}(h,a)$. The first two schemes produce an approximately admissible
set $\hat{\mathcal{A}}_{t}$ which varies from time step to time step; the
third directly uses the definition of admissibility in a history-based
implementation of the consistent Bellman operator.
Dropout. The _dropout_ method randomly adds each action $a$ to
$\hat{\mathcal{A}}_{t}$ with probability $\hat{\xi}(h_{t},a)$.
Masking. The _masking_ method uses an elimination threshold $c\in[0,1)$. The
set $\hat{\mathcal{A}}_{t}$ contains all actions $a$ whose estimated
admissibility is at least $c$:
$\hat{\mathcal{A}}_{t}:=\\{a:\hat{\xi}(h_{t},a)\geq c\\}.$
The masking method is a simplified version of ? (?)’s action elimination
algorithm, whose threshold is adaptively determined from a confidence
interval, itself derived from assuming a value function and admissibility
functions that can be expressed linearly in terms of some feature vector.
In both the dropout and masking methods, we use the action set
$\hat{\mathcal{A}}_{t}$ in lieu of the the full action set $\mathcal{A}$ when
selecting exploratory actions.
Consistent Q-learning for histories (CQLH). The third method leaves the action
set unchanged, but instead drives the action-values of purportedly
inadmissible actions to 0. This is done by adapting the consistent Bellman
operator (3) to the history-based setting. First, we replace the indicator
$\mathbb{I}_{\left[s_{t+1}\neq s_{t}\right]}$ by the probability
$\hat{\xi}_{t}:=\hat{\xi}(h_{t},a_{t})$. Second, we drive $Q(s_{t},a_{t})$ to
$0$ in the case when we believe the state is unchanged, following the
argumentation of (4). This yields a version of consistent Q-learning which is
adapted to histories, and makes use of the predicted admissibility:
$\displaystyle\delta^{\textsc{cqlh}}_{t}=\;$ $\displaystyle
r_{t}+\gamma\max_{a\in\mathcal{A}}Q(h_{t+1},a)\hat{\xi}_{t}$
$\displaystyle+\gamma Q(h_{t},a_{t})(1-\hat{\xi}_{t})-Q(h_{t},a_{t}).$ (5)
One may ask whether this method is equivalent to a belief-state average of
consistent Q-learning when $\hat{\xi}(h_{t},a_{t})$ is accurate, i.e. equals
$\xi(h_{t},a_{t})$. In general, this is not the case: the admissibility of an
action depends on the hidden state, which in turns influences the action-value
at the next step. As a result, the above method may underestimate action-
values when there is state aliasing (e.g., $\hat{\xi}(h_{t},a_{t})\approx
0.5$), and yields smaller action gaps than the state-based version when
$\hat{\xi}(h_{t},a_{t})=1$. However, when $a_{t}$ is known to be inadmissible
($\hat{\xi}(h_{t},a_{t})=0$), the methods do coincide, justifying its use as
an action gating scheme.
We implement these ideas using an auxiliary classifier $\Phi_{C}$. For each
action $a$, this classifier outputs the estimated probability
$\hat{\xi}(h_{t},a)$, parametrized as a sigmoid function. These probabilities
are learned from bandit feedback: after choosing $a$ from history $h_{t}$, the
agent receives a binary signal $e_{t}$ as to whether $a$ was admissible or
not. In our setting, learning this classifier is particularly challenging
because the agent must predict admissibility solely based on the history
$h_{t}$. As a point of comparison, using the information-gathering commands
look and inventory to establish the state, as proposed by ? (?), leads to a
simpler learning problem, but one which does not consider the full history.
The need to learn $\hat{\xi}(h_{t},a)$ from bandit feedback also encourages
methods that generalize across histories and textual descriptions.
## 4 A Synthetic IF Benchmark
Both score contextualisation and action gating are tailored to domains that
exhibit the structure typical of interactive fiction. To assess how useful
these methods are, we will make use of a synthetic benchmark based on the
TextWorld framework (? ?). TextWorld provides a reinforcement learning
interface to text-based games along with an environment specification language
for designing new environments. Environments provide a set of locations, or
_rooms_ , objects that can picked up and carried between locations, and a
reward function based on interacting with these objects. Following the genre,
special _key_ objects are used to access parts of the environment.
Our benchmark provides seven environments of increasing complexity, which we
call _levels_. We control complexity by adding new rooms and/or objects to
each successive level. Each level also requires the agent to complete a number
of subtasks (Table 1), most of which involve carrying one or more items to a
particular location. Reward is provided only when the agent completes one of
these subtasks. Thematically, each level involves collecting food items to
make a salad, inspired by the first TextWorld competition. Example objects
include an apple and a head of lettuce, while example actions include get
apple and slice lettuce with knife. Accordingly we call our benchmark
SaladWorld.
SaladWorld provides a graded measure of an agent architecture’s ability to
deal with both partial observability and large action spaces. Indeed,
completing each subtasks requires memory of what has previously been
accomplished, along with where different objects are. Together with this, each
level in the SaladWorld involves some amount of history-dependent
admissibility i.e the admissibility of the action depends on the history
rather than the state. For example, put lettuce on counter can only be
accomplished once take lettuce (in a different room) has happened. Keys pose
an additional difficulty as they do not themselves provide reward. As shown in
Table 1, the number of possible actions rapidly increases with the number of
objects in a given level. Even the small number of rooms and objects
considered here preclude the use of tabular representations, as the state
space for a given level is the exponentially-sized cross-product of possible
object and agent locations. In fact, we have purposefully designed SaladWorld
as a small challenge for IF agents, and even our best method falls short of
solving the harder levels within the allotted training time. Full details are
given in Table 2 in the appendix.
Table 1: Main characteristics of each level in our synthetic benchmark. Level | # Rooms | # Objects | # Sub-tasks | $|\mathcal{A}|$
---|---|---|---|---
1 | 4 | 2 | 2 | 8
2 | 7 | 4 | 3 | 15
3 | 7 | 4 | 3 | 15
4 | 9 | 8 | 4 | 50
5 | 11 | 15 | 5 | 141
6 | 12 | 20 | 6 | 283
7 | 12 | 20 | 7 | 295
Table 2: Subtasks information and scores possible for each level of the suite. Level | Subtasks | Possible Scores
---|---|---
1 | Following subtasks with reward and fulfilling condition: • 10 points when the agent first enters the vegetable market. • 5 points when the agent gets lettuce from the vegetable market and puts lettuce on the counter. | 10, 15
2 | All subtasks from previous level plus this subtask: • 5 points when the agent takes the blue key from open space, opens the blue door, gets tomato from the supermarket and puts it on the counter in the kitchen. | 5, 10, 15, 20
3 | All subtasks from level 1 plus this subtask: • 5 points when the agent takes the blue key from open space, goes to the garden, opens the blue door with the blue key, gets tomato from the supermarket and puts it on the counter in the kitchen. Remark: Level 3 game differs from Level 2 game in terms of number of steps required to complete the additional sub-task (which is greater in case of Level 3) | 5, 10, 15, 20
4 | All subtasks from previous level plus this subtask: • 5 points when the agent takes parsley from the backyard and knife from the cutlery shop to the kitchen, puts parsley into fridge and knife on the counter. | 5, 10, 15, 20, 25
5 | All subtasks from previous level plus this subtask: • 5 points when the agent goes to fruit shop, takes chest key, opens container with chest key, takes the banana from the chest and puts it into the fridge in the kitchen. | 5, 10, 15, 20, 25, 30
6 | All subtasks from previous level plus this subtask: • 5 points when the agent takes the red key from the supermarket, goes to the playroom, opens the red door with the red key, gets the apple from cookhouse and puts it into the fridge in the kitchen. | 5, 10, 15, 20, 25, 30, 35
7 | All subtasks from previous level plus this subtask: • 5 points when the agent prepares the meal. | 5, 10, 15, 20, 25, 30, 35, 40
## 5 Empirical Analysis
In the first set of experiments, we use SaladWorld to establish that both
score contextualisation and action gating provide positive benefits in the
context of IF domain. We then validate these findings on the celebrated text-
based game Zork used in prior work (?; ?).
Our baseline agent is the LSTM-DRQN agent (? ?) but with a different action
representation. We augment this baseline with either or both score
contextualisation and action gating, and observe the resulting effect on agent
performance in SaladWorld. We measure this performance as the fraction of
subtasks completed during an episode, averaged over time. In all cases, our
results are generated from 5 independent trials of each condition. To smooth
the results, we use moving average with a window of 20,000 training steps. The
graphs and the histograms report average $\pm$ std. deviation across the
trials.
Score contextualisation uses $K=5$ network heads; the baseline corresponds to
$K=1$. Each head is trained using the Adam optimizer (? ?) with a learning
rate $\alpha=0.001$ to minimize a Q-learning loss (? ?) with a discount factor
of $\gamma=0.9$. The auxiliary classifier $\Phi_{C}$ is trained with the
binary cross-entropy loss over the selected action’s admissibility (recall
that our agent only observes the admissibility function for the selected
action). Training is done using a balanced form of prioritized replay which we
found improves baseline performance appreciably. Specifically, we use the
sampling mechanism described in ? (?) with prioritization i.e we sample
$\tau_{p}$ fraction of episodes that had atleast one positive reward,
$\tau_{n}$ fraction with atleast one negative reward and $1-\tau_{p}-\tau_{n}$
from whole episodic memory $\mathcal{D}$. Section C.1 in the Appendix compares
the baseline agent with and without prioritization. For prioritization,
$\tau_{p}=\tau_{n}=0.25$.
Actions are chosen from the estimated admissible set $\hat{\mathcal{A}}_{t}$
according to an $\epsilon$-greedy rule, with $\epsilon$ annealed linearly from
1.0 to 0.1 over the first million training steps. To simplify exploration, our
agent further takes a forced look action every 20 steps. Each episode lasts
for a maximum $T$ steps. For Level $1$ game, $T=100$, whereas for rest of the
levels $T=200$.To simplify exploration, our agent further takes a forced look
action every 20 steps (Section C.1).Full details are given in the Appendix
(Section A).
Figure 3: Fraction of tasks solved by each method at the end of training for
1.3 million steps. The tabular agents, which do not take history into account,
perform quite poorly. LI stands for “look, inventory” (see text for details).
### 5.1 Score Contextualisation
Figure 4: Comparing whether score contextualisation as an architecture
provides a useful representation for learning to act optimally. Row 1 and 2
correspond to Level 1 and 2 respectively.
We first consider the effect of score contextualisation on our agents’ ability
to complete tasks in SaladWorld. We ask,
> Does score contextualisation mitigate the negative effects of partial
> observability?
We begin in a simplified setting where the agent knows the admissible set
$\mathcal{A}_{t}$. We call this setting _oracle gating_. This setting lets us
focus on the impact of contextualisation alone. We compare our score
contextualisation (SC) to the baseline and also to two “tabular” agents. The
first tabular agent treats the most recent feedback as state, and hashes each
unique description-action pair to a Q-value. This results in a memoryless
scheme that ignores partial observability. The second tabular agent performs
the information-gathering actions look and inventory to construct its state
description, and also hashes these to unique Q-values. Accordingly, we call
this the “LI-tabular” agent. This latter scheme has proved to be a successful
heuristic in the design of IF agents (? ?), but can be problematic in domains
where taking information-gathering actions can have negative consequences (as
is the case in Zork).
Figure 3 shows the performance of the four methods across SaladWorld levels,
after 1.3 million training steps. We observe that the tabular agents’
performance suffers as soon as there are multiple subtasks, as expected. The
baseline agent performs well up to the third level, but then shows
significantly reduced performance. We hypothesize that this occurs because the
baseline agent must estimate the hidden state from longer history sequences
and effectively learn an implicit contextualisation. Beyond the fourth level,
the performance of all agents suffers, suggesting the need for a better
exploration strategy, for example using expert data (?).
We find that score contextualisation performs better than the baseline when
the admissible set is unknown. Figure 4 compares learning curves of the SC and
baseline agents with oracle gating and using the full action set,
respectively, in the simplest of levels (Level 1 and 2). We find that score
contextualisation can learn to solve these levels even without access to
$\mathcal{A}_{t}$, whereas the baseline cannot. Our results also show that
oracle gating simplifies the problem, and illustrate the value in handling
inadmissible actions differently.
We hypothesize that score contextualisation results in a simpler learning
problem in which the agent can more easily learn to distinguish which actions
are relevant to the task, and hence facilitate credit assignment. Our result
indicates that it might be unreasonable to expect contextualisation to arise
naturally (or easily) in partially observable domains with large actions sets.
We conclude that score contextualisation mitigates the negative effects of
partial observability.
Figure 5: Fraction of tasks solved by each method at the end of training for
1.3 million steps. Except in Level 1, action gating by itself does not improve
end performance. Figure 6: Effectiveness of action gating with score
contextualisation in Level 3. Of the three methods, masking performs best.
### 5.2 Score Contextualisation with Learned Action Gating
The previous experiment (in particular, Figure 4) shows the value of
restricting action selection to admissible actions. With the goal in mind of
designing an agent that can operate from feedback alone, we now ask:
> Can an agent learn more efficiently when given bandit feedback about the
> admissibility of its chosen actions?
We address this question by comparing our three action gating mechanisms. As
discussed in Section 3.2, the output of the auxiliary classifier describes our
estimate of an action’s admissibility for a given history.
As an initial point of comparison, we tested the performance of the baseline
agent when using the auxiliary classifier’s output to gate actions. For the
masking method, we selected $c=0.001$ from a larger initial parameter sweep.
The results are summarized in Figure 5. While action gating alone provides
some benefits in the first level, performance is equivalent for the rest of
the levels.
However, when combined with score contextualisation (see Fig 6, 7), we observe
some performance gains. In Level 3 in particular, we almost recover the
performance of the SC agent with oracle gating. From our results we conclude
that masking with the right threshold works best, but leave as an open
question whether the other action gating schemes can be improved.
Figure 8 shows the final comparison between the baseline LSTM-DRQN and our new
agent architecture which incorporates action gating and score
contextualisation (full learning curves are provided in the appendix, Figure
14). Our results show that the augmented method significantly outperforms the
baseline, and is able to handle more complex IF domains. From level 4 onwards,
the learning curves in the appendix show that combining score
contextualisation with masking results in faster learning, even though final
performance is unchanged. We posit that better exploration schemes are
required for further progress in SaladWorld.
Figure 7: Fraction of tasks solved by each method at the end of training for
1.3 million steps. For first 3 levels, SC + Masking is better or equivalent to
SC. For levels 4 and beyond, better exploration strategies are required.
Figure 8: Score contextualisation and masking compared to the baseline agent.
We show the fraction of tasks solved by each method at the end of training for
1.3 million steps.
### 5.3 Zork
As a final experiment, we evaluate our agent architecture on the interactive
fiction Zork I, the first installment of the popular trilogy. Zork provides an
interesting point of comparison for our methods, as it is designed by and for
humans – following the ontology of ? (?), it is a domain which is both
_interesting_ and _independent_. Our main objective is to compare the
different methods studied with ? (?)’s AE-DQN agent. Following their
experimental setup, we take $\gamma=0.8$ and train for 2 million steps. All
agents use the smaller action set (131 actions). Unlike AE-DQN, however, our
agent does not use information-gathering actions (look and inventory) to
establish the state.
Figure 9 shows the corresponding learning curves. Despite operating in a
harder regime than AE-DQN, the score contextualizing agent reaches a score
comparable to AE-DQN, in about half of the training steps. All agents
eventually fail to pass the 35-point benchmark, which corresponds to a
particularly difficult in-game task (the “troll quest”) which involves a
timing element, and we hypothesize requires a more intelligent exploration
strategy.
Figure 9: Learning curves for different agents in Zork.
## 6 Related Work
RL applied to Text Adventure games: LSTM-DQN by ? (?) deals with parser-based
text adventure games and uses an LSTM to generate feedback representation. The
representation is then used by an action scorer to generate scores for the
action verb and objects. The two scores are then averaged to determine Q-value
for the state-action pair. In the realm of choice-based games, ? (?) uses two
separate deep neural nets to generate representation for feedback and action
respectively. Q-values are calculated by dot-product of these representations.
None of the above approaches deals with partial observability in text
adventure games.
Admissible action set learning: ? (?) approach the issue of learning
admissible set given context as a supervised learning one. They train their
model on (input, label) pairs where input is context (concatenation of
feedbacks by look and inventory) and label is the list of admissible commands
given this input. AE-DQN (? ?) employs an additional neural network to prune
in-admissible actions from action set given a state. Although the paper
doesn’t deal with partial observability in text adventure games, authors show
that having a tractable admissible action set led to faster convergence. ? (?)
work on bounding the action set through affordances. Their agent is trained
through tabular Q-Learning.
Partial Observability: ? (?) replace the shared MLP in ? (?) with an LSTM cell
to calculate context representation. However, they use concatenation of
feedbacks by look and inventory as the given state to make the game more
observable. Their work also doesn’t focus on pruning in-admissible actions
given a context. Finally, ? (?) deal with partial observability by
representing state as a knowledge graph and continuously updating it after
every game step. However, the graph update rules are hand-coded; it would be
interesting to see they can be learned during gameplay.
## 7 Conclusions and Future work
We introduced two algorithmic improvements for deep reinforcement learning
applied to interactive fiction (IF). While naturally rooted in IF, we believe
our ideas extend more generally to partially observable domains and large
discrete action spaces. Our results on SaladWorld and Zork show the usefulness
of these improvements. Going forward, we believe better contextualisation
mechanisms should yield further gains. In Zork, in particular, we hypothesize
that going beyond the 35-point limit will require more tightly coupling
exploration with representation learning.
## 8 Acknowledgments
This work was funded by the CIFAR Learning in Machines and Brains program.
Authors thank Compute Canada for providing the computational resources.
## References
* [2019] Ammanabrolu, P., and Riedl, M. 2019\. Playing text-adventure games with graph-based deep reinforcement learning. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 3557–3565. Minneapolis, Minnesota: Association for Computational Linguistics.
* [2018] Atkinson, T.; Baier, H.; Copplestone, T.; Devlin, S.; and Swan, J. 2018\. The text-based adventure AI competition. IEEE Transactions on Games.
* [2013] Bellemare, M. G.; Naddaf, Y.; Veness, J.; and Bowling, M. 2013\. The arcade learning environment: An evaluation platform for general agents. Journal of Artificial Intelligence Research 47:253–279.
* [2016a] Bellemare, M.; Srinivasan, S.; Ostrovski, G.; Schaul, T.; Saxton, D.; and Munos, R. 2016a. Unifying count-based exploration and intrinsic motivation. In Lee, D. D.; Sugiyama, M.; Luxburg, U. V.; Guyon, I.; and Garnett, R., eds., Advances in Neural Information Processing Systems 29. Curran Associates, Inc. 1471–1479.
* [2016b] Bellemare, M. G.; Ostrovski, G.; Guez, A.; Thomas, P. S.; and Munos, R. 2016b. Increasing the action gap: New operators for reinforcement learning. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, AAAI’16, 1476–1483. AAAI Press.
* [2018] Côté, M.-A.; Kádár, A.; Yuan, X.; Kybartas, B.; Barnes, T.; Fine, E.; Moore, J.; Hausknecht, M.; Asri, L. E.; Adada, M.; Tay, W.; and Trischler, A. 2018\. Textworld: A learning environment for text-based games. CoRR abs/1806.11532.
* [2017] Fulda, N.; Ricks, D.; Murdoch, B.; and Wingate, D. 2017\. What can you do with a rock? affordance extraction via word embeddings. Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence.
* [2018] Ghosh, D.; Singh, A.; Rajeswaran, A.; Kumar, V.; and Levine, S. 2018\. Divide-and-conquer reinforcement learning. ICLR 2018.
* [2015] Hausknecht, M., and Stone, P. 2015\. Deep recurrent q-learning for partially observable mdps. In AAAI Fall Symposium on Sequential Decision Making for Intelligent Agents (AAAI-SDMIA15).
* [2016] He, J.; Chen, J.; He, X.; Gao, J.; Li, L.; Deng, L.; and Ostendorf, M. 2016\. Deep reinforcement learning with a natural language action space. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1621–1630. Berlin, Germany: Association for Computational Linguistics.
* [2018] Icarte, R. T.; Klassen, T. Q.; Valenzano, R.; and McIlraith, S. A. 2018\. Using reward machines for high-level task specification and decomposition in reinforcement learning. In Proceedings of the International Conference on Machine Learning.
* [2015] Kingma, D. P., and Ba, J. 2015\. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR).
* [2017] Lample, G., and Chaplot, D. S. 2017\. Playing fps games with deep reinforcement learning. In Thirty-First AAAI Conference on Artificial Intelligence.
* [1995] McCallum, A. K. 1995\. Reinforcement learning with selective perception and hidden state. Ph.D. Dissertation, University of Rochester.
* [2015] Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A. A.; Veness, J.; Bellemare, M. G.; Graves, A.; Riedmiller, M.; Fidjeland, A. K.; Ostrovski, G.; et al. 2015\. Human-level control through deep reinforcement learning. Nature 518(7540):529.
* [2015] Narasimhan, K.; Kulkarni, T.; and Barzilay, R. 2015\. Language understanding for text-based games using deep reinforcement learning. Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.
* [2018] OpenAI. 2018\. Gym retro. https://github.com/openai/retro.
* [2019] Rakelly, K.; Zhou, A.; Quillen, D.; Finn, C.; and Levine, S. 2019\. Efficient off-policy meta-reinforcement learning via probabilistic context variables. ICML 2019.
* [2010] Silver, D., and Veness, J. 2010\. Monte-carlo planning in large pomdps. In Advances in Neural Information Processing Systems.
* [1998] Sutton, R. S., and Barto, A. G. 1998\. Reinforcement learning: An introduction. MIT Press.
* [2018] Tao, R. Y.; Côté, M.-A.; Yuan, X.; and Asri, L. E. 2018\. Towards solving text-based games by producing adaptive action spaces.
* [2019] Tessler, C.; Zahavy, T.; Cohen, D.; Mankowitz, D. J.; and Mannor, S. 2019\. Sparse imitation learning for text based games with combinatorial action spaces. The Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM) 2019.
* [1989] Watkins, C. J. C. H. 1989\. Learning from delayed rewards. Ph.D. Dissertation, Cambridge University, Cambridge, England.
* [2018] Yuan, X.; Côté, M.-A.; Sordoni, A.; Laroche, R.; des Combes, R. T.; Hausknecht, M.; and Trischler, A. 2018\. Counting to explore and generalize in text-based games.
* [2018] Zahavy, T.; Haroush, M.; Merlis, N.; Mankowitz, D. J.; and Mannor, S. 2018\. Learn what not to learn: Action elimination with deep reinforcement learning. In Proceedings of the 32Nd International Conference on Neural Information Processing Systems, NIPS’18, 3566–3577. USA: Curran Associates Inc.
## Appendix A Training Details
### A.1 Hyper-parameters
Training hyper-parameters: For all the experiments unless specified,
$\gamma=0.9$. Weights for the learning agents are updated every $4$ steps.
Agents with score contextualisation architecture have $K=5$ network heads.
Parameters of score contextualisation architecture are learned end to end with
Adam optimiser (? ?) with learning rate $\alpha=0.001$. To prevent imprecise
updates for the initial states in the transition sequence due to in-sufficient
history, we use updating mechanism proposed by ? (?). In this mechanism,
considering the transition sequence of length $l,o_{1},o_{2},\dots,o_{l}$,
errors from $o_{1},o_{2},\dots,o_{n}$ aren’t back-propagated through the
network. In our case, the sequence length $l=15$ and minimum history size for
a state to be updated $n=6$ for all experiments. Score contextualisation heads
are trained to minimise the Q-learning loss over the whole transition
sequence. On the other hand, $\Phi_{C}$ minimises the BCE (binary cross-
entropy) loss over the predicted admissibility probability and the actual
admissibility signal for every transition in the transition sequence. The
behavior policy during training is $\epsilon-$greedy over the admissible set
$\hat{\mathcal{A}}_{t}$. Each episode lasts for a maximum $T$ steps. For Level
$1$ game, we anneal $\epsilon=1$ to $0.1$ over $1000000$ steps and $T=100$.
For rest of the games in the suite, we anneal $\epsilon=1$ to $0.1$ over
$1000000$ steps and $T=200$.
Architectural hyper-parameters: In $\Phi_{R}$, word embedding size is $20$ and
the number of hidden units in encoder LSTM is $64$. For a network head $k$,
the number of hidden units in context LSTM is $512$; $\Phi_{A}(k)$ is a two
layer MLP: sizes of first and second layer are 128 and $|\mathcal{A}|$
respectively. $\Phi_{C}$ has the same configuration as $\Phi_{A}(k)$.
Figure 10: Comparing the effect of oracle gating versus learning admissibility
from bandit feedback. Learning is faster in case of oracle gating since agent
is given admissible action set resulting in overall better credit assignment.
Figure 11: Final comparison shows that our algorithmic enhancements improve
the baseline. We show number of tasks solved by each method at the end of
training for 1.3 million steps.
### A.2 Action Gating Implementation
For _dropout_ and _masking_ when selecting actions, we set
$Q(h_{t},a_{t})=-\infty$ for $a\notin\hat{\mathcal{A}}_{t}$. Since
$\hat{\xi}_{t}$ is basically an estimate for admissibility for action $a$
given history $h_{t}$, we use (5) to implement consistent Q value backups:
$Q(h_{t},a_{t})\leftarrow Q(h_{t},a_{t})+\alpha\delta^{\prime\prime}_{t}$
where
$\delta^{\prime\prime}_{t}=r_{t}+\gamma\big{[}\max_{a\in\mathcal{A}}Q(h_{t+1},a)\hat{\xi}_{t}+\\\
Q(h_{t+1},a_{t})(1-\hat{\xi}_{t})\big{]}-Q(h_{t},a_{t})$
We notice by using the above equation, that for an action $a$ inadmissible in
$s$, it’s value indeed reduces to $0$ over time.
### A.3 Baseline Modifications
We modify LSTM-DRQN (? ?) in two ways. First, we concatenate the
representations $\Phi_{R}(o_{t})$ and $\Phi_{R}(a_{t-1})$ before sending it to
the history LSTM, in contrast ? (?) concatenates the inputs $o_{t}$ and
$a_{t-1}$ first and then generates $\Phi_{R}([o_{t};a_{t-1}])$. Second, we
modify the action scorer as action scorer in the LSTM-DRQN could only handle
commands with two words.
## Appendix B Notations and Algorithm
Following are the notations important to understand the algorithm:
* •
$o_{t},r_{t},e_{t}:$ observation (i.e. feedback), reward and admissibility
signal received at time $t$.
* •
$a_{t}:$ command executed in game-play at time $t$.
* •
$u_{t}:$ cumulative rewared/score at time $t$.
* •
$\Phi_{R}:$ representation generator.
* •
$\Phi_{C}:$ auxiliary classifier.
* •
$K$ : number of network heads in score contextualisation architecture.
* •
$\mathcal{J}$: dictionary mapping cumulative rewards to network heads.
* •
$H(k):$ LSTM corresponding to network head $k$.
* •
$\Phi_{A}(k):$ Action scorer corresponding to network head $k$.
* •
$h_{t}:$ agent’s context/history state at time $t$.
* •
$T$: maximum steps for an episode.
* •
$p_{i}:$ boolean that determines whether +ve reward was received in episode
$i$.
* •
$q_{i}:$ boolean that determines whether -ve reward was received in episode
$i$.
* •
$\tau_{p}:$ fraction of episodes where $\exists t<T:r_{t}>0$
* •
$\tau_{n}:$ fraction of episodes where $\exists t<T:r_{t}<0$
* •
$l:$ sequence length.
* •
$n:$ minimum history size for a state to be updated.
* •
$\mathcal{A}:$ action set.
* •
$\hat{\mathcal{A}}_{t}:$ admissible set generated at time $t$.
* •
$I_{target}:$ update interval for target network
* •
$\epsilon:$ parameter for $\epsilon-$greedy exploration strategy.
* •
$\epsilon_{1}:$ softness parameter i.e. $\epsilon_{1}$ fraction of times
$\hat{\mathcal{A}}_{t}=\mathcal{A}$.
* •
$c:$ threshold parameter for action elimination strategy Masking.
* •
$G_{\textrm{max}}:$ maximum steps till which training is performed.
Full training procedure is listed in Algorithm 1.
Algorithm 1 General training procedure
1:function
ACT$(o_{t},a_{t-1},u_{t},h_{t-1},\mathcal{J},\epsilon,\epsilon_{1},c,\theta)$
2: Get network head $k=\mathcal{J}(u_{t})$.
3: $h_{t}\leftarrow$ LSTM $H(k)[w_{t},h_{t-1}]$.
4:
${Q}(h_{t},:,u_{t};\theta)\leftarrow\Phi_{A}(k)(h_{t});\hat{\xi}(h_{t},a;\theta)\leftarrow\Phi_{C}(h_{t})$.
5: Generate $\hat{\mathcal{A}}_{t}$ (see Section 3.2).
6: With probability
$\epsilon,a_{t}\leftarrow\textrm{Uniform}(\hat{\mathcal{A}}_{t})$, else
$a_{t}\leftarrow\textrm{argmax}_{a\in\hat{\mathcal{A}}_{t}}{Q}(h_{t},a,u_{t};\theta)$
7: return $a_{t},h_{t}$
8:end function
9:function TARGETS$(f,\gamma,\theta^{-})$
10:
$(a_{0},o_{1},a_{1},r_{2},u_{2},e_{2},o_{2},\dots,o_{l},a_{l},r_{l+1},e_{l+1},u_{l+1})\leftarrow
f;h_{b,0}\leftarrow 0$
11: Pass transition sequence through $H$ to get
$h_{b,1},h_{b,2},\dots,h_{b,l}$
12: $E_{b,i}\leftarrow\xi(h_{b,i},a_{i};\theta^{-})$
13:
$y_{b,i}\leftarrow\textrm{max}_{a\in\mathcal{A}}{Q}(h_{b,i+1},a,u_{b,i+1};\theta^{-})$
14: $y_{b,i}\leftarrow E_{b,i}y_{b,i}+(1-E_{b,i})\
{Q}(h_{b,i+1},a_{i},u_{b,i+1};\theta^{-})$ if using CQLH.
15: $y_{b,i}\leftarrow r_{i+1}$ if $o_{i}$ is terminal else $y_{b,i}\leftarrow
r_{i+1}+\gamma y_{b,i}$
16: return $y_{b,:},E_{b,:}$
17:end function
18:
19:Input:
$G_{\textrm{max}},I_{\text{look}},I_{\text{update}},\gamma,\epsilon_{1},\epsilon,c,K,\mathbb{I}_{\left[{}_{\text{using}\Phi_{C}}\right]},n$
20:Initialize episodic replay memory $\mathcal{D}$, global step counter
$G\leftarrow 0$, dictionary $\mathcal{J}=\\{\\}$.
21:Initialize parameters $\theta$ of the network, target network parameter
$\theta^{-}\leftarrow\theta$.
22:while $G<G_{\textrm{max}}$ do
23: Initialize score $u_{1}=0$, hidden State of $H$, $h_{0}=0$ and get start
textual description $o_{1}$ and initial command $a_{0}=\textrm{'look'}$. Set
$p_{k}\leftarrow 0,q_{k}\leftarrow 0$.
24: for $t\leftarrow 1$ to $T$ do
25: $a_{t},h_{t}\leftarrow$
ACT$(o_{t},a_{t-1},u_{t},h_{t-1},\mathcal{J},\epsilon,\epsilon_{1},c,\theta)$
26: $a_{t}\leftarrow$ ’look’ if $t\bmod 20==0$
27: Execute action $a_{t}$, observe $\\{r_{t+1},o_{t+1},e_{t+1}\\}$.
28: $p_{k}\leftarrow 1$ if $r_{t}>0$; $q_{k}\leftarrow 1$ if $r_{t}<0$;
$u_{t+1}\leftarrow u_{t}+r_{t}$
29: Sample minibatch of transition sequences $f$
30: $y_{b,:},E_{b,:}\leftarrow$ TARGETS$(f,\gamma,\theta^{-})$
31: Perform gradient descent on
$\mathcal{L}(\theta)=\sum_{i=j+n-1}^{j+l}[y_{b,i}-{Q}(h_{b,i},a_{i},u_{b,i};\theta)^{2}+\mathbb{I}_{\left[{}_{using\Phi_{C}}\right]}\
\textrm{BCE}(e_{i},E_{b,i})]$
32: $\theta^{-}\leftarrow\theta$ if $t\bmod I_{update}==0$
33: $G\leftarrow G+1$
34: End episode if $o_{t+1}$ is terminal.
35: end for
36: Store episode in $\mathcal{D}$.
37:end while
## Appendix C More Empirical Analysis
### C.1 Prioritised Sampling & Infrequent look
Our algorithm uses prioritised sampling and executes a look action every
$I_{look}=20$ steps. The baseline agent LSTM-DRQN follows this algorithm. We
now ask,
> Does prioritised sampling and an infrequent look play a significant role in
> the baseline’s performance?
Figure 12: Learning curves for baseline ablation study.
For this experiment, we compare the Baseline to two agents. The first agent is
the Baseline without prioritised sampling and the second is the one without an
infrequent look. Accordingly, we call them “No-priority (NP)” and “No-look
(NL)” respectively. We use Zork as the testing domain.
From Fig 12, we observe that the Baseline performs better than the NP agent.
This is because prioritised sampling helps the baseline agent to choose the
episodes in which rewards are received in, thus assigning credit to the
relevant states faster and overall better learning. In the same figure, the
Baseline performs slightly better than the NL agent. We hypothesise that even
though look command is executed infrequently, it helps the agent in
exploration and do credit assignment better.
### C.2 CQLH
Our algorithm uses CQLH implementation as described in Section A.2. An
important case that CQLH considers is $s_{t+1}=s_{t}$. This manifests in
$(1-\hat{\xi})$ term in equation (5). We now ask whether ignoring the case
$s_{t+1}=s_{t}$ worsen the agent’s performance?
Figure 13: Learning curves for CQLH ablation study.
For this experiment, we compare CQLH agent with the agent which uses this
error for update:
$\displaystyle\delta^{\prime\prime}_{t}=$ $\displaystyle
r_{t}+\gamma\max_{a\in\mathcal{A}}Q(h_{t+1},a)\hat{\xi}_{t}-Q(h_{t},a_{t}).$
Accordingly, we call this new agent as “alternate CQLH (ACQLH)” agent. We use
Zork as testing domain. From Fig 13, we observe that although ACQLH has a
simpler update rule, its performance seems more unstable compared to the CQLH
agent.
Figure 14: Learning curves for Score contextualisation (SC : red), Score
contextualisation + Masking (SC + Masking : blue) and Baseline (grey) for all
the levels of the SaladWorld. For the simpler levels i.e. level 1 and 2, SC
and SC + Masking perform better than Baseline. With difficult level 3, only SC
+ Masking solves the game. For levels 4 and beyond, we posit that better
exploration strategies are required. Figure 15: Game map shows Level 1 and 2.
Simpler levels help us test the effectiveness of score contextualisation
architecture. Figure 16: Game map shows the _progression_ from Level 3 to
Level 7 in terms of number of rooms and objects. Besides every object in the
map, there is a tuple which shows in which levels the object is available. We
observe that with successive levels, the number of rooms and objects increase
making it more difficult to solve these levels.
|
# Turbulence in Milky Way Star-Forming Regions Traced by Young Stars and Gas
Trung Ha Department of Physics, University of North Texas, Denton, TX 76203,
USA<EMAIL_ADDRESS>Yuan Li Department of Physics, University of North
Texas, Denton, TX 76203, USA Marina Kounkel Physics & Astronomy Department,
Vanderbilt University, PMB 401807, 2301 Vanderbilt Place, Nashville, TN 37235
Siyao Xu NASA Hubble Fellow Institute for Advanced Study, 1 Einstein Drive,
Princeton, NJ 08540, USA Hui Li NASA Hubble Fellow Department of Astronomy,
Columbia University, 550 West 120th Street, New York, NY 10027, USA Yong
Zheng Department of Astronomy, University of California, Berkeley, CA 94720,
USA
###### Abstract
The interstellar medium (ISM) is turbulent on all scales and in all phases. In
this paper, we study turbulence with different tracers in four nearby star-
forming regions: Orion, Ophiuchus, Perseus, and Taurus. We combine the
APOGEE-2 and Gaia surveys to obtain the full 6-dimensional measurements of
positions and velocities of young stars in these regions. The velocity
structure functions (VSFs) of the stars show a universal scaling of
turbulence. We also obtain H$\alpha$ gas kinematics in these four regions from
the Wisconsin H-Alpha Mapper. The VSFs of the H$\alpha$ are more diverse
compared to the stars. In regions with recent supernova activities, they show
characteristics of local energy injections and higher amplitudes compared to
the VSFs of stars and of CO from the literature. Such difference in amplitude
of the VSFs can be explained by the different energy and momentum transport
from supernovae into different phases of the ISM, thus resulting in higher
levels of turbulence in the warm ionized phase traced by H$\alpha$. In regions
without recent supernova activities, the VSFs of young stars, H$\alpha$, and
CO are generally consistent, indicating well-coupled turbulence between
different phases. Within individual regions, the brighter parts of the
H$\alpha$ gas tend to have a higher level of turbulence than the low-emission
parts. Our findings support a complex picture of the Milky Way ISM, where
turbulence can be driven at different scales and inject energy unevenly into
different phases.
††facilities: Gaia, Sloan, WHAM††software: astropy (Astropy Collaboration et
al., 2013, 2018), matplotlib (Hunter, 2007), numpy (van der Walt et al., 2011;
Harris et al., 2020), pandas (pandas development team, 2020; Wes McKinney,
2010), scipy (Virtanen et al., 2020), Sagitta (McBride et al., 2021)
## 1 Introduction
The interstellar medium (ISM) is multi-phase and turbulent. Interstellar
turbulence affects the velocity statistics on all scales from the diffuse
ionized gas to the dense molecular gas. Larson (1981) compiled the internal
velocity dispersion of nearby molecular clouds from the width of various
emission lines, and found a power-law relation between the velocity dispersion
and cloud size, with slope 0.38, between the 1/3 slope expected for turbulence
dominated by solenoidal motions (Kolmogorov, 1941) and the 1/2 slope expected
for shock-dominated turbulence (e.g., Galtier & Banerjee, 2011; Federrath,
2013).
Inside molecular clouds, massive O- and B-type stars create H II regions by
ionizing and heating the ISM with UV radiation (see, e.g., Spitzer, 1978).
Compression from supernova (SN) explosions can trigger further star formation
(Kennicutt & Evans, 2012). Interstellar processes such as SNe, stellar winds,
ionizing radiation, and cosmic rays all play a role in maintaining the
structure of the ISM, as well as in regulating star formation (Hennebelle &
Falgarone, 2012; Rathjen et al., 2021; Hu et al., 2022; Xu & Lazarian, 2022).
Of these mechanisms, SN explosions are perhaps the most significant source of
energy injection into the local ISM (Mac Low & Klessen, 2004; Padoan et al.,
2016).
Many numerical simulations have been carried out to study how SN explosions
affect the surrounding ISM and star formation in the last decade. In general,
they find that most of the kinetic energy from an SN explosion is transferred
to the low-density region of the ISM in both single- (see, e.g., Kim &
Ostriker, 2015; Martizzi et al., 2015; Li et al., 2015) and multiple-SN events
(Padoan et al., 2016), suggesting that turbulence may not be evenly driven in
different phases of the ISM.
Observational studies of ISM turbulence have been done with various gas
tracers. Measurements of turbulence through density spectrum were carried out
using the Wisconsin H-Alpha Mapper (WHAM; Haffner et al., 2003, 2010), which
revealed a turbulent cascade consistent with Kolmogorov theory (Chepurnov &
Lazarian, 2010). Burkhart et al. (2010) found that most of the H I in the
Small Magellanic Cloud (SMC) is subsonic or transonic, and concluded that
turbulence in the SMC is dominated by hydrodynamical pressure over magnetic
pressure.
Statistical measurements of velocity fluctuations in the neutral ISM can be
done using the Velocity Channel Analysis, the Velocity Coordinate Spectrum
(Lazarian & Pogosyan, 2000, 2006), the Delta-variance technique (Stutzki et
al., 1998; Ossenkopf & Mac Low, 2002), or the Principal Component Analysis
(Heyer & Brunt, 2004; Roman-Duval et al., 2011) based on spectroscopic data.
These analyses generally reveal power-law spectra of turbulent velocities
(Lazarian, 2009; Chepurnov et al., 2010).
Additionally, kinematics of Galactic H II regions have been studied using the
observations of ionized gas. Fabry-Perot interferometry was used to probe the
H$\alpha$ line in emission nebulae. Louise & Monnet (1970) found a velocity-
size relation that agrees with the 1/3 slope of Kolmogorov turbulence in M8,
while Roy & Joncas (1985) and Joncas & Roy (1986) found slopes of the
structure functions closer to 1/2, indicating turbulence in the supersonic
regime in both S142 and M17, respectively.
Closer to our solar neighborhood, studies into the nature of turbulence in the
Orion Nebula is an ongoing effort since the 1950s. von Hoerner (1951) observed
a root-mean-square difference in emission line velocities that increases with
larger angular separations. He reasoned that such relation could be explained
by the Kolmogorov theory if the effective depth of the nebula was sufficiently
small as to neglect projection effects. Subsequent studies, carried out by
Münch (1958) and Castaneda (1988), disputes this conclusion. Münch (1958)
found an optical depth smaller than the value postulated by von Hoerner
(1951), and attributed the failure of Kolmogorov theory to the compressibility
of turbulence in the Orion Nebula, while Castaneda (1988) argued against a
simple power-law relation, instead suggested energy injections at different
scales.
ISM turbulence has also been probed using “point sources” such as molecular
cores. Qian et al. (2012) developed statistical measurements of core-to-core
velocity dispersion over a range of length scales , i.e., the core velocity
dispersion technique. Qian et al. (2018) find signatures for a turbulent
power-law spectrum in the Taurus Molecular Cloud, and demonstrate that the
statistics of turbulent velocities are imprinted in the velocities of dense
cores (Xu, 2020).
One limitation of using gas tracers is they lack three-dimensional information
on the position and the velocity information is also limited to only the line-
of-sight. In Ha et al. (2021, hereafter Paper I), we proposed a new method to
measure turbulence in the ISM with young stars. Since stars are formed out of
turbulent molecular clouds, we expected these young stars to inherit the
turbulent kinematics from their natal clouds. Using six-dimensional position
and velocity information of $>1400$ stars in the Orion Complex, we found that
the first-order velocity structure function (VSF) of young stars in loose
groups exhibit characteristics of turbulence. Our method can therefore be used
to probe the turbulent properties of molecular clouds in addition to analysis
of gas kinematics and has been used in recent work in Krolikowski et al.
(2021); Zhou et al. (2022).
In this work, we apply a similar analysis to a larger sample of molecular
complexes, which include three of the nearest massive star-forming regions:
Ophiuchus, Perseus, and Taurus. Additionally, we analyze the kinematics of
H$\alpha$ gas from the WHAM survey to study the connection between stars and
the multi-phase ISM.
In Section 2, we describe the data used in this analysis, the reduction
pipeline for the H$\alpha$ kinematic survey, and how we compute the VSFs using
stars and H$\alpha$. In Section 3, we present the VSF of young stars and
multi-phase gas of all four regions. In Section 4, we discuss sources of
uncertainty associated with our analysis and the nature of turbulence in the
Milky Way ISM. We summarize out results in Section 5.
Figure 1: Top panels: positions of young stars color-coded by their distances
in Orion, Ophiuchus, Taurus-Perseus (from left to right). Background shows
H$\alpha$ flux from the WHAM survey (lighter is brighter). The orange and
green boxes denote the areas where H$\alpha$ is analyzed for Taurus and
Perseus, respectively. Bottom panels show the line-of-sight velocities of the
young stars (in the LSR frame) in these regions. The background shows the
H$\alpha$ line-of-sight centroid velocities. Table 1: Properties of Star-
Forming Regions
Region | Number | Number of | Average | GLON | GLAT | Distance | $v_{\rm LSR}$ range | $v_{\rm LSR,~{}err}$ cut
---|---|---|---|---|---|---|---|---
| of stars | stars used | distance (pc) | (°) | (°) | range (pc) | ($\mathrm{km~{}s}^{-1}$) | ($\mathrm{km~{}s}^{-1}$)
Orion | 2647 | 2468 | 410 | $190,217$ | $-28,-5$ | $300,500$ | $-10,40$ | $10$
Ophiuchus | 113 | 107 | 140 | $342,360$ | $12,28$ | $120,170$ | $-30,30$ | $2$
Perseus | 159 | 138 | 320 | $156,164$ | $-23,-15$ | $250,400$ | $-6,30$ | $10$
Taurus | 157 | 139 | 140 | $165,177$ | $-23,-11$ | $97,200$ | $-5,25$ | $5$
Note. — Number of stars are all the stars with 6D information. Number of stars
used are the stars in our analysis after excluding outliers. GLON (Galactic
Longitude) and GLAT (Galactic Latitude) ranges denote the borders of windows
used to compute the H$\alpha$ VSF. The distance ranges, $v_{\rm LSR}$ ranges,
and $v_{\rm LSR,~{}err}$ cuts are applied to the stars to exclude outliers.
## 2 Observations, Data reduction, and Methods
### 2.1 Observations and Data Analysis
#### 2.1.1 Stellar observations
We obtain the locations, parallax distances, and proper motions of stars from
the Gaia mission (Gaia Collaboration et al., 2016) in four nearby star-forming
regions: the Orion Molecular Cloud complex (Orion), the $\rho$ Oph dark cloud
(Ophiuchus), the Perseus molecular cloud traced by two clusters, NGC 1333 and
IC 348 (Perseus), and the Taurus molecular clouds (Taurus). The Gaia catalog
includes five-parameter astrometric measurements (plane-of-sky coordinates,
proper motions, and parallax distances) of over 1.46 billion stars in the
Milky Way. Only a very limited subset ($\sim 7$ million) has line-of-sight
velocities measured by Gaia. The five-dimensional astrometric data for our
four regions are obtained from Gaia’s third early data release (EDR3; Gaia
Collaboration et al., 2021).
The line-of-sight velocities of the stars in our study were observed in the
near infrared with the Apache Point Observatory Galactic Evolution Experiment
(APOGEE) spectrograph, mounted on the 2.5 m Sloan Foundation Telescope of the
Sloan Digital Sky Survey. APOGEE-2 Data Release 17 (DR17) collected high
resolution spectral image of over 657,000 stars in the Milky Way (Gunn et al.,
2006; Blanton et al., 2017; Abdurro’uf et al., 2022). Combining these two
surveys provides six-dimensional (6D) information (3D positions and 3D
velocities) of stars for the four molecular clouds.
The stars in our study were assigned to each of the four region by Kounkel &
Covey (2019). They performed a clustering analysis on Gaia DR2 sources in the
Milky Way using Python implementation of HDBSCAN (Hierarchical Density-Based
Spatial Clustering of Applications with Noise; McInnes et al., 2017). We
crossmatch this catalog with Gaia EDR3 and APOGEE-2 to obtain 6D astrometries
of the stars.
Then, we exclude a small number of stars whose parallax distances $d$, line-
of-sight velocities $v_{\rm LSR}$, and line-of-sight velocity errors $v_{\rm
LSR,~{}err}$ are more than 4$\sigma$ deviated from the mean of the
distribution. In Taurus, one star with very high reported proper motion
($v_{\rm LSR,~{}RA}>100\mathrm{~{}mas~{}yr}^{-1}$) is excluded. The ranges of
$d$ and $v_{\rm LSR}$, and the threshold of $v_{\rm LSR,~{}err}$ that we use,
as well as the total number of stars before and after the cut are specified in
Table 1. We note, however, that the exact cut of these ranges do not affect
the conclusion of this work.
#### 2.1.2 H$\alpha$ observations
The H$\alpha$ data in our study comes from the WHAM survey (Haffner et al.,
2003, 2010). We obtained the data cubes from the “DR1-v161116-170912” release,
which comprises more than 49000 spectra divided into an irregular grid with a
1° seeing resolution and a 0.25° sampling resolution. The intensity of the
H$\alpha$ emission is recorded in a 3-dimensional array encoding its Galactic
longitude, Galactic latitude, and line-of-sight velocity ($-170$ km
s${}^{-1}<v_{\rm LSR}<160$ km s-1). We fit a Gaussian distribution to the flux
profile along individual lines of sight. Then we take the velocity at the peak
of the distribution as the centroid velocity and the uncertainty of the fit as
the uncertainty. We also correct for Galactic rotation following the procedure
in Qu et al. (2020).
Figure 1 shows the H$\alpha$ map of the four regions in our analysis, with
stars overplotted as foreground dots. The top panels show the integrated line-
of-sight fluxes in log scale. In the foreground, each star is color-coded by
their distance from Earth. We note that although Perseus and Taurus are very
close to each other on the sky, they are physically separated by $\sim 150$ pc
along the line of sight. In the bottom panels, we plot the centroid velocity
of the H$\alpha$ in the background, and the line-of-sight velocities of stars
in the foreground.
### 2.2 Data analysis
We compute the first-order VSF of stars in each region. The VSF is related to
the kinetic energy power spectrum of a velocity field, and is expressed as a
two-point correlation function between the absolute velocity differences
$|\delta v|$ versus the physical separation $\ell$. For a turbulent velocity
field, we expect the VSF to have a characteristic power-law scaling with slope
1/3 for incompressible turbulence dominated by solenoidal motions (Kolmogorov,
1941), or a slope 1/2 for supersonic turbulence dominated by compressive
motions. For the Milky Way molecular clouds, Larson (1981) found a
relationship between cloud size and velocity dispersion to be
$\sigma=1.1~{}R^{0.38}$.
The 3D VSF of stars is computed as follows. First, the positions, distances,
radial velocities, and proper motions of stars are translated into
galactocentric Cartesian coordinates and velocities. Then, for each pair of
stars, the absolute velocity vector difference $|\delta\vec{v}|$ and the
distance $d$ between them is calculated. For individual clouds, these
distances are sorted into logarithmic bins of $\ell$, and $|\delta\vec{v}|$
values within a bin are averaged out to obtain
$\langle|\delta\vec{v}|\rangle$. To estimate uncertainties, we first generate
1000 random samples of stellar parameters following a Gaussian distribution
around the reported measurements and with a 1$\sigma$ scatter equal to the
reported uncertainties. Then at each $\ell$, we take the 1$\sigma$ value of
the 1000 $\langle|\delta\vec{v}|\rangle$ to be the uncertainty.
For the WHAM H$\alpha$ data, we first define a rectangular box containing the
stars in each of the four regions as specified in Table 1 and shown in Figure
1. Our results are not sensitive to the exact size of the box (see Appendix C
for detailed discussions). We compute the first-order VSF of each region using
the projected position-position-velocity information. The resultant VSF is
$\langle|\delta v|\rangle_{\rm proj}$ versus angular separations $\theta$.
Then, we assume a constant distance equal to the mean distance of stars in
each region (see Table 1) and translate $\theta$ (°) to $\ell_{\rm proj}$
(pc). Since three degrees of freedom are missing in this projected VSF
calculation (the proper motions of gas and its exact distance), we scale
$\langle|\delta v|\rangle_{\rm proj}$ and $\ell_{\rm proj}$ by a constant
factor, assuming an isotropic turbulent velocity field.
$\langle|\delta\vec{v}|\rangle=\sqrt{3}~{}\langle|\delta v|\rangle_{\rm
proj},$ (1) $\ell=\sqrt{\frac{3}{2}}~{}\ell_{\rm proj}.$ (2)
## 3 Results
Figure 2: The first-order VSFs of the stars in Orion (light blue), Ophiuchus
(magenta), Perseus (green), and Taurus (orange). For reference, we plot a
solid blue line for Larson’s relation, a black dotted-dashed line for
Kolmogorov turbulence, and a pink dashed line for supersonic turbulence. The
vertical lines represent the uncertainties in $\langle|\delta\vec{v}|\rangle$
at each $\ell$.
### 3.1 The Universality of ISM Turbulence Traced by Young Stars
We first examine the velocity statistics of the four star-forming regions:
Orion, Ophiuchus, Perseus, and Taurus. Figure 2 shows the VSFs traced by stars
of these four regions. For reference, we also plot the Larson’s law of
velocity dispersion vs. cloud size, the 1/3 slope of Kolmogorov turbulence,
and the 1/2 slope of supersonic turbulence. Immediately, we recognize a
remarkable universality among the VSFs of different regions. Over a large
dynamical range, the magnitude and slopes of the VSFs show a general agreement
with Larson’s law (see Table 2 for the best-fit slopes of the VSFs). This
suggests that the motion of young stars in the Milky Way show characteristics
of turbulence, and supports our previous findings in Paper I that turbulence
in the ISM can be traced with young stars.
At $\ell\lesssim 10$ pc, the slopes of the VSF of all four regions are flatter
than Larson’s law, as well as Kolmogorov (slope 1/3) turbulence. We discuss
the possible cause of this flattening in Section 4.1.
### 3.2 The VSFs of Individual Regions
Figure 3: The first-order 3D VSF of stars (red) and H$\alpha$ (olive) for
Orion, Ophiuchus, Perseus, and Taurus. The shaded region around the H$\alpha$
VSF is its associated uncertainty. At small $\ell$, the dots showing the VSF
traced by H$\alpha$ are not connected to indicate the seeing limit of the WHAM
survey ($\theta<1$°). The slopes of the VSFs of stars and of H$\alpha$ are
specified in Table 2. The VSF of H$\alpha$ in Ophiuchus can also be fitted
with a broken power-law, with slope $0.58\pm 0.24$ for $3~{}\mathrm{pc}<l<9$
pc and slope $0.29\pm 0.08$ for $9~{}\mathrm{pc}<l<30$ pc. For comparison, we
also include CO (violet) analysis from Qian et al. (2015, 2018) for three of
these regions.
Figure 3 shows the VSFs of young stars as well as gas in each of the four
regions: Orion, Ophiuchus, Perseus, and Taurus. The best-fit slopes of these
VSFs are lists in Table 2.
#### 3.2.1 Orion
The first-order VSF of Orion traced by stars has been studied in detail in
Paper I. We only reiterate some of the main points here and address the
differences caused by the addition of $\sim 1000$ stars compared with Paper I.
Throughout the inertial range, the magnitude of
$\langle|\delta\vec{v}|\rangle$ is higher than Larson’s law. At small scales
($1~{}\mathrm{pc}\lesssim\ell\lesssim 20$ pc), the number statistics is
dominated by stars in the Orion Nebula Cluster (ONC), which flattens the VSF
(see Section 4.1 for more discussion on the small-scale flattening). At ranges
from $20$ pc to $\sim 70$ pc, Orion D, which is a loose stellar association at
the front of the Orion Complex along our line of sight, dominates the
statistics, and it exhibits a power-law scaling, which indicates an imprint of
turbulence from their natal molecular clouds. The VSF of Orion without
considering the ONC would be steeper at small scales, which renders it more
in-line with the VSF of other regions (see Figure 2). At $\sim 70$ pc, there
is a bump in the structure function, and we proposed that this is an evidence
supporting the existence of a SN explosion at the heart of Orion 6 Myr ago
(Kounkel, 2020). Paper I used Gaia DR2 only. The newly added stars in this
study from Gaia EDR3 have larger overall uncertainty in their parametric
solutions. Therefore, the noise floor for Orion at small scales is boosted
from $\sim 3$ km s-1 to $\sim 5$ km s-1. This slightly elevates the overall
VSF compared to Paper I.
H$\alpha$ emission in Orion mainly comes from the Barnard’s loop and the
$\lambda$ Ori bubble (see Figure 1), which are thought to be created by SN
explosions (Kounkel, 2020). We see the effect of such energy injection in the
VSF traced by H$\alpha$. At $\ell\gtrsim 20$ pc, the amplitude of the
H$\alpha$ VSF is higher than the VSF of young stars. Over the whole dynamical
range we probe here, the H$\alpha$ VSF has a roughly constant slope of 1/2.
The VSF traced by H$\alpha$ shows a bump at $\ell\sim 150$ pc. This is
indicative of a local energy injection, similar to those seen in the VSF
traced by young stars. However, we caution that this scale is very close to
the size of the H$\alpha$ box, and the bump can also be an artifact (see,
e.g., Mohapatra et al., 2022). We discuss this effect in more details in
Appendix C. At very small scales ($\theta=1^{\rm o}$, corresponding to
$\ell\sim 7$ pc), our analysis is limited by the resolution of the WHAM survey
(Haffner et al., 2003). We discuss the seeing effect and other uncertainties
related to H$\alpha$ VSF in Section 4.2.
#### 3.2.2 Ophiuchus
The young stars in Ophiuchus has a VSF similar to that of the Orion Complex.
At small $\ell$, the structure function is flatter than 1/3, and then steepens
to $\sim$ 1/3 beyond 10 pc up to $\sim 25$ pc. The H$\alpha$ VSF in Ophiuchus
has slope close to 1/3 and a higher amplitude than the VSF of the stars. Both
VSFs show clear bumps at $\sim 25$ pc, which we interpret as evidence of local
energy injection.
In Orion, such an energy injection is associated with a past SN explosion as
we discuss in detail in Paper I. If the energy injection in Ophiuchus is also
caused by a SN, we estimate that the event took place around
$25\,\mathrm{pc}/v_{\rm region}\sim 2.5$ Myr ago, where we use the
characteristic H$\alpha$ velocity $v_{\rm region}\sim 10$ km s-1. Such an
estimate is broadly consistent with observations of runaway stars which trace
back to a SN explosion in the region $\sim 2$ Myr ago (Neuhäuser et al.,
2020).
We note that the scale of the bump is close to the size of the region
containing the young stars in our analysis. Therefore, it is technically
possible that the bump in the VSF of young stars is only an artifact caused by
the limited sample size. In Appendix A, we analyze the entire Upper Sco
region. We find that as long as we apply our analysis to young stars ($<2$
Myr), the results are consistent, while the VSFs of older stars do not show
this feature. In addition, the SN scenario is also supported by the bump in
the H$\alpha$ VSF.
The cold and dense molecular cores, where stars are born out of, emits CO.
Qian et al. (2015, 2018) applied the core-velocity-dispersion (CVD; Qian et
al., 2012) method to analyze turbulence in star-forming regions. We translate
their best-fit CVD to a VSF of CO by
$\mathrm{VSF}=\sqrt{\frac{2}{\pi}}~{}\mathrm{CVD}$.
In Ophiuchus, the amplitude of CO VSF (pink solid lines in Figure 3) is much
lower than the VSF of young stars and H$\alpha$. It is possible that the
recent energy injection has affected the warm phase of the ISM more than the
cold phase (see discussion in Section 4.3). It is also possible that the low
amplitude is due to spacial variation of ISM turbulence (see discussion in
Appendix C). Qian et al. (2015) only examined the smaller scale velocity
structures ($\ell\lesssim 10$ pc), where there is no reliable component that
we can compare against (due to relaxation of stars and resolution limit of
H$\alpha$).
#### 3.2.3 Perseus
At length scales below 10 pc, the VSF of young stars in Perseus is very
similar to that of Ophiuchus (see Figure 2). Since we include two clusters in
our analysis of Perseus: NGC 1333 and IC 348, the VSF is flattened from
internal dynamical relaxation. The existence of subgroups and sub-clusters are
likely the cause of the small wiggles in the VSF. At larger scales, the VSF is
generally consistent with Larson’s law.
The VSF traced by H$\alpha$ here has an amplitude similar to the VSF traced by
young stars but a steeper slope ($\sim$ 1/2). VSF traced by CO matches the
H$\alpha$ VSF at $\ell>6$ pc, but flattens on smaller scales due to projection
effects for thick clouds (Qian et al., 2015; Xu, 2020).
#### 3.2.4 Taurus
Taurus is overall very similar to Perseus. The VSFs of young stars, H$\alpha$
and CO generally agree with each other, as well as Larson’s Law.
There is no prominent bump detected in the VSF in both Perseus and Taurus,
which is consistent with the findings of Krolikowski et al. (2021) that there
has been no SN activities in the last 5 Myr in these regions. However, there
are wiggles scattered throughout the length scales, likely because Taurus
contains several small dense groups that are primarily homogeneously
distributed throughout the region.
Table 2: Slopes of the VSFs
Region | Slope of | Fitting | Slope of | Fitting
---|---|---|---|---
| stars VSF | range (pc) | H$\alpha$ VSF | range (pc)
Orion | $0.25\pm 0.01$ | $12-75$ | $0.55\pm 0.02$ | $9-150$
Ophiuchus | $0.34\pm 0.03$ | $7-28$ | $0.36\pm 0.03$ | $3-30$
Perseus | $0.22\pm 0.03$ | $5-80$ | $0.61\pm 0.08$ | $7-55$
Taurus | $0.34\pm 0.03$ | $7-70$ | $0.42\pm 0.12$ | $3-25$
Note. — Inside the fitting ranges, the VSFs are fitted with an equation:
$\langle|\delta\vec{v}|\rangle$ $=A\cdot\ell^{B}$, where $B$ is the slope. The
uncertainties reported with each slope is obtained from the covariance matrix,
taking into account the uncertainties in $\langle|\delta\vec{v}|\rangle$ shown
in Figure 3. We choose the fitting ranges such that the uncertainties are low
and there is a consistent trend in the slope of the VSF.
## 4 Discussion
### 4.1 Uncertainties and biases in turbulence traced by young stars
We discussed various biases and uncertainties associated with measuring
turbulence from stars in Paper I, including dynamical relaxation, binary
contamination, and drifting. In this section, we reiterate the main points and
discuss additional mechanisms that can affect the shape of the VSF.
Stellar interactions can cause dynamical relaxation, which erases the memory
of turbulence and flattens the VSF. Similar to Orion’s loose groups discussed
in Paper I, the relaxation times of Ophiuchus, Perseus, and Taurus are
estimated to be tens of Myrs, much longer than their age (for a detailed
discussion of the age of each region, see Wilking et al., 2008; Krolikowski et
al., 2021; Pavlidou et al., 2021). Thus dynamical relaxation has not affected
the overall shape of the VSF for the three regions. However, there are some
small, compact star clusters in these regions, which can contribute to the
flattening at small scales ($\ell\lesssim 10$ pc, see Figure 2).
Another source of uncertainty originates from the drifting of stars along
their natural trajectories over time. Stars born with high velocity
differences drift apart at faster rates than stars born with low velocity
differences, which steepens the VSF roughly over the the crossing time
($t_{\rm cross}$) of the cloud. Similar to the star groups in Orion that we
discussed in Paper I, $t_{cross}$ of Ophiuchus, Perseus, and Taurus are all
larger than their current age, and thus this effect is unlikely to have
significantly altered the shape of the structure functions.
Low-velocity binary systems and contamination from field stars can act like
noise in our analysis, which also tend to flatten the VSF as is discussed in
Paper I.
Additionally, other stellar feedback processes such as winds, jets, and
radiation can also inject energy on small scales (e.g., Gritschneder et al.,
2009; Offner & Arce, 2015; Li et al., 2019; Rosen et al., 2021; Grudić et al.,
2022). This can contribute to the flattening of the VSFs at small scales,
provided that these stars are massive ($>8~{}M_{\odot}$) and are within $\sim
1$ pc of each other (Rosen et al., 2021).
Stutz & Gould (2016) quantified an ejection mechanism through which protostars
decouple from the molecular gas. They found that the undulation of star-
forming filaments in Orion A “slingshots” protostars out of the filament at
$\sim 2.5$ km s-1, which is also at the level where the VSF flattens in
Ophiuchus, Perseus, and Taurus.
All sources of uncertainties listed above can contribute to the flattening of
the VSF at small scales ($\ell\lesssim 10$ pc). Therefore, we only consider
young stars to be reliable tracers of ISM turbulence above $\sim 10$ pc.
### 4.2 ISM Turbulence traced by H$\alpha$
In Figure 2 and Figure 3, we find that while the VSFs of stars are broadly
consistent across the regions, the VSFs of H$\alpha$ are more diverse.
In regions with recent supernova explosions (Orion and Ophiuchus), the VSF of
H$\alpha$ gas has a higher amplitude than that of stars. Whereas in Perseus
and Taurus, the H$\alpha$ flux is lower (Figure 1) and the H$\alpha$ VSF is
more in line with the VSFs of stars and CO.
Figure 4: VSF of four regions traced by H$\alpha$. Solid lines denote the VSF
computed for pixels with low emission, and dotted-dashed lines denote the VSF
computed for pixels with high emission.
Within each individual region, we divide the pixels in half based on the
H$\alpha$ brightness– the high-emission half with all the pixels brighter than
the median and the low-emission half with all the pixels dimmer than the
median. Figure 4 shows the VSFs of high-emission H$\alpha$ (dotted-dashed
line) and low-emission H$\alpha$ (solid line) within each region. In general,
the VSF of high-emission H$\alpha$ has a higher amplitude compared to that of
the low-emission H$\alpha$.
One possible explanation for this is that bright H$\alpha$ comes from photo-
ionization by young massive stars (Tacchella et al., 2022). Therefore, it
traces the most intense stellar feedback regions and shows a higher VSF
amplitude. Whereas the low-emission H$\alpha$ comes from primarily collisional
ionization. Note that Orion has its high- and low-emission H$\alpha$ VSFs
essentially overlapping for the most part. This is likely because star
formation in Orion is the most intense among all the regions studied here (see
Table 2 of Lada et al., 2010), and thus most of the H$\alpha$ is photo-ionized
by massive stars.
There are many uncertainties and potential biases associated with the
H$\alpha$ analysis. One uncertainty comes from atmospheric seeing. Although
the WHAM data cubes provide us with H$\alpha$ line-of-sight measurements every
0.25°, the seeing limit of the survey is reported as 1°. At separations
$\theta\leq 1$° (or $\ell_{\rm proj}\leq d\cdot 1$°), we treat the VSF traced
by H$\alpha$ as unresolved (represented by the unconnected dots in Figure 3).
We note that seeing might not significantly alter the overall VSF, as
suggested by Li et al. (2020). Future studies with mock simulated H$\alpha$
observations may help properly evaluate the effects of seeing.
There are also uncertainties with the distance to the H$\alpha$ gas. The
bright H$\alpha$ emissions in Orion and Ophiuchus are likely associated with
the star forming regions. However, since Perseus and Taurus do not emit
intense H$\alpha$ (see top right panel of Figure 1), it is uncertain whether
the centroid velocities found in these two regions are at the exact same
distance as the stars.
Another bias can come from the projection effects. When analysing the VSF of
H$\alpha$, we assume that all the gas is located at the same distance,
effectively treating it as a 2D structure. This can flatten the VSF for thick
clouds, as is discussed in detail in Qian et al. (2015); Xu (2020); Mohapatra
et al. (2022). All the H$\alpha$ VSFs in our analysis is either as steep as
1/3 or steeper. This suggests that either most of the H$\alpha$ comes from a
thin sheet of gas, or the true VSF has a even steeper slope.
Another potential bias related to projection is that we are trying to infer 3D
turbulence using projected 2D information. This can introduce uncertainties as
the ISM is magnetized. In the presence of magnetic fields, turbulence is no
longer isotropic (Seifried & Walch, 2015; Hu et al., 2021).
We find that our results are not sensitive to the rotation model (see detailed
discussions in Appendix B).
### 4.3 The Complex Nature of ISM Turbulence
The ISM is multi-phase in nature. The kinematic coupling between these phases
is not well understood (for a review of the multi-phase ISM model, see Cox,
2005). More specifically, it is suggested that each phase possesses varying
levels of turbulence, where ionized gas has been observed to have higher
velocity dispersion than molecular gas in both high- and low-redshift galaxies
(Girard et al., 2021; Ejdetjärn et al., 2021). In our study, we observe a
similar trend comparing the VSF of H$\alpha$ and other tracers of turbulence.
In Orion and Ophiuchus, there is a large deviation between the level of
turbulence seen by different tracers. The VSFs traced by H$\alpha$ show the
largest amplitude at large scales, followed by the VSFs traced by stars, and
finally by CO. One possible explanation is the difference in the transport of
energy and momentum from SNe into different phases in these regions, as
mentioned in Section 3.2. In an inhomogeneous medium such as the Milky Way
ISM, SNe carry shocked matter radially outward following the path of least
resistance. As a consequence, the cool dense phase traced by CO receives less
energy injection than the warm diffuse phase traced by H$\alpha$, resulting in
lower levels of turbulence in the corresponding VSF. Such behavior is also
seen in hydrodynamical simulations of single SN remnants (e.g., Martizzi et
al., 2015; Li et al., 2015).
In Perseus and Taurus, the VSFs traced by H$\alpha$ and CO show a good
agreement in both the amplitudes and slopes in these regions, except for the
range $\ell\lesssim 7$ pc in Perseus, which suffers from projection effect due
to the thickness of the cloud (see Section 4.2 of Qian et al., 2015, for a
detailed discussion). At larger $\ell$, the VSF traced by H$\alpha$ and stars
also closely follow each other. Such trend is consistent with the evolutionary
history of these regions with no clear imprint of recent SN explosions.
Without this important local driver of turbulence, different phases of gas are
well-coupled, and the young stars formed out of these molecular clouds also
largely retain the same memory of turbulence.
The difference of H$\alpha$ VSFs in different regions show that ISM turbulence
can be unevenly distributed spatially as a result of local drivers. Turbulence
can also be unevenly distributed between different phases of the ISM (e.g.,
Orion and Ophiuchus). Section 4.2 and Figure 4 further demonstrate that even
with the same gas tracer H$\alpha$, the level of turbulence can vary depending
on its brightness. In addition, we note that local energy injection from
supernova explosions is also intermittent temporally. Using the H$\alpha$ VSF
in Ophiuchus, we estimate the eddy turnover time at the driving scale to be:
$\frac{\ell}{v}\sim\frac{20-30\,\rm{pc}}{5\,\rm{km/s}}\sim 4-6\,\rm{Myr}.$ (3)
The H$\alpha$ VSF in Orion gives a similar estimate. Typically, it takes a few
eddy turnover times to establish a steady-state turbulent flow. On the other
hand, given the size of these regions, the average time interval between
supernovae is on the order of a few Myr (Narayan & Ostriker, 1990; Li et al.,
2015). Each event also injects energy as a pulse with a very short duration.
Therefore, both the “on” and “off” states of the driver are short compared
with the timescale required to achieve a steady state. Thus ISM turbulence in
molecular complexes is often not in a steady state.
## 5 Conclusions
In this work, we analyze the kinematics of young stars and gas in four nearby
star-forming regions: Orion, Ophiuchus, Perseus, and Taurus. For each region,
we compute the first-order VSF of young stars and H$\alpha$ gas, and compare
the results against each other, as well as analysis of CO molecular cores from
the literature.
We find that the VSFs traced by young stars in all four regions exhibit
similar properties largely in agreement with Larson’s law and reveal a
universal scaling of turbulence in the ISM. This also confirms our previous
finding in Ha et al. (2021) that young stars retain the memory of turbulence
in their natal molecular clouds, and thus can be used to measure the turbulent
kinematics of molecular clouds.
The VSFs of H$\alpha$ are more diverse, likely as a result of local energy
injection. In regions with recent SN activities such as Orion and Ophiuchus,
the VSFs of H$\alpha$ are higher in amplitude compared to those without. The
H$\alpha$ VSFs are also higher than the VSFs of stars and CO in the these
regions, while regions without recent SN explosions (Perseus and Taurus) show
well-coupled turbulence between different phases. Within each individual
region, the high-emission H$\alpha$ tends to show higher VSFs than the low-
emission H$\alpha$.
Our findings support a complex picture of the Milky Way ISM, where turbulence
can be driven at difference scales and inject energy unevenly into different
phases. Future observations with more tracers of turbulence and on more star
forming regions will help us better understand the diversity and complexity of
ISM turbulence. In addition, high-resolution numerical studies can help
further our understanding of the coupling between different phases and
relevant timescales.
## Acknowledgments
We thank Shmuel Bialy, John Forbes, Di Li, Mordecai-Mark Mac Low, Lei Qian,
Amiel Sternberg, and Yuan-Sen Ting for the helpful discussions. We thank
Zhijie Qu for sharing the Galactic rotation model. H.L. and S.X. are supported
by NASA through the NASA Hubble Fellowship grant HST-HF2-51438.001-A and HST-
HF2-51473.001-A, respectively, awarded by the Space Telescope Science
Institute, which is operated by the Association of Universities for Research
in Astronomy, Incorporated, under NASA contract NAS5-26555. Y.L. acknowledges
financial support from NSF grant AST-2107735 and the College of Science and
College of Engineering at UNT. Y.Z. is supported by NASA through a grant (HST-
AR-16640.001-A) from the Space Telescope Science Institute, which is operated
by the Association of Universities for Research in Astronomy, Incorporated,
under NASA contract NAS5-26555. The Wisconsin H$\alpha$ Mapper and its
H$\alpha$ Sky Survey have been funded primarily by the National Science
Foundation. The facility was designed and built with the help of the
University of Wisconsin Graduate School, Physical Sciences Lab, and Space
Astronomy Lab. NOAO staff at Kitt Peak and Cerro Tololo provided on-site
support for its remote operation.
## References
* Abdurro’uf et al. (2022) Abdurro’uf, Accetta, K., Aerts, C., et al. 2022, ApJS, 259, 35, doi: 10.3847/1538-4365/ac4414
* Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068
* Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123, doi: 10.3847/1538-3881/aabc4f
* Blanton et al. (2017) Blanton, M. R., Bershady, M. A., Abolfathi, B., et al. 2017, AJ, 154, 28, doi: 10.3847/1538-3881/aa7567
* Burkhart et al. (2010) Burkhart, B., Stanimirović, S., Lazarian, A., & Kowal, G. 2010, ApJ, 708, 1204, doi: 10.1088/0004-637X/708/2/1204
* Castaneda (1988) Castaneda, H. O. 1988, ApJS, 67, 93, doi: 10.1086/191268
* Chepurnov & Lazarian (2010) Chepurnov, A., & Lazarian, A. 2010, ApJ, 710, 853, doi: 10.1088/0004-637X/710/1/853
* Chepurnov et al. (2010) Chepurnov, A., Lazarian, A., Stanimirović, S., Heiles, C., & Peek, J. E. G. 2010, ApJ, 714, 1398, doi: 10.1088/0004-637X/714/2/1398
* Cox (2005) Cox, D. P. 2005, ARA&A, 43, 337, doi: 10.1146/annurev.astro.43.072103.150615
* Ejdetjärn et al. (2021) Ejdetjärn, T., Agertz, O., Östlin, G., Renaud, F., & Romeo, A. B. 2021, arXiv e-prints, arXiv:2111.09322. https://arxiv.org/abs/2111.09322
* Federrath (2013) Federrath, C. 2013, MNRAS, 436, 1245, doi: 10.1093/mnras/stt1644
* Gaia Collaboration et al. (2016) Gaia Collaboration, Prusti, T., de Bruijne, J. H. J., et al. 2016, A&A, 595, A1, doi: 10.1051/0004-6361/201629272
* Gaia Collaboration et al. (2021) Gaia Collaboration, Brown, A. G. A., Vallenari, A., et al. 2021, A&A, 649, A1, doi: 10.1051/0004-6361/202039657
* Galtier & Banerjee (2011) Galtier, S., & Banerjee, S. 2011, Phys. Rev. Lett., 107, 134501, doi: 10.1103/PhysRevLett.107.134501
* Girard et al. (2021) Girard, M., Fisher, D. B., Bolatto, A. D., et al. 2021, ApJ, 909, 12, doi: 10.3847/1538-4357/abd5b9
* Gritschneder et al. (2009) Gritschneder, M., Naab, T., Walch, S., Burkert, A., & Heitsch, F. 2009, ApJ, 694, L26, doi: 10.1088/0004-637X/694/1/L26
* Grudić et al. (2022) Grudić, M. Y., Guszejnov, D., Offner, S. S. R., et al. 2022, MNRAS, 512, 216, doi: 10.1093/mnras/stac526
* Gunn et al. (2006) Gunn, J. E., Siegmund, W. A., Mannery, E. J., et al. 2006, AJ, 131, 2332, doi: 10.1086/500975
* Ha et al. (2021) Ha, T., Li, Y., Xu, S., Kounkel, M., & Li, H. 2021, ApJ, 907, L40, doi: 10.3847/2041-8213/abd8c9
* Haffner et al. (2003) Haffner, L. M., Reynolds, R. J., Tufte, S. L., et al. 2003, ApJS, 149, 405, doi: 10.1086/378850
* Haffner et al. (2010) Haffner, L. M., Reynolds, R. J., Madsen, G. J., et al. 2010, in Astronomical Society of the Pacific Conference Series, Vol. 438, The Dynamic Interstellar Medium: A Celebration of the Canadian Galactic Plane Survey, ed. R. Kothes, T. L. Landecker, & A. G. Willis, 388. https://arxiv.org/abs/1008.0612
* Harris et al. (2020) Harris, C. R., Millman, K. J., van der Walt, S. J., et al. 2020, Nature, 585, 357, doi: 10.1038/s41586-020-2649-2
* Hennebelle & Falgarone (2012) Hennebelle, P., & Falgarone, E. 2012, A&A Rev., 20, 55, doi: 10.1007/s00159-012-0055-y
* Heyer & Brunt (2004) Heyer, M. H., & Brunt, C. M. 2004, ApJ, 615, L45, doi: 10.1086/425978
* Hu et al. (2022) Hu, Y., Federrath, C., Xu, S., & Mathew, S. S. 2022, MNRAS, doi: 10.1093/mnras/stac972
* Hu et al. (2021) Hu, Y., Lazarian, A., & Xu, S. 2021, ApJ, 915, 67, doi: 10.3847/1538-4357/ac00ab
* Hunter (2007) Hunter, J. D. 2007, Computing in Science Engineering, 9, 90, doi: 10.1109/MCSE.2007.55
* Joncas & Roy (1986) Joncas, G., & Roy, J. R. 1986, ApJ, 307, 649, doi: 10.1086/164450
* Kennicutt & Evans (2012) Kennicutt, R. C., & Evans, N. J. 2012, ARA&A, 50, 531, doi: 10.1146/annurev-astro-081811-125610
* Kim & Ostriker (2015) Kim, C.-G., & Ostriker, E. C. 2015, ApJ, 802, 99, doi: 10.1088/0004-637X/802/2/99
* Kolmogorov (1941) Kolmogorov, A. 1941, Akademiia Nauk SSSR Doklady, 30, 301
* Kounkel (2020) Kounkel, M. 2020, The Astrophysical Journal, 902, 122, doi: 10.3847/1538-4357/abb6e8
* Kounkel & Covey (2019) Kounkel, M., & Covey, K. 2019, AJ, 158, 122, doi: 10.3847/1538-3881/ab339a
* Krolikowski et al. (2021) Krolikowski, D. M., Kraus, A. L., & Rizzuto, A. C. 2021, AJ, 162, 110, doi: 10.3847/1538-3881/ac0632
* Lada et al. (2010) Lada, C. J., Lombardi, M., & Alves, J. F. 2010, ApJ, 724, 687, doi: 10.1088/0004-637X/724/1/687
* Larson (1981) Larson, R. B. 1981, MNRAS, 194, 809, doi: 10.1093/mnras/194.4.809
* Lazarian (2009) Lazarian, A. 2009, Space Science Reviews, 143, 357, doi: 10.1007/s11214-008-9460-y
* Lazarian & Pogosyan (2000) Lazarian, A., & Pogosyan, D. 2000, ApJ, 537, 720, doi: 10.1086/309040
* Lazarian & Pogosyan (2006) —. 2006, ApJ, 652, 1348, doi: 10.1086/508012
* Li et al. (2019) Li, H., Vogelsberger, M., Marinacci, F., & Gnedin, O. Y. 2019, MNRAS, 487, 364, doi: 10.1093/mnras/stz1271
* Li et al. (2015) Li, M., Ostriker, J. P., Cen, R., Bryan, G. L., & Naab, T. 2015, ApJ, 814, 4, doi: 10.1088/0004-637X/814/1/4
* Li et al. (2020) Li, Y., Gendron-Marsolais, M.-L., Zhuravleva, I., et al. 2020, ApJ, 889, L1, doi: 10.3847/2041-8213/ab65c7
* Louise & Monnet (1970) Louise, R., & Monnet, G. 1970, A&A, 8, 486
* Mac Low & Klessen (2004) Mac Low, M.-M., & Klessen, R. S. 2004, Reviews of Modern Physics, 76, 125, doi: 10.1103/RevModPhys.76.125
* Martizzi et al. (2015) Martizzi, D., Faucher-Giguère, C.-A., & Quataert, E. 2015, MNRAS, 450, 504, doi: 10.1093/mnras/stv562
* McBride et al. (2021) McBride, A., Lingg, R., Kounkel, M., Covey, K., & Hutchinson, B. 2021, AJ, 162, 282, doi: 10.3847/1538-3881/ac2432
* McInnes et al. (2017) McInnes, L., Healy, J., & Astels, S. 2017, The Journal of Open Source Software, 2, 205, doi: 10.21105/joss.00205
* Mohapatra et al. (2022) Mohapatra, R., Jetti, M., Sharma, P., & Federrath, C. 2022, MNRAS, 510, 2327, doi: 10.1093/mnras/stab3429
* Münch (1958) Münch, G. 1958, Reviews of Modern Physics, 30, 1035, doi: 10.1103/RevModPhys.30.1035
* Narayan & Ostriker (1990) Narayan, R., & Ostriker, J. P. 1990, ApJ, 352, 222, doi: 10.1086/168529
* Neuhäuser et al. (2020) Neuhäuser, R., Gießler, F., & Hambaryan, V. V. 2020, MNRAS, 498, 899, doi: 10.1093/mnras/stz2629
* Offner & Arce (2015) Offner, S. S. R., & Arce, H. G. 2015, ApJ, 811, 146, doi: 10.1088/0004-637X/811/2/146
* Ossenkopf & Mac Low (2002) Ossenkopf, V., & Mac Low, M. M. 2002, A&A, 390, 307, doi: 10.1051/0004-6361:20020629
* Padoan et al. (2016) Padoan, P., Pan, L., Haugbølle, T., & Nordlund, Å. 2016, ApJ, 822, 11, doi: 10.3847/0004-637X/822/1/11
* pandas development team (2020) pandas development team, T. 2020, pandas-dev/pandas: Pandas, latest, Zenodo, doi: 10.5281/zenodo.3509134
* Pavlidou et al. (2021) Pavlidou, T., Scholz, A., & Teixeira, P. S. 2021, MNRAS, 503, 3232, doi: 10.1093/mnras/stab352
* Qian et al. (2018) Qian, L., Li, D., Gao, Y., Xu, H., & Pan, Z. 2018, ApJ, 864, 116, doi: 10.3847/1538-4357/aad780
* Qian et al. (2012) Qian, L., Li, D., & Goldsmith, P. F. 2012, ApJ, 760, 147, doi: 10.1088/0004-637X/760/2/147
* Qian et al. (2015) Qian, L., Li, D., Offner, S., & Pan, Z. 2015, ApJ, 811, 71, doi: 10.1088/0004-637X/811/1/71
* Qu et al. (2020) Qu, Z., Bregman, J. N., Hodges-Kluck, E., Li, J.-T., & Lindley, R. 2020, ApJ, 894, 142, doi: 10.3847/1538-4357/ab774e
* Rathjen et al. (2021) Rathjen, T.-E., Naab, T., Girichidis, P., et al. 2021, MNRAS, 504, 1039, doi: 10.1093/mnras/stab900
* Roman-Duval et al. (2011) Roman-Duval, J., Federrath, C., Brunt, C., et al. 2011, ApJ, 740, 120, doi: 10.1088/0004-637X/740/2/120
* Rosen et al. (2021) Rosen, A. L., Offner, S. S. R., Foley, M. M., & Lopez, L. A. 2021, arXiv e-prints, arXiv:2107.12397. https://arxiv.org/abs/2107.12397
* Roy & Joncas (1985) Roy, J. R., & Joncas, G. 1985, ApJ, 288, 142, doi: 10.1086/162772
* Seifried & Walch (2015) Seifried, D., & Walch, S. 2015, MNRAS, 452, 2410, doi: 10.1093/mnras/stv1458
* Spitzer (1978) Spitzer, L. 1978, Physical processes in the interstellar medium, doi: 10.1002/9783527617722
* Stutz & Gould (2016) Stutz, A. M., & Gould, A. 2016, A&A, 590, A2, doi: 10.1051/0004-6361/201527979
* Stutzki et al. (1998) Stutzki, J., Bensch, F., Heithausen, A., Ossenkopf, V., & Zielinsky, M. 1998, A&A, 336, 697
* Tacchella et al. (2022) Tacchella, S., Smith, A., Kannan, R., et al. 2022, MNRAS, doi: 10.1093/mnras/stac818
* van der Walt et al. (2011) van der Walt, S., Colbert, S. C., & Varoquaux, G. 2011, Computing in Science Engineering, 13, 22, doi: 10.1109/MCSE.2011.37
* Virtanen et al. (2020) Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2020, Nature Methods, 17, 261, doi: 10.1038/s41592-019-0686-2
* von Hoerner (1951) von Hoerner, S. 1951, ZAp, 30, 17
* Wes McKinney (2010) Wes McKinney. 2010, in Proceedings of the 9th Python in Science Conference, ed. Stéfan van der Walt & Jarrod Millman, 56 – 61, doi: 10.25080/Majora-92bf1922-00a
* Wilking et al. (2008) Wilking, B. A., Gagné, M., & Allen, L. E. 2008, Star Formation in the $\rho$ Ophiuchi Molecular Cloud, ed. B. Reipurth, Vol. 5, 351
* Xu (2020) Xu, S. 2020, MNRAS, 492, 1044, doi: 10.1093/mnras/stz3092
* Xu & Lazarian (2022) Xu, S., & Lazarian, A. 2022, ApJ, 927, 94, doi: 10.3847/1538-4357/ac4dfd
* Zhou et al. (2022) Zhou, J.-X., Li, G.-X., & Chen, B.-Q. 2022, MNRAS, doi: 10.1093/mnras/stac900
## Appendix A Selection of the $\rho$ Oph dark cloud within the Upper Sco
region
Figure A.1: Top panel: positions of young stars color-coded by their age.
Stellar age is estimated using the Sagitta package developed by McBride et al.
(2021). Background shows H$\alpha$ flux from the WHAM survey. Bottom panel:
the VSF of stars in Upper Sco within three age brackets. Only stars younger
than 2 Myr show characteristic of SN energy injection.
In the main body of this paper, we select only stars within the actively star-
forming region Ophiuchus. Here, we show the analysis results when selecting
the entire Upper Sco region and discuss how the VSF of the stars depend on the
stellar age.
The top panel of Figure A.1 shows the spatial distribution of stars in Upper
Sco and their age. Most of the stars outside Ophiuchus are older than 4 Myr,
while stars inside Ophiuchus are much younger, primarily younger than 2 Myr.
The SN explosion in this region traced by runaway stars took place $\sim 2$
Myr ago (Neuhäuser et al., 2020), so the majority of stars in Upper Sco did
not receive energy injection from this event, and thus would not exhibit
characteristics of such energy injection in their VSF. Indeed, we see this
trend in the lower panel of Figure A.1. With the entire Upper Sco region taken
into account, the VSF shows only a slight bump at $\sim 25$ pc, while both the
VSF of stars older than 5 Myr and of stars between 2 and 5 Myr do not show any
discernible energy injection scale. Meanwhile, the VSF of stars younger than 2
Myr shows a clear energy injection scale at $\sim 25$ pc, similar to what we
obtain in the main paper. Because of these results, we elect to use only stars
categorized as members of Ophiuchus a majority of which are younger than 1
Myr, to analyze in the main body of this paper.
## Appendix B Effect of Galactic Rotation on the VSF of H$\alpha$
Figure B.2: The VSFs of H$\alpha$, with (solid lines) and without (dotted-
dashed lines) a correction to the Galactic rotation (Qu et al., 2020).
Figure B.3: Top panel: illustration of our H$\alpha$ window tests. The red box
is the smallest window within which all stars in Taurus belong. The two boxes
in green (the window used in the main analysis) and light-blue are each 4 deg2
bigger than the former. Inside the red box, we also test the variation of the
VSF within the region by dividing the window into four equal-sized boxes.
Bottom left panel: The VSF of Taurus traced by H$\alpha$, plotted across three
different window sizes. Bottom right panel: The H$\alpha$ VSFs of Taurus in
the four quadrants within the red window.
Figure B.2 shows the VSF traced by H$\alpha$ in four regions with and without
Galactic rotation correction. In three out of the four regions, when a
rotation profile is applied, the VSF does not significantly differ from the
VSF without a rotation correction up to $\sim 20$ pc. This behavior is
expected, since at small $\ell$ the rotation correction between two points are
very similar. Thus, $\delta\vec{v}$ between two points essentially cancel out
the rotation correction, and the shape of the VSF is preserved. For two points
separated by a large distance, the rotation correction applied to them can
differ more, and hence results in a shift in the VSF, as Figure B.2 shows.
However, the overall shifts in the VSFs due to rotation correction are
moderate and do not affect the conclusions of our work.
## Appendix C Effect of the size and location of the H$\alpha$ window
Another potential source of bias comes from the exact size and location of the
window chosen to calculate the VSF of H$\alpha$ within each region. In Section
3.2, we caution that the bumps seen in the VSF of H$\alpha$ is a potential
artifact since it is close to the size of the window. Such effect is noted as
the apparent driving scale in simulations of multi-phase turbulence, and is
equal to 1/2 of the box size (e.g., Mohapatra et al., 2022). To test whether
this effect also presents in the VSF of H$\alpha$, we select a solid angle
that encompasses all the stars in a region, and compute the VSF within this
window. We then compute the VSF for a window that is 4 deg2 larger, and repeat
this a second time (see top panel in Figure B.3).
In all four regions, the VSFs of H$\alpha$ are independent of the window size
up to $\ell\sim L/2$, where $L$ is the window size. In Taurus and Perseus, the
bump in the VSF shifts to larger $\ell$ when the window size increases. Thus
we conclude that in these regions, the VSF bump is likely caused by the box
limit rather than a real physical driving scale. As an example, the bottom
left panel of Figure B.3 shows the VSF of H$\alpha$ in Taurus as a function of
window size.
We also use Taurus as an example to illustrate the spatial variation of
turbulence on small scales. We divide the smallest box (8 deg2) in the
previous experiment into four equal-sized quadrants. The bottom right panel of
Figure B.3 shows the H$\alpha$ VSFs of these four quadrants in Taurus.
Although the slopes of these four VSFs are roughly consistent with one
another, their amplitudes vary by up to a factor of three, indicating
potentially large variations in turbulence between local patches of the ISM.
The VSF of CO is in the best agreement with the bottom-left window
($172^{\circ}<l<176^{\circ}$ and $-20^{\circ}<b<-16^{\circ}$), where many CO
cores were also identified (see Figure 2 of Qian et al., 2012, for detailed
locations of CO cores in Taurus).
|
# Toward Open Integrated Access and Backhaul
with O-RAN
Eugenio Moro1, Gabriele Gemmi2, Michele Polese3, Leonardo Maccari2, Antonio
Capone1, Tommaso Melodia3
This work was partially supported by NGIAtlantic.eu project within the
EUHorizon 2020 programme under Grant No. 871582, by the U.S. National Science
Foundation under Grant CNS-1925601, and by OUSD(R&E) through Army Research
Laboratory Cooperative Agreement Number W911NF-19-2-0221. The views and
conclusions contained in this document are those of the authors and should not
be interpreted as representing the official policies, either expressed or
implied, of the Army Research Laboratory or the U.S. Government. The U.S.
Government is authorized to reproduce and distribute reprints for Government
purposes notwithstanding any copyright notation herein. 1Department of
Electronics, Information and Bioengineering,
Polytechnic University of Milan, Italy
<EMAIL_ADDRESS>2Department of Environmental Sciences, Informatics
and Statistics,
Ca’ Foscari University of Venice, Italy.
<EMAIL_ADDRESS>3Institute for the Wireless Internet of Things,
Northeastern University, Boston, MA, U.S.A.
<EMAIL_ADDRESS>
###### Abstract
Millimeter wave (mmWave) communications has been recently standardized for use
in the fifth generation (5G) of cellular networks, fulfilling the promise of
multi-gigabit mobile throughput of current and future mobile radio network
generations. In this context, the network densification required to overcome
the difficult mmWave propagation will result in increased deployment costs.
Integrated Access and Backhaul (IAB) has been proposed as an effective mean of
reducing densification costs by deploying a wireless mesh network of base
stations, where backhaul and access transmissions share the same radio
technology. However, IAB requires sophisticated control mechanisms to operate
efficiently and address the increased complexity. The Open RAN paradigm
represents the ideal enabler of RAN intelligent control, but its current
specifications are not compatible with IAB. In this work, we discuss the
challenges of integrating IAB into the Open RAN ecosystem, detailing the
required architectural extensions that will enable dynamic control of 5G IAB
networks. We implement the proposed integrated architecture into the first
publicly-available Open-RAN-enabled experimental framework, which allows
prototyping and testing Open-RAN-based solutions over end-to-end 5G IAB
networks. Finally, we validate the framework with both ideal and realistic
deployment scenarios exploiting the large-scale testing capabilities of
publicly available experimental platforms.
###### Index Terms:
IAB, O-RAN, 5G, Colosseum
## I Introduction
Radio Access Network (RAN) densification is a key technique to boost the
coverage and performance metrics of current and future generations of mobile
radio networks [1]. However, these ultra-dense deployments come with increased
costs and complexity for provisioning wired backhaul to each base station [2].
To address this, the 3rd Generation Partnership Project (3GPP) has introduced
Integrated Access and Backhaul (IAB) in its Release 16 for NR [3]. With IAB,
the backhaul traffic is multiplexed on the air interface together with regular
User Equipments (UEs) access traffic. This effectively creates a wireless mesh
network of Base Stations (BSs) where only a few require an expensive wired
connection to the Core Network (CN) (i.e., the IAB-Donors). Hence the cost-
reduction potential through wireless relays (i.e., the IAB-Nodes) [4].
Additionally, IAB is especially relevant for millimeter wave (mmWave)-based
radio access, where inexpensive network densification is a fundamental
necessity [5].
While the standardization process has reached a sufficient maturity level, the
open challenges brought about by integrating access and backhaul are still
open. Consequently, IAB offers optimization opportunities at all layers of
communication abstraction. At the lowest levels, specialized IAB-aware
techniques are required to ensure a fair and effective resource allocation
among UEs and Mobile Terminations (MTs) [6, 7]. At the same time, backhaul and
access transmission multiplexing must be managed to minimize interference [8].
Furthermore, adaptive topology reconfiguration mechanisms must be provisioned
to maintain resiliency against link failures, traffic unbalances and anomalous
user distribution [9]. Overall, these sophisticated management procedures
require control primitives that go beyond what has been specified by 3GPP.
The unprecedented paradigm shift brought about by the O-RAN architecture,
developed by the O-RAN Alliance, promises to enable programmatic control of
RAN components through open interfaces and centralized control loops [10]. As
such, it is the ideal candidate to unlock the potential optimization and
management gains awaiting in IAB. However, the current O-RAN architecture is
tailored to traditional RAN deployments, and an extension to enable IAB
control is required. The first contribution of this work resides in a
discussion on how the O-RAN architecture, interfaces, and control loops can be
extended to IAB scenarios, with the ultimate goal of allowing large-scale,
data-driven control and management of 5th generation (5G) IAB networks.
Additionally, to foster prototyping and testing with IAB and O-RAN, we propose
a comprehensive framework where researchers can easily deploy an end-to-end
O-RAN-enabled IAB network with Over-The-Air (OTA) and hardware-in-the-loop
emulation capabilities. In line with O-RAN core concepts, our framework is
designed to be open, accessible and flexible by leveraging on open-source
software and Commercial Off-the-Shelf (COTS) hardware. The framework builds on
IABEST, the first large-scale accessible and open IAB testbed presented in
[11]. This testbed has been enriched to produce a complete O-RAN IAB
experimental solution, effectively replicating the proposed O-RAN-IAB
integrated architecture. In particular, IAB-Donors and IAB-Nodes have been
equipped with custom-developed agents for the so-called E2 and O1 standard
interfaces. These additions enable the controllers introduced by the O-RAN
architecture to manage IAB-Nodes, effectively representing the first publicly
available O-RAN-enabled IAB prototyping and testing solution.
To further facilitate experimental research activities, we have packaged and
integrated the entire framework into OpenRAN Gym, a publicly-available
research platform for data-driven O-RAN experimentation at scale [12]. Through
OpenRAN Gym, researchers can swiftly deploy and test the proposed framework
over large-scale and publicly available hardware experimental platforms, such
as the PAWR testbeds and Colosseum [13, 14]. Notably, we showcase how
Colosseum can be leveraged for large-scale IAB testing through hardware-in-
the-loop channel emulation to create sophisticated deployment scenarios. A
tutorial on how to deploy an O-RAN-driven IAB network, together with the
source code of all the framework, is available on the OpenRAN Gym
website.111https://openrangym.com/tutorials/iab-tutorial Finally, we use
Colosseum to validate the proposed framework numerically. In particular, we
test the attainable performance in a controlled radio scenario and in a more
realistic deployment in which we reconstruct a part of Florence, Italy.
The remainder of this paper is organized as follows. Section II analyses the
challenges of extending O-RAN to 5G IAB networks. Section III contains a
description of the proposed frameworks, focusing on the O-RAN extensions that
have been included in [11]. Section IV contains the results of the experiments
we performed to validate our framework by exploiting the large-scale testing
capabilities of Colosseum. Finally, Section V concludes the paper and
discusses future extensions.
Figure 1: IAB and O-RAN integrated architectures.
## II Integrating IAB in Open RAN
As discussed in Section I, IAB represents a scalable solution to the need for
backhaul in ultra-dense 5G and 6G deployments. At the same time, however, the
wireless backhaul introduces additional complexity to the network deployments:
new parameters and configurations that need to be tuned—and possibly, adapted
dynamically—to get the best performance out of the network and to seamlessly
adjust to updated conditions in the scenario and in the equipment status. For
example, it is possible to optimize the IAB network performance by properly
selecting the connectivity of IAB-Nodes to their parents [9], or by
appropriately allocating resources to backhaul and access flows sharing the
same air interface [6].
As for traditional RAN deployments with fiber-based backhaul [15], there is a
case to be made for providing IAB RAN equipment with primitives for flexible,
dynamic, data-driven programmatic control. This requires providing endpoints
to expose telemetry, measurements, and analytics from IAB-Nodes, as well as
parameters and control knobs to enable the optimization. So far, the Open RAN
paradigm has been successfully applied to non-IAB networks to achieve the same
goals, thanks to interfaces that give access to 3GPP Key Performance
Measurements (KPMs) and control parameters in the RAN nodes [16, 17]. The Open
RAN vision, which is being developed into technical specifications by the
O-RAN Alliance, includes controllers that run custom control loops, i.e., the
RAN Intelligent Controllers (RICs). The O-RAN Alliance has defined control
loops and related RICs that can operate at a time scale of 10 ms to 1 s (i.e.,
_near-real-time_) or more than 1 s (i.e., _non-real-time_) [18]. The near-
real-time, or near-RT, RIC is connected to the RAN nodes through the E2
interface, while the non-real-time RIC, which is part of the network Service
Management and Orchestration (SMO), interacts with the RAN through the O1
interface, as shown in the left part of Figure 1. Other interfaces from the
non-RT RIC/SMO include A1 to the near-RT RIC, for policy guidance and
Artificial Intelligence (AI)/Machine Learning (ML) model management, and the
O2 interface to the O-Cloud, which is an abstraction of the virtualization
infrastructure that can support the deployment of O-RAN functions. The use of
standard interfaces makes it possible to run even third-party applications in
the controllers, the so-called xApps and rApps for the Near Real-time RAN
Intelligent Controller (near-RT RIC) and Non-Real-Time Ran Intelligent
Controller (non-RT RIC), respectively.
The 3GPP already provides control and adaptation capabilities through the IAB
Backhaul Adaptation Protocol (BAP) layer, the F1 interface, and the Radio
Resource Control (RRC) layer across the IAB-Donor Central Unit (CU) and the
IAB-Node Distributed Unit (DU). How and when control and adaptation of such
configurations could be performed, however, is left to the vendor
implementation. This is where an extension of the O-RAN architecture to IAB
networks can play a role, exposing IAB-Donor and IAB-Node functions to the
RICs. These can leverage a centralized point of view on the RAN and a wealth
of analytics and information usually unavailable in the individual IAB-Donors
and Nodes. For IAB, this could translate into effective multi-donor
coordination with reduced interference and agile topology adaptation across
different IAB-Donor domains, and dynamic resource allocation with—for
example—data-driven proactive congestion identification and resolution across
access and backhaul links.
### II-A Extensions to Open RAN
Extending the O-RAN architecture and interfaces to IAB deployments, however,
presents some design and architectural challenges. Primarily, supporting O-RAN
interfaces in IAB-Nodes means either (i) terminating the interfaces at the
IAB-Donor; or (ii) transporting their data over the wireless backhaul. The
first option is simpler, does not require architectural updates, but at the
same time limits the control and reconfiguration to what is available in the
IAB-Donor, without insight on the IAB-Nodes. The second option, instead,
provides more granular access at the cost of additional complexity and
tunneling of data over the wireless backhaul.
The 3GPP already foresees performing SMO-like operations through the wireless
backhaul interface [19]. Therefore, in this paper and in the architecture
described in Figure 1 we consider the second option, which would provide
tighter and more effective integration between O-RAN and IAB deployments. In
general, the tunneling can be performed by encapsulating the O-RAN interfaces
payloads into dedicated bearers. Note that this requires some interaction
between functions of the control plane of the network and the transport in the
user plane, e.g., through a dedicated Packet Data Unit (PDU) session between a
local User Plane Function (UPF) in the IAB-Donor and in the IAB-Node MT. Then,
a local interface termination can be installed in the IAB-Node, as it would in
a traditional, fiber-equipped RAN node. The O-RAN traffic, in this case, would
be multiplexed with user data on the wireless backhaul resources, and it needs
to be properly prioritized to achieve the control goals while not harming
users’ performance or creating congestion.
Figure 2: Overview of the RAN architecture deployed over white-box hardware.
E2 extension for IAB. The extension of the E2 interface likely requires one or
multiple new, dedicated E2 Service Models (E2SMs). The E2SM represents the
semantic of the E2 interface, i.e., the RAN function with which an xApp in the
near-RT RIC interacts. For IAB, an extension of E2SM KPM [20] can be used to
expose performance metrics related to the MT, besides the DU. Another near-
real-time control target over E2 can include, for example, resource
partitioning between backhaul and access traffic, or dynamic Time Division
Duplexing (TDD) slot configuration to adapt to varying traffic on the access
and backhaul.
O1 extension for IAB. The O1 interface would connect the SMO to the IAB-Node,
e.g., to perform maintenance and updates of the components (MT and DU) of the
IAB-Node. Compared to E2 near-real-time control, the O1 interface would run
control loops at 1 s or more. Thus its traffic can be transported with lower
priority than the E2 traffic. This makes a case for dedicated bearers and
tunnels on the backhaul interface for _each_ of the O-RAN interfaces.
O2 extension for IAB. This interface can be used to integrate the IAB-Nodes as
resources in the O-Cloud. Compared to traditional virtualization
infrastructure for the O-Cloud, the IAB-Nodes are available—and reachable over
O2—only when a session is established from one IAB-Donor to the IAB-Node
itself.
## III An Experimental Framework for
IAB and O-RAN
Our proposed experimental framework packages the entire software chain
required to run the O-RAN-enabled IAB network described in Section II in a
multi-layer architecture. At the hardware level, our framework does not
present any specific requirement. Indeed, every software component can run on
COTS hardware like generic x86 machines and USRP Software-defined Radio (SDR).
On the other hand, some software components are customized or designed from
scratch to reproduce and support a 5G IAB network. In particular, we have
adapted OpenAirInterface (OAI), an open source 5G RAN framework [21], to
implement IAB-Donors, IAB-Nodes, and IAB-capable core functions. Additionally,
we have integrated agents for the E2 and O1 interfaces in the IAB-Donor and
IAB-Node, effectively implementing the architectural integration proposed in
Section II. These interfaces are used by the non-real-time and real-time RICs
packaged in our framework to control all the components of the deployed IAB
network. We now describe the aforementioned components, separating them into
the RAN and O-RAN domains.
### III-A RAN and Core Network Components
Figure 2 represents an overview of the radio access functional components that
enable end-to-end communication in our framework. In particular, we provide
the following: a minimal yet functional deployment of 5G CN functions,
software-defined IAB-Nodes and IAB-Donors and software-defined UEs.
IAB-Nodes and IAB-Donors. According to 3GPP specifications [4], an IAB-Donor
hosts a CU and multiple DUs. Similarly, IAB-Node is split into a DU and an MT.
Functionally, these have the task of enabling downstream and upstream
connectivity, respectively. At the time of writing, OAI’s main branch
implementation of the CU/DU functional split does not support multiple DUs
connected to a single CU [22]. This limitation is incompatible with the IAB
architecture. Consequently, we employ a full OAI Next Generation Node Base
(gNB) in place of both CU and DU. In other words, the IAB-Nodes and IAB-Donors
in our framework do not follow 3GPP split 2. Instead, these components are
deployed as monolithic gNBs. As for the MT, an open-source implementation is
currently unavailable. However, this component is functionally equivalent to a
UE, as it connects to upstream nodes using the same resources and protocols.
Consequently, we have selected OAI’s software-defined UE to act as MTs in the
proposed framework. This results in a system where a single IAB-Node is made
up of two concurrently running instances: an OAI gNB—acting as a DU—and an OAI
UE—acting as a MT. In the resulting architecture, IAB-Nodes are naturally
deployed over two separate machines, hosting the gNB and the UE, and connected
out-of-band as it is shown in Figure 2. Alternatively, the two software
components can run on a single x86 machine, provided that sufficient computing
power is available. While this architecture does not require any particular
modification to OAI’s implementations, we have added a signaling functionality
through which the IAB-Nodes or IAB-Donors can discern connected MTs from
connected UEs. This has been achieved through proper manipulation of the UE
Capability messages. Such information can be exploited to enable IAB-aware
optimization solutions in the gNB.
Core Network Functions. A minimal set of 5G CN functions have been included in
our framework: Network Repository Function (NRF), Access and Mobility
Management Function (AMF), Slicing Magangement Framework (SMF) and User Plane
Function (UPF), all based on the OAI 5G core implementation. All these
functions run as containers on a single x86 machine, as shown in Figure 2. Due
to the selected IAB system design, the UPF required modifications to enable
IAB operations. As previously mentioned, UEs acts as MTs in IAB-Nodes,
connecting to upstream nodes. The established GPRS Tunneling Protocol (GTP)
tunnels are then used to provide direct connectivity between the DU of the
node and the CN functions. In other words, MT-acting UEs relay the backhaul
traffic of the IAB-Nodes. However, OAI’s UPF implementation lacks support for
the required forwarding capability,222To the best of the authors’ knowledge,
there is no available open source implementation that supports this operating
mode. as any packet whose destination is not a UE is dropped. Therefore, we
have implemented a minimal version of framed routing [23] in OAI UPF, enabling
UEs to act as intermediate nodes.
User Equipment From the perspective of the UE, an IAB network deployed using
the components described above is entirely standard-compliant. As such, both
software-defined UEs (as shown in Figure 2) and COTS UEs can be used in the
proposed framework.
### III-B O-RAN Components
As mentioned in Section II, O-RAN defines a set of standardized and open
interfaces with which the RAN exposes data collection and control primitives
to the RICs. In the proposed framework, we have enabled IAB-Nodes and IAB-
Donors to be O-RAN-compatible by integrating software agents for the E2 and O1
interfaces into the codebase of OAI. Furthermore, our framework comprises a
near-RT RIC and a non-RT RIC.
E2 interface integration. The E2 interface is functionally split into two
protocols: E2AP—tasked with establishing a connection with the near-RT RIC—and
E2SM—which implements specific monitoring and control functionalities, namely
Service Models (SMs), as discussed in Section II. In the software
implementation we provide, E2AP has been adapted from O-RAN Alliance Software
Community reference implementation and, as such, it is entirely compliant with
O-RAN. On the other hand, the SMs provided by the O-RAN alliance are defined
using ASN.1: a powerful production-ready abstract description language which
is, however, cumbersome and challenging to use in the fast-paced research and
development environments targeted by our framework. In light of this, we
employ custom SM that are defined through Protocol Buffers (protobuf)—an
abstract definition language that is easier to handle and allows for fast
prototyping and testing, facilitating the development of IAB-aware control
solutions. Since the E2 interface is such that the E2SM messages are encoded
and decoded only in the RAN and xApp, the custom SM definitions are
transparent to the RIC, allowing our proposed solution to retain generic O-RAN
compliance. At the time of this writing, we have implemented a set of protobuf
messages that can be used to reproduce both the KPM and RAN Control (RC) SMs
[10]. These can be used to develop data collection and control xApps,
respectively.
(a) Linear IAB topology. $1$$2$$3$$4$$5$$10$$20$$30$$40$$50$Hops Throughput
[Mbps] DownlinkUplink (b) Throughput measurements for the linear topology.
$1$$2$$3$$4$$5$$20$$40$$60$$80$Hops RTT [ms] (c) RTT measures for the linear
topology.
Figure 3: Topology and results for the linear chain.
O1 interface integration. In order to properly manage all the different
aspects of networked elements, the O1 interface defines various Management
Services (MnS), which can be used either from the managed entities (the gNBs)
to report information back to the RIC or from the managing entity (the SMO and
the rApps running on it) to deploy configurations changes, transfer files or
update the software on the managed entities [24, 10]. Among all the different
MnS, we have focused our contribution on implementing the Heartbeat MnS, which
periodically transmits heartbeats; the Fault Supervision MnS, which reports
errors and events; and the Performance Assurance MnS, which streams
performance data. Those MnS have been integrated into the OAI codebase by
implementing a scheduler that, running on a dedicated thread, periodically
sends Virtual Network Function (VNF) Event Stream (VES) notifications in JSON
format over HTTP. This format and communication protocol has been chosen among
the different options defined in the standard, as it is widely known and
easily extendable by other researchers. As of now, our implementation reports
performance metrics, such as the throughput and information on the channel
quality between IAB-Nodes, and failure events, such as RRC or Uplink Shared
Channel (UL-SCH) failures, which can be used in rApps to monitor and optimize
the backhaul network. Provisioning MnS, which can be used by the rApps to
deploy configuration changes (e.g., topology optimizations), have not been
implemented by following the O1 specifications, as it would have needed major
reworks in the OAI codebase. Instead, we have taken advantage of IAB-Manager,
a software component we developed to orchestrate IAB experiments, as discussed
next.
IAB-Manager. IAB networks are generally expected to include several IAB-Nodes,
and the proposed framework can scale to such numbers. However, managing
experiments with tens or more RAN components can take time and effort. Indeed,
each component is potentially hosted by a dedicated machine, and setting up an
IAB deployment requires each one to be activated and configured according to a
sequence that starts from the CN functions and ends with the terminal IAB-
Nodes. To facilitate experimenting at such a large scale, we have developed
IAB-Manager [11]: a software component that can automate the IAB network
deployment and testing through a command line interface and an Application
Programming Interface (API). In particular, IAB-Manager is a single entrypoint
for controlling the entire experiment: network components and radio
environment setup (in case of wireless channel emulation), topology and
routing management and reconfiguration, automated testing, and result
collection. From a functional perspective, the manager connects to the
machines involved in the experimentation and configures them according to the
assigned roles. In particular, once the user specifies the roles, the manager
sequentially activates each network component until the final deployment is
ready for experimentation, greatly simplifying the setup phase. Additionally,
as previously mentioned, IAB-Manager executes the network configuration
changes mandated by the rApps.
RAN Intelligent Controllers. The proposed framework packages a near-RT RIC and
a non-RT RIC. Both are compliant with the standard and based on O-RAN Software
Community reference implementations.
## IV Validation and Results
Parameter | Value
---|---
Area Size for realistic deployment | 0.627 km2
gNB Density | 45 gNB/km2
IAB-donors/ IAB-nodes ratio | 1/10
Emulated center frequency | 28 GHz
Bandwidth | 40 MHz
Scheduler | 7 2 1
Subcarrier Spacing | 30khz
Colosseum Base loss | 50 dB
3GPP Channel Model | Urban Micro
MIMO layers | 1
TABLE I: Table of System Settings
This section focuses on validating our proposed framework from an experimental
perspective. In particular, we are interested in giving an initial
characterization of some fundamental Key Performance Indicators (KPIs) of the
deployments allowed by our IAB framework while validating its correct
functioning.
While the openness and flexibility of the software components are such that
the framework can run on generic hardware, we chose to base our validation
campaign on Colosseum [14]. The Colosseum testbed is a publicly available
large-scale testing platform with hardware-in-the-loop capabilities. It
comprises 128 Standard Radio Nodes (SRNs), each composed of a powerful x86
computing node and an USRP X310 SDR. All the components in the proposed
framework can be easily deployed on SRNs. Every SRN radio is interconnected by
an FPGA mesh that emulates arbitrary radio channels defined through tapered
delay models. With the capability of emulating complex scenarios of tens of
entities, Colosseum makes it possible to deploy large IAB networks over
complex propagation scenarios. As such, it represents an ideal validation
platform for our framework. Furthermore, Colosseum is open to the research
community, and the validation tools are made available, allowing interested
parties to start experimenting with a minimal initial effort.
### IV-A Experiments with a linear chain
We start by evaluating the performance of an IAB network deployed in a tightly
controlled scenario. To this end, we consider a 5-hop linear topology, as
shown in Figure 3(a). As detailed in Section II, each IAB-Node comprises an MT
and a DU, bringing this experiment’s overall radio node count to 10. In order
to characterize the upper-bound performance of the proposed framework, we
employ an ideal propagation scenario. Through properly manipulating
Colosseum’s channel emulator, a 0 dB pathloss model is selected for nodes
connected in the linear topology, and an infinite pathloss is set for all the
other channels, effectively suppressing any possible interference. In other
words, this radio scenario is equivalent to connecting the SDRs with coaxial
cables.333 Transmissions occur on band n78 with 106 Physical Resource Blocks
(PRBs) available, for a total of 40MHz bandwidth.
123$0$$10$$20$$30$$40$Hops Throughput [Mbps] (a) Downlink.
123$5$$10$$15$$20$Hops Throughput [Mbps] (b) Uplink. 123$20$$40$$60$$80$Hops
RTT [ms] (c) Round Trip Time.
Figure 5: Measurements for the realistic scenario.
Figure 3(b) shows the downlink and uplink TCP throughput against the number of
hops, as measured between the core network and the specific MT/UE. The first-
hop values of 47 Mbps in DL and 21 Mbps in UL represent the maximum throughput
attainable in the testing settings. This upper bound is far from the
theoretical maximum allowed by the available bandwidth. It is limited by
several factors that depend on the experimental platform, OAI software
implementation, and system design. Most notably, the strongest detractor to
the final throughput performance is given by the OAI implementation of the
software-defined UE, which is employed to build the MT. In particular, the OAI
UE is inefficient in reception and transmission, thus becoming a bottleneck
for the entire communication chain. Efforts are ongoing to improve the
performance and stability of this software framework. Furthermore, the
frameworks’ system design is such that each IP packet is encapsulated into as
many GTP packets as the number of hops. This increased overhead can cause
packet fragmentation with a further negative impact on the overall
performance. Furthermore, even if the emulated channel is set to a 0 dB
pathloss, Colosseum’s architecture includes an unavoidable base loss of 50 dB
[25] due to characteristics of the hardware architecture. This, together with
the aforementioned inefficiencies, make such that packet drops and subsequent
retransmissions happen also in this ideal scenario.
As the number of hops increase, the downlink throughput experiences a sharp
decrease before stabilizing on a per-hop loss of around 6 Mbps. The notable
throughput loss experienced at the second hop can be explained by observing
the standard deviation of the throughput, represented by the whiskers in
Figure 3(b). This value is at its maximum for the first hop, suggesting that
the first radio link is unstable due to the RX pipeline of the MT being
overwhelmed. This substantial variability is caused by packet loss and
retransmissions and internal buffer overflow, which negatively affect the
performance of the second hop, as it is noticeable in the numerical results.
At the same time, the second hop’s throughput standard deviation is lower, as
the decreased traffic volume causes less drops in the involved MTs. This
stabilizing effect propagates down the topology, as both the decreasing
standard deviation and the linear per-hop loss testify. On the other hand, the
uplink throughput is relatively stable and close to the upper bound, even at
the fourth hop. This is because the limited OAI UE performance and BS
scheduling process limits the uplink traffic volume, and the gNBs are far from
being overwhelmed. On the other hand, since the uplink throughput does not
significantly decrease from the maximum, the UE’s congestion level remains
relatively stable and high, as proven by the constant standard deviation
values.
Figure 4: Realistic deployment scenario in Florence, Italy. Donors are
represented in red, while IAB-Nodes are represented in yellow.
RTT is measured when the network is unloaded, that is when there is no traffic
flowing through the IAB network. As shown in Figure 3(c), the first hop
latency is around 11 ms. This value represents the base processing delay plus
a small fixed propagation delay that is, however, the same for each hop. As
the number of hops increases, the RTT experiences a linear increase comparable
with the first hop latency, as expected. This shows how the system does not
introduce any spurious latency when the network is unloaded. Finally, the
relatively higher RTT standard deviation of the last hop (as represented by
the whiskers in Figure 3(c)) suggests that multiple packet retransmissions are
required.
### IV-B Validation over realistic RF scenarios
After having validated the system performance in a controlled environment, we
move to more realistic urban scenarios, representing the typical deployment
environment of an IAB network. We start by selecting a densely populated area
of a city, from which we extract a precise 3D model of the environment. On top
of this, we apply a coverage planning heuristic to find the optimal locations
for the IAB-Nodes [26]. We then take advantage of a viewshed algorithm
implemented on GPU and the above-mentioned 3D models to evaluate the Line of
Sight (LoS) between each pair of locations and to produce the so-called
visibility graph [27]. Then, we characterize the propagation according to the
3GPP Urban Micro parametric model [28], and we produce a tapered-delay
representation of the communication channels, which Colosseum can then
emulate. This process has been carried out for several European cities and
four different scenarios are made
available.444https://colosseumneu.freshdesk.com/support/solutions/articles/61000303373-integrated-
access-and-backhaul-scenarios
Motivated by the fact that IAB is unanimously considered as a key enabler of
mmWave RAN [29], we are interested in providing an experimental solution that
enables testing in such conditions. While Colosseum is not directly capable of
operating at frequencies higher than 6 GHz, we can approximate these radio
scenarios by reproducing the most relevant propagation characteristics of
mmWaves, namely the extremely directive transmissions through beamforming and
the increased pathloss [30]. In particular, the pathloss between nodes that
are not directly connected in the provided topologies has been set to
infinite. The resulting suppression of inter-node interference might appear
too ideal at first. However, this is compatible with the highly directive
transmissions typical of mmWave, where interference in static conditions
(i.e., as in a backhaul IAB topology) can be practically neglected [31]. A
more refined mmWave channel emulation will be subject of future extensions. In
addition, since Colosseum’s channel emulation happens in base-band, we can
apply arbitrary pathloss independently of the radio frequency employed during
the experiments. Thanks to this flexibility, we could compute pathloss for a
carrier frequency of 28 GHz and apply them to LoS links. Nonetheless, the
scenarios made available to the Colosseum community are available for both 3.6
GHz and 28 GHz, both with and without inter-node interference suppression.
For the experimental evaluation presented in this work, we have selected a
scenario based on the city of Florence, Italy. Figure 4 shows both the urban
layout and the IAB deployment, which is extended over $0.7~{}\text{km}^{2}$
and comprises 21 nodes (3 of which are IAB-Donors). To determine which nodes
are going to become IAB-Donors, we have applied the group closeness centrality
metric [32] to the visibility graph. This centrality metric selects $k$ nodes
such that their distance to all the other nodes is minimized. Then, we have
determined the IAB topology as a Shortest-Path Forest computed over the
visibility graph of the area with the well-known Dijktra’s Algorithm. Similar
to what has been done for the previous analysis, we characterize the
throughput and latency at each hop in the network. In this case, however, the
different link lengths cause performance variations in the per-hop throughput
and latency. As such, we employ box plots to synthetically describe the
network performance statistics in Figure 5. In particular, consider Figure
5(a). Here the bottom and top edges of each box represent the first and third
quartile of the downlink throughput measurements taken at all the different
hops in the scenario. Similarly, the central marks indicate the median, and
the whiskers represent the extreme data points. The plotted values indicate
how the realistic pathloss introduced in the study scenario causes lower
performance than the ideal case previously analyzed, independently of the
considered hop. The same can be noted for the uplink throughput, as shown in
Figure 5(b). In both cases, the decreasing per-hop throughput trend is
conserved. However, the throughput variability is the same for the two
transmission directions. This is because, as opposed to the ideal scenario,
the link length now represents the main performance-determining factor. This
is testified by the significant distance between the first and third quartile
of the first hop in both downlink and uplink throughput, which is consistent
with the high variance of the first hop length in the topology of study. As
for the second and third hop, the relatively closer quartiles are motivated by
lower link length variations for these hops in the considered topology.
Finally, the upper whiskers represent the performance of the shortest links,
giving a further characterization of the system performance in this realistic
scenario.
Figure 5(c) shows the RTT statistic through the same plotting technique.
Differently from the throughput, the latency is not affected by the link
length variations in the considered scenario for the first two hops.
Additionally, the RTT increase at hops 1 and 2 is consistent with the one
experienced in the controlled scenario. On the other hand, the high RTT
variance of the third and last hop suggests a high probability of requiring
retransmissions along the IAB path.
## V Conclusions
In this work, we have discussed the motivations and challenges of integrating
IAB with O-RAN. On this matter, we have proposed possible architecture
extensions that enable dynamic control and data collection over 5G IAB
networks through O-RAN intelligent controllers. We have implemented the
proposed integrated architecture and packaged it into the first publicly
available experimental framework enabling at-scale testing and prototyping of
O-RAN-based solutions applied to IAB networks. The system comprises all the
software components required to establish end-to-end connectivity, plus
custom-developed E2 and O1 agents that allow software-defined IAB-Nodes to be
O-RAN-compliant. The framework is designed to be open and accessible and can
be deployed over COTS hardware. We numerically validated the framework
exploiting the large-scale testing capabilities of Colosseum, showing the
system’s performance over both an ideal linear topology and more sophisticated
realistic deployments. Finally, the framework has been packaged and released
into OpenRAN Gym and is available to the research community.
## References
* [1] S. Dang, O. Amin, B. Shihada, and M.-S. Alouini, “What should 6G be?” _Nature Electronics_ , vol. 3, no. 1, pp. 20–29, 2020.
* [2] N. Bhushan, J. Li, D. Malladi, R. Gilmore, D. Brenner, A. Damnjanovic, R. T. Sukhavasi, C. Patel, and S. Geirhofer, “Network densification: the dominant theme for wireless evolution into 5G,” _IEEE Communications Magazine_ , vol. 52, no. 2, pp. 82–89, 2014.
* [3] T. Inoue, “5G NR Release 16 and Millimeter Wave Integrated Access and Backhaul,” in _2020 IEEE Radio and Wireless Symposium (RWS)_ , 2020.
* [4] M. Polese, M. Giordani, T. Zugno, A. Roy, S. Goyal, D. Castor, and M. Zorzi, “Integrated Access and Backhaul in 5G mmWave Networks: Potential and Challenges,” _IEEE Communications Magazine_ , vol. 58, no. 3, pp. 62–68, 2020.
* [5] W. Feng, Y. Wang, D. Lin, N. Ge, J. Lu, and S. Li, “When mmWave Communications Meet Network Densification: A Scalable Interference Coordination Perspective,” _IEEE Journal on Selected Areas in Communications_ , vol. 35, no. 7, pp. 1459–1471, 2017.
* [6] Y. Zhang, V. Ramamurthi, Z. Huang, and D. Ghosal, “Co-optimizing performance and fairness using weighted pf scheduling and iab-aware flow control,” in _2020 IEEE Wireless Communications and Networking Conference (WCNC)_ , 2020\.
* [7] B. Zhang, F. Devoti, I. Filippini, and D. De Donno, “Resource allocation in mmWave 5G IAB networks: a reinforcement learning approach based on column generation,” _Computer Networks_ , vol. 196, p. 108248, 2021.
* [8] “Coordinated parallel resource allocation for integrated access and backhaul networks,” _Computer Networks_ , vol. 222, p. 109533, 2023.
* [9] S. Ranjan, P. Chaporkar, P. Jha, and A. Karandikar, “Backhaul-Aware Cell Selection Policies in 5G IAB Networks,” in _2021 IEEE Wireless Communications and Networking Conference (WCNC)_ , 2021.
* [10] M. Polese, L. Bonati, S. D’Oro, S. Basagni, and T. Melodia, “Understanding O-RAN: Architecture, Interfaces, Algorithms, Security, and Research Challenges,” _IEEE Communications Surveys & Tutorials_, pp. 1–1, 2023\.
* [11] E. Moro, M. Polese, I. Filippini, S. Basagni, A. Capone, and T. Melodia, “IABEST: An Integrated Access and Backhaul 5G Testbed for Large-Scale Experimentation,” in _Proceedings of the 28th Annual International Conference on Mobile Computing And Networking_ , ser. MobiCom ’22. Sydney, NSW, Australia: Association for Computing Machinery, 2022.
* [12] L. Bonati, M. Polese, S. D’Oro, S. Basagni, and T. Melodia, “OpenRAN Gym: AI/ML Development, Data Collection, and Testing for O-RAN on PAWR Platforms,” _Computer Networks_ , vol. 220, pp. 1–11, January 2023.
* [13] A. Gosain, “Platforms for Advanced Wireless Research: Helping Define a New Edge Computing Paradigm,” in _Proceedings of the 2018 on Technologies for the Wireless Edge Workshop_ , ser. WirelessEdge ’18. New Delhi, India: Association for Computing Machinery, 2018\.
* [14] L. Bonati, P. Johari, M. Polese, S. D’Oro, S. Mohanti, M. Tehrani-Moayyed, D. Villa, S. Shrivastava, C. Tassie, K. Yoder _et al._ , “Colosseum: Large-scale wireless experimentation through hardware-in-the-loop network emulation,” in _2021 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN)_. IEEE, 2021.
* [15] S. Niknam, A. Roy, H. S. Dhillon, S. Singh, R. Banerji, J. H. Reed, N. Saxena, and S. Yoon, “Intelligent O-RAN for Beyond 5G and 6G Wireless Networks,” in _2022 IEEE Globecom Workshops (GC Wkshps)_ , 2022.
* [16] H. Lee, J. Cha, D. Kwon, M. Jeong, and I. Park, “Hosting AI/ML Workflows on O-RAN RIC Platform,” in _2020 IEEE Globecom Workshops (GC Wkshps)_ , 2020\.
* [17] B. Brik, K. Boutiba, and A. Ksentini, “Deep learning for B5G open radio access network: Evolution, survey, case studies, and challenges,” _IEEE Open Journal of the Communications Society_ , vol. 3, pp. 228–250, 2022\.
* [18] 3GPP, “NR; NR and NG-RAN Overall description; Stage-2,” 3rd Generation Partnership Project (3GPP), Technical Report (TR) 38.300, 09 2022, version 17.2.0.
* [19] ——, “NG-RAN; Architecture description,” 3rd Generation Partnership Project (3GPP), Technical Specification (TS) 38.401, 09 2022, version 17.2.0.
* [20] O.-R. WG3, “O-RAN Near-Real-time RAN Intelligent Controller E2 Service Model (E2SM) KPM 2.0,” Open-RAN Alliance, Technical Specification (TS), 7 2021, version 2.0.0.
* [21] F. Kaltenberger, A. P. Silva, A. Gosain, L. Wang, and T.-T. Nguyen, “OpenAirInterface: Democratizing innovation in the 5G Era,” _Computer Networks_ , vol. 176, p. 107284, 2020.
* [22] (2021) F1 interface - Data plane status. [Online]. Available: https://gitlab.eurecom.fr/oai/openairinterface5g/-/wikis/f1-interface
* [23] 3GPP, “Interface between the Control Plane and the User Plane Nodes,” 3rd Generation Partnership Project (3GPP), Technical Specification (TS) 29.144, 6 2022, version 16.4.0. [Online]. Available: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3111
* [24] O-RAN Working Group 10, “O-RAN Operations and Maintenance Interface Specification 8.0,” O-RAN.WG10.O1-Interface.0-v08.00 Technical Specification, Oct 2022.
* [25] D. Villa, M. Tehrani-Moayyed, P. Johari, S. Basagni, and T. Melodia, “CaST: A Toolchain for Creating and Characterizing Realistic Wireless Network Emulation Scenarios,” in _Proceedings of the 16th ACM Workshop on Wireless Network Testbeds, Experimental Evaluation & Characterization_, ser. WiNTECH ’22. Sydney, NSW, Australia: Association for Computing Machinery, 2022.
* [26] G. Gemmi, R. Lo Cigno, and L. Maccari, “On Cost-Effective, Reliable Coverage for LoS Communications in Urban Areas,” _IEEE Transactions on Network and Service Management_ , vol. 19, no. 3, pp. 2767–2779, 2022.
* [27] G. Gemmi, R. L. Cigno, and L. Maccari, “On the Properties of Next Generation Wireless Backhaul,” _IEEE Transactions on Network Science and Engineering_ , vol. 10, no. 1, pp. 166–177, 2023.
* [28] 3GPP, “Study on channel model for frequencies from 0.5 to 100 GHz,” 3rd Generation Partnership Project (3GPP), Technical Specification (TS) 38.901, 01 2020, version 16.1.0. [Online]. Available: https://portal.3gpp.org/desktopmodules/Specifications/SpecificationDetails.aspx?specificationId=3173
* [29] M. Cudak, A. Ghosh, A. Ghosh, and J. Andrews, “Integrated Access and Backhaul: A Key Enabler for 5G Millimeter-Wave Deployments,” _IEEE Communications Magazine_ , vol. 59, no. 4, pp. 88–94, 2021.
* [30] S. Kutty and D. Sen, “Beamforming for millimeter wave communications: An inclusive survey,” _IEEE communications surveys & tutorials_, vol. 18, no. 2, pp. 949–973, 2015.
* [31] P. Fiore, E. Moro, I. Filippini, A. Capone, and D. D. Donno, “Boosting 5G mm-Wave IAB Reliability with Reconfigurable Intelligent Surfaces,” in _2022 IEEE Wireless Communications and Networking Conference (WCNC)_ , 2022\.
* [32] A. Bavelas, “Communication patterns in task-oriented groups,” _The journal of the acoustical society of America_ , vol. 22, no. 6, pp. 725–730, 1950\.
|
# Cluster analysis of homicide rates in the Brazilian state of Goiás from 2002
to 2014
Samuel Bruno da Silva Sousa Institute of Science and Technology
Federal University of São Paulo
São José dos Campos, SP - Brazil
<EMAIL_ADDRESS>Ronaldo de Castro Del-Fiaco Campus Henrique
Santillo
State University of Goiás
Anápolis, GO - Brazil
<EMAIL_ADDRESS>Lilian Berton Institute of Science and Technology
Federal University of São Paulo
São José dos Campos, SP - Brazil
<EMAIL_ADDRESS>
###### Abstract
Homicide mortality is a worldwide concern and has occupied the agenda of
researchers and public managers. In Brazil, homicide is the third leading
cause of death in the general population and the first in the 15-39 age group.
In South America, Brazil has the third highest homicide mortality, behind
Venezuela and Colombia. To measure the impacts of violence it is important to
assess health systems and criminal justice, as well as other areas. In this
paper, we analyze the spatial distribution of homicide mortality in the state
of Goiás, Center-West of Brazil, since the homicide rate increased from 24.5
per 100,000 in 2002 to 42.6 per 100,000 in 2014 in this location. Moreover,
this state had the fifth position of homicides in Brazil in 2014. We
considered socio-demographic variables for the state, performed analysis about
correlation and employed three clustering algorithms: K-means, Density-based
and Hierarchical. The results indicate the homicide rates are higher in cities
neighbors of large urban centers, although these cities have the best
socioeconomic indicators.
###### Index Terms:
Homicide, Violence, Spatial analysis, K-means clustering, Density-based
clustering, Hierarchical clustering
## I Introduction
Violence appears in the 21st century as one of the major social problems in
urban centers[15]. This phenomenon is noticed in several ways in different
spaces. It is not unusual to see news from all around the world reporting
violent acts every day. Both physical and psychological forms of violence may
change the relationships among people and the space they live[15]. We notice
violence in households, groups, communities, neighborhoods, cities, and so on.
Homicide is one of the most frightful forms of violence, and its occurrences
vary from a place to another.
In the most part of the world, the homicide rates are low, excluding war
zones. The homicide count in 2012[13] was around 437,000 victims in the world.
In 2017, the global homicide rate per 100,000 inhabitants was 6.2. 78% of the
victims were male, and the countries with the highest homicide rates were
Honduras (85.5), Venezuela (53.7), and Belize (44.7). On the other hand,
Monaco, Liechtenstein, and Andorra did not register any case of homicide which
made their rates per 100,000 inhabitants reach the value 0. Most of the
countries on the top of the list of homicide rates[13] were located in Latin
America or the Caribbean. These regions historically presented high homicide
rates, which is related to drug trafficking, corruption in governments, and
cartels operations[23].
In Brazil, the homicide rate per 100,000 inhabitants increased between 2002 to
2014. In 2002, the rate was 27.9 per 100,000, while in 2014 reached 29.7 per
100,000, according to data from Igarapé Institute[13]. The main reasons for
this increase are related to drug trafficking, corruption, and lack of
policing. Only in 2014, 59,681 people were murdered in Brazil. The country is
also responsible for more than 10% of the homicides in the world[23].
In some Brazilian states, the increase in homicide rates has been even higher.
For instance, in the state of Goiás, the homicide rate increased from 24.5 per
100,000 in 2002 to 42.6 per 100,000 in 2014, totalizing 24,300 victims in the
period[13]. The state of Goiás is located in the Center-West region of Brazil,
surrounding the Brazilian capital, Brasília. Its population was estimated at
around 6,7 million people in 2017[4], and its larger city is the capital,
Goiânia, with 1,4 million inhabitants. The state is formed by 246
municipalities whose social-economic features vary a lot from one to another.
Although the state concentrated the fifth highest homicide rate amongst the
Brazilian states in 2014, this theme is still under-researched.
Brazil covers 8.5 million $km^{2}$ (47% of South America), with an estimated
population of 190 million inhabitants in 2010[5]. The country is divided into
27 States with differing socioeconomic and health conditions. Recently
published studies in Brazil analyzed homicide and suicide mortality rates in
different states[24, 10, 11, 9, 20]. Despite the beliefs that homicides
cluster in specific regions[2], no papers using machine learning techniques
such as clustering algorithms to analyze homicide rates were found. Thus, this
paper introduces a new approach in this field.
Organizing data into groupings is an important mode of understanding and
learning. Cluster analysis can be defined as “the formal study of methods and
algorithms for grouping objects according to measured or perceived intrinsic
characteristics or similarity, without prior knowledge of the number of
clusters or any other information about their composition”[3]. Cluster
analysis does not use category labels that tag objects with prior identifiers,
such as class labels. The absence of category information distinguishes data
clustering (unsupervised learning) from classification (supervised
learning)[3].
The main contributions of this paper are: 1) evaluation of correlations
between homicide and socio-demographic variables for the state of Goiás,
Brazil; 2) employment of three clustering algorithms (K-means, Hierarchical
and Density-based clustering) to cluster the cities and to identify critical
areas for homicides; 3) pattern identification suggesting that, in Goiás, the
homicide rates are higher in cities neighbors of large urban centers, like
Brasília and Goiânia, despite these cities having the highest per capita
income in the country.
The rest of the paper is organized as follows: Section II presents related
work about homicide rate analysis in the world and in Brazil. Section III
presents the materials and methods employed in this work. The objectives,
datasets, and algorithms are described with details. Section IV presents the
results of our analysis. Finally, Section V presents the final remarks and
future works.
## II Related works
A plenty of works has been developed on homicide rate analysis since the last
half of the 20th century. The fields of sociology and criminology were the
first to start researching this theme. The main purposes of those works were
related to investigating whether demographic, economic, ecological, and social
variables maintained some correlations to the variation in homicide rates
across time and space[17]. Variables as resident racial segregation, racial
inequality, extreme poverty, social capital, and unemployment rate were used
for some well-succeeded findings[21].
In the United States of America, gunshot violence is responsible for about
34,000 deaths annually[8]. A paper published in 2017 used scan spatial
statistics to analyze clusters of the gunshot occurrences within the city of
Syracuse, New York. Amongst the results, it was noticed that the higher
violence rate was related to environmental and economic disparities[8].
In Central America, where the homicide rates are historically elevated, some
works have been developed on the analysis of possible causes. In El Salvador,
the clusters of homicides may be related to drug trafficking and organized
crime[7]. And in Mexico, the spatial variation of homicides was explained as
being linked to firearms possession, drug trafficking and social
exclusion[12].
Some papers analyzed mortality by homicide in Brazilian states. Souza et
al.[24] takes an ecological approach to the situation of homicides in all the
municipalities of Bahia for the male population aged 15 to 39, considering the
health macro-region. Lozada et al.[10] analyzed the homicide mortality trend
from 1979 to 2005 for males aged 15 to 49, living in the State of Paraná.
Souza et al.[11] made an ecological study which analyzed the violence-related
death records of women aged 10 years and older, in the Brazilian geographic
regions, between 1980 and 2014. Bando and Lester[9] evaluated correlations
between suicide, homicide and socio-demographic variables by an ecological
study for the Brazilian states. Peres et al.[20] described homicide mortality
in the municipality of São Paulo according to social-economic features of the
victims and type of weapon.
As far as we know no paper studied the homicide rates focused on the Goiás
state. Furthermore, we did not find any work using clustering algorithms to
analyze the spatial distribution of homicides. The techniques and tools
commonly used on this topic are: estimation of regression coefficients[17,
21]; spatial scan statistics[8]; SaTScan software methods[2]; Bayesian
approach with Monte Carlo Markov Chain algorithm[7]; Moran’s Global index[24];
descriptive statistics or correlation techniques[10, 9]; estimable functions
and negative binomial regression[11]; SSPS software tools[20]; and multiple
regression analysis (stepwise method)[12].
This paper aims to analyze spatial patterns of mortality by homicide for
better understanding violence in the Goiás state and to identify risky areas.
Due to the lack of use of clustering algorithms on this theme, this paper
proposes the use of K-means, Hierarchical and Density-based algorithms for
clustering, in an exploratory approach.
## III Materials and methods
This is an exploratory study of homicide mortality in the State of Goiás for
the period from 2002 to 2014. This period was chosen in function of data
availability. Goiás had an estimated population of 6,778,772 inhabitants in
2017 and has 246 municipalities, being recognized as the 12th most populated
state in Brazil[4]. The following sections present the objectives (III-A), the
methodology (III-B), the datasets used (III-C) and the evaluation measure
employed in the experiments (III-D).
### III-A Objectives
The general objective of this study is to analyze the spatial distribution of
mortality by homicide in the State of Goiás. The specific objectives can be
divided as follows:
* •
Collect data about mortality from homicide rates (MHR) in Goiás from 2002 to
2014.
* •
Collect demographic and socioeconomic characteristics of the various
territorial areas for the purpose of correlated with homicide rates.
* •
Apply cluster algorithms to group similar data, specifically, K-means,
Hierarchical and Density-based algorithms.
* •
Identify patterns in the clustering results to identify risky areas.
### III-B Methodology
The presence of clusters was evaluated by K-means, Hierarchical and Density-
based algorithms. The software used for building the database and for
computing the clusters was R.111https://www.r-project.org/
K-means clustering aims to partition $n$ data instances into K clusters in
which each instance belongs to the cluster with the nearest mean, serving as a
cluster’s prototype[16]. This is a very popular algorithm in data mining. It
proceeds by alternating between two steps, given an initial set of K centroids
$m_{1}$, $m_{2}$… $m_{n}$. The first step comprehends the assignment of each
instance to the cluster with the least squared Euclidean Distance. In the
second step, new means are calculated to be the centroids of the new clusters.
During the iterative process, it is common that some instances skip from one
cluster to another. The ending of the iterations occurs when the instances are
established and, at this time, the K-means are truly the means of the groups
formed[16].
Let $\mu_{k}$ be the centroid of cluster $c_{k}$. The squared error between
$\mu_{k}$ and the points in cluster $c_{k}$ is defined as in Equation 1.
$G(c_{k})=\sum_{\vec{x_{i}}\in c_{k}}||\vec{x_{i}}-\mu_{k}||^{2}$ (1)
The goal of K-means is to minimize the sum of the squared error overall K
clusters, as defined by Equation 2.
$G(C)=\sum_{k=1}^{K}\sum_{x_{i}\in c_{k}}||x_{i}-\mu_{k}||^{2}$ (2)
Automatically determining the number of clusters is a difficult problem in
data clustering. K-means is run independently for different values of K and
the partition that appears the most meaningful to the domain expert is
selected.
Hierarchical clustering is a method of cluster analysis which seeks to build a
hierarchy of clusters[22]. Its representation consists of a hierarchical
system (dendrogram) like the taxonomic hierarchy in biology. Strategies for
hierarchical clustering generally fall into two types:
* •
Agglomerative: This is a “bottom up” approach where each observation starts in
its own cluster and pairs of clusters are merged as one moves up the
hierarchy.
* •
Divisive: This is a “top-down” approach where all observations start in one
cluster, and splits are performed recursively as one moves down the hierarchy.
Formally, it is assumed that when there is a sequence of $n$ elements,
represented by values from 1 to $n$, there is also a sequence of clusterings
with length $m+1$ ($C_{0}$, $C_{1}$, $C_{2}$, …, $C_{m}$)[22]. Each cluster
has a number $\alpha_{i}$ with its value. The cluster $C_{0}$ is considered
the weakest clustering of $n$ elements ($\alpha_{0}$ = 0), and the cluster
$C_{m}$ is considered the strongest clustering. The numbers $\alpha_{i}$
increase ($\alpha_{i}$ $\leq$ $\alpha_{i+1}$) as well as the clusters $C_{i}$
($C_{i}$ $\leq$ $C_{i+1}$), which means each cluster $C_{i}$ is the merging
(or union) of clusters $C_{i-1}$[22]. In general, the merges and splits are
determined in a greedy manner. The results of hierarchical clustering are
usually presented in a dendrogram. The most common algorithms are Single-
linkage and Complete-linkage clustering.
Density-based clustering (DBSCAN) is one of the most common density-based
clustering algorithms. It was designed to discover clusters and noise in a
spatial dataset, requiring only one parameter for input whose value is
determined by the user[19]. This algorithm works as follows: given a set of
points in some space, it groups together points that are closely packed
together (points with many nearby neighbors), marking as outliers points that
lie alone in low-density regions (whose nearest neighbors are too far away).
The DBSCAN algorithm can be abstracted into the following steps:
* •
Find the $\epsilon$ neighbors of every point and identify the core points with
more than a predefined number of neighbors.
* •
Find the connected components of core points on the neighbor graph, ignoring
all non-core points.
* •
Assign each non-core point to a nearby cluster if the cluster is an $\epsilon$
neighbor, otherwise, signalize it as noise.
DBSCAN is an effective algorithm to discover clusters that present unusual
shapes even in large spatial databases[19].
### III-C Dataset
Data from 246 municipalities of the state of Goiás (Figure 1) were used as
analysis units.
Figure 1: Municipalities of the State of Goiás.
Source: http://www.sieg.go.gov.br/rgg/atlas/index.html/.
The data were collected between January and March of 2018 by downloads from
official websites of organizations maintained by the Brazilian government and
the United Nations. All data used in this analysis are publicly available for
download on the Internet. They are highly reliable once they are published by
governmental agencies and by internationally recognized organizations.
The data on demographics consist of the population estimate for each
municipality and its demographic density. They were extracted from the website
of the Brazilian Institute of Geography and Statistics (IBGE)[4]. The values
used in the query consisted of the municipalities’ names.
The data on mortality were extracted from the Mortality Information System
(SIM)[6]. This system is maintained and updated by the Brazilian Ministry of
Health with data from mortality in general. In the query, we aimed to extract
only the MHR, that included deaths by aggression, under the CID-10 codes X85
and Y09, and by legal intervention, under the codes Y35 and Y36. The selected
values in the query were: the area of influence of the State of Goiás, the
municipalities’ names, the period in the extraction was between the years 2002
and 2014, and the codes of CID-10 mentioned above.
The data about the Basic Education Development Index (IDEB)[14] were extracted
from the website of Instituto Nacional de Estudos e Pesquisas Educacionais
Anísio Teixeira (INEP), an agency from the Brazilian Ministry of Education.
This index measures the quality of basic education in Brazil, according to
international standards. The values for the index range from 0 to 10. In the
query, we selected to consult by: municipality, public (federal, state or
municipal) administrative dependency, series corresponding to the final years
of Secondary School (8th and 9th years in Brazil).
The data on social and economic features were collected from the website of
the United Nations Development Program (UNDP). In Brazil, UNDP develops an
atlas of development in the country. This atlas, named Human Development Atlas
in Brazil, covers all the Brazilian States and Federal District and its
municipalities. In this study, we collected the data that comprehended the
Municipal Human Development Index (MHDI)[25] and its variables, as life
expectancy, education index, the percentage of total income appropriated by
the 10% richest, and Gini index. These data were available to be downloaded in
a Microsoft Excel file.
Table I contains the name and description of all variables selected to compose
the dataset.
TABLE I: Description of the social-economic variables. Variable | Description
---|---
MHR | Total number of MHR since 2002 until 2014.
POPULATION | Population counting in IBGE 2010’s census.
DEMOGDENSITY | Municipality population by its total area.
IDEB2005 | Basic Education Development Index in 2005.
IDEB2007 | Basic Education Development Index in 2007.
IDEB2009 | Basic Education Development Index in 2009.
IDEB2011 | Basic Education Development Index in 2011.
IDEB2013 | Basic Education Development Index in 2013.
LIFEEXPECT | Life expectation in 2010.
GINI | Gini coefficient in 2010.
INRICHEST10 | Rate of the overall income held by richest 10%.
EDUCLEVEL | Education level of adult population in 2010.
MHDI | MHDI in 2010.
MHDIE | MHDI (Education) in 2010.
MHDIL | MHDI (Longevity) in 2010.
MHDII | MHDI (Income) in 2010.
For this analysis, the dataset was preprocessed and analyzed in R. We have
made use of Factoextra, Amap, FPC, Cluster, and DBSCAN packages.
### III-D Evaluation of results
The process of evaluating the results obtained from a clustering algorithm is
commonly called validation. There are two main types of grouping validation
indexes: i) external indexes, which compare the group structure discovered
with a previously known group structure; ii) internal indexes, which analyzes
the structure of groups discovered in relation to some criterion, such as, for
example, compactness or separability.
We employ the external index Silhouette, and the internal measures GAP
statistic and Sum of Squares Within (SSW) to validate the clustering results.
The Silhouette index (SIL) is calculated per data and the SIL of a group is
the average of the SIL of all the data in the group. And the clustering SIL is
the mean of the SIL of the groups. The higher the index value the better. It
is described by Equation 3.
$SIL=\frac{a_{i}-b_{i}}{max\\{a_{i},b_{i}\\}}$ (3)
where $a_{i}$ is the average distance from the data $i$ to all other data in
its group and $b_{i}$ is the minimum distance of the data $i$ to all other
data that do not belong to its group.
The GAP statistic compares the overall intra-cluster variation for a different
number of clusters with their expected values under a null reference data
distribution. The GAP statistic for a given value $k$ is defined as;
$GAP_{n}(k)=E^{*}_{n}log(W_{k})-log(w_{k}),$ (4)
where $E^{*}_{n}$ is an expectation in a sample of size $n$ from the reference
distribution, obtained by bootstrapping by computing the average $log(W_{k})$,
and by generating $B$ copies of the dataset. The GAP statistic measures the
deviation of the observed $W_{k}$ value from its expected value under a null
hypothesis.
SSW is a measure of compactness of a cluster structure, given by the distance
between the centroid of a cluster and the instances assigned to it. The SSW
equation has the form:
$SSW_{M}=\sum^{N}_{i=1}{||x_{i}-c_{Pi}||}^{2},$ (5)
where $x_{i}$ is an instance in each cluster whose centroid is $c_{Pi}$.
## IV Results and discussion
Section IV-A presents the analysis of the variables and their correlation with
homicide rate. Pearson’s, Spearman’s and Kendall’s indexes were employed to
measure correlation. Section IV-B presents the cluster analysis. Algorithms
Hierarchical, Density-based and K-means were employed to cluster the
municipalities. Section IV-C presents the discussion of the results.
### IV-A Social-economic variables analysis
From 2002 to 2014, 24,300 people were murdered in the State of Goiás[13]. The
number of MHR did not vary a lot between 2002 and 2007, but from 2007 to 2013
it increased year-to-year and reached the highest record in 2013 (Figure 2).
In 2014, it was noticed a slight decrease in the number of MHR, but this was
still the second highest record for the period.
Figure 2: MHR in the State of Goiás by year.
The Gini coefficient provides a measure of income distribution among
individuals or households within a territory, such as a country, a state, a
municipality, etc. A value 0 represents perfect equality, and a value 100
represents a total inequality. In our analysis, we used a normalized range of
Gini coefficient from 0 to 1, in the way it was presented on the data source.
Figure 3 presents the distribution of Gini coefficient among the
municipalities in the State of Goiás. Fewer than 20 municipalities have
presented values for Gini index under 0.4. Thus, it is possible to notice that
the State of Goiás presents high-income disparities among its municipalities
and within most of the municipalities too. For instance, the state’s largest
city, Goiânia, presented a Gini value of 0.58.
Figure 3: Distribution of Gini coefficient in Goiás (2010).
The educational variable EDUCLEVEL is related to the percent of the adult
population that has completed the last year of Secondary School. In most of
the municipalities, fewer than 60% of the adult population completed Secondary
School (Figure 4). This situation is due to the fact of the most of the
population lived in rural areas until the decade of 1970[18], when schools and
universities were concentrated in the largest cities. So the rural population
was aside from formal education.
Figure 4: Distribution of education level of adult population (2010).
The variable IDEB consists of the Basic Education Development Index. In this
study, we analyzed the values for IDEB in the last years of Secondary School
for public schools. The means for IDEB from 2003 to 2013, are shown in Figure
5. It is possible to notice that the values of IDEB were increasing through
the years, what means improvements in the quality of public basic education.
In 2013, the IDEB mean for the state reached 4.5. In this year, the
municipalities with the highest IDEB means were Perolândia (6.2), Córrego do
Ouro (5.9), and Ivolândia (5.9), that are all small cities. On the other hand,
the cities with the lowest IDEB means in 2013 were Águas Lindas de Goiás
(3.4), Água Fria de Goiás (3.5), and Cavalcante (3.6), where the first two are
located in the Surroundings of Federal District.
Figure 5: IDEB variation through the years.
In 2012 UNDP started calculating MHDI for all municipalities in Brazil. This
index is calculated the same way as HDI, which considers three dimensions
(income, education, and longevity). Despite having the seventh highest HDI in
Brazil (0.735) in 2010[4], the State of Goiás presents a large disparity in
the values of HDI among its municipalities (Figure 6). More than half of the
municipalities of Goiás have values for MHDI under 0.70, i.e., medium and low
human development.
Figure 6: Distribution of Municipal Human Development Index in Goiás (2010).
Figure 7: Regression lines between socio-economic variables and MHR (Color
online).
(a) Population
(b) Demographic density
(c) Life expectation
(d) IDEB index
(e) Educational level
(f) Gini index
(g) Income held by richest 10%
(h) MHDI
(i) MHDI (Education)
We also have calculated simple linear regression and local regression (LOWESS)
between the dataset variables. The results are shown in Figure 8(i). Simple
linear regression is a statistical approach to model a relationship in the
values of a dependent and one independent variable. This approach is commonly
used in forecasting, error reduction, and relationships explanation. LOWESS is
a non-parametric approach which makes use of a multiple linear method in a
model based in k-Nearest Neighbor, popularly employed in trends forecasting of
variables relationship.
The variables population (Sub-figure 7(a)) and demographic density (Sub-figure
7(b)) presented strong relationships with MHR values. In these sub-figures, it
is possible to notice that both simple regression lines (red lines) and LOWESS
lines (blue lines) assume rising values. Otherwise, the other variables are
shown in the Sub-figures 7(c) to 7(i) presented weak relationships with MHR
values. Both regression lines in these sub-figures assumed values equal or
close to 0. The lines for the variables MHDI (Longevity) and MHDI (Income)
also presented the values of simple and LOWESS regression close to 0.
Table II presents the Pearson’s, Spearman’s and Kendall’s coefficient
correlations between the social-economic variables considered and the MHR. The
values of the correlations are between -1 and +1. The signal indicates the
direction, whether the correlation is positive or negative, and the absolute
value of the variable indicates the strength of the correlation. Values
between 0.7 and 0.9 indicate a strong correlation. Notice that Population has
a strong correlation with MHR as indicated by the three coefficients.
Demographic Density also has a strong correlation, as indicated by the
Pearson’s coefficient.
TABLE II: Correlations between the total of MHR and social-economic variables. Variable | Pearson’s c. | Spearman’s c. | Kendall’s c.
---|---|---|---
POPULATION | 0.9915637 | 0.8932006 | 0.7385185
DEMOGDENSITY | 0.7262442 | 0.3446707 | 0.2389337
IDEB | -0.1440371 | -0.2019128 | -0.1417871
LIFEEXPECT | 0.09612802 | 0.1669409 | 0.1119476
GINI | 0.1134491 | 0.2435379 | 0.1748277
INRICHEST10 | 0.1039698 | 0.2740375 | 0.1939559
EDUCLEVEL | 0.3994967 | 0.4042891 | 0.287134
MHDI | 0.248962 | 0.1948997 | 0.133404
MHDIE | 0.206129 | 0.08632882 | 0.05902461
MHDIL | 0.096134 | 0.1656033 | 0.1120611
MHDII | 0.2649425 | 0.2717906 | 0.1852401
Before applying the cluster algorithms, we computed the distance between the
dataset columns (Figure 8), to visualize the (dis)similarity within the
dataset. It is possible to notice that values for demographic density,
population, and MHR are close to each other. At the same time, these variables
present values far from the educational, economic, and MHDI variables values,
showing to us that in the state of Goiás population growth was not accompanied
by educational and human development growth.
Figure 8: Dataset distance matrix (Color online).
### IV-B Cluster analysis
This section presents the results of applying the Hierarchical, Density-based
and K-means algorithms in the municipalities of Goiás. We validate the
clustering results with the Silhouette measure, GAP statistic, and SSW.
#### IV-B1 Hierarchical clustering
The hierarchical clustering algorithm, with single linkage method and
Manhattan distance, generated clusters with few municipalities. We have cut
the dendrogram into 3 partitions, which is shown in the Figure 9. The optimal
number of clusters to cut the tree was decided by the results of Silhouette
index and GAP statistic. We used the value which presented the biggest
difference to its successor, and the optimum GAP statistic is given by
$GAP_{(k)}>GAP_{(k+1)}-S_{k+1}$, being $S_{k+1}$ the standard deviation
multiplied by the square of 1 plus the inverse of B copies of the reference
dataset done by bootstrapping.
This algorithm assigned to the first cluster (in red) the largest city of the
state, Goiânia; assigned to the second cluster (in yellow) the city with the
highest demographic density, Valparaíso de Goiás; and assigned to the third
cluster (in purple) the other cities in the state, despite of the
dissimilarity between them.
Figure 9: Results of Agglomerative Hierarchical Algorithm (Color online).
#### IV-B2 Density-based clustering
The DBSCAN algorithm generates just one cluster and assigned as outliers 20
municipalities, amongst these outliers, there are: the largest cities of the
state, the municipality with the highest demographic density, and small cities
whose values for MHR are 0 or close to it. This algorithm did not generate a
meaningful result, because of the dissimilarity between the instances of the
dataset. Most of the municipalities assigned as outliers hold the highest
values for the variables population, demographic density and MHR. The results
of DBSCAN are shown in the Figure 10.
Figure 10: Results of DBSCAN (Color online). Figure 11: Results of K-means
clustering with K = 4 (Color online).
#### IV-B3 K-means clustering
We employed different K values (from 1 to 5) in the K-means clustering
algorithm. We also employed different distance measures, such as Euclidean,
Manhattan, Pearson, and Canberra. The one that provided the most meaningful
set of clusters was Canberra, once this measure is not sensible to outliers
presence. It was decided to start K with 5 because the state has 5 regions and
maybe this should influence the distribution of MHR. With K = 2, 3, and 5, the
results were not interesting. In every application, the algorithm assigned
Goiânia to the same cluster as small cities. The K = 4 value generated groups
more interesting as shown in Figure 11. The results signal different dynamics
in the distribution of MHR among the municipalities of Goiás. The results of
K-means algorithm are shown in Figure 11.
In cluster 1 (beige), there are 96 municipalities predominantly small cities.
The cluster presented the second lowest mean for the number of MHR and
variables, such as population, demographic density, life expectation, Gini
index, educational level of adult population, MHDI and its tree dimensions
(Education, Longevity, and Income). Otherwise, this cluster has the second
highest mean for income held by the richest 10% and the highest mean for IDEB
index amongst all the clusters.
In cluster 2 (pink), there are 108 municipalities, which are also small cities
close to the lowest means for population, demographic density, income held by
the richest 10%, educational level of adult population, MHDI and its three
dimensions and MHR amongst all the clusters. Furthermore, this cluster has the
second lowest IDEB and Gini index means. As the cities in this cluster are
predominantly small, far from the two metropolis in the region, the
educational level of the adult population tends to be low because, as cited in
Sec. IV-A, most of the state population was aside from formal education until
the decade of 1970.
In cluster 3 (black), there are 33 municipalities mid-sized like Rio Verde,
Cidade Ocidental, Catalão, Trindade, and Itumbiara; and some small cities.
This cluster presented the highest means of life expectation, Gini index,
income held by the richest 10%, MHDI, MHDI (Longevity), and MHDI (Income).
Furthermore, it was noticed in this cluster the second highest means for the
variables population, demographic density, educational level, MHDI
(Education), IDEB, and MHR.
The cluster 4 (purple) aggregates only 8 municipalities, all of them big or
mid-sized cities. In this cluster, there are the municipalities of Goiânia,
Aparecida de Goiânia, Anápolis, Valparaíso de Goiás, Águas Lindas de Goiás,
Senador Canedo, Novo Gama, and Luziânia. These cities compose the region
between the capital of the state of Goiás, Goiânia, and the capital of Brazil,
Brasília. The means of the population, demographic density, educational level,
MHDI (Education), and MHR were the highest amongst the 4 clusters. At the same
time, the cluster presented the second highest means for life expectation,
Gini index, MHDI, MHDI (Longevity), and MHDI (Income). The mean income held by
the richest 10% was the third highest. Although having the highest value for
the educational level of the adult population, the index which measures the
quality of basic education – IDEB – presented the lowest mean amongst all the
clusters. Furthermore, this cluster held 59.59% of the MHR in the state from
2002 until 2014.
In the Fig. 12, we plotted the results of K-means clustering. The small black
points represent the cluster 1 aforementioned. The yellow points represent the
cluster 2. The blue points represent the cluster 3, and the big red points
represent the cluster 4. It is possible to see that all the cities within the
fourth cluster are located close to the region largest cities, Goiânia and
Brasília, and Goiânia is also in this cluster.
Figure 12: Results of K-means clustering.
### IV-C Discussion
Applying clustering algorithms it was possible to identify clusters of
homicides in the state that are not merely random. Such clusters were
considered as critical areas for homicides. It was also possible to notice
different dynamics in the distribution of MHR in the state. The largest cities
presented most of the homicide cases, despite its income and human development
levels being higher than most of the other municipalities. On the other hand,
these large cities present low values for IDEB index, suggesting that although
having most people completing Elementary and Secondary School degrees, the
quality of teaching in public schools is still low.
We also computed Silhouette measure, GAP statistic, and SSW to validate the
clustering results (Table III). The values in bold represent the optimum
values for these measures. The optimum Silhouette value for each algorithm was
assumed as the value with the biggest difference to its successor in the
table. Optimum GAP statistic values were determined by selecting the lowest
which satisfied the condition: $GAP_{(k)}>GAP_{(k+1)}-S_{k+1}$. The optimal
values for SSW were assumed by analyzing the bend (knee) in the curves of SSW,
i.e., the SSW values in bold were the bends (knees) in their respective curves
of SSW.
TABLE III: Validation measures. Clustering algorithm | Silhouette | GAP | SSW
---|---|---|---
K-means (2) | 0.97 | 0.717 | 22.55
K-means (3) | 0.89 | 0.748 | 26.10
K-means (4) | 0.84 | 0.765 | 23.96
K-means (5) | 0.72 | 0.769 | 22.44
Hierarchical (2) | 0.97 | 0.722 | 50.00
Hierarchical (3) | 0.94 | 0.742 | 44.30
Hierarchical (4) | 0.86 | 0.731 | 40.00
Hierarchical (5) | 0.85 | 0.743 | 36.80
Density-based | 0.97 | 0.705 | 66.71
Measures such as Silhouette, GAP statistic, and SSW helps us to measure the
clustering quality under a mathematical perspective. As clustering is an
unsupervised task of data mining and Machine Learning, there is no label to be
predicted/classified, and no training and test phases, to lately measure the
error rate. So a meaningful clustering result is that one whose interpretation
makes sense for a human being [1].
Hierarchical and K-means clustering generates groups separating the large
cities from small ones. The Silhouette measure indicates high similarity
between the instances clustered. However, when the number of clusters was
increased, the Silhouette measure decreased. The value 4 for K in K-means
clustering was assumed as optimal because of GAP statistic value for K = 4 is
bigger than the value for K = 5, i.e., $GAP_{(k=4)}>GAP_{(k=5)}-S_{k=5}$.
K-means with K = 2 value achieves the highest Silhouette and the lowest SSW,
but this clustering generates one group with 147 small cities and another
group with 99 large cities. However, K = 4 gives a more interesting separation
splitting the outliers cities, and K = 5 presents a low value for Silhouette,
which suggests a weak clustering.
Density-based clustering resulted in just one cluster, and assigned as
outliers 20 municipalities, amongst them the state largest cities. The
Silhouette value for this results was also high, as well as the values for GAP
statistic and SSW.
Few studies bring information about all the municipalities from the State of
Goiás, especially in the interior regions. Our contribution emphasizes the
importance of the use of spatial analysis to understand violence indicators
and geographic distribution.
## V Conclusion
In this paper, we employed three clustering algorithms: Hierarchical, Density-
based and K-means to analyze spatial patterns of mortality by homicide in the
Brazilian state of Goiás.
The state’s three largest cities, Goiânia, Aparecida de Goiânia, and Anápolis,
had the highest numbers of MHR. The municipalities in the region known as
Surroundings of the Federal District also had elevated numbers of MHR. In the
other municipalities, the numbers were low. All clustering algorithms
separated these regions with high MHR from the others. However, K-means with K
= 4 results in a better separation.
The means for population, Gini index, MHDI, IDEB, income concentration, and
educational level of adult population varied among the municipalities as well
as the number of MHR. The highest IDEB values were found in the smallest
cities, which presented low homicide rates. In general, we noticed three
scenarios for the occurrence of MHR in the state of Goiás. The first is
characterized by a low occurrence of MHR in the small cities, as shown by the
clusters 1 and 2 in the Subsection IV-B. The second scenario refers to the
occurrence of MHR in mid-sized cities in municipalities with low demographic
density and high-income concentration, as shown by the cluster 3. The last
scenario is related to the occurrence of MHR in the state’s municipalities
with highest values of demographic density, low values of IDEB index, although
having high MHDI and income, and comprising the metropolitan areas of Goiânia
and Brasília, as noticed in the cluster 4.
The use of clustering algorithms in the analysis of the distribution of MHR
was an experimental approach, once we did not find studies that did it. In
general, the behavior of the algorithms was satisfactory. Hierarchical
clustering and DBSCAM provided results influenced principally by the
population size, although K-means, with K = 4, showed an interesting
separation whose quality was confirmed by its Silhouette measure, GAP
statistic, and SSW.
We expect this study to be useful to improve the knowledge about the
distribution of MHR in the State of Goiás and to aid the state’s government to
define risky regions, where policing should be more effective. This study is
also a source to know the variation of social-economic variables among the
state’s municipalities.
As future work other analysis employing different variables to explain the
rates of homicides in certain regions could be done, like precarious
socioeconomic conditions, social inequality, etc.
## VI Acknolowdgments
This work was conducted during a scholarship at ICT-UNIFESP – Institute of
Science and Technology of Federal University of São Paulo. Financed by CAPES –
Brazilian Federal Agency for Support and Evaluation of Graduate Education
within the Ministry of Education of Brazil.
## References
* [1] A. C. Lorena, J. Gama, K. Facelli, and A. C. P. L. F. de Carvalho, “Inteligência Artificial: uma abordagem de aprendizado de máquina,” Grupo Gen-LTC, 2011.
* [2] A. M. Zeoli, S. Grady, J.M. Pizarro, and C. Melde, “Modeling the movement of homicide by type to inform public health prevention efforts,” American journal of public health, Vol. 105, No. 10, pp. 2035-2041, 2015.
* [3] A. K. Jain, “Data clustering: 50 years beyond K-means,” Pattern recognition letters, Vol. 31, No. 8, pp. 651-666, 2010.
* [4] Brazilian Institute of Geography and Statistics. (2018, Mar.) Goiás - IBGE Cidade. [Online]. Available: https://cidades.ibge.gov.br/brasil/go.
* [5] Brazilian Institute of Geography and Statistics. (2018, Mar.) IBGE Censo 2010. [Online]. Available: https://censo2010.ibge.gov.br/sinopse/index.php?dados=4&uf=00.
* [6] Brazilian Ministry of Health. (2018, Mar.) TabbNet Win32 3.0: Mortalidade - Goiás. [Online]. Available: http://tabnet.datasus.gov.br/cgi/deftohtm.exe?sim/cnv/obt10GO.def.
* [7] C. Carcach, “A spatio-temporal model of homicide in El Salvador,” Crime science, Vol. 4, No. 1, pp. 20, 2015.
* [8] D. A. Larsen, S. Lane, T. Jennings-Bey, A. Haygood-El, K. Brundage, and R. A. Rubinstein, “Spatio-temporal patterns of gun violence in Syracuse, New York 2009-2015,” PLoS ONE, Vol. 12, No. 3, pp. e0173001, 2017.
* [9] D. H. Bando, and D. Lester, “An ecological study on suicide and homicide in Brazil,” Ciência & saúde coletiva, Vol. 19, pp. 1179-1189, 2014.
* [10] E. M. K Lozada, T. A. D. F. Mathias, S. M. D. Andrade, and T. Aidar, “Homicide mortality trend in the state of Paraná, Brazil, per Health District, 1979 to 2005,” Revista Brasileira de Epidemiologia, Vol. 12, No. 2, pp. 258-269, 2009.
* [11] E. R. de Souza, et al., “Homicides among women in the different Brazilian regions in the last 35 years: an analysis of age-period-birth cohort effects,” Ciência & Saúde Coletiva, Vol. 22, No. 9, pp. 2949-2962, 2017.
* [12] G. J. Gonzalez-Perez, M. G. Vega-lopez, C. E. Cabrera-pivaral, and A. Vega-Lopez, “P1-432 Violence and health: an epidemiological analysis of homicides in México, 1979–2008,” Journal of Epidemiology & Community Health, Vol. 65, No. 1, 2011.
* [13] Igarapé Institute. (2016, Mar.) Homicide Monitor. [Online]. Available: http://homicide.igarape.org.br.
* [14] Instituto Nacional de Estudos e Pesquisas Educacionais Anísio Teixeira. (2016, Mar.) IDEB - Resultados e Metas. [Online]. Available: http://ideb.inep.gov.br/resultado/home.seam?cid=2098801.
* [15] J. Laplante, “La violence, la peur et le crime,” University of Ottawa Press, 2001.
* [16] J. MacQueen, “Some methods for classification and analysis of multivariate observations,” Proceedings of the fifth Berkeley symposium on mathematical statistics and probability, Vol. 1, No. 14, 1967.
* [17] K. C. Land, P. L. McCall, and L. E. Cohen, “Structural covariates of homicide rates: Are there any invariances across time and social space?,” American journal of sociology, Vol. 95, No. 4, pp. 922-963, 1990.
* [18] Mauro Borges Institute. (2018, Mar.) Tabela 3 - Estado de Goiás, Centro-Oeste e Brasil: População por situação de domicílio – 1950, 1960, 1970, 1980, 1991, 2000, 2008 – 12. [Online]. Available: http://www.imb.go.gov.br/pub/Godados/2013/01-tab03.htm.
* [19] M. Ester, H. P. Kriegel, J. Sander, and X. Xu, “A density-based algorithm for discovering clusters in large spatial databases with noise,” Kdd. Vol. 96, No. 34, 1996.
* [20] M. F. T. Peres, D. Vicentin, M. B. Nery, et al., “Queda dos homicídios em São Paulo, Brasil: uma análise descritiva,” Revista Panamericana De Salud Publica, Vol. 29, No. 1, pp. 17-26, 2011.
* [21] P. L. McCall, K. C. Land, and K. F. Parker, “An empirical assessment of what we know about structural covariates of homicide rates: a return to a classic 20 years later,” Homicide Studies, Vol. 14, No. 3, pp. 219-243, 2010.
* [22] S. C. Johnson, “Hierarchical clustering schemes,” Psychometrika, Vol. 32, No. 3, pp. 241-254, 1967.
* [23] The Economist. (2016, Mar.) Daily chart: Revisiting the world’s most violent cities. [Online]. Available: https://www.economist.com/blogs/graphicdetail/2016/03/daily-chart-18.
* [24] T. O. de Souza, L. W. Pinto, and E. R. de Souza, “Spatial study of homicide rates in the state of Bahia, Brazil, 1996-2010,” Revista de saúde pública, Vol. 48, pp. 468-477, 2014.
* [25] United Nations Development Program. (2016, Mar.) Atlas do desenvolvimento humano no Brasil 2013. [Online]. Available: http://www.atlasbrasil.org.br/.
|
# All A-board: Sharing Educational Data Science Research With School Districts
Nabeel Gillani Northeastern UniversityBoston, MAUSA
<EMAIL_ADDRESS>, Doug Beeferman Massachusetts Institute of
TechnologyCambridge, MAUSA<EMAIL_ADDRESS>, Cassandra Overney Massachusetts
Institute of TechnologyCambridge, MAUSA<EMAIL_ADDRESS>, Christine Vega-
Pourheydarian Wellesley CollegeWellesley, MAUSA<EMAIL_ADDRESS>and
Deb Roy Massachusetts Institute of TechnologyCambridge, MAUSA
<EMAIL_ADDRESS>
(2023)
###### Abstract.
Educational data scientists often conduct research with the hopes of
translating findings into lasting change through policy, civil society, or
other channels. However, the bridge from research to practice can be fraught
with sociopolitical frictions that impede, or altogether block, such
translations—especially when they are contentious or otherwise difficult to
achieve. Focusing on one entrenched educational equity issue in US public
schools—racial and ethnic segregation—we conduct randomized email outreach
experiments and surveys to explore how local school districts respond to
algorithmically-generated school catchment areas (“attendance boundaries”)
designed to foster more diverse and integrated schools. Cold email outreach to
approximately 4,320 elected school board members across over 800 school
districts informing them of potential boundary changes reveals a large average
open rate of nearly 40%, but a relatively small click-through rate of 2.5% to
an interactive dashboard depicting such changes. Board members, however,
appear responsive to different messaging techniques—particularly those that
dovetail issues of racial and ethnic diversity with other top-of-mind issues
(like school capacity planning). On the other hand, media coverage of the
research drives more dashboard engagement, especially in more segregated
districts. A small but rich set of survey responses from school board and
community members across several districts identify data and operational
bottlenecks to implementing boundary changes to foster more diverse schools,
but also share affirmative comments on the potential viability of such
changes. Together, our findings may support educational data scientists in
more effectively disseminating research that aims to bridge educational
inequalities through systems-level change.
Education; Inequality; Segregation; Randomized Experiments
††journalyear: 2023††copyright: acmlicensed††conference: Proceedings of the
Tenth ACM Conference on Learning @ Scale; July 20–22, 2023; Copenhagen,
Denmark.††booktitle: Proceedings of the Tenth ACM Conference on Learning @
Scale (L@S ’23), July 20–22, 2023, Copenhagen, Denmark††price: 15.00††isbn:
979-8-4007-0025-5/23/07††doi: 10.1145/3573051.3593377††ccs: Applied computing
Education††ccs: Applied computing Sociology††ccs: Applied computing Decision
analysis††ccs: Applied computing Psychology
## 1\. Introduction
There is perhaps no place in America where failures of pluralism and democracy
are manifesting more clearly than in public schools. In recent years, US
school districts have become battlegrounds for political debates about in-
person schooling (Sawchuck, 2021), mask mandates (Harris, 2022), and critical
race theory, along with discussions about gender and sexuality (Natanson,
2021). Despite this climate, public schools have the capacity to help
cultivate common ground, more evenly distribute social capital, spark exposure
to and appreciation for divergent viewpoints, and foster other aspects of a
healthy, pluralistic society by bringing together students and families from
different backgrounds.
Yet across the US, school segregation along racial, cultural, and
socioeconomic lines remains rampant (Office, 2022) and continues to adversely
affect the academic performance and attainment of lower socioeconomic status
(SES) students and students of color (Reardon et al., 2018; Owens, 2018;
Reardon and Owens, 2014; Mayer, 2014). One reason for this is that segregation
may exacerbate inequities in how resources like experienced teachers and
advanced course offerings are distributed across schools (EdBuild, 2020;
Frankenberg et al., 2003; Orfield and Lee, 2005). Segregated schools may also
impede access to social networks, which may in turn limit access to the kinds
of “bridging social capital” that can help them access new resources, jobs,
and other quality-of-life-enhancing opportunities (Bell et al., 2017; Small,
2009; Burton and Welsh, 2015; Chetty et al., 2022). There is evidence that
integration can reduce inequalities in academic outcomes (Yeung, 2022; Quick,
2016; Johnson, 2011), especially when academic environments are re-designed to
be inclusive of an increasingly diverse student body (Burton and Welsh, 2015;
Bridges, 2016).
Despite such evidence, actual socioeconomic integration remains elusive across
many districts, largely because of racialized schooling preferences among
predominantly White and affluent families (Billingham and Hunt, 2016; Hailey,
2021; Hall and Hibel, 2017) and residential/school selection based on factors
related to income (Reardon and Bischoff, 2011; Hasan and Kumar, 2019). The
correlation between place of residence and place of schooling is reinforced by
“school attendance boundaries”, or school catchment areas that local education
agencies (“school districts”) draw to determine which neighborhoods are
assigned to (“zoned for”) which schools. Despite the growing popularity of
school choice policies, neighborhood-based assignment still accounts for the
vast majority of student assignment decisions across the US (of Education,
2021). Furthermore, choice policies can also exacerbate patterns of racial and
ethnic segregation (Monarrez et al., 2022; Whitehurt, 2017; Candipan,
2019)—despite how they, at least in theory, place of residence and schools.
Many choice systems also factor in geography in some form or fashion—for
example, through the priorities they assign students (Monarrez and Chien,
2021) or how they create zones that constrain choice sets for families (Campos
and Kearns, 2022). Boundaries, therefore, matter, and play an important role
in determining which children go to which schools and what the quality of
their ensuing educational experiences are.
Educational data scientists—an emerging class of researchers and practitioners
wielding computational methods to explore outstanding issues in education
(McFarland et al., 2021)—have expended great effort exploring how such methods
can improve online learning (Kizilcec et al., 2013; Gillani and Eynon, 2014),
personalized and adaptive tutoring (Ritter et al., 2007; Piech et al., 2015),
and other learner and teacher-centric practices. There appears to be fewer
studies, however, investigating how such methods might help address challenges
that are “upstream” to these: namely, social, organizational and contextual
barriers that might impede access to quality-of-life enhancing educational
opportunities. Attendance boundaries that exacerbate segregation often act as
such barriers. Yet they are malleable: school districts could, in theory,
redraw attendance boundaries in ways that seek to minimize racial and ethnic
segregation across schools. Doing so, however, is a politically challenging
task, and often runs the risk of increasing travel times or producing other
unpopular outcomes that face opposition from parents and other stakeholders
(Delmont, 2016; McMillan, 2018). It is also technically challenging: manual
map redrawing can be highly time-consuming given the large space of possible
configurations, and from the political redistricting literature, it is clear
that automated methods are challenging to develop and scale in their own right
(Gurnee and Shmoys, 2021; Becker and Solomon, 2020). There is an opportunity,
then, for educational data scientists to explore how computational methods
might help education policymakers—a stakeholder group not usually directly
targeted as an audience for pedagogy-focused research—foster more diverse and
integrated schools. This is, arguably, especially relevant for educational
data scientists interested in “Learning at Scale”, since the school assignment
policies that districts design and implement continue to shape the learning
experiences of millions of children across the US.
In this vein, a recent study developed computational redistricting algorithms
to explore how much school districts across the US might reduce racial and
ethnic segregation in elementary schools through boundary changes (Gillani et
al., 2023). The results, surprisingly, revealed the possibility of boundary
changes that might reduce segregation while also slightly reducing travel
times. These findings motivate this particular study: a follow-on effort to
explore to what extent US school districts might engage with the findings from
this research and take them into consideration as they conduct their own
boundary planning efforts. To investigate this, we adopt a multi-pronged
outreach strategy, combining cold email outreach to 4,320 elected school board
members across 803 districts with media coverage in order to invite leaders
and members of school districts to explore an online dashboard depicting
algorithmically-generated, diversity-promoting elementary school attendance
boundaries. Our focus on outreach seeks to address another gap, not just in
educational data science research, but academic research more broadly: better
understanding the conditions under which academic findings might be translated
into practice. The emerging field of “implementation science” is increasingly
exploring this question (Bauer et al., 2015), and recent studies in the
education domain have used school district outreach as a channel (Delevoye,
2021; Nakajima, 2021) for investigating how district leaders interpret and put
policy-relevant research to use. We aim to build on these bodies of work by
asking the following questions:
1. (1)
Which messaging and outreach strategies are more or less effective in
prompting school district leaders and community members in exploring potential
boundary changes that could help foster more racially/ethnically integrated
schools?
2. (2)
How do school district leaders and community members perceive these boundary
changes—namely, what potential merits and pitfalls do they see in both how
these boundaries were produced and their potential viability in practice?
We find that email open and click rates are highest when campaign content is
framed in terms of existing district objectives—like school capacity planning
and family satisfaction—which were suggested by district leaders through our
informal conversations with them. Furthermore, media coverage appears to be a
much more effective channel than email outreach in engaging audiences around
potential boundary changes in their school districts—especially districts that
exhibit higher levels of White/non-White segregation. Through an optional
survey available on the data dashboard, we obtain a more nuanced view of how
communities perceive such hypothetical, algorithmically-informed attendance
policy changes: many participants highlight aspects of their unique contexts
that would need to be considered in order for the research findings to be
adapted and eventually implemented in practice, like rapid shifts in
population due to immigration; school capacity constraints; and other factors.
Nevertheless, most respondents indicate more interest in the topic of
attendance boundary changes to foster diverse schools than they had prior to
learning about the study, and several express an interest in learning more
about how to implement some of the proposed changes in practice. While our
study focuses on the issue of changing attendance boundaries to foster more
racial and ethnic integration, we believe our findings may be relevant more
broadly to educational data scientists who wish to explore how findings from
research can be effectively communicated with key education stakeholders, and
eventually translated into effective socio-technical artifacts and practices
in learning settings.
## 2\. Related work
Our work builds upon work in school assignment planning for measuring and
promoting more diverse schools. It also builds upon recent efforts exploring
how to engage school district leaders and other policymakers to help them use
findings from academic research in their work. Most broadly, it is motivated
by and rooted in the emerging discipline of “implementation science” (Bauer et
al., 2015), which seeks to more precisely identify the contexts and conditions
under which research findings can be scaled and generalized to populations
beyond those included in the research. We briefly review these areas below.
### 2.1. School assignment for diversity
Across school districts, the lines between districts have been shown to
perpetuate racial and ethnic segregation across schools even more than those
within them (i.e., attendance boundaries) (Fiel, 2013). Unfortunately,
changing school district boundaries require collaboration across multiple
districts, and/or state-level intervention—both of which are subject to many
of the political frictions and challenges that accompany policy changes
requiring coordination across multiple governing bodies. Changing attendance
boundaries, on the other hand, generally falls under the purview of school
district leaders. Yet within-district boundary changes continue to be highly
contentious: parents may worry about impacts on values (Kane et al., 2005;
Bridges, 2016; Black, 1999), commute times (Frankenberg and Jacobsen, 2011;
Delmont, 2016), community cohesion (Bridges, 2016; Staff, 2019), and other
factors. These concerns can block boundary changes (Castro et al., 2022)
and/or spark families to move to other schools and districts (Reber, 2005;
Macartney and Singleton, 2018).
The present study does not explore how to address these parent/ community-
level factors, which are critical topics for future research on this issue.
Instead, it explores to what extent school district leaders are willing even
to engage with research on this contentious topic that suggests there might be
potentially palatable solutions within reach—and what their reactions to such
findings are. Given the positions of power that board members occupy, we
believe this is important to investigate, as identifying effective ways of
discussing the topic of boundary changes for diversity with district leaders
may illuminate key leverage points for supporting systemic policy change.
### 2.2. Engaging education policymakers in research findings
While still a relatively nascent area of inquiry, recent studies have started
to explore how to design and execute effective communications with school
districts to foster more engagement with, and policy decision-making based on,
evidence-based research. Two recent studies in particular come to mind. The
first conducted randomized outreach to over 2,000 education professionals
working across both state and local education agencies to assess both the
preferences these policymakers have for research evidence, and the extent to
which they update their beliefs about the effectiveness of various education
policies as they are presented with new evidence (Nakajima, 2021). The study
found that policymakers do indeed value research evidence, particularly when
they contain findings from multiple sites, large samples, and contexts similar
to their own—though they do not appear to prefer experimental studies over
observational ones. The study also found that policymakers are more likely to
update their beliefs about the effectiveness of education policies when
research evidence is presented with clear, accessible descriptions of its
research design. Together, these results offer valuable insights into how
clarity of communication and local contextualization of findings might help
increase policymakers’ engagement with education research.
A second study, conducted during the COVID-19 pandemic, explored the potential
agenda-setting role of cold email outreach to school district board members
and staff (Delevoye, 2021). The authors randomly assigned a subset of nearly
600 school districts leaders to receive an email containing a link to a memo
with research on emerging best practices around transportation in education
settings during the pandemic. They then analyzed video recordings and other
outputs from school board meetings to analyze whether those district leaders
who received emails were more likely to discuss transportation-related issues.
While the study did not report a positive treatment effect, it did find that
some of the words from the memo made it into some of the school boards’
discourse. It is possible that with a larger sample of districts, there may
have been more of an agenda-setting impact of such research-based outreach.
Our study builds on these efforts by exploring the types of cold email
outreach that are more or less effective at garnering engagement from school
district leaders vis-a-vis a particularly contentious and systemic issue: how
school attendance boundaries are drawn, and the impact these boundaries have
on issues of school segregation and diversity in their districts.
### 2.3. Implementation science
Yet another emerging area of inquiry is implementation science, or “the
scientific study of methods to promote the systematic uptake of research
findings and other evidence-based practices into routine practice …” (Bauer et
al., 2015). For example, development economists are increasingly exploring
implementation science through the lens of “scaling”, or investigating how
promising interventions conducted in relatively sand-boxed settings might be
effectively scaled across different contexts and environments (Mobarak, 2022).
Scaling successfully, of course, requires anticipating the complexities that
may come with scale—which, crucially, include potentially unanticipated or
indirect consequences (Mobarak, 2022).
Health researchers are also interested in implementation science; indeed, the
health domain has been a driver of this new discipline, perhaps because
evidence-based practices in healthcare take, on average, 17 years to be
implemented into routine general practice (Bauer et al., 2015). One
interesting emerging trend from this domain is the design of “hybrid research
designs”, or research projects that seek to both further understanding about
evidence-based practices (e.g., how well some practice improves a health
outcome—generally, the type of finding one might typically expect to uncover
by running a randomized control trial), as well as a
documentation/investigation of which practices and policies are most effective
in enabling such interventions to have this intended impact (Bauer et al.,
2015). Such hybrid designs offer the hope of both advancing basic scientific
inquiry as well as theories and knowledge about how the outputs of such
inquiry can actually serve people in different real-world contexts.
Many of these ideas have started to take hold in education research (Moir,
2018), as partly demonstrated by the district outreach studies in the prior
subsection. They are also beginning to spark new field-building and funding
priorities in education research (Foundation, 2022). For an issue as complex
as school assignment planning, particularly when intersected with a topic as
controversial as racial and ethnic segregation in schools, modeling
hypothetical boundary changes is just one small part—and perhaps the easiest
part—of investigating how to foster more diverse and integrated schools.
Deeply understanding the sociopolitical contexts of these districts, how to
capture the attention and interest of district leaders and community members,
and eliciting the conditions under which alternative boundaries may or may not
be viable are all critical pieces of translating such research into practice.
On their own, these activities do not actually identify conditions for
successful translations in the field, but rather, help identify potential
starting points for iterating on existing research to eventually help make it
more useful in practice. Our study, then, may offer helpful insights to
educational data scientists, public policy researchers, and other
practitioners interested in learning more about how to begin translating
potentially controversial, equity-promoting research findings into practice in
complex settings like school districts.
## 3\. Methods
### 3.1. Modeling and data dashboard
We begin with the datasets and rezoning algorithms used in (Gillani et al.,
2023). This preliminary study focused on elementary school segregation,
combining 2019/2020 student enrollment data by race and ethnicity across US
schools with 2020 Census demographics and 2021/2022 school attendance
boundaries to estimate counts of students living in each Census block and
attending each zoned elementary school (see (Gillani et al., 2023) for a more
thorough discussion of the datasets and some of the challenges involved in
making these estimations). It then developed and presented a redistricting
algorithm, building on top of open source tools from the operations research
community (OR-Tools, 2022; Perron and Didier, 2020), that produced
hypothetical attendance boundary redrawings to help foster more racially and
ethnically integrated schools. The redrawing problem was framed as an
assignment problem, namely, trying to find an alternative assignment of Census
blocks to schools that would minimize White/non-White segregation (as defined
by the dissimilarity index (Massey and Denton, 1988)), subject to constraints
around the maximum amount any family’s travel time might increase under the
new rezoning, as well as a cap on how much school capacities might increase.
Such redistricting problems are NP-hard and notoriously difficult to solve
given the large space of possible assignments; nevertheless, recent advances
in the operations research community, e.g. through innovations in methods like
constraint programming (Van Hentenryck, 1989; Perron and Didier, 2020), have
created powerful off-the-shelf solvers that can often produce high-quality
solutions to such problems.
(Gillani et al., 2023) simulated changes across 98 large school districts and
reported a median relative decrease in segregation of 12%, which would require
nearly 20% of students to switch schools. Perhaps most surprisingly, the study
showed that such reductions may be achieved with slight average decreases in
travel times—suggesting there may be modest changes that produce modest gains
in integration without requiring large travel-related sacrifices by families.
Using the datasets and algorithms from this study, we began by re-running the
authors’ redistricting algorithms for a larger set of over 4,000 school
districts—namely, those school districts that have at least two elementary
schools (and hence, different neighborhoods within the district zoned to
different schools). In total, we produced 4 different configurations across
each district by varying different parameter values for the aforementioned
travel time and school size constraints. Simulations were run on a university
computing cluster with a cutoff time of 5.5 hours each. We then produced a
dashboard similar to the one in (Gillani et al., 2023) to enable district
leaders and community members to explore hypothetical rezonings. The dashboard
was built using the Python-based Streamlit111https://streamlit.io/., a
lightweight framework for building and deploying data dashboards. Using
Streamlit limited some of our data tracking, logging, and customization
possibilities; on the other hand, it enabled us to rapidly prototype and
deploy a dashboard without requiring full-stack design and engineering.
Figure 1 shows screenshots of the dashboard, which served as the main
exploration platform for this study. To make the description of the project’s
aims and key findings as accessible as possible on the dashboard, we
informally sought feedback from a past school district superintendent involved
in boundary planning for integration; a current (at the time) school
assignment and planning director; and a data journalist who frequently writes
about issues relating to educational equity. We conducted Zoom calls with
these individuals, inviting them to explore the dashboard and share their
observations and questions (similar to a think-aloud protocol (Ericsson and
Simon, 1984)). Their feedback on clarity of wording (how we describe changes
in school-level demographics—and avoid esoteric segregation formulas in order
to make research findings more accessible (Nakajima, 2021)), framing of
diversity (making sure to not lose the focus on within-classroom learning
experiences) and data visualizations was invaluable and led us to make several
changes that we believe made the dashboard more accessible to a general
audience.
Figure 1. Key screenshots of rezoning dashboard, which can be found at
https://www.schooldiversity.org/: a) shows the view “above the fold” when
users land on the page and includes a description of why attendance boundaries
matter and what the objective of the dashboard is. It also provides dropdown
boxes for users to explore hypothetical boundary changes for different school
districts, and includes a verbal description of what the depicted changes for
any given district would mean in terms of changes in diversity across its
constituent elementary schools. b) shows a map view, where users can toggle
between boundaries proposed by the redistricting algorithm and current
“status-quo” boundaries. Further down the page and hidden by default is c),
which allows users to drill into specific schools to understand how their
demographics would change under the depicted rezoning—including cases where
rezonings move racial/ethnic proportions at some schools away from district-
levels instead of towards them, perhaps to achieve more balance at other
schools. Finally, further down is d), which shows how travel times would
change (on average) for different racial/ethnic groups under the rezoning.
### 3.2. Email campaign
Since school board members across many districts are elected officials, their
names and contact information are available through public ballot databases.
We collaborate with the XQ Institute222https://xqsuperschool.org/. to obtain
this information for over 115k elected officials across 13,233 school
districts—which constitute nearly all regular school districts across the
US333https://nces.ed.gov/programs/digest/d20/tables/dt20_214.10.asp.. Next,
via the elementary school attendance boundaries and school enrollment data
used in (Gillani et al., 2023), we identify the 1,000 school districts with
the largest elementary school enrollment that are also listed as having closed
enrollment (i.e., entirely boundary-assigned) elementary schools as our main
sample for outreach. Merging the district board member data onto this dataset
yields approximately 6,664 board members, 4,831 of which we have a valid email
address for. We then design and send emails using the email marketing platform
Mailchimp. Emails contain links unique to each recipient, and clicking the
link takes the recipient to a preliminary page on the dashboard that informs
them of the research study. The recipient is then invited to click through to
the main dashboard, where they are shown the dashboard depicted in Figure 1
with data specific to their district pre-selected.
As described in the Introduction by Research Question 1, we are interested in
better understanding which messaging techniques are more or less effective in
prompting school district leaders (board members) to explore and share
feedback on potential attendance boundary changes for school diversity. One
way to do this is to randomly assign board members to receive different
emails—e.g., emails with different subject lines—and then observe which ones
generate more engagement and eventual traffic to the data dashboard (we use
“email” and “subject line” below interchangeably, since email previews and
bodies were generally similar across runs and only adapted to reflect their
corresponding subject lines).
Randomly assigning board members within the same district to different email
conditions, however, runs the risk of treatment “spillover”—for example, if
one board member forwards an email to one of their colleagues—which would
limit our ability to identify the causal effect of any particular email on
engagement (Gerber and Green, 2012). To try and mitigate against this, we
design a series of cluster-randomized control trials, where we randomly assign
email conditions at the cluster (school district) level. There is still the
possibility that school board members across districts who are assigned to
different conditions forward received emails to one another, but we expect
this to be small, given their specific contexts (boundaries) are likely to be
different. To account for the fact that board members might respond
differently depending on the extent to which demographic imbalances are
actually an issue in their particular districts, we stratify districts into
quintiles according to their White/non-White dissimilarity indices, and
randomly assign those within each quintile to each possible email condition.
This increases the likelihood that districts receiving any of the email
conditions are well-balanced across the levels of segregation manifesting
across their schools.
We conduct these cluster randomized control trials in eight successive waves.
At the start of each wave, we begin with a set of hypotheses for which types
of subject lines would be most effective in driving open and click-through
rates. After each wave, we analyze engagement data (opens and click-throughs
to the dashboard) for each email. The next wave then contains emails informed
by well-performing ones in prior waves, plus those evaluating new hypotheses.
We also generate new hypotheses—based on observations and informal interviews
conducted with school board members who engaged in previous waves—to inform
the development of new subject lines. In the results section, we discuss some
of these interviews and the insights that emerged from them. Table 1 contains
some of our hypotheses and corresponding subject lines, with a full list
deferred to Appendix A. In total, we explored 14 subject lines across 17
campaigns, which were sent through 8 iterative waves. The median number of
recipients across campaigns was 205. Our final email included a subject line
that included aspects of well-performing emails in prior waves; we sent this
to 1,383 participants that remained in our total sample after running all
previous subject line experiments.
We note that due to the exploratory and iterative nature of this study, we did
not formally pre-register it in an Open Science repository. We acknowledge
this as a limitation of this work, as even exploratory studies and sequential
experimentation can be accommodated in existing pre-registration frameworks
(Nosek et al., 2018), and intend to rectify this in future studies of this
type.
Email subject line | Hypothesis
---|---
$20 for your views on diversifying attendance boundaries | A small incentive that board members can allocate to a teacher of their choice may increase engagement
Data science to improve learning experiences | Using a popular term (“data science”) and focusing on learning instead of diversity may increase engagement
More diversity with shorter commutes? | Stating the main finding of (Gillani et al., 2023) up front may increase engagement
Data science to decide which schools to open or close | Framing in terms of a problem school districts often face that triggers boundary planning may increase engagement
Table 1. Example email subject lines and corresponding hypotheses motivating
their inclusion across different email waves.
We use logistic regression with cluster robust standard errors to analyze the
causal effect of each subject line on open and click-through rates. The use of
cluster robust standard errors is typical in such types of experimental
designs and generally meant to account for correlations between participants
(board members) in the same cluster (district) that may impact results
(Cameron and Miller, 2015). We control for several variables that may also
affect the propensity for board members to engage with such outreach. The
regression specification is as follows:
$y^{o}_{i}=\beta_{0}+\beta_{1}\cdot S_{i}+\beta_{2}\cdot
R_{i}+\Sigma^{8}_{j=3}~{}\beta_{j}\cdot
D^{j}_{i}+\Sigma^{10}_{j=9}~{}\beta{j}\cdot C^{j}_{i}+\epsilon_{i}$
Where $i$ represents each recipient; $y^{o}_{i}$ represents the binary result
for each outcome $o\in\\{opened,clicked\\}$; $S_{i}$ represents the subject
line (campaign) randomly assigned to $i$; $R_{i}$ is an estimate of $i$’s
race/ethnicity using the rethnicity package in R (Xie, 2022); $D_{i}^{3:8}$
represent a set of district-level characteristics for the district $i$ belongs
to; $C_{i}^{9:10}$ represent campaign-specific variables—namely, the day of
the week and month (May or June) that the campaign was sent during, given
potential impacts of week and school year-level seasonalities on engagement;
$\beta_{0}$ is the intercept; and $\epsilon_{i}$ is the error term. In
particular, the variables $D_{i}^{3:8}$ represent: an indicator describing
whether the school board is whiter than the district, suggesting a lack of
racial representativeness in the district leadership; the district’s
White/non-White dissimilarity across elementary schools; the percentage of
students in the district who are White; the district’s urbanicity (urban,
suburban, small city, rural); the number of elementary students enrolled
across its schools; and its total number of elementary schools. Once the model
is fit, the coefficient $\beta_{1}$ indicates the causal effect of the
particular email subject line on open and click-through rates, and the other
coefficients indicate associations between various recipient, district, and
campaign-level variables and these outcomes measures.
### 3.3. Survey
We included an optional survey linked off of the data dashboard to invite
leaders and community members to share their feedback on the depicted
rezonings. The main objective of the survey was to help answer Research
Question 2, namely: “How do school district leaders and community members
perceive these boundary changes—namely, what potential merits and pitfalls do
they see in both how these boundaries were produced and their potential
viability in practice?”. Survey questions assessed respondents’ interest in
the topic of boundary planning; their views of the algorithms that produced
the hypothetical boundary scenarios; what they found interesting about the
depicted scenarios; what they found concerning about the depicted scenarios;
and space to leave any other questions or comments. A full list of questions
can be found in Appendix B.
## 4\. Results
### 4.1. Email campaign
We begin with a few descriptive statistics about the recipients and email
campaigns before exploring results from our regressions. Out of 4,831 board
members included in our sample, emails were delivered to nearly 90% (4,320);
approximately 10% of emails bounced (due to email address inaccuracies, being
classified as spam, or other reasons) and were not delivered to the intended
inbox. Two-thirds of school districts in our sample had school boards with a
Whiter racial composition than their elementary school student body. We caveat
this finding given potential biases and inaccuracies in race classification
(Lockhart et al., 2023) (the rethnicity package has an overall f1 score of
0.78 (Xie, 2022)). To the extent it is correct, however, it reflects a broader
trend of school boards failing to reflect the racial and ethnic compositions
of ever-diversifying student and family compositions (Samuels, 2020).
In terms of email engagement, 39% of recipients (1,686) opened the email they
received, but only 2.5% (106) clicked through to the dashboard. Analyzing log
data on the dashboard revealed that the vast majority of these click-throughs
did not make it past the IRB-required text included on the landing page that
describes the research study and data (see Appendix C for this text). Finally,
just under 1.5% of recipients (64) unsubscribed from receiving future emails.
Benchmarking open, click-through, and unsubscribe rates is generally difficult
as they may vary across industries, purpose/type of campaign, and many other
variables. Nevertheless, using average estimates produced by Mailchimp—the
platform we used for our email campaigns—an open rate of 39% vastly exceeds
their reported average of 23% 444https://mailchimp.com/resources/email-
marketing-benchmarks/.. Our observed 2.5% click-through rate, however, is
slightly below their reported average of 2.9%. Finally, while our 1.5%
unsubscribe rate is small in an absolute sense, it is an order of magnitude
higher than their average of 0.25%. We caveat these Mailchimp benchmark
numbers as only rough guideposts, given they represent averages over a wide
range of campaigns and audiences. Interestingly, our observed open rate is
nearly identical to the 40% open rate observed in (Delevoye, 2021), which
involved outreach to school board members during COVID, though their click-
through rate of 7% is much larger than ours. This may be due to the fact that
their call to action involved reading a memo sharing information about school
transportation during an evolving pandemic schooling context, while ours
contained an invitation to explore findings pertaining to a much more
politically contentious, and temporally less-pressing, issue.
Next, we analyze the results of our regressions to investigate Research
Question 1 (“Which messaging and outreach strategies are more or less
effective?”). First, we analyze the causal effect of different email subject
lines on the likelihood that recipients open our outreach emails, controlling
for the additional individual, district, and campaign-level variables
described in the Methods section. Figure 2(a) shows that campaigns with the
subject line “Data Science To Decide Which Schools To Open Or Close” and “Data
Science To Diversify Learning and Reduce Commuting Times” are significantly
more likely to be opened than a control email subject line that reads
“Diversify Learning”—selected for its simplicity. This is particularly
interesting because these subject lines were informed by conversations with
school board members in between campaign waves. After observing low dashboard
engagement following a few email campaign waves, we reached out to some of the
board members who were included in the outreach to see if they might be
willing to have a short conversation to share their views and feedback on our
project. We informally spoke with three board members across three different
districts in two states. The conversations were enlightening in many ways, but
a few key takeaways emerged. In one conversation, a school board member seemed
surprised that it might be possible to redraw attendance boundaries in ways
that might reduce segregation and travel times simultaneously, and believed
others may find this to be surprising as well. This motivated us to design the
“Data Science To Diversify Learning and Reduce Commuting Times” subject line
in an effort to create more of a “hook” to engage future recipients. In
another conversation, a school board member described the capacity planning
challenges they faced as their district’s population rapidly changed with
immigration. The board member described the need to determine which schools to
open or close (and where) as the main impetus for their ongoing boundary
planning efforts. Even though their board cared about school diversity, on its
own, the member highlighted it probably would not spark new boundary analyses.
The board member mentioned, however, that diversity might be an added
consideration as a part of such ongoing boundary planning efforts, and that
our tools and algorithms may help support such planning efforts by dovetailing
issues of diversity with issues that are of more pressing concern for
districts. This prompted us to design the subject line “Data Science To Decide
Which Schools To Open Or Close”, in an attempt to foreground and lead with a
concern that may be more top-of-mind (and less controversial) for school board
members than issues of segregation and diversity.
The results in Figure 2(a) also show that the higher the percentage of
elementary school student body in the district that is white, the more likely
they are to open our emails (regardless of campaign), and that suburban
district board members are slightly less likely to open our emails than those
in urban districts—in both cases, after controlling for all other variables in
the regression.
Figure 2. Associations between different regression variables and likelihood
of board members opening and clicking emails, respectively. The “Subject: …”
variables represent different email campaigns and the depicted values can be
interpreted as the magnitude of the causal effect of each campaign on open and
click outcomes, since participants were randomly assigned to campaigns.
Circles represent average associations (i.e., regression coefficient values,
which indicate log odds), and lines represent 95% confidence intervals.
Intervals that do not intersect zero indicate a statistically significant
association between that variable and its depicted outcome at an $\alpha$
level of 0.05. The results for categorical variables indicate the magnitude of
associations relative to a particular reference category: for BM Race, this is
“White”; for Urbanicity, it is “urban”; for day of week email was sent, it is
“Monday”; for month email was sent, it is “May”; and for the campaigns, it is
the control subject line of “Diversify Learning”.
Figure 2(b) shows results for email click-throughs to the dashboard (N=4,320;
results look virtually identical when conditioning on those who opened the
email). Despite the low number of click-throughs, the results suggest several
associations between email campaigns, individual/ district/ campaign-level
variables and click activity. For one, after controlling for the same
recipient and district-level characteristics as before, two campaigns offering
$20 to board members for their input on depicted boundaries generate lower
click-through rates than the control campaign with a subject line of
“Diversify Learning” and no incentive. While the body of the email specifies
that the $20 is meant to be donated to a school/teacher of their choosing, it
is possible that board members are weary of clicking through for fear of being
perceived as accepting a “gift”—a politically precarious practice—for their
participation. Mentioning “attendance boundaries” in the subject line and body
of the email also tends to generate lower click-through rates—perhaps because
of how controversial and unpopular the topic is555As we sought his feedback on
the dashboard, one former superintendent of a large suburban school district
shared that changing attendance boundaries is so unpopular that it often costs
school board members their jobs / opportunities for re-election. On the other
hand, the emails that produce higher open rates (as described earlier) also
appear to produce higher click-through rates.
Importantly, we highlight these results as exploratory and directional, and
far from conclusive. Given the number of variables in our regressions, there
is the possibility of an inflated Type I error rate (i.e., the likelihood of
detecting significant associations purely by chance) due to multiple
comparisons. Indeed, conservatively applying Bonferonni correction to correct
for the number of inferred coefficients in the email opens model renders all
predictors insignificant at $\alpha=0.05$. Predictors remain significant in
the email clicks model even after correction, but noting the small number of
overall clicks, we still encourage readers to interpret these results as
preliminary. With these caveats in mind, we observe district leaders’
engagement to be sensitive to how our outreach is framed and contextualized:
district leaders appear more likely to open and click through on emails that
use less controversial language that also appears to align with their a priori
student assignment and planning objectives. While these findings are
interesting and may inform future outreach efforts for this and other
projects, we note that the limited click-through rates coupled with even lower
rates of dashboard exploration due to large dropoff at the IRB study
description text render these cold email campaigns as ineffective channels for
engaging school districts in considering educational data science-informed
policies for fostering more diverse schools.
### 4.2. Media coverage
Observing low click-through rates and engagement with the dashboard, we turned
to another channel for reaching school districts: media coverage. There is a
vast literature on media effects, including how media coverage can shape
public discourse and opinion across a range of topics (King et al., 2017). We
reached out to an education journalist at the Hechinger Report, an outlet
covering Educational Inequality and Innovation, for advice on how to gain more
attention and engagement from district leaders. Our outreach was “almost
cold”: we had only previously spoken with the journalist once, almost 18
months earlier, about a separate research project. After demoing our dashboard
to the journalist, they decided to write a story describing the tool and how
the boundary scenarios were produced (Barshay, 2022). Stumbling across this
piece one week later, another journalist—a host of Georgia NPR’s All Things
Considered show—also published a piece highlighting the tool and how districts
might use it to inform their own diversity planning efforts (Biello, 2022).
We analyzed engagement with the dashboard in the 5 weeks following the first
story’s release to better understand how levels of exploration differ across
different types of districts represented on the dashboard. During this time,
the dashboard received over 3,700 visits, which contributed to the exploration
of boundary results for over 500 districts across 42 states. Given extremely
low levels of engagement with the dashboard prior to these stories, we believe
it is safe to attribute increased engagement with the dashboard during this
time period to the traffic that the stories likely drove.
Figure 3. Outputs of a negative binomial regression depicting associations
between the number of times a particular district is explored on the dashboard
and various district-level characteristics.
Figure 3 shows the associations inferred from a negative binomial regression
modeling the number of explorations per district on the dashboard as a
function of various district-level characteristics. Interestingly, urban
districts were more likely to be explored than suburban, small city, or rural
ones. Furthermore, larger districts (defined in terms of the number of
elementary schools in our data for them) were also more likely to be
explored—likely because they have larger populations overall. Perhaps most
notably, the strongest predictor of whether or not a district would be
explored, after controlling for all others, was its level of White/non-White
segregation: more segregated districts were more likely to be explored. This
suggests that media coverage was effective in drawing attention to those
districts where there is, arguably, great scope for change.
Given the broad nature of media coverage, it is likely that both district
leaders and community members visited and engaged with the dashboard during
the period following the articles’ publishing. Furthermore, given the
anonymous nature of the media-driven dashboard visits, it is impossible to
know who explored the dashboard in which ways, or from which districts.
Nevertheless, as the next section describes, a small set of people from both
sets of stakeholders took time to explore and share their reflections and
feedback on the premise and outputs of the boundary simulations.
### 4.3. Survey responses
In the months following the email campaigns and media coverage (from May
through October 2022), we collected approximately 11 responses via the
feedback survey linked off of the dashboard. Two survey responses came from
email recipients; the remainder, we believe, came from both district leaders
and community members who learned about the dashboard via the published
articles. Given that thousands of individuals explored the dashboard, 11
survey responses constitutes an extremely small—and hence, likely very
biased—sample of total possible respondents. Noting this selection bias, we
caution against making generalized claims from the input. Nevertheless, like
interviews, we believe the qualitative insights provided by these survey
responses shed invaluable light onto Research Question 2—“How do school
district leaders and community members perceive these boundary changes—namely,
what potential merits and pitfalls do they see in both how these boundaries
were produced and their potential viability in practice?”.
When asked how happy they would be with the depicted rezonings, approximately
an equal share of respondents expressed they would be somewhat/very happy and
somewhat/very unhappy (three each), with two responding “neutral”.
Interestingly, when asked how happy they believed other families would be with
the depicted rezonings, only one respondent indicated “somewhat happy”; eight
indicated believing families would be somewhat or very unhappy. This reflects
potential selection biases in who filled out the survey (i.e., those who may
be more open to systemic, equity-promoting changes in schools) and underscores
the controversial nature of boundary changes across many school districts. It
may also point to potential clashes between student assignment models and the
values of certain subsets of families (Robertson et al., 2021). When asked how
much they trusted the computer program that produced the rezonings, five
respondents indicated they sort of trusted it or trusted it completely, while
three indicated they don’t trust it, and two were unsure. While these
responses are mixed, it is interesting—and perhaps concerning—to see such high
levels of trust, especially since the dashboard provided no insights into how
the algorithm was working or which computational methods were applied to
produce the rezonings (though the media coverage offered a high-level
description). Only one out of 9 respondents disagreed that changing attendance
boundaries is a valuable strategy for promoting more diverse schools. Finally,
when asked how they felt about the idea of redrawing attendance boundaries to
increase diversity in their district’s schools after learning about this
project, only one respondent said they were less interested in the idea than
before. An approximately equal number indicated the same or more interest as
before (five and four, respectively), while one expressing same interest wrote
in that they felt more “…hopeful and inspired”.
Respondents’ open-ended feedback offered a number of rich insights. When asked
what they found surprising, interesting, or novel about the depicted
rezonings, several respondents shared interest in the possibility of fostering
more diverse schools without significant travel disruptions. For example, one
respondent commented that they felt “…Surprise that this could be done.
Surprise that the transportation impact is so small.” Similarly, another
respondent—a school district official–said that they “Love that it didn’t move
Children to Far from their old School.” Another school board member shared
that “I am interested to see that…there are improvements that could be made
without major changes in time/distance to schools”. One school board member
shared an optimistic view, commenting that “This is helpful information and a
worthy project, and I don’t want perfect to be the enemy of improvement. I
think there are practical solutions to improving racial diversity here.”
Yet respondents also had consistent suggestions for improvement. When asked
what they found troubling or concerning about the depicted rezonings, many
commented on the fact that the depicted changes to school demographics were
quite small. One school board member mentioned “The computer program only had
1 school out of 13 with a significant change which is not worth the headaches
rezoning causes. The other 12 barely changed if at all.” Another respondent
similarly questioned “Is it worth disrupting families for a small change.”
Several other respondents highlighted the dashboard’s failure to take into
account the nuances and details of their unique local context when depicting
hypothetical rezonings. One respondent described the complexity of redrawing
boundaries in the context of schools with special programs, sharing the
following: “I think the idea is great, but in a city like [redacted] there are
other complexities. We have several schools that are language immersion,
serving the local community. They are a 50/50 DLI [Dual Language Instruction]
model. This really influences how boundaries can move and what the actual
neighborhood space looks like. Additionally we have both k-5 and k-8 and this
changes how boundaries can move.” Another, a board member, commented on the
algorithm’s unrealistic school capacity constraints, highlighting that many of
the schools that the algorithm allocated more students to were already at or
over capacity (which is unknowable from the Department of Education’s (DOE)
Common Core of Data). Some described other shortcomings, like missing schools
(e.g. closed enrollment or newly built ones), a failure to account for
measures of community cohesion (like elementary-middle-high feeder patterns),
and other inaccuracies due to aggregated and dated DOE data.
Finally, we found it encouraging that several respondents shared their contact
information to continue the conversation and share additional feedback. One
board member wrote that they “…would like to chat with [us] about updating
[our] maps to include more accurate information about the schools”, ending
their note with the following question: “Are you able to talk with me
sometime???” One district official expressed interest in seeing results for
other schools, asking “When will the middle and high school data update”?
Finally, a community member wrote asking “How can I work to show this to my
county board and persuade them?” Despite the myriad of possible improvements
and iterations that would be required to make the depicted rezonings
applicable in practice, there appeared to be interest from at least a handful
of leaders and members across districts. We have followed up with respondents
to learn more about their unique contexts and are currently collaborating with
districts to translate the tools and methods from this research into practice.
## 5\. Discussion
Recall our original research questions:
1. (1)
Which messaging and outreach strategies are more or less effective in
prompting school district leaders and community members in exploring potential
boundary changes that could help foster more racially/ethnically integrated
schools?
2. (2)
How do school district leaders and community members perceive these boundary
changes—namely, what potential merits and pitfalls do they see in both how
these boundaries were produced and their potential viability in practice?
With respect to question 1, while we find that email campaigns using subject
lines that align diversity goals with other perennial district-level concerns
(like school capacity planning or travel times) yield larger open and click-
through rates, cold email outreach proves largely ineffective as a means for
prompting district leaders to explore hypothetical diversity-promoting
boundaries. Media coverage, however, generates much more
exploration—particularly among scenarios for districts that are more
segregated. We cannot say that media coverage is a more efficient method for
sharing educational data science research findings with districts; indeed, the
greater volume of exploration it prompts may simply be due to its wider reach.
Nevertheless, it is clear that in this particular setting, media coverage
resulting from (almost cold) outreach to an education journalist produced a
level of engagement that emails alone could not accomplish. We acknowledge
several limitations in our email campaign and analysis. For one, given the low
engagement volume (especially click-through rates) and differences in
engagement rates for emails with similar subject lines, it is important to
interpret all results as exploratory and suggestive. Replication efforts may
help illuminate the robustness of results, as might follow-on analyses that
account for other variables that could have introduced noise or otherwise
affected outcomes. While click-through rates were low, the high open rates
offer an interesting direction for future work. For example, instead of asking
education stakeholders to click through to a dashboard, the dashboard and/or
its most salient findings may be brought into the body of the email—increasing
the likelihood that recipients actually see it.
With respect to question 2, an important caveat of our largely survey
response-driven findings to this question is the extremely small number of
responses (11), especially when compared to the thousands of district data
explorations that occurred through the dashboard. We did not hear from the
vast majority of those who explored the dashboard, and so, cannot make general
claims about their perceptions of specific boundary changes. Nevertheless,
those who did respond generally indicated an interest in the idea of redrawing
school attendance boundaries to foster more diverse schools, and several asked
to have follow-on conversations to explore how these preliminary research
findings might be applied in their local settings. Many also had constructive
suggestions for improving the algorithms to make their outputs more useful in
practice. Admittedly, even though the response pool was small, given how
contentious and unpopular boundary changes are, we were surprised and
encouraged that anyone responded affirmatively at all. Our findings thus
suggest that there is hope for making progress on a problem as entrenched as
racial and ethnic segregation in schools, through means as controversial as
attendance boundary changes—even if just in a few school districts to begin
with. Indeed, sparking change in a few districts serving thousands or tens of
thousands of students can translate into scalable impacts on learning and
future life outcomes for students.
Looking ahead, we believe there are two under-explored opportunities for
educational data scientists to consider pursuing as they shape their research
agendas in the months and years ahead. The first involves conducting research
at the intersection of computation and education policy—particularly focusing
on systemic issues like segregation—in addition to the myriad of efforts
underway exploring the intersection of computation and learning. The growth of
educational data science practices like “learning engineering” (Thille, 2016)
offer new and exciting opportunities for using computational methods to
advance academic mastery, socioemotional understanding, and various other
cognitive and metacognitive skills. Yet learning is often shaped by political
and sociological forces before students even step foot in a school or
classroom. Through computational research projects like boundary modeling and
other policy-related efforts, educational data scientists can help study and
perhaps even shape these “upstream”, systemic forces that can silently, yet
powerfully, affect the opportunities afforded to students and families. The
second opportunity involves exploring which communication and engagement
strategies are most effective for translating research findings into practice.
Too often, the fruits of great research remain unrealized because the paths to
successful implementation are unclear. We believe there is a rich opportunity
for educational data scientists to explore intersections between their work
and “implementation science” in order to ensure that promising early research
is translated and further evaluated across different contexts. We hope our
study offers a useful, albeit preliminary, reference point for both of these
exciting future directions.
## 6\. Acknowledgements
Funding for this project was provided by the MIT Center for Constructive
Communication. The authors declare no competing interests. The research was
approved by MIT’s Institutional Review Board. We are grateful to Peter
Bergman, Kumar Chandra, Alvin Chang, Akeshia Craven-Howell, Rebecca Eynon, Tad
Hirsch, Melissa Krull, Justin Reich, Todd Rogers, Joshua Starr, district
leaders who responded to our outreach, and the anonymous reviewers for their
helpful comments and guidance on this project. We also thank Jill Barshay and
Peter Biello for their support in sharing the research through press channels.
Finally, we thank Lauren J. Bierbaum and Max Vargas at the XQ Institute for
sharing their input and enabling us to access public records for elected
school board members.
## References
* (1)
* Barshay (2022) Jill Barshay. 2022\. PROOF POINTS: Computer scientists create tool that can desegregate schools – and shorten bus routes. The Hechinger Report. https://hechingerreport.org/proof-points-computer-scientists-create-tool-that-can-desegregate-schools-and-shorten-bus-routes/
* Bauer et al. (2015) Mark S. Bauer, Laura Damschroder, Hildi Hagedorn, Jeffrey Smith, and Amy M. Kilbourne. 2015\. An introduction to implementation science for the non-specialist. _BMC Psychology_ 3, 32 (2015).
* Becker and Solomon (2020) Amariah Becker and Justin Solomon. 2020. Redistricting Algorithms. _arXiv: 2011.09504_ (2020).
* Bell et al. (2017) Alex M. Bell, Raj Chetty, Xavier Jaravel, Neviana Petkova, and John Van Reenen. 2017. Who Becomes an Inventor in America? The Importance of Exposure to Innovation. _NBER Working Paper No. 24062_ (2017).
* Biello (2022) Peter Biello. 2022\. Georgia Tech, MIT researchers develop tool for increasing racial integration in schools. All Things Considered, Georgia Public Broadcasting. https://www.gpb.org/news/2022/07/12/georgia-tech-mit-researchers-develop-tool-for-increasing-racial-integration-in
* Billingham and Hunt (2016) Chase M. Billingham and Matthew O. Hunt. 2016. School Racial Composition and Parental Choice: New Evidence on the Preferences of White Parents in the United States. _Sociology of Education_ 89, 2 (2016).
* Black (1999) Sandra E. Black. 1999\. Do Better Schools Matter? Parental Valuation of Elementary Education. _The Quarterly Journal of Economics_ 114, 2 (1999), 577–599.
* Bridges (2016) Kim Bridges. 2016\. Eden Prairie Public Schools: Adapting to Demographic Change in the Suburbs. The Century Foundation. https://tcf.org/content/report/eden-prairie-public-schools/
* Burton and Welsh (2015) Linda M. Burton and Whitney Welsh. 2015. Inequality and Opportunity: The Role of Exclusion, Social Capital, and Generic Social Processes in Upward Mobility. WT Grant Foundation. https://wtgrantfoundation.org/library/uploads/2016/01/Inequality-and-Opportunity-Burton-and-Welsh-William-T.-Grant-Foundation.pdf
* Cameron and Miller (2015) A. Colin Cameron and Douglas L. Miller. 2015. A Practitioner’s Guide to Cluster-Robust Inference. _The Journal of Human Resources_ 50, 2 (2015), 317–372.
* Campos and Kearns (2022) Christopher Campos and Caitlin Kearns. 2022. The Impact of Neighborhood School Choice: Evidence from Los Angeles’ Zones of Choice. _SSRN Working Paper_ (2022).
* Candipan (2019) Jennifer Candipan. 2019\. Neighbourhood change and the neighbourhood-school gap. _Urban Studies_ 56, 15 (2019).
* Castro et al. (2022) Andrene J. Castro, Genevieve Siegel-Hawley, Kim Bridges, and Shenita E. Williams. 2022\. Narratives of Race in School Rezoning: How the Politics of Whiteness Shape Belonging, Leadership Decisions, and School Attendance Boundaries. _AERA Open_ (2022).
* Chetty et al. (2022) Raj Chetty, Matthew O. Jackson, Theresa Kuchler, and Johannes et al. Stroebel. 2022. Social capital II: determinants of economic connectedness. _Nature_ 608 (2022), 122–134.
* Delevoye (2021) Angele Delevoye. 2021\. School boards in 2020: deliberation and decision-making in a crisis. _Working Paper_ (2021). https://drive.google.com/file/d/1-2rcwRetEJ6J8QeSb8JWF4hb4noWUuN5/view
* Delmont (2016) Matthew F. Delmont. 2016\. _Why Busing Failed: Race, Media, and the National Resistance to School Desegregation_. University of California Press.
* EdBuild (2020) EdBuild. 2020. Fault Lines: America’s Most Segregating School District Borders. EdBuild. https://edbuild.org/content/fault-lines/full-report.pdf
* Ericsson and Simon (1984) K. Anders Ericsson and Herbert A. Simon. 1984. _Protocol analysis: Verbal reports as data_. MIT Press.
* Fiel (2013) Jeremy E. Fiel. 2013\. Decomposing School Resegregation: Social Closure, Racial Imbalance, and Racial Isolation. _American Sociological Review_ 78, 5 (2013).
* Foundation (2022) WT Grant Foundation. 2022\. Improving the Use of Research Evidence. WT Grant Foundation. https://wtgrantfoundation.org/focus-areas/improving-use-research-evidence
* Frankenberg and Jacobsen (2011) Erica Frankenberg and Rebecca Jacobsen. 2011. The Polls–Trends: School Integration Polls. _Public Opinion Quarterly_ 75, 4 (2011), 788–811.
* Frankenberg et al. (2003) Erica Frankenberg, Chungmei Lee, and Gary Orfield. 2003\. A Multiracial Society with Segregated Schools: Are We Losing the Dream? The Civil Rights Project, Harvard University. https://civilrightsproject.ucla.edu/research/k-12-education/integration-and-diversity/a-multiracial-society-with-segregated-schools-are-we-losing-the-dream/frankenberg-multiracial-society-losing-the-dream.pdf
* Gerber and Green (2012) Alan S. Gerber and Donald Green. 2012. _Field Experiments: Design, Analysis, and Interpretation_. New York: W. W. Norton & Company.
* Gillani et al. (2023) Nabeel Gillani, Doug Beeferman, Christine Vega-Pourheydarian, Cassandra Overney, Pascal Van Hentenryck, and Deb Roy. 2023. Redrawing attendance boundaries to promote racial and ethnic diversity in elementary schools. _Forthcoming in Educational Researcher_ (2023). https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4387899
* Gillani and Eynon (2014) Nabeel Gillani and Rebecca Eynon. 2014. Communication patterns in massively open online courses. _The Internet and Higher Education_ 23 (2014), 18–26.
* Gurnee and Shmoys (2021) Wes Gurnee and David B. Shmoys. 2021. Fairmandering: A column generation heuristic for fairness-optimized political districting. _arXiv: 2103.11469_ (2021).
* Hailey (2021) Chantal A. Hailey. 2021\. Racial Preferences for Schools: Evidence from an Experiment with White, Black, Latinx, and Asian Parents and Students. _Sociology of Education_ (2021).
* Hall and Hibel (2017) Matthew Hall and Jacob Hibel. 2017. Latino Students and White Migration from School Districts, 1980-2010. _Social Problems_ 64, 4 (2017), 457–475.
* Harris (2022) Adam Harris. 2022\. School Boards are No Match for America’s Political Dysfunction. The Atlantic. https://www.theatlantic.com/politics/archive/2022/04/parents-partisan-school-board-meeting-covid/629669/
* Hasan and Kumar (2019) S. Hasan and A. Kumar. 2019\. Digitization and Divergence: Online School Ratings and Segregation in America. SSRN. https://ssrn.com/abstract=3265316
* Johnson (2011) Rucker C. Johnson. 2011\. Long-run Impacts of School Desegregation & School Quality on Adult Attainments. _NBER Working Paper No. 16664_ (2011).
* Kane et al. (2005) Thomas J. Kane, Douglas O. Staiger, and Stephanie K. Riegg. 2005\. School Quality, Neighborhoods and Housing Prices: The Impacts of school Desegregation. _NBER Working Paper No. 11347_ (2005).
* King et al. (2017) Gary King, Benjamin Schneer, and Ariel White. 2017\. How the news media activate public expression and influence national agendas. _Science_ 358, 6364 (2017), 776–780.
* Kizilcec et al. (2013) Rene F. Kizilcec, Chris Piech, and Emily Schneider. 2013\. Deconstructing disengagement: analyzing learner subpopulations in massive open online courses. In _Third International Conference on Learning Analytics and Knowledge_. 170–179.
* Lockhart et al. (2023) Jeffrey W. Lockhart, Molly M. King, and Chris Munsch. 2023\. Name-based demographic inference and the unequal distribution of misrecognition. _Nature Human Behaviour_ (2023).
* Macartney and Singleton (2018) Hugh Macartney and John D. Singleton. 2018. School boards and student segregation. _Journal of Public Economics_ 164 (2018), 165–182.
* Massey and Denton (1988) D. S. Massey and N. A. Denton. 1988. The Dimensions of Residential Segregation. _Social Forces_ 67, 2 (1988), 281–315.
* Mayer (2014) Susan Mayer. 2014\. How Economic Segregation Affects Children’s Educational Attainment. _Annual Review of Sociology_ 40 (2014), 199–218.
* McFarland et al. (2021) Daniel A. McFarland, Saurabh Khanna, Benjamin W. Domingue, and Zachary A. Pardos. 2021. Education Data Science: Past, Present, Future. _AERA Open_ (2021).
* McMillan (2018) Susan M. McMillan. 2018\. Common Practices in Changing School Attendance Zone Boundaries. Educational Data Systems. https://eddata.com/wp-content/uploads/2018/10/Common-Practices-in-Changing-School-Attendance-Zone-Boundaries.pdf
* Mobarak (2022) Ahmed Mushfiq Mobarak. 2022\. Assessing social aid: the scale-up process needs evidence, too. _Nature Comment_ (2022). https://www.nature.com/articles/d41586-022-03039-2
* Moir (2018) Taryn Moir. 2018\. Why Is Implementation Science Important for Intervention Design and Evaluation Within Educational Settings? _Frontiers in Education_ (2018).
* Monarrez and Chien (2021) Tomas Monarrez and Carina Chien. 2021. Dividing Lines: Racially Unequal School Boundaries in US Public School Systems. The Urban Institute. https://www.urban.org/sites/default/files/publication/104736/dividing-lines-racially-unequal-school-boundaries-in-us-public-school-systems.pdf
* Monarrez et al. (2022) Tomas Monarrez, Brian Kisida, and Matthew Chingos. 2022\. The Effect of Charter Schools on School Segregation. _American Economic Journal: Economic Policy_ 14, 1 (2022), 1–42.
* Nakajima (2021) Nozomi Nakajima. 2021\. Evidence-Based Decisions and Education Policymakers. _Working Paper_ (2021). https://nozominakajima.github.io/files/nakajima_policymaker.pdf
* Natanson (2021) Hannah Natanson. 2021\. How and why Loudoun County became the face of the nation’s culture wars. The Washington Post. https://www.washingtonpost.com/local/education/loudoun-critical-race-theory-transgender-rights/2021/07/05/3dab01b8-d4eb-11eb-ae54-515e2f63d37d_story.html
* Nosek et al. (2018) Brian A. Nosek, Charles R. Ebersole, Alexander C. DeHaven, and David T. Mellor. 2018. The preregistration revolution. _Proceedings of the National Academy of Sciences_ 115, 11 (2018), 2600–2606.
* of Education (2021) US Department of Education. 2021\. Fast Facts: Public school choice programs. Digest of Education Statistics, 2019 (NCES 2021-009). https://nces.ed.gov/fastfacts/display.asp?id=6
* Office (2022) U.S. Government Accountability Office. 2022\. K-12 Education: Student Population Has Significantly Diversified, but Many Schools Remain Divided Along Racial, Ethnic, and Economic Lines. U.S. Government Accountability Office. https://www.gao.gov/assets/gao-22-104737.pdf
* OR-Tools (2022) Google OR-Tools. 2022\. CP-SAT Solver. https://developers.google.com/optimization/cp/cp_solver
* Orfield and Lee (2005) Gary Orfield and Chungmei Lee. 2005. Why Segregation Matters: Poverty and Educational Inequality. The Civil Rights Project, Harvard University. https://civilrightsproject.ucla.edu/research/k-12-education/integration-and-diversity/why-segregation-matters-poverty-and-educational-inequality/orfield-why-segregation-matters-2005.pdf
* Owens (2018) Ann Owens. 2018\. Income Segregation between School Districts and Inequality in Students’ Achievement. _Sociology of Education_ 91, 1 (2018).
* Perron and Didier (2020) Laurent Perron and Frederic Didier. 2020. CPAIOR 2020 Master Class: Constraint Programming. Youtube. https://youtu.be/lmy1ddn4cyw
* Piech et al. (2015) Chris Piech, Jonathan Bassen, Jonathan Huang, Surya Ganguli, Mehran Sahami, Leonidas Guibas, and Jascha Sohl-Dickstein. 2015. Deep Knowledge Tracing. _Advances in Neural Information Processing Systems_ 28 (2015).
* Quick (2016) Kimberly Quick. 2016\. Chicago Public Schools: Ensuring Diversity in Selective Enrollment and Magnet Schools. The Century Foundation. https://tcf.org/content/report/chicago-public-schools/
* Reardon and Owens (2014) Sean Reardon and Ann Owens. 2014. 60 Years After Brown: Trends and Consequences of School Segregation. _Annual Review of Sociology_ 40 (2014), 199–218.
* Reardon and Bischoff (2011) Sean F. Reardon and K. Bischoff. 2011. Income Inequality and Income Segregation. _American Journal of Sociology_ 116, 4 (2011), 1092–1153.
* Reardon et al. (2018) Sean F. Reardon, D. Kalogrides, and K. Shores. 2018\. The Geography of Racial/Ethnic Test Score Gaps. _Stanford Center for Education Policy Analysis, Working Paper No. 16-10_ (2018).
* Reber (2005) Sarah J. Reber. 2005\. Court-Ordered Desegregation: Successes and Failures Integrating American Schools since Brown versus Board of Education. _Journal of Human Resources_ 40, 3 (2005), 559–590.
* Ritter et al. (2007) Steven Ritter, Jonna Kulikowich, Pui-Wa Lei, and Christy Mcguire. 2007. What Evidence Matters? A Randomized Field Trial of Cognitive Tutor Algebra I. _Frontiers in Artificial Intelligence and Applications_ 162, 13 (2007).
* Robertson et al. (2021) Samantha Robertson, Tonya Nguyen, and Niloufar Salehi. 2021\. Modeling Assumptions Clash with the Real World: Transparency, Equity, and Community Challenges for Student Assignment Algorithms. In _CHI ’21_.
* Samuels (2020) Christina A. Samuels. 2020\. Why School Board Diversity Matters. EdWeek. https://www.edweek.org/leadership/why-school-board-diversity-matters/2020/11
* Sawchuck (2021) Stephen Sawchuck. 2021\. Why School Boards Are Now Hot Spots for Nasty Politics. EdWeek. https://www.edweek.org/leadership/why-school-boards-are-now-hot-spots-for-nasty-politics/2021/07
* Small (2009) Mario L. Small. 2009\. _Unanticipated Gains: The Origins of Network Inequality in Everyday Life_. Oxford University Press.
* Staff (2019) Baltimore Sun Staff. 2019\. In Howard County, a ‘courageous’ plan to redraw school boundaries tests community’s commitment to diversity. The Baltimore Sun. https://www.baltimoresun.com/education/bs-md-howard-school-redistricting-20190906-xhzkmkf2zvgcxdkbd3vqdanblm-story.html
* Thille (2016) Candace M. Thille. 2016\. Bridging Learning Research and Teaching Practice for the Public Good: The Learning Engineer. TIAA Institute. https://origin-www.tiaainstitute.org/sites/default/files/presentations/2017-02/bridging_learning_research_and_teaching_practice.pdf
* Van Hentenryck (1989) Pascal Van Hentenryck. 1989\. _Constraint Satisfaction in Logic Programming_. MIT Press.
* Whitehurt (2017) Grover J. Whitehurt. 2017\. New evidence on school choice and racially segregated schools. https://www.brookings.edu/research/new-evidence-on-school-choice-and-racially-segregated-schools/
* Xie (2022) Fangzhou Xie. 2022\. rethnicity: An R package for predicting ethnicity from names. _SoftwareX_ 17 (2022), 317–372.
* Yeung (2022) Peter Yeung. 2022\. What Happened When France Sent Low-Income Kids to Wealthy Schools. The 74 Million. https://www.the74million.org/article/what-happened-when-france-sent-low-income-kids-to-wealthy-schools/
## Appendix A Email subject lines and corresponding hypotheses
Table 2 shows the different subject lines, motivating hypotheses, and waves
these emails were sent during.
Email subject line | Hypothesis | Wave(s)
---|---|---
Diversify Learning | Concise subject line that highlights project topic | Test, 1
Diversify Learning and get $20 for a school | A small incentive that board members can allocate to a teacher of their choice may increase engagement | 1
Data science to diversify attendance boundaries | “Data science” might pique interest, especially in the context of boundary planning | 2
$20 for your views on diversifying attendance boundaries | A small incentive that board members can allocate to a teacher of their choice may increase engagement | 2
Data science to improve learning experiences | “Data science”, focusing on learning instead of diversity may increase engagement | 2
Data science to diversify learning experiences | Same as above, but mentioning diversity and learning together | 2, 5
More diversity with shorter commutes? | Stating the main finding of (Gillani et al., 2023) up front may increase engagement | 3, 5
More diversity with shorter commutes? $20 for your thoughts! | Stating the main finding of (Gillani et al., 2023) up front and adding the incentive may increase engagement | 3
More diversity, less driving? | Pithy description of findings from (Gillani et al., 2023) | 4
2,000 pennies for your thoughts | Pithy version of incentive email | 4
Improve discussions about school diversity with families | Heard from districts that they find it difficult to talk with families about diversifying schools | 5
Data science to diversify learning and reduce commuting times | Combing learning, diversity, and shorter commute messages from earlier | 6
Data science to decide which schools to open or close | Framing in terms of a problem school districts often face that triggers boundary planning may increase engagement | 6
Data science to diversify learning and reduce commute times | Has elements of promising subject lines throughout runs | Final
Table 2. Email subject lines and corresponding hypotheses motivating their
inclusion across different email waves.
## Appendix B Dashboard survey questions
The optional survey linked off of the dashboard included the following
questions:
* •
How happy would you be with this rezoning? [Very happy / Somewhat happy /
Neutral / Somewhat unhappy / Very unhappy / Other:]
* •
How happy do you believe other families would be with this rezoning? [Very
happy / Somewhat happy / Neutral / Somewhat unhappy / Very unhappy / Other:]
* •
How much do you trust the computer program that created this rezoning? [I
trust it completely / I sort of trust it / I’m not sure if I trust it or not /
I don’t trust it / Other:]
* •
After learning about this project, how do you feel about the idea of redrawing
attendance boundaries to increase diversity in your district’s schools? [More
interested in the idea than before / Less interested in the idea than before /
About the same level of interest as before / Other:]
* •
How much do you agree with the following statement: ”I believe changing
attendance boundaries is a valuable strategy for promoting more diverse
schools in my district” [Strongly agree / Agree / Neutral / Disagree /
Strongly disagree / Other:]
* •
What’s something you found interesting, surprising, or novel about the
proposed rezoning, if anything? [Open-ended response]
* •
What’s something you found troubling or concerning about the proposed
rezoning, if anything? [Open-ended response]
* •
Do you have any other thoughts or reactions that you’d like to share? [Open-
ended response]
* •
Please share your contact information if you would like for a member of our
team to reach out to you to discuss this project further, including how it
might be applied to your school district.
## Appendix C Dashboard landing page IRB text
Email recipients who clicked through saw the following text on the dashboard
landing page.
> Hello! We are scientists at the Massachusetts Institute of Technology, and
> we’ve found that it may be possible to increase diversity across many
> districts’ schools without creating large disruptions for families. We
> invite school district leaders like yourself to explore these results and
> share your feedback through this website, which is part of a research study
> we are conducting in order to better understand how education leaders
> respond to different diversity policies. Exploring the dashboard is
> completely voluntary, and you can stop anytime. We don’t expect any
> discomforts from doing so, and hope you’ll find our results interesting and
> informative for your work. Additionally, there will be no public record or
> knowledge of your participation, so you can feel free to engage (or not
> engage!) in a way that’s best for you. If you have any questions or
> comments, feel free to email us at<EMAIL_ADDRESS>or contact
> MIT’s research office at<EMAIL_ADDRESS>Please click “Continue” to get
> started!
|
observer and source, respectively. The Einstein radius $\theta_{E}$ cannot be
obtained directly from the study of the magnification light curve of the
microlensing event181818In the rare case of perfect alignment
observer/lens/source it is possible to infer the angular Einstein radius
$\theta_{E}$ by using photometric information alone., hence other quantities
need to be defined. The only quantity that contains some information about the
physical properties of the event, especially about the lens itself, and can be
directly inferred from the light curve is instead the timescale $t_{E}$ of the
event, i.e. the time it takes the source to traverse an angular distance of
$\theta_{E}$ and can be written as follows:
$t_{E}=\frac{\theta_{E}}{\mu_{LS}}$ (XVI.3)
where $\mu_{LS}$ is the relative proper motion between lens and source. In a
microlensing event, $t_{E}$ can vary from a few days up to several hundreds
days. In long-duration microlensing events, the annual parallax tends to lead
to prominent departures in the photometric signature [1043, 1044], so that we
can also infer the microlensing parallax parameter $\pi_{E}$:
$\pi_{E}=\frac{\pi_{LS}}{\theta_{E}}$ (XVI.4)
From Eq. XVI.1 we know that the more massive the lens, the longer the duration
of the event itself. In this context, microlensing events involving massive
and compact objects such as BHs as lenses are the most likely to have
sufficient duration to detect the microlensing parallax effect using
photometry.
To obtain the mass of the lens, however, the photometric information is
unfortunately not sufficient, even when the parallax $\pi_{E}$ is known, as
$M_{L}$ and $\pi_{LS}$ are degenerate. The only way to disentangle these two
contributions and get an accurate estimate of the mass of the lens is by
monitoring and measuring the small astrometric shift of the position of the
source induced by the presence of the lens. This can be done using high-
precision astrometry from space, e.g. with HST (see an example later in the
text). Indeed, the astrometric shift ${\bf\delta(u)}$ is a two-dimensional
position angle which is proportional to the angular Einstein radius
$\theta_{E}$ according to the expression:
$\delta({\bf u})=\frac{{\bf u}}{u^{2}+2}\theta_{E}$ (XVI.5)
where the vector ${\bf u}$ defines the angular separation between lens and
source and it changes with time. By substituting Eq. XVI.1 in Eq. XVI.4, we
can derive an expression for the lens mass:
$M_{L}=\frac{\theta_{E}}{k\pi_{E}}$ (XVI.6)
where $\theta_{E}$ comes from precise, multi-epoch astrometry and $\pi_{E}$
comes from photometry. If the distance of the source is known, for example
because the microlensing event is detected in a specific region of the Galaxy
(e.g. the Galactic bulge, whose distance is known), then we can use this
information to constrain the distance of the lens as well, from Eq. XVI.2. If
this is not the case, we can still overcome the limitation if some
spectroscopic data of the source is available.
Searching for isolated stellar-mass BHs via gravitational microlensing is a
promising method as massive objects such as BHs are capable of producing
greater space-time distortions than any other medium-mass star, so they can be
distinguished. In addition, detailed Milky Way microlensing simulation codes
have recently demonstrated that when plotting different sources (stars, white
dwarfs, NS or BHs) in the $\pi_{E}$ \- $t_{E}$ plane, BHs tend to segregate in
a specific region of this parameter space, corresponding to long Einstein
crossing times and relatively low microlensing parallaxes (see Fig. 13 in
[1045]). This means that selecting microlensing events with such properties is
an ideal way to maximize the chances of detecting events involving BH lenses.
Figure 16: from Pietrzyński et al. [1046]: it illustrates the gravitational
microlensing produced by a BH. The deflection of the light of a background
star when it passes in front of a BH produces two images of the same star,
temporarily amplifying its brightness. These two images are too close to being
spatially resolved but the difference in brightness between them produces a
shift in the source’s position that can be measured using astrometry and can
provide strong constraints on the nature of the lens.
Over the past twenty years there has been a growing interest in microlensing
events and several microlensing survey programs have been designed for this
purpose, some of which are currently still in operation. The Optical
Gravitational Lensing Experiment (OGLE-IV; [1047, 1048]), the Expérience pour
la Recherche d’Objets Sombres (EROS; [1049]), the Microlensing Observations in
Astrophysics collaboration (MOA-II; [1050, 1051]), and the Korea Microlensing
Telescope Network (KMTNet; [1052]) are the best examples. These surveys carry
out photometric monitoring of several fields, both in the Milky Way (mostly
the Galactic bulge) and in Galactic satellites (for example the Magellanic
Clouds), allowing to construct light curves for an unlimited amount of
objects. Thanks to their precious work over many years, these surveys have
provided the scientific community not only with more than 30,000 microlensing
events, but also with a detailed and long list of classified variables, from
pulsators (e.g., RR-Lyrae or Cepheids) to eclipsing or ellipsoidal variables.
Unfortunately, it is more difficult to have accurate astrometric measurements
and it is only recently that people have started to have access to this type
of data, thanks to important international collaborations (e.g., HSTPROMO
[1053] which exploits high-resolution HST observations of globular clusters
across multiple epochs) or space missions (e.g. GAIA [1054], which instead
scans the sky multiple times to measure stars’ positions and proper
motions191919We note here that the astrometric power of GAIA is very limited
in crowded fields like in the direction of the Galactic bulge.), opening to
the concrete possibility of carrying out a search for isolated BHs through
astrometric microlensing in the Universe.
To date, only a few isolated BHs have claimed to be detected via this
technique, and interestingly, all of them have been found in the direction of
the Galactic bulge, where there is a higher rate of microlensing events due to
the high density of stars. It is interesting to note that, historically, all
the searches for isolated BHs have focused primarily on microlensing events
with $\pi_{E}$ $\sim$ 0.1 or higher, essentially excluding all those events we
now know are most likely produced by BHs. This could be one of the reasons for
the lack of isolated BH detections, despite the enormous observational effort
of the last few decades.
MACHO-96-BLG-5 and MACHO-98-BLG-6 are the first two microlensing events that
scientists have proposed to be linked to the presence of isolated BHs, with a
mass of the order of 6 $M_{\odot}$ [1055]. These events were first detected
and observed using long-term photometric monitoring [1056] as part of the
MACHO project, which used microlensing events to specifically search for
MAssive Compact Halo Objects [1057]. However, since these estimates were
derived by fitting only the light curves (an approach that may lead to
degeneracies), they were strongly influenced by the imposed Galactic model
prior. Abdurrahman et al. [1058] recently re-analyzed the data of these two
sources, complementing the light curves with new high-resolution near infrared
data. Their work strengthens the case for an isolated BH to be the lens in
both microlensing events, for relative lens-source proper motions above 0.81
mas/yr for MACHO-96-BLG-5, and 2.48 mas/yr for MACHO-98-BLG-6. A future
astrometric/spectroscopic examination of the sources will be crucial: 1) to
obtain an independent measurement of the main parameters of the sources (i.e.
lens-source parallax, lens mass); and 2) to search for every possible
contribution from a second object such as a non-BH compact object or a brown
dwarf lens. This will provide a definitive answer on the nature of these
objects.
The first and only unambiguous detection of an isolated stellar-mass BH in the
Galactic bulge, but more generally thus far in the Universe, has been recently
reported by two different groups, Lam et al. [1059] and Sahu et al. [1060],
almost at the same time. The microlensing event has been detected by MOA and
OGLE, has a long duration ($t_{E}$ $\approx$ 270 d) and is labeled as
MOA-2011-BLG-191 or OGLE-2011-BLG-0462. The magnification of the source caused
by the passage of the lens is shown in Figure 17 in two different ways: as a
sample of HST observations acquired from 2011 to 2017 where both the apparent
change in luminosity and the astrometric deflection can be visualized (top
panel) and as the photometric light curve of the source drawn over time by
multiple facilities (bottom panel). The two teams have used almost the same
set of multi-epoch observations from Hubble to obtain precise astrometry of
this magnified source and derived both the mass of the lens and its distance.
They both agree that the lens must be a compact source rather than a luminous
object, but they disagree on its current mass, thus on the interpretation of
its nature. Lam and colleagues [1059] estimated the source to be relatively
nearby, at a distance of 0.70-1.92 kpc and derived a mass of 1.6-4.4
$M_{\odot}$, consistent with either a NS or a BH, while Sahu and collaborators
[1060] measured a distance of 1.58 $\pm$ 0.18 kpc, implying an inferred lens
mass of 7.1 $\pm$ 1.3 $M_{\odot}$, only compatible with it being an isolated
BH.
Both groups noted an inconsistency between the best fit solution preferred by
the photometric data and the one from the astrometric data but they adopted
slightly different approaches to try and investigate this aspect further. Lam
et al. [1059] in particular used two different weights to analyze the results
of the microlensing models: a default weight (DW) which gave the same weight
to each measurement and uncertainty, and an equal weight (EW) where the two
data sets were weighted in the same way. Of the two, the DW solution, that
better agrees with the photometric data, is the one closer to the solution
found in the study by Sahu et al. [1060], although some differences still
remain. The EW solution, instead, is more in agreement with the astrometric
data but predicts a slightly lower distance for the lens, as well as a lower
mass, falling in the NS mass regime.
Shortly after this detection, Mereghetti et al. [1061] have started an
observational campaign with the Chandra, XMM-Newton and INTEGRAL satellites to
search for any X-ray emission coming from this object, with the aim of
independently binding its nature. They found that the inferred X-ray
luminosity of this object is consistent with the small radiative efficiency
expected for a BH more than for a NS. Upon reanalysis of the available
photometric and astrometric data, Mróz et al. [1062] were able to say the last
word on this story: according to them, the lens is indeed an isolated BH (with
mass M = 7.88 $\pm$ 0.82 $M_{\odot}$ at a distance of $D_{L}$= 1.49 $\pm$ 0.12
kpc). They investigated the presence of systematic errors both in the
photometric and/or astrometric data but they also evaluated the impact of the
presence of a bright star located $\approx$ 0.4” away from the microlensing
source (see Fig. 17) in the estimate of the astrometric shift of the source.
They came to the conclusion that, although in general the presence of a
blending source might produce erroneous results and need to be properly taken
into account, in this case systematic errors in the astrometric solution were
the primary cause of the low mass measured by the Lam et al. team.
The blind search for isolated BHs is still in an embryonic state, so any new
detections coming in the future will represent an important step forward in
the field. The third data release (DR3) of the GAIA satellite (more details
about the mission can be found in Section XVII.2.2), for example, already
contains 363 new microlensing events [1063] that no other surveys have
detected before and the prospect of combining both the light curves and the
astrometric information that GAIA will continue to provide until the end of
the mission looks extremely promising. The development of new dedicated tools
for searching for transient astrometric lensing events caused by dark compact
objects in GAIA DR4 will also greatly help in this regard [1064].
The detection of isolated BHs through gravitational microlensing will receive
a decisive boost in the future thanks to the advent of a new and advanced
facility, the Nancy Grace Roman Space Telescope and its Galactic Bulge Time
Domain Survey (GBTDS). With its wide field of view (0.28 square degrees),
super resolution (0.11 arcsec/pixel) and ability to simultaneously obtain
precise photometry and astrometry, Roman is expected to detect and allow the
characterization of several thousand microlensing events [1065, 1066]. It will
observe the Galactic bulge and other fields multiple times over the years,
allowing us to collect a rich sample of photometric and astrometric
measurements of known and unknown microlensing events. For such a reason,
Roman will enable the GBTDS to expand the catalog of isolated BHs by order(s)
of magnitude, yielding exquisite determinations of their masses, distances,
and proper motions, as well as opening up the opportunity to finally perform
BH population studies.
A recent work by Saradjan & Sahu [1067] have simulated a numerous sample of
microlensing events as seen by Roman, in order to effectively quantify the
role that this instrument will have in the future research for isolated BHs.
Since there is not yet an accurate mass function for isolated BHs based on
observations, they performed the simulations by considering several mass
functions (e.g. $dN/dM\propto M^{-0.5},\propto M^{-1}\leavevmode\nobreak\
\text{and}\propto M^{-2}$). The conclusion of the study is that, considering
the entire lifetime of the mission, the Roman telescope will be able to detect
3-4, 15-17 and 22-24 isolated BHs through astrometric microlensing (depending
of the mass function adopted) and will estimate their physical properties with
reasonable accuracy. In particular, they estimate that the relative errors on
physical parameters such as mass, lens-observer distance and proper motion
will be respectively less than 1, 5, 10%. Advances in this field will provide
important and independent insights into our understanding of the formation and
evolution of BHs.
Figure 17: combined from Sahu et al. [1060]: Top panel: The source of the
microlensing event MOA-11-191/OGLE-11-462 as observed in eight epochs of HST
observations (from 2011 to 2017). The source is highlighted with a white arrow
in the leftmost panel. All stars in the field have similar luminosity over the
epochs while the magnification of MOA-11-191/OGLE-11-462 in Epochs 1 and 2
(produced by the lens passage) is evident compared to the remaining epochs.
Epochs 2 and 3 are instead those where a maximum astrometric shift is
measured. Bottom panel: The photometric light curve of the microlensing event
MOA-11-191/OGLE-11-462 built by combining ground-based observational data
collected from different facilities is shown along with the best-fitting model
light curve. The instruments are labeled directly in the figure with different
colors. The shape of the magnification curve is characteristic of a
microlensing event.
### XVII Black Holes in binary systems with luminous companions
The second family of stellar-mass BHs consists of all those objects that orbit
a star in a binary system. The physics governing binary systems is rather
complicated and still not fully understood but for the purpose of
understanding this Section, it is important to mention some basic concepts. In
a binary system, each of the two components is surrounded by a teardrop-shaped
region, called Roche Lobe, where self-gravity dominates over the gravitational
pull of the companion. If the star is inside of this region, the material is
considered bound. Usually approximated as a sphere of the same volume, the
Roche Lobe have size which depends on the mass difference between the two
stars and on the orbital separation of the system, therefore on how far apart
the two components are from each other. According to [412], the size of a
Roche Lobe can be approximated to:
$\frac{r_{1}}{A}=\frac{0.49q^{2/3}}{0.6q^{2/3}+ln(1+q^{1/3})}$ (XVII.1)
where $r_{1}$ is the radius of the Roche Lobe of the primary component (the
more massive by convention), A is the angular separation of the system, while
$q$ is the mass ratio defined here as $M_{1}/M_{2}$. To obtain the Roche Lobe
size for the secondary component, it is sufficient to replace label 1 with
label 2 in the previous formula.
During the evolution of a binary system, both the radius of the stars and the
size of the Roche Lobes can vary, for several reasons. The first happens as
the stars evolve to different stages of their evolution (for example expanding
when evolving from the main sequence to the red giant branch), while the
latter occurs when the orbit of the binary shrinks or widens due to dynamical
interactions and close passages of third stars (especially in crowded fields)
or the mutual influence of the two stars. As soon as the size of the most
massive object (the one that evolves faster as predicted by stellar evolution
theory) exceeds that of its Roche Lobe, the material “located outside” the
Roche Lobe becomes unbound, triggering three possible scenarios. This material
can indeed: 1) escape the system completely, reducing the total mass of the
system; 2) orbit both stars; 3) fall onto the companion’s Roche Lobe, with
consequent mass transfer and accretion. Each of these events significantly
modifies the configuration of the binary. For those interested in the details,
a comprehensive review on binary systems can be found in Chapter 1.
Depending on the status of the system, it can be classified as either non-
interacting or interacting binary, with pretty significant differences in how
they can be observed. The simplest and most secure way to detect a BH (or a
NS) in a binary system with a star is if the two objects are actually
interacting, as the accretion of material from the star onto the compact
object can trigger the emission of energy at high-frequency (for example
X-rays or gamma-rays) but also at low-frequency (in radio) which can be
detected through tailored observations in these bands. Unfortunately this is
not the case if the BH is in a non-interacting system, also called a detached
configuration. As the word itself explains, this means that the companion has
not yet filled its Roche Lobe and has not yet begun to transfer material to
the BH, therefore they are in a quiescent state and their presence can only be
revealed using alternative techniques. These two configurations will be
discussed in more detail in two dedicated sections later.
#### XVII.1 Black Holes in interacting systems
In this Section we will discuss interacting black hole binary systems, their
observational history and their connection to gravitational wave physics.
X-ray binaries (XRBs) consist of a compact object (black hole or neutron star)
accreting matter from a gravitationally bound companion (or donor star). These
sources are multiwavelength objects and are most luminous in the X-ray regime,
with $\rm{L_{X}}>10^{36}$ erg/s. The relativistic accretion of material from
the companion star to the compact object is what powers this enormous energy
output. XRBs are the only class of binary system that can be
electromagnetically detected outside the Local Group, a small ($\sim$1 Mpc)
group of 50 galaxies that includes the Milky Way [see e.g. 1068, 1069], making
these sources very useful in the study of binary stellar systems.
X-ray binaries were some of the first X-ray sources to be detected and have
been actively studied for several decades [see e.g. 1070]. There are over 400
known XRBs in the Galaxy and the Magellanic Clouds [1071, 1072].
The study of X-ray binaries allow us to probe physics in extreme conditions;
from regions of high magnetic field strength to relativistic accretion and
strong gravity [see e.g. 1070]. They also help us to gain insight into other
astrophysical systems.
For instance, black hole X-ray binaries (BHBs) can allow us to explore the
physics of black holes on much shorter time scales than exhibited by
supermassive black holes. A study of BHBs with low accretion rates uncovered a
power law relation between radio ($L_{R}$) and X-ray luminosity ($L_{X}$),
with $L_{R}\sim L_{X}^{0.7}$ [1073, 1074]. Extending this work to include
supermassive black holes in active galactic nuclei, [1075, 1076] revealed a
relation between X-ray luminosity ($L_{X}$), radio luminosity ($L_{R}$) and
black hole mass $M$, often called the Fundamental Plane of black hole activity
(see Fig. 18), with
$\rm{log}\,\rm{L_{R}}\sim 0.6\,log\,\rm{L_{X}}+log\,M$ (XVII.2)
This relation suggests that for black holes with low accretion rates, the
physical processes governing how accreted material is converted into into
radiative energy could be universal over the entire black hole mass range,
from stellar mass black holes to supermassive black holes [see 1077, and
refernces therein].
Figure 18: Taken from [1075]. This is an edge-on view of the Fundamental Plane
of black hole activity, across a black hole mass range spanning 9 orders of
magnitude.
XRBs are predicted to be possible progenitors of some of the type of
gravitational wave sources that are already being detected [1078, 1079, 718,
and references therein]. They are also predicted to be sources of detectable
gravitational waves themselves [see e.g. 1080, and references therein]. In
light of this, there has been a renewed interest in XRBs from many branches of
astronomy.
In this work, we will be exclusively focusing on XRBs with a black hole
accretor (black hole X-ray binaries; BHBs). In the following sections, we will
go over the past, present and future of X-ray detectors and observatories. We
will discuss the types of X-ray binaries and their links to gravitational wave
physics, and give an overview of the observational signatures of black hole
binaries.
##### XVII.1.1 Development of electromagnetic detection methods for X-ray
binaries
The Earth’s atmosphere is opaque to X-rays from space. Thus, in order to
detect this high energy emission from astronomical objects, X-ray telescopes
must be placed above the atmosphere using balloons, rockets or satellites. The
first X-ray detection outside of the Earth’s atmosphere was from the our Sun’s
corona, in 1949, using Geiger counters launched on a sounding rocket [1081].
To date, over 60 X-ray instruments and telescopes have been used by
astronomers202020A list of all satellite and rocketborne X-ray missions can be
found at https://heasarc.gsfc.nasa.gov/docs/heasarc/missions/.
The first X-ray binary, and indeed the first extrasolar X-ray source, to be
discovered was Scorpius X-1 [1082]. In a mission lasting 350 s, an Aerobee
rocket equipped with proportional counter detectors observed what is still the
brightest apparent source of X-rays in the sky, after the Sun. Figure 19 shows
some early observational X-ray data of Scorpius X-1 taken in 1967 with
detectors on a sounding rocket. After considering the X-ray and optical data
for Scorpius X-1 it was concluded that it must be a neutron star accreting
from a binary companion [1083].
In 1964, another rocket equipped with Geiger counters observed a strong
Galactic X-ray source in the Cygnus constellation [1084]. This source, called
Cygnus X-1, would later be found to be a XRB containing the first compact
object widely accepted to be black hole.
Figure 19: Taken from [1085]. This data shows the X-ray photon count rate from
a three minute long observation taken with a rocket-borne detector in 1967 in
the direction of Scorpius X-1 and including the Galactic centre. We see
clearly that Scorpius X-1 is a very bright X-ray source.
The first dedicated X-ray satellite, _Uhuru_ was launched in 1970 [1086]. The
mission lasted for just over two years and a total of 339 X-ray sources were
detected [1087], including galaxy clusters, supernova remnants and binary star
systems (see Fig. 20). _Uhuru_ detections allowed us to understand that most
the X-ray stars that had been observed were in fact black holes or neutron
stars accreting from a binary companion. Indeed it was observations from
_Uhuru_ that helped to determine that Cygnus X-1 contained a black hole [1088,
1089].
Figure 20: Taken from [1087]. The X-ray sources from the fourth and final
_Uhuru_ source catalogue displayed in Galactic coordinates. The size of the
symbols is proportional to the logarithm of the peak source intensity.
The systematic study of XRBs became possible with the launch of the _Einstein_
X-ray Observatory [_Einstein_ ; 1090]. A space-based instrument launched in
November 1978, it was the first fully imaging X-ray satellite telescope.
_Einstein_ boasted an angular resolution of tens of arcseconds (compared to
the few arcminute resolution of the rocket borne telescopes at the time [see
e.g. 1091]), a field of view of a few arcminutes and a sensitivity nearly 1000
times better than previous instruments and operated in the 0.3–4 keV range.
Prior to _Einstein_ ’s launch, only six extragalactic X-ray binaries had been
detected (excluding the $\sim$60 XRBs in the Milky Way); five in the Large
Magellanic Cloud and one in the Small Magellanic Cloud [1092, and references
therein]. With _Einstein_ , over 100 XRBs or XRB candidates were found in 10
Local Group galaxies, making it possible to compare XRB populations in
different environments. For instance, it was found that high mass X-ray
binaries (see section XVII.1.2) were found preferentially in later type
galaxies whereas low mass X-ray binaries (see section XVII.1.2) were more
likely to be found in earlier types. This finding was not thought to be only
due to the increasing number of massive stars in later type galaxies.
_Einstein_ also revealed that a significant number of XRBs were associated
with globular clusters and that XRB host clusters were brighter than average [
e.g. 1093, see also sections XVII.1.2 and XVII.1.2].
The _Einstein Observatory_ (also called HEAO-2)) ushered in a new and exciting
age in X-ray astronomy. The three large X-ray satellite telescopes that
followed, the _Röntgen_ Satellite [_ROSAT_ ; launched in June 1990 1094] , the
_Advanced Satellite for Cosmology and Astrophysics_ [_ASCA_ ; launched in
Febraury 1993 1095] and the _Rossi X-ray Timing Explorer_ [_RXTE_ ; launched
in December 1995 1096] had instruments that provided the next big improvements
on Einstein. Each of these three telescopes were optimized for one aspect of
X-ray observations that complimented the other two. _ASCA_ , with it’s focus
on X-ray spectroscopy, had a large energy range and good energy resolution. It
was also the first X-ray telescope to make use of large CCD detectors. _ROSAT_
was designed for X-ray imaging, therefore it had a large field of view (2°)
and very good spatial resolution. _RXTE_ was a timing instrument and offered
microsecond time resolution. Figure 21 shows a selection of sources detected
by _ROSAT_ during the all-sky survey (see Table 1 for a comparison of various
telescope characteristics.)
Figure 21: Taken from https://www.mpe.mpg.de/971847/RH_all_sky_survey. This
image shows the 18 811 bright point sources detected by _ROSAT_ in the all-sky
survey. A total of 150 000 sources were detected by _ROSAT_ in this survey, 25
times more than in all previous X-ray satellites combined.
The next significant advancement in X-ray astronomy came with the launching of
the _Chandra_ X-ray Observatory and the _XMM-Newton_ Observatory. The
_Chandra_ X-ray Observatory was launched in 1999 [1097]. This telescope has
subarcsecond spatial resolution and an energy range of 0.1–10 keV. The _XMM-
Newton_ Observatory, launched later that same year, operates in a similar
energy range (0.1–15 keV). _XMM_ has a much coarser angular resolution ($\sim$
$15"$) than _Chandra_ , but has an effective area many times larger (4500 cm2
at 1 keV) [1098]. Figure 22 shows over 40,000 _XMM_ sources detected until the
end of 2013. Using these telescopes, it is now possible to carry detailed
studies of X-ray binaries, and other X-ray sources, in galaxies well outside
the Local Group [see e.g. 1099, 1100].
Figure 22: Taken from [1101]. This image shows the nearly 400 000 X-ray
sources detected by _XMM_ using public observations up to 2013 December 31.
With their complimentary specifications and their greatly improved
capabilities compared to predecessors, _Chandra_ and _XMM_ have been the main
work horse instruments for the study of X-ray binaries for over two decades.
Table 1 shows the comparison between the technical characteristics of
_Chandra_ , _XMM_ and previous X-ray satellites.
Table 1: A comparison of X-ray telescope characteristics. This table has been
adapted from Table 28 of the _XMM-Newton_ Users Handbook (https://xmm-
tools.cosmos.esa.int/external/xmm_user_support/documentation/uhb/xmmcomp.html).
Data are taken from this table unless otherwise indicated. See also
https://heasarc.gsfc.nasa.gov/docs/heasarc/missions/comparison.html for
comparisons of _Chandra_ , _XMM_ and other X-ray telescopes.
Telescope | Energy | Effective | Angular | Energy | Time | Orbital
---|---|---|---|---|---|---
| range | area | resolution | resolution | resolution | target
| (keV) | at 1 keV (cm2) | (arcsec) | at 1 keV (eV) | (s) | visibility (hour)
_Chandra_ | 0.1$-$10 | 555 (ACIS-S)212121For more information on _Chandra_ ’s ACIS instrument see https://cxc.harvard.edu/proposer/POG/html/chap6.html#tth_chAp6. | 0.5 | 1 (HETG)222222For more information on _Chandra_ ’s ACIS instrument see https://cxc.harvard.edu/proposer/POG/html/chap8.html. | 3.2232323taken from https://cxc.harvard.edu/cal/Acis/index.html. (ACIS) | 44.4
_XMM_ | $0.15-12$ | 4650 | 15 | 4 (RGS)242424Fore more infomation on _XMM_ ’s RGS instrument see https://www.cosmos.esa.int/web/xmm-newton/technical-details-rgs. | 0.07252525Taken from https://www.cosmos.esa.int/web/xmm-newton/technical-details-epic. (pn)262626For more information on _XMM_ ’s pn instrument see https://www.cosmos.esa.int/web/xmm-newton/technical-details-epic#5.1.2. | 36.7
_ASCA_ | $0.5-10$ | 350 | 174 | 100 | $6\times 10^{-5}$ (GIS)272727Taken from [1102] and see same source for more information of _ASCA_ ’s GIS instrument. | 0.9
_ROSAT_ | $0.1-2.4$ | 400 | 3.5 | 500 | $6\times 10^{-5}$ (HRI)282828Taken from https://heasarc.gsfc.nasa.gov/docs/rosat/hri.html and see same link for more information on _ROSAT_ ’s HRI instrument. | 1.3
_RXTE_ | $2-250$ | _n.a._ | _n.a._ | _n.a._ | $10^{-6}$ | 1
_Einstein_ 292929Data taken from Table 1 in[1090], see same source for more information on _Einstein_ instruments | $0.3-4$ | 300 | 2 | $10-100$ (FPCS) | 8$\times 10^{-6}$ (HRI) | $0.3-0.6$303030Taken from https://heasarc.gsfc.nasa.gov/docs/einstein/HEAO-B_Guest_Observers_Guide.pdf
There are currently about a dozen X-ray satellite observatories operating,
including _Chandra_ and _XMM_ 313131see
https://heasarc.gsfc.nasa.gov/docs/heasarc/missions/active.html for a list of
active X-ray observatories. This is a global effort with China, Europe, India,
Japan, Russia and the US all involved in launching and running missions, many
of them jointly [1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112,
1113]. A list of planned upcoming and future X-ray missions can be found at
https://heasarc.gsfc.nasa.gov/docs/heasarc/missions/upcoming.html.
##### XVII.1.2 Types of X-ray binaries
X-ray binaries are generally classified by the mass of the companion star, as
the nature of this companion determines the time scales and accretion mode of
the binary. Thus we have high mass X-ray binaries, low mass X-ray binaries
(see Fig. 23), intermediate mass X-ray binaries, depending on the mass of the
companion. XRBs can also be classified by their X-ray emission profiles, such
as ultraluminous X-ray sources and super soft X-ray sources (the latter are
sources with a white dwarf accretor and so will not be discussed further in
this).
###### High Mass X-ray Binaries
High mass X-ray binaries (HMXBs) are XRBs with a giant or subgiant donor star
with mass $>10$ M⊙, typically O or B main sequence stars [see e.g. 371] and
the optical luminosity is greater than the X-ray. HMXBs are associated with
young stellar populations and they are a good tracer of recent star formation
[see e.g. 1114, 1115, 1116].
There are two categories of HMXBs: standard HMXBs and Be star XRBs [371, and
references therein]. The standard HMXBs have supergiant O or B type companion
stars, are persistent X-ray sources, have low eccentricity ($e<0.1$) and short
orbital periods ($<10$ days). The mass transfer in these sources proceeds via
one of two processes:
* •
Roche lobe overflow. Roche lobe overflow (RLOF) occurs in a binary when the
companion star expands due to its evolution, filling its Roche lobe, the tear-
shaped surface of gravitational equipotential. Matter from the companion then
flows through the inner Lagrangian point (the point between the binary
components where their respective Roche lobes intersect, i.e. their respective
gravitational potentials are equal; see Fig. 23) onto the compact object [see
e.g. 1117].
* •
Wind-fed accretion. The O and B companion stars in HMXBs have dense, strong,
high-velocity stellar winds. The orbit of the compact object is inside this
stellar wind and the compact object accretes material directly from the
companion’s wind [1118, 1119].
The second class of HMXBs are Be X-ray binaries (BeXRBs), so called because
they have B-emission star (Be-star) companions, rapidly rotating stars with a
variable disk of circumstellar materal around its equator. These systems have
higher eccentricity ($\lesssim 0.2-0.5$) and wider orbits (P$\sim 10-35$ days)
than the standard HMXBs [1120]. In these systems, accretion occurs when the
compact object’s orbit intersects with the circumstellar disk of the Be-star;
rare instances of transient RLOF have also been observed. Due to the high
eccentricity of the compact object orbit and the variable nature of the
circumstellar disk, BeXRBs are transient and highly variable X-ray sources
[1118, 371].
Only 10 BH HXMBs or BH HMXB candidates are currently known, with 7 of these
being extragalactic sources [1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128,
1129, 1130, 1131, 1132, 1133]. These sources are of particular interest since
HMXBs are thought to be progenitors of some of the gravitational wave sources
we observe with instruments such as the advanced Laser Interferometer
Gravitational-Wave Observatory (a _LIGO_ 323232See https://www.ligo.org/ for
details on this instrument.), such as BBHs and BH-NS binaries [see e.g. 645,
291, 98, 1079, 465, and references therein. See also section XVIII].
However, the current models for the evolution of HMXBs to double degenerate
binaries that merge within a Hubble timescale depends on several poorly
constrained aspects of stellar and binary evolution. These uncertainties
include metallicity, the common envelop phase of binary evolution, star-
formation rates and massive star evolution [e.g. 1134, 718].
###### Low Mass X-ray Binaries
Low mass X-ray binaries (LMXBs) consist of an accreting compact object with a
low mass ($\lesssim$1 M⊙) companion [371]. The companion stars are faint and
rarely observed, with $\rm{L_{opt}}/\rm{L_{X}}\ll 0.1$ [1135]. Mass transfer
in LMXBs occurs via RLOF.
These systems have orbital periods of between 11 minutes and about 20 days.
LMXBs are associated with old stellar populations like those found in the
Galactic bulge, elliptical galaxies and globular clusters [371].
LMXB population size scales with the stellar mass and age of the host galaxy
[e.g. 1136, and references therein]. It was also found that younger stellar
populations have fewer faint LMXBs and more bright ones per unit mass than
older populations. LMXB populations can also be used to probe the nature and
distribution of dark matter in galaxies [see e.g. 1137, 1138].
_LMXBs in dense environments_
Globular clusters (GCs) are old ($\gtrsim 1$ Gyr), gravitationally bound
stellar clusters that orbit galaxies. These clusters are compact systems with
half light radii of a few parsecs, masses in the range $\sim 10^{4}-10^{6}$
M⊙, optical luminosities of MV = -5 to -10 and increasing stellar density
towards the cluster centre [e.g. 1139, 1140].
Globular clusters have a high rate of dynamical stellar interactions, making
these clusters very efficient at forming LMXBs. Binaries are formed in GCs via
such processes as tidal capture, exchange interactions and direct collisions
[1093, 1141, 1142, 1143, 1144, and references therein].
Ten percent of the LMXBs associated with the Milky Way reside in its GCs
[1072, and references therein]. Milky Way LMXBs are therefore more than two
orders of magnitude more likely to be found in a GC than in the field of the
Galaxy per unit stellar mass [see e.g. 1145]. For elliptical galaxies,
$20\%-70\%$ of LMXBs are found in GCs [see e.g. 1146, 1147, 1148, 1149].
LMXBs with hydrogen-poor donors and orbital periods of less than an hour are
called ultracompact X-ray binaries (UCXBs). Currently, there are about 30
UCXBs and UCXB candidates known in the Milky Way [1150]. UCXBs are fives time
more common in globular clusters than in the field of the Milky Way. Only one
black hole UCXB in a GC has been discovered to date [1151].
Future gravitational wave observatories, such as the Laser Interferometer
Space Antenna (_LISA_) will be able to detect binary black holes and black
hole UCXBs in Milky Way globular clusters and perhaps even in GCs as far out
at the Virgo Cluster [see e.g. 1152, 1153, 1154]. These gravitational wave
detections will provide robust mass estimates for the binary components,
thereby constraining GC black hole formation and evolutionary models.
Figure 23: Taken from [371]. The typical configurations for a high mass (left)
and low mass (right) X-ray binary (this sketch depicts neutron star binaries,
but black hole binaries have the same composition). The Roche lobe overflow
accretion process is depicted here for the LMXB and windfed accretion is shown
for the HMXB, although Roche lobe overflow can take place in HMXBs as well.
Note the difference in size and mass for two types of binary.
###### IMXBs
For completeness, we note that X-ray binaries with intermediate mass (IMXB)
companion stars (1–10 M⊙) do exist [see e.g. 1155]. Accretion onto the compact
object in an IMXB proceeds via Roche lobe overflow. Due to the relatively high
mass ratio between the compact object and donor star, this accretion phase
proceeds very rapidly, lasting only a few thousand years [see 371, and
references therein]. These short time scales make these systems difficult to
observe and only a small number of these systems are known. Additionally, the
evolution of IMXBs is thought to be a formation channel for low mass X-ray
binaries [see e.g. 1156, and references therein]. Thus IMXBs are not discussed
further in this work.
###### ULXs
Ultraluminous X-ray sources (ULXs) are extragalactic point-like X-ray sources
located outside of the centre of a galaxy, with an apparent peak X-ray
luminosity $>10^{39}$ erg/s, the Eddington luminosity of an 8 M⊙ black hole
[see 595, for a review] (see also XVII.1.3). With their high X-ray
luminosities, ULXs are important objects in the study of relativistic
accretion as they challenge our understanding of black hole formation and
relativistic accretion [1157, 595, 1158, 1159, 1160, and references therein].
Figure 24: Figure 2 from [1161], showing the X-ray contour map of the spiral
galaxy IC 342 overlaid on the Palomar plate image of the galaxy. The image
shows two bright point sources to the NE and SE of the central source with
estimated X-ray luminosities of $\gtrsim 2\times 10^{39}$ erg/s.
ULXs were first detected in the 1970s by the _Einstein_ Observatory [1162,
1163] (see Fig. 24). These sources were initially thought to be intermediate
mass black holes (IMBHs) with mass range $100-10^{5}$ M⊙, since the Eddington
luminosity, $L_{\rm X}$ scales with accretor mass, $M$, as
$\rm{L_{X}}\approx 1.3\times 10^{38}(M/M_{\odot})\,erg/s$ (XVII.3)
[1164, and references therein] (see also XVII.1.3). IMBHs are important
objects from a cosmological standpoint, as they are thought to be the seed
black holes from which the observed supermassive black holes in the universe
grow [1165, 1166, 1167]. Subsequent studies showed that the predicted
formation rates of accreting IMBHs cannot explain the entirety of the observed
population of ULXs [1168, 1169].
_ULXs in low-density density environments_
ULXs are more likely to be found in star-forming galaxies [1170] and younger
ellipticals with recent star formation [1171, 1172], suggesting that ULXs are
an extension of the high luminosity X-ray binary population, which scales with
both star-formation rate and stellar mass of the host galaxy [1173, 1174,
1116]. Using observational data from the _Chandra_ and _XMM-Newton_ X-ray
telescopes and cross-matching with optical catalogues, [1175] compiled a
catalogue of over 700 ULX sources and candidates. They found that ULXs are
preferentially located in spiral galaxies and areas with active star
formation, consistent with previous work (see Fig. 25).
Figure 25: Figure 9 from [1175], showing the ULX population size by host
galaxy Hubble type ($T_{H}$). The vast majority of ULXs in the catalogue are
found in spiral galaxies ($0<T_{H}<10$) with a spike in elliptical galaxies
($T_{H}<-3$). The blue histogram represents the distance and luminosity
complete sample and the red represents the bright ($\rm{L_{X}}$ $>5\times
10^{40}$ erg/s) subsample.
Today, ULXs are mostly thought to be the result of super-Eddington accretion
onto stellar mass black holes and neutron stars in X-ray binaries [1176, 1177,
1178, 1179, 1180]. There is a subset of extremely high luminosity sources, so
called hyperluminous X-ray sources (HLXs; $\rm{L_{X}}>10^{41}$ erg/s), that
are better explained by accretion onto higher mass black holes [1181, 1182,
1183].
[1184] compiled a catalogue of over 100 HLX candidates out to a redshift of
0.9 by combining X-ray and optical observations. This sample of HLXs is a
useful starting point to search for IMBHs electromagnetically and they
identified 22 sources as IMBH candidates.
###### ULXs in high-density environments
There is a small population of about 20 ULXs that resides in extragalactic
globular clusters (GCs) [1185, and references therein]. Unlike ULXs in the
field of galaxies, GC ULXS form via dynamical interactions, due to the high
stellar density of GCs. The donor stars for GC ULXs are also old, low mass and
hydrogen poor, compared to the high mass stars we find in star forming regions
in the field of galaxies. The presence of ULXs in GCs means these X-ray
sources can be easily separated from background AGN. An analysis of this
population of X-ray sources revealed that ULXs are preferentially found in
brighter GCs, but that the presence of a ULX in a GC is not correlated with
the optical colour of the cluster [1186, 1187].
The first globular cluster ULX source was discovered by [1188, see Fig. 26]
using _XMM_ data. The source flux decreased by a factor of seven from an
estimated peak flux of $\sim 4\times 10^{39}$ erg/s over the course of a few
hours. This rapid variability does not support the scenario that the X-ray
flux is from a superposition of neutrons stars in the GC. Indeed, rapid X-ray
flux variability is one of the ways we can distinguish a black hole X-ray
binary (BHB) from other X-ray sources in GCs [1189].
Figure 26: Taken from [1188], showing the X-ray light curve of XMMU
J122939.7+075333, the first ULX detected in a globular cluster (GC). The
source is located with in the GC RZ2019 of the giant elliptical galaxy NGC
44772. The light curve shows an estimated peak X-ray luminosity of $\sim
4\times 10^{39}$ erg/s.
BHBs are the most likely sources of GC ULXs. Thus the study of ULXs in GCs can
help shed light on the formation, evolution and merger history of black holes
in GCs [see e.g 1078]. However, it is important to note given the
extragalactic nature of these sources, we are unable to determine mass
estimates for the components of these X-ray binaries electromagnetically.
Fortunately, we can turn to gravitational wave detections to help solve this
problem.
Initially, it was thought that any black holes formed in GCs would either
cause the cluster to be destroyed [1190] or the BHs would be ejected via
dynamical interactions fairly early on in the cluster lifetime. [1191, 1192].
Globular clusters are now thought to be very efficient at forming binary black
hole systems (BBHs) and other binaries due to the high density of stars,
especially in the cluster core [see e.g. 1193, 1194]. Indeed, GCs are now
thought to be likely hosts of some of the gravitational waves sources observed
by a _LIGO_[see e.g. 1195]. As with other LMXBs in GCs, the study of GC ULXs
will be greatly enhanced with future gravitational wave observations [see e.g.
1196].
###### Finding black hole X-ray binaries
Figure 27: Taken from [1197], last updated in December 2021. The cumulative
histogram of stellar mass black hole candidates in transient XRBs. The red
histogram shows the BH candidates and the blue bars represent dynamically
confirmed BHs. The horizontal black lines indicate the lifetimes of the X-ray
observatories that have discovered SXTs during an X-ray outburst.
As of the end of 2022, there are over 60 BHs and BH candidates known in our
Galaxy [1197]. The vast majority are LMXBs detected during an outburst, where
the X-ray luminosty increases by at least two orders of magnitude for weeks to
months before returning to a quiescent state for between one and 50 years,
with L${}_{X}\sim 10^{32}$ erg/s [371]. These types of LMXBs are called soft
X-ray transients (SXTs). The first black hole SXT, A0620-00, was identified by
[1198] in the 1980s. The source was initially discovered in August 1975 when
it went into outburst, becoming the brightest extra-solar X-ray source ever
detected at the time [1199]. A0620-00 remained bright for several weeks before
returning to quiescence.
With most LMXBs, the optical flux from the low mass donor star is not detected
due to the very bright optical emission produced from the X-ray heated gas
around the compact object. In the case of soft X-ray transients, when the
X-ray source is in a quiescent X-ray emission state ($<10^{33}$ erg/s; see
next section), the X-ray heating and subsequent optical emission are
diminished to the point where the optical emission from the donor star becomes
detectable. The donor star optical spectrum can then be used to determine the
orbital parameters of the SXT, including the masses of the binary components
[see e.g. 1198, 371, and references therein], (see also XVII.2.1 later in this
chapter).
In the next sections, we discuss the accretion disk physics that leads to
observed X-ray outbursts as well as the X-ray emission states of black hole
binaries.
##### XVII.1.3 Accretion physics in X-ray binaries
###### The accretion disk
As mentioned before, the vast energy output of XRBs is the result of accretion
of material from the donor star onto the compact object. In the case of RLOF
matter moving from the donor to towards the black hole object will have too
much angular momentum to fall directly onto the compact object [1200]. Instead
the material flows into circular orbit around the black hole, since this is
the lowest energy orbit for a given amount of angular momentum [e.g. 1201].
Viscous processes within the accreting material then cause the angular
momentum and energy to be transferred outwards. The material spreads outwards
into a series of concentric annuli moving in circular, Keplarian orbits,
called an accretion disk.
The viscosity in accretion disks is not the typical molecular viscosity due to
interactions between particles. It widely thought that the viscosity is due to
the interactions between the weak magnetic fields within the accretion disk
and the accretion disk material; this is called the magneto-rotational
instability model [1202, 1203, 1204].
As the accretion material loses angular momentum and kinetic energy, it moves
further down the gravitational potential well of the black hole, releasing
gravitational potential energy. The amount of rest mass energy released as
radiation depends on the type of compact object. For black holes, up to 40% of
of the accreted material rest mass energy can be released as radiation [1205]
.
The conditions to form an accretion disk are not necessarily present for wind-
fed HMXBs. The stellar wind of O and B star companions do not carry as much
angular momentum as the material coming from a RLOF scenario. Only one wind-
fed HMXB has shown evidence of having formed an accretion disk [1206, 1207,
and references therein].
###### Properties of the accretion disk
The structure and radiation emission of the accretion disk is largely
determined by the rate at which material flows from the companion star to the
accretion disk and towards the black hole [1200]. This accretion rate
($\dot{M}$) is related to the disk luminosity by
$L_{disk}=\eta\dot{M}c^{2}\,erg/s,$ (XVII.4)
where $\eta$ is the efficiency with which gravitational potential energy
released as radiation and $c$ is the speed of light.
At a critical value of the accretion rate $\dot{M}_{crit}$, the resultant disk
luminosity produces enough of radiation pressure on the disk electrons to
balance the compact object’s gravitational force on the disk protons. For a
compact object of mass $M$,
$\dot{M}_{crit}=3\times 10^{8}\eta M/year.$ (XVII.5)
The resultant disk luminosity, called the Eddington luminosity ($L_{Edd}$) is
then,
$L_{Edd}=\frac{4\pi GMm_{P}c}{\sigma_{T}}\,erg/s,$ (XVII.6)
where $G$ is the gravitational constant, $m_{P}$ is the mass of a proton, and
$\sigma_{T}$ is the Thompson scattering cross section of an electron [see e.g.
1200].
We note that the Eddington luminosity is calculated under the assumptions that
the accretion disk is composed entirely of pure hydrogen that is fully ionised
and that the radiation is isotropic. Despite this, $L_{Edd}$ is used widely to
parameterise the luminosity of XRBs.
When $\dot{M}<\dot{M}_{crit}$, radiative processes cool the disk enough for it
to be geometrically thin, but optically thick, such that the disk scale
height, $H\ll R$, where $R$ is the disk radius [see e.g. 1200, 1205]. The thin
accretion disk temperature profile is then
$T(R)=\biggl{\\{}\frac{3GM\dot{M}}{8\pi\sigma
R^{3}}\left[1-\beta\left(\frac{R_{in}}{R}\right)\right]\biggr{\\}}^{1/4}\,K,$
(XVII.7)
where $R_{in}$ is the inner radius if the accretion disk, $\beta$ is a
dimensionless parameter determined by the boundary condition at $R_{in}$ and
$\sigma$ is the Stefan-Boltzmann constant.
The smallest, stable, Keplerian circular orbit that the accretion disk can
extend to depends on the type of compact object [1200]. For a BH, this is the
innermost stable circular orbit (ISCO), equal to 3 times the Schwarzschild
radius ($R_{Schwarzschild}=2\frac{GM}{c^{2}}$). The general relativistic
effects in the strong gravitational field of the BH prevent stable orbits
smaller than the ISCO.
Because the thin accretion disk rotates at roughly Keplerian velocities, half
the disk energy is released as radiation and the other half is changed to
kinetic energy. Black holes have no solid surface, instead they are surrounded
by an intangible boundary called an event horizon that encompasses the region
of spacetime that can no longer communicate with the universe outside it. Thus
the kinetic energy within the accretion disk is lost once it crosses the event
horizon [1208].
###### The multicolour disk model
The geometrically thin, optically thick accretion disk described above is most
commonly modeled as a superposition of black body spectra, with a maximum
temperature at the innermost radius ($T_{in}$), the so-called ”multi-colour
disk” (MCD) model [1209]. This a non-relativistic model that does not take
into account the effects of strong gravity near the inner disk radius
($R_{in}$). Yet this MCD approximation is widely used to parameterise XRB
spectra and is able to fit the spectral data well. The disk luminosity
($L_{disk}$) is then obtained by integrating over the radial distance of the
disk and we have
$L_{disk}=4\pi\sigma R_{in}^{2}T^{4}_{in}\,erg/s.$ (XVII.8)
For black holes, we have
$\rm{M}\propto T^{-2}_{in}L^{1/2}_{disk},$ (XVII.9)
therefore more massive black holes have a lower characteristic disk
temperature [e.g. 1210].
Merloni et al. (2000) [1211] do caution that MCD model is an
oversimplification of the accretion disk than can have significant impact on
disk parameters obtained when the model is applied. Taking into account such
processes as Doppler blurring and gravitational redshift, they showed that the
MCD model systematically underestimates the value of $R_{in}$.
###### Deviation from the standard accretion disk model
When $\dot{M}>\dot{M}_{crit}$, the accretion disk structure changes; it
becomes both optically and geometrically thick, a so-called ’slim’ disk. The
radial temperature profile then goes from $T(R)\propto R^{-3/4}$ in the thin
disk regime, to $T(R)\propto R^{-1/2}$ [and references therein 1212].
When the accretion rate is very low ($<0.01M_{crit}$), the energy released
from viscous processes is not radiated away as efficiently as in the standard
thin disk scenario and the energy is instead stored in the disk material. This
causes the inner disk to swell and become geometrically thick, optically thin
and become hot, with the electron temperatures of about 100 keV [1213, 1212].
As a result, the inner accretion disk moves away from the compact object and
the disk becomes truncated. The hot, optically thin accretion material and the
energy inside it are then advected towards the black hole. Only a small
fraction of the energy is released before the accretion material reaches the
event horizon and is lost. This regime is called the advection dominated
accretion flow [ADAF; 1214] and the radiative efficiency is less than 1%.
###### The disk instability model
For an accretion disk to have a stable flow of matter through it, the
accretion material must be fully ionised throughout the disk. This stable flow
will result in a persistently bright X-ray source. As mentioned earlier, XRBs
can display strong flux variability in the form out outbursts on timescales of
weeks to months. The most widely accepted explanation for this variability is
the disk instability model, whereby instabilities in the accretion disk are
caused by changing ionisation states of the accretion disk material [1215,
1216].
Over a range of accretion rates, the accretion disk is thought to be in either
a cool, low ionisation, low viscosity state or a hot, high ionisation, high
viscosity state. The accretion disk opacity is depends very sensitively on the
ionisation level of the accretion disk material and therefore on the disk
temperature. Thus small changes in disk temperature cause rapid changes
between the hot and cool states.
To ensure that the hydrogen in the accretion disk is fully ionised, the disk
temperature must be greater than 6500 K, the ionisation temperature for
hydrogen [see e.g. 1217]. The disk temperature declines with increasing disk
radius, therefore for a stable accretion disk, the outer disk temperature must
be greater than 6500 K.
If the outer accretion disk temperature is lower than 6500 K, the disk will
contain areas of neutral hydrogen and these will quickly spread throughout the
disk. Cool, neutral accretion material does not flow as as easily as ionised
matter, thus the accretion flow towards the black hole decreases. This
reduction in accretion rate then leads to a reduction in luminosity and also
changes the structure of the accretion disk.
The changes in $\dot{M}$ described by the disk instability model, and the
subsequent changes in the nature of the accretion disk, lead to very distinct
radiation emission states. In the next section, we discuss the salient XRB
properties that arise due to the variations in the accretion rate.
##### XVII.1.4 Electromagnetic observational signatures of black hole
binaries
###### Black hole binary X-ray emission states
Black hole binaries are known to transition between quasi-stable emission
states, where the X-ray spectrum is dominated by either thermal or non-thermal
emission. These emission states are a result of the disk instability model
described above [see 1218, for a full review of black hole emission states].
In the following sections I will describe the canonical BHB X-ray emission
states as well as the accretion flow and physical accretion disk features
related to them.
_The quiescent state_
As described in previous sections, BH LMXBs spend most of their lifetime in a
quiescent X-ray state, where $10^{30.5}<L_{X}<10^{33}$ erg/s. The accretion
rate is very low in this state, with $\dot{M}<0.01M_{crit}$ [see e.g. 1214,
1208]. In the quiescent state, the accretion disk does not extend to the ISCO.
Instead, observations of BHB show that the disk is truncated at a large inner
disk radius [1213, 1214, 1219, 1220]. The ADAF model is invoked to explain the
low luminosity of the quiescent state as discussed in the previous section.
The leading model for the origin of the X-ray emission in the quiescent state
is as follows: A corona of hot ($100\textendash 300$ keV) particles surrounds
the accretion disk and the ADAF [e.g. 1214, 371, and references therein]. The
X-ray emission in the quiescent state comes from soft (low energy) photons
originating in the accretion disk being Compton up-scattered by the corona,
giving rise to to non-thermal, hard emission with energies up to 100 keV. The
energy spectrum of this emission is well fit by a power law (PL) spectral
model, with
$F(E)\propto E^{-\Gamma},$ (XVII.10)
where F is the flux, E in the energy and $\Gamma$ is the photon specrtal
index. For the quiescent state $\Gamma$ is in the range $1.7-2.1$ [e.g. 1208].
_The thermal dominant state_
This emission state is characterised by a large ($>75\%$) thermal emission
contribution component and the accretion rate is estimated be $0.1-0.5$
$\dot{M}/\dot{M}_{crit}$. The spectrum can be modelled by a MCD, with inner
disk temperature of $0.7-1.5$ keV [1214, 1208]. In this state, there is also
sometimes a hard, non-thermal component to the spectrum, modelled by steep
power law, $2.1<\Gamma<4.8$. The flux variability in the thermal dominant
state is weak and the accretion disk is thought to be in the classic thin disk
configuration of [1200].
_The hard state_
The hard emission state is defined by hard, non-thermal emission, where
$>80\%$ of the emission in the $2-20$ keV range can be well modelled by a
power law with $1.7<\Gamma<2.1$ [1208]. A soft X-ray contribution is also
present, which can be modelled as a cooler disk emission component of
$T_{in}=0.1-0.2$ keV. For some BHBs, there is also an excess of hard emission
($20-100$ keV), thought to arise from the reflection of the PL component
emission off the inner accretion disk.
The hard state is sometimes regarded as a less extreme version of the
quiescent state and the accretion rate is thought lie between those of the
quiescent and thermal dominant state. As with the quiescent state, the inner
accretion disk in the hard state is truncated to a larger radius than the ISCO
[1214, 1208].
The nature of the accretion flow within the inner disk radius is still a
matter of debate. Several versions of the ADAF model have been put forward,
but no one set of models can entirely predict or replicate the observed X-ray
behaviour [see 1221, 1222, for reviews].
The hard emission state in BHBs also is associated with synchrotron radio
emission from the a quasi-steady, compact jet [1223, and references therein].
The jets emission disappears when the BHB transitions from the hard state to
the thermal dominant state and there is a strong correlation between the jet
launching mechanisms and the nature of the accretion disk.
_The steep power law state_
The steep power law state (SPL; formerly called the very high state) is
characterised by non-thermal emission which is well modelled by a power law
model with a spectral index, $\Gamma>2.4$. At least 20% of the overall flux is
from the SPL component. In this state, $L_{X}$ is typically $>0.2L_{Edd}$ and
the accretion rate is $\gtrsim 0.5\dot{M}_{crit}$ [see 1208, and references
therein].
The exact physical processes that give rise to the SPL are still being
debated. [1224] and [1225] present a model where the steep power law emission
is due to Compton upscattering of the lower energy accretion disk photons by
the high energy corona electrons. Later, [1226] put forward the following
scenario: high energy ($>$ MeV energies) photons produced via synchrotron
emission from relativistic electrons in highly magnetised areas near the ISCO.
These high energy seed photons are then down-scattered by electrons in the
corona to produce the SPL. Another model uses the expected nature of the inner
accretion disk at high luminosities to produce the steep power law emission
[1227]. At high luminosites, the inner accretion disk is radiation dominated,
resulting in optically thin material producing significant flux. As a
consequence, the gas temperature increases rapidly inwards, leading to locally
saturated Compton spectra with rapidly increasing peak energies, the sum of
which results in a steep power law spectral component.
###### Black hole binary timing signatures
X-ray binaries display a wide range of flux variability across the
electromagnetic spectrum and on various timescales. We have already mentioned
flux variations on timescales of years to decades (quiescent state), weeks to
months (X-ray outbursts) and days to minutes (orbital periods). In the X-ray
regime, flux variability can also occur on millisecond timescales. This
aperiodic, fast variability is an important tool in the study of the strong
gravity near compact objects [see 1228].
The spectral emission states and flux variability in XRBs are correlated, with
both arising due to the physical processes occurring in the inner accretion
disk. [1229] first postulated that accretion onto stellar mass compact objects
will result in millisecond timescale variability. This fast, quasi-periodic
variability was thought to be the result of clumps of accretion material
orbiting in the inner disk near the compact object [1230]. Observational
evidence of this short timescale variability was provided by the _RXTE_
mission, which had millisecond time resolution.
In BHBs, this variability is made up of longer (noise) and shorter timescale
features. The short timescale features are called quasi-periodic oscillations
(QPOs) with frequencies in the range 0.01 to 450 Hz. QPOs are modelled as
Lorentzian functions and have coherence parameters, $Q=\nu/FWHM\gtrsim 2$,
where $\nu$ is the centroid frequency [e.g. 1231, 1228]. QPOs are associated
with X-ray spectral state transitions as well as non-thermal emission states.
Quasi-periodic oscillations can be divided into two subgroups: low frequency
QPOs (LFGQPOs; $\sim 0.1\textendash 30$ Hz) and high frequency QPOs (HFQPOs;
$40\textendash 450$ Hz).
LFQPOs have been observed in BHBs in the SPL, hard and SPL to hard transition
states. The frequencies and amplitudes of LFQPOs have been found to be
correlated to both the PL and thermal components of these spectral states [see
e.g. 1232, 1233, 1234, 1235]. Thus studying LFQPOs can help us understand the
PL spectral component and LFQPOs data can be used to constrain models of non-
thermal emission in BHBs [see 1218].
HFQPOs are transient and have lower amplitudes than LFQPOs and they have been
observed to occur when the BHB is in the steep power law state. Their
frequencies correspond to that expected for accretion material orbiting near
the ISCO of a roughly 10 M⊙ BH. It is thought that HFQPOs are related to the
mass and spin of the BH. Combining HFQPO data with BH mass estimates are a
good way to constrain the spin of the BH [see §8.2.4 in 1218].
#### XVII.2 Black Holes in non-interacting systems
In this Section we will describe what are the main observational strategies
that can be used to unveil stellar-mass BHs which are part of non-interacting
(or quiescent333333In this Chapter we give the words ”non-interacting” and
”quiescent” the same meaning. In the literature you can find these words
referring to slightly different configurations or evolutionary phases.) binary
systems.
Not being able to rely on X-ray, gamma-ray or radio emission from these
sources, the only way we can unveil a BH in a quiescent binary is by measuring
the dynamics of the companion luminous star. There are two main indirect
techniques that can be used to unveil a BH-star binary: 1) spectroscopically
by measuring period changes on the spectral lines of the luminous companion or
2) astrometrically (for example with GAIA), by tracking the orbital motion of
a luminous star as a NS or BH pulls it around. We will explore these two
approaches by highlighting the observational studies that have been conducted
so far in the field.
##### XVII.2.1 Spectroscopic detection of quiescent Black Holes
In a binary system made of two stars, the reciprocal motion of the two
companions with respect to their mutual center of mass causes the line-of-
sight components of their velocities to vary in antiphase with each other
(also referred to as reflex motion) as a function of time, and this produces a
periodic signal, whose periodicity depends on the orbital period of the system
under investigation, directly linked to the distance between the two stars in
the system. The amplitude of the signal is instead related to the masses
involved in the binary. By measuring the change in wavelengths of a few lines
of interest and/or the full spectrum (i.e., Doppler shift) for the two stars
over a period of time, it is possible to construct a radial velocity curve.
Figure 28 illustrates the case of a stellar binary. Depending on their
positions relative to the binary’s center of mass, star 1 and star 2 (as
labeled in the figure) appear to move towards and away from the observer. This
motion makes the light from the stars appear slightly bluer when they are
moving towards the observer, and slightly redder when moving away. This is
what is called Doppler effect.
Figure 28: Schematic representation of the methodology used to detect
spectroscopic binaries. This is the case of a stellar binary. Depending on the
position of a star (star 1 or 2) relative to the binary’s center of mass it
may appear to move towards or away from the observer. In the first case the
light emitted by this star appears blueshifted, however in the second it
appears redshifted. By monitoring these shifts over time it is possible to
construct the radial velocity curve of both star 1 and star 2, independently.
Binaries detected using this method (also called radial velocity method) are
named spectroscopic binaries and can be divided in two groups: 1) Binaries
where the two stars are almost equally bright, called double-lined
spectroscopic binaries (or SB2) ; 2) Binaries where one component is
significantly brighter than the other, called single-lined spectroscopic
binaries (or SB1). While for SB2 binaries we can determine the radial
velocities of the two components ($v_{1}$ and $v_{2}$), for SB1 binaries we
can instead only measure the radial velocities of the brightest component
(usually called $v_{1}$).
The radial velocity or Doppler shift method has an important limitation. It
does not allow to determine the position of the binaries’ orbits in space,
thus the observed velocities $v$ are not the true velocities $v_{0}$ of the
stars but their velocities projected into the line of sight. These two
quantities are linked to each other via the following relationship:
$v=v_{0}\leavevmode\nobreak\ sin\leavevmode\nobreak\ i$ (XVII.11)
where the inclination $i$ is the angle between the line of sight and the
normal of the orbital plane. This directly affects the measurement of the
radial velocity semi-amplitude of the binary $K$, which is also defined as
$K=K_{0}sin\leavevmode\nobreak\ i$, where $K_{0}$ and $K$ represent the true
and observed amplitudes, respectively. By using a simple mathematical
formulation we can understand what are the main properties of a binary
obtainable with the radial velocity method, and what differentiates SB1 from
SB2 binaries. If we consider a binary system where the two stars move in a
circular orbit around their common center of mass, and we assume that the
radii of the orbits of the two components are $a_{1}$ and $a_{2}$,
respectively, by using the equation of the center of mass $m_{1}a_{1}$ =
$m_{2}a_{2}$ and considering $a=a_{1}+a_{2}$, we can rewrite the orbital
radius of the first component as a function of the masses involved in the
system as follows:
$a_{1}=\frac{a\leavevmode\nobreak\ m_{2}}{m_{1}+m_{2}}$ (XVII.12)
The true semi-amplitude velocity $K_{0}$ can instead be rewritten as:
$K_{0,1}=\frac{2\pi a_{1}}{P},$ (XVII.13)
where $P$ is the orbital period of the binary. By substituting $K_{0}$ from
the above equation, we obtain:
$K_{1}=\frac{2\pi a_{1}\leavevmode\nobreak\ sin\leavevmode\nobreak\ i}{P}$
(XVII.14)
Substituting Eq. XVII.12 in Eq. XVII.14, we then obtain:
$K_{1}=\frac{2\pi a}{P}\frac{m_{2}sin\leavevmode\nobreak\ i}{m_{1}+m_{2}}$
(XVII.15)
By solving the equation for the semi-major axis $a$ of the binary and by
applying the Third Kepler’s law, we derive the mass function of the binary
system:
$f(m)=\frac{m_{2}^{3}\leavevmode\nobreak\ sin\leavevmode\nobreak\
i^{3}}{(m_{1}+m_{2})^{2}}=\frac{K_{1}^{3}P}{2\pi G}$ (XVII.16)
This formula, which only applies to binaries in circular orbits, can be
generalized to take into account the effect of eccentricity $e$ as follows:
$f(m)=\frac{m_{2}^{3}\leavevmode\nobreak\ sin\leavevmode\nobreak\
i^{3}}{(m_{1}+m_{2})^{2}}=\frac{K_{1}^{3}P}{2\pi G}(1-e^{2})^{3/2}$ (XVII.17)
The right hand side of Eq. XVII.16 (or Eq. XVII.17) can be fully inferred from
radial velocity data in a spectroscopic binary, as $K_{1}$, $P$ and $e$ are
all measurable quantities. In an SB2 binary, where the spectral lines of both
components can be observed, we measure not only $K_{1}$ but also $K_{2}$ and
an expression similar to Eq. XVII.16 (or Eq. XVII.17 in the general case) can
be written, with the labels reversed. By using Eq. XVII.15 and the definition
of the center of mass, we obtain the following:
$\frac{K_{1}}{K_{2}}=\frac{a_{1}}{a_{2}}=\frac{m_{2}}{m_{1}}$ (XVII.18)
This means that in an SB2 binary we can derive the mass ratio of the two
components. However, since we do not have any direct information about
$sin\leavevmode\nobreak\ i$, we cannot obtain the true masses $m_{1}$ and
$m_{2}$. Only if the true mass of one of the two stars is determined via
alternative methods (e.g. using photometric data to derive the spectral energy
distribution of the star, hence its mass), then the mass of the companion is
immediately deduced.
The situation becomes significantly more complicated in SB1 binaries, where
only the spectral lines of the brightest component can be observed. In this
case the mass ratio of the visible vs unseen object in the binary is unknown
and the only information that can be retrieved is the system mass function in
Eq. XVII.17. Even if we know the mass of the bright component (that we call
here primary), we are not able to determine the true mass $m_{0,2}$ of the
secondary unseen star, but only $m_{2}=m_{0,2}\leavevmode\nobreak\
sin^{3}\leavevmode\nobreak\ i$. This is the so-called
$m\;sin\leavevmode\nobreak\ i$ degeneracy, i.e. the same value of $m_{2}$ can
be produced either by a low inclination $i$ (close to a face-on orbit,
$i=0^{\circ}$) and a high mass or by a high $i$ (close to an edge-on orbit,
$i=90^{\circ}$) and a low mass for the star. There are a few cases in which
this degeneracy between mass and inclination can eventually be broken and we
will discuss them briefly in the following. In all other cases scientists
determine a minimum mass for the secondary unseen star, by assuming that the
binary orbit is perfectly edge-on, thus by assuming that it has an orbital
inclination of $90^{\circ}$. This way $sin^{3}\leavevmode\nobreak\
(i=90^{\circ})$=1 and this term can be removed. Every other inclination value
in the range between $0^{\circ}$ and $90^{\circ}$ provides a larger value for
the true mass of the secondary unseen companion.
As anticipated, there is a way to break the degeneracy between mass and
inclination in spectroscopic binaries, and that happens when the light
modulation of the system as a function of time (light curve) is observed
together with its radial velocity curve. More specifically:
* •
In eclipsing binaries, i.e. when the system inclination is such that one star
passes in front of the other and partially obscures the light coming from it.
The shape and depth of the eclipses can be uniquely solved to give the
inclination and hence the true masses of the individual stars.
* •
In non-eclipsing close binary systems, or when one component is not seen, then
the ellipsoidal modulation of the seen component depends on the mass ratio and
the inclination. Together with the radial velocity curve, this can then give
unique masses for the components.
Promising BH-stellar binary candidates are those that meet two criteria: i)
the periodic signal due to the motion of the star can be clearly detected but
the reflex motion due to a visible companion is not observed; and ii) the
analysis of the radial velocity curve suggests the unseen component to be more
massive than the observed one. However, before misinterpreting the data, some
caution is required as there may be several reasons why a visible companion,
although present, does not appear in the composite spectra (see the discussion
in the next two Sections for details).
Spectroscopic surveys are one of the main strategy for dynamically detecting
non-interacting binaries. Any spectrograph can be used to perform these
studies, with the most useful instruments for this purpose being either multi-
object spectrographs (e.g. GIRAFFE at the Very Large Telescope - VLT) or
spectrographs equipped with an integral field unit (IFU) such as the VLT/MUSE.
Both can be used to obtain simultaneous spectra of many objects, hence
performing blind but efficient searches. By observing the same field for a
number of times, and extracting a spectrum for a significant number of sources
in the field of view, it is possible to identify those objects showing radial
velocity variations and among them, those orbiting massive (and likely) dark
companions. Unfortunately, the only surveys that have multi-epoch spectroscopy
that is rich enough for a time-series analysis are APOGEE (Apache Point
Observatory Galactic Evolution Experiment [1236]), LAMOST (Large Sky Area
Multi- Object Fiber Spectroscopic Telescope [1237]), RAVE (RAdial Velocity
Experiment [1238]) and Gaia-ESO [1239], which came into play only in recent
years. This limitation has severely hampered the detection of BH-stellar
binaries in the past via this method. More surveys (WEAVE [1240] and 4MOST
[1241]) will be in operation in the coming years, surely providing a boost in
the field, thanks to a significant increase in both the number and depth of
the observations.
As mentioned above, an important caveat to keep in mind in this type of study
is that, when only radial velocity measurements are available, the binary
system cannot be fully characterized and only a minimum companion mass can be
inferred for the unseen source. This is due to the fact that the inclination
of the system is unconstrained. When radial velocities are instead
complemented with additional pieces of information, for example, data showing
the photometric variability of the system as a function of time (i.e. light
curves), then these degeneracies can be removed and the current mass of the
two components can be finally estimated. Unlike spectroscopy, there are many
photometric surveys, some even quite dated, which scan the sky both in the
direction of the Milky Way and of Galactic satellites (e.g. the Large and
Small Magellanic Clouds). Most of them have been designed to study variable
stars in different regions of the Universe and for different purposes. Among
the most popular we find VVV [1242], VMC [1243], MACHO [1244], EROS-2 [1245],
ASAS-SN [1246] and TESS [1247]. As already mentioned in Section XVI, OGLE also
has to be added to the list, which over the last twenty years has been an
extraordinary resource of interesting objects.
When talking about dynamically detecting quiescent BHs in binaries, it is
important to make a distinction between two different types of environment
where they can be found. They can be located in high density fields, like the
central regions of massive star clusters (dense collections of thousands to
million of gravitationally bound stars) or in low density fields, which are
the most abundant in the Universe. Various mechanisms are at play in these two
environments, so the BH population they host can be very different. Below we
will describe the low and high density fields separately, highlighting the
main results on the detection of quiescent BHs in both.
###### Quiescent Black Holes in low-density fields
Our Universe is largely very low-density, populated by collisionless systems,
such as galaxies, where stars are mainly moving in the collective
gravitational field, and interactions or collisions between individual stars
are extremely rare if at all absent. When a stellar-mass BH is detected in
this type of environments, it is very likely that it might be the remnant of
the death of a massive star that formed either isolated or in a ”primordial”
binary that exploded as a supernova, rather than the product of a more complex
formation mechanism (e.g., hierarchical assembly of low-mass BHs). The
population of BHs that can be sampled by randomly observing the field is as
close as possible to an “undisturbed” distribution but does not necessarily
resemble what we can find in other types of environments.
Many studies in the literature have postulated that millions of BHs should
exist in our Galaxy [1248, 1249, 1035, 1036, 1037, 1250, 1038], and detecting
a good fraction of them would be extremely important not only to understand
the physical processes that trigger their formation but also to characterize
their properties (mass distribution, spin), to date completely unknown.
Unfortunately, until very recently, no quiescent non-interacting BH-stellar
binaries have been found in radial velocity searches. As previously
anticipated, one of the reasons is the lack of dedicated spectroscopic surveys
capable of scanning the sky a sufficient number of times to capture a
significant number of sources to analyze. Another, even more critical, is the
difficulty of interpreting the results coming from the few available.
Casares et al. [1251] reported the first detection of an X-ray quiescent
stellar-mass BH of 3-6 $M_{\odot}$ in the Milky Way, in a binary system with a
classical Be star (i.e., fast rotating B star that displays strong emission
lines) in a mass range between 10 and 16 $M_{\odot}$. The system, MWC 656, has
been identified as the candidate counterpart of a potential $\gamma$-ray
flare, hence it does not represent the best example of a fully dynamically
detected system. Subsequent radio and X-ray emissions have been found to be
associated with the system, potentially strengthening the case for this to be
a BH. A re-analysis of the spectral variability properties of the source,
however, has recently questioned this interpretation, suggesting a hot
subdwarf as the most likely companion [1252]. Another interesting object, AS
386, has been studied by Khokhlov et al. [1253], showing H$\alpha$ emission
and near-infrared excess. Thanks to low and high-resolution spectroscopy and
multi-band optical and near-infrared photometry the authors have suggested
that the system is as follows: a binary of a B-type star surrounded by
circumstellar gas and dust, exhibiting the B[e] phenomenon and a high mass
($\sim$7 $M_{\odot}$) invisible companion which could not be anything other
than a BH. This system is one of the very few to date that has not been
questioned.
The first detection of a Galactic non-interacting stellar-mass BH in a binary
system with a sensibly smaller luminous companion was reported by Thompson et
al. [1254]. Indeed, the invisible companion has an estimated mass of $\sim$3.3
$M_{\odot}$ and it is in a binary system with a giant star (2MASS
J05215658+4359220). The system has an orbital period of $\sim$83 days, a semi-
amplitude velocity K$\sim$44.6 km/s, near-zero eccentricity and an inclination
close to 90 degrees. The detection has been possible thanks to the analysis of
the radial velocity curve of the system provided by APOGEE and TRES,
complemented by multi-band ASAS-SN and TESS light curves. In Figure 29, left
panel, both radial velocity and light curves of the binary system are shown.
Given the uncertainties in the estimated mass of the invisible companion, it
is compatible with both a BH or a NS, falling in the so-called mass gap
between these two types of objects [1255]. Right panel of Figure 29 shows the
constraints on the mass of the compact object based on the properties of the
visible star (L, Teff) and on the predicted inclination of the system itself
based on the observed photometric variability. Van den Heuvel & Tauris [1256]
analyzed the same set of data available for the system but came to a rather
different conclusion. Apparently, if the estimated mass of the giant star is
lower by a factor of 3 from what suggested by [1254] ($\sim$ 1 $M_{\odot}$),
the system would likely be a tertiary instead of a binary system. In this
case, the unseen component would be compatible with a close binary system,
made of two low mass main sequence stars, which would produce no detectable
signals in the spectral energy distribution, as well as providing a simple
explanation why no X-ray emission is detected. To date, none of the two
scenarios have been investigated further, hence it is still unclear whether
this system contains or not a BH. If confirmed, it would be an extremely
important discovery as it would help building statistics on these systems.
Figure 29: from Thompson et al. [1254]: Left panel: The ASAS-SN light curves
of the binary in 4 different filters, and the TRES radial velocity curve used
to constrain the orbital properties of the system. Right panel: The mass of
the unseen companion (the compact object) as a function of the mass of the
giant. The blue vertical shaded areas define the allowable mass range for the
giant given its Teff and L, while the solid black curves correspond to
different inclinations of the orbits. The black and white dots identify the
more likely masses for both components in the binary.
Starting in 2019, a number of non-interacting BH candidates have been reported
in the literature, most of which with associated massive companions (i.e.
O-type and B-type stars). For most of those systems, alternative explanations
have been proposed that do not require the presence of a BH in them. While the
alternative interpretation has been confirmed by further analysis/data for a
number of these objects, additional investigations need to be performed for
others, in order to provide a firm conclusion on the nature of these systems.
We briefly discuss some of these individual systems here as they represent the
most recent claims of BHs found in the field, and even when it was later
proved that a BH did not exist in the systems, they are still very useful
examples that help us understand 1) how challenging these types of studies
are; 2) what are the main caveats that scientists need to be aware of when
dealing with these systems; and 3) what are the main tests to be performed to
be sure that a stellar-mass BH actually resides in the system.
\- LB-1 is a Galactic binary system. According to Liu et al. [1257], it is
made of a massive BH of $\sim$68 $M_{\odot}$ orbiting around a B-type star
with period P = 78.9 $\pm$ 0.3 d, semi-amplitude velocity K = 52.8 $\pm$ 0.7
km/s, and eccentricity e = 0.03 $\pm$ 0.01. The radial velocity curve of the
system was measured with a sample of 26 LAMOST, 21 GTC and 7 Keck observations
obtained over two years. By studying the motion of the B-type star and the
accompanying H$\alpha$ emission line, that they interpreted linked to an
accretion disk orbiting around the invisible companion, the authors concluded
that the invisible companion should be relatively massive, hence it could only
be a BH. When published, this finding drew much attention from the scientific
community as it was the most massive stellar BH ever reported (even more
massive than theoretical predictions), with strong implications for our
interpretations of gravitational wave (GW) signals. A number of groups
independently analyzed the system from different perspectives, using
alternative techniques, and the final answer came shortly after: LB-1 does not
contain a BH. Instead, it is a rare binary system of a stripped star (the
former mass donor) and a Be star rotating nearly at its critical velocity (the
former mass accretor) [1258, 777, 1259, 1260]. The high rotation of the B-type
star creates a rotationally supported equatorial disk responsible for the
H$\alpha$ emission that Liu and collaborators wrongly attributed to the
presence of a BH in the system. Moreover, when a star rotates very fast, its
spectral lines widen due to this effect, so that the lines of a presumed
second bright companion can be easily overlooked, making it impossible to
detect and separate the two contributions. The use of the shift-and-add
spectral disentangling technique ([1261, 1262]) was fundamental to separate
the flux contribution of the two sources from the composite spectrum and guide
to the correct interpretation of the data.
\- HR 6819 is a naked-eye Galactic binary system. Using FEROS echelle spectra
and TESS light curves, Rivinius et al. [1263] suggested that HR 6819 is a
hierarchical triple made of a classical Be star in a wide orbit with an
unconstrained period around an inner 40 d binary consisting of a B3 III star
and a BH in a circular orbit. They measured a radial-velocity semi-amplitude K
of 61.3 km/s for the inner star and by assuming a minimum mass of $\sim$5
$M_{\odot}$ for it, they deduced a mass for the unseen object $>$ 4.2
$M_{\odot}$, indeed compatible with a BH. Similarly to what happened for LB-1,
also HR 6819 triggered lots of investigations within the community which ended
up with different teams independently suggesting that the system is not a
triple, but instead a binary system consisting of a stripped B-type primary
and a rapidly-rotating Be star formed from a previous mass transfer ([1264,
1265]). This is basically the same kind of system proposed for LB-1, although
the interpretation of the two systems was initially very different, and the
presence of the companion in the spectra has been overlooked for exactly the
same reason. The final word on whether or not HR 6819 contains a BH came very
recently, when new high-angular resolution data from the VLTI/GRAVITY revealed
the presence of a stellar companion with an angular resolution of $\sim$1.2
mas from the primary, that moves on the plane of the sky over a timescale
compatible with the known spectroscopic period of 40 d [1266]. Thanks to
further data and more sophisticated tools, it has now been established that
neither LB-1 nor HR 6819 contain a BH.
In the past couple of years, two more claims of quiescent BH detections in the
Milky Way have been reported in the literature: V723 Mon (“the Unicorn”,
[1267]) and 2M04123153+6738486 (“the Giraffe”, [1268]). Both of them have been
proposed to be binary systems of giant stars with low-mass/faint companions,
which they interpreted as BHs of a few solar masses, around 2-3 $M_{\odot}$.
Masuda and collaborators [1269] also presented the results of further
investigations in support of the dark nature of the companion in “the
Unicorn”. El-Badry and colleagues [1270], on the other hand, have recently
proposed an alternative explanation for both systems, namely that they are
stripped giants with subgiant companions. The authors used a combination of
spectral disentangling and fitting of binary spectral models to investigate
the presence of any luminous companion in both systems. They applied the
spectral disentangling technique to the Keck/HIRES spectra available for “the
Unicorn” and “the Giraffe” and were able to actually identify the presence of
an additional stellar source in both systems. In “the Giraffe”, the companion
was also detected in the near-infrared APOGEE spectra. They found the
secondaries to contribute about half the light in the optical, while they
dominate at the bluest observed wavelengths.
HD 96670 in the Carina OB2 association is one of the last binary systems
claimed to contain a quiescent stellar-mass BH [1271]. It is a binary system
of a $\sim$22 $M_{\odot}$ O-type star orbiting around a $\sim$7 $M_{\odot}$ BH
with an orbital period of $\sim$5 d and possibly a third component. Similarly
to the case of MWC 656, an emission at high-frequency (in X-rays) was found to
be associated with the system triggering further photometric and spectroscopic
analyses, hence the detection, if later confirmed from further analyses, would
not come from a blind dynamical search.
Very recently, a new X-ray quiet BH (VFTS 243) has been discovered in the
field of the Large Magellanic Cloud (LMC), in the direction of the Tarantula
nebula, to be the companion of a massive (M$\sim$25 $M_{\odot}$) O-type star
[566]. The authors have used both radial velocity curves from the Fibre Large
Array Multi Element Spectrograph (FLAMES) at the VLT and OGLE light curves to
constrain the nature of the binary system. In particular, it has an orbital
period P$\sim$10.05 d, an almost circular orbit (e$\sim$0.017) and a semi-
amplitude velocity K$\sim$81 km/s. Shenar and collaborators have applied a
spectral disentangling technique to set an upper limit to the flux
contribution of any putative luminous companion and found that any luminous
companion would exceed this limit significantly. From the analysis of the
light curves and the very small ellipsoidal variations they estimated the
possible inclination of the system to be larger than 40 degrees and the mass
of the unseen companion to be of at least 9 $M_{\odot}$, which can only be a
BH. Under the same observational campaign targeting the Tarantula nebula in
the LMC, Shenar et al. [1272] have also looked at other SB1 binaries and they
have identified two other binaries hosting as many putative BH candidates.
Their nature, however, is much more uncertain and a proper study of these
systems is required.
Given the number of binaries with quiescent BHs that have been questioned in
the recent years, Mahy et al. [1273] have suggested a method to overcome this
problem in the future, which is based on the combination of 1) high-resolution
spectroscopy, 2) high-precision space-based photometry, and 3) state-of-the-
art spectral disentangling. These three ingredients together can allow to
actually constrain the nature of unseen companions in systems classified as
SB1 in the literature. To show the power of this approach, they selected a
sample of 32 Galactic binaries hosting an O-type or an early B-type star. This
sample contained also two systems that, from X-rays observations, are already
known to host a BH (Cyg X-1 [1274, 1275]) and a NS (HD 74194 [1276]),
respectively. Their analysis demonstrates that, when good data are available,
even very faint companions can be detected. They found a secondary luminous
companion in 17 out of 32 binaries in their sample, while in the remaining
systems they were not able to detect any secondary star, either because it was
too faint, or because it was dark. The two most promising candidates hosting a
BH were actually Cyg X-1 and HD 130298, where the estimated mass of the unseen
companion was more than 7 $M_{\odot}$. We know that Cyg X-1 hosts a BH since
many years now, while HD 130298 was presented as another system likely hosting
a quiescent BH candidate. Further multi-wavelength observations will confirm
the nature of the latter but thus far it represents one of the few promising
candidates in the literature. Two more binaries have been recently identified
by Fu et al. [1277] as potential hosts of as many compact objects, by
inspecting the GAIA DR3 catalog and modelling TESS light curves of the
targets, but even in this case there is no certainty about the nature of the
companions. Follow-up observations will establish this.
To conclude, currently no quiescent BHs in stellar binaries have been
unambiguously unveiled in the Milky Way field through a purely dynamical
approach based on radial velocity searches, other than that reported by
Khokhlov et al. [1253] and by Mahy et al. [1273] (high-mass companions). In HD
96670, the only other system not questioned so far, a counterpart at high
frequency has triggered the dynamical search, so technically we cannot
classify it as a binary with a dormant BH. The only extra-galactic quiescent
BH detected and confirmed to date is that recently reported in the LMC field
by Shenar et al. [566]. The near lack of BHs detected so far is interesting
but not unexpected, given that they are all individual detections based on
relatively small samples. It is even more justified if we consider how
challenging these studies are.
It is interesting to note, however, that most of the binary systems so far
proposed to host a dormant BH (confirmed or later disputed) have relatively
massive O- and B-type star companions. Due to the large masses involved, these
systems are by far the most difficult to model, thus being the most prone to
interpretation errors. On the other side, they are the ones with the highest
expected fraction of BH companions. Indeed, recent theoretical calculations
predict that about 3% of massive O or early-B stars in binary systems should
have a dormant BH as a companion [1278, 1279]. If this theory is correct,
large numbers of these binaries are actually hiding out in the Universe and it
may be only a matter of time before they are detected. However, it is
important to underline that O and B stars are very massive and short-lived
objects. The reason they are largely undetected may simply be an observational
limitation: their lifetime are short enough to make their observations
unlikely. An intriguing aspect to mention is also that many of these systems,
at some point in their evolution, may effectively become double degenerate
binaries (i.e., binaries with two compact objects such as BH-NS or BH-BH
binaries), thus the possible progenitors of GW sources.
In light of these theoretical studies [1278, 1279], the small fraction of
quiescent BHs detected so far in the field, the increasing number of
interacting BHs in binaries detected through X-ray or radio observations, and
the large number of GW signals produced by BH-NS or BH-BH binaries revealed by
the LIGO-VIRGO-KAGRA consortium since 2015 [1280] can probably be reconciled.
The main reasons for the apparent discrepancy we see in their numbers may
indeed be the following: 1) the latter detection methods are much more
efficient in identifying stellar-mass BHs as their signature is clear and
cannot be misinterpreted; 2) in comparison, radio and X-ray detectors tend to
scan a much larger area of the sky and for GWs facilities it is even more the
case. The study of these complementary detections (see the related sections of
this Chapter) is very instructive, since it is expected that the local density
of BHs in the field is made up of BHs randomly but almost equally distributed
in these three different groups: 1) non-interacting BHs; 2) interacting BHs;
3) BH-BH binaries, hence there is real hope that future investigations may
finally reveal the conspicuous number of these quiescent BHs that are now
missing.
###### Quiescent Black Holes in high-density fields
In a Universe predominantly dominated by low-density structures, high-density
fields such as star clusters are very special environments, where specific
mechanisms can occur, triggering interesting phenomena or the formation of
exotic objects (such as cataclysmic variables, blue straggler stars or
millisecond pulsars [1281, 1282, 1283]) that are not observed elsewhere. Star
clusters are a class of cosmic objects made up of $10^{3}-10^{6}$
gravitationally bound stars, which share nearly the same metallicity and age
among them, since they are born during the same burst of star formation (at
least as a first approximation, but see the review by [1284] for a recent
discussion about the multiple population phenomenon in star clusters).
Due to the very high number of stars located in a very limited volume, star
clusters are collisional systems, meaning that all stars during their
lifetimes experience several interactions with other cluster members, both
stellar or compact objects. These close interactions are the main engines of
their dynamical evolution and may lead to changes in the properties of the
clusters themselves with time but also trigger the formation and destruction
of binary systems, which are present in significant numbers within them. In
addition, there is observational evidence that more massive stars (e.g., main-
sequence OB stars) are found more preferentially in binaries or higher order
multiples than lighter ones (see [363, 95, 365]). Initially made up of many
massive stars, star clusters are the ideal places to look at for finding
stellar-mass BHs, both interacting and dormant. In this regard, massive star
clusters have recently been placed under the magnifying glass, after the
GW190521 event was detected [51, 50]. This was the first intermediate mass BH
discovered ($\sim$150 $M_{\odot}$), which formed from the merger of two
massive BHs (60 $M_{\odot}$ and 85 $M_{\odot}$, respectively). Interestingly,
these progenitor BHs might be themselves the product of a previous binary BH
merger or stellar collision (as single star evolution struggles to produce
such massive BHs; refer to Chapter 1, Section 1a, for an exhaustive discussion
of the topic). Two more GW events of this type (GW190426 and GW200220) have
been detected in recent years that involved very massive BH candidates, which
merged to produce final BHs of more than 100 $M_{\odot}$, or even close to 200
$M_{\odot}$. BHs in star clusters have the real chance to pair up dynamically,
thus high-density environments can actually be considered potential nurseries
of many GW sources.
Until very recently, the innermost regions of star clusters, both in the Milky
Way and in external galaxies, were completely inaccessible to perform
spectroscopic studies, due to observational limitations, so until 2018 no
dynamical detections of dormant stellar-mass BHs in clusters were claimed.
Thanks to spectrographs equipped with IFUs, such as MUSE at the VLT (the best
currently available in terms of field coverage and both spatial and spectral
resolution), scientists have started to observe the innermost regions of
clusters, trying to resolve and extract a spectra for every single source
detected. The central regions are the most interesting for this type of
science, as this is where massive and compact objects are most likely to be
found, due to a mechanism at play in star clusters called mass segregation,
i.e. more massive objects sink into the center while lighter sources tend to
lie on the outskirts.
Giesers et al. [1285] used MUSE multi-epoch spectroscopic data of the old
($\sim$ 12 Gyr) Galactic globular cluster NGC 3201 spanning a time baseline of
a few years to start a blind systematic survey of binaries and, among them,
those with likely dark companions. Performing a blind search is very powerful
for two main reasons: 1) we know that the highest chance to detect stellar-
mass BHs is in the center compared to the outskirts, but we do not know
exactly where they are in the cluster (at which direction / what distance from
the center); 2) we are not biased toward detecting only a specific sample of
objects, which always happens when handling targeted observations.
Using this novel technique, they were able to detect a stellar-mass BH in the
cluster, in an orbit rather eccentric (e$\sim$0.6) with a period of $\sim$167
d and a semi-amplitude velocity K$\sim$70km/s with a subgiant star companion.
Since the luminous star was part of the cluster, they were able to exploit the
advantage of knowing exactly the age, the metallicity, the distance and the
extinction in the direction of the cluster, to infer the mass of the luminous
star through a comparison with stellar evolutionary models (MIST, [152]). They
found a mass of $\sim$0.8 $M_{\odot}$ for the star, and thanks to the orbital
properties of the binary system, they derived a minimum mass of 4.36
$M_{\odot}$ for the unseen companion, which can be confidently claimed to be a
BH (see Figure 30). After analyzing the entire sample of binaries in NGC 3201,
they identified another BH candidate and a binary whose orbital solutions were
compatible with a BH or NS companion. These solutions have been published
along with the properties of several other stellar binaries in [1286].
This was a very important result for two reasons: 1) it represented the first
detection of non-interacting stellar-mass BHs in star clusters; 2) given the
old age of NGC 3201, these detections were a further confirmation that BHs can
actually reside in star clusters for a Hubble time or more. The latter is in
contrast with the general understanding of star clusters in the early 1990s.
In fact, it was believed that, due to interactions and dynamical processes,
any BH formed within them would be expelled, leaving star clusters essentially
devoid of these objects in a very short timescale (t $<$ 1 Gyr, [1191]). On
the other side, the results by Giesers and colleagues [1285, 1286] are in good
agreement with what more advanced and sophisticated cluster models have
actually predicted, i.e. star clusters are able to retain their BHs for longer
times, even longer than a Hubble time [1287, 1288, 1289, 1290, 1291].
Figure 30: from Giesers et al. [1285]: The radial velocity curve of the first
BH candidate detected in NGC 3201, from MUSE observations. The red solid line
represents the best fit solution for the system, thanks to a comparison with a
large library of possible orbits. The radial velocities of a nearby cluster
member showing no signs of variability (i.e. a single star) are presented in
blue for comparison.
Unlike in the field, uncovering the BH population of star clusters can
actually help to answer some crucial long-standing questions about stellar
astrophysics, such as: 1) how strong are natal kicks in clusters? 2) does a
merger event always produce a natal kick? 3) how likely BHs are to be ejected
from globular clusters through dynamical interactions? Answering these
questions would be important to understand, for example, if globular clusters
are the main factories of GWs, but also to investigate how likely it is that
BHs with masses larger than 100 $M_{\odot}$ (like GW190521) are formed (via
hierarchical merging of lighter stellar-mass BHs) and are held within globular
clusters, or whether denser environments with deeper gravitational potential
are required for this, such as nuclear star clusters [1292].
Until very recently, no other attempts have been made to reveal further
dormant BHs in star clusters using the radial velocity method. Saracino et al.
[1293] exploited the approach developed by Giesers and collaborators to look
for any evidence of stellar-mass BHs in a massive ($\sim 10^{5}$ $M_{\odot}$)
cluster of $\sim$100 Myr located in the LMC. They chose a young cluster, where
dynamical processes have not yet occurred, to study the initial distribution
of BHs in clusters, which is still completely unknown. In this work, the
authors claimed to have detected a binary system consisting of a BH of
$\sim$11 $M_{\odot}$ orbiting almost circularly a main-sequence B-type star
with a period of 5.04 d and a semi-amplitude of K $\sim$ 140.5$\pm 3$ km/s,
which they named NGC1850 BH1. This finding was the result of a combined
analysis of the radial velocity curve constructed from 17 multi-epoch
observations with MUSE and the I and V light curves provided by OGLE-IV. Since
the binary is a member of the cluster and its position in the CMD is
consistent with a main sequence turn-off star, the authors derived a mass for
the luminous object of $\sim$4.9 $M_{\odot}$ from the comparison with stellar
isochrones of age, metallicity, extinction and distance of NGC 1850.
Right after the detection, an alternative explanation has been proposed for
the binary system NGC1850 BH1 [1294, 1295], which does not require the
presence of a BH. Both groups suggested that this system is a post-mass
transfer binary in a rare state of its evolution and it consists of a bloated-
stripped star and a more massive (possibly rapidly rotating) B-type star. They
argued that the stripped star is about 5 times less massive than what Saracino
and collaborators assumed, but that its luminosity is still consistent with
that of the observed star. Furthermore, they predicted that the massive B-type
star would be expected to contribute 10 to 100% to the total flux in the MUSE
spectra and would be overlooked due to the rapid rotation of the star itself.
After a re-analysis of the MUSE spectra, the original authors [1296] confirmed
that the visible star has a lower mass compared to what initially thought but
they also made the following findings: 1) a larger radial velocity semi-
amplitude ($K_{\rm 2,revised}=176\pm 3$ km/s) than in the discovery paper,
which translates in a mass function $f_{\rm revised}$=2.83 $M_{\odot}$ for the
system; 2) the unseen source contributes less than 10% to the total light in
the visual; 3) an accretion disk seems to be present in NGC1850 BH1. These new
observational constraints indicate that, regardless of the mass of the visible
star, the unseen companion must be rather massive. Its nature, however, is
still unconstrained. This could either be a BH with an accretion disk or a
massive luminous star enshrouded in an optically think disk that almost
completely blocks its light. Further data and modelling are needed to provide
a definitive answer as to whether or not NGC1850 BH1 contains a BH.
A recent addition to the list of dormant BHs detected in star clusters is NGC
2004-115 [1297]. It lies on the periphery of the young cluster NGC 2004, on
the northern edge of the LMC. The authors used VLT/FLAMES and SALT-HRS
spectroscopic data covering a baseline of about 20 years to shed light on the
nature of the system: a triple system consisting of an inner binary with
period P=2.92 d, semi-amplitude velocity K$\sim$62.4 km/s and mass function
f(m)=0.07 $M_{\odot}$. By assuming the visible star is a B-type main sequence
of 8.6 $M_{\odot}$, they inferred a mass for the invisible companion greater
than 25 $M_{\odot}$. An upper limit on the X-ray luminosity coming from the
binary has been set by XMM-Newton and it is consistent with having a quiescent
BH in the system. The presence of a massive BH in NGC 2004-115 was later
questioned by [1298] as such a high mass results from an inferred orbital
inclination of only 9 degrees for the system, which is not consistent with the
modelling of both the OGLE and the MACHO light curves of the system. A triple
system in which the 25 $M_{\odot}$ BH is replaced by a 2-3 $M_{\odot}$
luminous companion, and with an outer Be star appears to be the only scenario
capable of reproducing the observed light curves and radial velocity curves of
the system simultaneously.
As part of a multi-epoch VLT/FLAMES spectroscopic campaign aimed at
constraining the binary fraction of OB stars in high density environments,
Banyard et al. [1299, 1300] have recently provided an important contribution
to the search for quiescent BHs in clusters. Indeed, they studied the
population of B stars in the young (7 Myrs) open cluster NGC 6231 in the Milky
Way, measuring an observed binary fraction of 33 $\pm$ 5% (which increases to
52 $\pm$ 8% if corrected for the observational biases) and characterised the
multiplicity properties of the identified binary systems. By integrating
medium-resolution optical spectroscopy, ground-based photometry and high-
cadence space-based photometry, the authors found that 7 of the 15 SB1 systems
in the cluster are indeed SB2 systems with mass ratios as low as q = 0.1. Of
the 8 remaining systems, they ruled out 4 ambiguous systems and identified a
sample of 4 binaries with a high probability of harboring a compact object.
They calculated the minimum mass for the unseen companions, finding that two
have masses between 1 and 3.5 $M_{\odot}$ and fall within the NS mass range,
while the other two have mass ranges between 2.5 to 8 $M_{\odot}$ and 1.6 to
26 $M_{\odot}$, respectively, and so are possible BH candidates. The most
massive unseen source is part of the CD-41 11038 system. However, as the
authors themselves mention in the paper, the detection limits of their
methodology do not exclude that some or all of these candidates have lower-
mass main sequence companions, stripped star companions or are part of triple
systems. Follow up photometric and/or interferometric observations are
therefore needed to answer this question.
To recap, only the two (possibly three) BH candidates in NGC 3201 are
currently recognized as robust non-interacting BH detections in star clusters
within the scientific community, but NGC 6231 appears to be another promising
source of compact objects, again underscoring the fact that crowded fields
such as clusters are ideal places to look for these degenerate sources. The
methods that Giesers and Banyard have developed together with their
collaborators are extremely promising and have the potential to lead to the
discovery of many more of these objects in other clusters if such a systematic
search is performed. The work by Giesers et al. is especially important in
light of the new instruments that will be available to the community in the
coming years, starting from HARMONI at the ELT in 2027 [1301]. This IFU
spectrograph will have a superb spatial resolution (up to 0.004 arcsec,
compared to 0.2 arcsec of MUSE-WFM) and thanks to the extremely large
collecting area will be able to observe and resolve the central regions of
star clusters more distant than the Magellanic Clouds into individual stars, a
science impossible to accomplish with current facilities. By overcoming the
crowding issue in nearby star clusters, scientists will be able to look for
non-interacting stellar-mass BHs in binaries where the luminous stars are much
fainter on the main sequence than has been possible hitherto.
To conclude and improve the readability of Section XVII.2.1 (a, b), we present
in Table 2 a summary of all the systems discussed, where those hosting
quiescent BHs are included, together with those subsequently questioned and
proven to be BH imposters.
Table 2: A summary of all systems claimed to date to host a quiescent BH is
presented below, including only those systems detected via spectroscopic
investigation. Both low-density and high-density environments are included,
and the systems are classified and catalogued based on the estimated mass of
the luminous companions. For completeness, confirmed BHs and BH imposters are
listed.
Name | Z | Envr. | Proposed | Alternative(s) | References
---|---|---|---|---|---
High - mass | | | | |
companions | | | | |
LB-1 | MW | Field | B + BH | Be + stripped (str) | Liu+[1257]; Shenar+[1259]; El-Badry[1260]
MWC 656 | MW | Field | Be + BH | Be + sdO ? | Casares+[1251]; Rivinius+[1252]
NGC 1850 BH1 | LMC | Cluster | B + BH | B + str ? | Saracino+[1293, 1296]; El-Badry+[1294];
| | | | | Stevance+[1295]
HR 6819 | MW | Field | (B+BH)+Be | Be + str | Rivinius+[1263]; Bodensteiner+[1264];
| | | | | El-Badry+[1265]; Frost+[1266]
NGC 2004 # 115 | LMC | Cluster | (B+BH)+Be | (B+A) + Be ? | Lennon+[1297]; El-Badry+[1298]
HD 96670 | MW | Field | (O+BH) + ? | \- - - | Gomez+[1271]
AS 386 | MW | Field | B[e] + BH | \- - - | Khokhlov+[1253]
VFTS 243 | LMC | Field | O + BH | \- - - | Shenar+[566]
HD 130298 | MW | Field | O + BH | \- - - | Mahy+[1273]
Low - mass | | | | |
companions | | | | |
NGC3201 #12560 | MW | Cluster | MS + BH | \- - - | Giesers+[1285]
NGC3201 #21859 | MW | Cluster | MS + BH | \- - - | Giesers+[1286]
NGC3201 #5132 | MW | Cluster | MS + BH/NS | \- - - | Giesers+[1286]
Unicorn (V723 Mon) | MW | Field | RG + BH | RG + str subgiant ? | Jayasinghe+[1267]; Masuda+[1269];
| | | | | El-Badry+[1270]
Giraffe (2M0412) | MW | Field | RG + BH | RG + str subgiant ? | Jayasinghe+[1268]; El-Badry+[1270]
2MASS | MW | Field | RG + BH/NS | RG + (MS+MS) ? | Thompson+[1254]; Van den Heuvel+
J05215658+4359220 | | | | | [1256]
##### XVII.2.2 Astrometric detection of quiescent Black Holes
As already explained in the previous Sections, the dynamical detection of a
stellar-BH binary totally relies on the possibility to obtain multi-epoch
spectroscopic data for the source of interest or for a very large sample of
objects in order to conduct a blind search. When they are not available,
alternative methods can instead be used to detect a quiescent BH orbiting
around a luminous companion. Indeed, binary systems can also be detected
“astrometrically”, by measuring the perturbations in the motion of the stars
by their unseen companions. This technique has been used previously to
identify “ordinary” binary systems and exoplanets, but can also be extended to
discover any kind of binaries and it is particularly useful in identifying
non-interacting compact binaries. Briefly, astrometric binaries are systems
made of two sources visible in astronomical observations as single sources.
Three are the main reasons why the companion source is not observed: it is
either too faint to be seen or too close to the primary to be resolved as two
stars. Alternatively, it is a compact remnant (a BH or a NS) and cannot be
observed due to its dark nature. What is observed in these cases is that the
primary has a telltale wavy motion across the sky inconsistent with it being a
single star. It instead wobbles about an unseen centre of mass, clearly
indicating that another object exists in the system and gravitationally
affects the motion of the bright, observed star. By analysing the motion of
this star over time it is possible to place important constraints on the
orbital properties of the system and on the masses involved.
As discussed in Section XVII.2.1, the dynamical detection of binaries through
radial velocity measurements allows for a robust estimate of the mass function
of the system, but unfortunately does not provide any constraints on its
inclination. This limitation is effectively exceeded by astrometry, as it also
allows measuring the binary’s orientation in space, including its inclination
angle. This means that if the mass of the luminous component can be derived
from independent analyses/data, this is sufficient to break the degeneracy
between the different parameters and have a complete characterization of the
system [1302]. In order to detect astrometric binaries, it is necessary to
measure the position on the sky of many objects and to trace their motion over
time. Moreover, this has to be done with extremely high precision in order to
pick up even very small displacements. In this respect, Figure 31 shows the
astrometric motion of a 1 $M_{\odot}$ visible star orbiting around an unseen
companion of 1.4 $M_{\odot}$ for different orbital periods (increasing from
left to right) and proper motion (increasing from top to bottom), from Andrews
et al. [1302]. The astrometric effects are amplified as the binary is placed
at a distance of only 35 pc.
Figure 31: from Andrews et al. [1302]: The astrometric motion of a luminous
star of 1 $M_{\odot}$ orbiting around an unseen companion of 1.4 $M_{\odot}$,
for different orbital periods (from 10 d to 10000 d) and two different proper
motions (10 mas/yr and 50 mas/yr). The binary is placed at a relatively close
distance of 35 pc to magnify any astrometric effects and the black dots show
the positions of the star over many GAIA epochs. For orbital periods much
shorter than a year, the orbital motion and parallax can be disentangled.
However, for P = 100-1000 d, the positions become extremely complex.
The Hipparcos space astrometry mission led by the European Space Agency (ESA)
was a pioneering project in this respect and it was able to pinpoint the
positions of more than 100,000 stars with high precision, opening up the
concrete possibility of detecting new binaries using this method. Several
astrometric binaries have been detected thanks to this mission in the
literature [1303, 1304, 1305] but there have been no claims of binaries
containing a non-interacting BH companion. Now that Hipparcos has been
replaced by the new ESA space astrometry mission GAIA, the ability to detect
quiescent BHs by analyzing the “wobble” in their proper motions over time is
more real than ever (see [1306] and [1307]). The GAIA satellite was launched
at the end of 2013 with the aim of providing 3D motion, luminosity,
temperature and composition of almost a billion stars in our Galaxy and
beyond, and so far it has been a revolution in many astronomical fields (e.g.
galactic archaeology).
Although the first GAIA astrometric orbital solutions for binary sources have
been published only in its third release in June 2022, many scientists have
started to analyze the data searching for a way to uncover astrometric BH
binaries well before this date. As an example, [1308] focused on the number of
OB + BH systems (i.e. BHs in binary systems with massive companions, $>$ 8
$M_{\odot}$) that GAIA would have been detected as binaries in the DR3 and
DR4. They took as a reference the 2nd Alma Luminous Star Catalogue (ALS II)
containing a sample of massive OB-type stars in GAIA and estimated that 77% of
the OB + BH binaries in ALS II would be detected as binaries in DR3, of which
89% would be unambiguously identifiable as OB + BH binaries. These systems
would cover an uncharted parameter space of long-period binaries (10 d $<$ P
$<$ 1000 d). They predicted that these numbers would increase to 85% and 82%,
respectively, by the end of the mission, as they would also detect systems
with both shorter and longer periods than those predicted in DR3. It is
interesting to note that, after an accurate cleaning procedure performed by
the GAIA team, most of these sources have been removed from the DR3 sample,
hence we do not have any information on these sources yet. However, GAIA DR3
still represents the game changer in this field, as scientists from around the
world have finally access to orbital solutions (periods, eccentricities, etc.)
for a sample of over $10^{5}$ astrometric binaries [1309, 1310, 1311]. The
analysis is still ongoing but this huge data set will grow and refine year by
year, with the concrete possibility of starting to identify which of these
astrometric binaries may actually contain massive invisible companions, i.e.
compact objects such as NSs or BHs.
Andrews et al. [1312] performed a first attempt in this direction: they
analyzed the whole sample of GAIA astrometric binaries under the conservative
assumption that the companions were dark and therefore the photocenters of the
systems followed the observed stars. They assumed the luminous stars to be of
1 $M_{\odot}$ and selected only those systems having a companion more massive
than 1.4 $M_{\odot}$, this to exclude any contamination from white dwarf-like
objects. After removing binaries with suspiciously long orbital periods
relative to the duration of the mission to date and applying further cuts,
they ended up with a sample of 24 binaries with companion masses ranging from
1.35 to 2.7 $M_{\odot}$, compatible with both NS or low-mass BH companions.
These systems show interesting characteristics, like periods of a few years
and orbital velocities of the order of $\sim$20 km/s. For at least 8 of these
objects a subsequent spectroscopic follow-up was performed and the inspection
of their spectra has confirmed that there is no evidence for a second luminous
component.
On the same line but using a different approach, Shahaf et al. [1313] also
looked at the sample of GAIA astrometric binaries searching for possible
compact object companions. Their triage technique, already proposed in [1314],
relies on a good knowledge of the mass-luminosity relation (for main-sequence
stars only), in order to discard all the single stars that for one reason or
another are contaminating the sample. They compiled a sample of 177 systems
with highly-probable non-luminous massive companions, only 20% of the total
sample provided by GAIA. Among them, they identified 8 BH candidates with
masses larger than 2.4 $M_{\odot}$. The most promising candidate among them
was soon after the subject of two independent spectroscopic follow-ups which
recently confirmed its dark and compact nature, estimating a mass for the BH
of approximately 10 $M_{\odot}$ [1315, 1316]. As in [1315], we refer to the
source as GAIA BH1.
GAIA BH1 represents the first BH identified via its astrometric orbital
solution from GAIA as well as the nearest BH ever detected (in the solar
neighborhood, d = 474 pc). Although a slight discrepancy in the predicted mass
of the BH between the two groups, they both confirm that it is in a binary
system with a slowly-rotating G dwarf (Teff = 5850 K, log g = 4.5, and M =
0.93 $M_{\odot}$) and the system has an orbital period of 185.6 d. The
predicted motion of GAIA BH1’s photocentre on the sky over a 6-yr window and
its predicted orbit are shown in Figure 32, from [1315]. This detection is
pretty robust, as consistent results have been obtained by using 1) GAIA
astrometric data alone; 2) spectroscopic data alone (acquired as a follow-up
program to the astrometric detection); or 3) a joint analysis of both
measurements. The latter obviously provided the most stringent constraints on
the properties of the system.
A very similar claim has been made almost at the same time from another group
[1317]. They analyzed a sample of $\sim$65,000 binaries with both time-
resolved spectroscopic and astrometric data in Gaia DR3 and identified a
peculiar system made of a red giant star (Gaia DR3 5870569352746779008) and an
unseen companion of a minimum mass $>$ 5.25 $M_{\odot}$. Although the mass of
the visible star was rather uncertain, they argued that the unseen source is a
BH candidate at a 99% confidence level, based on the orbital parameters of the
system. Their conclusions were entirely dependent on the reliability of the
Gaia solution.
El-Badry and collaborators [1318] obtained additional spectra for this
specific target, in the course of a broader program to spectroscopically
follow-up Gaia DR3 binaries suspected to contain compact objects. These data
confirmed the compact and dark nature of the invisible companion, which they
called Gaia BH2. They were also able to better constrain the properties of the
system itself. A bright (G = 12.3), nearby (d = 1.16 kpc) star on the lower
red giant branch with mass M = 1.07 $\pm$ 0.19 $M_{\odot}$ orbiting a BH of
mass $M_{BH}$=8.9 $\pm$ 0.3 $M_{\odot}$ with a moderate eccentricity e=0.52
and a relative low inclination angle i = 35°. Interestingly, this system has
the longest period (1277 $\pm$ 1 days) among those discovered so far, slightly
lower than what predicted by the Gaia solution (1352.25 $\pm$ 45.50 days) and
reported by Tanikawa and colleagues [1317].
Figure 32: from El-Badry et al. [1315]: Left: The predicted astrometric orbit
of the photocentre of GAIA BH1 on the sky over a period of 6 years. The black
line indicates the total motion due to proper motion, parallax, and orbital
motion while the dotted blue line shows the contribution due to parallax and
proper motion alone. The predicted motion of the source, with parallax and
proper motion removed is instead shown in the middle panel based on the joint
astrometry and RV solution and in the right panel by using the pure astrometry
solution. The joint analysis provides the most stringent constraints on the
system.
Despite extensive research, to date GAIA BH1 and BH2 appear to be the only two
quiescent BHs enclosed in the Gaia DR3 sample of binaries that can be
detected. Figure 33 shows where the population of known accreting BHs detected
via X-ray observations are located as a function of period and distance from
us. When Gaia BH1 and Gaia BH2 are included (large black points) they occupy a
very different parameter space based on their properties. They have
significantly longer periods and closer to Earth than the rest. We do not have
enough statistics to make any claims, but it appears that Gaia BH1 and BH2 may
be the tip of the iceberg of the hitherto totally unknown binary population of
wide-orbit black holes.
Figure 33: from El-Badry et al. [1318]: Gaia BH1 and BH2 (black points) are
compared to Galactic BHs detected via accretion (in X-rays) in the reference
plane of distance and orbital period. As indicated in the legend, the red and
blue dots represent LMXBs and HMXBs respectively, while the cyan and purple
dots identify detached BHs detected so far either in the field or in star
clusters. As can be seen, the systems discovered by GAIA are in a part of this
parameter space unexplored by other known BHs, i.e. with longer periods and
closer to Earth than the rest of BHs.
Gaia BH1 and Gaia BH2 unfortunately will not be progenitors of any GW signals
as the luminous companion of the compact object is a low-mass star and there
is no chance that the system may become a BH-NS binary at any point in time.
However, the study of these systems remains extremely important as it allows
us to build a more complete view of the actual population(s) of NSs and BHs
existing in the Universe and if (at all) their properties change as a function
of the environment.
The release of Gaia DR4 data in late 2025 as well as further investigations of
GAIA DR3 will probably reveal many other interesting systems (e.g. massive
white dwarfs and NSs). For what concerns BHs, based on recent theoretical
predictions, Chawla et al. [1319]) argued that the extended Gaia mission can
astrometrically resolve $\approx$ 30-300 detached BH-stellar binaries in the
Milky Way, with a high probability of finding even some compatible with being
|
# Navigation Turing Test (NTT): Learning to Evaluate Human-Like Navigation
Sam Devlin Raluca Georgescu Ida Momennejad Jaroslaw Rzepecki Evelyn Zuniga
Gavin Costello Guy Leroy Ali Shaw Katja Hofmann
###### Abstract
A key challenge on the path to developing agents that learn complex human-like
behavior is the need to quickly and accurately quantify human-likeness. While
human assessments of such behavior can be highly accurate, speed and
scalability are limited. We address these limitations through a novel
automated Navigation Turing Test (ANTT) that learns to predict human judgments
of human-likeness. We demonstrate the effectiveness of our automated NTT on a
navigation task in a complex 3D environment. We investigate six classification
models to shed light on the types of architectures best suited to this task,
and validate them against data collected through a human NTT. Our best models
achieve high accuracy when distinguishing true human and agent behavior. At
the same time, we show that predicting finer-grained human assessment of
agents’ progress towards human-like behavior remains unsolved. Our work takes
an important step towards agents that more effectively learn complex human-
like behavior.
Machine Learning, ICML
## 1 Introduction
Developing agents capable of learning complex human-like behaviors is a key
goal of artificial intelligence research (Winston, 1984). Progress in this
area would lead to new applications, from engaging video games (Soni &
Hingston, 2008), to intuitive virtual assistants (Cowan et al., 2017) and
robots (Scheutz et al., 2007). Yet, despite striking successes in training
_highly skilled_ agents (Vinyals et al., 2019), quantifying and making
progress towards learning _human-like_ behavior remains an open challenge.
We propose an automated Navigation Turing Test (automated NTT or ANTT), a
novel approach to learn to quantify the degree to which behavior is perceived
as human-like by humans. While human judges can be highly accurate when
judging human-like behavior, their scalability and speed are limited. As a
result, it is prohibitive to use human assessments in, e.g., iterations of
training machine learning-based agents. We aim to remove this bottleneck,
providing a scalable and accurate proxy of human assessments.
We demonstrate our NTT approach on the complex task of goal-directed
navigation in 3D space. This task is highly relevant and practical because:
(1) it has been studied from a wide range of perspectives in both biological
(Banino et al., 2018; de Cothi, 2020) and artificial (Alonso et al., 2020;
Chaplot et al., 2020) agents, so is relatively well understood; and (2)
artificial approaches achieve a high level of task success (Alonso et al.,
2020), so proficient agents can be obtained enabling us to focus on human-
likeness. While the majority of research on learning 3D navigation has focused
on task success, we show that a high success is not sufficient for human-like
behavior.
Our key contributions compared to prior work (Sec. 2) are: (1) we introduce an
automated NTT for assessing human-like behavior and investigate six
representative architectures that capture principal design choices of this
task (Sec. 3); (2) we devise methodology for validating our automated NTT
against a human Navigation Turing Test (HNTT) and empirically demonstrate its
effectiveness (Sec. 4); (3) our results show high agreement of ANTT with human
judges when comparing human behavior to state-of-the-art reinforcement
learning agents (Sec. 5). At the same time, we demonstrate that capturing
finer gradations of human-likeness in artificial agents remains an open
challenge. In addition to the potential to drive learning of more human-like
behavior, our work opens the door to a deeper understanding of human decision
making when assessing human-like behavior.
## 2 Related Work
The question of how to systematically assess whether an automated agent
behaves in a human-like manner was compellingly addressed by Alan Turing
(Turing, 1950). Initially called the “imitation game”, and now adapted and
widely known as “Turing Test”, it asks a human judge to identify which of two
respondents is human, and which is machine. Turing’s initial thought
experiment considered a setup where the judge could interact with each
respondent through an appropriate natural language interface. Since its
conception, the idea has been widely adapted to assess the degree of human-
likeness in a range of scenarios, including to assess human-like play in video
games (Yannakakis & Togelius, 2018). Closest to our focus, a Turing Test was
used to assess bot behavior in the video game Unreal Tournament as part of the
2K BotPrize Contest (Hingston, 2009) and human-like play in the Mario AI
Championship (Shaker et al., 2013).
While a Turing Test, if well designed, can provide strong ground truth
assessments of whether humans judge behavior as human-like, several drawbacks
exist. Most important from a machine learning perspective is the need to
collect human judgments for every iteration of an agent, making the approach
extremely costly. This makes it prohibitive for use in typical machine
learning settings. Consequently, previous work has proposed automated proxies
of human-likeness (Kirby et al., 2009; Karpov et al., 2013; de Cothi, 2020).
Domain specific approaches to detecting (non) human-like behavior
automatically often rely on heuristics or rules. In navigation scenarios this
can include methods for detecting mistakes such as “getting stuck” (Kirby et
al., 2009; Karpov et al., 2013). Kirby et al. (2009) focus on socially
acceptable robot navigation with experiments performed in simulated navigation
tasks (e.g., passing people in a hallway). Human preferences were explicitly
encoded as constraints (e.g., a preference for moving to the right to let
someone pass), and learned behaviors were evaluated in simulation, by
assessing how often a given preference was violated. Karpov et al. (2013)
tackle the problem of “believable” bot navigation in the context of the
BotPrize competition (Hingston, 2009). Believable behavior is assessed using
rules for detecting mistakes such as getting stuck in cyclic behavior,
collisions with other entities, failing to move, or falling off a map. A key
drawback of rule-based automation is that they require high manual effort by
domain experts to construct. In addition, due to this high effort, they are
necessarily coarse and unable to capture nuanced aspects of human-likeness.
Addressing both the limitations of scalability of human Turing Tests and rule-
based approaches, learned models can serve as a proxy measure of human-
likeness. If such a model can learn accurate predictions of human-likeness
given ground truth judgments, these approaches could be deployed rapidly and
at scale to drive the development and training of more human-like agents.
Closely aligned with this goal of our work is the recent development of
imitation learning approaches that rely on a discriminator to learn to
distinguish between learned imitation and demonstrated expert behavior (Ho &
Ermon, 2016; Peng et al., 2018). However, while closely aligned with the goals
of our work in principle, the current state of the art in this field has not
yet reached the level required for empirically strong performance on complex
tasks and human-like behavior, with many evaluated on relatively less complex
behaviors and imitating other automated agents. Our work has the potential to
pave the way towards learning complex human-like behavior, and provides
important insights into the key design choices required for capturing notions
of human-likeness. We include representative examples closest to this approach
as baselines (model SYM-FF and VIS-FF, Tbl. 1).
Closely related to our work is the approach taken by de Cothi et al. (2020) to
compare artificial agent behavior in a maze navigation task to human and rat
behavior. They used a support vector machine classifier trained on a small set
of expert designed features to distinguish between model-free, model-based,
and successor-representation-based trained agents. The trained classifier was
then applied to behavior of humans and rats to estimate which agent behavior
was closest to that of biological agents (de Cothi et al., 2020; de Cothi,
2020). We base our approach on this setup, but extend it to take as input
lower-level features (e.g., trajectories, images) to avoid the need for expert
designed feature extraction and better align with the line of work on
imitation learning that forms the second inspiration to our work.
While our approach makes no specific assumption about the type of task for
which we aim to detect human-like behavior, we focus on the task of navigation
in 3D space. This task is well suited as it has been widely studied from a
range of perspectives, including neuroscience (Iaria et al., 2003; Banino et
al., 2018; de Cothi, 2020) and machine learning (Alonso et al., 2020; Chaplot
et al., 2020). In addition, recent years have seen rapid progress on machine
learned agents that achieve high task success, allowing us to focus our
attention on whether the learned successful behavior is indeed human-like. A
driver of progress on learning to navigate has been the availability of
simulated environments and evaluation frameworks. Anderson et al. (2018)
survey recent progress and evaluation frameworks for 3D navigation. In line
with the majority of work in this area, the focus is on evaluating task
success. Recent work on learning to navigate with a machine learning focus
includes improvements of sample efficiency, e.g., through unsupervised
pretraining (Li et al., 2020) and object semantics (Chaplot et al., 2020).
Here we leverage the work of Alonso et al. (2020), as it most closely aligns
with our requirement for achieving high task success in a 3D game environment.
At the intersection of machine learning and neuroscience, machine learned
models have been proposed as models for studying biological behavior (Banino
et al., 2018; de Cothi, 2020). As before, the primary evaluation of these
focused on task success (Banino et al., 2018), with the notable exception of
de Cothi et al. (2020) which forms the starting point of our work.
## 3 Method
In this work we approach the task of predicting human judgments of whether an
observed behavior appears human-like as a classification task, as detailed in
Section 3.1. Our key focus is the question of which type of input space (Sec.
3.2) and model (Sec. 3.3) can best capture human judgments.111See Appendix A.1
for training details and hyperparameters.
### 3.1 Problem setup
We formalize the task of assessing human-likeness as binary classification. We
assume input data in the form of time-indexed trajectories $x_{0:T}$ of length
$T$, where $x_{t}\in\mathbb{R}^{d}$ denotes the $d$-dimensional observation at
time step $t$.
Our goal is to obtain a classifier that achieves high alignment with ground
truth human judgments (detailed in Sec. 4.4). Ground truth takes the form
$y^{*}:x^{(1)}\succ x^{(2)}$ to indicate that a single human judged the
trajectory represented by $x^{(1)}$ to be more human-like than $x^{(2)}$. We
obtain repeat measurements from several human judges for each pair. We focus
on two evaluation measures: (1) accuracy compared to human ground truth
(aggregated by majority vote), and (2) Spearman correlation (a non-parametric
rank correlation measure robust to outliers (Croux & Dehon, 2010)) compared to
a ranking of agent trajectories by the proportion of judges that assessed them
as human-like. Together, these capture coarse and fine-grained agreement
between our classification models and human judgments of human-likeness.
In this work, we assume that human ground truth judgments are only available
at test time. This is a realistic assumption given the resource intensive
nature of this type of data. To address this challenge, we propose training
our models with _proxy labels_ , which are derived from known agent identity
$\hat{y}$. Thus, our classifiers are trained to maximize accuracy when
predicting whether a trajectory was truly generated by a human player or an
artificial agent, using a binary cross entropy loss:
$L=-[\hat{y}_{H=1}\log(p_{H=1})+(1-\hat{y}_{H=1})\log(1-p_{H=1})]$, where we
denote by $\hat{y}_{H=1}$ the pseudo label indicating that trajectory was
truly generated by a human player, and by $p_{H=1}$ the probability that the
current model assigns to the trajectory having been generated by a human
player.
A key question is whether the resulting models can generalize to align well
with the degree to which human judges actually _perceive_ a trajectory as more
or less human. A key design choice that we expect to influence this ability to
capture human judgments is the input space the classifier operates on. We
discuss these choices in the next section.
### 3.2 Choice of input space
Classifier input can be represented in a variety of ways. For example,
presenting input in the form of video frames derived directly from the game
footage, is likely to align well with the information available to human
judges when they decide on whether a given behavior appears human-like. We
hypothesize that observation spaces that are well aligned with the information
available to human judges will result in higher accuracy on the test task of
capturing human judgments of human-like behavior. However, more complex
observations, such as video, may suffer from low amounts of training data,
leading to a potential trade-off between model fidelity and sample efficiency.
(a) Bar-code input.
(b) Sample of a top-down trajectory ($320\times 200$ pixels).
(c) Video frames ($320\times 200$)
Figure 1: Examples of the inputs used by our classifiers: (1) bar-code of a
video (used by BC-CNN), (1) top-down trajectory (TD-CNN), (1) video (frame
sequence, VIS-FF and VIS-GRU).
We consider the following choices (samples in Fig. 1):
SYM. _Symbolic_ observations provide a low-dimensional representation of each
step along the agent trajectory. Here, we use absolute agent location in 3D
space as it provides a minimal representation of agent state.
VIS. _Visual_ observations provide images of size 320x200 of the actual game
footage, rendered without overlays such as game score. This observation type
is expected to be well aligned with the information available to a human judge
when assessing whether a behavior appears human-like.
In our architectures, we explore feed-forward and recurrent models for
encoding observations, as detailed in Section 3.3. In addition, we consider
two compressed representations that abstract the respective observations in a
single image:
TD. _Top-down_ trajectory projection (a single 2D image per trajectory),
obtained by projecting the _symbolic_ representation (agent position) along
the ”up” direction (z-coordinate).
Name | Input (Sec. 3.2) | Encoder | Pre-Training | Notes
---|---|---|---|---
SYM-FF | _Symbolic_ | Feedforward Network | [None] | Equivalent to typical discriminator architectures in adversarial imitation learning (Ho & Ermon, 2016)
SYM-GRU | _Symbolic_ | GRU (Cho et al., 2014) | [None] |
VIS-FF | _Visual_ | VGG-16 (Simonyan & Zisserman, 2014) | Imagenet (CNN) | Equivalent to discriminators for adversarial imitation learning from visual observations (Li et al., 2017)
VIS-GRU | _Visual_ | CNN (VGG-16) + GRU | Imagenet (CNN) |
TD-CNN | _Top-down_ | CNN (VGG-16) | Imagenet (CNN) |
BC-CNN | _Bar-code_ | CNN (VGG-16) | Imagenet (CNN) |
Table 1: Overview of the ANTT model architectures investigated in this work.
See Sections 3.2 and 3.3 for details.
BC. _Bar-code_ encodes _visual_ observations (a single 2D image per trajectory
video). It is obtained by collapsing every video frame into a single line by
averaging along the y-dimension of the frame, and then stacking those lines
into a single image. If the input video had 600 frames of size 320x200, the
resulting bar-code has size 320x600.
We hypothesize that the compressed TD and BC representations capture key
patterns of human-likeness while providing higher information density than
symbolic or visual observations, leading to more effective classification.
### 3.3 Models
Following from our problem definition above (Section 3.1), we now detail our
proposed classification models. Each model tests specific hypotheses about how
to best learn to align with human judgments. A summary of all models is
provided in Table 1, with details as follows:
FF. Feedforward models extract features from individual states, without taking
into account the sequential structure along a trajectory. We construct this
type of model as a representative baseline in line with discriminators typical
of adversarial imitation learning (Ho & Ermon, 2016).
GRU. Recurrent models explicitly model sequence information and are expected
to more directly align with human assessments (human assessors view
observations in the form of videos, c.f., Section 4). We encode sequences
using Gated Recurrent Unit (GRU) networks (Cho et al., 2014).
CNN. Convolutional models are applied to image input (visual, top-down and
bar-code observations). We use a VGG-16 (Simonyan & Zisserman, 2014) pre-
trained on Imagenet (Deng et al., 2009) to extract visual features.
For all model training, input batches of individual observations (feedforward)
and sequences (GRU) are sampled uniformly from trajectories. At test time,
predictions are aggregated per trajectory using the majority vote. Learning
aggregation at the trajectory level is an important direction of research,
however improving simple and robust aggregation baselines with learned
approaches is challenging (Radford et al., 2018; Lan et al., 2020).
Figure 2: Validation accuracy over training epochs for our ANTT models (5-fold
cross validation). Symbolic models achieve relatively lower validation
accuracy, however, our later results show they are less prone to overfitting
(See Section. 5.2).
Before moving forward, we validate that the proposed models are effective on
the task they are trained to perform - distinguishing between behavior
generated by human vs artificial agents. The validation curves in Figure 2
show that all proposed models learn to distinguish between these on our
validation set. This confirms that we have a strong basis for evaluating which
of the proposed approaches matches ground truth provided by human assessors.
Our main results on this question will be presented in Section 5, after we
detail our approach to validating the classifiers in Section 4.
## 4 Validating ANTT by Human NTT
This section details our approach to validating the methods proposed for
automatically evaluating human-like behavior (Sec. 3) with a human NTT (HNTT).
The following components are required for this validation, and are discussed
in turn below. First, we detail the 3D navigation task on which human-like
behavior is assessed (Sec. 4.1). Second, we need to collect behavior data from
human players (Sec. 4.2) and artificial agents (Sec. 4.3). Third, we obtain
ground truth human judgments of human-likeness (Sec. 4.4).222See Appendices
for further details: A.2 (agent training), A.3 (human data collection) and A.4
(evaluation).
### 4.1 Definition of the 3D navigation task
Our experiments are conducted in a modern AAA video game.333We thank
[anonymized] for providing access to this game. A screenshot of the navigation
task is shown in Figure 3, together with a top-down view of the map. The task
is set up on a map that is representative in size and complexity for a modern
3D video game, and comparable to recent navigation experiments, such as the
large game map explored by Alonso et al. (2020). The players’ spawning
position is on an island outside the main map, connected by a set of 3
available jump areas. Agents must learn to navigate to a target location on
the main map, represented by the blue containers in the left screenshot in
Figure 3. On each navigation episode, the target spawns uniformly at random in
one of 16 possible locations. In preliminary tests we confirmed that human
players can successfully navigate to all goal locations within $26$ seconds on
average.
Figure 3: Screenshot of the 3D navigation task as perceived by human players
(left). Mini map of the level (right), indicating as a blue circle the
player’s spawning location (bottom) and in green the 16 possible goal
locations. The main map is approximately 500x320 meters in size and includes
complex 3D elements. The screenshot is not representative of the final game
play and visuals.
### 4.2 Collecting human navigation data
We now describe how we collected human navigation data.
Requirements. Data from a total of 7 human players was collected using the
replay capture engine in an experimental game build implementing the
navigation task (Section 4.1). Human players’ expertise with video games
varied, ranging from casual to advanced players. All players had high
familiarity with the game map and task, and recorded replays on machines that
met the system requirements of the experimental game build, including GPU
rendering support.
Parameters and Constraints. The navigation target was indicated on-screen in a
mini-map that was shown throughout the navigation task (Fig. 3). There was a
limited action space: agents and human players could only navigate (e.g.,
jump, attack and special abilities were disabled). The human players used
keyboard and mouse navigation only.
Video Generation. The episode replays were saved as MP4. A total of 140 human
recordings were collected, which we split into 100 videos (from 4 players) for
classifier training and validation, and 40 (3 remaining players) for testing.
Post-Processing. We applied post-processing edits to the videos for final
inclusion into the HNTT ground truth assessment (Section 4.4). These edits are
defined by a set of rules which were uniformly applied to all the videos. Rule
1: Mask identifying information (computer name tag). Rule 2: Cut all videos to
start the moment the character lands on the main map.444This is to focus our
current validation on ground-based navigation. Assessing human-like flying is
left for future work. Rule 3: Cut the last 1-3 seconds of the human player
videos to just before they stopped moving to correct for an effect of the
human data collection process, where human players had to manually end the
recording and thus appeared briefly “stopped” at the end of each episode.
### 4.3 Collecting agent navigation data
This section details training of the artificial agents used to generate agent
navigation data, including agent architectures, training process, and agent
navigation data collection.
Figure 4 shows our two reinforcement learning agent architectures: symbolic
and hybrid. Both are representative examples of current state-of-the-art
approaches to learning 3D navigation in video games, closely following very
recent work by Alonso et al. (2020). We chose this starting point as a
representative approach that learns effective navigation in a game similar to
ours, and select two observation spaces and architectures for which we expect
high task performance based on the ablation studies of Alonso et al. (2020).
Symbolic. The _symbolic_ agent (Fig. 4) observes a low-dimensional
representation of the goal (relative angle and distance) and game state (agent
position, average frame depth), processed using a feed forward network. We
hypothesize that this compact observation space enables sample-efficient
learning but relatively less human-like behavior.
Hybrid. The _hybrid_ agent (Fig. 4) receives the symbolic observation, as
above, and a 32x32 depth buffer of the game footage. Image features are
extracted using a CNN before concatenating with the output of the feed forward
layers that process symbolic features. Given access to (coarse) visual and
depth information, we hypothesize the agent to be better aligned with the type
of information used by human players and to consequently learn more human-like
behavior.
Both agents use a discrete action space: moving forwards, left and right to a
set of given degrees (30, 45, 90), as well as turning in place. Both agents
receive the same reward signal during training. This consists of a dense
reward for minimizing the distance between the goal location and the agent’s
position, a +1 reward for reaching the target, and a -1 penalty for dying, as
agents are able to jump off the edges of the map. In addition, a small per-
step penalty of $-0.01$ encourages efficient task completion. Episodes end
when agents reach the goal radius or after $3,000$ game ticks ($\approx 50$
seconds, $210$ agent steps).
The agents were trained using PPO (Schulman et al., 2017), implemented in
Tensorflow (Abadi et al., 2015). Behavior policy rollouts during training were
collected through a scalable distributed framework on $60$ parallel game
instances. In each training episode, the agent spawned in one of the 26
predefined locations, both on the main map and in the spawn area, with a
$34\%$ preference for the latter. This uniform domain randomization (Tobin et
al., 2017) was found sufficient for sample efficient learning in preliminary
experiments.
(a) Symbolic agent model architecture
(b) Hybrid agent model architecture Figure 4: Overview of the agent
architectures used in this study.
Before proceeding to the next stage, we verify that the selected agent models
successfully learn the navigation task. Figure 5 verifies that the agent
models used here are representative of the state of the art of learning to
navigate in complex 3D games. As shown, agents learn to reliably reach the
goal within the first 5M training steps. On average, the hybrid agent achieves
task completion 10% faster than the symbolic agent. This demonstrates that our
trained agents achieve the required level of task success, allowing us to
focus on whether the learned behavior is also human-like.
Figure 5: Time to goal over training steps (5 seeds each). Both models achieve
high task success. The high variance observed in the hybrid agent is due to
one seed that had a delay in convergence.
To capture agent navigation data we rolled out the model checkpoints after 20M
training steps. The rolled out policies are non-deterministic. The starting
position for all rollouts was the spawn area (as for human players). The
episodes were saved as videos and then sampled to match the navigation goals
in the selected human navigation samples. The agent navigation was post-
processed in line with the human navigation data (Sec. 4.2).
### 4.4 Collecting ground truth human judgments
We designed two behavioral human NTT (HNTT) studies to establish ground truth
based on collecting human judgments of agent vs. human navigation in the game.
Each study was administered in the form of an online survey.
IRB Approval. The survey was approved by our institution’s Institutional
Review Board (IRB), authorizing us to collect data from internal employees. An
IRB approved digital consent form was obtained before participation.
Participants. A total of 60 human assessors participated in the HNTT. We
recruited participants from internal mailing lists, none of the participants
were familiar with our study. Participants were required to be 18 years or
older. No additional demographic information was collected. Participants were
asked to indicate their familiarity with third-person action games and the
game title used in this study.
Study Design. Each HNTT study consisted of 10 Turing Test trials, in each of
which the stimuli consisted of two side-by-side videos, A and B, of an avatar
navigating a game environment (Figure 6). The task of the participant was to
watch each video and answer the question “Which video is more likely to be
human?”, explain their rationale, and indicate uncertainty of their decision
on a 5-point Likert scale varying from “extremely certain” to “extremely
uncertain”.
In each study all participants viewed the same ten HNTT trials with the same
stimuli, presented in randomized order per participant (within-subject
design). Videos A and B were selected from a set of short videos in which the
symbolic or hybrid agent (Section 4.3), or a human player (Section 4.2)
completed the navigation task (Section 4.1).
Stimuli and conditions. The presented video stimuli and trial conditions
differed between Study 1 and Study 2 according to Table 2, administered as
follows.
vs. | Human | Hybrid | Symbolic
---|---|---|---
Human | - | Study 1 | Study 2
Hybrid agent | Study 1 | - | Study 1,2
Symbolic agent | Study 2 | Study 1,2 | -
Table 2: Table of HNTT comparison conditions in Study 1 and Study 2. Each
study consisted of 6 human vs. agent comparison trials, and 4 trials comparing
the two agents.
Study 1. Human participants were asked to judge a human player vs. a hybrid
agent in 6 trials with side-by-side video pairs, and a symbolic agent vs. a
hybrid agent in 4 trials with side-by-side video pairs (Table 2).
Study 2. Human participants were asked to judge a human player vs. a symbolic
agent in 6 trials with side-by-side video pairs, and a symbolic agent vs. a
hybrid agent in 4 trials with side-by-side video pairs (Table 2).
The video stimuli used in the study were sampled uniformly at random from our
available recordings, matching the video pairs on task goal location. See
Figure 6 for a screenshot.
Figure 6: Example of a Human Navigation Turing Test (HNTT) trial. Videos are
not representative of actual game play and visuals.
## 5 Results
In this section we detail our main empirical results of two Navigation Turing
Test studies, obtained according to the validation approach detailed above
(Sec. 4).
### 5.1 HNTT: Human assessments of human-likeness
We analyzed human behavioral data from two human NTT (HNTT) studies, in each
of which participants observed 10 pairs of video stimuli, judged which is more
human-like, and marked their uncertainty. These results were used to establish
ground truth for automated NTT studies (Sec. 4.4). We first assessed the
extent to which human participants accurately judged videos of real human
players as more human-like than videos of RL agents (Fig. 7). Participants
accurately detected human players above chance in both Study 1 (mean=$0.84$,
std=$0.16$, n=$30$ throughout for both studies) and 2 (mean=$0.77$,
std=$0.16$). On average, participants were slightly more accurate in judging
humans as more human-like in Study 1 compared to Study 2 (Mann-Whitney
$U=341$, $p=0.046$). We then analyzed the reported uncertainty of
participants, reported on a scale of 1 (extremely certain) to 5 (extremely
uncertain), Fig. 8. In Study 1 participants indicated a mean certainty of
$2.1$ (std=$0.47$), corresponding to the level “somewhat certain”. In Study 2,
participants indicated an average certainty of $2.6$ (std=$0.84$), between
“somewhat certain” and “neutral”. The distributions in the two groups differed
significantly (Mann–Whitney $U=257.5$, $p=0.002$), indicating that
participants were less certain of their decisions in Study 2.
Figure 7: HNTT Results. Left: violin plots display proportion of videos with
true human behavior (ground truth) that were judged more human-like. Each
participant’s agreement with ground truth is a dot in the swarm plot. Right:
Reported uncertainty for the same judgment (1: extremely certain to 5:
extremely uncertain).
Next we compared how participants judged symbolic and hybrid agents (Sec.
4.3). In both studies, and in line with our prediction, human participants
judge the hybrid agent to be more human-like when directly compared with
symbolic agents (Fig. 8). At the same time, we observed substantial
variability among human judges. A comparison across studies reveals an
interesting pattern. In hybrid-symbolic comparisons of Study 1, on average
$0.78$ of all participants judged the hybrid agent to be more human-like
(std=$0.25$). This proportion is substantially higher than that of Study 2
(mean=$0.62$, std=$0.30$), with a statistically significant difference (Mann-
Whitney $U=328.0$, p=$0.013$).
Figure 8: Responses in direct comparisons of symbolic and hybrid agents in
Study 1 (left) and 2 (right). Y-axis: number of times a participant judged an
agent more human-like (out of $4$). X-axis: participants, ordered by how often
they chose the hybrid agent.
Classifier | Identity Accuracy | Human-Agent Accuracy | Human-Agent Rank | Hybrid-Symbolic Accuracy | Hybrid-Symbolic Rank
---|---|---|---|---|---
SYM-FF | $0.850$ $(0.062)$ | $0.850$ $(0.062)$ | $0.364$ $(0.043)$ | $0.475$ $(0.166)$ | $-0.244$ $(0.252)$
SYM-GRU | $0.850$ $(0.082)$ | $0.850$ $(0.082)$ | $0.173$ $(0.049)$ | $0.400$ $(0.200)$ | $-0.249$ $(0.210)$
VIS-FF | $0.633$ $(0.041)$ | $0.633$ $(0.041)$ | $-0.041$ $(0.160)$ | $0.225$ $(0.050)$ | $-0.165$ $(0.286)$
VIS-GRU | $0.767$ $(0.097)$ | $0.767$ $(0.097)$ | $0.220$ $(0.267)$ | $0.425$ $(0.127)$ | $-0.056$ $(0.331)$
TD-CNN | $0.583$ $(0.075)$ | $0.583$ $(0.075)$ | $0.222$ $(0.059)$ | $0.525$ $(0.094)$ | $-0.093$ $(0.149)$
BC-CNN | $0.717$ $(0.145)$ | $0.717$ $(0.145)$ | $-0.009$ $(0.131)$ | $0.475$ $(0.050)$ | $-0.095$ $(0.412)$
Table 3: Classifier accuracy and rank compared to human judgments on held-out
test data. All results are the mean (and standard deviation) from 5 repeats of
training the classifier with hyperparameter settings chosen by their average
validation accuracy in 5-fold cross-validation.
One interpretation of this finding is that participants’ judgments are non-
transitive: they judge humans more human-like when directly compared to hybrid
agents (Study 1), and less human-like when compared to the symbolic agent
(Study 2), while at the same time judging the symbolic agent less human-like
in a direct comparison to the hybrid agent. Another interpretation is that
this may be due to the specific videos used in Study 2. Discerning between
these hypotheses is an interesting direction for further investigation.
Overall, our analysis of human judgments show that, in line with expectations
and on average, videos of human navigation are judged more human-like than
those of artificial agents. In addition, as hypothesized in Sec. 4.3, the
hybrid agent is judged more human-like when directly compared with the
symbolic agent (Figure 8). These results validate our HNTT methodology,
enabling us to validate our classifiers (artificial judges) against the human
judgment ground truth reported in this section. Future behavioral studies can
explore individual differences in human judgments of both human and artificial
navigation.
### 5.2 Evaluating the ANTT
We now turn to our results of evaluating our proposed ANTT. We aim to identify
classification models (Sec. 3.3) that accurately capture human judgments of
human-likeness. All models were trained to distinguish human from agent
trajectories. Training and hyperparameter tuning was performed using 5-fold
cross validation on trajectories generated by agent checkpoints and human
players that were fully separate from those that generated test data. We
collected human judgments (Sec. 5.1) for test trajectories only.
Table 3 shows our results for all considered classifier models (see Table 1
for model details). First, we verify that all models generalize to the held
out training data. We assess this by comparing classifier predictions
(aggregated on the trajectory level by majority vote) against the true
identity of the agent (i.e., human or agent) termed identity accuracy. We find
that all models generalize to the test data, with results for the symbolic
models particular robust to overfitting. Results range from an accuracy of
$0.583$ for TD-CNN to $0.850$ for both symbolic models. These results also
validate our trajectory level aggregation by majority voting approach,
demonstrating that is achieves high accuracy when comparing to trajectory
level human judgments.
Our main results compare classifier models to human judgments, split into
human-agent accuracy and hybrid-symbolic accuracy (Table 3). As the majority
vote aggregation of human responses was equal to the true identity for all
questions that compared human videos with agent videos, we again observe the
same high accuracy on human-agent judgments. However, accuracy is low for
hybrid-symbolic pairs across all classifiers.
Our results are further confirmed when assessing human-agent and hybrid-
symbolic rank correlation between human judgments and model predictions. We
compute these using Spearman’s rank correlation (Croux & Dehon, 2010), as
follows. We aggregate human ground truth judgments for human-agent
(separately: symbolic-hybrid) pairs based on the percentage of human judges
that labeled a human (or hybrid agent in hybrid-symbolic) trajectory as more
human-like. This was compared to a ranking of model output logits. We observe
medium rank agreement and low variance for the symbolic and topdown CNN models
on human-agent rankings. However, rankings of all models on hybrid-symbolic
agent rankings are low (negative values indicate anti-correlated rankings) and
have high variance.
Our classification results demonstrate that classifier models do generalize
well to unseen test trajectories. Agreement with human judgments is high when
assessing the human-likeness of agents compared to human players. Highest
agreement is observed for symbolic models. This relative performance of
classifiers is unexpected, given that the symbolic classifiers observe a
representation that is very different from that used by human judges for
decision making. Based on our results, it is likely that the visual models
overfit given the richer representations and higher model capacity. In
addition, we find that current classifiers perform poorly when assessing the
degree of human-likeness of pairs of agents. Our results motivate important
follow up work to close this gap between matching human judgments on human-
agent vs agent-agent pairs.
## 6 Conclusion
We introduce the automated Navigation Turing Test (ANTT), a new approach to
automatically evaluating the human-likeness of navigation behavior. Our
results, comparing to a human NTT, show that a high level of agreement between
automated and human judgments can be achieved, opening the door to using
classification approaches as scalable proxies for driving the development of
more human-like agents. Best performance is achieved by classification models
that operate on compact representations of agent trajectories, in particular
_bar-code_ as well as symbolic representations. At the same time, we show that
detecting finer-grained human judgments that capture agents’ improvements
towards human-likeness remains an open challenge for future work. Our work
takes an important step on the dual goals of understanding human assessment of
human-like behavior and towards agents capable of learning complex human-like
behaviors.
## Acknowledgements
We thank Batu Aytemiz, Dave Bignell, Mikhail Jacob, Mingfei Sun, Robert
Loftin, Sebastian Nowozin, and Stephanie Reikine for their feedback and
valuable discussions. We are grateful to Ninja Theory and Microsoft for their
support. We thank the anonymous participants of our human NTT studies for
their time and effort.
## References
* Abadi et al. (2015) Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G. S., Davis, A., Dean, J., Devin, M., Ghemawat, S., Goodfellow, I., Harp, A., Irving, G., Isard, M., Jia, Y., Jozefowicz, R., Kaiser, L., Kudlur, M., Levenberg, J., Mané, D., Monga, R., Moore, S., Murray, D., Olah, C., Schuster, M., Shlens, J., Steiner, B., Sutskever, I., Talwar, K., Tucker, P., Vanhoucke, V., Vasudevan, V., Viégas, F., Vinyals, O., Warden, P., Wattenberg, M., Wicke, M., Yu, Y., and Zheng, X. TensorFlow: Large-scale machine learning on heterogeneous systems, 2015\.
* Alonso et al. (2020) Alonso, E., Peter, M., Goumard, D., and Romoff, J. Deep reinforcement learning for navigation in aaa video games. _arXiv preprint:2011.04764_ , 2020.
* Anderson et al. (2018) Anderson, P., Chang, A., Chaplot, D. S., Dosovitskiy, A., Gupta, S., Koltun, V., Kosecka, J., Malik, J., Mottaghi, R., Savva, M., et al. On evaluation of embodied navigation agents. _arXiv preprint:1807.06757_ , 2018.
* Banino et al. (2018) Banino, A., Barry, C., Uria, B., Blundell, C., Lillicrap, T., Mirowski, P., Pritzel, A., Chadwick, M. J., Degris, T., Modayil, J., et al. Vector-based navigation using grid-like representations in artificial agents. _Nature_ , 557(7705):429–433, 2018.
* Chaplot et al. (2020) Chaplot, D. S., Gandhi, D. P., Gupta, A., and Salakhutdinov, R. R. Object goal navigation using goal-oriented semantic exploration. _Advances in Neural Information Processing Systems_ , 33, 2020.
* Cho et al. (2014) Cho, K., Van Merriënboer, B., Bahdanau, D., and Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. _arXiv preprint:1409.1259_ , 2014.
* Cowan et al. (2017) Cowan, B. R., Pantidi, N., Coyle, D., Morrissey, K., Clarke, P., Al-Shehri, S., Earley, D., and Bandeira, N. ” what can i help you with?” infrequent users’ experiences of intelligent personal assistants. In _Proceedings of the 19th International Conference on Human-Computer Interaction with Mobile Devices and Services_ , pp. 1–12, 2017\.
* Croux & Dehon (2010) Croux, C. and Dehon, C. Influence functions of the spearman and kendall correlation measures. _Statistical methods & applications_, 19(4):497–515, 2010.
* de Cothi et al. (2020) de Cothi, W., Nyberg, N., Griesbauer, E.-M., Ghanamé, C., Zisch, F., Lefort, J., Fletcher, L., Newton, C., Renaudineau, S., Bendor, D., Grieves, R., Duvelle, É., Barry, C., and Spiers, H. J. Predictive maps in rats and humans for spatial navigation. _bioRxiv preprint: 2020.09.26.314815_ , 2020.
* de Cothi (2020) de Cothi, W. J. _Predictive maps in rats and humans for spatial navigation_. PhD thesis, UCL (University College London), 2020.
* Deng et al. (2009) Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In _2009 IEEE conference on computer vision and pattern recognition_ , pp. 248–255. Ieee, 2009.
* Hingston (2009) Hingston, P. A turing test for computer game bots. _IEEE Transactions on Computational Intelligence and AI in Games_ , 1(3):169–186, 2009.
* Ho & Ermon (2016) Ho, J. and Ermon, S. Generative adversarial imitation learning. In _Proceedings of the 30th International Conference on Neural Information Processing Systems_ , pp. 4572–4580, 2016.
* Iaria et al. (2003) Iaria, G., Petrides, M., Dagher, A., Pike, B., and Bohbot, V. D. Cognitive strategies dependent on the hippocampus and caudate nucleus in human navigation: variability and change with practice. _Journal of Neuroscience_ , 23(13):5945–5952, 2003.
* Karpov et al. (2013) Karpov, I. V., Schrum, J., and Miikkulainen, R. Believable bot navigation via playback of human traces. In _Believable Bots_ , pp. 151–170. Springer, 2013.
* Kirby et al. (2009) Kirby, R., Simmons, R., and Forlizzi, J. Companion: A constraint-optimizing method for person-acceptable navigation. In _RO-MAN 2009-The 18th IEEE International Symposium on Robot and Human Interactive Communication_ , pp. 607–612. IEEE, 2009.
* Lan et al. (2020) Lan, O., Huang, X., Lin, B. Y., Jiang, H., Liu, L., and Ren, X. Learning to contextually aggregate multi-source supervision for sequence labeling. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pp. 2134–2146, 2020.
* Li et al. (2020) Li, J., Wang, X., Tang, S., Shi, H., Wu, F., Zhuang, Y., and Wang, W. Y. Unsupervised reinforcement learning of transferable meta-skills for embodied navigation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 12123–12132, 2020.
* Li et al. (2017) Li, Y., Song, J., and Ermon, S. Inferring the latent structure of human decision-making from raw visual inputs. _arXiv preprint:1703.08840_ , 2, 2017.
* Peng et al. (2018) Peng, X. B., Kanazawa, A., Toyer, S., Abbeel, P., and Levine, S. Variational discriminator bottleneck: Improving imitation learning, inverse rl, and gans by constraining information flow. In _International Conference on Learning Representations_ , 2018.
* Radford et al. (2018) Radford, B. J., Richardson, B. D., and Davis, S. E. Sequence aggregation rules for anomaly detection in computer network traffic. _arXiv preprint:1805.03735_ , 2018.
* Scheutz et al. (2007) Scheutz, M., Schermerhorn, P., Kramer, J., and Anderson, D. First steps toward natural human-like hri. _Autonomous Robots_ , 22(4):411–423, 2007.
* Schulman et al. (2017) Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and Klimov, O. Proximal policy optimization algorithms. _arXiv preprint:1707.06347_ , 2017.
* Shaker et al. (2013) Shaker, N., Togelius, J., Yannakakis, G. N., Poovanna, L., Ethiraj, V. S., Johansson, S. J., Reynolds, R. G., Heether, L. K., Schumann, T., and Gallagher, M. The turing test track of the 2012 mario ai championship: entries and evaluation. In _2013 IEEE Conference on Computational Inteligence in Games (CIG)_ , pp. 1–8. IEEE, 2013.
* Simonyan & Zisserman (2014) Simonyan, K. and Zisserman, A. Very deep convolutional networks for large-scale image recognition. In _International Conference on Learning Representations (ICLR)_ , 2014.
* Soni & Hingston (2008) Soni, B. and Hingston, P. Bots trained to play like a human are more fun. In _2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence)_ , pp. 363–369. IEEE, 2008\.
* Tobin et al. (2017) Tobin, J., Fong, R., Ray, A., Schneider, J., Zaremba, W., and Abbeel, P. Domain randomization for transferring deep neural networks from simulation to the real world. In _2017 IEEE/RSJ international conference on intelligent robots and systems (IROS)_ , pp. 23–30. IEEE, 2017.
* Turing (1950) Turing, A. Computing machinery and intelligence. _Mind_ , 59(236):433–460, 1950.
* Vinyals et al. (2019) Vinyals, O., Babuschkin, I., Czarnecki, W. M., Mathieu, M., Dudzik, A., Chung, J., Choi, D. H., Powell, R., Ewalds, T., Georgiev, P., et al. Grandmaster level in starcraft ii using multi-agent reinforcement learning. _Nature_ , 575(7782):350–354, 2019.
* Winston (1984) Winston, P. H. _Artificial intelligence, an MIT perspective_. MIT Press, 1984.
* Yannakakis & Togelius (2018) Yannakakis, G. N. and Togelius, J. _Artificial intelligence and games_ , volume 2. Springer, 2018.
|
# cGraph: Graph Based Extensible Predictive Domain Threat Intelligence
Platform
Wathsara Daluwatta†, Ravindu De Silva†, Sanduni Kariyawasam†, Mohamed Nabeel‡,
Charith Elvitigala†, Kasun De Zoysa†, Chamath Keppitiyagama† †University of
Colombo School of Computing‡Qatar Computing Research InstituteSri Lanka, Qatar
(2022)
###### Abstract.
Ability to effectively investigate indicators of compromise and associated
network resources involved in cyber attacks is paramount not only to identify
affected network resources but also to detect related malicious resources.
Today, most of the cyber threat intelligence platforms are reactive in that
they can identify attack resources only after the attack is carried out.
Further, these systems have limited functionality to investigate associated
network resources. In this work, we propose an extensible predictive cyber
threat intelligence platform called cGraph that addresses the above
limitations. cGraph is built as a graph-first system where investigators can
explore network resources utilizing a graph based API. Further, cGraph
provides real-time predictive capabilities based on state-of-the-art inference
algorithms to predict malicious domains from network graphs with a few known
malicious and benign seeds. To the best of our knowledge, cGraph is the only
threat intelligence platform to do so. cGraph is extensible in that additional
network resources can be added to the system transparently.
malicious domains, belief propagation, graph inference, kubernetes, apache
spark
††copyright: acmcopyright††journalyear: 2022††conference: Online, hosted by
Lyon, France; April 25-29, 2022; The Web Conference 2022
## 1\. Introduction
Internet domains are the launch pad for many cyber attacks we observe
nowadays. One effective way to reduce the damage caused by such attacks is to
identify the domains involved early and take actions such as blocking, taking
down or sinkholing. Further, these malicious domains have underlying
associations with other Internet resources that help to uncover malicious
infrastructures. We observe that existing threat intelligence systems ,such as
Cisco Umbrella Investigate and Anomali, fail to take advantage of such
associations to detect stealthy malicious domains nor do they capture such
associations in a user-friendly graphical form that allows users to easily
explore related Internet resources. All these solutions are reactive in nature
and by the time security operation teams observe them the damage is already
done. A predictive solution would immensely assist in mitigating such therats.
Motivated by these gaps in the industry, we build a proactive threat
intelligence platform that can uncover malicious domains early and visualize
associated Internet resources in a graph format for easy analysis.
Building an investigation platform in this domain involves a unique set of
challenges. First, the amount of data one needs to process on a daily basis is
huge. On average, we observe around 850 million network records per day.
Second, the network data is highly dynamic. For example, domains change their
hosting IPs periodically, the ownership of IPs change over time, new domains
are created and existing domains cease to exist. Third, intelligence sources
such as Phishtank, VirusTotal and GSB, contain limited information about the
status of domain and they are slow to detect malicious domains (Nabeel et al.,
2020). This is due to the fact that most of such systems rely on either
website content or user accessing of these websites to ascertain whether they
are malicious or not. Providing timely and up-to-date intelligence is thus a
challenge on its own. Fourth, network data, though mostly structured, is
available in different data formats and these formats over time. Fifth,
collecting additional network data and intelligence seeds is often rate
limited and encounter frequent network failures. Sixth, storage and retrieval
of huge amounts of graph data, especially in community databases, are not
demonstrated to work reliably in practice.
In this work, we build a system called cGraph to address the above mentioned
limitations. We design to implement a scalable big data processing pipeline
that can ingest billions of disparate reports daily. In order to track the
changes over time, we build succinct snapshots of the whole Internet on a
daily basis. cGraph provides predictive intelligence by incorporating state of
the art graph inference based malicious domain prediction techniques (Nabeel
et al., 2020). In order to make the system oblivious to data format changes,
cGraph provides one level of indirection to ingest data as common schema into
the system. This allows one to incorporate data format changes without
requiring to change the data processing pipeline allowing to extend with
additional data sources transparently. Further, unlike prior work (Nabeel et
al., 2020) which utilizes static graphs to infer malicious domains, cGraph is
able to construct a graph around the domain in question on the fly and
generate real-time intelligence. This gives cGraph to predict the
maliciousness of any domain provided there are seed domains in the
neighborhood.
Demonstration Scenarios: cGraph assists in domain threat investigation as well
as identification of associated malicious domains 111Demonstrations of a
combosquatting domain and a benign domain investigation are available at
https://youtu.be/yu47CZkp_eY and https://youtu.be/waxczBk7vA8 respectively..
We show three use cases: (1) searching for impersonating domains and
discovering related malicious domains, (2) real-time reputation score for a
domain via a browser based plugin, and (3) investigation of the evaluation of
network resources over time around a domain under consideration.
## 2\. Heterogeneous Network Graph
### 2.1. Data Sources
The system ingests four types of data/intelligence sources to build the graph:
(1) Farsight Security PDNS (Farsight Security, Inc., 2022) \- Passive DNS data
observed from all around the world which covers the majority of the DNS
resolutions in the Internet., (2) VirusTotal (VT) (VirusTotal, Subsidiary of
Google, 2022) \- Reputation service that provides aggregated intelligence on
any URL, which is the most popular aggregated threat intelligence platform
that provides threat reports using more than 80 antivirus engines, blacklists,
and sandbox systems., (3) Maxmind (MaxMind, 2022) \- A comprehensive database
consisting of geolocations and ASN information of IP addresses (4) Alexa top 1
million domains (ale, 2022) \- Alexa is a widely used resource for search
engine optimization and capturing the status of domains.
In summary, we use Farsight passive DNS records to generate comprehensive
knowledge graphs and use Alexa Top 1 million domains, VirusTotal, and Maxmind
information to inject the graph with benign and malicious seed domains for
running belief propagation and predicting new malicious domains (Nabeel et
al., 2020).
### 2.2. Graph Modeling
To create the knowledge graph (KG), we employ the DNS record types A, NS, MX,
SOA, and CNAME, which include the most important information on domains.
Specifically, Type A records contain domain to IPv4 address resolutions, Type
NS records contain domain to nameserver resolutions, Type MX records contain
domain to mail server resolutions, and CNAME records provide domain to alias
domains resolutions. All records are created by having a domain name as their
pivot. We create a heterogeneous KG utilizing the above DNS records with
different edge types representing different record types and nodes
representing domains or IP addresses. Our intuition of building such a KG is
that attackers are likely to host their domains in the same infrastructure
where some of their blacklisted domains are hosted. This allows one to utilize
label propagation algorithms to predict malicious domains from a small seed.
We build a heterogeneous KG based on the following different types of DNS
records available in PDNS.
* •
Vertices: Apex, FQDNs, IPs, Name Servers, Mail Servers, CNAME, SOA
* •
Edges: Apex-IP, Apex-FQDNs, FQDNs-IP, Apex-NS, FQDNs-NS, Apex-MX, FQDNs-MX,
Apex-CNAME, FQDNs-CNAME, Apex-SOA, FQDNs-SOA
It is important to store the graph optimized for lookup as cGraph should be
able to retrieve the KG for inference as well as replaying of network
infrastructure over time quickly. Thus, the graph information is stored along
with timestamps at the granularity of a day.
## 3\. System Architecture
### 3.1. Overall System
The system consists of four main layers: data processing, database and caching
layer, middleware and API layer, and consumer applications. The data
processing layer is the component where all the data is cleaned and processed
at the first phase. The infrastructure layer orchestrates the resources to
insert/update the cleaned and reshaped data into the database cluster. Caching
layer handles the data caching and the middleware layer handles the real-time
inference process and intermediate data processing tasks. API layer connects
the backend to the consumer products, which are web application, browser
extension, and public developer API 222More information is available at
https://qcri.github.io/cgraph/..
Figure 1. Overall system from infrastructure point of view
All resources have been provisioned on top of VMware vSphere private cloud.
Kubernetes cluster has been installed on top of the provisioned VMs and an
abstracted computation resources layer has been created in order to manage all
computation resources together.
### 3.2. Graph Inference
Inference. Belief propagation (BP) is an iterative algorithm that passes the
belief iteratively. In each iteration t, belief propagation algorithm updates
the belief value on the node and passes the beliefs (messages) to its
neighbors based on the belief values it received from the previous iteration
t-1. As mentioned, this interactive process continues until each message value
passed by a node to a neighbor node converges to a stable value. The final
beliefs are extracted after convergence.
Execution. The main idea is to construct a KG on the fly and inject ground
truth labels to the KG from credible sources where VT, Alexa, and Maxmind and
then run the BP algorithm on top of the knowledge graph by using the labeled
values as initial beliefs. It propagates a probabilistic values for each
unlabeled domain indicating the maliciousness. For bravity, we explain the
execution for a selected example graph. The ground truth extraction process is
conducted in two stages; benign ground truth extraction and malicious ground
truth extraction.
For the benign ground truth extraction, Alexa Rank data is used. Alexa Rank
data feed for the latest seven consecutive days is taken and the domains with
Alexa Rank $\leq 10000$ is taken for each day. From this, the unique set of
domains is taken as the benign ground truth. Similar to prior research (De
Silva et al., 2021), we observe that domains with Alexa Rank $\leq 10000$
which appear consistently throughout a week are likely to be benign as
malicious domains do not live longer than a day or two in Alexa 10K.
For the malicious ground truth, the data from the daily feed of VirusTotal is
used. VirusTotal reports for latest seven consecutive days are taken and and
the reports with a VT positive count of at least 3 is extracted as malicious
ground truth as VT reports with 2 or less positive counts are known to be a
less reliable indicator of maliciousness (Nabeel et al., 2020). We manually
verify random samples of benign and malicious ground truth and ascertain that
they are indeed correctly labeled.
After seeding the KG with the ground truth domains, we run a multi-threaded
version of the BP algorithm which executes within a few seconds. Our
experiments show that our inference based classification achieves a high
precision of 92.17% and high recall of 99.41%. We observe that a domain’s
maliciousness score does not affect significantly by considering only its two
hop neighbors compared to the complete graph. Hence, for the real-time
inference on an unknown domain, we build the two-hop neighborhood KG and run
the inference algorithm to further improve the computational performance.
## 4\. Current Threat intelligent Solutions vs cGraph
We thoroughly evaluate the key threat intelligence solutions in the market
against cGraph, our threat intelligence platform: AT&T Cybersecurity,
VirusTotal Graph, Cisco Umbrella Investigate, and Anomali Threat Intelligence
Platform. Below, we provide a summary of key differentiation in our system:
(1) Dynamic graph-based visualization and search of Internet artifacts for
threat intelligence (all the above solutions lack an intuitive graph based
navigation), (2) The ability to integrate various data sources and expand the
network coverage and the ability to integrate any intelligence sources
seamlessly, (3) Predictive domain threat intelligence based on an efficient
graph inference algorithm (all above solutions provide only reactive
intelligence), and (4) The ability to traverse the network infrastructure
based on time and observe the changes in the physical layout as well as the
maliciousness of the domains in the network over time, which greatly assist
security operation teams in identifying root causes of attacks and the extent
of the damage.
## 5\. Demonstration
In this section, we show three common use cases of our system.
Searching for impersonating domains and related other malicious domains:
Attackers increasingly impersonate popular brands such as Apple, Paypal and
Microsoft. cGraph allows to search for any popular brands and easily identify
likely combosquatting domains.We demonstrate the use case of hunting for
combosuqatting domains of paypal.com.
Figure 2. Search results for the keyword paypal
Figure 2 shows the search results for the keyword paypal ranked based on VT
positive counts. An investigator may pick any of these domains to further
investigate. Assume that they select paypal-assist.com, which takes them to
the graph based interface shown in Figure 3. This view includes the domain
summary information, network graph, Alexa ranking and VT statistics over time,
IP geolocation information and historical hosting information.
Figure 3. Dashboard for the domain paypal-assit.com
The ability expand the network graph based on the investigator’s findings is
quite useful in practice. cGraph not only allows to expand any network node
when it has children. but also provides real-time predict intelligence on the
unknown domains in the network. Figure 4 shows the expansion of the IP address
47.254.170.24 that adds 4 additional domains to the graph which are hosted on
this IP address.
Figure 4. Expansion of the IP address 47.254.170.24
The inference graph view shows the maliciousness score of each domain in the
network graph calculated on the fly using the BP algorithm based on known seed
domains, paypal-assist.com in this case. Figure 5 shows that cGraph inference
identifies the domains google-standard.com, google-money.com, paypal-debit.com
and google-sale.com to be highly malicious.
Figure 5. Malicious domains identified by inference
Evaluation of maliciousness of any domain via the browser plugin: In this use
case, we show how cGraph API can be consumed to obtain reputation scores for
websites on the fly. While blacklists such as GSB that powers the Chrome
browser identify many malicious websites, they are either slow to react or
unable to make a judgement if the website content is cloaked from the
blacklisting service. cGraph’s reputation engine takes a different approach
and infers the maliciousness of domains based on the maliciousness of the
neighboring network resources such as sibling domains, name servers, CNAME
domains, IP addresses and certificates. As mentioned earlier, the reputation
score is computed dynamically based on the status of the network
infrastructure at the present moment.
Figure 6. Browser plugin: real time reputation score
Figure 6 shows the cGraph inference API in action when a user types the
website soaponline.org. The maliciousness score indicates how likely the
domain is malicious or benign. In this example, the maliciousness score of
0.87 indicates that soaponline.org is highly likely to be malicious.
Investigating the network over time: Network resources associated with domains
change over time due to both malicious as well as benign reasons. For example,
cloud IP rotations, CDN hostings and IP load balancing change IPs associated
with domains frequently. Attackers also move from one IP to another as a
technique to either resist take down or evade detection. cGraph timeline
feature allows to investigate the network graph over a time period. Coupled
with the real-time inference feature, one can observe how the reputation of
any of the domains change over time. Figure 7 shows the network graph on day 1
and the evolved network graph on day 7 respectively. This allows the
investigator to learn the fluxing behavior of domains and take necessary
actions.
Figure 7. Network timeline for day 1 (left) and day 7 (right)
## 6\. Conclusions
We propose, cGraph, which takes domain threat investigation a step forward.
cGraph can efficiently store and process large amount of disparate network
data, build network graph for any given time period and perform real-time
graph inference to predict domains as malicious or benign, compared to
reactive threat intelligence platforms available. We demonstrate three
frequent use cases of cGraph: domain search, real-time reputation computation
and network graph evolution along with the maliciousness scores of domains in
the network over time.
## References
* (1)
* ale (2022) 2022\. Alexa: The Web Information Company. https://www.alexa.com/topsites. Accessed January 2022.
* ano (2022) 2022\. Anomali Threat Intelligence Platform. https://www.anomali.com/. [Online].
* cis (2022) 2022\. Cisco Umbrella Investigate. https://umbrella.cisco.com/products/umbrella-investigate. [Online].
* De Silva et al. (2021) R. De Silva, M. Nabeel, C. Elvitigala, I. Khalil, T. Yu, and C. Keppitiyagama. 2021\. Compromised or Attacker-Owned: A Large Scale Classification and Study of Hosting Domains of Malicious URLs. Usenix Security.
* Farsight Security, Inc. (2022) Farsight Security, Inc. 2022\. DNS Database. https://www.dnsdb.info/.
* MaxMind (2022) MaxMind. 2022\. GeoLite2 Databases.
* Nabeel et al. (2020) M. Nabeel, I. M. Khalil, B. Guan, and T. Yu. 2020\. Following Passive DNS Traces to Detect Stealthy Malicious Domains Via Graph Inference. _ACM Trans. Priv. Secur._ 23, 4, Article 17 (July 2020), 36 pages. https://doi.org/10.1145/3401897
* VirusTotal, Subsidiary of Google (2022) VirusTotal, Subsidiary of Google. 2022\. VirusTotal. https://www.virustotal.com/.
|
# Black-Box Detection of Language Model Watermarks
Thibaud Gloaguen
Department of Mathematics
ETH Zurich
<EMAIL_ADDRESS>&Nikola Jovanović
Department of Computer Science
ETH Zurich
<EMAIL_ADDRESS>&Robin Staab
Department of Computer Science
ETH Zurich
&Martin Vechev
Department of Computer Science
ETH Zurich
###### Abstract
Watermarking has emerged as a promising way to detect LLM-generated text. To
apply a watermark an LLM provider, given a secret key, augments generations
with a signal that is later detectable by any party with the same key. Recent
work has proposed three main families of watermarking schemes, two of which
focus on the property of preserving the LLM distribution. This is motivated by
it being a tractable proxy for maintaining LLM capabilities, but also by the
idea that concealing a watermark deployment makes it harder for malicious
actors to hide misuse by avoiding a certain LLM or attacking its watermark.
Yet, despite much discourse around detectability, no prior work has
investigated if any of these scheme families are detectable in a realistic
black-box setting. We tackle this for the first time, developing rigorous
statistical tests to detect the presence of all three most popular
watermarking scheme families using only a limited number of black-box queries.
We experimentally confirm the effectiveness of our methods on a range of
schemes and a diverse set of open-source models. Our findings indicate that
current watermarking schemes are more detectable than previously believed, and
that obscuring the fact that a watermark was deployed may not be a viable way
for providers to protect against adversaries. We further apply our methods to
test for watermark presence behind the most popular public APIs: GPT4, Claude
3, Gemini 1.0 Pro, finding no strong evidence of a watermark at this point in
time.
## 1 Introduction
With the rapid increase in large language model (LLM) capabilities and their
widespread adoption, researchers and regulators alike are raising concerns
about their potential misuse for generating harmful content [1, 3]. To tackle
this issue, the idea of watermarking, a process of embedding a signal
invisible to humans into generated texts, is gaining significant traction.
#### Language model watermarking
More formally, in LLM watermarking [8, 9, 10, 6, 22, 21, 23, 19] we consider
the setting of a _model provider_ that offers black-box access to their
proprietary model $LM$ while ensuring that each generation $y$ in response to
a prompt $q$ can be reliably attributed to the model. To enable this, the
provider modifies the generations using a secret watermark key $\xi$, which a
corresponding watermark detector can later use to detect whether a given text
was generated by $LM$.
The prominent family of _Red-Green_ watermarks [8, 9] achieves this by, at
each step of generation, selecting a context-dependent pseudorandom set of
logits to be boosted, modifying the model’s next-token distribution. In
contrast to these, the recently proposed families of _Fixed-Sampling_ [10] and
_Cache-Augmented_ [22, 6] schemes have focused on developing watermarks that
aim to preserve the output distribution of the LLM, achieving such guarantees
in ideal theoretical settings. This goal can be motivated by two main
benefits: (i) it is a tractable proxy for preservation of capabilities of the
original LLM, and (ii) intuitively, the deployment of the watermark should
become inherently harder to detect for malicious actors that would otherwise
be able to avoid or attack the watermark [7, 15, 19].
#### Practical detectability
Yet, despite watermark detection being a key concern in current discourse,
there have so far been no rigorous investigations into the _practical
detectability_ of any of the three watermark scheme families. It thus remains
unclear if the distribution preservation guarantees of Fixed-Sampling and
Cache-Augmented schemes, that hold in ideal settings, offer any detectability
benefits in real-world deployments. Recent work has even argued that Red-Green
schemes, whose detectability has also never been thoroughly studied in
realistic settings, avoid some undesirable tradeoffs of distribution-
preserving watermarks, and are ultimately more practical [16].
#### This work: practical black-box watermark detection
In this work, for the first time, we comprehensively investigate the question
of watermark detection in practical real-world settings. Faithful to how LLMs
are exposed in current public deployments, we assume a minimal setting in
which the adversary cannot control any input parameters (e.g., temperature,
sampling strategy) apart from the prompt $q$, and does not receive any
additional information (e.g., log-probabilities) apart from the instruction-
tuned model’s textual response $y$. In this setting, the goal of the adversary
is to identify if the model is watermarked and determine which scheme was
used.
To this end, we propose rigorous and principled statistical tests for the
black-box detection of seven scheme variants of the three most prominent
watermark families. Our tests, illustrated in Fig. 1, are based on fundamental
properties of the three respective scheme families: Red-Green (Section 2),
Fixed-Sampling (Section 3), and Cache-Augmented (Section 4) watermarks. In our
extensive experimental evaluation, we find that in practice, both
distribution-modifying as well as distribution-preserving schemes are _easily
detectable_ by our tests, even in the most restricted black-box setting.
Highlighting their practicality, we further test several real-world LLM
deployments (GPT4, Claude 3, Gemini 1.0 Pro) and detect no watermarks,
suggesting that watermark deployment at this scale is an ongoing challenge.
#### Implications
While the impact of watermarking schemes on model capabilities (point (i)
above) remains an interesting avenue for future work, our findings question
whether non-detectability (point (ii) above) should be a key focus in the
development of LLM watermarks. Guided by our results, we suspect that avoiding
practical detectability may be an inherently hard problem given the
fundamental nature of watermarks. In light of this, we advocate for a more
diverse and practically oriented development and evaluation of LLM watermarks,
allowing for broader consideration of other relevant properties such as attack
robustness, text quality, and efficiency.
#### Main contributions
We make the following key contributions:
* •
We present the first principled and statistically rigorous tests for practical
black-box detection of popular LLM watermarks across the three prominent
scheme families: Red-Green (Section 2), Fixed-Sampling (Section 3), and Cache-
Augmented (Section 4).
* •
We confirm the effectiveness of all our tests in an extensive experimental
evaluation across seven schemes and five open-source models, and present
additional results that verify the applicability of our tests in real-world
settings (Sections 5.1 and 5.2).
* •
We further highlight the practicality of our tests by executing them on
several real-world black-box LLM deployments—GPT, Claude, and Gemini (Section
5.3).
## 2 Detecting Red-Green Watermarks
Figure 1: Overview of our key contribution. Given black-box textual access to
a language model, a client can query the model and run statistical tests to
rigorously test for presence of a watermark. In this example, both the test
for _Cache-Augmented_ watermarks (Section 4) and the test for _Fixed-Sampling_
watermarks (Section 3) fail, while the test for _Red-Green_ watermarks
(Section 2) successfully detects the watermark.
In this section, we introduce our statistical test for detecting watermarks
from the _Red-Green_ family, illustrated in Fig. 1 (bottom). To this end, we
exploit their core property: applying the watermark introduces a _context-
dependent logit bias_ , i.e., the model’s output distribution is biased in a
way that can greatly vary depending on the last few tokens that were generated
(_context_), yet is unaffected by the tokens preceding the context (_prefix_).
We first recall the background related to these schemes. We then introduce our
modeling assumptions, describe the querying strategy we use to obtain the data
for the test, and finally describe the test itself. In Section 5, we
experimentally confirm the effectiveness of the test in realistic settings and
study its query cost, and in App. C explore methods to estimate scheme
parameters once a watermark is confirmed.
#### Background
Assume a watermark key $\xi\in\mathbb{N}$, a pseudorandom function (PRF) $f$,
and a hashing function $H\colon\mathbb{N}\to\mathbb{N}$. At each generation
step $t$, a Red-Green watermark modifies the logits $l_{1:|V|}$ of tokens from
the vocabulary $V$ to promote a subset of tokens (_green tokens_) before
applying standard sampling. We consider two popular variants, LeftHash [8] and
SelfHash [9], both parametrized by $\delta,\gamma\in\mathbb{R}^{+}$, and using
$h=1$ and $h=3$ previous tokens as context, respectively. LeftHash seeds $f$
with $H(y_{t-1})\cdot\xi$, and uses it to split $V$ into $\gamma|V|$ green
tokens and remaining _red_ tokens. For each green $i$, it then increases
$l_{i}$ by $\delta$. SelfHash differs by seeding $f$ with
$\min(H(y_{t-3}),\ldots,H(y_{t}))\cdot H(y_{t})\cdot\xi$, effectively using a
context size of $4$ by including the token $y_{t}$ yet to be generated. For
both schemes, the watermark detector is based on observing a significant
number of green tokens in the input text of length $m$, above the expectation
of $\gamma m$. Other schemes from this family include, among else, varying the
aggregation function or the context size.
#### Modeling Red-Green watermarks
Assume a setting where the model chooses a token from some restricted set
$\Sigma\subset V$, following some _context_ $t_{2}$ (longer than the watermark
context $h$), which is preceded by a _prefix_ $t_{1}$. We discuss how to
construct such a setting shortly. To model the watermark, we assume the
following distribution for $p_{t_{1},t_{2}}(x)$, the probability the model
assigns to some $x\in\Sigma$:
$p_{t_{1},t_{2}}(x)=\frac{\exp\left((x^{0}+\delta_{t_{2}}(x)+\varepsilon_{t_{1},t_{2}}(x))/T\right)}{\sum_{w\in\Sigma}\exp\left((w^{0}+\delta_{t_{2}}(w)+\varepsilon_{t_{1},t_{2}}(w))/T\right)}.$
(1)
Here, we assume the existence of the _true logit_ $x^{0}$, modeling the model
bias towards $x$. The true logit is shifted by $\delta_{t_{2}}(x)$, a
$\delta$-Bernoulli random variable, where $\delta\in\mathbb{R}^{+}$ is the
watermark parameter introduced above. Finally, $\varepsilon_{t_{1},t_{2}}(x)$
for different $t_{1},t_{2}$ are iid symmetric error terms with mean $0$ and
variance $\sigma^{2}$. Applying the _logit_ function $p\to\log(p/(1-p))$ to
Eq. 1, we obtain:
$l_{t_{1},t_{2}}(x)=\frac{x^{0}}{T}+\frac{\delta_{t_{2}}(x)}{T}+\frac{\varepsilon_{t_{1},t_{2}}(x)}{T}-\log\left(\sum_{w\in\Sigma\setminus\\{x\\}}\exp\left(\frac{w^{0}}{T}+\frac{\delta_{t_{2}}(j)}{T}+\frac{\varepsilon_{t_{1},t_{2}}}{T}\right)\right).$
(2)
Approximating the log-sum-exp with a maximum, and WLOG setting $w^{0}=0$ for
$w$ which is the maximum in the log-sum-exp (as logits are shift-invariant),
the above simplifies to
$l_{t_{1},t_{2}}(x)=x^{0}/T+\delta^{\prime}_{t_{2}}(x)+\varepsilon^{\prime}_{t_{1},t_{2}}(x),$
(3)
where $\varepsilon^{\prime}_{t_{1},t_{2}}(x)$ absorbs the previous error
terms. The resulting $\delta^{\prime}_{t_{2}}(x)$ is a random variable that is
$0$ for unwatermarked models, and has $3$ possible values for watermarked
models: $\delta/T$ (where $\delta$ is the watermark parameter), if $x$ is the
only token from $\Sigma$ that is in the green partition of the vocabulary,
$-\delta/T$, if $x$ is in the red partition and some other tokens from
$\Sigma$ are in the green partition, and $0$ otherwise. Our test is based on
detecting cases of $\delta^{\prime}_{t_{2}}(x)\neq 0$ and checking that their
occurrence is indeed independent of $t_{1}$, distinguishing model variance
from the watermark bias.
#### Querying strategy
Recall that our goal is to steer a model into a setting (via an instruction,
as we have no access to the completion model) where it makes a choice from
some restricted set $\Sigma$. Importantly, we want to ensure enough
variability in the model’s choices to be able to observe the behavior specific
to a Red-Green watermark and perform efficient estimation (in terms of query
cost).
Assuming an upper bound $H$ on the context size $h$ of the watermark, we use
the following prompt template, parametrized by $t_{1}$ (an arbitrary length
string), $d$ (a single token), and a word list $\Sigma$.
⬇
f"Complete the sentence \"{$t1$} {$\underbrace{d\cdot H}_{t_{2}}$}\" using a
random word from: [{$\Sigma$}]."
Here, $t_{1}$ serves as the prefix, and $t_{2}=d\cdot H$ as the context. For a
watermarked model, changing $t_{2}$ is the only way to change the red-green
vocabulary split, which can greatly affect the model’s choice. For
unwatermarked models, while we often see bias towards some choices from
$\Sigma$, this bias will not strongly depend on $t_{2}$. This holds for all
context sizes $\leq H$ and most aggregation functions, making our setting and
the corresponding test directly applicable to different variants of Red-Green
schemes.
In our instantiation (illustrated in Fig. 1), we use $N$ different $t_{1}$ of
the form “I bought”, varying the verb, $M$ different values of $d$ from the
set of digits, a word list $\Sigma$ of four different plural fruit names, and
an example that uses different words to not bias the model towards a specific
choice. We empirically observe that this setup minimizes the chance that the
model collapses to a single choice or fails to follow the instruction, which
often occurred for similar settings based on repetition of words or numbers
outside of natural context.
#### Estimating the logits
To collect the data for the test, we query the model in two phases. First, we
choose different values of $\Sigma$ until we find one where the model does not
always make the same choice, inducing $Q_{1}$ total queries. We set $x$ to be
the most commonly observed word from $\Sigma$. Next, for each $(t_{1},t_{2})$
pair we query the model until we obtain $K$ valid responses (filtering out
failures to follow the instruction), inducing $Q_{2}$ additional queries.
We use these samples to estimate the model _logits_ corresponding to $x$ as
$\hat{l}_{t_{1},t_{2}}(x)=\log\frac{\hat{p}_{t_{1},t_{2}}(x)}{1-\hat{p}_{t_{1},t_{2}}(x)}$,
where $\hat{p}_{t_{1},t_{2}}(x)$ is the empirically estimated probability of
$x$ in the responses. The result of the adversary’s querying procedure is a
matrix $L_{N\times M}$ (visualized in Fig. 1) of such logit estimates.
#### Testing watermark presence
Finally, we describe the statistical test based on the logit estimates
$L_{N\times M}$. We first estimate $\sigma$, the standard deviation of
$\varepsilon^{\prime}_{t_{1},t_{2}}(x)$, as follows:
$\hat{\sigma}^{2}=\text{median}[\text{Var}_{t_{2}}(L)],$ (4)
using the empirical median to improve robustness to unpredictable behavior
caused by different $t_{1}$. Then, we calculate the following two binary
functions, which flag cases where we believe the model’s probability was
affected by a watermark:
$R_{x}(t_{1},t_{2})=\mathbbm{1}\\{\hat{l}_{t_{1},t_{2}}(x)-\text{median}[L]<-r\hat{\sigma}\\},$
(5)
and
$G_{x}(t_{1},t_{2})=\mathbbm{1}\\{\hat{l}_{t_{1},t_{2}}(x)-\text{median}[L]>r\hat{\sigma}\\},$
(6)
with $r\in\mathbb{R}^{+}$ a parameter of the test. In practice, to account for
model unpredictability, we use the empirical median conditioned on $t_{1}$ in
Eqs. 5 and 6. For simplicity, let us denote $t_{1}\in\\{1,\ldots,N\\}$ and
$t_{2}\in\\{1,\ldots,M\\}$. Let
$cnt_{x}(t_{2})=\max\left(\sum_{t_{1}=1}^{N}R_{x}(t_{1},t_{2}),\sum_{t_{1}=1}^{N}G_{x}(t_{1},t_{2})\right)$
count the number of consistently flagged values for fixed $t_{2}$. We define
the following test statistic:
$S_{x}(L)=\max_{t_{2}\in[M]}cnt_{x}(t_{2})-\min_{t_{2}\in[M]}cnt_{x}(t_{2}).$
(7)
The null hypothesis of our test is $\forall
t_{2}\colon~{}\delta^{\prime}_{t_{2}}(x)=0$, i.e., the model is not
watermarked. To obtain a $p$-value, we apply a Monte Carlo permutation test to
$S_{x}$, checking if the flagged values are correlated with $t_{2}$ in a way
that indicates a Red-Green watermark. Namely, we sample a set of permutations
$\sigma$ of the matrix $L$ uniformly at random, and calculate a 99% confidence
interval of $\Pr[S_{x}(\sigma(L))\geq S_{x}(L)]$, whose upper bound we take as
our $p$-value. When this value is small, we interpret that as evidence of a
watermark. Because Eq. 3 is permutation invariant when
$\delta^{\prime}_{t_{2}}(x)=0$, this ensures that the test does not reject
under the null. This completes our method for detection of Red-Green
watermarks.
## 3 Detecting Fixed-Sampling Watermarks
Unlike Red-Green watermarks, the recently proposed _Fixed-Sampling_ watermarks
do not modify the logit vectors during generation, so estimating the
probabilities of model outputs as above is not informative. Instead, the
sampling is fully determined by the rotation of the watermark key, making the
natural vector to exploit when detecting this watermark its _lack of
diversity_. Given a prompt for which an unwatermarked model is expected to
produce highly diverse outputs, we can use this observation to distinguish
between the two, as illustrated in Fig. 1 (middle).
As in Section 2, we start by introducing the needed background. We then
formally model the diversity of model outputs, discuss our querying strategy
that ensures our assumptions are met, and describe the resulting statistical
test. The effectiveness of our method is evaluated in Section 5.
#### Background
For Fixed-Sampling watermarks, the secret watermark key sequence $\xi$ of
length $n_{key}$ is cyclically shifted uniformly at random for each generation
to obtain $\bar{\xi}$, and the entry $\bar{\xi}_{t}$ is used to sample from
$l$. In the ITS variant, $\bar{\xi}_{t}$ is a pair $(u,\pi)\in[0,1]\times\Pi$,
where $\Pi$ defines the space of permutations of the vocabulary $V$. Given the
probability distribution $p$ over $V$, obtained by applying the softmax
function to $l$, ITS samples the token with the smallest index in the
permutation $\pi$ such that the CDF of $p$ with respect to $\pi$ is at least
$u$. In the EXP variant, $\bar{\xi}_{t}$ is a value $u\in[0,1]$, and we sample
the token $\operatorname*{arg\,min}_{i\in V}-\log(u)/p_{i}$. The detection,
for both variants, is based on testing the correlation between the text and
$\xi$ using a permutation test. As noted in Section 1, the key design goal of
ITS and EXP is that, in expectation w.r.t. $\xi$, they do not distort the
distribution of the responses. How close to this ideal is achieved in
practical implementations is the question we aim to answer.
#### Modeling the diversity
Let $U_{n}(q,t)$ denote a random variable that counts the number of unique
outputs to a fixed prompt $q$ in $n$ queries, each of length exactly $t$ in
tokens. We introduce the _rarefaction curve_ (visualized in Fig. 1) as
$R(n)=\mathbb{E}[U_{n}(q,t)].$ (8)
For suitable $q$ that enables diversity and large enough $t$, the
unwatermarked model exhibits $R(n)=n$. For a Fixed-Sampling watermark (both
ITS and EXP variants), the watermark key segment used for sampling is
determined by choosing a rotation of the key $\xi$ uniformly at random. As
choosing the same rotation for the same prompt and sampling settings will
always yield the same output, the number of unique outputs is at most equal to
the key size $n_{key}$. The probability that an output $i$ was not produced is
given by $(1-1/n_{key})^{n}$. By linearity of expectation, we have the
rarefaction curve
$R(n)=n_{\text{key}}\left(1-(1-\frac{1}{n_{\text{key}}})^{n}\right).$ (9)
#### Querying strategy
To estimate $R(n)$ of $LM$, we query it with a fixed prompt $q$, using
rejection sampling to discard short responses until we obtain a set of $N$
responses of length $t$ tokens (inducing $Q$ total queries). We then
repeatedly sample $n$ responses from this set to get a Monte-Carlo estimation
of $R(n)$. There are two key considerations. First, we need to ensure that we
are in a setting where an unwatermarked model would have $R(n)=n$. To do this,
we use the prompt ‘‘This is the story of’’ that reliably causes diverse
outputs, and set $t$ high enough to minimize the chance of duplicates. In
Section 5 we experimentally confirm that the number of unique outputs scales
exponentially with $t$, and investigate the effect of small temperatures.
Second, as larger $n_{key}$ make $R(n)$ closer to linear, we must ensure that
$n$ is large enough for our test to be reliable. To do this, we can set an
upper bound $\bar{n}_{key}$ on key size, and estimate the number of samples
needed for a given power by simulating the test—our experiments show that our
test succeeds even on values of $\bar{n}_{key}$ far above practical ones.
#### Testing watermark presence
Finally, to test for presence of a Fixed-Sampling watermark, we use a Mann-
Whitney U test to compare the rarefaction curve $R(n)=n$ with the observed
rarefaction data obtained as above. If the test rejects the null hypothesis,
we interpret this as evidence that the model is watermarked with a Fixed-
Sampling scheme. We confirm the effectiveness of our test in Section 5. In
App. C we discuss estimation of the watermark key size $n_{key}$ once the
presence of a Fixed-Sampling watermark is confirmed.
## 4 Detecting Cache-Augmented Watermarks
We proceed to the detection of Cache-Augmented watermarks. While the
underlying techniques used in these schemes are often similar to those of the
Red-Green and Fixed-Sampling families, we focus on a general approach that
exploits the presence of a cache on a fundamental level, and can be generally
applied to any Cache-Augmented scheme. Namely, we note that the cache
sometimes _reveals the true distribution_ of the model, which was also noted
in recent work [15] as an undesirable property for a watermarking scheme. This
implies that the distribution of choices is _cache-conditioned_ (as seen in
Fig. 1, top), which our adversary will exploit to detect the presence of a
watermark.
We first recall relevant background, and then discuss our method: we query the
model in two phases to probe the two cache-conditioned distributions, and
apply a statistical test to detect when they differ, which would indicate the
presence of a Cache-Augmented watermark.
#### Background
The watermarks that we consider in this section use previous $h$ tokens to
reweigh the distribution or apply deterministic sampling at each step. As a
key feature motivated by practical distribution preservation, these schemes
introduce a _cache_ of previously seen contexts. Namely, whenever
$y_{t-h:t-1}$ is already present in the cache, they ignore the watermarking
procedure and instead fall back to standard generation. The cache can be
either global or local (per user) and clears after a certain number $G$ of
generations. We consider three variants: $\delta$-reweight [6],
$\gamma$-reweight [6], and DiPmark [22] with parameter $\alpha$, the details
of which we defer to App. B.
#### Probing the true distribution
In the first phase of querying, our goal is to find a setting where the
distribution of the model under the watermark will differ from its true
distribution, and estimate the true distribution. For schemes we focus on,
this corresponds to a setting with two choices, where the model is not
significantly biased towards any of them. In particular, we use the following
prompt:
⬇
Pick a fruit between: {$f_{1}$} and {$f_{2}$}. Use the following format:
{$uc$}{$f_{example}$},
and modify $f_{1},f_{2},$ and $f_{example}\neq f_{1},f_{2}$ until we find a
setting where the model outputs the two choices roughly uniformly. Crucially,
we prefix the prompt with a randomly sampled sufficiently long string of
tokens $uc$. As $LM$ will repeat $uc$ before providing the answer, this
ensures that if a cache is present, after our first query (the result of which
we discard) the choice of the model will be made according to the true
distribution, as the relevant part of $uc$ was cached. Assuming WLOG that
$f_{1}$ is the more likely choice for the model, we query it $Q_{1}$ times
with the same input to obtain $\hat{p_{1}}$, the estimate of the true
probability of the model to pick $f_{1}$.
#### Probing the watermarked distribution
In the second phase, we query $LM$ with the same input, while ensuring that
the cache is reset between each query, i.e., the model will respond according
to the watermarked distribution. In case of a global cache, it is sufficient
to wait for a set amount of time—resetting the cache too infrequently is not a
realistic setting for a deployment, as it would on average lead to a weak
watermark. The uncommon prefix $uc$ ensures that no other user will
accidentally insert the same context into the cache. In case of a per-user
cache, we can either saturate the cache by asking diverse queries, or use
multiple user accounts. We query $LM$ $Q_{2}$ times and obtain $\hat{p_{2}}$,
the estimate of the probability of $f_{1}$ under the watermarked distribution.
#### Testing watermark presence
For unwatermarked models or those watermarked with a scheme from another
family, both of the previous steps were sampling from the same distribution,
thus for high enough $Q_{1}$ and $Q_{2}$ we expect $\hat{p_{1}}=\hat{p_{2}}$.
However, for all Cache-Augmented watermarking schemes, these probabilities
will differ, indicating that the cache has revealed the true distribution of
the model. To test this, we apply a Fischer’s exact test with the null
hypothesis $\hat{p_{1}}=\hat{p_{2}}$. If we observe a low $p$-value, we
interpret this as evidence that the model is watermarked with a Cache-
Augmented scheme. Our experiments in Section 5 demonstrate that this test is
robust to different scheme variants, and does not lead to false positives when
$LM$ is unwatermarked or uses a scheme from another family. In App. C we
discuss how to distinguish different variants of Cache-Augmented schemes.
## 5 Experimental Evaluation
Table 1: Main results of our watermark detection tests across different models
and watermarking schemes. We report median p-values across 100 repetitions of
the experiment, and for Red-Green schemes additionally over 5 watermarking
keys. p-values below $0.05$ (test passing) are highlighted in bold. $\delta$R
and $\gamma$R denote $\delta$-reweight and $\gamma$-reweight schemes,
respectively.
| | | | | Red-Green | | Fixed-Sampling | | Cache-Augmented
---|---|---|---|---|---|---|---|---|---
| | Unwatermarked | | LeftHash | SelfHash | | ITS/EXP | | DiPmark/$\gamma$R | $\delta$R
Model | Test | $T=$ $1.0$ | $T=$ $0.7$ | | $\delta,\gamma=$ $2,0.25$ | $\delta,\gamma=$ $4,0.5$ | $\delta,\gamma=$ $2,0.5$ | $\delta,\gamma=$ $4,0.25$ | | $n_{key}=$ $256$ | $n_{key}=$ $2048$ | | $\alpha=$ $0.3$ | $\alpha=$ $0.5$ |
Mistral 7B | R-G (Section 2) | 1.000 | 1.000 | | 0.000 | 0.000 | 0.000 | 0.000 | | 1.000 | 1.000 | | 1.000 | 1.000 | 1.000
Fixed (Section 3) | 0.938 | 0.938 | | 0.938 | 0.938 | 0.938 | 0.938 | | 1.3e-125 | 1.1e-9 | | 0.938 | 0.938 | 0.938
Cache (Section 4) | 0.570 | 0.667 | | 0.607 | 0.608 | 1.00 | 0.742 | | 0.638 | 0.687 | | 2.4e-4 | 2.1e-3 | 5.6e-27
Llama2 13B | R-G (Section 2) | 0.149 | 0.663 | | 0.000 | 0.000 | 0.000 | 0.000 | | 0.121 | 0.128 | | 0.149 | 0.149 | 0.149
Fixed (Section 3) | 0.968 | 0.868 | | 0.938 | 0.938 | 0.938 | 0.938 | | 5.5e-124 | 7.5e-8 | | 0.938 | 0.938 | 0.968
Cache (Section 4) | 0.708 | 0.573 | | 0.511 | 0.807 | 0.619 | 0.710 | | 0.518 | 0.692 | | 1.8e-2 | 5.3e-3 | 6.7e-32
Llama2 70B | R-G (Section 2) | 1.000 | 1.000 | | 0.000 | 0.020 | 0.000 | 0.000 | | 1.000 | 1.000 | | 1.000 | 1.000 | 1.000
Fixed (Section 3) | 0.938 | 0.526 | | 0.966 | 0.965 | 0.984 | 0.975 | | 4.2e-125 | 1.7e-8 | | 0.938 | 0.966 | 0.938
Cache (Section 4) | 0.596 | 0.620 | | 0.657 | 0.639 | 0.651 | 0.608 | | 0.463 | 0.818 | | 1.5e-3 | 4.4e-3 | 5.8e-28
In this section, we apply the tests introduced in Sections 2, 3 and 4 to a
wide range of models and watermarking schemes, and confirm their effectiveness
in detecting watermarks. In Section 5.1 we show that the adversary can
reliably detect the watermarking scheme used (if any) at a low cost, across a
wide range of practical settings. In Section 5.2, we provide further
experimental insights into the tests’ robustness. Finally, in Section 5.3, we
demonstrate that our tests can be directly applied to real-world LLM
deployments, revealing no significant evidence of any watermarking schemes at
this point in time.
### 5.1 Main Experiments: Detecting Watermarking Schemes
We perform our main set of experiments to verify the effectiveness of our
tests introduced in Sections 2, 3 and 4 in detecting watermarking schemes in
realistic scenarios.
#### Experimental setup
We run all our tests on five different models (Mistral-7B, Llama2-7B, -13B,
and -70B, and Llama3-8B), in eleven different scenarios. We here present
results of a subset of those, and defer the rest to App. A. In each scenario,
each model is either unwatermarked (where we vary the temperature) or
watermarked with a certain scheme from the three main families (where we vary
the particular scheme and its parameters). If our tests are reliable, we
expect to see low p-values only when the model is watermarked exactly with the
scheme family that we are testing for.
For Red-Green tests, we set $N=10,M=9,r=1.96$, a different $\Sigma$ per model
based on the first $Q1$ samples, use $100$ samples to estimate the
probabilities, and use $10000$ permutations in the test. For Fixed-Sampling
tests, we use $n=1000$ queries and set $t=50$. For Cache-Augmented tests, we
use $Q1=Q2=75$ and assume the cache is cleared between queries in the second
phase.
#### Results: reliable watermark detection
Our main results are shown in Table 1, where we report the median p-values for
each (model, test, scenario) tuple across 100 repetitions of each experiment.
Across all experiments, all three tests reject the null hypothesis (_the
specific watermark is not present_) at a 95% confidence level only when the
scheme from the target family is indeed applied to the model. This confirms
that our tests are reliable in detecting watermarks, and robust with respect
to the model and the specific parameters of the scheme, emphasizing our tests’
generalization to all schemes that are based on the same foundational
principles. In particular, our Red-Green tests are robust to the seeding
scheme, the logit bias $\delta$, and the green token fraction $\gamma$; our
Fixed-Sampling tests maintain high confidence even for unusually high values
of $n_{key}$ and those higher than the number of queries; finally, our Cache-
Augmented tests are robust to all details of the underlying scheme. Our
results also provide evidence that unrelated model modifications do not cause
false positives, as no test passes when the model is unwatermarked or
watermarked with a different scheme family.
Our tests do not incur significant costs for the adversary, making them easily
applicable in practice. While the cost of a particular run varies across
models and scenarios, we can estimate the average cost of the tests to be
below $20 for Red-Green, $1 for Fixed-Sampling, and $0.1 for Cache-Augmented
tests, assuming latest OpenAI GPT4o pricing.
### 5.2 Additional Experiments: Validating the Assumptions
Figure 2: _Left_ : distribution of bootstrapped p-values of the Red-Green test
on Llama2-13B with $(\delta,\gamma)=(2,0.25)$, for different sample sizes. We
see reliable results for $100$ or more samples. _Right_ : the diversity gap
$n-R(n)$ on log scale in different settings. Linear behavior means that
diversity scales exponentially with $t$, and we see that the assumption of
$R(n)=n$ can be easily met in practice.
We present two additional experiments to validate the assumptions made in our
tests and provide further insight into their behavior in practical settings.
#### Sampling in Red-Green tests
The Red-Green test (Section 2) relies on the sampling of model outputs to
estimate the probabilities of the model. As the resulting p-value is computed
assuming knowledge of true probabilities, this raises the question of the
impact of the sampling error on our results. To heuristically mitigate this,
we propose a bootstrapping procedure, where for fixed $(t_{1},t_{2})$, we
sample with replacement from a single set of model outputs, and report the
median p-value $p_{med}$ across such samples. In Fig. 2 (Left) we report the
resulting distribution of $p_{med}$, where one point corresponds to one
independent run. We see that already for $100$ samples per $(t_{1},t_{2})$ (as
used in Table 1), the p-value distribution is narrow and the false positive
rate due to sampling error is well controlled. This confirms that our test is
robust against sampling error and can be used in realistic settings, without
additional access to model logprobs. For computational reasons, we did not
apply this correction in Table 1, where we still observe reliable results in
the median case across experiment repetitions.
#### Diversity assumption in Fixed-Sampling tests
As detailed in Section 3, the Fixed-Sampling test relies on the unwatermarked
model being sufficiently diverse, i.e., for the number of unique outputs
$R(n)=\mathbb{E}[U_{n}(q,t)]$ after $n$ queries with prompt $q$ and response
length $t$, it should hold that $R(n)=n$. Our goal is to show that we can
easily choose $t$ such that this property holds across different settings.
To this end, we hypothesize that the number of unique outputs converges to $n$
exponentially fast as the response length $t$ increases. In particular, we
assume
$R(n)=n-\lfloor n\cdot\exp(-\alpha(T)t)\rfloor,$ (10)
where $\alpha(T)$ is a monotonic function of the temperature $T$. To verify
this hypothesis, we measure $n-R(n)$ on several models and temperatures, and
show the result on log scale in Fig. 2 (Right). If Eq. 10 holds, we expect the
resulting relationship to be linear, which is indeed confirmed by our results.
While $\alpha$ (the slope of the line) varies across settings, we see that a
bit over $200$ tokens would be sufficient for the line to drop to $0$ (not
visible on the log scale plot). This holds even in cases impractical for
deployment of Fixed-Sampling watermarks such as $T=0.4$ [10, 16], indicating
that $R(n)=n$ and our assumption is met, validating our p-values.
### 5.3 Detecting watermarks in deployed models
Table 2: The results of our watermark detection tests on popular black-box LLM
deployments.
| | GPT4 | Claude 3 | Gemini 1.0 Pro
---|---|---|---|---
R-G (Section 2) | | 0.998 | 0.638 | 0.683
Fixed (Section 3) | | 0.938 | 0.844 | 0.938
Cache (Section 4) | | 0.51 | 0.135 | 0.478
Finally, we demonstrate the applicability of the statistical tests introduced
in Sections 2, 3 and 4, by applying them to popular black-box LLM deployments:
GPT4, Claude 3, and Gemini 1.0 Pro. We use the same experimental setup as in
Section 5.1, and use the API access for efficiency reasons—we do not rely on
any additional capabilities, and our tests could be as easily run in the web
interface. For the Cache-Augmented tests, we assume a global cache, that if
present clears after $1000$ seconds. For the Fixed-Sampling test on Claude 3,
due to its lack of diversity, we used $t=75$ tokens per query to ensure the
test’s hypotheses are met.
#### Results
Our results in Table 2 show that the null hypothesis is not rejected for any
of the models and any of the watermark tests. This suggests that none of the
deployed models tested are watermarked at this point in time. The demonstrated
applicability of our tests makes it simple to monitor these deployments and
detect any watermarking schemes that may be introduced in the future.
## 6 Related Work
#### Language model watermarking
Besides the approaches by [6, 8, 10] introduced above, there are various
methods building on similar ideas. Hou et al., [5], Liu et al., 2024b [13],
Ren et al., [18] all apply variations of [8] on semantic information, while Gu
et al., [4] aims to distill a new model from the output of a watermarked
model. Similarly, Liu et al., 2024a [12] apply a Red-Green scheme using a
learned classifier instead of hash functions. A range of works on multi-bit
watermarking [21, 23] aim to not only watermark generated texts but encode
additional information in the watermark itself.
#### Attacks on language model watermarking
Attacks on LLM watermarks have so far been mainly investigated in terms of
scrubbing [7, 8, 19] (i.e., removal of a watermark) and spoofing [7, 19, 4]
(i.e., applying a watermark without knowing $\xi$). Notably, Jovanović et al.,
[7] showed that observing watermarked texts can facilitate both attacks on
various distribution-modifying schemes, disproving common robustness
assumptions [9]. However, using this and similar attacks as means of practical
watermark detection is infeasible, as they generally offer no way to quantify
the attack’s success—in contrast, we aim to provide rigorous statements about
scheme presence. Further, such attacks incur significantly higher query costs
than necessary for detection (as our work demonstrates), and in some cases
assume certain knowledge of the watermark parameters, a setting fundamentally
at odds with our threat model of black-box watermark detection.
The closest related work to ours is Tang et al., [20], that tackles the
problem of watermark detection in strictly simpler settings where the
adversary either has access to an unwatermarked counterpart of the target
model, or can access full model logits. Such knowledge is commonly not
available in practice, limiting the applicability of this approach. To the
best of our knowledge, no work has developed methods for detecting the
presence of a watermark in a realistic black-box setting.
#### Extracting data from black-box models
With many of the most potent LLMs being deployed behind restricted APIs, the
extraction of model details has been an active area of research. This
includes, e.g., the reconstruction of a black-box model tokenizer [17] as well
as the extraction of the hidden-dimensionality or the weights of the embedding
projection layer [2]. Naseh et al., [14] have shown practical attacks to
recover the decoding mechanism of non-watermarked black-box models. Given
access to output logits, Li and Fung, [11] have further demonstrated that it
is possible to train an inversion model that aims to recover the input prompt.
## 7 Conclusion, Impact and Limitations
In this paper, we have focused on the problem of detecting watermarks in large
language models (LLMs) given only black-box access. We developed rigorous
statistical tests for the three most prominent scheme families, and validated
their effectiveness on a wide range of schemes and real-world models. Our
results imply that protecting from detectability should not be the key focus
of LLM watermarks, and that other properties such as robustness to attacks,
text quality, and efficiency should be prioritized.
#### Broader Impact
While our work primarily enables malicious parties to more easily circumvent
attempts at tracing LLM-generated text, we believe the benefits of our work
outweigh those risks, as our conclusions help model providers calibrate their
expectations in terms of watermark detectability, and avoid a false sense of
security.
#### Limitations
One limitation of our tests is that they are restricted to the three scheme
families discussed in Sections 2, 3 and 4. These are indeed the most prominent
in the literature, and as our tests are based on fundamental properties of
these scheme families, they should generalize to more variants and
combinations of the underlying ideas. However, it is possible that a model
provider deploys a scheme based on an entirely novel idea, which our tests
would not be able to detect. Further, our p-values are based on several model
assumptions, such as symmetric error terms and perfect sampling. While we
validate that these assumptions are sufficiently met on several open-source
models, we cannot guarantee that all models adhere to them. Finally, another
conceptual limitation of our Red-Green test is that it does not take into
account the possibility that the red-green vocabulary split is the same (on
the observed domain) for all contexts. The probability of this event decreases
exponentially with the number of different contexts, and is thus unlikely to
affect the test’s performance in practice.
## References
* Bommasani et al., [2021] Bommasani, R., Hudson, D. A., Adeli, E., Altman, R. B., Arora, S., von Arx, S., Bernstein, M. S., Bohg, J., Bosselut, A., Brunskill, E., Brynjolfsson, E., Buch, S., Card, D., Castellon, R., Chatterji, N. S., Chen, A. S., Creel, K., Davis, J. Q., Demszky, D., Donahue, C., Doumbouya, M., Durmus, E., Ermon, S., Etchemendy, J., Ethayarajh, K., Fei-Fei, L., Finn, C., Gale, T., Gillespie, L., Goel, K., Goodman, N. D., Grossman, S., Guha, N., Hashimoto, T., Henderson, P., Hewitt, J., Ho, D. E., Hong, J., Hsu, K., Huang, J., Icard, T., Jain, S., Jurafsky, D., Kalluri, P., Karamcheti, S., Keeling, G., Khani, F., Khattab, O., Koh, P. W., Krass, M. S., Krishna, R., Kuditipudi, R., and et al. (2021). On the opportunities and risks of foundation models. CoRR.
* Carlini et al., [2024] Carlini, N., Paleka, D., Dvijotham, K. D., Steinke, T., Hayase, J., Cooper, A. F., Lee, K., Jagielski, M., Nasr, M., Conmy, A., Wallace, E., Rolnick, D., and Tramèr, F. (2024). Stealing part of a production language model. CoRR.
* Council of the European Union, [2024] Council of the European Union (2024). Proposal for a regulation of the european parliament and of the council laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts - analysis of the final compromise text with a view to agreement.
* Gu et al., [2024] Gu, C., Li, X. L., Liang, P., and Hashimoto, T. (2024). On the learnability of watermarks for language models. ICLR.
* Hou et al., [2023] Hou, A. B., Zhang, J., He, T., Wang, Y., Chuang, Y., Wang, H., Shen, L., Durme, B. V., Khashabi, D., and Tsvetkov, Y. (2023). Semstamp: A semantic watermark with paraphrastic robustness for text generation. arXiv.
* Hu et al., [2024] Hu, Z., Chen, L., Wu, X., Wu, Y., Zhang, H., and Huang, H. (2024). Unbiased watermark for large language models. ICLR.
* Jovanović et al., [2024] Jovanović, N., Staab, R., and Vechev, M. (2024). Watermark stealing in large language models. arXiv.
* Kirchenbauer et al., [2023] Kirchenbauer, J., Geiping, J., Wen, Y., Katz, J., Miers, I., and Goldstein, T. (2023). A watermark for large language models. In ICML.
* Kirchenbauer et al., [2024] Kirchenbauer, J., Geiping, J., Wen, Y., Shu, M., Saifullah, K., Kong, K., Fernando, K., Saha, A., Goldblum, M., and Goldstein, T. (2024). On the reliability of watermarks for large language models. ICLR.
* Kuditipudi et al., [2024] Kuditipudi, R., Thickstun, J., Hashimoto, T., and Liang, P. (2024). Robust distortion-free watermarks for language models. TMLR.
* Li and Fung, [2013] Li, Y. and Fung, P. (2013). Improved mixed language speech recognition using asymmetric acoustic model and language model with code-switch inversion constraints. In ICASSP.
* [12] Liu, A., Pan, L., Hu, X., Li, S., Wen, L., King, I., and Yu, P. S. (2024a). A private watermark for large language models. ICLR.
* [13] Liu, A., Pan, L., Hu, X., Meng, S., and Wen, L. (2024b). A semantic invariant robust watermark for large language models. ICLR.
* Naseh et al., [2023] Naseh, A., Krishna, K., Iyyer, M., and Houmansadr, A. (2023). On the risks of stealing the decoding algorithms of language models. CoRR.
* Pang et al., [2024] Pang, Q., Hu, S., Zheng, W., and Smith, V. (2024). Attacking LLM watermarks by exploiting their strengths. arXiv.
* Piet et al., [2023] Piet, J., Sitawarin, C., Fang, V., Mu, N., and Wagner, D. A. (2023). Mark my words: Analyzing and evaluating language model watermarks. arXiv.
* Rando and Tramèr, [2024] Rando, J. and Tramèr, F. (2024). The worst (but only) claude 3 tokenizer. https://github.com/javirandor/anthropic-tokenizer.
* Ren et al., [2023] Ren, J., Xu, H., Liu, Y., Cui, Y., Wang, S., Yin, D., and Tang, J. (2023). A robust semantics-based watermark for large language model against paraphrasing. arXiv.
* Sadasivan et al., [2023] Sadasivan, V. S., Kumar, A., Balasubramanian, S., Wang, W., and Feizi, S. (2023). Can ai-generated text be reliably detected? arXiv.
* Tang et al., [2023] Tang, L., Uberti, G., and Shlomi, T. (2023). Baselines for identifying watermarked large language models. CoRR.
* Wang et al., [2024] Wang, L., Yang, W., Chen, D., Zhou, H., Lin, Y., Meng, F., Zhou, J., and Sun, X. (2024). Towards codable text watermarking for large language models. ICLR.
* Wu et al., [2023] Wu, Y., Hu, Z., Zhang, H., and Huang, H. (2023). Dipmark: A stealthy, efficient and resilient watermark for large language models. arXiv.
* [23] Yoo, K., Ahn, W., and Kwak, N. (2024s). Advancing beyond identification: Multi-bit watermark for language models. NAACL.
## Appendix A Extending Experiments across Diverse Settings
Table 3: Additional results of our watermark detection tests across different
models and watermarking schemes. We report median p-values across 100
repetitions of the experiment, and for Red-Green schemes additionally over 5
watermarking keys. p-values below $0.05$ (test passing) are highlighted in
bold. $\delta$R and $\gamma$R denote $\delta$-reweight and $\gamma$-reweight
schemes, respectively.
| | | | | Red-Green | | Fixed-Sampling | | Cache-Augmented
---|---|---|---|---|---|---|---|---|---
| | Unwatermarked | | LeftHash | SelfHash | | ITS/EXP | | DiPmark/$\gamma$R | $\delta$R
Model | Test | $T=$ $1.0$ | $T=$ $0.7$ | | $\delta,\gamma=$ $2,0.25$ | $\delta,\gamma=$ $2,0.5$ | $\delta,\gamma=$ $4,0.25$ | $\delta,\gamma=$ $4,0.5$ | $\delta,\gamma=$ $2,0.25$ | $\delta,\gamma=$ $2,0.5$ | $\delta,\gamma=$ $4,0.25$ | $\delta,\gamma=$ $4,0.5$ | | $n=$ $256$ | $n=$ $2048$ | | $\alpha=$ $0.3$ | $\alpha=$ $0.5$ |
Mistral 7B | R-G (Section 2) | 1.000 | 1.000 | | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | | 1.000 | 1.000 | | 1.000 | 1.000 | 1.000
Fixed (Section 3) | 0.938 | 0.938 | | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | | 1.3e-125 | 1.1e-9 | | 0.938 | 0.938 | 0.938
Cache (Section 4) | 0.570 | 0.667 | | 0.607 | 0.639 | 0.632 | 0.608 | 1.000 | 1.000 | 0.742 | 0.824 | | 0.638 | 0.687 | | 2.4e-4 | 2.1e-3 | 5.6e-27
Llama2 13B | R-G (Section 2) | 0.149 | 0.663 | | 0.000 | 0.014 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | | 0.121 | 0.128 | | 0.149 | 0.149 | 0.149
Fixed (Section 3) | 0.968 | 0.868 | | 0.938 | 0.966 | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | 0.967 | | 5.5e-124 | 7.5e-8 | | 0.938 | 0.938 | 0.968
Cache (Section 4) | 0.708 | 0.573 | | 0.511 | 0.596 | 0.623 | 0.807 | 0.657 | 0.619 | 0.710 | 0.583 | | 0.518 | 0.692 | | 1.8e-2 | 5.3e-3 | 6.7e-32
Llama2 70B | R-G (Section 2) | 1.000 | 1.000 | | 0.000 | 0.000 | 0.000 | 0.020 | 0.000 | 0.020 | 0.000 | 0.000 | | 1.000 | 1.000 | | 1.000 | 1.000 | 1.000
Fixed (Section 3) | 0.938 | 0.526 | | 0.966 | 0.964 | 0.889 | 0.965 | 0.963 | 0.984 | 0.975 | 0.990 | | 4.2e-125 | 1.7e-8 | | 0.938 | 0.966 | 0.938
Cache (Section 4) | 0.596 | 0.620 | | 0.657 | 0.797 | 0.824 | 0.639 | 0.535 | 0.651 | 0.608 | 0.593 | | 0.463 | 0.818 | | 1.5e-3 | 4.4e-3 | 5.8e-28
Llama2 7B | R-G (Section 2) | 1.000 | 1.000 | | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | | 1.000 | 1.000 | | 1.000 | 1.000 | 1.000
Fixed (Section 3) | 0.993 | 0.988 | | 0.938 | 0.938 | 0.938 | 0.938 | 0.968 | 0.938 | 0.938 | 0.938 | | 7.6e-124 | 1.0e-8 | | 0.938 | 0.938 | 0.993
Cache (Section 4) | 0.604 | 0.602 | | 0.623 | 0.705 | 0.728 | 0.593 | 0.620 | 0.718 | 0.610 | 0.593 | | 0.476 | 0.588 | | 4.2e-6 | 4.5e-7 | 1.3e-21
Llama3 8B | R-G (Section 2) | 1.000 | 1.000 | | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | 0.000 | | 1.000 | 1.000 | | 1.000 | 1.000 | 1.000
Fixed (Section 3) | 0.938 | 0.938 | | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | 0.938 | 0.986 | 0.938 | | 1.2e-124 | 3.8e-8 | | 0.938 | 0.938 | 0.938
Cache (Section 4) | 0.734 | 0.504 | | 0.605 | 0.514 | 0.712 | 0.605 | 0.600 | 0.731 | 0.729 | 0.714 | | 0.618 | 0.605 | | 5.2e-5 | 3.2e-8 | 3.5e-18
#### Watermark detection with additional models and schemes
We extend the experiments from Section 5.1 using two additional models,
Llama2-7B and Llama3-8B, as well as more variations of the watermarking
schemes’ parameters to further assess the robustness of the tests. The
experimental setup for the additional results is consistent with the one
described in Section 5.1.
Our exhaustive results are presented in Table 3. The same conclusion applies
to the two additional models: the null hypothesis (_the specific watermark is
not present_) is rejected at a 95% confidence level only when the
corresponding watermarking scheme is applied to the model. These results
confirm that the modeling assumptions for each test are satisfied across a
wide range of models, indicating the tests’ relevance in practical scenarios.
#### Multiple keys in Red-Green watermarks
To demonstrate that the Red-Green test is robust to variations in the
watermarking scheme within the same watermark family, we consider the case of
Red-Green watermarks with multiple keys, where the key $\xi$ is uniformly
selected from a predefined pool of keys at each generation. Using $n$ keys
turns Eq. 3 into
$l_{t_{1},t_{2}}(x)=x^{0}/T+\delta^{\prime\prime}_{t_{2}}(x)+\varepsilon^{\prime}_{t_{1},t_{2}}(x),$
(11)
with $\delta^{\prime\prime}_{t_{2}}(x)$ in $\\{k\delta/(nT)\,|\,\forall
k\in\\{-n,...,n\\}\\}$ is obtained by averaging the variables
$\delta^{\prime}_{t_{2}}(x)$ over the set of keys. Despite modeling changes,
the core assumption of logit bias being conditioned on $t_{2}$ remains
unchanged. Therefore, we can apply the same test as in Section 2 to detect the
watermark. Consequently, we conducted the Red-Green test on both the LeftHash
and SelfHash variants using $n=3$ keys on three models (Llama2-13B, Llama2-70B
and Mistral-7B). Recent work [15] shows that using too many keys can lead to
other vulnerabilities.
Across all three models and scheme parameters, the null hypothesis (_the Red-
Green watermark is not present_) is rejected at a 95% confidence level, with
median p-values lower than $1e\text{-}4$ for each combination of model and
setting. It shows that the Red-Green test is robust even in settings that
slightly deviate from the original modeling considered in Section 2. It
emphasizes the test’s reliance on the foundational principles behind Red-Green
schemes rather than on specific implementation details.
## Appendix B Details of Cache-Augmented Schemes
We provide the details of the three Cache-Augmented watermarking schemes
considered in this work: $\delta$-reweight [6], $\gamma$-reweight [6], and
DiPmark [22], that were omitted from Section 4.
All three variants, at each generation step $t$, jointly hash the watermark
key $\xi\in\mathbb{Z}_{2}^{K}$ and the preceding context $y_{t-h:t-1}$
(commonly setting $h=5$) using SHA256, and use the result as a seed to sample
a _code_ $E_{t}\in P_{E}$ uniformly at random. Let $p$ denote the probability
distribution over $V$, obtained by applying the softmax function to the
logits. For the $\delta$-reweight variant, $P_{E}=[0,1]$, and the code $E_{t}$
is used to sample the token in $V$ with the smallest index, such that the CDF
of $p$ is at least $E_{t}$. For the $\gamma$-reweight variant and DiPmark,
$P_{E}$ is the space of permutations of $V$. For $\gamma$-reweight, we
transform $p$ to a new distribution $p^{\prime}$ by, for each token $i\in V$,
setting $p^{\prime}(i)=f_{2}(f_{1}(E_{t}(i)))-f_{2}(f_{1}(E_{t}(i)-1))$, where
we have $f_{1}(i^{\prime})=\sum_{j\in V}{\mathbbm{1}}(E_{t}(j)\leq
i^{\prime})p(j)$ and $f_{2}(v)=\max(2v-1,0)$, effectively dropping the first
half of the permuted CDF. For DiPmark, given parameter $\alpha\in[0,0.5]$,
this is generalized by using $f_{2}(v)=\max(v-\alpha,0)+\max(v-(1-\alpha),0)$,
recovering $\gamma$-reweight for $\alpha=0.5$. The former two variants perform
detection using a log-likelihood ratio test (requiring access to $LM$), while
DiPmark uses a model-independent test.
## Appendix C Estimating Scheme Parameters
In this section, we describe how, mostly leveraging the queries already
performed during the detection tests, we can estimate the main parameters of
the detected watermarking scheme. This demonstrates the soundness of our
modeling assumptions in Sections 2, 3 and 4 from which we derive all the
estimators.
### C.1 Estimation of Red-Green Watermarking Scheme Parameters
If the null hypothesis (_the model not being Red-Green watermarked_) is
rejected, we can then estimate the scheme-specific watermark parameters
$\delta$ and the context size $h$ using mostly the same data as used in the
test. First, we describe the estimators for both parameters, and then discuss
their practicality by analyzing their performance on multiple models.
#### Description of estimators
To estimate $\delta$, we establish a parametrized model based on Eq. 2 that
relies on our flagging function from Eq. 6, and additional estimates
$\hat{l}_{t_{1},t_{2}}(w)$ for every $w\in\Sigma$, computed on the same data
as above, requiring no additional queries. For each $w\in\Sigma$, we set:
$\hat{l}_{t_{1},t_{2}}(w)=\bar{w}_{t_{1}}+G_{w}(t_{1},t_{2})\bar{\delta}-\log\left(\sum_{w^{\prime}\in\Sigma\setminus\\{w\\}}\exp\left(\bar{w}^{\prime}_{t_{1}}+G_{w^{\prime}}(t_{1},t_{2})\bar{\delta}\right)\right),$
(12)
and set $\forall t_{1}$, $\bar{w}_{t_{1}}=0$ for a single $w\in\Sigma$, as
logits are shift-invariant. Fitting the parameters $\bar{\delta}$ and all
$\bar{w}_{t_{1}}$ by minimizing the mean squared error with gradient descent
allows us to recover $\delta/T$ as $\bar{\delta}$. If $T$ is known or
estimated separately, this term can be removed.
Let $h$ denote the context size, i.e., the number of _previous_ tokens
considered by the watermarking scheme. To estimate $h$, we use the same prompt
template as in Section 2, with a fixed prefix $t_{1}$ and digit $d$, but with
a varying $H$ and perturbation digit $d^{\prime}$ prepended to $d$.
⬇
f"Complete the sentence \"{$t1$} {$d^{\prime}$}{$d\cdot H$}\" using a random
word from: [{$\Sigma$}]."
The probability distribution of the model output will abruptly change when $H$
exceeds the context size $h$, as the change in $d^{\prime}$ will not alter the
red/green split of the vocabulary. By performing a series of pairwise Mood’s
tests on the estimated log-probabilities of $x$ we can find the largest $H$
for which $t_{1}$ is part of the context. This corresponds to the first value
of $H$ for which the test is rejected.
Estimating $\gamma$ is more challenging, as in contrast to $\delta$, this
parameter is not directly reflected in the logits but rather defines a more
global behavior of the scheme. This is particularly difficult for schemes with
self-seeding, as the rejection sampling interferes with the behavior of
$\gamma$. We leave further exploration of this problem to future work.
Figure 3: Estimation of $\delta$ for different models using LeftHash with
$\gamma=0.25$. The number of samples used increases from left to right, with
the leftmost plot assuming direct access to the log-probabilities. The
estimation is done on the same data as the test. Error bars are given by the
95% bootstrapped confidence interval with respect to the sampling of the model
outputs.
#### Experimental results
We computed the estimator for $\delta$ on the LeftHash variant with
$\gamma=0.25$ and varying $\delta$ from 0 to 4. The results are shown in Fig.
3, with the 95% confidence intervals reflecting the sampling error. The
estimator successfully estimates $\delta$ for all models with a sufficient
number of samples, using only the estimated output probabilities of the model.
It is also consistent across all models tested, suggesting that the model
assumptions in Eq. 1 are met in practice.
Estimating the context size for Red-Green schemes requires performing a new
attack once the model is flagged as watermarked. We estimate the context size
for three different models (Mistral-7B, Llama2-13B and Llama2-70B) using
LeftHash with $\delta=2$ and $\gamma=0.25$. The estimation process requires an
additional 5,000 queries, and the estimator successfully determines the
context size for all models. However, the estimator is less robust on the
SelfHash variant due to the self-seeding algorithm, which leads to higher
probabilities for tokens in $\Sigma$ being in the green vocabulary, and thus
diminishes the perturbation’s significance and resulting in false negatives in
Mood’s test. Therefore, the procedure stated above produces a lower bound for
the context size. To mitigate this issue, we use the estimator across 10
different $t_{2}$ and then consider the median of the $10$ estimators as our
final estimator. This estimator applied on the SelfHash variant with
$\delta=2$ and $\gamma=0.25$ is successful on all three models. It also does
not change the results on LeftHash and can be used as a more robust estimator
for $h$ in all cases, when the additional cost of $50,000$ queries is not
prohibitive.
### C.2 Estimation of Fixed-Sampling Watermarking Scheme Parameters
Our approach does not distinguish between the variant of the Fixed-Sampling
watermark used (ITS or EXP), as the diversity property that we exploit is
common to both. The only other relevant parameter of Fixed-Sampling schemes is
$n_{\text{key}}$. To estimate it, we use non-linear regression on the
rarefaction curve using (9) and the same data that we used for the presence
test, and compute the confidence intervals using bootstrapping.
Our results are given in Table 4. We see that the estimator is consistent
across different models and remains relatively precise even for values of
$n_{\text{key}}$ higher than the number of queries.
Table 4: Key length estimation for Fixed-Sampling watermarks using non-linear
regression on the rarefaction curve.
Key length | | Llama2-7B | Llama2-13B | Llama2-70B | Llama3-8B | Mistral-7B
---|---|---|---|---|---|---
256 | | $259\pm 0.6$ | $259\pm 0.5$ | $256\pm 0.5$ | $257\pm 0.5$ | $256\pm 0.6$
2048 | | $1978\pm 10$ | $2107\pm 12$ | $2006\pm 13$ | $2070\pm 14$ | $1831\pm 10$
### C.3 Estimation of Cache-Augmented Watermarking Scheme Parameters
For Cache-Augmented watermarks, we can estimate which scheme variant is
present, and if the variant is DiPmark, attempt to learn the value of $\alpha$
(recall that $\alpha=0.5$ corresponds to $\gamma$-reweight). To do this, we
use the same approach as in Section 4 to obtain $\hat{p_{1}}$ and
$\hat{p_{2}}$, where WLOG we assumed $p_{1}>0.5$. If we observe
$\hat{p_{2}}=0$ this directly implies a $\delta$-reweight watermark. If we
observe $\hat{p_{2}}\in(0,1)$, we learn the following: if
$\hat{p_{2}}=2\hat{p_{1}}-1$ then $\alpha>1-\hat{p_{1}}$, otherwise
$\alpha=|\hat{p_{1}}-\hat{p_{2}}|$. The bound in the first case can be further
improved with additional queries with different $p_{1}$. Finally, if we
observe $\hat{p_{2}}=1$ we repeat the whole procedure in total $K$ times,
following the same case distinction—if $\hat{p_{2}}=1$ repeats in all $K$
runs, we conclude that the model is watermarked with $\delta$-reweight.
Using the same parameters as the one for the test, we distinguish with 100%
accuracy between a DiPmark and a $\delta$-reweight watermark. However, the
estimation of $\alpha$ becomes unreliable for higher values of $\alpha$,
especially for smaller models. One of the reasons for this are the failures of
the model to follow the instruction, that are more common in the presence of
the uncommon prefix $uc$. While the detection test in Section 4 was robust to
such behavior, this does not hold for the estimation of $\alpha$.
|
# The Surprising Lack of Effect from Stellar Feedback on the Gas Stripping
Rate from Massive Jellyfish Galaxies
Nina Akerman,1,2 Stephanie Tonnesen,3 Bianca Maria Poggianti,1 Rory Smith,4
Antonino Marasco,1 Andrea Kulier,1 Ancla Müller,5 and Benedetta Vulcani,1
1INAF - Astronomical Observatory of Padova, vicolo dell’Osservatorio 5,
IT-35122 Padova, Italy
2Dipartimento di Fisica e Astronomia ‘Galileo Galilei’, Università di Padova,
vicolo dell’Osservatorio 3, IT-35122, Padova, Italy
3Flatiron Institute, CCA, 162 5th Avenue, New York, NY 10010, USA
4Departamento de Física, Universidad Técnica Federico Santa María, Vicuña
Mackenna 3939, San Joaquín, Santiago de Chile
5Ruhr University Bochum (RUB), Faculty of Physics and Astronomy, Astronomical
Institute, Universitätsstr. 150, 44801 Bochum, Germany E-mail:
<EMAIL_ADDRESS>
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
We study the role of star formation and stellar feedback in a galaxy being ram
pressure stripped on its infall into a cluster. We use hydrodynamical wind-
tunnel simulations of a massive galaxy ($M_{\text{star}}=10^{11}M_{\odot}$)
moving into a massive cluster ($M_{\text{cluster}}=10^{15}M_{\odot}$). We have
two types of simulations: with and without star formation and stellar
feedback, SF and RC respectively. For each type we simulate four realisations
of the same galaxy: a face-on wind, edge-on wind, $45^{\circ}$ angled wind,
and a control galaxy not subject to ram pressure. We directly compare the
stripping evolution of galaxies with and without star formation. We find that
stellar feedback has no direct effect on the stripping process, i.e. there is
no enhancement in stripping via a velocity kick to the interstellar medium
gas. The main difference between RC and SF galaxies is due to the indirect
effect of stellar feedback, which produces a smoother and more homogeneous
interstellar medium. Hence, while the average gas surface density is
comparable in both simulation types, the scatter is broader in the RC
galaxies. As a result, at the galaxy outskirts overdense clumps survive in RC
simulation, and the stripping proceeds more slowly. At the same time, in the
inner disc, underdense gas in the RC holes is removed faster than the smoothly
distributed gas in the SF simulation. For our massive galaxy, we therefore
find that the effect of feedback on the stripping rate is almost negligible,
independent of wind angle.
###### keywords:
methods: numerical – galaxies: evolution – galaxies: clusters: intracluster
medium – galaxies: ISM
††pubyear: 2023††pagerange: The Surprising Lack of Effect from Stellar
Feedback on the Gas Stripping Rate from Massive Jellyfish Galaxies–C
## 1 Introduction
A cluster galaxy may be subject to processes specific to high-density
environments. Among them is ram pressure stripping (RPS; Gunn & Gott, 1972) in
which the intracluster medium (ICM) interacts with the interstellar medium
(ISM) of the galaxy, removing it as the galaxy falls into the cluster
potential well. As a result, over time the galaxy quenches its star formation,
while the stripped gas forms tails trailing behind the galaxy disc (Gavazzi et
al., 2001; Sun et al., 2010, 2021; Fumagalli et al., 2014; Kenney et al.,
2014; Fossati et al., 2016; Jáchym et al., 2017, 2019; George et al., 2018;
Moretti et al., 2020a). Observational evidence (corroborated by simple
theoretical predictions, e.g. Gunn & Gott, 1972) suggests that as the gas is
stripped outside-in (Cortese et al., 2012; Merluzzi et al., 2016; Gullieuszik
et al., 2017; Fossati et al., 2018; Cramer et al., 2019; Vulcani et al., 2020)
and the disc is left truncated with a disturbed morphology (Koopmann & Kenney,
2004; Crowl & Kenney, 2008; Boselli et al., 2014; Fritz et al., 2017; Kenney
et al., 2015). The effects outlined above have long been observed in HI gas
(Cayatte et al., 1990; Kenney et al., 2004; Chung et al., 2007), which can be
ionised by in-situ star formation or other processes, giving rise to H$\alpha$
emitting-tails (Fumagalli et al., 2014; Merluzzi et al., 2016; Poggianti et
al., 2017; Bellhouse et al., 2017; Gullieuszik et al., 2017). Recent
observations showed that such tails also contain large amounts of molecular
gas, partly formed in-situ and partly stripped in the vicinity of the disc
(Lee et al., 2017; Jáchym et al., 2017, 2019; Lee & Chung, 2018; Moretti et
al., 2018, 2020a, 2020b).
Galaxy evolution under RPS has also been studied in cosmological and idealised
simulations. Some of the first simulations used a constant ICM wind (Schulz &
Struck, 2001; Roediger & Brüggen, 2006; Tonnesen & Bryan, 2009), though
Tonnesen (2019) showed that modelling ram pressure (RP) whose magnitude
increases with time (as experienced by a galaxy falling into a cluster) is
crucial for obtaining realistic stripping profiles of galaxies that can be
compared to the observed ones. Different angles at which the wind hits the
galaxy have also been modelled (Roediger & Brüggen, 2006, 2007; Jáchym et al.,
2007; Bekki, 2014; Steinhauser et al., 2016; Akerman et al., 2023), with a
conclusion that there is a weak dependence of the stripping process on the
inclination angle unless it is close to edge-on stripping. There are also some
simulations that include a magnetised ISM (Dursi & Pfrommer, 2008; Ruszkowski
et al., 2014; Tonnesen & Stone, 2014; Ramos-Martínez et al., 2018) and cosmic
rays (Farber et al., 2022), which generally find small changes to the global
stripping rate but more significant differences in the gas phase distribution.
Whether RP immediately quenches a galaxy or is able to cause some star
formation enhancement via gas compression has long been debated. The latest
observations show that some enhancement can indeed take place (Vulcani et al.,
2018; Roberts & Parker, 2020; Roberts et al., 2021, 2022a, 2022b). Simulations
agree as well, noting that the star formation rate (SFR) would be increased
only temporarily and that whether or not the SFR enhancement would happen
depends on many factors, such as galaxy mass, wind inclination angle and RP
strength (Tonnesen & Bryan, 2012; Bekki, 2014; Roediger et al., 2014;
Steinhauser et al., 2016; Ruggiero & Lima Neto, 2017; Lee et al., 2020).
Göller et al. (2023) find similar results in a large cosmological simulation
(TNG50), where SFR is enhanced in individual stripped galaxies at some point
of their evolution, although as a population RPS galaxies have low SFRs.
Although simulations have extensively studied the effect of RPS on the SFR, it
is still unclear how, in turn, star formation and stellar feedback (hereafter
just ‘feedback’ since this is the only type of feedback that we model)
influence gas stripping. Bahé & McCarthy (2015) show that feedback assists RP
in removing gas from a galaxy, with the effect being stronger for low-mass
galaxies with $M_{\text{star}}=(2-5)\times 10^{9}M_{\odot}$. Kulier et al.
(2023) also find that in the EAGLE simulation (Schaye et al., 2015) stripping
proceeds quicker due to the stellar feedback, and they note that the feedback
might be overestimated in EAGLE (see also Bahé et al., 2016) due to the
subgrid implementation and cosmological-scale resolution.
In this work, we study how star formation and, importantly, stellar feedback,
affect the process of RPS by comparing gas in galaxies from two sets of
simulations, a first set that includes radiative cooling, and a second set
that adds star formation and feedback prescriptions to the radiative cooling
simulations. We can naively imagine two scenarios. In the first one, feedback
might impede gas stripping by adding more thermal and nonthermal pressure
support to and above the ISM that will block the RP from making direct contact
with the disc (this will be most effective when the wind has a significant
component in the face-on direction). On the other hand, feedback might enhance
gas removal by moving ISM gas higher above and below the disc where it is less
tightly gravitationally bound. In this work we find that the impact of
feedback is more nuanced than either of the above simple sketches.
The paper is organised as follows. First, we outline our simulations in
Section 2. In Section 3 we calculate and compare the stripping rates of star-
forming versus radiative-cooling-only galaxies, and in Section 4 we explore
what drives the differences in RPS in these sets of simulations. We compare
our results to other works in Section 5 and draw conclusions in Section 6.
## 2 Methodology
We describe our simulations in detail in our previous work, Akerman et al.
(2023), and here will briefly highlight their main features. Using the
adaptive mesh refinement code Enzo (Bryan et al., 2014), we make a simulation
box with 160 kpc on a side, with 5 levels of refinement allowing for a maximum
resolution of 39 pc. The refinement criteria are Jeans length and baryon mass
of $\approx 7500M_{\odot}$. In Appendix A we perform a resolution test and
show how our refinement criteria resolve the galaxy disc.
The simulations include radiative cooling using the grackle library (Smith et
al., 2017) with metal cooling and the UV background by Haardt & Madau (Haardt
& Madau, 2012). The disc starts with metallicity $Z=1.0\;Z_{\odot}$ and the
ICM has $Z=0.3\;Z_{\odot}$.
We use star formation and stellar feedback recipes by Goldbaum et al. (2015,
2016). If the mass of a cell exceeds the Jeans mass and the minimum threshold
number density of $n_{\rm{thresh}}=10\,\rm{cm}^{-3}$, a stellar particle will
form with a minimum mass of $10^{3}M_{\odot}$, assuming a star formation
efficiency of 1 per cent. We also simulated galaxies from the same initial
conditions (described below) with $n_{\rm{thresh}}=3\,\rm{cm}^{-3}$ and
$30\,\rm{cm}^{-3}$ and found that galaxies with all three $n_{\rm{thresh}}$
evolve similarly in their star formation history. We study the effect that the
choice of the threshold number density might have on our results in Appendix
B, where we also outline the role of star formation itself in shaping the ISM
distribution. Thus, the parameter choice for the $n_{\rm{thresh}}$ does not
notably affect the results. The feedback includes momentum and energy input
from supernovae (SNe) (that also increase cell metallicity), ionising
radiation from young stars (heating up to $10^{4}$ K), and winds from evolved
massive stars. SNe have combined energy budget of $10^{51}$ erg; first the
terminal momentum input is added from the number of SNe expected at a given
time step, then any additional energy is added as thermal energy. The total
momentum of $3\times 10^{5}\text{ M}_{\odot}\text{ km }\text{s}^{-1}$ is
distributed equally among the 26 nearest neighbour cells. The maximum change
in velocity of any given cell is limited to 1000 km s-1.
We follow the setup by Roediger & Brüggen (2006) and Tonnesen & Bryan (2009),
modelling static potentials for the stellar disc and the spherical dark matter
halo, while calculating the self-gravity of the gas component at each time
step. For the Plummer–Kuzmin stellar disc (Miyamoto & Nagai, 1975) we use a
mass of $M_{\text{star}}=10^{11}M_{\odot}$, a scale length of
$r_{\text{star}}=5.94$ kpc and a scale height of $z_{\text{star}}=0.58$ kpc.
We model a Burkert profile for the dark matter halo (Burkert, 1995; Mori &
Burkert, 2000) with a core radius of $r_{\text{DM}}=17.36$ kpc111While an
Navarro–Frenk–White dark matter potential has been found by many to be more
appropriate for massive galaxies (e.g. Di Cintio et al., 2014), a Burkert
profile has been found to be a better match to some observational samples of
massive galaxies (e.g. Rodrigues et al., 2017; Li et al., 2020). Importantly
for this work, the Burkert potential we use leads to a good match between the
rotation curve of our simulated galaxy and JO201. To orient the reader, the
R200 and M200 of our halo are 340 kpc and 1.15$\times$1012 M⊙, respectively..
Initial parameters of the gaseous disc are the following: mass
$M_{\text{gas}}=10^{10}M_{\odot}$, scale length $r_{\text{gas}}=10.1$ kpc and
scale height $z_{\text{gas}}=0.97$ kpc. The values were chosen to match a
well-studied RPS galaxy, JO201 (Bellhouse et al., 2017, 2019). As with the
previous work, our intention here is to merely select realistic initial
conditions based on an observed stripped galaxy, not to model it exactly. A
more detailed comparison between our simulated galaxies and JO201 can be found
in Akerman et al. (2023), Section 7.1.
We fix the galaxy in the centre of the simulation box and add an ICM wind via
inflow boundary conditions (and outflow on the opposite side) to simulate RPS.
We include a time-varying (in density and velocity) ICM wind (Tonnesen, 2019),
and to find its parameters, we model a galaxy falling into a massive cluster
($M_{\text{cluster}}=10^{15}M_{\odot}$) following the procedure described in
Bellhouse et al. (2019) from a clustercentric radius of 1.9 Mpc and with an
initial velocity of $1785\;\mathrm{km\,s}^{-1}$. We assume hydrostatic
equilibrium of an isothermal ICM with constant temperature of $T=7.55\times
10^{7}$ K and a beta profile. From Rankine–Hugoniot jump conditions for Mach
number of 3 we find the pre-wind ICM conditions using the ICM wind parameters
at the initial radius.
Since our galaxy is fixed in space, we define a wind angle as the angle
between the wind direction and the galaxy rotation axis, and model three wind
angles: $0^{\circ}$ (a face-on wind that flows along the $z$-direction in our
simulated box, W0), $45^{\circ}$ (W45, in which the wind has equal components
along the $y$\- and $z$-directions), and $90^{\circ}$ (edge-on wind that flows
along the $y$-direction in our simulated box, W90). The wind angle is constant
throughout a simulation. As a control, we simulate a galaxy that does not
undergo RPS (no wind, NW).
We simulate galaxies that include only radiative cooling (denoted as RC) and
those that also include star formation and feedback (denoted as SF), thus
modelling in total eight galaxies: RCNW, RCW0, RCW45, RCW90 and SFNW, SFW0,
SFW45, SFW90.
We evolve our RC galaxies in isolation for 130 Myr, during which the gas cools
down and collapses into a thin disc. The wind starts inflowing through the
lower boundary of the simulation box, and we restart our simulation as four
separate runs: RCNW, RCW0, RCW45, RCW90. The wind reaches the galaxies after
70 Myr, taking 200 Myr in total for the galaxies to evolve before the onset of
RP. This amount of time allowed the majority of the gas in the disc
(particularly within the stripping radius of 15 kpc, discussed below) to
fragment into clouds. We repeat the procedure for the SF galaxies, allowing
them to evolve for 230 Myr in isolation in order to stabilise the SFR, such
that the variation of the SFR on a 5 Myr time-scale decreases to 5 per cent.
In total, the four galaxies (SFNW, SFW0, SFW45, SFW90) evolve for 300 Myr
before the beginning of RPS.
The RC simulations run for a total of 900 Myr and the SF simulations for a
full Gyr. We used 3.0M node-hours to simulate the four RC galaxies and 3.7M
node-hours for the four SF galaxies.
## 3 Stripping rate
Figure 1: Stripping rate of gas with metallicity $Z>0.35Z_{\odot}$ in a galaxy
disc (defined with a radius 30 kpc and a height of $\pm 2$ kpc from the disc
plane) as a function of time. From top to bottom, rows show data for different
wind angles, where W0 is a face-on stripped galaxy and W90 is stripped edge-
on. Each panel shows $\dot{M}$ for SF (solid) and RC (dashed) simulations.
Overplotted in black is the SFR for each galaxy normalised by the SFR in SFNW
at the same time step (right $y$-axis), with a solid (dotted) line plotting
the ratio above (below) one. Notice that for W0 and W45 (top and middle
panels) the left $y$-axis range is the same, while for W90 (bottom panel) it
is smaller. The right $y$-axis range is the same in all the panels.
We start by analysing differences in the gas distribution in the galactic
disc, a region defined as having a radius of $R=30$ kpc and $h=\pm 2$ kpc
height from the plane of the disc (we note that we also used a disc with
$h=\pm 5$ kpc height and found similar results). To quantify the effect that
the feedback has on the process of stripping we calculate the stripping rate:
$\dot{M}=\frac{\Delta M}{\Delta t}-\text{SFR},$ (1)
where $\Delta t=t_{i}-t_{i-1}=5$ Myr time step, $\Delta M=M_{i}-M_{i-1}$ is
mass difference of gas with metallicity $Z>0.35Z_{\odot}$. The metallicity cut
allows us to exclude the pure ICM, and the choice of this metallicity
threshold does not have a significant impact on our results, as we tested up
to a much more strict metallicity criterion of $0.7Z_{\odot}$. While equation
1 could be positive in the case of accretion from galaxy fountain flows or the
surrounding ICM, once the ICM wind begins removing gas from the galaxy
(shortly after RPS begins) $\dot{M}$ is always negative. Because we are
examining our simulations while stripping is occurring, we call $\dot{M}$ the
stripping rate. SFR is found at the same time step:
$\text{SFR}=\frac{M_{\text{star}}^{i}}{\Delta t},$ (2)
where $M_{\text{star}}^{i}$ is the total mass of stars born within the last
time step $\Delta t$, which is always 5 Myr. For RC galaxies, SFR is defined
to be zero. By subtracting SFR from the stripping rate we account for the gas
lost to star formation, thus the differences between stripping rates of RC and
SF galaxies should be due to the effect the feedback plays in RPS.
Figure 2: Top: SF to RC stripping rate (from Fig. 1) ratio as a function of
time, colour-coded by wind angle. The ratio is calculated for $\dot{M}<0$
only, to avoid confusion when gas is accreted at the beginning of the
simulation. The grey line of equal ratio serves to guide the eye. The W90
ratio starts at $\sim 150$ Myr because before that time the gas is not
stripped but accreted onto both of the galaxies. Bottom: Cumulative mass of
stripped gas (accounting for star formation) with metallicity
$Z>0.35Z_{\odot}$ as a function of time, for SF (solid) and RC (dashed)
simulations.
Fig. 1 shows the stripping rate, with each row for a different wind angle (W0
is a face-on stripped galaxy and W90 is stripped edge-on), solid lines
plotting SF galaxies and dashed lines plotting the RC ones. Note that here and
in other plots the $x$-axis (and later any time measurements) is ‘time since
RPS’, where $t=0$ denotes the start of RPS and excludes the 200 and 300 Myr
during which RC and SF galaxies, respectively, were evolving in isolation. We
also confirm that by the start of RPS the SF galaxy has only 5 per cent less
gas mass than the RC one, due to star formation, meaning that any variation in
the stripping process is not a result of a simple mass difference. We overplot
in black the SFR for each galaxy divided by the SFR in SFNW at the same
simulation time, with a solid (dotted) line indicating that the ratio is above
(below) one.
Once we have accounted for the gas loss due to star formation, the stripping
rates of RC and SF galaxies are very similar for any given wind angle. This
might be expected in SFW0 and SFW45 galaxies, since they slowly quench their
star formation after the onset of RPS due to the removal of gas in the
outskirts, following a brief slight enhancement of SFR during the initial disc
compression by the wind. Hence, the global role of feedback in these galaxies
will be decreasing with time, and they would become more and more similar to
their RC counterparts. Moreover, SFW0 and SFW45 (as well as RCW0 and RCW45)
are similar to each other. In Akerman et al. (2023) we show that under RP,
SFW0 and SFW45 evolve rather closely on the galaxy scale in terms of their gas
mass and quenching histories, so their comparable stripping rates (and the
striking difference from SFW90) are no surprise. Note that the small delay in
the peak stripping rates between W0 and W45 galaxies is simply due to the fact
that it takes the angled wind more time to reach the galaxy since it has a
larger distance to travel to the centre of the simulation box. At the peak
stripping rate of W0 and W45 (at $\sim$30–50 Myr) the RC galaxies are stripped
more rapidly than the SF galaxies, indicating that the small SFRs in our W0
and W45 simulations do not enhance stripping. We also note that the W0 maxima
are larger than those in W45, suggesting that more angled winds may strip
galaxies slower (in agreement with previous works such as Roediger & Brüggen,
2006). It is more surprising to see that even in SFW90, where the SFR is
increased by a factor of 2–3 due to RP, the feedback has almost no apparent
effect on the stripping rate. On the contrary, from 350–500 Myr after
stripping has begun RCW90 is stripped faster. We will go into the details of
where RPS removes mass from the galaxies in the next section.
To emphasise the variations in the stripping rates in the top panel of Fig. 2
we plot their ratio, $\dot{M}_{\text{SF}}$ over $\dot{M}_{\text{RC}}$. The
ratio is calculated for $\dot{M}<0$ only (i.e. outflow), to avoid confusion
when gas is accreted at the beginning of the simulation, as the ICM wind
starts mixing with the ISM prior to the gas removal. For W0 and W45 the ratio
is quite unstable at first, but between 100 to 200 Myr the RC galaxies are
stripped systematically faster. The apparent delay in the ratios of W0 and W45
is, again, due to the delay in stripping of the W45 galaxies. Finally, at 200
Myr the W0 galaxies reach the point of ‘stripping equilibrium’
($\dot{M}_{\text{SF}}/\dot{M}_{\text{RC}}\approx 1$) that lasts until 400 Myr
when $\geq 80$ per cent of these galaxies’ original gas has been stripped or
used for star formation. The W45 galaxies during most of this period have a
higher ratio than that of the W0 at identical times, and also higher than one.
However, the stripping rate itself is quite low in all the four galaxies, with
$\dot{M}\approx 7M_{\odot}\text{yr}^{-1}$ on average, which means that even
small differences in the stripping rate will be reflected in the ratio.
The differences in the stripping rates of SFW90 and RCW90 show a different
pattern, with a more clear trend in the ratio. We note that at the earliest
times we do not plot the ratio in the top panel because at the leading edge of
the galaxy, where the disc is hit by the wind, the ICM begins mixing with the
ISM at a rate higher than that of the gas removal on the opposite side of the
galaxy. Because of this, for the first 120 Myr the gas is not stripped but
accreted onto both of the galaxies. In RCW90 the total mass of the gas
accreted over this period equates to $6.5\times 10^{7}M_{\odot}$ or 0.67 per
cent of the galaxy mass at the start of RP, while in SFW90 it is $2\times
10^{8}M_{\odot}$ or 2 per cent of the galaxy mass. This accretion does not
affect the course of stripping. When the gas starts being removed at a steady
rate in both SFW90 and RCW90, the stripping rate is higher in the SF galaxy.
The ratio then starts gradually but steadily decreasing until 400 Myr, after
which it begins rising again.
The bottom panel in Fig. 2 shows the cumulative amount of stripped gas (this
is the integral of Fig. 1), i.e. the gas that was removed from a galaxy by
RPS, not the total gas mass lost by a galaxy (which would also include gas
used in star formation). This distinction is important to keep in mind, since
in the figure SFW0 and SFW45 have seemingly lost 15 per cent less gas than the
respective RC galaxies by the end of the simulation. This is due to the fact
that less gas was available to be stripped from these galaxies, because it was
used for star formation, and due to the difference in the stripping rates
during 30–50 Myr (to be discussed in Section 4). In combination, the two
panels in the figure illustrate that even though the SF-to-RC stripping rate
ratio in the W90 galaxies varies more in the first 500 Myr of RPS, while the
ratio in the W0 and W45 stays, in comparison, more constant, the actual mass
of stripped gas in the W90 galaxies is much lower. As also shown in Fig. 1,
before 200 Myr, the stripping rate in the W0 and W45 galaxies is much higher
than in the W90 galaxy, and even small disagreements between the RC and SF
stripping rates, when accumulated, lead to noticeable differences in the
amount of stripped gas. Conversely, RCW90 and SFW90 lose much less gas,
especially in the initial 100 Myr.
In none of the three galaxies can the behaviour of the stripping rate ratio be
easily tied to the presence of feedback. One could imagine that it would
either assist or impede the RPS. By expelling the gas high above the plane of
the disc (up to 15 kpc, see below), the feedback makes the gas less tightly
gravitationally bound and thus more easily stripped. On the other hand, the
feedback adds thermal and nonthermal pressure to the surrounding gas that in
the cases of face-on stripped W0 and W45 galaxies acts against the ICM wind.
In our simulations, even as the SFR in SFW0 and SFW45 quickly decreases, we
might expect the feedback to have a systemic effect on the stripping rate
prior to complete quenching (quenching only happens after 500 Myr). The role
of feedback is even less clear in the SFW90 galaxy, where the SFR is always
2–3 times higher than in SFNW galaxy, and hence, there is no obvious star-
formation-related reason for the long-term fluctuations of the stripping rate
ratio. To explain the behaviour we see, we now take a closer look at the gas
in their discs.
Figure 3: Mean cell-mass-weighted outward $z$-velocity, $v_{\text{out}}$,
(outward velocities $v_{\text{out}}>0$ and inward velocities
$v_{\text{out}}<0$) of ISM ($Z>0.35Z_{\odot}$) in a thin ($x=\pm 2$ kpc) slice
of the disc for RCNW (top) and SFNW (bottom), taken at 0 Myr. In the bottom
panel the black dashed lines indicate the galaxy boundaries, as we define them
with $z=\pm 2$ kpc.
## 4 What drives the differences between RC and SF galaxies
Here we directly compare the RC and SF simulations in the NW, W0 and W45, and
W90 cases separately. We begin by examining the impact of star formation and
feedback on an isolated galaxy, build a physical interpretation of how it will
affect stripping rates, and then use our understanding to examine the
differences in the stripping rates in detail.
Figure 4: Density projections for ISM ($Z>0.35Z_{\odot}$) in the central 5 kpc
of the disc, within $\pm 2$ kpc of the disc plane for RCNW (top) and SFNW
(bottom) at 0 Myr. While star formation and stellar feedback locally smooth
out the gas distribution in the disc, in RCNW the gas distribution remains
patchy, with holes between the dense clumps.
### 4.1 General impact of stellar feedback
We will start by analysing the impact that star formation and stellar feedback
have on gas in general, using our isolated simulations. To illustrate the
effect of SN feedback on gas flows around our simulated galaxies, in Fig. 3 we
plot the mean cell-mass weighted outward $z$-velocity, $v_{\text{out}}\equiv
v_{z}\text{sgn}(z)$ (outward velocities from the galaxy $v_{\text{out}}>0$ are
red and inward velocities to the galaxy $v_{\text{out}}<0$ are blue), of ISM
($Z>0.35Z_{\odot}$) in a thin ($x=\pm 2$ kpc) edge-on slice of the RCNW (top)
and SFNW (bottom) discs, taken at 0 Myr of RPS (200 Myr total evolution of
RCNW and 300 Myr for SFNW). As we begin the simulation, the initially
homogeneous ISM starts collapsing into a thin disc, as shown in the top panel
of Fig. 3, where the disc is about 400 pc thick. While there is no mechanism
to stop the collapse into dense clumps in RCNW (illustrated also with the blue
inward flows), in SFNW the feedback expels the SN gas from the galaxy via
galactic fountains. The bottom panel in Fig. 3 illustrates these fountains, as
they eject the gas (red) and it later falls back on to the galaxy (blue).
Along with ionising radiation from young stars and winds from evolved massive
stars the fountains act against collapse, and the SFNW galaxy disc becomes
much thicker compared to its RCNW counterpart. Although the outflows can eject
some material beyond 10 kpc from the disc plane, we highlight that there is
very little gas mass in these flows, with only about 0.1 per cent of the total
gaseous disc mass extending above 5 kpc from the disc plane.
In addition to this, star formation and stellar feedback also smooth out the
gas distribution in the disc, while in the RC galaxies there is no mechanism
to inject energy into cold clumps. As a result, the RC gas distribution is
patchy, with low-density holes between the dense clumps that would otherwise
turn into stars. To illustrate this, in Fig. 4 we plot the density projection
of ISM ($Z>0.35Z_{\odot}$) in the central 5 kpc of the disc within $\pm 2$ kpc
of the disc plane for the RCNW (top) and SFNW (bottom) at 0 Myr. We stress,
however, that the feedback is only able to smooth out the ISM locally and does
not cause radial mass redistribution across the whole disc.
Another feature present in RC galaxies is a ‘ring’, which manifests itself at
$\sim$200 Myr as a non-fragmented density distribution (therefore without
high-density clumps) located at $14\text{kpc}<r<20\text{kpc}$. It can be seen
in the top panel of Fig. 3 as a white ‘gap’ where the gas is not moving, and
is a result of a non-uniform collapse of the initial disc. The ring is a
numerical effect present in many galaxy-scale idealised simulations (e.g.
Goldbaum et al., 2015; Behrendt et al., 2015) that disappears with time (by
300 Myr of total simulation time or 0 Myr of RPS) in SF simulations as the gas
distribution becomes more homogeneous and less clumpy due to SN feedback. In
RCNW the ring disappears by 350 Myr of total simulation time, when the whole
galaxy disc collapses into dense clumps. Both the added clumpiness in the
central disc and the more uniform density distribution in the unfragmented
ring play a role in the comparative stripping rates of the RC and SF galaxies.
To summarise, star formation and stellar feedback homogenised the gas
distribution in the galaxy disc and prevented it from collapsing into the
highly dense clouds found in the inner regions of RCNW, with SN fountains
launching the gas as far as 15 kpc above the plane of the galaxy (Fig. 3).
### 4.2 The Importance of the Disc Gas Distribution
Figure 5: Surface density of ISM ($Z>0.35Z_{\odot}$) as a function of
galactocentric radius for RCW0 (top) and SFW0 (bottom) at 10 Myr. Surface
density is found within our galaxy disc boundaries $z\pm 2$ kpc (see the text
for details). The points are coloured by the mean cell-mass-weighted $v_{z}$.
The orange line plots average gas surface density as a function of radius. The
red line represents threshold surface density from the Gunn & Gott (1972)
model below which (hatched area) the gas could be stripped at the current RP.
This line illustrates that at the galaxy outskirts even the highest density
gas will be removed by the ICM wind, while at the centre the gas is protected.
To gain a physical understanding of how the gas distribution could affect a
galaxy’s gas stripping rate, we consider the gas distribution in the disc and
the outside-in nature of RPS, which can be quantified by calculating the
stripping radius at any time. Gunn & Gott (1972) define the stripping radius
as the radius at which the gravitational restoring force per unit area equals
the RP: $P_{\text{ram}}=2\pi G\Sigma_{\text{star}}\Sigma_{\text{gas}}$, where
$\Sigma_{\text{star}}$ is the stellar surface density and
$\Sigma_{\text{gas}}$ is the gas surface density. While other equations can be
used to calculate whether gas will be removed (e.g. McCarthy et al., 2008),
this simple criterion has been shown to match reasonably well with simulations
(Roediger & Brüggen, 2006; Steinhauser et al., 2016; Lee et al., 2020) and is
enough to guide our understanding.
Here, to find the restoring force we measure $\Sigma_{\text{gas}}$ as a
function of radius along the wind direction ($z$-axis), within $z=\pm 2$ kpc
(the galaxy boundaries). We have repeated the calculations for the disc height
of $z\pm 5$ kpc and found qualitatively similar results. We measure the
$\Sigma_{\text{stars}}$ only from the static stellar disc, since compared to
it the surface density of our star particles is negligible. Assuming
instantaneous stripping, this is the radius to which a galaxy would be
stripped at any given RP. In our galaxies, immediately after the onset of RP,
the stripping radius is $\sim 15$ kpc, and due to the increasing RP it drops
down to 10 kpc in the next 100 Myr. Note that in this paper we define the
stripping radius solely by comparing the gas surface density in our simulated
discs to the surface density at which the Gunn & Gott (1972) criterion would
predict stripping. It is not a measure of where the gas is actually stripped
from the disc in our simulation, and indeed the goal of this section is to
predict when and understand why simulated galaxies do not follow the simple
Gunn & Gott (1972) prescription.
The use of the stripping radius to calculate where the ISM will be removed
assumes that gas only varies in surface density as a function of radius, which
is clearly an oversimplification. Fig. 4 shows that in both SF and RC galaxies
gas fragments into dense clumps embedded in a lower density ISM. To understand
the impact of the level of homogeneity in the gas density distribution on gas
removal by RPS, in Fig. 5 we plot surface density of the ISM
($Z>0.35Z_{\odot}$) as a function of galactocentric radius for RCW0 (top) and
SFW0 (bottom) at 10 Myr. In order to use a more spatially-refined density
distribution in the disc, we select the area of $60\times 60\text{ kpc}^{2}$
around the galaxy centre and divide it into $1024\times 1024$ squares (we
confirm that the following results are independent of the resolution). We then
integrate the density in each square over the galaxy height of $z\pm 2$ kpc.
We find qualitatively similar results for the disc height of $z\pm 5$ kpc. In
contrast with finding an average surface density at a given radius, this
method allows us to capture high-density clumps and low-density holes in RCW0
and SFW0, while the choice of surface density instead of the volume/number
density informs us about the stripping using the Gunn & Gott (1972) criterion.
We colour the points by the mean cell-mass-weighted $v_{z}$. The orange line
plots average gas surface density as a function of radius. The red line
represents threshold surface density from the Gunn & Gott (1972) model. This
threshold surface density is found by equating the restoring force to the RP
and calculating for each radius the surface density below which (hatched area)
the gas could be stripped at the current RP. This line illustrates that at the
galaxy outskirts even gas with surface densities well above the average will
be removed by the ICM wind, while at the centre nearly all of the gas is
protected.
In this figure we first notice that the surface density in both the RC and SF
galaxies varies by orders of magnitude at every radius. Also, although the
average value of the surface density is quite similar in both galaxies, the RC
galaxy has a larger range of $\Sigma_{\text{gas}}$ at most radii, except for
in the ring region that we introduced in Section 4.1. This figure also shows
that SFW0 disc is more radially extended ($r>25$ kpc) than the RCW0 disc with
low $\Sigma_{\text{gas}}$ from stellar feedback. We again emphasise that this
is a local effect that does not cause mass redistribution over large radial
scales, nor does feedback cause the disappearance of the ring region that
happens in RCNW at a similar rate compared to SFNW – this naturally happens
over time due to fragmentation from radiative cooling. We also point out that
lower $\Sigma_{\text{gas}}$ tends to have a higher vz.
We can now discuss how the gas inhomogeneity in the discs will influence gas
stripping, starting at the galaxy outskirts. Beyond about 15 kpc, gas must
have a higher $\Sigma_{\text{gas}}$ than the average in order to remain
unstripped according to our criterion. In Fig. 5 we see that at the outermost
regions of the galaxy ($\sim$23 kpc), RCW0 has more gas at
$\Sigma_{\text{gas}}$ above the red line. Therefore, at early times, when the
outermost gas is removed, we would expect SFW0 to be stripped more quickly.
In the ring region of RCW0, at 14–20 kpc, the gas has not yet fragmented and
therefore largely lies below the red line. In the same region in SFW0, there
is gas at higher $\Sigma_{\text{gas}}$ that is likely to survive longer.
Therefore, in the outer region, we predict that gas homogeneity increases gas
stripping as more gas falls below the red line.
In the inner region we might expect the gas homogeneity to assist RPS as it
does in the disc outskirts, but Fig. 5 reveals a different picture. Within the
inner 10 kpc gas inhomogeneity leads to some regions of the disc having very
low $\Sigma_{\text{gas}}$. While the mass in these regions is low, the
extremely low $\Sigma_{\text{gas}}$ ISM in RCW0 has high velocity in the wind
direction ($v_{z}$). In fact, in the inner disc, the gas below the red line in
RCW0 has a higher $v_{z}$ than the same $\Sigma_{\text{gas}}$ ISM in SFW0.
Although there are similar amounts of mass below the gas surface density
stripping threshold, the higher velocities in the RC ISM result in higher mass
fluxes.
We highlight that this is a snapshot of the ISM $\Sigma_{\text{gas}}$
distribution, and this distribution will vary with time over the simulation.
The most obvious differences will be a lack of low $\Sigma_{\text{gas}}$ ISM
at large radii as that gas is removed, and that as RP increases, the red line
will shift to higher $\Sigma_{\text{gas}}$ values. However, all of the general
trends we have outlined here persist over time: stellar feedback continues to
allow for low $\Sigma_{\text{gas}}$ gas just beyond the bulk of the disc,
dense clumps survive at large radius, and low-density regions are accelerated
to high velocities. In the next subsections we will see how this plays out in
our simulations.
Figure 6: For the W0 galaxies at 10 Myr of RPS, from top to bottom: difference
between the azimuthally-averaged distributions of cell-mass-weighed velocity
$v_{z}$, vertical mass flux, and total cell mass in RCW0 and SFW0 galaxies, as
a function of cylindrical radius $r$ and $z$. We calculate the data only for
the stripped gas with $v_{z}>0$ to avoid confusion when plotting the
difference. The solid cyan line separates downwind and upwind halves of the
disc (wind moves upward in the $z$-direction), while the dashed cyan line
denotes the the $z=2$ kpc height at which we define the galactic disc
boundaries. The green line outlines the gas distribution in RCW0 to illustrate
that it is more compact compared to SFW0 at the same time step.
### 4.3 Face-on stripped galaxies
Now that we have built intuition for how we expect stripping to be influenced
by the ISM gas distribution, we will begin looking at RPS-galaxies, starting
with W0, as it is stripped the most rapidly. We measure the mass flux of gas
($Z>0.35Z_{\odot}$) in the galaxy disc ($R=30$ kpc and $h=\pm 2$ kpc) in RCW0
and SFW0 galaxies. Then, to understand the evolution of the SF-to-RC stripping
rate ratio we will map the difference in the mass flux distribution between
the two galaxies. Since in this case RPS is axially-symmetric, we can make use
of cylindrical coordinates, focusing only on variations as a function of
radius $r$ and $z$-height. Here, we select the data only for the cells with
upward outflows $v_{z}>0$ to avoid confusion when plotting the difference
between the two galaxies (see below), but for completeness, in Appendix C we
show maps for cells with downward inflows $v_{z}<0$. The mass flux is:
$\dot{M}=\sum_{i}\frac{m_{i}v_{z,i}}{dL},$ (3)
where $m_{i}$ is mass of the $i$-th cell, $v_{z,i}$ is the $z$ velocity
component and $dL=200$ pc is the height of a bin in Fig. 6, and the sum is
extended to all cells in a given $(r,z)$ bin.
We can now spatially compare the mass flux between the RC and SF galaxies. In
Fig. 6 we plot for the two W0 galaxies at 10 Myr from top to bottom:
difference between the $(r,z)$ maps of cell-mass weighted azimuthally-averaged
velocity $v_{z}$, vertical mass flux, and total cell mass in RCW0 and SFW0
galaxies, as a function of cylindrical radius $r$ and cylindrical $z$. The bin
size in each of the panels is $200\times 200\text{ pc}^{2}$. The goal of this
plot is to map which of the two galaxies, SFW0 or RCW0, has more mass flux in
the wind direction (central panel), and which of the mass flux components
contributes more to it. Each panel shows RC $-$ SF, so orange indicates higher
RC values, and purple indicates higher SF values. With the solid cyan line
($z=0$) we separate downwind and upwind halves of the disc, while the dashed
cyan line denotes the $z=2$ kpc height at which we define the galactic disc
boundaries. We also overplot a green contour that outlines the gas
distribution in RCW0 as it would appear if in this figure we plotted real RCW0
values222We note that these green contours are only denoting the edge of the
distribution of disc gas, defined as gas with a metallicity above 0.35. There
is gas in the cells beyond this contour, but it is ICM gas. instead of the
difference. This illustrates that RCW0 is more compact compared to SFW0 at the
same time step, since all gas outside the contour is only from the SFW0 disc.
As mentioned earlier, the SF-to-RC stripping rate ratio for W0 galaxies
changes periodically in the first 200 Myr. Here we look in detail at these
different periods and what causes the change, always referring to the top
panel of Fig. 2. At the beginning, up to 25 Myr, SFW0 is stripped faster,
illustrated in the middle panel of Fig. 6 with purple outflows beyond $r>10$
kpc. At $r<10$ kpc, however, the difference between the RC and SF mass fluxes
is quite small but tends towards higher mass flux in the RC simulation.
Interestingly, the inner region of the SF galaxy has stronger outflows from
feedback (Fig. 3), so those feedback-driven outflows do not seem to be
dominating the mass flux. Close to the disc plane we see enhanced mass flux in
the RC galaxy, which will result in larger RC stripping rates at later times.
Since stellar feedback expels gas in all directions, the gas disc of SFW0 is
much thicker and more radially extended (by $\sim 2$ kpc) than in RCW0, as
shown in Fig. 3 and in the bottom panel of Fig. 6. This added thickness does
not change the gas surface density, and therefore does not affect the
stripping rate. The feedback does not directly remove gas or assist RP in
doing so, as it would be with the gas getting an additional ‘kick’ to its
vertical velocity by galactic fountains, but, rather, enhanced stripping is a
result of star formation and stellar feedback homogenising the ISM (as
illustrated in Fig. 3 and 4), and radially extending the gas disc (as seen in
Fig. 5). We note that the homogeneity and the added radial extent persist in
the SF runs because star formation and feedback continually affect the ISM.
Although most of the gas is lost from the outskirts and the gas disc generally
shrinks as the RP continuously increases, these figures can also be used to
explain the difference in stripping rates in the inner disc. In the inner 10
kpc the SFW0 gas seems to be protected from RPS, as the mass-flux distribution
is dominated by the high-velocity gas from RCW0 (orange area in the top panel
of Fig. 6). The ICM wind can easily flow through the underdense regions
created by continuous cooling and gas fragmentation (Fig. 4 and 5). As it does
so, it rapidly accelerates the low density ISM, and it is this gas that can be
stripped even from the galaxy centre that we see in orange in the middle panel
of Fig. 6.
In addition to the enhanced mass flux from the SFW0 galaxy crossing $z=2$ kpc,
at 10 Myr we can see enhanced mass flux formed in the RCW0 galaxy in a narrow
strip directly above the $z=0$ line (Fig. 6, middle and bottom panels, in
orange). By 35 Myr, this stripped slice evolves into a whole arch as shown in
Fig. 7. This arch is the reason why the stripping rates are higher in RCW0
during 30–50 Myr and why at that time there is a peak in the top panel of Fig.
1. This peak is also present in SFW0 but is due to the gas removal at large
radii. The enhanced stripping of the RCW0 ring once again illustrates that, as
in the outskirts of SFW0, homogeneity of gas leads to more gas below the
stripping surface density (as there is an absence of dense clumps that are
present in this region in SFW0), which makes it easier to be stripped.
In the middle panel of Fig. 7 at $12\text{ kpc}<r<20\text{ kpc}$ we can see
SFW0 gas that begins to move (the dark purple area under the orange arch).
Later, during 55–85 Myr, as SFW0 is stripped outside-in, the same gas is
leaving the galaxy. Since by this time there is no gas left in RCW0 at these
radii (the gas arch seen here has passed beyond $z=2$ kpc), in this period the
stripping rate is stronger in the SF galaxy.
Figure 7: Same as Fig. 6 but for 35 Myr of RPS. At this time, the stripping
rate is dominated by gas removed from a more uniform, unfragmented low-density
region between $14\text{kpc}<r<20\text{kpc}$ in RCW0 galaxy.
The final change to the SF-to-RC stripping rate ratio happens at 90 Myr. With
a smooth ISM distribution, a galaxy subject to face-on RPS will be gradually
reduced in size up to its stripping radius. This largely occurs in SFW0, which
has a constant peak gas surface density at all radii as shown in Fig. 5.
Therefore, increasing RP smoothly strips it from the outside-in. In RCW0,
however, while the loosely-bound gas was easily stripped from the former ring
area, the dense clumps of gas in the outskirts at $r>20$ kpc remain until 90
Myr. We can see these dense clouds surviving at 35 Myr in the bottom panel of
Fig. 7 by the orange region of excess mass and in Fig. 5 as the high surface
density gas at $\sim$23 kpc. Hence, from 90 to 170 Myr while SFW0 has already
been reduced to 12 kpc radius, RCW0 once again dominates the stripping rate as
the remaining gas clumps in the outskirts are being quickly stripped in
addition to the gas at $\sim$12 kpc that is being stripped from both galaxies.
During 170–400 Myr we see similar stripping rates in the RC and SF galaxies.
By this point both the W0 galaxies have been reduced to the same stripping
radius (using the Gunn & Gott (1972) gas surface density definition), around
12 kpc, with the gas distribution in SFW0 still being 1–2 kpc more extended
than in RCW0. Because feedback continues to push some of the ISM to larger
radii, the SFW0 galaxy is stripped faster in the outskirts. On the other hand,
in the galaxy centre RCW0 is stripped faster due to the low surface density
regions in the gas distribution, making the overall picture of RPS similar to
what we saw during the first 30 Myr. This also agrees well with our intuition
derived from the gas surface density distributions seen in Fig. 5. Therefore,
feedback both enhances outer stripping by extending the ISM and depresses
inner stripping by homogenising the gas surface density. At this time in our
simulations these two effects balance each other out, and globally the
galaxies have almost identical stripping rates (Fig. 2, top panel).
From 400 Myr onward the SFW0 galaxy starts losing gas more and more slowly.
During this time its calculated stripping radius and measured gas disc radius
are both smaller than that of RCW0. As shown for earlier times by Fig. 6 and 7
the mass flux is determined mostly by the mass distribution: where there is
more mass, the mass flux is stronger, even if that gas is moving slower. By
400 Myr, SFW0 has lost gas not only to RPS, but also to star formation, which
is especially true in the inner 5 kpc, where most of the stars are formed.
This means that in the later part of the simulation there is simply very
little gas left in the centre of SFW0, and therefore RCW0 starts being
stripped faster.
As evident from the stripping rates in Fig. 1 and, consequently, from the mass
of the stripped gas in the bottom panel of Fig. 2, the W45 galaxies follow a
very similar evolution to the W0; something that was also shown by Roediger &
Brüggen (2006) and Akerman et al. (2023). It is not surprising then that the
SF-to-RC stripping rate ratio of W45 galaxies also follows the one for the W0
(Fig. 2, top panel). The main difference between the two ratios is due to the
time delay that, as explained above, stems from the fact that it takes an
inclined wind more time to reach a galaxy and begin the stripping process.
Another difference shown in the top panel of Fig. 2 is that during two
relatively short periods of time, 225–250 Myr and 305–375 Myr, SFW45 is
stripped faster than RCW45. This is another effect of the continuous collapse
of gas, as a result of which dense clumps form. These clumps are then pushed
in the shadow of the disc, where they are protected from the ICM wind. The
stripping rate in RCW45 drops and the SF-to-RC stripping rate ratio goes up.
In conclusion, the added inhomogeneities in RC galaxies result in more gas
both above and below the stripping surface density (Fig. 5). Dense gas is
harder to strip, as evidenced by the lower $v_{z}$ at high surface density. In
the galaxy outskirts, ISM inhomogeneities are able to enhance gas surface
density to the point where it cannot be stripped, slowing down the global
stripping rate. We see this in the delayed stripping of dense gas beyond 20
kpc in the RC galaxies relative to SF galaxies, as well as in the earlier
stripping of the homogeneous gas ring at 14–20 kpc in the RC galaxies. In the
galaxy centre, ISM inhomogeneities can result in regions with very low gas
surface density. These regions are stripped very quickly, which leads to
faster stripping rates in the central regions of RC galaxies.
Figure 8: For the W90 galaxies, $v_{y}$ mass flux of ISM ($Z>0.35Z_{\odot}$)
as a function of $z$ measured in $29.5\text{kpc}<r<30.5\text{kpc}$ at (from
left to right) 150, 400 and 550 Myr of RPS for SF (solid) and RC (dashed)
simulations. We select the data for the downwind part, $y>0$, and for the
stripped gas with $v_{y}>0$.
In this section we have focused on the early times when the RC and SF
stripping rates differ in order to understand the physical reason for the
distinct rates. However, generally, the lower stripping rates at the outskirts
of the RC galaxies balance the higher stripping rates in their centres,
leading to similar stripping rates in the SF and RC simulations.
### 4.4 Edge-on stripped galaxies
After having discussed the detailed evolution of the face-on stripped
galaxies, we examine the edge-on ones, RCW90 and SFW90. Since in this case the
feedback does not act parallel to the direction of the wind, its effect is
harder to predict. Still, we could expect RP from the edge-on ICM wind to
easily remove the gas expelled above (and below) the disc by the galaxy
fountains. Surprisingly, while this effect indeed takes place, it does not
dominate the stripping rates in SFW90.
To illustrate this, in Fig. 8 we show the $v_{y}$ mass flux (similar to eq. 3)
measured in $29.5\text{kpc}<r<30.5\text{kpc}$ as a function of $z$. We remind
the reader that in the edge-on case (W90) the wind moves along the
$y$-direction. Solid (dashed) lines plot SFW90 (RCW90) at three different
times (from left to right): 150 Myr, 400 Myr, and 550 Myr. We chose these time
steps to have three different SF-to-RC stripping rate ratios in the top panel
of Fig. 2, where at 150 Myr SFW90 is stripped faster, at 400 Myr RCW90
dominates and at 550 Myr the galaxies lose their gas at a similar rate. We
also select the data only for the downwind part, $y>0$, and for $v_{y}>0$
only. Thus, shown here is that while (for the most part) outside of the disc
plane there is more mass flux in SFW90 than in RCW90, the bulk of the
stripping at all times is in the galaxy plane.
Moreover, shown in the bottom panel of Fig. 1 is SFR of SFW90, enhanced
compared to SFNW on the whole duration of the simulation. This means that the
scenario outlined above of SN outflows being swept away by the ICM wind plays
out during all 700 Myr. Globally, evolution of the SF-to-RC stripping rate
ratio (Fig. 2, top panel) can be divided into two periods: from 120 to 400 Myr
the ratio goes down, indicating a slow but steady decrease in the relative
outflow rates of SFW90 compared to RCW90; from 400 to 700 Myr it goes up
again, meaning that now the SF galaxy gets stripped systematically faster.
This means that despite the ongoing stripping of the expelled feedback gas,
RCW90 manages to overcome the effect, and is even stripped faster than SFW90
from 330 to 530 Myr (as also evident in the middle panel of Fig. 8).
To understand the change in the stripping rate, we look at the gas
($Z>0.35Z_{\odot}$) distribution in the disc plane ($z\pm 2$ kpc), shown in
Fig. 9 by an $xy$ density projection for 250 Myr. Again, we calculate the data
only for the gas with $v_{y}>0$. This leads to a lack of gas at positive
$x$-values in the figure, where galaxy rotation leads to negative $v_{y}$. In
both panels, the dashed pink line denotes the outer boundary of $r=30$ kpc
with which we define the galactic disc. The yellow line outlines the gas
distribution in SFW90 to illustrate that it is more compact compared to RCW90
at the same time step. Also note that unlike in the W0 runs, where the gas
needs to travel up to $z=2$ kpc to be considered stripped, in W90 this path is
increased to $r=30$ kpc. Any change in the gas distribution in the galaxy disc
will affect the stripping rate only about 100 Myr later, since the gas needs
to travel 30 kpc on average to the downwind edge. This needs to be kept in
mind when analysing the plots, since differences in the immediate gas
distribution of RCW90 and SFW90 may not reflect the current SF-to-RC stripping
rate ratio.
Figure 9: Density (cell-volume weighted) distribution of gas ($z\pm 2$ kpc) in
RCW90 (left) and SFW90 (right) at 250 Myr. We calculate the data only for the
stripped gas with $v_{y}>0$. The dashed pink line denotes the upper boundary
of $r=30$ kpc with which we define the galactic disc. The yellow line outlines
the gas distribution in SFW90 to illustrate that it is more compact compared
to RCW90 at the same time step.
The gas distribution in RCW90 and SFW90 reveals that, just like in the face-on
stripped galaxies, the main factor contributing to the differences between SF
and RC galaxies is not direct removal of gas from the disc plane due to
feedback (such as galaxy fountains) but the homogenised gas distribution in
the disc. When hit by an edge-on ICM wind, the dense clouds are hard to
remove; at the same time, they shield the easily-stripped low-density gas.
Compared to this, the more homogeneous gas in SFW90 is being easily pushed as
evident in the top panel of Fig. 2. This is reflected in the more extended
distribution of gas in RCW90 compared to SFW90 – the higher density clouds in
RCW90 have been less affected by the ICM wind and therefore less gas has been
removed.
However, while the lower density ISM in SFW90 has been stripped more quickly
than the higher density clouds in RCW90, the ISM is not simply directly
stripped from the galaxy. As described above, the gas has a long way to travel
before being considered ‘lost’ by the galaxy (the longest path from the upwind
to the downwind side is $\sim 55$ kpc). On this course, the combination of
wind and galaxy rotation makes gas from different galactic radii collide with
each other to create a dense ‘arch’ (right panel of Fig. 9, at the leading
edge of the disc in the bottom left quadrant). At this point, the SFW90 galaxy
is also substantially smaller than RCW90, as outlined by the yellow line. As
the arch forms, its dense gas becomes harder and harder to strip, at which
point the SF-to-RC stripping rate ratio crosses the equality line. In the
bottom panel of Fig. 1 the decreasing $\dot{M}$ after 330 Myr indicates that
SFW90 runs out of easily stripped gas at this point. The remaining gas from
the arch is pushed into a big dense knot on the wind-facing side of the disc
about 10 kpc from the galaxy centre. At this point the stripping rate in SFW90
drops, while the RC galaxy maintains its stripping rate by removing the low-
density gas.
The steady pushing of gas to the galaxy centre also proceeds in RCW90, but
more slowly since the dense clumps can only be pushed by higher RP (which
occurs later in the simulations). By 400 Myr RCW90 repeats the fate of SFW90:
gas forms an even denser (compared to the SF simulation) knot in the galaxy
centre, and after that the stripping rate starts to drop (Fig. 1, bottom
panel).
In the next $\sim 150$ Myr, SFW90 slowly overtakes RCW90 in the stripping rate
(Fig. 2, top panel). The dense central knot will inevitably be destroyed by
the increasing RP (as it starts to at $\sim$550 Myr), and its gas stripped.
From $\sim 550$ Myr onward the two galaxies are stripped at about the same
rate.
In conclusion, the main differences between the stripping process in the edge-
on stripped galaxies, just like in the face-on ones, stem from the fact that
in the SF galaxies the gas is homogenised and thus more easy to push at first,
while in the RC galaxies it is patchy with dense clumps that are comparatively
harder to move. This however does not simply mean that SFW90 is always
stripped faster, because the edge-on RP can compress gas to higher densities
as well as directly remove gas, and once the high density region in the SF
galaxy matches that in the RC galaxy, stripping occurs at the same rate.
## 5 Discussion
Our simulations show that the stripping rates of star-forming and radiative-
cooling-only galaxies do not differ significantly, and that on galactic scales
the stellar feedback does not directly assist or impede RPS by driving
significant gas mass outside the disc plane (Fig. 1). While there are not many
papers that study the effect of stellar feedback on the process of stripping,
the ones that do, Bahé & McCarthy (2015) and Kulier et al. (2023), show
results different from ours. Both of these works explore cosmological smoothed
particle hydrodynamic simulations, the former with Galaxies-Intergalactic
Medium Interaction Calculations (GIMIC) and the latter with EAGLE data.
Bahé & McCarthy (2015) reveal two things. Firstly, the fraction of gas that
can be stripped from a galaxy declines with galaxy mass, while the fraction
that can be removed (but still remain bound to the galaxy halo) via galaxy
fountains, on the contrary, increases. If we consider their galaxies that are
the most similar to ours, with stellar masses $M_{\text{star}}\approx
10^{10.75}M_{\odot}$ in massive clusters with
$M_{\text{cluster}}=10^{15.2}M_{\odot}$, then it appears that RPS accounts for
60 per cent of all removed gas, while the rest is due to galactic fountains.
Secondly, they also show that the stellar feedback implemented in GIMIC (and
EAGLE) gave a prior ‘kick’ to a significant part of the stripped gas that
later allowed RPS to quickly remove it. Still, for the most massive galaxies
in the most massive clusters the amount of stripped gas when accounting for
the ‘kick’ is within the boundaries predicted by Gunn & Gott (1972) model.
Kulier et al. (2023) come to a similar conclusion, where they measure the
‘excess’ gas that is stripped due to the feedback and reveal that the amount
increases with RP. This indicates a synergy between RP and stellar feedback in
removing galactic gas.
These findings are, perhaps, similar to our own, but rather than the direct
effect of SN feedback, SFW0 is stripped faster at the galaxy outskirts due to
the feedback homogenising and radially extending gas in the galaxy disc, thus
lowering its maximum surface density. However, we also show that in a direct
comparison with a non-star-forming galaxy this effect proves to be
insignificant. It is worth emphasising again that both Bahé & McCarthy (2015)
and Kulier et al. (2023) use cosmological smoothed-particle hydrodynamics
simulations. By their nature such simulations cannot have high resolution and
sophisticated feedback physics of galaxy-scale adaptive-mesh-refinement
simulations. Due to these necessary technical limitations the effect of
stellar feedback might be overestimated. For example, in EAGLE the feedback is
ejective, with galactic fountains launching gas to large distances (Mitchell
et al., 2020); it is also known to create excessively large ‘holes’ in the ISM
(Bahé et al., 2016). In this light, the role of the homogenised ISM density
distribution in RPS is likely underestimated in these simulations.
Bahé & McCarthy (2015) may give us insight into the possible galaxy mass
dependence of our result. They find that while the highest-mass galaxies seem
to be the least affected by the stellar feedback when it comes to stripping,
the lower-mass galaxies $M_{\text{star}}<10^{10}M_{\odot}$ lose much more gas
than estimated by Gunn & Gott (1972). Muratov et al. (2015) in a cosmological
zoom-in simulation FIRE measured the mass loading factor (mass carried out by
stellar feedback per mass of newly formed stars) and found a decreasing trend
with galaxy stellar mass, although their galaxies do not reach the stellar
mass of our simulated galaxy. This means that low-mass galaxies lose more gas
via galaxy fountains, in a seeming agreement with Bahé & McCarthy (2015). In
this case, it is possible that stellar feedback would assist the gas removal
via RPS of low-mass galaxies. We stress that our specific result that SN
feedback has little effect on the stripping rate of galaxies may be quite mass
dependent. However, our result that the homogeneity of the ISM will have
important consequences on the radial distribution of stripping is general.
Another important note about this work is that we did not test a range of
feedback models, in which we would expect feedback implementations that
strongly eject mass to large distances from the disc to have a bigger effect
on the stripping rate of galaxies. However, we can compare our results with
other works studying SN feedback. Kim et al. (2020) conducted local
simulations of several regions in a Milky-Way-like galaxy with a resolution of
2 to 8 pc. Similar to Li & Bryan (2020) they find that the hot ($T>10^{6}$ K)
galactic outflows are much more energetic than the cool ($T<10^{4}$ K) ones,
which, conversely, are responsible for driving the mass out of a galaxy. We
can compare mass flux from different regions in SFNW that correspond to the
regions in Kim et al. (2020). We find that at the radius where our
$\Sigma_{\text{gas}}$ and $\Sigma_{\text{star}}$ are similar to those in Kim
et al. (2020) (R8), our mass fluxes are also comparable to theirs. This
therefore gives us confidence that the scale of our mass fluxes is reasonable,
and our galaxy should not have dramatically different mass fluxes that would
change our main results.
We also reiterate that these are purely hydrodynamical simulations, so we do
not include magnetic fields or cosmic rays. Magnetic fields have been shown in
simulations to broaden the density PDF (Burkhart et al., 2015), especially for
low-Mach number turbulence (Burkhart et al., 2009), and may decrease the SFR
(Federrath & Klessen, 2012; Federrath, 2015; Wibking & Krumholz, 2023). There
is also some observational evidence that enhanced magnetic fields may reduce
gas fragmentation (e.g. Añez-López et al., 2020). Simulations also indicate
that cosmic rays smooth the distribution of gas density in the ISM and the
circumgalactic medium (Butsky & Quinn, 2018; Butsky et al., 2020; Farber et
al., 2022). The impact of these sources of non-thermal pressure on RPS could
be an interesting avenue for future work.
## 6 Conclusions
In this paper we present a suite of wind-tunnel galaxy-scale simulations to
determine the role of star formation and stellar feedback in stripping a
galaxy of its gas via RP. The initial conditions are the same in each
simulation, and model a galaxy slightly more massive than the Milky Way
($M_{\text{star}}=10^{11}M_{\odot}$) falling into a massive cluster
($M_{\text{cluster}}=10^{15}M_{\odot}$). We have two sets of simulations:
radiative-cooling-only (denoted as RC) and with star formation and stellar
feedback (SF). In each of the two sets there are four galaxies, of which one
is not undergoing RPS (no wind, NW). The other three galaxies are on their
first infall into a cluster and have different galaxy-wind impact angle: face-
on (W0), edge-on (W90) and $45^{\circ}$ (W45). We compare the stripping
histories of a RC and SF galaxy at the same impact angle and find that:
1. 1.
Star formation and stellar feedback vertically puff up the galaxy disc,
expelling galactic fountains up to 15 kpc, and homogenise the ISM density
distribution. In RC galaxies, the gas continuously cools down and fragments
into dense clumps (Fig. 3 and 4), resulting in a broader distribution of gas
densities in the ISM. In addition, since the fountains eject gas in all
directions, including radially, at 0 Myr (the time when the stripping begins)
the radius of a SF galaxy is $\sim 2$ kpc bigger than that of a RC one.
2. 2.
Globally, the stripping rates are quite similar for RC and SF galaxies of the
same impact angle (Fig. 1), although since part of the gas is used up for star
formation, the SF galaxies lose less total gas through RPS (Fig. 2, bottom
panel). Comparing how the stripping rates change relative to each other (Fig.
2, top panel) reveals that the relative stripping rates of the SF and RC
galaxies invert multiple times over the course of the simulations.
3. 3.
The differences in the stripping rates of the RC and SF galaxies are related
to the differences in the density distributions in the galaxy discs. This is
because a large distribution in gas density results in a large distribution in
gas surface density (Fig. 5). While the average surface density is not
affected, more inhomogeniety in the ISM of RC galaxies results in more gas
both above and below the stripping surface density at which the restoring
force equals the RP. Gas that falls below the stripping surface density may be
more easily removed, while the gas above it may survive unstripped.
4. 4.
In the outskirts of W0 and W45 where the average surface density is below the
stripping surface density, inhomogeneity makes stripping slower by allowing
dense clouds to survive (Fig. 5 and 6).
5. 5.
At inner radii where the average surface density is higher than the stripping
surface density, inhomogeneity leads to more gas at lower surface densities,
allowing more stripping and faster acceleration of low density gas. We see
this in the enhanced stripping rates in RC in the central regions (Fig. 5, 6
and 7).
6. 6.
In the edge-on stripped SFW90, low-density galaxy fountain gas is stripped
faster by the ICM wind (Fig. 8). However, this has a very small effect on the
overall stripping rate, since the bulk of the gas is stripped in the galaxy
plane. Instead, we find that the homogenised gas in SFW90 is easier to push at
the beginning of RPS, thus allowing for the galaxy to be stripped faster (Fig.
9). Once this lower density gas has been removed or shifted towards the centre
of the galaxy (550 Myr), both RCW90 and SFW90 have almost identical stripping
rates.
We find that regardless of the wind impact angle stellar feedback has no
direct effect on the stripping process, such as giving an additional velocity
kick to the stripped gas. Instead, the indirect effect of homogenising the gas
is what makes the difference between star-forming and radiative-cooling only
galaxies.
An important insight we have gained from this comparison is that correctly
modelling the small-scale distribution of gas in a disc undergoing RPS is
important for understanding how quickly it will be stripped. We argue that
cosmological simulations, which may find that feedback can enhance gas removal
by RPS due to the strong ejection of gas particles to large distances from the
galaxy plane (e.g. Bahé & McCarthy, 2015; Kulier et al., 2023), cannot be used
for detailed comparisons with RPS galaxy populations because they do not
resolve the density range or distribution in the discs of satellite galaxies.
We find that a more smooth gas distribution will result in more (compared to
clumpy and underdense gas in RC simulation) stripping in regions with low
average restoring force versus RP, and less stripping in regions with high
average restoring force versus RP.
## Acknowledgements
This project has received funding from the European Research Council (ERC)
under the European Union’s Horizon 2020 research and innovation programme
(grant agreement No. 833824). The simulations were performed on the Frontera
supercomputer operated by the Texas Advanced Computing Center (TACC) with LRAC
allocation AST20007 and on the Stampede2 supercomputer operated by the Texas
Advanced Computing Center (TACC) through allocation number TG-MCA06N030. B. V.
acknowledges the Italian PRIN-Miur 2017 (PI A. Cimatti). We use the yt python
package (Turk et al., 2011) for data analysis and visualisation.
## Data Availability
The data underlying this article will be shared on reasonable request to the
corresponding author.
## References
* Añez-López et al. (2020) Añez-López N., et al., 2020, A&A, 644, A52
* Akerman et al. (2023) Akerman N., Tonnesen S., Poggianti B. M., Smith R., Marasco A., 2023, ApJ, 948, 18
* Bahé & McCarthy (2015) Bahé Y. M., McCarthy I. G., 2015, MNRAS, 447, 969
* Bahé et al. (2016) Bahé Y. M., et al., 2016, MNRAS, 456, 1115
* Behrendt et al. (2015) Behrendt M., Burkert A., Schartmann M., 2015, MNRAS, 448, 1007
* Bekki (2014) Bekki K., 2014, MNRAS, 438, 444
* Bellhouse et al. (2017) Bellhouse C., et al., 2017, ApJ, 844, 49
* Bellhouse et al. (2019) Bellhouse C., et al., 2019, MNRAS, 485, 1157
* Boselli et al. (2014) Boselli A., Cortese L., Boquien M., Boissier S., Catinella B., Gavazzi G., Lagos C., Saintonge A., 2014, A&A, 564, A67
* Bryan et al. (2014) Bryan G. L., et al., 2014, ApJS, 211, 19
* Burkert (1995) Burkert A., 1995, ApJ, 447, L25
* Burkhart et al. (2009) Burkhart B., Falceta-Gonçalves D., Kowal G., Lazarian A., 2009, ApJ, 693, 250
* Burkhart et al. (2015) Burkhart B., Collins D. C., Lazarian A., 2015, ApJ, 808, 48
* Butsky & Quinn (2018) Butsky I. S., Quinn T. R., 2018, ApJ, 868, 108
* Butsky et al. (2020) Butsky I. S., Fielding D. B., Hayward C. C., Hummels C. B., Quinn T. R., Werk J. K., 2020, ApJ, 903, 77
* Cayatte et al. (1990) Cayatte V., van Gorkom J. H., Balkowski C., Kotanyi C., 1990, AJ, 100, 604
* Chung et al. (2007) Chung A., van Gorkom J. H., Kenney J. D. P., Vollmer B., 2007, ApJ, 659, L115
* Cortese et al. (2012) Cortese L., et al., 2012, A&A, 544, A101
* Cramer et al. (2019) Cramer W. J., Kenney J. D. P., Sun M., Crowl H., Yagi M., Jáchym P., Roediger E., Waldron W., 2019, ApJ, 870, 63
* Crowl & Kenney (2008) Crowl H. H., Kenney J. D. P., 2008, AJ, 136, 1623
* Di Cintio et al. (2014) Di Cintio A., Brook C. B., Dutton A. A., Macciò A. V., Stinson G. S., Knebe A., 2014, MNRAS, 441, 2986
* Dursi & Pfrommer (2008) Dursi L. J., Pfrommer C., 2008, ApJ, 677, 993
* Farber et al. (2022) Farber R. J., Ruszkowski M., Tonnesen S., Holguin F., 2022, MNRAS, 512, 5927
* Federrath (2015) Federrath C., 2015, MNRAS, 450, 4035
* Federrath & Klessen (2012) Federrath C., Klessen R. S., 2012, ApJ, 761, 156
* Fossati et al. (2016) Fossati M., Fumagalli M., Boselli A., Gavazzi G., Sun M., Wilman D. J., 2016, MNRAS, 455, 2028
* Fossati et al. (2018) Fossati M., et al., 2018, A&A, 614, A57
* Fritz et al. (2017) Fritz J., et al., 2017, ApJ, 848, 132
* Fumagalli et al. (2014) Fumagalli M., Fossati M., Hau G. K. T., Gavazzi G., Bower R., Sun M., Boselli A., 2014, MNRAS, 445, 4335
* Gavazzi et al. (2001) Gavazzi G., Boselli A., Mayer L., Iglesias-Paramo J., Vílchez J. M., Carrasco L., 2001, ApJ, 563, L23
* George et al. (2018) George K., et al., 2018, MNRAS, 479, 4126
* Goldbaum et al. (2015) Goldbaum N. J., Krumholz M. R., Forbes J. C., 2015, ApJ, 814, 131
* Goldbaum et al. (2016) Goldbaum N. J., Krumholz M. R., Forbes J. C., 2016, ApJ, 827, 28
* Göller et al. (2023) Göller J., Joshi G., Rohr E., Zinger E., Pillepich A., 2023, arXiv e-prints, p. arXiv:2304.09199
* Gullieuszik et al. (2017) Gullieuszik M., et al., 2017, ApJ, 846, 27
* Gunn & Gott (1972) Gunn J. E., Gott J. R. I., 1972, ApJ, 176, 1
* Haardt & Madau (2012) Haardt F., Madau P., 2012, ApJ, 746, 125
* Jáchym et al. (2007) Jáchym P., Palouš J., Köppen J., Combes F., 2007, A&A, 472, 5
* Jáchym et al. (2017) Jáchym P., et al., 2017, ApJ, 839, 114
* Jáchym et al. (2019) Jáchym P., et al., 2019, ApJ, 883, 145
* Kenney et al. (2004) Kenney J. D. P., van Gorkom J. H., Vollmer B., 2004, AJ, 127, 3361
* Kenney et al. (2014) Kenney J. D. P., Geha M., Jáchym P., Crowl H. H., Dague W., Chung A., van Gorkom J., Vollmer B., 2014, ApJ, 780, 119
* Kenney et al. (2015) Kenney J. D. P., Abramson A., Bravo-Alfaro H., 2015, AJ, 150, 59
* Kim et al. (2020) Kim C.-G., et al., 2020, ApJ, 900, 61
* Koopmann & Kenney (2004) Koopmann R. A., Kenney J. D. P., 2004, ApJ, 613, 851
* Kulier et al. (2023) Kulier A., et al., 2023, arXiv e-prints, p. arXiv:2305.03758
* Lee & Chung (2018) Lee B., Chung A., 2018, ApJ, 866, L10
* Lee et al. (2017) Lee B., et al., 2017, MNRAS, 466, 1382
* Lee et al. (2020) Lee J., Kimm T., Katz H., Rosdahl J., Devriendt J., Slyz A., 2020, ApJ, 905, 31
* Li & Bryan (2020) Li M., Bryan G. L., 2020, ApJ, 890, L30
* Li et al. (2020) Li P., Lelli F., McGaugh S., Schombert J., 2020, ApJS, 247, 31
* McCarthy et al. (2008) McCarthy I. G., Frenk C. S., Font A. S., Lacey C. G., Bower R. G., Mitchell N. L., Balogh M. L., Theuns T., 2008, MNRAS, 383, 593
* Merluzzi et al. (2016) Merluzzi P., Busarello G., Dopita M. A., Haines C. P., Steinhauser D., Bourdin H., Mazzotta P., 2016, MNRAS, 460, 3345
* Mitchell et al. (2020) Mitchell P. D., Schaye J., Bower R. G., Crain R. A., 2020, MNRAS, 494, 3971
* Miyamoto & Nagai (1975) Miyamoto M., Nagai R., 1975, PASJ, 27, 533
* Moretti et al. (2018) Moretti A., et al., 2018, MNRAS, 480, 2508
* Moretti et al. (2020a) Moretti A., et al., 2020a, ApJ, 889, 9
* Moretti et al. (2020b) Moretti A., et al., 2020b, ApJ, 897, L30
* Mori & Burkert (2000) Mori M., Burkert A., 2000, ApJ, 538, 559
* Muratov et al. (2015) Muratov A. L., Kereš D., Faucher-Giguère C.-A., Hopkins P. F., Quataert E., Murray N., 2015, MNRAS, 454, 2691
* Poggianti et al. (2017) Poggianti B. M., et al., 2017, ApJ, 844, 48
* Ramos-Martínez et al. (2018) Ramos-Martínez M., Gómez G. C., Pérez-Villegas Á., 2018, MNRAS, 476, 3781
* Roberts & Parker (2020) Roberts I. D., Parker L. C., 2020, MNRAS, 495, 554
* Roberts et al. (2021) Roberts I. D., et al., 2021, A&A, 650, A111
* Roberts et al. (2022a) Roberts I. D., van Weeren R. J., Timmerman R., Botteon A., Gendron-Marsolais M., Ignesti A., Rottgering H. J. A., 2022a, A&A, 658, A44
* Roberts et al. (2022b) Roberts I. D., et al., 2022b, ApJ, 941, 77
* Rodrigues et al. (2017) Rodrigues D. C., del Popolo A., Marra V., de Oliveira P. L. C., 2017, MNRAS, 470, 2410
* Roediger & Brüggen (2006) Roediger E., Brüggen M., 2006, MNRAS, 369, 567
* Roediger & Brüggen (2007) Roediger E., Brüggen M., 2007, MNRAS, 380, 1399
* Roediger et al. (2014) Roediger E., Bruggen M., Owers M. S., Ebeling H., Sun M., 2014, MNRAS, 443, L114
* Ruggiero & Lima Neto (2017) Ruggiero R., Lima Neto G. B., 2017, MNRAS, 468, 4107
* Ruszkowski et al. (2014) Ruszkowski M., Brüggen M., Lee D., Shin M. S., 2014, ApJ, 784, 75
* Schaye et al. (2015) Schaye J., et al., 2015, MNRAS, 446, 521
* Schulz & Struck (2001) Schulz S., Struck C., 2001, MNRAS, 328, 185
* Smith et al. (2017) Smith B. D., et al., 2017, MNRAS, 466, 2217
* Steinhauser et al. (2016) Steinhauser D., Schindler S., Springel V., 2016, A&A, 591, A51
* Sun et al. (2010) Sun M., Donahue M., Roediger E., Nulsen P. E. J., Voit G. M., Sarazin C., Forman W., Jones C., 2010, ApJ, 708, 946
* Sun et al. (2021) Sun M., et al., 2021, NatAs, 6, 270
* Tasker & Bryan (2006) Tasker E. J., Bryan G. L., 2006, ApJ, 641, 878
* Tonnesen (2019) Tonnesen S., 2019, ApJ, 874, 161
* Tonnesen & Bryan (2009) Tonnesen S., Bryan G. L., 2009, ApJ, 694, 789
* Tonnesen & Bryan (2012) Tonnesen S., Bryan G. L., 2012, MNRAS, 422, 1609
* Tonnesen & Stone (2014) Tonnesen S., Stone J., 2014, ApJ, 795, 148
* Turk et al. (2011) Turk M. J., Smith B. D., Oishi J. S., Skory S., Skillman S. W., Abel T., Norman M. L., 2011, ApJS, 192, 9
* Vulcani et al. (2018) Vulcani B., et al., 2018, ApJ, 866, L25
* Vulcani et al. (2020) Vulcani B., et al., 2020, ApJ, 892, 146
* Wibking & Krumholz (2023) Wibking B. D., Krumholz M. R., 2023, MNRAS, 521, 5972
## Appendix A Refinement criteria and resolution test
Our simulation box has 160 kpc on a side, divided into $128\times 128\times
128$ cells. We use the adaptive approach to include additional 5 levels of
refinement, with each level having twice the resolution of the previous one,
and get a maximum resolution of 39 pc. Our refinement criteria are Jeans
length and baryon mass of $\approx 7500M_{\odot}$. In Fig. 10 we show how
using these criteria we are able to refine the galaxy discs ($R=30$ kpc and
$z\pm 2$ kpc) in RCNW (left) and SFNW (centre) and low-resolution SFNW (right,
see the text below). The cell sizes are colour-coded by cell mass. We select
the data at 10 Myr before RPS begins. Firstly, this plot shows that the chosen
refinement criteria are able to highly resolve the disc, since in all of the
galaxies the bulk of the mass is concentrated at the finest level. Secondly,
compared to RCNW, SFNW has more mass on levels 4 and 5 (156 and 78 pc,
respectively) since in this case the ISM disc within the $z\pm 2$ kpc
boundaries is much thicker (Fig. 3).
Figure 10: Cell sizes $dx$ (we have cubical cells, $dx=dy=dz$) as a function
of cylindrical radius $r$ colour-coded by cell mass in a galaxy disc ($R=30$
kpc and $z\pm 2$ kpc) in RCNW (left) and SFNW (centre) and low-resolution SFNW
(right) at 10 Myr before RPS begins.
Our galaxy, which is the best-resolved region in the simulation box, has a
maximum of $3.33\times 10^{7}$ of cells (of any resolution) in the SFNW run,
while the median number of cells over the 1 Gyr of total simulation time is
$2.86\times 10^{7}$. In the SF wind runs, during the first 100 Myr after the
ICM has just hit the galaxies, the peak cell number value for different wind
angles is $[3.36,3.39]\times 10^{7}$. After that, the number of cells only
continues to decrease with time, as the gas is being stripped and less and
less regions are left to be resolved. In RCNW, as the lack of stellar feedback
allows for the gas to collapse into a thin disc, the number of cells in the
galaxy disc is somewhat smaller with the maximum and median values being
$2.17\times 10^{7}$ and $1.76\times 10^{7}$, respectively.
Here we discuss how resolution may impact our results. We run a short
simulation that includes only 4 levels of refinement, with the finest cell
being 78 pc. In the central and right panels of Fig. 10 we can compare the
resolution distribution in the fiducial and low-resolution runs, respectively.
The bulk of the mass is located at the highest-resolution level, while the
mass distribution on levels 1 to 4 (1248 to 156 pc) is similar in both
simulations.
The main result of this paper, pictured in Fig. 5, is the role of the surface
density distribution in gas stripping. The resolution used, therefore, must
capture both the low- and high-density gas. In Fig. 11 we repeat Fig. 5 for
the fiducial SFNW (top) and low-resolution (bottom) runs 10 Myr before RPS
begins. Note that the colour bar range is different in this figure from Fig.
5, since before the RP, there is no high-velocity gas in the disc (in other
words, we are missing gas that is being rapidly stripped). Overall, the
surface density distribution in both galaxies looks quite similar, with some
extension to lower surface densities in the SFNW low-res run. Meanwhile, the
average surface densities are almost identical, as illustrated in the bottom
panel with the white dashed line that overplots the average in the fiducial
SFNW over the average in the low-resolution run. The similar average values
mean that at any radius the bulk of the mass is distributed at comparable
surface densities. We also note that this means that the calculated truncation
radius (for which the average surface density is used) is resolution-
independent.
Figure 11: Same as Fig. 5, but for the fiducial SFNW (top) and low-resolution
(bottom) runs 10 Myr before RPS begins. Note that the colour bar range is
different in this figure, since before RP there is no high-velocity gas in the
disc. In the bottom panel, the white dashed line overplots average surface
density in the fiducial SFNW over the average in the low-resolution run to
illustrate that they are almost identical.
## Appendix B The effect of star formation on density distribution
In this work we compared radiative-cooling-only simulations to the ones that
include star formation and stellar feedback. Here, we will consider the effect
that star formation alone has on the gas distribution in the galaxy. This will
help us separate the role of star formation from that of stellar feedback in
shaping the ISM.
We will start by outlining the findings of previous works that used, like us,
the Enzo code and considered the impact of star formation and feedback on the
ISM. For example, Tasker & Bryan (2006) run a suite of hydrodynamical
simulations of a Milky-Way-like galaxy. They study the effect that star
formation threshold density and star formation efficiency play in shaping the
Kennicutt-Schmidt law, and run radiative-cooling-only, star-formation-only and
simulations that include both star formation and stellar-feedback. Comparison
of the density-temperature gas distribution between their star-formation-only
and feedback-including simulations (their Figs 11 and 12) reveals that the
low-density-high-temperature ISM is produced (for the most part) and is most
strongly affected by the stellar feedback. This is evident from both the mass
and the volume distributions. Conversely, the high-density-low-temperature ISM
(‘molecular’ or star-forming gas) is affected mostly by star formation,
although some density redistribution by stellar feedback is also evident.
Goldbaum et al. (2016) also makes a comparison between star-formation-only and
feedback-including simulations. Although this is not the focus of their work,
they also show that it is the feedback that is responsible for smoothing the
density distribution of the multi-phase ISM by preventing much of the mass
from being bound into dense clouds. Therefore, star formation alone should not
cause the lack of the lowest surface density gas found when comparing SFW0 to
RCW0 in Fig. 5.
To confirm these previous works, we run a short simulation similar to SFW0 but
without stellar feedback. In order to save computational resources, we restart
RCW0 40 Myr prior to RPS (160 Myr from the beginning of the simulation) and
turn on only the star formation with the same parameters as in our fiducial
model. In Fig. 12 we repeat the Fig. 5 but for RCW0 (top) SFW0-no-stellar-
feedback (middle) and SFW0 (bottom) galaxies (50 Myr later, 10 Myr after the
ICM wind has hit the galaxy). As expected from previous studies, the star
formation alone is not enough to affect the ISM, especially its low-density
phase, and, in fact, the only noticeable difference from RCW0 is in slightly
lower density peaks as the dense gas can now form stars. The run with star
formation and feedback (bottom panel), however, has the lowest density peaks
and, particularly in the central regions, has a narrower range in gas surface
density. We conclude that it is the stellar feedback that creates the more
homogeneous multi-phase ISM in SFW0.
Figure 12: Same as Fig. 5, but for RCWO (top) SFW0-no-stellar-feedback
(middle) and SFW0 (bottom) runs at 10 Myr.
Another factor that could potentially influence our results is the value of
the star formation threshold density. As we explained in Section 2, according
to the chosen star formation recipe, if the mass of a cell exceeds the Jeans
mass and the minimum threshold number density, a stellar particle will form
with a minimum mass of $10^{3}M_{\odot}$, with a chosen star formation
efficiency of 1 per cent. In the fiducial SFNW (in this Appendix referred to
as SF10) we choose a threshold number density of
$n_{\rm{thresh}}=10\,\rm{cm}^{-3}$. We also simulate galaxies from the same
initial conditions with $n_{\rm{thresh}}=3\,\rm{cm}^{-3}$ (SF03) and
$30\,\rm{cm}^{-3}$ (SF30) for 300 Myr (the time at which the ICM wind hits the
galaxies in the fiducial star-forming runs).
We find that the differences in the ISM brought by the change of threshold
number density are overall minor. We illustrate this by once again repeating
Fig. 5. In Fig. 13 we plot surface density distribution, this time for the
SF03 (top), SF10 (middle) and SF30 (bottom) runs. The central regions, where
the bulk of star formation occurs, are quite similar in SFR and gas surface
density across all the density thresholds. The gas surface density average and
distributions are very similar in all cases until at the very edge of the SF03
disc where there is little ISM gas as it has been used in star formation. The
differences in the ISM brought by the change of threshold number density are
overall unsubstantial. We further prove this by plotting in Fig. 14 the volume
density histograms of all the cells in the ISM of the galaxy disc ($R=30$ kpc
and $z\pm 2$ kpc) for SF03 (blue), SF10 (yellow) and SF30 (red) runs, which
are very close to each other. Thus, we conclude that the choice of the number
density threshold for star formation would not have a significant influence
over our results.
Figure 13: Same as Fig. 5, but for the SF03
($n_{\rm{thresh}}=3\,\rm{cm}^{-3}$, top), SF10 (or SFNW,
$n_{\rm{thresh}}=10\,\rm{cm}^{-3}$, middle) and SF30
($n_{\rm{thresh}}=30\,\rm{cm}^{-3}$, bottom) runs at 300 Myr of total
evolution, at which point the wind would hit the galaxies in the fiducial
runs. Note that the colour bar range is different in this figure, since
without RP there is no high-velocity gas in the disc. Figure 14: For SF03
($n_{\rm{thresh}}=3\,\rm{cm}^{-3}$, blue), SF10 (or SFNW,
$n_{\rm{thresh}}=10\,\rm{cm}^{-3}$, yellow) and SF30
($n_{\rm{thresh}}=30\,\rm{cm}^{-3}$, red) runs, the volume density histogram
of gas ($Z>0.35Z_{\odot}$) in a galaxy disc ($R=30$ kpc and $z\pm 2$ kpc)
taken at 300 Myr of total evolution.
## Appendix C Gas fallback in face-on stripped galaxies
Here for completeness we show the gas fallback in the face-on stripped
galaxies. We repeat the plots from Section 4.3 but for $v_{z}<0$ so that Fig.
15 is for 10 Myr (same as Fig. 6) and Fig. 16 is for 35 Myr (same as Fig. 7).
To avoid confusion, we show the difference of absolute values of velocity (top
panels) and mass flux (middle panels), so that, like in the previous plots,
orange colour means faster gas, more mass flux in RCW0 and purple colour means
that SFW0 dominates.
In Akerman et al. (2023) we show that fallback is negligible compared to the
radial flows withing the galaxy when it is subject to a varying, ever-
increasing ICM wind. It is no surprise, then, to see that on the maps at both
10 and 35 Myr the mass flux is almost 0 (notice that the colour bar scale is
the same as in Fig. with $v_{z}>0$). Since in green we outline the gas
distribution in RCW0, we can easily compare it to Fig. 6 and 7 and find that
the gas with negative $v_{z}$ is several kpc less radially extended than the
gas with positive $v_{z}$. This means that there is no fallback in the
outskirts or in the regions with the most violent stripping.
We can only see some fallback in the inner 10 kpc where it is dominated by
SFW0. As explained in Section 4.3, in SFW0 this part of the galaxy is
protected from RPS, since it is located within the stripping radius and the
outflows driven by feedback are able to fall back to the galaxy in fountain
flows. On the other hand, through the big low-density holes in the RCW0 gas
distribution the ICM wind can move freely, dragging some of the ICM with it.
As we can see here, this gas is truly stripped and does not fall back on the
galaxy.
Figure 15: Same as Fig. 6 but for $v_{z}<0.$ Figure 16: Same as Fig. 7 but for
$v_{z}<0.$
|
\\[1.3cm]
\quad
\quad
\quad
\quad
\quad
\quad
\\[0.7cm]
\hline
\quad
\tikz{0.6}{
\draw[lgray,line width=1.5pt,->] (-1,0) -- (1,0);
\draw[lgray,line width=1.5pt,->] (0,-1) -- (0,1);
\node[left] at (-1,0) {\tiny $1$};\node[right] at (1,0) {\tiny $1$};
\node[below] at (0,-1) {\tiny $1$};\node[above] at (0,1) {\tiny $1$};
\quad
\quad
\tikz{0.6}{
\draw[lgray,line width=1.5pt,->] (-1,0) -- (1,0);
\draw[lgray,line width=1.5pt,->] (0,-1) -- (0,1);
\node[left] at (-1,0) {\tiny $1$};\node[right] at (1,0) {\tiny $1$};
\node[below] at (0,-1) {\tiny $0$};\node[above] at (0,1) {\tiny $0$};
\quad
\quad
\tikz{0.6}{
\draw[lgray,line width=1.5pt,->] (-1,0) -- (1,0);
\draw[lgray,line width=1.5pt,->] (0,-1) -- (0,1);
\node[left] at (-1,0) {\tiny $1$};\node[right] at (1,0) {\tiny $0$};
\node[below] at (0,-1) {\tiny $0$};\node[above] at (0,1) {\tiny $1$};
\quad
\\[1.3cm]
\quad
\quad
\quad
\quad
\quad
\quad
\\[0.7cm]
\hline
\end{tabular}
\end{align}
Observe that these weights are stochastic, that is,
∑_i_2, j_2 ℛ_y / x (i_1, j_1; i_2, j_2) = 1.
§.§ Partition Function $Z(x_1,\dots,x_N;y_1,\dots,y_N;k)$
Fix two alphabets $(x_1,\dots,x_N) \in \mathbb{C}^N$ and $(y_1,\dots,y_N) \in \mathbb{C}^N$, and an integer $k \in \mathbb{Z}_{\geq 0}$. Define the following partition function in the quadrant:
\begin{align}
\label{hybrid-pf}
\tikz{0.75}{
\foreach\y in {1,...,5}{
\draw[lgray,line width=1.5pt,->] (0,\y) -- (7,\y);
\foreach\x in {1,...,5}{
\draw[lgray,line width=1.5pt,->] (\x,0) -- (\x,7);
\draw[lgray,line width=4pt,->] (6,0) -- (6,6) -- (0,6);
%top labels
\node[above] at (5,7) {$1$};
\node[above] at (4,7) {$1$};
\node[above] at (3,7) {$1$};
\node[above] at (2,7) {$1$};
\node[above] at (1,7) {$1$};
%bottom labels
\node[below] at (6,0) {$k$};
\node[below] at (5,0) {$0$};
\node[below] at (4,0) {$0$};
\node[below] at (3,0) {$0$};
\node[below] at (2,0) {$0$};
\node[below] at (1,0) {$0$};
\node[below,circle,scale=0.75,draw] at (5,-0.75) {$y_N$};
\node[below] at (3.5,-0.75) {$\cdots$};
\node[below] at (2.5,-0.75) {$\cdots$};
\node[below,circle,scale=0.75,draw] at (1,-0.75) {$y_1$};
%left labels
\node[left] at (0,1) {$1$};
\node[left] at (0,2) {$1$};
\node[left] at (0,3) {$1$};
\node[left] at (0,4) {$1$};
\node[left] at (0,5) {$1$};
\node[left] at (0,6) {$k$};
\node[left,circle,scale=0.75,draw] at (-0.75,1) {$\overline{x}_1$};
\node[left] at (-0.75,2.5) {$\vdots$};
\node[left] at (-0.75,3.5) {$\vdots$};
\node[left,circle,scale=0.75,draw] at (-0.75,5) {$\overline{x}_N$};
%right labels
\node[right] at (7,1) {$0$};
\node[right] at (7,2) {$0$};
\node[right] at (7,3) {$0$};
\node[right] at (7,4) {$0$};
\node[right] at (7,5) {$0$};
\end{align}
Vertices in the topmost row and the rightmost column are of the form (<ref>); all other vertices are given by (<ref>). The variables associated to horizontal lines are reciprocated; namely, we set $\overline{x}_a = 1/x_a$, for all $1 \leq a \leq N$. Partition functions of the form (<ref>) were originally studied in [52], where they appeared in connection with Cauchy identities and as a one-parameter generalization of domain wall partition functions [33, 37].
We now show that (<ref>) admits a nice probabilistic interpretation. Fix parameters $t$, $(x_1,\dots,x_N)$ and $(y_1,\dots,y_N)$ such that each vertex (<ref>) within (<ref>) has real weight in the interval $[0,1]$. We may then associate to any configuration $\mathfrak{C}$ in the (finite) quadrant $[1,N] \times [1,N]$ (that does not include the thick arrow along the northeast boundary in (<ref>)) a probability weight $\mathbb{P}_{6{\rm v}}(\mathfrak{C})$, defined as the product of weights of all vertices within $\mathfrak{C}$.
Let $\mathfrak{h}(N,N)$ denote the height function assigned to the vertex $(N,N)$ within the stochastic six-vertex model. We have that
\begin{align}
\label{6v-E}
\prod_{a=1}^N
\cdot
\mathbb{E}_{6{\rm v}}
\left[
\frac{(t^{k+1};t)_{\infty}}{(t^{k+1+\mathfrak{h}(N,N)};t)_{\infty}}
\right],
\end{align}
where the expectation is taken with respect to the measure $\mathbb{P}_{6{\rm v}}$ defined above.
We begin by decomposing the partition function (<ref>) along the edges where vertices of the types (<ref>) and (<ref>) meet. This produces the equation
\begin{multline}
\label{Z-expand}
\\
= \sum_{\{i_1,\dots,i_N\} \in \{0,1\}^N}
\sum_{\{j_1,\dots,j_N\} \in \{0,1\}^N}
\mathbb{P}_{6{\rm v}}(i_1,\dots,i_N;j_1,\dots,j_N)
\end{multline}
where we have defined two new partition functions. The first is given by
\begin{align}
\label{square}
\mathbb{P}_{6{\rm v}}(i_1,\dots,i_N;j_1,\dots,j_N)
\tikz{0.8}{
\foreach\y in {1,...,5}{
\draw[lgray,line width=1.5pt,->] (0,\y) -- (6,\y);
\foreach\x in {1,...,5}{
\draw[lgray,line width=1.5pt,->] (\x,0) -- (\x,6);
%top labels
\node[above] at (5,6) {$i_N$};
\node[above] at (4,6) {$\cdots$};
\node[above] at (3,6) {$\cdots$};
\node[above] at (2,6) {$\cdots$};
\node[above] at (1,6) {$i_1$};
%bottom labels
\node[below] at (5,0) {$0$};
\node[below] at (4,0) {$0$};
\node[below] at (3,0) {$0$};
\node[below] at (2,0) {$0$};
\node[below] at (1,0) {$0$};
\node[below,circle,scale=0.75,draw] at (5,-0.75) {$y_N$};
\node[below] at (3.5,-0.75) {$\cdots$};
\node[below] at (2.5,-0.75) {$\cdots$};
\node[below,circle,scale=0.75,draw] at (1,-0.75) {$y_1$};
%left labels
\node[left] at (0,1) {$1$};
\node[left] at (0,2) {$1$};
\node[left] at (0,3) {$1$};
\node[left] at (0,4) {$1$};
\node[left] at (0,5) {$1$};
\node[left,circle,scale=0.75,draw] at (-0.75,1) {$\overline{x}_1$};
\node[left] at (-0.75,2.5) {$\vdots$};
\node[left] at (-0.75,3.5) {$\vdots$};
\node[left,circle,scale=0.75,draw] at (-0.75,5) {$\overline{x}_N$};
%right labels
\node[right] at (6,1) {$j_1$};
\node[right] at (6,2) {$\vdots$};
\node[right] at (6,3) {$\vdots$};
\node[right] at (6,4) {$\vdots$};
\node[right] at (6,5) {$j_N$};
\end{align}
where all vertices are of the type (<ref>); this quantity is the probability that a random configuration $\mathfrak{C}$ in the quadrant $[1,N] \times [1,N]$ has state $i_a \in \{0,1\}$ exiting vertically from vertex $(a,N)$ and state $j_a \in \{0,1\}$ exiting horizontally from vertex $(N,a)$, for all $1 \leq a \leq N$. The second is a tower of vertices of the type (<ref>) (which one may also view as a straightened version of the thick arrow along the northeast boundary in (<ref>)):
\begin{align}
\label{tower}
\tikz{0.85}{
\foreach\y in {1,...,6}{
\draw[lgray,line width=1.5pt,->] (0,\y) -- (2,\y);
\draw[lgray,line width=4pt,->] (1,0) -- (1,7);
%top labels
\node[above] at (1,7) {$k$};
%bottom labels
\node[below] at (1,0) {$k$};
%left labels
\node[left] at (0,1) {$j_1$};
\node[left] at (0,2) {$\vdots$};
\node[left] at (0,3) {$j_N$};
\node[left] at (0,4) {$i_N$};
\node[left] at (0,5) {$\vdots$};
\node[left] at (0,6) {$i_1$};
\node[left,circle,scale=0.75,draw] at (-0.75,1) {$\overline{x}_1$};
\node[left] at (-0.75,2) {$\vdots$};
\node[left,circle,scale=0.75,draw] at (-0.75,3) {$\overline{x}_N$};
\node[left,circle,scale=0.75,draw] at (-0.75,4) {$y_N$};
\node[left] at (-0.75,5) {$\vdots$};
\node[left,circle,scale=0.75,draw] at (-0.75,6) {$y_1$};
%right labels
\node[right] at (2,1) {$0$};
\node[right] at (2,2) {$0$};
\node[right] at (2,3) {$0$};
\node[right] at (2,4) {$1$};
\node[right] at (2,5) {$1$};
\node[right] at (2,6) {$1$};
\end{align}
In view of the arrow conservation property of the vertices (<ref>), it is easy to see that each internal vertical edge within (<ref>) admits a unique state such that the tower has non-vanishing weight. This allows us to compute $H_k(j_1,\dots,j_N;i_1,\dots,i_N)$ explicitly:
\begin{align}
\label{H-prod}
\prod_{a=1}^{N}
\mathcal{L}_{\overline{x}_a}(k+J_{a-1},j_a;k+J_a,0)
\mathcal{L}_{y_a}(k+I_a,i_a;k+I_{a-1},1),
\end{align}
where we have defined the partial sums $J_0 = I_0 = 0$, $J_a = \sum_{b=1}^{a} j_b$, $I_a = a-\sum_{b=1}^{a} i_b$ for $1 \leq a \leq N$, and where we note that $J_N = I_N$. From the table (<ref>), the weights
$\mathcal{L}_{\overline{x}_a}(k+J_{a-1},j_a;k+J_a,0)$ are all equal to $1$; the remaining terms in the product (<ref>) are given by
\begin{align*}
\mathcal{L}_{y_a}(k+I_a,i_a;k+I_{a-1},1)
y_a \cdot
\left\{
\begin{array}{ll}
1-t^{k+I_a}, & \qquad i_a=0,
\\
\\
1, & \qquad i_a =1.
\end{array}
\right.
\end{align*}
The product (<ref>) then simplifies to
\begin{align*}
\prod_{a=1}^{N} y_a
\cdot
\prod_{b=1}^{I_N} (1-t^{k+b})
\prod_{a=1}^{N} y_a
\cdot
\prod_{b=1}^{J_N} (1-t^{k+b}).
\end{align*}
In particular, $H_k(j_1,\dots,j_N;i_1,\dots,i_N)$ depends on $(i_1,\dots,i_N)$ and $(j_1,\dots,j_N)$ only via $I_N = J_N = \sum_{b=1}^{N} j_b$. In fact, returning to the quadrant (<ref>), we see that
$J_N = \mathfrak{h}(N,N)$; accordingly, one has
\begin{align*}
\prod_{a=1}^{N}
\cdot
\prod_{b=1}^{\mathfrak{h}(N,N)}
\\
\prod_{a=1}^{N}
\cdot
\prod_{b=1}^{\infty}
\frac{1-t^{k+b}}{1-t^{k+b+\mathfrak{h}(N,N)}}
\prod_{a=1}^{N}
\cdot
\frac{(t^{k+1};t)_{\infty}}{(t^{k+1+\mathfrak{h}(N,N)};t)_{\infty}}.
\end{align*}
Coming back to (<ref>), we then find that
\begin{multline*}
\\
\prod_{a=1}^{N}
\cdot
\sum_{\{i_1,\dots,i_N\} \in \{0,1\}^N}
\sum_{\{j_1,\dots,j_N\} \in \{0,1\}^N}
\mathbb{P}_{6{\rm v}}(i_1,\dots,i_N;j_1,\dots,j_N)
\frac{(t^{k+1};t)_{\infty}}{(t^{k+1+\mathfrak{h}(N,N)};t)_{\infty}}.
\end{multline*}
Conditioning on the possible values of $\mathfrak{h}(N,N)$, this may be written as
\begin{align*}
%\sum_{\{i_1,\dots,i_N\} \in \{0,1\}^N}
%\sum_{\{j_1,\dots,j_N\} \in \{0,1\}^N}
%\mathbb{P}_{6{\rm v}}(i_1,\dots,i_N;j_1,\dots,j_N)
\prod_{a=1}^{N}
\cdot
\sum_{m=0}^{N}
\frac{(t^{k+1};t)_{\infty}}{(t^{k+1+m};t)_{\infty}}
\mathbb{P}_{6{\rm v}}(\mathfrak{h}(N,N)=m),
\end{align*}
which proves the claim (<ref>).
§.§ Determinant Evaluation
Following Section 4.2 and Appendix B of [52], the partition function (<ref>) may be computed in closed form:
For any $N \geq 1$ and $k \geq 0$, one has
\begin{align}
\label{Z-det}
\frac{\prod_{i=1}^{N}y_i \prod_{1 \leq i,j \leq N} (1-x_i y_j)}
{\prod_{1 \leq i<j \leq N} (x_i-x_j)(y_i-y_j)}
\det_{1 \leq i,j \leq N}
\left[
\frac{1-t^{k+1}-t(1-t^k) x_i y_j}{(1-x_i y_j)(1-t x_i y_j)}
\right].
\end{align}
The proof relies on finding a list of properties of $Z(x_1,\dots,x_N;y_1,\dots,y_N;k)$ that determine it uniquely; one then shows that the right hand side of (<ref>) obeys the same properties. We list these properties below without derivation; for more information, we refer the reader to <cit.>.
* $Z(x_1,\dots,x_N;y_1,\dots,y_N;k)$ is symmetric in the alphabet $(x_1,\dots,x_N)$ and separately in the alphabet $(y_1,\dots,y_N)$;
* $\prod_{1 \leq i,j \leq N} (1-t x_i y_j) \cdot Z(x_1,\dots,x_N;y_1,\dots,y_N;k)$ is a polynomial in $x_N$ of degree $N$;
* Setting $x_N = 1/y_N$ we have
\begin{align*}
\Big|_{x_N = 1/y_N}
\cdot
\end{align*}
* Setting $x_i = 0$ for all $1 \leq i \leq N$, we have
\begin{align*}
\prod_{b=1}^{N}
\end{align*}
* For $N=1$, there holds
\begin{align*}
y_1 \cdot \left( \frac{1-t^{k+1}-t(1-t^k)x_1 y_1}{1-t x_1 y_1} \right).
\end{align*}
§.§ Schur Expectation
Having computed (<ref>) in determinant form, it is now easy to pass to its Schur expansion.
Define the following Schur measure with respect to two alphabets $(x_1,\dots,x_N)$ and $(y_1,\dots,y_N)$:
\begin{align}
\label{schur-meas}
\mathbb{P}_{\SM}(\lambda)
\prod_{1 \leq i,j \leq N}
(1-x_i y_j)
\cdot
\end{align}
For fixed $N \geq 1$ and $k \geq 0$ we then have the identity
\begin{align}
\label{Z-schur}
%\prod_{i=1}^{N}y_i \prod_{1 \leq i,j \leq N} (1-x_i y_j)
\prod_{i=1}^{N} y_i
\cdot
\mathbb{E}_{\SM}
\left[
\prod_{i=1}^{N}
\right],
\end{align}
where the expectation is taken with respect to the measure (<ref>).
We begin by manipulating the determinant present in (<ref>). One has
\begin{align*}
\det_{1 \leq i,j \leq N}
\left[
\frac{1-t^{k+1}-t(1-t^k) x_i y_j}{(1-x_i y_j)(1-t x_i y_j)}
\right]
\det_{1 \leq i,j \leq N}
\left[
\frac{1}{1-x_i y_j}
\frac{t^{k+1}}{1-t x_i y_j}.
\right],
\end{align*}
Replacing the two fractions on the right hand side by their corresponding geometric series, we obtain the identity
\begin{align*}
\det_{1 \leq i,j \leq N}
\left[
\frac{1-t^{k+1}-t(1-t^k) x_i y_j}{(1-x_i y_j)(1-t x_i y_j)}
\right]
%\det_{1 \leq i,j \leq N}
%(x_i y_j)^a
%t^a (x_i y_j)^a
\det_{1 \leq i,j \leq N}
\left[
\sum_{a=0}^{\infty}
(x_i y_j)^a
\right].
\end{align*}
To the latter we apply the Cauchy–Binet identity, which yields
\begin{align*}
\det_{1 \leq i,j \leq N}
\left[
\frac{1-t^{k+1}-t(1-t^k) x_i y_j}{(1-x_i y_j)(1-t x_i y_j)}
\right]
\sum_{0 \leq a_1 < \cdots < a_N}
\prod_{i=1}^{N}
\det_{1 \leq i,j \leq N}
\left[ x_i^{a_j} \right]
\det_{1 \leq i,j \leq N}
\left[ y_j^{a_i} \right];
\end{align*}
including Vandermonde factors on both sides of the equation, and making the change of summation indices $a_i = \lambda_{N-i+1}+i-1$, we then find that
\begin{multline*}
\prod_{1 \leq i<j \leq N}
\frac{1}{(x_i-x_j)(y_i-y_j)}
\cdot
\det_{1 \leq i,j \leq N}
\left[
\frac{1-t^{k+1}-t(1-t^k) x_i y_j}{(1-x_i y_j)(1-t x_i y_j)}
\right]
\\
\sum_{\lambda_1 \geq \cdots \geq \lambda_N \geq 0}
\
\prod_{i=1}^{N}
\end{multline*}
This recovers the claim (<ref>), after multiplying through by $\prod_{i=1}^{N} y_i \cdot \prod_{1 \leq i,j \leq N} (1-x_i y_j)$.
§.§ Final Match
Comparing (<ref>) and (<ref>), we have proved that
\begin{align}
\label{final-match}
\mathbb{E}_{6{\rm v}}
\left[
\frac{(t^{k+1};t)_{\infty}}{(t^{k+1+\mathfrak{h}(N,N)};t)_{\infty}}
\right]
\mathbb{E}_{\SM}
\left[
\prod_{i=1}^{N}
\right].
\end{align}
Both sides of (<ref>) are polynomial in $t^{k}$; since (<ref>) holds for all integer values $k \in \mathbb{Z}_{\geq 0}$ and the set $\{t^k\}_{k \geq 0}$ has a point of accumulation for $|t|<1$, we may extend the equality to all complex values by the analytic continuation $t^{k+1} = -\zeta \in \mathbb{C}$. This yields
\begin{align*}
\mathbb{E}_{6{\rm v}}
\left[
\frac{(-\zeta;t)_{\infty}}{(-\zeta t^{\mathfrak{h}(N,N)};t)_{\infty}}
\right]
\mathbb{E}_{\SM}
\left[
\prod_{i=1}^{N}
(1+\zeta t^{\lambda_i-i+N})
\right],
\end{align*}
which is precisely the result (<ref>) in the case $q=t$, $r_i=s_i=t^{-1/2}$ and $M=N$.
Extending this result to $M<N$ may be achieved by taking $y_{M+1} = \cdots = y_N = 0$ in (<ref>). On the side of the six-vertex model, this choice trivializes the contribution of the final $N-M$ thin vertical lines in the picture (<ref>) (as no arrows can enter them through the left), leading to a rectangular domain; on the side of the Schur expectation, this choice does not damage the measure (<ref>) in view of the stability property of Schur polynomials. A similar reduction is possible in the case $M>N$.
[1]
A. Aggarwal.
Current fluctuations of the stationary ASEP and six-vertex model.
Duke Math. J., 167(2):269–384, 2018.
[2]
A. Aggarwal and A. Borodin.
Phase transitions in the ASEP and stochastic six-vertex model.
Ann. Probab., 47(2):613–689, 2019.
[3]
A. Aggarwal, A. Borodin, and A. Bufetov.
Stochasticization of solutions to the Yang-Baxter equation.
Ann. Henri Poincaré, 20(8):2495–2554, 2019.
[4]
A. Aggarwal, A. Borodin, and M. Wheeler.
Colored fermionic vertex models and symmetric functions.
Preprint, arXiv:2101.01605.
[5]
D. Aldous and P. Diaconis.
Longest increasing subsequences: from patience sorting to the
Baik-Deift-Johansson theorem.
Bull. Amer. Math. Soc. (N.S.), 36(4):413–432, 1999.
[6]
G. Amir, I. Corwin, and J. Quastel.
Probability distribution of the free energy of the continuum directed
random polymer in $1+1$ dimensions.
Comm. Pure Appl. Math., 64(4):466–537, 2011.
[7]
J. Baik, G. Ben Arous, and S. Péché.
Phase transition of the largest eigenvalue for nonnull complex sample
covariance matrices.
Ann. Probab., 33(5):1643–1697, 2005.
[8]
J. Baik, P. Deift, and K. Johansson.
On the distribution of the length of the longest increasing
subsequence of random permutations.
J. Amer. Math. Soc., 12(4):1119–1178, 1999.
[9]
J. Baik and E. M. Rains.
Limiting distributions for a polynuclear growth model with external
J. Statist. Phys., 100:523–541, 2000.
[10]
G. Barraquand.
A phase transition for $q$-TASEP with a few slower particles.
Stoch. Proc. Appl., 125(7):2674–2699, 2015.
[11]
A. Borodin.
On a family of symmetric rational functions.
Adv. Math., 306:973–1018, 2017.
[12]
A. Borodin.
Stochastic higher spin six vertex model and Macdonald measures.
J. Math. Phys., 59(2):023301, 17, 2018.
[13]
A. Borodin, A. Bufetov, and M. Wheeler.
Between the stochastic six vertex model and Hall–Littlewood
Preprint, arXiv:1611.09486.
[14]
A. Borodin and I. Corwin.
Macdonald processes.
Probab. Theory Related Fields, 158(1-2):225–400, 2014.
[15]
A. Borodin, I. Corwin, and P. Ferrari.
Free energy fluctuations for directed polymers in random media in
$1+1$ dimension.
Comm. Pure Appl. Math., 67(7):1129–1214, 2014.
[16]
A. Borodin and V. Gorin.
Lectures on integrable probability.
In Probability and statistical physics in St. Petersburg,
volume 91 of Proc. Sympos. Pure Math., pages 155–214. Amer. Math.
Soc., Providence, RI, 2016.
[17]
A. Borodin and V. Gorin.
Moments match between the KPZ equation and the Airy point
SIGMA Symmetry Integrability Geom. Methods Appl., 12:Paper No.
102, 7, 2016.
[18]
A. Borodin, A. Okounkov, and G. Olshanski.
Asymptotics of Plancherel measures for symmetric groups.
J. Amer. Math. Soc., 13(3):481–515, 2000.
[19]
A. Borodin and G. Olshanski.
The ASEP and determinantal point processes.
Comm. Math. Phys., 353(2):853–903, 2017.
[20]
A. Borodin and S. Péché.
Airy kernel with two sets of parameters in directed percolation and
random matrix theory.
J. Stat. Phys., 132(2):275–290, 2008.
[21]
A. Borodin and L. Petrov.
Nearest neighbor Markov dynamics on Macdonald processes.
Adv. Math., 300:71–155, 2016.
[22]
A. Borodin and L. Petrov.
Higher spin six vertex model and symmetric rational functions.
Selecta Math. (N.S.), 24(2):751–874, 2018.
[23]
A. Borodin and M. Wheeler.
Coloured stochastic vertex models and their spectral theory.
Preprint, arXiv:1808.01866.
[24]
A. Bufetov and K. Matveev.
Hall-Littlewood RSK field.
Selecta Math. (N.S.), 24(5):4839–4884, 2018.
[25]
P. Calabrese, P. Le Doussal, and A. Rosso.
Free-energy distribution of the directed polymer at high temperature.
EPL (Europhysics Letters), 90(2):20002, apr 2010.
[26]
I. Corwin.
The Kardar-Parisi-Zhang equation and universality class.
Random Matrices Theory Appl., 1(1):1130001, 76, 2012.
[27]
I. Corwin, P. Ghosal, H. Shen, and L.-C. Tsai.
Stochastic PDE limit of the six vertex model.
Comm. Math. Phys., 375(3):1945–2038, 2020.
[28]
I. Corwin and L. Petrov.
The $q$-PushASEP: a new integrable model for traffic in $1+1$
J. Stat. Phys., 160(4):1005–1026, 2015.
[29]
I. Corwin and L. Petrov.
Stochastic higher spin vertex models on the line.
Comm. Math. Phys., 343(2):651–700, 2016.
[30]
V. Dotsenko.
Bethe ansatz derivation of the tracy-widom distribution for
one-dimensional directed polymers.
EPL (Europhysics Letters), 90(2):20003, apr 2010.
[31]
P. L. Ferrari and H. Spohn.
Random growth models.
In The Oxford handbook of random matrix theory, pages
782–801. Oxford Univ. Press, Oxford, 2011.
[32]
T. Imamura and T. Sasamoto.
Fluctuations of the one-dimensional polynuclear growth model with
external sources.
Nuclear Phys. B, 699(3):503–544, 2004.
[33]
A. G. Izergin.
Partition function of a six-vertex model in a finite volume.
In Doklady Akademii Nauk, volume 297, pages 331–333. Russian
Academy of Sciences, 1987.
[34]
K. Johansson.
Shape fluctuations and random matrices.
Comm. Math. Phys., 209(2):437–476, 2000.
[35]
K. Johansson.
Discrete orthogonal polynomial ensembles and the Plancherel
Ann. of Math. (2), 153(1):259–296, 2001.
[36]
M. Kardar, G. Parisi, and Y.-C. Zhang.
Dynamic scaling of growing interfaces.
Phys. Rev. Lett., 56:889–892, Mar 1986.
[37]
V. E. Korepin.
Calculation of norms of Bethe wave functions.
Communications in Mathematical Physics, 86(3):391–418, 1982.
[38]
P. P. Kulish, N. Y. Reshetikhin, and E. K. Sklyanin.
Yang-Baxter equations and representation theory. I.
Lett. Math. Phys., 5(5):393–403, 1981.
[39]
C. T. MacDonald, J. H. Gibbs, and A. C. Pipkin.
Kinetics of biopolymerization on nucleic acid templates.
Biopolymers, 6(1):1–25, 1968.
[40]
I. G. Macdonald.
Symmetric functions and Hall polynomials.
Oxford Classic Texts in the Physical Sciences. Oxford University
Press, New York, second edition, 2015.
[41]
V. V. Mangazeev.
On the Yang-Baxter equation for the six-vertex model.
Nuclear Phys. B, 882:70–96, 2014.
[42]
P. Meakin.
Fractals, scaling and growth far from equilibrium, volume 5 of
Cambridge Nonlinear Science Series.
Cambridge University Press, Cambridge, 1998.
[43]
A. Okounkov.
Infinite wedge and random partitions.
Selecta Math. (N.S.), 7(1):57–81, 2001.
[44]
Y. Pei.
A $q$-Robinson-Schensted-Knuth algorithm and a $q$-polymer.
Electron. J. Combin., 24(4):Paper No. 4.6, 38, 2017.
[45]
M. Prähofer and H. Spohn.
Universal distributions for growth processes in $1+1$ dimensions and
random matrices.
Phys. Rev. Lett., 84:4882–4885, May 2000.
[46]
M. Prähofer and H. Spohn.
Scale invariance of the PNG droplet and the Airy process.
J. Statist. Phys., 108(5-6):1071–1106, 2002.
[47]
J. Quastel.
Introduction to KPZ.
In Current developments in mathematics, 2011, pages 125–194.
Int. Press, Somerville, MA, 2012.
[48]
T. Sasamoto and H. Spohn.
Exact height distributions for the KPZ equation with narrow wedge
initial condition.
Nucl. Phys. B, 834(3):523–542, 2010.
[49]
F. Spitzer.
Interaction of Markov processes.
Adv. Math., 5:246–290 (1970), 1970.
[50]
C. A. Tracy and H. Widom.
Asymptotics in ASEP with step initial condition.
Comm. Math. Phys., 290(1):129–154, 2009.
[51]
B. Virag.
The heat and the landscape I.
Preprint, arXiv:2008.07241.
[52]
M. Wheeler and P. Zinn-Justin.
Refined Cauchy/Littlewood identities and six-vertex model
partition functions: III. Deformed bosons.
Advances in Mathematics, 299:543–600, 2016.
|
11institutetext: François Lamothe 22institutetext: Alain Haït 33institutetext:
Emmanuel Rachelson 44institutetext: ISAE-SUPAERO, Université de Toulouse
44email<EMAIL_ADDRESS>55institutetext: Claudio Contardo
66institutetext: Concordia University, Canada 77institutetext: Bernard Gendron
88institutetext: Université de Montréal, Canada
# On the integration of Dantzig-Wolfe and Fenchel decompositions via
directional normalizations
François Lamothe Alain Haït Emmanuel Rachelson Claudio Contardo Bernard
Gendron
###### Abstract
The strengthening of linear relaxations and bounds of mixed integer linear
programs has been an active research topic for decades. Enumeration-based
methods for integer programming like linear programming-based branch-and-bound
exploit strong dual bounds to fathom unpromising regions of the feasible
space. In this paper, we consider the strengthening of linear programs via a
composite of Dantzig-Wolfe and Fenchel decompositions. We provide geometric
interpretations of these two classical methods. Motivated by these geometric
interpretations, we introduce a novel approach for solving Fenchel sub-
problems and introduce a novel decomposition combining Dantzig-Wolfe and
Fenchel decompositions in an original manner. We carry out an extensive
computational campaign assessing the performance of the novel decomposition on
the unsplittable flow problem. Very promising results are obtained when the
new approach is compared to classical decomposition methods.
###### Keywords:
Mixed Integer Linear Programming Decomposition methods Unsplittable flows
Dantzig-Wolfe decomposition Fenchel decomposition
## Acknowledgments
This document is the result of a research project funded by the Centre
national d’études spatiales (CNES) and Thales Alenia Space. C. Contardo and B.
Gendron thank the Natural Sciences and Engineering Research Council of Canada
(NSERC) for its financial support under Discovery Grants no. 2020-06311 and
2017-06054.
## 1 Introduction
Enumeration-based algorithms are arguably the main algorithmic frameworks to
solve mixed-integer linear programs among which the branch-and-bound (B&B)
method land1960automatic (33) is perhaps the most efficient and versatile one.
It relies on the ability to compute primal and dual bounds on the value of the
solutions to the problem at hand. The method is most efficient when quick
computing of tight primal and dual bounds is available, resulting in short
enumerations. Traditionally, primal bounds are found via heuristics and dual
bounds via solving relaxations of the problem. The most classical such
relaxation is the so-called linear relaxation in which the integrality
constraints are ignored. The resulting problem, an ordinary linear program,
can be solved efficiently either in polynomial time via interior point methods
KHACHIYAN198053 (30, 38), or via a greedy method — namely the simplex method
dantzig1951maximization (19) — of exponential worst-case complexity, but very
efficient in practice borgwardt2012simplex (10).
A common practice to improve the linear relaxation of mixed integer linear
programs is to apply a decomposition method to a subset of the problem’s
constraints whose associated polyhedron does not have the integrality property
(some vertices of the polyhedron have non-integer components). The
decomposition tightens the polyhedron which in turn strengthens the linear
relaxation of the problem. One of the most classical decomposition methods is
the one by dantzig1960decomposition (20), known as the Dantzig-Wolfe
decomposition. It has proven successful in various applications which explains
the attention it has received over the years desaulniers2006column (22). The
Dantzig-Wolfe decomposition is closely related to the Lagrangian decomposition
which has also been successful in many practical applications. The main
difference between the Lagrangian and Dantzig-Wolfe decomposition is that in
the former only dual information is extracted and exploited in a sub-gradient
algorithm whereas the latter also extracts primal information which can be
embedded within an enumeration scheme, usually referred to as branch-and-price
barnhart1998branch (3). Because in this work we will consider column
generation and cutting plane procedures, we will focus on the Dantzig-Wolfe
decomposition.
Our analysis restricts to decompositions that are able to exploit integer
subproblems —as opposed to pure linear problems—, and for that reason, we do
not consider Benders decomposition benders1962partitioning (7). Fenchel
decomposition, on the other hand, is a cutting plane method similar to Benders
decomposition that can be applied when the sub-problem is integer. It is thus
able to improve the relaxation quality of mixed integer linear programs. This
technique has had success in several applications such as knapsack problems
boyd1993generating (11, 29), generalized assignment problems
avella2010computational (1), network design problems with unsplittable flow
chen2021exact (15) or even stochastic optimization problems ntaimo2013fenchel
(40, 4).
Although Fenchel and Dantzig-Wolfe decompositions have been largely studied in
the context of linear relaxation strengthening, several limitations of both
decomposition methods have been identified in the literature which may affect
their convergence. In particular, the Dantzig-Wolfe decomposition is known to
suffer from the degeneracy of its master problem (which happens when the
master problem admits several optimal dual solutions). This explains why a
considerable effort has been made by the scientific community to improve these
decomposition methods and overcome their weaknesses. This work is in this line
of contribution. We show that Dantzig-Wolfe and Fenchel decompositions can be
seen in a similar light as constructing inner and outer approximations of the
polyhedron being decomposed and that the synergy between the two
approximations may be beneficial for the entire approach when applied to some
large-scale mixed-integer linear programs.
We outline the main contributions of our work as follows:
* •
We provide geometric interpretations of Dantzig-Wolfe and Fenchel
decompositions which may fuel complementary insights on these two approaches
compared to purely analytical interpretations.
* •
We provide a critical overview of normalizations at the core of the separation
problems arising in Fenchel decomposition. The normalization impacts the type
of cut, its properties, and ultimately the convergence speed of the
decomposition method. We provide geometric interpretations of several types of
normalization and their properties.
* •
We introduce a novel approach to the Fenchel sub-problem when a directional
normalization is used. The proposed method possesses reduces the numerical
instabilities of a direct resolution approach commonly used. We show that the
new approach solves the Fenchel sub-problem in finitely many iterations.
* •
We introduce a new decomposition method inspired by both the Dantzig-Wolfe and
the Fenchel decompositions. The proposed method uses a Dantzig-Wolfe master
problem and a Fenchel master problem. A Fenchel sub-problem guided with a
directional normalization is used to coordinate the two master problems. The
resulting method is shown to perform especially well on instances presenting
high degrees of degeneracy. We provide a possible explanation of this
phenomenon based on our findings.
In the following, we start by giving in Section 2 the context in which we
apply the various decomposition methods together with some notation. Then, we
present in Section 3 a geometric interpretation of Dantzig-Wolfe and Fenchel
decomposition methods. Section 4 is dedicated to an overview of the resolution
of the Fenchel subproblem and in particular, the impact of the normalization
used. This concept will be used in the next section for the new methods
introduced in the paper. We then present in Section 5 a new approach to
solving the Fenchel subproblem. In Section 6, we introduce the new
decomposition method which integrates both Dantzig-Wolfe and Fenchel master
problems as well as a Fenchel subproblem. In Section 7, we illustrate our
method by applying it to the unsplittable flow problem. An experimental
evaluation of all proposed methods is then made in Section 8 on small and
medium-sized instances of the problem.
## 2 Context
In the remainder of this paper, we will apply decomposition methods to the
following general mixed integer linear program:
$\displaystyle(P)\quad$ $\displaystyle\max_{x}$ $\displaystyle cx$ (1a)
subject to $\displaystyle A_{1}x\leq b_{1}$ (1b) $\displaystyle A_{2}x\leq
b_{2}$ (1c) $\displaystyle x\in X$ (1d)
where $X$ is a product set $\mathbb{R}^{p}\times\mathbb{Z}^{q}$ of appropriate
dimension whose linear relaxation $\mathbb{R}^{p+q}$ will be noted $\bar{X}$.
Moreover, to explain the methods used, we will consider the following
polyhedra which are assumed to be bounded to simplify the explanations:
$\displaystyle LR_{1}=\\{x\in\bar{X}|A_{1}x\leq b_{1}\\}$ $\displaystyle
LR_{2}=\\{x\in\bar{X}|A_{2}x\leq b_{2}\\}$ $\displaystyle Q_{2}=conv(\\{x\in
X|A_{2}x\leq b_{2}\\})$
In order to bound from above the value of the optimal solution of a general
mixed integer linear program as ($P$) one can optimize its objective function
over a relaxation containing the set of valid solutions. Usually, the linear
relaxation of the set of solutions $LR_{1}\cap LR_{2}$ is used. However, one
might want a tighter relaxation to obtain a better upper bound. This can be
obtained by using the relaxation $LR_{1}\cap Q_{2}$ which still contains the
set of valid solutions while being included in $LR_{1}\cap LR_{2}$. As
illustrated in Figure 1, this relaxation usually returns strictly better
bounds than the linear relaxation when the polyhedron $Q_{2}$ is strictly
included in $LR_{2}$ which happens when the polyhedron $LR_{2}$ does not have
the integrality property (some of the vertices of the polyhedron have non-
integer coordinates). However, the drawback of $Q_{2}$ compared to $LR_{2}$ is
that one usually only has a representation of $Q_{2}$ with an exponential
number of variables or constraints which is not manageable directly by a
linear programming solver. In compensation, we assume to have at our disposal
an efficient algorithm to optimize a linear function on the polyhedron
$Q_{2}$. This algorithm is called the optimization oracle and solves the
following mixed integer linear program:
$\displaystyle(O)\quad$ $\displaystyle\max_{x}$ $\displaystyle\pi x$ subject
to $\displaystyle A_{2}x\leq b_{2}$ $\displaystyle x\in X$
where $x\mapsto\pi x$ is any linear function. Thus, to be able to optimize
over $LR_{1}\cap Q_{2}$, the goal of a decomposition method is to use this
oracle to compute an approximation of $Q_{2}$ of manageable size. In
particular, the Dantzig-Wolfe decomposition iteratively grows an inner
approximation of $Q_{2}$ while the Fenchel decomposition iteratively refines
an outer approximation of $Q_{2}$.
$LR_{1}$$LR_{2}$ (a) Polyhedra associated with the linear relaxation
$LR_{1}$$Q_{2}$ (b) Polyhedra obtained after reformulation by the
decomposition methods
Figure 1: Representation of solution spaces of the linear relaxation and the
models obtained from the decomposition methods
Decomposition methods are known to work very well when the matrix $A_{2}$ is
block diagonal because it enables decomposing the problem of optimizing over
$Q_{2}$ into several smaller problems (one per block of the matrix). For
instance, in the case of the application considered in this study, the
unsplittable flow problem, the capacity constraints lead to a block diagonal
structure. In this context, the constraints $A_{1}x\leq b_{1}$ would
correspond to the flow conservation constraints while the constraints
$A_{2}x\leq b_{2}$ would correspond to the capacity constraints. However, our
work makes use of a different rationale. Let us assume that the matrix $A_{2}$
is sparse with non-zero elements for only a few variables; this may happen for
instance when $A_{2}$ corresponds to one of the blocks of a block diagonal
matrix. With this choice, it is possible to apply decomposition methods even
in contexts where the main problem ($P$) does not have any block diagonal
structure. For clarity purposes, we will present the methods as if we were
decomposing only one polyhedron at a time. However, in practice, one would
decompose several polyhedra at the same time; for example, all the blocks of a
block diagonal matrix.
## 3 Geometric interpretation of Dantzig-Wolfe and Fenchel decompositions
In this section, we provide a geometric interpretation of both decomposition
methods with the potential to fuel new intuitions regarding their strengths
and weaknesses.
### 3.1 Dantzig-Wolfe decomposition
Instead of optimizing the objective function of mixed integer linear program
(P) over the solution set $LR_{1}\cap LR_{2}$ of its linear relaxation,
Dantzig-Wolfe decomposition allows for the optimization over the smaller set
$LR_{1}\cap Q_{2}$. Because one does not usually have a manageable description
of the set $Q_{2}$, it is necessary to compute an approximation of $Q_{2}$ in
order to optimize over the intersection $LR_{1}\cap Q_{2}$. To compute this
approximation, we assume the availability of an optimization oracle over
$Q_{2}$. The main idea in the Dantzig-Wolfe decomposition is to iteratively
grow an inner approximation of $Q_{2}$. One can create such an approximation
by obtaining a set of points $x_{i}$ belonging to $Q_{2}$ (usually extreme
points of $Q_{2}$) and setting the approximation as the convex envelope of
these points. We will denote this inner approximation with $\widehat{Q}_{2}$.
The Dantzig-Wolfe decomposition illustrated in Figure 2 proceeds as follows:
1. 1.
Initialize the approximation $\widehat{Q}_{2}$ with points of $Q_{2}$
2. 2.
Find the optimal solution $\widehat{x}$ over $LR_{1}\cap\widehat{Q}_{2}$
3. 3.
Search for a point of $Q_{2}$ whose addition to $\widehat{Q}_{2}$ may improve
the value of the optimal solution $\widehat{x}$
4. 4.
If such a point exist, add it to $\widehat{Q}_{2}$ and go to Step 2.
5. 5.
Else: $\widehat{x}$ is the optimal solution over $LR_{1}\cap Q_{2}$. Stop.
#### 3.1.1 Optimizing over $LR_{1}\cap\widehat{Q}_{2}$
To optimize over $LR_{1}\cap\widehat{Q}_{2}$, a master linear program is
created in which the condition $x\in\widehat{Q}_{2}$ must be enforced. This
can be done by rewriting $x$ as a convex combination of its extreme points
$x_{i}$. To that end, a variable $\lambda_{i}$ is introduced for each point
$x_{i}$ used to create $\widehat{Q}_{2}$. This variable represents the weight
of the vertex $x_{i}$ in the convex combination. The Dantzig-Wolfe master
linear program can then be written as follows:
$\displaystyle(DW)\quad$ $\displaystyle\max_{x,\lambda_{i}}$ $\displaystyle
c^{T}x$ subject to $\displaystyle A_{1}x\leq b_{1}$ $\displaystyle
x=\sum_{i\in I}\lambda_{i}x_{i}$ $\displaystyle\sum_{i\in I}\lambda_{i}=1$
$\displaystyle x\in\bar{X},~{}\lambda_{i}\in\mathbb{R}^{+},\quad\forall i\in
I$
(a) Iteration 1. Initialization of $\widehat{Q}_{2}$ with points of $Q_{2}$
and optimization over $LR_{1}\cap\widehat{Q}_{2}$. Black arrow: objective
function of the master problem
(b) A direction (red arrow) in which to grow the polyhedron $\widehat{Q}_{2}$
is created with the dual variables of the master problem. The farthest point
of $Q_{2}$ in this direction is found and added to $\widehat{Q}_{2}$.
(c) Iteration 2. Re-optimization over $LR_{1}\cap\widehat{Q}_{2}$. The
farthest point of $Q_{2}$ in the dual direction is found and added to
$\widehat{Q}_{2}$.
(d) Iteration 3. Even though the optimal solution is found, due to degeneracy,
a new point of $Q_{2}$ is found in the dual direction to strengthen the dual
bound.
(e) Iteration 4. Once the current solution is the farthest point of $Q_{2}$ in
the dual direction then it is optimal and the method stops.
Figure 2: Geometric illustration of the Dantzig-Wolfe decomposition
#### 3.1.2 Finding an improving point of $Q_{2}$
In the Dantzig-Wolfe decomposition, the master problem returns the farthest
point $x^{*}$ of $LR_{1}\cap\widehat{Q}_{2}$ in the direction $c$ and one
would like to know whether this point is also the optimal for $LR_{1}\cap
Q_{2}$ or if the inner approximation $\widehat{Q}_{2}$ needs to be improved.
The dual point of view of this statement, on which is based the Dantzig-Wolfe
subproblem, is that the bound $cx\leq cx^{*}$ is valid for
$LR_{1}\cap\widehat{Q}_{2}$ and we would like to know if it also holds true
for $LR_{1}\cap Q_{2}$.
Ideas from linear programming duality theory: The following concepts are
illustrated in Figure 3. Linear programming duality informs us that because
$LR_{1}\cap\widehat{Q}_{2}$ is the intersection of two polyhedra, the bound
$cx\leq cx^{*}$ can always be decomposed as the sum of two inequalities, one
valid for $LR_{1}$ and the other valid for $\widehat{Q}_{2}$. Furthermore, the
dual variables of the master problem yield such a decomposition of the optimal
bound for $LR_{1}\cap\widehat{Q}_{2}$. Indeed, let us denote $u$, $\pi$ and
$\pi_{0}$ the optimal dual variables of the constraints $A_{1}x\leq b_{1}$,
$x=\sum_{i}\lambda_{i}x_{i}$ and $\sum_{i}\lambda_{i}=1$, respectively. By
construction of the dual of the master problem (whose derivation we let to the
reader), $uA_{1}x\leq ub_{1}$ is valid for $LR_{1}$, $\pi x\leq\pi_{0}$ is
valid for $\widehat{Q}_{2}$ and these two inequalities sum to the optimal
bound $cx\leq cx^{*}$ over $LR_{1}\cap\widehat{Q}_{2}$ (note: the equality
$ub+pi_{0}=cx^{*}$ follows thanks to the strong duality theorem of linear
programming matouvsek2007understanding (37)).
Figure 3: The optimal bound (black) on the optimization direction (black
arrows) over the intersection (red fill) of $LR_{1}$ (grey) and
$\widehat{Q}_{2}$ (red) can be decomposed as the sum of a valid inequality
(blue) for $LR_{1}$ and a valid inequality (red) for $\widehat{Q}_{2}$.
The Dantzig-Wolfe subproblem: In order to show that the solution $x^{*}$ of
the master problem is not the farthest point of $LR_{1}\cap Q_{2}$ in the
direction $c$, one must at least prove that the bound $cx\leq cx^{*}$ implied
by the dual variables of the master problem is not valid for $LR_{1}\cap
Q_{2}$. However, if the inequality $\pi x\leq\pi_{0}$ were valid for $Q_{2}$,
because we know that $uA_{1}x\leq ub_{1}$ is valid for $LR_{1}$ and that these
two inequalities sum to the bound $cx\leq cx^{*}$ then this bound would be
valid for $LR_{1}\cap Q_{2}$. Thus, the goal of the sub-problem is to show
that the inequality $\pi x\leq\pi_{0}$ is not valid for $Q_{2}$. To that end,
the subproblem is tasked to find the farthest point of $Q_{2}$ in the
direction $\pi$. If this point violates $\pi x\leq\pi_{0}$ then we have found
a point of $Q_{2}$ violating the inequality. The point can then be added to
the master problem to grow the inner approximation $\widehat{Q}_{2}$ and at
least prevent the master problem from yielding the same dual variables again.
Otherwise, the current solution is optimal because the bound implied by the
dual variables is valid for $LR_{1}\cap Q_{2}$.
#### 3.1.3 A note on degeneracy
We have seen above that the dual variables of the master problem imply a bound
on the value of its objective function. This bound can be considered a
certificate that the current value of the master problem is optimal.
Meanwhile, a linear program such as a Dantzig-Wolfe master problem is said to
be degenerate when it has several dual optimal solutions. Each of these dual
solutions is a certificate of optimality for the current value of the master
problem. Thus, to improve the value of the master problem, one must invalidate
each of these certificates. However, the subproblem of the Dantzig-Wolfe
procedure only invalidates one of these certificates without any guarantee
about the one implied by the other dual solutions. Thus the Dantzig-Wolfe
decomposition is susceptible to having many iterations without improvement of
the objective function when the master problem is highly degenerate. This can
slow down the convergence of the method considerably.
### 3.2 Fenchel decomposition
In Fenchel decomposition, instead of growing an inner approximation of
$Q_{2}$, an outer approximation of the polyhedron $Q_{2}$ is refined to enable
the optimization over $LR_{1}\cap Q_{2}$. Such an outer approximation can use
any collection of inequalities valid for $Q_{2}$. The decomposition is
illustrated in Figure 4 and proceeds as follows:
1. 1.
Initialize an outer approximation $\widecheck{Q}_{2}$ with inequalities valid
for $Q_{2}$; typically one can take $\widecheck{Q}_{2}=LR_{2}$.
2. 2.
Optimize over $LR_{1}\cap\widecheck{Q}_{2}$ and recover a solution
$\widecheck{x}$.
3. 3.
Search for a cut separating $\widecheck{x}$ from $Q_{2}$.
4. 4.
If such a cut exists, add the cut to $\widecheck{Q}_{2}$ and go to Step 2.
5. 5.
Else: $\widecheck{x}$ is the optimal solution over $LR_{1}\cap Q_{2}$. Stop.
(a) Initialization of $\widecheck{Q}_{2}$ with $LR_{2}$ and optimization over
$LR_{1}\cap\widecheck{Q}_{2}$. Black arrow: objective function of the master
problem
(b) Creation of a cut separating $Q_{2}$ from the solution of the optimization
over $LR_{1}\cap\widecheck{Q}_{2}$
(c) After addition of the cut to $\widecheck{Q}_{2}$, re-optimization over
$LR_{1}\cap\widecheck{Q}_{2}$.
(d) Creation of a cut separating $Q_{2}$ from the solution of the optimization
over $LR_{1}\cap\widecheck{Q}_{2}$
(e) Once enough cuts have been added to the approximation $\widecheck{Q}_{2}$,
the solution found cannot be separated from $Q_{2}$ and is thus optimal.
Figure 4: Geometric illustration of the Fenchel decomposition
In the second step, in order to optimize over $LR_{1}\cap\widecheck{Q}_{2}$,
the following Fenchel master problem is used:
$\displaystyle(F)\quad$ $\displaystyle\max_{x}$ $\displaystyle c^{T}x$ subject
to $\displaystyle A_{1}x\leq b_{1}$ $\displaystyle\pi
x\leq\pi_{0}\quad\forall(\pi,\pi_{0})\in\mathcal{C}$ $\displaystyle
x\in\bar{X}$
where $\mathcal{C}$ is the set of cuts describing $\widecheck{Q}_{2}$.
The main challenge in the Fenchel decomposition is to generate a cut $\pi
x\leq\pi_{0}$ separating the solution of the Fenchel master problem
$\widecheck{x}$ from the polyhedron $Q_{2}$. The classical approach to
generating such cuts is based on a linear program as described in Section 4.
## 4 The Fenchel separation subproblem and its normalizations
In the Fenchel decomposition, a cut $\pi x\leq\pi_{0}$ separating the solution
of the Fenchel master problem $\widecheck{x}$ from the polyhedron $Q_{2}$ must
be found. Such a cut can be created by finding a solution of non-negative
value of the following separation linear program:
$\displaystyle(S)\quad$ $\displaystyle\max_{\pi,\pi_{0}}$
$\displaystyle\pi\widecheck{x}-\pi_{0}$ subject to $\displaystyle\pi
x_{i}\leq\pi_{0},\quad\forall x_{i}\in Q_{2}$
$\displaystyle\pi,\pi_{0}\in\mathbb{R}$
where the objective maximizes the violation of the generated cut by
$\widecheck{x}$ while the constraints ensure that the cut is valid for every
point of the polyhedron $Q_{2}$. However, this separation problem has too many
constraints to consider them all explicitly. It is thus initially solved with
a subset of its constraints ensuring that the generated cut is valid for a few
points of $Q_{2}$. New points will progressively be taken into account in the
constraints until the validity of the cut for the whole polyhedron can be
ensured. To check if a cut $\bar{\pi}x\leq\bar{\pi_{0}}$ is valid for $Q_{2}$,
we search for the point of $Q_{2}$ that most violates the cut. This can be
done with a call to the optimization oracle ($O$) for the linear function
associated with $\bar{\pi}$. If the point returned by ($O$) violates the cut,
it is added as a constraint to ($S$). Otherwise, the cut
$\bar{\pi}x\leq\bar{\pi_{0}}$ is the optimal solution of ($S$) and maximizes
the separation of the solution of the Fenchel master problem from $Q_{2}$.
Another issue preventing the direct resolution of the problem ($S$) is that
its solution space is a cone. Indeed, if a cut $\pi x\leq\pi_{0}$ is valid for
all the points of $Q_{2}$ then so is the cut $\alpha\pi x\leq\alpha\pi_{0}$
for any non-negative constant $\alpha$. Thus, the problem ($S$) is often
unbounded. To prevent this, a normalization process must be performed which
classically consists in adding a constraint to the separation problem ($S$)
that will make the set of possible coefficients of the generated cut bounded.
The choice of this normalization greatly impacts the generated cut as well as
the convergence speed of the Fenchel decomposition. In addition, the
normalization may have an effect on the quality of the cut obtained, for
instance by favoring (or not) the obtention of facets of the polyhedron
$Q_{2}$. As we will see in the following sections, the normalization also
impacts the dual problem of ($S$) which becomes “Find the point of $Q_{2}$
minimizing some criterion of proximity to $\widecheck{x}$". It is often more
intuitive to interpret the impact of a normalization on the dual and therefore
we use this approach for our geometric interpretations.
### 4.1 Normalization $\|\pi\|\leq 1$
boyd1995convergence (12) analyzes the normalization consisting in adding to
the separation problem the constraint $\|\pi\|\leq 1$ for any norm (i.e. a
function that satisfies the triangular inequality and for which for every
$x\in\mathbb{R}^{n}$ and $\lambda\in\mathbb{R}$ satisfies $\|\lambda
x\|=|\lambda|\|x\|$ and $\|x\|=0\Rightarrow x=0$). With this choice, we can
consider that the optimized quantity is
$\frac{\pi\widecheck{x}-\pi_{0}}{\|\pi\|}$, i.e. the distance from
$\widecheck{x}$ to the hyperplane $\pi x=\pi_{0}$. Note that this distance is
not measured in the norm $\|.\|$ but in its dual norm $\|.\|^{*}$:
$\|\lambda\|^{*}=\ max_{\|x\|\leq 1}\lambda x$. In particular, note that the
norms $\|.\|_{1}$ and $\|.\|_{\infty}$ are dual and that the norm $\|.\|_{2}$
is self-dual. When using this normalization, the dual problem ($D$) becomes
"Find the point of $Q_{2}$ closest to $\widecheck{x}$ in the sense of the dual
norm $\|.\|^{*}$". This point is located on the border of $Q_{2}$ and
corresponds to the point of contact between $Q_{2}$ and the smallest sphere
centered in $\widecheck{x}$ touching $Q_{2}$. The generated cut is then a
tangent cut to $Q_{2}$ and to the sphere at the point of contact. One
important result of the work of boyd1995convergence (12) is that the
polyhedron $Q_{2}$ can be computed in finite time with this normalization. On
the other hand, the generated cuts are not always facets of the polyhedron
$Q_{2}$. All these geometric interpretations are illustrated in the
$\|.\|_{2}$ norm in Figure 5.
$\widecheck{x}$ (a) Cut associated to a $\|.\|_{2}$ normalization: a facet of
$Q_{2}$ is generated
$\widecheck{x}$ (b) Cut associated to a $\|.\|_{2}$ normalization: the
generated cut is not a facet of $Q_{2}$
Figure 5: Example of cuts generated with a $\|.\|_{2}$ normalization
### 4.2 Normalizations guaranteeing the generation of facets
We are now interested in normalizations guaranteeing the generation of facets
of $Q_{2}$. To do this, we start by presenting a theorem linking the facets of
$Q_{2}$ and the extreme rays of the cone of the solutions of the separation
problem ($S$). This theorem can be found in conforti2019facet (17)
(Proposition 1). In this theorem, $cone(X)$ (resp. $vect(X)$) is the set of
all conical (resp. linear) combinations of elements of a set $X$.
###### Theorem 4.1
Let P be a nonempty polyhedron. Let $(\pi^{j}x=\pi_{0}^{j})_{j\in J^{=}}$ be a
non-redundant representation of the affine envelope of P and let
$(\pi^{j}x\leq\pi_{0}^{j})_{j\in J^{\leq}}$ be the set of facets of P. Then
the set of valid cuts for P is:
$\mathcal{C}(P)=cone\left(\begin{pmatrix}0\\\
1\end{pmatrix},\begin{pmatrix}\pi^{j}\\\ \pi_{0}^{j}\end{pmatrix}_{j\in
J^{\leq}}\right)+vect\left(\begin{pmatrix}\pi^{j}\\\
\pi_{0}^{j}\end{pmatrix}_{j\in J^{=}}\right)$
In addition, all these vectors are necessary for the previous representation.
This theorem states that apart from the trivial ray $(0,1)$, the extreme rays
of the cone of the solutions of ($S$) are associated with the facets of the
polyhedron $Q_{2}$. When the polyhedron is not fully dimensional (_i.e._
$J^{=}\neq\emptyset$), the cuts corresponding to $(\pi^{j},\pi_{0}^{j})_{j\in
J^{=}}$ are the improper faces of $Q_{2}$. Let us suppose for the following
that $Q_{2}$ is full-dimensional. This happens once all the improper faces
have been generated.
A sufficient condition ensuring that the cuts generated are facets of $Q_{2}$
is to use a normalization condition which can be applied by adding a unique
linear constraint to the separation problem. If only one linear constraint is
added to ($S$), then all the extreme points of the new solution space
correspond to old extreme rays and therefore to facets of $Q_{2}$. If the
problem ($S$) is indeed made bounded by the addition of the linear constraint,
then it suffices to find an optimal vertex to guarantee the construction of a
facet. More generally, it is possible to add several linear constraints, as
long as they do not intersect inside the cone of solution of ($S$). For
example, adding a constraint $-1\leq f(\pi,\pi_{0})\leq 1$, where $f$ is a
linear function, corresponds to adding two linear constraints that do not
intersect. The cuts generated will then induce facets of $Q_{2}$. When, for
reasons specific to the problem, we know that valid cuts satisfy $\pi\geq 0$
(for example if $Q_{2}$ is a knapsack polyhedron), then the condition
$\|\pi\|_{1}\leq 1$ reduces to $\sum_{i}\pi_{i}\leq 1$ and can be used to
generate facet inducing inequalities.
Normalization of $\pi_{0}$: A normalization guaranteeing the generation of
facets is $|\pi_{0}|\leq$ 1\. This normalization is studied in detail in
conforti2019facet (17). It makes the problem ($S$) bounded if and only if
$\widecheck{x}$ can be written as a combination of elements in $Q_{2}$ using
only non-negative coefficients. Indeed, let us look at the impact of
normalization on the dual problem ($D$) which becomes:
$\displaystyle(D)\quad$ $\displaystyle\min_{\lambda_{i},~{}z}$
$\displaystyle|z|$ subject to $\displaystyle\sum_{i\in
I}\lambda_{i}x_{i}=\widecheck{x}$ $\displaystyle\sum_{i\in I}\lambda_{i}=1+z$
$\displaystyle z\in\mathbb{R},~{}\lambda_{i}\in\mathbb{R}^{+},\quad\forall
i\in I$
In simple words, this problem can be interpreted as follows: find a
combination with non-negative coefficients of elements of $Q_{2}$ equal to
$\widecheck{x}$ whose sum of the coefficients is as close as possible to 1.
The separation problem ($S$) is bounded if and only if its dual ($D$) is
feasible. However, the problem ($D$) is feasible if and only if
$\widecheck{x}$ can be written as a combination with non-negative coefficients
of elements of $Q_{2}$. This condition is naturally reached when the origin is
in the interior of $Q_{2}$ since the set of combinations with non-negative
coefficients of elements of $Q_{2}$ is then the entire space. If we know a
point in the interior of $Q_{2}$, it is possible to ensure this condition by
translating the problem to place the origin on this interior point. Once the
origin is inside $Q_{2}$, the generated cut is a facet of $Q_{2}$ intersecting
the segment connecting $\widecheck{x}$ to the origin.
Directional normalization of $\pi$: A second normalization, studied by
bonami2003etude (9) and guaranteeing the generation of facets, consists in
bounding the coefficients of $\pi$ in a given direction using the constraint
$|(\widehat{x}-\widecheck{x})\pi|\leq 1$ where $\widehat{x}$ is an arbitrary
point in $\mathbb{R}^{n}$. In order to determine when this normalization makes
the problem ($S$) bounded, let us look at its impact on the dual problem ($D$)
which becomes:
$\displaystyle(D)\quad$ $\displaystyle\min_{\lambda_{i},~{}z}$
$\displaystyle|z|$ subject to $\displaystyle\sum_{i\in
I}\lambda_{i}x_{i}=z\widehat{x}+(1-z)\widecheck{x}$ $\displaystyle\sum_{i\in
I}\lambda_{i}=1$ $\displaystyle
z\in\mathbb{R},~{}\lambda_{i}\in\mathbb{R}^{+},\quad\forall i\in I$
This problem can be interpreted as follows: find the point of $Q_{2}$ closest
to $\widecheck{x}$ on the line containing $\widehat{x}$ and $\widecheck{x}$.
The problem ($S$) is bounded when its dual ($D$) is feasible which happens
when there is a point belonging to both $Q_{2}$ and to the previous line. This
condition can be fulfilled for instance by choosing $\widehat{x}$ equal to a
known point of $Q_{2}$. In this case, the generated cut is a facet of $Q_{2}$
intersecting the segment connecting $\widecheck{x}$ and $\widehat{x}$. This is
illustrated in Figure 6(a). Unlike the previous normalization, the known point
$\widehat{x}$ does not need to be in the interior of $Q_{2}$. However, note
that when it is on the border of $Q_{2}$, the optimal value of the problem
($S$) is bounded but the set of optimal solutions may be unbounded. Indeed,
let us consider the following example illustrated in Figure 6(b).
Example: Suppose that $\widehat{x}$ is located on a vertex of $Q_{2}$ and that
$\widecheck{x}$ is located such that the line $(\widecheck{x},\widehat{x})$
intersects $Q_{2}$ only in $\widehat{x}$. Let $\pi x\leq\pi\widehat{x}$ be an
optimal cut separating $\widecheck{x}$ from $Q_{2}$ thus passing through
$\widehat{x}$. Let $\pi^{\bot}$ be the projection of $\pi$ on the orthogonal
of the vector space induced by the vector $\widecheck{x}-\widehat{x}$. Finally
let us assume that the cut $\pi^{\bot}x\leq\pi^{\bot}\widehat{x}$ is valid for
$Q_{2}$. This is for instance the case in the 2D example illustrated in Figure
6(b) as the hyperplane $\pi^{\bot}x=\pi^{\bot}\widehat{x}$ is, in this case,
the line $(\widecheck{x},\widehat{x})$. Then for all $\alpha\geq 0$, the cut
$(\pi+\alpha\pi^{\bot})x\leq(\pi+\alpha\pi^{\bot})\widehat{x}$ is also an
optimal cut for the separation problem. Indeed, it is valid for $Q_{2}$ as a
non-negative combination of valid cuts for $Q_{2}$. Moreover, the violation of
this cuts by $\widecheck{x}$ is the same as the violation of $\pi
x\leq\pi\widehat{x}$ because by construction of $\pi^{\bot}$ the product
$\pi^{\bot}(\widecheck{x}-\widehat{x})$ is equal to zero. Thus, in this
example, the separation problem is bounded because its dual is feasible but
the set of optimal cuts is unbounded because the norm of
$\pi+\alpha\pi^{\bot}$ approaches infinity when $\alpha$ does the same.
Despite this unbounded set of solutions, only the facets of the polyhedron
$Q_{2}$ are vertices of the solution space of ($S$). Thus, if the algorithm
solving ($S$) always returns a vertex, the generated cut will always be a
facet of $Q_{2}$.
$\widecheck{x}$$\widehat{x}$ (a) Cut associated with directional normalization
$\pi$$\pi^{\bot}$$\widecheck{x}$$\widehat{x}$ (b) Different normals of optimal
cuts for the directional normalization: the norm of optimal normals can tend
to infinity
Figure 6: Examples of results obtained when using directional normalization.
The above normalizations have been presented several times in the literature
and are applicable in the general case. However, certain problems can admit
normalizations particularly adapted to their structure. One can find such
normalizations for example in the separation problems of disjunctive
programming balas2002lift (2). In addition, the unsplittable flow problem,
which will be used in our experimental study, admits a normalization that
seems natural and which be presented in Section 7.1.
## 5 A new approach for the Fenchel sub-problem
In this section, we present a new approach to solve the separation problem of
the Fenchel decomposition when the directional normalization presented in
Section 4 is used.
### 5.1 Presentation of the method
The proposed method is described in Algorithm 1 and illustrated in Figure 7.
Note that the cut associated with a directional normalization toward
$\widehat{x}$ is the same for $\widecheck{x}$ than for any other point of the
segment $(\widecheck{x},\widehat{x})$ not belonging to $Q_{2}$. The underlying
idea is to compute intermediate cuts using an alternative normalization. By
projecting a point $x^{\prime}$ (initially equal to $\widecheck{x}$) onto
these intermediate cuts, the algorithm gradually shifts the point $x^{\prime}$
along the segment $(\widecheck{x},\widehat{x})$ in the direction of
$\widehat{x}$. Once $x^{\prime}$ reaches the frontier of $Q_{2}$ the procedure
stops. Indeed, the last cut generated is the one associated with the
directional normalization toward $\widehat{x}$ since it contains the
intersection between the segment $(\widecheck{x},\widehat{x})$ and the
frontier of $Q_{2}$.
Algorithm 1 New approach for solving the Fenchel sub-problem
1:a polyhedron $Q_{2}$, a point $\widecheck{x}$ to separate from $Q_{2}$, a
point $\widehat{x}$ belonging to $Q_{2}$
2:a cut $C$ separating $\widecheck{x}$ from $Q_{2}$ and containing the
intersection point the frontier of $Q_{2}$ and the segment
$(\widecheck{x},\widehat{x})$, a list $LS$ of vertices of $Q_{2}$ satisfying
the cut $C$ to equality
3:Set $x^{\prime}$ equal to $\widecheck{x}$
4:while $x^{\prime}\notin Q_{2}$ do
5: $C$, $LS$ = Secondary_separation($Q_{2}$, $x^{\prime}$)
6: Set $x^{\prime}$ equal to the intersection point between the segment
$(\widecheck{x}$, $\widehat{x})$ and the cut $C$
7:return $C$, $LS$
$\widecheck{x}$$\widehat{x}$ (a) The goal is to separate the blue point using
a directional normalization with the red point as an interior point
$\widecheck{x}$$\widehat{x}$ (b) The separation problem is solved with a
secondary normalization (here the norm $\|.\|_{2}$)
$\widecheck{x}$$\widehat{x}$ (c) The point to be separated is projected on the
generated section
$\widecheck{x}$$\widehat{x}$ (d) The separation problem is solved again with
the secondary normalization (here the norm $\|.\|_{2}$)
$\widecheck{x}$$\widehat{x}$ (e) Once the point to be separated is projected
on the border of the polyhedron, the procedure stops.
Figure 7: Example of solving the Fenchel sub-problem for directional
normalization using a secondary normalization
The advantage of this iterative approach is that the linear separation program
($S$) is never directly solved with the directional normalization, which often
presents numerical instabilities. Indeed, for this normalization, the
separation problem ($S$) seeks a face of the polyhedron $Q_{2}$ intersecting
the segment $(\widecheck{x},\widehat{x})$. If, for example, this segment
intersected a facet of $Q_{2}$ by forming an angle close to $0$ with it then a
small error on the parameters of the segment or of the facet can induce a
large error on the position of the point of intersection. For this reason, the
problem ($S$) is sometimes too numerically unstable to be solved directly. The
new method presented above yields an alternative way of computing the cut
associated with the directional normalization without ever solving the
numerically unstable linear program of the direct method. Unfortunately, note
that this iterative algorithm requires the computation of the intersection
between the segment $(\widecheck{x},\widehat{x})$ and the cuts returned by the
secondary separation problem. If the segment is almost parallel to one of
these cuts, the computation of this point of intersection may still be
numerically unstable. However, this is only an intermediate point not returned
by the separation oracle. The results of the method are a cut and vertices
which are computed with a secondary normalization. In practice, this seems
sufficient as no numerical instabilities were found during our thorough
numerical campaign.
Our experimental section yields insights into the practical performance of
this new approach for the separation oracle. From a theoretical standpoint, we
provide proofs in A that the method converges in a finite number of iterations
in two cases: 1) when the secondary normalization always generates facets of
$Q_{2}$, and 2) when the secondary normalization is $\|\pi\|\leq 1$ for any
norm. These two cases cover most normalizations used in the literature
including all those mentioned in Section 4.
## 6 Coupling the Fenchel and Dantzig-Wolfe decompositions
In this section, we present a decomposition method that integrates Dantzig-
Wolfe and Fenchel decompositions. In our experimental campaign, we show that
the new method presents a much superior performance than Fenchel decomposition
alone, and is competitive against Dantzig-Wolfe on non-degenerate problems.
For degenerate problems, again the proposed method shows a superior perormance
to a classical Dantzig-Wolfe decomposition.
### 6.1 Presentation of the method
When generating a Fenchel cut, the primal variables of the separation sub-
problem ($S$) are the coefficients of the generated cut and the active
constraints correspond to the vertices of the separated polyhedron $Q_{2}$
verifying this cut at equality. Thus, the Fenchel sub-problem generates both
valid cuts for the polyhedron $Q_{2}$ and vertices of this polyhedron.
The main idea of the new decomposition method is to use two master problems
operating in tandem. The first corresponds to the Fenchel formulation ($F$) in
which the generated Fenchel cuts are added to improve an outer approximation
of $Q_{2}$. The second one is the Dantzig-Wolfe formulation ($DW$) in which
the vertices generated in the separation sub-problem are added to grow an
inner approximation of $Q_{2}$. One of the key points of the method is that
points $\widecheck{x}$ provided by the Fenchel master problem are separated
using directional normalization. This normalization requires the knowledge of
a point in the polyhedron $Q_{2}$, for which we use the solution $\widehat{x}$
computed by the Dantzig-Wolfe master problem.
The steps of the algorithm, illustrated in Figure 8, are as follows:
1. 1.
Initialize an inner approximation $\widehat{Q}_{2}$ of the polyhedron $Q_{2}$
as in the Dantzig-Wolfe decomposition and an outer approximation
$\widecheck{Q}_{2}$ as in the Fenchel decomposition.
2. 2.
Optimize over $LR_{1}\cap\widecheck{Q}_{2}$ using a Fenchel master problem: a
solution $\widecheck{x}$ is obtained whose value is an upper bound of the
problem.
3. 3.
Optimize over $LR_{1}\cap\widehat{Q}_{2}$ using a Dantzig-Wolfe master
problem: a solution $\widehat{x}$ is obtained whose value is a lower bound of
the problem.
4. 4.
If the two bounds are equal: end of the algorithm
5. 5.
With the Fenchel separation problem ($S$), separate the point $\widecheck{x}$
from $Q_{2}$ using directional normalization with $\widehat{x}$ as the
interior point: a cut is obtained as well as vertices of $Q_{2}$.
6. 6.
Add the cut to the outer approximation $\widecheck{Q}_{2}$ and the vertices to
the inner approximation $\widehat{Q}_{2}$. Then, go to Step 2.
(a) The outer approximation $\widecheck{Q}_{2}$ and the inner approximation
$\widehat{Q}_{2}$ are computed. The solutions $\widecheck{x}$ and
$\widehat{x}$ are computed by solving over the intersection of $LR_{1}$ and
the corresponding approximation.
(b) The solution $\widecheck{x}$ is separated from $Q_{2}$ using the
directional normalization toward $\widehat{x}$. This separation problem
returns a cut and vertices of the polyhedron $Q_{2}$.
(c) The cut is added to the outer approximation $\widecheck{Q}_{2}$ and the
vertices to the inner approximation $\widehat{Q}_{2}$.
(d) After enough iterations, the value of the solutions given by the two
approximations coincide and the method stops.
Figure 8: Illustration of the Dantzig-Wolfe-Fenchel decomposition
Although this new decomposition method uses ideas taken from both the Dantzig-
Wolfe and the Fenchel decompositions, the way in which they operate yields a
few remarkable observations and interpretations that are worth describing:
* •
Intuitively, the inner and outer approximations constructed disagree on the
location of the boundary of $Q_{2}$ on the segment
$(\widecheck{x},\widehat{x})$. The separation problem finds the exact position
of the frontier and gives information to the two approximations so that they
can approximate exactly this part of the frontier.
* •
Another way of looking at the method is to say that it gradually improves the
point $\widehat{x}$. At each iteration, the algorithm tests a direction of
potential improvement $\widecheck{x}-\widehat{x}$. The method then finds
either new vertices of $Q_{2}$ allowing to improve $\widehat{x}$ or a facet of
$Q_{2}$ passing through $\widehat{x}$ proving that it is not possible to
improve $\widehat{x}$ in this direction. This interaction between an exterior
point and an interior point of $Q_{2}$ is reminiscent of the in-out separation
proposed for the Benders decomposition ben2007acceleration (6).
* •
Compared to a classic Fenchel decomposition, this method concentrates its cut
generation around the point $\widehat{x}$. It thus refines the knowledge of
the polyhedron $Q_{2}$ around this point. On the other hand, a classical
Fenchel decomposition might spend multiple iterations searching for cuts in
regions of the solution space that end up being far from the optimum.
* •
Compared to a classic Dantzig-Wolfe method, this method devotes more time to
the resolution of its sub-problem which allows it to generate a greater number
of vertices to add to the master problem.
### 6.2 Degeneracy
The Dantzig-Wolfe decomposition is known to present convergence issues when
its master problem is highly degenerate. Our experiments reveal that the
method presented in this section does not suffer from the same issues. In the
following, we give a partial theoretical explanation of the absence of
degeneracy issues in the new method.
A linear program is said to be degenerate when it has multiple dual optima. We
have seen in Section 3.1 that each dual solution of the Dantzig-Wolfe master
problem implies a bound on the value of its objective function $cx\leq c_{0}$
that certifies the optimality of its current primal solution on the solution
set $LR_{1}\cap\widehat{Q}_{2}$. Thus, in order to improve the solution of the
master problem, one needs to be able to separate each of these dual solutions
from the polyhedron $LR_{1}\cap Q_{2}$. However, the sub-problem of the
Dantzig-Wolfe decomposition may only separate one such dual vector which
explains why the Dantzig-Wolfe decomposition is so affected by degeneracy.
The new proposed Dantzig-Wolfe-Fenchel decomposition seems to be unaffected by
the degeneracy of its Dantzig-Wolfe master problem. One possible explanation
of this behavior is supported in the following observation and proposition.
First, let us remark that the new decomposition neither computes nor uses any
dual information, therefore showing no sensitivity to the quality of the
duals. Second:
###### Proposition 1
At each iteration, the sub-problem invalidates either all the dual solutions
of the Dantzig-Wolfe master program or none of them.
###### Proof
The sub-problem finds the farthest point in $Q_{2}$ along the segment
$(\widehat{x},\widecheck{x})$. On the one hand, if this farthest point
coincides with $\widehat{x}$ then it does not invalidate any of the bounds
implied by the dual solutions because $\widehat{x}$ belongs to
$LR_{1}\cap\widehat{Q}_{2}$. On the other hand, suppose the farthest point
$x^{*}$ does not coincide with $\widehat{x}$. First, by construction, $x^{*}$
belongs to $Q_{2}$ but also to $LR_{1}$ because it is a convex combination of
$\widehat{x}$ and $\widecheck{x}$, both belonging to $LR_{1}$. Second,
assuming that $\widecheck{x}$ has a strictly better objective value than
$\widehat{x}$ (which is always the case except when the method is about to
terminate) then $x^{*}$ also has a strictly better objective value than
$\widehat{x}$. Thus, it must invalidate all the bounds implied by the dual
solutions of the Dantzig-Wolfe master problem (remember that these bounds
certify that no better solution than $\widehat{x}$ exists). ∎
Therefore, whether the Dantzig-Wolfe master problem admits multiple dual
optima or not does not influence the algorithm’s capacity to find a strictly
improving point.
Although we presented all the decomposition methods as if only one polyhedron
was decomposed at once (e.g. a block of a block diagonal matrix), in practice
several polyhedra are decomposed at the same time (e.g. all the blocks of a
block diagonal matrix). In this case, Proposition 1 does not hold. However,
the ideas discussed in its proof may still impact positively the practical
computations. In any case, the decomposition method still does not use any
dual information which renders it oblivious to the number of dual solutions of
its Dantzig-Wolfe master problem.
## 7 Application to the unsplittable flow problem
In the unsplittable flow problem (UFP), one is given a weighted directed graph
$\mathcal{G}=(V,A,c)$ where $c_{a}$ is the capacity of the arc $a$ for every
$a\in A$. We are also given a family $K$ of commodities, each composed of an
origin $o(k)$, a destination $d(k)$, and a demand $D_{k}$, for every $k\in K$.
Each commodity has to be routed from its origin to its destination through an
unique path. We consider the problem where the capacity constraints are soft,
meaning that they can be violated at a certain unit penalty. The objective is
to design routing paths for every commodity on the network so as to minimize
the sum of the violations of the arcs’ capacities.
The UFP is an extensively studied NP-hard variant of the classic maximum-flow
problem. It has multiple applications, as for instance in telecommunication
networks (_e.g._ optical networks, telecommunication satellites
rivano2002lightpath (18, 31)), and logistics farvolden1993primal (23). The
early work of belaidouni2007minimum (5) studied the UFP from a polyhedral
perspective, proposing cutting planes to strengthen the linear relaxation of a
three-index model that uses variables $x_{ijk}$ for every arc $(i,j)$ in the
network and every commodity $k$ to be transported. park2003integer (41) also
strengthen the linear relaxation by applying the Dantzig-Wolfe decomposition
to the capacity constraints of the UFP. The resulting relaxation is as strong
as if all the inequalities valid for the capacity constraints of the problem
were added. Thus, faster decomposition methods able to compute the Dantzig-
Wolfe linear relaxation could yield improvements in the resolution of the UFP.
We will consider an arc-path formulation where the meaning of the variables is
the following:
* •
$x_{pk}$ indicates whether commodity $k$ uses path $p$ to push its flow,
* •
$\Delta_{a}$ represents the overflow on arc $a$.
In addition to the decision variables, we also denote, for a given commodity
$k\in K$, $P_{k}$ the set of all $o(k)$-$d(k)$-paths in $G$. For every $p\in
P_{k}$ and arc $a\in A$, we define a constant $\alpha_{ap}$ that takes the
value 1 iff $p$ uses the arc $a$. The path formulation of the UFP is the
following:
$\displaystyle\min_{x_{pk},\Delta_{a}}$ $\displaystyle\sum_{a\in A}\Delta_{a}$
(9a) subject to $\displaystyle\sum_{p\in P_{k}}x_{pk}=1$ $\displaystyle\forall
k\in K,$ (9b) $\displaystyle\sum_{k\in K}\sum_{p\in P_{k},a\in
p}x_{pk}D_{k}\leq c_{a}+\Delta_{a}$ $\displaystyle\forall a\in A,$ (9c)
$\displaystyle x_{pk}\in\\{0,1\\},~{}\Delta_{a}\in\mathbb{R}^{+}$
$\displaystyle\forall p\in P_{k},~{}\forall k\in K,~{}\forall a\in A$ (9d)
The objective function minimizes the sum of the overflows on the arcs.
Equation (9b) ensures that exactly one path is chosen for each commodity.
Equation (9c) corresponds to the soft capacity constraints. It ensures that
any overflow on an arc $a$ is recorded on the corresponding variable
$\Delta_{a}$. The fact that $x_{pk}\in\\{0,1\\}$ ensures that the flow is
unsplittable.
The polyhedron associated with the capacity constraints does not have the
integrality property and its relaxation can thus be tightened with any of the
previously discussed decomposition methods. If we denote $f_{a}^{k}=\sum_{p\in
P_{k},a\in p}x_{pk}$, this polyhedron can be written as follows:
$\left\\{f_{a}^{1},...,f_{a}^{k}\in[0,1],o_{a}\in\mathbb{R}^{+}\bigg{|}\sum_{k\in
K}f_{a}^{k}D^{k}\leq c_{a}+o_{a}\right\\}.$
Studies have been carried out on the structure, the cut selection, and the
strengthening of the linear relaxation of this type of polyhedron by
Marchand19990 (36) as well as in the more general framework of linear programs
in mixed variables dash2011mixed (21, 24, 16). Moreover, optimization methods
on this polyhedron have been studied by buther2012reducing (14, 34, 46, 28,
35).
In the following, we present a specialized normalization for this polyhedron.
This normalization guarantees the generation of facets and will be used in our
implementation of the Fenchel decomposition and as secondary normalization in
the new procedure to solve the Fenchel sub-problem presented in Section 4.
Then, we describe our implementation of the oracle that optimizes a linear
function on the polyhedron associated with the capacity constraints of the
UFP.
### 7.1 Natural normalization for unsplittable flows
Let $\widecheck{f}=(\widecheck{f}^{k})_{k\in K}\in[0,1]^{|K|}$ be a flow
distribution for each commodity which induces an overflow $\widecheck{o}$ on a
given arc. In this section, we assume that the arc is fixed and will therefore
drop the arc index for the sake of simplicity. In the context of unsplittable
flows, vertices of the polyhedron $Q_{2}$ correspond to commodity patterns
which will be indexed by a superscript $g$. A naturally occurring question is:
how does this distribution break down into a combination of commodity patterns
inducing a minimum capacity overflow? This question can be solved using the
following linear program:
$\displaystyle(D)\quad$ $\displaystyle\min_{\lambda^{g},z}$ $\displaystyle z$
subject to $\displaystyle\sum_{g\in G}\lambda^{g}f^{g}=\widecheck{f}$
$\displaystyle\sum_{g\in G}\lambda^{g}o^{g}=\widecheck{o}+z$
$\displaystyle\sum_{g\in G}\lambda^{g}=1$
$\displaystyle\lambda^{g}\in\mathbb{R}^{+},~{}z\in\mathbb{R}^{+}\quad\forall
g\in G$
where $\lambda^{g}$ is the coefficient in the decomposition associated with a
commodity pattern $f^{g}$ inducing an overflow $o^{g}$.
Now the dual of this decomposition program is the following program:
$\displaystyle(S)\quad$ $\displaystyle\max_{\pi,\pi_{o}\pi_{0}}$
$\displaystyle\pi\widecheck{f}+\pi_{o}\widecheck{o}-\pi_{0}$ subject to
$\displaystyle\pi f^{g}+\pi_{o}o^{g}\leq\pi_{0},\quad\forall(f^{g},o^{g})\in
Q_{2}$ $\displaystyle\pi_{o}\leq-1$
$\displaystyle\pi,\pi_{0},\pi_{o}\in\mathbb{R}$
This program corresponds exactly to the problem ($S$) of separating the point
$\widecheck{x}=(\widecheck{f_{1}},...,\widecheck{f}_{|K|},\widecheck{o})$ with
a constraint of normalization imposing that the coefficient $\pi_{o}$
associated with the overflow variable satisfies $\pi_{o}\leq-1$. This
normalization constraint is what we will call the natural normalization for
the unsplittable flow problem. This normalization is very close to a
particular case of directional normalization for the direction
$\widehat{x}-\widecheck{x}=(0,...,0,1)$. Just like directional normalization,
the natural normalization guarantees the generation facets because it is
imposed using a single linear constraint.
### 7.2 Knapsack oracle resolution
All the decomposition methods presented in this work assume that there exists
an efficient algorithm capable of optimizing a linear function on the
polyhedron $Q_{2}$. In this section, we detail the problem solved by the
oracle in the context of unsplittable flows.
In the version of the unsplittable flow problem that we are studying, the
capacity constraints do not require that the flow of commodities respect the
capacities $c_{a}$ of the arcs. However, the overflow must be stored in a
variable $o_{a}$. Thus, the polyhedron of variables satisfying the soft
capacity constraint associated with the arc $a$ is written:
$\left\\{(f_{a}^{k}\in\\{0,1\\})_{a\in A,k\in
k},o_{a}\in\mathbb{R}^{+}\Bigg{|}\sum_{k\in K}f_{a}^{k}D^{k}\leq
c_{a}+o_{a}\right\\}.$
The optimization of a linear function whose coefficients are $(\pi^{k})_{k\in
K}$ and $-\pi_{o}$ on this polyhedron can be written as follows:
$\displaystyle(O_{a})\quad$ $\displaystyle\max_{f_{a}^{k},o_{a}}$
$\displaystyle\sum_{k\in K}\pi^{k}f_{a}^{k}-\pi_{o}o_{a}$ subject to
$\displaystyle\sum_{k\in K}f_{a}^{k}D^{k}\leq c_{a}+o_{a}$ $\displaystyle
f_{a}^{k}\in\\{0,1\\},~{}o_{a}\in\mathbb{R}^{+}$
This problem can be solved as a sequence of two 0-1 knapsack problems using a
case disjunction. This method was presented by buther2012reducing (14) and is
recalled in B. In our experiments, we use the MINKNAP algorithm proposed by
pisinger1997minimal (43) to solve the two associated knapsack problems.
## 8 Experimental study
In this section, we present an experimental comparison of different
decomposition methods. The datasets and code used in this section can be
accessed at https://github.com/TwistedNerves/decomposition_paper_code. The
code was written in Python 3 and the experiments carried on an Intel Core
i9-9900K 3.60 GHz $\times 16$ cores CPU, 60 Gbit of RAM, running Ubuntu 20.10.
### 8.1 Datasets
An instance of the unsplittable flow problem is composed of a graph and a list
of commodities. The method used to create instances in our experiments is the
one presented in lamothe2021randomized (32). All the graphs used are strongly
connected random graphs. To create demands for the commodities,
lamothe2021randomized (32) used two formulas. In this work, we use the formula
that creates mainly commodities with large demands because it tends to create
instances that are harder to solve. Moreover, in each instance, all the
commodities can be unsplittably routed without exceeding the arc capacities.
Therefore, the lower bound given by the linear relaxation is optimal. In order
to create an optimality gap in the instances, we slightly modify the
capacities of some arcs as follows a number of times equal to 100 times the
number of nodes:
* •
Randomly select the origin of a commodity.
* •
Randomly select two arcs coming out of this origin.
* •
Add 1 to the capacity of one arc and subtract 1 from the ability of the other
arc.
Because of the way instances are created, all outgoing arcs from each origin
node are saturated in the solutions without overflow while the other arcs are
often non-saturated. Therefore, in most cases, transferring some capacity
between outgoing arcs of origins does not change the value of the linear
relaxation. On the other hand, this transfer of capacity can have an impact on
the value of the best unsplittable solution. Indeed, there is no longer
necessarily a combination of commodities whose sum of demands is exactly equal
to the capacity of each arc. In this case, the best unsplittable solution has
a non-zero overflow.
Another change made to the instances is that the commodities have only access
to a restricted set of paths to push their flow. The restricted set of paths
of a commodity is chosen to be the k-shortest paths from the origin to the
destination of the commodity with k = 4. Because this study explores the
strengthening of the linear relaxation of the unsplittable flow problem
through its capacity constraints, this modification should not change the
relative behavior of the tested algorithms but does make the instances much
simpler to solve which enables the testing of the different algorithms on
larger instances.
##### The datasets
Three different datasets are used during the experiments in which ten
instances are generated for each value of the varying parameter.
* •
Low maximum demand dataset: this dataset considers strongly connected random
graphs from 50 nodes to 145 nodes. The maximum commodity demand is set at
$\widecheck{d}_{max}=$ 100 and the arc capacity at 1000. This choice of
maximum demand implies that a large number of commodities can pass through
each arch. However, in our tests, the optimal solution for these instances
often does not contain overflow. Therefore, these instances do not contain an
optimality gap. We hypothesize that the large number of commodities allows
them to rearrange themselves to exactly fill the capacity of each arc.
* •
High maximum demand dataset: this dataset considers strongly connected random
graphs from 145 nodes to 1000 nodes. The maximum commodity demand is set at
$\widecheck{d}_{max}=$ 1000 and the arc capacity at 1000. Because of this
maximum demand choice, these instances contain only a small number of
commodities. However, they generally have an optimality gap which allows us to
study the evolution of the lower bounds given by the algorithms.
* •
Size of capacities dataset: this dataset considers strongly connected random
graphs of 70 nodes. The maximum demand of the $\widecheck{d}_{max}$
commodities is fixed at $1/10$ of the common capacity of the arcs which varies
from 100 to 100,000. The knapsack problem is known to have algorithms that are
pseudo-polynomial in the capacity of the knapsack. One such algorithm is the
MINKNAP algorithm we use. In the case of unsplittable flows, the capacity of
the knapsack corresponds to the capacity of the arcs. The instances of this
dataset all have the same structure (same graph size, same size of commodities
relative to the capacity of the arcs) but varying arc capacities. This impacts
the resolution time of the MINKNAP algorithm.
### 8.2 Comparison of Dantzig-Wolfe variants
A large number of works have proposed variations and improvements to Dantzig-
Wolfe decomposition. Indeed, the method presented in Section 3.1 sometimes
suffers from a slow convergence which has been associated with the following
observations pessoa2013out (42):
* •
Dual oscillations: the dual variables $\pi$ used to generate the vertices of
$Q_{2}$ perform large oscillations and do not converge monotonically toward
their optimal value.
* •
The tailing-off effect: during the last iterations, the space of the dual
solutions is only marginally reduced and the dual bound progresses very
slowly.
* •
Degenerate primal and equivalent dual solutions: The master problem ($DW$) is
regularly degenerate because it has several dual optimal solutions. The method
iterates between equivalent dual solutions without making progress on the
value of the objective function.
In order to overcome these difficulties, stabilization methods for the dual
variables have been considered. These methods can be classified into three
categories pessoa2013out (42):
* •
Penalization: the penalization methods are best interpreted by considering the
dual of the problem ($DW$). In order to stabilize the dual variables $\pi$, a
penalty $f(\|\pi-\bar{\pi}\|)$ is added to the objective function of the dual.
In this penalty, $f()$ is an increasing function, which pushes $\pi$ to stay
close to a value $\bar{\pi}$ which evolves slowly during the algorithm.
Typically, $\bar{\pi}$ is one of the values taken by the dual variables during
the previous iterations. A widely studied special case is to penalize
proportionally to $\|\pi-\bar{\pi}\|_{2}^{2}$, which is done in the Bundle
methods briant2008comparison (13).
* •
Smoothing: in smoothing methods, the dual variables $\pi$ of the problem
($DW$) are not used directly in the sub-problem in order to generate new
vertices of $Q_{2}$. We note, for the iteration $j$ of column generation,
$\pi_{j}$ the values of the variables resulting from the dual of the problem
($DW$) and $\bar{\pi}_{j}$ the values used in the sub-problem. A smoothing
method proposed by neame2000nonsmooth (39) uses the following formula:
$\bar{\pi}_{j}=\alpha\bar{\pi}_{j-1}+(1-\alpha)\pi_{j}$. This method amounts
to adding an momentum effect to the dual variables. Another method, proposed
by wentges1997weighted (44), performs a convex combination with a fixed dual
value $\bar{\pi}$, i.e. $\bar{\pi}_{j}=\alpha\bar{\pi}+(1-\alpha)\pi_{j}$.
* •
Centralization: The idea of centralization methods is that it is more
efficient to use in the sub-problem dual values located inside the dual
polyhedron rather than on an extreme vertex of the dual polyhedron. On the
other hand, such interior points are more expensive to compute than extreme
points. The interior point used in the Primal-Dual Column Generation
gondzio2013new (26) is obtained by approximately solving the problem ($DW$) by
an interior point method. Another classic point is the analytical center used
in the analytical center cutting plane method goffin2002convex (25).
In order to have a suitable comparison for the decomposition methods we
proposed in this work, we experimentally compared the following three
variations of Dantzig-Wolfe decomposition:
DW: Dantzig-Wolfe decomposition method. No stabilization of the dual variables
is used. The lower bounds are computed using the dual variables and the value
of the solution of the knapsack sub-problems.
DW-momentum: similar to the DW method except that the dual variables are
stabilized by smoothing using the formula of neame2000nonsmooth (39),
$\bar{\pi}_{t}=\alpha\bar{\pi}_{t-1}+(1-\alpha)\pi_{t}$ with the coefficient
$\alpha$ set to $0.8$.
DW-interior-point: similar to the DW method except that the Dantzig-Wolfe
master problem is solved with an interior point method in order to return a
non-optimal but centered solution. To that end, we ask the gurobi (27) solver
to solve the linear program using an interior point method with a precision of
$10^{-3}$ and without using its crossover method. However, the first time the
sub-problem fails to generate a new negative reduced cost variable, the solver
Gurobi is reset to its default settings to ensure an exact computation of the
last reduced costs. With the default parameters, the generation of columns is
no longer stabilized.
These methods are compared in Figure 9. For the rest of the experiments, we
will use the variation based on the interior point solver as it always returns
the best results in our tests.
Figure 9: Compararison of Dantzig-Wolfe decomposition variations, Low maximum
demand dataset, 70 nodes
### 8.3 The decomposition methods studied
In the following, we experimentally compare the following decomposition
methods:
Fenchel: Fenchel decomposition method. The cuts generated are added to the
linear relaxation while the generated vertices are added to a Dantzig-Wolfe
formulation. The Fenchel sub-problem is solved with the natural normalization
presented in Section 7.1. Therefore, the two master problems do not act in
tandem.
DW-Fenchel: method combining the Fenchel and Dantzig-Wolfe decompositions
presented in Section 6, the cuts generated are added to the linear relaxation
while the generated vertices are added to a Dantzig-Wolfe formulation. The
Fenchel sub-problem is solved with directional normalization with the optimal
point of the Dantzig-Wolfe formulation as the interior point. The use of this
normalization couples the two formulations.
DW-Fenchel-iterative: similar to the DW-Fenchel method except that the Fenchel
sub-problem is solved using the iterative method presented in Section 5.
DW-interior-point: Dantzig-Wolfe decomposition introduced in Section 8.2 where
the master problem is solved with an interior point method in order to return
a non-optimal but centered solution.
### 8.4 Algorithms’ parameters
Authorized paths Because the focus of this work is on the capacity constraints
and not on how to generate the paths for each commodity, each commodity is
restrained to a small set of allowed paths. This set is made up of the four
shortest paths between the origin and destination of the commodity as well as
the path used to create the commodity in the method of lamothe2021randomized
(32).
Algorithm termination condition The decomposition methods considered are
stopped when the absolute difference between their bounds is $10^{-3}$.
Pre/post-processings for the sub-problem of Fenchel: Solving directly a
Fenchel subproblem is sometimes too computationally expensive to be integrated
into a decomposition method. However, the resolution time of this subproblem
can be greatly reduced with pre/post-processing steps. Indeed, boccia2008cut
(8) showed that it is possible to solve the Fenchel separation problem by
focusing on a sub-polyhedron of $Q_{2}$ of far lesser dimension which
decreases the computing time. However, the generated cut is not directly valid
for $Q_{2}$ and one must use a lifting procedure to create a cut valid for
$Q_{2}$. These concepts are explained in C. Together, the techniques of
Dimensionality reduction and Lifting induce a drastic decrease in the
resolution time of the Fenchel subproblem. This fact was confirmed by our
experiments and the results we present in our experimental study reflect this
algorithmic choice.
### 8.5 Experimental results
We now present the results of our experimental campaign for the different
decomposition methods presented in this work. In each figure, we display the
evolution of the lower and upper bounds achieved by the algorithms as a
function of the computational time (in seconds). Note that all the displayed
values are not directly the bounds but their deviation from the value of an
optimal solution of the Dantzig-Wolfe reformulation. The plotted curves
represent the average results of the algorithms aggregated on instances using
the same parameters while the confidence intervals at $95\%$ for the mean are
plotted in semi-transparency around the main curve. A problem encountered when
generating these curves is that the algorithms only return bounds at the end
of each of their iterations but these iterations take a variable time for the
same algorithm depending on the instance. It is therefore not possible to
directly aggregate the curves using the points given at the end of each
iteration because they do not correspond to the same computing time. To obtain
points on which we can appropriately average the values, the points defining
the curves are replaced with points sampled every ten seconds by considering
that the bounds evolve linearly between two iterations. The confidence
intervals are created using the statistical method called Bootstrapping with a
number of resamplings equal to 1000. Because the Bootstrapping method is
applied independently for every ten seconds of the curves the resulting
confidence intervals have jitters. These jitters can be interpreted as the
uncertainty on the bound of the confidence intervals due to the Bootstrapping
method.
Figure 10: Legend of the Figures 11 to 17
Figure 11: Low maximum demand dataset, 70 nodes
Figure 12: Low maximum demand dataset, 90 nodes
Figure 13: Low maximum demand dataset, 145 nodes
Figure 14: High maximum demand dataset, 250 nodes
Figure 15: High maximum demand dataset, 400 nodes
Figure 16: Size of commodities dataset, capacity 100
Figure 17: Size of commodities dataset, capacity 10000
Solving the Fenchel sub-problem with a secondary normalization. The new method
of solving the Fenchel sub-problem presented in Section 5 is used in the DW-
Fenchel-iterative method. This method appears to be slightly slower than the
DW-Fenchel method which uses a direct approach for the sub-problem. On the
other hand, the direct approach sometimes fails to solve the sub-problem
because of numerical instabilities which prevent the decomposition method from
converging. For instances with 145 nodes of the Low maximum demand dataset,
this happens every 10 to 20 instances. The 10 instances presented in Figure 13
did not suffer from instability in this set of experiments thus it does not
appear in the figure. However, we were able to identify seeds for which the
instability appears. Unfortunately, these seeds appear to be hardware-
dependent thus researchers trying to reproduce the results will have to find
their own seeds.
Impact of coupling the two master problems using directional normalization.
The DW-Fenchel and DW-Fenchel-iterative methods couple the Dantzig-Wolfe and
Fenchel master problems using a directional normalization in their Fenchel
sub-problem. The impact of this coupling can be studied by comparing these
methods to the Fenchel method whose only difference is to use the natural
normalization of the unsplittable flow problem in its sub-problem. As
illustrated in Figure 13, one notes that all the methods are similar during
the first iterations. However, the Fenchel method stalls rapidly. On the other
hand, this is not the case with the methods using directional normalization
which yield much better results. Our interpretation of this phenomenon is as
follows. There are many equivalent optimal solutions of the linear relaxation
of the unsplittable flow problem. When a cut is generated with the natural
normalization, it only cuts a subset of these solutions, and the new solution
to the Fenchel master problem is in a completely different place in the
solution space. The method then fails to cut all the solutions because of this
large number of symmetries. In contrast, the cuts generated using directional
normalization focus on the optimal solution of the Dantzig-Wolfe master
problem and try to prove its optimality. By focusing on a sub-part of the
solution space, this method avoids the problem of symmetries which improves
its convergence. This hypothesis is supported by the results of a preliminary
study on a variation of the unsplittable flow problem where a path is favored.
Indeed, because of the presence of a privileged path for each commodity, this
variant does not have as many symmetries. In this context, there is a smaller
difference between the methods based on the two normalizations.
Comparison between DW-interior-point and DW-Fenchel-iterative. In the case of
the low maximum demand dataset where the number of commodities is high, the
DW-Fenchel-iterative method behaves a lot better than the Dantzig-Wolfe
methods. We assume that this is because the large number of commodities
implies a greater degeneracy of the master problem which should less bother
the DW-Fenchel methods. Indeed, this degeneracy seems to be the cause of the
rather slow start of the Dantzig-Wolfe methods on these instances. In
contrast, for the high maximum demand dataset, the DW-interior-point and DW-
Fenchel-iterative methods show more similar results. On these instances, the
DW-Fenchel-iterative method shows a faster start of convergence, but slows
down at the end of convergence, in particular for the lower bound.
Impact of capacity size. The knapsack problem is known to have pseudo-
polynomial resolution algorithms in the knapsack capacity such as the MINKNAP
algorithm that we use. In the case of unsplittable flows, this capacity of the
knapsack corresponds to the capacity of the arcs. In Figure 17 and 17, we vary
the capacities of the arcs. Note that the results for capacities of 1000 are
given in Figure 11. This variation in capacities impacts the computation time
of the two methods making them slower. However, the DW-Fenchel-iterative
method is much more impacted because the methods having a Fenchel sub-problem
spend more time in their sub-problem than the Dantzig-Wolfe methods. This
emphasizes the fact that having a fast optimization oracle for the polyhedron
$Q_{2}$ is much more important for the methods based on a Fenchel sub-problem
than those based on a Dantzig-Wolfe sub-problem.
General comments. The new methods presented that couples the Dantzig-Wolfe and
Fenchel decompositions shows very promising results. In particular, they seems
to be far less affected by degeneracy than the Dantzig-Wolfe decomposition and
possess better convergence than the Fenchel decomposition. On the other hand,
they can end up converging slightly less rapidly than the Dantzig-Wolfe
decomposition on instances where degeneracy is not an issue. The new methods
are particularly effective when the optimization oracle ($O$) can be
implemented by a fast algorithm.
## 9 Conclusions
In this work we have revisited Dantzig-Wolfe and Fenchel decompositions for
some hard combinatorial problems with block structures. We have provided
geometrical and intuitive interpretations of several types of normalizations
used in the literature to stabilize the sub-problems associated. This
intuition has fueled the development of a novel methodology capable of
coupling both decomposition approaches acting in tandem via a directional
normalization. We have conducted a thorough computational campaign to
demonstrate the effectiveness of the newly proposed approach for the
unsplittable flow problem. We have observed that on problems suffering from
high degrees of degeneracy, the new approach is superior to its competitors.
Meanwhile, it is also competitive with the classical approaches on the less
degenerate cases. We also proposed a new approach to solve the Fenchel
subproblem with directional normalization by using an alternative
normalization. We provide theoretical guarantees for the finiteness of this
new approach for several classes of alternative normalizations and our
experimental campaign revealed that it presents far less numerical
instabilities.
A likely lead for future research will therefore be to investigate the
performance of this new method in different contexts than the unsplittable
flow problems. Moreover, one of the central points of this new method is the
use of directional normalization in the Fenchel sub-problem. It would be
interesting to use this normalization inside other decomposition methods.
## References
* (1) Pasquale Avella, Maurizio Boccia and Igor Vasilyev “A computational study of exact knapsack separation for the generalized assignment problem” In _Computational Optimization and Applications_ 45.3 Springer, 2010, pp. 543–555 DOI: 10.1007/s10589-008-9183-8
* (2) Egon Balas and Michael Perregaard “Lift-and-project for mixed 0–1 programming: recent progress” In _Discrete Applied Mathematics_ 123.1-3 Elsevier, 2002, pp. 129–154 DOI: 10.1016/S0166-218X(01)00340-7
* (3) Cynthia Barnhart et al. “Branch-and-Price: Column Generation for Solving Huge Integer Programs” In _Operations Research_ 46.3, 1998, pp. 316–329 DOI: 10.1287/opre.46.3.316
* (4) Eric Beier, Saravanan Venkatachalam, Luca Corolli and Lewis Ntaimo “Stage-and scenario-wise fenchel decomposition for stochastic mixed 0-1 programs with special structure” In _Computers & Operations Research_ 59 Elsevier, 2015, pp. 94–103 DOI: https://doi.org/10.1016/j.cor.2014.12.011
* (5) Meriema Belaidouni and Walid Ben-Ameur “On the minimum cost multiple-source unsplittable flow problem” In _RAIRO-Operations Research_ 41.3 EDP Sciences, 2007, pp. 253–273 DOI: 10.1051/ro:2007023
* (6) Walid Ben-Ameur and José Neto “Acceleration of cutting-plane and column generation algorithms: Applications to network design” In _Networks: An International Journal_ 49.1 Wiley Online Library, 2007, pp. 3–17 DOI: 10.1002/net.20137
* (7) JF Benders “Partitioning procedures for solving mixed-variables programming problems.” In _Numerische Mathematik_ 4, 1962, pp. 238–252 DOI: 10.1007/BF01386316
* (8) Maurizio Boccia, Antonio Sforza, Claudio Sterle and Igor Vasilyev “A cut and branch approach for the capacitated p-median problem based on Fenchel cutting planes” In _Journal of mathematical modelling and algorithms_ 7.1 Springer, 2008, pp. 43–58 DOI: 10.1007/s10852-007-9074-5
* (9) Pierre Bonami “Etude et mise en œuvre d’approches polyédriques pour la résolution de programmes en nombres entiers ou mixtes généraux”, 2003
* (10) Karl Heinz Borgwardt “The Simplex Method: a probabilistic analysis” Springer Science & Business Media, 2012 DOI: /10.1007/978-3-642-61578-8
* (11) E Andrew Boyd “Generating Fenchel cutting planes for knapsack polyhedra” In _SIAM Journal on Optimization_ 3.4 SIAM, 1993, pp. 734–750 DOI: 10.1137/0803038
* (12) E Andrew Boyd “On the convergence of Fenchel cutting planes in mixed-integer programming” In _SIAM Journal on Optimization_ 5.2 SIAM, 1995, pp. 421–435 DOI: 10.1137/0805021
* (13) Olivier Briant et al. “Comparison of bundle and classical column generation” In _Mathematical programming_ 113.2 Springer, 2008, pp. 299–344 DOI: 10.1007/s10107-006-0079-z
* (14) Marcel Büther and Dirk Briskorn “Reducing the 0-1 knapsack problem with a single continuous variable to the standard 0-1 knapsack problem” In _International Journal of Operations Research and Information Systems (IJORIS)_ 3.1 IGI Global, 2012, pp. 1–12 DOI: 10.4018/joris.2012010101
* (15) Liang Chen, Wei-Kun Chen, Mu-Ming Yang and Yu-Hong Dai “An exact separation algorithm for unsplittable flow capacitated network design arc-set polyhedron” In _Journal of Global Optimization_ Springer, 2021, pp. 1–31 DOI: 10.1007/s10898-020-00967-z
* (16) Vašek Chvátal, William Cook and Daniel Espinoza “Local cuts for mixed-integer programming” In _Mathematical Programming Computation_ 5.2 Springer, 2013, pp. 171–200 DOI: 10.1007/s12532-013-0052-9
* (17) Michele Conforti and Laurence A Wolsey ““Facet” separation with one linear program” In _Mathematical Programming_ 178.1 Springer, 2019, pp. 361–380 DOI: 10.1007/s10107-018-1299-8
* (18) D. Coudert and H. Rivano “Lightpath assignment for multifibers WDM networks with wavelength translators” In _Global Telecommunications Conference, 2002. GLOBECOM ’02. IEEE_ 3, 2002, pp. 2686–2690 vol.3 DOI: 10.1109/GLOCOM.2002.1189117
* (19) George B Dantzig “Maximization of a linear function of variables subject to linear inequalities” In _Activity analysis of production and allocation_ 13, 1951, pp. 339–347
* (20) George B Dantzig and Philip Wolfe “Decomposition principle for linear programs” In _Operations research_ 8.1 INFORMS, 1960, pp. 101–111 DOI: 10.1287/opre.8.1.101
* (21) Sanjeeb Dash “Mixed integer rounding cuts and master group polyhedra” In _Combinatorial Optimization_ IOS Press, 2011, pp. 1–32 DOI: 10.3233/978-1-60750-718-5-1
* (22) Guy Desaulniers, Jacques Desrosiers and Marius M Solomon “Column generation” Springer Science & Business Media, 2006 DOI: 10.1007/b135457
* (23) Judith M Farvolden, Warren B Powell and Irvin J Lustig “A primal partitioning solution for the arc-chain formulation of a multicommodity network flow problem” In _Operations Research_ 41.4 INFORMS, 1993, pp. 669–693 DOI: 10.1287/opre.41.4.669
* (24) Ricardo Fukasawa and Marcos Goycoolea “On the exact separation of mixed integer knapsack cuts” In _Mathematical programming_ 128.1 Springer, 2011, pp. 19–41 DOI: 10.1007/s10107-009-0284-7
* (25) Jean-Louis Goffin and Jean-Philippe Vial “Convex nondifferentiable optimization: A survey focused on the analytic center cutting plane method” In _Optimization methods and software_ 17.5 Taylor & Francis, 2002, pp. 805–867 DOI: 10.1080/1055678021000060829a
* (26) Jacek Gondzio, Pablo González-Brevis and Pedro Munari “New developments in the primal–dual column generation technique” In _European Journal of Operational Research_ 224.1 Elsevier, 2013, pp. 41–51 DOI: 10.1016/j.ejor.2012.07.024
* (27) LLC Gurobi Optimization “Gurobi Optimizer Reference Manual”, 2020 URL: http://www.gurobi.com
* (28) Yichao He et al. “Encoding transformation-based differential evolution algorithm for solving knapsack problem with single continuous variable” In _Swarm and Evolutionary Computation_ 50 Elsevier, 2019, pp. 100507 DOI: 10.1016/j.swevo.2019.03.002
* (29) Konstantinos Kaparis and Adam N Letchford “Separation algorithms for 0-1 knapsack polytopes” In _Mathematical programming_ 124.1 Springer, 2010, pp. 69–91 DOI: 10.1007/s10107-010-0359-5
* (30) L.G. Khachiyan “Polynomial algorithms in linear programming” In _USSR Computational Mathematics and Mathematical Physics_ 20.1, 1980, pp. 53–72 DOI: 10.1016/0041-5553(80)90061-0
* (31) François Lamothe et al. “Dynamic unsplittable flows with path-change penalties: New formulations and solution schemes for large instances” In _Computers & Operations Research_ 152, 2023, pp. 106–154 DOI: 10.1016/j.cor.2023.106154
* (32) François Lamothe et al. “Randomized rounding algorithms for large scale unsplittable flow problems” In _Journal of Heuristics_ , 2021 DOI: 10.1007/s10732-021-09478-w
* (33) AH Land and AG Doig “An Automatic Method of Solving Discrete Programming Problems” In _Econometrica_ 28.3, 1960, pp. 497–520 DOI: 10.1007/978-3-540-68279-0_5
* (34) Geng Lin, Wenxing Zhu and M Montaz Ali “An exact algorithm for the 0–1 linear knapsack problem with a single continuous variable” In _Journal of global optimization_ 50.4 Springer, 2011, pp. 657–673 DOI: 10.1007/s10898-010-9642-5
* (35) Hongtao Liu “An exact algorithm for the biobjective 0-1 linear knapsack problem with a single continuous variable” In _2017 18th International Conference on Parallel and Distributed Computing, Applications and Technologies (PDCAT)_ , 2017, pp. 81–85 IEEE DOI: 10.1109/PDCAT.2017.00022
* (36) Hugues Marchand and Laurence A Wolsey “The 0-1 knapsack problem with a single continuous variable” In _Mathematical Programming_ 85.1 Springer, 1999, pp. 15–33 DOI: 10.1007/s101070050044
* (37) Jiří Matoušek and Bernd Gärtner “Understanding and using linear programming” Springer, 2007 DOI: 10.1007/978-3-540-30717-4
* (38) Sanjay Mehrotra “On the implementation of a primal-dual interior point method” In _SIAM Journal on optimization_ 2.4 SIAM, 1992, pp. 575–601 DOI: 10.1137/0802028
* (39) Philip James Neame “Nonsmooth dual methods in integer programming” University of Melbourne, Department of MathematicsStatistics, 2000
* (40) Lewis Ntaimo “Fenchel decomposition for stochastic mixed-integer programming” In _Journal of Global Optimization_ 55.1 Springer, 2013, pp. 141–163 DOI: 10.1007/s10898-011-9817-8
* (41) Sungsoo Park, Deokseong Kim and Kyungsik Lee “An integer programming approach to the path selection problems” In _Proceedings of the International Network Optimization Conference INOC, Evry-Paris, France_ , 2003, pp. 448–453
* (42) Artur Pessoa, Ruslan Sadykov, Eduardo Uchoa and Francois Vanderbeck “In-out separation and column generation stabilization by dual price smoothing” In _International Symposium on Experimental Algorithms_ , 2013, pp. 354–365 Springer DOI: 10.1007/978-3-642-38527-8_31
* (43) David Pisinger “A minimal algorithm for the 0-1 knapsack problem” In _Operations Research_ 45.5 INFORMS, 1997, pp. 758–767 DOI: 10.1287/opre.45.5.758
* (44) Paul Wentges “Weighted Dantzig-Wolfe decomposition for linear mixed-integer programming” In _International Transactions in Operational Research_ 4.2 Elsevier, 1997, pp. 151–162 DOI: 10.1016/S0969-6016(97)00001-4
* (45) Laurence A Wolsey and George L Nemhauser “Integer and combinatorial optimization” John Wiley & Sons, 1999 DOI: 10.1002/9781118627372
* (46) Chenxia Zhao and Xianyue Li “Approximation algorithms on 0–1 linear knapsack problem with a single continuous variable” In _Journal of Combinatorial optimization_ 28.4 Springer, 2014, pp. 910–916 DOI: 10.1007/s10878-012-9579-3
## Appendix A Proof of the finite convergence for the iterative resolution of
the Fenchel sub-problem
In this appendix, we give proofs of the finite convergence for the method
presented in Section 5 for two types of secondary normalizations. First, when
the secondary normalization guarantees that a facet of $Q_{2}$ will be
generated by the linear program. Second, when the secondary normalization is
$\|\pi\|\leq 1$ where $\|.\|$ is any norm.
### A.1 Secondary normalization generating facets
In this section, we are interested in the termination of the proposed method
to solve the Fenchel sub-problem when the secondary normalization guarantees
that a facet of $Q_{2}$ will be generated by the separation linear program.
###### Theorem A.1
If the secondary normalization guarantees the generation of facets then the
method generates a cut associated with a directional normalization in finite
time.
###### Proof
We will show that in the worst case the method ends after the secondary
separation has generated all the facets of $Q_{2}$. For this, we show that
each facet of $Q_{2}$ is generated at most once. As the algorithm progresses,
the point $x^{\prime}$ separated during secondary separations advances along
the segment $(\widecheck{x},\widehat{x})$ in the direction of $\widehat{x}$.
Once a facet of $Q_{2}$ is generated, the point $x^{\prime}$ is projected onto
that facet along the segment $(\widecheck{x},\widehat{x})$. The future points
$x^{\prime}$ will therefore all satisfy the inequality associated with this
facet which can thus no longer be generated by the secondary separation
problem. Since a polyhedron has a finite number of facets, a facet
intersecting the segment $(\widecheck{x},\widehat{x})$ is generated in a
finite number of steps. Once this happens, the method ends after a single call
to the alternate separation problem. Indeed, the procedure places the point
$x^{\prime}$ on the point of intersection between this facet and the segment
$(\widecheck{x},\widehat{x})$. On the next iteration, the secondary separation
problem indicates that the point $x^{\prime}$ belongs to $Q_{2}$ and the
method stops. ∎
### A.2 Secondary normalization using any norm
We now present a proof of convergence when the secondary normalization is
$\|\pi\|\leq 1$ where $\|.\|$ is any norm. This proof uses the Lemmas 1 and 2
presented in boyd1995convergence (12). Previously, we recall the following
properties and notations. In the case of a normalization $\|\pi\|\leq 1$, the
dual problem of the separation problem is: find the point of $Q_{2}$
minimizing the distance $\|x-\widecheck{x}\|^{*}$ where $\|.\|^{*}$ is the
dual norm of $\|.\|$: $\|\lambda\|^{*}=\max_{\|x\|\leq 1}\lambda x$. The
solution point of the dual problem is denoted $x^{*}$ and always satisfies at
equality the cut generated by the separation problem. We now present the
lemmas used in the proof.
###### Lemma 1 (boyd1995convergence (12))
Let $\lambda^{*}x\leq\lambda^{*}x^{*}$ be the cut generated during the
separation of a point $\widehat{x}$ from a polyhedron $Q_{2}$ with the
normalization $\|\pi\|\leq 1$ where $x^{*}$ is the optimal solution of the
dual of the separation problem. Then $-\lambda^{*}$ is a sub-gradient of
$x\mapsto\|x-\widecheck{x}\|^{*}$ in $x^{*}$.
In the second lemma $\angle(\lambda,x)$ will denote the angle between the
vectors $\lambda$ and $x$ (the one lower than $\pi$ radiant).
###### Lemma 2 (boyd1995convergence (12))
For each norm, there exists an angle $\theta_{min}>0$ such that at any point
$x$ and for any sub-gradient $\lambda$ of this norm in $x$:
$\angle(\lambda,x)\geq\frac{\pi}{2}-\theta_{min}$
We now present the main theorem of this section.
###### Theorem A.2
If the secondary normalization is $\|\pi\|\leq 1$ for any norm then the method
generates a cut associated with a directional normalization in finite time.
###### Proof
Scheme of the proof: First we will show that once a face intersecting the
segment $(\widecheck{x},\widehat{x})$ has been generated the method ends after
a single call to the secondary separation problem. Secondly, we will show that
if another face of $Q_{2}$ is generated, the point $x^{\prime}$ advances more
than $\epsilon$ along the segment $(\widecheck{x},\widehat{x})$ toward
$\widehat{x}$ where $\epsilon$ is a strictly non-negative distance independent
of the iteration. Thus, this second case cannot happen more than
$\frac{\|\widehat{x}-\widecheck{x}\|_{2}}{\epsilon}$ times so the procedure
ends in a finite number of steps.
1) Suppose that at one iteration, a face intersecting the segment
$(\widecheck{x},\widehat{x})$ is generated. After generating the face, the
procedure places the point $x^{\prime}$ on the intersection point. At the next
iteration, the secondary separation problem indicates that the point
$x^{\prime}$ belongs to $Q_{2}$ and the method stops.
2) For the rest of this proof, we will denote by $x^{(i)}$ the point
$x^{\prime}$ separated during iteration $i$ of the algorithm. We will show
that if a face $F$ of $Q_{2}$ generated does not intersect the segment
$(\widecheck{x},\widehat{x})$, then the point $x^{\prime}$ advances along the
segment $(\widecheck{x},\widehat{x})$ a strictly non-negative distance
$d_{min}\sin(\theta_{min})$ independent of the face $F$, i.e.
$\|x^{(i+1)}-x^{(i)}\|_{2}\geq d_{min}\sin(\theta_{min})$. To that end, we
will use trigonometry on the triangle formed by the points associated with
$x^{(i)}$, $x^{*}$, and $x^{(i+1)}$ where $x^{*}$ is the dual optimal solution
of the secondary separation problem. This triangle will be denoted $\triangle
x^{(i)}x^{*}x^{(i+1)}$. We will use $\theta_{min}$ as lower bound for the
angle $\angle(x^{(i+1)}-x^{*},x^{(i)}-x^{*})$ and $d_{min}$ as lower bound for
the distance $\|x^{(i)}-x^{*}\|_{2}$.
2.1) Suppose that at iteration $i$, the secondary separation problem of
$x^{(i)}$ returns a primal-dual solution pair $(\lambda^{*},x^{*})$ which thus
correspond to the cut $\lambda^{*}x\leq\lambda^{*}x^{*}$. Recall that $x^{*}$
is a point of $Q_{2}$ satisfying the generated cut to equality. It is thus on
the generated face $F$. From Lemma 1, since the secondary normalization is
$\|\pi\|\leq 1$, the vector $-\lambda^{*}$ is a sub-gradient of
$\|x-x^{(i)}\|^{*}$ in $x^{*}$. Thus, according to Lemma 2, we have
$\angle(-\lambda^{*},x^{*}-x^{(i)})\leq\frac{\pi}{2}-\theta_{min}$ for a
$\theta_{min}>0$ depending on the used norm but neither on the face $F$ nor on
the iteration. This is equivalent to
$\frac{\pi}{2}-\angle(\lambda^{*},x^{(i)}-x^{*})\geq\theta_{min}$. After
generating the cut $\lambda^{*}x\leq\lambda^{*}x^{*}$, the procedure projects
the point $x^{(i)}$ on the hyperplane $\lambda^{*}x=\lambda^{*}x^{*}$ along
the segment $(\widecheck{x},\widehat{x})$. The result of this projection is
the point $x^{(i+1)}$. Since both $x^{(i+1)}$ and $x^{*}$ are point of the
hyperplane $\lambda^{*}x=\lambda^{*}x^{*}$, the vector $x^{(i+1)}-x^{*}$ is on
the hyperplane $\lambda^{*}x=0$. Let us consider the smallest angle between
$x^{(i)}-x^{*}$ and a point of the hyperplane $\lambda^{*}x=0$. Firstly, it is
smaller than the angle $\angle(x^{(i+1)}-x^{*},x^{(i)}-x^{*})$. Secondly, this
smallest angle can be expressed in terms of the normal $\lambda^{*}$ as
$\frac{\pi}{2}-\angle(\lambda^{*},x^{(i)}-x^{*})$ which is shown above to be
greater than $\theta_{min}$. Thus, we have shown that the angle
$\bm{\angle(x^{(i+1)}-x^{*},x^{(i)}-x^{*})}$ is greater than
$\bm{\theta_{min}}$.
2.2) Let $d_{min}$ be the distance in norm $\|.\|_{2}$ between the segment
$(\widecheck{x},\widehat{x})$ and the union of the faces of $Q_{2}$ that do
not intersect the segment $(\widecheck{x},\widehat{x})$. Since the generated
face $F$ does not intersect the segment $(\widecheck{x},\widehat{x})$, the
distance between the segment $(\widecheck{x},\widehat{x})$ and the face $F$ is
greater than $d_{min}$. However, $x^{(i)}$ belongs to the segment
$(\widecheck{x},\widehat{x})$ and $x^{*}$ to the face $F$ so the distance
$\bm{\|x^{(i)}-x^{*}\|_{2}}$ is greater than $\bm{d_{min}}$.
2.3) We will now show with simple trigonometry on the triangle $\triangle
x^{(i)}x^{*}x^{(i+1)}$ that $\|x^{(i+1)}-x^{(i)}\|_{2}\geq
d_{min}\sin(\theta_{min})$. First, the distance $\|x^{(i+1)}-x^{(i)}\|_{2}$ is
larger than the length of the altitude of the triangle $\triangle
x^{(i)}x^{*}x^{(i+1)}$ associated with $x^{(i)}$. However, the length of this
altitude is $\|x^{(i)}-x^{*}\|_{2}\sin(\angle(x^{(i)}-x^{*},x^{(i+1)}-x^{*}))$
which according to paragraph 2.1) and 2.2) is greater than
$d_{min}\sin(\theta_{min})$. Thus finally, we have:
$\bm{\|x^{(i+1)}-x^{(i)}\|_{2}\geq d_{min}\sin(\theta_{min})}$.
The distance $d_{min}\sin(\theta_{min})$ is strictly non-negative and
independent of the iteration of the algorithm which completes the proof. ∎
## Appendix B Solution of the knapsack oracle
All the decomposition methods presented in this work assume that there exists
an efficient algorithm capable of optimizing a linear function on the
polyhedron $Q_{2}$ which in our version of the unsplittable flow problem can
be written as:
$\\{(f_{a}^{k}\in\\{0,1\\})_{a\in A,k\in k}o_{a}\in\mathbb{R}^{+}|\sum_{k\in
K}f_{a}^{k}D_{k}\leq c_{a}+o_{a}\\}.$
Optimizing a linear function whose coefficients are $(\pi^{k})_{k\in K}$ and
$-\pi_{o}$ on this polyhedron can thus be done by following mixed integer
linear program:
$\displaystyle(O_{a})\quad$ $\displaystyle\max_{f_{a}^{k},o_{a}}$
$\displaystyle\sum_{k\in K}\pi^{k}f_{a}^{k}-\pi_{o}o_{a}$ subject to
$\displaystyle\sum_{k\in K}f_{a}^{k}D_{k}\leq c_{a}+o_{a}$ $\displaystyle
f_{a}^{k}\in\\{0,1\\},~{}o_{a}\in\mathbb{R}^{+}$
This problem can be addressed by solving two classic 0-1 knapsack problems
using a case disjunction. This method was presented by buther2012reducing (14)
but we recall it here. Consider the following disjunction: either the flow of
the commodities respects the capacity of the arc $a$, or the flow of the
commodities exceeds the capacity of the arc $a$. Finding the best solution in
the first case amounts to solving the following problem:
$\displaystyle\max_{f_{a}^{k}}$ $\displaystyle\sum_{k\in K}\pi^{k}f_{a}^{k}$
subject to $\displaystyle\sum_{k\in K}f_{a}^{k}D_{k}\leq c_{a}$ $\displaystyle
f_{a}^{k}\in\\{0,1\\}$
Indeed, we suppose that the coefficient $\pi_{o}$ is non-negative because
otherwise the problem $O_{a}$ would be unbounded. Since the flow of
commodities respects the capacity of the arc, the variable $o_{a}$ always
takes the value zero and can be removed from the problem. Note that in this
first case of the disjunction, the problem to be solved is a classic knapsack
problem. In the second case of disjunction, finding the best solution amounts
to solving the following problem:
$\displaystyle\max_{f_{a}^{k},o_{a}}$ $\displaystyle\sum_{k\in
K}\pi^{k}f_{a}^{k}-\pi_{o}o_{a}$ subject to $\displaystyle\sum_{k\in
K}f_{e}^{k}D_{k}\leq c_{a}+o_{a}$ $\displaystyle\sum_{k\in
K}f_{a}^{k}d^{k}\geq c_{a}$ $\displaystyle
f_{a}^{k}\in\\{0,1\\},~{}o_{a}\in\mathbb{R}^{+}$
Since we assume that $\pi_{o}$ is non-negative and that the flow of
commodities does not respect the capacity of the arcs, the variable $o_{e}$ is
always equal to $\sum_{k\in K}f_{a}^{k}D_{k}-c_{a}$. By performing the
replacement in the objective function, by replacing the variables $f_{a}^{k}$
by their complement $\bar{f}_{a}^{k}=1-f_{a}^{k}$ and by multiplying the
constraint $\sum_{k\in K}f_{a}^{k}D_{k}\geq c_{a}$ by -1 we get the following
reformulation:
$\displaystyle\max_{\bar{f}_{a}^{k}}$ $\displaystyle\sum_{k\in
K}(\pi_{o}D_{k}-\pi^{k})\bar{f}_{a}^{k}+C$ subject to $\displaystyle\sum_{k\in
K}\bar{f}_{a}^{k}D_{k}\leq\sum_{k\in K}D_{k}-c_{a}$
$\displaystyle\bar{f}_{a}^{k}\in\\{0,1\\}$
where the constant $C$ is equal to $\sum_{k\in
K}(\pi^{k}-\pi_{o}d^{k})-\pi_{o}c_{a}$. This reformulation shows that the
problem to be solved in the second case of the disjunction is also a classic
knapsack problem.
Note: in the experimental study of Section 8, in order to solve these two
knapsack problems, we use the MINKNAP algorithm proposed by
pisinger1997minimal (43). Since MINKNAP only accepts integer-valued weights
and profits, we multiply the dual vector $(\pi,\pi_{0},\pi_{o})$ by $10^{7}$
and truncate before invoking MINKNAP.
## Appendix C Pre/post-processing for Fenchel subproblems
Solving directly a Fenchel subproblem is sometimes too computationally
expensive to be integrated into a decomposition method. However, the
resolution time of this subproblem can be greatly reduced with pre/post-
processing steps. Indeed, boccia2008cut (8) showed that it is possible to
solve the Fenchel separation problem by focusing on a sub-polyhedron of
$Q_{2}$ of far lower dimension. However, the cut generated may not be valid
for $Q_{2}$ and one must use a lifting procedure to create a cut valid for
$Q_{2}$. These concepts are explained in the following.
Dimensionality reduction: This technique typically applies when the variables
$x$ of the initial problem ($P$) are binary. Instead of separating a point
$\widecheck{x}$ from $Q_{2}$, this point is separated from the sub-polyhedron
$Q_{2}^{f}(\widecheck{x})$ induced by the variables taking a value different
from their bounds in $\widecheck{x}$ (in the binary case, induced by variables
taking a fractional value in $\widecheck{x}$). More precisely,
$Q_{2}^{f}(\widecheck{x})=\\{x\in Q_{2}|x_{i}=\widecheck{x}_{i}\text{ if
}\widecheck{x}_{i}\in\\{0,1\\}\\}$. Indeed, $\widecheck{x}$ belongs to $Q_{2}$
if and only if it belongs to $Q_{2}^{f}(\widecheck{x})$ which implies that
$\widecheck{x}$ is separable from $Q_{2}$ if and only if it is separable from
$Q_{2}^{f}(\widecheck{x})$. Replacing $Q_{2}$ with $Q_{2}^{f}(\widecheck{x})$
in the separation problem has two advantages. First, the problem only
considers variables taking a fractional value in $\widecheck{x}$ which reduces
the number of dimensions of the problem and therefore speeds up its
resolution. Forcing certain variables to take a binary value in the
optimization oracle on $Q_{2}$ is also often easy to do. Second, if the
separation problem ($S$) is solved directly through constraint generation,
most of the computational time is used to generate constraints associated with
vertices of $Q_{2}$ not belonging to $Q_{2}^{f}(\widecheck{x})$ and which are
therefore mostly unnecessary. Reducing the search space to
$Q_{2}^{f}(\widecheck{x})$ allows the computation to concentrate on the part
of the space containing the most important vertices.
Lifting: Upon applying the dimensionality reduction described before, the cut
generated will be valid for $Q_{2}^{f}(\widecheck{x})$ but not necessarily for
$Q_{2}$. A procedure able to create a valid cut for $Q_{2}$ from one valid for
$Q_{2}^{f}(\widecheck{x})$ is the sequential lifting procedure
wolsey1999integer (45) that we will now describe. In the following, we will
name fixed variables, the variables eliminated from the separation problem
when replacing the polyhedron $Q_{2}$ by the polyhedron
$Q_{2}^{f}(\widecheck{x})$. For the fixed variables, the coefficients of the
generated cut are all zero. At each step of the sequential lifting procedure,
a new value for the coefficient of one of the fixed variables is computed so
that the cut becomes valid for the polyhedron where this variable is no longer
fixed. These new coefficients are optimal in the sense that if the initial cut
was a facet of the polyhedron $Q_{2}^{f}(\widecheck{x})$ then the cut
resulting from the sequential lifting is a facet of $Q_{2}$. We now present an
iteration of the sequential lifting procedure. Suppose therefore that
$Q_{2}^{f}(\widecheck{x})=\\{x\in Q_{2}|x_{1}=1\\}$; the procedure will then
contain only one iteration because only the variable $x_{1}$ is fixed. The cut
generated for $Q_{2}^{f}(\widecheck{x})$ is $\pi x\leq\pi_{0}$ where the
coefficient $\pi_{1}$ associated with the variable $x_{1}$ is zero. The
lifting procedure consists in creating a new cut $\gamma x\leq\gamma_{0}$ such
that the two cuts are identical when $x_{1}=1$ and such that the new cut is
valid for $Q_{2}$. Since $\pi x\leq\pi_{0}$ is already valid for
$Q_{2}^{f}(\widecheck{x})$, the two previous conditions can be written:
$\forall x\in\\{x|x_{1}=1\\},~{}\gamma x-\gamma_{0}=\pi x-\pi_{0}$ $\max_{x\in
Q_{2}~{}|~{}x_{1}=0}\gamma x\leq\gamma_{0}$
Let $e_{i}$ be the $i^{\text{th}}$ vector of the canonical basis. Applying the
first condition for $x=e_{1}$, we get $\gamma_{1}-\gamma_{0}=\pi_{1}-\pi_{0}$.
Moreover, by applying it in $x=e_{1}+e_{i}$ for all $i>1$, we obtain
$\gamma_{1}+\gamma_{i}-\gamma_{0}=\pi_{1}+\pi_{i}-\pi_{0}$ which becomes
$\gamma_{i}=\pi_{i}$ by subtracting the previous equality. We therefore know:
$\displaystyle\forall i>1,\gamma_{i}=\pi_{i}$
$\displaystyle\gamma_{1}=\gamma_{0}+\pi_{1}-\pi_{0}$
Thus, the value of $\gamma_{0}$ is the only element missing to know the value
of all the $\gamma_{i}$. By replacing $\gamma_{i}$ by $\pi_{i}$ for all $i>1$
in the second condition we get $\gamma_{0}\geq\max_{x\in\\{x\in
Q_{2}|x_{1}=0\\}}\pi x$. The smallest value of $\gamma_{0}$ satisfying the
second condition is, therefore, $\gamma_{0}=\max_{x\in\\{x\in
Q_{2}|x_{1}=0\\}}\pi x$ which can be computed with a call to the optimization
oracle on $Q_{2}$ where the variable $x_{1}$ set to $0$. It is therefore
possible to compute each lifted coefficient using a single call to the oracle
($O$).
|
AND
# Bounded-error constrained state estimation in presence of sporadic
measurements
Yasmina BECIS-AUBRY *The author is with Université d’Orléans, Laboratoire
PRISME EA 4229 (Univ. Orléans - INSA CVL). 63 av. de Lattre de Tassigny, 18020
Bourges Cedex, FRANCE. Tel. +33 2 48 23 84 78<EMAIL_ADDRESS>
###### Abstract
This contribution proposes a recursive set-membership method for the
ellipsoidal state characterization for linear discrete-time models with
additive unknown disturbances vectors, bounded by possibly degenerate
zonotopes and polytopes, impacting respectively, the state evolution equation
and the sporadic measurement vectors, which are expressed as linear inequality
and equality constraints on the state vector. New algorithms are designed
considering the unprecedented fact that, due to equality constraints, the
shape matrix of the ellipsoid characterizing all possible values of the state
vector is non invertible. The two main size minimizing criterions (volume and
sum of squared axes lengths) are examined in the time update step and also in
the observation updating, in addition to a third one, minimizing some error
norm and ensuring the input-to-state stability of the estimation error.
## 1 INTRODUCTION
There is no more need to praise the interests of the set-membership state
estimation techniques neither is there a necessity to recall how interesting
alternative they offer to conventional state estimation methods where the
statistical assumptions on the disturbances can not be satisfied in certain
practical situations, nor how increasing attention they are currently
receiving since the noises of their models are assumed only to be bounded.
Nevertheless, the stability question is rarely addressed in this kind of
estimation approach.
On the other hand, the constrained state filtering has been widely studied in
stochastic context [1], [2], [3]. In [4], constrained Kalman filter variations
were reexamined and an alternative derivation of the optimal constrained
Kalman filter for time variant systems was proposed. The literature is less
abundant on this subject when it comes to bounded error framework. LMI
techniques were employed for ellipsoidal set-membership constrained state
filtering with linear [5] and linearized [6] equalities. In [7], the authors
used a combined stochastic and set-membership uncertainty representation by
integrating, into the Kalman filter structure, ellipsoidal constraints on the
state vector as a relaxation of equality constraints. All these works faced a
same challenge, not arising here, in inverting the estimation error covariance
matrix, becoming inevitably singular, when dealing with equality constraints.
Most algorithms of set-membership techniques, when applied to linear discrete-
time systems, are composed of two stages, in the manner of Kalman filter: 1.
the time prediction consisting of the linear transformation of the set
containing the state vectors at previous time step, on one hand and the
Minkowski sum (cf. $\S$ 1.1.9.) of the set containing the processes errors, on
the other; followed by 2. the measurement correction phase, where an
intersection of the predicted set just obtained is performed with the one
provided by the measurements with bounded noises.
Four main approaches are sharing the research interests in this area: those
using intervals or boxes (i.e.balls for the $\infty-$norm), parallelotopes
(affine images of boxes, obtained by injective, usually bijective, maps) or
more general zonotopes (usually surjective, affine maps of boxes) and
ellipsoids (affine maps of balls for the $2-$norm).
The asset of the interval methods is the simplicity of its conceptualization
and the availability of libraries allowing the approximation of a large range
of linear and nonlinear functions. On the other hand, the recourse to interval
computing softwares including time costly operations such as subpavings and
contractors are inevitable to overcome the conservatism of such aligned with
the coordinates’ axes boxes [8], [9].
The zonotope has the attractive structure of a centrally symmetric convex
polytope where every facet is a parallelotope having a parallel congruent
facet. This class has the interesting quality to be closed under the Minkowski
sum (operated at the time prediction stage) resulting in the concatenation of
the generator vectors of the operands, thus inducing their inevitable growth.
The zonotopic approximation has the advantage of the accuracy. However, this
accuracy is again obtained at the expense of memory and time complexity.
As intervals, ellipsoids benefit from the simplicity of their formulation and
the constant minimal space storage required ($n(\frac{n}{2}+2)$ at each
iteration, where $n$ is the state dimension). Also, this class has, in common
with zonotopes, the useful property to be closed under the linear
transformation. The only drawback of choosing ellipsoids for the
characterization of the disturbances lies at the necessity of solving two
polynomial equations of order $n$ (the size of the state vector), at each time
step: one arising at the prediction and the other at the correction step,
while optimizing the ellipsoids overbounding, respectively, the Minkowski sum
and the intersection of ellipsoids111This equation can be avoided at the
prediction step, if minimizing the squared axes’ sum of the resulting
ellipsoid rather than its volume. [10, 11, 12, 13, 14]. It is nonetheless
important to note that the parallelotope is a particular zonotope and that the
ellipsoidal and parallelotopic overbounding methods are tightly related: the
ellipsoid of center $\bm{c}$ and of shape matrix $P$ is the unique minimum
volume ellipsoid (called Löwner-John ellipsoid) inscribed in the parallelotope
of exactly the same center and the same shape matrix. This ellipsoid also
circumscribes the parallelotope of the same center and of shape matrix
$\frac{1}{\sqrt{n}}P$. It is therefore easy to deduce the parallelotopic
characterization from the ellipsoidal one and vice versa.
This is the reason we opted for the ellipsoid to characterize only the set of
all possible values of the state vector. The disadvantages of using the
ellipsoid as bounding set for both process and disturbances vectors are
overcome by choosing zonotopes to bound the process disturbances and, since we
are interested here in the state estimation of linear discrete-time systems
subject not only to bounded disturbances but also to all kinds of linear
constraints on the state vector, the output disturbances vector is modeled by
a more general class of polyhedrons (intersection of halfspaces, each modeling
an inequality constraint on the state vector), which can be unbounded and/or
flattened, due to some possible hyperplane intersections (introduced by
equality constraints). In fact, among all the classes considered here, the
class of ellipsoids is the only one that is closed under the intersection with
a hyperplane. Neither intervals nor zonotopes have this advantageous feature
[15], which will be favorably exploited while processing the equality
constraints. Moreover, the choice of the ellipsoidal set for the state
characterization is the only one that guarantees an optimal (_w.r.t._ chosen
criterions) set, without any conservatism.
This work is the extension of the author’s recent one, [16], where only one
criterion could be chosen at each phase in the presence of equality
constraints: the SSAL (sum of squared axes lengths) at the prediction step and
the scaling parameter $\varsigma_{k}$ representing a weighted estimation error
norm in the worst noise case was minimized at the correction step. Since the
volume of the state bounding ellipsoid becomes zero in presence of equality
constraints, it cannot be minimized. To address this issue, a new notion of
pseudo-volume is introduced. Our concern here is to propose a complete, easily
applicable solution. For this purpose, the trace minimization approach during
the measurement correction stage, was also thoroughly examined. In fact, even
if the trace and determinant minimizations were widely used in literature, for
ellipsoidal set-membership approaches, [10, 12], especially in the parameter
identification framework, [17, 18], at our knowledge, no analytical ready-to-
use solution (without $n$${}^{\textrm{th}}$ order equations or LMI’s to solve)
was developed for the state estimation problem.
The paper is organized as follows. After this introduction, which is completed
by some notations and definitions, the second section recaptures the
constrained set-membership state estimation problem with sporadic
measurements, already formulated in [16]. The third section concerns the time
prediction stage where two size minimization criterions are considered:
pseudo-volume and SSAL (sum of squared axes lengths) the measurement
correction stage of the estimation algorithm is detailed in forth section
using three different optimization criterions. The algorithm’s convergence is
studied in the fifth section. Numerical simulations are presented in the sixth
and finally, a brief conclusion terminates the paper.
### 1.1 Notations and definitions
1. 1.
The symbol ${\,:=\,}$ (_resp_. ${\,=:\,}$) means that the Left Hand Side
(_resp_. RHS) _is defined to be equal to_ the Right Hand Side (_resp_. LHS).
Normal lowercase letters are used for scalars, capital letters for matrices,
bold lowercase letters for vectors and calligraphic capital letters for sets.
${\mathop{\rm I\mkern-3.5muR}}$, ${\mathop{\rm I\mkern-3.5muR}}^{*}$, $R_{+}$,
$R_{+}^{*}$ denote the sets of real, non-zero, nonnegative and positive
numbers _resp_. ${\mathop{\rm I\mkern-3.5muN}}$ and ${\mathop{\rm
I\mkern-3.5muN}}^{*}$ are the sets of nonnegative and positive integers
_resp_. $l,m,n,p,q\in{\mathop{\rm I\mkern-3.5muN}}$ designate vectors and
matrices dimensions. The subscript $k\in{\mathop{\rm I\mkern-3.5muN}}$ is the
discrete time step and $i,j\in{\mathop{\rm I\mkern-3.5muN}}^{*}$ are vector
and matrix component indices.
2. 2.
$x_{i}$ is the $i$${}^{\textrm{th}}$ component of the vector $x$. $a_{ij}$ is
the $i$${}^{\textrm{th}}$ row and $j$${}^{\textrm{th}}$ column element of
$A{\,:=\,}\big{[}a_{ij}\big{]}=[\bm{a}_{j}]\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n\times m}}$ and $\bm{a}_{j}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$ is its $j$${}^{\textrm{th}}$ column vector (if $n=0$
or $m=0$, $A$ is an empty matrix).
3. 3.
$\mbox{\boldmath$0_{n}$}\in{\mathop{\rm I\mkern-3.5muR}}^{n}$ and
$0_{n,m}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times m}}$ are vector and
matrix of zeros; $I_{n}{\,:=\,}[\bm{i}_{n_{1}}\ldots\bm{i}_{n_{n}}]$ is the
$n\times n$ identity matrix.
4. 4.
$A^{T}$, $A^{\dagger}$, $\operatorname*{rank}(A)$, $\mathpzc{Ker}(A)$ and
$\mathpzc{R}(A)$ stand _resp_. for the transpose, Moore-Penrose inverse, rank,
kernel and range of the matrix $A$. If A is square, $\operatorname*{tr}(A)$,
${\left|A\right|}$ and $A^{-1}$, are its trace, determinant and inverse (if
any) _resp_.
5. 5.
$\operatorname*{Diag}(x_{i})_{i=1}^{k}$ is a diagonal matrix where
$x_{1},\ldots,x_{k}$ are its diagonal elements.
6. 6.
A Symmetric matrix $M$ is Positive Definite, denoted by SPD or $M>0$ (_resp_.
Positive Semi-Definite or non-negative definite, denoted by SPSD or $M\geq 0$)
if and only if $\forall\mbox{\boldmath$x$}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\text{--}\\{{\mathbf{0}}\\}$,
$\mbox{\boldmath$x$}^{T}Mx>0$ (_resp_. $\mbox{\boldmath$x$}^{T}Mx\geq~{}0$).
This condition is met if and only if all its eigenvalues are real (because of
its symmetry) and positive (_resp_. non-negative). The matrix inequality $M>N$
(_resp_. $M\geq N$) means that $M-N>0$ (_resp_. $M-N\geq 0$).
7. 7.
$\left\|\mbox{\boldmath$x$}\right\|{\,:=\,}\left\|\mbox{\boldmath$x$}\right\|_{2}{\,:=\,}\sqrt{\mbox{\boldmath$x$}^{T}\mbox{\boldmath$x$}}$
is the 2-norm of the vector $x$ and
$\left\|A\right\|{\,:=\,}\left\|A\right\|_{2}{\,:=\,}\sup_{\bm{x}\neq{\mathbf{0}}}\frac{\left\|A\bm{x}\right\|}{\left\|\bm{x}\right\|}=$
$\sigma_{\max}(A)$, where $\sigma_{\max}(A)=\max_{i}\sigma_{i}(A)$ is the
largest singular value of $A$.
8. 8.
${\mathcal{B}}^{n}_{p}{\,:=\,}\\{\mbox{\boldmath$z$}\in{\mathop{\rm
I\mkern-3.5muR}}^{n}|\left\|\mbox{\boldmath$z$}\right\|_{p}\leq 1\\}$ is a
unit ball in ${\mathop{\rm I\mkern-3.5muR}}^{n}$ for the $p-$norm.
${\mathcal{B}}^{n}_{2}$ and ${\mathcal{B}}^{n}_{\infty}{\,:=\,}[-1,1]^{n}$ are
the centred unit hypersphere and hypercube/box _resp_.
9. 9.
${\mathcal{S}}_{1}\oplus{\mathcal{S}}_{2}{\,:=\,}\\{\mbox{\boldmath$x$}\in{\mathop{\rm
I\mkern-3.5muR}}^{n}|\mbox{\boldmath$x$}=\mbox{\boldmath$x$}_{1}+\mbox{\boldmath$x$}_{2},\mbox{\boldmath$x$}_{1}\in{\mathcal{S}}_{1},\mbox{\boldmath$x$}_{2}\in{\mathcal{S}}_{2}\\}$
is the Minkowski sum of the sets
${\mathcal{S}}_{1},{\mathcal{S}}_{2}\subset{\mathop{\rm I\mkern-3.5muR}}^{n}$
and
$\operatorname*{\oplus}_{i=1}^{m}{\mathcal{S}}_{i}{\,:=\,}{\mathcal{S}}_{1}\oplus\cdots\oplus{\mathcal{S}}_{m}$.
10. 10.
${\mathcal{E}}(\bm{c},P){\,:=\,}\\{x\in{\mathop{\rm I\mkern-3.5muR}}^{n}|\
(x-\bm{c})^{T}P^{-1}(x-\bm{c})\leq 1\\}$ is an ellipsoid in ${\mathop{\rm
I\mkern-3.5muR}}^{n}$, where $\bm{c}\in{\mathop{\rm I\mkern-3.5muR}}^{n}$ is
its center and $P\in{\mathop{\rm I\mkern-3.5muR}}^{n\times n}$ is a SPD matrix
that defines its shape, size and orientation in the ${\mathop{\rm
I\mkern-3.5muR}}^{n}$ space. When $P$ is SPSD not invertible, then the
ellipsoid is degenerate. It is as an affine transformation of matrix $M$
(where $M^{T}M=P$) of the unit Euclidean ball
${\mathcal{B}}_{2}^{n}$:${\mathcal{E}}(\bm{c},M^{T}M)=~{}\\{\mbox{\boldmath$x$}\in{\mathop{\rm
I\mkern-3.5muR}}^{n}|\
\mbox{\boldmath$x$}=\bm{c}+M\mbox{\boldmath$z$},\mbox{\boldmath$z$}\in{\mathcal{B}}_{2}^{p}\\}$.
11. 11.
${\mathcal{H}}(\bm{d},a){\,:=\,}\\{\mbox{\boldmath$x$}\in{\mathop{\rm
I\mkern-3.5muR}}^{n}|\mbox{\boldmath$x$}^{T}\bm{d}=a\\}$ is a hyperplane in
${{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$ of normal vector
$\bm{d}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$ and whose signed distance
from the origin is $\frac{{a}}{\left\|\bm{d}\right\|}$. Let also
${{\mathcal{G}}}(\bm{d},a){\,:=\,}\\{\mbox{\boldmath$x$}:\mbox{\boldmath$x$}^{T}\bm{d}\leq
a\\}$ be one of the two halfspaces into which the hyperplane divides the
${\mathop{\rm I\mkern-3.5muR}}^{n}$ space and ${{\mathcal{G}}}(-\bm{d},-a)$ is
the other one. Now let
${\mathcal{D}}(\bm{d},a){\,:=\,}{{\mathcal{G}}}(\bm{d},1+a)\cap{{\mathcal{G}}}(-\bm{d},1-a)$,
i.e., ${\mathcal{D}}(\bm{d},a){\,:=\,}\\{\mbox{\boldmath$x$}\in{\mathop{\rm
I\mkern-3.5muR}}^{n}:{\left|\mbox{\boldmath$x$}^{T}\bm{d}-a\right|}\leq 1\\}$
which is the strip of ${\mathop{\rm I\mkern-3.5muR}}^{n}$, of width
${2}{\left\|\bm{d}\right\|}^{-1}$, that can also be seen as a degenerate
unbounded ellipsoid or zonotope centered at ${\mathcal{H}}(\bm{d},a)$.
${\mathcal{P}}(C,\bm{d})=\bigcap_{i=1}^{m}{\mathcal{G}}(\bm{c}_{i},d_{i})$ is
a polyhedron.
12. 12.
${\mathcal{Z}}(\bm{c},L){\,:=\,}\\{\mbox{\boldmath$x$}\in{\mathop{\rm
I\mkern-3.5muR}}^{n}|\
\mbox{\boldmath$x$}=\bm{c}+L\mbox{\boldmath$z$},\mbox{\boldmath$z$}\in{\mathcal{B}}_{\infty}^{m}\\}$$=\operatorname*{\oplus}_{j=1}^{q}\\{t_{j}\bm{l}_{j},{\left|t_{j}\right|}\leq
1\\}\oplus\\{\bm{c}\\}$ is a zonotope of center $\bm{c}$, obtained by affine
transformation, of shape matrix $L\in{\mathop{\rm I\mkern-3.5muR}}^{n\times
m}$, of the unit box ${\mathcal{B}}_{\infty}^{m}$, where $m$ can be smaller,
equal to or greater than $n$. A zonotope is also a convex polyhedron with
centrally symmetric faces in all dimensions.
13. 13.
The support function of a set ${\mathcal{S}}\subset{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$ is $\rho_{\mathcal{S}}~{}:{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\rightarrow{\mathop{\rm I\mkern-3.5muR}}$,
$\displaystyle\mbox{\boldmath$x$}\mapsto\rho_{\mathcal{S}}(\mbox{\boldmath$x$}){\,:=\,}\sup_{\bm{u}\in{\mathcal{S}}}\mbox{\boldmath$u$}^{T}\mbox{\boldmath$x$}$.
${\mathcal{H}}\big{(}\mbox{\boldmath$x$},\rho_{\mathcal{S}}(\mbox{\boldmath$x$})\big{)}$
is the supporting hyperplane of ${\mathcal{S}}$ and
${\mathcal{S}}\subset{\mathcal{G}}\big{(}\mbox{\boldmath$x$},\rho_{\mathcal{S}}(\mbox{\boldmath$x$})\big{)}$.
$\displaystyle\rho_{{\mathcal{E}}(\bm{c},P)}(\mbox{\boldmath$x$})=\bm{c}^{T}\mbox{\boldmath$x$}+\sqrt{\mbox{\boldmath$x$}^{T}P\mbox{\boldmath$x$}}$
cf. [19].
## 2 PROBLEM FORMULATION
Consider the following linear discrete time system 0pt
$\displaystyle\mbox{\boldmath$x$}_{k}$
$\displaystyle=A_{k-1}\mbox{\boldmath$x$}_{k-1}+B_{k-1}\mbox{\boldmath$\tau$}_{k-1}+R_{k-1}\bm{w}_{k-1},\quad
k\in{\mathop{\rm I\mkern-3.5muN}}^{*}$ (1a) where
$\displaystyle\mbox{\boldmath$x$}_{0}$
$\displaystyle\in{\mathcal{E}}(\mbox{\boldmath$\hat{x}_{0}$},\varsigma_{0}P_{0}){\,=:\,}{\mathcal{E}}_{0}\subset{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\text{ and }\bm{w}_{k}\in{\mathcal{B}}_{\infty}^{m},$
(1b)
where $\mbox{\boldmath$x$}_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$,
$\mbox{\boldmath$\tau$}_{k}\in{\mathop{\rm I\mkern-3.5muR}}^{l}$ and
$\bm{w}_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{m}}$ are the unknown state
vector to be estimated, a known and bounded control vector and an unobservable
bounded process noise vector with unknown statistical characteristics,
_resp_.,
${\mathcal{E}}(\mbox{\boldmath$\hat{x}_{0}$},\varsigma_{0}P_{0}){\,=:\,}{\mathcal{E}}_{0}$
is a known ellipsoid (cf. $\S$ 1.1.10.), where
$\mbox{\boldmath$\hat{x}_{0}$}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$ is the
initial estimate of $\mbox{\boldmath$x$}_{k}$ at $k=0$,
$P_{0}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$ is a SPD matrix,
$\varsigma_{0}\in{\mathop{\rm I\mkern-3.5muR}}_{+}^{*}$ is a scaling positive
scalar, the product $\varsigma_{0}P_{0}$ is chosen as large as the confidence
in $\hat{x}_{0}$ is poor; ${\mathcal{B}}_{\infty}^{m}$ is the unit ball for
the $\infty-$norm in ${{{\mathop{\rm I\mkern-3.5muR}}}^{m}}$ (cf. $\S$
1.1.8.); $A_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$ and
$B_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times l}}$ are known state and
input matrices, _resp_. and $R_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times
m}}$ is the generator matrix defining the shape of the zonotope bounding the
unknown input vector
$\bar{\bm{w}}_{k}{\,:=\,}R_{k}\bm{w}_{k}\in{\mathcal{Z}}({\mathbf{0}}_{n},R_{k})$.
Now consider the output equation for the system (1):
$\displaystyle
F_{k}^{T}\mbox{\boldmath$x$}_{k}=\mbox{\boldmath$y$}_{k},\quad\mbox{\boldmath$y$}_{k}\in{\mathop{\rm
I\mkern-3.5muR}}^{p_{k}}$ (2a)
$\displaystyle\underaccent{\bar}{y}_{k_{i}}\leq{y_{k_{i}}}\leq\bar{y}_{k_{i}},\
i\in\mathscr{P}_{k}{\,:=\,}\\{1,\ldots,{p_{k}}\\}=\bar{\mathscr{D}}_{k}\cup\mathscr{H}_{k}\subset{\mathop{\rm
I\mkern-3.5muN}},$ (2b)
where, the output matrix
$F_{k}{\,:=\,}[\bm{f}_{k_{1}}\ldots,\bm{f}_{k_{p_{k}}}]\in{\mathop{\rm
I\mkern-3.5muR}}^{n\times{p_{k}}}$ is time varying and so is the number of its
columns, ${p}{\,:=\,}p_{k}\in{\mathop{\rm I\mkern-3.5muN}}$, which can be zero
sometimes (in the absence of measurements)222The output equation (2) can be
roughly derived from the one with measurements vector:
$\mbox{\boldmath$z$}_{k}=G_{k}^{T}\mbox{\boldmath$x$}_{k}+\mbox{\boldmath$v$}_{k}$,
subject to a noise vector belonging to a polyhedron:
$\mbox{\boldmath$v$}_{k}\in{\mathcal{P}}([C_{k}\
-C_{k}],[\bar{\bm{d}}_{k}^{T}\
-\underaccent{\bar}{\db}_{k}^{T}]^{T})\subset{\mathop{\rm
I\mkern-3.5muR}}^{2{p}}$, where
$\bar{\bm{y}}_{k}{\,:=\,}[\bar{{y}}_{k_{1}}\cdots\bar{{y}}_{k_{p}}]^{T}=~{}\\!\bar{\bm{d}}_{k}-C_{k}^{T}\mbox{\boldmath$z$}_{k}$,
$\underaccent{\bar}{\y}_{k}{\,:=\,}~{}\\![\underaccent{\bar}{{y}}_{k_{1}}\cdots\underaccent{\bar}{{y}}_{k_{p}}]^{T}=\underaccent{\bar}{\db}_{k}-C_{k}^{T}\mbox{\boldmath$z$}_{k}$
and $F_{k}=-G_{k}C_{k}$. . Indeed, the measurements are available in varying
amounts, at not all but only some sporadic, not a priori known, time steps
$k$. Three cases can be exhaustively enumerated: 1) for some
$i\in\mathscr{D}_{k}\subset\mathscr{P}_{k}$, both (finite and distinct) bounds
are available: $\underaccent{\bar}{y}_{k_{i}}<\bar{y}_{k_{i}}$; 2) for some
$i\in\mathscr{G}_{k}{\,:=\,}(\underaccent{\bar}{\mathscr{G}}_{k}\cup\bar{\mathscr{G}}_{k})\subset\mathscr{P}_{k}$,
only one bound, either $\bar{y}_{k_{i}}$ (if $i\in\bar{\mathscr{G}}_{k}$) or
$\underaccent{\bar}{y}_{k_{i}}$ (if $i\in\underaccent{\bar}{\mathscr{G}}_{k}$)
is available, in this case, the other (unavailable) bound is considered as
$\mp\infty$. 3) for some other $i\in\mathscr{H}_{k}\subset\mathscr{P}_{k}$,
the bounds are equal: $\underaccent{\bar}{y}_{k_{i}}=\bar{y}_{k_{i}}$. The
sets $\mathscr{D}_{k}$, $\bar{\mathscr{G}}_{k}$,
$\underaccent{\bar}{\mathscr{G}}_{k}$ and $\mathscr{H}_{k}$ form a partition
for $\mathscr{P}_{k}$:
$\mathscr{P}_{k}=\mathscr{D}_{k}\cup\bar{\mathscr{G}}_{k}\cup\underaccent{\bar}{\mathscr{G}}_{k}\cup\mathscr{H}_{k}$.
The measurement inequalities (2) can be rewritten as
0pt
$\displaystyle\bm{f}_{k_{i}}^{T}\mbox{\boldmath$x$}_{k}\leq\bar{y}_{k_{i}}\text{
and
}\underaccent{\bar}{y}_{k_{i}}\rightarrow-\infty\Leftrightarrow\mbox{\boldmath$x$}_{k}\in\bar{{\mathcal{G}}}_{k_{i}}{\,:=\,}{\mathcal{G}}(\bm{f}_{k_{i}},\bar{y}_{k_{i}}),$
$\displaystyle\text{ if }i\in\bar{\mathscr{G}}_{k},$ (3a)
$\displaystyle\bm{f}_{k_{i}}^{T}\mbox{\boldmath$x$}_{k}\geq\underaccent{\bar}{y}_{k_{i}}\text{
and
}\bar{y}_{k_{i}}\rightarrow+\infty\Leftrightarrow\mbox{\boldmath$x$}_{k}\in\underaccent{\bar}{\GC}_{k_{i}}{\,:=\,}{\mathcal{G}}(-\bm{f}_{k_{i}},-\underaccent{\bar}{y}_{k_{i}}),$
$\displaystyle\text{ if }i\in\underaccent{\bar}{\mathscr{G}}_{k},$ (3b)
$\displaystyle\bm{f}_{k_{i}}^{T}\mbox{\boldmath$x$}_{k}=\bar{y}_{k_{i}},\text{
and }\underaccent{\bar}{y}_{k_{i}}=\bar{y}_{k_{i}}\
\Leftrightarrow\mbox{\boldmath$x$}_{k}\in{\mathcal{H}}_{k_{i}}{\,:=\,}{\mathcal{H}}(\bm{f}_{k_{i}},\bar{y}_{k_{i}}),$
$\displaystyle\text{ if }i\in\mathscr{H}_{k},$ (3c)
$\displaystyle{\left|\tfrac{1}{\gamma_{k_{i}}}\bm{f}_{k_{i}}^{T}\mbox{\boldmath$x$}_{k}-{y_{k_{i}}}\right|}\leq
1\Leftrightarrow\mbox{\boldmath$x$}_{k}\in{\mathcal{D}}_{k_{i}}{\,:=\,}{\mathcal{D}}\Big{(}\tfrac{1}{\gamma_{k_{i}}}\bm{f}_{k_{i}},{y_{k_{i}}}\Big{)},\text{
otherwise }$ $\displaystyle(i\in\mathscr{D}_{k})$ (3d)
$\displaystyle\text{where
}\gamma_{k_{i}}{\,:=\,}\tfrac{\bar{y}_{k_{i}}-\underaccent{\bar}{y}_{k_{i}}}{2}\text{
and
}{y_{k_{i}}}{\,:=\,}\tfrac{\bar{y}_{k_{i}}+\underaccent{\bar}{y}_{k_{i}}}{2\gamma_{k_{i}}};$
(3e)
where ${\mathcal{G}}$, ${\mathcal{H}}$ and ${\mathcal{D}}$ are a halfspace, a
hyperplane and a strip resp. (cf. $\S$ 1.1.11.). The linear inequality
constraint on the state vector of the form (3a) (resp. (3b)) stands for a
measurement corrupted by an error whose only upper (_resp_. lower) bound is
known; the linear equality constraint on the state vector (3c) represent a
noiseless output; as for (3e), it acts as a measurement that is affected by a
bounded noise. All these three categories of outputs can obviously coexist,
i.e., occur at the same time-step $k$ and it is possible to add a forth one,
in which $\bar{y}_{k_{i}}\rightarrow+\infty$ and
$\underaccent{\bar}{y}_{k_{i}}\rightarrow+\infty$ meaning that no measurement
is available.
###### Assumptions 2.1
From now on, we assume that
1. 1.
all known matrices and vectors intervening in (1) and (3), as well as the SPD
$P_{0}$ and $\varsigma_{0}\in{\mathop{\rm I\mkern-3.5muR}}_{+}^{*}$ are
bounded;
2. 2.
all the columns of all the matrices intervening in (1) and those of $F_{k}$,
if any, are nonzero;
3. 3.
the matrix
$F_{k}({\mathscr{H}_{k}}){\,:=\,}[\bm{f}_{k_{i}}]_{i\in{\mathscr{H}_{k}}}$,
intervening in (3c), has full column rank (thus avoiding contradictory
constraints leading to an empty set);
###### Aims 2.2
We are intending here to design an estimator $\hat{\mbox{\boldmath$x$}}_{k}$
for the state vector $\mbox{\boldmath$x$}_{k}$ of the system (1)-(2), such
that,
1. 1.
a set (ellipsoid ${\mathcal{E}}_{k}$ of center
$\hat{\mbox{\boldmath$x$}}_{k}$) containing all possible values of the true
state vector $\mbox{\boldmath$x$}_{k}$ is quantified, at each time step
$k\in{\mathop{\rm I\mkern-3.5muN}}^{*}$ (standard requirement for a set-
membership approach);
2. 2.
the state estimate vector $\hat{\mbox{\boldmath$x$}}_{k}$ is _acceptable_ ,
i.e., it belongs to all the sets defined in (3).
3. 3.
under some conditions, the estimator $\hat{\mbox{\boldmath$x$}}_{k}$ is ISS,
(Input-to-State Stable, cf. Theorem 5.2). This is one of the distinguishing
features of the algorithm designed here.
The other distinguishing feature is that, unlike the other set-membership
techniques, such as those using exclusively intervals, zonotopes or polytopes,
the one detailed here delivers an optimal (_w.r.t._ some chosen criterions)
set, without any conservatism. Since the only measured information about the
true state vector $\mbox{\boldmath$x$}_{k}$ consists in its belonging to the
sets defined in (3), there is no better estimate than the one that belongs to
these sets. But such an estimator is not unique and is not necessarily stable
so the most suitable one will be chosen among the set of all possible
estimators by optimizing a given criterion.
Let
${\mathcal{E}}_{k}{\,:=\,}{\mathcal{E}}(\hat{\mbox{\boldmath$x$}}_{k},\varsigma_{k}P_{k})$
be the ellipsoid containing all possible values of the true state vector
$\mbox{\boldmath$x$}_{k}$. Note that the singular values of the shape matrix
$\varsigma_{k}P_{k}$ correspond to the semi-lengths of its axes, whose
directions are defined by the associated–orthogonal since $P_{k}$ is
symmetric–eigenvectors. In what follows, we have to determine the progression
law for the ellipsoid ${\mathcal{E}}_{k}$ (and thence for the state estimate
vector $\hat{\mbox{\boldmath$x$}}_{k}$) such that the aims $i$.–$iii$. are
fulfilled.
## 3 Time update (prediction stage)
In the two first paragraphs of this subsection, useful tools are established
in view of the development of the prediction algorithm in $\S$ 3.3.
### 3.1 Minkowski sum of an ellipsoid and a line segment
The lemma below gives the parametrized family of ellipsoids,
$\bar{{\mathcal{E}}}(\mu)$, that contains the Minkowski sum of the ellipsoid
$A{\mathcal{E}}$ (resulting from the endomorphism of matrix $A$ applied to the
ellipsoid ${\mathcal{E}}$) on one hand and the segment
${\mathcal{Z}}({\mathbf{0}}_{n},\bm{r}\bm{r}^{T})$, on the other.
###### Lemma 3.1
Let $\bm{c},\bm{u},\bm{r}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$, with
$\bm{r}\neq{\mathbf{0}}_{n}$, $A\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times
n}}$ and $P\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$ SPSD.
For any $\bm{x}\in{\mathcal{E}}{\,:=\,}{\mathcal{E}}(\bm{c},\varsigma P)$,
$\bm{w}\in{\mathcal{Z}}(\bm{u},\bm{r})$ and for any $\mu\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}$,
0pt $\displaystyle A\bm{x}+\bm{w}$
$\displaystyle\subset\bar{{\mathcal{E}}}(\mu){\,:=\,}{\mathcal{E}}\big{(}\bar{\bm{c}},\varsigma\bar{P}(\mu)\big{)},$
(4a) $\displaystyle\text{where }\qquad\qquad\qquad\qquad\bar{\bm{c}}$
$\displaystyle{\,:=\,}A\bm{c}+\bm{u},$ (4b) $\displaystyle\bar{P}(\mu)$
$\displaystyle{\,:=\,}(1+\mu){Q}+\tfrac{1+\mu}{\varsigma\mu}\bm{r}\bm{r}^{T}\text{,
where }{Q}{\,:=\,}APA^{T}.$ (4c)
Proof. cf. Appendix A.1. ❑
$\mu$ is a positive scalar parameter, chosen in such a way as to minimize the
size of $\bar{{\mathcal{E}}}(\mu)$, as detailed in the next paragraph.
### 3.2 Optimal values for the parameter $\bm{\mu}$
Now, the most telling two measures of the size of an ellipsoid, i.e.the volume
and the SSAL (sum of the squared axes lengths) will be minimized. Since the
eigenvalues of $\varsigma\bar{P}(\mu)$ are the squared semi-axes lengths of
$\bar{{\mathcal{E}}}(\mu)$, the former is proportional to their product, i.e.,
to ${\left|\varsigma\bar{P}(\mu)\right|}$ and the latter is equal to
$\varsigma\operatorname*{tr}\big{(}\bar{P}(\mu)\big{)}$.
#### 3.2.1 Pseudo-volume minimization
The equality constraints on the state vector, introduced by the measurements
$i~{}\\!\in\mathscr{H}_{k}$ and resulting in the intersection of the state
ellipsoid ${\mathcal{E}}_{k}$ with hyperplanes (studied in $\S$ 4.4), causes
the ellipsoid’s shape matrix $P_{k}$ to loose rank during the correction
stage, ensuing in its dimension reduction by zeroing some axes lengths and,
therefore, bestowing this ellipsoid a zero volume. Thereupon, we shall
introduce a generalized volume, the _pseudo-volume_ of an ellipsoid, when its
usual volume can be zero. Let us first recall that if
$P\in~{}\\!{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$ is SPD, then the
usual volume of an ellipsoid is
$\operatorname*{vol}\Big{(}{\mathcal{E}}(\bm{c},P)\Big{)}{\,=:\,}\dfrac{2\pi^{\frac{n}{2}}}{n\Gamma(\frac{n}{2})}{\left|P\right|}$,
where ${\left|P\right|}$ is the usual determinant of $P$ and $\Gamma$ denotes
the $\Gamma-$function [20].
###### Definition 3.1
For any SPSD matrix $P$ and any $\bm{c}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$, the pseudo-volume of the ellipsoid
${\mathcal{E}}\big{(}\bm{c},P)$ is proportional to the determinant of $P$:
$\displaystyle\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}(\bm{c},P)\big{)}$
$\displaystyle{\,=:\,}\operatorname*{vol}({\mathcal{B}}^{q}_{2}){\left|P\right|}_{{\dagger}},\text{
where
}\operatorname*{vol}({\mathcal{B}}^{q}_{2})=\tfrac{\pi^{\frac{q}{2}}}{\Gamma(\frac{q}{2}+1)},$
(5)
where $q{\,:=\,}\operatorname*{rank}(P)$ and
${\left|P\right|}_{{\dagger}}{\,=:\,}\displaystyle\lim_{t\rightarrow
0}\dfrac{{\left|P+tI_{n}\right|}}{t^{n-q}}$ is the pseudo-determinant of the
matrix $P$, i.e., the product of all its nonzero singular values. The pseudo-
volume of ${\mathcal{E}}(\bm{c},P)$ is nothing else than the volume of the
projection of ${\mathcal{E}}(\bm{c},P)$ onto $\mathpzc{R}(P)$.
###### Proposition 3.2
Let ${Q}_{+}{\,:=\,}b\big{(}Q+a\bm{r}\bm{r}^{T})$, $a,\>b\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}$; then
0pt
$\displaystyle\textit{i.}\qquad\qquad{q_{+}{\,:=\,}\operatorname*{rank}\big{(}{Q_{+}}\big{)}}$
$\displaystyle=\begin{cases}q{\,:=\,}\operatorname*{rank}(Q),&\text{ if
}\mbox{\boldmath$v$}={\mathbf{0}}_{n},\\\ q+1,&\text{ otherwise.}\end{cases}$
(6a) $\displaystyle\textit{ii.}\qquad\qquad{\left|{Q}_{+}\right|}_{{\dagger}}$
$\displaystyle=\begin{cases}b^{q}{\left|{Q}\right|}_{{\dagger}}\big{(}1+a\bm{r}^{T}\bm{u}\big{)},&\text{if
}\mbox{\boldmath$v$}={\mathbf{0}}_{n};\\\
b^{q+1}a\mbox{\boldmath$v$}^{T}\mbox{\boldmath$v$},&\text{otherwise;}\end{cases}$
(6b) 0pt $\displaystyle\text{where
}\qquad\bm{u}{\,:=\,}Q^{\dagger}\bm{r}\qquad\text{ and
}\qquad\mbox{\boldmath$v$}{\,:=\,}(I_{n}-QQ^{\dagger})\bm{r}.$ (6c)
Proof. cf. Appendix A.2. ❑
###### Theorem 3.3
$\bar{{\mathcal{E}}}(\mu)$ defined in (4) has the minimum pseudo-volume if
$\mu=~{}\\!\mu_{\text{v}}{\,:=\,}\arg\min_{\mu\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}(\bar{{\mathcal{E}}}(\mu)):$
0pt
$\displaystyle\mu_{\text{v}}{\,:=\,}\begin{cases}\tfrac{1}{2{q}}\sqrt{({q}-1)^{2}{h}^{2}+4{q}{h}}-\tfrac{{q}-1}{2{q}}{h},&\text{
if }\mbox{\boldmath$v$}={\mathbf{0}}_{n},\\\ \tfrac{1}{q},\text{ otherwise;
}\end{cases}$ (7a) $\displaystyle\text{where $\bm{u}$ and
$\mbox{\boldmath$v$}$ are defined in \eqref{eq_u_v_def},
}{q}{\,:=\,}\operatorname*{rank}(Q)\text{ and
}{h}{\,:=\,}\varsigma^{-1}\bm{r}^{T}\bm{u}.$ (7b)
Proof. cf. Appendix A.3. ❑
It is worth noting that he minimization of the volume of
$\bar{{\mathcal{E}}}(\mu)$ requires to determine the pseudo-inverse of the
matrix $P$. The result hereafter will serve to express
$\bar{P}(\mu)^{\dagger}$ by means of $Q^{\dagger}$, allowing to deduce it
recursively, without the need to compute it anew at each step. 0pt
###### Proposition 3.4
Let ${Q}_{+}{\,:=\,}b\big{(}Q+a\bm{r}\bm{r}^{T})$, $a\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}$; then
$\displaystyle{Q_{+}^{\dagger}}$
$\displaystyle=\tfrac{1}{b}\big{(}Q^{\dagger}+{\Delta}\big{)},\text{ with
}{\Delta}{\,:=\,}\begin{cases}\tfrac{1}{\left\|\bm{r}\right\|^{2}}\Big{(}\\!\tfrac{c}{\left\|\bm{r}\right\|^{2}}\mbox{\boldmath$v$}\mbox{\boldmath$v$}^{T}-\bm{u}\mbox{\boldmath$v$}^{T}-\mbox{\boldmath$v$}\bm{u}^{T}\\!\Big{)},&\\!\\!\\!\text{if
}\mbox{\boldmath$v$}\neq{\mathbf{0}}_{n};\\\
-\tfrac{1}{c}\bm{u}\bm{u}^{T},&\\!\\!\\!\text{otherwise.}\end{cases}$ (8a)
where $c{\,:=\,}\tfrac{1+a\bm{r}^{T}Q^{\dagger}\bm{r}}{a}$ and $\bm{u}$ and
$v$ are defined in (6c).
Proof. (6) can be obtained by applying Thm 1 and Thm 3 of [21]. ❑
#### 3.2.2 SSAL minimization
As for the minimization of the sum of the squared axes lengths of the
ellipsoid, _a.k.a._ the trace criterion, it is given directly by the following
theorem ensuing from the literature, where
$\operatorname*{ssal}\big{(}{\mathcal{E}}(\bm{c},\varsigma
P)\big{)}{\,:=\,}\varsigma\operatorname*{tr}(P)$.
###### Theorem 3.5 ([10])
$\bar{{\mathcal{E}}}(\mu)$ defined in (4) has the minimum SSAL if
$\mu=\mu_{\text{s}}$:
$\displaystyle\mu_{\text{s}}$
$\displaystyle{\,:=\,}\arg\min_{\mu\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}\operatorname*{ssal}\big{(}\bar{{\mathcal{E}}}(\mu)\big{)}=\arg\min_{\mu\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}\operatorname*{tr}\big{(}\bar{P}(\mu)\big{)}=\sqrt{\dfrac{\bm{r}^{T}\bm{r}}{\varsigma\operatorname*{tr}(APA^{T})}}.$
(9)
### 3.3 The time update algorithm
Let
${\mathcal{E}}_{k+1/k}{\,:=\,}{\mathcal{E}}(\hat{\bm{x}}_{k+1/k},\varsigma_{k}P_{k+1/k})$
be the ellipsoid including the “reachable set” of every possible value of
$\mbox{\boldmath$x$}_{k}\in{\mathcal{E}}_{k}{\,:=\,}{\mathcal{E}}(\hat{\mbox{\boldmath$x$}}_{k},\varsigma_{k}P_{k})$
that evolves according to the plant dynamics eq. (1a), subject to (1b). The
following theorem gives the parametrized family of ellipsoids
${\mathcal{E}}_{k+1/k}$ (of parameter $\bm{\mu}$) that contains
$A_{k}{\mathcal{E}}_{k}\oplus{\mathcal{Z}}({\mathbf{0}}_{n},{R}_{k})$.
###### Theorem 3.6 (Prediction stage)
If
$\mbox{\boldmath$x$}_{k}\in{\mathcal{E}}_{k}{\,:=\,}{\mathcal{E}}(\hat{\mbox{\boldmath$x$}}_{k},\varsigma_{k}P_{k})$
and $\mbox{\boldmath$x$}_{k}$ obeys to (1), then
$\forall\bm{\mu}{\,:=\,}(\mu_{1},\ldots,\mu_{m})^{T}\in]0,+\infty[^{m}$,
0pt
$\displaystyle\bm{x}_{k+1}\in{\mathcal{E}}(\hat{\bm{x}}_{k+1/k},\varsigma_{k}P_{k+1/k}){\,=:\,}{\mathcal{E}}_{k+1/k}{\,:=\,}{\mathcal{E}}_{{k+1/k}_{m}}$
0pt $\displaystyle\text{where }\quad{\mathcal{E}}_{{k+1/k}_{i}}$
$\displaystyle{\,:=\,}{\mathcal{E}}(\hat{\bm{x}}_{k+1/k},\varsigma_{k}{Q}_{k_{i}})\supseteq{\mathcal{E}}_{{k+1/k}_{i-1}}\oplus{\mathcal{Z}}({\mathbf{0}}_{n},{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T}),$
(10a) $\displaystyle\text{where }\quad\hat{\bm{x}}_{k+1/k}$
$\displaystyle{\,:=\,}A_{k}\hat{\mbox{\boldmath$x$}}_{k}+B_{k}\mbox{\boldmath$\tau$}_{k},$
(10b) $\displaystyle P_{k+1/k}$ $\displaystyle{\,:=\,}{Q}_{k_{m}},$ (10c)
$\displaystyle{Q}_{k_{0}}$ $\displaystyle{\,:=\,}A_{k}P_{k}A_{k}^{T};$ (10d)
$\displaystyle{Q}_{k_{i}}$
$\displaystyle{\,:=\,}(1+\mu_{i})\left({Q}_{k_{i-1}}+\tfrac{1}{\mu_{i}\varsigma_{k}}{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T}\right),\forall
i\in\\{1,\ldots,m\\};$ (10e)
${\bm{r}}_{k_{i}}$ (the $i$${}^{\textrm{th}}$ column of $R_{k}$) being the
generator vector of the zonotope containing all possible values of the process
noise $\bar{\bm{w}}_{k}$.
Proof. cf. Appendix A.4. ❑
Now, the results of $\S$ 3.2 are employed in order to express the optimal
predicted ellipsoid ${\mathcal{E}}_{k/k-1}$ according to the volume and trace
criterion respectively.
#### 3.3.1 Pseudo-volume minimization
###### Theorem 3.7
${\mathcal{E}}_{{k+1/k}_{i}}$ (cf. Thm 3.6) has the minimum pseudo-volume
if$\displaystyle\mu_{i}{\,:=\,}\mu_{k_{i}}{\,:=\,}\arg\min_{\mu_{i}\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}{\left|Q_{k_{i}}\right|}_{{\dagger}}=\mu_{\text{v}}$,
given by (7a), where 0pt
0pt $\displaystyle q$ $\displaystyle{\,:=\,}{q}_{{k+1/k}_{i-1}},\
h{\,:=\,}\varsigma_{k}^{-1}{\bm{r}}_{k_{i}}^{T}\bm{u},$ (11a)
$\displaystyle\bm{u}$
$\displaystyle{\,:=\,}\bar{Q}_{k_{i-1}}{\bm{r}}_{k_{i}},\
\mbox{\boldmath$v$}{\,:=\,}{\bm{r}}_{k_{i}}-Q_{k_{i-1}}\bm{u},$ (11b)
$\displaystyle{q}_{{k+1/k}_{i}}$
$\displaystyle{\,:=\,}\begin{cases}{q}_{{k+1/k}_{i-1}},&\text{ if
}\mbox{\boldmath$v$}={\mathbf{0}}_{n},\\\ {q}_{{k+1/k}_{i-1}}+1&\text{
otherwise;}\end{cases}$ (11c) $\displaystyle{{q}_{{k+1/k}_{0}}}$
$\displaystyle{\,:=\,}\begin{cases}{q}_{k}{\,:=\,}{q}_{k_{p}},&\text{ if
}\operatorname*{rank}(A_{k})=n;\\\ \operatorname*{rank}({Q}_{k_{0}}),&\text{
otherwise;}\end{cases}$ (11d) where
${q}_{k_{p}}{\,:=\,}\operatorname*{rank}(P_{k})$ (cf. (26q));
${q}_{{k+1/k}_{i}}\\!{\,:=\,}\operatorname*{rank}({Q}_{k_{i-1}})$,
$q_{k+1/k}{\,:=\,}{q}_{{k+1/k}_{m}}$ and
$\bar{Q}_{k_{i}}{\,:=\,}Q_{k_{i}}^{\dagger}$: $\displaystyle\bar{Q}_{k_{i}}$
$\displaystyle{\,:=\,}\tfrac{1}{1+\mu_{i}}\big{(}\bar{Q}_{k_{i-1}}+{\Delta_{k_{i}}}\big{)},\
i\in\\{1,\ldots,m\\},$ (11e) $\displaystyle\bar{Q}_{k_{0}}$
$\displaystyle{\,:=\,}Q_{k_{0}}^{\dagger}=\big{(}A_{k}P_{k}A_{k}^{T}\big{)}^{\dagger},$
(11f) $\displaystyle\text{with }\Delta_{k_{i}}$ $\displaystyle=\Delta\text{
given in \eqref{Delta_pseudo-inv-rec}, where
}\bm{r}{\,:=\,}{\bm{r}}_{k_{i}}\text{ and
}c{\,:=\,}{\varsigma_{k}\mu_{i}+{\bm{r}}_{k_{i}}^{T}\bm{u}}.$ (11g)
Proof. Direct consequence of Thm 3.3 and Proposition 3.4. ❑
Algorithm 1 resumes the last two theorems and computes
${\mathcal{E}}_{k+1/k}$, where
$\hat{\mbox{\boldmath$x$}}_{k}\rightarrow~{}\\!\mbox{\boldmath$x$}$,
$P_{k}\rightarrow~{}\\!P$, $\varsigma_{k}\rightarrow~{}\\!\varsigma$,
$q_{k}\rightarrow~{}\\!q$, $A_{k}\rightarrow~{}\\!A$,
$B_{k}\rightarrow~{}\\!B$,
$R_{k}=[{\bm{r}}_{k_{1}}\cdots{\bm{r}}_{k_{m}}]\rightarrow~{}\\!R$,
$\mbox{\boldmath$\tau$}_{k}\rightarrow~{}\\!\mbox{\boldmath$\tau$}$ and
$\bar{\mbox{\boldmath$x$}}\rightarrow~{}\\!\hat{\bm{x}}_{k+1/k}$,
$\bar{P}\rightarrow~{}\\!P_{k+1/k}$,
$\bar{\varsigma}\rightarrow~{}\\!\varsigma_{k+1/k}$,
$\bar{q}\rightarrow~{}\\!q_{k+1/k}{\,:=\,}{q}_{{k+1/k}_{m}}$. Let Predic_$\mu$
be a boolean variable set to “True” when choosing the pseudo-volume
minimization,
$\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}({\mathcal{E}}_{k/k-1_{i}})$,
and to “False” when minimizing the SSAL of ${\mathcal{E}}_{k/k-1_{i}}$.
Algorithm 1 Computation of the minimal pseudo-volume predicted ellipsoid
0: $x$, $P$, $\varsigma$, $A$, $B$, $R=[\bm{r}_{1}\cdots\bm{r}_{m}]$, $\tau$,
$q$
0: $\bar{\mbox{\boldmath$x$}}$, $\bar{P}$, $\bar{\varsigma}$, Predic_VolMin,
$\bar{q}$
1: $n\leftarrow$ size of $x$
2: $m\leftarrow$ number of columns of $R$
3: ${Q}{\,:=\,}APA^{T}$; {cf. (10d)}
4: $\bar{Q}{\,:=\,}{Q}^{\dagger}$; {cf. (11f)}
5: $q{\,:=\,}\operatorname*{rank}({Q})$; {this line can be skipped if
$\operatorname*{rank}({A_{k}})=n$; cf. (11d)}
6: for $i=1,\cdots,m$ do
7: $\bm{u}{\,:=\,}\bar{Q}\bm{r}_{i}$;
$\mbox{\boldmath$v$}{\,:=\,}\bm{r}_{i}-Q\bm{u}$; {cf. (11b)}
8: if $\mbox{\boldmath$v$}={\mathbf{0}}_{n}$ then
9: $h{\,:=\,}\varsigma\bm{r}_{i}^{T}\bm{u}$;
$\mu=\tfrac{1}{2{q}}\sqrt{({q}-1)^{2}{h}^{2}+4{q}{h}}-\tfrac{{q}-1}{2{q}}{h}$;
{cf. (11a), (7a) _resp_.}
10:
${\Delta}{\,:=\,}-\tfrac{1}{\varsigma\mu+\bm{r}_{i}^{T}\bm{u}}\bm{u}\bm{u}^{T}$;
{cf. (11g) and (8a)}
11: else
12: $\mu=\tfrac{1}{{q}}$; {cf. (7a)}
13: $q\leftarrow q+1$; {cf. (11d)}
14:
${\Delta}{\,:=\,}\tfrac{1}{\left\|\bm{r}_{i}\right\|^{2}}\bigg{(}\tfrac{{\varsigma\mu+\bm{r}_{i}^{T}\bm{u}}}{\left\|\bm{r}_{i}\right\|^{2}}\mbox{\boldmath$v$}\mbox{\boldmath$v$}^{T}-\bm{u}\mbox{\boldmath$v$}^{T}-\mbox{\boldmath$v$}\bm{u}^{T}\bigg{)}$;
{cf. (11g) and (8a)}
15: end if{cf. (7a)}
16:
${Q}{\,:=\,}(1+\mu)\big{(}{Q}+\tfrac{1}{\mu\varsigma}\bm{r}_{i}\bm{r}_{i}^{T}\big{)}$;
{cf. (10e)}
17: $\bar{Q}{\,:=\,}(1+\mu)^{-1}\big{(}\bar{Q}+{\Delta}\big{)}$; {cf. (11e)}
18: end for
19: $\bar{\mbox{\boldmath$x$}}\leftarrow
A\mbox{\boldmath$x$}+B\mbox{\boldmath$\tau$}$ {cf. (10b)};
$\bar{P}\leftarrow{Q}$ {cf. (10c)}; $\bar{\varsigma}\leftarrow\varsigma$;
$\bar{q}\leftarrow q$;Predic_VolMin $\leftarrow$ True;
###### Remark 3.1
It is worth noting that the volume minimization problem
$\displaystyle\arg\min_{\mu_{i}}{\left|{Q}_{k_{i}}\right|}_{{\dagger}}$ has an
explicit solution here. If the unknown input vector was bounded by an
ellipsoid, as was the case in [10, 12, 14, 22], rather than by an interval-
like set, such as a zonotope, $\displaystyle\mu_{{\text{v}}_{i}}$ would be the
unique positive root of an $n-$order polynomial. Nevertheless, considering
that the pseudo-inverse of a $n\times n$ matrix is needed at each time step
$k$, in line with (11f), the trace minimization is more appealing, at least
from the computational point of view.
#### 3.3.2 Sum of the squared axes’ lengths minimization
###### Theorem 3.8
${\mathcal{E}}_{k+1/k}$ defined in Thm 3.6 has the minimum SSAL, if
$\bm{\mu}{\,:=\,}\bm{\mu}_{k}{\,:=\,}\arg\min_{\mu_{i}\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}\operatorname*{ssal}({\mathcal{E}}_{k+1/k})=\bm{\mu}_{\text{s}}$,
0pt
$\displaystyle\text{where }\mu_{{\text{s}}_{i}}$
$\displaystyle{\,:=\,}\arg\min_{\mu_{i}\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}\operatorname*{tr}(Q_{k_{i}})=\sqrt{\dfrac{{\bm{r}}_{k_{i}}^{T}{\bm{r}}_{k_{i}}}{\varsigma_{k}\operatorname*{tr}(Q_{k_{i-1}})}},i\in\\{1,\cdots,m\\},$
(12)
and the recursive formula (10c)-(10e) becomes:
0pt $\displaystyle P_{k+1/k}$
$\displaystyle=\big{(}1+\tfrac{\bar{\mu}}{\bar{\mu}_{0}}\big{)}\big{(}Q_{k_{0}}+\tfrac{\bar{\mu}_{0}}{\varsigma_{k}}\tilde{R}_{k}\big{)},$
(13a) $\displaystyle{\text{where }}\qquad\bar{\mu}_{0}$
$\displaystyle{\,:=\,}\sqrt{\varsigma_{k}\operatorname*{tr}(Q_{k_{0}})}\text{
and $Q_{k_{0}}$ given in \eqref{P_pred0},}$ (13b)
$\displaystyle\bar{\mu}_{k}{\,:=\,}\sum_{i=1}^{m}{\bar{\mu}_{i}},\
\bar{\mu}_{i}{\,:=\,}{\left\|{\bm{r}}_{k_{i}}\right\|}\text{ and
}\tilde{R}_{k}{\,:=\,}\sum_{i=1}^{m}\bar{\mu}_{i}^{-1}{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T}=R_{k}{\operatorname*{Diag}\big{(}\tfrac{1}{\bar{\mu}}\big{)}_{i=1}^{m}}R_{k}^{T}.$
(13c)
Proof. cf. Appendix A.5. ❑
Algorithm 2 computes the value of $\bm{\mu}_{\text{s}}$ given in Thm 3.8 .
Algorithm 2 Computation of the minimal SSAL predicted ellipsoid
0: $x$, $P$, $\varsigma$, $A$, $B$, $R=[\bm{r}_{1}\cdots\bm{r}_{m}]$, $\tau$
0: $\bar{}\mbox{\boldmath$x$}$, $\bar{P}$, $\bar{\varsigma}$, Predic_$\mu$
1: $n\leftarrow$ size of $x$;
2: $m\leftarrow$ number of columns of $R_{k}$;
3: ${Q}{\,:=\,}AP_{k}A^{T}$; {cf. (10d)}
4: $\bar{\mu}_{0}{\,:=\,}\sqrt{\varsigma\operatorname*{tr}(Q)}$; {cf. (13b)}
5: $\bar{\mu}=0$;
6: $\tilde{R}=0_{n\times n}$;
7: for $i=1,\cdots,m$ do
8: $\epsilon{\,:=\,}{\left\|\bm{r}_{i}\right\|}$;
9: $\bar{\mu}\leftarrow\bar{\mu}+\epsilon$;
10: $\tilde{R}\leftarrow\tilde{R}+\frac{1}{\epsilon}\bm{r}_{i}\bm{r}_{i}^{T}$;
11: end for
12: $\bar{}\mbox{\boldmath$x$}\leftarrow
A\mbox{\boldmath$x$}+B\mbox{\boldmath$\tau$}$ {(10b)};
$\bar{P}\leftarrow\big{(}1+\tfrac{\bar{\mu}}{\bar{\mu}_{0}}\big{)}\big{(}Q+\tfrac{\bar{\mu}_{0}}{\varsigma}\tilde{R}\big{)}${(13a)};
$\bar{\varsigma}\leftarrow\varsigma$; Predic_$\mu$ $\leftarrow$ False;
###### Remark 3.2
When minimizing the SSAL of ${\mathcal{E}}_{k/k-1}$, it is not necessary to
compute the $m$ intermediate values of ${Q}_{k_{i}}$, given by the recursive
formula (10c)-(10e). $P_{k/k-1}$ can be computed directly using (13) instead.
Note also that all ${\bar{\mu}_{k_{i}}}$ are nonzero thanks to the assumption
2.
###### Remark 3.3
It is possible to minimize the weighted sum of the squared axes lengths of
${\mathcal{E}}_{k+1/k}$: $\operatorname*{tr}(CP_{k+1/k}C^{T})$, for any
$C\in{\mathop{\rm I\mkern-3.5muR}}^{n_{C}\times n}$, $n_{C}\in{\mathop{\rm
I\mkern-3.5muN}}^{*}$. In this case, instead of (12), the optimal value for
$\bm{\mu}$ would be
$\bm{\mu}_{k_{\text{s}}}{\,:=\,}\bm{\mu}_{\text{s}}{\,:=\,}[\mu_{{\text{s}}_{1}},\ldots,\mu_{{\text{s}}_{m}}]$,
where
$\displaystyle\mu_{{\text{s}}_{i}}$
$\displaystyle{\,:=\,}\sqrt{\frac{\mbox{\boldmath$r$}_{{k-1}_{i}}^{T}C^{T}C\mbox{\boldmath$r$}_{{k-1}_{i}}}{\varsigma_{k-1}\operatorname*{tr}(C{Q}_{k_{i-1}}C^{T})}},\
i\in\\{1,\cdots,m\\},$ (14a) and the equations (13c) would be replaced by
$\displaystyle\bar{\mu}_{k}$
$\displaystyle{\,:=\,}\sum_{i=1}^{m}{\bar{\mu}_{k_{i}}},\quad\bar{\mu}_{k_{i}}{\,:=\,}{\left\|C{\bm{r}}_{k_{i}}\right\|}\text{
and
}\tilde{R}_{k}{\,:=\,}\sum_{i=1}^{m}\bar{\mu}_{k_{i}}^{-1}C{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T}C^{T}.$
(14b)
Given the ellipsoid ${\mathcal{E}}_{k}$ at the time step $k$, Thm 3.6 provides
the predicted ellipsoid ${\mathcal{E}}_{k+1/k}$ whose center is given by (10b)
and whose shape matrix is given, up to the factor $\varsigma_{k}$, by the
recursive formula (10c)-(10e), which depends on $\bm{\mu}$; Thms 3.7 and 3.8
offer the optimal values for this parameter according to two criterions, whose
choice is left to the reader. The author will prefer the trace minimization,
for computational reason explained in Remark 3.1.
Discussions on how to choose the parameters $\mu_{i}$ to achieve more
optimality conditions for the ellipsoid bounding the reachable set for
continuous-time systems are proposed in [13] and [23]. See also survey books
[19], [11] for a more complete overview.
## 4 Measurement update (correction stage)
The dynamic state evolution equation (1) allowed to compute the predicted
ellipsoid ${\mathcal{E}}_{k/k-1}$ which contains all possible values of the
state vector $\mbox{\boldmath$x$}_{k}$ taking into account all the
measurements up to time step $k-1$ if any. Now, let us recall the other sets
containing $\mbox{\boldmath$x$}_{k}$, obtained from the measurements:
$\displaystyle\eqref{bounds}$
$\displaystyle\Leftrightarrow\mbox{\boldmath$x$}_{k}\in\bigcap_{i\in\mathscr{G}_{k}}{\mathcal{G}}_{k_{i}}\cap\bigcap_{i\in\mathscr{D}_{k}}{\mathcal{D}}_{k_{i}}\cap\bigcap_{i\in\mathscr{H}_{k}}{\mathcal{H}}_{k_{i}},\text{
if }{p}\neq 0.$ (15)
${p}{\,:=\,}{p_{k}}$ is the number of measurements at time-step $k$. It is
interesting to note that the intersection of half-spaces can be considered as
a possibly unbounded polyhedron and that the intersection of strips is a
zonotope:
$\displaystyle\bigcap_{i\in\mathscr{G}_{k}}{\mathcal{G}}_{k_{i}}$
$\displaystyle{\,=:\,}{\mathcal{P}}_{k}{\,:=\,}{\mathcal{P}}\big{(}[\bm{f}_{k_{i}}]_{i\in\mathscr{G}_{k}},[{y_{k_{i}}}]_{i\in\mathscr{G}_{k}}\big{)},$
(16) $\displaystyle\bigcap_{i\in\mathscr{D}_{k}}{\mathcal{D}}_{k_{i}}$
$\displaystyle{\,=:\,}{\mathcal{Z}}_{k}{\,:=\,}{\mathcal{Z}}^{{\mathcal{H}}}\big{(}[\bm{f}_{k_{i}}]_{i\in\mathscr{D}_{k}},[{y_{k_{i}}}]_{i\in\mathscr{G}_{k}}\big{)}.$
(17)
The correction stage consists in performing the intersection between
${\mathcal{E}}_{k/k-1}$ and the set (15), allowing to find
${\mathcal{E}}_{k}\supset{\mathcal{S}}_{k}$ in light of the current
measurements, where
$\displaystyle{\mathcal{S}}_{k}$
$\displaystyle{\,:=\,}\Big{(}\big{(}{\mathcal{E}}_{k/k-1}\cap\bigcap_{i\in\mathscr{G}_{k}}{\mathcal{G}}_{k_{i}}\big{)}\cap\bigcap_{i\in\mathscr{D}_{k}}{\mathcal{D}}_{k_{i}}\Big{)}\cap\bigcap_{i\in\mathscr{H}_{k}}{\mathcal{H}}_{k_{i}}$
$\displaystyle=\Big{(}\big{(}{\mathcal{E}}_{k/k-1}\cap{\mathcal{P}}_{k}\big{)}\cap{\mathcal{Z}}_{k}\Big{)}\cap\bigcap_{i\in\mathscr{H}_{k}}{\mathcal{H}}_{k_{i}}.$
(18)
It will be shown that this intersection is the one between
${\mathcal{E}}_{k/k-1}$ and the possibly degenerate (if
$\mathscr{H}_{k}\neq\emptyset$) zonotope. It does not result in an ellipsoid
in general and has to be circumscribed by such a set, which is the subject of
the upcoming paragraphs. We shall begin by working on the intersection
$\displaystyle{\mathcal{E}}_{k/k-1}\cap{\mathcal{G}}_{k_{i}}$ in $\S$ 4.1.
Secondly, we’ll be dealing with the intersection between an ellipsoid and a
strip in order to carry out the set obtained in $\S$ 4.1 and intersecting it
with $\displaystyle\bigcap_{i\in\mathscr{D}_{k}}{\mathcal{D}}_{k_{i}}$; $\S$
4.2 provides the optimal ellipsoid overbounding this intersection. Thirdly,
the intersection of an ellipsoid with a hyperplane will be presented in $\S$
4.4, in order to handle the intersection of the previously obtained ellipsoid
with $\displaystyle\bigcap_{i\in\mathscr{H}_{k}}{\mathcal{H}}_{k_{i}}$.
Finally, all these results will be compiled in a unique state estimation
algorithm in $\S$ 5.1.
### 4.1 Intersection of an ellipsoid with a halfspace
The intersection between the ellipsoid ${\mathcal{E}}_{k/k-1}$ obtained in
$\S$ 3 and the polyhedron ${\mathcal{P}}_{k}$ can be reformulated as the
intersection of ${\mathcal{E}}_{k/k-1}$ and a series of strips
${\mathcal{D}}_{k_{i}}$. To grasp this idea, take any closed convex set
${\mathcal{S}}$ and a hyperplane ${\mathcal{H}}$ intersecting it. The
intersection of ${\mathcal{S}}$ with a halfspace ${\mathcal{G}}$ delimited by
${\mathcal{H}}$ is nothing else that its intersection with the strip formed
between ${\mathcal{H}}$ and a support hyperplane of ${\mathcal{S}}$, parallel
to ${\mathcal{H}}$ and contained in ${\mathcal{G}}$. Now, if ${\mathcal{H}}$
doesn’t intersect ${\mathcal{S}}$, the latter is either a subset of
${\mathcal{G}}$ or lies outside of it, and if ${\mathcal{H}}$ is tangent to
${\mathcal{S}}$ (being its support hyperplane), then ${\mathcal{S}}$ is either
again a subset of ${\mathcal{G}}$ or it has only one point in common with it.
In the case where ${\mathcal{S}}$ is an ellipsoid and the intersecting
halfspace corresponds to the constraint (3a), the theorem below provides the
parameters of the intersecting strip. To obtain the intersection of an
ellipsoid with the halfspace given by the constraint (3b), it suffices to
replace $\bm{f}$ by $-\bm{f}$ and $\bar{y}$ by $\underaccent{\bar}{y}$:
###### Theorem 4.1 (ellipsoid-halfspace intersec.)
Let $\bm{c}\in~{}{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$,
$\bm{f}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\text{--}\\{{\mathbf{0}}_{n}\\}$, $P\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n\times n}}$ SPSD, $\varsigma\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}$ and $\bar{y}\in{\mathop{\rm I\mkern-3.5muR}}$.
If $\bar{y}<-\underaccent{\bar}{\rho}$, (case 1)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{G}}(\bm{f},\bar{y})$ $\displaystyle=\emptyset;$ (19b) else if
$\bar{y}\geq\bar{\rho}$, (case 2) $\displaystyle{\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{G}}(\bm{f},\bar{y})$
$\displaystyle={\mathcal{E}}(\bm{c},\varsigma P);$ (19c) else if
$\bar{y}=-\underaccent{\bar}{\rho}$, (case 3)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{G}}(\bm{f},\bar{y})$
$\displaystyle={\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{H}}(\bm{f},-\underaccent{\bar}{\rho})=\\{\bm{c}-\varsigma^{\frac{1}{2}}(\bm{f}^{T}P\bm{f})^{-\frac{1}{2}}P\bm{f}\\};$
(19d) else ($-\underaccent{\bar}{\rho}<\bar{y}<\bar{\rho}$), (case 4)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{G}}(\bm{f},\bar{y})$
$\displaystyle={\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{D}}(\tfrac{1}{\gamma}\bm{f},{y}),$ (19e) where
$\displaystyle\gamma$
$\displaystyle{\,:=\,}\tfrac{1}{2}(\bar{y}+\underaccent{\bar}{\rho})\text{ and
}{y}{\,:=\,}\tfrac{1}{2\gamma}(\bar{y}-\underaccent{\bar}{\rho})$ (19f)
$\displaystyle\underaccent{\bar}{\rho}$
$\displaystyle{\,:=\,}\rho_{{\mathcal{E}}(\bm{c},\varsigma
P)}(-\bm{f})=-\bm{c}^{T}\bm{f}+\sqrt{\varsigma\bm{f}^{T}P\bm{f}}\text{ ({\it
cf.\ }$\S$~{}\ref{subsec_notations} \ref{Support function})}$ (19g)
$\displaystyle\bar{\rho}$
$\displaystyle{\,:=\,}\rho_{{\mathcal{E}}(\bm{c},\varsigma
P)}(\bm{f})=\bm{c}^{T}\bm{f}+\sqrt{\varsigma\bm{f}^{T}P\bm{f}}.$ (19h)
Proof. cf. Appendix B.1. ❑
The figure 1 illustrates the above theorem in the case where $n=3$
($\mbox{\boldmath$x$}_{k}\in{\mathop{\rm I\mkern-3.5muR}}^{3}$). The halfspace
is depicted by the shaded region and its upper boundary is the colored
hyperplane.
Case 1:
when $\bar{y}-\bm{c}^{T}\bm{f}<-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$, the
ellipsoid ${\mathcal{E}}(\bm{c},\varsigma P)$ is outside the halfspace
${\mathcal{G}}(\bm{f},\bar{y})$ and their intersection is thus empty.
Case 2:
when $\bar{y}-\bm{c}^{T}\bm{f}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$, the
ellipsoid is entirely contained in the halfspace and their intersection is
nothing else than the ellipsoid ${\mathcal{E}}(\bm{c},\varsigma P)$ itself.
Case 3:
when $\bar{y}-\bm{c}^{T}\bm{f}=-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$, the
ellipsoid is tangent to the halfspace and the intersection is reduced to one
single point given by (19d) represented by a red “o” on the figure 1\- (Case
3).
Case 4:
when
${\left|\bar{y}-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$,
the intersection between the ellipsoid, ${\mathcal{E}}(\bm{c},\varsigma P)$
and the halfspace ${\mathcal{G}}(\bm{f},\bar{y})$ is the same as the
intersection of this ellipsoid with a strip
${\mathcal{D}}(\frac{1}{\gamma}\bm{f},y)$ (dark shaded area), bounded on one
hand by the hyperplane ${\mathcal{H}}(\bm{f},\bar{y})$ (cyan), the boundary of
${\mathcal{G}}(\bm{f},\bar{y})$, and on the other hand, by the support
hyperplane, ${\mathcal{H}}(-\bm{f},-\underaccent{\bar}{\rho})$ (violet) of the
ellipsoid.
(a) Case 1: $\bar{y}-\bm{c}^{T}\bm{f}<-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(b) Case 2: $\bar{y}-\bm{c}^{T}\bm{f}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(c) Case 3: $\bar{y}-\bm{c}^{T}\bm{f}=-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(d) Case 4:
${\left|\bar{y}-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
Figure 1: Illustration of the four cases of Theorem 4.1 for $n=3$
### 4.2 Ellipsoid bounding the intersection of an ellipsoid with a strip
In the previous paragraph, we showed that the incorporation of the
measurements $i\in\mathscr{G}_{k}$ result, as for those $i\in\mathscr{D}_{k}$,
from the intersection of the predicted ellipsoid with a zonotope, formulated
as an intersection of several strips. We need now to overbound this
intersection by an ellipsoid. To begin with, the theorem below presents a
family of parametrized ellipsoids (of parameter $\beta$) that contain an
ellipsoidal layer, coming out of the intersection of
${\mathcal{E}}(\bm{c},\varsigma P)$ with the strip ${\mathcal{D}}(\bm{f},y)$,
which can be considered–interestingly enough–as an ellipsoid unbounded in the
direction orthogonal to $\bm{f}$.
###### Theorem 4.2 (ellips./strip inters.)
Let $\bm{c}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$,
$\varsigma\in{\mathop{\rm I\mkern-3.5muR}}_{+}^{*}$, $y\in{\mathop{\rm
I\mkern-3.5muR}}$, $P\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$ SPSD
and $\bm{f}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\text{--}\\{{\mathbf{0}}_{n}\\}$,
if $y-1>\bar{\rho}\vee y+1<-\underaccent{\bar}{\rho}$, (case 1)
$\displaystyle{\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\varsigma P)$
$\displaystyle=\emptyset;$ (20b) else if $y+1\geq\bar{\rho}\wedge
y-1\leq-\underaccent{\bar}{\rho}$, (case 2)
$\displaystyle{\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\varsigma P)$
$\displaystyle={\mathcal{E}}(\bm{c},\varsigma P);$ (20c) else if
$y=-\underaccent{\bar}{\rho}-1$, (case 3.a)
$\displaystyle{\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\varsigma P)$
$\displaystyle={\mathcal{H}}(\bm{f},-\underaccent{\bar}{\rho})\cap{\mathcal{E}}(\bm{c},\varsigma
P)=\\{\bm{c}-\varsigma^{\frac{1}{2}}(\bm{f}^{T}P\bm{f})^{-\frac{1}{2}}P\bm{f}\\};\\!\\!\\!\\!\\!$
(20d) else if $y=\bar{\rho}+1$, (case 3.b)
$\displaystyle{\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\varsigma P)$
$\displaystyle={\mathcal{H}}(\bm{f},\bar{\rho})\cap{\mathcal{E}}(\bm{c},\varsigma
P)=\\{\bm{c}+\varsigma^{\frac{1}{2}}(\bm{f}^{T}P\bm{f})^{-\frac{1}{2}}P\bm{f}\\};$
(20e) else
$(-\underaccent{\bar}{\rho}<y+1<\bar{\rho}\vee-\underaccent{\bar}{\rho}<y-1<\bar{\rho}),\
\forall\beta\in]0,1[$, (case 4)
$\displaystyle{\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\varsigma
P)={\mathcal{D}}(\breve{\bm{f}},\breve{y})\cap{\mathcal{E}}(\bm{c},\varsigma
P)\subset{\mathcal{E}}\big{(}\breve{\bm{c}}(\beta),\breve{\varsigma}(\beta)\breve{P}(\beta)\big{)}{\,=:\,}\breve{{\mathcal{E}}}(\beta),$
(21b) 0pt $\displaystyle\text{where
}\breve{\bm{f}}{\,:=\,}\tfrac{1}{\gamma}\bm{f}\text{ and
}\breve{y}{\,:=\,}\tfrac{1}{2\gamma}(\bar{y}+\underaccent{\bar}{y})=\tfrac{1}{\gamma}(\bm{f}^{T}\bm{c}+\delta),$
(21c) $\displaystyle\breve{P}(\beta)$
$\displaystyle{\,:=\,}P-{\alpha\beta}P\bm{f}\bm{f}^{T}P=P-{\alpha\beta}\bm{\varphi}\bm{\varphi}^{T},\quad$
(21d) $\displaystyle\breve{\bm{c}}(\beta)$
$\displaystyle{\,:=\,}\bm{c}+{\alpha\beta}\delta
P\bm{f}=\bm{c}+{\alpha\beta}\delta\bm{\varphi},$ (21e)
$\displaystyle\breve{\varsigma}(\beta)$
$\displaystyle{\,:=\,}\varsigma+{\alpha\beta}\left(\tfrac{\gamma^{2}}{1-\beta}-\delta^{2}\right),$
(21f) $\displaystyle\alpha$
$\displaystyle{\,:=\,}\big{(}\bm{f}^{T}P\bm{f}\big{)}^{-1}=\big{(}\bm{f}^{T}\bm{\varphi}\big{)}^{-1}\text{
and }\bm{\varphi}{\,:=\,}P\bm{f},$ (21g) $\displaystyle\delta$
$\displaystyle{\,:=\,}\tfrac{1}{2}(\bar{y}+\underaccent{\bar}{y})-\bm{f}^{T}\bm{c}=\tfrac{1}{2}(\bar{y}+\underaccent{\bar}{y}-\bar{\rho}+\underaccent{\bar}{\rho}),$
(21h) $\displaystyle\gamma$
$\displaystyle{\,:=\,}\tfrac{1}{2}(\bar{y}-\underaccent{\bar}{y}),$ (21i)
$\displaystyle\bar{y}$ $\displaystyle{\,:=\,}\min(y+1,\bar{\rho})\text{ and
}\underaccent{\bar}{y}{\,:=\,}\max(y-1,-\underaccent{\bar}{\rho}),$ (21j)
and $\underaccent{\bar}{\rho}$ and $\bar{\rho}$ are defined in (19g)–(19h).
Proof. cf. Appendix B.2. ❑
(a) (1a): ${y-\bm{c}^{T}\bm{f}}<-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}-1$
(b) (1b): ${y-\bm{c}^{T}\bm{f}}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}+1$
(c) (2):
${\left|y-\bm{c}^{T}\bm{f}\right|}<1-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(d) (3a): ${y-\bm{c}^{T}\bm{f}}=-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}-1$
(e) (3b): ${y-\bm{c}^{T}\bm{f}}=\sqrt{\varsigma\bm{f}^{T}P\bm{f}}+1$
(f) (4a):
${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
and ${\left|y-1-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(g) (4b):
${\left|y+1-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
and ${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(h) (4c):
${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
and ${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
Figure 2: Illustration of Theorem 4.2 for $n=3$
The figure 2 illustrates the above theorem in the case where $n=3$. The strip
is depicted by the shaded region and its boundaries are the colored
hyperplanes.
Case 1:
when ${\left|y-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}+1$,
the strip is either below (cf. fig. 2.1a) or above (cf. fig. 2.1b) with empty
intersection.
Case 2:
when ${\left|y-\bm{c}^{T}\bm{f}\right|}<1-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$,
the ellipsoid is entirely contained in the strip (cf. fig. 2.2) and the
intersection is the ellipsoid ${\mathcal{E}}(\bm{c},\varsigma P)$ itself.
Case 3:
when ${\left|y-\bm{c}^{T}\bm{f}\right|}=\sqrt{\varsigma\bm{f}^{T}P\bm{f}}+1$,
the ellipsoid is tangent to the strip and the intersection is reduced to a
single point given by (20d) (represented by a red “o” on fig. 2.3a) or (20e)
(white “o” on fig. 2.3.b).
Case 4:
when ${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
or ${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$.
This case can be decomposed in three sub-cases (not appearing in the theorem):
4.a:
if ${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$ and
${\left|y-1-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$,
meaning that the hyperplane ${\mathcal{H}}(\bm{f},y+1)$ representing the upper
boundary (cyan) of the strip (the strip is depicted by all the shaded region
in fig. 2.4a) intersects the ellipsoid while the lower one
${\mathcal{H}}(\bm{f},y-1)$ (magenta) does not; the strip is then reduced in
width (to the new dark shaded strip) replacing the latter hyperplane by a
support hyperplane (violet) of the ellipsoid,
${\mathcal{H}}(\bm{f},\underaccent{\bar}{y})$, where
$\underaccent{\bar}{y}{\,:=\,}\bm{c}^{T}\bm{f}-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$.
4.b:
if ${\left|y+1-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$ and
${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$,
meaning that the hyperplane ${\mathcal{H}}(\bm{f},y-1)$ (green) representing
the lower boundary of the strip is intersecting the ellipsoid while the upper
one ${\mathcal{H}}(\bm{f},y+1)$ (magenta) is not; the strip is then reduced in
width replacing the latter by ${\mathcal{H}}(\bm{f},\bar{y})$ (yellow) where
$\bar{y}{\,:=\,}\bm{c}^{T}\bm{f}+\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$.
4.c:
if ${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$ and
${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$,
meaning that both the hyperplanes ${\mathcal{H}}(\bm{f},y-1)$ and
${\mathcal{H}}(\bm{f},y+1)$ representing _resp_. the lower and upper
boundaries of the strip are intersecting the ellipsoid.
(a) (4a):
${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
and ${\left|y-1-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(b) (4b):
${\left|y+1-\bm{c}^{T}\bm{f}\right|}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
and ${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(c) (4c):
${\left|y+1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
and ${\left|y-1-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
Figure 3: Illustration of the case 4 Theorem 4.2 for $n=2$
After the strip is reduced, if necessary (case 4.a and 4.b), the intersection
between the new strip and the ellipsoid is overbounded by a family of
parametrized, function of $\beta$, ellipsoids. The red ellipsoids depicted in
3.4a, 3.4b and 3.4c are only one possible occurrence of this family of
ellipsoids for a particular value of $\beta$ given in (22).
### 4.3 Optimal values for the parameter $\beta$
In this section, the optimal value of the weighting parameter $\beta$ with
respect to a judiciously chosen criterion will be derived.
Three optimization criteria have been used in literature for the parameter
identification. The main two involve set measures on the ellipsoid and were
first proposed by Fogel and Huang in their well-known pioneer paper [17], in
the parameter identification framework. They gave two optimal values for the
weighting parameter $\omega{\,:=\,}\tfrac{\alpha\beta}{1-\beta}$: the first
minimizing the determinant of $\tilde{\varsigma}(\omega)\tilde{P}(\omega)$
(where $\varsigma=1$) in Lemma B.2 and the second, its trace, thus optimizing
the volume and the sum, _resp_., of the squared semi-axes lengths of the
ellipsoid ${\mathcal{E}}\big{(}\tfrac{\omega}{\omega+\alpha}\big{)}$, defined
in (21). Five years later, Dasgupta and Huang, in [24], designed a modified
least-squares parameter identification algorithm with a forgetting factor,
where they used a weighting parameter $0\leq\nu\leq 1$ (the forgetting factor
being $1-\nu$), which could be roughly related to the one used in [17] by
$\nu=\tfrac{\omega}{\omega+1}$ and they introduced a new optimization
criterion for $\nu$ based on the minimization of $\tilde{\varsigma}$. Nayeri,
Deller, Liu et al. , actively studied all these aspects during the nineties
[25, 26, 27, 28, 29], proposing a set-membership stochastic approximation
identification algorithm and, in [18], a unified framework of the general
class of optimal bounding ellipsoid (OBE) for all the methods previously
cited, on the basis of the weighted least squares identification method. They
focused on the volume minimization criterion, for evident reasons.
Even if the convergence of such algorithms in the particular case of
unconstrained set-membership parameter identification framework333The
parameter identification can be seen, in a reductive manner, as a particular
case of the state estimation where the estimated vector is supposed to be
constant, i.e., $A_{k}=I_{n}$, $B_{k}=0_{n,l}$ and $R_{k}=0_{n,m}$ in (1a).
was proven for the three addressed criteria, each having its own interesting
properties, when it comes to the set-membership state estimation, it is
definitely an open issue.
#### 4.3.1 Minimization of the worst case weighted estimation error
In contrast with all the algorithms in the literature, [10, 11, 12, 13], that
minimize the size of the ellipsoid $\breve{{\mathcal{E}}}(\beta)$, the optimal
value of $\beta$ developed in this paragraph is obtained by minimizing some
quadratic measure of the estimation error vector444represented by its
candidate Lyapunov function $\breve{\mathscr{V}}_{\beta}(\mbox{\boldmath$x$})$
in the worst noise case, embodied by $\breve{\varsigma}$, in the manner of
[14, 22], inspired by some identification algorithms [24, 30, 31].
###### Theorem 4.3.1
Consider the case 4 of Thm 4.2 where $\breve{{\mathcal{E}}}(\beta)$,
$\beta\in]0,1[$, is given by (21), then $\breve{\varsigma}(\beta)$ defined in
(21f) satisfies
0pt $\displaystyle{\varsigma}(\beta)$
$\displaystyle=\\!\max_{\bm{x}\in{\mathcal{D}}(\breve{\bm{f}},\breve{y})\cap{\mathcal{E}}(\bm{c},\varsigma
P)}\\!\\!{\mathscr{V}}_{\breve{P}(\beta)}\big{(}\mbox{\boldmath$x$}-\breve{\bm{c}}(\beta)\big{)},\text{
where }{\mathscr{V}}_{{P}}(\bm{x}){\,:=\,}\bm{x}^{T}{P}^{\dagger}\bm{x},$
(22a)
$\displaystyle{\mathscr{V}}_{\breve{P}(\beta)}\big{(}\mbox{\boldmath$x$}-\breve{\bm{c}}(\beta)\big{)}-{\mathscr{V}}_{{P}}(\mbox{\boldmath$x$}-\bm{c})$
$\displaystyle\leq\breve{\varsigma}(\beta)-\varsigma=\tfrac{\alpha\beta\gamma^{2}}{1-\beta}-\alpha\beta{\delta^{2}}.$
(22b) and its minimum is achieved at $\displaystyle\beta_{\varsigma}$
$\displaystyle{\,:=\,}\arg\min_{\beta\in]0,1[}\breve{\varsigma}(\beta)=\begin{cases}1-\gamma{\left|\delta\right|}^{-1},&\text{if
}{\left|\delta\right|}>\gamma,\\\ 0,&\text{otherwise};\end{cases}$ (22c) where
$\breve{\bm{f}}$, $\breve{y}$, $\alpha$ and $\delta$ are defined in (21c),
(21g) and (21h). 0pt $\displaystyle\text{Furthermore, if
}{\left|\delta\right|}>\gamma,\text{ then
}\breve{\varsigma}(\beta_{\varsigma})<\varsigma\text{ and
}\breve{\varsigma}(\beta_{\varsigma})-\varsigma=-\alpha\beta_{\varsigma}^{2}{\delta}^{2}<0,$
(22d)
$\displaystyle\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}\breve{{\mathcal{E}}}(\beta_{\varsigma})\big{)}<\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}(\bm{c},\varsigma
P)\big{)}\text{ and
}\operatorname*{ssal}\big{(}\breve{{\mathcal{E}}}(\beta_{\varsigma})\big{)}<\operatorname*{ssal}\big{(}{\mathcal{E}}(\bm{c},\varsigma
P)\big{)}.$ (22e)
Proof. cf. Appendix B.3. ❑
The value of $\beta_{\varsigma}$ is resumed in Algorithm 3.
Algorithm 3 Optimal value of parameter $\beta$ minimizing
$\breve{\varsigma}(\beta)$
0: $\gamma$, $\delta$ {defined in (21)}
0: $\beta_{\varsigma}$
1: if ${\left|\delta\right|}>\gamma$ then
2: $\beta_{\varsigma}{\,:=\,}1-\gamma{\left|\delta\right|}^{-1}$; {cf. (22c)};
3: else
4: $\beta_{\varsigma}{\,:=\,}0$; {cf. (22c)};
5: end if
###### Remark 4.1
The representation of the output noise vector’s bounding set as an
intersection of strips, rather than as an ellipsoid, enables this optimization
problem to have an analytical solution.
As for fig. 2 (case 4), the figure 4 on the right shows the (red/small)
ellipsoid $\breve{{\mathcal{E}}}(\beta_{\varsigma})$ containing the
intersection of the (blue/big) ellipsoid ${\mathcal{E}}(\bm{c},\varsigma P)$
with the (colored) strip ${\mathcal{D}}(\bm{f},y)$, for $\beta_{\varsigma}$
given by (22c) and $n=2$. It is worth noting that with this value of
$\beta_{\varsigma}$, the center $\breve{\bm{c}}(\beta_{\varsigma})$, of
$\breve{{\mathcal{E}}}(\beta_{\varsigma})$ is the projection of $\bm{c}$ (the
center of ${\mathcal{E}}(\bm{c},\varsigma P)$), on the hyperplan representing
the nearest strip boundary, in the direction $P\bm{f}$.
Figure 4:
$\breve{{\mathcal{E}}}(\beta_{\varsigma})\supset{\mathcal{E}}(\bm{c},\varsigma
P)\cap{\mathcal{D}}(\bm{f},y)$
#### 4.3.2 Minimization of the ellipsoid’s volume
As already mentioned in Section 3, the ellipsoid can possibility have zero
axes lengths and, therefore, a zero volume and a noninvertible shape matrix.
To our knowledge, this issue was not addressed in set-membership state
estimation algorithms of the literature reviewed here.
The optimal value for the weighting parameter $\beta$ adopted in this paper
and intervening in (21), minimizing the pseudo-volume (cf. Definition 3.1) of
the ellipsoid $\breve{{\mathcal{E}}}(\beta)$, when its usual volume can be
zero, is derived in what follows.
###### Theorem 4.3.2
If teh ellipsoid $\breve{{\mathcal{E}}}(\beta)$ given by (21) of Thm 4.2, for
which ${q}{\,:=\,}\operatorname*{rank}(P)>1$, then
$\beta_{\text{v}}{\,:=\,}\arg\min_{\beta\in]0,1[}\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}\breve{{\mathcal{E}}}(\beta)\big{)}$
is the unique solution of the quadratic equation
$a_{2}\beta^{2}+a_{1}\beta+a_{0}=0$, where 0pt $\displaystyle a_{0}$
$\displaystyle{\,:=\,}{q}\alpha(\gamma^{2}-\delta^{2})-\varsigma,$ (23a)
$\displaystyle a_{1}$
$\displaystyle{\,:=\,}(2{q}+1)\alpha\delta^{2}+\varsigma-\gamma^{2}\alpha,$
(23b) $\displaystyle a_{2}$ $\displaystyle{\,:=\,}-({q}+1)\alpha\delta^{2},$
(23c) if $a_{0}<0$; otherwise $\beta_{\text{v}}=0$. 0pt
$\displaystyle\text{Furthermore, if $a_{0}<0$, then
}\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}\breve{{\mathcal{E}}}(\beta_{\text{v}})\big{)}<\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}(\bm{c},\varsigma
P)\big{)}.$ (23d)
Proof. cf. Appendix B.4. ❑
Algorithm 4 Optimal value of parameter $\beta$ minimizing the volume of
$\breve{\mathcal{E}}(\beta)$
0: $q$, $\alpha$, $\gamma$, $\delta$, $\varsigma$ {defined in (21)}
0: $\beta_{\textit{v}}$
1: $a_{0}{\,:=\,}{q}\alpha(\gamma^{2}-\delta^{2})-\varsigma$; {cf. (23a)};
2: if $a_{0}\geq 0$ then
3: $\beta_{\textit{v}}{\,:=\,}0$;
4: else
5: $a_{1}{\,:=\,}(2{q}+1)\alpha\delta^{2}+\varsigma-\gamma^{2}\alpha$; {cf.
(23b)};
6: $a_{2}{\,:=\,}-({q}+1)\alpha\delta^{2}$; {cf. (23c)};
7:
$\beta_{\textit{v}}{\,:=\,}\dfrac{-a_{1}-\sqrt{a_{1}^{2}-4a_{0}a_{2}}}{2a_{2}}$;
8: end if
#### 4.3.3 Minimization of the sum of the squared axes’ lengths
###### Theorem 4.3.3
If the ellipsoid $\breve{{\mathcal{E}}}(\beta)$ is given by (21) of Thm 4.2,
then
$\beta_{\text{s}}{\,:=\,}\arg\min_{\beta\in]0,1[}\operatorname*{tr}\big{(}\breve{\varsigma}(\beta)\breve{P}(\beta)\big{)}$
is the unique positive solution to the cubic equation $\displaystyle
b_{3}\beta^{3}+b_{2}\beta^{2}+b_{1}\beta+b_{0}{\,:=\,}0,$ (24a) if $b_{0}<0$;
$\beta_{\text{s}}=0$ otherwise; where $\displaystyle b_{0}$
$\displaystyle{\,:=\,}\nu\varsigma-\tau(\delta^{2}-\gamma^{2}),\
\tau{\,:=\,}\operatorname*{tr}(P)\text{ and
}\nu{\,:=\,}\bm{\varphi}^{T}\bm{\varphi}=\bm{f}^{T}P^{2}\bm{f},$ (24b)
$\displaystyle b_{1}$
$\displaystyle{\,:=\,}2\big{(}\tau\delta^{2}-\nu\varsigma+\alpha\nu(\delta^{2}-\gamma^{2})\big{)},$
(24c) $\displaystyle b_{2}$
$\displaystyle{\,:=\,}\nu\varsigma-\tau\delta^{2}+\alpha\nu(\gamma^{2}-4\delta^{2}),$
(24d) $\displaystyle b_{3}$ $\displaystyle{\,:=\,}2\alpha\nu\delta^{2}.\quad$
(24e) 0pt $\displaystyle\text{Furthermore, if $b_{0}<0$, then
}\operatorname*{ssal}\big{(}\breve{{\mathcal{E}}}(\beta_{\text{s}})\big{)}<\operatorname*{ssal}\big{(}{\mathcal{E}}(\bm{c},\varsigma
P)\big{)}.$ (24f)
Proof. cf. Appendix B.5. ❑
The computation of $\beta_{\text{s}}$ is elaborated in Algorithm 5.
Algorithm 5 Optimal value of the parameter $\beta$ minimizing the SSAL of
$\breve{\mathcal{E}}(\beta)$
0: $P$, $\alpha$, $\gamma$, $\delta$, $\varsigma$, $\bm{\varphi}$ {defined in
(21)}
0: $\beta_{\text{s}}$
1: $\tau{\,:=\,}\operatorname*{tr}(P)$;
$\nu{\,:=\,}\bm{\varphi}^{T}\bm{\varphi}$;
2: $b_{0}{\,:=\,}\nu\varsigma-\tau(\delta^{2}-\gamma^{2})$;
3: if $b_{0}>=0$, then
4: $\beta_{\text{s}}{\,:=\,}0$;
5: else
6:
$b_{1}{\,:=\,}2\big{(}\tau\delta^{2}-\nu\varsigma+\alpha\nu(\delta^{2}-\gamma^{2})\big{)}$;$b_{2}{\,:=\,}\nu\varsigma-\tau\delta^{2}+\alpha\nu(\gamma^{2}-4\delta^{2})$;
$b_{3}{\,:=\,}2\alpha\nu\delta^{2}$;
7: ${s}{\,:=\,}\tfrac{b_{1}}{b_{3}}-\tfrac{b_{2}^{2}}{3b_{3}^{2}}$;
${t}{\,:=\,}\tfrac{b_{0}}{b_{3}}-\tfrac{b_{1}b_{2}}{3b_{3}^{2}}+\tfrac{2b_{2}^{3}}{27b_{3}^{3}}$;
8:
${u}{\,:=\,}\left(\tfrac{{s}}{3}\right)^{3}+\big{(}\tfrac{{t}}{2}\big{)}^{2}$;
${v}{\,:=\,}\sqrt[3]{-\tfrac{{t}}{2}+\sqrt{{u}}}$;
${w}{\,:=\,}\sqrt[3]{-\tfrac{{t}}{2}-\sqrt{{u}}}$;
9: $\omega{\,:=\,}-\tfrac{1}{2}+\tfrac{1}{2}\sqrt{3}i$;
$\omega_{s}{\,:=\,}-\tfrac{1}{2}-\tfrac{1}{2}\sqrt{3}i$; { $\omega$ is the
cubic root of 1 and $\omega_{s}{\,:=\,}\omega^{2}$}
10: $\bar{\beta}_{1}{\,:=\,}{v}+{w}$;
$\bar{\beta}_{2}{\,:=\,}\omega{v}+\omega_{s}{w}$;
$\bar{\beta}_{3}{\,:=\,}\omega_{s}{v}+\omega{w}$;
11: $\beta_{1}{\,:=\,}\bar{\beta}_{1}-\tfrac{b_{2}}{3b_{3}}$;
$\beta_{2}{\,:=\,}\bar{\beta}_{2}-\tfrac{b_{2}}{3b_{3}}$;
$\beta_{3}{\,:=\,}\bar{\beta}_{3}-\tfrac{b_{2}}{3b_{3}}$;
12: $\beta_{\text{s}}{\,:=\,}\beta_{i}$, $i\in\\{1,2,3\\}$, _s.t._
$\beta_{i}\in{\mathop{\rm I\mkern-3.5muR}}^{*}_{+}$.
13: end if
### 4.4 Intersection of an ellipsoid with a hyperplane
In this paragraph, the equality-type constraint (3c) is examined. This
constraint on the state vector can be also viewed as a noiseless measurement,
_a.k.a._ pseudo-measurement. Let us examine the intersection of an ellipsoid
with a hyperplane. This intersection leads to a degenerate ellipsoid of lesser
dimension, whose shape matrix loses one rank with each intersecting (not
parallel) hyperplane.
The theorem below gives the expression of thusly obtained ellipsoid.
###### Theorem 4.3 (ellips./hyperplane inters.)
Let $\bm{c}\in~{}{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$, $P\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n\times n}}$ SPSD, $\varsigma\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}$, $\bm{f}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\text{--}\\{{\mathbf{0}}_{n}\\}$ and $y\in{\mathop{\rm
I\mkern-3.5muR}}$,
if $y>\bar{\rho}\vee y<-\underaccent{\bar}{\rho}$, (case 1)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma P)\cap{\mathcal{H}}(\bm{f},y)$
$\displaystyle=\emptyset;$ (25b) else if
$y=\bar{\rho}=-\underaccent{\bar}{\rho}$, (case 2)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma P)\cap{\mathcal{H}}(\bm{f},y)$
$\displaystyle={\mathcal{E}}(\bm{c},\varsigma P);$ (25c) else if
$y=-\underaccent{\bar}{\rho}$, (case 3a)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma P)\cap{\mathcal{H}}(\bm{f},y)$
$\displaystyle=\\{\bm{c}-\varsigma^{\frac{1}{2}}(\bm{f}^{T}P\bm{f})^{-\frac{1}{2}}P\bm{f}\\};$
(25d) else if $y=\bar{\rho}$, (case 3b)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma P)\cap{\mathcal{H}}(\bm{f},y)$
$\displaystyle=\\{\bm{c}+\varsigma^{\frac{1}{2}}(\bm{f}^{T}P\bm{f})^{-\frac{1}{2}}P\bm{f}\\};$
(25e) otherwise $(\text{if }-\underaccent{\bar}{\rho}<y<\bar{\rho})$ (case 4)
$\displaystyle{\mathcal{E}}(\bm{c},\varsigma P)\cap{\mathcal{H}}(\bm{f},y)$
$\displaystyle={\mathcal{E}}(\check{\bm{c}},\check{\varsigma}\check{P}),$
(25f) where $\displaystyle\check{\bm{c}}$
$\displaystyle{\,:=\,}\bm{c}+\alpha\delta{P\bm{f}},$ (25g)
$\displaystyle\check{P}$
$\displaystyle{\,:=\,}P-{\alpha}{P\bm{f}\bm{f}^{T}P},$ (25h)
$\displaystyle\check{\varsigma}$
$\displaystyle{\,:=\,}\varsigma-{\alpha}{\delta^{2}},$ (25i)
$\displaystyle\delta$ $\displaystyle{\,:=\,}y-\bm{f}^{T}\bm{c}$ (25j) and
where $\alpha$, $\underaccent{\bar}{\rho}$ and $\bar{\rho}$ are defined in
(21g), (19g) and (19h) _resp_.
Proof. cf. Appendix B.6. ❑
The figure 5 is self-explanatory. It is plain to see that the intersection of
a three-dimensional ellipsoid with an intersecting non parallel hyperplane is
a degenerate three-dimensional, corresponding to a two-dimensional, ellipsoid
and the good news is that there is no need to circumscribe it by an other one,
as it is done in case of intersection with strip or halfspace. Please note
that the case 2 happens only when the ellipsoid
${\mathcal{E}}(\bm{c},\varsigma P)$ is already degenerate and contained in the
hyperplane ${\mathcal{H}}(\bm{f},y)$, meaning that $y=\bm{c}^{T}\bm{f}$ and
that $\bm{f}\in\mathpzc{Ker}(P)$.
(a) (1a): $y-\bm{c}^{T}\bm{f}>\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(b) (1b): $y-\bm{c}^{T}\bm{f}<-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(c) (2):
${y-\bm{c}^{T}\bm{f}}\\!=\\!{\bm{f}^{T}P\bm{f}}\\!=\\!-{\bm{f}^{T}P\bm{f}}=0$
(d) (3a): $y-\bm{c}^{T}\bm{f}=\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(e) (3b): $y-\bm{c}^{T}\bm{f}=-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
(f) (4): ${\left|y-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
Figure 5: Illustration of the four cases of Theorem 4.3 for $n=3$
### 4.5 The output update algorithm
Hereafter, the measurement update part of the state estimation algorithm is
summarized. All the variables are depending on the time step $k$ even if the
subscript $k$ was skipped for some of them.
###### Theorem 4.4 (Correction stage)
Let
0pt
$\displaystyle\bar{y}_{k_{i}}{\,:=\,}\min(\bar{y}_{k_{i}},\bar{{\rho}}_{k_{i}})$
$\displaystyle\text{ and
}\underaccent{\bar}{y}_{k_{i}}{\,:=\,}\max(\underaccent{\bar}{y}_{k_{i}},-\underaccent{\bar}{{\rho}}_{k_{i}}),\
i\in\\{1,\ldots,{p}\\}.$ (26a) If $\mbox{\boldmath$x$}_{k}$ satisfies (1)
meeting (3), then $\displaystyle\mbox{\boldmath$x$}_{k}\in{\mathcal{S}}_{k}$
$\displaystyle\subseteq{\mathcal{E}}_{k}{\,:=\,}{\mathcal{E}}(\hat{\mbox{\boldmath$x$}}_{k},\varsigma_{k}P_{k}),$
(26b) 0pt ${\mathcal{S}}_{k}$ is defined in (18) and, $\forall
k\in{\mathop{\rm I\mkern-3.5muN}}^{*}$, $\displaystyle\varsigma_{k}$
$\displaystyle{\,:=\,}\varsigma_{k_{p}},P_{k}{\,:=\,}P_{k_{p}}\text{ and
}\hat{\mbox{\boldmath$x$}}_{k}{\,:=\,}\mbox{\boldmath$\hat{x}$}_{k_{p}};\quad\text{and
for $i\in\\{1,\ldots,{p}\\}$:}$ (26d) $\displaystyle P_{k_{i}}$
$\displaystyle{\,:=\,}P_{k_{i-1}}-{\alpha_{k_{i}}\beta_{k_{i}}}{{\bm{\varphi}}_{k_{i}}{\bm{\varphi}}_{k_{i}}^{T}},$
(26e) $\displaystyle\hat{\mbox{\boldmath$x$}}_{k_{i}}$
$\displaystyle{\,:=\,}\hat{\mbox{\boldmath$x$}}_{k_{i-1}}+{\alpha_{k_{i}}\beta_{k_{i}}}{\delta_{k_{i}}}{\bm{\varphi}}_{k_{i}},$
(26f) $\displaystyle\varsigma_{k_{i}}$
$\displaystyle{\,:=\,}\begin{cases}\varsigma_{k_{i-1}}-{\alpha_{k_{i}}\beta_{k_{i}}^{2}{\delta_{k_{i}}^{2}}},&\text{if
}\big{(}\underaccent{\bar}{y}_{k_{i}}=\bar{y}_{k_{i}}\wedge-\underaccent{\bar}{{\rho}}_{k_{i}}\neq\bar{{\rho}}_{k_{i}}\big{)}\vee\beta^{*}=\beta_{\varsigma},\\\
\varsigma_{k_{i-1}}+{\alpha_{k_{i}}\beta_{k_{i}}\bigg{(}\tfrac{\gamma_{k_{i}}^{2}}{1-\beta_{k_{i}}}-{\delta_{k_{i}}^{2}}\bigg{)}},&\text{otherwise;}\end{cases}$
(26g) $\displaystyle P_{k_{0}}$
$\displaystyle{\,:=\,}P_{k/k-1},\hat{\mbox{\boldmath$x$}}_{k_{0}}{\,:=\,}\hat{\mbox{\boldmath$x$}}_{k/k-1}\text{
and }\varsigma_{k_{0}}{\,:=\,}\varsigma_{k-1},$ (26h) where
$\hat{\mbox{\boldmath$x$}}_{k/k-1}\text{ and }P_{k/k-1}$ are computed
according to Thms 3.6 and 3.7 or 3.8, with
${q}_{k_{0}}{\,:=\,}q_{k/k-1}{\,:=\,}{q}_{{k/k-1}_{m}}$ computed at the
prediction stage, in (11d) (line 13 of Algorithm 1) and
${q}_{k}{\,:=\,}\operatorname*{rank}(P_{k}){\,:=\,}{q}_{k_{m}}$.
$\displaystyle\alpha_{k_{i}}$
$\displaystyle{\,:=\,}\begin{cases}\theta_{k_{i}}^{-1},&\text{if }{p_{k}}\neq
0\text{ and }\lambda_{k_{i}}\neq 0,\\\ 0,&\text{otherwise};\end{cases}$ (26i)
$\displaystyle\beta_{k_{i}}$ $\displaystyle{\,:=\,}\begin{cases}1,&\text{if
}\underaccent{\bar}{y}_{k_{i}}=\bar{y}_{k_{i}}\wedge-\underaccent{\bar}{{\rho}}_{k_{i}}\neq\bar{{\rho}}_{k_{i}},(i\in\mathscr{H}_{k}),\\\
\beta^{*},&\text{else if
}-\underaccent{\bar}{{\rho}}_{k_{i}}<\underaccent{\bar}{y}_{k_{i}}\vee\bar{y}_{k_{i}}<\bar{{\rho}}_{k_{i}},(i\in\mathscr{D}_{k}\cup\mathscr{G}_{k})\\\
0,&\text{otherwise;}\end{cases}$ (26j) where, either
$\beta^{*}=\beta_{\varsigma}$, cf. (22c) - Algo. 3,
$\beta^{*}=\beta_{\text{v}}$, given by Thm 4.3.2 \- Algo. 4 or
$\beta^{*}=\beta_{\text{s}}$, cf. Thm 4.3.3 \- Algo. 5; where
$\alpha{\,:=\,}\alpha_{k_{i}}$, $\gamma{\,:=\,}\gamma_{k_{i}}$,
$\delta{\,:=\,}{\delta_{k_{i}}}$, $\varsigma{\,:=\,}\varsigma_{k_{i-1}}$,
$\bm{\varphi}{\,:=\,}{\bm{\varphi}}_{k_{i}}$, $P{\,:=\,}P_{k_{i-1}}$ and
${q}{\,:=\,}{q}_{k_{i-1}}$and $\displaystyle{\delta_{k_{i}}}$
$\displaystyle{\,:=\,}\gamma_{k_{i}}{y_{k_{i}}}-\tfrac{1}{2}(\bar{{\rho}}_{k_{i}}-\underaccent{\bar}{{\rho}}_{k_{i}}),\quad{y_{k_{i}}}{\,:=\,}\tfrac{1}{2\gamma_{k_{i}}}({\bar{y}_{k_{i}}+\underaccent{\bar}{y}_{k_{i}}}),$
(26k) $\displaystyle\gamma_{k_{i}}$
$\displaystyle{\,:=\,}\tfrac{1}{2}(\bar{y}_{k_{i}}-\underaccent{\bar}{y}_{k_{i}}),$
(26l) $\displaystyle\theta_{k_{i}}$
$\displaystyle{\,:=\,}\bm{f}_{k_{i}}^{T}{\bm{\varphi}}_{k_{i}},$ (26m)
$\displaystyle{\bm{\varphi}}_{k_{i}}$
$\displaystyle{\,:=\,}P_{k_{i-1}}\bm{f}_{k_{i}},$ (26n)
$\displaystyle\bar{{\rho}}_{k_{i}}$
$\displaystyle{\,:=\,}\lambda_{k_{i}}+\bm{f}_{k_{i}}^{T}\hat{\mbox{\boldmath$x$}}_{k_{i-1}}\text{
and
}\underaccent{\bar}{{\rho}}_{k_{i}}{\,:=\,}2\lambda_{k_{i}}-\bar{{\rho}}_{k_{i}},$
(26o) $\displaystyle\lambda_{k_{i}}$
$\displaystyle{\,:=\,}(\varsigma_{k_{i-1}}\theta_{k_{i}})^{\frac{1}{2}};$
(26p) and, if ${\mathcal{E}}_{k/k-1}$ is computed according to Thms 3.6 and
3.7 by (10)-(11), then $\displaystyle{q}_{k_{i}}$
$\displaystyle\big{(}{\,:=\,}\operatorname*{rank}(P_{k_{i}})\big{)}{\,:=\,}\begin{cases}{q}_{k_{i-1}}-1,&\text{if
}\underaccent{\bar}{y}_{k_{i}}=\bar{y}_{k_{i}}\wedge-\underaccent{\bar}{{\rho}}_{k_{i}}\neq\bar{{\rho}}_{k_{i}}\wedge\alpha_{k_{i}}\neq
0,\\\ {q}_{k_{i-1}},&\text{otherwise.}\end{cases}$ (26q)
Proof. cf. Appendix B.7. ❑
The correction stage described in Thm 4.4 is resumed in Algorithm 6 and
${\mathcal{E}}_{k}$ is computed where
$\hat{\mbox{\boldmath$x$}}_{k/k-1}\rightarrow~{}\\!\bar{}\mbox{\boldmath$x$}$,
$P_{k/k-1}\rightarrow~{}\\!\bar{P}$,
$\varsigma_{k-1}\rightarrow~{}\\!\bar{\varsigma}$,
$q_{k/k-1}\leftarrow~{}\\!\bar{q}$ (from the prediction stage Algorithm 1 or
2), ${F}_{k}\rightarrow~{}\\!F=[\bm{f}_{i}]$,
$\bar{\bm{y}}_{k}\rightarrow~{}\\!\bar{\mbox{\boldmath$y$}}=~{}\\![\bar{y}_{i}]$,
${i\in\\{1,\dots,p\\}}$,
$\underaccent{\bar}{\y}_{k}\rightarrow~{}\\!\underaccent{\bar}{\yb}$ (as in
(3)) and $\mbox{\boldmath$x$}\rightarrow~{}\\!\hat{\mbox{\boldmath$x$}}_{k}$,
$P\rightarrow~{}\\!P_{k}$, $\varsigma\rightarrow~{}\\!\varsigma_{k}$. Note
that the updating of the rank of the shape matrix $P_{k_{i}}$, by (26q), is
done only when the volume minimization criterion is chosen at the prediction
step, by setting Predic_$\mu$ to “True”. Let Correc_$\beta$ $\in\\{0,1,2\\}$
set to 0, 1 or 2 when choosing the minimization of
$\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}({\mathcal{E}}_{k_{i}})$,
the SSAL of ${\mathcal{E}}_{k_{i}}$ or $\varsigma_{k_{i}}$, _resp_., during
the correction stage.
Algorithm 6 Computation of the corrected ellipsoid
0: $\bar{}\mbox{\boldmath$x$}$, $\bar{P}$, $\bar{\varsigma}$ ${F}$,
$\bar{\mbox{\boldmath$y$}}$,
$\underaccent{\bar}{\yb}=[\underaccent{\bar}{y}_{i}]_{i\in\\{1,\dots,p\\}}$,
Predic_$\mu$, Correc_$\beta$ , $\bar{q}$ (if Predic_$\mu$)
0: $x$, $P$, $\varsigma$, $q$ (if Predic_$\mu$)
1: $p\leftarrow$ number of columns of $F$;
2: $\mbox{\boldmath$x$}\leftarrow\bar{}\mbox{\boldmath$x$}$;
$P\leftarrow\bar{P}$; $\varsigma\leftarrow\bar{\varsigma}$;
$q\leftarrow\bar{q}$ (if Predic_$\mu$);
3: if $p\neq 0$ then
4: for $i=1,\ldots p$ do
5: $\bm{\varphi}{\,:=\,}P\bm{f}_{i}$;
$\theta{\,:=\,}\bm{f}_{i}^{T}\bm{\varphi}$; {cf. (26n)};
6: if $\theta\neq 0$ then
7: $\alpha{\,:=\,}\theta^{-1}$;
$\lambda{\,:=\,}(\varsigma\theta)^{\frac{1}{2}}$; {cf. (26i), (26n)};
8: $\bar{\rho}{\,:=\,}\lambda+\bm{f}_{i}^{T}\mbox{\boldmath$x$}$;
$\underaccent{\bar}{\rho}{\,:=\,}2\lambda-\bar{\rho}$; {cf. (26o)};
9: $\bar{y}{\,:=\,}\min(\bar{y}_{i},\bar{\rho})$;
$\underaccent{\bar}{y}{\,:=\,}\max(\underaccent{\bar}{y}_{i},-\underaccent{\bar}{\rho})$;
{cf. (26a)};
10:
$\delta{\,:=\,}\tfrac{1}{2}({\bar{y}+\underaccent{\bar}{y}}-\bar{\rho}+\underaccent{\bar}{\rho})$;
{cf. (26k)};
11: $\gamma{\,:=\,}\frac{1}{2}(\bar{y}-\underaccent{\bar}{y})$; {cf. (26l)};
12: if $\underaccent{\bar}{y}=\bar{y}\text{ and
}-\underaccent{\bar}{\rho}\neq\bar{\rho}$ then
13: $\beta=1$;
14: if Predic_$\mu$ $=$ True then
15: $q\leftarrow q-1$
16: end if
17: else if $-\underaccent{\bar}{\rho}<\underaccent{\bar}{y}$ or
$\bar{y}<\bar{\rho}$ then
18: if Correc_$\beta$ $=0$ then
19: Algorithm 4 (Input: $q$, $\alpha$, $\gamma$, $\delta$, $\varsigma$;
Output: $\beta_{\text{v}}$)
20: $\beta{\,:=\,}\beta_{\text{v}}$;
21: else if Correc_$\beta$ $=1$ then
22: Algorithm 5 (Input: $P$, $\alpha$, $\gamma$, $\delta$, $\varsigma$,
$\bm{\varphi}$; Output: $\beta_{\text{s}}$)
23: $\beta{\,:=\,}\beta_{\text{s}}$;
24: else
25: Algorithm 3 (Input: $\delta$, $\gamma$; Output: $\beta_{\varsigma}$ )
26: $\beta{\,:=\,}\beta_{\varsigma}$;
27: end if
28: else
29: $\beta=0$;
30: end if
31: $P\leftarrow P-{\alpha\beta}{\bm{\varphi}\bm{\varphi}^{T}}$;
$\mbox{\boldmath$x$}\leftarrow\mbox{\boldmath$x$}+{\alpha\beta}\delta\bm{\varphi}$;
$\varsigma\leftarrow\varsigma-{\alpha\beta^{2}\delta^{2}}$; {cf. (26e)–(26g)};
32: end if
33: end for
34: end if
###### Remark 4.2
This algorithm is of low computational complexity : $O(n^{2})$. Indeed, all
the operations are simple sums and products: they were optimized in this
regard and are thence suitable for systems with high dimensional state vector
(big $n$) and with many measurements (big ${p_{k}}$). The intermediate
variables $\alpha_{k_{i}}$, $\theta_{k_{i}}$, $\lambda_{k_{i}}$,
${\bm{\varphi}}_{k_{i}}$ were added on to perform redundant vector and matrix
operations only once. Thereby noticing that
$\bm{f}_{k_{i}}^{T}\hat{\mbox{\boldmath$x$}}_{k_{i-1}}=\tfrac{1}{2}(\bar{{\rho}}_{k_{i}}-\underaccent{\bar}{{\rho}}_{k_{i}})$
allows to determine
${\delta_{k_{i}}}{\,:=\,}\tfrac{1}{2}({\bar{y}_{k_{i}}+\underaccent{\bar}{y}_{k_{i}}})-\bm{f}_{k_{i}}^{T}\hat{\mbox{\boldmath$x$}}_{k_{i-1}}$
and
$\underaccent{\bar}{{\rho}}_{k_{i}}{\,:=\,}\lambda_{k_{i}}-\bm{f}_{k_{i}}^{T}\hat{\mbox{\boldmath$x$}}_{k_{i-1}}$
using addition of scalars, in (26k) and (26o) _resp_., rather than
multiplication of possibly high dimensional vectors.
###### Remark 4.3
In the case where $\mathscr{H}_{k}\neq\emptyset$, the matrix $P_{k_{i}}$ loses
rank with each intersecting hyperplane ${\mathcal{H}}_{k_{i}}$,
$i\in\mathscr{H}_{k}$, thusly entailing the progressive flattening of the
ellipsoid ${\mathcal{E}}_{k_{i}}$. Depending on the rank of the matrix $R_{k}$
(on which no assumption is made), the rank of $P_{k+1/k}$ can be recovered at
the time-update phase. The value of $\operatorname*{rank}(P_{k_{i}})$ is
needed at each $i=1\ldots,m,\ldots,m+{p_{k}}$ for each time step $k$, whenever
the volume of the ellipsoid ${\mathcal{E}}_{k}$ is minimized. So keeping track
of this parameter through simple relations, as (11d) during the time update
and (26q) during the observation update, spares the need to calculate it each
time, provided that $\operatorname*{rank}(A_{0}P_{0}A_{0}^{T})$ is given.
###### Remark 4.4
Setting either $\alpha_{k_{i}}=0$ or $\beta_{k_{i}}=0$ results in freezing
${\mathcal{E}}_{k_{i-1}}$, meaning that the corresponding measurements
$\bm{f}_{k_{i}},\underaccent{\bar}{y}_{k_{i}},\bar{y}_{k_{i}}$ do not bring
any useful information.
###### Remark 4.5
The cases 1 of Thms 4.1, 4.2 and 4.3 are not explicitly treated in this
theorem assuming that they can not occur since the intervening measurements
should be consistent with the system model; yet the case where the measurement
$\\{\bm{f}_{k_{i}},\underaccent{\bar}{y}_{k_{i}},\bar{y}_{k_{i}}\\}$ is
aberrant is implicitly considered, setting again either $\alpha_{k_{i}}=0$ or
$\beta_{k_{i}}=0$, preventing so the updating of the ellipsoid
${\mathcal{E}}_{k_{i-1}}$.
## 5 Algorithm properties and stability analysis
### 5.1 The overall state estimation algorithm
The time prediction stage given by Thm 3.6 with either Thm 3.7 or Thm 3.8 on
one hand, and on the other, the measurement correction phase, given by Thm 4.4
are concatenated to form the hole state estimation algorithm presented in
Algorithm 7, where $N$ is the number of samples.
Algorithm 7 Computation of the ellipsoid
${\mathcal{E}}(\hat{\mbox{\boldmath$x$}}_{k},\varsigma_{k}P_{k})$
0: $\hat{x}_{0}$, $\varsigma_{0}$, $P_{0}$, $N$, Predic_$\mu$, Correc_$\beta$
0: $\hat{\mbox{\boldmath$x$}}_{k}$, $\varsigma_{k}$, $P_{k}$,
$k\in\\{1,\ldots,N\\}$
1: $n\leftarrow$ size of $\hat{x}_{0}$
2: for $k=0,1,\ldots,N-1$ do
{// Prediction //}
3: if Predic_$\mu$ then
4: Algorithm 1 (Input: $\hat{\mbox{\boldmath$x$}}_{k}$, $P_{k}$,
$\varsigma_{k}$, $A_{k}$, $B_{k}$, $R_{k}$, $\mbox{\boldmath$\tau$}_{k}$,
Predic_$\mu$, $q_{k}$; Output $\hat{\bm{x}}_{k+1/k}$, $P_{k+1/k}$,
$\varsigma_{k+1/k}$, Predic_$\mu$, $q_{k+1/k}$)
5: else
6: Algorithm 2 (Input: $\hat{\mbox{\boldmath$x$}}_{k}$, $P_{k}$,
$\varsigma_{k}$, $A_{k}$, $B_{k}$, $R_{k}$,
$\mbox{\boldmath$\tau$}_{k}$;Output: $\hat{\bm{x}}_{k+1/k}$, $P_{k+1/k}$,
$\varsigma_{k+1/k}$, Predic_$\mu$)
7: end if
{// Correction //}
8: $k\leftarrow k+1$
9: Algorithm 6 (Input: $\hat{\mbox{\boldmath$x$}}_{k/k-1}$, $P_{k/k-1}$,
$\varsigma_{k/k-1}$, ${F}_{k}$, $\bar{\bm{y}}_{k}$,
$\underaccent{\bar}{\y}_{k}$, Predic_$\mu$, Correc_$\beta$ , $q_{k/k-1}$;
Output: $\hat{\mbox{\boldmath$x$}}_{k}$, $P_{k}$, $\varsigma_{k}$, $q_{k}$)
10: $P_{k}\leftarrow\frac{\varsigma}{\varsigma_{0}}P_{k}$;
$\varsigma_{k}\leftarrow\varsigma_{0}$;
11: end for
###### Remark 5.1
For more numerical stability and in order to avoid the explosion of the matrix
$P_{k}$, caused by the set summations at the prediction step, the assignments
(26h) corresponding to the line 10 of the Algorithm 7 can be replaced by
$P_{k}\leftarrow\frac{\varsigma_{k_{p}}}{\varsigma_{0}}P_{k_{p}}$,
$\bar{\varsigma}_{k}\leftarrow\frac{\varsigma_{0}\varsigma_{k_{p}}}{\bar{\varsigma}_{k-1}}$
and $\varsigma_{k}\leftarrow\varsigma_{0}$. Then $P_{k}$ would, by itself,
represent the shape of the ellipsoid ${\mathcal{E}}_{k}$ up to a constant
factor $\varsigma_{0}^{-1}$ and the new variable $\bar{\varsigma}_{k}$ is
introduced to keep track of the decreasing parameter $\varsigma_{k}$, with
$\bar{\varsigma}_{0}=\varsigma_{0}$.
The proposed algorithm is designed in such a way as to fulfill the
requirements 1. \- 3., expressed in the $\mathsection$2 and this is what will
be shown in this section.
### 5.2 Algorithm properties
In this paragraph, some algorithm’s properties are shown, while the stability
will be examined in the next.
###### Theorem 5.1
Consider the system (1) subject to (3) and its state estimation algorithm
given by Thms 3.6 and 3.7 or 3.8 on one hand and Thm 4.4, on the other.
1. 1.
If
$\mbox{\boldmath$x_{0}$}\in{\mathcal{E}}(\mbox{\boldmath$\hat{x}_{0}$},\varsigma_{0}P_{0})$,
then $\forall\bm{\beta}\in[0,1[^{p_{k}}$, $\forall\bm{\mu}\in]0,+\infty[^{m}$,
$\mbox{\boldmath$x$}_{k}\in{\mathcal{E}}(\hat{\mbox{\boldmath$x$}}_{k},\varsigma_{k}P_{k})$,
$\forall k\in{\mathop{\rm I\mkern-3.5muN}}^{*}$;
2. 2.
The vector $\hat{\mbox{\boldmath$x$}}_{k}$ is _acceptable_ , i.e., it
satisfies (3a)–(3e):
$\displaystyle\forall k\in{\mathcal{K}},\
\hat{\mbox{\boldmath$x$}}_{k}\in\bigcap_{i\in\mathscr{G}_{k}}{\mathcal{G}}_{k_{i}}\cap\bigcap_{i\in\mathscr{D}_{k}}{\mathcal{D}}_{k_{i}}\cap\bigcap_{i\in\mathscr{H}_{k}}{\mathcal{H}}_{k_{i}},$
(27a)
where ${\mathcal{K}}{\,:=\,}\\{k\in{\mathop{\rm I\mkern-3.5muN}}|{p_{k}}\neq
0\\}$.
3. 3.
$\varsigma_{k}$, defined in (26g), satisfies,
$\displaystyle\varsigma_{k}$
$\displaystyle=\max_{\bm{x}_{k}\in{\mathcal{S}}_{k}}\mathscr{V}_{{P}_{k}}(\mbox{\boldmath$x$}_{k}-\hat{\mbox{\boldmath$x$}}_{k}),\
\forall\beta^{*}\in]0,1[,$ (27b)
where ${\mathcal{S}}_{k}$ and $\mathscr{V}_{{P}}$ are given by (18) and (22a)
_resp_.
Moreover, if in (26j), $\beta^{*}=\beta_{\varsigma}$, given by (22c) - Algo.
3, then the sequence $\left\\{\varsigma_{k}\right\\}_{k\in{\mathop{\rm
I\mkern-3.5muN}}^{*}}$ is decreasing and convergent on ${\mathop{\rm
I\mkern-3.5muR}}_{+}$.
Proof. The proof of this lemma is detailed in the Appendix C.1. ❑
### 5.3 Stability analysis
The stability requirement 3. exploits the Input-to-State stability concept:
roughly speaking, for an ISS system, inputs that are bounded, “eventually
small”, or convergent, should lead to the state vector with the respective
property; and the ${\mathbf{0}}$-input system should be globally stable. We
shall now recall more formal definitions and results about the ISS concept.
Before doing so, let us recall some comparison functions, widely used in
stability analysis. A continuous function $\psi_{1}:{\mathop{\rm
I\mkern-3.5muR}}_{+}\rightarrow{\mathop{\rm I\mkern-3.5muR}}_{+}$ is called
positive definite if it satisfies $\psi_{1}(0)=0$ and $\psi_{1}(t)>0$,
$\forall t>0$. A positive definite function is of class $\mathscr{K}$ if it is
strictly increasing and of class $\mathscr{K}_{\infty}$ if it is of class
$\mathscr{K}$ and unbounded. A continuous function $\psi_{2}:{\mathop{\rm
I\mkern-3.5muR}}_{+}\rightarrow{\mathop{\rm I\mkern-3.5muR}}_{+}$ is of class
$\mathscr{L}$ if $\psi_{2}(t)$ is strictly decreasing to 0 as
$t\rightarrow\infty$ and a continuous function $\psi_{3}:{\mathop{\rm
I\mkern-3.5muR}}_{+}\times{\mathop{\rm
I\mkern-3.5muR}}_{+}\rightarrow{\mathop{\rm I\mkern-3.5muR}}_{+}$ is of class
$\mathscr{KL}$ if it is of class $\mathscr{K}$ in the first argument and of
class $\mathscr{L}$ in the second argument.
###### Definition 5.1 (based on [32])
The system 0pt
$\displaystyle\mbox{\boldmath$z$}_{k+1}$
$\displaystyle=f_{k}\big{(}\mbox{\boldmath$z$}_{k},\bm{u}_{k}\big{)},$ (28a)
$\displaystyle\text{where }\qquad f({\mathbf{0}},{\mathbf{0}})$
$\displaystyle={\mathbf{0}},$ (28b)
is globally input-to-state stable (ISS), if there exists a
$\mathscr{KL}$-function $\phi$ and a $\mathscr{K}$-function $\psi$ such that,
for each bounded input sequence
$\bm{u}_{[0,k]}{\,:=\,}\left\\{\bm{u}_{0},\ldots,\bm{u}_{k}\right\\}$ and each
$\mbox{\boldmath$z$}_{0}\in~{}{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$,
$\displaystyle\left\|\mbox{\boldmath$z$}(k,\mbox{\boldmath$z$}_{0},\bm{u}_{[0,k-1]})\right\|\leq\phi(\left\|\mbox{\boldmath$z$}_{0}\right\|,k)+\psi({\\!\sup_{i\in\\{1,\ldots,k-1\\}}\\!\left\|\bm{u}_{i}\right\|}),$
where $\mbox{\boldmath$z$}(k,\mbox{\boldmath$z$}_{0},\bm{u}_{[0,k-1]})$ is the
trajectory of the system (28), for the initial state
$\mbox{\boldmath$z$}_{0}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$ and the
input sequence $\bm{u}_{[0,k-1]}$.
###### Definition 5.2 ([32])
_A continuous function $\mathscr{V}:{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\rightarrow{\mathop{\rm I\mkern-3.5muR}}_{+}$ is an ISS-
Lyapunov function for the system (28), if both conditions 1 et 2 are met:_
1. 1.
there exists $\mathscr{K}_{\infty}$-functions $\psi_{1}$ and $\psi_{2}$ such
that for all $\mbox{\boldmath$z$}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n}}$,
$\displaystyle\psi_{1}(\left\|\mbox{\boldmath$z$}\right\|)\leq\mathscr{V}(\mbox{\boldmath$z$})\leq\psi_{2}(\left\|\mbox{\boldmath$z$}\right\|);$
(29)
2. 2.
there exists a $\mathscr{K}_{\infty}$\- function $\psi_{3}$ and a
$\mathscr{K}$-function $\chi$ such that for all $k\in{\mathop{\rm
I\mkern-3.5muN}}^{*}$, $\mbox{\boldmath$z$}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$ and all $\bm{u}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{m}}$,
$\displaystyle\mathscr{V}\big{(}f_{k}(\mbox{\boldmath$z$},\bm{u})\big{)}-\mathscr{V}(\mbox{\boldmath$z$})\leq-\psi_{3}(\left\|\mbox{\boldmath$z$}\right\|)+\chi(\left\|\bm{u}\right\|).$
(30)
###### Lemma 5.2 ([32])
The system (28) is ISS, if it admits a continuous ISS- Lyapunov function.
To prove the ISS stability, we will be using the candidate Lyapunov function
defined in (27b). For this purpose, the shape matrix $P_{k}$ should be bounded
above and below. Before studying the boundedness of $P_{k}$, we need to recall
the uniform controllability and observability notions:
###### Definition 5.3 (uniform observability and controllability)
Consider time-varying matrices $\bar{F}_{k}\in{\mathop{\rm
I\mkern-3.5muR}}^{n\times p}$, ${R}_{k}$ and invertible
${A}_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$. The pair
$\\{{A}_{k},\bar{F}_{k}^{T}\\}$ is _uniformly observable_ , if there exist
positive constants ${\varrho}_{1}$ and ${\varrho}_{2}$ and a positive integer
$h$, such that, for all $k\geq h$,
$\displaystyle{\varrho}_{1}I_{n}\leq{{\mathcal{O}}}_{k,k-h}\leq{\varrho}_{2}I_{n},$
(31)
where ${{\mathcal{O}}}_{k+l,k}$ is an observability gramian of length $l$: 0pt
$\displaystyle{{\mathcal{O}}}_{k+l,k}$
$\displaystyle{\,:=\,}\sum_{i=k}^{k+l}{\Phi}_{i,k}^{T}\bar{F}_{i}\bar{F}_{i}^{T}{\Phi}_{i,k}.$
(32) $\displaystyle\text{where \quad}{\Phi}_{k+l,k}$
$\displaystyle{\,:=\,}{A}_{k+l-1}\ldots{A}_{k},\text{ with
}{\Phi}_{k,k+l}={\Phi}_{k+l,k}^{-1}.$ (33)
The pair $\\{{A}_{k},{R}_{k}\\}$ is _uniformly controllable_ , if there exist
positive constants ${\varrho}_{3}$ and ${\varrho}_{4}$ and a positive integer
${h}$, such that, for all $k\geq h$,
$\displaystyle{\varrho}_{3}I_{n}\leq{{\mathcal{C}}}_{k,k-{h}}\leq{\varrho}_{4}I_{n},$
(34)
where ${{\mathcal{C}}}_{k+l,k}$ is a controllability gramian of length $h$:
$\displaystyle{{\mathcal{C}}}_{k+l,k}$
$\displaystyle{\,:=\,}\sum_{i=k}^{k+l-1}{\Phi}_{k,i+1}{R}_{i}{R}_{i}^{T}{\Phi}_{k,i+1}^{T}.$
(35)
It is needless to say that it is difficult to ensure the full rank for the
matrix sum ${{\mathcal{O}}}_{k,k-{h}}$ on a time window of constant length $h$
when dealing with sporadic measurements. The system (1)-(3) can therefore not
be uniformly observable. For this purpose, let us introduce the new
observability criterion for systems with sporadic measurements, by allowing
the observability gramian, used in _uniform observability_ to have a variable
length $\kappa_{k}(h)$ instead of the fixed one $h$:
###### Definition 5.4 (sporadic observability)
The pair $\\{{A}_{k},\bar{F}_{k}^{T}\\}$ is said _sporadically observable_ ,
if there exist positive constants $\bar{\varrho}_{1}$ and $\bar{\varrho}_{2}$
and a positive integer $h$, such that, for all $k\geq\kappa_{k}(h)$,
$\displaystyle\bar{\varrho}_{1}I_{n}\leq{{\mathcal{O}}}_{k,k-\kappa_{k}(h)}\leq\bar{\varrho}_{2}I_{n},$
where ${{\mathcal{O}}}_{k+l,k}$ is the observability gramian given in (32) and
${\Phi}_{k+l,k}$, defined in (33), is the state transition matrix associated
to ${A}_{k}\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$, which is assumed
to be _invertible_ ; $\kappa_{k}(h)\in{\mathop{\rm I\mkern-3.5muN}}$ is _s.t._
$\displaystyle\operatorname*{Card}\big{(}\\{i\in\bar{{\mathcal{K}}}|k-\kappa_{k}(h)\leq
i\leq k\\}\big{)}=h$ (36)
where $\operatorname*{Card}({\mathcal{S}})$ stands for the cardinality (number
of elements) of the set ${\mathcal{S}}$.
###### Lemma 5.3
Consider the system (1) subject to (3) and the matrix $P_{k}$ computed in line
with either ((10c)-(10e) and (11)) or (13) on one hand and (26), on the other.
Let $H_{k_{1}}\in~{}{\mathop{\rm I\mkern-3.5muR}}^{n\times{q}_{k}}$ whose
columns form an orthonormal basis for $\mathpzc{R}\big{(}P_{k}\big{)}$, where
${q}_{k}{\,:=\,}\operatorname*{rank}(P_{k})$ and let
$\bar{P}_{k}{\,:=\,}H_{k_{1}}P_{k}H_{k_{1}}^{T}$. If the pair
$\\{A_{k},\bar{F}_{k}^{T}\\}$ is _sporadically observable_ and
$\\{A_{k},R_{k}\\}$ is uniformly controllable, then there exist positive
finite numbers $\varrho_{k_{1}}$ and $\varrho_{k_{2}}$, _s.t._ for all
$k\geq\kappa_{k}(l)$,
$\displaystyle\varrho_{k_{1}}I_{{q}_{k}}$
$\displaystyle\leq\bar{P}_{k}\leq\varrho_{k_{2}}I_{{q}_{k}},$ (37)
$\kappa_{k}$ being given in (36) of Definition 5.4.
Proof. The proof of this lemma is detailed in Appendix C.2. ❑
The following theorem shows the stability of the estimation algorithm
according to the choice of the optimisation criterion made for $\beta_{k_{i}}$
in (26j).
###### Theorem 5.4
Consider the system (1) subject to (3) and its state estimation algorithm
given by either Thms 3.6 and 3.7 or 3.8, on one hand and 4.4, on the other.
Let
0pt $\displaystyle\bar{\mathscr{D}}_{k}$
$\displaystyle{\,:=\,}\bar{\mathscr{G}}_{k}\cup\underaccent{\bar}{\mathscr{G}}_{k}\cup\mathscr{D}_{k}=\mathscr{P}_{k}-\mathscr{H}_{k},\text{
{\it cf.\ }\eqref{bounds},}$ (38a) $\displaystyle\bar{F}_{k}$
$\displaystyle{\,:=\,}{F}_{k}(\bar{\mathscr{D}}_{k}){\,:=\,}[\bm{f}_{k_{i}}]_{i\in\bar{\mathscr{D}}_{k}}\in{\mathop{\rm
I\mkern-3.5muR}}^{n\times\bar{p}},$ (38b) $\displaystyle\bar{p}$
$\displaystyle{\,:=\,}\bar{p}_{k}{\,:=\,}\operatorname*{Card}(\bar{\mathscr{D}}_{k})\text{
and }\bar{{\mathcal{K}}}=\\{k\in{\mathop{\rm
I\mkern-3.5muN}}^{*}|\bar{p}_{k}\neq 0\\},$ (38c) $\displaystyle\text{setting
\quad}\bar{F}_{k}\bar{F}_{k}^{T}$ $\displaystyle{\,:=\,}0_{n,n},\quad\forall
k\notin\bar{{\mathcal{K}}}.$ (38d) For each time step $k\in{\mathop{\rm
I\mkern-3.5muN}}^{*}$ and each measurement $i\in\bar{\mathscr{D}}_{k}$,
consider $\beta^{*}$ the value of $\beta_{k_{i}}$ defined in (26j). If $A_{k}$
is invertible and the pairs $\\{A_{k},\bar{F}_{k}^{T}\\}$ and
$\\{A_{k},R_{k}\\}$ are _sporadically observable_ 555cf. Definition 5.4. and
uniformly controllable666cf. Definition 5.3. _resp_., then
1. 1.
the volume of ${\mathcal{E}}_{k}$ and all its axes lengths are bounded,
$\forall\beta^{*}\in]0,1[$, in (26j);
2. 2.
If $\beta^{*}=\beta_{\varsigma}$ given in (22c) of Thm 4.3.1 or Algo. 3, then
$\mathscr{V}_{{P}_{k}}:{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}\rightarrow{\mathop{\rm I\mkern-3.5muR}}_{+}$,
$\bm{x}\mapsto~{}\\!\mathscr{V}_{{P}_{k}}(\bm{x}){\,:=\,}\bm{x}^{T}P_{k}^{\dagger}\bm{x}$
is an ISS-Lyapunov function777cf. Definition 5.2 for the estimation error
$\tilde{\bm{x}}_{k}{\,:=\,}\mbox{\boldmath$x$}_{k}-\hat{\mbox{\boldmath$x$}}_{k}$,
which is ISS888cf. Lemma 5.2.
3. 3.
If $\beta^{*}=\beta_{\text{v}}$ given by Thm 4.3.2 and Algo 4, then
$\tfrac{1}{{v}_{k}}\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}_{k}\big{)}$
is nonincreasing and
$\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}_{k}\big{)}\leq{v}_{k}\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}_{0}\big{)}$;
4. 4.
If $\beta^{*}=\beta_{\text{s}}$ given by Thm 4.3.3 and Algo. 5, then
$\tfrac{1}{{s}_{k}}\operatorname*{ssal}\big{(}{\mathcal{E}}_{k}\big{)}$ is
nonincreasing and
$\operatorname*{ssal}\big{(}{\mathcal{E}}_{k}\big{)}\leq{{s}_{k}}\operatorname*{ssal}\big{(}{\mathcal{E}}_{0}\big{)}$;
where
$\displaystyle{v}_{k}$
$\displaystyle{\,:=\,}{\prod_{j=0}^{k-1}{\left|\bar{A}_{j}P_{j}P_{j}^{\dagger}\right|}_{{\dagger}}^{2}{\left|\bar{Q}_{j}^{\dagger}\bar{Q}_{j}+\big{(}\bar{Q}_{j}^{\dagger}+I_{n}-\bar{Q}_{j}^{\dagger}\big{)}W_{j}\right|}_{{\dagger}}},\
{v}_{0}=1;$ (38e) $\displaystyle{s}_{k}$
$\displaystyle{\,:=\,}{\prod_{j=0}^{k-1}\Big{(}\operatorname*{tr}\big{(}A_{j}A_{j}^{T}\big{)}+\tfrac{\operatorname*{tr}(W_{j})}{\operatorname*{tr}(P_{j})}\Big{)}},\
{s}_{0}=1;$ (38f) $\displaystyle\bar{Q}_{j}$
$\displaystyle{\,:=\,}\bar{A}_{j}P_{j}\bar{A}_{j}^{T};$ (38g)
$\displaystyle\bar{A}_{j}$ $\displaystyle{\,:=\,}\lambda_{j}A_{j},\text{ where
}\lambda_{j}{\,:=\,}\prod_{i=1}^{m}\sqrt{1+\mu_{i}}=\sqrt{\chi_{1}},\
\forall\mu_{i}\in]0,1[;$ (38h) $\displaystyle W_{j}$
$\displaystyle{\,:=\,}\tfrac{1}{\varsigma_{j}}R_{j}\operatorname*{Diag}\big{(}\tfrac{\chi_{i}}{{\mu_{i}}}\big{)}_{i=1}^{m}R_{j}^{T},\text{
where }\chi_{i}=\prod_{j=i}^{m}(1+\mu_{j}).$ (38i)
5. 5.
Now, let $R_{k}=0_{n,m}$, $\forall k\in{\mathop{\rm I\mkern-3.5muN}}$. If
$\beta^{*}=\beta_{\varsigma}$ $\mathscr{V}_{{P}_{k}}$ is a nonincreasing
Lyapunov function for the estimation error which is Lyapunov stable.
6. 6.
Furthermore, if $\left\|A_{k}\right\|\leq 1$, $\forall k\in{\mathop{\rm
I\mkern-3.5muN}}$, then999There is an error in the formulation of this point
in the paper submitted to L-CSS.
1. $i$.
all the axes lengths of ${\mathcal{E}}_{k}$ are nonincreasing, if
$\beta^{*}=\beta_{\varsigma}$;
2. $ii$.
$\operatorname*{\textstyle\operatorname*{vol}_{\dagger}\\!}\big{(}{\mathcal{E}}_{k}\big{)}$
is nonincreasing, if $\beta^{*}=\beta_{\text{v}}$;
3. $iii$.
$\operatorname*{ssal}\big{(}{\mathcal{E}}_{k}\big{)}$ is nonincreasing, if
$\beta^{*}=\beta_{\text{s}}$.
Proof. The proof is detailed in the Appendix C.3. ❑
## 6 NUMERICAL SIMULATIONS
First, for the sake of graphic illustration, the presented algorithm is
applied to a second order randomly generated system with coil-shaped input and
one (either strip or halfspace type) measurement, also randomly generated at
each time-step for $k=0,\cdots,100$. The figures 6(a) and 6(b) show the
evolution of the ellipsoid ${\mathcal{E}}_{k}$ for a stable model (the
eigenvalues of state matrix $A{\,:=\,}A_{k}$ are less than 1) and a model at
the stability limit (the eigenvalues $A{\,:=\,}A_{k}$ are 1), _resp_.
(a) Stable state model
(b) Unstable state model
Secondly, in order to evaluate the algorithm performances, the matrices of the
system model (1) and (3) are generated randomly for two values of the state
dimension: $n=10$ and $n=~{}100$, with $\varrho=q=r=s=\frac{n}{2}$, $m=n$,
$\pi=\frac{n}{5}$ and $\bm{\mu}=\bm{\mu}^{\operatorname*{tr}}_{k}$; the input
vector $B_{k-1}\mbox{\boldmath$\tau$}_{k-1}$ contains sine entries of random
magnitude and frequency. $P_{0}=100I_{n}$, $\varsigma_{0}=1$ and
$\hat{\mbox{\boldmath$x$}}_{k_{0}}$ randomly chosen on the boundary of
${\mathcal{E}}_{0}$. The measurements are available at all time steps
${\mathcal{K}}=\\{1,\ldots,N\\}$ in the case 1; at some randomly chosen time
steps, in case 2 and ${\mathcal{K}}=\emptyset$ in the case 3 (where only
prediction stage is performed without any measurement correction). For each
case, the simulations are run 25 times under MATLAB R2018b on Intel Core i7
(2.3GHz, 8G RAM), each one for a different system model and containing
$N=~{}100$ time steps. The results are summarized in Table 1. Let
$\overline{\varsigma\operatorname*{tr}(P)}{\,:=\,}\underset{k\in\\{1,\ldots,N\\}}{\operatorname*{mean}\varsigma_{k}\operatorname*{tr}(P_{k})}$:
the average sum of ${\mathcal{E}}_{k}$’s squared axes lengths,
$\overline{\left\|\tilde{\bm{x}}\right\|}{\,:=\,}\underset{k\in\\{1,\ldots,N\\}}{\operatorname*{mean}\left\|\tilde{\bm{x}}_{k}\right\|}$:
the mean estimation error vector norm and $T$: the average computational time
for the simulation horizon of $N$ time steps.
$n$ | ${\mathcal{K}}$ | $\dfrac{\varsigma_{0}\operatorname*{tr}(P_{N})}{\varsigma_{N}\operatorname*{tr}(P_{0})}$ | $\overline{\varsigma\operatorname*{tr}(P)}$ | $\dfrac{\left\|\tilde{\bm{x}}_{N}\right\|}{\left\|\tilde{\bm{x}}_{0}\right\|}$ | $\overline{\left\|\tilde{\bm{x}}\right\|}$ | $T(ms)$
---|---|---|---|---|---|---
$10$ | case 1 | 0.010 | 1 141 | 0.027 | 3.02 | 18
case 2 | 0.026 | 2 206 | 0.053 | 4.60 | 15
case 3 | 0.058 | 7 540 | 0.066 | 7.25 | 10
$100$ | case 1 | 0.712 | 7$\cdot 10^{5}$ | 0.264 | 26.68 | 1870
case 2 | 0.707 | 2$\cdot 10^{6}$ | 0.269 | 46.19 | 1321
case 3 | 5.774 | 5$\cdot 10^{6}$ | 0.702 | 71.16 | 700
Table 1: Simulation results
It is plain to see that the algorithm considered here exhibits better
performances for systems of rather smaller dimension but it can still be
fairly efficiently used with very high dimensional systems provided enough
measurements are available. Moreover, given its low running time, it can be
implemented online with such systems.
## 7 CONCLUSION
An ellipsoidal state bounding approach was proposed for discrete-time linear
dynamic models with sporadic measurements, corrupted by additive unknown
process and measurement disturbances and subject to equality constraints on
the state vector. The following aspects were of special concern in the design
of the state estimation algorithm: 1\. the optimization of its inherent
parameters ($\mu$ and $\beta$), while dealing with equality constraints that
induce rank loss of the ellipsoid’s shape matrix, thus making a difference
with the existing algorithms. 2\. the numerical stability and the
computational undemandingness, in terms of time and memory, what makes this
algorithm suitable for high dimensional systems. 3\. the formulation of the
conditions of convergence in spite of the sporadic character of the
measurements and the state constraints manifesting themselves episodically,
now and then, not necessarily together and not at all time steps. 4\. the easy
applicability of the proposed technique by presenting detailed and turnkey
pseudocodes.
## Appendix A Proofs of the results of section 3
### A.1 Proof of Lemma 3.1
The set containing every possible value of
$\bar{\bm{x}}{\,:=\,}A\bm{x}+\bm{\eta}$ can be schematized by $\bar{\bm{x}}\in
A{\mathcal{E}}\oplus{\mathcal{Z}}\subset\bar{{\mathcal{E}}}(\mu)$; where
$A{\mathcal{E}}{\,:=\,}{\mathcal{E}}(A\bm{c},\varsigma APA^{T})$ is the image
of the ellipsoid ${\mathcal{E}}{\,:=\,}{\mathcal{E}}(\bm{c},\varsigma P)$ by
the endomorphism of matrix $A$; and
$\bar{{\mathcal{E}}}(\mu){\,:=\,}{\mathcal{E}}\big{(}\bar{\bm{c}},\varsigma\bar{P}(\mu)\big{)}$
is the outer-bounding ellipsoid of the Minkowski sum (cf. $\mathsection$.1.9.)
of $A{\mathcal{E}}$ and the one-dimensional zonotope, i.e., interval segment,
that is also a one-dimensional ellipsoid:
${\mathcal{Z}}{\,:=\,}{\mathcal{Z}}(\bm{u},\bm{r})={\mathcal{E}}(\bm{u},\bm{r}\bm{r}^{T})$.
Now, the Minkowski sum of two ellipsoids ${\mathcal{E}}(\bm{c}_{1},P_{1})$ and
${\mathcal{E}}(\bm{c}_{2},P_{2})$ is not an ellipsoid, in general, yet can be
bounded by a parametrized ellipsoid [19]:
$\displaystyle{\mathcal{E}}(\bm{c},P(\mu))$
$\displaystyle\supset{\mathcal{E}}(\bm{c}_{1},P_{1})\oplus{\mathcal{E}}(\bm{c}_{2},P_{2}),\
\forall\mu\in{\mathop{\rm I\mkern-3.5muR}}_{+}^{*},$ (39a)
$\displaystyle\text{ where }\bm{c}$ $\displaystyle=\bm{c}_{1}+\bm{c}_{2}\text{
and }P(\mu)=(1+\mu)P_{1}+(1+\leavevmode\kern 1.00006pt\raise
2.15277pt\hbox{$1$}\kern-1.00006pt/\kern-1.49994pt\lower
1.07639pt\hbox{$\mu$})P_{2}.$ (39b)
Setting $P_{1}{\,:=\,}\varsigma APA^{T}$, $\bm{c}_{1}=A\bm{c}$,
$P_{2}{\,:=\,}\bm{r}\bm{r}^{T}$ and $\bm{c}_{2}=\bm{u}$ completes the proof.
### A.2 Proof of Proposition 3.2
$\operatorname*{rank}\big{(}{Q_{+}}\big{)}=\operatorname*{rank}(Q+a\bm{r}\bm{r}^{T})$.
0pt
1. 1.
Let $\bar{q}{\,:=\,}n-{q}$. There exists a unitary matrix $U$, _s.t._
$U^{T}QU=\Big{[}{{\Sigma}^{2}\atop 0_{\bar{q},{q}}}{0_{{q},\bar{q}}\atop
0_{\bar{q},\bar{q}}}\Big{]}$, where $\Sigma\in~{}\\!{{{\mathop{\rm
I\mkern-3.5muR}}}^{q\times q}}$ is a diagonal matrix with nonzero singular
values of ${Q}$ on its diagonal. Let
$\bar{\bm{r}}{\,:=\,}\sqrt{a}{\Sigma}^{-1}\big{[}I_{q}\
{\mathbf{0}}_{q,\bar{q}}\big{]}U^{T}\bm{r}\in~{}\\!{{{\mathop{\rm
I\mkern-3.5muR}}}^{q}}$ and
$\underaccent{\bar}{\rb}{\,:=\,}\sqrt{a}\big{[}{\mathbf{0}}_{\bar{q},q}\
I_{\bar{q}}\big{]}U^{T}\bm{r}\in{\mathop{\rm I\mkern-3.5muR}}^{\bar{q}}$.
$\displaystyle Q+a\bm{r}\bm{r}^{T}$ $\displaystyle=U\left[{\Sigma\atop
0_{\bar{q},{q}}}{0\atop
0}{\Sigma\bar{\bm{r}}\atop\underaccent{\bar}{\rb}}\right]\left[{\Sigma\atop
0_{\bar{q},{q}}}{0\atop
0}{\Sigma\bar{\bm{r}}\atop\underaccent{\bar}{\rb}}\right]^{T}U^{T}=U\left[{\Sigma\atop
0_{\bar{q},{q}}}{\Sigma\bar{\bm{r}}\atop\underaccent{\bar}{\rb}}\right]\left[{\Sigma\atop
0_{\bar{q},{q}}}{\Sigma\bar{\bm{r}}\atop\underaccent{\bar}{\rb}}\right]^{T}\\!U^{T},\\!\\!\\!$
(40)
$\displaystyle\operatorname*{rank}(Q+\bm{r}\bm{r}^{T})=\operatorname*{rank}\left(\left[{\Sigma\atop
0_{\bar{q},{q}}}{0_{{q},\bar{q}}\atop I_{\bar{q}}}\right]\left[{I_{q}\atop
0_{\bar{q},q}}{\bar{\bm{r}}\atop\underaccent{\bar}{\rb}}\right]\right)=\operatorname*{rank}\left(\left[{I_{q}\atop
0_{\bar{q},q}}{\bar{\bm{r}}\atop\underaccent{\bar}{\rb}}\right]\right)$. Then
if
$\mbox{\boldmath$v$}{\,:=\,}(I_{n}-QQ^{\dagger})\bm{r}\neq{\mathbf{0}}_{n}$,
which is the projection of $\bm{r}$ onto $\mathpzc{Ker}({Q})$, meaning that
$\underaccent{\bar}{\rb}\neq{\mathbf{0}}_{\bar{q}}$, implying that
$\operatorname*{rank}\big{(}{Q_{+}}\big{)}=\operatorname*{rank}(Q+a\bm{r}\bm{r}^{T})=~{}\\!q+1$;
otherwise $\operatorname*{rank}\big{(}{Q_{+}}\big{)}=~{}\\!q$.
2. 2.
From (40),
$\displaystyle Q+a\bm{r}\bm{r}^{T}$ $\displaystyle=U\left[{\Sigma\atop
0_{\bar{q},{q}}}{0_{{q},\bar{q}}\atop
I_{\bar{q}}}\right]\left[{I_{q}+\bar{\bm{r}}\bar{\bm{r}}^{T}\atop\underaccent{\bar}{\rb}\bar{\bm{r}}^{T}}{\bar{\bm{r}}\underaccent{\bar}{\rb}^{T}\atop\underaccent{\bar}{\rb}\underaccent{\bar}{\rb}^{T}}\right]\left[{\Sigma\atop
0_{\bar{q},{q}}}{0_{{q},\bar{q}}\atop I_{\bar{q}}}\right]^{T}U^{T}$
$\displaystyle{\left|Q+a\bm{r}\bm{r}^{T}\right|}_{{\dagger}}$
$\displaystyle={\left|\Sigma\right|}^{2}{\left|\left[\begin{smallmatrix}I_{q}&0_{q,\bar{q}}&{\mathbf{0}}_{q}\\\
0_{\bar{q},q}&I_{\bar{q}}&{\mathbf{0}}_{\bar{q}}\\\
\bar{\bm{r}}^{T}&\underaccent{\bar}{\rb}^{T}&1\end{smallmatrix}\right]\left[\begin{smallmatrix}I_{q}+\bar{\bm{r}}\bar{\bm{r}}^{T}&\bar{\bm{r}}\underaccent{\bar}{\rb}^{T}&\bar{\bm{r}}\\\
\underaccent{\bar}{\rb}\bar{\bm{r}}^{T}&\underaccent{\bar}{\rb}\underaccent{\bar}{\rb}^{T}&\underaccent{\bar}{\rb}\\\
0_{1,q}&0_{1,\bar{q}}&1\end{smallmatrix}\right]\left[\begin{smallmatrix}I_{q}&0_{q,\bar{q}}&{\mathbf{0}}_{q}\\\
0_{\bar{q},q}&I_{\bar{q}}&{\mathbf{0}}_{\bar{q}}\\\
-\bar{\bm{r}}^{T}&-\underaccent{\bar}{\rb}^{T}&1\end{smallmatrix}\right]\right|}_{{\dagger}}$
$\displaystyle={\left|\Sigma\right|}^{2}{\left|\left[\begin{smallmatrix}I_{q}&0_{q,\bar{q}}&\bar{\bm{r}}\\\
0_{\bar{q},q}&0_{\bar{q}}&\underaccent{\bar}{\rb}\\\
{\mathbf{0}}_{q}^{T}&-\underaccent{\bar}{\rb}^{T}&\bar{\bm{r}}^{T}\bar{\bm{r}}+\underaccent{\bar}{\rb}^{T}\underaccent{\bar}{\rb}+1\end{smallmatrix}\right]\right|}_{{\dagger}}=\begin{cases}{\left|\Sigma\right|}^{2}\big{(}\bar{\bm{r}}^{T}\bar{\bm{r}}+1\big{)},&\text{
if }\underaccent{\bar}{\rb}\neq{\mathbf{0}}_{\bar{q}}\\\
{\left|\Sigma\right|}^{2}\underaccent{\bar}{\rb}^{T}\underaccent{\bar}{\rb},&\text{
otherwise.}\end{cases}$ (41)
Now, noticing that
${\left|bQ_{+}\right|}_{{\dagger}}=\prod_{\sigma_{i}(Q_{+})\neq
0}b\sigma_{i}(Q_{+})=b^{q_{+}}{\left|Q_{+}\right|}_{{\dagger}}$, that
${\left|\Sigma\right|}^{2}={\left|Q\right|}_{{\dagger}}$ and since
$\underaccent{\bar}{\rb}^{T}\underaccent{\bar}{\rb}=a\mbox{\boldmath$v$}^{T}\mbox{\boldmath$v$}$
and $\bar{\bm{r}}^{T}\bar{\bm{r}}=a\bm{r}^{T}Q^{\dagger}\bm{r}$, the proof of
(6b) is completed.
### A.3 Proof of Theorem 3.3
The pseudo-volume of an ellipsoid being proportional to the pseudo-determinant
of its shape matrix and $\varsigma$ being considered as constant _w.r.t._
$\mu$, 0pt
$\displaystyle\mu_{\text{v}}{\,:=\,}\arg\min_{\mu\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}{\left|\bar{P}(\mu)\right|}_{{\dagger}}=\arg\min_{\mu\in{\mathop{\rm
I\mkern-3.5muR}}_{+}^{*}}\log{\left|Q\right|}_{{\dagger}}.$
According to Proposition 3.2, where $b{\,:=\,}(1+\mu)$ and
$a{\,:=\,}\tfrac{1}{\varsigma\mu}$, two cases are distinguished.
Case 1:
the projection of the vector $\bm{r}$ onto the nullspace of ${Q}$, embodied by
$v$ is a zero vector, in this case
$\operatorname*{rank}\big{(}\bar{P}(\mu)\big{)}=\operatorname*{rank}\big{(}Q_{+})=\operatorname*{rank}({Q})=q$
and
$\displaystyle\bar{P}(\mu)$
$\displaystyle{\,:=\,}(1+\mu)\left({Q}+\tfrac{1}{\varsigma\mu}\bm{r}\bm{r}^{T}\right)\Rightarrow{\left|\bar{P}(\mu)\right|}_{{\dagger}}={\left|Q\right|}_{{\dagger}}(1+\mu)^{q}\big{(}1+\tfrac{h}{\mu}\big{)}$
$\displaystyle\tfrac{\partial}{\partial\mu}\log{\left|\bar{P}(\mu)\right|}_{{\dagger}}=0$
$\displaystyle\Leftrightarrow\tfrac{{q}}{1+\mu}-\tfrac{h}{\mu(\mu+h)}=0\Leftrightarrow\
{q}\mu^{2}+h({q}-1)\mu-h=0.$ (42)
$\mu_{\text{v}}$ in (7a), is the only positive solution to the quadratic
equation (42).
Case 2:
the projection of $\bm{r}$ onto the nullspace of ${Q}$ is nonzero, thus
implying the rank incrementation of the shape matrix:
$\operatorname*{rank}\big{(}\bar{P}(\mu)\big{)}=q+1$ and
$\displaystyle{\left|\bar{P}(\mu)\right|}_{{\dagger}}={\left|Q\right|}_{{\dagger}}\left\|\mbox{\boldmath$v$}\right\|^{2}\tfrac{(1+\mu)^{q+1}}{\varsigma\mu}\Leftrightarrow\tfrac{\partial}{\partial\mu}\log{\left|\bar{P}(\mu)\right|}_{{\dagger}}=0$
$\displaystyle\Leftrightarrow\tfrac{{q}+1}{1+\mu}-\tfrac{1}{\mu}=0,$
this results in $\mu=\tfrac{1}{q}$ of (7a).
### A.4 Proof of Theorem 3.6
Conforming to (1a), the set containing every possible value of
$\mbox{\boldmath$x$}_{k}$ can be schematized by
$\displaystyle
A_{k}{\mathcal{E}}_{k}\oplus{\mathcal{Z}}(B_{k}\mbox{\boldmath$\tau$}_{k},R_{k})={\mathcal{E}}_{{k+1/k}_{0}}\oplus{\mathcal{Z}}({\mathbf{0}}_{n},R_{k}).$
(43)
Now, the zonotope ${\mathcal{Z}}({\mathbf{0}}_{n},R_{k})$, where
$R_{k}=\big{[}\mbox{\boldmath$r$}_{k_{1}}\cdots\mbox{\boldmath$r$}_{k_{m}}\big{]}$,
can be viewed as the sum of $m$ degenerate ellipsoids [33]:
$\displaystyle{\mathcal{Z}}({\mathbf{0}}_{n},R_{k})=\operatorname*{\oplus}_{i=1}^{m}{\mathcal{Z}}({\mathbf{0}}_{n},{\bm{r}}_{k_{i}})=\operatorname*{\oplus}_{i=1}^{m}{\mathcal{E}}({\mathbf{0}}_{n},{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T}).$
Applying Lemma 3.1 sequentially to
$\displaystyle\displaystyle{\mathcal{E}}_{k+1/k}\supset{\mathcal{E}}_{{k+1/k}_{0}}\oplus\left(\operatorname*{\oplus}_{i=1}^{m}{\mathcal{E}}({\mathbf{0}}_{n},{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T})\right),$
(44)
eventuates in (10).
### A.5 Proof of Theorem 3.8
Since the SSAL of an ellipsoid is the trace of its shape matrix,
$\displaystyle\mu_{{\text{s}}_{i}}=~{}\\!\arg\min_{\mu_{i}}\operatorname*{tr}(P_{{k+1/k}_{i}})$;
(12) is the direct corollary of Theorem 3.8. As for (13), it is a direct
consequence of the result [12] saying that the minimum trace ellipsoid
containing the Minkowski sum of $m+1$ ellipsoids is :
$\displaystyle{\mathcal{E}}(\bm{c},P)$
$\displaystyle{\,:=\,}\oplus_{i=0}^{m}{\mathcal{E}}(\bm{c}_{i},P_{i}),$ (45a)
where $\displaystyle\bm{c}{\,:=\,}\sum_{i=0}^{m}\bm{c}_{i}\text{ and }P$
$\displaystyle{\,:=\,}\Big{(}\sum_{i=0}^{m}\sqrt{\operatorname*{tr}(P_{i})}\Big{)}\Big{(}\sum_{i=0}^{m}\big{(}\sqrt{\operatorname*{tr}(P_{i})}\big{)}^{-1}{P_{i}}\Big{)}.$
(45b)
Then, after noticing that
$\displaystyle\operatorname*{tr}(\mbox{\boldmath$r$}_{{k}_{i}}\mbox{\boldmath$r$}_{{k}_{i}}^{T})=\mbox{\boldmath$r$}_{{k}_{i}}^{T}\mbox{\boldmath$r$}_{{k}_{i}}=\left\|\mbox{\boldmath$r$}_{{k}_{i}}\right\|^{2}$
(46) and that
$\displaystyle\sum_{i=1}^{m}\left({\sqrt{\operatorname*{tr}(\mbox{\boldmath$r$}_{{k}_{i}}\mbox{\boldmath$r$}_{{k}_{i}}^{T})}}\right)^{-1}{{\bm{r}}_{k_{i}}{\bm{r}}_{k_{i}}^{T}}=\tilde{R}_{k}$
(47)
(45) applied to
$\displaystyle{\mathcal{E}}_{{k+1/k}_{0}}\oplus\big{(}\oplus_{i=1}^{m}{\mathcal{E}}({\mathbf{0}}_{n},\mbox{\boldmath$r$}_{{k}_{i}}\mbox{\boldmath$r$}_{{k}_{i}}^{T}\big{)}$,
leads clearly to (13). It is also stated in [12] that such an ellipsoid is the
same that the one obtained sequentially in (9).
## Appendix B Proofs of the results of section 4
### B.1 Proof of Theorem 4.1
###### Definition B.1
The signed distance from a set ${\mathcal{S}}\subset{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$ to a vector $\mbox{\boldmath$x$}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$ is
$\displaystyle\zeta({\mathcal{S}},\mbox{\boldmath$x$}){\,:=\,}\max_{\left\|\bm{u}\right\|=1}\bm{u}^{T}\mbox{\boldmath$x$}-\rho_{\mathcal{S}}(\bm{u})$.
###### Proposition B.1 ([34])
The signed distance from an ellipsoid to a hyperplane is given by:
$\displaystyle\zeta\big{(}{\mathcal{E}}(\bm{c},P),{\mathcal{H}}(\bm{d},a)\big{)}{\,:=\,}\left\|\bm{d}\right\|^{-1}\Big{(}{\left|a-\bm{c}^{T}\bm{d}\right|}-\sqrt{\bm{d}^{T}P\bm{d}}\Big{)}.$
(48)
Let ${\mathcal{E}}{\,:=\,}{\mathcal{E}}(\bm{c},\varsigma P)$. The signed
distance from ${\mathcal{E}}$ to ${\mathcal{H}}(\bm{f},\bar{y})$ is
$\displaystyle\zeta{\,:=\,}\left\|\bm{f}\right\|^{-1}\left({\left|\bar{y}-\bm{c}^{T}\bm{f}\right|}-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}\right).$
(49)
* •
$\zeta\geq 0$ means that ${\mathcal{H}}(\bm{f},\bar{y})$ does not intersect
${\mathcal{E}}$ in more than one point:
1. 1.
if $\bm{c}^{T}\bm{f}>\bar{y}$, then
${\mathcal{E}}\subset~{}{\mathcal{G}}(-\bm{f},-\bar{y})$ and
${\mathcal{E}}\cap{\mathcal{G}}(\bm{f},\bar{y})=\emptyset\Leftrightarrow$
(19b);
2. 2.
if $\bm{c}^{T}\bm{f}\leq\bar{y}$, then
${\mathcal{E}}\subset{\mathcal{G}}(\bm{f},\bar{y})\Rightarrow{\mathcal{E}}\cap{\mathcal{G}}(\bm{f},\bar{y})={\mathcal{E}}\Leftrightarrow$
(19c);
3. 3.
if $\bm{c}^{T}\bm{f}-\bar{y}=\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$, then
${\mathcal{H}}(\bm{f},\bar{y})$ is tangent to ${\mathcal{E}}$ and
${\mathcal{E}}\cap{\mathcal{G}}(\bm{f},\bar{y})={\mathcal{E}}\cap{\mathcal{H}}(\bm{f},\bar{y})=\\{\check{\bm{c}}\\}$,
where $\check{\bm{c}}$ is given in (25e) and can be calculated using (25g),
letting $\delta\leftarrow\bar{y}-\bm{c}^{T}\bm{f}$, with
$\delta^{2}={\varsigma\bm{f}^{T}P\bm{f}}$, $\bm{f}\leftarrow\bm{f}$,
$P\leftarrow\varsigma^{-1}P$.
* •
If $\zeta\leq 0$, then ${\mathcal{H}}(\bm{f},\bar{y})$ intersects
${\mathcal{E}}$ and ${\mathcal{H}}(-\bm{f},\rho(-\bm{f}))$ is the ellipsoid’s
supporting hyperplane of normal vector $-\bm{f}$ which is contained in
${\mathcal{G}}(\bm{f},\bar{y})$. Indeed,
$\displaystyle\mbox{\boldmath$x$}\in{\mathcal{H}}(-\bm{f},\rho(-\bm{f}))$
$\displaystyle\Leftrightarrow\mbox{\boldmath$x$}^{T}\bm{f}-\bm{c}^{T}\bm{f}=-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}\leq\bar{y}-\bm{c}^{T}\bm{f}$
$\displaystyle\Rightarrow\mbox{\boldmath$x$}^{T}\bm{f}\leq\bar{y}\Leftrightarrow\mbox{\boldmath$x$}\in{\mathcal{G}}(\bm{f},\bar{y}).$
(50) Thence, ${\mathcal{G}}\big{(}{-\bm{f}},\rho(-\bm{f})\big{)}$ is its
supporting halfspace and
${\mathcal{E}}\subset{\mathcal{G}}\big{(}-\bm{f},\rho(-\bm{f})\big{)}$.
Therefore, $\displaystyle{\mathcal{E}}\cap{\mathcal{G}}(\bm{f},\bar{y})$
$\displaystyle=\Big{(}{\mathcal{G}}\big{(}{-\bm{f}},\rho(-\bm{f})\big{)}\cap{\mathcal{G}}(\bm{f},\bar{y})\Big{)}\cap{\mathcal{E}}$
(51)
${\left|\bar{y}-\bm{c}^{T}\bm{f}\right|}<\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$
means that $0<\bar{y}+\rho(-\bm{f})<2\sqrt{\varsigma\bm{f}^{T}P\bm{f}}$,
entailing, on one hand, $\displaystyle{\mathcal{G}}(\bm{f},\bar{y})$
$\displaystyle=\big{\\{}\mbox{\boldmath$x$}\big{|}\mbox{\boldmath$x$}^{T}\bm{f}\leq\bar{y}\big{\\}}=\bigg{\\{}\mbox{\boldmath$x$}\bigg{|}\frac{2}{\bar{y}+\rho(-\bm{f})}\mbox{\boldmath$x$}^{T}\bm{f}\leq\frac{2\bar{y}}{\bar{y}+\rho(-\bm{f})}\bigg{\\}}$
$\displaystyle={\mathcal{G}}(\gamma^{-1}\bm{f},{y}+1)$ (52)
$\displaystyle\text{and }{\mathcal{G}}(-\bm{f},\rho(-\bm{f}))$
$\displaystyle=\bigg{\\{}\mbox{\boldmath$x$}\bigg{|}-\frac{2}{\bar{y}+\rho(-\bm{f})}\mbox{\boldmath$x$}^{T}\bm{f}\leq\frac{2\rho(-\bm{f})}{\bar{y}+\rho(-\bm{f})}\bigg{\\}}$
$\displaystyle={\mathcal{G}}(-\gamma^{-1}\bm{f},-{y}+1),$ (53) on the other
hand. Finally, the proof (19e)–(19f) is achieved thusly:
$\displaystyle{\mathcal{E}}\cap{\mathcal{G}}(\bm{f},\bar{y})$
$\displaystyle={\mathcal{E}}\cap\Big{(}{\mathcal{G}}\big{(}\gamma^{-1}\bm{f},{y}+1\big{)}\cap{\mathcal{G}}\big{(}-\gamma^{-1}\bm{f},-{y}+1\big{)}\Big{)}$
(54)
$\displaystyle={\mathcal{E}}\cap{\mathcal{D}}\big{(}\gamma^{-1}\bm{f},{y}\big{)}.$
❑
### B.2 Proof of Theorem 4.2
Let ${\mathcal{E}}{\,:=\,}{\mathcal{E}}(\bm{c},\varsigma P)$. The signed
distance from the ellipsoid ${\mathcal{E}}$ to each of the two hyperplanes
${\mathcal{H}}(\bm{f},y\mp 1)$, bounding the strip ${\mathcal{D}}(\bm{f},y)$,
is
$\displaystyle\zeta{\,:=\,}\left\|\bm{f}\right\|^{-1}\big{(}\big{|}y\mp
1-\bm{f}^{T}\bm{c}\big{|}-\sqrt{\varsigma\bm{f}^{T}P\bm{f}}\big{)}.$ (55)
When $\zeta>0$, the ellipsoid doesn’t intersect any of both hyperplanes
meaning either that the ellipsoid is located outside the strip, in which case
(case 1), the intersection is empty or that it is situated between them i.e.,
contained in the strip (case 2). In the case 3, the interior of the ellipsoid
is outside the strip touching it in only one point and the case 3 of Thm 4.1
is then applicable:
${\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}={\mathcal{G}}(\bm{f},y+1)\cap{\mathcal{E}}$
(case 3.a) and
${\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}={\mathcal{G}}(-\bm{f},-y+1)\cap{\mathcal{E}}$
(case 3.b). In the case 4, where $\zeta\leq~{}0$, the intersection is not
empty. It is then possible to introduce the following lemma, based on the
results of [17] and [30]:
###### Lemma B.2
$\forall y\in{\mathop{\rm I\mkern-3.5muR}}$, $\bm{c}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$, $\bm{f}\in{\mathop{\rm I\mkern-3.5muR}}^{n}$,
$\sigma\in~{}{\mathop{\rm I\mkern-3.5muR}}_{+}^{*}$ and SPD
$P\in{{{\mathop{\rm I\mkern-3.5muR}}}^{n\times n}}$, if
${\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\sigma P)\neq\emptyset$, then
$\displaystyle\forall\omega\in{\mathop{\rm I\mkern-3.5muR}}_{+}^{*},$
$\displaystyle\quad{\mathcal{E}}(\tilde{\bm{c}}(\omega),\tilde{\sigma}(\omega)\tilde{P}\big{(}\omega)\big{)}\supset{\mathcal{D}}(\bm{f},y)\cap{\mathcal{E}}(\bm{c},\sigma
P),$ (56a) where $\displaystyle\tilde{P}(\omega)$
$\displaystyle{\,:=\,}P-\tfrac{\omega}{\omega\alpha+{1}}P{\bm{f}}{\bm{f}}^{T}P,\quad$
(56b) $\displaystyle\tilde{\bm{c}}(\omega)$
$\displaystyle{\,:=\,}\bm{c}+\tfrac{\omega}{\omega\alpha+{1}}\delta
P{\bm{f}}=\bm{c}+\omega\tilde{P}(\omega)^{-1}{\bm{f}}\delta,\quad$ (56c)
$\displaystyle\tilde{\sigma}(\omega)$
$\displaystyle{\,:=\,}\sigma+\omega(1-\tfrac{\alpha}{\omega+\alpha}\delta^{2}),$
(56d) $\displaystyle\delta$ $\displaystyle{\,:=\,}y-{\bm{f}}^{T}\bm{c}\text{
and $\alpha$ is given in \eqref{alpha_def}}.$ (56e)
This lemma is precisely the mono-output case of the “observation update” part
of Thm 1, [14]. (21d), (21e) and (21f) are obtained by setting
$\omega{\,:=\,}{\alpha\beta}({1-\beta})^{-1}$, thus
$\beta={\omega}({\alpha+\omega})^{-1}$. But before applying the lemma above,
it is suitable to reduce the strip ${\mathcal{D}}(\bm{f},y)$ in case where one
of the two hyperplanes does not intersect the ellipsoid ${\mathcal{E}}$, i.e.,
when either $y+1>\bar{\rho}$ or $y-1<-\underaccent{\bar}{\rho}$, by
translating the aforementioned hyperplane so that it becomes tangent to the
ellipsoid, as proposed in [35]. The new strip so obtained is
${\mathcal{D}}(\gamma^{-1}\bm{f},\breve{y})$, where $\gamma$ and
$\breve{y}{\,:=\,}y$ are given in (19f) and result from applying (case 4) of
Thm 4.1 to ${\mathcal{E}}\cap{\mathcal{G}}(-\bm{f},-y+1)$ and to
${\mathcal{E}}\cap{\mathcal{G}}(\bm{f},y+1)$.
### B.3 Proof of Theorem 4.3.1
Applying the generalization of the Sherman-Morrison formula to the pseudo-
inverse of the matrix (21d) (cf. Corollary 3.5 [38]), we can write
$\displaystyle\breve{P}(\beta)^{\dagger}=P^{\dagger}+\tfrac{\alpha\beta}{1-\beta}{P^{\dagger}P\bm{f}\bm{f}^{T}PP^{\dagger}}.$
(57)
Now, noticing that $(PP^{\dagger})^{T}=P^{\dagger}P$ and recalling that,
$\forall\bm{u}\in\mathpzc{R}(P)$, $\exists\mbox{\boldmath$v$}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$, $\bm{u}=P\mbox{\boldmath$v$}$, so
$PP^{\dagger}\bm{u}=PP^{\dagger}P\mbox{\boldmath$v$}=P\mbox{\boldmath$v$}=~{}\\!\bm{u}$,
in particular for $\bm{u}{\,:=\,}\mbox{\boldmath$x$}-\breve{\bm{c}}(\beta)$,
$\forall\mbox{\boldmath$x$}\in~{}\\!{\mathcal{E}}(\bm{c},\varsigma P)$, since
$\breve{\bm{c}}(\beta)\in{\mathcal{E}}(\bm{c},\varsigma P)$; replacing (57) in
(22a) leads then to
0pt
$\displaystyle{\mathscr{V}}_{\breve{P}(\beta)}\big{(}\bm{x}-\breve{\bm{c}}(\beta)\big{)}$
$\displaystyle{\,:=\,}\big{(}\bm{x}-\breve{\bm{c}}(\beta)\big{)}^{T}\Big{(}P^{\dagger}+\tfrac{\alpha\beta}{1-\beta}{P^{\dagger}P\bm{f}\bm{f}^{T}PP^{\dagger}}\Big{)}\big{(}\mbox{\boldmath$x$}-\breve{\bm{c}}(\beta)\big{)}$
$\displaystyle=\big{(}\bm{x}-\breve{\bm{c}}(\beta)\big{)}^{T}\Big{(}P^{\dagger}+\tfrac{\alpha\beta}{1-\beta}{\bm{f}\bm{f}^{T}}\Big{)}\big{(}\mbox{\boldmath$x$}-\breve{\bm{c}}(\beta)\big{)}.$
(58a) Inserting (21e) in (58a), we can show, by the mean of some standard
algebraic manipulations, that101010${\mathscr{V}}_{\breve{P}(\beta)}$ is
optimized on
${\mathcal{D}}(\breve{\bm{f}},\breve{y})\cap{\mathcal{E}}(\bm{c},\varsigma
P)$, it is then obvious that $\mbox{\boldmath$x$}\in\mathpzc{R}(P)$, since
$\bm{x}\in{\mathcal{E}}(\bm{c},\varsigma P)$., $\forall\bm{x}\in{{{\mathop{\rm
I\mkern-3.5muR}}}^{n}}$, _s.t._ $(\bm{x}-\bm{c})\in\mathpzc{R}(P)$ (since
$\bm{x}\in{\mathcal{E}}(\bm{c},\varsigma P)$),
$\displaystyle{\mathscr{V}}_{\breve{P}(\beta)}\big{(}\mbox{\boldmath$x$}-\breve{\bm{c}}(\beta)\big{)}$
|
# Independent Sets in Direct Products of Vertex-transitive Graphs
Huajun Zhang<EMAIL_ADDRESS>Department of Mathematics, Zhejiang Normal
University, Jinhua 321004, P.R. China Department of Mathematics, Shanghai
Normal University, Shanghai 200234, P.R. China
###### Abstract
The direct product $G\times H$ of graphs $G$ and $H$ is defined by:
$V(G\times H)=V(G)\times V(H)$
and
$E(G\times H)=\left\\{[(u_{1},v_{1}),(u_{2},v_{2})]:(u_{1},u_{2})\in
E(G)\mbox{\ and\ }(v_{1},v_{2})\in E(H)\right\\}.$
In this paper, we will prove that the equality
$\alpha(G\times H)=\max\\{\alpha(G)|H|,\alpha(H)|G|\\}$
holds for all vertex-transitive graphs $G$ and $H$, which provides an
affirmative answer to a problem posed by Tardif (Discrete Math. 185 (1998)
193-200). Furthermore, the structure of all maximum independent sets of
$G\times H$ are determined.
###### keywords:
direct product; primitivity; independence number; vertex-transitive
MSC: 05D05, 06A07
## 1 Introduction
Let $G$ and $H$ be two graphs. The direct product $G\times H$ of $G$ and $H$
is defined by
$\displaystyle V(G\times H)=V(G)\times V(H)$
and
$\displaystyle E(G\times
H)=\left\\{[(u_{1},v_{1}),(u_{2},v_{2})]:(u_{1},u_{2})\in E(G)\mbox{\ and\
}(v_{1},v_{2})\in E(H)\right\\}.$
It is easy to see this product is commutative and associative, and the product
of more than two graphs is well-defined. For a graph $G$, the products
$G^{n}=G\times G\times\cdots\times G$ is called the $n$-th powers of $G$.
An interesting problem is the independence number of $G\times H$. It is clear
that if $I$ is an independent set of $G$ or $H$, then the preimage of $I$
under projections is an independent set of $G\times H$, and so $\alpha(G\times
H)\geq\max\\{\alpha(G)|H|,\alpha(H)|G|\\}.$ It is natural to ask whether the
equality holds or not. In general, the equality does not hold for non-vertex-
transitive graphs (see [13]). So Tardif [17] posed the following problem.
###### Problem 1.1
(Tardif [17]) Does the equality
$\alpha(G\times H)=\max\\{\alpha(G)|H|,\alpha(H)|G|\\}$
hold for all vertex-transitive graphs $G$ and $H$?
Furthermore, it immediately raises another interesting problem:
###### Problem 1.2
When $\alpha(G\times H)=\max\\{\alpha(G)|H|,\alpha(H)|G|\\}$, is every maximum
independent set of $G\times H$ the preimage of an independent set of one
factor under projections?
If the answer is yes, we then say the direct product $G\times H$ is MIS-normal
(maximum-independent-set-normal). Furthermore, the direct products
$G_{1}\times G_{2}\times\cdots\times G_{n}$ is said to be MIS-normal if every
maximum independent set of it is the preimage of an independent set of one
factor under projections.
About these two problems, there are some progresses have been made for some
very special vertex-transitive graphs.
Let $n,r$ and $t$ be three integers with $n\geq r\geq t\geq 1$. The graph
$K(t,r,n)$ is defined by: whose vertices set is the set of all $r$-element
subsets of $[n]=\\{1,2,\ldots,n\\}$, and $A$ and $B$ of which are adjacent if
and only if $|A\cap B|<t$. If $n\geq 2r$, then $K(1,r,n)$ is the well-known
Kneser graph. The classical Erdős-Ko-Rado Theorem [8] states that
$\alpha(K(1,r,n))=\binom{n-1}{r-1}$ (where $n\geq 2r$), and Frankl [9] first
investigated the independence number of the direct products of Kneser graphs.
Subsequently, Ahlswede, Aydinian and Khachatrian investigated the general case
[2].
###### Theorem 1.3
Let $n_{i}\geq r_{i}\geq t_{i}$ for $i=1,2,\ldots,k$.
(i) (Frankl [9]) if $t_{1}=\cdots=t_{k}=1$ and
$\frac{r_{i}}{n_{i}}\geq\frac{1}{2}$ for $i=1,2,\ldots,k$, then
$\alpha\left(\prod_{1\leq i\leq
k}K(1,r_{i},n_{i})\right)=\max\left\\{\frac{r_{1}}{n_{1}},\frac{r_{2}}{n_{2}},\ldots,\frac{r_{k}}{n_{k}}\right\\}\prod_{1\leq
i\leq k}|K(1,r_{i},n_{i})|.$
(ii) (Ahlswede, Aydinian and Khachatrian [2])
$\alpha\left(\prod_{1\leq i\leq
k}K(t_{i},r_{i},n_{i})\right)=\max\left\\{\frac{\alpha(K(t_{i},r_{i},n_{i}))}{|K(t_{i},r_{i},n_{i})|}:1\leq
i\leq k\right\\}\prod_{1\leq i\leq k}|K(t_{i},r_{i},n_{i})|.$
The circular graph $Circ(r,n)$ ($n\geq 2r$) is defined by:
$V(Circ(r,n))=\mathbb{Z}_{n}=\\{0,1,\ldots,n-1\\}$
and
$E(Circ(r,n))=\left\\{(i,j):|i-j|\in\\{r,r+1,\ldots,n-r\\}\right\\}.$
It is well known that $\alpha(Circ(r,n))=r$. Mario and Juan [16] determined
the independence number of the direct products of circular graphs.
###### Theorem 1.4
(Mario and Juan [16]) Let $n_{i}\geq 2r_{i}$ for $i=1,2\ldots,k$. Then
$\alpha\left(\prod_{1\leq i\leq
k}Circ(r_{i},n_{i})\right)=\max\left\\{\frac{r_{1}}{n_{1}},\frac{r_{2}}{n_{2}},\ldots,\frac{r_{k}}{n_{k}}\right\\}\prod_{1\leq
i\leq k}n_{i}.$
For positive integers $n$, let $S_{n}$ denote the permutation group on $[n]$.
Two permutations $f$ and $g$ are said to be intersecting if there exists an
$i\in[n]$ such that $f(i)=g(i)$. We define a graph on $S_{n}$ as that two
permutations are adjacent if and only if they are not intersecting. For
brevity, this graph is also denoted by $S_{n}$. Deza and Frankl [7] first
obtained that $\alpha(S_{n})=(n-1)!$. Cameron and Ku [6] proved that each
maximum independent set of $S_{n}$ is a coset of the stabilizer of a point, to
which Larose and Malvenuto [14], Wang and Zhang [18] and Godsil and Meagher
[10] gave alternative proofs, respectively. Recently, Cheng and Wong [11]
further investigated the independence number and the MIS-normality of the
direct products of $S_{n}$.
###### Theorem 1.5
(Cheng and Wong[11]) Let $2\leq n_{1}=\cdots=n_{p}<n_{p+1}\leq\ldots,n_{q}$,
$1\leq p\leq q$. Then
$\alpha\big{(}S_{n_{1}}\times S_{n_{2}}\times\cdots\times
S_{n_{q}}\big{)}=(n_{1}-1)!\prod_{2\leq i\leq q}n_{i}!,$
and the direct products $S_{n_{1}}\times S_{n_{2}}\times\cdots\times
S_{n_{q}}$ is MIS-normal except for the following cases:
1. (i)
$n_{1}=\cdots=n_{p}<n_{p+1}=3\leq n_{p+2}\leq\cdots\leq n_{q}$;
2. (ii)
$n_{1}=n_{2}=3\leq n_{3}\leq\cdots\leq n_{q}$;
3. (iii)
$n_{1}=n_{2}=n_{3}\leq n_{4}\leq\cdots\leq n_{q}$.
In [15], Larose and Tardif investigated the relationship between projectivity
and the structure of maximum independent sets in powers of some vertex-
transitive graphs, and obtained the MIS-normality of the powers of Kneser
graphs and circular graphs.
###### Theorem 1.6
(Larose and Tardif [15]) Let $n$ and $r$ be two positive integers. If $n>2r$,
then both $K^{k}(1,r,n)$ and $Circ^{k}(r,n)$ are MIS-normal for all positive
integer $k$.
Besides the above results, Larose and Tardif [15] prove that if $G$ is vertex-
transitive, then $\alpha(G^{n})=\alpha(G)|V(G)|^{n-1}$ for all $n>1$. They
also ask whether or not $G^{n}$ is MIS-normal if $G^{2}$ is MIS-normal.
Recently, Ku and Mcmillan [12] gave an affirmative answer to this problem, and
we solved this problem in a more general setting [20].
In this paper we shall solve both Problem 1.1 and Problem 1.2. To state our
results we need to introduce some notations and notions.
For a graph $G$, let $I(G)$ denote the set of all maximum independent sets of
$G$. Given a subset $A$ of $V(G)$, we define
$N_{G}(A)=\\{b\in V(G):\mbox{$(a,b)\in E(G)$ for some $a\in A$}\\}$
$N_{G}[A]=N_{G}(A)\cup A\mbox{ and }\bar{N}_{G}[A]=V(G)-N_{G}[A].$
If $G$ is clear from the context, for simplicity, we will omit the index $G$.
In [20], by the so-called “No-Homomorphism” lemma of Albertson and Collins [1]
we proved the following result.
###### Proposition 1.7
([20]) Let $G$ be a vertex-transitive graph. Then, for every independent set
$A$ of $G$, $\frac{|A|}{|N_{G}[A]|}\leq\frac{\alpha(G)}{|V(G)|}$. Equality
implies that $|S\cap N_{G}[A]|=|A|$ for every $S\in I(G)$, and in particularly
$A\subseteq S$ for some $S\in I(G)$.
An independent set $A$ in $G$ is said to be imprimitive if $|A|<\alpha(G)$ and
$\frac{|A|}{|N[A]|}=\frac{\alpha(G)}{|V(G)|}$. And $G$ is called IS-
imprimitive if $G$ has an imprimitive independent set. In any other cases, $G$
is called _IS-primitive_. From definition we see that a disconnected vertex-
transitive graph $G$ is IS-imprimitive and hence an IS-primitive vertex-
transitive graph $G$ is connected.
The following Theorem is the main result of this paper.
###### Theorem 1.8
Let $G$ and $H$ be two vertex-transitive graphs with
$\frac{\alpha(G)}{|G|}\geq\frac{\alpha(H)}{|H|}$. Then
$\alpha(G\times H)=\alpha(G)|H|,$
and either:
1. (i)
$G\times H$ is MIS-normal, or
2. (ii)
$\frac{\alpha(G)}{|G|}=\frac{\alpha(H)}{|H|}$ and one of them is IS-
imprimitive, or
3. (iii)
$\frac{\alpha(G)}{|G|}>\frac{\alpha(H)}{|H|}$ and $H$ is disconnected.
We leave the proof of Theorem 1.8 to the next section, while in Section 3, we
discuss the MIS-normality of the direct products of more than two vertex-
transitive graphs.
## 2 Proof of Theorem 1.8
Let $S$ be a maximum independent set of $G\times H$. Then
$|S|\geq\alpha(G)|H|\geq|G|\alpha(H)$. We now prove $\alpha(G\times
H)\leq\alpha(G)|H|$.
For every $a\in G$, define
$X_{a}=\\{x\in H:(a,x)\in S\\}.$
Since $S$ is an independent set of $G\times H$, for each $x\in X_{a}$ and
$y\in X_{b}$, $(x,y)\not\in E(H)$ whenever $(a,b)\in E(G)$. In this case, we
say that $X_{a}$ and $X_{b}$ are cross-independent. This concept is equivalent
to cross-intersecting families in extremal set theory. We refer [19] for
details.
In the language of cross-intersecting families, Borg [3, 4, 5] introduce a
decomposition of $X_{a}$ as follows.
$X^{*}_{a}=\\{x\in X_{a}:N_{H}(x)\cap X_{a}=\emptyset\\},$
$X^{\prime}_{a}=\\{x\in X_{a}:N_{H}(x)\cap X_{a}\neq\emptyset\\}$
and
$X^{\prime}=\bigcup_{a\in V(G)}X^{\prime}_{a}.$
Clearly, $X_{a}^{*}$ is an independent set of $H$ for every $a\in V(G)$, and
$|S|=\sum_{a\in V(G)}|X_{a}|$. Here, the empty set is regarded as an
independent set.
We list all distinct $X^{*}_{a}$’s as $Y_{1},Y_{2},\ldots,Y_{k}$, and define
$B_{i}=\\{a\in V(G):X^{*}_{a}=Y_{i}\\},\ i=1,2,\ldots,k.$
We then obtain a partition of $V(G)$ as $V(G)=B_{1}\cup B_{2}\cup\cdots\cup
B_{k}$. Then
$\displaystyle|S|$ $\displaystyle=$ $\displaystyle\sum_{a\in
V(G)}|X_{a}|=\sum_{a\in V(G)}(|X_{a}^{*}|+|X^{\prime}_{a}|)=\sum_{i=1}^{k}\
\sum_{a\in B_{i}}|X_{a}^{*}|+\sum_{a\in V(G)}|X_{a}^{\prime}|$ (1)
$\displaystyle=$ $\displaystyle\sum_{i=1}^{k}|Y_{i}||B_{i}|+\sum_{x\in
X^{\prime}}|A_{x}|,$
where
$A_{x}=\\{a\in V(G):x\in X^{\prime}_{a}\\}.$
For every pair $a,b\in V(G)$, it is easy to verify that $(a,b)\not\in E(G)$ if
$X^{\prime}_{a}\cap X^{\prime}_{b}\neq\emptyset$. Therefore, $A_{x}$ is an
independent set of $G$. By Proposition 1.7 we have that
$|A_{x}|\leq\frac{\alpha(G)}{|V(G)|}|N_{G}[A_{x}]|,$ (2)
and equality holds if and only if $|A_{x}|=0$, or $|A_{x}|=\alpha(G)$, or
$A_{x}$ is an imprimitive independent set of $G$.
Suppose $x\in N_{H}[Y_{i}]=N_{H}(Y_{i})\cup Y_{i}$. If $x\in N_{H}(Y_{i})$,
then there exists $y\in Y_{i}$ such that $(x,y)\in E(H)$ and
$\\{(a,x),(b,y)\\}\subset S$ for any $b\in B_{i}$ and $a\in A_{x}$, hence
$(a,b)\not\in E(G)$ since $S$ is an independent set; if $x\in Y_{i}$, then for
each $a\in A_{x}$, there is a $z\in X_{a}$ with $(x,z)\in E(H)$ and
$\\{(a,z),(b,x)\\}\subset S$, yielding $(a,b)\not\in E(G)$. Thus proving that
$B_{i}\subseteq\bar{N}_{G}[A_{x}]$ if $x\in N_{H}[Y_{i}]$. From this it
follows that
$\sum_{i:x\in
N_{H}[Y_{i}]}|B_{i}|\leq|\bar{N}_{G}[A_{x}]|=|V(G)|-|N_{G}[A_{x}]|,$
i.e.,
$|N_{G}[A_{x}]|\leq|V(G)|-\sum_{i:x\in
N_{H}[Y_{i}]}|B_{i}|=\sum_{i:x\in\bar{N}_{H}[Y_{i}]}|B_{i}|.$ (3)
Note that
$X^{\prime}\subseteq\bigcup_{i=1}^{k}\bar{N}_{H}[Y_{i}].$ (4)
Together with (2), (3)and (4), we then obtain that
$\displaystyle\sum_{x\in
X^{\prime}}|A_{x}|\leq\frac{\alpha(G)}{|V(G)|}\sum_{x\in
X^{\prime}}\sum_{i:x\in\bar{N}_{H}[Y_{i}]}|B_{i}|$ (5) $\displaystyle\leq$
$\displaystyle\frac{\alpha(G)}{|V(G)|}\sum_{i=1}^{k}\sum_{x\in\bar{N}_{H}[Y_{i}]}|B_{i}|=\frac{\alpha(G)}{|V(G)|}\sum_{i=1}^{k}|B_{i}||\bar{N}_{H}[Y_{i}]|.$
Combining (1) and (5) gives that
$\displaystyle|S|$ $\displaystyle=$
$\displaystyle\sum_{i=1}^{k}|Y_{i}||B_{i}|+\sum_{x\in X^{\prime}}|A_{x}|$
$\displaystyle\leq$
$\displaystyle\sum_{i=1}^{k}|Y_{i}||B_{i}|+\frac{\alpha(G)}{|V(G)|}\sum_{i=1}^{k}|B_{i}||\bar{N}_{H}[Y_{i}]|$
$\displaystyle=$
$\displaystyle\sum_{i=1}^{k}|B_{i}|\left(\frac{\alpha(G)}{|V(G)|}|H|+|Y_{i}|-\frac{\alpha(G)}{|V(G)|}|N_{H}[Y_{i}]|\right)$
$\displaystyle=$
$\displaystyle\alpha(G)|H|+\sum_{i=1}^{k}|B_{i}|\left(|Y_{i}|-\frac{\alpha(G)}{|V(G)|}|N_{H}[Y_{i}]|\right)$
$\displaystyle\leq$ $\displaystyle\alpha(G)|H|.$
The last inequality follows from that
$\displaystyle|Y_{i}|-\frac{\alpha(G)}{|G|}|N_{H}[Y_{i}]|\leq|Y_{i}|-\frac{\alpha(H)}{|V(H)|}|N_{H}[Y_{i}]|\leq
0,$ (6)
by Proposition 1.7.
The maximum of $|S|$ implies that $|S|=\alpha(G)|H|$, from which it follows
that equalities (2), (3), (4) and (6) hold. Also, from Proposition 1.7,
equality (6) means that either $Y_{i}=\emptyset$, or
$\frac{\alpha(G)}{|G|}=\frac{\alpha(H)}{|V(H)|}$ and $Y_{i}$ is either
imprimitive or a maximum independent set of $H$ for $i=1,2,\ldots,k$.
We now prove that either $S$ is the preimages of projections of a maximum
independent set of $G$ or $H$, or (ii) or (iii) holds. There are two cases to
be considered.
Case 1: $\frac{\alpha(G)}{|G|}>\frac{\alpha(H)}{|H|}$. Then, equality (6)
means that $Y_{i}=\emptyset$ for all $i$, and so $X^{\prime}=V(H)$ by equality
(4). Hence, from equality (2) it follows that $A_{x}$ is a maximum independent
set of $G$ for all $x\in V(H)$. With this assumption we have that for any
$x,y\in V(H)$ with $(x,y)\in E(H)$, if $A_{x}\neq A_{y}$, there must exist
$a\in A_{x}$ and $b\in A_{y}$ with $(a,b)\in E(G)$ since both $A_{x}$ and
$A_{y}$ are maximum independent set, so $[(a,x),(b,y)]\in E(G\times H)$,
contradicting $\\{(a,x),(b,y)\\}\subset S$. Therefore, $A_{x}=A_{y}$ whenever
$(x,y)\in E(H)$, which implies that $S$ is the preimage of a maximum
independent set of $G$ under projections if $H$ is connected.
Case 2: $\frac{\alpha(G)}{|G|}=\frac{\alpha(H)}{|H|}$. Then, equality (6)
means that either $|Y_{i}|=0$ or $\alpha(H)$, or $Y_{i}$ is an imprimitive
independent set of $H$ for each index $i$. If $Y_{i}$ is an imprimitive
independent set of $H$ for some $i$, then $H$ is IS-imprimitive. If
$|Y_{i}|=\alpha(H)$ for all $i$, then $X_{a}=X_{a}^{*}$ is a maximum
independent set of $H$ for all $a\in V(G)$, and we can prove in the similar
way as in Case 1 that $S$ is the preimage of a maximum independent set of $H$
under projections if $G$ is connected. We now suppose that $|Y_{i}|=0$ for
some $i$. With this assumption, then equality (4) implies $X^{\prime}=V(H)$,
and then equality (3) means that either $A_{x}$ is either imprimitive or a
maximum independent set of $G$ for all $x\in V(H)$. If the former holds for
some $x\in V(H)$, we have that $H$ is IS-imprimitive; otherwise, the latter
holds for all $x\in V(H)$, and then we can prove in the similar way as in Case
1 that $S$ is the preimage of a maximum independent set of $G$ under
projections if $H$ is connected.
## 3 Concluding Remark.
Let $G_{1},G_{2},\ldots,G_{n}$ be $n$ non-empty vertex-transitive graphs, and
set $G=G_{1}\times G_{2}\times\cdots\times G_{n}$. From Theorem 1.8 it
immediately follows that
$\alpha(G)=\alpha(G_{1})\prod_{2\leq i\leq n}|G_{i}|.$
We now discuss the MIS-normality of $G$. For convenience, we say $G$ is MIS-
normal if $n=1$.
A graph $H$ is said to be non-empty if $E(H)\neq\emptyset$. It is well known
that if $H$ is a non-empty vertex-transitive graph, then
$\frac{\alpha(H)}{|H|}\leq\frac{1}{2}$, and equality holds if and only if $H$
is a bipartite graph.
Without loss of generality we may assume that
$\frac{1}{2}\geq\frac{\alpha(G_{1})}{|G_{1}|}=\cdots=\frac{\alpha(G_{\ell})}{|G_{\ell}|}>\frac{\alpha(G_{\ell+1})}{|G_{\ell+1}|}\geq\cdots\geq\frac{\alpha(G_{n})}{|G_{n}|}$,
and write $H_{0}=G_{1}\times\cdots\times G_{\ell}$ and $H_{i}=H_{i-1}\times
G_{\ell+i}$ for $i=1,\ldots,n-\ell$ subject to $n>\ell$. Then $G=H_{n-\ell}$
and with
$\frac{\alpha(H_{i-1})}{|H_{i-1}|}>\frac{\alpha(G_{\ell+i})}{|G_{\ell+i}|}$
for $i\geq 1$.
###### Proposition 3.1
Suppose $n>\ell$. Then $G$ is MIS-normal if and only if $H_{0}$ is MIS-normal
and $G_{\ell+1},\ldots,G_{n}$ are all connected.
Proof. Since $\alpha(G)=\alpha(H_{0})\prod_{i=\ell+1}^{n}|G_{i}|$, we have
that if $H_{0}$ is not MIS-normal, then $G$ is not MIS-normal. Furthermore, if
$G_{i}$ is not connected for for some $i\geq 1$, writing
$G_{i}=G_{i}^{\prime}\cup G_{i}^{\prime\prime}$, a union of disjoint
subgraphs, then, for all $I_{1},I_{2}\in I(H_{i-1})$ with $I_{1}\neq I_{2}$,
it is clear that $S=(I_{1}\times G_{i}^{\prime})\cup(I_{2}\times
G_{i}^{\prime\prime})\in I(H_{i})$, which is not a preimage of any independent
set of one factor under projections, i.e., $H_{i}$ is not MIS-normal, hence
$G$ is not MIS-normal.
Conversely, suppose $H_{0}$ is MIS-normal, and $G_{\ell+i}$ is connected for
$i\geq 1$. Since
$\frac{\alpha(H_{i-1})}{|H_{i-1}|}>\frac{\alpha(G_{\ell+i})}{|G_{\ell+i}|}$,
Theorem 1.8 implies that each maximal-sized independent set is of the form
$S\times G_{\ell+i}$, where $S\in I(H_{i-1})$, which means that $H_{i}$ is
MIS-normal for $i\geq 1$. We thus prove that $G$ is MIS-normal. ∎
We now discuss the case $n=\ell$, that is, each $G_{i}$ has the identical
independence ratio. To deal with this case we need a lemma as follows.
###### Lemma 3.2
Suppose that G is a vertex-transitive bipartite graph. Then $G$ is imprimitive
if and only if $G$ is disconnected.
Proof. It is clear that $G$ is imprimitive if $G$ is disconnected. On the
converse, if $G$ is imprimitive, then there is an imprimitive independent set
$A$ such that $\frac{|A|}{|N_{G}[A]|}=\frac{\alpha(G)}{|G|}=\frac{1}{2}.$ Set
$B=N_{G}(A)$. $|B|=|A|$ and $A\subseteq N_{G}(B)$ is clearly. If $N_{G}(B)\neq
A$, then we obtain that $\sum_{u\in A}d(u)\leq\sum_{v\in B}d(v)$, which
induces a contradiction. Hence $N_{G}(B)=A$, that is to say $G$ is
disconnected.
###### Proposition 3.3
Suppose that
$\frac{\alpha(G_{1})}{|G_{1}|}=\cdots=\frac{\alpha(G_{n})}{|G_{n}|}=\frac{\alpha(G)}{|G|}$.
Then $G$ is MIS-normal if and only if one of the following holds.
(i) $\frac{\alpha(G)}{|G|}<\frac{1}{2}$ and every $G_{i}$ is IS-primitive.
(ii) $\frac{\alpha(G_{1})}{|G_{1}|}=\frac{1}{2}$, $n=2$ and both $G_{1}$ and
$G_{2}$ are connected.
Proof. For $1\leq i\leq n$, set $\hat{G}_{i}=G_{1}\times\cdots\times
G_{i-1}\times G_{i+1}\times\cdots\times G_{n}$. Then $G=\hat{G}_{i}\times
G_{i}$ for $i=1,2,\ldots,n$. If $G_{i}$ is imprimitive, letting $A_{i}$ be an
imprimitive independent set of $G_{i}$, for every $I\in I(\hat{G}_{i})$, it is
easy to see that $S=(\hat{G}_{i}\times
A_{i})\cup(I\times\bar{N}_{G_{i}}[A_{i}])\in I(G)$, which is not a preimage of
any independent set of $\hat{G}_{i}$ or $G_{i}$ under projections, therefore,
$G$ is not MIS-normal. Conversely, if both $\hat{G}_{i}$ and $G_{i}$ are IS-
primitive, Theorem 1.8 implies that $G$ is MIS-normal. It remains to check
when $\hat{G}_{i}$ is IS-primitive. Summing up the above, $G$ is MIS-normal if
and only if both $\hat{G}_{i}$ and $G_{i}$ are IS-primitive. To complete the
proof, it remains to check when $\hat{G}_{i}$ is IS-primitive. We distinguish
two cases.
Case (i):$\frac{\alpha(G)}{|G|}<\frac{1}{2}$. In this case, Theorem 2.6 in
[20] says that if $G$ is MIS-normal, then both $\hat{G}_{i}$ and $G_{i}$ are
IS-primitive. The induction implies (i).
Case (ii): $\frac{\alpha(G_{1})}{|G_{1}|}=\frac{1}{2}$, i.e., every $G_{i}$ is
bipartite. From Lemma 3.2 it follows that $\hat{G}_{i}$ and $G_{i}$ is IS-
primitive if and only if both $\hat{G}_{i}$ and $G_{i}$ are connected.
However, it is well known that $\hat{G}_{i}$ is disconnected if $n>2$, thus
proving (ii). ∎
Combining Proposition 3.1 and Proposition 3.3 gives the following theorem.
###### Theorem 3.4
Let $G_{1},G_{2},\ldots,G_{n}$ be connected vertex-transitive graphs with
$\frac{1}{2}\geq\frac{\alpha(G_{1})}{|G_{1}|}=\cdots=\frac{\alpha(G_{\ell})}{|G_{\ell}|}>\frac{\alpha(G_{\ell+1})}{|G_{\ell+1}|}\geq\cdots\geq\frac{\alpha(G_{n})}{|G_{n}|}$,
where $n\geq 2$ and $1\leq\ell\leq n$. Then $G_{1}\times
G_{2}\times\cdots\times G_{n}$ is MIS-normal if and only if one of the
following holds:
(i) $\frac{\alpha(G_{1})}{|G_{1}|}<\frac{1}{2}$ and
$G_{1},G_{2},\ldots,G_{\ell}$ are all IS-primitive whenever $\ell>1$.
(ii) $\frac{\alpha(G_{1})}{|G_{1}|}=\frac{1}{2}$ and $\ell\leq 2$.
Acknowledgement The author is greatly indebted to Professor J. Wang for giving
useful comments, suggestions and helps that have considerably improved the
manuscript.
## References
* [1] M.O. Albertson and K.L. Collins, Homomorphisms of $3$-chromatic graphs, Discrete Math., 54 (1985) 127-132.
* [2] R. Ahlswede, H. Aydinian and L.H. Khachatrian, The Intersection Theorem for Direct Products, European J. Combin., 19 (1998) 649-661.
* [3] P. Borg, A short proof of a cross-intersection theorem of Hilton, Discrete Math., 309 (2009) 4750-4753.
* [4] P. Borg, Cross-intersecting families of permutations, J. Combin. Theory Ser. A, 117 (2010) 483-487.
* [5] P. Borg and I. Leader, Multiple cross-intersecting families of signed sets, J. Combin. Theory Ser. A, 117 (2010) 583-588.
* [6] P.J. Cameron and C.Y. Ku, Intersecting families of permutations, European J. Combin., 24 (2003) 881-890.
* [7] M. Deza and P. Frankl, On the maximum number of permutations with given maximal or minimal distance, J. Combin. Theory Ser. A, 22 (1977) 352-362.
* [8] P. Erdős, C. Ko and R. Rado, Intersection theorems for systems of finite sets, Quart. J. Math. Oxford Ser., 2 (12) (1961) 313-318.
* [9] P. Frankl, An Erdős-Ko-Rado Theorem for direct products, European J. Combin., 17 (1996) 727-730.
* [10] C. Godsil and K. Meagher, A new proof of the Erdős-Ko-Rado theorem for intersecting families of permutations, Eurpean J. Combin., 30 (2008) 404-414.
* [11] C.Y. Ku and T.W.H. Wong, Intersecting families in the alternating group and direct product of symmetric groups, Electron. J. Combin., 14 (2007).
* [12] C.Y. Ku and B.B. Mcmillan, Independent sets of maximal size in tensor powers of vertex-transitive graphs, J. Graph Theory, 60 (2009) 295-301.
* [13] P.K. Jha and S. Klavz̆ar, Independence in direct-product graphs, Ars Combin., 50 (1998) 53-60.
* [14] B. Larose and C. Malvenuto, Stable sets of maximal size in Kneser-type graphs, European J. Combin., 25 (2004) 657-673.
* [15] B. Larose and C. Tardif, Projectivity and independent sets in powers of graph, J. Graph Theory, 40 (2002) 162-171.
* [16] V.P. Mario and V. Juan, Independence and coloring properties of direct products of some vertex-transitive graphs, Discrete Math., 306 (2006) 2275-2281.
* [17] C. Tardif, Graph products and the chromatic difference sequence of vertex-transitive graphs, Discrete Math., 185 (1998) 193-200.
* [18] J. Wang and S.J. Zhang, An Erdős-Ko-Rado-Type Theorem in Coxeter Groups, Eurpean J. Combin., 29 (2008) 1112-1115.
* [19] J. Wang and H.J. Zhang, Cross-intersecting families and primitivity of symmetric systems, submitted.
* [20] H.J. Zhang, Primitivity and independent sets in direct products of vertex-transitive graphs, J. Graph Theory, to appear.
|
# How will electric vehicles affect traffic congestion and energy consumption:
an integrated modelling approach
Artur Grigorev1, Tuo Mao1, Adam Berry1, Joachim Tan2, Loki Purushothaman1,
Adriana-Simona Mihaita1 1 Most authors are with the University of Technology
in Sydney, Faculty of Engineering and IT, School of Computer Science: 61
Broadway Ultimo, NSW, Australia. Corresponding authors contact:
<EMAIL_ADDRESS><EMAIL_ADDRESS>This
work was sponsored by AEMO the Australian Energy Market Operator. The authors
are highly grateful for the guidance and support of Joachim Tan, Adrian
Grantham, Greg Staib and Andrew Turley. Joachim Tan is a Principal Manager in
the Australian Energy Market Operators and main lead of the UTS AEMO project
collaboration.3 The authors are highly grateful for the guidance and data
support of Transport for NSW, Australia, more specifically Virginie Vernin,
Bill Chen, Blake Xu and Marwan Daizli.
###### Abstract
This paper explores the impact of electric vehicles (EVs) on traffic
congestion and energy consumption by proposing an integrated bi-level
framework comprising of: a) a dynamic micro-scale traffic simulation suitable
for modelling current and hypothetical traffic and charging demand scenarios
and b) a queue model for capturing the impact of fast charging station use,
informed by traffic flows, travel distances, availability of charging
infrastructure and estimated vehicle battery state of charge. To the best of
our knowledge, this paper represents the first integrated analysis of
potential traffic congestion and energy infrastructure impacts linked to EV
uptake, based on real traffic flows and the placement and design of existing
fast-charging infrastructure. Results showcase that the integrated queue-
energy-transport modelling framework can predict correctly the limitations of
the EV infrastructure as well as the traffic congestion evolution. The
modelling approach identifies concrete pain points to be addressed in both
traffic and energy management and planning. The code for this project can be
found at : https://github.com/Future-Mobility-Lab/EV-charging-impact
###### Index Terms:
electric vehicles, traffic simulation modelling, queue modelling, recharging
scenario evaluation, fast charge impact modelling
## I Motivation
The adoption of electric vehicles (EVs) at large scale around the globe is
largely dependent on two key factors: charging accessibility for its consumers
and availability of sufficient electrical grid capacity to support vehicle
charging, particularly where fast charging infrastructure is necessary. Under
ideal conditions for the consumer, access to an EV charging station should be
just as convenient as current access to a petrol station. Several studies have
been conducted in recent years to model and identify suitable locations for
such charging stations [2], [3]. The majority of these studies are based on
the behavioural pattern of combustion engine (CE) drivers. However, the
behaviour of electric vehicle (EV) drivers is significantly different in terms
of location selection and their waiting time to charge [2], [4]. Multiple
factors can influence the decision of EV drivers to choose a specific location
to charge, among which we cite: the travel time towards the closest EV
station, the state of charge of the battery (SoC) while driving, the waiting
time in queue at the EV station, the total charging time that a driver needs
to plan different activities around while the car is charging. The
stochasticity of this charging behaviour makes the modelling of EVs a hard
problem to solve, which may delay evidence-based infrastructure planning at a
government level and reduce investment confidence for commercial
infrastructure and electric vehicle businesses. Therefore, the inability to
charge conveniently in the traffic network coupled with limited charging stops
makes the problem challenging from both operational and planning perspectives.
Even more challenging, evaluating the grid impact of EV adoption comes along
not only with a consumer random behaviour affected as well by traffic
congestion variation, but with the physical constraints of both infrastructure
availability and maximal energy capacity of each serviced area.
These challenges represent the motivation of this study which to the best of
our knowledge is the first to propose an integrated bi-level framework
comprised of a dynamic traffic simulation modelling coupled with queueing
behaviour modelling for estimating the impact of traffic demand and congestion
on EV fast charging behaviour and vice-versa.
### I-A Related works
Figure 1: Bi-level framework of transport and energy modelling containing: A)
Upper Level focused on a hybrid transport and queue modelling approach for
quantifying EV charging on traffic congestion and vice-versa and B) Lower
level focus on building a home-and-destination charging model using substation
capacity data and the daily and dynamic energy load data in the study area.
Several studies [5]-[6], have been conducted to forecast the electricity
demand that will be created by the penetration of EVs into the market in the
future. There are also works to determine and forecast the impact this
electricity demand will have on the existing power grids [7]. However, the
majority of these studies either orient themselves towards analysing the
impact of the electricity consumption on the grid using hypothetical
assumptions of the charging locations or very broad estimations of the major
attraction points in specific areas (p-median or location-allocation problem)
such as shopping malls, schools, hospitals and workplaces [8]. Alternatively,
some studies that have considered the user equilibrium (UE) approach to
optimise the route choice so that the travel time is minimized when recharging
[9], [10]. However, these studies have several assumptions such as that
congestion will not affect the electricity consumption or that the path
lengths are within battery driving range. However, congestion represents an
important factor to be taken into consideration in the EV impact modelling,
especially when the fast charging stations reach their plug-in capacity. Some
studies have developed queuing models that take into consideration the station
capacity, the charging time and EV waiting time for example [11]; others have
considered as well the maximal charging demand which satisfies the user’s
maximal tolerable queuing time [12] or have proposed a global queue system
that would dispatch EVs to available charging stations [13]. While attention
was given to parameters describing the charging situations inside the EV
stations, studies did not take into consideration for example the battery
state of charge (SoC) when routing decisions are being made, nor the travel
time affected by increased demand. Other queuing approaches can be found in
[14] where they took consideration of the fact that charging stations can have
a different number of plugs with various charging capabilities.
In [15] there is a tentative approach to couple together simulation and
optimisation techniques for optimising a taxi fleet and similar approaches
have been proposed which consider for example the planning of routes according
to an activity-based model [16] or agent-based simulation [17]. We believe,
however, that there remains a key gap in the use of traffic simulation models
to explore the interrelationship between charging activity, traffic behaviour
and road congestion outside of strictly fleet-based settings. Similarly,
embedding EV queue modelling at each station in the simulation model remains
an under-explored research topic.
From the above discussions, we identify several gaps in the current
literature. Firstly, several studies do not take into consideration future
vehicle demand increases and how this may impact EV fast charging
accessibility, queuing times, traffic volumes and emergent congestion related
to the use of fast charging infrastructure. The inter-dependency between
congestion, demand and queue spill-back from EV charging stations should be
modelled together for both current and future traffic scenarios. Secondly, the
number of available plug-in slots and their charging rates should be embedded
in the queue model and constraints should be applied around realistic charging
queue lengths; this is to avoid assumptions of infinite queue models which can
accept all incoming EV requests. Thirdly, several studies propose an
optimisation of charging stations based on assumption that people charge
either at home or at their destination, whereas in reality, the charging
behaviour is dynamic and can include mid-route charging stops as well based on
battery SoC, charging availability and convenience (i.e. charging at a
shopping centre); this is a factor that would need to be embedded in any queue
modelling approach. Finally, analysis of electric vehicle behaviour should
include the electrical grid impact, as the shape and extent of charging demand
will inevitably impact electrical network infrastructure planning and
operation, particularly where charging demand is aligned with peak electrical
demand.
### I-B Paper Contributions
In our work we aim to address previous gaps by proposing an integrated
framework of traffic modelling with stochastic queue behaviour (the upper-
level which can be used for en-route charging behavioural modelling showcased
in this paper) and a data-driven energy consumption estimation (the lower-
level for home and work charging estimation which we will present in a future
paper extension). The approach has been applied on the Northern Sydney urban
area and is built using realistic data sets in terms of: traffic demand,
vehicle counting, EV charging stations (locations, number of plugs, type of
plugs, etc.), electricity zone substation capacity data and daily electricity
load profiles for those substations (sufficient for determining peak
electricity demand times).
Firstly, we propose an integrated traffic simulation and queue behavioural
model for capturing not only the dynamic nature of traffic congestion but also
the queue impact around existing EV charging stops. The microscopic traffic
simulation is calibrated against real-world traffic flow counts and is used to
evaluate traffic congestion under different EV and charging station settings.
We also explore the potential impact of hypothetical future demand scenarios.
The queue model which works on a first-in-first-out approach is tailored for
each station and ingests several simulation outputs (travel time to reach the
station, calibrated flow, travel distance, calibrated OD demand), as well as
the SoC of the batteries, the number of available charging plugs and their
charging rates. We denote this model as
$EV-Q(M/M/n_{ep}^{S_{i}}/N_{c}^{S_{i}})$ where $N_{c}^{S_{i}}$ is the maximum
number of car slots available at each station (for both waiting and charging)
and $n_{ep}^{S_{i}}$ is the number of plugs (chargers) for each station. The
EV-Q model generates queue length estimations which are then fed to the micro
traffic simulation model for congestion impact analysis. Secondly, we explore
the behaviour based on both observed road demand from 2016 and two
hypothetical scenarios: one with a $15\%$ increase in road use and another
with a $30\%$ increase. Across all scenarios, we explore impacts of different
levels of assumed EV penetration. Lastly, for all simulations, we estimate
total electrical demand associated with fast charging, enabling a review of
potential electrical infrastructure impacts.
The paper is organised as follows: Section II introduces the bi-level
framework for our integrated queue, energy and transport simulation modelling,
Section III presents a case study evaluation on the Northern Sydney area
including traffic simulation modelling (construction, calibration and
validation), queue modelling (construction and validation) and EV impact
evaluation across different penetration rates and demand scenarios. The last
section is reserved for conclusions, limitations of this work and further
directions.
## II Methodology
Figure 2: Network layout of the North Sydney traffic simulation model (NSTM)
including all existent electric vehicles charging stations (marked as red).
The bi-level framework we propose in this paper is illustrated in Fig. 1 and
is comprised of two major parts: A) the Upper level construction with the
purpose of building a hybrid model for analysing the quantified impact of EV
charging station availability on traffic congestion and vice-versa, and B) the
Lower level construction which focuses on building a home-and-destination
charging model using substation capacity data and the daily and dynamic energy
load data in the study area. Due to lack of space the current paper will only
include the upper level modelling and selected results. Each model comprises
of several modules coloured differently, with dedicated entry data streams and
output measures, summarized here below.
Figure 3: EV charging representation using a queue modelling approach of
vehicles arriving at each charging stations.
### II-A Transport simulation modelling
In order to evaluate the impact of current and future traffic demand on the
electric vehicle fast charging, one needs to be able to evaluate realistic
travel distances, travel times and traffic flows for the road network under
various dynamic traffic conditions. The traffic simulation we are building is
a microscopic dynamic model which is comprised of all vehicle movements, from
any internal origins to their destinations and intermediary charging stops
along the route.
We select North Sydney as an application area (see Fig. 2) and refer to our
simulation model as NSTM. All notations are provided in Table I. The area has
been selected due to its large business hubs, major train stations, new metro
line and major commuting trips to/from Sydney CBD on a daily basis. The NSTM
model consumes in its input: a) the network layout with real road structure
definition, b) OD demand matrices ($OD(T_{r})$), c) traffic signal control
definitions and d) the placement and characteristics of real-world EV charging
stations deployed in Northern Sydney ($S_{i},i\in\\{1,..25\\}$). Overall the
simulation model contains the daily trips of almost 100,000 people (according
to 2016 Australian Census data), 25 EV charging stations, 41 permanent traffic
count stations, 240 SCATS controlled intersections, 1,919 traffic flow
detectors and 2,750 road sections.
The micro-simulation model goes through several procedures of initialisation,
calibration and validation including: i) an initial static traffic assignment
(STA) procedure of the original traffic demand to available routes, ii) a
static OD adjustment (SOD) using permanent traffic flow counts, iii) a daily
demand profiling for AM and PM peaks, iv) a Dynamic OD adjustment (DOD) using
detector flow counts on an hourly basis for 24-hour time interval which helps
us to calibrate the initial demand matrix, v) a dynamic user equilibrium (DUE)
assignment which is validated against both permanent and detector flow counts,
and finally vi) a microscopic simulation run consuming on input as well the
estimated queue lengths on each road section in the vicinity of the EV
charging stops (this is an output of the EV-Q model described in the next sub-
section).
The major outputs of the micro NSTM model are the following: a) a calibrated
and more realistic OD demand matrix $OD_{c}(T_{r})$, b) the total travel times
for AM/PM peaks, c) the calibrated traffic flows and finally d) the average
travel distance (the last two used as inputs to the EV-Q charging model for
each station).
The NSTM is used for testing the impact of various scenarios: a baseline
scenario based on observed traffic demand for 2016, and two hypothetical
growth scenarios, with a $15\%$ and $30\%$ increase in total demand,
respectively. The scenarios allow us to explore how changes in traffic demand
interact with electric vehicle penetration rates and EV infrastructure
availability to effect road use, queuing times, congestion and electricity
demand.
### II-B Queue modelling of EV charging
While the traffic simulation model has the objective of providing dynamic
insights on the traffic congestion in the neighbourhood, it does not have the
capability of replicating the queueing behaviour of drivers that need EV
charging. Therefore, by using simulation outputs such as car average travel
distance and traffic flow on road sections, we develop a queue model to
quantify the impact of the waiting behaviour around existing EV stations and
feed this behaviour back to the simulation model. The traffic simulation and
the queue models are connected to each other and run in parallel consuming
each-other outputs. Fig. 3 presents the overall queue modelling framework
(EV-Q) which we further detail.
Table I: Notations in use for EV modelling.
Variable Definition $NSTM$ North Sydney Traffic Model, $NSTM_{c}$ North Sydney
Traffic Model calibrated, $T_{r}$ the 1h time interval during a 24-hour period
where $r\in\\{1...24\\}$; $OD(T_{r})$ the Origin-Destination matrix containing
the number of assigned car trips in the network in time period $T_{r}$
$OD_{c}(T_{r})$ the calibrated Origin-Destination matrix $Tf_{k}(T_{r})$ the
traffic flow for each road section $k$ in the network during time interval
$T_{r}$ $Avg_{d}(T_{r})$ the average travel distance of cars during $T_{r}$
$TT_{tot}^{e}(T_{r})$ the car total time during $T_{r}$ and specific $EV_{p}$
$N$ the total number of EV stations in the network, $S_{i},i\in\\{1,..,N\\}$
the EV station ID, $N_{c}^{S_{i}}$ the maximal number of cars inside a station
$EV_{p}$ the EV penetration rate in the network $\lambda^{S_{i}}$ mean-arrival
time of cars at an EV station $n_{ep}^{S_{i}}$ number of electric plugs per
station $B_{c}(T_{r})[kWh]$ battery size of arriving cars during $T_{r}$
$SoC^{S_{i}}$ the initial state of charge for vehicles arriving at station
$S_{i}$ $\rho^{S_{i}}$ the service density of a station $S_{i}$
$\overline{Q^{S_{i}}}(T_{r})$ (O1) Mean queue length of an EV station $S_{i}$
per hour $\overline{WT_{q}^{S_{i}}}(T_{r})$ (O2) Mean waiting time in queue at
an EV station $S_{i}$ $\mu_{t}^{S_{i}}(T_{r})$ (O3) Mean service time to
charge at an EV station $S_{i}$ $WT^{S_{i}}(T_{r})$ (O4) Total time spent
overall at an EV station $S_{i}$ $E^{S_{i}}(T_{r})$ (O5) Total energy
consumption of an EV station $S_{i}$ $MaxQ^{S_{i}}(T_{r})$ (O6) Maximum
recorded queue length of an EV station $S_{i}$ $MaxW_{q}^{S_{i}}(T_{r})$ (O7)
Maximum waiting time in queue at an EV station $S_{i}$ $MaxWT^{S_{i}}(T_{r})$
(O8) Maximum time spent overall at an EV station $S_{i}$ $MaxE^{S_{i}}(T_{r})$
(O9) Maximal energy consumption of an EV station $S_{i}$
Let $N_{c}(T_{r})$ be the total number of cars in the traffic network during
time interval $(T_{r})\in\\{0,1,2,...23h\\}$, $EV_{p}$ the percentage of
vehicles that will need to charge and $S$ the total number of EV charging
stations (25 in our study according to the real world set-up [18]). Each EV
station has a specific number of plugs (chargers) which we denote
$n_{ep}^{S_{i}}$ and which follows a real set-up arrangement in our network (7
stations have only 1 plug available, 3 have 4 plugs, 2 have 3 plugs, 10 have 2
plugs, 1 has 7 plugs and only 1 has 10 plugs available). We denote the queue
model as $EV-Q(M/M/n_{ep}^{S_{i}}/N_{c}^{S_{i}})$ where the first M stands for
EV arrivals following a Gaussian distribution with mean arrival time at each
station $\lambda^{S_{i}}$, the second M represents the distribution of the
charging times with $\mu^{S_{i}}$ mean-service time and $N_{c}^{S_{i}}$ is the
maximum number of car slots available at each station (for both waiting and
charging). For this paper, we follow the queue modelling procedures from [14,
19] which are using the Markov Chain theory for probability calculation of
charging at specific plugs of a station, with the assumption that the number
of waiting cars inside the stations are limited to a maximum threshold.
The cars arrive at each EV charging station following a normal distribution
with mean-inter-arrival (mIA) times denoted as $\lambda^{S_{i}}$ (this is
calculated from real data-streams of vehicles captured by the traffic count
detectors in the vicinity of each EV station in our network; mIA is calculated
hourly from traffic flow with a variation of $\pm 20\%$ following normal
distribution; therefore this is estimated based on simulation of traffic flow
at vicinity of the charging stations). For fairness of charging, the arriving
cars are being charged on a First-Come-First-Served (FCFS) approach to the
next available plug in the network (statistics on emergent queues are stored
in O1-Fig. 3). The service time (O3 in Fig. 3) is then calculated individually
for each arriving car by taking into consideration the car State of Charge
($SoC^{S_{i}}$) as follows:
$(O3):\mu^{S_{i}}(T_{r})=\frac{SoC^{S_{i}}(T_{r})*B_{c}(T_{r})}{Pp^{S_{i}}}$
(1)
where $SoC^{S_{i}}(T_{r})$ represents the state of charge for arriving cars,
$B_{c}(T_{r})$ is the battery size and $Pp^{S_{i}}$ is the charging rate of
the fast charger of a station $S_{i}$ (which follows the real EV station
information with the specification that majority have a 22kW capability [18]).
We assume that the initial State-of-Charge will follow a normal distribution
between $20\%-80\%$ according to efficient Li-Ion battery use [20]. Meanwhile
the battery size can take several values which have been used for sensitivity
and robustness analysis: $B_{c}(T_{r})[kWh]\in\\{50,58,66,74,82,90,98,106\\}$,
according to a 10% difference in battery size compared to Tesla Model 3
(Standard Range) [21]. Based on arrival and service times, one can then
express the EV charging density $\rho^{S_{i}}$ as:
$\rho^{S_{i}}(T_{r})=\lambda^{S_{i}}(T_{r})/\mu^{S_{i}}(T_{r})$ (2)
Based on the above formulations and the Markov Chain stationary conditions
regarding the matrix of transition probabilities $P$ and the stationary
distribution ($\vec{\pi}\cdot P=0$), we can first calculate the stationary
distribution (the probability that a vehicle will charge at a specific plug in
the EV station) in a recursive form as follows:
$\pi_{k}^{S_{i}}=\left\\{\begin{array}[]{ll}\frac{1}{k!}\left(\rho^{S_{i}}(T_{r})\right)^{k}\pi_{0}^{S_{i}}&\mbox{if
}\ k\leq{n_{ep}^{S_{i}}}\\\
\left(\frac{1}{n_{ep}^{S_{i}}}\right)^{(k-n_{ep}^{S_{i}})}\cdot\frac{1}{{n_{ep}^{S_{i}}}!}{\left(\rho^{S_{i}}(T_{r})\right)^{{n_{ep}^{S_{i}}}k}}\pi_{0}^{S_{i}}&\mbox{if
}\ k\geq{n_{ep}^{S_{i}}}\end{array}\right.$ (3)
By further replacing in the total probability condition ($\sum_{j}\pi_{j}=1$),
one can easily calculate the $\pi_{0}$ probability that no EV will charge at a
specific station $S_{i}$; Implicitly the average queue length (O1 in Fig. 3)
and the average waiting time (O2 in Fig. 3) can then be estimated based on the
surplus of vehicles waiting to charge, the capacity of a charging station and
the probability of these vehicles to arrive at a station $S_{i}$.
$\displaystyle(O1):Q^{S_{i}}(T_{r})$
$\displaystyle=\sum_{k=n_{ep}^{S_{i}}}^{N_{c}^{S_{i}}}(k-n_{ep}^{S_{i}})\cdot\pi_{k}^{S_{i}}$
(4) $\displaystyle(O2):\overline{WT_{q}^{S_{i}}}(T_{r})$
$\displaystyle=\frac{Q^{S_{i}}(T_{r})}{\lambda\left(1-\pi_{k}^{S_{i}}\right)}$
(5)
Other outputs of the EV-Q model can then be calculated on an hourly basis as
follows:
$\displaystyle(O4):WT^{S_{i}}(T_{r})$
$\displaystyle=W_{q}^{S_{i}}(T_{r})+\mu^{S_{i}}(T_{r})$ (6)
$\displaystyle(O5):E^{S_{i}}(T_{r})$
$\displaystyle=PP^{S_{i}}\cdot\mu^{S_{i}}(T_{r})$ (7)
where $WT^{S_{i}}(T_{r})$ represents the total waiting time to charge at a
station $S_{i}$ (O4 in Fig. 3), and $E^{S_{i}}(T_{r})$ is the total energy
consumption to charge during a specific time interval at a given EV station
(O5 in Fig. 3). Other model outputs are extracted such as minimum and maximal
queue length, waiting time and energy consumption (O6-O9). The EV-Q model
outputs are further used in the global traffic-energy simulation modelling
framework: a) the average queue length as an input in the NSTM calibration
process to realistically simulate the impact of queue waiting to charge even
outside of the EV stations and b) the energy consumption estimation for the
impact of fast charging on electricity distribution networks.
## III North Sydney Case Study
### III-A Simulation calibration and validation
Figure 4: Validation of NSTM model for the AM peak via a) Trip Length
Distribution and b) $R^{2}=0.92$ metric.
After the initial construction of the NSTM model, the 200-pair OD matrix
$OD(T_{r})$ is calibrated by using permanent traffic flow counts from 2016
records (marked in green in Fig. 10 from the supplementary material [1]). The
$OD(T_{r})$ matrices are split to contain $93\%$ of regular car trips and
$7\%$ of truck trips in the network according to real-life records. After the
calibration, the dynamic and hourly demand matrices are then validated against
traffic counts received from lane detectors placed at signalised intersections
(showcased in Fig. 11 and Fig. 12 from the supplementary material [1]). The
results for the AM-peak NTSM calibration and validation are provided in Fig.
4, where the Original versus Adjusted Trip Length distributions follow very
similar trends with less than $2\%$ difference between them (Fig. 4a));
similarly, the $R^{2}$ metric calculated between simulated and the real
traffic demand scored 0.92 in accuracy which validates once again the good
calibration of the model (Fig. 4b)). The NSTM calibration and validation for
the PM peak indicates similar performance of $R^{2}=0.92$ and are provided in
the complementary material [1]-Section -B. We make the observation that while
we validated the NSTM model against real-world traffic data from 2016, our two
hypothetical growth scenarios (OD15 and OD30) cannot be similarly validated
and therefore are best thought of as “what-if” scenarios. The outputs of the
simulation model are further used as inputs in both the queue modelling and
energy impact evaluation, as detailed in following section.
### III-B Queue Model results
The traffic flow $Tf_{k}(T_{r})$ and the average travel distance
$Avg_{d}(T_{r})$ have been used as inputs in the EV-Q model with the aim of
calculating the percentages of EVs needing recharging at each station, and the
travel distance they had to travel before recharging. The daily fluctuations
of traffic flow in the network can greatly influence the charging needs (see
Fig. 20 in [1]), and the number of EVs that will seek recharging based on the
nearest station.
Therefore, we apply the EV-Q model under several scenarios which allows an
extensive simulation of current and potential future traffic conditions, by
varying several input parameters such as the battery size of the vehicles
$B_{c}(T_{r})[kWh]$, the penetration rates of EVs that need to charge
$EV_{p}\in\\{0.01,0.02,..0.05,0.1,0.12,...0.2,0.21...0.3,0.5,1,2,5\\}[\%]$ as
well as our three traffic demand scenarios. Overall, the EV-Q modelling allows
us to analyse more than 90,000 different scenarios with modeling precision
down to a second-by-second.
Figure 5: The impact of EV penetration rates on queue lengths, waiting times,
service times and energy consumption.
The nine outputs of the EV-Q summarised in Table I are aggregated hourly and
used for analysing the energy impact when drivers decide to apply an in-route
charging approach (samples of O1-O5 are represented already in Fig. 3 and
several aggregations by station/area are extracted). A detailed selection of
averaged outputs from the EV-Q model is provided in Fig. 5, where we choose
$B_{c}=82kWh$ and vary $EV_{p}$ between $\\{0.01,0.02,..0.3\\}$ (higher
variations are presented in [1]-Fig. 21); we analyse the mean queue length
(O1-Fig. 5a)), the mean waiting time (O2-Fig. 5b)), the mean service time
(O3-Fig. 5c)) and the mean energy consumption (O5-Fig. 5d)). On each figure we
also plot the impact of the number of plugs on the EV-Q model; for example, in
Fig. 5a), the mean queue length recorded across EV stations with one plug
(blue dots and their confidence interval) indicates that cars start to queue
from an $EV_{p}=0.03\%$ and can easily reach 20 cars if $EV_{p}=0.3\%$;
stations with 7 charging plugs only start to accumulate cars in the queue from
$EV_{p}=0.21\%$ while reaching a mean of 10 cars in the queue for
$EV_{p}=0.3\%$; the best performing stations are the ones with 10 charging
plugs which present zero queue accumulation until $EV_{p}=1\%$ (see [1]-Fig.
22). The mean waiting time at stations (Fig. 5b)) is highly influenced by the
available number of fast charging plugs and for even a low $EV_{p}=0.04\%$ it
could reach 1 hour easily. The results highlighted in Fig. 5 indicate that
current fast charging infrastructure deployed in our case study area is likely
insufficient for even low levels of EV uptake and utilisation. Even assuming
that only $0.15\%$ of passing vehicles choose to recharge at a station,
average waiting times balloon to unrealistic and unsustainable levels
(exceeding 2 hours) in all but ten-plug stations. Separately, note that total
electricity consumption is highly sensitive to the placement of charging
infrastructure. The average total energy consumption of 7 plug stations, which
are located in high-volume traffic locations, far outstrip the consumption of
the ten plug stations located in lower volume areas (if $0.15\%$ of passing
vehicles charge, the 7 plug stations more than triple the energy consumption
levels of the 10 plug stations).
The average service time (O3) is under 1 hour for the majority of stations
regardless of the number of EVs waiting in the queue (see trend of Fig. 5c))
while the energy consumption is highly dependent on the traffic flow recorded
at specific stations and the utilisation of that station - for example in Fig.
5d) the 7-plug stations are consuming almost 5 times more than even stations
with ten plugs (at $EV_{p}=0.3\%$). The findings reveal not only the
limitation of the current EV infrastructure in the area, but also allow to
make predictions and planning decisions based on the popularity of specific
areas in the network, their capacity, traffic demand, etc.
### III-C EV impact on congestion
The mean queue results $\overline{Q^{S_{i}}}(T_{r})[veh]$ of the EV-Q
modelling is further utilised in the NSTM model to evaluate the impact that
queue spill-back outside of EV stations would have on the overall traffic
congestion in the network. Fig. 6 presents the daily total travel time spent
by all cars in the network, across all EV penetration rates (Ox axis), for
OD2016 (blue), OD15 (orange) and OD30 (grey). Assuming the traffic demand
remains constant in 2016, the total travel time would increase by almost
$4.9\%$ if the EV penetration rate would reach $5\%$ (see increase from point
A to B). More specifically, this represents an extra of almost $1074$ days
spent in traffic by all vehicles (equivalent to an extra 6.18 hours spent in
traffic for each driver - a highly penalising scenario). OD15 seems to add an
even higher burden: an $8.7\%$ increase from point C to D translates in $11.6$
extra hours of time spent in traffic per year. The OD30 demand scenario
translates in a $12.1\%$ increase assuming $EV_{p}=0.01$ remains the same, an
unrealistic assumption given the global EV uptake trajectories (point E versus
A).
Figure 6: Travel time evolution across several traffic demand and EV
penetration rates for OD2016, OD15, OD30 scenarios.
### III-D EV impact on energy consumption
While increasing the number of plugs can significantly reduce the waiting time
for drivers, the energy impact will also increase considerably; Fig. 7
showcases the regression trends of daily total energy $E^{S_{i}}$(O5) versus
$n_{ep}^{S_{i}}$; while stations with small $n_{ep}^{S_{i}}$ will not affect
the energy consumption (for $n_{ep}^{S_{i}}<2$ and $EV_{p}=5\%$ the
$E^{S_{i}}<1200kWh$), the stations with high number of plugs (7,9,10) will
induce almost a 4-times higher energy consumption in the network
($E^{S_{i}}=4,500kWH$ for $EV_{p}=5\%$); this can have a significant load on
the electric network and the current EV station capacity.
Figure 7: Energy consumption vs. the number of plugs for OD2016. Figure 8:
Energy consumption estimations for OD2016, OD15, OD30.
Furthermore, in Fig. 8 we show that our demand increase assumptions for OD15
and OD30 translates in a saturation of the average daily energy consumption
after an $EV_{p}=2\%$ (reaching 1400kWH) but follows similar trends across all
traffic demands that we tested; this is mostly related to: a) the saturation
of the traffic network when the demand increases considerably and vehicles
would queue everywhere in the network and b) the EV infrastructure limitation
which we assumed remains the same for our experiments despite the traffic
demand would increase even by $30\%$ in the next years. There is however a
significant impact on the energy consumption when switching from
$EV_{p}=0.01\%$ to $EV_{p}=0.1\%$ which translates in an extra 1000kWh
consumed daily on average across all stations. The impact of future EV
increases and how this affects differently all EV stations, with an additional
comparison is provided in [1]-Section -E and Fig. 23. An important aspect is
that EV stations with higher number of plugs (7,9,10) will be the ones that
will increase almost three times more their energy consumption when compared
to smaller stations (up to 4 plugs).
### III-E EV impact on state grid
Figure 9: 2016 state-level weekday energy consumption versus total estimated
EV consumption across all stations in the EV-Q model.
For analysing the EV impact on the electricity grid infrastructure, we compare
the shape of aggregated EV charging demand with the shape of state-wide
electricity demand in Fig. 9. The results suggest that EV fast charging may
exacerbate existing electricity peak demand issues. Indeed, during the
electricity peak demand period, modelled aggregate EV charging demand is
between $91\%$ and $95\%$ of its daily peak. In many ways, the coincidence of
electrical and charging demand is unsurprising: peak electrical demand in New
South Wales is driven by residential energy use, particularly in the period
following the end of the working day; traffic volume (and thus probable fast
charger use) is similarly linked to working patterns and transit from work to
home.
## IV Conclusion
This paper presents an integrated bi-level framework of dynamic traffic
modelling, data-driven queue and energy modelling with the purpose of
evaluating the EV impact on both the traffic network and the energy
consumption of the studies area. Results reveal existing limitations of the EV
charging infrastructure which would not be able to handle even slight
increases of EV rates, as well as significant travel time and waiting time
degradation for larger EV penetration rates rates at each station level. The
proposed bi-level framework was validated against real traffic and energy data
sources and is currently being used for an extended analysis of this case
study.
## References
* [1] online supplement, “Appendix: How will electric vehicles affect traffic congestion and energy consumption: an integrated modelling approach,” 2020, https://bit.ly/32gLSsY.
* [2] B. Csonka and C. Csiszár, “Determination of charging infrastructure location for electric vehicles,” _Transportation Research Procedia_ , vol. 27, pp. 768–775, 2017, 20th EURO Working Group on Transportation Meeting, EWGT 2017, 4-6 September 2017, Budapest, Hungary. [Online]. Available: https://bit.ly/3mNMNuc
* [3] Y. Motoaki, “Location-allocation of electric vehicle fast chargers—research and practice,” _World Electric Vehicle Journal_ , vol. 10, no. 1, 2019. [Online]. Available: https://www.mdpi.com/2032-6653/10/1/12
* [4] D. Chakraborty, D. S. Bunch, J. H. Lee, and G. Tal, “Demand drivers for charging infrastructure-charging behavior of plug-in electric vehicle commuters,” _Transportation Research Part D: Transport and Environment_ , vol. 76, pp. 255–272, 2019. [Online]. Available: https://bit.ly/3abRFo2
* [5] H. Moon, S. Y. Park, C. Jeong, and J. Lee, “Forecasting electricity demand of electric vehicles by analyzing consumers’ charging patterns,” _Transportation Research Part D: Transport and Environment_ , vol. 62, pp. 64–79, 2018. [Online]. Available: https://bit.ly/3uQcWvd
* [6] K. Qian, C. Zhou, M. Allan, and Y. Yuan, “Modeling of load demand due to ev battery charging in distribution systems,” _IEEE Transactions on Power Systems_ , vol. 26, no. 2, pp. 802–810, 2011.
* [7] L. I. Dulău and D. Bică, “Effects of electric vehicles on power networks,” _Procedia Manufacturing_ , vol. 46, pp. 370–377, 2020, 13th International Conference Interdisciplinarity in Engineering, INTER-ENG 2019, 3–4 October 2019, Targu Mures, Romania. [Online]. Available: https://bit.ly/2RFHjX7
* [8] M. Ghamami, Y. M. Nie, and A. Zockaie, “Planning charging infrastructure for plug-in electric vehicles in city centers,” _International Journal of Sustainable Transportation_ , vol. 10, no. 4, pp. 343–353, 2016. [Online]. Available: https://doi.org/10.1080/15568318.2014.937840
* [9] F. He, Y. Yin, and S. Lawphongpanich, “Network equilibrium models with battery electric vehicles,” _Transportation Research Part B: Methodological_ , vol. 67, pp. 306–319, 2014. [Online]. Available: https://bit.ly/3mLdH60
* [10] M. Xu, Q. Meng, and K. Liu, “Network user equilibrium problems for the mixed battery electric vehicles and gasoline vehicles subject to battery swapping stations and road grade constraints,” _Transportation Research Part B: Methodological_ , vol. 99, pp. 138–166, 2017. [Online]. Available: https://bit.ly/3g8pvxT
* [11] J. Zhu, Y. Li, J. Yang, X. Li, S. Zeng, and Y. Chen, “Planning of electric vehicle charging station based on queuing theory,” _The Journal of Engineering_ , vol. 2017, pp. 1867–1871, 2017.
* [12] Fang Lu and Guowei Hua, “A location-sizing model for electric vehicle charging station deployment based on queuing theory,” in _2015 International Conference on Logistics, Informatics and Service Sciences (LISS)_ , 2015, pp. 1–5.
* [13] S. Su, H. Li, and D. W. Gao, “Optimal planning of charging for plug-in electric vehicles focusing on users’ benefits,” _Energies_ , vol. 10, no. 7, 2017. [Online]. Available: https://www.mdpi.com/1996-1073/10/7/952
* [14] D. Xiao, S. An, H. Cai, J. Wang, and H. Cai, “An optimization model for electric vehicle charging infrastructure planning considering queuing behavior with finite queue length,” _Journal of Energy Storage_ , vol. 29, p. 101317, 2020. [Online]. Available: https://bit.ly/2PXAVdg
* [15] J. Jung, J. Y. Chow, R. Jayakrishnan, and J. Y. Park, “Stochastic dynamic itinerary interception refueling location problem with queue delay for electric taxi charging stations,” _Transportation Research Part C: Emerging Technologies_ , vol. 40, pp. 123–142, 2014. [Online]. Available: https://bit.ly/3e3eXNU
* [16] M. Usman, L. Knapen, A.-U.-H. Yasar, T. Bellemans, D. Janssens, and G. Wets, “Optimal recharging framework and simulation for electric vehicle fleet,” _Future Generation Computer Systems_ , vol. 107, pp. 745–757, 2020. [Online]. Available: https://bit.ly/32ihjCU
* [17] M. Klein, L. Lüpke, and M. Günther, “Home charging and electric vehicle diffusion: Agent-based simulation using choice-based conjoint data,” _Transportation Research Part D: Transport and Environment_ , vol. 88, p. 102475, 2020. [Online]. Available: https://bit.ly/3slF3kn
* [18] Electric vehicle Council, “EV Charger Map,” https://electricvehiclecouncil.com.au/about-ev/charger-map/, 2021, online; accessed 22 February18/03/2021.
* [19] Z. Tian, W. Hou, X. Gu, F. Gu, and B. Yao, “The location optimization of electric vehicle charging stations considering charging behavior,” _SIMULATION_ , vol. 94, no. 7, pp. 625–636, 2018. [Online]. Available: https://doi.org/10.1177/0037549717743807
* [20] S. Castano, L. Gauchia, E. Voncila, and J. Sanz, “Dynamical modeling procedure of a li-ion battery pack suitable for real-time applications,” _Energy Conversion and Management_ , vol. 92, pp. 396–405, 2015.
* [21] F. Lambert, “Tesla confirms new 82 kWh battery pack in Model 3, thanks to new cells,” https://bit.ly/3tk2jR7, 2020, [Online; accessed 20-March-2021].
This document is accompanying the submission How will electric vehicles affect
traffic congestion and energy consumption: an integrated modelling approach.
The information in this document complements the submission, and it is
presented here for completeness reasons. It is not required for understanding
the main paper, nor for reproducing the results.
### -A NSTM model representation
Figure 10: North Sydney Transport network layout including the connection of
centroids (main entry points of vehicles in the network; marked as circles
with blue interior) and permanent traffic count locations (marked in green
dots). Figure 11: Layout of detector traffic flow. Figure 12: Sample
representation of a signalised intersection including signal controls, phases,
turnings and detectors for vehicle counting per each turning.
### -B NSTM - PM peak validation
The model validation and calibration has been done as well for the PM peak
hours as represented in figures below.
Figure 13: Validation of PM-NSTM model via Trip Length Distribution Figure
14: Validation of the PM-NSTM model with $R^{2}=0.91$ metric.
While the PM peak distribution after calibration indicates a lower number of
adjusted vehicles travelling in the network, the trend profile still follows
the original profiling of the data set. In addition Fig. 14 indicates a high
$R^{2}=0.91$ which is a proof of model well calibrated, despite few cases of
over and under calibrated traffic counts.
### -C NSTM - car arrivals and mean distance travelled
One of the outputs of the NSTM model is the total number of arrivals and
departures from each area during AM and PM peaks. For modelling a home to
destination charging, the number of arriving cars in each urban area is
extremely important in order to be able to estimate the total number of EVs
that will require charging at specific times. Fig. 15 and Fig. 17 indicates
the AM and PM car arrivals for each of the urban Local Governmental Area in
our case study which allowed us to identify the top most attractive locations.
For example, Dalley is a small catchment, but this area connects two parts of
the city through Harbour Bridge.
Figure 15: NTSM car arrival by area - AM peak. Figure 16: NTSM mean distance
travelled - AM peak. Figure 17: NTSM car arrival by area - PM peak. Figure 18:
NTSM mean distance travelled - PM peak.
### -D Capacity and loading of 10 most popular EV stations
Figure 19: Zone substation area catchment.
The Zone substation catchment area helped to identify the mapping of EV
stations to the regional energy consumption and the analysis of the energy
load for the top 10 most popular stations. Findings have revealed that 7 out
of these 10 stations are operating at more than $75\%$ of their capacity,
which already reveals current EV infrastructure limitation for further EV
uptake.
Figure 20: Load and capacity of 10 most popular EV stations.
### -E Energy impact of various scenarios
Figure 21: Raw results for EV-Q model outputs for OD2016, OD15, OD30
aggregated across all stations. Figure 22: Mean queue variation for 10 plug EV
stations. Figure 23: Energy consumption of all stations vs $EV_{p}$,
|
# Subfrequency search at a laser as a signal of dark matter sector particles
Hye-Sung Lee Department of Physics, KAIST, Daejeon 34141, Korea Jiheon Lee
Department of Physics, KAIST, Daejeon 34141, Korea Jaeok Yi Department of
Physics, KAIST, Daejeon 34141, Korea
(January 2022)
###### Abstract
We introduce a new method to search for the dark matter sector particles using
laser light. Some dark matter particles may have a small mixing or interaction
with a photon. High-power lasers provide substantial test grounds for these
hypothetical light particles of exploding interests in particle physics. We
show that any light source can also emit a subfrequency light as a new physics
signal. Searching for this subfrequency light at a laser can be a simple and
effective way to investigate the new light particles, even in tabletop optics
experiments.
## I Introduction
One of the characteristic trends in particle physics research in the past
decade or two is investigating the very light dark matter sector particles.
While the standard model (SM) is the utmost successful model of the unveiled
particles and their interactions, cosmological observations indicate that 95%
of the Universe consists of nonluminous matter and energy, which cannot be
explained by the SM Aghanim _et al._ (2020). The mysterious dark sector may
have various kinds of particles as the SM sector, and the two sectors may have
a feeble interaction channel referred to portals Essig _et al._ (2013).
This tiny size of interactions allows extremely light mass for the dark matter
sector particles that often require an ingenious strategy to probe. It brings
the other fields such as optics and condensed matter physics into a new
physics search originating from the dark matter sector, permitting genuinely
interdisciplinary studies Geraci _et al._ (2018); Bartram _et al._ (2021);
Melcón _et al._ (2020); Trickle _et al._ (2020); Ehret _et al._ (2010);
Spector (2017); Robilliard _et al._ (2007); Chou _et al._ (2008); Afanasev
_et al._ (2009); Ballou _et al._ (2015); Cameron _et al._ (1993).
Among the popular dark sector particles are an axion ($a$) and a dark photon
($\gamma^{\prime}$). They are exceptional in the sense that they mix or couple
to a photon ($\gamma$), which chiefly distinguishes the dark sector from the
visible sector made of the SM particles. These particles have been searched
for using this connection to the light in the lab experiments as well as in
the astrophysical observations. (See Refs. Zyla _et al._ (2020); Kim and
Carosi (2010); Fabbrichesi _et al._ (2021) for the reviews on the related
subjects.)
Figure 1: Diagrams showing how the subfrequency photon can be emitted through
the combined dark sector portals ($\varepsilon$ and
$G_{a\gamma\gamma^{\prime}}$) when there is a source of a photon.
In this paper, we make one key observation regarding a photon and the dark
matter sector and study its implication in one experimental setup, which is a
laser experiment. Our observation is following. Whenever there is a source
that emits light, the same source may also emit subfrequency light due to the
connections of a photon to some dark sector particles. A dark photon may
couple to the electromagnetic current because of the kinetic mixing with a
photon. It decays to an axion and a lower energy (subfrequency) photon, which
can be the new physics signal. This is a very general property, which is
depicted in Fig. 1. As there is a great number of sources of light, the
potential applications of this possibility are far reaching. We find the
optical laser system is a simple and effective way to study this new physics.
Although there have been a few laser experiments to search for new physics,
this one is qualitatively different from any of them in generating and
detecting the new physics signal.
The rest of this paper is organized as followings. In Sec. II, we describe the
relevant Lagrangian for the new physics scenario. In Sec. III, we describe how
the subfrequency light can be produced from the dark photon decay. In Sec. IV,
we describe the kinematics of the relevant process. In Sec. V, we propose a
new laser experiment and the sensitivity of the experiment. In Sec. VI, we
discuss the general properties of the new idea and experiment. In Sec. VII, we
summarize our findings and depict outlooks.
## II New physics
The SM particles do not carry a charge of the dark sector gauge symmetry
$U(1)_{\text{dark}}$. The vector portal ($\varepsilon$) Holdom (1986) is a
dimensionless parameter, which is a small kinetic mixing between the
$U(1)_{\text{QED}}$ and the $U(1)_{\text{dark}}$. Through this
$\gamma$-$\gamma^{\prime}$ mixing, the dark photon can couple to the
electromagnetic current with a suppressed coupling constant $\varepsilon e$ in
contrast to the electromagnetic coupling constant $e$.
The dark axion portal ($G_{a\gamma\gamma^{\prime}}$) Kaneta _et al._ (2017)
is a nonrenormalizable $a$-$\gamma$-$\gamma^{\prime}$ vertex, whose actual
size depends on the details of the underlying theory. (By the axion, we mean
the axionlike particle in general, not necessarily related to the strong $CP$
problem Peccei and Quinn (1977); Weinberg (1978); Wilczek (1978).) Many
implications of this portal have been found in various contexts Choi _et al._
(2017); Kitajima _et al._ (2018); deNiverville _et al._ (2018); Arias _et
al._ (2021); Deniverville _et al._ (2021); Hook _et al._ (2021); Ge _et
al._ (2021).
The relevant part of the Lagrangian in the mass eigenstate is
$\mathcal{L}\sim-\big{(}A^{\text{phys}}_{\mu}+\varepsilon{A^{\prime}}^{\text{phys}}_{\mu}\big{)}J_{\text{em}}^{\mu}+\frac{G_{a\gamma\gamma^{\prime}}}{2}aF^{\text{phys}}_{\mu\nu}\tilde{F}_{\text{phys}}^{\prime\mu\nu}$
(1)
where $J_{\text{em}}^{\mu}$ is the electromagnetic current,
$F^{\text{phys}}_{\mu\nu}=\partial_{\mu}A^{\text{phys}}_{\nu}-\partial_{\nu}A^{\text{phys}}_{\mu}$
is the field strength tensor for the photon field $(A_{\mu}^{\text{phys}})$,
and ${F^{\prime}}^{\text{phys}}_{\mu\nu}$ is for the dark photon field
$({A^{\prime}}_{\mu}^{\text{phys}})$. For the dark photon heavier than the
axion, $\gamma^{\prime}\to a\gamma$ decay can occur with the
$G_{a\gamma\gamma^{\prime}}$. Although it is possible to consider the axion
decay into a pair of photons in the presence of the axion portal
($G_{a\gamma\gamma}$) allowing $\gamma^{\prime}\to
a\gamma\to\gamma\gamma\gamma$, we treat things as model independent as
possible and assume the axion portal is zero in this paper. In this sense, the
number of photon signals in our analysis will be conservative. The full
Lagrangian and the extended Maxwell’s equations in the presence of various
dark sector portals can be found in Ref. Huang and Lee (2019).
## III Subfrequency light from the dark photon decay
In the following, we use the number flux of laser photon $N_{\gamma}$, defined
as $P/\omega$, where $P$ is the laser power and $\omega$ is the frequency. At
the laser, a dark photon with mass $m_{\gamma^{\prime}}$ can be produced
through the suppressed coupling to the electromagnetic current. Alternatively,
this dark photon can be understood in terms of the photon-dark photon
oscillations in a similar fashion to the neutrino flavor oscillations, and it
has been searched for in the light shining through the wall (LSW) experiments
Ehret _et al._ (2010); Spector (2017); Robilliard _et al._ (2007); Chou _et
al._ (2008); Afanasev _et al._ (2009); Ballou _et al._ (2015); Cameron _et
al._ (1993). The largest mass of the dark photon that we can produce from the
optical laser is the laser frequency of the eV scale
($m_{\gamma^{\prime}}<\omega$).
In the presence of a lighter axion, the dark photon may decay into a photon
and an axion with a decay rate
$\Gamma\equiv\Gamma_{\gamma^{\prime}\rightarrow
a\gamma}=\frac{G_{a\gamma\gamma^{\prime}}^{2}}{96\pi}m_{\gamma^{\prime}}^{3}\bigg{(}1-\frac{m_{a}^{2}}{m_{\gamma^{\prime}}^{2}}\bigg{)}^{3}.$
(2)
The photon produced from this dark photon decay, which we call the signal
photon, has a smaller frequency than the original laser frequency.
In order to demonstrate the production of the signal photon, we investigate
the photon-dark photon oscillation together with the decay of massive dark
photons. We define the massless eigenstate as a photon mass eigenstate
$\ket{\gamma}$ and massive eigenstate as a dark photon mass eigenstate
$\ket{\gamma^{\prime}}$. The mass eigenstates propagate as plane waves of the
form
$\displaystyle\ket{\gamma(t,L)}$ $\displaystyle=e^{-i(\omega t-\omega
L)}\ket{\gamma}\,,$ (3) $\displaystyle\ket{\gamma^{\prime}(t,L)}$
$\displaystyle=e^{-i(\omega t-pL)}e^{-m_{\gamma^{\prime}}\Gamma
t/2\omega}\ket{\gamma^{\prime}}\,,$ (4)
where $p=\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}$ is the momentum of the
dark photon. The Lorentz factor $\gamma=\omega/m_{\gamma^{\prime}}$ reflects
the time dilation of the decay. Here, $t$ and $L$ are measured in the lab
frame.
When the photon is produced in the laser, it is produced in a flavor
eigenstate $\ket{A}$ and its orthogonal pair is $\ket{X}$. They are related to
the mass eigenstates $\ket{\gamma}$ and $\ket{\gamma^{\prime}}$ by $2\times 2$
unitary matrix.
$\begin{bmatrix}A\\\
X\end{bmatrix}=\begin{bmatrix}\sqrt{1-\varepsilon^{2}}&\varepsilon\\\
-\varepsilon&\sqrt{1-\varepsilon^{2}}\end{bmatrix}\begin{bmatrix}\gamma\\\
\gamma^{\prime}\end{bmatrix}$ (5)
Let our state emitted from the laser be $\ket{\psi(t,L)}$. Since the initial
state $\ket{\psi(0,0)}=\ket{A}$, the state after time $t$ and length $L$ of
propagation is
$\displaystyle\ket{\psi(t,L)}$
$\displaystyle=\sqrt{1-\varepsilon^{2}}e^{-i(\omega t-\omega L)}\ket{\gamma}$
$\displaystyle+\varepsilon e^{-i(\omega t-pL)-m_{\gamma^{\prime}}\Gamma
t/2\omega}\ket{\gamma^{\prime}}\,.$ (6)
We can obtain a probability that the initial $\ket{A}$ turns into $\ket{X}$
after propagation as
$\displaystyle\mathcal{P}_{A\rightarrow X}(L)$
$\displaystyle=|\braket{X}{\psi(t,L)}|^{2}$
$\displaystyle=\varepsilon^{2}\bigg{[}1+e^{-\frac{m_{\gamma^{\prime}}\Gamma
t}{\omega}}-2e^{-\frac{m_{\gamma^{\prime}}\Gamma t}{2\omega}}\cos(L\Delta
p)\bigg{]}$
$\displaystyle=\varepsilon^{2}\bigg{[}1+e^{-\frac{m_{\gamma^{\prime}}\Gamma
L}{p}}-2e^{-\frac{m_{\gamma^{\prime}}\Gamma L}{2p}}\cos(L\Delta p)\bigg{]}$
(7)
where $\Delta p=\omega-p=\omega-\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}$ and
the terms with $\mathcal{O}(\varepsilon^{4})$ are neglected. (We assume small
kinetic mixing.) We used $L=vt$ and $p=\omega v$ in the last line. Likewise, a
simple calculation gives the probability that $X$ results in decay products.
$\displaystyle\mathcal{P}_{X\rightarrow\text{decay}}(L)$
$\displaystyle=1-\mathcal{P}_{X\rightarrow A}(L)-\mathcal{P}_{X\rightarrow
X}(L)$
$\displaystyle=(1-\varepsilon^{2})\left[1-e^{-\frac{m_{\gamma^{\prime}}\Gamma
L}{p}}\right]$ (8)
Here, the oscillation behavior is suppressed with
$\mathcal{O}(\varepsilon^{4})$.
As we will describe in Sec. V, although it is not the only possible option,
our proposed setup is basically the LSW-type experiment in which there are two
baselines of length $L_{1}$ and $L_{2}$ that are separated by a “wall” in
between. So the laser light initially in the $\ket{A}$ travels the $L_{1}$,
then only the $\ket{X}$ component of the light travels the $L_{2}$ (as a
photon cannot go through the wall) decaying into an axion and a subfrequency
photon, which we measure at the end of $L_{2}$.
Therefore, the flux of the subfrequency signal photon is given as
$\displaystyle N_{\text{sub}}$ $\displaystyle=\mathcal{P}_{A\rightarrow
X}(L_{1})\mathcal{P}_{X\rightarrow\text{decay}}(L_{2})N_{\gamma}$
$\displaystyle=\varepsilon^{2}N_{\gamma}\bigg{[}1-e^{-\frac{m_{\gamma^{\prime}}\Gamma
L_{2}}{p}}\bigg{]}$
$\displaystyle\times\bigg{[}1+e^{-\frac{m_{\gamma^{\prime}}\Gamma
L_{1}}{p}}-2e^{-\frac{m_{\gamma^{\prime}}\Gamma L_{1}}{2p}}\cos(L_{1}\Delta
p)\bigg{]}.$ (9)
In this paper, we work on a “simple setup” described by the following
conditions.
1. (i)
$m_{a}\ll m_{\gamma^{\prime}}<\omega$.
2. (ii)
$\dfrac{m_{\gamma^{\prime}}\Gamma
L_{i}}{\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}}\ll 1$ (for $i=1,\ 2$).
3. (iii)
$L_{1}\Delta p\gg 2\pi$.
These are reasonable assumptions for the optical experiment in the laboratory
scale for the following reasons.
Condition (i) reflects that the dark photon should be produced from the laser
($m_{\gamma^{\prime}}<\omega$) and decay into a much lighter axion ($m_{a}\ll
m_{\gamma^{\prime}}$) and a subfrequency photon. The latter ensures the
subfrequency photon has a sizable number of events and a visible light
frequency range that can be detected with a typical photon detector, although
we will briefly discuss how the non-negligible axion mass ($m_{a}\sim
m_{\gamma^{\prime}}$) can change the result in Sec. VI.
Condition (ii) is the requirement to put the relevant exponential expressions
in linear form of $L$ as in Eq. (11), which is convenient for analysis. Under
this condition, the probability of the dark photon decay during the
propagation is much less than 1. Using Eq. (2), this can be rewritten as
$G_{a\gamma\gamma^{\prime}}^{2}\ll
96\pi\frac{\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}}{m_{\gamma^{\prime}}^{4}}\frac{1}{L_{2}}\,.$
(10)
This can be easily satisfied for the typical lab-size experiment scales if
$m_{\gamma^{\prime}}\ll\omega$. The left-hand side of condition (ii) diverges
when $m_{\gamma^{\prime}}\sim\omega$, but it hardly occurs because of the
finite line width of a laser. Thus only a tiny fraction of the dark photon can
have an enhancement effect, even if the $m_{\gamma^{\prime}}$ happens to
coincide the laser frequency.
Condition (iii) allows us to neglect the oscillation in
$\mathcal{P}_{\gamma\rightarrow\gamma^{\prime}}$. If $L_{1}\Delta p$ is much
larger than $2\pi$, the cosine function in
$\mathcal{P}_{\gamma\rightarrow\gamma^{\prime}}$ is highly oscillatory. The
linewidth of laser in the setup can alter $\Delta p$, and its small change can
alter the signal size uncontrollably. To treat this oscillation, we take the
average value of the cosine function in Eq. (9). Even in the low mass region,
this condition can be satisfied by our interested parameter region for
$L_{1}>1\text{m}$.
In the simple setup described above, the signal photon ratio can be obtained
by
$\frac{N_{\text{sub}}}{N_{\gamma}}=\frac{K^{2}}{48\pi}\frac{m_{\gamma^{\prime}}^{4}}{\sqrt{\omega^{2}-m^{2}_{\gamma^{\prime}}}}L$
(11)
where $L=L_{2}$ and the effective portal coupling $K$ is given as a product of
the two portal couplings
$K\equiv\varepsilon\,G_{a\gamma\gamma^{\prime}}.$ (12)
The fraction of the subfrequency light is given by two model parameters $K$,
$m_{\gamma^{\prime}}$, the laser frequency $\omega$, and a geometry factor
$L$. The factor of 2 difference between Eqs. (2) and (11) comes from the fact
that both mass eigenstates generate the $X$ state.
## IV Kinematics
Consider the rest frame of a dark photon, which decays into a subfrequency
signal photon and an axion. In our simple setup where an axion is considered
nearly massless, both the photon and axion have the same energy
$m_{\gamma^{\prime}}/2$. Since there is no preference of direction in the rest
frame, the probability density ($dp/d\Omega$) that the signal photon heads to
a specific solid angle $d\Omega$ is given by
$\frac{dp}{d\Omega}=\frac{1}{4\pi},\quad\int\frac{dp}{d\Omega}d\Omega=1.$ (13)
By denoting the solid angle with the polar angle $\theta_{\text{rest}}$ and
azimuthal angle $\phi_{\text{rest}}$, we have
$\int\frac{dp}{d\Omega}\sin\theta_{\text{rest}}d\theta_{\text{rest}}d\phi_{\text{rest}}=\int\frac{\sin\theta_{\text{rest}}}{2}d\theta_{\text{rest}}=1,$
(14)
where we used azimuthal symmetry in the first equality. Using this relation,
we can define the probability density function with respect to
$\theta_{\text{rest}}$:
$\frac{dp}{d\theta_{\text{rest}}}=\frac{\sin\theta_{\text{rest}}}{2}$ (15)
Let the $z$ axis be the direction a dark photon travels in. In the lab frame,
the dark photon has the energy $\omega=\gamma m_{\gamma^{\prime}}$, where
$\gamma=1/\sqrt{1-\beta^{2}}$ is the Lorentz factor. Thus, the lab frame
boosts with velocity $\beta=\sqrt{1-(m_{\gamma^{\prime}}/\omega)^{2}}$. Then
the energy $E$ and polar angle $\theta_{\text{lab}}$ of a signal photon in the
lab frame are
$\displaystyle E$ $\displaystyle=$
$\displaystyle\frac{\omega}{2}(1+\beta\cos\theta_{\text{rest}}),$ (16)
$\displaystyle\tan\theta_{\text{lab}}$ $\displaystyle=$
$\displaystyle\frac{\sin\theta_{\text{rest}}}{\gamma(\cos\theta_{\text{rest}}+\beta)}.$
(17)
If $\beta\ll 1$ ($m_{\gamma^{\prime}}\approx\omega)$, then $E$ is not
dispersive. But $\theta_{\text{lab}}$ gets close to $\theta_{\text{rest}}$ and
the signal photons spread almost isotropically. On the other hand, $E$ is
highly dispersive if $\beta\approx 1$ ($m_{\gamma^{\prime}}\ll\omega)$. For a
fixed $\omega$, as $m_{\gamma^{\prime}}$ increases, the dispersion of $E$
decreases, but that of $\theta_{\text{lab}}$ increases. We can check these
dispersions by computing the probability density with respect to $E$ and
$\theta_{\text{rest}}$ for each. They are given by
$\frac{dp}{dE}=\frac{dp}{d\theta_{\text{rest}}}\left|\frac{d\theta_{\text{rest}}}{dE}\right|,\quad\frac{dp}{d\theta_{\text{lab}}}=\frac{dp}{d\theta_{\text{rest}}}\left|\frac{d\theta_{\text{rest}}}{d\theta_{\text{lab}}}\right|.$
(18)
These probability distributions and dispersive behaviors are illustrated in
Fig. 2.
Figure 2: Probability density function with respect to the energy $E$ of the
subfrequency signal photon and the angle in the lab frame
$\theta_{\text{lab}}$. As the dark photon mass $m_{\gamma^{\prime}}$
increases, the energy of the signal photon is focused around $\omega/2$ but
angular distribution gets more disperse.
One of the interesting features of the energy dispersion, as we can find in
Fig. 2, is that the distribution is uniform. This probability distribution
function is given by
$\frac{dp}{dE}=\frac{1}{\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}}~{}\text{
for }\frac{\omega}{2}(1-\beta)<E<\frac{\omega}{2}(1+\beta).$ (19)
The energy dispersion also shows that the frequency of resulting signal is
smaller than the laser frequency $\omega$, which justifies our naming
“subfrequency.” However, one should note that despite the dispersion of $E$, a
signal photon whose energy is much smaller than $\omega/2$ is produced very
little and in the backward direction as one can see from Eqs. (16)–(17).
Because of the angular dispersion of the subfrequency photon, the location and
the size of the detector could matter in detecting the signal. We want to
estimate how much signal can be collected in a circular detector of radius $a$
located at $L$ from the dark photon production. Often, a lens is installed in
front of a detector to focus the beam into the small detector, then the “$a$”
can be replaced by the radius of the lens as an effective detector size in the
conservative sense.
Figure 3: $\eta_{\text{kin}}$ of the subfrequency new physics signal in
various experimental setups. The angular dispersion of the signal photon is
large for high dark photon mass, and it weakens $\eta_{\text{kin}}$ unless a
suitable measure is taken (as in waveguide). Dotted curves denote the
waveguide with 98.5% reflectivity coating, while solid ones denote 100%. The
solid blue and solid purple curves (two curves with 100% reflectivity)
coincide. A small kink in the dotted curves are due to the limited wavelength
range of the mirror. (See texts in Sec. V for details.)
Let $\ell$ be the distance from the detector to the point a decay occurs. In
our simple setup, we can have
$\frac{d^{2}p}{d\ell
d\theta_{\text{lab}}}=\frac{1}{L}\frac{dp}{d\theta_{\text{lab}}}=\frac{1}{L}\frac{dp}{d\theta_{\text{rest}}}\left|\frac{d\theta_{\text{rest}}}{d\theta_{\text{lab}}}\right|.$
(20)
The probability that the photon reaches the detector is
$\eta_{\text{kin}}=\int_{0}^{L}d\ell\int_{0}^{\tan^{-1}(a/\ell)}d\theta_{\text{lab}}\
\frac{d^{2}p}{d\ell d\theta_{\text{lab}}},$ (21)
which is straightforward to calculate.
For the later use, we will briefly discuss substitution between
$\theta_{\text{lab}}$ and $E$. In many cases, calculations become much easier
by changing the variable from $\theta_{\text{lab}}$ to $E$ since $dp/dE$ is
uniform. The relations between $E$ and $\theta_{\text{lab}}$ are given by
$\displaystyle\theta_{\text{lab}}(E)$ $\displaystyle=\cos^{-1}\frac{2E\omega-
m_{\gamma^{\prime}}^{2}}{2E\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}},$ (22)
$\displaystyle E(\theta_{\text{lab}})$
$\displaystyle=\frac{m_{\gamma^{\prime}}^{2}\left(\omega+\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}\cos\theta_{\text{lab}}\right)}{2\left(\omega^{2}\sin^{2}\theta_{\text{lab}}+m_{\gamma^{\prime}}^{2}\cos^{2}\theta_{\text{lab}}\right)}.$
(23)
Using this substitution, Eq. (21) can be rewritten as
$\displaystyle\eta_{\text{kin}}$
$\displaystyle=\int_{0}^{L}d\ell\int^{E(\theta_{\text{lab}}=0)}_{E(\theta_{\text{lab}}=\tan^{-1}(a/\ell))}dE\
\frac{d^{2}p}{d\ell dE}$
$\displaystyle=\frac{1}{\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}}\int_{0}^{L}d\ell\int^{E(\theta_{\text{lab}}=0)}_{E(\theta_{\text{lab}}=\tan^{-1}(a/\ell))}dE$
$\displaystyle=\int_{0}^{L}\frac{m_{\gamma^{\prime}}^{2}\ell(\ell-\sqrt{\ell^{2}+a^{2}})+a^{2}\omega(\omega+\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}})}{2L(\ell^{2}m_{\gamma^{\prime}}^{2}+a^{2}\omega^{2})}d\ell.$
(24)
In Fig. 3, we show $\eta_{\text{kin}}$ for various experimental setups. Some
of these setups are the existing LSW experiments that are partly sensitive to
our subfrequency new physics scenario. We used the parameter values specified
in Table 1. For $m_{\gamma^{\prime}}\approx\omega$, only a tiny fraction of
the signal photon can be detected. ALPS II has the lowest $\eta_{\text{kin}}$
due to its smallest $a/L$ ratio. It is clear how important to collect
dispersive signal photons. As we will detail later in this paper, our new
design adopts a waveguide that can fix this, and at least half of the signal
photons can reach the detector in ideal case. (See waveguide in Fig. 3.)
| a (mm) | $L$ (m) | $\omega$ (eV) | $N_{\gamma}$ (Hz) | $N_{\text{pass}}$ | $\eta_{\text{eff}}$ | $N_{d}$ (Hz) | $t_{s}$ (h)
---|---|---|---|---|---|---|---|---
Waveguide | $-$ | $1,\ 100$ | $1.17$ | $1.6\times 10^{20}$ | $5000$ | $0.54$ | $10^{-6}$ | $480$
ALPS II | $8.75$111The ALPS II uses an optic suitable to collect a 17.5 mm diameter beam Bähre _et al._ (2013). | $100$ | $1.17$ | $1.6\times 10^{20}$ | $5000$ | $0.95$ | $10^{-6}$ | $480$
ALPS | $7$222Because we are unaware of the radius of the lens, we take the maximum of the vacuum tube size that can be inserted in the HERA dipole magnet Ehret _et al._ (2009). | $7.6$ | $2.33$ | $2.6\times 10^{21}$ | $1$ | $0.9$ | $0.0018$ | $27$
GammeV | $25.5$333We take the size of the lens in front of the PMT Chou _et al._ (2008). | $7.2$ | $2.33$ | $6.6\times 10^{23}$ | $1$ | $0.25$ | $130$ | $24$
SPring-8 | $6$444We take the radius of crystal in Ge detector Inada _et al._ (2013). | $0.654$ | $7270\sim 26000$ | $4.3\times 10^{12}\sim 8.9\times 10^{13}$ | $1$ | $0.23\sim 0.83$ | $0.0019\sim 0.0142$ | $5.28\sim 8.88$
Table 1: Specifications of the experimental setups we use in our analysis. The
LSW experiments (GammeV, ALPS, ALPS II) are partly sensitive to the sub-
frequency new physics scenario. Our proposed experimental setup (Waveguide)
adopts a waveguide. The parameters for the LSW experiments including
$\eta_{\text{eff}}$ were taken from Refs. Ehret _et al._ (2010); Flambaum
_et al._ (2019); Bastidon _et al._ (2015); Inada _et al._ (2013). Some of
the parameter values may not properly reflect the actual experiments. For
instance, the $\eta_{\text{eff}}$ of the LSW experiments would change if the
frequency-dependence is properly applied.
## V New laser experiment
Figure 4: Proposed layout (waveguide). The red is the photon beam, and the
black is the dark photon beam. The mirror (M) reflects the laser beam
(photon), and only the dark photon can pass the mirror and decay to the
subfrequency signal photon. (We do not show the axion in the figure.) The
dashed line is the waveguide, which protects the signal from dispersion. A
lens is placed in front of the photon detector to focus the signal light into
the detector.
Figure 4 shows a schematic diagram of the laser experiment setup that can
explore the subfrequency new physics scenario. A small portion of laser light
converts to the dark photon, which subsequently decays into the photon and
axion. Because of the different frequencies, we can distinguish the
subfrequency signals from the original laser beam in several ways (for
instance, we can adopt a prism to distinguish the subfrequency signal light
from the original laser light), but here we adopt a rather simple method. We
use a mirror (M) to reflect the laser beam. Then only the dark photon beam
passes the mirror. The decay can occur anywhere in the propagation of the dark
photon, but the decay signal made before the M cannot pass the mirror. Thus,
the decay length ($L$) should be taken as the distance between the M and the
detector.
The Fabry-Perot cavity before the M is used to amplify the dark photon
transition rate through the mirrors, and the cavity amplification ratio
($\eta_{\text{cav}}$) is given by Ahlers _et al._ (2007)
$\eta_{\text{cav}}=(N_{\text{pass}}+1)/2$ (25)
where $N_{\text{pass}}$ is the number of the reflections of the laser beam
inside the cavity. This depends on the cavity factors and is limited by the
damage threshold of cavity mirrors at which the surface coating on the mirror
can be damaged. The authors of the Ref. Bähre _et al._ (2013) limit the
maximum intensity in the cavity mirror to 500 kW/cm2 (150 kW for 5 mm minimum
beam radius in ALPS II). One has a wide choice for the laser intensity and the
cavity amplification as long as the circulating power inside the cavity does
not exceed the damage threshold. We also take the 150 kW for our maximum
circulating power ($\omega N_{\gamma}N_{\text{pass}}<150$ kW).
To avoid losing signal photons from the angular dispersion described earlier,
we use a waveguide that can collect and guide all the signal photons in the
forward direction toward the detector, which will save at least half of the
signals. A long hollow pipe/rectanguloid with the inside coated by a highly
reflective metal would be suitable. The silver-coated mirror has high
reflectivity in the optical to near IR range, commercial mirror with more than
98.5$\%$ reflectivity in the 1100 to 20000 nm range is available. (In fact,
there is a reduced but still nonzero reflectivity below the 1100 nm wavelength
range, but we take it zero for the simplicity.) We choose the radius of
waveguide as the same as that of lens in ALPS II since we also want to use a
lens.
In the presence of the waveguide, one should consider the loss of the signal
due to the reflections at the mirror. Thus, for the waveguide setup,
$\eta_{\text{kin}}$ should be modified to
$\eta_{\text{kin}}=\frac{1}{\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}}\int_{0}^{L}d\ell\int^{E(\theta_{\text{lab}}=0)}_{E(\theta_{\text{lab}}=\pi/2)}dE\;\mathcal{R}(\theta_{\text{lab}}(E),\ell)\,.$
(26)
The $\mathcal{R}(\theta_{\text{lab}}(E),\ell)$ is a surviving fraction of the
photon, and we obtain it as
$\mathcal{R}(\theta_{\text{lab}}(E),\ell)=r(E)^{\frac{\ell}{2R}\tan\theta_{\text{lab}}+\frac{1}{2}}\,,$
(27)
where ${R}$ is the radius of the waveguide and $r$ is the reflectivity of the
mirror. The power of $r$ is the number of the reflections each signal photon
experiences. This expression slightly overcounts the reflection by a decimal,
but it is negligible when $r$ is close to $1$. The
$\mathcal{R}(\theta_{\text{lab}},\ell)$ is always less than 1 and
$\mathcal{R}(\theta_{\text{lab}},\ell)\rightarrow 1$ as $r\rightarrow 1$.
Figure 5: The averaged detector efficiency ($\bar{\eta}_{\text{eff}}$) with
the dark photon mass ($m_{\gamma^{\prime}}$) for various TES devices when
$\omega$=1.17 eV. We take tungsten (W) TES Lita _et al._ (2008),
titanium/silver (Ti/Au) TES Niwa _et al._ (2017), hafnium (Hf) TES Lita _et
al._ (2010) for the waveguide with $98.5\%$ reflectivity with $L=1$ m (dashed
curves), $L=100$ m (solid curves). Various TES are quite sensitive in the
given mass range.
In general, the SNR of the single photon detection is given by
$\text{SNR}=\frac{N_{s}\sqrt{t_{s}}}{\sqrt{N_{s}+N_{d}}}$ (28)
where $N_{s}$ ($N_{d}$) is the number of signal photon (noise) per second, and
$t_{s}$ is the total measurement time. To be sensitive to small coupling $K$,
it is required for the detector to have an extremely low noise. Also, one
needs to consider a detector efficiency ($\eta_{\text{eff}}$), which is a
ratio of the detected number of the photon to the incident photon to the
detector.
In our subfrequency search scenario, as we discussed in Sec. IV, the detector
needs to be sensitive to the frequency around $\omega/2$. Thus, the transition
edge sensor (TES) which has high sensitivity and low noise in the near IR to
the optical range, can be a suitable detector. Figure 5 shows the efficiency
of TES devices for the waveguide experiment with $L=1$ m and $L=100$ m. Since
the detector efficiency mainly depends on the energy of the incident signal,
one needs to consider the energy and angular dispersion of the subfrequency
signal.
We take the averaged detector efficiency as
$\bar{\eta}_{\text{eff}}=\frac{1}{\eta_{\text{kin}}}\frac{1}{\sqrt{\omega^{2}-m_{\gamma^{\prime}}^{2}}}\\\
\int_{0}^{L}d\ell\int^{E(\theta_{\text{lab}}=0)}_{E(\theta_{\text{lab}}=\pi/2)}dE\;\eta_{\text{eff}}(E)\mathcal{R}(\theta_{\text{lab}}(E),\ell)\,.$
(29)
When $m_{\gamma^{\prime}}\ll\omega$, the subfrequency signal is widely
dispersed in the energy domain (see Fig. 2). In that limit, the frequency-wide
average makes the sensitivity of each TES flat as shown in Fig. 5. In
contrast, when $m_{\gamma^{\prime}}\sim\omega$, the subfrequency signals focus
at $\omega/2$ and has a large angular dispersion. Therefore, the local shape
of the $\eta_{\text{eff}}(E)$ and the signal loss by the waveguide is fed into
the curve of $\bar{\eta}_{\text{eff}}$ near the $\omega/2$.
Finally, the ratio of the detectable signal photon number flux to the laser
flux is
$\frac{N_{s}}{N_{\gamma}}=\eta\frac{N_{\text{sub}}}{N_{\gamma}}=\eta\frac{K^{2}}{48\pi}\frac{m_{\gamma^{\prime}}^{4}}{\sqrt{\omega^{2}-m^{2}_{\gamma^{\prime}}}}L\,,$
(30)
with
$\eta\equiv\eta_{\text{kin}}\eta_{\text{cav}}\bar{\eta}_{\text{eff}}\,.$ (31)
## VI Discussions
So far, we considered only the case when $m_{a}\ll m_{\gamma^{\prime}}$. Non-
negligible axion mass has the following effects. The decay rate [see Eq. (2)]
and the energy of the each subfrequency photon decrease as
$\Gamma\rightarrow(1-m_{a}^{2}/m_{\gamma^{\prime}}^{2})^{3}\Gamma$,
$E\rightarrow(1-m_{a}^{2}/m_{\gamma^{\prime}}^{2})E$. These imply that the
subfrequency signal becomes weaker and no longer lies in the near IR to the
optical range as $m_{a}\rightarrow m_{\gamma^{\prime}}$. This would require
microwave or radio frequency detectors, which we do not consider in this
paper.
Figure 6: 1$\sigma$ C.L. exclusion/sensitivity limits in the mass
($m_{\gamma^{\prime}}$) and coupling ($K$) parameter space for various
experimental setups. For waveguides, we chose the same optical parameters and
the measurement time ($t_{s}=20$ days) as the ALPS II, which is the state-of-
the-art experiment for the high-power, low-signal measurement (see the
parameters and caveats in Table 1). Figure 7: A comparison of 1$\sigma$ C.L.
sensitivity limits for the waveguides of the $98.5\%$ (dashed curves) and
$100\%$ (solid curves) reflectivity coating. The signal loss from the
reflection is more critical when the length of waveguide is larger, since the
number of the reflection is proportional to the length.
Figure 6 shows the exclusion limits and the future sensitivities of various
experimental setups. The waveguide sensitivity curves are for the ideal,
100$\%$ reflectivity case. The comparison to the realistic case of the
$98.5\%$ reflectivity is shown in Fig. 7. We made a cutoff for each curve at
$m_{\gamma^{\prime}}=0.99\omega$. This is far below the spectrum of the laser
due to its linewidth. Also, the limit (10) is satisfied by this cutoff. Since
the decay rate of the dark photon (thus the signal) increases with its mass,
the experiment is particularly sensitive when the $m_{X}$ is close to the
laser frequency $\omega$, which can be seen from the negative slopes of
curves. In contrast, the subfrequency signal is more dispersive in the angular
domain as $m_{\gamma^{\prime}}\rightarrow\omega$. This reduces the signal
reaching the detector (see Fig. 3).
The new setup with a waveguide can excel the early LSW experiments (GammeV,
ALPS) by orders of magnitude. Even tabletop experiments ($L=1$ m) can
outperform them. ALPS II is an ongoing LSW experiment, which has the most
extended baseline ($L=100$ m). This long baseline clearly improves its
sensitivity, although it is not using a waveguide. However, the ALPS II gets
the most stringent degradation from the angular dispersion due to its large
baseline, as one can see from its gentle slope. As the curves show, using a
waveguide for a given $L$ significantly improves the sensitivity, especially
in the high mass region.
A rough estimate for the QCD axion can be done, if one assumes
$G_{a\gamma\gamma}\sim G_{a\gamma\gamma^{\prime}}$. The
experimental/observational bound is $G_{a\gamma\gamma}<10^{-10}$ GeV-1 for the
QCD axion of lighter than the eV scale Zyla _et al._ (2020). Our sensitivity
study in $K$ shown in Fig. 6 indicates the QCD axion cannot be detected with
the current technology even if we take the largest possible kinetic mixing
$\varepsilon\sim 1$.
The sensitivity would further increase if we can (i) increase the circulating
laser power, which is currently limited by the damage threshold of the mirrors
in the cavity (i.e., improve the mirror coating), (ii) reduce the detector
background noise, (iii) extend the baseline for the dark photon decay
(accompanied by a waveguide). It can also reach a higher mass using the higher
frequency light source. (For instance, see the mass coverage of GammeV and
ALPS, which use relatively high-frequency lasers.)
Especially, an x-ray source can cover high mass region where the decay rate is
large. In Fig. 8, we show the constraints from the null result of the light
shining through wall experiment SPring-8 using the x-ray Inada _et al._
(2013). Although a waveguide cannot be not used for the x-ray due to the high
transmittance, it already gives a sizable constraints.
Figure 8: 1$\sigma$ C.L. exclusion limits in the mass ($m_{\gamma^{\prime}}$)
and coupling ($K$) parameter space for SPring-8 experiments Inada _et al._
(2013). SPring-8 experiments consist of a series of LSW experiments with
various frequency x-ray sources. The sawtooth shape comes from the combination
of a few experiments.
One of the critical differences of the proposed experiment from the other
laser-based new physics searches is that it exploits the decay of a new
particle. Because of this property, it results in several differences such as
the angular deviation of the signal, need of sizeable decay length, no need of
wall in principle (we could use a prism instead of the wall), and different
signal frequency from the original laser frequency. These distinctions allow a
different way of optimization from the LSW, such as the waveguide we take.
## VII Summary and outlook
Our study shows a simple setup of an eV-scale optical laser and a photon
detector can do the entire process from the production of a new particle, its
decay into the SM particles, and the new physics signal detection. It was
considered possible only in high-energy accelerator experiments not long ago.
It is striking to note the whole setup can be as small as a tabletop
experiment of a 1-meter scale.
One of the crucial points is that the test of the combined portal
$K=\varepsilon G_{a\gamma\gamma^{\prime}}$ cannot be obtained just by
multiplying the bounds on $\varepsilon$ and $G_{a\gamma\gamma^{\prime}}$
obtained independently. For instance, if we use only $\varepsilon$, the dark
photon lighter than 1 MeV would not practically decay Pospelov _et al._
(2008). Also, the $\varepsilon$ bound from the stellar objects can be weakened
due to the reconversion of the dark photon with the combined portal. If we use
only $G_{a\gamma\gamma^{\prime}}$, on the other hand, then the dark photon
production will be greatly suppressed because it relies on an off-shell
photon. Therefore, we need a dedicated study for the implications and
constraints of the combined portal $K$. In this sense, it can be taken as a
practically new portal, which we name the “subfrequency portal.”
The essential part of this portal is that any source of light can produce
subfrequency light as a signal of new physics. While we focused on the new
physics search in the laser experiments, which cover the visible light
frequency, plus some x-ray examples, there are plenty of possible extensions
such as microwave cavities and isotope spectrum spectroscopy. Also, the
astrophysical/cosmological data from the stellar objects and cosmic microwave
backgrounds are wonderful places to study. Rich physics implications are
warranted.
###### Acknowledgements.
J. L. thanks J. Jung and H. Kim for helpful conversations about the optical
devices. This work was supported in part by the National Research Foundation
of Korea (Grant No. NRF-2021R1A2C2009718).
## References
* Aghanim _et al._ (2020) N. Aghanim _et al._ (Planck), Astron. Astrophys. 641, A6 (2020), [Erratum: Astron.Astrophys. 652, C4 (2021)], arXiv:1807.06209 [astro-ph.CO] .
* Essig _et al._ (2013) R. Essig _et al._ , in _Community Summer Study 2013: Snowmass on the Mississippi_ (2013) arXiv:1311.0029 [hep-ph] .
* Geraci _et al._ (2018) A. A. Geraci _et al._ (ARIADNE), Springer Proc. Phys. 211, 151 (2018), arXiv:1710.05413 [astro-ph.IM] .
* Bartram _et al._ (2021) C. Bartram _et al._ (ADMX), Phys. Rev. Lett. 127, 261803 (2021), arXiv:2110.06096 [hep-ex] .
* Melcón _et al._ (2020) A. A. Melcón _et al._ (CAST), JHEP 21, 075 (2020), arXiv:2104.13798 [hep-ex] .
* Trickle _et al._ (2020) T. Trickle, Z. Zhang, and K. M. Zurek, Phys. Rev. Lett. 124, 201801 (2020), arXiv:1905.13744 [hep-ph] .
* Ehret _et al._ (2010) K. Ehret _et al._ , Phys. Lett. B 689, 149 (2010), arXiv:1004.1313 [hep-ex] .
* Spector (2017) A. Spector (ALPS), in _12th Patras Workshop on Axions, WIMPs and WISPs_ (2017) pp. 133–136, arXiv:1611.05863 [physics.ins-det] .
* Robilliard _et al._ (2007) C. Robilliard, R. Battesti, M. Fouche, J. Mauchain, A.-M. Sautivet, F. Amiranoff, and C. Rizzo, Phys. Rev. Lett. 99, 190403 (2007), arXiv:0707.1296 [hep-ex] .
* Chou _et al._ (2008) A. S. Chou, W. C. Wester, III, A. Baumbaugh, H. R. Gustafson, Y. Irizarry-Valle, P. O. Mazur, J. H. Steffen, R. Tomlin, X. Yang, and J. Yoo (GammeV (T-969)), Phys. Rev. Lett. 100, 080402 (2008), arXiv:0710.3783 [hep-ex] .
* Afanasev _et al._ (2009) A. Afanasev, O. K. Baker, K. B. Beard, G. Biallas, J. Boyce, M. Minarni, R. Ramdon, M. Shinn, and P. Slocum, Phys. Lett. B 679, 317 (2009), arXiv:0810.4189 [hep-ex] .
* Ballou _et al._ (2015) R. Ballou _et al._ (OSQAR), Phys. Rev. D 92, 092002 (2015), arXiv:1506.08082 [hep-ex] .
* Cameron _et al._ (1993) R. Cameron _et al._ , Phys. Rev. D 47, 3707 (1993).
* Zyla _et al._ (2020) P. A. Zyla _et al._ (Particle Data Group), PTEP 2020, 083C01 (2020).
* Kim and Carosi (2010) J. E. Kim and G. Carosi, Rev. Mod. Phys. 82, 557 (2010), [Erratum: Rev.Mod.Phys. 91, 049902 (2019)], arXiv:0807.3125 [hep-ph] .
* Fabbrichesi _et al._ (2021) M. Fabbrichesi, E. Gabrielli, and G. Lanfranchi, _The Physics of the Dark Photon_ (Springer Cham, 2021).
* Holdom (1986) B. Holdom, Phys. Lett. B 166, 196 (1986).
* Kaneta _et al._ (2017) K. Kaneta, H.-S. Lee, and S. Yun, Phys. Rev. Lett. 118, 101802 (2017), arXiv:1611.01466 [hep-ph] .
* Peccei and Quinn (1977) R. D. Peccei and H. R. Quinn, Phys. Rev. Lett. 38, 1440 (1977).
* Weinberg (1978) S. Weinberg, Phys. Rev. Lett. 40, 223 (1978).
* Wilczek (1978) F. Wilczek, Phys. Rev. Lett. 40, 279 (1978).
* Choi _et al._ (2017) K. Choi, H. Kim, and T. Sekiguchi, Phys. Rev. D 95, 075008 (2017), arXiv:1611.08569 [hep-ph] .
* Kitajima _et al._ (2018) N. Kitajima, T. Sekiguchi, and F. Takahashi, Phys. Lett. B 781, 684 (2018), arXiv:1711.06590 [hep-ph] .
* deNiverville _et al._ (2018) P. deNiverville, H.-S. Lee, and M.-S. Seo, Phys. Rev. D 98, 115011 (2018), arXiv:1806.00757 [hep-ph] .
* Arias _et al._ (2021) P. Arias, A. Arza, J. Jaeckel, and D. Vargas-Arancibia, JCAP 05, 070 (2021), arXiv:2007.12585 [hep-ph] .
* Deniverville _et al._ (2021) P. Deniverville, H.-S. Lee, and Y.-M. Lee, Phys. Rev. D 103, 075006 (2021), arXiv:2011.03276 [hep-ph] .
* Hook _et al._ (2021) A. Hook, G. Marques-Tavares, and C. Ristow, JHEP 06, 167 (2021), arXiv:2105.06476 [hep-ph] .
* Ge _et al._ (2021) S.-F. Ge, X.-D. Ma, and P. Pasquini, Eur. Phys. J. C 81, 787 (2021), arXiv:2104.03276 [hep-ph] .
* Huang and Lee (2019) F. P. Huang and H.-S. Lee, Int. J. Mod. Phys. A 34, 1950012 (2019), arXiv:1806.09972 [hep-ph] .
* Bähre _et al._ (2013) R. Bähre _et al._ , JINST 8, T09001 (2013), arXiv:1302.5647 [physics.ins-det] .
* Ehret _et al._ (2009) K. Ehret _et al._ (ALPS), Nucl. Instrum. Meth. A 612, 83 (2009), arXiv:0905.4159 [physics.ins-det] .
* Inada _et al._ (2013) T. Inada, T. Namba, S. Asai, T. Kobayashi, Y. Tanaka, K. Tamasaku, K. Sawada, and T. Ishikawa, Physics Letters B 722, 301 (2013).
* Flambaum _et al._ (2019) V. V. Flambaum, I. B. Samsonov, and H. B. Tran Tan, Phys. Rev. D 99, 115019 (2019), arXiv:1904.02271 [hep-ph] .
* Bastidon _et al._ (2015) N. Bastidon, D. Horns, and A. Lindner (2015) arXiv:1509.02064 [physics.ins-det] .
* Ahlers _et al._ (2007) M. Ahlers, H. Gies, J. Jaeckel, J. Redondo, and A. Ringwald, Phys. Rev. D 76, 115005 (2007), arXiv:0706.2836 [hep-ph] .
* Lita _et al._ (2008) A. E. Lita, A. J. Miller, and S. W. Nam, Opt. Express 16, 3032 (2008).
* Niwa _et al._ (2017) K. Niwa, T. Numata, K. Hattori, and D. Fukuda, Scientific Reports 7 (2017), 10.1038/srep45660.
* Lita _et al._ (2010) A. Lita, B. Calkins, L. Pellouchoud, A. Miller, and S. Nam, Proceedings of SPIE - The International Society for Optical Engineering 7681 (2010), 10.1117/12.852221.
* Pospelov _et al._ (2008) M. Pospelov, A. Ritz, and M. B. Voloshin, Phys. Rev. D 78, 115012 (2008), arXiv:0807.3279 [hep-ph] .
|
# Modeling of dendritic solidification and numerical analysis of the phase-
field approach to model complex morphologies in alloys
Kunal Bhagat Department of Mechanical Engineering, University of Wisconsin-
Madison Shiva Rudraraju
Department of Mechanical Engineering, University of Wisconsin-Madison;
corresponding author
(Preprint)
## Abstract
Dendrites are one of the most widely observed patterns in nature and occur
across a wide spectrum of physical phenomena. In solidification and growth
patterns in metals and crystals, the multi-level branching structures of
dendrites pose a modeling challenge, and a full resolution of these structures
is computationally demanding. In the literature, theoretical models of
dendritic formation and evolution, essentially as extensions of the classical
moving boundary Stefan problem exist. Much of this understanding is from the
analysis of dendrites occurring during the solidification of metallic alloys,
as this is critical for understanding microstructure evolution during metal
manufacturing processes (MAM) that involve solidification of a liquid melt.
Motivated by the problem of modeling microstructure evolution from liquid
melts of pure metals and alloys during MAM, we developed a comprehensive
numerical framework for modeling a large variety of dendritic structures that
are relevant to metal solidification. In this work, we present a numerical
framework encompassing the modeling of Stefan problem formulations relevant to
dendritic evolution using a phase-field approach and a finite element method
implementation. Using this framework, we model numerous complex dendritic
morphologies that are physically relevant to the solidification of pure melts
and binary alloys. The distinguishing aspects of this work are - a unified
treatment of both pure-metals and alloys; novel numerical error estimates of
dendritic tip velocity; and the convergence of error for the primal fields of
temperature and the order parameter with respect to numerical discretization.
To the best of our knowledge, this is a first of its kind study of numerical
convergence of the phase-field equations of dendritic growth in a finite
element method setting. Further, we modeled various types of physically
relevant dendritic solidification patterns in 2D and 3D computational domains.
## 1 Introduction
Dendrites are tree-like patterns with complex multi-level branch structures
that are observed across a wide spectrum of physical phenomena - from snow
flakes to river basins; from bacterial colonies to lungs and vascular systems;
and ubiquitously in solidification and growth patterns in metals and crystals.
In the context of solidification problems involving a pure metal or metallic
alloys, dendritic “trees” with primary and secondary branches grow from a
nucleation point or a field perturbation. These solidification dendrites often
initiate at random nucleation points or at domain boundaries at the initial
stage, followed by anisotropic interface growth during the intermediate stage,
eventually leading to grain formation and coarsening. The theoretical
foundations of dendritic solidification lie in the classical Stefan problem, a
moving boundary problem, that describes the evolution of a solid-liquid phase
front [1]. The two-phase interface motion is obtained by solving the heat
equation in each phase, coupled with an evolving interface boundary condition
(Stefan condition) that explicitly sets the velocity of the moving interface.
While the evolution of dendrites in pure melts with no dissolved solutes is
well represented by the Stefan problem, for melts with dissolved solutes (like
alloys), mass diffusion in the constituent phases should also be accounted for
in the governing equations.
Many analytical and numerical approaches exist for treating the Stefan
problem, albeit under various simplifying assumptions on the problem geometry,
interface geometry and boundary conditions. In the numerical domain, front
tracking methods [2, 3] are useful in solving straight and curved interface
evolution in one-dimensional problems. Other popular methods that explicitly
track moving interface are the Landau transformation and the finite element
mesh-based moving node techniques. These methods are more suited when the
movement of the interface is not far from the initial position [4], but
bookkeeping of the interface movement can be an arduous task, especially for
anisotropic interphase growth conditions. The level set method is another
popular numerical technique that is extensively used to solve moving interface
problems, including dendrite solidification using the Stefan problem [5].
The phase-field method has gained popularity in the last two decades as an
alternate diffuse interface numerical technique to model solidification
processes. The phase-field model modifies the equations represented by the
Stefan problem by introducing an order parameter to distinguish between the
solid and liquid phases. Further, the previously sharp interface is
represented as a diffuse interface using this order parameter without
sacrificing much of the accuracy. The earliest of the simplified isotropic
phase-field model has been proposed by Fix [6], and Collins and Levine [7].
Caginalp [8, 9] and Kobayashi [10] introduced basic anisotropy into the phase-
field model and numerically predicted dendritic patterns seen in the
solidification of the pure melt. They also modeled the formation of the side
branching by introducing noise into the phase-field models.
The potential of the so far surveyed phase-field models was limited due to
constraints on the lattice size, undercooling, interface width, capillary
length, and interface kinetics. A modified thin interface phase-field model
proposed by Karma and Rappel [11] overcame these limitations. The dendrite
shape and tip velocities of the pure melt obtained from the thin interface
phase-field model were in excellent agreement with the steady-state results
reported in theoretical studies. Karma and Rappel [12] studied dendrite side-
branching by introducing a microscopic thermal noise, and the side branch
characteristics were in line with the linear WKB theory based stability
analysis. Plapp and Karma [13] proposed a hybrid technique where the phase-
field model and diffusion-based Monte-Carlo algorithm were used in conjunction
to simulate dendritic solidification at the low undercooling in two and three-
dimensional geometry.
The phase-field model for the solidification of a binary alloy was presented
by Warren and Boettinger [14]. Realistic patterns showing the primary and the
secondary arms of the dendrite were simulated using the interface thickness
far different from the asymptotic limit of the sharp interface models.
Loginova et al. [15] was the first to include the non-isothermal effects in
binary alloy solidification. However, the computations were time-consuming. An
artificial solute trapping was observed in the simulations with higher
undercooling. This artificial solute trapping was the result of diffuse
interface thickness. Karma [16] developed a modified thin interface phase-
field model for alloy solidification. Compared to the previous models, this
model was applicable for varied interface thicknesses, zero non-equilibrium
effects at the interface, and nearly zero diffusivity in the solid.
The thin interface phase-field model by Ramirez et al. [17] considered both
heat and solute diffusion along with zero kinetics at the interface. The
interface thickness used was an order of magnitude less than the radius of
curvature of the interface. Echebarria et al. [18] modeled numerous numerical
test cases and studied the convergence of the thin interface phase-field
model. They discussed the final form of anti-trapping current that is often
used in many alloy solidification models [16, 17, 19]. When the solid
diffusivity of the alloys is non-negligible, the form of anti-trapping current
previously used needs modification and this was rigorously derived in the work
of Ohno and Matsuura [20]
With the substantial review tracing Stefan problem and important theoretical
development of the phase-field model, we review literature focusing on the
numerical implementations of these models and error analysis. Several
important studies leveraged the finite-element and finite-difference method
based numerical schemes and their application to the phase-field model of
solidification. These studies focused on the accuracy and stability of
proposed numerical schemes. In this regard, the work of Feng and Prohl [21] is
important. They made use of fully discrete finite element methods and obtained
optimal error bounds in relation to the interface thickness for a phase-field
method of solidification. Using the error estimates they showed convergence of
finite element schemes to the solution to the phase field models in the limit
of the sharp interface. Gonzalez-Ferreiro et al. [22] presented a finite
element discretization in space and mid-point discretization in time of a
phase-field model that is consistent with both the thermodynamics laws. The
thermodynamically consistent numerical model was developed to have better
dendrite resolution and higher order accuracy in time. A first-order accurate
in time and energy stable numerical method was presented by Chen and Yang
[23]. The proposed method was applied to coupled Allen-Cahn, heat diffusion,
and modified Navier-Stokes equations. They resort to techniques that decoupled
these three equations with the use of implicit-explicit schemes in their
numerical implementation.
Better spatial and temporal accuracy and smaller error estimates were reported
in the work of Kessler and Scheid [24]. They applied a finite element method
to the phase-field model of binary alloy solidification where the error
convergence numerical tests were done on a physical example of Ni-Cu alloy
solidification with a simpler and non-branched solidification structure. The
adaptive meshing technique was used in many numerical studies of
solidification using a phase-field model. Hu et al. [25] presented a multi-
mesh adaptive finite-element based numerical scheme to solve the phase-field
method for the pure-melt solidification problem. The accuracy of their method
was tested with its ability to predict dendrite tip velocities with a range of
undercooling conditions. Work of Rosam et al. [26] using the mesh and time
adaptive fully implicit numerical scheme utilizing finite-difference method
applied to binary alloy solidification is very relevant. Such highly space-
time adaptive techniques saved computational time. Their focus was on the
implicit and explicit schemes error estimates using only derived variables
such as dendrite tip position, radius, and velocity. However, error analysis
reported in the literature did not explore the effect of basis continuity
(i.e, $C^{n}$-continuous basis) on capturing dendrite kinetics and
morphologies, and the convergence studied reported did not explicitly study
the error in the primal fields, i.e., temperature and phase-field order
parameter. As part of this manuscript, these two aspects of error analysis
will also be addressed.
A review by Tourret et al. [27] highlights several studies that applied the
phase-field methods to model solidification and obtained good comparisons with
the experimental results. Present challenges and extensions of phase-field
models were also discussed in their review. Wang et al. [28] discovered the
relation between lower and upper limit primary dendrite arm spacing to inter-
dendritic solute distribution and inter-dendritic undercooling by solving
phase-field models using finite-element method. Fallah et al. [29] work showed
the suitability of phase-field models coupled with heat transfer models to
reproduce experimentally known complex dendrite morphology and accurate
dendrite size of Ti-Nb alloys thereby demonstrating the suitability of the
numerical methods to laser deposition and industrial scale casting process.
More recently, microstructure evolution processes involving competitive grain
growth of columnar dendrites and the grain growth along the converging grain
boundary were modeled using the phase-field model by Tourret and Karma [30],
and Takaki et al. [31]. In manufacturing processes like welding and molding,
phase-field models were used to study solidification cracking susceptibility
in Al-Mg alloys by Geng et al. [32], and dendrite morphology in the melt pool
of Al-Cu alloys by Farzadi et al. [33]. Integrated phase-field model and
finite element methods are also used to study microstructure evolution
especially formation of laves phase in additive manufacturing of IN718 [34].
Phase-field models finds extensive applications in modeling rapid
solidification conditions prevalent in the metal additive manufacturing [35,
36, 37, 38, 39].
A lot of existing studies on solidification and related phase-field models
have focused only on either pure metals or alloys. However, in this work, we
present a unified treatment of both pure metals and alloys. We discuss
classical Stefan problems relevant to both these types of solidification, and
present their phase-field formulations and numerical implementations. Further,
we present novel numerical error estimates of dendritic tip velocity, and the
convergence of error for the primal fields of temperature and order parameter
with respect to the numerical discretization. Lastly, using this numerical
framework, various types of physically relevant dendritic solidification
patterns like single equiaxed, multi-equiaxed, single columnar and multi-
columnar dendrites are modeled in two-dimensional and three-dimensional
computational domains.
Following this literature review and summary of the advancements in numerical
modeling of solidification and dendritic growth, we now present an overview of
this manuscript. In Section 2 we present a detailed discussion of the
numerical models used in this work to model dendritic growth. Essentially, we
look at the appropriate Stefan problem formulations and their phase-field
representations for modeling the solidification problem of pure melts and
binary alloys. Then, in Section 3, we discuss the numerical framework and its
computational implementation. Further, we present a numerical error analysis
and the simulation results of various 2D and 3D dendritic solidification
problems. Finally, we present the concluding remarks in Section 4.
## 2 Numerical models of dendritic growth
In this section, we discuss the Stefan problem for modeling solidification of
a pure metal and a binary alloy. We elaborate on the complexities of the
Stefan problem and the physics associated with solidification. The Stefan
problem is then re-written as a phase-field formulation. The primary goal of
using a phase-field approach is to model the interface growth kinetics,
particularly the velocity of interface motion and the complex morphologies of
dendrites that occur during solidification, without the need to explicitly
track the dendritic interfaces. The interface growth can be influenced by the
surface anisotropy, heat diffusion, mass diffusion, interface curvature, and
the interface attachment kinetics. In Sections 2.1-2.2, the classical Stefan
problem and its phase-field representation relevant to solidification of pure
melts is presented. Then, in Sections 2.3-2.4, the extensions needed in the
Stefan problem and its phase-field representation for modeling solidification
of binary alloys are discussed.
In general, a pure melt is a single-component material (pure solvent) without
any solutes. Addition of one or more solute components results in an alloy. In
this work, we specifically consider a binary alloy (two-component alloy with a
solvent and one solute), as a representative of multi-component alloys. And as
will be shown, during the solidification of a binary alloy, we model the
diffusion of this solute concentration in the solid and liquid phases.
### 2.1 Stefan problem for modeling solidification of a pure melt
The Stefan problem describing the solidification of an undercooled pure metal
is presented. The set of Equations 1a-1c mathematically model the
solidification of an undercooled pure melt. Equations 2a-2c represent the same
governing equations in a non-dimensional form. Equation 1a models heat
conduction in the bulk liquid and the bulk solid regions of the pure metal. At
the solid-liquid interface, the balance of the heat flux from either of the
bulk regions is balanced by the freezing of the melt, and thus additional
solid bulk phase is formed. In other words, the solid-liquid interface moves.
The energy balance at the interface is given by the Equation 1b. The
temperature of the interface is not fixed and can be affected by multiple
factors such as the undercooling effect on the interface due to its curvature
(Gibbs-Thomson effect) and the interface attachment kinetics. These are
captured in the equation 1c.
$\frac{\partial T(\bm{x},t)}{\partial
t}=\frac{k}{c_{p}}\nabla^{2}T,\quad\bm{x}\in\Omega^{s},\Omega^{l}$ (1a)
$\bm{\nu}_{n}L=k\big{(}\partial_{n}T(\bm{x})\lvert^{+}-\;\partial_{n}T(\bm{x})\lvert^{-}\big{)},\quad\bm{x}\in\Gamma$
(1b) $T_{m}-T_{i}=\Delta T_{\Gamma_{g}}+\Delta
T_{\mu_{k}}=\Gamma_{g}\bm{\kappa}+\frac{\bm{\nu}_{n}}{\mu_{k}},\quad\bm{x}\in\Gamma$
(1c)
Here, $k$ and $c_{p}$ are thermal conductivity and specific heat capacity for
a pure solid or liquid metal. $T_{m}$ $\bm{\nu}_{n}$, $L$ is the melting
temperature, interface velocity, and latent heat of the metal.
$\partial_{n}T\lvert^{+}$ and $\partial_{n}T\lvert^{-}$ is the temperature
gradient normal to the interface in the liquid and the solid phase
respectively. $\Gamma_{g}$, $\bm{\kappa}$, and $\mu_{k}$ are the Gibbs-Thomson
coefficient, the curvature of the interface, and interface-attachment
coefficient, respectively. We re-write these equations in their non-
dimensional form, using the scaled temperature $u=\frac{c_{p}(T-T_{m})}{L}$
[4].
$\frac{\partial u(\bm{x},t)}{\partial
t}=D\nabla^{2}u,\quad\bm{x}\in\Omega^{s},\Omega^{l}$ (2a)
$\bm{\nu}_{n}=D\big{(}\partial_{n}u(\bm{x})\lvert^{+}-\;\partial_{n}u(\bm{x})\lvert^{-}\big{)},\quad\bm{x}\in\Gamma$
(2b)
$u^{*}=-d(\mathbf{n})\bm{\kappa}-\beta(\mathbf{n})\bm{\nu}_{n},\quad\bm{x}\in\Gamma$
(2c)
$D=\frac{k}{c_{p}}$ is the non-dimensional thermal diffusivity in the solid
and liquid phases. $d(\mathbf{n})=\gamma(\mathbf{n})T_{m}c_{p}/L^{2}$ is the
capillary length, $\gamma(\mathbf{n})$ is the surface tension, $\mathbf{n}$ is
the unit vector denoting normal to the interface, $\beta(\mathbf{n})$ is the
kinetic coefficient, and $\partial_{n}u\lvert^{+}$ and
$\partial_{n}u\lvert^{-}$ are the non-dimensional derivatives normal to the
interface in the solid and liquid region, respectively [40]. The driving force
for this solidification is the initial undercooling, i.e., the initial
temperature of the liquid melt below its melting temperature. The non-
dimensional undercooling is given by $\Delta=-c_{p}(\frac{Tm-T_{\infty}}{L})$.
The Stefan problem described using Equation 1 does not have a known analytical
solution, and its numerical implementation, as is the case with free boundary
problems, is challenging. Numerical methods like time-dependent boundary
integral formulations [11], a variational algorithm with zero or non-zero
interface kinetics, and methods that require the book-keeping of the
solidifying front and low grid anisotropy conditions have been used in the
past. Modeling correct dendritic growth involves getting the correct operating
dendritic tip conditions, mainly the tip radius and the tip velocity.
Variationally derived phase-field models, like the one described in Section
2.2, are now the method of choice to model complex dendritic solidification
problems.
### 2.2 Phase-field model describing the solidification of a pure melt
In this section, we describe the phase-field method developed by Plapp and
Karma [13] to model solidification in a pure metal. The governing equations
for the model are derived using variational principles. A phenomenological
free energy expression, $\Pi\left[\phi,u\right]$, in terms of a phase-field
order parameter ($\phi$) and temperature ($u$), is defined in Equation 3. This
functional form defines the thermodynamic state of a system, and has two terms
- first is $f(\phi,u)$, the bulk energy term, and second is an interface term
given by $\frac{1}{2}\lambda^{2}(\bm{n})|\bm{\nabla}\phi|^{2}$. The magnitude
of the interface term is controlled by the interface parameter, $\lambda$, and
the interface term is positive in the diffuse solid-liquid interface due to a
non-zero gradient of the order parameter; elsewhere, the interface term is
zero.
$\Pi\left[\phi,u\right]=\int_{\Omega}\left[f(\phi,u)+\frac{1}{2}\lambda^{2}(\mathbf{n})|\bm{\nabla}\phi|^{2}\right]\leavevmode\nobreak\
dV$ (3a) $f(\phi,u)=-\frac{1}{2}\phi^{2}+\frac{1}{4}\phi^{4}+\xi
u\phi\left(1-\frac{2}{3}\phi^{2}+\frac{1}{5}\phi^{4}\right)$ (3b)
The bulk energy term $f(\phi,u)$ in Equation 3b is given by the double well
potential that has a local minima at $\phi=1$ and $\phi=-1$.
$\Omega=\Omega_{s}\cup\Omega_{l}$ represents total volume encompassing the
bulk solid, the bulk liquid and the interface region. The value of $u$ then
tilts the equilibrium value of $f(\phi,u)$. The form of $f(\phi,u)=g(\phi)+\xi
uh(\phi)$ is given below. $\xi$ is the coupling parameter in the double well
function. The governing equations for the phase-field model of solidification
are obtained by minimizing the above functional with respect to the primal
fields, $\phi$ and $u$. This results in the partial differential equations
presented below:
$\frac{\partial u(\bm{x},t)}{\partial
t}=\bm{\nabla}.\Big{(}D\bm{\nabla}\frac{\delta\Pi}{\delta
u}\Big{)},\quad\bm{x}\in\Omega$ (4a) $\frac{\partial\phi(\bm{x},t)}{\partial
t}=-K_{\phi}\frac{\delta\Pi}{\delta\phi},\quad\bm{x}\in\Omega$ (4b)
Taking the variational derivatives results in,
$\frac{\partial u}{\partial
t}=D\nabla^{2}u+\frac{1}{2}\frac{\partial\phi}{\partial
t},\quad\bm{x}\in\Omega$ (5a)
$\displaystyle\tau(\bm{n})\frac{\partial\phi}{\partial t}=-\frac{\partial
f}{\partial\phi}+\bm{\nabla}\cdot\big{(}\lambda^{2}(\bm{n})\bm{\nabla}\phi\big{)}+\frac{\partial}{\partial\bm{x}}\left[|\bm{\nabla}\phi|^{2}\lambda(\bm{n})\frac{\partial\lambda(\bm{n})}{\partial\left(\frac{\partial\phi}{\partial\bm{x}}\right)}\right]+\frac{\partial}{\partial\bm{y}}\left[|\bm{\nabla}\phi|^{2}\lambda(\bm{n})\frac{\partial\lambda(\bm{n})}{\partial\left(\frac{\partial\phi}{\partial\bm{y}}\right)}\right],\bm{x}\in\Omega$
(5b)
where $\Omega$ is the problem domain. Anisotropy in the interface energy
representation is considered. Thus, $\lambda(\bm{n})=\lambda_{0}a_{s}(\bm{n})$
and $\tau(\bm{n})=\tau_{0}a_{s}^{2}(\bm{n})$, where anisotropy parameter is
defined as
$a_{s}=(1-3\epsilon_{4})\Big{[}1+\frac{4\epsilon_{4}}{(1-3\epsilon_{4})}\frac{(\partial_{x}\phi)^{4}+(\partial_{y}\phi)^{4}}{|\phi|^{4}}\Big{]}=1+\epsilon_{4}cos(m\theta)$.
The equivalence is easily established by taking
$\tan(\theta)=\frac{\partial\phi}{\partial y}/\frac{\partial\phi}{\partial
x}$. $\epsilon_{4}$ is the strength of anisotropy and m=4 corresponds to an
anisotropy with four-fold symmetry. The interface between the solid and liquid
phases is diffuse in the phase-field representation of a Stefan problem of
solidification, but in the asymptotic limit of $\lambda(\bm{n})\to 0$, the
Stefan problem is recovered [8, 9].
#### 2.2.1 Weak formulation
We now pose the above governing equations in their weak (integral) form. This
formulation is used to solve these equations within a standard finite element
method framework.
Find the primal fields $\\{u,\phi\\}$, where,
$\displaystyle u$
$\displaystyle\in\mathscr{S}_{u},\quad\mathscr{S}_{u}=\\{u\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ u=\leavevmode\nobreak\ u^{\prime}\leavevmode\nobreak\
\forall\leavevmode\nobreak\ \textbf{X}\in\Gamma^{u}\\},$ $\displaystyle\phi$
$\displaystyle\in\mathscr{S}_{\phi},\quad\mathscr{S}_{\phi}=\\{\phi\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ \phi=\leavevmode\nobreak\
\phi^{\prime}\leavevmode\nobreak\ \forall\leavevmode\nobreak\
\textbf{X}\in\Gamma^{\phi}\\}$
such that,
$\displaystyle\forall\leavevmode\nobreak\ w_{u}$
$\displaystyle\in\mathscr{V}_{u},\quad\mathscr{V}_{u}=\\{w_{u}\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ w_{u}=0\leavevmode\nobreak\ \forall\leavevmode\nobreak\
\textbf{X}\in\Gamma^{u}\\},$ $\displaystyle\forall\leavevmode\nobreak\
w_{\phi}$
$\displaystyle\in\mathscr{V}_{\phi},\quad\mathscr{V}_{\phi}=\\{w_{\phi}\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ w_{\phi}=0\leavevmode\nobreak\
\forall\leavevmode\nobreak\ \textbf{X}\in\Gamma^{\phi}\\}$
we have,
$\int_{\Omega}w_{u}\Big{(}\frac{\partial u}{\partial
t}-\frac{1}{2}\frac{\partial\phi}{\partial
t}\Big{)}\,dV+\int_{\Omega}D\big{(}\bm{\nabla}w_{u}.\bm{\nabla}u\big{)}\,dV=0$
(6a)
$\displaystyle\int_{\Omega}w_{\phi}\Big{(}\tau\frac{\partial\phi}{\partial
t}+\frac{\partial
f}{\partial\phi}\Big{)}\,dV+\int_{\Omega}\lambda^{2}\bm{\nabla}w_{\phi}.\bm{\nabla}\phi\,dV+\int_{\Omega}|\bm{\nabla}\phi|^{2}\lambda\frac{\partial
w_{\phi}}{\partial\bm{x}}\frac{\partial\lambda}{\partial\left(\frac{\partial\phi}{\partial\bm{x}}\right)}\,dV$
(6b) $\displaystyle+\int_{\Omega}|\bm{\nabla}\phi|^{2}\lambda\frac{\partial
w_{\phi}}{\partial\bm{y}}\frac{\partial\lambda}{\partial\left(\frac{\partial\phi}{\partial\bm{y}}\right)}\,dV=0$
here, $\Omega=\Omega_{s}\cup\Omega_{l}$, where $\Omega_{s}$ is the solid phase
and $\Omega_{l}$ is the liquid phase, respectively. The variations,
$\omega_{u}$ and $\omega_{\phi}$, belong to $\text{H}^{1}(\Omega)$ \- the
Sobolev space of functions that are square-integrable and have a square-
integrable derivatives.
### 2.3 Stefan problem for modeling solidification of a binary alloy
The Stefan problem describing the solidification of a binary alloy is
described in this section. As mentioned earlier, an alloy is a multi-component
system. A binary alloy is a specific case of a multi-component alloy with only
one solute component. A binary alloy consists of a solute that has a
composition of $c$, and a solvent that has a composition of $(1-c)$. During
the solidification of a binary alloy, we model the diffusion of the solute
concentration across the solid and the liquid phase. Mass diffusion is
governed by Equation 7a. Near the vicinity of the solid-liquid interface, the
composition of the solute in the solid and the liquid phase is dictated by the
binary alloy phase diagram given in the Figure 1a. The difference in the
concentration of the solute across the solid and liquid phase is balanced by
the mass diffusion flux. This is captured in Equation 7b. The temperature of
the interface is not fixed and can be affected by multiple factors. The
undercooling on the interface can be due to its curvature (Gibbs-Thomson
effect), the interface attachment kinetics, cooling rate ($\dot{T}$), and the
applied temperature gradient. This is mathematically represented in Equation
7c. Using the phase-diagram, we can write this condition in-terms of
composition, $c_{l}$ at the interface. We neglect the contribution from the
attachment kinetics.
$c_{l}=c^{0}_{l}-\frac{\Gamma_{g}\bm{\kappa}+Gy+\dot{T}t}{|m|}$ where $c$ is
the solute composition in either the solid or the liquid phase. $D$ is the
mass diffusivity of alloy in the liquid phase, $c_{l}$ and $c^{0}_{l}$ are the
solute concentration at the interface for the temperature $T_{i}$ and $T_{0}$
respectively. $\dot{T}=G\bm{\nu}_{p}$ is the cooling rate. We scale the solute
concentration into a variable $u_{c}=\frac{c-c^{0}_{l}}{c^{0}_{l}(1-k)}$ which
represents supersaturation. k is the partition coefficient which represents
the ratio of solid composition to the liquid composition.
$\frac{\partial c(\bm{x},t)}{\partial
t}=D\nabla^{2}c(\bm{x},t),\quad\bm{x}\in\Omega^{s},\Omega^{l}$ (7a)
$c_{l}(1-k)\bm{\nu}_{n}=-D\partial_{n}c(\bm{x},t)\lvert^{+},\quad\bm{x}\in\Gamma$
(7b)
$T_{m}-T_{i}=\Gamma\bm{\kappa}+\frac{\bm{\nu}_{n}}{\mu_{k}}+Gy+\dot{T}t,\quad\bm{x}\in\Gamma$
(7c)
here, $\Gamma$ is the solid-liquid interface that separates the solid phase,
$\Omega^{s}$, and the liquid phase, $\Omega^{l}$.
(a) (b)
Figure 1: Directional solidification in a binary alloy. Shown are (1a) a
simplified phase diagram of a binary alloy, and (1b) a schematic of a
directional solidification process. In the case of directional solidification,
typically, the melt is forced to solidify at a constant pulling velocity under
a fixed temperature gradient. The dendritic structures in this case evolve at
the solid-liquid interface, via a Mullins-Sekerka instability, when the
pulling velocity is above a critical value.
Rewriting the Equations 7a, 7b and 7c in-terms of the supersaturation,
$u_{c}$, we get,
$\frac{\partial u_{c}(\bm{x},t)}{\partial
t}=D\nabla^{2}u_{c}(\bm{x},t),\quad\bm{x}\in\Omega^{s},\Omega^{l}$ (8a)
$(1+(1-k)u^{*}_{c})\bm{\nu}_{n}=-D\partial_{n}u_{c}(\bm{x},t)\lvert^{+},\quad\bm{x}\in\Gamma$
(8b) $u^{*}_{c}=-d_{0}\bm{\kappa}+\dot{\theta}t-\gamma z,\quad\bm{x}\in\Gamma$
(8c)
Here $d_{0}=\frac{\Gamma_{g}}{\Delta T_{0}}$ is the chemical capillary length
, $\dot{\theta}=\frac{-\dot{T}}{\Delta T_{0}}$ is the non-dimensional cooling
rate and $\gamma=\frac{G}{\Delta T_{0}}$ is non-dimensional thermal gradient
and $\Delta T_{0}=|m|(1-k)c^{0}_{l}$ [41].
As can be expected, an analytical solution for the moving boundary problem in
Equations 8 is not known. In the next subsection, we discuss a phase-field
representation of this problem that is extensively used to model dendritic
solidification of binary alloys.
### 2.4 Phase-field model describing the solidification of a binary alloy
In this section, we describe the phase-field model for the solidification of a
binary alloy. Following the variational procedure adopted in the case of a
pure melt, we write the free energy functional form, $\Pi[\phi,c,T]$, for a
binary alloy. The free energy in this case has the additional dependence on
the solute composition, $c$, along with the order paramter, $\phi$, and the
temperature, $T$. This functional form, given by Equation 9, takes into
account bulk free energy of the system $f(\phi,T_{m})$, internal energy and
entropy $f_{AB}(\phi,c,T)$, and the interface term
$\frac{1}{2}\sigma|\bm{\nabla}\phi|^{2}$, where $\sigma\propto\lambda^{2}$.
$\Pi[\phi,c,T]=\int_{\Omega}\left[f(\phi,T_{m})+f_{AB}(\phi,c,T)+\frac{1}{2}\sigma|\bm{\nabla}\phi|^{2}\right]\leavevmode\nobreak\
dV$ (9)
here $\Omega=\Omega_{s}\cup\Gamma\cup\Omega_{l}$ represents the total volume
encompassing the bulk solid, the bulk liquid and the interface region. The
governing equations for the model are obtained by writing the diffusion
equations relevant to a conserved quantity - the solute composition, and a
non-conserved quantity - the order-parameter.
$\frac{\partial c(\bm{x},t)}{\partial
t}=\bm{\nabla}.\Big{(}M(\phi,c)\bm{\nabla}\frac{\delta\Pi}{\delta
c}\Big{)},\quad\bm{x}\in\Omega$ (10a) $\frac{\partial\phi(\bm{x},t)}{\partial
t}=-K_{\phi}\frac{\delta\Pi}{\delta\phi},\quad\bm{x}\in\Omega$ (10b) where
$M(\phi,c)$ is the solute mobility, and $K_{\phi}$ is the order parameter
mobility. It is convenient to scale the composition, c and write it as
$u_{c}=\frac{1}{1-k}\big{(}\frac{2c}{c^{0}_{l}(1-\phi+k(1+\phi))}-1\big{)}$.
The parameter $u_{c}$ can be understood as a concentration undercooling
parameter analogous to $u$ in Section 2.2. This change of variable enables us
to draw similarities with the earlier form of the phase-field model described
in Equations 5. $u_{c}$ is a continuous variable across the solid and liquid
region unlike the solute concentration, $c$ which is discontinuous. We present
the final form of the phase-field model in Equations 10c-10d, without the
accompanying mathematical derivation. For interested readers, a detailed
derivation can be found in [18].
$\displaystyle\Big{(}\frac{1+k}{2}-\frac{1-k}{2}\phi\Big{)}\frac{\partial
u_{c}}{\partial
t}=\bm{\nabla}\cdot\Big{(}\tilde{D}\big{(}\frac{1-\phi}{2}\big{)}\bm{\nabla}u_{c}-\bm{J}_{at}\Big{)}+\Big{(}1+(1-k)u_{c}\Big{)}\frac{1}{2}\frac{\partial\phi}{\partial
t},\quad\bm{x}\in\Omega$ (10c)
$\displaystyle\begin{split}\Big{(}1-(1-k)\frac{y-\tilde{\bm{\nu}}t}{\tilde{l}_{T}}\Big{)}a^{2}_{s}(\mathbf{n})\frac{\partial\phi}{\partial
t}=\bm{\nabla}\cdot\left(a^{2}_{s}(\mathbf{n})\bm{\nabla}\phi\right)+\frac{\partial}{\partial\bm{x}}\Big{(}|\bm{\nabla}\phi|^{2}a_{s}(\mathbf{n})\frac{\partial
a_{s}(\mathbf{n})}{\partial\left(\frac{\partial\phi}{\partial\bm{x}}\right)}\Big{)}\\\
+\frac{\partial}{\partial\bm{y}}\Big{(}|\bm{\nabla}\phi|^{2}a_{s}(\mathbf{n})\frac{\partial
a_{s}(\mathbf{n})}{\partial\left(\frac{\partial\phi}{\partial\bm{y}}\right)}\Big{)}+\phi-\phi^{3}-\xi(1-\phi^{2})^{2}\big{(}u_{c}+\frac{y-\tilde{\bm{\nu}}t}{\tilde{l}_{T}}\big{)}\end{split},\quad\bm{x}\in\Omega$
(10d)
The solute diffusivity in the solid region is small as compared to the solute
diffusivity in the liquid region. Thus mass diffusion is neglected in the
solid region. $\tilde{D}$ is the non-dimensional diffusivity in the liquid
phase. In the directional solidification process, dendrite growth is
influenced by the presence of a temperature gradient $G$ and pulling velocity
$\tilde{\bm{\nu}}_{p}$. This can be realized by the the term
$\frac{y-\tilde{\bm{\nu}}t}{\tilde{l}_{T}}$ in the Equation 10d, where
$\tilde{l}_{T}=\frac{|m|(1-k)c^{0}_{l}}{G}$ is called the thermal length. A
modified form of Equation 10c was suggested by Ohno and Matsuura[20] where the
mass diffusion in the solid phase is not neglected, and $D_{s}$ is comparable
to $D_{l}$. This results in the modification to the diffusivity part and anti-
trapping flux current as seen in the Equation 10e. If we substitute k=1 and
$D_{s}=D_{l}$ in this equation, Equation 10c is recovered.
$\begin{split}\Big{(}\frac{1+k}{2}-\frac{1-k}{2}\phi\Big{)}\frac{\partial
u_{c}}{\partial
t}=\bm{\nabla}\cdot\Bigg{(}\tilde{D}\big{(}\frac{1-\phi}{2}+k\frac{1+\phi}{2}\frac{D_{s}}{D_{l}}\big{)}\bm{\nabla
u_{c}}-(1-k\frac{D_{s}}{D_{l}})\bm{J}_{at}\Bigg{)}\\\
+\Big{(}1+(1-k)u_{c}\Big{)}\frac{1}{2}\frac{\partial\phi}{\partial
t}\end{split}$ (10e)
#### 2.4.1 Weak formulation
We now pose the above governing equations in their weak (integral) form. This
formulation is used to solve these equations within a standard finite element
framework.
Find the primal fields $\\{u_{c},\phi\\}$, where,
$\displaystyle u_{c}$
$\displaystyle\in\mathscr{S}_{u_{c}},\quad\mathscr{S}_{u_{c}}=\\{u_{c}\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ u_{c}=\leavevmode\nobreak\
u_{c}^{\prime}\leavevmode\nobreak\ \forall\leavevmode\nobreak\
\textbf{X}\in\Gamma^{u_{c}}\\},$ $\displaystyle\phi$
$\displaystyle\in\mathscr{S}_{\phi},\quad\mathscr{S}_{\phi}=\\{\phi\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ \phi=\leavevmode\nobreak\
\phi^{\prime}\leavevmode\nobreak\ \forall\leavevmode\nobreak\
\textbf{X}\in\Gamma^{\phi}\\}$
such that,
$\displaystyle\forall\leavevmode\nobreak\ w_{u_{c}}$
$\displaystyle\in\mathscr{V}_{u_{c}},\quad\mathscr{V}_{u_{c}}=\\{w_{u_{c}}\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ w_{u_{c}}=0\leavevmode\nobreak\
\forall\leavevmode\nobreak\ \textbf{X}\in\Gamma^{u_{c}}\\},$
$\displaystyle\forall\leavevmode\nobreak\ w_{\phi}$
$\displaystyle\in\mathscr{V}_{\phi},\quad\mathscr{V}_{\phi}=\\{w_{\phi}\in\text{H}^{1}(\Omega)\leavevmode\nobreak\
|\leavevmode\nobreak\ w_{\phi}=0\leavevmode\nobreak\
\forall\leavevmode\nobreak\ \textbf{X}\in\Gamma^{\phi}\\}$
we have,
$\displaystyle\int_{\Omega}w_{u_{c}}\Bigg{(}\Big{(}\frac{1+k}{2}-\frac{1-k}{2}\phi\Big{)}\frac{\partial
u_{c}}{\partial
t}-\Big{(}1+(1-k)u_{c}\Big{)}\frac{1}{2}\frac{\partial\phi}{\partial
t}\Bigg{)}\,dV$ (11a)
$\displaystyle+\int_{\Omega}\tilde{D}\big{(}\frac{1-\phi}{2}\big{)}\bm{\nabla}w_{u_{c}}.\big{(}\bm{\nabla}u_{c}-\bm{J}_{at}\big{)}\,dV=0$
$\displaystyle\int_{\Omega}w_{\phi}\Bigg{(}\Big{(}1-(1-k)\frac{y-\tilde{\bm{\nu}}t}{\tilde{l}_{T}}\Big{)}a^{2}_{s}(\mathbf{n})\frac{\partial\phi}{\partial
t}-\Big{(}\phi-\phi^{3}-\xi(1-\phi^{2})^{2}\big{(}u_{c}+\frac{y-\tilde{\bm{\nu}}t}{\tilde{l}_{T}}\big{)}\Big{)}\Bigg{)}\,dV$
(11b)
$\displaystyle+\int_{\Omega}a_{s}^{2}\bm{\nabla}w_{\phi}.\bm{\nabla}\phi\,dV+\int_{\Omega}|\bm{\nabla}\phi|^{2}a_{s}\frac{\partial
w_{\phi}}{\partial\bm{x}}\frac{\partial
a_{s}}{\partial\left(\frac{\partial\phi}{\partial\bm{x}}\right)}\,dV+\int_{\Omega}|\bm{\nabla}\phi|^{2}a_{s}\frac{\partial
w_{\phi}}{\partial\bm{y}}\frac{\partial
a_{s}}{\partial\left(\frac{\partial\phi}{\partial\bm{y}}\right)}\,dV=0$ here,
$\Omega=\Omega_{s}\cup\Omega_{l}$, where $\Omega_{s}$ is the solid phase and
$\Omega_{l}$ is the liquid phase, respectively. The variations,
$\omega_{u_{c}}$ and $\omega_{\phi}$, belong to $\text{H}^{1}(\Omega)$ \- the
Sobolev space of functions that are square-integrable and have a square-
integrable derivatives.
## 3 Numerical implementation and results
This section outlines the numerical implementation of the phase-field models
described in Section 2.2 and 2.4. In Section 3.1, we cover the computational
implementation. In Section 3.2, we summarize the input model parameters used
in the simulation of dendritic growth in a pure metal and a binary alloy, and
list the non-dimensional parameters used in these models. In Sections 3.3,
using various dendritic morphologies modeled in this work, we identify
important geometric features of dendrites that are tracked in the dendritic
shape studies presented later. In Sections 3.4.1 we simulate the classical
four fold symmetric dendrite shape occurring in undercooled pure melt and use
it as a basis for convergence studies of dendritic morphology and dendritic
tip velocity. Later, in Sections 3.5-3.8 , we model the evolution of complex
2D and 3D dendrite morphologies. These simulations cover various cases of
solidification ranging from a single equiaxed dendrite in a pure liquid melt
to a multi-columnar dendrites growing in a binary alloy. Finally, evolution of
equiaxed dendrite in 3D is demonstrated.
### 3.1 Computational implementation
The phase-field formulations presented in this work are solved using
computational implementations of two numerical techniques: (1) A
$C^{0}$-continuous basis code based on the standard Finite Element Method
(FEM); the code base is an in-house, C++ programing language based, parallel
code framework with adaptive meshing and adaptive time-stepping - build on top
of the deal.II open source Finite Element library [42], (2) A
$C^{1}$-continuous basis code based on the Isogeometric Analysis (IGA) method;
the code base is an in-house, C++ programing language based, parallel code
framework - build on top of the PetIGA open source IGA library [43]. As can be
expected, the IGA code base can model $C^{0}$-continuous basis also, but does
not currently support some capabilities like adaptive meshing that are needed
for 3D dendritic simulations. Both codes support a variety of implicit and
explicit time stepping schemes. The complete code base of the current work is
made available to the wider research community as an open source library [44].
### 3.2 Material properties and non-dimensional quantities in solidification
We now discuss the specific material properties and non-dimensional parameters
used in the modeling of pure metal and binary alloys. First, we consider the
case of dendrite growth in an undercooled melt of pure metal, whose governing
equations are given by Equations 5a and 5b. Undercooling of the liquid melt is
responsible for driving dendrite growth in this case and there is no external
imposition of a temperature gradient or a dendritic growth velocity. This is
commonly referred to as free dendritic growth. The set of non-dimensional
input parameters used in this model and their numerical values (adopted from
Karma and Rappel [45]) are summarized in Table 1. Here, $\tilde{D}$ is the
thermal diffusivity of heat conduction, $\epsilon_{4}$ is the strength of
interface surface energy anisotropy where the subscript indicates the assumed
symmetry of the dendritic structure, and in this simulation, we consider a
four-fold symmetry. The choice of the length-scale parameter $\lambda_{0}$,
identified as the interface thickness, and the time-scale parameter $\tau_{0}$
set the spatial and temporal resolution of the dendritic growth process. $\xi$
is a coupling parameter representing the ratio of interface thickness and
capillary length.
Non-dimensional parameter | Numerical value
---|---
Thermal diffusivity, $\tilde{D}$ | 1
Anisotropy strength, $\epsilon_{4}$ | 0.05
Characteristic length-scale, $\lambda_{0}$ | 1
Characteristic time-scale, $\tau_{0}$ | 1
Coupling parameter, $\xi$ | 1.60
Table 1: Input parameters for the phase-field model of solidification of a
pure metal. The length-scale parameter $\lambda_{0}$ and the time-scale
parameter $\tau_{0}$ set the spatial and temporal resolution of the dendritic
growth process.
Parameter | Numerical value (units)
---|---
Liquid phase diffusivity, $D_{l}$ | $2.4\times 10^{-9}$ $m^{2}s^{-1}$
Solid phase diffusivity, $D_{s}$ | $1.15\times 10^{-12}$ $m^{2}s^{-1}$
Partition coefficient, $k$ | 0.14
Anisotropy strength, $\epsilon_{4}$ | 0.05
Gibbs-Thomson coefficient, $\Gamma_{g}$ | $2.36\times 10^{-7}$ m-K
Liquidus slope, $m_{l}$ | -3.5 $K(wt\%)^{-1}$
Initial concentration, $c_{0}$ | 4wt%
Thermal gradient, $G$ | 3700 $Km^{-1}$
Pulling velocity, $\bm{\nu}_{p}$ | $1.35\times 10^{-5}$ $ms^{-1}$
Interface thickness, $\lambda_{0}$ | $1.058\times 10^{-6}$ m
Model constant, $a_{1}$ | 0.8839
Model constant, $a_{2}$ | 0.6267
Table 2: Material properties and phase-field model parameters for a binary
alloy Al-4wt.% Cu
Input parameters | Expression
---|---
Chemical capillary length, $d_{0}$ | $\frac{k\Gamma}{mC_{0}(k-1)}$
Ratio of interface thickness to capillary length, $\Xi$ | $\frac{\lambda_{0}}{d_{0}}$
Non-dimensional thermal diffusivity, $\tilde{D}$ | $\frac{D_{l}\tau_{0}}{\lambda_{0}^{2}}$
Non-dimensional pulling velocity, $\tilde{\bm{\nu}}$ | $\frac{\bm{\nu}\tau_{0}}{\lambda_{0}}$
Characteristic thermal length-scale, $\tilde{L_{T}}$ | $\frac{|m_{l}|(1-k)c_{0}}{kG{\lambda_{0}}}$
Characteristic time-scale, $\tau_{0}$ | $\frac{a_{1}a_{2}\xi\lambda_{0}^{2}}{D_{l}}$
Coupling parameter, $\xi$ | $a_{1}\Xi$
Table 3: Input parameters for the phase-field model of solidification of a
binary metallic alloy. The numerical values of the input parameters are
computed using the expressions shown here that are in terms of the properties
listed in Table 3.
Next we discuss the case of dendritic growth in binary alloys. As mentioned
earlier, chemically, alloys are distinct from pure melts due to the presence
of solute atoms of one or more alloying materials. During the solidification
of alloys, solute diffusion plays a critical role. The coupled governing
equations of solute diffusion and phase-field are given by the Equations
10c-10e. The solid-liquid interface is constrained with the imposition of the
temperature gradient, $G$, and the pulling velocity, $\bm{\nu}_{p}$. We solve
for two primal fields - composition undercooling, $u_{c}$, and the phase-field
order parameter, $\phi$. The input parameters used in the numerical model are
given in Tables 3-3. The parameters listed in the Table 3 are the required
material properties for modeling dendritic growth in a binary alloy (adopted
from Zhu et al.[46]): mass diffusivities $D_{l}$ and $D_{s}$ in the solid and
liquid phase, Gibbs-Thomson coefficient $\Gamma_{g}$, thermal gradient, $G$
and pulling velocity $\nu_{p}$. Since here we only model a generic dilute
binary alloy, the temperature and composition phase diagram of a physical
alloy is approximated by a linear relationship. This approximation is possible
if we assume the alloy to be a dilute binary alloy. This linearization of a
temperature-composition relationship results in two more parameters: the
partition coefficient $k$ and the slope of the liquidus line $m_{l}$. An
optimal choice of the phase-field interface thickness is made keeping in mind
the computational tractability of the final numerical model. Table 3 lists the
expressions used to calculate the input parameters relevant to the model.
### 3.3 Primary geometric features and shape variations in dendrite
morphologies
Dendrites occur in a great variety of morphologies and orientations. One of
most common, yet fascinating, occurrences of dendritic growth is in the
formation of snowflakes that are dendritic structures with a distinct six-fold
geometric symmetry. In more practical applications, as in the solidification
of metallic alloys, dendritic shapes can be broadly classified as “columnar”
or “equiaxed”. Columnar dendrites grow into a positive temperature gradient,
and the rate of growth is controlled by the constitutional undercooling of the
liquid melt ahead of the dendritic tip. On the other hand, growth of an
equiaxed dendrites in pure melt is controlled by the temperature undercooling
of the surrounding liquid melt.
The schematic shown in Figure 2 depicts these possible variations in dendritic
shapes, and also identifies important geometric features that are
traditionally used by the Materials Science and Metallurgy community to
quantify dendritic morphologies. At the top end of the schematic, shown is a
equiaxed dendrite with a six-fold symmetry (one-half of the symmetric dendrite
shown) that is typical of snowflakes and in certain alloy microstructures. At
the bottom, shown are columnar dendrites with their primary arms oriented
parallel and inclined at a $30^{\circ}$ angle to the vertical axis. These
primary arms develop secondary arms as they evolve, and the separation of the
primary arms (referred to as the primary dendrite arm spacing or PDAS) and
secondary arms (referred to as the secondary dendrite arm spacing or SDAS) are
important geometric characteristics of a given dendrite shape. Further,
towards the right side of the schematic is a columnar dendrite colony growing
homogeneously oriented along the horizontal axis, and shown towards the left
side of the schematic are two time instances (separated by an interval
$\Delta$t) of an equiaxed dendrite’s interface evolution and the movement of
the dendritic tip over this time interval is $\Delta\bm{x}$. The dendritic tip
velocity, a very important measure of dendritic kinetics, is then given by
$\bm{{\nu}}_{\text{tip}}=\frac{d\bm{x}}{dt}$. It is to be noted that all these
morphologies were produced by the authors of this manuscript in simulations of
various solidification conditions using the numerical framework presented in
this work, and are but one demonstration of the capabilities of this framework
in modeling complex dendritic morphologies.
Figure 2: Schematic showing important geometric features used to quantify
dendritic morphologies, and some typical variations of shapes and orientations
in columnar and equiaxed dendrites. Shown are the primary arms and primary
dendrite arm spacing (PDAS), secondary arms and secondary dendrite arm spacing
(SDAS), and the movement of the dendritic tip by a distance $\Delta$x over a
time interval $\Delta$t. Dendritic tip velocity is given by
$\bm{\nu}_{\text{tip}}$. All these shapes were generated using the
solidification numerical models presented in this work.
### 3.4 Single equiaxed dendrite evolution
Growth of a single equiaxed dendrite in an undercooled melt of pure metal is a
classical solidification problem, and much of traditional analysis of
dendritic growth is based on reduced order models for this problem. The phase-
field governing equations are given by Equations 6a-6b, and the primal fields
solved for are the temperature undercooling $u$ and the order parameter
$\phi$. These equations are solved using both a $C^{0}$ basis and a $C^{1}$
basis as part of the study of convergence with respect to variations in the
spatial discretization. Input parameters used in this model have been listed
in Section 3.2. Zero flux Boundary Conditions are considered on all the
boundaries for the fields $u$ and $\phi$, and the Initial Condition is a
uniformly undercooled liquid melt with $\Delta=-0.75$ everywhere. The phase-
field order parameter is set to $\phi=-1.0$ everywhere, except a very small
circular region at the center representing a solid seed formed due to
nucleation with $\phi=1.0$ in the seed. Recollect that, in all the models,
$\phi=1.0$ represents the bulk solid and $\phi=-1.0$ represents the bulk
liquid region and any value in between represents the diffuse interface
between these two regions. The non-dimensional size of the domain is 500x500
and it is uniformly discretized into elements with a length measure of $\Delta
x=2.0$. The non-dimensional time step for the implicit time stepping scheme
considered in this case is $\Delta t=0.04$.
The growth of dendrites out of the nucleation point at the center is dictated
by the degree of undercooling of the liquid melt. As solidification begins,
the surrounding undercooled liquid melt is at lower temperature than the solid
seed. As a result, the solid seed diffuses heat into the liquid region.
Theoretically, the solid-liquid interface of an undercooled pure melt is
always unstable, so small local numerical perturbations on circular solid seed
interface begin to grow. Due to the assumption of strong anisotropy (four-fold
symmetry) of the interface, perturbations to the interface grow rapidly along
the favorable orientations as compared to the non-favorable orientations.
Figure 3a represents $u$ at a particular time, t = 840. Diffusion takes place
over the entire stretch of the solid and liquid regions. From t = 0 to 840,
$u$ changes from the initial value of -0.75 as per the solution of the heat
diffusion equation. It should be noted that the bulk of the liquid region and
the solid dendrite are at a uniform temperature, and the temperature
variations are mostly localized to the vicinity of the interface. The bulk
liquid melt region shown in blue is at a low temperature and the bulk solid
dendrite in red is at a high temperature, as seen in Figure 3a. The level set
of $\phi=0$ is identified as the distinct solid-liquid interface and the
evolution of this interface is plotted in Figure 3b as a series of $\phi=0$
level sets at different time instances. As can be seen, at t=0, the circular
seed at the center of the domain, over time, evolves into a four-fold
symmetric interface. At time, t = 840, we see a fully developed equiaxed
dendrite shape. We also simulated this Initial Boundary Value Problem on
$C^{1}$-continuous mesh and found identical spatiotemporal growth of the
equiaxed dendrite.
(a) Temperature contour at t=840 (b) Time evolution of the interface
Figure 3: Growth of a single equiaxed dendrite in an undercooled melt of pure
metal. Shown are the (3a) Temperature contour $(u)$ at t = 840, and (3b)
Temporal evolution of the equiaxed dendrite interface, delineated by the level
set $\phi=0$, at different time instances.
#### 3.4.1 Numerical convergence studies
In this section, we look at the accuracy of the interface propagation
numerical results for the single equiaxed dendrite case presented above. We do
this by comparing the velocity of propagation of the dendritic tip with known
analytical solutions of tip velocity for this problem, and also study the
convergence of the tip velocity with mesh size and increasing continuity of
the numerical basis.
##### Velocity of the dendritic tip:
As can be seen in Figure 3, the phase-field model of solidification of a pure
metal successfully produces the freely growing dendrite shape and four-fold
symmetry as expected. However, predicting quantifiable features of the
dendritic evolution would provide a better validation of the proposed
numerical framework. In this context, the dendritic tip propagation velocity
(see Figure 2) for single equiaxed dendrites has been widely used in
analytical studies of dendritic propagation. In the literature, implicit
expressions for radius and velocity of the dendritic tip have been obtained
using semi-analytical approaches, and they provide estimates of
$\bm{\nu}_{tip}$. One of the more popular semi-analytical approaches to obtain
the tip velocity employs a Green’s function (GF) approach. The Green’s
function is an integral kernel that is widely used to solve of certain classes
of differential equations. We will not discuss this approach here, but only
use the dendritic tip velocity values predicted by this approach. These values
are denoted by $\bm{\nu}_{tip}(GF)$. Interested readers are referred to Karma
and Rappel[45] for details of the GF approach.
The model parameters used in this study are $\Delta=-0.65$,
$\epsilon_{4}=0.05$, $\tilde{D}=1$, $d_{0}/\lambda_{0}=0.544$,
$\lambda_{0}=1$, $\tau_{0}=1$, $\xi=1.596$. The numerical domain size and time
step size, in dimensionless units, are 500x500 and $\Delta t=0.04$,
respectively. For these values, the Green’s function approach predicts an
equilibrium tip velocity, $\bm{\nu}_{tip}(GF)=0.0469$. Freely growing equiaxed
dendrite in an infinite domain, under solidification conditions held constant
over time, attains a fixed equilibrium tip velocity. In this study, we
consider the effect of two discretization parameters, namely: (1) Comparison
of the numerical dendritic tip velocity ($\bm{\nu}_{tip}$) to the equilibrium
tip velocity obtained from Green’s function approach $\bm{\nu}_{tip}(GF)$
(Figure 4), and (2) Convergence of the tip velocity ($\bm{\nu}_{tip}$) with
mesh refinement, and its dependence on the order of continuity of the basis
(Figure 5).
The method of estimating the dendrite tip velocity from the numerical
simulations using the computational frameworks described above is described
briefly. Simulations were performed using both a $C^{0}$-continuous basis and
a $C^{1}$-continuous basis, to study the effect of basis continuity when
resolving fine dendritic morphologies. The dendritic tip velocity measurement
is done manually as a post-processing of the phase-field order parameter
contours of the numerical simulations. At each time instance, the tip of the
zero level set ($\phi=0$) of the order parameter is manually marked, and the
velocity of tip propagation at each time instance is computed using a simple
finite difference estimate of the derivative
$\bm{\nu}_{tip}=\frac{d\bm{x}}{dt}$. Two sources of error can potentially
cause minor variations to the computed velocity values. The first source of
error arises from not picking the exact tip location in subsequent time
snapshots as the dendrite interface shape is evolving with time. The second
source of error arises due to the use of only $C^{0}$ interpolation of the
nodal solutions inside the elements by the visualization tool. This error is
significant for $C^{1}$-basis, as the results are $C^{1}$-continuous but the
visualization of the results inside the elements is only $C^{0}$-continuous.
As can be seen in Figure 4, $\bm{\nu}_{tip}$ starts to approach the Green’s
function value, $\bm{\nu}_{tip}(GF)$, across all the simulations, but shows
better convergence to the Green’s function value for the $C^{1}$-continuous
basis. Further, we observe the percentage error of $\bm{\nu}_{tip}$, given by
Error$=\frac{|\bm{\nu}_{tip}-\bm{\nu}_{tip}(GF)|}{\bm{\nu}_{tip}(GF)}$,
consistently reduces with a decrease in the mesh size across both
$C^{0}$-continuous and $C^{1}$-continuous basis, as shown in Figures 5(a) and
5(b). This is an important numerical convergence result, as to the best of our
knowledge, this is the first time optimal convergence rates have been
demonstrated for non-trivial dendritic shape evolution with respect to mesh
refinement. The convergence rates are demonstrated in the plot in Figures
5(a), where the rate of convergence (slope of the percentage Error vs mesh
size plot) is close to two for a $C^{0}$-continuous basis, and close to three
for a $C^{1}$-continuous basis. While obtaining theoretical optimal converge
rates with respect to h-refinement of the discretization for this coupled
system of parabolic partial differential equations is very challenging, one
can observe that the convergence rates shown are comparable to the optimal
convergence rates of classical heat conduction and mass diffusion partial
differential equations [47]. This completes the estimation of error in
$\bm{\nu}_{tip}$, which is a measure of localized error in the order parameter
field, as the dendritic interface is given by the zero level set of the order
parameter field. Now we look at estimation of errors over the entire domain
for both the temperature field and the order parameter field.
$500$$1{,}000$$1{,}500$$2{,}000$$4.5$$5$$5.5$$6$$6.5$$7$$\cdot 10^{-2}$Time
$\bm{\nu}_{tip}$ (a) $\Delta x$=2.0
$500$$1{,}000$$1{,}500$$2{,}000$$4.5$$5$$5.5$$6$$6.5$$7$$7.5$$\cdot
10^{-2}$Time $\bm{\nu}_{tip}$ $C^{0}$ basis$C^{1}$ basisGreen’s function value
(b) $\Delta x$=1.0
Figure 4: Comparison of the numerical and analytical (Green’s function
approach) dendritic tip velocities over time during growth of a single
equiaxed dendrite for two different mesh sizes ($\Delta x$), and for both
$C^{0}$-continuous and $C^{1}$-continuous basis.
$10^{-0.1}$$10^{0}$$10^{0.1}$$10^{0.2}$$10^{0.3}$$10^{0}$$10^{1}$2131$y_{0}=2.04x_{0}+0.35$$y_{1}=3.72x_{1}-0.48$$y_{2}=4.11x_{2}-0.51$${\Delta
x}$ Error $C^{0}$ basis$C^{1}$ basis (a) Convergence of error in dendritic tip
velocity with mesh refinement
$C^{0}$ basis$C^{1}$ basis$0$$2$$4$$6$$8$$10$Error $\Delta x$ = 1.0 (b)
Comparison of error in dendritic tip velocity for a mesh size of $\Delta x$ =
1.0
Figure 5: Convergence of the percentage error in dendritic tip velocity, given
by Error=$\frac{|\bm{\nu}_{tip}-\bm{\nu}_{tip}(GF)|}{\bm{\nu}_{tip}(GF)}$,
with mesh refinement and its dependence on the order of continuity of the
basis. Subplot (a) shows optimal convergence of the error with respect to mesh
refinement in a log-log plot, and Subplot (b) shows the difference in
percentage errors between a $C^{0}$-continuous basis and a $C^{1}$-continuous
basis for one mesh size.
##### Numerical error estimates for the temperature and order parameter
fields:
For this study, we consider three different mesh refinements of a
$C^{0}$-continuous basis. For each of these refinements, the computational
domain size is 160x160, chosen time step is $\Delta t=0.12$, and the other
input parameters are as given in Table 1. As the Initial Condition, a small
nucleation seed was considered, and this seed grows into the surrounding
undercooled melt. For the problem of interest, there are no known analytical
solutions for the fields $u$ or $\phi$. As mentioned in the earlier
subsection, the only analytical result available is for the equilibrium
dendritic tip velocity. Thus, we follow a computational approach to obtain the
reference solution of both the primal fields. A separate problem is solved on
a very fine mesh ($\Delta x=$ 0.2) using the popular open source phase field
library PRISMS-PF [48], and the primal field solutions for this very fine mesh
are considered the reference solutions. It is to be noted that one of the
authors (S.R.) in this work was the lead developer of the PRISMS-PF library,
and this library is being successfully used by the wider Materials Science
community to solve numerous phase-field problems.
Using this numerical reference solution, error in the temperature and order
parameter fields of the solutions obtained for the numerical formulations
presented in this work, is calculated by comparing the fields and their
gradients using two types of norms relevant to the problem, and for three
different mesh refinements at one instance of time (time = 450) when the
dendritic shape is well established. The two error norms used in this analysis
are the $L_{2}$-norm and the $H_{1}$-norm, and these are defined as follows:
$|u-\hat{u}|_{L_{2}}=\sqrt{\int_{\Omega}|u_{h}-\hat{u}|^{2}\leavevmode\nobreak\
dV}$ (12a)
$|u-\hat{u}|_{H_{1}}=\sqrt{\int_{\Omega}\Big{(}|u_{h}-\hat{u}|^{2}+|\nabla
u_{h}-\nabla\hat{u}|^{2}\Big{)}\leavevmode\nobreak\ dV}$ (12b)
Here, $u_{h}$ is the finite element solution under consideration and $\hat{u}$
is the reference numerical solution. The integral is defined over the entire
numerical domain, so the estimated error, unlike the previous error in tip
velocity, is not localized to any particular region in the domain. The
convergence of this error in these two norms with mesh refinement is shown in
Figure 6 on a log-log scale. As can be seen from these plots, we observe good
convergence in both the norms for both the primal fields, with the rates of
convergence being in the vicinity of 1.5 to 2.0. We are not aware of any
theoretical optimal error estimates in any of the relevant error norms for
this problem, so this numerical strategy to obtain error estimates is seen as
filing this vital gap by providing numerical convergence estimates, thus
permitting the numerical analyses of these governing equations.
$10^{-0.2}$$10^{0}$$10^{0.2}$$10^{0.4}$$10^{0.5}$$10^{1}$$10^{1.5}$11.80y=1.80x+0.85$\mathbf{\Delta
x}$ $|\bm{u}-\hat{\bm{u}}|_{L_{2}}$ (a) $L^{2}$-norm of the temperature
$10^{-0.2}$$10^{0}$$10^{0.2}$$10^{0.4}$$10^{0.5}$$10^{1}$$10^{1.5}$11.79y=1.79x+0.86$\mathbf{\Delta
x}$ $|\bm{u}-\hat{\bm{u}}|_{H_{1}}$ Numerical dataLinear Fit (b) $H^{1}$-norm
of the temperature
$10^{-0.2}$$10^{0}$$10^{0.2}$$10^{0.4}$$10^{1}$$10^{2}$11.59y=1.59x+1.45$\mathbf{\Delta
x}$ $|\bm{\phi}-\hat{\bm{\phi}}|_{L^{2}}$ (c) $L^{2}$-norm of the order
parameter
$10^{-0.2}$$10^{0}$$10^{0.2}$$10^{0.4}$$10^{1.5}$$10^{2}$11.50y=1.50x+1.50$\mathbf{\Delta
x}$ $|\bm{\phi}-\hat{\bm{\phi}}|_{H^{1}}$ Numerical dataLinear Fit (d)
$H^{1}$-norm of the order parameter
Figure 6: Error estimates and order of convergence for the primal fields
$(u,\phi)$ shown on log-log plots. The error norms used in this study are the
$L_{2}$-norm and the $H_{1}$-norm of $(u,\phi)$. The linear fit expression is
also given for each plot.
### 3.5 Growth of multiple equiaxed dendrites
In Section 3.4, we discussed one particular type of dendritic growth, namely
the evolution of a single equiaxed dendrite during the solidification of pure
metal. Now we look at the evolution of multiple equiaxed dendrites during the
solidification of a more complex binary alloy solution. The input parameters
for the model are listed in Table 3. Growth of multi-equiaxed dendrites during
solidification of a binary alloy is governed by the Equations 11a-11b. The
primal fields in this problem are the non-dimensional composition, $u_{c}$,
and the phase-field order parameter, $\phi$. As an Initial Condition for this
problem, we consider multiple small nucleation seeds randomly distributed in
the numerical domain. The seed locations are completely random, and generated
using a random number generator. To ensure free equiaxed dendritic growth,
thermal gradient $G$ and pulling velocity $\bm{\nu}_{p}$ is taken to be zero.
The numerical domain of size 600x600 is subdivided into uniform elements with
a length measure of $\Delta x=1.2$. The time step size is $\Delta t=0.005$.
The average alloy composition in the entire domain is taken to be $c_{0}=4\%$,
which corresponds to $u_{c}=0$. Nucleation seeds of the solid phase are
randomly placed such that $\phi=1$ inside of the seed and $\phi=-1$ outside of
the seed. Zero flux Boundary Conditions are considered for both the fields.
During the metal casting process, many solid nuclei are randomly formed near
cavities or defects on the surface of a mold. These nuclei then start to grow
in an equiaxial manner with neighboring nuclei competing with each other. This
competition can shunt the growth of dendrites approaching each other. Solution
contours of the fields $u_{c}$ and $\phi$ are shown in the Figure 7. Here,
Subfigure 7a shows the solute composition in the numerical domain at t = 40.
Initially, the entire domain is at an alloy concentration of 4%. Due to
negligible mass diffusivity in the solid phase, solute diffusion primarily
takes place in the liquid region. A solute composition of $c_{0}=4\%$ is
highest near the interface, and as one moves away from the interface, the
solute concentration decreases. This is apparent in all the dendritic
interfaces shown in the Subfigure 7a. Solute composition inside the solid is
given by $c_{0}=0.56$ and it stays constant due to low negligible diffusivity
in the solid region. Subfigure 7b demonstrates the competition between the
dendrites growing out of the various seeds, the interaction of the dendritic
tips, and the growth of secondary dendrite arms. Material properties of an
alloy are often correlated with fine-scale microstructure length scales (at
the grain scale and sub-grain scale), and the secondary dendrite arm spacing
(SDAS) is one of the important microstructural properties of alloys.
(a) Composition (b) Zero level sets of the order parameter
Figure 7: Growth of multiple equiaxed dendrites. Shown are the (7a) contours
of solute composition, c, at time t = 40, and (7b) Time evolution of multiple
equiaxed dendrite interfaces at time t = 0, 25 and 40. The dendrite interfaces
are the zero level sets of the order parameter, $\phi=0.0$.
### 3.6 Growth of single columnar dendrites
In this section, we demonstrate the growth of columnar dendrites, growing from
a single seed during the solidification of a binary alloy. The numerical model
and input parameters used in this case are identical to the previous Section
3.5. The notable difference from the model in Section 3.5 to the current model
is the imposition of constraints like non-zero thermal gradient $G$ and
pulling velocity $\bm{\nu}_{p}$. Initially, we place a single solid seed at
the bottom of the numerical domain. The entire domain is at a solute
composition of $c_{0}=4\%$ which corresponds to the composition undercooling
of $u_{c}=0$. Zero flux Boundary Conditions are enforced on both the primal
fields, $u_{c}$ and $\phi$. Figure 8 depicts the evolution of the composition
field, and the dendritic interface as a function of time. The temperature
gradient (G) applied is in the vertical direction. As opposed to the
solidification in an undercooled pure metal, the liquid phase in this type of
solidification is at a higher temperature than the dendrite in a solid phase.
Any numerically localized perturbation to the solid-liquid interface will grow
as per the degree of constitutional undercooling. For the given temperature
gradient $G$ and pulling velocity $\bm{\nu}_{p}$, the solid-liquid interface
is unstable. This is evident from the interface $\phi=0$ evolution captured in
the Subfigure 8b. At t = 0, the solid seed starts to grow. By t = 35, the
solid seed has developed four primary branches due to the four-fold symmetry
in the anisotropy considered. The growth of only three primary arms is
captured in the simulation and the fourth arm pointing downward is not a part
of the computation. These three primary arms have developed secondary arms of
their own as seen in contour. By t = 40, the initial seed has grown into a
larger size dendrite and several of its secondary arms are growing in the
columnar fashion aligning with the direction in which constraints are applied,
i.e, vertically. The phenomenon of secondary arms growing in the vertical
direction becomes more obvious by t = 45 and t = 50 as seen in the Subfigure
8b. The diffusion of the solute is governed by the same mechanism discussed in
Section 3.5. Like Subfigure 7a, in this case as well, the solute composition
is maximum near the dendrite interface and it decreases as one moves away from
the interface. This is clearly shown in the Subfigure 8a.
(a) Composition (b) Zero level sets of the order parameter
Figure 8: Growth of a single columnar dendrite. Shown are the (8a) contours of
solute composition, c, at different time instances, and (8b) Time evolution of
single columnar dendrite interfaces at times t = 30 to 50.
### 3.7 Growth of multiple columnar dendrites
Columnar dendrites growing out of a flat boundary are the next type of
dendrites modeled. This type of dendrites occur in directional solidification
processes which are of significance in manufacturing of alloy components by
use of molds, where often the dendrite growth is primarily columnar. The
characteristic of the columnar dendrites are the growth of a colony of
dendrites along the direction of the imposed temperature gradient and the
pulling velocity. These dendrites then move at a steady-state tip velocity
which is determined by the pulling velocity. The other geometric features of
columnar dendrites which are typically observed in the experimental studies
include primary arms, secondary arms, spacing between arms, growth competition
between the neighboring dendrites, and the interaction of the diffusion field
between neighboring dendrites. The growth of multi-columnar dendrites is
governed by the same governing equations that were used to model the
solidification of a binary alloy in the previous Section 3.4-3.5, and so are
the input parameters. However for the Initial Condition, we model an initial
thin layer of solid solute on one of the boundaries, and set $\phi=1$ inside
the initial solute layer thickness and $\phi=-1$ in the rest of the domain.
Non-dimensional composition undercooling is set to $u_{c}=-1$ everywhere in
the domain. For the Boundary Conditions, zero flux conditions are enforced on
the boundaries for both the primal fields $u_{c}$ and $\phi$. Figure 9 depicts
the results of multiple columnar dendritic growth. As can be seen, the phase-
field contour interface ($\phi=0$) at time, t= 240, fills the entire domain
with a colony of columnar dendrites. The initial planar interface with a
solute composition seeds the dendrites that then compete to evolve into this
columnar dendritic structure. The emergence of seven primary arms of the
dendrites, as seen in Subfigure 9b, is the result of the competition for
growth among several more initial primary arms of dendrites. Some of these
underdeveloped arms can be seen in the lower portion of the figure. The
primary arm spacing between these seven primary arms appears to converge to a
constant value with the progress of the dendrites. Another feature of the
columnar dendrites that we were able to capture is the growth of the secondary
dendrite arms and their orientation. The concentration of the solute in the
solid region is nearly uniform at a value of 0.56. The region surrounding
these solid dendrites consists of the liquid melt where the highest
concentration is about 4.0. Due to mass diffusion, there exists a gradient in
the solute concentration in the liquid melt as can be seen in the Subfigure
9a.
(a) Composition (b) Zero level sets of the order parameter
Figure 9: Growth of multiple columnar dendrites. Shown are the (9a) Contours
of the solute composition, $c$, at time t = 240, and (9b) Growth profile of
multiple columnar dendrite interfaces at time t = 240.
### 3.8 Three dimensional growth of single equiaxed dendrite
We now simulate the phase-field model described by the Equations 5a-5b in
three dimensions. The initial solidification conditions consist of a small
spherical seed placed in a $500\times 500\times 500$ computational domain with
an adaptive mesh and the smallest element length measure of $\Delta x=2.0$ and
uniform time step $\Delta t=0.12$. The seed is placed in an undercooled melt
of pure metal. The input phase-field parameters used in the simulation are
listed in Table 1. The evolution of a spherical seed into a four-fold
symmetric dendrite structure is shown in the Figure 10. The fully grown
dendrite in blue at the end of the evolution has six primary arms as dictated
by the surface anisotropy. The distribution of an undercooling temperature $u$
on a fully grown dendrite, shown in red, can also be seen in the figure. Three
dimensional implementation of phase-field models of dendrites is
computationally very demanding, due to the need to resolve the fine scale
primary-arm and secondary-arm structures. To make this problem computationally
tractable, we use local mesh adaptivity (h-adaptivity), implicit time-stepping
schemes with adaptive time-step control, and domain decomposition using MPI.
Figure 10: Evolution of a 3D single equiaxed dendrite. The dendritic
structures shown in blue are the time evolution of the contours of the phase-
field parameter, $\phi$, and the dendritic structure shown in red is the
contour of the undercooling temperature, $u$, at one time instance.
## 4 Conclusion
In this work, we present a detailed development of the requisite numerical
aspects of dendritic solidification theory, and the computational
implementation of the corresponding phase-field formulations to model
dendritic growth in pure metals and binary alloys. A wide variety of
physically relevant dendritic solidification patterns are modeled by solving
the governing equations under various initial conditions and boundary
conditions. To validate the numerical framework, we simulate the classical
four-fold symmetric dendrite shape occurring in undercooled pure melt and
compare the numerically computed dendritic tip velocities with the
corresponding analytical values obtained using a Green’s function approach.
Further, this problem is used as a basis for performing error convergence
studies of dendritic tip velocity and dendritic morphology (primal fields of
temperature and order parameter). Further, using this numerical framework,
various types of dendritic solidification patterns like multi-equiaxed, single
columnar and multi-columnar dendrites are modeled in two-dimensional and
three-dimensional computational domains.
The distinguishing aspects of this work are - a unified treatment of both
pure-metals and alloys; novel numerical error estimates of dendritic tip
velocity; and the convergence of error for the primal fields of temperature
and the order parameter with respect to the numerical discretization. To the
best of our knowledge, this is a first of its kind study of numerical
convergence of the phase-field equations of dendritic growth in a finite
element method setting, and a unified computational framework for modeling a
variety of dendritic structures relevant to solidification in metallic alloys.
## Acknowledgement
The authors would like to thank Prof. Dan Thoma (University of Wisconsin-
Madison) and Dr. Kaila Bertsch (University of Wisconsin-Madison; now at
Lawrence Livermore National Laboratory) for very useful discussions on
dendritic growth and microstructure evolution in the context of additive
manufacturing of metallic alloys.
## Declarations
Conflict of interest: The authors declare that they have no conflict of
interest.
## References
* [1] L.I. Rubinstein. The stefan problem, transl. math. Monographs, 27:327–3, 1971.
* [2] Gunter H Meyer. The numerical solution of stefan problems with front-tracking and smoothing methods. Applied Mathematics and Computation, 4(4):283–306, 1978.
* [3] Guillermo Marshall. A front tracking method for one-dimensional moving boundary problems. SIAM journal on scientific and Statistical Computing, 7(1):252–263, 1986.
* [4] Jonathan A Dantzig and Michel Rappaz. Solidification: -Revised & Expanded. EPFL press, 2016.
* [5] S Chen, B Merriman, Smereka Osher, and P Smereka. A simple level set method for solving stefan problems. Journal of Computational Physics, 135(1):8–29, 1997.
* [6] G Fix. Phase field method for free boundary problems, in; free boundary problems, a. Fasanao and M. Primicerio, eds., Pit-mann, London, 1983.
* [7] Joseph B Collins and Herbert Levine. Diffuse interface model of diffusion-limited crystal growth. Physical Review B, 31(9):6119, 1985.
* [8] Gunduz Caginalp. An analysis of a phase field model of a free boundary. Archive for Rational Mechanics and Analysis, 92(3):205–245, 1986\.
* [9] G Caginalp. Stefan and hele-shaw type models as asymptotic limits of the phase-field equations. Physical Review A, 39(11):5887, 1989.
* [10] Ryo Kobayashi. Modeling and numerical simulations of dendritic crystal growth. Physica D: Nonlinear Phenomena, 63(3-4):410–423, 1993.
* [11] Alain Karma and Wouter-Jan Rappel. Phase-field method for computationally efficient modeling of solidification with arbitrary interface kinetics. Physical review E, 53(4):R3017, 1996.
* [12] Alain Karma and Wouter-Jan Rappel. Phase-field model of dendritic sidebranching with thermal noise. Physical review E, 60(4):3614, 1999.
* [13] Mathis Plapp and Alain Karma. Multiscale finite-difference-diffusion-monte-carlo method for simulating dendritic solidification. Journal of Computational Physics, 165(2):592–619, 2000.
* [14] James A Warren and William J Boettinger. Prediction of dendritic growth and microsegregation patterns in a binary alloy using the phase-field method. Acta Metallurgica et Materialia, 43(2):689–703, 1995.
* [15] Irina Loginova, Gustav Amberg, and John Ågren. Phase-field simulations of non-isothermal binary alloy solidification. Acta materialia, 49(4):573–581, 2001.
* [16] Alain Karma. Phase-field formulation for quantitative modeling of alloy solidification. Physical Review Letters, 87(11):115701, 2001.
* [17] JC Ramirez, C Beckermann, As Karma, and H-J Diepers. Phase-field modeling of binary alloy solidification with coupled heat and solute diffusion. Physical Review E, 69(5):051607, 2004.
* [18] Blas Echebarria, Roger Folch, Alain Karma, and Mathis Plapp. Quantitative phase-field model of alloy solidification. Physical Review E, 70(6):061604, 2004.
* [19] Robert F Almgren. Second-order phase field asymptotics for unequal conductivities. SIAM Journal on Applied Mathematics, 59(6):2086–2107, 1999.
* [20] Munekazu Ohno and Kiyotaka Matsuura. Quantitative phase-field modeling for dilute alloy solidification involving diffusion in the solid. Physical Review E, 79(3):031603, 2009.
* [21] Xiaobing Feng and Andreas Prohl. Analysis of a fully discrete finite element method for the phase field model and approximation of its sharp interface limits. Mathematics of computation, 73(246):541–567, 2004.
* [22] B Gonzalez-Ferreiro, Héctor Gómez, and Ignacio Romero. A thermodynamically consistent numerical method for a phase field model of solidification. Communications in Nonlinear Science and Numerical Simulation, 19(7):2309–2323, 2014.
* [23] Chuanjun Chen and Xiaofeng Yang. Efficient numerical scheme for a dendritic solidification phase field model with melt convection. Journal of Computational Physics, 388:41–62, 2019.
* [24] Daniel Kessler and J-F Scheid. A priori error estimates of a finite-element method for an isothermal phase-field model related to the solidification process of a binary alloy. IMA journal of numerical analysis, 22(2):281–305, 2002.
* [25] Xianliang Hu, Ruo Li, and Tao Tang. A multi-mesh adaptive finite element approximation to phase field models. Communications in Computational Physics, 5(5):1012–1029, 2009.
* [26] Jan Rosam, Peter K Jimack, and Andy Mullis. A fully implicit, fully adaptive time and space discretisation method for phase-field simulation of binary alloy solidification. Journal of Computational Physics, 225(2):1271–1287, 2007.
* [27] Damien Tourret, Hong Liu, and Javier LLorca. Phase-field modeling of microstructure evolution: Recent applications, perspectives and challenges. Progress in Materials Science, 123:100810, 2022.
* [28] Zhijun Wang, Junjie Li, Jincheng Wang, and Yaohe Zhou. Phase field modeling the selection mechanism of primary dendritic spacing in directional solidification. Acta Materialia, 60(5):1957–1964, 2012.
* [29] V Fallah, M Amoorezaei, N Provatas, SF Corbin, and A Khajepour. Phase-field simulation of solidification morphology in laser powder deposition of ti–nb alloys. Acta Materialia, 60(4):1633–1646, 2012.
* [30] Damien Tourret and Alain Karma. Growth competition of columnar dendritic grains: A phase-field study. Acta Materialia, 82:64–83, 2015.
* [31] Tomohiro Takaki, Munekazu Ohno, Takashi Shimokawabe, and Takayuki Aoki. Two-dimensional phase-field simulations of dendrite competitive growth during the directional solidification of a binary alloy bicrystal. Acta Materialia, 81:272–283, 2014.
* [32] Shaoning Geng, Ping Jiang, Xinyu Shao, Gaoyang Mi, Han Wu, Yuewei Ai, Chunming Wang, Chu Han, Rong Chen, Wei Liu, et al. Effects of back-diffusion on solidification cracking susceptibility of al-mg alloys during welding: A phase-field study. Acta Materialia, 160:85–96, 2018.
* [33] A Farzadi, Minh Do-Quang, S Serajzadeh, AH Kokabi, and Gustav Amberg. Phase-field simulation of weld solidification microstructure in an al–cu alloy. Modelling and Simulation in Materials Science and Engineering, 16(6):065005, 2008.
* [34] X Wang, PW Liu, Y Ji, Y Liu, MH Horstemeyer, and L Chen. Investigation on microsegregation of in718 alloy during additive manufacturing via integrated phase-field and finite-element modeling. Journal of Materials Engineering and Performance, 28(2):657–665, 2019.
* [35] Matthew R Rolchigo, Michael Y Mendoza, Peyman Samimi, David A Brice, Brian Martin, Peter C Collins, and Richard LeSar. Modeling of ti-w solidification microstructures under additive manufacturing conditions. Metallurgical and Materials Transactions A, 48(7):3606–3622, 2017\.
* [36] Supriyo Ghosh, Li Ma, Nana Ofori-Opoku, and Jonathan E Guyer. On the primary spacing and microsegregation of cellular dendrites in laser deposited ni–nb alloys. Modelling and simulation in materials science and engineering, 25(6):065002, 2017.
* [37] Xibing Gong and Kevin Chou. Phase-field modeling of microstructure evolution in electron beam additive manufacturing. Jom, 67(5):1176–1182, 2015.
* [38] Seshadev Sahoo and Kevin Chou. Phase-field simulation of microstructure evolution of ti–6al–4v in electron beam additive manufacturing process. Additive manufacturing, 9:14–24, 2016.
* [39] Trevor Keller, Greta Lindwall, Supriyo Ghosh, Li Ma, Brandon M Lane, Fan Zhang, Ursula R Kattner, Eric A Lass, Jarred C Heigel, Yaakov Idell, et al. Application of finite element, phase-field, and calphad-based methods to additive manufacturing of ni-based superalloys. Acta materialia, 139:244–253, 2017.
* [40] Vladimir Slavov, Stefka Dimova, and Oleg Iliev. Phase-field method for 2d dendritic growth. In International Conference on Large-Scale Scientific Computing, pages 404–411. Springer, 2003.
* [41] Hieram Neumann-Heyme. Phase-field modeling of solidification and coarsening effects in dendrite morphology evolution and fragmentation. PhD thesis, Technical University of Dresden, Dresden, Technical University of Dresden, Dresden, 8 2017. An optional note.
* [42] Daniel Arndt, Wolfgang Bangerth, Bruno Blais, Marc Fehling, Rene Gassmöller, Timo Heister, Luca Heltai, Uwe Köcher, Martin Kronbichler, Matthias Maier, Peter Munch, Jean-Paul Pelteret, Sebastian Proell, Konrad Simon, Bruno Turcksin, David Wells, and Jiaqi Zhang. The deal.II library, version 9.3. Journal of Numerical Mathematics, 29(3):171–186, 2021.
* [43] V.M. Calo N. Collier, L. Dalcin. PetIGA: High-performance isogeometric analysis. arxiv, (1305.4452), 2013. http://arxiv.org/abs/1305.4452.
* [44] K. Bhagat. Phase-field based dendritic modeling. https://github.com/cmmg/dendriticGrowth, 2022.
* [45] Alain Karma and Wouter-Jan Rappel. Quantitative phase-field modeling of dendritic growth in two and three dimensions. Physical review E, 57(4):4323, 1998.
* [46] Chang-sheng Zhu, Sheng Xu, Li Feng, Dan Han, and Kai-ming Wang. Phase-field model simulations of alloy directional solidification and seaweed-like microstructure evolution based on adaptive finite element method. Computational Materials Science, 160:53–61, 2019.
* [47] M Bieterman and I Babuška. The finite element method for parabolic equations. Numerische Mathematik, 40(3):373–406, 1982.
* [48] Stephen DeWitt, Shiva Rudraraju, David Montiel, W Beck Andrews, and Katsuyo Thornton. Prisms-pf: A general framework for phase-field modeling with a matrix-free finite element method. npj Computational Materials, 6(1):1–12, 2020.
|
# A note on an effective bound for the gonality conjecture
Alexander Duncan, Wenbo Niu, Jinhyung Park Department of Mathematical
Sciences, University of Arkansas, Fayetteville, AR 72701, USA
<EMAIL_ADDRESS>Department of Mathematical Sciences, University of Arkansas,
Fayetteville, AR 72701, USA<EMAIL_ADDRESS>Department of Mathematical
Sciences, KAIST, 291 Daehak-ro, Yuseong-gu, Daejeon 34141, Republic of Korea
<EMAIL_ADDRESS>
###### Abstract.
The gonality conjecture, proved by Ein–Lazarsfeld, asserts that the gonality
of a nonsingular projective curve of genus $g$ can be detected from its
syzygies in the embedding given by a line bundle of sufficiently large degree.
An effective result obtained by Rathmann says that any line bundle of degree
at least $4g-3$ would work in the gonality theorem. In this note, we improve
the degree bound to $4g-4$ with two exceptional cases.
###### Key words and phrases:
gonality conjecture, syzygies of an algebraic curve, symmetric product of an
algebraic curve
###### 2020 Mathematics Subject Classification:
14Q20, 13A10
W. Niu was supported by the Simons Collaboration Grants for Mathematicians
J. Park was partially supported by the National Research Foundation (NRF)
funded by the Korea government (MSIT) (NRF-2021R1C1C1005479).
## 1\. Introduction
We work over the field $\mathbf{C}$ of complex numbers. Let $C$ be a
nonsingular projective curve of genus $g$, and $B$ and $L$ be line bundles on
$C$. Suppose that $L$ is globally generated, and write
$S=\operatorname{Sym}H^{0}(L)$. The associated section module
$R(C,B;L):=\bigoplus_{q\geq 0}H^{0}(B\otimes L^{q})$
is a finitely generated graded $S$-module. It admits a minimal graded free
resolution over $S$:
$\cdots\longrightarrow E_{p}\longrightarrow\cdots\longrightarrow
E\longrightarrow E_{1}\longrightarrow E_{0}\longrightarrow
R(C,B;L)\longrightarrow 0.$
Each graded free $S$-module $E_{p}$ in the resolution has the form
$E_{p}=\bigoplus_{q\in\mathbb{Z}}K_{p,q}(C,B;L)\otimes S(-p-q),$
where $K_{p,q}(C,B;L)$ is the Koszul cohomology group defined as the
cohomology at the middle of the Koszul-type complex
$\wedge^{p+1}H^{0}(L)\otimes H^{0}(B\otimes
L^{q+1})\longrightarrow\wedge^{p}H^{0}(L)\otimes H^{0}(B\otimes
L^{q})\longrightarrow\wedge^{p-1}H^{0}(L)\otimes H^{0}(B\otimes L^{q-1}).$
A particularly interesting case is when $B=\mathscr{O}_{C}$ and $L$ has large
degree. The Koszul groups $K_{p,q}(C;L):=K_{p,q}(C,\mathscr{O}_{C};L)$ give a
minimal free resolution of the section ring $R(C;L):=R(C,\mathscr{O}_{C};L)$.
Note that if $C\subseteq\mathbf{P}(H^{0}(L))$ is projectively normal, then
$R(C;L)$ is the homogeneous coordinate ring of $C$ in $\mathbf{P}(H^{0}(L))$.
If $L$ is nonspecial, then $K_{p,q}(C;L)$ vanishes for $q\geq 3$ and thus the
minimal free resolution essentially consists of two strands $K_{p,1}(C;L)$ and
$K_{p,2}(C;L)$. If $L$ has degree larger than $2g$, Green’s $(2g+1+p)$ theorem
([Gr84, Theorem 4.a.1], see also [GL88]) determines the strand $K_{p,2}(C;L)$
using the notion of $N_{p}$ property. The Green–Lazarsfeld gonality conjecture
[GL86] together with their nonvanishing theorem [Gr84, Appendix] predicts the
shape of the strand $K_{p,1}(C;L)$. This conjecture was proved by Ein-
Lazarsfeld [EL15] when $L$ has sufficiently large degree. Ranthmann then
showed an effective result (see [Ra16, Theorem 1.1]): if $\deg L\geq 4g-3$,
then
$K_{p,1}(C;L)\neq 0\Longleftrightarrow 1\leq p\leq\deg
L-g-\operatorname{gon}(C),$
where $\operatorname{gon}(C)$ is the gonality of $C$ which by definition is
the minimal degree of pencils on $C$. As pointed out in [Ra16], although the
degree bound above is not expected to be optimal, there is an example of a
plane quartic curve showing the degree bound $4g-4$ does not work. In this
short note, we investigate the failure of the gonality conjecture when $\deg
L=4g-4$. The main result is the following.
###### Theorem 1.1.
Let $C$ be a nonsingular projective curve of genus $g\geq 2$, and $L$ be a
line bundle on $C$ with $\deg L\geq 4g-4$. Then
$K_{p,1}(C;L)\neq 0\Longleftrightarrow 1\leq p\leq\deg
L-g-\operatorname{gon}(C)$
unless $L=\omega^{2}_{C}$ and either $g=2$ or $C$ is a plane quartic curve. In
the exceptional cases, $K_{\deg L-g-\operatorname{gon}(C)+1,1}(C;L)\neq 0$ but
$K_{\deg L-g-\operatorname{gon}(C)+2,1}(C;L)=0$.
An easy application of the theorem gives a uniform picture of syzygies of
pluricanonical embedding of curves, especially the second power of the
canonical divisor. It has been a long standing interest to understand the
syzygies of canonical curves. The shape of the minimal free resolution of
$R(C;\omega_{C})$ was predicted in Green’s conjecture [Gr84, Conjecture 5.1].
It was verified by Voisin [Voi1, Voi2] for general curves, but it is still
widely open in general. For pluricanonical embedding
$C\subseteq\mathbf{P}(H^{0}(\omega_{C}^{k}))$, the picture of syzygies turns
out to be complete, and we give a summary here. Let $C$ be a curve of genus
$g\geq 2$ and gonality $\operatorname{gon}(C)$. Put $L:=\omega^{k}_{C}$ and
write $r:=h^{0}(L)-1$. For $k\geq 3$, Greens’s $(2g+1+p)$-theorem and
Rathmann’s effective gonality theorem give the result that
$K_{p,1}(C;\omega^{k}_{C})\neq 0\Longleftrightarrow 1\leq p\leq
r-\operatorname{gon}(C).$
For $L=\omega^{2}_{C}$ and $r=3g-3$, Green’s $(2g+1+p)$-theorem and Theorem
1.2 gives us the following two cases
1. (1)
If either $g=2$ ($r=\operatorname{gon}(C)=2$), or $C$ is a plane quartic curve
($r=5$ and $\operatorname{gon}(C)=3$), then
$K_{p,1}(C;\omega^{2}_{C})\neq 0\Longleftrightarrow 1\leq p\leq
r-\operatorname{gon}(C)+1.$
2. (2)
Otherwise,
$K_{p,1}(C;\omega^{2}_{C})\neq 0\Longleftrightarrow 1\leq p\leq
r-\operatorname{gon}(C).$
In the setting of Theorem 1.1, Green–Lazarsfeld’s nonvanishing theorem [Gr84,
Appendix] shows that $K_{p,1}(C;L)\neq 0$ for $1\leq p\leq\deg
L-g-\operatorname{gon}(C)$. To prove the theorem, it is sufficient to prove
that $K_{\deg L-g-\operatorname{gon}(C)+1,1}(C;L)=0$. By the duality theorem
[Gr84, Theorem 2.c.6],
$K_{\deg
L-g-\operatorname{gon}(C)+1,1}(C;L)=K_{\operatorname{gon}(C)-2,1}(C,\omega_{C};L)^{\vee}.$
Notice that $\omega_{C}$ is $(\operatorname{gon}(C)-2)$-very ample. Recall
that $B$ is $p$-very ample if the restriction map on global sections
$H^{0}(B)\rightarrow H^{0}(B|_{\xi})$ is surjective for every effective
divisor $\xi$ of degree $p+1$, (in other words, $\xi$ imposes independent
conditions on the global sections of $B$). As in [EL15] and [Ra16], it is
natural to study more generally vanishing of $K_{p,1}(C,B;L)$ when $B$ is a
$p$-very ample line bundle and $\deg L\geq\deg B+2g-2$. The main result of
[Ra16] says that if $H^{1}(C,L\otimes B^{-1})=0$, then $K_{p,1}(C,B;L)=0$. For
our purpose, we only need to consider the case that $L=B\otimes\omega_{C}$.
Theorem 1.1 can be deduced from the following:
###### Theorem 1.2.
Let $C$ be a nonsingular projective curve of genus $g\geq 0$, $B$ be a
$p$-very ample line bundle on $C$, and $L:=B\otimes\omega_{C}$.
1. (1)
If $h^{0}(B)\geq p+3$, then $K_{p,1}(C,B;L)=0$.
2. (2)
If $h^{0}(B)=p+2$, then $K_{p,1}(C,B;L)=S^{p}H^{0}(\omega_{C})$.
3. (3)
If $h^{0}(B)=p+1$, then $K_{p,1}(C,B;L)=0$.
The idea to prove the theorem is to use the kernel bundles on the symmetric
products of the curve. We follow the approach introduced by Voisin [Voi1,
Voi2] and then used by Ein–Lazarsfeld [EL15], Rathmann [Ra16], and many others
to conduct a computation of Koszul cohomology groups on the symmetric products
of the curve. To be concrete, in our case,
$K_{p,1}(C,B;L)=H^{1}(C_{p+1},M_{p+1,B}\otimes N_{p+1,L}),$
where $M_{p+1,B}$ is the kernel bundle of the evaluation map
$H^{0}(C,B)\otimes\mathscr{O}_{C_{p+1}}\to E_{p+1,B}$ of the tautological
bundle $E_{p+1,B}$ and $N_{p+1,L}$ is a line bundle on $C_{p+1}$. More
generally, we establish the following vanishing:
$H^{i}(C_{p+1},\wedge^{k}M_{p+1,B}\otimes N_{p+1,L})=0~{}~{}\text{ for $i>0$}$
when $h^{0}(B)\geq p+k+2$. We hope that our results and methods may shed
lights on the similar problems for higher dimensional varieties.
Acknowledgments. The authors would like to thank Lawrence Ein for suggestions
and comments.
## 2\. Preliminaries
Let us start with setting up notations used throughout the paper. Let $C$ be a
nonsingular projective curve of genus $g$. For any $p\geq 0$, denote by
$C_{p+1}$ the $(p+1)$-th symmetric product. Write $U_{p+1}=C_{p}\times C$ to
be the universal family over $C_{p+1}$. One has a commutative diagram
$\textstyle{U_{p+1}=C_{p}\times C\
\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\sigma_{p+1}}$$\scriptstyle{j}$$\textstyle{C_{p+1}\times
C\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\pi_{p+1}}$$\textstyle{C_{p+1}}$
in which $\pi_{p+1}$ is the projection map, $j$ is an embedding defined by
$j(\xi,x)=(\xi+x,x)$, and $\sigma_{p+1}=\pi_{p+1}|_{U_{p+1}}$ so that
$\sigma_{p+1}(\xi,x)=\xi+x$. Write $pr\colon U_{p+1}\rightarrow C$ to be the
projection map to $C$.
###### Definition 2.1.
Let $B$ be a line bundle on $C$. For $p\geq 0$, define
$E_{p+1,B}=\sigma_{p+1,*}(pr^{*}B)\text{ and }N_{p+1,B}=\det E_{p+1,B}.$
###### Remark 2.2.
For basic properties of the vector bundles $E_{p+1,B}$ and the line bundle
$N_{p+1,B}$, we refer the reader to the paper [ENP20]. Here we mention that
$N_{p+1,B}=S_{p+1,B}(-\delta_{p+1})$, where $S_{p+1,B}$ is the invariant
descend of
$B^{\boxtimes p+1}=\underbrace{B\boxtimes\cdots\boxtimes B}_{\text{$p+1$
times}}$
on $C^{p+1}$ to $C_{p+1}$ under the action of the permutations group
$\mathfrak{S}_{p+1}$ on $C^{p+1}$ and
$\mathscr{O}_{C_{p+1}}(-\delta_{p+1})=N_{p+1,\mathscr{O}_{C}}$.
Let $B$ be a $p$-very ample line bundle on $C$. As the fiber of $E_{p+1,B}$
over $\xi\in C_{p+1}$ is $H^{0}(B|_{\xi})$, the evaluation map
$H^{0}(B)\otimes\mathscr{O}_{C_{p+1}}\rightarrow E_{p+1,B}$ on global sections
is surjective. Define $M_{p+1,B}$ to be the kernel bundle of the evaluation
map. We obtain a short exact sequence
$0\longrightarrow M_{p+1,B}\longrightarrow
H^{0}(B)\otimes\mathscr{O}_{C_{p+1}}\longrightarrow E_{p+1,B}\longrightarrow
0.$
The following vanishing theorem about the kernel bundle $M_{p+1,B}$ is an
immediate consequence of Ranthmann’s vanishing theorem on Cartesian products
of the curve.
###### Proposition 2.3.
Let $B$ be a $p$-very ample line bundle on $C$, and $L$ be a globally
generated line bundle on $C$ such that $h^{1}(L)=h^{1}(L\otimes B^{-1})=0$.
Then one has
$H^{k}(C_{p+1},\wedge^{m}M_{p+1,B}\otimes N_{p+1,L})=0,\text{ for all
}k>0,m>0.$
###### Proof.
By [Ra16, Theorem 3.1], one has the vanishing
$H^{k}(C^{p+1},q^{*}(\wedge^{m}M_{p+1,B}\otimes N_{p+1,L}))=0\text{ for all
}k>0,m>0,$
where $q\colon C^{p+1}\rightarrow C_{p+1}$ is the quotient map. Since $q$ is
finite, $\mathscr{O}_{C_{p+1}}$ is a direct summand of
$q_{*}\mathscr{O}_{C^{p+1}}$. Thus by projection formula,
$\wedge^{m}M_{p+1,B}\otimes N_{p+1,L}$ is a direct summand of
$q_{*}(q^{*}(\wedge^{m}M_{p+1,B}\otimes N_{p+1,L}))$, from which the result
follows. ∎
Next we prove a crucial property of kernel bundle $M_{p+1,B}$, which is
important for us to use the inductive argument.
###### Proposition 2.4.
Let $B$ be a $p$-very ample line bundle. There is a short exact sequence
$0\longrightarrow\sigma^{*}_{p+1}M_{p+1,B}\longrightarrow
M_{p,B}\boxtimes\mathscr{O}_{C}\longrightarrow(\mathscr{O}_{C_{p}}\boxtimes
B)(-U_{p})\longrightarrow 0.$
###### Proof.
Denote by $\alpha\colon
M_{p,B}\boxtimes\mathscr{O}_{C}\rightarrow(\mathscr{O}_{C_{p}}\boxtimes
B)(-U_{p})$ the morphism appeared on the right hand side of the sequence. We
first show that it is surjective. Indeed, choose any $\xi\in C_{p}$, and
consider the fiber $C=\\{\xi\\}\times C\subseteq C_{p}\times C$ over $\xi$.
Restricting $\alpha$ to this fiber yields the evaluation map
$\alpha_{\xi}:H^{0}(B(-\xi))\otimes\mathscr{O}_{C}\longrightarrow B(-\xi).$
Since $B$ is $p$-very ample and $\xi$ has degree $p$, it follows that
$B(-\xi)$ is $0$-very ample and thus globally generated. Hence $\alpha_{\xi}$
is surjective. This means that $\alpha$ is surjective.
Next we consider the following fiber product diagram
$\begin{CD}C_{p}\times C\times C@>{\bar{\sigma}}>{}>C_{p+1}\times C\ \supseteq
U_{p+1}\\\ @V{\bar{\pi}}V{}V@V{}V{\pi_{p+1}}V\\\ U_{p+1}=C_{p}\times
<EMAIL_ADDRESS>
On $C_{p}\times C\times C$, we have two divisors $D_{0}$ and $D_{1}$ defined
in the way that $D_{0}$ is the image of
$C_{p}\times C\longrightarrow C_{p}\times C\times
C,\quad(\xi,x)\longmapsto(\xi,x,x),$
and $D_{1}$ is the image of
$C_{p-1}\times C\times C\longrightarrow C_{p}\times C\times
C,\quad(\xi,y,x)\longmapsto(\xi+x,y,x).$
Observe that
$\bar{\sigma}^{*}U_{p+1}=D_{0}+D_{1}\text{ and }D_{0}\cap D_{1}=C_{p-1}\times
C.$
It is easy to check that
$\sigma^{*}_{p+1}M_{p+1,B}=\bar{\pi}_{*}(pr^{*}B(-D_{0}-D_{1})\text{ and
}M_{p,B}\boxtimes\mathscr{O}=\bar{\pi}_{*}pr^{*}B(-D_{1}),$
where $pr:C_{p}\times C\times C\rightarrow C$ is the projection to the right
hand side component $C$. Now we can form a short exact sequence on
$C_{p}\times C\times C$,
$0\longrightarrow\mathscr{O}(-D_{0}-D_{1})\longrightarrow\mathscr{O}(-D_{1})\longrightarrow\mathscr{O}_{D_{0}}(-D_{1})\longrightarrow
0.$
Note that $\mathscr{O}_{D_{0}}(-D_{1})=\mathscr{O}_{C_{p}\times C}(-U_{p})$.
Tensoring the short exact sequence with $pr^{*}B$ and then pushing it down to
$C_{p}\times C$, we obtain the desired short exact sequence. ∎
###### Remark 2.5.
The proof above shows that for any line bundle $B$ (not necessarily $p$-very
ample), one has a short exact sequence
$0\longrightarrow
pr^{*}B(-U_{p})\longrightarrow\sigma^{*}_{p+1}E_{p+1,B}\longrightarrow
E_{p,B}\boxtimes\mathscr{O}_{C}\longrightarrow 0$
on the universal family $U_{p+1}$.
## 3\. Proofs of Main Results
In this section, we prove the main results of the paper – Theorems 1.1 and
1.2. We keep using the notations introduced in Section 2. On the universal
family $U_{p+1}$, consider the short exact sequence
(3.0.1)
$0\longrightarrow\mathscr{O}_{U_{p+1}}\longrightarrow\mathscr{O}_{U_{p+1}}(U_{p})\longrightarrow\mathscr{O}_{U_{p}}(U_{p})\longrightarrow
0$
associated to the divisor $U_{p}$. The normal sheaf
$\mathscr{O}_{U_{p}}(U_{p})$ of $U_{p}$ in $U_{p+1}$ can be expressed as
$\mathscr{O}_{U_{p}}(U_{p})\cong(\mathscr{O}_{C_{p-1}}\boxtimes\omega^{-1}_{C})(U_{p-1}).$
Let $L$ be a line bundle on $C$. Tensoring $pr^{*}L$ with the short exact
sequence (3.0.1), we obtain a short exact sequence
$0\longrightarrow\mathscr{O}_{C_{p}}\boxtimes
L\longrightarrow(\mathscr{O}_{C_{p}}\boxtimes
L)(U_{p})\longrightarrow(\mathscr{O}_{C_{p-1}}\boxtimes
L\otimes\omega^{-1}_{C})(U_{p-1})\longrightarrow 0$
on $U_{p+1}$. Pushing it down to $C_{p}$ by the projection map
$\pi_{p}:U_{p+1}\rightarrow C_{p}$ yields a connecting map $\delta$ in the
associated long exact sequence
$0\longrightarrow
H^{0}(L)\otimes\mathscr{O}_{C_{p}}\longrightarrow\pi_{p,*}((\mathscr{O}_{C_{p}}\boxtimes
L)(U_{p}))\longrightarrow\sigma_{p,*}((\mathscr{O}_{C_{p-1}}\boxtimes
L\otimes\omega^{-1}_{C})(U_{p-1}))\stackrel{{\scriptstyle\delta}}{{\longrightarrow}}\cdots\hskip
56.9055pt$ $\hskip
199.16928pt\cdots\stackrel{{\scriptstyle\delta}}{{\longrightarrow}}H^{1}(L)\otimes\mathscr{O}_{C_{p}}\longrightarrow
R^{1}\pi_{p,*}((\mathscr{O}_{C_{p}}\boxtimes L)(U_{p}))\longrightarrow 0,$
where $\sigma_{p}$ is the restriction of $\pi_{p}$ onto the divisor $U_{p}$.
To understand the connecting map $\delta$, we consider its dual map
$\delta^{\vee}$ by applying
$\operatorname{\mathscr{H}om}(-,\mathscr{O}_{C_{p+1}})$. It is easy to
calculate that
$\big{(}\sigma_{p,*}((\mathscr{O}_{C_{p-1}}\boxtimes
L\otimes\omega^{-1}_{C})(U_{p-1}))\big{)}^{\vee}=\sigma_{p,*}(\mathscr{O}_{C_{p-1}}\boxtimes
L^{-1}\otimes\omega_{C})=E_{p,L^{-1}\otimes\omega_{C}}.$
Then the dual map $\delta^{\vee}$ turns out to be the evaluation map
$H^{0}(L^{-1}\otimes\omega_{C})\otimes\mathscr{O}_{C_{p}}\stackrel{{\scriptstyle\delta^{\vee}}}{{\longrightarrow}}E_{p,L^{-1}\otimes\omega_{C}}.$
We shall only need the special case that $L=\omega_{C}$. In this case, the map
$\delta^{\vee}$ splits $E_{p,\mathscr{O}_{C}}$ by the trace map. As a
consequence of the splitting, we have
$(\sigma_{p,*}\mathscr{O}_{U_{p}})^{\vee}\cong\sigma_{p,*}(\mathscr{O}_{U_{p}}(U_{p-1}))\cong\mathscr{O}_{C_{p}}\oplus\mathscr{K}_{p},$
where the direct summand $\mathscr{K}_{p}$ is the kernel sheaf of the
connecting map $\delta$ fitting into a short exact sequence
$0\longrightarrow
H^{0}(\omega_{C})\otimes\mathscr{O}_{C_{p}}\longrightarrow\pi_{p,*}((\mathscr{O}_{C_{p}}\boxtimes\omega_{C})(U_{p}))\longrightarrow\mathscr{K}_{p}\longrightarrow
0.$
###### Theorem 3.1.
Let $B$ be a $p$-very ample line bundle on $C$. Consider a line bundle
$L:=B\otimes\omega_{C}$. Suppose that $h^{0}(B)\geq p+k+2$ for $k\geq 1$. Then
(3.1.1)
$H^{i}(U_{p+1},\sigma^{*}_{p+1}(\wedge^{k}M_{p+1,B})\otimes(N_{p,L}\boxtimes
L))=0\quad\text{ for }i>0.$
As a consequence, one has
$H^{i}(C_{p+1},\wedge^{k}M_{p+1,B}\otimes N_{p+1,L})=0\quad\text{ for }i>0.$
###### Proof.
First observe that by [ENP20, Lemma 3.5],
$\mathscr{O}_{C_{p+1}}(-\delta_{p+1})$ is a direct summand of the vector
bundle
$\sigma_{p+1,*}(\mathscr{O}_{C_{p}}(-\delta_{p})\boxtimes\mathscr{O}_{C})$.
Thus the bundle
$\sigma_{p+1,*}(\sigma_{p+1}^{*}(\wedge^{k}M_{p+1,B})\otimes(N_{p,L}\boxtimes
L))\cong\wedge^{k}M_{p+1,B}\otimes
S_{p+1,L}\otimes\sigma_{p+1,*}(\mathscr{O}_{C_{p}}(-\delta_{p})\boxtimes\mathscr{O}_{C})$
contains $\wedge^{k}M_{p+1,B}\otimes N_{p+1,L}$ as a direct summand. Since
$\sigma_{p+1}$ is a finite map, the second vanishing statement in the theorem
would follow from the first one. Thus in the sequel, it suffices to show the
first vanishing statement (3.1.1).
To this end, we use the short exact sequence in Lemma 2.4 to yield a locally
free resolution of $\sigma^{*}_{p+1}(\wedge^{k}M_{p+1,B})$ as follows:
$\cdots\longrightarrow(\wedge^{k+2}M_{p,B}\boxtimes
B^{-2})(2U_{p})\longrightarrow(\wedge^{k+1}M_{p,B}\boxtimes
B^{-1})(U_{p})\longrightarrow\sigma^{*}_{p+1}(\wedge^{k}M_{p+1,B})\longrightarrow
0.$
Tensoring it with $N_{p,L}\boxtimes L$ gives rise to a resolution
$\cdots\longrightarrow(\wedge^{k+2}M_{p,B}\otimes N_{p,L})\boxtimes(L\otimes
B^{-2})(2U_{p})\longrightarrow(\wedge^{k+1}M_{p,B}\otimes
N_{p,L})\boxtimes(L\otimes B^{-1})(U_{p})\longrightarrow\cdots\hskip
56.9055pt$ $\hskip
227.62204pt\cdots\longrightarrow\sigma^{*}_{p+1}(\wedge^{k}M_{p+1,B})\otimes(N_{p,L}\boxtimes
L)\longrightarrow 0.$
We make the following claim:
###### Claim 3.1.2.
One has
$R^{t}pr_{*}\Big{(}(\wedge^{k+j}M_{p,B}\otimes N_{p,L})\boxtimes(L\otimes
B^{-j})(jU_{p})\Big{)}=0~{}~{}\text{ for }t\geq 1,j\geq 2,$
where $pr\colon U_{p+1}\rightarrow C$ is the projection map.
Proof of Claim. For a point $x\in C$, the restriction of the sheaf
$(\wedge^{k+j}M_{p,B}\otimes N_{p,L})\boxtimes(L\otimes B^{-j})(jU_{p})$ onto
the fiber $pr^{-1}(x)\cong C_{p}$ equals $\wedge^{k+j}M_{p,B}\otimes
N_{p,L(jx)}$, and $H^{t}(\wedge^{k+j}M_{p,B}\otimes N_{p,L(jx)})=0$ for $t>0$
by Proposition 2.3. Thus the claimed vanishing holds by base change.
By the claim above and using Larry spectral sequence
$H^{s}(R^{t}pr_{*}((\wedge^{k+j}M_{p,B}\otimes N_{p,L})\boxtimes(L\otimes
B^{-j})(jU_{p})))\Rightarrow H^{s+t}(U_{p+1},(\wedge^{k+j}M_{p,B}\otimes
N_{p,L})\boxtimes(L\otimes B^{-j})(jU_{p})),$
we see that
$H^{i}(U_{p+1},(\wedge^{k+j}M_{p,B}\otimes N_{p,L})\boxtimes(L\otimes
B^{-j})(jU_{p}))=0,\text{ for }i\geq 2,j\geq 2.$
Thus chasing through the resolution of
$\sigma^{*}_{p+1}(\wedge^{k}M_{p+1,B})\otimes(N_{p,L}\boxtimes L)$, in order
to prove the vanishing (3.1.1), the only left thing is to show the case when
$j=1$, i.e., to show
(3.1.3) $H^{i}(U_{p+1},(\wedge^{k+1}M_{p,B}\otimes
N_{p,L})\boxtimes\omega_{C}(U_{p}))=0,$
where we use the fact $L\otimes B^{-1}\cong\omega_{C}$. To do this, we tensor
$(\wedge^{k+1}M_{p,B}\otimes N_{p,L})\boxtimes\omega_{C}$ with the short exact
sequence (3.0.1). Pushing down the resulting sequence to $C_{p}$ by the
projection map $\pi_{p}:U_{p+1}\rightarrow C_{p}$, we obtain a long exact
sequence
$0\longrightarrow\wedge^{k+1}M_{p,B}\otimes N_{p,L}\otimes
H^{0}(\omega_{C})\longrightarrow\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\otimes\pi_{p,*}(\mathscr{O}_{C_{p}}\boxtimes\omega_{C}(U_{p}))\longrightarrow\ldots\hskip
56.9055pt$ $\hskip 71.13188pt\ldots\longrightarrow\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\otimes\sigma_{p,*}(\mathscr{O}_{U_{p}}(U_{p-1}))\stackrel{{\scriptstyle\delta}}{{\longrightarrow}}\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\longrightarrow\cdots$ $\hskip
227.62204pt\cdots\stackrel{{\scriptstyle\phantom{b}}}{{\longrightarrow}}R^{1}\pi_{p,*}(\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\boxtimes\omega_{C}(U_{p}))\longrightarrow 0.$
As in the discussion located before the theorem, the connecting map $\delta$
splits. This means that $R^{1}\pi_{p,*}(\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\boxtimes\omega_{C}(U_{p}))=0$ and $\wedge^{k+1}M_{p,B}\otimes N_{p,L}$
is a direct summand of $\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\otimes\sigma_{p,*}(\mathscr{O}_{U_{p}}(U_{p-1}))$. Thus we reduce the
vanishing (3.1.3) to showing the vanishing
(3.1.4) $H^{i}(C_{p},\wedge^{k+1}M_{p,B}\otimes
N_{p,L}\otimes\sigma_{p,*}(\mathscr{O}_{U_{p}}(U_{p-1})))=0.$
Observe that
$N_{p,L}\otimes\sigma_{p,*}(\mathscr{O}_{U_{p}}(U_{p-1}))=\sigma_{p,*}(N_{p-1,L}\boxtimes
L).$
By projection formula, the vanishing (3.1.4) would follow from the following
vanishing
(3.1.5)
$H^{i}(U_{p},\sigma^{*}_{p}(\wedge^{k+1}M_{p,B})\otimes(N_{p-1,L}\boxtimes
L))=0.$
Repeating this argument and noticing that $B$ is $(p-1)$-very ample with
$h^{0}(B)\geq(p-1)+(k+1)+2$, we finally reduce the problem to showing the
vanishing
$H^{i}(C,\wedge^{k+p}M_{B}\otimes L)=0,$
Here we write $M_{B}=M_{1,B}$ for simplicity. The only nontrivial case is when
$i=1$. Write $b=\operatorname{rank}M_{B}$ and notice that $\det
M^{\vee}_{B}=B$. By Serre duality,
$H^{1}(C,\wedge^{k+p}M_{B}\otimes L)\cong H^{0}(C,\omega_{C}\otimes\det
M^{\vee}_{B}\otimes\wedge^{b-1-k-p}M_{B}\otimes
L^{-1})^{\vee}=H^{0}(C,\wedge^{b-1-k-p}M_{B})^{\vee}.$
Now as $\wedge^{b-1-k-p}M_{B}$ is a direct summand of $\otimes^{b-1-k-p}M_{B}$
and the latter has no global sections, we conclude
$H^{1}(C,\wedge^{k+p}M_{B}\otimes L)=0$ as desired. This completes the proof.
∎
###### Proposition 3.2.
Let $B$ be a $p$-very ample line bundle on a curve $C$. Consider a line bundle
$L=B\otimes\omega_{C}$.
1. (1)
If $h^{0}(B)=p+k+1$ for $k\geq 1$. Then
$H^{i}(C_{p+1},\wedge^{k}M_{p+1,B}\otimes
N_{p+1,L})=H^{i}(C_{p+1},S_{p+1,\omega_{C}})=S^{p+1-i}H^{0}(\omega_{C}).$
2. (2)
If $h^{0}(B)=p+k$ for $k\geq 1$, then $\wedge^{k}M_{p+1,B}=0$, and therefore
$H^{i}(C_{p+1},\wedge^{k}M_{p+1,B}\otimes N_{p+1,L})=0.$
###### Proof.
For (1), since $M_{p+1,B}$ has rank $k$ and $\wedge^{k}M_{p+1,B}\cong
N^{-1}_{p+1,B}\cong S_{p+1,B^{-1}}(\delta_{p+1})$, we compute
$\wedge^{k}M_{p+1,B}\otimes N_{p+1,L}\cong S_{p+1,L\otimes B^{-1}}\cong
S_{p+1,\omega_{C}}.$
The result then follows from [ENP20, Lemma 3.7]. For (2), since $M_{p+1,B}$
has rank $k-1$, the result follows immediately. ∎
We will only need Theorem 3.1 and Proposition 3.2 for the case $k=1$. In the
following proposition, we classify when a $p$-very ample line bundle $B$ can
have $h^{0}(B)\leq p+2$.
###### Proposition 3.3.
Let $B$ be a $p$-very ample line bundle on $C$, and $p\geq 0$.
1. (1)
$h^{0}(B)=p+1$ if and only if either $p=0$ and $B=\mathscr{O}_{C}$ or $p\geq
1$, $C=\mathbf{P}^{1}$ and $B=\mathscr{O}_{\mathbf{P}^{1}}(p)$.
2. (2)
$h^{0}(B)=p+2$ if and only if one of the following cases holds.
1. (i)
$g=0$, $p\geq 0$ and $B=\mathscr{O}_{\mathbf{P}^{1}}(p+1)$.
2. (ii)
$g=1$, $p\geq 0$ and $\deg B=p+2$.
3. (iii)
$g\geq 2$, either $p=0$ and $B$ is a base point free pencil, or $p=1$ and
$C\subseteq\mathbf{P}(H^{0}(B))$ is a plane curve of degree $\geq 4$.
###### Proof.
(1) If $p=0$, then $B$ is a globally generated line bundle with
$H^{0}(C,B)=1$. Then $B=\mathscr{O}_{C}$ since the only section of $B$ is
nowhere vanishing. Assume $p\geq 1$, so $B$ is very ample and gives an
embedding of $C$ into the space $\mathbf{P}^{p}=\mathbf{P}(H^{0}(B))$. As $B$
is $p$-very ample, any $p+1$ points of $C$ will span the whole space
$\mathbf{P}^{p}$, which means that the degree of $C$ would be smaller than
$p$. But $C$ is also nondegenerate in $\mathbf{P}^{p}$ and thus has degree
$\geq p$. Hence $C$ has degree exactly $p$, and therefore, it is a rational
normal curve.
(2) Since (i) and (ii) are obvious, we only need to prove (iii). If $p=0$,
then $B$ is a base point free pencil. Assume that $p\geq 2$. Take $p-1$ points
$x_{1},\ldots,x_{p-1}$ of $C$, and put $D:=x_{1}+\cdots+x_{p-1}$. Since $B$ is
a $p$-very ample, we see that $B(-D)$ is very ample with $h^{0}(B(-D))=3$ and
$h^{1}(B(-D+x_{1}))=h^{1}(B(-D))=h^{1}(B)$. This means $C$ is a plane curve of
some degree $d\geq 4$ embedded by $B(-D)$ into $\mathbf{P}^{2}$, and thus, the
canonical line bundle $\omega_{C}$ has the form $\omega_{C}=(B(-D))^{d-3}$ by
the adjunction formula. By duality, the equality
$h^{1}(B(-D+x_{1}))=h^{1}(B(-D))$ is the same as the equality
$h^{0}((B(-D))^{d-4}(-x_{1}))=h^{0}((B(-D))^{d-4})$, which is impossible
because $B(-D)$ is very ample. Thus we conclude $p=1$ and
$C\subseteq\mathbf{P}(H^{0}(B))$ is a plane curve of degree $\geq 4$. ∎
Recall that the gonality of $C$ captures the positivity of the canonical line
bundle $\omega_{C}$. More precisely, $\operatorname{gon}(C)\geq p+2$ if and
only if $\omega_{C}$ is $p$-very ample. In particular,
$\operatorname{gon}(C)=\max\\{p+2\mid\text{$\omega_{C}$ is $p$-very
ample}\\}.$
We can compare the gonality with the genus. The following proposition may be
well-known.
###### Corollary 3.4.
Assume that $g\geq 2$. Then $g\geq\operatorname{gon}(C)$, and the equality
holds if and only if either $g=2$ or $C$ is a plane quartic curve.
###### Proof.
Since $g\geq 2$, it follows that $\operatorname{gon}(C)\geq 2$. Write
$\operatorname{gon}(C)=p+2$. Then $\omega_{C}$ is $p$-very ample. Applying
Proposition 3.3 to the case $B=\omega_{C}$, we see that $g\geq p+2$ and the
equality holds if either $g=2$ (i.e., $g=\operatorname{gon}(C)=2$), or $C$ is
a plane curve of $g=3$ which is a plane quartic curve (i.e.,
$g=\operatorname{gon}(C)=3$). ∎
###### Proof of Theorem 1.2.
In (1) and (2), $B$ is ample and thus $h^{1}(L)=0$. This implies
$h^{1}(C_{p+1},N_{p+1,L})=0$ and thus [EL15, Lemma 1.1] yields
$K_{p,1}(C,B;L)=H^{1}(C_{p+1},M_{p+1,B}\otimes N_{p+1,L}).$
So the assertion (1) follows from Theorem 3.1 by taking $k=1$, and the
assertion (2) follows from Proposition 3.2 by taking $k=1$. For the assertion
(3), if $p=0$, then $B=\mathscr{O}_{C}$ and then $K_{0,1}(C;\omega_{C})=0$ by
definition of Koszul cohomology group. If $p\geq 1$, then by Proposition 3.3,
$C=\mathbf{P}^{1}$ and $B=\mathscr{O}_{\mathbf{P}^{1}}(p)$. Then
$K_{p,1}(\mathbf{P}^{1},\mathscr{O}_{\mathbf{P}^{1}}(p);\mathscr{O}_{\mathbf{P}^{1}}(p-2))=0$
by a direct computation. ∎
###### Corollary 3.5.
Assume that $g\geq 2$. Let $B$ be a $p$-very ample line bundle on $C$, and $L$
be a line bundle on $C$. Suppose that $\deg(L\otimes B^{-1})\geq 2g-2$. Then
one has
$K_{p,1}(C,B;L)=0$
unless $L=B\otimes\omega_{C}$ and either $(1)$ $p=0$ and $B$ is a base point
free pencil, or $(2)$ $p=1$ and $C\subseteq\mathbf{P}(H^{0}(B))$ is a plane
curve. In the exceptional cases, $K_{p,1}(C,B;L)\neq 0$ but
$K_{p-1,1}(C,B;L)=0$.
###### Proof.
If $L\otimes B^{-1}\neq\omega_{C}$, then $h^{1}(L\otimes B^{-1})=0$ so that
one can use Rathmann’s theorem [Ra16, Theorem 1.1] to get the desired result.
Assume that $L\otimes B^{-1}=\omega_{C}$. By Theorem 1.2, $K_{p,1}(C,B;L)=0$
if $h^{0}(B)\neq p+2$, and $K_{p,1}(C,B;L)\neq 0$ if $h^{0}(B)=p+2$. In the
latter case, $K_{p-1,1}(C,B;L)=0$ by Theorem 1.2 since $B$ is $(p-1)$-very
ample and $h^{0}(B)=(p-1)+3$. However, if $h^{0}(B)=p+2$, then Proposition 3.3
shows that either $(1)$ $p=0$ and $B$ is a base point free pencil, or $(2)$
$p=1$ and $C\subseteq\mathbf{P}(H^{0}(B))$ is a plane curve. ∎
###### Proof of Theorem 1.1.
By Green–Lazarsfeld’s nonvanishing theorem [Gr84, Appendix] and the duality
theorem [Gr84, Theorem 2.c.6], we only need to know when
$K_{\operatorname{gon}(C)-2,1}(C,\omega_{C};L)=0$ vanishes. As $\omega_{C}$ is
$(\operatorname{gon}(C)-2)$-very ample, the theorem follows from Corollary
3.5. ∎
## References
* [AN10] Marian Aprodu and Jan Nagel. Koszul Cohomology and Algebraic Geometry. University Lecture Series, 52, Amer. Math. Soc., Providence, RI, 2010.
* [EL15] Lawrence Ein and Robert Lazarsfeld. The gonality conjecture on syzygies of algebraic curves of large degree. Publ. Math. Inst. Hautes Études Sci., 122:301–313, 2015.
* [ENP20] Lawrence Ein, Wenbo Niu, and Jinhyung Park. Singularities and syzygies of secant varieties of nonsingular projective curves. Invent. Math., 222(2):615–665, 2020.
* [Gr84] Mark Green. Koszul cohomology and the geometry of projective varieties. J. Differential Geom., 19(1):125–171, 1984.
* [GL86] Mark Green and Robert Lazarsfeld. On the projective normality of complete linear series on an algebraic curve. Invent. Math., 83(1):73–90, 1986.
* [GL88] Mark Green and Robert Lazarsfeld. Some results on the syzygies of finite sets and algebraic curves. Compositio Math., 67(3):301–314, 1988.
* [La04] Robert Lazarsfeld. Positivity in algebraic geometry. I. Classical setting: line bundles and linear series. II. Positivity for vector bundles, and multiplier ideals. Springer-Verlag, 2004
* [Ra16] Juergen Rathmann. An effective bound for the gonality conjecture.
* [Voi1] Claire Voisin. Green’s generic syzygy conjecture for curves of even genus lying on a K3 surface. J. Eur. Math. Soc. (JEMS) 4 (2002), 363–404.
* [Voi2] Claire Voisin. Green’s canonical syzygy conjecture for generic curves of odd genus. Compositio Math. 141 (2005), 1163–1190.
|
# Gender Biases in Error Mitigation by Voice Assistants
Amama Mahmood<EMAIL_ADDRESS>Johns Hopkins University3400 N. Charles
StBaltimoreMarylandUSA21218 and Chien-Ming Huang<EMAIL_ADDRESS>Johns Hopkins University3400 N. Charles StBaltimoreMarylandUSA21218
(2018)
###### Abstract.
Commercial voice assistants are largely feminized and associated with
stereotypically feminine traits such as warmth and submissiveness. As these
assistants continue to be adopted for everyday uses, it is imperative to
understand how the portrayed gender shapes the voice assistant’s ability to
mitigate errors, which are still common in voice interactions. We report a
study (N=40) that examined the effects of voice gender (feminine, ambiguous,
masculine), error mitigation strategies (apology, compensation) and
participant’s gender on people’s interaction behavior and perceptions of the
assistant. Our results show that AI assistants that apologized appeared warmer
than those offered compensation. Moreover, male participants preferred
apologetic feminine assistants over apologetic masculine ones. Furthermore,
male participants interrupted AI assistants regardless of perceived gender
more frequently than female participants when errors occurred. Our results
suggest that the perceived gender of a voice assistant biases user behavior,
especially for male users, and that an ambiguous voice has the potential to
reduce biases associated with gender-specific traits.
error mitigation, smart speaker, gender stereotypes, ambiguous voices, voice
assistance
††copyright: acmcopyright††journalyear: 2018††doi: XXXXXXX.XXXXXXX††ccs:
Human-centered computing Empirical studies in HCI††ccs: Computing
Methodologies Artificial intelligence
Figure 1. In this project, we investigate how the portrayed gender of a voice
assistant, error mitigation strategy, and participant gender impact people’s
perceptions of and behavior towards the AI assistant in a collaborative task.
This figure illustrates that users react to errors made by voice assistants,
to apology or compensation offered by voice assistants to mitigate negative
effects of those errors, and to other aspects of interaction in general.
[Voice assistants elicit non-verbal reactions from user in different
conversational scenarios]The figure consists of images of users reacting to
three different conversational scenarios. There are total of 9 pictures. Each
picture has a person sitting on a chair conversing with the smart speaker
placed on a table in front of them. The The first three pictures show
reactions of people to errors made by AI agent. The reactions are laughter by
a male participant in the first picture, laughter by a female participant in
the second picture, and a male participant frowning in the third picture. Next
two pictures show reactions to mitigation statements that are; a frown by a
male participant and laughter by a female participant. Next 4 pictures show
reactions to other aspects of interactions which are; a female participant
laughing, a female participant nodding, a female participant smiling, and a
male participant raising eyebrows.
## 1\. Introduction
Intelligent voice assistants are increasingly being used for a variety of
assistive tasks in various personal and professional settings. For example,
voice assistants, such as Alexa and Siri, are capable of supporting people
while performing functional tasks, such as retrieving information, setting
alarms and reminders, sending texts, and making calls; and engaging in social
tasks, such as telling jokes, and playing music and games (e.g., “Simon
says”). Moreover, voice assistants are positioned to play an essential role in
future collaborative learning and work (Baker et al., 2021; Paraiso and
Barthès, 2006; Sáiz-Manzanares et al., 2020; Schmid et al., 2022).
These modern voice assistants are powered by data-driven deep learning models
to recognize speech commands and user intent to ensure smooth collaboration.
While becoming powerful, these data-driven models are susceptible to errors
(Pearl, 2016), and when not mitigated and managed, these errors are likely to
result in users distrusting and halting the collaboration. As a result,
growing research has studied various strategies for mitigating AI errors. For
instance, apology as an error mitigation strategy from an agent makes it
likeable (Mahmood et al., 2022) and perceived as polite (Lee et al., 2010),
whereas compensation increases users’ satisfaction after error recovery (Lee
et al., 2010).
On top of experiencing imperfect interactions with voice assistants, users
tend to personify (Habler et al., 2019) and characterize (Abercrombie et al.,
2021) these assistants as feminine (UNESCO, 2019). Such gender
characterization is detrimental (for instance, it welcomes sexualised comments
and insults (Curry and Rieser, 2019; Curry et al., 2020)) because it fortifies
women’s amenable and unassertive role in the society (Loideain and Adams,
2020; Guzman, 2016). Existing gender biases due to association of stereotypes
in human-agent interactions (Ahn et al., 2022; Lee et al., 2019) raise the
following research questions that we aim to investigate: “how gender biases
may potentially manifest when users interact with voice assistants,
particularly in the context of erroneous interactions” , and “how the
effectiveness of error mitigation strategies may be modulated by the portrayed
gender of voice assistants”. The growing availability of (imperfect) voice
assistants in digital services and the rooted gender biases in our society
propel our inquiry into the relationships among the portrayed gender of a
voice assistant, its mitigation strategy for errors, and the gender of the
user when the user is interacting with a voice assistant.
In this work, we experimentally manipulated the gender of voice assistant
(ambiguous, feminine, and masculine) and error mitigation strategy (apology
and compensation), and studied how these factors along with participant gender
(female (women) and male (men)) may influence people’s perceptions of and
behavior toward erroneous voice assistants. We developed a mock smart speaker
and conducted an in-person experiment where participants interacted with voice
assistants to complete a series of shopping tasks (Figure 1). In this study,
our emphasis is on the “assistance” scenarios, where commercial voice
assistants excel in tasks such as online shopping, setting reminders,
retrieving information, and creating lists. Therefore, we select the online
shopping task as it exemplifies a typical assistance-type task performed by
these voice assistants and has the potential for a longer, and multi-turn
interaction between user and the voice assistant.
Our experiment reveals several findings identifying nuanced dynamics among the
portrayed gender of a voice assistant, mitigation strategy, and participant
gender:
* •
Participants reacted more socially, with a higher number of non-verbal
reactions, to voice assistants that offer apology rather than compensation
when mitigating errors.
* •
Male participants interrupted voice assistants, regardless of their portrayed
gender and use of mitigation strategy, more often than female participants.
* •
Male participants interacted more socially with “apologetic” voice assistants
that were portrayed as feminine compared to their masculine counterparts.
Overall, we also observed that male participants reacted more socially to the
feminine assistants than masculine ones and that apology promotes the
perception of warmth for voice assistants while compensation increases
participants’ satisfaction of service recovery, and perception of assistants’
competence.
## 2\. Background and Related Work
Given that the primary focus of this paper is to explore gender biases, error
mitigation strategies, and their intersection in user interactions with voice
assistants, we first present related work on various error mitigation
strategies for repairing human-agent relationships damaged by inevitable
disruptions. We then provide an overview of previous studies that examine the
stereotypes associated with gendered AI agents, with a specific emphasis on
voice assistants.
### 2.1. Error Mitigation Strategies in Human-Agent Interaction
It is crucial to repair relationships between agents111In our discussion of
prior work, we have defined agents in a broad sense with respect to their
embodiment. These agents encompass a wide range of forms, such as physical
robots, AI aids, virtual avatars, voice assistants, and other similar
entities. and humans after service breakdowns for continued use of technology.
Mitigation strategies can help increase people’s acceptance of faulty agents
(Lee et al., 2010). Moreover, these mitigation strategies can have varying
effects on different dimensions of people’s perceptions of the agents. For
instance, a robot that apologizes for its mistake is considered more likeable
and increases people’s willingness to use it (Cameron et al., 2021); however,
it is not perceived equally capable to a robot that demonstrates the knowledge
of why the error occurred and how the error can be fixed (Cameron et al.,
2021). Within apologies, a neutral apology, as opposed to humorous one, is
found to be more sincere and effective in repairing human-agent relationship
after errors made by smart speakers (Ge et al., 2019). For voice-based
interactions, combining acceptance of blame for the error and seriousness of
tone while apologizing results in better user satisfaction after recovery from
errors (Mahmood et al., 2022); moreover, such agents are perceived to be more
intelligent and likeable compared to the agents that lacked either seriousness
or acceptance of blame (Mahmood et al., 2022).
While apologies create an “emotional shift toward forgiveness” to promote
reconciliation and regain trust during relationship repair in both human-human
interactions (Govier and Verwoerd, 2002) and human-agent interactions (Lee et
al., 2010), other mitigation strategies such as compensation may offer better
user satisfaction while recovering from service failures. For instance,
compensating with a free drink on a robot’s failure to bring correct drink to
the user resulted in higher service recovery satisfaction than acknowledging
and apologizing for the error (Lee et al., 2010). For deeper insights into the
comparative aspects of apology and compensation strategies, we can look into
the decades of research on mitigation strategies in repairing broken human-
human relationships, which has a stronger foundation compared to research on
human-agent interactions. For instance, human interactions in e-commerce show
that the type of trust violation (competence- or integrity-based) also affects
whether apologizing with internal or external attribution of blame, and
compensating equal to or more than the loss would be more effective in
regaining trust (Cui et al., 2018). Even in purely financial settings,
apologies can facilitate preservation of relationships especially in the case
of undercompensation that fails to redress the loss of the affected party;
moreover, not surprisingly, overcompensation also repairs the relationship
better than under or equal compensation (Haesevoets et al., 2013). Since in
human-human interactions, apology is identified as a relational while
compensation is termed as a financial (utilitarian) strategy (Haesevoets et
al., 2013), we may find similar preferences towards apologies and compensation
offered by intelligent agents. In the domain of human-robot interaction, we
see that users’ orientation towards service can affect how the error
mitigation strategies are perceived. People who want to maintain amicable
relationship with service providers (stronger relational orientation) prefer
the apology strategy and have as much of negative impression for compensation
strategy as for no strategy (Lee et al., 2010). However, people who care more
about quality of service than relationship with providers (stronger
utilitarian orientation) find the compensation strategy to be more appealing
(Lee et al., 2010). However, it is unclear how the preference towards apology
and compensation might hold for voice assistants that do not have
anthropomorphic physical embodiment. In this work, we seek to understand how
these two mitigation strategies compare in a voice-based interactions
contextualized in an online shopping task using a smart speaker.
### 2.2. Gender Stereotypes in Human-Agent Interaction
This section aims to provide an overview of the relevant literature on gender
stereotypes and biases during human-agent interactions; in particular, we
review the stereotypes and biases in 1) the broader literature on human-agent
interactions, and 2) specific research focused on voice assistants.
Additionally, we discuss the research regarding the utilization of different
voice variations for assistants as a means of combating gender stereotypes and
biases.
#### 2.2.1. Gender Stereotypes in Human-Robot and Human-AI Interactions
Design of “masculine” and “feminine” AI agents and robots may introduce gender
based stereotypes. The terms “masculine” or “feminine” when referring to
agents and robots encompass various methods that may contribute to their
gendering. These methods may include manipulating their appearance, voice, and
the use of specific names and pronouns in experiment instructions and surveys.
Such gendering can be influenced by one or a combination of these factors.
Thus, when we refer to an agent as “masculine” or “feminine”, it means an
agent that may be perceived to look, sound or act as a stereotypical male or
female.
Prior work in HRI supports that masculine robots are considered better at
stereotypical male tasks while feminine ones are perceived to be friendlier
and warmer (Nomura, 2017). Even with minimal gender markers, such as their
voices or names, people tend to react in alignment with gender stereotypes
seen in human interactions. For instance, robots with feminine names or voices
are expected to be emotionally intelligent (Chita-Tegmark et al., 2019).
Additionally, personality of the agent, perceived as feminine or masculine,
influences perception of trustworthiness, reliability and competence. Context
and task at hand further shape how the gendered agents are perceived in these
dimensions. For instance, a robot with a feminine personality reflected
through agreeable and warm statements may be considered less trustworthy,
reliable, and competent as opposed to one with masculine personality that is
dominant, confident, and assertive, in a male-oriented task (e.g., dispatching
taxis (Kraus et al., 2018)).
The perception of individuals’ competence is often associated with factors
such as intelligence, efficiency, and skill, while warmth is more related to
friendliness, sincerity, and helpfulness (Fiske et al., 2007). Thus,
considering masculine agents to be more competent while assigning higher
warmth to feminine agents (Ahn et al., 2022), depicts association of gender
stereotypes based on agent characteristics. Such biased perceptions also have
implications for consumer preferences when interacting with AI agents in
relation to different types of products. For example, in a chatbot-based
interaction, masculine AI is preferred while recommending utilitarian products
(Ahn et al., 2022) (such as headphones for work from home), which are
practical and require competence from the seller to successfully convince
buyers of their value (Bennett and Hill, 2012). On the contrary, feminine AI
fairs better in promoting hedonic products (related to happiness: drawing
supplies) (Ahn et al., 2022), where warmth in the salesperson’s personality
influences purchasing decisions (Bennett and Hill, 2012).
In summary, various factors such as agent appearance, voice, personality,
task, and context may significantly influence how the perceived gender of an
agent may cause stereotyping. For this study, our focus is primarily on
examining the interaction between gendered voices of AI agents and error
mitigation strategies in assistance-type task scenarios. We aim to understand
how apology (relational/hedonic strategy) and compensation
(financial/utilitarian strategy) interact with gendered voices of AI agents to
influence user’s perceptions.
#### 2.2.2. Stereotypes Associated with Gendering of Voice Assistants
Most voice assistants in our present digital world are gendered as female; for
example, Apple’s Siri, Amazon’s Alexa, and Google voice assistant all have
feminine voices set as default. These agents are intentionally feminized and
marketed as “fembots” to increase their familiarity among users (Phan, 2019;
Strengers and Kennedy, 2021; Woods, 2018). Gendering of commercial voice
assistants poses societal harm because association of female gender by default
to these personal voice assistants reinforces women’s submissive role of
taking and fulfilling orders in the society and welcomes harassment, abuse,
and discriminatory behavior (Loideain and Adams, 2020; Guzman, 2016; Curry et
al., 2020; Curry and Rieser, 2019). In recent work, concerns about
stereotypical assignment of female gender to agent voices have been raised,
and power dynamic between the user and the woman-like voice assistants have
been discussed (Hwang et al., 2019). A report by United Nations (UN) (UNESCO,
2019) states that despite what the creators claim, “nearly all of these
assistants have been feminized—in name, in voice, in patterns of speech and in
personality. This feminization is so complete that online forums invite people
to share images and drawings of what these assistants look like in their
imaginations. Nearly all of the depictions are of young, attractive women.”
The influence of context and task on gender biases. Prior research has
yielded mixed results on the influence of task on gender biases in voice
assistants. On one hand, gender biases exist in the preferences for voice
assistants and are dependent on contextual settings and task at hand. For
example, a masculine voice is often favored for informative tasks, while a
feminine voice is preferred for social interactions (Lee et al., 2019).
Similarly, feminine voice assistant gained more trust in automation system
targeted for home while masculine voice assistant gained more trust in office
setting (Damen and Toh, 2019). Moreover, in health-critical situations, users
exhibit greater trust in feminine voice assistants compared to their masculine
counterparts (Goodman and Mayhorn, 2023). However, on the other hand, type of
task may not always affect the perception of voice assistants. For instance,
no gender biases were seen while performing functional (information retrieval
and mathematics calculations) and social (humours questions) tasks with Siri
(Lee et al., 2021). Such contradictory evidence on how task, especially
assistive tasks such as controlling automated devices (Damen and Toh, 2019)
and retrieving information (Lee et al., 2021), may affect the perception of
voice assistants motivated our choice of an assistance-type task—online
shopping—in our investigation of gender biases in error mitigation by voice
assistants.
The influence of participant gender on gender biases. In addition to context
and task, user’s own gender may shape their perceptions of voice assistants.
For instance, prior work has demonstrated effect of participant’s gender on
their perception of trust (Lee et al., 2021). Participants tend to trust
voices of the same gender more, even when uncertain about the accuracy of the
advice or suggestion being given by voice assistants such as Siri (Lee et al.,
2021). Additionally, female participants have shown a higher level of
skepticism compared to male participants, regardless of whether the voice they
were interacting with belonged to a masculine or feminine assistant (Lee et
al., 2021). On the contrary, when interacting with animated pedagogical
agents, female students showed a preference for a matched gender, whereas male
students rated the female agent more favorably (Ozogul et al., 2013).
Similarly, other studies indicate varied gender preferences (Krämer et al.,
2016; Kim et al., 2007), suggesting existence of potential gender stereotypes
and biases. Therefore, in this work, we also look at the role of gender of
users, in addition to gender of assistants, in association of stereotypes with
voice assistants.
#### 2.2.3. Intervention for Mitigating Gender Stereotypes and Biases
To address the issue of stereotypes and biases associated with gendered
voices, researchers have explored various methods to mitigate their influence.
One approach discussed to mitigate or eliminate the association of stereotypes
with gendered voices is the creation of ambiguous voices through the
manipulation of voice characteristics, such as pitch. Research has shown that
there were no discernible differences in user trust formation between voice
assistants with gendered voices, and those with claimed ambiguous voices
(Tolmeijer et al., 2021); hence, ambiguous voices can potentially be used
commercially in place of gendered assistants as a tool to reduce biases.
However, it is worth mentioning that the gender ambiguous voice employed in
the aforementioned study is unable to completely eradicate implicit biases,
which are still evident similar to the biases associated with the feminine
voice(Tolmeijer et al., 2021), drawing our attention to lack of empirical
evidence for effectiveness of ambiguous voices to culminate gender biases,
implicit or explicit. Hence, in this work, in addition to masculine and
feminine sounding assistants, we also study effects of ambiguous sounding
assistants on people’s perceptions of voice assistants in an online shopping
task using a smart speaker.
#### 2.2.4. Gaps in Methodological Approach for Empirical Studies on Gender
Biases
Within the context of empirical studies on gender biases, a significant gap
exists in the methodological approach employed. Most of prior studies on
exploration of gender stereotypes and biases in voice assistants are based on
online reviews (Phan, 2019), crowd sourced evaluations (Curry and Rieser,
2019), case studies (Woods, 2018), surveys (Obinali, 2019; Curry et al.,
2020), chat bots, and online interfaces (Damen and Toh, 2019; Tolmeijer et
al., 2021; Goodman and Mayhorn, 2023). While there are a few exceptions, such
as studies conducted with phone-based voice assistant Siri (Lee et al., 2021),
or using off-the-shelf speakers controlled by experimenters (Habler et al.,
2019) , these studies primarily depend on subjective evaluations through
questionnaires to gain insights into stereotypes. In our work, we strive to
address the methodological gap by introducing a simulated physical smart
speaker platform through the implementation of the Wizard of Oz (WoZ) paradigm
for the voice assistant. Moreover, we augment the subjective measures with an
analysis of users’ behavioral data to investigate implicit expression of
biases associated with gender stereotypes.
### 2.3. Hypotheses
Based on the prior work on mitigation strategies to repair human-agent
relationship, and understanding various factors that impact gender stereotypes
in human-agent interaction outlined above, we propose the following
hypotheses:
* •
Hypothesis 1. Compensation, as a mitigation strategy, improves perceptions of
service recovery satisfaction and competence more than apology while apology
increases perceptions of warmth of the assistants. Hypothesis 1 is informed by
previous research showing that 1) compensation is a financial strategy while
apology is a relational strategy (Haesevoets et al., 2013), and 2) apologies
make an agent or robot more likeable (Cameron et al., 2021) and appear more
polite (Lee et al., 2010), while compensation results in better service
recovery satisfaction (Lee et al., 2010).
* •
Hypothesis 2. Feminine sounding assistants are perceived warmer than masculine
sounding assistants whereas masculine assistants are considered more
competent. Hypothesis 2 is informed by previous work showing that masculine AI
agent is rated more competent while feminine AI agent is more likeable and
warm (Ahn et al., 2022; Nomura, 2017; Chita-Tegmark et al., 2019). In addition
to testing this hypothesis, we will explore if the gender ambiguous voice can
culminate these biases.
* •
Hypothesis 3. Assistants with feminine voice gain better user perceptions when
they apologize for their mistake while assistants with masculine voice are
perceived more positively when they compensate for their mistake. Hypothesis 3
is formulated based on evidence from prior work in human-human interactions
showing that 1) typically women are more apologetic than men (Turiman et al.,
2013), 2) men are characterized to have authoritative and practical roles
while women are considered better at submissive roles (Powell et al., 2002;
Koenig et al., 2011); and 3) apology is a relational strategy while offering
compensation is a financial (utilitarian) strategy (Haesevoets et al., 2013)
and may be considered as a feminine and a masculine trait, respectively.
Moreover, we explore where gender ambiguous voices would fall on this spectrum
and if ambiguous assistants can reduce gender biases faced by AI agents while
still being perceived as effective.
Lastly, as discussed above, prior work on how participant gender may influence
perception of masculine or feminine agents and robots has shown conflicting
findings (Lee et al., 2021; Mitchell et al., 2011; Siegel et al., 2009). Thus,
in this study we also consider possible influence of participants’ gender on
their perceptions of gendered voice assistants while mitigating AI errors.
Moreover, we conduct this study in the context of an assistance-type task and
contextualize our results in that space.
## 3\. Methods
### 3.1. Study Design and Experimental Task
To test our hypotheses, we conducted an in-person experiment considering three
factors: error mitigation strategy—(apology vs. compensation), gendered voice
of AI assistants—(ambiguous vs. feminine vs. masculine), and participant
gender—(female vs. male). Error mitigation strategy and gendered voice of AI
assistants were within-subjects factors while participant gender was a
between-subjects factor.
Our experiment consisted of six online shopping tasks; for each task,
participants were asked to collaborate with one of the six AI voice assistants
to complete the task. We presented the six tasks as beta testing of a voice
assistant developed for our prototype smart speaker. Each task involved
working with a different voice assistant on a different shopping list. The six
AI assistants were named Galaxy, Dot, Cloud, Mist, Phoenix, Gamma. We picked
non-human and gender neutral names to reduce any biases in participant’s
preconception of the assistant’s gender. The order of the assistant names and
shopping lists was same for all the participants. For instance, the assistant
in the first session is always named Galaxy and the participant ordered items
for a party. Balanced Latin square is used to counterbalance the six
experimental conditions.
In each task, the user was given a list of five items to order through a
speech-based interaction with the AI assistant (Fig. 2). The user would prompt
the assistant to order a specific item and confirm the addition of item to the
cart once the assistant presents the user with the item it found. The
assistant would let the user know after adding the item to the cart. The user
could also ask the assistant to show more options. The AI assistant would make
a mistake on either the second, third, or fourth item on the list. In this
study, we considered intent recognition errors, which are described as
“recognized but incorrectly” (Pearl, 2016). The predetermined errors in this
study were the use of homonyms—words that sound the same but may mean
different things. For example, “flour” could be heard as flour or flower. As
there are only a few possible items, mostly two, associated with the chosen
homonyms, the assistant recovered from the error in the second attempt by
suggesting correct item.
The user could report a mistake by verbally indicating that the item suggested
by the assistant is not the correct one when the assistant asked if it should
add the item to the cart. On indication of error by the user, the assistant
prompted them to try again. On the second try, the assistant fixed the mistake
by indicating the correct item. After fixing the error, the assistant
performed the recovery strategy by initiating either apology or offering
compensation based on the experimental condition. After ordering the five
items on the list, the user was directed to solve a riddle while they “waited
for the items to arrive”. The experimenter handed over the ordered items
printed on a piece of paper with title and images to the user while they are
solving the riddle. The user was prompted to see if the ordered items were
correct. The user could start a return process with the assistant in case any
of the arrived items was incorrect. The assistant would initiate the return
and notify that the item was returned and refunded. It was up to the user if
they wanted to order the required item after returning the incorrect one or
not. The assistant either apologized or offered compensation based on
experimental condition once the return process was complete. Next, we describe
the implementation details of our mock smart speaker used in the experiment.
Figure 2. Experimental setup. The participant interacts with the smart speaker
using voice commands. The wired speaker is connected to the laptop. The LED
lights ring on the speaker is activated by the Arduino once either of the two
applications running on the laptop signals that assistant or participant
speech is detected. The experimenter accesses the web interface on the laptop
to control the speech of the mock smart speaker using TeamViewer on the
Desktop (WoZ). The participant and the experimenter are in the same room
separated by a physical partition.
[Experimental set up explaining WoZ interface and interaction with user.]The
figure shows our experimental set up. To the left of the figure, the
experimenter is shown with the desktop computer. To the right of a physical
partition which separates the participant from the experimenter, the
participant is shown with a wired speaker connected to Arduino and the laptop.
To the very right, a picture of actual participant interacting with the smart
speaker is shown where the person is sitting in front of the speaker which is
placed on a table. The WoZ remote control is indicated between the desktop and
the laptop. The experimenter simulates the assistant speech by controlling the
laptop next to the wired speaker remotely through the desktop computer.
### 3.2. Mock Smart Speaker Setup
We developed a mock smart speaker (Fig. 2) that can be controlled by an
experimenter through Wizard of Oz (WoZ) paradigm. The speaker consisted of
following components:
* •
Arduino controlled LED lights ring for signaling when the assistant was
speaking or listening. LED lights were activated for signaling assistant’s
speech as soon as the experimenter played any audio on behalf of the
assistant. For listening, the the LED light ring was activated on wake words
which were either the name of the assistant, “Hello”, “Hey”, or “Hi” for the
first time the assistant is addressed. Once the interaction started, if any
speech from participant was detected right after the wake work or the
assistant’s turn of speaking (assistant waits for the participant’s response
after talking every time), the LED ring lighted up indicating that the speaker
was listening to them. Each assistant had a different colored ring (blue,
purple, green, pink, orange, and yellow) which was decided by the row of Latin
square the participant was assigned to. We coded this behavior using two
applications in $C\\#$; 1) to detect audio output from the assistant, and 2)
to detect and recognize the participant’s speech. On detection of either
participant or assistant’s speech (as described above), the respective
application will send a signal to Arduino to activate the lights ring. Both
applications are running on the laptop (see Fig. 2).
* •
Wired speakers for playing audio were connected to the laptop (see Fig. 2. The
laptop powered and controlled the wired speaker, and the Arduino. The laptop
was controlled remotely by the experimenter using
TeamViewer222https://www.teamviewer.com/en-us/ on a desktop computer. The
laptop screen was closed so that it is not visible to the participant. An iPad
is also placed next to the wired speaker for the participants to complete
puzzles and surveys throughout the study.
* •
Wizard of Oz interface is used to control the smart speaker’s audio. The
interface website, hosted on GitHub, dictated the flow of experiment and was
running on the laptop connected to the wired speakers. The experimenter
operated the website remotely using TeamViewer on the desktop computer to
control the assistant on the mock smart speaker using WoZ paradigm. The
experimenter could click on various buttons to activate assistant responses to
the participant. Apart from standard responses for each item in each shopping
list such as “Okay, I found Duracell Double A batteries. Should I add it
cart”, “Let’s try that again” for fixing the error, and “The item has been
added to your cart”, the web interface also had generic responses to handle a
participant’s divergent requests. For instance, statements such as “I am doing
alright. How I can help you today?” to respond to the participant’s questions
about how the assistant is doing. The simple responses such as “Yes”, “No”,
“Hi”, and “How can I help you today?” were also included. For requests that
could not be fulfilled by the speaker such as the participant asking to read
out all the items in their cart, the assistant responded with “Unfortunately,
that is currently beyond my capabilities.” and presented a possible solution
to continue the conversation, “but you can ask me if a specific item is in the
cart or not?”. To handle such specific queries, we also added “Yes, that item
is already in the cart” and “No, that item is not in the cart yet, would you
like me to add it?”. If the participant requests to remove any particular item
from the cart, the assistant obliged and responded, “The item has been removed
from the cart.” If the participant forgot to ask the voice assistant to check
out at the end of the task, the assistant prompted the participant by asking
“Would you like to check out?” On the participant’s request or confirmation to
check out, the assistant said “Okay, checking out. Your items should arrive
shortly”. To handle return process, there were a few statements at the
experimenter’s disposal such as “Okay, I can start a return process for the
item”, “Which item would you like to return?”, and “Your item has been
returned and a refund has been issued” in addition to the error mitigation
statement.
Next, we discuss design of our manipulations for error mitigation strategies
and gendered voices.
### 3.3. Manipulation of Error Mitigation Strategies
In this study, we focused on two error mitigation strategies, namely apology
and compensation; the strategy of apology has been employed in prior research
with a similar online shopping context (Mahmood et al., 2022), whereas the
strategy of compensation is relevant given the nature of the shopping task.
* •
Apology. The design of our apology strategy followed a similar design of error
responses in a prior study on online shopping with a smart speaker in a
simulated online experiment (Mahmood et al., 2022); in particular, the study
showed that an apology offered in a serious tone with the acceptance of blame
for the AI error is preferred over the apologies that lacked seriousness,
acceptance of blame, or both. In our study, the agent apologized for the AI
error by saying “I am sorry for the inconvenience. I confused the items
because there are multiple items for this keyword. From time to time, I have
difficulty distinguishing among such items.”
* •
Compensation. For the compensation strategy, the agent offered a 10 dollar
credit to the user by saying “I will give you a 10 dollar credit that can be
applied to your future purchases with us for the inconvenience caused.”
### 3.4. Manipulation of Gendered Voices
We manipulated the gender of AI voice to sound feminine, masculine and
ambiguous. We chose the feminine and masculine voices from two free text-to-
speech websites “Sound of Text”333https://soundoftext.com/ and
“wideo”444https://wideo.co/text-to-speech/ (US Jack Bailey), respectively. The
voices had American accent, and default pitch and speed. Inspired from prior
work (Tolmeijer et al., 2021), we generated gender ambiguous voices by
shifting the pitch of the above feminine and masculine voices by 2 or 3
semitones where shifting by one semitone is equivalent to shifting by 5.946%.
We used an online pitch shifting tool555https://onlinetonegenerator.com/pitch-
shifter.html. The masculine voice was increased by 2 and 3 semitones while the
feminine voice was lowered by 2 and 3 semitones to obtain 4 candidates for
gender ambiguous voices. To select a gender ambiguous voice, we conducted two
pilot studies:
#### 3.4.1. Pilot Study 1: Picking 2 gender ambiguous voices from 4
candidates
To pick one ambiguous voice between the shifted masculine voices and one
between the shifted feminine voices, we conducted an online poll with nine
participants who listened to the 4 voices and were asked to choose two
ambiguous voices, one from each gender category. Moreover, they were asked to
rank the voices from most ambiguous to least ambiguous. Six of the 9
participants picked feminine-3 semitones to be more ambiguous among shifted
feminine voices which is consistent with the results from prior study
(Tolmeijer et al., 2021). Whereas, seven of the 9 participants picked the
masculine+2 semitones as more ambiguous from the shifted masculine voices.
Overall, masculine+2 semitones was picked by eight participants to be in top 2
when they were asked to rank the voices for most ambiguous to least ambiguous
voices. While feminine-3 semitones was picked by four participants to be in
top 2. Based on the results from this poll, we kept masculine+2 semitones
(masculine-ambiguous) and feminine-3 semitones (feminine-ambiguous) as two
candidates for gender ambiguous voice for the next pilot study.
#### 3.4.2. Pilot Study 2: Picking 1 gender ambiguous voice from 2 candidates
and validating the manipulation of all gendered AI voices.
To verify that the gendered voices, especially ambiguous voices, are perceived
as intended regardless of the nature of a task, we ran a within-subjects study
with two factors: voice gender—(feminine, masculine feminine-ambiguous,
masculine-ambiguous) and nature of task—(stereotypically masculine, neutral,
stereotypically feminine). For the manipulation of task nature, we picked
following three statements narrated by the voice agents:
* •
Stereotypically masculine task (fixing a car). “If the car’s charging port is
not working, check to see if the fuse for the power outlet has blown”.
* •
Neutral task (defining a term). “A recession is a period of temporary economic
decline during which trade and industrial activity are reduced”.
* •
Stereotypically feminine task (nursing). “The doctor will be with you shortly.
Meanwhile, please fill out the consent forms”.
We recruited 16 participants (gender: 9 female, 7 male; age:
$M=23.33,SD=2.61$) through convenience sampling from the local community. The
participants were presented with 12 conditions in a randomized order. In each
condition, the participants listened to the voice of one of the AI agents
saying one of the above statements and then rated the agent in terms of
perceived gender (1 being feminine, 3 being ambiguous and 5 being masculine)
and humanlikeness (1 being robotic and 5 being human) on 5-point scales.
Furthermore, participants were asked to give a name to the agent. Two coders
independently assigned gender (masculine, feminine, or ambiguous) to the names
provided by the participants. In case of disagreement between two coders, they
discussed and reached a consensus on the perceived gender based on the name
given to the agent.
The data collected from the second pilot study was analyzed to select the
voices for the voice assistants. To validate the manipulation of perceived
gender of the AI voice, we compared the rating of perceived gender to the
baseline, the value 3 (ambiguous) on a 5-point scale. One-way analysis of
variance (ANOVA) revealed that the AI voice had a significant main effect on
perceived gender, $F(4,235)=146.04,p<.001$. Pairwise comparisons using
Dunnett’s test with control (baseline: perceived gender = 3, ambiguous)
revealed that all the voices except masculine-ambiguous ($M=3.25,SD=0.84$) is
significantly different from the baseline. The masculine voice
($M=4.52,SD=0.74$) is significantly higher on perceived gender (5-point scale,
1 being feminine, 3 being ambiguous, and 5 being masculine) than baseline
($M=3,SD=0$), $p<.001$ while the feminine voice ($M=1.14,SD=.36$), $p<.001$,
and feminine-ambiguous ($M=2.65,1.02$), $p=.046$, voices are significantly
lower on perceived gender than baseline. These results verified that our
manipulation for masculine and feminine voices was adequate. A chi-square test
of independence was performed to examine the effect of gender of AI voice and
the perceived gender of the names given by participants. The relation between
these variables was significant ($\chi^{2}(6,N=192)=169.21,p<.001$). Our
masculine-ambiguous AI voice was given gender ambiguous names ($66.7\%$) more
often than the feminine-ambiguous AI voice ($62.5\%$). Based on these results,
we concluded that masculine-ambiguous voice was perceived to be more gender
ambiguous than the feminine-ambiguous voice. As a result, we used the
masculine-ambiguous voice as a gender ambiguous voice in our main study.
Furthermore, on running two-way repeated measures ANOVA with voice gender and
nature of task as factors, there is no significant main effect of nature of
task on perceived gender. A two-way repeated measures ANOVA revealed no
significant main effect of gender and nature of task on humanlikeness.
However, there is a significant interaction effect between voice gender and
nature of task, $F(6,90)=2.46,p=.030,\eta^{2}=.141$; Tukey’s HSD post hoc test
shows no significant pairwise comparisons.
### 3.5. Measures
For the main study, we used a range of metrics to measure user perceptions of
agent characteristics, subjective evaluations of the voice assistants, and
user behavior towards the assistants.
#### 3.5.1. Perceptions of Agent Characteristics
* •
Perceived gender. To check if our manipulation of the gender of AI voices was
perceived as intended, we asked the participants to rate the voice gender on a
5-point scale (1 being feminine, 3 being ambiguous, and 5 being masculine).
* •
Humanlikeness. We asked the participants to rate humanlikness of the agent on
a 5-point scale (1 being robotic and 5 being human).
#### 3.5.2. Subjective Measures of User Experience and Perceptions of AI
* •
Service recovery satisfaction (Three items; Cronbach’s $\alpha=.91$). We used
three questions (“The AI assistant was responsive to any errors.”, “I am happy
with how the error was handled” and “In my opinion, the AI assistant provided
a satisfactory response to the error”) as informed by prior work on apology
(Roschk and Kaiser, 2013) in the domain of consumer services (Tax et al.,
1998; Maxham III and Netemeyer, 2003) to measure service recovery
satisfaction.
* •
Warmth (Four items; Cronbach’s $\alpha=87$). We asked the participants to rate
their impression of the agent on these dimensions: 1) Cold – Warm, 2)
Unfriendly – Friendly, 3) Insincere – Sincere, and 4) Boastful – Modest.
* •
Competence (Three items; Cronbach’s $\alpha=.86$). We asked the participants
to rate their impression of the agent on these dimensions: 1) Unskillful –
Skillful, 2) Impractical – Practical, and 3) Foolish – Intelligent as informed
by prior work on gender stereotypes on AI recommendations (Ahn et al., 2022)
and social cognition dimensions (Fiske et al., 2007) similar to ‘warmth’.
#### 3.5.3. Behavioral Measures
To understand participants’ behaviors toward voice assistants during their
interactions, we asked two human coders to annotate the interaction videos and
extracted the following behavioral measures based on the annotation. The two
coders annotated same 10% of the participant videos for inter-coder
reliability; the rest of the videos were 50:50 split between the two coders.
For dichotomous variables, percentage agreement was reported for assessing
inter-coder reliability. For continuous variables, Intraclass Correlation
Coefficient (ICC) estimates and their 95% confident intervals were calculated
and reported using SPSS assuming single measurement, consistency, two-way
mixed-effects model. ICC coefficients between 0.75 and 0.90, and greater than
0.90 are indicative of good and excellent reliability, respectively (Koo and
Li, 2016).
* •
Participant interruption during error (Dichotomous). This measure encodes
whether the participant interrupted the assistant after realizing that an
error had occurred while the assistant was still talking. Essentially, this
measure indicates if the participant started talking before the assistant
stopped. For instance, during the statement with error, “Okay, I found red
archery bow. Should I add it to the cart?”, the interruption would be recorded
if the participant starts talking before the assistant finishes saying “Should
I add it to the cart?”. There was a 100% agreement between the two coders.
* •
Non-verbal reaction to error (Dichotomous). This measure encodes whether the
participant reacted to the error non-verbally (e.g., changes in face and/or
upper body; see Fig. 1 for example reactions). There was a 100% agreement
between the two coders.
* •
Verbal response to mitigation statement (Dichotomous). We look at whether the
participant responds verbally to the error mitigation statement (compensation
or apology). For instance, participants saying “Thank you” and “It’s okay”
would would count as verbal responses to error mitigation strategy. There was
a 100% agreement between the two coders.
* •
Non-verbal response to mitigation statement (Dichotomous). We look at whether
the participant reacts non-verbally (changes in face and/or upper body; see
Fig. 1 for example responses) to the error mitigation statement (compensation
or apology). There was a 100% agreement between the two coders.
* •
Number of other interruptions (interruptions that were not concerned with the
error). Interruptions are defined as when the participant and the assistant
interrupts or talks over each other. This measure counts the number of
interruptions, apart form those to error as mentioned above, during the
interaction, trying to capture the overall interaction fluency. A good degree
of reliability was found between 2 raters, $ICC(3,1)=.835[.644,.928]$.
* •
Number of reactions (verbal and non-verbal). Verbal reactions are recorded
when the participant says phrases other than direct response to the speaker
such as “Thank you”, “You’re good” etc. Non-verbal reactions, such as smile,
laugh, and raising eye brows, are recorded when the participant’s face and/or
upper body (visible in video recordings) changes as reaction to the
assistant’s speech (see Fig. 1 for example non-verbal reactions). Involuntary
movements were not recorded as non-verbal reactions. We do not include verbal
and non-verbal responses to error in this measure because we are interested in
understanding, in general, when nothing unexpected occurs how social the
interaction of the participant with the smart speaker is. Hence, we focus only
on the conversational aspect of the interaction, which includes responses to
mitigation statements. A good degree of reliability was found between 2 raters
for number of verbal reactions, $ICC(3,1)=.837[.648,.929]$. Similarly, a good
degree of reliability was also found between 2 raters for number of non-verbal
reactions, $ICC(3,1)=.896[.766,.955]$.
### 3.6. Procedure
This study consisted of five phases:
1. (1)
Introduction and consent. At the start of the study, participants were
provided with a brief description of the study. The description stated that
participants would be ordering items from shopping lists while collaborating
with six AI assistants to test our smart speaker prototype. The participation
was voluntary; participants agreed to continue the study by signing the
consent form approved by our Institutional Review Board (IRB).
2. (2)
Pre-study questionnaire. The participants filled a pre-study questionnaire
inspired from prior work (Bradley and LaFleur, 2016) to evaluate users’
utilitarian or relational orientation regarding shopping experience.
3. (3)
Experimental task: simulated shopping. Participants were assigned one of the
rows in a Latin square of order 6, dictating the order of experimental tasks.
They interacted with the AI assistant on the mock smart speaker. Participants
were prompted to do a puzzle while waiting for their order to arrive. After
the experimenter handed them the items (printed on a piece of paper) that they
ordered, participants verified if the items were correct, and started a return
with the speaker if needed.
4. (4)
Perception survey. After interacting with the AI assistant, participants
completed a questionnaire about their perceptions of the AI assistant. They
continued onto the next condition and repeated phases 3 and 4.
5. (5)
Post-study questionnaire. After completing all the conditions, participants
filled out a post-study demographics questionnaire.
The study was approved by our IRB. The study took approximately 45 minutes to
complete. The participants were compensated with a $15 USD gift card for their
participation in the study.
### 3.7. Participants
We recruited 43 participants using university mailing lists and flyers around
the campus. Two of the participants were excluded from data analysis because
they failed to identify the targeted error correctly in 5 out of 6 trials. A
total of 41 participants (21 females (women), 19 males (men), 1 non-binary)
were considered for data analysis. The participants were aged between 18 to 58
($M=26.20,SD=7.92$) and had a variety of education backgrounds, including
computer science, engineering and technology, healthcare, life sciences, media
sciences, music, and education. Twenty seven participants indicated their
ethnicity to be Asian, 2 African American, 6 Caucasian, 5 Spanish origin of
any race, and 1 Caucasian and Spanish origin of any race.
Table 1. Results of perceived gender of agent voice(1: feminine, 3: ambiguous, and 5: masculine) and humanlikeness. Mixed-model repeated measures ANOVA for voice gender, mitigation strategy, and participant gender. ${}^{\ast}p\leq 0.05$, ${}^{\ast\ast}p\leq 0.01$, and ${}^{\ast\ast\ast}p\leq 0.001$ Perceived Voice Gender (1: feminine 3: ambiguous 5: masculine) | | |
---|---|---|---
Variables | $F$ | $p$ | $\eta_{p}^{2}$
Voice Gender | $F(2,76)=323.55$ | $<.001^{\ast\ast\ast}$ | $.895$
Feminine ($M=1.14,SD=0.44$) ¡ Masculine ($M=4.89,SD=0.34$) | | $<.001^{\ast\ast\ast}$ |
Feminine ($M=1.14,SD=0.44$) ¡ Ambiguous ($M=3.93,SD=1.18$) | | $<.001^{\ast\ast\ast}$ |
Ambiguous ($M=3.93,SD=1.18$) ¡ Masculine ($M=4.89,SD=0.34$) | | $<.001^{\ast\ast\ast}$ |
Mitigation Strategy | $F(1,38)=0.15$ | $.700$ | $.004$
Participant Gender | $F(1,38)=5.96$ | $.019^{\ast}$ | $.136$
Male ($M=3.17,SD=1.82$) ¡ Female ($M=3.45,SD=1.70$) | | |
Participant Gender $\times$ Voice Gender | $F(2,76)=2.86$ | $.063$ | $.070$
Participant Gender $\times$ Mitigation Strategy | $F(1,38)=3.77$ | $.072$ | $.083$
Voice Gender $\times$ Mitigation Strategy | $F(2,76)=1.69$ | $.192$ | $.043$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,76)=0.52$ | $.598$ | $.013$
Humanlikeness | | |
Variables | $F$ | $p$ | $\eta_{p}^{2}$
Voice Gender | $F(2,76)=21.18$ | $<.001^{\ast\ast\ast}$ | $.358$
Masculine ($M=2.96,SD=1.06$) ¡ Feminine ($M=3.36,SD=1.02$) | | $.003^{\ast\ast}$ |
Ambiguous ($M=2.59,SD=1.13$) ¡ Feminine ($M=3.36,SD=1.02$) | | $<.001^{\ast\ast\ast}$ |
Ambiguous ($M=3.93,SD=1.18$) ¡ Masculine ($M=2.96,SD=1.06$) | | $.007^{\ast\ast}$ |
Mitigation Strategy | $F(1,38)=0.21$ | $.646$ | $.006$
Participant Gender | $F(1,38)=0.90$ | $.348$ | $.023$
Participant Gender $\times$ Voice Gender | $F(2,76)=1.26$ | $.291$ | $.032$
Participant Gender $\times$ Mitigation Strategy | $F(1,38)=3.56$ | $.067$ | $.086$
Voice Gender $\times$ Mitigation Strategy | $F(2,76)=0.48$ | $.620$ | $.013$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,76)=0.54$ | $.583$ | $.014$
## 4\. Results
Our data analysis included a total of 246 trials from 41 participants (six
trials for six conditions per participant). In 9 of the 246 trials, the
intended errors were not correctly identified. To handle these missing values
for the questionnaire measures, first, we analyzed whether these values are
missing completely at random (MCAR) or not. Little’s MCAR test (Little, 1988)
was not significant suggesting that it is safe to assume that data is MCAR,
$\chi^{2}(197,N=43)=158.28$, $p=.98$. Multiple Imputation (MI) (Rubin, 2004)
is one of the optimal techniques to handle the missing data and gives unbiased
results under MCAR (van Ginkel et al., 2020). Thus, before proceeding with our
analysis, we replaced the missing data using MI in SPSS by computing five
imputations and pooling the results, taking into account variation across
these imputations.
For the analyses reported in this paper, we included participant gender as a
between-subjects factor to explore the nuanced relationships among voice
gender, mitigation strategy, and participant gender. Due to the significant
imbalance in participant gender, with only one participant who self reported
having a non-binary gender, we excluded this participant in the following
analyses. Thus, a total of 40 participants (21 females (women), 19 males
(men)) aged between 18 to 58 ($M=26.23,SD=8.08$) were included for analysis.
For the results reported below, we used mixed-model repeated measures analysis
of variance (ANOVA) unless stated otherwise. The gender of AI voice, error
mitigation strategy, and participant’s gender were set as the fixed effect and
participants as a random effect. Voice gender and error mitigation strategy
are within-subjects factor while participant gender is between-subjects
factor. For all the statistical tests reported below, $p<.05$ is considered as
a significant effect. We only describe significant results below. Complete
results are presented in Tables 1, 2, 3, and 4. We follow Cohen’s guidelines
on effect size and considered $\eta_{p}^{2}=0.01$ a small effect,
$\eta_{p}^{2}=0.06$ a medium effect, and $\eta_{p}^{2}=0.14$ a large effect
(Cohen, 1988).
Figure 3. Results of perceived gender and humalikeness of the assistant.
Mixed-model repeated measures ANOVAs were conducted to discover effects of
voice gender—ambiguous (Amb.), feminine (Fem.), masculine (Masc.)—, mitigation
strategy—apology (Apol.), compensation (Comp.)—, and participant (PCP)
gender—female, male— on participants’ perceptions of agent characteristics.
All pairwise comparisons were conducted using Fisher’s LSD method with
Bonferroni correction. Error bars represent standard error (SE) and only the
significant comparisons ($p<.05$) are highlighted.
[Results of perceived gender and humalikeness of the assistant.]The figure
shows results of perceived gender and humalikeness of the assistant. The
figure further consists of 2 sub graphs. All are bar plots and y-axis is
ranged from 1 to 5. The first of the 2 sub graphs from left to right further
consists of 3 bar plots and shows results on perceived gender which is the
y-axis for all three bar plots. One means most feminine and 5 means most
masculine while 3 means ambiguous. First of the 3 bar plots has voice gender
on x-axis and shows significant main effect of voice gender on perceived
gender of assistants. The p-value of all the significant comparisons discussed
is less than .001 unless stated otherwise. Feminine assistant with a mean of
1.1 is rated more feminine than masculine assistant with a mean of 4.9. While,
ambiguous with mean of about 3.9 is also rated more masculine than feminine
and more feminine than masculine assistant. Second of the 3 bar plots has
mitigation strategy on x-axis and shows no significant main effect of
mitigation strategy on assistant’s perceived gender. Assistants employing
compensation and apology strategy have mean of about 3.5. Third of the 3 bar
plots has participant gender on x-axis and shows significant effect of
participant gender on assistants’ perceived gender with a p-value of.019. Male
participants with a mean of 3.2 perceived the assistants’ gender to be lower
on the scale which is more feminine than female participants with a mean of
3.4. The figure shows results of perceived gender and humalikeness of the
assistant. The figure further consists of 2 sub graphs. All are bar plots and
y-axis is ranged from 1 to 5. The second of the 2 sub graphs from left to
right further consists of 3 bar plots and shows results on perceived
humanlikeness which is the y-axis for all three bar plots. One means least
humanlike and 5 means most humanlike. First of the 3 bar plots has voice
gender on x-axis and shows significant main effect of voice gender on
perceived gender of assistants. Feminine assistant with a mean of 3.4 is rated
more humanlike than masculine assistant with a mean of 3.0 and a p-value of
.003. While, ambiguous with mean of about 2.6 is rated less humanlike as
compared to masculine with a p-value of .007 and feminine assistants with a
p-value of less than .001. Second of the 3 bar plots has mitigation strategy
on x-axis and shows no significant main effect of mitigation strategy on
assistant’s perceived humanlikeness. Assistants employing compensation and
apology strategy have mean of about 2.9 and 3.0 respectively. Third of the 3
bar plots has participant gender on x-axis and shows no significant main
effect of participant gender on assistants’ perceived humalikeness. Male
participants have a mean of 3 and female participants have a mean of 3.2.
### 4.1. Perceptions of Agent Characteristics
Fig. 3 and Table 1 summarize our results for perceptions of agent
characteristics.
#### 4.1.1. Perceived Voice Gender
We saw a significant main effect of the voice gender on participants’
perception of agent’s gender confirming that the manipulation of gendered
voices for assistants in the study was successful (Fig. 3 a and Table 1). The
agent with feminine voice was rated more feminine than the agents with
masculine and ambiguous voice. Masculine voice was rated more masculine than
ambiguous voice. Moreover, we observed a significant main effect of
participant gender on their perceptions of agent’s gender. Male participants
perceived assistants’ gender to be lower on the scale (more feminine) than
female participants.
#### 4.1.2. Humanlikeness
We found a significant main effect of voice gender on participants’ perception
of humanlikeness of the agent (Fig. 3 b and Table 1). The agent with feminine
voice was perceived more humanlike than the agent with masculine voice.
Ambiguous voice was perceived more robotic than both feminine and masculine
voice.
Table 2. Results of subjective measures: service recovery satisfaction, warmth, and competence. Mixed-model repeated measures ANOVA for voice gender, mitigation strategy, and participant gender. ${}^{\ast}p\leq 0.05$, ${}^{\ast\ast}p\leq 0.01$, and ${}^{\ast\ast\ast}p\leq 0.001$ Service Recovery Satisfaction | | |
---|---|---|---
Variables | $F$ | $p$ | $\eta_{p}^{2}$
Voice Gender | $F(2,76)=0.99$ | $.377$ | $.025$
Mitigation Strategy | $F(1,38)=4.81$ | $.034^{\ast}$ | $.112$
Apology ($M=4.22,SD=0.84$) ¡ Compensation ($M=4.47,SD=0.76$) | | |
Participant Gender | $F(1,38)=1.93$ | $.173$ | $.048$
Participant Gender $\times$ Voice Gender | $F(2,76)=.310$ | $.720$ | $.039$
Participant Gender $\times$ Mitigation Strategy | $F(1,38)=3.77$ | $.060$ | $.090$
Voice Gender $\times$ Mitigation Strategy | $F(2,76)=0.19$ | $.831$ | $.005$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,76)=0.29$ | $.712$ | $.008$
Warmth | | |
Variables | $F$ | $p$ | $\eta_{p}^{2}$
Voice Gender | $F(2,76)=3.93$ | $.024^{\ast}$ | $.094$
Ambiguous ($M=3.79,SD=0.82$) ¡ Feminine ($M=4.03,SD=0.70$) | | $.027^{\ast}$ |
Mitigation Strategy | $F(1,38)=6.39$ | $.016^{\ast}$ | $.144$
Compensation ($M=3.80,SD=0.77$) ¡ Apology ($M=3.98,SD=0.74$) | | |
Participant Gender | $F(1,38)=0.96$ | $.334$ | $.025$
Participant Gender $\times$ Voice Gender | $F(2,76)=1.52$ | $.225$ | $.039$
Participant Gender $\times$ Mitigation Strategy | $F(1,38)=0.05$ | $.826$ | $.001$
Voice Gender $\times$ Mitigation Strategy | $F(2,76)=0.42$ | $.656$ | $.011$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,76)=1.953$ | $.149$ | $.023$
Competence | | |
Variables | $F$ | $p$ | $\eta_{p}^{2}$
Voice Gender | $F(2,76)=3.56$ | $.033^{\ast}$ | $.086$
Masculine ($M=3.83,SD=0.85$) ¡ Feminine (($M=4.00,SD=0.84$)) | | $.021^{\ast}$ |
Mitigation Strategy | $F(1,38)=6.39$ | $.016^{\ast}$ | $.144$
Compensation ($M=3.80,SD=0.77$) ¡ Apology ($M=3.98,SD=0.74$) | | |
Participant Gender | $F(1,38)=1.93$ | $.173$ | $.048$
Participant Gender $\times$ Voice Gender | $F(2,76)=1.20$ | $.308$ | $.030$
Participant Gender $\times$ Mitigation Strategy | $F(1,38)=0.00$ | $.997$ | $.000$
Voice Gender $\times$ Mitigation Strategy | $F(2,76)=0.18$ | $.794$ | $.005$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,76)=0.29$ | $.712$ | $.008$
### 4.2. Subjective Measures
Fig. 4 and Table 2 summarize our results for subjective measures.
#### 4.2.1. Service Recovery Satisfaction
We observed a significant main effect of mitigation strategy of participants’
satisfaction of service recovery; in particular, compensation resulted in
better service recovery satisfaction than apology (Fig. 4 a and Table 2).
Figure 4. Results of participants’ service recovery satisfaction, and
perceived warmth and competence of the assistant. Mixed-model repeated
measures ANOVAs were conducted to discover effects of voice gender—ambiguous
(Amb.), feminine (Fem.), masculine (Masc.)—-, mitigation strategy—apology
(Apol.), compensation (Comp.)—, and participant (PCP) gender—female, male— on
subjective measures. All pairwise comparisons were conducted using Fisher’s
LSD method with Bonferroni correction. Error bars represent standard error
(SE) and only the significant comparisons ($p<.05$) are highlighted.
[Results of participants’ service recovery satisfaction, and perceived warmth
and competence of the assistant.]The figure shows results of participants’
service recovery satisfaction, and perceived warmth and competence of the
assistant. The figure further consists of 3 sub graphs. All are bar plots and
y-axis is ranged from 1 to 5. The first of the 3 sub graphs from left to right
further consists of 3 bar plots and shows results on service recovery
satisfaction which is the y-axis for all three bar plots. First of the 3 bar
plots has voice gender on x-axis and shows no significant main effect of voice
gender on service recovery satisfaction. the mean for ambiguous is 4.2, and
for feminine and masculine sounding assistants is closer to 4.3. Second of the
3 bar plots has mitigation strategy on x-axis and shows a significant main
effect of mitigation strategy on service recover satisfaction with a p-value
of .034. Assistant employing compensation strategy with a mean of about 4.5 is
rated better than the assistant that apologizes with a mean of 4.2. Third of
the 3 bar plots has participant gender on x-axis and shows no significant main
effect of participant gender on service recovery satisfaction. Male
participants have a mean of 4.2 and female participants have a mean of 4.3.
The second of the 3 sub graphs from left to right further consists of 3 bar
plots and shows results on warmth which is the y-axis for all three bar plots.
First of the 3 bar plots has voice gender on x-axis and shows a significant
difference between ambiguous sounding and Feminine sounding assistants with a
p-value of 0.027. Feminine sounding assistant with a mean of about 4.0 is
perceived to be warmer than ambiguous sounding one with a mean of 3.8.
Masculine assistant has a mean warmth of 3.9. Second of the 3 bar plots has
mitigation strategy on x-axis and shows a significant main effect of
mitigation strategy on warmth with a p-value of .016. Assistant employing
apology strategy with a mean of about 4.0 is rated better than the assistant
that compensates with a mean of 3.8. Third of the 3 bar plots has participant
gender on x-axis and shows no significant main effect of participant gender on
assistants’ perceived warmth. Male participants have a mean of about 3.8 and
female participants have a mean of about 4.0. The third of the 3 sub graphs
from left to right further consists of 3 bar plots and shows results on
competence which is the y-axis for all three bar plots. First of the 3 bar
plots has voice gender on x-axis and shows a significant difference between
feminine sounding and masculine sounding assistants with a p-value of 0.021.
Feminine sounding assistant with a mean of about 4.0 is perceived to be more
competent than masculine sounding one with a mean of 3.8. Ambiguous assistant
has a mean competence of 3.9. Second of the 3 bar plots has mitigation
strategy on x-axis and shows no significant main effect of mitigation strategy
on warmth. Assistant employing compensation strategy has a mean of about 4.0
and the assistant that apologizes with a mean of about 3.8. Third of the 3 bar
plots has participant gender on x-axis and shows no significant main effect of
participant gender on assistants’ perceived competence. Male participants have
a mean of about 3.7 and female participants have a mean of about 4.2.
#### 4.2.2. Warmth
We observed a a significant main effect of voice gender on participants’
perception of warmth of agent. The agent with a feminine voice was perceived
warmer than the agent with an ambiguous voice (Fig. 4 b and Table 2).
#### 4.2.3. Competence
We observed a significant main effect of voice gender on participants’
perception of competence of voice assistant (Fig. 4 c and Table 2). The
assistant with a feminine voice was perceived more competent than the
assistant with a masculine voice. Moreover, there was a significant main
effect of the mitigation strategy on perceived competence; assistants offered
compensation were perceived to be warmer than assistants that apologized.
Table 3. Results of behavioral analysis: Related to error and error mitigation strategy. ${}^{\ast}p\leq 0.05$, ${}^{\ast\ast}p\leq 0.01$, and ${}^{\ast\ast\ast}p\leq 0.001$ Participant interruption during error | |
---|---|---
Variables | $\chi^{2}$ | $p$
Voice Gender | $\chi^{2}(2)=1.66$ | $.435$
Participant Gender | $\chi^{2}(1)=4.93$ | $.026^{\ast}$
Female($9.6\%$) ¡ Male($19.8\%$) $95\%$ CI: $[1.10,4.89]$ | |
Non-verbal reaction to error | |
Variables | $\chi^{2}$ | $p$
Voice Gender | $\chi^{2}(2)=3.08$ | $.857$
Participant Gender | $\chi^{2}(1)=0.03$ | $.857$
Verbal response to error mitigation strategy | |
Variables | $\chi^{2}$ | $p$
Voice Gender | $\chi^{2}(2)=0.44$ | $.802$
Mitigation Strategy | $\chi^{2}(1)=0.09$ | $.766$
Participant Gender | $\chi^{2}(1)=0.06$ | $.812$
Non-verbal response to error mitigation strategy | |
Variables | $\chi^{2}$ | $p$
Voice Gender | $\chi^{2}(2)=1.50$ | $.471$
Mitigation Strategy | $\chi^{2}(1)=0.01$ | $.937$
Participant Gender | $\chi^{2}(1)=3.32$ | $.069$
Figure 5. Results of interruption (Intrp.) related behavioral metrics.
Generalized Estimating Equations (GEE) was used to fit a logistic regression
to discover effects of voice gender, and participant gender on participants’
interruption to error. Moreover, a mixed model ANOVA was conducted to study
effects of voice gender—ambiguous, feminine, masculine—-, mitigation
strategy—apology (Apol.), compensation (Comp.)—, and participant (PCP)
gender—female, male— on number of interruptions. All pairwise comparisons were
conducted using Tukey’s HSD method. Error bars represent standard error (SE)
and only the significant comparisons ($p<.05$) are highlighted.
[Results of interruption related behavioral metrics]The figures has results on
interruption related user behavior metrics. There are 3 sub graphs. First sub
graph from left to right further consists of two mosaic plots. Generalized
Estimating Equations (GEE) to fit a repeated measures logistic regression was
run for this analysis of interrupted error. First of the two mosaic plots has
participant gender on x-axis and interrupted error (dichotomous variable)
which shows a significant difference with p-value of .026 between female and
male participants. Female participants interrupted 12 times while did not
interrupt at all 113 times where as male participants interrupted 21 times and
did not interrupt 84 times. The second mosaic plot has portrayed AI gender on
x-axis while y-axis is the same which is interrupted error. It shows no
significant effect of AI gender on whether participants interrupted during
mitigation statement delivery. For ambiguous assistant there was an
interruption in 11 of total 76 trials, for feminine assistant in 14 of 80
trials and for masculine assistant in 8 of 75 trials. The second and third
graphs from left to right show results from a mixed model ANOVA to study
effects of our factors on number of interruptions. The y-axis for both plots
is average count of interruptions with a range of 0 to 0.8. The second from
left to right has two bar plots. First of the two bar plots is for female
participants only and the x-axis is mitigation strategies showing no
significant interaction effect between the two and mean for both apology and
compensation is around 0.1. For the second of 2 bar graphs, mitigation
strategies only for male participants are on x-axis. It shows that apologetic
assistants had lesser interruptions with mean about 0.2 than compensating
assistants with mean about 0.4 when looking at male participants only. The
p-value for this interaction effect is .047. The third and the last graph from
left to right has mitigation strategies on x-axis only for ambiguous sounding
assistants. Apologetic assistants with mean of about 0.1 were interrupted less
than compensating ones with mean of about 0.3 when looking at ambiguous
sounding assistants only. The p-value for this interaction effect is .018.
Table 4. Results of behavioral analysis: Relevant to overall interaction. ${}^{\ast}p\leq 0.05$, ${}^{\ast\ast}p\leq 0.01$, and ${}^{\ast\ast\ast}p\leq 0.001$ Number of interruptions | |
---|---|---
Variables | $F$ | $p$
Voice Gender | $F(2,75.89)=0.16$ | $.849$
Mitigation Strategy | $F(1,38.06)=3.45$ | $.071$
Compensation ($M=3.80,SD=0.77$) ¡ Apology ($M=3.98,SD=0.74$) |
Participant Gender | $F(1,38.30)=0.49$ | $.489$
Participant Gender $\times$ Voice Gender | $F(2,75.89)=0.78$ | $.464$
Participant Gender $\times$ Mitigation Strategy | $F(1,38.06)=4.67$ | $.037^{\ast}$
Male, Apology ($M=0.17,SD=0.64$) ¡ | |
Male, compensation ($M=0.34,SD=0.90$) | | $.047^{\ast}$
Voice Gender $\times$ Mitigation Strategy | $F(2,75.29)=3.80$ | $.027^{\ast}$
Ambiguous, Apology ($M=0.08,SD=0.36$) | |
¡ Ambiguous, Compensation ($M=0.33,SD=0.77$) | | $.018^{\ast}$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,75.29)=0.58$ | $.562$
Number of other verbal reactions | |
Variables | $F$ | $p$
Voice Gender | $F(2,63.99=0.04$ | $.960$
Mitigation Strategy | $F(1,38.35)=2.59$ | $.116$
Participant Gender | $F(1,37.34)=0.73$ | $.397$
Participant Gender $\times$ Voice Gender | $F(2,63.99)=0.95$ | $.391$
Participant Gender $\times$ Mitigation Strategy | $F(1,38.35)=0.27$ | $.606$
Voice Gender $\times$ Mitigation Strategy | $F(2,66.31)=2.10$ | $.130$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,66.31)=0.34$ | $.710$
Number of other non-verbal reactions | |
Variables | $F$ | $p$
Voice Gender | $F(2,73.65)=2.80$ | $.067$
Mitigation Strategy | $F(1,35.1)=7.14$ | $.011^{\ast}$
Compensation ($M=0.40,SD=0.77$) ¡ Apology ($M=0.74,SD=1.27$) | |
Participant Gender | $F(1,38.30)=1.88$ | $.178$
Participant Gender $\times$ Voice Gender | $F(2,73.65)=4.54$ | $.014^{\ast}$
Male, Masculine ($M=0.53,SD=0.90$) ¡ Male, Feminine ($M=1.08,SD=1.62$) | | $.029^{\ast}$
Participant Gender $\times$ Mitigation Strategy | $F(1,35.10)=.69$ | $.411$
Voice Gender $\times$ Mitigation Strategy | $F(2,76.79)=0.19$ | $.828$
Participant Gender $\times$ Voice Gender & Mitigation Strategy | $F(2,76.79)=4.83$ | $.011^{\ast}$
Male, Masculine, Apology ($M=0.61,SD=0.98$) | |
¡ Male, Feminine, Apology ($M=1.47,SD=1.98$) | | $.030^{\ast}$
### 4.3. Behavioral Analysis
We did not include the nine trials in which participants did not recognize the
error (one trial from a female participant and eight trials from male
participants). As a result, the following analyses included 231 trials. Tables
3 and 4 summarize the results of behavioral analysis.
#### 4.3.1. Participant interruption during error (Dichotomous variable:
0/Yes, 1/No)
For this analysis, we did not consider error mitigation as a factor because
the participants did not experience the mitigation strategy until after the
error was fixed. In other words, whether the participants interrupted the
agent during the error was not affected by the mitigation strategy. We set up
Generalized Estimating Equations (GEE) to fit a logistic regression to examine
whether gender of voice and participant gender had an effect on whether the
participant interrupted the error or not. The model estimated that the odds
for male participants ($19.8\%$) interrupting the assistant after realizing
the error has occurred while assistant was still talking are approximately
2.32 times higher than female participants ($9.6\%$) interrupting the
assistant (Fig. 5 a and Table 3).
Figure 6. Results of number of other non-verbal reactions (RXN). Mixed-model
repeated measures ANOVA was conducted to discover effects of voice
gender—ambiguous (Amb.), feminine (Fem.), masculine (Masc.)—, mitigation
strategy—apology (Apol.), compensation (Comp.)—, and participant (PCP)
gender—female, male— on number of other non-verbal reactions. All pairwise
comparisons were conducted using Tukey’s HSD method. Error bars represent
standard error (SE) and only the significant comparisons ($p<.05$) are
highlighted.
[Results of number of other non-verbal reactions]There are four bar charts for
mixed model repeated measures ANOVAs to show effects of our factors on number
of other non-verbal reactions. The range of y-axis is 0 to 1.6 for all graphs.
First bar chart from left to right further consists of 3 bar charts. The first
among the three bar charts shows voice gender on the x-axis to present that
there is no significant main effect of gender of AI voice on number of other
non-verbal reactions. The mean for ambiguous sounding assistant is around 0.4,
female sounding assistant is around 0.6 and male one is around 0.5. The second
of the 3 bar charts shows mitigation strategy on x-axis to present that there
is significant main effect of gender on the number of non-verbal reactions,
p-value is .025 where mean for apology is about 0.6 and is higher than mean
for compensation which is about 0.4. the third of the 3 bar charts shows
participant gender on the x-axis and shows no significant effect of
participant gender on count of other non-verbal reactions. The mean for female
participants is around 0.4 while mean for male participants is around 0.6. The
second bar chart from left to right further consists of one bar chart for
showing interaction of voice gender and participant gender on the number of
other non-verbal reactions. It only shows plot for error mitigation strategies
for male participants. Thus, the x-axis is ambiguous with mean of about 0.6,
female with mean of about 0.9 and male assistants with mean of about 0.4 for
male participants only. There is only a significant difference between female
and male assistants with female having higher mean than male assistants while
looking at male participants only. The p-value for this interaction is .032.
The third bar chart from left to right further consists of 3 bar charts. It
only shows all the graphs for male participants only. The first of the three
bar charts shows mitigation strategies for male voice assistants only, on
x-axis. There is no significant difference between apology with a mean of
about 0.8 and compensation strategy with a mean of about 0.6 when it comes to
number of non-verbal reactions. The second of the three bar charts shows
mitigation strategies for only female voice assistants on x-axis while the
third of the three bar charts only shows mitigation strategies for male voice
assistants. There is only one significant interaction effect between these two
bar charts which is that male participants react non-verbally more to
apologetic female assistants with mean of about 1.2 than apologetic male
assistants with mean of about 0.4. The p-value for this interaction is .042.
Mean for female voice assistants that compensate is about 0.6 and for male
voice assistant that compensated is about 0.4.
#### 4.3.2. Non-verbal reaction to error (Dichotomous variable: 0/Yes, 1/No)
For this measure, we also did not considering error mitigation as factor as
well because the participants did not interact with this manipulation till
after error was fixed. Hence, whether the participants reacted non-verbally to
the agent during the error was not affected by the mitigation strategy. We set
up GEE to fit a logistic regression to examine whether voice gender and
participant gender influenced non-verbal reaction to error. We saw no
significant results.
#### 4.3.3. Verbal response to error mitigation strategy (Dichotomous
variable: 0/Yes, 1/No)
We set up GEE to fit a logistic regression to examine whether voice gender,
mitigation strategy, and participant gender influenced verbal response to
error mitigation strategy. We did not see any significant effects.
#### 4.3.4. Non-verbal response to error mitigation strategy (Dichotomous
variable: 0/Yes, 1/No)
We set up GEE to fit a logistic regression to examine whether voice gender,
mitigation strategy, and participant gender influenced non-verbal response to
error mitigation strategy. No significant effects were observed.
#### 4.3.5. Number of interruptions
A mixed repeated measures ANOVA with voice gender and mitigation as within-
subjects factors, and participant gender as a between-subjects factor revealed
a significant interaction effect of participant gender and mitigation on count
of interruptions. Post-hoc comparisons using Tukey’s HSD revealed that there
were more interruptions between male participants and agents compensating than
the agents apologizing, regardless of assistant’s gender (see Fig. 5 b and
Table 4). Moreover, a significant interaction effect of voice gender and
mitigation on count of interruptions was also observed. For ambiguous sounding
agents, there were more interruptions when they compensated rather than
apologized for the errors (see Fig. 5 c and Table 4).
#### 4.3.6. Number of other verbal reactions
Using mixed model repeated measures ANOVA with voice gender and mitigation
strategy as within-subjects factors, and participant gender (male, female) as
a between-subjects factor, we observed no significant effects.
#### 4.3.7. Number of other non-verbal reactions
Fig. 6 and Table 4 summarize our results for number of other non-verbal
reactions. Using mixed model repeated measures ANOVA with voice gender and
mitigation strategy as within-subjects factors, and participant gender (male,
female) as a between-subjects factor, we observed a significant main effect of
mitigation strategy on the number of non-verbal responses. Participants
reacted non-verbally more to agents in the apology condition than in the
compensation condition. There was a significant interaction effect between
participant gender and the voice gender of AI. Tukey’s HSD shows that male
participants reacted non-verbally more to agents with feminine voice than to
agents with masculine voices. Similarly, a significant interaction effect was
found between participant gender, mitigation strategy, and the voice gender of
AI. Tukey’s HSD shows that male participants reacted non-verbally more to
apology coming from the agent with feminine voice than from the agent with
masculine voice.
## 5\. Discussion
Mitigating AI errors during recovery from service failures is essential to
restoring and maintaining human-AI collaboration. AI agents in tasks of
assistance nature are mostly defaulted to manifesting feminine voices and
characteristics, which may reinforce biases because of association of gender
stereotypes. Users’ gender may also influence perceptions of gendered AI
agents. Gender ambiguous voices have been explored as a way to mitigate these
biases. This work explored how error mitigation strategy (apology vs.
compensation), gender of AI assistant voice (feminine vs. ambiguous vs.
masculine), and participant’s gender (female vs. male) may affect
participants’ satisfaction with the service recovery, perception of, and their
behavior around voice assistants during an assistive task exemplified as an
online shopping task. Below, we discuss our results, their implications,
limitations of this study, and future work.
### 5.1. Apology Communicates Warmth and Compensation Leads to User
Satisfaction
Our hypothesis 1 predicts that apology during recovery from errors by voice
assistants portrays the assistant as warmer while offered compensation creates
greater service recovery satisfaction and perceptions of competence. Our
results support this hypothesis (Fig. 4 and Table 2). Prior work has
demonstrated that humans exhibit social reactions in response to
conversational agents when these agents exhibit social personality such as
engaging in small talk, displaying gender, and age characteristics, using
gestures, or expressing facial expressions (Feine et al., 2019). In our study,
we see that voice assistants that were apologetic —exhibiting a social
personality— elicited more non-verbal reactions than the ones that offered
compensation —exhibiting a utilitarian personality— (Fig. 6 a and Table 4).
For example, one participant laughed 13 times when interacting with assistants
that apologized while only 5 times for assistants that compensated. These non-
verbal reactions such as smile, laugh, and nod are reflective of social
reactions. Since, apologizing or offering compensation can be seen as
indications of distinct personalities or traits communicated through speech,
the level of non-verbal user reactions can serve as an indicator of users’
perceptions of voice assistant’s personality.
However, further looking into the reactions, their valence may vary for
different participants. For instance, one participant reacted a total of six
times to apologetic agents in which two of the reactions were “nodding”, two
were “shaking head”, and two were “hands in the air” gestures indicating
confusion. While both of his reactions to compensating agents were “hands in
the air” gestures indicating confusion. The apologetic agents were able to
elicit a greater variety of expressions with both positive and negative
connotations. Though may be participant dependent, social reactions can be
informative of users’ perceptions of agent’s personality traits such as warmth
and competence. This observation has design implications for using social
reactions to understand users’ perceptions of assistants and is similar to the
prior findings illustrating that social reactions from people are indicative
of their surprise and experiences of unexpected events (e.g., errors) during
their interactions with robots (Stiber et al., 2022, 2023). However, further
research is needed to study how the valence of non-verbal reactions may be
used to model and estimate users’ perceptions of AI agents, which could then
be used to adapt the agents’ behaviors towards achieving desirable interaction
outcomes.
### 5.2. Portrayed Voice Gender Influences User Perceptions of the Voice
Assistant
Our hypothesis 2 predicts that masculine voice assistants are viewed as more
competent while feminine voice assistants are considered warmer. Our results
do not support this hypothesis since we observe that the feminine assistants
are perceived as more competent than masculine assistants (Fig. 4 and Table
2). We speculate that this preference of feminine voice assistants in the
context of online shopping assistance may be influenced by the broader
societal stereotypes associating women with competence in serving supportive
roles in assistance-type tasks (Powell et al., 2002; Koenig et al., 2011). It
is noteworthy that commonly available voice assistants such as Siri, Alexa,
and Cortana are marketed as “personal assistants” and typically employ
feminine voices (UNESCO, 2019); such de facto voice setting has shown to
attract abusive behavior (Curry and Rieser, 2019; Curry et al., 2020). Such
stereotypes have been translated to the preference of feminine voice
assistants in tasks that are “perceived” as domestic chores such as shopping
for groceries, schools supplies, and more (Lee et al., 2019). While digital
personal assistance should not be inherently associated with a specific
gender, when a gendered voice (e.g., feminine voice) is assigned to a voice
assistant, user perceptions of the online shopping task may align with the
gender stereotypes and biases seen in prior work (Lee et al., 2019; Ahn et
al., 2022; Damen and Toh, 2019; Tay et al., 2014). Hence, the association of
competence with feminine agents for assistance-type tasks found in our study
possibly hint at implicit preferences and biases driven by gender stereotypes.
The text-to-speech systems are evolving to generate increasingly realistic and
humanlike voices for AI agents (Seaborn et al., 2021). In this context, our
study revealed that the feminine voice is perceived to be more humanlike
($M=3.36$, 1 being robotic and 5 being humanlike) as compared to masculine
($M=2.96$) and ambiguous one ($M=2.59$). We speculate that the higher
perception of humanlikeness for feminine voices may again be attributed to
users’ familiarity with feminine assistants in real life, such as Alexa, Siri,
and Bixby. However, as voice assistants become more humanlike due to addition
of social cues to their behavior (Bucher et al., 2022), biases encountered by
these assistants could potentially reinforce gender preferences and biases in
actual human interactions. Therefore, designers and developers must carefully
consider the portrayed gender of voices while designing AI agents for specific
tasks to support people effectively while minimizing potential negative
consequences in both human-agent and human-human interactions.
### 5.3. Men Show More Biased Behavior
Our results highlight the impact of user gender on their behavior towards
gendered voice assistants while repairing human-agent relationship using
different mitigation strategies. Male participants generally reacted non-
verbally (socially) more to feminine assistants than masculine ones (Fig. 6 b
and Table 4). For example, a male participant smiled at feminine assistants
more than at masculine assistants. Similarly, another male participant
presented various types of reactions (two nodes, one smile, and one frown) to
feminine assistants whereas he only frowned once at the masculine assistants.
Moreover, male participants registered more non-verbal reactions to
“apologetic” agents when they sounded feminine as opposed to masculine (Fig. 6
c and Table 4). For instance, a male participant laughed six times when
interacting with feminine sounding apologetic assistant while only one time to
the masculine counterpart. Similarly, another male participant nodded twice,
smiled once, and raised an eyebrow (frowned) once in reaction to the feminine
sounding apologetic assistant, whereas no reaction was given to the masculine
sounding apologetic assistant. As previously discussed in Section 5.1, we see
that the valence and variety of reactions may vary depending on the individual
participant. Future research could explore how such diverse reactions
contribute to the reflection of biased behavior in assistance scenarios.
Prior work has also shown similar discrepancy in behavior between men and
women. Men found robots with feminine voices more trustworthy and engaging,
leading them to donate more money. On the other hand, women show no preference
between robots with feminine or masculine voices (Siegel et al., 2009).
Consistent with these prior findings, our results imply that male participants
engage more with feminine assistants, especially during apologetic
interactions which highlights the importance of considering user gender and
its potential impact on interactions with gendered voice assistants,
especially considering that these assistants will inevitably make mistakes in
their interactions with people in complex, real-world settings.
We further see biased behavior towards voice assistants by men—regardless of
the gender of voice assistants and error mitigation strategy, male
participants were found to interrupt the assistants on realization of error
before the assistant stopped talking almost twice as many times as female
participants would (Fig. 5 a and Table 3). These interruptions can be seen as
impolite behavior (Hutchby, 2008), possibly stemming from implicit power
dynamics where men are typically in authoritative roles (Powell et al., 2002).
Previous research in human-agent interactions, has also indicated that men
show less inclination to assist the agent, such as in the case of an AI study
companion (Lubold et al., 2016). Such biased behavior towards voice assistants
by men highlights the need for further examination of implicit power dynamics
and their impact on human-agent interactions.
We also observe that that male participants ($M=3.17$), as opposed to their
female counterparts ($M=3.45)$, overall, perceived the voice assistants to be
more feminine regardless of the portrayed gender of the voice and mitigation
strategy (Fig. 3 a and Table 1). Although this perception may arise from a
greater familiarity with “feminine commercial voice assistants” as discussed
in section 5.2, it can still have a negative impact on interactions with AI
assistants. Hence, it is of utmost importance to further understand impacts of
association of any gender to voice assistants, and interaction of agents’ and
users’ gender before continuing to create gendered voice assistants aimed to
assist and support people.
Despite male participants interrupting the voice assistants more during error,
we see that the interruptions between the male participants and the voice
assistants were fewer in number when the assistants were apologizing for
errors instead of compensating the participants (Fig. 5 b and Table 4). A
possible explanation for this finding is that it may be considered rude to
interrupt while someone is apologizing. Thus, voice assistants can potentially
utilize apologies to deter people from interrupting, thereby fostering more
polite, smooth, and respectful interactions. This finding implies that we can
tailor agent behavior such as using apology as a mitigation strategy to elicit
positive attitude from people in face of imperfect interactions.
### 5.4. Gender Ambiguous Voices as a Plausible Means to Reduce Biases
Gender ambiguous voices (e.g., “Q, the
genderless”666https://www.genderlessvoice.com/) is presented as a tool to
culminate gender stereotypes and biases in prior work on voice assistants
(Tolmeijer et al., 2021). Therefore, in addition to our hypotheses 2 and 3, we
explore if an ambiguous voice can eliminate or reduce the effects of
stereotypes associated with gendered voices for AI agents. Our results show
that ambiguous sounding assistants are perceived less warm compared to
feminine ones (Fig. 4 b and Table 2), suggesting that ambiguous voices can
potentially be employed to isolate gender stereotypes around warm personality
(feminine agents or robots are considered warmer than masculine ones (Ahn et
al., 2022; Nomura, 2017)) for voice assistants. This finding contradicts a
previous study where feminine and ambiguous voices are not disassociated with
stereotypical feminine traits such as family-orientation, delicacy, and
sensitivity (Tolmeijer et al., 2021). The distinct perception of warmth of our
gender-ambiguous voice, as compared to feminine one, indicates the potential
for using ambiguity of gender in voice to create less biased voice
interactions with AI agents.
Looking into behavioral measures, the interruptions between users and
ambiguous voice assistants were more frequent when the assistants compensated
instead of apologizing for the error (Fig. 5 c and Table 4). Fewer
interruptions generally reflect increased politeness (Hutchby, 2008).
Therefore, by employing a relational strategy like apologizing, ambiguous
assistants were able to elicit politeness and courtesy from users without
significant interruptions, despite being perceived as more robotic than
gendered assistants. Therefore, a design implication of our work is that
ambiguous voices for AI agents have the potential to maintain certain
benefits, such as promoting politeness and attentive listening, while reducing
the association with stereotypical feminine traits (such as warmth and
humanlikness) when the agent’s role is assistive in nature.
Our study additionally shows that ambiguous sounding assistants are perceived
to be the most robotic when compared to feminine and masculine ones. This
result, in combination with prior work showing that majority of people
preferred a robotic voice ($32.4\%$), followed by feminine ($20.6\%$) and
masculine ($11.8\%$) voice for conversational assistants (Curry et al., 2020),
highlights potential of gender ambiguous voices being welcomed by users.
However, further research is needed to establish the effectiveness of
ambiguous voices in a variety of assistance-type tasks; this work only serves
as an initial exploration in this direction.
### 5.5. Limitations and Future work
This work has a few limitations that can inform future work. First, online
shopping with voice assistants is novel to most people because of limited
capabilities and hindrance in adoption of the smart speakers for functional
tasks with higher risks than putting alarms and reminders, controlling smart
appliances, information retrieval, and playing music. Second, the experiment
was conducted in a lab setting where participants might not be fully
comfortable to interact freely with the speaker. Future work should
investigate the integration of voice assistants into people’s personal
physical spaces to better understand the organic interaction between humans
and agents in commonplace and everyday tasks.
Third, limited voices were explored for masculine, feminine and ambiguous
agents in this study. Future work can focus on understanding various
characteristics of voice generation towards creation and testing of genderless
and ambiguous voices. Moreover, there are other aspects, such as personality,
names assigned, and representations in form of color and appearance, to design
of voice assistants than just voice that may influence perception of their
gender (Sutton, 2020).
Fourth, our participants do not represent the full spectrum of gender, age,
ethnicity, and education. Thus, our findings and implications should be
interpreted with this context in mind. Specifically, due to large imbalance in
gender of participants, we could not include non-binary participant gender in
our analyses. Therefore, effects of gendered and ambiguous voices on non-
binary population is unexplored. Future research should aim to recruit
participants representative of all genders to understand interaction of users’
and assistants’ gender. Moreover, age of users may affect how they interact
with voice assistants as these assistants are likely to make more errors when
interacting with older adults due to the diminished ability in their speech
such as difficulty in constructing a structured sentence for a command (Kim et
al., 2021).
Overall, future research should explore voice-based interactions with virtual
assistants to assist people in realistic settings and how various aspects of
agent design such as voice, personality, and style of speech may be
manipulated to reduce biases caused due to the association of stereotypes
based on humanlike characteristics, such as race and gender, projected onto AI
agents.
## 6\. Conclusion
The growing use of intelligent voice assistants, their susceptibility to
errors, and the potential reinforcement of gender biases in user interactions
motivated us to study how gender portrayal through gendered voice influences
error mitigation strategies employed by these voice assistants. Our results
reveal that people exhibit higher social engagement, as evidenced by non-
verbal reactions, when interacting with voice assistants that apologize rather
than compensate for errors, possibly because apologetic voice assistants are
perceived as warmer while compensating ones are viewed as more competent.
Moreover, we observe that men prefer interacting with feminine voice
assistants that apologize for mistakes during assistive tasks. Furthermore,
our study reveals a notable gender difference in interruption patterns—men
interrupted voice assistants more frequently than women when errors occurred,
indicating potential sociocultural and power dynamics at play. Apart from
illustrating biases in users’ perceptions and behaviors towards gendered voice
assistants, our study highlights the potential of using ambiguous voices to
mitigate the association of stereotypical feminine traits, such as warmth,
with AI agents designed to assist people. This work takes a step towards
designing characteristics and behavior of voice assistants to foster
interactions that are engaging, while simultaneously addressing potential
gender biases, with the ultimate goal of enhancing the overall user
experience.
## References
* (1)
* Abercrombie et al. (2021) Gavin Abercrombie, Amanda Cercas Curry, Mugdha Pandya, and Verena Rieser. 2021. Alexa, Google, Siri: What are Your Pronouns? Gender and Anthropomorphism in the Design and Perception of Conversational Assistants. _arXiv preprint arXiv:2106.02578_ (2021).
* Ahn et al. (2022) Jungyong Ahn, Jungwon Kim, and Yongjun Sung. 2022. The effect of gender stereotypes on artificial intelligence recommendations. _Journal of Business Research_ 141 (2022), 50–59.
* Baker et al. (2021) Matthew Baker, Xiaohui Hu, Gennaro De Luca, and Yinong Chen. 2021. Intelligent Voice Instructor-Assistant System for Collaborative and Interactive Classes. _Journal of Artificial Intelligence and Technology_ 1, 2 (2021), 121–130.
* Bennett and Hill (2012) Aronté Marie Bennett and Ronald Paul Hill. 2012. The universality of warmth and competence: A response to brands as intentional agents. _Journal of Consumer Psychology_ 22, 2 (2012), 199–204.
* Bradley and LaFleur (2016) Gregory T Bradley and Elizabeth K LaFleur. 2016. Toward the development of hedonic-utilitarian measures of retail service. _Journal of Retailing and Consumer Services_ 32 (2016), 60–66.
* Bucher et al. (2022) Andreas Bucher, Dario Staehelin, Mateusz Dolata, and Gerhard Schwabe. 2022. Form Follows Function: Designing For Tensions Of Conversational Agents In Service Encounters. (2022).
* Cameron et al. (2021) David Cameron, Stevienna de Saille, Emily C Collins, Jonathan M Aitken, Hugo Cheung, Adriel Chua, Ee Jing Loh, and James Law. 2021. The effect of social-cognitive recovery strategies on likability, capability and trust in social robots. _Computers in Human Behavior_ 114 (2021), 106561.
* Chita-Tegmark et al. (2019) Meia Chita-Tegmark, Monika Lohani, and Matthias Scheutz. 2019. Gender effects in perceptions of robots and humans with varying emotional intelligence. In _2019 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI)_. IEEE, 230–238.
* Cohen (1988) Jacob Cohen. 1988. Statistical power analysis for the behavioral sciences. _England: Routledge_ (1988).
* Cui et al. (2018) Yubao Cui, Xuehe Zhang, Xiaobao Peng, and Jianxun Chu. 2018. How to use apology and compensation to repair competence-versus integrity-based trust violations in e-commerce. _Electronic Commerce Research and Applications_ 32 (2018), 37–48.
* Curry and Rieser (2019) Amanda Cercas Curry and Verena Rieser. 2019. A crowd-based evaluation of abuse response strategies in conversational agents. _arXiv preprint arXiv:1909.04387_ (2019).
* Curry et al. (2020) Amanda Cercas Curry, Judy Robertson, and Verena Rieser. 2020. Conversational assistants and gender stereotypes: Public perceptions and desiderata for voice personas. In _Proceedings of the Second Workshop on Gender Bias in Natural Language Processing_. 72–78.
* Damen and Toh (2019) Nicole Damen and Christine Toh. 2019. Designing for trust: Understanding the role of agent gender and location on user perceptions of trust in home automation. _Journal of Mechanical Design_ 141, 6 (2019).
* Feine et al. (2019) Jasper Feine, Ulrich Gnewuch, Stefan Morana, and Alexander Maedche. 2019. A taxonomy of social cues for conversational agents. _International Journal of Human-Computer Studies_ 132 (2019), 138–161.
* Fiske et al. (2007) Susan T Fiske, Amy JC Cuddy, and Peter Glick. 2007. Universal dimensions of social cognition: Warmth and competence. _Trends in cognitive sciences_ 11, 2 (2007), 77–83.
* Ge et al. (2019) Xiang Ge, Dan Li, Daisong Guan, Shihui Xu, Yanyan Sun, and Moli Zhou. 2019. Do smart speakers respond to their errors properly? A study on human-computer dialogue strategy. In _International Conference on Human-Computer Interaction_. Springer, 440–455.
* Goodman and Mayhorn (2023) Kylie L Goodman and Christopher B Mayhorn. 2023. It’s not what you say but how you say it: Examining the influence of perceived voice assistant gender and pitch on trust and reliance. _Applied Ergonomics_ 106 (2023), 103864.
* Govier and Verwoerd (2002) Trudy Govier and Wilhelm Verwoerd. 2002. The promise and pitfalls of apology. _Journal of social philosophy_ 33, 1 (2002), 67–82.
* Guzman (2016) Andrea L Guzman. 2016. Making AI safe for humans: A conversation with Siri. In _Socialbots and their friends_. Routledge, 85–101.
* Habler et al. (2019) Florian Habler, Valentin Schwind, and Niels Henze. 2019. Effects of Smart Virtual Assistants’ Gender and Language. In _Proceedings of Mensch und Computer 2019_. 469–473.
* Haesevoets et al. (2013) Tessa Haesevoets, Chris Reinders Folmer, David De Cremer, and Alain Van Hiel. 2013. Money isn’t all that matters: The use of financial compensation and apologies to preserve relationships in the aftermath of distributive harm. _Journal of Economic Psychology_ 35 (2013), 95–107.
* Hutchby (2008) Ian Hutchby. 2008. Participants’ orientations to interruptions, rudeness and other impolite acts in talk-in-interaction. _Journal of Politeness Research-language Behaviour Culture_ 4 (08 2008), 221–241. https://doi.org/10.1515/JPLR.2008.011
* Hwang et al. (2019) Gilhwan Hwang, Jeewon Lee, Cindy Yoonjung Oh, and Joonhwan Lee. 2019. It sounds like a woman: Exploring gender stereotypes in South Korean voice assistants. In _Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems_. 1–6.
* Kim et al. (2021) Sunyoung Kim et al. 2021\. Exploring How Older Adults Use a Smart Speaker–Based Voice Assistant in Their First Interactions: Qualitative Study. _JMIR mHealth and uHealth_ 9, 1 (2021), e20427.
* Kim et al. (2007) Yanghee Kim, Amy L Baylor, and Entong Shen. 2007. Pedagogical agents as learning companions: the impact of agent emotion and gender. _Journal of Computer Assisted Learning_ 23, 3 (2007), 220–234.
* Koenig et al. (2011) Anne M Koenig, Alice H Eagly, Abigail A Mitchell, and Tiina Ristikari. 2011. Are leader stereotypes masculine? A meta-analysis of three research paradigms. _Psychological bulletin_ 137, 4 (2011), 616.
* Koo and Li (2016) Terry K Koo and Mae Y Li. 2016. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. _Journal of chiropractic medicine_ 15, 2 (2016), 155–163.
* Krämer et al. (2016) Nicole C Krämer, Bilge Karacora, Gale Lucas, Morteza Dehghani, Gina Rüther, and Jonathan Gratch. 2016. Closing the gender gap in STEM with friendly male instructors? On the effects of rapport behavior and gender of a virtual agent in an instructional interaction. _Computers & Education_ 99 (2016), 1–13.
* Kraus et al. (2018) Matthias Kraus, Johannes Kraus, Martin Baumann, and Wolfgang Minker. 2018. Effects of gender stereotypes on trust and likability in spoken human-robot interaction. In _Proceedings of the eleventh international conference on language resources and evaluation (LREC 2018)_.
* Lee et al. (2010) Min Kyung Lee, Sara Kiesler, Jodi Forlizzi, Siddhartha Srinivasa, and Paul Rybski. 2010. Gracefully mitigating breakdowns in robotic services. In _2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI)_. IEEE, 203–210.
* Lee et al. (2019) Sanguk Lee, Rabindra Ratan, and Taiwoo Park. 2019. The voice makes the car: Enhancing autonomous vehicle perceptions and adoption intention through voice agent gender and style. _Multimodal Technologies and Interaction_ 3, 1 (2019), 20.
* Lee et al. (2021) Sun Kyong Lee, Pavitra Kavya, and Sarah C Lasser. 2021. Social interactions and relationships with an intelligent virtual agent. _International Journal of Human-Computer Studies_ 150 (2021), 102608.
* Little (1988) Roderick JA Little. 1988. A test of missing completely at random for multivariate data with missing values. _Journal of the American statistical Association_ 83, 404 (1988), 1198–1202.
* Loideain and Adams (2020) Nora Ni Loideain and Rachel Adams. 2020. From Alexa to Siri and the GDPR: the gendering of virtual personal assistants and the role of data protection impact assessments. _Computer Law & Security Review_ 36 (2020), 105366.
* Lubold et al. (2016) Nichola Lubold, Erin Walker, and Heather Pon-Barry. 2016. Effects of voice-adaptation and social dialogue on perceptions of a robotic learning companion. In _2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI)_. IEEE, 255–262.
* Mahmood et al. (2022) Amama Mahmood, Jeanie W Fung, Isabel Won, and Chien-Ming Huang. 2022. Owning Mistakes Sincerely: Strategies for Mitigating AI Errors. In _CHI Conference on Human Factors in Computing Systems_. 1–11.
* Maxham III and Netemeyer (2003) James G Maxham III and Richard G Netemeyer. 2003. Firms reap what they sow: the effects of shared values and perceived organizational justice on customers’ evaluations of complaint handling. _Journal of Marketing_ 67, 1 (2003), 46–62.
* Mitchell et al. (2011) Wade J Mitchell, Chin-Chang Ho, Himalaya Patel, and Karl F MacDorman. 2011. Does social desirability bias favor humans? Explicit–implicit evaluations of synthesized speech support a new HCI model of impression management. _Computers in Human Behavior_ 27, 1 (2011), 402–412.
* Nomura (2017) Tatsuya Nomura. 2017. Robots and gender. _Gender and the Genome_ 1, 1 (2017), 18–25.
* Obinali (2019) Chidera Obinali. 2019. The perception of gender in voice assistants. In _SAIS 2019 Proceedings of the Southern Association for Information Systems Conference_. 22–23.
* Ozogul et al. (2013) Gamze Ozogul, Amy M Johnson, Robert K Atkinson, and Martin Reisslein. 2013. Investigating the impact of pedagogical agent gender matching and learner choice on learning outcomes and perceptions. _Computers & Education_ 67 (2013), 36–50.
* Paraiso and Barthès (2006) Emerson Cabrera Paraiso and Jean-Paul A Barthès. 2006. An intelligent speech interface for personal assistants in R&D projects. _Expert Systems with Applications_ 31, 4 (2006), 673–683.
* Pearl (2016) Cathy Pearl. 2016. _Designing voice user interfaces: Principles of conversational experiences_. ” O’Reilly Media, Inc.”.
* Phan (2019) Thao Phan. 2019. Amazon Echo and the aesthetics of whiteness. _Catalyst: Feminism, Theory, Technoscience_ 5, 1 (2019), 1–38.
* Powell et al. (2002) Gary N Powell, D Anthony Butterfield, and Jane D Parent. 2002. Gender and managerial stereotypes: have the times changed? _Journal of management_ 28, 2 (2002), 177–193.
* Roschk and Kaiser (2013) Holger Roschk and Susanne Kaiser. 2013. The nature of an apology: An experimental study on how to apologize after a service failure. _Marketing Letters_ 24, 3 (2013), 293–309.
* Rubin (2004) Donald B Rubin. 2004. _Multiple imputation for nonresponse in surveys_. Vol. 81. John Wiley & Sons.
* Sáiz-Manzanares et al. (2020) María Consuelo Sáiz-Manzanares, Raúl Marticorena-Sánchez, and Javier Ochoa-Orihuel. 2020. Effectiveness of using voice assistants in learning: A study at the time of COVID-19. _International journal of environmental research and public health_ 17, 15 (2020), 5618.
* Schmid et al. (2022) Damaris Schmid, Dario Staehelin, Andreas Bucher, Mateusz Dolata, and Gerhard Schwabe. 2022. Does Social Presence Increase Perceived Competence? Evaluating Conversational Agents in Advice Giving Through a Video-Based Survey. _Proceedings of the ACM on Human-Computer Interaction_ 6, GROUP (2022), 1–22.
* Seaborn et al. (2021) Katie Seaborn, Norihisa P Miyake, Peter Pennefather, and Mihoko Otake-Matsuura. 2021. Voice in human–agent interaction: a survey. _ACM Computing Surveys (CSUR)_ 54, 4 (2021), 1–43.
* Siegel et al. (2009) Mikey Siegel, Cynthia Breazeal, and Michael I Norton. 2009. Persuasive robotics: The influence of robot gender on human behavior. In _2009 IEEE/RSJ International Conference on Intelligent Robots and Systems_. IEEE, 2563–2568.
* Stiber et al. (2022) Maia Stiber, Russell Taylor, and Chien-Ming Huang. 2022. Modeling human response to robot errors for timely error detection. In _2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_. IEEE, 676–683.
* Stiber et al. (2023) Maia Stiber, Russell H Taylor, and Chien-Ming Huang. 2023. On using social signals to enable flexible error-aware HRI. In _Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction_. 222–230.
* Strengers and Kennedy (2021) Yolande Strengers and Jenny Kennedy. 2021. _The smart wife: Why Siri, Alexa, and other smart home devices need a feminist reboot_. MIT Press.
* Sutton (2020) Selina Jeanne Sutton. 2020. Gender ambiguous, not genderless: Designing gender in voice user interfaces (VUIs) with sensitivity. In _Proceedings of the 2nd conference on conversational user interfaces_. 1–8.
* Tax et al. (1998) Stephen S Tax, Stephen W Brown, and Murali Chandrashekaran. 1998. Customer evaluations of service complaint experiences: implications for relationship marketing. _Journal of marketing_ 62, 2 (1998), 60–76.
* Tay et al. (2014) Benedict Tay, Younbo Jung, and Taezoon Park. 2014. When stereotypes meet robots: the double-edge sword of robot gender and personality in human–robot interaction. _Computers in Human Behavior_ 38 (2014), 75–84.
* Tolmeijer et al. (2021) Suzanne Tolmeijer, Naim Zierau, Andreas Janson, Jalil Sebastian Wahdatehagh, Jan Marco Marco Leimeister, and Abraham Bernstein. 2021. Female by Default?–Exploring the Effect of Voice Assistant Gender and Pitch on Trait and Trust Attribution. In _Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems_. 1–7.
* Turiman et al. (2013) Syamimi Turiman, Amelia Leong, and Fauziah Hassan. 2013. Are Men More Apologetic Than Women? _Pertanika Journal of Social Sciences & Humanities_ 21, 3 (2013).
* UNESCO (2019) EQUALS Skills Coalition UNESCO. 2019. I’d blush if I could: closing gender divides in digital skills through education. https://unesdoc.unesco.org/ark:/48223/pf0000367416.page=1. Accessed: 2021-09-02.
* van Ginkel et al. (2020) Joost R van Ginkel, Marielle Linting, Ralph CA Rippe, and Anja van der Voort. 2020. Rebutting existing misconceptions about multiple imputation as a method for handling missing data. _Journal of personality assessment_ 102, 3 (2020), 297–308.
* Woods (2018) Heather Suzanne Woods. 2018. Asking more of Siri and Alexa: Feminine persona in service of surveillance capitalism. _Critical Studies in Media Communication_ 35, 4 (2018), 334–349.
|
# Ground state preparation with shallow variational warm-start
Youle Wang<EMAIL_ADDRESS>Chenghong Zhu<EMAIL_ADDRESS>Mingrui Jing<EMAIL_ADDRESS>Xin Wang<EMAIL_ADDRESS>Institute
for Quantum Computing, Baidu Research, Beijing 100193, China
###### Abstract
Preparing the ground states of a many-body system is essential for evaluating
physical quantities and determining the properties of materials. This work
provides a quantum ground state preparation scheme with shallow variational
warm-start to tackle the bottlenecks of current algorithms, i.e., demand for
prior ground state energy information and lack of demonstration of efficient
initial state preparation. Particularly, our methods would not experience the
instability for small spectral gap $\Delta$ during pre-encoding the phase
factors since our methods involve only $\widetilde{O}(1)$ factors while
$\widetilde{O}(\Delta^{-1})$ is requested by the near-optimal methods. We
demonstrate the effectiveness of our methods via extensive numerical
simulations on spin-$1/2$ Heisenberg models. We also show that the shallow
warm-start procedure can process chemical molecules by conducting numerical
simulations on the hydrogen chain model. Moreover, we extend research on the
Hubbard model, demonstrating superior performance compared to the prevalent
variational quantum algorithms.
## I Introduction
The ground state is one crucial quantum state appearing in condensed matter
physics, quantum chemistry, and combinatorial optimisation. Physical reality
tends to force a quantum system to its ground state at low temperature, which
can exist with stability Feynman and Cohen (1956). Therefore, preparing the
exact form of the ground state becomes significant for evaluating the
expectations of physical quantities and determining the properties of
materials Sakurai and Commins (1995). Classical methods have raised success in
predicting matter behaviours. Based on quantum mechanics, traditional methods,
including, Hartree-Fock (HF) method Sherrill (2009), full configuration
interaction Knowles and Handy (1989), density functional method Parr (1980),
and Monte Carlo simulations Blankenbecler _et al._ (1981) have boosted
developments of quantum chemistry by providing accurate eigenstate
evaluations. However, the exponential growth of Hilbert space and multi-
determinant interactions Cade _et al._ (2020); Dagotto (1994); Nelson _et
al._ (2019) demand an explosive amount of memories leading to a restricted
efficiency regime and forcing the problem into an NP-complete complexity
Barahona (1982). The fermionic statistics also raise the sign problem, which
pulls down the convergence speed of approximation methods Gilks _et al._
(1995); Del Moral _et al._ (2006); Scalapino (2006); LeBlanc _et al._
(2015).
Quantum computer, originally proposed by Feynman Feynman (1982), is naturally
embedded in the quantum Hilbert space, indicating promising and competitive
near-term applications, including solving ground states of different physical
models and molecules Peruzzo _et al._ (2014); Wecker _et al._ (2015). The
typical results from variational quantum eigensolver (VQE) Peruzzo _et al._
(2014) demonstrated an excellent perspective for solving the ground state
energies of small molecules Kandala _et al._ (2017). Advanced variational
ansatzes have been designed and achieved better performance predicting the
physical properties and electronic behaviours Wecker _et al._ (2015); Dagotto
(1994), such as charge and spin densities, metal-insulator transitions, and
dynamics of distinct models, which could further energise the industrial
development Sapova and Fedorov (2022) covering new material design Ostaszewski
_et al._ (2021) and studies on high-temperature superconductivity Dallaire-
Demers _et al._ (2019); Cade _et al._ (2020). Another category, namely,
adiabatic quantum algorithms has practically solved the ground state of the
Ising model, demonstrating the potential of handling the combinatorial
optimisation problems Johnson _et al._ (2011).
One of the most promising technical routes to prepare the ground state is to
utilise phase estimation. Under a reasonable assumption that the initial state
$\rho$ has a non-zero overlap $\gamma$ regarding the exact ground state and
the ground state energy (within the required precision) is known, one can
directly apply phase estimation Kitaev (1995) together with amplitude
amplification Brassard _et al._ (2002) can project out the ground state from
$\rho$ with a probability proportional to $\gamma^{2}$. However, phase
estimation usually demands an extensive ancilla system, making the realisation
of preparing high-fidelity states challenging. Recently, Ge _et al._ (2019)
proposed a method to exponentially and polynomially improve the run-time on
the allowed error $\varepsilon$ and the spectral gap $\Delta$ and the initial
state overlap $\gamma$ dependencies, respectively, where $\Delta$ is the
difference between the ground and the first excited state energies. The same
authors also discussed that polylogarithmic scaling in $\varepsilon^{-1}$
could be attained using the method proposed by Poulin and Wocjan (2009)
requiring more ancilla qubits.
Quantum singular value transformation (QSVT) Gilyén _et al._ (2019) is
another promising route for realising the ground state preparation. Notably,
Lin and Tong (2020) have shown a method using constant number of ancilla
qubits to achieve a quantum complexity
$\mathcal{O}(\Delta^{-1}\gamma^{-1}\log(\varepsilon^{-1}))$111The
normalisation factor $\alpha$ in the block encoding is omitted.. Their
algorithms use the block encoding input model of the Hamiltonian Berry _et
al._ (2015) rather than the evolution operator used in phase estimation. They
also provide lower bounds of quantum complexity to show the algorithms’
dependence optimality on the spectral gap and initial overlap. Subsequently,
Dong _et al._ (2022) proposed a technique called “quantum eigenvalue
transformation of unitary matrices” (QET-U), which is closely related to
quantum signal processing Low and Chuang (2017a) and QSVT Gilyén _et al._
(2019). With QET-U and controlled Hamiltonian evolution, they further reduce
the ancilla qubits to two while reaching the same asymptotic scaling in
$\varepsilon$, $\Delta$, $\gamma$ as Lin and Tong (2020). Remarkably, both Lin
and Tong (2020) and Dong _et al._ (2022) assume an upper bound $\mu$ of
ground state energy at priority. If no such $\mu$ is known, it can be obtained
by first employing ground state energy estimation Abrams and Lloyd (1999);
Higgins _et al._ (2007); Berry _et al._ (2009); Poulin and Wocjan (2009);
Knill _et al._ (2007); Lin and Tong (2020); Dong _et al._ (2022); Lin and
Tong (2022); Ding and Lin (2022); O’Brien _et al._ (2019); Wang _et al._
(2019); Somma (2019); Ge _et al._ (2019); Wan _et al._ (2022); Wang _et
al._ (2022a) with precision $\mathcal{O}(\Delta)$, which would increase the
cost of state preparation.
Fig 1: The outline of our ground state preparation with shallow variational
warm-start involving (left) the three applicable physical models of Heisenberg
spin-$1/2$ chain model, chemical molecular model and Fermi-Hubbard model could
be efficiently solved via our scheme; (middle) the two main components of our
scheme of (a) shallow-depth VQE as the warm-start and (b) the ground state
preparation algorithms provided the warm-start initial state; (right) our
demonstrations from testing (top) warm-start with a massive number of models
and (bottom) the principle of the ground state preparation. The sign function
$F(x)$ is simulated via the ground state preparation circuit.
To solve ground-state problems in physics and chemistry, we notice the demand
for overcoming limitations set by the current algorithms, such as the prior
information of the ground-state energy and an efficient initial state
preparation. This paper addresses these issues using shallow-depth VQE and the
recent advances in quantum phase estimation Wang _et al._ (2022b).
Intuitively, one would expect to prepare the ground state with very high
fidelity when estimating the ground state energy. However, it cannot be
realised using the conventional phase estimation due to the imperfect residual
states, as discussed in Poulin and Wocjan (2009); Ge _et al._ (2019).
Fortunately, this expectation can be fulfilled by the algorithm of Wang _et
al._ (2022b), which uses only one ancilla qubit and can apply to block
encoding and Hamiltonian evolution input models. We show that measuring the
ancilla qubit can estimate the ground state energy up to the desired precision
with a probability $\approx\gamma^{2}$. Setting the precision as
$\mathcal{O}(\Delta)$, the ground state can be prepared up to exponentially
high precision at a cost
$\widetilde{\mathcal{O}}(\gamma^{-2}\Delta^{-1})$222Notation
$\widetilde{\mathcal{O}}$ hides the logarithmic factors.. As a notable
property, the prior ground state energy information is no longer necessary.
Methods | Input | #Phase | Query | #Ancilla Qubits | Need prior | Need amp.
---|---|---|---|---|---|---
Model | Factors | Complexity | GSE info.? | amplif.?
This work | HE | $\widetilde{\mathcal{O}}(1)$ | $\widetilde{\mathcal{O}}(\Delta^{-1}\gamma^{-2})$ | $1$ | No | No
BE | $\widetilde{\mathcal{O}}(\alpha\Delta^{-1}\gamma^{-2})$ | 2
Dong _et al._ (2022) | HE | $\widetilde{\mathcal{O}}(\Delta^{-1})$ | $\widetilde{\mathcal{O}}(\Delta^{-1}\gamma^{-2})$ | 1 | Yes | No
$\widetilde{\mathcal{O}}(\Delta^{-1}\gamma^{-1})$ | 2 | Yes
Lin and Tong (2020) | BE | $\widetilde{\mathcal{O}}(\Delta^{-1})$ | $\widetilde{\mathcal{O}}(\alpha\Delta^{-1}\gamma^{-1})$ | $\mathcal{O}(1)$ | Yes | Yes
Ge _et al._ (2019) | HE | — | $\widetilde{\mathcal{O}}(\Delta^{-1}\gamma^{-1})$ | $\mathcal{O}\left(\log(\Delta^{-1})+\log\log(\varepsilon^{-1})\right)$ | Yes | Yes
Table 1: Comparison to the current quantum ground-state preparation
algorithms. $\Delta$ is a lower bound of the spectral gap of the Hamiltonian.
Here, we use $\gamma=\sqrt{\operatorname{tr}(\rho\Pi)}$ to represent the
overlap between the initial and ground states, where $\Pi$ denotes a
projection onto the subspace generated by ground states. “GSE” represents the
ground state energy, and “amp. amplif.” is the abbreviation of amplitude
amplification. “HE” means Hamiltonian evolution $U=e^{iH}$, and “BE” stands
for block encoding $U_{H}=\left[\begin{smallmatrix}H/\alpha&\cdot\\\
\cdot&\cdot\end{smallmatrix}\right]$, where $\alpha$ is the normalisation
factor. If $U_{H}$ is constructed by the linear-combination-of-unitaries
technique Childs _et al._ (2018), the ancillas of our method can be further
reduced to 1. The number of ancillas of Lin and Tong (2020) is at least three,
which are used to block-encode the projector by QSVT. The ancillas for
implementing amplitude amplification need to be counted as well.
In our algorithms, we apply a sequence of single-qubit rotations on the
ancilla qubit and interleave the power of the input models. The structures of
QSVT and QET-U are similar to ours but use block encoding and Hamiltonian
evolution, respectively. Rotation angles, usually referred to as the phase
factors, are chosen so that the circuit can simulate a specific function, such
as sign function Lin and Tong (2020); Wang _et al._ (2022b) and shifted sign
function Dong _et al._ (2022). The number of phase factors is crucial for
preparing a high-fidelity ground state. In Lin and Tong (2020); Dong _et al._
(2022), the number scales as $\mathcal{O}(\Delta^{-1}\log(\varepsilon^{-1}))$,
given the spectral gap $\Delta$ and an approximation error $\varepsilon$.
However, classically computing the factors would suffer from numerical
instability, especially when $\Delta$ is small. A convex-optimisation-based
method is proposed in Dong _et al._ (2022) to generate a near-optimal
approximation. In contrast, our algorithms advantageously require
$\widetilde{\mathcal{O}}(1)$ number of factors, endowing better applicability
in practice than Lin and Tong (2020); Dong _et al._ (2022). In particular,
our numerical experiments have used circuits with fixed phase factors to
prepare ground states of more than 200 random Hamiltonian instances. To better
understand our algorithms, we provide a schematic diagram in Fig. 1 and
summarize the comparisons to Ge _et al._ (2019); Lin and Tong (2020); Dong
_et al._ (2022) in Table 1.
Although algorithms of Lin and Tong (2020); Dong _et al._ (2022) have better
scaling in $\gamma$, the scheme relies on amplitude amplification incurring
more ancilla qubits and deeper circuits. The quantum complexity can be
compensated by preparing an initial state with an overlap larger than 0.1
since the quadratic scaling in overlap would make no big difference. To this
end, we have conducted extensive zero-noise numerical simulations to apply a
depth-three VQE to warm-start ground-state preparation of many-body systems
and find that a shallow-depth VQE could play a good overlap initialiser.
Firstly, we randomly test 10000 two-body interactions, spin-$1/2$ Heisenberg-
type models for every case $\leq$ 10 qubits, 200 and 100 models for the cases
of 12 and 14 qubits, respectively. We observe that the optimised overlaps
highly cluster between 0.4 and 1, and ground states with the fidelity of
almost one are prepared via our algorithms. Secondly, we investigate the
shallow-depth warm-start on molecular ground state preparation offering
satisfactory performances for small bond distances. Thirdly, we apply our
methods to solve the Hubbard model which the 2D case is recognised as a
classical challenge LeBlanc _et al._ (2015); Scalapino (2006); hub (2013);
Cade _et al._ (2020) and quantumly demonstrated only by VQE-based methods
Cade _et al._ (2020); Wecker _et al._ (2015). Our scheme derives extremely
accurate charge-spin densities and the chemical potentials of a specific 1D
model, reaching an absolute error around $10^{-5}$ beyond the chemical
accuracy with respect to the full configuration interaction (FCI) method.
Organisation. In Sec. II, we give technical details of algorithms for ground
state preparation and provide a method to analyse the success probability. In
Sec. III, we show extensive numerical results that demonstrate the efficacy of
a constant-depth VQE in processing Heisenberg models up to 14 qubits. We also
show correctness by applying the algorithms to the optimised states of VQE. In
Sec. IV, we numerically examine the constant-depth warm-start on a hydrogen
chain model. In Sec. V, we use our algorithms to study the Fermi-Hubbard
models and accurately evaluate the model’s physical properties. Finally, we
summarize our results and discuss potential applications in Sec. VI.
## II Ground state preparation
In this section, we present ground-state preparation algorithms. The core of
our approaches has two steps: firstly, prepare an initial state with a
considerable overlap, such as states generated by shallow-depth VQE, Gibbs-VQE
Wang _et al._ (2021), or Hartree-Fock states (which will be discussed in
numerical experiments); secondly, use a quantum algorithm to project out the
ground state from the initial state with high probability. The projection can
be realised using the quantum algorithmic tool, namely the quantum phase
search algorithm (QPS) Wang _et al._ (2022b), which can classify the
eigenvectors of a unitary operator. Using QPS, the probability of sampling the
ground state is highly approximate to the initial overlap. Therefore, by
repeating these two steps sufficient times, one can obtain the desired state
with arbitrarily high precision.
Motivated by the previous works Dong _et al._ (2022); Ge _et al._ (2019);
Lin and Tong (2020), we establish the algorithms for the ground state
preparation with two different Hamiltonian input modes: one is the real-time
Hamiltonian evolution $U=e^{iH}$, and the other uses the block encoding
operator Low and Chuang (2019), which is a unitary matrix
$U_{H}=\left[\begin{smallmatrix}H/\alpha&\cdot\\\
\cdot&\cdot\end{smallmatrix}\right]$ with a scaled Hamiltonian in the upper-
left corner. Since the size of $U_{H}$ is larger than that of $H$, it would
require an ancilla system to implement $U_{H}$ on quantum computers. A common
strategy to block-encode the Hamiltonian of a many-body system is the linear-
combination-of-unitaries (LCU) technique Childs _et al._ (2018). We,
therefore, could regard the LCU-type block encoding as the input of our
algorithms. To better describe our algorithms, we first make some assumptions:
the ancilla system for implementing $U_{H}$ is composed of $m$ qubits, the
spectrum of the matrix $H$ is restricted in the region $(-\pi,\pi)$, the
spectral gap of $H$ is larger than $\Delta>0$, and $\|H\|\leq\alpha$ regarding
the spectral norm. With these, the results of the algorithms are summarized in
Theorem 1.
###### Theorem 1 (Ground state preparation)
Suppose a Hamiltonian of the form
$H=\sum_{j=0}^{2^{n}-1}\lambda_{j}\op{\psi_{j}}{\psi_{j}}$, where
$-\pi\leq\lambda_{0}\leq\lambda_{1}\leq\ldots\leq\lambda_{2^{n}-1}\leq\pi$.
Let $\Delta>0$ be a lower bound of the spectral gap. Assume access to a
quantum state $\rho$ that has a non-zero overlap
$\gamma=\sqrt{\operatorname{tr}(\rho\Pi)}$, where $\Pi$ denotes a projection
operator onto the ground state subspace. Then there exists a quantum algorithm
that prepares the ground state with a precision of at least $1-\varepsilon$.
* •
If the Hamiltonian is accessed through a Hamiltonian evolution operator
$U=e^{iH}$, then the total queries to controlled $U$ and $U^{\dagger}$ is
$\mathcal{O}(\gamma^{-2}\Delta^{-1}\log(\varepsilon^{-1}\log(\Delta^{-1})))$.
The number of used ancilla qubits is $1$ and used copies of $\rho$ is
$\mathcal{O}(\gamma^{-2})$.
* •
If the Hamiltonian is accessed through a block encoding operator
$U_{H}=\left[\begin{smallmatrix}H/\alpha&\cdot\\\
\cdot&\cdot\end{smallmatrix}\right]$ that satisfies $\left\langle
0^{m}\middle|U_{H}\middle|0^{m}\right\rangle=H/\alpha$, then the total queries
to controlled $U_{H}$ and $U_{H}^{\dagger}$ is
$\mathcal{O}(\alpha\gamma^{-2}\Delta^{-1}\log(\varepsilon^{-1}\log(\alpha\Delta^{-1})))$.
The number of used ancilla qubits is $2$ and used copies of $\rho$ is
$\mathcal{O}(\gamma^{-2})$.
The trace distance of quantum states characterises the precision, and the
notation $|0^{m}\rangle$ really indicates $|0^{m}\rangle\otimes I$, where we
have omitted the identity operator on the main system.
Next, we concretely discuss the algorithm using Hamiltonian evolution in Sec.
II.1. We provide the detailed procedure of finding the ground state via QPS
and the analysis of the output probability. The algorithm using block encoding
is further discussed in Sec. II.2. Since the analysis of the algorithm’s
correctness is similar from the previous literature, we mainly discuss the
differences from using the Hamiltonian evolution.
### II.1 Algorithm using real-time Hamiltonian evolution
Given access to the controlled versions of the unitary $U=e^{iH}$ and
$U^{\dagger}=e^{-iH}$, we construct the quantum circuit of QPS as shown in
Fig. 2.
$\Qcircuit@C=0.2em@R=1em{\lstick{|0\rangle}&\gate{R_{z}(\phi_{L})}\gate{R_{y}(\theta_{L})}\ctrl{1}\gate{R_{z}(\phi_{L-1})}\gate{R_{y}(\theta_{L-1})}\ctrlo{1}\gate{R_{z}(\phi_{L-2})}\gate{R_{y}(\theta_{L-2})}\qw\cdots\ctrlo{1}\gate{R_{z}(\phi_{0})}\gate{R_{y}(\theta_{0})}\gate{R_{z}(\omega)}\qw\meter\\\
\lstick{\rho}/^{n}\qw\qw\gate{U}\qw\qw\gate{U^{\dagger}}\qw\qw\qw\cdots\gate{U^{\dagger}}\qw\qw\qw\qw}$
Fig 2: Quantum circuit for quantum phase search. The unitary $U$ would be
adaptively changed with its power. Parameters $\bm{\phi}$, $\bm{\theta}$, and
$\omega$ are chosen such that the circuit can implement a trigonometric
polynomial transformation to each phase of the unitary operator. The integer
$L$ is the order of the polynomial.
There is an ancilla system consisting of only one qubit, which is applied with
a sequence of single-qubit rotations. Meanwhile, several controlled unitaries
are interleaved between rotations. The phase factors $\bm{\phi}$,
$\bm{\theta}$, and $\omega$ are chosen such that the circuit implements a
trigonometric polynomial transformation to each phase of the unitary, i.e.,
$f\mathrel{\mathop{\mathchar
58\relax}}\lambda\mapsto\sum_{j=-L}^{L}c_{j}e^{ij\lambda}$, where each
$c_{j}\in\mathbb{C}$ and satisfies $\sum_{j}|c_{j}|\leq 1$. For a certain
trigonometric polynomial, the corresponding factors can be classically
computed beforehand Dong _et al._ (2021); Wang _et al._ (2022b). The action
of running the circuit to an eigenvector of the unitary $U$ is presented in
the lemma below.
###### Lemma 2 (Quantum phase evolution Wang _et al._ (2022b))
Given a unitary operator
$U=\sum_{j=0}^{2^{n}-1}e^{i\lambda_{j}}|\psi_{j}\rangle\\!\langle\psi_{j}|$,
for any trigonometric polynomial $f(x)=\sum_{j=-L}^{L}c_{j}e^{ijx}$ with
$\sum_{j}|c_{j}|\leq 1$, there exist parameters $\omega\in{{\mathbb{R}}}$ and
$\bm{\theta},\bm{\phi}\in{{\mathbb{R}}}^{L+1}$ such that the circuit in Fig. 2
transforms the initial state $|0,\psi\rangle$ into
$\left(\sqrt{(1+f(\lambda))/2}|0\rangle+\sqrt{(1-f(\lambda))/2}|1\rangle\right)|\psi\rangle$,
where $|\psi\rangle$ denotes an eigenvector of $U$ with a phase
$\lambda\in(-\pi,\pi)$.
For QPS, the trigonometric polynomial $f(x)=\sum_{j=-L}^{L}c_{j}e^{ijx}$ is
chosen to approximate the sign function. Technically, supposing a constant
$\kappa\in(0,1/2)$, we could find a $f(x)$ with order
$L\in\mathcal{O}(\kappa^{-1}\log(\varepsilon^{-1}))$ such that
$|f(x)-1|\leq\varepsilon$ for all $x\in(\kappa,\pi-\kappa)$ and
$|f(x)+1|\leq\varepsilon$ for all $x\in(-\pi+\kappa,-\kappa)$ Low and Chuang
(2017b); Wang _et al._ (2022b). Particularly, given the relation between the
order $L$ and precision $\varepsilon$, we could choose $\varepsilon$ to be
exponentially small at the expense of a modest increase in the circuit depth.
Therefore, eigenvectors fall in the region
$(-\pi+\kappa,-\kappa)\cup(\kappa,\pi-\kappa)$ are labelled by states 0 or 1,
and the collapsed state from the measured ancilla qubit could indicate the
label. For instance, for any state
$|\psi\rangle=\sum_{j}\alpha_{j}|\psi_{j}\rangle$, the circuit of QPS
transforms the initial state $|0,\psi\rangle$ into the following form.
$\displaystyle|0\rangle\sum_{j}\alpha_{j}\sqrt{\frac{1+f(\lambda_{j})}{2}}|\psi_{j}\rangle+|1\rangle\sum_{j}\alpha_{j}\sqrt{\frac{1-f(\lambda_{j})}{2}}|\psi_{j}\rangle.$
(1)
We readily know that $|0\rangle$ is associated with eigenvectors in the region
$(-\kappa,\pi+\kappa)$, and $|1\rangle$ is associated with eigenvectors in the
region $(-\pi-\kappa,\kappa)$. The only possible source of error comes from
the intrinsic sign function approximating error from the searching procedures
discussed previously, which raises negligible effect. This property enables
one to efficiently locate the region where the ground state energy falls
through a binary search.
#### Rough ground state energy search
By executing the binary search, the region containing the ground state energy
would shrink gradually, and the probability of outputting the ground state is
analyzed later. After each iteration, the underlying region is cut into two
halves, and one of them is chosen, where the post-measurement state falls in,
for the next iteration. Since the region’s size decreases by nearly half, the
search could halt in a logarithmic time, producing an exponentially-small
region. When the final region of the ground state is attained, any point in
the final region could be taken as an estimate of ground state energy. In
particular, if the estimation accuracy is smaller than half of the spectral
gap, the ground state would eventually fall into a region whose size is
smaller than the spectral gap, and excited states would be filtered out. At
that time, the post-measurement state on the primary system is the desired
ground state.
To better understand the algorithm, we sketch the search procedure below. For
remark, the final post-measurement state is not the desired ground state. We
defer the discussion of improving the estimation accuracy.
1. 1.
Input: unitary $U=e^{iH}$, constant $\kappa$ (irrelevant to $H$), error
tolerance $\varepsilon$, and state $\rho$ that non-trivially overlaps with the
ground state.
2. 2.
Compute parameters $\bm{\phi}$, $\bm{\theta}$, and $\omega$ according to
parameters $\kappa$ and $\varepsilon$.
3. 3.
Choose the initial region $(x_{0},x_{1})$, where $x_{0}=-\pi$ and $x_{1}=\pi$.
4. 4.
Calculate the middle point of the region $(x_{0},x_{1})$, which is $x_{\rm
mid}=\frac{x_{0}+x_{1}}{2}$.
5. 5.
Set the circuit of QPS using controlled $\tilde{U}=e^{-ix_{\rm mid}}U$ and
$\tilde{U}^{\dagger}$ and the calculated parameters from the step 2.
6. 6.
Input the initial state $\op{0}{0}\otimes\rho$ and run the circuit.
7. 7.
Measure the ancilla qubit of the circuit and update the region according to
the measurement outcome: when $x_{1}-x_{0}>2\pi-2\kappa$, execute the update
$\left\\{\begin{matrix}(x_{0},x_{1})\leftarrow(x_{\rm
mid}-\kappa,x_{1}+\kappa),&\textrm{if the outcome is 0;}\\\\[5.0pt]
(x_{0},x_{1})\leftarrow(x_{0}-\kappa,x_{\rm mid}+\kappa),&\textrm{if the
outcome is 1.}\end{matrix}\right.$ (2)
When $x_{1}-x_{0}\leq 2\pi-2\kappa$, execute the update
$\left\\{\begin{matrix}(x_{0},x_{1})\leftarrow(x_{\rm
mid}-\kappa,x_{1}),&\textrm{if the outcome is 0;}\\\\[5.0pt]
(x_{0},x_{1})\leftarrow(x_{0},x_{\rm mid}+\kappa),&\textrm{if the outcome is
1.}\end{matrix}\right.$ (3)
8. 8.
Go to step 4 and repeat till the region converges.
9. 9.
Output: final region $(x_{0},x_{1})$.
After executing the rough search procedure, the size of the region would
converge to $2\kappa$. Specifically, after the first iteration, the size of
the region is $\kappa+\pi/2$. Since then, the region is updated as
$(\theta-\kappa,x_{1})$ or $(x_{0},\theta+\kappa)$, and the size of the new
region is $\kappa+(x_{1}-x_{0})/2$. Then we can inductively find the size of
the final region. Let $(x_{0}^{(t)},x_{1}^{(t)})$ denote the region in the
$t$-iteration and $Q$ denote the total number of iterations, then we have
$\displaystyle|x_{1}^{(Q)}-x_{0}^{(Q)}|$
$\displaystyle=\kappa+\frac{|x_{1}^{(Q-1)}-x_{0}^{(Q-1)}|}{2}=\kappa+\frac{1}{2}\left[\kappa+\frac{|x_{1}^{(Q-2)}-x_{0}^{(Q-2)}|}{2}\right]=\cdots=2\kappa+\frac{\pi}{2^{Q}}.$
(4)
#### Improving search accuracy
Suppose the ground state energy falls in the region
$(x_{0}^{(Q)},x_{1}^{(Q)})$, then the above procedure would return a rough
estimate of the ground state energy up to precision $\kappa+\pi/2^{Q+1}$.
However, the estimation precision is not guaranteed to be smaller than
$\Delta/2$. Thus more processes are needed to improve the precision. It is
realised by applying the power of the unitary operator that amplifies the
phases of the post-measurement state. By repeatedly running the circuit with
the unitary power, the ground state energy is efficiently constrained to a
smaller region. To be specific, firstly, run the rough search procedure once
with unitary $U^{(0)}=U$, attaining a region of size $2\kappa+\pi/2^{Q}$. For
clarity, denote this region by $(\zeta_{0}^{(1)},\zeta_{1}^{(1)})$ and
$\bar{\kappa}=\kappa+\pi/2^{Q+1}$, where
$|\zeta_{1}^{(1)}-\zeta_{0}^{(1)}|=2\bar{\kappa}$. Let
$\lambda^{(1)}=(\zeta_{0}^{(1)}+\zeta_{1}^{(1)})/2$ denote the middle point of
the region. Secondly, consider a modified unitary
$U^{(1)}=(e^{-i\lambda^{(1)}}U^{(0)})^{\lfloor 1/\bar{\kappa}\rfloor}$333The
notation $\lfloor x\rfloor$ denotes the floor function that gives as output
the greatest integer less than or equal to $x$.. For $U^{(1)}$, phases of
eigenvectors in the post-measurement state are rescaled to $(-1,1)$. Now,
running the rough search procedure with the new unitary operator would give a
new region $(\zeta_{0}^{(2)},\zeta_{1}^{(2)})$.
Executing the above two steps would exponentially fast improve the estimation
accuracy compared with the rough searching scheme. For state
$|\psi_{0}\rangle$, let $\lambda_{0}$ denote its original phase in $U$. Its
phase in unitary $U^{(1)}$ is $(\lambda_{0}-\lambda^{(1)})\lfloor
1/\bar{\kappa}\rfloor$ and falls in the region
$(\zeta_{0}^{(2)},\zeta_{1}^{(2)})$. Let
$\lambda^{(2)}=(\zeta_{0}^{(2)}+\zeta_{1}^{(2)})/2$ denote the middle point of
the second region $(\zeta_{0}^{(2)},\zeta_{1}^{(2)})$. Then we readily give an
inequality that characterises the estimation error
$\left|\lambda^{(2)}-(\lambda_{0}-\lambda^{(1)})\lfloor
1/\bar{\kappa}\rfloor\right|\leq\bar{\kappa}$. Rewrite this inequality as
$\left|\lambda_{0}-\left(\lambda^{(1)}+\lambda^{(2)}/\lfloor
1/\bar{\kappa}\rfloor\right)\right|\leq\bar{\kappa}/\lfloor
1/\bar{\kappa}\rfloor\leq\lfloor 1/\bar{\kappa}\rfloor^{-2}$. From this
equation, we can see that the first two steps of the scheme give an estimate
of the phase with an error of $\lfloor 1/\bar{\kappa}\rfloor^{-2}$. So,
inductively repeating the above procedure leads to a sequence
$(\lambda^{(1)},\lambda^{(2)},\ldots)$ that could be taken as an estimate of
the ground state energy. Assuming repeats $j$ times, the estimate is given by
$\displaystyle\left|\lambda_{0}-\left(\lambda^{(1)}+\frac{\lambda^{(2)}}{\lfloor
1/\bar{\kappa}\rfloor}+\frac{\lambda^{(3)}}{\lfloor
1/\bar{\kappa}\rfloor^{2}}+\cdots+\frac{\lambda^{(j)}}{\lfloor
1/\bar{\kappa}\rfloor^{j-1}}\right)\right|\leq\frac{\bar{\kappa}}{\lfloor
1/\bar{\kappa}\rfloor^{j-1}}\leq\lfloor 1/\bar{\kappa}\rfloor^{-j}.$ (5)
Clearly, reaching the halt condition only requires repeating at most
$\mathcal{O}(\log(1/\Delta)/\log(\lfloor 1/\bar{\kappa}\rfloor))$ times.
#### Output probability analysis
Due to many intermediate measurements in QPS, the search procedure is
conducted randomly. The event that the ground state is output would occur with
a probability. However, how the analysis of this probability is missing in
Wang _et al._ (2022b). We use an example to explain how to analyse the output
probability, which can be used in more general cases.
Example. If we collect all measurement outcomes together to represent a
trajectory that the ground state moves along, the search process of QPS can be
depicted using a binary tree as shown in Fig. 3. The root node represents the
initial state $\rho$, and other nodes represent the post-measurement state
with a non-zero overlap concerning the ground state. Each leaf node in the
tree represents a possible searched ground state from the program. Each
trajectory consists of four nodes, meaning that QPS finds the ground state
after three iterations. In addition, the line goes left, meaning that the
measurement outcome is 0, and goes right, meaning that the measurement outcome
is 1.
$\rho$$\rho_{1,1}$$\rho_{2,1}$$\rho_{3,1}$$\rho_{2,2}$$\rho_{3,2}$$\rho_{1,2}$$\rho_{2,3}$$\rho_{3,3}$$\rho_{3,4}$
Fig 3: An example of a binary tree diagram for sketching phase search. The
node with notation $\rho_{l,k}$ represents the $k$th post-measurement state at
$l$ iteration.
By the law of total probability, the probability of finding the ground state
is the sum of the probability of following each trajectory. We take the most
left trajectory that reaches the leaf node $\rho_{3,1}$ as an example to
explain how to compute the probability of a trajectory. Let
$\Pr[\rho_{i,j}|\rho_{p,l}]$ denote the conditional probability of reaching
node $\rho_{i,j}$ from node $\rho_{p,l}$. Clearly, the probability that
$\rho_{p,l}$ transits to $\rho_{i,j}$ only depends on the state $\rho_{p,l}$.
Then the probability of reaching the leaf node $\rho_{3,1}$ from the root node
can be given by
$\displaystyle\Pr[\rho_{3,1}]=\Pr[\rho_{3,1}|\rho_{2,1}]\times\Pr[\rho_{2,1}|\rho_{1,1}]\times\Pr[\rho_{1,1}|\rho].$
(6)
To represent probability $\Pr[\rho_{1,1}|\rho]$, we define linear maps
$\mathcal{M}_{(1,1)}^{(0)}$ that transform the input density matrix $\rho$ to
an un-normalised post-measurement state, where $k\in\\{0,1\\}$ indicates the
measurement outcome. Recalling Theorem 6 of Wang _et al._ (2022b), the
measurement probabilities are
$\Pr[0]=\frac{1}{2}(1+\sum_{j}\left\langle\psi_{j}\middle|\rho\middle|\psi_{j}\right\rangle
f(\lambda_{j}))$ and
$\Pr[1]=\frac{1}{2}(1-\sum_{j}\left\langle\psi_{j}\middle|\rho\middle|\psi_{j}\right\rangle
f(\lambda_{j}))$. Then the action of $\mathcal{M}_{(1,1)}^{(k)}$ is described
as follows.
$\displaystyle\mathcal{M}_{(1,1)}^{(0)}\mathrel{\mathop{\mathchar
58\relax}}\rho=\sum_{l}p_{l}\op{\phi_{l}}{\phi_{l}}\mapsto\sum_{l}p_{l}\op{\widetilde{\phi}_{l}^{(0)}}{\widetilde{\phi}_{l}^{(0)}},\quad\mathrm{where}\quad|\widetilde{\phi}_{l}^{(0)}\rangle=\sum_{j}\sqrt{\frac{1+f(\lambda_{j})}{2}}\langle\psi_{j}|\phi_{l}\rangle|\psi_{j}\rangle.$
(7) $\displaystyle\mathcal{M}_{(1,1)}^{(1)}\mathrel{\mathop{\mathchar
58\relax}}\rho=\sum_{l}p_{l}\op{\phi_{l}}{\phi_{l}}\mapsto\sum_{l}p_{l}\op{\widetilde{\phi}_{l}^{(1)}}{\widetilde{\phi}_{l}^{(1)}},\quad\mathrm{where}\quad|\widetilde{\phi}_{l}^{(1)}\rangle=\sum_{j}\sqrt{\frac{1-f(\lambda_{j})}{2}}\langle\psi_{j}|\phi_{l}\rangle|\psi_{j}\rangle.$
(8)
Note that $|\widetilde{\phi}_{l}^{(0)}\rangle$ and
$|\widetilde{\phi}_{l}^{(1)}\rangle$ are not quantum states but represent
column vectors. For other probabilities, the maps are defined similarly. That
is, the map that transforms $\rho_{i,j}$ to $\rho_{p,l}$ is denoted by
$\mathcal{M}_{(i,j)\to(p,l)}^{(0)}$. With these notations, the probabilities
are re-expressed below.
$\displaystyle\Pr[\rho_{1,1}|\rho]$
$\displaystyle=\operatorname{tr}\left(\mathcal{M}_{(1,1)}^{(0)}(\rho)\right)$
(9) $\displaystyle\Pr[\rho_{2,1}|\rho_{1,1}]$
$\displaystyle=\operatorname{tr}\left(\mathcal{M}_{(1,1)\to(2,1)}^{(0)}(\rho_{1,1})\right)=\frac{\operatorname{tr}\left(\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\rho)\right)}{\operatorname{tr}\left(\mathcal{M}_{(1,1)}^{(0)}(\rho)\right)}$
(10) $\displaystyle\Pr[\rho_{3,1}|\rho_{2,1}]$
$\displaystyle=\operatorname{tr}\left(\mathcal{M}_{(2,1)\to(3,1)}^{(0)}(\rho_{2,1})\right)=\frac{\operatorname{tr}\left(\mathcal{M}_{(2,1)\to(3,1)}^{(0)}\circ\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\rho)\right)}{\operatorname{tr}\left(\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\rho)\right)}$
(11)
Eventually, the probability $\Pr[\rho_{3,1}]$ can be written as
$\displaystyle\Pr[\rho_{3,1}]=\operatorname{tr}\left(\mathcal{M}_{(2,1)\to(3,1)}^{(0)}\circ\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\rho)\right).$
(12)
As shown in Eq. (7), the circuit of QPS allocates weight to each eigenvector
conditional on the state of the ancilla qubit. Specifically, each eigenvector
$|\psi_{j}\rangle$ is allocated with a weight $\sqrt{(1+f(\lambda_{j}))/2}$ if
the ancilla qubit is $|0\rangle$; otherwise, the weight is
$\sqrt{(1-f(\lambda_{j}))/2}$. Since $f(x)$ is an approximation of the sign
function in the sense that $|f(x)-\mathrm{sign}(x)|\leq\varepsilon$ for all
$x\in(-\pi+\kappa,-\kappa)\cup(\kappa,\pi-\kappa)$, and the approximation
error $\varepsilon$ could be exponentially small, we can see that some
eigenvectors are filtered from the un-normalised post-measurement state.
$\begin{array}[]{cc}|\widetilde{\phi}_{l}^{(0)}\rangle=\sum_{j\mathrel{\mathop{\mathchar
58\relax}}\lambda_{j}\not\in(-\pi+\kappa,-\kappa)}\sqrt{\frac{1+f(\lambda_{j})}{2}}\langle\psi_{j}|\phi_{l}\rangle|\psi_{j}\rangle,&|\widetilde{\phi}_{l}^{(1)}\rangle=\sum_{j\mathrel{\mathop{\mathchar
58\relax}}\lambda_{j}\not\in(\kappa,\pi-\kappa)}\sqrt{\frac{1-f(\lambda_{j})}{2}}\langle\psi_{j}|\phi_{l}\rangle|\psi_{j}\rangle.\end{array}$
(13)
Hence, after the operation of
$\mathcal{M}_{(2,1)\to(3,1)}^{(0)}\circ\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}$,
the initial state $\rho$ is filtered, and only the ground state remains. Using
the linearity of the map $\mathcal{M}$ and trace, we can see that
$\displaystyle\Pr[\rho_{3,1}]$
$\displaystyle=\sum_{ljk}p_{l}\langle\psi_{j}|\phi_{l}\rangle\langle\phi_{l}|\psi_{k}\rangle\operatorname{tr}\left(\mathcal{M}_{(2,1)\to(3,1)}^{(0)}\circ\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\op{\psi_{j}}{\psi_{k}})\right)$
(14) $\displaystyle=\sum_{l,j\mathrel{\mathop{\mathchar
58\relax}}\Pi|\psi_{j}\rangle\neq 0,\atop k\mathrel{\mathop{\mathchar
58\relax}}\Pi|\psi_{k}\rangle\neq
0}p_{l}\langle\psi_{j}|\phi_{l}\rangle\langle\phi_{l}|\psi_{k}\rangle\operatorname{tr}\left(\mathcal{M}_{(2,1)\to(3,1)}^{(0)}\circ\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\op{\psi_{j}}{\psi_{k}})\right)$
(15)
$\displaystyle=\operatorname{tr}\left(\mathcal{M}_{(2,1)\to(3,1)}^{(0)}\circ\mathcal{M}_{(1,1)\to(2,1)}^{(0)}\circ\mathcal{M}_{(1,1)}^{(0)}(\Pi\rho\Pi)\right).$
(16)
where $\Pi$ denotes the projection operator onto the subspace of the ground
state. The second equality holds since excited eigenvectors would be assigned
zero weight. Conducting the same process would give expressions for the rest
probabilities $\Pr[\rho_{3,2}]$, $\Pr[\rho_{3,3}]$, and $\Pr[\rho_{3,4}]$.
Finally, simple calculations would immediately give the probability of finding
the ground state.
$\displaystyle\Pr[\rho_{3,1}]+\Pr[\rho_{3,2}]+\Pr[\rho_{3,3}]+\Pr[\rho_{3,4}]=\operatorname{tr}(\rho\Pi).$
(17)
To consider the error of approximating the sign function, the above discussion
can be reduced to the case where the initial state is an eigenvector. Then the
output probability is highly approximate to $\operatorname{tr}(\rho\Pi)$. We
refer the interested readers to Wang _et al._ (2022b) for more complexity
analysis.
For a general case, we can sketch the search process as a binary tree and
calculate the probability of finding the ground state in the same manner.
Hence, we conclude that the probability is highly approximate to the initial
state overlap under the assumption that the Hamiltonian evolution is simulated
very accurately.
### II.2 Algorithm using block encoding
When the input mode is a block encoding operator
$U_{H}=\left[\begin{smallmatrix}H/\alpha&\cdot\\\
\cdot&\cdot\end{smallmatrix}\right]$ of the Hamiltonian, QPS is similarly
employed to prepare the target ground state. The discussion on probability and
correctness is similar to the previous section. In contrast, there are several
differences in using the Hamiltonian evolution. Firstly, implementing $U_{H}$
requires an ancilla system; thus, making the circuit wider. Secondly, QPS
outputs the eigenvalue and eigenvector of the block encoding rather than the
Hamiltonian; hence further processing is demanded. Specifically, when the QPS
finishes, computing the cosine function of the output gives an estimation of
the eigenvalue of the Hamiltonian. Measure the ancilla system of the final
output state to achieve the ground state once the post-measurement state has
the outcomes all zeros; otherwise, one has to restart the QPS procedure with
the initial state or the post-measurement state and halt until all-zero
outcomes are observed. Thirdly, the estimation accuracy is set smaller than
$\Delta/2\alpha$. The corresponding circuit is shown in Fig. 4. More
discussions on implementing block encoding in QPS can be found in Wang _et
al._ (2022b).
$\Qcircuit@C=0.2em@R=1em{\lstick{|0\rangle}&\gate{R_{z}(\phi_{L})}\gate{R_{y}(\theta_{L})}\ctrl{1}\gate{R_{z}(\phi_{L-1})}\gate{R_{y}(\theta_{L-1})}\ctrlo{1}\gate{R_{z}(\phi_{L-2})}\gate{R_{y}(\theta_{L-2})}\qw\cdots\ctrlo{1}\gate{R_{z}(\phi_{0})}\gate{R_{y}(\theta_{0})}\gate{R_{z}(\omega)}\qw\meter\\\
\lstick{|0^{m}\rangle}/^{m}\qw\qw\multigate{1}{U_{H}}\qw\qw\multigate{1}{U_{H}^{\dagger}}\qw\qw\qw\cdots\multigate{1}{U_{H}^{\dagger}}\qw\qw\qw\qw\\\
\lstick{\rho}/^{n}\qw\qw\ghost{U_{H}}\qw\qw\ghost{U_{H}^{\dagger}}\qw\qw\qw\cdots\ghost{U_{H}^{\dagger}}\qw\qw\qw\qw}$
Fig 4: Block-encoding based QPS circuit. $U_{H}$ is a block encoding of the
Hamiltonian. The second line represents the block-encoding’s ancilla system of
size $m$. The third line stands for the main system with $n$ qubits.
A general block encoding cannot be directly employed to prepare the ground
state unless it satisfies an invariant property. When the invariant property
is satisfied, eigenvalues of $U_{H}$ and $H$ can be connected via a cosine
function. Specifically, the action of the block encoding can be described as
$U_{H}|0^{m},\psi_{j}\rangle=\lambda_{j}/\alpha|0^{m},\psi_{j}\rangle+\sqrt{1-(\lambda_{j}/\alpha)^{2}}|\bot_{j}\rangle$,
where $|\bot_{j}\rangle$ denotes a normalised column vector satisfying
$(\op{0^{m}}{0^{m}}\otimes I)|\bot_{j}\rangle=0$. Let $\mathcal{H}_{j}$ denote
a two-dimensional plane spanned by vectors
$\\{|0^{m},\psi_{j}\rangle,|\bot_{j}\rangle\\}$. If $U_{H}$ is invariant in
$\mathcal{H}_{j}$, i.e., $U_{H}\mathrel{\mathop{\mathchar
58\relax}}\mathcal{H}_{j}\to\mathcal{H}_{j}$ is a rotation, we could easily
find the relation between $\lambda_{j}$ and the corresponding phase
$\pm\tau_{j}$ of $U_{H}$, which is $\lambda_{j}/\alpha=\cos(\pm\tau_{j})$. If
$U_{H}$ does not satisfy the invariant property, this issue can be addressed
by using the “qubitization” technique Low and Chuang (2019). The core of
qubitization is constructing a new block encoding operator using one more
ancilla qubit and querying the controlled-$U_{H}$ and its inverse operator
once. Furthermore, the LCU-type block encoding satisfies the aforementioned
invariant property; hence, qubitization is no longer needed in this case.
To output an eigenvector of $U_{H}$ corresponding to the ground state, the
estimation accuracy is set smaller than $\Delta/2\alpha$. Note that the cosine
function is monotonically increasing and decreasing in the region $(-\pi,0)$
and $(0,\pi)$, respectively. We can know that the phase gap of the unitary is
larger than the scaled spectral gap $\Delta/\alpha$. Thus setting accuracy as
$\Delta/2\alpha$ suffices to isolate the ground state. Suppose we have
attained an eigenvector, e.g., $(|0^{m},\psi_{0}\rangle\pm
i|\bot_{0}\rangle)/\sqrt{2}$. Directly measuring on the ancilla system of the
block encoding, the post-measurement state is the ground state if the outcomes
are all zeros. The probability of this happening is exactly $1/2$. If all-zero
outcomes are not observed, one can restart the whole procedure. Or one can
send the post-measurement state into QPS again. Afterwards, the state will
collapse to either $(|0^{m},\psi_{0}\rangle+i|\bot_{0}\rangle)/\sqrt{2}$ or
$(|0^{m},\psi_{0}\rangle-i|\bot_{0}\rangle)/\sqrt{2}$. This process is
repeated until the outcomes are all zeros. Particularly, the failure
probability will fast decay as $(1/2)^{r}$, where $r$ means the number of
repeats.
## III Application to Heisenberg Hamiltonians
The general Heisenberg-typed model is arguably one of the most commonly used
models in the research of quantum magnetism and quantum many-body physics that
researching in the model’s ground state indicates the characteristics of spin
liquid and inspires crystal topology design Yan _et al._ (2011). The
Hamiltonian can be expressed as
$H=\sum_{\langle
i,j\rangle}(J_{x}S_{i}^{x}S_{j}^{x}+J_{y}S_{i}^{y}S_{j}^{y}+J_{z}S_{i}^{z}S_{j}^{z})+\sum_{i}h_{z}S_{i}^{z},$
(18)
with $\langle i,j\rangle$ depends on the specific lattice structure,
$J_{x},J_{y},J_{z}$ describe the spin coupling strength respectively in the
$x$,$y$,$z$ directions and $h_{z}$ is the magnetic field applied along the $z$
direction. Classical approaches involving Bethe ansatz and thermal limitations
Bonechi _et al._ (1992); Faddeev (1996); Rojas _et al._ (2001) have been
developed and the model can be exactly solved. Besides, variational methods,
in particular, VQE provides an alternative practical approach.
For no-noise numerical experiments, we evaluate the performance of constant-
depth VQE on randomly generated Hamiltonians with shuffled spin coupling
strengths. Specifically, we consider the 1D model with adjacent interactions
and randomly assign $J_{xyz}$ coefficients in the interval $[-1,1]$ while
ignoring magnetic fields involving periodic boundary conditions. Furthermore,
we utilise two commonly used parameterised quantum circuits or ansatzes namely
hardware efficient ansatz (HEA) Kandala _et al._ (2017) and alternating
layered ansatz (ALT) Cerezo _et al._ (2021), as illustrated in Fig. 5. And
for all simulations, we randomly initialise circuit parameters and perform 200
optimisation iterations using Adam optimiser with 0.1 learning rate.
$\Qcircuit<EMAIL_ADDRESS>\qw\gate{R_{y}}\gate{R_{z}}\targ\ctrl{1}\qw\gate{R_{y}}\gate{R_{z}}\qw\\\
\qw\gate{R_{y}}\gate{R_{z}}\qw\targ\ctrl{1}\gate{R_{y}}\gate{R_{z}}\qw\\\
\qw\gate{R_{y}}\gate{R_{z}}\qw\qw\targ\gate{R_{y}}\gate{R_{z}}\qw\inputgroupv{1}{4}{0.4em}{3em}{n=4\quad}\gategroup{1}{5}{4}{9}{1.2em}{--}\\\
\\\ \quad\times p}$
$\Qcircuit<EMAIL_ADDRESS>\qw\gate{R_{y}}\qw\ctrl{0}\gate{R_{y}}\ctrl{1}\gate{R_{y}}\qw\qw\\\
\qw\gate{R_{y}}\qw\ctrl{1}\gate{R_{y}}\ctrl{0}\gate{R_{y}}\qw\qw\\\
\qw\gate{R_{y}}\qw\ctrl{0}\gate{R_{y}}\qw\qw\qw\qw\inputgroupv{1}{4}{0.4em}{3em}{n=4\quad}\gategroup{1}{5}{4}{8}{2.4em}{--}\\\
\\\ \times p}$
Fig 5: Four-qubit templates for the experimental ansatz (with $p$ repeated
circuit block) diagrams of the left: Hardward Efficient ansatz (HEA) and the
right: Alternating Layered ansatz (ALT).
#### Algorithm effectiveness via 1D Heisenberg model
First, we investigate the performance of initial state preparation as the
circuit depth increases. In our simulation, we randomly sample 200 $8$-qubit
Heisenberg Hamiltonians to demonstrate the initial state preparation by using
ALT ansatz with circuit depth $d\in\\{1,3,5,7,9\\}$. As shown in Fig. 6, we
categorise the trained overlaps into ranges $(0-0.4)$, $(0.4-1.0)$,
$(0.6-1.0)$, and $(0.8-1.0)$ and designate them with purple, blue, brown and
orange lines, respectively. We observe that as the circuit depth increases,
the probability of $(0.6-1.0)$ interval rises from $26\%$ to $98\%$ and
$(0.8-1.0)$ overlap interval rises from $0$ to $58.5\%$, while $(0-0.4)$
interval drops from $12.5\%$ to $1\%$. We also find that VQE with circuit
depth can generate the initial state within $(0.4-1.0)$ overlap range with at
least $97\%$ confidence. This finding indicates that depth-3 ansatz can
reliably prepare the initial state for the ground state preparation of the 1D
Heisenberg model and a deeper-depth circuit may further boost the overlap. As
depth-3 ansatz is already capable of achieving satisfactory performance, we
choose it for additional investigations.
Fig 6: The left figure shows the overlap range prepared with ALT ansatz depth
$[1,3,4,5,7]$ for 200 8-qubit Heisenberg Hamiltonians. The purple, blue, brown
and orange lines indicate the intervals $(0.0-0.4)$, $(0.4-1.0)$, $(0.6-1.0)$
and $(0.8-1.0)$ in which the sampled probability of overlaps that prepared
shallow-depth VQE may fall. The right figure shows the overlap prepared by
depth-3 VQE with HEA ansatz and our algorithm. The blue zone represents
overlap generated by shallow-depth VQE, while the red portion represents
overlap produced by our method.
We then demonstrate the validity of our ground state preparation algorithms by
applying the depth-3 VQE for preparing the initial state. The experimental
results are shown in Fig. 6. There are 200 8-qubit Heisenberg Hamiltonians
being sampled, and HEA ansatz with circuit depth 3 is used. The blue region
depicts the overlap generated by a randomly initialised depth-3 VQE, while the
red line represents the overlap boosted by our algorithm given the initial
states prepared before. We can see that the shallow-depth VQE provides an
initial state with a minimum overlap of 0.5 with the ground state. Then, by
inputting the optimised state into our program, we can effectively prepare the
ground state with a fidelity error of less than $10^{-4}$ regarding the true
ground state. For remark, we use the Hamiltonian evolution operator $U=e^{iH}$
and set the estimation accuracy as $1/3$ of the spectral gap.
Previous two experiments demonstrate the depth-3 VQE has satisfactory
performance in the 8-qubit Heisenberg model. To further understand the
scalability of shallow-depth VQE in other system sizes, we conduct extensive
noiseless numerical simulations for system sizes up to 14. Specifically, we
sample 10000 randomly generated Heisenberg Hamiltonians for each qubit size
between 4 to 10, and sample 200 and 100 random Hamiltonians for qubit sizes 12
and 14. As a comparison, ALT and HEA ansatz are both used. The experimental
results are depicted in Fig. 7.
Fig 7: The frequency of overlap prepared by using depth-3 VQE ansatz of ALT
(up) and HEA (down) for varied system sizes. The x-axis labels the overlap
regarding the randomly generated Heisenberg Hamiltonians’ ground states. For
every system with a size less than or equal to 10 qubits, $10000$ random
Hamiltonians are sampled, and for qubit sizes 12 and 14, $200$ and $100$
random Hamiltonians are sampled, respectively. The result shows the efficacy
of depth-3 VQE of two ansatz templates in preparing initial states with an
overlap greater than 0.4 for the intermediate-scaled Heisenberg models.
According to the numerical results, the shallow-depth VQE is shown to be
effective for Heisenberg Hamiltonians. The majority of the results indicate
that the prepared state can have a greater overlap with the ground state than
0.4, making it an acceptable initial state for our algorithm. Also, as the
system size of Heisenberg Hamiltonian increases, the prepared states have a
trend that approximately converged to an overlap between 0.5 and 0.75 and the
full width at half maximum of the sampled probability distribution expands,
implying that, with a randomly initialised shallow-depth parameterised quantum
circuit, the optimised state can be prepared with substantial overlap to the
ground state with high probability. Particularly, our results show that the
technical concern from Ge _et al._ (2019); Dong _et al._ (2022); Lin and
Tong (2020) can be fulfilled by a shallow-depth VQE, indicating the
foreseeable possibility of realising the warm-start applications on large
systems.
Exceptionally, the experiments show that the overlap is less than 0.1 for a
small number of Hamiltonians. We alternatively utilise the Gibbs-VQE Wang _et
al._ (2021) method. In comparison to the shallow-depth VQE, Gibbs-VQE is less
prone to prepare state with small overlap and has a more steady performance.
More discussions can be found in Appendix. A. Furthermore, due to the
computational limitations, we are limited to observe the phenomenon up to the
14-qubit Heisenberg model. To further study the effectiveness of shallow-depth
VQE, the Ising model is considered with system sizes greater than 20. A
detailed discussion can be found in Appendix. B. Finally, barren plateaus
McClean _et al._ (2018) is a significant obstacle to the efficient use of
VQE. It refers to the gradient of the cost function vanishing exponentially
with the number of qubits for a randomly initialised PQC with sufficient
depth. We show that shallow-depth VQE will not be affected by barren plateaus
and is scalable to large-scale models, which can be referred to Appendix. C.
## IV Application to chemical molecules
Fig 8: The molecular geometry of the hydrogen chain $H_{4}$ with $4$ atoms
with equal spacing $R$. The right shows the qubit representation of the spin
orbitals, and the subscripts label the qubit indices in the quantum circuit.
The four spin-up orbitals are in the $|1\rangle$ state while no spin-down
orbitals are occupied, representing the HF initial state of $H_{4}$.
Previous results have demonstrated the effectiveness of our algorithms for
solving the Heisenberg model under intermediate-size situations and have the
potential for large-size applications. Nevertheless, the systematic analysis
of the ground state preparation performances with shallow-depth warm-start
stays ambiguous for other physical models. In this section, we aim to extend
the usage of our algorithm into the quantum chemical regime by numerically
demonstrating the warm-start performance on handling the small molecular model
while taking the well-known Hartree-Fock (HF) state as a standard.
To our best knowledge, both VQE and HF state methods own abilities to prepare
the high-accurate ground states Dong _et al._ (2021); Lin and Tong (2020);
Wecker _et al._ (2015) that can be used in quantum chemistry calculations for
determining molecular geometries, and ground energies Bartlett and Stanton
(1994); Peruzzo _et al._ (2014); Kandala _et al._ (2017). In the HF
approximation, electrons are treated independently and indistinguishably Omar
(2005) that interact with nuclei potential and an electronic mean field Seeger
and Pople (1977). Quantum chemistry assumes discrete molecular orbitals
determining the wave function of each electron appearing in the molecule.
General molecular Hamiltonian fermionic configuration can be represented in a
second quantisation form,
$H=\sum_{pq}h_{pq}c_{p}^{\dagger}c_{q}+\frac{1}{2}\sum_{pqrs}h_{pqrs}c_{p}^{\dagger}c_{q}^{\dagger}c_{r}c_{s},$
(19)
where $c_{p}^{\dagger},c_{p}$ are the creation and annihilation operators,
respectively. The coefficients $h_{pq}$ and $h_{pqrs}$ denote the one- and
two-electron Coulomb integrals Fermann and Valeev (2020). In the quantum
computing regime, the molecular electronic wave function can be represented in
the occupation number basis that each qubit defines a spin-orbital. By adding
one fermion (electron) into the system, each orbital can be in an occupied
state $|1\rangle$ or a non-occupied state $|0\rangle$ represented in the
computational basis. The single Slater-determinant state of $N$ orbitals can
be defined as,
$|\psi_{\theta}\rangle=U_{\theta}|\eta\rangle,\quad
U_{\theta}=\exp{\left(\sum_{j,k=1}^{N}\theta_{jk}c_{j}^{\dagger}c_{k}\right)}$
(20)
where $|\eta\rangle=c_{\eta}^{\dagger}\cdots c_{1}^{\dagger}|0\rangle$ in the
core orbital basis. The Hartree-Fock state is the Slater-determinant having
the lowest energy. Preparing the HF state classically requires a large amount
of computational resource Quantum _et al._ (2020). Even with quantum
technologies Tubman _et al._ (2019); Kivlichan _et al._ (2018), one have to
resolve the ansatz design for finding a basis to achieve relatively low
circuit complexity Babbush _et al._ (2018). Besides, the extreme number of
terms in the molecular Hamiltonian, writing in molecular orbital basis, could
raise scalable challenging to simulate and measure which requires
factorisation strategies Quantum _et al._ (2020); Berry _et al._ (2019).
(a)
(b)
Fig 9: Figures showing the ground state energy estimation and ground state
preparation via shallow-depth (depth-3) VQE (blue), deep-depth (depth-50) VQE
(green) and Gibbs-VQE (depth-3) (orange) on $H_{4}$ chain with bond distance
$R$ equals to $0.5$ (a) and $3.0$ (b). The above diagrams illustrate the
energy variations through the optimisation iterations, while the red lines
represent the exact values. The bottom shows the corresponding prepared state
overlaps, while the red shows the HF states’ overlaps.
We specifically consider the 1D hydrogen chain of $4$ atoms at fixed equal
spacing $R$ containing $4$ electrons. Each hydrogen atom has one molecular
orbital involving two spin orbitals that, in total, demands $8$ qubits for
VQE. We choose hydrogen chain $H_{4}$ as the objective since its simplicity
matches our current computational resource and its historical meaning as a
benchmark for quantum chemistry Motta _et al._ (2017). We use openfermion
built-in Psi4 kernel to compute the fermionic Hamiltonian components (19) of
$h_{pq}$ and $h_{pqrs}$ via self-consistent field method using the STO-3G
basis. We perform the simulations with $R=0.5$ and $3.0$ a.u. corresponding to
the strong and weak interaction regimes, respectively. Before running the VQE,
we need to map the fermionic Hamiltonian into the qubit representation via the
Jordan-Wigner transformation Seeley _et al._ (2012). During all simulations,
the VQE ansatz is chosen to be the HEA optimised in $200$ iterations of $0.1$
fixed learning rate by the Adam optimiser.
#### Comparing VQE and HF initial state on hydrogen chain
Results have been illustrated in Fig. 9. The top two diagrams illustrate the
energy curves while the bottom shows the corresponding output states’ overlap
regarding the exact ground states. The decreasing energy curves’ observation
in the left plot matches the growth in the optimised overlap for all the cases
of shallow-depth VQE (blue) and HF state (red) in (a) bottom. As the energy
converges to around $-1.65$ Ha in (a) top, both the two schemes reach an
overlap of almost one ($0.995$) in (a) bottom. Intuitively, bond distance
$R=0.5$ a.u. creates strong electron-electron interaction making the mean
field a reasonable approximation so that the HF state obtained a nearly
perfect fidelity. This also explains the reduced overlap from the HF results
reaching about $0.54$ when $R=3.0$ a.u. in (b) bottom since the approximation
deviates. Counterintuitively, the VQE results, in blue, experience a distinct
decline in the overlap though the corresponding energy values almost reach the
ground of $-1.868$ Ha of error $10^{-3}$ compared to the exact diagonalisation
results. However, increasing the depth to $50$ for VQE ansatz (green) could
gradually improve the overlap to $0.46$ within $200$ optimisation steps. We
also include the Gibbs-VQE (yellow) stated in the previous sections as a
reference scheme that showed similar behaviours from the shallow-depth VQE.
We conclude that the shallow-depth VQE may be a scalable and preferable warm-
starting strategy for molecular ground state preparation, showing great
performance of generating high-overlap initial states for short bond
distances. Interestingly, for larger bond distances, both methods imply
reduced initial overlap values. As the mean field decoupled, the accuracy of
the HF state decreases. Besides, the energy gaps between eigenstates shrunk
traps the VQE into the local minimum Wierichs _et al._ (2020) since the close
energy values carrying by the excited states could disturb the minimisation
processes leading to a zero overlap by the orthogonality. Deeper VQE circuit
could mitigate the effect but with a slow growth of the overlap. Therefore,
the ground energy estimation still performs well for governed by the
variational principle, however, might be insufficient for the efficient
initial state preparation. The fundamental pictures of the effect and the
coping strategies remain open for extending the usage of our method on
molecular models.
## V Application to Fermi-Hubbard models
Researching the Fermi-Hubbard model is one of the fundamental priorities in
condensed matter physics covering metal-insulator transitions and high-
temperature superconductivity Cade _et al._ (2020); Dagotto (1994). However,
the model involves a wide range of correlated electrons requiring multi-
determinant calculations and numerous computational resources, hence impeding
the classical exploration in the area. Therefore, designing the quantum method
for the model stays as a compelling application for quantum computing. The
potential of solving the 1D Hubbard model has been proposed by Dong _et al._
(2022) while only the VQE-typed algorithms have been examined via both
numerical simulations and real devices Cheuk _et al._ (2016); Wecker _et
al._ (2015); Lin and Tong (2020), as far as we know. In this section, we aim
to provide a first numerical demonstration for the quantum ground state
preparation scheme on the performance of predicting and estimating the
physical properties of a classically challenging Fermi-Hubbard model other
than commonly utilised VQE Stanisic _et al._ (2022); Quantum _et al._
(2020).
Fig 10: Qubit layout for implementing our $1\times 5$ Hubbard instances. The
lower illustrate the model’s physical site configuration where each site is
associated with an index $1\leq i\leq 5$ labelling from left to right. Every
site can hold a pair of spin-up and -down electrons which requires two qubits
to encode all possible spin orbitals. Each qubit has perfect connections with
the other qubits.
For an $n_{x}\times n_{y}$ square lattice physical system, for example, a
metallic crystal, each lattice point, called a site, is assigned with an
index. The Hubbard model Hamiltonian has a fermionic, second quantisation
form,
$H=-J\sum_{\langle
i,j\rangle,\sigma}(a^{\dagger}_{i\sigma}a_{j\sigma}+a^{\dagger}_{j\sigma}a_{i\sigma})+U\sum_{i}n_{i\uparrow}n_{i\downarrow}+H_{\text{local}},$
(21)
where $a^{\dagger}_{i\sigma},a_{i\sigma}$ are fermionic creation and
annihilation operators; $n_{i\sigma}=a^{\dagger}_{i\sigma},a_{i\sigma}$, are
the number operators; the notation $\langle i,j\rangle$ associates adjacent
sites in the $n_{x}\times n_{y}$ rectangular lattice;
$\sigma\in\\{\uparrow,\downarrow\\}$ labels the spin orbital. The first term
in Eq. (21) is the hopping term with $J$ being the tunnelling amplitude, and
$U$ in the second term is the on-site Coulomb repulsion. The last term defines
the local potential from nuclear-electron interaction. The local interaction
term is determined by the Gaussian form Wecker _et al._ (2015),
$H_{\text{local}}=\sum_{j=1}\sum_{\nu=\uparrow,\downarrow}\epsilon_{j,\nu}n_{j,\nu};\quad\epsilon_{j,\nu}=-\lambda_{\nu}e^{-\frac{1}{2}(j-m_{\nu})^{2}/\sigma_{\nu}^{2}}.$
(22)
We consider an $1\times 5$ lattice model with $J=2,U=3$ and
$\lambda_{\uparrow,\downarrow}=3,0.1$, $m_{\uparrow,\downarrow}=3,3$. The
standard deviation for both spin-up and -down potential is set to $1$. Such a
setup guarantees a charge-spin symmetry around the centre site ($i=3$) of the
entire system.
As before, representing the Hubbard Hamiltonian on a quantum computer in the
qubit representation requires fermionic encoding via efficient JW
transformation. For a 1D Hubbard model with $N$ sites, based on the Pauli
exclusion principle Deutsch (1951), each site can contain at most a pair of
spin-up and -down, two electrons. In total, there are $2N$ possible electronic
orbitals represented by the same number of qubits which covers all possible
orbitals for the system. The qubit layout of our five-qubit model has been
shown in Fig. 10 where all pairs of qubits are connected perfectly during
simulations. We study the medium-scaled model having non-degenerate ground
space. The fermionic Hamiltonian and qubitic transformations could be derived
using openfermion and paddle quantum python libraries. The VQE uses an HEA
circuit of depth $3$ and $50$ approaching realistic applications, and we
choose the Adam optimiser (fixed learning rate $0.1$) as before training the
circuit within $200$ iterations.
#### Evaluating physical quantities of Hubbard model
Fig 11: The (a) charge $\rho^{+}$ and (b) spin $\rho^{-}$ densities evaluated
from different produced ground states of our stated Hubbard model ($5$ sites)
by direct diagonalisation (green dashed), shallow-depth VQE (blue solid),
deep-depth VQE (orange solid) and our algorithm (red star) methods. The
shallow- and deep-depth VQE access HEA circuit with depth $3$ and $50$,
respectively. In both cases, the learning rate is set to $0.1$, and the total
number of iterations is set to $200$. The Adam optimiser powers all the
training procedures.
The first discussion is about predicting the charge-spin behaviours of the
given solid system. We study the global ground state of the Hubbard model
allowing the accessibility of every orbital with a sufficient number of
electrons in the qubit representation. After deriving the ground state of the
Hubbard Hamiltonian, we could compute the expectation value of different
physical observables, for example, charge and spin densities defined as,
$\rho^{\pm}_{j}=\langle n_{j\uparrow}\rangle\pm\langle
n_{j\downarrow}\rangle,$ (23)
where $j$ is the target site index. The densities could indicate the model’s
electronic and magnetic structures which further relate to the metal-metal
bonds of given materials Epsztein _et al._ (2018) and can be experimentally
demonstrated via electron paramagnetic resonance experiments Jacob and Reiher
(2012), respectively. The density diagrams illustrate charge-spin
distributions which could be used in designing solid structures of materials.
We compare the evaluated ground state densities from different preparation
schemes shown in Fig. 11. As we could see both the densities show a
centralized feature that matches our model configuration. The initial state of
our algorithm is prepared via shallow-depth VQE deriving the corresponding
density values (red stars) in the figure. A dramatic observation here shows
our scheme outperforms the traditional VQE methods, in which the derived
densities match the exact results while deep-depth VQE produces a visible gap
regarding the true values (green dashed). Notice that a deeper circuit
increases the VQE accuracy of both evaluated densities due to the better
expressibility. However, such a circuit approximates 2-design which
significantly drops the convergence speed of VQE. As a result, with the deep-
depth circuit, VQE could only prepare a defective ground state and predict
approximate densities (orange) within 200 iterations.
Besides, our algorithm only requires a non-zero-overlap initial state coming
from the shallow-depth VQE experiencing no trainability predicament. The
searched state should theoretically be the exact ground state and predict
exact density results. We show our method could derive high-precision charge-
spin densities which can be used to verify the theoretical developments on
Luttinger liquid and predict separation dynamics Arute _et al._ (2020).
Fig 12: Chemical potential $\mu$ calculations of Hubbard ground states from
different occupation numbers $N_{\text{occ}}$. The green dashed and blue solid
lines illustrate the exact and shallow-depth VQE-prepared ground states,
respectively. The stars label our algorithm’s results by initialising the
shallow-depth states. The subplot at the left corner indicates the
corresponding potential values’ absolute error $\varepsilon(\mu)$ (log-scale),
where the green shaded area marks the points satisfying chemical accuracy
regarding the FCI results.
Apart from that, chemical potential is another physical quantity determining
the energy variations during chemical and electronic interactions Stanisic
_et al._ (2022). In the second quantisation, the potential fields are
considered with discrete orbital states. Chemical potential describes the
energy requirement by adding or subtracting an electron from the system. Such
an electron would occupy an orbital state or leave a state empty,
respectively. The value can be computed via subtracting the ground state
energies of different occupation numbers (number of orbitals that are filled
with particles), having the following expression,
$\mu(N_{\text{occ}})=E(N_{\text{occ}})-E(N_{\text{occ}}-1).$ (24)
In the simulations, we derive the Hubbard Hamiltonian with different
occupation numbers by classically applying fermionic projection maps to the
previous global model by $\tilde{H}=P_{N_{\text{occ}}}HP_{N_{\text{occ}}}$.
The resulting $\tilde{H}$ could have image space spanning by the standard
basis elements with a fixed number of $|1\rangle$’s, and hence restrict the
number of electrons in the system. Such a projection can also be realised by
so-called the number-preserving ansatz introduced in Anselmetti _et al._
(2021). Previous literature works on the deep HVA-based ansatzes ($30$ layers)
and directly prepare the Hubbard ground states using VQE Stanisic _et al._
(2022). We instead, run the shallow-depth VQE on each $\tilde{H}$ to prepare
an initial state followed by our searching algorithm to locate the ground
states of each occupation number with constant scaled phase factor. The
corresponding energies can be simultaneously estimated from our algorithm.
As from Fig. 12, the algorithm could successfully extract the ground state
from initial shallow-depth VQE’s states. By zooming into the subplot
illustrating the absolute error of the chemical potentials at each
$N_{\text{occ}}$. Our estimated values obtain an absolute error of almost
$10^{-5}$ which have all reached the chemical precision with respect to the
FCI calculations via noiseless simulations in the green-shaded area. The
observation significantly proves the effectiveness of our method on other
high-accurate condensed matter calculations which inspires that our scheme
could promote a variety of physical research and provide a new paradigm for
studying and predicting hard-to-solve model behaviours from both classical and
quantum horizons.
## VI Conclusion and outlook
In this paper, we provide quantum algorithms to prepare ground states and
demonstrate the correctness and effectiveness through extensive numerical
simulations on Heisenberg models and chemical molecules. We also apply our
algorithms to estimate physical quantities of interest, such as charge and
spin densities and chemical potential of Fermi-Hubbard models having no
generally efficient classical methods, to our best knowledge. Only VQE and its
variants have been applied and demonstrated with numerical and practical
experiments from recent literature. A notable property of our algorithms is
that the ground state can be prepared simultaneously with the ground state
energy estimation, thus requiring no prior information on the ground state
energy. In contrast, existing ground state preparation algorithms involving
quantum phase estimation have to estimate a proper ground state energy
beforehand Ge _et al._ (2019), leading to indefinite practicability. For
other works Lin and Tong (2020); Dong _et al._ (2022) based on QSVT and QET-U
algorithms, a reasonable upper bound estimate of the ground state energy is
pre-requested. Since such information is usually not known as priorities, one
has to expensively estimate the ground state energy before preparing the
state. Another worth-noting point is that our algorithms only require to
compute phase factors with a scaling $\widetilde{\mathcal{O}}(1)$ while QSVT
and QET-U based algorithms require $\widetilde{\mathcal{O}}(\Delta^{-1})$,
which means our algorithms have advantages in avoiding challenging phase
factor evaluations when $\Delta$ is small.
Besides algorithmic improvements, we also contribute to showing that shallow-
depth VQE can be a favourable method for warm-starting the ground state
preparation of the intermediate-scaled many-body systems. In contrast, the
previous works Ge _et al._ (2019); Dong _et al._ (2022); Lin and Tong
(2020); Dong _et al._ (2022) all assume the state with a considerable overlap
is available and can be reused. Our scheme provides a potential paradigm for
using shallow-depth VQE to solve challenging physical models. By conducting
simulations to a 10-qubit Hubbard model, our method outperforms the VQE-based
evaluations of the charge and spin densities and the chemical potentials.
Meanwhile, we efficiently and qualitatively predict the ground state density
behaviours of the potential centralised model and derive extremely accurate
chemical potentials of distinct occupations up to an absolute error of
$10^{-5}$. Such evaluations have reached the chemical accuracy of the FCI
results, which is foreseeable to own advantages to exploring high-temperature
superconductivity and quantum chemistry problems. Our simulations have
demonstrated the potential of our algorithm on solving the 1D Hubbard model
and prepared for further research on the classical challenging 2D models Cheuk
_et al._ (2016).
Ground state preparation is significant to the fundamental research in physics
and chemistry. To establish a synthetic quantum solution for this task, there
are still some remaining problems to resolve in future. Since our works can
not offer suggestions on the depth of the warm-start ansatz, a theoretical
guarantee of reaching the large initial state overlap using sufficient-depth
ansatzes stays required. Recall that the algorithms are based on the prior
information on the spectral gap, especially, when the spectral gap is small,
it would be better to amplify it economically. Hence, an efficient and
effective method to compute the spectral gap and even amplify the gap is
necessary for applying quantum computers in solving ground-state problems. In
addition to addressing many-body systems, we believe our techniques can also
be used for combinatorial optimization problems e.g. the travelling
salesperson problem. Based on the results in this paper, we believe that
shallow variational warm-start would find more applications in physics,
chemistry, and quantum machine learning, which is worth further studying in
theory and experiments.
## Acknowledgement
Y. W., C. Z., and M. J. contributed equally to this work. Part of this work
was done when Y. W., C. Z., and M. J. were research interns at Baidu Research.
Y. W. acknowledges support from the Baidu-UTS AI Meets Quantum project. The
authors would like to thank Lloyd C.L. Hollenberg and Haokai Zhang for the
valuable comments and Lei Zhang for the helpful discussion.
## References
* Feynman and Cohen (1956) R. P. Feynman and Michael Cohen, “Energy spectrum of the excitations in liquid helium,” Phys. Rev. 102, 1189–1204 (1956).
* Sakurai and Commins (1995) J. J. Sakurai and Eugene D. Commins, “Modern quantum mechanics, revised edition,” American Journal of Physics 63, 93–95 (1995), https://doi.org/10.1119/1.17781 .
* Sherrill (2009) C. David Sherrill, “An introduction to hartree-fock molecular orbital theory,” (2009).
* Knowles and Handy (1989) Peter J. Knowles and Nicholas C. Handy, “A determinant based full configuration interaction program,” Computer Physics Communications 54, 75–83 (1989).
* Parr (1980) Robert G Parr, “Density functional theory of atoms and molecules,” in _Horizons of Quantum Chemistry: Proceedings of the Third International Congress of Quantum Chemistry Held at Kyoto, Japan, October 29-November 3, 1979_ (Springer, 1980) pp. 5–15.
* Blankenbecler _et al._ (1981) Richard Blankenbecler, DJ Scalapino, and RL Sugar, “Monte carlo calculations of coupled boson-fermion systems. i,” Physical Review D 24, 2278 (1981).
* Cade _et al._ (2020) Chris Cade, Lana Mineh, Ashley Montanaro, and Stasja Stanisic, “Strategies for solving the fermi-hubbard model on near-term quantum computers,” Physical Review B 102 (2020), 10.1103/physrevb.102.235122.
* Dagotto (1994) Elbio Dagotto, “Correlated electrons in high-temperature superconductors,” Rev. Mod. Phys. 66, 763–840 (1994).
* Nelson _et al._ (2019) James Nelson, Rajarshi Tiwari, and Stefano Sanvito, “Machine learning density functional theory for the hubbard model,” Physical Review B 99 (2019), 10.1103/physrevb.99.075132.
* Barahona (1982) Francisco Barahona, “On the computational complexity of ising spin glass models,” Journal of Physics A: Mathematical and General 15, 3241 (1982).
* Gilks _et al._ (1995) W. R. Gilks, N. G. Best, and K. K. C. Tan, “Adaptive rejection metropolis sampling within gibbs sampling,” Journal of the Royal Statistical Society: Series C (Applied Statistics) 44, 455–472 (1995), https://rss.onlinelibrary.wiley.com/doi/pdf/10.2307/2986138 .
* Del Moral _et al._ (2006) Pierre Del Moral, Arnaud Doucet, and Ajay Jasra, “Sequential monte carlo samplers,” Journal of the Royal Statistical Society: Series B (Statistical Methodology) 68, 411–436 (2006).
* Scalapino (2006) Douglas J. Scalapino, “Numerical studies of the 2d hubbard model,” arXiv: Strongly Correlated Electrons , 495–526 (2006).
* LeBlanc _et al._ (2015) J. P. F. LeBlanc, Andrey E. Antipov, Federico Becca, Ireneusz W. Bulik, Garnet Kin-Lic Chan, Chia-Min Chung, Youjin Deng, Michel Ferrero, Thomas M. Henderson, Carlos A. Jiménez-Hoyos, E. Kozik, Xuan-Wen Liu, Andrew J. Millis, N. V. Prokof’ev, Mingpu Qin, Gustavo E. Scuseria, Hao Shi, B. V. Svistunov, Luca F. Tocchio, I. S. Tupitsyn, Steven R. White, Shiwei Zhang, Bo-Xiao Zheng, Zhenyue Zhu, and Emanuel Gull (Simons Collaboration on the Many-Electron Problem), “Solutions of the two-dimensional hubbard model: Benchmarks and results from a wide range of numerical algorithms,” Phys. Rev. X 5, 041041 (2015).
* Feynman (1982) Richard P. Feynman, “Simulating physics with computers,” International Journal of Theoretical Physics 21, 467–488 (1982).
* Peruzzo _et al._ (2014) Alberto Peruzzo, Jarrod McClean, Peter Shadbolt, Man-Hong Yung, Xiao-Qi Zhou, Peter J. Love, Alán Aspuru-Guzik, and Jeremy L. O’Brien, “A variational eigenvalue solver on a photonic quantum processor,” Nature Communications 5 (2014), 10.1038/ncomms5213.
* Wecker _et al._ (2015) Dave Wecker, Matthew B. Hastings, and Matthias Troyer, “Progress towards practical quantum variational algorithms,” Physical Review A 92 (2015), 10.1103/physreva.92.042303.
* Kandala _et al._ (2017) Abhinav Kandala, Antonio Mezzacapo, Kristan Temme, Maika Takita, Markus Brink, Jerry M. Chow, and Jay M. Gambetta, “Hardware-efficient variational quantum eigensolver for small molecules and quantum magnets,” Nature 549, 242–246 (2017).
* Sapova and Fedorov (2022) Mariia D. Sapova and Aleksey K. Fedorov, “Variational quantum eigensolver techniques for simulating carbon monoxide oxidation,” Communications Physics 5, 199 (2022).
* Ostaszewski _et al._ (2021) Mateusz Ostaszewski, Lea M. Trenkwalder, Wojciech Masarczyk, Eleanor Scerri, and Vedran Dunjko, “Reinforcement learning for optimization of variational quantum circuit architectures,” in _Advances in Neural Information Processing Systems_, Vol. 34, edited by M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (Curran Associates, Inc., 2021) pp. 18182–18194.
* Dallaire-Demers _et al._ (2019) Pierre-Luc Dallaire-Demers, Jonathan Romero, Libor Veis, Sukin Sim, and Alán Aspuru-Guzik, “Low-depth circuit ansatz for preparing correlated fermionic states on a quantum computer,” Quantum Science and Technology 4, 045005 (2019).
* Johnson _et al._ (2011) M. W. Johnson, M. H. S. Amin, S. Gildert, T. Lanting, F. Hamze, N. Dickson, R. Harris, A. J. Berkley, J. Johansson, P. Bunyk, E. M. Chapple, C. Enderud, J. P. Hilton, K. Karimi, E. Ladizinsky, N. Ladizinsky, T. Oh, I. Perminov, C. Rich, M. C. Thom, E. Tolkacheva, C. J. S. Truncik, S. Uchaikin, J. Wang, B. Wilson, and G. Rose, “Quantum annealing with manufactured spins,” Nature 473, 194–198 (2011).
* Kitaev (1995) A. Yu. Kitaev, “Quantum measurements and the abelian stabilizer problem,” (1995).
* Brassard _et al._ (2002) Gilles Brassard, Peter Hoyer, Michele Mosca, and Alain Tapp, “Quantum amplitude amplification and estimation,” Contemporary Mathematics 305, 53–74 (2002).
* Ge _et al._ (2019) Yimin Ge, Jordi Tura, and J Ignacio Cirac, “Faster ground state preparation and high-precision ground energy estimation with fewer qubits,” Journal of Mathematical Physics 60, 022202 (2019).
* Poulin and Wocjan (2009) David Poulin and Pawel Wocjan, “Preparing ground states of quantum many-body systems on a quantum computer,” Physical Review Letters 102 (2009), 10.1103/physrevlett.102.130503.
* Gilyén _et al._ (2019) András Gilyén, Yuan Su, Guang Hao Low, and Nathan Wiebe, “Quantum singular value transformation and beyond: exponential improvements for quantum matrix arithmetics,” in _Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing_ (ACM, 2019).
* Lin and Tong (2020) Lin Lin and Yu Tong, “Near-optimal ground state preparation,” Quantum 4, 372 (2020).
* Berry _et al._ (2015) Dominic W. Berry, Andrew M. Childs, and Robin Kothari, “Hamiltonian simulation with nearly optimal dependence on all parameters,” in _2015 IEEE 56th Annual Symposium on Foundations of Computer Science_ (IEEE, 2015).
* Dong _et al._ (2022) Yulong Dong, Lin Lin, and Yu Tong, “Ground-state preparation and energy estimation on early fault-tolerant quantum computers via quantum eigenvalue transformation of unitary matrices,” PRX Quantum 3 (2022), 10.1103/prxquantum.3.040305.
* Low and Chuang (2017a) Guang Hao Low and Isaac L. Chuang, “Optimal hamiltonian simulation by quantum signal processing,” Physical Review Letters 118 (2017a), 10.1103/physrevlett.118.010501.
* Abrams and Lloyd (1999) Daniel S. Abrams and Seth Lloyd, “Quantum algorithm providing exponential speed increase for finding eigenvalues and eigenvectors,” Physical Review Letters 83, 5162–5165 (1999).
* Higgins _et al._ (2007) B. L. Higgins, D. W. Berry, S. D. Bartlett, H. M. Wiseman, and G. J. Pryde, “Entanglement-free heisenberg-limited phase estimation,” Nature 450, 393–396 (2007).
* Berry _et al._ (2009) D. W. Berry, B. L. Higgins, S. D. Bartlett, M. W. Mitchell, G. J. Pryde, and H. M. Wiseman, “How to perform the most accurate possible phase measurements,” Physical Review A 80 (2009), 10.1103/physreva.80.052114.
* Knill _et al._ (2007) Emanuel Knill, Gerardo Ortiz, and Rolando D. Somma, “Optimal quantum measurements of expectation values of observables,” Phys. Rev. A 75, 012328 (2007).
* Lin and Tong (2022) Lin Lin and Yu Tong, “Heisenberg-limited ground-state energy estimation for early fault-tolerant quantum computers,” PRX Quantum 3 (2022), 10.1103/prxquantum.3.010318.
* Ding and Lin (2022) Zhiyan Ding and Lin Lin, “Even shorter quantum circuit for phase estimation on early fault-tolerant quantum computers with applications to ground-state energy estimation,” (2022).
* O’Brien _et al._ (2019) Thomas E O’Brien, Brian Tarasinski, and Barbara M Terhal, “Quantum phase estimation of multiple eigenvalues for small-scale (noisy) experiments,” New Journal of Physics 21, 023022 (2019).
* Wang _et al._ (2019) Daochen Wang, Oscar Higgott, and Stephen Brierley, “Accelerated variational quantum eigensolver,” Phys. Rev. Lett. 122, 140504 (2019).
* Somma (2019) Rolando D Somma, “Quantum eigenvalue estimation via time series analysis,” New Journal of Physics 21, 123025 (2019).
* Wan _et al._ (2022) Kianna Wan, Mario Berta, and Earl T. Campbell, “Randomized quantum algorithm for statistical phase estimation,” Physical Review Letters 129 (2022), 10.1103/physrevlett.129.030503.
* Wang _et al._ (2022a) Guoming Wang, Daniel Stilck-Franca, Ruizhe Zhang, Shuchen Zhu, and Peter D Johnson, “Quantum algorithm for ground state energy estimation using circuit depth with exponentially improved dependence on precision,” arXiv preprint arXiv:2209.06811 (2022a).
* Wang _et al._ (2022b) Xin Wang, Youle Wang, Zhan Yu, and Lei Zhang, “Quantum phase processing: Transform and extract eigen-information of quantum systems,” arXiv preprint arXiv:2209.14278 (2022b).
* Childs _et al._ (2018) Andrew M Childs, Dmitri Maslov, Yunseong Nam, Neil J Ross, and Yuan Su, “Toward the first quantum simulation with quantum speedup,” Proceedings of the National Academy of Sciences 115, 9456–9461 (2018).
* hub (2013) “The hubbard model at half a century,” Nature Physics , 523 (2013).
* Wang _et al._ (2021) Youle Wang, Guangxi Li, and Xin Wang, “Variational quantum gibbs state preparation with a truncated taylor series,” Physical Review Applied 16 (2021), 10.1103/physrevapplied.16.054035.
* Low and Chuang (2019) Guang Hao Low and Isaac L. Chuang, “Hamiltonian simulation by qubitization,” Quantum 3, 163 (2019).
* Dong _et al._ (2021) Yulong Dong, Xiang Meng, K. Birgitta Whaley, and Lin Lin, “Efficient phase-factor evaluation in quantum signal processing,” Physical Review A 103 (2021), 10.1103/physreva.103.042419.
* Low and Chuang (2017b) Guang Hao Low and Isaac L Chuang, “Hamiltonian simulation by uniform spectral amplification,” arXiv preprint arXiv:1707.05391 (2017b).
* Yan _et al._ (2011) Simeng Yan, David A. Huse, and Steven R. White, “Spin-liquid ground state of the $\langle s\rangle$ = 1/2 kagome heisenberg antiferromagnet,” Science 332, 1173–1176 (2011).
* Bonechi _et al._ (1992) F Bonechi, E Celeghini, R Giachetti, E Sorace, and M Tarlini, “Heisenberg XXZ model and quantum galilei group,” Journal of Physics A: Mathematical and General 25, L939–L943 (1992).
* Faddeev (1996) LD Faddeev, “How algebraic bethe ansatz works for integrable model,” arXiv preprint hep-th/9605187 (1996).
* Rojas _et al._ (2001) Onofre Rojas, S. M. de Souza, E. V. Corrêa Silva, and M. T. Thomaz, “Thermodynamics of the Limiting Cases of the XXZ Model Without Bethe Ansatz,” Brazilian Journal of Physics 31, 577–582 (2001).
* Cerezo _et al._ (2021) M. Cerezo, Akira Sone, Tyler Volkoff, Lukasz Cincio, and Patrick J. Coles, “Cost function dependent barren plateaus in shallow parametrized quantum circuits,” Nature Communications 12 (2021), 10.1038/s41467-021-21728-w.
* McClean _et al._ (2018) Jarrod R. McClean, Sergio Boixo, Vadim N. Smelyanskiy, Ryan Babbush, and Hartmut Neven, “Barren plateaus in quantum neural network training landscapes,” Nature Communications 9 (2018), 10.1038/s41467-018-07090-4.
* Bartlett and Stanton (1994) Rodney J Bartlett and John F Stanton, “Applications of post-hartree—fock methods: A tutorial,” Reviews in computational chemistry , 65–169 (1994).
* Omar (2005) Yasser Omar, “Indistinguishable particles in quantum mechanics: an introduction,” Contemporary Physics 46, 437–448 (2005).
* Seeger and Pople (1977) Rolf Seeger and John A. Pople, “Self‐consistent molecular orbital methods. xviii. constraints and stability in hartree–fock theory,” The Journal of Chemical Physics 66, 3045–3050 (1977), https://doi.org/10.1063/1.434318 .
* Fermann and Valeev (2020) Justin T. Fermann and Edward F. Valeev, “Fundamentals of molecular integrals evaluation,” (2020).
* Quantum _et al._ (2020) Google AI Quantum, Collaborators*†, Frank Arute, Kunal Arya, Ryan Babbush, Dave Bacon, Joseph C Bardin, Rami Barends, Sergio Boixo, Michael Broughton, Bob B Buckley, _et al._ , “Hartree-fock on a superconducting qubit quantum computer,” Science 369, 1084–1089 (2020).
* Tubman _et al._ (2019) Norm Tubman, Carlos Mejuto Zaera, Jeffrey Epstein, Diptarka Hait, Daniel Levine, William Huggins, Zhang Jiang, Jarrod McClean, Ryan Babbush, Martin Head-Gordon, and Birgitta Whaley, “Postponing the orthogonality catastrophe: efficient state preparation for electronic structure simulations on quantum devices,” in _APS March Meeting Abstracts_ , APS Meeting Abstracts, Vol. 2019 (2019) p. E42.004.
* Kivlichan _et al._ (2018) Ian D. Kivlichan, Jarrod McClean, Nathan Wiebe, Craig Gidney, Alán Aspuru-Guzik, Garnet Kin-Lic Chan, and Ryan Babbush, “Quantum simulation of electronic structure with linear depth and connectivity,” Phys. Rev. Lett. 120, 110501 (2018).
* Babbush _et al._ (2018) Ryan Babbush, Nathan Wiebe, Jarrod McClean, James McClain, Hartmut Neven, and Garnet Kin-Lic Chan, “Low-depth quantum simulation of materials,” Phys. Rev. X 8, 011044 (2018).
* Berry _et al._ (2019) Dominic W. Berry, Craig Gidney, Mario Motta, Jarrod R. McClean, and Ryan Babbush, “Qubitization of Arbitrary Basis Quantum Chemistry Leveraging Sparsity and Low Rank Factorization,” Quantum 3, 208 (2019).
* Motta _et al._ (2017) Mario Motta, David M. Ceperley, Garnet Kin-Lic Chan, John A. Gomez, Emanuel Gull, Sheng Guo, Carlos A. Jiménez-Hoyos, Tran Nguyen Lan, Jia Li, Fengjie Ma, Andrew J. Millis, Nikolay V. Prokof’ev, Ushnish Ray, Gustavo E. Scuseria, Sandro Sorella, Edwin M. Stoudenmire, Qiming Sun, Igor S. Tupitsyn, Steven R. White, Dominika Zgid, and Shiwei Zhang (Simons Collaboration on the Many-Electron Problem), “Towards the solution of the many-electron problem in real materials: Equation of state of the hydrogen chain with state-of-the-art many-body methods,” Phys. Rev. X 7, 031059 (2017).
* Seeley _et al._ (2012) Jacob T. Seeley, Martin J. Richard, and Peter J. Love, “The bravyi-kitaev transformation for quantum computation of electronic structure,” The Journal of Chemical Physics 137, 224109 (2012), https://doi.org/10.1063/1.4768229 .
* Wierichs _et al._ (2020) David Wierichs, Christian Gogolin, and Michael Kastoryano, “Avoiding local minima in variational quantum eigensolvers with the natural gradient optimizer,” Physical Review Research 2 (2020), 10.1103/physrevresearch.2.043246.
* Cheuk _et al._ (2016) Lawrence W Cheuk, Matthew A Nichols, Katherine R Lawrence, Melih Okan, Hao Zhang, Ehsan Khatami, Nandini Trivedi, Thereza Paiva, Marcos Rigol, and Martin W Zwierlein, “Observation of spatial charge and spin correlations in the 2d fermi-hubbard model,” Science 353, 1260–1264 (2016).
* Stanisic _et al._ (2022) Stasja Stanisic, Jan Lukas Bosse, Filippo Maria Gambetta, Raul A. Santos, Wojciech Mruczkiewicz, Thomas E. O’Brien, Eric Ostby, and Ashley Montanaro, “Observing ground-state properties of the Fermi-Hubbard model using a scalable algorithm on a quantum computer,” Nature Commun. 13, 5743 (2022), arXiv:2112.02025 [quant-ph] .
* Deutsch (1951) Martin Deutsch, “Introductory nuclear physics.” Science 113, 443–443 (1951).
* Epsztein _et al._ (2018) Razi Epsztein, Evyatar Shaulsky, Nadir Dizge, David M. Warsinger, and Menachem Elimelech, “Role of ionic charge density in donnan exclusion of monovalent anions by nanofiltration,” Environmental Science & Technology 52, 4108–4116 (2018), pMID: 29510032, https://doi.org/10.1021/acs.est.7b06400 .
* Jacob and Reiher (2012) Christoph R. Jacob and Markus Reiher, “Spin in density-functional theory,” International Journal of Quantum Chemistry 112, 3661–3684 (2012).
* Arute _et al._ (2020) Frank Arute, Kunal Arya, Ryan Babbush, Dave Bacon, Joseph C Bardin, Rami Barends, Andreas Bengtsson, Sergio Boixo, Michael Broughton, Bob B Buckley, _et al._ , “Observation of separated dynamics of charge and spin in the fermi-hubbard model,” arXiv preprint arXiv:2010.07965 (2020).
* Anselmetti _et al._ (2021) Gian-Luca R Anselmetti, David Wierichs, Christian Gogolin, and Robert M Parrish, “Local, expressive, quantum-number-preserving vqe ansätze for fermionic systems,” New Journal of Physics 23, 113010 (2021).
* Egger _et al._ (2021) Daniel J. Egger, Jakub Mareč ek, and Stefan Woerner, “Warm-starting quantum optimization,” Quantum 5, 479 (2021).
* Nakaji and Yamamoto (2021) Kouhei Nakaji and Naoki Yamamoto, “Expressibility of the alternating layered ansatz for quantum computation,” Quantum 5, 434 (2021).
Supplementary Material for
Ground state preparation with shallow variational warm-start
## Appendix A Initial State Preparation with Gibbs-VQE
The experiments in Fig. 7 show that it is unsuccessful for a small number of
Hamiltonians. This may be caused by the local minima problem, which we re-
initialise all the circuit parameters and re-run the circuit again may
significantly amplify the overlap. To mitigate this issue, we consider
applying variational Gibbs-state preparationWang _et al._ (2021) (Gibbs-VQE)
alternatively.
We begin by briefly reviewing the background of the Gibbs state and the
protocol Gibbs-VQE. For a Hamiltonian with an $n$-qubit system, the Gibbs
state $\rho_{G}$ at temperature $T$ is defined as follows:
$\rho_{G}=\frac{e^{-\beta H}}{\operatorname{tr}(e^{-\beta H})},$ (A.1)
where $-\beta H$ is the matrix exponential of matrix $-\beta H$.
$\beta=\frac{1}{kT}$ is the inverse temperature of the system and $k$ is the
Boltzmann’s constant. To prepare a Gibbs state we use the protocol as
introduced in Wang _et al._ (2021). The main idea is to consider the
variational principle of the system’s free energy, where parameterised quantum
circuits are used to minimize the free energy. Specifically, we set the loss
function as $L=L_{1}+L_{2}+L_{3}$, where
$L_{1}=\operatorname{tr}{H_{\rho_{B}}},\quad
L_{2}=2\beta^{-1}\operatorname{tr}{\rho_{B}^{2}},\quad
L_{3}=\frac{-\beta^{-1}(\operatorname{tr}{\rho_{B}^{2}}+3)}{2}.$ (A.2)
Fig S1: The optimised overlap using shallow-depth VQE and shallow-depth Gibbs-
VQE. For each sampled Hamiltonian, 10 optimisations are performed to determine
the overlap between the optimised and ground states. The blue circle indicates
the overlap prepared with shallow-depth VQE and the red diamond represents the
case with the Gibbs-VQE.
To compare the performance of Gibbs-VQE and shallow-depth VQE, 200 random
Hamiltonians are sampled and 10 random initialisations of circuit parameters
are performed for each Hamiltonian. For both simulations of two methods, we
employ depth-3 HEA ansatz and 200 Adam optimiser optimisation iterations. The
experiment results are shown in Fig. S1. The blue circle indicates the overlap
prepared by shallow-depth VQE, and the red diamond indicates the overlap
prepared by shallow-depth Gibbs-VQE. Simulation results indicate that Gibbs-
VQE has a more steady performance than shallow-depth VQE and is less prone to
become stuck in local minima. It can be seen that $\sim 11\%$ number of
shallow-VQE prepare the overlap less than 0.4, whereas, for Gibbs-VQE, the
ratio reduces to $\sim 2\%$. Thus, Gibbs-VQE is superior for solving physical
and chemical problems because it generates initial states with reliable
overlap.
## Appendix B Initial State Preparation with Tensor Product Ansatz
The process that initialises an approximating state for a quantum algorithm
via classical methods is usually called a warm-start quantum method. By
inheriting the initial state, quantum algorithm could require fewer resources
for solving the problem. Famous result, for example, warm-start QAOA Egger
_et al._ (2021) combining Goemans-Williamson (GW) randomized rounding
algorithm, guarantees the performance of solving the max-cut problem with
${\cal O}(n)$ qubits and low-depth quantum circuits.
From the previous section III, we observe the efficiency of initial state
preparation on Heisenberg Hamiltonians through shallow-depth VQE. Due to the
computational limitations, we are limited to observe the phenomenon on
Hamiltonians up to $14$-qubit. To further investigate the efficacy of shallow-
depth VQE, we use the tensor product ansatz to examine the max-cut problem.
Specifically, the max-cut problem can be naturally mapped to the QUBO
Hamiltonian. A general QUBO Hamiltonian has an expression,
$H_{\text{QUBO}}=\frac{1}{2}\sum_{p\neq
q}Z_{p}Z_{q}-\frac{n}{2}\mathbb{I}^{\otimes n}.$ (B.1)
This is closely related to the continuous relaxation of the max-cut problem
which can be approximately solved using semidefinite programming and GW
rounding. Such a technique generates the initial state for the problem
Hamiltonian, which can be efficiently prepared by applying a single layer of
$R_{Y}$ rotations. This indicates the possibility of the approximating ground
state of specific Hamiltonian models via shallow-depth quantum circuits.
Taking the max-cut problem as a starting point, we could have the following
observation.
###### Lemma S1
For a product state ansatz composed with only one layer of single-qubit
$R_{y}$ rotation, s.t., $|\psi(\theta)\rangle=\bigotimes
R_{y}(\theta_{p})|0\rangle$, we have
$\langle\psi(\theta)|H|\psi(\theta)\rangle=\frac{1}{2}\sum_{p\neq
q}\langle\psi(\theta)|Z_{p}Z_{q}|\psi(\theta)\rangle-\frac{n}{2}\langle\psi(\theta)|\mathbb{I}^{\otimes
n}|\psi(\theta)\rangle=\frac{1}{2}\sum_{p\neq
q}\cos{(\theta_{p})}\cos{(\theta_{q})}-\frac{n}{2}.$ (B.2)
As from above, we could work out the analytic gradient for any k-local
Hamiltonian with the product state ansatz $|\psi(\theta)\rangle$. For QUBO
case, setting cost
$C(\bm{\theta})=\langle\psi(\theta)|H_{\text{QUBO}}|\psi(\theta)\rangle$, we
could easily derive,
$\partial_{\alpha}C=-\sin{(\theta_{\alpha})}\sum_{p\neq\alpha}\cos{(\theta_{p})}$
We then use QUBO Hamiltonian and the result in Lemma S1 to further investigate
the performance of shallow-depth VQE. With powerful classical machine learning
techniques, we could classically derive the optimal rotation angles via
minimizing the cost (B.2). The preparation of the approximating initial state
is then implemented by applying the corresponding rotations on $|0\rangle$. In
this experiment, we investigate QUBO Hamiltonians with a size of 15 to 22
qubits and calculate the overlap using a brute-force technique. The experiment
results are shown in Fig. S2.
Fig S2: The highest overlap between the optimised state and one of the ground
states of different given QUBO Hamiltonian models (each model probably has
degenerate ground state space) vs. the system sizes. For the number of qubits
between 15 to 19 and 20 to 22, we sample $500$ and $100$ random max-cut
graphs’ qubo Hamiltonians respectively. The true ground states of each model
were first determined using brute force algorithm and stored in classical
memory. Each model will then be processed in VQE with the tensor product
ansatz and the optimisation was done w.r.t the analytic form Eq. (B.2) to
produce an optimised state.
As we seen from the diagram, the x-axis represents the size of QUBO
Hamiltonian and y-axis is the overlap between the initial state prepared by
analytic form according to E.q. B.2 and the ground states. The probability of
getting high overlap states regarding one of the ground states is slightly
higher than trapping into the excited states (local minima). The experiment
trend suggests that shallow-depth VQE may continue to perform well on large-
scale Hamiltonians.
## Appendix C Numerical Simulations on the Gradient of Shallow-Depth Circuit
The scalability is a significant obstacle to the efficient use of variational
quantum eigensolver. Numerous studies have demonstrated that the gradient of
the cost function vanishes exponentially with the number of qubits for a
randomly initialised PQC with sufficient depth McClean _et al._ (2018). This
phenomenon is termed the barren plateaus (BP). It has been shown that hardware
efficient ansatz with deep depth leads to exponentially vanishing gradients
McClean _et al._ (2018). While Alternating Layered Ansatz, a specific
structure of the HEA ansatz Cerezo _et al._ (2021), has been demonstrated
that when circuit depth is shallow, it does not exhibit the gradient vanishing
problem Nakaji and Yamamoto (2021). As we explore using the HEA and ALT ansatz
as an overlap initialiser, it is crucial to determine if these two ansatz with
shallow depth suffer from the gradient vanishing issue.
As shallow-depth ALT ansatz is BP-free, we choose to use the depth-3 HEA
ansatz as an illustration and conduct numerical experiments to sample the
variance of the gradient. The experiment results are shown in Fig. S3. Similar
to the experimental setting described in McClean _et al._ (2018), we analyse
the Pauli $ZZ$ operator acting on the first and second qubits such that
$H=Z_{1}Z_{2}$. The gradient is then sampled on the first parameter
$\theta_{1,1}$ by taking 1000 samples and the system size is evaluated up to
18 qubits.
Fig S3: Semi-log plot displaying the sample variance of the gradient of a two-
local Pauli term as a function of the number of qubits. The blue line
represents exponential decay as a function of the number of qubits, the dark
blue line represents the logarithmic decay and the orange line indicates that
the shallow-depth VQE do not exhibit exponential decay.
As shown in the figure, the orange line indicates the sample variance of the
gradient of the shallow-depth VQE while the blue and dark blue lines depict
exponential and logarithmic decay, respectively. Additionally, we identify the
blue area below the exponential decay line as the BP area and the purple area
above the logarithmic decay as BP-free. It is evident that the depth-3 HEA
ansatz does not suffer from the gradient vanishing problem. Therefore, it is
scalable to use shallow-depth VQE for initial state preparation.
|
# ERASING THE VARIABLE: EMPIRICAL FOREGROUND DISCOVERY FOR GLOBAL $21$ cm
SPECTRUM EXPERIMENTS
Eric R. Switzer11affiliation: NASA Goddard Space Flight Center, Greenbelt, MD
20771, USA 22affiliation: Canadian Institute for Theoretical Astrophysics,
University of Toronto, 60 St. George St., Toronto, ON, M5S 3H8, Canada and
Adrian Liu33affiliation: Department of Astronomy, UC Berkeley, Berkeley, CA
94720, USA 44affiliation: Berkeley Center for Cosmological Physics, UC
Berkeley, Berkeley, CA 94720, USA<EMAIL_ADDRESS>
###### Abstract
Spectral measurements of the $21$ cm monopole background have the promise of
revealing the bulk energetic properties and ionization state of our universe
from $z\sim 6-30$. Synchrotron foregrounds are orders of magnitude larger than
the cosmological signal, and are the principal challenge faced by these
experiments. While synchrotron radiation is thought to be spectrally smooth
and described by relatively few degrees of freedom, the instrumental response
to bright foregrounds may be much more complex. To deal with such
complexities, we develop an approach that discovers contaminated spectral
modes using spatial fluctuations of the measured data. This approach exploits
the fact that foregrounds vary across the sky while the signal does not. The
discovered modes are projected out of each line of sight of a data cube. An
angular weighting then optimizes the cosmological signal amplitude estimate by
giving preference to lower-noise regions. Using this method, we show that it
is essential for the passband to be stable to at least $\sim 10^{-4}$. In
contrast, the constraints on the spectral smoothness of the absolute
calibration are mainly aesthetic if one is able to take advantage of spatial
information. To the extent it is understood, controlling polarization to
intensity leakage at the $\sim 10^{-2}$ level will also be essential to
rejecting Faraday rotation of the polarized synchrotron emission.
###### Subject headings:
dark ages, reionization, first stars – methods: data analysis – methods:
statistical
## 1\. Introduction
One of the richest yet least understood narratives in cosmology is the
formation of the complex structure that we see today out of the simple initial
conditions implied by the cosmic microwave background (CMB). The first
luminous objects are thought to have formed at $z\sim 20-30$ through collapse
in $10^{6}-10^{8}\,{\rm M}_{\odot}$ halos (Barkana & Loeb, 2001; Bromm, 2013).
The radiation from these objects heated and then reionized the intergalactic
medium (IGM). There are several sources of complementary information about the
evolution of ionization in this epoch.
The CMB temperature anisotropy is damped by the total Thomson depth to free
electrons. The Planck collaboration has used this effect, combined with a
constraint on the scalar amplitude from gravitational lensing, to measure the
optical depth through reionization (Planck Collaboration et al., 2013). In
addition, Thomson scattering through the reionization epoch generates a unique
polarization signature on large angular scales. WMAP has measured the total
optical depth using this polarization signature (Bennett et al., 2013). These
are integral constraints on the free electron abundance and can be translated
into a central reionization redshift of $10.6\pm 1.1$ (Bennett et al., 2013).
Once the IGM is highly ionized, it is transparent to Ly-$\alpha$ photons.
Absorption measurements along sight lines to high-redshift quasars indicate
that reionization must have ended by $z<6$ (Fan et al., 2006). Absorption
saturates at low abundance, so these should be taken as bounds on the end of
reionization, which could still have been largely complete at redshifts higher
than $6$.
Recently, two methods have been developed to place much more direct bounds on
the duration of reionization. The ionization process is thought to be
spatially patchy, as local sources of radiation blow ionized bubbles that
coalesce into the fully reionized IGM. CMB photons scattering from this patchy
screen produce an additional kinetic Sunyaev-Zel’dovich anisotropy appearing
most clearly at $\ell>3000$, where the primary CMB is negligible (Knox et al.,
1998). Upper limits on this effect translate into a model-dependent upper
bound on the duration of reionization Zahn et al. (2012) and hold the promise
of direct detection of patchy structure in the near future.
The patchy structure of reionization can also be observed directly in three
dimensions using emission of neutral hydrogen through its $21$ cm line
(Furlanetto et al., 2006). Recent bounds from GMRT (Paciga et al., 2013), MWA
(Dillon et al., 2014), and PAPER (Parsons et al., 2014; Pober et al., 2013)
are marching down to the expected level of fluctuations, in parallel to
efforts at LOFAR (van Haarlem et al., 2013). An alternative to measuring the
$21$ cm anisotropy is to measure the signal of its global emission (or
absorption at earlier times) (Pritchard & Loeb, 2008, 2010), which reveals the
bulk energetic properties and ionization state of the universe during
reionization and preceding epochs when the first luminous structures were
forming. Global $21$ cm experiments include EDGES (Bowman & Rogers, 2010) and
SCI-HI Voytek et al. (2014) (which have both reported bounds),
LEDA111http://www.cfa.harvard.edu/LEDA/science.html and the proposed DARE
mission (Burns et al., 2012).
The frequencies of interest in these global studies are $\sim 50-200$ MHz, and
fiducial theoretical models suggest a maximum contrast of $\sim 100$ mK
relative to the synchrotron emission of the galaxy, which can vary $\sim
10^{2}-10^{5}$ K across the sky and frequency range. Astrophysical synchrotron
emission is thought to be fully described by a handful of spectrally smooth
functions that can be distinguished from the variation of the global
reionization signal.
Extremely bright foregrounds make the measurement susceptible to instrumental
systematics. For example, if an instrument with a $1\%$ perturbation to the
spectral calibration observed a $500$ K power law, subtraction of a pure power
law would leave a $5$ K residual, significantly larger than the signal.
Through careful instrumental design, the level of instrumental systematics may
be controlled but generally cannot be nulled entirely.
Here, we develop methods that can be used to constrain the cosmological signal
in these heavily contaminated data. Following the monopole nature of the
signal, experiments to date have mapped the sky with very large beams (Bowman
& Rogers, 2010; Voytek et al., 2014). However, a unique trait of the
foregrounds is that they vary across the sky, while the signal is constant.
This regime is the opposite of the situation normally found in analysis of
small signals, where a modulated signal is pulled out of foregrounds. Liu et
al. (2013) (henceforth L13) proposed that experiments seeking to measure the
global signal should also resolve the sky with an instrumental beam. This
allows selective weighting against regions of high contamination and allows
for the use of angular correlation information to reject foregrounds. The
additional spatial resolution yields higher fidelity recovery of the
cosmological $21$ cm spectrum.
Here, we extend the ideas in L13 to a method that uses the spatial
fluctuations of foregrounds in the data to discover contaminated spectral
modes. A similar idea has been employed successfully in $21$ cm large-scale
structure measurements Switzer et al. (2013); Masui et al. (2013); Chang et
al. (2010) at $z\sim 1$ and has been suggested for cleaning ionospheric
contamination (Vedantham et al., 2014).
Discovery of foreground spectral modes in the measured data makes the method
more robust to assumptions about the foreground contamination. For example,
now if the instrumental passband has a $1\%$ ripple, the largest foreground
mode discovered in our foreground cleaning method will also self-consistently
exhibit this ripple. Generally, instrumental systematics take relatively clean
and smooth functions of frequency from synchrotron emission and convert them
into a more complex structure that requires additional spectral functions to
describe. We argue that the primary goal in instrumental design should be to
prevent proliferation of bright, new foreground modes in the data. Each new
foreground degree of freedom produced by instrumental response to foregrounds
results in more signal loss and makes discovery of the signal more ambiguous.
The methods described here of (1) using spatial variation to discover spectral
foreground modes, which can then be projected out and (2) down-weighting known
spatial areas of high contamination (the galaxy) provide the strongest methods
for recovering the global $21$ cm signal in the absence of additional prior
information about the foregrounds or instrumental response. While the
algorithm of mode subtraction and angular weighting is intuitive, we develop
it from the ground up to expose several implicit choices and possible
pitfalls.
Recently, Bernardi et al. (2014) argued that a dipole gain pattern can be
calibrated in an interferometric array. However, additional variations in
spectral response due to factors such as the analog-to-digital converter,
reflection, and signal loss after the antenna were not included. Our goal here
is to understand how data analysis can be made more robust to this class of
instrumental response (or any other source of foreground covariance), or
alternately how tightly certain instrumental tolerances must be constrained.
In Section 2 we review the basic properties of the global signal and describe
our foreground model. Section 3 builds up the estimator for joint foreground
and signal estimation using a simplified model of spectra along independent
sight lines. Section 4 considers implications of this model for passband
calibration. Section 5 develops spatial weights, and Section 6 combines the
estimators with spatial and spectral weights. Section 7 describes a number of
considerations for using the methods developed here and for global $21$ cm
signal estimation in general. We discuss telescope beam width, the foreground
monopole, how aggressively foreground modes should be removed, and
susceptibility to Faraday rotation. We also consider mode removal of the pre-
reionization absorption feature and extensions of the simple template
amplitude constraint considered throughout. We summarize our conclusions in
Section 8.
## 2\. Global Signal and Foreground Models
From radiative transport arguments, the brightness temperature of $21$ cm
radiation is
$T_{b}\approx 27(1-\bar{x}_{i})\left(\frac{T_{S}-T_{\rm
CMB}}{T_{S}}\right)\left(\frac{1+z}{10}\right)^{1/2}\,{\rm mK},$ (1)
where $\bar{x}_{i}$ is the mean ionization fraction, $T_{S}$ is the spin
temperature of the hyperfine transition, and $T_{\rm CMB}$ is the CMB
temperature. The basic physics of the spin-temperature coupling is well
understood Pritchard & Loeb (2010), but the detailed astrophysical processes
that determine the coupling strength and the gas temperature are still
conjectural.
For $z>200$, the universe is dense enough that electron collisional
interactions drive the spin temperature to the gas temperature, which is
cooling faster than the CMB. This produces absorption. By $z\sim 30$, the
universe is sufficiently rarefied that the spin temperature is better coupled
to the CMB bath, and absorption is expected to subside. Once the first
luminous objects form, these produce radiation that drives the Wouthuysen-
Field (Wouthuysen, 1952; Field, 1958) coupling of $T_{S}$ again to the gas
temperature, leading to a second absorption feature. Then, X-ray heating of
the gas can drive the spin temperature above the CMB temperature, leading to
emission. As these luminous processes proceed and increase, they also ionize
the IGM, which causes the signal to disappear as $1-\bar{x}_{i}\rightarrow 0$
(Pritchard & Loeb, 2010).
We will find it convenient to have a reionization model with a small number of
parameters, rather than a full spectrum. For concreteness, we will spend most
of the paper focusing on the evolution of the brightness temperature as the
universe is reionized, rather than the absorption dip as heating begins.
However, our methods are applicable at all redshifts, and in Section 7.6 we
will briefly discuss the pre-reionization dip.
At the beginning of the reionization epoch, it is thought that the spin
temperature is strongly coupled to the gas and that the gas is heated, driving
$21$ cm emission. As the ionization proceeds, this emission dies away. A
simple way of parameterizing this is (Pritchard & Loeb, 2010)
$\bar{x}_{i}(z)=\frac{1}{2}\left[1+\tanh\left(\frac{z_{r}-z}{\Delta
z}\right)\right],$ (2)
so that the brightness temperature scales as (Figure 1)
$T_{b}(z)=\frac{27}{2}\left(\frac{1+z}{10}\right)^{1/2}\left[1+\tanh\left(\frac{z_{r}-z}{\Delta
z}\right)\right]\,{\rm mK}$ (3)
where we have assumed that $T_{S}\gg T_{\rm CMB}$.
Figure 1.— Brightness temperature of the global $21$ cm emission in a
simplified two-parameter model of reionization. The central redshift is taken
from the WMAP9 constraint at $z=10.6$, and several representative values for
the duration of reionization are shown.
Our model for the diffuse intensity (Stokes-$I$) of the foregrounds is based
on an extended version of the Global Sky Model (GSM) of de Oliveira-Costa et
al. (2008) that is developed in L13. The original GSM used a principal
component analysis to extrapolate and interpolate between previous galactic
surveys at a wide range of frequencies. It is based on three spectral
components and is accurate down to $\approx 10^{\circ}$. L13 extends this
model by adding additional (less smooth) spectral eigenmodes as well as a
population of power-law sources with randomly distributed spectral indices,
drawn from the $dN/dS$ brightness distribution of Di Matteo et al. (2002). The
goal of adding these components is to boost the rank of the foreground
covariance to reflect spectral fluctuations that can be expected on the real
sky. These cannot be probed in the original rank-3 GSM. An example map of our
foreground model at $80\,\textrm{MHz}$ is shown in Figure 2.
Figure 2.— Foreground model at 80 MHz, displayed as $\log(T)$.
In Section 7.7, we will additionally consider the impact of Faraday rotation
and polarized foregrounds. Because this is somewhat peripheral to the
development of our data analysis methods in the next few sections, we will
defer our discussion of our foreground polarization model until then.
## 3\. Global $21$ cm Observations Along Independent Lines of Sight
A global $21$ cm experiment that maps the sky will have a finite number of
resolution elements across its field, defined by the beam. In each one of
these resolution elements, it observes the spectrum of incoming radiation. A
reasonable approximation is to split the survey into $N_{\theta}$ spectra of
length $N_{\nu}$ along all of the angular resolution elements. Let
$\boldsymbol{y}_{i}$ be that vector, where $i$ indexes the sight line ($1$ to
$N_{\theta}$) and the length of the vector is $N_{\nu}$. In this section, we
will build up infrastructure and intuition using this simplified case of
$N_{\theta}$ independent and identically distributed sight lines. Section 6
extends this to a method that treats realistic foregrounds with proper angular
correlations. Because angular correlations become largely irrelevant when
considering wide-beam experiments with essentially no angular sensitivity,
much of the intuition developed here is directly transferable to the analysis
of such experiments.
In writing an estimator, there are many choices for aspects of the global
signal that could be estimated. The goal could be to constrain (1) an
arbitrary spectrum of the cosmological $21$ cm evolution, (2) some modes of
its variation that relate to physical parameters, (3) some amplitudes that are
based on external information about the expected redshift of reionization, or
(4) the amplitude of a template of a provisional global $21$ cm signal. For
our purposes here, it is simplest to develop estimators for the amplitude of
an assumed $21$ cm signal template. Initially, experiments will simply be
stepping down in bounds and seeking some evidence of $z\sim 20$ absorption and
heating or a $z\sim 11$ reionization signal—an amplitude constraint could
provide the simplest, clear indication of a cosmological signal. Section 7.5
considers other estimation regimes in light of the methods developed here.
Let the assumed template of the global signal be a vector $\boldsymbol{x}$ of
length $N_{\nu}$ and normalized to have a maximum of $1$. Multiply by some
amplitude $\alpha$ to get the $21$ cm signal. Then the observed spectrum is
the sum of signal, thermal noise $\boldsymbol{n}_{{\rm inst},i}$, and
foregrounds $\boldsymbol{n}_{{\rm fg},i}$:
$\boldsymbol{y}_{i}=\alpha\boldsymbol{x}+\boldsymbol{n}_{{\rm
fg},i}+\boldsymbol{n}_{{\rm inst},i}.$ (4)
Initially we will assume that the foregrounds and noise are identically and
independently normally distributed along each line of sight $i$,
$\boldsymbol{n}_{{\rm fg},i}\sim N(0,\mathbf{\Sigma}_{\rm fg})$ and
$\boldsymbol{n}_{{\rm inst},i}\sim N(0,\mathbf{\Sigma}_{\rm inst})$. In
reality, the foreground field is strongly correlated in angle and non-
Gaussian. These issues will be examined in subsequent sections. In contrast,
thermal noise is uncorrelated between sight lines and normally distributed:
$\boldsymbol{n}_{{\rm inst},i}\sim N(0,\mathbf{\Sigma}_{\rm inst})$ will
remain an excellent approximation. Throughout, the $N_{\nu}\times N_{\nu}$
matrix $\mathbf{\Sigma}\equiv\mathbf{\Sigma}_{\rm inst}+\mathbf{\Sigma}_{\rm
fg}$ will refer to the total $(\nu,\nu^{\prime})$ covariance.
### 3.1. Known Covariance
If the foregrounds and thermal noise are drawn from a total, known covariance
$\mathbf{\Sigma}$, then the maximum likelihood estimate for the template
amplitude is
$\hat{\alpha}_{\rm
ML}=(\boldsymbol{x}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{\bar{y}}$
(5)
where
$\boldsymbol{\bar{y}}=N_{\theta}^{-1}\sum_{i=1}^{N_{\theta}}\boldsymbol{y}_{i}$
(6)
is the mean spectrum along all lines of sight.
This can be understood as “deweighting” the foregrounds
($\mathbf{\Sigma}^{-1}\boldsymbol{\bar{y}}$), projecting onto the signal
template ($\boldsymbol{x}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{\bar{y}}$), and
then applying a normalization to account for the weights. The estimated
amplitude is normally distributed, and its error is
${\rm Var}(\hat{\alpha}_{\rm
ML})|_{\mathbf{\Sigma}}=(\boldsymbol{x}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{x})^{-1}.$
(7)
To convert this into a more intuitive cleaning process, take the
eigendecomposition of the covariance
$\mathbf{\Sigma}=\boldsymbol{V}\mathbf{\Lambda}\boldsymbol{V}^{T}$. The
cleaning operation $\mathbf{\Sigma}^{-1}\boldsymbol{\bar{y}}$ is then
$\boldsymbol{V}\mathbf{\Lambda}^{-1}\boldsymbol{V}^{T}\boldsymbol{\bar{y}}$.
Here $\boldsymbol{V}^{T}$ projects the data onto a basis where the
contaminated modes $\boldsymbol{v}_{j}$ (the $j$’th row of $\boldsymbol{V}$)
are orthonormal. Those modes are weighted against
$\lambda_{j}^{-1}=[\mathbf{\Lambda}^{-1}]_{jj}$, where the highest variance,
most contaminated modes are down-weighted. Then, $\boldsymbol{V}$ projects
this down-weighted basis back onto the spectral basis.
In the usual assumptions, the foregrounds are not full rank, even if the
spectral covariance $\boldsymbol{\Sigma}$—which contains instrumental noise—is
full rank. Often, the foregrounds are instead described as some set of
contaminated modes $\boldsymbol{v}_{j}$ where $j$ ranges from one to the
number of contaminated modes $N_{\rm fg}$, or
$\boldsymbol{n}_{{\rm fg},i}=\sum_{j}^{N_{\rm fg}}a_{i,j}\boldsymbol{v}_{j}.$
(8)
If the foreground covariance is described by only $N_{\rm fg}$ highly
contaminated modes, a robust cleaning method is to null those modes entirely,
setting $\lambda\rightarrow\infty$ artificially (Liu & Tegmark, 2012). This is
equivalent to fitting and subtracting those spectral modes from each line of
sight, as
$\boldsymbol{\bar{y}}_{\rm clean}=\sum_{j}^{N_{\rm
fg}}(1-\boldsymbol{v}_{j}\boldsymbol{v}_{j}^{T})\boldsymbol{\bar{y}},$ (9)
where the $\boldsymbol{v}_{i}$ are normalized so that
$\boldsymbol{v}^{T}_{i}\boldsymbol{v}_{i}=1$.
Here we have assumed that the foreground spectral functions are known in
advance and can be projected out. This is essentially the same as arguments
for subtracting polynomials or power laws along the line of sight. A crucial
difference is that there is no assumption of smoothness—if the foregrounds
were known to have a particular spectral shape, that could be represented in
the vectors $\boldsymbol{v}_{i}$. The essential property of the foregrounds
that allows for their removal is not that they are smooth, but that they are
described by few functions. In the limit that the number of orthogonal
functions removed approaches $N_{\nu}$, all of the signal is removed because
$\\{\boldsymbol{v}_{i},i=1\ldots N_{\nu}\\}$ spans the space, assuming
orthogonality of the $\boldsymbol{v}_{i}$. The rank of bright foregrounds is
the primary determinant of cosmological signal that can be extracted.
Subtraction of bright foregrounds has proven to be significantly more
challenging in practice due to instrumental effects (Switzer et al., 2013).
Figure 3 shows an example of $0.1\%$ calibration for a power-law foreground
spectrum similar in amplitude to the one reported by Voytek et al. (2014) and
scaled to the slightly different frequencies of interest here. This represents
the case where the $\boldsymbol{v}_{i}$ mode removed assumes a pure power law
but the actual measurement reflects the instrument’s response to
$\boldsymbol{v}_{i}$. Residuals are considerably larger than the signal.
Figure 3.— Impact of calibration uncertainty. Top: a power-law foreground
spectrum similar to the one observed in Voytek et al. (2014), multiplied by a
$0.1\%$ calibration error. Bottom: residuals when a smooth power-law spectrum
is subtracted from data with a calibration error. For contrast, the global
signal at these redshifts is expected to be $\sim 30$ mK. Also note that the
calibration error will not integrate down thermally.
One approach to treating the passband is to control it through instrumental
design and to measure it very precisely prior to the experiment. In principle,
this could be achieved by long integrations on a calibration reference that is
known to be smooth. This carries the challenge of calibrating on a source
other than the sky. An alternative is to recognize that the foregrounds are a
bright spectral reference, viewed through the same survey program as the
primary science. One calibration scheme would then consist of integrating down
on bright synchrotron emission and renormalizing the spectrum under the
assumption that the emission is spectrally smooth (Voytek et al., 2014).
An alternative to calibrating the instrument to meet the smooth mode functions
$\\{\boldsymbol{v}_{i}\\}$ is to adjust the mode functions to reflect
imperfections in the instrument. In the next section we develop a method for
discovering foreground spectral mode functions in the data based on their
spatial variations. The central idea is that the cosmological signal is a
monopole spectrum, so the spectrum of anything that fluctuates across the sky
must reflect some foreground modes (hence the title “erasing the variable”).
If there are monopole foreground components, they are formally
indistinguishable from the cosmological $21$ cm monopole in this picture (in
the absence of prior information).
Measurement of $\\{\boldsymbol{v}_{i}\\}$ modes within the data may also
discover unanticipated or unconstrained instrumental response systematics.
Examples include (1) a passband that varies with time or pointing, which
causes a response to synchrotron radiation that varies across the sky, (2)
polarization leakage, which can cause Faraday (spectral) rotation that varies
across the sky, (3) terrestrial interference that varies with instrument
pointing, and (4) unmodeled frequency dependence of the beam. Even if the
foregrounds were intrinsically a simple rank-1 spectral function, the
instrument response will tend to proliferate the rank of the foreground
covariance. This limits the ultimate sensitivity to the global signal. While
many of these sources of systematics may be tightly controlled by the
instrument’s construction, residual imperfections will not be known in advance
and could be discovered by this method.
We will assume throughout that the true synchrotron foregrounds are described
by a small number of modes $N_{\rm fg,true}$. The instrument observes these
bright foregrounds and records spectral data that may need some $N_{\rm
fg}\geq N_{\rm fg,true}$ functions to describe the contamination.
### 3.2. Covariance from the Measurements
This section considers the joint determination of the global signal amplitude
$\alpha$ and the contaminated foreground modes. The statistical methods
described and extended here were developed by Rao (1967). The central idea in
these methods is to begin with a poor, contaminated estimator for the $21$ cm
signal amplitude. This initial estimator can then be cleaned up by projecting
out correlations with the foregrounds.
To reflect the reality that often one does not know the covariance a priori,
first consider a simple estimator that does not use any information from the
spectral covariance. This will serve as our starting point for constructing an
estimator that simultaneously estimates the spectral covariance from the data.
Again, let $\boldsymbol{x}$ be the spectral template of the cosmological
signal that we would like to constrain. The simplest (and arguably worst)
estimator we could develop for the amplitude is
$T(\parallel\boldsymbol{x})=(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\boldsymbol{\bar{y}},$
(10)
which takes the dot product of the $21$ cm template and observed data
($\boldsymbol{x}^{T}\boldsymbol{\bar{y}}$) and normalizes through
$(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}$. We refer to this as
$T(\parallel\boldsymbol{x})$ because it estimates the amplitude of the
component of $\boldsymbol{\bar{y}}$ parallel to the signal $\boldsymbol{x}$.
This is exactly the estimator that would be used if there were no frequency
correlations in the noise. What makes this a very poor estimator is that it is
strongly correlated with the foreground modes and will therefore contain
strong foreground residuals. In contrast, Equation (5) in the previous section
used the known frequency covariance to deweight the foreground contamination.
We can begin to improve upon our simple estimator by constructing spectral
modes that contain no cosmological signal by design. Using a Gram-Schmidt
process, we can make a set of vectors $\boldsymbol{z}_{i}$ that span the space
orthogonal to the theoretical template of the cosmological signal
$\boldsymbol{x}$, so that $\boldsymbol{z}_{i}^{T}\boldsymbol{x}=0$. There are
$N_{\nu}-1$ such vectors, and the choice of these vectors is completely
arbitrary at this point. Any spectrum can be represented as a sum
$\boldsymbol{y}=\alpha\boldsymbol{x}+\sum_{i}\beta_{i}\boldsymbol{z}_{i}$ (11)
because $\\{\boldsymbol{x},\boldsymbol{z}_{i}\\}$ span the
$N_{\nu}$-dimensional space. We can pack these vectors $\boldsymbol{z}_{i}$
into a matrix $\boldsymbol{Z}$ and write an estimator for all of the spectral
information orthogonal to $\boldsymbol{x}$,
$T(\perp\boldsymbol{x})=\boldsymbol{Z}^{T}\boldsymbol{\bar{y}}$. This
represents all of the components of $\boldsymbol{\bar{y}}$ orthonormal to the
signal and is dominated by foregrounds. (Note that we could choose to
normalize this for general vectors $\boldsymbol{z}$ and find
$(\boldsymbol{Z}^{T}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}\boldsymbol{\bar{y}}$.
This would make the equations more complex but would not modify the estimator
for the cosmological quantity.)
The covariance between $T(\parallel\boldsymbol{x})$ and
$T(\perp\boldsymbol{x})$ is
$\displaystyle{\rm Cov}_{\parallel,\perp}$ $\displaystyle=$
$\displaystyle\left(\begin{array}[]{cc}\boldsymbol{C}_{\parallel,\parallel}&\boldsymbol{C}_{\parallel,\perp}\\\
\boldsymbol{C}_{\perp,\parallel}&\boldsymbol{C}_{\perp,\perp}\end{array}\right)$
(14) $\displaystyle=$
$\displaystyle\left(\begin{array}[]{cc}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\Sigma}\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}&(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\Sigma}\boldsymbol{Z}\\\
\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}&\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{Z}\end{array}\right).$
(17)
The off-diagonal terms
$(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\Sigma}\boldsymbol{Z}$
represent correlations between $T(\parallel\boldsymbol{x})$ and
$T(\perp\boldsymbol{x})$ caused by the $(\nu,\nu^{\prime})$ covariance of
foregrounds in $\mathbf{\Sigma}$. If the noise terms were only thermal
fluctuations, $\mathbf{\Sigma}=\sigma_{\rm inst}^{2}\mathbf{1}$, and by the
construction $\boldsymbol{x}^{T}\boldsymbol{Z}=0$, the off-diagonal terms
vanish. In other words, an estimator $T(\perp\boldsymbol{x})$ that is known to
contain only contaminants (and no cosmological signal) is correlated with our
estimate of the cosmological signal, so the latter must be contaminated.
With knowledge of spectral correlations, we can form an improved, adjusted
estimator
$\hat{\alpha}=T(\parallel\boldsymbol{x})-\boldsymbol{C}_{\parallel,\perp}\boldsymbol{C}_{\perp,\perp}^{-1}T(\perp\boldsymbol{x})$
(18)
that projects foreground frequency correlations out of the
$T(\parallel\,\boldsymbol{x})$ estimator. Intuitively, this estimator
instructs us to make an estimate $T(\perp\boldsymbol{x})$ of the portion of
the measurement that is known to contain only foregrounds. Off-diagonal
elements of the covariance matrix then allow the level of foreground leakage
(into our estimate $T(\parallel\boldsymbol{x})$ of the cosmological signal) to
be predicted and subtracted off. A similar technique was used recently in a
power spectrum analysis of PAPER data (Parsons et al., 2014).
So far, we have assumed that $\mathbf{\Sigma}$ is known, but the same
covariance adjustment can be performed with respect to an estimated
$\mathbf{\hat{\Sigma}}$. Without perfect knowledge of $\mathbf{\Sigma}$, the
error bars grow but the estimator remains unbiased. The $(\nu,\nu^{\prime})$
covariance can be estimated empirically using
$\mathbf{\hat{\Sigma}}=(N_{\theta}-1)^{-1}\sum_{i=1}^{N_{\theta}}(\boldsymbol{y}_{i}-\boldsymbol{\bar{y}})(\boldsymbol{y}_{i}-\boldsymbol{\bar{y}})^{T}$
(19)
Writing out Equation (18),
$\hat{\alpha}=(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}(1-\mathbf{\Pi})\boldsymbol{\bar{y}}$
(20)
where
$\mathbf{\Pi}=\mathbf{\hat{\Sigma}}\boldsymbol{Z}(\boldsymbol{Z}^{T}\mathbf{\hat{\Sigma}}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}$.
Remarkably, under our current assumptions where $\boldsymbol{Z}$ spans the
rest of the $N_{\nu}$ dimensions, this estimator is equivalent to (see
Appendix A)
$\hat{\alpha}=(\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{\bar{y}},$
(21)
which is precisely the same as Equation (5), except with the estimated
spectral covariance $\mathbf{\hat{\Sigma}}$ replacing the known covariance
$\mathbf{\Sigma}$.
Gleser & Olkin (1972) develop the same result through a completely independent
method. They write down the joint likelihood of the covariance and mean
(cosmological signal), where the covariance is Wishart-distributed and the
mean is normally distributed. They then maximize the likelihood and find the
same result, showing that the choice
$\mathbf{\Sigma}\rightarrow\mathbf{\hat{\Sigma}}$ coincides with the maximum
likelihood.
In summary, the proposed procedure is to
* •
Find the sample mean and $(\nu,\nu^{\prime})$ covariance across observed lines
of sight.
* •
Invert the measured covariance and use it to down-weight contaminated modes in
the data.
* •
Find the inner product of the cleaned data and the signal template, then
normalize.
While the maximum of the likelihood has the same form as the case where the
covariance is known, the distribution of the estimated $21$ cm global signal
template amplitude is no longer Gaussian, and is generally broader. These
changes are due to the fact that the data are also used for the covariance
estimation, which uses up degrees of freedom in determining the foreground
modes. Rao (1967) shows that
${\rm
Var}(\hat{\alpha})=\frac{N_{\theta}-1}{N_{\theta}-1-r}[\boldsymbol{C}_{\parallel,\parallel}-\boldsymbol{C}_{\parallel,\perp}\boldsymbol{C}_{\perp,\perp}^{-1}\boldsymbol{C}_{\perp,\parallel}],$
(22)
where $r$ is the rank of the additional degrees of freedom that are estimated.
In our case here, $r=N_{\nu}-1$ ($\boldsymbol{Z}$ spans the rest of the
spectral space) so that
${\rm
Var}(\hat{\alpha})=\left(1+\frac{N_{\nu}-1}{N_{\theta}-N_{\nu}}\right)(\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{x})^{-1},$
(23)
where we have also plugged in the explicit forms for
$\boldsymbol{C}_{\parallel,\parallel}$,
$\boldsymbol{C}_{\parallel,\perp}$,$\boldsymbol{C}_{\perp,\perp}$, and
$\boldsymbol{C}_{\perp,\parallel}$ from Equation (14). We see that the
covariance is enhanced relative to the case of perfect foreground covariance
knowledge (Equation (7)) by a factor related to the number of frequencies and
independent lines of sight. The error diverges when $N_{\theta}$ approaches
$N_{\nu}$ from above. The rank $r(\mathbf{\hat{\Sigma}})\leq{\rm
min}(N_{\theta},N_{\nu})$, and for $N_{\theta}\gg N_{\nu}$, the covariance is
well measured. This suggests that the optimal limit is to have many more
resolution elements than spectral bins.
There is little instrumental limitation on the number of spectral bins
$N_{\nu}$ over the bands observed. In contrast, there is a hard limit on the
number of “independent” lines of sight that can be observed, $N_{\theta}$. At
the frequencies of interest for the global $21$ cm signal, the beam size tends
to be large by diffraction. Angular correlations and noise generally limit the
number of independent samples of the spectrum (see Section 7.1).
The requirement $N_{\theta}>N_{\nu}$ stems from the fact that
$\mathbf{\hat{\Sigma}}$ is trying to estimate a full $N_{\nu}$ rank
covariance. In general, the sensitivity to $\alpha$ should not depend on the
number of spectral bins in the survey, once they become fine enough to resolve
the signal. The solution is to instead estimate some number $N_{\rm fg}$ of
contaminated spectral functions, less than the number of spectral bins. Then
the estimation becomes independent of the $N_{\nu}$, so long as
$N_{\nu}>N_{\rm fg}$.
### 3.3. Empirically Determined Foregrounds of Limited Rank
The key step in the covariance adjustment scheme of the previous section was
the formation and application of
$\mathbf{\Pi}=\mathbf{\hat{\Sigma}}\boldsymbol{Z}(\boldsymbol{Z}^{T}\mathbf{\hat{\Sigma}}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}$,
which modeled contaminated modes in the original “poor” estimator for later
subtraction. The modes formed a basis $\boldsymbol{Z}$ orthogonal to the
signal $\boldsymbol{x}$. The specific choice of vectors in $\boldsymbol{Z}$
was arbitrary so long as they were orthonormal (for simplicity) and spanned
the spectral subspace orthogonal to the signal. For concreteness, we suggested
forming this basis blindly using a Gram-Schmidt process. While formally a
solution to our problem, this is not a particularly efficient way to implement
our recipe in practice, for each of the resulting basis vectors will be a
linear combination of noise and foreground modes. On the other hand, the
foregrounds should be describable by a small number of modes (Liu & Tegmark,
2012; Bernardi et al., 2014). A sensible alternative would therefore be to
intelligently partition the basis $\boldsymbol{Z}$ into two sub-bases, one
that is of a relatively low rank $N_{\rm fg}$ containing the foregrounds and
another of rank $N_{\nu}-N_{\rm fg}-1$ consisting of the remaining noise-
dominated modes. Computationally, this means that rather than estimating the
full $N_{\nu}$ rank covariance from the data, we only need to estimate a
subset of contaminated modes and their amplitudes. This limits the number of
degrees of freedom $r$ that need to be estimated from the data, which as we
saw from Equation (22) is crucial for keeping the final error bars small.
In the rank-restricted foreground approximation, the data along a line of
sight are the cosmological signal plus some amplitudes $\boldsymbol{\beta}$
times foreground modes $\boldsymbol{F}$ plus thermal noise, as
$\boldsymbol{y}=\alpha\boldsymbol{x}+\boldsymbol{F}\boldsymbol{\beta}+\boldsymbol{n}_{t}\Rightarrow\mathbf{\Sigma}=\boldsymbol{F}\mathbf{\Gamma}\boldsymbol{F}^{T}+\sigma_{\rm
inst}^{2}\boldsymbol{1},$ (24)
where we have taken the thermal noise to be stationary for simplicity and let
$\mathbf{\Gamma}$ be the eigenvalue matrix of the foreground spectral
covariance
$\boldsymbol{F}\langle\boldsymbol{\beta}\boldsymbol{\beta}^{T}\rangle\boldsymbol{F}^{T}$,
quantifying the power of the foreground modes $\boldsymbol{F}$. The amplitude-
ordered eigenvalue spectrum of the total covariance $\mathbf{\Sigma}$ in this
case would show some large contaminant amplitudes followed by a noise floor
set by $\sigma$.
Unlike $\boldsymbol{Z}$ of the previous section, the foreground mode vectors
in $\boldsymbol{F}$ alone do not span the entire spectral subspace orthogonal
to the cosmological signal. The remaining portion of the space is spanned by
basis vectors that describe instrumental noise, under the assumptions of
Equation (24). These vectors can be formed, like before, using a Gram-Schmidt
process (this time relative to the signal $\boldsymbol{x}$ and foregrounds
$\boldsymbol{F}$) and packed into a matrix $\boldsymbol{G}$. The observed
spectrum $\boldsymbol{y}$ can then be written as
$\boldsymbol{y}=\alpha\boldsymbol{x}+\sum_{j=1}^{N_{\rm
fg}}\beta_{j}\boldsymbol{f}_{j}+\sum_{j=N_{\rm
fg}+1}^{N_{\nu}}\gamma_{j}\boldsymbol{g}_{j},$ (25)
which is to be contrasted with Equation (11).
Again, we can form poor estimators for the amplitudes of the signal
($\hat{\alpha}=(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\boldsymbol{\bar{y}}$),
foreground modes
($\boldsymbol{\hat{\beta}}=\boldsymbol{F}^{T}\boldsymbol{\bar{y}}$), and
thermal noise modes
($\boldsymbol{\hat{\gamma}}=\boldsymbol{G}^{T}\boldsymbol{\bar{y}}$). To do
so, however, we must first specify how the foreground modes $\boldsymbol{F}$
are identified and defined, and in the following section we outline two
different methods for this.
### 3.4. Two Outlooks on Separation of Signal and Foregrounds
The first method is closely related to the treatment that we have presented so
far. In particular, we considered modes in the matrices $\boldsymbol{Z}$ (full
rank) or $\boldsymbol{F}$ (finite rank) that represent components of the data
that are orthogonal to the cosmological signal. There is, however, nothing
preventing real foregrounds from being parallel to the signal, and generally
$\boldsymbol{n}_{\rm fg}$ will be the sum of $\boldsymbol{n}_{{\rm
fg},\parallel\boldsymbol{x}}$ (foregrounds parallel to the cosmological
signal) and $\boldsymbol{n}_{{\rm fg},\perp\boldsymbol{x}}$ (foregrounds
perpendicular to the signal). For example, if the signal and foregrounds share
a slowly varying spectral component, it will fall in $\boldsymbol{n}_{{\rm
fg},\parallel\boldsymbol{x}}$.
An alternative is to form foreground spectral modes that best describe the
covariance of the fluctuating terms on the sky without regard for the
cosmological signal. Then, the signal has a piece that is parallel to those
foreground modes and a piece that is perpendicular, or
$\boldsymbol{x}=\boldsymbol{x}_{\parallel\boldsymbol{F}}+\boldsymbol{x}_{\perp\boldsymbol{F}}$.
In equations, these two methods are
$\displaystyle\textrm{ Method 1:}~{}~{}~{}~{}\boldsymbol{y}$ $\displaystyle=$
$\displaystyle\boldsymbol{x}+\boldsymbol{n}_{{\rm
fg},\parallel\boldsymbol{x}}+\boldsymbol{n}_{{\rm
fg},\perp\boldsymbol{x}}~{}~{}~{}~{}\textrm{delete:}~{}~{}\boldsymbol{n}_{{\rm
fg},\perp\boldsymbol{x}}$ $\displaystyle\textrm{ Method
2:}~{}~{}~{}~{}\boldsymbol{y}$ $\displaystyle=$
$\displaystyle\boldsymbol{x}_{\perp\boldsymbol{F}}+\boldsymbol{x}_{\parallel\boldsymbol{F}}+\boldsymbol{n}_{\rm
fg}~{}~{}~{}~{}\textrm{delete:}~{}~{}\boldsymbol{x}_{\parallel\boldsymbol{F}}+\boldsymbol{n}_{\rm
fg}.$
In the first method, we can develop a cleaning operation that
removes/deweights $\boldsymbol{n}_{{\rm fg},\perp\boldsymbol{x}}$, that is,
components of the foreground variance orthogonal to the signal. This estimator
does not touch the signal by design, so it is guaranteed to have no
cosmological signal loss. However, substantial foregrounds will remain in the
estimated signal with this rather conservative approach. In the second method,
we can develop cleaning that removes any foreground spectral modes that vary
spatially on the sky along with any component of the signal parallel to that.
If implemented carelessly, this aggressive method will entail cosmological
signal loss, although the problem is rectifiable.
Method 1 is described in Rao (1967) and is a simply a slight modification of
the prescription in Section 3.2. Method 2 is the one we develop here and
advocate for global $21$ cm signal recovery.
We demonstrate the differences in approach using simple simulations. These
have $N_{\theta}$ lines of sight and a number of power laws in the synchrotron
emission (to limit the rank). The indices of these power laws are
$\\{-2.1,-2.5,-2.9\\}$ and the amplitude is uniformly distributed and
positive. Here, $-2.5$ is a characteristic synchrotron index at these
frequencies Rogers & Bowman (2012); Voytek et al. (2014), and the two spanning
it give additional modes in our initial toy model. Again, as amplitudes we
take representative values from the recent observations of Voytek et al.
(2014) and scale to the slightly higher frequencies of interest here. For the
purposes here we will neglect thermal noise and assume the integration is deep
relative to the cosmological signal. Figure 4 shows one realization of the
foreground model. This model is only meant to be pedagogical, showing the main
point of the algorithms here for independent sight lines in a simplified
setting with three spectral modes. Beginning in Section 5 we will use the full
extended GSM model data cubes described in Section 2.
Figure 4.— Simple simulation of 10 lines of sight with a random combination of
three power laws.
Figure 5.— Modes discovered from the $(\nu,\nu^{\prime})$ covariance of simple
simulations with three synchrotron spectral indices (Figure 4). These include
only the foregrounds and simple signal model of Equation (3) with $\Delta
z=0.5$ and $z_{r}=10.6$. Top left: the largest three eigenvectors of the
covariance $\mathbf{\hat{\Sigma}}$ that fully describe the foregrounds. Top
right: the filtered data
$(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{\bar{y}}$. The input signal
in red is largely recovered after cleaning the three foreground eigenmodes in
the top left. In this toy model, the foregrounds are removed entirely, and
only the signal (minus components parallel to the subtracted foregrounds) is
left. Bottom left: the largest three eigenvectors of the restricted covariance
$\mathbf{\hat{\Sigma}}|_{\boldsymbol{x}}=[1-\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}]\mathbf{\hat{\Sigma}}$.
These are all orthogonal to the cosmological signal. Bottom right: the data
after subtracting the restricted eigenvectors [or
$(1-\boldsymbol{F}_{\perp}\boldsymbol{F}_{\perp}^{T})\boldsymbol{\bar{y}}$]
are exactly the signal template shape by the construction of the mode
functions. The amplitude, however, is clearly dominated by residual
foregrounds (note that the signal is too small to be seen on this vertical
scale). These are the foreground components parallel to the cosmological
signal. While this procedure is immune from loss of cosmological signal, it is
insufficiently aggressive when dealing with foregrounds, and the final result
is strongly influenced by the foreground covariance. We note that the apparent
recovery of the reionization step’s location in the bottom-right subplot is
spurious, for it is almost entirely driven by our choice of which theoretical
template we use. The location of the “step” in the top-right subplot, on the
other hand, is truly constrained by the data, although signal loss means that
it no longer appears as a clean step.
#### 3.4.1 Method 1: Foreground Modes Orthogonal to Signal
To find foreground spectral modes that are orthogonal to the cosmological
signal $\boldsymbol{x}$, we can find the largest eigenmodes of the restricted
covariance
$\mathbf{\hat{\Sigma}}\biggl{|}_{\perp\boldsymbol{x}}=[1-\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}]\mathbf{\hat{\Sigma}},$
(26)
which in practice will be eigenmodes that represent foregrounds. By
construction, any eigenmode $\boldsymbol{f}_{i}$ of
$\mathbf{\hat{\Sigma}}\bigl{|}_{\perp\boldsymbol{x}}$ will be orthogonal to
$\boldsymbol{x}$, i.e. $\boldsymbol{f}_{i}^{T}\boldsymbol{x}=0$. Let
$\boldsymbol{F}_{\perp}$ hold the restricted eigenvectors
$\boldsymbol{f}_{i}$. The top-left subplot of Figure 5 shows the unrestricted
eigenvectors of an input foreground simulation, while the bottom-left subplot
shows the restricted eigenvectors relative to a signal template.
Having identified our foreground eigenvectors, we can construct our estimator
for the signal amplitude $\alpha$ using the methods introduced in Section 3.2.
Once again, we first form a poor initial estimator of the signal, which we
then adjust using the covariance between the poor estimator and the
foregrounds (which is nonzero because a substantial foreground contamination
remains in the estimate of the signal). The covariance of this estimator takes
the form
${\rm
Cov}_{\hat{\alpha},\boldsymbol{\hat{\beta}}}=\left(\begin{array}[]{cc}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}&(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}\boldsymbol{F}_{\perp}\\\
\boldsymbol{F}^{T}_{\perp}\mathbf{\hat{\Sigma}}\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}&\boldsymbol{F}^{T}_{\perp}\mathbf{\hat{\Sigma}}\boldsymbol{F}_{\perp}\end{array}\right),$
(27)
which looks exactly like Equation (14) but with all occurrences of
$\boldsymbol{Z}$ replaced by $\boldsymbol{F}_{\perp}$. This is unsurprising
because the modes in $\boldsymbol{G}$ represent a diagonal piece of the
covariance, $\sigma_{\rm inst}^{2}\boldsymbol{1}$, and thus are not correlated
in frequency with either the foreground modes or the signal.
Projecting out the piece of the signal estimator that is correlated with the
foregrounds,
$\displaystyle\hat{\alpha}$ $\displaystyle=$
$\displaystyle(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}(1-\mathbf{\Pi})\boldsymbol{\bar{y}}$
(28) $\displaystyle\mathbf{\Pi}$ $\displaystyle=$
$\displaystyle\mathbf{\hat{\Sigma}}\boldsymbol{F}_{\perp}(\boldsymbol{F}_{\perp}^{T}\mathbf{\hat{\Sigma}}\boldsymbol{F}_{\perp})^{-1}\boldsymbol{F}_{\perp}^{T}.$
(29)
That is, project foreground-contaminated modes in the data and then dot
against the signal template. The operation $\mathbf{\Pi}$ puts the signal in
the basis of foreground modes, weights those by their covariance, and moves
back to the spectral space. Note that $\boldsymbol{F}_{\perp}$ does not span
the rest of the spectral space like $\boldsymbol{Z}$ in the previous section.
Here, Equation (28) is not equivalent to
$\hat{\alpha}=(\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{\bar{y}}$.
By construction, our estimator suffers from no loss of cosmological signal
(despite the projecting-out of foregrounds) because
$(1-\mathbf{\Pi})\boldsymbol{x}=\boldsymbol{x}-\mathbf{\hat{\Sigma}}\boldsymbol{F}_{\perp}(\boldsymbol{F}_{\perp}^{T}\mathbf{\hat{\Sigma}}\boldsymbol{F}_{\perp})^{-1}\boldsymbol{F}_{\perp}^{T}\boldsymbol{x}=\boldsymbol{x}$.
In the last step, we used the fact that
$\boldsymbol{F}_{\perp}^{T}\boldsymbol{x}=0$ through the choice of forming
$\boldsymbol{F}_{\perp}$ with eigenvectors restricted to be orthogonal to
$\boldsymbol{x}$, Equation (26).
The error on $\hat{\alpha}$ is
$\displaystyle{\rm Var}(\hat{\alpha})$ $\displaystyle\propto$
$\displaystyle[\boldsymbol{C}_{\parallel,\parallel}-\boldsymbol{C}_{\parallel,\perp}\boldsymbol{C}_{\perp,\perp}^{-1}\boldsymbol{C}_{\perp,\parallel}]$
(30) $\displaystyle\propto$
$\displaystyle(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}(1-\mathbf{\Pi})\mathbf{\hat{\Sigma}}\boldsymbol{x}(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}.$
(31)
Here $\boldsymbol{x}^{T}(1-\mathbf{\Pi})\mathbf{\hat{\Sigma}}\boldsymbol{x}$
can be interpreted as projecting all of the foreground modes orthogonal to the
signal out of the $(\nu,\nu^{\prime})$ covariance and finding the noise with
respect to that residual covariance. The residual covariance originates from
foreground components parallel to the signal and from thermal noise—any
$(\nu,\nu^{\prime})$ covariance component with a nonzero dot product into the
signal $\boldsymbol{x}$ will contribute to the variance of $\hat{\alpha}$.
This property can be seen in Fig 5. The recovered signal has exactly the shape
of the signal template $\boldsymbol{x}$, but the amplitude is dominated by
foregrounds. This feature is clearly undesirable in the regime where
foregrounds vastly exceed the signal.
Another undesirable property of this estimator is that the quoted error ${\rm
Var}(\hat{\alpha})$ relies on the foregrounds being normally distributed. Real
foregrounds are strongly non-Gaussian, and we would like to avoid describing
those components in the errors of $\hat{\alpha}$. We can make a simple
modification to the estimator in Rao (1967) to treat both of these
shortcomings.
#### 3.4.2 Method 2: Signal Component Orthogonal to Foregrounds
In some sense, the method described in Section 3.4.1 was too conservative. By
limiting our labeling of foregrounds to modes that are orthogonal to the
signal, we arrived at an estimator with no formal signal loss, but one in
which substantial foreground residuals remained in the final answer. Here we
consider an estimator that more aggressively projects out foregrounds. The
result will be a lossy treatment, but we will also show how this can be
rectified.
Instead of isolating the foregrounds orthogonal to the signal, consider the
component of the signal that is orthogonal to the foregrounds. The signal is
then partitioned as
$\boldsymbol{x}=\boldsymbol{x}_{\parallel\boldsymbol{F}}+\boldsymbol{x}_{\perp\boldsymbol{F}}$
(32)
For this decomposition to be meaningful, we must once again decide on a
definition for our foreground modes. We let $\boldsymbol{F}$ be the largest
foreground eigenvectors of $\mathbf{\hat{\Sigma}}$ (without the orthogonality
restrictions of the previous section). In the calculation for
$\mathbf{\hat{\Sigma}}$, the mean $\boldsymbol{\bar{y}}$ contains the
cosmological signal and is subtracted out. For this reason, the cosmological
signal does not perturb the $(\nu,\nu^{\prime})$ covariance or its eigenmodes.
This is in contrast to inhomogeneous $21$ cm mapping, where the cosmological
signal is stochastic, cannot be subtracted from the covariance, and does
perturb the eigenvectors (Switzer et al., 2013; Masui et al., 2013). (The
foreground modes can optionally be isolated from thermal noise by forming the
$(\nu,\nu^{\prime})$ cross-variance of maps acquired at two different times,
as was done in Switzer et al. 2013.)
To project the identified foreground modes out of the estimator, we apply
$1-\mathbf{\Pi}=1-\boldsymbol{F}\boldsymbol{F}^{T}$. This operation splits the
spectrum into a foreground-contaminated and a (theoretically) foreground-clean
subspace as
$\left(\begin{array}[]{c}\mathbf{\Pi}\\\
1-\mathbf{\Pi}\end{array}\right)\boldsymbol{y}=\left(\begin{array}[]{c}\mathbf{\Pi}\\\
1-\mathbf{\Pi}\end{array}\right)\alpha\boldsymbol{x}+\left(\begin{array}[]{c}\boldsymbol{n}_{\rm
fg}\\\ \mathbf{0}\end{array}\right)+\boldsymbol{n}_{\rm inst}.$ (33)
Unlike the previous estimator Equation (28), however, the foreground modes are
no longer constructed to be orthogonal to the signal. By projecting out the
foregrounds, then, it is likely that some cosmological signal will be lost.
Mathematically, $(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}$ equals
$\boldsymbol{x}_{\perp\boldsymbol{F}}$, not $\boldsymbol{x}$, and the
estimator
$\hat{\alpha}=(\boldsymbol{x}^{T}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}(1-\mathbf{\Pi})\boldsymbol{\bar{y}}$
is an incorrectly normalized estimator of the signal amplitude (i.e., it
suffers from a multiplicative bias). Correcting this, we propose the revised
estimator
$\hat{\alpha}=\frac{\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{\bar{y}}}{\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}}.$
(34)
In the error calculation from Equation (31), the key term
$(1-\mathbf{\Pi})\mathbf{\hat{\Sigma}}$ now reduces to $\mathbf{1}\sigma_{\rm
inst}^{2}$, under the assumption that all foregrounds are removed to a good
approximation and only Gaussian thermal noise remains. While the foregrounds
may be strongly non-Gaussian, we will assume that the residuals after the
modes $\boldsymbol{F}$ have been subtracted are Gaussian and due to the
thermal noise floor. This would need to be verified in an analysis of real
data.
The overall error is now
${\rm Var}(\hat{\alpha})\sim\left(1+\frac{N_{\rm fg}}{N_{\theta}-1-N_{\rm
fg}}\right)\frac{\sigma_{\rm
inst}^{2}}{[\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}]^{2}}$
(35)
To get some rough intuition, assume that the foreground modes all have about
the same overlap with the cosmological signal. Then
$\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}\propto
N_{\nu}-N_{\rm fg}$. As $N_{\rm fg}$ approaches $N_{\nu}$, the foreground
modes do a progressively better job of spanning the $N_{\nu}$ spectral space.
In the limit that they fully span the space (there is no signal
distinguishable from foregrounds), the signal is also nulled and the error
diverges. Alternately, as $N_{\rm fg}$ approaches $N_{\theta}$, the foreground
discovery process uses up all of the spatial degrees of freedom and the error
diverges. In this formulation, then, we are self-consistently including the
possibility of signal loss in our error analysis.
## 4\. Constraints on Passband Calibration
The abstract design guidance suggested by this method is that (1) the
experiment should not increase the rank of the foreground covariance (keep
$N_{\rm fg}$ small) and (2) if there are foregrounds with $N_{\rm fg}$
spectral degrees of freedom, at least $N_{\theta}=N_{\rm fg}$ samples of its
variations need to be observed if those foregrounds are to be subtracted.
A time-varying passband calibration is the worst instrumental systematic in
this sense. Here each line of sight sees a slightly different foreground
spectrum and so requires a different spectral function for cleaning. In the
limit that each line of sight can have a different foreground, $N_{\rm
fg}\rightarrow N_{\theta}$ and the errors diverge. In contrast, a constant
passband error only modulates the foreground covariance spectral functions,
but it does not increase the rank of its covariance. Its impact is more
aesthetic, can be corrected by recalibrating off the brightest mode, and does
not fundamentally limit the investigation. (This is proven in Section 6 using
the specific notation developed there.)
Figure 6 shows the 10 simulated lines of sight with a constant $10\%$
calibration error. Here we multiply the spectrum along each line of sight by a
constant $1+\boldsymbol{\delta}$, where $\boldsymbol{\delta}$ is $N_{\nu}$
long and normally distributed. Fitting a smooth power law to data with this
constant $1+\boldsymbol{\delta}$ calibration error would leave unacceptable
residuals, but the modes are able to discover the constant calibration error
and subtract the foreground with modest signal loss. This parameterization of
a calibration error is a matter of convenience. In simulations where the
errors $\boldsymbol{\delta}$ are a smoothly varying function of frequency or
even have the same shape as the signal template, the conclusions are the same.
(In the pathological case that the calibration error interacts with the
foreground power law to produce modes identical to the cosmological template,
indeed signal would be lost. This loss is self-consistently accounted for in
the method and could be discovered by visual inspection of the modes.)
In contrast, Figure 7 shows that even a $0.1\%$ _variable_ passband
calibration error spreads foreground variance into many modes that are all
significantly larger than the signal ($10^{-5}$ there). To implement a
variable passband, we multiply each line of sight by a different
$1+\boldsymbol{\delta}_{i}$, representing a change in the instrument response
with time.
The rule of thumb here is that if foregrounds are $R$ times larger than the
signal, then the per-pointing calibration needs to be stable to $\sim 1/R$.
Averaging over sight lines tends to relax this constraint. In the simulations
shown in Section 7.1, we find that the stability must be better than $\sim
10^{-4}$ to have no effect. The constraint degrades rapidly above that. In the
other limit of extremely high stability, we find that once calibration
stability is suppressed below the thermal noise, there is no further gain in
foreground discrimination.
An important point is that the method developed here can discover any response
to foregrounds that varies across the sky (non-monopole). This does not
necessarily need to originate from passband calibration or differential
polarization response—these are just plausible sources of error from
instrumental response.
Figure 6.— Toy model showing that an absolute passband calibration error that
does not vary with time does not impact primordial signal recovery. Top left:
observations of the spectrum across 10 lines of sight with a stable $10\%$
passband calibration error between bins (here neglecting thermal noise). The
input synchrotron is taken from a mixture of three power laws, so it is
described by three spectral modes in the data. Bottom left: foreground modes
discovered in the data. Because of passband calibration error, these modes are
not smooth. A constant calibration error does not increase the rank of the
three input power laws. Right: recovered cosmological signal compared to the
input signal, and the cosmological signal with the three modes on the left
removed. The fixed calibration error is cosmetic. Figure 7.— Residual
foregrounds as a function of modes removed in a toy model with three
foreground degrees of freedom. Amplitudes are normalized to one for zero modes
removed and in rms rather than variance. In the case with no variable
calibration error, the variance in the spectrum quickly reaches thermal noise
after three modes are removed. If the passband calibration varies between
lines of sight, the residuals become significantly worse. Generally, for
foregrounds $10^{5}$ times larger than the signal, the passband should be
stable to $\sim 10^{-5}$. Passband calibration instability spreads the
variance of the three intrinsic foreground modes across many more degrees of
freedom.
## 5\. Incorporating Spatial Information
The previous discussion assumed that the foregrounds are stationary and
independent between different lines of sight. It is for this reason that the
first step in the prescription outlined above was to form $\overline{y}$, the
unweighted average of all lines of sight—with stationary statistics, there is
no reason to prefer one direction of the sky over another, and an unweighted
average provides the best signal-to-noise ratio. This is the implicit
assumption that is being made by experiments that average over large regions
of sky at once (such as single-dipole experiments) and produce a single
spectrum as their measurement. The results in preceding sections therefore
also apply to experiments with no angular sensitivity, except without that
angular information, error bars and foreground cleaning methods cannot be
informed by the data itself and must be derived a priori.
In reality, foregrounds are neither stationary nor independent. The
synchrotron brightness varies across the sky (violating stationarity), and in
addition the foregrounds are known to be spatially correlated (violating
independence).
Angular information can be leveraged in several ways. First, nonstationarity
of the foregrounds allows the estimator to down-weight parts of the sky that
are known to be particularly noisy, such as the galactic center. Second,
correlation information allows foreground properties in one part of the sky to
be inferred from observations of another part of the sky. For example, in the
unrealistically extreme limit of perfectly correlated foregrounds, a
measurement of foregrounds in any part of the sky automatically allows a
perfect prediction of foreground brightness in any other part of the sky.
To take advantage of angular information, L13 considered an optimal estimator
given full-sky maps at all frequencies of interest as well as a known
$N_{\nu}N_{\rm pix}\times N_{\nu}N_{\rm pix}$ covariance $\mathbf{N}$ between
all pixels and all frequencies. Let $\boldsymbol{d}$ be a vector of length
$N_{\nu}N_{\theta}$ containing the measured sky maps. Here we use $N_{\rm
pix}$ to denote the number of pixel indices in a full-sky HEALPix map (Górski
et al., 2005). For reasons we will see in Section 7.1, this also prevents
confusion between the $N_{\theta}$ independent sight lines in the previous
section and the present discussion, which is made more complex by beam
convolution.
The observations $\boldsymbol{d}$ are related to the global $21$ cm spectrum
$\boldsymbol{s}$ via the measurement equation
$\boldsymbol{d}=\boldsymbol{A}\boldsymbol{s}+\boldsymbol{n},$ (36)
where
$\boldsymbol{n}\equiv\mathbf{n}_{\textrm{fg}}+\mathbf{n}_{\textrm{inst}}$ is
the generalized noise, containing $\mathbf{n}_{\textrm{fg}}$ and
$\mathbf{n}_{\textrm{inst}}$ as the foreground and instrumental noise
contribution to $\boldsymbol{d}$, respectively. The $N_{\rm pix}N_{\nu}\times
N_{\nu}$ matrix $\boldsymbol{A}$ is given by
$\boldsymbol{1}\otimes\boldsymbol{e}_{0}$, where $\otimes$ is the Kronecker
product, $\boldsymbol{1}$ is an $N_{\nu}\times N_{\nu}$ identity matrix, and
$\boldsymbol{e}_{0}$ is the spatial monopole, i.e., an $N_{\rm pix}$-long
vector of 1 s. Its function is to copy, at every frequency, the value of the
global spectrum $\boldsymbol{s}$ to all of the spatial pixels comprising the
sky map. The maximum-likelihood estimate $\hat{\mathbf{s}}_{\textrm{ML}}$ for
the full global spectrum is given by
$\hat{\boldsymbol{s}}_{\textrm{ML}}=(\boldsymbol{A}^{T}\boldsymbol{N}^{-1}\boldsymbol{A})^{-1}\boldsymbol{A}^{T}\boldsymbol{N}^{-1}\boldsymbol{d},$
(37)
where
$\boldsymbol{N}\equiv\langle\boldsymbol{n}\boldsymbol{n}^{T}\rangle-\langle\boldsymbol{n}\rangle\langle\boldsymbol{n}\rangle^{T}$
is the covariance matrix of the generalized noise, which is assumed to be
known. To facilitate comparisons with previous work, we will begin by
considering the estimator $\hat{\boldsymbol{s}}$ of the full global spectrum.
Section 6 returns to the template amplitude constraint.
### 5.1. Spatial Weighting to Recover the Spectrum
L13 used a series of analytical manipulations to show that incorporating
angular information via Equation (37) could in principle lead to large
reductions in foreground contamination. In practice, however, this
prescription is difficult to implement for a number of reasons. First, one may
not possess sufficiently accurate models of the instrument and the foregrounds
to write down the matrix $\boldsymbol{N}$, placing a priori expectations on
the instrument response. This matrix cannot be derived from the data (unlike
$\boldsymbol{\Sigma}$ from previous sections) because it has dimensions
$N_{\nu}N_{\theta}\times N_{\nu}N_{\theta}$ and therefore contains many more
degrees of freedom than the number of measurements $N_{\nu}N_{\theta}$.
Moreover, even if the matrix is somehow known, its large size makes its
inversion in Equation (37) computationally challenging.
To deal with these challenges, consider a modified recipe where the sky maps
are dealt with frequency by frequency (unlike in Equation (37) where all
frequencies and all lines of sight are mixed together) and an estimator for
the global signal at each frequency is formed by spatially averaging with
nontrivial weights (unlike in Equation (6), where different lines of sight
were equally weighted). Let $\boldsymbol{d}_{\beta}$ be a vector of length
$N_{\rm pix}$ that represents the map within $\boldsymbol{d}$ corresponding to
the $\beta$th frequency channel. The global signal at this frequency channel
can then be estimated by computing the weighted average
$\hat{\boldsymbol{s}}_{\beta}=\boldsymbol{w}_{\beta}^{T}\boldsymbol{d}_{\beta},$
(38)
where $\boldsymbol{w}_{\beta}$ is a vector of length $N_{\rm pix}$ with weight
appropriate for the $\beta$th frequency channel. Throughout this paper, we
adopt a convention where summations are written explicitly. Repeated indices
therefore do _not_ imply summations. The variance of this estimator—which we
will seek to minimize—is
$\textrm{Var}(\hat{\boldsymbol{s}}_{\beta})\equiv\langle\hat{\boldsymbol{s}}_{\beta}^{2}\rangle-\langle\hat{\boldsymbol{s}}_{\beta}\rangle^{2}=\boldsymbol{w}_{\beta}^{T}\mathbf{\Phi}_{\beta}\boldsymbol{w}_{\beta},$
(39)
where $\mathbf{\Phi}_{\beta}$ is the $N_{\rm pix}\times N_{\rm pix}$ _spatial_
covariance matrix of the map at the $\beta$th frequency. For our final
estimator to be correctly normalized, we require the weights at each frequency
to sum to unity, i.e., $\boldsymbol{w}_{\beta}^{T}\boldsymbol{e}_{0}=1$. To
obtain the minimum-variance solution for our weights, we use a Lagrange
multiplier $\mu$ to impose our normalization constraint and minimize the
quantity
$L=\boldsymbol{w}_{\beta}^{T}\mathbf{\Phi}_{\beta}\boldsymbol{w}_{\beta}-\mu\boldsymbol{w}_{\beta}^{T}\boldsymbol{e}_{0}.$
(40)
Doing so gives
$\boldsymbol{w}_{\beta}=\frac{\mathbf{\Phi}^{-1}_{\beta}\boldsymbol{e}_{0}}{\boldsymbol{e}_{0}^{T}\mathbf{\Phi}^{-1}_{\beta}\boldsymbol{e}_{0}},$
(41)
and therefore
$\hat{\boldsymbol{s}}_{\beta}=\frac{\boldsymbol{e}_{0}^{T}\mathbf{\Phi}^{-1}_{\beta}\boldsymbol{d}_{\beta}}{\boldsymbol{e}_{0}^{T}\mathbf{\Phi}^{-1}_{\beta}\boldsymbol{e}_{0}}.$
(42)
In words, these weights tell us to whiten the sky maps at each frequency using
the inverse spatial covariance before summing together different lines of
sight and normalizing. Because the whitening takes the form of a nondiagonal
matrix multiplication, this operation not only down-weights brighter parts of
the sky but also makes use of angular correlation information to better
estimate the monopole signal. Tegmark et al. (2003) tackled the transpose of
this problem for CMB foreground removal. Note that in modeling
$\boldsymbol{\Phi}_{\beta}$, it is essential to capture the nonstationary
nature of our galaxy’s foreground emission. It is insufficient, for example,
to describe the galaxy using an angular power spectrum alone, which assumes
statistical isotropy. This would not only be physically unrealistic but would
also make our weights constant across the sky (a fact that can be derived by
applying Parseval’s theorem to our expression for $\boldsymbol{w}_{\beta}$),
defeating the purpose of using spatial weights in the first place.
### 5.2. Spatial Weights Within a Separable Covariance Model
The spatial weight recipe that we have just specified can also be considered a
special limit of the maximum-likelihood estimator, albeit with one small
modification. Suppose that the full $N_{\nu}N_{\theta}\times
N_{\nu}N_{\theta}$ covariance $\boldsymbol{N}$ is separable so that we can
write it as $\boldsymbol{\Sigma}\otimes\boldsymbol{\Phi}$, where
$\boldsymbol{\Sigma}$ is the $N_{\nu}\times N_{\nu}$ spectral covariance from
previous sections, $\boldsymbol{\Phi}$ is an $N_{\rm pix}\times N_{\rm pix}$
spatial covariance, and $\otimes$ is the Kronecker product.
Using the identities
$(\boldsymbol{G}\otimes\boldsymbol{H})^{-1}\equiv\boldsymbol{G}^{-1}\otimes\boldsymbol{H}^{-1}$
and
$(\boldsymbol{G}\otimes\boldsymbol{H})(\boldsymbol{J}\otimes\boldsymbol{K})\equiv\boldsymbol{G}\boldsymbol{J}\otimes\boldsymbol{H}\boldsymbol{K}$,
we have
$\boldsymbol{A}^{T}\boldsymbol{N}^{-1}=\boldsymbol{\Sigma}^{-1}\otimes\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}$
(43)
and
$(\boldsymbol{A}^{T}\boldsymbol{N}^{-1}\boldsymbol{A})^{-1}=\boldsymbol{\Sigma}\otimes(\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}\boldsymbol{e}_{0})^{-1}.$
(44)
Inserting these into Equation (37), the final estimator in this separable
approximation is then
$\hat{\boldsymbol{s}}_{\beta}\Bigg{|}_{\textrm{sep}}=\frac{\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}\boldsymbol{d}_{\beta}}{\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}\boldsymbol{e}_{0}},$
(45)
which is almost identical to Equation (42). If we slightly relax the
assumption of perfect separability by allowing $\boldsymbol{\Phi}$ to acquire
a frequency dependence (so that
$\boldsymbol{\Phi}\rightarrow\boldsymbol{\Phi}_{\beta}$), the correspondence
becomes exact. Including this frequency-dependent nonseparability also allows
us to incorporate the fact that the instrumental beam will generally broaden
toward lower frequencies, in addition to intrinsic nonseparability of the
foregrounds. Section 6.1 discusses nonseparability caused by the differences
between foregrounds and thermal noise.
Interestingly—but unsurprisingly—the spectral covariance $\boldsymbol{\Sigma}$
drops out of the optimal estimator for the global spectrum once separability
is invoked. As discussed in L13, this is due to the fact that once we
spatially average over the maps, we are left with a spectrum consisting of
$N_{\nu}$ numbers. If our final goal is to measure a cosmological spectrum,
that is also of length $N_{\nu}$, the constraint that our estimator be
correctly normalized and unbiased means that there is nothing left to do.
Mathematically, this manifested itself in our derivation when the factors of
$\boldsymbol{\Sigma}$ in Equations (43) and (44) canceled each other out when
forming
$(\boldsymbol{A}^{T}\boldsymbol{N}^{-1}\boldsymbol{A})^{-1}\boldsymbol{A}^{T}\boldsymbol{N}^{-1}$.
If we return to our method in previous sections, however, and attempt to
constrain the amplitude $\alpha$ of a theoretical template, the spectral
covariance $\boldsymbol{\Sigma}$ re-enters the discussion. When constraining
$\alpha$, one is seeking to measure a single number from $N_{\nu}$
measurements, which in general ought to be combined nonuniformly if these
measurements have different error properties (as captured by
$\boldsymbol{\Sigma}$).
## 6\. Analyzing Global $21$ cm Data Cubes
In this section we combine the finite-rank, in situ estimate of spectral
foregrounds with a model of angular correlations. The end product will be an
estimator for the amplitude of a cosmological signal template that can be
applied to experimental global $21$ cm data cubes. The angular part of the
estimator uses assumed spatial distributions to improve the amplitude
estimate. In contrast, the central aspect of this estimator is that it does
not assume a set of _spectral_ modes or spectral covariance of the data.
It is notationally convenient to write the $N_{\rm pix}N_{\nu}$-long data
matrix $\boldsymbol{d}$ by reshaping it into an $N_{\nu}\times N_{\rm pix}$
matrix $\boldsymbol{Y}$. This “data cube” is a stack of maps at all of the
observed frequencies. The original data vector $\boldsymbol{d}$ can be
recovered through the serializing “vec()” operation ${\rm
vec}(\boldsymbol{Y})=\boldsymbol{d}$. This data cube form naturally
accommodates weighting operations by the separable covariance. A left-
multiplication acts on the spectral direction, and a right-multiplication acts
on the spatial direction.
The essential weighting operation on the full data set is
$\boldsymbol{N}^{-1}\boldsymbol{d}$, which under the separable covariance
assumption becomes
$\boldsymbol{N}^{-1}\boldsymbol{d}=(\mathbf{\Sigma}\otimes\mathbf{\Phi})^{-1}\boldsymbol{y}=\mathbf{\Sigma}^{-1}\boldsymbol{Y}\mathbf{\Phi}^{-1}$
(46)
using identities of the Kronecker and vec operations.
A model for the noise on $\boldsymbol{Y}$ is to draw some matrix
$\boldsymbol{E}$ with shape $N_{\nu}\times N_{\rm pix}$ of normal deviates
from $N(\mu=0,\sigma^{2}=1)$ and then correlate the spectral and spatial parts
as $\mathbf{\Sigma}^{1/2}\boldsymbol{E}\mathbf{\Phi}^{1/2}$ to give a noise
realization drawn from the covariance $\mathbf{\Sigma}\otimes\mathbf{\Phi}$.
The data model in this presentation is
$\boldsymbol{Y}=\alpha\boldsymbol{x}\boldsymbol{e}_{0}^{T}+\mathbf{\Sigma}^{1/2}\boldsymbol{E}\mathbf{\Phi}^{1/2}.$
(47)
This is a specialized case of the “growth curve” model (Kollo & von Rosen,
2005), reviewed in Appendix B. We can transfer the separable method developed
in Section 5.2 to the present case of constraining the template amplitude
$\alpha$ by letting $\boldsymbol{s}\rightarrow\alpha$ and
$\boldsymbol{A}=\boldsymbol{1}\otimes\boldsymbol{e}_{0}\rightarrow\boldsymbol{x}\otimes\boldsymbol{e}_{0}$
(because we are now assuming a form for $\boldsymbol{x}$ as a theoretical
template). Here, the maximum-likelihood estimate for the amplitude of some
spectral template $\boldsymbol{x}$ is
$\hat{\alpha}=(\boldsymbol{x}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{x})^{T}\boldsymbol{x}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{Y}\mathbf{\Phi}^{-1}\boldsymbol{e}_{0}(\boldsymbol{e}_{0}^{T}\mathbf{\Phi}^{-1}\boldsymbol{e}_{0})^{-1}.$
(48)
This applies the familiar maximum-likelihood estimators for the spatial
monopole (with respect to covariance $\mathbf{\Phi}$) and spectral template
(with respect to covariance $\mathbf{\Sigma}$) to the right and left side of
our data cube $\boldsymbol{Y}$.
Following Section 3.2, we can seek to replace $\mathbf{\Sigma}$ with a
$(\nu,\nu^{\prime})$ covariance measured in the data, $\mathbf{\hat{\Sigma}}$.
In Section 3.2, there was no information distinguishing different lines of
sight, so the mean subtraction in Equation (19) simply used the unweighted
average of $\boldsymbol{y}_{i}$. We now have information about the angular
covariance and can improve the mean spectrum estimation by including the
spatial weight, as $\boldsymbol{Y}_{\rm
sub}^{T}=(1-\boldsymbol{e}_{0}\boldsymbol{w}^{T})\boldsymbol{Y}^{T}$. Here
$\boldsymbol{w}=\boldsymbol{\Phi}^{-1}\boldsymbol{e}_{0}(\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}\boldsymbol{e}_{0})^{-1}$
is an $N_{\rm pix}$-long weight map. Another way of looking at the operation
$(1-\boldsymbol{e}_{0}\boldsymbol{w}^{T})$ is that it subtracts the monopole
spectrum out of the map so that the eigenvectors of $\mathbf{\hat{\Sigma}}$
will only contain spatially variable spectral modes, not the monopole.
Procedurally, one (1) applies the angular weight $\boldsymbol{\Phi}^{-1}$, (2)
dots against the spatial monopole template $\boldsymbol{e}_{0}$, (3) accounts
for the impact of the spatial weight through
$(\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}\boldsymbol{e}_{0})^{-1}$, and
4) projects this spectral template as a monopole across the map (under
$\boldsymbol{e}_{0}$) and subtracts this from the measured map. Using the
mean-subtracted map, we can find the weighted covariance as
$\mathbf{\hat{\Sigma}}=\boldsymbol{Y}_{\rm
sub}\boldsymbol{\Phi}^{-1}\boldsymbol{Y}_{\rm sub}^{T}~{}~{}{\rm
with}~{}~{}\boldsymbol{Y}_{\rm
sub}^{T}=(1-\boldsymbol{e}_{0}\boldsymbol{w}^{T})\boldsymbol{Y}^{T}.$ (49)
The estimator for the monopole spectrum template amplitude is then
$\hat{\alpha}=(\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{x})^{-1}\boldsymbol{x}^{T}\mathbf{\hat{\Sigma}}^{-1}\boldsymbol{Y}\boldsymbol{w}$
(50)
Kollo & von Rosen (2005) show that this choice of weight yields the maximum-
likelihood $\hat{\alpha}$. The inverse $\mathbf{\Phi}^{-1}$ of the $N_{\rm
pix}\times N_{\rm pix}$ spatial covariance appears explicitly only in the
construction of the sample variance. Elsewhere, it is collapsed onto the
single simple weight map $\boldsymbol{w}$. Section 6.2 describes choices for
the form of $\boldsymbol{\Phi}$. There we advocate a diagonal form of
$\boldsymbol{\Phi}$ based on thermal noise, whose amplitude is determined by a
position-dependent sky temperature and integration time. Diagonal
$\boldsymbol{\Phi}$ makes estimation of $\mathbf{\hat{\Sigma}}^{-1}$
computationally simple.
We will argue in Section 6.2 that the angular weights only need to treat the
residual foregrounds and thermal noise variations across the sky. We choose to
let $\mathbf{\hat{\Sigma}}^{-1}$ do the legwork of determining the spectral
modes to remove from the data itself. In this role,
$\mathbf{\hat{\Sigma}}^{-1}$ accounts for instrumental systematics, whereas
$\mathbf{\Phi}^{-1}$ can improve the estimate of $\hat{\alpha}$ by using a
foreground model to down-weight noisier regions.
Similar to the discussion in Section 3.3, we can replace the
$\mathbf{\hat{\Sigma}}^{-1}$ operation by a projection that removes a finite
number of contaminated spectral modes. This is equivalent to (1) identifying a
data subspace parallel to foregrounds that is contaminated and fully removed
and (2) keeping a data subspace orthogonal to foregrounds.
Let $\boldsymbol{F}$ contain the largest eigenvectors of the sample covariance
$\mathbf{\hat{\Sigma}}$. Then
$\hat{\alpha}=\frac{\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})}{\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}}\boldsymbol{Y}\boldsymbol{w}.$
(51)
The denominator
$\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}$ is a
scalar normalization.
This is analogous to Equation (34), except that instead of a simple
$\boldsymbol{\bar{y}}$, each spatial slice of the map $\boldsymbol{Y}$ is
weighted to find the mean as $\boldsymbol{Y}\boldsymbol{w}$.
Using this notation, we can easily prove that constant passband calibration
errors do not increase the rank of foreground spectral modes. Let
$1+\delta\boldsymbol{c}$ be an $N_{\nu}$-long vector that represents constant
miscalibration of the passband. If a constant calibration error multiplies all
lines of sight, $\boldsymbol{C}={\rm diag}(1+\delta\boldsymbol{c})$ modifies
the data as $\boldsymbol{C}\boldsymbol{Y}$. So long as the response is not
zeroed out at some frequencies, $\boldsymbol{C}$ is invertible, and ${\rm
rank}(\boldsymbol{C}\mathbf{\hat{\Sigma}}\boldsymbol{C}^{T})={\rm
rank}(\mathbf{\hat{\Sigma}})$ for invertible $\boldsymbol{C}$. Intuitively, a
constant passband error can be thought of as just a per-frequency rescaling of
units. Because such a rescaling does not affect the rank of the foregrounds,
there is no increase in the number of foreground modes that need to be
constrained. Our self-discovery scheme for spectral modes will therefore be
just as effective when constant passband miscalibrations are present.
### 6.1. Deriving the Final Estimator: A Two-component Covariance
In deriving Equation (51), the key assumption that we made was that the total
covariance is separable into spatial and spectral components. This confuses
foregrounds and thermal noise, which each have very different angular
distributions. For example, we know that a synchrotron spectral mode is
associated with broad spatial distributions of galactic emission. In contrast,
thermal noise is always uncorrelated across angular pixels, but may change in
variance across the sky. In the limit that spectral foreground subtraction is
perfect, only Gaussian thermal noise will remain. In contrast, the separable
assumption would dictate that it has approximately the same angular
correlations as the galaxy. This issue ultimately translates into the angular
weight $\boldsymbol{w}$ that should be applied.
The solution to this problem is to let the total covariance be nonseparable
but comprised of independently separable foreground and noise components.
Here, $\boldsymbol{N}=\boldsymbol{N}_{f}+\boldsymbol{N}_{g}$ where
$\boldsymbol{N}_{f}=\mathbf{\Sigma}_{f}\otimes\mathbf{\Phi}_{f}$ and
$\boldsymbol{N}_{g}=\mathbf{\Sigma}_{g}\otimes\mathbf{\Phi}_{g}$. For
concreteness, imagine that the “$f$” component consists of modeled foreground
modes in $\boldsymbol{F}$, and the “$g$” component consists of any residuals
after the projection operation. This is exactly analogous to the
$\boldsymbol{F}/\boldsymbol{G}$ split in Rao (1967) described in Section 3.3.
If our discovered foreground modes happen to perfectly capture the true
foregrounds, then the “g” component would consist only of Gaussian thermal
noise. In general, however, we need not make this assumption.
The central problem with the model
$\boldsymbol{N}=\mathbf{\Sigma}\otimes\mathbf{\Phi}$ is that
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$ only acts on the $\mathbf{\Sigma}$ part
but not on $\mathbf{\Phi}$. In contrast, in the two-component model, when
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$ acts on the data, the remaining
covariance is only $\mathbf{\Sigma}_{g}\otimes\mathbf{\Phi}_{g}$, e.g., it
correctly ascribes $\mathbf{\Phi}_{f}$’s spatial correlations to the modes
that $\boldsymbol{F}$ removes.
We can now reassess the $\boldsymbol{w}$ in Equation (51) that yields the
minimum-variance estimate of $\alpha$. To clarify the discussion, let
$\boldsymbol{q}\equiv[\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}]^{-1}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}$.
The first step of Equation (51) in this notation is to form
$\boldsymbol{q}^{T}\boldsymbol{Y}$, which can be interpreted as a series of
estimators for $\alpha$, one for each line of sight.
With our two-component covariance, a little algebra reveals that the variance
to minimize is given by
$\textrm{Var}(\hat{\alpha})=(\boldsymbol{q}^{T}\boldsymbol{\Sigma}_{f}\boldsymbol{q})(\boldsymbol{w}^{T}\boldsymbol{\Phi}_{f}\boldsymbol{w})+(\boldsymbol{q}^{T}\boldsymbol{\Sigma}_{g}\boldsymbol{q})(\boldsymbol{w}^{T}\boldsymbol{\Phi}_{g}\boldsymbol{w}).$
(52)
The first term is zero by construction because
$\boldsymbol{q}^{T}\boldsymbol{\Sigma}_{f}\propto(\boldsymbol{1}-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{F}=0$.
Minimizing the remaining term subject to the constraint that the weights sum
to unity requires minimizing the quantity
$L=(\boldsymbol{q}^{T}\boldsymbol{\Sigma}_{g}\boldsymbol{q})(\boldsymbol{w}^{T}\boldsymbol{\Phi}_{g}\boldsymbol{w})-\mu\boldsymbol{w}^{T}\boldsymbol{e}_{0},$
(53)
where $\mu$ is again a Lagrange multiplier. Because
$\boldsymbol{q}^{T}\boldsymbol{\Sigma}_{g}\boldsymbol{q}$ is just a constant,
it can be absorbed into $\mu$ by a simple rescaling. One then sees that $L$ is
identical to Equation (40), except that $\boldsymbol{\Phi}_{g}$ (the spatial
covariances of the residuals) takes the place of $\boldsymbol{\Phi}_{\beta}$
(the full spatial covariance at the $\beta$th frequency). We may thus import
our previous solution for the spatial weights to find
$\boldsymbol{w}=\boldsymbol{\Phi}^{-1}_{g}\boldsymbol{e}_{0}(\boldsymbol{e}_{0}^{T}\boldsymbol{\Phi}^{-1}_{g}\boldsymbol{e}_{0})^{-1}$
(54)
In conclusion, we see that a simple reinterpretation of Equation (51) is all
that is required to accommodate a two-component covariance model. The new
weights $\boldsymbol{w}$ reflect the spatial covariance of residuals after the
frequency mode subtraction rather than the spatial covariance of the sky
itself. In spirit, this is reminiscent of the approach in L13, where a best-
guess model of the foregrounds was first subtracted from the sky maps before
the variance of the residuals was minimized. The difference here is that our
best-guess cleaning is based on the data rather than on a model. In addition,
the estimator operates with separable spectral and spatial steps that are
computationally trivial (a feature that we have managed to retain even with
the _non_ separable two-component covariance of this section).
### 6.2. Angular Weighting
Angular weighting is applied after the majority of the foreground covariance
has been removed by the spectral mode subtraction
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$. In the language of the previous
section, it is not the full angular covariance $\boldsymbol{\Phi}$ of the
foregrounds that acts as a weight, but instead just $\boldsymbol{\Phi}_{g}$,
the angular covariance of the residuals.
If instrumental systematics are well controlled, the eigenvalue spectrum of
the $\nu-\nu^{\prime}$ covariance will fall rapidly and settle to a thermal
noise floor. Once the full complement of spectral modes down to the noise
floor are removed, the map would then be dominated by thermal noise. The
optimal angular weight in that case corresponds to dividing by the foreground
template squared. Rather than suppressing any foregrounds, this operation
simply weighs more strongly against regions of higher thermal noise due to
brighter foregrounds that have boosted the $T_{\rm sys}$ (assuming foregrounds
dominate the system temperature in these frequencies).
Our primary argument throughout this paper has been that foreground models
should not be trusted to define spectral modes for foreground cleaning because
they will invariably be too simple due to instrumental effects. In the case of
angular weighting of thermal noise, though, the operation of dividing by the
foreground template simply needs to be a best-guess thermal noise weight.
Slight misspecification reduces the optimality of $\hat{\alpha}$ but does not
introduce bias.
In experimental reality, the eigenvalue spectrum may fall more gradually and
merge into the noise floor. This could be caused by the calibration, beam and
polarization effects described in Section 7. As the eigenvalue falls to the
thermal noise floor, the thermal noise spectral modes are strongly mixed with
foreground modes, making the foregrounds difficult to cleanly separate.
Further, Section 7.4 describes how the number of foreground modes to remove
may be ambiguous because of thermal noise. An $\ell$-by-$\ell$ weighting of
the residuals then has the potential to improve the estimate of $\hat{\alpha}$
by down-weighting angular correlations that are consistent with residual
foregrounds. Detecting some angular power spectrum in the
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$-cleaned map that is measurably
different from Gaussian thermal noise could indicate that a better weight
$w_{\ell}$ is possible.
Given that the integration depth of an experiment is at the level of the
cosmological signal (which is much smaller than the foregrounds), by
definition most of the foreground spectral modes will be well measured.
Deleting this set of spectral modes will remove the vast majority of
foregrounds. Even if there are residual foregrounds mixed with the thermal
noise, we expect that the optimality of the angular weight would not depend
strongly on the assumed foreground spatial template.
Following L13, we split the problem of nonstationarity and angular correlation
of the angular covariance by writing
$\boldsymbol{\Phi}_{g}={\rm diag}(\boldsymbol{m})\boldsymbol{Q}{\rm
diag}(\boldsymbol{m})^{T},$ (55)
where $\boldsymbol{m}$ is the foreground mean map. This is equivalent to
dividing the foreground by the expected mean and then encoding the angular
correlations of that normalized map in the matrix $\boldsymbol{Q}$. Let these
correlations be full sky, isotropic, and diagonal. If $\boldsymbol{T}$ is a
matrix that converts from real space to spherical harmonics, then the
correlations will be diagonal
$\mathbf{\Lambda}=\boldsymbol{T}\boldsymbol{Q}\boldsymbol{T}^{T}$, where
$\mathbf{\Lambda}$ is just some function of $\ell$. The operation
$\mathbf{\Phi}^{-1}$ can be interpreted algorithmically as
* •
Divide each spatial slice of the data by a foreground model map (a synchrotron
template).
* •
Transform to spherical harmonic space and deweight $\ell$-by-$\ell$ by
$w_{\ell}$ (defined below).
* •
Transform back to map space and divide by the foreground model map again.
A simple, two-parameter model for the correlations is to let the residuals
have some amplitude $\xi$ relative to thermal noise and correlation length
$\sigma$, as
$w_{\ell}=\left[\xi
e^{-\sigma^{2}\ell(\ell+1)/2}+e^{\theta_{b}\ell(\ell+1)}\right]^{-1}.$ (56)
Here we have assumed that the input data are deconvolved by the beam before
applying angular weights (dividing by $B_{\ell}$). The term
$e^{\theta_{b}\ell(\ell+1)}$ describes thermal noise in that case. By picking
the $\sigma$ and $\xi$ to give the lowest error, we may give preference to
parameters where the error spuriously scatters low. A better alternative is to
use the known instrumental thermal noise to inform $\xi$ and assume residuals
in the data are correlated across the beam scale $\theta_{b}$ (not determining
$w_{l}$ based on the error of $\hat{\alpha}$). Ideally, the residuals would be
dominated by nearly Gaussian thermal noise, and $w_{\ell}=1$ would be a
sufficiently optimal weight.
Figure 8.— Top: spatial weights at $80\,\textrm{MHz}$ for noise-dominated
residuals ($\xi=0.5$; $\sigma=20^{\circ}$). The main purpose of spatial
weighting in this regime is to average down noise, hence the weight is
positive across the sky. Bright regions of the sky (which cause more
instrumental noise for sky-noise-dominated instruments) are down-weighted.
Bottom: spatial weights at $80\,\textrm{MHz}$ for foreground-dominated
residuals ($\xi=50$; $\sigma=20^{\circ}$). In this regime, the spatial weights
play a role in foreground subtraction, and thus the weights are both positive
and negative, as different parts of the sky are differenced to mitigate
residual foregrounds. The beam size assumed in both cases is
$\theta_{b}=10^{\circ}$. The weights are defined to sum to one, but only the
shape, not normalization, is relevant.
In Figure 8 we show example spatial weights computed from Equations (54),
(55), and (56) in two different regimes. The top panel shows a noise-dominated
case ($\xi=0.5$), and the bottom panel shows a foreground residual-dominated
case ($\xi=50$). In the noise-dominated case, the spatial weights are designed
to simply average down the noise, and are therefore all positive. The galactic
plane and the bright point sources are seen to receive almost zero weight
because they contribute the most thermal noise in a sky-noise-dominated
instrument. In the foreground-dominated case, the spatial weights attempt to
further suppress residual foreground contamination. There are thus both
negative and positive weights because different parts of the sky can be
differenced to subtract off foregrounds. In both regimes, the right half of
the sky is given more weight because the galaxy is dimmer there (see Figure
2).
An important property of our spatial weighting scheme is that the
normalization is irrelevant and only the shape matters. This is a particularly
attractive property because most models of the sky at our frequencies of
interest are based on interpolations and extrapolations from other
frequencies. The amplitude may have large uncertainties, but the shape
information is likely to be much more reliable. If one is particularly
confident about the available shape information, it may be prudent to go one
step further and set $w_{\ell=0}=0$ by hand. Examining Equations. (55) and
(56) reveals that this projects out any component of the measured sky that has
precisely the same spatial shape as our foreground model. Note that even
though this is the $\ell=0$ weight, we do not destroy the global monopole
signal that we seek to measure because the $w_{\ell}$ act in a prewhitened
space _after_ dividing by $\boldsymbol{m}$. The global signal therefore takes
the form $1/\boldsymbol{m}$ and is no longer just the $\ell=0$ mode. It is
instead spread out over a wide range of $\ell$ values. Setting $w_{\ell=0}=0$
then sacrifices the sensitivity in one of the modes within this range,
resulting in slightly increased error bars as one deviates slightly from the
optimal prescription described above. However, this may be a cost that is
worth bearing for the sake of robustness in an aggressive campaign against
foreground systematics.
### 6.3. Summary of the Proposed Algorithm
The final algorithm suggested here for analyzing global $21$ cm data cubes is
* •
Use a spatially weighted average to project the monopole spectrum out of the
map and find the $(\nu,\nu^{\prime})$ sample covariance.
* •
Find the largest eigenvectors of the sample covariance and project them out of
the map.
* •
Combine lines of sight using a prescription for the angular correlation of the
residuals (which conservatively is to divide by the synchrotron template
squared).
* •
Dot this against the signal template to find the amplitude and perform error
analysis.
While this algorithm is intuitive, our goal here has been to describe some of
the implicit choices: (1) discovering frequency modes within the data (and
implications for errors), (2) choosing to remove the part of the signal
parallel to the foreground modes, and (3) assuming a two-component separable
form for the thermal noise and foregrounds, respectively. The next section
describes several considerations for using estimators of this type and
challenges in global $21$ cm signal measurement.
## 7\. Considerations for Analyzing Global $21$ Maps
In the next few subsections, we will consider (1) the notion of resolution
elements and the choice of beam size, (2) combined passband stability and
resolution considerations, (3) challenges in error estimation from residual
monopoles, (4) determination of the number of modes to remove, (5) extensions
to simple amplitude constraints, (6) applications of our methods to the pre-
reionization absorption feature, and (7) complications in foreground
mitigation that result from Faraday rotation.
### 7.1. Resolution Elements and Foreground Modes
In Section 3, the data were spectra of independent lines of sight. There, the
number of foreground modes removed ($N_{\rm fg}$) compared to the number of
sight lines ($N_{\theta}$) entered the errors as $\sim(N_{\rm
fg}-N_{\theta})^{-1}$. The sense of $N_{\theta}$ is more complex in our
present case of an $N_{\nu}\times N_{\rm pix}$ $\boldsymbol{Y}$ data cube
where the spatial part of the sky signal is convolved by an instrumental
beam,.
Intuition suggests that the number of independent sight lines is approximately
the number of beam spots on the sky, but this is incorrect. If each spatial
slice of a noiseless full-sky map is written in terms of its spherical
harmonics $T_{\ell m}$, convolution by the beam is multiplication by
$B_{\ell}$. Unless the beam falls to zero somewhere, this operation is
invertible and so does not modify the rank of $\mathbf{\hat{\Sigma}}$. The
rank of $\boldsymbol{Y}$ is $N_{\rm fg}$ so long as $N_{\rm pix}>N_{\rm fg}$
and $N_{\nu}>N_{\rm fg}$. Hence, all of the spectral modes can still be
recovered from the noiseless beam-convolved map.
With wider beams, higher eigenvalue spectral eigenmodes (typically the less-
smooth modes) are washed out by the beam. These modes do not disappear
entirely, but their amplitude diminishes. We can develop a rough rule for this
effect by finding the eigenvalue spectrum of full-rank white noise that is
convolved by a given beam size. A fitting form for the suppression of
eigenvalues (normalized to give no suppression for infinitely fine beams) is
$b_{n}\approx 10^{-(\theta_{\rm FWHM}/300^{\circ})^{2}f_{\rm sky}^{-1}n}$, as
a function of eigenvalue $n$ and full width at half max (FWHM) of the beam.
At first sight, it would appear that a more rapidly falling eigenvalue
spectrum is beneficial because it means that the foregrounds are better
described by fewer covariance modes. When the suppression occurs because of
beam convolution, however, it can be problematic in the presence of
instrumental noise. With instrumental noise, the information about a mode may
be suppressed below the noise floor. That foreground can therefore no longer
be cleanly detected and removed and may contaminate the signal and lead to
larger errors. Such detection and removal is potentially crucial because
higher eigenvalues can still greatly exceed the signal (even after beam
suppression) because of the hierarchy of scales.
We therefore come to the conclusion that even if angular resolution does not
impose a hard limit on the number of foreground modes that may be discovered
(as suggested in the case of independent sight lines), it does put a practical
limit on the efficacy of the cleaning. Note that the eigenvalues of
$\mathbf{\Sigma}=\mathbf{\Sigma}_{\rm fg}+\mathbf{\Sigma}_{\rm inst}$ will be
different from $\mathbf{\Sigma}_{\rm fg}$. Adding noise has the effect of
mixing the high-variance foreground modes with the thermal noise modes, making
it harder to cleanly separate foregrounds.
A rule of thumb for finding the number of modes discoverable before they are
confused by noise is to plot the eigenvalue spectrum of noiseless foregrounds
and the thermal noise, then find where they intersect. Modes some margin above
this point will still be well measured. Consider a normalized eigenvalue
spectrum of foregrounds that follows $\lambda_{n}\approx b_{n}\cdot 10^{-an}$
(Liu & Tegmark, 2012). The eigenvalues of thermal noise are relatively flat in
$n$ in comparison. Finding the intersection with the noise level, the number
of foreground modes that can be learned in the data is
$N_{\rm fg}\sim\log({\rm FNR})\left[\left(\frac{\theta_{\rm
FWHM}}{300^{\circ}}\right)^{2}f_{\rm sky}^{-1}+a\right]^{-1},$ (57)
where FNR is the foreground-to-noise ratio (in map variance). In our simple
models here, $a$ may be a factor of a few, for only a few modes in total. In
this idealized setting, the rather lenient scale of $300^{\circ}$ in our
fitting form for $b_{n}$ means that an instrument could cover the full sky at
poor resolution and still discover all of the foreground modes. In reality,
instrumental effects will drive a higher number of required modes. For
example, Switzer et al. (2013) had to remove tens of modes to clean
foregrounds even though only approximately four intrinsic synchrotron modes
were expected.
An advantageous factor here is that the sky signal is convolved by the beam
whereas instrumental thermal noise is not. Hence, thermal noise can be
estimated within the data using $N_{\rm pix}$ independent samples. In an ideal
experiment, all of the foreground modes can be discovered and subtracted in
$\boldsymbol{F}$ so that only thermal noise remains. Then errors could be
assessed from the map’s thermal noise, and signal loss could be corrected
against the subtracted modes as
$\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}$.
It is generally beneficial to cover larger areas with finer resolution to
better constrain and discover foreground modes. Higher resolution also helps
focus deweighting of particular contaminated sky regions. An important
exception to this occurs when the number of foreground modes scales with the
FWHM or sky fraction. Both Faraday rotation and passband instability will
generally increase the number of modes needed for studies on larger areas with
smaller beams. Each new beam that is observed could see a new rotation measure
or be observed with a different passband.
### 7.2. Summarizing the Effects of Passband Stability and Angular Resolution
Figure 9.— Error of the recovered cosmological template amplitude as a
function of calibration stability and instrumental resolution where the
fiducial amplitude is $\alpha=1$. The thermal noise is set so that in the
limit of no calibration or beam effects the amplitude can be constrained to
$5\sigma$. Here we consider foregrounds with four intrinsic spectral degrees
of freedom. The constraint degrades for FWHM above $\sim 50^{\circ}$ and
$10^{-4}$ stability in passband calibration. At low resolution, the modes
cannot be discovered. Calibration stability worse than the thermal noise
produces many more modes that must be constrained and removed. The Monte Carlo
estimates in the upper right become very noisy, and we saturate the color
scale. Figure 10.— Spectral eigenmodes of the foreground model described in
Section 2. The second, third, and fourth modes are $3\times 10^{-4}$, $8\times
10^{-7}$, and $6\times 10^{-8}$ down from the first mode (in variance),
respectively.
Figure 9 summarizes two of the main points of our paper. Here, we use the
galaxy model described in Section 2 to perform a Monte Carlo simulation for
the errors on $\alpha$ as a function of calibration stability and instrument
resolution. Figure 10 shows the four largest modes of this extended GSM model.
All further modes of the model itself (without noise) are negligible at the
level of machine precision. Note that in contrast to the simple pedagogical
models in Section 3 (with three modes), the full data cube simulation is more
representative of the synchrotron sky.
The sense of the passband calibration stability in our Monte Carlo is pixel to
pixel (e.g., not beam convolved), so it is not affected by the beam
convolution operation. We set the thermal noise level to produce a $5\sigma$
detection of $\alpha$ in the absence of beam and calibration effects. For
beams that are too large, not all foreground modes can be recovered and
subtracted. In the simulations here with four modes, the constraint on
$\alpha$ worsens gradually as the FWHM exceeds $\sim 50^{\circ}$. For
calibration stabilities worse than $\sim 10^{-4}$, each line of sight responds
differently to bright foregrounds, so each requires a new spectral function to
remove. As soon as these passband calibration perturbations exceed the noise
floor, they rapidly degrade the constraint.
Decreasing the thermal noise through longer integration times loosens the beam
resolution requirement (because foreground modes are better measured). It does
not modify the passband stability requirement. Passband stability is driven by
the separation in scales between the foregrounds and the cosmological signal
and the fact that each line of sight has a different response to the
foregrounds. Even if the modes are well measured relative to thermal noise,
they quickly overwhelm the signal.
Figure 11.— Spectral eigenmodes of foregrounds in a diffraction-limited
instrument where a Gaussian ${\rm FWHM}\propto\lambda$. Here, the beam’s
frequency dependence mixes spatial structure into spectral structure. This
dramatically proliferates the number of spectral modes that are discovered and
need to be removed, resulting in the loss of essentially all of the signal. A
treatment for this effect in a mapping experiment is to convolve to a common
resolution using the measured beams. Imperfections in the beam model will then
generically produce additional modes of the spectral covariance.
Throughout, we have assumed that the angular resolution at all frequencies is
constant. This is difficult to arrange experimentally. More typically, ${\rm
FWHM}\propto\lambda$ by diffraction. In the frequency range considered here, a
diffraction-limited FWHM would change by a factor of 2.5. Variation of the
beam will mix spatial structure into frequency structure because different
beams at each frequency see different angular contamination patterns (Morales
et al., 2012). Implicitly, we have assumed that the beam is known well enough
to convolve all of the frequency slices to a common resolution. In contrast,
if no correction is performed, beam effects will proliferate the number of
modes that need to be removed. Figure 11 shows the first $10$ bright spectral
modes from the simulations described in Section 2, which originally had four
modes. These new spectral modes describe the spatial structure that is mixed
into the spectral direction. If these spectral modes had to be removed, most
of the signal would also be projected out. The frequency-dependent beam must
therefore be corrected to achieve efficient cleaning. Errors in the beam model
executing this correction then translate into new spectral contamination
modes.
The tolerance of beam models is experiment-dependent and is beyond the scope
of the present work but must be carefully considered. We emphasize that
frequency-dependent beam effects impact any global $21$ cm experiment, whether
they choose to take advantage of angular information or not. The new spectral
mixing modes reflect a best effort of our algorithm to describe a new spectral
covariance that must be cleaned to reach the $21$ cm monopole. Careful
experimental design and measurement of the beam will control these covariance
modes (which remain inaccessible in a single-beam measurement).
Here we have limited the intrinsic foreground rank to four. In reality, it may
be discovered that foregrounds on the sky have higher rank, pushing the
resolution constraint. Also, we assume a full sky here—an experiment covering
a smaller fraction of the sky would need $f_{\rm sky}^{-1}$ more resolution in
our simple approximations here.
### 7.3. Challenges in Error Estimation
In the absence of prior information, monopole contamination is
indistinguishable from the signal. This contamination may arise from a
synchrotron monopole or from an additive $T_{\rm rx}(\nu)$ of the receiver
that is temporally constant but has some spectral structure. The foreground
modes are estimated from the $(\nu,\nu^{\prime})$ sample covariance
$\mathbf{\hat{\Sigma}}$ of the map with the mean spectrum removed. The modes
in $\boldsymbol{F}$ represent the frequency components of the spatially
fluctuating part of the foregrounds, and the operation
$1-\boldsymbol{F}\boldsymbol{F}^{T}$ is uninformed by a monopole signal.
The monopole foregrounds may have significant overlap with the spectral
functions of the spatially fluctuating foregrounds. In our simulations here,
the foregrounds are randomly drawn from a fixed set of spectral modes. For
example, if there is a spatially varying synchrotron foreground with a
spectral index of $-2$, that synchrotron emission will also be subtracted from
the monopole by $1-\boldsymbol{F}\boldsymbol{F}^{T}$ .
In reality, we cannot rule out the possibility of a foreground monopole that
remains even after cleaning the spatially-varying synchrotron modes. Further,
the spectral pattern of a constant instrumental $T_{\rm rx}(\nu)$ will never
be discoverable by mapping across the sky and could be confused with the
global signal. These present serious challenges for rigorous error estimation.
Performing the angular operations on the right of Equation (51) first, the
noise terms of $\hat{\alpha}$ can be isolated as
$\hat{\alpha}=\alpha|_{\rm
true}+\frac{\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})}{\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{x}}(\boldsymbol{n}_{\rm
fg}+\boldsymbol{n}_{\rm inst})$ (58)
where $\boldsymbol{n}_{\rm fg}$ and $\boldsymbol{n}_{\rm inst}$ are the
foregrounds and thermal noise of the monopole through
$\boldsymbol{Y}\boldsymbol{w}$. We lump the constant $T_{\rm rx}(\nu)$ in with
$\boldsymbol{n}_{\rm fg}$ rather than introducing a new term.
While the thermal noise contribution
$\propto\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{n}_{\rm
inst}$ has a well-defined distribution and can easily be included in the
errors,
$\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{n}_{\rm
fg}$ is much more challenging. If
$\boldsymbol{x}^{T}(1-\boldsymbol{F}\boldsymbol{F}^{T})\boldsymbol{n}_{\rm
fg}$ were always positive, then $\hat{\alpha}$ could be interpreted as an
upper bound. This practice is common in treating foregrounds to the power
spectrum because their contribution is always positive in the quadratic
quantity Switzer et al. (2013); Parsons et al. (2014); Dillon et al. (2014);
Paciga et al. (2013). Here, we have no guarantee that the residual foreground
monopole dotted into the signal is positive. Recall that even though
synchrotron radiation may be a strong power law, the residuals here will
largely represent unknown instrumental factors that could dot with an
arbitrary sign with the signal template. It has long been known that higher-
order polynomials lead to better subtractions. This suggests that residuals
will generally produce two-sided errors. Further, any monopole residual
foregrounds are likely to be non-Gaussian, inheriting from the non-Gaussianity
of the bright foregrounds.
These “monopole” foregrounds do not strictly need to be a monopole on the
sky—they only have to be constant over the area of the survey, if that only
covers part of the sky. This again bolsters the argument that global $21$ cm
experiments should observe as much of the sky as possible with modest
resolution to maximize their sensitivity to spatial variations that could lead
to significantly cleaner signal recovery.
Barring any prior information about the foregrounds or instrumental response,
the best ways to immunize the analysis against this response are (1) to
document that the spectral cube cleaned with
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$ is consistent with Gaussian thermal
noise and (2) to cover a large area of the sky with modest resolution to
capture additional spatial variation of foregrounds. Additional information
about spectral smoothness and instrument response can be used (Bowman &
Rogers, 2010; Voytek et al., 2014) to further clean the spectral monopole.
Quoted errors would be based only on thermal noise, and boosted to account for
signal lost in the cleaning process, Equation (35). Another possibility would
be to conduct, for example, surveys of the north and south celestial poles
with different receiver architectures. If these were performed independently
and had the same result up to thermal errors, it could lend some confidence
that any residual monopoles in the respective surveys are small compared to
noise.
### 7.4. Determining the Number of Modes to Remove
Figure 12.— Fraction of the root mean squared of the reionization signal
remaining after the first four foreground modes are removed. The signal is
given by Equation (3) for a simple reionization scenario with central redshift
$z_{r}$ and width $\Delta z$. The first mode is a power law (see Figure 10),
and most of the signal remains for these reionization scenarios. The other
three modes carry more redshift information and tend to express signal
variations at high redshift better. In all cases, as the signal becomes
smoother ($\Delta z$ increasing), it becomes harder to differentiate from the
foreground modes.
Figure 12 shows the fractional rms of the cosmological reionization signal
remaining after removing the four foreground modes of the extended GSM
(Section 2), shown in Figure 10. We use the operation
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$ described previously. The foreground
spectral functions overlap strongly with the signal, especially for slower,
spectrally smoother reionization histories (high $\Delta z$ in the plot).
Nulling four foregrounds causes significant signal loss over a range of
reionization models. Any instrumental response that increases the rank of
these foregrounds will result in additional signal loss from the new modes it
introduces.
In the case of these simulations, we know the precise number of spectral
degrees of freedom. In experimental practice, the number of contaminated modes
cannot be determined directly. This can be made especially ambiguous by an
eigenvalue spectrum that drops slowly rather than reaching a clear noise
floor.
A central challenge of any global $21\,\textrm{cm}$ signal estimator is
determining the number of modes to remove (Rao, 1967; Kollo & von Rosen,
2005). In similar methods applied to the $21$ cm autopower (Switzer et al.,
2013), an advantage is that the bias from foregrounds is purely additive. In
this case, an experiment can report the constraint with errors as a function
of modes removed. This will generically fall as more contaminated modes are
removed. This decline will level out if the foreground have been discovered.
As further modes are removed, the errors will increase due to cosmological
signal loss, which is self-consistently included in our formalism. A
reasonable prescription for a bound is to find the number of modes that gives
the most stringent upper bound.
In the global spectrum case, foreground residuals after $N$ spectral modes are
removed can have either sign in the dot product with the signal template. This
two-sided bias should stabilize as most of the foreground modes are
discovered. For more modes removed, the errors will grow, but the central
value should remain stable if most of the foreground covariance is described
and removed. The summary results of a global $21$ cm experiment should include
plots of the spectral modes removed and characterization of the constraint as
a function of modes removed.
Both this issue and the residual monopole in Section 7.3 cannot be addressed
rigorously. An irreducible challenge of global $21$ cm measurement is
determining whether any foreground bias remains. The methods here provide
excellent prospects for cleaning the data and guiding an experiment, but they
do not solve these central issues.
### 7.5. Extending the Spectral Template
We have discussed constraints on the amplitude of a cosmological $21$ cm
signal template that is known in advance. The primary goal of the current
generation of instruments is discovery of this amplitude. Subsequent
experiments will then constrain the spectrum itself. The methods developed
here can also be applied to that case.
There is a continuum of models between the template estimate here and a full
spectrum. One possibility would be to use prior information about the redshift
of reionization and estimate two amplitudes, $\alpha(z<z_{r})$ and
$\alpha(z>z_{r})$, and then examine the significance of a step feature.
Another option would be to constrain a fixed number of spectral spline
amplitudes. In the limit that each frequency bin has a different amplitude,
the methods developed here can find the component of that signal orthogonal to
the identified foreground modes.
The choice between models must juggle (1) use of prior information, (2)
discriminatory power, and (3) insight about the trajectory of the global $21$
cm signal. This model selection problem is deferred to future work.
### 7.6. Application to Pre-reionization Absorption
Figure 13.— Same as Figure 12, except that we now consider a pre-reionization
absorption feature parameterized by a Gaussian. The duration is given as the
$\sigma$ width of the Gaussian. Note that significantly more signal remains.
Throughout the paper, we have considered the global $21$ cm signal from
reionization. This choice was primarily pedagogical, to have a simple step
around which the foregrounds are subtracted. The maximum contrast of this
feature is $\sim 30$ mK, and for even modest $\Delta z$ it becomes smooth and
more easily subtracted by the foreground modes. In contrast, in the period
prior to reionization, the $21$ cm is thought to go into absorption at $\sim
100$ mK (Pritchard & Loeb, 2010). Here, the spin temperature is strongly
radiatively coupled to the cold gas. However, while synchrotron radiation
contaminating $z\sim 10$ $21$ cm radiation around the time of reionization is
$\sim 500$ K (averaged across the sky), the contamination of an absorption
signature at $z\sim 20$ is $\sim 3000$ K.
While the synchrotron emission from higher redshift $21$ cm emission may be
greater, the absorption signature has the additional feature that it is two-
sided, both falling and rising. This generally makes foreground modes less
parallel to the signal and leads to less signal loss. For our purposes here,
we can again develop a simple model where $T_{b}$ is a Gaussian function
parameterized by a central redshift and width $\sigma_{z}$. Figure 13 is
analogous to Figure 12 except that it considers this era. Generally much more
cosmological signal remains orthogonal to the foreground modes.
### 7.7. Susceptibility to Faraday Rotation
In addition to the Stokes-$I$ component of the foregrounds, synchrotron
emission is also polarized. The polarized emission is subject to Faraday
rotation as it traverses various column densities of free electrons in
magnetic fields. If a receiver were designed to be purely sensitive to
Stokes-$I$, then this spectral fluctuation could be ignored. Generally though,
a receiver will have some level of response to polarized signals, and Faraday
rotation produces a signal that oscillates in frequency (Moore et al., 2013).
The rotation angle is $({\rm RM})\cdot\lambda^{2}$, where RM is the rotation
measure. Even for modest RM, the rotation angle can vary rapidly over our
band. This is especially problematic for the global $21$ cm signal because the
rotation measure varies over the sky, increasing the number of contaminated
modes in the data.
Extragalactic sources are subject to Faraday rotation through the screen of
the entire Milky Way. This has been measured recently by Oppermann et al.
(2012, 2014), and is shown in Figure 14. Each point source will generally have
a different rotation measure and require a new degree of freedom to describe.
Bernardi et al. (2013) has conducted deep, wide-field observations of
polarization of galactic synchrotron emission at $189$ MHz. They find only one
source out of 70 with $S>4$ Jy with polarization at the level of $1.8\%$ in
$2400\,{\rm deg}^{2}$, with all others falling below the $2\%$ polarization
fraction.
Polarization and Faraday rotation of galactic synchrotron emission is more
complex because synchrotron emission is interspersed with the Faraday screen.
Bernardi et al. (2013) find peak polarized emission at $\sim 13$ K and
rotation measures mostly below $10~{}{\rm rad}\,{\rm m}^{-2}$. The low
rotation measures and polarization fraction are thought to be due to a
depolarization horizon (Landecker et al., 2010; Uyaniker et al., 2003) within
the galaxy, a distance beyond which most of the emission is depolarized along
the line of sight.
The global $21$ cm experiments to date have been single element, making it
difficult to make a precise comparison with interferometric galaxy
polarization surveys conducted at similar frequencies (Pen et al., 2009;
Bernardi et al., 2013). Vinyaikin et al. (1996) measured $3.5\pm 1.0$ K at
$88$ MHz ($24^{\circ}$ FWHM, phased array) and $2.15\pm 0.25$ K at $200$ MHz
($8^{\circ}$ FWHM, single dish) for the value of $\sqrt{Q^{2}+U^{2}}$ toward
the north celestial pole.
To assess the galactic polarization field, we use the model of Waelkens et al.
(2009), which self-consistently simulates the polarized emission and rotation
measure. Figure 14 shows the eigenvalue spectrum of the Stokes-$Q$ spectral
modes. Unlike Stokes-$I$, polarization is spectrally much more complex, and
its covariance is spread among many modes. Ideally, all of these modes would
be suppressed well below the noise level of the experiment so they do not need
to be estimated and subtracted. Based on the measured amplitudes of the
polarized signals above, suppose that polarization is $10^{-3}$ of intensity.
To achieve a $10^{5}$ suppression, the polarization leakage must be kept below
$1\%$.
Figure 14.— Top: galactic rotation measure map of Oppermann et al. (2014) in
units of ${\rm rad}\,{\rm m}^{-2}$, with color range truncated at $\pm
100~{}{\rm rad}\,{\rm m}^{-2}$ to show more structure out of the plane.
Bottom: the eigenvalue spectrum of Stokes-$Q$ emission in the galactic model
of Waelkens et al. (2009), normalized to one. This component falls off very
slowly because the rotation measure introduces many degrees of freedom that
vary across the sky.
## 8\. Conclusion
Measurements of the global $21$ cm signal are very challenging due to (1)
astrophysical foregrounds and their interaction with instrumental systematics,
(2) terrestrial radio interference, and (3) the ionosphere. Here we have
examined the first issue, especially in regard to instrumental response.
Following $z\sim 1$ $21$ cm literature (Switzer et al., 2013; Masui et al.,
2013; Chang et al., 2010), we develop a new method where foregrounds can be
jointly estimated with the monopole spectrum. This relies on the fact that
foregrounds vary across the sky while the cosmological signal is constant.
This idea also extends L13, who argue that surveys with moderate angular
resolution covering much of the sky are able to better discriminate between
the monopole $21$ cm signal and foregrounds.
The key observation arising from our cleaning method is that the instrument
should be designed to minimize the generation of new spectral degrees of
freedom from the foregrounds. For example, if each line of sight has a
slightly different passband calibration, it also requires a new spectral
degree of freedom to describe the foregrounds there. In this sense, the
instrument “spreads out” the variance over modes that ultimately require more
aggressive cleaning and signal loss. In contrast, a constant passband
calibration error does not increase the rank of the foreground spectral
covariance, and its effect is primarily aesthetic. This is fortuitous because
obtaining a smooth spectral calibration of an instrument at these frequencies
would require very large, expensive structures that are black in radio
wavelengths. Simply requiring that galactic foregrounds be smooth to a few
percent provides a sufficient passband calibration for signal recovery. In
contrast, significant effort must be put into maintaining passband stability
to at least $\sim 10^{-4}$. This may require using much of the instrument’s
sensitivity to integrate against a stable reference.
Polarization to intensity leakage is another example of an instrumental
systematic that increases the rank of the foreground covariance. It allows
spectral oscillations in the Faraday-rotating polarization signal to
contaminate the spectral intensity measurement. Because of depolarization, sky
polarization is only a fraction of the intensity, and our estimates of the
instrumental constraints are more lax, $<10^{-2}$.
In attempting to measure the global $21\,\textrm{cm}$ signal, one is faced
with bright contaminating foregrounds that can interact with instrumental
systematics in nontrivial ways, overwhelming the cosmological signal that one
seeks to measure. In this paper, we have developed methods that bear
similarities to various previously proposed intuitive data-analysis
techniques. However, our methods arise from a rigorous, self-consistent
framework that allows unknown and unanticipated foreground properties to be
derived from real data. Such a framework also provides guidance for instrument
design. If design requirements can be adequately met, high-significance
measurements of the global $21\,\textrm{cm}$ signal will be possible,
providing direct access to the rich and complex astrophysics of the first
luminous objects and reionization.
E.S. thanks Ue-Li Pen for conversations and feedback, and for stimulating the
approach in collaboration on $z\sim 1$ $21$ cm data analysis from the Green
Bank Telescope. A.L. thanks Matt McQuinn for conversations. We thank Harish
Vedantham for comments and the Ohio $21$ cm workshop for fostering
conversations that initiated the project. A.L. acknowledges support from NSF
grants AST-0804508, AST-1129258, and AST-1125558.
## Appendix A A. Relation Between Covariance Adjustment and the Maximum-
Likelihood Linear Estimator
In Section 3.2, we claim that the estimator
$\boldsymbol{m}=(\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{x}$
is equivalent to the projection operation,
$\boldsymbol{m}=(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}(1-\mathbf{\Pi})\boldsymbol{y}$
where
$\mathbf{\Pi}=\mathbf{\Sigma}\boldsymbol{Z}(\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}$.
The two essential conditions here are that
$\boldsymbol{X}^{T}\boldsymbol{Z}=0$ (that the signal and nonsignal vectors
are orthogonal) and that the vectors in $\boldsymbol{X}$ and $\boldsymbol{Z}$
together span the space. The operations $\boldsymbol{X}^{T}$ and
$\boldsymbol{Z}^{T}$ can be understood as projecting onto the signal basis and
the everything-but-signal basis, respectively. This proof follows Lemma 2b of
Rao (1967). Writing out the terms, we would like to prove
$(\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}=(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}-(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}\mathbf{\Sigma}\boldsymbol{Z}(\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}.$
(A1)
Multiplying on the right by $\boldsymbol{X}$ is trivially true because
$\boldsymbol{Z}^{T}\boldsymbol{X}=0$ and the other terms are the identity
matrix. This checks just the $\boldsymbol{X}$ subspace of the equality. To
prove general equality we need to check the $\boldsymbol{Z}$ subspace.
Multiply on the right by $\boldsymbol{Z}$ and from the left by
$\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{X}$ (which will allow
simplification)
$\displaystyle\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{Z}=-(\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{X})(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}\mathbf{\Sigma}\boldsymbol{Z}(\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}\boldsymbol{Z}$
(A2)
$\displaystyle\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{Z}=-(\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{X})(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}\mathbf{\Sigma}\boldsymbol{Z}[(\boldsymbol{Z}^{T}\boldsymbol{Z})^{-1}(\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{Z})]^{-1}$
(A3)
$\displaystyle\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{Z}[(\boldsymbol{Z}^{T}\boldsymbol{Z})^{-1}(\boldsymbol{Z}^{T}\mathbf{\Sigma}\boldsymbol{Z})]=-(\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}\boldsymbol{X})(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}\mathbf{\Sigma}\boldsymbol{Z}$
(A4)
$\displaystyle\boldsymbol{X}^{T}\mathbf{\Sigma}^{-1}[\boldsymbol{Z}(\boldsymbol{Z}^{T}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}+\boldsymbol{X}(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}]\mathbf{\Sigma}\boldsymbol{Z}=0.$
(A5)
Because $\boldsymbol{X}$ and $\boldsymbol{Z}$ span the space,
$\boldsymbol{Z}(\boldsymbol{Z}^{T}\boldsymbol{Z})^{-1}\boldsymbol{Z}^{T}+\boldsymbol{X}(\boldsymbol{X}^{T}\boldsymbol{X})^{-1}\boldsymbol{X}^{T}=\boldsymbol{1}$,
then noting that $\boldsymbol{X}^{T}\boldsymbol{Z}=0$ proves the identity. In
many of the methods in the paper, the projection $\mathbf{\Pi}$ is reduced to
$(1-\boldsymbol{F}\boldsymbol{F}^{T})$, completely projecting out a set of
modes. The modes in $\boldsymbol{F}$ clearly cannot span the space. Further,
we let the modes $\boldsymbol{F}$ overlap with the signal. These choices make
the estimator less formally optimal but more robust to very bright, non-
|
[a]Andrés E. Rentería-Olivo
# Grover’s Quantum Search Algorithm of Causal Multiloop Feynman Integrals.
###### Abstract
A proof-of-concept application of a quantum algorithm to multiloop Feynman
integrals in the Loop-Tree Duality (LTD) framework is applied to a
representative four-loop topology. Bootstrapping causality in the LTD
formalism, is a suitable problem to address with quantum computers given the
straightforward possibility to encode the two on-shell states of a propagator
on the two states of a qubit. A modification of Grover’s quantum search
algorithm is developed and the quantum algorithm is successfully implemented
on IBM Quantum and QUTE simulators.
## 1 Introduction
High-precision theoretical predictions for current and future high-energy
particle colliders require novel techniques to deal with the higher orders in
perturbation theory of scattering amplitudes. In this context, the loop-tree
duality (LTD) [1, 3, 2] formalism exhibits interesting mathematical properties
of high potential to overcoming current limitations. A remarkable property of
LTD is the possibility of representing the causal nature of Feynman diagrams
and scattering amplitudes, leading to an intuitive understanding of the
singular structure of loop integrals. This manifestly causal representation
[3, 4, 5] allows to replace the original multiloop Feynman diagrams into a
class of multiloop topologies defined by collapsing propagators into edges
where the two on-shell states of propagators are naturally encoded by the two
states of a qubit, leading to explore the application of quantum algorithms,
for instance, a Grover’s quantum search algorithm [6] and a Variational
Quantum Eigensolver [7] approach. In this work we focus on the former
approach.
## 2 Loop-Tree Duality and Causality
The causal representation of scattering amplitudes in the LTD formalism is
obtained through the calculation of nested residues. In Refs. [2, 3] it is
shown that scattering amplitudes can be written as
${\cal
A}_{D}^{(L)}=\int_{\vec{\ell}_{1}\ldots\vec{\ell}_{L}}\frac{1}{x_{n}}\sum_{\sigma\in\Sigma}{\cal
N}_{\sigma}\,\prod_{i=1}^{n-L}\,\frac{1}{\lambda_{\sigma(i)}^{h_{\sigma(i)}}}\
+(\lambda^{+}\leftrightarrow\lambda^{-})\leavevmode\nobreak\ ,$ (1)
with $x_{n}=\prod_{i}2q_{i,0}^{(+)}$, $h_{\sigma(i)}=\pm 1$,
$\cal{N}_{\sigma}$ a numerator determined by the interaction vertices of a
specific theory and $\int_{\vec{\ell_{s}}}=-\mu^{4-d}(2\pi)^{1-d}\int{\rm
d}^{d-1}\ell_{s}$, the integration measure in the loop three-momentum space.
The Eq. (1) only involves denominators with positive on-shell energies
$q_{i,0}^{(+)}=(\vec{q}_{i}^{\,2}+m_{i}^{2}-\imath 0)^{1/2}$, added together
in same-sign combinations in the so-called causal propagators,
$1/\lambda_{\sigma(i)}^{\pm}$, with
$\lambda_{\sigma(i)}^{\pm}\equiv\lambda_{p}^{\pm}=\sum_{i\in
p}q_{i,0}^{(+)}\pm k_{p,0}\leavevmode\nobreak\ ,$ (2)
where $\sigma(i)$ stands for a partition $p$ of the set of on-shell energies
and the orientation of the energy components of the external momenta,
$k_{p,0}$. The causal structure of $\lambda_{p}^{\pm}$ is defined by the sign
of $k_{p,0}$ when the propagators in the partition $p$ are set on-shell. Each
causal propagator is in a one-to-one correspondence with any possible
threshold singularity of the amplitude, which contains overlapped thresholds
that are known as causal entangled thresholds. The combinations of entangled
causal propagators represent causal thresholds that can occur simultaneously,
which are collected in the set $\Sigma$.
## 3 Causal query of multiloop Feynman integrals: the four-eloop case
We develop a modified version [6] of Grover’s algorithm [8] for querying the
causal configurations from the LTD framework. We take as example the four-
eloop topology, which consists in four loops made of edges (eloops) with a
central four-particle interaction vertex and we implement it on IBM
Quantum***https://quantum-computing.ibm.com/. The implementation of this
quantum algorithm requires four registers, namely, $|{e}\rangle$,
$|{c}\rangle$, $|{a}\rangle$ and $|{\it out}\rangle$. In the first register,
$|{e}\rangle$, we encode the state of the $N=2^{n}$ edges of the topology and
initialise it with a uniform superposition by applying Hadamard gates,
$|{e}\rangle=H^{\otimes n}|{0}\rangle$. The second register, $|{c}\rangle$,
encodes the comparison between two adjacent edges $e_{i}$ and $e_{j}$ in
binary Boolean clauses defined by
$\displaystyle
c_{ij}\equiv(e_{i}=e_{j})\qquad\text{and}\qquad\bar{c}_{ij}\equiv(e_{i}\neq
e_{j})\leavevmode\nobreak\ .$ (3)
Each clause $\bar{c}_{ij}$ is implemented through two CNOT gates, each taking
as control the corresponding qubits $e_{i}$ and $e_{j}$ respectively, and both
taking as target the same qubit in the register $|{c}\rangle$. The clause
$c_{ij}$ requires an extra NOT gate on the target qubit on the register
$|{c}\rangle$. The third register, $|{a}\rangle$, stores the eloop clauses
that probe the adjacent edges that compose a cyclic circuit within the
diagram. Specifically for the four-eloop topology, taking the conventional
orientation of the edges in Fig. (1), the clauses are given by
$\displaystyle a_{0}^{(4)}$ $\displaystyle=\neg(c_{01}\wedge c_{12}\wedge
c_{23})\leavevmode\nobreak\ ,$ $\displaystyle a_{1}^{(4)}$
$\displaystyle=\neg(\bar{c}_{05}\wedge\bar{c}_{45})\leavevmode\nobreak\ ,$
$\displaystyle a_{2}^{(4)}$
$\displaystyle=\neg(\bar{c}_{16}\wedge\bar{c}_{56})\leavevmode\nobreak\ ,$
$\displaystyle a_{3}^{(4)}$
$\displaystyle=\neg(\bar{c}_{27}\wedge\bar{c}_{67})\leavevmode\nobreak\ ,$
$\displaystyle a_{4}^{(4)}$
$\displaystyle=\neg(\bar{c}_{34}\wedge\bar{c}_{47})\leavevmode\nobreak\ .$ (4)
Each $a_{i}^{(4)}$ requires a multicontrolled Toffoli gate that takes as
control its specific qubits $|{c_{ij}}\rangle$ and as target a qubit in the
register $|{a}\rangle$ followed by a NOT gate. The last register, $|{\it
out}\rangle$, requires a single qubit initialised in the
$|{-}\rangle=HX|{0}\rangle$ and is used as the Grover’s marker. It stores the
output of the quantum algorithm and is implemented through a multicontrolled
Toffoli gate taking as control all the qubits from the register $|{a}\rangle$
and, if required, a qubit from the register $|{e}\rangle$. The oracle is
defined as
$\displaystyle U_{w}|{e}\rangle|{c}\rangle|{a}\rangle|{\it
out}\rangle=(-1)^{f(a,e)}|{e}\rangle|{c}\rangle|{a}\rangle|{\it
out}\rangle\leavevmode\nobreak\ .$ (5)
When the causal conditions are satisfied, $f(a,e)=1$, and marks the
corresponding states; otherwise, if $f(a,e)=0$, they are left unchanged. For
the four-eloop topology, the required Boolean marker is given by
$\displaystyle f^{(4)}(a,e)=\left(\leavevmode\nobreak\
\bigwedge_{i=0}^{4}a_{i}^{(4)}\leavevmode\nobreak\ \right)\wedge
e_{0}\leavevmode\nobreak\ .$ (6)
After storing the marked states in $|{\it out}\rangle$, the registers
$|{a}\rangle$ and $|{c}\rangle$ are rotated back to the state $|{0}\rangle$ by
applying the oracle operations in inverse order. The last step in the
algorithm is the amplification of the marked states by applying the diffuser
opertator on the register $|{e}\rangle$. We use the diffuser defined on IBM
Quantum†††https://qiskit.org.
missing
$e_{0}$
missing
$e_{1}$
missing
$e_{2}$
missing
$e_{3}$
missing
$e_{4}$
missing
$e_{5}$
missing
$e_{6}$
missing
$e_{7}$
Figure 1: Left: Four-eloop topology with a fixed conventional orientation.
Right: corresponding quantum circuit for the four-eloop topology.
## 4 Conclusions
The quantum algorithm developed in this work allowed us to successfully
identify the causal singular configurations of the four-eloop topology in the
LTD framework. The algorithm was successfully implemented in IBM Quantum and
QUTE simulators‡‡‡https://qute.ctic.es. The result of the quantum algorithm is
used to bootstrap the causal representation in LTD of representative multiloop
topologies, allowing us a better understanding of this approach and enabling
us to explore other techniques, such as a Variation Quantum approach, This
challenging problem is closely related to the identification of directed
acyclic graphs in graph theory, which can find application in other fields
beyond particle physics.
## 5 Acknowledgements
I would like to thank the team that collaborated in the development of this
work, G. Rodrigo, S. Ramírez-Uribe, G. F. R. Sborlini and L. Vale-Silva. To
IBM Quantum for the access to IBMQ and also to Fundación Centro Tecnológico de
la Información y Comunicación (CTIC) for the access to Quantum Testbed (QUTE).
This work is supported by the PRE2018-085925 grant from the Consejo Superior
de Investigaciones Científicas (CSIC) of the Spanish Government,
MCIN/AEI/10.13039/501100011033, and Grant No. PID2020-114473GB-I00.
## References
* [1] S. Catani, T. Gleisberg, F. Krauss, G. Rodrigo and J.-C. Winter, _JHEP_ 09 (2008) 065, [0804.3170].
* [2] J. de Jesús Aguilera-Verdugo, F. Driencourt-Mangin, R. J. Hernández-Pinto, J. Plenter, R. M. Prisco, N. S. Ramírez-Uribe, A. E. Rentería-Olivo, G. Rodrigo, G. Sborlini and W. J. Torres Bobadilla, et al. Symmetry 13 (2021) no.6, 1029 [2104.14621 [hep-ph].
* [3] J. J. Aguilera-Verdugo, F. Driencourt-Mangin, R. J. Hernández-Pinto, J. Plenter, S. Ramirez-Uribe, A. E. Renteria Olivo, G. Rodrigo, G. F. R. Sborlini, W. J. Torres Bobadilla and S. Tracz, Phys. Rev. Lett. 124 (2020) no.21, 211602 [2001.03564 [hep-ph]].
* [4] J. J. Aguilera-Verdugo, R. J. Hernandez-Pinto, G. Rodrigo, G. F. R. Sborlini and W. J. Torres Bobadilla, _JHEP_ 01 (2021) 069, [2006.11217].
* [5] S. Ramírez-Uribe, R. J. Hernández-Pinto, G. Rodrigo, G. F. R. Sborlini and W. J. Torres Bobadilla, _JHEP_ 04 (2021) 129, [2006.13818].
* [6] S. Ramírez-Uribe, A. E. Rentería-Olivo, G. Rodrigo, G. F. R. Sborlini and L. Vale Silva, JHEP 05 (2021) 100, [2105.08703].
* [7] G. Clemente, A. Crippa, K. Jansen, S. Ramírez-Uribe, A. E. Rentería-Olivo, G. Rodrigo, G. F. R. Sborlini and L. Vale Silva, [2210.13240].
* [8] L. K. Grover, _Phys. Rev. Lett._ 79 (1997) 325--328, [quant-ph/9706033].
|
# Task-Free Continual Learning via
Online Discrepancy Distance Learning
Fei Ye and Adrian G. Bors
Department of Computer Science
University of York
York, YO10 5GH, UK
<EMAIL_ADDRESS>
###### Abstract
Learning from non-stationary data streams, also called Task-Free Continual
Learning (TFCL) remains challenging due to the absence of explicit task
information. Although recently some methods have been proposed for TFCL, they
lack theoretical guarantees. Moreover, forgetting analysis during TFCL was not
studied theoretically before. This paper develops a new theoretical analysis
framework which provides generalization bounds based on the discrepancy
distance between the visited samples and the entire information made available
for training the model. This analysis gives new insights into the forgetting
behaviour in classification tasks. Inspired by this theoretical model, we
propose a new approach enabled by the dynamic component expansion mechanism
for a mixture model, namely the Online Discrepancy Distance Learning (ODDL).
ODDL estimates the discrepancy between the probabilistic representation of the
current memory buffer and the already accumulated knowledge and uses it as the
expansion signal to ensure a compact network architecture with optimal
performance. We then propose a new sample selection approach that selectively
stores the most relevant samples into the memory buffer through the
discrepancy-based measure, further improving the performance. We perform
several TFCL experiments with the proposed methodology, which demonstrate that
the proposed approach achieves the state of the art performance.
## 1 Introduction
Continual learning (CL) and its extension to lifelong learning, represents one
of the most desired functions in an artificial intelligence system,
representing the capability of learning new concepts while preserving the
knowledge of past experiences [32]. Such an ability can be used in many real-
time applications such as robotics, health investigative systems, autonomous
vehicles [20] or for guiding agents exploring artificial (meta) universes,
requiring adapting to a changing environment. Unfortunately, modern deep
learning models suffer from a degenerated performance on past data after
learning novel knowledge, a phenomenon called catastrophic forgetting [13].
A popular attempt to relieve forgetting in CL is by employing a small memory
buffer to preserve a few past samples and replay them when training on a new
task [1, 6]. However, when there are restrictions on the available memory
capacity, memory-based approaches would suffer from degenerated performance on
past tasks, especially when aiming to learn an infinite number of tasks.
Recently, the Dynamic Expansion Model (DEM) [51] has shown promising results
in CL, aiming to guarantee optimal performance by preserving the previously
learnt knowledge through the parameters of frozen components trained on past
data, while adding a new component when learning a novel task. However, such
approaches require knowing where and when the knowledge associated with a
given task is changed, which is not always applicable in a real environment.
In this paper, we address a more realistic scenario, called Task-Free
Continual Learning (TFCL) [3], where task identities are not available while
the model can only access a small batch of samples at a given time. Most
existing CL methods requiring the task label can be adapted to TFCL by
removing the task information dependency. For instance, memory-based
approaches can store a few past samples from the data stream at each training
time and replay them during later training steps [8, 12]. However, such an
approach requires to carefully design the sample selection criterion to avoid
memory overload. The key challenge for the memory-based approach is the
negative backward transfer caused by the stored samples that interfere with
the model’s updating with incoming samples [6]. This issue can be relieved by
DEM in which previously learnt samples are preserved into frozen components
and do not interfere with the learning of probabilistic representations of new
data [24, 38]. However, these approaches do not provide any theoretical
guarantees and there are no studies analysing the trade-off between the
model’s generalization and its complexity under TFCL.
Recent attempts have provided the theoretical analysis for CL from different
perspectives including the risk bound [46, 51], NP-hard problem [17], Teacher-
Student framework [23, 58] and game theory [37]. However all these approaches
require strong assumptions, such as defining the task identities, which is not
available in TFCL. This inspires us to bridge the gaps between the underlying
theory and the algorithm implementation for TFCL. We propose a theoretical
classification framework, which provides new insights in the forgetting
behaviour analysis and guidance for algorithm design addressing catastrophic
forgetting. The primary motivation behind the proposed theoretical framework
is that we can formulate forgetting as a generalization error in the domain
adaptation theory. Based on this analysis we extend the domain adaptation
theory [29] to derive time-dependent generalization risk bounds, explicitly
explaining the forgetting process at each training step.
Inspired by the theory, we devise the Online Discrepancy Distance Learning
(ODDL) method which introduces a new expansion mechanism based on the
discrepancy distance estimation for implementing TFCL. The proposed expansion
mechanism detects the data distribution shift by evaluating the variance of
the discrepancy distance during the training. This model enables a trade-off
mechanism between the model’s generalization and complexity. We also propose a
new sample selection approach based on the discrepancy-based criterion, which
guides storing diverse samples with respect to the already learnt knowledge,
further improving performance. Our contributions are :
* •
This paper is the first research study to propose a new theoretical framework
for TFCL, which provides new insights into the forgetting behaviour of the
model in classification tasks.
* •
Inspired by the theoretical analysis, we develop a novel dynamic expansion
approach, which ensures a compact model architecture enabled by optimal
performance.
* •
We propose a new sample selection approach that selects appropriate data
samples for the memory buffer, further improving performance.
* •
The proposed method achieves state of the art results on TFCL benchmarks,
## 2 Related works
Continual learning
defines a learning paradigm which aims to learn a sequence of tasks without
forgetting. Catastrophic forgetting is a major challenge in continual
learning. One of the most popular approaches to relieve forgetting is by
imposing a regularization loss within the optimization procedure [7, 11, 13,
16, 19, 25, 26, 31, 34, 35, 40, 41, 57], where the network’s parameters which
are important to the past tasks are penalized when updating. Another kind of
approaches for continual learning focuses on the memory system, which usually
employs a small memory buffer [1, 5, 6, 28, 36, 44, 59] to store a few past
data or trains a generator to provide the replay samples when learning new
tasks [38, 43, 46, 47, 52, 58, 53]. However, these approaches usually rely on
knowing the task information, which is not applicable in TFCL.
Task-free continual learning
is a special scenario in CL where a model can only see one or very few samples
in each training step/time without having any task labels. Using a small
memory buffer to store past samples has shown benefits for TFCL and was
firstly investigated in [3, 54, 56]. This memory replay approach was then
extended by employing Generative Replay Mechanisms (GRMs) for training both a
Variational Autoencoder (VAEs) [15] and a classifier, where a new retrieving
mechanism is used to select specific data samples, called the Maximal
Interfered Retrieval (MIR), [2]. The Gradient Sample Selection (GSS) [1] is
another sample selection approach that treats sample selection as a
constrained optimization reduction. More recently, a Learner-Evaluator
framework is proposed for TFCL, called the Continual Prototype Evolution
(CoPE) [8] which stores the same number of samples for each class in the
memory in order to ensure the balance replay. Another direction for the
memory-based approaches is to edit the stored samples which would increase the
loss in the upcoming model updates, called the Gradient based Memory EDiting
(GMED) [12], which can be employed in the existing CL models to further
enhance their performance.
Dynamic expansion models
aim to automatically increase the model’s capacity to adapt to new tasks by
adding new hidden layers and units. Such an approach, called the Continual
Unsupervised Representation Learning (CURL) [38], dynamically builds new
inference models when meeting the expansion criterion. However, since CURL
still requires a GRM to relieve forgetting, it would lead to a negative
knowledge transfer when updating the network’s parameters to adapt to a new
task. This issue can be addressed by using Dirichlet processes by adding new
components while freezing all previously learnt members, in a model called the
Continual Neural Dirichlet Process Mixture (CN-DPM), [24]. However, the
expansion criterion used by these approaches relies on the change of the loss
when training each time, which does not have any theoretical guarantees.
## 3 Theoretical analysis of TFCL
In this section, we firstly introduce learning settings and notations, and
then we analyze the forgetting behaviour for a single as well as for a dynamic
expansion model by deriving their Generalization Bounds (GBs).
### 3.1 Preliminary
Let $\mathcal{X}$ be the input space and $\mathcal{Y}$ represent the output
space which is $\\{-1,1\\}$ for binary classification and
$\\{1,2,\dots,n^{\prime}\\},n^{\prime}>2$ for multi-class classification. Let
$\mathcal{D}^{T}_{i}=\\{{\bf x}^{T}_{j},y^{T}_{j}\\}^{N^{T}_{i}}_{j=1}$ and
$\mathcal{D}^{S}_{i}=\\{{\bf x}^{S}_{j},y^{S}_{j}\\}^{N^{S}_{i}}_{j=1}$
represent the training and testing sets for the $i$-th dataset where ${\bf
x}^{T}_{j}\in\mathcal{X}$ and $y^{T}_{j}\in{\mathcal{Y}}$ are the image and
the associated ground truth label. $N^{T}_{i}$ and $N^{S}_{i}$ are the total
number of samples for ${\mathcal{D}}^{T}_{i}$ and ${\mathcal{D}}^{S}_{i}$,
respectively. In the paper, we mainly focus on the task-free class-incremental
learning, described as follows.
Definition 1.
(Data stream.)
For a given $t$-th training dataset $\mathcal{D}^{S}_{t}$ with $C^{S}_{t}$
data categories, let us consider a data stream $\mathcal{S}$ which consists of
samples $\mathcal{D}^{S}_{t,j}$ from each category, expressed by
$S=\bigcup_{j=1}^{C^{S}_{t}}\mathcal{D}^{S}_{t,j}$. Let
${\mathcal{D}}^{T}_{t,j}$ represent the set of samples drawn from the $j$-th
category of ${\mathcal{D}}^{T}_{t}$. Let ${\mathbb{P}}^{T}_{t,j}$ and
${\mathbb{P}}^{S}_{t,j}$ represent the distributions for
${\mathcal{D}}^{T}_{t,j}$ and ${\mathcal{D}}^{S}_{t,j}$, respectively. Let
${\mathbb{P}}^{T,{\mathcal{X}}}_{t,j}$ represent the marginal distribution
over $\mathcal{X}$.
Definition 2.
(Learning setting.)
Let $\mathcal{T}_{i}$ represent the $i$-th training step. For a given data
stream $\mathcal{S}$, we assume that there are a total of $n$ training steps
for $\mathcal{S}$, where each training step $\mathcal{T}_{i}$ is associated
with a small batch of paired samples $\\{{\bf X}^{b}_{i},{\bf Y}^{b}_{i}\\}$
drawn from $\mathcal{S}$, expressed by ${\mathcal{S}}=\bigcup_{i=1}^{n}\\{{\bf
X}_{i}^{b},{\bf Y}^{b}_{i}\\},\;\;\\{{\bf X}_{i}^{b},{\bf
Y}^{b}_{i}\\}\cap\\{{\bf X}_{j}^{b},{\bf Y}^{b}_{j}\\}=\varnothing$, where
$i\neq j$ and a model (classifier) can only access $\\{{\bf X}^{b}_{i},{\bf
Y}^{b}_{i}\\}$ at $\mathcal{T}_{i}$ while all previous batches are not
available. After finishing all training steps, we evaluate the classification
accuracy of the model on the testing set $\mathcal{D}^{T}_{t}$. In the
following, we define the model and the memory buffer.
Definition 3.
(Model and memory.)
Let us consider $h$ a model implemented by a classifier, and
$\mathcal{H}=\\{h\,|\,h\colon{\mathcal{X}}\to\mathcal{Y}\\}$ the space of
classifiers. Let $\mathcal{M}_{i}$ be a memory buffer at $\mathcal{T}_{i}$. We
assume that $\mathcal{M}_{i}$ randomly removes samples from the memory buffer
while continually adding new data to its memory at each training step. Let
$\mathbb{P}_{{\mathcal{M}}_{i}}$ represent the probabilistic representation of
${\mathcal{M}}_{i}$ and $|{\mathcal{M}}_{i}|$ its cardinality.
### 3.2 Measuring the distribution shift
In TFCL, the distance between the target domain (testing set) and the source
domain (memory) would be dynamically changed during each training step. We can
use the discrepancy distance [29] to measure this gap through the analysis of
the model’s risk.
Definition 4.
(The risk.)
For a given distribution ${\mathbb{P}}^{S}_{t,j}$, the risk of a model $h$ is
defined as
${\mathcal{R}}\big{(}h,{\mathbb{P}}^{S}_{t,j}\big{)}\buildrel\Delta\over{=}{\mathbb{E}}_{\\{{\bf
x},y\\}\sim{\mathbb{P}}^{S}_{t,j}}\big{[}{\mathcal{L}}\big{(}y,h({\bf
x})\big{)}\big{]}$ where
$\mathcal{L}\colon{\mathcal{Y}}\times{\mathcal{Y}}\to[0,1]$ is the loss
function.
Definition 5.
(Discrepancy distance.)
For two given distributions ${{\mathbb{P}}}^{T}_{t,j}$ and
${{\mathbb{P}}}^{S}_{t,j}$, the discrepancy distance $\mathcal{L}_{d}$ is
defined on two marginals as :
$\displaystyle{\mathcal{L}}_{d}\big{(}{\mathbb{P}}^{T,{\mathcal{X}}}_{t,j},{{\mathbb{P}}}^{S,{\mathcal{X}}}_{t,j}\big{)}$
$\displaystyle\buildrel\Delta\over{=}\mathop{\sup}\limits_{\left(h,h^{\prime}\right)\in\mathcal{H}^{2}}\Big{|}{\mathbb{E}}_{{\bf
x}\sim{\mathbb{P}}^{T,{\mathcal{X}}}_{t,j}}\big{[}{\mathcal{L}}\big{(}h({\bf
x}),h^{\prime}({\bf x})\big{)}\big{]}-{\mathbb{E}}_{{\bf
x}\sim{\mathbb{P}}^{S,{\mathcal{X}}}_{t,j}}\big{[}{\mathcal{L}}\big{(}h({\bf
x}),h^{\prime}({\bf x})\big{)}\big{]}\Big{|}\,,$ (1)
where $\\{h,h^{\prime}\\}\in\mathcal{H}$. In practice, the discrepancy
distance ${\mathcal{L}}_{d}(\cdot,\cdot)$ can be estimated as the upper bound
based on the Rademacher complexity which is used in the domain adaptation
theory as a measure of richness for a particular hypothesis space [30, 60].
Corollary 1.
[29] For two given domains ${{\mathbb{P}}}^{T,{\mathcal{X}}}_{t,j}$ and
${{\mathbb{P}}}^{S,{\mathcal{X}}}_{t,j}$, let $U_{\mathcal{P}}$ and
$U_{\mathbb{P}}$ represent sample sets of sizes $m_{\mathcal{P}}$ and
$m_{\mathbb{P}}$, drawn independently from
${{\mathbb{P}}}^{T,{\mathcal{X}}}_{t,j}$ and
${{\mathbb{P}}}^{S,{\mathcal{X}}}_{t,j}$. Let
${\widehat{\mathbb{P}}}^{T,{\mathcal{X}}}_{t,j}$ and
${\widehat{\mathbb{P}}}^{S,{\mathcal{X}}}_{t,j}$ represent the empirical
distributions for $U_{\mathcal{P}}$ and $U_{\mathbb{P}}$. Let
$\mathcal{L}({\bf x^{\prime}},{\bf x})=|{\bf x^{\prime}}-{\bf x}{|^{1}}$ be a
loss function (L1-Norm), satisfying
$\forall({\bf{x}},{\bf{x^{\prime}}})\in\mathcal{X},\mathcal{L}({\bf{x}},{\bf{x^{\prime}}})>M$,
where $M>0$. Then, with the probability $1-\delta$, we have :
$\displaystyle{\mathcal{L}_{d}}({{\mathbb{{P}}}}^{T,{\mathcal{X}}}_{t,j},{{\mathbb{{P}}}}^{S,{\mathcal{X}}}_{t,j})$
$\displaystyle\leq{\mathcal{L}_{d}}(\widehat{\mathbb{P}}^{T,{\mathcal{X}}}_{t,j},\widehat{\mathbb{P}}^{S,{\mathcal{X}}}_{t,j})+C^{\star}+3M\left({\sqrt{\frac{{\log\left({\frac{4}{\delta}}\right)}}{{2{m_{\mathcal{P}}}}}}+\sqrt{\frac{{\log\left({\frac{4}{\delta}}\right)}}{{2{m_{\mathbb{P}}}}}}}\right),$
(2)
where $C^{\star}=4q\left({{{\mathop{\rm
Re}\nolimits}}_{{U_{\mathcal{P}}}}}\left(\mathcal{H}\right)+{{{\mathop{\rm
Re}\nolimits}}_{{U_{\mathbb{P}}}}}\left(\mathcal{H}\right)\right)$ and ${\rm
Re}_{U_{\mathcal{P}}}({\mathcal{H}})$ is the Rademacher complexity (Appendix-B
from Supplemental Material (SM)). Let
${\mathcal{L}_{\widehat{d}}}({{\mathbb{{P}}}}^{T,{\mathcal{X}}}_{t,j},{{\mathbb{{P}}}}^{S,{\mathcal{X}}}_{t,j})$
represent the Right-Hand Side (RHS) of Eq. (2). We also assume that
${\mathcal{L}}\colon\mathcal{Y}\times\mathcal{Y}\to[0,1]$ is a symmetric and
bounded loss function
$\forall(y,y^{\prime})\in{\mathcal{Y}^{2}},{\mathcal{L}}(y,y^{\prime})\leq{U^{\prime}}$,
and ${\mathcal{L}}(\cdot,\cdot)$ obeys the triangle inequality, where
${U^{\prime}}$ is a positive number.
### 3.3 GB for a single model
Based on the definitions from Section 3.2, we firstly derive the GB that can
describe the learning process of a single model under TFCL.
Theorem 1.
Let $\mathcal{P}_{i}$ represent the distribution of all visited training
samples (including all previous batches) drawn from $\mathcal{S}$ at
$\mathcal{T}_{i}$. Let
$h_{\mathcal{P}_{i}}=\arg\min_{h\in\mathcal{H}}{\mathcal{R}}(h,{\mathcal{P}}_{i})$
and
$h_{{\mathcal{M}}_{i}}=\arg\min_{h\in\mathcal{H}}{\mathcal{R}}(h,{\mathbb{P}}_{{\mathcal{M}}_{i}})$
represent the ideal classifiers for ${\mathcal{P}}_{i}$ and
${\mathbb{P}}_{{\mathcal{M}}_{i}}$, respectively. We derive the GB between
$\mathcal{P}_{i}$ and ${\mathbb{P}}_{{\mathcal{M}}_{i}}$, based on the results
from Corollary 1 :
$\displaystyle{\mathcal{R}}\big{(}h,{\mathcal{P}}_{i}\big{)}$
$\displaystyle\leq{\mathcal{R}}\big{(}h,h_{{\mathcal{M}}_{i}},{\mathbb{P}}_{{\mathcal{M}_{i}}}\big{)}+{\mathcal{L}}_{\widehat{d}}\big{(}{\mathcal{P}}^{\mathcal{X}}_{i},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}+\eta\big{(}{\mathcal{P}}_{i},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}\,,$
(3)
where $\eta\big{(}{\mathcal{P}}_{i},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}$
is the optimal combined error
${\mathcal{R}}(h_{{\mathcal{P}}_{i}},h_{{\mathcal{M}}_{i}},{{\mathcal{P}}_{i}})+{\mathcal{R}}(h_{{\mathcal{P}}_{i}},h_{{\mathcal{P}}_{i}}^{\star},{{\mathcal{P}}_{i}})$
where
${\mathcal{R}}(h_{{\mathcal{P}_{i}}},h_{{\mathcal{M}}_{i}},{\mathbb{P}}_{{\mathcal{M}}_{i}})$
is the risk, expressed by ${\mathbb{E}}_{{{\bf
x}}\sim{\mathbb{P}}_{{\mathcal{M}}_{i}}}[{\mathcal{L}}(h_{{\mathcal{P}}_{i}}({\bf
x}),h_{{\mathcal{M}}_{i}}({\bf x}))]$ and $h^{\star}_{{\mathcal{P}}_{i}}$ is
the true labeling function for ${\mathcal{P}}_{i}$.
The proof is provided in Appendix-A from the SM. Compared to the GB used in
the domain adaptation [29], Theorem 1 provides an explicit way to measure the
gap between the model’s prediction and the true labels in each training step
(${\mathcal{T}}_{i}$). During the initial training stages (when $i$ is very
small), the memory ${\mathcal{M}}_{i}$ can store all previous samples and GB
is tight. However, as the number of training steps increases, the discrepancy
distance
${\mathcal{L}}_{\widehat{d}}\big{(}{\mathcal{{P}}}^{\mathcal{X}}_{i},{\mathbb{{P}}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}$
would increase because ${\mathcal{M}}_{i}$ would lose the knowledge about
previously learnt samples. This can lead to a degenerated performance on
${\mathcal{P}}_{i}$, corresponding to the forgetting process. Next we extend
Theorem 1 to analyze the generalization performance on testing sets.
Theorem 2.
For a given target domain ${\mathbb{P}}^{T}_{t,j}$, we derive the GB for a
model at the training step ${\mathcal{T}}_{i}$ :
$\displaystyle{\mathcal{R}}\big{(}h,{\mathbb{P}}^{T}_{t,j}\big{)}$
$\displaystyle\leq{\mathcal{R}}\big{(}h,h_{{\mathcal{M}}_{i}},{\mathbb{P}}_{{\mathcal{M}_{i}}}\big{)}+{\mathcal{L}}_{\widehat{d}}\big{(}{\mathbb{P}}^{T,{\mathcal{X}}}_{t,j},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}+\eta\big{(}{\mathbb{P}}^{T}_{t,j},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}\,,$
(4)
The proof is similar to that for Theorem 1. We can observe that the
generalization performance on a target domain ${\mathbb{P}}^{T}_{t,j}$, by a
model $h$ is relying mainly on the discrepancy distance between
${\mathbb{P}}^{T,{\mathcal{X}}}_{t,j}$ and ${\mathbb{P}}_{{\mathcal{M}}_{i}}$.
In practice, we usually measure the generalization performance of $h$ on
several data categories where each category is represented by a different
underlying distribution. In the following, we extend Theorem 2 for multiple
target distributions.
Lemma 1.
For a given data stream $S=\bigcup_{j=1}^{C^{S}_{t}}{\mathcal{D}^{S}_{t,j}}$
consisting of samples from ${\mathcal{D}}^{S}_{t}$, let
${\mathcal{D}}^{T}_{t}$ be the corresponding testing set and
${\mathbb{P}}^{T}_{t,j}$ represent the distribution of samples for the $j$-th
category from ${\mathcal{D}}^{T}_{t}$, we derive the GB for multiple target
domains as:
$\displaystyle\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}{\mathcal{R}}\big{(}h,{\mathbb{P}}^{T}_{t,j}\big{)}\Big{\\}}$
$\displaystyle\leq\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}{\mathcal{R}}\big{(}h,h_{{\mathcal{M}}_{i}},{\mathbb{P}}_{{\mathcal{M}_{i}}}\big{)}+{\mathcal{L}}_{\widehat{d}}\big{(}{\mathbb{P}}^{T,{\mathcal{X}}}_{t,j},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}+\eta\big{(}{\mathbb{P}}^{T}_{t,j},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}\Big{\\}}\,,\vspace*{-0.4cm}$
(5)
where $C^{T}_{t}$ represents the number of testing data streams.
Remarks.
Lemma 1 had the following observations : 1) The optimal performance of the
model $h$ on the testing set can be achieved by minimizing the discrepancy
distance between each target domain ${\mathbb{P}}^{T,{\mathcal{X}}}_{t,j}$ and
the distribution ${\mathbb{P}}_{{\mathcal{M}}_{i}}$ at the training step
${\mathcal{T}}_{i}$. 2) [17]employs the set theory to theoretically
demonstrate that a perfect memory is crucial for CL. In contrast, we evaluate
the memory quality using the discrepancy distance in Eq. (5), which provides a
practical way to investigate the relationship between the memory and
forgetting behaviour of existing approaches [6, 8] at each training step
without requiring any task information (See more details in Appendix-D from
SM). 3) [51]introduces a similar risk bound for forgetting analysis, which
still requires the task information. In contrast, the proposed GB can be used
in a more realistic CL scenario. Moreover, we provide the theoretical analysis
for component diversity (Appendix-C from SM), which is missing in [51].
### 3.4 GB for the dynamic expansion mechanism
As discussed in Lemma 1, a memory of fixed capacity would lead to degenerated
performance on all target domains. The other problem for a single memory
system is the negative backward transfer [27] in which the performance of the
model is decreased due to samples being drawn from entirely different
distributions [14]. A Dynamic Expansion Model (DEM) can address these
limitations from two aspects : 1) DEM relieves the negative transfer by
preserving the previously learnt knowledge into a frozen component from a
mixture system; 2) DEM would achieve better generalization performance under
TFCL by allowing each component to model one or only a few similar underlying
data distributions. We derive GB for DEM and show the advantages of DEM for
TFCL.
Definition 7.
(Dynamic expansion mechanism.)
Let ${\mathcal{G}}$ represent a dynamic expansion model and ${G}_{j}$
represent the $j$-th component in ${\mathcal{G}}$, implemented by a single
classifier. ${\mathcal{G}}$ starts with training its first component during
the initial training phase and would add new components during the following
training steps. In order to overcome forgetting, only the newly created
component is updated each time, while all previously trained components have
their parameters frozen.
Theorem 3.
For a given data stream ${\mathcal{S}}=\\{{\bf X}^{b}_{1},\cdots,{\bf
X}^{b}_{n}\\}$, let ${\mathcal{P}}_{(i,j)}$ represent the distribution of the
$j$-th training batch ${\bf X}^{b}_{j}$ (visited) drawn from ${\mathcal{S}}$
at $\mathcal{T}_{i}$. We assume that ${\mathcal{G}}=\\{G_{1},\cdots,G_{c}\\}$
trained $c$ components at $\mathcal{T}_{i}$. Let
${\mathcal{T}}=\\{{\mathcal{T}}_{k_{1}},\cdots,{\mathcal{T}}_{k_{c}}\\}$ be a
set of training steps, where $G_{j}$ was frozen at $\mathcal{T}_{k_{j}}$. We
derive the GB for $\mathcal{G}$ at ${\mathcal{T}}_{i}$ as:
$\displaystyle\frac{1}{i}\sum\nolimits_{j=1}^{i}\big{\\{}{\mathcal{R}}\big{(}h,{\mathcal{P}}_{(i,j)}\big{)}\big{\\}}$
$\displaystyle\leq\frac{1}{i}\sum\nolimits_{j=1}^{i}\big{\\{}{\rm
F}_{S}\big{(}{\mathcal{P}}_{(i,j)},{\mathcal{G}}\big{)}\big{\\}}\,,$ (6)
where ${\rm F}_{S}(\cdot,\cdot)$ is the selection function, defined as :
$\displaystyle{\rm F}_{S}\big{(}{\mathcal{P}}_{(i,j)},{\mathcal{G}}\big{)}$
$\displaystyle\buildrel\Delta\over{=}\mathop{\min}\limits_{k_{1},\cdots,k_{c}}\Big{\\{}{\mathcal{R}}\big{(}h,h_{{\mathcal{M}}_{k_{i}}},{\mathbb{P}}_{{\mathcal{M}_{k_{i}}}}\big{)}+{\mathcal{L}}_{\widehat{d}}\big{(}{\mathcal{P}}^{\mathcal{X}}_{(i,j)},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{k_{i}}}\big{)}+\eta\big{(}{\mathcal{P}}_{(i,j)},{\mathbb{P}}_{{\mathcal{M}}_{k_{i}}}\big{)}\Big{\\}}\,,$
(7)
where ${\mathbb{P}}_{{\mathcal{M}_{k_{i}}}}$ represents the memory
distribution at $\mathcal{T}_{k_{i}}$. The proof is provided in Appendix-B
from SM. It notes that ${\rm F}_{S}(\cdot,\cdot)$ can be used for arbitrary
distributions. We assume an ideal model selection in Eq. (7), where always the
component with the minimal risk is chosen. DEM can achieve the minimal upper
bound to the risk (Left Hand Side (LHS) of Eq. (6)) when comparing with a
single model (Theorem 3),. Then, we derive the GB for analyzing the
generalization performance of ${\mathcal{G}}$ on multiple target
distributions.
Lemma 2.
For a given data stream $S=\bigcup^{C^{S}_{t}}_{j}{\mathcal{D}^{S}_{t,j}}$
consisting of samples from ${\mathcal{D}}^{S}_{t}$, we have a set of target
sets $\\{{\mathcal{D}}^{T}_{t,1},\cdots,{\mathcal{D}}^{T}_{t,C^{T}_{t}}\\}$,
where each ${\mathcal{D}}^{T}_{t,j}$ contains $C^{b}_{t,j}$ batches of
samples. Let ${\mathbb{P}}^{T}_{t,j}(d)$ represent the distribution of the
$d$-th batch of samples in ${\mathcal{D}}^{T}_{t,j}$. We assume that
$\mathcal{G}$ consists of $c$ components trained on samples from
$\mathcal{S}$, at ${\mathcal{T}}_{i}$. We derive the GB for multiple target
domains as :
$\displaystyle\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}\sum\nolimits_{d=1}^{C^{b}_{t,j}}{\mathcal{R}}\big{(}h,{\mathbb{P}}^{T}_{t,j}(d)\big{)}\Big{\\}}$
$\displaystyle\leq\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}\sum\nolimits_{d=1}^{C^{b}_{t,j}}\big{\\{}{\rm
F}_{S}\big{(}{\mathbb{P}}^{T}_{t,j}(d),{\mathcal{G}}\big{)}\big{\\}}\Big{\\}}.$
(8)
Remark.
We have several observations from Lemma 2 : 1) The generalization performance
of $\mathcal{G}$ is relying on the discrepancy distance between each target
distribution ${\mathbb{P}}^{T}_{t,j}$ and the memory distribution
${\mathbb{P}}_{{\mathcal{M}}_{k_{i}}}$ of the selected component (Also see
details in Appendix-C from SM). 2) Eq. (8) provides the analysis of the
trade-off between the model’s complexity and generalization for DEM [24, 38].
By adding new components, $\mathcal{G}$ would capture additional information
of each target distribution and thus improve its performance. On the other
hand, the selection process ensures a probabilistic diversity of stored
information, aiming to capture more knowledge with a minimal number of
components.
In practice, we usually perform the model selection for ${\mathcal{G}}$ by
using a certain criterion that only accesses the testing samples without task
labels. Therefore, we introduce a selection criterion ${\widehat{\rm
F}}(\cdot,\cdot)$, implemented by comparing the sample log-likelihood :
$\displaystyle{\widehat{\rm
F}}\big{(}{\mathbb{P}}^{T}_{t,j}(d),{\mathcal{G}}\big{)}$
$\displaystyle\buildrel\Delta\over{=}\mathop{\arg\max}\limits_{k_{1},\cdots,k_{c}}\big{\\{}{\mathbb{E}}_{{\bf
x}\sim{\mathbb{P}}^{T}_{t,j}(d)}[\hat{f}({\bf x},G_{k_{j}})]\big{\\}},$ (9)
where $j=\\{1,\cdots,c\\}$ and ${\hat{f}}(\cdot,\cdot)$ is a pre-defined
sample-log likelihood function. Then Eq. (9) is used for model selection:
$\displaystyle{\widehat{\rm
F}}_{S}\big{(}{\mathbb{P}}_{t,j}^{T}(d),{\mathcal{G}}\big{)}$
$\displaystyle=\big{\\{}{\mathcal{R}}\big{(}h,h_{{\mathcal{M}}_{s}},{\mathbb{P}}_{{\mathcal{M}_{s}}}\big{)}+{\mathcal{L}}_{\widehat{d}}\big{(}{\mathbb{P}}^{T,{\mathcal{X}}}_{t,j}(d),{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{s}}\big{)}+\eta\big{(}{\mathbb{P}}^{T}_{t,j}(d),{\mathbb{P}}_{{\mathcal{M}}_{s}}\big{)}\,|\,$
$\displaystyle s={\widehat{\rm
F}}\big{(}{\mathbb{P}}^{T}_{t,j}(d),{\mathcal{G}}\big{)}\big{\\}}\,.$ (10)
We rewrite Eq. (8) using Eq. (3.4), resulting in :
$\displaystyle\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}$
$\displaystyle\sum\nolimits_{d=1}^{C^{b}_{t,j}}{\mathcal{R}}\big{(}h,{\mathbb{P}}^{T}_{t,j}(d)\big{)}\Big{\\}}\leq\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}\sum\nolimits_{d=1}^{C^{b}_{t,j}}\big{\\{}{\widehat{\rm
F}}_{S}\big{(}{\mathbb{P}}_{t,j}^{T}(d),{\mathcal{G}}\big{)}\big{\\}}\Big{\\}}\,.$
(11)
Compared with an ideal solution (Eq. (8)), Eq. (11) would involve the extra
error terms caused by the selection process (Eq. (3.4)), expressed as
$\sum\nolimits_{j=1}^{C^{T}_{t}}\Big{\\{}\sum\nolimits_{d=1}^{C^{b}_{t,j}}\big{\\{}{{\widehat{\rm
F}}}_{S}\big{(}{\mathbb{P}}_{t,j}^{T}(d),{\mathcal{G}}\big{)}-{{\rm
F}}_{S}\big{(}{\mathbb{P}}_{t,j}^{T}(d),{\mathcal{G}}\big{)}\big{\\}}\Big{\\}}$.
In Section 4, we introduce a new CL framework according to the theoretical
analysis.
### 3.5 Parallels with other studies defining lifelong learning bounds
In this section, we discuss the differences between the results in this paper
and those from other studies proposing lifelong learning bounds. The bound
(Theorem 1) in [30] assumes that the model is trained on all previous samples,
which is not practical under a TFCL setting. In contrast, in Theorem 1 from
this paper, the model is trained on a memory buffer, which can be used in the
context of TFCL. In addition, the bounds in [30] mainly provide the
theoretical guarantees for the performance when learning a new data
distribution (Theorem 1 from [30]), which does not provide an analysis of the
model’s forgetting. In contrast, this paper studies the forgetting analysis of
our model and its theoretical developments can be easily extended for
analyzing the forgetting behaviour of a variety of continual learning methods,
while this cannot be said about the study from [30] (See Appendix-D from
supplemental material).
When comparing with [33], our theoretical analysis does not rely on explicit
task boundaries, which is a more realistic assumption for TFCL. In addition,
[33] employs the KL divergence to measure the distance between two
distributions, which would require knowing the explicit distribution form and
is thus hard to evaluate in practice during the learning. However, our theory
employs the discrepancy distance, which can be reliably estimated. Moreover,
the study from [33] does not develop a practical algorithm to be used
according to the theoretical analysis. In contrast, our theory provides
guidelines for the algorithm design under the TFCL assumption. Finally,
inspired by the proposed theoretical analysis, this paper develops a
successful continual learning approach for TFCL.
Theorem 1 in [4], similarly to [30], assumes that the model is trained on
uncertain data sets over time, which would not be suitable for TFCL since the
model can not access previously learnt samples under TFCL. In contrast,
Theorem 1 in our theory represents a realistic TFCL scenario in which the
model is trained on a fixed-length memory buffer and is evaluated on all
previously seen samples. Therefore, the analysis for the forgetting behaviour
of the model under TFCL relies on Theorem 1 and its consequences. In addition,
the study from [4], similarly to [30, 33], only provides a theoretical
guarantee for a single model. In contrast, we extend our theoretical analysis
to the dynamic expansion model, which motivates us to develop a novel
continual learning approach for TFCL. Moreover, this paper is the first work
to provide the theoretical analysis for the diversity of knowledge recorded by
different components (See Appendix-C from supplemental material). This
analysis indicates that by maintaining the knowledge diversity among different
components we ensure a good trade-off between the model’s complexity and
generalization performance, thus providing invaluable insights into algorithm
design for TFCL.
## 4 Methodology
### 4.1 Network architecture
Each component $G_{j}\in{\mathcal{G}}$ consists of a classifier $h_{j}$
implemented by a neural network $f_{\varsigma_{j}}({\bf x})$ with trainable
parameters $\varsigma_{j}$, and a variational autoencoder (VAE) model
implemented by an encoding network ${\rm enc}_{\omega_{j}}$ as well as the
corresponding decoding network ${\rm dec}_{\theta_{j}}$, with trainable
parameters $\\{\omega_{j},\theta_{j}\\}$. Due to its robust generation and
inference mechanisms, VAE has been widely used in many applications [48, 55,
49, 50, 51, 61]. This paper employs a VAE for discrepancy estimation and
component selection. The loss function for the $j$-th component at
$\mathcal{T}_{i}$ is defined as :
$\displaystyle{\mathcal{L}}_{class}(G_{j},{\mathcal{M}}_{i})\buildrel\Delta\over{=}\frac{1}{{|{\mathcal{M}}_{i}|}}\sum\nolimits_{t=1}^{|{\mathcal{M}}_{i}|}\big{\\{}{\mathcal{L}}_{ce}\big{(}h_{j}({\bf
x}_{t}),{y}_{t}\big{)}\big{\\}}\,,$ (12)
$\displaystyle{\mathcal{L}}_{VAE}(G_{j},{\mathcal{M}}_{i})\buildrel\Delta\over{=}$
$\displaystyle{\mathbb{E}_{{q_{{\omega_{j}}}}({\bf{z}}\,|\,{\bf{x}})}}\left[{\log{p_{{\theta_{j}}}}({\bf{x}}_{t}\,|\,{\bf{z}})}\right]-{D_{KL}}\left[{{q_{{\omega_{j}}}}({\bf{z}}\,|\,{\bf{x}}_{t})\,||\,p({\bf{z}})}\right],$
(13)
where $\\{{\bf x}_{t},y_{t}\\}\sim\mathcal{M}_{i}$, where $|\mathcal{M}_{i}|$
is the memory buffer size, and ${\mathcal{L}}_{ce}(\cdot)$ is the cross-
entropy loss. $\mathcal{L}_{VAE}(\cdot,\cdot)$ is the VAE loss [15],
${p_{{\theta_{j}}}}({\bf{x}}_{t}\,|\,{\bf{z}})$ and
${q_{{\omega_{j}}}}({\bf{z}}\,|\,{\bf{x}}_{t})$ are the encoding and decoding
distributions, implemented by ${\rm enc}_{\omega_{j}}$ and ${\rm
dec}_{\theta_{j}}$, respectively. We also implement $\hat{f}(\cdot,\cdot)$ in
Eq. (9) by $-\mathcal{L}_{VAE}(\cdot,\cdot)$ for the component selection at
the testing phase. The training algorithm for the proposed ODDL consists of
three stages (See more information in Appendix-E from SM). In the initial
training stage, we aim to learn the initial knowledge of the data stream and
preserve it into a frozen component ${G}_{1}$, which can provide information
for the dynamic expansion and sample selection evaluation in the subsequent
learning. During the evaluator training stage we train the current component
as the evaluator that estimates the discrepancy distance between each
previously learnt component and the memory buffer, providing appropriate
signals for the model expansion. In the sample selection stage, we train the
current component with new data, while we aim to promote knowledge diversity
among components.
In the following, we firstly propose a new dynamic expansion mechanism based
on the discrepancy criterion. Then we introduce a new sample selection
approach that can further improve the performance of the model. Finally, we
provide the detailed algorithm implementation.
### 4.2 Discrepancy based mixture model expansion
From Lemma 2, we observe that the probabilistic diversity of trained
components in $\mathcal{G}$ can ensure a compact network architecture while
maintaining a good generalization performance (See Theorem 4 in Appendix-C
from SM). In order to achieve this, we maximize the discrepancy between each
trained component of ${\mathcal{G}}$ and the current memory buffer, during the
training, by using the discrepancy distance (notations are defined in Theorem
3 ), expressed as :
$\displaystyle{\mathcal{M}}^{\star}=\mathop{\arg\max}\limits_{{\mathcal{M}}_{i}}\sum\nolimits_{j}^{\rm{c}}\big{\\{}{\mathcal{L}}_{\widehat{d}}({\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{k_{j}}},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}})\big{\\}}\,,$
(14)
where $i=k_{c}+1,\cdots,n$ represents the index of the training steps and
$k_{c}$ is the $c$-th component trained at ${\mathcal{T}}_{k_{c}}$.
${\mathcal{M}}^{\star}$ is the optimal memory. Eq. (14) can be seen as a
recursive optimization problem when ${\mathcal{G}}$ dynamically adds new
components ($c$ is increased) during the training. In order to provide a
practical way to solve Eq. (14) while balancing the model’s complexity and
performance, we derive an expansion criterion based on the discrepancy
distance :
$\displaystyle\min\Big{\\{}{\mathcal{L}}_{\widehat{d}}\big{(}{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{k_{j}}},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}\,|\,j=1,\cdots,c\Big{\\}}\geq\lambda\,,$
(15)
where $\lambda\in[0,4]$ is an architecture expansion threshold. If the current
memory distribution ${\mathbb{P}}_{{\mathcal{M}}_{i}}$ is sufficiently
different from each component (satisfies Eq. (15)), ${\mathcal{G}}$ will add a
new component to preserve the knowledge of the current memory
${\mathcal{M}}_{i}$, while also encouraging the probabilistic diversity among
the trained components.
In the following, we describe the implementation. We start by training the
first component $G_{1}$ which consists of a classifier $h_{1}$ (classification
task) and a VAE model $v_{1}$ (model selection at the testing phase). We also
train an additional component called the evaluator $G_{e}=\\{h_{e},v_{e}\\}$,
at the initial training stage, which aims to capture the future information
about the data stream. Once the memory $\mathcal{M}_{j}$ reaches its maximum
size at $\mathcal{T}_{j}$, we freeze the weights of the first component to
preserve the previously learnt knowledge while continually training the
evaluator during the following training steps. Since the evaluator continually
captures the knowledge from the current memory ${\mathcal{M}}_{i}$, we check
the model expansion using Eq. (15) at ${\mathcal{T}}_{i}$ :
$\displaystyle{\mathcal{L}}^{\star}_{d}\big{(}{\mathbb{P}}^{{\mathcal{X}}}_{{\mathcal{M}}_{k_{1}}},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}\geq\lambda\,.$
(16)
It notes that we can not access the previously learnt memory distribution
${\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{k_{1}}}$ and we approximate it by
using the auxiliary distribution ${\mathbb{P}}^{{\mathcal{X}}}_{v^{j}_{1}}$
formed by samples drawn from $v_{1}^{j}$ of $G_{1}^{j}$, where the superscript
$j$ represents the first component finishing the training at
${\mathcal{T}}_{j}$. ${\mathcal{L}}^{\star}_{d}(\cdot,\cdot)$ is the estimator
of the discrepancy distance, achieved by $h^{i}_{e}$ and $h^{j}_{1}$.
$\displaystyle{\mathcal{L}}^{\star}_{d}\big{(}{\mathbb{P}}^{{\mathcal{X}}}_{v^{j}_{1}},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}$
$\displaystyle\buildrel\Delta\over{=}\Big{|}{\mathbb{E}}_{{\bf
x}\sim{\mathbb{P}}^{{\mathcal{X}}}_{v^{j}_{1}}}\big{[}{\mathcal{L}}\big{(}h^{j}_{1}({\bf
x}),h^{i}_{e}({\bf x})\big{)}\big{]}-{\mathbb{E}}_{{\bf
x}\sim{\mathbb{P}}_{{\mathcal{M}}_{i}}}\big{[}{\mathcal{L}}\big{(}h^{j}_{1}({\bf
x}),h^{i}_{e}({\bf x})\big{)}\big{]}\Big{|}.$ (17)
If Eq. (16) is fulfilled, then we add $G^{i}_{e}$ into the model $\mathcal{G}$
while building a new evaluator $(G^{i+1}_{e}=G_{3}^{i+1})$ at
$\mathcal{T}_{i+1}$; otherwise, we train $G^{i}_{e}\to G^{i+1}_{e}$ at the
next training step, $\mathcal{T}_{i+1}$. Furthermore, for satisfying the
diversity of knowledge in the components from the mixture model (See Lemma 2),
we clear the memory $\mathcal{M}_{i}$ when performing the expansion in order
to ensure that we would learn non-overlapping distributions during the
following training steps. We can also extend Eq. (16) to the expansion
criterion for $\mathcal{G}=\\{G_{1},\cdots,G_{s-1}\\}$ that has already
preserved $(s-1)$ components, illustrated in Fig. 1:
$\displaystyle\min\Big{\\{}{\mathcal{L}}^{\star}_{d}\big{(}{\mathbb{P}}^{{\mathcal{X}}}_{v^{k_{j}}_{j}},{\mathbb{P}}^{\mathcal{X}}_{{\mathcal{M}}_{i}}\big{)}\,|\,j=1,\cdots,s-1\Big{\\}}\geq\lambda\,,\vspace{-20pt}$
(18)
where ${\mathbb{P}}^{{\mathcal{X}}}_{v^{k_{j}}_{j}}$ is the distribution of
the generated samples by $G^{k_{j}}_{j}$ and denote
$S_{j}={\mathcal{L}}^{\star}_{d}\big{(}{\mathbb{P}}^{{\mathcal{X}}}_{v^{k_{j}}_{j}},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}$.
### 4.3 Sample selection
Figure 1: We generate the knowledge by using the VAE ($V_{j}$) of each
previous component $G_{j},j=1,\cdots,s-1$, which is used to evaluate the
discrepancy distance
$S_{j}={\mathcal{L}}^{\star}_{d}\big{(}{\mathbb{P}}^{{\mathcal{X}}}_{v_{j}},{\mathbb{P}}_{{\mathcal{M}}_{i}}\big{)}$
at ${\mathcal{T}}_{i}$ (Eq. (17)) between $G_{j}$ and the memory buffer. Then
we use these discrepancy scores $\\{S_{1},\cdots,S_{s-1}\\}$ to check the
model expansion (Eq. (18))
According to Lemma 2, the probabilistic diversity of the knowledge accumulated
in the components is crucial for the performance. In the following, we also
introduce a novel sample selection approach from the memory buffer that
further encourages component diversity. The primary motivation behind the
proposed sample selection approach is that we desire to store in the memory
buffer those samples that are completely different from the data used for
training the other components. This mechanism enables the currently created
component to capture a different underlying data distribution. To implement
this goal, we estimate the discrepancy distance on a pair of samples as the
diversity score :
$\displaystyle{\cal L}_{d}^{s}({\cal G},{{\bf{x}}_{c}}){\rm{}}$
$\displaystyle\buildrel\Delta\over{=}\frac{1}{s}\sum\nolimits_{t=1}^{s}\big{|}{\cal
L}(h_{t}^{k_{t}}({\bf{x}}_{c}),h_{e}^{i}({\bf{x}}_{c}))-{\cal
L}(h_{t}^{k_{t}}({{\bf{x}}^{k_{t}}_{t}}),h_{e}^{i}({\bf{x}}^{k_{t}}_{t}))\big{|}\,,$
(19)
where ${\bf x}_{c}$ and ${\bf{x}}^{k_{t}}_{t}$ are the $c$-th sample from
$\mathcal{M}_{i}$ and the generated one drawn from $G^{k_{t}}_{t}$,
respectively. $h_{t}^{k_{t}}({\bf{\cdot}})$ is the classifier of
$G^{k_{t}}_{t}$ in $\mathcal{G}$. Eq. (19) evaluates the average discrepancy
distance between the knowledge generated by each trained component and the
stored samples, which guides for the sample selection at the training step
(${\mathcal{T}}_{i}$) as :
$\displaystyle\vspace{-10pt}{\mathcal{M}}_{i}=\bigcup\nolimits^{|{{\mathcal{M}}}^{\prime}_{i}|-b}_{j=1}{{\mathcal{M}}}^{\prime}_{i}[j]\,,$
(20)
where ${{\mathcal{M}}}^{\prime}_{i}$ is the sorted memory that satisfies the
condition
${\mathcal{L}}^{s}_{d}({\mathcal{G}},{{\mathcal{M}}}^{\prime}_{i}[a])>{\mathcal{L}}^{s}_{d}({\mathcal{G}},{{\mathcal{M}}}^{\prime}_{i}[q])$
for $a<q$. ${{\mathcal{M}}}^{\prime}_{i}[j]$ represents the $j$-th stored
sample and $b=10$ is the batch size. We name the approach with sample
selection as ODDL-S.
(a) Risk on the training set. (b) Risk on all target sets. (c) Results on
MNIST.
Figure 2: The forgetting analysis of a single model and DEM in (a) and (b)
where the batch size is 64 in the data stream. Data distribution shift and
increasing the number of components in ODDL-S during the training in (c).
### 4.4 Algorithm implementation
We provide the pseudocode ( Algorithm 1 ) in Appendix-E from SM. The
algorithm has three main stages :
* •
(Initial training).
We start by building two components $\\{G_{1},G_{2}\\}$ where we only add
$G_{1}$ in $\mathcal{G}$ and consider $G_{2}=G_{e}$ as the Evaluator. $G_{1}$
and $G_{2}$ are trained jointly using Eq. (13) and Eq. (12) during the initial
training stage until the memory $\mathcal{M}_{j}$ reaches its maximum size
$|{\mathcal{M}}|^{max}$ at a certain training step ${\mathcal{T}}_{j}$. Then
we freeze the first component $G_{1}$ and perform the second stage.
* •
(Evaluator training).
In this stage, we only update the Evaluator using Eq. (13) and Eq. (12) on
${\mathcal{M}}_{j+1}$ at ${\mathcal{T}}_{j+1}$. If
$|\mathcal{M}_{j+1}|\geq|\mathcal{M}|^{max}$, then we evaluate the discrepancy
distance using Eq. (17) to check the expansion (Eq. (18)). If the expansion
criterion is satisfied then we add $G_{e}$ to $\mathcal{G}$ and build a new
Evaluator while cleaning up the memory $\mathcal{M}_{j+1}$, otherwise, we
perform the sample selection.
* •
(Sample selection).
We evaluate the diversity score for each stored sample in $\mathcal{M}_{j+1}$
using Eq. (19). We then perform the sample selection for $\mathcal{M}_{j+1}$
using Eq. (20) and return back to the second stage.
## 5 Experiments
We perform the experiments to address the following research questions: 1)
What factors would cause the model’s forgetting, and how to explain such
behaviour? 2) How efficient is the proposed ODDL-S under TFCL benchmarks? 3)
How important is each module in OODL-S?
In this experiment, we adapt the TFCL setting from [8] which employs several
datasets including Split MNIST [22], Split CIFAR10 [18] and Split CIFAR100
[18]. The detailed information for datasets, hyperparameters and network
architectures is provided in Appendix-F from the supplementary material. The
code is available at https://dtuzi123.github.io/ODDL/.
Table 1: The accuracy of various continual learning models for five independent runs. Methods | Split MNIST | Split CIFAR10 | Split CIFAR100
---|---|---|---
finetune* | 19.75 $\pm$ 0.05 | 18.55 $\pm$ 0.34 | 3.53 $\pm$ 0.04
GEM* | 93.25 $\pm$ 0.36 | 24.13 $\pm$ 2.46 | 11.12 $\pm$ 2.48
iCARL* | 83.95 $\pm$ 0.21 | 37.32 $\pm$ 2.66 | 10.80 $\pm$ 0.37
reservoir* | 92.16 $\pm$ 0.75 | 42.48 $\pm$ 3.04 | 19.57 $\pm$ 1.79
MIR* | 93.20 $\pm$ 0.36 | 42.80 $\pm$ 2.22 | 20.00 $\pm$ 0.57
GSS* | 92.47 $\pm$ 0.92 | 38.45 $\pm$ 1.41 | 13.10 $\pm$ 0.94
CoPE-CE* | 91.77 $\pm$ 0.87 | 39.73 $\pm$ 2.26 | 18.33 $\pm$ 1.52
CoPE* | 93.94 $\pm$ 0.20 | 48.92 $\pm$ 1.32 | 21.62 $\pm$ 0.69
ER + GMED${\dagger}$ | 82.67 $\pm$ 1.90 | 34.84 $\pm$ 2.20 | 20.93 $\pm$ 1.60
ERa \+ GMED${\dagger}$ | 82.21 $\pm$ 2.90 | 47.47 $\pm$ 3.20 | 19.60 $\pm$ 1.50
CURL* | 92.59 $\pm$ 0.66 | - | -
CNDPM* | 93.23 $\pm$ 0.09 | 45.21 $\pm$ 0.18 | 20.10 $\pm$ 0.12
Dynamic-OCM | 94.02 $\pm$ 0.23 | 49.16 $\pm$ 1.52 | 21.79 $\pm$ 0.68
ODDL | 94.85 $\pm$ 0.02 | 51.48 $\pm$ 0.12 | 26.20 $\pm$ 0.72
ODDL-S | 95.75 $\pm$ 0.05 | 52.69 $\pm$ 0.11 | 27.21 $\pm$ 0.87
### 5.1 Empirical results for the forgetting analysis
In this section, we investigate the forgetting behaviour of the model
according to the proposed theoretical framework. Firstly, we train a single
classifier $h$ under Split MNIST database, as a baseline, with a memory buffer
of the maximum size of 2000, and we randomly remove a batch of stored samples
(batch size is 10) when the memory is full. Then we estimate
${\mathcal{R}}(h,{\mathcal{P}}_{i})$ (target risk on all visited training
samples),
${\mathcal{R}}(h,h_{{\mathcal{M}}_{i}},{\mathbb{P}}_{{\mathcal{M}}_{i}})$
(source risk on the memory). We plot the results in Fig. 2-a, where “Random
(Source risk)” represents the source risk of a single classifier. The results
show that the source risk always keeps stable, and the target risk is small
for a few initial training steps since the memory can capture all information
of visited samples. However, as the number of training steps grows, the target
risk is increased, which is caused by the memory that loses previous samples,
theoretically explained in Theorem 1. We also evaluate the risk of the
baseline on all testing sets (target risk) and plot the results in Fig. 2-b
where it is observed that the single model always leads to a large target risk
during the training.
We evaluate the source and target risks for the proposed ODDL under Split
MNIST and plot the results in Fig. 2. The performance of ODDL on the
distribution ${\mathcal{P}}_{i}$ (Target risk) does not degenerate during the
whole training phase. At the same time, the baseline tends to increase the
target risk on ${\mathcal{P}}_{i}$ as the training steps increase, as shown in
Fig. 2-a. Finally, the risk on all target distributions (all categories in the
testing set) from each training step is shown in Fig. 2-b, where the proposed
ODDL minimizes the target risk as gaining more knowledge during the training.
In contrast, the baseline invariably leads to a large target risk even when
the number of training steps increases. These results show that ODDL can
relieve forgetting and achieve better generalization than the random approach
on all target sets.
### 5.2 Results on TFCL benchmark
We provide the results in Tab. 1 where * and ${\dagger}$ denote the results
cited from [8] and [12], respectively. We compare with several baselines
including: finetune that directly trains a classifier on the data stream, GSS
[1], Dynamic-OCM [54], MIR [2], Gradient Episodic Memory (GEM) [27],
Incremental Classifier and Representation Learning (iCARL) [39], Reservoir
[45], CURL, CNDPM, CoPE, ER + GMED and ERa \+ GMED [12] where ER is the
Experience Replay [42] and ERa is ER with data augmentation. The number of
components in ODDL-S and ODDL for Split MNIST, Split CIFAR10 and Split
CIFAR100 is 7, 9, 7, respectively. The proposed approach outperforms CNDPM,
which uses more parameters, on all datasets and achieves state-of-the-art
performance.
Table 2: Classification accuracy for 20 runs when testing various models on Split MImageNet and Permuted MNIST. Methods | Split MImageNet | Permuted MNIST
---|---|---
ERa | 25.92 $\pm$ 1.2 | 78.11 $\pm$ 0.7
ER + GMED | 27.27 $\pm$ 1.8 | 78.86 $\pm$ 0.7
MIR+GMED | 26.50 $\pm$ 1.3 | 79.25 $\pm$ 0.8
MIR | 25.21 $\pm$ 2.2 | 79.13 $\pm$ 0.7
CNDPM | 27.12 $\pm$1.5 | 80.68 $\pm$ 0.7
ODDL | 27.45 $\pm$ 0.9 | 82.33 $\pm$ 0.6
ODDL-S | 28.68 $\pm$ 1.5 | 83.56 $\pm$ 0.5
In the following, we also evaluate the performance of the proposed approach on
the large-scale dataset (MINI-ImageNet [21]), and Permuted MNIST [9]. We split
MINI-ImageNet into 20 tasks (See details in Appendix-F1 from SM), namely Split
MImageNet. We follow the setting from [2] where the maximum memory size is
$10K$, and a smaller version of ResNet-18 [10] is used as the classifier. The
hyperparameter $\lambda$ used for learning Split MINI-ImageNet and Permuted
MNIST is equal to 1.2 and 1.5, respectively. We compare with several state-of-
the-art methods under Split MImageNet, reported in Tab. 2, where the results
of other baselines are cited from [12]. These results show that the proposed
ODDL-S outperforms different baselines under the challenging dataset.
### 5.3 Ablation study
We investigate whether the proposed discrepancy-based criterion can provide
better signals for the expansion of ODDL-S. We train ODDL-S on Split MNIST,
where we record the variance of tasks and the number of components in each
training step. We plot the results in Fig. 2-c where “task” represents the
number of tasks in each training step. We observe that the proposed
discrepancy-based criterion can detect the data distribution shift accurately,
allowing ODDL-S to expand the network architecture each time when detecting
the data distribution shift. This also encourages the proposed ODDL-S to use a
minimal number of components while achieving optimal performance, as discussed
in Lemma 2. We also provide the analysis of how to maximize the memory bound
in Appendix-F2.2, while more ablation study results are provided in
Appendix-F.2 from the supplementary material.
## 6 Conclusion
In this paper, we develop a novel theoretical framework for Task-Free
Continual Learning (TFCL), by defining a statistical discrepancy distance.
Inspired by the theoretical analysis, we propose the Online Discrepancy
Distance Learning enabled by a memory buffer sampling (ODDL-S) model, which
trades off between the model’s complexity and performance. The memory buffer
sampling mechanism ensures the information diversity learning. The proposed
theoretical analysis provides new insights into the model’s forgetting
behaviour during each training step of TFCL. Experimental results on several
TFCL benchmarks show that the proposed ODDL-S achieves state-of-the-art
performance.
## References
* [1] R. Aljundi, M. Lin, B. Goujaud, and Y. Bengio. Gradient based sample selection for online continual learning. In Advances in Neural Information Processing Systems (NeurIPS), arXiv preprint arXiv:1903.08671, 2019.
* [2] Rahaf Aljundi, Eugene Belilovsky, Tinne Tuytelaars, Laurent Charlin, Massimo Caccia, Min Lin, and Lucas Page-Caccia. Online continual learning with maximal interfered retrieval. In Advances in Neural Information Processing Systems (NeurIPS), arXiv preprint arXiv:1908.04742, 2019.
* [3] Rahaf Aljundi, Klaas Kelchtermans, and Tinne Tuytelaars. Task-free continual learning. In Proc. of IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pages 11254–11263, 2019.
* [4] Verónica Álvarez, Santiago Mazuelas, and José Antonio Lozano. Minimax classification under concept drift with multidimensional adaptation and performance guarantees. In Proc. International Conference on Machine Learning (ICML), vol. PMLR 162, pages 486–499, 2022.
* [5] Jihwan Bang, Heesu Kim, YoungJoon Yoo, Jung-Woo Ha, and Jonghyun Choi. Rainbow memory: Continual learning with a memory of diverse samples. In Proc. of IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), pages 8218–8227, 2021.
* [6] A. Chaudhry, M. Rohrbach, M. Elhoseiny, T. Ajanthan, P. Dokania, P. H. S. Torr, and M.’A. Ranzato. On tiny episodic memories in continual learning. arXiv preprint arXiv:1902.10486, 2019.
* [7] W. Dai, Q. Yang, G. R. Xue, and Y. Yu. Boosting for transfer learning. In Proc. Int Conf. on Machine Learning (ICML), pages 193–200, 2007\.
* [8] Matthias De Lange and Tinne Tuytelaars. Continual prototype evolution: Learning online from non-stationary data streams. In Proc. of the IEEE/CVF Int. Conf. on Computer Vision (ICCV), pages 8250–8259, 2021.
* [9] Ian J Goodfellow, Mehdi Mirza, Da Xiao, Aaron Courville, and Yoshua Bengio. An empirical investigation of catastrophic forgetting in gradient-based neural networks. In arXiv preprint arXiv:1312.6211, 2014.
* [10] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proc. of IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2016.
* [11] G. Hinton, O. Vinyals, and J. Dean. Distilling the knowledge in a neural network. In Proc. NIPS Deep Learning Workshop, arXiv preprint arXiv:1503.02531, 2014.
* [12] Xisen Jin, Arka Sadhu, Junyi Du, and Xiang Ren. Gradient-based editing of memory examples for online task-free continual learning. In Advances in Neural Information Processing Systems (NeurIPS), arXiv preprint arXiv:2006.15294, 2021.
* [13] H. Jung, J. Ju, M. Jung, and J. Kim. Less-forgetting learning in deep neural networks. In Proc. AAAI Conf. on Artificial Intelligence, volume 32, pages 3358–3365, 2018.
* [14] Zixuan Ke, Bing Liu, Nianzu Ma, Hu Xu, and Lei Shu. Achieving forgetting prevention and knowledge transfer in continual learning. Advances in Neural Information Processing Systems, 34, 2021.
* [15] D. P. Kingma and M. Welling. Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114, 2013.
* [16] J. Kirkpatrick, R. Pascanu, N. Rabinowitz, J. Veness, G. Desjardins, A. A. Rusu, K. Milan, J. Quan, T. Ramalho, A. Grabska-Barwinska, D. Hassabis, C. Clopath, D. Kumaran, and R. Hadsell. Overcoming catastrophic forgetting in neural networks. Proc. of the National Academy of Sciences (PNAS), 114(13):3521–3526, 2017.
* [17] Jeremias Knoblauch, Hisham Husain, and Tom Diethe. Optimal continual learning has perfect memory and is NP-hard. In Proc. Int. Conf. on Machine Learning (ICML), vol PMLR 119, pages 5327–5337, 2020.
* [18] Alex Krizhevsky and Geoffrey Hinton. Learning multiple layers of features from tiny images. Technical report, Univ. of Toronto, 2009.
* [19] Richard Kurle, Botond Cseke, Alexej Klushyn, Patrick van der Smagt, and Stephan Günnemann. Continual learning with Bayesian neural networks for non-stationary data. In Int. Conf. on Learning Representations (ICLR), 2020.
* [20] Sampo Kuutti, Richard Bowden, Yaochu Jin, Phil Barber, and Saber Fallah. A survey of deep learning applications to autonomous vehicle control. IEEE Transactions on Intelligent Transportation Systems, 22(2):712–733, 2020.
* [21] Ya Le and Xuan Yang. Tiny imageNet visual recognition challenge. Technical report, Univ. of Stanford, 2015.
* [22] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proc. of the IEEE, 86(11):2278–2324, 1998.
* [23] Sebastian Lee, Sebastian Goldt, and Andrew Saxe. Continual learning in the teacher-student setup: Impact of task similarity. In Proc. Int. Conf. on Machine Learning (ICML), vol. PMLR 139, pages 6109–6119, 2021.
* [24] Soochan Lee, Junsoo Ha, Dongsu Zhang, and Gunhee Kim. A neural Dirichlet process mixture model for task-free continual learning. In Int. Conf. on Learning Representations (ICLR), arXiv preprint arXiv:2001.00689, 2020.
* [25] Z. Li and D. Hoiem. Learning without forgetting. IEEE Trans. on Pattern Analysis and Machine Intelligence, 40(12):2935–2947, 2017.
* [26] Sen Lin, Li Yang, Deliang Fan, and Junshan Zhang. TRGP: Trust region gradient projection for continual learning. In Int. Conf. on Learning Representations (ICLR), arXiv preprint arXiv:2202.02931, 2022.
* [27] David Lopez-Paz and Marc’Aurelio Ranzato. Gradient episodic memory for continual learning. In Advances in Neural Information Processing Systems, pages 6467–6476, 2017.
* [28] Divyam Madaan, Jaehong Yoon, Yuanchun Li, Yunxin Liu, and Sung Ju Hwang. Representational continuity for unsupervised continual learning. In Int. Conf. on Learning Representations (ICLR), arXiv preprint arXiv:2110.06976, 2022.
* [29] Yishay Mansour, Mehryar Mohri, and Afshin Rostamizadeh. Domain adaptation: Learning bounds and algorithms. In Proc. Conf. on Learning Theory (COLT), arXiv preprint arXiv:2002.06715, 2009.
* [30] Mehryar Mohri and Andres Munoz Medina. New analysis and algorithm for learning with drifting distributions. In International Conference on Algorithmic Learning Theory, pages 124–138. Springer, 2012.
* [31] Cuong V Nguyen, Yingzhen Li, Thang D Bui, and Richard E Turner. Variational continual learning. In Int. Conf. on Learning Representations (ICLR), arXiv preprint arXiv:1710.10628, 2018.
* [32] G. I. Parisi, R. Kemker, J. L. Part, C. Kanan, and S. Wermter. Continual lifelong learning with neural networks: A review. Neural Networks, 113:54–71, 2019.
* [33] Anastasia Pentina and Christoph H Lampert. Lifelong learning with non-iid tasks. In Proc. Advances in Neural Information Processing Systems (NIPS), pages 1540–1548, 2015.
* [34] R. Polikar, L. Upda, S. S. Upda, and Vasant Honavar. Learn++: An incremental learning algorithm for supervised neural networks. IEEE Trans. on Systems Man and Cybernetics, Part C, 31(4):497–508, 2001.
* [35] Mozhgan PourKeshavarzi, Guoying Zhao, and Mohammad Sabokrou. Looking back on learned experiences for class/task incremental learning. In Int. Conf. on Learning Representations (ICLR), 2022.
* [36] Ameya Prabhu, Philip HS Torr, and Puneet K Dokania. GDumb: A simple approach that questions our progress in continual learning. In Proc. European Conference on Computer Vision (ECCV), vol. LNCS 12347, pages 524–540, 2020.
* [37] Krishnan Raghavan and Prasanna Balaprakash. Formalizing the generalization-forgetting trade-off in continual learning. Advances in Neural Information Processing Systems, 34, 2021.
* [38] Dushyant Rao, Francesco Visin, Andrei A. Rusu, Yee Whye Teh, Razvan Pascanu, and Raia Hadsell. Continual unsupervised representation learning. In Proc. Neural Inf. Proc. Systems (NIPS), pages 7645–7655, 2019\.
* [39] Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H Lampert. iCaRL: Incremental classifier and representation learning. In Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), pages 2001–2010, 2017.
* [40] B. Ren, H. Wang, J. Li, and H. Gao. Life-long learning based on dynamic combination model. Applied Soft Computing, 56:398–404, 2017.
* [41] Hippolyt Ritter, Aleksandar Botev, and David Barber. Online structured Laplace approximations for overcoming catastrophic forgetting. In Advances in Neural Information Processing Systems (NeurIPS), volume 31, pages 3742–3752, 2018.
* [42] David Rolnick, Arun Ahuja, Jonathan Schwarz, Timothy P. Lillicrap, and Gregory Wayne. Experience replay for continual learning. In Advances in Neural Information Processing Systems 34 (NeurIPS), pages 348–358, 2019.
* [43] H. Shin, J. K. Lee, J. Kim, and J. Kim. Continual learning with deep generative replay. In Advances in Neural Inf. Proc. Systems (NIPS), pages 2990–2999, 2017.
* [44] Shengyang Sun, Daniele Calandriello, Huiyi Hu, Ang Li, and Michalis Titsias. Information-theoretic online memory selection for continual learning. In Int. Conf. on Learning Representations (ICLR), arXiv preprint arXiv:2204.04763, 2022.
* [45] Jeffrey S Vitter. Random sampling with a reservoir. ACM Transactions on Mathematical Software (TOMS), 11(1):37–57, 1985\.
* [46] Fei Ye and Adrian G. Bors. Learning latent representations across multiple data domains using lifelong VAEGAN. In Proc. European Conf. on Computer Vision (ECCV), vol. LNCS 12365, pages 777–795, 2020.
* [47] Fei Ye and Adrian G. Bors. Lifelong learning of interpretable image representations. In Proc. Int. Conf. on Image Processing Theory, Tools and Applications (IPTA), pages 1–6, 2020.
* [48] Fei Ye and Adrian G Bors. Mixtures of variational autoencoders. In Proc. Int. Conf. on Image Processing Theory, Tools and Applications (IPTA), pages 1–6, 2020.
* [49] Fei Ye and Adrian G. Bors. InfoVAEGAN: Learning joint interpretable representations by information maximization and maximum likelihood. In Proc. IEEE Int. Conf. on Image Processing (ICIP), pages 749–753, 2021.
* [50] Fei Ye and Adrian G Bors. Learning joint latent representations based on information maximization. Information Sciences, 567:216–236, 2021.
* [51] Fei Ye and Adrian G. Bors. Lifelong infinite mixture model based on knowledge-driven Dirichlet process. In Proc. of the IEEE Int. Conf. on Computer Vision (ICCV), pages 10695–10704, 2021.
* [52] Fei Ye and Adrian G. Bors. Lifelong mixture of variational autoencoders. IEEE Transactions on Neural Networks and Learning Systems, pages 1–14, 2021.
* [53] Fei Ye and Adrian G. Bors. Lifelong twin generative adversarial networks. In Proc. IEEE Int. Conf. on Image Processing (ICIP), pages 1289–1293, 2021.
* [54] Fei Ye and Adrian G Bors. Continual variational autoencoder learning via online cooperative memorization. arXiv preprint arXiv:2207.10131, 2022.
* [55] Fei Ye and Adrian G. Bors. Deep mixture generative autoencoders. IEEE Transactions on Neural Networks and Learning Systems, 33(10):5789–5803, 2022.
* [56] Fei Ye and Adrian G Bors. Learning an evolved mixture model for task-free continual learning. arXiv preprint arXiv:2207.05080, 2022.
* [57] Fei Ye and Adrian G. Bors. Lifelong generative modelling using dynamic expansion graph model. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8):8857–8865, Jun. 2022.
* [58] Fei Ye and Adrian G. Bors. Lifelong teacher-student network learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(10):6280–6296, 2022.
* [59] Jaehong Yoon, Divyam Madaan, Eunho Yang, and Sung Ju Hwang. Online coreset selection for rehearsal-based continual learning. In Int. Conf. on Learning Representations (ICLR), arXiv preprint arXiv:2106.01085, 2022.
* [60] Yuchen Zhang, Tianle Liu, Mingsheng Long, and Michael Jordan. Bridging theory and algorithm for domain adaptation. In Proc. International Conference on Machine Learning (ICML), vol. PMLR 97, pages 7404–7413, 2019.
* [61] Shengjia Zhao, Jiaming Song, and Stefano Ermon. InfoVAE: Balancing learning and inference in variational autoencoders. In Proc. AAAI Conf. on Artif. Intel., volume 33, pages 5885–5892, 2019.
|
111 Mathematics Subject Classification (2020). Primary: 34L40; Secondary:
34L20. Keywords: Dirac operator, degenerate boundary conditions, spectrum
On the spectrum of $2\times 2$ Dirac operator
with degenerate boundary conditions
Alexander Makin
Peoples Friendship University of Russia
117198, Miklukho-Maklaya str. 6, Moscow, Russia
> We study the spectral problem for the Dirac operator with degenerate
> boundary conditions and a complex-valued summable potential. Sufficient
> conditions are found under which the spectrum of the problem under
> consideration coincides with the spectrum of the corresponding unperturbed
> operator.
1\. Introduction
In the present paper, we study the Dirac system
$B\mathbf{y}^{\prime}+V\mathbf{y}=\lambda\mathbf{y},$ $None$
where
$B=\begin{pmatrix}0&1\\\ -1&0\end{pmatrix},\quad V=\begin{pmatrix}p(x)&q(x)\\\
q(x)&-p(x)\end{pmatrix},$
the functions $p,q\in L_{1}(0,\pi)$, with two-point boundary conditions
$\begin{array}[]{c}U_{1}(\mathbf{y})=a_{11}y_{1}(0)+a_{12}y_{2}(0)+a_{13}y_{1}(\pi)+a_{14}y_{2}(\pi)=0,\\\
U_{2}(\mathbf{y})=a_{21}y_{1}(0)+a_{22}y_{2}(0)+a_{23}y_{1}(\pi)+a_{24}y_{2}(\pi)=0,\end{array}$
$None$
where the coefficients $a_{jk}$ are arbitrary complex numbers, and rows of the
matrix
$A=\begin{pmatrix}a_{11}&a_{12}&a_{13}&a_{14}\\\
a_{21}&a_{22}&a_{23}&a_{24}\end{pmatrix}$
are linearly independent.
We consider the operator
$\mathbb{L}\mathbf{y}=B\mathbf{y}^{\prime}+V\mathbf{y}$ as a linear operator
on the space $\mathbb{H}=L_{2}(0,\pi)\oplus L_{2}(0,\pi)$, with the domain
$D(\mathbb{L})=\\{\mathbf{y}\in
W_{1}^{1}[0,\pi]:\,\mathbb{L}\mathbf{y}\in\mathbb{H}$, $U_{j}(\mathbf{y})=0$
$(j=1,2)\\}$.
Denote by $J_{jk}$ the determinant composed of the $j$th and $k$th columns of
the matrix $A$. Denote $J_{0}=J_{12}+J_{34}$, $J_{1}=J_{14}-J_{23}$,
$J_{2}=J_{13}+J_{24}$.
Boundary conditions (2) are called degenerate if
$J_{1}=J_{2}=0;\quad J_{0}=0,\quad J_{1}+iJ_{2}\neq 0,\quad
J_{1}-iJ_{2}=0;\quad J_{0}=0,\quad J_{1}+iJ_{2}=0,\quad J_{1}-iJ_{2}\neq 0,$
otherwise they are nondegenerate.
There is an enormous literature related to the spectral theory for Dirac
operators with nondegenerate boundary conditions. The case of degenerate
conditions has been investigated much less although in the last decade
interest in the study of these spectral problems has increased significantly
[see 1-9 and the references therein]. The main goal of present paper is to
establish conditions on the potential $V$ under which the spectrum of the
problem under consideration with degenerate boundary conditions coincides with
the spectrum of the corresponding unperturbed operator (3)
$B\mathbf{y}^{\prime}=\lambda\mathbf{y},\quad U(\mathbf{y})=0.$ $None$
2\. Main results
Theorem 1. Suppose the following conditions are valid
$J_{14}=J_{23}=J_{13}+J_{24}=0,$ $None$
$p(\pi-x)=-p(x),\quad q(\pi-x)=q(x),$ $None$
where $0\leq x\leq\pi$. Then the spectrum of problem (1) (2) coincides with
the spectrum of the corresponding unperturbed operator (3).
Proof. First of all, we rewrite system (1) in scalar form
$\left\\{\begin{array}[]{rcl}y_{2}^{\prime}+p(x)y_{1}+q(x)y_{2}=\lambda
y_{1}\\\ -y_{1}^{\prime}+q(x)y_{1}-p(x)y_{2}=\lambda y_{2}.\\\
\end{array}\right.$ $None$
Denote by
$E(x,\lambda)=\begin{pmatrix}e_{11}(x,\lambda)&e_{12}(x,\lambda)\\\
e_{21}(x,\lambda)&e_{22}(x,\lambda)\end{pmatrix}$ $None$
the matrix of the fundamental solution system to system (1) with boundary
condition $E(\frac{\pi}{2},\lambda)=I$, where $I$ is the unit matrix. It is
well known that
$e_{11}(x,\lambda)e_{22}(x,\lambda)-e_{12}(x,\lambda)e_{21}(x,\lambda)=1$
$None$
for any $x,\lambda$.
Substituting the first column of matrix (7) in system (6), we obtain
$\left\\{\begin{array}[]{rcl}e_{21}^{\prime}(x,\lambda)+p(x)e_{11}(x,\lambda)+q(x)e_{21}(x,\lambda)=\lambda
e_{11}(x,\lambda)\\\
-e_{11}^{\prime}(x,\lambda)+q(x)e_{11}(x,\lambda)-p(x)e_{21}(x,\lambda)=\lambda
e_{21}(x,\lambda).\\\ \end{array}\right.$ $None$
Replacing $x=\pi-t$ in relations (9), we find
$\left\\{\begin{array}[]{rcl}-e_{21}^{\prime}(\pi-t,\lambda)+p(\pi-t)e_{11}(\pi-t,\lambda)+q(\pi-t)e_{21}(\pi-t,\lambda)=\lambda
e_{11}(\pi-t,\lambda)\\\
e_{11}^{\prime}(\pi-t,\lambda)+q(\pi-t)e_{11}(\pi-t,\lambda)-p(\pi-t)e_{21}(\pi-t,\lambda)=\lambda
e_{21}(\pi-t,\lambda).\\\ \end{array}\right.$ $None$
It follows from (5) and (10) that
$\left\\{\begin{array}[]{rcl}-e_{21}^{\prime}(\pi-t,\lambda)-p(t)e_{11}(\pi-t,\lambda)+q(t)e_{21}(\pi-t,\lambda)=\lambda
e_{11}(\pi-t,\lambda)\\\
e_{11}^{\prime}(\pi-t,\lambda)+q(t)e_{11}(\pi-t,\lambda)+p(t)e_{21}(\pi-t,\lambda)=\lambda
e_{21}(\pi-t,\lambda).\\\ \end{array}\right.$ $None$
Denote $z_{2}(t,\lambda)=e_{11}(\pi-t,\lambda)$,
$z_{1}(t,\lambda)=e_{21}(\pi-t,\lambda)$. It follows from (11) that
$\left\\{\begin{array}[]{rcl}z_{2}^{\prime}+p(t)z_{1}+q(t)z_{2}=\lambda
z_{1}\\\ -z_{1}^{\prime}+q(t)z_{1}-p(t)z_{2}=\lambda z_{2}.\\\
\end{array}\right.$ $None$
Obviously, $z_{2}(\frac{\pi}{2},\lambda)=1$, $z_{1}(\frac{\pi}{2},\lambda)=0$,
therefore, by virtue of the uniqueness of the solution to the Cauchy problems
for systems (9) and (12) $z_{2}(t,\lambda)=e_{22}(t,\lambda)$,
$z_{1}(t,\lambda)=e_{21}(t,\lambda)$, hence, we obtain
$e_{22}(t,\lambda)=e_{11}(\pi-t,\lambda),\quad
e_{21}(t,\lambda)=e_{21}(\pi-t,\lambda).$
The last relations imply
$e_{21}(\pi)=e_{12}(0),\quad e_{12}(\pi)=e_{21}(0),\quad
e_{11}(\pi)=e_{22}(0),\quad e_{22}(\pi)=e_{11}(0).$ $None$
The eigenvalues of problem (1), (2) are the roots of the characteristic
equation
$\Delta(\lambda)=0,$
where
$\Delta(\lambda)=\left|\begin{array}[]{cccc}U_{1}(E^{[1]}(\cdot,\lambda))&U_{1}(E^{[2]}(\cdot,\lambda))\\\
U_{2}(E^{[1]}(\cdot,\lambda))&U_{2}(E^{[2]}(\cdot,\lambda))\\\
\end{array}\right|,$
$E^{[k]}(x,\lambda)$ is the $k$th column of matrix (7). Simple computations
together with relations (4), (8), and (13) show that
$\Delta(\lambda)=\left|\begin{array}[]{cccc}a_{11}e_{11}(0)+a_{12}e_{21}(0)+a_{13}e_{11}(\pi)+a_{14}e_{21}(\pi)&a_{11}e_{12}(0)+a_{12}e_{22}(0)+a_{13}e_{12}(\pi)+a_{14}e_{22}(\pi)\\\
a_{21}e_{11}(0)+a_{22}e_{21}(0)+a_{23}e_{11}(\pi)+a_{24}e_{21}(\pi)&a_{21}e_{12}(0)+a_{22}e_{22}(0)+a_{23}e_{12}(\pi)+a_{24}e_{22}(\pi)\\\
\end{array}\right|$
$\begin{array}[]{c}=[a_{11}e_{11}(0)+a_{12}e_{21}(0)+a_{13}e_{11}(\pi)+a_{14}e_{21}(\pi)][a_{21}e_{12}(0)+a_{22}e_{22}(0)+a_{23}e_{12}(\pi)+a_{24}e_{22}(\pi)]\\\
-[a_{21}e_{11}(0)+a_{22}e_{21}(0)+a_{23}e_{11}(\pi)+a_{24}e_{21}(\pi)][a_{11}e_{12}(0)+a_{12}e_{22}(0)+a_{13}e_{12}(\pi)+a_{14}e_{22}(\pi)]\\\
\\\
=[a_{11}e_{11}(0)+a_{12}e_{21}(0)+a_{13}e_{22}(0)+a_{14}e_{12}(0)][a_{21}e_{12}(0)+a_{22}e_{22}(0)+a_{23}e_{21}(0)+a_{24}e_{11}(0)]\\\
-[a_{21}e_{11}(0)+a_{22}e_{21}(0)+a_{23}e_{22}(0)+a_{24}e_{12}(0)][a_{11}e_{12}(0)+a_{12}e_{22}(0)+a_{13}e_{21}(0)+a_{14}e_{11}(0)]\\\
\\\
=\par[a_{11}e_{11}(0)a_{21}e_{12}(0)+a_{11}e_{11}(0)a_{22}e_{22}(0)+a_{11}e_{11}(0)a_{23}e_{21}(0)+a_{11}e_{11}(0)a_{24}e_{11}(0)\\\
+a_{12}e_{21}(0)a_{21}e_{12}(0)+a_{12}e_{21}(0)a_{22}e_{22}(0)+a_{12}e_{21}(0)a_{23}e_{21}(0)+a_{12}e_{21}(0)a_{24}e_{11}(0)\\\
+a_{13}e_{22}(0)a_{21}e_{12}(0)+a_{13}e_{22}(0)a_{22}e_{22}(0)+a_{13}e_{22}(0)a_{23}e_{21}(0)+a_{13}e_{22}(0)a_{24}e_{11}(0)\\\
+a_{14}e_{12}(0)a_{21}e_{12}(0)+a_{14}e_{12}(0)a_{22}e_{22}(0)+a_{14}e_{12}(0)a_{23}e_{21}(0)+a_{14}e_{12}(0)a_{24}e_{11}(0)]\\\
-[a_{21}e_{11}(0)a_{11}e_{12}(0)+a_{21}e_{11}(0)a_{12}e_{22}(0)+a_{21}e_{11}(0)a_{13}e_{21}(0)+a_{21}e_{11}(0)a_{14}e_{11}(0)\\\
+a_{22}e_{21}(0)a_{11}e_{12}(0)+a_{22}e_{21}(0)a_{12}e_{22}(0)+a_{22}e_{21}(0)a_{13}e_{21}(0)+a_{22}e_{21}(0)a_{14}e_{11}(0)\\\
+a_{23}e_{22}(0)a_{11}e_{12}(0)+a_{23}e_{22}(0)a_{12}e_{22}(0)+a_{23}e_{22}(0)a_{13}e_{21}(0)+a_{23}e_{22}(0)a_{14}e_{11}(0)\\\
+a_{24}e_{12}(0)a_{11}e_{12}(0)+a_{24}e_{12}(0)a_{12}e_{22}(0)+a_{24}e_{12}(0)a_{13}e_{21}(0)+a_{24}e_{12}(0)a_{14}e_{11}(0)]\\\
\\\ =\par
e_{11}(0)e_{22}(0)(a_{11}a_{22}+a_{13}a_{24}-a_{21}a_{12}-a_{23}a_{14})+e^{2}_{11}(0)(a_{11}a_{24}-a_{21}a_{14})+e^{2}_{22}(0)(a_{13}a_{22}-a_{23}a_{12})\\\
+e_{12}(0)e_{21}(0)(a_{12}a_{21}+a_{14}a_{23}-a_{24}a_{13}-a_{22}a_{11})+e_{11}(0)e_{12}(0)(a_{11}a_{21}+a_{14}a_{24}-a_{21}a_{11}-a_{24}a_{14})\\\
+e_{11}(0)e_{21}(0)(a_{11}a_{23}+a_{12}a_{24}-a_{21}a_{13}-a_{22}a_{14})\\\
+e^{2}_{21}(0)(a_{12}a_{23}-a_{22}a_{13})+e^{2}_{12}(0)(a_{14}a_{21}-a_{24}a_{11})\\\
+e_{21}(0)e_{22}(0)(a_{12}a_{22}+a_{13}a_{23}-a_{22}a_{12}-a_{23}a_{13})\\\
+e_{22}(0)e_{12}(0)(a_{13}a_{21}+a_{14}a_{22}-a_{23}a_{11}-a_{24}a_{12})\\\
\\\ =\par
e_{11}(0)e_{22}(0)(J_{12}+J_{34})+e^{2}_{11}(0)J_{14}+e^{2}_{22}(0)J_{32}\\\
+e_{12}(0)e_{21}(0)(J_{21}+J_{43})+e_{11}(0)e_{21}(0)(J_{13}+J_{24})\\\
+e^{2}_{21}(0)J_{23}+e^{2}_{12}(0)J_{41}+e_{22}(0)e_{12}(0)(J_{31}+J_{42})\\\
\\\
=\par[e_{11}(0)e_{22}(0)-e_{12}(0)e_{21}(0)](J_{12}+J_{34})+e^{2}_{11}(0)J_{14}+e^{2}_{22}(0)J_{32}+e^{2}_{21}(0)J_{23}+e^{2}_{12}(0)J_{41}\\\
+[e_{11}(0)e_{21}(0)-e_{22}(0)e_{12}(0)](J_{13}+J_{24)})\\\ \\\ =\par
J_{12}+J_{34}+[e^{2}_{11}(0)-e^{2}_{12}(0)]J_{14}+[e^{2}_{21}(0)-e^{2}_{22}(0)]J_{23}+[e_{11}(0)e_{21}(0)-e_{22}(0)e_{12}(0)](J_{13}+J_{24})=J_{12}+J_{34}.\par\par\par\par\par\end{array}$
This completes the proof.
Remark 1. Condition (4) holds if, for example,
$A=\begin{pmatrix}1&0&0&b\\\ 0&1&b&0\end{pmatrix}\quad\mbox{or}\quad
A=\begin{pmatrix}0&b&1&0\\\ b&0&0&1\end{pmatrix}.$ $None$
In both cases $J_{12}+J_{34}=1-b^{2}$, hence, if $b^{2}\neq 1$ the spectrum is
empty, and if $b^{2}=1$ the spectrum fills all complex plane. Notice, that if
$b=0$ conditions (14) are the Cauchy boundary conditions. The Cauchy problem
has no spectrum for any potential $V$.
Remark 2. Let us consider system (1) with boundary conditions
$\tilde{A}=\begin{pmatrix}1&0&0&1\\\ 0&1&1&0\end{pmatrix}.$ $None$
Denote by
$\tilde{E}(x,\lambda)=\begin{pmatrix}c_{1}(x,\lambda)&-s_{2}(x,\lambda)\\\
s_{1}(x,\lambda)&c_{2}(x,\lambda)\end{pmatrix}$
the matrix of the fundamental solution system to system (1) with boundary
condition $\tilde{E}(0,\lambda)=I$, where $I$ is the unit matrix. It is easy
to see that $J_{14}=J_{23}=J_{12}+J_{34}=0,J_{13}=1,J_{24}=-1$. Trivial
computation shows that the characteristic determinant of problem (1), (15) can
be reduced to the form
$\tilde{\Delta}(\lambda)=s_{1}(\pi,\lambda)-s_{2}(\pi,\lambda).$
Let condition (5) hold. From Theorem 1, we get that any complex $\lambda$ is
an eigenvalue of problem (1), (15), hence, $\tilde{\Delta}(\lambda)\equiv 0$,
therefore, $s_{1}(\pi,\lambda)\equiv s_{2}(\pi,\lambda)$.
## References
* [1]
* [2] A. P. Kosarev, A. A. Shkalikov, _Spectral asymptotics of solutions of a $2\times 2$ system of first-order ordinary differential equations,_ Math. Notes. 110 (2021) 967-971.
* [3] A. A. Lunyov and M. M. Malamud, _On Spectral Synthesis for Dissipative Dirac Type Operators,_ Integr. Equ. Oper. Theory 90 (2014), 79-106.
* [4] A. A. Lunyov, M. M. Malamud, _On the completeness and Riesz basis property of root subspaces of boundary value problems for first order systems and applications,_ J. Spectr. Theory. 5 (2015), 17-70.
* [5] A. A. Lunyov and M. M. Malamud, _On the Riesz basis property of the root vector system for Dirac-type $2\times 2$ systems,_ Dokl. Math. 90 (2014), 556-561.
* [6] A. A. Lunyov, M. M. Malamud, _On the Riesz basis property of root vectors system for $2\times 2$ Dirac type operators,_ J. Math. Anal. Appl. 441 (2016), 57-103.
* [7] A. A. Lunyov, M. M. Malamud, _On transformation operators and Riesz basis property of root vectors system for $n\times n$ Dirac type operators. Application to the Timoshenko beam model,_ arXiv:2112.07248.
* [8] A. A. Lunyov, M. M. Malamud, _On the trace formulas and completeness property of root vectors systems for $2\times 2$ Dirac type operators,_ arXiv:2312.15933.
* [9] A. S. Makin, _On the completeness of root function system of the Dirac operator with two-point boundary conditions_ , Math. Nachr. (2024), accepted.
* [10] A. S. Makin, _On the completeness of root function system of the $2\times 2$ Dirac operators with non-regular boundary conditions,_ arXiv:2401.02232.
email<EMAIL_ADDRESS>
|
Subsets and Splits
Filtered Text Samples
Retrieves 100 samples of text containing the specific phrase "You are a helpful assistant", providing limited insight into the dataset.
Helpful Assistant Text Samples
Returns a limited set of rows containing the phrase 'helpful assistant' in the text, providing basic filtering of relevant entries.