text
stringlengths
11
9.77k
label
stringlengths
2
104
For Huntington disease, identification of brain regions related to motor impairment can be useful for developing interventions to alleviate the motor symptom, the major symptom of the disease. However, the effects from the brain regions to motor impairment may vary for different groups of patients. Hence, our interest is not only to identify the brain regions but also to understand how their effects on motor impairment differ by patient groups. This can be cast as a model selection problem for a varying-coefficient regression. However, this is challenging when there is a pre-specified group structure among variables. We propose a novel variable selection method for a varying-coefficient regression with such structured variables. Our method is empirically shown to select relevant variables consistently. Also, our method screens irrelevant variables better than existing methods. Hence, our method leads to a model with higher sensitivity, lower false discovery rate and higher prediction accuracy than the existing methods. Finally, we found that the effects from the brain regions to motor impairment differ by disease severity of the patients. To the best of our knowledge, our study is the first to identify such interaction effects between the disease severity and brain regions, which indicates the need for customized intervention by disease severity.
statistics
We investigate the exact WKB method for the quantum Seiberg-Witten curve of 4d $N=2$ pure $SU(3)$ Yang-Mills, in the language of abelianization. The relevant differential equation is a third-order equation on $\mathbb{CP}^1$ with two irregular singularities. Exact WKB analysis leads us to consider new Darboux coordinates on a moduli space of flat SL(3,$\mathbb{C}$)-connections. In particular, in the weak coupling region we encounter coordinates of higher length-twist type generalizing Fenchel-Nielsen coordinates. The Darboux coordinates are conjectured to admit asymptotic expansions given by the formal quantum periods series; we perform numerical analysis supporting this conjecture.
high energy physics theory
The closed-form expression for the quantum partition function of the improved Tietz oscillator is obtained using the Voronoi summation formula.
physics
For spectrally negative L\'evy processes, adapting an approach from \cite{BoLi:sub1} we identify joint Laplace transforms involving local times evaluated at either the first passage times, or independent exponential times, or inverse local times. The Laplace transforms are expressed in terms of the associated scale functions. Connections are made with the permanental process and the Markovian loop soup measure.
mathematics
A set of vertices $W$ of a graph $G$ is a resolving set if every vertex of $G$ is uniquely determined by its vector of distances to $W$. In this paper, the Maker-Breaker resolving game is introduced. The game is played on a graph $G$ by Resolver and Spoiler who alternately select a vertex of $G$ not yet chosen. Resolver wins if at some point the vertices chosen by him form a resolving set of $G$, whereas Spoiler wins if the Resolver cannot form a resolving set of $G$. The outcome of the game is denoted by $o(G)$ and $R_{\rm MB}(G)$ (resp. $S_{\rm MB}(G)$) denotes the minimum number of moves of Resolver (resp. Spoiler) to win when Resolver has the first move. The corresponding invariants for the game when Spoiler has the first move are denoted by $R'_{\rm MB}(G)$ and $S'_{\rm MB}(G)$. Invariants $R_{\rm MB}(G)$, $R'_{\rm MB}(G)$, $S_{\rm MB}(G)$, and $S'_{\rm MB}(G)$ are compared among themselves and with the metric dimension ${\rm dim}(G)$. A large class of graphs $G$ is constructed for which $R_{\rm MB}(G) > {\rm dim}(G)$ holds. The effect of twin equivalence classes and pairing resolving sets on the Maker-Breaker resolving game is described. As an application $o(G)$, as well as $R_{\rm MB}(G)$ and $R'_{\rm MB}(G)$ (or $S_{\rm MB}(G)$ and $S'_{\rm MB}(G)$), are determined for several graph classes, including trees, complete multi-partite graphs, grid graphs, and torus grid graphs.
mathematics
In this paper, by performing a general Kaluza-Klein (KK) decomposition, we obtain a gauge invariant effective action for a bulk massless $q$-form field on a $p$-brane with codimension two. There appear four types of KK modes: two $(q-1)$-forms and one $(q-2)$-form in addtion to the ordinary $q$-form, which are essential for the gauge invariance. Due to the two extra dimensions, we find eight Schr\"{o}dinger-like equations for the four modes and their mass spectra are closely related. Moreover, via this decomposition mechanism, the Hodge duality in the bulk naturally induces four coupled dualities on the brane, which guarantees that the physical equavalence of bulk dual fields is preserved under the dimensional reduction.
high energy physics theory
Background: Neighbourhood social fragmentation and socioeconomic deprivation seem to be associated with suicide mortality. However, results are inconclusive, which might be because dynamics in the social context are not well-represented by administratively bounded neighbourhoods at baseline. We used individualized neighbourhoods to examine associations between suicide mortality, social fragmentation, and deprivation for the total population as well as by sex and age group. Methods: Using a nested case-control design, all suicides aged 18-64 years between 2007 and 2016 were selected from longitudinal Dutch register data and matched with 10 random controls. Indices for social fragmentation and deprivation were calculated annually for 300, 600, and 1,000 metre circular buffers around each subject's residential address. Results: Suicide mortality was significantly higher in neighbourhoods with high deprivation and social fragmentation. Accounting for individual characteristics largely attenuated these associations. Suicide mortality remained significantly higher for women living in highly fragmented neighbourhoods in the fully adjusted model. Age-stratified analyses indicate associations with neighbourhood fragmentation among women in older age groups (40-64 years) only. Among men, suicide risk was lower in fragmented neighbourhoods for 18-39-year-olds and for short-term residents. In deprived neighbourhoods, the suicide risk was lower for 40-64-year-old men and long-term residents. Associations between neighbourhood characteristics and suicide mortality were comparable across buffer sizes. Conclusion: Our findings suggest that next to individual characteristics, the social and economic context within which people live may both enhance and buffer the risk of suicide.
statistics
A photonic integrated circuit (PIC) comprised of an 11 cm multimode speckle waveguide, a 1x32 splitter, and a linear grating coupler array is fabricated and utilized to receive 2 GHz of radio-frequency (RF) signal bandwidth from 2.5 to 4.5 GHz using compressive sensing (CS). Incoming RF signals are modulated onto chirped optical pulses which are input to the multimode waveguide. The multimode waveguide produces the random projections needed for CS via optical speckle. The time-varying phase and amplitude of two test RF signals between 2.5 and 4.5 GHz are successfully recovered using the standard penalized $l_1$-norm method. The use of a passive PIC serves as an initial step towards the miniaturization of a compressive sensing RF receiver.
physics
The inelastic dark matter model is one kind of popular models for the light dark matter (DM) below $O(1)$ GeV. If the mass splitting between DM excited and ground states is small enough, the co-annihilation becomes the dominant channel for thermal relic density and the DM excited state can be long-lived at the collider scale. We study scalar and fermion inelastic dark matter models for $ {\cal O}(1) $ GeV DM at Belle II with $ U(1)_D $ dark gauge symmetry broken into its $Z_2$ subgroup. We focus on dilepton displaced vertex signatures from decays of the DM excited state. With the help of precise displaced vertex detection ability at Belle II, we can explore the DM spin, mass and mass splitting between DM excited and ground states. Especially, we show scalar and fermion DM candidates can be discriminated and the mass and mass splitting of DM sector can be determined within the percentage of deviation for some benchmark points. Furthermore, the allowed parameter space to explain the excess of muon $(g-2)_\mu$ is also studied and it can be covered in our displaced vertex analysis during the early stage of Belle II experiment.
high energy physics phenomenology
Informative features play a crucial role in the single image super-resolution task. Channel attention has been demonstrated to be effective for preserving information-rich features in each layer. However, channel attention treats each convolution layer as a separate process that misses the correlation among different layers. To address this problem, we propose a new holistic attention network (HAN), which consists of a layer attention module (LAM) and a channel-spatial attention module (CSAM), to model the holistic interdependencies among layers, channels, and positions. Specifically, the proposed LAM adaptively emphasizes hierarchical features by considering correlations among layers. Meanwhile, CSAM learns the confidence at all the positions of each channel to selectively capture more informative features. Extensive experiments demonstrate that the proposed HAN performs favorably against the state-of-the-art single image super-resolution approaches.
electrical engineering and systems science
High-resolution observations of the solar photosphere have recently revealed the presence of elongated filamentary bright structures inside sunspot umbrae. These features, which have been called umbral filaments (UFs), differ in morphology, evolution, and magnetic configuration from light bridges that are usually observed to intrude in sunspots. To study an UF observed in the leading sunspot of active region NOAA 12529, we have analyzed spectro-polarimetric observations taken in the photosphere with the spectropolarimeter (SP) aboard the Hinode satellite. High-resolution observations in the upper chromosphere and transition region taken with the IRIS telescope and observations acquired by SDO/HMI and SDO/AIA have been used to complement the spectro-polarimetric analysis. The results obtained from the inversion of the Hinode/SP measurements allow us to discard the hypothesis that UFs are a kind of light bridge. In fact, we find no field-free or low-field strength region cospatial to the observed UF. In contrast, we detect in the structure Stokes profiles that indicate the presence of strong horizontal fields, larger than 2500 G. Furthermore, a significant portion of the UF has opposite polarity with respect to the hosting umbra. In the upper atmospheric layers, we observe filaments being cospatial to the UF in the photosphere. We interpret these findings as suggesting that the UF could be the photospheric manifestation of a flux rope hanging above the sunspot, which triggers the formation of penumbral-like filaments within the umbra via magneto-convection.
astrophysics
The Reeh-Schlieder theorem, with the time-slice axiom of quantum field theory, may be used to recover the information which falls into a black hole. Analyticity of quantum fields in states with finite energy plays the crucial role. In AdS spacetime, our argument based on the Reeh-Schlieder theorem is consistent with the argument that there is no information loss because of the AdS/CFT correspondence.
high energy physics theory
With the increased use of social media platforms by people across the world, many new interesting NLP problems have come into existence. One such being the detection of sarcasm in the social media texts. We present a corpus of tweets for training custom word embeddings and a Hinglish dataset labelled for sarcasm detection. We propose a deep learning based approach to address the issue of sarcasm detection in Hindi-English code mixed tweets using bilingual word embeddings derived from FastText and Word2Vec approaches. We experimented with various deep learning models, including CNNs, LSTMs, Bi-directional LSTMs (with and without attention). We were able to outperform all state-of-the-art performances with our deep learning models, with attention based Bi-directional LSTMs giving the best performance exhibiting an accuracy of 78.49%.
computer science
We compute the helicity-dependent strange quark distribution in the proton in the framework of chiral effective theory. Starting from the most general chiral SU(3) Lagrangian that respects Lorentz and gauge invariance, we derive the complete set of hadronic splitting functions at the one meson loop level, including the octet and decuplet rainbow, tadpole, Kroll-Ruderman and octet-decuplet transition configurations. By matching hadronic and quark level operators, we obtain generalized convolution formulas for the quark distributions in the proton in terms of hadronic splitting functions and quark distributions in the hadronic configurations, and from these derive model-independent relations for the leading nonanalytic behavior of their moments. Within the limits of parameters of the Pauli-Villars regulators derived from inclusive hyperon production, we find that the polarized strange quark distribution is rather small and mostly negative.
high energy physics phenomenology
The project of \emph{"quantum spacetime phenomenology"} focuses on searching pragmatically for the Planck scale quantum features of spacetime. Among these features is the existence of a characteristic length scale addressed commonly by effective approaches to quantum gravity (QG). This characteristic length scale could be realized, for instance and simply, by generalizing the standard Heisenberg uncertainty principle (HUP) to a \emph{"generalized uncertainty principle"} (GUP). While usually it is expected that phenomena belonging to the realm of QG are essentially probable solely at the so-called Planck energy, here we show how a GUP proposal containing the most general modification of coordinate representation of the momentum operator could be probed by a \emph{"cold atomic ensemble recoil experiment"} (CARE) as a low energy quantum system. This proposed atomic interferometer setup has advantages over the conventional architectures owing to the enclosure in a high finesse optical cavity which is supported by a new class of low power consumption integrated devices known as \emph{"micro-electro-opto-mechanical systems"} (MEOMS). The proposed system comprises of a micro mechanical oscillator instead of spherical confocal mirrors as one of the components of high finesse optical cavity. In the framework of a bottom-up QG phenomenological viewpoint and by taking into account the measurement accuracy realized for the fine structure constant (FSC) from the Rubidium ($^{87}$Rb) CARE, we set some constraints as upper bounds on the characteristic parameters of the underlying GUP. In the case of superposition of the possible GUP modification terms, we managed to set a tight constraint as $0.999978<\lambda_0<1.00002$ for the dimensionless characteristic parameter.
high energy physics phenomenology
$\psi$-epistemic interpretations of quantum theory maintain that quantum states only represent incomplete information about the physical states of the world. A major motivation for this view is the promise to provide a reasonable account of state update under measurement by asserting that it is simply a natural feature of updating incomplete statistical information. Here we demonstrate that all known epistemic ontological models of quantum theory in dimension $d\geq3$, including those designed to evade the conclusion of the PBR theorem, cannot represent state update correctly. Conversely, interpretations for which the wavefunction is real evade such restrictions despite remaining subject to long-standing criticism regarding physical discontinuity, indeterminism and the ambiguity of the Heisenberg cut. This revives the possibility of a no-go theorem with no additional assumptions, and demonstrates that what is usually thought of as a strength of epistemic interpretations may in fact be a weakness.
quantum physics
Despite recent developments, there are a number of conceptual issues on the hadronic light-by-light (HLbL) contribution to the muon $(g-2)$ which remain unresolved. One of the most controversial ones is the precise way in which short-distance constraints get saturated by resonance exchange, particularly in the so-called Melnikov-Vainshtein (MV) limit. In this paper we address this and related issues from a novel perspective, employing a warped five-dimensional model as a tool to generate a consistent realization of QCD in the large-$N_c$ limit. This approach differs from previous ones in that we can work at the level of an effective action, which guarantees that unitarity is preserved and the chiral anomaly is consistently implemented at the hadronic level. We use the model to evaluate the inclusive contribution of Goldstone modes and axial-vector mesons to the HLbL. We find that both anomaly matching and the MV constraint cannot be fulfilled with a finite number of resonances (including the pion) and instead require an infinite number of axial-vector states. Our numbers for the HLbL point at a non-negligible role of axial-vector mesons, which is closely linked to a correct implementation of QCD short-distance constraints.
high energy physics phenomenology
At present, the majority of the proposed Deep Learning (DL) methods provide point predictions without quantifying the models uncertainty. However, a quantification of the reliability of automated image analysis is essential, in particular in medicine when physicians rely on the results for making critical treatment decisions. In this work, we provide an entire framework to diagnose ischemic stroke patients incorporating Bayesian uncertainty into the analysis procedure. We present a Bayesian Convolutional Neural Network (CNN) yielding a probability for a stroke lesion on 2D Magnetic Resonance (MR) images with corresponding uncertainty information about the reliability of the prediction. For patient-level diagnoses, different aggregation methods are proposed and evaluated, which combine the single image-level predictions. Those methods take advantage of the uncertainty in image predictions and report model uncertainty at the patient-level. In a cohort of 511 patients, our Bayesian CNN achieved an accuracy of 95.33% at the image-level representing a significant improvement of 2% over a non-Bayesian counterpart. The best patient aggregation method yielded 95.89% of accuracy. Integrating uncertainty information about image predictions in aggregation models resulted in higher uncertainty measures to false patient classifications, which enabled to filter critical patient diagnoses that are supposed to be closer examined by a medical doctor. We therefore recommend using Bayesian approaches not only for improved image-level prediction and uncertainty estimation but also for the detection of uncertain aggregations at the patient-level.
electrical engineering and systems science
Using the diagrammatic approach, here we study how spin-orbit coupling (SOC) affects the fermion-dimer and dimer-dimer scattering lengths in the Born approximation, and benchmark their accuracy with the higher-order approximations. We consider both isotropic and Rashba couplings in three dimensions, and show that the Born approximation gives accurate results in the $1/(m \alpha a_s) \ll -1$ limit, where $m$ is the mass of the fermions, $\alpha$ is the strength of the SOC, and $a_s$ is the $s$-wave scattering length between fermions. This is because the higher-loop contributions form a perturbative series in the $1/(m \alpha a_s) < 0$ region that is controlled by the smallness of the residue $Z$ of the dimer propagator. In sharp contrast, since $Z$ grows with the square-root of the binding energy of the dimer in the $1/(m \alpha a_s) > 0$ region, all of the higher-loop contributions are of similar order.
condensed matter
This paper studies massive access in cell-free massive multi-input multi-output (MIMO) based Internet of Things and solves the challenging active user detection (AUD) and channel estimation (CE) problems. For the uplink transmission, we propose an advanced frame structure design to reduce the access latency. Moreover, by considering the cooperation of all access points (APs), we investigate two processing paradigms at the receiver for massive access: cloud computing and edge computing. For cloud computing, all APs are connected to a centralized processing unit (CPU), and the signals received at all APs are centrally processed at the CPU. While for edge computing, the central processing is offloaded to part of APs equipped with distributed processing units, so that the AUD and CE can be performed in a distributed processing strategy. Furthermore, by leveraging the structured sparsity of the channel matrix, we develop a structured sparsity-based generalized approximated message passing (SS-GAMP) algorithm for reliable joint AUD and CE, where the quantization accuracy of the processed signals is taken into account. Based on the SS-GAMP algorithm, a successive interference cancellation-based AUD and CE scheme is further developed under two paradigms for reduced access latency. Simulation results validate the superiority of the proposed approach over the state-of-the-art baseline schemes. Besides, the results reveal that the edge computing can achieve the similar massive access performance as the cloud computing, and the edge computing is capable of alleviating the burden on CPU, having a faster access response, and supporting more flexible AP cooperation.
computer science
Unlike the situation with gain and phase margins in robust stabilization, the problem to determine an exact maximum delay margin is still an open problem, although extensive work has been done to establish upper and lower bounds. The problem is that the corresponding constraints in the Nyquist plot are frequency dependent, and encircling the point $s=-1$ has to be done at sufficiently low frequencies, as the possibility to do so closes at higher frequencies. In this paper we present a new method for determining a sharper lower bound by introducing a frequency-dependent shift. The problem of finding such a bound simultaneously with gain and phase margin constraints is also considered. In all these problems we take an analytic interpolation approach.
mathematics
We consider the problem of partitioning effectively a given symmetric (and irreflexive) rational relation R into two asymmetric rational relations. This problem is motivated by a recent method of embedding an R-independent language into one that is maximal R-independent, where the method requires to use an asymmetric partition of R. We solve the problem when R is realized by a zero-avoiding transducer (with some bound k): if the absolute value of the input-output length discrepancy of a computation exceeds k then the length discrepancy of the computation cannot become zero. This class of relations properly contains all recognizable, all left synchronous, and all right synchronous relations. We leave the asymmetric partition problem open when R is not realized by a zero-avoiding transducer. We also show examples of total wordorderings for which there is a relation R that cannot be partitioned into two asymmetric rational relations such that one of them is decreasing with respect to the given word-ordering.
computer science
The anti-triplet charmed baryon decays with the light scalar mesons are rarely measured, whereas the recent observation of the Cabibbo-favored $\Lambda_c^+\to \Lambda\eta\pi^+$ decay hints a possible $\Lambda_c^+\to\Lambda a_0(980)^+,a_0(980)^+\to \eta\pi^+$ process. We hence study the $\Lambda_c^+\to\Lambda a_0(980)^+$ decay. Particularly, it is found that the final state interaction can give a significant contribution, where $\Sigma^{+}(1385)$ and $\eta$ in $\Lambda_c^+\to \Sigma^{+}(1385)\eta$ by exchanging a charged pion are transformed as $\Lambda$ and $a_0(980)^+$, respectively. Accordingly, we predict ${\cal B}(\Lambda_c^+\to\Lambda a_0(980)^+)=(1.7^{+2.8}_{-1.0}\pm 0.3)\times 10^{-3}$, accessible to the BESIII, BELLEII and LHCb experiments.
high energy physics phenomenology
Generalized uncertainty principle puts forward the existence of the shortest distances and/or maximum momentum at the Planck scale for consideration. In this article, we investigate the solutions of a two-dimensional Duffin-Kemmer-Petiau (DKP) oscillator within an external magnetic field in a minimal length (ML) scale. First, we obtain the eigensolutions in ordinary quantum mechanics. Then, we examine the DKP oscillator in the presence of an ML for the spin-zero and spin-one sectors. We determine an energy eigenvalue equation in both cases with the corresponding eigenfunctions in the non-relativistic limit. We show that in the ordinary quantum mechanic limit, where the ML correction vanishes, the energy eigenvalue equations become identical with the habitual quantum mechanical ones. Finally, we employ the Euler-Mclaurin summation formula and obtain the thermodynamic functions of the DKP oscillator in the high-temperature scale.
physics
We assess the variance of the post-collapse evolution remnants of compact, massive, low-metallicity stars, under small changes in the degrees of rotation and magnetic field of selected pre-supernova cores. These stellar models are commonly considered progenitors of long gamma-ray bursts. The fate of the proto-neutron star (PNS) formed after collapse, whose mass may continuously grow due to accretion, critically depends on the poloidal magnetic field strength at bounce. Should the poloidal magnetic field be sufficiently weak, the PNS collapses to a black hole (BH) within a few seconds. Models on this evolutionary track contain promising collapsar engines. Poloidal magnetic fields smooth over large radial scales (e.g. dipolar fields) or slightly augmented with respect to the original pre-supernova core yield long-lasting PNSs. In these models, BH formation is avoided or staved off for a long time, hence, they may produce proto-magnetars (PMs). Some of our PM candidates have been run for $\lesssim 10\,$s after core bounce, but they have not entered the Kelvin-Helmholtz phase yet. Among these models, some display episodic events of spin-down during which we find properties broadly compatible with the theoretical expectations for PMs ($M_{PNS} \approx 1.85\,M_\odot - 2.5\,M_\odot$, $\bar{P}_{PNS} \approx 1.5 - 4\,$ms, and $b^{\rm surf}_{PNS} \lesssim 10^{15}\,$G) and their very collimated supernova ejecta has nearly reached the stellar surface with (still growing) explosion energies $\gtrsim 2\times 10^{51}\,$erg.
astrophysics
The fundamental nature of dark matter is entirely unknown. A compelling candidate is Twin Higgs mirror matter, invisible hidden-sector cousins of the Standard Model particles and forces. This generically predicts mirror neutron stars, degenerate objects made entirely of mirror nuclear matter. We find their structure using realistic equations of state, robustly modified based on first-principle quantum chromodynamic calculations. We predict their detectability with gravitational waves and binary pulsars, suggesting an impressive discovery potential and ability to probe the dark sector.
astrophysics
Many versions of cross-validation (CV) exist in the literature; and each version though has different variants. All are used interchangeably by many practitioners; yet, without explanation to the connection or difference among them. This article has three contributions. First, it starts by mathematical formalization of these different versions and variants that estimate the error rate and the Area Under the ROC Curve (AUC) of a classification rule, to show the connection and difference among them. Second, we prove some of their properties and prove that many variants are either redundant or "not smooth". Hence, we suggest to abandon all redundant versions and variants and only keep the leave-one-out, the $K$-fold, and the repeated $K$-fold. We show that the latter is the only among the three versions that is "smooth" and hence looks mathematically like estimating the mean performance of the classification rules. However, empirically, for the known phenomenon of "weak correlation", which we explain mathematically and experimentally, it estimates both conditional and mean performance almost with the same accuracy. Third, we conclude the article with suggesting two research points that may answer the remaining question of whether we can come up with a finalist among the three estimators: (1) a comparative study, that is much more comprehensive than those available in literature and conclude no overall winner, is needed to consider a wide range of distributions, datasets, and classifiers including complex ones obtained via the recent deep learning approach. (2) we sketch the path of deriving a rigorous method for estimating the variance of the only "smooth" version, repeated $K$-fold CV, rather than those ad-hoc methods available in the literature that ignore the covariance structure among the folds of CV.
statistics
The zero-temperature limit of the backgammon model under resetting is studied. The model is a balls-in-boxes model whose relaxation dynamics is governed by the density of boxes containing just one particle. As these boxes become rare at large times, the model presents an entropy barrier. As a preliminary step, a related model with faster relaxation, known to be mapped to a symmetric random walk, is studied by mapping recent results on diffusion with resetting onto the balls-in-boxes problem. Diffusion with an absorbing target at the origin (and diffusion constant equal to one), stochastically reset to the unit position, is a continuum approximation to the dynamics of the balls-in-boxes model, with resetting to a configuration maximising the number of boxes containing just one ball. In the limit of a large system, the relaxation time of the balls-in-boxes model under resetting is finite. The backgammon model subject to a constant resetting rate is then studied using an adiabatic approximation.
condensed matter
A critical bottleneck of massive multiple-input multiple-output (MIMO) system is the huge training overhead caused by downlink transmission, like channel estimation, downlink beamforming and covariance observation. In this paper, we propose to use the channel state information (CSI) of a small number of antennas to extrapolate the CSI of the other antennas and reduce the training overhead. Specifically, we design a deep neural network that we call an antenna domain extrapolation network (ADEN) that can exploit the correlation function among antennas. We then propose a deep learning (DL) based antenna selection network (ASN) that can select a limited antennas for optimizing the extrapolation, which is conventionally a type of combinatorial optimization and is difficult to solve. We trickly designed a constrained degradation algorithm to generate a differentiable approximation of the discrete antenna selection vector such that the back-propagation of the neural network can be guaranteed. Numerical results show that the proposed ADEN outperforms the traditional fully connected one, and the antenna selection scheme learned by ASN is much better than the trivially used uniform selection.
electrical engineering and systems science
Antiferromagnetism (AF) such as Neel ordering is often closely related to Coulomb interactions such as Hubbard repulsion in two-dimensional (2D) systems. Whether Neel AF ordering in 2D can be dominantly induced by electron-phonon couplings (EPC) has not been completely understood. Here, by employing numerically-exact sign-problem-free quantum Monte Carlo (QMC) simulations, we show that optical Su-Schrieffer-Heeger (SSH) phonons with frequency $\omega$ and EPC constant $\lambda$ can induce AF ordering for a wide range of phonon frequency $\omega>\omega_c$. For $\omega<\omega_c$, a valence-bond-solid (VBS) order appears and there is a direct quantum phase transition between VBS and AF phases at $\omega_c$. The phonon mechanism of the AF ordering is related to the fact that SSH phonons directly couple to electron hopping whose second-order process can induce an effective AF spin exchange. Our results shall shed new lights to understanding AF ordering in correlated quantum materials.
condensed matter
We present multi-color high-resolution imaging of the host galaxy of the dwarf Seyfert UGC 06728. As the lowest-mass black hole to be described with both a direct mass constraint and a spin constraint, UGC 06728 is an important source for comparison with black hole evolutionary models, yet little is known about the host galaxy. Using Hubble Space Telescope imaging in the optical and near-infrared, we find that UGC 06728 is a barred lenticular (SB0) galaxy with prominent ansae at the ends of the bar. We cleanly separated the AGN from the resolved galaxy with two-dimensional image decompositions, thus allowing accurate surface brightness profiles to be derived in all filters from the outer edge of the galaxy all the way into the nucleus. Based on a sample of 51 globular cluster candidates identified in the images, the globular cluster luminosity function predicts a distance to UCG 06728 of $32.5\pm3.5$ Mpc. Combining the galaxy photometry with the distance estimate, we derive a starlight-corrected AGN luminosity, the absolute magnitude of the galaxy, and a constraint on the galaxy stellar mass of $\log M_{\star}/M_{\odot}=9.9\pm0.2$.
astrophysics
The Gaussian graphical model (GGM) has become a popular tool for analyzing networks of psychological variables. In a recent paper in this journal, Forbes, Wright, Markon, and Krueger (FWMK) voiced the concern that GGMs that are estimated from partial correlations wrongfully remove the variance that is shared by its constituents. If true, this concern has grave consequences for the application of GGMs. Indeed, if partial correlations only capture the unique covariances, then the data that come from a unidimensional latent variable model ULVM should be associated with an empty network (no edges), as there are no unique covariances in a ULVM. We know that this cannot be true, which suggests that FWMK are missing something with their claim. We introduce a connection between the ULVM and the GGM and use that connection to prove that we find a fully-connected and not an empty network associated with a ULVM. We then use the relation between GGMs and linear regression to show that the partial correlation indeed does not remove the common variance.
statistics
We show that a rational function $f$ of degree $>1$ on the projective line over an algebraically closed field that is complete with respect to a non-trivial and non-archimedean absolute value has no potentially good reductions if and only if the Berkovich Julia set of $f$ is uniformly perfect. As an application, a uniform regularity of the boundary of each Berkovich Fatou component of $f$ is also established.
mathematics
Aiming at maximizing the achievable sum-rate of wideband multiuser mmWave massive MIMO systems, the hybrid precoding is studied. Since each computation of the achievable sum-rate can be performed only after the analog precoder and digital precoder are both determined, the maximization of the achievable sum-rate has intractable computational complexity. By introducing the interference free (IF) achievable sum-rate, the design of the analog and digital precoders can be decoupled. To avoid the beam conflict and maximize the IF achievable sum-rate, a Hungarian-based codeword selection algorithm is proposed for the analog precoding design. Simulation results verify the effectiveness of the proposed scheme and show that better performance can be achieved compared with existing schemes.
electrical engineering and systems science
After reviewing some of the fundamental aspects of Drinfel'd doubles and Poisson-Lie T-duality, we describe the three-dimensional isotropic rigid rotator on $SL(2,\mathbb{C})$ starting from a non-Abelian deformation of the natural carrier space of its Hamiltonian description on $T^*SU(2) \simeq SU(2) \ltimes \mathbb{R}^3$. A new model is then introduced on the dual group $SB(2,\mathbb{C})$, within the Drinfel'd double description of $SL(2,\mathbb{C})=SU(2) \bowtie SB(2,\mathbb{C})$. The two models are analyzed from the Poisson-Lie duality point of view, and a doubled generalized action is built with $TSL(2,\mathbb{C})$ as carrier space. The aim is to explore within a simple case the relations between Poisson-Lie symmetry, Doubled Geometry and Generalized Geometry. In fact, all the mentioned structures are discussed, such as a Poisson realization of the $C$-brackets for the generalized bundle $T \oplus T^*$ over $SU(2)$ from the Poisson algebra of the generalized model. The two dual models exhibit many features of Poisson-Lie duals and from the generalized action both of them can be respectively recovered by gauging one of its symmetries.
high energy physics theory
In this paper, we present a novel statistical model, $\textit{the generalized-Gaussian-Rician}$ (GG-Rician) distribution, for the characterization of synthetic aperture radar (SAR) images. Since accurate statistical models lead to better results in applications such as target tracking, classification, or despeckling, characterizing SAR images of various scenes including urban, sea surface, or agricultural, is essential. The proposed statistical model is based on the Rician distribution to model the amplitude of a complex SAR signal, the in-phase and quadrature components of which are assumed to be generalized-Gaussian distributed. The proposed amplitude GG-Rician model is further extended to cover the intensity SAR signals. In the experimental analysis, the GG-Rician model is investigated for amplitude and intensity SAR images of various frequency bands and scenes in comparison to state-of-the-art statistical models that include $\mathcal{K}$, Weibull, Gamma, and Lognormal. In order to decide on the most suitable model, statistical significance analysis via Kullback-Leibler divergence and Kolmogorov-Smirnov statistics are performed. The results demonstrate the superior performance and flexibility of the proposed model for all frequency bands and scenes and its applicability on both amplitude and intensity SAR images. The Matlab package is available at https://github.com/oktaykarakus/GG-Rician-SAR-Image-Modelling.
electrical engineering and systems science
We investigate the phenomenological aspects of non-perturbative baryon- and lepton-number-violating processes at hadron colliders. Such processes, induced by instanton/sphaleron configurations of the electroweak gauge fields, are believed to play a crucial role in the generation of baryon asymmetry in the early Universe at finite temperature. On the other hand, at colliders (that represent the zero-temperature high-energy regime) the rate and observability of such processes are still under debate. Motivated by current theoretical considerations, we construct a modern event generator within the general-purpose Herwig Monte Carlo framework, that aims to capture the most relevant features of the dominant processes. We perform a detailed phenomenological analysis focussing on the Large Hadron Collider, at 13 TeV proton-proton centre-of-mass energy, a potential high-energy upgrade at 27 TeV and the proposed Future Circular Collider (FCC-hh) at 100 TeV. We derive constraints on the expected rates for various parametrisations of our model. We find that all three colliders are capable of providing meaningful information on the nature of instanton/sphaleron-induced processes at various energy scales.
high energy physics phenomenology
We define a set of fully Lorentz-invariant wave packets and show that it spans the corresponding one-particle Hilbert subspace, with a manifestly Lorentz-invariant completeness relation (resolution of identity). The position-momentum uncertainty relation for this Lorentz-invariant wave packet deviates from the ordinary Heisenberg uncertainty principle, and reduces to it in the non-relativistic limit.
high energy physics theory
In this paper, we consider a hierarchical control based DC microgrid (DCmG) equipped with unknown input observer (UIO) based detectors, and investigate false data injection (FDI) attacks against the secondary control layer. First, we point out that the fundamental limitation of the UIO-based detector comes from the unknown inputs, whose dynamics can be utilized to construct zero trace undetectable (ZTU) attacks causing zero impact on detection residuals. Furthermore, we extend ZTU attacks to nonzero trace undetectable (NTU) attacks by utilizing the system noise, under which the detection residuals are still bounded by certain detection thresholds. Then, by approximating primary control loops as unit gains, we theoretically analyze attack impacts of NTU attacks on voltage balancing and current sharing in the DCmG. Moreover, we propose a countermeasure against ZTU and NTU attacks by observing the average PCC voltage deviation. Finally, extensive simulations are conducted in Simulink/PLECS to validate theoretical results and effectiveness of the countermeasure.
electrical engineering and systems science
Nonnegative Matrix Factorization (NMF), first proposed in 1994 for data analysis, has received successively much attention in a great variety of contexts such as data mining, text clustering, computer vision, bioinformatics, etc. In this paper the case of a symmetric matrix is considered and the symmetric nonnegative matrix factorization (SymNMF) is obtained by using a penalized nonsymmetric minimization problem. Instead of letting the penalizing parameter increase according to an a priori fixed rule, as suggested in literature, we propose a heuristic approach based on an adaptive technique. Extensive experimentation shows that the proposed algorithm is effective.
mathematics
The Milky Way's metal-poor stars are nearby ancient objects that are used to study early chemical evolution and the assembly and structure of the Milky Way. Here we present reliable metallicities of $\sim280,000$ stars with $-3.75 \lesssim$ [Fe/H] $\lesssim -0.75$ down to $g=17$ derived using metallicity-sensitive photometry from the second data release (DR2) of the SkyMapper Southern Survey. We use the dependency of the flux through the SkyMapper $v$ filter on the strength of the Ca II K absorption features, in tandem with SkyMapper $u,g,i$ photometry, to derive photometric metallicities for these stars. We find that metallicities derived in this way compare well to metallicities derived in large-scale spectroscopic surveys, and use such comparisons to calibrate and quantify systematics as a function of location, reddening, and color. We find good agreement with metallicities from the APOGEE, LAMOST, and GALAH surveys, based on a standard deviation of $\sigma\sim0.25$dex of the residuals of our photometric metallicities with respect to metallicities from those surveys. We also compare our derived photometric metallicities to metallicities presented in a number of high-resolution spectroscopic studies to validate the low metallicity end ([Fe/H] $< -2.5$) of our photometric metallicity determinations. In such comparisons, we find the metallicities of stars with photometric [Fe/H] $< -2.5$ in our catalog show no significant offset and a scatter of $\sigma\sim$0.31dex level relative to those in high-resolution work when considering the cooler stars ($g-i > 0.65$) in our sample. We also present an expanded catalog containing photometric metallicities of $\sim720,000$ stars as a data table for further exploration of the metal-poor Milky Way.
astrophysics
We present a modelling framework for multi-target tracking based on possibility theory and illustrate its ability to account for the general lack of knowledge that the target-tracking practitioner must deal with when working with real data. We also introduce and study variants of the notions of point process and intensity function, which lead to the derivation of an analogue of the probability hypothesis density (PHD) filter. The gains provided by the considered modelling framework in terms of flexibility lead to the loss of some of the abilities that the PHD filter possesses; in particular the estimation of the number of targets by integration of the intensity function. Yet, the proposed recursion displays a number of advantages such as facilitating the introduction of observation-driven birth schemes and the modelling the absence of information on the initial number of targets in the scene. The performance of the proposed approach is demonstrated on simulated data.
statistics
The modular bootstrap program for 2d CFTs could be seen as a systematic exploration of the physical consequences of consistency conditions at the elliptic points and at the cusp of their toruspartition function. The study at $\tau=i$, the elliptic point stabilized by the modular inversion $S$, was initiated by Hellerman, who found a general upper bound for the most relevant scaling dimension $\Delta$. Likewise, analyticity at $\tau=i\infty$, the cusp stabilized by the modular translation $T$, yields an upper bound on the twist gap. Here we study consistency conditions at $\tau=\exp[2i\pi/3]$, the elliptic point stabilized by $S T$. We find a much stronger upper bound in the large-c limit, namely $\Delta<\frac{c-1}{12}+0.092$, which is very close to the minimal mass threshold of the BTZ black holes in the gravity dual of $AdS_3/CFT_2$ correspondence.
high energy physics theory
Calogero-Sutherland models of $N$ identical particles on a circle are deformed away from hermiticity but retaining a $\cal PT$ symmetry. The interaction potential gets completely regularized, which adds to the energy spectrum an infinite tower of previously non-normalizable states. For integral values of the coupling, extra degeneracy occurs and a nonlinear conserved supersymmetry charge enlarges the ring of Liouville charges. The integrability structure is maintained. We discuss the $A_{N-1}$-type models in general and work out details for the cases of $A_2$ and $G_2$.
high energy physics theory
Transition discs are expected to be a natural outcome of the interplay between photoevaporation (PE) and giant planet formation. Massive planets reduce the inflow of material from the outer to the inner disc, therefore triggering an earlier onset of disc dispersal due to PE through a process known as Planet-Induced PhotoEvaporation (PIPE). In this case, a cavity is formed as material inside the planetary orbit is removed by PE, leaving only the outer disc to drive the migration of the giant planet. We investigate the impact of PE on giant planet migration and focus specifically on the case of transition discs with an evacuated cavity inside the planet location. This is important for determining under what circumstances PE is efficient at halting the migration of giant planets, thus affecting the final orbital distribution of a population of planets. For this purpose, we use 2D FARGO simulations to model the migration of giant planets in a range of primordial and transition discs subject to PE. The results are then compared to the standard prescriptions used to calculate the migration tracks of planets in 1D planet population synthesis models. The FARGO simulations show that once the disc inside the planet location is depleted of gas, planet migration ceases. This contradicts the results obtained by the impulse approximation, which predicts the accelerated inward migration of planets in discs that have been cleared inside the planetary orbit. These results suggest that the impulse approximation may not be suitable for planets embedded in transition discs. A better approximation that could be used in 1D models would involve halting planet migration once the material inside the planetary orbit is depleted of gas and the surface density at the 3:2 mean motion resonance location in the outer disc reaches a threshold value of $0.01\,\mathrm{g\,cm^{-2}}$.
astrophysics
We elaborate on the recent proposal that intersection numbers of certain cohomology classes on the moduli space of genus-zero Riemann surfaces with punctures compute tree-level scattering amplitudes in quantum field theories. The relevant cohomology classes are twisted by representations of the fundamental group that describes how punctures braid around each other on the Riemann surface, which can be used to link it to the space of kinematic invariants. Intersection numbers of said cohomology classes, whose representatives we call twisted forms, can be shown to fully localize on the boundaries of the moduli space, which are in one-to-one map with Feynman diagrams. We prove that when twisted forms are logarithmic, their intersection numbers have a simple expansion in terms of trivalent Feynman diagrams allowing only for massless propagators on the internal and external lines. For physical applications one also needs to study non-logarithmic forms as they are responsible for propagation of massive states. We utilize the natural fibre bundle structure of the moduli space, which allows for a direct access to the boundaries, to introduce recursion relations for intersection numbers that "integrate out" puncture-by-puncture. The resulting recursion involves only linear algebra of certain matrices describing braiding properties of the moduli space and evaluating one-dimensional residues, thus paving a way for explicit analytic computations of scattering amplitudes. Together with the previous reformulation of the tree-level S-matrix of string theory in terms of twisted forms, the results of this work complete a unified geometric framework for studying scattering amplitudes from genus-zero Riemann surfaces. We show that a web of dualities between different homology and cohomology groups allows for deriving a host of identities among various types of amplitudes computed from the moduli space.
high energy physics theory
Speckle noise suppression is a challenging and crucial pre-processing stage for higher-level image analysis. In this work, a new attempt has been made using telegraph total variation equation and fuzzy set theory for speckle noise suppression. The intuitionistic fuzzy divergence (IFD) function has been used to distinguish between edges and noise. To the best of the author's knowledge, most of the studies on multiplicative speckle noise removal process focus on only diffusion-based filters, and little attention has been paid to the study of fuzzy set theory. The proposed approach enjoy the benefits of both telegraph total variation equation and fuzzy edge detector, which is not only robust to noise but also preserves image structural details. Moreover, we establish the existence and uniqueness of a weak solution of the regularized version of the proposed model using Schauder fixed point theorem. With the proposed model, despeckling is carried out on natural and Synthetic Aperture Radar (SAR) images. The experimental results of the proposed model are reported, which found better in terms of noise suppression and detail/edge preservation, with respect to the existing approaches.
electrical engineering and systems science
We construct a numerical solution of the recently-derived large-$N_c \& N_f$ small-$x$ helicity evolution equations with the aim to establish the small-$x$ asymptotics of the quark helicity distribution beyond the large-$N_c$ limit explored previously in the same framework. (Here $N_c$ and $N_f$ are the numbers of quark colors and flavors.) While the large-$N_c$ helicity evolution involves gluons only, the large-$N_c \& N_f$ evolution includes contributions from quarks as well. We find that adding quarks to the evolution makes quark helicity distribution oscillate as a function of $x$. Our numerical results in the large-$N_c \& N_f$ limit lead to the $x$-dependence of the flavor-singlet quark helicity distribution which is well-approximated by \begin{align} \Delta \Sigma (x, Q^2)\bigg|_{\mbox{large-}N_c \& N_f} \sim \left( \frac{1}{x} \right)^{\alpha_h^q} \, \cos \left[ \omega_q \, \ln \left( \frac{1}{x} \right) + \varphi_q \right]. \end{align} The power $\alpha_h^q$ exhibits a weak $N_f$-dependence, and, for all $N_f$ values considered, remains very close to $\alpha_h^q (N_f=0) = (4/\sqrt{3}) \sqrt{\alpha_s N_c/(2 \pi)}$ obtained earlier in the large-$N_c$ limit. The novel oscillation frequency $\omega_q$ and phase shift $\varphi_q$ depend more strongly on the number of flavors $N_f$ (with $\omega_q =0$ in the pure-glue large-$N_c$ limit). The typical period of oscillations for $\Delta \Sigma$ is rather long, spanning many units of rapidity. We speculate whether the oscillations we find are related to the sign variation with $x$ seen in the strange quark helicity distribution extracted from the data.
high energy physics phenomenology
We study the reduction of classical strings rotating in the deformed three-sphere truncation of the double Yang-Baxter deformation of the $\hbox{AdS}_3 \times \hbox{S}^3 \times \hbox{T}^4$ background to an integrable mechanical model. The use of the generalized spinning-string ansatz leads to an integrable deformation of the Neumann-Rosochatius system. Integrability of this system follows from the fact that the usual constraints for the Uhlenbeck constants apply to any deformation that respects the isometric coordinates of the three-sphere. We construct solutions to the system in terms of the underlying ellipsoidal coordinate. The solutions depend on the domain of the deformation parameters and the reality conditions of the roots of a fourth order polynomial. We obtain constant-radii, giant-magnon and trigonometric solutions when the roots degenerate, and analyze the possible solutions in the undeformed limit. In the case where the deformation parameters are purely imaginary and the polynomial involves two complex-conjugated roots, we find a new class of solutions. The new class is connected with twofold giant-magnon solutions in the degenerate limit of infinite period.
high energy physics theory
Slim accretion disks idea emerged over 30 years ago as an answer to several unsolved problems. Since that time there was a tremendous increase in the amount of observational data where this model applies. However, many critical issues on the theoretical side remain unsolved, as they are inherently difficult. This is the issue of the disk stability under the radiation pressure, the role of the magnetic field in the energy transfer inside the disk and the formation (or not) of a warm corona, and outflows. Thus the progress has to be done both through further developments of the model and through the careful comparison to the observational data.
astrophysics
Current burning issues in stellar physics, for both hot and cool stars, concern their magnetism. In hot stars, stable magnetic fields of fossil origin impact their stellar structure and circumstellar environment, with a likely major role in stellar evolution. However, this role is complex and thus poorly understood as of today. It needs to be quantified with high-resolution UV spectropolarimetric measurements. In cool stars, UV spectropolarimetry would provide access to the structure and magnetic field of the very dynamic upper stellar atmosphere, providing key data for new progress to be made on the role of magnetic fields in heating the upper atmospheres, launching stellar winds, and more generally in the interaction of cool stars with their environment (circumstellar disk, planets) along their whole evolution. UV spectropolarimetry is proposed on missions of various sizes and scopes, from POLLUX on the 15-m telescope LUVOIR to the Arago M-size mission dedicated to UV spectropolarimetry.
astrophysics
This paper is concerned with matching feature vectors in a one-to-one fashion across large collections of datasets. Formulating this task as a multidimensional assignment problem with decomposable costs (MDADC), we develop extremely fast algorithms with time complexity linear in the number $n$ of datasets and space complexity a small fraction of the data size. These remarkable properties hinge on using the squared Euclidean distance as dissimilarity function, which can reduce ${n \choose 2}$ matching problems between pairs of datasets to $n$ problems and enable calculating assignment costs on the fly. To our knowledge, no other method applicable to the MDADC possesses these linear scaling and low-storage properties necessary to large-scale applications. In numerical experiments, the novel algorithms outperform competing methods and show excellent computational and optimization performances. An application of feature matching to a large neuroimaging database is presented. The algorithms of this paper are implemented in the R package matchFeat available at https://github.com/ddegras/matchFeat.
statistics
A large number of binary black holes (BBHs) with longer orbital periods are supposed to exist as progenitors of BBH mergers recently discovered with gravitational wave (GW) detectors. In our previous papers, we proposed to search for such BBHs in triple systems through the radial-velocity modulation of the tertiary orbiting star. If the tertiary is a pulsar, high precision and cadence observations of its arrival time enable an unambiguous characterization of the pulsar -- BBH triples located at several kpc, which are inaccessible with the radial velocity of stars. The present paper shows that such inner BBHs can be identified through the short-term R{\o}mer delay modulation, on the order of $10$ msec for our fiducial case, a triple consisting of $20~M_\odot$ BBH and $1.4~M_\odot$ pulsar with $P_\mathrm{in}=10$ days and $P_\mathrm{out}=100$ days. If the relativistic time delays are measured as well, one can determine basically all the orbital parameters of the triple. For instance, this method is applicable to inner BBHs of down to $\sim 1$ hr orbital periods if the orbital period of the tertiary pulsar is around several days. Inner BBHs with $\lesssim 1$ hr orbital period emit the GW detectable by future space-based GW missions including LISA, DECIGO, and BBO, and very short inner BBHs with sub-second orbital period can be even probed by the existing ground-based GW detectors. Therefore, our proposed methodology provides a complementary technique to search for inner BBHs in triples, if exist at all, in the near future.
astrophysics
A framework for relativistic thermodynamics and statistical physics is built by first exploiting the symmetries between energy and momentum in the derivation of the Boltzmann distribution, then using Einstein's energy-momentum relationship to derive a PDE for the partition function. It is shown that the extended Boltzmann distribution implies the existence of an inverse four-temperature, while the form of the partition function PDE implies the existence of a quantizable field theory of classical statistics, with hints of an associated gravity like gauge theory. An adaptation of the framework is then used to derive a thermodynamic certainty relationship.
condensed matter
Bayesian modelling enables us to accommodate complex forms of data and make a comprehensive inference, but the effect of partial misspecification of the model is a concern. One approach in this setting is to modularize the model, and prevent feedback from suspect modules, using a cut model. After observing data, this leads to the cut distribution which normally does not have a closed-form. Previous studies have proposed algorithms to sample from this distribution, but these algorithms have unclear theoretical convergence properties. To address this, we propose a new algorithm called the Stochastic Approximation Cut algorithm (SACut) as an alternative. The algorithm is divided into two parallel chains. The main chain targets an approximation to the cut distribution; the auxiliary chain is used to form an adaptive proposal distribution for the main chain. We prove convergence of the samples drawn by the proposed algorithm and present the exact limit. Although SACut is biased, since the main chain does not target the exact cut distribution, we prove this bias can be reduced geometrically by increasing a user-chosen tuning parameter. In addition, parallel computing can be easily adopted for SACut, which greatly reduces computation time.
statistics
In this paper we study AdS-Schwarzschild black holes in four and five dimensions in dRGT minimally coupled to a cloud of strings and investigate the effects of this string cloud on the thermodynamics of the black holes. It is observed that the entropy of the string cloud and massive terms does not affect the black hole entropy. The observations about four dimensions indicate that the massive term in the presence of external string cloud can not exhibit Van der Waals like behavior for AdS-Schwarzschild black holes and therefore there is only the Hawking-Page phase transition. In contrast, in five dimensions, the graviton mass modifies this behavior through the third massive term, so that a critical behavior and second order phase transition is deduced. Also, the Joule-Thomson effect is not observed. The black hole stability conditions are also studied in four and five dimensions and a critical value for the string cloud parameter is presented. In five dimensions a degeneracy between states for extremal black holes is investigated. After studying black holes as thermodynamic systems, we consider such systems as heat engines, and finally the efficiency of them is calculated.
high energy physics theory
A method for density-based topology optimization of heat exchangers with two fluids is proposed. The goal of the optimization process is to maximize the heat transfer from one fluid to the other, under maximum pressure drop constraints for each of the fluid flows. A single design variable is used to describe the physical fields. The solid interface and the fluid domains are generated using an erosion-dilation based identification technique, which guarantees well-separated fluids, as well as a minimum wall thickness between them. Under the assumption of laminar steady flow, the two fluids are modelled separately, but in the entire computational domain using the Brinkman penalization technique for ensuring negligible velocities outside of the respective fluid subdomains. The heat transfer is modelled using the convection-diffusion equation, where the convection is driven by both fluid flows. A stabilized finite element discretization is used to solve the governing equations. Results are presented for two different problems: a two-dimensional example illustrating and verifying the methodology; and a three-dimensional example inspired by shell-and-tube heat exchangers. The optimized designs for both cases show an improved heat transfer compared to the baseline designs. For the shell-and-tube case, the full freedom topology optimization approach is shown to yield performance improvements of up to 113% under the same pressure drop.
physics
We show that products of the isotopic substitution reactions in experimentally accessible molecules such as NaK, RbCs, and SrF are cold according to their translational energy below hundreds of mK. For these chemical reactions, molecular products may occupy only the lowest rotational states. We also discuss the possibility of controlling the chemical reactions by the electric field in ultracold mixtures of molecules and atoms with low kinetic energy release, where one of the constituent atoms of colliding molecule is replaced by its isotope. This letter opens new avenues in investigating the branching ratios of chemical reactions in ultracold conditions.
physics
Nanographenes with zigzag edges are predicted to manifest non-trivial pi-magnetism resulting from the interplay of hybridization of localized frontier states and Coulomb repulsion between valence electrons. This provides a chemically tunable platform to explore quantum magnetism at the nanoscale and opens avenues toward organic spintronics. The magnetic stability in nanographenes is thus far limited by the weak magnetic exchange coupling which remains below the room temperature thermal energy. Here, we report the synthesis of large rhombus-shaped nanographenes with zigzag periphery on gold and copper surfaces. Single-molecule scanning probe measurements unveil an emergent magnetic spin-singlet ground state with increasing nanographene size. The magnetic exchange coupling in the largest nanographene, determined by inelastic electron tunneling spectroscopy, exceeds 100 meV or 1160 K, which outclasses most inorganic nanomaterials and remarkably survives on a metal electrode.
condensed matter
When constructing a model to estimate the causal effect of a treatment, it is necessary to control for other factors which may have confounding effects. Because the ignorability assumption is not testable, however, it is usually unclear which set of controls is appropriate, and effect estimation is generally sensitive to this choice. A common approach in this case is to fit several models, each with a different set of controls, but it is difficult to reconcile inference under the multiple resulting posterior distributions for the treatment effect. Therefore we propose a two-stage approach to measure the sensitivity of effect estimation with respect to control specification. In the first stage, a model is fit with all available controls using a prior carefully selected to adjust for confounding. In the second stage, posterior distributions are calculated for the treatment effect under nested sets of controls by propagating posterior uncertainty in the original model. We demonstrate how our approach can be used to detect the most significant confounders in a dataset, and apply it in a sensitivity analysis of an observational study measuring the effect of legalized abortion on crime rates.
statistics
We study the phenomenology of the minimal $(2,2)$ inverse-seesaw model supplemented with Abelian flavour symmetries. To ensure maximal predictability, we establish the most restrictive flavour patterns which can be realised by those symmetries. This setup requires adding an extra scalar doublet and two complex scalar singlets to the Standard Model, paving the way to implement spontaneous CP violation. It is shown that such CP-violating effects can be successfully communicated to the lepton sector through couplings of the scalar singlets to the new sterile fermions. The Majorana and Dirac CP phases turn out to be related, and the active-sterile neutrino mixing is determined by the active neutrino masses, mixing angles and CP phases. We investigate the constraints imposed on the model by the current experimental limits on lepton flavour-violating decays, especially those on the branching ratio $BR(\mu\rightarrow e \gamma)$ and the capture rate $CR(\mu-e,{\rm Au})$. The prospects to further test the framework put forward in this work are also discussed in view of the projected sensitivities of future experimental searches sensitive to the presence of heavy sterile neutrinos. Namely, we investigate at which extent upcoming searches for $\mu\rightarrow e \gamma$, $\mu \rightarrow 3e$ and $\mu-e$ conversion in nuclei will be able to test our model, and how complementary will future high-energy collider and beam-dump experiments be in that task.
high energy physics phenomenology
We study the Coulomb drag between two strange-metal layers using the Einstein-Maxwell-Dilaton model from holography. We show that the low-temperature dependence of the drag resistivity is $\rho_D \propto T^4$, which strongly deviates from the quadratic dependence of Fermi liquids. We also present numerical results at room temperature, using typical parameters of the cuprates, to provide an estimate of the magnitude of this effect for future experiments. We find that the drag resistivity is enhanced by the plasmons characteristic of the two-layer system.
condensed matter
The structure theorem is established which shows that an arbitrary multi-mode bosonic Gaussian observable can be represented as a combination of four basic cases, the physical prototypes of which are homodyne and heterodyne, noiseless or noisy, measurements in quantum optics. The proof establishes connection between the description of Gaussian observable in terms of the characteristic function and in terms of density of the probability operator-valued measure (POVM) and has remarkable parallels with treatment of bosonic Gaussian channels in terms of their Choi-Jamiolkowski form. Along the way we give the ``most economical'', in the sense of minimal dimensions of the quantum ancilla, construction of the Naimark extension of a general Gaussian observable. It is also shown that the Gaussian POVM has bounded operator-valued density with respect to the Lebesgue measure if and only if its noise covariance matrix is nondegenerate.
quantum physics
We reconsider the problem of the critical behavior of a three-dimensional $O(m)$ symmetric magnetic system in the presence of random anisotropy disorder with a generic trimodal random axis distribution. By introducing $n$ replicas to average over disorder it can be coarse-grained to a $\phi^{4}$-theory with $m \times n$ component order parameter and five coupling constants taken in the limit of $n \to 0$. Using a field theory approach we renormalize the model to two-loop order and calculate the $\beta$-functions within the $\varepsilon$ expansion and directly in three dimensions. We analyze the corresponding renormalization group flows with the help of the Pad\'e-Borel resummation technique. We show that there is no stable fixed point accessible from physical initial conditions whose existence was argued in the previous studies. This may indicate an absence of a long-range ordered phase in the presence of random anisotropy disorder with a generic random axis distribution.
condensed matter
In this paper, we propose a scheme to generate entanglement between two distant qubits (two-level atom) which are separately trapped in their own (in general) non-Markovian dissipative cavities by utilizing entangling swapping. We consider the case in which the qubits can move along their cavity axes rather than a static state of motion. We first examine the role of movement of the qubit by studying the entropy evolution for each subsystem. We calculate the average entropy over the initial states of the qubit. Then by performing a Bell state measurement on the fields leaving the cavities, we swap the entanglement between qubit-field in each cavity into qubit-qubit and field-field subsystems. We use the entangling power to measure the average amount of swapped entanglement over all possible pure initial states. Our results are presented in two weak and strong coupling regimes. Our results illustrate the positive role of the movement of the qubits on the swapped entanglement. It is revealed that by considering certain conditions for the initial state of qubits, it is possible to achieve a maximally long-leaving stationary entanglement (Bell state) which is entirely independent of the environmental variables as well as the velocity of qubits. This happens when the two qubits have the same velocities.
quantum physics
Sorted $L_1$ penalization estimator (SLOPE) is a regularization technique for sorted absolute coefficients in high-dimensional regression. By arbitrarily setting its regularization weights $\lambda$ under the monotonicity constraint, SLOPE can have various feature selection and clustering properties. On weight tuning, the selected features and their clusters are very sensitive to the tuning parameters. Moreover, the exhaustive tracking of their changes is difficult using grid search methods. This study presents a solution path algorithm that provides the complete and exact path of solutions for SLOPE in fine-tuning regularization weights. A simple optimality condition for SLOPE is derived and used to specify the next splitting point of the solution path. This study also proposes a new design of a regularization sequence $\lambda$ for feature clustering, which is called the quasi-spherical and octagonal shrinkage and clustering algorithm for regression (QS-OSCAR). QS-OSCAR is designed with a contour surface of the regularization terms most similar to a sphere. Among several regularization sequence designs, sparsity and clustering performance are compared through simulation studies. The numerical observations show that QS-OSCAR performs feature clustering more efficiently than other designs.
statistics
Jiangmen Underground Neutrino Observatory (JUNO) is designed to determine the neutrino mass hierarchy using a 20 kton liquid scintillator detector. To calibrate detector boundary effect, the Guide Tube Calibration System (GTCS) has been designed to deploy a radioactive source along a given longitude on the outer surface of the detector. In this paper, we studied the physics case of this system via simulation, which leads to a mechanical design.
physics
There has been a growing interest in realizing quantum simulators for physical systems where perturbative methods are ineffective. The scalability and flexibility of circuit quantum electrodynamics (cQED) make it a promising platform to implement various types of simulators, including lattice models of strongly-coupled field theories. Here, we use a multimode superconducting parametric cavity as a hardware-efficient analog quantum simulator, realizing a lattice in synthetic dimensions with complex hopping interactions. The coupling graph, \textit{i.e.} the realized model, can be programmed \textit{in situ}. The complex-valued hopping interaction further allows us to simulate, for instance, gauge potentials and topological models. As a demonstration, we simulate a plaquette of the bosonic Creutz ladder. We characterize the lattice with scattering measurements, reconstructing the experimental Hamiltonian and observing emerging topological features. This platform can be easily extended to larger lattices and different models involving other interactions.
quantum physics
We discuss a recent test of the performance of the kinetic energy partition method (KEP) through its application to two separable quantum-mechanical models. We argue that one of the benchmark models is exceedingly simple for testing any realistic approximate method and that almost any reasonable approach yields better results. In the second example our exact benchmark eigenvalues disagree considerably with those chosen by the authors for comparison, which casts doubts on the accuracy of their KEP approach.
quantum physics
Recently Bauer $\textit{et al.}$ arXiv:1909.13022 proposed ANUBIS, an auxiliary detector to be installed in one of the shafts above the ATLAS or CMS interaction point, as a tool to search for long-lived particles. Here, we study the sensitivity of this proposal for long-lived heavy neutral leptons (HNLs) in both minimal and extended scenarios. We start with the minimal HNL model where both production and decay of the HNLs are mediated by active-sterile neutrino mixing, before studying the case of right-handed neutrinos in a left-right symmetric model. We then consider a $U(1)_{B-L}$ extension of the SM. In this model HNLs are produced from the decays of the mostly SM-like Higgs boson, via mixing in the scalar sector of the theory. In all cases, we find that ANUBIS has sensitivity reach comparable to the proposed MATHUSLA detector. For the minimal HNL scenario, the contributions from $W$'s decaying to HNLs are more important at ANUBIS than at MATHUSLA, extending the sensitivity to slightly larger HNL masses at ANUBIS.
high energy physics phenomenology
Large-scale magnetic field is believed to play a key role in launching and collimating jets/outflows. It was found that advection of external field by a geometrically thin disk is rather inefficient, while the external weak field may be dragged inwards by fast radially moving tenuous or/and hot gas above the thin disk. We investigate the field advection in a thin (cold) accretion disk covered with hot corona, in which turbulence is responsible for the angular momentum transfer of the gas in the disk and corona. The radial velocity of the gas in the corona is significantly higher than that in the thin disk. Our calculations show that the external magnetic flux is efficiently transported inwards by the corona, and the field line is strongly inclined towards the disk surface, which help launching outflows. The field configurations are consistent with those observed in the numerical simulations. The strength of the field is substantially enhanced in the inner region of the disk (usually several orders of magnitude higher than the external field strength), which is able to drive a fraction of gas in the corona into outflows. This mechanism may be useful in explaining the observational features in X-ray binaries and active galactic nuclei. Our results may help understanding the physics of the magneto-hydrodynamic (MHD) simulations.
astrophysics
Non-variable OH/IR stars are thought to have just left the asymptotic giant branch (AGB) phase. In this conventional picture, they must still show strong circumstellar extinction caused by the dust ejected during the AGB phase, and the extinction is expected to decrease over time because of the dispersal of the circumstellar dust after the cessation of the stellar mass loss. The reduction of the extinction makes the stars become apparently brighter and bluer with time especially in the near-infrared (NIR) range. We look for such long-term brightening of non-variable OH/IR stars by using 2MASS, UKIDSS, and OAOWFC survey data. As a result, we get multi-epoch NIR data taken over a 20-year period (1997-2017) for 6 of 16 non-variable OH/IR stars, and all six objects are found to be brightening. The K-band brightening rate of five objects ranges from 0.010 to 0.130 mag yr$^{-1}$, which is reasonably explained with the conventional picture. However, one OH/IR star, OH31.0-0.2, shows a rapid brightening, which cannot be explained only by the dispersal of the dust shell. Multi-color (J-, H-, and K-band) data are obtained for three objects, OH25.1-0.3, OH53.6-0.2, and OH77.9+0.2. Surprisingly, none of them appears to have become bluer, and OH53.6-0.2 is found to have been reddened with a rate of 0.013 mag yr$^{-1}$ in (J-K). Our findings suggest other mechanisms such as rapid changes in stellar properties (temperature or luminosity) or a generation of a new batch of dust grains.
astrophysics
The dynamics of coalescing compact binaries can be affected by the environment in which the systems evolve, leaving detectable signatures into the emitted gravitational signal. In this paper we investigate the ability of gravitational-wave detectors to constrain the nature of the environment in which compact binaries merge. We parametrize a variety of environmental effects by modifying the phase of the gravitational signal emitted by black hole and neutron star binaries. We infer the bounds on such effects by current and future generation of interferometers, studying their dependence on the binary's parameters. We show that the strong dephasing induced by accretion and dynamical friction can constraint the density of the surrounding medium to orders of magnitude below that of accretion disks. Planned detectors, such as LISA or DECIGO, will be able to probe densities typical of those of dark matter.
astrophysics
Although the 125 GeV Higgs boson discovered at the LHC is often heralded as the origin of mass, it may not in fact be the origin of Yukawa couplings. In alternative models, Yukawa couplings may instead arise from a seesaw type mechanism involving the mixing of Standard Model (SM) chiral fermions with new vector-like fermions, controlled by the vacuum expectation value (VEV) of a new complex Higgs singlet field $\langle \Phi \rangle$. For example, the largest third family $(t,b)$ quark Yukawa couplings may be forbidden by a $U(1)'$ gauge or global symmetry, broken by $\langle \Phi \rangle$, and generated effectively via mixing with a vector-like fourth family quark doublet $(T,B)$. Such theories predict a new physical Higgs singlet $\phi$, which we refer to as the Yukon, resulting from $\langle \Phi \rangle$, in the same way that the Higgs boson $h^0$ results from $\langle H\rangle$. In a simplified model we discuss the prospects for discovering the Yukon $\phi$ in gluon-gluon fusion production, with $(t,b)$ and $(T,B)$ quarks in the loops, and decaying in the channels $\phi\rightarrow \gamma\gamma, Z\gamma$ and $\phi\rightarrow tT\rightarrow tth^0,ttZ$. The potential for discovery of the Yukon $\phi$ is studied at present or future hadron colliders such as the LHC (Run 3), HL-LHC, HE-LHC and/or FCC. For example, we find that a 300-350 GeV Yukon $\phi$ could be accessed at LHC Run 3 in the di-photon channel in the global model, providing a smoking gun signature of the origin of Yukawa couplings.
high energy physics phenomenology
The main aim of this present paper is to present a new extension of the fractional derivative operator by using the extension of Beta function recently defined by Shadab et al.[19]. Moreover, we establish some results related to the newly defined modified fractional derivative operator such as Mellin transform and relations to extended hypergeometric and Appell's function via generating functions.
mathematics
We obtain Poisson equations satisfied by elliptic modular graph functions with four links. Analysis of these equations leads to a non--trivial algebraic relation between the various graphs.
high energy physics theory
This paper is devoted to study the Cauchy problem for the fractional dissipative BO equations $u_t+\mathcal{H}u_{xx}-(D_x^{\alpha}-D_x^{\beta})u+uu_x=0$, $0< \alpha < \beta$. When $1<\beta <2$, we prove GWP in $H^s(\mathbb{R})$, $s>-\beta/4$. For $\beta\geq 2$, we show GWP in $H^s(\mathbb{R})$, $s>\max\{3/2-\beta , \, -\beta/2\}$. We establish that our results are sharp in the sense that the flow map $u_0\mapsto u$ fails to be $C^2$ in $H^s(\mathbb{R})$, for $s<-\beta/2$, and it fails to be $C^3$ in $H^s(\mathbb{R})$ when $s<\min\{3/2-\beta , \, -\beta/4\}$. When $0< \beta<1$, we show ill-posedness in $H^s(\mathbb{R})$, $s\in \mathbb{R}$. Finally, if $\beta >3/2$, we prove GWP in $H^s(\mathbb{T})$, $s>\max\{3/2-\beta , \, -\beta/2\}$, and we deduce lack of $C^2$ regularity in $H^s(\mathbb{T})$ when $s<-\beta/2$, in particular we get sharp results when $\beta \geq 3$.
mathematics
Fine particulate matter (PM2.5) is a mixture of air pollutants that has adverse effects on human health. Understanding the health effects of PM2.5 mixture and its individual species has been a research priority over the past two decades. However, the limited availability of speciated PM2.5 measurements continues to be a major challenge in exposure assessment for conducting large-scale population-based epidemiology studies. The PM2.5 species have complex spatial-temporal and cross dependence structures that should be accounted for in estimating the spatiotemporal distribution of each component. Two major sources of air quality data are commonly used for deriving exposure estimates: point-level monitoring data and gridded numerical computer model simulation, such as the Community Multiscale Air Quality (CMAQ) model. We propose a statistical method to combine these two data sources for estimating speciated PM2.5 concentration. Our method models the complex relationships between monitoring measurements and the numerical model output at different spatial resolutions, and we model the spatial dependence and cross dependence among PM2.5 species. We apply the method to combine CMAQ model output with major PM2.5 species measurements in the contiguous United States in 2011.
statistics
The $\mathcal{N}{=}\,4$ supersymmetric $\mathrm{U}(2)$-spin hyperbolic Calogero-Sutherland model with odd matrix fields is examined. Explicit form of the $\mathcal{N}{=}\,4$ supersymmetry generators is derived. The Lax representation for the dynamics of the $\mathcal{N}{=}\,4$ hyperbolic $\mathrm{U}(2)$-spin Calogero-Sutherland system is found. The reduction to the $\mathcal{N}{=}\,4$ supersymmetric spinless hyperbolic Calogero-Sutherland system is established.
high energy physics theory
Galaxy clusters observed through the thermal Sunyaev-Zeldovich (tSZ) effect are a recent cosmological probe. The precision on the cosmological constraints is affected mainly by the current knowledge of cluster physics, which enters the analysis through the scaling relations. Here we aim to study one of the most important sources of systematic uncertainties, the mass bias, $b$. We have analysed the effects of a mass-redshift dependence, adopting a power-law parametrisation. We applied this parametrisation to the combination of tSZ number counts and power spectrum, finding a hint of redshift dependence that leads to a decreasing value of the mass bias for higher redshift. We tested the robustness of our results for different mass bias calibrations and a discrete redshift dependence. We find our results to be dependent on the clusters sample that we are considering, in particular obtaining an inverse (decreasing) redshift dependence when neglecting $z<0.2$ clusters. We analysed the effects of this parametrisation on the combination of cosmic microwave background (CMB) primary anisotropies and tSZ galaxy clusters. We find a preferred constant value of mass bias, having $(1-b) =0.62 \pm 0.05$. The corresponding value of $b$ is too high with respect to weak lensing and numerical simulations estimations. Therefore we conclude that this mass-redshift parametrisation does not help in solving the remaining discrepancy between CMB and tSZ clusters observations.
astrophysics
In this paper, we have attempted to study the behaviour of the family wise error rate (FWER) for Bonferroni's procedure and false discovery rate (FDR) of the Benjamini-Hodgeberg procedure for simultaneous testing problem with equicorrelated normal observations. By simulation study, we have shown that F.W.E.R. is a concave function for small no. of hypotheses and asymptotically becomes a convex function of the correlation. The plots of F.W.E.R. and F.D.R. confirms that if non-negative correlation is present, then these procedures control the type-I error rate at a much smaller rate than the desired level of significance. This confirms the conservative nature of these popular methods when correlation is present and provides a scope for improvement in power by appropriate adjustment for correlation.
statistics
For systems consisting of distinguishable particles, there exists an agreed upon notion of entanglement which is fundamentally based on the possibility of addressing individually each one of the constituent parties. Instead, the indistinguishability of identical particles hinders their individual addressability and has prompted diverse, sometimes discordant definitions of entanglement. In the present review, we provide a comparative analysis of the relevant existing approaches, which is based on the characterization of bipartite entanglement in terms of the behaviour of correlation functions. Such a a point of view provides a fairly general setting where to discuss the presence of non-local effects; it is performed in the light of the following general consistency criteria: i) entanglement corresponds to non-local correlations and cannot be generated by local operations; ii) when, by "freezing" suitable degrees of freedom, identical particles can be effectively distinguished, their entanglement must reduce to the one that holds for distinguishable particles; iii) in absence of other quantum resources, only entanglement can outperform classical information protocols. These three requests provide a setting that allows to evaluate strengths and weaknesses of the existing approaches to indistinguishable particle entanglement and to contribute to the current understanding of such a crucial issue. Indeed, they can be classified into five different classes: four hinging on the notion of particle and one based on that of physical modes. We show that only the latter approach is consistent with all three criteria, each of the others indeed violating at least one of them.
quantum physics
An overview of pseudo-Dirac neutrino framework is given starting from general spinor phenomenology. The framework is then tested by simulation of oscillations for T2K experiment parameters. Two possible derivations of oscillation parameters are indicated to have the same result.
high energy physics phenomenology
We examine the stability of mixed-symmetry superconducting states with broken time-reversal symmetry in spatial-symmetry-broken systems, including chiral states, on the basis of the free-energy functional derived in the weak-coupling theory. We consider a generic a_1 + i a_2 wave state, with a_1 and a_2 being different symmetry indices such as (a_1,a_2) = (d,s), (p_x,p_y), and (d,d').The time-reversal symmetry of the mixed-symmetry state with the a_1- and a_2-wave components is broken when the phases of these components differ, and such a state is called the time-reversal-symmetry breaking (TRSB) state. However, their phases are equated by Cooper-pair scattering between these components if it occurs; i.e., when the off-diagonal elements S_{a_1 a_2} = S_{a_2 a_1} of the scattering matrix are nonzero, they destabilize the TRSB state. Hence, it has often been believed that the TRSB state is stable only in systems with a spatial symmetry that guarantees S_{a_1 a_2}=0. We note that, contrary to this belief, the TRSB state can remain stable in systems without the spatial symmetry when the relative phase shifts so that S_{a_1 a_2} = 0 is restored, which results in a distorted TRSB (a_1 + a_2) + i a_2 wave state. Here, note that the restoration of S_{a_1 a_2} = 0 does not imply that the symmetry of the quasi-particle energy E_k is recovered. This study shows that such stabilization of the TRSB state occurs when the distortion is sufficiently small and \Delta_{a_1} \Delta_{a_2} is sufficiently large, where \Delta_a is the amplitude of the a-wave component in the TRSB state in the absence of the distortion. We clarify the manner in which the shift in the relative phase eliminates S_{a_1 a_2} and prove that such a state yields a free-energy minimum. We also propose a formula for the upper bound of the degree of lattice distortion, below which the TRSB state can be stable.
condensed matter
Airlines use different boarding policies to organize the queue of passengers waiting to enter the airplane. We analyze three policies in the many-passenger limit by a geometric representation of the queue position and row designation of each passenger and apply a Lorentzian metric to calculate the total boarding time. The boarding time is governed by the time each passenger needs to clear the aisle, and the added time is determined by the aisle-clearing time distribution through an effective aisle-clearing time parameter. The non-organized queues under the common random boarding policy are characterized by large effective aisle-clearing time. We show that, subject to a mathematical assumption which we have verified by extensive numerical computations in all realistic cases, the average total boarding time is always reduced when slow passengers are separated from faster passengers and the slow group is allowed to enter the airplane first. This is a universal result that holds for any combination of the three main governing parameters: the ratio between effective aisle-clearing times of the fast and the slow group, the fraction of slow passenger, and the congestion of passengers in the aisle. Separation into groups based on aisle-clearing time allows for more synchronized seating, but the result is non-trivial, as the similar fast-first policy -- where the two groups enter the airplane in reverse order -- is inferior to random boarding for a range of parameter settings. The asymptotic results conform well with discrete-event simulations with realistic number of passengers, and both the slow-first and the fast-first policies have the ability to perform unboundedly better than random boarding. Parameters based on empirical data, with hand luggage as criteria for separating passengers into the slow group, give a 13\% reduction in total boarding time for slow first compared to random boarding.
physics
The shape of the local bubble is modeled in the framework of the thin layer approximation. The asymmetric shape of the local bubble is simulated by introducing axial profiles for the density of the interstellar medium, such as exponential, Gaussian, inverse square dependence and Navarro--Frenk--White. The availability of some observed asymmetric profiles for the local bubble allows us to match theory and observations via the observational percentage of reliability. The model is compatible with the presence of radioisotopes on Earth.
astrophysics
In this article we exploit Ruelle-type spectral functions and analyze the Verma module over Virasoro algebra, boson-fermion correspondence, the analytic torsion, the Chern-Simons and $\eta$ invariants, as well as the generation function associated to dimensions of the Hochschild homology of the crossed product $\mathbb{C}[S_n]\ltimes \mathcal{A}^{\otimes n}$ ($\mathcal{A}$ is the $q$-Weyl algebra). After analysing the Chern-Simons and $\eta$ invariants of Dirac operators by using irreducible $SU(n)$-flat connections on locally symmetric manifolds of non-positive section curvature, we describe the exponential action for the Chern-Simons theory.
high energy physics theory
We present new exact solutions of the Landau-Lifshitz and higher-order Landau-Lifshitz equations describing particle motion, with radiation reaction, in intense electromagnetic fields. Through these solutions and others we compare the phenomenological predictions of different equations in the context of the conjectured `radiation-free direction' (RFD). We confirm analytically in several cases that particle orbits predicted by the Landau-Lifshitz equation indeed approach the RFD at extreme intensities, and give time-resolved signals of this behaviour in radiation spectra.
high energy physics phenomenology
Future active distribution grids (ADGs) will incorporate a plethora of Distributed Generators (DGs) and other Distributed Energy Resources (DERs), allowing them to provide ancillary services in grid-connected mode and, if necessary, operate in an islanded mode to increase reliability and resilience. In this paper, we investigate the ability of an ADG to provide frequency control (FC) in grid-connected mode and ensure reliable islanded operation for a pre-specified time period. First, we formulate the operation of the grid participating in European-type FC markets as a centralized multi-period optimal power flow problem with a rolling horizon of 24 hours. Then, we include constraints to the grid-connected operational problem to guarantee the ability to switch to islanded operation at every time instant. Finally, we explore the technical and economic feasibility of offering these services on a balanced low-voltage distribution network. The results show that the proposed scheme is able to offer and respond to different FC products, while ensuring that there is adequate energy capacity at every time step to satisfy critical load in the islanded mode.
mathematics
The aim of presented research is to design a nanodevice based on a gate-defined quantum dot within a MoS$_2$ monolayer in which we confine a single electron. By applying control voltages to the device gates we modulate the confinement potential and force intervalley transitions. The present Rashba spin-orbit coupling additionally allows for spin operations. Moreover, both effects enable the spin-valley SWAP. The device structure is modeled realistically, taking into account feasible dot-forming potential and electric field that controls the Rasha coupling. Therefore, by performing reliable numerical simulations, we show how by electrically controlling the state of the electron in the device, we can obtain single- and two-qubit (thus universal) gates in a spin-valley two-qubit system. Through simulations we investigate possibility of implementation of two qubits \textit{locally}, based on single electron, with an intriguing feature that two-qubit gates are easier to realize than single ones.
condensed matter
In high-dimensional statistics, variable selection is an optimization problem aiming to recover the latent sparse pattern from all possible covariate combinations. In this paper, we propose a novel optimization method to solve the exact $L_0$-regularized regression problem (a.k.a. best subset selection). We reformulate the optimization problem from a discrete space to a continuous one via probabilistic reparameterization. Within the framework of stochastic gradient descent, we propose a family of unbiased gradient estimators to optimize the $L_0$-regularized objective and a variational lower bound. Within this family, we identify the estimator with a non-vanishing signal-to-noise ratio and uniformly minimum variance. Theoretically, we study the general conditions under which the method is guaranteed to converge to the ground truth in expectation. In a wide variety of synthetic and semi-synthetic data sets, the proposed method outperforms existing variable selection methods that are based on penalized regression and mixed-integer optimization, in both sparse pattern recovery and out-of-sample prediction. Our method can find the true regression model from thousands of covariates in a couple of seconds. a
statistics
We study production of scalar dark matter via the freeze--in mechanism in the relativistic regime, focussing on the simplest Higgs portal model. We derive the corresponding relativistic reaction rates based on the Bose--Einstein statistics taking into account the thermal mass effects as well as the change in the Higgs degrees of freedom at the electroweak phase transition. The consequent constraints on the Higgs portal coupling are obtained.
high energy physics phenomenology
Online social media platforms are turning into the prime source of news and narratives about worldwide events. However,a systematic summarization-based narrative extraction that can facilitate communicating the main underlying events is lacking. To address this issue, we propose a novel event-based narrative summary extraction framework. Our proposed framework is designed as a probabilistic topic model, with categorical time distribution, followed by extractive text summarization. Our topic model identifies topics' recurrence over time with a varying time resolution. This framework not only captures the topic distributions from the data, but also approximates the user activity fluctuations over time. Furthermore, we define significance-dispersity trade-off (SDT) as a comparison measure to identify the topic with the highest lifetime attractiveness in a timestamped corpus. We evaluate our model on a large corpus of Twitter data, including more than one million tweets in the domain of the disinformation campaigns conducted against the White Helmets of Syria. Our results indicate that the proposed framework is effective in identifying topical trends, as well as extracting narrative summaries from text corpus with timestamped data.
computer science
We report numerical simulations of the nonlinear dynamics of Josephson vortices driven by strong dc currents in layered superconductors. Dynamic equations for interlayer phase differences in a stack of coupled superconducting layers were solved to calculate a drag coefficient $\eta(J)$ of the vortex as a function of the perpendicular dc current density $J$. It is shown that Cherenkov radiation produced by a moving vortex causes significant radiation drag increasing $\eta(v)$ at high vortex velocities $v$ and striking instabilities of driven Josephson vortices moving faster than a terminal velocity $v_c$. The steady-state flux flow breaks down at $v>v_c$ as the vortex starts producing a cascade of expanding vortex-antivortex pairs evolving into either planar macrovortex structures or branching flux patterns propagating both along and across the layers. The pair production triggered by a rapidly moving vortex is most pronounced in a stack of underdamped planar junctions where it can occur at $J>J_s$ well below the interlayer Josephson critical current density. Both $v_c$ and $J_s$ were calculated as functions of the quasiparticle damping parameter, and the dc magnetic field applied parallel to the layers. The effects of vortex interaction on the Cherenkov instability of moving vortex chains and lattices in annular stacks of Josephson junctions were considered. It is shown that a vortex driven by a current density $J>J_s$ in a multilayer of finite length excites self-sustained large-amplitude standing waves of magnetic flux, resulting in temporal oscillations of the total magnetic moment. We evaluated a contribution of this effect to the power $W$ radiated by the sample and showed that $W$ increases strongly as the number of layers increases.
condensed matter
Cylinders in projective varieties play an important role in connection with unipotent group actions on certain affine algebraic varieties. The previous work due to Dubouloz and Kishimoto deals with the condition for a del Pezzo fibration to contain a vertical cylinder. In the present work, as a generalization in the sense of singularities, we shall determine the condition under which a del Pezzo fibration with canonical singularities admits a vertical cylinder by means of degree and type of singularities found on the corresponding generic fiber.
mathematics
We prove the centrality of $\mathrm{K}_2 (\mathsf{F}_4, \,R)$ for an arbitrary commutative ring $R$. This completes the proof of the centrality of $\mathrm K_2(\Phi,\, R)$ for any root system $\Phi$ of rank $\geq 3$. Our proof uses only elementary localization techniques reformulated in terms of pro-groups. Another new result of the paper is the construction of a crossed module on the canonical homomorphism $\mathrm{St}(\Phi, R) \to \mathrm{G}_\mathrm{sc}(\Phi, R)$, which has not been known previouly for exceptional $\Phi$.
mathematics
Let $1\le p<\infty$. A symmetric space $X$ on $[0,1]$ is said to be $p$-disjointly homogeneous (resp. restricted $p$-disjointly homogeneous) if every sequence of normalized pairwise disjoint functions from $X$ (resp. characteristic functions) contains a subsequence equivalent in $X$ to the unit vector basis of $l_p$. Answering a question posed recently, we construct, for each $1\le p<\infty$, a restricted $p$-disjointly homogeneous symmetric space, which is not $p$-disjointly homogeneous. Moreover, we prove that the property of $p$-disjoint homogeneity is preserved under Banach isomorphisms.
mathematics
Birational rowmotion is a discrete dynamical system on the set of all positive real-valued functions on a finite poset, which is a birational lift of combinatorial rowmotion on order ideals. It is known that combinatorial rowmotion for a minuscule poset has order equal to the Coxeter number, and exhibits the file homomesy phenomenon for refined order ideal cardinality statistic. In this paper we generalize these results to the birational setting. Moreover, as a generalization of birational promotion on a product of two chains, we introduce birational Coxeter-motion on minuscule posets, and prove that it enjoys periodicity and file homomesy.
mathematics
This paper models stress softening during cyclic loading and unloading of an elastomer. The paper begins by remodelling the primary loading curve to include a softening function and goes on to derive non-linear transversely isotropic constitutive equations for the elastic response, stress relaxation, residual strain and creep of residual strain. These ideas are combined with a transversely isotropic version of the Arruda-Boyce eight-chain model to develop a constitutive relation that is capable of accurately representing the Mullins effect during cyclic stress-softening for a transversely isotropic, hyperelastic material, in particular a carbon-filled rubber vulcanizate. Keywords: Mullins effect, stress-softening, hysteresis, stress relaxation, residual strain, creep of residual strain, transverse isotropy. MSC codes: 74B20, 74D10, 74L15
condensed matter
In an effective field theory approach to gravity, the Einstein-Hilbert action is supplemented by higher derivative terms. In the absence of matter, four derivative terms can be eliminated by a field redefinition. We use the Euclidean action to calculate analytically the corrections to thermodynamic quantities of the Kerr solution arising from terms with six or eight derivatives. The eight derivative terms make a non-negative correction to the entropy (at fixed mass and angular momentum) if their coefficients have appropriate signs. The correction from the six derivative terms does not have a definite sign.
high energy physics theory