text
stringlengths 11
9.77k
| label
stringlengths 2
104
|
---|---|
Understanding the nature of the excitation spectrum in quantum spin liquids is of fundamental importance, in particular for the experimental detection of candidate materials. However, current theoretical and numerical techniques have limited capabilities, especially in obtaining the dynamical structure factor, which gives a crucial characterization of the ultimate nature of the quantum state and may be directly assessed by inelastic neutron scattering. In this work, we investigate the low-energy properties of the $S=1/2$ Heisenberg model on the triangular lattice, including both nearest-neighbor $J_1$ and next-nearest-neighbor $J_2$ super-exchanges, by a dynamical variational Monte Carlo approach that allows accurate results on spin models. For $J_2=0$, our calculations are compatible with the existence of a well-defined magnon in the whole Brillouin zone, with gapless excitations at $K$ points (i.e., at the corners of the Brillouin zone). The strong renormalization of the magnon branch (also including roton-like minima around the $M$ points, i.e., midpoints of the border zone) is described by our Gutzwiller-projected state, where Abrikosov fermions are subject to a non-trivial magnetic $\pi$-flux threading half of the triangular plaquettes. When increasing the frustrating ratio $J_2/J_1$, we detect a progessive softening of the magnon branch at $M$, which eventually becomes gapless within the spin-liquid phase. This feature is captured by the band structure of the unprojected wave function (with $2$ Dirac points for each spin component). In addition, we observe an intense signal at low energies around the $K$ points, which cannot be understood within the unprojected picture and emerges only when the Gutzwiller projection is considered, suggesting the relevance of gauge fields for the low-energy physics of spin liquids. | condensed matter |
The planar Landau system which describes the quantum mechanical motion of a charged particle in a plane with a uniform magnetic field perpendicular to the plane, is explored within pedagogical settings aimed at the beginning graduate level. The system is known to possess the Euclidean symmetry in two dimensions with central extension $\bar{E}(2)$. In this paper, we revisit the well-known energy eigenvalues of the system, known as the Landau levels, by exploiting the related $\bar{e}(2)$ symmetry algebra. Specifically, we utilize the Casimir operator and the commutation relations of the generators of the $\bar{E}(2)$ group. More importantly, an algebraic formalism on this topic based on Schwinger's oscillator model of angular momentum is also presented. The dimensions of irreducible representations of the $\bar{E}(2)$ group and their implications on the degeneracy of Landau levels is discussed. | quantum physics |
A novel domain-decomposition least-squares Petrov-Galerkin (DD-LSPG) model-reduction method applicable to parameterized systems of nonlinear algebraic equations (e.g., arising from discretizing a parameterized partial-differential-equations problem) is proposed. In contrast with previous works, we adopt an algebraically non-overlapping decomposition strategy rather than a spatial-decomposition strategy, which facilitates application to different spatial-discretization schemes. Rather than constructing a low-dimensional subspace for the entire state space in a monolithic fashion, the methodology constructs separate subspaces for the different subdomains/components characterizing the original model. In the offline stage, the method constructs low-dimensional bases for the interior and interface of components. In the online stage, the approach constructs an LSPG ROM for each component and enforces strong or weak compatibility on the 'ports' connecting them. We propose four different ways to construct reduced bases on the interface/ports of subdomains and several ways to enforce compatibility across connecting ports. We derive a posteriori and a priori error bounds for the DD-LSPG solutions. Numerical results performed on nonlinear benchmark problems in heat transfer and fluid dynamics demonstrate that the proposed method performs well in terms of both accuracy and computational cost, with different choices of basis and compatibility constraints yielding different performance profiles. | mathematics |
Ultrathin sheets of transition metal dichalcogenides (MX$ _2$) with charge density waves (CDWs) is increasingly gaining interest as a promising candidate for graphene-like devices. Although experimental data including stripe/quasi-stripe structure and hidden states have been reported, the ground state of ultrathin MX$ _2$ compounds and, in particular, the origin of anisotropic (stripe and quasi-stripe) CDW phases is a long-standing problem. Anisotropic CDW phases have been explained by Coulomb interaction between domain walls and inter-layer interaction. However, these models assume that anisotropic domain walls can exist in the first place. Here, we report that anisotropic CDW domain walls can appear naturally without assuming anisotropic interactions: We explain the origin of these phases by topological defect theory (line defects in a two-dimensional plane) and interference between harmonics of macroscopic CDW wave functions. We revisit the McMillan-Nakanishi-Shiba model for monolayer 1$T$-TaS$ _2$ and 2$H$-TaSe$ _2$ and show that CDWs with wave vectors that are separated by $120^\circ$ (i.e. the three-fold rotation symmetry of the underlying lattice) contain a free-energy landscape with many local minima. Then, we remove this $120^\circ$ constraint and show that free energy local minima corresponding to the stripe and quasi-stripe phase appear. Our results imply that Coulomb interaction between domain walls and inter-layer interaction may be secondary factors for the appearance of these phases. Furthermore, this model can predict new CDW phases, hence it may become the basis to study CDW further. We anticipate our results to be a starting point for further study in two-dimensional physics, such as explanation of "Hidden CDW states", study the interplay between supersolid symmetry and lattice symmetry, and application to other van der Waals structures. | condensed matter |
In this work we analyze F-theory and Type IIB orientifold compactifications to study $\alpha '$-corrections to the four-dimensional, $\mathcal{N} = 1$ effective actions. In particular, we obtain corrections to the K\"ahlermoduli space metric and its complex structure for generic dimension originating from eight-derivative corrections to eleven-dimensional supergravity. We propose a completion of the $G^ 2 R^3$ and $(\nabla G)^2R^2$-sector in eleven-dimensions relevant in Calabi--Yau fourfold reductions. We suggest that the three-dimensional, $\mathcal{N}=2$ K\"ahler coordinates may be expressed as topological integrals depending on the first, second, and third Chern-forms of the divisors of the internal Calabi--Yau fourfold. The divisor integral Ansatz for the K\"ahler potential and K\"ahler coordinates may be lifted to four-dimensional, $\mathcal{N} = 1$ F-theory vacua. We identify a novel correction to the K\"ahler potential and coordinates at order $ \alpha'^2$, which is leading compared to other known corrections in the literature. At weak string coupling the correction arises from the intersection of $D7$-branes and $O7$-planes with base divisors and the volume of self-intersection curves of divisors in the base. In the presence of the conjectured novel $\alpha'$-correction resulting from the divisor interpretation the no-scale structure may be broken. Furthermore, we propose a model independent scenario to achieve non-supersymmetric AdS vacua for Calabi-Yau orientifold backgrounds with negative Euler-characteristic. | high energy physics theory |
In this stu dy, two dimensional ( 2D correlation analysis wa s utilized for achieving a significant improvement in the signal to noise (S/N) ratio of laser induced breakdown spectroscopy (LIBS) data . Time resolved LIBS spectra of metallic, bimetal lic targets and the normal LIBS spectra of bimetallic targets with varying compositions were used for the detailed analysis . The diagonal of the matrix in the synchronous spectra was used to demonstrate the improvement in the S/N ratio. An improvemen t in the peak intensities by few order s of magnitude accompanied by suppression in the noise was observed. The correlations between LIBS peaks were also visualized using 2 D plots. Correlation strengths of atomic transitions were visualized in Aluminum (Al), Copper (Cu), and Brass whereas correlation strengths of atomic, atomic, and ionic transitions were visualized in Au Ag bimetallic targets with different compositions (Au30Ag70, Au50Ag50, Au80 Ag20). The improved spectra were subsequently used in the principal component analysis for classification studies of four compositions of bimetallic targets Au20Ag80, Au30Ag70, Au50 Ag50, and Au80Ag20 The variance of the first three principal components were found to be improved from the analysis. The accumulated percentage of explained variance of 95 was achieved with the first three components from improved spectra whereas only ~ 80 was achieved with the regular LIBS spectra from PCA studies | physics |
The orientation completion problem for a fixed class of oriented graphs asks whether a given partially oriented graph can be completed to an oriented graph in the class. Orientation completion problems have been studied recently for several classes of oriented graphs, yielding both polynomial time solutions and NP-completeness results. Local tournaments are a well-structured class of oriented graphs that generalize tournaments and their underlying graphs are intimately related to proper circular-arc graphs. Proper interval graphs are precisely those which can be oriented as acyclic local tournaments. It has been proved that the orientation completion problems for local tournaments and acyclic local tournaments are both polynomial time solvable. In this paper we identify the obstructions for acyclic local tournament orientation completions. These are in a sense minimal partially oriented graphs that cannot be completed to acyclic local tournaments. Our description of the obstructions imply that they can be recognized in polynomial time. In a companion paper we will determine all obstructions for local tournament orientation completions. | mathematics |
In recent years, many Blockchain based frameworks for transacting commodities on a congestible network have been proposed. In particular, as the number of controllable grid connected assets increases, there is a need for a decentralized, coupled economic and control mechanism to dynamically balance the entire electric grid. Blockchain based Transactive Energy (TE) systems have gained significant momentum as an approach to sustain the reliability and security of the power grid in order to support the flexibility of electricity demand. What is lacking in these designs, however, is a mechanism that physically verifies all the energy transactions, to keep the various inherently selfish players honest. In this paper, we introduce a secure peer-to-peer network mechanism for the physical validation of economic transactions cleared over a distributed ledger. The framework is $\textit{secure}$ in the sense that selfish and malicious agents that are trying to inject false data into the network are prevented from adversely affecting the optimal functionality of the verification process by detecting and isolating them from the communication network. Preliminary simulations focusing on TE show the workings of this framework. | electrical engineering and systems science |
It is well-know that deciding consistency for normal answer set programs (ASP) is NP-complete, thus, as hard as the satisfaction problem for classical propositional logic (SAT). The best algorithms to solve these problems take exponential time in the worst case. The exponential time hypothesis (ETH) implies that this result is tight for SAT, that is, SAT cannot be solved in subexponential time. This immediately establishes that the result is also tight for the consistency problem for ASP. However, accounting for the treewidth of the problem, the consistency problem for ASP is slightly harder than SAT: while SAT can be solved by an algorithm that runs in exponential time in the treewidth k, it was recently shown that ASP requires exponential time in k \cdot log(k). This extra cost is due checking that there are no self-supported true atoms due to positive cycles in the program. In this paper, we refine the above result and show that the consistency problem for ASP can be solved in exponential time in k \cdot log({\lambda}) where {\lambda} is the minimum between the treewidth and the size of the largest strongly-connected component in the positive dependency graph of the program. We provide a dynamic programming algorithm that solves the problem and a treewidth-aware reduction from ASP to SAT that adhere to the above limit. | computer science |
Certifying that quantum devices behave as intended is crucial for quantum information science. Here, methods are developed for certification of both state preparation devices and measurement devices based on prepare-and-measure experiments with independent devices. The experimenter assumes the independence of the devices and knowledge of the Hilbert space dimension. Thus no precise characterisation of any part of the experiment is required. The certification is based on a randomised version of unambiguous state discrimination and targets the class of state ensembles corresponding to quantum $t$-designs of any size and any dimension. These quantum designs are sets of states over which the average of any $t$-degree polynomial equals its average over all pure states, and they accommodate many of the most useful discrete structures in quantum information processing. Furthermore, it is shown that the same experiments also certify the detection efficiency of the measurement devices, as well as their non-projective nature. The presented methods can readily be implemented in experiments. | quantum physics |
An $n\overset{p}{\mapsto}m$ random access code (RAC) is an encoding of $n$ bits into $m$ bits such that any initial bit can be recovered with probability at least $p$, while in a quantum RAC (QRAC), the $n$ bits are encoded into $m$ qubits. Since its proposal, the idea of RACs was generalized in many different ways, e.g. allowing the use of shared entanglement (called entanglement-assisted random access code, or simply EARAC) or recovering multiple bits instead of one. In this paper we generalize the idea of RACs to recovering the value of a given Boolean function $f$ on any subset of fixed size of the initial bits, which we call $f$-random access codes. We study and give protocols for $f$-random access codes with classical ($f$-RAC) and quantum ($f$-QRAC) encoding, together with many different resources, e.g. private or shared randomness, shared entanglement ($f$-EARAC) and Popescu-Rohrlich boxes ($f$-PRRAC). The success probability of our protocols is characterized by the \emph{noise stability} of the Boolean function $f$. Moreover, we give an \emph{upper bound} on the success probability of any $f$-QRAC with shared randomness that matches its success probability up to a multiplicative constant (and $f$-RACs by extension), meaning that quantum protocols can only achieve a limited advantage over their classical counterparts. | quantum physics |
A violation of mirror symmetry in the $\eta\to\pi^+\pi^-\pi^0$ Dalitz plot has long been recognized as a signal of C and CP violation. Here we show how the isospin of the underlying C- and CP-violating structures can be reconstructed from their kinematic representation in the Dalitz plot. Our analysis of the most recent experimental data reveals, for the first time, that the C- and CP-violating amplitude with total isospin $I=2$ is much more severely suppressed than that with total isospin $I=0$. | high energy physics phenomenology |
Stochastic recurrent neural networks with latent random variables of complex dependency structures have shown to be more successful in modeling sequential data than deterministic deep models. However, the majority of existing methods have limited expressive power due to the Gaussian assumption of latent variables. In this paper, we advocate learning implicit latent representations using semi-implicit variational inference to further increase model flexibility. Semi-implicit stochastic recurrent neural network(SIS-RNN) is developed to enrich inferred model posteriors that may have no analytic density functions, as long as independent random samples can be generated via reparameterization. Extensive experiments in different tasks on real-world datasets show that SIS-RNN outperforms the existing methods. | computer science |
Exploring the nature of exotic multiquark candidates such as the $X(3872)$ plays a pivotal role in understanding quantum chromodynamics (QCD). Despite significant efforts, consensus on their internal structures is still lacking. As a prime example, it remains a pressing open question to decipher the $X(3872)$ state between two popular exotic configurations: a loose hadronic molecule or a compact tetraquark. We demonstrate a novel approach to help address this problem by studying the $X(3872)$ production in heavy ion collisions, where a hot fireball with ample light as well as charm (anti-)quarks is available for producing the exotics. Adopting a multiphase transport model (AMPT) for describing such collisions and implementing appropriate production mechanism of either molecule or tetraquark picture, we compute and compare a series of observables for $X(3872)$ in Pb-Pb collisions at the Large Hadron Collider. We find the fireball volume plays a crucial role, leading to a 2-order-of-magnitude difference in the $X(3872)$ yield and a markedly different centrality dependence between hadronic molecules and compact tetraquarks, thus offering a unique opportunity for distinguishing the two scenarios. We also make the first prediction of $X(3872)$ elliptic flow coefficient to be tested by future experimental measurements. | high energy physics phenomenology |
Aims: We survey the transmission spectrum of WASP-121 b for line-absorption by metals and molecules at high spectral resolution, and elaborate on existing interpretations of the optical transmission spectrum observed with HST/STIS and WFC3. Methods: We use the cross-correlation technique and direct differential spectroscopy to search for sodium and other neutral and ionised atoms, TiO, VO and SH in high-resolution transit spectra obtained with the HARPS spectrograph. We inject models assuming chemical and hydrostatic equilibrium with varying temperature and composition to enable model comparison, and employ two bootstrap methods to test the robustness of our detections. Results: We detect neutral Mg, Na, Ca, Cr, Fe, Ni and V, which we predict exists in equilibrium with a significant quantity of VO, supporting earlier observations by HST/WFC3. Non-detections of Ti and TiO support the hypothesis that Ti is depleted via a cold-trap mechanism as has been proposed in the literature. Atomic line depths are under-predicted by hydrostatic models by a factor of 1.5 to 8, confirming recent findings that the atmosphere is extended. We predict the existence of significant concentrations of gas-phase TiO$_2$, VO$_2$ and TiS, which could be important absorbers at optical and NIR wavelengths in hot Jupiter atmospheres, but for which accurate line-list data is currently not available. We find no evidence for absorption by SH, and find that inflated atomic lines can plausibly explain the slope of the transmission spectrum observed in the NUV with HST/STIS. The Na D lines are significantly broadened and show a difference in their respective depths of 15 scale heights, which is not expected from isothermal hydrostatic theory. | astrophysics |
Atmospheric escape from close-in Neptunes and hot Jupiters around sun-like stars driven by extreme ultraviolet (EUV) irradiation plays an important role in the evolution of exo-planets and in shaping their ensemble properties. Intermediate and low mass stars are brightest at EUV wavelengths at the very end of their lives, after they have expelled their envelopes and evolved into hot white dwarfs. Yet the effect of the intense EUV irradiation of giant planets orbiting young white dwarfs has not been assessed. We show that the giant planets in the solar system will experience significant hydrodynamic escape caused by the EUV irradiation from the white dwarf left behind by the Sun. A fraction of the evaporated volatiles will be accreted by the solar white dwarf, resulting in detectable photospheric absorption lines. As a large number of the currently known extra-solar giant planets will survive the metamorphosis of their host stars into white dwarfs, observational signatures of accretion from evaporating planetary atmospheres are expected to be common. In fact, one third of the known hot single white dwarfs show photospheric absorption lines of volatile elements, which we argue are indicative of ongoing accretion from evaporating planets. The fraction of volatile contaminated hot white dwarfs strongly decreases as they cool. We show that accretion from evaporating planetary atmospheres naturally explains this temperature dependence if more than 50 per cent of hot white dwarfs still host giant planets. | astrophysics |
The four-component cat state represents a particularly useful quantum state for realizing fault-tolerant continuous variable quantum computing. While such encoding has been experimentally generated and employed in the microwave regime, the states have not yet been produced in the optical regime. Here we propose a simple linear optical circuit combined with photon counters for the generation of such optical four-component cat states. This work might pave the way for the first experimental generation of fault-tolerant optical continuous variable quantum codes. | quantum physics |
The relationship between quantum physics and discrete mathematics is reviewed in this article. The Boolean functions unitary representation is considered. The relationship between Zhegalkin polynomial, which defines the algebraic normal form of Boolean function, and quantum logic circuits is described. It is shown that quantum information approach provides simple algorithm to construct Zhegalkin polynomial using truth table. Developed methods and algorithms have arbitrary Boolean function generalization with multibit input and multibit output. Such generalization allows us to use many-valued logic (k-valued logic, where k is a prime number). Developed methods and algorithms can significantly improve quantum technology realization. The presented approach is the baseline for transition from classical machine logic to quantum hardware. | quantum physics |
We illustrate the potential of a very high energy lepton collider (from 10 to 30 TeV center of mass energy) to explore new physics indirectly in the vector boson fusion double Higgs production process and in direct diboson production at high energy. Double Higgs production is found to be sensitive to the anomalous Higgs trilinear coupling at the percent level, and to the Higgs compositeness $\xi$ parameter at the per mille or sub-per mille level thanks to the measurement of the cross-section in the di-Higgs high invariant mass tail. High energy diboson (and tri-boson) production is sensitive to Higgs-lepton contact interaction operators at a scale of several tens or hundred TeV, corresponding to a reach on the Higgs compositeness scale well above the one of any other future collider project currently under discussion. This result follows from the unique capability of the very high energy lepton collider to measure Electroweak cross-sections at 10 TeV energy or more, where the effect of new physics at even higher energy is amplified. The general lesson is that the standard path towards precision physics, based on measurements of high-statistics processes such as single and double Higgs production, is accompanied at the very high energy lepton collider by a second strategy based on measurements at the highest available energy. | high energy physics phenomenology |
We derive the leading exponential finite volume corrections in two dimensional integrable models for non-diagonal form factors in diagonally scattering theories. These formulas are expressed in terms of the infinite volume form factors and scattering matrices. If the particles are bound states then the leading exponential finite-size corrections ($\mu$-terms) are related to virtual processes in which the particles disintegrate into their constituents. For non-bound state particles the leading exponential finite-size corrections (F-terms) come from virtual particles traveling around the finite world. In these F-terms a specifically regulated infinite volume form factor is integrated for the momenta of the virtual particles. The F-term is also present for bound states and the $\mu$-term can be obtained by taking an appropriate residue of the F-term integral. We check our results numerically in the Lee-Yang and sinh-Gordon models based on newly developed Hamiltonian truncations. | high energy physics theory |
The purpose of this paper is to present a comprehensive study of a coherent feedback network where the main component consists of two distant double quantum dot (DQD) qubits which are directly coupled to a cavity. This main component has recently been physically realized (van Woerkom, {\it et al.}, Microwave photon-mediated interactions between semiconductor qubits, Physical Review X, 8(4):041018, 2018). The feedback loop is closed by cascading this main component with a beamsplitter. The dynamics of this coherent feedback network is studied from three perspectives. First, an analytic form of the output single-photon state of the network driven by a single-photon state is derived; in particular, it is observed that coherent feedback elongates considerably the interaction between the input single photon and the network. Second, excitation probabilities of DQD qubits are computed when the network is driven by a single-photon input state. Moreover, if the input is vacuum but one of the two DQD qubits is initialized in its excited state, the explicit expression of the state of the network is derived, in particular, it is shown that the output field and the two DQD qubits can form an entangled state if the transition frequencies of two DQD qubits are equal. Finally, the exact form of the pulse shape is obtained by which the single-photon input can fully excite one of these two DQD qubits at any controllable time, which may be useful in the construction of $2$-qubit quantum gates. | quantum physics |
Recent pump-probe experiments on underdoped cuprates and similar systems suggest the existence of a transient superconducting state above $\mathrm{T}_c$. This poses the question how to reliably identify the emergence of long-range order, in particular superconductivity, out-of-equilibrium. We investigate this point by studying a quantum quench in an extended Hubbard model and by computing various observables, which are used to identify (quasi-)long-range order in equilibrium. Our findings imply that, in contrast to current experimental studies, it does not suffice to study the time evolution of the optical conductivity to identify superconductivity. In turn, we suggest to utilize time-resolved ARPES experiments to probe for the formation of a condensate in the two-particle channel. | condensed matter |
Quantum computers based on superconducting circuits are experiencing a rapid development, aiming at outperforming classical computers in certain useful tasks in the near future. However, the currently available chip fabrication technologies limit the capability of gathering a large number of high-quality qubits in a single superconducting chip, a requirement for implementing quantum error correction. Furthermore, achieving high connectivity in a chip poses a formidable technological challenge. Here, we propose a hybrid digital-analog quantum algorithm that enhances the physical connectivity among qubits coupled by an arbitrary inhomogeneous nearest-neighbour Ising Hamiltonian and generates an arbitrary all-to-all Ising Hamiltonian only by employing single-qubit rotations. Additionally, we optimize the proposed algorithm in the number of analog blocks and in the time required for the simulation. These results take advantage of the natural evolution of the system by combining the flexibility of digital steps with the robustness of analog quantum computing, allowing us to improve the connectivity of the hardware and the efficiency of quantum algorithms. | quantum physics |
The problem of estimating the size of a population based on a subset of individuals observed across multiple data sources is often referred to as capture-recapture or multiple-systems estimation. This is fundamentally a missing data problem, where the number of unobserved individuals represents the missing data. As with any missing data problem, multiple-systems estimation requires users to make an untestable identifying assumption in order to estimate the population size from the observed data. Approaches to multiple-systems estimation often do not emphasize the role of the identifying assumption during model specification, which makes it difficult to decouple the specification of the model for the observed data from the identifying assumption. We present a re-framing of the multiple-systems estimation problem that decouples the specification of the observed-data model from the identifying assumptions, and discuss how log-linear models and the associated no-highest-order interaction assumption fit into this framing. We present an approach to computation in the Bayesian setting which takes advantage of existing software and facilitates various sensitivity analyses. We demonstrate our approach in a case study of estimating the number of civilian casualties in the Kosovo war. Code used to produce this manuscript is available at https://github.com/aleshing/revisiting-identifying-assumptions. | statistics |
We introduce the software blocks_3d for computing four-point conformal blocks of operators with arbitrary Lorentz representations in 3d CFTs. It uses Zamolodchikov-like recursion relations to numerically compute derivatives of blocks around a crossing-symmetric configuration. It is implemented as a heavily optimized, multithreaded, C++ application. We give performance benchmarks for correlators containing scalars, fermions, and stress tensors. As an example application, we recompute bootstrap bounds on four-point functions of fermions and study whether a previously observed sharp jump can be explained using the "fake primary" effect. We conclude that the fake primary effect cannot fully explain the jump and the possible existence of a "dead-end" CFT near the jump merits further study. | high energy physics theory |
The NICER collaboration recently published a joint estimate of the mass and the radius of PSR J0030+0451, derived via X-ray pulse-profile modeling. Raaijmakers et al. (2019) explored the implications of this measurement for the dense matter equation of state (EOS) using two parameterizations of the high-density EOS: a piecewise-polytropic model, and a model based on the speed of sound in neutron stars. In this work we obtain further constraints on the EOS following this approach, but we also include information about the tidal deformability of neutron stars from the gravitational wave signal of the compact binary merger GW170817. We compare the constraints on the EOS to those set by the recent measurement of a 2.14 solar mass pulsar, included as a likelihood function approximated by a Gaussian, and find a small increase in information gain. To show the flexibility of our method, we also explore the possibility that GW170817 was a neutron star-black hole merger, which yields weaker constraints on the EOS. | astrophysics |
Despite significant efforts over the last few years to build a robust automatic speech recognition (ASR) system for different acoustic settings, the performance of the current state-of-the-art technologies significantly degrades in noisy reverberant environments. Convolutional Neural Networks (CNNs) have been successfully used to achieve substantial improvements in many speech processing applications including distant speech recognition (DSR). However, standard CNN architectures were not efficient in capturing long-term speech dynamics, which are essential in the design of a robust DSR system. In the present study, we address this issue by investigating variants of large receptive field CNNs (LRF-CNNs) which include deeply recursive networks, dilated convolutional neural networks, and stacked hourglass networks. To compare the efficacy of the aforementioned architectures with the standard CNN for Wall Street Journal (WSJ) corpus, we use a hybrid DNN-HMM based speech recognition system. We extend the study to evaluate the system performances for distant speech simulated using realistic room impulse responses (RIRs). Our experiments show that with fixed number of parameters across all architectures, the large receptive field networks show consistent improvements over the standard CNNs for distant speech. Amongst the explored LRF-CNNs, stacked hourglass network has shown improvements with a 8.9% relative reduction in word error rate (WER) and 10.7% relative improvement in frame accuracy compared to the standard CNNs for distant simulated speech signals. | electrical engineering and systems science |
The primary analysis of randomized cancer screening trials for cancer typically adheres to the intention-to-screen principle, measuring cancer-specific mortality reductions between screening and control arms. These mortality reductions result from a combination of the screening regimen, screening technology and the effect of the early, screening-induced, treatment. This motivates addressing these different aspects separately. Here we are interested in the causal effect of early versus delayed treatments on cancer mortality among the screening-detectable subgroup, which under certain assumptions is estimable from conventional randomized screening trial using instrumental variable type methods. To define the causal effect of interest, we formulate a simplified structural multi-state model for screening trials, based on a hypothetical intervention trial where screening detected individuals would be randomized into early versus delayed treatments. The cancer-specific mortality reductions after screening detection are quantified by a cause-specific hazard ratio. For this, we propose two estimators, based on an estimating equation and a likelihood expression. The methods extend existing instrumental variable methods for time-to-event and competing risks outcomes to time-dependent intermediate variables. Using the multi-state model as the basis of a data generating mechanism, we investigate the performance of the new estimators through simulation studies. In addition, we illustrate the proposed method in the context of CT screening for lung cancer using the US National Lung Screening Trial (NLST) data. | statistics |
State preparation is a process encoding the classical data into the quantum systems. Based on quantum phase estimation, we propose the specific quantum circuits for a deterministic state preparation algorithm and a probabilistic state preparation algorithm. To discuss the gate complexity in these algorithms, we decompose the diagonal unitary operators included in the phase estimation algorithms into the basic gates. Thus, we associate the state preparation problem with the decomposition problem of the diagonal unitary operators. We analyse the fidelities in the two algorithms and discuss the success probability in the probabilistic algorithm. In this case, we explain that the efficient decomposition of the corresponding diagonal unitary operators is the sufficient condition for state preparation problems. | quantum physics |
The marked interplay between the crystalline, electronic, and magnetic structure of atomically thin magnets has been regarded as the key feature for designing next-generation magneto-optoelectronic devices. In this respect, a detailed understanding of the microscopic interactions underlying the magnetic responses of these crystals is of primary importance. Here, we combine model Hamiltonians with multi-reference configuration interaction wavefunctions to accurately determine the strength of the spin couplings in the prototypical single-layer magnet CrI$_3$. Our calculations identify the (ferromagnetic) Heisenberg exchange interaction $J = -1.44$ meV as the dominant term, being the inter-site magnetic anisotropies substantially {weaker}. We also find that single-layer CrI$_3$ features an out-of-plane easy axis ensuing from a single-ion anisotropy $A = -0.10$ meV, and predict $g$-tensor in-plane components $g_{xx} = g_{yy} = 1.90$ and out-of-plane component $g_{zz} = 1.92$. In addition, we assess the performance of a dozen widely used density functionals against our accurate correlated wavefunctions {calculations} and available experimental data, thereby establishing reference results for future first-principles investigations. Overall, our findings offer a firm theoretical ground to experimental observations. | condensed matter |
We study supervisor localization for timed discrete-event systems under partial observation and communication delay in the Brandin-Wonham framework. First, we employ timed relative observability to synthesize a partial-observation monolithic supervisor; the control actions of this supervisor include not only disabling action of prohibitible events (as that of controllable events in the untimed case) but also "clock-preempting" action of forcible events. Accordingly we decompose the supervisor into a set of partial-observation local controllers one for each prohibitible event, as well as a set of partial-observation local preemptors one for each forcible event. We prove that these local controllers and preemptors collectively achieve the same controlled behavior as the partial-observation monolithic supervisor does. Moreover, we propose channel models for inter-agent event communication and impose bounded and unbounded delay as temporal specifications. In this formulation, there exist multiple distinct observable event sets; thus we employ timed relative coobservability to synthesize partial-observation decentralized supervisors, and then localize these supervisors into local controllers and preemptors. The above results are illustrated by a timed workcell example. | computer science |
In 1977 Korchinski presented a new type of shock discontinuity in conservation laws. These singular solutions were coined $\delta$-shocks since there is a time dependent Dirac delta involved. A naive description is that such $\delta$-shock is of the overcompressive type: a two-family shock wave the four characteristic lines of which impinge into the shock itself. In this work, we open the fan of solutions by studying two-family waves without intermediate constant states but, possessing central rarefactions and also comprising $\delta$-shocks. | mathematics |
We summarise some of our recent works on $L_\infty$-algebras and quasi-groups with regard to higher principal bundles and their applications in twistor theory and gauge theory. In particular, after a lightning review of $L_\infty$-algebras, we discuss their Maurer-Cartan theory and explain that any classical field theory admitting an action can be reformulated in this context with the help of the Batalin-Vilkovisky formalism. As examples, we explore higher Chern-Simons theory and Yang-Mills theory. We also explain how these ideas can be combined with those of twistor theory to formulate maximally superconformal gauge theories in four and six dimensions by means of $L_\infty$-quasi-isomorphisms, and we propose a twistor space action. | high energy physics theory |
We present a new approach to derive the connectivity properties of pairwise interacting n-body systems in thermal equilibrium. We formulate an integral equation that relates the pair connectedness to the distribution of nearest neighbors. For one-dimensional systems with nearest-neighbor interactions, the nearest-neighbor distribution is, in turn, related to the pair correlation function g through a simple integral equation. As a consequence, for those systems, we arrive at an integral equation relating g to the pair connectedness, which is readily solved even analytically if g is specified analytically. We demonstrate the procedure for a variety of pair-potentials including fully penetrable spheres as well as impenetrable spheres, the only two systems for which analytical results for the pair connectedness exist. However, the approach is not limited to nearest-neighbor interactions in one dimension. Hence, we also outline the treatment of external fields and long-ranged interactions, and we illustrate how the formalism can applied to higher-dimensional systems using the three-dimensional ideal gas as an example. | condensed matter |
We have developed a new method for fabrication of ultra-low noise silver-silver chloride (Ag-AgCl) reference electrode with dramatically high stability, useful in biomedical precise measurements. Low noise level and high stability are the most important features when evaluating reference electrode for example when used as a half-cell at electrochemical cells. The noise was measured for bare Ag-AgCl electrode pairs in a saline electrolyte without any junction. The noise spectral density for such measurements increases at low frequencies due to flicker (1/f) noise. Thermal noise from the real part of the electrode impedance is always lower than 1/f noise and is shown as offset voltage in noise spectrum. Electrode noise is highly correlated to its offset voltage. Moreover, the variance of the potential difference in our measurements is correlated to the noise level of electrodes. In this paper, offset voltage spectrum and the variance of data while doing measurements are proposed for identification of noise in our electrodes. The results presented suggest a new process for fabrication of electrode pairs with differential noise potential of 16 nanovolt root mean square at frequency of 1 Hz. | physics |
Diagnosis of pulmonary lesions from computed tomography (CT) is important but challenging for clinical decision making in lung cancer related diseases. Deep learning has achieved great success in computer aided diagnosis (CADx) area for lung cancer, whereas it suffers from label ambiguity due to the difficulty in the radiological diagnosis. Considering that invasive pathological analysis serves as the clinical golden standard of lung cancer diagnosis, in this study, we solve the label ambiguity issue via a large-scale radio-pathomics dataset containing 5,134 radiological CT images with pathologically confirmed labels, including cancers (e.g., invasive/non-invasive adenocarcinoma, squamous carcinoma) and non-cancer diseases (e.g., tuberculosis, hamartoma). This retrospective dataset, named Pulmonary-RadPath, enables development and validation of accurate deep learning systems to predict invasive pathological labels with a non-invasive procedure, i.e., radiological CT scans. A three-level hierarchical classification system for pulmonary lesions is developed, which covers most diseases in cancer-related diagnosis. We explore several techniques for hierarchical classification on this dataset, and propose a Leaky Dense Hierarchy approach with proven effectiveness in experiments. Our study significantly outperforms prior arts in terms of data scales (6x larger), disease comprehensiveness and hierarchies. The promising results suggest the potentials to facilitate precision medicine. | electrical engineering and systems science |
This paper develops two orthogonal contributions to scalable sparse regression for competing risks time-to-event data. First, we study and accelerate the broken adaptive ridge method (BAR), an $\ell_0$-based iteratively reweighted $\ell_2$-penalization algorithm that achieves sparsity in its limit, in the context of the Fine-Gray (1999) proportional subdistributional hazards (PSH) model. In particular, we derive a new algorithm for BAR regression, named cycBAR, that performs cyclic update of each coordinate using an explicit thresholding formula. The new cycBAR algorithm effectively avoids fitting multiple reweighted $\ell_2$-penalizations and thus yields impressive speedups over the original BAR algorithm. Second, we address a pivotal computational issue related to fitting the PSH model. Specifically, the computation costs of the log-pseudo likelihood and its derivatives for PSH model grow at the rate of $O(n^2)$ with the sample size $n$ in current implementations. We propose a novel forward-backward scan algorithm that reduces the computation costs to $O(n)$. The proposed method applies to both unpenalized and penalized estimation for the PSH model and has exhibited drastic speedups over current implementations. Finally, combining the two algorithms can yields $>1,000$ fold speedups over the original BAR algorithm. Illustrations of the impressive scalability of our proposed algorithm for large competing risks data are given using both simulations and a United States Renal Data System data. | statistics |
We theoretically investigate the effects of Coulomb interaction, at the level of unscreened Hartree-Fock approximation, on third harmonic generation of undoped graphene in an equation of motion framework. The unperturbed electronic states are described by a widely used two-band tight binding model, and the Coulomb interaction is described by the Ohno potential. The ground state is renormalized by taking into account the Hartree-Fock term, and the optical conductivities are obtained by numerically solving the equations of motion. The absolute values of conductivity for third harmonic generation depend on the photon frequency $\Omega$ as $\Omega^{-n}$ for $\hbar\Omega<1$, and then show a peak as $3\hbar\Omega$ approaches the renormalized energy of the $M$ point. Taking into account the Coulomb interaction, $n$ is found to be $5.5$, which is significantly greater than the value of $4$ found with the neglect of the Coulomb interaction. Therefore the Coulomb interaction enhances third harmonic generation at low photon energies -- for our parameters $\hbar\Omega<0.8$~eV -- and then reduces it until the photon energy reaches about $2.1$~eV. The effect of the background dielectric constant is also considered. | condensed matter |
We propose a formula for the exact central charge of a B-type D-brane that is expected to hold in all regions of the Kahler moduli space of a Calabi-Yau. For Landau-Ginzburg orbifolds we propose explicit expressions for the mathematical objects that enter into the central charge formula. We show that our results are consistent with results in FJRW theory and the hemisphere partition function of the gauged linear sigma model. | high energy physics theory |
In this work we study the fluctuation and dissipation of a string attached to a brane in a deformed and backreated AdS-Schwarzschild spacetime. This space is a solution of Einstein-dilaton equations and contains a conformal exponential factor $\exp(k/r^2)$ in the metric. We consider the backreaction contributions coming only from the exponential warp factor on the AdS-Schwarzschild black hole, where the string and brane are in the probe approximation. Within this Lorentz invariant holographic model we have computed the admittance, the diffusion coefficient, the two-point functions and the regularized mean square displacement $s^2_{reg}$. From this quantity we obtain the diffuse and ballistic regimes characteristic of the Brownian motion. From the two-point functions and the admittance, we also have checked the well know fluctuation-dissipation theorem in this set up. | high energy physics theory |
Neural networks are among the most powerful nonlinear models used to address supervised learning problems. Similar to most machine learning algorithms, neural networks produce point predictions and do not provide any prediction interval which includes an unobserved response value with a specified probability. In this paper, we proposed the $k$-fold prediction interval method to construct prediction intervals for neural networks based on $k$-fold cross validation. Simulation studies and analysis of 10 real datasets are used to compare the finite-sample properties of the prediction intervals produced by the proposed method and the split conformal (SC) method. The results suggest that the proposed method tends to produce narrower prediction intervals compared to the SC method while maintaining the same coverage probability. Our experimental results also reveal that the proposed $k$-fold prediction interval method produces effective prediction intervals and is especially advantageous relative to competing approaches when the number of training observations is limited. | statistics |
Fact-checking is the task of verifying the veracity of claims by assessing their assertions against credible evidence. The vast majority of fact-checking studies focus exclusively on political claims. Very little research explores fact-checking for other topics, specifically subject matters for which expertise is required. We present the first study of explainable fact-checking for claims which require specific expertise. For our case study we choose the setting of public health. To support this case study we construct a new dataset PUBHEALTH of 11.8K claims accompanied by journalist crafted, gold standard explanations (i.e., judgments) to support the fact-check labels for claims. We explore two tasks: veracity prediction and explanation generation. We also define and evaluate, with humans and computationally, three coherence properties of explanation quality. Our results indicate that, by training on in-domain data, gains can be made in explainable, automated fact-checking for claims which require specific expertise. | computer science |
Superconductor-ferromagnet-superconductor Josephson junctions are known to exist in the $0$ and $\pi$ states with the transitions between them controlled by the temperature and ferromagnetic interlayer thickness. We demonstrate that these transitions can be controlled also by the external magnetic field directed perpendicular to the layers. By varying the ratio of diffusion coefficients in superconducting and ferromagnetic layers, these field-controlled transitions can be made detectable for arbitrary large value of the exchange energy in the ferromagnet. We also show that the $0$-$\pi$ transitions in the perpendicular field can be observed as the specific features of the flux-flow conductivity dependencies on the ferromagnetic thickness in accordance with recent experimental results. | condensed matter |
In this paper, we propose a novel effective light-weight framework, called LightTrack, for online human pose tracking. The proposed framework is designed to be generic for top-down pose tracking and is faster than existing online and offline methods. Single-person Pose Tracking (SPT) and Visual Object Tracking (VOT) are incorporated into one unified functioning entity, easily implemented by a replaceable single-person pose estimation module. Our framework unifies single-person pose tracking with multi-person identity association and sheds first light upon bridging keypoint tracking with object tracking. We also propose a Siamese Graph Convolution Network (SGCN) for human pose matching as a Re-ID module in our pose tracking system. In contrary to other Re-ID modules, we use a graphical representation of human joints for matching. The skeleton-based representation effectively captures human pose similarity and is computationally inexpensive. It is robust to sudden camera shift that introduces human drifting. To the best of our knowledge, this is the first paper to propose an online human pose tracking framework in a top-down fashion. The proposed framework is general enough to fit other pose estimators and candidate matching mechanisms. Our method outperforms other online methods while maintaining a much higher frame rate, and is very competitive with our offline state-of-the-art. We make the code publicly available at: https://github.com/Guanghan/lighttrack. | computer science |
To make precise the sense in which the operational predictions of quantum theory conflict with a classical worldview, it is necessary to articulate a notion of classicality within an operational framework. A widely applicable notion of classicality of this sort is whether or not the predictions of a given operational theory can be explained by a generalized-noncontextual ontological model. We here explore what notion of classicality this implies for the generalized probabilistic theory (GPT) that arises from a given operational theory, focusing on prepare-measure scenarios. We first show that, when mapping an operational theory to a GPT by quotienting relative to operational equivalences, the constraint of explainability by a generalized-noncontextual ontological model is mapped to the constraint of explainability by an ontological model. We then show that, under the additional assumption that the ontic state space is of finite cardinality, this constraint on the GPT can be expressed as a geometric condition which we term simplex-embeddability. Whereas the traditional notion of classicality for a GPT is that its state space be a simplex and its effect space be the dual of this simplex, simplex-embeddability merely requires that its state space be embeddable in a simplex and its effect space in the dual of that simplex. We argue that simplex-embeddability constitutes an intuitive and freestanding notion of classicality for GPTs. Our result also has applications to witnessing nonclassicality in prepare-measure experiments. | quantum physics |
We study fermion number non-conservation (or chirality breaking) in Abelian gauge theories at finite temperature. We consider the presence of a chemical potential $\mu$ for the fermionic charge, and monitor its evolution with real-time classical lattice simulations. This method accounts for short-scale fluctuations not included in the usual effective magneto-hydrodynamics (MHD) treatment. We observe a self-similar decay of the chemical potential, accompanied by an inverse cascade process in the gauge field that leads to a production of long-range helical magnetic fields. We also study the chiral charge dynamics in the presence of an external magnetic field $B$, and extract its decay rate $\Gamma_5 \equiv -{d\log \mu\over dt}$. We provide in this way a new determination of the gauge coupling and magnetic field dependence of the chiral rate, which exhibits a best fit scaling as $\Gamma_5 \propto e^{11/2}B^2$. We confirm numerically the fluctuation-dissipation relation between $\Gamma_5$ and $\Gamma_{\rm diff}$, the Chern-Simons diffusion rate, which was obtained in a previous study. Remarkably, even though we are outside the MHD range of validity, the dynamics observed are in qualitative agreement with MHD predictions. The magnitude of the chiral/diffusion rate is however a factor $\sim 10$ times larger than expected in MHD, signaling that we are in reality exploring a different regime accounting for short scale fluctuations. This discrepancy calls for a revision of the implications of fermion number and chirality non-conservation in finite temperature Abelian gauge theories, though not definite conclusion can be made at this point until hard-thermal-loops (HTL) are included in the lattice simulations. | high energy physics theory |
We present the first Hubble diagram of superluminous supernovae (SLSNe) out to a redshift of two, together with constraints on the matter density, $\Omega_{\rm M}$, and the dark energy equation-of-state parameter, $w(\equiv p/\rho)$. We build a sample of 20 cosmologically useful SLSNe~I based on light curve and spectroscopy quality cuts. We confirm the robustness of the peak decline SLSN~I standardization relation with a larger dataset and improved fitting techniques than previous works. We then solve the SLSN model based on the above standardisation via minimisation of the $\chi^2$ computed from a covariance matrix which includes statistical and systematic uncertainties. For a spatially flat $\Lambda$CDM cosmological model, we find $\Omega_{\rm M}=0.38^{+0.24}_{-0.19}$, with a rms of 0.27 mag for the residuals of the distance moduli. For a $w_0w_a$CDM cosmological model, the addition of SLSNe~I to a `baseline' measurement consisting of Planck temperature together with type Ia supernovae, results in a small improvement in the constraints of $w_0$ and $w_a$ of 4\%. We present simulations of future surveys with 868 and 492 SLSNe I (depending on the configuration used) and show that such a sample can deliver cosmological constraints in a flat $\Lambda$CDM model with the same precision (considering only statistical uncertainties) as current surveys that use type Ia supernovae, while providing a factor 2-3 improvement in the precision of the constraints on the time variation of dark energy, $w_0$ and $w_a$. This paper represents the proof-of-concept for superluminous supernova cosmology, and demonstrates they can provide an independent test of cosmology in the high-redshift ($z>1$) universe. | astrophysics |
Ultrafast ultrasound (US) revolutionized biomedical imaging with its capability of acquiring full-view frames at over 1 kHz, unlocking breakthrough modalities such as shear-wave elastography and functional US neuroimaging. Yet, it suffers from strong diffraction artifacts, mainly caused by grating lobes, side lobes, or edge waves. Multiple acquisitions are typically required to obtain a sufficient image quality, at the cost of a reduced frame rate. To answer the increasing demand for high-quality imaging from single-shot acquisitions, we propose a two-step convolutional neural network (CNN)-based image reconstruction method, compatible with real-time imaging. A low-quality estimate is obtained by means of a backprojection-based operation, akin to conventional delay-and-sum beamforming, from which a high-quality image is restored using a residual CNN with multi-scale and multi-channel filtering properties, trained specifically to remove the diffraction artifacts inherent to ultrafast US imaging. To account for both the high dynamic range and the radio frequency property of US images, we introduce the mean signed logarithmic absolute error (MSLAE) as training loss function. Experiments were conducted with a linear transducer array, in single plane wave (PW) imaging. Trainings were performed on a simulated dataset, crafted to contain a wide diversity of structures and echogenicities. Extensive numerical evaluations demonstrate that the proposed approach can reconstruct images from single PWs with a quality similar to that of gold-standard synthetic aperture imaging, on a dynamic range in excess of 60 dB. In vitro and in vivo experiments show that trainings performed on simulated data translate well to experimental settings. | electrical engineering and systems science |
Processing quantum information on continuous variables requires a highly nonlinear element in order to attain universality. Noise reduction in processing such quantum information involves the use of a nonlinear phase state as a non-Gaussian ancilla. A necessary condition for a nonlinear phase state to implement a nonlinear phase gate is that noise in a selected nonlinear quadrature should decrease below the level of classical states. A reduction of the variance in this nonlinear quadrature below the ground state of the ancilla, a type of nonlinear squeezing, is the resource embedded in these non-Gaussian states and a figure of merit for nonlinear quantum processes. Quantum optomechanics with levitating nanoparticles trapped in nonlinear optical potentials is a promising candidate to achieve such resources in a flexible way. We provide a scheme for reconstructing this figure of merit, which we call nonlinear squeezing, in standard linear quantum optomechanics, analysing the effects of mechanical decoherence processes on the reconstruction and show that all mechanical states which exhibit reduced noise in this nonlinear quadrature are nonclassical. | quantum physics |
This paper provides a new process which integrates an inventive problem solving method into one modern software development program, making it part of the software development process. The research question is how to improve software development process which tech startups could adopt with minor project management skills. The Systematic Innovation Mounted Software Development Process, a combination of Agile and Systematic Innovation, provides an alternative development process which is targeted to adapt idea generation into software products. The intuitive project management framework helps technology driven companies to manage their software projects more effectively. The implication and aim of this research are providing the guideline to help the entrepreneurs for managing their project properly. The Systematic Innovation model helps to generate new ideas and innovative ways to solve problems with the collaboration with the existing Agile model. The new software development process and associated techniques could impact the current software development industry significantly, especially software startup companies, because these powerful tools can help reduce managerial workloads of the companies and give them more time to remain focused on their key technologies. | computer science |
The influence of the defect position on absorption and emission in the cholesteric liquid crystal with an isotropic defect inside is studied. It is shown that for non-diffracting circularly polarized incident light absorption/emission is maximum if the defect is in the centre of the system; and for diffracting circularly polarized incident light absorption/emission is maximum if the defect is shifted from the centre of the system to its left border from where light is incident. The influence of anisotropic absorption in the cholesteric liquid crystal layer on photonic states of density was investigated, too. | physics |
In this paper the Rayleigh-Taylor instability (RTI) along the density interfaces of gravity-current fronts is analyzed. Both the location and the spanwise wavenumber of the most unstable mode determined by the local dispersion relation agree with those of the strongest perturbation obtained from numerical simulations, suggesting that the original formation mechanism of lobes and clefts at the current front is RTI. Furthermore, the predictions of the semi-infinite RTI model, i.e. the original dominating spanwise wavenumber of the Boussinesq current substantially depends on the Prandtl number and has a 1/3 scaling law with the Grashof number, are confirmed by the three-dimensional numerical simulations. | physics |
Recent pulsar timing data reported by the NANOGrav collaboration indicates the existence of a stochastic gravitational wave (GW) background at a frequency $f\sim 10^{-8}~\rm Hz$. We show that a dark sector consisting of a Standard Model (SM) gauge singlet fermion $\chi$ and a singlet scalar $\phi$, both charged under a $Z_4$ symmetry, is capable of generating such a low frequency GW via strong first order phase transition (SFOPT) through the modification of the standard cosmological history, where we assume faster-than-usual expansion at pre-BBN times driven by a new cosmological species $\varphi$ whose energy density red-shifts with the scale factor as $\rho_\varphi\propto a^{-\left(4+n\right)}$. Depending on the choice of the fast expansion parameters, reheat temperature and effective scale of the theory, it is also possible to address correct dark matter (DM) relic abundance via freeze-in. We show that a successful first order phase transition explaining NANOGrav results together with PLANCK observed DM abundance put bound on the fast expansion parameters requiring $n\lesssim 4$ to explain both. | astrophysics |
We generalize the textbook Kronig-Penney model to realistic conditions for a quantum-particle moving in the quasi-one-dimensional (quasi-1D) waveguide, where motion in the transverse direction is confined by a harmonic trapping potential. Along the waveguide, the particle scatters on an infinite array of regularized delta potentials. Our starting point is the Lippmann-Schwinger equation, which for quasi-1D geometry can be solved exactly, based on the analytical formula for the quasi-1D Green's function. We study the properties of eigen-energies as a function of particle quasi-momentum, which form band structure, as in standard Kronig-Penney model. We test our model by comparing it to the numerical calculations for an atom scattering on an infinite chain of ions in quasi-1D geometry. The agreement is fairly good and can be further improved by introducing energy-dependent scattering length in the regularized delta potential. The energy spectrum exhibits the presence of multiple overlapping bands resulting from excitations in the transverse direction. At large lattice constants, our model reduces to standard Kronig-Penney result with one-dimensional coupling constant for quasi-1D scattering, exhibiting confinement-induced resonances. In the opposite limit, when lattice constant becomes comparable to harmonic oscillator length of the transverse potential, we calculate the correction to the quasi-1D coupling constant due to the quantum interference between scatterers. Finally, we calculate the effective mass for the lowest band and show that it becomes negative for large and positive scattering lengths. | quantum physics |
Hierarchically hyperbolic spaces (HHSs) are a large class of spaces that provide a unified framework for studying the mapping class group, right-angled Artin and Coxeter groups, and many 3--manifold groups. We investigate strongly quasiconvex subsets in this class and characterize them in terms of their contracting properties, relative divergence, the coarse median structure, and the hierarchical structure itself. Along the way, we obtain new tools to study HHSs, including two new equivalent definitions of hierarchical quasiconvexity and a version of the bounded geodesic image property for strongly quasconvex subsets. Utilizing our characterization, we prove that the hyperbolically embedded subgroups of hierarchically hyperbolic groups are precisely those which are almost malnormal and strongly quasiconvex, producing a new result in the case of the mapping class group. We also apply our characterization to study strongly quasiconvex subsets in several specific examples of HHSs. We show that while many commonly studied HHSs have the property that that every strongly quasiconvex subset is either hyperbolic or coarsely covers the entire space, right-angled Coxeter groups exhibit a wide variety of strongly quasiconvex subsets. | mathematics |
In this paper we study shells of matter and black holes on the expanding bubbles realizing de Sitter space, that were proposed in arXiv:1807.01570. The explicit solutions that we find for the black holes, can also be used to construct Randall-Sundrum braneworld black holes in four dimensions. | high energy physics theory |
The integrable nonlinear Schr\"odinger equation (NLSE) is a fundamental model of nonlinear science which also has important consequences in engineering. The powerful framework of the periodic inverse scattering transform (IST) provides a description of the nonlinear phenomena modulational instability and Fermi-Pasta-Ulam-Tsingou (FPUT) recurrence in terms of exact solutions. It associates the complex nonlinear dynamics with invariant nonlinear spectral degrees of freedom that may be used to encode information. While optical fiber is an ideal testing ground of its predictions, maintaining integrability over sufficiently long distances to observe recurrence, as well as synthesizing and measuring the field in both amplitude and phase on the picosecond timescales of typical experiments is challenging. Here we report on the experimental realization of FPUT recurrence in terms of an exact space-time-periodic solution of the integrable NLSE in a testbed for optical communication experiments. The complex-valued initial condition is constructed by means of the finite-gap integration method, modulated onto the optical carrier driven by an arbitrary waveform generator and launched into a recirculating fiber loop with periodic amplification. The measurement with an intradyne coherent receiver after a predetermined number of revolutions provides a non-invasive full-field characterization of the space-time dynamics. The recurrent space-time evolution is in close agreement with theoretical predictions over a distance of 9000 km. Nonlinear spectral analysis reveals an invariant nonlinear spectrum. The space-time scale exceeds that of previous experiments on FPUT recurrence in fiber by three orders of magnitude. | electrical engineering and systems science |
By using conformal symmetry we unify the standard model of particle physics with gravity in a consistent quantum field theory which describes all the fundamental particles and forces of nature. | high energy physics phenomenology |
For hydrological applications, such as urban flood modelling, it is often important to be able to simulate sub-daily rainfall time series from stochastic models. However, modelling rainfall at this resolution poses several challenges, including a complex temporal structure including long dry periods, seasonal variation in both the occurrence and intensity of rainfall, and extreme values. We illustrate how the hidden Markov framework can be adapted to construct a compelling model for sub-daily rainfall, which is capable of capturing all of these important characteristics well. These adaptations include clone states and non-stationarity in both the transition matrix and conditional models. Set in the Bayesian framework, a rich quantification of both parametric and predictive uncertainty is available, and thorough model checking is made possible through posterior predictive analyses. Results from the model are interpretable, allowing for meaningful examination of seasonal variation and medium to long term trends in rainfall occurrence and intensity. To demonstrate the effectiveness of our approach, both in terms of model fit and interpretability, we apply the model to an 8-year long time series of hourly observations. | statistics |
The dispersion of a tracer in a fluid flow is influenced by the Lagrangian motion of fluid elements. Even in laminar regimes, the irregular chaotic behavior of a fluid flow can lead to effective stirring that rapidly redistributes a tracer throughout the domain. When the advected particles possess a finite size and nontrivial shape, however, their dynamics can differ markedly from passive tracers, thus affecting the dispersion phenomena. Here we investigate the behavior of neutrally buoyant particles in 2-dimensional chaotic flows, combining numerical simulations and laboratory experiments. We show that depending on the particles shape and size, the underlying Lagrangian coherent structures can be altered, resulting in distinct dispersion phenomena within the same flow field. Experiments performed in a two-dimensional cellular flow, exhibited a focusing effect in vortex cores of particles with anisotropic shape. In agreement with our numerical model, neutrally buoyant ellipsoidal particles display markedly different trajectories and overall organization than spherical particles, with a clustering in vortices that changes accordingly with the aspect ratio of the particles. | physics |
The rules governing the essentially algebraic notion of a category with families have been observed (independently) by Steve Awodey and Marcelo Fiore to precisely match those of a representable natural transformation between presheaves. This provides us with a natural, functorial description of essentially algebraic objects which are used to model dependent type theory; following Steve Awodey, we call them 'natural models'. We can view natural models from several different viewpoints, of which we focus on three in this thesis. First, natural models are essentially algebraic, meaning that they can be described by specifying operations between sorts, subject to equational axioms; this allows us to assemble natural models into a category with certain beneficial properties. Second, since natural models are natural transformations between presheaves, they are morphisms in a locally cartesian closed category, meaning that they can be regarded as polynomials. Third, since natural models admit interpretations of dependent type theory, we can use them to provide a functorial semantics. | mathematics |
We present the first $\mathrm{o}(n)$-space polynomial-time algorithm for computing the length of a longest common subsequence. Given two strings of length $n$, the algorithm runs in $\mathrm{O}(n^{3})$ time with $\mathrm{O}\left(\frac{n \log^{1.5} n}{2^{\sqrt{\log n}}}\right)$ bits of space. | computer science |
We study the non-linear beam splitter in matter-wave interferometers using ultracold quantum gases in a double-well configuration in presence of non-local interactions inducing inter-well density-density coupling, as they can be realized, e.g., with dipolar gases. We explore this effect after considering different input states, in the form of either coherent, or Twin-Fock, or NOON states. We first review the non-interacting limit and the case in which only the local interaction is present, including the study of sensitivity near the self-trapping threshold. Then, we consider the two-mode model in the presence of inter-well interactions and consider the scaling of the sensitivity as a function of the non-local coupling strength. Our analysis clearly shows that non-local interactions can compensate the degradation of the sensitivity induced by local interactions, so that they may be used to restore optimal sensitivity. | condensed matter |
Cosmic rays propagate through the Galaxy and encounter systems that may trap them temporarily, as well as magnetic field structures that induce chaotic behavior on their trajectories. In particular, this is the case for particles that propagate in the local interstellar medium and interact with the heliospheric magnetic field before being detected on Earth. As a consequence, the observed cosmic-ray arrival direction distribution is affected by the heliosphere as long as their gyro-radius is smaller or comparable with the heliospheric size, i.e. in the TeV energy range. The chaotic nature of those cosmic-ray particle trajectories that are temporarily trapped inside the heliosphere can be characterized using the Finite-Time Lyapunov Exponents. Specifically, we will delve into the case of particles trapped in a heliospheric-inspired toy model of a static magnetic bottle configuration and the presence of temporal magnetic perturbations. In this work, we also suggest that a time-variability could prove to be important in the understanding of the TeV cosmic-ray anisotropy at Earth. | astrophysics |
Quantum generalizations of R\'enyi's entropies are a useful tool to describe a variety of operational tasks in quantum information processing. Two families of such generalizations turn out to be particularly useful: the Petz quantum R\'enyi divergence $\bar{D}_{\alpha}$ and the minimal quantum R\'enyi divergence $\widetilde{D}_{\alpha}$. Moreover, the maximum quantum R\'enyi divergence $\widehat{D}_{\alpha}$ is of particular mathematical interest. In this Master thesis, we investigate relations between these divergences and their applications in quantum information theory. Our main result is a reverse Araki-Lieb-Thirring inequality that implies a new relation between the minimal and the Petz divergence, namely that $\alpha \bar{D}_{\alpha}(\rho \| \sigma) \leqslant \widetilde{D}_{\alpha}(\rho \| \sigma)$ for $\alpha \in [0,1]$ and where $\rho$ and $\sigma$ are density operators. This bound suggests defining a "pretty good fidelity", whose relation to the usual fidelity implies the known relations between the optimal and pretty good measurement as well as the optimal and pretty good singlet fraction. In addition, we provide a new proof of the inequality $\widetilde{D}_{1}(\rho \| \sigma) \leqslant \widehat{D}_{1}(\rho \| \sigma)\, ,$ based on the Araki-Lieb-Thirring inequality. This leads to an elegant proof of the logarithmic form of the reverse Golden-Thompson inequality. | quantum physics |
Precision measurements of the Higgs mass have become a powerful constraint on models of physics beyond the standard model. We revisit supersymmetric models with Dirac gauginos and study the contributions to the Higgs mass. We calculate the leading two-loop corrections to the SM-like Higgs mass by constructing a series of EFTs and iteratively integrating out heavy particles. We then apply these calculations to a variety of scenarios, including a simple Dirac gluino, and split Dirac models of supersymmetry. We present the detailed formulae for threshold corrections and compare with previous results, where available. In general, the contributions are small, but the additional precision allows us to make more concrete statements about the relevant scales in Dirac SUSY models. | high energy physics phenomenology |
Quantifying the evolution of stellar extreme ultraviolet (EUV, 100 -- 1000 $\overset{\circ}{A}$) emission is critical for assessing the evolution of planetary atmospheres and the habitability of M dwarf systems. Previous studies from the HAbitable Zones and M dwarf Activity across Time (HAZMAT) program showed the far- and near-UV (FUV, NUV) emission from M stars at various stages of a stellar lifetime through photometric measurements from the Galaxy Evolution Explorer (GALEX). The results revealed increased levels of short-wavelength emission that remain elevated for hundreds of millions of years. The trend for EUV flux as a function of age could not be determined empirically because absorption by the interstellar medium prevents access to the EUV wavelengths for the vast majority of stars. In this paper, we model the evolution of EUV flux from early M stars to address this observational gap. We present synthetic spectra spanning EUV to infrared wavelengths of 0.4 $\pm$ 0.05 M$_{\odot}$ stars at five distinct ages between 10 and 5000 Myr, computed with the PHOENIX atmosphere code and guided by the GALEX photometry. We model a range of EUV fluxes spanning two orders of magnitude, consistent with the observed spread in X-ray, FUV, and NUV flux at each epoch. Our results show that the stellar EUV emission from young M stars is 100 times stronger than field age M stars, and decreases as t$^{-1}$ after remaining constant for a few hundred million years. This decline stems from changes in the chromospheric temperature structure, which steadily shifts outward with time. Our models reconstruct the full spectrally and temporally resolved history of an M star's UV radiation, including the unobservable EUV radiation, which drives planetary atmospheric escape, directly impacting a planet's potential for habitability. | astrophysics |
The inconsistent Hubble constant values derived from cosmic microwave background (CMB) observations and from local distance-ladder measurements may suggest new physics beyond the standard $\Lambda$CDM paradigm. It has been found in earlier works that, at least phenomenologically, non-standard recombination histories can reduce the $\gtrsim 4\sigma$ Hubble tension to $\sim 2\sigma$. Following this path, we vary physical and phenomenological parameters in RECFAST, the standard code to compute ionization history of the universe, to explore possible physics beyond standard recombination. We find that the CMB constraint on the Hubble constant is sensitive to the Hydrogen ionization energy and $2s \rightarrow 1s$ two-photon decay rate, both of which are atomic constants, and is insensitive to other details of recombination. Thus, the Hubble tension is very robust against perturbations of recombination history, unless exotic physics modifies the atomic constants during the recombination epoch. | astrophysics |
We study the early time and coarsening dynamics in the Light-Heavy model, a system consisting of two species of particles ($light$ and $heavy$) coupled to a fluctuating surface (described by tilt fields). The dynamics of particles and tilts are coupled through local update rules, and are known to lead to different ordered and disordered steady state phases depending on the microscopic rates. We introduce a generalized balance mechanism in non-equilibrium systems, namely $bunchwise~balance$, in which incoming and outgoing transition currents are balanced between groups of configurations. This allows us to exactly determine the steady state in a subspace of the phase diagram of this model. We introduce the concept of $irreducible~sequences$ of interfaces and bends in this model. These sequences are non-local, and we show that they provide a coarsening length scale in the ordered phases at late times. Finally, we propose a $local$ correlation function ($\mathcal{S}$) that has a direct relation to the number of irreducible sequences, and is able to distinguish between several phases of this system through its coarsening properties. Starting from a totally disordered initial configuration, $\mathcal{S}$ displays an initial linear rise and a broad maximum. As the system evolves towards the ordered steady states, $\mathcal{S}$ further exhibits power law decays at late times that encode coarsening properties of the approach to the ordered phases. Focusing on early time dynamics, we posit coupled mean-field evolution equations governing the particles and tilts, which at short times are well approximated by a set of linearized equations, which we solve analytically. Beyond a timescale set by a lattice cutoff and preceding the onset of coarsening, our linearized theory predicts the existence of an intermediate power-law stretch, which we also find in simulations of the ordered regime of the system. | condensed matter |
We present new variants of the Two Higgs-Doublet Model where all Yukawa couplings with physical Higgs bosons are controlled by the quark mixing matrices of both chiralities, as well as, in one case, the ratio between the two scalar doublets' vacuum expectation values. We obtain these by imposing approximate symmetries on the Lagrangian which, in one of the cases, clearly reveals the model to be the electroweak remnant of the Minimal Left-Right Symmetric Model. We also argue for the benefits of the bidoublet notation in the Two Higgs-Doublet Model context for uncovering new models. | high energy physics phenomenology |
The deformation and breakup of droplets in air flows is important in many applications of spray and atomization processes. However, the shear effect of airflow has never been reported. In this study, the deformation and breakup of droplets in the shear flow of air is investigated experimentally using high-speed imaging, digital image processing, and particle image velocimetry. We identify a new breakup mode of droplets, i.e., the butterfly breakup, in which the strong aerodynamic pressure on the lower part of the droplet leads to the deflection of the droplet and then the formation of a butterfly-shaped bag. A regime map of the droplet breakup is produced, and the transitions between different modes are obtained based on scaling analysis. The elongation and the fragmentation of the droplet rim are analyzed, and the results show that they are significantly affected by the shear via the formation and the growth of nodes on the rim. | physics |
Exploring the extent of model dependence, we study effects of certain Ansaetze for the T-dependence of the correction term in the QCD topological susceptibility. The one producing unwanted effects on results at T > 0 in the eta'-eta complex in the usual limit of isospin symmetry, is largely cured from its peculiar behavior and brought into agreement with the other, when breaking of isospin symmetry is allowed and the realistic quark mass ratio m_u/m_d = 0.50 is adopted. | high energy physics phenomenology |
Quantum Channel Discrimination (QCD) presents a fundamental task in quantum information theory, with critical applications in quantum reading, illumination, data-readout and more. The extension to multiple quantum channel discrimination has seen a recent focus to characterise potential quantum advantage associated with quantum enhanced discriminatory protocols. In this paper, we study thermal imaging as an environment localisation task, in which thermal images are modelled as ensembles of Gaussian phase insensitive channels with identical transmissivity, and pixels possess properties according to background (cold) or target (warm) thermal channels. Via the teleportation stretching of adaptive quantum protocols, we derive ultimate limits on the precision of pattern classification of abstract, binary thermal image spaces, and show that quantum enhanced strategies may be used to provide significant quantum advantage over known optimal classical strategies. The environmental conditions and necessary resources for which advantage may be obtained are studied and discussed. We then numerically investigate the use of quantum enhanced statistical classifiers, in which quantum sensors are used in conjunction with machine learning image classification methods. Proving definitive advantage in the low loss regime, this work motivates the use of quantum enhanced sources for short-range thermal imaging and detection techniques for future quantum technologies. | quantum physics |
Recently, \citet{vitral2021does} detected a central concentration of dark objects in the core-collapsed globular cluster NGC 6397, which could be interpreted as a subcluster of stellar-mass black holes. However, it is well established theoretically that any significant number of black holes in the cluster would provide strong dynamical heating and is fundamentally inconsistent with this cluster's core-collapsed profile. Claims of intermediate-mass black holes in core-collapsed clusters should similarly be treated with suspicion, for reasons that have been understood theoretically for many decades. Instead, the central dark population in NGC 6397 is exactly accounted for by a compact subsystem of white dwarfs, as we demonstrate here by inspection of a previously published model that provides a good fit to this cluster. These central subclusters of heavy white dwarfs are in fact a generic feature of core-collapsed clusters, while central black hole subclusters are present in all {\em non\/}-collapsed clusters. | astrophysics |
This article is motivated by studying the interaction of magnetic moments in high entropy alloys (HEAs), which plays an important role in guiding HEA designs in materials science. While first principles simulations can capture magnetic moments of individual atoms, explicit models are required to analyze their interactions. This is essentially an inverse covariance matrix estimation problem. Unlike most of the literature on graphical models, the inverse covariance matrix possesses inherent structural constraints encompassing node types, topological distance of nodes, and partially specified conditional dependence patterns. The present article is, to our knowledge, the first to consider such intricate structures in graphical models. In particular, we introduce graph constraints to formulate these structures that arise from domain knowledge and are critical for interpretability, which leads to a Bayesian conditional autoregressive model with graph constraints (CARGO) for structured inverse covariance matrices. The CARGO method enjoys efficient implementation with a modified Gauss-Seidel scheme through proximity operators for closed-form posterior exploration. We establish algorithmic convergence for the proposed algorithm under a verifiable stopping criterion. Simulations show competitive performance of CARGO relative to several other methods and confirm our convergence analysis. In a novel real data application to HEAs, the proposed methods lead to data-driven quantification and interpretation of magnetic moment interactions with high tractability and transferability. | statistics |
A potential motion of ideal incompressible fluid with a free surface and infinite depth is considered in two-dimensional geometry. A time-dependent conformal mapping of the lower complex half-plane of the auxiliary complex variable $w$ into the area filled with fluid is performed with the real line of $w$ mapped into the free fluid's surface. The fluid dynamics can be fully characterized by the motion of the complex singularities in the analytical continuation of both the conformal mapping and the complex velocity. We consider the short branch cut approximation of the dynamics with the small parameter being the ratio of the length of the branch cut to the distance between its center and the real line of $w$. We found that the fluid dynamics in that approximation is reduced to the complex Hopf equation for the complex velocity coupled with the complex transport equation for the conformal mapping. These equations are fully integrable by characteristics producing the infinite family of solutions, including the pairs of moving square root branch points. The solutions are compared with the simulations of the fully nonlinear Eulerian dynamics giving excellent agreement even when the small parameter approaches about one. | physics |
We prove that any function in $GSBD^p(\Omega)$, with $\Omega$ a $n$-dimensional open bounded set with finite perimeter, is approximated by functions $u_k\in SBV(\Omega;\mathbb{R}^n)\cap L^\infty(\Omega;\mathbb{R}^n)$ whose jump is a finite union of $C^1$ hypersurfaces. The approximation takes place in the sense of Griffith-type energies $\int_\Omega W(e(u)) \,\mathrm{d}x +\mathcal{H}^{n-1}(J_u)$, $e(u)$ and $J_u$ being the approximate symmetric gradient and the jump set of $u$, and $W$ a nonnegative function with $p$-growth, $p>1$. The difference between $u_k$ and $u$ is small in $L^p$ outside a sequence of sets $E_k\subset \Omega$ whose measure tends to 0 and if $|u|^r \in L^1(\Omega)$ with $r\in (0,p]$, then $|u_k-u|^r \to 0$ in $L^1(\Omega)$. Moreover, an approximation property for the (truncation of the) amplitude of the jump holds. We apply the density result to deduce $\Gamma$-convergence approximation \emph{\`a la} Ambrosio-Tortorelli for Griffith-type energies with either Dirichlet boundary condition or a mild fidelity term, such that minimisers are \emph{a priori} not even in $L^1(\Omega;\mathbb{R}^n)$. | mathematics |
We propose a new method to compute asymptotics of periods using tropical geometry, in which the Riemann zeta values appear naturally as error terms in tropicalization. Our method suggests how the Gamma class should arise from the Strominger-Yau-Zaslow conjecture. We use it to give a new proof of (a version of) the Gamma Conjecture for Batyrev pairs of mirror Calabi-Yau hypersurfaces. | mathematics |
We construct a ring homomorphism comparing the tautological ring, fixing a point, of a closed smooth manifold with that of its stabilisation by $S^{2a} \times S^{2b}$. | mathematics |
Undoubtedly, one of the most important issues in computer science is intelligent speech recognition. In these systems, computers try to detect and respond to the speeches they are listening to, like humans. In this research, presenting of a suitable method for the diagnosis of Persian phonemes by AI using the signal processing and classification algorithms have tried. For this purpose, the STFT algorithm has been used to process the audio signals, as well as to detect and classify the signals processed by the deep artificial neural network. At first, educational samples were provided as two phonological phrases in Persian language and then signal processing operations were performed on them. Then the results for the data training have been given to the artificial deep neural network. At the final stage, the experiment was conducted on new sounds. | computer science |
As a necessary step in constructing elliptic matrix models, which preserve the superintegrability property $<char>\sim {\rm char}$, we suggest an elliptic deformation of the peculiar loci $p_k^{\Delta_n}$, which play an important role in precise formulation of this property. The suggestion is to define the $p_k^{\Delta_n}$-loci as elliptic functions with the right asymptotics at $\tau\to i\infty$. If this hypothesis is correct, one can move to substituting the Schur and Macdonald functions in the role of characters by the elliptic GNS and study their behavior at the deformed topological and $\Delta$ loci. | high energy physics theory |
In an abstract sense, quantum data hiding is the manifestation of the fact that two classes of quantum measurements can perform very differently in the task of binary quantum state discrimination. We investigate this phenomenon in the context of continuous variable quantum systems. First, we look at the celebrated case of data hiding 'against' the set of local operations and classical communication. While previous studies have placed upper bounds on its maximum efficiency in terms of the local dimension and are thus not applicable to continuous variable systems, we tackle this latter case by establishing more general bounds that rely solely on the local mean photon number of the states employed. Along the way, we perform a quantitative analysis of the error introduced by the non-ideal Braunstein--Kimble quantum teleportation protocol, determining how much two-mode squeezing and local detection efficiency is needed in order to teleport an arbitrary local state of known mean energy with a prescribed accuracy. Finally, following a seminal proposal by Winter, we look at data hiding against the set of Gaussian operations and classical computation, providing the first example of a relatively simple scheme that works with a single mode only. The states employed can be generated from a two-mode squeezed vacuum by local photon counting; the larger the squeezing, the higher the efficiency of the scheme. | quantum physics |
We extend the dictionary between the BPS spectrum of Heterotic strings and the one of F-/M-theory compactifications on $K3$ fibered Calabi-Yau 3-folds to cases with higher rank non-Abelian gauge groups and in particular to dual pairs between Heterotic CHL orbifolds and compactifications on Calabi-Yau 3-folds with a compatible genus one fibration. We show how to obtain the new supersymmetric index purely from the Calabi-Yau geometry by reconstructing the Noether-Lefschetz generators, which are vector-valued modular forms. There is an isomorphism between the latter objects and vector-valued lattice Jacobi forms, which relates them to the elliptic genera and twisted-twined elliptic genera of six- and five-dimensional Heterotic strings. The meromorphic Jacobi forms generate the dimensions of the refined cohomology of the Hilbert schemes of symmetric products of the fiber and allow us to refine the BPS indices in the fiber and therefore to obtain, conjecturally, actual state counts. Using the properties of the vector-valued lattice Jacobi forms we also provide a mathematical proof of the non-Abelian weak gravity conjecture for F-/M-theory compactified on this general class of fibered Calabi-Yau 3-folds. | high energy physics theory |
A dynamical quantum phase transition can occur during time evolution of sudden quenched quantum systems across a phase transition. It corresponds to the nonanalytic behavior at a critical time of the rate function of the quantum state return amplitude, analogous to nonanalyticity of the free energy density at the critical temperature in macroscopic systems. A variety of many-body systems can be represented in momentum space as a spin-1/2 state evolving on the Bloch sphere, where each momentum mode is decoupled and thus can be simulated independently by a single qubit. Here, we report the observation of a dynamical quantum phase transition in a superconducting qubit simulation of the quantum quench dynamics of many-body systems. We take the Ising model with a transverse field as an example for demonstration. In our experiment, the spin state, which is initially polarized longitudinally, evolves based on a Hamiltonian with adjustable parameters depending on the momentum and strength of the transverse magnetic field. The time evolving quantum state is read out by state tomography. Evidence of dynamical quantum phase transitions, such as paths of time evolution states on the Bloch sphere, non-analytic behavior of the dynamical free energy and the emergence of Skyrmion lattice in momentum-time space, is observed. The experimental data agrees well with theoretical and numerical calculations. The experiment demonstrates for the first time explicitly the topological invariant, both topologically trivial and non-trivial, for dynamical quantum phase transitions. Our results show that the quantum phase transitions of this class of many-body systems can be simulated successfully with a single qubit by varying certain control parameters over the corresponding momentum range. | quantum physics |
High critical temperature superconductivity often occurs in systems where an antiferromagnetic order is brought near $T=0K$ by slightly modifying pressure or doping. CaKFe$_4$As$_4$ is a superconducting, stoichiometric iron pnictide compound showing optimal superconducting critical temperature with $T_c$ as large as $38$ K. Doping with Ni induces a decrease in $T_c$ and the onset of spin-vortex antiferromagnetic order, which consists of spins pointing inwards to or outwards from alternating As sites on the diagonals of the in-plane square Fe lattice. Here we study the band structure of CaK(Fe$_{0.95}$Ni$_{0.05}$)$_4$As$_4$ (T$_c$ = 10 K, T$_N$ = 50 K) using quasiparticle interference with a Scanning Tunneling Microscope (STM) and show that the spin-vortex order induces a Fermi surface reconstruction and a fourfold superconducting gap anisotropy. | condensed matter |
This paper is dedicated to the "50 Years of the Relevance Problem" - a long-neglected topic that begs attention from practical statisticians who are concerned with the problem of drawing inference from large-scale heterogeneous data. | statistics |
Context: Code smells (CS) tend to compromise software quality and also demand more effort by developers to maintain and evolve the application throughout its life-cycle. They have long been catalogued with corresponding mitigating solutions called refactoring operations. Objective: This SLR has a twofold goal: the first is to identify the main code smells detection techniques and tools discussed in the literature, and the second is to analyze to which extent visual techniques have been applied to support the former. Method: Over 83 primary studies indexed in major scientific repositories were identified by our search string in this SLR. Then, following existing best practices for secondary studies, we applied inclusion/exclusion criteria to select the most relevant works, extract their features and classify them. Results: We found that the most commonly used approaches to code smells detection are search-based (30.1%), and metric-based (24.1%). Most of the studies (83.1%) use open-source software, with the Java language occupying the first position (77.1%). In terms of code smells, God Class (51.8%), Feature Envy (33.7%), and Long Method (26.5%) are the most covered ones. Machine learning techniques are used in 35% of the studies. Around 80% of the studies only detect code smells, without providing visualization techniques. In visualization-based approaches several methods are used, such as: city metaphors, 3D visualization techniques. Conclusions: We confirm that the detection of CS is a non trivial task, and there is still a lot of work to be done in terms of: reducing the subjectivity associated with the definition and detection of CS; increasing the diversity of detected CS and of supported programming languages; constructing and sharing oracles and datasets to facilitate the replication of CS detection and visualization techniques validation experiments. | computer science |
Spatial documentation is exponentially increasing given the availability of Big IoT Data, enabled by the devices miniaturization and data storage capacity. Bayesian spatial statistics is a useful statistical tool to determine the dependence structure and hidden patterns over space through prior knowledge and data likelihood. Nevertheless, this modeling class is not well explored as the classification and regression machine learning models given their simplicity and often weak (data) independence supposition. In this manner, this systematic review aimed to unravel the main models presented in the literature in the past 20 years, identify gaps, and research opportunities. Elements such as random fields, spatial domains, prior specification, covariance function, and numerical approximations were discussed. This work explored the two subclasses of spatial smoothing global and local. | statistics |
We report on our realization of a high power holmium doped fiber laser, together with the validation of our numerical simulation of the laser. We first present the rare absolute measurements of the physical parameters that are mandatory to model accurately the laser-holmium interactions in our silica fiber. We then describe the realization of the clad-pumped laser, based on a triple-clad large mode area holmium (Ho) doped silica fiber. The output signal power is 90 W at 2120 nm, with an efficiency of about 50% with respect to the coupled pump power. This efficiency corresponds to the state of the art for clad-pumped Ho-doped fiber lasers in the 100 W power class. By comparing the experimental results to our simulation, we demonstrate its validity, and use it to show that the efficiency is limited, for our fiber, by the non-saturable absorption caused by pair induced quenching between adjacent holmium ions. | physics |
We give sufficient conditions for Input-to-State Stability in $C^{1}$ norm of general quasilinear hyperbolic systems with boundary input disturbances. In particular the derivation of explicit Input-to-State Stability conditions is discussed for the special case of $2\times 2$ systems. | mathematics |
In this paper, it is shown that in interaction with an oscillating bubbles cluster the fluid becomes inhomogeneous. The radial variation of the acoustic refractive index of the fluid generates an acoustic lens with spherical symmetry. When the focal length of the lens associated with the bubbles cluster is equal to the length of the position vector then the bubbles cluster behaves like a dumb hole. Unlike the properties of a bubble, the cluster, by compression under the action of attractive internal forces, can shrink its radius to become a dumb hole. | physics |
This paper proposes a framework for safe reinforcement learning that can handle stochastic nonlinear dynamical systems. We focus on the setting where the nominal dynamics are known, and are subject to additive stochastic disturbances with known distribution. Our goal is to ensure the safety of a control policy trained using reinforcement learning, e.g., in a simulated environment. We build on the idea of model predictive shielding (MPS), where a backup controller is used to override the learned policy as needed to ensure safety. The key challenge is how to compute a backup policy in the context of stochastic dynamics. We propose to use a tube-based robust NMPC controller as the backup controller. We estimate the tubes using sampled trajectories, leveraging ideas from statistical learning theory to obtain high-probability guarantees. We empirically demonstrate that our approach can ensure safety in stochastic systems, including cart-pole and a non-holonomic particle with random obstacles. | electrical engineering and systems science |
This paper explores the problem of clustering ensemble, which aims to combine multiple base clusterings to produce better performance than that of the individual one. The existing clustering ensemble methods generally construct a co-association matrix, which indicates the pairwise similarity between samples, as the weighted linear combination of the connective matrices from different base clusterings, and the resulting co-association matrix is then adopted as the input of an off-the-shelf clustering algorithm, e.g., spectral clustering. However, the co-association matrix may be dominated by poor base clusterings, resulting in inferior performance. In this paper, we propose a novel low-rank tensor approximation-based method to solve the problem from a global perspective. Specifically, by inspecting whether two samples are clustered to an identical cluster under different base clusterings, we derive a coherent-link matrix, which contains limited but highly reliable relationships between samples. We then stack the coherent-link matrix and the co-association matrix to form a three-dimensional tensor, the low-rankness property of which is further explored to propagate the information of the coherent-link matrix to the co-association matrix, producing a refined co-association matrix. We formulate the proposed method as a convex constrained optimization problem and solve it efficiently. Experimental results over 7 benchmark data sets show that the proposed model achieves a breakthrough in clustering performance, compared with 12 state-of-the-art methods. To the best of our knowledge, this is the first work to explore the potential of low-rank tensor on clustering ensemble, which is fundamentally different from previous approaches. | computer science |
Invoking increasingly higher dimension operators to encode novel UV physics in effective gauge and gravity theories traditionally means working with increasingly more finicky and difficult expressions. We demonstrate that local higher derivative supersymmetric-compatible operators at four-points can be absorbed into simpler higher-derivative corrections to scalar theories, which generate the predictions of Yang-Mills and Gravity operators by suitable replacements of color-weights with color-dual kinematic weights as per Bern-Carrasco-Johansson double-copy. We exploit that Jacobi-satisfying representations can be composed out of other Jacobi-satisfying representations, and show that at four-points only a small number of building blocks are required to generate the predictions of higher-derivative operators. We find that this construction saturates the higher-derivative operators contributing to the four-point supersymmetric open and closed-string tree amplitudes, presenting a novel representation of the four-point supersymmetric open string making this structure manifest, as well as identifying the only four additional gauge-invariant building blocks required to saturate the four-point bosonic open string. | high energy physics theory |
This paper formalize the existence's proof of first-integrals for any second order ODE, allowing to discriminate periodic orbits. Up to the author's knowledge, such a powerful result is not available in the literature providing a tool to determine periodic orbits/limit cycles in the most general scenario. | mathematics |
The Third Generation Partnership Project (3GPP) has recently published its Release 16 that includes the first Vehicle to-Everything (V2X) standard based on the 5G New Radio (NR) air interface. 5G NR V2X introduces advanced functionalities on top of the 5G NR air interface to support connected and automated driving use cases with stringent requirements. This paper presents an in-depth tutorial of the 3GPP Release 16 5G NR V2X standard for V2X communications, with a particular focus on the sidelink, since it is the most significant part of 5G NR V2X. The main part of the paper is an in-depth treatment of the key aspects of 5G NR V2X: the physical layer, the resource allocation, the quality of service management, the enhancements introduced to the Uu interface and the mobility management for V2N (Vehicle to Network) communications, as well as the co-existence mechanisms between 5G NR V2X and LTE V2X. We also review the use cases, the system architecture, and describe the evaluation methodology and simulation assumptions for 5G NR V2X. Finally, we provide an outlook on possible 5G NR V2X enhancements, including those identified within Release 17. | electrical engineering and systems science |
We theoretically study the quantum-coherence-enhanced subradiance in a chiral-coupled atomic chain with nonreciprocal decay channels. The collective radiation in this one-dimensional (1D) nanophotonics system results from the resonant dipole-dipole interactions (RDDI) in 1D reservoirs, which allow infinite-range couplings between atoms. When single photon interacts with part of the atomic chain from a side excitation, the subradiant decay can be further reduced when highly correlated states are initially excited. The excitation plateau in the decay process can emerge due to the ordered population exchanges, which presents one distinctive signature of long-range and light-induced atom-atom correlations. Multiple time scales of the decay behaviors also show up due to multiple scattering of light transmissions and reflections in the chain. We further investigate the effect of atomic position fluctuations, and find that the cascaded scheme with uni-directional coupling is more resilient to the fluctuations, while the overall decay constant can be increased due to large deviations. Our results present a fundamental study on the subradiance and light-induced atom-atom correlations in such 1D nanophotonics platforms, and offer rich opportunities in potential applications of quantum storage of photons. | quantum physics |
Under shear, a system of particles changes its contact network and becomes unstable as it transitions between mechanically stable states. For hard spheres at zero pressure, contact breaking events necessarily generate an instability, but this is not the case at finite pressure, where we identify two types of contact changes: network events that do not correspond to instabilities and rearrangement events that do. The relative fraction of such events is constant as a function of system size, pressure and interaction potential, consistent with our observation that both nonlinearities obey the same finite-size scaling. Thus, the zero-pressure limit of the nonlinear response is highly singular. | condensed matter |
We calculate the next-to-leading order (NLO) quantum chromodynamics (QCD) correction to the exclusive process $\gamma+\gamma\to J/\psi+J/\psi$ in the framework of non-relativistic QCD (NRQCD) factorization formalism. The cross sections at the SuperKEKB electron-positron collider, as well as at the future colliders, like the Circular Electron Positron Collider (CEPC) and the International Linear Collider (ILC), are evaluated. Numerical result indicates that the process will be hopefully to be seen by the Belle II detector within the next decade. | high energy physics phenomenology |
We show the simple Hurwitz space $\mathcal{H}_{g,d}$ has trivial rational Picard group for $d>g-1$ and is uniruled for $d>g+1$. | mathematics |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.