text
stringlengths
11
9.77k
label
stringlengths
2
104
In analogy to the development of fluorescent proteins, innovative tools for screening optoacoustic cell labels could lead to tailored protein labels for OA, imparting novel ways to visualize biological structure and function. Optoacoustic imaging emerges towards a highly promising modality for life sciences and medical practise with advantageous capabilities such as great accessible depth, and 3D studying of living tissue. The development of novel labels with molecular specificity could significantly enhance the optoacoustic contrast, specificity, and sensitivity and allow optoacoustic to interrogate tissues not amenable to the fluorescence method. We report on an optoacoustic flow cytometer (OAFC) prototype, developed for screening optoacoustic reporter genes. The cytometer concurrently records light scattering for referencing purposes. Since recording light scattering is completely independent from OA, we believe it to be a more reliable referencing method than e.g. fluorescence or ultrasound-backscatter. Precise characterization of our OAFC prototype showcases its ability to optoacoustically characterize objects in-flow that are in the size range of single cells. We apply the OAFC to distinguish individual E. coli cells based on optoacoustic properties of their expressed chromoproteins read in-flow using microfluidic arrangements and achieved precisions over 90%. We discuss how the light scattering referenced OAFC method offers a critical step towards routine measurement of optoacoustic properties of single-cells and could pave the way for identifying genetically encoded optoacoustic reporters, by transferring working concepts of the fluorescence field.
physics
Synthetic Magnetic Resonance (MR) imaging predicts images at new design parameter settings from a few observed MR scans. Model-based methods, that use both the physical and statistical properties underlying the MR signal and its acquisition, can predict images at any setting from as few as three scans, allowing it to be used in individualized patient- and anatomy-specific contexts. However, the estimation problem in model-based synthetic MR imaging is ill-posed and so regularization, in the form of correlated Gaussian Markov Random Fields, is imposed on the voxel-wise spin-lattice relaxation time, spin-spin relaxation time and the proton density underlying the MR image. We develop theoretically sound but computationally practical matrix-free estimation methods for synthetic MR imaging. Our evaluations demonstrate excellent ability of our methods to synthetize MR images in a clinical framework and also estimation and prediction accuracy and consistency. An added strength of our model-based approach, also developed and illustrated here, is the accurate estimation of standard errors of regional means in the synthesized images.
statistics
51 Oph is a Herbig Ae/Be star that exhibits strong near-infrared CO ro-vibrational emission at 2.3 micron, most likely originating in the innermost regions of a circumstellar disc. We aim to obtain the physical and geometrical properties of the system by spatially resolving the circumstellar environment of the inner gaseous disc. We used the second-generation VLTI/GRAVITY to spatially resolve the continuum and the CO overtone emission. We obtained data over 12 baselines with the auxiliary telescopes and derive visibilities, and the differential and closure phases as a function of wavelength. We used a simple LTE ring model of the CO emission to reproduce the spectrum and CO line displacements. Our interferometric data show that the star is marginally resolved at our spatial resolution, with a radius of 10.58+-2.65 Rsun.The K-band continuum emission from the disc is inclined by 63+-1 deg, with a position angle of 116+-1 deg, and 4+-0.8 mas (0.5+-0.1 au) across. The visibilities increase within the CO line emission, indicating that the CO is emitted within the dust-sublimation radius.By modelling the CO bandhead spectrum, we derive that the CO is emitted from a hot (T=1900-2800 K) and dense (NCO=(0.9-9)x10^21 cm^-2) gas. The analysis of the CO line displacement with respect to the continuum allows us to infer that the CO is emitted from a region 0.10+-0.02 au across, well within the dust-sublimation radius. The inclination and position angle of the CO line emitting region is consistent with that of the dusty disc. Our spatially resolved interferometric observations confirm the CO ro-vibrational emission within the dust-free region of the inner disc. Conventional disc models exclude the presence of CO in the dust-depleted regions of Herbig AeBe stars. Ad hoc models of the innermost disc regions, that can compute the properties of the dust-free inner disc, are therefore required.
astrophysics
We investigate a learning algorithm in the context of nominal automata, an extension of classical automata to alphabets featuring names. This class of automata captures nominal regular languages; analogously to the classical language theory, nominal automata have been shown to characterise nominal regular expressions with binders. These formalisms are amenable to abstract modelling resource-aware computations. We propose a learning algorithm on nominal regular languages with binders. Our algorithm generalises Angluin's L* algorithm with respect to nominal regular languages with binders. We show the correctness and study the theoretical complexity of our algorithm.
computer science
SMUTHI is a python package for the efficient and accurate simulation of electromagnetic scattering by one or multiple wavelength-scale objects in a planarly layered medium. The software combines the T-matrix method for individual particle scattering with the scattering matrix formalism for the propagation of the electromagnetic field through the planar interfaces. In this article, we briefly introduce the relevant theoretical concepts and present the main features of SMUTHI. Simulation results obtained for several benchmark configurations are validated against commercial software solutions. Owing to the generality of planarly layered geometries and the availability of different particle shapes and light sources, possible applications of SMUTHI include the study of discrete random media, meta-surfaces, photonic crystals and glasses, perforated membranes and plasmonic systems, to name a few relevant examples at visible and near-visible wavelengths.
physics
Activation of quantum capacity is a surprising phenomenon according to which the quantum capacity of a certain channel may increase by combining it with another channel with zero quantum capacity. Superactivation describes an even more particular occurrence, in which both channels have zero quantum capacity, but their composition has a nonvanishing one. We investigate these effects for all single-mode phase-insensitive Gaussian channels, which include thermal attenuators and amplifiers, assisted by a two-mode positive-partial-transpose channel. Our result shows that activation phenomena are special but not uncommon. We can reveal superactivation in a broad range of thermal attenuator channels, even when the transmissivity is quite low. This means that we can transmit quantum information reliably through very noisy Gaussian channels having zero quantum capacity. We further show that no superactivation is possible for entanglement-breaking Gaussian channels in physically relevant circumstances by proving the non-activation property of the coherent information of bosonic entanglement-breaking channels with finite input energy.
quantum physics
A recent work [Y. Huang and B.-Q. Ma, Commun. Phys. {\bf 1}, 62 (2018)] associated all four PeV neutrinos observed by IceCube to gamma-ray bursts (GRBs), and revealed a regularity which indicates a Lorentz violation scale $E_{\rm LV}=(6.5\pm0.4)\times10^{17}$ GeV with opposite sign factors $s=\pm 1$ between neutrinos and antineutrinos. The association of "time delay" and "time advance" events with neutrinos and antineutrinos (or vice versa) is only a hypothesis since the IceCube detector cannot tell the chirality of the neutrinos, and further experimental tests are needed to verify this hypothesis. We derive the values of the CPT-odd Lorentz violating parameters in the standard-model extension (SME) framework, and perform a threshold analysis on the electron-positron pair emission of the superluminal neutrinos (or antineutrinos). We find that different neutrino/antineutrino propagation properties, suggested by Y. Huang and B.-Q. Ma, can be described in the SME framework with both Lorentz invariance and CPT symmetry violation, but with a threshold energy constraint. A viable way on testing the CPT symmetry violation between neutrinos and antineutrinos is suggested.
high energy physics phenomenology
We develop a theory of classical complexity. We study the relations between classical complexity and entropy, and conjecture that in an isolated system, classical absolute complexity always tends to grow, until it reaches its maximum. We calculate some exact closed-form expressions of the growth of average classical complexity over time in some concrete models, and gain further insights of both classical and quantum complexity by using the theory of Markov chains.
high energy physics theory
In this paper the exact linear relation between the leading eigenvector of the unnormalized modularity matrix and the eigenvectors of the adjacency matrix is developed. Based on this analysis a method to approximate the leading eigenvector of the modularity matrix is given, and the relative error of the approximation is derived. A complete proof of the equivalence between normalized modularity clustering and normalized adjacency clustering is also given. Some applications and experiments are given to illustrate and corroborate the points that are made in the theoretical development.
statistics
We study the four-point function of the lowest-lying half-BPS operators in the ${\cal N} =4$ $SU(N)$ super-Yang-Mills theory and its relation to the flat-space four-graviton amplitude in type IIB superstring theory. We work in a large-$N$ expansion in which the complexified Yang-Mills coupling $\tau$ is fixed. In this expansion, non-perturbative instanton contributions are present, and the $SL(2, \mathbb{Z})$ duality invariance of correlation functions is manifest. Our results are based on a detailed analysis of the sphere partition function of the mass-deformed SYM theory, which was previously computed using supersymmetric localization. This partition function determines a certain integrated correlator in the undeformed ${\cal N} = 4$ SYM theory, which in turn constrains the four-point correlator at separated points. In a normalization where the two-point functions are proportional to $N^2-1$ and are independent of $\tau$ and $\bar \tau$, we find that the terms of order $\sqrt{N}$ and $1/\sqrt{N}$ in the large $N$ expansion of the four-point correlator are proportional to the non-holomorphic Eisenstein series $E({\scriptstyle \frac{3}{2}},\tau,\bar\tau)$ and $E({\scriptstyle \frac{5}{2}},\tau,\bar\tau)$, respectively. In the flat space limit, these terms match the corresponding terms in the type IIB S-matrix arising from $R^4$ and $D^4 R^4$ contact interactions, which, for the $R^4$ case, represents a check of AdS/CFT at finite string coupling. Furthermore, we present striking evidence that these results generalize so that, at order $N^{\frac{1}{2}-m}$ with integer $m \ge 0$, the expansion of the integrated correlator we study is a linear sum of non-holomorphic Eisenstein series with half-integer index, which are manifestly $SL(2,\mathbb{Z})$ invariant.
high energy physics theory
Analyzing the properties of dust and its evolution in the early phases of star formation is crucial to put constraints on the collapse and accretion processes as well as on the pristine properties of planet-forming seeds. We analyze PdBI observations at 1.3 and 3.2 mm for 12 Class 0 protostars obtained as part of the CALYPSO survey and derive dust emissivity index (beta) profiles as a function of the envelope radius at 200-2000 au scales. Most of the protostellar envelopes show low dust emissivity indices decreasing toward the central regions. The decreasing trend remains after correction of the (potentially optically-thick) central region emission, with surprisingly low beta (lower than 1) values across most of the envelope radii of NGC1333-IRAS4A, NGC1333-IRAS4B, SVS13B, and Serpens-SMM4. We discuss the various processes that could explain such low and varying dust emissivity indices at envelope radii 200-2000 au. Our observations of extremely low dust emissivity indices could trace the presence of large (mm-size) grains in Class 0 envelopes and a radial increase of the dust grain size toward the inner envelope regions. While it is expected that large grains in young protostellar envelopes could be built via grain growth and coagulation, we stress that the typical timescales required to build mm grains in current coagulation models are at odds with the youth of our Class 0 protostars. Additional variations in the dust composition could also partly contribute to the low beta we observe. The steepness of the beta radial gradient depends strongly on the envelope mass, which might favor a scenario where large grains are built in high density protostellar disks and transported to the intermediate envelope radii, for example with the help of outflows and winds.
astrophysics
We study the ring extensions R \subseteq T having the same set of prime ideals provided Nil(R) is a divided prime ideal. Some conditions are given under which no such T exist properly containing R. Using idealization theory, the examples are also discussed to strengthen the results.
mathematics
As a green and secure wireless transmission way, secure spatial modulation (SM) is becoming a hot research area. Its basic idea is to exploit both the index of activated transmit antenna and amplitude phase modulation (APM) signal to carry messages, improve security, and save energy. In this paper, we reviewed its crucial techniques: transmit antenna selection (TAS), artificial noise (AN) projection, power allocation (PA), and joint detection at desired receiver. To achieve the optimal performance of maximum likelihood (ML) detector, a deep-neural-network (DNN) joint detector is proposed to jointly infer the index of transmit antenna and signal constellation point with a lower-complexity. Here, each layer of DNN is redesigned to optimize the joint inference performance of two distinct types of information: transmit antenna index and signal constellation point. Simulation results show that the proposed DNN method performs 3dB better than the conventional DNN structure and is close to ML detection in the low and medium signal-to-noise ratio regions in terms of the bit error rate (BER) performance, but its complexity is far lower-complexity compared to ML. Finally, three key techniques TAS, PA, and AN projection at transmitter can be combined to make SM a true secure modulation.
electrical engineering and systems science
We present a general formalism to write the decay amplitude for multibody reactions with explicit separation of the rotational degrees of freedom, which are well controlled by the spin of the decay particle, and dynamic functions on the subchannel invariant masses, which require modeling. Using the three-particle kinematics we demonstrate the proposed factorization, named the Dalitz-plot decomposition. The Wigner rotations, that are subtle factors needed by the isobar modeling in the helicity framework, are simplified with the proposed decomposition. Consequently, we are able to provide them in an explicit form suitable for the general case of arbitrary spins. The only unknown model-dependent factors are the isobar lineshapes that describe the subchannel dynamics. The advantages of the new decomposition are shown through three examples relevant for the recent discovery of the exotic charmonium candidate $Z_c(4430)$, the pentaquarks $P_c$, and the intriguing $\Lambda_c^+\to pK^-\pi^+$ decay.
high energy physics phenomenology
We define a new concept of "mistake" strategies and actions for strategic-form and extensive-form games, analyze the relationship to prior main game-theoretic solution concepts, study algorithms for computation, and explore practicality. This concept has potential applications to cybersecurity, for example detecting whether a human player is illegally using real-time assistance in games like poker.
computer science
A scalp-recording electroencephalography (EEG)-based brain-computer interface (BCI) system can greatly improve the quality of life for people who suffer from motor disabilities. Deep neural networks consisting of multiple convolutional, LSTM and fully-connected layers are created to decode EEG signals to maximize the human intention recognition accuracy. However, prior FPGA, ASIC, ReRAM and photonic accelerators cannot maintain sufficient battery lifetime when processing real-time intention recognition. In this paper, we propose an ultra-low-power photonic accelerator, MindReading, for human intention recognition by only low bit-width addition and shift operations. Compared to prior neural network accelerators, to maintain the real-time processing throughput, MindReading reduces the power consumption by 62.7\% and improves the throughput per Watt by 168\%.
electrical engineering and systems science
We compute next-to-leading order virtual two-loop corrections to the process $gg\to ZZ$ in the low- and high-energy limits, considering the contributions with virtual top quarks. Analytic results for all 20 form factors are presented including expansion terms up to $1/m_t^{12}$ and $m_t^{32}$. We use a Pad\'e approximation procedure to extend the radius of convergence of the high-energy expansion and apply this approach to the fini\ te virtual next-to-leading order corrections.
high energy physics phenomenology
Transmission line failures in power systems propagate and cascade non-locally. In this work, we propose a fast-timescale distributed control strategy that offers strong guarantees in both the mitigation and localization of line failures. Specifically, we leverage the properties of network bridge-block decomposition and a frequency regulation method called the unified control. If the balancing areas over which the unified control operates coincide with the bridge-blocks of the network, the proposed strategy drives the post-contingency system to a steady state where the impact of initial line outages is localized within the areas where they occurred whenever possible, stopping the cascading process. When the initial line outages cannot be localized, the proposed control strategy provides a configurable design that progressively involves and coordinates more balancing areas. We compare the proposed control strategy with the classical Automatic Generation Control (AGC) on the IEEE 118-bus and 2736-bus test networks. Simulation results show that our strategy greatly improves overall reliability in terms of the $N-k$ security standard, and localizes the impact of initial failures in majority of the simulated contingencies. Moreover, the proposed framework incurs significantly less load loss, if any, compared to AGC, in all our case studies.
electrical engineering and systems science
In the paper, we study the generalized Crewther Relation (GCR) between the Adler function ($D$) and Gross-Llewellyn Smith sum rule for the polarized deep-inelastic electron scattering ($C^{\rm GLS}$) at the leading twist by using the single-scale approach of the principle of maximum conformality (PMC). The PMC adopts the renormalization group equation to set the correct running behavior of the strong coupling constant for a pQCD approximant. After applying the PMC, the perturbative series becomes scale independent, and a precise fixed-order pQCD prediction can be achieved. We observe that the PMC conformal series also leads to a scheme-independent GCR, or equivalently the commensurate scale relation (CSR) between the effective charges of $D$ and $C^{\rm GLS}$. Similarly, we can obtain a CSR between the effective charges of $C^{\rm GLS}$ and the $R$ ratio of $e^+e^-$ annihilation cross section, which is also scheme-independent and allows us to fix the effective charge of $C^{\rm GLS}$ from the $R$-ratio experimental data. Thus by using the PMC single-scale approach, a precise test of QCD theory without renormalization scheme and scale ambiguities can be achieved.
high energy physics phenomenology
Specific low-bitrate coding strategies are examined through their effect on LQ control performance. By limiting the subject to these methods, we are able to identify principles underlying coding for control; a subject of significant recent interest but few tangible results. In particular, we consider coding the quantized output signal deploying period-two codes of differing delay-versus-accuracy tradeoff. The quantification of coding performance is via the LQ control cost. The feedback control system comprises the coder-decoder in the path between the output and the state estimator, which is followed by linear state-variable feedback, as is optimal in the memoryless case. The quantizer is treated as the functional composition of an infinitely-long linear staircase function and a saturation. This permits the analysis to subdivide into estimator computations, seemingly independent of the control performance criterion, and an escape time evaluation, which ties the control back into the choice of quantizer saturation bound. An example is studied which illustrates the role of the control objective in determining the efficacy of coding using these schemes. The results mesh well with those observed in signal coding. However, the introduction of a realization-based escape time is a novelty departing significantly from mean square computations.
mathematics
It is common in the study of a dizzying array of soft matter systems to perform agent-based simulations of particles interacting via conservative and often short-ranged forces. In this context, well-established algorithms for efficiently computing the set of pairs of interacting particles have established excellent open-source packages to efficiently simulate large systems over long time scales -- a crucial consideration given the separation in time- and length-scales often observed in soft matter. What happens, though, when we think more broadly about what it means to construct a neighbor list? What if interactions are non-reciprocal, or if the "range" of an interaction is determined not by a distance scale but according to some other consideration? As the field of soft and active matter increasingly considers the properties of living matter -- from the cellular to the super-organismal scale -- these questions become increasingly relevant, and encourage us to think about new physical and computational paradigms in the modeling of active matter. In this chapter we examine case studies in the use of non-metric interactions.
condensed matter
Real-world complex systems always interact with each other, which causes these systems to collapse in an avalanche or cascading manner in the case of random failures or malicious attacks. The robustness of multilayer networks has attracted great interest, where the modeling and theoretical studies of which always rely on the concept of multilayer networks and percolation methods. A straightforward and tacit assumption is that the interdependence across network layers is strong, which means that a node will fail entirely with the removal of all links if one of its interdependent neighbours fails. However, this oversimplification cannot describe the general form of interactions across the network layers in a real-world multilayer system. In this paper, we reveal the nature of the avalanche disintegration of general multilayer networks with arbitrary interdependency strength across network layers. Specifically, we identify that the avalanche process of the whole system can essentially be decomposed into two microscopic cascading dynamics in terms of the propagation direction of the failures: depth penetration and scope extension. In the process of depth penetration, the failures propagate from layer to layer, where the greater the number of failed nodes is, the greater the destructive power that will emerge in an interdependency group. In the process of scope extension, failures propagate with the removal of connections in each network layer. Under the synergy of the two processes, we find that the percolation transition of the system can be discontinuous or continuous with changes in the interdependency strength across network layers, which means that sudden system-wide collapse can be avoided by controlling the interdependency strength across network layers.
physics
We study the decay properties of Tonks-Girardeau gases escaping from a double-well potential. Initially, the gases are constrained between two infinite $\delta$ barriers with an on-center $\delta$-split potential. The strength of one of the obstacles is suddenly reduced, and the particles start to tunnel to the open space. Using the resonance expansion method, we derive the single-term approximate expression for the $N$-particle survival probability and demonstrate its effectiveness in both the exponential and nonexponential regimes. We also predict a parity effect and provide physical insights into its nature at different stages of the time evolution. We conclude that only the initial phase of the decay of the many-particle state may comply with an exponential law. The decay properties are dramatically affected by the presence of the split barrier. Our results reveal the overall decay mechanism of unstable Tonks-Girardeau gases from single and double quantum wells.
condensed matter
The problem of characterizing a multivariate distribution of a random vector using examination of univariate combinations of vector components is an essential issue of multivariate analysis. The likelihood principle plays a prominent role in developing powerful statistical inference tools. In this context, we raise the question: can the univariate likelihood function based on a random vector be used to provide the uniqueness in reconstructing the vector distribution? In multivariate normal (MN) frameworks, this question links to a reverse of Cochran's theorem that concerns the distribution of quadratic forms in normal variables. We characterize the MN distribution through the univariate likelihood type projections. The proposed principle is employed to illustrate simple techniques for assessing multivariate normality via well-known tests that use univariate observations. The presented testing strategy can exhibit high and stable power characteristics in comparison to the well-known procedures in various scenarios when observed vectors are non-MN distributed, whereas their components are normally distributed random variables. In such cases, the classical multivariate normality tests may break down completely. KEY WORDS: Characterization, Goodness of fit, Infinity divisible, Likelihood, Multivariate normal distribution, Projection, Quadratic form, Test for multivariate normality.
statistics
Lattice-based cryptography is one of the leading proposals for post-quantum cryptography. The Shortest Vector Problem (SVP) is arguably the most important problem for the cryptanalysis of lattice-based cryptography, and many lattice-based schemes have security claims based on its hardness. The best quantum algorithm for the SVP is due to Laarhoven [Laa16 PhD] and runs in (heuristic) time $2^{0.2653d + o(d)}$. In this article, we present an improvement over Laarhoven's result and present an algorithm that has a (heuristic) running time of $2^{0.2570 d + o(d)}$ where $d$ is the lattice dimension. We also present time-memory trade-offs where we quantify the amount of quantum memory and quantum random access memory of our algorithm. The core idea is to replace Grover's algorithm used in [Laa16 PhD] in a key part of the sieving algorithm by a quantum random walk in which we add a layer of local sensitive filtering.
quantum physics
Vision, as an inexpensive yet information rich sensor, is commonly used for perception on autonomous mobile robots. Unfortunately, accurate vision-based perception requires a number of assumptions about the environment to hold -- some examples of such assumptions, depending on the perception algorithm at hand, include purely lambertian surfaces, texture-rich scenes, absence of aliasing features, and refractive surfaces. In this paper, we present an approach for introspective vision for obstacle avoidance (IVOA) -- by leveraging a supervisory sensor that is occasionally available, we detect failures of stereo vision-based perception from divergence in plans generated by vision and the supervisory sensor. By projecting the 3D coordinates where the plans agree and disagree onto the images used for vision-based perception, IVOA generates a training set of reliable and unreliable image patches for perception. We then use this training dataset to learn a model of which image patches are likely to cause failures of the vision-based perception algorithm. Using this model, IVOA is then able to predict whether the relevant image patches in the observed images are likely to cause failures due to vision (both false positives and false negatives). We empirically demonstrate with extensive real-world data from both indoor and outdoor environments, the ability of IVOA to accurately predict the failures of two distinct vision algorithms.
computer science
In the novel stoichiometric iron-based material RbEuFe$_{4}$As$_{4}$ superconductivity coexists with a peculiar long-range magnetic order of Eu 4f states. Using angle-resolved photoemission spectroscopy, we reveal a complex three dimensional electronic structure and compare it with density functional theory calculations. Multiple superconducting gaps were measured on various sheets of the Fermi surface. High resolution resonant photoemission spectroscopy reveals magnetic order of the Eu 4f states deep into the superconducting phase. Both the absolute values and the anisotropy of the superconducting gaps are remarkably similar to the sibling compound without Eu, indicating that Eu magnetism does not affect the pairing of electrons. A complete decoupling between Fe- and Eu-derived states was established from their evolution with temperature, thus unambiguously demonstrating that superconducting and a long range magnetic orders exist independently from each other. The established electronic structure of RbEuFe$_{4}$As$_{4}$ opens opportunities for the future studies of the highly unorthodox electron pairing and phase competition in this family of iron-based superconductors with doping.
condensed matter
Models of particle physics that feature phase transitions typically provide predictions for stochastic gravitational wave signals at future detectors and such predictions are used to delineate portions of the model parameter space that can be constrained. The question is: how precise are such predictions? Uncertainties enter in the calculation of the macroscopic thermal parameters and the dynamics of the phase transition itself. We calculate such uncertainties with increasing levels of sophistication in treating the phase transition dynamics. Currently, the highest level of diligence corresponds to careful treatments of the source lifetime; mean bubble separation; going beyond the bag model approximation in solving the hydrodynamics equations and explicitly calculating the fraction of energy in the fluid from these equations rather than using a fit; and including fits for the energy lost to vorticity modes and reheating effects. The lowest level of diligence incorporates none of these effects. We compute the percolation and nucleation temperatures, the mean bubble separation, the fluid velocity, and ultimately the gravitational wave spectrum corresponding to the level of highest diligence for three explicit examples: SMEFT, a dark sector Higgs model, and the real singlet-extended Standard Model (xSM). In each model, we contrast different levels of diligence in the calculation and find that the difference in the final predicted signal can be several orders of magnitude. Our results indicate that calculating the gravitational wave spectrum for particle physics models and deducing precise constraints on the parameter space of such models continues to remain very much a work in progress and warrants care.
high energy physics phenomenology
By combining H$\alpha$ flux measurements from the Sloan Digital Sky Survey (SDSS) with UV flux observations from the Galaxy Evolution Explorer (GALEX), we examine the environmental dependence (through central/satellite distinction) of the rapid quenching and rejuvenation of galaxies. H$\alpha$ emissions trace the most massive stars, thereby indicating star-formation on timescales of $\sim 10$ Myr, while UV emission traces star-formation on timescales of $\sim 100$ Myr. These varying timescales are exploited to probe the most recent star-formation histories of galaxies. In this work, we define a class of transient galaxies which have UV emission typical of star-formation but negligible H$\alpha$ emission. We find that the occurrence of these transients has a strong stellar mass dependence in both the satellite and central population. However, while at stellar masses greater than $\sim 10^{10} M_\odot$ they occur with equal frequency regardless of environmental class, at lower stellar masses they are more common in satellites only, with an excess of about 1 percentage point across all low stellar mass galaxies. These satellite transients also have a strong halo mass and group-centric radial dependence suggesting they are driven by an environmental process. Finally, we select a sample of galaxies with H$\alpha$ emission but not UV emission which could contain short-timescale rejuvenating galaxies. These rejuvenating candidates are few in number and do not have a strong difference in their occurrence rate in centrals or satellites. These unique probes point to an environmental quenching mechanism which occurs on short timescales after the satellite has been in the group environment for a significant time - consistent with 'delayed-then-rapid' quenching.
astrophysics
Video microscopy has a long history of providing insights and breakthroughs for a broad range of disciplines, from physics to biology. Image analysis to extract quantitative information from video microscopy data has traditionally relied on algorithmic approaches, which are often difficult to implement, time consuming, and computationally expensive. Recently, alternative data-driven approaches using deep learning have greatly improved quantitative digital microscopy, potentially offering automatized, accurate, and fast image analysis. However, the combination of deep learning and video microscopy remains underutilized primarily due to the steep learning curve involved in developing custom deep-learning solutions. To overcome this issue, we introduce a software, DeepTrack 2.0, to design, train and validate deep-learning solutions for digital microscopy. We use it to exemplify how deep learning can be employed for a broad range of applications, from particle localization, tracking and characterization to cell counting and classification. Thanks to its user-friendly graphical interface, DeepTrack 2.0 can be easily customized for user-specific applications, and, thanks to its open-source object-oriented programming, it can be easily expanded to add features and functionalities, potentially introducing deep-learning-enhanced video microscopy to a far wider audience.
electrical engineering and systems science
We study the inclusive J/psi production at large transverse momenta at lepton-hadron colliders in the limit when the exchange photon is quasi real, also referred to as photoproduction. Our computation includes the leading-P_T leading-v next-to-leading alpha_s corrections. In particular, we consider the contribution from J/psi plus another charm quark, by employing for the first time in quarkonium photoproduction the variable-flavour-number scheme. We also include a QED-induced contribution via an off-shell photon which remained ignored in the literature and which we show to be the leading contribution at high P_T within the reach of the EIC. In turn, we use our computation of J/psi+charm to demonstrate its observability at the future EIC and the EIC sensitivity to probe the non-perturbative charm content of the proton at high x.
high energy physics phenomenology
Modification of the gap at the Dirac point (DP) in antiferromagnetic (AFM) axion topological insulator MnBi$_2$Te$_4$ and its electronic and spin structure has been studied by angle- and spin-resolved photoemission spectroscopy (ARPES) under laser excitation with variation of temperature (9-35~K), light polarization and photon energy. We have distinguished both a large (62-67~meV) and a reduced (15-18~meV) gap at the DP in the ARPES dispersions, which remains open above the N\'eel temperature ($T_\mathrm{N}=24.5$~K). We propose that the gap above $T_\mathrm{N}$ remains open due to short-range magnetic field generated by chiral spin fluctuations. Spin-resolved ARPES, XMCD and circular dichroism ARPES measurements show a surface ferromagnetic ordering for large-gap sample and significantly reduced effective magnetic moment for the reduced-gap sample. These effects can be associated with a shift of the topological DC state towards the second Mn layer due to structural defects and mechanical disturbance, where it is influenced by a compensated effect of opposite magnetic moments.
condensed matter
In this paper, the sum secure degrees of freedom (SDoF) of the $K$-user Multiple Input/Single Output (MISO) Broadcast Channel with Confidential Messages (BCCM) and alternating Channel State Information at the Transmitter (CSIT) is investigated. In the MISO BCCM, a $K$-antenna transmitter (TX) communicates toward $K$ single-antenna receivers (RXs), so that message for RX $k$ is kept secret from RX $j$ with $j<k$. For this model, we consider the scenario in which the CSI of the RXs from $2$ to $K$ is instantaneously known at the transmitter while CSI of RX $1$ is known at the transmitter (i) instantaneously for half of the time and (ii) with a unit delay for the remainder of the time. We refer to this CSIT availability as \emph{alternating} CSIT. Alternating CIST has been shown to provide synergistic gains in terms of SDoF and is thus of a viable strategy to ensure secure communication by simply relying on the CSI feedback strategy. Our main contribution is the characterization of sum SDoF for this model as $SDoF_{\rm sum}= (2K-1)/2$. Interestingly, this $SDoF_{\rm sum}$ is attained by a rather simple achievability in which the TX uses artificial noise to prevent the decoding of the message of the unintended receivers at RX $1$. For simplicity first, the proof for the case $K=3$ is discussed in detail and after that, we have presented the results for any number of RXs.
computer science
Extreme multi-label text classification (XMTC) aims to tag a text instance with the most relevant subset of labels from an extremely large label set. XMTC has attracted much recent attention due to massive label sets yielded by modern applications, such as news annotation and product recommendation. The main challenges of XMTC are the data scalability and sparsity, thereby leading to two issues: i) the intractability to scale to the extreme label setting, ii) the presence of long-tailed label distribution, implying that a large fraction of labels have few positive training instances. To overcome these problems, we propose GNN-XML, a scalable graph neural network framework tailored for XMTC problems. Specifically, we exploit label correlations via mining their co-occurrence patterns and build a label graph based on the correlation matrix. We then conduct the attributed graph clustering by performing graph convolution with a low-pass graph filter to jointly model label dependencies and label features, which induces semantic label clusters. We further propose a bilateral-branch graph isomorphism network to decouple representation learning and classifier learning for better modeling tail labels. Experimental results on multiple benchmark datasets show that GNN-XML significantly outperforms state-of-the-art methods while maintaining comparable prediction efficiency and model size.
computer science
Bilayer graphene with small internal twist angles develops large scale moir\'e patterns with flat energy bands hosting both correlated insulating states and superconductivity. The large system size and intricate band structure have however hampered investigations into the properties of the superconducting state. Here, we use full-scale atomistic modeling with local electronic interactions and find a highly inhomogeneous superconducting state with nematic ordering on both the atomic and moir\'e lattice length scales. More specifically, we obtain locally anisotropic real-valued d-wave pairing with a nematic vector winding throughout the moir\'e pattern, generating a three-fold ground state degeneracy. Despite the d-wave nature, the superconducting state has a full energy gap, which we further tie to a {\pi}-phase interlayer coupling. The superconducting nematicity is easily detected through signatures in the local density of states. Our results show not only that atomistic modeling is essential for twisted bilayer graphene but also that the superconducting state is necessarily distinctly different from that of the high-temperature cuprate superconductors, despite similarity in electronic interactions.
condensed matter
A widely used method for parameterizing the outcomes of common envelopes (CEs) involves defining an ejection efficiency, $\bar{\alpha}_{\mathrm{eff}}$, that represents the fraction of orbital energy used to unbind the envelope as the orbit decays. Given $\bar{\alpha}_{\mathrm{eff}}$, a prediction for the post-CE orbital separation is possible with knowledge of the energy required to unbind the primary's envelope from its core. Unfortunately, placing observational constraints on $\bar{\alpha}_{\mathrm{eff}}$ is challenging as it requires knowledge of the primary's structure at the onset of the common envelope phase. Numerical simulations have also had difficulties reproducing post-CE orbital configurations as they leave extended, but still bound, envelopes. Using detailed stellar interior profiles, we calculate $\bar{\alpha}_{\mathrm{eff}}$ values for a matrix of primary-companion mass pairs when the primary is at maximal extent in its evolution. We find that the ejection efficiency is most sensitive to the properties of the surface-contact convective region (SCCR). In this region, the convective transport timescales are often short compared to orbital decay timescales, thereby allowing the star to effectively radiate orbital energy and thus lower $\bar{\alpha}_{\mathrm{eff}}$. The inclusion of convection in numerical simulations of CEs may aid ejection without the need for additional energy sources as the orbit must shrink substantially further before the requisite energy can be tapped to drive ejection. Additionally, convection leads to predicted post-CE orbital periods of less than a day in many cases, an observational result that has been difficult to reproduce in population studies where $\bar{\alpha}_{\mathrm{eff}}$ is taken to be constant. Finally, we provide a simple method to calculate $\bar{\alpha}_{\mathrm{eff}}$ if the properties of the SCCR are known.
astrophysics
It is well established that quantum criticality is one of the most intriguing phenomena which signals the presence of new states of matter. Without prior knowledge of the local order parameter, the quantum information metric (or fidelity susceptibility) can indicate the presence of a phase transition as well as it measures distance between quantum states. In this work, we calculate the distance between quantum states which is equal to the fidelity susceptibility in quantum model for a time-dependent system describing a two-level atom coupled to a time-driven external field. As inspired by the Landau-Zener quantum model, we find in the present work information metric induced by fidelity susceptibility. We, for the first time, derive a higher-order rank-3 tensor as a third-order fidelity susceptibility. Having computed quantum noise function in this simple time-dependent model we show that the noise function eternally lasts long in our model.
quantum physics
Indoor localization has become an important issue for wireless sensor networks. This paper presents a zoning-based localization technique that uses WiFi signals and works efficiently in indoor environments. The targeted area is composed of several zones, the objective being to determine the zone of the sensor using an observation model based on statistical learning.
statistics
This note examines the safety verification of the solution of Ito stochastic differential equations using the notion of stochastic zeroing barrier function. The main tools in the proposed method include Ito calculus and the concept of stochastic invariant set.
electrical engineering and systems science
In this work, we propose a fully discrete energy stable scheme for the phase-field moving contact line model with variable densities and viscosities. The mathematical model consists of a Cahn-Hilliard equation, a Navier-Stokes equation and the generalized Navier boundary condition for the moving contact line. A scalar auxiliary variable is adopted to transform the governing system into an equivalent form, allowing the double well potential to be treated semi-explicitly. A stabilization term is added to balance the explicit nonlinear term originating from the surface energy at fluid-solid interface. A pressure stabilization method is used to decouple the computation of velocity and pressure. Some subtle implicit-explicit treatments are adopted to deal with convention and stress terms. We establish a rigorous proof of energy stability for the proposed time-marching scheme. Then a finite difference method on staggered grids is used to spatially discretize the constructed time-marching scheme. We further prove that the fully discrete scheme also satisfies the discrete energy dissipation law. Numerical results demonstrate accuracy and energy stability of the proposed scheme. Using our numerical scheme, we analyze the contact line dynamics through a shear flow driven droplet sliding case. Three-dimensional droplet spreading is also investigated on a chemically patterned surface. Our numerical simulation accurately predicts the expected energy evolutions and it successfully reproduces expected phenomena that an oil droplet contracts inwards on a hydrophobic zone and spreads outwards quickly on a hydrophilic zone.
physics
Our current knowledge of cosmic star-formation history during the first two billion years (corresponding to redshift z >3) is mainly based on galaxies identified in rest-frame ultraviolet light. However, this population of galaxies is known to under-represent the most massive galaxies, which have rich dust content and/or old stellar populations. This raises the questions of the true abundance of massive galaxies and the star-formation-rate density in the early universe. Although several massive galaxies that are invisible in the ultraviolet have recently been confirmed at early epochs, most of them are extreme starbursts with star-formation rates exceeding 1000 solar masses per year, suggesting that they are unlikely to represent the bulk population of massive galaxies. Here we report submillimeter (wavelength 870um) detections of 39 massive star-forming galaxies at z > 3, which are unseen in the spectral region from the deepest ultraviolet to the near-infrared. With a space density of about $2 \times 10^{-5}$ per cubic megaparsec (two orders of magnitudes higher than extreme starbursts) and star-formation rates of 200 solar masses per year, these galaxies represent the bulk population of massive galaxies that have been missed from previous surveys. They contribute a total star-formation-rate density ten times larger than that of equivalently massive ultraviolet-bright galaxies at z >3. Residing in the most massive dark matter halos at their redshifts, they are probably the progenitors of the largest present-day galaxies in massive groups and clusters. Such a high abundance of massive and dusty galaxies in the early universe challenges our understanding of massive-galaxy formation.
astrophysics
Background: This study aims to evaluate the Chicago Teen Pregnancy Prevention Initiative delivery optimization outcomes given policy-neutral and policy-focused approaches to deliver this program to at-risk teens across the City of Chicago. Methods: We collect and compile several datasets from public sources including: Chicago Department of Public Health clinic locations, two public health statistics datasets, census data of Chicago, list of Chicago public high schools, and their Locations. Our policy-neutral approach will consist of an equal distribution of funds and resources to schools and centers, regardless of past trends and outcomes. The policy-focused approaches will evaluate two models: first, a funding model based on prediction models from historical data; and second, a funding model based on economic and social outcomes for communities. Results: Results of this study confirms our initial hypothesis, that even though the models are optimized from a machine learning perspective, there is still possible that the models will produce wildly different results in the real-world application. Conclusions: When ethics and ethical considerations are extended beyond algorithmic optimization to encompass output and societal optimization, the foundation and philosophical grounding of the decision-making process become even more critical in the knowledge discovery process.
computer science
Recent advancements in recurrent neural network (RNN) research have demonstrated the superiority of utilizing multiscale structures in learning temporal representations of time series. Currently, most of multiscale RNNs use fixed scales, which do not comply with the nature of dynamical temporal patterns among sequences. In this paper, we propose Adaptively Scaled Recurrent Neural Networks (ASRNN), a simple but efficient way to handle this problem. Instead of using predefined scales, ASRNNs are able to learn and adjust scales based on different temporal contexts, making them more flexible in modeling multiscale patterns. Compared with other multiscale RNNs, ASRNNs are bestowed upon dynamical scaling capabilities with much simpler structures, and are easy to be integrated with various RNN cells. The experiments on multiple sequence modeling tasks indicate ASRNNs can efficiently adapt scales based on different sequence contexts and yield better performances than baselines without dynamical scaling abilities.
computer science
We present a novel approach to generate higher-order initial conditions (ICs) for cosmological simulations that take into account the distinct evolution of baryons and dark matter. We focus on the numerical implementation and the validation of its performance, based on both collisionless N-body simulations and full hydrodynamic Eulerian and Lagrangian simulations. We improve in various ways over previous approaches that were limited to first-order Lagrangian perturbation theory (LPT). Specifically, we (1) generalize nth-order LPT to multi-fluid systems, allowing 2LPT or 3LPT ICs for two-fluid simulations, (2) employ a novel propagator perturbation theory to set up ICs for Eulerian codes that are fully consistent with 1LPT or 2LPT, (3) demonstrate that our ICs resolve previous problems of two-fluid simulations by using variations in particle masses that eliminate spurious deviations from expected perturbative results, (4) show that the improvements achieved by going to higher-order PT are comparable to those seen for single-fluid ICs, and (5) demonstrate the excellent (i.e., few per cent level) agreement between Eulerian and Lagrangian simulations, once high-quality initial conditions are used. The rigorous development of the underlying perturbation theory is presented in a companion paper. All presented algorithms are implemented in the Monofonic Music-2 package that we make publicly available.
astrophysics
In this study the Leidenfrost temperature during spray cooling of very hot substrates is experimentally measured. The spray parameters, i.e. the drop diameters and velocities and the mass flux, are very accurately measured. Astonishingly, the measured Leidenfrost temperature is independent of any of the spray impact parameters, but is determined exclusively by the materials of the liquid and the substrate. The mechanism of film boiling is explained by the formation of a fast propagating vaporizing front, when the inertial forces in the associated liquid flow are comparable with the viscous stresses. This leads to a theoretical prediction for the Leidenfrost temperature which agrees well with the experimental data.
physics
We prove that for any isometric action of a group on a unit sphere of dimension larger than one, the quotient space has diameter zero or larger than a universal dimension-independent positive constant.
mathematics
In 1984 Bedford and McMullen independently introduced a family of self-affine sets now known as \emph{Bedford-McMullen carpets}. Their work stimulated a lot of research in the areas of fractal geometry and non-conformal dynamics. In this survey article we discuss some aspects of Bedford-McMullen carpets, focusing mostly on dimension theory.
mathematics
The discovery of a non-zero rate for a lepton flavor violating decay mode of the Higgs boson would definitely be an indication of New Physics. We review the prospects for such signal in Two Higgs Doublet Models, in particular for Higgs boson decays into $\tau \mu$ final states. We will show that this scenario contains all the necessary ingredients to provide large flavor violating rates and still be compatible with the stringent limits from direct searches and low-energy flavor experiments.
high energy physics phenomenology
Hysteresis exhibited by current-voltage characteristics during field-emission experiments is often considered undesirable in terms of practical applications. However, this is an appealing effect for the purposes of memristive devices. We developed a two-stage model to describe hysteretic characteristics, with a particular focus on the system which includes a cathode made of a single-layered graphene sheet on a substrate. In addition to hysteresis, the current-voltage curves display also an unusual self-crossing behavior. The presented memristive model can be used for quantitative descriptions of different hysteretic characteristics such as abrupt changes and self-crossings and for understanding (and modeling) the processes associated with field emission from plane graphene emitters.
condensed matter
Automation driving techniques have seen tremendous progresses these last years, particularly due to a better perception of the environment. In order to provide safe yet not too conservative driving in complex urban environment, data fusion should not only consider redundant sensing to characterize the surrounding obstacles, but also be able to describe the uncertainties and errors beyond presence/absence (be it binary or probabilistic). This paper introduces an enriched representation of the world, more precisely of the potential existence of obstacles through an evidential grid map. A method to create this representation from 2 very different sensors, laser scanner and stereo camera, is presented along with algorithms for data fusion and temporal updates. This work allows a better handling of the dynamic aspects of the urban environment and a proper management of errors in order to create a more reliable map. We use the evidential framework based on the Dempster-Shafer theory to model the environment perception by the sensors. A new combination operator is proposed to merge the different sensor grids considering their distinct uncertainties. In addition, we introduce a new long-life layer with high level states that allows the maintenance of a global map of the entire vehicle's trajectory and distinguish between static and dynamic obstacles. Results on a real road dataset show that the environment mapping data can be improved by adding relevant information that could be missed without the proposed approach.
computer science
The field of optomechanics provides us with several examples of quantum photon-phonon interface. In this paper, we theoretically investigate the generation and manipulation of quantum correlations in a microfabricated optomechanical array. We consider a system consisting of localized photonic and phononic modes interacting locally via radiation pressure at each lattice site with the possibility of hopping of photons and phonons between neighboring sites. We show that such an interaction can correlate various modes of a driven coupled optomechanical array with well-chosen system parameters. Moreover, in the linearized regime of Gaussian fluctuations, the quantum correlations not only survive in the presence of thermal noise, but may also be generated thermally. We find that these optomechanical arrays provide a suitable platform for quantum simulation of various many-body systems.
quantum physics
In quantum nanoelectronics, time-dependent electrical currents are built from few elementary excitations emitted with well-defined wavefunctions. However, despite the realization of sources generating quantized numbers of excitations, and despite the development of the theoretical framework of time-dependent quantum electronics, extracting electron and hole wavefunctions from electrical currents has so far remained out of reach, both at the theoretical and experimental levels. In this work, we demonstrate a quantum tomography protocol which extracts the generated electron and hole wavefunctions and their emission probabilities from any electrical current. It combines two-particle interferometry with signal processing. Using our technique, we extract the wavefunctions generated by trains of Lorentzian pulses carrying one or two electrons. By demonstrating the synthesis and complete characterization of electronic wavefunctions in conductors, this work offers perspectives for quantum information processing with electrical currents and for investigating basic quantum physics in many-body systems.
condensed matter
Empirical optimal transport (OT) plans and distances provide effective tools to compare and statistically match probability measures defined on a given ground space. Fundamental to this are distributional limit laws and we derive a central limit theorem for the empirical OT distance of circular data. Our limit results require only mild assumptions in general and include prominent examples such as the von Mises or wrapped Cauchy family. Most notably, no assumptions are required when data are sampled from the probability measure to be compared with, which is in strict contrast to the real line. A bootstrap principle follows immediately as our proof relies on Hadamard differentiability of the OT functional. This paves the way for a variety of statistical inference tasks and is exemplified for asymptotic OT based goodness of fit testing for circular distributions. We discuss numerical implementation, consistency and investigate its statistical power. For testing uniformity, it turns out that this approach performs particularly well for unimodal alternatives and is almost as powerful as Rayleigh's test, the most powerful invariant test for von Mises alternatives. For regimes with many modes the circular OT test is less powerful which is explained by the shape of the corresponding transport plan.
statistics
Ambitwistor strings are chiral (holomorphic) strings whose target is the space of complex null geodesics, ambitwistor space. We introduce twistor representations of ambitwistor space in 6 and 5 dimensions. In 6d the twistor representation is naturally conformally invariant. Anomaly cancellation leads to models that describe biadjoint scalar amplitudes and certain conformally invariant gauge and gravity theories, respectively of 4$^{\rm th}$ and 6$^{\rm th}$ order. There are three such models, reflecting triality for the conformal group SO(8) associated to these 6d models. On reduction to five dimensions, gauge anomaly cancellation requires supersymmetry and the resulting models describe maximally supersymmetric Yang-Mills and gravity. The twistor representation of these ambitwistor strings lead to formul{\ae} for maximally supersymmetric gauge and gravity amplitudes based on the polarized scattering equations in 5d, found earlier by the first two authors.
high energy physics theory
We present a new, elementary way to obtain axially symmetric Gaussian processes on the sphere, in order to accommodate for the directional anisotropy of global climate data in geostatistical analysis.
statistics
We present a detailed X-ray spectral analysis of 1152 AGNs selected in the Chandra Deep Fields (CDFs), in order to identify highly obscured AGNs ($N_{\rm H} > 10^{23}\ \rm cm^{-2}$). By fitting spectra with physical models, 436 (38%) sources with $L_{\rm X} > 10^{42}\ \rm erg\ s^{-1}$ are confirmed to be highly obscured, including 102 Compton-thick (CT) candidates. We propose a new hardness-ratio measure of the obscuration level which can be used to select highly obscured AGN candidates. The completeness and accuracy of applying this method to our AGNs are 88% and 80%, respectively. The observed logN-logS relation favors cosmic X-ray background models that predict moderate (i.e., between optimistic and pessimistic) CT number counts. 19% (6/31) of our highly obscured AGNs that have optical classifications are labeled as broad-line AGNs, suggesting that, at least for part of the AGN population, the heavy X-ray obscuration is largely a line-of-sight effect, i.e., some high-column-density clouds on various scales (but not necessarily a dust-enshrouded torus) along our sightline may obscure the compact X-ray emitter. After correcting for several observational biases, we obtain the intrinsic NH distribution and its evolution. The CT-to-highly-obscured fraction is roughly 52% and is consistent with no evident redshift evolution. We also perform long-term (~17 years in the observed frame) variability analyses for 31 sources with the largest number of counts available. Among them, 17 sources show flux variabilities: 31% (5/17) are caused by the change of NH, 53% (9/17) are caused by the intrinsic luminosity variability, 6% (1/17) are driven by both effects, and 2 are not classified due to large spectral fitting errors.
astrophysics
We study the real-time dynamics of a periodic $XY$ system exposed to a composite field comprised of a constant homogeneous magnetic and a quantized circularly polarized electromagnetic fields. The interaction between the quantized mode and spin-magnetic moments is modeled by the Dicke Hamiltonian. The rotating wave approximation is applied and the conditions for its validity are discussed. It is shown that if initially all of the excitations are contained in the field, then in the regime of large detuning, the main evolutionary effect involves oscillations of the excitations between the zero-momentum modes of the chain and the field. Accordingly, the reduced photon number and magnetization per site reveal a sort of oscillatory behavior. Effective Hamiltonians describing the short-time dynamics of the present model for small number of excitations and large detuning are introduced. The resonance case is considered in the context of photon emission from the chain initially prepared in the (partially) excited state. In particular, it is demonstrated, in the framework of a specific example, that the superradiant behavior shows up at the beginning of the emission, when we have an initial state with a maximally excited $XY$ chain. Possible applications of the model to problems such as spin chain and $J$-aggregate in a single-mode cavity are discussed.
condensed matter
Capsule Networks attempt to represent patterns in images in a way that preserves hierarchical spatial relationships. Additionally, research has demonstrated that these techniques may be robust against adversarial perturbations. We present an improvement to training capsule networks with added robustness via non-parametric kernel methods. The representations learned through the capsule network are used to construct covariance kernels for Gaussian processes (GPs). We demonstrate that this approach achieves comparable prediction performance to Capsule Networks while improving robustness to adversarial perturbations and providing a meaningful measure of uncertainty that may aid in the detection of adversarial inputs.
statistics
We review the concepts and the present state of theoretical studies of spin-imbalanced superfluidity, in particular the elusive Fulde-Ferrell-Larkin-Ovchinnikov (FFLO) state, in the context of ultracold quantum gases. The comprehensive presentation of the theoretical basis for the FFLO state that we provide is useful also for research on the interplay between magnetism and superconductivity in other physical systems. We focus on settings that have been predicted to be favourable for the FFLO state, such as optical lattices in various dimensions and spin-orbit coupled systems. These are also the most likely systems for near-future experimental observation of the FFLO state. Theoretical bounds, such as Bloch's and Luttinger's theorems, and experimentally important limitations, such as finite-size effects and trapping potentials, are considered. In addition, we provide a comprehensive review of the various ideas presented for the observation of the FFLO state. We conclude our review with an analysis of the open questions related to the FFLO state, such as its stability, superfluid density, collective modes and extending the FFLO superfluid concept to new types of lattice systems.
condensed matter
For addressing the remarkable difference between neutrino and quark mixings, high-scale mixings relations (HSMR) or unification (HSMU) hypotheses were proposed. These neutrino phenomenology frameworks have been explored with respect to bounds from neutrino oscillations and relevant cosmological experiments. However there are caveats with regards to assessing the models' compatibility with data in a statiscally robust and convergent manner. These can be addressed by using Bayesian algorithms. Global fits of the models to experimental constraints can readily yield compatible parameter regions, including Majorana phases. The posterior samples could be used for studying correlations between neutrino oservables and prospects for updates of related experiments.
high energy physics phenomenology
In a one-dimensional (1D) disordered potential, quantum interferences leading to Anderson lo-calization are ubiquitous, such that all wave-functions are exponentially localized. Moreover, no phase transition toward delocalization is expected in 1D. This behavior is strongly modified in the presence of a bias force. We experimentally study this case, launching a non-interacting 39 K Bose-Einstein condensate in a 1D disordered potential induced by a far-off-resonance laser speckle, while controlling a bias force. In agreement with theoretical predictions, we observe a transition between algebraic localization and delocalization as a function of our control parameter that is the relative strength of the disorder against the bias force. We also demonstrate that the initial velocity of the wave-packet only plays a role through an effective disorder strength due to the correlation of the disorder. Adding a bias force is a quite natural way to probe the transport properties of quantum systems, a subject of broad interest that can be in particular addressed with atomic quantum gases thanks to their high degree of control and versatility [1]. For example, Bloch oscillations has been measured through the addition of a constant force to atoms in periodic potential induced by an optical lattice [2]. A force applied to a harmonic trap is equivalent to a trap displacement. The response to such a displacement permits to reveal the fluid or insulating behavior of atomic systems. In 1D interacting Bose gases, the pinning transition by an optical lattice [3] or the insulating transition in quasi-disordered optical lattice [4, 5] have been studied in this manner. More recently, transport in quantum gases is also studied in junction-type setup more analogous to condensed-matter systems: two reservoirs with different chemical potentials are connected through a constriction. For example, in a gas of fermions, the quantization of conductance through a quantum point contact [6] and the superfluid to normal transition in a disordered thin film have been observed [7]. In our work, we focus on the transport of non-interacting particles in disordered media. Without a bias force, quantum interferences between multiple paths lead to Anderson localization [8] whose signature is an exponential decay in space of single particle wave-function [9]. This phenomenon is ubiquitous in wave/quantum physics and it has been observed in many physical contexts [10] including condensed-matter [11] and ultra-cold atoms [12-14]. One-dimensional truly disordered systems are always localized [15], contrary to the 3D case where a phase transition between localized and extended single particle wave-functions takes place as a function of the disorder strength [16-18]. The localization properties of 1D disordered systems are modified in the presence of a bias force. Theoretical studies predict a transition from algebraic localization to delocalization as a function of a single control non-dimensional parameter $\alpha$ which is the ratio of the force to the disorder strength [19, 20]. Physically, $\alpha$ is the relative energy gain $\Delta$E/E of a particle of energy E when moving over a localization length. Interestingly, in a 1D white noise disorder, this quantity is independent of E as the localization length is proportional to E. If $\alpha$ is small, the force does not considerably change the localization behavior of the particle while for large $\alpha$ its dynamics is severely affected leading to delocalization. This localization-delocalization transition is predicted in the infinite time limit for white noise disorder [20]. In a correlated disorder, as the one produced from a far-off-resonance laser speckle [21], the situation is more complicated. Speckles have no Fourier component beyond a spatial frequency 2k c. As a consequence, back-scattering and localization are not expected in the framework of Born approximation for atoms with wavevectors k > k c [12, 22]. Since localized wave-functions always have a small fraction at long distance corresponding to large energies and momenta in the presence of a bias force, we thus expect correlation-induced delocalization at infinite time. However, signatures of the algebraic localization-delocalization transition are predicted to be observable at transient times [20]. In this paper, we report on the observation of the algebraic localization-delocalization transition with cold-atoms propagating in a one dimensional disordered potential in the presence of a controlled bias force. We experimentally show that the non-dimensional parameter $\alpha$ is the only relevant parameter to describe the transition. We notice that the initial velocity of the quantum wave packet only plays a role through the correlation of the disordered potential, showing that the transition is in-trinsically energy independent. In the localized regime, we demonstrate an algebraic decay of the density and measure the corresponding decay exponent as a function of $\alpha$. At large disorder strength, a saturation of the expo
physics
We consider various methods and techniques for measuring electron magnetization and susceptibility, which are used in experimental condensed matter physics. The list of considered methods for macroscopic measurements includes magnetomechanic, electromagnetic, modulation-type, and also thermodynamic methods based on the chemical potential variation. We also consider local methods of magnetic measurements based on the spin Hall effects and NV-centers. Several scanning probe magnetometers-microscopes are considered, such as magnetic resonance force microscope, SQUID-microscope, and Hall microscope. The review focuses on the spin magnetization measurements of electrons in non-magnetic materials and artificial systems, particularly, in low-dimensional electron systems in semiconductors and in nanosystems, which came to the forefront in recent years.
condensed matter
The global increase in materials consumption calls for innovative materials, with tailored performance and multi-functionality, that are environmentally sustainable. Composites from renewable resources offer solutions to fulfil these demands but have so far been dominated by hybrid petrochemicals-based matrixes reinforced by natural fillers. Here, we present biological matrix composites with properties comparable to wood and commercial polymers. The biocomposites are obtained from cultured, undifferentiated plant cells, dehydrated and compressed under controlled conditions, forming a lamellar microstructure. Their stiffness and strength surpass that of commercial plastics of similar density, like polystyrene, and low-density polyethylene, while being entirely biodegradable. The properties can be further tuned varying the fabrication process. For example, filler particles can be integrated during fabrication, to vary the mechanical response or introduce new functionalities.
physics
We consider long-lived relic particles as the source of the PeV-scale neutrinos detected at the IceCube observatory over the last six years. We derive the present day neutrino flux, including primary neutrinos from direct decays, secondary neutrinos from electroweak showering, and tertiary neutrinos from re-scatters off the relic neutrino background. We compare the high-energy neutrino flux prediction to the most recently available datasets and find qualitative differences to expected spectra from other astrophysical processes. We utilize electroweak corrections to constrain heavy decaying relic abundances, using measurements impacted by electromagnetic energy injection, such as light element abundances during Big Bang nucleosynthesis, cosmic microwave background anisotropies, and diffuse $\gamma$-ray spectra. We compare these abundances to those necessary to source the IceCube neutrinos and find two viable regions in parameter space, ultimately testable by future neutrino, $\gamma$-ray, and cosmic microwave background observatories.
high energy physics phenomenology
Panoptic segmentation aims to perform instance segmentation for foreground instances and semantic segmentation for background stuff simultaneously. The typical top-down pipeline concentrates on two key issues: 1) how to effectively model the intrinsic interaction between semantic segmentation and instance segmentation, and 2) how to properly handle occlusion for panoptic segmentation. Intuitively, the complementarity between semantic segmentation and instance segmentation can be leveraged to improve the performance. Besides, we notice that using detection/mask scores is insufficient for resolving the occlusion problem. Motivated by these observations, we propose a novel deep panoptic segmentation scheme based on a bidirectional learning pipeline. Moreover, we introduce a plug-and-play occlusion handling algorithm to deal with the occlusion between different object instances. The experimental results on COCO panoptic benchmark validate the effectiveness of our proposed method. Codes will be released soon at https://github.com/Mooonside/BANet.
computer science
We describe the dynamics of quantum discord of two interacting spin-1/2 subjected to controllable time-dependent magnetic fields. The exact time evolution of discord is given for various input mixed states consisting of classical mixtures of two Bell states. The quantum discord manifests a complex oscillatory behaviour in time and is compared with that of quantum entanglement, measured by concurrence. The interplay of the action of the time-dependent magnetic fields and the spin-coupling mechanism in the occurrence and evolution of quantum correlations is examined in detail.
quantum physics
Proximity effects induced in the 2D Dirac material graphene potentially open access to novel and intriguing physical phenomena. Thus far, the coupling between graphene and ferromagnetic insulators has been experimentally established. However, only very little is known about graphene's interaction with antiferromagnetic insulators. Here, we report a low temperature study of the electronic properties of high quality van der Waals heterostructures composed of a single graphene layer proximitized with $\alpha$-RuCl$_3$. The latter is known to become antiferromagnetically ordered below 10 K. Shubnikov de Haas oscillations in the longitudinal resistance together with Hall resistance measurements provide clear evidence for a band realignment that is accompanied by a transfer of electrons originally occupying the graphene's spin degenerate Dirac cones into $\alpha$-RuCl$_3$ band states with in-plane spin polarization. Left behind are holes in two separate Fermi pockets, only the dispersion of one of which is distorted near the Fermi energy due to spin selective hybridization with these spin polarized $\alpha$-RuCl$_3$ band states. This interpretation is supported by our DFT calculations. An unexpected damping of the quantum oscillations as well as a zero field resistance upturn close to the N$\'e$el temperature of $\alpha$-RuCl$_3$ suggests the onset of additional spin scattering due to spin fluctuations in the $\alpha$-RuCl$_3$.
condensed matter
We carry out three dimensional SPH simulations to show that a migrating giant planet strongly suppresses the spiral structure in self-gravitating discs. We present mock ALMA continuum observations which show that in the absence of a planet, spiral arms due to gravitational instability are easily observed. Whereas in the presence of a giant planet, the spiral structures are suppressed by the migrating planet resulting in a largely axisymmetric disc with a ring and gap structure. Our modelling of the gas kinematics shows that the planet's presence could be inferred, for example, using optically thin 13C16O. Our results show that it is not necessary to limit the gas mass of discs by assuming high dust-to-gas mass ratios in order to explain a lack of spiral features that would otherwise be expected in high mass discs.
astrophysics
Pre-trained Transformer-based neural language models, such as BERT, have achieved remarkable results on varieties of NLP tasks. Recent works have shown that attention-based models can benefit from more focused attention over local regions. Most of them restrict the attention scope within a linear span, or confine to certain tasks such as machine translation and question answering. In this paper, we propose a syntax-aware local attention, where the attention scopes are restrained based on the distances in the syntactic structure. The proposed syntax-aware local attention can be integrated with pretrained language models, such as BERT, to render the model to focus on syntactically relevant words. We conduct experiments on various single-sentence benchmarks, including sentence classification and sequence labeling tasks. Experimental results show consistent gains over BERT on all benchmark datasets. The extensive studies verify that our model achieves better performance owing to more focused attention over syntactically relevant words.
computer science
We demonstrate that tailored laser beams provide a powerful means to make quantum vacuum signatures in strong electromagnetic fields accessible in experiment. Typical scenarios aiming at the detection of quantum vacuum nonlinearities at the high-intensity frontier envision the collision of focused laser pulses. The effective interaction of the driving fields mediated by vacuum fluctuations gives rise to signal photons encoding the signature of quantum vacuum nonlinearity. Isolating a small number of signal photons from the large background of the driving laser photons poses a major experimental challenge. The main idea of the present work is to modify the far-field properties of a driving laser beam to exhibit a field-free hole in its center, thereby allowing for an essentially background free measurement of the signal scattered in the forward direction. Our explicit construction makes use of a peculiar far-field/focus duality.
quantum physics
We present a general framework to systematically study the derivative expansion of asymptotically safe quantum gravity. It is based on an exact decoupling and cancellation of different modes in the Landau limit, and implements a correct mode count as well as a regularisation based on geometrical considerations. It is applicable independent of the truncation order. To illustrate the power of the framework, we discuss the quartic order of the derivative expansion and its fixed point structure as well as physical implications.
high energy physics theory
Twin-field quantum key distribution (TF-QKD) and its variant protocols are highly attractive due to the advantage of overcoming the rate-loss limit for secret key rates of point-to-point QKD protocols. For variations of TF-QKD, the key point to ensure security is switching randomly between a code mode and a test mode. Among all TF-QKD protocols, their code modes are very different, e.g. modulating continuous phases, modulating only two opposite phases, and sending or not sending signal pulses. Here we show that, by discretizing the number of global phases in the code mode, we can give a unified view on the first two types of TF-QKD protocols, and demonstrate that increasing the number of discrete phases extends the achievable distance, and as a trade-off, lowers the secret key rate at short distances due to the phase post-selection.
quantum physics
Accurate knowledge of the distribution system topology and parameters is required to achieve good voltage controls, but this is difficult to obtain in practice. This paper develops a model-free approach based on the surrogate model and deep reinforcement learning (DRL). We have also extended it to deal with unbalanced three-phase scenarios. The key idea is to learn a surrogate model to capture the relationship between the power injections and voltage fluctuation of each node from historical data instead of using the original inaccurate model affected by errors and uncertainties. This allows us to integrate the DRL with the learned surrogate model. In particular, DRL is applied to learn the optimal control strategy from the experiences obtained by continuous interactions with the surrogate model. The integrated framework contains training three networks, i.e., surrogate model, actor, and critic networks, which fully leverage the strong nonlinear fitting ability of deep learning and DRL for online decision making. Several single-phase approaches have also been extended to deal with three-phase unbalance scenarios and the simulation results on the IEEE 123-bus system show that our proposed method can achieve similar performance as those that use accurate physical models.
electrical engineering and systems science
Implications are explored of promoting non-conformal scale-invariant theories to conformal theories by nonlinearly realizing the missing symmetry. Properties of the associated Nambu-Goldstone mode imply that conformal invariance cannot be spontaneously broken to scale invariance in unitary theories and that, as well known, scale invariant unitary theories in two dimensions are also conformal. The promoted theories have only conformal primaries and no descendants. The (non-)decoupling of the Nambu-Goldstone mode is explicitly shown in examples of scale invariant theories that are actually (not) conformal.
high energy physics theory
Lattice defects play a key role in determining the properties of crystalline materials. Probing the 3D lattice strains that govern their interactions remains a challenge. Bragg Coherent Diffraction Imaging (BCDI) allows strain to be measured with nano-scale 3D resolution. However, it is currently limited to materials that form micro-crystals. Here we introduce a new technique that allows the manufacture of BCDI samples from bulk materials. Using tungsten as an example, we show that focussed ion beam (FIB) machining can be used to extract, from macroscopic crystals, micron-sized BCDI samples containing specific pre-selected defects. To interpret the experimental data, we develop a new displacement-gradient-based analysis for multi-reflection BCDI. This allows accurate recovery of the full lattice strain tensor from samples containing multiple dislocations. These new capabilities open the door to BCDI as a microscopy tool for studying complex real-world materials.
condensed matter
In this white paper, we recommend the European Space Agency plays a proactive role in developing a global collaborative effort to construct a large high-contrast imaging space telescope, e.g. as currently under study by NASA. Such a mission will be needed to characterize a sizable sample of temperate Earth-like planets in the habitable zones of nearby Sun-like stars and to search for extraterrestrial biological activity. We provide an overview of relevant European expertise, and advocate ESA to start a technology development program towards detecting life outside the Solar system.
astrophysics
The vertical component of the magnetic field ($B_\perp$) was found to reach a constant value at the umbra-penumbra boundary of stable sunspots in a recent statistical study of Hinode/SP data. The objective of this work is to verify the existence of a constant value for $B_\perp$ at the umbra-penumbra boundary from ground-based data in the near-infrared wavelengths and to determine its value for GREGOR Infrared Spectrograph (GRIS) data. This is the first statistical study on the Jurcak criterion with ground-based data, and we compare it with the results from Hinode and HMI data. Eleven spectropolarimetric data sets from the GRIS slit-spectograph containing stable sunspots were selected from the GRIS archive (sdc.leibniz-kis.de). SIR inversions including a polarimetric straylight correction are used to produce maps of the magnetic field vector using the Fe 1564.8 and 1566.2 nm lines. Averages of $B_\perp$ along umbra-penumbra boundaries are analyzed for the 11 data sets. Geometric differences between contours at the resulting $B_\perp^{\rm const}$ value and contours in intensity, $\Delta P$, are calculated. Averaged over the 11 sunspots, we find a value of $B_\perp=1787\pm100$ gauss. Contours at $B_\perp=B_\perp^{\rm const}$ and contours calculated in intensity maps match from a visual inspection and the geometric distance $\Delta P$ was found to be on the order of 2 pixels. Furthermore, the standard deviation between different data sets of averages along umbra--penumbra contours is smaller for $B_\perp$ than for $B_\parallel$ by a factor of 2.4. Our results provide further support to the Jurcak criterion with the existence of an invariable value $B_\perp^{\rm const}$ at the umbra--penumbra boundary. We also found that the geometric difference, $\Delta P$, between intensity contours and contours at $B_\perp=B_\perp^{\rm const}$ acts as an index of stability for sunspots.
astrophysics
Boosting is a popular machine learning algorithm in regression and classification problems. Boosting can combine a sequence of regression trees to obtain accurate prediction. In the presence of outliers, traditional boosting, based on optimizing convex loss functions, may show inferior results. In this article, a unified robust boosting is proposed for more resistant estimation. The method utilizes a recently developed concave-convex family for robust estimation, composite optimization by conjugation operator, and functional decent boosting. As a result, an iteratively reweighted boosting algorithm can be conveniently constructed with existing software. Applications in robust regression, classification and Poisson regression are demonstrated in the R package ccboost.
statistics
We propose a first-order autoregressive model for dynamic network processes in which edges change over time while nodes remain unchanged. The model depicts the dynamic changes explicitly. It also facilitates simple and efficient statistical inference such as the maximum likelihood estimators which are proved to be (uniformly) consistent and asymptotically normal. The model diagnostic checking can be carried out easily using a permutation test. The proposed model can apply to any network processes with various underlying structures but with independent edges. As an illustration, an autoregressive stochastic block model has been investigated in depth, which characterizes the latent communities by the transition probabilities over time. This leads to a more effective spectral clustering algorithm for identifying the latent communities. Inference for a change point is incorporated into the autoregressive stochastic block model to cater for possible structure changes. The developed asymptotic theory as well as the simulation study affirms the performance of the proposed methods. Application with three real data sets illustrates both relevance and usefulness of the proposed models.
statistics
In this paper, unmanned aerial vehicles (UAVs) and intelligent reflective surface (IRS) are utilized to support terahertz (THz) communications. To this end, the joint optimization of UAV's trajectory, the phase shift of IRS, the allocation of THz sub-bands, and the power control is investigated to maximize the minimum average achievable rate of all the users. An iteration algorithm based on successive Convex Approximation with the Rate constraint penalty (CAR) is developed to obtain UAV's trajectory, and the IRS phase shift is formulated as a closed-form expression with introduced pricing factors. Simulation results show that the proposed scheme significantly enhances the rate performance of the whole system.
electrical engineering and systems science
Using high-resolution player tracking data made available by the National Football League (NFL) for their 2019 Big Data Bowl competition, we introduce the Expected Hypothetical Completion Probability (EHCP), a objective framework for evaluating plays. At the heart of EHCP is the question "on a given passing play, did the quarterback throw the pass to the receiver who was most likely to catch it?" To answer this question, we first built a Bayesian non-parametric catch probability model that automatically accounts for complex interactions between inputs like the receiver's speed and distances to the ball and nearest defender. While building such a model is, in principle, straightforward, using it to reason about a hypothetical pass is challenging because many of the model inputs corresponding to a hypothetical are necessarily unobserved. To wit, it is impossible to observe how close an un-targeted receiver would be to his nearest defender had the pass been thrown to him instead of the receiver who was actually targeted. To overcome this fundamental difficulty, we propose imputing the unobservable inputs and averaging our model predictions across these imputations to derive EHCP. In this way, EHCP can track how the completion probability evolves for each receiver over the course of a play in a way that accounts for the uncertainty about missing inputs.
statistics
We prove existence of all possible bi-axisymmetric near-horizon geometries of 5-dimensional minimal supergravity. These solutions possess the cross-sectional horizon topology $S^3$, $S^1\times S^2$, or $L(p,q)$ and come with prescribed electric charge, two angular momenta, and a dipole charge (in the ring case). Moreover, we establish uniqueness of these solutions up to an isometry of the symmetric space $G_{2(2)}/SO(4)$.
high energy physics theory
Nonreciprocal devices are indispensable for building quantum networks and ubiquitous in modern communication technology. Here, we use optomechanical interaction and linearly-coupled interaction to realize optical nonreciprocal transmission in a double-cavity optomechanical system. The scheme relies on the interference between the two interactions. We derive the essential conditions to realize perfect optical nonreciprocity in the system, and analyse the properties of optical nonreciprocal transmission and the output fields from mechanical mode. These results can be used to control optical transmission in quantum information processing.
quantum physics
Fully inverted atoms placed at exactly the same location synchronize as they deexcite, and light is emitted in a burst (known as "Dicke's superradiance"). We investigate the role of finite interatomic separation on correlated decay in mesoscopic chains, and provide an understanding in terms of collective jump operators. We show that the superradiant burst survives at small distances, despite Hamiltonian dipole-dipole interactions. However, for larger separations, competition between different jump operators leads to dephasing, suppressing superradiance. Collective effects are still significant for arrays with lattice constants of the order of a wavelength, and lead to a photon emission rate that decays nonexponentially in time. We calculate the two-photon correlation function and demonstrate that emission is correlated and directional, as well as sensitive to small changes in the interatomic distance. These features can be measured in current experimental setups, and are robust to realistic imperfections.
quantum physics
Minimal paths are regarded as a powerful and efficient tool for boundary detection and image segmentation due to its global optimality and the well-established numerical solutions such as fast marching method. In this paper, we introduce a flexible interactive image segmentation model based on the Eikonal partial differential equation (PDE) framework in conjunction with region-based homogeneity enhancement. A key ingredient in the introduced model is the construction of local geodesic metrics, which are capable of integrating anisotropic and asymmetric edge features, implicit region-based homogeneity features and/or curvature regularization. The incorporation of the region-based homogeneity features into the metrics considered relies on an implicit representation of these features, which is one of the contributions of this work. Moreover, we also introduce a way to build simple closed contours as the concatenation of two disjoint open curves. Experimental results prove that the proposed model indeed outperforms state-of-the-art minimal paths-based image segmentation approaches.
computer science
Indeterminacy associated with the probing of quantum state is commonly expressed via spectral distances (metric) featured in the outcomes of repeated experiments. Here we treat it as an effective abundance (measure) of the distinct outcomes instead. The resulting $\mu$-uncertainties are described by the effective number theory [1] whose central result, the existence of a minimal count, leads to a well-defined notion of intrinsic (irremovable) uncertainty. We derive $\mu$-uncertainty formulas for arbitrary set of commuting operators, including the cases with continuous spectra. The associated entropy-like characteristics ($\mu$-entropies) convey how many degrees of freedom are effectively involved in a given measurement process. In order to construct quantum $\mu$-entropies, we are led to quantum effective numbers designed to count independent (mutually orthogonal) states effectively comprising a density matrix. This concept is basis-independent and provides for a measure-based characterization of entanglement.
quantum physics
In their standard form Gaussian processes (GPs) provide a powerful non-parametric framework for regression and classificaton tasks. Their one limiting property is their $\mathcal{O}(N^{3})$ scaling where $N$ is the number of training data points. In this paper we present a framework for GP training with sequential selection of training data points using an intuitive selection metric. The greedy forward selection strategy is devised to target two factors - regions of high predictive uncertainty and underfit. Under this technique the complexity of GP training is reduced to $\mathcal{O}(M^{3})$ where $(M \ll N)$ if $M$ data points (out of $N$) are eventually selected. The sequential nature of the algorithm circumvents the need to invert the covariance matrix of dimension $N \times N$ and enables the use of favourable matrix inverse update identities. We outline the algorithm and sequential updates to the posterior mean and variance. We demonstrate our method on selected one dimensional functions and show that the loss in accuracy due to using a subset of data points is marginal compared to the computational gains.
statistics
We explore how to improve the hybrid model description of the particles originating from the wake that a jet produced in a heavy ion collision leaves in the droplet of quark-gluon plasma (QGP) through which it propagates, using linearized hydrodynamics on a background Bjorken flow. Jet energy and momentum loss described by the hybrid model become currents sourcing linearized hydrodynamics. By solving the linearized hydrodynamic equations numerically, we investigate the development of the wake in the dynamically evolving droplet of QGP, study the effect of viscosity, scrutinize energy-momentum conservation, and check the validity of the linear approximation. We find that linearized hydrodynamics works better in the viscous case because diffusive modes damp the energy-momentum perturbation produced by the jet. We calculate the distribution of particles produced from the jet wake by using the Cooper-Frye prescription and find that both the transverse momentum spectrum and the distribution of particles in azimuthal angle are similar in shape in linearized hydrodynamics and in the hybrid model. Their normalizations are different because the momentum-rapidity distribution in the linearized hydrodynamics analysis is more spread out, due to sound modes. Since the Bjorken flow has no transverse expansion, we explore the effect of transverse flow by using local boosts to add it into the Cooper-Frye formula. After including the effects of transverse flow in this way, the transverse momentum spectrum becomes harder: more particles with transverse momenta bigger than $2$ GeV are produced than in the hybrid model. Although we defer implementing this analysis in a jet Monte Carlo, as would be needed to make quantitative comparisons to data, we gain a qualitative sense of how the jet wake may modify jet observables by computing proxies for two example observables: the lost energy recovered in a cone of varying open angle, and the fragmentation function. We find that linearized hydrodynamics with transverse flow effects added improves the description of the jet wake in the hybrid model in just the way that comparison to data indicates is needed. Our study illuminates a path to improving the description of the wake in the hybrid model, highlighting the need to take into account the effects of both transverse flow and the broadening of the energy-momentum perturbation in spacetime rapidity on particle production.
high energy physics phenomenology
This paper addresses the practical aspects of quantum algorithms used in numerical integration, specifically their implementation on Noisy Intermediate-Scale Quantum (NISQ) devices. Quantum algorithms for numerical integration utilize Quantum Amplitude Estimation (QAE) (Brassard et al., 2002) in conjunction with Grovers algorithm. However, QAE is daunting to implement on NISQ devices since it typically relies on Quantum Phase Estimation (QPE), which requires many ancilla qubits and controlled operations. To mitigate these challenges, a recently published QAE algorithm (Suzuki et al., 2020), which does not rely on QPE, requires a much smaller number of controlled operations and does not require ancilla qubits. We implement this new algorithm for numerical integration on IBM quantum devices using Qiskit and optimize the circuit on each target device. We discuss the application of this algorithm on two qubits and its scalability to more than two qubits on NISQ devices.
quantum physics
We provide an explicit technical framework for proving very general two-weight commutator estimates in arbitrary parameters. The aim is to both clarify existing literature, which often explicitly focuses on two parameters only, and to extend very recent results to the full generality of arbitrary parameters. More specifically, we study two-weight commutator estimates -- Bloom type estimates -- in the multi-parameter setting involving weighted product BMO and little BMO spaces, and their combinations.
mathematics
The Sun is the most studied of stars and a laboratory of fundamental physics. However, the understanding of our star is stained by the solar modelling problem which can stem from various causes. We combine inversions of sound speed, an entropy proxy and the Ledoux discriminant with the position of the base of the convective zone and the photospheric helium abundance to test combinations of ingredients such as equation of state, abundance and opacity tables. We study the potential of the inversions to constrain ad-hoc opacity modifications and additional mixing in the Sun. We show that they provide constraints on these modifications to the ingredients and that the solar problem likely occurs from various sources and using phase shifts with our approach is the next step to take.
astrophysics
We prove a new upper bound on the second moment of Maass form symmetric square L-functions defined over Gaussian integers. Combining this estimate with the recent result of Balog-Biro-Cherubini-Laaksonen, we improve the error term in the prime geodesic theorem for the Picard manifold.
mathematics
Dedicated surveys using different detection pipelines are being carried out at multiple observatories to find more Fast Radio Bursts (FRBs). Understanding the efficiency of detection algorithms and the survey completeness function is important to enable unbiased estimation of the underlying FRB population properties. One method to achieve end-to-end testing of the system is by injecting mock FRBs in the live data-stream and searching for them blindly. Mock FRB injection is particularly effective for machine-learning-based classifiers, for which analytic characterisation is impractical. We describe a first-of-its-kind implementation of a real-time mock FRB injection system at the upgraded Molonglo Observatory Synthesis Telescope (UTMOST) and present our results for a set of 20,000 mock FRB injections. The injections have yielded clear insight into the detection efficiencies and have provided a survey completeness function for pulse width, fluence and DM. Mock FRBs are recovered with uniform efficiency over the full range of injected DMs, however the recovery fraction is found to be a strong function of the width and Signal-to-Noise (SNR). For low widths ($\lesssim 20$ ms) and high SNR ($\gtrsim$ 9) the recovery is highly effective with recovery fractions exceeding 90%. We find that the presence of radio frequency interference causes the recovered SNR values to be systematically lower by up to 20% compared to the injected values. We find that wider FRBs become increasingly hard to recover for the machine-learning-based classifier employed at UTMOST. We encourage other observatories to implement live injection set-ups for similar testing of their surveys.
astrophysics
Automatic speech recognition (ASR) is a relevant area in multiple settings because it provides a natural communication mechanism between applications and users. ASRs often fail in environments that use language specific to particular application domains. Some strategies have been explored to reduce errors in closed ASRs through post-processing, particularly automatic spell checking, and deep learning approaches. In this article, we explore using a deep neural network to refine the results of a phonetic correction algorithm applied to a telesales audio database. The results exhibit a reduction in the word error rate (WER), both in the original transcription and in the phonetic correction, which shows the viability of deep learning models together with post-processing correction strategies to reduce errors made by closed ASRs in specific language domains.
electrical engineering and systems science
Motivated by the aero-acoustic feedback loop phenomenon in high speed free jets and impinging jets, a thorough examination of a POD (Proper Orthogonal Decomposition)-Galerkin method to determine the average convection velocity of coherent structures in the shear layer is presented in this paper. The technique is shown to be applicable to both time resolved as well as time unresolved data, if the data set meets certain requirements. Using a detailed sensitivity analysis on a synthetic data set, a quantitative estimate on the required time resolution for the technique has been found, which can be useful for both experimental, as well as numerical studies investigating the aero-acoustic feedback loop in high speed flows. Moreover, some innovative ways to apply the technique are also demonstrated using a simulated data set, showing the effectiveness of the technique to any general problem in supersonic jets, heat transfer, combustion or other areas in fluid mechanics, where an advection process can be identified.
physics
We present the results of numerical simulations and experimental studies about the effects of resonant and random excitations on proton losses, emittances, and beam distributions in the Large Hadron Collider (LHC). In addition to shedding light on complex nonlinear effects, these studies are applied to the design of hollow electron lenses (HEL) for active beam halo control. In the High-Luminosity Large Hadron Collider (HL-LHC), a considerable amount of energy will be stored in the beam tails. To control and clean the beam halo, the installation of two hollow electron lenses, one per beam, is being considered. In standard electron-lens operation, a proton bunch sees the same electron current at every revolution. Pulsed electron beam operation (i.e., different currents for different turns) is also considered, because it can widen the range of achievable halo removal rates. For an axially symmetric electron beam, only protons in the halo are excited. If a residual field is present at the location of the beam core, these particles are exposed to time-dependent transverse kicks and to noise. We discuss the numerical simulations and the experiments conducted in 2016 and 2017 at injection energy in the LHC. The excitation patterns were generated by the transverse feedback and damping system, which acted as a flexible source of dipole kicks. Proton beam losses, emittances, and transverse distributions were recorded as a function of excitation patterns and strengths. The resonant excitations induced rich dynamical effects and nontrivial changes of the beam distributions, which, to our knowledge, have not previously been observed and studied in this detail. We conclude with a discussion of the tolerable and achievable residual fields and proposals for further studies.
physics
We propose a fast potential splitting Markov Chain Monte Carlo method which costs $O(1)$ time each step for sampling from equilibrium distributions (Gibbs measures) corresponding to particle systems with singular interacting kernels. We decompose the interacting potential into two parts, one is of long range but is smooth, and the other one is of short range but may be singular. To displace a particle, we first evolve a selected particle using the stochastic differential equation (SDE) under the smooth part with the idea of random batches, as commonly used in stochastic gradient Langevin dynamics. Then, we use the short range part to do a Metropolis rejection. Different from the classical Langevin dynamics, we only run the SDE dynamics with random batch for a short duration of time so that the cost in the first step is $O(p)$, where $p$ is the batch size. The cost of the rejection step is $O(1)$ since the interaction used is of short range. We justify the proposed random-batch Monte Carlo method, which combines the random batch and splitting strategies, both in theory and with numerical experiments. While giving comparable results for typical examples of the Dyson Brownian motion and Lennard-Jones fluids, our method can save more time when compared to the classical Metropolis-Hastings algorithm.
physics
We propose Regularized Learning under Label shifts (RLLS), a principled and a practical domain-adaptation algorithm to correct for shifts in the label distribution between a source and a target domain. We first estimate importance weights using labeled source data and unlabeled target data, and then train a classifier on the weighted source samples. We derive a generalization bound for the classifier on the target domain which is independent of the (ambient) data dimensions, and instead only depends on the complexity of the function class. To the best of our knowledge, this is the first generalization bound for the label-shift problem where the labels in the target domain are not available. Based on this bound, we propose a regularized estimator for the small-sample regime which accounts for the uncertainty in the estimated weights. Experiments on the CIFAR-10 and MNIST datasets show that RLLS improves classification accuracy, especially in the low sample and large-shift regimes, compared to previous methods.
computer science
In order to describe few-body scattering in the case of the Coulomb interaction, an approach based on splitting the reaction potential into a finite range part and a long range tail part is presented. The solution to the Schr\"odinger equation for the long range tail is used as an incoming wave in an inhomogeneous Schr\"odinger equation with the finite range potential. The resulting equation with asymptotic outgoing waves is then solved with the exterior complex scaling. The potential splitting approach is illustrated with calculations of scattering processes in the H${}^+$ -- H${}^+_2$ system considered as the three-body system with one-state electronic potential surface.
physics
We investigate finite and infinite nested square root formulas convergent to unity.
mathematics