text
stringlengths
11
9.77k
label
stringlengths
2
104
We propose a two qubit experiment for validating tunable antiferromagnetic $XX$ interactions in quantum annealing. Such interactions allow the time-dependent Hamiltonian to be non-stoquastic, and the instantaneous ground state can have negative amplitudes in the computational basis. Our construction relies on how the degeneracy of the Ising Hamiltonian's ground states is broken away from the end point of the anneal: above a certain value of the antiferromagnetic $XX$ interaction strength, the perturbative ground state at the end of the anneal changes from a symmetric to an antisymmetric state. This change is associated with a suppression of one of the Ising ground states, which can then be detected using solely computational basis measurements. We show that a semiclassical approximation of the annealing protocol fails to reproduce this feature, making it a candidate `quantum signature' of the evolution.
quantum physics
The Orthogonal Frequency Division Multiplexing (OFDM) is one of the most widely adopted schemes in wireless technologies such as Wi-Fi and LTE due to its high transmission rates, and the robustness against Intersymbol Interference (ISI). However, OFDM is highly sensitive to synchronism errors, which affects the orthogonality of the carriers. We analyzed several synchronization algorithms based on the correlation of the preamble symbols through the implementation in Software-Defined Radio (SDR) using the Universal Software Radio Peripheral (USRP). Such an implementation was performed in three stages: frame detection, comparing the autocorrelation output and the average power of the received signal; time synchronism, where the cross-correlation based on the short and long preamble symbols was implemented; and the frequency synchronism, where the Carrier Frequency Offset (CFO) added by the channel was detected and corrected. The synchronizer performance was verified through the USRP implementation. The results serve as a practical guide to selecting the optimal synchronism scheme and show the versatility of the USRP to implement digital communication systems efficiently.
electrical engineering and systems science
We study the interplay between mass deformations and unoriented projections of super-conformal quiver gauge theories resulting from D3-branes at (toric) Calabi-Yau singularities. We focus on simple orbifold cases ($\mathbb{C}^3/\mathbb{Z}_3$ and $\mathbb{C}^3/\mathbb{Z}_4$) and their non-orbifold descendants. This allows us to generalize the construction rules and clarify points that have been previously overlooked. In particular we spell out the conditions of anomaly cancellations as well as super-conformal invariance that typically require the introduction of flavour branes, which in turn may spoil toric symmetry. Finally, we discuss duality cascades in this context and the interplay between Seiberg/toric duality and unoriented projection with (or without) mass deformations.
high energy physics theory
The coding space in quantum communication could be expanded to high-dimensional space by using orbital angular momentum (OAM) states of photons, as both the capacity of the channel and security are enhanced. Here we present a novel approach to realize high-capacity quantum key distribution (QKD) by exploiting OAM states. The innovation of the proposed approach relies on a unique type of entangledphoton source which produces entangled photons with OAM randomly distributed among high order Fiboncci-like numbers and a new physical mechanism for efficiently sharing keys. This combination of entanglement with mathematical properties of high order Fibonacci sequences provides the QKD protocol which is immune to photon-number-splitting attacks and allows secure generation of long keys from few photons. Unlike other protocols, reference frame alignment and active modulation of production and detection bases are unnecessary.
quantum physics
Configuring hybrid precoders and combiners is a major challenge to deploy practical mmWave communication systems. Prior work addresses the problem of designing hybrid precoders and combiner, yet focusing on finding solutions under a total transmit power constraint. The design of hybrid precoders and combiners in practical system, is constrained, however, by a per antenna transmit power, since each antenna element in the array is connected to a power amplifier (PA) that has to operate within its linear region. In this paper, we focus on the problem of hybrid precoding and combining with per-antenna power constraints, and under a frequency-selective bandlimited channel model. We first propose an all-digital solution to this problem, and develop a hybrid precoding and combining strategy that aims at matching this solution by minimizing the chordal distance between the all-digital precoders (combiners) and their hybrid approximations. Finally, since minimizing this metric does not guarantee that the final spectral efficiency will be maximized, we optimize the resulting spectral efficiency taking into account the per-antenna power constraints. Simulation results show the effectiveness of our all-digital and hybrid solutions, while emphasizing the differences with respect to the corresponding solution under a total power constraints. As shown in our numerical results, the proposed all-digital solution performs similarly to the case in which a total power constraint is considered. Further, our proposed hybrid solution is also shown to exhibit near-optimum performance, and the influence of different system parameters is also shown, thereby showing the suitability of our proposed framework to deploy practical mmWave MIMO systems.
electrical engineering and systems science
The variational autoencoder is a well defined deep generative model that utilizes an encoder-decoder framework where an encoding neural network outputs a non-deterministic code for reconstructing an input. The encoder achieves this by sampling from a distribution for every input, instead of outputting a deterministic code per input. The great advantage of this process is that it allows the use of the network as a generative model for sampling from the data distribution beyond provided samples for training. We show in this work that utilizing batch normalization as a source for non-determinism suffices to turn deterministic autoencoders into generative models on par with variational ones, so long as we add a suitable entropic regularization to the training objective.
computer science
We propose an efficient transfer Bayesian optimization method, which finds the maximum of an expensive-to-evaluate black-box function by using data on related optimization tasks. Our method uses auxiliary information that represents the task characteristics to effectively transfer knowledge for estimating a distribution over target functions. In particular, we use a Gaussian process, in which the mean and covariance functions are modeled with neural networks that simultaneously take both the auxiliary information and feature vectors as input. With a neural network mean function, we can estimate the target function even without evaluations. By using the neural network covariance function, we can extract nonlinear correlation among feature vectors that are shared across related tasks. Our Gaussian process-based formulation not only enables an analytic calculation of the posterior distribution but also swiftly adapts the target function to observations. Our method is also advantageous because the computational costs scale linearly with the number of source tasks. Through experiments using a synthetic dataset and datasets for finding the optimal pedestrian traffic regulations and optimal machine learning algorithms, we demonstrate that our method identifies the optimal points with fewer target function evaluations than existing methods.
statistics
Extra-large massive multiple-input multiple-output (XL-MIMO) systems is a new concept, where spatial non-stationarities allow activate a high number of user equipments (UEs). This paper focuses on a grant-based random access (RA) approach in the novel XL-MIMO channel scenarios. Modifications in the classical Strongest User Collision Resolution (SUCRe) protocol have been aggregated to explore the visibility regions (VRs) overlapping in XL-MIMO. The proposed grant-based RA protocol takes advantage of this new degree of freedom for improving the number of access attempts and accepted UEs. As a result, the proposed grant-based protocol for XL-MIMO systems is capable of reducing latency in the pilot allocation step.
electrical engineering and systems science
A typical f-electron Kondo lattice system Ce exhibits the well-known isostructural transition, the so-called gamma-alpha transition, accompanied by an enormous volume collapse. Most interestingly, we have discovered that a topological-phase transition also takes place in elemental Ce, concurrently with the gamma-alpha transition. Based on the dynamical mean-field theory approach combined with density functional theory, we have unravelled that the non-trivial topology in alpha-Ce is driven by the f-d band inversion, which arises from the formation of coherent 4f band around the Fermi level. We captured the formation of the 4f quasi-particle band that is responsible for the Lifshitz transition and the non-trivial Z2 topology establishment across the phase boundary. This discovery provides a concept of 'topology switch' for topological Kondo systems. The 'on' and 'off' switching knob in Ce is versatile in a sense that it is controlled by available pressure (around 1 GPa) at room temperature.
condensed matter
We address the nature of phase transitions in periodically driven systems coupled to a bath. The latter enables a synchronized non-equilibrium Floquet steady state at finite entropy, which we analyse for rapid drives within a non-equilibrium RG approach. While the infinitely rapidly driven limit exhibits a second order phase transition, here we reveal that fluctuations turn the transition first order when the driving frequency is finite. This can be traced back to a universal mechanism, which crucially hinges on the competition of degenerate, near critical modes associated to higher Floquet Brillouin zones. The critical exponents of the infinitely rapidly driven system -- including a new, independent one -- can yet be probed experimentally upon smoothly tuning towards that limit.
condensed matter
An omega-meson extension of the Skyrme model - without the Skyrme term but including the pion mass - first considered by Adkins and Nappi is studied in detail for baryon numbers 1 to 8. The static problem is reformulated as a constrained energy minimisation problem within a natural geometric framework and studied analytically on compact domains, and numerically on Euclidean space. Using a constrained second-order Newton flow algorithm, classical energy minimisers are constructed for various values of the omega-pion coupling. At high coupling, these Skyrmion solutions are qualitatively similar to the Skyrmions of the standard Skyrme model with massless pions. At sufficiently low coupling, they show similarities with those in the lightly bound Skyrme model: the Skyrmions of low baryon number dissociate into lightly bound clusters of distinct 1-Skyrmions, and the classical binding energies for baryon numbers 2 through 8 have realistic values.
high energy physics theory
A sensitivity analysis in an observational study tests whether the qualitative conclusions of an analysis would change if we were to allow for the possibility of limited bias due to confounding. The design sensitivity of a hypothesis test quantifies the asymptotic performance of the test in a sensitivity analysis against a particular alternative. We propose a new, non-asymptotic, distribution-free test, the uniform general signed rank test, for observational studies with paired data, and examine its performance under Rosenbaum's sensitivity analysis model. Our test can be viewed as adaptively choosing from among a large underlying family of signed rank tests, and we show that the uniform test achieves design sensitivity equal to the maximum design sensitivity over the underlying family of signed rank tests. Our test thus achieves superior, and sometimes infinite, design sensitivity, indicating it will perform well in sensitivity analyses on large samples. We support this conclusion with simulations and a data example, showing that the advantages of our test extend to moderate sample sizes as well.
statistics
With growing interest in mmWave connectivity for UAVs, a basic question is whether networks intended for terrestrial users can provide sufficient aerial coverage as well. To assess this possibility, the paper proposes a novel evaluation methodology using generative models trained on detailed ray tracing data. These models capture complex propagation characteristics and can be readily combined with antenna and beamforming assumptions. Extensive simulation using these models indicate that standard (street-level and downtilted) base stations at typical microcellular densities can indeed provide satisfactory UAV coverage. Interestingly, the coverage is possible via a conjunction of antenna sidelobes and strong reflections. With sparser deployments, the coverage is only guaranteed at progressively higher altitudes. Additional dedicated (rooftop-mounted and uptilted) base stations strengthen the coverage provided that their density is comparable to that of the standard deployment, and would be instrumental for sparse deployments of the latter.
electrical engineering and systems science
We have analyzed the behavior of a mobile quantum impurity in a bath formed by a two-leg bosonic ladder by a combination of field theory (Tomonaga-Luttinger liquid) and numerical (Density Matrix Renormalization Group) techniques. Computing the Green's function of the impurity as a function of time at different momenta, we find a power law decay at zero momentum, which signals the breakdown of any quasi-particle description of the impurity motion. We compute the exponent both for the limits of weak and strong impurity-bath interactions. At small impurity-bath interaction, we find that the impurity experiences the ladder as a single channel one-dimensional bath, but effective coupling is reduced by a factor of $\sqrt 2$, thus the impurity is less mobile in the ladder compared to a one dimensional bath. We compared the numerical results for the exponent at zero momentum with a semi-analytical expression that was initially established for the chain and find excellent agreement without adjustable parameters. We analyze the dependence of the exponent in the transverse hopping in the bath and find surprisingly an increase of the exponent at variance with the naive extrapolation of the single channel regime. We study the momentum dependence of the impurity Green's function and find that, as for the single chain, two different regime of motion exist, one dominated by infrared metatrophy and a more conventional polaronic behavior. We compute the critical momentum between these two regimes and compare with prediction based on the structure factor of the bath. In the polaronic regime we also compute numerically the lifetime of the polaron. Finally we discuss how our results could be measured in cold atomic experiments.
condensed matter
We study Jordan-Lie inner ideals of finite dimensional associative algebras and the corresponding Lie algebras and prove that they admit Levi decompositions. Moreover, we classify Jordan-Lie inner ideals satisfying a certain minimality condition and show that they are generated by pairs of idempotents.
mathematics
The induction hardening behavior of a new, hot-rolled 0.4 wt.% carbon steel with the two different starting microstructures of upper and lower bainite has been simulated using a Gleeble 3800. The effect of heating rate in the range 1 - 500 {\deg}C/s on austenite grain size distribution has been rationalized. Dilatometry together with Scanning Electron Microscopy combined with Electron Backscatter Diffraction analyses and thermodynamic simulations provide insight into the austenite formation mechanisms that operate at different heating rates. Two main mechanisms of austenite formation during re-austenitization were identified: diffusional and diffusionless (massive). At conventional (1-5 {\deg}C/s) and fast (10-50 {\deg}C/s) heating rates the austenite formation mechanism and kinetics are controlled by diffusion, whereas at ultrafast heating rates (100-500 {\deg}C/s) the formation of austenite starts by diffusion control, but is later overtaken by a massive transformation mechanism. Comprehensive thermodynamic descriptions of the influence of cementite on austenite formation are discussed. The finest austenite grain size and the highest final hardness are achieved with a lower bainite starting microstructure processed with a heating rate of 50 {\deg}C/s to an austenitization temperature of 850 {\deg}C followed by cooling at 60 {\deg}C/s. Keywords Induction Hardening, Heating Rate, Cementite Dissolution, Prior Austenite Grain Size, Dilatometry
condensed matter
Pd-intercalated ErTe$_3$ is studied as a model system to explore the effect of "intertwined" superconducting and charge density wave (CDW) orders. Despite the common wisdom that superconductivity emerges only when CDW is suppressed, we present data from STM and AC susceptibility measurements that show no direct competition between CDW order and superconductivity. Both coexist over most of the intercalation range, with uniform superconductivity over length scales that exceed the superconducting coherence length. This is despite persisting short-range CDW order and increased scattering from the Pd intercalation. While superconductivity is insensitive to local defects in either of the bi-directional CDWs, vestiges of the Fermi-level distortions are observed in the properties of the superconducting state.
condensed matter
We give a computational interpretation to an abstract instance of Zorn's lemma formulated as a wellfoundedness principle in the language of arithmetic in all finite types. This is achieved through G\"odel's functional interpretation, and requires the introduction of a novel form of recursion over non-wellfounded partial orders whose existence in the model of total continuous functionals is proven using domain theoretic techniques. We show that a realizer for the functional interpretation of open induction over the lexicographic ordering on sequences follows as a simple application of our main results.
computer science
Can deep learning solve multiple tasks simultaneously, even when they are unrelated and very different? We investigate how the representations of the underlying tasks affect the ability of a single neural network to learn them jointly. We present theoretical and empirical findings that a single neural network is capable of simultaneously learning multiple tasks from a combined data set, for a variety of methods for representing tasks -- for example, when the distinct tasks are encoded by well-separated clusters or decision trees over certain task-code attributes. More concretely, we present a novel analysis that shows that families of simple programming-like constructs for the codes encoding the tasks are learnable by two-layer neural networks with standard training. We study more generally how the complexity of learning such combined tasks grows with the complexity of the task codes; we find that combining many tasks may incur a sample complexity penalty, even though the individual tasks are easy to learn. We provide empirical support for the usefulness of the learning bounds by training networks on clusters, decision trees, and SQL-style aggregation.
computer science
We report a family of two-dimensional hybrid perovskites (2DHPs) based on phenethylammonium lead iodide ((PEA)$_2$PbI$_4$) that show complex structure in their low-temperature excitonic absorption and photoluminescence (PL) spectra as well as hot exciton PL. We replace the 2-position (ortho) H on the phenyl group of the PEA cation with F, Cl, or Br to systematically increase the cation's cross-sectional area and mass and study changes in the excitonic structure. These single atom substitutions substantially change the observable number of and spacing between discrete resonances in the excitonic absorption and PL spectra and drastically increase the amount of hot exciton PL that violates Kasha's rule by over an order of magnitude. To fit the progressively larger cations, the inorganic framework distorts and is strained, reducing the Pb-I-Pb bond angles and increasing the 2DHP band gap. Correlation between the 2DHP structure and steady-state and time-resolved spectra suggests the complex structure of resonances arises from one or two manifolds of states, depending on the 2DHP Pb-I-Pb bond angle (as)symmetry, and the resonances within a manifold are regularly spaced with an energy separation that decreases as the mass of the cation increases. The uniform separation between resonances and the dynamics that show excitons can only relax to the next-lowest state are consistent with a vibronic progression caused by a vibrational mode on the cation. These results demonstrate that simple changes to the structure of the cation can be used to tailor the properties and dynamics of the confined excitons without directly modifying the inorganic framework.
physics
The propagation of acoustic or elastic waves in artificial crystals, including the case of phononic and sonic crystals, is inherently anisotropic. As is known from the theory of periodic composites, anisotropy is directly dictated by the space group of the unit cell of the crystal and the rank of the elastic tensor. Here, we examine effective velocities in the long wavelength limit of periodic acoustic and elastic composites as a function of the direction of propagation. We derive explicit and efficient formulas for estimating the effective velocity surfaces, based on second-order perturbation theory, generalizing the Christofell equation for elastic waves in solids. We identify strongly anisotropic sonic crystals for scalar acoustic waves and strongly anisotropic phononic crystals for vector elastic waves. Furthermore, we observe that under specific conditions, quasi-longitudinal waves can be made much slower than shear waves propagating in the same direction.
physics
Time-varying optical media, whose dielectric properties are actively modulated in time, introduce a host of novel effects in the classical propagation of light, and are of intense current interest. In the quantum domain, time-dependent media can be used to convert vacuum fluctuations (virtual photons) into pairs of real photons. We refer to these processes broadly as ``dynamical vacuum effects'' (DVEs). Despite interest for their potential applications as sources of quantum light, DVEs are generally very weak, providing many opportunities for enhancement through modern techniques in nanophotonics, such as using media which support excitations such as plasmon and phonon polaritons. Here, we present a theory of DVEs in arbitrary nanostructured, dispersive, and dissipative systems. A key element of our framework is the simultaneous incorporation of time-modulation and ``dispersion'' through time-translation-breaking linear response theory. We propose a highly efficient scheme for generating entangled surface polaritons based on time-modulation of the optical phonon frequency of a polar insulator. We show that the high density of states, especially in hyperbolic polaritonic media, may enable high-efficiency generation of entangled phonon-polariton pairs. More broadly, our theoretical framework enables the study of quantum light-matter interactions in time-varying media, such as spontaneous emission, and energy level shifts.
physics
The singlet majoron model of seesaw neutrino mass is appended by one dark Majorana fermion singlet $\chi$ with $L=2$ and one dark complex scalar singlet $\zeta$ with $L=1$. This simple setup allows $\chi$ to obtain a small radiative mass anchored by the same heavy right-handed neutrinos, whereas the one-loop decay of the standard-model Higgs boson to $\chi \chi + \bar{\chi} \bar{\chi}$ provides the freeze-in mechanism for $\chi$ to be the light dark matter of the Universe.
high energy physics phenomenology
We study how emojis are used to express solidarity in social media in the context of two major crisis events - a natural disaster, Hurricane Irma in 2017 and terrorist attacks that occurred on November 2015 in Paris. Using annotated corpora, we first train a recurrent neural network model to classify expressions of solidarity in text. Next, we use these expressions of solidarity to characterize human behavior in online social networks, through the temporal and geospatial diffusion of emojis. Our analysis reveals that emojis are a powerful indicator of sociolinguistic behaviors (solidarity) that are exhibited on social media as the crisis events unfold.
computer science
Purpose: Manual medical image segmentation is an exhausting and time-consuming task along with high inter-observer variability. In this study, our objective is to improve the multi-resolution image segmentation performance of U-Net architecture. Approach: We have proposed a fully convolutional neural network for image segmentation in a multi-resolution framework. We used U-Net as the base architecture and modified that to improve its image segmentation performance. In the proposed architecture (mrU-Net), the input image and its down-sampled versions were used as the network inputs. We added more convolution layers to extract features directly from the down-sampled images. We trained and tested the network on four different medical datasets, including skin lesion photos, lung computed tomography (CT) images (LUNA dataset), retina images (DRIVE dataset), and prostate magnetic resonance (MR) images (PROMISE12 dataset). We compared the performance of mrU-Net to U-Net under similar training and testing conditions. Results: Comparing the results to manual segmentation labels, mrU-Net achieved average Dice similarity coefficients of 70.6%, 97.9%, 73.6%, and 77.9% for the skin lesion, LUNA, DRIVE, and PROMISE12 segmentation, respectively. For the skin lesion, LUNA, and DRIVE datasets, mrU-Net outperformed U-Net with significantly higher accuracy and for the PROMISE12 dataset, both networks achieved similar accuracy. Furthermore, using mrU-Net led to a faster training rate on LUNA and DRIVE datasets when compared to U-Net. Conclusions: The striking feature of the proposed architecture is its higher capability in extracting image-derived features compared to U-Net. mrU-Net illustrated a faster training rate and slightly more accurate image segmentation compared to U-Net.
electrical engineering and systems science
Plasma wakefield acceleration is a method for accelerating particle beams using electromagnetic fields that are orders of magnitude larger than those found in conventional radio frequency cavities. The core component of a plasma wakefield accelerator is the plasma source, which ranges from millimeter-scale gas jets used in laser-driven experiments, to the ten-meter-long rubidium cell used in the AWAKE experiment. The density of the neutral gas is a controlled input to the experiment, but the density of the plasma after ionization depends on many factors. AWAKE uses a high-energy proton beam to drive the plasma wakefield, and the wakefield acts back on the proton bunch by modulating it at the plasma frequency. We infer the plasma density by measuring the frequency of modulation of the proton bunch, and we measure the evolution of the density versus time by varying the arrival of the proton beam with respect to the ionizing laser pulse. Using this technique, we uncover a microsecond-long period of a stable plasma density followed by a rapid decay in density. The stability of the plasma after ionization has implications for the design of much longer vapor cells that could be used to accelerate particle beams to extremely high energies.
physics
Employing uplift formulae, we uplift supersymmetric $AdS_6$ black holes from $F(4)$ gauged supergravity to massive type IIA and type IIB supergravity. In massive type IIA supergravity, we obtain supersymmetric $AdS_6$ black holes asymptotic to the Brandhuber-Oz solution. In type IIB supergravity, we obtain supersymmetric $AdS_6$ black holes asymptotic to the non-Abelian T-dual of the Brandhuber-Oz solution. For the uplifted black hole solutions, we calculate the holographic entanglement entropy. In massive type IIA supergravity, it precisely matches the Bekenstein-Hawking entropy of the black hole solutions.
high energy physics theory
Satellite images are often contaminated by clouds. Cloud removal has received much attention due to the wide range of satellite image applications. As the clouds thicken, the process of removing the clouds becomes more challenging. In such cases, using auxiliary images such as near-infrared or synthetic aperture radar (SAR) for reconstructing is common. In this study, we attempt to solve the problem using two generative adversarial networks (GANs). The first translates SAR images into optical images, and the second removes clouds using the translated images of prior GAN. Also, we propose dilated residual inception blocks (DRIBs) instead of vanilla U-net in the generator networks and use structural similarity index measure (SSIM) in addition to the L1 Loss function. Reducing the number of downsamplings and expanding receptive fields by dilated convolutions increase the quality of output images. We used the SEN1-2 dataset to train and test both GANs, and we made cloudy images by adding synthetic clouds to optical images. The restored images are evaluated with PSNR and SSIM. We compare the proposed method with state-of-the-art deep learning models and achieve more accurate results in both SAR-to-optical translation and cloud removal parts.
electrical engineering and systems science
We argue that impact velocities between dust grains with sizes less than $\sim 0.1$ $\mu m$ in molecular cloud cores are dominated by drift arising from ambipolar diffusion. This effect is due to the size dependence of the dust coupling to the magnetic field and the neutral gas. Assuming perfect sticking in collisions up to $\approx 50$ m/s, we show that this effect causes rapid depletion of small grains - consistent with starlight extinction and IR/microwave emission measurements, both in the core center ($n \sim 10^{6}$ cm$^{-3}$) and envelope ($n \sim 10^{4}$ cm$^{-3}$). The upper end of the size distribution does not change significantly if only velocities arising from this effect are considered. We consider the impact of an evolved dust size distribution on the gas temperature, and argue that if the depletion of small dust grains occurs as would be expected from our model, then the cosmic ray ionization rate must be well below $10^{-16}$ s$^{-1}$ at a number density of $10^{5}$ cm$^{-3}$.
astrophysics
Let $X$ be a compact complex manifold in the Fujiki class $\mathscr{C}$. We study the compactification of $\operatorname{Aut}^0(X)$ given by its closure in Barlet cycle space. The boundary points give rise to non-dominant meromorphic self-maps of $X$. Moreover convergence in cycle space yields convergence of the corresponding meromorphic maps. There are analogous compactifications for reductive subgroups acting trivially on $\operatorname{Alb} X$. If $X$ is K\"ahler, these compactifications are projective. Finally we give applications to the action of $\operatorname{Aut}(X)$ on the set of probability measures on $X$. In particular we obtain an extension of Furstenberg lemma to manifolds in the class $\mathscr{C}$.
mathematics
We introduce a nonequilibrium grand-canonical ensemble defined by considering the stationary state of a driven system of particles put in contact with a nonequilibrium particle reservoir. At odds with its equilibrium counterpart, or with purely formal constructions of a grand-canonical ensemble, this physically-motivated construction yields a grand-canonical distribution that depends on the details of the contact dynamics between the system and the reservoir. For non-interacting driven particles, a grand-canonical chemical potential can still be defined, although this chemical potential now differs from that of the reservoir. However, in the general case, the usual exponential factor (in the particle number) defining the grand-canonical chemical potential, is replaced by the exponential of a non-linear function of the density, this function being proportional to the volume. This case is illustrated explicitly on a one-dimensional lattice model. Although a grand-canonical chemical potential can no longer be defined in this case, it is possible for a subclass of contact dynamics to generalize the equilibrium fluctuation-response relation by introducing a small external potential difference between the system and the reservoir.
condensed matter
Automated medical image classification with convolutional neural networks (CNNs) has great potential to impact healthcare, particularly in resource-constrained healthcare systems where fewer trained radiologists are available. However, little is known about how well a trained CNN can perform on images with the increased noise levels, different acquisition protocols, or additional artifacts that may arise when using low-cost scanners, which can be underrepresented in datasets collected from well-funded hospitals. In this work, we investigate how a model trained to triage head computed tomography (CT) scans performs on images acquired with reduced x-ray tube current, fewer projections per gantry rotation, and limited angle scans. These changes can reduce the cost of the scanner and demands on electrical power but come at the expense of increased image noise and artifacts. We first develop a model to triage head CTs and report an area under the receiver operating characteristic curve (AUROC) of 0.77. We then show that the trained model is robust to reduced tube current and fewer projections, with the AUROC dropping only 0.65% for images acquired with a 16x reduction in tube current and 0.22% for images acquired with 8x fewer projections. Finally, for significantly degraded images acquired by a limited angle scan, we show that a model trained specifically to classify such images can overcome the technological limitations to reconstruction and maintain an AUROC within 0.09% of the original model.
electrical engineering and systems science
In this paper, we investigate a prescribed-time and fully distributed Nash Equilibrium (NE) seeking problem for continuous-time noncooperative games. By exploiting pseudo-gradient play and consensus-based schemes, various distributed NE seeking algorithms are presented over either fixed or switching communication topologies so that the convergence to the NE is reached in a prescribed time. In particular, a prescribed-time distributed NE seeking algorithm is firstly developed under a fixed graph to find the NE in a prior-given and user-defined time, provided that a static controller gain can be selected based on certain global information such as the algebraic connectivity of the communication graph and both the Lipschitz and monotone constants of the pseudo-gradient associated with players' objective functions. Secondly, a prescribed-time and fully distributed NE seeking algorithm is proposed to remove global information by designing heterogeneous dynamic gains that turn on-line the weights of the communication topology. Further, we extend this algorithm to accommodate jointly switching topologies. It is theoretically proved that the global convergence of those proposed algorithms to the NE is rigorously guaranteed in a prescribed time based on a time function transformation approach. In the last, numerical simulation results are presented to verify the effectiveness of the designs.
electrical engineering and systems science
This paper gives a free entropy theoretic perspective on amenable absorption results for free products of tracial von Neumann algebras. In particular, we give the first free entropy proof of Popa's famous result that the generator MASA in a free group factor is maximal amenable, and we partially recover Houdayer's results on amenable absorption and Gamma stability. Moreover, we give a unified approach to all these results using $1$-bounded entropy. We show that if $\mathcal{M} = \mathcal{P} * \mathcal{Q}$, then $\mathcal{P}$ absorbs any subalgebra of $\mathcal{M}$ that intersects it diffusely and that has $1$-bounded entropy zero (which includes amenable and property Gamma algebras as well as many others). In fact, for a subalgebra $\mathcal{P} \leq \mathcal{M}$ to have this absorption property, it suffices for $\mathcal{M}$ to admit random matrix models that have exponential concentration of measure and that "simulate" the conditional expectation onto $\mathcal{P}$.
mathematics
We investigate numerically and theoretically the effect of spatial disorder on two-dimensional split-step discrete-time quantum walks with two internal "coin" states. Spatial disorder can lead to Anderson localization, inhibiting the spread of quantum walks, putting them at a disadvantage against their diffusively spreading classical counterparts. We find that spatial disorder of the most general type, i.e., position-dependent Haar random coin operators, does not lead to Anderson localization but to a diffusive spread instead. This is a delocalization, which happens because disorder places the quantum walk to a critical point between different anomalous Floquet-Anderson insulating topological phases. We base this explanation on the relationship of this general quantum walk to a simpler case more studied in the literature and for which disorder-induced delocalization of a topological origin has been observed. We review topological delocalization for the simpler quantum walk, using time evolution of the wave functions and level spacing statistics. We apply scattering theory to two-dimensional quantum walks and thus calculate the topological invariants of disordered quantum walks, substantiating the topological interpretation of the delocalization and finding signatures of the delocalization in the finite-size scaling of transmission. We show criticality of the Haar random quantum walk by calculating the critical exponent $\eta$ in three different ways and find $\eta$ $\approx$ 0.52 as in the integer quantum Hall effect. Our results showcase how theoretical ideas and numerical tools from solid-state physics can help us understand spatially random quantum walks.
quantum physics
We compute a presentation of the fundamental group of a higher-rank graph using a coloured graph description of higher-rank graphs developed by the third author. We compute the fundamental groups of several examples from the literature. Our results fit naturally into the suite of known geometrical results about $k$-graphs when we show that the abelianisation of fundamental group is the homology group. We end with a calculation which gives a non-standard presentation of the fundamental group of the Klein bottle to the one normally found in the literature.
mathematics
We study the equivalence principle, regarded as the cornerstone of general relativity, by analyzing the deformation observable of black hole shadows. Such deformation can arise from new physics and may be expressed as a phenomenological violation of the equivalence principle. Specifically, we assume that there is an additional background vector field that couples to the photons around the supermassive black hole. This type of coupling yields impact on the way the system depends on initial conditions, and affects the black hole shadow at different wavelengths by a different amount, and therefore observations of the shadow in different wavelengths could constrain such couplings. This can be tested by future multi-band observations. Adopting a specific form of the vector field, we obtain constraints on model parameters from Event Horizon Telescope observations and measurements of gas/stellar orbits.
astrophysics
The recent experimental measurements on $pp$ collisions at $\sqrt{s}=5.02 \,\rm TeV$ have shown a very large abundance of heavy baryon production corresponding to a ratio of $\Lambda_c/D^0 \sim 0.6$, about one order of magnitude larger than what measured in $e^+e^-$, $ep$ collisions and even in $pp$ collisions at LHC, but at forward rapidity. We apply for the first time a quark coalescence plus fragmentation approach, assuming the formation of Hot QCD matter at finite temperature. An approach that have predicted a $\Lambda_c/D \sim O(1)$ in AA collisions at RHIC energy. We calculate the heavy baryon/meson ratio and the $p_T$ spectra of charmed hadrons with and without strangeness content: $D^{0}$, $D_{s}$, $\Lambda_{c}^{+}$, $\Xi_c$ and $\Omega_c$ in $pp$ collisions at top LHC energies, finding a satisfactory prediction of the measured $\Lambda_{c}^{+}/D^0$ and the $\Xi_c/D^0$ without any specific tuning of parameters to $pp$ collisions. At variance with other approaches a coalescence approach predicts also a significant production of $\Omega_c$ such that $\Omega_c/D^0 \sim O(10^{-1})$ .
high energy physics phenomenology
The control of traffic signals is fundamental and critical to alleviate traffic congestion in urban areas. However, it is challenging since traffic dynamics are complicated in real-world scenarios. Because of the high complexity of the optimisation problem for modelling the traffic, experimental settings of existing works are often inconsistent. Moreover, it is not trivial to control multiple intersections properly in real complex traffic scenarios due to its vast state and action space. Failing to take intersection topology relations into account also results in inferior solutions. To address these issues, in this work we carefully design our settings and propose a new dataset including both synthetic and real traffic data in more complex scenarios. Additionally, we propose a novel baseline model with strong performance. It is based on deep reinforcement learning with an encoder-decoder structure: an edge-weighted graph convolutional encoder to excavate multi-intersection relations; and an unified structure decoder to jointly model multiple junctions in a comprehensive manner, which significantly reduces the number of the model parameters. By doing so, the proposed model is able to effectively deal with the multi-intersection traffic optimisation problem. Models are trained/tested on both synthetic and real maps and traffic data with the Simulation of Urban Mobility (SUMO) simulator. Experimental results show that the proposed model surpasses multiple competitive methods.
electrical engineering and systems science
The estimation of causal treatment effects from observational data is a fundamental problem in causal inference. To avoid bias, the effect estimator must control for all confounders. Hence practitioners often collect data for as many covariates as possible to raise the chances of including the relevant confounders. While this addresses the bias, this has the side effect of significantly increasing the number of data samples required to accurately estimate the effect due to the increased dimensionality. In this work, we consider the setting where out of a large number of covariates $X$ that satisfy strong ignorability, an unknown sparse subset $S$ is sufficient to include to achieve zero bias, i.e. $c$-equivalent to $X$. We propose a common objective function involving outcomes across treatment cohorts with nonconvex joint sparsity regularization that is guaranteed to recover $S$ with high probability under a linear outcome model for $Y$ and subgaussian covariates for each of the treatment cohort. This improves the effect estimation sample complexity so that it scales with the cardinality of the sparse subset $S$ and $\log |X|$, as opposed to the cardinality of the full set $X$. We validate our approach with experiments on treatment effect estimation.
statistics
In CRESST-III, 10 cryogenic detector modules optimized for low energy thresholds were operated for almost two years (May 2016 - February 2018). Together with this document we are publishing data from the best performing detector module which has a nuclear recoil threshold of 30.1eV. With this data-set we were able to set limits on the cross-section for spin-dependent and spin-independent elastic scattering of dark matter particles off nuclei at dark matter masses down to 160MeV/c$^2$. We publish the energies of all events after data selection as well as of all events within the acceptance region for dark-matter searches. In this document we describe how to use these data sets.
astrophysics
We present for the first time an experimental demonstration on the energy allocation of commodity Wi-Fi signals in a programmable and inexpensive way. To that end, we design an electronically-programmable phase-binary coding metasurface, working at the 2.4GHz Wi-Fi frequency band, to manipulate dynamically and arbitrarily the spatial distribution of commodity Wi-Fi signals. Meanwhile, an efficient algorithm is developed to find the optimal coding sequence of the programmable metasurface such that the spatial energy of commodity Wi-Fi signals can be instantly controlled in a desirable way. Selected experimental results based on an IEEE 802.11n commercial Wi-Fi protocol have been provided to demonstrate the performance of the developed proof-of-concept system in enhancing the commodity Wi-Fi signals dynamically and arbitrarily. It could be expected that the proposed strategy will pave a promising way for wireless communications, future smart home, and so on.
physics
Neuromorphic computing takes inspiration from the brain to create energy efficient hardware for information processing, capable of highly sophisticated tasks. In this article, we make the case that building this new hardware necessitates reinventing electronics. We show that research in physics and material science will be key to create artificial nano-neurons and synapses, to connect them together in huge numbers, to organize them in complex systems, and to compute with them efficiently. We describe how some researchers choose to take inspiration from artificial intelligence to move forward in this direction, whereas others prefer taking inspiration from neuroscience, and we highlight recent striking results obtained with these two approaches. Finally, we discuss the challenges and perspectives in neuromorphic physics, which include developing the algorithms and the hardware hand in hand, making significant advances with small toy systems, as well as building large scale networks.
computer science
Bitcoin introduced delegation of control over a monetary system from a select few to all who participate in that system. This delegation is known as the decentralization of controlling power and is a powerful security mechanism for the ecosystem. After the introduction of Bitcoin, the field of cryptocurrency has seen widespread attention from industry and academia, so much so that the original novel contribution of Bitcoin i.e. decentralization, may be overlooked, due to decentralizations assumed fundamental existence for the functioning of such cryptoassets. However recent studies have observed a trend of increased centralization in cryptocurrencies such as Bitcoin and Ethereum. As this increased centralization has an impact the security of the blockchain, it is crucial that it is measured, towards adequate control. This research derives an initial taxonomy of centralization present in decentralized blockchains through rigorous synthesis using a systematic literature review. This is followed by iterative refinement through expert interviews. We systematically analyzed 89 research papers published between 2009 and 2019. Our study contributes to the existing body of knowledge by highlighting the multiple definitions and measurements of centralization in the literature. We identify different aspects of centralization and propose an encompassing taxonomy of centralization concerns. This taxonomy is based on empirically observable and measurable characteristics. It consists of 13 aspects of centralization classified over six architectural layers Governance Network Consensus Incentive Operational and Application. We also discuss how the implications of centralization can vary depending on the aspects studied. We believe that this review and taxonomy provides a comprehensive overview of centralization in decentralized blockchains involving various conceptualizations and measures.
computer science
In this paper, we report the results of our participation in the TREC-COVID challenge. To meet the challenge of building a search engine for rapidly evolving biomedical collection, we propose a simple yet effective weighted hierarchical rank fusion approach, that ensembles together 102 runs from (a) lexical and semantic retrieval systems, (b) pre-trained and fine-tuned BERT rankers, and (c) relevance feedback runs. Our ablation studies demonstrate the contributions of each of these systems to the overall ensemble. The submitted ensemble runs achieved state-of-the-art performance in rounds 4 and 5 of the TREC-COVID challenge.
computer science
Creation of electrons and positrons from light alone is a basic prediction of quantum electrodynamics, but yet to be observed. Here we show that it is possible to create ${>}10^8$ positrons by dual laser irradiation of a structured plasma target, at intensities of $2 \times 10^{22} \mathrm{W}\mathrm{cm}^{-2}$. In contrast to previous work, the pair creation is primarily driven by the linear Breit-Wheeler process ($\gamma\gamma \to e^+ e^-$), not the nonlinear process assumed to be dominant at high intensity, because of the high density of $\gamma$ rays emitted inside the target. The favorable scaling with laser intensity of the linear process prompts reconsideration of its neglect in simulation studies, but also permits positron jet formation at intensities that are already experimentally feasible. Simulations show that the positrons, confined by a quasistatic plasma magnetic field, may be accelerated by the lasers to energies $> 200$ MeV.
physics
I propose a discrete synchronization model of finite d-level systems and discuss what happens once superposition of states is allowed. The model exhibits various asymptotic behaviors that depend on the initial state. In particular, two antagonistic phenomena can occur: a quantum-to-classical transition and entanglement generation. Next, I generalize this model and show that it is possible to phase-lock a periodic dynamics of a single qubit to a periodic dynamics of a qudit stimulus.
quantum physics
Many video enhancement algorithms rely on optical flow to register frames in a video sequence. Precise flow estimation is however intractable; and optical flow itself is often a sub-optimal representation for particular video processing tasks. In this paper, we propose task-oriented flow (TOFlow), a motion representation learned in a self-supervised, task-specific manner. We design a neural network with a trainable motion estimation component and a video processing component, and train them jointly to learn the task-oriented flow. For evaluation, we build Vimeo-90K, a large-scale, high-quality video dataset for low-level video processing. TOFlow outperforms traditional optical flow on standard benchmarks as well as our Vimeo-90K dataset in three video processing tasks: frame interpolation, video denoising/deblocking, and video super-resolution.
computer science
First- and second-harmonic dielectric susceptibilities are maidenly studied on Samarium Orthoferrite of mesoscopic/500 nm and nanoscopic/55 nm grainsizes. Magneto-electrically coupled to the antiferromagnetic and spin-reorientation transitions, fundamental and harmonic dielectricity consistently reflect the global/local polarization effects of crystallite-size dependent electrical orderings. Bulk and incipient ferroelectricity respectively in nanoscopic and mesoscopic crystallites concur the higher-temperature antiferromagnetic ordering (T_N ~670 K). Upon the spin-reorientation transition at lower-temperature (T_SR ~470 K), re-entrant relaxor state in the nano-crystallites and bulk-like/temperature-windowed ferroelectricity in the meso-crystallites emerge. In the nano-crystallites, magneto-electric signature of interfacial spins' de-pinning (T_SP ~540 K) is exclusively revealed by the scaled-harmonics.
condensed matter
We study the luminescence dynamics of telecom wavelength InAs quantum dots grown on InP(111)A by droplet epitaxy. The use of the ternary alloy InAlGaAs as a barrier material leads to photon emission in the 1.55 $\mu$m telecom C-band. The luminescence decay is well described in terms of the theoretical interband transition strength without the impact of nonradiative recombination. The intensity autocorrelation function shows clear anti-bunching photon statistics. The results suggest that our quantum dots are useful for constructing a practical source of single photons and quantum entangled photon pairs.
condensed matter
Recently, topologically engineered photonic structures have garnered significant attention as their eigenstates may offer a new insight on photon manipulation and an unconventional route for nanophotonic devices with unprecedented functionalities and robustness. Herein, we present lasing actions at all hierarchical eigenstates that can exist in a topologically designed single two-dimensional (2D) photonic crystal (PhC) platform: 2D bulk, one-dimensional edge, and zero-dimensional corner states. In particular, multiple topological eigenstates are generated in a hierarchical manner with no bulk multipole moment. The unit cell of the topological PhC structure is a tetramer composed of four identical air holes perforated into an InGaAsP multiple-quantum-well epilayer slab. A square area of a topologically nontrivial PhC structure is surrounded by a topologically trivial counterpart, resulting in multidimensional eigenstates of one bulk, four side edges, and four corners within and at the boundaries. Spatially resolved optical excitation spontaneously results in lasing actions at all nine hierarchical topological states. Our experimental findings may provide insight into the development of sophisticated next-generation nanophotonic devices and robust integration platforms.
physics
We investigate analyticity properties of correlation functions in conformal field theories (CFT) in the Wightman formulation. The goal is to determine domain of holomorphy of permuted Wightman functions. We focus on crossing property of three-point functions. The domain of holomorphy of a pair of three-point functions is determined by appealing to Jost's theorem and by adopting the technique of analytic completion. This program paves the way to address the issue of crossing for the four-point functions on a rigorous footing.
high energy physics theory
We study conformal blocks for thermal one-point-functions on the sphere in conformal field theories of general dimension. These thermal conformal blocks satisfy second order Casimir differential equations and have integral representations related to AdS Witten diagrams. We give an analytic formula for the scalar conformal block in terms of generalized hypergeometric functions. As an application, we deduce an asymptotic formula for the three-point coeffcients of primary operators in the limit where two of the operators are heavy.
high energy physics theory
Edge-preserving filters play an essential role in some of the most basic tasks of computational photography, such as abstraction, tonemapping, detail enhancement and texture removal, to name a few. The abundance and diversity of smoothing operators, accompanied by a lack of methodology to evaluate output quality and/or perform an unbiased comparison between them, could lead to misunderstanding and potential misuse of such methods. This paper introduces a systematic methodology for evaluating and comparing such operators and demonstrates it on a diverse set of published edge-preserving filters. Additionally, we present a common baseline along which a comparison of different operators can be achieved and use it to determine equivalent parameter mappings between methods. Finally, we suggest some guidelines for objective comparison and evaluation of edge-preserving filters.
electrical engineering and systems science
We explore the use of Array-RQMC, a randomized quasi-Monte Carlo method designed for the simulation of Markov chains, to reduce the variance when simulating stochastic biological or chemical reaction networks with $\tau$-leaping. The task is to estimate the expectation of a function of molecule copy numbers at a given future time $T$ by the sample average over $n$ sample paths, and the goal is to reduce the variance of this sample-average estimator. We find that when the method is properly applied, variance reductions by factors in the thousands can be obtained. These factors are much larger than those observed previously by other authors who tried RQMC methods for the same examples. Array-RQMC simulates an array of realizations of the Markov chain and requires a sorting function to reorder these chains according to their states, after each step. The choice of sorting function is a key ingredient for the efficiency of the method. We illustrate this by comparing various choices. The expected number of reactions of each type per step also has an impact on the efficiency gain.
statistics
We study a generalized clock model on the simple cubic lattice. The parameter of the model can be tuned such that the amplitude of the leading correction to scaling vanishes. In the main part of the study we simulate the model with $Z_8$ symmetry. At the transition, with increasing length scale, $O(2)$ symmetry emerges. We perform Monte Carlo simulations using a hybrid of local Metropolis and cluster algorithms of lattices with a linear size up to $L=512$. The field variable requires less memory and the updates are faster than for a model with $O(2)$ symmetry at the microscopic level. Our finite size scaling analysis yields accurate estimates for the critical exponents of the three-dimensional XY-universality class. In particular we get $\eta=0.03810(8)$, $\nu=0.67169(7)$, and $\omega=0.789(4)$. Furthermore we obtain estimates for fixed point values of phenomenological couplings and critical temperatures.
condensed matter
Coherent steering of a quantum state, induced by a sequence of weak measurements, has become an active area of theoretical and experimental study. For a closed steered trajectory, the underlying phase factors involve both geometrical and dynamical terms. Furthermore, considering the reversal of the order of the measurement sequence, such a phase comprises a symmetric and an antisymmetric term. Superseding common wisdom, we show that the symmetric and the antisymmetric components do not correspond to the dynamical and geometrical parts respectively. Addressing a broad class of measurement protocols, we further investigate the dependence of the induced phases on the measurement parameters (e.g., the measurement strength). We find transitions between different topologically distinct sectors, defined by integer-valued winding numbers, and show that the transitions are accompanied by diverging dephasing. We propose experimental protocols to observe these effects.
quantum physics
The insulator-to-metal transition (IMT) in vanadium dioxide (VO2) can enable a variety of optics applications, including switching and modulation, optical limiting, and tuning of optical resonators. Despite the widespread interest in optics, the optical properties of VO2 across its IMT are scattered throughout the literature, and are not available in some wavelength regions. We characterized the complex refractive index of VO2 thin films across the IMT for free-space wavelengths from 300 nm to 30 {\mu}m, using broadband spectroscopic ellipsometry, reflection spectroscopy, and the application of effective-medium theory. We studied VO2 thin films of different thickness, on two different substrates (silicon and sapphire), and grown using different synthesis methods (sputtering and sol gel). While there are differences in the optical properties of VO2 synthesized under different conditions, they are relatively minor compared to the change resulting from the IMT, most notably in the ~2 - 11 {\mu}m range where the insulating phase of VO2 has relatively low optical loss. We found that the macroscopic optical properties of VO2 are much more robust to sample-to-sample variation compared to the electrical properties, making the refractive-index datasets from this article broadly useful for modeling and design of VO2-based optical and optoelectronic components.
physics
We theoretically investigate the radial-spin-wave induced magnetic vortex switching in the presence of Dzyaloshinskii-Moriya interaction (DMI). From micromagnetic simulations, we observe a circular-to-radial vortex phase transition by increasing the DMI strength. The radial spin-wave excitation spectrum for each magnetization configuration is analyzed, showing that the frequency of spin-wave mode with a given radial node number monotonically increases (decreases) with the DMI parameter of the radial (circular) vortex. Interestingly, we find that the DMI can significantly facilitate the polarity switching of the circular vortex driven by radial spin waves. Our work provides a new insight into the DMI effect on the vortex dynamics and is helpful for designing fast all-magnonic memory devices.
condensed matter
Object encoding and identification is crucial for many robotic tasks such as autonomous exploration and semantic relocalization. Existing works heavily rely on the tracking of detected objects but difficult to recall revisited objects precisely. In this paper, we propose a novel object encoding method based on a graph of key-points. To be robust to the number of key-points detected, we propose a feature sparse encoding and object dense encoding method to ensure that each key-point can only affect a small part of the object descriptors, leading it robust to viewpoint changes, scaling, occlusion, and even object deformation. In the experiments, we show that it achieves superior performance for object identification than the state-of-the art algorithm and is able to provide reliable semantic relocalization. It is a plug-and-play module and we expect that it will play an important role in the robotic applications.
computer science
We present analysis of the rate of giant radio pulses (GPs) emission from the Crab pulsar (B0531+21). Results of our 9 years daily observations with the Large Phased Array radio telescope of Pushchino Radio Astronomy Observatory at 111 MHz were used. Limited sample of 8753 strong individual pulses in 2004 observational sessions was further analysed. It was shown that the observed monthly averaged rate of GPs emission was highly unstable during the entire span of observations and changes by about two orders of magnitude for high-energy pulses. Data were further analysed to search for the possible connection between pulsar glitches and the process of GP emission. We have found a significant increase in the rate of emission of high-energy GPs after MJD 58064, when the largest glitch ever observed in the Crab pulsar was happened. Although considerable changes in GPs emission rate could have been caused by the propagation effects in the nebula itself, we have found that the pulsar had demonstrated high degree of intrinsic irregularity of high-energy pulses emission over long time intervals.
astrophysics
We study the effect of periodic boundary conditions on chiral symmetry breaking and its restoration in Quantum Chromodynamics. As an effective model of the effective potential for the quark condensate, we use the quark-meson model, while the theory is quantized in a cubic box of size $L$. After specifying a renormalization prescription for the vacuum quark loop, we study the condensate at finite temperature, $T$, and quark chemical potential, $\mu$. We find that lowering $L$ leads to a catalysis of chiral symmetry breaking. The excitation of the zero mode leads to a jump in the condensate at low temperature and high density, that we suggest to interpret as a gas-liquid phase transition that takes place between the chiral symmetry broken phase (hadron gas) and chiral symmetry restored phase (quark matter). We characterize this intermediate phase in terms of the increase of the baryon density, and of the correlation length of the fluctuations of the order parameter: for small enough $L$ the correlation domains occupy a substantial portion of the volume of the system, and the fluctuations are comparable to those in the critical region. For these reasons, we dub this phase as the {\it subcritical liquid}. The qualitative picture that we draw is in agreement with previous studies based on similar effective models. We also clarify the discrepancy on the behavior of the critical temperature versus $L$ found in different models.
high energy physics phenomenology
Simulating a fermionic system on a quantum computer requires encoding the anti-commuting fermionic variables into the operators acting on the qubit Hilbert space. The most familiar of which, the Jordan-Wigner transformation, encodes fermionic operators into non-local qubit operators. As non-local operators lead to a slower quantum simulation, recent works have proposed ways of encoding fermionic systems locally. In this work, we show that locality may in fact be too strict of a condition and the size of operators can be reduced by encoding the system quasi-locally. We give examples relevant to lattice models of condensed matter and systems relevant to quantum gravity such as SYK models. Further, we provide a general construction for designing codes to suit the problem and resources at hand and show how one particular class of quasi-local encodings can be thought of as arising from truncating the state preparation circuit of a local encoding. We end with a discussion of designing codes in the presence of device connectivity constraints.
quantum physics
In this work, we propose and study annotated code search: the retrieval of code snippets paired with brief descriptions of their intent using natural language queries. On three benchmark datasets, we investigate how code retrieval systems can be improved by leveraging descriptions to better capture the intents of code snippets. Building on recent progress in transfer learning and natural language processing, we create a domain-specific retrieval model for code annotated with a natural language description. We find that our model yields significantly more relevant search results (with absolute gains up to 20.6% in mean reciprocal rank) compared to state-of-the-art code retrieval methods that do not use descriptions but attempt to compute the intent of snippets solely from unannotated code.
computer science
The Price model, the directed version of the Barab\'{a}si-Albert model, produces a growing directed acyclic graph. We look at variants of the model in which directed edges are added to the new vertex in one of two ways: using cumulative advantage (preferential attachment) choosing vertices in proportion to their degree, or with random attachment in which vertices are chosen uniformly at random. In such networks, the longest path is well defined and in some cases is known to be a better approximation to geodesics than the shortest path. We define a reverse greedy path and show both analytically and numerically that this scales with the logarithm of the size of the network with a coefficient given by the number of edges added using random attachment. This is a lower bound on the length of the longest path to any given vertex and we show numerically that the longest path also scales with the logarithm of the size of the network but with a larger coefficient that has some weak dependence on the parameters of the model.
physics
We characterize a semiconductor external cavity diode laser whose optical feedback is provided by a guided mode resonance filter (GMRF). We focus on the spectral properties. The wavelength of operation falls in the telecom range (1506 nm). The GMRF acting both as a wavelength intracavity filter and feedback mirror allows for a compact laser design. The single-mode operation is verified in a wide range of driving currents. We finely tune the cavity length to adjust the frequency by 14 GHz without mode-hops in agreement with the expected free-spectral range of the resonator $\sim$20 GHz. The compactness of the cavity allows fast frequency sweeps when modulating the current (90 MHz/mA at 100 kHz, the modulation bandwidth). The frequency noise (366 kHz white-noise contribution) is also analysed to evaluate the potential of our design for high-resolution applications.
physics
Within the chiral unitary approach and with the constraints of heavy quark spin symmetry, we study the coupled channel interactions of ${\bar D}^{(*)}\Sigma_c^{(*)}$ channels, close to whose thresholds three pentaquark-like $P_c$ states have been reported by the LHCb Collaboration. In the present work, we take into account the contributions of pion exchanges via box diagrams to the interaction potentials, and therefore lift the degeneracy in the masses of ${\bar D}^*\Sigma_c^{(*)}$ spin multiplets. Fitting the $J/\psi p$ invariant mass distributions in the $\Lambda_b^0 \to J/\psi K^- p$ decay, we find that the LHCb pentaquark states can not be reproduced in the direct $J/\psi p$ production in the $\Lambda_b^0$ decay, and can only be indirectly produced in the final state interactions of the $\Lambda_b^0$ decay products, ${\bar D}^*\Sigma_c^{(*)}$, which further supports the nature of these states as $\bar{D}\Sigma_c$ molecules. Based on the fit results obtained, we study the partial decay widths/branching ratios to other decay channels, $\bar{D}^* \Lambda_c$, $\bar{D} \Lambda_c$, and $\eta_c N$, and the corresponding invariant mass distributions. The resonances with $J^P=\frac{1}{2}^-$, $P_c(4312)$, $P_c(4440)$ and the one of $\bar{D}^* \Sigma_c^*$ around 4500 MeV, have large partial decay width into $\eta_c N$, and thus, can be easily seen in the $\eta_c N$ invariant mass distributions. By contrast, the states with $J^P=\frac{3}{2}^-$, $P_c(4457)$, the (predicted) narrow $P_c(4380)$ and the bound state of $\bar{D}^* \Sigma_c^*$ with a mass of about 4520 MeV, do not decay into $\eta_c N$. Therefore, the $\eta_c N$ channel should be studied in future to provide further insights into the nature of these states, especially that of the $P_c(4440)$ and $P_c(4457)$.
high energy physics phenomenology
We introduce a new family of rank metric codes: Low Rank Parity Check codes (LRPC), for which we propose an efficient probabilistic decoding algorithm. This family of codes can be seen as the equivalent of classical LDPC codes for the rank metric. We then use these codes to design cryptosystems \`a la McEliece: more precisely we propose two schemes for key encapsulation mechanism (KEM) and public key encryption (PKE). Unlike rank metric codes used in previous encryption algorithms -notably Gabidulin codes - LRPC codes have a very weak algebraic structure. Our cryptosystems can be seen as an equivalent of the NTRU cryptosystem (and also to the more recent MDPC \cite{MTSB12} cryptosystem) in a rank metric context. The present paper is an extended version of the article introducing LRPC codes, with important new contributions. We have improved the decoder thanks to a new approach which allows for decoding of errors of higher rank weight, namely up to $\frac{2}{3}(n-k)$ when the previous decoding algorithm only decodes up to $\frac{n-k}{2}$ errors. Our codes therefore outperform the classical Gabidulin code decoder which deals with weights up to $\frac{n-k}{2}$. This comes at the expense of probabilistic decoding, but the decoding error probability can be made arbitrarily small. The new approach can also be used to decrease the decoding error probability of previous schemes, which is especially useful for cryptography. Finally, we introduce ideal rank codes, which generalize double-circulant rank codes and allow us to avoid known structural attacks based on folding. To conclude, we propose different parameter sizes for our schemes and we obtain a public key of 3337 bits for key exchange and 5893 bits for public key encryption, both for 128 bits of security.
computer science
We show that the various higher Segal conditions of Dyckerhoff and Kapranov can all be characterized in purely categorical terms by higher excision conditions (in the spirit of Goodwillie-Weiss manifold calculus) on the simplex category $\Delta$ and the cyclic category $\Lambda$.
mathematics
We study the spin-polarized spectral properties of Yu-Shiba-Rusinov resonance states induced by magnetic impurities in 2- and 3-dimensional nematic superconductors: few layer Bi$_2$Te$_3$ grown on FeTe$_{0.55}$Se$_{0.45}$ (2-dimensional) and Cu$_x$Bi$_2$Se$_3$ (3-dimensional). We focus on the relationship between pairing symmetry and the topograph of spin-polarized spectroscopy. We calculate the spin-polarized local density of states (SP LDOS) and the corresponding Fourier transformation using the $T$-matrix method for both the 2- and 3-dimensional materials. Various situations with different impurity orientations and different SP LDOSs have been investigated. We find that, like the quasiparticle interference spectrum, the SP LDOS can be applied to distinguish other pairings which preserve the threefold rotation symmetry of $D_3$ point group and nematic pairings in these materials.
condensed matter
The IoT consists of a lot of devices such as embedded systems, wireless sensor nodes (WSNs), control systems, etc. It is essential for some of these devices to protect information that they process and transmit. The issue is that an adversary may steal these devices to gain a physical access to the device. There is a variety of ways that allows to reveal cryptographic keys. One of them are optical Fault Injection attacks. We performed successful optical Fault Injections into different type of gates, in particular INV, NAND, NOR, FF. In our work we concentrate on the selection of the parameters configured by an attacker and their influence on the success of the Fault Injections.
computer science
We consider a linear power flow model with interval-bounded nodal power injections and limited line power flows. We determine the minimal number of power injections to control based on a minimal set of measurements, such that the overall system is feasible for all assignments of the non-controlled power injections. For the important case where the possible measurements are the nodal power injections, we show that the problem can be solved efficiently as a mixed-integer linear program (MILP). When also line power flows are considered as potential measurements, we derive an iterative, greedy algorithm that provides a feasible, but potentially conservative solution. We apply the developed algorithms to both a small microgrid and a modified version of the IEEE 118 bus test power system. We show that in both cases a sparse solution in terms of the number of required controllers and measurements can be obtained. Moreover, the number of required measurements can be reduced significantly if line flow measurements are considered additionally to nodal power injections.
electrical engineering and systems science
We define a theta operator on p-adic vector-valued modular forms on unitary groups of arbitrary signature, over a quadratic imaginary field in which p is inert. We study its effect on Fourier-Jacobi expansions and prove that it extends holomorphically beyond the {\mu}-ordinary locus, when applied to scalar-valued forms.
mathematics
Motivated by weak ferromagnetism (FM) in a $\tau$-type molecular conductor ($\tau$-MC), we examine its mechanism using a two-band extended Hubbard model. Applying the random phase approximation, we elucidate the uniform spin and charge fluctuations between unit cells in the presence of on-site and off-site interactions. Applying the mean-field approximation, we find the ordered state mixing with antiferromagnetism (AFM), weak FM, and charge ordering (CO) components in each unit cell: we classify this state as ferrimagnetism (FIM). We reveal the phase diagrams in the interaction and interaction-temperature spaces. The former shows that the off-site interaction induces FIM from pure AFM and the latter shows that lowering the temperature stabilizes FIM. To clarify the stabilization mechanism of the phases, we focus on the microscopic nature of the ordered states, including the band structure, Fermi surface, and density of states. We find that the FIM state is obtained from mixing features of AFM, CO, and FM; therefore, the emergence of FIM requires both the on-site and off-site interactions. Then, we discuss the effect of lowering the temperature and predict that the AFM gap assists the emergence of FIM based on AFM. This FIM state is possibly related to the observation of the weak FM in $\tau$-MC.
condensed matter
Superconducting Nb3Sn films can be synthesized by controlling the atomic concentration of Sn. Multilayer sequential sputtering of Nb and Sn thin films followed by high temperature annealing is considered as a method to fabricate Nb3Sn films, where the Sn composition of the deposited films can be controlled by the thickness of alternating Nb and Sn layers. We report on the structural, morphological and superconducting properties of Nb3Sn films fabricated by multilayer sequential sputtering of Nb and Sn films on sapphire substrates followed by annealing at 950 {\deg}C for 3 h. We have investigated the effect of Nb and Sn layer thickness and Nb:Sn ratio on the properties of the Nb3Sn films. The crystal structure, surface morphology, surface topography, and film composition were characterized by X-ray diffraction (XRD), scanning electron microscopy (SEM), atomic force microscopy (AFM), and energy dispersive X-ray spectroscopy (EDS). The results showed Sn loss from the surface due to evaporation during annealing. Superconducting Nb3Sn films of critical temperature up to 17.93 K were fabricated.
physics
Starting from a continuum-model description, we develop a microscopic weak-coupling theory for superconductivity in twisted double-bilayer graphene. We study both electron-phonon and entirely electronic pairing mechanisms. In each case, the leading superconducting instability transforms under the trivial representation, $A$, of the point group $C_3$ of the system, while the subleading pairing phases belong to the $E$ channel. We explicitly compute the momentum dependence of the associated order parameters and find that the leading state has no nodal points for electron-phonon pairing but exhibits six sign changes on the Fermi surface if the Coulomb interaction dominates. On top of these system-specific considerations, we also present general results relevant to other correlated graphene-based moir\'e superlattice systems. We show that, irrespective of microscopic details, triplet pairing will be stabilized if the collective electronic fluctuations breaking the enhanced $\text{SU}(2)_+ \times \text{SU}(2)_-$ spin symmetry of these systems are odd under time reversal, even when the main $\text{SU}(2)_+ \times \text{SU}(2)_-$-symmetric part of the pairing glue is provided by phonons. Furthermore, we discuss the disorder sensitivity of the candidate pairing states and demonstrate that the triplet phase is protected against disorder on the moir\'e scale.
condensed matter
Using new VLT/XShooter spectral observations we analyse the physical properties of five z~0.3-0.4 confirmed LyC leakers. Strong resonant MgII 2796,2803 emission lines (I(2796,2803)/I(Hbeta)=10-38 per cent) and non-resonant FeII* 2612,2626 emission lines are observed in spectra of five and three galaxies, respectively. We find high electron densities Ne~400cm-3, significantly higher than in typical low-z, but comparable to those measured in z~2-3 star-forming galaxies. The galaxies have a mean value of logN/O=-1.16, close to the maximum values found for star-forming (SF) galaxies in the metallicity range of 12+logO/H=7.7-8.1. All 11 low-z LyC emitting galaxies found by Izotov et al. (2016, 2018), including the ones considered in the present study, are characterised by high EW(Hbeta)~200-400A, high ionisation parameter (log(U)=-2.5 to -1.7), high average ionising photon production efficiency \xi= 10^{25.54} Hz erg-1 and hard ionising radiation. On the BPT diagram we find the same offset of our leakers from low-$z$ main-sequence SFGs as that for local analogues of LBGs and extreme SF galaxies at z~2-3. We confirm the effectiveness of the HeI emission lines diagnostics proposed by Izotov et al. (2017) in searching for LyC leaker candidates and find that their intensity ratios correspond to those in a median with low neutral hydrogen column density N(HI)=10^{17}-5x10^{17} cm-2 that permit leakage of LyC radiation, likely due to their density-bounded HII regions.
astrophysics
This research article suggests that there are significant benefits in exposing demand planners to forecasting methods using matrix completion techniques. This study aims to contribute to a better understanding of the field of forecasting with multivariate time series prediction by focusing on the dimension of large commercial data sets with hierarchies. This research highlights that there has neither been sufficient academic research in this sub-field nor dissemination among practitioners in the business sector. This study seeks to innovate by presenting a matrix completion method for short-term demand forecast of time series data on relevant commercial problems. Albeit computing intensive, this method outperforms the state of the art while remaining accessible to business users. The object of research is matrix completion for time series in a big data context within the industry. The subject of the research is forecasting product demand using techniques for multivariate hierarchical time series prediction that are both precise and accessible to non-technical business experts. Apart from a methodological innovation, this research seeks to introduce practitioners to novel methods for hierarchical multivariate time series prediction. The research outcome is of interest for organizations requiring precise forecasts yet lacking the appropriate human capital to develop them.
statistics
The goal of this note is to explore the behavior of effective action in the SYK model with general continuous global symmetries. A global symmetry will decompose the whole Hamiltonian of a many-body system to several single charge sectors. For the SYK model, the effective action near the saddle point is given as the free product of the Schwarzian action part and the free action of the group element moving in the group manifold. With a detailed analysis in the free sigma model, we prove a modified version of Peter-Weyl theorem that works for generic spin structure. As a conclusion, we could make a comparison between the thermodynamics and the spectral form factors between the whole theory and the single charge sector, to make predictions on the SYK model and see how symmetry affects the chaotic behavior in certain timescales.
high energy physics theory
In a recent note I argued that the holographic origin of gravitational attraction is the quantum mechanical tendency for operators to grow under time evolution. In a followup the claim was tested in the context of the SYK theory and its bulk dual---the theory of near-extremal black holes. In this paper I give an improved version of the size-momentum correspondence and show that Newton's laws of motion are a consequence. Operator size is closely related to complexity. Therefore one may say that gravitational attraction is a manifestation of the tendency for complexity to increase. The improved version of the size-momentum correspondence can be justified by the arguments of Lin, Maldacena, and Zhao constructing symmetry generators for the approximate symmetries of the SYK model.
high energy physics theory
In this paper, we propose to address the issue of the lack of strongly labeled data by using pseudo strongly labeled data that is approximated using Convolutive Nonnegative Matrix Factorization (CNMF). Using this pseudo strongly labeled data, we then train a new architecture combining Convolutional Neural Network (CNN) with Macaron Net (MN), which we term it as Convolutional Macaron Net (CMN). As opposed to the Mean-Teacher approach which trains two similar models synchronously, we propose to train two different CMNs synchronously where one of the models will provide the frame-level prediction while the other will provide the clip level prediction. Based on our proposed framework, our system outperforms the baseline system of the Detection and Classification of Acoustic Scenes and Events (DCASE) 2020 Challenge Task 4 by a margin of over 10%. By comparing with the first place of the challenge which utilize a combination of CNN and Conformer, our system also marginally wins it by 0.3%.
electrical engineering and systems science
Possibility to communicate between spatially separated regions, without even a single photon passing between the two parties, is an amazing quantum phenomenon. The possibility of transmitting one value of a bit in such a way, the interaction-free measurement, was known for quarter of a century. The protocols of full communication, including transmitting unknown quantum states were proposed only few years ago, but it was shown that in all these protocols the particle was leaving a weak trace in the transmission channel, the trace larger than the trace left by a single particle passing through the channel. This made the claim of counterfactuality of these protocols at best controversial. However, a simple modification of these recent protocols eliminates the trace in the transmission channel making all these protocols counterfactual.
quantum physics
Here, we provide a reappraisal of potential LLSVPs compositions based on an improved mineralogical model including, for instance, the effects of alumina. We also systematically investigate the effects of six parameters: FeO and Al$_{2}$O$_{3}$ content, proportion of CaSiO$_{3}$ and bridgmanite (so that the proportion of ferropericlase is implicitly investigated), Fe$^{3+}$/$\sum$Fe and temperature contrast between far-field mantle and LLSVPs. From the 81 millions cases studied, only 79000 cases explain the seismic observations. Nevertheless, these successful cases involve a large range of parameters with, for instance, FeO content between 12--25~wt\% and Al$_{2}$O$_{3}$ content between 3--17~wt\%. We then apply a principal component analysis (PCA) to these cases and find two robust results: (i) the proportion of ferropericlase should be low ($<$6vol\%); (ii) the formation of Fe$^{3+}$-bearing bridgmanite is much more favored than other iron-bearing phases. Following these results, we identify two end-member compositions, Bm-rich and CaPv-rich, and discuss their characteristics. Finally, we discuss different scenarios for the formation of LLSVPs and propose that investigating the mineral proportion produced by each scenario is the best way to evaluate their relevance. For instance, the solidification of a primitive magma ocean may produce FeO and Al$_{2}$O$_{3}$ content similar to those suggested by our analysis. However, the mineral proportion of such reservoirs is not well-constrained and may contain a larger proportion of ferropericlase than what is allowed by our results.
physics
Squeezed light is a quantum resource that can improve the sensitivity of optical measurements. However, existing sources of squeezed light generally require high powers and are not amenable to portability. Here we theoretically investigate an alternative technique for generating squeezing using degenerate four-wave-mixing in atomic vapors. We show that by minimizing excess noise, this technique has the potential to generate measurable squeezing with low powers attainable by a small diode laser. We suggest experimental techniques to reduce excess noise and employ this alternative nonlinear optical process to build a compact, low-power source of squeezed light.
quantum physics
A method to establish a qubit decomposition of a general qudit state is presented. This new representation allows a geometrical depiction of any qudit state in the Bloch sphere. Additionally, we show that the nonnegativity conditions of the qudit state imply the existence of quantum correlations between the qubits which compose it. These correlations are used to define new inequalities which the density matrices components must satisfy. The importance of such inequalities in the reconstruction of a qudit state is addressed. As an example of the general procedure the qubit decomposition of a qutrit system is shown, which allows a classification of the qutrit states by fixing their invariants ${\rm Tr}(\hat{\rho}^2)$, ${\rm Tr}(\hat{\rho}^3)$.
quantum physics
We investigate analytically and numerically the steady-state entanglement and coherence of two coupled qubits each interacting with a local boson or fermion reservoir, based on the Bloch-Redfield master equation beyond the secular approximation. We find that there is non-vanishing steady-state coherence in the nonequilibrium scenario, which grows monotonically with the nonequilibrium condition quantified by the temperature difference or chemical potential difference of the two baths. The steady-state entanglement in general is a non-monotonic function of the nonequilibrium condition as well as the bath parameters in the equilibrium setting. We also find that weak inter-qubit coupling and high base temperature or chemical potential of the baths can strongly suppress the steady-state entanglement and coherence, regardless of the strength of the nonequilibrium condition. On the other hand, the energy detuning of the two qubits, when used in a compensatory way with the nonequilibrium condition, can lead to significant enhancement of the steady-state entanglement in some parameter regimes. In addition, the qubits typically have a stronger steady-state entanglement when coupled to fermion baths exchanging particle with the system than boson baths exchanging energy with the system under similar conditions. We also discussed the possible experimental realization of measuring the steady state entanglement and coherence for coupled qubits systems in nonequilibrium environments. These results offer some general guidelines for optimizing the steady-state entanglement and coherence in the coupled qubit system and may find potential applications in quantum information technology.
quantum physics
In this paper we study the existence of solution for the following class of nonlocal problems \[ L_0u =f(x,u)+g(x) , \ \mbox{in} \ \Omega, \] where $\Omega \subset \mathbb{R}^{N}$, $N\geq 1$, is a bounded connected open, $g \in C(\overline{\Omega})$, $f:\overline{\Omega} \times \mathbb{R} \to \mathbb{R}$ are function, and $L_0 : C(\overline{\Omega}) \to C(\overline{\Omega})$ is a nonlocal dispersal operator. Using a sub-supersolution method and the degree theory for $\gamma$-Condensing maps, we have obtained a result of the Ambrosetti-Prodi type, that is, we obtain a necessary condition on $g$ for the non-existence of solutions, the existence of at least one solution, and the existence of at least two distinct solutions.
mathematics
Fundamental bounds on the performance of monochromatic scattering-cancellation and field-zeroing cloaks made of prescribed linear passive materials occupying a predefined design region are formulated by projecting field quantities onto a sub-sectional basis and applying quadratically constrained quadratic programming. Formulations are numerically tested revealing key physical trends as well as advantages and disadvantages between the two classes of cloaks. Results show that the use of low-loss materials with high dielectric contrast affords the highest potential for effective cloaking.
physics
Detecting changed regions in paired satellite images plays a key role in many remote sensing applications. The evolution of recent techniques could provide satellite images with very high spatial resolution (VHR) but made it challenging to apply image coregistration, and many change detection methods are dependent on its accuracy.Two images of the same scene taken at different time or from different angle would introduce unregistered objects and the existence of both unregistered areas and actual changed areas would lower the performance of many change detection algorithms in unsupervised condition.To alleviate the effect of unregistered objects in the paired images, we propose a novel change detection framework utilizing a special neural network architecture -- Generative Adversarial Network (GAN) to generate many better coregistered images. In this paper, we show that GAN model can be trained upon a pair of images through using the proposed expanding strategy to create a training set and optimizing designed objective functions. The optimized GAN model would produce better coregistered images where changes can be easily spotted and then the change map can be presented through a comparison strategy using these generated images explicitly.Compared to other deep learning-based methods, our method is less sensitive to the problem of unregistered images and makes most of the deep learning structure.Experimental results on synthetic images and real data with many different scenes could demonstrate the effectiveness of the proposed approach.
electrical engineering and systems science
Spectrum sensing is one of the means of utilizing the scarce source of wireless spectrum efficiently. In this paper, a convolutional neural network (CNN) model employing spectral correlation function which is an effective characterization of cyclostationarity property, is proposed for wireless spectrum sensing and signal identification. The proposed method classifies wireless signals without a priori information and it is implemented in two different settings entitled CASE1 and CASE2. In CASE1, signals are jointly sensed and classified. In CASE2, sensing and classification are conducted in a sequential manner. In contrary to the classical spectrum sensing techniques, the proposed CNN method does not require a statistical decision process and does not need to know the distinct features of signals beforehand. Implementation of the method on the measured overthe-air real-world signals in cellular bands indicates important performance gains when compared to the signal classifying deep learning networks available in the literature and against classical sensing methods. Even though the implementation herein is over cellular signals, the proposed approach can be extended to the detection and classification of any signal that exhibits cyclostationary features. Finally, the measurement-based dataset which is utilized to validate the method is shared for the purposes of reproduction of the results and further research and development.
electrical engineering and systems science
By embedding a $\cal PT$-symmetric (pseudo-Hermitian) system into a large Hermitian one, we disclose the relations between $\cal{PT}$-symmetric Hamiltonians and weak measurement theory. We show that the amplification effect in weak measurement on a conventional quantum system can be used to effectively simulate a local broken $\cal PT$-symmetric Hamiltonian system, with the pre-selected state in the $\cal PT$-symmetric Hamiltonian system and its post-selected state resident in the dilated Hamiltonian system.
quantum physics
Classical open systems with balanced gain and loss, i.e. parity-time ($\mathcal{PT}$) symmetric systems, have attracted tremendous attention over the past decade. Their exotic properties arise from exceptional point (EP) degeneracies of non-Hermitian Hamiltonians that govern their dynamics. In recent years, increasingly sophisticated models of $\mathcal{PT}$-symmetric systems with time-periodic (Floquet) driving, time-periodic gain and loss, and time-delayed coupling have been investigated, and such systems have been realized across numerous platforms comprising optics, acoustics, mechanical oscillators, optomechanics, and electrical circuits. Here, we introduce a $\mathcal{PT}$-symmetric (balanced gain and loss) system with memory, and investigate its dynamics analytically and numerically. Our model consists of two coupled $LC$ oscillators with positive and negative resistance, respectively. We introduce memory by replacing either the resistor with a memristor, or the coupling inductor with a meminductor, and investigate the circuit energy dynamics as characterized by $\mathcal{PT}$-symmetric or $\mathcal{PT}$-symmetry broken phases. Due to the resulting nonlinearity, we find that energy dynamics depend on the sign and strength of initial voltages and currents, as well as the distribution of initial circuit energy across its different components. Surprisingly, at strong inputs, the system exhibits self-organized Floquet dynamics, including $\mathcal{PT}$-symmetry broken phase at vanishingly small dissipation strength. Our results indicate that $\mathcal{PT}$-symmetric systems with memory show a rich landscape.
quantum physics
We consider a certain class of multiplicative functions $f: \mathbb N \rightarrow \mathbb C$ and study the distribution of zeros of Dirichlet polynomials $F_N(s)= \sum_{n\le N} f(n)n^{-s}$ corresponding to these functions. We prove that the known non-trivial zero-free half plane for Dirichlet polynomials associated to this class of multiplicative functions is optimal. We also introduce a characterization of elements in this class based on a new parameter depending on the Dirichlet series $F(s) = \sum_{n=1}^\infty f(n) n^{-s}$. In this context, we obtain non-trivial regions in which the associated Dirichlet polynomials do have zeros.
mathematics
Recent technological progress in the development of Unmanned Aerial Vehicles (UAVs) together with decreasing acquisition costs make the application of drone fleets attractive for a wide variety of tasks. In agriculture, disaster management, search and rescue operations, commercial and military applications, the advantage of applying a fleet of drones originates from their ability to cooperate autonomously. Multi-Agent Reinforcement Learning approaches that aim to optimize a neural network based control policy, such as the best performing actor-critic policy gradient algorithms, struggle to effectively back-propagate errors of distinct rewards signal sources and tend to favor lucrative signals while neglecting coordination and exploitation of previously learned similarities. We propose a Multi-Critic Policy Optimization architecture with multiple value estimating networks and a novel advantage function that optimizes a stochastic actor policy network to achieve optimal coordination of agents. Consequently, we apply the algorithm to several tasks that require the collaboration of multiple drones in a physics-based reinforcement learning environment. Our approach achieves a stable policy network update and similarity in reward signal development for an increasing number of agents. The resulting policy achieves optimal coordination and compliance with constraints such as collision avoidance.
computer science
We have witnessed in the past decade the observation of a puzzling cosmic-ray excess at energies larger than $10$ GeV. The AMS-02 data published this year has new ingredients such as the bump around $300$ GeV followed by a drop at $800$ GeV, as well as smaller error bars. Adopting the background used by the AMS-02 collaboration in their analysis, one can conclude that previous explanations to the new AMS-02 such as one component annihilating and decaying dark matter as well as pulsars seem to fail at reproducing the data. Here, we show that in the right-handed neutrino portal might reside the answer. We discuss a decaying two-component dark matter scenario where the two-body decay products are right-handed neutrinos that have their decay pattern governed by the type I seesaw mechanism. This setup provides a very good fit to data, for example, for a conservative approach including just statistical uncertainties leads to $\chi^2/d.o.f \sim 2.3$ for $m_{DM_1}=2150$ GeV with $\tau_{1}=3.78 \times 10^{26}$ s and $m_{DM_2}=300$ with $\tau_{2}=5.0 \times 10^{27}$ s for $M_N=10$ GeV, and, in an optimistic case, including systematic uncertainties, we find $\chi^2/d.o.f \sim 1.12$, for $M_N = 10$ GeV, with $m_{DM_1}=2200$ GeV with $\tau_{1}=3.8 \times 10^{26}$ s and $m_{DM_2}=323$ GeV with $\tau_{2}=1.68 \times 10^{27}$ s.
high energy physics phenomenology
Measure transport underpins several recent algorithms for posterior approximation in the Bayesian context, wherein a transport map is sought to minimise the Kullback--Leibler divergence (KLD) from the posterior to the approximation. The KLD is a strong mode of convergence, requiring absolute continuity of measures and placing restrictions on which transport maps can be permitted. Here we propose to minimise a kernel Stein discrepancy (KSD) instead, requiring only that the set of transport maps is dense in an $L^2$ sense and demonstrating how this condition can be validated. The consistency of the associated posterior approximation is established and empirical results suggest that KSD is competitive and more flexible alternative to KLD for measure transport.
statistics
Estimation of the number of components (or order) of a finite mixture model is a long standing and challenging problem in statistics. We propose the Group-Sort-Fuse (GSF) procedure---a new penalized likelihood approach for simultaneous estimation of the order and mixing measure in multidimensional finite mixture models. Unlike methods which fit and compare mixtures with varying orders using criteria involving model complexity, our approach directly penalizes a continuous function of the model parameters. More specifically, given a conservative upper bound on the order, the GSF groups and sorts mixture component parameters to fuse those which are redundant. For a wide range of finite mixture models, we show that the GSF is consistent in estimating the true mixture order and achieves the $n^{-1/2}$ convergence rate for parameter estimation up to polylogarithmic factors. The GSF is implemented for several univariate and multivariate mixture models in the R package GroupSortFuse. Its finite sample performance is supported by a thorough simulation study, and its application is illustrated on two real data examples.
statistics
We develop a novel approach for a Time Projection Chamber (TPC) concept suitable for deployment in kilotonne scale detectors, with a charge-readout system free from reconstruction ambiguities, and a robust TPC design that reduces high-voltage risks while increasing the coverage of the light collection system. This novel concept could be deployed as a Far Detector module in the Deep Underground Neutrino Experiment (DUNE) neutrino-oscillation experiment. For the charge-readout system, we use the charge-collection pixels and associated application-specific integrated circuits currently being developed for the liquid argon (LAr) component of the DUNE Near Detector design, ArgonCube. In addition, we divide the TPC into a number or shorter drift volumes, reducing the total voltage used to drift the ionisation electrons, and minimising the stored energy per TPC. Segmenting the TPC also contains scintillation light, allowing for precise trigger localisation and a more expansive light-readout system. Furthermore, the design opens the possibility of replacing or upgrading components. These augmentations could substantially improve reliability and sensitivity, particularly for low energy signals, in comparison to a traditional monolithic LArTPCs with projective charge-readout.
physics
We show that a pair of field theory monodromies in which the shift symmetry is broken by small, well motivated deformations, naturally incorporates a mechanism for cancelling off radiative corrections to the cosmological constant. The lighter monodromy sector plays the role of inflation as well as providing a rigid degree of freedom that acts as a dynamical counterterm for the cosmological constant. The heavier monodromy sector includes a rigid dilaton that forces a global constraint on the system and the cancellation of vacuum energy loops occurs at low energies via the sequestering mechanism. This suggests that monodromy constructions in string theory could be adapted to incorporate mechanisms to stabilise the cosmological constant in their low energy descriptions.
high energy physics theory
As a specific proportional hazard rates model, sequential order statistics can be used to describe the lifetimes of load-sharing systems. Inference for these systems needs to account for small sample sizes, which are prevalent in reliability applications. By exploiting the probabilistic structure of sequential order statistics, we derive exact finite sample inference procedures to test for the load-sharing parameters and for the nonparametrically specified baseline distribution, treating the respective other part as a nuisance quantity. This improves upon previous approaches for the model, which either assume a fully parametric specification or rely on asymptotic results. Simulations show that the tests derived are able to detect deviations from the null hypothesis at small sample sizes. Critical values for a prominent case are tabulated.
statistics