text
stringlengths
11
9.77k
label
stringlengths
2
104
The main purpose of this work is to identify the general quadratic operator which is invariant under the action of Linear Canonical Transformations (LCTs) in the framework of a relativistic quantum theory. LCTs are known in signal processing and optics as the transformations which generalize certain useful integral transforms. In quantum theory, they can be identified as the linear transformations which keep invariant the canonical commutation relations characterizing the coordinates and momenta operators. In this paper, LCTs corresponding to a general pseudo-Euclidian space are considered. Explicit calculations are performed for the monodimensional case to identify the corresponding LCT invariant operator then multidimensional generalizations of the obtained results are deduced. It was noticed that the introduction of a variance-covariance matrix, of coordinate and momenta operators, facilitate the establishment of a pseudo-orthogonal representation of LCTs which is useful for the identification of the invariant quadratic operator especially in the multidimensional case. According to the calculations carried out, the LCT invariant operator is a second order polynomial of the coordinates and momenta operators. The coefficients of this polynomial depend on the mean values and the statistical variances-covariances of these coordinates and momenta operators themselves. The eigenstates of the LCT invariant operator are also identified with it and some of the main possible applications of the obtained results are discussed.
quantum physics
To model the interior of a black hole, a study is made of a spin system with long-range random four-spin couplings that exhibits quantum chaos. The black hole limit corresponds to a system where the microstates are approximately degenerate and equally likely, corresponding to the high temperature limit of the spin system. At the leading level of approximation, reconstruction of bulk physics implies that local probes of the black hole should exhibit free propagation and unitary local evolution. We test the conjecture that a particular mean field Hamiltonian provides such a local bulk Hamiltonian by numerically solving the exact Schrodinger equation and comparing the time evolution to the approximate mean field time values. We find excellent agreement between the two time evolutions for timescales smaller than the scrambling time. In earlier work, it was shown bulk evolution along comparable timeslices is spoiled by the presence of the curvature singularity, thus the matching found in the present work provides evidence of the success of this approach to interior holography. The numerical solutions also provide a useful testing ground for various measures of quantum chaos and global scrambling. A number of different observables, such as entanglement entropy, out-of-time-order correlators, and trace distance are used to study these effects. This leads to a suitable definition of scrambling time, and evidence is presented showing a logarithmic variation with the system size.
high energy physics theory
In this thesis we study the momentum space approach to the solution of the CWI's of CFT's in higher dimensions. Our work's goal is to illustrate the essential steps needed to build tensor correlators starting from the scalar solutions, for 3-point functions. In the case of 4-point functions, our attention is centred around scalar correlators for which the CWI's are sufficient to isolate the unique solution if we enhance the symmetry with the addition of a dual conformal symmetry. Dual conformal symmetry in momentum space is obtained once the momentum variables are rewritten in a dual form, as the difference of coordinate-like variables and treated as ordinary correlators in such variables, mirroring the action of coordinate space. This enhancement of the symmetry is sufficient to fix the solutions also for such correlators. The solution of the conformal constraints are given in terms of triple-$K$ integrals and are expressed in terms of a set of constants, specific for each correlator and spacetime dimension. We present a discussion of the intermediate steps in the description of two nontrivial correlators, the $TTO$ and the $TTT$, in a more pedagogical way, offering details that could help extend such methods to higher point function.
high energy physics theory
Eliashberg's foundational theory of superconductivity is based on the application of Migdal's approximation, which states that vertex corrections to first order electron-phonon scattering are negligible if the ratio between phonon and electron energy scales is small. The resulting theory incorporates the first Feynman diagrams for electron and phonon self-energies. However, the latter is most commonly neglected in numerical analyses. Here we provide an extensive study of full-bandwidth Eliashberg theory in two and three dimensions, where we include the full back reaction of electrons onto the phonon spectrum. We unravel the complex interplay between nesting properties, Fermi surface density of states, renormalized electron-phonon coupling, phonon softening, and superconductivity. We propose furthermore a scaling law for the maximally possible critical temperature $T_c^{\textrm{max}}\propto\lambda (\Omega ) \sqrt{\Omega_0^2-\Omega^2}$ in 2D and 3D systems, which embodies both the renormalized electron-phonon coupling strength $\lambda(\Omega)$ and softened phonon spectrum $\Omega$. Also, we analyze for which electronic structure properties a maximal $T_c$ enhancement can be achieved.
condensed matter
Human dynamics and sociophysics build on statistical models that can shed light on and add to our understanding of social phenomena. We propose a generative model based on a stochastic differential equation that enables us to model the opinion polls leading up to the UK 2017 and 2019 general elections, and to make predictions relating to the actual result of the elections. After a brief analysis of the time series of the poll results, we provide empirical evidence that the gamma distribution, which is often used in financial modelling, fits the marginal distribution of this time series. We demonstrate that the proposed poll-based forecasting model may improve upon predictions based solely on polls. The method uses the Euler-Maruyama method to simulate the time series, measuring the prediction error with the mean absolute error and the root mean square error, and as such could be used as part of a toolkit for forecasting elections.
physics
Teaching by direct models in science has been weakening the learning process of the students, because the real problems in engineering are not solved by direct models instead commonly they are solved by inverse models. On the other hand, one of the most relevant topics in the course of waves and particle physics oriented for the forming engineers; it's the subject of simple harmonic motion forced damping, which many physical phenomena can be explained as the quality factor Q and the resonance frequency of an oscillatory forced system. In order to capture the attention of students and give an application to this issue. We have developed an experimental setup to take measurements of electric current, voltages from capacitor and inductor for different frequencies and resistances, once the experimental data were collected to study the behavior of the electrical current inside the circuit and find out the RLC parameters with an inverse model. Finally, we want to show the process in detail how parameters of the system (Resistance, Inductance and Capacitance values) are very relevant in this kind of systems, from the results obtained by experimental measurements of voltage, current and angle of phase shift, where this was achieved by implementing an indirect method described in this document, so that can be applied to studies of more complex systems such as a motor where such parameters may be unknown.
physics
We report the discovery of a set of four-point, two-factor, free-ranging, putatively IMSPE-optimal designs with a pair of twin points, in the statistical design of computer experiments, under Gaussian-process, fixed-Gaussian-covariance parameter, and zero-nugget assumptions. We conjecture this is the set of free-ranging, twin-point designs with the smallest number of degrees of freedom.
statistics
Traffic systems are complex systems that exhibit non-stationary characteristics. Therefore, the identification of temporary traffic states is significant. In contrast to the usual correlations of time series, here we study those of position series, revealing structures in time, i.e. the rich non-Markovian features of traffic. Considering the traffic system of the Cologne orbital motorway as a whole, we identify five quasi-stationary states by clustering reduced rank correlation matrices of flows using the $k$-means method. The five quasi-stationary states with nontrivial features include one holiday state, three workday states and one mixed state of holidays and workdays. In particular, the workday states and the mixed state exhibit strongly correlated time groups shown as diagonal blocks in the correlation matrices. We map the five states onto reduced-rank correlation matrices of velocities and onto traffic states where free or congested states are revealed in both space and time. Our study opens a new perspective for studying traffic systems. This contribution is meant to provide a proof of concept and a basis for further study.
physics
This white paper highlights the crucial and urgent synergies required between WFIRST, Subaru Hyper Suprime-Cam or other >25m-class telescopes galaxy observations and SKA 21cm measurements to constrain the nature of reionization (ionization history and topology) and its sources.
astrophysics
In a nonlinear three-wave mixing process, the interacting waves can accumulate an adiabatic geometric phase (AGP) if the nonlinear coefficient of the crystal is modulated in a proper manner along the nonlinear crystal. This concept was studied so far only for the case in which the pump wave is much stronger than the two other waves, hence can be assumed to be constant. Here we extend this analysis for the fully nonlinear process, in which all three waves can be depleted and we show that the sign and magnitude of the AGP can be controlled by the period, phase and duty cycle of the nonlinear modulation pattern. In this fully nonlinear interaction, all the states of the system can be mapped onto a closed surface. Specifically, we study a process in which the eigenstate of the system follows a circular rotation on the surface. Our analysis reveals that the AGP equals to the difference between the total phase accumulated along the circular trajectory and that along its vertical projection, which is universal for the undepleted (linear) and depleted (nonlinear) cases. Moreover, the analysis reveals that the AGPs in the processes of sum-frequency generation and difference-frequency generation have opposite chirality. Finally, we utilize the AGP in the fully nonlinear case for splitting the beam into different diffraction orders in the far field.
physics
The blue straggler stars (BSSs) are main-sequence (MS) stars, which have evaded stellar evolution by acquiring mass while on the MS. The detection of extremely low mass (ELM) white dwarf (WD) companions to two BSSs and one yellow straggler star (YSS) from our earlier study using UVIT/ASTROSAT, as well as WD companions to main-sequence stars (known as blue lurkers) suggest a good fraction of post-mass transfer binaries in M67. Using deeper UVIT observations, here we report the detection of another blue lurker in M67, with an ELM WD companion. The post-mass transfer systems with the presence of ELM WDs, including BSSs, are formed from Case A/B mass transfer and are unlikely to show any difference in surface abundances. We find a correlation between the temperature of the WD and the $v\ sin i$ of the BSSs. We also find that the progenitors of the massive WDs are likely to belong to the hot and luminous group of BSSs in M67. The only detected BSS+WD system by UVIT in the globular cluster NGC 5466, has a normal WD and suggests that open cluster like environment might be present in the outskirts of low density globular clusters.
astrophysics
We discuss the linear hydrodynamic response of a two-dimensional active chiral compressible fluid with odd viscosity. The viscosity coefficient represents broken time-reversal and parity symmetries in the 2D fluid and characterizes the deviation of the system from a passive fluid. Taking into account the hydrodynamic coupling to the underlying bulk fluid, we obtain the odd viscosity-dependent mobility tensor, which is responsible for the non-reciprocal hydrodynamic response to a point force. Furthermore, we consider a finite-size disk moving laterally in the 2D fluid and demonstrate that the disk experiences a non-dissipative lift force in addition to the dissipative drag one.
physics
We give a construction of a large first-order definable family of subrings of finitely generated fields $K$ of any characteristic. We deduce that for any such $K$ there exists a first-order sentence $\varphi_K$ characterising $K$ in the class of finitely generated fields, i.e. such that for any finitely generated field $L$ we have $L \models \varphi_K$ if and only if $L \cong K$. This answers a question considered by Pop and others. In characteristic two, our results depend on resolution of singularities, whereas they are unconditional in all other characteristics.
mathematics
In this paper, we propose a robust Kalman filtering framework for systems with probabilistic uncertainty in system parameters. We consider two cases, namely discrete time systems, and continuous time systems with discrete measurements. The uncertainty, characterized by mean and variance of the states, is propagated using conditional expectations and polynomial chaos expansion framework. The results obtained using the proposed filter are compared with existing robust filters in the literature. The proposed filter demonstrates better performance in terms of root mean squared error and rate of convergence.
electrical engineering and systems science
Using Arakelov geometry, we compute the partition function of the noncompact free boson at genus two. We begin by compiling a list of modular invariants which appear in the Arakelov theory of Riemann surfaces. Using these quantities, we express the genus two partition function as a product of modular forms, as in the well-known genus one case. We check that our result has the expected obstruction to holomorphic factorization and behavior under degeneration.
high energy physics theory
This paper proposes a deep neural network approach for predicting multiphase flow in heterogeneous domains with high computational efficiency. The deep neural network model is able to handle permeability heterogeneity in high dimensional systems, and can learn the interplay of viscous, gravity, and capillary forces from small data sets. Using the example of carbon dioxide (CO2) storage, we demonstrate that the model can generate highly accurate predictions of a CO2 saturation distribution given a permeability field, injection duration, injection rate, and injection location. The trained neural network model has an excellent ability to interpolate and to a limited extent, the ability to extrapolate beyond the training data ranges. To improve the prediction accuracy when the neural network model needs to extrapolate, we propose a transfer learning (fine-tuning) procedure that can quickly teach the neural network model new information without going through massive data collection and retraining. Based on this trained neural network model, a web-based tool is provided that allows users to perform CO2-water multiphase flow calculations online. With the tools provided in this paper, the deep neural network approach can provide a computationally efficient substitute for repetitive forward multiphase flow simulations, which can be adopted to the context of history matching and uncertainty quantification.
computer science
In competing event settings, a counterfactual contrast of cause-specific cumulative incidences quantifies the total causal effect of a treatment on the event of interest. However, effects of treatment on the competing event may indirectly contribute to this total effect, complicating its interpretation. We previously proposed the separable effects (Stensrud et al, 2019) to define direct and indirect effects of the treatment on the event of interest. This definition presupposes a treatment decomposition into two components acting along two separate causal pathways, one exclusively outside of the competing event and the other exclusively through it. Unlike previous definitions of direct and indirect effects, the separable effects can be subject to empirical scrutiny in a study where separate interventions on the treatment components are available. Here we extend and generalize the notion of the separable effects in several ways, allowing for interpretation, identification and estimation under considerably weaker assumptions. We propose and discuss a definition of separable effects that is applicable to general time-varying structures, where the separable effects can still be meaningfully interpreted, even when they cannot be regarded as direct and indirect effects. We further derive weaker conditions for identification of separable effects in observational studies where decomposed treatments are not yet available; in particular, these conditions allow for time-varying common causes of the event of interest, the competing events and loss to follow-up. For these general settings, we propose semi-parametric weighted estimators that are straightforward to implement. As an illustration, we apply the estimators to study the separable effects of intensive blood pressure therapy on acute kidney injury, using data from a randomized clinical trial.
statistics
The ability to make optimal decisions under uncertainty remains important across a variety of disciplines from portfolio management to power engineering. This generally implies applying some safety margins on uncertain parameters that may only be observable through a finite set of historical samples. Nevertheless, the optimized decisions must be resilient to all probable outcomes, while ideally providing some measure of severity of any potential violations in the less probable outcomes.It is known that the conditional value-at-risk (CVaR) can be used to quantify risk in an optimization task, though may also impose overly conservative margins. Therefore, this paper develops a means of co-controlling the value-at-risk (VaR) level associated with the CVaR to guarantee resilience in probable cases while providing a measure of the average violation in less probable cases. To further combat uncertainty, the CVaR and VaR co-control is extended in a distributionally robust manner using the Wasserstein metric to establish an ambiguity set constructed from finite samples, which is guaranteed to contain the true distribution with a certain confidence.
electrical engineering and systems science
We designed a superposition calculus for a clausal fragment of extensional polymorphic higher-order logic that includes anonymous functions but excludes Booleans. The inference rules work on $\beta\eta$-equivalence classes of $\lambda$-terms and rely on higher-order unification to achieve refutational completeness. We implemented the calculus in the Zipperposition prover and evaluated it on TPTP and Isabelle benchmarks. The results suggest that superposition is a suitable basis for higher-order reasoning.
computer science
In this paper, we extend our PrInDT method (Weihs, Buschfeld 2021) towards additional undersampling of one of the predictors. This helps us to handle multiple unbalanced data sets, i.e. data sets that are not only unbalanced with respect to the class variable but also in one of the predictor variables. Beyond the advantages of such an approach, our study reveals that the balanced accuracy in the full data set can be much lower than in the predictor undersamples. We discuss potential reasons for this problem and draw methodological conclusions for linguistic studies.
statistics
For a symplectic manifold satisfying some topological condition,we define a special class of modules over the deformation quantization algebra. For any two such modules we construct an infinity local system of morphisms. We construct such special module starting from a Lagrangian submanifold satisfying a topological condition. We compare the result to Morse theory, to the microlocal category of sheaves recently defined by Tamarkin, and to the Fukaya category of the two-dimensional torus. Several appendices explain the motivations that come from asymptotic analysis of pseudo-differential operators and distributions.
mathematics
In recent work Alexandre, Ellis, Millington and Seynaeve have extended the Goldstone theorem to non-Hermitian Hamiltonians that possess a discrete antilinear symmetry such as $PT$. They restricted their discussion to those realizations of antilinear symmetry in which all the energy eigenvalues of the Hamiltonian are real. Here we extend the discussion to the two other realizations possible with antilinear symmetry, namely energies in complex conjugate pairs or Jordan-block Hamiltonians that are not diagonalizable at all. In particular, we show that under certain circumstances it is possible for the Goldstone boson mode itself to be one of the zero-norm states that are characteristic of Jordan-block Hamiltonians. While we discuss the same model as Alexandre et al. our treatment is quite different, though their main conclusion that one can have Goldstone bosons in the non-Hermitian case remains intact. In their paper Alexandre et al. presented a variational procedure for the action in which the surface term played an explicit role, to thus suggest that one has to use such a procedure in order to establish the Goldstone theorem in the non-Hermitian case. However, by taking certain fields that they took to be Hermitian to actually either be anti-Hermitian or be made so by a similarity transformation, we show that we are then able to obtain a Goldstone boson using a completely standard variational procedure. Since we use a standard variational procedure we can readily extend our analysis to a continuous local symmetry by introducing a gauge boson. We show that when we do this the gauge boson acquires a non-zero mass by the Higgs mechanism in all realizations of the antilinear symmetry, except the one where the Goldstone boson itself has zero norm, in which case, and despite the fact that the continuous local symmetry has been spontaneously broken, the gauge boson remains massless.
high energy physics theory
A new ro-vibrational line list for the ground electronic state of the main isotopologue of acetylene, $^{12}$C$_2$H$_2$, is computed as part of the ExoMol project. The aCeTY line list covers the transition wavenumbers up to 10,000 cm$^{-1}$ ($ \lambda >1$ $\mu$m), with lower and upper energy levels up to 12,000 cm$^{-1}$ and 22,000 cm$^{-1}$ considered, respectively. The calculations are performed up to a maximum value for the vibrational angular momentum, $K_{\rm max}=L_{\rm max}$ = 16, and maximum rotational angular momentum, $J$ = 99. Higher values of $J$ were not within the specified wavenumber window. The aCeTY line list is considered to be complete up to 2200 K, making it suitable for use in characterising high-temperature exoplanet or cool stellar atmospheres. Einstein-A coefficients, which can directly be used to calculate intensities at a particular temperature, are computed for 4.3 billion (4,347,381,911) transitions between 5 million (5,160,803) energy levels. We make comparisons against other available data for $^{12}$C$_2$H$_2$, and demonstrate this to be the most complete line list available. The line list is available in electronic form from the online CDS and ExoMol databases.
astrophysics
We compute the four-derivative corrections to the geometry, extremality bound, and thermodynamic quantities of AdS-Reissner-Nordstr{\"o}m black holes for general dimensions and horizon geometries. We confirm the universal relationship between the extremality shift at fixed charge and the shift of the microcanonical entropy, and discuss the consequences of this relation for the Weak Gravity Conjecture in AdS. The thermodynamic corrections are calculated using two different methods: first by explicitly solving the higher-derivative equations of motion and second, by evaluating the higher-derivative Euclidean on-shell action on the leading-order solution. In both cases we find agreement, up to the addition of a Casimir energy in odd dimensions. We derive the bounds on the four-derivative Wilson coefficients implied by the conjectured positivity of the leading corrections to the microcanonical entropy of thermodynamically stable black holes. These include the requirement that the coefficient of Riemann-squared is positive, meaning that the positivity of the entropy shift is related to the condition that $c - a$ is positive in the dual CFT. We discuss implications for the deviation of $\eta/s$ from its universal value and a potential lower bound.
high energy physics theory
Replica geometries are not rigid when gravity is dynamical. We numerically construct $1<n\leq 2$ replica saddles in $2d$ gravity coupled to a CFT and compare the resulting Renyi entropies with the field theory result.
high energy physics theory
The ability to form images through hair-thin optical fibres promises to open up new applications from biomedical imaging to industrial inspection. Unfortunately, deployment has been limited because small changes in mechanical deformation (e.g. bending) and temperature can completely scramble optical information, distorting images. Since such changes are dynamic, correcting them requires measurement of the fibre transmission matrix (TM) in situ immediately before imaging. TM calibration typically requires access to both the proximal and distal facets of the fibre simultaneously, which is not feasible during most realistic usage scenarios without compromising the thin form factor with bulky distal optics. Here, we introduce a new approach to determine the TM of multi-mode fibre (MMF) or multi-core fibre (MCF) in a reflection-mode configuration without access to the distal facet. A thin stack of structured metasurface reflectors is used at the distal facet to introduce wavelength-dependent, spatially heterogeneous reflectance profiles. We derive a first-order fibre model that compensates these wavelength-dependent changes in the TM and show that, consequently, the reflected data at 3 wavelengths can be used to unambiguously reconstruct the full TM by an iterative optimisation algorithm. We then present a method for sample illumination and imaging following TM reconstruction. Unlike previous approaches, our method does not require the TM to be unitary making it applicable to physically realistic fibre systems. We demonstrate TM reconstruction and imaging first using simulated non-unitary fibres and noisy reflection matrices, then using much larger experimentally-measured TMs of a densely-packed MCF, and finally on an experimentally-measured multi-wavelength set of TMs recorded from a MMF. Our findings pave the way for online transmission matrix calibration in situ in hair-thin optical fibres
physics
We study the link between parton dynamics in the collinear limit and the logarithmically enhanced terms of the groomed jet mass distribution, for jets groomed with the modified mass-drop tagger (mMDT). While the leading-logarithmic (LL) result is linked to collinear evolution with leading-order splitting kernels, here we derive the NLL structure directly from triple-collinear splitting kernels. The calculation we present is a fixed-order calculation in the triple-collinear limit, independent of resummation ingredients and methods. It therefore constitutes a powerful cross-check of the NLL results previously derived using the SCET formalism and provides much of the insight needed for resummation within the traditional QCD approach.
high energy physics phenomenology
We investigate the steady state properties arising from the open system dynamics described by a memoryless (Markovian) quantum collision model, corresponding to a master equation in the ultra-strong coupling regime. By carefully assessing the work cost of switching on and off the system-environment interaction, we show that only a coupling Hamiltonian in the energy-preserving form drives the system to thermal equilibrium, while any other interaction leads to non-equilibrium steady states that are supported by steady-state currents. These currents provide a neat exemplification of the housekeeping work and heat. Furthermore, we characterize the specific form of system-environment interaction that drives the system to a steady-state exhibiting coherence in the energy eigenbasis, thus, giving rise to families of states that are non-passive.
quantum physics
We present a new vacuum solution of Einstein's equations describing the near horizon region of two neutral, extreme (zero-temperature), co-rotating, non-identical Kerr black holes. The metric is stationary, asymptotically near horizon extremal Kerr (NHEK), and contains a localized massless strut along the symmetry axis between the black holes. In the deep infrared, it flows to two separate throats which we call "pierced-NHEK" geometries: each throat is NHEK pierced by a conical singularity. We find that in spite of the presence of the strut for the pierced-NHEK geometries the isometry group SL(2,R)xU(1) is restored. We find the physical parameters and entropy.
high energy physics theory
The line bundle $12M-D$ on the perfect compactification $A_g^P$ is nef; we show here that in positive characteristic it is semi-ample and that in all characteristics its exceptional locus is the closure of the locus of abelian varieties with an elliptic factor.
mathematics
When suppressing the itinerant antiferromagnetism in chromium by doping with the isostructual itinerant ferromagnet iron, a dome of spin-glass behavior emerges around a putative quantum critical point at an iron concentration $x \approx 0.15$. Here, we report a comprehensive investigation of polycrystalline samples of Fe$_{x}$Cr$_{1-x}$ in the range $0.05 \leq x \leq 0.30$ using x-ray powder diffraction, magnetization, ac susceptibility, and neutron depolarization measurements, complemented by specific heat and electrical resistivity data for $x = 0.15$. Besides antiferromagnetic ($x < 0.15$) and ferromagnetic regimes ($0.15 \leq x$), we identify a dome of reentrant spin-glass behavior at low temperatures for $0.10 \leq x \leq 0.25$ that is preceded by a precursor phenomenon. Neutron depolarization indicates an increase of the size of ferromagnetic clusters with increasing $x$ and the Mydosh parameter $\phi$, inferred from the ac susceptibility, implies a crossover from cluster-glass to superparamagnetic behavior. Taken together, these findings consistently identify Fe$_{x}$Cr$_{1-x}$ as an itinerant-electron system that permits to study the evolution of spin-glass behavior of gradually varying character in unchanged crystalline environment.
condensed matter
The short-time Fourier transform (STFT) usually computes the same number of frequency components as the frame length while overlapping adjacent time frames by more than half. As a result, the number of components of a spectrogram matrix becomes more than twice the signal length, and hence STFT is hardly used for signal compression. In addition, even if we modify the spectrogram into a desired one by spectrogram-based signal processing, it is re-changed during the inversion as long as it is outside the range of STFT. In this paper, to reduce the number of components of a spectrogram while maintaining the analytical ability, we propose the frequency-undersampled STFT (FUSTFT), which computes only half the frequency components. We also present the inversions with and without the periodic condition, including their different properties. In simple numerical examples of audio signals, we confirm the validity of FUSTFT and the inversions.
electrical engineering and systems science
This paper reviews the NTIRE 2020 challenge on real world super-resolution. It focuses on the participating methods and final results. The challenge addresses the real world setting, where paired true high and low-resolution images are unavailable. For training, only one set of source input images is therefore provided along with a set of unpaired high-quality target images. In Track 1: Image Processing artifacts, the aim is to super-resolve images with synthetically generated image processing artifacts. This allows for quantitative benchmarking of the approaches \wrt a ground-truth image. In Track 2: Smartphone Images, real low-quality smart phone images have to be super-resolved. In both tracks, the ultimate goal is to achieve the best perceptual quality, evaluated using a human study. This is the second challenge on the subject, following AIM 2019, targeting to advance the state-of-the-art in super-resolution. To measure the performance we use the benchmark protocol from AIM 2019. In total 22 teams competed in the final testing phase, demonstrating new and innovative solutions to the problem.
electrical engineering and systems science
Let $u$ be a weak solution of the free boundary problem $$\mathcal L u=\lambda_0 \mathcal H^1\lfloor\partial\{u>0\}, u\ge 0,$$ where $\mathcal L u={\text{div}}(g(\nabla u)\nabla u)$ is a quasilinear elliptic operator and $g(\xi)$ is a given function of $\xi$ satisfying some structural conditions. We prove that the free boundary $\partial\{ u>0\}$ is continuously differentiable in $\mathbb R^2$, provided that $\partial\{ u>0\}$ has locally finite connectivity. Moreover, we show that the free boundaries of weak solutions with finite $\it{Morse \ index}$ must have finite connectivity. The weak solutions are locally Lipschitz continuous and non-degenerate stationary points of the Alt-Caffarelli type functional $J[u]=\int_{\Omega}F(\nabla u)+Q^2\chi_{\{ u>0\}}$. The full regularity of the free boundary is not fully understood even for the {\it minimizers} of $J[u]$ in the simplest case $g(\xi)=|\xi|^{p-2}, p>1$, partly because the methods from the classical case $p=2$ cannot be generalized to the full range of $p$. Our method, however, is very geometric and works even for the $ stationary\ points$ of the functional $J[u]$ for a large class of nonlinearities $F$.
mathematics
In this paper, we suggest a method of a complete calculation of the Sommerfeld effect in both the s-wave and p-wave annihilation situations. The basic idea is to uniformly consider the equal-time Beth-Salpeter wave functions of both the dark matter and the final state particles, then short-distance interactions can be parametrized as a cross-term considering the equations. Solving these equations will result in the complete formulas of the cross sections with the Sommerfeld effects.
high energy physics phenomenology
We provide a framework for understanding the gapless Kitaev spin liquid (KSL) in the language of tensor network(TN). Without introducing Majorana fermion, most of the features of the KSL including the symmetries, gauge structure, criticality and vortex-freeness are explained in a compact TN representation. Our construction reveals a hidden string gas structure of the KSL. With only two variational parameters to adjust, we obtain an accurate KSL ansatz with the bond dimension D = 8 in a compact form, where the energy is about 0.007% higher than the exact one. In addition, the opening of gap and non-Abelian phase driven by a magnetic field are naturally understood in our construction.
condensed matter
Deepening our knowledge of the partonic content of nucleons and nuclei represents a central endeavour of modern high-energy and nuclear physics, with ramifications in related disciplines such as astroparticle physics. There are two main scientific drivers motivating these investigations of the partonic structure of hadrons. On the one hand, addressing fundamental open issues in our understanding in the strong interactions such as the origin of the nucleon mass, spin, and transverse structure; the presence of heavy quarks in the nucleon wave function; and the possible onset of novel gluon-dominated dynamical regimes. On the other hand, pinning down with the highest possible precision the substructure of nucleons and nuclei is a central component for theoretical predictions in a wide range of experiments, from proton and heavy ion collisions at the Large Hadron Collider to ultra-high energy neutrino interactions at neutrino telescopes. This Article presents a succinct non-technical overview of our modern understanding of the quark, gluon, and photon substructure of nucleons and nuclei, focusing on recent trends and results and discussing future perspectives for the field.
high energy physics phenomenology
In this paper, we simultaneously determine the optimal sensor precision and the observer gain, which achieves the specified accuracy in the state estimates. Along with the unknown observer gain, the formulation parameterizes the scaling of the exogenous inputs that correspond to the sensor noise. Reciprocal of this scaling is defined as the sensor precision, and sparseness is achieved by minimizing the $l_1$ norm of the precision vector. The optimization is performed with constraints guaranteeing specified accuracy in state estimates, which are defined in terms of $\mathcal{H}_2$ or $\mathcal{H}_{\infty}$ norms of the error dynamics. The results presented in this paper are applied to the linearized longitudinal model of an F-16 aircraft.
electrical engineering and systems science
Inverse problems consist of recovering a signal from a collection of noisy measurements. These are typically cast as optimization problems, with classic approaches using a data fidelity term and an analytic regularizer that stabilizes recovery. Recent Plug-and-Play (PnP) works propose replacing the operator for analytic regularization in optimization methods by a data-driven denoiser. These schemes obtain state of the art results, but at the cost of limited theoretical guarantees. To bridge this gap, we present a new algorithm that takes samples from the manifold of true data as input and outputs an approximation of the projection operator onto this manifold. Under standard assumptions, we prove this algorithm generates a learned operator, called Wasserstein-based projection (WP), that approximates the true projection with high probability. Thus, WPs can be inserted into optimization methods in the same manner as PnP, but now with theoretical guarantees. Provided numerical examples show WPs obtain state of the art results for unsupervised PnP signal recovery.
computer science
Both, human appreciation of music and musical genres, transcend time and space. The universality of musical genres and associated musical scales is intimately linked to the physics of sound and the special characteristics of human acoustic sensitivity. In this series of articles, we examine the science underlying the development of the heptatonic scale, one of the most prevalent scales of the modern musical genres, both western and Indian.
physics
A new framework is proposed to study rank-structured matrices arising from discretizations of 2D and 3D elliptic operators. In particular, we introduce the notion of a graph-induced rank structure (GIRS) which aims to capture the fine low rank structures which appear in sparse matrices and their inverses in terms of the adjacency graph $\mathbb{G}$. We show that the GIRS property is invariant under inversion, and hence any effective representation of the inverse of GIRS matrices would lead to effective solvers. Starting with the observation that sequentially semi-separable (SSS) matrices form a good candidate for representing GIRS matrices on the line graph, we propose two extensions of SSS matrices to arbitrary graphs: Dewilde-van der Veen (DV) representations and $\mathbb{G}$-semi-separable ($\mathbb{G}$-SS) representations. It is shown that both these representations come naturally equipped with fast solvers where the solve complexity is commensurate to fast sparse Gaussian elimination on the graph $\mathbb{G}$, and $\mathbb{G}$-SS representations have a linear time multiplication algorithm. We show the construction of these representations to be highly nontrivial by determining the minimal $\mathbb{G}$-SS representation for the cycle graph $\mathbb{G}$. To obtain a minimal representation, we solve an exotic variant of a low-rank completion problem.
mathematics
We examine the Regge limit of holographic 4-point correlation functions in AdS$_3\times S^3$ involving two heavy and two light operators. In this kinematic regime such correlators can be reconstructed from the bulk phase shift accumulated by the light probe as it traverses the geometry dual to the heavy operator. We work perturbatively -- but to arbitrary orders -- in the ratio of the heavy operator's conformal dimension to the dual CFT${}_2$'s central charge, thus going beyond the low order results of arXiv:1812.03120 and arXiv:2007.12118. In doing so, we derive all-order relations between the bulk phase shift and the Regge limit OPE data of a class of heavy-light multi-trace operators exchanged in the cross-channel. Furthermore, we analyse two examples for which the relevant 4-point correlators are known explicitly to all orders: firstly the case of heavy operators dual to AdS${}_3$ conical defect geometries and secondly the case of non-trivial smooth geometries representing microstates of the two-charge D1-D5 black hole.
high energy physics theory
Fe$_{1+x}$Te is a two dimensional van der Waals antiferromagnet that becomes superconducting on anion substitution on the Te site. The parent phase of Fe$_{1+x}$Te is sensitive to the amount of interstitial iron situated between the iron-tellurium layers displaying collinear magnetic order coexisting with low temperature metallic resistivity for small concentrations of interstitial iron $x$ and helical magnetic order for large values of $x$. While this phase diagram has been established through scattering [see for example E. E. Rodriguez $\textit{et al.}$ Phys. Rev. B ${\bf{84}}$, 064403 (2011) and S. R\"ossler $\textit{et al.}$ Phys. Rev. B ${\bf{84}}$, 174506 (2011)], recent scanning tunnelling microscopy measurements [C. Trainer $\textit{et al.}$ Sci. Adv. ${\bf{5}}$, eaav3478 (2019)] have observed a different magnetic structure for small interstitial iron concentrations $x$ with a significant canting of the magnetic moments along the crystallographic $c$ axis of $\theta$=28 $\pm$ 3$^{\circ}$. In this paper, we revisit the magnetic structure of Fe$_{1.09}$Te using spherical neutron polarimetry and scanning tunnelling microscopy to search for this canting in the bulk phase and compare surface and bulk magnetism. The results show that the bulk magnetic structure of Fe$_{1.09}$Te is consistent with collinear in-plane order ($\theta=0$ with an error of $\sim$ 5$^{\circ}$). Comparison with scanning tunnelling microscopy on a series of Fe$_{1+x}$Te samples reveals that the surface exhibits a magnetic surface reconstruction with a canting angle of the spins of $\theta=29.8^{\circ}$. We suggest that this is a consequence of structural relaxation of the surface layer resulting in an out-of-plane magnetocrystalline anisotropy. The magnetism in Fe$_{1+x}$Te displays different properties at the surface when the symmetry constraints of the bulk are removed.
condensed matter
Future generation wireless networks are targeting the convergence of fixed, mobile and broadcasting systems with the integration of satellite and terrestrial systems towards utilizing their mutual benefits. Satellite Communications (Sat- Com) is envisioned to play a vital role to provide integrated services seamlessly over heterogeneous networks. As compared to terrestrial systems, the design of SatCom systems require a different approach due to differences in terms of wave propagation, operating frequency, antenna structures, interfering sources, limitations of onboard processing, power limitations and transceiver impairments. In this regard, this letter aims to identify and discuss important modeling and design aspects of the next generation High Throughput Satellite (HTS) systems. First, communication models of HTSs including the ones for multibeam and multicarrier satellites, multiple antenna techniques, and for SatCom payloads and antennas are highlighted and discussed. Subsequently, various design aspects of SatCom transceivers including impairments related to the transceiver, payload and channel, and traffic-based coverage adaptation are presented. Finally, some open topics for the design of next generation HTSs are identified and discussed.
electrical engineering and systems science
We propose two-loop neutrino mass model with gauged hidden $U(1)$ symmetry, and discuss a Majorana type of dark matter candidate that has semi-annihilation processes in the relic density as well as lepton flavor violations and muon anomalous magnetic moment. Also, we demonstrate global analysis to satisfy neutrino oscillation data, lepton flavor violations, and relic density of dark matter candidate and show that semi-annihilation modes play a crucial role in finding observed relic density.
high energy physics phenomenology
We have used the capability of the MUSE instrument to explore the impact of stellar feedback at large scales in Haro 11, a galaxy under extreme starburst condition and one of the first galaxies where Lyman continuum (LyC) has been detected. Using Ha, [OIII] and [OI] emission lines from deep MUSE observations, we have constructed a sequence of velocity-dependent maps of the Ha emission, the state of the ionised gas and a tracer of fast shocks. These allowed us to investigate the ionisation structure of the galaxy in 50 kms^2 bins over a velocity range of -400 to 350 kms. The ionised gas in Haro 11 is assembled by a rich arrangement of structures, such as superbubbles, filaments, arcs and galactic ionised channels, whose appearances change drastically with velocity. The central star forming knots and the star forming dusty arm are the main engines that power the strong mechanical feedback in this galaxy, although with different impact on the ionisation structure. Haro 11 appears to leak LyC radiation in many directions. We found evidence of a kpc-scale fragmented superbubble, that may have cleared galactic-scale channels in the ISM. Additionally, the southwestern hemisphere is highly ionised in all velocities, hinting at a density bound scenario. A compact kpc-scale structure of lowly ionised gas coincides with the diffuse Lya emission and the presence of fast shocks. Finally, we find evidence that a significant fraction of the ionised gas mass may escape the gravitational potential of the galaxy.
astrophysics
We read with interest the article by di Cecco et al. (2018), but have reservations about the usefulness of the latent class model specifically for estimating overcoverage. In particular, we question the interpretation of the parameters of the fitted latent class model.
statistics
Over the last decade of machine learning, convolutional neural networks have been the most striking successes for feature extraction of rich sensory and high-dimensional data. While learning data representations via convolutions is already well studied and efficiently implemented in various deep learning libraries, one often faces limited memory capacity and insufficient number of training data, especially for high-dimensional and large-scale tasks. To overcome these limitations, we introduce a network architecture using a self-adjusting and data dependent version of the Radon-transform (linear data projection), also known as x-ray projection, to enable feature extraction via convolutions in lower-dimensional space. The resulting framework, named PiNet, can be trained end-to-end and shows promising performance on volumetric segmentation tasks. We test proposed model on public datasets to show that our approach achieves comparable results only using fractional amount of parameters. Investigation of memory usage and processing time confirms PiNet's superior efficiency compared to other segmentation models.
electrical engineering and systems science
When developing autonomous systems, engineers and other stakeholders make great effort to prepare the system for all foreseeable events and conditions. However, these systems are still bound to encounter events and conditions that were not considered at design time. For reasons like safety, cost, or ethics, it is often highly desired that these new situations be handled correctly upon first encounter. In this paper we first justify our position that there will always exist unpredicted events and conditions, driven among others by: new inventions in the real world; the diversity of world-wide system deployments and uses; and, the non-negligible probability that multiple seemingly unlikely events, which may be neglected at design time, will not only occur, but occur together. We then argue that despite this unpredictability property, handling these events and conditions is indeed possible. Hence, we offer and exemplify design principles that when applied in advance, can enable systems to deal, in the future, with unpredicted circumstances. We conclude with a discussion of how this work and a broader theoretical study of the unexpected can contribute toward a foundation of engineering principles for developing trustworthy next-generation autonomous systems.
computer science
Domain Adaptation (DA) has the potential to greatly help the generalization of deep learning models. However, the current literature usually assumes to transfer the knowledge from the source domain to a specific known target domain. Domain Agnostic Learning (DAL) proposes a new task of transferring knowledge from the source domain to data from multiple heterogeneous target domains. In this work, we propose the Domain-Agnostic Learning framework with Anatomy-Consistent Embedding (DALACE) that works on both domain-transfer and task-transfer to learn a disentangled representation, aiming to not only be invariant to different modalities but also preserve anatomical structures for the DA and DAL tasks in cross-modality liver segmentation. We validated and compared our model with state-of-the-art methods, including CycleGAN, Task Driven Generative Adversarial Network (TD-GAN), and Domain Adaptation via Disentangled Representations (DADR). For the DA task, our DALACE model outperformed CycleGAN, TD-GAN ,and DADR with DSC of 0.847 compared to 0.721, 0.793 and 0.806. For the DAL task, our model improved the performance with DSC of 0.794 from 0.522, 0.719 and 0.742 by CycleGAN, TD-GAN, and DADR. Further, we visualized the success of disentanglement, which added human interpretability of the learned meaningful representations. Through ablation analysis, we specifically showed the concrete benefits of disentanglement for downstream tasks and the role of supervision for better disentangled representation with segmentation consistency to be invariant to domains with the proposed Domain-Agnostic Module (DAM) and to preserve anatomical information with the proposed Anatomy-Preserving Module (APM).
electrical engineering and systems science
We generalise the work of 1810.11442 for the case of AdS$_7$/CFT$_6$. Starting from the 2-equivalent charge, 3-equivalent rotation non-extremal black-hole solution in 7D gauged supergravity, we consider the supersymmetric and then the extremal limit and evaluate the associated thermodynamic quantities. Away from extremality, the black-hole solution becomes complex. The entropy is then given by the Legendre transform of the on-shell action with respect to two complex chemical potentials subject to a constraint. At the conformal boundary we derive the dual background and evaluate the corresponding partition function for the $A_{N-1}$ 6D (2,0) theory at large $N$ in a Cardy-like limit. This is carried out via a 5D $\mathcal N=2$ super Yang-Mills calculation on $S^5$. The gravitational on-shell action is found to be exactly reproduced by the boundary partition function at large $N$. We argue that this agreement puts strong constraints on the form of possible higher-derivative corrections to the 5D gauge theory that is used in the $S^5$ evaluation.
high energy physics theory
Grade prediction for future courses not yet taken by students is important as it can help them and their advisers during the process of course selection as well as for designing personalized degree plans and modifying them based on their performance. One of the successful approaches for accurately predicting a student's grades in future courses is Cumulative Knowledge-based Regression Models (CKRM). CKRM learns shallow linear models that predict a student's grades as the similarity between his/her knowledge state and the target course. A student's knowledge state is built by linearly accumulating the learned provided knowledge components of the courses he/she has taken in the past, weighted by his/her grades in them. However, not all the prior courses contribute equally to the target course. In this paper, we propose a novel Neural Attentive Knowledge-based model (NAK) that learns the importance of each historical course in predicting the grade of a target course. Compared to CKRM and other competing approaches, our experiments on a large real-world dataset consisting of $\sim$1.5 grades show the effectiveness of the proposed NAK model in accurately predicting the students' grades. Moreover, the attention weights learned by the model can be helpful in better designing their degree plans.
computer science
This paper presents a brand new nonparametric density estimation strategy named the best-scored random forest density estimation whose effectiveness is supported by both solid theoretical analysis and significant experimental performance. The terminology best-scored stands for selecting one density tree with the best estimation performance out of a certain number of purely random density tree candidates and we then name the selected one the best-scored random density tree. In this manner, the ensemble of these selected trees that is the best-scored random density forest can achieve even better estimation results than simply integrating trees without selection. From the theoretical perspective, by decomposing the error term into two, we are able to carry out the following analysis: First of all, we establish the consistency of the best-scored random density trees under $L_1$-norm. Secondly, we provide the convergence rates of them under $L_1$-norm concerning with three different tail assumptions, respectively. Thirdly, the convergence rates under $L_{\infty}$-norm is presented. Last but not least, we also achieve the above convergence rates analysis for the best-scored random density forest. When conducting comparative experiments with other state-of-the-art density estimation approaches on both synthetic and real data sets, it turns out that our algorithm has not only significant advantages in terms of estimation accuracy over other methods, but also stronger resistance to the curse of dimensionality.
statistics
The temporal Talbot effect supports the generation of RF signals with high purity in optical domain. This allows for example a remote RF generation without the need for costly high-end electronic circuits and resource sharing. However, one of the most interesting features of the approach is its inherent ability to suppress phase noise during the carrier generation process. Besides the comb laser source, the dispersive element is the key component in this upconversion scheme. Ideally, it provides the right amount of dispersion over the whole spectral range of the comb. Broader combs are expected to allow a higher degree of phase noise suppression. The simulation of phase noise of an RF tone generated by an optical scheme is challenging in terms of computational and memory effort. Therefore, a simulation tool taking advantage of the properties of the comb and the Talbot effect had to be developed. In this paper, the dependency of the phase noise suppression on the dispersion characteristic is investigated yielding an important input to the design of these elements. First, the implemented simulation tool and the simulation parameters ensuring a correct estimation of the phase noise before and after dispersion influence are introduced. Then, the effect of different dispersion characteristics on the phase noise behavior is analyzed. It could be shown and explained how the phase noise at different offset frequencies to the carrier is affected differently depending on the dispersion characteristic and also the comb width. Design rules for the required comb width as well as the acceptable variation of the produced dispersion from the ideal value depending on the demands for the phase noise improvements could be developed and will be presented.
electrical engineering and systems science
In this paper we describe our efforts to make a bidirectional Congolese Swahili (SWC) to French (FRA) neural machine translation system with the motivation of improving humanitarian translation workflows. For training, we created a 25,302-sentence general domain parallel corpus and combined it with publicly available data. Experimenting with low-resource methodologies like cross-dialect transfer and semi-supervised learning, we recorded improvements of up to 2.4 and 3.5 BLEU points in the SWC-FRA and FRA-SWC directions, respectively. We performed human evaluations to assess the usability of our models in a COVID-domain chatbot that operates in the Democratic Republic of Congo (DRC). Direct assessment in the SWC-FRA direction demonstrated an average quality ranking of 6.3 out of 10 with 75% of the target strings conveying the main message of the source text. For the FRA-SWC direction, our preliminary tests on post-editing assessment showed its potential usefulness for machine-assisted translation. We make our models, datasets containing up to 1 million sentences, our development pipeline, and a translator web-app available for public use.
computer science
We study the effect of inhomogeneous environments on the swimming direction of the microalgae \textit{Chlamydomonas Reinhardtii} (CR) in the presence of a light stimulus. Positive or negative phototaxis describe the ability of microorganisms to bias their swimming towards or away from a light source. Here we consider microswimmers with negative phototaxis in a microfluidic device with a microfabricated square lattice of pillars as obstacles. We measured a mean deflection of microswimmers that shows an interesting nonlinear dependence on the direction of the guiding light beam with respect to the symmetry axes of the pillar lattice. By simulating a model swimmer in a pillar lattice and analyzing its scattering behavior, we identified the width of the reorientation distribution of swimmers to be also crucial for the nonlinear behavior of the swimmer deflection. On the basis of these results we suggest in addition an analytical model for microswimmers, where the pillar lattice is replaced by an anisotropic scattering medium, that depends only on a scattering rate and the width of the reorientation distribution of swimmers. This flexible and handy model fits the experimental results as well. The presented analysis of the deflection of light guided swimmers through pillar lattice may be used for separating swimmers having different reorientation distributions.
condensed matter
The formalism of relativistic partial wave expansion is developed for four-point celestial amplitudes of massless external particles. In particular, relativistic partial waves are found as eigenfunctions to the product representation of celestial Poincar\'e Casimir operators with appropriate eigenvalues. The requirement of hermiticity of Casimir operators is used to fix the corresponding integral inner product, and orthogonality of the obtained relativistic partial waves is verified explicitly. The completeness relation, as well as the relativistic partial wave expansion follow. Example celestial amplitudes of scalars, gluons, gravitons and open superstring gluons are expanded on the basis of relativistic partial waves for demonstration. A connection with the formulation of relativistic partial waves in the bulk of Minkowski space is made in appendices.
high energy physics theory
We prove the Lipman-Zariski conjecture for complex surface singularities with $p_g - g - b \le 2$. Here $p_g$ is the geometric genus, $g$ is the sum of the genera of the exceptional curves and $b$ is the first Betti number of the dual graph. This improves on a previous result of the second author. As an application, we show that a compact complex surface with locally free tangent sheaf is smooth as soon as it admits two generically linearly independent twisted vector fields and its canonical sheaf has at most two global sections.
mathematics
Cross-project defect prediction (CPDP) aims to predict defects of projects lacking training data by using prediction models trained on historical defect data from other projects. However, since the distribution differences between datasets from different projects, it is still a challenge to build high-quality CPDP models. Unfortunately, class imbalanced nature of software defect datasets further increases the difficulty. In this paper, we propose a transferlearning oriented minority over-sampling technique (TOMO) based feature weighting transfer naive Bayes (FWTNB) approach (TOMOFWTNB) for CPDP by considering both classimbalance and feature importance problems. Differing from traditional over-sampling techniques, TOMO not only can balance the data but reduce the distribution difference. And then FWTNB is used to further increase the similarity of two distributions. Experiments are performed on 11 public defect datasets. The experimental results show that (1) TOMO improves the average G-Measure by 23.7\%$\sim$41.8\%, and the average MCC by 54.2\%$\sim$77.8\%. (2) feature weighting (FW) strategy improves the average G-Measure by 11\%, and the average MCC by 29.2\%. (3) TOMOFWTNB improves the average G-Measure value by at least 27.8\%, and the average MCC value by at least 71.5\%, compared with existing state-of-theart CPDP approaches. It can be concluded that (1) TOMO is very effective for addressing class-imbalance problem in CPDP scenario; (2) our FW strategy is helpful for CPDP; (3) TOMOFWTNB outperforms previous state-of-the-art CPDP approaches.
computer science
Let $L$ be an $n$-component link ($n>1$) with pairwise nonzero linking numbers in a rational homology $3$-sphere $Y$. Assume the link complement $X:=Y\setminus\nu(L)$ has nondegenerate Thurston norm. In this paper, we study when a Thurston norm-minimizing surface $S$ properly embedded in $X$ remains norm-minimizing after Dehn filling all boundary components of $X$ according to $\partial S$ and capping off $\partial S$ by disks. In particular, for $n=2$ the capped-off surface is norm-minimizing when $[S]$ lies outside of a finite set of rays in $H_2(X,\partial X;\mathbb{R})$. When $Y$ is an integer homology sphere this gives an upper bound on the number of surgeries on $L$ which may yield $S^1\times S^2$. The main techniques come from Gabai's proof of the Property R conjecture and related work.
mathematics
We review the hypothetical interactions predicted beyond the Standard Model which could be constrained by using the results of tabletop laboratory experiments. These interactions are described by the power-type potentials with different powers, Yukawa potential, other spin-independent potentials, and by the spin-dependent potentials of different kinds. In all these cases the current constraints on respective hypothetical interactions are considered which follow from the Casimir effect and some other tabletop physics. The exotic particles and constraints on them are discussed in the context of problems of the quantum vacuum, dark energy, and the cosmological constant.
high energy physics phenomenology
We show that a polynomial H(N) of degree N of a harmonic oscillator hamiltonian allows us to devise a fully solvable continuous quantum system for which the first N discrete energy eigenvalues can be chosen at will. In general such a choice leads to a re-ordering of the associated energy eigenfunctions of H such that the number of their nodes does not increase monotonically with increasing level number. Systems H have certain universal features, we study their basic behaviours.
quantum physics
We consider a hybrid method to simulate the return time to the initial state in a critical-case birth--death process. The expected value of this return time is infinite, but its distribution asymptotically follows a power-law. Hence, the simulation approach is to directly simulate the process, unless the simulated time exceeds some threshold and if it does, draw the return time from the tail of the power law.
statistics
Studies of nucleated dwarf galaxies can constrain the scenarios for the formation and evolution of nuclear star clusters (NSC) in low-mass galaxies and give us insights on the origin of ultra compact dwarf galaxies (UCDs). We report the discovery of a NSC in the dwarf galaxy KKs58 and investigate its properties together with those of another NSC in KK197. Both NSCs are hosted by dwarf elliptical galaxies of the Centaurus group. Combining ESO VLT MUSE data with photometry from VLT FORS2, CTIO Blanco DECam, and HST ACS, as well as high-resolution spectroscopy from VLT UVES, we analyse the photometric, kinematic and stellar population properties of the NSCs and their host galaxies. We confirm membership of the NSCs based on their radial velocities and location close to the galaxy centres. We also confirm the membership of two globular clusters (GCs) and detect oblate rotation in the main body of KK197. Based on high signal-to-noise spectra taken with MUSE of the NSCs of both KKs58 and KK197 we measure low metallicities, [Fe/H] = $-1.75 \pm 0.06$ dex and [Fe/H] = $-1.84 \pm 0.05$ dex, and stellar masses of $7.3 \times 10^5 M_\odot$ and $1.0 \times 10^6 M_\odot$, respectively. Both NSCs are more metal-poor than their hosts that have metallicities of $-1.35 \pm 0.23$ dex (KKs58) and $-0.84 \pm 0.12$ dex (KK197). This can be interpreted as NSC formation via the in-spiral of GCs. The masses, sizes and metallicities of the two NSCs place them among other NSCs, but also among the known UCDs of the Centaurus group. This indicates that NSCs might constitute the progenitors of a part of the low-mass UCDs, although their properties are almost indistinguishable from typical GCs.
astrophysics
Recently, a QCD sum-rule analysis of tetraquark molecular states has been published, having the objective of demonstrating that our previously formulated tetraquark-adequate QCD sum rules are not correct. This comment brings to the attention of the reader inconsistencies in that article and explains some subtle details of QCD sum rules for exotic states.
high energy physics phenomenology
Simultaneous concurrence of extreme values across multiple climate variables can result in large societal and environmental impacts. Therefore, there is growing interest in understanding these concurrent extremes. In many applications, not only the frequency but also the magnitude of concurrent extremes are of interest. One way to approach this problem is to study the distribution of one climate variable given that another is extreme. In this work we develop a statistical framework for estimating bivariate concurrent extremes via a conditional approach, where univariate extreme value modeling is combined with dependence modeling of the conditional tail distribution using techniques from quantile regression and extreme value analysis to quantify concurrent extremes. We focus on the distribution of daily wind speed conditioned on daily precipitation taking its seasonal maximum. The Canadian Regional Climate Model large ensemble is used to assess the performance of the proposed framework both via a simulation study with specified dependence structure and via an analysis of the climate model-simulated dependence structure.
statistics
Organic-inorganic metal halide perovskites have recently attracted increasing attention as highly efficient light harvesting materials for photovoltaic applications. However, the precise control of crystallization and morphology of organometallic perovskites deposited from solution, considered crucial for enhancing the final photovoltaic performance, remains challenging. In this context, here, we report on growing microcrystalline deposits of CH3NH3PbI3 (MAPbI3), by one-step solution casting on cylinde-shaped quartz substrates (rods). We show that the substrate curvature has a strong influence on morphology of the obtained polycrystalline deposits of MAPbI3. Although the crystalline width and length markedly decreased for substrates with higher curvatures, the photoluminescence (PL) spectral peak positions did not significantly evolve for MAPbI3 deposits on substrates with different diameters. The crystalline size reduction and denser coverage of microcrystalline MAPbI3 deposits on cylinder-shaped substrates with higher curvatures were attributed to two major contributions, both related to the annealing step of the MAPbI3 deposits. In particular, the diameter-dependent variability of the heat capacities and the substrate curvature-enhanced solvent evaporation rate seemed to contribute the most to the crystallization process and the resulting morphology changes of MAPbI3 deposits on cylinder-shaped quartz substrates with various diameters. The longitudinal geometry of cylinder-shaped substrates provided also a facile solution for checking the PL response of the deposits of MAPbI3 exposed to the flow of various gaseous media, such as oxygen, nitrogen and argon. Overall, the approach reported herein inspires novel, cylinder-shaped geometries of MAPbI3 deposits, which can find applications in low-cost photo-optical devices, including gas sensors.
condensed matter
Integration testing is a very important step in software testing. Existing methods evaluate the stubbing cost for class integration test orders by considering only the interclass direct relationships such as inheritance, aggregation, and association, but they omit the interclass indirect relationship caused by control coupling, which can also affect the test orders and the stubbing cost. In this paper, we introduce an integration test order strategy to consider control coupling. We advance the concept of transitive relationship to describe this kind of interclass dependency and propose a new measurement method to estimate the complexity of control coupling, which is the complexity of stubs created for a transitive relationship. We evaluate our integration test order strategy on 10 programs on various scales. The results show that considering the transitive relationship when generating class integration test orders can significantly reduce the stubbing cost for most programs and that our integration test order strategy obtains satisfactory results more quickly than other methods.
computer science
Recent studies of gamma-ray, cosmic-ray and radio data put stringent constraints on the fraction of primordial black holes (PBHs) in our universe. In this article, we propose a new indirect method in using the X-ray luminosity data of cool-core clusters to constrain the evaporating PBH fraction for the monochromatic, log-normal and power-law mass distributions. The present results show that the amount of evaporating PBHs only constitutes a minor component of dark matter for a large parameter space. The constraints are consistent with and close to that obtained from other cosmic-ray and multi-wavelength observations.
astrophysics
We propose a novel calibration method for computer simulators, dealing with the problem of covariate shift. Covariate shift is the situation where input distributions for training and test are different, and ubiquitous in applications of simulations. Our approach is based on Bayesian inference with kernel mean embedding of distributions, and on the use of an importance-weighted reproducing kernel for covariate shift adaptation. We provide a theoretical analysis for the proposed method, including a novel theoretical result for conditional mean embedding, as well as empirical investigations suggesting its effectiveness in practice. The experiments include calibration of a widely used simulator for industrial manufacturing processes, where we also demonstrate how the proposed method may be useful for sensitivity analysis of model parameters.
statistics
Discovering interaction effects on a response of interest is a fundamental problem faced in biology, medicine, economics, and many other scientific disciplines. In theory, Bayesian methods for discovering pairwise interactions enjoy many benefits such as coherent uncertainty quantification, the ability to incorporate background knowledge, and desirable shrinkage properties. In practice, however, Bayesian methods are often computationally intractable for even moderate-dimensional problems. Our key insight is that many hierarchical models of practical interest admit a particular Gaussian process (GP) representation; the GP allows us to capture the posterior with a vector of O(p) kernel hyper-parameters rather than O(p^2) interactions and main effects. With the implicit representation, we can run Markov chain Monte Carlo (MCMC) over model hyper-parameters in time and memory linear in p per iteration. We focus on sparsity-inducing models and show on datasets with a variety of covariate behaviors that our method: (1) reduces runtime by orders of magnitude over naive applications of MCMC, (2) provides lower Type I and Type II error relative to state-of-the-art LASSO-based approaches, and (3) offers improved computational scaling in high dimensions relative to existing Bayesian and LASSO-based approaches.
statistics
Using the technique of enrichment of contractive type mappings by Krasnoselskij averaging, introduced in [Berinde, V., {\it Approximating fixed points of enriched nonexpansive mappings by Krasnoselskij iteration in Hilbert spaces}, Carpathian J. Math. {\bf 35} (2019), no. 3, 277-288.], we introduce the class of enriched Chatterjea contractions and prove general fixed point theorems for such contractions in the setting of a Banach space. Examples to illustrate the richness of the new class of contractions and the relationship between enriched Banach contractions, enriched Kannan contractions and enriched Kannan contractions are also given.
mathematics
Beilinson-Bernstein localization realizes representations of complex reductive Lie algebras as monodromic $D$-modules on the "basic affine space" $G/N$, a torus bundle over the flag variety. A doubled version of the same space appears as the horocycle space describing the geometry of the reductive group $G$ at infinity, near the closed stratum of the wonderful compactification $\overline{G}$, or equivalently in the special fiber of the Vinberg semigroup of $G$. We show that Beilinson-Bernstein localization for $U\mathfrak g$-bimodules arises naturally as the specialization at infinity in $\overline{G}$ of the $D$-modules on $G$ describing matrix coefficients of Lie algebra representations. More generally, the asymptotics of matrix coefficient $D$-modules along any stratum of $\overline{G}$ are given by the matrix coefficient $D$-modules for parabolic restrictions. This provides a simple algebraic derivation of the relation between growth of matrix coefficients of admissible representations and $\mathfrak n$-homology. The result is an elementary consequence of the compatibility of localization with the degeneration of affine $G$-varieties to their asymptotic cones; analogous results hold for the asymptotics of the equations describing spherical functions on symmetric spaces.
mathematics
Let $V$ be any vector space of multivariate degree-$d$ homogeneous polynomials with co-dimension at most $k$, and $S$ be the set of points where all polynomials in $V$ {\em nearly} vanish. We establish a qualitatively optimal upper bound on the size of $\epsilon$-covers for $S$, in the $\ell_2$-norm. Roughly speaking, we show that there exists an $\epsilon$-cover for $S$ of cardinality $M = (k/\epsilon)^{O_d(k^{1/d})}$. Our result is constructive yielding an algorithm to compute such an $\epsilon$-cover that runs in time $\mathrm{poly}(M)$. Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models with hidden variables. These include density and parameter estimation for $k$-mixtures of spherical Gaussians (with known common covariance), PAC learning one-hidden-layer ReLU networks with $k$ hidden units (under the Gaussian distribution), density and parameter estimation for $k$-mixtures of linear regressions (with Gaussian covariates), and parameter estimation for $k$-mixtures of hyperplanes. Our algorithms run in time {\em quasi-polynomial} in the parameter $k$. Previous algorithms for these problems had running times exponential in $k^{\Omega(1)}$. At a high-level our algorithms for all these learning problems work as follows: By computing the low-degree moments of the hidden parameters, we are able to find a vector space of polynomials that nearly vanish on the unknown parameters. Our structural result allows us to compute a quasi-polynomial sized cover for the set of hidden parameters, which we exploit in our learning algorithms.
computer science
In this paper, we present a model in which an up-type vector-like quark (VLQ) is charged under a new $U(1)_d$ gauge force which kinetically mixes with the SM hypercharge. The gauge boson of the $U(1)_d$ is the dark photon, $\gamma_d$. Traditional searches for VLQs rely on decays into Standard Model electroweak bosons $W,Z$ or Higgs. However, since no evidence for VLQs has been found at the Large Hadron Collider (LHC), it is imperative to search for other novel signatures of VLQs beyond their traditional decays. As we will show, if the dark photon is much less massive than the Standard Model electroweak sector, $M_{\gamma_d}\ll M_Z$, for the large majority of the allowed parameter space the VLQ predominately decays into the dark photon and the dark Higgs that breaks the $U(1)_d$ . That is, this VLQ is a `maverick top partner' with nontraditional decays. One of the appeals of this scenario is that pair production of the VLQ at the LHC occurs through the strong force and the rate is determined by the gauge structure. Hence, the production of the dark photon at the LHC only depends on the strong force and is largely independent of the small kinetic mixing with hypercharge. This scenario provides a robust framework to search for a light dark sector via searches for heavy colored particles at the LHC.
high energy physics phenomenology
Upon treating the whole closed-string massless NS-NS sector as stringy graviton fields, Double Field Theory may evolve into `Stringy Gravity'. In terms of an $\mathbf{O}(D,D)$ covariant differential geometry beyond Riemann, we present the definitions of the off-shell conserved stringy Einstein curvature tensor and the on-shell conserved stringy Energy-Momentum tensor. Equating them, all the equations of motion of the massless sector are unified into a single expression, $G_{AB}{=8\pi G} T_{AB}$, carrying $\mathbf{O}(D,D)$ vector indices, which we dub `the Einstein Double Field Equations'.
high energy physics theory
In this paper, by using the recently compiled set of 120 intermediate-luminosity quasars (ILQSO) observed in a single-frequency VLBI survey, we propose an improved model-independent method to probe cosmic curvature parameter $\Omega_k$ and make the first measurement of the cosmic curvature referring to a distant past, with redshifts up to $z\sim 3.0$. Compared with other methods, the proposed one involving the quasar data achieves constraints with higher precision in this redshift range. More importantly, our results indicate that the measured $\Omega_k$ is in good agreement with zero cosmic curvature, implying that there is no significant deviation from a flat Universe. Finally, we investigate the possibility of testing $\Omega_k$ with a much higher accuracy using quasars observed in the future VLBI surveys. It is shown that our method could provide a reliable and tight constraint on the prior $\Omega_k$ and one can expect the zero cosmic curvature to be established at the precision of $\Delta\Omega_k\sim 10^{-2}$ with 250 well-observed radio quasars.
astrophysics
We prove that the Lawson surface $\xi_{g,1}$ in Lawson's original notation, which has genus $g$ and can be viewed as a desingularization of two orthogonal great two-spheres in the round three-sphere ${\mathbb{S}}^3$, has index $2g+3$ and nullity $6$ for any genus $g\ge2$. In particular $\xi_{g,1}$ has no exceptional Jacobi fields, which means that it cannot `flap its wings' at the linearized level and is $C^1$-isolated.
mathematics
Entanglement is a central subject in quantum information theory. Due to its genuine relativistic behavior and fundamental nature, high-energy colliders are attractive systems for the experimental study of quantum information. We propose the detection of entanglement between the spins of top-antitop quark pairs at the LHC, representing the first proposal of entanglement detection in a pair of quarks, and also the entanglement observation at the highest energy scale so far. We show that entanglement can be observed by direct measurement of the angular separation between the leptons arising from the decay of the top-antitop pair. The detection can be achieved with high statistical significance, using the current data recorded at the LHC. In addition, we develop a simple protocol for the quantum tomography of the top-antitop pair, providing a new experimental tool to test theoretical predictions for its quantum state. Our work explicitly implements canonical experimental techniques in quantum information in a two-qubit high-energy system, paving the way to use high-energy colliders to also study quantum information theory.
quantum physics
Because of the efficiency of modeling fuzziness and vagueness, Z-number plays an important role in real practice. However, Z-numbers, defined in the real number field, lack the ability to process the quantum information in quantum environment. It is reasonable to generalize Z-number into its quantum counterpart. In this paper, we propose quantum Z-numbers (QZNs), which are the quantum generalization of Z-numbers. In addition, seven basic quantum fuzzy operations of QZNs and their corresponding quantum circuits are presented and illustrated by numerical examples. Moreover, based on QZNs, a novel quantum multi-attributes decision making (MADM) algorithm is proposed and applied in medical diagnosis. The results show that, with the help of quantum computation, the proposed algorithm can make diagnoses correctly and efficiently.
quantum physics
Optical coherence tomography offers astounding opportunities to image the complex structure of living tissue, but lacks functional information. We present dynamic full-field optical coherence tomography to image living human induced pluripotent stem cell-derived retinal organoids non-invasively. Colored images with an endogenous contrast linked to organelle motility are generated, with sub-micrometer spatial resolution and millisecond temporal resolution, opening an avenue to identify specific cell types in living tissue via their function.
physics
We argue that the static non-linear Hall conductivity can always be represented as a vector in two-dimensions and as a pseudo-tensor in three-dimensions independent of its microscopic origin. In a single band model with a constant relaxation rate this vector or tensor is proportional to the Berry curvature dipole \cite{Sodemann_2015}. Here, we develop a quantum Boltzmann formalism to second order in electric fields. We find that in addition to the Berry Curvature Dipole term, there exist additional disorder mediated corrections to the non-linear Hall tensor that have the same scaling in impurity scattering rate. These can be thought of as the non-linear counterparts to the side-jump and skew-scattering corrections to the Hall conductivity in the linear regime. We illustrate our formalism by computing the different contributions to the non-linear Hall conductivity of two-dimensional tilted Dirac fermions.
condensed matter
It is expensive and time-consuming to collect sufficient labeled data for human activity recognition (HAR). Domain adaptation is a promising approach for cross-domain activity recognition. Existing methods mainly focus on adapting cross-domain representations via domain-level, class-level, or sample-level distribution matching. However, they might fail to capture the fine-grained locality information in activity data. The domain- and class-level matching are too coarse that may result in under-adaptation, while sample-level matching may be affected by the noise seriously and eventually cause over-adaptation. In this paper, we propose substructure-level matching for domain adaptation (SSDA) to better utilize the locality information of activity data for accurate and efficient knowledge transfer. Based on SSDA, we propose an optimal transport-based implementation, Substructural Optimal Transport (SOT), for cross-domain HAR. We obtain the substructures of activities via clustering methods and seeks the coupling of the weighted substructures between different domains. We conduct comprehensive experiments on four public activity recognition datasets (i.e. UCI-DSADS, UCI-HAR, USC-HAD, PAMAP2), which demonstrates that SOT significantly outperforms other state-of-the-art methods w.r.t classification accuracy (9%+ improvement). In addition, our mehtod is 5x faster than traditional OT-based DA methods with the same hyper-parameters.
electrical engineering and systems science
The performance of low-density parity-check (LDPC) codes at high signal-to-noise ratios (SNRs) is known to be limited by the presence of certain sub-graphs that exist in the Tanner graph representation of the code, for example trapping sets and absorbing sets. This paper derives a lower bound on the frame error rate (FER) of any LDPC code containing a given problematic sub-graph, assuming a particular message passing decoder and decoder quantization. A crucial aspect of the lower bound is that it is code-independent, in the sense that it can be derived based only on a problematic sub-graph and then applied to any code containing it. Due to the complexity of evaluating the exact bound, assumptions are proposed to approximate it, from which we can estimate decoder performance. Simulated results obtained for both the quantized sum-product algorithm (SPA) and the quantized min-sum algorithm (MSA) are shown to be consistent with the approximate bound and the corresponding performance estimates. Different classes of LDPC codes, including both structured and randomly constructed codes, are used to demonstrate the robustness of the approach.
computer science
Magnetic fluid hyperthermia (MFH), the procedure of raising the temperature of tumor cells using magnetic nanoparticles (MNPs) as heating agents, has proven successful in treating some types of cancer. However, the low heating power generated under physiological conditions makes necessary a high local concentration of MNPs at tumor sites. Here, we report how the in vitro heating power of magnetically soft MnFe$_2$O$_4$ nanoparticles can be enhanced by intracellular low-dimensional clusters through a strategy that includes: a) the design of the MNPs to retain N\'eel magnetic relaxation in high viscosity media, and b) culturing MNP-loaded cells under magnetic fields to produce elongated intracellular agglomerates. Our direct in vitro measurements demonstrated that the specific loss power (SLP) of elongated agglomerates ($SLP=576\pm33$ W/g) induced by culturing BV2 cells in situ under a dc magnetic field was increased by a factor of 2 compared to the $SLP=305\pm25$ W/g measured in aggregates freely formed within cells. A numerical mean-field model that included dipolar interactions quantitatively reproduced the SLPs of these clusters both in phantoms and in vitro, suggesting that it captures the relevant mechanisms behind power losses under high-viscosity conditions. These results indicate that in situ assembling of MNPs into low-dimensional structures is a sound possible way to improve the heating performance in MFH.
condensed matter
Context. Stellar internal magnetic fields have recently been shown to leave a detectable signature on period spacing patterns of gravity modes. Aims. We investigate the effect of the obliquity of a mixed (poloidal and toroidal) dipolar internal fossil magnetic field with respect to the rotation axis on the frequency of gravity modes in rapidly rotating stars. Methods. We use the traditional approximation of rotation to compute non-magnetic modes, and a perturbative treatment of the magnetic field to compute the corresponding frequency shifts. We apply the new formalism to HD 43317, a magnetic, rapidly rotating, slowly pulsating B-type star, whose field has an obliquity angle of about 80{\deg}. Results. We find that frequency shifts induced by the magnetic field on high-radial-order gravity modes are larger with increasing obliquity angle, when the magnetic axis is closer to the equatorial region, where these modes are trapped. The maximum value is reached for an obliquity angle of 90{\deg}. This trend is observed for all mode geometries. Conclusions. Our results predict that the signature of an internal oblique dipolar magnetic field is detectable using asteroseismology of gravity modes.
astrophysics
Given two arbitrary sequences of denoisers for block lengths tending to infinity we ask if it is possible to construct a third sequence of denoisers with an asymptotically vanishing (in block length) excess expected loss relative to the best expected loss of the two given denoisers for all clean channel input sequences. As in the setting of DUDE [1], which solves this problem when the given denoisers are sliding block denoisers, the construction is allowed to depend on the two given denoisers and the channel transition probabilities. We show that under certain restrictions on the two given denoisers the problem can be solved using a straightforward application of a known loss estimation paradigm. We then show by way of a counter-example that the loss estimation approach fails in the general case. Finally, we show that for the binary symmetric channel, combining the loss estimation with a randomization step leads to a solution to the stated problem under no restrictions on the given denoisers.
computer science
Thin molecular films under model conditions are often exploited as benchmarks and case studies to investigate the electronic and structural changes occurring on the surface of metallic electrodes. Here we show that the modification of a metallic surface induced by oxygen adsorption allows the preservation of the geometry of a molecular adlayer, giving access to the determination of molecular orbital symmetries by means of near-edge x-ray absorption fine structure spectroscopy, NEXAFS. As a prototypical example, we exploited Nickel Tetraphenyl Porphyrin molecules deposited on a bare and on an oxygen pre-covered Cu(100) surface. We find that adsorbed atomic oxygen quenches the charge transfer at the metal-organic interface but, in contrast to a thin film sample, maintains the ordered adsorption geometry of the organic molecules. In this way, it is possible to disentangle {\pi}* and {\sigma}* symmetry orbitals, hence estimating the relative oscillator strength of core level transitions directly from the experimental data, as well as to evaluate and localize the degree of charge transfer in a coupled system. In particular, we neatly single out the {\sigma}* contribution associated with the N 1s transition to the mixed N 2px,y-Ni 3dx2-y2 orbital, which falls close to the leading {\pi}*-symmetry LUMO resonance.
condensed matter
Young massive clusters play an important role in the evolution of their host galaxies, and feedback from the high-mass stars in these clusters can have profound effects on the surrounding interstellar medium. The nuclear starburst in the nearby galaxy NGC253 at a distance of 3.5 Mpc is a key laboratory in which to study star formation in an extreme environment. Previous high resolution (1.9 pc) dust continuum observations from ALMA discovered 14 compact, massive super star clusters (SSCs) still in formation. We present here ALMA data at 350 GHz with 28 milliarcsecond (0.5 pc) resolution. We detect blueshifted absorption and redshifted emission (P-Cygni profiles) towards three of these SSCs in multiple lines, including CS 7$-$6 and H$^{13}$CN 4$-$3, which represents direct evidence for previously unobserved outflows. The mass contained in these outflows is a significant fraction of the cluster gas masses, which suggests we are witnessing a short but important phase. Further evidence of this is the finding of a molecular shell around the only SSC visible at near-IR wavelengths. We model the P-Cygni line profiles to constrain the outflow geometry, finding that the outflows must be nearly spherical. Through a comparison of the outflow properties with predictions from simulations, we find that none of the available mechanisms completely explains the observations, although dust-reprocessed radiation pressure and O star stellar winds are the most likely candidates. The observed outflows will have a very substantial effect on the clusters' evolution and star formation efficiency.
astrophysics
We here consider the subset simulation method which approaches a failure event using a decreasing sequence of nested intermediate failure events. The method resembles importance sampling, which actively explores a probability space by conditioning the next evaluation on the previous evaluations using a Markov chain Monte Carlo (MCMC) algorithm. A Markov chain typically requires many steps to estimate the target distribution, which is impractical with expensive numerical models. Therefore, we propose to approximate each step of a Markov chain locally with Gaussian process (GP) regression. Benchmark examples of reliability analysis show that local approximations significantly improve overall efficiency of subset simulation. They reduce the number of expensive limit-state evaluations by over $80\%$. However, GP regression becomes computationally impractical with increasing dimension. Therefore, to make our use of a GP feasible, we employ the partial least squares (PLS) regression, a gradient-free reduction method, locally to explore and utilize a low-dimensional subspace within a Markov chain. Numerical experiments illustrate a significant computational gain with maintained sufficient accuracy.
statistics
A novel approach for the state-specific enantiomeric enrichment and the spatial separation of enantiomers is presented. Our scheme utilizes techniques from strong-field laser physics, specifically an optical centrifuge in conjunction with a static electric field, to create a chiral field with defined handedness. Molecular enantiomers experience unique rotational excitation dynamics and this can be exploited to spatially separate the enantiomers using electrostatic deflection. Notably, the rotational-state-specific enantiomeric enhancement and its handedness is fully controllable. To explain these effects, we introduce the conceptual framework of $field\text{-}induced~diastereomers$ of a chiral molecule and perform robust quantum mechanical simulations on the prototypical chiral molecule propylene oxide (C$_3$H$_6$O), for which ensembles with an enantiomeric excess of up to $30~\%$ were obtained.
physics
The radio:X-ray correlation that characterises accreting black holes at all mass scales - from stellar mass black holes in binary systems to super-massive black holes powering Active Galactic Nuclei - is one of the most important pieces of observational evidence supporting the existence of a connection between the accretion process and the generation of collimated outflows - or jets - in accreting systems. Although recent studies suggest that the correlation extends down to low luminosities, only a handful of stellar mass black holes have been clearly detected, and in general only upper limits (especially at radio wavelengths) can be obtained during quiescence. We recently obtained detections of the black hole X-ray binary GX 339--4 in quiescence using the MeerKAT radio telescope and Swift X-ray Telescope instrument onboard the Neil Gehrels Swift Observatory, probing the lower end of the radio:X-ray correlation. We present the properties of accretion and of the connected generation of jets in the poorly studied low-accretion rate regime for this canonical black hole XRB system.
astrophysics
High levels of X-ray and UV activity on young M dwarfs may drive rapid atmospheric escape on temperate, terrestrial planets orbiting within the liquid water habitable zone. However, secondary atmospheres on planets orbiting older, less active M dwarfs may be stable and present more promising candidates for biomarker searches. We present new HST and Chandra observations of Barnard's Star (GJ 699), a 10 Gyr old M3.5 dwarf, acquired as part of the Mega-MUSCLES program. Despite the old age and long rotation period of Barnard's star, we observe two FUV ($\delta_{130}$ $\approx$ 5000s; $E_{130}$ $\approx$ 10$^{29.5}$ erg each) and one X-ray ($E_{X}$ $\approx$ 10$^{29.2}$ erg) flares, and estimate a high-energy flare duty cycle (defined here as the fraction of the time the star is in a flare state) of $\sim$ 25\%. A 5 A - 10 $\mu$m SED of GJ 699 is created and used to evaluate the atmospheric stability of a hypothetical, unmagnetized terrestrial planet in the habitable zone ($r_{HZ}$ $\sim$ 0.1 AU). Both thermal and non-thermal escape modeling indicate (1) the $quiescent$ stellar XUV flux does not lead to strong atmospheric escape: atmospheric heating rates are comparable to periods of high solar activity on modern Earth, and (2) the $flare$ environment could drive the atmosphere into a hydrodynamic loss regime at the observed flare duty cycle: sustained exposure to the flare environment of GJ 699 results in the loss of $\approx$ 87 Earth atmospheres Gyr$^{-1}$ through thermal processes and $\approx$ 3 Earth atmospheres Gyr$^{-1}$ through ion loss processes, respectively. These results suggest that if rocky planet atmospheres can survive the initial $\sim$ 5 Gyr of high stellar activity, or if a second generation atmosphere can be formed or acquired, the flare duty cycle may be the controlling stellar parameter for the stability of Earth-like atmospheres around old M stars.
astrophysics
In this paper we formulate and solve a robust least squares problem for a system of linear equations subject to quantization error in the data matrix. Ordinary least squares fails to consider uncertainty in the operator, modeling all noise in the observed signal. Total least squares accounts for uncertainty in the data matrix, but necessarily increases the condition number of the operator compared to ordinary least squares. Tikhonov regularization or ridge regression is frequently employed to combat ill-conditioning, but requires parameter tuning which presents a host of challenges and places strong assumptions on parameter prior distributions. The proposed method also requires selection of a parameter, but it can be chosen in a natural way, e.g., a matrix rounded to the 4th digit uses an uncertainty bounding parameter of 0.5e-4. We show here that our robust method is theoretically appropriate, tractable, and performs favorably against ordinary and total least squares.
mathematics
Computer-aided diagnosis has become a necessity for accurate and immediate coronavirus disease 2019 (COVID-19) detection to aid treatment and prevent the spread of the virus. Numerous studies have proposed to use Deep Learning techniques for COVID-19 diagnosis. However, they have used very limited chest X-ray (CXR) image repositories for evaluation with a small number, a few hundreds, of COVID-19 samples. Moreover, these methods can neither localize nor grade the severity of COVID-19 infection. For this purpose, recent studies proposed to explore the activation maps of deep networks. However, they remain inaccurate for localizing the actual infestation making them unreliable for clinical use. This study proposes a novel method for the joint localization, severity grading, and detection of COVID-19 from CXR images by generating the so-called infection maps. To accomplish this, we have compiled the largest dataset with 119,316 CXR images including 2951 COVID-19 samples, where the annotation of the ground-truth segmentation masks is performed on CXRs by a novel collaborative human-machine approach. Furthermore, we publicly release the first CXR dataset with the ground-truth segmentation masks of the COVID-19 infected regions. A detailed set of experiments show that state-of-the-art segmentation networks can learn to localize COVID-19 infection with an F1-score of 83.20%, which is significantly superior to the activation maps created by the previous methods. Finally, the proposed approach achieved a COVID-19 detection performance with 94.96% sensitivity and 99.88% specificity.
electrical engineering and systems science
The concept of intelligent transportation systems (ITS) is considered to be a highly promising area of research due to its diversity of unique features. It is based mainly on the wireless vehicular network (WVN), where vehicles can perform sophisticated services such as sharing real-time safety information. To ensure high-quality service, WVN needs to solve the security challenges like eavesdropping, where malicious entities try to intercept the confidential transmitted signal. In this paper, we are going to provide a security scheme under the Double kappa-mu Shadowed fading. Our solution is based on the use of a friendly jammer that will transmit an artificial noise (AN) to jam the attacker's link and decrease its eavesdropping performances. To evaluate the efficiency of our solution, we investigated the outage probability for two special cases: Nakagami-m and Rician shadowed while taking into consideration the density of the blockage and the shadowing effects. We also studied the average secrecy capacity via deriving closed-form expressions of the ergodic capacity at the legitimate receiver and the attacker for the special case: Nakagami-m fading distribution.
electrical engineering and systems science
Diamond, a well-known wide-bandgap insulator, becomes a low-temperature superconductor upon substitutional doping of carbon with boron. However, limited boron solubility and significant lattice disorder introduced by boron doping prevent attaining the theoretically-predicted high-temperature superconductivity. Here we present an alternative co-doping approach, based on the combination of ionic gating and boron substitution, in hydrogenated thin films epitaxially grown on (111)- and (110)-oriented single crystals. Gate-dependent electric transport measurements show that the effect of boron doping strongly depends on the crystal orientation. In the (111) surface, it strongly suppresses the charge-carrier mobility and moderately increases the gate-induced doping, while in the (110) surface it strongly increases the gate-induced doping with a moderate reduction in mobility. In both cases the maximum total carrier density remains below $2{\cdot}10^{14}\,$cm$^{-2}$, three times lower than the value theoretically required for high-temperature superconductivity. Density-functional theory calculations show that this strongly orientation-dependent effect is due to the specific energy-dependence of the density of states in the two surfaces. Our results allow to determine the band filling and doping-dependence of the hole scattering lifetime in the two surfaces, showing the occurrence of a frustrated insulator-to-metal transition in the (110) surface and of a re-entrant insulator-to-metal transition in the (111) surface.
condensed matter
The knowledge and thus characterization of the temporal modes of quantum light fields is important in many areas of quantum physics ranging from experimental setup diagnosis to fundamental-physics investigations. Recent results showed how the auto-correlation function computed from continuous-wave homodyne measurements can be a powerful way to access the temporal mode structure. Here, we push forward this method by providing a deeper understanding and by showing how to extract the amplitude and phase of the temporal mode function with reduced experimental resources. Moreover, a quantitative analysis allows us to identify a regime of parameters where the method provides a trustworthy reconstruction, which we illustrate experimentally.
quantum physics
Using dispersive representations of the nucleon gravitational form factors, the results for their absorptive parts from chiral effective field theory in curved space-time, and the mechanical stability conditions, we obtain a model independent inequality for the value of the gravitational $D(t)$ form factor at zero momentum transfer (Druck-term). In particular, the obtained inequality leads to a conservative bound on the Druck-term in the chiral limit $D \leq -0.95(9)$. This bound implies the restriction on the low-energy constant $c_8$ of the effective chiral action for nucleons and pions in the presence of an external gravitational field, $c_8\leq -1.1(1)$ GeV$^{-1}$. For the physical pion mass we obtain a model independent bound $D\leq -0.20(2)$.
high energy physics phenomenology
We discuss exterior and classical interior alternatives for evaluating fluid flow induced forces on bodies. The discussion aims at a reduction of the total shape derivative, achieved through a decoupling of control and objective in the exterior approach. In this case, geometric as well as convective contributions to the shape derivative vanish. Convective contributions depend on primal physics and may disappear, which is not the case for geometric components. The latter can be interpreted as curvatures immanent to industrial applications. The remaining local derivative of the objective functional can be determined efficiently with an adjoint system, that differs to the classical approach in its boundary conditions only and resembles an ALE strategy. A two-dimensional flow exposed to gravity illustrates the features of the exterior approach, whereby carefully derived derivatives from a second order Finite-Difference study were used to validate the results.
physics