text
stringlengths
11
9.77k
label
stringlengths
2
104
In this paper, we study the problem of precision matrix estimation when the dataset contains sensitive information. In the differential privacy framework, we develop a differentially private ridge estimator by perturbing the sample covariance matrix. Then we develop a differentially private graphical lasso estimator by using the alternating direction method of multipliers (ADMM) algorithm. The theoretical results and empirical results that show the utility of the proposed methods are also provided.
statistics
The Orion Nebula Cluster (ONC) is the nearest site of ongoing massive star formation, which allows us to study the kinematics and dynamics of the region in detail and constrain star formation theories. Using HST ACS/WFPC2/WFC3IR and Keck II NIRC2 data, we have measured the proper motions of 701 stars within an $\sim6'\times6'$ field of view around the center of the ONC. We have found more than 10 escaping star candidates, concentrated predominantly at the core of the cluster. The proper motions of the bound stars are consistent with a normal distribution, albeit elongated North-South along the Orion filament, with proper motion dispersions of $(\sigma_{\mu,\alpha^*}, \sigma_{\mu,\delta}) = (0.83\pm0.02,\,1.12\pm0.03)$ mas yr$^{-1}$ or intrinsic velocity dispersions of $(\sigma_{v,\alpha^*}, \sigma_{v,\delta}) = (1.57\pm0.04,\,2.12\pm0.06)$ km s$^{-1}$ assuming a distance of 400 pc to the ONC. The cluster shows no evidence for tangential-to-radial anisotropy. Our velocity dispersion profile agrees with the prediction from the observed stellar + gas density profile from Da Rio et al. (2014), indicating that the ONC is in virial equilibrium. This finding suggests that the cluster was formed with a low star formation efficiency per dynamical timescale based on comparisons with current star formation theories. Our survey also recovered high-velocity IR sources BN, x, and n in the BN/KL region. The estimated location of the first two sources $\sim500$ years ago agrees with that of the radio source I, consistent with their proposed common origin from a multi-stellar disintegration. However, source n appears to have a small proper motion and is unlikely to have been involved in the event.
astrophysics
Catastrophic loss data are known to be heavy-tailed. Practitioners then need models that are able to capture both tail and modal parts of claim data. To this purpose, a new parametric family of loss distributions is proposed as a gamma mixture of the generalized log-Moyal distribution from Bhati and Ravi (2018), termed the generalized log-Moyal gamma distribution (GLMGA). We discuss the probabilistic characteristics of the GLMGA, and statistical estimation of the parameters through maximum likelihood. While the GLMGA distribution is a special case of the GB2 distribution, we show that this simpler model is effective in regression modelling of large and modal loss data. A fire claim data set reported in Cummins et al. (1990) and a Chinese earthquake loss data set are used to illustrate the applicability of the proposed model.
statistics
In advanced mission concepts with high levels of autonomy, spacecraft need to internally model the pose and shape of nearby orbiting objects. Recent works in neural scene representations show promising results for inferring generic three-dimensional scenes from optical images. Neural Radiance Fields (NeRF) have shown success in rendering highly specular surfaces using a large number of images and their pose. More recently, Generative Radiance Fields (GRAF) achieved full volumetric reconstruction of a scene from unposed images only, thanks to the use of an adversarial framework to train a NeRF. In this paper, we compare and evaluate the potential of NeRF and GRAF to render novel views and extract the 3D shape of two different spacecraft, the Soil Moisture and Ocean Salinity satellite of ESA's Living Planet Programme and a generic cube sat. Considering the best performances of both models, we observe that NeRF has the ability to render more accurate images regarding the material specularity of the spacecraft and its pose. For its part, GRAF generates precise novel views with accurate details even when parts of the satellites are shadowed while having the significant advantage of not needing any information about the relative pose.
computer science
A rice disease diagnosis LINE Bot System from paddy field images was presented in this paper. An easy-to-use automatic rice disease diagnosis system was necessary to help rice farmers improve yield and quality. We targeted on the images took from the paddy environment without special sample preparation. We used a deep learning neural networks technique to detect rice disease in the images. We purposed object detection model training and refinement process to improve the performance of our previous rice leaf diseases detection research. The process was based on analyzing the model's predictive results and could be repeatedly used to improve the quality of the database in the next training of the model. The deployment model for our LINE Bot system was created from the selected best performance technique in our previous paper, YOLOv3, trained by refined training data set. The performance of deployment model was measured on 5 target classes by average mAP improved from 82.74% in previous paper to 89.10%. We purposed Rice Disease LINE Bot system used this deployment model. Our system worked automatically real-time to suggest primary rice disease diagnosis results to the users in the LINE group. Our group included of rice farmers and rice disease experts, and they could communicate freely via chat. In the real LINE Bot deployment, the model's performance measured by our own defined measurement Average True Positive Point was 78.86%. It took approximately 2-3 seconds for detection process in our system servers.
electrical engineering and systems science
This article presents a flow visualization method for wind-waves, as well as a technique to measure flow field on two sides of interface by stereoscopic particle image velocimetry (PIV) simultaneously. The new flow visualization method applied a special illumination setup to enhance reflection and weaken refraction on interface, which improves contrast and detail performance of photos. The concatenation of flow visualization photo along flow direction is able to demonstrate the scenario from capillary waves to gravity-capillary waves at the early stage of wind-waves. After that, time-resolved stereoscopic PIV is employed in vertical planes on both sides of interface with two camera pairs. Two kinds of tracking particles with different scattering performance are spread in air or water separately. Then interface detection algorithm is designed to separate different particles and determine interface position. Adaptive min-max normalization and near-wall image preprocessing is used to improve near-wall and main stream PIV result. Finally, clip arts of results show the flow field in the same time in air and water near interface.
physics
We briefly show how the use of topological spaces and $\sigma$-algebras in physics can be rederived and understood as the fundamental requirement of experimental verifiability. We will see that a set of experimentally distinguishable objects will necessarily be endowed with a topology that is Kolmogorov (i.e. $T_0$) and second countable, which both puts constraints on well-formed scientific theories and allows us to give concrete physical meaning to the mathematical constructs. These insights can be taken as a first step in a general mathematical theory for experimental science. This work is an overview of some of the results of Assumptions of Physics, a project that aims to identify a handful of physical principles from which the basic laws can be rigorously derived (see https://assumptionsofphysics.org ).
physics
Preparation technology, the structure determination, multiband luminescence and nonlinear optical properties of the chalcogenide glasses are subject of present work. The glass samples with two different Er content were prepared by classical two-stage melt-quenching method. Glass state and morphology were confirmed by X-ray and EDS techniques. Influence of the Erbium doping on the luminescence and NLO properties was investigated.
physics
The performance requirements for fault-tolerant quantum computing are very stringent. Qubits must be manipulated, coupled, and measured with error rates well below 1%. For semiconductor implementations, silicon quantum dot spin qubits have demonstrated average single-qubit Clifford gate error rates that approach this threshold, notably with error rates of 0.14% in isotopically enriched $^{28}$Si/SiGe devices. This gate performance, together with high-fidelity two-qubit gates and measurements, is only known to meet the threshold for fault-tolerant quantum computing in some architectures when assuming that the noise is incoherent, and still lower error rates are needed to reduce overhead. Here we experimentally show that pulse engineering techniques, widely used in magnetic resonance, improve average Clifford gate error rates for silicon quantum dot spin qubits to 0.043%,a factor of 3 improvement on previous best results for silicon quantum dot devices. By including tomographically complete measurements in randomised benchmarking, we infer a higher-order feature of the noise called the unitarity, which measures the coherence of noise. This in turn allows us to theoretically predict that average gate error rates as low as 0.026% may be achievable with further pulse improvements. These fidelities are ultimately limited by Markovian noise, which we attribute to charge noise emanating from the silicon device structure itself, or the environment.
condensed matter
Until recently, deep steganalyzers in spatial domain have been all designed for gray-scale images. In this paper, we propose WISERNet (the wider separate-then-reunion network) for steganalysis of color images. We provide theoretical rationale to claim that the summation in normal convolution is one sort of "linear collusion attack" which reserves strong correlated patterns while impairs uncorrelated noises. Therefore in the bottom convolutional layer which aims at suppressing correlated image contents, we adopt separate channel-wise convolution without summation instead. Conversely, in the upper convolutional layers we believe that the summation in normal convolution is beneficial. Therefore we adopt united normal convolution in those layers and make them remarkably wider to reinforce the effect of "linear collusion attack". As a result, our proposed wide-and-shallow, separate-then-reunion network structure is specifically suitable for color image steganalysis. We have conducted extensive experiments on color image datasets generated from BOSSBase raw images and another large-scale dataset which contains 100,000 raw images, with different demosaicking algorithms and down-sampling algorithms. The experimental results show that our proposed network outperforms other state-of-the-art color image steganalytic models either hand-crafted or learned using deep networks in the literature by a clear margin. Specifically, it is noted that the detection performance gain is achieved with less than half the complexity compared to the most advanced deep-learning steganalyzer as far as we know, which is scarce in the literature.
computer science
We introduce an exactly solvable one-dimensional potential that supports both bound and/or resonance states. This potential is a generalization of the well-known 1D Morse potential where we introduced a deformation that preserves the finite spectrum property. On the other hand, in the limit of zero deformation, the potential reduces to the exponentially confining potential well introduced recently by A. D. Alhaidari. The latter potential supports infinite spectrum which means that the zero deformation limit is a critical point where our system will transition from the finite spectrum limit to the infinite spectrum limit. We solve the corresponding Schrodinger equation and obtain the energy spectrum and the eigenstates using the tridiagonal representation approach.
quantum physics
The appearance of scalar/moduli fields in the early universe, as motivated by string theory, naturally leads to non-thermal "moduli cosmology". Such cosmology provides a consistent framework where the generation of radiation, baryons, and dark matter can occur while maintaining successful Big Bang Nucleosynthesis and avoiding the cosmological moduli problem. We present a relatively economical construction with moduli cosmology, building on a variety of string-inspired components (e.g. supersymmetry, discrete symmetries, Green-Schwarz anomaly cancellation). We address a range of outstanding problems of particle physics and cosmology simultaneously, including the fermion mass hierarchy and flavor puzzle, the smallness of neutrino masses, baryogenesis and dark matter. Our setup, based on discrete $\mathrm{Z}_{12}^{R}$ symmetry and anomalous $\mathrm{U}(1)_A$, is void of the usual issues plaguing the Minimal Supersymmetric Standard Model, i.e. the $\mu$-problem and the overly-rapid proton decay due to dimension-4,-5 operators. The model is compatible with $\mathrm{SU}(5)$ Grand Unification. The smallness of Dirac neutrino masses is automatically established by requiring the cancellation of mixed gravitational-gauge anomalies. The decay of the moduli field provides a common origin for the baryon number and dark matter abundance, explaining the observed cosmic coincidences, $\Omega_{B} \sim \Omega_{DM}$.
high energy physics phenomenology
Recent practical approaches for the use of current generation noisy quantum devices in the simulation of quantum many-body problems have been dominated by the use of a variational quantum eigensolver (VQE). These coupled quantum-classical algorithms leverage the ability to perform many repeated measurements to avoid the currently prohibitive gate depths often required for exact quantum algorithms, with the restriction of a parameterized circuit to describe the states of interest. In this work, we show how the calculation of zero-temperature dynamic correlation functions defining the linear response characteristics of quantum systems can also be recast into a modified VQE algorithm, which can be incorporated into the current variational quantum infrastructure. This allows for these important physical expectation values describing the dynamics of the system to be directly converged on the frequency axis, and they approach exactness over all frequencies as the flexibility of the parameterization increases. The frequency resolution hence does not explicitly scale with gate depth, which is approximately twice as deep as a ground state VQE. We apply the method to compute the single-particle Green's function of ab initio dihydrogen and lithium hydride molecules, and demonstrate the use of a practical active space embedding approach to extend to larger systems. While currently limited by the fidelity of two-qubit gates, whose number is increased compared to the ground state algorithm on current devices, we believe the approach shows potential for the extraction of frequency dynamics of correlated systems on near-term quantum processors.
quantum physics
Cognitive Radio (CR) networks presents a paradigm shift aiming to alleviate the spectrum scarcity problem exasperated by the increasing demand on this limited resource. It promotes dynamic spectrum access, cooperation among heterogeneous devices, and spectrum sharing. Spectrum sensing is a key cognitive radio functionality, which entails scanning the RF spectrum to unveil underutilised spectral bands for opportunistic use. To achieve higher data rates while maintaining high quality of service QoS, effective wideband spectrum sensing routines are crucial due to their capability of achieving spectral awareness over wide frequency range(s)\ and efficiently harnessing the available opportunities. However, implementing wideband sensing under stringent size, weight, power and cost requirements (e.g., for portable devices) brings formidable design challenges such as addressing potential prohibitively high complexity and data acquisition rates. This article gives a survey of various wideband spectrum sensing approaches outlining their advantages and limitations; special attention is paid to approaches that utilise sub-Nyquist sampling techniques. Other aspects of CR such as cooperative sensing and performance requirements are briefly addressed. Comparison between sub-Nyquist sensing approaches is also presented.
statistics
We propose a novel electron acceleration mechanism, which we call stochastic shock drift acceleration (SSDA), that extends the standard shock drift acceleration (SDA) for low-energy electrons at a quasi-perpendicular shock to include the effect of stochastic pitch-angle scattering. We demonstrate that the steady-state energy spectrum of electrons accelerated within the shock transition region becomes a power-law in the limit of strong scattering. The spectral index is independent of the pitch-angle scattering coefficient. On the other hand, the maximum energy attainable through the mechanism scales linearly with the pitch-angle scattering coefficient. These results have been confirmed by Monte Carlo simulations that include finite pitch-angle anisotropy. We find that the theory can reasonably well explain in-situ observations of quasi-perpendicular Earth's bow shock. Theoretical scaling law suggests that the maximum energy increases in proportion to the square of the shock speed, indicating that the thermal electrons may be accelerated up to mildly relativistic energies by the SSDA at quasi-perpendicular supernova remnant shocks. Therefore, the mechanism provides a plausible solution to the long-standing electron injection problem.
astrophysics
This study explores how thermal disequilibrium during melt-infiltration and melt-rock interaction may modify the continental lithosphere from beneath. Using an idealized 1D model of thermal disequilibrium between melt-rich channels and the surrounding melt-poor material, I estimate heat exchange across channel walls during channelized melt transport at the lithosphere-asthenosphere boundary (LAB). For geologically-reasonable values of the volume fraction of channels ($\phi$), relative velocity across channel walls ($v$), channel spacing ($d$), and the timescale of episodic melt-infiltration ($\tau$), model results suggest disequilibrium heating may contribute $>$ $10^{-3}$ W/m$^3$ to the LAB heat budget. During episodic melt-infiltration, a thermal reworking zone (TRZ) associated with spatio-temporally varying disequilibrium heat exchange forms at the LAB. The TRZ grows by the transient migration of a disequilibrium-heating front at material-dependent velocity, reaching a maximum steady-state width $\delta\sim$ $\left[\phi vd^{-2}\tau^{2} \right]$. The spatio-temporal scales associated with establishment of the TRZ are comparable with those inferred for the migration of the LAB based on geologic observations within continental intra-plate settings, such as the western US.
physics
In this paper we construct a class of hairy static black holes of higher dimensional Einstein-Skyrme theories with the cosmological constant $\Lambda \le 0$ whose scalar is an $SU(2)$ valued field. The spacetime is set to be conformal to $ \mathcal{M}^4 \times \mathcal{N}^{N-4}$ where $\mathcal{M}^4$ and $\mathcal{N}^{N-4}$ are a four dimensional spacetime and a compact Einstein $(N-4)$-dimensional submanifold for $N \ge 5$, respectively, whereas $N=4$ is the trivial case. We discuss the behavior of solutions near the boundaries, namely, near the (event) horizon and in the asymptotic region. Then, we establish local-global existence of black hole solutions and show that black holes with finite energy exist if their geometries are asymptotically flat. At the end, we perform a linear stability analysis using perturbative method and give a remark about their stability.
high energy physics theory
The detection of exoplanets orbiting other stars has revolutionized our view of the cosmos. First results suggest that it is teeming with a fascinating diversity of rocky planets, including those in the habitable zone. Even our closest star, Proxima Centauri, harbors a small planet in its habitable zone, Proxima b. With the next generation of telescopes, we will be able to peer into the atmospheres of rocky planets and get a glimpse into other worlds. Using our own planet and its wide range of biota as a Rosetta stone, we explore how we could detect habitability and signs of life on exoplanets over interstellar distances. Current telescopes are not yet powerful enough to characterize habitable exoplanets, but the next generation of telescopes that is already being built will have the capabilities to characterize close-by habitable worlds. The discussion on what makes a planet a habitat and how to detect signs of life is lively. This review will show the latest results, the challenges of how to identify and characterize such habitable worlds, and how near-future telescopes will revolutionize the field. For the first time in human history, we have developed the technology to detect potential habitable worlds. Finding thousands of exoplanets has taken the field of comparative planetology beyond the Solar System.
astrophysics
We discuss an extended easy quantum group formalism, with the Schur-Weyl theoretic Kronecker symbols being as general as possible, and allowed to take values in $\{-1,0,1\}$, and more generally in $\mathbb T\cup\{0\}$. Our study includes an axiomatization of the theory, some structure and classification results, and the study of the basic examples of quantum unitary and reflection groups, with both positive and negative results.
mathematics
Let $G$ be a finite simple graph on $n$ vertices. Let $J_G \subset K[x_1, \ldots, x_n]$ be the cover ideal of $G$. In this article, we obtain syzygies, Betti numbers and Castelnuovo-Mumford regularity of $J_G^s$ for all $s \geq 1$ for certain classes of graphs $G$.
mathematics
In a surprising recent work, Lemke Oliver and Soundararajan noticed how experimental data exhibits erratic distributions for consecutive pairs of primes in arithmetic progressions, and proposed a heuristic model based on the Hardy--Littlewood conjectures containing a large secondary term, which fit very well the data. In this paper, we study consecutive pairs of sums of squares in arithmetic progressions, and develop a similar heuristic model based on the Hardy--Littlewood conjecture for sums of squares, which also explain the biases in the experimental data. In the process, we prove several results related to averages of the Hardy--Littlewood constant in the context of sums of two squares.
mathematics
We propose a method for spatially re-routing single photons or light in a coherent state with small average photon number by purely electronic means, i.e. without using mechanical devices such as micro-mirror arrays. The method is based on mapping the quantum state of the incoming light onto a spin-wave in an atomic vapor as is done in quantum memories of light. Then the wavevector of the spin-wave is modified in a controlled way by an applied magnetic field gradient or an AC Stark dressing of the atoms. Finally, by re-applying the same control beam as for storing, the signal pulse is released in a new direction that depends on the deflected wavevector of the spin-wave. We show by numerical simulation that efficiencies can be achieved for arbitrary deflection angles in the plane that are comparable with simple photon storage and re-emission in forward direction, and propose a new method for reducing decoherence in the quantum memory. In a reasonable parameter regime, the re-routing should be achievable on a time-scale on the order of micro-seconds.
quantum physics
In this work we propose a many-body Hamiltonian construction which introduces only a single separate energy scale of order $\Theta(1/N^{2+\delta})$, for a small parameter $\delta>0$, and for $N$ terms in the target Hamiltonian. In its low-energy subspace, the construction can approximate any normalized target Hamiltonian $H_\mathrm{t}=\sum_{i=1}^N h_i$ with norm ratios $r=\max_{i,j\in\{1,\ldots,N\}}\|h_i\| / \| h_j \|=O(\exp(\exp(\mathrm{poly} n)))$ to within relative precision $O(N^{-\delta})$. This comes at the expense of increasing the locality by at most one, and adding an at most poly-sized ancilliary system for each coupling; interactions on the ancilliary system are geometrically local, and can be translationally-invariant. As an application, we discuss implications for QMA-hardness of the local Hamiltonian problem, and argue that "almost" translational invariance-defined as arbitrarily small relative variations of the strength of the local terms-is as good as non-translational-invariance in many of the constructions used throughout Hamiltonian complexity theory. We furthermore show that the choice of geared limit of many-body systems, where e.g. width and height of a lattice are taken to infinity in a specific relation, can have different complexity-theoretic implications: even for translationally-invariant models, changing the geared limit can vary the hardness of finding the ground state energy with respect to a given promise gap from computationally trivial, to QMAEXP-, or even BQEXPSPACE-complete.
quantum physics
In this article, we propose a new variational approach to learn private and/or fair representations. This approach is based on the Lagrangians of a new formulation of the privacy and fairness optimization problems that we propose. In this formulation, we aim at generating representations of the data that keep a prescribed level of the relevant information that is not shared by the private or sensitive data, while minimizing the remaining information they keep. The proposed approach (i) exhibits the similarities of the privacy and fairness problems, (ii) allows us to control the trade-off between utility and privacy or fairness through the Lagrange multiplier parameter, and (iii) can be comfortably incorporated to common representation learning algorithms such as the VAE, the $\beta$-VAE, the VIB, or the nonlinear IB.
statistics
Liquid dropout and retention in gas-condensate reservoirs, specially in the near wellbore region, obstruct gas flowing paths and impact negatively the produced fluid volume and composition. Yet, condensate banking forecasting is commonly inaccurate, as experiments seldom reproduce reservoir extreme conditions and complex fluid composition, while most pore-scale models oversimplify the physics of phase transitions between gas and condensate. To address this gap, a fully implicit isothermal compositional pore-network model for gas and condensate flow is presented. The proposed pore-networks consist of 3D structures of pores connected by constricted circular capillaries. Hydraulic conductances are calculated for the capillaries, which can exhibit single-phase flow or two-phase annular flow, according to local gas and liquid saturations, or be blocked by a liquid bridge, when capillary forces overcome viscous forces. A PT-flash based on the Peng-Robinson EoS is performed at control volumes defined for the pores at each time step, updating the phases properties. Flow analyses were carried based on coreflooding experiments reported in the literature, with matching fluid composition and flow conditions, and approximated pore-space geometry. Predicted and measured relative permeability curves showed good quantitative agreement, for two values of interfacial tension and three values of gas flow velocity.
physics
We propose a novel mean field games (MFGs) based GAN(generative adversarial network) framework. To be specific, we utilize the Hopf formula in density space to rewrite MFGs as a primal-dual problem so that we are able to train the model via neural networks and samples. Our model is flexible due to the freedom of choosing various functionals within the Hopf formula. Moreover, our formulation mathematically avoids Lipschitz-1 constraint. The correctness and efficiency of our method are validated through several experiments.
computer science
We describe the order type of range sets of compact ultrametrics and show that an ultrametrizable infinite topological space $(X, \tau)$ is compact iff the range sets are order isomorphic for any two ultrametrics compatible with the topology $\tau$. It is also shown that an ultrametrizable topology is separable iff every compatible with this topology ultrametric has at most countable range set.
mathematics
We investigate the AdS/CFT correspondence for quiver gauge theories realized on D3-branes put on abelian orbifolds by using the superconformal index. We assume that on the gravity side the finite $N$ corrections of the index are reproduced by D3-branes wrapped on three particular three-cycles in the internal space ${\cal Y}$, the abelian orbifold of $\boldsymbol{S}^5$. We first establish the relation between baryonic charges on the gauge theory side and the D3-brane wrapping numbers and holonomies on D3-branes. Then we confirm our proposal by comparing the results of localization for gauge theories and the results on the AdS side including the contributions of D3-branes and excitation on them for many examples. We only focus on the leading finite $N$ corrections starting from $q^N$, and leave the sub-leading corrections starting at $q^{kN}$ ($k\geq2$) as a task for the future. We find complete agreement for the leading corrections in all examples.
high energy physics theory
In this note, we discuss the impact of the recent Belle result on the various theoretical explanations of the $R_D$ and $R_{D^*}$ anomalies. The pure tensor explanation, which was strongly disfavoured by the measurements of $F_L^{D^*}$ and high-$p_T$ $p \, p \to \tau \, \nu$ searches before Moriond, is now completely allowed because of reduction of the experimental world-average. Moreover, the pure right-chiral vector solution (involving right-chiral neutrinos) has now moved into the $2\sigma$ allowed range of the LHC $p \, p \to \tau \, \nu$ searches. We also critically re-examine the bound on $\mathcal{B}(B_c^- \to \tau^- \bar{\nu}_\tau)$ from LEP data and show that the bound is considerably weaker than the number $10\%$ often used in the recent literature.
high energy physics phenomenology
We investigate a quantum spatial search problem on fractal lattices, such as Sierpinski carpets and Menger sponges. In earlier numerical studies of the Sierpinski gasket, the Sierpinski tetrahedron, and the Sierpinski carpet, conjectures have been proposed for the scaling of a quantum spatial search problem finding a specific target, which is given in terms of the characteristic quantities of a fractal geometry. We find that our simulation results for extended Sierpinski carpets and Menger sponges support the conjecture for the ${\it optimal}$ number of the oracle calls, where the exponent is given by $1/2$ for $d_{\rm s} > 2$ and the inverse of the spectral dimension $d_{\rm s}$ for $d_{\rm s} < 2$. We also propose a scaling hypothesis for the ${\it effective}$ number of the oracle calls defined by the ratio of the ${\it optimal}$ number of oracle calls to a square root of the maximum finding probability. The form of the scaling hypothesis for extended Sierpinski carpets is very similar but slightly different from the earlier conjecture for the Sierpinski gasket, the Sierpinski tetrahedron, and the conventional Sierpinski carpet.
quantum physics
Oblique plane microscopy (OPM) is a single objective light-sheet microscopy which performs three dimensional (3D) imaging by axial scan of the generated light-sheet. Recently, multiple techniques for lateral scan of the generated light-sheet in OPM have emerged. However, their suitability for geometrically distortion free 3D imaging, which essentially requires a constant tilt light-sheet scan, has not been evaluated. In this work, we use a geometrical optics approach and derive analytical relationship for the amount of tilt variance in planar mirror based scanned oblique plane illumination (SOPi) arrangement. We experimentally validate the derived relationship and use it to arrive at an optimized scanner geometry and to understand its associated limitations. We also discuss the effects of scanning on optical aberrations and 3D field of view in optimized, tilt invariant, lateral scanning OPM systems.
physics
We investigate the characteristics of $\sigma$, $f_{0}(980)$, and $a_{0}(980)$ with the formalism of chiral unitary approach. With the dynamical generation of them, we make a further study of their properties by evaluating the couplings, the compositeness, the wave functions and the radii. We also research their properties in the single channel interactions, where the $a_{0}(980)$ can not be reproduced in the $K\bar{K}$ interactions with isospin $I=1$ since the potential is too weak. In our results, the states of $\sigma$ and $f_{0}(980)$ can be dynamically reproduced stably with varying cutoffs both in the coupled channel and the single channel cases. We find that the $\pi\eta$ components is much important in the coupled channel interactions to dynamically reproduce the $a_{0}(980)$ state, which means that $a_{0}(980)$ state can not be a pure $K\bar{K}$ molecular state. We obtain their radii as: $|\langle r^2 \rangle|_{f_0(980)} = 1.80 \pm 0.35$ fm, $|\langle r^2 \rangle|_{\sigma} = 0.68 \pm 0.05$ fm and $|\langle r^2 \rangle|_{a_0(980)} = 0.94 \pm 0.09$ fm. Based on our investigation results, we conclude that the $f_{0}(980)$ state is mainly a $K\bar{K}$ bound state, the $\sigma$ state a resonance of $\pi\pi$ and the $a_{0}(980)$ state a loose $K\bar{K}$ bound state. From the results of the compositeness, they are not pure molecular states and have something non-molecular components, especially for the $\sigma$ state.
high energy physics phenomenology
We address the dynamics of continuous-time quantum walk (CTQW) on planar 2D lattice graphs, i.e. those forming a regular tessellation of the Euclidean plane (triangular, square, and honeycomb lattice graphs). We first consider the free particle: on square and triangular lattice graphs we observe the well-known ballistic behavior, whereas on the honeycomb lattice graph we obtain a sub-ballistic one, although still faster than the classical diffusive one. We impute this difference to the different amount of coherence generated by the evolution and, in turn, to the fact that, in 2D, the square and the triangular lattices are Bravais lattices, whereas the honeycomb one is non-Bravais. From the physical point of view, this means that CTQWs are not universally characterized by the ballistic spreading. We then address the dynamics in the presence of a perpendicular uniform magnetic field and study the effects of the field by two approaches: (i) introducing the Peierls phase-factors, according to which the tunneling matrix element of the free particle becomes complex, or (ii) spatially discretizing the Hamiltonian of a spinless charged particle in the presence of a magnetic field. Either way, the dynamics of an initially localized walker is characterized by a lower spread compared to the free particle case, the larger is the field the more localized stays the walker. Remarkably, upon analyzing the dynamics by spatial discretization of the Hamiltonian (vector potential in the symmetric gauge), we obtain that the variance of the space coordinate is characterized by pseudo-oscillations, a reminiscence of the harmonic oscillator behind the Hamiltonian in the continuum, whose energy levels are the well-known Landau levels.
quantum physics
We study a generalization of the Mermin-Peres magic square game to arbitrary rectangular dimensions. After exhibiting some general properties, these rectangular games are fully characterized in terms of their optimal win probabilities for quantum strategies. We find that for $m \times n$ rectangular games of dimensions $m,n \geq 3$ there are quantum strategies that win with certainty, while for dimensions $1 \times n$ quantum strategies do not outperform classical strategies. The final case of dimensions $2 \times n$ is richer, and we give upper and lower bounds that both outperform the classical strategies. Finally, we apply our findings to quantum certified randomness expansion to find the noise tolerance and rates for all magic rectangle games. To do this, we use our previous results to obtain the winning probability of games with a distinguished input for which the devices give a deterministic outcome, and follow the analysis of C. A. Miller and Y. Shi [SIAM J. Comput. 46, 1304 (2017)].
quantum physics
The universal asymptotic amplitude ratio between the gyration radius and the hydrodynamic radius of self-avoiding walks is estimated by high-resolution Monte Carlo simulations. By studying chains of length of up to $N = 2^{25} \approx 34 \times 10^6$ monomers, we find that the ratio takes the value $R_{\mathrm{G}}/R_{\mathrm{H}} = 1.5803940(45)$, which is several orders of magnitude more accurate than the previous state of the art. This is facilitated by a sampling scheme which is quite general, and which allows for the efficient estimation of averages of a large class of observables. The competing corrections to scaling for the hydrodynamic radius are clearly discernible. We also find improved estimates for other universal properties that measure the chain dimension. In particular, a method of analysis which eliminates the leading correction to scaling results in a highly accurate estimate for the Flory exponent of $\nu = 0.58759700(40)$.
condensed matter
Titan's surface was revealed by Cassini's instruments, showing the presence of liquid hydrocarbons in lakes, and features like dry riverbed. In order to study the sediment transport in Titan's channels and to map distribution of the water-ice signature in these terrains, we use a radiative transfer model to retrieve the surface albedo, after we estimated VIMS error via an original method. We also establish a criteria related to the intensity of the water ice signature. The tuning of the radiative transfer model shows that the fractal dimension of Titan's aerosols is higher than previously thought, around 2.3 - 2.4. We find spots of increasing signal of water ice downstream, at the margins of Tui Regio, that could correspond to alluvial fans, deltas or crater rims. We also observe a very low water ice signal on Tui Regio, with a positive gradient between the central region and the boundary of the area, possibly due to the thickness variation of an evaporitic layer. The riverbeds show within the error bars a decreasing grain size from the top to the bottom of the channels.
astrophysics
We consider a free boundary problem of the Navier--Stokes equations in the three-dimensional Euclidean space with moving contact line, where the 90$^\circ$-contact angle condition is posed. We show that for given $T > 0$ the problem is local well-posed on $(0, T)$ provided that the initial data are small. In contrast to the strategy in Wilke (2013), we study the transformed problem in an $L^p$-in-time and $L^q$-in-space setting, which yields the optimal regularity of the initial data.
mathematics
Ultra-wideband technology has emerged in recent years as a robust solution for localization in GNSS denied environments. In particular, its high accuracy when compared to other wireless localization solutions is enabling a wider range of collaborative and multi-robot application scenarios, being able to replace more complex and expensive motion-capture areas for use cases where accuracy in the order of tens of centimeters is sufficient. We present the first survey of UWB-based localization focused on multi-UAV systems and heterogeneous multi-robot systems. We have found that previous literature reviews do not consider in-depth the challenges in both aerial navigation and navigation with multiple robots, but also in terms of heterogeneous multi-robot systems. In particular, this is, to the best of our knowledge, the first survey to review recent advances in UWB-based (i) methods that enable ad-hoc and dynamic deployments; (ii) collaborative localization techniques; and (iii) cooperative sensing and cooperative maneuvers such as UAV docking on mobile platforms. Finally, we also review existing datasets and discuss the potential of this technology for both localization in GNSS-denied environments and collaboration in multi-robot systems.
electrical engineering and systems science
This paper studies a variant of multi-player reach-avoid game played between intruders and defenders. The intruder team tries to score by sending as many intruders as possible to the target area, while the defender team tries to minimize this score by intercepting them. Specifically, we consider the case where the defenders are constrained to move on the perimeter of the target area. Since it is challenging to directly solve the multi-player game due to the high dimensionality of the joint state space, we leverage the solutions to smaller scale problems. First, we solve the one vs. one game, for which existing works either rely on numerical approaches or make simplifying assumptions (e.g., circular perimeter, or equal speed). This paper accommodates target areas with any arbitrary convex shapes and provides analytical solution which lends itself to a useful geometric interpretation. We also provide a detailed discussion on the optimality of the derived strategies. Secondly, we solve the two vs. one game to introduce a cooperative pincer maneuver, where a pair of defenders team up to capture an intruder that cannot be captured by either one of the defender individually. Finally, we introduce how the aforementioned building blocks are used in three different assignment-based defense strategies.
electrical engineering and systems science
We propose a next-to-minimal supersymmetric Standard Model (NMSSM) extended by an $\mathbb{A}_{4}\times \boldsymbol{Z}_{3}$ flavor symmetry and three right-handed neutrinos providing a detailed study of the neutrino sector and a solution to the domain wall problem. In this proposal, neutrino masses are generated through Type I seesaw mechanism while the mixing angles are described by the trimaximal mixing realized using the NMSSM singlet S and only two flavon fields. The phenomenology of neutrino parameters is studied for normal and inverted mass hierarchies. In particular, we numerically evaluated the observables related to neutrino masses and mixing, namely, $\sum m_{i}$, $m_{ee}$, $m_{\nu_{e}}$, and $\delta_{CP}$ where we find that the ranges of $m_{ee}$ and $m_{\nu_{e}}$ are accessible by current and future experiments while the obtained ranges of $\sum m_{i}$ and $\delta_{CP}$ lie within the current experimental data. Another attractive feature we discussed in this paper is the circumvention of the domain wall problem induced by the spontaneous breaking of the $\mathbb{A}_{4}\times \boldsymbol{Z}_{3}$ discrete symmetry. We first showed that the domain walls in the charged lepton sector occur at high energy scale leading to unproblematic domain walls, while in the neutrino sector they are inevitable. Then, to solve this problem, we reconsidered the well-known approach that relies on the explicit breaking of the discrete symmetry through the insertion of Planck-suppressed operators induced by supergravity. \keywords{Neutrino physics, Discrete flavor symmetry,Trimaximal mixing, Domain walls}
high energy physics phenomenology
Locally, the atomic structure in well annealed amorphous silicon appears similar to that of crystalline silicon. We address here the question whether a point defect, specifically a vacancy, in amorphous silicon also resembles that in the crystal. From density functional theory calculations of a large number of nearly defect free configurations, relaxed after an atom has been removed, we conclude that there is little similarity. The analysis is based on formation energy, relaxation energy, bond lengths, bond angles, Vorono\"i volume, coordination, atomic charge and electronic gap states. All these quantities span a large and continuous range in amorphous silicon and while the removal of an atom leads to the formation of one to two bond defects and to a lowering of the local atomic density, the relaxation of the bonding network is highly effective, and the signature of the vacancy generally unlike that of a vacancy in the crystal.
condensed matter
A recent refinement of Ker\'ekj\'art\'o's Theorem has shown that in $\mathbb R$ and $\mathbb R^2$ all $\mathcal C^l$-solutions of the functional equation $f^n =\textrm{Id}$ are $\mathcal C^l$-linearizable, where $l\in \{0,1,\dots \infty\}$. When $l\geq 1$, in the real line we prove that the same result holds for solutions of $f^n=f$, while we can only get a local version of it in the plane. Through examples, we show that these results are no longer true when $l=0$ or when considering the functional equation $f^n=f^k$ with $n>k\geq 2$.
mathematics
We propose a new method for multivariate response regressions where the elements of the response vector can be of mixed types, for example some continuous and some discrete. Our method is based on a model which assumes the observable mixed-type response vector is connected to a latent multivariate normal response linear regression through a link function. We explore the properties of this model and show its parameters are identifiable under reasonable conditions. We propose an algorithm for approximate maximum likelihood estimation that works "off-the-shelf" with many different combinations of response types, and which scales well in the dimension of the response vector. Our method typically gives better predictions and parameter estimates than fitting separate models for the different response types and allows for approximate likelihood ratio testing of relevant hypotheses such as independence of responses. The usefulness of the proposed method is illustrated using simulations and through three data examples.
statistics
The result of removing of heavy non-equal mass particles from the theory can be described, at low energy, by the effective action, which is a series in inverse-square powers of the mass. We propose a new efficient tool to calculate the leading terms of this series based on the Schwinger proper-time method. Unequal masses give rise to a large number of effective vertices describing the explicit flavour symmetry breaking effects with well-defined coupling constants. Our method is pertinent to the theory with explicit and spontaneous chiral symmetry breaking, chiral gauge theory, standard and beyond standard model effective field theory, the theory of critical phenomena, cosmology, etc.
high energy physics theory
All networks can be analyzed at multiple scales. A higher scale of a network is made up of macro-nodes: subgraphs that have been grouped into individual nodes. Recasting a network at higher scales can have useful effects, such as decreasing the uncertainty in the movement of random walkers across the network while also decreasing the size of the network. However, the task of finding such a macroscale representation is computationally difficult, as the set of all possible scales of a network grows exponentially with the number of nodes. Here we compare various methods for finding the most informative scale of a network, discovering that an approach based on spectral analysis outperforms greedy and gradient descent-based methods. We then use this procedure to show how several structural properties of preferential attachment networks vary across scales. We describe how meso- and macroscale representations of networks can have significant benefits over their underlying microscale, which include properties such as increase in determinism, a decrease in degeneracy, a lower entropy rate of random walkers on the network, an increase in global network efficiency, and higher values for a variety of centrality measures than the microscale.
computer science
We report on a discrete-time quantum walk that uses the momentum of ultra-cold rubidium-87 atoms as the walk space and two internal atomic states as the coin degree of freedom. Each step of the walk consists of a coin toss (a microwave pulse) followed by a unitary shift operator (a resonant ratchet pulse). We carry out a comprehensive experimental study on the effects of various parameters, including the strength of the shift operation, coin parameters, noise, and initialization of the system on the behavior of the walk. The walk dynamics can be well controlled in our experiment; potential applications include atom interferometry and engineering asymmetric walks.
quantum physics
We study a fluid-structure interaction problem describing movement of a rigid body inside a bounded domain filled by a viscous fluid. The fluid is modelled by the generalized incompressible Naiver-Stokes equations which include cases of Newtonian and non-Newtonian fluids. The fluid and the rigid body are coupled via the Navier slip boundary conditions and balance of forces at the fluid-rigid body interface. Our analysis also includes the case of the nonlinear slip condition. The main results assert the existence of strong solutions, in an $L^p-L^q$ setting, globally in time, for small data. The proof essentially uses the maximal regularity property of the associated linear system which is obtained by proving the $\mathcal{R}$-sectoriality of the corresponding operator. Moreover, in the Newtonian case, we also prove the exponential stability of the system.
mathematics
Legged robots need to make contact with irregular surfaces, when operating in unstructured natural terrains. Representing and perceiving these areas to reason about potential contact between a robot and its surrounding environment, is still largely an open problem. This paper introduces a new framework to model and map local rough terrain surfaces, for tasks such as bipedal robot foot placement. The system operates in real-time, on data from an RGB-D and an IMU sensor. We introduce a set of parametrized patch models and an algorithm to fit them in the environment. Potential contacts are identified as bounded curved patches of approximately the same size as the robot's foot sole. This includes sparse seed point sampling, point cloud neighborhood search, and patch fitting and validation. We also present a mapping and tracking system, where patches are maintained in a local spatial map around the robot as it moves. A bio-inspired sampling algorithm is introduced for finding salient contacts. We include a dense volumetric fusion layer for spatiotemporally tracking, using multiple depth data to reconstruct a local point cloud. We present experimental results on a mini-biped robot that performs foot placements on rocks, implementing a 3D foothold perception system, that uses the developed patch mapping and tracking framework.
computer science
MRI offers outstanding soft tissue contrast that may reduce uncertainties in target and organ-at-risk delineation and enable online adaptive image-guided treatment. Spatial distortions resulting from non-linearities in the gradient fields and non-uniformity in the main magnetic field must be accounted for across the imaging field-of-view to prevent systematic errors during treatment delivery. This work presents a modular phantom and software application to characterize geometric distortion (GD) within the large field-of-view MRI images required for radiation therapy simulation. The modular phantom is assembled from a series of rectangular foam blocks containing high-contrast fiducial markers in a known configuration. The modular phantom design facilitates transportation of the phantom between different MR scanners and MR-guided linear accelerators and allows the phantom to be adapted to fit different sized bores or coils. The phantom was evaluated using a 1.5T MR-guided linear accelerator (MR-Linac) and 1.5T and 3.0T diagnostic scanners. Performance was assessed by varying acquisition parameters to induce image distortions in a known manner. Imaging was performed using T1 and T2 weighted pulse sequences with 2D and 3D distortion correction algorithms and the receiver bandwidth (BW) varied as 250-815 Hz/pixel. Phantom set-up reproducibility was evaluated across independent set-ups. The software was validated by comparison with a non-modular phantom. Average geometric distortion was 0.94+/-0.58 mm for the MR-Linac, 0.90+/-0.53 mm for the 1.5 T scanner, and 1.15+/-0.62 mm for the 3.0T scanner, for a 400 mm diameter volume-of-interest. GD increased, as expected, with decreasing BW, and with the 2D versus 3D correction algorithm. Differences in GD attributed to phantom set-up were 0.13 mm or less. Differences in GD for the two software applications were less than 0.07 mm.
physics
We define a supersymmetric quantum mechanics of fermions that take values in a simple Lie algebra. We summarize what is known about the spectrum and eigenspaces of the Laplacian which corresponds to the Koszul differential d. Firstly, we concentrate on the zero eigenvalue eigenspace which coincides with the Lie algebra cohomology. We provide physical insight into useful tools to compute the cohomology, namely Morse theory and the Hochschild-Serre spectral sequence. We list explicit generators for the Lie algebra cohomology ring. Secondly, we concentrate on the eigenspaces of the supersymmetric quantum mechanics with maximal eigenvalue at given fermion number. These eigenspaces have an explicit description in terms of abelian ideals of a Borel subalgebra of the simple Lie algebra. We also introduce a model of Lie algebra valued fermions in two dimensions, where the spaces of maximal eigenvalue acquire a cohomological interpretation. Our work provides physical interpretations of results by mathematicians, and simplifies the proof of a few theorems. Moreover, we recall that these mathematical results play a role in pure supersymmetric gauge theory in four dimensions, and observe that they give rise to a canonical representation of the four-dimensional chiral ring.
high energy physics theory
First microscopic theory for electron-phonon energy exchange in Anderson insulators is developed. The major contribution to the cooling power as a function of electron temperature is shown to be directly related to the correlation function of the local density of electron states at small energy difference argument. In Anderson insulators not far from localization transition, this correlation function is strongly enhanced by wave-function's multi-fractality and, additionally, by the presence of Mott's resonant pairs of localized states. The theory we develop explains huge enhancement of the cooling power observed in insulating Indium Oxide films as compared to predictions of the theory previously developed for disordered metals. Our results open the way to predict the conditions appropriate for the observation of Many Body Localization transition those presence in electronic insulators was advocated in the seminal paper by Basko, Aleiner and Altshuler (2006) but have not been convincingly demonstrated yet.
condensed matter
A high-precision charge measurement can be achieved by the area integration of a digitized quasi-Gaussian signal after the signal passes through the shaper and analog-to-digital converter (ADC). The charge measurement contains an error due to the uncertainty of the first sampled point of a signal waveform. To reduce the error, we employ a time-to-digital converter (TDC) to measure the uncertainty precisely, and we design correction algorithms to improve the resolution of the charge measurement. This work includes analysis and simulations of the proposed algorithms and implementation of them in an FPGA device. Besides, the tests are also conducted to evaluate the performance of the correction method. Test results indicate that the resolution of the charge measurement is successfully improved from 0.231% to 0.126% by using a signal from the shaping circuit (with the amplitude of 2 V, and leading and trailing edges of about 80 ns and 280 ns, respectively) digitized at the sampling rate of 62.5 Msps.
physics
In wall-bounded flows, the laminar regime remain linearly stable up to large values of the Reynolds number while competing with nonlinear turbulent solutions issued from finite amplitude perturbations. The transition to turbulence of plane channel flow (plane Poiseuille flow) is more specifically considered via numerical simulations. Previous conflicting observations are reconciled by noting that the two-dimensional directed percolation scenario expected for the decay of turbulence may be interrupted by a symmetry-breaking bifurcation favoring localized turbulent bands. At the other end of the transitional range, a preliminary study suggests that the laminar-turbulent pattern leaves room to a featureless regime beyond a well defined threshold to be determined with precision.
physics
Audio Sentiment Analysis is a popular research area which extends the conventional text-based sentiment analysis to depend on the effectiveness of acoustic features extracted from speech. However, current progress on audio sentiment analysis mainly focuses on extracting homogeneous acoustic features or doesn't fuse heterogeneous features effectively. In this paper, we propose an utterance-based deep neural network model, which has a parallel combination of Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) based network, to obtain representative features termed Audio Sentiment Vector (ASV), that can maximally reflect sentiment information in an audio. Specifically, our model is trained by utterance-level labels and ASV can be extracted and fused creatively from two branches. In the CNN model branch, spectrum graphs produced by signals are fed as inputs while in the LSTM model branch, inputs include spectral features and cepstrum coefficient extracted from dependent utterances in audio. Besides, Bidirectional Long Short-Term Memory (BiLSTM) with attention mechanism is used for feature fusion. Extensive experiments have been conducted to show our model can recognize audio sentiment precisely and quickly, and demonstrate our ASV is better than traditional acoustic features or vectors extracted from other deep learning models. Furthermore, experimental results indicate that the proposed model outperforms the state-of-the-art approach by 9.33\% on Multimodal Opinion-level Sentiment Intensity dataset (MOSI) dataset.
electrical engineering and systems science
This paper conducts a comparative study of proximal gradient methods (PGMs) and proximal DC algorithms (PDCAs) for sparse regression problems which can be cast as Difference-of-two-Convex-functions (DC) optimization problems. It has been shown that for DC optimization problems, both General Iterative Shrinkage and Thresholding algorithm (GIST), a modified version of PGM, and PDCA converge to critical points. Recently some enhanced versions of PDCAs are shown to converge to d-stationary points, which are stronger necessary condition for local optimality than critical points. In this paper we claim that without any modification, PGMs converge to a d-stationary point not only to DC problems but also to more general nonsmooth nonconvex problems under some technical assumptions. While the convergence to d-stationary points is known for the case where the step size is small enough, the finding of this paper is valid also for extended versions such as GIST and its alternating optimization version, which is to be developed in this paper. Numerical results show that among several algorithms in the two categories, modified versions of PGM perform best among those not only in solution quality but also in computation time.
mathematics
Triple junction (InGaP/GaAs/Ge) and single junction (SJ) solar cells were irradiated with electrons, protons and neutrons. The degradation of remaining factors was analyzed as function of the induced Displacement Damage Dose (DDD) calculated by means of the SR-NIEL (Screened Relativistic Non Ionizing\ Energy Loss) approach. In particular, the aim of this work is to analyze the variation of the solar cells remaining factors due to neutron irradiation with respect to those previously obtained with electrons and protons. The current analysis confirms that the degradation of the $P_{max}$ electrical parameter is related by means of the usual semi-empirical expression to the displacement dose, independently of type of the incoming particle. $I_{sc}$ and $V_{oc}$ parameters were also measured as a function of the displacement damage dose. Furthermore, a DLTS analysis was carried out on diodes - with the same epitaxial structure as the middle sub-cell - irradiated with neutrons.
physics
The Anderson model with decoherence features a temporal evolution from localized eigenstates to a uniform spatial distribution bar any interference features. We discuss the growth and decay of pronounced interference peaks on transient time-scales and develop an analytic understanding for the emergence of these peaks.
quantum physics
This paper aims at developing a clustering approach with spectral images directly from the compressive measurements of coded aperture snapshot spectral imager (CASSI). Assuming that compressed measurements often lie approximately in low dimensional subspaces corresponding to multiple classes, state of the art methods generally obtains optimal solution for each step separately but cannot guarantee that it will achieve the globally optimal clustering results. In this paper, a low-rank subspace representation (LRSR) algorithm is proposed to perform clustering on the compressed measurements. In addition, a subspace structured norm is added into the objective of low-rank representation problem exploiting the fact that each point in a union of subspaces can be expressed as a sparse linear combination of all other points and that the matrix of the points within each subspace is low rank. Simulation with real dataset illustrates the accuracy of the proposed spectral image clustering approach.
electrical engineering and systems science
In the framework of the modified chromo-magnetic interaction model, we perform a systematical study on the mass spectrums of the ground pentaquark states with $qqqq\bar{Q}$, $qqqQ\bar{q}$, $QQQQ\bar{q}$, $QQQQ\bar{Q}$, and $QQQq\bar{Q}$, $(Q=c,b; q=n,s; n=u,d)$ configurations. The isospin-color-spin wave functions satisfying Pauli principle for each type of ground pentaquark states are constructed. With the help of experimental data, we estimate their possible mass spectrums in two different schemes. Based on our results, we present a detailed analysis on the mass spectrums and decay behaviors for the discussed pentquark states. We hope that our study will be helpful to experimentally search for such types of the exotic pentaquark states in the future.
high energy physics phenomenology
Device-independent quantum cryptography allows security even if the devices used to execute the protocol are untrusted - whether this is due to unknown imperfections in the implementation, or because the adversary himself constructed them to subvert the security of the protocol. While device-independence has seen much attention in the domain of quantum key distribution, relatively little is known for general protocols. Here we introduce a new model for device-independence for two-party protocols and position verification in the noisy-storage model. For the first time, we show that such protocols are secure in the most general device-independent model in which the devices may have arbitrary memory, states and measurements. In our analysis, we make use of a slight modification of a beautiful new tool developed in [arXiv:1607.01796] called "Entropy Accumulation Theorem". What's more, the protocols we analyze use only simple preparations and measurements, and can be realized using any experimental setup able to perform a CHSH Bell test. Specifically, security can be attained for any violation of the CHSH inequality, where a higher violation merely leads to a reduction in the amount of rounds required to execute the protocol.
quantum physics
We generalize the Bartels-Ermolaev-Ryskin approach for the $g_1$ structure function at small-$x$ to determine the small-$x$ asymptotic behavior of the orbital angular momentum distributions in QCD. We present an exact analytical solution of the evolution equation in the double logarithmic approximation and discuss its implications for the proton spin problem.
high energy physics phenomenology
Random geometric graphs are a popular choice for a latent points generative model for networks. Their definition is based on a sample of $n$ points $X_1,X_2,\cdots,X_n$ on the Euclidean sphere~$\mathbb{S}^{d-1}$ which represents the latent positions of nodes of the network. The connection probabilities between the nodes are determined by an unknown function (referred to as the "link" function) evaluated at the distance between the latent points. We introduce a spectral estimator of the pairwise distance between latent points and we prove that its rate of convergence is the same as the nonparametric estimation of a function on $\mathbb{S}^{d-1}$, up to a logarithmic factor. In addition, we provide an efficient spectral algorithm to compute this estimator without any knowledge on the nonparametric link function. As a byproduct, our method can also consistently estimate the dimension $d$ of the latent space.
statistics
The primary objective of this scholarly work is to develop two estimation procedures - maximum likelihood estimator (MLE) and method of trimmed moments (MTM) - for the mean and variance of lognormal insurance payment severity data sets affected by different loss control mechanism, for example, truncation (due to deductibles), censoring (due to policy limits), and scaling (due to coinsurance proportions), in insurance and financial industries. Maximum likelihood estimating equations for both payment-per-payment and payment-per-loss data sets are derived which can be solved readily by any existing iterative numerical methods. The asymptotic distributions of those estimators are established via Fisher information matrices. Further, with a goal of balancing efficiency and robustness and to remove point masses at certain data points, we develop a dynamic MTM estimation procedures for lognormal claim severity models for the above-mentioned transformed data scenarios. The asymptotic distributional properties and the comparison with the corresponding MLEs of those MTM estimators are established along with extensive simulation studies. Purely for illustrative purpose, numerical examples for 1500 US indemnity losses are provided which illustrate the practical performance of the established results in this paper.
statistics
Alignment of OCS, CS$_2$ and I$_2$ molecules embedded in helium nanodroplets is measured as a function of time following rotational excitation by a non-resonant, comparatively weak ps laser pulse. The distinct peaks in the power spectra, obtained by Fourier analysis, are used to determine the rotational, B, and centrifugal distortion, D, constants. For OCS, B and D match the values known from IR spectroscopy. For CS$_2$ and I$_2$, they are the first experimental results reported. The alignment dynamics calculated from the gas-phase rotational Schr\"{o}dinger equation, using the experimental in-droplet B and D values, agree in detail with the measurement for all three molecules. The rotational spectroscopy technique for molecules in helium droplets introduced here should apply to a range of molecules and complexes.
physics
Superconducting Josephson junction qubits constitute the main current technology for many applications, including scalable quantum computers and thermal devices. Theoretical modeling of such systems is usually done within the two-level approximation. However, accurate theoretical modeling requires taking into account the influence of the higher excited states without limiting the system to the two-level qubit subspace. Here, we study the dynamics and control of a superconducting transmon using the numerically exact stochastic Liouville-von Neumann equation approach. We focus on the role of state leakage from the ideal two-level subspace for bath induced decay and single-qubit gate operations. We find significant short-time state leakage due to the strong coupling to the bath. We quantify the leakage errors in single-qubit gates and demonstrate their suppression with DRAG control for a five-level transmon in the presence of decoherence. Our results predict the limits of accuracy of the two-level approximation and possible intrinsic constraints in qubit dynamics and control for an experimentally relevant parameter set.
quantum physics
One of the most accurate methods for solving the time-dependent Schr\"{o}dinger equation uses a combination of the dynamic Fourier method with the split-operator algorithm on a tensor-product grid. To reduce the number of required grid points, we let the grid move together with the wavepacket, but find that the na\"ive algorithm based on an alternate evolution of the wavefunction and grid destroys the time reversibility of the exact evolution. Yet, we show that the time reversibility is recovered if the wavefunction and grid are evolved simultaneously during each kinetic or potential step; this is achieved by using the Ehrenfest theorem together with the splitting method. The proposed algorithm is conditionally stable, symmetric, time-reversible, and conserves the norm of the wavefunction. The preservation of these geometric properties is shown analytically and demonstrated numerically on a three-dimensional harmonic model and collinear model of He-H$_{2}$ scattering. We also show that the proposed algorithm can be symmetrically composed to obtain time-reversible integrators of an arbitrary even order. We observed $10000$-fold speedup by using the tenth- instead of the second- order method to obtain a solution with a time discretization error below $10^{-9}$. Moreover, using the adaptive grid instead of the fixed grid resulted in a 64-fold reduction in the required number of grid points in the harmonic system and made it possible to simulate the He-H$_{2}$ scattering for six times longer, while maintaining reasonable accuracy. Applicability of the algorithm to high-dimensional quantum dynamics is demonstrated using the strongly anharmonic eight-dimensional H\'{e}non--Heiles model.
quantum physics
The dynamics of both global and local vortices with non-Abelian orientational moduli is investigated in detail. Head-on collisions of these vortices are numerically simulated for parallel, anti-parallel and orthogonal internal orientations where we find interesting dynamics of the orientational moduli. A detailed study of the inter-vortex force is provided and a phase diagram separating Abelian and non-Abelian vortex types is constructed. Some results on scatterings with non-zero impact parameter and multi-vortex collisions are included.
high energy physics theory
This is a self-contained tour of the Conley index and connection matrices. The starting point is Conley's fundamental theorem of dynamical systems. There is a short stop at the necessary topological background, before we proceed to the basic properties of the index. Then, the itinerary passes through the construction of connection matrices with a panoramic view of the applications: detect heteroclinic orbits arising in delay differential equations, and partial differential equations of parabolic type. The ride will be filled with examples and figures.
mathematics
It is known that every two-qubit unitary operation has Schmidt rank one, two or four, and the construction of three-qubit unitary gates in terms of Schmidt rank remains an open problem. We explicitly construct the gates of Schmidt rank from one to seven. It turns out that the three-qubit Toffoli and Fredkin gate respectively have Schmidt rank two and four. As an application, we implement the gates using quantum circuits of CNOT gates and local Hadamard and flip gates. In particular, the collective use of three CNOT gates can generate a three-qubit unitary gate of Schmidt rank seven in terms of the known Strassen tensor from multiplicative complexity. Our results imply the connection between the number of CNOT gates for implementing multiqubit gates and their Schmidt rank.
quantum physics
A fast Bayesian method that seamlessly fuses classification and hypothesis testing via discriminant analysis is developed. Building upon the original discriminant analysis classifier, modelling components are added to identify discriminative variables. A combination of cake priors and a novel form of variational Bayes we call reverse collapsed variational Bayes gives rise to variable selection that can be directly posed as a multiple hypothesis testing approach using likelihood ratio statistics. Some theoretical arguments are presented showing that Chernoff-consistency (asymptotically zero type I and type II error) is maintained across all hypotheses. We apply our method on some publicly available genomics datasets and show that our method performs well in practice for its computational cost. An R package VaDA has also been made available on Github.
statistics
This activity was created within the framework of the "Space for Education" project, which ams at experiencing physical principles on the basis of topics related to space travel. Artificial satellites are suitable as application-oriented examples to explain the effect of gravity on their orbits. This working material deals with the orbit of the International Space Station (ISS) around the Earth. In simple calculations and representations, the students learn how the orbits of artificial satellites are created and which characteristic velocities occur. They compare their ISS results with those of geostationary satellites and discover applications of this particular orbit. Additional materials at: https://www.haus-der-astronomie.de/raum-fuer-bildung ----- Diese Aktivit\"at wurde im Rahmen des Projekts "Raum f\"ur Bildung" erstellt, welches physikalische Prinzipien anhand der Raumfahrt erlebbar macht. K\"unstliche Satelliten eignen sich als anwendungsnahe Beispiele, um die Wirkung der Gravitation auf ihre Bahn n\"aher zu erl\"autern. Dieses Arbeitsmaterial behandelt dazu exemplarisch den Orbit der Internationalen Raumstation (ISS) um die Erde. In einfachen Rechnungen und Darstellungen erfahren die Sch\"ulerinnen und Sch\"uler, wie Bahnen k\"unstlicher Satelliten zustande kommen und welche charakteristischen Geschwindigkeiten dabei auftreten. Sie vergleichen ihre Ergebnisse zur ISS mit denen von geostation\"aren Satelliten und entdecken Anwendungen dieses besonderen Orbits. Weitere Materialien unter: https://www.haus-der-astronomie.de/raum-fuer-bildung
physics
The $J$-equation proposed by Donaldson is a complex Hessian quotient equation on K\"ahler manifolds. The solvability of the $J$-equation is proved by Song-Weinkove to be equivalent to the existence of a subsolution. It is also conjectured by Lejmi-Szekelyhidi to be equivalent to a stability condition in terms of holomorphic intersection numbers as an analogue of the Nakai-Moishezon criterion in algebraic geometry. The conjecture is recently proved by Chen under a stronger uniform stability condition. In this paper, we establish a Nakai-Moishezon type criterion for pairs of K\"ahler classes on analytic K\"ahler varieties. As a consequence, we prove Lejmi-Szekelyhidi's original conjecture for the $J$-equation. We also apply such a criterion to obtain a family of constant scalar curvature K\"ahler metrics on smooth minimal models.
mathematics
To realize fault-tolerant quantum computing, it is necessary to store quantum information in logical qubits with error correction functions, realized by distributing a logical state among multiple physical qubits or by encoding it in the Hilbert space of a high-dimensional system. Quantum gate operations between these error-correctable logical qubits, which are essential for implementation of any practical quantum computational task, have not been experimentally demonstrated yet. Here we demonstrate a geometric method for realizing controlled-phase gates between two logical qubits encoded in photonic fields stored in cavities. The gates are realized by dispersively coupling an ancillary superconducting qubit to these cavities and driving it to make a cyclic evolution depending on the joint photonic state of the cavities, which produces a conditional geometric phase. We first realize phase gates for photonic qubits with the logical basis states encoded in two quasiorthogonal coherent states, which have important implications for continuous-variable-based quantum computation. Then we use this geometric method to implement a controlled-phase gate between two binomially encoded logical qubits, which have an error-correctable function.
quantum physics
The proton spin crisis still remains an unsolved problem in particle physics. It is suggested in the literature that the gauge invariant spin and orbital angular momentum of the quark and the gauge invariant total angular momentum of the gluon contribute to the proton spin. However, in this paper we find a new contribution to the proton spin. This new contribution is due to the boundary surface term in the total angular momentum conservation equation in QCD which is non-zero due to the confinement of the quarks and gluons inside the finite size proton. We derive the non-perturbative formula of this new contribution to the proton spin which can be calculated by using the lattice QCD method.
high energy physics phenomenology
It has been proven by Serre, Larsen-Pink and Chin, that over a smooth curve over a finite field, the monodromy groups of compatible semi-simple pure lisse sheaves have "the same" $\pi_0$ and neutral component. We generalize their results to compatible systems of semi-simple lisse sheaves and overconvergent $F$-isocrystals over arbitrary smooth varieties. For this purpose, we extend the theorem of Serre and Chin on Frobenius tori to overconvergent $F$-isocrystals. To put our results into perspective, we briefly survey recent developments of the theory of lisse sheaves and overconvergent $F$-isocrystals. We use the Tannakian formalism to make explicit the similarities between the two types of coefficient objects.
mathematics
In this paper, we obtain a version of Ekeland's variational principle for interval-value functions by means of the Dancs-Hegedus-Medvegyev theorem [14]. We also derive two versions of Ekeland's variational principle involving the generalized Hukuhara Gateaux differentiability of interval-valued functions as well as a version of Ekeland's variational principle for interval-valued bifunctions. Finally, we apply these new versions of Ekeland's variational principle to fixed point theorems, to interval-valued optimization problems, to the interval-valued Mountain Pass Theorem, to noncooperative interval-valued games, and to interval-valued optimal control problems described by interval-valued differential equations.
mathematics
We report high steady-state nuclear polarization of 1 torr $^3$He gas nuclei via metastability-exchange optical pumping at magnetic fields above 2 T. The introduction of highly polarized $^3$He gas into Brookhaven's Electron Beam Ion Source would enable a new, polarized $^3$He ion source for use at the Relativistic Heavy Ion Collider and a future Electron-Ion Collider facility. By adapting recent developments in high field metastability-exchange optical pumping for higher pressure gas, we have successfully polarized 1 torr $^3$He sealed cells in the EBIS solenoid. Through careful manipulation of the RF discharge parameters, polarizations above 80% were attained at 2, 3 and 4 T, with 89% being reached at 3 T with a 664 s relaxation time.
physics
Metal sulfides are emerging as an important class of materials for photocatalytic applications, because of their high photo responsive nature in the wide visible light range. CdS in this class of materials, have a direct band gap of 2.4 eV, have gained special attention due to the relative position of its conduction band minimum, which is very close to the energies of the reduced protons. However, the photogenerated holes in the valence band of CdS are prone to oxidation and destroy its structure during photocatalysis. Thus constructing a CdS based heterostructure would be an effective strategy for improving the photocatalytic performance. In this work we have done a detail theoretical investigation based on hybrid density functional theory calculation to get insight into the energy band structure, mobility and charge transfer across the CdS/CdSe heterojunction. The results indicate that CdS/CdSe forms type-II heterostructure that has several advantages in improving the photocatalytic efficiency under visible light irradiation.
condensed matter
In a consistent top-down approach based on orbifold compactifications, modular and traditional flavor symmetries combine nontrivially to the so-called eclectic flavor symmetry. We extend this scheme from two extra dimensions, discussed previously, to the six extra dimensions of string theory. By doing so, new insights on the nature of CP and its spontaneous breaking emerge. Moreover, we identify a new interpretation of R-symmetries as unbroken remnants from modular symmetries that are associated with geometrically stabilized complex structure moduli. Hence, all symmetries (i.e. modular, traditional flavor, CP and R) share a common origin in string theory: on a technical level, they are given by outer automorphisms of the Narain space group. The eclectic top-down approach leads to a very restrictive scheme with high predictive power. It remains a challenge to connect this with existing bottom-up constructions of modular flavor symmetry.
high energy physics theory
In this paper we consider a shape optimization problem for the minimization of the erosion, that is caused by the impact of inert particles onto the walls of a bended pipe. Using the continuous adjoint approach, we formally compute the shape derivative of the optimization problem, which is based on a one-way coupled, fully Eulerian description of a monodisperse particle jet, that is transported in a carrier fluid. We validate our approach by numerically optimizing a three-dimensional pipe segment with respect to a single particle species using a gradient descent method, and show, that the erosion rates on the optimized geometry are reduced with respect to the initial bend for a broader range of particle Stokes numbers.
mathematics
This regularly updated survey provides an overview of public resources that offer medical images and metadata of COVID-19 cases. The purpose of this survey is to simplify the access to open COVID-19 image data resources for all scientists currently working on the coronavirus crisis.
electrical engineering and systems science
Anisotropic outgassing from comets exerts a torque sufficient to rapidly change the angular momentum of the nucleus, potentially leading to rotational instability. Here, we use empirical measures of spin changes in a sample of comets to characterize the torques and to compare them with expectations from a simple model. Both the data and the model show that the characteristic spin-up timescale, $\tau_s$, is a strong function of nucleus radius, $r_n$. Empirically, we find that the timescale for comets (most with perihelion 1 to 2 AU and eccentricity $\sim$0.5) varies as $\tau_s \sim 100 r_n^{2}$, where $r_n$ is expressed in kilometers and $\tau_s$ is in years. The fraction of the nucleus surface that is active varies as $f_A \sim 0.1 r_n^{-2}$. We find that the median value of the dimensionless moment arm of the torque is $k_T$ = 0.007 (i.e. $\sim$0.7\% of the escaping momentum torques the nucleus), with weak ($<$3$\sigma$) evidence for a size dependence $k_T \sim 10^{-3} r_n^2$. Sub-kilometer nuclei have spin-up timescales comparable to their orbital periods, confirming that outgassing torques are quickly capable of driving small nuclei towards rotational disruption. Torque-induced rotational instability likely accounts for the paucity of sub-kilometer short-period cometary nuclei, and for the pre-perihelion destruction of sungrazing comets. Torques from sustained outgassing on small active asteroids can rival YORP torques, even for very small ($\lesssim$1 g s$^{-1}$) mass loss rates. Finally, we highlight the important role played by observational biases in the measured distributions of $\tau_s$, $f_A$ and $k_T$.
astrophysics
This thesis is devoted to the study of the deformation and rigidity of infinite dimensional Lie algebras which are not subject to the Hochschild-Serre factorization theorem. In particular, we consider $bms_{3}$, Virasoro-Kac-Moody and $bms_{4}$ algebras and their central extensions which are respectively obtained as asymptotic and/or near horizon symmetry algebras for Einstein gravity on $3d$ flat, AdS$_{3}$ and $4d$ flat spacetimes. We also explore possible deformations of the Maxwell-BMS algebra, which is obtained as asymptotic symmetry algebra of the Chern-Simons gravity theory invariant under the $2+1$ dimensional Maxwell algebra. We find that these algebras are not rigid and can be deformed into new non isomorphic infinite dimensional (family of) algebras. We study these deformations by direct computations and also by cohomological analysis. We then classify all the algebras obtained through deformation of these algebras as well as all possible central extensions thereof. We propose/conjecture an extension of the Hochschild-Serre factorization theorem for infinite dimensional algebras as well as introducing a new notion of rigidity for family algebras obtained through deformation. We also explore physical realizations and significance of the family of algebras we obtain through the deformation procedure.
high energy physics theory
Combined use of PET and dual-energy CT provides complementary information for multi-parametric imaging. PETenabled dual-energy CT combines a low-energy x-ray CT image with a high-energy &\gamma&-ray CT (GCT) image reconstructed from time-of-flight PET emission data to enable dual-energy CT material decomposition on a PET/CT scanner. The maximumlikelihood attenuation and activity (MLAA) algorithm has been used for GCT reconstruction but suffers from noise. Kernel MLAA exploits an x-ray CT image prior through the kernel framework to guide GCT reconstruction and has demonstrated substantial improvements in noise suppression. However, similar to other kernel methods for image reconstruction, the existing kernel MLAA uses image intensity-based features to construct the kernel representation, which is not always robust and may lead to suboptimal reconstruction with artifacts. In this paper, we propose a modified kernel method by using an autoencoder convolutional neural network (CNN) to extract an intrinsic feature set from the x-ray CT image prior. A computer simulation study was conducted to compare the autoencoder CNN-derived feature representation with raw image patches for evaluation of kernel MLAA for GCT image reconstruction and dual-energy multimaterial decomposition. The results show that the autoencoder kernel MLAA method can achieve a significant image quality improvement for GCT and material decomposition as compared to the existing kernel MLAA algorithm. A weakness of the proposed method is its potential over-smoothness in a bone region, indicating the importance of further optimization in future work. The codes is available on https://github.com/SiqiLi1020/Autoencoder- Kernel-MLAA.
physics
The clustering property of an equilibrium bipartite correlation is one of the most general thermodynamic properties in non-critical many-body quantum systems. Herein, we consider the thermalization properties of a system class exhibiting the clustering property. We investigate two regimes, namely, regimes of high and low density of states corresponding to high and low energy regimes, respectively. We show that the clustering property is connected to several properties on the eigenstate thermalization through the density of states. Remarkably, the eigenstate thermalization is obtained in the low-energy regime with sparse density of states, which is typically seen in gapped systems. For the high-energy regime, we demonstrate the ensemble equivalence between microcanonical and canonical ensembles even for subexponentially small energy shell with respect to the system size, which eventually leads to the weak version of eigenstate thermalization.
condensed matter
We define a new variant of exclusion processes in discrete time that has jump probabilities that depend on the last jump performed. In a particular limit for the jump probabilities and in suitable scaling limits for space and time, we compute the time evolution of the particle density starting from an arbitrary initial configuration, with closed boundary conditions. The core of the argument is the analysis of the time evolution of the moments. Numerical results are compared with the prediction and give excellent agreement.
condensed matter
We study the size of the largest rectangle containing no point of a given point set in the two-dimensional torus, the dispersion of the point set. A known lower bound for the dispersion of any point set of cardinality $n\ge 2$ in this setting is $2/n$. We show that if $n$ is a Fibonacci number then the Fibonacci lattice has dispersion exactly $2/n$ meeting the lower bound. Moreover, we completely characterize integration lattices achieving the lower bound and provide insight into the structure of other optimal sets. We also treat related results in the nonperiodic setting.
mathematics
An inevitable consequence of the global power system transition towards nearly 100% renewable-based generation is the loss of conventional bulk generation by synchronous machines, their inertia, and accompanying frequency and voltage control mechanisms. This gradual transformation of the power system to a low-inertia system leads to critical challenges in maintaining system stability. Novel control techniques for converters, so-called grid-forming strategies, are expected to address these challenges and replicate functionalities that so far have been provided by synchronous machines. This article presents a low-inertia case study that includes synchronous machines and converters controlled under various grid-forming techniques. In this work 1) the positive impact of the grid-forming converters on the frequency stability of synchronous machines is highlighted, 2) a qualitative analysis which provides insights into the frequency stability of the system is presented, 3) we explore the behavior of the grid-forming controls when imposing the converter dc and ac current limitations, 4) the importance of the dc dynamics in grid-forming control design as well as the critical need for an effective ac current limitation scheme are reported, and lastly 5) we analyze how and when the interaction between the fast grid-forming converter and the slow synchronous machine dynamics can contribute to the system instability
electrical engineering and systems science
EOSIO is one typical public blockchain platform. It is scalable in terms of transaction speeds and has a growing ecosystem supporting smart contracts and decentralized applications. However, the vulnerabilities within the EOSIO smart contracts have led to serious attacks, which caused serious financial loss to its end users. In this work, we systematically analyzed three typical EOSIO smart contract vulnerabilities and their related attacks. Then we presented EOSFuzzer, a general black-box fuzzing framework to detect vulnerabilities within EOSIO smart contracts. In particular, EOSFuzzer proposed effective attacking scenarios and test oracles for EOSIO smart contract fuzzing. Our fuzzing experiment on 3963 EOSIO smart contracts shows that EOSFuzzer is both effective and efficient to detect EOSIO smart contract vulnerabilities with high accuracy.
computer science
We report the results from \textit{AstroSat} observations of the transient Galactic black hole X-ray binary MAXI J1535-571 during its hard-intermediate state of the 2017 outburst. We systematically study the individual and joint spectra from two simultaneously observing \textit{AstroSat} X-ray instruments, and probe and measure a number of parameter values of accretion disc, corona and reflection from the disc in the system using models with generally increasing complexities. Using our broadband ($1.3-70$ keV) X-ray spectrum, we clearly show that a soft X-ray instrument, which works below $\sim 10-12$ keV, alone cannot correctly characterize the Comptonizing component from the corona, thus highlighting the importance of broadband spectral analysis. By fitting the reflection spectrum with the latest version of the \textsc{relxill} family of relativistic reflection models, we constrain the black hole's dimensionless spin parameter to be $0.67^{+0.16}_{-0.04}$. We also jointly use the reflection spectral component (\textsc{relxill}) and a general relativistic thin disc component (\texttt{Kerrbb}), and estimate the black hole's mass and distance to be $10.39_{-0.62}^{+0.61} M_{\odot}$ and $5.4_{-1.1}^{+1.8}$ kpc respectively.
astrophysics
We define a three-dimensional quantum theory of gravity as the holographic dual of the Liouville conformal field theory. The theory is consistent and unitary by definition. The corresponding theory of gravity with negative cosmological constant has peculiar properties. The quantum theory has no normalisable AdS3 vacuum. The model contains primary black holes with zero spin. All states can be interpreted as black holes dressed with boundary gravitons. There is a unique universal interaction between these states consistent with unitarity and the conformal symmetry of the model. This theory of gravity, though conceptually isolated from other models of quantum gravity, is worth scrutinising.
high energy physics theory
Quantization of neural networks has become common practice, driven by the need for efficient implementations of deep neural networks on embedded devices. In this paper, we exploit an oft-overlooked degree of freedom in most networks - for a given layer, individual output channels can be scaled by any factor provided that the corresponding weights of the next layer are inversely scaled. Therefore, a given network has many factorizations which change the weights of the network without changing its function. We present a conceptually simple and easy to implement method that uses this property and show that proper factorizations significantly decrease the degradation caused by quantization. We show improvement on a wide variety of networks and achieve state-of-the-art degradation results for MobileNets. While our focus is on quantization, this type of factorization is applicable to other domains such as network-pruning, neural nets regularization and network interpretability.
computer science
We characterize the mechanisms of vortex pinning in a superfluid thin film described by the two-dimensional Gross-Pitaevskii equation. We consider a vortex "scattering experiment" whereby a single vortex in a superfluid flow interacts with a circular pinning potential. By an analogy with linear dielectrics, we develop an analytical hydrodynamic approximation that predicts vortex trajectories, the vortex fixed point, and the unpinning superfluid velocity beyond which the vortex cannot be trapped. We then solve the Gross-Pitaevskii equation to validate this model, and build a phase portrait of vortex pinning. We identify two different dynamical pinning mechanisms, marked by distinctive phonon emission signatures: firstly a "fall-on" regime enabled by acoustic radiation, and secondly a "pair-creation" regime, mediated by vortex dipoles nucleated within the pin. Pinning potentials with a size on the order of the healing length are found to be optimal for vortex capture. Our results will be useful in mitigating the deleterious effects of drag due to vortices in superfluid channels, in analogy to maximising supercurrents in type-II superconductors.
condensed matter
We make the first steps towards generalizing the theory of stochastic block models, in the sparse regime, towards a model where the discrete community structure is replaced by an underlying geometry. We consider a geometric random graph over a homogeneous metric space where the probability of two vertices to be connected is an arbitrary function of the distance. We give sufficient conditions under which the locations can be recovered (up to an isomorphism of the space) in the sparse regime. Moreover, we define a geometric counterpart of the model of flow of information on trees, due to Mossel and Peres, in which one considers a branching random walk on a sphere and the goal is to recover the location of the root based on the locations of leaves. We give some sufficient conditions for percolation and for non-percolation of information in this model.
statistics
Numerous advantages may accrue to ET from the use of probes for the conveyance of information relative to the alternative of beaming information electromagnetically from its home star system. However, probes would benefit from a trans-galactic communications network to transmit information back to their progenitor civilization(s) and throughout the Galaxy. A model for such trans-galactic communications has previously been proposed, comprised of probes linked together by relay nodes. The current paper refines that model, specifying two types of probes or nodes, both part of a communications lattice. The first type of cellular probe (CP) is statically associated with individual stars, and the second type is a scout CP that traverses the galaxy surveying successive star systems rather than permanently residing within any one of them. CPs would communicate with adjacent CPs and, ultimately, through the network of CPs, with their progenitor civilization(s). The question of whether Oumuamua is such a flyby scout CP is considered. Search strategies for CPs are suggested.
physics
We describe and validate a novel data-driven approach to the real time detection and classification of traffic anomalies based on the identification of atypical fluctuations in the relationship between density and flow. For aggregated data under stationary conditions, flow and density are related by the fundamental diagram. However, high resolution data obtained from modern sensor networks is generally non-stationary and disaggregated. Such data consequently show significant statistical fluctuations. These fluctuations are best described using a bivariate probability distribution in the density-flow plane. By applying kernel density estimation to high-volume data from the UK National Traffic Information Service (NTIS), we empirically construct these distributions for London's M25 motorway. Curves in the density-flow plane are then constructed, analogous to quantiles of univariate distributions. These curves quantitatively separate atypical fluctuations from typical traffic states. Although the algorithm identifies anomalies in general rather than specific events, we find that fluctuations outside the 95\% probability curve correlate strongly with the spikes in travel time associated with significant congestion events. Moreover, the size of an excursion from the typical region provides a simple, real-time measure of the severity of detected anomalies. We validate the algorithm by benchmarking its ability to identify labelled events in historical NTIS data against some commonly used methods from the literature. Detection rate, time-to-detect and false alarm rate are used as metrics and found to be generally comparable except in situations when the speed distribution is bi-modal. In such situations, the new algorithm achieves a much lower false alarm rate without suffering significant degradation on the other metrics. This method has the additional advantage of being self-calibrating.
statistics
Consider a general machine learning setting where the output is a set of labels or sequences. This output set is unordered and its size varies with the input. Whereas multi-label classification methods seem a natural first resort, they are not readily applicable to set-valued outputs because of the growth rate of the output space; and because conventional sequence generation doesn't reflect sets' order-free nature. In this paper, we propose a unified framework--sequential set generation (SSG)--that can handle output sets of labels and sequences. SSG is a meta-algorithm that leverages any probabilistic learning method for label or sequence prediction, but employs a proper regularization such that a new label or sequence is generated repeatedly until the full set is produced. Though SSG is sequential in nature, it does not penalize the ordering of the appearance of the set elements and can be applied to a variety of set output problems, such as a set of classification labels or sequences. We perform experiments with both benchmark and synthetic data sets and demonstrate SSG's strong performance over baseline methods.
computer science
We introduce the Action Transformer model for recognizing and localizing human actions in video clips. We repurpose a Transformer-style architecture to aggregate features from the spatiotemporal context around the person whose actions we are trying to classify. We show that by using high-resolution, person-specific, class-agnostic queries, the model spontaneously learns to track individual people and to pick up on semantic context from the actions of others. Additionally its attention mechanism learns to emphasize hands and faces, which are often crucial to discriminate an action - all without explicit supervision other than boxes and class labels. We train and test our Action Transformer network on the Atomic Visual Actions (AVA) dataset, outperforming the state-of-the-art by a significant margin using only raw RGB frames as input.
computer science
The use of flat or weakly informative priors is popular due to the objective a priori belief in the absence of strong prior information. In the case of the Weibull model the improper uniform, equal parameter gamma and joint Jeffrey's priors for the shape parameter are popular choices. The effects and behaviors of these priors have yet to be established from a modeling viewpoint, especially their ability to reduce to the simpler exponential model. In this work we propose a new principled prior for the shape parameter of the Weibull model, originating from a prior on the distance function, and advocate this new prior as a principled choice in the absence of strong prior information. This new prior can then be used in models with a Weibull modeling component, like competing risks, joint and spatial models, to mention a few. This prior is available in the R-INLA for use, and is applied in a joint longitudinal-survival model framework using the INLA method.
statistics
This study develops a federated learning (FL) framework overcoming largely incremental communication costs due to model sizes in typical frameworks without compromising model performance. To this end, based on the idea of leveraging an unlabeled open dataset, we propose a distillation-based semi-supervised FL (DS-FL) algorithm that exchanges the outputs of local models among mobile devices, instead of model parameter exchange employed by the typical frameworks. In DS-FL, the communication cost depends only on the output dimensions of the models and does not scale up according to the model size. The exchanged model outputs are used to label each sample of the open dataset, which creates an additionally labeled dataset. Based on the new dataset, local models are further trained, and model performance is enhanced owing to the data augmentation effect. We further highlight that in DS-FL, the heterogeneity of the devices' dataset leads to ambiguous of each data sample and lowing of the training convergence. To prevent this, we propose entropy reduction averaging, where the aggregated model outputs are intentionally sharpened. Moreover, extensive experiments show that DS-FL reduces communication costs up to 99% relative to those of the FL benchmark while achieving similar or higher classification accuracy.
computer science