Project description:Complex networks abound in physical, biological and social sciences. Quantifying a network's topological structure facilitates network exploration and analysis, and network comparison, clustering and classification. A number of Wiener type indices have recently been incorporated as distance-based descriptors of complex networks, such as the R package QuACN. Wiener type indices are known to depend both on the network's number of nodes and topology. To apply these indices to measure similarity of networks of different numbers of nodes, normalization of these indices is needed to correct the effect of the number of nodes in a network. This paper aims to fill this gap. Moreover, we introduce an f-Wiener index of network G, denoted by Wf(G). This notion generalizes the Wiener index to a very wide class of Wiener type indices including all known Wiener type indices. We identify the maximum and minimum of Wf(G) over a set of networks with n nodes. We then introduce our normalized-version of f-Wiener index. The normalized f-Wiener indices were demonstrated, in a number of experiments, to improve significantly the hierarchical clustering over the non-normalized counterparts.
Project description:It is often of interest to decompose the total effect of an exposure into a component that acts on the outcome through some mediator and a component that acts independently through other pathways. Said another way, we are interested in the direct and indirect effects of the exposure on the outcome. Even if the exposure is randomly assigned, it is often infeasible to randomize the mediator, leaving the mediator-outcome confounding not fully controlled. We develop a sensitivity analysis technique that can bound the direct and indirect effects without parametric assumptions about the unmeasured mediator-outcome confounding.
Project description:BackgroundSufficient-cause interaction is a type of interaction that has received much attention recently. The sufficient component cause model on which the sufficient-cause interaction is based is however a non-identifiable model. Estimating the interaction parameters from the model is mathematically impossible.MethodsIn this paper, I derive bounding formulae for sufficient-cause interactions under the assumption of no redundancy.ResultsTwo real data sets are used to demonstrate the method (R codes provided). The proposed bounds are sharp and sharper than previous bounds.ConclusionsSufficient-cause interactions can be quantified by setting bounds on them.
Project description:Many monostable reaction-diffusion equations admit one-dimensional travelling waves if and only if the wave speed is sufficiently high. The values of these minimum wave speeds are not known exactly, except in a few simple cases. We present methods for finding upper and lower bounds on minimum wave speed. They rely on constructing trapping boundaries for dynamical systems whose heteroclinic connections correspond to the travelling waves. Simple versions of this approach can be carried out analytically but often give overly conservative bounds on minimum wave speed. When the reaction-diffusion equations being studied have polynomial nonlinearities, our approach can be implemented computationally using polynomial optimization. For scalar reaction-diffusion equations, we present a general method and then apply it to examples from the literature where minimum wave speeds were unknown. The extension of our approach to multi-component reaction-diffusion systems is then illustrated using a cubic autocatalysis model from the literature. In all three examples and with many different parameter values, polynomial optimization computations give upper and lower bounds that are within 0.1% of each other and thus nearly sharp. Upper bounds are derived analytically as well for the scalar reaction-diffusion equations.
Project description:Complex networks are ubiquitous in biological, physical and social sciences. Network robustness research aims at finding a measure to quantify network robustness. A number of Wiener type indices have recently been incorporated as distance-based descriptors of complex networks. Wiener type indices are known to depend both on the network's number of nodes and topology. The Wiener polarity index is also related to the cluster coefficient of networks. In this paper, based on some graph transformations, we determine the sharp upper bound of the Wiener polarity index among all bicyclic networks. These bounds help to understand the underlying quantitative graph measures in depth.
Project description:The geometry of polynomials explores geometrical relationships between the zeros and the coefficients of a polynomial. A classical problem in this theory is to locate the zeros of a given polynomial by determining disks in the complex plane in which all its zeros are situated. In this paper, we infer bounds for general polynomials and apply classical and new results to graph polynomials namely Wiener and distance polynomials whose zeros have not been yet investigated. Also, we examine the quality of such bounds by considering four graph classes and interpret the results.
Project description:Uncertain data are observations that cannot be uniquely mapped to a referent. In the case of uncertainty due to incompleteness, possibility theory can be used as an appropriate model for processing such data. In particular, granular counting is a way to count data in presence of uncertainty represented by possibility distributions. Two algorithms were proposed in literature to compute granular counting: exact granular counting, with quadratic time complexity, and approximate granular counting, with linear time complexity. This paper extends approximate granular counting by computing bounds for exact granular count. In this way, the efficiency of approximate granular count is combined with certified bounds whose width can be adjusted in accordance to user needs.
Project description:In order to understand the identity of the Central American species of the genus Phaenonotum Sharp, 1882, the type specimens of the species described by Sharp (1882) deposited in the David Sharp collection in the Natural History Museum in London have been re-examined. The following species are redescribed: Phaenonotum apicale Sharp, 1882, Phaenonotum collare Sharp, 1882, Phaenonotum dubium Sharp, 1882 (confirmed as junior synonym of Phaenonotum exstriatum (Say, 1835)), Phaenonotum laevicolle Sharp, 1882, Phaenonotum rotundulum Sharp, 1882 and Phaenonotum tarsale Sharp, 1882. Lectotypes are designated for Phaenonotum apicale, Phaenonotum collare, Phaenonotum rotundulum and Phaenonotum tarsale. External diagnostic characters and morphology of male genitalia are illustrated. A table summarizing diagnostic characters allowing the identification of the species is provided.
Project description:In recent years, mass spectrometry-based metabolomics has increasingly been applied to large-scale epidemiological studies of human subjects. However, the successful use of metabolomics in this context is subject to the challenge of detecting biologically significant effects despite substantial intensity drift that often occurs when data are acquired over a long period or in multiple batches. Numerous computational strategies and software tools have been developed to aid in correcting for intensity drift in metabolomics data, but most of these techniques are implemented using command-line driven software and custom scripts which are not accessible to all end users of metabolomics data. Further, it has not yet become routine practice to assess the quantitative accuracy of drift correction against techniques which enable true absolute quantitation such as isotope dilution mass spectrometry. We developed an Excel-based tool, MetaboDrift, to visually evaluate and correct for intensity drift in a multi-batch liquid chromatography - mass spectrometry (LC-MS) metabolomics dataset. The tool enables drift correction based on either quality control (QC) samples analyzed throughout the batches or using QC-sample independent methods. We applied MetaboDrift to an original set of clinical metabolomics data from a mixed-meal tolerance test (MMTT). The performance of the method was evaluated for multiple classes of metabolites by comparison with normalization using isotope-labeled internal standards. QC sample-based intensity drift correction significantly improved correlation with IS-normalized data, and resulted in detection of additional metabolites with significant physiological response to the MMTT. The relative merits of different QC-sample curve fitting strategies are discussed in the context of batch size and drift pattern complexity. Our drift correction tool offers a practical, simplified approach to drift correction and batch combination in large metabolomics studies.
Project description:The ability to intervene in disease progression given a person's disease history has the potential to solve one of society's most pressing issues: advancing health care delivery and reducing its cost. Controlling disease progression is inherently associated with the ability to predict possible future diseases given a patient's medical history. We invoke an information-theoretic methodology to quantify the level of predictability inherent in disease histories of a large electronic health records dataset with over half a million patients. In our analysis, we progress from zeroth order through temporal informed statistics, both from an individual patient's standpoint and also considering the collective effects. Our findings confirm our intuition that knowledge of common disease progressions results in higher predictability bounds than treating disease histories independently. We complement this result by showing the point at which the temporal dependence structure vanishes with increasing orders of the time-correlated statistic. Surprisingly, we also show that shuffling individual disease histories only marginally degrades the predictability bounds. This apparent contradiction with respect to the importance of time-ordered information is indicative of the complexities involved in capturing the health-care process and the difficulties associated with utilising this information in universal prediction algorithms.