Principled statistical methods for high-dimensional correlation networks. This project aims to develop a novel and principled approach for building correlation networks. Correlation networks aim to identify the most significant associations present in modern massive datasets, and have numerous applications, ranging from the biomedical and environmental sciences to the social sciences. Nodes of such networks represent features, and edges represent associations, or the lack thereof. Current method ....Principled statistical methods for high-dimensional correlation networks. This project aims to develop a novel and principled approach for building correlation networks. Correlation networks aim to identify the most significant associations present in modern massive datasets, and have numerous applications, ranging from the biomedical and environmental sciences to the social sciences. Nodes of such networks represent features, and edges represent associations, or the lack thereof. Current methods are not readily scalable to modern ultra-high dimensional settings, and do not account for uncertainty in the estimated associations. This project will develop a principled, highly scalable methodology for building such networks, which incorporates uncertainty quantification. Emphasis is placed on modern ultra-high dimensional settings in which differentiating a true correlation from a spurious one is a notoriously difficult task.Read moreRead less
Discovery Early Career Researcher Award - Grant ID: DE180101252
Funder
Australian Research Council
Funding Amount
$343,450.00
Summary
Statistical theory and algorithms for joint inference of complex networks. This project aims to address the challenges in jointly modelling complex networks by applying an integrated approach encompassing statistical theory, computation, and applications. The project expects to contribute to core statistical methodology development for complex inference and generate new knowledge in the fields of genomics, neuroscience, and social science through in-depth analyses of large-scale multilayered net ....Statistical theory and algorithms for joint inference of complex networks. This project aims to address the challenges in jointly modelling complex networks by applying an integrated approach encompassing statistical theory, computation, and applications. The project expects to contribute to core statistical methodology development for complex inference and generate new knowledge in the fields of genomics, neuroscience, and social science through in-depth analyses of large-scale multilayered network data. Expected outcomes include enhanced theoretical and computational frameworks for probabilistic network models to better utilise the power of multiple observations. This should foster international and interdisciplinary collaborations and add significant value to the rapidly progressing field of networks research.Read moreRead less
Inference for Hawkes processes with challenging data. The Hawkes processes are statistical models for the analysis of high-impact event sequences, such as bushfires, earthquakes, infectious diseases, and cyber attacks. When the times and/or marks are missing for some events or when the data is otherwise incomplete, it is challenging to fit these models and perform diagnostic checks on the fitted models. This project aims to develop novel statistical methods to fit these models in the presence of ....Inference for Hawkes processes with challenging data. The Hawkes processes are statistical models for the analysis of high-impact event sequences, such as bushfires, earthquakes, infectious diseases, and cyber attacks. When the times and/or marks are missing for some events or when the data is otherwise incomplete, it is challenging to fit these models and perform diagnostic checks on the fitted models. This project aims to develop novel statistical methods to fit these models in the presence of incomplete data and to check the goodness-of-fit of the fitted models. The expected outcomes include publications documenting these methods and software packages implementing them. The primary benefits include the advancement of statistical methodology and the training of junior research personnel. Read moreRead less
Fast flexible feature selection for high dimensional challenging data. The project aims to provide new frameworks for fast flexible feature selection and appropriate modelling of heterogeneous data through structural varying-coefficient regression models. The outcomes will be a series of new statistical methods and concepts enabling more powerful modelling of complex bioscience data. The project will create the science for building reliable statistical models taking model uncertainty into accoun ....Fast flexible feature selection for high dimensional challenging data. The project aims to provide new frameworks for fast flexible feature selection and appropriate modelling of heterogeneous data through structural varying-coefficient regression models. The outcomes will be a series of new statistical methods and concepts enabling more powerful modelling of complex bioscience data. The project will create the science for building reliable statistical models taking model uncertainty into account, impacting how results will be interpreted, and with accompanying software. This will be a significant improvement in the assessment of model confidence in the food and health research priority areas including areas such as meat science, Huntington’s disease, and kidney transplantation.Read moreRead less
Fast approximate inference methods: new algorithms, applications and theory. This project aims to develop new algorithms and theory for fast approximate inference and lay down infrastructure to aid future extensions. Fast approximate inference methods are a principled and extensible means of fitting large and complex statistical models to big data sets. They come into their own in applications where speed is paramount and traditional approaches are not feasible. The project aims to lead to prac ....Fast approximate inference methods: new algorithms, applications and theory. This project aims to develop new algorithms and theory for fast approximate inference and lay down infrastructure to aid future extensions. Fast approximate inference methods are a principled and extensible means of fitting large and complex statistical models to big data sets. They come into their own in applications where speed is paramount and traditional approaches are not feasible. The project aims to lead to practical outcomes from better business decision-making for insurance data warehouses, to improved medical imaging technology.Read moreRead less
New methods for modelling real-world extremes. This project aims to develop new theory and methods for analysing and predicting extreme values observed in real-world processes. Many existing techniques are limited by convenient mathematical assumptions that commonly do not hold in practice: dependence at asymptotic levels, process stationarity, and that the observed data are direct measurements of the process of interest. As a result, using these techniques may produce undesirable results. Expec ....New methods for modelling real-world extremes. This project aims to develop new theory and methods for analysing and predicting extreme values observed in real-world processes. Many existing techniques are limited by convenient mathematical assumptions that commonly do not hold in practice: dependence at asymptotic levels, process stationarity, and that the observed data are direct measurements of the process of interest. As a result, using these techniques may produce undesirable results. Expected outcomes of this project include theoretically justified data analysis techniques that can accurately model extreme values seen in the real world. Project benefits include more realistic analyses of nationally important applications in climate, bushfire insurance risk, and anomaly detection.Read moreRead less
Feature Learning for High-dimensional Functional Time Series. This project aims to develop new methods and theories for common features on high-dimensional functional time series observed in empirical applications. The significance includes addressing a key gap in adaptive and efficient feature learning, improving forecasting accuracy and understanding forecasting-driven factors comprehensively for empirical data. Expected outcomes involve advances in big data theory and easy-to-implement algori ....Feature Learning for High-dimensional Functional Time Series. This project aims to develop new methods and theories for common features on high-dimensional functional time series observed in empirical applications. The significance includes addressing a key gap in adaptive and efficient feature learning, improving forecasting accuracy and understanding forecasting-driven factors comprehensively for empirical data. Expected outcomes involve advances in big data theory and easy-to-implement algorithms for applied researchers. This project benefits not only advanced manufacturing by finding optimal stopping time for wood panel compression, but also superior forecasting for mortality in demography, climate data in environmental science, asset returns in finance, and electricity consumption in economics. Read moreRead less
Reliable and accurate statistical solutions for modern complex data. This project aims to develop novel methods for reliable and accurate statistical modelling with modern, complex correlated and error-prone data. The project expects to make significant strides towards future-proofing statistical data analysis, equipping practitioners with a suite of robust and computationally efficient methods which provide confidence in the stability and reproducibility of results obtained, while offering guar ....Reliable and accurate statistical solutions for modern complex data. This project aims to develop novel methods for reliable and accurate statistical modelling with modern, complex correlated and error-prone data. The project expects to make significant strides towards future-proofing statistical data analysis, equipping practitioners with a suite of robust and computationally efficient methods which provide confidence in the stability and reproducibility of results obtained, while offering guarantees on their transferability over a range of populations. This will provide important benefits as they are applied in predicting endangered marine species for fisheries conservation, and in enhancing our national understanding of the relationship between education achievement and financial success. Read moreRead less
Technology-Driven and Scalable Regression Methodology, Computing and Theory. Regression is a mainstay of data analysis, statistics, machine learning and data science but is in continual need of enhancement in the face of technological change. Scalability and flexibility for the handling of non-linear signals are fundamental to the practical utility of new regression methodology. Several streams of research aimed at confronting data from specific technologies as well as generic types of data are ....Technology-Driven and Scalable Regression Methodology, Computing and Theory. Regression is a mainstay of data analysis, statistics, machine learning and data science but is in continual need of enhancement in the face of technological change. Scalability and flexibility for the handling of non-linear signals are fundamental to the practical utility of new regression methodology. Several streams of research aimed at confronting data from specific technologies as well as generic types of data are proposed. The project is to be networked with researchers in the United States of America and aims to have Australia-based researchers providing leadership in terms of methodological, theoretical, computational and software development.Read moreRead less
Innovative statistical methods for analysing high-dimensional counts. The aim is to develop fast, modern statistical methods for analysing high dimensional data in ecology at large scales, in particular, for visualising, classifying and predicting ecological communities. The benefit of the project is a set of multivariate tools that can be used to better understand biodiversity and its response to environmental drivers, a challenging statistical problem. The proposed methods for analysing high d ....Innovative statistical methods for analysing high-dimensional counts. The aim is to develop fast, modern statistical methods for analysing high dimensional data in ecology at large scales, in particular, for visualising, classifying and predicting ecological communities. The benefit of the project is a set of multivariate tools that can be used to better understand biodiversity and its response to environmental drivers, a challenging statistical problem. The proposed methods for analysing high dimensional data can provide insight into large scale questions in ecology, such as automated identification of biogeographic boundaries. The expected outcome is a powerful statistical toolset for model-based analysis of high dimensional data, introducing modern multivariate approaches to a high-impact area of ecology.Read moreRead less