Trans-dimensional and Approximate Bayesian Computation. Many applied scientists in Australia, particularly those in the biological, medical and environmental sciences are now interested in incorporating Bayesian statistical methodologies into their research.
The development of more generic and efficient Bayesian statistical methods will not only benefit applied statisticians but also the more occasional users of statistics in other disciplinary areas. The success of this project will enhance Au ....Trans-dimensional and Approximate Bayesian Computation. Many applied scientists in Australia, particularly those in the biological, medical and environmental sciences are now interested in incorporating Bayesian statistical methodologies into their research.
The development of more generic and efficient Bayesian statistical methods will not only benefit applied statisticians but also the more occasional users of statistics in other disciplinary areas. The success of this project will enhance Australia's reputation as a strong contributor to the development of Bayesian methodologies. Two PhD students will also be provided training in computational Bayesian statistics.Read moreRead less
New Bayesian methodology for understanding complex systems using hidden Markov models and expert opinion, environmental, robotics and genomics applications. This project aims to merge four areas of intense international interest in describing complex systems: hidden Markov models and mixtures, semi-parametric and nonparametric approaches, true combination of expert opinion with data, and new Bayesian computational methods based on perfect sampling and particle sampling. The project will signific ....New Bayesian methodology for understanding complex systems using hidden Markov models and expert opinion, environmental, robotics and genomics applications. This project aims to merge four areas of intense international interest in describing complex systems: hidden Markov models and mixtures, semi-parametric and nonparametric approaches, true combination of expert opinion with data, and new Bayesian computational methods based on perfect sampling and particle sampling. The project will significantly contribute to statistical methodology and its ability to inform about real-world problems. A strong focus on applications to genomics, robotics and environmental modelling will bring immediate research and monetary benefit for industry. Expected outcomes include enhanced cross-disciplinary and international linkages, publications, industry-funded projects and highly trained graduates.Read moreRead less
ARC Centre of Excellence for Mathematical and Statistical Frontiers of Big Data, Big Models, New Insights. In today's world, massive amounts of data in a variety of forms are collected daily from a multitude of sources. Many of the resulting data sets have the potential to make vital contributions to society, business and government, as well as impact on international developments, but are so large or complex that they are difficult to process and analyse using traditional tools. The aim of this ....ARC Centre of Excellence for Mathematical and Statistical Frontiers of Big Data, Big Models, New Insights. In today's world, massive amounts of data in a variety of forms are collected daily from a multitude of sources. Many of the resulting data sets have the potential to make vital contributions to society, business and government, as well as impact on international developments, but are so large or complex that they are difficult to process and analyse using traditional tools. The aim of this Centre is to create innovative mathematical and statistical models that can uncover the knowledge concealed within the size and complexity of these big data sets, with a focus on using the models to deliver insight into problems vital to the Centre's Collaborative Domains: Healthy People, Sustainable Environments and Prosperous Societies.Read moreRead less
Discovery Early Career Researcher Award - Grant ID: DE130101670
Funder
Australian Research Council
Funding Amount
$370,410.00
Summary
Scalable Bayesian model selection for massive data sets. This project will develop highly innovative, efficient and ultimately effective methodology for Bayesian model selection for large-scale problems which commonly arise in biostatistics and bioinformatics. The resulting methodology will dramatically reduce the duration of analyses in these areas from days or weeks to minutes or hours.
New approaches to predictive modelling of high-dimensional count data to study climate impacts on ecological communities. This project will lay methodological foundations for future studies of potential impacts of climate change on ecological communities. A flexible new toolset of predictive modelling approaches will be developed, capable of handling all common data types, which fit easy-to-interpret models, and which are more powerful than currently used methods.
Principled statistical methods for high-dimensional correlation networks. This project aims to develop a novel and principled approach for building correlation networks. Correlation networks aim to identify the most significant associations present in modern massive datasets, and have numerous applications, ranging from the biomedical and environmental sciences to the social sciences. Nodes of such networks represent features, and edges represent associations, or the lack thereof. Current method ....Principled statistical methods for high-dimensional correlation networks. This project aims to develop a novel and principled approach for building correlation networks. Correlation networks aim to identify the most significant associations present in modern massive datasets, and have numerous applications, ranging from the biomedical and environmental sciences to the social sciences. Nodes of such networks represent features, and edges represent associations, or the lack thereof. Current methods are not readily scalable to modern ultra-high dimensional settings, and do not account for uncertainty in the estimated associations. This project will develop a principled, highly scalable methodology for building such networks, which incorporates uncertainty quantification. Emphasis is placed on modern ultra-high dimensional settings in which differentiating a true correlation from a spurious one is a notoriously difficult task.Read moreRead less
Frontiers in Data Science: Analysing Distributions as Data. This project aims to develop the statistical foundations of a new approach to analysing large and complex data, based on building distributional approximations of the data, which can then be analysed by standard statistical methods. The need to analyse very large and complex datasets has become a vital part of everyday life, particularly in the analysis of national problems in public health, environmental pollution, computer network sec ....Frontiers in Data Science: Analysing Distributions as Data. This project aims to develop the statistical foundations of a new approach to analysing large and complex data, based on building distributional approximations of the data, which can then be analysed by standard statistical methods. The need to analyse very large and complex datasets has become a vital part of everyday life, particularly in the analysis of national problems in public health, environmental pollution, computer network security and climate extremes. The project expects to change our way of thinking in how to be smarter about what data we use (and collect) for analysis, rather than relying on brute force analysis of large datasets. The project is expected to transform the knowledge base of the discipline, and the resulting techniques will enable across-the-board research advances for many industries and disciplines.Read moreRead less
New methods for modelling real-world extremes. This project aims to develop new theory and methods for analysing and predicting extreme values observed in real-world processes. Many existing techniques are limited by convenient mathematical assumptions that commonly do not hold in practice: dependence at asymptotic levels, process stationarity, and that the observed data are direct measurements of the process of interest. As a result, using these techniques may produce undesirable results. Expec ....New methods for modelling real-world extremes. This project aims to develop new theory and methods for analysing and predicting extreme values observed in real-world processes. Many existing techniques are limited by convenient mathematical assumptions that commonly do not hold in practice: dependence at asymptotic levels, process stationarity, and that the observed data are direct measurements of the process of interest. As a result, using these techniques may produce undesirable results. Expected outcomes of this project include theoretically justified data analysis techniques that can accurately model extreme values seen in the real world. Project benefits include more realistic analyses of nationally important applications in climate, bushfire insurance risk, and anomaly detection.Read moreRead less
Building models for complex data. The purpose of this project is to better understand the process of building statistical models and construct new methods for building models for particular kinds of complex data. The expected outcomes include a new way of thinking about model building and practical tools which together enable us to get more value out of analysing complex data.
Semiparametric Regression for Streaming Data. Semiparametric regression converts large and complex data-sets into interpretable summaries from which sound decisions can be made. This project tackles semiparametric regression analysis of streaming data - where the data are so voluminous that they may not be storable in standard computer memory and therefore need to be processed rapidly on arrival and then discarded. Effective solutions necessitate a rethinking of semi-parametric regression and ne ....Semiparametric Regression for Streaming Data. Semiparametric regression converts large and complex data-sets into interpretable summaries from which sound decisions can be made. This project tackles semiparametric regression analysis of streaming data - where the data are so voluminous that they may not be storable in standard computer memory and therefore need to be processed rapidly on arrival and then discarded. Effective solutions necessitate a rethinking of semi-parametric regression and new approaches will be developed. The project will also develop novel theory and methodology for robotics applications. It will allow analysis of streaming and massive data sets that would not be possible using currently available methods, opening up new applications.Read moreRead less