Advanced Mixture Models for the Analysis of Modern-Day Data. Extracting key information from huge data sets is critical to the scientific successes of the future. This project will develop novel mixture models that can be used directly to analyse complex and high-dimensional data sets that may consist of thousands of variables observed on only a limited number of entities. In order to handle the challenging problems arising in the latter situation. This project develops mixtures of factor models ....Advanced Mixture Models for the Analysis of Modern-Day Data. Extracting key information from huge data sets is critical to the scientific successes of the future. This project will develop novel mixture models that can be used directly to analyse complex and high-dimensional data sets that may consist of thousands of variables observed on only a limited number of entities. In order to handle the challenging problems arising in the latter situation. This project develops mixtures of factor models with options for skew distributions that can be used to effectively analyse such data. Key applications include the domains of bioinformatics, biostatistics, business, data mining, economics, finance, image analysis, marketing, and personalised medicine, among many others.Read moreRead less
Joint clustering and matching of multivariate samples across objects. The project will provide a novel and very effective approach to the clustering of multivariate samples on objects, say patients, that automatically matches the sample clusters across the objects. A key application is the matching of biologically relevant cell subtypes across patients for use in the study and the clinical diagnosis and prognosis of cancer.
Expanding the role of mixture models in statistical analyses of big data. This project aims to develop theoretical procedures to scale inference and learning algorithms to analyse big data sets. It will develop analytic tools and algorithms to analyse big data sets which classical methods of inference cannot analyse directly due to the data’s complexity or size. This will accelerate the progress of scientific discovery and innovation, leading, for example, to new fields of inquiry; to an increas ....Expanding the role of mixture models in statistical analyses of big data. This project aims to develop theoretical procedures to scale inference and learning algorithms to analyse big data sets. It will develop analytic tools and algorithms to analyse big data sets which classical methods of inference cannot analyse directly due to the data’s complexity or size. This will accelerate the progress of scientific discovery and innovation, leading, for example, to new fields of inquiry; to an increase in understanding from studies on human and social processes and interactions; and to the promotion of economic growth and improved health and quality of life. Such applications should lead to breakthrough discoveries and innovation in science, engineering, medicine, commerce, education and national security.Read moreRead less
A new approach to fast matrix factorization for the statistical analysis of high-dimensional data. Some form of dimension reduction is essential in order to extract meaningful information from huge data sets. For this purpose we provide a novel and very fast approach to the factorization of the data matrix. It has wide applicability for improving the quality and validity of research in science and medicine and in most industries in Australia.
Large-Scale Statistical Inference: Multiple Testing. Multiple testing procedures are among the most important statistical tools for the analysis of modern data. This project aims to develop new methods for providing more powerful simultaneous tests while controlling the proportion of false positive conclusions. They are proposed to be derived by the novel pooling of information in individual attribute based contrasts to produce a Weighted Individual attribute-Specific Contrast (WISC) based stati ....Large-Scale Statistical Inference: Multiple Testing. Multiple testing procedures are among the most important statistical tools for the analysis of modern data. This project aims to develop new methods for providing more powerful simultaneous tests while controlling the proportion of false positive conclusions. They are proposed to be derived by the novel pooling of information in individual attribute based contrasts to produce a Weighted Individual attribute-Specific Contrast (WISC) based statistic. They will also exploit contextual information. They are expected to be of direct application to the problem of testing for no differences between two or more classes, as in the detection of differential expression in bioinformatics. Other key applications are expected to include biomedicine, economics, finance, genetics, and neuroscience.Read moreRead less
Discovery Early Career Researcher Award - Grant ID: DE170101134
Funder
Australian Research Council
Funding Amount
$360,000.00
Summary
Feasible algorithms for big inference. This project aims to develop algorithms for computationally-intensive statistical tools to analyse Big Data. Big Data is ubiquitous in science, engineering, industry and finance, but needs special machine learning to conduct correct inferential analysis. Computational bottlenecks make many tried-and-true tools of statistical inference inadequate. This project will develop tools including false discovery rate control, heteroscedastic and robust regression an ....Feasible algorithms for big inference. This project aims to develop algorithms for computationally-intensive statistical tools to analyse Big Data. Big Data is ubiquitous in science, engineering, industry and finance, but needs special machine learning to conduct correct inferential analysis. Computational bottlenecks make many tried-and-true tools of statistical inference inadequate. This project will develop tools including false discovery rate control, heteroscedastic and robust regression and mixture models, via Big Data-appropriate optimisation and composite-likelihood estimation. It will make open, well-documented, and accessible software available for the scalable and distributable analysis of Big Data. The expected outcome is a suite of scalable algorithms to analyse Big Data.Read moreRead less
International Networks in Applied Bayesian Statistics: improving Australia''s knowledge through intelligent data analysis and modelling. National benefits of this project are fourfold: (i) new international networks between Australia, Southern Africa, France and USA in the priority area of mathematical sciences; (ii) state-of-the-art Bayesian statistical methods for integrating and analyzing non-standard data and diverse information sources, including expert opinion, in order to solve complex pr ....International Networks in Applied Bayesian Statistics: improving Australia''s knowledge through intelligent data analysis and modelling. National benefits of this project are fourfold: (i) new international networks between Australia, Southern Africa, France and USA in the priority area of mathematical sciences; (ii) state-of-the-art Bayesian statistical methods for integrating and analyzing non-standard data and diverse information sources, including expert opinion, in order to solve complex problems in environment, industry, health, defence; (iii) direct contribution to solution of global environmental problems, specifically water quality, threatened species and environmental risk; (iv) superior training of the next generation of the global community of researchers in applied statistics.Read moreRead less
Australian Laureate Fellowships - Grant ID: FL150100150
Funder
Australian Research Council
Funding Amount
$2,413,112.00
Summary
Bayesian learning for decision making in the big data era. Bayesian learning for decision making in the big data era: This fellowship project aims to develop new techniques in evidence-based learning and decision-making in the big data era. Big data has arrived, and with it a huge global demand for statistical knowledge and skills to analyse these data for improved learning and decision-making. This project will seek to address this need by creating a step-change in knowledge in Bayesian statist ....Bayesian learning for decision making in the big data era. Bayesian learning for decision making in the big data era: This fellowship project aims to develop new techniques in evidence-based learning and decision-making in the big data era. Big data has arrived, and with it a huge global demand for statistical knowledge and skills to analyse these data for improved learning and decision-making. This project will seek to address this need by creating a step-change in knowledge in Bayesian statistics and translating this knowledge to real-world challenges in industry, environment and health. The new big data statistical analysts trained through the project could also create much needed capacity at national and international levels.Read moreRead less
Scalable and Robust Bayesian Inference for Implicit Statistical Models. This project aims to develop the next generation of efficient methods for fitting complex simulation-based statistical models to data. Practitioners and scientists are interested in such implicit models to enable discoveries, produce accurate predictions and inform decisions under uncertainty. However, the associated computational cost has restricted researchers to implicit models that must have a small number of parameters ....Scalable and Robust Bayesian Inference for Implicit Statistical Models. This project aims to develop the next generation of efficient methods for fitting complex simulation-based statistical models to data. Practitioners and scientists are interested in such implicit models to enable discoveries, produce accurate predictions and inform decisions under uncertainty. However, the associated computational cost has restricted researchers to implicit models that must have a small number of parameters and be well specified, impeding scientific progress. This project will develop new computational methods and algorithms for implicit models that scale to high dimensions and are robust to misspecification. Benefits will arise from the more routine use of implicit models in epidemiology, biology, ecology and other fields.Read moreRead less
Advances in Sequential Monte Carlo Methods for Complex Bayesian Models. This project aims to develop efficient statistical algorithms for parameter estimation of complex stochastic models that currently cannot be handled. Parameter estimation is an essential component of mathematical modelling for answering scientific questions and revealing new insights. Current parameter estimation methods can be inefficient and require too much user intervention. This project will develop novel Bayesian alg ....Advances in Sequential Monte Carlo Methods for Complex Bayesian Models. This project aims to develop efficient statistical algorithms for parameter estimation of complex stochastic models that currently cannot be handled. Parameter estimation is an essential component of mathematical modelling for answering scientific questions and revealing new insights. Current parameter estimation methods can be inefficient and require too much user intervention. This project will develop novel Bayesian algorithms that are optimally automated and efficient by exploiting ever-improving parallel computing devices. The new methods will allow practitioners to process realistic models, enabling new scientific discoveries in a wide range of disciplines such as biology, ecology, agriculture, hydrology and finance.Read moreRead less