Identifying technological trajectories using machine learning algorithms. This project aims to improve our understanding of why scientific knowledge progresses in certain directions and what causes it to grow faster or slower across fields. The project will create new neural-network machine-learning algorithms to scan patent and scientific article texts (specifications and claims) for natural language concepts. The results will potentially be used by patent offices to improve their own database ....Identifying technological trajectories using machine learning algorithms. This project aims to improve our understanding of why scientific knowledge progresses in certain directions and what causes it to grow faster or slower across fields. The project will create new neural-network machine-learning algorithms to scan patent and scientific article texts (specifications and claims) for natural language concepts. The results will potentially be used by patent offices to improve their own database search, by business analytics companies to reveal new technologies and potential collaborators, and by academic economists to understand how knowledge travels and accumulates.
Read moreRead less
Investigation and Development of Parallel Large Scale Record Linkage Techniques. Record linkage aims at matching records of the same entity (like customer or patient) in large (administrative) databases. The outcomes of the proposed research will improve current techniques in terms of efficiency, accuracy and the need for human intervention. Through experimental studies and stochastic modelling the performance of traditional and new methods for data cleaning, standardisation and linkage will be ....Investigation and Development of Parallel Large Scale Record Linkage Techniques. Record linkage aims at matching records of the same entity (like customer or patient) in large (administrative) databases. The outcomes of the proposed research will improve current techniques in terms of efficiency, accuracy and the need for human intervention. Through experimental studies and stochastic modelling the performance of traditional and new methods for data cleaning, standardisation and linkage will be assessed. The effect of the statistical dependency of attribute values will be studied. New methods using clustering for blocking large datasets, and predictive models including interaction terms will be implemented, analysed and evaluated on high-performance computers and office-based PC clusters.
Read moreRead less