There are massive amounts of data presented to the internet in real or near-to-real-time that allow monitoring of economic and societal conditions, amongst other things. This has been used to some effect in the automated monitoring of stock fluctuations, used to inform algorithmic trading. The evaluation of data-sources and curation of these on the basis of predictive power is an area requiring exploration. The project here will focus on a multitude of data sources that would allow the real-time evaluation of geo-political conditions around the globe, with the intention to predict various market shifts and impending political flash-points. Data will be captured by a wide range of sources, including multiple languages, print, audio and video. Text-mining methods will be developed to generate topic models and monitoring these over time, including those emergent. Methods will be explored that evaluate predictive performance for the purposes of curating data-sources and selection of modelling techniques.
The project deals with large data issues, data-mining/machine-learning methods, cloud-computing and interaction with APIs. The project will be heavily computational in either R or python, with potentially compiled languages for computational bottlenecks. The ideal candidate would have a good grasp of practical computing and statistics.
Applicants should have a good first degree in mathematics, statistics or another discipline (e.g., biology, computer science), with substantial statistical component. A masters-level degree is an advantage. Requirement details available here.
Applications are accepted all year round.
Application details are available here.
If you need help in responding to this opportunity, please write your question in the comment box below.
If you need to receive similar opportunities like this one, please follow this blog by pressing the follow button in the upper right corner of this page.