|
ICCS 2014 Keynote Lectures
|
Towards Scalable Mathematical Methods and Algorithms for Extreme Scale Computing | |
Abstract: Novel mathematics and mathematical modelling approaches together with scalable algorithms are needed to enable key applications at extreme-scale. This is especially true as HPC systems continue to scale up in compute node and processor core count. At the moment computational scientists are at the critical point/threshold of novel mathematics development as well as large-scale algorithm development and re-design and implementation that will affect most of the application areas. Therefore the talk will focus on the mathematical and algorithmic challenges and approaches towards exascale and beyond. It will present ideas and some novel mathematical methods and approaches for extreme-scale systems, such as stochastic and hybrid methods that in turn lead to scalable scientific algorithms with minimal or no global communication, hiding network and memory latency, have very high computation/communication overlap, have no synchronization points. Examples will be drawn from diverse application areas. |
A Science of Cities | |
Abstract: The study of cities immediately brings up most of the complex questions inherent to understanding human social systems in terms of their extreme complexity and evolvability. Yet, over the last few years increasing accessibility of data and a more interdisciplinary focus have started to pay off and reveal the nature of cities as complex adaptive systems, describable by quantitative laws about their general organization and dynamics. In this talk I will review the current empirical understanding of the structure and dynamics of cities in terms of their land use, social organization and mobility costs. I will show how this evidence can be synthesized in terms of a small set of quantitative principles of self-organization and how, in this light, we can understand the type of information processing cities perform in human societies. I will conclude with a set of open challenges for a deeper understanding of cities as computational networks that when addressed will form the basis of a mature science of cities and of other complex adaptive systems. |
Automating Computational Materials Discovery through Model-Integrated Computing | |
Abstract: One of the key elements of President Obama’s manufacturing competitiveness initiative is the Materials Genome Initiative (MGI). Using high accuracy validated computational materials methods to predict and screen materials for specific properties, the MGI’s goal is to reduce by a factor of two both the time and the cost to bring new materials to the market. The development of computational materials screening tools for hard materials (metals, alloys, ceramics, etc) is relatively easy compared to the same task for soft materials (liquids, colloids, polymers, foams, gels, granular materials, and soft biomaterials), since in the former case the energy scales involved are large by comparison to thermal energy (~kBT where kB is Boltzmann’s constant and T is temperature) and the atoms/molecules are typically in crystalline lattice with no mesoscopic structuring. By contrast, for soft materials the energy scales are comparable to thermal energy, weak dispersion interactions frequently dominate, and soft matter often self-organizes into mesoscopic physical structures that are much larger than the microscopic scale yet much smaller than the macroscopic scale, and the macroscopic properties result from the mesoscopic structure. Consequently, to perform computational screening and design for soft materials involves complex molecular simulations, frequently on large petascale computing platforms and in some cases performed at multiple levels of detail. These simulations have complex workflows that experts understand and perform routinely; however, to translate these workflows into a set of tasks that can be automated, combined with other tasks and embedded within hierarchies of application-hardened engineering methodologies (such as stochastic optimization), we employ the discipline of model integrated computing (MIC). Using MIC, we develop a “meta-programming” tool that enables the establishment of a domain specific modeling language for the creation, synthesis and execution of simulation workflows. We illustrate this within two specific materials design problems: tethered nanoparticle systems and lubrication systems. |
The Fourth Paradigm: Data-Intensive Scientific Discovery with the Cloud | |
Abstract: There is broad recognition within the scientific community that the emerging data deluge will fundamentally alter disciplines in areas throughout academic research. A wide variety of scientists—biologists, chemists, physicists, astronomers, engineers – will require tools, technologies, and platforms that seamlessly integrate into standard scientific methodologies and processes. “The Fourth Paradigm” refers to the data management techniques and the computational systems needed to manipulate, visualize, and manage large amounts of scientific data. This talk will illustrate the challenges researchers will face, the opportunities these changes will afford, and the resulting implications for data-intensive researchers. It will also cover how Cloud Computing can be used to new insights in Scientific Discovery. |
Scaling and compression of regulatory information in integrated complex systems: lessons from genomics | |
Abstract: At 150 Whole Human Genomes every 3 days, scaling up to handle Illumina's HiSeq X Ten population-scale genome sequencing factory has many implications. Our group in the Kinghorn Centre for Clinical Genomics at the Garvan Institute are addressing the challenges of analysing and storing 18000 human genomes each year, together with our plans for building a Genome Variant Store to correlate genome variants with phenotypical information. My talk will include core software systems including our LIMS used in our sequencing laboratory, SeqWare for the running and management of analytical pipelines as well as Alfred, our own software application, that oversees everything from sample receipt to the management of Quality Control events and the return of results to users in both research and clinical settings. I will also discuss our software development process of Continuous Integration that allows us to update our software systems multiple times daily, while ensuring that upgrades are not introducing new and unforeseen problems. Finally, I intend discussing our infrastructure, both in-house and external, which we will use in order to ensure the project's success. |
Computational science in conservation planning: potential and limitations | |
Abstract: The field of systematic conservation planning began, without computers, in 1983 but quickly developed a dependence on machines for rapid processing of large data sets. Since then the conceptual and technical expectations of computer analyses have grown exponentially, and algorithms have been developed for many, though not all problems. Dimensionality presents challenges, if not for analysis then for data elicitation. The leading edges of computation in conservation concern visualization and dynamic interaction. Perhaps the greatest limitations of computing in the field relate to interactions with people. People are affected by conservation decisions and increasingly expecting a role in decisions. This requires optimality to be redefined to refer to socially, as well as environmentally optimal outcomes. |
Modelling cancer as a transcriptional landscape | |
Abstract: More than half a century ago, the British developmental biologist and philosopher Conrad Hal Waddington introduced the landscape as a metaphor of how cells differentiate into different types of tissues. This is now recognised as “probably the most famous and most powerful metaphor in developmental biology”. Nonetheless it has remained unclear how such a surface might be computed from actual cell-state data, and if so whether it could be informative or predictive about real-life biology. Here I explore Hopfield neural networks with genome-wide gene expression data as a computational model of Waddington’s landscape. I discuss concepts of state, trajectory and attractors, and present visualisations of the Hopfield surface for subtypes of breast cancer. |
Research Computing Centre |