Keynote Speakers (in alphabetic order)
University of Connecticut, USMicrosoftTokyo Institute of Technology, JPChinese Academy of Sciences, CNConsiglio Nazionale delle Ricerche, ITUniversity of Warwick, UK
Jeffrey Craig Hoch
University of Connecticut, US
NMRbox is a new software platform for biomolecular nuclear magnetic resonance (NMR) spectroscopy designed to tame complexity and foster computational reproducibility. Computation is crucial for applications of NMR that provide insights into biomolecular structure, dynamics, interactions, and stability, and have translational applications including diagnostics and drug discovery. Spectrum analysis, parametric data reduction, computer--‐aided assignment, spectral databases, and molecular structure computation and validation are just some of the tasks dependent on sophisticated software, and have all been vital for extending the size and complexity of biomolecular systems amenable to NMR. The amount and variety of NMR software is already large, and it is rapidly expanding, making it more difficult to discover new software, to support a full complement of software on a diverse set of computer platforms, or to maintain legacy packages that are no longer actively developed. Communication between software packages also becomes more challenging as new packages proliferate. Facile discovery of new software and removing barriers to interoperability are essential for NMR software to be fully utilized. The broad aim of NMRbox is to simplify and integrate dissemination, maintenance, support, interoperation, and application of a range of widely used NMR data processing and analysis software packages. NMRbox exploits modern virtualization technologies to provide high availability as well as software persistence. By facilitating the use and persistence of advanced software, NMRbox will advance the application of biomolecular NMR for challenging applications in biomedicine, including structural biology, drug discovery, and metabolomics. The approaches used in NMRbox to address challenges in computational reproducibility have potential application to other fields. Visit nmrbox.org.
Jeffrey Craig Hoch received an A.B. (Chemistry, 1976) from Boston University, and A.M. (Chemistry, 1978) and Ph.D. (Physical Chemistry, 1983) degrees from Harvard University, working with Christopher Dobson and Nobel Laureate Martin Karplus. From 1983 to 2003 he headed the Biophysical Chemistry group at the Rowland Institute for Science in Cambridge, MA. In 2003 he moved to the University of Connecticut Medical School, where he is Professor of Molecular Biology and Biophysics, Director of the Gregory P. Mullen NMR Structural Biology Facility, and Director of NMRbox: National Center for NMR Data Processing and Analysis. His research interests are in the development and application of nuclear magnetic resonance spectroscopy for investigation of biologically important molecules. He was elected to the Connecticut Academy of Science and Engineering in 2012.
Ulrich Homann is a Distinguished Engineer in the Cloud and Enterprise business at Microsoft. As part of the senior engineering leadership team, Homann is responsible for enterprise scenarios and focus across the cloud and enterprise platform working on the most complex and challenging customer and partner opportunities driving structured growth for the $26 billion p.a. business. Previously Homann was the Chief Architect for Microsoft worldwide enterprise services. As part of the office of the CTO, he was responsible for the management and enterprise architecture of key strategic projects all over the world. Homann also defined and drove the technical strategy for emerging technologies and trends such as cloud or in-memory computing, big data, etc.. He was responsible for setting adoption and support strategy for the key technologies and its relevancy in the worldwide services business as well developing new senior technical talent. Formerly Homann played a key role in the newly formed Platforms, Technology and Strategy Group at Microsoft. As solutions architect, Homann was responsible for the development and strategy for Microsoft in the area of business architecture and its relationship to solution architecture, software tools and infrastructure. Previously, Ulrich was program manager in the Enterprise Program Management responsible for the development relationship with key partners, mainly SAP AG but also players such as Baan, JD Edwards, Siebel, etc. Ulrich also served as part of Microsoft Consulting Services, where he was responsible for several key distributed database application projects in Germany. Prior to joining Microsoft in 1991, Ulrich worked for several small consulting companies, where he designed and developed distributed systems. Ulrich has more than 25 years of experience in the systems industry. He has spent most of his career using well-defined applications and architectures to simplify and streamline the development of business applications. His dedication to this pursuit is driven by a passion to improve product planning by working closely with customers and their applications. Ulrich holds a bachelor of computer science from the University of Cologne, Germany.
The Inevitable End of Moore’s Law beyond Exascale will Result in Data and HPC Convergence and More
Tokyo Institute of Technology, JP
The so-called “Moore’s Law”, by which the performance of the processors will increase exponentially by factor of 4 every 3 years or so, is slated to be ending in 10-15 year timeframe due to the lithography of VLSIs reaching its limits around that time, and combined with other physical factors. This is largely due to the transistor power becoming largely constant, and as a result, means to sustain continuous performance increase must be sought otherwise than increasing the clock rate or the number of floating point units in the chips, i.e., increase in the FLOPS. The promising new parameter in place of the transistor count is the perceived increase in the capacity and bandwidth of storage, driven by device, architectural, as well as packaging innovations: DRAM-alternative Non-Volatile Memory (NVM) devices, 3-D memory and logic stacking evolving from VIAs to direct silicone stacking, as well as next-generation terabit optics and networks. The overall effect of this is that, the trend to increase the computational intensity as advocated today will no longer result in performance increase, but rather, exploiting the memory and bandwidth capacities will instead be the right methodology. However, such shift in compute-vs-data tradeoffs would not exactly be return to the old vector days, since other physical factors such as latency will not change. As such, performance modeling to account for the evolution of such fundamental architectural change in the post-Moore era would become important, as it could lead to disruptive alterations on how the computing system, both hardware and software, would be evolving towards the future. We are now in the process of launching such innovative projects for the future of computing in Japan
Satoshi Matsuoka has been a Full Professor at the Global Scientific Information and Computing Center (GSIC), a Japanese national supercomputing center hosted by the Tokyo Institute of Technology, since 2001. He received his Ph. D. from the University of Tokyo in 1993. He is the leader of the TSUBAME series of supercomputers, including TSUBAME2.0 which was the first supercomputer in Japan to exceed Petaflop performance and became the 4th fastest in the world on the Top500 in Nov. 2010, as well as the recent TSUBAME-KFC becoming #1 in the world for power efficiency for both the Green 500 and Green Graph 500 lists in Nov. 2013. He is also currently leading several major supercomputing research projects, such as the MEXT Green Supercomputing, JSPS Billion-Scale Supercomputer Resilience, as well as the JST-CREST Extreme Big Data. He has written over 500 articles according to Google Scholar, and chaired numerous ACM/IEEE conferences, most recently the overall Technical Program Chair at the ACM/IEEE Supercomputing Conference (SC13) in 2013. He is a fellow of the ACM and European ISC, and has won many awards, including the JSPS Prize from the Japan Society for Promotion of Science in 2006, awarded by his Highness Prince Akishino, the ACM Gordon Bell Prize in 2011, the Commendation for Science and Technology by the Minister of Education, Culture, Sports, Science and Technology in 2012, and recently the 2014 IEEE-CS Sidney Fernbach Memorial Award, the highest prestige in the field of HPC.
Chinese Academy of Sciences, CN
Kai Nan is a professor of computer science and Deputy Director at the Computer Network Information Center (CNIC) of the Chinese Academy of Sciences (CAS). He is also Director of China Science and Technology Network (CSTNET) since 2012. He received his B.S. in electronics engineering from Peking University in 1996, and his M.S. and Ph.D. in computer science from the Institute of Computing Technology of CAS in 1999 and 2006, respectively.
His research interests include distributed computing, cloud computing and CSCW. He has extensive experience in construction and operation of large scale information systems and Internet services. Since 2007, he has led the development of the Duckling Software, which is ever an Internet-based collaboration toolkit and now a cloud platform service, mainly for scientific research teams. The website www.escience.cn, a hub of services powered by Duckling, has had hundreds of thousands users in Chinese academia and grows. In 2012-2015, he directs the Science and Technology Cloud project funded by CAS.
Dr. Nan serves as Chair of APAN China and serves on Standing Board of the Internet Society of China, Steering Committee of the Pacific Rim Application and Grid Middleware Assembly (PRAGMA) and Steering Committee of the IEEE International Conference on eScience.
Luca Pezzati (male). Physicist and Optics Specialist. Researcher at INO-CNR. He started the Art Diagnostic Group of INO and the “Optical Metrology Lab for the Diagnostic of Cultural Heritage” at OPD-Opificio delle Pietre Dure in Florence in 1998. Currently: director of the INO Research Unit in Lecce; coordinator of the EU project IPERION CH – integrated Platform for the European Research Infrastructure on Cultural Heritage (http://iperionch.eu) and national coordinator for Italy of DARIAH ERIC (http://www.dariah.eu). He has managed many research projects for CNR in the field of optical technologies applied to Cultural Heritage. See his full profile at http://it.linkedin.com/pub/luca-pezzati/6/15b/550/
University of Warwick, UK
Dr Emma Uprichard is Associate Professor and Deputy Director at the Centre for Interdisciplinary Methodologies, University of Warwick, UK. She is founder and co-director of the Warwick Q-Step Centre (funded by the Nuffield Foundation, ESRC and HEFCE); this is one of fifteen Q-Step Centres across the UK, developed as a strategic response to the shortage of quantitatively-skilled social science graduates. She is currently PI of the ESRC Seminar Series, Complexity and Method in the Social Sciences. Her work is driven by the methodological challenge of studying complex social systems across time and space. Her substantive research interests revolve around studying change and continuity in relation to: big data for policy; cities; food; time; childhood; and time-space methodologies. She is currently preparing a book on ‘Time and Method’ (Routledge).