- Location: School of Engineering, Featheringill Hall - Room 298
- Room: 298
- Contact: Andreas Berlind
- Email: firstname.lastname@example.org
- Website: https://www.facebook.com/vanderbiltdatascience/
- Audience: Free and Open to the Public
Title: Big and Large Data is Busting Out All Over
Every information and communication technology (ICT) system is designed under specific assumptions, and the field is defined by and operates under a more-or-less common set of them. One important class of assumptions defines how much data of various kinds is available and how much information or value can be derived from it. Recent trends in storage, networking and processing have led to an increase in the volume of data and in the value that can be derived from it, exceeding a number of different conventional assumptions, and giving rise to the study of “Big Data”. In this talk I will draw the distinction between data that is simply high volume (what I term “Large Data”) and data from which high value can be derived (the commonly used meaning of “Big Data”). I will also examine other ICT assumptions that are being violated in practice, specifically in wide area distributed systems. These new conditions are creating unexpected challenges and new opportunities in future data management, communication and computing infrastructure. Examples include High Traffic Content Delivery, Smart Cities, Disaster Recovery, Digital Divide, Cyberphysical Systems and Vehicular Data Management.
Micah Beck began his research career in distributed operating systems at BellLaboratories, and received his Ph.D. in Computer Science from Cornell University (1992) in the area of parallelizing compilers. He then joined the faculty of the CS Department at the University of Tennessee, where he is currently an Associate Professor working in architecture for distributed and high-performance computing, networking and storage. He is known for his work in Data Logistics, having done early work in Web Caching and Replication and Content Delivery Networks, which motivated the Logistical Networking architecture and the Data Logistics Toolkit open source service stack. He was a cofounder of the Internet2 Distributed Storage Infrastructure project, Lokomo Distributed Systems (a startup company) and the Logistical Networking and Computation Laboratory (at UT Knoxville).