Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.”

Slides:



Advertisements
Similar presentations
© 2013 IBM Corporation Complexity must be conquered (Data is more than “just” big) Laura Haas IBM Almaden Research Center 1.
Advertisements

Moore’s Law Kyle Doran Greg Muller Casey Culham May 2, 2007.
CMSC 611: Advanced Computer Architecture Performance Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
2.6 Moore’s Law U2C6.
Quantum Computers. Overview Brief History Computing – (generations) Current technology Limitations Theory of Quantum Computing How it Works? Applications.
Lecture 1: Performance EEN 312: Processors: Hardware, Software, and Interfacing Department of Electrical and Computer Engineering Spring 2013, Dr. Rozier.
 Recall:  3.1: We looked at how the price elasticity of demand measures the responsiveness of consumers to a change in a product’s price  Now:  3.2:
CS 104 – Fall 2011 Exploring Computer Science Lecture 2: August 29, Blown to Bits - CS Unplugged.
Limits and Horizon of Computing Post silicon computing.
Moore’s Law By Cameron Waters. Who’s Law? Gordon Moore Born January 3 rd 1929 Member of Institute of Electrical and Electronics Engineers (IEEE) Co-Founder.
SUPPLY CHAIN MANAGEMENT SYSTEMS Part I. 7-2 LEARNING OUTCOMES 1.List and describe the components of a typical supply chain 2.Define the relationship between.
Models in Science.
Chapter VI What should I know about the sizes and speeds of computers?
DNA Computing. What is it?  “DNA computing is a branch of computing which uses DNA, biochemistry, and molecular biology hardware, instead of the traditional.
CSC 108H: Introduction to Computer Programming Summer 2012 Marek Janicki.
CSC 108H: Introduction to Computer Programming Summer 2011 Marek Janicki.
Integration Lower sums Upper sums
Encryption with Keys and Passwords
The Study of Computer Science Chapter 0
Vocabulary byte - The technical term for 8 bits of data.
Digital Aerospace and Defense Build, Service, and Fly Better
AP CSP: Cleaning Data & Creating Summary Tables
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.” Moore’s.
AP CSP: What is Big Data?.
Vocabulary Prototype: A preliminary sketch of an idea or model for something new. It’s the original drawing from which something real might be built or.
Vocabulary Algorithm - A precise sequence of instructions for processes that can be executed by a computer Low level programming language: A programming.
Introduction to e-Commerce
Vocabulary byte - The technical term for 8 bits of data.
UNIT 2 – CHAPTER 1 – LESSON 1 DIGITAL INFORMATION.
Vocabulary byte - The technical term for 8 bits of data.
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.” Moore’s.
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.” Moore’s.
Vocabulary Prototype: A preliminary sketch of an idea or model for something new. It’s the original drawing from which something real might be built or.
Inspirational Physical Education
for the Offline and Computing groups
Section 2: Science as a Process
The Study of Computer Science Chapter 0
Section 1: The Nature of Science
UNIT 4 - BIG DATA AND PRIVACY
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.” Moore’s.
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.” Moore’s.
Vocabulary byte - The technical term for 8 bits of data.
AP CSP: Bytes, File Sizes, and Text Compression
RAPID RESEARCH DATA INNOVATIONS
What is Product “X”? Name a Product “X” have you recently bought to replace an older version, where the new one was much: Better Faster Cheaper Computer.
What was Earth’s temperature like in the past?
School of Information Management Nanjing University China
Internet Marketing & e-Commerce Ward Hanson Kirthi Kalyanam Requests for permission to copy any part of the material should be addressed to: PERMISSIONS.
Limits and Horizon of Computing
LESSON 12 - Loops and Simulations
The Study of Computer Science
Welcome to LSP 120 Dr. Curt M. White.
CMSC 611: Advanced Computer Architecture
CHAPTER 12 More About Regression
RAPID RESEARCH DATA INNOVATIONS
Section 1: The Nature of Science
Developments Key Revision Points.
Unit 4: Lesson 1 & 2- What is Big Data
Section 1: The Nature of Science
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.” Moore’s.
CMSC 611: Advanced Computer Architecture
CHAPTER 12 More About Regression
Section 1: The Methods of Science
Using data innovations to improve
Transformations Limit Continuity Derivative
CHAPTER 12 More About Regression
Future of Computers Moore’s Law.
U3L2 The Need For Algorithms
U3L4 Using Simple Commands
Welcome.
Presentation transcript:

Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.”

CS Principles U4L10 – What is Big Data

U4L10 – What is Big Data Objectives SWBAT: Identify sources of data produced, used, and consumed by a web application. Given a tool that provides access to a large dataset, explain the kinds of problems such a tool could solve. Use a tool that provides access to “big data” and investigate its sources. Explain that new techniques are necessary to store, manage, transmit, and process data at the scale it is currently being produced

Big Data Understanding the types of things that can be captured in data and anticipating the types of innovations or new knowledge that can be built upon this data is increasingly the role of the computer scientist. A first step toward understanding big data is a survey of how big data is already being used to learn and solve problems across numerous disciplines. The scale of big data makes it hard to “see” sometimes, and techniques for looking at, working with, and understanding data change once the data is “big.” Everything, from how it’s stored to how it’s processed to how it's visualized, is a little different once you enter the realm of big data.

Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.”

The exponential growth of data: Moore’s Law A often-cited “law” in computer science is “Moore’s law.”  In 1965, a computer chip designer named Gordon Moore predicted that the number of transistors one could fit on a chip would double every 18 months or so. That prediction more or less held true, and the implication is that, since about 1970, computers have gotten twice as fast, at half the cost, every 18 months or so.  With some small differences, the same is true for data storage capacity.  With more and more machines that are faster and faster, the amount of data being pushed around, saved, and processed is growing exponentially. As you can see from the chart, the amount of data flying around is growing exponentially, doubling every two years or so. Here’s a way to think about how fast this is: The world will produce as much digital data over the next 2 years, as currently existed in all of humanity prior to that. And it will do the same the 2 years after that. And so on. That’s a lot!  We need to keep Moore’s Law in mind as we plan for the future.

Moore’s Law There is a principle in computer science known as Moore’s Law. Wikipedia: Moore’s Law It is not a law of nature or mathematics but simply a surprisingly accurate prediction that was made a long time ago. In 1965, a computer chip designer named Gordon Moore predicted that the number of transistors one could fit on a chip would double every 18 months or so. Amazingly, that prediction has more or less held true to the present day! The result is that since about 1970, computers have gotten twice as fast, at half the cost, roughly every 1.5-2 years With some small differences, the same is true for data storage capacity. This is extraordinarly fast growth - we call it exponential growth. With more and more machines that are faster and faster, the amount of data being pushed around, saved, and processed is growing exponentially. This is so fast that it’s hard to fathom and even harder to plan for. For example: If the average hard drive today is 1 TB and you are planning for something or 6 years away, you should expect that average hard drives will be 8-10 TB. Key Takeaway: We need to keep Moore’s Law in mind as we plan for the future.

U4L10 Reflection Prompt: “So, after your explorations what do you think “big data” actually means? What makes it “big” as opposed to not?“ Consider the following in your response: What kinds of fields and industries are there that aren’t affected by big data? Are there any areas of life that are impossible to be affected by big data? What are some of the challenges of working with big data? What’s different about working with big data versus other data we’ve seen so far?