Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture 23 Intelligent Digital Forensics October 22, 2007.

Slides:



Advertisements
Similar presentations
SDMX in the Vietnam Ministry of Planning and Investment - A Data Model to Manage Metadata and Data ETV2 Component 5 – Facilitating better decision-making.
Advertisements

DIGIDOC A web based tool to Manage Documents. System Overview DigiDoc is a web-based customizable, integrated solution for Business Process Management.
Database System Concepts and Architecture
CAPTURE SOFTWARE Please take a few moments to review the following slides. Please take a few moments to review the following slides. The filing of documents.
Transaction.
Guide to Computer Forensics and Investigations, Second Edition
Information Retrieval in Practice
Xyleme A Dynamic Warehouse for XML Data of the Web.
Aki Hecht Seminar in Databases (236826) January 2009
Statistical Relational Learning for Link Prediction Alexandrin Popescul and Lyle H. Unger Presented by Ron Bjarnason 11 November 2003.
Chapter 14 The Second Component: The Database.
Data and Applications Security Developments and Directions Dr. Bhavani Thuraisingham The University of Texas at Dallas Secure Knowledge Management: and.
Overview of Search Engines
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas File Systems and Forensics Tools September 20, 2013.
COEN 252 Computer Forensics Forensic Duplication of Hard Drives.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #12 Intelligent Digital Forensics September 30, 2009.
Databases & Data Warehouses Chapter 3 Database Processing.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Review for Final Exam November 19, 2010.
MDC Open Information Model West Virginia University CS486 Presentation Feb 18, 2000 Lijian Liu (OIM:
By N.Gopinath AP/CSE. Why a Data Warehouse Application – Business Perspectives  There are several reasons why organizations consider Data Warehousing.
CLARIN tools for workflows Overview. Objective of this document  Determine which are the responsibilities of the different components of CLARIN workflows.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
A summary of the report written by W. Alink, R.A.F. Bhoedjang, P.A. Boncz, and A.P. de Vries.
Aurora: A Conceptual Model for Web-content Adaptation to Support the Universal Accessibility of Web-based Services Anita W. Huang, Neel Sundaresan Presented.
Katanosh Morovat.   This concept is a formal approach for identifying the rules that encapsulate the structure, constraint, and control of the operation.
Ihr Logo Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization Turban, Aronson, and Liang.
1 The BT Digital Library A case study in intelligent content management Paul Warren
Multimedia Databases (MMDB)
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Introduction to the Course August 24, 2011.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #8 Computer Forensics Data Recovery and Evidence Collection September.
PLoS ONE Application Journal Publishing System (JPS) First application built on Topaz application framework Web 2.0 –Uses a template engine to display.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #8 Guest Lecture September 21, 2009.
Summary Marie Yarbrough. Introduction History of Image Forgery Method Segmentation Classification Common-Sense Reasoning Conclusion.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Introduction to the Course August 29, 2014.
Knowledge Representation of Statistic Domain For CBR Application Supervisor : Dr. Aslina Saad Dr. Mashitoh Hashim PM Dr. Nor Hasbiah Ubaidullah.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Introduction to the Course August 20, 2007.
Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #30 Network Forensics (Revisited) November 7, 2007.
Data and Applications Security Developments and Directions Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #22 Secure Web Information.
Computer Systems & Architecture Lesson 4 8. Reconstructing Software Architectures.
1 Technology in Action Chapter 11 Behind the Scenes: Databases and Information Systems Copyright © 2010 Pearson Education, Inc. Publishing as Prentice.
A radiologist analyzes an X-ray image, and writes his observations on papers  Image Tagging improves the quality, consistency.  Usefulness of the data.
Using and modifying plan constraints in Constable Jim Blythe and Yolanda Gil Temple project USC Information Sciences Institute
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Using Text Mining and Natural Language Processing for.
Data and Applications Security Developments and Directions Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture #15 Secure Multimedia Data.
Selective and Intelligent Imaging Using Digital Evidence Bags.
INTRODUCTION TO GIS  Used to describe computer facilities which are used to handle data referenced to the spatial domain.  Has the ability to inter-
Using automation to enhance the process of Digital Forensic analysis Daniel Walton School of Computer and Information Science
Digital Video Library Network Supervisor: Prof. Michael Lyu Student: Ma Chak Kei, Jacky.
Detecting False Captioning Using Common Sense Reasoning James Byrd.
COEN 252 Computer Forensics Forensic Duplication of Hard Drives.
Forensic Investigation Techniques Michael Jones. Overview Purpose People Processes Michael Jones2Digital Forensic Investigations.
Copyright © 2016 Pearson Education, Inc. Modern Database Management 12 th Edition Jeff Hoffer, Ramesh Venkataraman, Heikki Topi CHAPTER 11: BIG DATA AND.
Introduction: Databases and Database Systems Lecture # 1 June 19,2012 National University of Computer and Emerging Sciences.
Information Retrieval in Practice
Databases (CS507) CHAPTER 2.
Fundamentals of Information Systems, Sixth Edition
INTRODUCTION TO GEOGRAPHICAL INFORMATION SYSTEM
Data and Applications Security Developments and Directions
Introduction Multimedia initial focus
Data and Applications Security Developments and Directions
Digital Forensics Dr. Bhavani Thuraisingham
Data Warehouse.
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Database System Concepts and Architecture.
Data Warehousing and Data Mining
ece 627 intelligent web: ontology and beyond
Dr. Bhavani Thuraisingham The University of Texas at Dallas
Data and Applications Security Developments and Directions
The ultimate in data organization
Data and Applications Security Developments and Directions
Presentation transcript:

Digital Forensics Dr. Bhavani Thuraisingham The University of Texas at Dallas Lecture 23 Intelligent Digital Forensics October 22, 2007

Reading for Monday October 22 l l XIRAF – XML-based indexing and querying for digital forensics l Selective and intelligent imaging using digital evidence bags l l Detecting false captioning using common-sense reasoning

Outline l Review of Lectures l Discussion of the papers on Intelligence Digital Forensics l Appendix: feature extraction

Review of Lectures l Auditing and Forensic Analysis l Richard T. Snodgrass, Stanley Yao and Christian Collberg, "Tamper Detection in Audit Logs," In Proceedings of the International Conference on Very Large Databases, Toronto, Canada, August–September 2004, pp. 504–515. l Kyri Pavlou and Richard T. Snodgrass, "Forensic Analysis of Database Tampering," in Proceedings of the ACM SIGMOD International Conference on Management of Data (SIGMOD), pages , Chicago, June, l Additional paper for reading: Kyri Pavlou and Richard. T. Snodgrass, "The Pre-images of Bitwise AND Functions in Forensic Analysis,'' TimeCenter TR 87, October, l

Abstract of Paper 1 l This paper describes a novel, XML-based approach towards managing and querying forensic traces extracted from digital evidence. This approach has been implemented in XIRAF, a prototype system for forensic analysis. XIRAF systematically applies forensic analysis tools to evidence files (e.g., hard disk images). Each tool produces structured XML annotations that can refer to regions (byte ranges) in an evidence file. XIRAF stores such annotations in an XML database, which allows us to query the annotations using a single, powerful query language (XQuery). XIRAF provides the forensic investigator with a rich query environment in which browsing, searching, and predefined query templates are all expressed in terms of XML database queries

Introduction l Framework for forensic analysis called XIRAF l A clean separation between feature extraction and analysis - Features extracted are stored in XML format l A single, XML-based output format for forensic analysis tools l The use of XML database technology for storing and querying the XML output of analysis tools.

XIRAF Framework l Consists of three components l Feature extraction manager - Features are extracted from BLOBs (Binary large objects) using feature extraction tools - Output of the tools are coded in XML for the forensics analyzer l Tool repository - Tools are wrapped (e.g., object wrappers) l Storage subsysystem - Stores BLOBs and XML annotations - XQuery used to query XML data

Forensic Applications l Authors have implemented following applications - Timeline browser: Through web browser examiner can look at data/time of interest - Photo search l Search for images satisfying certain conditions - Child pornography detection l Using hashing carried out matching

Summary and Directions l The separation of feature extraction and analysis brings benefits to both phases. XIRAF extracts features automatically, which is essential when processing large input sets. l The use of XML as a common, intermediate output format for tools allows the integration of the output of diverse, independent tools that produce similar information. This handles both the heterogeneity present in the input data (e.g., different browser types) and with the diversity of forensic analysis tools. l These benefits are demonstrated both by the timeline browser and by child pornography detection program. l By storing extracted features in an XML database system one can analyze those features using a single, general-purpose, powerful query language. In addition, we benefit automatically from advances that are made in the area of XML database systems l Directions: Use semantic web technologies?

Abstract of Paper 2 l This paper defines what selective imaging is, and the types of selective imaging that can be performed. This is contrasted with intelligent imaging and the additional capabilities that have to be built into an imager for it to be ‘intelligent’. A selective information capture scenario is demonstrated using the digital evidence bag (DEB) storage format. A DEB is a universal container for digital evidence from any source that allows the provenance to be recorded and continuity to be maintained throughout the life of the investigation. The paper concludes by defining the ‘ultimate test’ for an intelligent and selective imager

Selective Imaging l Selective imaging is a term that is generally associated with the decision not to acquire all the possible information during the capture process. l It is now recognized that ‘partial or selective file copying may be considered as an alternative’ when it may not be practical to acquire everything. l Techniques include manual selection, semi-automatic selection, automatic selection

Intelligent Imaging l Include the domain experts in the imaging process l How do you go about capturing the knowledge of the technical experts that are familiar with digital technical complexities and legal domain experts and combine them? l How do you know that you have captured everything relevant to the case under investigation or have not missed evidence of other offences?

Digital Evidence Bags l Both selective and intelligent imaging techniques offer many more options and capabilities than current bit stream imaging. l There are currently no commercial tools that perform selective imaging and adequately record the provenance of the selected information. l Furthermore, no method has existed that captured the criteria or method used by the examiner in deciding what to acquire. For example, was an arbitrary manual selection used or was information captured based on category of information, file extensions, file signature or hash set. l Authors solution to these problems is by the use of the digital evidence bag (DEB) format. A DEB is a universal container for digital information from any source. It allows the provenance of digital information to be recorded and continuity to be maintained throughout the life of the exhibit. l Additionally, DEBs may be encapsulated within other DEBs. This feature differentiates the DEB structure from that used by current monolithic formats commonly in use.

“The Ultimate Test” l The method and storage container used must be able to store sufficient information about the provenance of the information captured such that when the information is restored it is identical to that which would have been acquired should a bit stream image have been taken.

Summary and Directions l The methodology described and demonstrated by the authors is claimed to be a big improvement over bit stream imaging methods currently used. l Directions - Better selection methods, more accurate?

Abstract of Paper 3 l Detecting manipulated images has become an important problem in many domains (including medical imaging, forensics, journalism and scientific publication) largely due to the recent success of image synthesis techniques and the accessibility of image editing software. Many previous signal-processing techniques are concerned about finding forgery through simple transformation (e.g. resizing, rotating, or scaling), yet little attention is given to examining the semantic content of an image, which is the main issue in recent image forgeries. Here, the authors present a complete workflow for finding the anomalies within images by combining the methods known in computer graphics and artificial intelligence. They first find perceptually meaningful regions using an image segmentation technique and classify these regions based on image statistics. We then use AI common-sense reasoning techniques to find ambiguities and anomalies within an image as well as perform reasoning across a corpus of images to identify a semantically based candidate list of potential fraudulent images. Their method introduces a novel framework for forensic reasoning, which allows detection of image tampering, even with nearly flawless mathematical techniques.

Introduction l Detecting manipulated images has become an important problem in many domains l Many previous signal-processing techniques are concerned about finding forgery through simple transformation (e.g. resizing, rotating, or scaling), l Need to examine the semantic content of an image l Authors present a complete workflow for finding the anomalies within images by combining the methods known in computer graphics and artificial intelligence

Introduction l In Photo fakery, photo manipulation techniques may fall into four categories: l Deletion of details: removing scene elements l Insertion of details: adding scene elements l Photomontage: combining multiple images l False captioning: misrepresenting image content

Technical Approach l Authors find perceptually meaningful regions using an image segmentation technique and classify these regions based on image statistics. l They then use AI common-sense reasoning techniques to find ambiguities and anomalies within an image as well as perform reasoning across a corpus of images to identify a semantically based candidate list of potential fraudulent images. l They claim their method introduces a novel framework for forensic reasoning, which allows detection of image tampering, even with nearly flawless mathematical techniques.

Technical Approach l Image Segmentation - Segment the source into regions of importance - Compare across images in a corpus l Classification - Segment based classification l Common sense reasoning - Handles classification ambiguities

Summary and Directions l Introduces a hybrid method for image forensics. - Given a subset of a corpus as a suspicious candidate set, analyze the candidates through specific metrics that are optimized to find fakery given the image’s qualitative classification. This use of common-sense reasoning goes l Directions - To integrate the facts discovered in a photo corpus to help identify what evidence may be missing as well as what fact might be unique to this scenario. - ??

Appendix l Image Annotation and Classification: Lecture #24