Wir schaffen Wissen – heute für morgen Paul Scherrer Institut Timo Korhonen Improvements to Indexing Tool (Channel Archiver) ‏ EPICS Meeting, BNL 2010.

Slides:



Advertisements
Similar presentations
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Advertisements

Instructor: Sazid Zaman Khan Lecturer, Department of Computer Science and Engineering, IIUC.
May 7, A Real Problem  What if you wanted to run a program that needs more memory than you have?
Wir schaffen Wissen – heute für morgen Babak Kalantari, PSI MRF workshop, Prague, Eli Beamlines Paul Scherrer Institut SwissFEL Timing System.
Wir schaffen Wissen – heute für morgen Paul Scherrer Institut Controls tools for operation management at the Swiss Light Source at the EPICS collaboration.
CSC 213 – Large Scale Programming. Today’s Goals  Look at how Dictionary s used in real world  Where this would occur & why they are used there  In.
Nov COMP60621 Concurrent Programming for Numerical Applications Lecture 6 Chronos – a Dell Multicore Computer Len Freeman, Graham Riley Centre for.
Memory Management (II)
CS 333 Introduction to Operating Systems Class 18 - File System Performance Jonathan Walpole Computer Science Portland State University.
SECTIONS 13.1 – 13.3 Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin SECONDARY STORAGE MANAGEMENT.
SECTIONS 13.1 – 13.3 Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin SECONDARY STORAGE MANAGEMENT.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts Amherst Operating Systems CMPSCI 377 Lecture.
BTREE Indices A little context information What’s the purpose of an index? Example of web search engines Queries do not directly search the WWW for data;
Optimizing RAM-latency Dominated Applications
EPICS Archiving Appliance Test at ESS
Modularizing B+-trees: Three-Level B+-trees Work Fine Shigero Sasaki* and Takuya Araki NEC Corporation * currently with 1st Nexpire Inc.
Computing hardware CPU.
From the ChannelArchiver to the Best Ever Archive Utility, Yet July 2009.
July, 2001 High-dimensional indexing techniques Kesheng John Wu Ekow Otoo Arie Shoshani.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
Chord & CFS Presenter: Gang ZhouNov. 11th, University of Virginia.
Stanford Linear Accelerator Center R. D. Hall1 EPICS Collaboration Mtg Oct , 2007 Oracle Archiver Past Experience Lessons Learned for Future EPICS.
Wir schaffen Wissen – heute für morgen Gateway (Redux) PSI - GFA Controls IT Alain Bertrand Renata Krempaska, Hubert Lutz, Matteo Provenzano, Dirk Zimoch.
Building a Real Workflow Thursday morning, 9:00 am Greg Thain University of Wisconsin - Madison.
1 Components of the Virtual Memory System  Arrows indicate what happens on a lw virtual address data physical address TLB page table memory cache disk.
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
Wir schaffen Wissen – heute für morgen Paul Scherrer Institut EPICS V4 Archiver Service and Matlab client Timo Korhonen.
Wir schaffen Wissen – heute für morgen Paul Scherrer Institut Windows-based camera IOC with a SMB branch Helge Brands.
Week 7 : Chapter 7 Agenda SQL 710 Maintenance Plan:
DMBS Internals I. What Should a DBMS Do? Store large amounts of data Process queries efficiently Allow multiple users to access the database concurrently.
Computer Systems Week 14: Memory Management Amanda Oddie.
Marwan Al-Namari Hassan Al-Mathami. Indexing What is Indexing? Indexing is a mechanisms. Why we need to use Indexing? We used indexing to speed up access.
Wir schaffen Wissen – heute für morgen Electronic Logbooks Revisited Paul Scherrer Institute Andreas Lüdeke Presented at the ARW 2011.
CS333 Intro to Operating Systems Jonathan Walpole.
Swap Space and Other Memory Management Issues Operating Systems: Internals and Design Principles.
Lecture Topics: 11/24 Sharing Pages Demand Paging (and alternative) Page Replacement –optimal algorithm –implementable algorithms.
Introduction: Memory Management 2 Ideally programmers want memory that is large fast non volatile Memory hierarchy small amount of fast, expensive memory.
Multilevel Caches Microprocessors are getting faster and including a small high speed cache on the same chip.
M.Kersten MonetDB, Cracking and recycling Martin Kersten CWI Amsterdam.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
EPICS Release 3.15 Bob Dalesio May 19, Features for 3.15 Support for large arrays Channel access priorities Portable server replacement of rsrv.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
ORACLE & VLDB Nilo Segura IT/DB - CERN. VLDB The real world is in the Tb range (British Telecom - 80Tb using Sun+Oracle) Data consolidated from different.
DMBS Internals I February 24 th, What Should a DBMS Do? Store large amounts of data Process queries efficiently Allow multiple users to access the.
DMBS Internals I. What Should a DBMS Do? Store large amounts of data Process queries efficiently Allow multiple users to access the database concurrently.
Wir schaffen Wissen – heute für morgen Babak Kalantari, PSI MRF workshop, Prague, Eli Beamlines Paul Scherrer Institut Synchronous DAQ using.
R. Krempaska, October, 2013 Wir schaffen Wissen – heute für morgen Controls Security at PSI Current Status R. Krempaska, A. Bertrand, C. Higgs, R. Kapeller,
Lecture Topics: 11/22 HW 7 File systems –block allocation Unix and NT –disk scheduling –file caches –RAID.
Wir schaffen Wissen – heute für morgen 30. Mai 2016PSI,30. Mai 2016PSI, Paul Scherrer Institut The PSI web interface to the EPICS channel archiver Gaudenz.
1 Contents Memory types & memory hierarchy Virtual memory (VM) Page replacement algorithms in case of VM.
WIR SCHAFFEN WISSEN – HEUTE FÜR MORGEN Transition to user operation Didier Voulot :: Paul Scherrer Institut SwissFEL Commissioning Workshop, 22 March2016.
WIR SCHAFFEN WISSEN – HEUTE FÜR MORGEN Intra Pulse Phase Control Manuel Brönnimann :: Paul Scherrer Institut LLRF Workshop 2015.
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
Disk Cache Main memory buffer contains most recently accessed disk sectors Cache is organized by blocks, block size = sector’s A hash table is used to.
Jonathan Walpole Computer Science Portland State University
ECE232: Hardware Organization and Design
Memory COMPUTER ARCHITECTURE
A Real Problem What if you wanted to run a program that needs more memory than you have? September 11, 2018.
Lecture 16: Data Storage Wednesday, November 6, 2006.
FileSystems.
Steve Ko Computer Sciences and Engineering University at Buffalo
Introduction to Computers
Mirjam van Daalen, (Stephan Egli, Derek Feichtinger) :: Paul Scherrer Institut Status Report PSI PaNDaaS2 meeting Grenoble 6 – 7 July 2016.
Sanuja Dabade & Eilbroun Benjamin CS 257 – Dr. TY Lin
Steve Ko Computer Sciences and Engineering University at Buffalo
CSCE 313 – Introduction to UNIx process
Files Management – The interfacing
Database Recovery 1 Purpose of Database Recovery
THE GOOGLE FILE SYSTEM.
Cache writes and examples
Presentation transcript:

Wir schaffen Wissen – heute für morgen Paul Scherrer Institut Timo Korhonen Improvements to Indexing Tool (Channel Archiver) ‏ EPICS Meeting, BNL 2010

10/11/10EPICS Meeting, BNL, Currently four different archive servers are in use. –SLS Accelerator data: slsmcarch (machine archive server; HP, Xenon quadcore 2.66 GHz, 32 GB RAM)‏ Long Term: since January 2001; channels; 70 GB Medium Term: 6 months; channels; 120 GB Short Term Archiver: 14 days; channels; 114G GB Post Mortem Archiver: Stores the last famous words Total available disc space for data: 500 GB –SLS Beamline data: slsblarch (beamline archive server; HP, AMD Opteron dualcore 1.8 GHz; 6 GB RAM)‏ Long and short term archivers for every beamline (total 29 Engines)‏ Short term archivers store data up to 12 months Total amount of data: 163 GB / 384 GB Channel Archiver at PSI

10/11/10EPICS Meeting, BNL, archive servers (cont) – PSI (office) data: gfaofarch‏ –Long Term Archiver: Stores data since January 2006 –Medium and Short Term Archivers ZHE Cyclotron High Energy –Long (since April 2008) –Medium and short term –SwissFEL: felarch1 (HP, Quadcore 2.66 GHz, 10 G RAM) Small teststand OBLA –638 channels, 2.1 Terabytes! »Waveforms, images FIN250 test injector –LT, MT and ST (.6, 7.9 and 464 GB) Channel Archiver at PSI

10/11/10EPICS Meeting, BNL, –The archive engines are running stable –The problems we have had are on the retrieval side –Indexing is used to speed up retrieval Indexes on daily files Master index on the whole archived data –We need the performance The SwissFEL test machine is going to produce a lot of data –Waveforms, images –We need to archive more than in a production machine –For us, there is no need for (immediate) change We would like to keep the channel archiver going –Updates, bugfixes –Retrieval tools »Waveform viewer, etc have been developed »Matlab export would be welcome Indexing tools need work Channel Archiver at PSI

10/11/10EPICS Meeting, BNL, Background –The ArchiveIndexTool is used at PSI in the night between Saturday and Sunday each week to create master indexes for the midterm archive. –Indexing is essential for good retrieval performance –The tool produces many errors when run on the EPICS archive indices to produce or to update the master index. Disclaimer: I know very little about this, I just tell what the people who work on this have reported. –Involved people: Gaudenz Jud (archiver maintenance, operation and development)‏ Hans-Christian Stadler (PSI IT, Scientific Computing) is investigating the issue together with Gaudenz Index Tool improvements

10/11/10EPICS Meeting, BNL, Findings so far: –After investigating an error log: From the code it is clear that the ArchiveEngine and the ArchiveIndexTool are not supposed to be used concurrently on the same indices. Running them concurrently does produce errors – but not those we see in production. –the errors seem to only occur on the production machine, when there is a high load and a lot of disk activity. –try a quick fix: a retry mechanism on the highest level. All index files are closed and reopened after a delay. This quick fix seems to work so far. Index Tool improvements

10/11/10EPICS Meeting, BNL, Observations: –The RTree implementation does not allow concurrent read/write access. It might be possible to arrange the file operations in a way that allows concurrent access when the index is stored on a strictly POSIX compliant file system. –The RTree implementation has a RTree node "cache" that only grows. Nodes are never evicted from the cache. I'm implementing a new LRU node cache with a fixed number of entries to see if this reduces system load. –The RTree implementation uses many small disk operations (see example code above). A reimplementation should use large disk transfers. –The RTree implementation is like a B-Tree, but does not adjust the node size to the disk sector size for improved I/O performance.

10/11/10EPICS Meeting, BNL, Observations (continued): –The RTree implementation is not optimal for the use case seen at SLS, where data is inserted at the end only. This leads to a reduced fill level of the nodes. The RTree maintains the invariant, that only the root node may be filled less than 1/2. In addition to that data is moved between nodes too often, leading to many random accesses on disk. A reimplementation should feature a datastructure that is optimal for appends at the end.

10/11/10EPICS Meeting, BNL, Conclusions so far: –Finding out the real reason for the errors is a time consuming process. The real reason for the errors has not yet been identified. –the offsets zu Data structures in index get corrupted. However, it is not clear where. – Because the corruption only happens when the load on the production system is high, logical errors in the normal execution path can be almost certainly excluded. –The experience so far suggests that a new implementation of the RTree Code could solve a number of problems

10/11/10EPICS Meeting, BNL, Thank you for your attention!