DataGrid is a project funded by the European Commission under contract IST-2000-25182 QAG activity report On behalf of the QAG

Slides:



Advertisements
Similar presentations
Tony Doyle GridPP2 Proposal, BT Meeting, Imperial, 23 July 2003.
Advertisements

Current methods for negotiating firewalls for the Condor ® system Bruce Beckles (University of Cambridge Computing Service) Se-Chang Son (University of.
4 th DataGRID Project Conference, Paris, 5 March 2002 Testbed Software Test Plan I. Mandjavidze on behalf of L. Bobelin – CS SI; F.Etienne, E. Fede – CPPM;
Configuration Management
Software Quality Assurance Plan
Software Failure: Reasons Incorrect, missing, impossible requirements * Requirement validation. Incorrect specification * Specification verification. Faulty.
Software Quality Assurance Inspection by Ross Simmerman Software developers follow a method of software quality assurance and try to eliminate bugs prior.
The CrossGrid project Juha Alatalo Timo Koivusalo.
27-29 September 2002CrossGrid Workshop LINZ1 USE CASES (Task 3.5 Test and Integration) Santiago González de la Hoz CrossGrid Workshop at Linz,
Introduction to z/OS Basics © 2006 IBM Corporation Chapter 8: Designing and developing applications for z/OS.
Chapter 3: The Project Management Process Groups
EC Review – 01/03/2002 – G. Zaquine – Quality Assurance – WP12 – CS-SI – n° 1 DataGrid Quality Assurance Gabriel Zaquine Quality Engineer - WP12 – CS-SI.
Handouts Software Testing and Quality Assurance Theory and Practice Chapter 11 System Test Design
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 DataGrid Quality Assurance On behalf.
New Challenges in Cloud Datacenter Monitoring and Management
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Andrew McNab - Manchester HEP - 26 June 2001 WG-H / Support status Packaging / RPM’s UK + EU DG CA’s central grid-users file grid “ping”
EGEE is a project funded by the European Union under contract IST JRA1 Testing Activity: Status and Plans Leanne Guy EGEE Middleware Testing.
Objectives Understand the basic concepts and definitions relating to testing, like error, fault, failure, test case, test suite, test harness. Explore.
© 2012 IBM Corporation Rational Insight | Back to Basis Series Chao Zhang Unit Testing.
RUP Implementation and Testing
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
EMI INFSO-RI EMI SA2 Report Quality Assurance Alberto Aimar (CERN) SA2 WP Leader.
C. Loomis – Testbed Status – 28/01/2002 – n° 1 Future WP6 Tasks Charles Loomis January 28, 2002
EMI INFSO-RI EMI Quality Assurance Processes (PS ) Alberto Aimar (CERN) CERN IT-GT-SL Section Leader EMI SA2 QA Activity Leader.
EMI INFSO-RI SA2 - Quality Assurance Alberto Aimar (CERN) SA2 Leader EMI First EC Review 22 June 2011, Brussels.
Crossgrid kick-off meeting, Cracow, March 2002 Santiago González de la Hoz, IFIC1 Task 3.5 Test and Integration (
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
DataGrid WP1 Massimo Sgaravatto INFN Padova. WP1 (Grid Workload Management) Objective of the first DataGrid workpackage is (according to the project "Technical.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 WP1 activity, achievements and plans.
Grid Workload Management Massimo Sgaravatto INFN Padova.
(Spring 2015) Instructor: Craig Duckett Lecture 10: Tuesday, May 12, 2015 Mere Mortals Chap. 7 Summary, Team Work Time 1.
HiLumi LHC is co-funded by the EU FP7 Capacities Programme, Grant Agreement Svet Stavrev (EU Projects Office, CERN) Administrative Manager 17.
GLite – An Outsider’s View Stephen Burke RAL. January 31 st 2005gLite overview Introduction A personal view of the current situation –Asked to be provocative!
JRA Execution Plan 13 January JRA1 Execution Plan Frédéric Hemmer EGEE Middleware Manager EGEE is proposed as a project funded by the European.
LCG EGEE is a project funded by the European Union under contract IST LCG PEB, 7 th June 2004 Prototype Middleware Status Update Frédéric Hemmer.
INFSO-RI JRA 1 Testbed Management Technologies Alain Roy (University of Wisconsin-Madison, USA) ETICS 2 Final Review Brussels - 11 May 2010.
DataGRID PTB, Geneve, 10 April 2002 Testbed Software Test Plan Status Laurent Bobelin on behalf of Test Group.
“Reporting Procedure for I3” Seadatanet kick-off Heraklion, June 8-9, 2006 Lorenza Saracco, European Commission DG Research.
INFSO-RI Enabling Grids for E-sciencE Information and Monitoring Status and Plans Plzeň, 10 July 2006 Steve Fisher/RAL.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
8 th CIC on Duty meeting Krakow /2006 Enabling Grids for E-sciencE Feedback from SEE first COD shift Emanoil Atanassov Todor Gurov.
JRA2: Quality Assurance Overview EGEE is proposed as a project funded by the European Union under contract IST JRA.
Assoc. Prof. Dr. Ahmet Turan ÖZCERİT.  System and Software  System Engineering  Software Engineering  Software Engineering Standards  Software Development.
G.Govi CERN/IT-DB 1 September 26, 2003 POOL Integration, Testing and Release Procedure Integration  Packages structure  External dependencies  Configuration.
A. Aimar - EP/SFT LCG - Software Process & Infrastructure1 SPI Software Process & Infrastructure for LCG Project Overview LCG Application Area Internal.
European Middleware Initiative (EMI) The Software Engineering Model Alberto Di Meglio (CERN) Interim Project Director.
Yannick Patois - Datagrid Software Repository Presentation - March, n° 1 Datagrid Software Repository Presentation CVS, packages and automatic.
EMI INFSO-RI Software Quality Assurance in EMI Maria Alandes Pradillo (CERN) SA2.2 Task Leader.
Summary from WP 1 Parallel Section Massimo Sgaravatto INFN Padova.
Accounting in DataGrid HLR software demo Andrea Guarise Milano, September 11, 2001.
DataGrid is a project funded by the European Commission under contract IST EDG Baseline API Document Document build description and current.
JRA1 Meeting – 09/02/ Software Configuration Management and Integration EGEE is proposed as a project funded by the European Union under contract.
EMI INFSO-RI SA2: Quality Assurance Status Report Alberto Aimar(SA2) SA2 Leader EMI First EC Review 22 June 2011, Brussels.
WP3 WP3 at Budapest 2/9/2002 Steve Fisher / RAL. WP3 Steve Fisher/RAL - 2/9/2002WP3 at Budapest2 Summary News –EDG Retreat –EDG Tutorials –Quality –Release.
Maite Barroso – WP4 Workshop – 10/12/ n° 1 -WP4 Workshop- Developers’ Guide Maite Barroso 10/12/2002
Comments on SPI. General remarks Essentially all goals set out in the RTAG report have been achieved. However, the roles defined (Section 9) have not.
EGEE is a project funded by the European Union under contract IST Report from the PTF Fabrizio Pacini Datamat S.p.a. Milan, IT-CZ JRA1 meeting,
Project Management PTM721S
Regional Operations Centres Core infrastructure Centres
Software Engineering (CSI 321)
EGEE Middleware Activities Overview
DataGrid Quality Assurance
EO Applications Parallel Session
Chapter 18 Maintaining Information Systems
Leanne Guy EGEE JRA1 Test Team Manager
Introduction to Software Engineering
Presentation transcript:

DataGrid is a project funded by the European Commission under contract IST QAG activity report On behalf of the QAG

QAG Activity Report - n° 2 Summary u Bugzilla anomalies follow-up u Bugzilla procedure amendment u Release procedure QA checklist u Final deliverables & Proposal for the guidelines of the final deliverables report u Performance indicators Definition & Implementation

QAG Activity Report - n° 3 Bugzilla anomalies follow-up until mid September Better use of Bugzilla

QAG Activity Report - n° 4 Bugzilla: Zoom on release 2 MTTR(Mean Time To Repair) during the period No critical bug

QAG Activity Report - n° 5 Bugzilla procedure amendment u The main improvement concerns the fact that the bug is automatically driven back until the originator. NEW Originator Component owner ITeam member Originator The bug is accepted by the component owner The bug is solved with a code modification in cvs, tested, and the affected RPM is autobuilt. ASSIGNED The bug is kept in this state till the code is included in a tagged release to be deployed in the testbed. The IT member must write the release tag in the log. RESOLVED The originator will wait till the deployment of the release and then check if the bug is solved. If not, reopen the bug. CLOSED VERIFIED BUG STATUSPrerequisites for status change:Assigned to:

QAG Activity Report - n° 6 Release procedure QA checklist u Describes the basic checks that has to be fulfilled for a component release verification. u Adhere to the developers’ guide: e.g. Public interfaces should be in separate packages and versioned independently; package naming; Everything in central CVS; Validate if the correct version of the RPM is the flaged one in autobuild status page; Validate if the RPM is in RMP repository. u Documentation : Verify Licence file; Readme file; Install file; Authors & Maintainers file; Dependencies file; Post-install doc; Code doc (Is there javadoc, doxygen,….); End user documentation; Installation guide; Man pages (There must be a man page for each executable and you may provide one for configuration files). u Tests: test report according to the test plans including numbers on performance and scalability. Verify if each item of the test plan is covered in the test report. Give the test report’s URL. u Easily configurable to allow a setup/usage in release 2.0 mode. Verify if new functionalities are optional.  Api changes (and new apis) must follow the api change procedure. If the API changed: give the list; has ATF procedure been followed

QAG Activity Report - n° 7 QA checklist: release 2.x u VOMS - Slot: Starts week of September 1 st, ends September 12 th n QAG checklist delivered on September 12 th n u LCMAPS - Slot: Starts September 10 th, ends September 19 th n QAG checklist delivered on September 10 th n u RLS - Slot: Starts September 16 th, ends September 25 th n QAG checklist delivered on September 19 th n

QAG Activity Report - n° 8 Deliverables for PM33 u Title: User documentation for release 2.x (the production one) n D6.7=(D1.6,2.5,3.5,4.5,5.5,6.7) n This deliverable is presented as a set of links gathering together the various user guides from WPs (1,2,3,4,5,6,7): s Installation guide s Users guide s Developer guide u Editor: Eric Fede u Timescale: n The various link should be sent to Eric before 6/10/2003 n The review will start 6/10/2003 n The document should be sent to EU before the end of October u Reviewers: Users Guides: WP8 (D. Boutigny), Installation Guides (WP6): Site Administrators

QAG Activity Report - n° 9 QR11 Quarterly report for PM33 u D12.15: Quarterly report – Period: July to September 2003 u Timescale: n Should be sent to Gabriel before 14/10/2003 n The document should be sent to EU before the end of October

QAG Activity Report - n° 10 Deliverables for PM36 u Timescale: n The review process will start 24/11/2003 n PTB approval 15/12/2003 n The document should be sent to EU before the end of January 2004 u Moderators & reviewers: To be nominated D1.7Final evaluation report D2.6Final evaluation report D3.6Final evaluation report D4.6Final evaluation report D5.6Final evaluation report D6.8Final evaluation of testbed operation D7.4Final Report on network infrastructure and services D7.7Security report on final project releases D8.4Report on results of HEP applications at Run #2 and Final Application Report D9.5EO application processing testbed demonstration and final report D10.4Final report including report on the 2nd bio-testbed release D11.7Final conference D11.9Contribution to international standards

QAG Activity Report - n° 11 Final Quarterly & Annual report u D12.16: QR12 Quarterly report – Period: October to December 2003 u D12.19: Third Annual report u Timescale: n Should be sent to Gabriel before 14/01/2004 n Should be sent to EU before the end of January 2004

QAG Activity Report - n° 12 Proposal for the guidelines of the final deliverables report u The motto of the deliverables should be: 'what has been achieved - what needs to be done in future' u If applicable, the deliverables should relate to the first 2 deliverables of the project (Dx.1:Current technology report, Dx.2: Design documents), and report fulfilments and deviations (explaining why) of the original plans. u A thorough evaluation of the WP achievements, again if possible using criteria defined in the first 2 deliverables should also be part of it. u Finally, a discussion of future developments, in particular pointing out what seems to be a promising way and what not, is needed. u So, there should be three major chapters: n Achievements n Evaluation n Future Directions

QAG Activity Report - n° 13 There is a need to quantify how well grid projects are working Both in terms of providing the user with the service they require and in terms of utilising the resources available to it an efficient manner. Traditional benchmarking is inadequate, its works well in a homogeneous stable environment. Less well in \heterogeneous dynamic environment. Rough draft at Performance and definitions of Efficiency

QAG Activity Report - n° 14 To this end we described a set of “efficiencies” Properties of the efficiencies: -efficiency of 1.0 should reflect the ideal case. -efficiency of 0.0 should reflect that it does not work. -(Ideally) the efficiencies should be linear. Performance and definitions of Efficiency

QAG Activity Report - n° 15 Performance and definitions of Efficiency These efficiencies are designed to test how well the infrastructure is providing what is needed. These efficiencies should be general/non-architecture specific as possible. Essential if they are going to be used to compare between grid projects or the same project at different stages of its evolution. Obviously these efficiencies will depend on the nature of the job These should be measured during real operation when things really do go wrong. Results shown will be from the 1.4 Testbed with Jobs run through the Imperial College RB. Lots of help by Paul Crosby from UCL

QAG Activity Report - n° 16 User Efficiencies Most importantly, the jobs should complete returning the correct output. Easily turned into an efficiency: This should be measured applications testers, however can be estimated from the information in the LB. Sadly, this was only ~0.6 for release 1.4

QAG Activity Report - n° 17 User Efficiencies The reasons for failure were: -Information systems could not cope with the scale of resources causing jobs not to be matched even when resources were available -Known problem with the compiler version that was used causing problems with the jobs database. Cleaning caused all current jobs to be lost. -Limit of version and method of use of CondorG. If reached, cleaning caused all running jobs to be lost -Many smaller problems However what is important is to the user is the base figure of 0.6

QAG Activity Report - n° 18 User Efficiencies Hopefully, in release 2. Then a more subtle definition is required. “From a user perspective an efficiency of 1 corresponds to all their jobs running immediately on resources of sufficient scale and which have instant access to all the data required by the jobs.” “It is important to note that even a very large, centralised computing centre would not achieve perfect efficiency as there will always be a certain overhead taken to by the batch system to process the job etc.”

QAG Activity Report - n° 19 In a heterogeneous system there are many subtle problems with this definition. It is also open to user stupidity in their Job Definition However, it was the best working definition that we come up come up with. Captures most inefficiencies. User Efficiencies We have chosen:

QAG Activity Report - n° 20 Still many very short “Hallo World” jobs which are impossible to run efficiently… For short jobs… User Efficiencies

QAG Activity Report - n° 21 However for longer jobs… Still some jobs with low efficiencies… User Efficiencies

QAG Activity Report - n° 22 User Efficiencies As it becomes more meaningful to measure the performance from the lbserver. Hope to dump and collect the contents of these databases from the different lb machines.

QAG Activity Report - n° 23 System Efficiency System efficiency is a quantification of how well the system is utilising the resources available to it. Could be used to quantify differences between middleware projects or to show funding agencies that a distributed system really does use resources efficiently!

QAG Activity Report - n° 24 System Efficiency System Efficiency is harder to define than user efficiency, but one possible candidate for a system efficiency of 1 would be the case when all the resources are in the same location and requests for the resources are instantly dealt with, up to the limit that all the available resources were being used. In the case when the system was oversubscribed this would correspond to the fractional CPU usage and would still be meaningful for systems where this was not the case.

QAG Activity Report - n° 25 System Efficiency I.e. Perhaps an integrated value over a given period is more meaningful…

QAG Activity Report - n° 26 System Efficiency This is harder to measure, not only requires knowledge of what resources are available but also: -What resources satisfy the users needs (as specified in the jdl) -What resources is the user allowed to run on -Etc etc

QAG Activity Report - n° 27 System Efficiency To avoid assumption we (Paul Crosby) measured for a series of his own jobs. Submitted in bursts and monitoring what resources were available to him via MDS. Clearly not always 100%

QAG Activity Report - n° 28 Non-efficiency metrics Users… “build it and they will come…” But if it doesn’t work well, they will soon go away again…

QAG Activity Report - n° 29 Conclusions -Changes in Bugzilla procedure and usage -Release verification procedure -Still lots of deliverables -We are trying to quantify performance