Validation related issues Auger-Lecce, 10 November 2009  BuildBot- introduction   Site Wide Installation  Issues related to Install/Config/Valid.

Slides:



Advertisements
Similar presentations
Low level CASE: Source Code Management. Source Code Management  Also known as Configuration Management  Source Code Managers are tools that: –Archive.
Advertisements

SubVersioN – the new Central Service at DESY by Marian Gawron.
Linux Operations and Administration
Introduction to SharePoint Development with VS2010 Paul Yuknewicz Lead Program Manager
INTRODUCTION TO WEB DATABASE PROGRAMMING
Introduction to HP LoadRunner Getting Familiar with LoadRunner >>>>>>>>>>>>>>>>>>>>>>
Abstract The automated multi-platform software nightly build system is a major component in the ATLAS collaborative software organization, validation and.
Dynodroid: An Input Generation System for Android Apps
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
1 Lecture 19 Configuration Management Software Engineering.
1 Apache. 2 Module - Apache ♦ Overview This module focuses on configuring and customizing Apache web server. Apache is a commonly used Hypertext Transfer.
NICOS System of Nightly Builds for Distributed Development Alexander Undrus CHEP’03.
INFSOM-RI Juelich, 10 June 2008 ETICS - Maven From competition, to collaboration.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
Version control Using Git Version control, using Git1.
CMSBrownBag,05/29/2007 B.Mangano How to “use” CMSSW on own Linux Box and be happy In this context “use” means: - check-out pre-compiled CMSSW code - run.
Microsoft ® Business Solutions–Navision ® 4.0 Development II - C/SIDE Solution Development Day 5.
Computer Science and Engineering The Ohio State University  Widely used, especially in the opensource community, to track all changes to a project and.
1Offline Weekly Meeting May May 2006 AliRoot Build Integration and (Testing) System Peter Hristov Vagner Morais.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
Write-through Cache System Policies discussion and A introduction to the system.
CSE 219 Computer Science III CVS
DPA follow-up review, Malargüe, Nov/2000 DPA Data Processing and Analysis Task Follow-up Review Malargüe, November 14, 2000.
1 / 22 AliRoot and AliEn Build Integration and Testing System.
A Technical Validation Module for the offline Auger-Lecce, 17 September 2009  Design  The SValidStore Module  Example  Scripting  Status.
G4MICE Status and Plans 1M.Ellis - CM24 - RAL - 31st May 2009  Firstly, a correction to the agenda:  I failed to spot a mistake in the agenda that I.
Configuration Database David Forrest 15th January 2009 CM23, HIT, Harbin.
GAAIN Virtual Appliances: Virtual Machine Technology for Scientific Data Analysis Arihant Patawari USC Stevens Neuroimaging and Informatics Institute July.
Apache JMeter By Lamiya Qasim. Apache JMeter Tool for load test functional behavior and measure performance. Questions: Does JMeter offers support for.
Migration from SL4 to SL5 at CC-Lyon J ü rgen Brunner.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
AliRoot survey P.Hristov 11/06/2013. Offline framework  AliRoot in development since 1998  Directly based on ROOT  Used since the detector TDR’s for.
Grid Deployment Enabling Grids for E-sciencE BDII 2171 LDAP 2172 LDAP 2173 LDAP 2170 Port Fwd Update DB & Modify DB 2170 Port.
CPSC 871 John D. McGregor Change management Module 2 Session 3.
Module 12: Configuring and Managing Storage Technologies
Technical Validation The Technical Validation is a testing framework of the AUGER offline to monitor the code development process. It is not a validation.
UHCS 2005, slide 1 About Continuous Integration. UHCS 2005, slide 2 Why do you write Unit Test ? Improve quality/robustness of your code Quick feedback.
MICE CM28 Oct 2010Jean-Sebastien GraulichSlide 1 Detector DAQ o Achievements Since CM27 o DAQ Upgrade o CAM/DAQ integration o Online Software o Trigger.
Yannick Patois - Datagrid Software Repository Presentation - March, n° 1 Datagrid Software Repository Presentation CVS, packages and automatic.
Linux Operations and Administration
(1) Introduction to Continuous Integration Philip Johnson Collaborative Software Development Laboratory Information and Computer Sciences University of.
Unit 17: SDLC. Systems Development Life Cycle Five Major Phases Plus Documentation throughout Plus Evaluation…
SPI NIGHTLIES Alex Hodgkins. SPI nightlies  Build and test various software projects each night  Provide a nightlies summary page that displays all.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen Budapest
Geant4 is a toolkit to simulate the passage of particles through matter, and is widely used in HEP, in medical physics and for space applications. Ongoing.
Transition to SVN server: follow up P.Hristov 24/01/2008.
BESIII Offline Software Development Environment Ma qiumei * Development environment * Configuration & management tool * Software development.
INFSO-RI Enabling Grids for E-sciencE Using of GANGA interface for Athena applications A. Zalite / PNPI.
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
1 G4UIRoot Isidro González ALICE ROOT /10/2002.
Basic Setup of OpenMeetings Version Three Styles for Setup Style1 : Starting from Scratch – Starting from svn check out and compiling with ant.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Monitoring Dynamic IOC Installations Using the alive Record Dohn Arms Beamline Controls & Data Acquisition Group Advanced Photon Source.
AliRoot survey: Calibration P.Hristov 11/06/2013.
Testing framework for BlackFin on uClinux Grace Pan.
AAF tips and tricks Arsen Hayrapetyan Yerevan Physics Institute, Armenia.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
ANALYSIS TRAIN ON THE GRID Mihaela Gheata. AOD production train ◦ AOD production will be organized in a ‘train’ of tasks ◦ To maximize efficiency of full.
Computer & Network Security
Getting & Running EdgeX Docker Containers
Integrating ArcSight with Enterprise Ticketing Systems
Integrating ArcSight with Enterprise Ticketing Systems
Installation of the ALICE Software
Maven 04 March
NA61/NA49 virtualisation:
Work report Xianghu Zhao Nov 11, 2014.
Building and Testing using Condor
X in [Integration, Delivery, Deployment]
JENKINS TIPS Ideas for making your life with Jenkins easier
Lecture 34: Testing II April 24, 2017 Selenium testing script 7/7/2019
Presentation transcript:

Validation related issues Auger-Lecce, 10 November 2009  BuildBot- introduction   Site Wide Installation  Issues related to Install/Config/Valid  Updates on ValidationTests  Conclusions and Outlook

BuildBot – Introduction BuildBot is the system used in Auger to automate the compile/test cycle to validate code changes. By automatically rebuilding and testing the tree each time something has changed, build problems are pinpointed quickly. By running the builds on a variety of platforms, developers who do not have the facilities to test their changes everywhere before checkin will at least know shortly afterwards whether they have broken the build or not. The overall goal is to reduce tree breakage and provide a platform to run tests or code-quality checks. The Validation environment uses BuildBot as “testing automated framework”. Buildbot works in a master/slave daemons scheme. The master receives notification changes from the SVN server and tells the buildslaves to checkout, build and test the code. Multiple slaves can run on different platforms. The slaves report their results to the master, which posts them in a waterfall display and sends an to the appropriate person(s) in case problems are found.

pbsfarm Setting up BuildBot slaves on our nodes allow to automatically test the build/test process on our system platforms. A system virtual machine provides a complete System Platform which supports the execution of a complete Operating System (OS). On pbsfarm, 2 system virtual machines have been set up.: auger-le64.le.infn.it Operating System: Scientific Linux 4.7 Architecture: 64bit(x86-64) auger-le32.le.infn.it Operating System: Scientific Linux 4.7 Architecture: 32bit(i386) They emulates the pbsfarm real nodes used for simulation/reconstruction. The idea behind is to have BuildBot running on it, using a “site-wide” installation.

Site Wide Installation Using APE. Installation done from the virtual machines and located under nexus06. For using it: In your.bashrc includes For 64 bit architecture: export PATH=/nfs/argo/nexus06/gabriella/AugerOffline64Last/ape-0.98/:${PATH} export APERC=/nfs/argo/nexus06/gabriella/AugerOffline64Last/ape-0.98/ape.rc For 32 bit architecture export PATH=/nfs/argo/nexus06/gabriella/AugerOffline32Last/ape-0.98/:${PATH} export APERC=/nfs/argo/nexus06/gabriella/AugerOffline32Last/ape-0.98/ape.rc At log, for configuring the environment you need to do a: eval `ape sh Externals` (for setting only the Externals) eval `ape sh Offline` (for offline settings) NOTE It works also for tcsh using eval `ape csh Externals` eval `ape csh Offline` after setting in.tcshrc the equivalent of export (setenv PATH.... setenv APERC)

Problems during Aires build/install. (ape-0.98) In #ape-0.98/ape.rc... [package Aires] fc = g77... Should allow the setting of g77 as compiler in use, but it does not work. The compilation stops since the gfortran (default compiler) is not found. I manually changed the compiler setting directly in build/Aires/2-8-4a/config (mods FortCompile=“g77” ) and then I entered the command build/Aires/2-8-4a/doinstall 0 Apparently things were OK but in the auger-offline-config the build of Aires introduces a set of libraries in the system area, that address a boost installed in the system that conflicts with the Boost in external, crash at run-time. Solved changing manually the auger-offline-config. TRAC (#34) It is MANDATORY to have $APERC set Installation/Configuration

After a few rounds of validation on le-32 le-64 (see waterfall In some cases the StandardApplications are very slow (particularly on le-32) and the buildbot-master kills the application otherwise lasting forever. The problem seems to be worse since a few days. Apparently no mods related. The StandardApplications run involves full Sd simulation, starting from a Corsika air- shower, and randomize the core position on the array. It can sometimes happens that a core lands very close to a tank. In such a case an enormous amount of particles is run through Geant (... It is not worth simulating them in such details since those stations are in any case “saturated”...).... Only a luck of luckiness sequence?!... (Notice the SdSim events are never reproduced)... The example and standardApplication running shows a difference between le-32 le- 64. In (le-64) severals: FDTriggerSimulatorOG:MakeMirrorEvent...TAnalysedPixelData::Analyse() – found invalid 0x7f pattern! Seen also by Mariangela- Present also in other 64 bits build machines (see example in waterfall)- Requests to Tom Paul... +Ralf Ulrich...+ Michael Unger and Steffen Mueller...(FDEventLib... responsability)... + HJ Mathes..... Validation

Mods for Module Sequences: (used StandardApplications -data Reconstruction- as reference) PLEASE CHECK! ValidationTests EventFileReaderOG EventCheckerOG FdCalibratorOG FdPulseFinderOG PixelSelectorOG FdSDPFinderOG FdAxisFinderOG FdApertureLightOG FdProfileReconstructo rKG RecDataLister RecDataWriter EventFileExporterOG SValidStore FRec EventFileReaderOG EventCheckerOG SdCalibratorOG SdEventSelectorOG SdPlaneFitOG LDFFinderOG SdEventPosteriorSelectorOG SdRecPlotterOG RecDataLister RecDataWriter EventFileExporterOG SValidStore SRec with this Module Sequences: the code is working. To do- update the input- event before commit

IO work- Main idea: checking that new releases of Offline can read files produced with older versions. How to approach this: Trigger the BuildBot build on EventIO change. As Input – A list of reference Events with different versions A script running a read test A script running the hybrid Simulation+Reconstruction + writing the Event + reco/sim test. ValidationTests TAG 1 I/O TAG 2 I/O TAG …I/O DEV N I/O TAG N-1 I/O Code 1 Code 2 Code... Code N-1 Code DEV Sim Rec ref

Conclusions and Outlook 2 BuildBot(slave) have been set up. They allows to automatically test the build/test process on the system platforms we use. The use of a site wide installation from emulating node machines running BuildBot maximize the pinpointing of problems from our side. (The build is from the trunk with fixed externals). An offline reference is available. Possible evolution of virtualization- Worker Node on demand for GRID(?) – A possible conservative approach: Check feasibility then do OK/ then propose to the collaboration. What is the status of porting OFFLINE on GRID? The First issues from Installation, Buildbot setting and Validation are under study. For the old SREC FREC Validation tests. The ModuleSequence has been modified in order to update. Code working Feedback needed!