Download presentation
Presentation is loading. Please wait.
2
Denis Caromel 1 Denis Caromel, et al. http://ProActive.ObjectWeb.org OASIS Team INRIA -- CNRS - I3S -- Univ. of Nice Sophia-Antipolis, IUF September 4 th,, DAPSYS, Debrecen, Hungary Bridging Distributed and Multi-Core Computing 1.Background: INRIA, OASIS, ActiveEon 2.Active Objects 3.Components 4.GUI: IC2D 5.Applications and Perspectives: SOA+GRID
3
Denis Caromel 2 Background
4
Denis Caromel 3 INRIA and OASIS Team Computer Science and Control 8 Centers all over France Workforce: 3 800 186 Million Euro annual budget Strong in standardization committees: – IETF, W3C, ETSI, … Strong Industrial Partnerships Foster company foundation: 90 startups so far - Ilog (Nasdaq, Euronext) - … - ActiveEon A joint team between: – INRIA, University of Nice- CNRS Created in 1999 Started the ProActive Parallel Suite Over 40 persons Distributed and Parallel: From Multi-cores to Enterprise GRIDs
5
Denis Caromel 4 Startup Company Born of INRIA Co-developing, Providing support for Open Source ProActive Parallel SuiteProActive Parallel Suite Winner 80/1000 applications (Minister of Research Contest) Several Customers (Worldwide: Boston USA, etc.)
6
Denis Caromel 5 Written in Java Features: 1. Eclipse GUI 2. Parallel+Dist. Progr. 3. Scheduling & Grids Used in production by industry
7
Denis Caromel 6 ProActive Parallel Suite (1) Open Source + PROFESSIONAL SUPPORT
8
Denis Caromel 7 ProActive Parallel Suite: GUI
9
Denis Caromel 8 ProActive Parallel Suite: GUI
10
Denis Caromel 9
11
10 ProActive Parallel Suite: Deploy
12
Denis Caromel 11 ProActive Parallel Suite: Deploy
13
Denis Caromel 12 Deploy on Various Kinds of Infrastructures Internet Clusters Parallel Machine Large Equipment Internet Job management for embarrassingly parallel application (e.g. SETI) Internet Servlets EJBsDatabases
14
Denis Caromel 13 Desktop Grid P2P: AO Overlay Network
15
Denis Caromel 14
16
Denis Caromel 15
17
Denis Caromel 16 Scheduler and Resource Manager: User Interface
18
Denis Caromel 17 Scheduler: User Interface
19
Denis Caromel 18 Video 1: Scheduler, Resource Manager
20
Denis Caromel 19 The multiple GRIDs Scientific Grids Enterprise Grids Internet Grids, (miscalled P2P grid) Intranet Desktop Grids Grid Too much associated with Internet Grid (Multi-Organizations) Too many Promises, Too Early My belief : however, several factors will make it coming slowly
21
Denis Caromel 20 What Future for GRIDs ? Enterprise Grids: End of Moore Law Multi-cores to Distribution Reduce Electric Consumption Optimize and Share Resources SOA Architectures Dynamic Placement/Scheduling Challenges: Programming Model
22
Denis Caromel 21 ProActive Parallel Suite: Program
23
Denis Caromel 22 ProActive Parallel Suite: Program
24
Denis Caromel 23 ProActive Parallel Suite: Program
25
Denis Caromel 24 2. Distributed and Parallel Active Objects ProActive Programming
26
Denis Caromel 25 A ProActive : Active objects Proxy Java Object A ag = newActive (“A”, […], VirtualNode) V v1 = ag.foo (param); V v2 = ag.bar (param);... v1.bar(); //Wait-By-Necessity V Wait-By-Necessity is a Dataflow Synchronization JVM A Active Object Future Object Request Req. Queue Thread v1 v2 ag WBN!
27
Denis Caromel 26 ProActive: Inter- to Intra- Synchronization SequentialMultithreadedDistributed Synchronizations, Behavior: not dependent upon the physical location (mapping of activities)
28
Denis Caromel 27 ProActive: First-Class Futures SequentialMultithreadedDistributed Synchronizations, Behavior: not dependent upon the physical location (mapping of activities)
29
Denis Caromel 28 Wait-By-Necessity: First Class Futures ba Futures are Global Single-Assignment Variables V= b.bar () c c c.gee (V) v v b
30
Denis Caromel 29 Standard system at Runtime: No Sharing NoC: Network On Chip Proofs of Determinism
31
Denis Caromel 30 Calculus ASP: Asynchronous Sequential Processes
32
Denis Caromel 31 Proofs in GREEK ASP Confluence and Determinacy Future updates can occur at any time, Mobility does not change behavior
33
Denis Caromel 32 TYPED ASYNCHRONOUS GROUPS
34
Denis Caromel 33 A Creating AO and Groups Typed Group Java or Active Object A ag = newActiveGroup (“A”, […], VirtualNode) V v = ag.foo(param);... v.bar(); //Wait-by-necessity V Group, Type, and Asynchrony are crucial for Cpt. and GRID JVM
35
Denis Caromel 34 Broadcast and Scatter JVM ag cg ag.bar(cg); // broadcast cg ProActive.setScatterGroup(cg) ; ag.bar(cg); // scatter cg c1 c2 c3 c1 c2 c3 c1 c2 c3 c1 c2 c3 c1 c2 c3 c1 c2 c3 s c1 c2 c3 s Broadcast is the default behavior Use a group as parameter, Scattered depends on rankings
36
Denis Caromel 35 Dynamic Dispatch Group JVM ag cg c1 c2 c3 c4 c5 c6 c7 c8c0 c9c1 c2 c3 c4 c5 c6 c7 c8c0 c9 c1 c2 c3 c4 c5 c6 c7 c8c0 c9 Slowest Fastest ag.bar(cg);
37
Denis Caromel 36 Object-Oriented SPMD Single Program Multiple Data Motivation Cluster / GRID computing SPMD programming for many numerical simulations Use enterprise technology (Java, Eclipse, etc.) for Parallel Computing Able to express most of MPI’s Collective Communications: broadcast reduce scatter allscatter gather allgather and Barriers, Topologies.
38
Denis Caromel 37 OO SPMD A ag = newSPMDGroup (“A”, […], VirtualNode) // In each member myGroup.barrier (“2D”); // Global Barrier myGroup.barrier (“vertical”); // Any Barrier myGroup.barrier (“north”,”south”,“east”,“west”); A Still, not based on raw messages, but Typed Method Calls ==> Components
39
Denis Caromel 38 Parallel, Distributed, Hierarchical 3. Components Composing
40
Denis Caromel 39 Objects to Distributed Components (1) Typed Group Java or Active Object ComponentIdentity Cpt = newActiveComponent (params); A a = Cpt ….getFcInterface ("interfaceName"); V v = a.foo(param); V A Example of component instance JVM Truly Distributed Components IoC: Inversion Of Control (set in XML)
41
Denis Caromel 40 GCM: Grid Component Model GCM Being defined in the NoE CoreGRID (42 institutions) Open Source ObjectWeb ProActive implements a preliminary version of GCM Service Oriented: NESSI relation GridCOMP takes: GCM as a first specification, ProActive as a starting point, and Open Source reference implementation. The vision: GCM to be the IT Service GSM
42
Denis Caromel 41 GridCOMP Partners
43
GCM Scopes and Objectives: Grid Codes that Compose and Deploy No programming, No Scripting, … No Pain Innovation: Abstract Deployment Composite Components Multicast and GatherCast MultiCast GatherCast
44
Denis Caromel 43 ProActive Parallel Suite
45
Denis Caromel 44 ProActive Parallel Suite
46
Denis Caromel 45 ProActive Parallel Suite (1)
47
Denis Caromel 46 4. IC2D Interactive Control & Debug for Distribution Eclipse GUI for the GRID
48
Denis Caromel 47 GUI in ProActive Parallel Suite
49
Denis Caromel 48 GUI in ProActive Parallel Suite
50
Denis Caromel 49 Programmer Interface for Monitoring Debugging Optimizing
51
Denis Caromel 50
52
Denis Caromel 51 IC2D
53
Denis Caromel 52 Video 2: IC2D Monitoring, Debugging, Optimizing
54
Denis Caromel 53 Ongoing Work: 3D View in IC2D
55
Denis Caromel 54 5. Example of ProActive Applications
56
Denis Caromel 55 Sylvain Cussat-Blanc, Yves Duthen – IRIT TOULOUSE Artificial Life Generation ApplicationJ+1J+5J+6J+7 Version ProActive 251 300 CPUs Initial Application 1 PC56h52 => Crash! ProActive Version 300 CPUs19 minutes Developpement of artificial creatures
57
Denis Caromel 56 Sylvain Cussat-Blanc, Yves Duthen – IRIT TOULOUSE Artificial Life Generation
58
Denis Caromel 57 JECS : 3D Electromagnetism Radar Reflection on Planes
59
Denis Caromel 58 Code Coupling : Vibro Acoustic (courtesy of EADS)
60
59 NAS Parallel Benchmarks Designed by NASA to evaluate benefits of high performance systems Strongly based on CFD 5 benchmarks (kernels) to test different aspects of a system 2 categories : communication intensive and computation intensive
61
Denis Caromel 60 Communication Intensive CG Kernel (Conjugate Gradient) Floating point operations Eigen value computation High number of unstructured communications 12000 calls 570 MB sent 1 min 32 65 % comms Data density distribution Message density distribution
62
Denis Caromel 61 Communication Intensive CG Kernel (Conjugate Gradient)
63
Denis Caromel 62 Enterprise IT: Software Tests Amadeus (Opodo, Air France, KLM, Lufthansa): 500 programmers 50 machines with ProActive to execute Dist. Regression Tests in the production chain
64
Denis Caromel 63 Parallel BLAST with ProActive (1) together with Mario Leyton Basic Local Alignment Search Tool for rapid sequence comparison BLAST developed by NCBI (National Center for Biotechnology Information) Standard native code package, no source modification! With PPS Skeletons parallelization and distribution added to the application A seamless deployment on all Grid platforms is obtained: Input Files are automatically copied to computational nodes at Job submission Result Files will be copied on client host BLAST Skeleton program using the Divide and Conquer skeleton: Division of Database based on conditions (Nb. Nodes, Size, etc.)
65
Denis Caromel 64 Speedup of Distributed BLAST on Grid5000
66
Denis Caromel 65 Monte Carlo Simulations, Non-Linear Physics, INLN
67
Denis Caromel 66 Matlab and Scilab Grid Interface
68
Denis Caromel 67 Mikros Image: Post Production
69
Denis Caromel 68 New Project: AGOS
70
Denis Caromel 69 AGOS Grid Architecture for SOA Partners and Solutions Building a Platform for Agile SOA with Grid Use Cases
71
Denis Caromel 70 AGOS: What for ? AGOS Objectives: Create an architecture and environment for integration of –SOA business management with –GRID IT management Well fitted for data intensive and computational intensive applications: –Enact sub-parts of a BPEL workflow on dynamically allocated resource E.g.: Financial Simulations, Insurance, Revenue Management, BIO, HPC Full dynamic scheduling of Services on GRIDs in the future Integrated Management of SLO, SLA, QoS: –Bottom to top –Dynamic enforcement: Adaptive behavior
72
Denis Caromel 71 Summary
73
Denis Caromel 72 Multi-Core to Distributed Concurrency + Parallelism Multi-Cores + Distribution
74
Denis Caromel 73 Conclusion: Why does it scale? Thanks to a few key features: Connection-less, RMI+JMS unified Messages rather than long-living interactions ACTIVE OBJECTS --- GROUPS --- COMPONENTS
75
Denis Caromel 74 Conclusion: Why does it Compose? Thanks to a few key features: Because it Scales: asynchrony ! Because it is Typed: RMI with interfaces ! First-Class Futures: No unstructured Call Backs and Ports ACTIVE OBJECTS --- GROUPS --- COMPONENTS
76
Denis Caromel 75 Summary-Perspective: Comprehensive Toolkit Programming: Models & Tools Parallel: Multi-Core & Distributed
77
Denis Caromel 76 ProActive/ GCM Specifications for Components Services SLA QoS Open the way to Soft.+Serv. EU Industry with Clouds & Utilities, DAAS ProActive/ GCM Specifications for Components Services SLA QoS Open the way to Soft.+Serv. EU Industry with Clouds & Utilities, DAAS
78
Denis Caromel 77 Multi-Active Object in 1 Address Space for Multi-cores
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.