Vania Boccia | Università degli Studi di Napoli "Federico II" (original) (raw)

Papers by Vania Boccia

Research paper thumbnail of SCoPE@Scuola: (In)-formative Paths on Topics Related with High Performance, Parallel and Distributed Computing

Euro-Par 2017: Parallel Processing Workshops, 2018

Research paper thumbnail of GaaS 2.0: The New Release Based on OpenStack with Features Implemented with OCCI

Research paper thumbnail of User Recruitment and Support: A Viral Marketing Approach Based on the Word-of-Mouth Influence

Research paper thumbnail of The Belle II Simulation Campaign at ReCaS

Research paper thumbnail of Monitoring the ReCaS Project Resources

Research paper thumbnail of An Approach to Forecast Queue Time in Adaptive Scheduling: How to Mediate System Efficiency and Users Satisfaction

International Journal of Parallel Programming, 2016

The minimisation of the total cost of ownership is hard to be faced by the owners of large scale ... more The minimisation of the total cost of ownership is hard to be faced by the owners of large scale computing systems, without affecting negatively the quality of service for the users. Modern datacenters, often included in distributed environments, appear to be “elastic”, i.e., they are able to shrink or enlarge the number of local physical or virtual resources, also by recruiting them from private/public clouds. This increases the degree of dynamicity, making the infrastructure management more and more complex. Here, we report some advances in the realisation of an adaptive scheduling controller (ASC) which, by interacting with the datacenter resource manager, allows an effective and an efficient usage of resources. In particular, we focus on the mathematical formalisation of the ASC’s kernel that allows to dynamically configure, in a suitable way, the datacenter resources manager. The described formalisation is based on a probabilistic approach that, starting from both a hystorical resources usage and on the actual users request of the datacenter resources, identifies a suitable probability distribution for queue time with the aim to perform a short term forecasting. The case study is the SCoPE datacenter at the University of Naples Federico II.

Research paper thumbnail of Infrastructure Monitoring for Distributed Tier1: The ReCaS Project Use-Case

2014 International Conference on Intelligent Networking and Collaborative Systems, 2014

The management of distributed e-infrastructures introduces additional requirements in the monitor... more The management of distributed e-infrastructures introduces additional requirements in the monitoring system. The main issue is to enable users to achieve the awareness of the global status, and to present the information in function of the role: simple user, system admin, manager. In this work we present the solution created for the ReCaS infrastructure that federates four data centers in south of Italy. Thanks to the introduction of a data exchange schema, we integrated four heterogeneous and independent subsystems into a single monitoring dashboard. The data collection system allowed us to obtain aggregate metrics by giving a full instantaneous overview of the global infrastructure. Finally, thanks to the user profiling different access views are available, associated to the different class of consumers.

Research paper thumbnail of Monitoring and Migration of a PETSc-based Parallel Application for Medical Imaging in a Grid computing PSE

IFIP The International Federation for Information Processing

Research paper thumbnail of A Grid Enabled PSE for Medical Imaging: Experiences on MedIGrid

18th IEEE Symposium on Computer-Based Medical Systems (CBMS'05)

Public reporting burden for this collection of information is estimated to average 1 hour per res... more Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instruction, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington headquarters Services, Directorate for Information Operations and Reports,

Research paper thumbnail of Insertion of PETSc in the NEMO stack software driving NEMO towards exascale computing

2014 International Conference on High Performance Computing & Simulation (HPCS), 2014

This paper addresses the scientific challenges related to high level implementation strategies wh... more This paper addresses the scientific challenges related to high level implementation strategies which steer the NEMO (Nucleus for European Modelling of the Ocean) code toward the effective exploitation of the opportunities offered by exascale systems. We consider, as case studies, two components of the NEMO ocean model (OPA-Ocean PArallelization): the Sea Surface Height equation solver and the Variational Data Assimilation module. The advantages rising from the insertion of consolidated scientific libraries in the NEMO code are highlighted: such advantages concern both the “software quality” improvement (see the software quality parameters like robustness, portability, resilience, etc.) and the reduction of time spent for software development and maintenance. Finally, we consider the Shallow Water equations as a toy model for NEMO ocean model to show how the use of PETSc objects predisposes the application to gain a good level of scalability and efficiency when the most suitable level of abstraction is used.

Research paper thumbnail of Driving NEMO Towards Exascale: Introduction of a New Software Layer in the NEMO Stack Software

SSRN Electronic Journal, 2013

This paper addresses scientific challenges related to high level implementation strategies that l... more This paper addresses scientific challenges related to high level implementation strategies that leads NEMO to effectively use of the opportunities of exascale systems. We consider two software modules as proof-of-concept: the Sea Surface Height equation solver and the Variational Data Assimilation system, which are components of the NEMO ocean model (OPA). Advantages rising from the introduction of consolidated scientific libraries in NEMO are highlighted: such advantages concern both the "software quality" improvement (see the software quality parameters like robustness, portability, resilence, etc.) and time reduction of software development 02 Centro Euro-Mediterraneo sui Cambiamenti Climatici

Research paper thumbnail of Workpackage 9 Grid-enabled Scientific Libraries

Research paper thumbnail of Recent advances of MedIGrid PSE in an LCG/gLite environment

Research paper thumbnail of Computing at SuperB

Proceedings of 36th International Conference on High Energy Physics — PoS(ICHEP2012)

The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s-1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.

Research paper thumbnail of Realizzazione di un sistema dicheckpointing/migration, in ambientedistribuito, per la fault tolerance di applicazioniparallele basate su PETSc

Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Uni... more Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Universit`a di Catania, Universit`a di Napoli Federico II, Seconda Universit`a di Napoli, Universit`a di Palermo, Universit`a di Salerno ...

Research paper thumbnail of ASC: an adaptative scheduling controller

Research paper thumbnail of An Approach to Model Resources Rationalisation in Hybrid Clouds through Users Activity Characterisation

Research paper thumbnail of GaaS: Customized Grids in the Clouds

Lecture Notes in Computer Science, 2013

Cloud Computing has been widely adopted as a new paradigm for providing resources because of the ... more Cloud Computing has been widely adopted as a new paradigm for providing resources because of the advantages it brings to both users and providers. Even if it was firstly targeted at enterprises wishing to reduce their equipment management costs, it has been rapidly recognized as both an enabler for new applications and as a mean to allow enterprises of all sizes at running high demanding applications. Recently, Cloud Providers are trying to attract new applications, such as scientific ones, that today already benefit from distributed environment like Grids. This work presents a way to remove the paradigm mismatch between Cloud and Grid Computing, enabling the use of Cloud-provided resources with well-established Grid-like interfaces, avoiding the need for users to learn new resources access and use models. The proposed approach is validated through the development of a prototype implementation and its integration in a working Grid environment.

Research paper thumbnail of Computing at SuperB

The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s -1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.

Research paper thumbnail of ECCO: An Integrated Solution for Environment Compatible COmputing Systems

2014 International Conference on Intelligent Networking and Collaborative Systems, 2014

Research paper thumbnail of SCoPE@Scuola: (In)-formative Paths on Topics Related with High Performance, Parallel and Distributed Computing

Euro-Par 2017: Parallel Processing Workshops, 2018

Research paper thumbnail of GaaS 2.0: The New Release Based on OpenStack with Features Implemented with OCCI

Research paper thumbnail of User Recruitment and Support: A Viral Marketing Approach Based on the Word-of-Mouth Influence

Research paper thumbnail of The Belle II Simulation Campaign at ReCaS

Research paper thumbnail of Monitoring the ReCaS Project Resources

Research paper thumbnail of An Approach to Forecast Queue Time in Adaptive Scheduling: How to Mediate System Efficiency and Users Satisfaction

International Journal of Parallel Programming, 2016

The minimisation of the total cost of ownership is hard to be faced by the owners of large scale ... more The minimisation of the total cost of ownership is hard to be faced by the owners of large scale computing systems, without affecting negatively the quality of service for the users. Modern datacenters, often included in distributed environments, appear to be “elastic”, i.e., they are able to shrink or enlarge the number of local physical or virtual resources, also by recruiting them from private/public clouds. This increases the degree of dynamicity, making the infrastructure management more and more complex. Here, we report some advances in the realisation of an adaptive scheduling controller (ASC) which, by interacting with the datacenter resource manager, allows an effective and an efficient usage of resources. In particular, we focus on the mathematical formalisation of the ASC’s kernel that allows to dynamically configure, in a suitable way, the datacenter resources manager. The described formalisation is based on a probabilistic approach that, starting from both a hystorical resources usage and on the actual users request of the datacenter resources, identifies a suitable probability distribution for queue time with the aim to perform a short term forecasting. The case study is the SCoPE datacenter at the University of Naples Federico II.

Research paper thumbnail of Infrastructure Monitoring for Distributed Tier1: The ReCaS Project Use-Case

2014 International Conference on Intelligent Networking and Collaborative Systems, 2014

The management of distributed e-infrastructures introduces additional requirements in the monitor... more The management of distributed e-infrastructures introduces additional requirements in the monitoring system. The main issue is to enable users to achieve the awareness of the global status, and to present the information in function of the role: simple user, system admin, manager. In this work we present the solution created for the ReCaS infrastructure that federates four data centers in south of Italy. Thanks to the introduction of a data exchange schema, we integrated four heterogeneous and independent subsystems into a single monitoring dashboard. The data collection system allowed us to obtain aggregate metrics by giving a full instantaneous overview of the global infrastructure. Finally, thanks to the user profiling different access views are available, associated to the different class of consumers.

Research paper thumbnail of Monitoring and Migration of a PETSc-based Parallel Application for Medical Imaging in a Grid computing PSE

IFIP The International Federation for Information Processing

Research paper thumbnail of A Grid Enabled PSE for Medical Imaging: Experiences on MedIGrid

18th IEEE Symposium on Computer-Based Medical Systems (CBMS'05)

Public reporting burden for this collection of information is estimated to average 1 hour per res... more Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instruction, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington headquarters Services, Directorate for Information Operations and Reports,

Research paper thumbnail of Insertion of PETSc in the NEMO stack software driving NEMO towards exascale computing

2014 International Conference on High Performance Computing & Simulation (HPCS), 2014

This paper addresses the scientific challenges related to high level implementation strategies wh... more This paper addresses the scientific challenges related to high level implementation strategies which steer the NEMO (Nucleus for European Modelling of the Ocean) code toward the effective exploitation of the opportunities offered by exascale systems. We consider, as case studies, two components of the NEMO ocean model (OPA-Ocean PArallelization): the Sea Surface Height equation solver and the Variational Data Assimilation module. The advantages rising from the insertion of consolidated scientific libraries in the NEMO code are highlighted: such advantages concern both the “software quality” improvement (see the software quality parameters like robustness, portability, resilience, etc.) and the reduction of time spent for software development and maintenance. Finally, we consider the Shallow Water equations as a toy model for NEMO ocean model to show how the use of PETSc objects predisposes the application to gain a good level of scalability and efficiency when the most suitable level of abstraction is used.

Research paper thumbnail of Driving NEMO Towards Exascale: Introduction of a New Software Layer in the NEMO Stack Software

SSRN Electronic Journal, 2013

This paper addresses scientific challenges related to high level implementation strategies that l... more This paper addresses scientific challenges related to high level implementation strategies that leads NEMO to effectively use of the opportunities of exascale systems. We consider two software modules as proof-of-concept: the Sea Surface Height equation solver and the Variational Data Assimilation system, which are components of the NEMO ocean model (OPA). Advantages rising from the introduction of consolidated scientific libraries in NEMO are highlighted: such advantages concern both the "software quality" improvement (see the software quality parameters like robustness, portability, resilence, etc.) and time reduction of software development 02 Centro Euro-Mediterraneo sui Cambiamenti Climatici

Research paper thumbnail of Workpackage 9 Grid-enabled Scientific Libraries

Research paper thumbnail of Recent advances of MedIGrid PSE in an LCG/gLite environment

Research paper thumbnail of Computing at SuperB

Proceedings of 36th International Conference on High Energy Physics — PoS(ICHEP2012)

The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s-1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.

Research paper thumbnail of Realizzazione di un sistema dicheckpointing/migration, in ambientedistribuito, per la fault tolerance di applicazioniparallele basate su PETSc

Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Uni... more Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Universit`a di Catania, Universit`a di Napoli Federico II, Seconda Universit`a di Napoli, Universit`a di Palermo, Universit`a di Salerno ...

Research paper thumbnail of ASC: an adaptative scheduling controller

Research paper thumbnail of An Approach to Model Resources Rationalisation in Hybrid Clouds through Users Activity Characterisation

Research paper thumbnail of GaaS: Customized Grids in the Clouds

Lecture Notes in Computer Science, 2013

Cloud Computing has been widely adopted as a new paradigm for providing resources because of the ... more Cloud Computing has been widely adopted as a new paradigm for providing resources because of the advantages it brings to both users and providers. Even if it was firstly targeted at enterprises wishing to reduce their equipment management costs, it has been rapidly recognized as both an enabler for new applications and as a mean to allow enterprises of all sizes at running high demanding applications. Recently, Cloud Providers are trying to attract new applications, such as scientific ones, that today already benefit from distributed environment like Grids. This work presents a way to remove the paradigm mismatch between Cloud and Grid Computing, enabling the use of Cloud-provided resources with well-established Grid-like interfaces, avoiding the need for users to learn new resources access and use models. The proposed approach is validated through the development of a prototype implementation and its integration in a working Grid environment.

Research paper thumbnail of Computing at SuperB

The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s -1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.

Research paper thumbnail of ECCO: An Integrated Solution for Environment Compatible COmputing Systems

2014 International Conference on Intelligent Networking and Collaborative Systems, 2014