Vania Boccia | Università degli Studi di Napoli "Federico II" (original) (raw)
Papers by Vania Boccia
Euro-Par 2017: Parallel Processing Workshops, 2018
International Journal of Parallel Programming, 2016
The minimisation of the total cost of ownership is hard to be faced by the owners of large scale ... more The minimisation of the total cost of ownership is hard to be faced by the owners of large scale computing systems, without affecting negatively the quality of service for the users. Modern datacenters, often included in distributed environments, appear to be “elastic”, i.e., they are able to shrink or enlarge the number of local physical or virtual resources, also by recruiting them from private/public clouds. This increases the degree of dynamicity, making the infrastructure management more and more complex. Here, we report some advances in the realisation of an adaptive scheduling controller (ASC) which, by interacting with the datacenter resource manager, allows an effective and an efficient usage of resources. In particular, we focus on the mathematical formalisation of the ASC’s kernel that allows to dynamically configure, in a suitable way, the datacenter resources manager. The described formalisation is based on a probabilistic approach that, starting from both a hystorical resources usage and on the actual users request of the datacenter resources, identifies a suitable probability distribution for queue time with the aim to perform a short term forecasting. The case study is the SCoPE datacenter at the University of Naples Federico II.
2014 International Conference on Intelligent Networking and Collaborative Systems, 2014
The management of distributed e-infrastructures introduces additional requirements in the monitor... more The management of distributed e-infrastructures introduces additional requirements in the monitoring system. The main issue is to enable users to achieve the awareness of the global status, and to present the information in function of the role: simple user, system admin, manager. In this work we present the solution created for the ReCaS infrastructure that federates four data centers in south of Italy. Thanks to the introduction of a data exchange schema, we integrated four heterogeneous and independent subsystems into a single monitoring dashboard. The data collection system allowed us to obtain aggregate metrics by giving a full instantaneous overview of the global infrastructure. Finally, thanks to the user profiling different access views are available, associated to the different class of consumers.
IFIP The International Federation for Information Processing
18th IEEE Symposium on Computer-Based Medical Systems (CBMS'05)
Public reporting burden for this collection of information is estimated to average 1 hour per res... more Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instruction, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington headquarters Services, Directorate for Information Operations and Reports,
2014 International Conference on High Performance Computing & Simulation (HPCS), 2014
This paper addresses the scientific challenges related to high level implementation strategies wh... more This paper addresses the scientific challenges related to high level implementation strategies which steer the NEMO (Nucleus for European Modelling of the Ocean) code toward the effective exploitation of the opportunities offered by exascale systems. We consider, as case studies, two components of the NEMO ocean model (OPA-Ocean PArallelization): the Sea Surface Height equation solver and the Variational Data Assimilation module. The advantages rising from the insertion of consolidated scientific libraries in the NEMO code are highlighted: such advantages concern both the “software quality” improvement (see the software quality parameters like robustness, portability, resilience, etc.) and the reduction of time spent for software development and maintenance. Finally, we consider the Shallow Water equations as a toy model for NEMO ocean model to show how the use of PETSc objects predisposes the application to gain a good level of scalability and efficiency when the most suitable level of abstraction is used.
SSRN Electronic Journal, 2013
This paper addresses scientific challenges related to high level implementation strategies that l... more This paper addresses scientific challenges related to high level implementation strategies that leads NEMO to effectively use of the opportunities of exascale systems. We consider two software modules as proof-of-concept: the Sea Surface Height equation solver and the Variational Data Assimilation system, which are components of the NEMO ocean model (OPA). Advantages rising from the introduction of consolidated scientific libraries in NEMO are highlighted: such advantages concern both the "software quality" improvement (see the software quality parameters like robustness, portability, resilence, etc.) and time reduction of software development 02 Centro Euro-Mediterraneo sui Cambiamenti Climatici
Proceedings of 36th International Conference on High Energy Physics — PoS(ICHEP2012)
The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s-1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.
Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Uni... more Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Universit`a di Catania, Universit`a di Napoli Federico II, Seconda Universit`a di Napoli, Universit`a di Palermo, Universit`a di Salerno ...
Lecture Notes in Computer Science, 2013
Cloud Computing has been widely adopted as a new paradigm for providing resources because of the ... more Cloud Computing has been widely adopted as a new paradigm for providing resources because of the advantages it brings to both users and providers. Even if it was firstly targeted at enterprises wishing to reduce their equipment management costs, it has been rapidly recognized as both an enabler for new applications and as a mean to allow enterprises of all sizes at running high demanding applications. Recently, Cloud Providers are trying to attract new applications, such as scientific ones, that today already benefit from distributed environment like Grids. This work presents a way to remove the paradigm mismatch between Cloud and Grid Computing, enabling the use of Cloud-provided resources with well-established Grid-like interfaces, avoiding the need for users to learn new resources access and use models. The proposed approach is validated through the development of a prototype implementation and its integration in a working Grid environment.
The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s -1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.
2014 International Conference on Intelligent Networking and Collaborative Systems, 2014
Euro-Par 2017: Parallel Processing Workshops, 2018
International Journal of Parallel Programming, 2016
The minimisation of the total cost of ownership is hard to be faced by the owners of large scale ... more The minimisation of the total cost of ownership is hard to be faced by the owners of large scale computing systems, without affecting negatively the quality of service for the users. Modern datacenters, often included in distributed environments, appear to be “elastic”, i.e., they are able to shrink or enlarge the number of local physical or virtual resources, also by recruiting them from private/public clouds. This increases the degree of dynamicity, making the infrastructure management more and more complex. Here, we report some advances in the realisation of an adaptive scheduling controller (ASC) which, by interacting with the datacenter resource manager, allows an effective and an efficient usage of resources. In particular, we focus on the mathematical formalisation of the ASC’s kernel that allows to dynamically configure, in a suitable way, the datacenter resources manager. The described formalisation is based on a probabilistic approach that, starting from both a hystorical resources usage and on the actual users request of the datacenter resources, identifies a suitable probability distribution for queue time with the aim to perform a short term forecasting. The case study is the SCoPE datacenter at the University of Naples Federico II.
2014 International Conference on Intelligent Networking and Collaborative Systems, 2014
The management of distributed e-infrastructures introduces additional requirements in the monitor... more The management of distributed e-infrastructures introduces additional requirements in the monitoring system. The main issue is to enable users to achieve the awareness of the global status, and to present the information in function of the role: simple user, system admin, manager. In this work we present the solution created for the ReCaS infrastructure that federates four data centers in south of Italy. Thanks to the introduction of a data exchange schema, we integrated four heterogeneous and independent subsystems into a single monitoring dashboard. The data collection system allowed us to obtain aggregate metrics by giving a full instantaneous overview of the global infrastructure. Finally, thanks to the user profiling different access views are available, associated to the different class of consumers.
IFIP The International Federation for Information Processing
18th IEEE Symposium on Computer-Based Medical Systems (CBMS'05)
Public reporting burden for this collection of information is estimated to average 1 hour per res... more Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instruction, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington headquarters Services, Directorate for Information Operations and Reports,
2014 International Conference on High Performance Computing & Simulation (HPCS), 2014
This paper addresses the scientific challenges related to high level implementation strategies wh... more This paper addresses the scientific challenges related to high level implementation strategies which steer the NEMO (Nucleus for European Modelling of the Ocean) code toward the effective exploitation of the opportunities offered by exascale systems. We consider, as case studies, two components of the NEMO ocean model (OPA-Ocean PArallelization): the Sea Surface Height equation solver and the Variational Data Assimilation module. The advantages rising from the insertion of consolidated scientific libraries in the NEMO code are highlighted: such advantages concern both the “software quality” improvement (see the software quality parameters like robustness, portability, resilience, etc.) and the reduction of time spent for software development and maintenance. Finally, we consider the Shallow Water equations as a toy model for NEMO ocean model to show how the use of PETSc objects predisposes the application to gain a good level of scalability and efficiency when the most suitable level of abstraction is used.
SSRN Electronic Journal, 2013
This paper addresses scientific challenges related to high level implementation strategies that l... more This paper addresses scientific challenges related to high level implementation strategies that leads NEMO to effectively use of the opportunities of exascale systems. We consider two software modules as proof-of-concept: the Sea Surface Height equation solver and the Variational Data Assimilation system, which are components of the NEMO ocean model (OPA). Advantages rising from the introduction of consolidated scientific libraries in NEMO are highlighted: such advantages concern both the "software quality" improvement (see the software quality parameters like robustness, portability, resilence, etc.) and time reduction of software development 02 Centro Euro-Mediterraneo sui Cambiamenti Climatici
Proceedings of 36th International Conference on High Energy Physics — PoS(ICHEP2012)
The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s-1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.
Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Uni... more Page 1. DOTTORATO DI RICERCA in SCIENZE COMPUTAZIONALI E INFORMATICHE Ciclo XXI Consorzio tra Universit`a di Catania, Universit`a di Napoli Federico II, Seconda Universit`a di Napoli, Universit`a di Palermo, Universit`a di Salerno ...
Lecture Notes in Computer Science, 2013
Cloud Computing has been widely adopted as a new paradigm for providing resources because of the ... more Cloud Computing has been widely adopted as a new paradigm for providing resources because of the advantages it brings to both users and providers. Even if it was firstly targeted at enterprises wishing to reduce their equipment management costs, it has been rapidly recognized as both an enabler for new applications and as a mean to allow enterprises of all sizes at running high demanding applications. Recently, Cloud Providers are trying to attract new applications, such as scientific ones, that today already benefit from distributed environment like Grids. This work presents a way to remove the paradigm mismatch between Cloud and Grid Computing, enabling the use of Cloud-provided resources with well-established Grid-like interfaces, avoiding the need for users to learn new resources access and use models. The proposed approach is validated through the development of a prototype implementation and its integration in a working Grid environment.
The development of a computing model for the next generation of Super Flavour Factories, like Sup... more The development of a computing model for the next generation of Super Flavour Factories, like SuperB and SuperKEKB, presents significant challenges. With a nominal luminosity above 10 36 cm-2 s -1 , we estimate that, after few years of operation, the size of the data sample will be of the order of 500 PB and the amount of CPU required to process it will be close to 5000 KHep-Spec06 (the new HEP-wide benchmark for measuring CPU performance). The new many and multi core technologies need to be effectively exploited in order to manage very large data set and this has a potential large impact on the computing model for SuperB. In addition, the computing resources available to SuperB, as is already the case for LHC experiments, will be distributed and accessed through a Grid or a cloud infrastructure and a suite of efficient and reliable tools needs to be provided to the users. A dedicated research program to explore these issues is in progress and it is presented here.
2014 International Conference on Intelligent Networking and Collaborative Systems, 2014