Kokou Yetongnon - Academia.edu (original) (raw)
Uploads
Papers by Kokou Yetongnon
Abstract. Developing efficient techniques for comparing XML-based documents becomes essential in ... more Abstract. Developing efficient techniques for comparing XML-based documents becomes essential in the database and information retrieval communities. Various algorithms for comparing hierarchically structured data, e.g. XML documents, have been proposed in the literature. Most of them make use of techniques for finding the edit distance between tree structures, XML documents being modeled as ordered labeled trees. Nevertheless, a thorough investigation of current approaches led us to identify several unaddressed structural similarities, i.e. sub-tree related similarities, while comparing XML documents. In this paper, we provide an improved comparison method to deal with such resemblances. Our approach is based on the concept of tree edit distance, introducing the notion of commonality between sub-trees. Experiments demonstrate that our approach yields better similarity results with respect to alternative methods, while maintaining quatratic time complexity. 1.
This paper describes the Distributed Object Kernel (DOK) system, which provides extended database... more This paper describes the Distributed Object Kernel (DOK) system, which provides extended database functionalities over heterogeneous and distributed multimedia databases. The DOK object model provides mechanisms for representing di erent forms of information, including structured (e.g. tables and objects) and non-structured (e.g. images and voice) information. In addition, the DOK object model allows a self description of both structural and behavioral aspects of heterogeneous database applications. A set of metaobjects describe the information about objects, the messages they can process, and the way messages are to be executed in a distributed and heterogeneous environment. The DOK logical architecture has a clear separation of concerns, involving a coordination layer, a task layer and a database layer. Each of these layers contains a set of agents specialised in performing speci c functions related to the cooperative computation. When a user requests information about a DOK appli...
The concepts of object oriented data models aim towards modeling of application objects close to ... more The concepts of object oriented data models aim towards modeling of application objects close to the user's view. Yet developers of applications relying on object oriented database management systems are facing problems resulting from the limitations of object oriented data models to describe adequately the full range of possible associations between objects and between processes. This paper focuses on the conceptual data modeling needs of object oriented database applications, and proposes an object+relationship model, ERC+, which meets database application requirements by merging traditional semantic data models features with object oriented capabilities such as structural object orientation, inheritance, and object identity. It is an extended entity-relationship model specifically designed to support complex object description and to allow multiple perceptions of objects. The ERC+ model provides the foundations for an integrated environment of tools called SUPER that has been...
Proceedings of the 13th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management
2012 Eighth International Conference on Signal Image Technology and Internet Based Systems, 2012
Proceedings of New Trends in Intelligent Information Processing and Web Mining Iis 2003, 2003
... In Proceedings of the 2nd World Congress on Expert Systems, Lisbon,/Estoril, Portugal, Januar... more ... In Proceedings of the 2nd World Congress on Expert Systems, Lisbon,/Estoril, Portugal, January 1994. 2. Djamal Benslimane, Mohand-Said Hacid, Evimaria Terzi, and Farouk Toumani. A class-based logic language for ontologies. In In the proceedings of the Fifth Page 492. ...
Lecture Notes in Computer Science, 1999
ABSTRACT
Web Enabled Systems Integration, 2003
XML-Based Toolkit for Interoperability of Web Information Systems 91 Chapter V XML-Based Toolkit ... more XML-Based Toolkit for Interoperability of Web Information Systems 91 Chapter V XML-Based Toolkit for Interoperability of Web Information Systems Christophe Nicolle and Kokou Yétongnon University ofBourgogne, France ABSTRACT This chapter presents a methodology and ...
The entity-relationship (ER) approach still lacks sound and complete theoretical foundations for ... more The entity-relationship (ER) approach still lacks sound and complete theoretical foundations for modern, more sophisticated, ER models. We already proposed an algebra as a basis for data manipulation languages for such an advanced ER model, called ERC. The aim of this paper is to formally define a companion ERC calculus. We first investigate a general calculus, which is more natural (straight derivation from the model), but too powerful as it also allows expressions which are semantically inconsistent with the existing database. We then define what restrictions should be imposed, and how, to lead to a calculus whose expressive power is equivalent to the one of the ERC algebra.
This paper describes the Distributed Object Kernel DOK system, which provides extended database f... more This paper describes the Distributed Object Kernel DOK system, which provides extended database functionalities over heterogeneous and distributed multimedia databases. The DOK object model provides mechanisms for representing di erent forms of information, including structured e.g. tables and objects and non-structured e.g. images and voice information. In addition, the DOK object model allows a self , description of both structural and behavioral aspects of heterogeneous database applications. A set of metaobjects describe the information about objects, the messages they can process, and the way messages are to be executed in a distributed and heterogeneous environment. The DOK logical architecture has a clear separation of concerns, involving a coordination layer, a task layer and a database layer. Each of these layers contains a set of agents specialised in performing speci c functions related to the cooperative computation. When a user requests information about a DOK application, the coordination agents identify and delegate the sub-parts of the query to appropriate agents of the task-based layer. These specialised agents, in turn, use the agents of the database layer to access the required information to process the di erent subparts of the global request.
Proceedings of the Sixth Ieee International Symposium on Cluster Computing and the Grid, May 16, 2006
P2P systems represent a large portion of the Internet traffic which makes the data discovery of g... more P2P systems represent a large portion of the Internet traffic which makes the data discovery of great importance to the user and the broad Internet community. Hence, the power of a P2P system comes from its ability to provide an efficient search service. In this paper we address the problem of similarity search in a Hybrid Overlay P2P Network which organizes data and peers in a high dimensional feature space. Data and peers are described by a set of features and clustered using a density-based algorithm. We experimentally evaluate the effectiveness of the similarity-search using uniform and zipf data distribution.
Iswc, 2006
This paper focuses on P2P based data management and semantic mediation. We propose an approach ba... more This paper focuses on P2P based data management and semantic mediation. We propose an approach based on a P2P for semantic interoperability of information sources that aims to combine the advantages of semantic mediation and peer-to-peer systems. It is based on a pure P2P with super peer architecture consisting of two types of peers. The super peer contains a reference ontology, which provides a common ontology (CO) of the domain. The peer contains export schema (ES), which represent local data. The approach based on semantic agreement between CO and ES, which called half agreement (HA). The halfagreements utilize for discovery sources and exchange information among peers.
Revue Des Nouvelles Technologies Informatiques, 2011
http://le2i.cnrs.fr Résumé. Dans le domaine biomédical, la protéomique est confrontée à des sourc... more http://le2i.cnrs.fr Résumé. Dans le domaine biomédical, la protéomique est confrontée à des sources de données de plus en plus nombreuses et à des volumes de données très importants du fait de la multiplication des technologies dites à haut débit. L'hétérogénéité de la provenance des données implique de fait une hétérogénéité dans la représentation et le contenu de ces données. Les données peuvent aussi se révéler incorrectes ce qui engendre des erreurs sur les conclusions des expériences protéomiques. Notre approche a pour objectif de garantir la qualité initiale des données lors de leur importation dans un système d'information dédié à la protéomique. Elle est basée sur le couplage entre des modèles représentant les sources et le système protéomique, et des ontologies utilisées comme médiatrices entre les modèles. Les différents contrôles que nous proposons de mettre en place garantissent la validité des domaines de valeurs, la sémantique et la cohérence des données lors de l'importation. 1. Ces systèmes sont plus connus sous leur nom anglais : Laboratory Information Management System (LIMS)
Abstract. Developing efficient techniques for comparing XML-based documents becomes essential in ... more Abstract. Developing efficient techniques for comparing XML-based documents becomes essential in the database and information retrieval communities. Various algorithms for comparing hierarchically structured data, e.g. XML documents, have been proposed in the literature. Most of them make use of techniques for finding the edit distance between tree structures, XML documents being modeled as ordered labeled trees. Nevertheless, a thorough investigation of current approaches led us to identify several unaddressed structural similarities, i.e. sub-tree related similarities, while comparing XML documents. In this paper, we provide an improved comparison method to deal with such resemblances. Our approach is based on the concept of tree edit distance, introducing the notion of commonality between sub-trees. Experiments demonstrate that our approach yields better similarity results with respect to alternative methods, while maintaining quatratic time complexity. 1.
This paper describes the Distributed Object Kernel (DOK) system, which provides extended database... more This paper describes the Distributed Object Kernel (DOK) system, which provides extended database functionalities over heterogeneous and distributed multimedia databases. The DOK object model provides mechanisms for representing di erent forms of information, including structured (e.g. tables and objects) and non-structured (e.g. images and voice) information. In addition, the DOK object model allows a self description of both structural and behavioral aspects of heterogeneous database applications. A set of metaobjects describe the information about objects, the messages they can process, and the way messages are to be executed in a distributed and heterogeneous environment. The DOK logical architecture has a clear separation of concerns, involving a coordination layer, a task layer and a database layer. Each of these layers contains a set of agents specialised in performing speci c functions related to the cooperative computation. When a user requests information about a DOK appli...
The concepts of object oriented data models aim towards modeling of application objects close to ... more The concepts of object oriented data models aim towards modeling of application objects close to the user's view. Yet developers of applications relying on object oriented database management systems are facing problems resulting from the limitations of object oriented data models to describe adequately the full range of possible associations between objects and between processes. This paper focuses on the conceptual data modeling needs of object oriented database applications, and proposes an object+relationship model, ERC+, which meets database application requirements by merging traditional semantic data models features with object oriented capabilities such as structural object orientation, inheritance, and object identity. It is an extended entity-relationship model specifically designed to support complex object description and to allow multiple perceptions of objects. The ERC+ model provides the foundations for an integrated environment of tools called SUPER that has been...
Proceedings of the 13th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management
2012 Eighth International Conference on Signal Image Technology and Internet Based Systems, 2012
Proceedings of New Trends in Intelligent Information Processing and Web Mining Iis 2003, 2003
... In Proceedings of the 2nd World Congress on Expert Systems, Lisbon,/Estoril, Portugal, Januar... more ... In Proceedings of the 2nd World Congress on Expert Systems, Lisbon,/Estoril, Portugal, January 1994. 2. Djamal Benslimane, Mohand-Said Hacid, Evimaria Terzi, and Farouk Toumani. A class-based logic language for ontologies. In In the proceedings of the Fifth Page 492. ...
Lecture Notes in Computer Science, 1999
ABSTRACT
Web Enabled Systems Integration, 2003
XML-Based Toolkit for Interoperability of Web Information Systems 91 Chapter V XML-Based Toolkit ... more XML-Based Toolkit for Interoperability of Web Information Systems 91 Chapter V XML-Based Toolkit for Interoperability of Web Information Systems Christophe Nicolle and Kokou Yétongnon University ofBourgogne, France ABSTRACT This chapter presents a methodology and ...
The entity-relationship (ER) approach still lacks sound and complete theoretical foundations for ... more The entity-relationship (ER) approach still lacks sound and complete theoretical foundations for modern, more sophisticated, ER models. We already proposed an algebra as a basis for data manipulation languages for such an advanced ER model, called ERC. The aim of this paper is to formally define a companion ERC calculus. We first investigate a general calculus, which is more natural (straight derivation from the model), but too powerful as it also allows expressions which are semantically inconsistent with the existing database. We then define what restrictions should be imposed, and how, to lead to a calculus whose expressive power is equivalent to the one of the ERC algebra.
This paper describes the Distributed Object Kernel DOK system, which provides extended database f... more This paper describes the Distributed Object Kernel DOK system, which provides extended database functionalities over heterogeneous and distributed multimedia databases. The DOK object model provides mechanisms for representing di erent forms of information, including structured e.g. tables and objects and non-structured e.g. images and voice information. In addition, the DOK object model allows a self , description of both structural and behavioral aspects of heterogeneous database applications. A set of metaobjects describe the information about objects, the messages they can process, and the way messages are to be executed in a distributed and heterogeneous environment. The DOK logical architecture has a clear separation of concerns, involving a coordination layer, a task layer and a database layer. Each of these layers contains a set of agents specialised in performing speci c functions related to the cooperative computation. When a user requests information about a DOK application, the coordination agents identify and delegate the sub-parts of the query to appropriate agents of the task-based layer. These specialised agents, in turn, use the agents of the database layer to access the required information to process the di erent subparts of the global request.
Proceedings of the Sixth Ieee International Symposium on Cluster Computing and the Grid, May 16, 2006
P2P systems represent a large portion of the Internet traffic which makes the data discovery of g... more P2P systems represent a large portion of the Internet traffic which makes the data discovery of great importance to the user and the broad Internet community. Hence, the power of a P2P system comes from its ability to provide an efficient search service. In this paper we address the problem of similarity search in a Hybrid Overlay P2P Network which organizes data and peers in a high dimensional feature space. Data and peers are described by a set of features and clustered using a density-based algorithm. We experimentally evaluate the effectiveness of the similarity-search using uniform and zipf data distribution.
Iswc, 2006
This paper focuses on P2P based data management and semantic mediation. We propose an approach ba... more This paper focuses on P2P based data management and semantic mediation. We propose an approach based on a P2P for semantic interoperability of information sources that aims to combine the advantages of semantic mediation and peer-to-peer systems. It is based on a pure P2P with super peer architecture consisting of two types of peers. The super peer contains a reference ontology, which provides a common ontology (CO) of the domain. The peer contains export schema (ES), which represent local data. The approach based on semantic agreement between CO and ES, which called half agreement (HA). The halfagreements utilize for discovery sources and exchange information among peers.
Revue Des Nouvelles Technologies Informatiques, 2011
http://le2i.cnrs.fr Résumé. Dans le domaine biomédical, la protéomique est confrontée à des sourc... more http://le2i.cnrs.fr Résumé. Dans le domaine biomédical, la protéomique est confrontée à des sources de données de plus en plus nombreuses et à des volumes de données très importants du fait de la multiplication des technologies dites à haut débit. L'hétérogénéité de la provenance des données implique de fait une hétérogénéité dans la représentation et le contenu de ces données. Les données peuvent aussi se révéler incorrectes ce qui engendre des erreurs sur les conclusions des expériences protéomiques. Notre approche a pour objectif de garantir la qualité initiale des données lors de leur importation dans un système d'information dédié à la protéomique. Elle est basée sur le couplage entre des modèles représentant les sources et le système protéomique, et des ontologies utilisées comme médiatrices entre les modèles. Les différents contrôles que nous proposons de mettre en place garantissent la validité des domaines de valeurs, la sémantique et la cohérence des données lors de l'importation. 1. Ces systèmes sont plus connus sous leur nom anglais : Laboratory Information Management System (LIMS)