Lutz Schubert | Universität zu Köln (original) (raw)

Papers by Lutz Schubert

Research paper thumbnail of Proceedings of the 2013 international workshop on Multi-cloud applications and federated clouds

Research paper thumbnail of SLA(サービスレベル一致)支援の資源管理に向けて

Lect Notes Comput Sci, 2006

Research paper thumbnail of Proceedings of the Workshop on International Transit Turnkey and Joint Development. Session 5: Environmental and Risk Management Considerations

Transportation research circular, Mar 1, 1998

Session highlights are as follows: (1) Innovative approaches in environmental management, such as... more Session highlights are as follows: (1) Innovative approaches in environmental management, such as performance based financial incentives, collaboration with the community with regards to mitigation measures, and agreements with agencies that address the review process, minimizing paperwork and cost to all parties, should be pursued. (2) Unresolved environmental issues, since they result in higher risks which in turn result in higher costs, should be avoided. An effort should be made to understand risk probabilities. A proactive environmental management process should be established early in a turnkey procurement process. (3) In a turnkey procurement process, risks must be clearly explicit. Different methods are available to manage risk. A methodology for managing risk was presented. A flow-chart process was also presented to help in the process of identifying and managing risk. The different types of risks were discussed with suggested strategies to manage them. (4) In an effort to foster and direct community participation in transit projects, the Livable Communities Initiative Program was established by the Federal Transit Administration in 1994. A video was presented showing how communities got involved and challenged projects through legal actions in the Los Angeles area. Public Participation/Community Participation should have an active role in every phase of a transit project. (5) The industry is currently developing insurance policies that would protect design liability. Currently, sureties are applicable only in the construction phase of a project. They are also investigating the possibility of Phase Contracting where the owner accepts the risk in the design phase and sureties cover the construction phase. (6) The contractors' perspective in Risk Allocation is: How well companies identify, manage, and mitigate risk directly impact their bottom line. Risk allocation must be balanced between the public and private sector. Risks must be placed where they can best be managed. (7) Bay Area Rapid Transit's experience with environmental considerations in the extension to the San Francisco International Airport was explained. (8) The consultant's perspective with regards to environmental and risk management is: The role of a consultant is to identify clients that have a sound commitment to complete projects. One must know the playing field, understanding clients, stakeholders, and regulatory framework. One must push the limits by understanding best practices, relating innovation to established objectives, and valuing good design.

Research paper thumbnail of MyThOS — Scalable OS Design for Extremely Parallel Applications

2016 Intl IEEE Conferences on Ubiquitous Intelligence & Computing, Advanced and Trusted Computing, Scalable Computing and Communications, Cloud and Big Data Computing, Internet of People, and Smart World Congress (UIC/ATC/ScalCom/CBDCom/IoP/SmartWorld), 2016

Many-core architectures trade single-thread performance for a larger number of cores. Scalable th... more Many-core architectures trade single-thread performance for a larger number of cores. Scalable throughput can be attained only by a high degree of parallelism, minimized synchronization. Whilst this is achievable for many applications, the operating system still introduces bottlenecks through non-local sharing, synchronization,, message passing. A particular challenge for highly dynamic applications, for example invasive HPC applications, elastic compute clouds, is the management of short-living application threads, processes. This paper discusses OS architecture choices based on microkernel, multikernel, distributed systems designs, our development experience in the context of the MyThOS project. Initial experiments show a much faster thread creation, activation compared to monolithic systems like Linux while providing a more flexible protection, threading model that is better suited for dynamic scenarios. However, despite significant progress in the overall domain of operating systems, the design space for scalable many-core operating systems is yet to be fully explored.

Research paper thumbnail of How Cloud Computing, IoT and Multicore Systems Affect Software Engineering Principles

2018 32nd International Conference on Advanced Information Networking and Applications Workshops (WAINA), 2018

The world has moved from a single core, local application execution to distributed, parallel, soc... more The world has moved from a single core, local application execution to distributed, parallel, social, heterogeneous etc. Standard ways of programming thus do not hold anymore and all software engineers struggle with the growing complexity that equally affects performance, maintainability and, worst of all, cost. The paper presents a concept arising from the EC Cloud Computing expert group between 2012 and 2016 as a means to move away from the hazards of traditional, Turing based computing concepts to an abstract and more natural way of programming. The concept builds up on three major building blocks, namely "Information", "Intention" and "Incentive" to define the behaviour of an application in a fashion that can abstract away from the actual way of executing an algorithm, thus allowing for a larger degree of adaptation, distribution and scale.

Research paper thumbnail of Best Practice and Definitions of Formalisation and Formalism

Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjectiv... more Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjective combination of different attainments as there are intuition, experience, information, education, decision, power of persuasion and so on, which are selected, compared and balanced against each other, which are transformed, interpreted, and used in reasoning, also to infer further knowledge. Therefore, not all the knowledge can be explicitly formalised. Knowledge and content are multi-and inter-disciplinary long-term targets and values. In practice, powerful and secure information technology can support knowledge-based works and values." "Computing means methodologies, technological means, and devices applicable for universal automatic manipulation and processing of data and information. Computing is a practical tool and has well defined purposes and goals."

Research paper thumbnail of Best Practice and Definitions of Data Value

Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjectiv... more Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjective combination of different attainments as there are intuition, experience, information, education, decision, power of persuasion and so on, which are selected, compared and balanced against each other, which are transformed, interpreted, and used in reasoning, also to infer further knowledge. Therefore, not all the knowledge can be explicitly formalised. Knowledge and content are multi-and inter-disciplinary long-term targets and values. In practice, powerful and secure information technology can support knowledge-based works and values." "Computing means methodologies, technological means, and devices applicable for universal automatic manipulation and processing of data and information. Computing is a practical tool and has well defined purposes and goals."

Research paper thumbnail of The impact of probability on archaeological interpretation

INTERNATIONAL CONFERENCE OF NUMERICAL ANALYSIS AND APPLIED MATHEMATICS ICNAAM 2019, 2020

Interpretation of the archaeological record is determined by the occurrence and distribution of f... more Interpretation of the archaeological record is determined by the occurrence and distribution of finds and their (relative and absolute) distribution over space. Comparison and categorization of finds strongly depends on their characteristics and the significance assigned to them. By default this is a very erroneous process leading to debate about individual finds, their significance and of course their interpretation. In this paper we will discuss the (objective) impact of probability on interpretation of data by measuring and comparing the impact of probability on metrics such as centrality.

Research paper thumbnail of Proceedings of the 2013 International Workshop on Software Development Lifecycle for Mobile

Proceedings of the 2013 International Workshop on Software Development Lifecycle for Mobile - DeMobile 2013, 2013

Research paper thumbnail of Best Practice and Definitions of Data-centric and Big Data : Science, Society, Law, Industry, and Engineering

The term Big Data is refering to data, which is larger and/or more complex than conventionally ha... more The term Big Data is refering to data, which is larger and/or more complex than conventionally handled with storage and computing installations. Data use with associated application scenarios can be categorised by volume, velocity, variability, vitality, veracity,. .. associated with the data."

Research paper thumbnail of Modelling and Reasoning for Indirect Sensing over Discrete-time via Markov Logic Networks

With the always increasing availability of sensor devices, there is constant unseen monitoring of... more With the always increasing availability of sensor devices, there is constant unseen monitoring of our environment. A physical activity has an impact on more sensor modalities than we could imagine. It is so vivid that distinctive patterns in the data look almost interpretable. Such knowledge, which is innate to humans, ought to be encoded and reason upon declaratively. We demonstrate the power of Markov Logic Networks for encoding uncertain knowledge to discover interesting situations from the observed evidence. We formally relate distinguishable patterns from the sensor data with knowledge about the environment and generate a rule basis for verifying and explaining occurred phenomena. We demonstrate an implementation on a real dataset and present our results.

Research paper thumbnail of Challenges for Advanced Applications in Archaeology What IT can still learn from humanities

Computer science frequently considers much of humanities and in particular archaeology "triv... more Computer science frequently considers much of humanities and in particular archaeology "trivial". The "hard challenges" were defined by sciences, such as physics. Yet these "soft" domains have been and are struggling with challenges that still exceed computational capabilities and that cannot be solved with current approaches. On the other hand, climatological models, remote sensing, agent modelling etc. all can benefit from archaeological data and approaches. In this paper we review how current computer science is insufficient to address the challenges posed in an archaeological context.

Research paper thumbnail of Reconstructing Stratigraphy

Research paper thumbnail of On Group Theory and Interpretable Time Series Primitives

Advanced Data Mining and Applications, 2022

Research paper thumbnail of Scotty: Fast a priori Structure-based Extraction from Time Series

2021 IEEE International Conference on Big Data (Big Data), 2021

The recognition and extraction of data-driven patterns is a challenging task. In vast amounts of ... more The recognition and extraction of data-driven patterns is a challenging task. In vast amounts of data, suitable techniques should prepare data that match the user's intentions. For example, "an increasing daily price of a stock, is generally followed by a sudden fall due to the endorsed profit gains". Such a statement should effortlessly be transferred as input to a data mining task, without the need for an extensive training phase building an extremely complicated model and the usage of threshold-based approaches. We propose a method for extracting lexical representations directly from the raw data, enabling others not directly compatible with real-valued data. We provide evidence that our method is fast and accurate. We use as evaluation a preliminary step for a classification task compared to state-of-the-art classifiers applying it on a publicly available dataset. Index Terms-time series representation, shapeoids, symbolic aggregate approximation, shape representation 1 According to the Oxford Dictionary, the word a priori means "Relating to or denoting reasoning or knowledge which proceeds from theoretical deduction rather than from observation or experience". We use innate knowledge to describe a trendline (e. g., the mental model of a flat line).

Research paper thumbnail of Data collection framework: project deliverable D4.1, revision 2

Research paper thumbnail of Challenges for Operating Systems arising from Manycore Architectures and their Solutions in MyThOS

While heading towards exascale computing by resource replication the coordination of consumers, i... more While heading towards exascale computing by resource replication the coordination of consumers, i.e. throughput oriented processes and applications, becomes even more challenging for the involved operating and runtime systems. Highly elastic and parallel HPC applications require clever strategical decisions for the decomposition and placement of computations and particular functions, an almost non-disruptive task and process coordination, fast allocation and clean-up of ressources, and dynamic reconfiguration and adaptation mechanisms of the management and computing infrastructure. It turns out that modern operating systems and runtime environments only partly fulfill these requirements, which leads to suboptimal resource utilization and throughput and hinders the overall performance gain by effective parallelization. This talk will summarize experiences and solutions which were gathered and developed during the course of the MyThOS project [1]. The project was funded by the German ...

Research paper thumbnail of Towards Commonsense Reasoning in AAL Environments

Research paper thumbnail of Best Practice and Definitions of Data Sciences : Beyond Statistics

Research paper thumbnail of Reengineering for parallelism in heterogeneous parallel platforms

The Journal of Supercomputing, 2018

Research paper thumbnail of Proceedings of the 2013 international workshop on Multi-cloud applications and federated clouds

Research paper thumbnail of SLA(サービスレベル一致)支援の資源管理に向けて

Lect Notes Comput Sci, 2006

Research paper thumbnail of Proceedings of the Workshop on International Transit Turnkey and Joint Development. Session 5: Environmental and Risk Management Considerations

Transportation research circular, Mar 1, 1998

Session highlights are as follows: (1) Innovative approaches in environmental management, such as... more Session highlights are as follows: (1) Innovative approaches in environmental management, such as performance based financial incentives, collaboration with the community with regards to mitigation measures, and agreements with agencies that address the review process, minimizing paperwork and cost to all parties, should be pursued. (2) Unresolved environmental issues, since they result in higher risks which in turn result in higher costs, should be avoided. An effort should be made to understand risk probabilities. A proactive environmental management process should be established early in a turnkey procurement process. (3) In a turnkey procurement process, risks must be clearly explicit. Different methods are available to manage risk. A methodology for managing risk was presented. A flow-chart process was also presented to help in the process of identifying and managing risk. The different types of risks were discussed with suggested strategies to manage them. (4) In an effort to foster and direct community participation in transit projects, the Livable Communities Initiative Program was established by the Federal Transit Administration in 1994. A video was presented showing how communities got involved and challenged projects through legal actions in the Los Angeles area. Public Participation/Community Participation should have an active role in every phase of a transit project. (5) The industry is currently developing insurance policies that would protect design liability. Currently, sureties are applicable only in the construction phase of a project. They are also investigating the possibility of Phase Contracting where the owner accepts the risk in the design phase and sureties cover the construction phase. (6) The contractors' perspective in Risk Allocation is: How well companies identify, manage, and mitigate risk directly impact their bottom line. Risk allocation must be balanced between the public and private sector. Risks must be placed where they can best be managed. (7) Bay Area Rapid Transit's experience with environmental considerations in the extension to the San Francisco International Airport was explained. (8) The consultant's perspective with regards to environmental and risk management is: The role of a consultant is to identify clients that have a sound commitment to complete projects. One must know the playing field, understanding clients, stakeholders, and regulatory framework. One must push the limits by understanding best practices, relating innovation to established objectives, and valuing good design.

Research paper thumbnail of MyThOS — Scalable OS Design for Extremely Parallel Applications

2016 Intl IEEE Conferences on Ubiquitous Intelligence & Computing, Advanced and Trusted Computing, Scalable Computing and Communications, Cloud and Big Data Computing, Internet of People, and Smart World Congress (UIC/ATC/ScalCom/CBDCom/IoP/SmartWorld), 2016

Many-core architectures trade single-thread performance for a larger number of cores. Scalable th... more Many-core architectures trade single-thread performance for a larger number of cores. Scalable throughput can be attained only by a high degree of parallelism, minimized synchronization. Whilst this is achievable for many applications, the operating system still introduces bottlenecks through non-local sharing, synchronization,, message passing. A particular challenge for highly dynamic applications, for example invasive HPC applications, elastic compute clouds, is the management of short-living application threads, processes. This paper discusses OS architecture choices based on microkernel, multikernel, distributed systems designs, our development experience in the context of the MyThOS project. Initial experiments show a much faster thread creation, activation compared to monolithic systems like Linux while providing a more flexible protection, threading model that is better suited for dynamic scenarios. However, despite significant progress in the overall domain of operating systems, the design space for scalable many-core operating systems is yet to be fully explored.

Research paper thumbnail of How Cloud Computing, IoT and Multicore Systems Affect Software Engineering Principles

2018 32nd International Conference on Advanced Information Networking and Applications Workshops (WAINA), 2018

The world has moved from a single core, local application execution to distributed, parallel, soc... more The world has moved from a single core, local application execution to distributed, parallel, social, heterogeneous etc. Standard ways of programming thus do not hold anymore and all software engineers struggle with the growing complexity that equally affects performance, maintainability and, worst of all, cost. The paper presents a concept arising from the EC Cloud Computing expert group between 2012 and 2016 as a means to move away from the hazards of traditional, Turing based computing concepts to an abstract and more natural way of programming. The concept builds up on three major building blocks, namely "Information", "Intention" and "Incentive" to define the behaviour of an application in a fashion that can abstract away from the actual way of executing an algorithm, thus allowing for a larger degree of adaptation, distribution and scale.

Research paper thumbnail of Best Practice and Definitions of Formalisation and Formalism

Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjectiv... more Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjective combination of different attainments as there are intuition, experience, information, education, decision, power of persuasion and so on, which are selected, compared and balanced against each other, which are transformed, interpreted, and used in reasoning, also to infer further knowledge. Therefore, not all the knowledge can be explicitly formalised. Knowledge and content are multi-and inter-disciplinary long-term targets and values. In practice, powerful and secure information technology can support knowledge-based works and values." "Computing means methodologies, technological means, and devices applicable for universal automatic manipulation and processing of data and information. Computing is a practical tool and has well defined purposes and goals."

Research paper thumbnail of Best Practice and Definitions of Data Value

Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjectiv... more Knowledge and Computing (Delegates and other contributors) "Knowledge is created from a subjective combination of different attainments as there are intuition, experience, information, education, decision, power of persuasion and so on, which are selected, compared and balanced against each other, which are transformed, interpreted, and used in reasoning, also to infer further knowledge. Therefore, not all the knowledge can be explicitly formalised. Knowledge and content are multi-and inter-disciplinary long-term targets and values. In practice, powerful and secure information technology can support knowledge-based works and values." "Computing means methodologies, technological means, and devices applicable for universal automatic manipulation and processing of data and information. Computing is a practical tool and has well defined purposes and goals."

Research paper thumbnail of The impact of probability on archaeological interpretation

INTERNATIONAL CONFERENCE OF NUMERICAL ANALYSIS AND APPLIED MATHEMATICS ICNAAM 2019, 2020

Interpretation of the archaeological record is determined by the occurrence and distribution of f... more Interpretation of the archaeological record is determined by the occurrence and distribution of finds and their (relative and absolute) distribution over space. Comparison and categorization of finds strongly depends on their characteristics and the significance assigned to them. By default this is a very erroneous process leading to debate about individual finds, their significance and of course their interpretation. In this paper we will discuss the (objective) impact of probability on interpretation of data by measuring and comparing the impact of probability on metrics such as centrality.

Research paper thumbnail of Proceedings of the 2013 International Workshop on Software Development Lifecycle for Mobile

Proceedings of the 2013 International Workshop on Software Development Lifecycle for Mobile - DeMobile 2013, 2013

Research paper thumbnail of Best Practice and Definitions of Data-centric and Big Data : Science, Society, Law, Industry, and Engineering

The term Big Data is refering to data, which is larger and/or more complex than conventionally ha... more The term Big Data is refering to data, which is larger and/or more complex than conventionally handled with storage and computing installations. Data use with associated application scenarios can be categorised by volume, velocity, variability, vitality, veracity,. .. associated with the data."

Research paper thumbnail of Modelling and Reasoning for Indirect Sensing over Discrete-time via Markov Logic Networks

With the always increasing availability of sensor devices, there is constant unseen monitoring of... more With the always increasing availability of sensor devices, there is constant unseen monitoring of our environment. A physical activity has an impact on more sensor modalities than we could imagine. It is so vivid that distinctive patterns in the data look almost interpretable. Such knowledge, which is innate to humans, ought to be encoded and reason upon declaratively. We demonstrate the power of Markov Logic Networks for encoding uncertain knowledge to discover interesting situations from the observed evidence. We formally relate distinguishable patterns from the sensor data with knowledge about the environment and generate a rule basis for verifying and explaining occurred phenomena. We demonstrate an implementation on a real dataset and present our results.

Research paper thumbnail of Challenges for Advanced Applications in Archaeology What IT can still learn from humanities

Computer science frequently considers much of humanities and in particular archaeology "triv... more Computer science frequently considers much of humanities and in particular archaeology "trivial". The "hard challenges" were defined by sciences, such as physics. Yet these "soft" domains have been and are struggling with challenges that still exceed computational capabilities and that cannot be solved with current approaches. On the other hand, climatological models, remote sensing, agent modelling etc. all can benefit from archaeological data and approaches. In this paper we review how current computer science is insufficient to address the challenges posed in an archaeological context.

Research paper thumbnail of Reconstructing Stratigraphy

Research paper thumbnail of On Group Theory and Interpretable Time Series Primitives

Advanced Data Mining and Applications, 2022

Research paper thumbnail of Scotty: Fast a priori Structure-based Extraction from Time Series

2021 IEEE International Conference on Big Data (Big Data), 2021

The recognition and extraction of data-driven patterns is a challenging task. In vast amounts of ... more The recognition and extraction of data-driven patterns is a challenging task. In vast amounts of data, suitable techniques should prepare data that match the user's intentions. For example, "an increasing daily price of a stock, is generally followed by a sudden fall due to the endorsed profit gains". Such a statement should effortlessly be transferred as input to a data mining task, without the need for an extensive training phase building an extremely complicated model and the usage of threshold-based approaches. We propose a method for extracting lexical representations directly from the raw data, enabling others not directly compatible with real-valued data. We provide evidence that our method is fast and accurate. We use as evaluation a preliminary step for a classification task compared to state-of-the-art classifiers applying it on a publicly available dataset. Index Terms-time series representation, shapeoids, symbolic aggregate approximation, shape representation 1 According to the Oxford Dictionary, the word a priori means "Relating to or denoting reasoning or knowledge which proceeds from theoretical deduction rather than from observation or experience". We use innate knowledge to describe a trendline (e. g., the mental model of a flat line).

Research paper thumbnail of Data collection framework: project deliverable D4.1, revision 2

Research paper thumbnail of Challenges for Operating Systems arising from Manycore Architectures and their Solutions in MyThOS

While heading towards exascale computing by resource replication the coordination of consumers, i... more While heading towards exascale computing by resource replication the coordination of consumers, i.e. throughput oriented processes and applications, becomes even more challenging for the involved operating and runtime systems. Highly elastic and parallel HPC applications require clever strategical decisions for the decomposition and placement of computations and particular functions, an almost non-disruptive task and process coordination, fast allocation and clean-up of ressources, and dynamic reconfiguration and adaptation mechanisms of the management and computing infrastructure. It turns out that modern operating systems and runtime environments only partly fulfill these requirements, which leads to suboptimal resource utilization and throughput and hinders the overall performance gain by effective parallelization. This talk will summarize experiences and solutions which were gathered and developed during the course of the MyThOS project [1]. The project was funded by the German ...

Research paper thumbnail of Towards Commonsense Reasoning in AAL Environments

Research paper thumbnail of Best Practice and Definitions of Data Sciences : Beyond Statistics

Research paper thumbnail of Reengineering for parallelism in heterogeneous parallel platforms

The Journal of Supercomputing, 2018

Research paper thumbnail of Interpolating 3d stratigraphic information from written excavation reports

Early excavation reports of around 1950 already employed meticulous recording techniques, yet fai... more Early excavation reports of around 1950 already employed meticulous recording techniques, yet fail to respect a lot of aspects concerning information that would help in generating relational information about a site’s organization at different times, such as recurring occupation of sites during specific seasons. The interested archaeologists either have to rely on their ability to mentally visualize the information or painstakingly map it out on paper and try to generate some layout information this way. Such information is however typically not shared further and rarely takes additional concerns such as geological constraints into consideration. In this paper we present an approach that generates a rough / indicative 3d model of the stratigraphic layout of an excavation on the basis of stratigraphic profiles (along the excavation walls) and potential planar maps (insofar as they exist). The approach can take the location of finds into consideration, given that the according data (coordinates and stratigraphic layer they were found in) is reported.
We propose and discuss different approaches to align the 3d model with the profile data. Obviously, the model will fail to reproduce any irregularities that were not recorded by the excavators, but we will show how geological information can be taken into consideration to improve the regular(!) properties of the stratigraphic layout, even without further information given. The paper discusses whether such models are sufficient for archaeological discussions with a particular reference to site occupation and usage.
Full paper to be published