Professor Paul Krause

Professor in Complex Systems
BSc, PhD, CMath, FIMA
+44 (0)1483 689861
32 BB 02
Friday 10am - 12pm
Contact: Sue Trendell
+44 (0)1483 688907

Academic and research departments

Department of Computer Science.


University roles and responsibilities

  • Programme Lead for MSc in Data Science
  • Senior Academic Misconduct Officer
  • CS Rep on HPC Stakeholders Committee

    My publications


    Lutfiah Al Turk, Su Wang, Paul Krause, James Wawrzynski, George M. Saleh, Hend Alsawadi, Abdulrahman Zaid Alshamrani, Tunde Peto, Andrew Bastawrous, Jingren Li, Hongying Lilian Tang (2020)Evidence Based Prediction and Progression Monitoring on Retinal Images from Three Nations, In: Translational Vision Science & Technology9(2)44 Association for Research in Vision and Ophthalmology

    Purpose: The aim of this work is to demonstrate how a retinal image analysis system, DAPHNE, supports the optimization of diabetic retinopathy (DR) screening programs for grading color fundus photography. Method: Retinal image sets, graded by trained and certified human graders, were acquired from Saudi Arabia, China, and Kenya. Each image was subsequently analyzed by the DAPHNE automated software. The sensitivity, specificity, and positive and negative predictive values for the detection of referable DR or diabetic macular edema were evaluated, taking human grading or clinical assessment outcomes to be the gold standard. The automated software’s ability to identify co-pathology and to correctly label DR lesions was also assessed. Results: In all three datasets the agreement between the automated software and human grading was between 0.84 to 0.88. Sensitivity did not vary significantly between populations (94.28%–97.1%) with specificity ranging between 90.33% to 92.12%. There were excellent negative predictive values above 93% in all image sets. The software was able to monitor DR progression between baseline and follow-up images with the changes visualized. No cases of proliferative DR or DME were missed in the referable recommendations. Conclusions: The DAPHNE automated software demonstrated its ability not only to grade images but also to reliably monitor and visualize progression. Therefore it has the potential to assist timely image analysis in patients with diabetes in varied populations and also help to discover subtle signs of sight-threatening disease onset. Translational Relevance: This article takes research on machine vision and evaluates its readiness for clinical use.

    Paul Krause, AR Razavi, Sotiris Moschoyiannis, A Marinos (2009)Stability and Complexity in Digital Ecosystems, In: 2009 3RD IEEE INTERNATIONAL CONFERENCE ON DIGITAL ECOSYSTEMS AND TECHNOLOGIESpp. 85-90 IEEE

    In this paper we explore the concept of ldquoecosystemrdquo as a metaphor in the development of the digital economy. We argue that the modelling of social ecosystems as self-organising systems is also relevant to the study of digital ecosystems. Specifically, that centralised control structures in digital ecosystems militate against emergence of innovation and adaptive response to pressures or shocks that may impact the ecosystem. We hope the paper will stimulate a more holistic approach to gaining empirical and theoretical understanding of digital ecosystems.

    Norman Fenton, Martin Neil, William Marsh, Peter Hearty, Lukasz Radlinski, Paul Krause (2007)Project Data Incorporating Qualitative Facts for Improved Software Defect Prediction, In: Third International Workshop on Predictor Models in Software Engineering (PROMISE'07: ICSE Workshops 2007)

    To make accurate predictions of attributes like defects found in complex software projects we need a rich set of process factors. We have developed a causal model that includes such process factors, both quantitative and qualitative. The factors in the model were identified as part of a major collaborative project. A challenge for such a model is getting the data needed to validate it. We present a dataset, elicited from 31 completed software projects in the consumer electronics industry, which we used for validation. The data were gathered using a questionnaire distributed to managers of recent projects. The dataset will be of interest to other researchers evaluating models with similar aims. We make both the dataset and causal model available for research use.

    Nick F. Ryman-Tubb, Paul Krause, Wolfgang Garn (2018)How Artificial Intelligence and machine learning research impacts payment card fraud detection: A survey and industry benchmark, In: Engineering Applications of Artificial Intelligence76pp. 130-157 Elsevier

    The core goal of this paper is to identify guidance on how the research community can better transition their research into payment card fraud detection towards a transformation away from the current unacceptable levels of payment card fraud. Payment card fraud is a serious and long-term threat to society (Ryman-Tubb and d’Avila Garcez, 2010) with an economic impact forecast to be $416bn in 2017 (see Appendix A).1 The proceeds of this fraud are known to finance terrorism, arms and drug crime. Until recently the patterns of fraud (fraud vectors) have slowly evolved and the criminals modus operandi (MO) has remained unsophisticated. Disruptive technologies such as smartphones, mobile payments, cloud computing and contactless payments have emerged almost simultaneously with large-scale data breaches. This has led to a growth in new fraud vectors, so that the existing methods for detection are becoming less effective. This in turn makes further research in this domain important. In this context, a timely survey of published methods for payment card fraud detection is presented with the focus on methods that use AI and machine learning. The purpose of the survey is to consistently benchmark payment card fraud detection methods for industry using transactional volumes in 2017. This benchmark will show that only eight methods have a practical performance to be deployed in industry despite the body of research. The key challenges in the application of artificial intelligence and machine learning to fraud detection are discerned. Future directions are discussed and it is suggested that a cognitive computing approach is a promising research direction while encouraging industry data philanthropy.

    A Guidotti, V Icolari, D Tarchi, A Vanelli-Coralli, SK Sharma, E Lagunas, S Maleki, S Chatzinotas, J Grotz, J Krause, E Corbel, B Evans, P Thompson (2015)Spectrum Awareness and Exploitation for Cognitive Radio Satellite Communications, In: 2015 EUROPEAN CONFERENCE ON NETWORKS AND COMMUNICATIONS (EUCNC)pp. 16-20
    Y Zheng, J Zhou, P Krause, S Latifi (2007)A model checking based test case generation framework for web services, In: International Conference on Information Technology, Proceedingspp. 715-720
    S de Lusignan, J Cashman, N Poh, G Michalakidis, A Mason, T Desombre, P Krause (2012)Conducting Requirements Analyses for Research using Routinely Collected Health Data: a Model Driven Approach., In: Stud Health Technol Inform180pp. 1105-1107

    Background: Medical research increasingly requires the linkage of data from different sources. Conducting a requirements analysis for a new application is an established part of software engineering, but rarely reported in the biomedical literature; and no generic approaches have been published as to how to link heterogeneous health data. Methods: Literature review, followed by a consensus process to define how requirements for research, using, multiple data sources might be modeled. Results: We have developed a requirements analysis: i-ScheDULEs - The first components of the modeling process are indexing and create a rich picture of the research study. Secondly, we developed a series of reference models of progressive complexity: Data flow diagrams (DFD) to define data requirements; unified modeling language (UML) use case diagrams to capture study specific and governance requirements; and finally, business process models, using business process modeling notation (BPMN). Discussion: These requirements and their associated models should become part of research study protocols.

    Doaa Sinnari, Paul Krause, Maysoon Abulkhair (2020)Effects of E-Games on the Development of Saudi Children with Attention Deficit Hyperactivity Disorder Cognitively, Behaviourally and Socially: An Experimental Study, In: Universal Access in Human-Computer Interaction: Methods, Technologies and Users Springer

    Attention Deficit Hyperactivity Disorder (ADHD) is a set of behavioural characteristics disorder, such as inattentiveness, hyperactivity and/or impulsiveness. It can affect people with different intelligent abilities, and it may affect their academic performance, social skills and generally, their lives. Usually, symptoms are not clearly recognized until the child enters school, most cases are identified between the ages 6 to 12. In the kingdom of Saudi Arabia (KSA), ADHD is a widely spread disorder among young children. Usually, they suffer from distraction and lack of focus, and hyperactivity, which reduce their academic achievements. As technology have been used in classrooms to facilitate the information delivery for students, and to make learning fun; some of these technologies have actually been applied in many schools in KSA with normal students, but unfortunately no studies were reported by the time of writing this paper. Specifically, there are no studies done for using any type of technology to help Saudi students with ADHD reaching up their peers academically. Because of that, our focus in this study is to investigate the effect of using technology, particularly e-games, to improve Saudi children with ADHD cognitively, behaviourally and socially. As well as evaluating the interaction between those children with the game interface. Thus, the investigation done through exploring the interaction of web-based games that runs on Tablets. The respondents are 17 ADHD children aged from 6–12 in classroom settings. The study involves focussing on interface of the games stimulate different executive functions in the brain, which is responsible for the most important cognitive capacities, such as: Sustained Attention, Working Memory, and Speed of Processing. Ethnographic method of research was used, which involved observing students’ behaviour in classroom, to gather information and feedback about their interaction with the application. National Institutes of Health (NIH) tests were used in pre- and post- intervention to measure improvements in attention, processing speed and working memory. Students’ test scores of main school subjects were taken pre- and post-intervention to measure enhancement in academic performance. Results show that using the application significantly improve cognitive capacities for participants, which affected their academic grades in Math, English and Science, as well as its positive influence on their behaviour. In addition, the application’s interface was found easy to use and subjectively pleasing. As a conclusion, the application considered effective and usable.

    P Krause, S de Lusignan (2010)Procuring interoperability at the expense of usability: a case study of UK National Programme for IT assurance process., In: Studies in Health Technology and Informatics: Seamless care, safe care: the challenges of interoperability and patient safety in health care: Proceedings of the EFMI Special Topic Conference155pp. 143-149

    The allure of interoperable systems is that they should improve patient safety and make health services more efficient. The UK's National Programme for IT has made great strides in achieving interoperability; through linkage to a national electronic spine. However, there has been criticism of the usability of the applications in the clinical environment.

    D Bryant, P Krause (2008)A review of current defeasible reasoning implementations, In: KNOWLEDGE ENGINEERING REVIEW23(3)pp. 227-260 CAMBRIDGE UNIV PRESS
    Sotiris Moschoyiannis, A Razavi, Paul Krause (2010)Transaction Scripts: Making Implicit Scenarios Explicit, In: Electronic Notes in Theoretical Computer Science238(6)pp. 63-79 Elsevier

    We describe a true-concurrent approach for managing dependencies between distributed and concurrent coordinator components of a long-running transaction. In previous work we have described how interactions specified in a scenario can be translated into a tuples-based behavioural description, namely vector languages. In this paper we show how reasoning against order-theoretic properties of such languages can reveal missing behaviours which are not explicitly described in the scenario but are still possible. Our approach supports the gradual refinement of scenarios of interaction into a complete set of behaviours that includes all desirable orderings of execution and prohibits emergent behaviour of the transaction. Crown Copyright © 2010.

    Y Zheng, P Krause, H Mei (2006)Asynchronous semantics and anti-patterns for interacting web services, In: QSIC 2006: Sixth International Conference on Quality Software, Proceedingspp. 74-81
    A Marinos, A Razavi, Sotiris Moschoyiannis, Paul Krause, E Damiani, J Zhang, R Chang (2009)RETRO: A Consistent and Recoverable RESTful Transaction Model, In: 2009 IEEE INTERNATIONAL CONFERENCE ON WEB SERVICES, VOLS 1 AND 2pp. 181-188 IEEE Computer Society
    Y Zheng, J Zhou, P Krause, P Muller, P Liggesmeyer, E Maehle (2007)Analysis of BPEL data dependencies, In: SEAA 2007: 33rd EUROMICRO Conference on Software Engineering and Advanced Applications, Proceedingspp. 351-358
    Sotiris Moschoyiannis, A Marinos, Paul Krause (2010)Generating SQL queries from SBVR rules, In: Lecture Notes in Computer Science: Semantic Web Rules6403pp. 128-143 Springer

    Declarative technologies have made great strides in expressivity between SQL and SBVR. SBVR models are more expressive that SQL schemas, but not as imminently executable yet. In this paper, we complete the architecture of a system that can execute SBVR models. We do this by describing how SBVR rules can be transformed into SQL DML so that they can be automatically checked against the database using a standard SQL query. In particular, we describe a formalization of the basic structure of an SQL query which includes aggregate functions, arithmetic operations, grouping, and grouping on condition. We do this while staying within a predicate calculus semantics which can be related to the standard SBVR-LF specification and equip it with a concrete semantics for expressing business rules formally. Our approach to transforming SBVR rules into standard SQL queries is thus generic, and the resulting queries can be readily executed on a relational schema generated from the SBVR model.

    Vikash Kumar Singh, Sajal Mukhopadhyay, Fatos Xhafa, Paul Krause (2020)A Quality-Assuring, Combinatorial Auction Based Mechanism for IoT-Based Crowdsourcing, In: Advances in Edge Computing: Massive Parallel Processing and Applications35pp. 148-177 IOS Press

    In this chapter, we study some research issues from IoT-based crowdsourcing in strategic setting. We have considered the scenario in the IoT-based crowdsourcing, where there are multiple task requesters and multiple IoT devices as task executors. Each task requester has multiple tasks, with the tasks having start and finish times. Based on the start and finish times, the tasks are to be distributed into different slots. On the other hand, in each slot, each IoT device requests for the set of tasks that it wants to execute along with the valuation that it will charge in exchange of its service. Both the requested set of tasks and the valuations are private informations. Given such scenario, the objective is to allocate the subset of IoT devices to the tasks in a non-conflicting manner with the objective of maximizing the social welfare. For the purpose of determining the unknown quality of the IoT devices we have utilized the concept of peer grading. Therefore, we have designed a truthful mechanism for the problem under investigation that also allows us to have the true information about the quality of the IoT devices.

    Doaa Sinnari, Paul Krause, Maysoon Abulkhair (2018)Effectiveness and Usability of a Developed Collaborative Online Tool for Children with ADHD, In: International Conference on ArtsIT, Interactivity and Game Creation

    This study evaluated the effectiveness and usability of a developed collaborative online tool (chit-chat) for children with Attention Deficit Hyperactivity Disorder (ADHD). We studied whether this tool influenced children’s Knowledge and experience exchange, motivation, behavioral abilities and social skills while using another learning tool, ACTIVATE. A total of seven Saudi children with ADHD aged from 6 to 8 years were assigned to the collaborative intervention using iPads. They were asked to play mini games that positively affect children with ADHD cognitively and behaviorally, then chat using our developed collaborative online tool, for three sessions. Progress points were measured and quantitatively analyzed before and after the intervention, thematic analysis was applied on the qualitative data. Participants showed improvements in overall performance when using the learning tool ACTIVATE. Ecollaboration was found to be effective to children with ADHD and positively influencing their knowledge, experience, motivation and social skills.

    AR Razavi, Sotiris Moschoyiannis, Paul Krause (2007)A coordination model for distributed transactions in Digital Business EcoSystems, In: Proceedings of the 2007 Inaugural IEEE-IES Digital EcoSystems and Technologies Conference, DEST 2007pp. 159-164 IEEE

    In this paper we present a model for coordinating distributed long running and multi-service transactions in Digital Business EcoSystems. The model supports various forms of service composition, which are translated into a tuples-based behavioural description that allows to reason about the required behaviour in terms of ordering, dependencies and alternative execution. The compensation mechanism warranties consistency, including omitted results, without breaking local autonomy. The proposed model is considered at the deployment level of SOA, rather than the realisation level, and is targeted to business transactions between collaborating SMEs as it respects the loose-coupling of the underlying services. © 2007 IEEE.

    A Marinos, P Krause (2009)Using SBVR, REST and Relational Databases to develop Information Systems native to the Digital Ecosystem, In: 2009 3RD IEEE INTERNATIONAL CONFERENCE ON DIGITAL ECOSYSTEMS AND TECHNOLOGIESpp. 424-429
    A Razavi, Paul Krause, Sotiris Moschoyiannis (2010)Digital Ecosystems: challenges and proposed solutions, In: N Antonopoulos, G Exarchakos, M Li, A Liotta (eds.), Handbook of research on P2P and grid systems for service-oriented computing: Models, Methodologies and Applicationspp. 1003-1031 Information Science Reference - Imprint of: IGI Global Publishing
    A Razavi, S Moschoyiannis, P Krause (2008)A self-organising environment for evolving business activities, In: Proceedings of the 3rd International Multi-Conference Computing in the Global Information Technologypp. 277-283

    In this paper we are concerned with providing support for business activities in moving from value chains to value networks. We describe a fully distributed P2P architecture which reflects the dynamics of business processes that are not governed by a single organisation. The temporary virtual networks of long-term business transactions are used as the building block of the overall scale-free business network. The design is based on dynamically formed permanent clusters resulting in a topology that is highly resilient to failures (and attacks) and is capable of reconfiguring itself to adapt to changes in business models and respond to global failures of conceptual hubs. This fosters an environment where business communities can evolve to meet emerging business opportunities and achieve sustainable growth within a digital ecosystem.

    Paul J. Krause, Bernd Freimut, Witold Suryn (2002)New directions in measurement for software quality controlpp. 129-143

    Assessing and controlling software quality is still an immature discipline. One of the reasons for this is that many of the concepts and terms that are used in discussing and describing quality are overloaded with a history from manufacturing quality. We argue in this paper that a quite distinct approach is needed to software quality control as compared with manufacturing quality control. In particular, the emphasis in software quality control is in design to fulfill business needs, rather than replication to agreed standards. We will describe how quality goals can be derived from business needs. Following that, we will introduce an approach to quality control that uses rich causal models, which can take into account human as well as technological influences. A significant concern of developing such models is the limited sample sizes that are available for eliciting model parameters. In the final section of the paper we will show how expert judgment can be reliably used to elicit parameters in the absence of statistical data. In total this provides an agenda for developing a framework for quality control in software engineering that is freed from the shackles of an inappropriate legacy.

    Sotiris Moschoyiannis, AR Razavi, YY Zheng, Paul Krause (2008)Long-running Transactions: semantics, schemas, implementation, In: Proceedings of 2nd IEEE International Conference on Digital Ecosystems and Techonologiespp. 208-215

    In this paper we describe a formal model for the distributed coordination of long-running transactions in a Digital Ecosystem for business, involving Small and Medium Enterprises (SMEs). The proposed non-interleaving model of interaction-based service composition allows for communication between internal activities of transactions. The formal semantics of the various modes of service composition are represented by standard xml schemas. The current implementation framework uses suitable asynchronous message passing techniques and reflects the design decisions of the proposed model for distributed transactions in digital ecosystems.

    Fan Zhang, David Povey, Paul J. Krause (2007)Protein Attributes Microtuning System (PAMS): an effective tool to increase protein structure prediction by data purification, In: 2007 Inaugural IEEE-IES Digital EcoSystems and Technologies Conferencepp. 565-570

    Given the expense of more direct determinations, using machine-learning schemes to predict a protein secondary structure from the sequence alone remains an important methodology. To achieve significant improvements in prediction accuracy, the authors have developed an automated tool to prepare very large biological datasets, to be used by the learning network. By focusing on improvements in data quality and validation, our experiments yielded a highest prediction accuracy of protein secondary structure of 90.97%. An important additional aspect of this achievement is that the predictions are based on a template-free statistical modeling mechanism. The performance of each different classifier is also evaluated and discussed. In this paper a protein set of 232 protein chains are proposed to be used in the prediction. Our goal is to make the tools discussed available as services in part of a digital ecosystem that supports knowledge sharing amongst the protein structure prediction community.

    PJ Krause, E Perez-Minana, J Thornton (2012)Bayesian Networks for the management of Greenhouse Gas emissions in the British agricultural sector, In: Environmental Modelling and Software35pp. 132-148 Elsevier

    Recent years have witnessed a rapid rise in the development of deterministic and non-deterministic models to estimate human impacts on the environment. An important failing of these models is the difficulty that most people have understanding the results generated by them, the implications to their way of life and also that of future generations. Within the field, the measurement of greenhouse gas emissions (GHG) is one such result. The research described in this paper evaluates the potential of Bayesian Network (BN) models for the task of managing GHG emissions in the British agricultural sector. Case study farms typifying the British agricultural sector were inputted into both, the BN model and CALM, a Carbon accounting tool used by the Country Land and Business Association (CLA) in the UK for the same purpose. Preliminary results show that the BN model provides a better understanding of how the tasks carried out on a farm impact the environment through the generation of GHG emissions. This understanding is achieved by translating the emissions information into their cost in monetary terms using the Shadow Price of Carbon (SPC), something that is not possible using the CALM tool. In this manner, the farming sector should be more inclined to deploy measures for reducing its impact. At the same time, the output of the analysis can be used to generate a business plan that will not have a negative effect on a farm's capital income.

    Yongyan Zheng, Paul J. Krause (2007)Automata Semantics and Analysis of BPEL, In: 2007 Inaugural IEEE-IES Digital EcoSystems and Technologies Conferencepp. 147-152

    Web service is an emerging paradigm for distributed computing. In order to verify web services rigorously, it is important to provide a formal semantics for flow-based web service languages such as BPEL. A suitable formal model should cover most features of BPEL. The existing formal models either abstract from data, cover a simple subset of BPEL, or omit the interactions between BPEL activities. This paper presents Web Service Automata, an extension of Mealy machines, to fulfil the formal model requirements of the web service domain. Secondly, the paper analyses the control handling and data handling of BPEL, so that these can be verified in a clear manner.

    Lee-Onn Mak, Paul J. Krause (2006)Detection & Management of Concept Drift, In: 2006 International Conference on Machine Learning and Cyberneticspp. 3486-3491

    The ability to correctly detect the location and derive the contextual information where a concept begins to drift is essential in the study of domains with changing context. This paper proposes a Top-down learning method with the incorporation of a learning accuracy mechanism to efficiently detect and manage context changes within a large dataset. With the utilisation of simple search operators to perform convergent search and JBNC with a graphical viewer to derive context information, the identified hidden context are shown with the location of the disjoint points, the contextual attributes that contribute to the concept drift, the graphical output of the true relationships between these attributes and the Boolean characterisation which is the context.

    AR Razavi, SK Moschoyiannis, PJ Krause (2007)Concurrency Control and Recovery Management for Open e-Business Transactions, In: AA McEwan, S Schneider, W Ifill, P Welch (eds.), WOTUG-30: COMMUNICATING PROCESS ARCHITECTURES 200765pp. 267-285
    N Fenton, M Neil, W Marsh, P Hearty, D Marquez, P Krause, R Mishra (2007)Predicting software defects in varying development lifecycles using Bayesian nets, In: INFORMATION AND SOFTWARE TECHNOLOGY49(1)pp. 32-43 ELSEVIER SCIENCE BV
    Matthew J Sansom, Nicolas Salazar, Paul Krause (2017)MindBeat Quintet

    MindBeat: Kinetifying thought through movement and sound. MindBeat is a software developed at the University of Surrey that facilitates collaborative thinking within a multi-disciplinary set-up. The project involved the development of an electronic space that enabled an academic ensemble to carry out an 'ideas improvisation'. Five academics from very different disciplines were invited to post short 'beats' (texts made up of no more than 3-4 sentences), around a predefined question: what makes multidisciplinary collaborations work? The beats developed in time into a progressive thread of ideas. The aim of the software was to track the emergence, development and decline of new ideas within a multidisciplinary environment, and also to be able to understand the patterns that emerge in this process by representing the ideas visually as coloured squares. The MindBeat software was launched in June 2012 as part of an electronic theatre production of Peter Handke's 'Offending the Audience'. The five Surrey academics played the parts remotely by feeding Handke's text onto the Mindbeat website as part of a three-day durational performance. An open audience was then invited to interact with the play's five voices by sitting at one of five iMac stations set up in the studio space. These five computer monitors showed the text broken down into coloured square patterns. The audience could open the text by clicking onto the coloured squares, which would reveal the short text or beat. They could then add a comment or thought to the original text. The audience's participation produced almost 500 additional beats, creating an alternative version to the Handke script. The Mindbeat software visualised this ideation as a complex pattern of coloured squares. The installation featured generative video and generative electronic music played live throughout the entire three-days. Using the colour and shape patterns of the ideas-exchange as their score, the musicians shared the software visualisation as a basis for their durational sonic improvisation.

    Lee-Onn Mak, Paul Krause (2006)Detection and Management of Concept Drift, In: Proceedings of the Fifth International Conference on Machine Learning and Cyberneticspp. 3486-3491 Institute of Electrical and Electronics Engineers

    The ability to correctly detect the location and derive the contextual information where a concept begins to drift is essential in the study of domains with changing context. This paper proposes a Top-down learning method with the incorporation of a learning accuracy mechanism to efficiently detect and manage context changes within a large dataset. With the utilisation of simple search operators to perform convergent search and JBNC with a graphical viewer to derive context information, the identified hidden context are shown with the location of the disjoint points, the contextual attributes that contribute to the concept drift, the graphical output of the true relationships between these attributes and the Boolean characterisation which is the context.

    A Razavi, A Marinos, Sotiris Moschoyiannis, Paul Krause (2009)Recovery management in RESTful Interactions, In: Proceedings of 3rd IEEE International Conference on Digital Ecosystems and Technologiespp. 436-441 IEEE

    With REST becoming a dominant architectural paradigm for web services in distributed systems, more and more use cases are applied to it, including use cases that require transactional guarantees. We believe that the loose coupling that is supported by RESTful transactions, makes this currently our preferred interaction style for digital ecosystems (DEs). To further expand its value to DEs, we propose a RESTful transaction model that satisfies both the constraints of recoverable transactions and those of the REST architectural style. We then show the correctness and applicability of the model.

    NA Manaf, Sotiris Moschoyiannis, Paul Krause (2015)Service Choreography, SBVR, and Time, In: Proceedings CONCUR 2015 - FOCLASA, EPTCS 201pp. 63-77

    We propose the use of structured natural language (English) in specifying service choreographies, focusing on the what rather than the how of the required coordination of participant services in realising a business application scenario. The declarative approach we propose uses the OMG standard Semantics of Business Vocabulary and Rules (SBVR) as a modelling language. The service choreography approach has been proposed for describing the global orderings of the invocations on interfaces of participant services. We therefore extend SBVR with a notion of time which can capture the coordination of the participant services, in terms of the observable message exchanges between them. The extension is done using existing modelling constructs in SBVR, and hence respects the standard specification. The idea is that users - domain specialists rather than implementation specialists - can verify the requested service composition by directly reading the structured English used by SBVR. At the same time, the SBVR model can be represented in formal logic so it can be parsed and executed by a machine.

    D Bryant, P Krause (2006)An implementation of a lightweight argumentation engine for agent applications, In: M Fisher, W VanDerHoek, B Konev, A Lisitsa (eds.), LOGICS IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS4160pp. 469-472
    Emily Jennings, Simon de Lusignan, Georgios Michalakidis, Paul Krause, Frank Sullivan, Harshana Liyanage, Brendan C. Delaney (2018)An instrument to identify computerised primary care research networks, genetic and disease registries prepared to conduct linked research: TRANSFoRm International Research Readiness (TIRRE) survey, In: Journal of Innovation in Health Informatics25 BCS, The Chartered Institute for IT

    Purpose The Translational Research and Patients safety in Europe (TRANSFoRm) project aims to integrate primary care with clinical research whilst improving patient safety. The TRANSFoRm International Research Readiness survey (TIRRE) aims to demonstrate data use through two linked data studies and by identifying clinical data repositories and genetic databases or disease registries prepared to participate in linked research. Method The TIRRE survey collects data at micro-, meso- and macro-levels of granularity; to fulfil data, study specific, business, geographical and readiness requirements of potential data providers for the TRANSFoRm demonstration studies. We used descriptive statistics to differentiate between demonstration-study compliant and non-compliant repositories. We only included surveys with >70% of questions answered in our final analysis, reporting the odds ratio (OR) of positive responses associated with a demonstration-study compliant data provider. Results We contacted 531 organisations within the Eurpean Union (EU). Two declined to supply information; 56 made a valid response and a further 26 made a partial response. Of the 56 valid responses, 29 were databases of primary care data, 12 were genetic databases and 15 were cancer registries. The demonstration compliant primary care sites made 2098 positive responses compared with 268 in non-use-case compliant data sources [OR: 4.59, 95% confidence interval (CI): 3.93–5.35, p < 0.008]; for genetic databases: 380:44 (OR: 6.13, 95% CI: 4.25–8.85, p < 0.008) and cancer registries: 553:44 (OR: 5.87, 95% CI: 4.13–8.34, p < 0.008).Conclusions TIRRE comprehensively assesses the preparedness of data repositories to participate in specific research projects. Multiple contacts about hypothetical participation in research identified few potential sites.

    A Marinos, P Krause (2010)Towards the web of models: A rule-driven RESTful architecture for distributed systems, In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)6403 Lpp. 251-258
    AR Razavi, Sotiris Moschoyiannis, Paul Krause (2008)A scale-free business network for digital ecosystems, In: Proceedings of 2nd IEEE International Conference on Digital Ecosystems and Technologiespp. 241-246 IEEE

    The aim of this paper is to facilitate e-business transactions between small and medium enterprises (SMEs), in a way that respects their local autonomy, within a digital ecosystem. For this purpose, we distinguish transactions from services (and service providers) by considering virtual private transaction networks (VPTNs) and virtual service networks (VSNs). These two virtual levels are optimised individually and in respect to each other. The effect of one on the other, can supply us with stability, failure resistance and small-world characteristics on one hand and durability, consistency and sustainability on the other hand. The proposed network design has a dynamic topology that adapts itself to changes in business models and availability of SMEs, and reflects the highly dynamic nature of a digital ecosystem.

    MW Shields, S Moschoyiannis, PJ Krause (2010)Behavioural Presentations and an Automata Theory of Components, In: Proceedings of Real-time and Embedded Systems (RTES 2010)
    NF Ryman-Tubb, P Krause (2011)Neural Network Rule Extraction to Detect Credit Card Fraud, In: L Iliadis, C Jayne (eds.), ENGINEERING APPLICATIONS OF NEURAL NETWORKS, PT I363pp. 101-110

    This conference addresses the question of embodiment, in a digital-era context, and how bodies can be computed; how they can be constituted via algorithmic or other such formal processes. The focus of the conference is on practices of embodiment and computation in a number of key cultural practices, most notably: computer animation, interaction design, and live performance (digital dance-theatre).

    A Marinos, Sotiris Moschoyiannis, Paul Krause (2009)Towards a RESTful infrastructure for digital ecosystems, In: Proceedings of the International Conference on Management of Emergent Digital EcoSystems, MEDES '09pp. 340-344
    PJ Krause, A Marinos (2009)An SBVR framework for RESTful Web Applications, In: Lecture Notes in Computer Science: Rule Interchange and Applications5858pp. 144-158

    We propose a framework that can be used to produce functioning web applications from SBVR models. To achieve this, we begin by discussing the concept of declarative application generation and examining the commonalities between SBVR and the RESTful architectural style of the web. We then show how a relational database schema and RESTful interface can be generated from an SBVR model. In this context, we discuss how SBVR can be used to semantically describe hypermedia on the Web and enhance its evolvability and loose coupling properties. Finally, we show that this system is capable of exhibiting process-like behaviour without requiring explicitly defined processes.

    Sotiris Moschoyiannis, Paul Krause, MW Shields (2009)A True-Concurrent Interpretation of Behavioural Scenarios. In Proc. of ETAPS 2007 - Formal Foundations of Embedded Software and Component-Based Software Architectures (FESCA'07), In: Electronic Notes in Theoretical Computer Science203(7)pp. 3-22 Elsevier

    We describe a translation of scenarios given in UML 2.0 sequence diagrams into a tuples-based behavioural model that considers multiple access points for a participating instance and exhibits true-concurrency. This is important in a component setting since different access points are connected to different instances, which have no knowledge of each other. Interactions specified in a scenario are modelled using tuples of sequences, one sequence for each access point. The proposed unfolding of the sequence diagram involves mapping each location (graphical position) onto the so-called component vectors. The various modes of interaction (sequential, alternative, concurrent) manifest themselves in the order structure of the resulting set of component vectors, which captures the dependencies between participating instances. In previous work, we have described how (sets of) vectors generate concurrent automata. The extension to our model with sequence diagrams in this paper provides a way to verify the diagram against the state-based model.

    F Xhafa, J Wang, X Chen, JK Liu, J Li, P Krause (2014)An efficient PHR service system supporting fuzzy keyword search and fine-grained access control, In: SOFT COMPUTING18(9)pp. 1795-1802 SPRINGER
    J. C. Jacobs, J. H. van Moll, P. J. Krause, R. J. Kusters, J. J. M. Trienekens (2003)Effects of virtual development on product quality: exploring defect causes, In: Eleventh Annual International Workshop on Software Technology and Engineering Practicepp. 6-15

    This paper explores the effects of virtual development on product quality, from the viewpoint of "conformance to specifications". Specifically, causes of defect injection and non- or late-detection are explored. Because of the practical difficulties of obtaining hard project-specific defect data, an approach was taken that relied upon accumulated expert knowledge. The accumulated expert knowledge based approach was found to be a practical alternative to an in-depth defect causal analysis on a per-project basis. Defect injection causes seem to be concentrated in the requirements specification phases. Defect dispersion is likely to increase, as requirements specifications are input for derived requirements specifications in multiple, related sub-projects. Similarly, a concentration of causes for the non- or late detection of defects was found in the Integration Test phases. Virtual development increases the likelihood of defects in the end product because of the increased likelihood of defect dispersion, because of new virtual development related defect causes, and because causes already existing in co-located development are more likely to occur.

    A Razavi, A Marinos, Sotiris Moschoyiannis, Paul Krause (2009)RESTful Transactions Supported by the Isolation Theorems, In: WEB ENGINEERING, PROCEEDINGS5648pp. 394-409 Springer

    With REST becoming the dominant architectural paradigm for web services in distributed systems, more and more use cases are applied to it, including use cases that require transactional guarantees. We propose a RESTful transaction model that satisfies both the constraints of transactions and those of the REST architectural style. We then apply the isolation theorems to prove the robustness of its properties on a formal level.

    A Razavi, Sotiris Moschoyiannis, Paul Krause (2009)An open digital environment to support business ecosystems, In: Peer-to-Peer Networking and Applications2(4)pp. 367-397 SPRINGER

    We present a Peer-to-Peer network design which aims to support business activities conducted through a network of collaborations that generate value in different, mutually beneficial, ways for the participating organisations. The temporary virtual networks formed by long-term business transactions that involve the execution of multiple services from different providers are used as the building block of the underlying scale-free business network. We show how these local interactions, which are not governed by a single organisation, give rise to a fully distributed P2P architecture that reflects the dynamics of business activities. The design is based on dynamically formed permanent clusters of nodes, the so-called Virtual Super Peers (VSPs), and this results in a topology that is highly resilient to certain types of failure (and attacks). Furthermore, the proposed P2P architecture is capable of reconfiguring itself to adapt to the usage that is being made of it and respond to global failures of conceptual hubs. This fosters an environment where business communities can evolve to meet emerging business opportunities and achieve sustainable growth within a digital ecosystem.

    H Liyanage, S de Lusignan, ST Liaw, CE Kuziemsky, F Mold, P Krause, D Fleming, S Jones (2014)Big Data Usage Patterns in the Health Care Domain: A Use Case Driven Approach Applied to the Assessment of Vaccination Benefits and Risks. Contribution of the IMIA Primary Healthcare Working Group., In: Yearb Med Inform9pp. 27-35 Schattauer

    BACKGROUND: Generally benefits and risks of vaccines can be determined from studies carried out as part of regulatory compliance, followed by surveillance of routine data; however there are some rarer and more long term events that require new methods. Big data generated by increasingly affordable personalised computing, and from pervasive computing devices is rapidly growing and low cost, high volume, cloud computing makes the processing of these data inexpensive. OBJECTIVE: To describe how big data and related analytical methods might be applied to assess the benefits and risks of vaccines. METHOD: We reviewed the literature on the use of big data to improve health, applied to generic vaccine use cases, that illustrate benefits and risks of vaccination. We defined a use case as the interaction between a user and an information system to achieve a goal. We used flu vaccination and pre-school childhood immunisation as exemplars. RESULTS: We reviewed three big data use cases relevant to assessing vaccine benefits and risks: (i) Big data processing using crowdsourcing, distributed big data processing, and predictive analytics, (ii) Data integration from heterogeneous big data sources, e.g. the increasing range of devices in the "internet of things", and (iii) Real-time monitoring for the direct monitoring of epidemics as well as vaccine effects via social media and other data sources. CONCLUSIONS: Big data raises new ethical dilemmas, though its analysis methods can bring complementary real-time capabilities for monitoring epidemics and assessing vaccine benefit-risk balance.

    D Bryant, Paul Krause, Sotiris Moschoyiannis, M Fisher, W VanDerHoek, B Konev, A Lisitsa (2006)A tool to facilitate agent deliberation. In Proc. of 10th European Conference on Logics in Artificial Intelligence (JELIA'06), In: Lecture Notes in Computer Science4160pp. 465-468 Springer

    In this paper we present a prototype of a tool that demonstrates how existing limitations in ensuring an agent’s compliance to an argumentation-based dialogue protocol can be overcome. We also present the implementation of compliance enforcement components for a deliberation dialogue protocol, and an application that enables two human participants to engage in an efficiently moderated dialogue, where all inappropriate utterances attempted by an agent are blocked and prevented from inclusion within the dialogue.

    Sotiris Moschoyiannis, Paul Krause, P Georgiou (2012)An animation tool for exploring transactions in a de, In: IEEE International Conference on Digital Ecosystems and Technologies IEEE

    The concept of a digital ecosystem (DE) has been used to explore scenarios in which multiple online services and resources can be accessed by users without there being a single point of control. In previous work we have described how the so-called transaction languages can express concurrent and distributed interactions between online services in a transactional environment. In this paper we outline how transaction languages capture the history of a long-running transaction and highlight the benefits of our true-concurrent approach in the context of DEs. This includes support for the recovery of a long-running transaction whenever some failure is encountered. We introduce an animation tool that has been developed to explore the behaviours of long-running transactions within our modelling environment. Further, we discuss how this work supports the declarative approach to the development of open distributed applications. © 2012 IEEE.

    Nihal Esam, Areej Abbas, Paul Krause (2020)Towards Empowering Hearing Impaired Students' Skills in Computing and Technology, In: International Journal of Advanced Computer Science and Applications8(1) SAI Organization

    Studies have shown that deaf and hearing-impaired students have many difficulties in learning applied disciplines such as Medicine, Engineering, and Computer Programming. This study aims to investigate the readiness of deaf students to pursue higher education in applied sciences, more specifically in computer science. This involves investigating their capabilities in computer skills and applications. Computer programming is an integral component in the technological field that can facilitate the development of further scientific advances. Devising a manner of teaching the deaf and hearing-impaired population will give them an opportunity to contribute to the technology sector. This would allow these students to join the scientific world when otherwise; they are generally unable to participate because of the limitations they encounter. The study showed that deaf students in Jeddah are eager to continue their higher education and that a large percentage of these students are keen on studying computer science, particularly if they are provided with the right tools.

    PJ Krause, N Sabry (2013)Optimal Green Virtual Machine Migration Model, In: International Journal of Business Data Communications and Networking9(3)pp. 35-52

    Cloud computing provides the opportunity to migrate virtual machines to “follow-the-green” data centres. That is, to migrate virtual machines between green data centres on the basis of clean energy availability, to mitigate the environmental impact of carbon footprint emissions and energy consumption. The virtual machine migration problem can be modelled to maximize the utility of computing resources or minimizing the cost of using computing resources. However, this would ignore the network energy consumption and its impact on the overall CO2 emissions. Unless this is taken into account the extra data traffic due to migration of data could then cause an increase in brown energy consumption and eventually lead to an unintended increase in carbon footprint emissions. Energy consumption is a key aspect in deploying distributed service in cloud networks within decentralized service delivery architectures. In this paper, the authors address an optimiza- tion view of the problem of locating a set of cloud services on a set of sites green data centres managed by a service provider or hybrid cloud computing brokerage. The authors’ goal is to minimize the overall network energy consumption and carbon footprint emissions for accessing the cloud services for any pair of data centres i and j. The authors propose an optimization migration model based on the development of integer linear programming (ILP) models, to identify the leverage of green energy sources with data centres and the energy consumption of migrating VMs.

    Sotiris Moschoyiannis, PJ Krause (2015)True Concurrency in Long-running Transactions for Digital Ecosystems, In: Fundamenta Informaticae138(4)pp. 483-514 IOS PRESS
    PJ Krause, N Fenton, M Neil, W Marsh, P Hearty, L Radlinski (2008)On the effectiveness of early life cycle defect prediction with Bayesian Nets, In: Empirical Software Engineering: an international journal13(5)pp. 499-537 Springer

    Standard practice in building models in software engineering normally involves three steps: collecting domain knowledge (previous results, expert knowledge); building a skeleton of the model based on step 1 including as yet unknown parameters; estimating the model parameters using historical data. Our experience shows that it is extremely difficult to obtain reliable data of the required granularity, or of the required volume with which we could later generalize our conclusions. Therefore, in searching for a method for building a model we cannot consider methods requiring large volumes of data. This paper discusses an experiment to develop a causal model (Bayesian net) for predicting the number of residual defects that are likely to be found during independent testing or operational usage. The approach supports (1) and (2), does not require (3), yet still makes accurate defect predictions (an R 2 of 0.93 between predicted and actual defects). Since our method does not require detailed domain knowledge it can be applied very early in the process life cycle. The model incorporates a set of quantitative and qualitative factors describing a project and its development process, which are inputs to the model. The model variables, as well as the relationships between them, were identified as part of a major collaborative project. A dataset, elicited from 31 completed software projects in the consumer electronics industry, was gathered using a questionnaire distributed to managers of recent projects. We used this dataset to validate the model by analyzing several popular evaluation measures (R 2, measures based on the relative error and Pred). The validation results also confirm the need for using the qualitative factors in the model. The dataset may be of interest to other researchers evaluating models with similar aims. Based on some typical scenarios we demonstrate how the model can be used for better decision support in operational environments. We also performed sensitivity analysis in which we identified the most influential variables on the number of residual defects. This showed that the project size, scale of distributed communication and the project complexity cause the most of variation in number of defects in our model. We make both the dataset and causal model available for research use.

    Yongyan Zheng, Jiong Zhou, Paul Krause (2007)An Automatic Test Case Generation Framework for Web Services, In: Journal of Software (JSW)2pp. 64-77

    BPEL (Business Process Execution Language) as a de-facto standard for web service orchestration has drawn particularly attention from researchers and industries. BPEL is a semi-formal flow language with complex features such as concurrency and hierarchy. To test a model thoroughly, we need to cover different execution scenarios. As is well known, it is tedious, time consuming, and error prone to design test cases manually, especially for complex modelling languages. Hence, it is desirable to apply existing model-based-testing techniques in the domain of web services. We proposed WSA (Web Service Automata) to be the operational semantics for BPEL. Based on WSA, we propose a model checking based test case generation framework for BPEL. The SPIN and NuSMV model checkers are used as the test generation engine, and the conventional structural test coverage criteria are encoded into LTL and CTL temporal logic. State coverage and transition coverage are used for BPEL control flow testing, and all-du-path coverage is used for BPEL data flow testing. Two levels of test cases can be generated to test whether the implementation of web services conforms to the BPEL behaviour and WSDL interface models. The generated test cases are executed on the JUnit test execution engine.

    Sotiris Moschoyiannis, Paul Krause, D Bryant, P McBurney (2009)Verifiable Protocol Design for Agent Argumentation Dialogues, In: 2009 3RD IEEE INTERNATIONAL CONFERENCE ON DIGITAL ECOSYSTEMS AND TECHNOLOGIESpp. 630-635 IEEE

    We describe a formal approach to protocol design for dialogues between autonomous agents in a digital ecosystem that involve the exchange of arguments between the participants. We introduce a vector language-based representation of argumentation protocols, which captures the interplay between different agentspsila moves in a dialogue in a way that (a) determines the legal moves that are available to each participant, in each step, and (b) records the dialogue history. We use UML protocol state machines (PSMs) to model a negotiation dialogue protocol at both the individual participant level (autonomous agent viewpoint) and the dialogue level (overall interaction viewpoint). The underlying vector semantics is used to verify that a given dialogue was played out in compliance with the corresponding protocol.

    To define the key concepts which inform whether a system for collecting, aggregating and processing routine clinical data for research is fit for purpose.

    A Marinos, P Krause (2009)What, not How: A generative approach to service composition, In: 2009 3RD IEEE INTERNATIONAL CONFERENCE ON DIGITAL ECOSYSTEMS AND TECHNOLOGIESpp. 430-435
    RM Hierons, K Bogdanov, JP Bowen, R Cleaveland, J Derrick, J Dick, M Gheorghe, M Harman, K Kapoor, P Krause, G Luettgen, AJH Simons, S Vilkomir, MR Woodward, H Zedan (2009)Using Formal Specifications to Support Testing, In: ACM COMPUTING SURVEYS41(2)ARTN 9 ASSOC COMPUTING MACHINERY
    H Liyanage, P Krause, S De Lusignan (2015)Using ontologies to improve semantic interoperability in health data., In: J Innov Health Inform22(2)pp. 309-315

    The present-day health data ecosystem comprises a wide array of complex heterogeneous data sources. A wide range of clinical, health care, social and other clinically relevant information are stored in these data sources. These data exist either as structured data or as free-text. These data are generally individual person-based records, but social care data are generally case based and less formal data sources may be shared by groups. The structured data may be organised in a proprietary way or be coded using one-of-many coding, classification or terminologies that have often evolved in isolation and designed to meet the needs of the context that they have been developed. This has resulted in a wide range of semantic interoperability issues that make the integration of data held on these different systems changing. We present semantic interoperability challenges and describe a classification of these. We propose a four-step process and a toolkit for those wishing to work more ontologically, progressing from the identification and specification of concepts to validating a final ontology. The four steps are: (1) the identification and specification of data sources; (2) the conceptualisation of semantic meaning; (3) defining to what extent routine data can be used as a measure of the process or outcome of care required in a particular study or audit and (4) the formalisation and validation of the final ontology. The toolkit is an extension of a previous schema created to formalise the development of ontologies related to chronic disease management. The extensions are focused on facilitating rapid building of ontologies for time-critical research studies.

    Paul Krause (2020)Information Theory and Medical Decision Making, In: Phillip Scott, Nicolette de Keizer, Andrew Georgiou (eds.), Volume 263: Applied Interdisciplinary Theory in Health Informatics263pp. 23-34 IOS Press

    Information theory has gained application in a wide range of disciplines, including statistical inference, natural language processing, cryptography and molecular biology. However, its usage is less pronounced in medical science. In this chapter, we illustrate a number of approaches that have been taken to applying concepts from information theory to enhance medical decision making. We start with an introduction to information theory itself, and the foundational concepts of information content and entropy. We then illustrate how relative entropy can be used to identify the most informative test at a particular stage in a diagnosis. In the case of a binary outcome from a test, Shannon entropy can be used to identify the range of values of test results over which that test provides useful information about the patient’s state. This, of course, is not the only method that is available, but it can provide an easily interpretable visualization. The chapter then moves on to introduce the more advanced concepts of conditional entropy and mutual information and shows how these can be used to prioritise and identify redundancies in clinical tests. Finally, we discuss the experience gained so far and conclude that there is value in providing an informed foundation for the broad application of information theory to medical decision making.

    Fatos Xhafa, Burak Kilic, Paul Krause (2019)Evaluation of IoT stream processing at edge computing layer for semantic data enrichment, In: Future Generation Computer Systems105pp. 730-736 Elsevier

    The fast development of Internet of Things (IoT) computing and technologies has prompted a decentralization of Cloud-based systems. Indeed, sending all the information from IoT devices directly to the Cloud is not a feasible option for many applications with demanding requirements on real-time response, low latency, energy-aware processing and security. Such decentralization has led in a few years to the proliferation of new computing layers between Cloud and IoT, known as Edge computing layer, which comprises of small computing devices (e.g. Raspberry Pi) to larger computing nodes such as Gateways, Road Side Units, Mini Clouds, MEC Servers, Fog nodes, etc. In this paper, we study the challenges of processing an IoT data stream in an Edge computing layer. By using a real life data stream set arising from a car data stream as well as a real infrastructure using Raspberry Pi and Node-Red server, we highlight the complexities of achieving real time requirements of applications based on IoT stream processing.

    MW Shields, S Moschoyiannis, PJ Krause (2010)Primes in Component Languages, In: Proceedings of Real-time and Embedded Systems (RTES 2010)
    D Bryant, PJ Krause, GAW Vreeswijk (2006)Argue tuProlog: A Lightweight Argumentation Engine for Agent Applications, In: PE Dunne, TJM BenchCapon (eds.), COMPUTATIONAL MODELS OF ARGUMENT144pp. 27-32
    Anjan Bandyopadhyay, Vikash Kumar Singh, Sajal Mukhopadhyay, Ujjwal Rai, Fatos Xhafa, Paul Krause (2020)Matching IoT Devices to the Fog Service Providers: A Mechanism Design Perspective, In: Sensors (Basel, Switzerland)20(23)6761 MDPI

    In the Internet of Things (IoT) + Fog + Cloud architecture, with the unprecedented growth of IoT devices, one of the challenging issues that needs to be tackled is to allocate Fog service providers (FSPs) to IoT devices, especially in a game-theoretic environment. Here, the issue of allocation of FSPs to the IoT devices is sifted with game-theoretic idea so that utility maximizing agents may be benign. In this scenario, we have multiple IoT devices and multiple FSPs, and the IoT devices give preference ordering over the subset of FSPs. Given such a scenario, the goal is to allocate at most one FSP to each of the IoT devices. We propose mechanisms based on the theory of mechanism design without money to allocate FSPs to the IoT devices. The proposed mechanisms have been designed in a flexible manner to address the long and short duration access of the FSPs to the IoT devices. For analytical results, we have proved the economic robustness, and probabilistic analyses have been carried out for allocation of IoT devices to the FSPs. In simulation, mechanism efficiency is laid out under different scenarios with an implementation in Python.

    AR Razavi, Sotiris Moschoyiannis, Paul Krause (2007)Concurrency control and recovery management for open e-business transactions, In: Proc. of Communicating Process Architectures: Concurrent Systems Engineering Series65pp. 267-285 IOS Press

    Concurrency control mechanisms such as turn-taking, locking, serialization, transactional locking mechanism, and operational transformation try to provide data consistency when concurrent activities are permitted in a reactive system. Locks are typically used in transactional models for assurance of data consistency and integrity in a concurrent environment. In addition, recovery management is used to preserve atomicity and durability in transaction models. Unfortunately, conventional lock mechanisms severely (and intentionally) limit concurrency in a transactional environment. Such lock mechanisms also limit recovery capabilities. Finally, existing recovery mechanisms themselves afford a considerable overhead to concurrency. This paper describes a new transaction model that supports release of early results inside and outside of a transaction, decreasing the severe limitations of conventional lock mechanisms, yet still warranties consistency and recoverability of released resources (results). This is achieved through use of a more flexible locking mechanism and by using two types of consistency graph. This provides an integrated solution for transaction management, recovery management and concurrency control. We argue that these are necessary features for management of long-term transactions within "digital ecosystems" of small to medium enterprises.

    AA Allinjawi, HA Al-Nuaim, P Krause (2014)An Achievement Degree Analysis Approach to Identifying Learning Problems in Object-Oriented Programming, In: ACM TRANSACTIONS ON COMPUTING EDUCATION14(3)ARTN 20 ASSOC COMPUTING MACHINERY
    AR Razavi, PJ Krause, A Strømmen-Bakhtiar (2010)From Business Ecosystems towards Digital Business Ecosystems, In: 4th IEEE International Conference on Digital Ecosystems and Technologies - Conference Proceedings of IEEE-DEST 2010, DEST 2010pp. 290-295

    This paper is an introduction to Business Ecosystems and its improvements as represented in Digital Business Ecosystem. In traditional non-digital economic ecosystems, the hubs, once established are static. It is difficult to create new hubs and as such these established hubs enjoy considerable power in the marketplace. Indeed, in many instances, these hubs become the main impediments for the growth of smaller nodes. In other instances, they try to misuse their powers by creating monopolies or oligopolies. In contrast, it is hoped that Digital Business Ecosystems, by virtue of their loosely-coupled and self-organizing properties, will help Small and Medium Enterprises to create a fully distributed network, sidestepping the above mentioned problems of dominance of the large hubs in the market. The authors try to provide a brief introduction to Business Ecosystems and focus on the reasons and advantages of such a move/transition towards Digital Business Ecosystems. © 2010 IEEE.

    Anjan Bandyopadhyay, Fatos Xhafa, Saurav Mallik, Paul Krause, Sajal Mukhopadhyay, Vikash Kumar Singh, Ujjwal Maulik (2019)A Framework for Allocation of IoT Devices to the Fog Service Providers in Strategic Setting, In: Leonard Barolli, Peter Hellinckx, Juggapong Natwichai (eds.), Advances on P2P, Parallel, Grid, Cloud and Internet Computing96pp. 340-351 Springer, Cham

    In the IoT+Fog+Cloud architecture, with the ever increasing growth of IoT devices, allocation of IoT devices to the Fog service providers will be challenging and needs to be addressed properly in both the strategic and non-strategic settings. In this paper, we have addressed this allocation problem in strategic settings. The framework is developed under the consideration that the IoT devices (e.g. wearable devices) collecting data (e.g. health statistics) are deploying it to the Fog service providers for some meaningful processing free of cost. Truthful and Pareto optimal mechanisms are developed for this framework and are validated with some simulations.

    L-O Mak, P Krause (2006)Detection & management of concept drift, In: Proceedings of 2006 International Conference on Machine Learning and Cybernetics, Vols 1-7pp. 3486-3491
    AR Razavi, PJ Malone, Sotiris Moschoyiannis, B Jennings, Paul Krause (2007)A distributed transaction and accounting model for digital ecosystem composed services, In: 2007 Inaugural IEEE international conference on digital ecosystems and technologiespp. 215-218 IEEE Computer Society
    SJ Webb, T Hanser, B Howlin, P Krause, JD Vessey (2014)Feature combination networks for the interpretation of statistical machine learning models: application to Ames mutagenicity, In: JOURNAL OF CHEMINFORMATICS6ARTN 8pp. ?-? BIOMED CENTRAL LTD
    PJ Krause, PC Liang (2016)Smartphone-based Real-time Indoor Location Tracking with One-meter Precision, In: IEEE Journal of Biomedical and Health Informatics20(3)pp. 756-762 IEEE

    Monitoring the activities of daily living of the elderly at home is widely recognized as useful for the detection of new or deteriorating health conditions. However, the accuracy of existing indoor location tracking systems remains unsatisfactory. The aim of this study was, therefore, to develop a localization system that can identify a patient's real-time location in a home environment with maximum estimation error of 2 m at a 95% confidence level. A proof-of-concept system based on a sensor fusion approach was built with considerations for lower cost, reduced intrusiveness, and higher mobility, deployability, and portability. This involved the development of both a step detector using the accelerometer and compass of an iPhone 5, and a radio-based localization subsystem using a Kalman filter and received signal strength indication to tackle issues that had been identified as limiting accuracy. The results of our experiments were promising with an average estimation error of 0.47 m. We are confident that with the proposed future work, our design can be adapted to a home-like environment with a more robust localization solution.

    A Marinos, S Moschoyiannis, PJ Krause (2010)An SBVR to SQL Compiler., In: M Palmirani, MO Shafiq, E Francesconi, F Vitali (eds.), RuleML Challenge649