Gør som tusindvis af andre bogelskere
Tilmeld dig nyhedsbrevet og få gode tilbud og inspiration til din næste læsning.
Ved tilmelding accepterer du vores persondatapolitik.Du kan altid afmelde dig igen.
IMPROVE stands for "e;Information Technology Support for Collaborative and Distributed Design Processes in Chemical Engineering"e; and is a large joint project of research institutions at RWTH Aachen University. This volume summarizes the results after 9 years of cooperative research work. The focus of IMRPOVE is on understanding, formalizing, evaluating, and, consequently, improving design processes in chemical engineering. In particular, IMPROVE focuses on conceptual design and basic engineering, where the fundamental decisions concerning the design or redesign of a chemical plant are undertaken. Design processes are analyzed and evaluated in collaboration with industrial partners.
As software systems become ubiquitous, the issues of dependability become more and more crucial. Given that solutions to these issues must be considered from the very beginning of the design process, it is reasonable that dependability is addressed at the architectural level. This book was born of an effort to bring together the research communities of software architectures and dependability.This state-of-the-art survey contains expanded and peer-reviewed papers based on the carefully selected contributions to two workshops: the Workshop on Architecting Dependable Systems (WADS 2007), organized at the 2007 International Conference on Dependable Systems and Networks (DSN 2007), held in Edinburgh, UK in June 2007 and the Third Workshop on the Role of Software Architecture for Testing and Analysis (ROSATEA 2007) organized as part of a federated conference on Component-Based Software Engineering and Software Architecture (CompArch 2007), held in Medford, MA, USA in July 2007. It also contains a number of invited papers written by recognized experts in the area. The 14 papers are organized in topical sections on critical infrastructures, rigorous design/fault tolerance, and verification and validation.
On behalf of the Organizing Committee we are pleased to present the p- ceedings of the 2008 Symposium on Component-Based Software Engineering (CBSE). CBSE is concerned with the development of software-intensivesystems from independently developed software-building blocks (components), the - velopment of components, and system maintenance and improvement by means of component replacement and customization. CBSE 2008 was the 11th in a series of events that promote a science and technology foundation for achieving predictable quality in software systems through the use of software component technology and its associated software engineering practices. Wewerefortunateto haveadedicatedProgramCommitteecomprisingmany internationallyrecognizedresearchersandindustrialpractitioners.Wewouldlike to thank the members of the Program Committee and associated reviewers for their contribution in making this conference a success. We received 70 subm- sions and each paper was reviewed by at least three Program Committee m- bers (four for papers with an author on the Program Committee). The entire reviewing process was supported by the Conference Management Toolkit p- vided by Microsoft. In total, 20 submissions were accepted as full papers and 3 submissions were accepted as short papers.
The Workshop on Self-sustaining Systems (S3) is a forum for the discussion of topics relating to computer systems and languages that are able to bootstrap, implement, modify, and maintain themselves. One property of these systems is that their implementation is based onsmall but powerfulabstractions;examples include (amongst others) Squeak/Smalltalk, COLA, Klein/Self, PyPy/Python, Rubinius/Ruby,andLisp.Suchsystemsaretheenginesoftheirownreplacement, giving researchers and developers great power to experiment with, and explore future directions from within, their own small language kernels. S3 took place on May 15-16, 2008 at the Hasso-Plattner-Institute (HPI) in Potsdam, Germany. It was an exciting opportunity for researchers and prac- tioners interested in self-sustaining systems to meet and share their knowledge, experience, and ideas for future research and development. S3 provided an - portunity for a community to gather and discuss the need for self-sustainability in software systems, and to share and explore thoughts on why such systems are needed and how they can be created and deployed. Analogies were made, for example, with evolutionary cycles, and with urban design and the subsequent inevitable socially-driven change. TheS3participantsleftwithagreatersenseofcommunityandanenthusiasm for probing more deeply into this subject. We see the need for self-sustaining systems becoming critical not only to the developer's community, but to e- users in business, academia, learning and play, and so we hope that this S3 workshop will become the ?rst of many.
The 2008 International Symposium on Rule Interchange and Applications (RuleML th 2008), collocated in Orlando, Florida, with the 11 International Business Rules - rum, was the premier place to meet and to exchange ideas from all fields of rules te- nologies. The aim of RuleML 2008 was both to present new and interesting research results and to show successfully deployed rule-based applications. This annual sym- sium is the flagship event of the Rule Markup and Modeling Initiative (RuleML). The RuleML Initiative (www.ruleml.org) is a non-profit umbrella organization of several technical groups organized by representatives from academia, industry and government working on rule technologies and applications. Its aim is to promote the study, research and application of rules in heterogeneous distributed environments such as the Web. RuleML maintains effective links with other major international societies and acts as intermediary between various 'specialized' rule vendors, appli- tions, industrial and academic research groups, as well as standardization efforts from, for example, W3C, OMG, and OASIS.
Wearepleasedtopresenttheproceedingsofthe14thMontereyWorkshop,which tookplaceSeptember10-13,2007inMonterey,CA,USA. Inthispreface,wegive the reader an overview of what took place at the workshop and introduce the contributions in this Lecture Notes in Computer Science volume. A complete introduction to the theme of the workshop, as well as to the history of the Monterey Workshop series, can be found in Luqi and Kordon's "e;Advances in Requirements Engineering: Bridging the Gap between Stakeholders' Needs and Formal Designs"e; in this volume. This paper also contains the case study that many participants used as a problem to frame their analyses, and a summary of the workshop's results. The workshop consisted of three keynote talks, three panels, presentations of peer-reviewed papers, as well as presentations of various position papers by the participants. The keynote speakers at this year's workshop were Daniel Berry, Aravind Joshi, and Lori Clarke. Each of their talks was used to set the tone for the p- sentations and discussions for that particular day. Daniel Berry presented an overview of the needs and challenges of natural language processing in requi- ments engineering, with a special focus on ambiguity in his talk "e;Ambiguity in Natural Language Requirements. "e; Aravind Joshi provided an overview of current natural language processing research in discourse analysis in the talk "e;Some Recent Developments in Natural Language Processing. "e; Finally, Lori Clarke showed how to combine formal requirements speci?cation with natural language processing to cope with the complex domain of medical information processes in "e;Getting the Details Right.
This volume contains the proceedings of the International Middleware Con- rence, held in Leuven, Belgium during December 1-4, 2008. This year marked the ninth rendition of this annual conference in its current format initially - opted in 1998, aspiring to serve as the premier venue focusing exclusively on important middleware results. A lot has happened over the intervening 10-year span. Middleware has become pervasive in an increasinglyinterconnected world, with its concepts now securely embedded in the notional architectures driving forward the information age. The conference continues to be a focal point for important new initiatives and results for new generations of middleware. With each succeeding year, it has become an even more competitive publishing venue, further extending its selectivity this year as well. Of the 117 submissions, - ly 21 were able to receive acceptance invitations, for an acceptance rate of less than18%. Competitiveindeed. Additionally,thesesubmissionscontinuetocome from all over the globe, spanning authors from 23 countries. A truly world-wide endeavor. But other things have also changed as we turn the corner on the ?rst decade of this conference. In many ways, middleware has achieved signi?cant success where it really counts: in terms of technical innovations, advanced capabilities, successful ideas, and ?elded systems which permeate society, industry, gov- nment, and academia. With this success comes maturity.
This volume includes the proceedings of all the ?ve workshops that were held as part of the International Conference on Service-Oriented Systems 2007 (ICSOC 2007): - Mashups 2007: First International Workshop on Web APIs and Services Mashups - NFPSLA-SOC 2007: Workshop on Non-Functional Properties and Service Level Agreements in Service-Oriented Computing - SeMSoc 2007: Second International SeMSoC Workshop Business-Oriented Aspects Concerning Semantics and Methodologies in Service-Oriented Computing - TSOA 2007: First International Workshop on Telecom Service-Oriented Architectures - WESOA2007:ThirdInternationalWorkshoponEngineeringService-Oriented Applications: Analysis, Design and Composition These workshops o?ered interesting presentations and discussions on a wide rangeofhottopicsinservice-orientedcomputing:developmentofmashups,m- agement of non-functional properties and service level agreements, engineering approaches, and semantic methodologies. Also, the crucial application domain of telecom services and service architectures was investigated. The ?ve workshops were selected out of eight submissions. During the - lection process we encouraged the merging of workshop proposals with similar scope. We are grateful to all workshop organizers. Without their valuable ideas and support the workshop sessions at ICSOC as well as this volume would not have been possible. We also thank all authors for their active participation in the events, for the quality of their papers, and for being patient with us during the long process of compilation of this volume.
Declarative languages have long promised the ability to rapidly create easily maintainable software for complex applications. The International Symposium of Practical Aspects of Declarative Languages (PADL) provides a yearly - rum for presenting results on the principles the implementations and especially the applications of declarative languages. The PADL symposium held January 19-20, 2009 in Savannah, Georgia was the 11th in this series. This year 48 papers were submitted from authors in 17 countries. The P- gram Committee performed outstandingly to ensure that each of these papers submitted to PADL 2009 was thoroughly reviewed by at least three referees in a short period of time. The resulting symposium presented a microcosm of how the current generation of declarative languages are being used to address real applications, along with on-going work on the languages themselves. The program also included two invited talks, "e;Inspecting and Preferring Abductive Models"e; by Luis Moniz Pereira and "e;Applying Declarative Languages to C- mercial Hardware Design"e; by Je? Lewis. Regular papers presented a variety of applications, including distributed applications over networks, network veri?- tion, user interfaces, visualization in astrophysics, nucleotide sequence analysis and planning under incomplete information. PADL 2009 also included ongoing work on the declarative languages themselves. Multi-threaded and concurrent Prolog implementation was addressed in several papers, as were innovations for tabling in Prolog and functional arraysin Haskell. Recent applications have also sparked papers on meta-predicates in Prolog and a module system for ACL2.
The growing complexity of modern software systems increases the di?culty of ensuring the overall dependability of software-intensive systems. Complexity of environments, in which systems operate, high dependability requirements that systems have to meet, as well as the complexity of infrastructures on which they rely make system design a true engineering challenge. Mastering system complexity requires design techniques that support clear thinking and rigorous validation and veri?cation. Formal design methods help to achieve this. Coping with complexity also requires architectures that are t- erant of faults and of unpredictable changes in environment. This issue can be addressed by fault-tolerant design techniques. Therefore, there is a clear need of methods enabling rigorous modelling and development of complex fault-tolerant systems. This bookaddressessuchacuteissues indevelopingfault-tolerantsystemsas: - Veri?cation and re?nement of fault-tolerant systems - Integrated approaches to developing fault-tolerant systems - Formal foundations for error detection, error recovery, exception and fault handling - Abstractions, styles and patterns for rigorousdevelopment of fault tolerance - Fault-tolerant software architectures - Development and application of tools supporting rigorous design of depe- able systems - Integrated platforms for developing dependable systems - Rigorous approaches to speci?cation and design of fault tolerance in novel computing systems TheeditorsofthisbookwereinvolvedintheEU(FP-6)projectRODIN(R- orous Open Development Environment for Complex Systems), which brought together researchers from the fault tolerance and formal methods communi- 1 ties. In 2007 RODIN organized the MeMoT workshop held in conjunction with the Integrated Formal Methods 2007 Conference at Oxford University.
Models have become essential for supporting the development, analysis and e- lution of large-scale and complex IT systems. Models allow di?erent views, p- spectives and elements of a system to be captured rigorously and precisely, thus allowing automated tools to manipulate and manage the models. In a full-?edged model-driven engineering (MDE) process, the transformations developed and - pressed between models are also key. Model transformations allow the de?nition and implementation of the operations on models, and also provide a chain that enables the automated development of a system from its corresponding m- els. Model transformations are already an integral part of any model-driven approach, and there are a number of available model transformation languages, tools, and supporting environments; some of these approaches are now approa- ing maturity. Nevertheless, much work remains: the research community and industry need to better understand the foundations and implications of model transformations, such as the key concepts and operators supporting transfor- tion languages, their semantics, and their structuring mechanisms and properties (e. g. , modularity, composability and parametrization). The e?ect of using model transformations on organizations and development processes - particularly when applied to ultra-large scale systems, or in distributed enterprises - is still not clear. These issues, and others related to the speci?cation, design, implemen- tion, analysis and experimentation with model transformation, are the focus of these proceedings. The Second International Conference on Model Transformation (ICMT 2009) was held in late June 2009 in Zurich, Switzerland.
Formal methods for development of computer systems have been extensively studied over the years. A range of semantic theories, speci?cation languages, design techniques, and veri?cation methods and tools have been developed and applied to the construction of programs used in critical applications. The ch- lenge now is to scale up formal methods and integrate them into engineering - velopment processes for the correct and e?cient construction and maintenance of computer systems in general. This requires us to improve the state of the art on approaches and techniques for integration of formal methods into industrial engineering practice, including new and emerging practice. The now long-established series of International Conferences on Formal - gineering Methods brings together those interested in the application of formal engineering methods to computer systems. Researchers and practitioners, from industry, academia, and government, are encouraged to attend and to help - vance the state of the art. This volume contains the papers presented at ICFEM 2009, the 11th International Conference on Formal Engineering Methods, held during December 9-11, in Rio de Janeiro, Brazil.
This edition marks the tenth Middleware conference. The ?rst conference was held in the Lake District of England in 1998, and its genesis re?ected a growing realization that middleware systems were a unique breed of distributed system requiring their own rigorous research and evaluation. Distributed systems had been around for decades, and the Middleware conference itself resulted from the combination of three previous conferences. But the attempt to build common platforms for many di?erent applications requireda unique combinationofhi- level abstraction and low-level optimization, and presented challenges di?erent from building a monolithic distributed system. Since that ?rst conference, the notion of what constitutes "e;middleware"e; has changed somewhat, and the focus of research papers has changed with it. The ?rst edition focused heavily on distributed objects as a metaphor for building systems, including six papers with "e;CORBA"e; or "e;ORB"e; in the title. In f- lowing years, the conference broadened to cover publish/subscribe messaging, peer-to-peer systems, distributed databases, Web services, and automated m- agement, among other topics. Innovative techniques and architectures surfaced in workshops, and expanded to become themes of the main conference, while changes in the industry and advances in other research areas helped to shape research agendas. This tenth edition includes papers on next-generation pl- forms (such as stream systems, pervasive systems and cloud systems), managing enterprise data centers, and platforms for building other platforms, among o- ers.
This book constitutes the proceedings of the Second Technology Conference on Performance Evaluation and Benchmarking, TPCTC 2010, held in conjunction with the 36th International Conference on Very Large Data Bases, VLDB 2010, in Singapore, September 13-17, 2010. The 14 full papers and two keynote papers were carefully selected and reviewed from numerous submissions. This book considers issues such as appliance; business intelligence; cloud computing; complex event processing; database optimizations; data compression; energy and space efficiency, green computing; hardware innovations; high speed data generation; hybrid workloads; very large memory systems; and virtualization.
Tilmeld dig nyhedsbrevet og få gode tilbud og inspiration til din næste læsning.
Ved tilmelding accepterer du vores persondatapolitik.