Software developers and service operators need log files to identify issues, detect anomalies and trace execution behaviour. The amount of generated log data is increasing, and often log files need to be kept for longer periods of time due to regulations. To preserve logs in a cost-efficient manner, they are typically compressed, at certain cost for running the compression, and then stored in long-term archives, again at certain cost per size-duration products. The goal is decrease both cost components, but there are certain trade-offs, for instance a highly efficient compression that consumes a lot of CPU but leads to better compression ratios, consuming less storage capacity as a result. The decision which compression tools and parameters to use is usually hardcoded. We present a smart knowledge-based advisor service to query goal-based adaptive compression commands to maximise savings.
ESOCC 2020 – the 8th European Conference on Service-Oriented and Cloud Computing – was originally scheduled to happen in April on the wonderful isle of Crete. Due to the pandemic circumstances, it will now run from September 28 to 30 as online event. We will highlight some of our recent work in an invited talk on «The Great Puzzle: Cloud Functions, Data, Services, Less Servers and More Insights».
Check out the conference website for more information on the technical programme, workshops and European R&D project presentations, and for pointers on how to participate.
Automation is one of the key concerns in cloud environments. The need to introduce effort-saving automation around the process of bringing new applications to powerful cloud environments ranges from developer tooling over testing and deployment to operational concerns. According to Nokia’s Eric Bauer, application service efficiency is the ratio of service output produced to resource input consumed, and automation can significantly reduce the input effort.
For several years, we have conducted research on the design, implementation and evaluation of microservice-based applications, as well as on the assessment of characteristics of the constituent software artefacts. Yet we were so far not present in the first two editions of the International Conference on Microservices. Needless to say, we are now correcting this for the third edition of the conference with a talk on Syn.
In the context of our «Smart Cities and Regions Services Enablement» efforts, space (and to some extent time) are important dimensions. First, the digital transformation has an inherent spatial component. While the research application field is pragmatically scoped to cities and regions, indeed it spans a wider spectrum from households, quarters, districts to countries and even supranational entities. The recent wave of «surface digitalisation» has primarily affected mobile citizens (pandemic apps) and workers (video conferencing in home offices) around the world. This increased the surface over the previous one that for most citizens encompassed e-banking, e-ticketing and e-tax declarations, with various degrees of voluntariness.
For the past five years, Zurich University of Applied Sciences has hosted the Service Prototyping Lab to investigate new ways to design, prototype, implement and deploy SaaS and related cloud application concepts. We have worked with many companies from all over Switzerland to come up with innovative solutions together. We still continue this way, but we also want to reflect technological change and the evolving requirements of our research and innovation partners as well as our students in education. Therefore, we are working on reflecting this evolution also in naming, and gradually move towards a positioning as leading research partner in Switzerland around the topic of Distributed Application Computing Paradigms.
As we have recently been granted Google Cloud Research Credits for the investigation of Serverless Data Integration, we continue our exploration of open and public data. This HOWTO-style blog post presents the application domain of financial analytics and explains how to run a cloud function to achieve elastically scalable analytics. Although there are no research results to report yet, it raises a couple of interesting challenges that we or other computer scientists should work on in the future.
Cloud applications are typically designed as coupled microservices and deployed in managed containerised form. Industry trends around container build processes, deployment packages, management platforms and abstractions (e.g. cloud functions) are still fast-paced. Developers and operators need to be able to tell good from bad practices based on automatically determined metrics. Assuming they participate in this tutorial, they will learn how to do that on a hands-on level. We introduce approaches and open source tools for quantitative assessment of containers and other microservice technologies and ecosystems. On the research side, we explain how this blends with policy-driven deployments, trusted cloud execution and data science opportunities.
The three-hours tutorial will be offered at the CLOSER 2020 conference (originally scheduled in Prague, now online) in the afternoon on May 7. Registration information is available from the conference website.
Self-management is an important property of software services to increase the degree of exploiting benefitial characteristics of underlying runtime systems. Whether such services run in a managed cloud environment, on a device or somewhere else in the computing continuum, there may always be limitations in the managing runtime platform that a complementary or overarching application-level management can help to overcome. Using a Python Flask-based web service as example, this research blog post informs about our ongoing investigations into two specific self-management aspects: runtime resilience and feistiness.