Winter Term 2016/17

In our colloquium at the Department of Computer Science, national and international guests as well as department members present their research. Our guest speakers present impacting topics across various areas of the discipline. The colloquium series is held every semester and also includes inaugural and farewell lectures of the department's professors.

The colloquium is a noteworthy event for all graduate students as the talks are seen as a key part of their education at HSZG. The talks are also open to the public and outside attendance is welcome.

For news about the colloquium, the undefinedEAD-Lunch talks and seminars please feel free to subscribe to (Register here:!forum/ead-public)


19.10.2016 Prof. Frank Häßler "Notstromversorgung? Aber sicher! - Die Umsetzung höchster Anforderungen der funktionalen Sicherheit in der Softwareentwicklung am Beispiel eines kerntechnisch zertifizierten Notstromdiesels."

Die Versorgungssicherheit in Kraftwerken, auch bei einem Ausfall der externen Stromversorgung, wird unter anderem durch Notstromaggregate gewährleistet. Ein fortlaufender Betrieb des Kraftwerks oder ein kontrollierter Abschaltvorgang wird hierdurch ermöglicht. Sichere und hochverfügbare Regler sowie Steuergeräte, entwickelt nach den weltweit höchsten Standards im Bereich der funktionalen Sicherheit, werden bei Dieselmotoren in diesem Bereich verwendet.

Dargestellt wird praktische Anwendung der Normen IEC 60880 sowie IEC 61508 (SIL3) für die Softwareentwicklung. Nach einer kurzen Einführung in die Grundlagen der Entwicklungsstandard werden die einzelnen Teilbereiche des Softwarentwicklungsprozesses beleuchtet. Hierbei geht es sowohl um die technischen und nichttechnischen Aspekte. Im technischen Part liegt die Schwerpunktsetzung auf Methodik und Werkzeugen, im nichttechnischen Teil stehen die Anforderungen an den Prozess im Vordergrund.


09.09.2016 Prof. Thomas Weise " An Introduction"


Optimization algorithms have become a standard tool in many application areas such as management, logistics, engineering, design, chemistry, and medicine. They provide close-to-optimal approximate solutions for computationally hard problems within feasible time. This field has grown and evolved for the past 50 years and has several top-level journals dedicated to it. Research in optimization is focused on reducing the algorithm runtime and increasing the result quality. For such research to succeed and publications to have true impact on the real world, we need to be able to


  • analyze the performance of an algorithm, to
  • analyze the influence of different features of an optimization problem on its hardness, and to
  • compare the performance of different algorithms in a fair and sound fashion.

Many optimization methods are anytime algorithms, meaning that they start with a (usually bad) guess about the solution and step-by-step improve their approximation quality. All evolutionary algorithms, all local search algorithms (such as Simulated Annealing and Tabu Search), all swarm intelligence methods for optimization (such as ant colony and particle swarm optimization), CMA-ES and memetic algorithms, but also several exact and deterministic methods such as branch and bound belong into this class, just to name a few.


The comparison and evaluation of anytime algorithms must consider the whole runtime behavior of the algorithms in order to avoid misleading conclusions. Also, performance data has to be gathered from multiple independent runs on multiple different benchmark instances. It is easy to see that a thorough analysis and comparison of optimization algorithms is complicated and cumbersome. We present an open source software which can do this for you. You gather the data from your experiments, the software analyzes it. Our goal is to support researchers and practitioners as much as possible by automating the evaluation of experimental results. The software does not require any programming, just your benchmarking data. It imposes no limits, neither on the type of algorithms to be compared nor on the type of problem they are benchmarked on. Our software produces human-readable conclusions and reports in either XHTML or LaTeX format. You can freely select and configure different diagram types and group your data according to different aspects to get a better understanding of the behavior of your algorithm. Figures are styled for direct re-use in journals such as IEEE Transactions or conference proceedings such as LNCS. The software is dockerized, meaning that you can directly apply it with minimal installation effort.


We demonstrate the utility of this software on the example of the investigation of six primitive heuristics on the Maximum Satisfiability Problem (MAX-SAT). Similar examples are provided for download for numerical optimization and the Traveling Salesman Problem.


More information can be found at




1. Thomas Weise, Raymond Chiong, Ke Tang, Jörg Lässig, Shigeyoshi Tsutsui, Wenxiang Chen, Zbigniew Michalewicz, and Xin Yao. Benchmarking Optimization Algorithms: An Open Source Framework for the Traveling Salesman Problem. IEEE Computational Intelligence Magazine (CIM), 9(3):40-52, August 2014. Featured article and selected paper at the website of the IEEE Computational Intelligence Society ( doi:10.1109/MCI.2014.2326101


2. Thomas Weise, Yuezhong Wu, Raymond Chiong, Ke Tang, and Jörg Lässig. Global versus Local Search: The Impact of Population Sizes on Evolutionary Algorithm Performance. Journal of Global Optimization. accepted 12 February, 2016, published first online: 23 February, 2016. doi:10.1007/s10898-016-0417-5

Letzte Änderung:10. November 2016

Direktlinks & Suche