Design Science Benchmarking – A Structured Approach to Make Supervised Machine Learning Algorithms Comparable

Julian Armin Dormehl

Abstract

Current developments, including high availability of data and ever rising computing power, constantly enable new approaches in the field of artificial intelligence. By using algorithms from machine learning, an instance can iteratively learn from data and perform cognitive tasks. As a large amount of machine learning algorithms already exists today and is still increasing rapidly, researchers, data scientists, and machine learning engineers must choose which algorithm to apply and optimize to solve their individual problem. In most cases, the selection of a specific algorithm seems to be highly prediction performance motivated as well as dependent on situational tendencies of the user. For use in non-productive environments, these intentions would initially be sufficient. However, for operational applications these tendencies do not provide satisfactory solutions, as they represent a one-sided perspective and an unstructured approach. Thus, when developing productive applications, no methodological comparison is made between machine learning algorithms, that considers factors of the data basis, the operational view and the explainability of a machine learning model, in addition to the commonly used metrics for evaluating predictive per-formance. To close this gap we develop two artifacts, first a structured benchmarking procedure for the comparison of supervised machine learning algorithms, a particular paradigm of machine learning we focus on, second a list of criteria, implemented in the benchmarking model, for iden-tifying the most appropriate supervised machine learning algorithm from a holistic perspective following a design science research approach. To provide robust, practical, and user-friendly artifacts, we validate our results in a four-step approach, by, for example, conducting a discussion with research experts, which is prospectively completed by a real-world application of the model. Our results contribute to a structured, generic procedure that supports the benchmarking of supervised machine learning algorithms and provides users with benchmarking-relevant dimensions to identify the most appropriate supervised machine learning algorithm for their individual use-case.

Topics
Supervised Machine Learning Algorithm Selection Multi-criteria Benchmarking Structured Benchmarking Approach
Research Methods
Design Science Research

Publication Data

Author: Julian Armin Dormehl
Thesis Type: Bachelor's Thesis
Pages: 46
Language: English
DOI:
About the Author:
Major / Study Program: Industrial Engineering
Primary Field of Study: Business & Information Systems Engineering
Additional Study Interests: Machine Learning,Energy transition in times of digitalization,Digital business models
License: CC BY-NC-SA 4.0
Date of Publication: 10/25/22
Status: Available
Date of Grading: 03/15/22
Institution: University of Augsburg (University of Augsburg, Germany)

Endorsements

# Name Details Endorsement
1
Dr. Simon Wenninger
Supervisor
Researcher
10/20/22
01:00:00 AM

Thesis Documents and Supplemental Materials

11/30/22 08:16:48 AM
# Description Type Upload Date Location
1 Thesis Document PDF (4.91MB) 10/19/22 01:00:00 AMIPFS Download Raw