aforementioned lack in mind, we assume that the
considered company will conduct comparatively
basic experiments on the MaaS platform.
Consequently, expert features of the MaaS solution
like 2.19, 6.2 and 13.1 are not of particular
importance and will be assigned minor weights.
Limited resources are another impact factor for
the evaluation. We suppose that SMEs neither have
sufficient reserve assets, nor enough staff for distinct
advance development. Instead, SMEs need decent
return on investment in a comparatively short time
period. Hence, the MaaS solution should either be
low in price or generate quick net product. Criteria
fulfilling these requirements will be assigned major
weights (e.g. 3.2, 3.3).
Lastly, like every other company, SMEs must deal
with showstoppers. Such criteria will be assigned the
highest possible weight. Apart from universal
showstoppers, like non-compliance with locally
applicable law, we identified one noteworthy MaaS-
specific criterion: reliability or the correctness of the
system’s output (15.1). In machine learning,
estimating model performance for unseen data is a
complex task. Small or biased datasets can
complicate this task even more. Combined with non-
expert users, this bears the pitfall of overestimating
model performance which, in turn, can have
economic consequences. Next to mere model
performance in terms of known evaluation metrics,
there are additional risks such as estimators learning
side issues instead of focusing on relevant aspects of
the data. In image classification, for example, there
are known cases in which this leads to unexpected
behavior of image classifiers (Han S. Lee, 2017).
Therefore, in our scenario it is utterly important for
MaaS solutions to feature robust model performance
estimates such as cross validation on the one hand and
provide model insight (possibly with methods of the
field of explainable artificial intelligence) on the other
hand. Thereby, the risk of misuse by non-expert users
can be reduced.
The three aforementioned assumptions, together
with additional considerations (see a sample in Table
2), lead to the weights column. By normalizing, as
described in Section 2.3, we obtain the final list
4
exhibiting the most import criteria which is adjusted
to the domain MaaS as well as to the target audience
SMEs.
4 CONCLUSIONS
We presented a transferable methodology for
deriving a criteria catalogue for software solutions. It
can be directly applied as is or used as an inspiration
for problems alike. As such, different software
solutions (of the same scope) on the market can
objectively be compared, so that the optimal solution
can be found for a business.
To allow for this, we proposed that two
independent dimensions – domain knowledge and
industry context – need to be encoded into a template
criteria catalogue (which we also compiled). The first
dimension ensures that the software solution indeed
solves the technical problem one faces, the second
dimension attests, that it is in line with business
strategy and branch context. Followingly, one could
consider domain knowledge, as a bottom-up process,
reflecting the skill of specialists, while industry
context mirrors a top-down process, reflecting the
market understanding of decision makers.
The method is formalized in a three-layer model
with two-layer connections in between. Because the
first layer is a general list of criteria, blended from
many sources, it should offer a software-agnostic
basis for tackling decision problems. The transition to
the second layer encodes domain knowledge and the
connection to the third layer encodes industry
context. Connections were presented as a step
sequence, accompanied by examples, to additionally
illustrate the approach.
The resulting catalogue in layer three can then be
used for the comparison of software solutions: Every
software solution under scrutiny, is assessed in every
criterion, yielding a final matching score.
The theoretically derived model – here presented
in an abstract scenario – of course needs further real-
world validation. Thus, additional examination is
strongly suggested, and we plan to investigate this in
an empirical study.
REFERENCES
Brand, K. (2017, April 27). www.xing.com. Retrieved from
https://www.xing.com/news/insiders/articles/wesentlic
he-kriterien-fur-die-auswahl-von-it-losungen-705949
Ekanayaka, Y., Currie, W., & Seltsikas, P. (2003).
Evaluating application service providers.
Benchmarking.
Evans, J. (1989). Bias in human reasoning: Causes and
consequences. Lawrence Erlbaum Associates, Inc.
Godse, M., & Mulik, S. (2009). An Approach for Selecting
Software-as-a-Service (SaaS) Product. 2009 IEE
International Conference on Cloud Computing.
Han S. Lee, A. A. (2017, September). Why Do Deep Neural
Networks Still Not Recognize These Images?: A
Qualitative Analysis on Failure Cases of ImageNet