is addressed more frequently than others: perfor-
mance, addressed in ∼ 80% of the primary stud-
ies. More effort should be directed to other qual-
ity attributes since desired levels of QoS are usu-
ally achieved through a combination of quality at-
tributes (Software & Systems Engineering Standards
Committee, 1998). Furthermore, different primary
studies report on different factors that affect quality
attributes and, consequently, the overall QoS. Virtual-
ization is covered in more than ∼ 50% of the primary
studies, and it is mostly mentioned as a factor affect-
ing performance. This result was expected as virtu-
alization is a ground technology that enables the use
of cloud computing and improves the utility of phys-
ical machines. Furthermore, a notable number of pri-
mary studies investigate data storage architecture as
a factor affecting QoS and more specifically, security.
This was expected as numerous enterprises are mov-
ing their data to the cloud because of its simplicity,
but fear the security aspect. We believe additional re-
search must be conducted in how virtualization solu-
tions affect security because each virtualization layer
can be a target to malicious attacks and compromise
the entire cloud infrastructure. Moreover, our study
revealed that only ∼ 24% of the primary studies re-
ported on QoS with respect to a particular domain,
therefore, it is our belief that more research efforts
should be spent on investigating QoS in specific in-
dustrial domains.
Regarding quality metrics (RQ3), the majority
of primary studies focuses on three quality metrics:
response time, resource utilization and make span.
These results do not imply that the remaining quality
metrics have not received enough attention from the
research community, but only that they have a more
restricted application in the assessment of quality at-
tributes mentioned in the selected primary studies.
The aforementioned quality metrics are, to a large ex-
tent, used to assess performance. However, ∼ 18% of
quality metrics have two occurrences in the primary
studies, while ∼ 43% have only one. Furthermore,
only one primary study contributed by proposing met-
rics for the assessment of specific quality attributes.
This makes us believe that more research should be
conducted on the definition of more specific and en-
hanced quality metrics. Moreover, the mapping be-
tween quality attributes and quality metrics showed
that ∼ 67% of the quality attributes mentioned in the
primary studies were not assessed by any quality met-
ric. We foresee a need for further research targeting
these quality attributes with the goal of proposing and
identifying quality metrics to assess them.
Regarding objectives (RQ4) of the primary stud-
ies, we found that the majority were interested in
improving QoS. We could discuss how the research
community could have focused more on other ob-
jectives such as measuring or defining QoS, rather
than on its improvement, but being that QoS is cru-
cial for the adoption of cloud technologies, we un-
derstand the rationale behind this. Moreover, the ap-
proaches that strive to improve QoS involve the wide
use of scheduling techniques and mechanisms. How-
ever, our suggestion would be to focus on providing
scheduling solutions that take into consideration mul-
tiple QoS attributes rather than a single one. Further-
more, we observed a noteworthy focus on using secu-
rity mechanisms to improve QoS.
7 THREATS TO VALIDITY
External Validity. One major threat that could limit
the generalizability of our study is having a set of
primary studies that is not representative enough of
the research on this topic. To mitigate this threat, we
conducted an automatic search on two of the largest
and most complete electronic databases in software
engineering, which was complemented with a fully-
recursive backward snowballing activity to eventually
enrich the initial corpus of primary studies. Another
threat that can jeopardize the external validity of our
study is the exclusion of primary studies in languages
other than English. Regardless, being that English is
the de-facto standard language used for scientific pa-
pers, this threat is negligible. Moreover, we only tar-
geted peer-reviewed publications as they are expected
to provide scientific work of certain quality assessed
by peers, thus excluding gray literature.
Internal Validity. To mitigate biases regarding the
degree of influence of external variables on the de-
sign of the study, we rigorously defined and validated
a detailed study protocol that follows the guidelines
proposed by Petersen et al. (2015), Kitchenham and
Charters (2007), and Wohlin et al. (2012). Further-
more, we defined a classification framework for the
extraction of data from the primary studies, that was
iteratively revised. The keywording process was also
used to transform qualitative data into quantitative
data, while quantitative data was analyzed using de-
scriptive statistics, making the data analysis validity
threats minimal. Finally, a complete replication pack-
age
1
is available for the independent replication of our
study.
Construct Validity. To mitigate the threat of hav-
ing a set of primary studies that is not representa-
tive enough of the population defined by the research
questions, we conducted an automatic search on two
electronic databases and complemented the search
CLOSER 2021 - 11th International Conference on Cloud Computing and Services Science
158