SlideShare a Scribd company logo
1 of 13
Download to read offline
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
DOI : 10.5121/ijsea.2013.4504 49
STABLENESS MEASUREMENT MODEL: A
MATHEMATICAL APPROACH FOR MEASURING THE
STABILITY OF HANDHELD APPLICATION USAGE
Amalina Farhi Ahmad Fadzlah1
1
Department of Computer Science, Faculty of Defence Science and Technology,
Universiti Pertahanan Nasional Malaysia, 57000 Kuala Lumpur, Malaysia
amalina.farhi@upnm.edu.my
ABSTRACT
This study is designed to develop a mathematical model for measuring the stability of handheld application
usage namely Stableness Measurement Model (SMM). This model outlined a series of formulas based on
the total number of eleven stability measures (i.e. eight stability metrics and three stability attributes)
which are identified as having associated and contributed towards measuring the stability of handheld
application usage. This model is valuable as an alternative evaluation technique to be used for measuring
and ensuring the stability of handheld application usage.
KEYWORDS
Stability, model, measure, handheld, application
1. INTRODUCTION
Stability is one of the most fundamental and important of all usable and useful software
characteristics. The term stability means making the condition of software of being resistant to
change of position or condition with which not easily moved or disturbed [1]. Other definition
described stability as quality or attribute of being firm and steadfast to software that bare on the
provision of right or agreed results or effects with continuous function well in an acceptable
period [2]. In this paper, the term stability in the perspective of useful and usable can be defined
as the degree to which making the condition of software of being stable or steady in relation to
correct or complete as well as effort and time, that reflects the real world object or event being
described, based on the users’ needs and requirements. The fewer failures and times taken to
complete tasks that are observed the more stable an application is.
Stability normally plays an important factor for all software quality elements. Over the past few
decades, several researches for assessing and evaluating stability of software have been
mentioned. The international standard, ISO/IEC 9126 [2], described stability as quality sub
attribute to software that bare on the provision of the ease with which a product can be
maintained in order to improve reliability. In the other hand, stability correlates with the metrics
which measure attributes of the software that indicate about the risk of unexpected effects as a
result of modification [3][4][5]. Some researchers also classify stability as an essential
characteristic for evaluating and assessing the usability of software [6][7][8][9]. Within the
domain of handheld software, several researchers have proposed to explore the concept of
stability [10][11][12][13].
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
50
2. MATERIALS AND METHODS
This study outlined three main questions: 1) what stability measures are really important; 2) what
is the rank of each stability measure; and 3) what is the weight of each stability measure, towards
measuring the stability of handheld application usage. In order to answer these questions, a
questionnaire survey, namely Investigating Stability Measures for Handheld Application Usage
was conducted among handheld device users and a total number of two hundred nineteen
respondents responded.
These stability measures were classified into three hierarchical levels of metrics, attributes and
criterions. Metrics are described as the lowest hierarchy level. The main objective of the metrics
is to identify measurable data for the purpose of measuring the stability of handheld application
usages. The middle hierarchy level is described as attributes, whereas the highest is described as
criterion (i.e. stability of handheld application usage). This hierarchy which brings together three
different stability levels of metrics, attributes and criterion is as detailed below (Table 1).
Table 1. Stability hierarchy level
Hierarchy Description
Metric The lowest hierarchy level; A collection of measurable stability data
expressed in units
Attribute The middle hierarchy level; A collection of metrics which belongs to a
class of stability measures
Criterion The highest hierarchy level; A collection of attributes for measuring the
stability of handheld application usage
A total number of eleven stability measures, with a number of eight stability metrics and three
stability attributes, were outlined as having associated towards measuring the stability of
handheld application usage. The definition of each stability measure is as depicted below (Table
2).
Table 2. Stability measures and descriptions
Measure Description
Information Speed*
Capability in handling information per time
Lateral Position*
Capability in positioning objects per time
Optimal Solution*
Capability in solving tasks per time
Data Entered**
The number of data entered per time
Errors Corrected**
The number of errors corrected per time
Focuses Distracted**
The number of focuses distracted per time
Lines Read**
The number of lines read per time
Links Explored**
The number of links explored per time
Paths Traversed**
The number of paths traversed per time
Steps Navigated**
The number of steps navigated per time
Targets Located**
The number of targets located per time
Legend of the table:
*
attribute determines the criterion (i.e. stability of handheld application usage)
**
metric determines the attribute
Data collected from the questionnaire is entered in the Statistical Package for the Social Sciences
(SPSS) for the analysis process as well as to classify the stability measures into the hierarchical
structure of metrics, attributes and criterion. This brings together two parts of evaluation tests:
Pearson’s Chi-square test and the Spearman’s Rho test. Pearson’s Chi-Square test was conducted
to measure the amount of association between two different stability measures in two different
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
51
hierarchy levels and the Spearman’s Rho test was conducted to comprehend the relationship
strength between two different stability measures in two different hierarchy levels.
The model for measuring the stability of handheld application usage is specifically developed
using a conceptual framework, namely Stableness Measurement Framework (SMF) (Figure 1).
This framework brings together different stability measures in different hierarchy levels. As
illustrated below, the metric determines the attribute and the attribute determines the criterion (i.e.
stability as criterion). Each level represents interaction with other level and the impact to one
another to measure the stability of the desired handheld application usage. This can be explained
as either none, one or more metrics to represent a single attribute.
The combination of these metrics could be represented as the components that contributed to only
one attribute. And finally, these attributes are used to support in the calculation of the criterion
that can be concluded as directly affected the stability of handheld application usage. This is the
case at every level in which could be represented as an M-1 relationship. For example, metric M1
… Mn are the input to attribute AA and criterion CC is an output for the attribute AA. Consider if
the value of metric M1, M2, … , Mn-1 or Mn increases so as the value of attribute AA and criterion
CC. Again, if the value of metric M1, M2, … , Mn-1 or Mn decreases so as the value of attribute AA
and criterion CC.
3. RESULTS AND DISCUSSIONS
3.1. Results of Association Test
Result of association test reported that metrics of Data Entered (M = 4.52, SD = .738), Errors
Corrected (M = 4.10, SD = .979) and Focuses Distracted (M = 3.65, SD = 1.027) were
contributed towards measuring the stability of handheld application usage with p < .001. Results
also showed that metrics of Links Explored (M = 4.25, SD = .780), Lines Read (M = 4.08, SD =
.992) and Paths Traversed (M = 3.95, SD = .912) were contributed towards attribute Optimal
Solution with p < .001. Meanwhile, metrics of Steps Navigated (M = 4.04, SD = .905) and
Targets Located (M = 3.69, SD = 1.038)were also found contributed towards measuring the
stability of handheld application usage with p < .001.
Figure 1. Stableness Measurement Framework (SMF)
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
52
Finally, result of the association test also stated that the attributes of Information Speed (M =
4.34, SD = .811), Lateral Position (M = 4.27, SD = .734) and Optimal Solution (M = 4.27, SD =
.806) were found contributed towards measuring the stability of handheld application usage, with
p < .001. As a result, a total number of eleven stability measures (i.e. eight stability metrics and
three stability attributes) were identified having associated and contributed towards measuring the
stability of handheld application usage (Table 3).
Table 3. Result of association test
Stability Measures Mean
Stability Attributes
Information Speed 4.34
Lateral Position 4.27
Optimal Solution 4.27
Stability Metrics
Data Entered 4.52
Errors Corrected 4.10
Focuses Distracted 3.65
Lines Read Speed 4.08
Links Explored 4.25
Paths Traversed 3.95
Steps Navigated 4.04
Targets Located 3.69
Results from the association test were further ranked to prioritize the level of importance of each
stability measure towards measuring the overall stability of handheld application usage (Table 4).
Table 4. Rank of stability measures
Stability Measures Rank
Stability Attributes
Information Speed 1
Lateral Position 2
Optimal Solution 3
Stability Metrics
Data Entered 1
Links Explored 2
Errors Corrected 3
Lines Read 4
Steps Navigated 5
Paths Traversed 6
Targets Located 7
Focuses Distracted 8
3.2. Results of Relationship Test
Result of the relationship test revealed that there was a moderate and positive linear relationship
between metrics Data Entered (R = .346), Errors Corrected (R = .251) and Lines Read (R = .298)
towards attribute Information Speed with p < .001. Results also found that the coefficient value
of metrics Targets Located (R = .528) and Focuses Distracted (R = .470) were moderate and
positive linear relationship towards attribute Lateral Position with p < .001. Metrics Links
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
53
Explored (R = .333), Steps Navigated (R = .385) and Paths Traversed (R = .410) were also
reported to have a moderate and positive linear relationship between attribute Optimal Solution
with p < .001.
Finally, the relationship test also indicated the correlation strength between attributes Information
Speed (R = .306), Lateral Position (R = .311) and Optimal Solution (R = .298) resulted having a
moderate and positive linear relationship towards measuring the stability of handheld application
usage with p < .001. Based on the result of the relationship test, out of the total number of eleven
stability measures, seven (i.e. five stability metrics and two stability attributes) were identified
having moderate and positive linear relationship, three (i.e. two stability metrics and one stability
attributes) were identified having low and positive linear relationship, while only one stability
matric reported having high and positive linear relationship towards measuring the stability of
handheld application usage (Table 5).
Table 5. Result of relationship test
Stability Measures Relationship
Attributes contributed towards criterion (attribute  criterion)
Information Speed  Stability Moderate, Positive
Lateral Position  Stability Moderate, Positive
Optimal Solution  Stability Low, Positive
Metrics contributed towards attribute (metric  attribute)
Data Entered  Information Speed Moderate, Positive
Links Explored  Optimal Solution Moderate, Positive
Errors Corrected  Information Speed Low, Positive
Lines Read  Information Speed Low, Positive
Steps Navigated  Optimal Solution Moderate, Positive
Paths Traversed  Optimal Solution Moderate, Positive
Targets Located  Lateral Position High, Positive
Focuses Distracted  Lateral Position Moderate, Positive
Legend of the table: Correlation is significant at the 0.001 level (2-tailed) and range in the value
of +1 to -1
Results from the relationship test were further analysed to obtain the value of weightage of each
stability measure towards measuring the overall stability of handheld application usage (Table 6).
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
54
Table 6. Weight of stability measures
Stability Measures Weight
Attributes contributed towards criterion (attribute  criterion)
Information Speed  Stability .306
Lateral Position  Stability .311
Optimal Solution  Stability .298
Metrics contributed towards attribute (metric  attribute)
Data Entered  Information Speed .346
Errors Corrected  Information Speed .251
Lines Read  Information Speed .298
Targets Located  Lateral Position .528
Focuses Distracted  Lateral Position .470
Links Explored  Optimal Solution .333
Steps Navigated  Optimal Solution .385
Paths Traversed  Optimal Solution .410
The findings derived from the analysis of both association test and relationship test produced lists
of codes to represent each stability metric and attribute, presented as Mm•Aa•CSTB and Aa•CSTB
respectively (Table 7). Symbolized as Mm, M represents the stability metrics while m represents
the sequential series (m-th) of the stability metric such as 1, 2, …, m. Similarly, symbolized as
Aa, A represents the stability attribute while a represents the sequential series (a-th) of the stability
attribute such as 1, 2, …, a. Finally, symbolized as CSTB, C represents the stability criterion in
which STB represents the abbreviation of the stability.
Table 7. Code of stability measures
Stability Measures Code
Attributes contributed towards criterion (attribute  criterion) Aa• CSTB
Information Speed  Stability A1•CSTB
Lateral Position  Stability A2• CSTB
Optimal Solution  Stability A3•CSTB
Metrics contributed towards attributes (metrics  attributes) Mm•Aa•CSTB
Data Entered  Information Speed M1•A1•CSTB
Errors Corrected  Information Speed M2•A1•CSTB
Lines Read  Information Speed M3•A1•CSTB
Targets Located  Lateral Position M1•A2•CSTB
Focuses Distracted  Lateral Position M2•A2•CSTB
Links Explored  Optimal Solution M1•A3•CSTB
Steps Navigated  Optimal Solution M2•A3•CSTB
Paths Traversed  Optimal Solution M3•A3•CSTB
Furthermore, findings derived from the analysis of both association test and relationship test also
produced lists of codes to represent the weight of each stability metric and attribute, presented as
ωATTm and ωCRTa respectively (Table 8). ω represents the symbol of weights, meanwhile
symbolized as ATTm, m represents the sequential series (m-th) of the stability metric such as 1, 2,
…, m that contributed towards particular attribute, ATT, and finally symbolized as CRTa, a represents
the sequential series (a-th) of the stability attribute such as 1, 2, …, a that contributed towards
measuring the stability of handheld application usage, in which CRT represents the abbreviation of
the stability, coded as STB.
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
55
Table 8. Weight code of stability measures
Stability Measures Code
Attributes contributed towards criterion (attribute  criterion) ωCRTa
Information Speed  Stability ωSTB1
Lateral Position  Stability ωSTB2
Optimal Solution  Stability ωSTB3
Metrics contributed towards attributes (metrics  attributes) ωATTm
Data Entered  Information Speed ωIS1
Errors Corrected  Information Speed ωIS2
Lines Read  Information Speed ωIS3
Targets Located  Lateral Position ωLP1
Focuses Distracted  Lateral Position ωLP2
Links Explored  Optimal Solution ωOS1
Steps Navigated  Optimal Solution ωOS2
Paths Traversed  Optimal Solution ωOS3
4. STEADINESS MEASUREMENT MODEL
The analysis of association and relationship tests results the development of a model for
measuring the stability of handheld applications usage, namely Stableness Measurement Model
(SMM) (Figure 2).
Data Entered
(M1•A1•CSTB)
ωIS1=.346
Information
Speed
(A1•CSTB)
STABILITY
(CSTB)
Errors Corrected
(M2•A1•CSTB)
ωIS2=.251 ωSTB1=.306
Lines Read
(M3•A1•CSTB)
ωIS3=.298
Targets Located
(M1•A2•CSTB)
ωLP1=.528
Lateral
Position
(A2•CSTB)Focuses Distracted
(M2•A2•CSTB)
ωLP2=.470 ωSTB2=.311
Links Explored
(M1•A3•CSTB)
ωOS1=.333
Optimal Solution
(A3•CSTB)
Steps Navigated
(M2•A3•CSTB)
ωOS2=.385 ωSTB3=.298
Paths Traversed
(M3•A3•CSTB)
ωOS3=.410
Figure 2. Stableness Measurement Model (SMM)
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
56
4.1. Measuring the Metrics
In order to measure the stability of handheld application usage, score for each metric can be
formulated and calculated generally as the proportion of the difference between number of
expected and actual activities occurred per time out of the total number of estimated activities
occurred per time. Hence can be represented as
Stability Metric
(M1…m•A1…a•CSTB)
=
Number of actual activities occurred per time – Number
of expected activities occurred per time (1)
Total number of expected activities occurred
Detail representation for measuring stability metrics Data Entered (M1•A1•CSTB), Errors
Corrected (M2•A1•CSTB) and Lines Read (M3•A1•CSTB) that contribute towards attribute
Information Speed (A1•CSTB), thus can be referred as
Data Entered
(M1•A1•CSTB)
=
Number of actual data entered per time – Number of
expected data entered per time
(1.1a)
Total number of expected data entered per time
Errors Corrected
(M2•A1•CSTB)
=
Number of actual errors corrected per time – Number
of expected errors corrected per time (1.2a)
Total number of expected errors corrected per time
Lines Read
(M3•A1•CSTB)
=
Number of actual lines read per time – Number of
expected lines read per time (1.3a)
Total number of expected lines read per time
Detail representation for measuring stability metrics Targets Located (M1•A2•CSTB) and Focuses
Distracted (M2•A2•CSTB) that contribute towards attribute Lateral Position (A2•CSTB), thus can be
referred as
Targets Located
(M1•A2•CSTB)
=
Number of actual targets located per time – Number of
expected targets located per time (1.1b)
Total number of expected targets located per time
Focuses Distracted
(M2•A2•CSTB)
=
Number of actual focuses distracted per time – Number
of expected focuses distracted per time (1.2b)
Total number of expected focuses distracted per time
Detail representation for measuring stability metrics Links Explored (M1•A3•CSTB), Steps
Navigated (M2•A3•CSTB) and Paths Traversed (M3•A3•CSTB) that contribute towards attribute
Optimal Solution (A3•CSTB), thus can be referred as
Links Explored
(M1•A3•CSTB)
=
Number of actual links explored per time – Number of
expected links explored per time (1.1c)
Total number of expected links explored per time
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
57
Steps Navigated
(M2•A3•CSTB)
=
Number of actual steps navigated per time – Number of
expected steps navigated per time (1.2c)
Total number of expected steps navigated per time
Paths Traversed
(M3•A3•CSTB)
=
Number of actual paths traversed per time – Number of
expected paths traversed per time (1.3c)
Total number of expected paths traversed per time
4.2. Measuring the Attributes
Score for each stability attribute can be formulated and calculated generally as the proportion of
the accumulated product of attribute weight and the metric value out of the total of accumulated
weight for each stability attribute. Hence can be represented as
Stability Attribute
(A1…a•CSTB)
=
m = max(m)
ωATTm (M1…m•A1…a•CSTB)
(2a)
∑
m = 1
m = max(m)
ωATTm∑
m = 1
which can be further expanded as
Stability Attribute
(A1…a•CSTB)
=
ωATT1 (M1•A1…a•CSTB)
+
(2b)
ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m)
ωATT2 (M2•A1…a•CSTB)
+ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m)
…
ωATTmax(m)–1 (Mmax(m)–1•A1…a•CSTB)
+ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m)
ωATTmax(m) (Mmax(m)•A1…a•CSTB)
ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m)
Detail representations for measuring stability attribute Information Speed (A1•CSTB) that
contribute towards measuring the stability of handheld application usage can be referred as
Information Speed
(A1•CSTB)
=
m = 3
ωISm (Mm•A1•CSTB)
(2.1a)
∑
m = 1
m = 3
ωISm∑
m = 1
hence can be further expanded as
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
58
Information Speed
(A1•CSTB)
=
ωIS1 (M1•A1•CSTB)
+
(2.1b)
ωIS1 + ωIS2 + ωIS3
ωIS2 (M2•A1•CSTB)
+ωIS1 + ωIS2 + ωIS3
ωIS3 (M3•A1•CSTB)
ωIS1 + ωIS2 + ωIS3
which involved the proportion of the accumulated product of weight and value of each stability
metrics Data Entered (ωIS1[=.346] x M1•A1•CSTB), Errors Corrected (ωIS2[=.251] x M2•A1•CSTB)
and Lines Read (ωIS3[=.298] x M3•A1•CSTB) that contribute towards attribute Information Speed
(A1•CSTB) divide by the total of attribute weights (ωIS1 + ωIS2 + ωIS3 = .346 + .251 + .298 = 0.895).
Detail representations for measuring stability attribute Lateral Position (A2•CSTB) that contribute
towards measuring the stability of handheld application usage can be referred as
Lateral Position
(A2•CSTB)
=
m = 2
ωLPm (Mm•A2•CSTB)
(2.2a)
∑
m = 1
m = 2
ωLPm∑
m = 1
hence can be further expanded as
Lateral Position
(A2•CSTB)
=
ωLP1 (M1•A2•CSTB)
+
(2.2b)ωLP1 + ωLP2
ωLP2 (M2•A2•CSTB)
ωLP1 + ωLP2
which involved the proportion of the accumulated product of weight and value of each stability
metrics Targets Located (ωLP1[=.528] x M1•A2•CSTB) and Focuses Distracted (ωLP2[=.470] x
M2•A2•CSTB) that contribute towards attribute Lateral Position (A2•CSTB) divide by the total of
attribute weights (ωLP1+ ωLP2 = .528 + .470 = 0.998).
Detail representations for measuring stability attribute Optimal Solution (A3•CSTB) that contribute
towards measuring the stability of handheld application usage can be referred as
Optimal Solution
(A3•CSTB)
=
m = 3
ωOSm (Mm•A3•CSTB)
(2.3a)
∑
m = 1
m = 3
ωOSm∑
m = 1
hence can be further expanded as
Optimal Solution
(A3•CSTB)
=
ωOS1 (M1•A3•CSTB)
+ (2.3b)ωOS1 + ωOS2 + ωOS3
ωOS2 (M2•A3•CSTB) +
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
59
ωOS1 + ωOS2 + ωOS3
ωOS3 (M3•A3•CSTB)
ωOS1 + ωOS2 + ωOS3
which involved the proportion of the accumulated product of weight and value of each stability
metrics Links Explored (ωOS1[=.333] x M1•A3•CSTB), Steps Navigated (ωOS2[=.385] x
M2•A3•CSTB) and Paths Traversed (ωOS3[=.410] x M3•A3•CSTB) that contribute towards attribute
Optimal Solution (A3•CSTB) divide by the total of attribute weights (ωOS1 + ωOS2 + ωOS3 = .333 +
.385 + .410 = 1.128).
4.3. Measuring the Criterion
Score for Stability (CSTB) can be formulated and calculated generally as the proportion of the
accumulated product of criterion weight and the attribute value out of the total of accumulated
weights for each stability criterion. Hence can be represented as
Stability
(CSTB)
=
a = 3
ωSTBa (A1…a•CSTB)
(3a)
∑
a = 1
a = 3
ωSTBa∑
a = 1
which can be further expanded as
Stability
(CSTB)
=
ωSTB1 (A1•CSTB)
+
(3b)
ωSTB1 + ωSTB2 + ωSTB3
ωSTB2 (A2•CSTB)
+ωSTB1 + ωSTB2 + ωSTB3
ωSTB4 (A4•CSTB)
ωSTB1 + ωSTB2 + ωSTB3
which involved the proportion of the accumulated product of weight and value of each stability
attributes Information Speed (ωSTB1[=.306] x A1•CSTB), Lateral Position (ωSTB2[=.311] x A2•CSTB)
and Optimal Solution (ωSTB3[=.298] x A3•CSTB) that contribute towards measuring the stability of
handheld application (CACC) divide by the total of criterion weights (ωSTB1 + ωSTB2 + ωSTB3 = .306 +
.311 + .298 = 0.915).
Score for Stability (CSTB) can be further analysed according to five distinct classifications as
described below (Table 9). Prioritizing the stability of handheld application usage can be done by
converting the values into words or sentences with which evaluators from various backgrounds
and understanding can interpret the information accurately and comprehensively.
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
60
Table 9. Prioritizing stability level
Level Score (CSTB) Description
1 0.000 ≤ CSTB < 0.200 Most badly absence or shortage of a desirable usage
quality that attains stability level of unable to perform
comprehensively
2 0.200 ≤ CSTB < 0.400 Lack of a desirable usage quality that attains stability
level of the least excellent
3 0.400 ≤ CSTB < 0.600 Average of a desirable usage quality that can be tolerable
to consider good enough
4 0.600 ≤ CSTB < 0.700 Complete the specific requirements of a desirable usage
quality that achieves stability level of almost in a state of
being practical
5 0.800 ≤ CSTB ≤ 1.000 Fulfil all the requirements of a desirable usage quality that
achieves stability level of very high distinction of
proficiency
5. CONCLUSIONS
The model developed not only reveals the stability between handheld users and its application but
also provide a better understanding on the relationship of these factors. In addition, this model
can be established as a concrete evaluation technique for measuring the stability of handheld
application usage. For the future, it is recommended to evaluate cases between the stability
model and the actual handheld applications. With extensive experiences, stability measures
might change and additional new criteria could be included in the future work. Therefore, the
model developed need to be refined practically through many applications in the real work
environment.
REFERENCES
[1] Fayad, M. E. & Altman, A., (2001) “An introduction to software stability: Studying the design and
development that produces stable (or unstable) software”, Communications of the ACM, Vol. 44, No.
9, pp 95 – 98.
[2] ISO/IEC 9126, (2004) “Software engineering – Product quality. Part 1: 2001 – Parts 2 to 4”,
International Organization for Standardization, Geneva.
[3] Hina, M. D., (2012) “Quality attribute – Driven software architecture of a pervasive multimodal
computing system’, AMAIUB – CCS Research Journal, Vol. 2, No. 1.
[4] Sharma A. K., Kalia, A. & Singh, H., (2012) “The software quality assurance framework for CBSD”,
International Journal of Advanced Research in Computer Science and Software Engineering, Vol. 2,
No. 11, pp 161 – 170.
[5] Khaled, O. M. & Hosny, H. M., (2008) “A quality attribute-based approach for applying performance
design patterns”, International Journal of Software Engineering, Vol. 1, No. 1, pp 101 – 131.
[6] Hassenzahl, M., (2004) “The interplay of beauty, goodness, and usability in interactive products”,
Human-Computer Interaction, Vol. 19, No. 4, pp 319 – 349.
[7] Min, W., Jie, L., Hong, J., Ning, L., Wenzhao, W. & Zhenjie, H., (2011) “Evaluation on influence
factor of mechanism usability based on analytical prey incidence process”, Proceedings of the 2011
International Conference on Consumer Electronics, Communications and Networks (CECNet), pp
2200 – 2203.
[8] Pan, W., (2011) “Quantifying the stability of software systems via simulation in dependency
networks”, World Academy of Science, Engineering and Technology, Vol. 60, pp 1513 – 1520.
[9] Tuch, A., Roth, S., Hornbæk, K., Opwisa, K., & Bargas-Avilaa, J., (2012) “Is beautiful really usable?
Toward understanding the relation between usability, aesthetics, and affect in HCI”, Computers in
Human Behavior, Vol. 28, No. 5, pp 1596 – 1607.
International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013
61
[10] Li, Z. & Bai, X., (2010) “Influences of perceived risk and system usability on the adoption of mobile
banking service”, Proceedings of the Third International Symposium on Computer Science and
Computational Technology (ISCSCT), pp 51 – 54.
[11] Yang, P., Qin, G., Yan, Y., Wang, H. & Zhang, L., (2011) “On the channel usability of wireless mesh
networks: When stability plays with you”, International Journal of Ad Hoc and Ubiquitous
Computing, Vol. 8, No. 1/2, pp 64 – 77.
[12] Nicolau, H. & Jorge, J., (2012) “Touch typing using thumbs: Understanding the effect of mobility and
hand posture”, Proceedings of the 2012 ACM Annual Conference on Human Factors in Computing
Systems (CHI 2012), New York, USA, ACM, pp 2683 – 2686.
[13] Oliveira, R., Cherubini, M. & Oliver, N., (2012) “Influence of usability on customer satisfaction: A
case study on mobile phone services”, Proceedings of the I–UxSED CEUR Workshop, pp 14 – 19.
Authors
Thirteen years of extensive experience in teaching and mentoring graduate students at
public and private higher education institutions and eleven years of comprehensive
experience in research and development specifically on handheld – based software
application systems with interest on the field of Software Engineering, specializing on
software quality, software certification, software management and human computer
interaction.

More Related Content

What's hot

Unit 4- Software Engineering System Model Notes
Unit 4- Software Engineering System Model Notes Unit 4- Software Engineering System Model Notes
Unit 4- Software Engineering System Model Notes arvind pandey
 
International Journal on Web Service Computing (IJWSC)
International Journal on Web Service Computing (IJWSC)International Journal on Web Service Computing (IJWSC)
International Journal on Web Service Computing (IJWSC)ijwscjournal
 
A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...
A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...
A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...ijesajournal
 
A Level Training Set with both a Computer-Based Control and a Compact Control...
A Level Training Set with both a Computer-Based Control and a Compact Control...A Level Training Set with both a Computer-Based Control and a Compact Control...
A Level Training Set with both a Computer-Based Control and a Compact Control...ijesajournal
 
A Systematic Mapping Review of Software Quality Measurement: Research Trends,...
A Systematic Mapping Review of Software Quality Measurement: Research Trends,...A Systematic Mapping Review of Software Quality Measurement: Research Trends,...
A Systematic Mapping Review of Software Quality Measurement: Research Trends,...IJECEIAES
 

What's hot (9)

Unit 4- Software Engineering System Model Notes
Unit 4- Software Engineering System Model Notes Unit 4- Software Engineering System Model Notes
Unit 4- Software Engineering System Model Notes
 
International Journal on Web Service Computing (IJWSC)
International Journal on Web Service Computing (IJWSC)International Journal on Web Service Computing (IJWSC)
International Journal on Web Service Computing (IJWSC)
 
Unit I
Unit IUnit I
Unit I
 
A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...
A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...
A LEVEL TRAINING SET WITH BOTH A COMPUTER- BASED CONTROL AND A COMPACT CONTRO...
 
A Level Training Set with both a Computer-Based Control and a Compact Control...
A Level Training Set with both a Computer-Based Control and a Compact Control...A Level Training Set with both a Computer-Based Control and a Compact Control...
A Level Training Set with both a Computer-Based Control and a Compact Control...
 
Object oriented analysis and design unit- iii
Object oriented analysis and design unit- iiiObject oriented analysis and design unit- iii
Object oriented analysis and design unit- iii
 
Object oriented analysis and design unit- v
Object oriented analysis and design unit- vObject oriented analysis and design unit- v
Object oriented analysis and design unit- v
 
Object oriented analysis and design unit- ii
Object oriented analysis and design unit- iiObject oriented analysis and design unit- ii
Object oriented analysis and design unit- ii
 
A Systematic Mapping Review of Software Quality Measurement: Research Trends,...
A Systematic Mapping Review of Software Quality Measurement: Research Trends,...A Systematic Mapping Review of Software Quality Measurement: Research Trends,...
A Systematic Mapping Review of Software Quality Measurement: Research Trends,...
 

Viewers also liked

Some practical considerations and a
Some practical considerations and aSome practical considerations and a
Some practical considerations and aijseajournal
 
E government maturity models a comparative study
E government maturity models a comparative studyE government maturity models a comparative study
E government maturity models a comparative studyijseajournal
 
A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...
A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...
A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...ijseajournal
 
EA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONS
EA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONSEA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONS
EA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONSijseajournal
 
Code coverage based test case selection and prioritization
Code coverage based test case selection and prioritizationCode coverage based test case selection and prioritization
Code coverage based test case selection and prioritizationijseajournal
 
Design and develop of en e learning content based on multimedia game
Design and develop of en e learning content based on multimedia gameDesign and develop of en e learning content based on multimedia game
Design and develop of en e learning content based on multimedia gameijseajournal
 
A SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHM
A SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHMA SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHM
A SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHMijcisjournal
 
News Alert - DTAA & IPPA Approved by Mauritius and Gabon
News Alert - DTAA & IPPA Approved by Mauritius and GabonNews Alert - DTAA & IPPA Approved by Mauritius and Gabon
News Alert - DTAA & IPPA Approved by Mauritius and GabonIntuit Consultancy
 
English homework
English homeworkEnglish homework
English homeworkMiHa ElLa
 
Editing
EditingEditing
Editingfryr
 
Prgramacion actividades juvenud oct nov-dic 2013
Prgramacion actividades juvenud oct nov-dic 2013Prgramacion actividades juvenud oct nov-dic 2013
Prgramacion actividades juvenud oct nov-dic 2013juanfra105
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...IEEEGLOBALSOFTTECHNOLOGIES
 
8 October Daily Technical Trader
8 October Daily Technical Trader 8 October Daily Technical Trader
8 October Daily Technical Trader QNB Group
 
ErgoForma - компрессионный трикотаж
ErgoForma - компрессионный трикотажErgoForma - компрессионный трикотаж
ErgoForma - компрессионный трикотажChromeo93
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...IEEEGLOBALSOFTTECHNOLOGIES
 
Notebook
NotebookNotebook
Notebookspy343
 

Viewers also liked (20)

4213ijsea05
4213ijsea054213ijsea05
4213ijsea05
 
Some practical considerations and a
Some practical considerations and aSome practical considerations and a
Some practical considerations and a
 
4213ijsea07
4213ijsea074213ijsea07
4213ijsea07
 
4213ijsea06
4213ijsea064213ijsea06
4213ijsea06
 
E government maturity models a comparative study
E government maturity models a comparative studyE government maturity models a comparative study
E government maturity models a comparative study
 
A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...
A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...
A FUZZY MATHEMATICAL MODEL FOR PEFORMANCE TESTING IN CLOUD COMPUTING USING US...
 
EA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONS
EA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONSEA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONS
EA-MDA MODEL TO RESOLVE IS CHARACTERISTIC PROBLEMS IN EDUCATIONAL INSTITUTIONS
 
Code coverage based test case selection and prioritization
Code coverage based test case selection and prioritizationCode coverage based test case selection and prioritization
Code coverage based test case selection and prioritization
 
Design and develop of en e learning content based on multimedia game
Design and develop of en e learning content based on multimedia gameDesign and develop of en e learning content based on multimedia game
Design and develop of en e learning content based on multimedia game
 
A SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHM
A SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHMA SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHM
A SECURE BLOCK PERMUTATION IMAGE STEGANOGRAPHY ALGORITHM
 
News Alert - DTAA & IPPA Approved by Mauritius and Gabon
News Alert - DTAA & IPPA Approved by Mauritius and GabonNews Alert - DTAA & IPPA Approved by Mauritius and Gabon
News Alert - DTAA & IPPA Approved by Mauritius and Gabon
 
English homework
English homeworkEnglish homework
English homework
 
Editing
EditingEditing
Editing
 
Prgramacion actividades juvenud oct nov-dic 2013
Prgramacion actividades juvenud oct nov-dic 2013Prgramacion actividades juvenud oct nov-dic 2013
Prgramacion actividades juvenud oct nov-dic 2013
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Mining contracts for business events an...
 
reference from Margit
reference from Margitreference from Margit
reference from Margit
 
8 October Daily Technical Trader
8 October Daily Technical Trader 8 October Daily Technical Trader
8 October Daily Technical Trader
 
ErgoForma - компрессионный трикотаж
ErgoForma - компрессионный трикотажErgoForma - компрессионный трикотаж
ErgoForma - компрессионный трикотаж
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT Collaboration in multicloud computing e...
 
Notebook
NotebookNotebook
Notebook
 

Similar to STABLENESSMEASUREMENTMODEL: A MATHEMATICAL APPROACH FORMEASURING THE STABILITY OF HANDHELD APPLICATION USAGE

Guidelines to Understanding Design of Experiment and Reliability Prediction
Guidelines to Understanding Design of Experiment and Reliability PredictionGuidelines to Understanding Design of Experiment and Reliability Prediction
Guidelines to Understanding Design of Experiment and Reliability Predictionijsrd.com
 
Unique fundamentals of software
Unique fundamentals of softwareUnique fundamentals of software
Unique fundamentals of softwareijcsit
 
Contributors to Reduce Maintainability Cost at the Software Implementation Phase
Contributors to Reduce Maintainability Cost at the Software Implementation PhaseContributors to Reduce Maintainability Cost at the Software Implementation Phase
Contributors to Reduce Maintainability Cost at the Software Implementation PhaseWaqas Tariq
 
Thesis Part II EMGT 699
Thesis Part II EMGT 699Thesis Part II EMGT 699
Thesis Part II EMGT 699Karthik Murali
 
ANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEM
ANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEMANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEM
ANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEMIAEME Publication
 
A metrics suite for variable categorizationt to support program invariants[
A metrics suite for variable categorizationt to support program invariants[A metrics suite for variable categorizationt to support program invariants[
A metrics suite for variable categorizationt to support program invariants[IJCSEA Journal
 
A Software Measurement Using Artificial Neural Network and Support Vector Mac...
A Software Measurement Using Artificial Neural Network and Support Vector Mac...A Software Measurement Using Artificial Neural Network and Support Vector Mac...
A Software Measurement Using Artificial Neural Network and Support Vector Mac...ijseajournal
 
SOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEW
SOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEWSOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEW
SOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEWijseajournal
 
Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.Alexander Decker
 
Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.Alexander Decker
 
quasi-uniform theta graph
quasi-uniform theta graphquasi-uniform theta graph
quasi-uniform theta graphMangaiK4
 
Principles of Health Informatics: Evaluating medical software
Principles of Health Informatics: Evaluating medical softwarePrinciples of Health Informatics: Evaluating medical software
Principles of Health Informatics: Evaluating medical softwareMartin Chapman
 
Lecture3
Lecture3Lecture3
Lecture3soloeng
 
Milestone One Company Identification You have been hired as a.docx
Milestone One Company Identification You have been hired as a.docxMilestone One Company Identification You have been hired as a.docx
Milestone One Company Identification You have been hired as a.docxARIV4
 
Class quality evaluation using class quality scorecards
Class quality evaluation using class quality scorecardsClass quality evaluation using class quality scorecards
Class quality evaluation using class quality scorecardsIAEME Publication
 

Similar to STABLENESSMEASUREMENTMODEL: A MATHEMATICAL APPROACH FORMEASURING THE STABILITY OF HANDHELD APPLICATION USAGE (20)

Ijcatr04051006
Ijcatr04051006Ijcatr04051006
Ijcatr04051006
 
Guidelines to Understanding Design of Experiment and Reliability Prediction
Guidelines to Understanding Design of Experiment and Reliability PredictionGuidelines to Understanding Design of Experiment and Reliability Prediction
Guidelines to Understanding Design of Experiment and Reliability Prediction
 
Ijetcas14 340
Ijetcas14 340Ijetcas14 340
Ijetcas14 340
 
Unique fundamentals of software
Unique fundamentals of softwareUnique fundamentals of software
Unique fundamentals of software
 
Contributors to Reduce Maintainability Cost at the Software Implementation Phase
Contributors to Reduce Maintainability Cost at the Software Implementation PhaseContributors to Reduce Maintainability Cost at the Software Implementation Phase
Contributors to Reduce Maintainability Cost at the Software Implementation Phase
 
Thesis Part II EMGT 699
Thesis Part II EMGT 699Thesis Part II EMGT 699
Thesis Part II EMGT 699
 
ANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEM
ANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEMANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEM
ANALYZABILITY METRIC FOR MAINTAINABILITY OF OBJECT ORIENTED SOFTWARE SYSTEM
 
A metrics suite for variable categorizationt to support program invariants[
A metrics suite for variable categorizationt to support program invariants[A metrics suite for variable categorizationt to support program invariants[
A metrics suite for variable categorizationt to support program invariants[
 
Using dematel to analyze the quality characteristics of mobile
Using dematel to analyze the quality characteristics of mobileUsing dematel to analyze the quality characteristics of mobile
Using dematel to analyze the quality characteristics of mobile
 
A Software Measurement Using Artificial Neural Network and Support Vector Mac...
A Software Measurement Using Artificial Neural Network and Support Vector Mac...A Software Measurement Using Artificial Neural Network and Support Vector Mac...
A Software Measurement Using Artificial Neural Network and Support Vector Mac...
 
SOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEW
SOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEWSOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEW
SOFTWARE CODE MAINTAINABILITY: A LITERATURE REVIEW
 
Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.
 
Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.Constructing a musa model to determine priority factor of a servperf model.
Constructing a musa model to determine priority factor of a servperf model.
 
quasi-uniform theta graph
quasi-uniform theta graphquasi-uniform theta graph
quasi-uniform theta graph
 
Principles of Health Informatics: Evaluating medical software
Principles of Health Informatics: Evaluating medical softwarePrinciples of Health Informatics: Evaluating medical software
Principles of Health Informatics: Evaluating medical software
 
Lecture3
Lecture3Lecture3
Lecture3
 
Milestone One Company Identification You have been hired as a.docx
Milestone One Company Identification You have been hired as a.docxMilestone One Company Identification You have been hired as a.docx
Milestone One Company Identification You have been hired as a.docx
 
1841 1843
1841 18431841 1843
1841 1843
 
1841 1843
1841 18431841 1843
1841 1843
 
Class quality evaluation using class quality scorecards
Class quality evaluation using class quality scorecardsClass quality evaluation using class quality scorecards
Class quality evaluation using class quality scorecards
 

Recently uploaded

#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024BookNet Canada
 
Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Neo4j
 
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024BookNet Canada
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Science&tech:THE INFORMATION AGE STS.pdf
Science&tech:THE INFORMATION AGE STS.pdfScience&tech:THE INFORMATION AGE STS.pdf
Science&tech:THE INFORMATION AGE STS.pdfjimielynbastida
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsAndrey Dotsenko
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr LapshynFwdays
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
Unlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsUnlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsPrecisely
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentationphoebematthew05
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 

Recently uploaded (20)

#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptxVulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
 
Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024
 
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Science&tech:THE INFORMATION AGE STS.pdf
Science&tech:THE INFORMATION AGE STS.pdfScience&tech:THE INFORMATION AGE STS.pdf
Science&tech:THE INFORMATION AGE STS.pdf
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
Unlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power SystemsUnlocking the Potential of the Cloud for IBM Power Systems
Unlocking the Potential of the Cloud for IBM Power Systems
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentation
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 

STABLENESSMEASUREMENTMODEL: A MATHEMATICAL APPROACH FORMEASURING THE STABILITY OF HANDHELD APPLICATION USAGE

  • 1. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 DOI : 10.5121/ijsea.2013.4504 49 STABLENESS MEASUREMENT MODEL: A MATHEMATICAL APPROACH FOR MEASURING THE STABILITY OF HANDHELD APPLICATION USAGE Amalina Farhi Ahmad Fadzlah1 1 Department of Computer Science, Faculty of Defence Science and Technology, Universiti Pertahanan Nasional Malaysia, 57000 Kuala Lumpur, Malaysia amalina.farhi@upnm.edu.my ABSTRACT This study is designed to develop a mathematical model for measuring the stability of handheld application usage namely Stableness Measurement Model (SMM). This model outlined a series of formulas based on the total number of eleven stability measures (i.e. eight stability metrics and three stability attributes) which are identified as having associated and contributed towards measuring the stability of handheld application usage. This model is valuable as an alternative evaluation technique to be used for measuring and ensuring the stability of handheld application usage. KEYWORDS Stability, model, measure, handheld, application 1. INTRODUCTION Stability is one of the most fundamental and important of all usable and useful software characteristics. The term stability means making the condition of software of being resistant to change of position or condition with which not easily moved or disturbed [1]. Other definition described stability as quality or attribute of being firm and steadfast to software that bare on the provision of right or agreed results or effects with continuous function well in an acceptable period [2]. In this paper, the term stability in the perspective of useful and usable can be defined as the degree to which making the condition of software of being stable or steady in relation to correct or complete as well as effort and time, that reflects the real world object or event being described, based on the users’ needs and requirements. The fewer failures and times taken to complete tasks that are observed the more stable an application is. Stability normally plays an important factor for all software quality elements. Over the past few decades, several researches for assessing and evaluating stability of software have been mentioned. The international standard, ISO/IEC 9126 [2], described stability as quality sub attribute to software that bare on the provision of the ease with which a product can be maintained in order to improve reliability. In the other hand, stability correlates with the metrics which measure attributes of the software that indicate about the risk of unexpected effects as a result of modification [3][4][5]. Some researchers also classify stability as an essential characteristic for evaluating and assessing the usability of software [6][7][8][9]. Within the domain of handheld software, several researchers have proposed to explore the concept of stability [10][11][12][13].
  • 2. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 50 2. MATERIALS AND METHODS This study outlined three main questions: 1) what stability measures are really important; 2) what is the rank of each stability measure; and 3) what is the weight of each stability measure, towards measuring the stability of handheld application usage. In order to answer these questions, a questionnaire survey, namely Investigating Stability Measures for Handheld Application Usage was conducted among handheld device users and a total number of two hundred nineteen respondents responded. These stability measures were classified into three hierarchical levels of metrics, attributes and criterions. Metrics are described as the lowest hierarchy level. The main objective of the metrics is to identify measurable data for the purpose of measuring the stability of handheld application usages. The middle hierarchy level is described as attributes, whereas the highest is described as criterion (i.e. stability of handheld application usage). This hierarchy which brings together three different stability levels of metrics, attributes and criterion is as detailed below (Table 1). Table 1. Stability hierarchy level Hierarchy Description Metric The lowest hierarchy level; A collection of measurable stability data expressed in units Attribute The middle hierarchy level; A collection of metrics which belongs to a class of stability measures Criterion The highest hierarchy level; A collection of attributes for measuring the stability of handheld application usage A total number of eleven stability measures, with a number of eight stability metrics and three stability attributes, were outlined as having associated towards measuring the stability of handheld application usage. The definition of each stability measure is as depicted below (Table 2). Table 2. Stability measures and descriptions Measure Description Information Speed* Capability in handling information per time Lateral Position* Capability in positioning objects per time Optimal Solution* Capability in solving tasks per time Data Entered** The number of data entered per time Errors Corrected** The number of errors corrected per time Focuses Distracted** The number of focuses distracted per time Lines Read** The number of lines read per time Links Explored** The number of links explored per time Paths Traversed** The number of paths traversed per time Steps Navigated** The number of steps navigated per time Targets Located** The number of targets located per time Legend of the table: * attribute determines the criterion (i.e. stability of handheld application usage) ** metric determines the attribute Data collected from the questionnaire is entered in the Statistical Package for the Social Sciences (SPSS) for the analysis process as well as to classify the stability measures into the hierarchical structure of metrics, attributes and criterion. This brings together two parts of evaluation tests: Pearson’s Chi-square test and the Spearman’s Rho test. Pearson’s Chi-Square test was conducted to measure the amount of association between two different stability measures in two different
  • 3. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 51 hierarchy levels and the Spearman’s Rho test was conducted to comprehend the relationship strength between two different stability measures in two different hierarchy levels. The model for measuring the stability of handheld application usage is specifically developed using a conceptual framework, namely Stableness Measurement Framework (SMF) (Figure 1). This framework brings together different stability measures in different hierarchy levels. As illustrated below, the metric determines the attribute and the attribute determines the criterion (i.e. stability as criterion). Each level represents interaction with other level and the impact to one another to measure the stability of the desired handheld application usage. This can be explained as either none, one or more metrics to represent a single attribute. The combination of these metrics could be represented as the components that contributed to only one attribute. And finally, these attributes are used to support in the calculation of the criterion that can be concluded as directly affected the stability of handheld application usage. This is the case at every level in which could be represented as an M-1 relationship. For example, metric M1 … Mn are the input to attribute AA and criterion CC is an output for the attribute AA. Consider if the value of metric M1, M2, … , Mn-1 or Mn increases so as the value of attribute AA and criterion CC. Again, if the value of metric M1, M2, … , Mn-1 or Mn decreases so as the value of attribute AA and criterion CC. 3. RESULTS AND DISCUSSIONS 3.1. Results of Association Test Result of association test reported that metrics of Data Entered (M = 4.52, SD = .738), Errors Corrected (M = 4.10, SD = .979) and Focuses Distracted (M = 3.65, SD = 1.027) were contributed towards measuring the stability of handheld application usage with p < .001. Results also showed that metrics of Links Explored (M = 4.25, SD = .780), Lines Read (M = 4.08, SD = .992) and Paths Traversed (M = 3.95, SD = .912) were contributed towards attribute Optimal Solution with p < .001. Meanwhile, metrics of Steps Navigated (M = 4.04, SD = .905) and Targets Located (M = 3.69, SD = 1.038)were also found contributed towards measuring the stability of handheld application usage with p < .001. Figure 1. Stableness Measurement Framework (SMF)
  • 4. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 52 Finally, result of the association test also stated that the attributes of Information Speed (M = 4.34, SD = .811), Lateral Position (M = 4.27, SD = .734) and Optimal Solution (M = 4.27, SD = .806) were found contributed towards measuring the stability of handheld application usage, with p < .001. As a result, a total number of eleven stability measures (i.e. eight stability metrics and three stability attributes) were identified having associated and contributed towards measuring the stability of handheld application usage (Table 3). Table 3. Result of association test Stability Measures Mean Stability Attributes Information Speed 4.34 Lateral Position 4.27 Optimal Solution 4.27 Stability Metrics Data Entered 4.52 Errors Corrected 4.10 Focuses Distracted 3.65 Lines Read Speed 4.08 Links Explored 4.25 Paths Traversed 3.95 Steps Navigated 4.04 Targets Located 3.69 Results from the association test were further ranked to prioritize the level of importance of each stability measure towards measuring the overall stability of handheld application usage (Table 4). Table 4. Rank of stability measures Stability Measures Rank Stability Attributes Information Speed 1 Lateral Position 2 Optimal Solution 3 Stability Metrics Data Entered 1 Links Explored 2 Errors Corrected 3 Lines Read 4 Steps Navigated 5 Paths Traversed 6 Targets Located 7 Focuses Distracted 8 3.2. Results of Relationship Test Result of the relationship test revealed that there was a moderate and positive linear relationship between metrics Data Entered (R = .346), Errors Corrected (R = .251) and Lines Read (R = .298) towards attribute Information Speed with p < .001. Results also found that the coefficient value of metrics Targets Located (R = .528) and Focuses Distracted (R = .470) were moderate and positive linear relationship towards attribute Lateral Position with p < .001. Metrics Links
  • 5. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 53 Explored (R = .333), Steps Navigated (R = .385) and Paths Traversed (R = .410) were also reported to have a moderate and positive linear relationship between attribute Optimal Solution with p < .001. Finally, the relationship test also indicated the correlation strength between attributes Information Speed (R = .306), Lateral Position (R = .311) and Optimal Solution (R = .298) resulted having a moderate and positive linear relationship towards measuring the stability of handheld application usage with p < .001. Based on the result of the relationship test, out of the total number of eleven stability measures, seven (i.e. five stability metrics and two stability attributes) were identified having moderate and positive linear relationship, three (i.e. two stability metrics and one stability attributes) were identified having low and positive linear relationship, while only one stability matric reported having high and positive linear relationship towards measuring the stability of handheld application usage (Table 5). Table 5. Result of relationship test Stability Measures Relationship Attributes contributed towards criterion (attribute  criterion) Information Speed  Stability Moderate, Positive Lateral Position  Stability Moderate, Positive Optimal Solution  Stability Low, Positive Metrics contributed towards attribute (metric  attribute) Data Entered  Information Speed Moderate, Positive Links Explored  Optimal Solution Moderate, Positive Errors Corrected  Information Speed Low, Positive Lines Read  Information Speed Low, Positive Steps Navigated  Optimal Solution Moderate, Positive Paths Traversed  Optimal Solution Moderate, Positive Targets Located  Lateral Position High, Positive Focuses Distracted  Lateral Position Moderate, Positive Legend of the table: Correlation is significant at the 0.001 level (2-tailed) and range in the value of +1 to -1 Results from the relationship test were further analysed to obtain the value of weightage of each stability measure towards measuring the overall stability of handheld application usage (Table 6).
  • 6. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 54 Table 6. Weight of stability measures Stability Measures Weight Attributes contributed towards criterion (attribute  criterion) Information Speed  Stability .306 Lateral Position  Stability .311 Optimal Solution  Stability .298 Metrics contributed towards attribute (metric  attribute) Data Entered  Information Speed .346 Errors Corrected  Information Speed .251 Lines Read  Information Speed .298 Targets Located  Lateral Position .528 Focuses Distracted  Lateral Position .470 Links Explored  Optimal Solution .333 Steps Navigated  Optimal Solution .385 Paths Traversed  Optimal Solution .410 The findings derived from the analysis of both association test and relationship test produced lists of codes to represent each stability metric and attribute, presented as Mm•Aa•CSTB and Aa•CSTB respectively (Table 7). Symbolized as Mm, M represents the stability metrics while m represents the sequential series (m-th) of the stability metric such as 1, 2, …, m. Similarly, symbolized as Aa, A represents the stability attribute while a represents the sequential series (a-th) of the stability attribute such as 1, 2, …, a. Finally, symbolized as CSTB, C represents the stability criterion in which STB represents the abbreviation of the stability. Table 7. Code of stability measures Stability Measures Code Attributes contributed towards criterion (attribute  criterion) Aa• CSTB Information Speed  Stability A1•CSTB Lateral Position  Stability A2• CSTB Optimal Solution  Stability A3•CSTB Metrics contributed towards attributes (metrics  attributes) Mm•Aa•CSTB Data Entered  Information Speed M1•A1•CSTB Errors Corrected  Information Speed M2•A1•CSTB Lines Read  Information Speed M3•A1•CSTB Targets Located  Lateral Position M1•A2•CSTB Focuses Distracted  Lateral Position M2•A2•CSTB Links Explored  Optimal Solution M1•A3•CSTB Steps Navigated  Optimal Solution M2•A3•CSTB Paths Traversed  Optimal Solution M3•A3•CSTB Furthermore, findings derived from the analysis of both association test and relationship test also produced lists of codes to represent the weight of each stability metric and attribute, presented as ωATTm and ωCRTa respectively (Table 8). ω represents the symbol of weights, meanwhile symbolized as ATTm, m represents the sequential series (m-th) of the stability metric such as 1, 2, …, m that contributed towards particular attribute, ATT, and finally symbolized as CRTa, a represents the sequential series (a-th) of the stability attribute such as 1, 2, …, a that contributed towards measuring the stability of handheld application usage, in which CRT represents the abbreviation of the stability, coded as STB.
  • 7. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 55 Table 8. Weight code of stability measures Stability Measures Code Attributes contributed towards criterion (attribute  criterion) ωCRTa Information Speed  Stability ωSTB1 Lateral Position  Stability ωSTB2 Optimal Solution  Stability ωSTB3 Metrics contributed towards attributes (metrics  attributes) ωATTm Data Entered  Information Speed ωIS1 Errors Corrected  Information Speed ωIS2 Lines Read  Information Speed ωIS3 Targets Located  Lateral Position ωLP1 Focuses Distracted  Lateral Position ωLP2 Links Explored  Optimal Solution ωOS1 Steps Navigated  Optimal Solution ωOS2 Paths Traversed  Optimal Solution ωOS3 4. STEADINESS MEASUREMENT MODEL The analysis of association and relationship tests results the development of a model for measuring the stability of handheld applications usage, namely Stableness Measurement Model (SMM) (Figure 2). Data Entered (M1•A1•CSTB) ωIS1=.346 Information Speed (A1•CSTB) STABILITY (CSTB) Errors Corrected (M2•A1•CSTB) ωIS2=.251 ωSTB1=.306 Lines Read (M3•A1•CSTB) ωIS3=.298 Targets Located (M1•A2•CSTB) ωLP1=.528 Lateral Position (A2•CSTB)Focuses Distracted (M2•A2•CSTB) ωLP2=.470 ωSTB2=.311 Links Explored (M1•A3•CSTB) ωOS1=.333 Optimal Solution (A3•CSTB) Steps Navigated (M2•A3•CSTB) ωOS2=.385 ωSTB3=.298 Paths Traversed (M3•A3•CSTB) ωOS3=.410 Figure 2. Stableness Measurement Model (SMM)
  • 8. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 56 4.1. Measuring the Metrics In order to measure the stability of handheld application usage, score for each metric can be formulated and calculated generally as the proportion of the difference between number of expected and actual activities occurred per time out of the total number of estimated activities occurred per time. Hence can be represented as Stability Metric (M1…m•A1…a•CSTB) = Number of actual activities occurred per time – Number of expected activities occurred per time (1) Total number of expected activities occurred Detail representation for measuring stability metrics Data Entered (M1•A1•CSTB), Errors Corrected (M2•A1•CSTB) and Lines Read (M3•A1•CSTB) that contribute towards attribute Information Speed (A1•CSTB), thus can be referred as Data Entered (M1•A1•CSTB) = Number of actual data entered per time – Number of expected data entered per time (1.1a) Total number of expected data entered per time Errors Corrected (M2•A1•CSTB) = Number of actual errors corrected per time – Number of expected errors corrected per time (1.2a) Total number of expected errors corrected per time Lines Read (M3•A1•CSTB) = Number of actual lines read per time – Number of expected lines read per time (1.3a) Total number of expected lines read per time Detail representation for measuring stability metrics Targets Located (M1•A2•CSTB) and Focuses Distracted (M2•A2•CSTB) that contribute towards attribute Lateral Position (A2•CSTB), thus can be referred as Targets Located (M1•A2•CSTB) = Number of actual targets located per time – Number of expected targets located per time (1.1b) Total number of expected targets located per time Focuses Distracted (M2•A2•CSTB) = Number of actual focuses distracted per time – Number of expected focuses distracted per time (1.2b) Total number of expected focuses distracted per time Detail representation for measuring stability metrics Links Explored (M1•A3•CSTB), Steps Navigated (M2•A3•CSTB) and Paths Traversed (M3•A3•CSTB) that contribute towards attribute Optimal Solution (A3•CSTB), thus can be referred as Links Explored (M1•A3•CSTB) = Number of actual links explored per time – Number of expected links explored per time (1.1c) Total number of expected links explored per time
  • 9. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 57 Steps Navigated (M2•A3•CSTB) = Number of actual steps navigated per time – Number of expected steps navigated per time (1.2c) Total number of expected steps navigated per time Paths Traversed (M3•A3•CSTB) = Number of actual paths traversed per time – Number of expected paths traversed per time (1.3c) Total number of expected paths traversed per time 4.2. Measuring the Attributes Score for each stability attribute can be formulated and calculated generally as the proportion of the accumulated product of attribute weight and the metric value out of the total of accumulated weight for each stability attribute. Hence can be represented as Stability Attribute (A1…a•CSTB) = m = max(m) ωATTm (M1…m•A1…a•CSTB) (2a) ∑ m = 1 m = max(m) ωATTm∑ m = 1 which can be further expanded as Stability Attribute (A1…a•CSTB) = ωATT1 (M1•A1…a•CSTB) + (2b) ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m) ωATT2 (M2•A1…a•CSTB) +ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m) … ωATTmax(m)–1 (Mmax(m)–1•A1…a•CSTB) +ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m) ωATTmax(m) (Mmax(m)•A1…a•CSTB) ωATT1 + … + ωATTmax(m)–1 + ωATTmax(m) Detail representations for measuring stability attribute Information Speed (A1•CSTB) that contribute towards measuring the stability of handheld application usage can be referred as Information Speed (A1•CSTB) = m = 3 ωISm (Mm•A1•CSTB) (2.1a) ∑ m = 1 m = 3 ωISm∑ m = 1 hence can be further expanded as
  • 10. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 58 Information Speed (A1•CSTB) = ωIS1 (M1•A1•CSTB) + (2.1b) ωIS1 + ωIS2 + ωIS3 ωIS2 (M2•A1•CSTB) +ωIS1 + ωIS2 + ωIS3 ωIS3 (M3•A1•CSTB) ωIS1 + ωIS2 + ωIS3 which involved the proportion of the accumulated product of weight and value of each stability metrics Data Entered (ωIS1[=.346] x M1•A1•CSTB), Errors Corrected (ωIS2[=.251] x M2•A1•CSTB) and Lines Read (ωIS3[=.298] x M3•A1•CSTB) that contribute towards attribute Information Speed (A1•CSTB) divide by the total of attribute weights (ωIS1 + ωIS2 + ωIS3 = .346 + .251 + .298 = 0.895). Detail representations for measuring stability attribute Lateral Position (A2•CSTB) that contribute towards measuring the stability of handheld application usage can be referred as Lateral Position (A2•CSTB) = m = 2 ωLPm (Mm•A2•CSTB) (2.2a) ∑ m = 1 m = 2 ωLPm∑ m = 1 hence can be further expanded as Lateral Position (A2•CSTB) = ωLP1 (M1•A2•CSTB) + (2.2b)ωLP1 + ωLP2 ωLP2 (M2•A2•CSTB) ωLP1 + ωLP2 which involved the proportion of the accumulated product of weight and value of each stability metrics Targets Located (ωLP1[=.528] x M1•A2•CSTB) and Focuses Distracted (ωLP2[=.470] x M2•A2•CSTB) that contribute towards attribute Lateral Position (A2•CSTB) divide by the total of attribute weights (ωLP1+ ωLP2 = .528 + .470 = 0.998). Detail representations for measuring stability attribute Optimal Solution (A3•CSTB) that contribute towards measuring the stability of handheld application usage can be referred as Optimal Solution (A3•CSTB) = m = 3 ωOSm (Mm•A3•CSTB) (2.3a) ∑ m = 1 m = 3 ωOSm∑ m = 1 hence can be further expanded as Optimal Solution (A3•CSTB) = ωOS1 (M1•A3•CSTB) + (2.3b)ωOS1 + ωOS2 + ωOS3 ωOS2 (M2•A3•CSTB) +
  • 11. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 59 ωOS1 + ωOS2 + ωOS3 ωOS3 (M3•A3•CSTB) ωOS1 + ωOS2 + ωOS3 which involved the proportion of the accumulated product of weight and value of each stability metrics Links Explored (ωOS1[=.333] x M1•A3•CSTB), Steps Navigated (ωOS2[=.385] x M2•A3•CSTB) and Paths Traversed (ωOS3[=.410] x M3•A3•CSTB) that contribute towards attribute Optimal Solution (A3•CSTB) divide by the total of attribute weights (ωOS1 + ωOS2 + ωOS3 = .333 + .385 + .410 = 1.128). 4.3. Measuring the Criterion Score for Stability (CSTB) can be formulated and calculated generally as the proportion of the accumulated product of criterion weight and the attribute value out of the total of accumulated weights for each stability criterion. Hence can be represented as Stability (CSTB) = a = 3 ωSTBa (A1…a•CSTB) (3a) ∑ a = 1 a = 3 ωSTBa∑ a = 1 which can be further expanded as Stability (CSTB) = ωSTB1 (A1•CSTB) + (3b) ωSTB1 + ωSTB2 + ωSTB3 ωSTB2 (A2•CSTB) +ωSTB1 + ωSTB2 + ωSTB3 ωSTB4 (A4•CSTB) ωSTB1 + ωSTB2 + ωSTB3 which involved the proportion of the accumulated product of weight and value of each stability attributes Information Speed (ωSTB1[=.306] x A1•CSTB), Lateral Position (ωSTB2[=.311] x A2•CSTB) and Optimal Solution (ωSTB3[=.298] x A3•CSTB) that contribute towards measuring the stability of handheld application (CACC) divide by the total of criterion weights (ωSTB1 + ωSTB2 + ωSTB3 = .306 + .311 + .298 = 0.915). Score for Stability (CSTB) can be further analysed according to five distinct classifications as described below (Table 9). Prioritizing the stability of handheld application usage can be done by converting the values into words or sentences with which evaluators from various backgrounds and understanding can interpret the information accurately and comprehensively.
  • 12. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 60 Table 9. Prioritizing stability level Level Score (CSTB) Description 1 0.000 ≤ CSTB < 0.200 Most badly absence or shortage of a desirable usage quality that attains stability level of unable to perform comprehensively 2 0.200 ≤ CSTB < 0.400 Lack of a desirable usage quality that attains stability level of the least excellent 3 0.400 ≤ CSTB < 0.600 Average of a desirable usage quality that can be tolerable to consider good enough 4 0.600 ≤ CSTB < 0.700 Complete the specific requirements of a desirable usage quality that achieves stability level of almost in a state of being practical 5 0.800 ≤ CSTB ≤ 1.000 Fulfil all the requirements of a desirable usage quality that achieves stability level of very high distinction of proficiency 5. CONCLUSIONS The model developed not only reveals the stability between handheld users and its application but also provide a better understanding on the relationship of these factors. In addition, this model can be established as a concrete evaluation technique for measuring the stability of handheld application usage. For the future, it is recommended to evaluate cases between the stability model and the actual handheld applications. With extensive experiences, stability measures might change and additional new criteria could be included in the future work. Therefore, the model developed need to be refined practically through many applications in the real work environment. REFERENCES [1] Fayad, M. E. & Altman, A., (2001) “An introduction to software stability: Studying the design and development that produces stable (or unstable) software”, Communications of the ACM, Vol. 44, No. 9, pp 95 – 98. [2] ISO/IEC 9126, (2004) “Software engineering – Product quality. Part 1: 2001 – Parts 2 to 4”, International Organization for Standardization, Geneva. [3] Hina, M. D., (2012) “Quality attribute – Driven software architecture of a pervasive multimodal computing system’, AMAIUB – CCS Research Journal, Vol. 2, No. 1. [4] Sharma A. K., Kalia, A. & Singh, H., (2012) “The software quality assurance framework for CBSD”, International Journal of Advanced Research in Computer Science and Software Engineering, Vol. 2, No. 11, pp 161 – 170. [5] Khaled, O. M. & Hosny, H. M., (2008) “A quality attribute-based approach for applying performance design patterns”, International Journal of Software Engineering, Vol. 1, No. 1, pp 101 – 131. [6] Hassenzahl, M., (2004) “The interplay of beauty, goodness, and usability in interactive products”, Human-Computer Interaction, Vol. 19, No. 4, pp 319 – 349. [7] Min, W., Jie, L., Hong, J., Ning, L., Wenzhao, W. & Zhenjie, H., (2011) “Evaluation on influence factor of mechanism usability based on analytical prey incidence process”, Proceedings of the 2011 International Conference on Consumer Electronics, Communications and Networks (CECNet), pp 2200 – 2203. [8] Pan, W., (2011) “Quantifying the stability of software systems via simulation in dependency networks”, World Academy of Science, Engineering and Technology, Vol. 60, pp 1513 – 1520. [9] Tuch, A., Roth, S., Hornbæk, K., Opwisa, K., & Bargas-Avilaa, J., (2012) “Is beautiful really usable? Toward understanding the relation between usability, aesthetics, and affect in HCI”, Computers in Human Behavior, Vol. 28, No. 5, pp 1596 – 1607.
  • 13. International Journal of Software Engineering & Applications (IJSEA), Vol.4, No.5, September 2013 61 [10] Li, Z. & Bai, X., (2010) “Influences of perceived risk and system usability on the adoption of mobile banking service”, Proceedings of the Third International Symposium on Computer Science and Computational Technology (ISCSCT), pp 51 – 54. [11] Yang, P., Qin, G., Yan, Y., Wang, H. & Zhang, L., (2011) “On the channel usability of wireless mesh networks: When stability plays with you”, International Journal of Ad Hoc and Ubiquitous Computing, Vol. 8, No. 1/2, pp 64 – 77. [12] Nicolau, H. & Jorge, J., (2012) “Touch typing using thumbs: Understanding the effect of mobility and hand posture”, Proceedings of the 2012 ACM Annual Conference on Human Factors in Computing Systems (CHI 2012), New York, USA, ACM, pp 2683 – 2686. [13] Oliveira, R., Cherubini, M. & Oliver, N., (2012) “Influence of usability on customer satisfaction: A case study on mobile phone services”, Proceedings of the I–UxSED CEUR Workshop, pp 14 – 19. Authors Thirteen years of extensive experience in teaching and mentoring graduate students at public and private higher education institutions and eleven years of comprehensive experience in research and development specifically on handheld – based software application systems with interest on the field of Software Engineering, specializing on software quality, software certification, software management and human computer interaction.