Engineering and Applied Science Letter
ISSN: 2617-9709 (Online) 2617-9695 (Print)
DOI: 10.30538/psrp-easl2020.0033
Risk evaluation in information systems using continuous and discrete distribution laws
Ajit Singh\(^1\), Amrita Prakash
Department of Computer Science, Patna Women’s College Bihar, India.; (A.S & A.P)
Abstract
Keywords:
1. Introduction
At the present stage of the development of society, which is characterized by the intensive introduction of information systems in virtually areas of activity, issues related to the assessment of the risks that occur during their operation are of particular importance. When analyzing and assessing risks, issues related to the definition of distribution laws are of the greatest importance. The given work is devoted to the construction of distribution laws.
In the modeling of information systems, risk is a random variable and is described by a probability distribution on a given set [1, 2, 3]. In contrast to experiments conducted in physics, where the possibility of their multiple conduct, the conditions of the functioning of information systems are characterized by a constant impact of negative external influences and are constantly changing [4], and consequently the repetition of the experiment under the same conditions is practically impracticable. The laws of probability distribution of risk events, as a rule, do not correspond to the law of the normal Gaussian distribution [5, 6].
2. Construction of continuous distribution laws with the maximum entropy
Entropy coefficient is often used [7, 8] with the classification of distribution laws of random continuous value (RV) with number characteristics.The entropy value does not depend on shift parameter, to simple computation let's consider, that it is equal to zero. Firstly we need to find distribution law from unilateral laws of distribution of unlimited RV, for which entropy value (2) reaches the maximum with the following limitations imposed on probability density \(p(x)\):
From (9), it follows that if only first absolute central moment exists \(( \nu = 1) \) then Laplace distribution has the biggest entropy; if there are two moments \(( \nu = 2)\) then Gaussian law and if all direct moments exist \((\nu \rightarrow \infty )\) then uniform law. Indeed, the limiting case for (9) is a uniform law \(p(x) = 0.5 \beta^-1,~ -\beta < x < \beta \). So, if all direct moments exist then uniform law has the biggest entropy from bilateral symmetry distribution laws of RV. Considered private cases of bilateral laws with the maximum entropy coincide with already known laws (Laplace and Gaussian) which have maximum entropy that confirms the correctness of received results.
From analysis of the received expressions (8) and (9), it follows that for increasing the amount of information about evaluating parameters of distribution laws with big length (with long "tails") with the help of a method of moments is necessary to use direct moments of lesser order, including fractional order. If the parameters of distribution laws with lesser length are used then it is necessary to use direct moment of higher order.
Let's find from unilateral distribution laws of unlimited RV distribution law with which entropy value H reaches maximum with the following limitations imposed on probability density \(p(x)\):Let's define the bilateral distribution laws of RV for which entropy value \(H\) reaches the maximum with the following limitations imposed on probability density \(p(x)\)
From the analysis of expressions (15) and (18), it follows that exponential transformation of RV leads to transformation of form parameter \(\nu\) in scale parameter and \(\beta\) parameter in shift parameter.
Finally let's define such distribution law from unilateral distribution laws of unlimited RV for which the value of entropy H reaches maximum with the following limitations imposed on probability density \(p(x)\):
In general case, if RV \(\eta\) connected with RV \(\eta \square\) by a ratio \(y =f(x)\) and known PDD \(p(y)\) of continuous RV \(\xi\) , then PDD \(p(x)\) can be found by a method of functional transformation with the help of expression:
3. Distributions arising in the analysis of the sequence of independent tests with three outputs
Next, consider the development of a probabilistic model of a sequence of independent trials with three outcomes which becomes particularly important in the formation of estimates of the information security of information processing systems [10].During the test, it is taken into account that its result is either event A or the opposite event C. The probability of event A in any test is independent of the outcomes of all other tests (the tests are independent) and equal to the probability (this is ensured by the same set of conditions for each test). This scheme of tests was first considered by J. Bernoulli and bears his name [11, 12, 13, 14]. The probability \(P_A(k)\) of the fact that event \(A\) in \(N\) tests will come precisely \(k\) times (\(k = 1,2 , \dots , N\)) is defined by Bernoulli's formula [13, 14, 15]:
- If \( p_1 = p_2 = p < 0.5\), then
\begin{align}\label{equ34} P(k) = (1 - 2p) ^N \times \sum_{i = |k|}^{N} \frac{N!}{(N - i)!} \times \biggl(\frac{p}{1 - 2p} \biggr)^i \times \frac{0.5[1+ (-1)^{i + |k|}]}{\Gamma [0.5 (i + k) + 1] \Gamma [0.5 (i -k) + 1]} \end{align}(34)\begin{align}\label{equ35} P(k) = \frac{(2N)!}{(N - k)! (N+ k)!} \times p^(N + k) (1 - p)^{(N - k)}, ~ k = -N, ~ -(N - 1), \dots , N \end{align}(35)
- Let's view limiting case for distribution (33), when probability of coming value \(C\) is aims to zero, i.e., \((p_1 + p_2) \rightarrow 1\). In this case every test will end in two outputs: either coming of event \(A\) with the probability \((1 p)\), or event \(B\) with the probability \(p\). Those outputs can be matched discrete random value, which takes two values: -1, if event \(A\) happened and 1, if event \(B\) happened. In this probability distribution (33), the result can be transformed to distribution:
\begin{align}\label{equ36} P(k) = (0.5 N! [1 + (-1)^{N + |k|}]) \times (\Gamma [0.5 (N + k ) + 1] \Gamma [0.5 (N - k ) + 1 ]^-1) \times \biggl( \frac{p}{1 - p} \biggr)^{0.5k} (p (1 - p))^{0.5N} \end{align}(36)
- Let's view the second limiting case for distribution (33), when probability of coming event \(A\) aims to zero, i.e. \(p_1 \rightarrow 0\). In this case every test will end in two outputs: either coming of event \(C\) with a probability \((i - p)\), or event \(B\) with a probability \(p\). Those outputs can be matched random discrete value, which takes two values: 0, if event \(C\) happened and 1, if event \(B\) happened. This probability distribution (33) as a result of limiting transition transforms is the binomial distribution (24) and that's why received probability distribution (33) can be called generalized Bernoulli's formula, or bilateral binomial distribution.
-
Let's view the third limiting case for distribution (33), when \(p_1 \rightarrow 0, ~ p_2 \rightarrow 0, ~ N \rightarrow \infty\), and products \(Np_1, ~ Np_2\) aim to some positive constant values \(\lambda_1\), \(\lambda_2\) (i.e. \(Np_1 \rightarrow \lambda_1, ~ Np_2 \rightarrow \lambda_2 \) ). This probability distribution (33) in result of limiting transition transforms is the probability distribution either
\begin{align}\label{equ37} P(k) = exp ( -\lambda_1 - \lambda_2) \biggl( \sqrt{\frac{\lambda_2}{\lambda_1}} \biggr)^k \times \sum_{i = |k|}^{\infty} \frac{0.5[1+ (-1)^{i + |k|}] \sqrt{\lambda_1 \lambda_2}^i}{\Gamma [0.5 (i + k) + 1] \Gamma [0.5 (i -k) + 1]} \end{align}(37)\begin{align}\label{equ38} P(k) = exp ( -\lambda_1 - \lambda_2) \times \biggl( \sqrt{\frac{\lambda_2}{\lambda_1}}\biggr)^k I_{|k|} (2 \sqrt{\lambda_1 \lambda_2}), ~ -\infty < k < \infty \end{align}(38)
Probability \(P_A(k)\) of fact, that event \(A\) in \(N\) tests will come \(k\) times can be also found from formula (33), or it's private cases (34), (35), (36), (37) or (38). In this we suppose that \(P_A(k) = P(k), k = -1, -2, \dots , -N\).
Probability \(P_C\) of coming event \(C\) in \(N\) tests can be found using formula (33), or it's private cases (35), (36), (37) or (38). In that we suppose, that \(P_C = P(0)\). Probability \(P_C\) matches to probability of fact, that in \(N\) cases events \(A\) and \(B\) won't come.
Let's view the example. Two symmetric coins are being thrown for ten rimes. In every throw three outputs are possible: two "eagles" with probability 0.25; two "tails of coin" with probability 0.25 and "eagle and tail of coin" with probability 0.5. It's necessary to find: 1) probability of fact, that precisely five times two "eagles" drop; 2) probability \(P_{tt}\) of fact, that precisely three times two "tails of coin" drop; 3) probability \(P_{et}\) of fact, that precisely five times two "eagles" and three "tales of coin" drop. In the match with example's condition we have \begin{align*} p_1 = p_2 = p = 0.25,\,\,\,\,\,\,\,\,\,\,\,\, N = 10; ~ p_{ee} = P_A (-5),\,\,\,\,\,\,\,\,\,\,\,\, P_{tt} = P_B (3),\,\,\,\,\,\,\,\,\,\,\,\, P_{et } = P_A (-5) P_B (3). \end{align*} i.e., \(p_1 = p_2\), then we use expression (9) as a counting formula. With it's help we find, that either
Choosing from bilateral binomial, Poisson's and negative binomial distributions, we can use following properties of those distributions: Binomial - \(K_a M_2 < 1\), Poisson's - \(K_e M_2 = 1 \), Negative binomial - \(K_e M_2 > 1\).
So, there was developed probability model for sequence of independent tests with three outputs, were received expressions for it's general number characteristics, and also for calculating the probabilities of coming matched events precisely k times. It was shown, that limiting cases of received bilateral distributions are binomial, negative binomial and Poisson's distributions.
4. Conclusion
The following results are obtained in this paper- Generalized expressions for one-way and two-way continuous distribution laws with maximum entropy depending on the number of existing power, exponential or logarithmic moments. With their help, one can more reasonably choose the a priori distribution under the conditions of a priori uncertainty in the analysis of the risks of information systems. From the analysis of expression (23) and its particular cases (2), (11), (14), (17), (20) at the appropriate values \(q(x)\) it follows that in the general case the entropy depends also on the type of moments used to determine the numerical characteristics of the distribution law.
- Probabilistic model for a sequence of independent trials with three outcomes, which acquire special significance in the formation of information security assessments of information systems. Expressions for its basic numerical characteristics are obtained. It is shown that the limiting cases of the obtained two-way distributions are the binomial, negative binomial and Poisson distributions.
Acknowledgments
The authors would like to express their gratitude to University of Lagos for providing the enabling environment to conduct this research work.References
- Burkov, V. N., Novikov, D. A., & Shchepkin, A. V. (2015). Control mechanisms for ecological-economic systems. Springer International Publishing.[Google Scholor]
- Kornilov, S.N., & Kornilova, M.M. (2017). Information processing technique for analyzing the operation of freight stations of Russian Railways. Modern Problems of the Transport Complex of Russia , 4(1), 49-52.[Google Scholor]
- Broderick, J. S. (2006). ISMS, security standards and security regulations. Information Security Technical Report, 11(1), 26-31.[Google Scholor]
- Otten, K., & Debons, A. (1970). Towards a metascience of information: Informatology. Journal of the American Society for Information Science, 21(1), 89-94.[Google Scholor]
- Kulba, V., Bakhtadze, N., Zaikin, O., Shelkov, A., & Chernov, I. (2017). Scenario analysis of management processes in the prevention and the elimination of consequences of man-made disasters. Procedia Computer Science, 112, 2066-2075.[Google Scholor]
- Schulz, V. L., Kul'ba, V. V., Shelkov, A. B., & Chernov, I. V. (2013). Methods of scenario analysis of threats to the effective functioning of systems of organizational management. Trends and Management, 1(1), 6-30.[Google Scholor]
- Gromov, N., & Kazakov, V. (2012). Review of AdS/CFT integrability, chapter III. 7: Hirota dynamics for quantum integrability. Letters in Mathematical Physics, 99(1-3), 321-347.[Google Scholor]
- Moustafa, N., & Slay, J. (2016). The evaluation of Network Anomaly Detection Systems: Statistical analysis of the UNSW-NB15 data set and the comparison with the KDD99 data set. Information Security Journal: A Global Perspective, 25(1-3), 18-31.[Google Scholor]
- Jones, M. R., & Karsten, H. (2008). Giddens's structuration theory and information systems research. MIS quarterly, 32(1), 127-157.[Google Scholor]
- Gromov, Y. Y., Karpov, I. G., Minin, Y. V., & Ivanova, O. G. (2016). Generalized probabilistic description of homogeneous flows of events for solving informational security problems. Journal of Theoretical & Applied Information Technology, 87(2), 250-254.[Google Scholor]
- Chung, K. L., & Zhong, K. (2001). A course in probability theory. Academic press.[Google Scholor]
- Levin, B. R. (1989). Theoretical bases of statistical radio engineering. Moscow: Radio and telecommunication.[Google Scholor]
- Prokhorov, Y. V. (1999). Probability and Mathematical Statistics (Encyclopedia). Moscow: Bolshaya Rossiyskaya encyclopaedia Publisher. [Google Scholor]
- Kai, T., & Tomita, K. (1980). Statistical mechanics of deterministic chaos: The case of one-dimensional discrete process. Progress of Theoretical Physics, 64(5), 1532-1550.[Google Scholor]
- Gardiner, C. W. (1985). Handbook of stochastic methods (Vol. 3, pp. 2-20). Berlin: springer. [Google Scholor]