American Journal of Mathematics and Statistics

p-ISSN: 2162-948X    e-ISSN: 2162-8475

2012;  2(1): 20-24

doi: 10.5923/j.ajms.20120201.05

Measuring Complex Networks

Angel Garrido

Department of Fundamental Mathematics, Faculty of Sciences,UNED, PaseoSendadel Rey 9, Madrid, 28040, Spain

Correspondence to: Angel Garrido , Department of Fundamental Mathematics, Faculty of Sciences,UNED, PaseoSendadel Rey 9, Madrid, 28040, Spain.

Email:

Copyright © 2012 Scientific & Academic Publishing. All Rights Reserved.

Abstract

This paper analyzes some aspects of Fuzziness and Uncertainty Measures. We would need new ways to obtain some more adequate conditions, or restrictions, to modeling from vague pieces of information. Many times, classical fuzzy measure proceeds from Physics; for instance, from Thermodynamics. Furthermore, it was adapted, so creating new constructs such as Information Theory, with very crucial concepts, such as Symmetry and Entropy. The Hungarian mathematician Alfred Rényi shows that we may propose very different and valid entropy measures, according our purposes. So, it will be very necessary to clarify some different types of measures, and also their mutual relationships. For these reasons, we attempt to obtain a generalization and to improve the classification of some interesting fuzzy measures, as may be the Entropy on Complex Networks.

Keywords: Mathematics, Entropy, Symmetry, Information Theory, Complex Networks

Cite this paper: Angel Garrido , "Measuring Complex Networks", American Journal of Mathematics and Statistics, Vol. 2 No. 1, 2012, pp. 20-24. doi: 10.5923/j.ajms.20120201.05.

1. Introduction

Statistical entropy is a probabilistic measure of uncertainty, or ignorance about data. Information is a measure of a reduction in that uncertainty. But must be avoided according to Frank Lambert[39]the interpretation of entropy as disorder. So, instead proposing the notion as “dispersal of energy”.
We define I in terms of the probability p, by these
Properties of Information Measure, I:
1) I (p) ≥ 0; i.e. information content will be a non-negative quantity.
2) I (1) = 0; i.e. if an event has probability 1, we get no information from the occurrence of the event.
3) If two independent events occur, the information we get from observing the events is the sum of both information.
4) Information measure must be continuous, and also a monotonic function of the probability: slight changes in probability should result in slight changes in I.
From an advanced mathematical viewpoint[37,38,44], Kaufmann (1975) introduced the so-called “index of fuzziness” as a normal distance, using the functional form of Shannon´s entropy to define a discrete fuzzy set, A. Then, the Kaufmann´s entropy of A will be
Also must be mentioned[24-27] the very remarkable Iosifescu-Theodorescu Entropy (denoted IT-entropy, 1961), which, for instance, when it is defined on a stationary and homogeneous Markov Chain.For instance, we can obtainthe IT- entropy as a measure of the mobility for social processes as multiple Markov Chains. It would be justified, because the IT-entropy of a Markov Chain reaches a maximum when X(r) is independent of all their precedent states.
Yager (1979), and Higashi and Klir (1983), showed the entropy measure as the difference between two fuzzy sets. More concretely, it will be the difference between a fuzzy set and its complement, which is also a fuzzy set.
Pal, N. R., and Pal, S. K. (1989) introduce a definition for probabilistic entropy, in this case based on an exponential gainfunction, and used it as basis for defining this measure of fuzziness.
Bhandari and Pal (1993) use Renyi´s entropy to define a non-probabilistic entropy, or measure of fuzziness, for fuzzy sets.

2. Entropy

Given a discrete random variable, X, with an associated probability distribution, P (x), we will define[17,18] the Entropy of X by
H (x) = - p(x) ln p(x) =
= p(x) ln (1/p(x)) = E [ln (1/p(x)]
Therefore, such H is a measure of the quantity of information that we receive, when it is sent towards us.
The logarithmic base will be arbitrary. If b = 2, it is measured in bits. The more logical election, because the system of interest, is binary. If b = 10, will be in dits. And if b = e, will be in nats.
The Shannon Entropy is a measure of the average information content one is missing when one does not know the value of the random variable. These concepts proceed from the Shannon´s 1948 very famous paper and their derivations[46,47,49], i.e. the so-called “Mathematical Theory of Communication”. So, it represents an absolute limit on the best possible lossless compression of any communication, under certain constraints, treating messages to be encoded as a sequence of independent and identically distributed (i.i.d.) random variables. The information that we receive from an observation is equal to the degree to which uncertainty is reduced. So,
I = H (before) – H (after)
Among their main properties, we have the Continuity. The measure H should be continuous, in the sense that changing the values of the probabilities by a verysmall amount, should only change the H value by a small amount.
Maximality: The measure H will be maximal, if all the outcomes are equally likely, i.e. the Uncertainty is highest when all the possible events are equiprobable,
Hn (p1, p2, …, pn) ≤ Hn (1/n, 1/n, …, 1/n)
And H will increase with the number of outcomes
Hn (1/n, 1/n, …, 1/n) < Hn+1 (1/n+1, 1/n+1, …, 1/n+1)
Additivity: The amount of entropy should be independent of how the process is considered, as being divided into parts. Such functional relationship characterizes the entropy of a system respect to their sub-systems. It demands that the entropy of every system can be identified, and then, computed from the entropies of theirsub-systems.
The notion of Intuitionistic Fuzzy Set (in acronym, IFS) was introduced by Atanassov (1983), and then developed by authors as Hung & Yang, among others. Recall that an Intuitionistic Set is an incompletely known set[3-5].
An IFS must represent the degrees of their membership and non-membership, with a degree of hesitancy. For this reason, they have widely used in many applied areas.
A very frequent measure of fuzziness is the entropy of Fuzzy Sets, which were first mentioned by Zadeh (1965). But recently two new definitions have been proposed by Szmidt&Kacprzyk (2001), and by Burillo&Bustince(1996). The first one is a non-probabilistic entropy measure, which departs on a geometric interpretation of an IFS. And by the second, it would be possible to measure the degree of intuitionism of an IFS.
Also we can to generalize from IFS to a Neutrosophic Set, abridgedly N-Set, concept due to Smarandache (1995). It is a set which generalizes[3-5] many existing classes of sets. In particular by the so-denoted, FS, and its first generalization, IFS.
Let U be a universe of discourse, and M a set included in U. An element x from U is noted, with respect to the set M, as x (T, I, F). And it belongs to M in the following way: it is T % in the set (membership degree), I % indeterminate (unknown if it is in the set), and F % not in the set (non-membership degree). Here T, I, F components are real standard/non- standard subsets, included in the non-standard unit interval,
]-0, 1+[representing truth, indeterminacy and falsity percentages, respectively.
It is possible to define a measure of Neutrosophic H, or N-Entropy, as the summation of the respective entropies of three subsets, T, I and F, by
HNS = HT + HI + HF
The maximum entropy principle (expressed asMEP,by acronym) is a postulate about a universal feature of any probability assignment on a given set of propositions: events, hypotheses, etc.
Let some testable information about a probability distribution function be given. Consider the set of all trial probability distributions that encode this information. Then, the probability distribution that maximizes the information entropy is the true probability distribution with respect to the testable information prescribed.
E. T. Jaynessuggested[28-32] that the Entropy, denoted by S in Statistical Mechanics, and H (Information Entropy), are basically the same thing. Consequently, Statistical mechanics should be seen just as a particular application of a general tool of Logical Inference and Information Theory.
Given testable information, the maximum entropy procedure consists of seeking the probability distribution which maximizes the information entropy, H, subject to the constraints of the information. This constrained optimization problem will be typically solved using the analytical method well-known as Lagrange Multipliers.
Entropy maximization (max H) with no testable Information takes place under a single constraint: the sum of the probabilities must be one. Under this constraint, the maximum entropy probability distribution is the well-known as U (uniform distribution). The MEP can, thus, be considered as a generalization of the classical Principle of Indifference (due to Laplace), also known as the Principle of Insufficient Reason.
We may also consider the Metric Entropy, also called Kolmogorov Entropy, or Kolmogorov-Sinai Entropy, in acronym K - S Entropy[35,36].
In a dynamical system[6-10], the metric entropy is equal to zero for non-chaotic motion, and is strictly greater than zero for chaotic motion.
In Thermodynamics, Prigogine entropy[45] is a very often used term referring to the splitting of Entropy into two variables, one being that which is "exchanged" with the surroundings, and the other being a result of “internal” processes. The Entropy of a System is an Extensive Property, i.e.
If the system consists of several parts, the total entropy is equal to the sum of the entropies of each part, and
The change in entropy can be split into two parts.

3. Entropy on Networks

Graph entropywill be a functional on a graph, G = (V, E), with P a probability distribution on its node (or vertex) set, V. It will be denoted by GE. It was a concept initially introduced by Janos Körner, as solution of a coding problem formulated on IT. Because its sub-additivity, it has become a useful tool in proving some lower bounds results in Computational Complexity Theory[1,2].
The search for exact additivity has produced certain interesting combinatorial structures. One of such results is the characterization of perfect graphs by the additivity of GE. Such GE-function is convex. It tends to +∞ on the boundary of the non-negative orthant of Rn. And monotonically to -∞ along rays from the origin. So, such minimum is always achieved and finite.
Recall that a Perfect Graph is one in which the chromatic number of every induced subgraphequalsthe size of the largest clique of that subgraph. Because in any graph the clique number provides a lower bound for the chromatic number.
We may distinguish now four of such structural models.
Thus, we can classify as Regular Networks, Random Net- works, Small-World Networks, and Scale-free Networks.
In the Regular Network, each node is connected to all other nodes. I.e. they are fully connected. Because of such a type of structure, they have the lowest path length (L), and the lowest diameter (D), being
L = D = 1
Also, they have the highest clustering coefficient. So, it holds C = 1. Furthermore, the highest possible number of edges
card (E) =[(n (n-1)) / 2]
As related to Random Graphs (RGs), we can say that each pair of nodes is connected with probability p.
They have a low average path length, according to
L ≈ (ln n) / (ln) ln n, for n >> 1
And it is because the total network may be covered in steps, from which
nexp L
Moreover, they possess a low clustering coefficient, when the graph is sparse,
C = ( / n) << 1such that the probability of each pair of neighbouring nodes to be connected is precisely equal to p.
The Small-World effect is observed on a network when it has a low average path length, i.e.
L << n, for n >> 1
Recallthat the now famous "six degrees of separation” is also known as the small-world phenomenon. This initially surprising hypothesis was firstly formulated by the writer Frygies Karinthy in 1929.Then, it was tested by Stanley Milgram in 1967. So, Small-World models share with Random Graphs some common features, such asPoisson or Binomial degree distribution, near to the Uniform distribution; network size: it does not grow; each node has approximately the same number of edges, i.e. it shows a homogeneous nature.Such models show the low average path length typical of Random Graphs,
Lln n, for n >> 1
And also such model gives the usual high clustering coefficient of Regular Lattices, givingC ≈ .75,for k >> 1. Therefore, the WS-models have a small-world structure, being well clustered.
The Random Graphs coincides on the small-world structure, but they are poorly clustered. This model (Watts- Strogatz) has a peak degree distribution of Poisson type. With reference to the above last model (Scale-Free Network), this appears when the degree distribution follows a Power- Law
P (k) expk(- γ)
Whereγis a constant whose value is typically on thedomain
2 <γ< 3
In such a case, there exist a small number of highly connected nodes, called Hubs, which are the tail of the distribution.
On the other hand, the greatest majority of the sets of nodes have very few connections, representing the head of such distribution.Such a model was introduced by Albert- LászloBarabási, and simultaneously by Réka Albert (1999). But there exist certain historical precedents, as George Udny Yule (1925), Herbert Simon (1955), and Derek de Solla Price (1976), among others.
Some of their principal features may be these: Non- homogeneous nature, in the sense that some (few) nodes have many edges from them, and the remaining nodes only have very few edges, or links; as related to the network size, it continuously grows; and regarding to the connectivity distribution, it obeys a Power-Law distribution.

4. Modularity

A very notable example of Scale-Free Network may be the World Wide Web (WWW). It will be described by acollection of sub-networks. In such case, the nodes correspond to their pages, whereas the edges correspond to their hyperlinks. Also related to the Web graph characteristics, we may notice the Scale Invariance, as a very important feature of such notable complex network.
Community structure (also called Modularity) is a very frequent characteristic in many real networks. Therefore, it has become a key problem in the study of networked systems. Giving out its deterministic definition will be a nontrivial problem for the complexity of networks.
A concept of modularity Q can be used as a valid measure for community structure.Some current models have proposed to capture the basic topological evolution of complex networks by the hypothesis that highly connected nodes increase their connectivity faster than their less connected peers, a phenomenon denoted as PA (preferential attachment). So, we will found a class of models that view networks as evolving dynamical systems, rather than static graphs.
Most evolving network models are based on two essential hypotheses, growth and preferential attachment. The growth suggests that networks continuously expand through the addition of new nodes and links between the nodes. While the preferential attachment states that the rate with which a node with k links acquires new links is a monotonically increasing function of k.
The modularity function, Q, provides a way to determine if a partition is valid to decipher the community structure in a network.Maximization of such modularity function, over all the possible partitions of a network, is in fact a highly effective method.
An important case in community detection is that some nodes may not belong to a single community, and then placing them into more than one group is more reasonable.
The set of such nodes can provide a “fuzzy” categorization, and hence, they may take a special role[18], such as signal transduction in biological networks.

5. Conclusions

Therefore, we conclude for the moment our analysis on Entropy and another interesting Fuzzy Measures; in particular, when they are applied on Complex Networks. In such general vision we have showed some new entropy measures.

References

[1]  Aczél, and Daróczy: On Measures of Information and Their Characterizations.Maths.In Sci. And Eng.,Vol 115. Academic Press, New York, 1975
[2]  Adler, Konheim, and Mc Andrew: “Topological Entropy”. Transactions of AMS, Vol. 114, No. 2, pp. 309-319
[3]  Atanassov: “Intuitionistic Fuzzy Sets”. Fuzzy Set and Systems (FSS), Vol. 20, pp. 87-96, 1986
[4]  Atanassov: Intuitionistic Fuzzy Sets: Theory and Applications.Physica-Verlag, Heidelberg, 1999
[5]  Atanassov: “New operators defined over Intuitionistic Fuzzy Sets”. FSS.Vol. 61, pp. 137-142, 1994
[6]  Burbea, and Rao: “Entropy Differential Metric, Distance and Divergence Measures in Probabilistic Spaces:A Unified Approach”.J. Multi. Analysis, 12, pp. 575-596, 1982
[7]  Burbea, and Rao: “On the Convexity of some divergence measures based on Entropy functions”. IEEE Transactions on Information Theory, IT-28, pp. 489-495, 1982
[8]  Burillo and Bustince: “Entropy on Intuitionistic Fuzzy Sets, and interval-valued fuzzy sets”. FSS, Vol. 78: 305-316, 1996
[9]  Chakrabarty: “Shannon Entropy: Axiomatic Characterization and Application”. Intern.Journal of Maths.and Math. Sci., Vol. 17, pp. 2847-2854, 2005
[10]  Dubois and Prade: Fundamentals of Fuzzy Sets. Series: The Handbooks of Fuzzy Sets. Vol.7. Springer-Kluwer, 2000
[11]  Dumitrescu, D.: “Entropy of a fuzzy process”. FSS.Vol. 55, pp. 169-177, 1993
[12]  Dumitrescu, D.: “Fuzzy measures and the entropy of fuzzy partitions”.J. Math. Anal.Appl. Vol. 175, pp. 359-373, 1993
[13]  Dumitrescu, D.: “Entropy of a fuzzy dynamical system”, FSS, Vol. 70, pp. 45-57, 1995
[14]  Dumitrescu, D.; Haloiu, E.; and Dumitrescu, A.: “Generators of fuzzy dynamical systems”. FSS, Vol. 113, pp. 447-452, 2000
[15]  Fan, and Xie: “Distance measure and induced fuzzy entropy”. FSS, Vol. 104, No.2, pp. 305-314, 1999
[16]  Fisher: Collected Papers of R. A. Fisher.In five Volumes. University of Adelaide, 1971-1974
[17]  Garmendia, L.: “The Evolution of the Concept of Fuzzy Measure”. Studies in Computational Intelligence, Vol. 5, pp. 186-200, 2005
[18]  Garrido, A., and Postolica, V.: Modern Optimization. Matrix-Rom Ed. Bucharest, 2011
[19]  Hartley: "A More Symmetrical Fourier Analysis Applied to Transmission Problems“. Proceedings of the IRE, Vol. 30, pp. 144–150, 1942
[20]  Hartley: "A New System of Logarithmic Units", Proceedingsof the IRE, Vol. 43, No. 1, 1955
[21]  Havrda, and Charvát: “Quantification method of classification processes”. Kybernetica, Vol. 3, pp. 30-35, 1967
[22]  Hung, and Yang: “Fuzzy Entropy on Intuitionistic Fuzzy Sets”.Int. J. of Intell.Systems.Vol. 21, pp. 443-451, 2006
[23]  Hung: “A note on entropy of intuitionistic fuzzy sets”. Int. J. of Uncertainty, Fuzziness and Knowledge-Based Systems.Vol. 11, Issue 5, pp. 627-633, 2003
[24]  Iosifescu, M.: Finite Markov Processes and their Applications.Wiley Series in Probability and Mathematical Statistics.Dover, 2007
[25]  Iosifescu, and Grigorescu:Dependence with Complete Connections and its Applications. Cambridge University Press, Cambridge, 1990
[26]  Iosifescu, and Theodorescu: “On the entropy of chains with complete connections”.C. Ac. Romane 11, pp. 821–824, 1961
[27]  Iosifescu, M.: “Sampling Entropy for Random Homogeneous Systems with Complete Connections”.Ann. Math. Stat., Vol. 36, pp. 1433-1436, 1965
[28]  Jaynes, E. T.: “Information Theory and Statistical Mechanics”, in Physical Review.Vol. 106, No. 4, pp. 620-630, 1957
[29]  Jaynes, E. T.: “Prior Probabilities”, in IEEE Transactions on Systems Science and Cybernetics, SSC-4, p. 227, 1968
[30]  Jaynes, E. T.: “Information Theory and StatisticalMechanics”, in Physical Review.Vol. 106, No. 4, pp. 620-630, 1957
[31]  Jaynes, E. T.: “Monkeys, Kangaroos and N”, in Maximum Entropy, and Bayesian Methods in AppliedStatistics.Cambridge University Press, 1986
[32]  Jaynes, E. T.: Probability Theory: The Logic of Science.CUP, Cambridge University Press, 2003
[33]  Jeffreys: Theory of probability.Oxford University Press, 1948
[34]  Kerridge: “Inaccuracy and Inference”. J. Royal Statistical Society, Ser. B, 23, pp. 184-194, 1961
[35]  Kolmogorov, A.N.: “New Metric Invariant of Transitive Dynamical Systems and Endomorphisms of Lebesgue Spaces”. Doklady of Russian Academy of Sciences (RAS), Vol. 119, No. 5, pp. 861-864, 1958
[36]  Kolmogorov, A.N.: “Entropy per unit time as a metric invariant of automorphism”. Dokladyof Russian Academy of Sciences (RAS). Vol. 119, No. 5, pp. 861-864, 1958
[37]  Kullback, S.: Information Theory and Statistics, Wiley, 1959
[38]  Kullback, S., and Leibler, R.: “On Information and Sufficiency”. Ann. Math. Statist., 22, pp. 79-86, 1951
[39]  Lambert: “Entropy is Simple, Qualitative Journal of Chemical Ed., Vol. 79, pp. 1241-1246, 2002[it is also disposable online].
[40]  Luca, and Termini: “A definition of non-probabilistic entropy in the setting of Fuzzy Set theory”.Paper at Information and Control, Vol. 20, pp. 301-312, 1972
[41]  Marcus, S.: Words and Languages Everywhere. Polimetrica. Bucharest, 2007
[42]  Nyquist: "Certain factors affecting telegraph speed“. Bell System Technical Journal, Vol. 3, pp. 324-346, 1924
[43]  Preda, V., and Balcau, C.: “Homogeneous stationary multiple Markov Chains with maximum I-T Entropy”. Rev. RoumaineMaths. Pures Appl., Vol. 53, 1, pp. 55-61, 2008
[44]  Preda, V., and Balcau, C.: Entropy Optimization with Applications. Editura Academiei Romane. Bucharest, 2011
[45]  Prigogine: Etude Thermodynamique des PhenomenesIrreversibles.Dunod, Paris, 1947
[46]  Rényi, A.: “On measures of information and entropy”. Proc. of the 4th Berkeley Symposium on Mathematics, Statistics, and Probability, pp. 547-561, 1961.
[47]  Shannon, C.E.: “A Mathematical Theory of Communication”, Bell System Techn. J., pp. 379-423 and 623-656, 1948
[48]  Sibson: “Information Ratius”.Z. Wahrsverw. Geb., Vol. 14, pp. 149-160, 1969
[49]  Simonyi, G.: “Graph Entropy: A Survey”. DIMACS Series in Discrete Mathematics and Theor.Computer Science
[50]  Singh: “On Sharma-Taneja´s entropy of Type (α,β)”, at Kybernetika, vol. 9, Nr. 6, download
[51]  Smarandache F.:A Unifying Field inLogics. Neutrosophy: Neutrosophic Probability, Set, and Logic,AmericanResearch Press, Rehoboth, 1999
[52]  Szmidt, and Kacprzyk: “Entropy for Intuitionistic Fuzzy Sets”.FSS, Vol. 118, pp. 467-477, 2001
[53]  Szmidt, and Kacprzyk: “Distances between Intuitionistic Fuzzy Sets”. FSS, Vol. 118, pp. 467-477, 2001
[54]  Taneja:“On axiomatic characterization of entropy of type (α,β)”. Appl. of Maths.Vol. 22, Issue 6, pp. 409-417, 1977
[55]  Taneja:Generalized Information Measures and Their Applications.On-line book, 2001
[56]  Titchener, Nicolescu, Staiger, Gulliver, and Speidel: “Deterministic Complexity and Entropy”. FundamentaInformaticae, Vol. 64, Issue 1-4, Contagious Creativity. In Honor of the 80th Birthday of Prof. Solomon Marcus, 2004
[57]  Titchener: “A measure of information”. Proc. Data Compression Conference, pp. 353-362, Snowbird, UT, 2000
[58]  Titchener: “A deterministic theory of Complexity, Information and Entropy”. IEEE Information Theory Workshop. San Diego, CA, 1998
[59]  Tribus, and Irvine: “Energy and Information”.Scientific American 225, No. 3: 179-188, 1971
[60]  Tsallis, Mendes, and Plastino: “The role of constraints within generalized non-extensive statistics”. Physica, Vol. 261, pp. 534-554, 1998
[61]  Volkenstein: Entropy and Information. Series: Progress in Mathematical Physics, Vol. 57, BirkhäuserVerlag, 2009
[62]  Wang, and Klir: Fuzzy Measure Theory. Plenum Press, New York, 1992
[63]  Wang, and Klir: Generalized Measure Theory. SpringerVerlag, Berlin-New York, 2008
[64]  Yager: “On the measure of fuzziness and negation”. Int. J. Gral.Systems, Vol.5, pp. 189-200, 1979
[65]  You, and Gao: “Maximum entropy membership functions for discrete fuzzy variables”. Information Sciences, 179, Issue No. 14, pp. 2353-2361, 2009
[66]  Yu, and Hu: “Entropies of fuzzy indiscernibility relation, and their operations”.Int. J. of Uncertainty, Fuzziness and Knowledge-Based Systems.Vol. 12, Issue 5, pp. 575-589, 2005