Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Assessing the Robustness of Thermoeconomic Diagnosis of Fouled Evaporators: Sensitivity Analysis of the Exergetic Performance of Direct Expansion Coils
Next Article in Special Issue
Cauchy Principal Value Contour Integral with Applications
Previous Article in Journal
Hierarchical Decomposition Thermodynamic Approach for the Study of Solar Absorption Refrigerator Performance
Previous Article in Special Issue
Wavelet Entropy-Based Traction Inverter Open Switch Fault Diagnosis in High-Speed Railways
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Entropy and Fractal Antennas

by
Emanuel Guariglia
1,2
1
Department of Physics “E. R. Caianiello”, University of Salerno, Via Giovanni Paolo II, 84084 Fisciano, Italy
2
Division of Applied Mathematics, Mälardalen University, 721 23 Västerås, Sweden
Entropy 2016, 18(3), 84; https://doi.org/10.3390/e18030084
Submission received: 20 December 2015 / Revised: 29 February 2016 / Accepted: 29 February 2016 / Published: 4 March 2016
(This article belongs to the Special Issue Wavelets, Fractals and Information Theory I)
Figure 1
<p>Rényi and Shannon entropies for a binomial distribution with N = 20: they converge for <math display="inline"> <mrow> <mi>α</mi> <mo>→</mo> <mn>1</mn> </mrow> </math>, in accord with Equation (<a href="#FD3-entropy-18-00084" class="html-disp-formula">3</a>). The computation of both entropies was done for <math display="inline"> <mrow> <mi>b</mi> <mo>=</mo> <mn>2</mn> </mrow> </math>.</p> ">
Figure 2
<p>Kolmogorov entropy for 1D-regular, chaotic-deterministic and random systems. The attractor is the classical Lorentz attractor [<a href="#B27-entropy-18-00084" class="html-bibr">27</a>] with <math display="inline"> <mrow> <mi>σ</mi> <mo>=</mo> <mn>10</mn> </mrow> </math>, <math display="inline"> <mrow> <mi>R</mi> <mo>=</mo> <mn>28</mn> </mrow> </math>, <math display="inline"> <mrow> <mi>b</mi> <mo>=</mo> <mn>8</mn> <mo>/</mo> <mn>3</mn> </mrow> </math> and initial values <math display="inline"> <mrow> <mi>x</mi> <mo>(</mo> <mn>0</mn> <mo>)</mo> <mo>=</mo> <mi>z</mi> <mo>(</mo> <mn>0</mn> <mo>)</mo> <mo>=</mo> <mn>0</mn> </mrow> </math>, <math display="inline"> <mrow> <mi>y</mi> <mo>(</mo> <mn>0</mn> <mo>)</mo> <mo>=</mo> <mn>1</mn> </mrow> </math>, while the random motion is given by a 2D-random walk from <math display="inline"> <mrow> <mo>-</mo> <mn>1</mn> <mo>-</mo> <mi>j</mi> </mrow> </math> to <math display="inline"> <mrow> <mn>1</mn> <mo>+</mo> <mi>j</mi> </mrow> </math> of 500 elements in which <span class="html-italic">j</span> is the imaginary unit.</p> ">
Figure 3
<p>Graph of the <math display="inline"> <mrow> <msub> <mi mathvariant="script">M</mi> <mi>s</mi> </msub> <mrow> <mo>(</mo> <mi>A</mi> <mo>)</mo> </mrow> </mrow> </math>, where <span class="html-italic">A</span> is a bounded subset of the Euclidean metric space <math display="inline"> <msup> <mrow> <mi>R</mi> </mrow> <mi>n</mi> </msup> </math>. It takes only two possible values, and the Hausdorff–Besicovitch dimension of <span class="html-italic">A</span> is given by the value of <span class="html-italic">s</span> in which there is the jump from <span class="html-italic">∞</span> to zero.</p> ">
Figure 4
<p>Here, the first steps of the box-counting procedure about England’s coastline are represented.</p> ">
Figure 5
<p>Here, the von Kock curve (on the left) and the middle third Cantor set (on the right) are shown: <math display="inline"> <msub> <mi>A</mi> <mn>0</mn> </msub> </math> is the initiator of length equal to one in both cases; in the generator <math display="inline"> <msub> <mi>A</mi> <mn>1</mn> </msub> </math> for the von Kock curve, the middle third of the unit interval is replaced by the other two sides of an equilateral triangle, while that of the middle third Cantor set is obtained removing the middle third of the interval.</p> ">
Figure 6
<p>Archimedean spiral antenna (on the left) and commercial log-periodic dipole antenna of 16 elements (on the right) [<a href="#B7-entropy-18-00084" class="html-bibr">7</a>].</p> ">
Figure 7
<p>A Sierpinski triangle (on the left) and a Hilbert curve (on the right) are shown: as in <a href="#entropy-18-00084-f005" class="html-fig">Figure 5</a>, <math display="inline"> <msub> <mi>A</mi> <mn>0</mn> </msub> </math> is the initiator, and <math display="inline"> <msub> <mi>A</mi> <mn>1</mn> </msub> </math> is the generator. The Sierpinski triangle is constructed using the iterated function system (IFS), while the other construction is that of David Hilbert. The two relative antennas are shown below with their feed points.</p> ">
Figure 8
<p>Examples of non-fractal antennas that offer similar performance over their fractal counterparts. Three different non-fractal antennas are presented above: they outperform their fractal counterparts, while the current distribution on the Sierpinski gasket antenna at the first three resonance frequencies is shown in the middle of the page. The modified Parany antenna (starting from the classical Sierpinski gasket antenna) is represented below.</p> ">
Figure 9
<p>The Rényi entropy <math display="inline"> <msub> <mi>H</mi> <mi>α</mi> </msub> </math> of a Sierpinski gasket (<a href="#entropy-18-00084-f007" class="html-fig">Figure 7</a>) with <math display="inline"> <mrow> <mi>α</mi> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </math>: the plot shows us that Hartley entropy <math display="inline"> <msub> <mi>H</mi> <mn>0</mn> </msub> </math> is an upper to both Shannon entropy <math display="inline"> <msub> <mi>H</mi> <mn>1</mn> </msub> </math> and collision entropy <math display="inline"> <msub> <mi>H</mi> <mn>2</mn> </msub> </math>. The main limit of this procedure is clearly the precision of the triangulation.</p> ">
Versions Notes

Abstract

:
The entropies of Shannon, Rényi and Kolmogorov are analyzed and compared together with their main properties. The entropy of some particular antennas with a pre-fractal shape, also called fractal antennas, is studied. In particular, their entropy is linked with the fractal geometrical shape and the physical performance.

1. Introduction

Classically, the concept of entropy arises from the analysis of physical problems in statistical physics and thermodynamics. Since the beginning, it was a measure of uncertainty in a physical system, and based on this, C.E. Shannon [1] proposed to extend this concept for the analysis of complexity in signals, thus giving rise to the emerging information theory [2]. Several year later than Shannon, A. Rényi showed that a valid measure of entropy has to be defined in accordance with a measure of diversity [3]. A step forward in this direction was given by A. Kolmogorov (1958), who used the concept of entropy to define a fundamental measure for chaotic evolution and loss of information in the course of time [4].
Indeed, this entropy is an extension of a known concept in information entropy (for time-dependent processes), which is used to characterize dynamical systems between regular, chaotic and purely random evolution.
Chaotic motions and, in particular, attractors can be also described by iterative maps, which belong to the fundamental methods of fractal geometry.
Chaos, complexity and fractals have many common features, and recently, they have attracted the interest of scholars for their application in science and engineering.
Fractal sets are abstract objects that cannot be physically implemented. However, some related geometries known as pre-fractals have been shown to be very useful in engineering and applied science [5,6]. In particular, some fractal models have been used to design some fractal antenna with very special properties: about one-tenth of a wavelength (p. 231, [7]) and a pre-fractal geometrical configuration.
An antenna is a complex device, characterized by different parameters (resonant frequency, gain, directivity, radiation pattern, etcetera), which define the performance of the radiator. The chaoticity of the fractal antenna will be studied in the following by an entropy measure based on the computation of the fractal dimension, according to the analysis of a radiating structure given by [8,9].
Since the Rényi entropy H α and generalized fractal dimension D α are connected by a well-known relation (see Equation (25)), in this paper, the quantity D α was used to compute the Rényi entropy H α of a pre-fractal structure and to describe the electromagnetic behavior of an antenna together with the corresponding performance.
The results of Best [10,11] show how antenna geometry alone (pre-fractal or otherwise) is not a significant factor to determine the performance of small antennas. Yet, this may be a good clue.
In the literature, there are only a few articles about how the self-similarity property of a pre-fractal radiator can influence its performance (see again [8,9]).
In order to investigate in this direction, the values of Rényi entropy H α through Equation (25) for the classical Sierpinski gasket were determined. The quantity D α was numerically estimated (see Paragraph 4.3).

2. Concept of Entropy

There are basically three definitions of entropy in this article. The Kolmogorov entropy K, which measures the chaoticity of a dynamical system (Chapter 6, [12]), can be estimated as the Rényi entropy H α . In information theory, the Shannon entropy is a special case of Rényi entropy for α = 1 .
Definition 1 
(Shannon entropy). The Shannon entropy [1,13] of a discrete-type RV X is defined as:
H ( X ) - i = 1 N p i log b p i
where N is the number of possible states and p i is the probability of the event { X = x i } , and it is assumed p i > 0 ; the most common used values are b = 2 and b = e .
This entropy may be defined in a simple way for a continuous-type RV, as well [13]. Yet, this is not the scope of this article. As is well known, Shannon entropy satisfies different properties, which will not be treated herein [14].
Moreover, it is possible to show that it represents the measure of uncertainty about a suitable partition [13].
A first generalization of this kind of entropy is the so-called Rényi entropy [3,15]: it represents one family of functionals to describe the uncertainty or randomness of a given system.
Definition 2 
(Rényi entropy). Let α be a positive real number. The Rényi entropy of order α is defined as [3]:
H α ( X ) 1 1 - α log b i = 1 N p i α α > 0 , α 1 H ( X ) α = 1
where X is discrete-type RV and p i > 0 is the probability of the event { X = x i } , and the most common used values are the same as the Shannon entropy.
If the events { X = x i } are equiprobable, then H α is maximal and H α ( X ) = log b N for every α > 0 , i.e., the so-called Hartley entropy H 0 = log b i = 1 N p i 0 = log b N [16]. It is clear that they do not depend on the probability, but only on the number of events with non-zero probability.
In order to understand the meaning of the last definition, it is necessary to observe that at α = 1 , the quantity:
1 1 - α log b i = 1 N p i α
generates the indeterminate form 0 0 . By L’Hôpital’s rule [17], it is easy to show that:
lim α 1 1 1 - α log b i = 1 N p i α = - i = 1 N p i log b p i
i.e., the Shannon entropy, so lim α 1 H α ( X ) = H 1 ( X ) = H ( X ) , as shown in Figure 1.
Therefore, the Rényi entropy may be considered as a generalization of the Shannon entropy. It can be shown that the Rényi entropies decrease as a function of α [3].
Let x ( t ) = [ x 1 ( t ) , . . . , x d ( t ) ] be the trajectory of a dynamical system on a strange attractor. Let the d-dimensional phase space be partitioned into boxes of size l d and sampled at the discrete time intervals τ. Let p i 0 , . . . , i n be the joint probability that the trajectory x ( t = 0 ) is in the box i 0 , x ( t = τ ) is in the box i 1 , ..., and x ( t = n τ ) is in the box i n [18,19]; for example:
p i 0 , . . . , i n = P { x ( t = 0 ) i 0 ; x ( t = τ ) i 1 ; . . . ; x ( t = n τ ) i n }
According to Equation (1), the quantity:
K n = - i 0 , . . . , i n p i 0 , . . . , i n log p i 0 , . . . , i n
gives the expected amount of information needed to locate the system on a special trajectory i 0 * , . . . , i n * , i.e., if it is known a priori that our system was in i 0 * , . . . , i n * , then K n + 1 - K n is the necessary information to predict in which box i n + 1 * this system will be included. Using the language of information theory, this means that K n + 1 - K n measures the loss of information for the system from n to n + 1 (Chapter 6, [12]). Therefore:
n = 0 N - 1 ( K n + 1 - K n ) = K N - K N - 1 + K N - 1 - K N - 2 + . . . + K 1 - K 0 = K N - K 0 = - i 0 , . . . , i N p i 0 , . . . , i N log p i 0 , . . . , i N + p i 0 log p i 0 = - i 0 , . . . , i N - 1 p i 0 , . . . , i N - 1 log p i 0 , . . . , i N - 1
The definition of this new kind of entropy will be provided at this point.
Definition 3 
(Kolmogorov entropy). The Kolmogorov entropy K is defined as [19]:
K lim τ 0 lim l 0 lim N 1 N τ n = 0 N - 1 ( K n + 1 - K n ) = - lim τ 0 lim l 0 lim N 1 N τ i 0 , . . . , i N - 1 p i 0 , . . . , i N - 1 log p i 0 , . . . , i N - 1
where l and τ have the same meaning as above.
From this definition, it can be seen immediately that it is nothing other than the average rate of loss of information. K is independent of the particular partition (thanks to the limit l 0 ). Figure 2 reveals how K represents a measure of chaos: indeed (Chapter 6, [12]):
K = for random systems C > 0 for chaotic-deterministic systems 0 for regular motion
where the definition of a random system can be found in Chapter 3 of [20]. By describing the chaos in a dynamic system, the Kolmogorov entropy is expected to be strongly connected with the Lyapunov exponent λ; see [21]. For more information about the theoretical aspects of entropy, its generalizations and entropy-like measures, which can be used to measure the complexity of a system, see [22,23,24,25,26].

3. Remarks on Fractal Geometry

A fractal is characterized by the property that each enlargement of this set reveals further details, so it has a structure that is too irregular to be described by a classic mathematical theory (even if a fractal can often be described recursively). Furthermore:
  • it is self-similar, i.e., each very small portion of it is exactly or approximately similar to itself (this property has to be understood in the statistical or approximated sense, because a random element can be introduced in the construction of the fractal);
  • it is a space-filling curve [28].

3.1. Hausdorff–Besicovitch and Box-Counting Dimensions

Among the different definitions of fractal dimensions in use, the Hausdorff–Besicovitch dimension is probably the most important, even if it is not usually used for experimental procedures to find the fractal dimensions of real objects.
Fractal dimensions are very important because they provide a measure of the degree to which new details are revealed at different scales. For example, the fractal dimension of the coastline of Great Britain is about 1 . 2 (Chapter 2, [29]). In order to define the Hausdorff–Besicovitch dimension, some remarks on fractal geometry are given [5,6].
Theorem 4 
(See Figure 3). If n Z + and A is a bounded subset of Euclidean metric space R n , then there exists a unique number D H [ 0 , n ] , such that:
M s ( A ) = if s < D H and s [ 0 , ) 0 if s > D H and s [ 0 , )
Proof. 
See Chapter 3 in [6]. ☐
Definition 5 
(Hausdorff–Besicovitch dimension). Under the hypotheses of Theorem (4), the correspondent real number D H present in Equation (8) is called the Hausdorff–Besicovitch dimension of A, and it is generally indicated with D H ( A ) .
From this last definition, it follows that the Hausdorff–Besicovitch dimension of a bounded subset A R n is a non-negative number D H , such that:
M s ( A ) = if 0 s < D H ( A ) 0 if s > D H ( A )
Therefore, the Hausdorff measure of A, i.e., M s ( A ) , might be equal to zero, infinity or such that 0 < M s ( A ) < . In Figure 3, the plot of M s ( A ) is presented as a function of s, which shows us that D H ( · ) is the critical value of the variable s in the jump of M s ( A ) from to zero.
At this point, the definition of the fractal set can be provided. It is to be recalled that D T ( A ) D H ( A ) , where D T ( A ) represents the topological dimension of the bounded subset A of R n (p. 3, [30]).
Definition 6 
(Fractal set). A bounded subset A R n is fractal (in the sense of Mandelbrot) if it holds that D T ( A ) < D H ( A ) , where the difference D H ( A ) - D T ( A ) is called the fractal degree of A.
The Hausdorff–Besicovitch dimension is not particularly useful in engineering or applied sciences, because its calculation is not very easy, so another definition of the fractal dimension more suitable to compute the fractal dimension for problems of mathematical modeling was introduced [6,31].
Definition 7 
(Box-counting dimension). Let ( X , d ) be a metric space and A H ( X ) , where H ( X ) denotes the space of non-empty compact subsets of X. Let N δ ( A ) , δ > 0 be the smallest number of closed balls of radius δ needed to cover A. The lower and upper box-counting dimensions of A, denoted D U B ( A ) and D L B ( A ) , respectively, are defined as:
D U B ( A ) lim δ 0 sup log N δ ( A ) log ( 1 / δ )
D L B ( A ) lim δ 0 inf log N δ ( A ) log ( 1 / δ )
When D U B ( A ) = D L B ( A ) , the following limit exists and is called the box-counting dimension of A, denoted D B ( A ) :
D B ( A ) lim δ 0 log N δ ( A ) log ( 1 / δ ) = - lim δ 0 log N δ ( A ) log δ
The box-counting dimension of an object does not exactly have to be equal to the Hausdorff–Besicovitch dimension, even though they can be really close at times. This new definition of the fractal dimension is given by the minimum number of objects needed to cover the fractal set.
In Figure 4, it is shown how the box-counting dimension works to compute the length of England’s coastline: looking at the first iterations, the meaning of the adjective box-counting is clear.
In general, the limit in Equation (10) might not exist; hence, the box-counting and Hausdorff–Besicovitch dimensions are linked by the following relation: D H ( A ) D L B ( A ) D U B ( A ) (Chapter 3, [6]).

3.2. Iterated Function System and Pre-Fractals

It is to be recalled that a contraction on a metric space ( X , d ) is a transformation f : X X , such that:
d f ( x ) , f ( y ) s d ( x , y ) , x , y X
where the number s, called the contractivity factor for f, belongs to ( 0 , 1 ) .
The famous contraction mapping theorem states that every contraction f on a complete metric space ( X , d ) has exactly one fixed point x f X , and the sequence of iterations f n ( x ) n 0 converges to x f , x X (pp. 76–77, [5]).
Clearly any contraction is continuous. If the equality holds in Equation (11), f is called a contracting similarity, because it transforms sets into geometrically similar sets.
It is now time to give the definition of an important procedure concerning fractals.
Definition 8 
(Iterated function system). The iterated function system (IFS) is a couple { X , F } , where F : H ( X ) H ( X ) is defined through a finite family of contractions S i i = 1 m on the complete metric space ( X , d ) , with m 2 , and H ( X ) denotes again the space of non-empty compact subsets of X. Moreover, the set A H ( X ) is called the attractor (or sometimes invariant set) for the IFS if:
A = F ( A ) = i = 1 m S i ( A )
Technically speaking, the operator F given by Equation (12) is called the Hutchinson operator associated with the IFS { S 1 , S 2 ,..., S m } [5,32]. From the definition above, it is clear that the attractor for the IFS is also its unique fixed point. This is the fundamental property of an IFS, because this attractor is often a fractal. An IFS has a unique (non-empty compact) attractor (Chapter 9, [6]), but its introduction brings with it two main problems: the first one shows the way to represent a given set as the attractor of some IFS, while the second is to reconstruct the IFS starting from its attractor (p. 126, [6]).
Both of these two problems can often be solved by inspection, especially if F has a self-similar structure (see Figure 5).
For the majority of the fractals suitable for an application in antenna theory, the thesis of the Moran–Hutchinson theorem (pp. 130–132, [6]) holds true, so:
j = 1 m c j s = 1
where A is the attractor of the IFS with contraction factors c 1 , ..., c m and s = D H ( A ) = D B ( A ) .
This theorem provides us the possibility to compute the fractal dimension of many self-similarity fractals. Indeed, let us consider the von Koch curve and the middle third Cantor set (see Figure 5): for the first one, it is:
4 1 3 s = 1 s k o c h = log 4 log 3 1 . 26186
while for the other set, we get:
2 1 3 s = 1 s c a n t = log 2 log 3 0 . 63093
IFS can be applied to all self-similarity structures, especially for the simulation of real objects with fractal properties, like fractal antennas.
It is well known that fractals are only mathematical abstractions (because it is impossible to iterate indefinitely in the real word). In addition, numerical simulations show how the fractal modeling in antenna theory provides substantial advantages within a certain value of the iteration m * (typically, for fractal antennas, it is not greater than six). Beyond this value, the benefits are negligible. It is clear that all self-similar structures in nature are nothing other than fractals arrested at a prefixed iteration, i.e., pre-fractals (geometrical objects characterized by a finite number of fractal iterations).

4. Fractal Antennas

In order to minimize the antenna size holding a high radiation efficiency, a fractal approach to model its geometrical configuration can be considered.
The two fundamental properties of a fractal (i.e., self-similarity and space-filling) allow fractal antennas to have an efficient miniaturization and multiband characteristics.
The well-known log-periodic antennas, introduced by DuHamel and Isbell around the 1950s and closely paralleling the independent-frequency concept [7], might be considered the first fractal antenna of history. Another example of a self-similar antenna discovered in the same period is the spiral antenna (see Figure 6).
However, their true origin may be traced back to 1988, when Nathan L. Cohen, a Boston University radio astronomer, published a paper about this new type of antenna [33].
Fractal antennas have not only a large effective length, but also a simple matching circuit, thanks to the contours of their geometrical shape, which is able to generate a capacity or an inductance. For instance, a quarter-wavelength monopole may be transformed into a smaller antenna using the von Koch curve (see Figure 5).
A big part of the research on fractal antennas has been done by Fractal Antenna Systems Inc., an American company founded by Cohen.
Carles Puente Baliarda (Polytechnic University of Catalonia) was the first to treat these antennas as multiband antennas. In 1998, he won the award of “innovative IT products with evident market potential” due to his pioneering research in fractal antennas (for a total of € 200,000), while he and his company (Fractus S.A.) were the finalists for the European Inventor Award 2014, showing the great potentials of these antennas.
In 2011, 9 . 7 billion fractal-based antenna units were supplied worldwide (a report by BCC Research).

4.1. Sierpinski Gasket and Hilbert Antenna

The Sierpinski triangle T can be constructed from an equilateral triangle by the removal of inverted equilateral triangles (see Figure 7). It is a fractal and attractive fixed set. Considering Figure 7, it is:
j = 1 3 c j s = 3 1 2 s = 1
since all of the contraction factors c 1 , c 2 , c 3 are equal to 1 2 . Therefore:
s = D H ( T ) = D B ( T ) = log 3 log 2 1 . 58496
This fractal may be also generated by an IFS (Chapter 9, [6]).
There exist different versions of the Sierpinski triangle. The shape can be modified in many ways, and they are often used in engineering and applied sciences.
The Sierpinski (gasket) antenna belongs to the class of multiband fractal antennas based on the Sierpinski triangle. The classical Sierpinski dipole is shown in Figure 7. It is probably a fractal antenna with more applications, from wireless communication systems (GSM, UMTS and WLAN) through RF MEMS (radio frequency microelectromechanical system) to get to space probe design and ANN (artificial neural network) theory [34,35].
The famous Hilbert curve is a continuous fractal space-filling curve, i.e., it fills the plane without leaving any gaps. Hilbert introduced it as a modification of the Peano curve [28].
There are important differences between these two curves. Indeed, it is not possible to construct the Hilbert curve H through the IFS (while for the other one, this procedure is applicable). The reason is that the steps in the Hilbert curve’s construction are not self-similar, i.e., they are not divisible in a number of parts similar to the initial figure.
The original construction of the Hilbert curve is extraordinarily elegant: it starts with a square A 0 , while in the first step (see Figure 7), the curve A 1 connects the centers of the quadrants by three line segments (having a size of one). In the second step, four copies (reduced by 1 / 2 ) of this initial stage are made and placed into the quarters (see again Figure 7). In this way, the first copy is clockwise rotated and the last one counter-clockwise rotated by 90 degrees. After this, the start and end points of these four curves are connected using three line segments (of a size of 1 / 2 ), and we call the resulting curve A 2 .
In the third iteration, the scaling is done again by 1 / 2 , and four copies are placed into the quadrants of the square (as in the first step). They are again connected by three line segments (of a size of 1 / 4 ) obtaining A 3 , and so on.
In Figure 7, it can be noticed that each successive stage consists of four copies of the previous one, connected with additional line segments. Therefore, the curve is scaled down by the ratio 1 2 , and four copies are made; so:
j = 1 4 c j s = 4 1 2 s = 1
hence:
1 2 s = 1 4 s = D H ( H ) = D B ( H ) = 2
Naturally, the topological dimension of H is one, since it consists only of line segments. Therefore, the Hilbert curve is a fractal for all intents and purposes.
An alternative procedure to IFS is that of the so-called L-systems [36].
In Figure 7, a Hilbert dipole is also shown, where the feed source point is placed at the point of symmetry for these two pre-fractals.
The Hilbert antenna is especially used in spatial communications, like RF MEMS design [37] and, generally speaking, in each (telecommunication) system where the space available for the antenna is limited [38].

4.2. The Results of Best and HRC Conditions

A fractal approach is not the only way to miniaturize an antenna; indeed, there exist few particular non-fractal modifications of the classical von Koch dipole that could have the same performance [10].
In addition, it is clear that fractal geometry does not uniquely translate the electromagnetic behavior of the antenna. The geometrical configuration alone (fractal or non-fractal) could not be the only significant factor that determines the resonant behavior of wire antennas: indeed, a fractal configuration does not represent alone a guarantee of the highest antenna efficiency [11].
The same applies to the loop antennas. It is well known that the main advantage of the fractal loop antennas is that they have a high radiation resistance on a “small” physical area. In Figure 8 (top side), three examples of non-fractal antennas are shown. They offer similar or, in some cases, improved performance over their fractal-antenna counterparts, like the Minkowski antenna. The reason is that the radiation resistance of an electrically-small loop, given by [10]:
R r 31 , 200 A r e a l o o p λ 2 2 Ω
where λ is the working wavelength, is generally not valid for a loop antenna with complex geometry. However, there are few small non-fractal loop antennas with similar or better performance than their fractal counterparts (see Figure 8, top side).
In order to investigate the significance of self-similarity in determining the multiband behavior of the fractal antennas, Steven R. Best has presented a comparison of the multiband behavior of the Sierpinski gasket and several modified gaskets where the major portions of the self-similar structure were modified or eliminated [11].
His numerical simulations reveal that many of the self-similar fractal gap structures can be eliminated from the Sierpinski gasket, without modifying its multiband behavior.
Best showed how the total self-similar fractal gap structure is not the primary factor that determines the multiband behavior, because the Sierpinski gasket and modified Parany gasket antenna have the same behavior [11]. Therefore, for all of the Sierpinski-based antennas, the multiband behavior depends on the small isosceles trapezia located in the center of the modified Parany gasket antenna, as shown in Figure 8 (bottom side).
It would seem that some non-fractal geometries could be a good substitute for their fractal counterparts, but this is manifestly untrue.
Indeed, the results obtained by Best represent only a very few special cases, and these antennas do not belong to a family of radiators. Furthermore, the so-called HCR conditions can be considered [39] (they provide a necessary and sufficient condition for all frequency independent antennas). This criterion reveals that an antenna satisfies this property if and only if the radiating structure is both self-similar and origin symmetric about a point. It is clear that some non-fractal radiators might belong to this second one, potentially giving them the same performance of a fractal antenna that is non-symmetric about a point.

4.3. The Entropy of a Fractal Antenna

In addition to the box-counting dimension, another convenient way to estimate the fractal dimension is the so-called generalized fractal dimension (or Rényi dimension) D α , given by [40]:
D α ( X ) 1 α - 1 lim δ 0 log b i = 1 N p i α log b δ α > 0 , α 1 lim δ 0 i = 1 N p i log b p i log b δ α = 1
In this definition, N = N ( δ ) is the total number of boxes with p i > 0 , where also here, the most commonly-used values are b = 2 and b = e .
Considering the definition of Rényi entropy Equation (7), it is clear that:
D α = - lim δ 0 H α ( δ ) log b δ
As α 0 ,
D 0 = lim δ 0 log b i = 1 N 1 log b 1 δ = lim δ 0 log b N log b 1 δ
which is nothing but the fractal dimension. The numerator of the last equation is simply the Hartley entropy. It can be shown similarly, as for the definition of Rényi entropy, that
lim α 1 D α ( X ) = D 1 ( X )
Therefore, the dimension D α is a generalization of D 1 , which is called the information dimension. Indeed, D 1 characterizes the information required for the determination of the point location in some cell i.
According to Equation (21), it also follows that:
D 2 = lim δ 0 log b i = 1 N p i 2 log b δ
This quantity is called the correlation dimension, because it is very useful to detect chaotic behavior.
Taking still into account Equation (21), D α is clearly a nonincreasing function of α, i.e., D α D α at α > α : in particular, D 0 D 1 D 2 .
Therefore, the generalized fractal dimension D α provides a direct measurement of the fractal properties of an object. Several values of the momentum order α correspond to well-known generalized dimensions.
Equation (22) cannot be applied practically, and it is only possible to get an approximation fixing a small value of δ, but strictly greater than zero. Therefore, in applied sciences and engineering, Equation (22) becomes:
H α = - D α log b δ = D α log b ϵ
where H α = H α ( δ ) .
This equation shows us that the entropy of a region of size δ is a function of the box-counting fractal dimension ϵ = 1 / δ : the entropies of analyzed regions (with size δ) can be calculated from the three spatial dimensions through Equation (25) [16].
Right now, the Rényi entropy has to be computed for the geometric configuration of each fractal antenna. It is easy to create an algorithm for its computation using Equation (25). This procedure consists of the classical algorithm for numerical estimating D α of affine RIFS-invariant measures; see [41,42]. The Rényi entropy will be computed through Equation (25), considering the logarithm of the cell size (see Figure 9 below).
With this procedure, it is possible to compute the entropy for a fractal radiator, but it must be completely modified for each class of fractal antennas.
However, the general definition of entropy for a small antenna, in order to better understand how the chaoticness of the structure may affect its performance, remains an open problem.

5. Conclusions

The main fractals used in antenna theory have been examined, and the explanation as to how the Rényi entropy can be computed through the generalized fractal dimension of a set has been provided. Therefore, the entropy of a fractal antenna can be calculated accordingly, in which the input data are essentially the complexity of the shape, as described in this article for the Sierpinski gasket antenna. The numerical estimation of the generalized fractal dimension D α (see [41]) is based both on an integration technique over the fractal measures and a triangulation method. It yields good results for the range α 0 in an efficient and robust manner. Clearly, the computation of D α requires the aid of the box-counting algorithm. Furthermore, the developed code is general enough to be used on arbitrary geometrical shapes; hence, this algorithm may be applied both for other fractal antennas and for quantum dynamical systems in order to evaluate their chaoticness.

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANN
Artificial neural network
GSM
Global System for Mobile Communications
HCR
Hohlfeld–Cohen–Rumsey
IFS
Iterated function system
IT
Information technology
MSC
The current 2010 Mathematics Subject Classification
RF MEMS
Radio frequency microelectromechanical system
RIFS
Recurrent iterated function system
RV
Random variable
UMTS
Universal Mobile Telecommunications System
WLAN
Wireless local area network

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Sys. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
  2. Garrido, A. Classifying entropy measures. Symmetry 2011, 3, 487–502. [Google Scholar] [CrossRef]
  3. Rényi, A. On measures of information and entropy. In Proceedings of the Fourth Berkeley Symposium on Mathematics, Statistics and Probability, Berkeley, CA, USA, 20 June–30 July 1960; Volume 1, pp. 547–561.
  4. Kolmogorov, A.N. A new metric invariant of transient dynamical systems and automorphisms in Lebesgue spaces. Dokl. Akad. Nauk SSSR (NS) 1958, 119, 861–864. [Google Scholar]
  5. Barnsley, M.F. Fractals Everywhere: New edition; Cambridge University Press: San Diego, CA, USA, 2012. [Google Scholar]
  6. Falconer, K.J. Fractal geometry: Mathematical Foundations and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  7. Balanis, C.A. Antenna Theory: Analysis and Design; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  8. Ram, R.J.; Sporer, R.; Blank, H.-S.; Maccarini, P.; Chang, H.-C.; York, R.A. Chaos in microwave antenna arrays. IEEE MTT-S Int. Microw. Symp. Dig. 1996, 3, 1875–1878. [Google Scholar]
  9. Valdivia, J.A. The Physics of High Altitude Lightning. Ph.D. Thesis, The University of Maryland, College Park, MD, USA, 1997; pp. 48–50. [Google Scholar]
  10. Best, S.R. A Discussion on the Significance of Geometry in Determining the Resonant Behavior of Fractal and Other Non-Euclidean Wire Antennas. IEEE Antennas Propag. Mag. 2003, 45, 9–28. [Google Scholar] [CrossRef]
  11. Best, S.R. Operating Band Comparison of the Perturbated Sierpinski and Modified Parany Gasket Antennas. IEEE Antennas Wirel. Propag. Lett. 2002, 1, 35–38. [Google Scholar] [CrossRef]
  12. Schuster, H.G.; Just, W. Deterministic Chaos: An Introduction; Wiley-VCH: Weinheim, Germany, 2005. [Google Scholar]
  13. Papoulis, A.; Pillai, S.U. Probability, Random Variables and Stochastic Processes; McGraw-Hill: New York, NY, USA, 2002; Chapter 14. [Google Scholar]
  14. Csiszár, I. Axiomatic characterization of information measures. Entropy 2008, 10, 261–273. [Google Scholar] [CrossRef]
  15. Takens, F.; Verbitski, E. Generalized entropies: Rényi and correlation integral approach. Nonlinearity 1998, 11, 771–782. [Google Scholar] [CrossRef]
  16. Zmenskal, O.; Dzik, P.; Vesely, M. Entropy of fractal systems. Comput. Math. Appl. 2013, 66, 135–146. [Google Scholar] [CrossRef]
  17. Marsden, J.E.; Hoffman, M.J. Elementary Classical Analysis; W. H. Freeman and Company: San Francisco, CA, USA, 1993. [Google Scholar]
  18. Amigó, J.M.; Keller, K.; Unakafova, V.A. On entropy, entropy-like quantities, and applications. Discret. Contin. Dyn. Syst. B 2015, 20, 3301–3343. [Google Scholar] [CrossRef]
  19. Zanette, D.H. Generalized Kolmogorov entropy in the dynamics of the multifractal generation. Physica A 1996, 223, 87–98. [Google Scholar] [CrossRef]
  20. Bhattacharya, R.; Majumdar, M. Random Dynamical Systems: Theory and Applications; Cambridge University Press: New York, NY, USA, 2007. [Google Scholar]
  21. Pesin, Y.B. Lyapunov characteristic exponents and smooth ergodic theory. Uspeki Mat. Nauk 1977, 32, 55–112. [Google Scholar] [CrossRef]
  22. Falniowski, F. On the connections of generalized entropies with Shannon and Kolmogorov-Sinai entropies. Entropy 2014, 11, 3732–3753. [Google Scholar] [CrossRef]
  23. Crutchfield, J.P.; Feldman, D.P. Regularities unseen, randomness observed: Levels of entropy convergence. Chaos 2003, 13, 25–54. [Google Scholar] [CrossRef] [PubMed]
  24. Ferenczi, S. Measure-theoretic complexity of ergodic systems. Isr. J. Math. 1997, 100, 189–207. [Google Scholar] [CrossRef]
  25. Blanchard, F.; Host, B.; Maass, A. Topological complexity. Ergod. Theory Dyn. Syst. 2000, 20, 641–662. [Google Scholar] [CrossRef]
  26. Galatolo, S. Global and local complexity in weakly chaotic systems. Discret. Contin. Dyn. Syst. 2003, 9, 1607–1624. [Google Scholar] [CrossRef]
  27. Farmer, D.; Crutchfield, J.; Froehling, H.; Packard, N.; Shaw, R. Power spectra and mixing properties of strange attractors. Ann. N. Y. Acad. Sci. 1980, 375, 453–472. [Google Scholar] [CrossRef]
  28. Tricot, C. Curves and Fractal Dimension; Springer: New York, NY, USA, 1995. [Google Scholar]
  29. Mandelbrot, B.B. The Fractal Geometry of Nature; W. H. Freeman and Company: New York, NY, USA, 1982. [Google Scholar]
  30. Hata, M.; Kigami, J.; Yamaguti, M. Mathematics of Fractals; American Mathematical Society: Providence, RI, USA, 1997. [Google Scholar]
  31. Addison, P.S. Fractal and Chaos: An Illustrated Course; Institute of Physics Publishing: London, UK, 1997. [Google Scholar]
  32. Hutchinson, J.E. Fractals and self similarity. Indiana Univ. Math. J. 1981, 30, 713–747. [Google Scholar] [CrossRef]
  33. Cohen, N.L. Fractal’s New Era in Military Antennas Design. Available online: http://defenseelectronicsmag.com/site-files/defenseelectronicsmag.com/files/archive/rfdesign.com/mag/508RFDSF1.pdf (accessed on 4 March 2016).
  34. Hwang, K.C. A Modified Sierpinski Fractal Antenna for Multiband Application. IEEE Antennas Wirel. Propag. Lett. 2007, 6, 357–360. [Google Scholar] [CrossRef]
  35. Puente-Baliarda, C.; Romeu, J.; Pous, R.; Cardama, A. On the behavior of the Sierpinski multiband fractal antenna. IEEE Antennas Propag. 1998, 46, 517–524. [Google Scholar] [CrossRef]
  36. Peitgen, H.; Jürgens, H.; Saupe, D. Chaos and Fractal: New Frontiers in Science; Springer: New York, NY, USA, 2004. [Google Scholar]
  37. Vinoy, K.J.; Varadan, V.K. Design of reconfigurable fractal antennas and RF-MEMS for spaced-based communication systems. Smart Mater. Struct. 2001, 10, 1211–1223. [Google Scholar] [CrossRef]
  38. Krzysztofik, W.J. Fractal Geometry in Electromagnetics Applications—From Antenna to Metamaterials. Microw. Rev. 2013, 19, 3–14. [Google Scholar]
  39. Hohlfeld, R.G.; Cohen, N.L. Self-Similarity and the Geometric Requirements for Frequency Independence in Antennae. Fractals 1999, 7, 79–84. [Google Scholar] [CrossRef]
  40. Sheluhin, O.I.; Smolskiy, S.M.; Osin, A.V. Self-Similar Processes in Telecommunications; John Wiley & Sons: Chichester, UK, 2007. [Google Scholar]
  41. Martyn, T. A method for numerical estimation of generalized Rényi dimensions of affine Recurrent IFS invariant measures. In Thinking in Patterns: Fractals and Related Phenomena in Nature; Novak, M.N., Ed.; World Scientific Publishing: Singapore, Singapore, 2004; pp. 79–90. [Google Scholar]
  42. Słomczyński, W.; Kwapień, J.; Życzkowski, K. Entropy computing via integration over fractal measures. Chaos 2000, 10, 180–188. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Rényi and Shannon entropies for a binomial distribution with N = 20: they converge for α 1 , in accord with Equation (3). The computation of both entropies was done for b = 2 .
Figure 1. Rényi and Shannon entropies for a binomial distribution with N = 20: they converge for α 1 , in accord with Equation (3). The computation of both entropies was done for b = 2 .
Entropy 18 00084 g001
Figure 2. Kolmogorov entropy for 1D-regular, chaotic-deterministic and random systems. The attractor is the classical Lorentz attractor [27] with σ = 10 , R = 28 , b = 8 / 3 and initial values x ( 0 ) = z ( 0 ) = 0 , y ( 0 ) = 1 , while the random motion is given by a 2D-random walk from - 1 - j to 1 + j of 500 elements in which j is the imaginary unit.
Figure 2. Kolmogorov entropy for 1D-regular, chaotic-deterministic and random systems. The attractor is the classical Lorentz attractor [27] with σ = 10 , R = 28 , b = 8 / 3 and initial values x ( 0 ) = z ( 0 ) = 0 , y ( 0 ) = 1 , while the random motion is given by a 2D-random walk from - 1 - j to 1 + j of 500 elements in which j is the imaginary unit.
Entropy 18 00084 g002
Figure 3. Graph of the M s ( A ) , where A is a bounded subset of the Euclidean metric space R n . It takes only two possible values, and the Hausdorff–Besicovitch dimension of A is given by the value of s in which there is the jump from to zero.
Figure 3. Graph of the M s ( A ) , where A is a bounded subset of the Euclidean metric space R n . It takes only two possible values, and the Hausdorff–Besicovitch dimension of A is given by the value of s in which there is the jump from to zero.
Entropy 18 00084 g003
Figure 4. Here, the first steps of the box-counting procedure about England’s coastline are represented.
Figure 4. Here, the first steps of the box-counting procedure about England’s coastline are represented.
Entropy 18 00084 g004
Figure 5. Here, the von Kock curve (on the left) and the middle third Cantor set (on the right) are shown: A 0 is the initiator of length equal to one in both cases; in the generator A 1 for the von Kock curve, the middle third of the unit interval is replaced by the other two sides of an equilateral triangle, while that of the middle third Cantor set is obtained removing the middle third of the interval.
Figure 5. Here, the von Kock curve (on the left) and the middle third Cantor set (on the right) are shown: A 0 is the initiator of length equal to one in both cases; in the generator A 1 for the von Kock curve, the middle third of the unit interval is replaced by the other two sides of an equilateral triangle, while that of the middle third Cantor set is obtained removing the middle third of the interval.
Entropy 18 00084 g005
Figure 6. Archimedean spiral antenna (on the left) and commercial log-periodic dipole antenna of 16 elements (on the right) [7].
Figure 6. Archimedean spiral antenna (on the left) and commercial log-periodic dipole antenna of 16 elements (on the right) [7].
Entropy 18 00084 g006
Figure 7. A Sierpinski triangle (on the left) and a Hilbert curve (on the right) are shown: as in Figure 5, A 0 is the initiator, and A 1 is the generator. The Sierpinski triangle is constructed using the iterated function system (IFS), while the other construction is that of David Hilbert. The two relative antennas are shown below with their feed points.
Figure 7. A Sierpinski triangle (on the left) and a Hilbert curve (on the right) are shown: as in Figure 5, A 0 is the initiator, and A 1 is the generator. The Sierpinski triangle is constructed using the iterated function system (IFS), while the other construction is that of David Hilbert. The two relative antennas are shown below with their feed points.
Entropy 18 00084 g007
Figure 8. Examples of non-fractal antennas that offer similar performance over their fractal counterparts. Three different non-fractal antennas are presented above: they outperform their fractal counterparts, while the current distribution on the Sierpinski gasket antenna at the first three resonance frequencies is shown in the middle of the page. The modified Parany antenna (starting from the classical Sierpinski gasket antenna) is represented below.
Figure 8. Examples of non-fractal antennas that offer similar performance over their fractal counterparts. Three different non-fractal antennas are presented above: they outperform their fractal counterparts, while the current distribution on the Sierpinski gasket antenna at the first three resonance frequencies is shown in the middle of the page. The modified Parany antenna (starting from the classical Sierpinski gasket antenna) is represented below.
Entropy 18 00084 g008
Figure 9. The Rényi entropy H α of a Sierpinski gasket (Figure 7) with α = 0 , 1 , 2 : the plot shows us that Hartley entropy H 0 is an upper to both Shannon entropy H 1 and collision entropy H 2 . The main limit of this procedure is clearly the precision of the triangulation.
Figure 9. The Rényi entropy H α of a Sierpinski gasket (Figure 7) with α = 0 , 1 , 2 : the plot shows us that Hartley entropy H 0 is an upper to both Shannon entropy H 1 and collision entropy H 2 . The main limit of this procedure is clearly the precision of the triangulation.
Entropy 18 00084 g009

Share and Cite

MDPI and ACS Style

Guariglia, E. Entropy and Fractal Antennas. Entropy 2016, 18, 84. https://doi.org/10.3390/e18030084

AMA Style

Guariglia E. Entropy and Fractal Antennas. Entropy. 2016; 18(3):84. https://doi.org/10.3390/e18030084

Chicago/Turabian Style

Guariglia, Emanuel. 2016. "Entropy and Fractal Antennas" Entropy 18, no. 3: 84. https://doi.org/10.3390/e18030084

APA Style

Guariglia, E. (2016). Entropy and Fractal Antennas. Entropy, 18(3), 84. https://doi.org/10.3390/e18030084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop