Professional Documents
Culture Documents
17qk1a0402 Major Documentation PDF 2
17qk1a0402 Major Documentation PDF 2
BACHELOR OF TECHNOLOGY
in
By
A.SOWMYA 17QK1A0402
2017-2021
JYOTHISHMATHI INSTITUTE OF TECHNOLOGICAL SCIENCES
CERTIFICATE
This is to certify that this project work entitled “Machine Learning
Techniques for Optimizing Design of Double T-shaped Monopole
Antenna” is a bonafide workcarried out by A.SOWMYA (17QK1A0402)
in partial fulfillment of the requirements for the award of Bachelor of
Technology in Electronics and Communication Engineering during the
period of 2017-2021 underthe guidance and supervision.
A.SOWMYA
17QK1A0402
ACKNOWLEDGEMENT
Karimnagar for providing the college facilities for the completion of the
project.
Finally, I thank to all the faculty members, management and supporting
staff of ECE department, and friends for their kind co- operation and
valuable help for completing the project. My special thanks to my father
Mr. A. KAMALAKAR RAO and my mother Mrs.A.SARITHA for their
love,understanding and encouragement throughout my Project. Finally, I
thank Almighty God for always being there to guide me through thick and
thin.
A.SOWMYA
17QK1A0402
ABSTRACT
i
TABLE OF CONTENTS
CHAPTER NO. TITLE PAGE NO.
ABSTRACT I
LIST OF FIGURES IV
LIST OF TABLES V
LIST OF ABBREVIATIONS VI
1 INTRODUCTION 1
1.1 Machine Learning 3
1.1.1 History 5
1.1.2 Machine Learning techniques 7
1.1.3 Applications 13
1.1.4 Advantages and Disadvantages 15
1.1.5 Future Prosperity of Machine Learning 18
2 LITERATURE REVIEW 19
ANSYS HFSS 23
Ansys HFSS for antenna design 24
Applications 25
3 METHDOLOGY 27
3.1 Monopole antenna 27
Advantages of Monopole antenna 29
3.1.1
3.1.2 Drawbacks of Monopole antenna 30
3.1.3 Applications 31
3.1.4 Proposed antenna design parameters 32
ii
54
5
REFERENCES
iii
LIST OF FIGURES
iv
LIST OF TABLES
v
LIST OF ABBREVIATIONS
vi
SIR Shuttle Imaging Radar
SME Solar Mesospheric Explorer
ATMS Antenna Test and Measurement Society
vii
CHAPTER 1
INTRODUCTION
The upcoming era of internet of things, (IoT) has enabled an immense growth in the
demand of application-specific antennas, which are needed for almost all electronic
devices. Hence, the requirement of a smart and efficient way of an-tenna designing
has become inevitable. Current antenna design relies heavily on the designers
empirical experiences and EM simulations. Traditional methods are inherently
inefficient and computationally intensive, making them impractical when there are
a large number of antenna design parameters to be optimized such as for 3-D printed
antennas [1]. To address challenges for designing complex 3-D structures, machine
learning (ML) techniques may be highly beneficial. ML has been widely used as an
indispensable data analysis and decision-making tool in a broad range of
applications ranging from hand-written digit recognition [2] to human genomics [3].
Researchers have also explored optimization of antenna structures by applying
heuristic optimization techniques like genetic algorithms, particle swarm
optimization to antenna designs [4]–[6] as well, but these algorithms search for the
optimal solution by analyzing the output on individual data points and generating
new and possibly better search directions until a global maxima or minima is
identified. On the other hand, machine learning (ML) refers to all techniques and
Optimization algorithms of analyzing the data and finding the hidden mathematical
relation in data such that we can relate the input behavior to the output behavior and
make future predictions or decisions using this relationship. The main advantage of
using ML techniques is once we have the relational model, we can predict the output
for any data point rather than aiming for global optimal and minima points only. This
property is very beneficial when we want to use the same data set for multiple
different goals. As described in [7], there is some early work on applying machine
1
learning techniques for antenna analysis and synthesis [8]–[11]. In [8], the
performance of support vector machines (SVMs) is investigated for designing of a
rectangular patch antenna and a rectangular patch array while in [9] methodology to
use SVMs for linear and nonlinear beamforming and parameter design for array and
electromagnetic applications is mentioned. Another ML technique, Artificial Neural
Network has also been applied in this field [10]-[11]. In [12], clustering method is
used to find the optimum position for shorting posts in microstrip patch array design
so as to achieve acceptable bandwidth, scan angle and polarization. As observed
from here, some studies for using ML techniques for antenna design optimization
have been conducted but a detailed analysis and systematic comparison of various
ML techniques for antennas have not been reported. The main contribution of this
work is to fill the gap by presenting new classes of ML-based methods for automated
antenna design optimization, evaluating their performance in terms of prediction
accuracy and robustness, and making comparisons with EM simulations. Our
discovery suggests that ML is a promising choice to provide automated,
computational feasible and practically effective approaches for antenna design. The
ultimate goal of this study is to further extend the proposed ideas to more complex
design of antennas and develop scalable and efficient algorithms to tackle
computational challenges, by handling a large number of design parameters. In this
paper, we propose using machine learning techniques for antenna design
optimization, and particularly consider artificial neural networks (ANNs), least
absolute shrinkage and selection operator (lasso), and k-nearest neighbor (kNN).
The feasibility of these new approaches for antenna design is demonstrated by their
applications to optimize a reference double T-shaped monopole antenna [13] as
shown in Fig. 1.
2
The initial work related to this was reported in [14] where only
the lasso technique was employed and targeted optimization function was weighted
sum of fractional bandwidths of two bands of the reference antenna. In this paper,
we use finer optimization function and explore more machine learning techniques
to optimize the performance of the reference antenna. This helps in achieving
better results as compared to [13] and [14]
3
1.1 MACHINE LEARNING
Machine learning involves computers discovering how they can perform tasks
without being explicitly programmed to do so. It involves computers learning from
data provided so that they carry out certain tasks. For simple tasks assigned to
computers, it is possible to program algorithms telling the machine how to execute
all steps required to solve the problem at hand; on the computer's part, no learning
is needed. For more advanced tasks, it can be challenging for a human to manually
create the needed algorithms. In practice, it can turn out to be more effective to help
4
the machine develop its own algorithm, rather than having human programmers
specify every needed step.[7]
1.1.1 HISTORY
The term machine learning was coined in 1959 by Arthur Samuel, an American
IBMer and pioneer in the field of computer gaming and artificial intelligence.[8][9]
A representative book of the machine learning research during the 1960s was the
Nilsson's book on Learning Machines, dealing mostly with machine learning for
pattern classification.[10] Interest related to pattern recognition continued into the
1970s, as described by Duda and Hart in 1973.[11] In 1981 a report was given on
using teaching strategies so that a neural network learns to recognize 40 characters
(26 letters, 10 digits, and 4 special symbols) from a computer terminal.[12]
Tom M. Mitchell provided a widely quoted, more formal definition of the algorithms
studied in the machine learning field: "A computer program is said to learn from
experience E with respect to some class of tasks T and performance measure P if its
performance at tasks in T, as measured by P, improves with experience E."[13] This
5
definition of the tasks in which machine learning is concerned offers a fundamentally
operational definition rather than defining the field in cognitive terms. This follows
Alan Turing's proposal in his paper "Computing Machinery and Intelligence", in
which the question "Can machines think?" is replaced with the question "Can
machines do what we (as thinking entities) can do?".[14]
Modern day machine learning has two objectives, one is to classify data based on
models which have been developed, the other purpose is to make predictions for
future outcomes based on these models. A hypothetical algorithm specific to
classifying data may use computer vision of moles coupled with supervised learning
in order to train it to classify the cancerous moles. Where as, a machine learning
algorithm for stock trading may inform the trader of future potential predictions.[15]
Prior to lasso, the most widely used method for choosing covariates was stepwise
selection. That approach only improves prediction accuracy in certain cases, such as
when only a few covariates have a strong relationship with the outcome. However,
in other cases, it can increase prediction error.
At the time, ridge regression was the most popular technique for improving
prediction accuracy. Ridge regression improves prediction error by shrinking the
sum of the squares of the regression coefficients to be less than a fixed value in order
to reduce over fitting, but it does not perform covariate selection and therefore does
6
not help to make the model more interpretable.
Lasso achieves both of these goals by forcing the sum of the absolute value of the
regression coefficients to be less than a fixed value, which forces certain
coefficients to zero, excluding them from impacting prediction. This idea is similar
to ridge regression, which also shrinks the size of the coefficients, however Ridge
Regression tends to set far fewer coefficients to zero.
The history of Artificial Intelligence (AI) began in antiquity, with myths, stories and
rumors of artificial beings endowed with intelligence or consciousness by master
craftsmen. The seeds of modern AI were planted by classical philosophers who
attempted to describe the process of human thinking as the mechanical manipulation
of symbols. This work culminated in the invention of the programmable digital
computer in the 1940s, a machine based on the abstract essence of mathematical
reasoning. This device and the ideas behind it inspired a handful of scientists to begin
seriously discussing the possibility of building an electronic brain.
7
substantial amount about the estimator. These include its relationship to
ridge regression and best subset selection and the connections between
lasso coefficient estimates and so-called soft thresholding. It also reveals
that (like standard linear regression) the coefficient estimates do not need
to be unique if covariates are collinear.
8
Fig1.1.1 Lasso geometry
LASSO Geometry
LASSO forms a diamond shape in the plot for its constraint region, as shown in the
image below. The diamond shape includes corners, unlike the circular shape formed
by ridge regression. The proximity of the first point to the corner shows that the
model comes with one coefficient, which is equal to zero.
The ridge regression constraints region forms a circular shape that includes no
corners similar to that formed by the LASSO constraints region when plotted. The
ridge regression coefficients can, therefore, not be equal to zero.
Artificial neural networks (ANNs), usually simply called neural networks (NNs), are
computing systems vaguely inspired by the biological neural networks that
constitute animal brains.
9
Fig 1.1.2 An artificial neural network is an interconnected group of nodes,
Inspired by a simplification of neurons in a brain. Here, each circular
node represents an artificial neuron and an arrow represents a connection
from the output of one artificial neuron to the input of another.
10
connections are called edges. Neurons and edges typically have a weight that adjusts
as learning proceeds. The weight increases or decreases the strength of the signal at
a connection. Neurons may have a threshold such that a signal is sent only if the
aggregate signal crosses that threshold. Typically, neurons are aggregated into
layers. Different layers may perform different transformations on their inputs.
Signals travel from the first layer (the input layer), to the last layer (the output layer),
possibly after traversing the layers multiple times.
11
Fig 1.1.3 Example of k-NN classification .The test sample (green dot) should be classified either to blue
squares or to red triangle. If k=3 (solid line circle) it is assigned to the red triangle because there are 2
triangle and only 1 square inside the inner circle. If k=5 (dashed line circle) it is assigned to the blue
squares(3squares vs.2 triangle inside the outer circle).
Both for classification and regression, a useful technique can be to assign weights to
the contributions of the neighbors, so that the nearer neighbors contribute more to
the average than the more distant ones. For example, a common weighting scheme
consists in giving each neighbor a weight of 1/d, where d is the distance to the
neighbor.[5]
The neighbors are taken from a set of objects for which the class (for k-NN
classification) or the object property value (for k-NN regression) is known. This can
be thought of as the training set for the algorithm, though no explicit training step is
12
required.
A peculiarity of the k-NN algorithm is that it is sensitive to the local structure ofdata.
1.1.3 APPLICATIONS
➢ SPEECH RECOGNITION
Speech recognition is the translation of spoken words into the text. It is also known
as computer speech recognition or automatic speech recognition. Here, a software
application can recognize the words spoken in an audio clip or file, and then
subsequently convert the audio into a text file. The measurement in this application
can be a set of numbers that represent the speech signal. We can also segment the
speech signal by intensities in different time-frequency bands.
Speech recognition is used in the applications like voice user interface, voice
searches and more. Voice user interfaces include voice dialing, call routing, and
appliance control. It can also be used a simple data entry and the preparation of
structured documents.
13
recognition algorithms from 2002 and 100 times more accurate than those from
1995. Some of the algorithms were able to outperform human participants in
recognizing faces and could uniquely identify identical twins.
➢ Medical Diagnosis
Machine learning can be used in the techniques and tools that can help in the
diagnosis of diseases. It is used for the analysis of the clinical parameters and their
combination for the prognosis example prediction of disease progression for the
extraction of medical knowledge for the outcome research, for therapy planning and
patient monitoring. These are the successful implementations of the machine
learning methods. It can help in the integration of computer-based systems in the
healthcare sector.
➢ Financial Services
Machine learning has a lot of potential in the financial and banking sector. It is the
driving force behind the popularity of the financial services. Machine learning can
help the banks, financial institutions to make smarter decisions. Machine learning
can help the financial services to spot an account closure before it occurs. It can also
track the spending pattern of the customers. Machine learning can also perform the
market analysis. Smart machines can be trained to track the spending patterns. The
14
algorithms can identify the tends easily and can react in real time.
➢ Prediction Systems
Machine learning can also be used in the prediction systems. Considering the loan
example, to compute the probability of a fault, the system will need to classify the
available data in groups. It is defined by a set of rules prescribed by the analysts.
Once the classification is done, we can calculate the probability of the fault. These
computations can compute across all the sectors for varied purposes. Making
predictions is one of the best machine learning applications.
➢ Regression
We can also implement machine learning in the regression as well. In regression, we
can use the principle of machine learning to optimize the parameters. It can also be
used to decrease the approximation error and calculate the closest possible outcome.
We can also use the machine learning for the function optimization. We can also
choose to alter the inputs in order to get the closest possible outcome.
1.1.4 ADVANTAGES
15
histories of its users to help cater to the right products, deals, and reminders relevant
to them. It uses the results to reveal relevant advertisements to them.
➢ Continuous Improvement
As ML algorithms gain experience, they keep improving in accuracy and efficiency.
This lets them make better decisions. Say you need to make a weather forecast
model. As the amount of data you have keeps growing, your algorithms learn to
make more accurate predictions faster.
1.1.4 DISADVANTAGES
➢ Data Acquisition
Machine Learning requires massive data sets to train on, and these should be
inclusive/unbiased, and of good quality. There can also be times where they must
16
wait for new data to be generated.
➢ High error-susceptibility
Machine Learning is autonomous but highly susceptible to errors. Suppose you train
an algorithm with data sets small enough to not be inclusive. You end up with biased
predictions coming from a biased training set. This leads to irrelevant advertisements
being displayed to customers. In the case of ML, such blunders can set off a chain
of errors that can go undetected for long periods of time. And when they do get
noticed, it takes quite some time to recognize the source of the issue, and even longer
to correct it.
➢ Interpretation
When the algorithms help in all these processes and give a resulting output. This
given output must be checked for any errors and the correction operation should be
followed to get the desired accuracy. And during the selection of this algorithm, we
must select that algorithm which you require for the purpose.
17
1.1.5 FUTURE PROSPERITY OF MACHINE LEARNING
While machine learning algorithms have been around for decades, they've attained
new popularity as artificial intelligence has grown in prominence. Deep learning
models, in particular, power today's most advanced AI applications.
18
CHAPTER 2
LITERATURE REVIEW
• H. Xin and M. Liang, “3-D printed microwave and THz devices using
polymer jetting techniques,” Proc. IEEE, vol. 105, no. 4, pp. 737-755, Apr.
2017. 3-D additive manufacturing (AM) offers unprecedented flexibility in
the realization of complicated 3-D structures. Polymer jetting is one of the
promising 3-D AM techniques that utilizes photosensitive polymers as the
build material and is capable of precisely printing electromagnetic (EM)
components up into the THz range. In this paper, important design and
implementation aspects of polymer-jetting-based 3-D-printed EM
components are discussed. A number of 3-D-printable polymer materials and
their broadband EM characterization from GHz to THz are introduced. Design
methodologies specific for 3-D-printed antennas and other EM components
are presented. As examples, various 3-D-printed devices operating from GHz
to THz frequency, including electromagnetic crystals (EMXT), waveguide,
horn antenna, gradient index (GRIN) lenses, as well as 3-D AM-enabled new
designs, such as millimeter wave (mmW)/THz, reflect array antennas,
computer-generated THz holograms, and so on are reviewed. Moreover,
current limitations and possible future improvements of the polymer jetting
technique for EM applications are discussed. This type of 3-D AM technique
is likely to enable many novel antenna and circuit architectures as well as
various interesting 3-D metamaterial structures.
19
variable-fidelity simulations and surrogate models”, IEEE Trans.
Antennas Propag., vol. 61, no. 12, pp. 5931-5939, Dec. 2013. A
computationally-efficient procedure for multi-objective design of antenna
structures is presented. Our approach exploits the multi-objective
evolutionary algorithm (MOEA) working with a fast antenna surrogate model
obtained with kriging interpolation of coarse-discretization simulation data.
Response correction techniques are subsequently applied to refine the designs
obtained by MOEA. Our methodology allows us to obtain-at a low
computational cost-a set of designs corresponding to various trade-offs
between the antenna size and the refection coefficient. Two illustration
examples are considered: (i) an UWB monocone with two objectives being
reduction of the antenna size and minimization of the antenna reflection
coefficient in the bandwidth of interest, and (ii) a planar Yagi antenna with
the objectives being an increase of the end-fire gain and minimization of the
reflection coefficient, both in the bandwidth of interest.
ANSYS HFSS
23
problems ranging in detail and scale from passive IC components to extremely large-
scale EM analyses such as automotive radar scenes for ADAS systems. Its reliable
automatic adaptive mesh refinement lets you focus on the design instead of spending
time determining and creating the best mesh. This automation and guaranteed
accuracy differentiates HFSS from all other EM simulators, which require manual
user control and multiple solutions to ensure that the generated mesh is suitable and
accurate. With Ansys HFSS, the physics defines the mesh rather than the mesh
defining the physics.Ansys HFSS is the premier EM tool for R&D and virtual design
prototyping. It reduces design cycle time and boosts your product’s reliability and
performance. Beat the competition and capture your market with Ansys HFSS.
ANSYS simulation technology enables you to predict with confidence that your
products will thrive in the real world. Customers trust our software to help ensure
the integrity of their products and drive business success through innovation.
Industry Standard Full Wave, Electromagnetic Field Simulation HFSS sets the gold-
standard for accuracy, advanced solver and high-performance computing
technology, making it the ‘go to’ tool for engineers designing high-frequency and
high-speed electronics found in communication systems, radar systems, satellites,
smart phones and tablet devices. Rigorous Validation Sign-off quality high-
frequency EM results that allow customers to simulate and go straight to
manufacturing. With HFSS, engineers can extract parasitic parameters (S, Y, Z),
visualize 3D electromagnetic fields (near- and far-field) and generate Full-Wave
SPICE™ models that link to circuit simulations. Easy to Use, Versatile and Fast
Features such as automatic adaptive meshing, versatile design entry and advanced
high-performance computing technology put analyst-quality solvers in the hands of
the designer.
24
Ansys HFSS for Antenna Design
This application brief describes antenna design using Ansys HFSS, the industry
leading 3D Electromagnetic (EM) simulation tool for high frequency and high speed
electronic components. It highlights several antenna-related applications with
emphasis on antennas on or around other Structures. With multiple simulation
technologies and powerful automated adaptive mesh refinement Providing gold
standard accuracy, HFSS can help antenna designers who are constantly Challenged
with implementing designs across more and more frequency bands inside a smaller
And smaller footprint. With these additional technical challenges along with the ever
shrinking time to market, Simulation with HFSS is a must-have in the antenna design
and integration process.
Applications
The original ISM specifications envisioned that the bands would be used primarily
for non communication purposes, such as heating. The bands are still widely used
for these purposes. For many people, the most commonly encountered ISM device
is the home microwave oven operating at 2.45 GHz which uses microwaves to cook
food. Industrial heating is another big application area; such as induction heating,
microwave heat treating, plastic softening, and plastic welding processes. In medical
settings, shortwave and microwave diathermy machines use radio waves in the ISM
bands to apply deep heating to the body for relaxation and healing. More recently
hyperthermia uses microwaves to heat tissue to kill cancer cells [1] and [8].
However, as detailed below, the increasing congestion of the radio spectrum, the
increasing sophistication of microelectronics, and the attraction of unlicensed use,
has in recent decades led to an explosion of uses of these bands for short range
25
communication systems for wireless devices, which are now by far the largest uses
of these bands. These are sometimes called "non ISM" uses since they do not fall
under the originally envisioned "industrial", "scientific", and "medical" application
areas. One of the largest applications has been wireless networking(WiFi). The
bIEEE 802.11wireless networking protocols, the standards on which almost all
wireless systems are based, use the ISM bands. Virtually all laptops ,tablet
computers, computer printers and cellphones now have 802.11wireless modems
using the 2.4 and 5.7 GHz ISM bands. Bluetooth is another networking technology
using the 2.4 GHz band, which can be problematic given the probability of
interference. [9]Near field communication devices such as proximity cards and
contactless smart cards use the lower frequency 13 and 27 MHz ISM bands. Other
short range devices using the ISM bands are:wireless microphones ,monitors, garage
openers, wireless doorbells,keyless entry systems for vehicles,radio control
channels, RFID systems for merchandise, and tracking systems.
27
CHAPTER 3
METHODOLOGY
28
antennas 5/8 = 0.625 wavelength long are also popular, because at this length a
monopole radiates a maximum amount of its power in horizontal directions. The
monopole antenna was invented in 1895 by radio pioneer Guglielmo Marconi; for
this reason it is sometimes called the Marconi antenna.[4][5][6] Common types of
monopole antenna are the whip, rubber ducky, helical, random wire, umbrella,
inverted-L and T-antenna, inverted-F, folded unipole antenna, mast radiator, and
ground plane antennas. The load impedance of the quarter-wave monopole is half
that of the dipole antenna or 37.5+j21.25 ohms.
Common types of monopole antenna are the whip, rubber ducky, helical, random
wire, umbrella, inverted-L and T-antenna, inverted-F, folded unipole antenna, mast
radiator, and ground plane antennas.
29
➢ 3.1.2 The Advantages of Monopole Antennas
1. As the monopole antenna gets longer and the ground losses are reduced, the
efficiency of the antenna gets better. Vertical monopole antennas can achieve
efficiencies of up to 80%.
2. A vertical monopole antenna can be used for any frequency shorter than two
thirds of the wavelength.
3. Monopole antennas are a simple omnidirectional antenna that takes up far less
space than an array of wheel antennas stacked on top of each other. The monopole
antenna can handle communications in any direction except straight up above the
antenna.
4. Monopole antennas are easy to build and install. Passive monopoles are cheap to
make and rugged.
5. If working with frequencies over 800 MHz, the monopole antenna can be made
out of the trace on a printed circuit board; this is standard for cell phones.
6. A monopole antenna has relatively high reactive impedance over most of its
frequency range. Put an active amplifier with ah high input impedance, and you
can transform the impedance without losing any sensitivity.
The “torus” shape doesn’t extend to the top of the antenna because the voltage
increases as it travels up the antenna. In reality, the signal is sent from the bottom
two thirds of the antenna.
Metal objects and the ground itself can cause signal reflections, so you may get a
signal that is both horizontally and vertically polarized.
Because you are radiating equally in all directions, you have equally poor radiation
in all directions. The “torus” shape doesn't extend to the top of the antenna because
the voltage increases as it travels up the antenna.
As the monopole antenna gets longer and the ground losses are reduced, the
efficiency of the antenna gets better. Vertical monopole antennas can achieve
efficiencies of up to 80%. A vertical monopole antenna can be used for any
frequency shorter than two thirds of the wavelength
31
3.1.4 APPLICATIONS
➢ Military Applications:
In the military sector, high-velocity aircraft, missiles rockets and spacecrafts need
low-profile and light weight antennas for conformal mounting outside their surfaces.
The micro strip antenna is best suited for this requirement, As these antenna would
neither 13 disturb the aerodynamic flow nor protrude inward to disrupt other already
crowded space. The uses of micro strip antennas in the military sector have been
numerous. These include functions such as altimeters, beacons,guidance fusing,
telemetry command, communication radar, ECM, ECCM, GPS and so on. [53] and
[54]
➢ Space Applications:
In the space sector, numerous applications of microstrip antennas have been
implemented. To name a few, the following space programs have used microstrip
antennas as arrays. Earth Limb Measurement Satellite (ELMS), International Sun
Earth Explorer (ISEE), SEASET, Shuttle Imaging Radar (SIR), GEOSTAR, Solar
Mesospheric Explorer (SME) and Mars Pathfinder. [54]
32
3.1.5 Proposed antenna design parameters
Fig 3.1.2 Layout and design parameters of the referenced dual-band double – T shaped
monopole antenna
falling in the band of interest in order to calculate the FOM for a given design and
the same is represented mathematically
as follows:
where, f represents the frequency and S11(f) is the reflection coefficient value at that
frequency. While collecting the sample points, these parameters take values within
34
the following range of sample space, defined as: l21 ∈ [6.3, 7.3] , l22 ∈ [6.3, 7.3],
w1 ∈ [1, 3.5], w2 ∈ [1, 3.5], w ∈ [1, 3.5], with each parameter
taking a step size of 0.5 (all units are mm). In the antenna design process, these five
design parameters are input variables and FOM is the output or response variable.
The training data are represented by {(Xi , Yi), i = 1, 2, ..N}, where the input is X =
(l21, l22, w1,w2 , w) T ∈ χ or its transformation. The output Y is the value of figure
of merit (FOM). The goal is to learn a behavioral model based on the training set to
best describe the relationship between the expected F OM (F OM \ ) and the design
parameters. We represent this behavioral model by:
F OM \ = f (l21, l22, w1, w2, w) + , (2) where, is the error term and the function f is
a flexible mapping, which can be linear or nonlinear, continuous or discontinuous
function that is based on the main effects or including two-way interaction effects of
the input parameters. Machine learning methods are then used to search for the best
f ∈ F, a class of candidate models, to describe the relationship between X and Y and
make future predictions. To search for optimal design parameters based on the
obtained machine learning model, a very fine grid over the entire space χ with a step
size 0.1 mm (instead of 0.5 mm in the training set) is generated, consisting of a total
of 2,126,696 design points. The F OM values at all the design points are then
computed and the design parameter values that give the maximal F OM value are
identified. The performance of this analysis is verified by comparing the predicted
F OM at a set of test points with its actual value obtained from the HFSS simulation.
In the following subsections, the three machine learning techniques (namely, Least
absolute shrinkage and selection operator, Artificial neural network and k-nearest
neighbor) that have been used in this work to obtain the behavioral model f, are
explained.
35
Least Absolute Shrinkage and Selection Operator (lasso)
The lasso technique is a sparse regression and predictor selection algorithm that
estimates a linear model subject to some conditions. Basically it searches the value
of regression coefficients that can minimize the residual sum of squares, subject to
the condition that the absolute values of the re- gression coefficients are less than a
constant [16]. Denote the training data by (mk,nk), k = 1, 2, .., N, where mk = {mk1,
..., mkp} T
are p–dimensional predictor variables and nk are associated responses. The linear
estimation model predicts the response for a given input mk, as
nck = α+E p βpmkp. In a simple linear regression problem, the values of ( α, β) are
the difference between the actual response value nk and the estimated value (nck)
as defined below:
The standard regression analysis includes all the predictor variables in the model
fitting, irrespective of the magnitudes of their effects on the output. If some predictor
variables are not informative for prediction, these variables are regarded
36
as unimportant and should be removed from the final model, in order to improve
both prediction accuracy and model interpretability. This can be done by lasso,
which calculates the estimate (α, β) by minimizing the sum of residual squares
subject to the constraint Pp |βp| ≤ t. Here, t ≥ 0 is a tuning parameter, which controls
the amount of shrinkage that is applied to the estimates. In our analysis, we fit the
linear model by using the function lm() in R [17], which is a public-domain statistical
computing and graphics software. To increase model flexibility and capture any
nonlinear relationship between the predictors and the response, we have also
considered and compared various forms of nonlinear transformation of the raw
design parameters, including log, exponential, quadratic, cubic, etc., and finally
decided that a quadratic transformation is the most appropriate. To implement the
lasso procedure, we used the function lasso() in R. For this lasso model, the five
design parameters and their quadratic transformations are used as the predictor
variables and the F OM is used as the response variable. The optimal value of the
tuning parameter t is determined by the 5-fold cross validation method [18]. The
final fitted model predicted by the lasso is as follows:
37
3.1.3 ANN Architecture based on MLP
39
C. k–nearest neighbor (kNN)
40
CHAPTER 4
Using the behavioral models obtained from the artificial neural network model,
lasso model, and kNN model, respectively, we identify the optimal design
parameters which produce the highest F OM value. For this purpose, we predict the
F OM values for all possible combinations of values of the five design parameters
and locate the maximum F OM point. In particular, the five antenna design
parameters are varied in the following range: l21 ∈ [6.3, 7.3] , l22 ∈ [6.3, 7.3], w1 ∈
[1, 3.5], w2 ∈ [1, 3.5], w ∈ [1, 3.5], with each parameter having a step size of 0.1
(all units are in mm). simulation tool and summarized in Table I. The first column
of the top three rows in this table shows the design parameters value that is obtained
by each ML technique to give maximum F OM. The next three columns list the
corresponding maxi- mum F OM value predicted by each ML technique and the last
column is the F OM obtained from HFSS if the design parameters mentioned in the
first column are substituted in HFSS design. For comparison the design parameters
value of the reference antenna as mentioned in [13] are also stated in the first column
of the last row of Table I and in next columns of the fourth row, the corresponding
F OM value predicted by each technique as well as HFSS is shown. The fifth row in
this table mentions the total computation time corresponding to each of the ML
technique i.e. the time it took to train the ML model and then search the
corresponding best set of design parameters from that model. In the following
subsections, the results of Table I for each ML techniques are discussed in more
detail.
41
Table 4.1 Predicted FOM, design parameter values and computation time comparison
42
lasso Results
The lasso model is trained with 20 parameters in the input space, consists of five
original design parameters and their square as well as their cross product terms. Five-
fold cross validation [16] is used to determine the optimal value of the tuning
parameter. Though lasso is known to be able to produce sparse solutions, for this
case none of the coefficients becomes zero (as observed from Equation (4)), which
implies that all the predictors have non-trivial effects on the output performance. For
the model predicted by lasso as represented in Equation (4), the sensitivity analysis
is performed by perturbing few coefficients and the results are mentioned in Table
II. It can be observed from Table II that maximum value of F OM and the design
parameter values for which this maximum occurs for the perturbed model, remains
almost the same to the values predicted from the actual model. This proves the
robustness of the lasso model against slight random perturbations
43
Table 4.2 Sensitivity Analysis of predicted lasso model: Predicted FOM and design parameter values
when coefficients of Equation (4) are perturbed
44
The maximum F OM predicted by the lasso model is mentioned in first row of
Table I, which can be achieved by choosing the design parameter values as: l21 =
7.3 mm, l22 = 6.3 mm, w1 = 1 mm, w2 = 3.5 mm and w = 3.5mm. The trained
lasso model also predicts the F OM value at this design location equal to 298.56.
To verify the results, a HFSS simulation for these values of design parameters is
done. The corresponding value of F OM obtained from HFSS is 298.06
The results obtained from neural networks are more precise than lasso. For neural
network analysis, 70% of data was used for training and 15% each is used for testing
and validation, respectively. The input layer consists of five nodes, each representing
one design parameter, a hidden layer with 5 hidden nodes, and the output layer of
one single node, representing the F OM. As observed from the second row of Table
I, the maximum F OM is expected to occur at l21 = 7.3 mm, l22 = 6.3 mm, w1 = 1
mm, w2 = 3.5 mm and w = 3.5mm by the ANN model and the corresponding F OM
value is predicted to be equal to 297.68. The F OM value for these design parameters
is also checked through the HFSS simulation; the resulting value is 298.06. It can be
observed from here that the results obtained from lasso and neural network are very
close to each other.
Predicted as well as simulated F OM values with respect to change in (a) l21, (b)
l22, (c) w1, (d) w2 and (e) w respectively while keeping other four constants to the
values shown on top of each plot and legend shown on bottom-left.
48
(b) FOM value with respect to change in l
49
(C) FOM values with respect to change in w1
50
(e) FOM values with respect to change in w3
51
CHAPTER 5
CONCLUSION
In this work three machine learning techniques namely least absolute shrinkage and
selection operator (lasso), artificial neural network (ANN) and k-nearest neighbors
(kNN) are used to automatically identify the optimal values of the design parameters
for a reference antenna where it can provide the best performance in terms of
bandwidth of two bands. The brief description about these techniques is first
presented in this paper and then how these techniques are applied to a reference
double T-shaped monopole antenna is explained. With the help of these ML
techniques, performance of the reference antenna is analyzed for 2126696 design
points within a few seconds by learning from the training dataset of 450 data points
only. Compared to kNN, both neural network and lasso give more accurate
predictions in our study. In summary, these new methods are much more efficient
than traditional method of EM simulation optimization for achieving optimal
antenna design. The results obtained from this research imply that ML techniques
have the power to revolutionize EM simulation technology. Due to computational
power limits of EM tools it is challenging and time consuming to optimize complex
antenna designs like 3-D antenna structures involving a large number of design
parameters. This problem can be addressed by incorporating ML techniques in
simulation tools. The ultimate goal of this work is to further generalize machine
learning methods for complex design structures such as structures manufactured by
3D printing technology. For example, 3D printing enables designs with large
numbers of degree of freedom and therefore optimizing all parameters through EM
simulations is both tedious and computationally intensive.
52
Our preliminary results have shown that machine learning techniques may be able
to enable versatile and potentially automated design of antennas which will be
beneficial for anumber of applications including internet of things
53
REFERENCES
[1] H. Xin and M. Liang, “3-D printed microwave and THz devices using polymer jetting
techniques,” Proc. IEEE, vol. 105, no. 4, pp. 737-755, Apr. 2017.
[2] T. Hastie and P. Simard, “Metrics and Models for Handwritten Character Recognition,”
Statistical Science, vol. 13, no. 1, pp. 54-65, 1998.
[8] Z. Zheng, X. Chen and K. Huang, “Application of support vector machines to the antenna
design,” Int. J. RF Microw. Computer-Aided Eng., vol. 21, no. 1, pp. 85-90, 2010
[9] M. M.-Ramon, and C. Christodoulou, “Support Vector Machines for Antenna Array
Processing and Electromagnetics,” Morgan Claypool Publishers, 2006.
[11] K. C. Lee, and T. N. Lin, “Application of neural networks to analyses of nonlinearly loaded
54
antenna arrays including mutual coupling effects,” IEEE Trans Antennas Propag., vol. 53, no. 3,
pp. 11261132, 2005.
[12] G. Dadashzadeh, M. Kargar, Y. Torabi, and B. Rahmati, “Broad-Band and Wide Scan Phased
Array Element Design Using Data Mining,” Appl. Comput. Electromagn. Soc. J., vol. 31, no. 3,
pp. 244-251, 2016.
[13] Y. L. Kuo and K. L. Wong, “Printed double-T monopole antenna for 2.4/5.2 GHz dual-band
WLAN operations,” IEEE Trans. Antennas Propag., vol. 51, no. 9, pp. 2187-2192, Sept. 2003.
[14] Y. Sharma, J. Wu, H. Xin, and H. H. Zhang , “Sparse linear regression for optimizing design
parameters of double T-shaped monopole antennas,” 2017 IEEE International Symposium on
Antennas and Propagation & USNC/URSI National Radio Science Meeting, San Diego, CA,
2017, pp. 347-348.
[17] R: A Language and Environment for Statistical Computing, R Foundation for Statistical
Computing, Vienna, Austria, 2013.
[18] T. Hastie, R. Tibshirani, and J. Friedman, “The elements of statistical learning: data mining,
inference and prediction,” Springer, 2nd ed., 2016.
[21] J. -S. R. Jang, “Self-learning fuzzy controllers based on temporal back- propagation,” IEEE
Trans. Neural Networks, vol. 3, no. 5, pp. 714-723,1992.
[22] M. T. Hagan and M. B. Menhaj, “Training feedforward networks with the Marquardt
algorithm,” IEEE Trans. Neural Networks, vol. 5, no. 6, pp. 989-993, 1994.
[24] J. -M. Wu, “Multilayer Potts Perceptrons with Levenberg-Marquardt Learning,” IEEE Trans.
Neural Networks, vol. 19, no. 12, pp. 2032-2043, Dec. 2008.
[25] H. Demuth and H.M. Beale, “Neural Network Toolbox for Use with Matlab– User’s Guide
Verion 3.0,” 1998.
[27] A. Beygelzimer, S. Kakadet, J.Langford, S. Arya, D. Mount and S. Li, “Fast Nearest
Neighbor Search Algorithms and Applications,” R package, version1.1.2.1,20
56