You are on page 1of 4

Continuous-time Optimization using Sub-threshold

Current-mode Growth Transform Circuits


Ahana Gangopadhyay† , Oindrila Chatterjee† and Shantanu Chakrabartty
Department of Electrical and Systems Engineering
Washington University in St. Louis, MO, USA 63130
Email: shantanu@wustl.edu

Abstract—Analog circuits have long been used for solving were proposed in the canonical nonlinear programming circuit
various mathematical optimization problems due to their lower of Chua and Lin [4], the constrained optimization circuit
latency when compared to their digital counterparts. This paper of Platt [5], the penalty function based approach proposed
presents a novel continuous-time analog optimization circuit
based on a growth transform-based fixed-point algorithm. The by Lillo et al. [6] and many more. These circuits had
circuit uses translinear MOSFET elements to implement the mul- considerably high settling times (in orders of milliseconds),
tiplication and normalization functions using only 5 transistors, primarily because of the capacitances present in the circuit
whereas continuous-time updates and recursion are implemented for implementing the required constraint functions. Later in
using current mirrors. The circuit does not require any additional Vichik’s work [7], a similar concept was utilized for solving
components to enforce optimization constraints and naturally
converges to a steady-state solution corresponding to a local linear and quadratic programming problems by using an FPAA
minimum of an objective function. We show that the proposed (Field Programmable Analog Array) based implementation,
circuit is generic enough to encompass a multitude of objective which drastically reduced the computation time to the order
functions simply by changing the external circuitry, and the of microseconds. This was later extended in [8] to obtain
power dissipation of circuit can be adjusted according to the nanosecond range computation time with reasonably high
desired latency. For this paper, we present simulation results for
specific forms of quadratic and linear cost functions with tunable accuracy using switched capacitor circuits. However, the above
coefficients, subject to a normalization constraint, and the results methods suffer from a number of limitations. First, they im-
show excellent match to floating-point software simulation results. plement the Lagrangian function corresponding to the original
constrained optimization problem, and use the KKT conditions
Index Terms—analog optimization, growth transform, sub- for arriving at the equilibrium point by essentially solving the
threshold, translinear principle.
dual problem. The notion of strong duality is thus implicitly
I. I NTRODUCTION assumed, which may not be true for all classes of optimization
problems. Secondly, they have significantly higher power dissi-
Physical processes occurring in nature have an universal
pation, which scales with the number of optimization variables
tendency to move towards the most stable (minimum-energy)
involved. Additionally, parasitic capacitances involved in the
state over a constraint manifold, defined by the values taken
nonlinear programming implementation as well as in the
by the process variables. Since such processes are essentially
switched capacitor based implementation significantly degrade
analog in nature, exploiting analog circuits for performing such
the accuracy of the solution.
tasks is an obvious choice. Moreover, analog circuits are ca-
In this paper, we propose a continuous-time analog opti-
pable of processing real-time optimization problems and con-
mization circuit which uses multiplicative updates based on
verge to the steady-state solution almost quasi-instantaneously
Growth Transforms for solving generic optimization problems
because of the constant interaction between all parts of the
involving Lipschitz continuous cost functions over a scaled
circuit, in contrast to digital circuits having higher latency
probabilistic domain [9]. Sub-threshold operation ensures that
owing to the sequential updates they implement, which scales
power dissipation is extremely low, whereas inherent conser-
with the size of the optimization problem.
vation constraints present in the circuit alleviate the need for
Analog circuits were first used for solving linear and
designing separate circuitry for enforcing KKT conditions.
quadratic optimization problems in [1] by computing the
Additionally, since the proposed circuit does not implement
steady-state currents and node potentials in electrical circuits
an equivalent Lagrangian for the constrained optimization
consisting of resistors, voltage and current sources and diodes
problem unlike existing analog optimization topologies, the
for implementing the equality and inequality constraints. This
assumption of strong duality is not necessary. In Sections II
was later extended to more general nonlinear optimization
and III, we present the Growth transform updates, followed
problems which were applied for solving neural network based
by a circuit-level implementation of the same. In Section
algorithms [2], [3]. Modifications to the above topologies
IV, we show circuit simulation results for solving different
† Both the first authors Ahana Gangopadhyay and Oindrila Chatterjee forms of quadratic and linear optimization problems subject
contributed equally to this paper. to a normalization constraint, and compare it with results

978-1-5386-7392-8/18/$31.00 ©2018 IEEE 246

Authorized licensed use limited to: J.R.D. Tata Memorial Library Indian Institute of Science Bengaluru. Downloaded on August 22,2023 at 07:57:14 UTC from IEEE Xplore. Restrictions apply.
Fig. 1. Circuit-level implementation. (A) Schematic of the Growth transform based analog optimization circuit. The i−th block Mi implements Growth
∂H
transform updates on the i-th variable xi , given the corresponding gradient term for optimizing a cost function H({xi }). Updated currents are normalized
∂xi
at node A, w.r.t. the normalization current I0 . (B) and (C) External circuitry for implementing quadratic and linear objective functions respectively.

from software. In Section V, we conclude the paper with III. C IRCUIT I MPLEMENTATION
a brief discussion about the proposed framework and future We use the Growth transform minimization algorithm in
extensions. conjunction with the translinear principle for designing a
current-mode analog optimization circuit. By construction, the
II. OVERVIEW OF G ROWTH T RANSFORMS circuit will be capable of minimizing a generic Lipschitz con-
tinuous objective function subject to a conservation constraint
Growth transforms based on the Baum-Eagon inequality naturally implemented by the circuit. The design is essentially
[10] define a continuous mapping σ : D → D for minimizing canonical, consisting of individual update blocks for the opti-
a continuous function H on a continuously differentiable mization variables, which perform the multiplicative updates
N
xi = 1} ⊂ RN ,
P
manifold D = {xi : xi ≥ 0 and locally on each variable in the optimization problem, and sends
i=1 the updated variable for subsequent normalization.
where
∂H Fig.1(A) shows a circuit-level implementation of the i−th
xni (− + λ) submodule Mi , where the pins Pvar and Pder are connected to
∂xni
σ(xni ) = xn+1
i = , (1) an external circuit that governs the type of cost function to be
P n ∂H
xj (− n + λ) implemented by selecting the particular form for the derivative
j ∂xj term. We consider all of the transistors to be in sub-threshold,
which implies the following relationship between the drain
with λ being chosen to ensure that xn+1 i ≥ 0 ∀i. Originally current and the gate to source voltage:
applied for maximizing any arbitrary polynomial H({xi })
with real coefficients defined over a probabilistic domain, they κVg − Vs
Ids = KIs exp( ), (3)
were later extended for optimizing rational polynomials [11] UT
and more generic nonlinear analytic functions [12] over any where:
real domain having box constraints [13]. 2µCox UT2 κVth
Is = specific current = exp(− ).
In [9], a continuous-time variant of the Growth transform κ UT
updates was proposed, where each variable xi dynamically K = aspect ratio = W/L.
evolves with a time constant τ according to the following UT = thermal voltage.
equation: Vth = threshold voltage.
dxi (t) κ = gate coefficient.
τ + xi (t) = σ(xi (t)), (2)
dt Under the assumption that κ ' 1, this can be approximated
and the solution converges to the optimal point of H in steady as follows:
Vgs
state. Ids = KIs exp( ). (4)
UT

247

Authorized licensed use limited to: J.R.D. Tata Memorial Library Indian Institute of Science Bengaluru. Downloaded on August 22,2023 at 07:57:14 UTC from IEEE Xplore. Restrictions apply.
αi x2i .
P
The transistors N1 , N2 and N3 of each block, along which corresponds to the cost function H({xi }) =
with the normalization circuit, implement translinear current i
For demonstrating the circuit functionality, we chose a sub-
multiplication-normalization, such that the following relation circuit consisting of 4 variables, and set the voltage sources
holds: Vi and Vdi at the four blocks so as to produce the following
Vg|N 3i (t) arbitrary quadratic cost function
exp(κ )
dxi (t) UT
τ + xi (t) = I0 , ∀i, j = 1, . . . , N. Hquadratic = 2.2255x21 + x22 + 0.6703x23 + 1.2214x24 (8)
dt P Vg|N 3j (t)
exp(κ )
j UT which the circuit minimizes over the domain D. Fig. 2(A)
(5) shows the time-evolution of individual currents xi when the
Since Vg|N 3 (t) = Vgs|N 1i (t) + Vgs|N 2i (t), we arrive at the output pins are connected to the respective voltage sources at
following relation: time = 0. Here, the normalization current I0 = 100 nA, so
prior to t = 0, the currents are equally divided among the four
∂H
xi (t)(− + λ) update blocks (i.e., 25 nA). When the different voltage sources
dxi (t) ∂xi (t) are applied, the circuit applies growth-transform updates and
τ + xi (t) = I0 . (6)
dt P ∂H reaches the new steady-state with a latency of < 10µs.
xj (t)(− + λ)
j ∂xj (t) For the next set of experiments, we show that the operation
Inside each sub-module, the updated current xi is mirrored to of the quadratic circuit remains stable when we sweep one of
N2 at each instant to be used for the next update. The time- the coefficients (say, α1 ) over a large range, while keeping the
constant τ of the circuit is determined by the gate capacitances other coefficients fixed at 1. Fig. 2(B) shows the comparison
at the mirror and parasitic capacitances. A second mirror between the normalized steady-state solution obtained from
is used at pin Pvar , which gives the instantaneous value SPICE simulations, and the corresponding software simula-
of the variable at each block. The current λ can be tuned tions for different values of α1 .
to ensure that the current flowing through N 1 is always It is also possible to solve L1 − regularized least
P 1 squares
nonnegative. Since Eq. 6 exactly correponds to the growth optimization problems of the form H({xi }) = 2 (αi xi −
i
transform updates in Eq. 1, it follows that the proposed circuit yi )2 by connecting a constant current sink yi to pin Py . The
reaches the optimal solution corresponding to the cost function SPICE simulation results corresponding to
H in steady-state. Moreover, since KCL is enforced at each
1
node of the circuit at each instant, convergence of the solution Hrls = ((x1 − 30)2 + (x2 − 20)2 + (x3 − 40)2 + (x4 − 25)2 )
to the steady state is almost instantaneous, leading to high 2
(9)
speed operations, while sub-threshold operation ensures very are shown in Fig.2(C), which conform to software simulations
low power dissipation. with an average accuracy of 97.71% and an average latency
IV. E XPERIMENTAL R ESULTS of 50 µs across the four variables.
Although the proposed circuit is generic enough to im- B. Linear objective function
plement any Lipschitz continuous function by connecting
different external circuitry to the pins Pvar and Pder , in this The proposed framework is also capable of solving
P a linear
paper we present simulation results corresponding to the min- programming problem of the form H({xi }) = αi xi subject
i
imization of quadratic and linear objective functions subject to a normalization constraint, by connecting a constant current
to the normalization constraint. The circuit in Fig. 1(A) was sink αi to the pin Pder , and connecting Pvar to ground through
simulated using 0.5µm CMOS process SPICE parameters, and T1 and T2 , as shown in Fig.1(C). Since we are minimizing a
the results were compared against the equivalent optimization linear cost function under a simple sum constraint, the solution
problem implemented on floating-point software. will converge to a boundary point of the domain, with the
entire current being drawn by the block Mi having the lowest
A. Quadratic objective function
value of αi . This conforms to our SPICE simulation results
Using the proposed circuit, we can implement an objective shown in Fig.2(D), for the following simple cost function
∂H
function which is quadratic in xi , by making ∂x i
a linear
function of xi through a current mirror as shown in Fig. Hlinear = 15x1 + 20x2 + 10x3 + 25x4 , (10)
1(B). Moreover, we can make the weights (coefficients) of the
in which case, x3 converges to I0 as expected, whereas the
quadratic terms different simply by connecting different volt-
other variables draw no current. Here too, the average latency
age sources Vi and Vdi to the pins Pvar and Pder respectively,
is less than 10 µs.
such that Vi , Vdi  Vdd . Sub-threshold operation ensures that
the following relationship holds between the currents through V. C ONCLUSIONS
T1 and T2
! In this paper, we present a novel sub-threshold current-mode
∂H Vi − Vdi analog optimization circuit where the optimization variables
= αi xi , αi = exp , (7)
∂xi UT implement Growth transform-based multiplicative updates in

248

Authorized licensed use limited to: J.R.D. Tata Memorial Library Indian Institute of Science Bengaluru. Downloaded on August 22,2023 at 07:57:14 UTC from IEEE Xplore. Restrictions apply.
Fig. 2. (A) Solution of the quadratic objective function Hquadratic . (B) Comparison between the normalized steady-state values obtained from SPICE and
from software simulations, for a quadratic optimization problem by sweeping α1 while keeping α2 , α3 , α4 = 1. (C) Solution of the quadratic objective
function Hrls . (D) Solution of the linear objective function Hlinear .

parallel, for optimizing a generic Lipschitz continuous objec- [2] D. Tank and J. Hopfield, “Simple ‘neural’optimization networks: An
tive function subject to a normalization constraint. Simula- a/d converter, signal decision circuit, and a linear programming circuit,”
IEEE transactions on circuits and systems, vol. 33, no. 5, pp. 533–541,
tion results indicate that the proposed circuit produces high 1986.
accuracy solutions with significantly low latency and very [3] M. P. Kennedy and L. O. Chua, “Neural networks for nonlinear
low power dissipation, while simultaneously alleviating the programming,” IEEE Transactions on Circuits and Systems, vol. 35,
no. 5, pp. 554–562, 1988.
need for separate circuitry for enforcing the KKT conditions, [4] L. Chua and G.-N. Lin, “Nonlinear programming without computation,”
unlike other analog optimization frameworks. Additionally, IEEE Transactions on Circuits and Systems, vol. 31, no. 2, pp. 182–188,
existing topologies for analog optimization require separate 1984.
[5] J. C. Platt, “Analog circuits for constrained optimization,” in Advances
circuitry to optimize different polynomial cost functions, and in Neural Information Processing Systems, pp. 777–784, 1990.
circuit parameters are usually hard-wired for solving a cost [6] W. E. Lillo, M. H. Loh, S. Hui, and S. H. Zak, “On solving con-
function with fixed coefficients. We show that the proposed strained optimization problems with neural networks: A penalty method
approach,” IEEE Transactions on neural networks, vol. 4, no. 6, pp. 931–
circuit can implement quadratic and linear objective functions 940, 1993.
with easily tunable coefficients for individual terms, simply by [7] S. Vichik, Quadratic and linear optimization with analog circuits.
connecting different external circuitry to the canonical building University of California, Berkeley, 2015.
[8] K. Deems, “High speed analog circuit for solving optimization prob-
block. This can also be extended easily to implement other lems.” MS thesis. University of California in Berkeley, 2015. url:
types of nonlinear cost functions. The proposed framework http://www. eecs. berkeley. edu/Pubs/TechRpts/2015/EECS-2015-133.
is capable of minimizing a time-varying cost function in real- pdf.
[9] O. Chatterjee and S. Chakrabartty, “Decentralized global optimization
time, and is moreover scalable to larger problem sizes, as long based on a growth transform dynamical system model,” IEEE Transac-
as the sub-threshold operation of each individual update block tions on Neural Networks and Learning Systems, 2018.
is ensured. Additionally, because of the inherently modular [10] L. E. Baum and G. Sell, “Growth transformations for functions on
manifolds,” Pacific Journal of Mathematics, vol. 27, no. 2, pp. 211–
structure of the framework, faults in any part of the circuit 227, 1968.
can be easily isolated, or variables can be added or taken [11] P. Gopalakrishnan, D. Kanevsky, A. Nadas, and D. Nahamoo, “A
out of the optimization problem without changing the overall generalization of the baum algorithm to rational objective functions,”
in Acoustics, Speech, and Signal Processing, 1989. ICASSP-89., 1989
architecture. International Conference on, pp. 631–634, IEEE, 1989.
[12] D. Kanevsky, “Extended baum transformations for general func-
VI. ACKNOWLEDGMENT tions,” in Acoustics, Speech, and Signal Processing, 2004. Proceed-
This work was supported by the National Science Founda- ings.(ICASSP’04). IEEE International Conference on, vol. 1, pp. I–821,
IEEE, 2004.
tion under Grant CSR 1405273. [13] A. Gangopadhyay, O. Chatterjee, and S. Chakrabartty, “Extended poly-
R EFERENCES nomial growth transforms for design and training of generalized support
vector machines,” IEEE transactions on neural networks and learning
[1] J. B. Dennis, Mathematical programming and electrical networks. PhD systems, 2017.
thesis, Massachusetts Institute of Technology, 1959.

249

Authorized licensed use limited to: J.R.D. Tata Memorial Library Indian Institute of Science Bengaluru. Downloaded on August 22,2023 at 07:57:14 UTC from IEEE Xplore. Restrictions apply.

You might also like