You are on page 1of 24

Course / Program Information

1.1 Course/Module Information


University/College/
University of Sierra Leone Fourah Bay College
Institute
Faculty Engineering and Architecture
Department Electrical and Electronics
Advance Digital System Design Module Code:
Module Title: Credit Hours: 3
II EENG 424
Semester: First Semester Year: 4
Duration: 15 Weeks
Prof J
Phone #
Redwood- Office Rm #: ELEC
Lecturer 076/670/904 ; Instructor’s Email:
Sawyerr/Dr. RM 005
076323708 Osehsylrs1952@gm
Samba Sesay
ail.com &
Days Time Lecture Room Office Hours
sambasey@gmail.co
Wednesday: 14:00 –
m
16:00
Monday 09:00 – 10:00 EEB 10
Wednesday: 14:00 –
16:00
Wednesday 09:00 – 11:00 EEB 10
Pre-requisites:
EENG 414 Digital System Design III
1.2 Course / Module Details
Course Objectives:
This course is designed to enable students to acquire knowledge on Programmable logic devices,
Arithmetic circuits, Circuit Reliability and Fault diagnosis.
Learning Outcomes:
After the completion of the course students will be able to:
1. To design digital circuits using programming logic devices
2. Demonstrate the function of adder/subtracter circuits using gates
3. To Analyze and determine the reliability of circuit components
4. To perform circuit fault diagnosis and Testing
Module Contents:
1 PROGRAMMABLE LOGIC DEVICES
 Describe the Architecture and design procedures of various logic devices [ PROMS, PLAs,
PALs, PLDs]
 Implementation of logic functions using programmable logic devices
 Implementation of combinational and sequential logic using programmable logic devices
 Design of Seven-Segment Display using programmable logic devices
2 ARITHMETIC CIRCUITS
 Design Half Adder [ Block diagram, truth table, Sum and Carry equations, Implementation]
 Design Full Adder [ Block diagram, truth table, Sum and Carry equations, Implementation]
 Design of Multi-bit (parallel) Adders using cascade of Full Adders
 Explain the reasons for Look-Ahead Carry Generators
 Design Adders using Look-Ahead Carry Generators
 Design Half Subtracter
[ Block diagram, truth table, Difference and Borrow equations, Implementation]
 Design Full Subtracter
[ Block diagram, truth table, Difference and Borrow equations, Implementation]
 Design of Multi-bit (parallel) Subtracter
 Design of Multipliers
 Design of Dividers
 Design of Comparators
3 RELIABILITY
 Failure Rate Curve – Burn-in
 Useful life and Wear-out
 Mean time between failures-MTBF and Mean time to failures-MTTF
 Availability and the factors that affect it.
 Type of failures
 Reliability function (series, parallel and series-parallel)
 Factors that affect reliability
 Maintainability
4 FAULT DIAGNOSIS AND TESTING
 Fault detection and Location
 Gate sensitivity
 Fault-test for a 2-input AND gate
 Undetectable faults
 Bridging faults
 Fault detection table
 Two-level circuit fault detection in AND/OR circuit
 Two-level circuit fault detection in OR/AND circuit
 Boolean difference
 Testing Techniques
 Designing for Testability
Teaching and Learning Activities/Method of Delivery:
 Lectures
 Laboratory work
 Group work
 Problem Solving
 Tutorial
Assessment Strategy:

Assessment Method Point Value (Contribution to Total Mark (%)


Attendance 5
Class Work 5
Assignments 10
Group Work 10
Laboratory/Workshop Report 20
Final Exam 50
Total Point Value 100%
Recommended Texts / References / Resources:
1. Brian Holdsworth, “Digital Logic Design”, Third Edition
1. D C Green, “Digital Electronics Technology”, Third Edition.
2. Victor P. Nelson, H. Troy Nagle, Bill D. Carroll and J. David Irwin “Digital Logic Circuit Analysis
and Design”, Prentice Hall.
Attendance Policy and in – Class Expectations (turning in assignments late, respect for other
students, use of technology during class time, etc.)
References Books:

Grading Criteria:
A ≥ 70% B+ = 65-69% B= 60-64% B- = 55-59% C+ = 50-54% C=46-49% C- = 40-45% D = 35-39% E= 30-
34% F= 0-29%
Discrimination and Sexual Harassment Policy:
The University/Institution of is committed to fostering an environment free from discrimination,
including sexual or gender-based harassment or misconduct.  It is the policy of the University to
maintain an academic and work environment free of discrimination, including harassment.  The
University prohibits discrimination and harassment against any person because of age, ancestry, color,
disability or handicap, national origin, race, religious creed, sex, sexual orientation or gender identity.
Behaviors including sexual harassment, sexual misconduct, dating violence and stalking, as well as
retaliation for reporting any of these acts are not tolerated and will be punishable according to law.
Any of these incidences reported to the Dean of Faculty, Registrar or Deputy Registrar will be dealt
with sternly according to University code of conduct.
Disability Policy:
Student

Develop Disciplinary Knowledge of Theories and Concepts


Assessments

Develop ability to Design and Implement Research of data


Apply Theories, Concepts, Methods and Skills to Problem

Design/ Development of Solutions to Complex Problem

Ethical Understanding of The Engineer and Society


The University/Institution is committed to: complying with all relevant legislation regarding access and
equity for people with disabilities; providing services and support for students with disabilities to
Course Number and enable them to participate fully and independently in the academic, cultural and social life of the
Name: university; providing access for staff with disabilities to enable them to participate fully and
EENG 424:

Use of Modern Tools in the Discipline


independently in all aspects of their work and career development.
Advanced Digital Teaching Assessment

Ability to communicate effectively


System Design II & Methods

Develop Critical Thinking Skills


Learning
Activitie
s
(Identify
)

Analysis

Course Outcomes
1. To design digital Lecture, I D D D Classwork,
circuits using laboratory Assignment,
programming work Final Exams,
logic devices Lab Reports
2. Demonstrate the Lecture, I D D I Classwork,
function of laboratory Assignment,
adder/subtracter work, Final Exams,
circuits using Problem Lab Reports
gates Solving
3. To Analyze and Lecture, D D D I Classwork,
determine the Problem Assignment,
reliability of Solving, Final Exams
circuit Tutorials
components
4. To perform Lecture, I D D A Classwork,
circuit fault Problem Assignment,
diagnosis and
FOURAH BAY COLLEGE
INSTITUTION

ELECTRICAL AND ELECTRONIC ENGINEERING


DEPARTMENT

ADVANCED DIGITAL SYSTEMS DESIGN


MODULE TITLE

EENG424
MODULE CODE

PROF JONAS A S REDWOOD-SAWYERR


MODULE LECTURER
Mobile : +232 76 670904, +232 88 001019, +232 25 273401
CONTACTS Email : jasredwood@usl.edu.sl, abiosehrs@gmail.com, jredwood_sawyerr@yahoo.com

LECTURE HOURS/LAB. 3-0-3


PRACTICALS/CREDIT HOURS
Examination will account for 70% of the overall grade whilst continuous assessment will account
ASSESSMENT for 30%. Question sets at the end of each lecture must be returned for grading one week after
receipt of lectures and will account for 10% of the continuous assessment grade.
This module draws from a number of concepts met in earlier years of the programme and seeks to
show how these engineering tools can be used in digital systems design. You may wish to revise
Modules in Digital Systems at Year IV so as to better cope with the discussions as these concepts
OBJECTIVES AND OUTCOMES will be assumed during our lectures.

References have been made to a number of text books which are listed in the section for Further
Reading, and modifications made for a coherent and reader-friendly presentation with worked
examples and assignments provided.
MODULE DESCRIPTION
Below is a list of the topics to be covered.

 RELIABILITY: Failure Rate Curve – Burn-in; Useful life and Wear-out; Systems Reliability
(R(t) = exp(-λt), where λ is defined as the Failure Rate
 Mean time between failures-MTBF and Mean time to failures-MTTF; Mean Time To Repair-
MTTR; Availability and the factors that affect it; Type of failures; Reliability function (series,
parallel and series-parallel); Factors that affect reliability; Maintainability.
 FAULT DIAGNOSIS AND TESTING: Fault detection and Location; Gate sensitivity; Fault-
test for a 2-input AND gate; Undetectable faults; Bridging faults; Fault detection table; Fault
Library; Two-level circuit fault detection in AND/OR circuit; Two-level circuit fault detection
in OR/AND circuit; Boolean difference; Testing Techniques; Designing for Testability.
References
1. http://www.reliabilityeducation.com/ReliabilityPredictionBasics.pdf
SUGGESTED 2. http://www.mtl-inst.com/images/uploads/datasheets/App_Notes/AN9030.pdf
READINGREFERENCE 3. http://en.wikipedia.org/wiki/Failure_rate
TEXTS/MANUALS/WEBSITES 4. G C Loveday,1989. Electronic Testing and Fault Diagnosis. Longman Scientific & Technical.
ISBN 0-582-03865-0

01 – Systems Reliability analysis


LECTURE NUMBER
LECTURE DURATION (HOURS) 01
At the end of this lecture the students should be able to :
1. Derive the relationship between Reliability and failure rate
SPECIFIC INSTRUCTIONAL 2. Explain the practical relationship between failure rate and reliability as shown in the ‘Bath-
OBJECTIVES AND LEARNING Tub’ curve of Reliability
OUTCOMES 3. Derive and explain the meaning of the figures of merit MTTR and MTBF in network
analysis
4. Derive and explain the meaning of Availability and the discussion relating this to the
MTTR and MTBF
5. Solve problems relating to these concepts.

Welcome to Module EENG514, Digital Systems Design.

1.1 Reliability

The probability of the survival of components in an electronic system or of the sub-systems that constitute the system or that of the system as
a whole within a given period of observation is referred to as its Reliability R(t). We shall now derive the relationship of this figure of merit
and the failure rate, which is a factor usually provided by the manufacturers.

Derivation

Let us consider the degradation of N identical components under stress conditions, i.e. functioning under conditions of pressure, temperature,
humidity, etc.
Let S(t) denote the number of surviving components, i.e. the number of components still operating at time t, after the start of the ageing
experiment.
Let F(t) represent the number of components that have failed up to time t.

Thus the probability of survival of the components also known as the Reliability R(t) is given as

S (t )
R ( t )=
N
...1.1
The probability of failure of the components also known as the Un-reliability Q(t) is given by

F (t )
Q (t)=
N
...1.2
Now

S(t) + F(t) = N ..1.3

Hence

R(t) + Q(t) = 1 ...1.4

The failure rate also known as the hazard rate Z(t) is defined as the Number of failures per unit time compared with the number of surviving
components and given as
1 dF(t)
Z ( t )= .
S ( t ) dt

...1.5

Let us now discuss the relationship known as the Bath-tub curve of Failure rate in practical systems by referring to Figure 1.0

Failure Rate (Hazard Rate

Early Life Period Constant Failure Rate(λ) Wear-Out Period

‘Bath-tub’ curve of Failure Rate


Random failures

Figure 1.0. Variation of Useful Life Period


Failure Rate with time.

Time(hrs)

‘Burn-In’ Period

At the end of a production line some of the components might have inherent deficiencies which could lead to their failure during the early run
of the system. There could be defective diodes or other electronic components used in the production since components are normally
received through bulk purchase which may not have 100% perfect items. The system is normally subjected to a period of run-time referred to
as ‘Burn-in’ which is meant to identify such manufacturer’s defective components which would be removed thereby reducing the failure rate
of the system. The presence of these components leads to a significant failure rate which reduces speedily during the early life period of the
system.
During the normal operating period of the system, referred to as its Useful Life, the failure rate is a constant given as λ. This status is
maintained for the period shown until the increased ageing period where the normal wear and tear of the system leads to a rise in the failure
rate. This period is referred to as the ‘Wear Out’ period in the curve of failure rate.

Returning to the derivation let us assume operation of the components is in the useful life period with a constant failure rate λ, hence let

Z(t) = λ ...1.6

Hence we obtain

S ( t ) N −F(t ) F (t )
R ( t )= = =1−
N N N

giving
dR(t ) −1 dF (t)
=
dt N dt

Or

dF (t) dR(t)
=−N
dt dt
...1.7
Substituting Eqns. (1.6) and (1.7) into (1.5), we obtain

−N dR(t )
λ= .
S (t) dt

−1 dR(t )
λ= .
R (t ) dt
Noting that R(t)=S(t)/N
i.e.
−dR(t)
λdt =
R (t )

Integrating we obtain,
t t R (t )
dR (t) dR(t)
λ ∫ dt=−∫ =∫
0 0 R(t) 1 R(t )

Noting that at t=0, S(t) = N, giving R(t) = S(t)/N = 1 and at time t, Reliability is R(t), hence the limits of integration

i.e.
λ [t ]t0 =−log e [R (t ) ]1R(t )

Simplifying further we obtain

λt =−[ log e R ( t ) −log e 1 ]


i.e.
−λt=log e R(t)

Giving
R ( t )=exp ⁡( −λt )
..1.8

This is the expression defining the exponential relationship between the Reliability and Failure Rate, where λ is usually expressed as
percentage failures per 1000 hours or as failures per hour.

( λt )2 ( λt )3
Now R(t) = 1 - λt + − +…
2! 3!
For λt very small ( < 0.02), we can write

R(t) = 1- λt ..1.9

For a system containing k types of components, each with a failure rate, λi, where i = 1, 2, 3...., k, the overall system failure rate is given by
k
λ overall =∑ N i λ i
i=1
...1.10

1.2 Reliability and Mean (Average)Time Between Failure (MTBF)

The average time a system will run or operate before failures occur is referred to as the Mean Time Between Failure, MTBF. This is a more
useful figure of merit to define a system. It has units of hours and is defined as follows :

MTBF=∫ R ( t ) dt
0
...1.11
This can be seen as the area under the Reliability curve with respect to time, t.
i.e.

MTBF=∫ exp ⁡(−λt) dt
0

1 ∞ 1
= ¿− [exp−λt ]0 =
λ λ
...1.12
i.e the MTBF is the reciprocal of the Failure Rate of a system or component.

NB
If λ is the number of failures per hour then the MTBF has units of hours.

Example 1.1

Let the number of components in a system be 4000 and let the failure rate be given by 0.02% per 1000 hours. Calculate the MTBF of the
system.

Solution

From this data, N = 4000.

This implies that the average number of failures per hour is given by:

0.02 1
Nλ= . .4000=8 x 10−4 failures per hour
100 1000

The MTBF is therefore given as


1 1
MTBF= = =1250 hours ≅ 7 weeks
Nλ (failures per hour ) 8 x 10−4

Returning to the analysis we note that from Eqns (1.8) and (1.12), we obtain

R ( t )=exp ⁡( −λt )

−t
¿ exp ⁡( )
MTBF
...1.13
For t = MTBF, i,e the operating or constraint time equals the MTBF we obtain

R(t) = 36.8%.

i.e. a system with a MTBF of 100 hrs has only a 36.8% chance of running for 100 hrs without failure. In other words the system has a 63.2%
chance of failure in 100 hours.

This then explains the definition of Reliability as given in the BS 42000 Part 2 Specifications as the ability of an item to perform a required
function (without failure) under stated conditions for a stated period of time, where the item could be a component, an instrument or a
system.

Figure 1.2 provides a sketch of the graph of R(t) against time measured as a factor of the MTBF.

R(t)

1.0

0.5
0.36

t
MTBF 2MTBF 3MTBF
Figure 1.2. Reliability versus time in MTBF

Combining Eqns 1.9 and 1.12 we obtain

R(t) = 1- λt
t
¿ 1−
MTBF

This implies that


t
MTBF=
1−R(t )
...1.14

When single components that cannot be repaired make up a system are being discussed then the figure of merit used is the Mean Time To
Fail (MTTF) instead of the MTBF.

Example 1.2

A 1st generation computer contains 10,000 valves, each with a failure rate of 0.5% per 1000 hrs. Find the period of 99% reliability.

Solution

Let N = No. of valves = 10,000


Let λ = Failure rate = 0.5% per 1000 hours

i.e the individual failure rate is given by

0.005 −6
λ= =5 x 10 h ours
1000

We can define the Mean Time To Fail of the individual components especially when these cannot be repaired as the MTBF will apply to the
system as a whole.

We therefore have the MTTF of the valves as

1 1
MTTF= = =200,000 hours ≅ 23 years
λ 5 x 10−6

Although this seems quite impressive, if however the failure of any one valve renders the failure of the system then we obtain the following :

The overall failure rate λoverall = N λ = 104 x 5 x 10-6 = 5 x 10-2


We note that

R(t) = 1 - λovt

Hence we obtain the period of 99% system Reliability as

1−R(t ) 1−0.99 0.01


t= = = ≅ 12 minutes
λov λ ov 5 x 10
−2

Also we note

1 1
MTBF= = =20 hours
λov 5 x 10−2

If the number of such components constituting the system is further increased then the MTBF is further decreased.

NB
1. The MTBF is an average value observed over a given period.
2. The term MTTF is usually applied to components that cannot be repaired while the term MTBF applies to systems or instruments
that can be repaired

Example

A maintenance engineer observes the following periods (in hours) between failures of four subsystems in a plant given in Table Q1.

Table for Example .


Subsystem T1(x103) T2(x103) T3(x103) T4(x103) MTBF Failure Rate
(#/hour)
1 40 40 55 60
2 40 30 45 50
3 25 25 45 40
4 30 35 40 45

1.3 Maintainability

The Maintainability of a system is the probability that a failed system will be restored to its working condition within a specified time. This
figure of merit can be seen as the probability of isolating and repairing a fault in a system within a given time.

If the rate of Repair is represented by μ, then we can define the Mean Time To Repair a system , MTTR as the reciprocal of μ, i.e.

1
MTTR=
μ
...1.15
We can therefore write the Maintainability M(t) as
−t
M (t )=1−exp (−μt ) =1−exp ⁡( )
MTTR
...1.16

where t is defined as the permissible time constraint for the maintenance action to take place, i.e. the time to restore the system.

This factor is really a function of the design of the equipment or system, the level of skilled workers, availability of spare parts and
appropriate tools and testing equipment.

This is the one of three figures of merit of systems referred to as the RAM factors, i.e. Reliability, Availability and Maintainability. The
Maintainability of a system can be viewed as a measure of the time required to restore a certain percentage of all system failures or the
probability (P) of restoring a failed system to its working condition within a time T. The Availability will be discussed shortly. The
Maintainability relationship is depicted in Fig.1.3

Fig. 1.3. P(t) versus t(Time to repair)

Source : www.reliabilityanalytics.com. Posted 3 September 2011. Move cursor to the image for the full link address.

M(t) versus R(t)

There are a number of comparisons between R(t) and M(t). Let us now consider these factors as shown in Table 1.0
Table 1.0. Comparison of R(t) and M(t)
Reliability Maintainability Comments
R ( t )=exp ⁡( −λt ) M (t )=1−exp (−μt ) M(t) = the probability
of successfully
completing a repair
job within a time t.
This formula is similar
to the Probability of
failure Q(t) or system
Unreliability
discussed earlier
where Q(t) = 1- R(t)
and µ replaces λ in the
Reliability expression.
t= time to failure t = time to restore
λ = failure rate µ = repair rate

1.4. Maintainable Systems

In general, maintainable systems must have a predictable MTTR for various fault conditions. The system’s repair time consists of

a. A passive repair time and


b. An active repair time.

The passive repair time is determined by time taken by the engineers to travel to the site plus tea breaks, and any other stoppage time taken
during the maintenance of the equipment, T1
The Active Repair time is influenced by the system’s design and consists of the following:
(Elicit class contribution before reverting to notes)

I. T2, defined as the time between the occurrence of a failure and the system user or operator, becoming aware that the fault has
occurred.
II. T3, defines the time needed to detect the fault and isolate the components responsible
III. T4 is defined as the time required to replace the faulty component
IV. T5 is defined as the time needed to verify that the fault has been removed and the system is fully functioning
i.e.
System repair time = f(T1,T2,T3,T4,T5)
5
System repair time=∑ T i
1
...1.17
1.5 The Availability of a system
This is defined as the probability that the system will be functioning according to expectations at any time during its scheduled working
period.

Derivation

System up time
Availability=
Total functional period

System up time
¿
System up time + System downtime

System up time
¿
System up time +( No . of failures x MTTR)

System up time
¿
System up time +(System up time x λ x MTTR)
1
¿ ¿
1+ λ x MTTR ¿

MTBF
Availability=
MTBF+ MTTR
...1.18
Where
1
λ=
MTBF

Alternatively the Availability can be expressed in terms of the MTBF and System’s Mean Down Time (MDT) as follows.

The Mean Up Time of a system is the Mean Time Between failure = MTBF
The Mean Down Time is represented by MDT
Hence the Availability of the system can be written as :

System up time
Availability=
Total functional period

System up time
¿
System up time + System downtime

MTBF
¿
MTBF + MTD

NB.
The MTTR may not always be the same as the MDT for the following reasons:

1. The Failure may not have been detected at the time it occurred.
2. The repair work may not start soon after it was reported
3. The equipment may not be put in operation immediately after the repairs have been carried out.

The Availability can only be accurately determined when the true Mean Down Time of the system is factored into the equation. Using the
MTTR assumes that the system is repaired immediately it becomes faulty. Whichever expression is used, the true time for which the system
is not in use or unavailable must be used in calculating the Availability.

Table 1.0 Scenarios on Equation 1.18

Conditions Conditions Availabilit MTBF


Comments Availability=
of MTBF on MTTR y MTBF+ MTTR

Low Low Low Not too desirable, as faults occur frequently even though they
are repaired relatively quickly. This will impact on the
integrity and reliability of the system
Low High Low Very undesirable. Here the faults occur frequently and the
time to repair is high. The system down time is therefore
high.
High Low High This is the most desirable as the failure rate is low and the
faults are cleared quickly when they appear, hence the
Availability is high.
High High An MTBF that is high implies a failure rate that is low, i.e.
faults occur rarely. However, with a high average time to
repair, the system down time when faults do occur will be
high, even though they seldom occur. This will impact
negatively on the system’s availability. This is not very
desirable, though a better option than Cases 1and 2.

Assignment 1.0

By referring to Table 1.0 above and practical examples discuss the implications of combinations of levels of MTTR and MTBF in
determining the examples’ Availability values.

MODE AND DEADLINE FOR SUBMISSION OF


ASSIGNMENT Soft copy due One week from date of posting

You might also like