You are on page 1of 12

Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

www.elsevier.com/locate/soildyn

Common problems in automatic digitization of


strong motion accelerograms
M.D. Trifunac*, V.W. Lee, M.I. Todorovska
Civil Engineering Department, University of Southern California, Los Angeles, CA 90089-2531, USA

Abstract
Common problems encountered in automatic digitization of strong motion accelerograms, recorded on film, are presented and discussed.
These include synchronization of the time scale for the three components of motion, non-uniform film speed, trace following in case of
scratches or trace crossings, distortions from high contrast preprocessing of the scanned image, and trace “rotation” resulting from rotated
position of the scanned film record. Procedures for correcting or eliminating these problems are suggested. The image processing hardware
has developed so much during the past 20 years, that at present it exceeds the technical requirements for processing strong motion
accelerograms. The problems described in this paper result from lack of training of the operators and lack of quality control in the process,
which still seems to be esoteric and highly specialized. This situation may have been caused by the low demand by the engineering profession
for high quality and large volume of strong motion data. q 1999 Elsevier Science Ltd. All rights reserved.
Keywords: Automatic digitization of accelerograms; Accelerogram data processing; Strong motion data; Accelerogram image processing

1. Introduction capabilities, there have been dramatic improvements in


CPU speed and in hard disk capacity. Even more dramatic
The hardware and software for automatic digitization of has been the reduction of cost of the hardware system and
strong motion accelerograms benefited from the rapid reduction of operator time since 1970s.
growth of the image processing technology since the Judging by the current trends in computer and informa-
1970s. Modern scanners were first used to scan large tion technologies, the possibilities for further improvements
number of accelerograms in the late 1970s. The first system of the digitization system and for new developments are
ran on a Data General NOVA computer and used OPRO- essentially unlimited. However, advanced hardware and
NICS rotating drum scanner. Although the resolution of this software alone do not guarantee high quality of the digitized
scanner was 12.5 × 12.5 mm (2032 dpi), it was actually data. Critical for high quality processed data remains to be
operated at a four times smaller resolution of 50 × 50 mm an experienced operator and rigorous quality control. We
(508 dpi) which was found to be optimal for this type of have found recently many errors in commercially processed
application. The cost of hardware for this system in 1977 records of the 1994 Northridge earthquake and one after-
was US$180 000 [1]. About 10 years later, personal compu- shock, which we believe are due to operator inexperience,
ters (PCs) and inexpensive flat-bed scanners (at first, HP false expectation that the software will perform all the tasks
ScanJet II Plus, with 300 dpi optical resolution) appeared, perfectly and automatically, and lack of quality control.
and the automatic digitization software was modified to Errors caused by lack of operator knowledge of the empiri-
work on a PC (first on IBM PC AT; [2]). Today this system cal procedures in the automatic trace following software,
is running on a Pentium II PC with HP ScanJet 4c (600 dpi bad operator judgement and lack of quality control can be
optical resolution). The cost of the most modern hardware so severe that the quality of data from automatic digitizers
system is under US$5000. Fig. 1 illustrates the progress in may be inferior even to the old hand digitized accelerograms
hardware capabilities (scanner resolution, CPU speed, capa- [3]. Errors of such nature may be very common in digitized
city of hard disk storage, typical scanning time, typical strong motion data worldwide, due to the large volume of
operator time, and cost of one system). In the hardware recorded analog data, and pressure to reduce the backlogs.
Empirical modeling of strong ground motion amplitudes
and studies aiming to interpret the physical nature of strong
* Corresponding author. Tel.: 1 1-213-740-0570; fax: 1 1-213-744-
1426. motion depend on the availability of a large number of
E-mail address: trifunac@usc.edu (M.D. Trifunac) uniformly and accurately processed accelerograms. The
0267-7261/99/$ - see front matter q 1999 Elsevier Science Ltd. All rights reserved.
PII: S0267-726 1(99)00018-4
520 M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

algorithms cannot be implemented [4–7]. Transducer char-


acteristics are not calibrated periodically in field conditions;
in many instances, nominal values are used, or 10–20 years
old values supplied by the manufacturer [8]. Cross-referen-
cing data files with detailed information on the earthquake
source and recording site characteristics is often not avail-
able or is incomplete.
The purpose of this paper is to point out some common
problems encountered in automatic digitization of acceler-
ograms, to describe their origin and how to avoid them, and
to illustrate their appearance in digitized records. The
intended audience are both those that produce and those
that use strong motion data, the former to recognize and
avoid these problems, and the latter to be aware of these
problems in analyzing the data. The presented examples
come from digitization of real strong motion accelerograms,
and are typical of errors we found in commercially digitized
recordings of the 17 January, 1994, Northridge, California,
earthquake …ML ˆ 6:4† and its 20 March, 1994, aftershock at
the Van Norman complex of the Los Angeles Department of
Water and Power [9]. We redigitized these records for
compatibility with records of the same earthquake in this
area, which we digitized earlier. We will refer to the two
versions of a record as “old” and “new.” In this paper, we
Fig. 1. Evolution in the hardware capabilities for automatic digitization of
accelerograms, cost reduction and operator time. compare the two versions and try to explain the differences.
For better understanding of the discussed problems, we
describe briefly our system for automatic digitization and
amount of data for such analyses is growing, as more instru- the principles around which the software was designed, and
ments are deployed and more earthquakes are recorded. For refer the readers to references which offer more detail. We
example, we had 186 uniformly processed three-component could find no detailed description of the system used to
accelerograms in the mid-1970s, about 550 in the early process the “old” versions, in professional journals or in
1980s, and about 2000 in the early 1990s. With such a technical reports. So, in many instances we only guess
large number of records, digitized and processed by many how the “old” version was digitized.
different organizations, it is becoming increasingly difficult
to control the quality and uniformity of the data. At present,
the digitized strong motion data worldwide is of non- 2. Automatic digitization procedures
uniform quality and often lacks complete supporting docu-
mentation. Recording strong motion is a long term process, The main tasks in automatic digitization of accelerograms
changes in personnel responsible for network operation and are: (1) converting the film image into a digital bitmap with
data processing are inevitable, and it is often impossible to a defined binary or gray-scale level for each pixel (corre-
recover the original information. Unless this practice sponding to the optical density of the film record); (2) creat-
changes soon, it will become difficult to recover and ing line segments out of the bitmap by automatic trace
preserve data it took so long to record, and to carry out following (given a threshold gray level and minimum
large scale regression analyses and many new specialized trace width in pixels, specified by the operator); (3) inter-
studies. active editing of the set of line segments created automati-
Contrary to the popular perception, the bottleneck in data cally by the computer programs performing the previous
processing and dissemination is not in the digitization of task; and (4) trace concatenation, in case of long records
analog records, but in verification, quality control, inclusion and several separately scanned pages, and writing the trace
in a database, and the lack of clearly defined priorities. coordinates (in the scanner coordinate system) and other
Groups which gather, process and disseminate data are information (e.g. scanning resolution and trace type) into
rarely involved in large scale data analyses, and so do not a disk file, in format recognizable by the software for further
have first hand experience of what constitutes complete and processing (involving instrument and baseline corrections,
necessary data. Thus the supporting information of [10,11]). These tasks are performed by four computer
processed accelerograms is often incomplete. For example, programs with respective generic names: Film, Trace, TV
instrument characteristics are rarely supplied in adequate and Scribe [1]. These programs have evolved with the
detail, so that more advanced instrument correction evolution of hardware. Numerous major changes and
M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530 521

used by LeTrace. Most of the modifications of this program


have aimed to help the operator make decisions and ease the
editing process as much as possible. Most of the examples
discussed in the next section (Section 3) of this paper have
been extracted from various operations performed by this
program. The last program, LeScribe, can be executed in an
automatic or interactive mode. The former is used for short,
one page, film records. The latter is used for multiple page
records and allows operator intervention and visual control
Fig. 2. First 10 s of the 1994 Northridge earthquake accelerogram at of joining the trace segments from different pages.
Sylmar Converter Station-East, Valve Hall (ground floor). Fig. 2 shows an example of a film record (the first 10 s of
an accelerogram record of the Northridge earthquake at the
improvements have been added in the late 1980s [2] and Van Norman Complex). It has three acceleration traces (L, V
again in 1995/96, but the sequence of tasks has not changed. and T), one baseline (B) and two two-pulse-per-second
Our current software package, LeAuto, consists of (2PPS) traces (the segmented lines at the top and bottom).
computer programs LeFilm, LeTrace, LeTV and LeScribe. It is seen that the T-trace intersects at many places with the
LeFilm is interactive, menu driven program which drives baseline. Significant operator intervention and use of a
the scanner to produce a gray-scale bitmap image of the film special automatic digitization software (employing dynamic
record (256 levels of gray is the default mode; the other two optimization of threshold levels and monitoring the top or
modes are binary black-and-white and gray-scale at 16 bottom edges of a trace) had to be employed to digitize this
levels of gray). The operator interactively chooses a mini- trace.
mum size rectangular window containing all the necessary
information to be saved for further processing, and an opti-
mum threshold level and minimum trace width for the auto- 3. Common problems encountered in automatic
matic trace following by the next program. The operator digitization
also defines the starting points for each trace and trace
type. Program LeTrace performs the trace following and 3.1. Delays in digitized time series for different components
is entirely automatic. Program LeTV is entirely interactive of motion
and menu driven. The trace editing is the most time consum-
ing part of the procedure. The program efficiency and the The position of the first digitized point of an acceleration
quality of the output depend on the operator training and trace determines the origin of its time coordinates. Because
experience. The operator intervention consists of deletion of of the common trigger mechanism, all the three traces (L, V
line segments resulting from imperfections of the film and T) start to be recorded simultaneously, but if the first
record (scratches and dirt particles), joining consecutive point is not digitized properly, there will be a time delay in
segments of a particular trace, adding and deleting points the digitized time series. Synchronization of the origin time
in a line segment, and also completely redigitizing selected and of the running time scales for all the recorded compo-
portions of traces at a threshold level different from the one nents in a film record is crucial for many applications, for

Fig. 3. The 1994 Northridge earthquake accelerogram at Sylmar Converter Station-East (free-field): (a) first 2 s; and (b) beginning of the vertical trace.
522 M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

due to the fact that the traces are weak immediately after
trigger (while the light bulb warms up; during ,0.1 s), and
the position of the first digitized point changes with different
choice of threshold level. This is illustrated in Figs. 3 and 4.
Part (a) of Fig. 3 shows the first 2 s of a film record and part
(b) shows an enlargement of the beginning of the V-trace.
Fig. 4 (an illustration of an image in LeTV) shows the
position of the first digitized point and the bitmap (raw
digitized data) for different choices of threshold levels of
gray (between 200 and 240). It is seen that for lower thresh-
old level, the trace width increases and the trace beginning
“moves” to the left, while the opposite is true when the
threshold level is increased. This situation is further compli-
cated by the fact that the trace “darkness” on the film, trans-
lated to trace “thickness” after scanning, is in general
different for each trace on the same film, and depends not
only on time after trigger but also on the amplitude of
recorded motions (larger amplitude results in “lighter”, i.e.
Fig. 4. Illustration of a bitmap image of a trace beginning for different “thinner” trace for any given constant threshold level). The
threshold levels (200, 215 and 235), and the corresponding outcomes of
automatic digitization.
examples in Figs. 3 and 4 are from a record of the 1994
Northridge earthquake; gray threshold levels of 180–240
were common in digitization of the Northridge records.
example in inverse analyses of the earthquake source Because of the above, choosing one common threshold
mechanism (“random” time delays in the digitized data level for the entire accelerogram (which was the default
affect the numerical stability of the inversion), applications procedure in the old versions of program LeTrace) will
that require linear combination of the recorded components result in variations of the origin time by at least several
of motion, such as computation of the radial and transfer pixels. Other parameters specified for the trace following,
components of motion (affect the accuracy of peak ampli- e.g. minimum trace width (usually selected as 2–4 pixels for
tudes in the rotated directions; [12]), in analyses of building digitization with 600 dpi) will also contribute errors to the
response from wave propagation viewpoint, or in correction selection of the first point, and the overall uncertainty of the
of accelerograms for cross-axis sensitivity and transducer origin of the time axes can exceed 0.01–0.02 s (2–5 pixels
misalignment (these corrections are meaningless unless for 600 dpi digitization resolution). One way to reduce these
the three components are synchronized; [6–8,13]). errors is intervention by the operator who can choose
The difficulty in selecting the first point to be digitized is manually the first point, but this is time consuming and is

Fig. 5. The vertical trace of the 1994 Northridge earthquake accelerogram at Sylmar Converter Station-Valve Group 1–6 (basement): (a) the beginning; (b) the
beginning enlarged by macro-photography, showing gradual build up of optical density; (c) the beginning enlarged by a repeated Xerox process and the “old”
digitized version of the record. The dashed line shows the “old” digitization.
M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530 523

Fig. 6. Beginning of the accelerogram of the 20 March, 1994, Northridge aftershock at Sylmar Converter Station, Valve Group 7: (a) ground floor; and (b) free-
field site. Two digitized versions, the “old” and the “new”, are shown (by a dashed and a solid line), and the time delay between the two.

subjective. To increase the efficiency of this step and to operator intervention. An example of a circumstance that
eliminate operators subjectivity, since 1994 we have used a can complicate this task is when the gap between the end of
new special purpose algorithm which determines automati- the previous record and the onset of the current record is too
cally the “best” starting point for each acceleration trace. short or the traces overlap, so that they appear as continuous
This algorithm was tested on hundreds of accelerograms, on the scanned image. When such traces are displaced (i.e.
and was found to be successful (error less than one pixel) the trace starts with a large amplitude), this problem is
in about 95% of cases). Difficult cases would still require eliminated. An illustration of such “overlap” is shown in
Fig. 5. The dashed line in part (c) is an example of an
inaccurately digitized trace (0.05 s was omitted in the begin-
ning).
Fig. 6 shows an example of raw data images and inaccu-
rately (arbitrarily) digitized trace beginnings (the dashed
line, referred to as “old” digitization) and trace beginning
digitized with the help of the new features of program
LeFilm, LeTV and LeTrace. In this example the magnitude
of the error for the “old” digitization is difficult to interpret,
because of the lack of apparent reasoning that guided the
operator. For example, it is not clear why the “old” digitiza-
tion of the L trace in Fig. 6(b) starts early with a “ramp” of
, 0.05 s, in a manner not related to the raw data. Parts (a)
and (b) show, respectively, motions recorded at the ground
floor of a structure and in the free-field (approximately 25 m
away towards southwest), both recorded by the same multi-
channel recorder (CR-1), on the same film and with
common trigger time. Such records are invaluable for soil-
structure interaction studies, and for analyses of differential
motions between the two closely spaced points. Errors as
those seen in Fig. 6 (in the origin times of different compo-
nents) make the “old” digitization useless for detailed
studies, and misleading for an unsuspecting user. “Conclu-
sions” that could result from such data would be, e.g. that
the high frequency strong motion amplitudes are not corre-
lated at separation distance of 25 m, that the two sites have
different soil properties (i.e. different wave velocities), that
Fig. 7. Same as Fig. 6 but for the accelerograms at Los Angeles Dam, West
Abutment. The difference between the two digitized versions in choosing the wave train approached the two stations from a different
the beginning is significant. In the “old” version, the beginning seems to be direction (i.e. different phase velocities) and so on. Another
chosen arbitrarily. example is shown in Fig. 7 (Los Angeles dam, right abutment
524 M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

record of the Northridge earthquake). It is seen that, in the


“old” digitization (dashed line), the operator started to digi-
tize the L, V and T traces 0.135, 0.190 and 0.175 s after
trigger.
Many such errors go undetected, because their amplitudes
are small enough that visual comparison of raw data output
from digitization with original film cannot detect the discre-
pancies, and because hardly any records happen to be digi-
tized more than once, independently, allowing such
comparisons. The errors illustrated in Fig. 6(a) and (b)
and in particular the errors shown in Fig. 7 show negligence
and lack of quality control. The cumulative investment of
time and of resources in the analysis of the recorded motions
and of their effects on different structures is significant. In
the light of these conditions the above described errors
cannot be tolerated.

3.2. Trace following

3.2.1. Non-uniform optical density


Ideally, the traces on the film should be thin, sharp and
uniformly dark, resulting in a scanned trace image which is
Fig. 8. Appearance of the scanned image of the record and the outcome of
continuous. In reality, the optical density of the trace varies
automatic trace following for different choices of threshold level. This
example corresponds to an image scanned at 600 dpi resolution with 256 with the trace amplitude, and there are “random” variations
level gray scale. Notice the spurious peak in the digitized data due to noise resulting from imperfections of the instrument or handling
in the scanned image for lower threshold level (190). For higher threshold of the film. The scanned trace image is, in general, lighter
levels (e.g. 230), the trace image becomes discontinuous. For clarity, the for larger trace amplitudes, but at the very peak it is darker.
horizontal scale is stretched three times relative to the vertical.
The reason for the former is that, for uniform film speed and
light bulb intensity, the light beam moves faster and travels
longer distances as it exposes the film, and for the latter is
that the trace segments just below the peak partially overlap
due to finite trace thickness, and the vertical trace velocity is
zero at the peak, so that the light beam exposes the film
longer. Typical imperfections include scratches or dust,
poor focussing of the traces, fogged or dirty mirrors and
lenses, variable darkness of the film background, etc. and
result in “random” fluctuations of trace darkness.
The variations in darkness (optical density) of the film
image result in variations of gray levels of the scanned
pixels. The “width” of the scanned trace image depends
on the chosen threshold level, and can be controlled by
choosing this level. Pixels with smaller gray level than the
threshold will be considered as “white” (background) in the
case of dark traces on white or transparent background.
Traces are “wider” for lower threshold level, and become
discontinuous for sufficiently high threshold level. Fig. 8
illustrates these variations. It shows a bitmap image of an
acceleration trace for three threshold levels of gray, 190,
220 and 230 (600 dpi scanning resolution, 256 levels of
gray). It is seen that, for threshold level 230, trace bitmap
image is narrow and discontinuous. For threshold level 190,
near the acceleration peaks, the trace bitmap image is wide
and continuous, and has rough boundaries (to the right of the
Fig. 9. Beginning of the vertical trace of the 1994 Northridge at Sylmar negative peak). The rough boundaries are caused by random
Converter Station-East (free-field), and two digitized versions of the record. variations in the trace and background optical densities, and
The “old” version is delayed by 0.053 s, and has a spurious peak near 0.4 s. may lead to spurious peaks in the digitized data.
M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530 525

traces are thinner but continuous. If the bitmap trace image


is discontinuous, not enough measured information is
included in the estimation, and if threshold level is too
low (thick trace) the signal is estimated from measurements
which are too noisy. The “optimum” threshold level for the
illustration in Fig. 8 appears to be near 220. However, this
level will not be the optimal one along the entire length of
the record, because the trace darkness on the film, and
consequently the thickness of the scanned trace image,
depend on the signal amplitude. The operator adopts a
threshold level for automatic trace following which is “opti-
mal” for most parts of the signal, and in the finishing phase
corrects the detectable errors due to too high or too low
threshold levels.
Fig. 9 shows another example of a spurious peak. Such
peaks are difficult to detect, because many local acceleration
peaks do have similar “double” peaks. These spurious peaks
can be detected only by detailed and time consuming visual
inspection of all digitized signals versus the scanned image
(raw data) for various threshold levels.
A related type of a problem occurs when there is a
“white” spot in the middle of the scanned trace image, illu-
strated in Fig. 10(a). The automatic trace following algo-
rithm will follow the wider branch of the trace, producing a
pulse like error in the estimated signal. A related and often
overlooked problem occurs when the trace has a one sided
Fig. 10. An illustration of “pulses” in the digitized data due to scratches on indentation (Fig. 10(b)) or when a dust particle, a dark
the film and nonuniform optical density of a trace.
scratch or a dark shadow crosses the trace (Fig. 10(c)).
Digitization errors illustrated in Fig. 10 are very common,
The estimated amplitude of the digitized signal (y-coor- difficult to detect and occur almost exclusively for data
dinate) by the automatic trace following algorithm, is at the scanned with 600 dpi resolution. Such errors were very
middle of the vertical cross-section of the bitmap trace rare for scanning with 300 dpi. For higher resolutions, e.g.
image, and depends on the chosen threshold level. The 1200 dpi, these errors will multiply and may be responsible
signal is estimated more accurately from the bitmap if the for most of the operators time in manual verification and
correcting the data in LeTV.
Intersection of an acceleration with a fixed trace
(“straight” line created by a fixed mirror) is usually not
difficult to interpret automatically, particularly if the trace
following algorithm can memorize, as well as control the
variations of the slope of the estimated signal based on prior
knowledge of the type of trace followed. However, situa-
tions when the baseline is unusually thick and the accelera-
tion trace has a peak near the baseline are difficult to handle
automatically, and the best approach is to check for such
problems and fix them manually in the editing phase of
program LeTV. Fig. 11 shows an example of two spurious
peaks in the digitized signal resulting from such a situation,
which have remained undetected in commercially digitized
record of the Northridge earthquake.

3.2.2. Smoothing of the high frequency peaks


Another common problem with automatic trace following
is “low-pass filtering” of high frequency peaks in the
Fig. 11. An illustration of errors in digitized data near intersecting accel- recorded signal. This is caused by merging and partial over-
eration and fixed traces: the transverse trace of the 1994 Northridge record lapping of the trace just below the peak, due to finite thick-
at Sylmar Converter Station-East, free-field, ,9.5 s after trigger. ness of the light beam exposing the film. In the immediate
526 M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

only by reducing the trace width on the film (by careful


focussing), and by increasing the film speed beyond 1 cm/s.
The contradictory requirements calling for lower thresh-
old level, so that the trace image is continuous, and for
higher threshold level, so that peak amplitudes are not
reduced, can be solved simultaneously if variable threshold
levels are used in automatic trace following. This requires
sophisticated and reliable software, and highly skilled
operators. For these reasons, we run this software only in
the manual mode in program LeTV. Such a software is
generally not available, and usually a compromise must be
made by choosing some intermediate threshold level. This
leads to (1) smoothing of the amplitudes of high frequency
peaks (“low-pass filtering”) where the chosen threshold
level is too low (Fig. 12), and (2) missing variations in the
signal where the chosen threshold level is too high (e.g.
ascending and descending line segments in the areas with
large peaks; see Figs. 5(c) and 12). The latter causes high
frequency large peak amplitudes to be connected by straight
lines because the darkness of the portion in between is
below the working threshold level. Fig. 12(b) illustrates
outcomes of automatic trace following for different thresh-
old levels in the region outlined in Fig. 12(a). It is seen that,
for threshold level ,200, the high frequency acceleration
(,20 Hz wave) is reduced, and it almost disappears for
threshold level 150 or lower. For threshold levels 200–
230, the trace image between the peaks is progressively
lost (due to too high working threshold level) and is “digi-
Fig. 12. (a) The “old” and “new” digitized versions near the beginning of
tized” as a straight line (e.g. see also “old” digitization in
the vertical trace of the 1994 Northridge record at Sylmar Converter Fig. 5(c), between 0.4 and 0.6 s).
Station, Valve Hall Floor. The difference between the two versions is
probably due to smoothing effects of the automatic trace following algo- 3.2.3. Distortions due to high-contrast preprocessing of the
rithm, more pronounced for lower threshold levels. (b) Illustration of differ- film image
ences in data digitized by the LeAuto software caused by different threshold
levels, for a portion of the trace shown in part (a). The image was scanned
Many of the above problems can be reduced or eliminated
with 600 dpi resolution and 256 level gray scale. by working with 256 levels of gray, versus 16 levels or black
and white image only. This, however, requires large capa-
city and fast access hard disk, as well as a fast CPU, and was
vicinity of the peak, the resulting trace thickness is difficult or impractical to implement in the past, when only
increased, but only on the side towards the zero-acceleration 286, 386 or 486 PCs were available. Simple, well recorded
line, and the amplitude of the estimated signal at the peak is accelerations can be processed successfully by using only
biased towards smaller values. This problem is pronounced 16 levels of gray (0—“white”; 15—“black”). However,
for vertical acceleration traces, which usually contain more digitizing a binary black-and-white scanned image is very
high frequencies than the horizontal accelerations, espe- difficult, will inevitably require extensive photographic or
cially in the near-field of moderate and large earthquakes. digital enhancement of contrast, and will lead to situations
High frequency, large vertical accelerations result in large which are very difficult to process correctly. Moderate
amplitude, rapidly oscillating traces with low optical enhancement of contrast can be useful (sometimes it is
density (e.g. see the vertical traces in Figs. 2 and 3(a)). essential), but it may also lead to complicated distortions
They require lower than average threshold levels for auto- of the traces, to the extent that is not acceptable. Such
matic trace following, otherwise the trace is not continuous. distortions of the bitmap are illustrated in Fig. 13. Part (a)
On the contrary, lowering the threshold level will increase shows a segment of a record of the 20 March, 1994, North-
the filtering of high frequencies, and to avoid this effect in ridge earthquake aftershock (M ˆ 5.3) copied directly from
portions of the signal with intermediate and small accelera- the film. In part (b), an enlarged portion of the vertical
tions (e.g. see horizontal acceleration traces in Figs. 2 and acceleration trace is shown, after successive Xeroxing of
3(a)), the threshold level should be as high as possible. It the record (equivalent to multiple high-contrast processing
should be mentioned here that this problem cannot be solved of the film image). It can be seen that this process introduced
by reducing the pixel size. Its consequences can be diminished characteristic elongation of peak amplitudes (see also the
M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530 527

Fig. 13. (a) Segment of the 20 March aftershock record at Sylmar Converter Station-East, free-field, scanned from film. (b) An enlarged portion of the vertical
trace after repeated Xeroxing, showing associated distortions. Also the “old” and “new” versions of digitized data are shown. The differences may be due to
high contrast preprocessing of the scanned image in producing the “old” version.

vertical traces in Fig. 6(a) and (b) in which the film image two-pulses-per-second signal (2PPS), recorded along the
was enlarged by successive Xeroxing). Fig. 13(b) shows top and bottom edges of the film or paper. About 20 years
comparison of two independent digitizations of this segment ago, one of these relays was converted to work with a local
which differ significantly near the peaks. It appears that the clock which produces a binary code of the Julian day, hour,
“old” digitized version has distortions near peaks of the minute and second every 10 s. At first, it was believed that
same type as the repeatedly Xeroxed image shown below. absolute trigger time was not necessary for recorded strong
The “new” version, digitized by the LeAuto software ground motion [14], but since it was first introduced in the
(600 dpi resolution and 256 levels of gray), shows no such early 1970s [15], it opened many new possibilities for
distortions. We do not know the exact reason for the distor- advanced wave propagation studies in strong motion seis-
tion in the “old” version of this record, and we have never mology.
encountered such distortions in our digitization of strong The 2PPS signal accuracy is believed to be ,1%, but it is
motion records. We speculate that this record was digitized rarely calibrated. Since the early 1970s, we have assumed
after excessive high-contrast enhancement of the bitmap that the time coordinate of analog records is scaled more
image by software, or by a lithographic photo process accurately using 2PPS pulses (generated electronically)
prior to scanning. These types of distortions should have rather than by the nominal speed of the film (driven
been detected in the quality control phase of the job. mechanically), and have used it to correct for minor varia-
tions in the film speed [1–3]. Exceptions are records for
3.3. Non-uniform film speed which the 2PPS and absolute clock relays malfunctioned
simultaneously (e.g. [16]). Figs. 7, 9, 12(a) and 13 show
The nominal film speed for typical strong motion accel- differences in the digitized accelerations when the time is
erographs is 1 cm/s (SMA-1, CR-1). For the M02 accelero- scaled using the digitized 2PPS signal (“new”) and when we
graph (New Zealand) which records on a 35 mm film, the assume uniform film speed was used (“old”). The difference
actual film speed is 1.5 cm/s, and this is equivalent to 3 cm/s is manifested by time dependent delays between the two
for a 70 mm film (SMA-1). Increasing the film speed records.
improves the resolution and accuracy of digitization of Occasionally the film speed may experience abrupt
high frequency accelerations not only in time, but also in changes and stalls. This is caused by friction in the film
amplitude (see the discussion on the low-pass filtering of driving mechanism, friction in the film cassette or by faulty
high frequency and large amplitude accelerations in Section motors, and in general cannot be corrected uniquely. Dura-
3.2). The old AR-240 accelerograph (which recorded the tion of short stalls can be estimated by measuring the short-
Pacoima Dam accelerogram during the 1971 San Fernando, ening of distance between consecutive pulses of the 2PPS
California, earthquake) had a recording speed of 2 cm/s signal. Approximate corrections of the digitized data then
(equivalent to 0.5 cm/s on a 70 mm film; [14]). can be performed by inserting “gaps” into the scanned
Most instruments have one or two relays which produce a bitmap image, and recreating manually the missing portion
528 M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

Fig. 14. Illustration of a record imperfectly aligned with the scanner. There is an angle a between the coordinate system of the scanner (XSCAN –YSCAN) and the
accelerogram coordinate system (xREC –yREC). The two fiducial marks can be used to evaluate a and correct for it.

of traces [17]. A description of processing an accelerogram this angle is by marking and digitizing at least one pair of
with many stalls can be found in [16]. fiducial points onto one of the fixed traces, evaluating a
from the position of these marks, and rotating back the
3.4. Rotation of the digitized traces digitized signals for the same angle. The accuracy of this
correction is sufficient for typical acceleration records. It
Because the record is digitized directly from the scanned may be limited by the noise in the scanned image of the
image, it is important that the film or paper original is well baseline only for very short records (less than 5–10 s long).
aligned with the scanner. Perfect alignment would mean that While marking and digitizing fiducial points is esential for
the time axis of the record is parallel to one of the scanner long records that require scanning multiple pages (the fedu-
axes. In reality, however, this is difficult to achieve, and the cial marks are used to match the independently digitized
two are off by some small but finite angle, a . The problem pages), it is often neglected by some operators for one
of imperfect alignment is one of the oldest problems recog- page records, and in those cases correction for the angle a
nized in digitization and processing of strong motion accel- is not performed. The associated distortions for a real record
erograms [10]. By careful placement of the original onto the are illustrated in Fig. 15. It shows a segment of the trans-
scanner or the digitizing table, angle a can be kept small, verse acceleration trace of the record in Fig. 14. The “old”
i.e. of the order of 18 (Fig. 14). One simple way to correct for digitized trace appears to be rotated clockwise relative to the
“new” trace by a , 0:98. This is most obvious near the
positive and negative peaks. Assuming that the scanners
used for the “old” and “new” digitized versions were both
accurate, and knowing that the “new” version was corrected
for a rotation a , the difference between the “old” and “new”
versions could be explained as an error in the “old” version
due to imperfect alignment of the original record with the
scanner (a , 18). If this record was photographically
enlarged before scanning, the observed rotation could also
be explained as distortion caused by imperfections of the
lens, or by non-parallel planes of the negative and of the
projected image.

4. Discussion and conclusions

Selected most common problems encountered in auto-


matic digitization of accelerograms recorded on film were
illustrated and discussed. This was preceded by a brief
review of the changes in hardware and software capabilities
since the 1970s. Differences between two digitized versions
Fig. 15. Differences in the “old” and “new” versions of a digitized record
showing relative rotation by an angle a (see Fig. 14). The “new” version of the same original records were discussed and explained in
was corrected for this angle with the help of the fiducial marks shown in most of the presented examples (the “old” version was digi-
Fig. 14. tized commercially for the LADWP in 1994, and the “new”
M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530 529

version was digitized by the authors of this paper using the (high frequency errors) due to imperfections on the film,
LeAuto software package). The problems discussed include such as scratches and dust (these imperfections are “not
errors in identifying the origin of common time coordinates seen” or are “smoothed out” by larger pixels, for lower
for the three recorded components of motion, smoothing of scanning resolutions e.g. 300 dpi).
the high frequency peaks as well as of amplitude variations We conclude that the governing factor for high quality
between peaks (both in segments of the record with moder- digitized accelerograms is rigorous quality control and
ate and large signal amplitudes), spurious peaks and noise in experienced and conscientious operator. Intelligent software
the digitized signal due to imperfections of the film image or as well as the increased hardware capabilities significantly
specific circumstances (e.g. intersection of traces), possible speed up the process and reduce the labor cost. However, no
distortions due to high contrast preprocessing of the scanned software will take care of all of the difficulties. Continuous
image, scaling of the time coordinate and rotation of the and systematic quality control by the operator, at all the
digitized signal. The errors due to all of these problems phases of the process, will always remain the determining
can be either eliminated or significantly reduced, automati- factor for the quality of the end product.
cally by intelligent algorithms (performing accurate identi-
fication of trigger times for each trace individually and
Acknowledgements
correction for common time scale fluctuations) or manually
by operator intervention.
The authors thank Ron Tognazzini and Craig Davis of the
The most important factor determining the quality of the
Los Angeles Department of Water and Power for making
processed data is the experience of the operator and rigorous
available the original films and the commercially processed
quality control of the outcome of the automatic trace follow-
data of the 1994 Northridge earthquake records, used to
ing. An experienced operator can avoid most of these
illustrate the problems discussed in this paper.
problems by carefully using the software. Finally, critical
comments and suggestions by experienced operators are
invaluable for further improvement of the automatic digiti- References
zation software. The efficiency and reliability of this soft-
ware can be improved only by close interaction between the [1] Trifunac MD, Lee VW. Automatic digitization and processing of
programmer and an experienced operator. strong motion accelerograms: Parts I and II. Department of Civil
The digitization of accelerograms recorded on film can be Engineering Report No. 79-15, University of Southern California,
Los Angeles, California, 1979.
viewed as an estimation problem using noisy measurements, [2] Lee VW, Trifunac MD. Automatic digitization and processing of
and the result is nonunique. Too high threshold level results accelerograms using PC. Department of Civil Engineering Report
in exclusion of recorded information in the estimation No. 90-03, University of Southern California, Los Angeles, Califor-
process, and too low threshold level results in including nia, 1990.
too much noise in the estimation process. We showed that [3] Trifunac MD, Lee, VW. Routine computer processing of strong
motion accelerograms. Report EERL 73-03, California Institute of
systematic biases in the estimate due to this “noise” (e.g. Technology, Pasadena, California, 1973.
spurious peaks and additional pulses, and smoothing of the [4] Novikova EI, Trifunac MD. Instrument correction for the coupled
sharp peaks) can be eliminated or at least abated by careful transducer–galvanometer system. Department of Civil Engineering
choice of the threshold level. Scanning the image with a 256 Report No. 91-02, University of Southern California, Los Angeles,
level gray-scale is highly recommended, while digitization California, 1991.
[5] Novikova EI, Trifunac MD. Digital instrument response correction
from a binary black-and-white bitmap image is discouraged. for the force balance accelerometer. Earthquake Spectra
A major improvement of the software for automatic trace 1992;8(3):429–442.
following would be a self learning algorithm with adaptive [6] Todorovska, MI, Novikova EI, Trifunac MD, Ivanović SS. Correction
threshold level. Such an algorithm has been incorporated in for misalignment and cross axis sensitivity of strong earthquake
the editing phase of our procedures (LeTV). Within this motion recorded by SMA-1 accelerographs. Department of Civil
Engineering Report No. 95-06, University of Southern California,
phase, the algorithm is used to digitize a trace segment in Los Angeles, California, 1995.
a window defined manually by the operator. Much work [7] Todorovska MI. Cross-axis sensitivity of accelerographs with pendu-
remains to be done to include this algorithm in the fully lum like transducers—mathematical model and the inverse problem.
automatic trace following procedure (LeTrace). Earthquake Engineering and Structural Dynamics 1998;27(10):1031–
There is a common misperception that the accuracy of 1051.
[8] Todorovska MI, Novikova EI, Trifunac MD, Ivanović SS. Advanced
estimating the high frequencies in a record will increase sensitivity calibration of the Los Angeles strong motion array. Earth-
significantly by scanning the film record at a higher resolu- quake Engineering and Structural Dynamics 1998;27(10):1053–
tion. This is true only up to a limit. For example, the 1068.
“smoothing” of the amplitudes of the sharp peaks can be [9] Lindvall–Richter–Benuska Associates. Processed LADWP power
eliminated only by better focussing of the light beam (possi- system strong motion records from the Northridge, California, Earth-
quake of 17 January, 1994. Report LRB 007-027, prepared for the Los
ble only up to a certain degree) or by increasing the film Angeles Department of Water and Power, Los Angeles, California,
speed. Our experience shows that, for higher scanning reso- 1995.
lution (600 dpi), the digitized signal detects more noise [10] Trifunac MD. Zero baseline correction of strong motion accelerograms.
530 M.D. Trifunac et al. / Soil Dynamics and Earthquake Engineering 18 (1999) 519–530

Bulletin of the Seismological Society of America 1971;61:1201– editor. Earthquake engineering, Englewood Cliffs, NJ: Prentice Hall,
1211. 1970.
[11] Trifunac MD. A note on correction of strong motion accelerograms [15] Dielman RJ, Hanks TC, Trifunac MD. An array of strong motion
for instrument response. Bulletin of the Seismological Society of accelerographs in Bear Valley, California. Bulletin of the Seismolo-
America 1972;62:401–409. gical Society of America 1975;65:1–12.
[12] Todorovska MI, Trifunac MD. Amplitudes, polarity and time of peaks [16] Trifunac MD, Todorovska MI, Lee VW. The Rinaldi strong motion
of strong ground motion during the Northridge, California Earth- accelerogram of the Northridge, California, Earthquake of 17 January,
quake. Soil Dynamics and Earthquake Engineering 1997;16(4): 1994. Earthquake Spectra 1998;14(1):225–239.
235–58. [17] Lee VW, Trifunac MD. Current developments in data processing of
[13] Wong HL, Trifunac MD. Effects of cross-axis sensitivity and misa- strong motion accelerograms, Department of Civil Engineering
lignment on response of mechanical optical accelerographs. Bulletin Report No. 84-01, University of Southern California, Los Angeles,
of the Seismological Society of America 1977;67:929–956. California, 1984.
[14] Hudson DE. Ground motion measurements Chapter 6. In: Wiegel RL,

You might also like