Professional Documents
Culture Documents
System
Sizing Guide
for all platforms
Release 1.1
November 2001
DOC3-SYSIZEGD-1101
h11KTKH1mTh1dd1d
Copyright 2000, 2001
Documentum, Inc.
6801 Koll Center Parkway
Pleasanton, CA 94566
All Rights Reserved.
Documentum
, Documentum Administrator,
Documentum Developer Studio, Documentum Web Development Kit,
Documentum WebCache, Documentum ContentCaster, AutoRender
Pro, Documentum iTeam, Documentum Reporting Gateway,
Documentum Content Personalization Services, Documentum Site Delivery
Services, Documentum Content Authentication Services, Documentum
DocControl Manager, Documentum Corrective Action Manager,
DocInput, Documentum DocViewer, Virtual Document Manager,
Docbasic
, Documentum WorkSpace
,
Documentum SmartSpace
resources.
Active User In
Transaction
A connected user who is currently waiting for a response to a
request from Documentum Server. An example is a user who
is logged into Desktop Client and waiting for the
Docbase View window to open. An active user in
transaction consumes the most Documentum Server
resources of any user state, including CPU, RAM, network
throughput, and disk throughput.
Active User Out of
Transaction
A connected user who is not currently waiting for a
Documentum Server response to a request. An example is a
user who is logged into WorkSpace and viewing a document
in a word processor. While the word processor displays the
document, the Documentum Server does not receive any
more requests from the user until the user is finished viewing.
An active user out of transaction consumes fewer
Documentum Server resources than an active user in
transaction, but more than an inactive user. RAM is the
primary resource that an active user out of transaction
consumes.
Overview of System Sizing
Terminology Used in This Guide
Documentum System Sizing Guide 15
Activity Time-out A Documentum Server feature that conserves server-side
resources. When a connected user has not made a request of
the Documentum Server within a specified time limit, the
server transparently frees the connection to free up unused
OS and DBMS resources. The next time the user makes a
request of the Documentum Server, the request is handled
automatically without requiring the user to login again. The
activity time-out counter is reset after each completed user
request of the Documentum Server.
Bandwidth Refer to Network Throughput.
Bottleneck A resource that limits performance. Examples are CPU, RAM,
network throughput and disk throughput.
Connected User A user who is currently logged into a Docbase.
Connecting User A user who has requested a Docbase connection but is not yet
logged in.
Database Server
(RDBMS instance)
SQL Relational Database Management System required as
part of the Documentum Docbase. Used to store
Documentum object attribute information.
Docbase The dynamic document and Web page repository accessed by
the Documentum Server. The Docbase stores a document or
Web page as an object that encapsulates the documents
native content together with its attributes, including
information about the documents relationships, associated
versions, renditions, formats, workflow, and security.
DocPage Server Documentum Server version 3.x
Documentum Server Software used to service incoming and outgoing document
management requests for data in the Docbase. Different
versions carry different product names:
s eContent Server refers to Documentum Server, version
4.2.
s e-Content Server refers to version 4.1.
s EDM Server refers to version 4.0.
s DocPage Server refers to version 3.x.
Table 1-1 Definition of Common Terms
Term Definition
Overview of System Sizing
Terminology Used in This Guide
16 Documentum System Sizing Guide
Disk Throughput Number of bytes per unit of time transferred to or from the
disk subsystem during read or write operations.
e-Content Server Documentum Server version 4.1
eContent Server Documentum Server version 4.2.
EDM Server Documentum Server version 4.0.
HTTP Server (Web
Server)
Software required to service HTTP requests by a Web browser
from a file system or from Documentum RightSite
.
Inactive User A connected user who has reached activity time-out. Inactive
users do not consume Documentum Server resources
Named User A user for whom a user profile is defined in the Docbase. Each
user profile is stored as a dm_user object in the Docbase.
Network Latency The delay in response to a network request due to the time it
takes for a byte of data to traverse the network and travel
from the client to the server and back again. Latency depends
on the distance between the client and server, how many
pieces of equipment are in between, and the types of
communication lines.
Network
Throughput
The number of bytes per unit of time that can flow between a
client and server. Throughput is also referred to as
bandwidth.
Physical Memory Total RAM dedicated to a physical computer system.
RightSite Server technology required to coordinate document
management requests between an HTTP server and the
Documentum eContent Server. This component is required
for all Documentum Web products including Intranet
Client and other web-content-based applications. RightSite
must be physically installed on the same host as the HTTP
server.
Table 1-1 Definition of Common Terms
Term Definition
Overview of System Sizing
Terminology Used in This Guide
Documentum System Sizing Guide 17
Transactions Requests from the client that have a response from the server.
The client must wait for the response before it can continue.
For example, the client might send Update the Check Out
field with my name and the server sends back Done.
Multiple application transactions can occur for specific user-
level functions
Transformation
Engine
The facility within Documentum Server that automatically
transforms content in one format to another format.
The transformation engine uses a supported converters to
perform the transformation. Through the transformation
engine, you can:
s Transform one word processing file format to another
word processing file format
s Transform one graphic image format to another graphic
image format
s Transform one kind of format to another kind of format
for example, changing a raster image format to a page
description language format
Some of the converters are supplied with the Documentum
system; others must be purchased separately.
User States Named users may be in one of the following activity states:
connected user, active user, or inactive user. Active users may
be divided into two categories: active user in transaction
and active user out of transaction. Understanding the
different user states can have a beneficial impact on system
sizing because they vary in resource consumption due to the
activity time-out feature. (For more information, refer to
User Connection States and Resource Consumption on
page 2-2.)
Virtual Memory A service provided by the operating system (and hardware)
that allows each process to operate as if it has exclusive access
to all memory (0 to 2, typically). However, a process only
needs a small amount of this memory to perform its activities.
This small amount, called the process working set, is actually
kept in memory. The operating system manages sharing of
physical memory among the various working sets.
Table 1-1 Definition of Common Terms
Term Definition
Overview of System Sizing
Terminology Used in This Guide
18 Documentum System Sizing Guide
Documentum System Sizing Guide 21
2
Deriving Workload Requirements 1
This chapter introduces workloads, discusses two concepts that are key to
determining workload requirements, and describes the workloads used in the
benchmark tests provided by Documentum. The following topics are
included:
s What Is a Workload? on page 2-1
s Determining the Workload for a Site on page 2-2
s User Connection States and Resource Consumption on page 2-2
s The Busy Hour on page 2-6
s Response Time Expectations on page 2-9
s Using the Derived Workload on page 2-9
s The Documentum Workloads on page 2-9
s Comparing and Contrasting the Workloads on page 2-23
s Operations Not Included in Workloads on page 2-26
Note: Benchmark results are described in Chapter 4, Server Configuration and
Sizing. Detailed benchmark reports are available in Kpool.
What Is a Workload?
A workload is a usage pattern for a group of users. For example, checking
documents out of the Docbase and checking them in at a later time represents
a simple usage pattern for the users called contributors (because they
contribute content to the Docbase). However, typical workloads involve more
than simple check ins and check outs. Typically, a workload also includes
activities such as navigating through folders, viewing documents,
participating in workflows, constructing or publishing virtual documents,
and so forth.
Deriving Workload Requirements
Determining the Workload for a Site
22 Documentum System Sizing Guide
Determining the Workload for a Site
Before you attempt to define the workload for a site, you should understand
two principles: the relationship between user connection states and resources
consumed by a workload, and the concept of the busy hour. User Connection
States and Resource Consumption on page 2-2 defines the user connection
states and describes how they differ in resource consumption. The Busy
Hour on page 2-6 defines the concept of the busy hour.
To estimate a workload for a site, you must obtain the following information:
s What Documentum products are in use
s The estimated number of users who are connected during the busy hour
s The estimated number of active users during the busy hour
s What Docbase operations are performed and how often each is performed
s The number, size, and content profile of documents in the Docbase
Because it can be especially difficult to identify the Docbase operations and
how often each will be performed, it is strongly recommended that you use
the Documentum Sizing Spreadsheet. The spreadsheet makes some standard
assumptions about Docbase operations based on the user category
(contributor or consumer) and products in use (the workload column in
which you enter the information).
User Connection States and Resource
Consumption
User connection states affect the amounts of resources consumed by a
workload. There are four user connection states, and the resources consumed
by a user in each state differ. The four states are:
s Connecting
s Active - in transaction
s Active - out of transaction
s Inactive
Deriving Workload Requirements
User Connection States and Resource Consumption
Documentum System Sizing Guide 23
A connecting user is a user who is requesting a connection with the Docbase. A
connecting user consumes CPU, network, memory, and disk and swap space.
An active user - in transaction is a user who is connected to the Docbase and is
currently waiting for a response to a request from eContent Server. An active -
in transaction user consumes CPU, network, memory, and disk and swap
space.
An active user - out of transaction is a user who is connected to the Docbase but
is not currently waiting for a response from eContent Server. An active - out of
transaction user consumes server memory and swap space.
An inactive user is an active user - out of transaction whose Docbase session
has timed out. An inactive user consumes only memory on the client machine.
Figure 2-1 illustrates the user connection states and their relationship with
resource consumption.
Deriving Workload Requirements
User Connection States and Resource Consumption
24 Documentum System Sizing Guide
Figure 2-1 User Connection States and Resource Consumption
Users consume server CPU mainly during session establishment and when
they initiate a request to eContent Server. When an active user is not initiating
a request, only server memory and some operating system networking
resources are consumed. When a user is inactive, the server resources are
reclaimed for other purposes. Only the client machine will consume some
memory resources in this state (essentially remembering where the session
should resume when the session returns to the Active state).
Connecting User
Active User
In Transaction
Active User
Out of Transaction
Consumes Server memory
and swap space
Inactive User
Consumes memory
only on client machine
Consumes CPU, network,
memory, disk, and swap
space
Session Up
Inactivity
Time-out
Reached
User Initiates an
Action
Consumes CPU,
network, memory,
disk, and swap space
Deriving Workload Requirements
User Connection States and Resource Consumption
Documentum System Sizing Guide 25
Inactive Connections and Resource Consumption
Both eContent Server and the RightSite Server free inactive connections.
Freeing inactive connections reduces the memory demands on the system and
minimizes the number of concurrent DBMS sessions. By default, eContent
Server frees inactive connections after 5 minutes of inactivity and the
RightSite Server does so after 30 minutes of inactivity.
When eContent Server frees an inactive connection, the server disconnects
from the DBMS and kills the process (Unix) or thread (Windows NT) that
corresponds to the inactive connection. When the RightSite Server frees an
inactive connection, the server kills the process or thread associated with the
connection.
The freed sessions can be re-established. With eContent Server, the session is
re-established transparently when the user initiates another command. With
RightSite, the user must login again (for named sessions). However, when a
session is restarted, there is a startup cost that includes operations such as
reconnecting to the DBMS, resetting caches, and so forth.
Inactive time-out trades off CPU time for reduced memory and concurrent
session requirements. That is, stopping and restarting a session repeatedly
uses more CPU than leaving the session connected continuously. However,
disconnecting the session frees memory for other uses and reduces the
maximum number of active database sessions needed.
RightSite Server Connection States
From eContent Servers viewpoint, the RightSite Server is a user. RightSite
Server connections to eContent Server go through state transitions with
associated resource use similar to any other user connecting or connected to
eContent Server.
An active RightSite Server that is not processing a request consumes only
memory and swap space. When the RightSite Server requests a Docbase
connection or processes requests, it consumes all of the major types of
resources (CPU, network, memory, disk, and swap space).
Deriving Workload Requirements
The Busy Hour
26 Documentum System Sizing Guide
The Busy Hour
The busy hour is the hour during the day in which the largest number of
operations and active sessions occur. Even in the busy hour, however, the total
amount of activity is only a percentage of the total possible activity.
To illustrate this, consider the telephone world. Suppose that ABC Telephone
Company has 1 million telephones installed in a given calling area, and in that
area the busy hour is from 11:00 a.m. to 12:00 noon. Assuming that the average
phone call lasts 2 minutes, the busy hour could theoretically involve 30
million calls1 million phones used to make 30 calls each within that hour. In
reality, only a percentage of the phones are used during the busy hour and the
calls vary in duration and occurrence. Users do not typically make repeated
2-minute phone calls. They make a call of some duration, hang up, and
engage in some other activity, and they may or may not make another call.
Figure 2-2 illustrates the busy hour and the assumption that activity during
the busy hour is only a percentage of total possible activity.
Deriving Workload Requirements
The Busy Hour
Documentum System Sizing Guide 27
Figure 2-2 Telephone Busy Hour
The ABC Communications Company sizes the back end of its phone system to
accommodate the real-world busy hour use of the phones, not the upper limit
of theoretical use.
Applying the analogy to Documentum systems, the number of phones is the
number of users (or seats) in a Documentum installation. The number of
phone owners actually making phone calls is the number of users during the
busy hour. The phone conversations are the active sessions established
between a user and the Docbase.
Just as only a percentage of installed telephones are used during the busy
hour and used only intermittently, only a percentage of Docbase users are
logged-in during the busy hour and only a percentage are making requests.
Because eContent Server frees inactive sessions to save on resource
consumption, only a percentage of the logged-in users served in the busy hour
have active sessions at any one time. Figure 2-3 shows the relative proportion
of busy hour users and active users to licensed users. Proportions will vary
from site to site.
0
10
20
30
40
50
60
8:00 9:00 10:00 11:0012:00 1:00 2:00 3:00 4:00 5:00 6:00 7:00
1
10
100
1000
10000
100000
1000000
10000000
100000000
The Busy-Hour calls
Lots of Telephones
Lots & Lots of potential
2-min calls that could be
made by those phones.
K
c
a
l
l
s
p
e
r
h
o
u
r
Deriving Workload Requirements
The Busy Hour
28 Documentum System Sizing Guide
Figure 2-3 Licensed Users Versus Busy-Hour Users Versus Currently Active Users
In general, it is best to try to size the server systems for the busy hour. But
when is it? And how can one estimate it for an application that has not been
deployed yet? A bit of guess-timating is typically required. Typical sites
estimate that 20 to 30 percent of the licensed users (in a full deployment)
request service during the busy hour. Testing has shown about 20 percent of
the users served in one hour are active at any point in the hour (assuming
users make random requests to the Docbase throughout the hour). In the
absence of any data, these are reliable proportions to use on the sizing
spreadsheet.
Note: Because RightSite waits longer to time out its session, there are typically
more RightSite active sessions than eContent Server active sessions at any one
time.
Licensed users Users per busy hour Avg Active users
Deriving Workload Requirements
Response Time Expectations
Documentum System Sizing Guide 29
Response Time Expectations
Response times are an important criterion for judging the effectiveness of the
system or service that is deployed. Response times should match or better
user expectations.
When users are asked about desired response times, they typically respond
two to three seconds if they have no information about the document size or
format. However, users also expect that it will take longer to check out a 10-
megabyte document than one that is 10K bytes. They also expect that it will
take longer to publish a virtual document with 1000 parts than it will to
publish one with 10 parts.
We recommend determining the major components of the workload and
corresponding expectations for response time.
Using the Derived Workload
After you have determined the workload at your site, compare it to the
workloads described in the following section, The Documentum
Workloads. After you determine which workload most closely matches your
sites workload, you can fill in the appropriate columns in the Documentum
Sizing Spreadsheet with your information.
You can also examine the benchmark test results reported in Chapter 4, Server
Configuration and Sizing, for those tests conducted using the workload that
matches your workload. Using these may also help you determine your
configuration requirements.
The Documentum Workloads
This section describes the workloads used in the benchmark tests conducted
by Documentum. The following three workloads are included in this section:
s The iTeam 2.2 Workload on page 2-10
s The WebPublisher 4.1 Workload on page 2-16
Deriving Workload Requirements
The Documentum Workloads
210 Documentum System Sizing Guide
s The Load and Delete Workload on page 2-22
Appendix A describes four additional workloads:
s The EDMI Workload on page A-1
s The Web Site Workload on page A-6
s The Document Find and View Workload on page A-9
s The Online Customer Care Workload on page A-9
The iTeam 2.2 Workload
iTeam is a Documentum application that provides a collaborative framework
for developing projects. It groups together the documents, resources, news,
and discussions for a project and allows users to easily reuse these for future
projects. The iTeam workload simulates the activities of iTeam users.
iTeam is Web-based and uses the following Documentum products:
s Documentum eContent Server
s Documentum RightSite Server
s Documentum Web Development Kit (WDK)
s Docbasic
Pentium III
Xeon
TM
processors at 550 MHz
s 6 PCI slots: 2 64-Bit/66 MHz; 3 64-Bit/33MHz; 1 32-bit/33 MHz
The Web-tier machines were Proliant DL360s. Each machine has, at most, two
800MHz CPUs, 4GB of memory, and two internal drives set up in a mirrored
pair. The DL360s are 1u machines that fit 42 to a standard rack.
The Proliant 8500 was the RDBMS server and file server for this test. Its
features include:
s Up to eight 700MHz Pentium III Xeon processors (1M or 2M L2 Caches)
s Up to 16GB of 100MHz SDRAM memory (only 4 GB addressable by NT)
s Multi-peer 64-bit PCI buses including 66MHz PCI slots
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
410 Documentum System Sizing Guide
The tests employed 550Mhz processors and the machine had two Compaq
disk storage arrays attached to it. Table 4-3 shows the results of the tests.
Sun/Solaris Sizing Information
This section describes the following Sun machines and the benchmarks run on
them:
s Sun Enterprise 450 on page 4-11
s Sun Enterprise 6500 and 4500 on page 4-12
Table 4-3 Compaq Sizing Data for iTeam Workload
Configuration Users/busy hour Notes
s 8500-8 (RDBMS)
s 2 x 6400R-4 (eContent)
s 4 x DL360-2 (Web)
1600 The database and file server could
have easily been run on a
4-processor 8500
The database was the bottleneck in
this test (single process address
space limitation).
s 8500-8 (RDBMS)
s 1 x 6400R-2 (eContent)
s 2 x DL360-2 (Web)
400 This is a derived configuration
based on the test results with a
4-processor 6400R.
The eContent Server host was the
bottleneck in this test (Figure 4-6).
s 8500-8 (RDBMS)
s 1 x 6400R-4 (eContent)
s 2 x DL360-2 (Web)
800 The eContent Server host was the
bottleneck in this test (Figure 4-6).
s 8500-8 (RDBMS)
s 1 x 6400R-4 (eContent)
s 1 x DL360-2 (Web)
500 The Web server host was the
bottleneck in this test (Figure 4-5).
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
Documentum System Sizing Guide 411
Sun Enterprise 450
The Sun Enterprise 450 is the top-of-the-line workgroup server from Sun. It
can have up to four 400MHz Ultra Sparc II processors (each with 4MB E-cache
memory). The E450 can have up to 182 GB of internal storage capacity and up
to 6 TB of external storage. It has 6 PCI buses providing up to 1GB/sec I/O
throughput.
Table 4-4 shows the results when the EDMI workload is run on configurations
with the Sun Enterprise 450.
Table 4-5 shows the results of the Anonymous RightSite Website Workload
run on configurations with the Sun Enterprise 450.
Notes:
s The Sun E450 can have up to four 400MHz Ultra Sparc II processors. The
system Sun-E450-8 is two E450s with 4 processors each. Sun-E450-12
represents three E450s with 4 processors each.
s Values marked with an asterisk (*) are from actual runs. All other values
are estimated based on the actual runs.
Table 4-4 EDMI Workload on Sun Enterprise 450 with 400 MHz Ultra Sparc II Processor
System and Number of
CPUs
Users/Hour on
Host-based System
(Figure 4-2)
Users/Hour
on an N-tier
System
(Figure 4-3)
Users/Hour on
N-tier with
Focus on
eContent Server
(Figure 4-7)
Users/Hour on N-
tier with Focus on
Web-Server
(Figure 4-5)
Sun-E450-2 200 - 700 500
Sun-E450-4 450* - 1500 900
Sun-E450-8 (2 servers) - 1000 - 1500*
Sun-E450-12 (3 servers) - 1500* - -
Table 4-5 Anonymous RightSite Website Workload on Sun Enterprise 450 with 400MHz Ultra
Sparc II Processors
System and Number of CPUs Users/Hour on Host-Based System (Figure 4-2)
Sun-E450-2 1000
Sun-E450-4 2000*
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
412 Documentum System Sizing Guide
s The benchmark tests were run with Solaris 2.6 and EDMS 98 (v 3.1.6)
s In the 1500 EDMI users/hour test, each of the two E450s was about 60
percent busy. It is likely that the RightSite Application server results can be
increased by 30 percent.
s The metric for the Website benchmark can also be stated in terms of HTTP
Gets serviced per hour. Each Website user performs five STATIC_HTML
operations, and each STATIC_HTML operation, in turn, performs five
HTTP gets. Consequently, a 2000 Website users/hour run generates 50,000
HTTP gets per hour (each with an average HTTP get response time of 550
msecs).
Sun Enterprise 6500 and 4500
The Sun Enterprise E6500 is the top-of-the-line mid-range server from Sun. It
can have up to 30 x 336 MHz Ultra Sparc II processors (each with 4MB E-cache
memory). The E6500 can have up to 375 GB of storage capacity in the internal
cabinet and up to 10 TB of external storage. It can have 16 system boards,
which are either I/O boards or CPU/memory boards (2 CPUs per board).
Each I/O board has four PCI channels. The system on which the tests were
conducted had a total of 26 physical slots (the slots used to house the CPU/
memory system boards are also used to house the I/O boards).
The Sun Enterprise 4500 is the mid-range server from Sun. The Enterprise
4500 can have up to 14 CPUs at 336 MHz for each processor (or 8 system
boards).
See http://www.sun.com/servers/ for more information.
It is assumed that, with respect to the Documentum application, given equal
numbers of CPUs, memory, and disk, these systems will perform in an
identical fashion. That is, a 14-CPU E4500 will achieve the same performance
as a 14-CPU E6500. Therefore, in the sizing tables (Table 4-6 and Table 4-7)
they are treated as the same.
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
Documentum System Sizing Guide 413
Figure 4-6 shows the sizing figures when the EDMI workload was run on
configurations with the Sun Enterprise 4500 and 6500
Table 4-6 EDMI Workload on Sun Enterprise 4500 and 6500 with 336MHz Ultra Sparc II
Processors
System and Number of
CPUs
Users/Hour on
Host-based
System
(Figure 4-2)
Users/Hour on an
N-tier System
(Figure 4-3)
(refer to the Notes,
the fourth bullet)
Users/Hour on N-
tier with Focus on
eContent Server
and RDBMS
(Figure 4-7)
Users/Hour on N-
tier with Focus on
Web Server
(Figure 4-5)
Sun-E6500/E4500-2 225 475 475
Sun-E6500/E4500-4 425 950 950
Sun-E6500/E4500-6 650 1425 1425
Sun-E6500/E4500-8 850 950 1900 1900
Sun-E6500/E4500-10 1075 1200 2350 2350
Sun-E6500/E4500-12 1300 1425 2825 2825
Sun-E6500/E4500-14 1500* 1650 3300 3300
Sun-E6500/E4500-16 1650 1900 3775 3775
Sun-E6500/E4500-18 1850 2125 4250* 4250*
Sun-E6500/E4500-20 2050 2350
Sun-E6500/E4500-22 2250 2600
Sun-E6500/E4500-24 2825
Sun-E6500/E4500-26 3050
Sun-E6500/E4500-28 3300
Sun-E6500/E4500-30 3550
Sun-E6500/E4500-32 3775
Sun-E6500/E4500-34 4025
Sun-E6500/E4500-36 4250
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
414 Documentum System Sizing Guide
Table 4-7 shows the sizing figures when a host-based Web site workload was
run on configurations with the Sun Enterprise 4500 and 6500.
Notes:
s Values marked with an asterisk (*) are from actual runs. All other values
are estimated based on the actual runs.
s The host-based EDMI test that achieved 2,250 EDMI users/hour was
actually run on a 26-processor machine. However, even at the peak, the
average CPU utilization was around 65 percent. We believe that if the
system is reduced to 22 CPUs, response times will be maintained and the
utilization will be 80 percent.
s The 3-tier 4,250 result was actually run on a 3-tier configuration of two
E4500s and a single E6500 with 54 total CPUs. However, from an analysis
of the CPU utilization we feel that the response times could have been
maintained on a system with 18 CPUs on both the eContent Server/DBMS
machine and the RightSite application server machine (for a total of 36
CPUs).
s The scores on the Website 4 processor test are 40 percent higher than those
shown in Table 4-6. This difference is due to the fact that the DBMS server
were from different vendors.
s In all cases, we take the largest number of users tested with and
extrapolate downward for lower CPUs. The largest number is not
necessarily the limit of that server technology; it is the largest number
tested in a lab situation.
Table 4-7 Web Site Workload CPU Sizing for Sun Enterprise 4500 & 6500 with 336MHz Ultra
Sparc II Processors
System and Number of CPUs Users/Busy Hour
Sun-E6500/E4500-2 1300
Sun-E6500/E4500-4 2800
Sun-E6500/E4500-6 4300
Sun-E6500/E4500-8 5800
Sun-E6500/E4500-10 7300
Sun-E6500/E4500-12 8800*
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
Documentum System Sizing Guide 415
s The metric for the Web site benchmark can also be stated in terms of HTTP
Gets serviced per hour. Each Website user performs five STATIC_HTML
operations. Each STATIC_HTML operation, in turn, performs five HTTP
gets. Consequently, an 8,800 Web site users/hour run generates 220,000
HTTP gets per hour (each with an average HTTP get response time of 300
msecs).
IBM, Windows NT, and AIX Sizing Information
This section contains sizing information for the following machines:
s IBM Netfinity 7000 M10 on page 4-15
s IBM AIX Systems: S7A and F50 on page 4-16
IBM Netfinity 7000 M10
The IBM NF7000M10 is the top-of-the-line Intel server from IBM. It can have
up to 4 x 400 MHz Pentium II Xeon processors. The NF7000 M10 has 6 PCI
slots and can support up to 54 GB of internal storage and 5 TB of external
storage. The system on which the tests were conducted had a total of 4
physical CPUs, 4GB of memory, and 2 EXP10 disk arrays.
The EXP10 array can hold up to ten 9GB drives. Two I/O channels are used to
interface with the two controllers on the array. The array supports various
levels of RAID.
Table 4-8 lists the sizing figures for the IBM Netfinity 7000 M10.
Table 4-8 EDMI Workload on IBM Netfinity 7000 M10
System and Number of
CPUs
Users/Hour
on Host-
based System
(Figure 4-2)
Users/Hour on
an N-tier
System
(Figure 4-3)
Users/Hour on N-
tier with Focus on
eContent Server
and RDBMS
(Figure 4-7)
Users/Hour on N-
tier with Focus on
Web Server
(Figure 4-5)
IBM-NF7000M10-2 250 500 500
IBM-NF7000M10-4 500* 500 900* 1000*
IBM-NF7000M10-8 900*
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
416 Documentum System Sizing Guide
Notes:
s Values marked with an asterisk (*) are from actual runs. All other values
are estimated based on the actual runs.
s There are some performance differences between certain DBMS vendors
on Windows NT.
s In all cases, we take the largest number of users tested with and
extrapolate downward for lower CPUs. The largest number is not
necessarily the limit of that server technology; it is the largest number
tested in a lab situation.
s These tests were conducted with Windows NT 4.0 SP4.
IBM AIX Systems: S7A and F50
The IBM S7A is a high-end RS6000 AIX server from IBM. Its highlights are:
s Standard configuration
Microprocessor: 4-way 125 MHz RS64-I or 262 MHz RS64II (upgrade
only)
Level 2 (L2) cache: 4MB for 125 MHz processors; 8MB for 262 MHz
processors
RAM (memory): 512M
Media bays: 3 (2 available)
Expansion slots: 14 PCI (11 available)
PCI bus width: 32- and 64-bit
Memory slots: 20
s AIX operating system
Version 4.3 for 125 MHz processors and Version 4.3.1 for 262 MHz
processors
s System expansion
SMP configurations: Up to 2 additional 4-way processors
RAM: Up to 32GB
Internal PCI slots: Up to 56 per system
Internal media bays: Up to 12 per system
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
Documentum System Sizing Guide 417
Internal disk bays: Up to 48 (hot-swappable)
Internal disk storage: Up to 436.8GB
External disk storage: Up to 1.3TB SCSI; up to 14.0TB SSA
The IBM F50 is a lower-end Enterprise server. Its highlights are:
s Standard configuration
Microprocessors: 166 MHz or 332 MHz PowerPC 604e with X5 cache
Level 2 (L2) cache: 256KB ECC
RAM (memory): 128MB ECC Synchronous DRAM
Disk/media bays: 18 (1 used)/4 (2 used)
I/O expansion slots: 9 (7 PCI, 2 PCI/ISA)
PCI bus widths: 2 32-bit and 1 64-bit
Memory slots: 2
s AIX operating system
Version 4.2.1 or Version 4.3
s System expansion
SMP configurations: To 2, 3 or 4 166 MHz or 332 MHz processors
(cannot be mixed)
RAM: Up to 3GB
Internal disk storage: Up to 172.8GB (163.8GB hot-swappable)
External disk storage: Up to 4.8TB SCSI-2; up to 3.5TB SSA
Table 4-9 shows the CPU sizing for the Tested IBM/AIX Servers.
Table 4-9 EDMI Workload on IBM/AIX Servers
Configuration
in the format:
No. of Servers x System-No. of Processors
Users/Busy Hour
s 1 x S7A-4 (RDBMS)
s 1 x F50-4 (eContent)
s 3 x Netserver-2 (Web)
3000
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
418 Documentum System Sizing Guide
Notes:
s The 4-tier tests used three IBM NF7000M10s as RightSite application
servers. Each 7000M10 had four 400MHz Pentium II Xeon processors (for a
total of 20 CPUs used between Unix and NT).
s The tested F50/S7A combination equipped both systems with four
processors. The DBMS server ran on the S7A (AIX 4.3 operating system)
and the eContent Server ran on the F50 (AIX 4.2 operating system).
s The notation F50/S7A-4 means that both systems have two processors.
s The S7A was set up to have only 4 processors, to match the processing
power of the new, less expensive IBM H70 (4-processor machine).
HP Windows NT and HP-UX Servers
This section discusses the following machines:
s HP NT/Intel Servers on page 4-18
s HP-UX Servers on page 4-20
HP NT/Intel Servers
The HP NETSERVER LXR 8000 is the top-of-the-line Intel server from HP. It
can have up to 4 x 400 MHz Pentium II Xeon processors (with up to 1M cache
per processor) and upgrades to 8-way multiprocessing and future processors.
The NETSERVER LXR 8000 has 10 full length PCI slots: four 64-bit hot-swap
slots, five 32-bit slots, and one shared 32-bit PCI/ISA slot. It can handle up to 8
GB of physical memory. The system on which tests were conducted had a total
of 4 physical CPUs and 4GB of memory and connected to a disk array model
30/FC through two fiber-optic interfaces.
s 1 x S7A-2(RDBMS),
s 1 x F50-2 (eContent)
s 3 x Netserver-2 (Web)
1500
Table 4-9 EDMI Workload on IBM/AIX Servers
Configuration
in the format:
No. of Servers x System-No. of Processors
Users/Busy Hour
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
Documentum System Sizing Guide 419
HP has another comparable server called the LH4. This server is also a
4-processor Intel-based system. It only lacks some of the expandability of the
LXR 8000. The LH4 can only go up to 4 GB of memory.
The HP Lpr is a two-processor rack-mounted server that can have up to 1 GB
of memory. The processors used in the tests were 600Mhz. Each server is 2U in
size, and a standard rack can hold 20 servers.
Table 4-10 lists the sizing results for the HP LXR8000 and LH4 when the iTeam
workload is run.
Table 4-11 lists the sizing results for the Lpr/LH4 N-tier test.
Table 4-10 iTeam Workload on HP LXR8000 & LH4
System and Number of CPUs Users per Busy Hour on Host-Based System
(Figure 4-2)
HP-LH4-2 100
HP-LH4-4 200*
Table 4-11 iTeam Workload on Lpr/LH4, N-tier Test
Configuration
in the format:
No. of Servers x System-No. of
Processors
Users per Busy
Hour
Notes
s 1 x LH4-2 (RDBMS),
s 1 x Lpr-2 (eContent)
s 2 x Lpr-2 (Web)
400 -
s 1 x LH4-2 (RDBMS)
s 1 x Lpr-2 (eContent)
s 1 x Lpr-2 (Web)
200 The Lpr can only have 1GB
of memory and this limits
the number of RightSite
server connections. The
2-processor Lpr was
memory bound, not CPU
bound.
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
420 Documentum System Sizing Guide
Table 4-12 lists the sizing results for the EDMI workload on HP LXR8000 and
LH4 machines.
Note:
s Values marked with an asterisk (*) are from actual runs. All others are
estimates based on the actual runs.
HP-UX Servers
Two types of HP-UX servers have been tested with Documentum: the V2600
and the K580.
The V2600 machine is a high-end HP-UX server with the following features:
s Up to 32 CPUs (a maximum of 16 were used in the tests)
s Up to 32 GB of memory
s Up to 28 2X PCI slots
s System-wide throughput of up to 15.36 GB with HPs HyperPlane crossbar
technology
s Up to 19-GB I/O throughput
The K580 machine has the following features:
s Up to six-way symmetric multiprocessing
s Single-level, large, full-processor-speed 2-MB/2-MB and 1-MB/1-MB
instruction/data caches
s Up to 37 I/O slots with optional I/O expansion cabinets
s Four internal Fast/Wide Differential SCSI-2 disk storage bays
Table 4-12 EDMI Workload on HP LXR8000 and LH4
System and
Number of CPUs
Users per Hour
On Host-based
System
(Figure 4-2)
Users per Hour
on an
N-tier System
(Figure 4-3)
Users per Hour on
N-tier with Focus
on eContent and
RDBMS
(Figure 4-7)
Users per Hour on N-
tier with Focus on
Web-Server
(Figure 4-5)
HP-LH4-2 250 - 500 500
HP-LH4-4 500* 500 900* 1000*
HP-LH4-8 (2 servers) - 900* - -
Server Configuration and Sizing
Server Sizing Results from Benchmark Tests
Documentum System Sizing Guide 421
s Up to 30 TB of total disk capacity using optional expansion cabinets
Table 4-13 shows the sizing results for the HP-V2600 using the online
customer care workload. (The online customer care workload is described in
The Online Customer Care Workload on page A-9.)
Table 4-14 shows the sizing results for the HP-K580 running the Document-
Find-and-View workload. (The Document-Find-and-View workload is
described in The Document Find and View Workload on page A-9.)
Notes:
s The Document-Find-and-View workload is different from the EDMI
workload in that it is client-server based (not Web-based). It also has a
subset of the EDMI operations (folder searching and attribute searching). It
is read-only.
s Values marked with an asterisk (*) are from actual runs. All others are
estimates based on the actual runs.
Table 4-13 Online Customer Care Workload on V2600 Machines
System and Number of CPUs Users/Hour on Host-based System (Figure 4-2)
HP-V2600-4 500
HP-V2600-8 1000
HP-V2600-16 2000
Table 4-14 Document Find and View Workload on K580 Machines
System and Number of CPUs Users/Hour on Host-based System (Figure 4-2)
HP-K580-2 1300
HP-K580-4 2600
HP-K580-6 4025*
Server Configuration and Sizing
Other CPU-Related Notes
422 Documentum System Sizing Guide
Other CPU-Related Notes
Here are some other guidelines and recommendations:
s When feasible, dedicate a separate server to the Documentum installation.
s Do not run Documentum on the same physical server as an ERP
(Enterprise Resource Planning) system. Do not install Documentum on the
PDC (Primary Domain Controller), the NIS (Network Information Services
Master), a file or print server, or another application server.
s Documentum supports eContent Server and RightSite on all Microsoft-
certified Windows NT Server hardware vendors. Omission of any
Windows NT Server hardware vendor from this document is due to lack of
space and is not intended to imply any lack of support for that vendors
Windows NT Server hardware.
s Contact your chosen hardware vendor to select the best server for your
immediate and future needs. Note that the term Enterprise Servers refer to
business, not scientific, application servers. Workgroup Servers may also
be acceptable for development, testing, workgroup, and other less
demanding configurations. Consult the release notes for the specific
product for the exact hardware on which a product is certified.
The following web sites may be useful references for researching hardware
vendor and associated product information:
HP UNIX Enterprise Servers http://www.enterprisecomputing.hp.com/
Sun Enterprise Servers http://www.sun.com/servers
IBM RS/6000 Servers http://www.rs6000.ibm.com/hardware
IBM NT servers http://www.pc.ibm.com/us/netfinity/
NT Server OS http://www.microsoft.com/ntserver
HP NT Server http://www.hp.com/netserver/
COMPAQ NT Server http://www.compaq.com/products/servers
Server Configuration and Sizing
Sizing Server Memory
Documentum System Sizing Guide 423
Sizing Server Memory
This section discusses how to size server memory. It includes the following
topics:
s Overview of the Sizing Process on page 4-23
s Key Concepts Relating to Memory Use on page 4-25
s Estimating Physical Memory Usage on page 4-28
s Estimating Paging File Space on page 4-30
s Additional Considerations on page 4-31
Overview of the Sizing Process
Memory sizing considers two system components: physical memory and
paging (or swap) file space.
To size physical memory, you must determine how much memory process
working sets are going to consume. Aside from the DBMS, you should be
primarily concerned with the non-shared pages of process working sets.
(Key Concepts Relating to Memory Use on page 4-25 defines process
working sets and how they are managed within physical and virtual
memory.)
To size the paging file, you consider the virtual memory allocated.
If you run out of either, problems can arise. If you run out of physical memory,
then lots of I/O could occur to the paging (swap) file, which leads to poor
performance. If you run out of space in the paging file, commands may fail.
Although some operating systems clearly distinguish between the two
problems, most typically give out messages that are confusing and vague.
Figure 4-8 illustrates the steps to take to determine memory and swap space
needs.
Server Configuration and Sizing
Sizing Server Memory
424 Documentum System Sizing Guide
Figure 4-8 High-Level Steps to Determining Memory and Swap Space Needs
The information gathered in step one is used to obtain the estimates for
physical memory and paging file space.
Oversizing memory is strongly recommended because most memory use in a
Documentum deployment is attributed to the server caches in the system. The
caches enhance the performance of various operations, and more efficient
operations mean better response times. Consequently, it is better to oversize
memory than risk undersizing it.
Estimating Physical Memory Usage on page 4-28 and Estimating Paging
File Space on page 4-30 contain guidelines for estimating physical and
paging or swap file memory. Some general guidelines are listed in
Additional Considerations on page 4-31.
For some examples of memory calculations, refer to Examples of Memory
Calculation on page 4-32.
1. Determine Operating System.
2. Estimate Active users
3. Estimate Number of documents
Estimate the amount of physical memory
required
Estimate the amount of swap or paging
space required
Server Configuration and Sizing
Sizing Server Memory
Documentum System Sizing Guide 425
Key Concepts Relating to Memory Use
This section discusses two key concepts:
s Virtual and physical memory
s Cache memory use
Understanding these concepts is crucial to accurate memory estimation.
Virtual and Physical Memory
Virtual memory is a service provided by the operating system (and hardware)
that allows each process to operate as if it had exclusive access to all physical
memory. However, a process only needs a small amount of the virtual
memory to perform its activities. This small amount, called the process working
set, is the actual amount of physical memory used by the process. The
operating system manages the sharing of physical memory among the various
working sets.
Physical memory is a limited resource. When the operating system wants to
conserve physical memory or manage a situation in which all working sets
wont fit into physical memory, it moves the excess pages into a paging file.
Additionally, although a process may think it has exclusive access to all of
virtual memory, the operating system transparently shares many of the read-
only portions (such as program instructions).
Figure 4-9 illustrates the relationships between physical memory, virtual
memory, and process working sets.
Server Configuration and Sizing
Sizing Server Memory
426 Documentum System Sizing Guide
Figure 4-9 Real Memory vs. Virtual Memory
Cache Memory Usage
On a typical Documentum server software installation (RightSite, eContent
Server, and a DBMS server), memory is used most heavily by various caches.
A cache is a memory area used to hold some data or object so the software can
avoid performing an expensive operation (read data from disk, from network,
and so forth).
As a particular process grows, it typically fills its caches with information,
making its operations less expensive. (An administrator can control the
maximum size of some caches, but others are sized automatically.) This
trade-off between performance and cache size means that excessive memory
use by a cache is not always bad. The most important sizing task is to ensure
that cache needs do not outstrip the available physical memory.
Shared Pages
(EXE, RD only
or real shared
memory)
Private
Pages
Private
Pages
Process #1
Process #2
Physical
Memory
Process
Working
Virtual Memory: An
abstraction provided by OS and
hardware
The Paging File or Swap file.
May have to reserve space for all
virtual memory allocated
(depending on OS) . Holds pages
that have been pulled out of real
memory.
Server Configuration and Sizing
Sizing Server Memory
Documentum System Sizing Guide 427
DBMS Caches
The DBMS data cache is generally the most dominant cache. (Its size is under
administrative control.) The DBMS uses the data cache to minimize the
number of disk I/Os that it must perform for data and index rows. It is
significantly less expensive for a DBMS to find 100 rows in a data cache than
to find them on disk. A production server system with many documents will
likely need hundreds of Mbytes (perhaps even one or more Gbytes) of
memory for this cache to ensure acceptable performance. Sizing the DBMS
cache generously reduces disk I/Os significantly.
Several DBMS servers also have caches for SQL statements that are executed
repeatedly. These caches conserve the number of CPU cycles needed for
operations such as security validation, parsing, and execution plan
preparation. It is typically good to give these caches plenty of memory. Check
the RDBMS documentation for more details.
eContent Server Caches
eContent Server uses several caches to enhance performance for operations
such as DBMS interactions, CPU cycles, and network operations. Most of the
caches are small (less than 1M byte) and bounded by the number of objects
they can contain.
The global type cache is the most dominant of eContent Servers caches. The
global type cache holds structures that provide eContent Server with fast
access to the DBMS tables that make up a types instances. The size of this
cache is limited by the number of types in the Docbase. The amount of real
memory consumed is determined by how many instances and types are
accessed. Although this cache is called the global type cache, it primarily
supports per-session access. Each eContent Server process, or thread, has its
own copy.
If the process working set of your eContent Server is larger than the memory
estimates listed in Table 4-15, your installation is probably using more custom
types than those used in the capacity testing.
RightSite Server Caches and Work Areas
RightSites memory use is dominated by:
s DMCL object cache
Server Configuration and Sizing
Sizing Server Memory
428 Documentum System Sizing Guide
s Docbasic compiled code
s Temporary intermediate Dynamic HTML construction memory
The DMCL object cache requires memory to store recently referenced Docbase
objects. Its size is bounded by the maximum number of objects that can be
stored in it (the number is set by an environment variable).
The Docbasic compiled-code memory area contains the pre-compiled
Docbasic code for the dynamic HTML executed by RightSite. The memory
used is, at most, equal to the space used by the on-disk cache of pre-compiled
Docbasic routines.
The temporary, intermediate dynamic HTML construction memory is the
memory used by RightSite to construct a dynamic HTML page. RightSite
makes heavy use of memory when constructing dynamic pages, and the
larger the number of dynamic pages accessed, the more memory is needed.
All of these areas can grow Mbytes in size depending on the workload.
Estimating Physical Memory Usage
This section contains guidelines for estimating physical memory needs for a
server machine.
User Connection Memory Requirements
Table 4-15 lists the physical memory estimates for a single-user connection for
eContent Server and RightSite on Unix and Windows NT. These estimates are
based on observations of WebPublisher 4.2 and iTeam 4.2. Actual memory
needs will vary depending on the complexity of the workload.
Table 4-15 Estimated Physical Memory Needed per Connection
Server Solaris and AIX HP-UX Windows NT
eContent Server 10M Bytes 10M Bytes 10M Bytes
RightSite Server 20M Bytes 20M Bytes 20M Bytes
Server Configuration and Sizing
Sizing Server Memory
Documentum System Sizing Guide 429
DBMS Memory Requirements
When estimating memory usage for the DBMS, consider giving the DBMS
hundreds of Mbytes of memory (perhaps even several Gbytes if the Docbase
has several million documents). Because eContent Server supports a general
purpose query language, some queries can result in DBMS table scans. By
having more memory, you minimize the number of disk I/Os needed to
execute those queries.
Table 4-16 describes the DBMS memory requirements for Docbases of various
sizes. The values in this table are presented for planning use only. The actual
requirements may vary, depending on the object types and number of objects
in the Docbase.
Operating systems generally require special tuning to support multi-giga
bytes of physical memory for the DBMS. Table 4-17 lists some examples of the
required tuning.
Table 4-16 DBMS Memory Recommendations by Docbase Size
Number of Documents Planning Ranges of
Document Metadata Size
(DBMS disk space)
Minimum Recommended
Memory Size for DBMS
1,000,000 4G to 5G 500M to 1GB
2,000,000 8G to 10G > 1GB
3,000,000 12G to 15G > 1GB
4,000,000 16G to 20G > 2GB
5,000,000 20G to 25G > 2GB
Table 4-17 Special Considerations for Supporting Large Memory
DBMS and Platform To Achieve Required Tuning
Oracle on Unix A DBMS buffer cache >2GB Oracle executable must be
relinked/rebased. See Oracle
documentation for more
details.
Oracle on Solaris A DBMS buffer cache >1GB Shared memory parameters
must be set properly in the
system file.
Server Configuration and Sizing
Sizing Server Memory
430 Documentum System Sizing Guide
Operating System Memory Requirements
The operating system buffer cache also uses physical memory. This cache is
used to store content file blocks as they are read (or written) to disk. On some
versions of Unix, this cache can be explicitly sized. It should be small for
installations with content files that are small. For deployments with large
content files, it should be large.
On Windows NT, set up the operating system to favor process working sets
over the buffer cache. The buffer cache and the memory set aside for process
working sets are dynamically sized, and an administrator can configure how
conflict between the two areas is resolved. The buffer cache must have less
priority than the process working sets or an anomalous situation could arise
that forces the DBMS or eContent Server out of memory to make way for the
file cache. This will lead to very poor performance.
Additionally, if the RDBMS is Microsoft SQL Server, you can put the tempdb
in physical memory (called temp DB in RAM). Doing so may lead to some
performance gain and is useful especially when there is sufficient memory
available. See the SQL Server product information for more details.
Estimating Paging File Space
Determining the space required for the paging file is almost as important as
determining the required physical memory. Table 4-18 lists the paging file
space recommendations for eContent Server and the RightSite Server.
Oracle on Windows
NT
A DBMS buffer cache >2GB Windows NT Enterprise must
be booted with the /3GB flag in
the boot.ini.
Table 4-17 Special Considerations for Supporting Large Memory
DBMS and Platform To Achieve Required Tuning
Table 4-18 Recommended Paging File Space per Active Connection
Server Unix Windows NT
eContent Server 11M to 20M Bytes 4G
RightSite Server 6M to 12M Bytes 4G
Server Configuration and Sizing
Sizing Server Memory
Documentum System Sizing Guide 431
The actual amount of paging file space required for each active connection
depends on the number of Documentum object types that are created.
Implementations that create many customized types might require more
paging file space per connection than shown in Table 4-18.
The maximum page file size for each drive letter on Windows NT is 4GB.
Always size the paging file to this maximum for each drive in the server
unless it interferes with the operating systems ability to create a crash dump.
If the paging file area is improperly sized, errors can occur that appear to be
memory errors. On Unix, you can use vmstat to detect an out-of-memory or
out-of-paging file condition. On Windows NT, you can use the performance
monitor, the task manager, or post-error from an error popup to detect these
conditions.
Additional Considerations
Keep in mind these additional guidelines when you size memory:
s On Solaris 2.6, the disk space for /tmp is treated like a process and its
working set. This means that /tmp will consume some physical memory
and swap space. You may need to increase some of the swap space
estimates if there is heavy use of the /tmp file system.
s Consider the impact of the following business-related factors on memory
utilization:
End of month, quarter, year peaks
Company growth over three years
Batch processing (system administration jobs)
Use of AutoRender Pro and the Transformation Engine
Users connecting from other sites
Integrations with other systems (such as SAP R3, PeopleSoft, and so
forth.)
s Manage the risk of the unknown by adding a safety margin to your
calculation
s Research maximum RAM capacity per server. If you may need more
physical RAM in the future, do not buy a server box with maximum RAM
currently installed. If it is easy to buy physical RAM, then plan for the
Server Configuration and Sizing
Examples of Memory Calculation
432 Documentum System Sizing Guide
worst case and buy for the best. Your choice of a server model may be
governed by the memory requirements calculated for your active
Documentum users rather than by the CPU requirements.
Examples of Memory Calculation
The three examples in this section use the following equation as a basis for
memory estimates:
Memory = Base Memory + DBMS Memory +
(Per User Memory for Documentum x Number of Active Users)
Example One
This example is based on the following assumptions:
s All components (eContent Server, HTTP Server, RightSite, RDBMS) reside
on one server (host-based).
s The Docbase will have up to 500,000 objects within the next two years.
s The maximum number of active users will be 50 users (20 percent of 250
users per hour).
Table 4-19 shows the memory calculations for the first example.
Table 4-19 Memory Calculation Example 1
Software Components Minimum Memory
Capacity Required
Operating system 128 MB
RDBMS 500 MB
eContent Server 32 MB
RightSite and HTTP (required only for Web clients or
RightSite applications)
32 MB
50 Active Users x (10 MB + 20 MB) 1.5 GB
TOTAL Estimated Server Memory Requirements 2.2 GB
Server Configuration and Sizing
Examples of Memory Calculation
Documentum System Sizing Guide 433
Example Two
This example is based on the following assumptions:
s eContent Server and the RDBMS are on one server; RightSite and the
HTTP Server are on a second server.
s The Docbase will have up to 500,000 objects within the next two years.
s The maximum number of active users will be 50 users (20 percent of 250
users per hour).
Table 4-20 shows the calculations for the first server, and Table 4-21 shows the
calculations for the second server.
Table 4-20 Memory Calculation for Example 2, First Server (eContent Server and RDBMS)
Software Component Required Minimum Memory
Capacity
Operating system 128 MB
RDBMS 500 MB
eContent Server 32 MB
50 Active Users x 10 MB 500 MB
TOTAL Estimated Server 1 Memory
Requirements
1.2 GB
Table 4-21 Memory Calculation for Example 2, Second Server (RightSite and HTTP Server)
Software Component Required Minimum Memory
Capacity
Operating system 128 MB
HTTP Server 32 MB
RightSite Server 32 MB
50 Active Users x 20 MB 1 GB
TOTAL Estimated Server 2 Memory
Requirements
1.2 GB
Server Configuration and Sizing
Examples of Memory Calculation
434 Documentum System Sizing Guide
The aggregate minimum required memory for both servers is 2.4 GB.
Example Three
This example is based on the following assumptions:
s Three servers: One for the Documentum eContent Server, another for the
RDBMS, and another server for the HTTP server software and RightSite.
s The Docbase will have up to 500,000 objects within the next two years
s The maximum number of active users will be 50 users (20 percent of 250
users per hour)
Table 4-22 shows the calculations for the first server; Table 4-23 shows the
calculations for the second server; and Table 4-24 shows the calculations for
the third server.
Table 4-22 Memory Calculation for Example 3, First Server (eContent Server)
Software Component Required Minimum Memory
Capacity
Operating system 128 MB
eContent Server 32 MB
50 Active Users x 10 MB 500 MB
TOTAL Estimated Memory Requirement
for the eContent Server Machine
660 MB (round up to 700 MB)
Table 4-23 Memory Calculation for Example 3, Second Server (RDBMS)
Software Component Required Minimum Memory
Capacity
Operating system 128 MB
RDBMS 500 MB
TOTAL Estimated Memory Requirement
for the RDBMS Machine
628 MB (round up to 768 MB)
Server Configuration and Sizing
Sizing Server Disk Capacity
Documentum System Sizing Guide 435
The aggregate estimated minimum memory for all three servers is 2.3 GB.
Sizing Server Disk Capacity
This section describes how to size server disk capacity and provides
guidelines and information to help you with that task.
Figure 4-10 summarizes the disk capacity sizing process.
Table 4-24 Memory Calculation for Example 3, Third Server (HTTP and RightSite Servers)
Software Component Required Minimum Memory
Capacity
Operating system 128 MB
HTTP Server 32 MB
RightSite Server 32 MB
50 Active Users x 20 MB 1 GB
TOTAL Estimated Memory Requirement
for the HTTP and RightSite Server
Machine
1.2 GB
Server Configuration and Sizing
Sizing Server Disk Capacity
436 Documentum System Sizing Guide
Figure 4-10 Disk Storage Sizing Process
There are numerous inputs to this process, and it is important to obtain correct
estimates and information for them. If the inputs are incorrect and disk space
is sized incorrectly, the server will perform very badly. Fixing problems
related to disk space is difficult after the fact. The inputs to the process
include:
s The amount of physical space needed for all file stores
s The recovery characteristics for all file store areas
s The disk access needs for each file store area
After you obtain the needed inputs, you can determine which disk
configuration will best suit your needs.
1. Determine Operating System.
2. Estimate Active users & users per
hour
3. Estimate Number of documents
4. Est. # of renditions and versions
5. Frequency of Full text Searches,
case-insensitive searches, and
dump/load operations
Estimate the amount of physical space
required on disk
Estimate the amount of drives needed to
provide sufficient disk access capacity.
Establish recovery requirements for each
Data storage area
Server Configuration and Sizing
Sizing Server Disk Capacity
Documentum System Sizing Guide 437
The next section Key Concepts for Disk Sizing, reviews some of the factors
that affect decisions about disk sizing and configuration. Disk Striping and
RAID Configurations on page 4-40 describes a variety of common disk
configurations. Disk Storage Areas on page 4-42 describes the
characteristics of the disk storage areas used in a Documentum deployment
and makes some disk sizing and configuration recommendations for each.
Key Concepts for Disk Sizing
To correctly size disk space, it is important that you understand several key
concepts:
s Disk Space and Disk Access Capacity
s Effect of Table Scans, Indexes, and Cost-based Optimizers on I/O
s DBMS Buffer Cache Memory Effect on Disk I/Os
Disk Space and Disk Access Capacity
To size disks for a Documentum installation, you must size both the required
disk space and the disk access capacity. Sizing disk space entails ensuring that
there is sufficient room to store the permanent and temporary data. Sizing
access capacity entails ensuring that there are sufficient disk spindles (or
arms) to gather all of the data in a reasonable time frame.
To illustrate the difference, suppose there is a 4 GB DBMS table, which can fit
on a single 9 GB drive. Suppose also that the DBMS needs to scan the entire
table. If the DBMS uses 16K blocks, then 262,000 disk I/Os are needed to scan
the table. If the scan takes place over a single disk drive (at 10 msec an I/O), it
takes 44 minutes to scan the table (assuming that no pre-fetching occurs).
However, if the table is striped over 10 drives in a RAID 0 stripe, the scan
might actually only take 4 minutes if one read could hit all 10 drives (by small
stripe units). Notice that ten 9 GB drives offer 10 times the space needed for
the table; however, meeting the space requirements did not necessarily meet
the response time requirements.
Server Configuration and Sizing
Sizing Server Disk Capacity
438 Documentum System Sizing Guide
Effect of Table Scans, Indexes, and Cost-based Optimizers on I/O
The biggest demands on disks typically come from large table scans. A table
scan occurs when a DBMS reads all of the data in a table looking for a few
rows. In a table scan, the DBMS does not use an index to shorten the lookup
effort. Table scans are not always bad (in some cases using an index can
actually hurt performance). However, if the table is large enough or the
amount of physical RAM is small enough, table scans generate enormous
amounts of disk I/O.
Note: A database index is a data structure that allows the DBMS to locate
some records efficiently without having to read every row in the table.
Documentum maintains many indexes on tables and even allows the
Administrator to define additional indexes on their underlying tables.
In general, large table scans should not appear in your workload, but there are
some operations that will result in a table scan. For example, the DBMS cannot
use an index for a case-insensitive attribute search, which leads to a table scan.
If your applications contain operations that result in large table scans which
cant be tuned away, it is important to size the servers disks properly to get
best performance.
Tuning with the Optimizer
Some table scans are unavoidable (case-insensitive search); others are the
result of query optimization problems by the DBMS vendors. Some vendors,
such as Oracle, actually support multiple modes of query optimization. One
popular mode used frequently by the Documentum Server relies on the
Oracle rule-based optimizer. This optimizer picks a data access plan based on a
sequence of rules, not on the statistics of the tables (the number of rows). As
Docbases get larger, using a rule-based optimizer can lead to some costly
mistakes that cause table scans. On large Docbases, a cost-based optimizer
might deliver better access plans because it can determine table size and
column value distributions. However, cost-based optimizers are not
guaranteed to pick the best data access plan. Even with the table statistics,
they can mistakenly pick an access plan that leads to more disk I/O.
Its not too difficult to switch between the Oracle rule-based optimizer and a
cost-based optimizer (like ALL_ROWS). (The global parameter is set in the
init.ora file.) If you use a cost-based optimizer but the tables have no statistics,
Server Configuration and Sizing
Sizing Server Disk Capacity
Documentum System Sizing Guide 439
Oracle uses the rule-based optimizer instead. Therefore, in Oracle, tables
could be selectively moved from rule-based to cost-based optimizing if the
site doesnt generate statistics for the tables.
Databases like Sybase and the Microsoft SQL Server always use cost-based
optimization and need to have statistics for the various tables.
DBMS Buffer Cache Memory Effect on Disk I/Os
The DBMS data cache (illustrated in Figure 4-11) holds images of rows that
were read from or written to disk. If the DBMS can find the data rows in the
cache, it does not need to read from the disk. Consequently, there is an inverse
relationship between the amount of data in the cache and the demands on the
disk.
Figure 4-11 Illustration of a DBMS Cache
Theoretically, if memory is large enough and the Docbase small enough,
nearly all of the Docbase attribute information may fit in memory, almost
eliminating the requirement for disk I/O. However, in reality, all attribute
information for large Docbases cannot fit in physical memory. So, the DBMS
will keep some information and remove other information from the cache as
needed. A Least Recently Used (LRU) algorithm keeps the data most often
referenced in the cache and lets the least referenced go. This also helps
minimize disk I/O because the most referenced data stays in the cache.
Is data in
cache?
Yes! No
need for disk
I/O
No! then go
out to disk
and get it.
DBMS data cache
Server Configuration and Sizing
Sizing Server Disk Capacity
440 Documentum System Sizing Guide
Table 4-25 shows disk I/O ranges for some Docbase sizes based on the EDMI
workload. The workload does not have any table scans; however, this
represents disk I/O statistics for various database vendors under different
memory configurations.
Disk Striping and RAID Configurations
Disk striping distributes a piece of data over separate disks. When the data is
requested, all of the disks participate in returning the data.
It might seem that retrieving data from multiple drives would take longer
than retrieving data from one drive. However, given that the data is typically
large (many Kbytes), it is unlikely that retrieving the data from a single disk
could be accomplished in a single operation anyway. If you stripe the data, the
drives work in parallel, thus allowing the operation to happen much more
quickly.
The striping logic of an operating system (or of disk array firmware) makes a
group of disks appear as a single disk or volume. The operating system will
break up a disk request into multiple disk requests across multiple drives.
Figure 4-12 illustrates how striping works.
Table 4-25 Ranges of Disk I/Os vs. Memory Used vs. Docbase Sizes vs. Users/Hour
Docbase Size
(Megabytes)
Disk I/Os per second DBMS Memory
(Gigabytes)
EDMI Users/Hour
1 60 - 200 1 - 2 900 - 1500
2 - 2.5 60 - 300 1 - 2 800 - 3000
5 410 - 2000 1.5 - 3 2000+ - 4000+
Server Configuration and Sizing
Sizing Server Disk Capacity
Documentum System Sizing Guide 441
Figure 4-12 Disk Striping Concept
A key component of disk striping performance is the size of the data stripe
(data block). The smaller the stripe, the more parallel drives that might used
for a single I/O. The more parallel drives, the better the performance.
However, if the stripe is too small, the overhead of dealing with the stripe
exceeds any performance gains from the striping. If the stripe is too large, then
I/Os are likely to queue up on a single drive.
To illustrate, using an extreme example of poor stripe size from the DBMS
world, suppose an administrator has many individual disks and stripes the
data by creating multiple table space files across the independent disks. The
administrator puts a single, large, sequential portion of the table space on each
drive. If a request is made for a portion of the table, it is likely that the I/Os
are going to be concentrated on a single drive in an uneven fashion.
In general, RAID0 (striping without parity), as described above, outperforms
tablespace or DBMS device-level disk striping. However, RAID0 has some
disadvantages. If a single drive fails, the entire stripe set fails. That is, four
disks in a RAID 0 stripe set (or logical drive) have a shorter mean time before
failure (MTBF) than a single drive, because any one of the four physical drives
can bring the logical drive down.
Striping Logic
Striping Logic
Retrieve the data in 4
sequential I/Os
Retrieve the data in 4
parallel I/Os
Server Configuration and Sizing
Sizing Server Disk Capacity
442 Documentum System Sizing Guide
There are two major ways to protect performance yet maintain reliability:
mirroring and striping with parity. With mirroring (or RAID 1), the data is
written to two drives. If one drive fails, the data can be read from the other.
When data is striped over a set of mirrored drive pairs, the configuration is
called RAID1+0 or RAID10.
In striping with parity (or RAID5), parity information is written out in
addition to the data. The parity information can be used to recreate data if a
drive fails. The parity information for a write operation is always written to a
drive that does not contain the data that generated the parity code. The parity
information can be written to any drive in the configuration that doesnt
contain the data that generated the parity code. For example, suppose there
are four drives on which data is striped. One write might put data on drives 1,
2, and 3 with the parity information on drive 4. Another write might put the
data on drives 1 and 2, with the parity information on drive 3.
The disadvantage of a RAID1+0 configuration is the cost of the additional
drives. The disadvantage of a RAID5 configuration is the extra I/Os needed to
write out the parity information. In general, the access penalty for RAID5 is
fairly severe for DBMS files. It can provide decent performance for Docbase
content.
Disk Storage Areas
A Documentum installation has many different disk storage areas that need
sizing. Table 4-26 outlines these storage areas and some of their characteristics.
Table 4-26 I/O Characteristics for Documentum
Data Area Used For Size Recovery
Characteristics
I/O Activity Advice
DBMS and
Documentum
Server backup
Backup copies
of metadata and
content files
Same size as
content plus
DBMS data
typically Gbytes
Hard to recover Sequential
writes, but
infrequent
usage
RAID 5
Documentum
content and
fulltext index
Actual online
content plus full
text index
Gbytes Hard to recover Random read/
write for small
files and
sequential for
large files
RAID5 or
mirrored
pairs bound
into a RAID 0
stripe (RAID
1+0)
Server Configuration and Sizing
Sizing Server Disk Capacity
Documentum System Sizing Guide 443
Documentum
eContent Server
temp storage
area
Intermediate file
transfer area
Typically less
than 100M bytes
Easy to recover Random read/
write for small
files and
sequential for
large files
RAID5 is
acceptable
DBMS
transaction logs
Ensuring DBMS
operations
remain stable
after a failure
Mbytes to
Gbytes
Hard to recover Sequential
writes
Mirrored
pairs bound
into a RAID 0
stripe (RAID
1+0)
DBMS data &
index
Holding the
document meta-
data
Gbytes Hard to recover Random read/
write
RAID 1+0
preferred
DBMS temp &
Oracle rollback
segments
Temp is used for
DBMS sorting
and worktables;
rollback seg-
ments are used
for transaction
aborts
Mbytes to
Gbytes
Easier to recover
and rebuild than
DBMS data and
indexes
Sequential and
random writes
(some reads)
Mirrored
pairs bound
into a RAID 0
stripe (RAID
1+0)
RightSite temp
& log storage
area
Temp is used as
an intermediate
file transfer
area; log storage
area stores log
files
Hundreds of
Mbytes
Easy to recover Random read
and write for
small files and
sequential for
large files
RAID5 is
acceptable
RightSite DMCL
cache
Per session file
cache (what
would have
been on client
machines in a
client server
environment)
Hundreds of
Mbytes
Easy to recover Random read
and write for
small files and
sequential for
large files
RAID5 is
acceptable
Table 4-26 I/O Characteristics for Documentum
Data Area Used For Size Recovery
Characteristics
I/O Activity Advice
Server Configuration and Sizing
Sizing Server Disk Capacity
444 Documentum System Sizing Guide
Disk Space Sizing
This section provides a formula for calculating disk space needs.
Physical Disk Requirements of the Documentum Software
Components
Table 4-27 lists the disk space requirements for the servers in the system. The
figures in this table are based on test Docbase scenarios. The actual
requirements for a particular Docbase will vary based on such factors as
subtype depth, number of attributes, size of each attribute value, number of
repeating values, and overhead for non-document objects.
RightSite
Docbasic
compiled disk
cache
Pre-compiled
Docbasic scripts
Mbytes Easy to recover Random read
and write for
small files and
sequential for
large files
RAID5 is
acceptable
Internet Server
log files
Per-operation
logging by
Internet Server
Tens of Mbytes Easy to recover Random read
and write for
small files and
sequential for
large files
RAID5 is
acceptable
OS paging/
swap files
Holds the pages
of a process
working set no
longer needed
in memory
Gbytes Painful to
recover (OS will
have hard time
to boot)
Mostly
sequential
writes
Suggest RAID
1+0
Table 4-26 I/O Characteristics for Documentum
Data Area Used For Size Recovery
Characteristics
I/O Activity Advice
Table 4-27 Physical Disk Requirements for Server Software in Documentum System
Server Requirement
eContent Server 175 MB
RDBMS (not including Table Space) 100 MB
Server Configuration and Sizing
Sizing Server Disk Capacity
Documentum System Sizing Guide 445
Typical Disk Space Calculation Model for Content and Attribute
Data
You can use the following formula to estimate disk usage for document
content and the metadata:
5K Per Object x Number of saved versions stored in the RDBMS table
(Document Object Data) +
Document Size x Number of saved versions stored in the Docbase
(Document Content) +
Rendition Size x Number of saved versions stored in the Docbase
(Renditions) +
30% of the original document size x Number of saved versions stored in
the Docbase (Fulltext Indexes) +
2.5K x Number of saved versions stored in the Docbase (Annotations)
Expressed mathematically, the formula is:
((5k * number of versions) +
(Document Size * number of versions) +
(Rendition Size * number of versions) +
((30% * Document Size) * number of versions) +
(2.5 * number of versions)) * Total Number of Documents
= Total Disk Capacity
(Note that not all documents require versions, renditions, annotations or full-
text indexes. You can configure a system to prune version and rendition trees
and annotations).
Use the Excel Spreadsheet to automatically calculate the estimate.
Base Documentum Objects
(including RDBMS Table Space)
50 MB
RightSite 60 MB
Table 4-27 Physical Disk Requirements for Server Software in Documentum System
Server Requirement
Server Configuration and Sizing
Database License Sizing
446 Documentum System Sizing Guide
Additional Considerations
Documentum does not impose additional overhead on content storage in the
file system beyond the actual size of the content file unless the contents are
full-text indexed.
If a Docbase is participating in object replication, there must be available disk
space to execute the requirements of the replication job.
Similarly, disk space must be available when a dump and load is performed.
Additional References
System Administrators Guide, Chapter 11, Tools
Docbase Design Principles Customer Training Course.
If your organization has an existing Docbase, you can generate pertinent
information regarding disk utilization (and more) by executing eContent
Servers State of Docbase system administration tool. This tool is described in
the System Administrators Guide, Chapter 11, Tools.
Database License Sizing
RDBMS user licenses are an important component of the cost of a deployed
system. Database vendors typically have three different licensing schemes:
s Per seat licensing (fee for each possible named user of the system)
s Concurrent user licensing (fee for the maximum number of concurrent
users on a system)
s Per CPU licensing (fee per CPU used during production use)
Per CPU licensing is common when associated with Internet applications
In most cases, Documentum works well with concurrent user licensing.
Because user interaction with the Documentum environment is likely to be
erratic (get document, read document) and eContent Server closes inactive
sessions, the number of concurrent DBMS users is determined by the number
of active Documentum sessions. The number of active Documentum sessions
Server Configuration and Sizing
Certified Database and HTTP Server Versions
Documentum System Sizing Guide 447
is a percentage of the number of users supported per hour. For example, in the
EDMI workload, the number of active Documentum sessions was about 17 to
20 percent of the total EDMI users supported per hour.
Consequently, you can estimate the number of active sessions. Increase this
value to account for peak periods (for example, estimated number + 20
percent extra).
Some notes on concurrent user licensing and Documentum:
s For some implementations (for example, Oracle), multiple DBMS sessions
will be created for some user activities. Consequently, the actual number of
concurrent DBMS users will be greater than the number of active
Documentum users. Typically, the actual number is approximately one and
half times the number of active users. In these cases, however, the DBMS
vendors typically do not charge extra for multiple sessions per
Documentum user.
s Although shortening the eContent Server client session time-out reduces
the number of active sessions at any one time, it also causes more frequent
reconnections, which drives response time up. For remote users, this
penalty might be severe. You might actually need to increase the client
session timeout to keep remote users happy.
For anonymous RightSite users, per-CPU licensing might be more
appropriate. In this case, the number of active Documentum sessions created
by the pool of anonymous RightSite servers is typically fairly small. However,
each of those active servers might perform numerous operations. This fits well
with a per-CPU licensing scheme.
Certified Database and HTTP Server Versions
Please refer to the eContent Server release notes and the RightSite release
notes for information about the certified RDBMS versions and HTTP server
versions.
Server Configuration and Sizing
Certified Database and HTTP Server Versions
448 Documentum System Sizing Guide
Documentum System Sizing Guide 51
5
Server Network Configuration
Guidelines 1
This chapter contains guidelines for sizing and configuring the network
between multiple Documentum sites. The following topics are covered:
s Overview of Network Sizing on page 5-1
s Key Concepts for Network Sizing on page 5-2
s Making the Decision: Localizing Traffic or Buying More Bandwidth on
page 5-6
s Additional Specific Network Recommendations on page 5-12
Overview of Network Sizing
Sizing the network between Documentum sites is principally a matter of:
s Sizing bandwidth needs for servers and users
s Determining locations for servers
First, you must understand network bandwidth needs. Sometimes the need
for bandwidth between remote users and their server is so great that it makes
more sense to relocate the server (or just content) closer to users to improve
response times and drive down telecommunications costs.
The main issue that affects remote user response time is content size relative to
available bandwidth. The second issue is the number of operations that have
to take place between a client and the server and how much data is generated
during those operations.
Figure 5-1 illustrates the steps for configuring your network resources.
Server Network Configuration Guidelines
Key Concepts for Network Sizing
52 Documentum System Sizing Guide
Figure 5-1 Steps To Configure Network Resources
Key Concepts for Network Sizing
Before you can make decisions about sizing and configuring the network, you
must understand several key concepts about networks and how they affect
sizing. These key concepts are:
s Bandwidth and Latency on page 5-3
s Bandwidth Needs and Response Time on page 5-4
1. Estimate Number of users/hour
2. Estimate operations per user
3. Estimate the bytes per operation
4. Estimate document size
5. Determine locations of users
Determine network demand per user
community as if they had their own
Determine possible geographic
locations for servers
Rework network load based on
geographic locations. Adjust if
network demand outstrips budget.
Server Network Configuration Guidelines
Key Concepts for Network Sizing
Documentum System Sizing Guide 53
Bandwidth and Latency
Bandwidth describes the available throughput of the various components in a
network. It is usually measured in Bytes/sec or bits/sec. The factors affecting
bandwidth are:
s Transmission speed
s The amount of external traffic using the media
To illustrate how external traffic affects bandwidth, lets look at a single-lane
freeway with a speed limit of 60 miles per hour. Optimal bandwidth allows
the cars to go at the speed limit (that is, a car could cover 60 miles in an hour).
Actual bandwidth is likely to be far less due to rush hour traffic or accidents.
For example, the traffic on the freeway may have an average speed of 20 miles
per hour during rush hour. Available bandwidth is diminished by external
traffic forces, in this case, additional cars.
Data transfer latency is the time it takes to send data between two hosts. The
factors affecting latency are:
s Processing time on the respective host machines
s Propagation delay across the transmission media
s Available bandwidth for each network between the two hosts
s Processing time for all routers and gateways between the two hosts
To continue with the above example, latency is the time it takes to get from
one place to another. The distance from Pleasanton, California to the Golden
Gate bridge in San Francisco might only be about 50 miles, but the delays
caused by the toll bridge between Oakland and San Francisco and the various
traffic lights in San Francisco are likely to make that trip take longer than 1
hour (Figure 5-2).
Server Network Configuration Guidelines
Key Concepts for Network Sizing
54 Documentum System Sizing Guide
Figure 5-2 Example Of Bandwidth vs. Latency
When you apply the concepts of latency and bandwidth to sizing a real-world
network, you must be sensitive to the various components in the network and
how the components affect the response time of a Documentum
implementation.
Bandwidth Needs and Response Time
There are two ways to provision network resources based on network load.
You can:
s Optimize for average network demand
s Optimize for online response time
To optimize for average demand, you need only to determine the number of
bytes transferred in a busy hour and ensure that there is enough bandwidth to
meet that demand.
Unfortunately, optimizing for network demand often leaves online users with
poor response time. For example, suppose only 5 Mbytes of data must be
transferred between two sites in one hour. A 56K bps link should provide
sufficient bandwidth for users to get good response time (5M bytes per hour is
only 20 percent of the total amount of bytes that could be transferred on a 56K
bps link in 1 hour). However, that isnt the case. Network demands of online
users are characterized by small bursts separated by long pauses, and
response will be judged as good or bad by how long it takes to service one of
those bursts. Figure 5-3 illustrates the nature of the network demand.
50 miles distance
65 miles/hour on HW 580
(Bandwidth)
25 miles/hour on HW 101
in San Francisco
Golden Gate Bridge
In San Francisco
Pleasanton, CA
Latency for trip from Pleasanton to Golden Gate Bridge >= 1 hour
Server Network Configuration Guidelines
Key Concepts for Network Sizing
Documentum System Sizing Guide 55
Figure 5-3 Example of Bursty Network Load Caused by Online Users
As an example, suppose that a particular command (such as the display of a
dynamically rendered Web page) transfers 80,000 bytes of data. With a 56K
bps line, the command will take 11 seconds to complete at best and will be
judged by the users as a command with poor response time. At 256K bps, the
response can go down to approximately 2.5 seconds and be considered a
command with acceptable performance. Now suppose that 25 users issue the
above command 10 times in 1 hour. Their average network demand for the
hour is 4.8M bytes (80,000 bytes x 6 commands x 10 users), which is only 20
percent of the 56K bps bandwidth and 4 percent of a 256K bps bandwidth.
Although this seems to indicate that an enormous amount of bandwidth is
needed for users to achieve any level of decent performance, it turns out that
because there are large pauses in the use of the line, more online users can
share bandwidth due to the random nature of their requests. That is, the 256K
bps line could also likely serve an additional 70 users at the same number of
requests with good response time (which would drive the average usage of
the line to about 30%).
Typically, it is not cost effective to provide so much bandwidth so that all
commands can complete in 3 seconds or less. Rather, it is important to try to
ensure that the most frequently performed operations have good response
times. For example, if users log into their application only once an hour, then a
login that takes 10 seconds over some amount of bandwidth is much less
Example of bursty Network load caused by
online users
0
0.2
0.4
0.6
0.8
1
time (secs)
%
b
a
n
d
w
i
d
t
h
u
s
e
d
unused
bandwidth
longer the burst, the poorer the
response time
Server Network Configuration Guidelines
Making the Decision: Localizing Traffic or Buying More Bandwidth
56 Documentum System Sizing Guide
annoying than some other command that takes 10 seconds and is run 10 times
in the hour for each user. Determining how much bandwidth to allocate for a
particular Documentum application focuses on trying to make the most
frequent commands run the most quickly. This can be achieved by adding
more bandwidth or by choosing a Documentum distributed option to localize
network requests. Making the Decision: Localizing Traffic or Buying More
Bandwidth, discusses those decisions.
Making the Decision: Localizing Traffic or Buying
More Bandwidth
When you are sizing and configuring a network between Documentum sites,
the goal is to achieve good user response time at minimum cost. This section
describes the trade offs between bandwidth and a variety of Documentum
options, including:
s Remote Web servers
s Content servers
s Object replication
For all options, there is a trade off between purchasing more bandwidth and
localizing access by putting a server at the remote site. No single formula can
be applied to make the evaluation. Bandwidth costs differ by region and by
proximity to service and telephone facilities. There are also different choices
for the types of server machines to put at the remote site, and maintenance
and staffing costs for that software and hardware will also differ by region.
Lets illustrate the cost trade offs with a hypothetical example. Suppose a
remote office supports 25 users. Analysis determines that without any special
servers at that site, users need a bandwidth of 700K bps to achieve good
response time. However, locating a special server at the remote office would
reduce the traffic demands so that a bandwidth of 128K bps would provide
good response time.
The assumed costs associated with the remote server machine include:
s Server host (2 CPUs, 1GB memory, 2 internal disks, monitor, keyboard,
tape drive, and rack): $15,000
s Software costs (OS, etc): $1,500
Server Network Configuration Guidelines
Making the Decision: Localizing Traffic or Buying More Bandwidth
Documentum System Sizing Guide 57
s Tape backup costs: $200 per year
s Other administrative costs: $1,000 one-time fee for training remote admin
(who already supports the users at the remote site) + $2,000 one-time fee
for initial setup
In this example, the remote server machine requires fairly little day-to-day
administration, and the administration can be done remotely from the central
site. Additionally, the example assumes that the service will be in production
for three years and that power costs are negligible. These assumptions bring
the total charge for the remote machine to about $20,000.
Lets also assume that the remote office has access to a frame-relay service
provided by the local telephone company and that the frame-relay service
charges include both port prices and access facility charges. These port prices
are shown in Table 5-1.
Table 5-2 shows the access facility charges.
Table 5-1 Example Frame Relay Port Prices
Speed Installation Fee Monthly Fee
56K bps $354.67 $70.93
128K bps $354.67 $141.87
384K bps $354.67 $378.32
1.544M bps $354.67 $472.90
37M bps $1,418.69 $4,539.79
Table 5-2 Example Frame Access Facility Charges
Speed Installation Fee Monthly Fee
56K bps $597.37 $47.41
128K bps $600.69 $165.94
384K bps $600.69 $165.94
1.544M bps $600.69 $165.94
Server Network Configuration Guidelines
Making the Decision: Localizing Traffic or Buying More Bandwidth
58 Documentum System Sizing Guide
Table 5-3 shows the cost of setting up and using frame relay from the central
site to the remote office for bandwidths of 1.544M bps and 128K bps.
In this example, a remote server saves the company $3,000 over three years.
Such a savings is less than 10 percent. If the remote office is in a different
country or a cost-effective frame-relay service cant be used, then using the
remote server could add up to significantly more savings.
The following sections discuss how to determine the bandwidth needs for a
variety of deployment options. The Documentum Sizing Spreadsheet also
contains information about this topic.
More Bandwidth or Remote Web Servers
Documentum is an N-tier server architecture, and each tier employs a
different protocol with different networking characteristics. In Web-based
deployments, the browser-to-Web server protocol (HTTP/HTML) is likely to
be the most verbose. Depending on the application, the HTML could be 20
times more verbose than the corresponding DMCL. Consequently, if remote
users are centralized in a single office, it might be best to locate a remote Web
server in that office to achieve better response times. Doing so takes
advantage of the fact that HTML requires more bandwidth than the
corresponding Documentum DMCL protocol between the Web tier and
eContent Server. Because there is little state to maintain on a Web or
application server machine, the cost to use a remote Web or application server
machine is low.
Table 5-3 Total Cost for Three-Year Period for Two Line Sizes
Charges 128K bps 1.544M bps
Port installation fee $354.00 $354.00
Port charge for 3 years $5,076.00 $16,992.00
Access installation fee $600.00 $600.00
Access charge for 3 years $5,940.00 $5,940.00
Total for a single site $11,970.00 $23,886.00
Total for both sites $23,940.00 $47,772.00
Server Network Configuration Guidelines
Making the Decision: Localizing Traffic or Buying More Bandwidth
Documentum System Sizing Guide 59
Figure 5-4 Bandwidth and Remote Web Servers
Content Transfer Response Time: More Bandwidth or Content
Servers
Suppose that the Web server is located at the remote site, but performance is
still poor (or expected to be poor) given the bandwidth provided. The next
item to improve is the time required to transfer content (or files). You can add
more network bandwidth (for example, upgrade a 56K bps link to a 128K bps
link) or add a Documentum content server at the remote site to localize
content access. Deciding which strategy to employ will depend on the relative
costs of each, and the costs are likely to differ based on geography.
Verbose HTML from
Web/App server top Browser
Less verbose
DMCL between
Web/App server
and eContent
Verbose HTML
from Web/App
server top Browser
Less verbose
DMCL between
Web/App server
and eContent
Server Network Configuration Guidelines
Making the Decision: Localizing Traffic or Buying More Bandwidth
510 Documentum System Sizing Guide
Using Documentum content servers is quite attractive when bandwidth is
very expensive and users typically access large files. The administration and
hardware needs of a Documentum content server are fairly small. A content
server only requires additional disk space for content (beyond the needs of a
regular Web or application server) and it can be administered remotely.
Figure 5-5 illustrates the use of content servers.
Figure 5-5 Bandwidth and Remote Content Servers
DMCL + Content Transfer
Docbase is remote.
Response time can be
improved by
increasing bandwidth
DMCL only
Content is local and
only operation traffic
is sent to the remote
eContent server.
Content served locally by Content
Server Network Configuration Guidelines
Making the Decision: Localizing Traffic or Buying More Bandwidth
Documentum System Sizing Guide 511
Operation Response Time: More Bandwidth or Replication
The trade off between bandwidth and local server machines also applies to
DMCL operations issued from remote application servers or from remote
Desktop Client users. That is, although content and Web access are moved to
the remote site, users will still get poor response time. In such cases,
Documentum object replication might be the best solution for the enterprise.
With object replication, all or part of a Docbase is replicated to the remote site.
The replication process happens during off-peak hours, and consequently,
remote users almost always interact with their local servers and get great
response time. The costs of object replication include setting up a remote
RDBMS, the overhead of replication administration, and the overhead of
replica update latencies (nightly updates). Figure 5-6 illustrates how
bandwidth and object replication interact.
Figure 5-6 Bandwidth and Object Replication
Replication update at night
Fast online access of replicated data during the day
Server Network Configuration Guidelines
Additional Specific Network Recommendations
512 Documentum System Sizing Guide
Additional Specific Network Recommendations
s Keep roundtrip ping time between servers in a distributed server
configuration to below 250 milliseconds. Roundtrip ping time measures
network latency.
s Land-based communication is better than satellite communication due to
the physical distance data travels using satellite communication.
s Consider adding additional network interface cards to each server to
prevent network saturation.
s To handle large images across the network, consider a second network
card and higher speed media (for example, 100Mbit Ethernet not 10Mbit)
s Place the WAN between the client and eContent Server or the Web Browser
and RightSite (if possible).
s Do not place a WAN between eContent Server and the RDBMS (if
possible).
s Determine whether you can use Documentums network compression
feature to optimize network performance between the client and server.
Documentum System Sizing Guide 61
6
Sizing for Client Applications 1
This chapter describes sizing considerations and requirements for
Documentum clients. The following topics are covered:
s Sizing for Desktop Client on page 6-1
s Sizing for AutoRender Pro on page 6-5
s System Requirements for Client Products on page 6-6
Sizing for Desktop Client
CPU speed and memory are the main resources that must be sized properly
for Documentum Desktop Client. It is also important to size the disk space
and network resources correctly, as these can have a profound impact on
response time.
CPU Speed
Desktop Client offers a wide array of operations and functionality. Generally,
the richer the feature set, the more CPU processing is required. A typical large
enterprise will deploy a range of functionality over a PC population that
varies from very slow (for example, 166Mhz) to extremely fast (for example,
800Mhz). There is a direct relationship between the speed of a machine and
the number of features that can run on the machine with acceptable response
times. Figure 6-1 illustrates this relationship.
Sizing for Client Applications
Sizing for Desktop Client
62 Documentum System Sizing Guide
Figure 6-1 CPU Needs vs. Features used for Desktop Client
Basic usage represents a fixed set of basic operations carried out on
documents that are not extraordinarily large. For example, navigating folders
and checking documents out and in are fairly basic operations and should be
achieved with acceptable performance on a 166MHz server. However, if the
folders in the navigation path have a large number (hundreds) of subfolders
and documents or the documents checked out and in are many megabytes in
size, the 166MHz server probably wont be fast enough to get acceptable
response time.
Table 6-1 shows some example response times for a set of basic office
integration operations on two different servers. (Response times were
measured for the steady state invocation, not the initial operation.)
160Mhz 300MHz 600Mhz
Basic usage: Navigation,
checkout/in, workflow, no OLE
linked documents
Additional usage scenarios:
Business Policy, etc
Advanced usage: OLE Link
processing, XML, large document
processing, custom validations
Table 6-1 Response Time on Two CPUs
Operation Response Time (in seconds) on
400Mhz CPU 166Mhz CPU
Launch App 0.42 1.93
Open Dialog Box 1.58 3.73
Open Small Doc (70K bytes) 2.73 4.71
Sizing for Client Applications
Sizing for Desktop Client
Documentum System Sizing Guide 63
Simple usage of custom validations (user-defined attribute integrity checks
made when importing a document or changing its attributes) is unlikely to
provide good response time with slower machines. And the more advanced
the validation, the more processing is required.
Advanced usage features are those that not only provide more sophisticated
functionality (such as the conversion of OLE linked documents or XML
documents into Documentum virtual documents) but also process large
documents (large in size or in the number of virtual documents created). For
example, the response time for an XML document chunked into thousands of
nodes will be much higher than the response time for one chunked into a few
nodes. Similarly, checking in a document with hundreds of OLE links will
take longer than checking in a document with a single OLE link. Finally,
checking in a 40 Mbyte Powerpoint document takes longer than checking in a
2 Mbyte Powerpoint document.
As the documents get larger and more disk I/O is performed at the PC, disk
performance becomes a larger issue. SCSI drives are typically better choices
for PCs that will be using advanced features than EISA drives.
Component Initialization and Steady State Processing
Some operations take longer the first time they occur in a Desktop Client
session than on second or subsequent executions, due to component
initialization. Those operations display dialog boxes that use data that must
be initialized the first time the dialog box is displayed. Once the start-up
penalty is paid, the data is cached for the duration of the session.
Save Small Doc (70K bytes) 2.74 4.05
Open Medium Doc (300K bytes) 3.99 5.81
Save Medium Doc (300K bytes) 3.01 5.56
Open Large Doc (4M bytes) 4.98 8.58
Save Large Doc (4M bytes) 5.40 12.71
Table 6-1 Response Time on Two CPUs
Operation Response Time (in seconds) on
400Mhz CPU 166Mhz CPU
Sizing for Client Applications
Sizing for Desktop Client
64 Documentum System Sizing Guide
Response times when an operation must initialize components is longer than
when the operation is executing in a steady state (using cached data). Table 6-
2 lists some example initialization and steady-state response times for a set of
operations (on a 400MHz CPU).
If initialization response times on a system are unacceptable, you can reduce
them by increasing the CPU speed of the Desktop Client machine. (Note that
this will also reduce the steady-state response time. There will always be some
difference between initialization and steady-state response times.)
Memory Resource Needs
This section describes memory resources needed for Documentum Desktop
Client and integrations.
The base Explorer integration (using basic features) can run using 64M bytes.
However, actual memory resource needs will vary, depending on:
s The level of features used
s The type of applications used
s Size of the documents being processed and the folders being navigated
Using advanced features, such as converting OLE links to documents, could
require 96 Mbytes and more of memory.
When using the MS Office integrations, the memory required is about 10
Mbytes for each application invoked. (The Office integrations integrate
Desktop Client with MS Word, PowerPoint, and Excel.) If all three are invoked
and are running at the same time, the system requires an extra 30 Mbytes of
memory.
Table 6-2 Example Response Times Comparing Initialization and Steady State Operations
Operation Response Time (in seconds)
First 2nd and Subsequent
Open Dialog Box 10.40 1.58
Open Sm Doc 6.63 2.73
Save Sm Doc 3.65 2.74
Sizing for Client Applications
Sizing for AutoRender Pro
Documentum System Sizing Guide 65
Sizing for AutoRender Pro
AutoRender Pro services rendering requests issued by Documentum clients to
convert documents from one format into another. For example,
WebPublisher uses AutoRender Pro to automatically convert MS Word
documents to HTML when they are checked in. AutoRender Pro is a
single-threaded server application that runs on a server machine separate
from eContent Server. Figure 6-2 illustrates how AutoRender Pro works with
eContent Server and documents.
Figure 6-2 AutoRender Pro in Conjunction with a Docbase
The factors affecting response time for a single rendering job are:
s The CPU speed and disk access speed of the AutoRender Pro server
s The complexity of the work
AutoRender Server machine
Convert
Document
Converted
Document
checked into
Docbase
Queue of
Pending
Rendering
Requests
Document
brought to
AutoRender
Server disk
EContent server machine
Sizing for Client Applications
System Requirements for Client Products
66 Documentum System Sizing Guide
To convert a document, AutoRender Pro must parse and rewrite the
document to a different format. The conversion is CPU intensive, and
document size, complexity (graphics and so forth), and format affect the
amount of resources needed for the job. If the document is large, physical disk
I/Os occur on the AutoRender drives as the documents temporary version is
created. In fact, performance studies show that during the conversion process
both the CPU and the disk are quite busy. If the CPU is slow or if the drives or
their controllers have poor response time, rendering jobs will take longer to
process. Consequently, using good disk controllers, fast drives, and fast CPUs
(> 600MHz) for the AutoRender Pro server machines is strongly
recommended.
Memory requirements are minimal (refer to the release notes for actual
figures).
Multiple AutoRender Pro servers can be set up for a single Docbase. Each
server pulls work requests off the same work queue. Adding more servers
increases the rendering capacity, allowing more users to enter requests
simultaneously while maintaining the desired response time.
Note: It is possible for a single Auto Render Pro server to support multiple
Docbases.
System Requirements for Client Products
Refer to the current product release notes or installation guides for
information about system requirements and certification information. That
documentation is available on the Documentum ftp site, the Documentum
Support Web site, and in the Documentation Library in dm_notes.
Documentum System Sizing Guide A1
A
Additional Workloads A
This appendix contains descriptions of workloads not included in Chapter 2.
It includes the following topics:
s The EDMI Workload on page A-1
s The Web Site Workload on page A-6
s The Document Find and View Workload on page A-9
s The Online Customer Care Workload on page A-9
s Comparing and Contrasting the Workloads on page A-14
s Operations Not Included in Workloads on page A-17
The EDMI Workload
The EDMI workload represents a common software configuration for
Web-based Documentum deployments. The workload uses the following
products:
s eContent Server
s RightSite
Server
s Docbasic
s SmartSpace Intranet Client
The users in this workload are named users accessing the Docbase through
Internet browsers and SmartSpace Intranet Client. The software architecture
for the workload is illustrated in Figure A-1.
Additional Workloads
The EDMI Workload
A2 Documentum System Sizing Guide
Figure A-1 EDMI Software Architecture
This architecture is capable of supporting parallel multi-tier servers (multiple
RightSite and eContent Servers). The architecture allows maximum scalability
and adaptability by ensuring that storage, network, and server capacity can be
added to increase performance and throughput as more clients are added to
the network. A company can easily scale any tier to accommodate growth and
change. The architecture also provides the flexibility for optimizing server and
client hardware.
Workload Scenario
In this workload, a large number of named users work with documents
representing standard operating procedures, work instructions, human
resource notes, corporate Web pages, or other information that must be read
prior to performing some task.
There are three different groups of users: contributors, coordinators, and
consumers. Contributors access and modify documents and submit modified
documents for review using a workflow. Coordinators are a type of
contributor with coordinating workflow tasks. Consumers only read the
documents. Consumers sometimes access documents by logging onto the
system explicitly and at other times only access public Web pages.
WEB
Server
(e.g.,
Microsoft
or
Netscape)
Documentum
RightSite
Server
Documentum
DocPage
Server
RDBMS
OS file
System
Documentum
SmartSpace
(in Browser)
Documentum
SmartSpace
(in Browser)
Browser to
Dyn & static
Web pages
Additional Workloads
The EDMI Workload
Documentum System Sizing Guide A3
Typically, in the type of Web-based deployment modeled by this workload,
the largest number of users are consumers. Consequently, in the workload, 80
percent of the user population are consumers and 20 percent are contributors.
Workload Operations
The operations in the workload include:
s Locating a document through a folder search and viewing it
s Checking out and checking in documents
s Workflow processing (Inbox, routing, and so forth)
s Virtual document processing (publishing)
s Accessing static Web pages stored in the Docbase
s Accessing dynamic Web pages that query the Docbase
Table A-1 lists the operations in the workload.
Table A-1 Operations in the EDMI Workload
Operation Description
CONN Starts a Docbase session through SmartSpace
Intranet.
FOLDER_SRCH Searches folder by folder, eventually displaying a
selected document.
STATIC_HTML Accesses a sequence of Web pages. 20 percent of
the pages are dynamic. 80 percent of the pages are
static (stored in the Docbase).
VDM_PUBLISH Constructs a virtual document from its
components and then displays the document to
the user. Each document has 10 components of 2K
each.
VIEW_INBOX Displays the users Inbox.
CHECKOUT_DOC Checks out a previously selected and viewed
document.
CHECKIN_DOC Checks in an edited document.
Additional Workloads
The EDMI Workload
A4 Documentum System Sizing Guide
Workload Response Time Requirements
When a benchmark test is run, the primary metric obtained is the number of
users who can be supported with acceptable response times.
In the EDMI workload, each user type (contributor, coordinator, consumer)
performs a specific number of random tasks (operations) at random times
during the hour, and the response times for these tasks are measured. Each
task typically consists of dynamically generating several HTML screens from
RightSite.
The interval between a users requests affects performance and response time
because Documentum frees a user connection that does not have any activity
after some amount of time (typically two to five minutes, settable by the
administrator). Re-establishing the session (which happens transparently
when work is initiated on an idle session) consumes more CPU resources.
Simulating this behavior in the test more accurately models the real word.
The acceptable response time is, generally, no more than one to two seconds
per screen. Table A-2 lists the response time requirements for the EDMI
workload. Table A-3 shows a sample set of results.
SUBMIT_REVIEW Submits a modified document to a review
workflow, for a technical and compliance review.
This is a contributors task.
ASSIGN_REVIEW Assigns a document to be reviewed for technical
or regulatory compliance. This is a coordinators
task.
FORWARD_REVIEW Forwards a document to the next activity in the
review workflow.
Table A-1 Operations in the EDMI Workload
Operation Description
Additional Workloads
The EDMI Workload
Documentum System Sizing Guide A5
Table A-2 Response Time Requirements for the EDMI Workload
Operation Number of
Screens
Acceptable
Response Time
per Screen
(in seconds)
Total
Acceptable
Average
Response Time
(in seconds)
CONN 3 2 6
FOLDER_SRCH 5 2 10
STATIC_HTML 5 1 5
VDM_PUBLISH 1 6 6
VIEW_INBOX 1 4 4
CHECKOUT_DOC 2 2 4
CHECKIN_DOC 3 2 6
SUBMIT_REVIEW 3 3 9
ASSIGN_REVIEW 3 3.3 10
FORWARD_REVIEW 2 2.5 5
Table A-3 Example Response Times for the EDMI Workload
Operation Average
Response
Time
(in seconds)
Total
Acceptable
Average
Response Time
(in seconds)
Total
Operations in
One Hour
CONN 2.83 6 2268
FOLDER_SRCH 7.26 10 3860
STATIC_HTML 1.87 5 2981
VDM_PUBLISH 2.3 6 3025
VIEW_INBOX 0.65 4 1742
Additional Workloads
The Web Site Workload
A6 Documentum System Sizing Guide
Workload Scaling
The Docbase size is increased as the number of users in the workload
increases. To support larger numbers of users per hour, many more
documents are preloaded. There are at least 650 documents in the Docbase for
supported user. However, in most benchmarks with this workload, there are
well over 2,000 documents per user. Consequently, as more users are added to
the test, the queries become more expensive to perform. The documents range
in size from 2 Kbytes to 1 Mbyte.
The Web Site Workload
The Web Site workload simulates a common deployment scenario for some of
our Web-based implementations. It uses eContent Server and RightSite Server
and incorporates RightSites ability to retrieve HTML content from the
Docbase transparently for the user. Companies often use RightSite in this
manner to obtain better security and version control for their Web content.
The user population is entirely anonymous. Anonymous users dont need to
provide any security credentials in order to access a page. (Tighter security
can be provided by RightSite named users. The EDMI workload uses named
users.)
CHECKOUT_DOC 1.63 4 1397
CHECKIN_DOC 5.7 6 805
SUBMIT_REVIEW 7.17 9 428
ASSIGN_REVIEW 9.17 10 155
FORWARD_REVIEW 2.64 5 241
Table A-3 Example Response Times for the EDMI Workload
Operation Average
Response
Time
(in seconds)
Total
Acceptable
Average
Response Time
(in seconds)
Total
Operations in
One Hour
Additional Workloads
The Web Site Workload
Documentum System Sizing Guide A7
User access starts at a dynamically generated home page, EDMI_home.htm.
This page includes an HREF reference to all the root-level static Web pages.
(Static Web pages are pages that are stored in the Docbase.) The home page
dynamically constructs the next level of references and serves them as a page
to the user.
The static Web pages are all stored in the cabinet /Website in a folder structure
that is several layers deep. The pages are linked to each other through
<HREF> tags in a hierarchical structure. The structure models the real-world
way of storing Web pages in separate directories (grouped by applications, for
example). The HREF references are five levels deep, with four references per
level. Consequently, in a large Docbase, each root page ultimately points,
directly and indirectly, to thousands of pages.
Each static web page consists of 3000 bytes of random text, followed by up to
four references to other web pages. The pages at the bottom level dont
reference any pages. (Figure A-2, in the next section, illustrates the Web page
structure.)
Workload Operations
There is only one operation in this workload: STATIC_HTML. The
STATIC_HTML operation consists of a sequence of Web page accesses. In 20
percent of the accesses, the page is dynamically generated. In the remaining 80
percent of the accesses, the page is fetched from the Docbase.
The sequence of access operations starts from the home page. A reference is
picked at random from each page, and the operation moves to the next page
until the bottom is reached. Figure A-2 illustrates the access strategy. The
boxes in bold represent the total number of Web pages accessed by one cycle
(six Web page references).
Additional Workloads
The Web Site Workload
A8 Documentum System Sizing Guide
Figure A-2 Static Web Page Example
Workload Response Times
When a benchmark test is run, the primary metric obtained is the number of
users who can be supported with acceptable response times.
Each user performs five tasks. Each task consists of an HTML screen that is
dynamically generated from RightSite plus the additional static pages. The
tasks are performed at random times and the response time is measured.
Acceptable response time is, in general, considered to be no more than one or
two seconds per screen. Table A-4 shows the response time requirements for
the workload.
The Documentum client session time-out has little effect on this workload
because there is a steady stream of activity to only a few anonymous servers
in the RightSite pool. However, their random activities do affect the CPU
EDMI
home
Root
Of
group #1
Root
Of
group
Root
Of
group #3
2
nd
Level
group
3rd
Level
group
4th
Level
group
5th Level
group #2
2
nd
Level
group #2
2
nd
Level
group #2
3rd Level
group #2
3rd Level
group #2
4th Level
group #2
4th Level
group #2
5th
Level
group
5th Level
group #2
Table A-4 Response Time Requirements for the Web Site Workload
Operation Number of
Screens
Acceptable
Response Time per
screen (in seconds)
Total Acceptable
Average Response
Time (in seconds)
STATIC_HTML 5 1 5
Additional Workloads
The Document Find and View Workload
Documentum System Sizing Guide A9
resources used, because RightSite spawns more servers to handle the excess
load when all anonymous servers are busy. Random user activity helps ensure
that more servers are spawned during a run, which ensures that the test
models real world activity peaks more accurately.
Workload Scaling
The Docbase size is increased as the number of users in the workload
increases. There are at least 40 Web pages in the Docbase for each anonymous
user supported. However, this workload is usually part of a configuration
used in the EDMI benchmark and, consequently, the static Web pages are only
20 percent of the total number of documents stored in the Docbase. The Web
pages are all 3K bytes.
The Document Find and View Workload
The Document Find and View workload is essentially a small subset of the
EDMI workload, exercised in a client-server environment using WorkSpace
rather than in a Web environment using SmartSpace Intranet. Users locate
documents using attribute searches or by navigating through folders. After a
document is found, it is displayed to the user. Each document is 400K bytes.
The Online Customer Care Workload
The online customer care workload demonstrates Documentum interactive
performance for a large number of users on a very large Docbase (100,000,000
content objects). This workload uses out-of-the-box SmartSpace with a few
customizations. The customizations are centered on querying and setting
custom attributes.
The workload includes the following Documentum products commonly used
in Web-based deployments:
s eContent Server
s RightSite
s Docbasic
Additional Workloads
The Online Customer Care Workload
A10 Documentum System Sizing Guide
s SmartSpace Intranet Client
Workload Operations
In this workload, a large number of users work with documents representing
customer and supplier correspondence or standard operating procedures that
must be read prior to performing some task. The users access the Docbase
through Internet browsers.
The operations mimic the content management needs of businesses such as
insurance or financial services for online customer care. These needs are
defined by large volumes of images (customer correspondence or policy
agreements) and large groups of users who read the documents or modify
their attributes. The operations include:
s Locating a document (through a folder search)
s Inserting documents and notes into the Docbase and viewing them
s Workflow processing (Inbox, routing, and so forth)
s Accessing dynamic Web pages that query the Docbase
Table A-5 lists the basic operations in the workload.
Table A-5 Online Customer Care Workload Operations
Operation Description
CONN Starts a Docbase session using SmartSpace Intranet.
CREATE_DOCUMENT Creates a new document and opens the documents
editor for the user.
CREATE_DOCUMENT_F Checks in the new document.
FOLDER_SRCH Searches through a folder hierarchy for a document.
This is performed by office users.
FORWARD_REVIEW Forwards a document for review in a workflow.
This operation occurs when a workflow or data
entry user completes a task.
QUERY_SRCH This represents a Docbase search based on a
customer ID.
Additional Workloads
The Online Customer Care Workload
Documentum System Sizing Guide A11
In the workload, the activities performed by a user are made up of the actions
described in Table A-5 grouped in a way that is meaningful for the users
function. Table A-6 describes the user functions and their associated activities.
SET_PROPERTIES Sets the classification attributes of a TIFF document.
Data entry users perform this operation.
VIEW_DOCUMENT Displays a TIFF, Word, text, or Excel document for a
user to view.
VIEW_INBOX Selects a users Inbox icon and displays the next set
of tasks in the Inbox.
Table A-5 Online Customer Care Workload Operations
Operation Description
Table A-6 User Activities in the Online Customer Care Workload
User Function Description of User Activity
Data entry user Data entry users examine correspondence that was scanned
into the Docbase to ensure that attributes are correctly set.
The data entry users work on documents that have been
routed to them through their Inboxes.
A data entry users activity includes the following
operations:
s Check inbox for items to re-attribute (VIEW_INBOX)
s View TIFF image noted by next inbox item
(VIEW_DOCUMENT)
s Change the attributes to fit desired descriptions and
categories (SET_PROPERTIES)
s Complete workflow task (FORWARD_REVIEW)
Additional Workloads
The Online Customer Care Workload
A12 Documentum System Sizing Guide
Workflow users Workflow users review and sometimes annotate documents
routed to them. (The documents are routed prior to the busy
hour.) A workflow users activity includes the following
operations:
s Check inbox for work (VIEW_INBOX)
s View document designated by next item in the inbox
(VIEW_DOCUMENT)
s Complete the task (FORWARD_REVIEW)
s For 30 percent of the users, create a text note associated
with the document (CREATE_DOCUMENT &
CREATE_DOCUMENT_F)
Workflow users continue activities until all items in their
Inboxes are processed.
Office users Office users create MS Office Word and Excel documents. An
office users activity includes the following operations:
s Create Word document (CREATE_DOCUMENT &
CREATE_DOCUMENT_F)
s Create Excel document (CREATE_DOCUMENT &
CREATE_DOCUMENT_F)
s Navigate some folders (FOLDER_SEARCH)
s View a text document (VIEW_DOCUMENT)
Branch users Branch users search for documents using a policy number or
customer ID and then view the documents. A branch users
activity includes the following operations:
s Query for document based on customer ID
(QUERY_SRCH)
s View TIFF image (VIEW_DOCUMENT)
Table A-6 User Activities in the Online Customer Care Workload
User Function Description of User Activity
Additional Workloads
The Online Customer Care Workload
Documentum System Sizing Guide A13
Workload Response Time Requirements
When a benchmark test is run, the primary metric obtained is the number of
users who can be supported with acceptable response times.
Each user type performs a specific number of random tasks at random times
during the hour, and the response time for these tasks are measured. Each task
typically consists of several HTML screens that are dynamically generated
from RightSite. Acceptable response time is, in general, considered to be no
more than one to two seconds per screen.
The interval between a users requests affects performance and response time
because Documentum frees a user connection that does not have any activity
after some amount of time (typically two to five minutes, settable by the
administrator). Re-establishing the session (which happens transparently
when work is initiated on an idle session) consumes more CPU resources.
Simulating this behavior in the test more accurately models the real word.
Another important metric is obtained when response times are measured for
operations on the documents in the Docbase. For tests using this workload,
the Docbase is loaded with 100,000,000 TIFF images, to model a multi-user
environment after years of use. To preserve space but still have a sufficient
number of database rows, 90 percent of the content objects have zero length.
Call Center Users Call center users staff phone banks, taking customer calls.
The user receives a call and queries the customers records
based on the customers ID. Sometimes, the call center user
must put some additional information in the Docbase in
response to the call.
s Query for a document based on customer ID
(QUERY_SRCH)
s View TIFF image (VIEW_DOCUMENT)
s For 25 percent of the activities, create a text document
associated with this image (CREATE_DOCUMENT &
CREATE_DOCUMENT_F)
s Otherwise query again once complete.
Table A-6 User Activities in the Online Customer Care Workload
User Function Description of User Activity
Additional Workloads
Comparing and Contrasting the Workloads
A14 Documentum System Sizing Guide
When the benchmark test starts, an array of policy numbers corresponding to
documents with non-zero length content is loaded. All queries for documents
choose from this array. In this way, only documents with content are queried.
Content sizes range from 50K to 500K bytes.
Table A-7 lists the response time requirements for the workload.
Comparing and Contrasting the Workloads
This section compares and contrasts the workloads in terms of the software
architecture, the usage patterns modeled in each, and the resulting resource
consumption.
Software Architecture
The EDMI and Web site workloads use an HTTP thin-client or 4-tier
architecture. In an HTTP thin-client architecture, Documentum DMCL (client
library) processing occurs on the machine that hosts RightSite and the Internet
Server. This is in contrast to the 3-tier architecture, in which client library
Table A-7 Response Time Requirements for the Online Customer Care Workload
Operation Number of
Screens
Acceptable
Response Time per
Screen (in seconds)
Total Acceptable
Average Response
Time (in seconds)
CONN 3 2 6
CREATE_DOCUMENT 6 2 12
CREATE_DOCUMENT_F 5 2 10
FOLDER_SRCH 4 2 8
FORWARD_REVIEW 2 2 4
QUERY_SRCH 3 2 6
SET_PROPERTIES 2 2 4
VIEW_DOCUMENT 1 2 2
VIEW_INBOX 2 2 4
Additional Workloads
Comparing and Contrasting the Workloads
Documentum System Sizing Guide A15
processing occurs on the users PC. With HTTP thin-client architecture, very
little work actually happens on the client machine (all users are assumed to be
using browsers sending HTTP). That is, RightSite performs those operations
that, in a 3-tier architecture, are performed on the hundreds (or even
thousands) of client PCs. Figure A-3 illustrates the difference between 3-tier
architecture and HTTP thin-client architecture.
Figure A-3 Client-Library 3-Tier Architecture vs 4-Tier (HTTP thin-client) Architecture
Usage Models and Resource Consumption
The EDMI and Anonymous RightSite Web Site workload usage models are
opposites.
In the EDMI workload, all the users are named users. Named users provide a
user name and password and then are authenticated and provided with some
exclusive resources. In particular, a separate RightSite process is created for
each named user.
The tasks in the EDMI workload operate primarily on dynamic Web pages.
The named users use RightSite with SmartSpace to generate dynamic pages in
80 percent of the accesses and fetch static pages from the Docbase in 20
percent of the accesses.
Documentum DMCL operations
(3 Tier Mode)
Hundreds to Thousands of Individual user PCs
Documentum DocPage Server
Thin-client HTTP
Internet Server + Documentum RightSite
On Centralized Middle-tier Server
(Documentum DMCL Operations)
Documentum DocPage Server
Hundreds to Thousands of Individual user PCs
Additional Workloads
Comparing and Contrasting the Workloads
A16 Documentum System Sizing Guide
In the Web site workload, all users are anonymous users. Anonymous users
dont provide a name or password; they share the anonymous login
configured with RightSite. On the resource side, anonymous users share a
pool of RightSite processes, rather than having their own resources.
The tasks in the Website workload operate primarily on static Web pages. The
anonymous users use RightSite only (SmartSpace is not used) to generate
dynamic pages in 20 percent of the accesses and to fetch static pages in 80
percent of the accesses. This dynamic-to-static Web page profile is opposite
the one used in the EDMI workload.
The difference in the dynamic-to-static Web page profiles means that the
EDMI workload makes much heavier demands on the RightSite Server than
the Website workload does. In benchmarks using the EDMI workload, the
RightSite Server will consume more CPU than any other piece of server
software included in the benchmarks because the dynamic-to-static page ratio
makes heavy demands on the RightSite Server. Figure A-4 illustrates the
relationship between CPU consumption and dynamic-to-static ratio.
Figure A-4 Relationship Between Workloads and RightSite CPU Consumption
The difference in the user profiles means that the EDMI workload will
consume more CPU and memory resources per user than the Web site
workload, because named users consume more resources than anonymous
users.
Additional Workloads
Operations Not Included in Workloads
Documentum System Sizing Guide A17
Document Find and View Workload
The Document Find and View workload is client-server and named. RightSite
is not part of that workload. In addition, although the EDMI workload is Web-
based, in most cases the RightSite portion can be factored out of the data to
allow you to size this workload based on a client/server model.
Operations Not Included in Workloads
The following operations are not included in the workloads described in this
appendix:
s Creating PDF renditions
s Creating or searching full-text indexes
s Default SSI attribute searching (case insensitive attribute searches)
s Deleting documents from the Docbase
s Dumping and loading a Docbase
s Distributed content operations and object replication
s Operations on objects in turbo storage
If these are part of your workload, you may want to increase the expected
resource consumption for your workload.
Additional Workloads
Operations Not Included in Workloads
A18 Documentum System Sizing Guide
Documentum System Sizing Guide Index-1
I NDE X
A
active user 1-4
active user - in transaction 1-4, 2-3
active user - out of transaction 1-4, 2-3
activity timeout 1-5
AIX F50 (IBM server) 4-17
AIX S7A (IBM server) 4-16
Anonymous RightSite Web Site workload.
See Web Site workload
AutoRender Pro
sizing 6-5
in WCM Edition 3-19
availability considerations, for system 3-12
B
backup capacity and sizing 3-12
bandwidth
defined 1-5, 5-3
response times, affect on 5-4
vs localizing 5-6
vs remote Web servers 5-8
benchmark tests
Document Find and View workload on
HP K580 machines 4-21
EDMI workload
on AIX servers 4-17
on IBM Netfinity 7000 M10 4-15
on LXR8000 & LH4 4-20
on Sun Enterprise 450 4-11
on Sun Enterprise 6500/4500 4-13
focus for N-tier tests 4-5
hardward configurations 4-2
iTeam workload
on Compaq servers 4-10
on HP LXR8000 & LH4 4-19
on Lpr/LH4 4-19
Online Customer Care workload on
V2600 4-21
result tables, interpreting 4-6
Web site workload
on Sun Enterprise 6500/4500 4-14
Web Site workload on Sun Enterprise
450 4-11
bottleneck, defined 1-5
busy hour
active sessions, estimating 2-8
defined 2-6
C
caches
affect on performance 4-26
database 4-27
eContent Server 4-27
memory use 4-26
RightSite 4-27
capacity planning. See system sizing
cluster, eContent Server 3-8
Compaq servers
described 4-9
Web site for 4-22
configurations
host-based vs multi-tiered 3-11
connected user 1-5
connecting user 1-5, 2-3
connection states. See user connection states
content
replication, described 3-15
servers 3-14
transfers, response times 5-9
ContentCaster 3-20
cost-based optimizers 4-38
CPU usage
Documentum server 2-4
Documentum workloads 2-24
RightSite Server 2-5
Index-2 Documentum System Sizing Guide
D
data caches, database 4-27
database
caches
described 4-27
disk I/O and 4-39
license sizing 4-46
memory requirements 4-29
scaling 3-10
server 1-5
DBMS. See databases
Desktop Client sizing
CPU speed 6-1
dmcl operations 6-3
memory 6-4
disk access capacity 4-37
disk capacity sizing
overview 4-35
process inputs 4-36
query optimizers and 4-38
space vs access capacity 4-37
tables scans, affect on 4-38
disk I/O
database cache and 4-39
of disk storage areas 4-42
disk space sizing
formula for 4-45
general notes 4-46
server software requirements 4-44
space vs access speed 4-37
disk storage areas
sizing 4-42
disk striping
defined 4-40
with parity 4-42
without parity 4-41
disk throughput 1-6
distributed storage areas 3-14
DL360 (Compaq server) 4-9
DMCL
object cache 4-28
response times 5-11
Docbase size and memory requirements
4-29
Docbasic compiled-code memory area 4-28
DocBrokers
load balancing 3-10
scaling 3-10
DocPage Server 1-5
Document Find and View workload
described A-9
on HP K580 machines 4-21
Documentum Server
defined 1-5
transformation engine 1-7
Documentum Sizing Spreadsheet 1-3, 2-2
Documentum workloads
Docbase usage patterns 2-25
Document Find and View A-9
EDMI A-1
iTeam 2-10
Load and Delete workload 2-22
Online Customer Care A-9
operations not included 2-26, A-17
resource consumption 2-24, A-15
software architecture 2-23, A-14
Web Site A-6
WebPublisher 2-16
dynamic HTML, memory use 4-28
dynamic Web pages, affects on scaling 3-20
E
e-Content Server 1-6
eContent Server
caches 4-27
cluster 3-8
defined 1-6
load balancing 3-8
scaling 3-7
Documentum System Sizing Guide Index-3
server set 3-8
editions
Portal 3-21
Web Content Management 3-17
EDM Server 1-6
EDMI workload
on AIX servers 4-17
described A-1
execution scenario A-2
on IBM Netfinity 7000 M10 4-15
on LXR8000 & LH4 4-20
operations A-3
response time requirements A-4
scaling A-6
on Sun Enterprise 450 4-11
on Sun Enterprise 6500/4500 4-13
Enterprise 450 (Sun server) 4-11
Enterprise 4500 (Sun server) 4-12
Enterprise 6500 (Sun server) 4-12
Enterprise Resource Planning system,
Documentum and 4-22
errors in system sizing 1-3
F
F50 (AIX server) 4-17
failover
operating system solutions 3-13
partitioning and 3-5
federations 3-15
G
global type cache 4-27
H
hardware configurations
in benchmarks 4-2
choosing configuration 3-11
host-based configurations 3-11
HP servers
K580 4-20
LH4 4-19
Lpr 4-19
NETSERVER LXR 8000 4-18
V2600 4-20
Web sites for 4-22
HTTP Server 1-6
I
IBM servers
AIX F50 4-17
AIX S7A 4-16
Netfinity 7000 M10 4-15
Web site for 4-22
inactive users
resource consumption 2-5
inactive users, defined 1-6, 2-3
indexes and disk capacity 4-38
iTeam in Portal edition 3-21
iTeam workload
on Compaq servers 4-10
execution scenario 2-11
on HP LXR8000 & LH4 4-19
on Lpr/LH4 4-19
operations 2-12
purpose 2-10
resource consumption 2-24
response times
examples 2-16
requirements 2-15
task performance and 2-14
scaling 2-13
K
K580 (HP-UX server) 4-20
L
latency
defined 5-3
network 1-6
LH4 (HP server) 4-19
Index-4 Documentum System Sizing Guide
Load and Delete 2-22
load balancing
DocBrokers 3-10
eContent Server 3-8
network 3-6
transparent 3-5
Lpr (HP server) 4-19
M
memory
physical 1-6
virtual 1-7
memory sizing
cache usage 4-26
database requirements 4-29
for Desktop Client 6-4
Docbase size, affect on 4-29
dynamic HTML 4-28
equation for 4-32
examples 4-32 to 4-35
guidelines, general 4-31
for MS Office integrations 6-4
operating system 4-30
operating system requirements 4-29
oversizing 4-24
overview 4-23
paging file 4-30
RightSite 4-27
user connection requirements 4-28
virtual memory 4-25
metadata retrieval, affects on scaling 3-20
Microsoft Windows NT
Web site 4-22
mirroring, in disk striping 4-42
MS Office integrations memory use 6-4
multi-site deployments
administrative overhead 3-16
deployment options, list of 3-14
network bandwidth 3-16
multi-tiered configurations
advantages 3-11
availability and 3-12
N
name servers. See DocBrokers
named user 1-6
Netfinity 7000 M10 (IBM Windows NT
machine) 4-15
NETSERVER LXR 8000 (HP server) 4-18
network
latency 1-6
load balancer 3-6
throughput 1-6
network sizing
bandwidth
cost of 3-16
vs localizing 5-6
vs response time 5-4
for content transfer speed 5-9
for DMCL operations 5-11
guidelines, general 5-12
overview 5-1
N-tier configurations
benchmark test focus 4-5
described 4-3
O
object replication, described 3-15
Online Customer Care workload
described A-9
on HP V2600 machines 4-21
operations A-10
response time requirements A-13
user activities A-11
operating systems
memory requirements 4-30
physical memory, tuning 4-29
Documentum System Sizing Guide Index-5
P
paging file
memory use, estimating 4-30
out of memory detection 4-31
sizing 4-23
usage 4-25
on Windows NT 4-31
parity, in disk striping 4-41
partitioning
and failover 3-5
RDBMS 3-10
scaling out 3-3
scaling up 3-2
Web tier software 3-6
performance and cache use 4-26
physical memory
cache usage 4-26
calculation examples 4-32 to 4-35
defined 1-6
global type cache 4-27
large memory support 4-29
operating system, configuring 4-30
process working set 4-25
sizing 4-23
user connection requirements 4-28
Portal Edition, scaling 3-21
primary domain controller, Documentum
and 4-22
process working sets
defined 4-25
on Windows NT 4-30
Proliant servers (Compaq) 4-9
Q
query optimizers, affects on table scans
4-38
R
RAID configurations 4-40
reference links, described 3-15
relational databases. See databases
remote Web servers
defined 3-14
vs bandwidth 5-8
resource consumption
busy hour 2-6
Documentum workloads 2-24, A-15
dynamic and static web pages A-16
inactive users 2-5
process working set 4-25
RightSite 2-5
user connection states and 2-2
response times
Desktop client 6-3
examples
iTeam workload 2-16
WebPublisher workload 2-21
requirements
EDMI workload A-4
iTeam workload 2-15
Online Customer Care workload
A-13
Web Site workload A-8
WebPublisher workload 2-20
user expectations 2-9
vs bandwidth needs 5-4
RightSite
active sessions, estimating 2-8
described 1-6
DMCL object cache 4-28
user connection states 2-5
rule-based optimizers 4-38
S
S7A (AIX machine) 4-16
scaling
across sites 3-14
databases 3-10
DocBrokers 3-10
dynamic Web access and 3-20
Index-6 Documentum System Sizing Guide
eContent Server 3-7
iTeam workload 2-13
out 3-3, 3-4, 3-5
Portal Edition 3-21
trends affecting 3-1 to 3-5
up 3-2, 3-5
Web Content Management Edition
3-17, 3-19
Web tier software 3-5
server machines
Compaq 4-9
Sun Solaris 4-10
server sizing
CPU use
RightSite and WDK/App server
4-5
guidelines, general 4-22
overview 4-1
server CPU use 2-4
server software
configurations, compared A-14
host-based configurations 4-3
N-tier configurations 4-3
partitioning
across multiple hosts 3-3, 3-4
on one host 3-2
reuse 3-2
Site Delivery Services 3-20
Solaris memory sizing 4-31
spreadsheet, sizing 1-3, 2-2
striping. See disk striping
Sun servers
Enterprise 450 4-11
Enterprise 6500 & 4500 4-12
Web site for 4-22
swap file. See paging file
system sizing
AutoRender Pro 6-5
benchmark result tables, interpreting
4-6
benchmark test results 4-4 to 4-21
configuration constraints 4-22
database license sizing 4-46
Desktop Client 6-1
disk capacity 4-35
disk space sizing formula 4-45
disk space vs access capacity 4-37
disk storage areas 4-42
disk striping 4-40
Documentum Spreadsheet 1-3, 2-2
glossary of terms 1-4
hardware configurations 3-11
for high availability 3-12
memory calculation examples 4-32 to
4-35
mistakes, common 1-3
network
bandwidth needs 5-4
overview 5-1
overview 1-1
paging file 4-30
requirements of 1-2
scaling
across sites 3-14
databases 3-10
DocBrokers 3-10
Web tier software 3-5
server
guidelines 4-22
memory 4-23
process overview 4-1
software requirements 4-44
software reuse and 3-2
table scans, affect on 4-29
user deployment considerations 3-4
virtual memory 4-25
workloads
defined 2-1
estimating 2-2
Documentum System Sizing Guide Index-7
T
table scans
disk capacity, affect on 4-38
query optimizers and 4-38
throughput, defined 1-6
transactions 1-7
transformation engine 1-7
U
user connection states
active user 1-4
active user - in transaction 1-4, 2-3
active user - out of transaction 1-4, 2-3
connected user 1-5
connecting user 1-5, 2-3
described 1-7
inactive user 1-6
inactive users
defined 2-3
memory requirements 4-28
paging file use 4-30
resource consumption 2-2
RightSite Server 2-5
V
V2600 (HP-UX server) 4-20
virtual memory 1-7, 4-25
vmstat UNIX utility 4-31
W
WCM Edition. See Web Content
Management Edition
Web Content Management Edition
AutoRender Pro and 3-19
described 3-17
scaling requirements 3-19
Site Delivery Services and 3-20
WebPublisher and 3-19
Web Site workload
described A-6
operations A-7
response time requirements A-8
scaling A-9
on Sun Enterprise 450 4-11
on Sun Enterprise 6500/4500 4-14
Web tier software, scaling 3-5
WebCache 3-20
WebPublisher 3-19
WebPublisher workload
operations
list of 2-18
overview 2-17
purpose 2-16
resource consumption 2-24
response times
examples 2-21
requirements 2-20
task performance and 2-20
scaling 2-19
Windows NT paging file 4-31
workloads
See also Documentum workloads
active sessions, estimating 2-8
busy hour 2-6
defined 2-1
estimating 2-2
response time expectations 2-9
use in sizing 2-9
Index-8 Documentum System Sizing Guide