Networking Computers from Different Geographical positions.

Oracle Golden Gate
an overview
Oracle Golden Gate delivers low-impact, real-time, data acquisition, distribution and delivery across heterogeneous systems. Using this technology, it enables costeffective and low-impact real-time data integration and continuous availability of solutions. Oracle Golden Gate is the leading real-time data integration software available in the industry. The software moves transactional data across heterogeneous database, hardware and operating systems environment with minimal impact. The software platform captures, routes, transforms and delivers data in real-time enabling organizations to maintain continuous uptime for critical applications during planned and unplanned outages. Additionally it moves data from transaction processing environments to read-only reporting databases and analytical applications for accurate, timely reporting and improved Business Intelligence for the enterprise.

It has the ability to feed third party ETL solutions. It can be deployed unidirectional or bi-directional in multiple topologies including one to many. open systems/open source. High performance and low overhead even at high volumes. Moving only committed transaction with compression minimizes network overhead. Isolation and Durability) of transactions during the data movement. . many to one and cascading. Applies the data in the same order it was commit in the source database. Transaction Integrity : Maintains ACID (Atomicity.Features of Golden Gate High Performance : Continuous capture and delivery of data from sources to targets with end-to-end total latency of only sub-seconds. Reliability : Enables guaranteed delivery and data integrity after interruptions / failures. Flexibility / Extensibility : Capture and delivery of data between a variety of relational. and legacy databases on all major platforms. Low-Impact : Log-based capture and queueing changed data outside of the DBMS results in negligible overhead source system. many to many. Consistency.

Heterogeneity : Supports all major databases and platforms allowing companies to use the same product for all their real-time data integration and continuous data availability needs.Benefits Improved Business Sight : Enables fresh data for better decision making by feeding analytical systems from OLTP systems with sub-second latency. . Continuous System Availability : Eliminates planned and unplanned outages for mission critical systems to allow uninterrupted business operations. Reduced Risks : Offers data integrity and reliability between source and target systems while providing resilience against network / site outages. Lowered Integration Costs : Reads data base log files and moves only committed transactions to minimize overhead on the infrastructure while augmenting existing data integration investments.

It s modular architecture gives you the flexibility to extract and replicate selected data records. transformation and custom processing features of Golden Gate you can support numerous business requirements as follows: ‡ Business Continuance and high availability ‡ Initial Load and database migration ‡ Data Integration ‡ Decision Support and data warehousing . heterogeneous platforms across the enterprise.OGG Supported processing methods and databases: Golden Gate enables the exchange and manipulation of data at the transaction level among multiple. transactional changes across variety of topologies. With this flexibility and the filtering.

.

Golden Gate is composed of the following components: ‡ ‡ ‡ ‡ ‡ ‡ ‡ Extract Data Pump Replicat Trails or Extract Files Checkpoints Manager Collector .

Logical Architecture of the Golden Gate .

Extract extracts a current set of data directly from their source objects. Extract extracts transactional changes made to data (i. inserts. updates and deletes) after the initial synchronization has taken place.e. Change Synchronization : To keep source data synchronized with another set of data. Extract obtains the data from a data source that can be one of the following : 1) The database transaction logs (such as oracle redo logs). This method is also known as log based extraction..Overview of Extract The Extract process runs on the source system and is the capture mechanism of the Golden Gate. . When processing transactional data changes. Extract can be configured in two ways as follows: Initial Loads : For Initial Data Loads.

dsc.*. Or two Extract processes could extract and transmit in parallel to 2 Replicat processes to minimize target latency when the databases are large. Sample Extract parameter file EXTRACT capt USERID ggs. MGRPORT 7809 RMTTRAIL /ggs/dirdat/aa TABLE fin.*. PURGE RMTHOST sysb. Multiple extract processes can operate on different objects at the same time for example one process could continuously extract transactional data changes and stream them to a decision-support database. while another process performs batch extracts for periodic reporting.2) A Golden Gate Vendor Access Module (VAM). . PASSWORD ********* DISCARDFILE /ggs/capt. The VAM is a communication layer that passes data changes and transaction metadata to the Extract process. TABLE sales.

When the connectivity is restored. If the network on the target system becomes unavailable. there is nowhere on the source system to store data that Extract continuously extracts into memory. preventing the abend.Overview of Data pumps : A Data pump is a secondary extract group within the source Golden Gate configuration. In a typical configuration that includes a data pump. the primary Extract could run out of memory and abnormally end. a data pump can be configured for either online or batch processing. If a data pump is not used. however the primary Extract group writes to a trail on the source system. Extract must send data to a remote trail on the target. Like a primary extract group. However with a trail and data pump on the source system. The data pump reads this trail and sends the data across the network to a remote trail on the target. . captured data can be moved to disk. Protection against network and target failures: In a basic Golden Gate configuration. the data pump extracts the data from the source trail and sends it to the target system. Using data pump is one of the best practice. Reasons for using data pump are as follows: 1. with only a trail on the target system.

. 4) Synchronizing one source with multiple target : When sending data to multiple target systems. Dividing the storage load between the source and target systems reduces the need for massive amounts of space on the target systems to accommodate data arriving from multiple sources. 3) Consolidating data from many sources to a central target : When synchronizing multiple source databases with a central target database. and then use another data pump or the Replicat group to perform the second transformation. you can configure a data pump to perform the first transformation either on the source system or on the target system. data can still be sent through other targets. you can store expected data on each source system and use data pumps on each of those systems to send the data to a trail on the target system.2) You are implementing several phases of data filtering or transformation : When using complex filtering or data transformation configurations. you can configure data pumps on the source system for each target. If network connectivity to any of the targets fails.

Replicat can apply data to target objects or route them to a high speed bulk-load utility. We can use multiple Replicat processes with multiple Extract processes in parallel to increase throughput. A delay may be desirable to control data arrival across different time zones.Overview of Replicat : The Replicat system runs on the target system. Replicat reads extracted data changes that are specified in the Replicat configuration. Replicat can be configured as one of the following : 1) Initial Loads : For Initial Loads. statement caches and local database access. and then it replicates them to the target database. 2) Change Synchronization : To maintain synchronization Replicat applies extracted transactional changes to target objects using native database calls. You can delay replicat so that it waits a specific amount of time before applying data to the target database. .

TARGET fin. balance = bal.teller. branch = branch). PASSWORD **** SOURCEDEFS /ggs/dirdef/defs DISCARDFILE /ggs/deliv. WHERE (branch = ³LA´).acctab.telltab. MAP fin.teller.telltab. IGNOREINSERTS MAP fin. COLMAP (account = acct.account.Sample Replicat parameter file REPLICAT deliv USERID ggs. TARGET fin. PURGE GETINSERTS MAP fin. . WHERE (branch = ³NY´). TARGET fin.dsc.

Only one extract process can write to a trail. Golden Gate stores those changes temporarily on disk in a series of files called Trail. Replicat : Reads a trail to apply change data to the target database. . For example. With those processes separated. depending on how you configure Golden Gate. A Trail can exist on the source or target system. The use of Trail also allows extraction and replication activities to occur independently of each other. you have more choices for how data is delivered. and transfer it to the target system or to the next Golden Gate process downstream in the Golden Gate configuration. Read and Write process to Trail : The primary extract process writes to a trail. If needed.Overview of Trail : To support the continuous extraction and replication of supported database changes. Processes that read the trail are : Data dump extract : Extracts data from a local trail for further processing. intead of extracting and replicating changes continuously but store them in the Trail for replication to the target later whenever the target layer need them.

All file names in a trail begin with the same two characters. How processes write to Trail : To maxize throughput. Extract operates in two modes : Append mode and Overwrite mode. . each name is appended with a unique. However.Trail Maintenance: Trail files are created as needed during processing. Transactional order is preserved. Ex. trails are stored in the dirdat sub-directory of the Golden Gate directory. Golden Gate writes data to the trail in universal data format. six-digit serial (unique) number from 000000 to 999999. By default. and to minimize Input/Output load on the system. D:\ggs\dirdat\tr000009. a proprietary format which allows it to be exchanged rapidly and accurately among heterogeneous databases. By default. extracted data is sent into and out of a trail in large blocks. The file header contains the information about the trail file itself.0 each file of a trail contains a file header record that is stored at the beginning of the file. Trail file format : As of Golden Gate version 10. By default each file in a trail is 10MB in size. the data can be written in other formats that are compatible with different applications. As the files are created.

In this sense.Each data record in a Golden Gate trail file contains a header area and data area. The file or files are created automatically during the run. The extract file typically is a single file but can be configured to roll over into multiple files in anticipation of limitations on the size of a file that are imposed by the operating system. The same versioning features that apply to trails also apply to extract files. it is similar to a trail. Overview of Extract files : When processing a one-time run. such as an initial load or a batch run that synchronizes transactional changes Golden Gate stores the extracted changes in a extract file instead of a trail. and the data area contains the actual data values that were extracted. . except that check points are not recorded. The header contains Information about the transaction environment.

Thus if Golden Gate needs to re-read something that is already sent to the target system (for example in case of a process failure) checkpoints enable accurate recovery to the point where a new transaction starts. A batch process can be re-run from its start point. Checkpoints work with inter process acknowledgements to prevent messages from being lost in the network. and Golden Gate resume processing. A checkpoint system is used for Extract and Replicat processes that operate continuously. whereas continuous processing requires the support for planned or unplanned interruptions that is provided by checkpoints. . Extract creates checkpoints for its positions in the data source and in the trail. Checkpoints ensure that database changes marked for synchronization are extracted by Extract and replicat and they prevent redundant processing. The read checkpoint of a process is always synchronized with write checkpoint. Replicat creates checkpoints for its position in the trail.Overview of Checkpoints : Checkpoints store current read and write positions of a process to disk for recovery purpuses. but it is not required for batch mode.

for example when throughput slows down or synchronization latency increases. Manager performs the following functions : Monitor and restarts Golden Gate processes. One manager can process can control many Extract or Replicat processes. Manager must be running on each system in the Golden Gate configuration before Extract or Replicat can be started and manager must remain running while those processes are running so that resource management functions are performed. in addition to a standard checkpoint file. Overview of Manager : Manager is the control process of Golden Gate. . Maintain trail files and logs. Report errors and events.Checkpoint information is maintained in checkpoint files within the dirchk subdirectory of the Golden Gate directory. Issue threshold reports. Optionally Replicat checkpoints can be maintained in a checkpoint table within the target database. Receive and route user requests from the user interface. Allocate data storage space.

This is known as static collector. Collector receives extracted database changes that are sent across the TCP/IP network and it writes them to a trail or extract file. it can receive information from only one Extract process. several extract processes can share one collector. We can run collector manually. When a dynamic collector is used. When manager starts collector the process is known as Dynamic collector. . However a one-to-one ratio is optimal. Manager starts collector automatically when a network connection required.Overview of Collector : Collector is a process that runs in the back ground on the target system. so there must be a dynamic collector for each extract that you use. When a static collector is used.

Sign up to vote on this title
UsefulNot useful