Professional Documents
Culture Documents
FINANCE TRANSFORMATION
TARDIS (Transparent Actuarial Reporting Database Insight – UNISURE)
SEWPZG628T DISSERTATION
By:
Sagar Agrawal
2013HW70753
SEWPZG628T DISSERTATION
FINANCE TRANSFORMATION
TARDIS (Transparent Actuarial Reporting Database Insight – UNISURE)
By:
Sagar Agrawal
2013HW70753
Acknowledgements
I am highly indebted to Priyanka Katare for all the guidance and constant
supervision as well as for providing necessary information regarding the project &
also for the support in completing the project.
I would like to express my gratitude towards my examiners, Miss Sona Jain & Mr.
Naman Pandey for their kind co-operation and encouragement which help me in
completion of this project.
Contents
1 Introduction..........................................................................................................................................................9
1.1 Purpose.........................................................................................................................................................9
1.2 Business Background..................................................................................................................................10
1.3 Approach.....................................................................................................................................................11
1.4 Scope...........................................................................................................................................................13
4 Archival Strategy...............................................................................................................................................25
6 Non-Functional Requirements..........................................................................................................................27
8 Source Data.........................................................................................................................................................29
8.1 Source Data Specification...........................................................................................................................29
8.2 Extract Process...........................................................................................................................................30
8.3 Staging Source Data...................................................................................................................................30
8.4 Source Data Dictionary..............................................................................................................................31
8.5 Identify the Movement Logic.......................................................................................................................34
9 Database Overview............................................................................................................................................35
1 Introduction
1.1 Purpose
This document is a Finance Transformation deliverable and represents the interface macro
design for the UNISURE, covering the movement and transformation of data from the UNISURE
source system extract through to the Persistent Actuarial Database (PAD) for Product Families
Unitized NP Pensions, Unitized WP Pensions and Term Assurance. It will serve as the major
source input into the development of off-shore Technical specification for the same functional
scope.
This macro design is one of a number of Finance Transformation designs ultimately concerned
with the production of in-force and movement Model Point Files (MPFs) for consumption by
the Prophet application. For simplicity, the scope of this particular design is highlighted in
yellow in the diagram below:
P1L70 –Product
Families 1 to3
P1L70 – Product
Families 4 to 13
UNISURE
In-Force
and
AR
Movement
Model
Paymaster
(Refresh)
PAD Point Files
for Prophet
Alpha
Administrator
Other Sources
(manual policy
feeds)
Figure 1
This design therefore covers the extraction, transformation and loading of data items from the
UNISURE extracts into the target PAD database. The data transformed and stored is that
required to fulfil the stated requirements of the Finance Transformation Prophet Enhancement
team for the valuation of Product Families Unitized NP Pensions, Unitized WP Pensions and
Term Assurance.
The onward transformation of UNISURE data into Model Point Files (MPFs) for consumption by
Prophet will be covered within a separate macro design.
To meet this challenge, CUSTOMER is planning to transform its Finance function and the way it
serves its customers. The Finance Transformation programme has been established to address
this need through delivering the following outcomes.
Significantly simplify very complex data feeds and so reduce complexity in Prophet
modelling and reducing reconciliation effort; and
The PAD forms a key foundation of the Solution Architecture developed to support these
programmed outcomes, providing:
Data transformations which are documented, agreed and visible to the business via
Informatica’s Meta Data Manager tool
Below are the product families to which Unisure data goes as modal point files.
Figure 2
1.3 Approach
This macro design has been generated from the following primary inputs:
2 – The specification of required signed off business logic based on stakeholder engagement
with the Finance Data Team and consequent validation with the reporting teams
3 – The representation of that business logic in ETL mappings, applied using the following
design principles to the physical and logical models of the PAD and its data transformations:
4 – Specific design activities in accordance with the principles listed above to determine a) the
physical PAD model, and b) whether a particular piece of logic should be applied between
source extract and PAD or between PAD and MPF. The scope of this design excludes PAD to
MPF data transformations for now. It would be included in other few weeks.
1.4 Scope
1.4.1 In-Scope
Consume the UNISURE source extract and populate the PAD according to stated
requirements and agreed business rules. The stated requirements are explicitly taken
from Prophet IFAD for Unitized NP Pensions, Unitized WP Pensions and Term Assurance.
Application of errors and controls processing in line with the Errors and Controls macro
design
Movements capture for Non New business and Non Incremental business.
Any Model Point File data production (In force and Movement MPF, GMPF, XMPF).
The production of any Data marts or the consumption of any Prophet results
Any interaction with Prophet Automation (the processes which control the interface
between the PAD and Prophet)
Any storing/archiving of source extracts once these files have been read into the PAD
staging tables
Reporting of controls data, trend analysis and reconciliation. This design supports the
storage of that data and the subsequent macro design will incorporate the reporting
functionality.
Assumption Justification
to the stage tables
Approach 2:
A slim long temporary table will be created
to store all the surrogate keys for the data
corresponding to the current run/month
and later a table-minus-table is computed
between the PAD table and this temporary
table to infer the logical deletes from the
source.
This temporary table would be partitioned
based on the mapping identifier and source
to facilitate a partition truncate once the
logical deletes have been made.
PERL SCRIPT
Target Files
UNISURE Source Files (139 MPF’s)
Inbound Staging Intermediate Tables
(Fixed width ASCII) PAD Outbound Staging
on Informatica
to store UNISURE
UNIX server data sets
GOMD
Figure 3
The description of the above components and process flow is given below.
GENERIC OMD
Generic OMD is an CUSTOMER scheduling tool. As part of GOMD, a PERL script is the basic
driver to invoke and execute each ETL job throughout the process as explained in the OMD
Macro design document.
OMD Operational Meta Data table will contain data about batch identifiers, for instance: source
system name, received date of the source file, processed date and statuses of each processing
state (such as Pre-PAD Data Stage successful, Loaded to PAD, Post-PAD Data Stage Successful,
MPF created). The data into this table inserted/updated by PERL script only. There will not be
any direct interface to this table from Informatica mapping.
e.g.: PAD is associated with a Batch Identifier, enabling roll back to be handled by the Batch
Identifier when a session fails abruptly. It is expected that a particular ETL job should be re-
started on failure. More details can be found in the OMD Macro Design Document.
UNPROCESSED DATA:
Error handling and controls data structures and processes are represented in the Errors and
Controls macro design. The source specific implementation of that generic design is contained
in Section Error: Reference source not found of this document.
LOOKUP DATA:
Reference data is stored in multiple database tables. These tables will be looked up and
required reference values will be retrieved during the processes ETL-3 and ETL-4 which loads
data into PAD and outbound staging tables respectively.
ETL-1:
It consists of Informatica mapping/s which reads the data from reference files/tables and loads
into respective lookup database tables.
SOURCE:
CUSTOMER Valuation system on monthly basis will generate UNISURE extract files. The ETL
source will be the UNISURE source extract files which are generated by this valuation system.
The same files will be copied onto Informatica UNIX server at specified location from where the
Informatica service reads and processes the source data.
ICMS Commission system would be generating the extract on monthly basis. The same would
be copied onto Informatica server on the Unix environment at specified location and then
picked up by the Informatica service to processes it.
Asset Share Factors data would be used for referencing the Fund, Asset Share and Unit Price
calculations
ETL-2:
ETL-2 consists of Informatica mappings which read the data from source UNISURE files and
loads into inbound staging table. During the process if any source record violates the business
or database constraint rule, that record will be populated into unprocessed data table using the
process described in the Error Handling Macro Design.
INBOUND STAGING:
Inbound staging table has similar structure to the source data, with the addition of meta data
fields (e.g. Extract Date, Source File Name, Batch Identifier).
Where required, Inbound Staging will contain the data from the current and prior extract (i.e.
two consequent months of data), allowing movements to be detected via the delta between
two extracts, where movement events or transactions are not provided on the source extract.
The stage table will be partitioned in order to enable historical data to be easily deleted (via
truncating the partition) when no longer required.
ETL-3:
ETL -3 consists of Informatica mappings which reads the data from inbound staging table and
loads into Intermediate tables specific to UNISURE. These intermediate tables will contain the
Data sets derived from the UNISURE inbound staging data.
ETL-4:
ETL -4 consists of Informatica mappings which reads the data from inbound staging table and
loads into PAD table. During the process if any record violates the business or database
constraint rule that record will be populated into unprocessed data table according to the Error
Handling Macro Design.
PAD:
PAD (Persistent Actuarial Database) consists of multiple database tables with predefined
relationships.
The following stages are explicitly out of scope for this design:
ETL-5:
ETL-5 consists of Informatica mappings which read the data from the PAD and loads into
outbound staging table. More details are provided as a part of PAD-MPF Macro.
OUTBOUND STAGING:
Data will be staged in the outbound staging before creating the MPF files. More details are
provided as a part of PAD-MPF Macro.
ETL-6:
ETL-6 consists of mappings that will create the MPF files. More details are provided as a part of
PAD-MPF Macro.
TARGET:
The target is Model Point ASCII CSV files with column and business header information. More
details are provided as a part of PAD-MPF Macro.
Term Description
Term Description
Figure Description
Figure 1 Holistic view of UNISURE in Finance Transformation Solution
Figure 2 ETL Architecture flow
Figure 3 Data flow from Source to PAD
4 Archival Strategy
PAD.
6 Non-Functional Requirements
All Non-functional requirements are captured in Requisite Pro. They include response times,
data storage, tracking of data, data lineage, access levels and security etc.
In addition to the documents referenced above, the following ‘Issue Resolution’ requirement
was captured during the design review:
"In the event that reference data used to load a given source system data feed into the PAD is
found to be corrupt, incorrect or incomplete then it must be possible to re-supply or amend the
reference data and re-run the staging to PAD mapping to resolve the errors. In practice, this
would mean:
Policies that had previously been written to the Unprocessed Data table being successfully
loaded to the PAD
Policies that had previously been loaded to the PAD being written to the Unprocessed Data
table
Policies that had previously been written to the PAD being again written to the PAD but with
changed data values
The re-running of Staging to PAD processes will first require the manual deletion of inserted
records on the basis of batch identifier, and the re-setting of record end dates and current flags
for those records which have been updated. This is expected to be a manual BAU process,
though the necessary SQL statements are expected to be provided to aid testing.
Any PAD batch activities which are dependent on the successful load of this source system feed
(e.g. MPF production) should not commence until users confirm that the data loaded to the
PAD for the given source system is of sufficient quality (i.e. that the Unprocessed Data has been
examined and any reprocessing of the kind described above has taken place). This requirement
should be handled by the creation of GOMD meta data and the configuration of GOMD for
Finance Transformation."
Figure 4
8 Source Data
8.1 Source Data Specification
The UNISURE source data is transmitted in fixed width ASCII files. Data is split across 3 extracts,
each containing multiple files.
Interface
Description
Name
The extract will transmit the below 3 files on a monthly basis
NUCBL.DESIGNER.DATA.TCBDLEL.UNLOAD
ICMS NUCBL.DESIGNER.DATA.TCBCNDL.UNLOAD
commission NUCBL.DESIGNER.DATA.TCBDEAL.UNLOAD
shapes(3 files) Note:- ICMS is descoped because Unisure is dealing with Existing
Business whereas ICMS is dealing with New Business
For future enhancement, We are just loading data up to Stage
Unisure to The extract will transmit the below 20 files on a monthly basis
valuation NUWJL.FTP.UN.MHLY.FILE01.CRExxxxx.DEL500
inforce NUWJL.FTP.UN.MHLY.FILE02.CRExxxxx.DEL500
interface(20 NUWJL.FTP.UN.MHLY.FILE03.CRExxxxx.DEL500
files) NUWJL.FTP.UN.MHLY.FILE04.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE05.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE06.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE07.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE08.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE09.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE10.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE11.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE12.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE13.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE14.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE15.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE16.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE17.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE18.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE19.CRExxxxx.DEL500
NUWJL.FTP.UN.MHLY.FILE20.CRExxxxx.DEL500
NUWJL.FTP.UN.WKLY.CRExxxxx.DEL500
The extract will transmit the below 3 files on a monthly basis
Asset share NUVRP.PRODUS.P.PRICES.CREYYMM.INP.G0001V00
factors and NUVRP.PRODUS.P.SPRICE.CREYYMM.INP.G0001V00
W/P fund NUVRP.PRODUS.P.FTGIPP.CREYYMM.INP.G0001V00
ID's(5 files) NUVRP.PRODUS.P.FTGNG.CREYYMM.INP.G0001V00
NUVRP.PRODUS.P.FTGNGG.CREYYMM.INP.G0001V00
The Inforce and Movement files come in two sets, one set without headers and footers until
August – 2009 and would be termed as Pre-DFR files the files that arrive after August -2009 will
have headers and footers and would be termed as POST DFR files and the files would be
processed by two different ETL processes.
The metadata stored in the staging tables facilitates the identification of the source file and
batch identifier from which the policy got loaded.
Using Transformation logic (based on assumptions made out of available UNISURE SAS code,
and validated by CUSTOMER) the 3 UNISURE source extract files provided in section 13.1 of this
document will be merged and 10 data sets will be created. These data sets will be stored in
intermediate tables for further processing to PAD.
WH_STG_UNISURE
WH_STG_UNISURE_ICMS_DLEL
WH_STG_UNISURE_ICMS_DEAL
WH_STG_UNISURE_ICMS_CNDL
WH_STG_UNISURE_MOV
WH_STG_UNISURE_ASF_GNGG
WH_STG_UNISURE_ASF_GIPP
WH_STG_UNISURE_ASF_GNG
In the process of loading the data from inbound staging to intermediate tables the data is split
and grouped which is discussed in detail in the Data Mapping sheet in section 15.
Code
Variable Source Name
polref Policy number *
inrref Increment reference *
covref Coverage reference *
prdref Product ref
prdref Product ref
prdvr_no Product version no.
schemeno Scheme number
prefix Prefix
territ Territory
p_lcon Life contingency
The Activity table (WH_PAD_ACTVTY) will hold all the data related to movements for a
particular policy. The data mapping sheet for the target WH_PAD_ACTVTY table contains details
of how to populate this table.
9 Database Overview
A fully comprehensive overview of the database is covered in the Physical Design Model (PDM)
document. The following are the list of the tables that are applicable for the source to PAD
macro design of UNISURE.
This section will have a mapping sheet which will provide the following mappings
In the mappings, Extract Date is the last day of the month for which extract is provided. For
example for August 2017’s extract date is 31/08/2017, for September 2017 it is 30/09/2017.
Plan of Work:
No
Planned
of Specific Deliverable in terms
Tasks to be done Duration Status
task of project
(Weeks)
s
Requirement
1 2.5 IFAD’s, Solution design Completed
Gathering
2 Analysis & Design 2.5 Design Doc & Mapping sheets Completed