Home
[1 ] Oracle® Fusion Middleware
Contents
1. The Server Console Purge Results rules only apply to jobs run with a run label that is those run in Server Console with an assigned run label or from the Command Line using the runopsjob command tis possible to configure rules in Server Console to purge data after a set period of time See the Result Purge Rules topic in Enterprise Data Quality Online Help for further details These purge rules only apply to Server Console results 2 8 1 Purging Match Decision Data Match Decision data is not purged along with the rest of the results data Match Decisions are preserved as part of the audit trail which documents the way in which the output of matching was handled If it is necessary to delete Match Decision data for example during the development process the following method should be used 1 Open the relevant Match processor 2 Click Delete Manual Decisions 3 Click OK to permanently delete all the Match Decision data or Cancel to return to the main screen Note The Match Decisions purge takes place immediately However it will not be visible in the Match Results until the Match process is re run This final stage of the process updates the relationships to reflect the fact that there are no longer any decisions stored against them For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 28 Using Oracle Enterprise Data Quality Creating Processo
2. Configurable matching library provides flexible matching functionality for the majority of needs Extensible matching library ensures the optimal matching algorithms can be deployed x Allows import and export of match decisions Provides a complete audit trail of review activity decisions and review comments 3 2 Clustering Concept Guide Clustering is a necessary aspect of matching required to produce fast match results by creating intelligent first cuts through data sets in order that the matching processors do not attempt to compare every single record in a table with every single other record a process that would not be feasible in terms of system performance Clustering is also vital to Real time matching to allow new records to be matched against existing records in a system without the need for EDQ to hold a synchronized copy of all the records in that system Clusters Rather than attempt to compare all records with each other matching in EDQ takes place within Clusters which are created by a clustering process Clustering does not attempt to compare any records in a data set Rather it creates clusters for a data set by manipulating one or more of the identifier values used for matching on a record by record basis Records with a common manipulated value cluster key fall within the same cluster and will be compared with each other in matching Records that are in different clusters will not be compar
3. To set up an Export of a Staged Data table 1 Right click on Exports in the Project Browser and select New Expott o M E N S New Results Book Export OLH E H a Testi ra Copy Referenc fa i nana Hy Data Stor 2 Published 5 Package k a eee A Tm 1 14 Using Oracle Enterprise Data Quality Exporting Data Prepared exports 2 Select the required Staged Data table 3 Select the required Data Store 4 Ifthe table in the Data Store is not being created during export select the required table in the Data Store Note The contents of this table will be overwritten by default See Export Options below for further details 5 If required map the columns in the Staged Data table to columns in the target table How should the staged data columns be mapped to the export table columns Personumber Tithe Forenames Sumame Haonorific Gender DateO firth Address 1 Ack eps Posi gvm PersonPostroode County Map By Mame 20 of 20 attributes have maooings Note Click on the Auto button to automatically map the columns in the order they are presented 6 When the mapping is complete click Next and if required change the default Name of the Export the default name comprises the Staged Data table and the name of the Data Store 7 Either run the Export straight away or save the configuration for running later for
4. a Setting Inputs Setting Options Setting Output Attributes Setting Results Views Setting Output Filters Setting Dashboard Publication Options a Setting a Custom Icon 2 9 2 Setting Inputs The inputs required by the processor are calculated automatically from the configuration of the base processors Note that where many of the base processors use the same configured input attribute s only one input attribute will be created for the new processor However if required you can change or rename the inputs required by the processor in the processor design view or make an input optional To do this click on the Processor Setup icon at the top of the Canvas then select the Inputs tab In the case above two input attributes are created Title and Forenames as these were the names of the distinct attributes used in the configuration of the base processors The user chooses to change the External Label of one of these attributes from Forenames to Forename to make the label more generic and chooses to make the Forename input optional Processor Setup Published options Inputs Type External Label Required STRING Title Forenames STRING Forename F lat ia ee eee a T R ll Note that if an input attribute is optional and the user of the processor does not map an attribute to it the attribute value will be treated as Null in the logic of the processor 2 30 Using Oracle Enterprise
5. The options of a transformation allow you to vary the way the identifier value is transformed The available options are different for each transformation For example the following options may be configured for the First N Characters transformation a Number of characters the number of characters to select Characters to ignore the number of characters to skip over before selection Using clustering The best clustering configuration will depend upon the data used in matching and the requirements of the matching process Where many identifiers are used for a given entity it may be optimal to use clusters on only one or two of the identifiers for example to cluster people into groups by Surname and approximate Date of Birth for example Year of Birth but without creating clusters on First Name or Post Code though all these attributes are used in the matching process Again this depends on the source data and in particular on the quality and completeness of the data in each of the attributes For accurate matching results the attributes used by cluster functions require a high degree of quality In particular the data needs to be complete and correct Audit and transformation processors may be used prior to matching in order to ensure that attributes that are desirable for clustering are populated with high quality data Note that it is common to start with quite a simple clustering configuration for example when match
6. Column selection To change the set of columns that are displayed on the Event Log click the Select Columns button on the top left of the Event Log area The Select Columns dialog is displayed Select or deselect the columns as required and click OK or save or Cancel to abandon the changes Alternatively click Default to restore the default settings Select columns Sub Type Event Time Duration Severity E Job Name Task Type Task Name F User Name Note that Severity is a rarely used column it is currently set to 50 for tasks or jobs that completed correctly and 100 for tasks or jobs that raised an error or a warning Opening an event Double clicking to open an event will reveal further detail where it is available Opening a Task will display the Task Log showing any messages that were generated as the task ran INFO 13 Oct 2011 11 49 20 Validation complete 0 ms WFO 13 Oct 2011 11 49 21 process 83 job 73 114 starting execution of subprocess 1 in batch 1 O 13 Oct 2011 11 49 21 process 83 job 73 114 starting execution of subprocess 1 in batch 0 O 13 Oct 2011 11 49 21 process 83 job 73 114 starting execution of subprocess 1 in batch 2 O 13 Oct 2011 11 49 21 process 83 job 73 114 starting execution of subprocess 1 in batch 3 O 13 Oct 2011 11 49 22 process 83 job 73 114 subprocess 1 batch 3 complete INFO 13 Oct 2011 11 49 22 process 83 job 73 114 subprocess
7. P ntan dig a a salen NO ite nah The Show Characters option is also useful when working with very long fields such as descriptions that may be difficult to view fully in the Results Browser For example the below screenshot shows some example data captured from car advertisements Lar Ads Staged data BE BHR OCHS EG Title RESUITS BrOWSeT uade T 5 Doors Manual Hatchback Petrol 21 631 miles Aruba Blue 1 Ow 29 900 miles Excellent condition Full service history 5 dr hatchback 5 Doors Manual Estate Petrol 12 385 miles Metallic Grey 3x3 poi 3 Doors Manual Hatchback Petrol 24 730 miles Indigo Blue Metalli Beautiful example full ground up restoration in the last 10 years Ext 3 Doors Manual Hatchback Petrol 34 000 miles Metallic Blue 20 Manual Hatchback Petrol 35 232 miles Silver 1 Owner s Air con 3 Doors Manual Hatchback Petrol 37 605 miles Metallic Pearl Blac 5 Doors Manual Estate Diesel 33 000 miles GREY ABS Adjustabl l suto 5w Die 5 Door 4 4 Blue Diesel MSA Alloy wheels Audio 2005 54 Reg NISSAN Almera 1 8 SE 2003 HONDA CIVIC 1 6 VTEC Inspire 5dr Hatchback Spec Eds 2004 54 Reg RENAULT SCENIC 1 4 AUTHENTIQUE SDR AC Volkswagen Polo MK4 Hatchback 3 Or 1198 cc 1 25 A C 65 BHP 1978 MG MGB Roadster 2004 54 Reg SEAT Ibiza 1 2 5 2004 54 Reg Citroen C3 1360 cc 1 44 Desire 5dr 2005 55 Reg RENAULT CLIO 1 2 Rush 3dr 2004 54 Reg CITROEN Xsara Picas
8. Click OK to save changes However if a Trigger is deleted in error click Cancel instead For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 4 Job Notifications A Job may be configured to send a notification email to a user a number of specific users or a whole group of users each time the Job completes execution This allows EDQ users to monitor the status of scheduled jobs without having to log on to EDQ Emails will also only be sent if valid SMTP server details have been specified in the mail properties file in the notification smtp subfolder of the oedq_local_home directory The same SMTP server details are also used for Issue notifications For more information see Administering Enterprise Data Quality Server The default notification template default txt is found in the EDQ config notification jobs directory To configure additional templates copy this file and paste it into the same directory renaming it and modifying the content as required The name of the new file will appear in the Notification template field of the Email Notification Configuration dialog How To 2 13 Job Notifications 2 4 1 Configuring a Job Notification 1 Open the Job and click the Configure Notification button on the Job Canvas toolbar The Email Notification Configuration dialog is displayed Enabled ry Notification template Default template Groups Administrat
9. Ifa real time process includes a processor that is unsuitable for real time response processing it will raise an exception when the first record or message is received The supported execution types of each processor are listed in the help page for the processor Note that real time response processes may use much or all of the same logic as a batch process 2 23 3 Real time monitoring Real time monitoring processes are designed to check data quality at the point of data entry but not return a response to the calling application so that there is no extra burden on the user modifying the data on the source system As there is no need to return a response there are fewer restrictions on what the EDQ process can do for example it may include profiling processors that work on all the records for the period of time that the monitoring process runs Real time monitoring processes have the following characteristics They read data from a real time provider such as the inbound interface of a Web Service How To 2 75 Customizing Processor Icons a They may include any of the processors in the processor library including those such as Duplicate Check that are not suitable for real time response processes Note Ifa real time monitoring process contains processors that are designed to process whole batches and are therefore not suitable for real time response processes it should be run in Normal mode and not
10. Investigating a Process W Input Attribute Finder X Input Attributes Candidates Jump To Creator MAME Proper eee Al ADDRESS ADDRESS1 Proper e ADDRESS2 40DRES52 Proper ADDRES53 40DRES53 Proper Cancel There are 2 options available Jump to Creator this takes you to the processor which created the selected attribute Highlight Contributors this highlights all the processors including the Reader if relevant which had an influence on the value of the selected input attribute The contributory processors are highlighted in yellow In the example below the Concatenate processor had one of its input attributes searched in order to determine which path contributed to the creation of that attribute Reader E Unknown Parse Concate nate b gt _ gt Dia Lower Case Ae on aa ol ek M ae Ln Tem ee a p s T m m T o m e m e amh Ta Candidates The candidates tab in the Input Attributes Search enables the same functionality to take place on any of the attributes which exist in the processor configuration The attributes do not necessarily have to be used as inputs in the processor configuration 2 12 3 Clear Search Highlight This action will clear all the highlights made by the Highlight Contributors option For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help How To 2 41 Previewing P
11. Live you can display all events for live systems as follows Filter gt Events T l a gt Search x Job a I How To 2 47 Using the Event Log Note Date time filters The final set of filters on the right hand side of the screen allow you to filter the list of events by date and time A Date picker is provided to make it easier to specify a given date Note that although only the most recent events are shown when accessing the Event Log it is possible to apply filters to view older events if required Events may be filtered by their start times and or by their end times For example to see all Jobs and Tasks but not System Tasks that completed in the month of Note November 2008 apply filters as follows Sub Type Status Status System Task 4 Cancel Request End Start Search af DiOdt 20ii 14 06 FA To 31 Oct 2011 1406 H Fy uration a je v Vai 2 48 Using Oracle Enterprise Data Quality The Project Name column is not displayed by default To change the view to see it click the Select Columns button on the left hand side and check the Project Name box Events are never deleted from the history by EDQ though they are stored in the repository and may be subject to any custom database level archival or deletion policies that have been configured on the repository database Using the Event Log
12. Lo Cluster First 4 surname meta first 3 postcode Cluster Name First 4 surname meta first 3 postcode Home tel last 6 Identifiers amp Transformations EE Surname 7 Upper Case 79 Trim Whitespace 2 Metaphone Lu ff First N Characters i Postcode 72 Upper Case 729 Trim Whitespace 1 First N Characters Add Identifier Override Defaults For each record in the matching process therefore the value for the Surname identifier will be transformed converted to upper case all whitespace removed a metaphone value generated and trimmed to the first 4 characters and then concatenated with the transformed value from the Postcode identifier converted to upper case all whitespace removed and trimmed to the first 3 characters Note that the concatenation of the identifier values after transformation is not configured and occurs automatically So using the above cluster configuration cluster keys will be generated as follows Surname Postcode Cluster Key Matthews CB13 7AG MOSCB1 Foster CB4 1YG PSTRCB4 JONES SW11 3QB JNSSW1 Jones swll 3qb JNSSW1 This means that the last two records would be in the same cluster and would be compared with each other for a possible match but would not be compared with the other two records Advanced Features 3 7 Clustering Concept Guide Transformations in clustering Transformations in clustering allow you to normalize space case spelling and other differen
13. Views a t Processes on fi Parse Customers 2 Lan tl Profile Customers table re Ba mF Peara t a ans E E p 2 Double click on the Reader to configure it to read data from Staged Data such as a Snapshot a View or a real time data provider 3 Select the data that you want to read and the attributes from the data that are relevant to your process 4 Add a processor from the Tool Palette to the process by clicking on a processor and dragging it to the Canvas 5 Connect the processor to the data from the Reader by joining the arrows 1 6 Using Oracle Enterprise Data Quality Adding Processors Quickstats Profler a Pattems Proiler Reader Data Types Profler ae os en a i o a ee ee 6 Configure the processor by selecting its input attributes Attributes Options Dashboard Notes Input Atbribubes Selected Athribubes Qubput Atbribubes E Data B CU_4ACCOUNT CU_ACCOUNT string A TITLE A TILE B NAME 9 NAME is B GENDER al GENDER string BUSINESS a A BUSINESS string P ADDRESS A A ADDRESS string ADDRESS 4 A ADDRESS string P ADORESSS A A ADDRESS string P POSTCODE A A POSTCODE str A EMAL 41 EMAIL str A ACELOR al A ACC_MGR strina B OT _PURCHASED a DT_PURCHASED string a A DT_ACC_OPEN string Dap Added Flags E S Incompatible A DT_ACC_OPEN All Processors Search a x The blue arrow icons indicate that t
14. a a4 a M i cL a a djdd Standardizedvame al Namelype ity State Country a City Country Code Country Redgistration al 4 a Country Operating A Address Line 1 Street A Postcode a When Case Management is in use all the active data fields must be mapped No highlighting is used in this case The following screenshot shows an example of configuring the attributes in the Decision data to the Decision Key of a simple Deduplicate match processor W Decisions General Settings Mapped Fields Related Record Customer Customer Input Fields Decision Fields Input Fields Decision Fields A First Name s A Fname First Name s RelatedFname F hd ame fare A Address Line 1 Street A Address Line 1 Street A RelatedAddress1 U Postcode A ReBtedPostode v W Daorsi anos Note that the decisions data needs attributes for both records in the relationship in this case record and related record 2 17 4 Step 4 Importing the decisions Once the Decisions sub processor has been correctly configured you can click the OK button to save the configuration The match decisions can then be imported by running the matching process How To 2 55 Exporting Match Decisions If you are performing a one off import of decisions check that the decisions have been imported correctly and then delete the Reader for the decisions data If you want to continue importing decisions as part o
15. and optional Description for example Created from Customers table and choose whether or not to edit the data now 8 If you choose to edit the data now add or delete any entries in the Reference Data using the Reference Data Editor 9 Click OK to finish The Reference Data set now appears under your project in the Project Browser and is ready for use in processors for example in a List Check 1 4 Adding an Issue To add an Issue based on your results for example to tag an item of interest or to create an action for another user to follow up on 1 Right click on the data in the Results Browser and select Create Issue 1 4 Using Oracle Enterprise Data Quality Adding a Project Note Gadd 7 BE BBE O66 FEE Input Field __ Record Total With Data Without Data Singleton Duplicates CUNO 2010 2005 z 2002 r CU_ACCOUNT 2010 2003 7 004 i TITLE 2010 1872 13 Ctrl C NAME 2010 2009 1 GENDER 5010 1851 j4 Create Reference Data BUSINESS 2010 1673 33 Add to Reference Data ADDRESS1 2010 1998 12 ADDRESS 2010 1920 ey Create ISSUE A CUO Reena nT Mai ee O31 patna sant Clabes AF A ete eet le a oe Note The issue will be linked to the specific process and processor where it was created so that another user can quickly find the related data 2 Adda Description of the issue 3 Optionally assign the issue to yourself or another user and specify the Action needed if
16. configured inputs Published processors appear both in the Tool Palette for use in processes and in the Project Browser so that they can be packaged for import onto other EDQ instances How To 2 35 Publishing Processors Note The icon of the processor may be customized before publication This also allows you to publish processors into new families in the Tool Palette To publish a configured processor use the following procedure 1 Right click on the processor and select Publish Processor The following dialog is displayed o Publish Processor Hame Summarize Name Verification Description Type Scope Template Processor Project Level Reference Processor System level Locked Reference Processor Other Options Package Reference Data with processor y 2 Inthe Name field enter a name for the processor as it will appear on the Tool Palette 3 If necessary enter further details in the Description field 4 Select the Published processor Type Template Reference or Locked Reference 5 Select the Scope Project the processor is available in the current project only or System the processor is available for use in all projects on the system 6 If you want to package the associated Reference Data with this published processor select the Package Reference Data with processor checkbox Note Options that externalized on the published processor always require Refere
17. example as part of the scheduled execution of a job or process 1 10 1 1 Export Options The Export Task defines the source and shape of the data that will be exported and the target for the export such as an output text file an existing table in a database to which the data is mapped or a new table in a database The user chooses how the Export behaves when adding the Export Task to a job Getting Started 1 15 Exporting Data Prepared exports Outpu G Task Configuration Task Type Export Detail Output Customers to Outputlustomers csv Enabled Export Option The options are a Delete current data and insert default EDQ deletes all the current data in the target table or file and inserts the in scope data in the export For example if it is writing to an external database it will truncate the table and insert the data or if it is writing to a file it will recreate the file Append to current data EDQ does not delete any data from the target table or file but adds the in scope data in the export When appending to a UTF 16 file use the UTF 16LE or UTF 16 BE character set to prevent a byte order marker from being written at the start of the new data Replace records using primary key EDQ deletes any records in the target table that also exist in the in scope data for the export determined by matching primary keys and then inserts the in scope data Notes When an Export is run a
18. 1 2 2 2 1 3 2 1 4 aed 2 8 2 8 1 2 9 2 9 1 2 9 2 2 9 3 2 9 4 2 9 5 2 9 6 2 9 2 9 8 2 10 2 10 1 2 10 2 2 10 3 2 11 2 11 1 2 11 2 2 12 2 12 1 2 12 2 2 12 3 2 13 2 14 2 14 1 2 14 2 2 14 3 Editing A OID socactass cuazentaseseaoscducs O AONAN 2 10 B Kia tag Wl 6 0 peepee sree eerie erento E ra et ns ae arse Perera vee eno eiee eer ron eee rane ere me rer 2 10 Job Gany as Right Click Met ot istcaseshsncaestes set satet atastathencruaduin adesanaa acetates aeeats 2 10 Editing and Configuring JOD PRASES vijcpin oc noenee a 2 11 Usim JOR THEOCR casera ati snes aa naan ease aan 2 11 Conine regeis eari E pe tem e ua oasia en toca a ela ebeea aetna 2 12 Deletine a Trigger trom a Jobers n E ale unee A anidias dasa 2 13 FO NON Call ONS vce 0x saeco sas neccsccenecsaplorte se oaescesoannays A 2 13 Onn or 18a OD Notifica orare aieiaa E 2 14 Default Notiication COnent xsincisnansasansinen A eas 2 14 Optimizing JOR Perrin ia nce saia a E E A A 2 16 General Perlormance OPON Se o E AE A 2 16 Processor specific Performance Options seesseserserseserserstssrstesesreseserstesrserseeseneeseseeses 2 20 Pit lishing to the Dashboard asrcccsaniennie a a E tcice date eatay lat 2 23 PACA in Tanida a A A E en 2 24 Packacino ODI CCUG ai E E E ta aiaea osndncs 2 24 Filtering ancl Packaging imone a N E T A 2 25 Opening a package file and importing its contents sseesssesesersesersersessesreerseesrsersees 2 26 Work
19. 1 batch 1 complete IFO 13 Oct 2011 11 49 22 process 83 job 73 114 subprocess 1 batch 2 complete INFO 13 Oct 2011 11 49 23 process 83 job 73 114 subprocess 1 batch 0 complete IFO 13 Oct 2011 11 49 25 process 33 job 73 114 finished execution in 5 07s How To 2 49 Reviewing Matching Results Note Messages are classified as INFO WARNING or SEVERE An INFO message is for information purposes and does not indicate a problem A WARNING message is generated to indicate that there could be an issue with the process configuration or data but this will not cause the task to error SEVERE messages are generated for errors in the task For Jobs if a notification email was configured on the job the notification email will be displayed in a web browser when opening the completed event for the Job Jobs with no notifications set up hold no further information Exporting data from the Event Log It is possible to export the viewable data in the Event Log to a CSV file This may be useful if you are in contact with Oracle Support and they require details of what has run on the server To export the current view of events click Export to CSV This will launch a browser on the client for where to write the CSV file Give the file a name and click Export to write the file For more information see Enterprise Data Quality Online Help 2 16 Reviewing Matching Results One of the main principles behind
20. 19349 48586 richardt 31 Jul 2009 10 32 44 Pending 48580 9657 richardt 31 Jul 2009 10 33 00 Probable match pending further information 9657 48580 richardt 31 Jul 2009 10 33 07 Match 48565 36823 richardt 31 Jul 2009 10 33 12 Verified by email 48565 36823 richardt 31 Jul 2009 10 33 18 Match 48514 26045 richardt 31 Jul 2009 10 33 25 Verified by phone call 48514 26045 POE s lagi eg yee AeA An gigponnin nett tagptie Namnaina a E e A A gt ien goa natnn a 2 56 Using Oracle Enterprise Data Quality Externalizing Configuration Settings Note that all the relationship data at the time of the decision or comment is copied to the decision or comment record in order to keep a complete audit trail of the data to which the decision or comment applies As above the data stored for manual match decisions and comments is a little different but these are output together as a review comment may relate to a match decision If required the comment and decision records may be split in downstream processing for example by performing a No Data Check on the RuleDecision attribute All records with No Data in this attribute are review comments and all records with data in this attribute are manual match decisions Written decisions may be exported to a persistent external data store processed in another process or imported to another match process working on the same data if required see Importing Match Decisions For more information see Ente
21. EDQ installation 2 19 2 Externalizing Match Processors The Externalize option is displayed on the Sub Processor window of each Match processor Open this window by double clicking on the Match processor on the Canvas Externalizing Match processors allows the following settings to be changed dynamically at runtime a Which Clusters are enabled disabled The cluster limit for each cluster The cluster comparison limit for each cluster a Which match rules are enabled disabled a The order in which match rules are executed The priority score associated with each match rule The decision associated with each rule There are two stages to externalizing Match processor options Select Match processor options to externalize Configure the externalized Match processor options at the process level Note The second stage is optional but recommended to ensure the externalized option has a meaningful name 2 19 2 1 Select Match processor options to externalize 1 Click Externalize on the Sub Processor window The Externalization dialog is displayed 2 Select the properties to externalize with the check boxes next to the Match Properties listed 3 Click OK to accept or Close to abandon 2 19 2 2 Configure externalized Match processor options at the process level Externalized options have a default generic name Use the following procedure to assign a different name 1 Right click on the relevan
22. TITLE 2010 1872 0 0 136 935 1 NAME 2010 2009 0 0 1 9 9 j GENDER 2010 1561 0 0 149 92 6 BUSINESS 2010 16 3 0 0 337 83 2 ADDRESS1 2010 1997 1 0 12 39 4 ADDRESS2 2010 1920 0 0 90 05 5 ADDRESS3 2010 1031 0 0 979 aL 3 POSTCODE 2010 1767 0 0 243 87 9 EMAIL 2010 1932 0 0 7a 96 1 ACC_MGR 2010 1992 0 0 16 9 1 DT_ PURCHASED 2010 0 0 1994 16 99 2 DT_ACC OPEN 2010 0 0 1944 16 99 2 A pment on antl Ale aah A ne th Al nn am ah en tein vat Fn tt em nt 1 8 Using Oracle Enterprise Data Quality Adding Processors 10 Drill down on the metrics to see the relevant data for the metric Having successfully created a process you can now begin to create Reference Data for the validation and transformation of your data 1 6 1 Ambiguous Latest Attribute When selecting the input attributes for a processor it is possible that some attributes may have an ambiguous latest version This happens whenever an attribute exists on two different paths and has been transformed on either or both of these paths Where this happens the top level attribute or latest version will be greyed out and will not be selectable In this case you need to select the specific version of the attribute that you want to use by expanding on the attribute and viewing all its possible versions For example the Upper Case processor below has been configured with 2 input paths from 2 different processors The Replace processor transforms a title attribute
23. Width Text Bile POriats nersini tees ene teatariavadenene eerie 1 10 1 7 1 Files containing no new line characters cc cece ese ceseseeeeseeeeesseeseesesseesseeneaeens 1 11 1 8 Connecting toa Data Store vicars oceans een a 1 12 1 9 Creatine AT rO CE anri E E a N EEE 1 14 110 Expor ne Data Prepared XpPOrts enecinrd tenie aE E TANER 1 14 1 10 1 Staged Data EXPO Enine isase E O R 1 14 1 10 2 Resulte Dook T XPO R eroan EO AA 1 17 it R nga Prepared EXPO en n a aN scatters Oa E nS 1 17 1 11 1 Manually runnin aI CX PO Eas nansa aa a a a mateo 1 17 1 11 2 R nning ane portas Part OF ajo D annsna N 1 18 2 How To 2 1 Execution OPHONS aana E O 2 2 2 1 1 Snaps HOla A E eae mieteeuns 2 2 Chee ETOCS aisr cess occas Meee at ntact diate let letra ae reet abla Date h lead anal ee sles edeaee hash taesaats 2 2 2 1 3 External TAS KS senses aec Riri sag pictaishager andes sngsnsea nb sagtaca den tesanagassaasataame aseunepedesoniemeadedvasyaineveniase 2 7 2 1 4 DONS eae ess ie Pe E E ea ug abet atias ea Roba vacate iets ase OTTE OTS 2 2 1 5 Results BOOK EX PON Serrar staat aud Ws ueniesass os lst easel ale dasa E A 2 8 2 1 6 UST OL Shai E atten aie EE E E E Meaty en ames ee eae eRe 2 8 2 2 Creating and Managing ODS ssisccvlevestedsansent neil aawan saci awe E O 2 8 2 2 1 Creating a JOD eppen n niger gancsnmegnaaee magne arabanuaanoed nee maatedonns 2 8 2 2 2 2 2 3 2 2 4 2 2 9 2 3 2 3 1 2 3 2 2 4 2 4 1 2 4 2 2 5 2 5 1 29 2 2 6 2 1 2
24. _dause_sql Snapshot SOL E snapshot_sql Data Store Access file E access_file Use project specific landing area fic_landing_area 2 19 5 Externalizing External Tasks To externalize an External Task right click on the task in the Project Browser and select Externalize Select and edit the options as required Click OK to save or Cancel to abandon The fields on the right of all Externalization dialogs contain the default name used for each externalization attribute This is the label used to reference the externalized attribute in any third party tool or file To change any field name check the box to the left and edit the field as required Note Avoid the use of spaces in attribute names 2 19 5 1 External Task Externalization Dialog For External Tasks that run commands the following dialog is displayed How To 2 63 Externalizing Configuration Settings Externalization of External Task External Task Options Command E command Working Directory 7 working_directory Arguments E arguments 2 19 5 2 File Download Externalization Dialog If an External Task is a File Download the following dialog is displayed Extemalization of File Download Download Task Options URL Username Password Proxy Host Proxy Port Proxy Username Proxy Password Filename Oo O0 0 0 0 0 u username password proxy host proxy port woxy_Username roxy password filename 0l U
25. a Snapshot 3 Select the Staged Data Data Interface Reference Data or Real time Data Provider that you want to use in the process or do not select anything if you want to configure the Reader in the process later Note It may be that you do not want to stage the data you are analyzing that is you may want to stream the data directly from the source This can be done by selecting the Staged Data configuration and changing the Process Execution Preferences of the process 4 Select whether or not to add Profiling processors to the process straight away This may be useful if you are analyzing the data for the first time 5 Give the process a Name and an optional Description 6 Click Finish 1 2 Adding a Snapshot To add a Snapshot of data from a connected data store 1 Right click on Staged Data in the Project Browser and select New Snapshot HED Dynamic Configuration Demo 5 H Data Stores ie PI OAV Test Data E Orade Demo 9 Service Management D Dat Oe New Snap Ot a TRPIO aibo aw Stage 2 Select the data store that you want to create the snapshot from or add a new data store if the desired data store is not on the list 3 Select the table or view to snapshot or you may specify SQL to snapshot a new view of the data 4 Select the columns from the table or view that you want to include in the snapshot and how to enable sorting and filtering on the snapshot By default
26. also possible to stream data to an export target by configuring a Writer in a process to write to a Data Interface and configuring an Export Task that maps from the Data Interface to an Export target 1 Prepare coshoner data Eg OH e rite New Phase IndividualbataPr epared oe EntityDataPrepar II IIT CUST j CUST Entities I ad Build clusters ndexed Export of pare n Ral na inl lle B a anhaa apm OV aii e Note that Exports to Client side data stores are not available as tasks to run as part of a job They must be run manually from the EDQ Director Client as they use the client to connect to the data store 2 5 1 2 Minimized results writing Minimizing results writing is a different type of streaming concerned with the amount of Results Drilldown data that EDQ writes to the repository from processes Each process in EDQ runs in one of three Results Drilldown modes All all records in the process are written in the drilldowns Sample a sample of records are written at each level of drilldown None metrics only are written no drilldowns will be available All mode should be used only on small volumes of data to ensure that all records can be fully tracked in the process at every processing point This mode is useful when 2 18 Using Oracle Enterprise Data Quality Optimizing Job Performance processing small data sets or when
27. and Enterprise Data Quality Online Help 2 Packaging Most objects that are set up in Director can be packaged up into a configuration file which can be imported into another EDQ server using the Director client application This allows you to share work between users on different networks and provides a way to backup configuration to a file area The following objects may be packaged a Whole projects Individual processes Reference Data sets of all types Notes Data Stores Staged Data configurations Data Interfaces Export configurations Job Configurations Result Book configurations External Task definitions a Web Services Published Processors Note As they are associated with specific server users issues cannot be exported and imported nor simply copied between servers using drag and drop 2 7 1 Packaging objects To package an object select it in the Project Browser right click and select Package For example to package all configuration on a server select the Server in the tree or to package all the projects on a server select the Projects parent node and select Package in the same way 2 24 Using Oracle Enterprise Data Quality Packaging You can then save a Director package file with a dxi extension on your file system The package files are structured files that will contain all of the objects selected for packaging For example if
28. and adding a soundex or metaphone transformation may make the clusters too large In this case an additional cluster may be required If an additional cluster is configured on another identifier EDQ will create entirely new clusters and perform the same matching process on them For example you could choose to cluster on the first three digits of a post code such that all records with a post code of CB4 will be in a single cluster for matching purposes It is also possible to cluster the same identifier twice using different clustering configurations with different transformations To do this create two clusters on the same identifier and configure different transformations to work on the identifier values The more clusters there are the more likely matching is to detect matches However clusters should still be used sparingly to ensure that matching performance does not suffer The more clusters that exist the more records matching has to compare 3 6 Using Oracle Enterprise Data Quality Clustering Concept Guide Composite clusters Composite clusters allow a more sensitive and efficient way of dividing up data sets into clusters using multiple identifiers In this case different parts of each identifier are combined to create a single cluster key that is used to group records for matching For example when matching customers the following cluster might be configured using a combination of Surname and Postcode identifiers
29. can create it at the System level outside of any specific project in the same way as above The next step is Adding a Snapshot of data from your data store Getting Started 1 13 Creating a Project 1 9 Creating a Project To create a Project for working on a specific set or sets of data 1 From the menu select File New Project or 2 Right click on Projects in the Project Browser and select New Project of localhost dnadmin om ree AMD New Project on y 3 Follow the steps in the wizard giving the project a Name and an optional Description 4 Assign the user permissions required for the new Project By default all user groups will have access to the Project The new project is created and visible in the Project Browser You may want to Add a Note to the project to share amongst project users See Adding a Project Note for more information Otherwise the next step is Connecting to a Data Store 1 10 Exporting Data Prepared exports There are two sources of data for prepared exports Data from Staged Data and data from Results Books Note It is also possible to create an ad hoc export to Excel directly from the Results Browser 1 10 1 Staged Data Export Once a Staged Data table has been created with a Writer processor an Export to write it to a Data Store can be created The Export may then be manually run or executed when the process that generates the Staged Data table is run
30. default option This is recommended for most normal runs of a process With this option selected a sample of records will be made available for every drilldown generated by the process This ensures that you can explore results as you will always see some records when drilling down but ensures that excessive amounts of data are not written out by the process None means that the process will still produce metrics but drilldowns to the data will be unavailable This is recommended if you want the process to run as quickly as possible from source to target for example when running data cleansing processes that have already been designed and tested a Publish to Dashboard This option sets whether or not to publish results to the Dashboard Note that in order to publish results you first have to enable dashboard publication on one or more audit processors in the process 2 4 Using Oracle Enterprise Data Quality Execution Options 2 1 2 3 Run Modes To support the required Execution Types EDQ provides three different run modes If a process has no readers that are connected to real time providers it always runs in Normal mode If a process has at least one reader that is connected to a real time provider the mode of execution for a process can be selected from one of the following three options Normal mode In Normal mode a process runs to completion on a batch of records The batch of records is defined by the Reader
31. during processing are included Results Book Export Tasks The notification displays the status of the results book export task in the execution of the job The possible statuses are a FINISHED the results book export ran to completion a CANCELLED the job was canceled by a user during the results book export task a ERROR the results book export failed to complete due to an error Details of any errors encountered during processing are included External Tasks The notification displays the status of the external task in the execution of the job The possible statuses are FINISHED the external task ran to completion CANCELLED the job was canceled by a user during the external task ERROR the external task failed to complete due to an error Details of any errors encountered during processing are included How To 2 15 Optimizing Job Performance Example Notification The screenshot below shows an example notification email using the default email template Project OLH Job My job Status FINISHED Phase Download Reference Data Task Type Name Status Warnings Errors EXTERNALTASKS Download Nasdaq Trader Data FINISHED Phase Snapshot data Snapshot Status Records Warnings Errors Snapshot Customer Data FINISHED 1500 Phase Profile Customer Data Process Status Readers Writers Warnings Errors Profile Customer FINISHED Reader Records Writer Records Records found without cluster Snapshot
32. in production you may want to avoid the step of copying data into the repository as you always want to use the latest set of records in 2 16 Using Oracle Enterprise Data Quality Optimizing Job Performance your source system and because you know you will not require users to drill down to results To stream data into a process and therefore bypass the process of staging the data in the EDQ repository create a job and add both the snapshot and the process as tasks Then click on the staged data table that sits between the snapshot task and the process and disable it The process will now stream the data directly from the source system Note that any selection parameters configured as part of the snapshot will still apply o Papae customer data EJ New Phase 7 Customer Data Ja Customer IT CUST IndividualDataPr Ji Prepare epared TY fs EntityDataPrepar ed he a P eee ee ae a ee 2 oe a _ Note that any record selection criteria snapshot filtering or sampling options will still apply when streaming data Note also that the streaming option will not be available if the Data Store of the Snapshot is Client side as the server cannot access it Streaming a snapshot is not always the quickest or best option however If you need to run several processes on the same set of data it may be more efficient to snapshot the data as the first task of a job and then run the dependent pro
33. intelligent sort and filter enablement is used This means that results based on the snapshot may be sorted or filtered using any column s provided the snapshot is under a certain size set by the system administrator If the snapshot is above that size results based on it cannot be sorted or filtered by any column though users will be prompted to enable sorting and filtering on specific columns if they attempt to do it using the Results Browser Alternatively you can switch off intelligent sort and filter enablement and manually select the columns that you enable for sorting and filtering The default threshold above which sorting and filtering will be disabled for snapshots when using intelligent sort and filter enablement is 10 million cells so for example a snapshot with 500 000 rows and 15 columns 7 500 000 cells would have sorting and filtering enabled but a snapshot with 500 000 rows and 25 columns 12 500 000 cells would have sorting and filtering disabled 1 2 Using Oracle Enterprise Data Quality Adding Reference Data 8 9 Note Itis advisable to select all columns The columns to work with in a given process can be a subset of these Optionally filter the table or view to snapshot a subset of it or you may write your own SQL WHERE clause Optionally sample the selected data for example the first n records the first n records after an offset or 1 record in every 100 Optionally perform no d
34. matching in EDQ is that not every match and output decision can be made automatically Often the fastest and most efficient way to determine if possible matches should be considered as the same and what their merged output should be where appropriate is to look through the matching records manually and make manual decisions EDQ provides two different applications for reviewing match results Match Review and the Case Management The review application is chosen as part of the match processor configuration and determines the way in which the results are generated For any given processor therefore the two applications are mutually exclusive The selection of the review application also changes some of the options available to the match processor It is possible to change the configuration of a match processor at any time to switch between review applications but the processor must be re run before any results are available to the new application The review application to be used should be chosen based upon the requirements of the process downstream of the match processor and the strengths of each review application which are detailed in the following sections 2 16 1 Match Review The Match Review application is a lightweight review application which requires no configuration It supports the manual review of match results and merge results the persistence of manual match decisions between runs and the import and export of manual ma
35. merged process Reader List Gheck E gt Transfonred gt Uniranstornmed gt Invalid i i a i Fi Hs en amet diia eO ra ees at ool 3 SOR on min y Fn le en s Sin ii s o Sa er ae ae ih EEN Jh er pr ae i The Upper Case Processor configuration would appear as follows with the latest version of the title attribute greyed out to indicate that it is ambiguous and therefore not available for use In this case you need to select one of the specific attributes listed under the title attribute ie Upper Case Attributes Options Dashboard Notes Input Attributes Selected Attributes Gutput Attributes title Replaced Eh Added 1 he title Upper E deed Flags iD k iD title Replaced i H A Firstname string 7 m fe Added E Flags a i gt Incompatible d a m aitan a adiis L i n ee a S one lanl en oa a ee ee ee A ie a J F ae ae a i F This scenario will commonly occur while configuring the Merge Attributes processor as this is often used to unite separate processing paths Getting Started 1 9 Configuring Fixed Width Text File Formats 1 7 Configuring Fixed Width Text File Formats When you define a new data store that connects to a fixed width text file the New Data Store wizard will prompt you to define the names and sizes of the data fields in the file Data in a fixed width text file is arranged in rows and columns wit
36. or stakeholders in your data quality project can monitor data quality as it is checked on a periodic basis Results are optionally published on process execution To set up an audit processor to publish its results when this option is used you must configure the processor to publish its results To do this use the following procedure 1 Double click on the processor on the Canvas to bring up its configuration dialog 2 Select the Dashboard tab Note In Parse this is within the Input sub processor 3 Select the option to publish the processor s results 4 Select the name of the metric as it will be displayed on the Dashboard 5 Choose how to interpret the processor s results for the Dashboard that is whether each result should be interpreted as a Pass a Warning or a Failure How To 2 23 Packaging Once your process contains one or more processors that are configured to publish their results to the Dashboard you can run the publication process as part of process execution To publish the results of a process to the Dashboard 1 From the Toolbar click the Process Execution Preferences button 2 On the Process tab select the option to Publish to Dashboard 3 Click the Save amp Run button to run the process When process execution is complete the configured results will be published to the Dashboard and can be made available for users to view For more information see Understanding Enterprise Data Quality
37. processes or when revisiting existing processes it is often desirable to investigate how the process was set up and investigate errors if they exist 2 12 1 Invalid Processor Search Where many processors exist on the canvas you can be assisted in discovering problems by using the Invalid Processor Search right click on the canvas This option is only available where one or more processors have errors on that canvas How To 2 39 Investigating a Process Hyt H cory Fa Paste Corl Daspe Dehete mout Atu sA Clear Search Highlight i f A aims E E r a q n DO E e el ee ee ey re pana gh mna a l P EE N E T ER E e y a O a E ee T This brings up a dialog showing all the invalid processors connected on the canvas Details of the errors are listed for each processor as they are selected The Jump To Processor option takes the user to the selected invalid processor on the canvas so the problem can be investigated and corrected from the processor configuration W Invalid Processors QAS Summary Results Attribute qas outcome summary does not exist Cancel 2 12 2 Input Attribute Search You may list the attributes which are used as inputs to a processor by right clicking on that processor and selecting Input Attribute Search The icons used indicate whether the attribute is the latest version of an attribute or a defined attribute 2 40 Using Oracle Enterprise Data Quality
38. processor Derive Gender from Forename Once we have selected the options that we want to publish we can choose how these will be labeled on the new processor To do this click on Processor Setup button at the top of the canvas and rename the options For example we might label the two options published above Title Gender Map and Forename Gender Map ig Processor Setup Published options con Editor a F Forename Gender Map GE Gender from Forename boon L 4 Derive Gender From Forename Reference Data Value Map lg Title Gender Map GE Gender from Title gang Derive Gender From Title En Data Value Map 2 9 4 Setting Output Attributes The Output Attributes of the new processor are set to the output attributes of any one but only one of the member processors By default the final member processor in the sequence is used for the Output Attributes of the created processor To use a different member processor for the output attributes click on it and select the Outputs icon on the toolbar 2 32 Using Oracle Enterprise Data Quality Creating Processors i Add Gender Master KA The member processor used for Outputs is marked with a green shading on its output side q 5 verge to create best Gender Note Attributes that appear in Results Views are always exposed as output attributes of the new processor You may need to add a member processor to profile or check the output at
39. see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 22 Publishing to Case Management Case Management is enabled in the Advanced Options dialog of Match processors To enable Case Management for a processor open it and click the Advanced Options link Advanced Options PAssign Relationship Review O Review Results Assign Merged Review Match Individuals Configure Bulk Review Rules View Match Statistics Delete Realtime Review Results Delete Manual Decisions Externalize ee a a a l ee iah In the Advanced Options dialog select Case Management in the Review System drop down field Match groups share working records Cluster size limit Cluster comparison limit Match Group Size Limit 5000 Review Group Size Limit 5000 Allow null dusters Use review relationships in match groups E Token attribute prefix Clustered _ Sort Filter Enable Sort Filter Relationship Decision Trigger Review System Note The Use review relationships in match groups check box is disabled when Case Management is selected and the Decision Key tab is replaced by the Case Source tab How To 2 73 Execution Types Further configuration is required before Case Management can be used with this processor Configure a case source for the processor using the Case Source tab Map the input data streams to the data sources defined in the case source When these steps are com
40. tasks as required on the Job Canvas 2 10 Using Oracle Enterprise Data Quality Using Job Triggers 2 2 5 Editing and Configuring Job Phases Phases are controlled using a right click menu The menu is used to rename delete disable configure copy and paste Phases OC H TABIRA Rename Delete ivi Disable Configure Copy Paste ath il sleeps E ie ne The using the Add Delete Up and Down controls at the bottom of the Phase list t Tt For more details on editing and configuring phases see Creating a Job above and also the Using Job Triggers topic For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 3 Using Job Triggers Job Triggers are used to start or interrupt other Jobs Two types of triggers are available by default Run Job Triggers used to start a Job a Shutdown Web Services Triggers used to shut down real time processes Further Triggers can be configured by an Administrator such as sending a JMS message or calling a Web Service They are configured using the Phase Configuration dialog an example of which is provided below How To 2 11 Using Job Triggers Phase Configuration Name Initiate Real time Services Enabled Execution Condition Exeu te on success Clear Error E Triggers Q Real time Entity Clean Real time Entity Cluster Real time Entity Match Real time Individual C
41. the button next to each option is marked with green to indicate that it is externalized 6 When all the options are externalized as required click OK to save or Cancel to abandon 2 19 1 2 Rename externalized options Externalized processor options are used overridden in jobs at the process level Each externalized processor option for the process has a default name Use the following procedure to assign a different name 1 Right click on the relevant processor and select Externalize The Process Externalization dialog is displayed This displays all the externalized options available in the process 2 Ensure the Enable Externalization check box is selected 3 Click on the required option in the top area of the dialog The bottom area shows the processor s that the option is linked to and allows you to link directly to the processor s in the Canvas 2 58 Using Oracle Enterprise Data Quality Externalizing Configuration Settings 4 Click Rename 5 Click OK to save or Cancel to abandon Note The name you give to the externalized option is the name that must be used when overriding the option value either in a Run Profile or from the command line using the syntax for such override options For a full guide to this syntax complete with examples of overriding externalized option values see the instructions in the template properties file that is provided in the oedq_local_ home runprofiles directory of your
42. the family icon use the right side of the screen Note that when publishing a processor it will be published into the selected group or a new family created if it does not yet exist 5 For both processor and family icons a dialog is launched showing the server image library You can either select an existing image or create a new image 6 If adding a new image a dialog is shown allowing you to browse for or drag and drop an image resize it and enter a name and optional description Once an image has been created on a server it is added to the server s image library and available whenever customizing an icon The image library on a server can be accessed by right clicking on a server in the Project Browser and selecting Images For more information see Enterprise Data Quality Online Help 2 76 Using Oracle Enterprise Data Quality Publishing Result Views to Staged Data 2 25 Enabling Case Management The Case Management and Case Management Administration applications are published on the EDQ Launchpad by default However if they are not present they must be added to the Launchpad and User Groups assigned to them For further information see the Launchpad Configuration and Application Permissions topics in Enterprise Data Quality Online Help 2 26 Publishing Result Views to Staged Data EDQ can publish or write top level Results Views of processors to Staged Data Note Top level here means the fir
43. the same if the second is simply missing its house number Making a decision as to whether we should treat these as the same depends on factors such as a What is the information being used for Do we have any other information that will help us make a decision Advanced Features 3 1 Matching Concept Guide Effective matching requires tools that are much more sophisticated than traditional data analysis techniques that assume a high degree of completeness and correctness in the source data It also demands that the business context of how the information is to be used is included in the decision making process For example should related individuals at the same address be considered as one customer or two How EDQ solves the problem EDQ provides a set of matching processors that are suited to the most common business problems that require matching The matching processors use a number of logical stages and simple concepts that correspond with the way users think about matching Identifiers Rather than forcing users to express matching rules at the field by field level EDQ s matching processors exploit the powerful concept of Identifiers Identifiers allow the user to map related fields into real world entities and deliver a range of key benefits a Where similar information is stored in different applications or databases any naming differences between fields can be overcome by mapping identifiers This is illustra
44. within the Project Browser and either double click it or right click and select Edit 2 The Job is displayed in the Job Canvas Edit the Phases and or Tasks as required 3 Click File gt Save 2 2 3 Deleting a Job Deleting a job does not delete the processes that the Job contained and nor does it delete any of the results associated with it However if any of the processes contained in the Job were last run by the Job the last set of results for that process will be deleted This will result in the processors within that process being marked as out of date To delete a Job either a select it in the Project Browser and press the Delete key or a right click the job and select Delete Remember that it is not possible to delete a Job that is currently running 2 2 4 Job Canvas Right Click Menu There are further options available when creating or editing a job accessible via the right click menu Select a task on the canvas and right click to display the menu The options are as follows a Enabled If the selected task is enabled there will be a checkmark next to this option Select or deselected as required Configure Task This option displays the Configure Task dialog For further details see the Running Jobs Using Data Interfaces topic Delete Deletes the selected task Open Opens the selected task in the Process Canvas Cut Copy Paste These options are simply used to cut copy and paste
45. you package a project all its subsidiary objects data stores snapshot configurations data interfaces processes reference data notes and export configurations will be contained in the file Note Match Decisions are packaged with the process containing the match processor to which they are related Similarly if a whole process is copied and pasted between projects or servers its related match decisions will be copied across If an individual match processor is copied and pasted between processes however any Match Decisions that were made on the original process are not considered as part of the configuration of the match processor and so are not copied across 2 2 Filtering and Packaging It is often useful to be able to package a number of objects for example to package a single process in a large project and all of the Reference Data it requires in order to run There are three ways to apply a filter a To filter objects by their names use the quick keyword NameFilter option at the bottom of the Project Browser To filter the Project Browser to show a single project hiding all other projects right click on the Project and select Show Selected Project Only a To filter an object such as a process or job to show its related objects right click on the object and select Dependency Filter and either Items used by selected item to show other objects that are used by the selected object such as t
46. Comment fields are optional The Auto Map button attempts to map the fields automatically by looking for any of the default names for these fields The default names for the fields are as follows as these are the names created when writing out relationships from a match processor in EDQ Decision Field Decision Field Decision User Field Decision Timestamp Field Comment Field Comment User Field Comment Timestamp Field Data Stream Name Related Data Stream Name Default Auto Name RuleDecision ReviewedBy ReviewDate Comment CommentBy CommentDate DataStreamName RelatedDataStreamName How To 2 53 Importing Match Decisions Decision Field Default Auto Name Match Rule Name MatchRule State Expiry Timestamp Expiry Date This means that if you are importing decision data back into the same match processor that generated the relationships for review you should be able to use Auto Map to map all the decision fields automatically Note The State Expiry Timestamp field is only shown for processors using Case Management to handle reviews The second section of the first tab allows you to specify the actual decision values of the configured Decision field ManualDecision in the example above and how they map to the decision values understood by EDQ In the case above the field values are the same as those used by EDQ Match No Match Possible Match and Pending The third section allows you to specify the n
47. Customer Profiled Customer oys 1500 1500 Data Data Phase Export data and results book Task Type Name Status Warnings Errors EXPORT Profiled Customer Data to Customer DB FINISHED RESULT_EXPORT Profile Customer Data Results FINISHED For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 5 Optimizing Job Performance This topic provides a guide to the various performance tuning options in EDQ that can be used to optimize job performance 2 5 1 General Performance Options There are four general techniques applicable to all types of process that are available to maximize performance in EDQ Click on the headings below for more information on each technique a Data streaming a Minimized results writing a Disabling Sorting and Filtering 2 5 1 1 Data streaming The option to stream data in EDQ allows you to bypass the task of staging data in the EDQ repository database when reading or writing data A fully streamed process or job will act as a pipe reading records directly from a data store and writing records to a data target without writing any records to the EDQ repository Streaming a Snapshot When running a process it may be appropriate to bypass running your snapshot altogether and stream data through the snapshot into the process directly from a data store For example when designing a process you may use a snapshot of the data but when the process is deployed
48. Data Quality Creating Processors Note It is also possible to change the Name of each of the input attributes in this screen which means their names will be changed within the design of the processor only without breaking the processor if the actual input attributes from the source data set in current use are different This is available so that the configuration of the member processors matches up with the configuration of the new processor but will make no difference to the behavior of the created processor 2 9 3 Setting Options The processor design page allows you to choose the options on each of the member processors that you want to expose or publish for the processor you are creating In our example above we want the user to be able to select their own Reference Data sets for mapping Title and Forename values to Gender values as for example the processor may be used on data for anew country meaning the provided Forename to Gender map would not be suitable To publish an option open the member processor in the processor design page select the Options tab and tick the Show publishing options box at the bottom of the window You can then choose which options to publish If you do not publish an option it will be set to its configured value and the user of the new processor will not be able to change it unless the user has permission to edit the processor definition There are two ways to publish option
49. E Missierna heey a S E 2 41 Previewing Published Results Views ecrinncnorrererriii re ia RE EEEE 2 42 Using the Res NS DEO W SOE sareore onea e cat nacadreaationelearaat enue eabetaneuan doe suntintin A atstncaan 2 43 Show Results in New Window sessiooni tai ceca Aidt aes eee eaten eee 2 43 Show Characterscenar a A T Man Aaa Hause patina 2 43 De lech Colin eae a E E un E TS 2 45 210 Using tne Event LOG ne E Meneame 2 45 2 15 1 B 2427 lt 0 Bd whk f 210 be seme ner eee eer ote nr set ee mr ese Tenner eT reer aerR oor ace ones me ena neT raert re roan 2 46 2 15 2 DOEV Cl SCLC ON E EE E E A aes ids E naan atontaalss 2 46 2 15 3 PENTO OV CIES aa Sarat asta ahc vali c uate sand EOR 2 46 PNG Review Mar hine TRESS iee accra sce eat ra A E T A 2 50 2 16 1 Mate PREVI WE E ere E EN ATO 2 50 2 16 2 Case Mana eri eiit sera S A EE RN 2 51 217 Tmportne Match Decision erase a A A 2 51 2 17 1 Step 1 Connect the decisions data into the match processor sesessssesesesesessereses 2 51 2 17 2 Step 2 Specifying the attributes that hold the new decision data ccceee 2 52 2 17 3 Step o Mapping the decisions Gata Tields erais arn eE R 2 54 2 17 4 Step t mporn tne CECISIONS si siniie gu peliaen a ANA 2 55 218 Exporting Watch Decision S en ena aol meas ma ceaebicnur nese eet 2 56 2 19 Extermalizine Contieuration Serine S sesca atioi nein iin A 2 57 2 19 1 Externalizing Processor OPTIONS Sasima e aia a a EA S 2 57 2 19 2 Ext ermalizine Mat
50. Externalization box The Externalized Options area shows the options that are available for externalization 2 60 Using Oracle Enterprise Data Quality Externalizing Configuration Settings To delete an option select it and click Delete under the Externalized Options area To rename an option select it and click Rename Edit the name in the Rename pop up and click OK The Externalized Tasks area shows the selected option next to the Task it is associated with If an option is associated with more than Task it is listed once for each Task The example dialog above shows that the Enabled option is associated with the UK Customers and US Customers tasks To disassociate an option from a Task select it in this area and click Delete 2 19 4 Externalizing Snapshots Only snapshots from server side Data Stores can be externalized Note that externalizing snapshots allows external users to override not only configuration of the snapshot itself such as the table name sample size and SQL WHERE clause but also options of the Data Store the snapshot is being read from for example the file name or database connection details Use the following procedure 1 Right click on the required snapshot in the Project Browser 2 Select Externalize The Snapshot Externalization dialog is displayed Note that both Snapshot and Data Store options are available 3 Select an option by checking the corresponding box The field on t
51. ORACLE Oracle Fusion Middleware Using Oracle Enterprise Data Quality 12c 12 2 1 E65907 01 October 2015 Describes how to use the Oracle Enterprise Data Quality applications Oracle Fusion Middleware Using Oracle Enterprise Data Quality 12c 12 2 1 E65907 01 Copyright 2015 Oracle and or its affiliates All rights reserved This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws Except as expressly permitted in your license agreement or allowed by law you may not use copy reproduce translate broadcast modify license transmit distribute exhibit perform publish or display any part in any form or by any means Reverse engineering disassembly or decompilation of this software unless required by law for interoperability is prohibited The information contained herein is subject to change without notice and is not warranted to be error free If you find any errors please report them to us in writing If this is software or related documentation that is delivered to the U S Government or anyone licensing it on behalf of the U S Government then the following notice is applicable U S GOVERNMENT END USERS Oracle programs including any operating system integrated software any programs installed on the hardware and or documentation delivered to U S Government end users are commercial computer s
52. Title values to Gender values and one to map Forename values to Gender values Finally you want three output attributes TitleGender NameGender and BestGender from the processor To do this you need to start by configuring the member processors you need or you may have an existing process from which to create a processor For example the screenshot below shows the use of 5 processors to add a Gender attribute as follows 1 Derive Gender from Title Enhance from Map 2 Split Forename Make Array from String 3 Get first Forename Select Array Element 4 Derive Gender from Forename Enhance from Map 5 Merge to create best Gender Merge Attributes H gt oF H gt a Gay eme Gy gt gt gt gt Ay bares gt D gt Not enhanced D D D gt Not enhanced d Derive Gender 2 Split Forename 3 Get irst 4 Derive Gender 5 Merge to create from Title Forename from Forename best Gender eee ee hak aml ee te a a ae ee ca nahm 5 j How To 2 29 Creating Processors To make these into a processor select them all on the Canvas right click and select Make Processor This immediately creates a single processor on the Canvas and takes you into a processor design view where you can set up how the single processor will behave From the processor design view you can set up the following aspects of the processor though note that in many cases you will be able to use the default settings
53. a Quality Online Help 2 17 Importing Match Decisions EDQ provides functionality allowing you to import decision data and apply it to possible matches within the Match processor Importing decisions may be a one off activity or an ongoing part of the matching process For example if you are migrating your matching processes into EDQ from a different application it will be desirable to include all the previous decisions as part of the migration Alternatively if the review process is carried out externally decision import may be used routinely to bring the external decisions back into EDQ It is also possible to import match decisions that were exported from another process or instance of EDQ for example from another match process working on the same data All match decisions in EDQ refer to a pair of possibly matching records For an imported decision to be applied therefore it must be possible to match it to a relationship which was created by the match processor with a Review rule If you import a decision for a pair of records that the processor has identified as a definite match or are not matched at all the decision will not be applied unless and until the matching logic is changed so that a Review relationship is created for the records Note that it is possible to import decisions and use Case Management or Match Review to create manual decisions If a decision with a later date time than any manual decisions for a relations
54. ability and Extensibility EDQ comes with a highly configurable and tuneable library of matching algorithms allowing users to refine their matching process to achieve the best possible results with their data In addition EDQ provides the ability to define new matching algorithms and approaches The best matching functions depend entirely on the problem being addressed and the nature of the data being matched All key elements of the matching process can use extended components For example a What we use to identify the records How we compare the records How we transform and manipulate the data to improve the quality of comparisons The combination of configurability and extensibility ensures that the optimal solution can be deployed in the shortest possible time See the Extending EDQ topic in the Enterprise Data Quality Online Help for more information about adding matching extensions into the application 3 4 Using Oracle Enterprise Data Quality Clustering Concept Guide Key Features Key features of matching in EDQ include Matching for any type of data Guidance through the configuration of the matching process User definable data identification comparison and match rules Business rule based matching Multi user manual review facility for both match and merge decisions Remembers manual decisions when new data is presented Automates the production of output from the matching process
55. ame values expected in the decisions data for the data streams used by the match processor This is especially important for match processors with multiple data sets in order to match each imported decision with the correct records from the correct data streams In the case above the match processor is a Deduplicate processor working on a single data stream This means that all the records both records in each relationship are from the same data stream Customer Data As the Name Customer Data is specified above all the decisions data being imported must have the value Customer Data in both the DataStream and RelatedDataStream attributes 2 17 3 Step 3 Mapping the decisions data fields Next use the Mapped Fields tab of the Decisions sub processor to map the fields in the imported decisions data to the necessary fields in the match relationships The requirements for this mapping differ depending on whether you are using Match Review or Case Management to process your potential matches a When Match Review is in use only the decision data fields which correspond to the Decision Key fields in the matching data need to be mapped For ease of reference these fields are highlighted in red in the mapping tab 2 54 Using Oracle Enterprise Data Quality Importing Match Decisions General Settings Mapped Fields Record Customer Data Customer Date Input Fields Decision Fields Name Mame MameStandardized
56. anced Micro Devices UNIX is a registered trademark of The Open Group This software or hardware and documentation may provide access to or information about content products and services from third parties Oracle Corporation and its affiliates are not responsible for and expressly disclaim all warranties of any kind with respect to third party content products and services unless otherwise set forth in an applicable agreement between you and Oracle Oracle Corporation and its affiliates will not be responsible for any loss costs or damages incurred due to your access to or use of third party content products or services except as set forth in an applicable agreement between you and Oracle Contents Prle eer ae A ee eee eee ae oe ame er eee near eee eee eee ae ee eee ee vii FA Ge aig Gl cee en ee me ee oO CPR ce RO ROTI Te PRT TS SE eT ee DRC PEE SE a TET vii DOCUMENTATION Accessibile E a E E A wunae aus vii PREV ES O a a a vii CONC E1 1 ssl access EEEE EPOE AEEA Rae heals ih ee Saas Shasta AEE viii 1 Getting Started 1 1 PRG II SA roceSS carea a N A a Satna SeeunSenaens 1 1 1 2 Addins d IANO S OU ascent acta e T S 1 2 1 3 Addins Relerence 1 QUA csras ote herald ac oath ates ctaae a A e da ncgttoteiceoadath test 1 3 1 4 PRN ATINS AN BSU Craii O E A E EEEN 1 4 1 5 Adda a Proce NO nane a N A nasa aoe au aioe 1 5 1 6 Addn Proces SOL Snan A E T IS T TA 1 6 1 6 1 Ambig ous Latest Atribute ceres a a E N 1 9 1 7 Confieurmne Fixed
57. anvas and then select Processor Help or left click on a processor on the canvas or tool palette and then press F1 Conventions vill The following text conventions are used in this document Convention Meaning boldface Boldface type indicates graphical user interface elements associated with an action or terms defined in text or the glossary italic Italic type indicates book titles emphasis or placeholder variables for which you supply particular values monospace Monospace type indicates commands within a paragraph URLs code in examples text that appears on the screen or text that you enter 1 Getting Started This chapter provides information on the basic operations you will perform when using Director This chapter has the following sections Adding a Process Adding a Snapshot Adding Reference Data Adding an Issue Adding a Project Note Adding Processors Configuring Fixed Width Text File Formats Connecting to a Data Store Creating a Project Exporting Data Prepared exports Running a Prepared Export 1 1 Adding a Process To add a Process to analyze data from a snapshot 1 2 z E Dynamic Configuration Demo From the menu select File New Process or Right click on Processes in the Project Browser and select New Process 8 4 Data Stores 9 OAV Test Data HI Oracle Demo HI Service Management H O Staged Data D0 Data Interfaces Getting Started 1 1 Adding
58. any 4 Click Save to save the issue The issue is added and available from the Issue Manager If the issue was assigned to another user that user will be notified of the outstanding issue immediately if he she is logged on 1 5 Adding a Project Note To add a Note to a project for example to attach a project plan or to share some key information amongst project users 1 Right click Notes in the Project Browser and select New Note E Dynamic Configuration Demo H A Data Stores f py Staged Data Qi Data Interfaces ft Processes re Reference Data aS Results Books ngri External Tasks EE Jobs Ef Exports 2 Give the note a Title Getting Started 1 5 Adding Processors 3 Add detail to the note or leave blank if all you need is a Title and one or more attachments 4 Browse your file system to add a file attachment to the note or drag and drop files onto the indicated area 5 Click Save The note is created and visible in the Project Browser 1 6 Adding Processors EDQ comes with a library of processors for processing your data To add a processor to your process 1 Ensure your process is open on the Canvas ProjectBrowser SSO Profile Customers table 4 localhost dnadmin EEE Projects ma t Customer data migration m Reader Quick stats Profler 5 00 Customer DE jg a Data Stores Pattems Proiler a G Staged Data a
59. as a standalone task in Director by right clicking on the Export and selecting Run it always runs in Delete current data and insert mode Delete current data and insert and Replace records using primary key modes perform Delete then Insert operations not Update It is possible that referential integrity rules in the target database will prevent the deletion of the records therefore causing the Export task to fail Therefore in order to perform an Update operation instead Oracle recommends the use of a dedicated data integration product such as Oracle Data Integrator 2 1 5 Results Book Exports When a Results Book Export is configured to run as part of a job there is a single option to enable or disable the export allowing you to retain the same configuration but temporarily disable the export if required 2 1 6 Triggers Triggers are specific configured actions that EDQ can take at certain points in processing Before Phase execution in a Job After Phase execution in a Job For more information see Using Triggers in Administering Oracle Enterprise Data Quality and the Advanced options for match processors topic in Enterprise Data Quality Online Help 2 2 Creating and Managing Jobs This topic covers Creating a Job Editing a Job Deleting a Job Job Canvas Right Click Menu Editing and Configuring Job Phases Note alt is not possible to edit or delete a Job that is currently running Always check t
60. ata normalization For more information see the No Data Handling topic in Enterprise Data Quality Online Help Give the snapshot a Name and choose whether or not to run it immediately Click Finish to confirm the addition of the snapshot The snapshot is created and visible in the project browser It is now ready to be run by Right click Run Snapshot or used in a process and run later The snapshot may also be streamed that is used as a way of selecting the records to be processed from a Data Store directly that is without copying them into the repository The next step is Adding a Process for example to profile the data in the snapshot 1 3 Adding Reference Data To add a set of Reference Data to use in processors for the validation or transformation of data 1 Right click on Reference Data in the Project Browser and select New Reference Data If you want the Reference Data to be a lookup onto Staged or External Data or an alternative lookup onto an existing Reference data set rather than a new set of Data choose New Lookup ei f Dynamic Configuration Demo aa Lilie Laka ee i m ie ae Data Stores a H Staged Data 00 Data Interfaces H t Processes 5 Result New Reference Data H Extern New Lookup aot Or create the Reference Data using the data in the Results Browser by selecting some data values right clicking and selecting Create Reference Data For example
61. ata types 2 19 6 1 Example of the Export Externalization dialog for an Access database 8 Exteralization of Export a Export Table name E table_name Create new table create_new_table new table name New table name Data Store Access file E jaccess_file Use project specific landing area fic_landing_area cancel 2 19 6 2 Example of the Export Externalization dialog for a Delimited Text file Note that in this case the only options available are those of the target file itself The configuration of the export within the job determines whether or not a new file is written or if an existing file is appended to this setting cannot be externalized How To 2 65 Managing Data Interfaces Externalization of Export Data Store File in server work area Entity Output _c Use project specific landing area fic_landing_area Treat first row as header E row_as_header Field separator Entity_Output_C Quote character E guote_character Number of columns to assume imns_to_assume Encoding Entity_Output_c Skip lines at start E p_lines_at_start 2 20 Managing Data Interfaces This topic covers Adding a Data Interface Editing a Data Interface Creating Data Interface Mappings Deleting Data Interfaces 2 20 1 Adding a Data Interface 2 66 1 Right click the Data Interfaces node in the Project Browser and select New Data Interface The Data Interfa
62. ce Set mresa re rence records Staged matching i staged Enhance Ference Set 2 Staged wiki i ie n d T Sel a Seg mal ee i ak mAT O y hoek a Ee i a ma d T A i nae Preparing a process for real time reference matching For a real time reference matching process to perform correctly it must first be run in Prepare mode This allows it to compare the inbound records quickly against the reference sets and issue a response Running a process in Prepare mode will ensure that all the cluster keys for the reference data sets have been produced To run a process in Prepare mode set up a job and add the relevant process as a task Click on the process to set the configuration details Click on the Run Mode tab and select Prepare Details Type Process Detail R T reference matching 1 Enabled Readers Process Run mode Writers Normal Prepare Interval New interval after record Re preparing reference data Real time matching uses a cached and prepared copy of the reference data This means that when updates are made to reference data the snapshot must be re run and the data re prepared before the updates can be propagated to the matching process 3 12 Using Oracle Enterprise Data Quality Real Time Matching Re preparing reference data involves 1 Stopping the real time match process 2 Refreshing the snapshot of the reference data 3 Running the real time match process in pre
63. ce dialog is displayed es Example Project E HY Data Stores H Data Erte t Pi H New Data Interface a i 2 Add the attributes that you require in the Data Interface or paste a copied list of attributes To create a Data Interface from Staged or Reference Data 1 Right click on the object in the Project Browser and select Create Data Interface Mapping The New Interface Mappings dialog is displayed Using Oracle Enterprise Data Quality Managing Data Interfaces 2 Click New Data Interface to create an interface with the attributes and data types of the Staged or Reference Data selected 2 20 2 Editing a Data Interface 1 Right click on the Data Interface in the Project Browser 2 Select Edit The Data Interface Attributes dialog is displayed 3 Edit the attributes as required Click Finish to save or Cancel to abandon 2 20 3 Creating Data Interface Mappings Data Interface mappings are required so that data can be bound into or out of a process or a job To create a mapping either a right click on the Data Interface select Mappings and click the button on the Data Interface Mappings dialog to start the New Mapping wizard or a right click on a Staged Data or Reference Data set select Create Data Interface Mapping and click the required Data Interface in the New Data Interface Mappings dialog Note Data sources are either mapped as Data In or Data Out d
64. ce mappings have been specified for each data interface in a job both the mappings and the objects that they bind to appear in the job This means the job can now be run See Example Job containing two Data Interfaces below 2 21 2 Linking processes with a Data Interface It is possible to link two or more processes that contain Data Interfaces provided one is configured as a reader and the other as a writer 1 Add both processes to the job as in the following example Parse Names Batch aA RT Process Chaining Job no EJ How to chain processes EJ OCH 7T BlBIAAR ay New Phase 1 Parse Person Name 2 Add Cluster Keys DI Person DI Person DI Person DDI Person ae H PEHA H Ferson H Person H 2 Click and drag the connector arrow from the first process to the second The processes will be linked 2 70 Using Oracle Enterprise Data Quality Running Jobs Using Data Interfaces Parse Names Batch EJ RT Process Chaining Job no Fa 2 Fow fo chain processes EJ OCaml eT ABIS A a Hew Phase 1 Parse Person Name 2 Add Cluster Keys pD EA Ja Pile Eee 7 DI Person Pw DI Person 7 Person ae Person Person 2 Person f a a ee A a a ae eal 2 21 3 Chaining Processes in a Real Time Job Because of the way EDQ handles data streams within real time jobs there is a limitation on how processes should be chained together As long as only one data stream is maintained within
65. ces between values that are essentially the same enabling the creation of clusters for records that are only similar rather than identical in their identifier values For example a Name identifier may use either a Metaphone or a Soundex transformation during clustering such that similar sounding names are included in the same cluster This allows matching to work where data may have been misspelt For example with a Soundex transformation on a Surname identifier the surnames Fairgrieve and Fairgreive would be in the same cluster so that matching will compare the records for possible duplicates The valid transformations for an identifier vary depending on the Identifier Type for example there are different transformations for Strings as for Numbers For example the following are some of the transformations available for String identifiers Make Array from String splits up values into each separate word using a delimiter and groups by each word value For example JOHN and SMITH will be split from the value JOHN SMITH if a space delimiter is used First N Characters selects the first few characters of a value For example MATT from MATTHEWS Generate Initials generates initials from an identifier value For example IBM from Internal Business Machines EDQ comes with a library of transformations to cover the majority of needs It is also possible to add custom transformations to the system
66. cesses If the source system for the snapshot is live it is usually best to run the snapshot as a separate task in its own phase so that the impact on the source system is minimized Streaming an Export It is also possible to stream data when writing data to a data store The performance gain here is less as when an export of a set of staged data is configured to run in the same job after the process that writes the staged data table the export will always write records as they are processed whether or not records are also written to the staged data table in the repository However if you know you do not need to write the data to the repository you only need to write the data externally you can bypass this step and save a little on performance This may be the case for deployed data cleansing processes or if you are writing to an external staging database that is shared between applications for example when running a data quality job as part of a larger ETL process using an external staging database to pass the data between EDQ and the ETL tool How To 2 17 Optimizing Job Performance To stream an export create a job and add the process that writes the data as a task and the export that writes the data externally as another task Then disable the staged data that sits between the process and the export task This will mean that the process will write its output data directly to the external target Note It is
67. ch Proces OLS sasra a R A 2 59 2 19 3 Exea JODS sienen E a es 2 60 2 19 4 PX CCEMGIIZ INS NaP eNO Serra a A SEE A eee 2 61 2 19 5 Externalizine External Tasko ere arian Sic bla tess E A E aihtaaetaa eae 2 63 2 19 6 Externa W711 EXPOS e E E S 2 64 220 Manac me Data Ter Ice Seerne E E E A TEO aA 2 66 2 20 1 Adding a Waite Interlace a O E A 2 66 2 20 2 Editino a Data Te Rta Ce sruni a a T N 2 67 2 20 3 Creatine Data Interlace Mappine Sesniminsieiiioia tual tamer E A 2 67 2 20 4 Deleting Data Titer aces as r AOE 2 68 221 Junie Jobs Using Data Interfaces medenina a E E E aialss 2 69 2 21 1 Conk gurine a Data latertace mka OD s isetarastsaisacediatastiancaescbcsaeomatesangneredslutesaasseantenaes 2 69 2 21 2 Linking processes witha Data Inter tace eteenedoneisor onr o n E vents 2 70 2 21 3 Chainine Processes ina Real Time JoBnasriniieeini A 2 71 2 21 4 Example Job containing two Data Interfaces sssesessesessessssssserseeseseesessrseeseserseeersess 2 71 2 22 Publishing to Case Manacemient swvscnisiiscieststediasmneuktaiertietieiuetadveadtatisetiiuouads 2 73 P29 EX CCUTOIN VCS sasaisses can digo tice tints tackot caus da E O E EOE tolae Mogae 2 74 2 23 1 Pat eion A E sad raulea ceases tens nas 2 74 2 23 2 Real Mime Ves POnSe teens ar tanseapau a arn la eee des 2 75 2 23 3 Realtime VOR LOLI sesen wnueantdey T E 2 75 Cet Kuston Processor ICONS sisuna urate el aipraasee N A 2 76 eco Enabling Case Management si csscicvissssacesaeneatarwnreusaic
68. configuration and any further sampling options that have been set in the process execution preferences or job options Prepare mode Prepare mode is required when a process needs to provide a real time response but can only do so where the non real time parts of the process have already run that is the process has been prepared Prepare mode is most commonly used in real time reference matching In this case the same process will be scheduled to run in different modes in different jobs the first job will prepare the process for real time response execution by running all the non real time parts of the process such as creating all the cluster keys on the reference data to be matched against The second job will run the process as a real time response process probably in Interval mode Interval mode In Interval mode a process may run for a long period of time or even continuously but will write results from processing in a number of intervals An interval is completed and a new one started when either a record or time threshold is reached If both a record and a time threshold are specified then a new interval will be started when either of the thresholds is reached As Interval mode processes may run for long periods of time it is important to be able to configure how many intervals of results to keep This can be defined either by the number of intervals or by a period of time For example the following options might be s
69. cords from the process 2 1 2 1 Readers For each Reader in a process the following option is available Sample The Sample option allows you to specify job specific sampling options For example you might have a process that normally runs on millions of records but you might want to set up a specific job where it will only process some specific records that you want to check such as for testing purposes Specify the required sampling using the option under Sampling and enable it using the Sample option 2 2 Using Oracle Enterprise Data Quality Execution Options The sampling options available will depend on how the Reader is connected For Readers that are connected to real time providers you can limit the process so that it will finish after a specified number of records using the Count option or you can run the process for a limited period of time using the Duration option For example to run a real time monitoring process for a period of 1 hour only Reader DedupAccount Filter Config Sampling Options C Al Ocot C Duration For Readers that are connected to staged data configurations you can limit the process so that it runs only on a sample of the defined record set using the same sampling and filtering options that are available when configuring a Snapshot For example to run a process so that it only processes the first 1000 records from a data source Source A Source A Filter Configuratio
70. debugging a complex process using a small number of records Sample mode is suitable for high volumes of data ensuring that a limited number of records is written for each drilldown The System Administrator can set the number of records to write per drilldown by default this is 1000 records Sample mode is the default None mode should be used to maximize the performance of tested processes that are running in production and where users will not need to interact with results To change the Results Drilldown mode when executing a process use the Process Execution Preferences screen or create a Job and click on the process task to configure it For example the following process is changed to write no drilldown results when it is deployed in production Details Type Process Detail Match individuals 1 Enabled Readers Process Writers Use Intelligent Execution Publish to Dashboard F Enable Sort Filter in Match Results Drill Down OAL Sample None 2 5 1 3 Disabling Sorting and Filtering When working with large data volumes it can take a long time to index snapshots and written staged data in order to allow users to sort and filter the data in the Results Browser In many cases this sorting and filtering capability will not be needed or only needed when working with smaller samples of the data The system applies intelligent sorting and filtering enablement where it will enable sorting and filtering when work
71. dxi 5 BD Press E BO Compliance fel w Customer migration E O E g we Service Management system a Reference Data Dris E Data Stores ee ee eee ee ee eel ee tg me Tn ale em oe Te i pean Bi mal Neem a me e n Note that when multiple objects are imported from a package file and there are name conflicts with existing objects in the target location a conflict resolution screen is shown allowing you to change the name of the object you are importing ignore the object and so use the existing object of the same name or to overwrite the existing object with the one in the package file You can choose a different action for each object with a name conflict If you are importing a single object and there is a name conflict you cannot overwrite the existing object and must either cancel the import or change the name of the object you are importing Once you have completed copying across all the objects you need from a package file you can close it by right clicking on it and selecting Close Package File 2 26 Using Oracle Enterprise Data Quality Purging Results Opened package files are automatically disconnected at the end of each client session 2 7 4 Working with large package files Some package files may be very large for example if large volumes of Reference Data are included in the package When working with large package files it is quicker to copy the file to the server s landingarea for files and open
72. e if the previous phase did not complete successfully a Execute on success default the Phase will only execute if all previous Phases have executed successfully Execute regardless the Phase will execute regardless of whether previous Phases have succeeded or failed Note If an error occurs in any phase the error will stop all Execute on success phases unless an Execute regardless or Execute on failure phase runs with the Clear Error button checked runs first To clear or leave unchanged an error state in the Job If a job phase has been in error an error flag is applied Subsequent phases set to Execute on success will not run unless the error flag is cleared using this option The default state is unchecked How To 2 9 Creating and Managing Jobs Field Type Description Triggers N A To configure Triggers to be activated before or after the Phase has run For more information see Using Job Triggers 6 Click OK to save the settings 7 Click and drag Tasks from the Tool Palette configuring and linking them as required 8 To add more Phases click the Add Job Phase button at the bottom of the Phase area Phase order can be changed by selecting a Phase and moving it up and down the list using the Move Phase buttons To delete a Phase click the Delete Phase button 9 When the Job is configured as required click File gt Save 2 2 2 Editing a Job 1 To edit a Job locate it
73. e of an attribute on the right hand side Note that the matching is case insensitive Ifthe Data Interface contains attributes that do not exist in a mapped data source that is read into a process these attributes are treated as having Null values in all processes For Data Out mappings the Data Interface is displayed on the left and vice versa for Data In mappings However this can be changed by clicking Swap Lists For mappings which can be either Data In or Out the default option is to have the Data Interface on the left 2 20 4 Deleting Data Interfaces 1 Right click on the Data Interface and select Delete The following dialog is displayed 2 68 Using Oracle Enterprise Data Quality Running Jobs Using Data Interfaces Dependencies Ei Mortgage Data affects 8 Profile Mortgage Data The Delete dialog shows which linked or dependent objects are affected by the deletion if any 2 Click Yes to proceed or No to cancel For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 21 Running Jobs Using Data Interfaces When a process including a Data Interface is run either as a standalone process or as part of a job the user must configure how the Data Interface reads or writes data Note Data Interfaces are used in readers or writers in processes Therefore the Mappings available during configuration will vary depending on how the Data Interface i
74. eal time matching In order to enable real time matching you must configure a process containing a match processor and use a single working input to the match processor only Real time processes may be run in Interval mode In Interval mode the processes run on a continuous basis and write their results to the results database at regular intervals See Execution Options A real time process can be run in Normal mode but in this case The process will only write results when the process is cancelled with the option to keep results ticked Note In many cases there is no requirement to write results since all results may be persisted externally for example in a log file If the process needs to be prepared this will always happen first meaning the process will not be able to respond to requests until this is complete Real time consumer and provider interfaces must also be configured to communicate with an external system using either JMS or Web Services Note For real time reference matching via a Web Service you can use EDQ s Web Service capability to generate the real time consumer and provider interfaces automatically If you are using the EDQ Customer Data Services Pack pre configured Web Services are provided 3 4 Parsing Concept Guide Why parsing is needed An important aspect of data being fit for purpose is the structure it is found in Often the structure itself is not suitable f
75. ed records use the other tick boxes on the same part of the screen Streaming Inputs Batch matching processes require a copy of the data in the EDQ repository in order to compare records efficiently As data may be transformed between the Reader and the match processor in a process and in order to preserve the capability to review match results if a snapshot used in a matching process is refreshed match processors always generate their own snapshots of data except from real time inputs to work from For large data sets this can take some time Where you want to use the latest source data in a matching process therefore it may be advisable to stream the snapshot rather than running it first and then feeding the data into a match processor which will generate its own internal snapshot effectively copying the data twice See Streaming a Snapshot above Cache Reference Data for Real Time Match processes It is possible to configure Match processors to cache Reference Data on the EDQ server which in some cases should speed up the Matching process You can enable caching of Reference Data in a real time match processor in the Advanced Options of the match processor For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 6 Publishing to the Dashboard EDQ can publish the results of Audit processors and the Parse processor to a web based application Dashboard so that data owners
76. ed to a Job in the same project When an External Task is configured to run as part of a job there is a single Enabled option Enabling or Disabling the Enable export option allows you to retain a job definition but to enable or disable the export of data temporarily 2 1 4 Exports When an Export is configured to run as part of a job the export may be enabled or disabled allowing you to retain a Job definition but to enable or disable the export of data temporarily and you can specify how you want to write data to the target Data Store from the following options Delete current data and insert default EDQ deletes all the current data in the target table or file and inserts the in scope data in the export For example if it is writing to an external database it will truncate the table and insert the data or if it is writing to a file it will recreate the file Append to current data EDQ does not delete any data from the target table or file but adds the in scope data in the export When appending to a UTF 16 file use the UTF 16LE or UTF 16 BE character set to prevent a byte order marker from being written at the start of the new data Replace records using primary key EDQ deletes any records in the target table that also exist in the in scope data for the export determined by matching primary keys and then inserts the in scope data How To 2 7 Creating and Managing Jobs Note When an Export is run
77. ed together and clusters containing a single record will not be used in matching This is illustrated below for a single cluster on a Name column using a Make Array from String transformation to cluster all values from the Name identifier that are separated by a space Advanced Features 3 5 Clustering Concept Guide Identifier Values Clusters Review Groups Simpson 4 e g Names Mr Mike Simpson 2 The clustering process is therefore crucial to the performance and functionality of the matching process If the clusters created are too large matching may have too many comparisons to make and run slowly If on the other hand the clusters are too small some possible matching records may have been missed by failing to cluster similar values identically Depending on the matching required clustering must be capable of producing common cluster keys for records that are slightly different For example it is desirable for records containing identifier values such as Oracle and Oracle Ltd to fall within the same cluster and therefore be compared against each other Multiple clusters EDQ supports the use of multiple clusters On a given data set clustering on a single identifier value is often insufficient as matches may exist where the identifier value is slightly different For instance a Surname cluster with no transformations will be unreliable if some of the data in the Surname field is known to be misspelt
78. edconcodantassascisaeteunseaseatns pioaveicummeeeieden 2 77 2 26 Publishing Result Views to Staged Data is sxncccesutsex eves wceninsasgienaetantaareateceredeact onan saceaeeaeesc 2 77 2 26 1 Published Res lts Tic ator cj cacsiccadscavpialeceiveriiacseaunydetansapasd tend R ES 2 78 3 Advanced Features 3 1 Marcina Concept Cle isaac par e ta Maan eins hein da eases aa emeaeaaiass 3 1 3 2 Clusterin Concept EdE aides neds Sitar a a O E E enti 3 5 3 3 Keak Time Matching serrr i a tia atone ate attee saatetha nto 3 10 3 4 Farsine C OmCe Pi Ude saraaa nsan E E T E NE une anaate 3 14 vi Audience Preface This document describes how to use Oracle Enterprise Data Quality This document is intended for Data Analysts or Data Stewards who are using Oracle Enterprise Data Quality Tip Oracle recommends that the reader read this guide in conjunction with the content in Enterprise Data Quality Online Help Documentation Accessibility For information about Oracle s commitment to accessibility visit the Oracle Accessibility Program website at http www oracle com pls topic lookup ctx acc amp 1id docacc Access to Oracle Support Oracle customers that have purchased support have access to electronic support through My Oracle Support For information visit http www oracle com pls topic lookup ctx acc amp id info or visit http www oracle com pls topic lookup ctx acc amp id trs if you are hearing impaired Related Documents For m
79. ent is added to the Event Log whenever a Job Task or System Task either starts or finishes Tasks are run either as part of Jobs or individually instigated using the Director UI The following types of Task are logged a Process Snapshot a Export Results Export External Task a File Download The following types of System Task are logged OFB a System Task meaning Optimize for Browse this optimizes written results for browsing in the Results Browser by indexing the data to enable sorting and filtering of the data The OFB task will normally run immediately after a Snapshot or Process task has run but may also be manually instigated using the EDQ client by right clicking on a set of Staged Data and selecting Enable Sort Filter or by a user attempting to sort or filter on a non optimized column and choosing to optimize it immediately DASHBOARD a System Task to publish results to the Dashboard This runs immediately after a Process task has been run with the Publish to Dashboard option checked 2 15 2 Server selection If the Director UI is connected to multiple servers you can switch servers using the Server drop down field in the top left hand corner If Server Console UI is connected to multiple servers select the required server in the tab list at the top of the window 2 15 3 Filtering events Quick filters Quick filter options are made available to filter by Event Type Status and Tas
80. epending on their type Staged Data Web Services and JMS can be either Data In or Data Out Reference Data and Snapshots can only be configured as Data In Data Interface Mappings wizard Once the data source or target is selected the Mapping area is displayed on the New Data Interface Mappings dialog How To 2 67 Managing Data Interfaces New Data Interface Mappings Specify the mapping to from the Data Interface Data Interface Mortgage Data Attributes Mortgage Min Max ID r gt Input Field Loan start date gt Total Number Loan end date E Minimum Length Mortgage type Minimum Length Count Term gt Minimum Length r Loan amount gt _ _ Mamun Length Loan amount currency Maximum Length Count Property value gt Maximum Length Property value currency gt Minimum Count Property value date gt Minimum Yo E Macamum Count lanai h YYVYVYVY VVVVVF 3 of 50 atinbutes have mappings This mapping can be used as either a data provider or a data consumer Notes sData can be mapped by Type or by Name Map by type maps the attributes on the left hand side to the lined up attribute on the right hand side as long as the Data Type String String Array Number Number Array Date or Date Array of both attributes is the same Map by name attempts to match the name of each attribute on the left hand side to the nam
81. er a snapshot or written staged data a given process a given job or all results for a given project can be purged For example it is possible to purge the results of old projects from the server while keeping the configuration of the project stored so that its processes can be run again in the future Note that the results for a project process job or set of staged data are automatically purged if that project process job or staged data set is deleted If required this means that project configurations can be packaged the package import tested and then How To 2 27 Purging Results deleted from the server The configurations can then be restored from the archive at a later date if required To purge results for a given snapshot set of written staged data process job or project 1 Right click on the object in the Project Browser The purge options are displayed 2 Select the appropriate Purge option If there is a lot of data to purge the task details may be visible in the Task Window Note sPurging data from Director will not purge the data in the Results window in Server Console Similarly purging data in Server Console will not affect the Director Results window Therefore if freeing up disc space it may be necessary to purge data from both The EDQ Purge Results commands only apply to jobs run without a run label that is those run from Director or from the Command Line using the runjobs command
82. eractively from the GUI by right clicking on an object in the Project Browser and selecting Run or as part of a scheduled Job The tasks have different execution options Click on the task below for more information Snapshots Processes a External Tasks such as File Downloads or External Executables Exports of data a Results Book Exports In addition when setting up a Job it is possible to set Triggers to run before or after Phase execution When setting up a Job tasks may be divided into several Phases in order to control the order of processing and to use conditional execution if you want to vary the execution of a job according to the success or failure of tasks within it 2 1 1 Snapshots When a Snapshot is configured to run as part of a job there is a single Enabled option which is set by default Disabling the option allows you to retain a job definition but to disable the refresh of the snapshot temporarily for example because the snapshot has already been run and you want to re run later tasks in the job only 2 1 2 Processes There are a variety of different options available when running a process either as part of a job or using the Quick Run option and the Process Execution Preferences Readers options for which records to process a Process options for how the process will write its results Run Modes options for real time processes a Writers options for how to write re
83. et for a real time response process that runs on a continuous basis starting a new interval every day How To 2 5 Execution Options Normal Prepare G Interval New interval after records New interval after days w Save all intervals Keep a maximum of intervals Keep intervals For a maximum of days w Browsing Results from processing in Interval mode When a process is running in Interval mode you can browse the results of the completed intervals as long as they are not too old according to the specified options for which intervals to keep The Results Browser presents a simple drop down selection box showing the start and end date and time of each interval By default the last completed interval is shown Select the interval and browse results Interval 2 processed 0 records View Results 13 Nov 2007 15 42 06 15 43 19 wa 13 Nov 2007 15 42 06 15 43 19 13 Nov 2007 15 40 54 15 41 33 m Noy 2007 15 39 50 15 40 54 Eh ia alan T 1 E r Bos no A eee inh er ECAP EE SA maT m te If you have the process open when a new set of results becomes available you will be notified in the status bar New Results Available You can then select these new results using the drop down selection box 2 1 2 4 Writers For each Writer in a process the following options are available Write Data This option sets whether or not the writer will run that is for writers that wri
84. f the regular running of the match process the reader for the decisions data should remain connected You can refresh the decisions data in the same way as any other data source in EDQ by rerunning the snapshot or by streaming the data directly into the matching process Note that if a decision has already been imported it will not be re imported even if it exists in the decisions data Imported decisions are treated in the same way as decisions made using EDQ and are therefore visible in Case Management or in the Match Review application and will also affect the Review Status summary of the match processor For more information see Enterprise Data Quality Online Help 2 18 Exporting Match Decisions All relationship review activity conducted in EDQ manual match decisions and review comments can be written out of the match processor on which the review was based This may be useful for any of the following reasons To store a complete audit trail of review activity lo export match decisions for import into another process for example a new match process working on the same data To enable data analysis on review activity Manual match decisions and review comments are written out of a match process on the Decisions output filter from each match processor The data can be seen directly in the Data View of the match processor and can be connected up to any downstream processor such as a Writer to stage the data N
85. fications Where a snapshot task has a FINISHED status the number of records snapshotted is displayed Details of any warnings and errors encountered during processing are included Process Tasks The notification displays the status of the process task in the execution of the job The possible statuses are a FINISHED the process ran to completion CANCELLED the job was canceled by a user during the process task WARNING the process ran to completion but one or more warnings were generated a ERROR the process failed to complete due to an error Record counts are included for each Reader and Writer in a process task as a check that the process ran with the correct number of records Details of any warnings and errors encountered during processing are included Note that this may include warnings or errors generated by a Generate Warning processor Export Tasks The notification displays the status of the export task in the execution of the job The possible statuses are a STREAMED the export was optimized for performance by running the data directly out of a process and writing it to the data target FINISHED the export ran to completion as an independent task CANCELLED the job was canceled by a user during the export task a ERROR the export failed to complete due to an error Where an export task has a FINISHED status the number of records exported is displayed Details of any errors encountered
86. from the results of a Frequency profiler select a results tab Select the desired values right click and Create Reference Data Getting Started 1 3 Adding an Issue J i d Rey Create Issue d Prof 1 0 1 Col i lt 0 1 j PEE Pe E T a Note You can use the normal windows Shift select and Control select options to select data Take care not to drill down when attempting to select the desired values Alternatively to select all the loaded data in the results browser for a given column Control select the column at the top for example the Value column in the screenshot above Press Escape to de select all selected data 3 If you are adding new Reference Data rather than a new Lookup onto existing Reference Data define the columns that you require in the Reference Data For example for a simple list of values define a single column If you would like the Reference Data Editor to add a uniqueness constraint so that duplicate entries cannot be created select the Unique option on the column 4 Select the column or columns that you want to use when performing lookups on the data 5 Select the column or columns that you want to use when returning values from a lookup 6 Optionally select the Category of Reference Data that you want to create if you are creating Reference Data of a specific type such as a list of regular expressions 7 Give the Reference Data a Name for example Valid Titles
87. g see the Clustering Concept Guide Comparisons Comparisons are replaceable algorithms that compare identifier values with each other and deliver comparison results The nature of result delivered is dependent on the comparison For example a comparison result could be simply True match or False no match or may be a percentage value indicating match strength Match rules Match rules offer a way to interpret comparison results according to their business significance Any number of ordered rules can be configured in order to interpret the comparison results Each rule can result in one of three decisions a Match a No match Advanced Features 3 3 Matching Concept Guide a Review manual review required to confirm or deny the match The use of match rules form a rule table across all comparisons to determine the decision of the match operation for example O No Match O Close Match C Equivalent No Data Using pre constructed match processes EDQ is designed to allow you to construct new matching processes quickly and easily rather than depending on pre configured matching processes that are not optimized for your data and specific matching requirements and which will be more difficult to modify However in some cases a matching template can be used to learn how matching in EDQ works and in order to provide very fast initial results to give an indication of the level of duplication in your data Configur
88. h one entry per row Each column has a fixed width specified in characters which determines the maximum amount of data it can contain No delimiters are used to separate the fields in the file Instead smaller quantities of data are padded with spaces to fill the allotted space such that the start of a given column can always be specified as an offset from the beginning of a line The following file snippet illustrates characteristics common to many flat files It contains information about cars and their owners but there are no headings to the columns in the file and no information about the meaning of the data In addition the data has been laid out with a single space between each column for readability cars txt Notepad File Edit Format eQ060601 BMw Jones 45 500 20050601 BMw o002 2003 Chau 5 200 2O060102 Lexus o003 2006 smith 29 565 20070930 Mazda o004 2007 Mukherjee 35 000 20090909 Toyota 0005 2009 Barker 400 View Help i ee CO see aai ai In order to parse the data in a fixed width text file correctly EDQ needs to be informed of the column sizes implicit in that file This is done in the New Data Store wizard and can be edited as part of the data store settings later if required When you first enter the data store configuration screen for a fixed width text file the columns table is empty In the following screenshot it has been populated with the mapping information for some of the columns in ou
89. he Reference Data used by a selected Process or Items using selected item to show objects that use the selected object such as any Jobs that use a selected Process Whenever a filter has been applied to the Project Browser a box is shown just above the Task Window to indicate that a filter is active For example the below screenshot shows an indicator that a server that has been filtered to show only the objects used by the Parsing Example process localhost Show items used by Parsing Example 4 You can then package the visible objects by right clicking on the server and selecting Package This will only package the visible objects To clear the filter click on the x on the indicator box In some cases you may want to specifically exclude some objects from a filtered view before packaging For example you may have created a process reading data from a data interface with a mapping to a snapshot containing some sample data When you package up the process for reuse on other sets of data you want to publish the process and its data interface but exclude the snapshot and the data store To exclude the snapshot and the data store from the filter right click on the snapshot and select Exclude From Filter The data store will also be excluded as its relationship to the process is via the snapshot As packaging always packages the visible objects only the snapshot and the data store will not be included in the package H
90. he Job status before attempting to change it Snapshot and Export Tasks in Jobs must use server side data stores not client side data stores 2 2 1 Creating a Job 1 2 Expand the required project in the Project Browser Right click the Jobs node of the project and select New Job The New Job dialog is displayed 2 8 Using Oracle Enterprise Data Quality Creating and Managing Jobs 3 Enter a Name and if required Description then click Finish The Job is created and displayed in the Job Canvas Example New Job EJ OC eTA GIKA Mew Phase Pre na cee Cle a Address Clean Ja J Fa a Dae 100K Entity Clean a2 E bai Entity Match Batch Ja d Address Standardize Entity Cluster Real time Entity Match Real time a Fntitv Name and Address Clean Search 4 Right click New Phase in the Phase list and select Configure 5 Enter a name for the phase and select other options as required Field Type Enabled Checkbox Description To enable or disable the Phase Default state is checked enabled Note The status of a Phase can be overridden by a Run Profile or with the runopsjob command on the EDQ Command Line Interface Execution Condition Drop down list Clear Error Checkbox To make the execution of the Phase conditional on the success or failure of previous Phases The options are a Execute on failure the phase will only execut
91. he latest version of the attribute will be used as the input This is especially important when transformation processors have been used Notes sFor Profiling processors it is common to analyze the data in all attributes to discover issues of interest about the data a Once a processor is correctly configured it no longer appears with a blue background Getting Started 1 7 Adding Processors 7 Once you have connected the set of processors that you want to use click on the Quick Run process button on the Toolbar to run the process and look at the results O 6 AKO S u Profile Customers Table Data Types Proiler inai smen dih E E a 7 a a ie et ab o 8 The Canvas background changes to blue to show you that the process is running Also the process icon in the Project Browser turns green so that other users connected to the same host can see that it is running Note The process is locked and cannot be edited while it is running 9 When the process has finished the processors no longer appear with a shaded background and you can browse on the results for each processor by clicking on the processor on the Canvas and viewing its results in the Results Browser Data Types Profler Ca 7 bE L8H OOA Input Field Total Number Text Format Numeric Format Date time Format Null Values Consistency CU_ACCOUNT 2010 2003 0 0 i 99 7
92. he right is enabled 4 If required enter a custom name for the option in this field 5 Click OK to save changes or Cancel to abandon 2 19 4 1 Snapshot Externalization Dialog The fields on the right of all Externalization dialogs contain the default name used for each externalization attribute This is the label used to reference the externalized attribute in any third party tool or file To change any field name check the box to the left and edit the field as required Note Avoid the use of spaces in attribute names The fields in this dialog vary depending on whether the Data Store is a file or a database How To 2 61 Externalizing Configuration Settings Delimited text file example table_name sample size sample percentage sample_otfset ascending File in server work area E ver_work_area Use project specific landing area fic_landing_area Treat first row as header ow _as header Field separator field_separator Quote character quote _character Number of columns to assume Imns to assume Encoding encoding Skip lines at start p lines _at start 2 62 Using Oracle Enterprise Data Quality Externalizing Configuration Settings Access database example Externalization of Snapshot Snapshot Table name E table_name Sample size E sample_size Sample percentage sample_percentage Sample offset E sample_offset Ascending E ascending WHERE dause SOL F where
93. he same name if they do exist See also Running a Prepared Export 1 11 Running a Prepared Export There are two ways of running a prepared export of a Staged Data table or of a Results Book 1 Manually run the export 2 Run the export as part of a job 1 11 1 Manually running an export Right click on the named Export configuration in the Project Browser and select Run OLH E A Data Stores H G Staged Data H t3 Processes H E Reference Data aS Results Books a External Tasks H Jobs a f Exports E Export Screening results a z 7 Profiled Customer a A E Getting Started 1 17 Running a Prepared Export The export process will run immediately and you can monitor its progress in the Project Browser Note If you are exporting to a Client side Data Store the Export must be run manually 1 11 2 Running an export as part of a job If you are exporting to a server side Data Store you can run the export as part of a larger job and schedule that job for execution at some point in the future 1 2 Right Click on the Job Configuration in the Project Browser and select New Job This brings up the Job Configuration dialog Change the default name of the job to a descriptive name Note that the Tool Palette now changes to show a list of tasks that may be run in jobs Drag and drop the Export Task from the Tool Palette to add it to the job The job may also include other
94. hip is imported it will be treated as the latest decision If a decision with exactly the same date time as a manual decision is imported the manual decision is given precedence over the imported decision The remainder of this help topic provides a step by step guide to importing match decisions 2 17 1 Step 1 Connect the decisions data into the match processor To import match decisions use a Reader to read in the imported decision data and connect the data into the Decisions input port to the match processor How To 2 51 Importing Match Decisions Working Data B gt gt Groups i lt Erence Data 4 gt Relationships i hpr DE De cisions D gt D gt Ive ged Matgi Individuals Si ions ae Ah megaman engpchtememntiom nA Mita ten a E Note As Group and Merge does not allow manual decisions on possible matches it does not have a Decisions input port This data must include all the identifying information for the relationship required to match it to the corresponding relationship in the EDQ matching process that is attributes for each attribute that is included in the configurable Decision Key for the match processor for both records in the relationship and fields for the decision review user and review date time It is also possible to import comments made when reviewing possible matches 2 17 2 Step 2 Specifying the attributes that hold the new decision data When a set of dec
95. iguration Ciidakaicars no newlines bt Encoding 150 6859 1 han mileage at purchase Fixed width records Record size z 1 8 Connecting to a Data Store To connect to a new Data Store in order to process a set of data 1 Right click on Data Stores within your project in the Project Browser and select New Data Store 1 12 Using Oracle Enterprise Data Quality Connecting to a Data Store f Dynamic Configuration Demo TE celina tent E a I atlt c _ l 10 11 E E Data Stores St 3 New Data i AQ vz Select the category of data store that you want to connect to Database Text file XML file MS Office file or Other if you want to specify connection details using JDBC or ODBC Select where the data will be accessed from the server or the client Select the type of data store that you want to connect to for example for databases select the type of database for example Oracle SQL Server MS Access etc Specify the connection details to the data For example a For a client side Access database browse to the mdb file on the local file system For a client side Text file browse to the directory that contains the text file on the local file system For fixed width text files you must also define the fields that are present in the file a Fora server side file Access Text Excel or XML enter the name of
96. in Interval mode The supported execution types of each processor are listed in the help page for the processor a They either write no results or write their results to a staged data table and or external database or file The process completes when it is stopped or when a configured time or record threshold is reached They may be executed either in Normal mode for a limited period of time or processing a limited number of records or in Interval mode For more information see Enterprise Data Quality Online Help 2 24 Customizing Processor Icons It is possible to customize the icon for any processor instance in EDQ This is one way of distinguishing a configured processor which may have a very specific purpose from its generic underlying processor For example a Lookup Check processor may be checking data against a specific set of purchased or freely available reference data and it may be useful to indicate that reference data graphically in a process The customization of a processor icon is also useful when creating and publishing new processors When a processor has been published its customized icons becomes the default icon when using the processor from the Tool Palette To customize a processor icon 1 Double click on a processor on the Canvas 2 Select the Icon amp Family tab 3 To change the processor icon which appears at the top right of the image use the left side of the screen 4 To change
97. ing people group records using the first 5 letters of a Surname 3 8 Using Oracle Enterprise Data Quality Clustering Concept Guide attribute converted to upper case that yields fairly large clusters with hundreds of records in many of the groups After the match process has been further developed and perhaps applied to the full data sets rather than samples it is possible to improve performance by making the clustering configuration more sensitive for example by grouping records using the first 5 letters of a Surname attribute followed by the first 4 characters of a Postcode attribute This will have the effect of making the clusters smaller and reducing the total number of comparisons that need to be performed When matching on a number of identifiers where some of the key identifiers contain blanks and nulls It is generally better to use multiple clusters rather than a single cluster with large groups Note All No Data whitespace characters are always stripped from cluster keys after all user specified clustering transformations are applied and before the clustering engine finishes For example if you use the Make Array from String transformation and split data on spaces the values Jim lt space gt lt carriage return gt Jones and Jim lt space gt Jones would both create the cluster values Jim and Jones The former would not create the cluster value lt carriage return gt Jones This is in order that the user d
98. ing with large package files arneu tonais n a A ENE 2 27 Copying DELW CEI SOL Y CLS sees sessesenis iiei i Ea ET 2 27 Purgine Cereb dolar rene Oncor Renee etry er Reger ere eee een eee Teer eae 2 27 Purging Match Decision ater sas ceca cies asec sash aa A ante cle tonnsitectas NE 2 28 Creatine ROC CSS OLS ase ssp aca nels wots tues van Mcgee A inane aaa caus 2 29 Creating a processor from a sequence of configured processors ceeeeeeeeeeeeees 2 29 petine MPS nesepi i 2 30 Seting OPHONS angran a E E sae eats heeanane etetesial 2 31 SSCS UCU EAU ULES anisini a a n a 2 32 DELS Resule VIEW Surenin E a day nains 2 33 DEIN OUMU FIETS iaa a n 2 34 Setting Dashboard Publication Options sssesssessseessseessssessrsresesresesersresrssrseeseseeseseeses 2 34 Seting a C ustom Ionien e a R 2 35 Publishing FrOCESSOF Seoni i a a E A EAN 2 35 Editing a Published T roceSsOr rases A OE 2 36 Attaching Help to published Processors eisini r aon veos ten tervioouens 2 37 Publishing processors into Lames sccsssuccssscsnsy otanassdcy ohenaerensediiawdnavs ehangatgunstgodsasectauessineaas 2 37 Using Published PLOGeSSOr Success toad asirerse aceon os ei aera eae eres 2 38 FPOTIMISSLOING a O sa pantscents vanicheeon E 2 39 Unlockine 4 Keterence Published Processoromaineiionia e 2 39 IVES noate d PROCESS toe sisal a O E Wenn E N 2 39 TVA el Erocessor ocardiosrasairi ai o tude sete A E E EA 2 39 PUCA LMDUtE SEAN CI raira a A A E tastes 2 40 Clear cearc h Hie DH
99. ing with smaller data sets but will disable sorting and filtering for large data sets However you can choose to override these settings for example to achieve maximum throughput when working with a number of small data sets Snapshot Sort Filter options When a snapshot is created the default setting is to Use intelligent Sort Filtering options so that the system will decide whether or not to enable sorting and filtering based on the size of the snapshot For more information see Adding a Snapshot However If you know that no users will need to sort or filter results that are based on a snapshot in the Results Browser or if you only want to enable sorting or filtering at the point when the user needs to do it you can disable sorting and filtering on the snapshot when adding or editing it How To 2 19 Optimizing Job Performance To do this edit the snapshot and on the third screen Column Selection uncheck the option to Use intelligent Sort Filtering and leave all columns unchecked in the Sort Filter column Alternatively if you know that sorting and filtering will only be needed on a sub selection of the available columns use the tick boxes to select the relevant columns Disabling sorting and filtering means that the total processing time of the snapshot will be less as the additional task to enable sorting and filtering will be skipped Note that if a user attempts to sort or filter results based on a colum
100. ion Note that if the processor outputs multiple top level Results Views as in the Frequency Profiler example above the Publish Results dialog shows multiple tabs one for each view You can choose to publish any or all of the processor views 2 26 1 Published Results Indicator Processors in a process that are configured to write one or more Results Views to Staged Data and where publishing is currently enabled are indicated using an overlay icon on the Canvas as shown below Quicks ats Profiler 2 78 Using Oracle Enterprise Data Quality Publishing Result Views to Staged Data Note If the Staged Data set that a Results View is being written to is deleted or renamed the indicator turns red to indicate an error This is treated in the same way as if the processor s configuration is in error that is the process cannot run Staged Results Views and the Server Console UI By default all data that is snapshotted or staged during the execution of a job in the Server Console UI or using the runopsjob command from the Command Line Interface is available for view in the Server Console Results Browser by users with the appropriate permissions This includes any Results Views that you choose to stage in processes that are run in the Job However it is possible to override the visibility of a specific Staged Data set in the Server Console UI using an override setting in a Run Profile For more information see Enterp
101. isions data is connected in to a match processor as above an additional sub processor appears when the match processor is opened Match Individuals Input Identify Cluster a a i a E N a Double click on the Decisions sub processor to map the decisions data into the matching process The first tab allows you to set the fields in the connected decisions data with the actual decision and comment values that you want to import along with details such as user name and date time The following screenshot shows an example of a correct configuration for importing match decision data 2 52 Using Oracle Enterprise Data Quality Importing Match Decisions i Decisions General Settings Mapped Fields Decision Field Decision User Field Decision Timestamp Field Comment Field Comment User Field Comment Timestamp Field Data Stream Mame Related Data Stream Name Match Rule Mame State Expiry Timestamp Match Status Match A ManualDecision we A Review User we A ReviewDateTime t CommentUser A CommentDateTime Datastream RelatedDataStream A MakchRule we A ExpiryDate a Field Value wt wt wt wt NoMatch homad Fossible Match Possible Match Penig Pening Dataset Customer Data Auto Map Match Mame Customer Data The first section in the screenshot above shows the attributes in the imported decision data that will be used to capture the decision data Note that all the
102. isons between records that are not likely to match are not performed Finding the right balance may involve some trial and error for example assessment of the difference in match statistics when clusters are widened perhaps by using fewer characters of an identifier in the cluster key or narrowed perhaps by using more characters of an identifier in a cluster key or when a cluster is added or removed The following two general guidelines may be useful If you are working with data with a large number of well populated identifiers such as customer data with address and other contact details such as email addresses and phone numbers you should aim for clusters with a maximum size of 20 for every million records and counter sparseness in some identifiers by using multiple clusters rather than widening a single cluster If you are working with data with a small number of identifiers for example where you can only match individuals or entities based on name and approximate location wider clusters may be inevitable In this case you should aim to standardize enhance and correct the input data in the identifiers you do have as much as possible so that your clusters can be tight using the data available For large volumes of data a small number of clusters may be significantly larger For example Oracle Watchlist Screening uses a cluster comparison limit of 7m for some of the clustering methods used when screening names against Sanction
103. k Type Simply select the values that you want to include in the filter using Control Select to select multiple items and click on the Run Filter button on the bottom left of the screen to filter the events For example 2 46 Using Oracle Enterprise Data Quality Using the Event Log Chri hoeener g Aap OL Doers la Auge aL Customers 13 Aug 30 13 Aug 20 oe JARMA n Apl no 13 Aug 21 i3 Aug 2i 12 2 Address Sea 12 Aug 21 Pos 12 Aug 2011 12 Aug 30 Address Pos lo fug 20L Address Sem 12 Aug 20 L2Auge 201 Address Clean 15 Jul 2013 15 Jul 3013 i2 hWi M13 Person and 12 hi M 13 Indeadual d 12 Jul 2003 lt Geocoding 5 12 Jul 2015 Address Yeri 12 2013 1000 records 1000 records 1000 records 12 du a3 12 2013 13 3 2013 as I i 12 Wi 2013 OFW 1F 2 E 05 3 2013 5 1 2013 05 2017 a 0J hi Mi3 E Free text filters search ahead Further free text filtering options are available to filter by Project Name Job Name Task Name and User Name These are free text so that you can enter partial names into the fields You can enter a partial name into any of these fields provided the object contains the partial name it will be displayed though note that matching is case sensitive For example if you use a naming convention where all projects working on live systems have a name including the word
104. lean Triggers can be set before or after a Phase A Before Trigger is indicated by a blue arrow above the Phase name and an After Trigger is indicated by a red arrow below it For example the following image shows a Phase with Before and After Triggers i b Clean Customers TF Triggers can also be specified as Blocking Triggers A Blocking Trigger prevents the subsequent Trigger or Phase beginning until the task it triggers is complete 2 3 1 Configuring Triggers 1 Right click the required Phase and select Configure The Phase Configuration dialog is displayed 2 Inthe Triggers area click the Add Trigger button under the Before Phase or After Phase list as required The Select Trigger dialog is displayed 2 12 Using Oracle Enterprise Data Quality Job Notifications a2 JN 0 um e p Select Tigger oO Externalized job options mapping oO Snapshot Clean and Match Customers Select the Trigger type in the drop down field Select the specific Trigger in the list area Click OK If required select the Blocking checkbox next to the Trigger Set further Triggers as required When all the Triggers have been set click OK 2 3 2 Deleting a Trigger from a Job 1 2 Right click the required Phase and select Configure In the Phase Configuration dialog find the Trigger selected for deletion and click it Click the Delete Trigger button under the list of the selected Trigger The Trigger is deleted
105. lette and all processors that have been published and which use that family icon will appear in that family The family will have the same name as the name given to the family icon For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help How To 2 37 Using Published Processors 2 11 Using Published Processors Published processors are located in either the individual project or system level as shown below Project Browser H localhost dnadmin Er sy Projects l 5 0 Published Processor Summarize Address Verification i Staged Data G0 Data Interfaces H t Processes Titl od O Jobs aa Exports Gessa E Notes H Test 1 There are three types of Published processor Template These processors can be reconfigured as required m Reference These processors inherit their configuration from the original processor They can only be reconfigured by users with the appropriate permission to do so They are identified by a green box in the top left corner of the processor icon a Locked Reference These processors also inherit the configuration from the original processor but unlike standard Reference processors this link cannot be removed They are identified by a red box in the top left corner of the processor icon These processors can be used in the same way as any other processor either added from the Project Browser or Tool Palette by clicking and draggi
106. mple or filter records in the Reader in a process all processors will re execute regardless of the Intelligent Execution setting as the process will be running on a different set of records Enable Sort Filter in Match processors This option means that the specified Sort Filter enablement settings on any match processors in the process accessed via the Advanced Options on each match processor will be performed as part of the process execution The option is selected by default When matching large volumes of data running the Sort Filter enablement task to allow match results to be reviewed may take a long time so you may want to defer it by de selecting this option For example if you are exporting matching results externally you may want to begin exporting the data as soon as the matching process is finished rather than waiting until the Enable Sort Filter process has run You may even want to over ride the setting altogether if you know that the results of the matching process will not need to be reviewed Results Drill Down This option allows you to choose the level of Results Drill Down that you require All means that drilldowns will be available for all records that are read in to the process This is only recommended when you are processing small volumes of data up to a few thousand records when you want to ensure that you can find and check the processing of any of the records read into the process Sample is the
107. mum performance is required from a Parse processor it should be run in Parse mode rather than Parse and Profile mode This is particularly true for any Parse processors with a complete configuration where you do not need to investigate the classified and unclassified tokens in the parsing output The mode of a parser is set in its Advanced Options For even better performance where only metrics and data output are required from a Parse processor the process that includes the parser may be run with no drilldowns see Minimized results writing above When designing a Parse configuration iteratively where fast drilldowns are required it is generally best to work with small volumes of data When working with large volumes of data an Oracle results repository will greatly improve drilldown performance 2 20 Using Oracle Enterprise Data Quality Optimizing Job Performance 2 5 2 2 Matching performance options The following techniques may be used to maximize matching performance Optimized Clustering Matching performance may vary greatly depending on the configuration of the match processor which in turn depends on the characteristics of the data involved in the matching process The most important aspect of configuration to get right is the configuration of clustering in a match processor In general there is a balance to be struck between ensuring that as many potential matches as possible are found and ensuring that redundant compar
108. n Sampling Options Filter Options _ Oall Count 1000 percentage Sampling Offset Sampling Order Descending Ascending The Sampling Options fields are as follows All Sample all records Count Sample n records This will either be the first n records or last n records depending on the Sampling Order selected Percentage Sample n of the total number of records Sampling Offset The number of records after which the sampling should be performed Sampling Order Descending from first record or Ascending from last How To 2 3 Execution Options Note Ifa Sampling Offset of for example 1800 is specified for a record set of 2000 only 200 records can be sampled regardless of the values specified in the Count or Percentage fields 2 1 2 2 Process The following options are available when running a process either as part of the Process Execution Preferences or when running the process as part of a job Use Intelligent Execution Intelligent Execution means that any processors in the process which have up to date results based on the current configuration of the process will not re generate their results Processors that do not have up to date results are marked with the rerun marker For more information see the Processor States topic in Enterprise Data Quality Online Help Intelligent Execution is selected by default Note that if you choose to sa
109. n that has not been enabled the user will be presented with an option to enable it at that point Staged Data Sort Filter options When staged data is written by a process the server does not enable sorting or filtering of the data by default The default setting is therefore maximized for performance If you need to enable sorting or filtering on written staged data for example because the written staged data is being read by another process which requires interactive data drilldowns you can enable this by editing the staged data definition either to apply intelligent sort filtering options varying whether or not to enable sorting and filtering based on the size of the staged data table or to enable it on selected columns by selecting the corresponding Sort Filter checkboxes Match Processor Sort Filter options It is possible to set sort filter enablement options for the outputs of matching See Matching performance options 2 5 2 Processor specific Performance Options In the case of Parsing and Matching a large amount of work is performed by an individual processor as each processor has many stages of processing In these cases options are available to optimize performance at the processor level Click on the headings below for more information on how to maximize performance when parsing or matching data Parsing performance options Matching performance options 2 5 2 1 Parsing performance options When maxi
110. nabled data store it may be that the EDQ client does not have all the fonts installed to view the data correctly on screen though note that the data will still be processed correctly by the EDQ server In this case it is useful to inspect the characters by right clicking on a character or a string containing an unusual character and selecting the Show Characters option For example the below screenshot shows the Character Profiler processor working on some Unicode data with a multi byte character selected where the client does not have the required font installed to display the character correctly The character therefore appears as two control characters How To 2 43 Using the Results Browser CdS 7 28 B8HH COG m Record Count g FLIS XSUe0 tay E 949 x3b5 186 151 z 12471 x30b7 184 176 1 12455 x30a7 183 171 e 12524 x30ec 182 172 Ei 26031 x65af 181 162 250 xfa 178 137 A 955 176 141 z ieee Bs 176 115 Al 49334 xc2de 174 160 ono 3071 xbed 174 102 g 1508 x5e4 173 147 W 1064 x428 172 150 5 351 x15f 166 110 a 940 x3ac 165 156 ay 12496 x30d0 164 147 etl A BR tas ae a R H Mec iik If you right click on the character and use the Show Characters option EDQ can tell you the character range of the character in the Unicode specification In the case above the character is in the Tamil range W Show Characters Char OO Dec 30271 Hex xO0bed Display oo
111. nce Data to be made available either in the project or at system level Options that are not externalized on the published processor can either have their Reference Data supplied with the published processor the default behavior with this option selected or can still require Reference Data to be made available For example to use a standard system level Reference Data set 2 10 1 Editing a Published Processor Published processors can be edited in the same way as a normal processor although they must be republished once any changes have been made 2 36 Using Oracle Enterprise Data Quality Publishing Processors If a Template Published processor is edited and published only subsequent instances of that processor will be affected as there is no actual link between the original and any instances If a Reference or Locked Reference Published processor is reconfigured all instances of the process will be modified accordingly However if an instance of the processor is in use when the original is republished the following dialog is displayed 2 10 2 Attaching Help to published processors It is possible to attach Online Help before publishing a processor so that users of it can understand what the processor is intended to do The Online Help must be attached as a zip file containing an file named index htm or index html which will act as the main help page for the published processors Other html pages as well as image
112. nformation see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 14 Using the Results Browser The EDQ Results Browser is designed to be easy to use In general you can click on any processor in an EDQ process to display its summary view or views in the Results Browser The Results Browser has various straight forward options available as buttons at the top just hover over the button to see what it does However there are a few additional features of the Results Browser that are less immediately obvious a Show Results in New Window a Show Characters a Select Column Headers 2 14 1 Show Results in New Window It is often useful to open a new window with the results from a given processor so that you can refer to the results even if you change which processor you are looking at in EDQ For example you might want to compare two sets of results by opening two Results Browser windows and viewing them side by side To open anew Results Browser right click on a processor in a process and select Show results in new window The same option exists for viewing Staged Data either snapshots or data written from processes from the right click menu in the Project Browser 2 14 2 Show Characters On occasion you might see unusual characters in the Results Browser or you might encounter very long fields that are difficult to see in their entirety For example if you are processing data from a Unicode e
113. ng it to the Project Canvas For further information see the Published Processors topic in Enterprise Data Quality Online Help 2 38 Using Oracle Enterprise Data Quality Investigating a Process 2 11 1 Permissions The creation and use of published processors are controlled by the following permissions Published Processor Add The user can publish a processor Published Processor Modify This permission in combination with the Published Processor Add permission allows the user to overwrite an existing published processor Published Processor Delete The user can delete a published processor Remove Link to Reference Processor The user can unlock a Reference published processor See the following section for further details 2 11 2 Unlocking a Reference Published Processor If a user has the Remove Link to Reference Processor permission they can unlock a Reference Published processor To do this use the following procedure 1 Right click on the processor 2 Select Remove link to Reference Processor The following dialog is displayed Are you sure 3 Click OK to confirm The instance of the processor is now disconnected from the Reference processor and therefore will not be updated when the original processor is For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 12 Investigating a Process During the creation of complex
114. ns at the process level a Externalized processor options are configured at the process level and therefore are available for all other processors within the process Externalized options within a process must have unique names However it is possible to use the same name for an externalized option in other processes How To 2 57 Externalizing Configuration Settings The second stage is optional but Oracle recommends that meaningful names be set to ensure that externalized options can be easily identified 2 19 1 1 Select processor options to externalize 1 Inthe Canvas area double click on the required processor The Processor Configuration dialog is displayed 2 Select the Options tab 3 Click the Show Externalized Options button on the bottom right of the dialog An Externalize button is displayed next to each option Failure Lookup Check Lookup Check Minimum number of matches 1 Unlimited maximum matches Yes No Maximum number of matches 1 Lookup reference data Lookup Record Id From Rule Id ass gt x Hide externalized options gt gt OK Cancel 4 For each option that needs to be externalized 1 Click the Externalize button The Externalize dialog is displayed 2 Check the box in the dialog A default label is assigned 3 Edit the label if required or if available select another from the drop down list 4 Click OK 5 When externalized
115. oes not always have to consider how to cope with different forms of whitespace in the data when clustering Reviewing the clustering process In EDQ the clusters used in matching can be reviewed in order to ensure they are created to the required level of granularity This is possible using the views of clusters created when the match processor has been run with an clustering configuration The Results Browser displays a view of the clusters generated with their cluster keys Gat TEE BBE OG i Cluster Group size CustDB Customers KLREECL 12 12 PLEC4 10 10 KLREECS g PRNEC1 KLEC4 a TFSEC4 7 PNEC1 7 HSEC4 Fi KMRNEC 4 Fi KLEC1 7 ATMSEC4 7 7 7 6 6 5 6 D wo ANTREC4 ANTREC1 PERECI ATRTEC2 ATRTECI ATMSEC1 RSEC1 ee ee ee ee T mn mh mW mH ww wl d d d Md oS oo A Input CustDB Customers Clusters a Ee j kan ll rst 4 surname meta first 3 postcode Sm mhe The list of clusters can be sorted by the cluster key value in order to see similar groups that possibly ought not be distinct Advanced Features 3 9 Real Time Matching By drilling down it is possible to see the constituent records within each cluster from each input data set For example the 9 records from the Customers data set with a cluster key of KLRKEC3 above are O uHe 7 BE HE ooo Beale Family Parse Title_ Parse Given Parse Middle Parse Address Address Clark Mrs Helen MB Church Ves
116. oftware pursuant to the applicable Federal Acquisition Regulation and agency specific supplemental regulations As such use duplication disclosure modification and adaptation of the programs including any operating system integrated software any programs installed on the hardware and or documentation shall be subject to license terms and license restrictions applicable to the programs No other rights are granted to the U S Government This software or hardware is developed for general use in a variety of information management applications It is not developed or intended for use in any inherently dangerous applications including applications that may create a risk of personal injury If you use this software or hardware in dangerous applications then you shall be responsible to take all appropriate fail safe backup redundancy and other measures to ensure its safe use Oracle Corporation and its affiliates disclaim any liability for any damages caused by use of this software or hardware in dangerous applications Oracle and Java are registered trademarks of Oracle and or its affiliates Other names may be trademarks of their respective owners Intel and Intel Xeon are trademarks or registered trademarks of Intel Corporation All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International Inc AMD Opteron the AMD logo and the AMD Opteron logo are trademarks or registered trademarks of Adv
117. olumn header in the Results Browser will sort the data by that column However if you control click on the column header hold down the Ctrl key and click on the header you can select all the visible loaded data in that column in the Results Browser This is useful for example to copy all loaded rows or to use them to create or add to reference data using the right click option Note that the Results Browser only loads 100 records by default so you may want to use the Load All Data button before selecting the column header Multiple column headers can be selected in the same way For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 15 Using the Event Log The Event Log provides a complete history of all jobs and tasks that have run on an EDQ server By default the most recent completed events of all types are shown in the log However you can filter the events using a number of criteria to display the events that you want to see It is also possible to tailor the Event Log by changing the columns that are displayed in the top level view Double clicking on an event will display further information where it is available How To 2 45 Using the Event Log The displayed view of events by any column can be sorted as required However older events are not displayed by default so a filter must be applied before sorting before they can be viewed 2 15 1 Logged Events An ev
118. on dynamic That is they are updated on a regular basis but not constantly accessed and updated by multiple users for example watch lists rather than CRM data Reference matching is a single stage process Incoming records from a working real time source are matched against snapshots of reference records from one or more staged sources A writer on the process then returns the output of the match processor back to the calling system Note that the output returned may be any of the forms of output from the match processor If you want only to link the new records to the reference sets you can write back the Relationships output If you want to enhance the new records by merging in data from the matching reference records you can use the merge rules in a match processor and write back the Merged or Enhanced data output Note The Reference Data in a real time matching process can be cached and interrogated in memory on the EDQ server if required See the Cache reference records for real time processes topic in Enterprise Data Quality Online Help for further information Advanced Features 3 11 Real Time Matching In the following example the links to reference records are written Real time reference matching g Working Data Real tirne s gt Groups 1 Working Data E g E Relationships Fe Erence Data gt gt Enhanced B Indexed o Wihte links to ee P E E ee A Reference Refren
119. options see the Advanced options for match processors topic in Enterprise Data Quality Online Help 3 3 Real Time Matching EDQ provides two different types of real time matching Real time duplication prevention for matching against dynamically changing data for example working data in an application Real time reference matching for matching against slowly changing data for example reference lists This topic provides a general guide to how real time matching works in EDQ Note If you want to use EDQ for real time duplicate prevention and or data cleansing with Oracle Siebel UCM or CRM Oracle provides a standard connector for this purpose Real time duplicate prevention EDQ s real time duplicate prevention capability assumes that the data being matched is dynamic and therefore changes regularly for example customer data in a 3 10 Using Oracle Enterprise Data Quality Real Time Matching widely used CRM system For this reason EDQ does not copy the working data Instead the data is indexed on the source system using key values generated using an EDQ cluster generation process Real time duplicate prevention occurs in two stages Clustering and Matching Clustering In the Clustering stage EDQ receives a new record on its real time interface and assigns it cluster keys using a process that has been enabled for real time execution It returns the cluster keys for the record using a Writer The
120. or However not all the available outputs may be needed in your process For example you should disable Merged Output if you only want to identify sets of matching records Note that disabling any of the outputs will not affect the ability of users to review the results of a match processor To disable Match or Alert Groups output 1 Open the match processor on the canvas and open the Match sub processor 2 Select the Match or Alert Groups tab at the top 3 Uncheck the option to Generate Match Groups report or to Generate Alert Groups report Or if you know you only want to output the groups of related or unrelated records use the other tick boxes on the same part of the screen To disable Relationships output 1 Open the match processor on the canvas and open the Match sub processor 2 Select the Relationships tab at the top 3 Uncheck the option to Generate Relationships report 2 22 Using Oracle Enterprise Data Quality Publishing to the Dashboard Or if you know you only want to output some of the relationships such as only Review relationships or only relationships generated by certain rules use the other tick boxes on the same part of the screen To disable Merged Output 1 Open the match processor on the canvas and open the Merge sub processor 2 Uncheck the option to Generate Merged Output Or if you know you only want to output the merged output records from related records or only the unrelat
121. or the needs of the data For example The data capture system does not have fields for each distinct piece of information with a distinct use leading to user workarounds such as entering many distinct pieces of information into a single free text field or using the wrong fields for information which has no obvious place for example placing company information in individual contact fields The data needs to be moved to a new system with a different data structure Duplicates need to be removed from the data and it is difficult to identify and remove duplicates due to the data structure for example key address identifiers such as the Premise Number are not separated from the remainder of the address Alternatively the structure of the data may be sound but the use of it insufficiently controlled or subject to error For example Users are not trained to gather all the required information causing issues such as entering contacts with data cheats rather than real names in the name fields The application displays fields in an illogical order leading to users entering data in the wrong fields 3 14 Using Oracle Enterprise Data Quality Parsing Concept Guide Users enter duplicate records in ways that are hard to detect such as entering inaccurate data in multiple records representing the same entity or entering the accurate data but in the wrong fields These issues all lead to poor data qualit
122. ore information about EDQ see the following documents in the Oracle Enterprise Data Quality documentation set EDQ Documentation Library The following publications are provided to help you install and use EDQ a Oracle Fusion Middleware Release Notes for Enterprise Data Quality a Oracle Fusion Middleware Installing and Configuring Enterprise Data Quality a Oracle Fusion Middleware Administering Enterprise Data Quality a Oracle Fusion Middleware Understanding Enterprise Data Quality a Oracle Fusion Middleware Integrating Enterprise Data Quality With External Systems a Oracle Fusion Middleware Securing Oracle Enterprise Data Quality vii a Oracle Enterprise Data Quality Address Verification Server Installation and Upgrade Guide Oracle Enterprise Data Quality Address Verification Server Release Notes Find the latest version of these guides and all of the Oracle product documentation at https docs oracle com Online Help Online help is provided for all Oracle Enterprise Data Quality user applications It is accessed in each application by pressing the F1 key or by clicking the Help icons The main nodes in the Director project browser have integrated links to help pages To access them either select a node and then press F1 or right click on an object in the Project Browser and then select Help The EDQ processors in the Director Tool Palette have integrated help topics as well To access them right click on a processor on the c
123. ors Data Analysts Data Stewards Executives Match Reviewers Review Managers Check the Enabled box Select the Notification template from the drop down list Click to select the Users and Groups to send the notification to To select more than one User and or Group hold down the CTRL key when clicking Click OK Note Only users with valid email addresses will receive emails For users that are managed internally to EDQ a valid email address must be configured in User Administration For users that are managed externally to EDQ for example in WebLogic or an external LDAP system a valid mail attribute must be configured 2 4 2 Default Notification Content The default notification contains summary information of all tasks performed in each phase of a job as follows Snapshot Tasks The notification displays the status of the snapshot task in the execution of the job The possible statuses are STREAMED the snapshot was optimized for performance by running the data directly into a process and staging as the process ran FINISHED the snapshot ran to completion as an independent task CANCELLED the job was canceled by a user during the snapshot task WARNING the snapshot ran to completion but one or more warnings were generated for example the snapshot had to truncate data from the data source ERROR the snapshot failed to complete due to an error 2 14 Using Oracle Enterprise Data Quality Job Noti
124. ote that the match processor must be rerun after review activity takes place to write all decisions and comments to the Decisions output Also it is not possible to export Case Management decisions via a match processor The import export options within Case Management Administration for workflows and case sources must be used instead The following screenshot shows an example of the Decisions data output written out using a Writer ReviewDate RuleDecision CommentBy CommentDate Comment Customer DB ID Customer DB RelatedID nickf 31 Jul 2009 10 29 37 Match 48718 5907 nickf 31 Jul 2009 10 29 48 Verified by phone call 48718 5907 nickf 31 Jul 2009 10 29 56 No Match 4026 24735 mikem 27 Jul 2009 09 07 17 Match 49064 14952 nickf 31 Jul 2009 10 30 05 Husband and wife 24735 4026 nickf 31 Jul 2009 10 30 13 Match 24691 42360 nickf 31 Jul 2009 10 30 31 Match 43455 24688 nickf 31 Jul 2009 10 30 58 Moved house Verified by NCOA check 24688 43455 nickf 31 Jul 2009 10 31 05 Match 48633 6537 nickf 31 Jul 2009 10 31 15 Moved house Verified by NCOA check 48633 6537 nickf 31 Jul 2009 10 31 20 Match 48594 13147 nickf 31 Jul 2009 10 31 28 Moved house Verified by NCOA check 48594 13147 nickf 31 Jul 2009 10 31 33 Match 48589 48271 nickf 31 Jul 2009 10 30 25 Verified by email 24691 42360 nickf 31 Jul 2009 10 31 38 Verified by phone call 48589 48271 richardt 31 Jul 2009 10 32 32 Pending 19349 48586 richardt 31 Jul 2009 10 32 38 Pending further information
125. oundaries of the surrounding columns are tightly defined no extraneous data will be included in the data set Columns do not have to be specified in the same order as they occur in the file The column order specified here will be reflected in any snapshots created from the data source The buttons to the right of the columns table can be used to add or remove records or move the selected record up or down in the list 1 7 1 Files containing no new line characters By default it is assumed that fixed width files will be formatted as already described with a new line separating one row from the next However some files do not use new line characters to separate rows Data will then appear like this in a text editor Getting Started 1 11 Connecting to a Data Store P cars_no_newlines txt Notepad File Edit Format wiew Help 20060601 Briw OOO 1999 Jones 45 500 40050601 i Brew 0002 2003 Chau r3 200 40060102 Lexus 0003 2006 smith 25 365 20070930 Mazda ood 2007 Mukherjee 35 000 20090909 Toyota OO05 2009 Barker 400 K LN E ee ne a In this case the width of the whole record must also be specified as part of the data store configuration so that EDQ can correctly subdivide the data into rows To do this a Check the Fixed width records checkbox underneath the columns table and Specify the total record size in characters in the Record size box Edit Data Store Fixed width text Files Conf
126. ow To 2 25 Packaging 2 7 3 Opening a package file and importing its contents To open a Director package file either right click on an empty area of the Project Browser with no other object selected in the browser and select Open Package File or select Open Package File from the File menu Then browse to the dxi file that you want to open The package file is then opened and visible in the Director Project Browser in the same way as projects The objects in the package file cannot be viewed or modified directly from the file but you can copy them to the EDQ host server by drag and drop or copy and paste in the Project Browser You can choose to import individual objects from the package or may import multiple objects by selecting a node in the file and dragging it to the appropriate level in your Projects list This allows you to merge the entire contents of a project within a package into an existing project or for example to merge in all the reference data sets or processes only For example the following screenshot shows an opened package file with a number of projects all exported from a test system The projects are dragged and dropped into the new server by dragging them from the package file to the server a dn dt 0504 mikem E My Issues fe BD Prose Gl we Customer data assessment fa BO Compliance ce 8 Customer migration a g Reference Data e 4 Data Stores D All Projects From Test Area
127. pare mode 4 Re starting the real time process Triggers which can start and stop other jobs can be used to create a job which includes all the above phases A Stop re prepare and restart job would appear as follows Stop re prepare and restart EJ Real time CIW check FJ On e ite Stop and retake snapshot d Real time OJ check F3 a DIW _ Countries DIW _ Countries Re prepare DIW Individuals DJW Entities All INIT DIW DJW _Entities_In Prepare an dexed DJW _Individuals _Indexzed Restart oO Real time GIW check 4 P r s ei gape ae a ee E OU ee Ree T a A This job consists of three phases The first stops the real time match process and re runs the reference data snapshot The second runs the match process in Prepare mode The third re starts the real time match process in Interval mode Triggers are configured to run either at the beginning of a phase or at the end so it would be possible to include the restart trigger at the end of the second phase just as the stop trigger is placed at the beginning of the first phase However placing the restart trigger in its own phase means you can configure the restart phase to run only if the re prepare phase is successful see the Using Job Triggers topic in Administering Enterprise Data Quality for further information Advanced Features 3 13 Parsing Concept Guide Enabling r
128. plete the Match Review links associated with the match processor will also be disabled Advanced Options screen Ind ii duals View Match Statistics For more information see Understanding Enterprise Data Quality and Enterprise Data Quality Online Help 2 23 Execution Types 2 23 1 Batch EDQ is designed to support three main types of process execution a Batch Real time response Real time monitoring However in order to preserve the freedom to design processes independently of their intended execution mode and switch between types easily execution options are configured at the process or job level Processes are normally initially designed as batch processes as a set of data is normally required in order to test that a process produces the required results even if it will eventually be deployed as a real time process Batch processes have the following characteristics They read data from staged data configurations such as snapshots They may include any of the processors in the processor library including those such as Duplicate Check that are not suitable for real time response processes a They either write no results or write their results to a staged data table and or external database or file a They are executed in Normal mode That is the process completes when the batch has been completely processed 2 74 Using Oracle Enterprise Data Quality Execution Types 2 23 2 Real time re
129. processors topic in Enterprise Data Quality Online Help and as a process or job level override How To 2 21 Optimizing Job Performance To override the individual settings on all match processors in a process and disable the sorting filtering and review of match results untick the option to Enable Sort Filter in Match processors in a job configuration or process execution preferences r BoA Process Configuration Details Match individuals Type Process Detail Match individuals 1 Ja Enabled i J CUST_Individual Match Ri Readers Process Writers s Indexed individuals Use Intelligent Execution Publish to Dashboard C Enable Sort Filter in Match _ Results Drill Down OM G Sample None m ai i a a en x a a mm ie ee ee ee d ee ep a T m qo mete eee oe _ a i T a ee er ja aTr h z a I ee i Te F m Minimizing Output Match processors may write out up to three types of output Match or Alert Groups records organized into sets of matching records as determined by the match processor If the match processor uses Match Review it will produce Match Groups whereas if if uses Case Management it will produce Alert Groups Relationships links between matching records Merged Output a merged master record from each set of matching records By default all available output types are written Merged Output cannot be written from a Link process
130. r sample file 1 10 Using Oracle Enterprise Data Quality Configuring Fixed Width Text File Formats New Data Store Data Store Configuration Enter connection details For data File Cidatalcars txt al Encoding 150 68659 1 w ad Mame purchase date brand registration year mileage at purchase Fixed width records Record size ail Each column is described to EDQ by its starting position and width in characters Each column is also assigned a name which is used in data snapshots and downstream processing so that the data can be identified Names are defined by the user at the time the data store is defined and should be descriptive for maximum downstream usability Notice that the positions of the data columns are defined in terms of start point and width Note also that the first character on a line is at position 1 not zero Providing a width and a starting point for each column means that EDQ does not assume that one column continues right up until the start of the next with the result that Any spaces that have been included in the file for readability such as a single space between columns can automatically be bypassed x Itis not necessary to define mappings for every column in the file If un needed columns exist they can simply be omitted from the column definitions in the data store configuration For example we have not included the third column from the file in our mappings but because the b
131. ribute or across a number of attributes A String of data may be represented using a number of different token patterns Selection The process by which the Parse processor attempts to select the best explanation of the data using a tuneable algorithm where a record has many possible explanations or token patterns Resolution The categorization of records with a given selected explanation token pattern with a Result Pass Review or Fail and an optional Comment Resolution may also resolve records into a new output structure using rules based on the selected token pattern Summary of the EDQ Parse Processor The following diagram shows a summary of the way the EDQ Parse processor works Input data Syntactic analyst of data Split data into smallest unit base tokens Semantic analysis of data Assign meaning to tokens Examine token sequences fornew classified tokens Select the best description of the data where possible Resolve data to its desired structure and give a result Output data and metadata See the help pages for the EDQ Parse Processor in Enterprise Data Quality Online Help for full instructions on how to configure it 3 16 Using Oracle Enterprise Data Quality
132. rise Data Quality Online Help How To 2 79 Publishing Result Views to Staged Data 2 80 Using Oracle Enterprise Data Quality 3 Advanced Features This chapter provides an introduction to the Advanced Features of EDQ This chapter includes the following sections a Matching Concept Guide Clustering Concept Guide Real Time Matching a Parsing Concept Guide 3 1 Matching Concept Guide Why people need matching The need to match and reconcile information from one or more business applications can arise in numerous ways For example a Reconciliation of data held in multiple systems Mergers or acquisitions resulting in duplicate systems and information Migration of data to new systems with a desire to eliminate duplicates An identified need to improve the quality of system information by matching against a trusted reference set Why matching can be complex Defining whether records should match each other is not always simple Consider the following two records CUNO CU ACCOUNT TITLE NAME _ ADDRESS ADDRESS2 ADDRESS3 _ POSTCODE 10782 95 15134 SH Mr Victor CARSON 1A Spire Road Glover Road EstEast 9 Washington NES 3E5 15906 96 21229 FB Mz J CARSON Spire Road Glover Estate East WASHINGTON Tyne amp Wear NES 35 They are different in every database field but on inspection there are clearly similarities between the records For example They share a common surname The address could be
133. rprise Data Quality Online Help 2 19 Externalizing Configuration Settings It is possible to expose a number of configuration settings that are initially configured using the Director application such that they can be overridden at runtime either by users of the EDQ Server Console user application or by external applications that call EDQ jobs using its Command Line Interface This allows users of third party applications that are integrated with EDQ to change job parameters to meet their requirements For example users might use a standard EDQ job on several different source files specifying the input file name as an option at runtime along with a Run Label used to store results separately for each run of the job Job phases are automatically externalized that is the Director user does not need to configure anything to allow phases to be enabled or disabled at runtime using a Run Profile The other points at which options can be externalized are below Click on each point for further information a amp Externalizing Processor Options Externalizing Match Processors amp xternalizing Jobs amp xternalizing Snapshots Server side only a amp Externalizing External Tasks amp xternalizing Exports Server side only 2 19 1 Externalizing Processor Options There are two stages to externalizing processor options Select the processor options to externalize x Provide meaningful names for the externalized optio
134. rs 2 9 Creating Processors In addition to the range of data quality processors available in the Processor Library EDQ allows you to create and share your own processors for specific data quality functions There are two ways to create processors Using an external development environment to write a new processor see the Extending EDQ topic in Enterprise Data Quality Online Help for more details Using EDQ to create processors read on in this topic for more details 2 9 1 Creating a processor from a sequence of configured processors EDQ allows you to create a single processor for a single function using a combination of a number of base or member processors used in sequence Note that the following processors may not be included in a new created processor a Parse a Match Group and Merge Merge Data Sets A single configured processor instance of the above processors may still be published however in order to reuse the configuration Processor creation example To take a simple example you may want to construct a reusable Add Gender processor that derives a Gender value for individuals based on Title and Forename attributes To do this you have to use a number of member processors However when other users use the processor you only want them to configure a single processor input Title and Forename attributes however they are named in the data set and select two Reference Data sets one to map
135. ructured Names El a ee ee How To 2 71 Running Jobs Using Data Interfaces 2 3 the Data Interface Data Interface Unstructured Names Unetruchured Hemes ES 3 Parse Unstructured Names caer C Then double click the input and output Data Interfaces to select the Mappings For the input Data Interface select a web service input that has been mapped to 7 Data Interface Unstructured Names source Sampling Configure Mappi a ch Customer Data to Unstructured Names Real Time Unstructured Name Parse Input For the output Data Interface select both a Batch staged data mapping and a Real Time web service output mapping using the dialog below Data Interface Structured Names O Structured Names scheduled 2 72 Using Oracle Enterprise Data Quality Enahini Data Interface Structured Names H Structured Names Enable Sort Filter oon C Candidates Batch Customers Parsed to Strux Ree Time Name Parse Response Click OK to save The job appears as follows and is now ready to be run or Publishing to Case Management 3 Parse Unstructured Names Data InterFace Structured Data InterFace Unstructured Names b gt Names Structured Names H Unstructured Names ni Name Farse Output 4 parse_names parse_names a Poe s atl te ten ao lee nm A mal el te ey Rie eed nm EE eet ls el Rie ala iD meal ty oi For more information
136. s Publish as New this exposes the option as a new option on the processor you are creating Use an existing published option if any this allows a single published option to be shared by many member processors For example the user of the processor can specify a single option to Ignore Case which will apply to several member processors Note If you do not publish an option that uses Reference Data the Reference Data will be internally packaged as part of the configuration of the new processor This is useful where you do not want end users of the processor to change the Reference Data set In our example we open up the first member processor Derive Gender from Title and choose to publish as new the option specifying the Reference Data set used for mapping Title values to Gender values How To 2 31 Creating Processors 1 Derive Gender from Title Enhance from Map Bg Attributes Options Notes Value Map id Match Options Ignore case Yes No C Publish as Ne Match list by Whole value w Publish as ie Show publishing options Note above that the Match Options are not published as exposed options meaning the user of the processor will not be able to change these We then follow the same process to publish the option specifying the Reference Data set used for mapping Forename values to Gender values on the fourth
137. s may be included in the zip file and embedded in or linked from the main help page This is designed so that a help page can be designed using any HTML editor saved as an HTML file called index htm and zipped up with any dependent files To do this right click the published processor and select Attach Help This will open a file browsing dialog which is used to locate and select the file Note The Set Help Location option is used to specify a path to a help file or files rather than attaching them to a processor This option is intended for Solutions Development use only If a processor has help attached to it the help can be accessed by the user by selecting the processor and pressing F1 Note that help files for published processors are not integrated with the standard EDQ Online Help that is shipped with the product so are not listed in its index and cannot be found by search 2 10 3 Publishing processors into families It is possible to publish a collection of published processors with a similar purpose into a family on the Tool Palette For example you may create a number of processors for working with a particular type of data and publish them all into their own family To do this you must customize the family icon of each processor before publication and select the same icon for all the processors you want to publish into the same family When the processor is published the family icon is displayed in the Tool Pa
138. s List In this case you should still aim for clusters with a maximum size of around 500 records if possible bearing in mind that every record in the cluster will need to be compared with every other record in the cluster so for a single cluster of 500 records there will be 500 x 499 249500 comparisons performed See the Clustering Concept Guide for more information about how clustering works and how to optimize the configuration for your data Disabling Sort Filter options in Match processors By default sorting filtering and searching are enabled on all match results to ensure that they are available for user review However with large data sets the indexing process required to enable sorting filtering and searching may be very time consuming and in some cases may not be required If you do not require the ability to review the results of matching using the Review Application and you do not need to be able to sort or filter the outputs of matching in the Results Browser you should disable sorting and filtering to improve performance For example the results of matching may be written and reviewed externally or matching may be fully automated when deployed in production The setting to enable or disable sorting and filtering is available both on the individual match processor level available from the Advanced Options of the processor for details see Sort Filter options for match processors in the Advanced options for match
139. s a standalone task in Director by right clicking on the Export and selecting Run it always runs in Delete current data and insert mode Delete current data and insert and Replace records using primary key modes perform Delete then Insert operations not Update It is possible that referential integrity rules in the target database will prevent the deletion of the records therefore causing the Export task to fail Therefore in order to perform an Update operation instead Oracle recommends the use of a dedicated data integration product such as Oracle Data Integrator 1 16 Using Oracle Enterprise Data Quality Running a Prepared Export 1 10 2 Results Book Export Results Books can also be exported to a Data Store either manually or as part of a scheduled job Note Results Book Exports always perform a Delete current data and insert operation To set up an Export of a Results Book 1 Right click on Exports in the Project Browser and select New Results Book Export 2 Select the Results Book 3 Select the Result Pages to export and the number of rows to export Entering no value will cause all the records to be exported 4 Select the Data Store to export to Note If writing to a database a Results Book Export always perform a Delete current data and insert operation i e it creates tables corresponding to the names of the Results Pages if they do not exist and overwrites tables of t
140. s implemented 2 21 1 Configuring a Data Interface in a job When a process containing a Data Interface is added to a job it will appear as in the following example Cy Dais interface Examoe E OGB T 89a AK 3 Parse Unstructured Mantes Data Interface Unstructured Date Interface Shructured Hames Hames Unstructured Names Ve z a Structured Mames e ee e EE f EO N ee ee ee ee eee ee ee ee ee es eee eT T a ee Eee eee oe r How To 2 69 Running Jobs Using Data Interfaces Any Data Interfaces that appear in the job must be configured in order for the job to run To configure each Data Interface 1 Double click on the Data Interface The Configure Task dialog is displayed Configure Task i 1 os 3 Parse Unstructured Names Enabled F Ga Reader Sample Data Interface Unstructured Names el Unstructured Hames Data Aber tace linsinoctuted Hamet ih j Da Linghuctiired Hames maong GA Asb Cueiome Daia fo Uneinuctured Hames Rea Tire Urari ed Mane Parse Ingut teo Bk mmda Bee yi iii dela n ii ii ee i a ah aaae Ree ie ig Aa alo ln ante Ea Yr a Ae a a k a ah ae ean RS iea E a et aa 2 Select the required Mapping in the drop down field Note Itis possible to configure multiple mappings for a writer for example to write data to two difference staged data sets but only a single mapping for a reader 3 Click OK to save or Cancel to abandon Once Data Interfa
141. se Project Specific Folder _specific_folder Allow SSL Hostname Mismatch yame_mismatch Ignore Invalid SSL Certificate _ssl_certificate 2 19 6 Externalizing Exports To externalize an Export right click on the Export in the Project Browser and select Externalize Select and edit the options as required Click OK to save or Cancel to abandon The fields on the right of all Externalization dialogs contain the default name used for each externalization attribute This is the label used to reference the externalized attribute in any third party tool or file 2 64 Using Oracle Enterprise Data Quality Externalizing Configuration Settings To change any field name check the box to the left and edit the field as required Note Avoid the use of spaces in attribute names Externalizing exports allows external users to override not only configuration of the export itself such as the table name to write to but also options of the Data Store the export is writing to for example the file name or database connection details Note It is not possible to externalize the configuration that maps Staged Data attributes to target database columns Furthermore if the export is not set up to create a new table and you want to change the target table that it writes to dynamically at runtime the target table must have exactly the same structure as the one used in the export definition with the same columns and d
142. so 2 0 HDi Desire 2 IgE JEEP GRAN EHEROI EE AACR A belek bel 2 44 Using Oracle Enterprise Data Quality Using the Event Log The Full Column Widths button will widen the columns to show the full data but in this case there is too much data to show on the width of a screen To see the FullDescription field as wrapped text it is possible to right click on the rows you want to view and use the Show Characters option You can then click on the arrow at the top right of the screen to show each value in a text area and use the arrows at the bottom of the screen to scroll between records W Show Characters Viewing item 1 of 10 5 Doors Manual Hatchback Petrol 21 631 miles Aruba Blue 1 Owner ABS Adjustable seats Adjustable steering column wheel Air conditioning Central locking Climate Control Computer Driver airbag Electric windows Head restraints Folding rear seats Immobiliser Park distance control Passenger airbag Power assisted steering Rear armrest Remote locking Side airbags Radio CD Rear headrests Insurance Group 4 350 Dec Comment 0053 Basic Latin number 0032 Basic Latin whitespace 0068 Basic Latin uppercase letter 0111 Basic Latin lowercase letter 0111 Basic Latin lowercase letter 0114 Basic Latin lowercase letter 0115 Basic Latin lowercase letter 0044 xO02 Basic Latin Basic Latin whitespace M 0077 Basic Latin uppercase letter 2 14 3 Select Column Headers Clicking on the c
143. source system then receives this message selects candidate records using a pre keyed table and feeds back to EDQ a set of all the records that have share one of the cluster keys with the driving record Normally it is advisable to generate many key values per record as is also the case for batch matching Matching In the Matching stage EDQ receives the message containing the records that share a cluster key with the input record It then uses a match processor with all records presented as working records on a single input to select the definite and possible matches to the new record from these candidate match records and assign them a score that can be used to sort the matches The matching results are then returned on a real time interface and this response is handled externally to determine how to update the system Possible responses include declining the new duplicate record merging the new record with its duplicate record s or adding the new record but including a link to the duplicate record or records The following is an example of a real time matching process gl Contact dedupe real time E j F a fel 4 F j Ea p D Db J f 5 je amp 4 i a E u i Get all raasd records l and Sonne m Mi ai ka aa Real time reference matching EDQ s real time reference matching implementation matches new records against one or many reference sets The data in the reference sets is assumed to be n
144. sponse Real time response processes are designed to be called as interactive services to protect data quality at the point of data entry A real time response process may perform virtually any data quality processing including checking cleaning and matching data Profiling would not normally be used in a real time response process Real time response processes have the following characteristics They read data from a real time provider such as the inbound interface of a Web Service Note A real time response process may also include Readers connected to staged data configurations such as snaphots for example when Real time reference matching in this case the process must be executed in Prepare mode before processing requests a They write data to a real time consumer such as the outbound interface of a Web Service Note A real time response process may include Writers connected to staged data configurations for example to write a full audit trail of all records processed and their responses These writers will not write any results until the process stops regardless of any interval settings a They are typically executed in Normal mode and do not write out results but may be executed in Interval mode allowing results to be written out while the process runs continuously a They should not include processors that are unsuitable for real time response processing such as Duplicate Check Note
145. st summary view of results Interim results views that are accessed by drilling down on the top level results views cannot be published Data views also cannot be published in this way data in a process is written to Staged Data using a Writer The publishing of Results Views to Staged Data has three purposes To export Results Views to a target Data Store To use Results View data in further processing for example in a Lookup To allow users of the Server Console UI to view selected process results Published Results Views are written to Staged Data on process execution To set up a Results View to be written to Staged Data 1 Select the processor in the Canvas to see its results in the Results Browser 2 Click to publish the Results View to Staged Data This brings up the Publish Results dialog How To 2 77 Publishing Result Views to Staged Data Publish Results Staged data Last Name Frequency Profiler Results Delete Published View Using this dialog you can a Specify or change the name of the Staged Data set that you want to write to a Change the attributes of the Staged Data set to use different attribute names from the generic Results View names that appear on the left Remove attributes from the Staged Data set if there are attributes in the Results View that you do not want to write out Switch the publishing of a Results View on or off without losing the configurat
146. t processor and select Externalize The Process Externalization dialog is displayed This displays all the externalized options in the process including any externalized match processor options How To 2 59 Externalizing Configuration Settings Ensure the Enable Externalization check box is selected Click on the required option in the top area of the dialog The bottom area shows the name of the processor it originates from and allows you to link directly to that processor in the Canvas Click Rename Enter the new name in the Rename dialog Click OK to save or Cancel to abandon 2 19 3 Externalizing Jobs Tasks within Jobs contain a number of settings that can be externalized To externalize a setting on a Task 1 2 3 4 5 Right click on the Task and select Configure Task Click the Externalize button next to the required setting Select the checkbox in the Externalize pop up A default name for the setting is displayed which can be edited if required Click OK These settings are then managed from the Job Externalization dialog To open the dialog click the Job Externalization button on the Job Canvas tool bar G Jeb Externalizaton Extermalized Opbons J Enable Externaization P snapshots GeshboardouDtesh gq drida Matchoustomers Ceancustomers Delete Externalized Tasks Enabled Enabled Externalization for a Job can be enabled or disabled by checking or clearing the Enable
147. tasks for example to re run snapshots run processes and external jobs See the Jobs topic in Enterprise Data Quality Online Help for further details For each export task selected define how the export should be performed A further option allows you to disable the export Click Run to run the job straight away or click Schedule to schedule the job to run at a later date or time 1 18 Using Oracle Enterprise Data Quality 2 How To This chapter provides information on how to perform certain key tasks in EDQ These are most useful when you already understand the basics of the product This chapter includes the following sections Execution Options Creating and Managing Jobs Using Job Triggers Job Notifications Optimizing Job Performance Publishing to the Dashboard Packaging Purging Results Creating Processors Publishing Processors Using Published Processors Investigating a Process Previewing Published Results Views Using the Results Browser Using the Event Log Reviewing Matching Results Importing Match Decisions Exporting Match Decisions Externalizing Configuration Settings Managing Data Interfaces Running Jobs Using Data Interfaces Publishing to Case Management Execution Types Customizing Processor Icons Enabling Case Management How To 2 1 Execution Options a Publishing Result Views to Staged Data 2 1 Execution Options EDQ can execute the following types of task either int
148. tch decisions If key information is changed between processor runs any previously saved match decisions for that relationship are invalidated and the relationship is raised for re review The Match Review application is suitable for implementations where a small number of reviewers in the same business unit share the review work It does not support 2 50 Using Oracle Enterprise Data Quality Importing Match Decisions group distribution or assignment of review work and offers basic review management capabilities 2 16 2 Case Management The Case Management application offers a customizable workflow oriented approach to match reviewing Case Management groups related review tasks alerts together to form cases Cases and alerts can both exist in a number of states with permitted transitions between them Case Management also supports automatic escalation of alerts on a time out basis or based on other configurable attributes Case Management supports assignment of work to individual users or to a group of users and more sophisticated review management capabilities than those offered by Match Review Case Management must be explicitly activated for each match processor that wants to use it and extra configuration is required to support the generation of cases and alerts Case Management does not support merge reviews For more information on the Case Management system see the Using Case Management topic in Enterprise Dat
149. te to stage data de selecting the option will mean that no staged data will be written and for writers that write to real time consumers de selecting the option will mean that no real time response will be written This is useful in two cases 1 You want to stream data directly to an export target rather than stage the written data in the repository so the writer is used only to select the attributes 2 6 Using Oracle Enterprise Data Quality Execution Options to write In this case you should de select the Write Data option and add your export task to the job definition after the process 2 You want to disable the writer temporarily for example if you are switching a process from real time execution to batch execution for testing purposes you might temporarily disable the writer that issues the real time response a Enable Sort Filter This option sets whether or not to enable sorting and filtering of the data written out by a Staged Data writer Typically the staged data written by a writer will only require sorting and filtering to be enabled if it is to be read in by another process where users might want to sort and filter the results or if you want to be able to sort and filter the results of the writer itself The option has no effect on writers that are connected to real time consumers 2 1 3 External Tasks Any External Tasks File Downloads or External Executables that are configured in a project can be add
150. ted below CUST_REF NAME_PREFIX First nar FIRST NAME SURNAME INITIALS GENDER N LAST_NAME ADDR1 ADDRESS_LINE1 ADDR2 ADDRESS _LINE2 ADDR3 LOCALITY ADDR4 REGION POSTCODE POSTAL_CODE TEL_NO TELEPHONE EMAIL_ADDRESS EMAIL For specific types of identifier such as a person s name EDQ s extensibility also allows the introduction of new identifier types and associated comparisons These new identifier types allow many fields to be mapped to a single identifier allowing structural differences to be dealt with once and then ignored In this case matching rules can be simple but powerful as they are working at the entity level rather than the field level As a result the configuration is quicker to define and easier to understand This is illustrated below 3 2 Using Oracle Enterprise Data Quality Matching Concept Guide CUST_REF TITLE FIRST NAME name f J a INITIALS UERN LAST_NAME oom 00 ddres lt Locau REGION POSTCODE POSTAL_CODE Z b gt Olo s olo ad A LA un ERG i gt 2 m mm W Clustering Clustering is a necessary part of matching used to divide up data sets into clusters so that a match processor does not attempt to compare every record with every other record In EDQ you can configure many clusters using many identifiers in the same match processor so that you are not reliant on the data having pre formed cluster keys To read more about clusterin
151. th a traffic light icon to indicate that it is the Dashboard Processor E All gt Valid gt Un known gt Invalid Publish Results to Dashboard Note In most cases it is advisable to use the same member processor for Results Views Output Filters and Dashboard Publication options for consistent results when using the new processor This is particularly true when designing a processor designed to check data 2 9 8 Setting a Custom Icon You may want to add a custom icon to the new processor before publishing it for others to use This can be done for any processor simply by double clicking on the processor outside of the processor design view and selecting the Icon amp Group tab See the Customizing Processor Icons for more details Once you have finished designing and testing your new processor the next step is to publish it for others to use For more information see Enterprise Data Quality Online Help 2 10 Publishing Processors Configured single processors can be published to the Tool Palette for other users to use on data quality projects It is particularly useful to publish the following types of processor as their configuration can easily be used on other data sets Match processors where all configuration is based on Identifiers Parse processors where all configuration is based on mapped attributes Processors that have been created in EDQ where configuration is based on
152. the DXI file from the server Copying objects from the package file will then be considerably quicker To open a package file in this way first copy the DXI file to the server landing area Then using the Director client right click on the server in the Project Browser and select Open Server Package File You then need to type in the name of the file into the dialog If the file is stored in a subfolder of the landingarea you will need to include this in the name For example to open a file called MDM dxi that is held within a DXI subfolder of the landingarea 2 7 5 Copying between servers If you want to copy objects between EDQ servers on the same network you can do this without packaging objects to file To copy objects such as projects between two connected EDQ servers connect to each server and drag and drop the objects from one server to the other Note To connect to another server select File menu New Server The default port to connect to EDQ using the client is 9002 For more information see Enterprise Data Quality Online Help 2 8 Purging Results EDQ uses a repository database to store the results and data it generates during processing All Results data is temporary in the sense that it can be regenerated using the stored configuration which is held in a separate database internally In order to manage the size of the Results repository database the results for a given set of staged data eith
153. the file as it exists or will exist in the server landing area including the file suffix It is possible to use a project specific landing area to enforce isolation between data in different projects Administrators will need to setup a landing area for the projects which require the use of this functionality Again for fixed width text files you must also define the fields that are present in the file For a Database specify the Database host Port number if not using the default port number Database Name User Name Password and Schema if different from the default Schema for the User For a database accessed via a JNDI connection specify the JNDI name For any other type of data that can be accessed via an ODBC bridge connector specify the ODBC DSN JDBC URL User name and Password For any Other type of data that can be accessed via JDBC specify the Driver Class Name JDBC URL User name and Password If you want to check the connection to the new data store use the Test button Note that it is possible to specify connection details to a file that is not yet present such as a file to be created by an export task in EDQ Note Connecting non native types of data source requires some knowledge of JDBC connectivity Give the data store a Name and click Finish The new data stories are now configured and visible in the Project Browser Alternatively if the data store is going to be shared across projects you
154. the new processor u Make Add Gender processor Reader Processor lt iili Frocess Processor X T Value i F BestGender TitkeGender NameGender e a T _ m A Mam o a 2 9 6 Setting Output Filters The Output Filters of the new processor are set to those of any one and only one of the member processors By default the final member processor in the sequence is used for the Output Filters of the created processor To use a different member processor click on it and select the Filter button on the toolbar O e HEEBS9IL S g IETA A gt The selected Output Filters are colored green in the processor design view to indicate that they will be exposed on the new processor i Enhanced i Not enhanced 4 Denve Gender from Forename 2 9 7 Setting Dashboard Publication Options The Dashboard Publication Options of the new processor are set to those of any one and only one of the member processors 2 34 Using Oracle Enterprise Data Quality Publishing Processors If you require results from your new processor to be published to the Dashboard you need to have an Audit processor as one of your member processors To select a member processor as the Dashboard processor click on it and select the Dashboard icon on the toolbar O HESF9G S S gt 35 ay amp onm ee ee E a ee oe T Tie niie n a a o The processor is then marked wi
155. the process chain from request to response EDQ will be able to reconcile responses to requests and the real time service will work as expected However if the data stream is split and then merged further down the process chain EDQ will be unable to reconcile the response to the request Therefore the first Web Service request sent will cause the job to fail The error and log message generated will contain the following text failed to transmit ws packet no request found matching this response An example of such a chain is shown in the following screenshot 1 Parse Person Name 2 Add Cluster Keys 3b Transform DI Person gt ra JJ DI Person DI Person Jd DI Person DI Person Jd DI Person p Ferson H Person H Person PersonMatch 2 React a Person fj 2 Person gg 4 Merge back driving record DI Person 3 Match Individuals Ferson H l iii 1 p DI Person y Dr Matches y DI Matches 9 le PersonMatch Ferson H J Matches H Matches H 2 21 4 Example Job containing two Data Interfaces In this example job a process is used that both reads from and writes to Data Interfaces The user selects mappings to allow the process to run in real time but also to log its real time responses to staged data 1 First create the job and drag the process onto the Canvas 3 Parse Unstructured Names Data Interface Unstructured y Data Interface Structured Names Names H Unstructured Names E d St
156. tibule 80 Leadenhall Street d Clark Mrs Helen MB Offices 122 Leadenhall Street Y Clark Mrs Helen Mec Physiotheropist 107 Leadenhall Street Clark Mrs Helen MB Retail Unit 107 Leadenhall Street 4 Clark Mrs Helen ME Snack Bar 104 106 Leadenhall Stri Clarke Mr Gerald 5 Flat 1 25 Savage Gardens Clarke Mrs Susan 35 Eastcheap Clark Mr Louis SD 8 Crosby Square 1 Clarke Ms F Offices 73 Pane ip ee ee T E T r In this way an expert user can inspect and tune the clustering process to produce the optimal results in the quickest time For example if the clusters are too big and matching performance is suffering extra clusters could be created or a cluster configuration may be made tighter If on the other hand the user can see that some possible matches are in different clusters the clustering options may need to be changed to widen the clusters Note With some clustering strategies large clusters are created This is often the case for example if there are a large number of null values for an identifier creating a large cluster with a NULL cluster key If a cluster contains more than a configurable number of records or will lead to a large number of comparisons being performed it can be skipped to save the performance of the matching engine The default maximum size of a cluster is 500 records and it is also possible to limit the maximum number of comparisons that should be performed for each cluster To change these
157. tributes that you want to expose and set it as the Results Processor see below to ensure that you see only the output attributes that you require in the new processor and not for example input attributes to a transformation processor Alternatively if you do not require a Results View you can unset it and the exposed output attributes will always be those of the Outputs processor only 2 9 5 Setting Results Views The Results Views of the new processor are set to those of any one but only one of the member processors By default the final member processor in the sequence is used for the Results of the created processor To use a different member processor for the results views click on it and select the Results icon on the toolbar Tote Gender attributes Note that in some cases you may want to add a member processor specifically for the purpose of providing Results Views In our example we may want to add a Frequency Profiler of the three output attributes TitleGender ForenameGender and BestGender so that the user of a new processor can see a breakdown of what the Add Gender How To 2 33 Creating Processors processor has done To do this we add a Frequency Profiler in the processor design view select the three attributes as inputs select it as our Results Processor and run it If we exit the processor designer view we can see that the results of the Frequency Profiler are used as the results of
158. ture in order to make it fit for purpose Token A token is a piece of data that is recognized as a unit by the Parse processor using rules A given data value may consist of one or many tokens A token may be recognized using either syntactic or semantic analysis of the data Tokenization The initial syntactic analysis of data in order to split it into its smallest units base tokens using rules Each base token is given a tag such as lt A gt which is used to represent unbroken sequences of alphabetic characters Base Token An initial token as recognized by Tokenization A sequence of Base Tokens may later be combined to form a new Token in Classification or Reclassification Classification Semantic analysis of data in order to assign meaning to base tokens or sequences of base tokens Each classification has a tag such as Building and a classification level Valid or Possible that is used when selecting the best understanding of ambiguous data Token Check A set of classification rules that is applied against an attribute in order to check for a specific type of token Reclassification An optional additional classification step which allows sequences of classified tokens and unclassified base tokens to be reclassified as a single new token Advanced Features 3 15 Parsing Concept Guide Term Definition Token Pattern An explanation of a String of data using a pattern of token tags either in a single att
159. ublished Results Views 2 13 Previewing Published Results Views It is possible to preview the results of running a process before it is run or to disable or delete specific results views This is done with the Published Results Views dialog _ Published Results Views Processors with published results views Quickstats Profiler Frequency Profile Tite City State Publish View Name Target Customer address city i Customer City Frequencies Customer title i Customer Tite Frequencies To open the dialog either a Click the Published Results Views button on the Process toolbar E Erge EES or right click on a processor in the Process and select Published Results Views in the menu Fy Remove Link to Reference Processor 5 gt Externalize Published Results Views T eee ee ee ee E E R E This dialog is divided into two areas Processors with published results views Lists all the processors in the process that have published views a Views Lists the views of the processor currently selected in the Processors with published results views area Published views can be selected or deselected by checking or unchecking the Publish checkbox Alternatively they can be deleted by selecting the view and clicking Delete If a view is deleted by accident click Cancel on the bottom right corner of the dialog to restore it 2 42 Using Oracle Enterprise Data Quality Using the Results Browser For more i
160. y which may in many cases be costly to the business It is therefore important for businesses to be able to analyze data for these problems and to resolve them where necessary The EDQ Parser The EDQ Parse processor is designed to be used by developers of data quality processes to create packaged parsers for the understanding and transformation of specific types of data for example Names data Address data or Product Descriptions However it is a generic parser that has no default rules that are specific to any type of data Data specific rules can be created by analyzing the data itself and setting the Parse configuration Terminology Parsing is a frequently used term both in the realm of data quality and in computing in general It can mean anything from simply breaking up data to full Natural Language Parsing NLP which uses sophisticated artificial intelligence to allow computers to understand human language A number of other terms are also frequently used related to parsing Again these can have slightly different meanings in different contexts It is therefore important to define what we mean by parsing and its associated terms in EDQ Please note the following terms and definitions Term Definition Parsing In EDQ Parsing is defined as the application of user specified business rules and artificial intelligence in order to understand and validate any type of data en masse and if required improve its struc
Download Pdf Manuals
Related Search
Related Contents
TOTAL SCHUTZ Philips Wall light 34082/11/76 manual de instrucciones manual de instruções DjVu Document Sound Bar Kelly Motor Controller - Electrathon of Tampa Bay ASUS Transformer Pad FHD 10 TF303CL 16GB 3G 4G Blue Sundtek DM500 HD USB PCB User Guide Aperçu - Mairie de Sainte Operating manual Copyright © All rights reserved.
Failed to retrieve file