These procedures are followed in the process of data cleansing
Data Parsing: Parsing data extract one or more individual data elements from source files, which are then filled in the target files.
Data Correction: We fix any data inaccuracies using appropriate tools or secondary data sources.
Data Standardization: We change the data to fit the stipulated formats which could be business standards and/or customized rules.
Data Matching: We perform data matching on processed data to check for any duplication according to the business rules.
Data Consolidation: Merged processed data is analyzed to establish relationships and data which are found to be related is presented as a consolidated representation.
POST-MIGRATION TESTING
After the migration stage is done, black box testing/acceptance testing is done against business cases, use cases, and user stories. The same data subsets that were utilized during pre-migration testing are also utilized during post-migration testing. This activity is done in a test environment as opposed to the production environment.
Post-Migration Activities:
Evaluate how many records can be processed in a set amount of time (the throughput). Through this testing, we aim to verify that the downtime determined in advance will be sufficient.
Check the completeness and the accuracy of the migrated data.
Check the data record counts of the migrated data just to ascertain that all the data is actually migrated. Record count match only displays if data migration is completed or not. It does not disclose internal aspects like redundancy, validation, integrity, and other data attributes.
Check that the migration of contents is done according to the outline. We conduct this type of testing through sampling of data or automation tools.
Along with testing the known business workflows, our testers carry out the following tests. We also employ negative testing approaches primary to ensuring that data cleansing occurs at runtime within the system.
Input invalid data: We attempt to breach the validation rules utilizing boundary value analysis, equivalence partitioning, and error guessing. Validation at the user interface level should corroborate with the imposition of validation at the database level. If no validation is enforced or if the validation between the database and the user interface do not correspond, we treat this as a defect.
Skip over required data: We try to move on to subsequent steps of a use case in the system before all data requirements are completed.
Verifying data access locks: It must not be possible to allow multiple users access to the same new record within the database simultaneously.
The risk might be minimal as normal validation procedures are typically handled by the database system, but this testing still must be conducted anywhere deemed necessary.
Conversions testing: This process entails verifying that the data is accurately displayed upon request from the database(s) according to the requesting user’s access privileges.
Non-functional testing: These are other tests that we conduct outside of functional boundaries which include: security testing, performance testing, usability testing, etc.
USER ACCEPTANCE TESTING
After the post-migration testing is performed, we carry out user acceptance testing. We expect some level of functional weaknesses due to the consolidation of the migrated data during earlier stages, and this mitigation strategy is effective because the previous phases can overlook these subtle differences. It serves to enable interfacing with the legacy data in the claimed system prior to pushing it for production.
PRODUCTION MIGRATION
Even with successful execution of all the above phases, we acknowledge that there is no guarantee that the production process will be seamless. Some of the challenges/areas that we take care of at this stage are:
Procedural slips. Production system configuration errors.
If automated tools for software testing were utilized in the above phases, we prefer to execute those scripts here again. Otherwise, we perform sample or summary verification.
DATA QUALITY TOOLS
Here are some of the areas DQ tools can assist with,
Cost and Scope Analysis
Migration Scenario
Data Quality Governance
Sequence of Activities for Migration Execution
Self-Validation of Migrated Data on the Target System
Continuous Improvement of Data Quality
The following is a partial list of systems used for integration and migration of data: