Tagarchief: Testing

Analyzing and Testing Data Environments – Common Issues

Analyzing and Testing Data Environments – Common Issues

Armando Dörsek © 2023

Introduction

When dealing with data centric systems, e.g. when analyzing and testing data warehouse or business intelligence environments, you’ll probably come across various errors. These errors can be due to insufficient analysis of the source data, developing the wrong transformations or even by using the system in the wrong way. To explain and categorize these common mistakes, you’ll find a summary of common mistakes below. 

Types of errors

  1. Homonyms and Synonyms in columns/field names
  2. Technical differences between systems
  3. Wrong usage of source systems
  4. Fields/Columns filled with technically correct values but not providing any real business
  5. History issues
  6. Wrong mappings
  7. Wrong transformations

Homonyms/Synonyms

When various concepts are known under the same description, we call this a “homonym”. For instance, an “area” may be used to depict a quantity of a two-dimensional surface (e.g. measured in square meters) but it may also refer to a region (e.g. a neighborhood, borough or zone).

When two different names are used for the same concept, we call these “synonyms”. For instance, when analyzing concepts like “contact” and “client”, it is important to figure out if the meaning of the concepts are 100% identical or that slight differences apply, based on the usage in the source systems. E.g. for some departments, contacts may be treated as a customer only after they have actually put in their first order with the company.

When analyzing, building and testing interfaces between systems, data integration and reports, this is something to take a good look at.

Technical Differences between Systems

When transferring (loading, moving) data from one system to another and integrating these values, issues may arise due to technical differences between systems. In the design, development and test of the interfaces and data integration, this needs attention.

Character Sets

One example is the different implementation of treating characters, i.e. storing specific characters in a different way on the various systems. Computers are good at storing numbers (1’s and 0’s) but for storing texts, character sets were developed. These may be incomplete and not able to store all relevant characters – leading to simplifying texts. E.g. the ASCII character set is able to store “plain English” characters but to store all letters, punctuation and symbols of other languages, the Unicode character set is required. 

Physical Storage of Data

Related to the issue above, is the difference in storing files on different systems like mainframes, Windows based system and a Unix/Linux based systems. Files may be treated in a different way where it comes to technical features like “end of line” characters and the way specific characters are stored.

Differences in Data Types

When storing data, you often get to choose from different data types, like character, numbers, dates and currencies. Various Database management systems (DBMS’s) and integration tools treat data types in different ways. These differences may be inherent to the system, others may be configured by technical staff. E.g. if a source system only supports storing text and target systems have ways to store numbers, dates and texts, some kind of transformation is needed. The analysis of the possibilities in the source systems is crucial: is “7,001” to be treated the same as “7.001”? Data profiling tools can be helpful. Make sure ask the right questions during review of the specifications, to test variations in the unit- and system tests and to have “rich” acceptance test sets that cover all kinds of different values in the source systems.

Wrong use in/of Source Systems

Users are extremely resourceful when it comes to using their tools (i.e. our source systems) in the most efficient manner. This will sometimes lead to using in a way it was not meant to, e.g. by filling “free format” fields with data that should have been stored elsewhere. Or by team A using field X in a different way than team B, making it difficult to analyze the data in that certain field (X).

Values are Technically Correct – but have low Business Value

When analyzing data sets, you may come across tables which contain a lot of “generic” of default values. Consider a table with customers having a column called “MarketSegment”. If most hold the value “99 Various Market Segments”, the business value of having such a field in a report is quite low.

When analyzing data in order to build meaningful reports, this should be mentioned to end users in an early stage – as they will probably record this as a finding or bug at the time of an acceptance test. 

The following situation is even worse: when users decide to enter blanks or certain values like “X” or “0,0” in fields that should have been filled in a correct manner. In this case, the value is not only too generic (like with “99 Various Market Segments”) but plain wrong. If not addressed, it may lead to wrong business decisions. E.g. when car crashes are not registered on their precise location, due to users entering wrong geo locations, dangerous streets and crossings may not get improved in time.

History Issues

Some source systems are designed is such a way that only the current situation is stored, and does not retain “history”. Should history be required this may lead to (at least) two kinds of issues.

The first issue arises when end users get creative in storing information that they want to keep in fields that weren’t meant for this purpose, like in a “Remarks” field. Should data analysts overlook the data in this field, e.g. because the technical documentation says it’s “just for remarks”, this may lead to issues. Data profiling should be applied, to find out if these situations occur in the source systems. Even if the issue is addressed, retrieving the relevant data from such fields can be hard and error prone.

The second issue happens when there is uncertainty about the actual validity of a value, as the original values get overwritten in the source systems. Is the shown value “the latest value” or the “original value”? Input from the user community is required.

Note: A solution may consist of developers building a system for “change data capture” (CDC), registering each and every change in the source system and figuring out what to do with it based on technical and business rules.

In data warehousing, history is often kept by the use of Slowly Changing Dimensions type 2 (SCD2). Tables have columns added with a begin- and end date, showing the period in which the situation is valid. When an attribute changes value, a new row is added with a new begin date (and the old record gets “closed”). More variants of SCD exist, and when developing solutions based on SCD, errors can be introduced as well. Be sure to have a test plan in place to cover the basic situations when working with SCD.

Wrong Mappings

Mappings can be implemented in a wrong way, in many ways.

First, due to a human errors, columns may be mapped to other columns than intended (in the design). E.g. mapping “StudentName” in the student database to “BANKACCOUNT” instead of “LASTNAME”. Replacing the manual mapping with more automated mappings tackle many of these issues and many of the wrong mappings are easily discovered by a visual check. Should a mapping hold many similar fields, this may become more difficult, but even so: checking it is very valuable, as solving issues in Acceptance Tests are more costly than bugs that are found early in the process.

A second type of error does not concern mapping the actual fields, but concerns the contents of the fields. E.g. when mapping dates from TEXT and DATE fields to other DATE and DATETIME, extra attention is required from developers and testers. Does “20230305” stand for 3 May 2023 or 5 March 2023? Are DATETIME fields (e.g. “20230305000000”) actually holding date/time with much precision – or only the date?

If several systems hold similar values, these may be integrated, but small differences are lurking around the corner: if system A will store “sex” coded as 1 (male), 2 (female) and 0 (unknown and other) and system B will store texts “boy”, “girl”, “man”, “woman”, “military”, “disabled”, “other” and <blank>, then business rules will need to be applied to make sure that the data in Data Marts and reports will be meaningful.


Suggestions

This is just a limited number of issues that may occur. You’ll probably have found some of these, and others, yourself. Feel free to share your experiences and report issues that you’re experiencing in the field of data.

Training “Certified Data & Analytics Tester” (CDAT) – Live online

CERTIFIED DATA & ANALYTICS TESTER – 2 DAYS TRAINING (ENGLISH SPOKEN)
Due to the increasing demand for specialized education regarding testing of Data Warehouses (DWH), Business Intelligence (BI) and Analytics, we are hosting the unique training “Certified Data & Analytics Tester”. 

Since the start of this training in 2018, we have welcomed many groups of (future) data & analytics testing consultants. The training is well received and we are proud of an average rating of 8/10! At the end of this page, you’ll find the scheduled ‘open’ courses to which you can enroll. Would you rather participate as a group? You can also book an in-company course. Please contact us and we will answer your questions and/or help you to enroll in one of the courses.

Introduction

Today everybody recognizes the importance and added value of data for your Business model, if applied well. Whole Industries are changing: think of companies like Amazon, Uber, Netflix and Walmart. Data is high-prioritized at more and more organizations strategy agendas. They hire (interim) professionals who can collect, analyze, model and visualize (Big) Data. Both these companies and their clients trust on reliable, complete, correct and on-time delivered information. Not just one time, but continually. Data professionals can provide this trust by using methods from the field of quality assurance and testing. So, where to start? In this practical course you will get insights to a structured testing approach in complex Data & Analytics projects.

Set up

During this 2-day course at foundation level you will learn about Data & Analytics; what is it and how do to test in a Data & Analytics environment. This test method is based on international test-standards (ISTQB) with a specific and practical translation to a Data (Analytics) environment.

The course is set up in such a way that enables both data professionals and professionals to share ‘a common language’ regarding testing in Data & Analytics environments. Accordingly, the student will learn how to embed structured testing to reach and improve the quality goals.

The course combines theory with practice and contains many exercises which connects the student with practice and delivers the most applicability.

Subjects

After this training, you will:

  • be able to define a vision and test strategy – and to translate it to an efficient test approach in data-oriented environments and -projects.
  • have become familiar with test specification techniques applicable to data & analytics projects.
  • be able to distinguish between different data quality attributes, measure these and you will be able to use data profiling techniques
  • have become familiar with the specifics of the test environments and privacy-aspects of data its usage in data & analytics testing projects..

Program

Day 1:

  • Towards a common Basis: Business Intelligence, Data warehousing en Data & Analytics
  • Testing and the Test Process
  • Risk Based Testing (RBT)
  • Tester’s Skills Matrix
  • Testing of Reports & Dashboards

Day 2:

  • Testing Transformations (ETL)
  • Testing Completeness of Data
  • Data Quality and Data Profiling
  • Test Environments
  • Privacy Aspects
  • Test tools

During the course, links between theory and practice are made and supported by examples of test plans, test design techniques and test reports according to ISTQB standards.

Requirements

The course will be provided (live) on line through MS Teams.

No specific education and/or training is required. Affinity with IT and/or business knowledge of Data & Analytics is an advantage, like: Data Warehousing, ETL, Data Migrations, Business Intelligence or Analytics. Knowledge of ISTQB is an advantage but not required. On the second day, attendees need to have access to a laptop with he possibility to install (open source) software.

Target Audience

The course is designed for test engineers, test coordinators, business analysts, data leads, data warehouse developers and BI-consultants.

Exam

The course can be concluded with an on line test (multiple choice), hosted by Brightest. An exam voucher will be provided. After successfully passing the exam, the attendee will be rewarded with the certificate ‘DAU Certified Data & Analytics Tester (Foundation Level)’.

Literature

At start of the course you will receive a copy of the slides, the syllabus/reader and several materials to support you during the course and in your daily practice.

Trainers

Armando Dörsek

Armando Dörsek is an ISTQB- and TMap® Next certified test manager. He has over 20 years of experience in ICT as a developer, coach and test consultant of which the last decade in Business Intelligence, Data Warehousing and/or Analytics focused environments. Recent clients were in the area of health insurance, banking, retail and government (tax, law enforcement). Projects are often large, complex and have a company-wide impact.

Rogier Ammerlaan

Rogier has been active for many years in the disciplines of software quality and agile development in financial institutions and more technical environments like greenhouse automation. After graduating on the topic of CMM he specialized in software testing, taking up several roles like tester, test coordinator and test manager in large (international) projects. Rogier is a near full time trainer in the field of testing and test improvement, scrum/agile and robotics.

Duration

2 days:

21 and 22 September
9 AM – 5 PM

Price

€ 1.195,- ex. VAT (including materials and one exam (€ 200,- ex. VAT))

Contact

Armando Dörsek, see Contact Form.

See Course Dates and Enroll

    Yes, I'd like to enroll in the following course:

    Please contact me!

     

    Training “Certified Data & Analytics Tester” (CDAT) in Utrecht

    Certified Data & Analytics Tester (CDAT)

    Tijdens deze praktisch opgezette training krijg je inzicht in het gestructureerd testen in complexe Data & Analytics projecten en omgevingen. Na de cursus ben je in staat op een doordachte teststrategie op te zetten voor data warehouse- , business intelligence- en analytics trajecten, deze uit te voeren en er in heldere woorden over te rapporteren.

    De inhoud van de training sluit aan op ISTQB en biedt een specifieke toepassing in de Data (Analytics) omgeving. De training leidt op voor het examen van Data Analytics United (DAU), dat thuis kan worden afgenomen via Brightest en PearsonVue.

    Deze 2-daagse training DAU CDAT (Foundation Level) wordt verzorgd in het Nederlands (met Engelstalige slides en artikelen) en de trainingsdagen lopen van 9.00 tot 17.00 uur (inclusief pauze). Voor een lunch wordt uiteraard gezorgd.

    De kosten van de training bedragen € 1195,- (incl. één Brightest examenvoucher).

    Zie voor meer details de beschrijving van de training.

    Neem voor meer informatie contact op met Armando Dörsek van Verified of schrijf je meteen in via onderstaand formulier:

      Ja, ik neem graag deel aan de volgende training:

      Neem alstublieft contact met mij op!

      Training “Certified Data & Analytics Tester” (CDAT) – Utrecht Live & Online

      Certified Data & Analytics Tester (Foundation Level)

      Ben je een tester of ontwikkelaar in een Analytics omgeving, of wil je daar juist deel van gaan uitmaken? Neem dan deel aan deze praktisch opgezette training, waarin je inzicht krijgt in het gestructureerd testen in complexe Data & Analytics projecten en omgevingen.

      Na het afronden van Certified Data & Analytics Tester (CDAT) kun je een  teststrategie opzetten voor data warehouse- , business intelligence- en analytics trajecten, deze uitvoeren en er in heldere woorden over rapporteren. Als je het examen haalt (1 uur, 45 vragen, 26 goed) dan ontvang je ook de bijbehorende Credly-badge.

      De inhoud van de training sluit aan op ISTQB en biedt een specifieke toepassing in de Data (Analytics) omgeving.

      Deze tweedaagse training wordt verzorgd in het Nederlands (met Engelstalige slides en artikelen) en de trainingsdagen lopen van 9.00 tot 17.00 uur, inclusief pauze. Voor diegenen die live deelnemen staat uiteraard een lunch klaar.

      De kosten van de training bedragen € 1195,- (incl. examen t.w.v.  € 190,-).

      Zie voor meer details de beschrijving van de training.

      Neem voor meer informatie contact op met Armando Dörsek van Verified of schrijf je meteen in via onderstaand formulier:

        Ja, ik neem graag deel aan de volgende training:

        Neem alstublieft contact met mij op!

         

        Training “Certified Data & Analytics Tester” – Nederlands gesproken – 25 mei 2021 e.v.

        Uitverkocht
        Deze training is VOL. Graag verwijzen we je naar één van de andere data.

        Tijdens deze praktisch opgezette training krijg je inzicht in het gestructureerd testen in complexe Data & Analytics projecten en omgevingen. Je kunt na afloop een teststrategie opzetten voor data warehouse- , business intelligence- en analytics trajecten, deze uitvoeren en er in heldere bewoording over rapporteren.

        De inhoud van de training sluit aan op ISTQB en biedt een specifieke toepassing in de Data (Analytics) omgeving.

        Deze training CDAT Foundation Level wordt verzorgd in 4 sessies in het Nederlands (met Engelstalige slides en artikelen) op de avonden van:

        • 25 mei 2021
        • 1 juni 2021
        • 8 juni 2021
        • 15  juni 2021

        Zie voor meer details de beschrijving van de training.

        Training “Certified Data & Analytics Tester” (CDAT) – English spoken – 3 On Line Sessions

        Data & Analytics United – Certified Data & Analytics Tester  (DAU – CDAT)

        As the volume and importance of data grows – we aim to increase the quality of information delivery. One way to achieve this is by having a well designed testing strategy for data centric products.

        “DAU Certified Data & Analytics Tester” is a training that has been designed for test engineers, test coordinators, business analysts, data leads, data warehouse developers and business intelligence consultants.

        Please join us during 3 online sessions (of 4 hours each) and prepare yourself for the DAU exam.

        Training days
        • 19-4-2021 18:00-22:00 CET
        • 20-4-2021 18:00-22:00 CET
        • 21-4-2021 18:00-22:00 CET
        Subjects

        After having drawn a common view on the world of Data & Analytics by exploring the concepts of Business Intelligence, Data Warehousing and Data & Analytics, we’ll dive into testing and test processes, focussing on the specifications required when testing data and analytics solutions. ​

        We’ll explore Risk Based Testing and the special skills that are required testers dealing with data and analytics.

        We will tackle the concept of how to test reports and dashboards, completeness, data quality as well as testing business rules and ETL (extract-, transform and load).

        ​We will explore differences between data & analytics testing and conventional testing, the specifics of Test Environments when dealing with Data & Analytics, Test Data, Ethics and Privacy Law (e.g. GDPR).

        Participants learn how to apply Data Profiling in order to create better test cases and to cover the most relevant Data Quality attributes.

        Exam

        This training will prepare candidates for the DAU Certified Data & Analytics exam (exam voucher worth € 200,- included).

        Price

        € 1.195 excluding VAT incl. Exam Voucher


          Yes, I'd like to enroll in the following course:

          Please contact me!

          Training “Certified Data & Analytics Tester” (CDAT) in Berlin/Online

          Data & Analytics United – Certified Data & Analytics Tester  (DAU – CDAT)

          As the volume and importance of data grows – we aim to increase the quality of information delivery. One way to achieve this is by having a well designed testing strategy for data centric products.

          “DAU Certified Data & Analytics Tester” is a training that has been designed for test engineers, test coordinators, business analysts, data leads, data warehouse developers and business intelligence consultants.

          Please join us during two days in Berlin (Germany) and get ready to take the DAU-CDAT exam.

          Training days
          • 10 May 2021  09:00-17:00 CET
          • 11 May 2021  09:00-17:00 CET
          ​Subjects

          After having drawn a common view on the world of Data & Analytics by exploring the concepts of Business Intelligence, Data Warehousing and Data & Analytics, we’ll dive into testing and test processes, focussing on the specifications required when testing data and analytics solutions. ​

          We’ll explore Risk Based Testing and the special skills that are required testers dealing with data and analytics.

          We will tackle the concept of how to test reports and dashboards, completeness, data quality as well as testing business rules and ETL (extract-, transform and load).

          ​We will explore differences between data & analytics testing and conventional testing, the specifics of Test Environments when dealing with Data & Analytics, Test Data, Ethics and Privacy Law (e.g. GDPR).

          Participants learn how to apply Data Profiling in order to create better test cases and to cover the most relevant Data Quality attributes.

          Exam

          This training will prepare candidates for the DAU Certified Data & Analytics Tester exam (voucher included).

          Price

          EUR 1.195 excluding VAT incl. Exam Voucher


            Yes, I'd like to enroll in the following course:

            Please contact me!

            Training “Certified Data & Analytics Tester” (CDAT) – English spoken – Berlin (Germany)

            Data & Analytics United – Certified Data & Analytics Tester  (DAU – CDAT)

            As the volume and importance of data grows – we aim to increase the quality of information delivery. One way to achieve this is by having a well designed testing strategy for data centric products.

            “DAU Certified Data & Analytics Tester” is a training that has been designed for test engineers, test coordinators, business analysts, data leads, data warehouse developers and business intelligence consultants.

            Please join us during two days in Berlin and prepare yourself for the DAU exam.

            Training days
            • 12-7-2021 09:00-17:00 CET
            • 13-7-2021 09:00-17:00 CET
            ​Subjects

            After having drawn a common view on the world of Data & Analytics by exploring the concepts of Business Intelligence, Data Warehousing and Data & Analytics, we’ll dive into testing and test processes, focussing on the specifications required when testing data and analytics solutions. ​

            We’ll explore Risk Based Testing and the special skills that are required testers dealing with data and analytics.

            We will tackle the concept of how to test reports and dashboards, completeness, data quality as well as testing business rules and ETL (extract-, transform and load).

            ​We will explore differences between data & analytics testing and conventional testing, the specifics of Test Environments when dealing with Data & Analytics, Test Data, Ethics and Privacy Law (e.g. GDPR).

            Participants learn how to apply Data Profiling in order to create better test cases and to cover the most relevant Data Quality attributes.

            Exam

            This training will prepare candidates for the DAU Certified Data & Analytics exam, for which you will receive an exam voucher (first exam included). You can take the online or paper based exam through Brightest.

            Price

            EUR 1.195 excluding VAT incl. exam Voucher


              Yes, I'd like to enroll in the following course:

              Please contact me!

              Training “Certified Data & Analytics Tester” (CDAT) – Online – English Spoken

              Data & Analytics United – Certified Data & Analytics Tester  (DAU – CDAT)

              As the volume and importance of data grows – we aim to increase the quality of information delivery. One way to achieve this is by having a well designed testing strategy for data centric products.

              “DAU Certified Data & Analytics Tester” is a training that has been designed for test engineers, test coordinators, business analysts, data leads, data warehouse developers and business intelligence consultants.

              Please join us in 4 weekly online sessions of ± 3 hours each and prepare yourself for the DAU exam.

              Training days
              • 18-2-2021 14:00-17:00 CET
              • 25-2-2021 14:00-17:00 CET
              • 4-3-2021 14:00-17:00 CET
              • 11-3-2021 14:00-17:00 CET
              Requirements

              You will need a laptop or tablet which supports a internet browser to join the MS Teams sessions. An invite will be sent shortly before the first training session.

              ​Subjects

              We’ll start to explore the concepts of Business Intelligence, Data Warehousing and Data & Analytics after which we’ll dive into testing and test processes, focussing on the specifications required when testing data and analytics solutions. ​

              We’ll explore Risk-Based Testing and the special skills that are required of the tester when dealing with data and analytics.

              We will tackle the concept of how to test reports and dashboards, completeness, data quality as well as testing business rules and ETL (extract-, transform and load).

              ​We will explore differences between data & analytics testing and conventional testing, the specifics of Test Environments when dealing with Data & Analytics, Test Data and Privacy and Ethics Aspects (GDPR).

              Participants learn how to apply Data Profiling in order to create better test cases and to cover the most relevant Data Quality attributes.

              Exam

              This training will prepare candidates for the DAU Certified Data & Analytics exam. An exam voucher can be purchased upon enrollment or after the training (EUR 200,-).

              Price

              EUR 995,- excluding VAT (21%).

              Exam voucher sold separately.


                Yes, I'd like to enroll in the following course:

                Please contact me!

                Online Training DaU “Certified Data and Analytics Tester” (CDAT) – English – special introductory price

                Starting 14 October 2020 Rogier Ammerlaan and Armando Dörsek will provide the online training for “Certified Data and Analytics Tester (CDAT)” against a Special Introductory Price* of EUR 685,- excl. VAT (incl. exam). 

                The training consists of four online blocks of 3 hours each, after which you should be able to take the Brightest DaU exam (1st exam included).

                The dates and times are:
                14 October – 14:00-17:00 (CET)
                20 October- 14:00-17:00 (CET)
                4 November- 14:00-17:00 (CET)
                11 November- 14:00-17:00 (CET)

                Note: These times will enable people from the Americas and India  to attend, too.

                If interested, do not hesitate to contact us for more information and/or an application form.

                *) Special Introductory Price applicable for new entries only, on and after 8 October 2020, booked through Verified.nl. This training was held often in a live setting before (at EUR 995,- ex. VAT, exam at EUR 190,-) and this is the first online version of the CDAT training.  

                Training “Certified Data & Analytics Tester” (CDAT) – Noord Nederland (met Startgarantie)

                Certified Data & Analytics Tester (Foundation Level)

                De training CDAT op 18 en 19 mei in Drachten heeft Startgarantie!

                Ga je een datawarehouse, data-migratie, dashboard of rapportage testen? Bereid je dan voor op deze klus en volg de training Data & Analytics Tester (Foundation Level).

                Tijdens deze praktisch opgezette training krijg je inzicht in het gestructureerd testen in complexe Data & Analytics projecten en omgevingen. Na de cursus ben je in staat op een doordachte teststrategie op te zetten voor data warehouse- , business intelligence- en analytics trajecten, deze uit te voeren en er in heldere woorden over te rapporteren.

                De inhoud van de training sluit aan op ISTQB en biedt een specifieke toepassing in de Data (Analytics) omgeving. De training leidt op voor het examen van Data Analytics United, dat thuis kan worden afgenomen via Brightest en PearsonVue.

                Deze 2-daagse training DAU CDAT (Foundation Level) wordt verzorgd in het Nederlands (met Engelstalige slides en artikelen) en de trainingsdagen lopen van 9.00 tot 17.00 uur (inclusief pauze). Voor een lunch wordt uiteraard gezorgd.

                De kosten van de training bedragen € 1195,- (incl. één Brightest examenvoucher).

                Zie voor meer details de beschrijving van de training.

                Neem voor meer informatie contact op met Armando Dörsek van Verified of schrijf je meteen in.

                  Ja, ik neem graag deel aan de volgende training:

                  Neem alstublieft contact met mij op!

                  Data & Analytics Testing Tools Summary (2)

                  Testing of Data & Analytics environments is increasingly supported by test tools. At Verified, we wonder what your experience is with one of these tools – and if you propose certain tools that have helped you ?

                  Please let us know, by leaving a message under this post, through the Contact FormLinkedIn or Twitter or contact us in any way you prefer. If you like, we can reply with a listing of more than 15 tools that should help testers and analysts  in testing and quality issues. 

                  The previous post on Test Tools can be found here.

                  Note: We are not sponsored in any way by the tool providers.

                  RTTS QuerySurge

                  Querysurge is a commercial product made for testing ETL, where comparison of source and target tables is the main object.

                  It offers dashboards, query wizards, various prefabricated queries for comparison of data, text search etc.

                  A trial version is available: https://www.querysurge.com/compare-trial-options

                  Note that for BI Testing, an add on is available. Connectors are available for Cognos, Tableau, SAP Business Objects, Microstrategy and OBIEE.  Features include comparison of reports when upgrading or migrating, querying report meta data.

                  ICEDQ

                  iCEDQ is a commercial ETL Test Tool, which runs in the cloud, on premise or hybrid.

                  It offers a rule engine module, which features:

                  • Checksum rules (e.g. row counts)
                  • Reconciliation rules (e.g. missing records)
                  • Validity rules (SCD Type 2, Format Checks)
                  • Script rules (e.g. execute DML/DDL, pre and post actions).

                  iCEDQ currently integrates with Jenkins, TFS, Jira, HP ALM and ServiceNow.

                  iCEDQ offers trials through their website:

                  https://icedq.com/download-icedq-trial

                  Bitwise QualiDI

                  QualiDI focuses on ETL testing from sources to targets.

                  It provides data validation and comparison of data, has a centralized repository of requirements, test cases, test results and has a API trigger based automation of test execution.

                  There is Business Rules Engine, enabling non-technical users to write rules instead of queries. It integrates with Bitwise Test Data Management Suite for providing test data.

                  QualiDi is a commercial product, a 30 day trial is available

                  Source: https://www.bitwiseglobal.com/innovations/qualidi/

                  SSIS Tester

                  SSIS Tester is offered by bytesoftwo and focuses on testing Microsoft SSIS.

                  SSIS Tester is a testing framework built on top of SQL Server Integration Services. It enables you to test packages, tasks and precedence constraints. It supports two types of tests:

                  1. unit tests
                  2. integration tests.

                  SSIS Tester helps you to develop you ETL process in the test driven manner and to identify errors early in the development process. It offers “data taps” and fake source or targe components, enabling developers to work independently. Tests target packages or particular tasks.

                  More information on SSIS Tester on MSDN can be found here:

                  http://msdn.microsoft.com/en-us/magazine/dn342874.aspx

                  SSIS Tester offers a free 21-day trial version:

                  http://www.bytesoftwo.com/download.html

                  Next 4 Tools

                  In a few weeks I will publish the next Blog post, asking for new experiences. Thanks for reading this post!

                  Omdat deze tekst gericht is op een internationaal publiek, heb ik ervoor gekozen om de bovenstaande tekst in het Engels op te stellen. Stelt u prijs op een vertaling naar het Nederlands? Neemt u dan aub contact met ons op.

                  Data & Analytics Testing Tools Summary (1)

                  Introduction

                  During the training Certified Data & Analytics Tester (Foundation Level) we provide a list of tools that can be used in testing or monitoring Business Intelligence and Data Warehouse products.

                  Below you can find the first 4 tools of this list, with a short description and links to the tool providers.

                  We are curious for your experiences when using these tools. What have they brought to your (testing) team? What is a Big Plus of using the tool – and with which features are you less impressed? Is an active user community available, or a help desk that makes a difference? Please let us know, by leaving a message under this post, through the Contact Form, LinkedIn or Twitter.

                  Note: We are not being sponsored by one of the products or their integration partners.

                  Data Cleaner

                  Data Cleaner is software which is used for Data Profiling.

                  It can also be used in the process of creating logical and physical test cases as it offers extensive support to analyze and browse through data sets.

                  Data Cleaner also offers features to clean, transform and export data sets. Rules can be built and reused.

                  Data Cleaner is free, additional (paid) features are available.

                  It can be downloaded to your desktop (Windows, Mac) environment from sourceforge or the community website:

                  https://datacleaner.org/

                  SAS Unit

                  SAS Unit is a framework for the unit testing of SAS programs (SAS EG 9.2, 9.3, 9.4).

                  Amongst others, SAS Unit offers features to assess test coverage of SAS Macros, running tests in batch mode, performance testing of SAS programs.

                  SAS Unit uses assertions to test the values of macro variables, contents of SAS data sets, relations between data sets, the existence of files incl. log files and – messages.

                  SAS Unit is open source software and built from SAS macros.

                  https://sourceforge.net/projects/sasunit/

                  Informatica DVO

                  Informatica offers a testing solution called Data Validation Option for PowerCenter (DVO).

                  The DataValidator for PowerCenter product was originally developed by a company DVO SOFTWARE. It is now available under the Informatica brand as Informatica PowerCenter Data Validation Option.

                  DVO is a custom tool built on top of Informatica PowerCenter. DVO integrates with the Informatica PowerCenter Repository and Integration Services and enables developers and business analysts to create rules to test the data being transformed during the data integration process.

                  Informatica mentions the following ETL Testing Use Cases: Production Validation Testing, Source to Target Testing and Application upgrades.

                  Amongst others, it features:

                  • An easy-to-use GUI interface to test the rules created for data validations for multiple projects.
                  • No programming skills needed to create validation tests.
                  • A repository with reporting capabilities to provide a complete audit trail of all tests and their results.
                  • Reading data definitions from PowerCenter metadata repositories and can easily deal with data definition changes.

                  Note that DVO tests data only, not mappings or workflows. Testing mappings is unit testing, which is different from data validation.

                  As DVO is a commercial product it may be attractive to Informatica Power Center users. At this moment, no free trial seems available (Feb 9, 2020).

                  www.informatica.com

                  Datagaps ETL Validator

                  Datagaps offers several products, e.g. ETL Validator, BI Validator, Data Flow and Test Data Manager.

                  The Datagaps ETL Validator comes pre-packaged with an ETL engine capable of extracting and comparing millions of records from multiple data sources while executing test cases in parallel.

                  ETL Validator also has a Visual Test Case Builder with drag-and-drop capabilities and a Query Builder that enables defining tests without manual typing.

                  Key features include enterprise collaboration, flat file testing, data profile testing, baselining, data quality testing, and database metadata testing.

                  Datagaps offers a free 30-day trial.

                  https://www.datagaps.com/etl-testing-tools/etl-validator-download

                  Next 4 Tools

                  The next page shows another 4 tools.

                  Omdat deze tekst gericht is op een internationaal publiek, heb ik ervoor gekozen om de bovenstaande tekst in het Engels op te stellen. Stelt u prijs op een vertaling naar het Nederlands? Neemt u dan aub contact met ons op.

                  Training Data & Analytics Testing (CDAT) – Zwolle

                  Tijdens deze praktisch opgezette training “Certified Data & Analytics Testing” krijg je inzicht in het gestructureerd testen in complexe Data & Analytics projecten en omgevingen. Je kunt na afloop een teststrategie opzetten voor data warehouse- , business intelligence- en analytics trajecten en deze uitvoeren.

                  De inhoud van de training is gebaseerd op o.a. ISTQB met een specifieke vertaling voor de Data (Analytics) omgeving.

                  Deze 3-daagse training wordt verzorgd in het Nederlands (materialen in het Engels). De trainingsdagen zijn van 9.00 tot 17.00 uur (inclusief pauze). Voor een lunch wordt uiteraard gezorgd.

                  De kosten bedragen € 1495,- (incl. examen á € 200,-).

                  Zie voor meer details de beschrijving van de training.

                  Neem voor meer informatie contact op met Armando Dörsek van Verified of schrijf je direct in via de website van trainingspartner Ammerlaan IT Training & Advies.

                  Training Data & Analytics Tester (CDAT) – Utrecht

                  Tijdens deze praktisch opgezette training krijg je inzicht in het gestructureerd testen in complexe Data & Analytics projecten en omgevingen. Je kunt na afloop een teststrategie opzetten voor data warehouse- , business intelligence- en analytics trajecten, deze uitvoeren en er in heldere woorden over rapporteren.

                  De inhoud van de training sluit aan op ISTQB en biedt een specifieke toepassing in de Data (Analytics) omgeving.

                  Deze 3-daagse training wordt verzorgd in het Nederlands (met Engelstalige slides en – artikelen). De trainingsdagen zijn van 9.00 tot 17.00 uur, inclusief pauze. Voor een lunch wordt uiteraard gezorgd.

                  De kosten van de training bedragen € 1495,- (incl. examen á € 200,-).

                  Zie voor meer details de beschrijving van de training.

                  Neem voor meer informatie contact op met Armando Dörsek van Verified of schrijf je meteen in op de website van trainingspartner Ammerlaan Training en Advies.

                  Testen & Toetsen in de wereld van BI, Datawarehousing en Analytics