-
IF then ELSE Logic
Can I write IF (Some Condition)THEN (Valid) Else (Invalid) Logic in Aperture? So That in the Output Result set I can view both Valid and Invalid Result set
-
Has anybody used Aperture to anonomise/create synthetic data?
I have a requirement to create anonymised/synthetic data My initial thought, for the name, would be to swap forenames with a simple lookup table and create surnames by a simple letter replacement (this would be consistent where a client name may appear several times) Similarly, create false emails (using the…
-
How do we expand on the provided base blocking keys and rule sets on Aperture Data Studio?
Hi team I am asking this question on behalf of one of our Credit Services team members. They are looking to create Blocking keys and Rules that accommodate for a text string (Drivers License) and date of birth, but is running into issues. The rules that they are currently using are part of the attachments. Is there an easy…
-
Explore option on Report Object
Hi, Am using a report object for dashboard and the explorer is not showing to be able to filter
-
Single Customer View (SCV) user research interview invitation
Hi all, At Experian we are always looking to improve our technology and processes to give you the best customer experience possible. The Single Customer View (SCV) use case is one of the key focus areas for our Product Team, and we need your help. We would like to understand the jobs that need to be completed by different…
-
Creating new schedule with cron expression ?
Hi, I try to make new schedule every month for the time intervals from 10th to 15th and 25th to 30th. Schedules --> Create schedule --> Could someone show me how I can utilize cron expression? Or are there any other tips to do that. I know what cron expression syntax mean, I just need some example how to do that. Can…
-
XML files
HI, I'm trying to read selected fields from an XML file pulling out, for example, text between the <email>.........<email> indicators. Can Aperture do this? eg by using Substrings, pointers and arrays in some way? Nige
-
How to Execute Parallel Workflows (without Custom Steps or Powershell)
Hi all, The Setup: Aperture V2.2.8 on Windows Server 2012 R2 Current Situation: I have to run 8 different Workflows on a Daily basis, but some of those workflows share the same tables. Current Solution: I created a Workflow that has the "Allow auto-refresh" turned on and has all the tables that I am going to use on the…
-
Convert Pass and Fail to 0 and 1
Hi, Is the a quick way to transform Pass and Fail validations to numbers 1 and 0 respectively Regards,
-
How to convert unix timestamps to human-readable date?
Hi, Does any one know how to convert Unix timestamps to human-readable date? Example Unix timestamp: 1568732257 --> wanted format: 2021-09-17 14:57:37 (Year-Month-Date Hour:Min:Sec)
-
4009: Exception occurred
When filtering on a field I get an error 4009: Exception occurred
-
XML driver for Aperture
Hi, Does anybody on the forum know if it is possible to read semi-structured XML data into Aperture? If so, how do we do this? (and is it 'out of the box'?) Nige
-
ℹ️ Scoreboard Charts for Dashboards
Charts can be created as bar charts, line charts, pie charts etc. These charts can be collated together to form a Dashboard that users of the Space or Consumer users can use to monitor progress and data trends. There is a type of chart called a Scoreboard, which can be used to highlight important values or metrics A…
-
Notifications in Aperture - email content
Hi everyone! I have a question related to the email content on the notifications section. We are trialling this so that we can notify people in the wider community when their rules have been run, and if there are any failing rows. After some trial and error, I've found that at the moment the only way to get the email…
-
Calling subroutines
Hi, When there are more that 1 outputs from a subroutine is it possible to distinguish them in the calling program? Also, is it possible to retain the zoom level rather than it being reset to the default all the time? Thanks Nige
-
Aurora database connection?
Hi, Our client MDM has been built on an aurora database platform. Is there a connector for Aperture to this source? Thanks Nige
-
Salesforce - User connection properties
Hi all, If we add a new column to a Salesforce table and refresh the data, it doesn't bring the ones that are recently added. To be able to get the latest schema of the tables, we can add a connection property: Are there any other properties that are useful while creating the Salesforce External system connection?
-
Find Duplicates Language for Notepad++
If you're using Notepad++ as your text editor, it can be helpful to have an interpreter to highlight key words for matching rules for Find Duplicates and have auto completion suggestions. Attached is are two xml files which allow for this. Adding Language To add the new language option, open Notepad++ and select Language…
-
Data source and "Filter regular expression" - How it works?
Hi, does anyone know how System -> External systems -> "Filter regular expression" function works? I added Schema and Table Pattern and they worked fine, but I don't understand how this Filter regular expression have to add, if I have to do some special definitions. Connection type is JDBC. Can anyone help or know some…
-
Undo and Redo
In Aperture Data Studio Version 1, we had the Undo and Redo menu options while creating or editing Functions. This was however not available while creating or editing Workflows. In the earlier days of Version 2, while there was the concept of Actions that you can remove or apply easily within the View creation or grid…
-
Scheduling in Aperture Data Studio Version 1 vs Version 2
Many Aperture Data Studio customers process large volumes of data. Workflows are likely to be scheduled as overnight jobs or scheduled to be run over the weekends so that much of the waiting time happens outside the core business hours. There may also be a need to run multiple workflows, one after another automatically. In…
-
Transform node in Aperture 2
Hi, I'm trying to use the Transform step in Aperture 2 and am finding that, with the limited number of columns viewable in Aperture 2 (10/20/50 rows etc) compared to the whole table being visible in Aperture 1, handling wide tables is particularly fiddly eg simply moving a column from the end of a wide table to just after…
-
Identification of an individual flat in a building using Address Validate
Hi, I've checked the following address (which is valid with an appropriate postcode) using Address Validate Addr 1 :16 Wythburn Court Addr2 : 34 Seymour Place Addr3 : London Postcode : W1H 7NS However, because the "Flat" is missing from the address, the 16 at the front is ignored and I get a "3-tentative full match"…
-
Date formatting
I've lost the month in this date format - any suggestions? (not a lot to go on in the help here) Thanks @Nigel Light
-
Snowflake
We have been experimenting with Snowflake technology and were wondering whether anybody has experience of using this as a source for Aperture. Nigel
-
SAML anyone?
Hi everyone. Our engineering team is currently working on a feature for Single sign on enterprise authentication with SAML. There are some technical questions that they are seeking some input on in order for us to best address our client's needs. Which Identity Provider (IdP) do you use for SAML? Does the Identity Provider…
-
Sharing of useful functions to support contact data validation
Hi all, I've found myself using a few functions on a regular basis so thought I'd share them on here in case they are of benefit to anyone else. The attached .dmx file contains the following functions: Free from PCI info (checks an input does not contain 16-digit numbers either together or separated by hyphens or spaces…
-
Using Format Phone Number
I was wondering how everyone uses the function Format Phone Number in Aperture. For myself, I am to keep it simple but wish to avoid error messages being shown for non-phone numbers. Below is a set of various data that may exist in the customer's dataset: If you use Format Phone by itself, you will receive some errors for…
-
What is the best way to check for the presence of a number value in a field?
We often have a requirement from clients to check for numbers in a name field or for name in a phone number field. What is the easiest way to do this?
-
Extracting titles and honorifics from names (advanced)
In this earlier post I described how you might want to extract titles or honorifics from a name string, this might be important for a marketing campaign where to you wish to personalize your communication by saying dear sir or madam. Unfortunately the data isn't always as we want it and sometimes we land up with names of…
-
Automated Data Tagging in v2.1
In v2.1 you now have the ability to automatically tag your data using the predelivered data tagging fingerprints. You can of course still manually tag your data but automatically adding data tags hopefully makes the data tagging process easier and certainly makes using the email phone and address validation steps easier.…
-
Data Stewardship Priorities
When you think about Data Stewardship, what does that really mean to you and what do you think matters most? Before you answer the poll, you might want to do some research, it's a jungle out there! https://www.dataversity.net/what-is-data-stewardship/ Either way, in the context of what you do, what you know and what you…
-
Extract Title from Names
During the course of the data preparation that you might be doing on names, you might have occasion to pull out the Title or honorific associated with the name. In this example we assume all names have an honorific, if some don't you need to use conditional logic which i will cover separately. You have at least two…
-
Datasetdropzone directory changes in v2.1.2
Hi all, I’m writing this post to inform you all on the upcoming change to the datasetdropzone directory in v2.1.2. Prior to v2.1.2, the datasetdropzone folders are only segregated by environment. In v2.1.2, the datasetdropzone directory is now segregated by space as well. What changes can you expect for the datasetdropzone…
-
Export directory changes in v2.1.2
Hi all, I’m writing this post to inform you all on the upcoming change to the Export directory in v2.1.2. Prior to v2.1.2, there is only 1 export directory and all files generated from the export step will be exported to this single folder regardless of the environment or space. What changes can you expect for the Export…
-
Salesforce tables have different columns after upgrading from Data Studio v1 to v2
If you've previously used Data Studio v1 to load data from your Salesforce instance, and have upgraded to v2, you will have noticed that the way we show column names from Salesforce tables, and the columns we display by default, has changed. This is because the JDBC driver we use internally to connect to Salesforce has…
-
Upgrading to a new release for Aperture Data Studio v1 versus v2
Up until now, we have had 23 releases for Aperture Data Studio v1. In order to upgrade from one release to a newer one, for example from v1.5.0 to v1.6.3, there are a series of steps you need to follow: Check the job log for active or running jobs. Stop the services. Backup your repository by locating, copying and saving…
-
Obfuscate/unobfuscate
It might be a useful feature to have an 'obfuscation' node to run text fields through an obfuscation algorithm (for display purposes eg in reporting). It might also be useful to be able to reverse this by having an 'unobfuscate' node, using the same algorithm in reverse to allow the original data to be examined for DQ…
-
Tell us what we can do to improve your Aperture Data Studio administration experience
Hi everyone, We have created a survey for you to share with us your opinion on your Aperture Data Studio administration experience. Your answers will allow us to shape our product to better meet your expectations. The survey should not take more than 10 minutes of your time. Start the Survey Thank you very much for your…
-
Attention all V1 & V2 users...
Today we have launched a new sub category, Migrating from V1 to V2, which contains posts to aid your transition to V2, as well as access to a migration tool. We will be adding more posts over the coming days and weeks so stay tuned for tips and tricks to aid your migration. Please note, you must be signed in to the…
-
Triggering workflow executions upon file arrival in Aperture Data Studio v1 versus v2
Aperture Data Studio is commonly used for operational data quality management. This would mean that you are regularly pulling up-to-date information from multiple transactional systems used within your organization such that you can centrally perform data quality checks and improvements. This prepares the data for further…
-
Ensuring you always return the same duplicate ID with Find Duplicates
Find Duplicates returns consistent Cluster ID after each run. Consistency here applies to the sequence order of the input file. So if the same file is submitted to Find Duplicates, in the same order, then the Cluster ID will be the same on each run. In some cases you may not be able to guarantee the order of the records…
-
Profiling Data for the first time and changing the underlying data
This short video illustrates the ease with which you can define a dataset, profile the data to dynamically refresh via a view and then change the contents of the dataset without having to define it all over again
-
Edge
Hi, Overnight Edge (the company default browser) has been upgraded - does this improve things for Aperture or is Chrome still the recommendation? Regards Nige
-
Phased deployment and secure collaboration in Aperture Data Studio v1 versus v2
In Aperture Data Studio v1, there isn’t an effective way to compartmentalize objects such as datasets and workflows to allow phased development, testing, rollout and rollback in case of any issues. You also do not have granular control over what you choose to share with other users or define what each user can do. This…
-
Open Data Sources by Country
A few open data sources. Feel free to add more. United States: http://www.data.gov/ Austria: https://www.data.gv.at/ - municipality and state level Germany: https://www.govdata.de/ The Netherlands https://data.overheid.nl (in Dutch) Argentina http://datos.gob.ar European Data Portal: http://www.europeandataportal.eu/…
-
Fixed Width file parsing in Aperture Data Studio v1 versus v2
Fixed width files may be one of the common file formats that you have come across and need to process. You will often find that this would be more challenging as compared to a comma separated or Excel file. Here is an example of a small section in a layout to be submitted to the IRS (Internal Revenue Service) of the United…
-
JDBC connection
Hi Can you please let me know if there is/planned to be a JDBC connection for Optymyze? Thanks Nige
-
Function to calculate the number of working days between two dates
Recently I wanted to create a function in Data Studio to calculate the number of working days between two dates, similar to Excel's NETWORKDAYS function. First, a couple of comments: Weekend days are defined as Saturday and Sunday. Being able to handle alternative weekend days (like the Excel NETWORKDAYS.INTL function)…
-
How to systematically sample with a fixed sample size and varying population size
I was recently asked how to build a Data Studio workflow that would output a systematically sampled set of records with a fixed sample size (e.g. 1000 records), where the number of records in the input will vary. Here's my approach, let me know if you have a better one! Requirement We want to sample every nth record from…
-
What is the status of data catalog integrations, for example Collibra?
What is the current status of Aperture to Collibra-integration? https://marketplace.collibra.com/profiles/experian/ Do you need to have Mulesoft ESB to integrate to Collibra as suggested in the marketplace documents or will there be also other ways to integrate these two products?
-
Adding a variable comment as a "call to action" for data quality issues found
This is an experiment on handling HR data quality issues using a validation workflow in Aperture Data Studio v2.X. I built some simple HR data quality rules for FTE, Gender and Employment Category. With a validation step, i am able to obtain the following results for each of the rules. I then want to add a comment as a…
-
Setting up custom Find Duplicates rules
@Carolyn asks I would like to set up a custom Find Duplicates - Not sure how to set up Find Duplicates Rules and Blocking Keys When you select the find duplicates step you tell it the fields (match columns) you need to use. I have 5 lines in my address. Line 1 and 2 make up the address - so might be Unit 1/28 and second…
-
AWS database connection issues to Aperture (v1.6.4)
Hi, Does anybody have any suggestions why a database can arbitrarily flip-flop between connecting then disconnecting to Aperture? (I'm using AWS and there are no similar problems using native SQL Server) It is baffling our technical team here and they are just wondering if it is an application issue (as nothing else is…
-
APIs in Aperture Data Studio v2
In v1 there were a number of APIs for the automation of UI as outlined here under the REST API guide Are there similar APIs in v2?
-
Grouping in V2 and differences with V1
if you're working with financial data with Aperture Data Studio, one of the things that you might want to do, is to rollup columnar values and produce aggregates. When you produce aggregates you don't necessarily want to see all the columns or all the rows. Certainly if you see all the columns, the challenge can be…
-
Reading files without headers - then scheduling the job
Hi We would like to use Aperture (currently using v1.6.2) to read a series of files that are sent on a regular basis. This files are supplied without headers. Building the original solution, we successfully manually loaded the files and applied headers using 'Preview and Configure', setting the '1st line header' flag and…
-
The title field is required
Body is 1 character too short.
-
Code going missing in a v1.x job?
Hi Are any other users experiencing the frustration of sections of code going missing? We (@Pavan Rao and myself) are currently using v1.6.2 (though we have experienced similar issues in previous versions) and have found links to Union nodes (and the whole node itself - plus subsequent Export nodes) frustratingly…
-
2020 Global data management research
Experian recently released its annual global data management benchmark report. This year we surveyed over 1,100 data practitioners and business leaders globally to review data quality and data management trends. We tried to understand if organizations are indeed data-driven and found that most still have a long journey…
-
Using a network path for the server import or export directory
It may be desirable to have Data Studio v1 or v2 import and export data from a network path. This will allow users who do not have access to the physical machine have access to their relevant data folders. Note: Doing this incorrectly may cause Data Studio to fail to start up, or fail to read/write from those folders. Use…
-
Data Views - an abstracted view of your data sources
An entirely new concept of data views based on datasets is introduced in V2.0. Views enable designers of workflows and Space administrators to distill and abstract data from datasets in such a way as to ensure that other workflow designers, data consumers and users of a given Space or View, can only see curated data. Data…
-
Handling Gmail email addresses for duplicate identification
The problem: Gmail allows multiple e-mail addresses for a single account With over 1 billion active users since 2016, Gmail is likely to be a large part of any consumer dataset. Being able to accurately resolve these individuals becomes more problematic when Gmail allows for variants of an e-mail address to be used. An…
-
User Spaces - a way to organize your work
An entirely new concept of user spaces is introduced in V2.0. Spaces enable designers of workflows to categorise their work and group objects together in such a way that they can create self-contained Data Studio solutions where all related objects are packaged together. Spaces are akin to project folders but are also able…
-
Environments - giving your Data Studio installation more flexibility with sandboxes
One of the characteristics of business-critical systems is that they need to be secured in such a way that experimentation and sandbox type activities as well as new development work, is isolated from the operational daily use of the system. With v2.0 a new feature is introduced to support the implementation of…
-
Democratizing and controlling data access in v2.0
Data Studio v2.0 provides a sophisticated and secure architecture for exposing data to the enterprise in a controlled way from Excel spreadsheets, .CSV, .PSV, .SAS7BDAT, .TXT. Datasets can also be loaded from cloud storage environments like Azure blobs, Amazon S3 and Google Big Query and RedShift and Bigdata environments…
-
Connecting to an SMTP email server that does not require authentication
Data Studio lets you configure an SMTP server connection, which can then be used to send email notifications when particular events occur. Some SMTP servers do not require authentication: username and password details do not need to be provided. In this case, when you click the "Test SMTP Configuration" button after…
-
Scheduling workflows using the Data Studio REST API
Aperture Data Studio is intended to be invoked from third party schedulers , like Windows Task Scheduler, Cron or even triggered from workflows within other applications. The method for doing this is to use the Aperture Data Studio REST API to execute a workflow from within a script, and trigger that script from the…
-
Hello world...
🥂 our developers have been hard at work and we are stoked about this new release. Prod your account manager to learn more!
-
Inferring Gender from Title or Honorific and Gender Reversal
Sometimes it might be useful to determine the gender of a person from their honorific, so for example deciding that the greeting should be Sir or Madam depending on the Mr or Mrs or Miss honorific. The interesting aspect of this inferral task is that in Data Studio there are a good many ways that you can achieve this…
-
Impact of snapshots on memory
Hi, I am using a job that has several input snapshots defined in it, though I only connect 2 of them each time I run the job. it seems to be using a lot of storage (and slowing Aperture down) - is this correct? Something to do with caching? I am thinking, keep the number of snapshots in a job to a minimum. Does this apply…
-
Data Studio v1.6.2 is now available!
This maintenance release resolves several issues caused when executing workflows using the REST API. Try it out. New server settings provide the ability to disable connectivity and permissions checks on a JDBC source when using a loaded data table. Find out more.
-
Is it possible to 'pull' data from Aperture Data Studio into Excel?
Hi, Whilst there is the option to export data from Data Studio into Excel, is it possible to pull data into Excel from Data Studio using an edit-bar drop-down? If so, are there any licence considerations to do this? Thanks Nige
-
Approach for Matching Product data (product description)
Yesterday @MiteshKhatri @Akshay Davis @Katya Jermolina and I were having a discussion about approaches for matching product name information using Data Studio, so I've put together the below summary of the approach in case it helps anyone else. Before I go through the steps I want to call out that there are a variety of…
-
Extracting the country from an unstructured address using a lookup list of countries
The problem that you may need to solve is that you have data that is effectively a contiguous string of unstructured data and you want to determine perhaps the country from that data in order to either Address Match or filter your records. A basic approach is to simply use an extract Matches function within the Transform…
-
Accesing data from quick sense files
Hi everyone, I'm new in data aperture and I have a client who calculates KPIs in a quick sense report that he would like to centralize in a automated way. Is it possible to do it with Aperture Data Studio? Regards, Alfredo
-
Cross-system and cross-table relationship discovery automation
In our customer cases data quality issues are often cross-system inconsistencies. Technically, customer rows might not have been brought from CRM to MDM or DW for some technical or business reason. Additionally, data might have been removed or changed in one system leading to inconsistencies between systems. Automated…
-
Sequence count
Hi, I can't find a sequence count in Aperture - is there one planned in Version 2.x? Nige
-
Data Studio and MDM
It sometimes gets asked how we co-exist with an MDM system. We have many customers using Data Studio alongside their MDM - to improve it, monitor it and even to help build it in the first place. Here's a summary: Data Studio has no permanent data store. It is designed to ingest, enrich, monitor, cleanse, de-dupe and verify…
-
Use of SQL to refine source data prior to loading into Aperture
Hi, I was just wondering whether other users would find a Pandora-like feature to pre-filter prior to loading data useful? (I certainly would - especially when loading large database tables only to immediately filter-out the majority of the historic data) Similarly, who would find running SQL queries to load data into…
-
Availability of source file/table metadata
Hi, We are using Aperture to compare profile statistics from 2 tables from different SQL databases following an upgrade to SQL Server. Is the source database/table name available so we can include it in a report (without having to create a constant value each time we change sources)? Thanks Nige
-
Extracting a specific piece of information from an address using regular expressions and lookups
Data Studio has several built-in functions to allow you parse strings of text and extract a particular substring that contains the information you need. A common example would be to extract a postal code, state or country from a postal address, where the whole address is a single comma-separated string. Typically this…
-
How to trigger an e-mail notification if a job fails
Data Studio has a range of default notification options which allow you to send a custom email to defined users when certain important events occur. One common use for this is to have Data Studio send an email when a workflow execution fails. To do this we'll be using Notifications and Event Data Items. First, you'll need…
-
Checking data like social security numbers (SSN) for correctness
Social Security Numbers, National Identity numbers and the like, on face value are simply strings of data, made up with letters or numbers and segmented with hyphens. There is however inherent intelligence in many of these identifiers and even check digits. You'll find regular expressions to validate these kinds of…
-
Checking data like social security numbers (SSN) for correctness
Social Security Numbers, National Identity numbers and the like, on face value are simply strings of data, made up with letters or numbers and segmented with hyphens. There is however inherent intelligence in many of these identifiers and even check digits. You'll find regular expressions to validate these kinds of…
-
Sharing database connections across teams
What's the recommended approach to sharing tables/views and connections to databases across a wider team? Should we create new connections per user, share tables with specific users or create credentials for other users on the same connection? I'm interested in the security aspect but also the performance impact of each of…
-
Getting started with Regular Expressions in Data Studio
A regular expression (from here on we’ll call them “regex” for short) is a well-known and powerful syntax for describing a search pattern in a simple string of text. You can think of regex as a way of defining advanced wildcards: In Windows file explorer you might search for *.txt to return all txt files, or in SQL you…
-
Tips to improve workflow performance
If you ever find that a workflow is taking a long time to execute, you might want to think of ways to improve it. Below are some practical tips and examples that will help you to achieve the best performance out of your workflows. Workflow Structure First of all, it is advisable to break down large workflows into smaller…
-
Using a discriminant for Find Duplicates clustering
Find Duplicates will use fuzzy matching to link records, however, in some cases you may have a discriminant field you wish to use to break clusters. In this post we cover how to make use of these within a workflow. The simple scenario In this scenario, we are processing transaction records and matching on name, mailing…
-
Data Studio v1.6.0 is now available!
The latest version of Data Studio introduces new features, existing functionality enhancements, performance improvements and a number of bug fixes. Try it out. Some of the new features include: A number of enhancements to the Validate addresses step: Ability to select predefined custom layouts for output columns, giving…
-
Creating a new data connection to Salesforce from Data Studio via JDBC
It is likely that quite soon after you start to use Data Studio you will want to connect Data Studio to databases and systems of record. This is quickly and relatively easily done using the JDBC connector component delivered as a part of the application. Details of how to add a custom JDBC driver are explained in the…
-
Changing source file schemae and data definitions
Legacy system files often come without a schema or data definition embedded in the file, and if there is one, it might not be one that you are particularly happy with. One of the capabilities of Data Studio, is the support of using a Data Definition Language (DDL) file or COB file to accompany your file upload. DDL and COB…
-
JDBC Performance issues can simply be about security
In Data Studio you have a number of different ways of accessing data. File Based methods Data connection based methods (JDBC or special connector) API or programmatically Leveraging snapshot data JDBC is a high performance way of getting data into Data Studio fast! Getting JDBC up and running is straightforward and…
-
Refresh your sources
Occasionally you may find that the staged sources fall out of synch in how Data Studio renders them in the UI. If you feel this is the case, you can explicitly refresh the sources from the UI by right mouse clicking on the ☰ under the My Files area and selecting refresh
-
How to resolve a unmapped source on a workflow
In a previous tip I suggested you might want to export and import your workflow between systems. I also mentioned that one of the prerequisites for this to work seamlessly, was for you to have the same or similar configuration and sources defined in both systems. It will happen, that there are source differences. So how do…
-
Data Wrangling - is it so bad?
An interesting perspective from Pete Aven of Marklogic, popped up in my feed this week, written on Medium.com and enitled "Data Wrangling is Bad"; Aven describes how potentially we're all Data Wranglers, that it is not a good thing, should not be embraced or accepted. In reality though, do we have a choice?
-
Why might my Find Duplicates results look different?
If I use the Find Duplicates step on its own, in some instances I get more clusters of records (clusterIDs) than if I use Data Studio with the Address matching step. I attach an example that illustrates this using the test data delivered with the application Why would that be?
-
Do businesses run on premium data? New study assesses variables in data quality tools
Lisa Ehrlinger from Johannes Kepler Universität Linz Linz, Austria, and her team have identified 667 data quality tools on the market, and they have narrowed that number down to 13 for detailed testing and analysis based on their domain independence, non-specificity, and availability free or on a trial basis. While the…
-
Jobs with large lookup files
At St. James's Place, prior to highlighting whether client details need to be quality checked, we need to establish whether the client has a current fund holding (as this is where we realise maximum business benefit in correcting the data) To do this we are faced with loading very large table(s) ie >100Million rows which…
-
Dealing with PII
Talking to a prospect today about PII - identifying it on data loaded into the system and then processing it - eg reporting its presence. Comments appreciated Steve