SlideShare a Scribd company logo
1 of 17
What is Kleber?
A non technical look into the
platform that will change the
way you think about
data quality

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
What is DataTools Kleber?
Let’s begin.

Capture

First things first. If you are a technical person please contact us and we will send
you all the architecture, methods calls, functions, services information and
otherwise foreign to non technical people jargon you can handle. If you are a
business owner, operations manager, marketing guru, non technical PM or
otherwise just wanting to get things fixed, read on!
Kleber is the culmination of almost 20 years providing data quality software. It
binds together all the discrete elements required to deliver true data quality into a
single platform. Further still it harmoniously unifies each of the commonly required
individual processes into a single process that can be easily implemented from
within a website or application.
To delver true data quality, the process outlined in this document can be
implemented in real-time for on boarding of information and in a batch method to
process all the data already residing within your organisation.

Create

Parse

Enhance

Verify

Match

Repair
Format

Read on to learn about each step in the Kleber process and how these come together as a complete data processing solution.

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
“I want quality data. I’ve been told I need a Data Capture solution.”
It’s a great place to start.
“To assist in fast and accurate capture of contact information”

Data Capture solutions make the capture of contact information within your websites and
applications quick and easy. It’s also a great place to get into DataTools Kleber.
Predictive Search technologies (otherwise known as Type Assist) leverage advanced keystroke
reduction capabilities to predict the information being entered and minimise the number of
keystrokes required.
This makes entering of information not only faster and more accurate but also more
pleasurable for the user, reducing abandonment and enhancing the user experience.
Because Predictive Search is highly intuitive it requires no user training and can be used by new
users with ease.
Kleber offers predictive capture for physical addresses, first names, surnames, business
names and email addresses.

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
“I already use data capture. How is Kleber of any benefit to me?”
Handling Exceptions.
“Every data capture process creates exceptions that must be handled. This is
typically the most complex and costly aspect of any data capture project”
Despite advanced logic being used in Predictive Capture technologies
not all data will be captured correctly.
Even the most sophisticated data capture solutions will have some
percentage of error or information that could not be captured.

Data Capture solutions have been sold as the
golden panacea for data quality issues. The
reality is these technologies only handle the
“simple fixes”.
True data quality requires a mechanism for
handling the few percent of truly complicated
expectations these technologies create.

It is this data that pollutes databases and causes complex and costly
exception handling process throughout the information workflow.
DataTools Kleber makes handing of these exceptions easy by ensuring
even poorly captured data is repaired and cleansed before being stored.

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
“OK. So I can fix the exceptions, but that’s still a lot of work, right?”
Wrong! One Call does it all.

Automated exception handling via

“Handling exceptions is as simple as making a single call to Kleber .
Kleber will then repair and clean the data turning it into useful information
and return it to you”

• Parsing to split the captured data into
discrete fields

All of the individual and complex data processing methods have been
combined together in a single function within Kleber.
By calling this function the data will be parsed, verified, repaired and
formatted. Match keys will also be appended as will various other types
of information to provide insight into the data and deliver operational
efficiencies.

You can call each of these function individually if you want to, but all the
hard work has already been done, so why not leave it to us.

• Verification of data to ensure its accuracy
• Repair of data through results of verification
• Formatting of the data to ensure it is
suitable for use
• Matching to compare data and find
duplicates

Using Kleber at this point will eliminate the need for complex downstream exception handling.

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
“I don’t have Data Capture technologies. Do I need to implement these first?”
No. It will provider a better user experience, but you don’t have to use it.
Simply send the data as you store it now.
“Use your standard data entry forms to capture the data and send it as is or process all the data in your database as a batch.”
If you don’t have a data capture solution that’s OK. You can use your
existing fields for users to enter data. Once all the data has been entered
simply send it to Kleber and get the clean, verified and nicely formatted
result.
Data Capture technologies simplify the entry of data and make the
process faster and more pleasurable for the user.
We can provide these too!
Kleber can also be used in a Batch fashion whereby an entire list or
database can be processed.

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
“So how does it work?”
It’s complicated – Well it is: but we’ve made it simple.
“There is a logical process through which all data can be taken, each step of the
process building on the previous on to ultimately deliver true data quality and insight .”

Capture
Data is captured,
whether it be
though Predictive
technologies or
traditional data
entry.

Parse
Parsing breaks
the data down into
its smallest parts
ready for other
downstream
processes.

Copyright DataTools 2014

Verify
Parsed data is
compared to
“sources of truth”
to verify the
accuracy of the
information.

Repair
Results from the
parsing and
verification are
used to repair the
data, making it
clean and useful.

www.datatools.com.au

Format
The cleansed data
elements can then
be put together in
a variety of ways
to ensure the data
meets specific
downstream
requirements.

Match
Match Keys are
appended to the
data in order to
compare records
against each other
and find
duplicates.

Enhance
Other information,
is appended to the
data in order to
deliver true insight.
E.g.
Geospatial data.

Addressing Data Quality
“Why do I need to parse my data?”
It all starts with parsing.
“Parsing is the foundation of quality data. Without quality parsing all other
data processes will fail.”
Parsing is the process of splitting data apart into its smallest parts in a way
that makes sense. It is typically done behind the scenes and, as such, its
importance is often underestimated.
Proper parsing is often “faked” by many vendors who simply take
variations of a record and reference this against a known source of truth
until they get a match. This is inefficient and highly dependant on the
quality of the source of truth.
Kleber provides true parsing. After all, DataTools has been delivering
advanced data parsing technologies for almost 20 years!

Ex

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
“What does verification involve?”
Referencing data against a “Source of Truth”.
“Compare your data against a catalogue the most accurate and
comprehensive datasets available.”
Verification, or Validation as it is otherwise referred to, is the process of
comparing the data you have against the an authoritative third party
dataset.
Kleber makes verification of data against third party datasets easy as
many of the common data sets are embedded within the Kleber
platform.
Using the results from a validation process it can be determined what
information is correct and what needs to be repaired.

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
“But doesn’t verification give the same result as parsing?”
Yes & No: For accurate and clean information the results will be similar.
For problem data no amount of verification will deliver a clean result.
This is where Kleber really shines. Validation is only as good as the parsing
technology that underpins it. Remember, parsing is the foundation on
which all other data process are built.
Properly parsed data can still be repaired and cleaned without being
validated, so good parsing is vital. As with the exception handling that
Data Capture may create, so too can replying solely on validation
technologies. What to you do with data that doesn’t validate?

Beware of vendors who promote validation as a means of cleaning data. It
is easy to compare data against a structured source of truth. It is entirely
more complex to clean and make sense of poorly structured data. This is
the domain of a good parser such as that provided within Kleber.

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
“OK. So I can use the verified information to repair my data?”
Yes. But there is more…
“Both verified and un-verified information can be repaired.”
After verifying a record, or part thereof, against a known source of truth, any
discrepancies between the original record and the reference source can be
intelligently examined and the data repaired.
E.g. updating a street type from St to Dr.

This is particularly useful for address information where reference datasets
such as the Australia Post Postal Address File (PAF) can be used to verify and
repair data.

Remember these processes can happen for a single record during data entry or
across an entire database. Imagine repairing all your data in a single process?

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
“But what about records that can’t be verified? How do I repair them?”
That’s where it comes back to smart parsing.
“Remember parsing is the foundation to true data quality.”
Traditionally, to repair (i.e. clean) data records that could not be verified a separate manual exception handling process would
need to be undertaken. This is a time consuming and costly process and in most cases the result would be inconclusive.
The smart parsing capabilities of Kleber enable even unverified data to be cleansed and repaired sufficiently to facilitate
further downstream processes such as matching and data appending. This significantly reduces the number of exceptions and
time associated in dealing with these.
Quite often when an a business is looking for “Data Quality” what they typically mean is “Data Repair”. It is only due to the prolific
and highly successful promotion of data validation technologies that businesses look to validation as a source of data quality.

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
“So now I have clean data?”
Yes. However it may not be in the format required.
“Format of data is just as important as its cleanliness and validity.”
Formatting of data is vital to it being usable by a business. Different systems, process and
communications channels require data to be presented in different formats. Data must be
prepared in a way that each system will accept it. After all, you can’t push a round peg into a
square hole!
Presentation of an address on an invoice is very different to that required for exchange of
address data with 3PL providers or with other areas within the same business.
Because of Kleber’s advanced parsing capabilities data can be put together in almost any
format or structure. Imagine always having a perfectly formatted output for each specific
system or communications channel!

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
“It’s been parsed, verified, repaired and formatted – Is it clean data now?”
Yes, it’s clean. But its not true data quality.
“Data Quality is more than just the cleanliness, validity and format of an
individual record. Its these things across all the data in its entirety.”
Congratulations. You now have a beautifully formatted (and possibly verified) data
record – or you might have several of them, or millions, or more! Now how many of
these are unique?
Kleber includes industry leading data matching technologies that allow you to not
only find duplicates in your own database, but also allows you to compare records
between databases.
Kleber matching technologies can be used on full records to identify unique individuals
or on partial records to identify matches at specific levels. Kleber can append a “match
key” automatically during data processing if required.

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
“Are we there yet?”
That depends on how much you want to know!
“Appending certain data can provide insights that would otherwise be unknown.
Uplifting data will improve the completeness of the data, making it more valuable.”

There is a potentially limitless amount of data out there that can be appended to
your existing data to deliver insights and value. Geospatial data to plot a point on
a map, affluence and behavioural data for marketing and even psychographic and
cultural profiling data to predict trends.
Missing data, such as an phone number or current address can be added using
any one of the various data services providers Kleber makes accessible.
The ability to append and uplift data with this other useful information is
dependant on having clean, accurate and properly structured data. The Kleber
process makes all this possible.

Capture

Copyright DataTools 2014

Parse

Verify

Repair

www.datatools.com.au

Format

Match

Enhance

Addressing Data Quality
I think I get it now!
Great, but let’s recap.

Capture

Kleber provides all the necessary (if not vital) components to achieve data quality!
Create

Parse

It provides all of this in ONE simple to implement process (one call)

It can be used in both real-time (data entry) and batch (existing data)
Quality parsing is vital to all other data processes

Enhance

Verify

Validation is not parsing, but good parsing means better validation

Data Quality needn't be difficult, especially if you use Kleber.
For further information regarding Kleber and how it can benefit your business
please contact one of the friendly staff at DataTools on +61 2 9687 4666.

Match

Repair
Format

P.S. We’ve only scratched the surface!

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality
Thank You
www.datatools.com.au
info@datatools.com.au
+61 2 9687 4666

Copyright DataTools 2014

www.datatools.com.au

Addressing Data Quality

More Related Content

What's hot

Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017
Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017
Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017Caserta
 
Setting Up the Data Lake
Setting Up the Data LakeSetting Up the Data Lake
Setting Up the Data LakeCaserta
 
Moving Past Infrastructure Limitations
Moving Past Infrastructure LimitationsMoving Past Infrastructure Limitations
Moving Past Infrastructure LimitationsCaserta
 
What Data Do You Have and Where is It?
What Data Do You Have and Where is It? What Data Do You Have and Where is It?
What Data Do You Have and Where is It? Caserta
 
Making Big Data Easy for Everyone
Making Big Data Easy for EveryoneMaking Big Data Easy for Everyone
Making Big Data Easy for EveryoneCaserta
 
Big Data Analytics on the Cloud
Big Data Analytics on the CloudBig Data Analytics on the Cloud
Big Data Analytics on the CloudCaserta
 
You're the New CDO, Now What?
You're the New CDO, Now What?You're the New CDO, Now What?
You're the New CDO, Now What?Caserta
 
Benefits of the Azure Cloud
Benefits of the Azure CloudBenefits of the Azure Cloud
Benefits of the Azure CloudCaserta
 
Data_Harmonization_ClearStory
Data_Harmonization_ClearStoryData_Harmonization_ClearStory
Data_Harmonization_ClearStoryWilliam Davis
 
Balancing Data Governance and Innovation
Balancing Data Governance and InnovationBalancing Data Governance and Innovation
Balancing Data Governance and InnovationCaserta
 
Slides: Moving from a Relational Model to NoSQL
Slides: Moving from a Relational Model to NoSQLSlides: Moving from a Relational Model to NoSQL
Slides: Moving from a Relational Model to NoSQLDATAVERSITY
 
Big Data's Impact on the Enterprise
Big Data's Impact on the EnterpriseBig Data's Impact on the Enterprise
Big Data's Impact on the EnterpriseCaserta
 
Agile Leadership: Guiding DataOps Teams Through Rapid Change and Uncertainty
Agile Leadership: Guiding DataOps Teams Through Rapid Change and UncertaintyAgile Leadership: Guiding DataOps Teams Through Rapid Change and Uncertainty
Agile Leadership: Guiding DataOps Teams Through Rapid Change and UncertaintyTamrMarketing
 
Mastering Customer Data on Apache Spark
Mastering Customer Data on Apache SparkMastering Customer Data on Apache Spark
Mastering Customer Data on Apache SparkCaserta
 
Ureason jules oudmans
Ureason jules oudmansUreason jules oudmans
Ureason jules oudmansBigDataExpo
 
Using Machine Learning & Spark to Power Data-Driven Marketing
Using Machine Learning & Spark to Power Data-Driven MarketingUsing Machine Learning & Spark to Power Data-Driven Marketing
Using Machine Learning & Spark to Power Data-Driven MarketingCaserta
 
Defining and Applying Data Governance in Today’s Business Environment
Defining and Applying Data Governance in Today’s Business EnvironmentDefining and Applying Data Governance in Today’s Business Environment
Defining and Applying Data Governance in Today’s Business EnvironmentCaserta
 
Best Practices for Big Data Analytics with Machine Learning by Datameer
Best Practices for Big Data Analytics with Machine Learning by DatameerBest Practices for Big Data Analytics with Machine Learning by Datameer
Best Practices for Big Data Analytics with Machine Learning by DatameerDatameer
 
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Tristan Baker
 
Journey to Cloud Analytics
Journey to Cloud Analytics Journey to Cloud Analytics
Journey to Cloud Analytics Datavail
 

What's hot (20)

Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017
Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017
Creating a DevOps Practice for Analytics -- Strata Data, September 28, 2017
 
Setting Up the Data Lake
Setting Up the Data LakeSetting Up the Data Lake
Setting Up the Data Lake
 
Moving Past Infrastructure Limitations
Moving Past Infrastructure LimitationsMoving Past Infrastructure Limitations
Moving Past Infrastructure Limitations
 
What Data Do You Have and Where is It?
What Data Do You Have and Where is It? What Data Do You Have and Where is It?
What Data Do You Have and Where is It?
 
Making Big Data Easy for Everyone
Making Big Data Easy for EveryoneMaking Big Data Easy for Everyone
Making Big Data Easy for Everyone
 
Big Data Analytics on the Cloud
Big Data Analytics on the CloudBig Data Analytics on the Cloud
Big Data Analytics on the Cloud
 
You're the New CDO, Now What?
You're the New CDO, Now What?You're the New CDO, Now What?
You're the New CDO, Now What?
 
Benefits of the Azure Cloud
Benefits of the Azure CloudBenefits of the Azure Cloud
Benefits of the Azure Cloud
 
Data_Harmonization_ClearStory
Data_Harmonization_ClearStoryData_Harmonization_ClearStory
Data_Harmonization_ClearStory
 
Balancing Data Governance and Innovation
Balancing Data Governance and InnovationBalancing Data Governance and Innovation
Balancing Data Governance and Innovation
 
Slides: Moving from a Relational Model to NoSQL
Slides: Moving from a Relational Model to NoSQLSlides: Moving from a Relational Model to NoSQL
Slides: Moving from a Relational Model to NoSQL
 
Big Data's Impact on the Enterprise
Big Data's Impact on the EnterpriseBig Data's Impact on the Enterprise
Big Data's Impact on the Enterprise
 
Agile Leadership: Guiding DataOps Teams Through Rapid Change and Uncertainty
Agile Leadership: Guiding DataOps Teams Through Rapid Change and UncertaintyAgile Leadership: Guiding DataOps Teams Through Rapid Change and Uncertainty
Agile Leadership: Guiding DataOps Teams Through Rapid Change and Uncertainty
 
Mastering Customer Data on Apache Spark
Mastering Customer Data on Apache SparkMastering Customer Data on Apache Spark
Mastering Customer Data on Apache Spark
 
Ureason jules oudmans
Ureason jules oudmansUreason jules oudmans
Ureason jules oudmans
 
Using Machine Learning & Spark to Power Data-Driven Marketing
Using Machine Learning & Spark to Power Data-Driven MarketingUsing Machine Learning & Spark to Power Data-Driven Marketing
Using Machine Learning & Spark to Power Data-Driven Marketing
 
Defining and Applying Data Governance in Today’s Business Environment
Defining and Applying Data Governance in Today’s Business EnvironmentDefining and Applying Data Governance in Today’s Business Environment
Defining and Applying Data Governance in Today’s Business Environment
 
Best Practices for Big Data Analytics with Machine Learning by Datameer
Best Practices for Big Data Analytics with Machine Learning by DatameerBest Practices for Big Data Analytics with Machine Learning by Datameer
Best Practices for Big Data Analytics with Machine Learning by Datameer
 
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021
 
Journey to Cloud Analytics
Journey to Cloud Analytics Journey to Cloud Analytics
Journey to Cloud Analytics
 

Viewers also liked

Enhanced Social Accountability through Open Access to Data
Enhanced Social Accountability through Open Access to DataEnhanced Social Accountability through Open Access to Data
Enhanced Social Accountability through Open Access to DataSoren Gigler
 
Special Report: Geocoding...achieving the highest accuracy
Special Report: Geocoding...achieving the highest accuracySpecial Report: Geocoding...achieving the highest accuracy
Special Report: Geocoding...achieving the highest accuracyJohn Woloshen
 
Linear referencing 2014
Linear referencing 2014Linear referencing 2014
Linear referencing 2014Atiqa khan
 
[Android] Maps, Geocoding and Location-Based Services
[Android] Maps, Geocoding and Location-Based Services[Android] Maps, Geocoding and Location-Based Services
[Android] Maps, Geocoding and Location-Based ServicesNikmesoft Ltd
 
Study: The Future of VR, AR and Self-Driving Cars
Study: The Future of VR, AR and Self-Driving CarsStudy: The Future of VR, AR and Self-Driving Cars
Study: The Future of VR, AR and Self-Driving CarsLinkedIn
 

Viewers also liked (7)

Social media monitoring tactics
Social media monitoring tacticsSocial media monitoring tactics
Social media monitoring tactics
 
Bridging the gap
Bridging the gapBridging the gap
Bridging the gap
 
Enhanced Social Accountability through Open Access to Data
Enhanced Social Accountability through Open Access to DataEnhanced Social Accountability through Open Access to Data
Enhanced Social Accountability through Open Access to Data
 
Special Report: Geocoding...achieving the highest accuracy
Special Report: Geocoding...achieving the highest accuracySpecial Report: Geocoding...achieving the highest accuracy
Special Report: Geocoding...achieving the highest accuracy
 
Linear referencing 2014
Linear referencing 2014Linear referencing 2014
Linear referencing 2014
 
[Android] Maps, Geocoding and Location-Based Services
[Android] Maps, Geocoding and Location-Based Services[Android] Maps, Geocoding and Location-Based Services
[Android] Maps, Geocoding and Location-Based Services
 
Study: The Future of VR, AR and Self-Driving Cars
Study: The Future of VR, AR and Self-Driving CarsStudy: The Future of VR, AR and Self-Driving Cars
Study: The Future of VR, AR and Self-Driving Cars
 

Similar to DataTools Kleber. Powerful data quality is a single, simple to implement process.

Don’t Make Bad Data an Excuse
Don’t Make Bad Data an ExcuseDon’t Make Bad Data an Excuse
Don’t Make Bad Data an ExcuseConnexica
 
Big Data Matching - How to Find Two Similar Needles in a Really Big Haystack
Big Data Matching - How to Find Two Similar Needles in a Really Big HaystackBig Data Matching - How to Find Two Similar Needles in a Really Big Haystack
Big Data Matching - How to Find Two Similar Needles in a Really Big HaystackPrecisely
 
Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)
Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)
Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)Denodo
 
Expert Big Data Tips
Expert Big Data TipsExpert Big Data Tips
Expert Big Data TipsQubole
 
Automation in data migration and data validation
Automation in data migration and data validationAutomation in data migration and data validation
Automation in data migration and data validationruchabhandiwad
 
OberservePoint - The Digital Data Quality Playbook
OberservePoint - The Digital Data Quality  PlaybookOberservePoint - The Digital Data Quality  Playbook
OberservePoint - The Digital Data Quality PlaybookObservePoint
 
Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...
Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...
Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...Dana Gardner
 
Data quality and data profiling
Data quality and data profilingData quality and data profiling
Data quality and data profilingShailja Khurana
 
Data cleansing steps you must follow for better data health
Data cleansing steps you must follow for better data healthData cleansing steps you must follow for better data health
Data cleansing steps you must follow for better data healthGen Leads
 
How to Scale your Analytics in a Maturing Organization
How to Scale your Analytics in a Maturing OrganizationHow to Scale your Analytics in a Maturing Organization
How to Scale your Analytics in a Maturing OrganizationKissmetrics on SlideShare
 
How Can You Implement DataOps In Your Existing Workflow?
How Can You Implement DataOps In Your Existing Workflow?How Can You Implement DataOps In Your Existing Workflow?
How Can You Implement DataOps In Your Existing Workflow?Enov8
 
Data Driven Testing Is More Than an Excel File
Data Driven Testing Is More Than an Excel FileData Driven Testing Is More Than an Excel File
Data Driven Testing Is More Than an Excel FileMehmet Gök
 
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida  Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida CLARA CAMPROVIN
 
Using Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales GoalsUsing Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales GoalsIrshadKhan682442
 
Using Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales GoalsUsing Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales GoalsWilliamJohnson288536
 
Using Data Lakes To Sail Through Your Sales Goals
Using Data Lakes To Sail Through Your Sales GoalsUsing Data Lakes To Sail Through Your Sales Goals
Using Data Lakes To Sail Through Your Sales GoalsKevinJohnson667312
 
WHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEM
WHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEMWHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEM
WHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEMRajaraj64
 
FirstEigen Brochure- All clouds.pdf
FirstEigen Brochure- All clouds.pdfFirstEigen Brochure- All clouds.pdf
FirstEigen Brochure- All clouds.pdfarifulislam946965
 
Building The Agile Database
Building The Agile DatabaseBuilding The Agile Database
Building The Agile Databaseelliando dias
 

Similar to DataTools Kleber. Powerful data quality is a single, simple to implement process. (20)

Don’t Make Bad Data an Excuse
Don’t Make Bad Data an ExcuseDon’t Make Bad Data an Excuse
Don’t Make Bad Data an Excuse
 
Big Data Matching - How to Find Two Similar Needles in a Really Big Haystack
Big Data Matching - How to Find Two Similar Needles in a Really Big HaystackBig Data Matching - How to Find Two Similar Needles in a Really Big Haystack
Big Data Matching - How to Find Two Similar Needles in a Really Big Haystack
 
Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)
Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)
Accelerate Self-Service Analytics with Virtualization and Visualisation (Thai)
 
Expert Big Data Tips
Expert Big Data TipsExpert Big Data Tips
Expert Big Data Tips
 
Automation in data migration and data validation
Automation in data migration and data validationAutomation in data migration and data validation
Automation in data migration and data validation
 
OberservePoint - The Digital Data Quality Playbook
OberservePoint - The Digital Data Quality  PlaybookOberservePoint - The Digital Data Quality  Playbook
OberservePoint - The Digital Data Quality Playbook
 
Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...
Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...
Complex Carrier Network Performance Data on Vertica Yields Performance and Cu...
 
Data quality and data profiling
Data quality and data profilingData quality and data profiling
Data quality and data profiling
 
Data cleansing steps you must follow for better data health
Data cleansing steps you must follow for better data healthData cleansing steps you must follow for better data health
Data cleansing steps you must follow for better data health
 
How to Scale your Analytics in a Maturing Organization
How to Scale your Analytics in a Maturing OrganizationHow to Scale your Analytics in a Maturing Organization
How to Scale your Analytics in a Maturing Organization
 
How Can You Implement DataOps In Your Existing Workflow?
How Can You Implement DataOps In Your Existing Workflow?How Can You Implement DataOps In Your Existing Workflow?
How Can You Implement DataOps In Your Existing Workflow?
 
Data Driven Testing Is More Than an Excel File
Data Driven Testing Is More Than an Excel FileData Driven Testing Is More Than an Excel File
Data Driven Testing Is More Than an Excel File
 
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida  Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
Jet Reports es la herramienta para construir el mejor BI y de forma mas rapida
 
Using Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales GoalsUsing Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales Goals
 
Using Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales GoalsUsing Data Lakes to Sail Through Your Sales Goals
Using Data Lakes to Sail Through Your Sales Goals
 
Using Data Lakes To Sail Through Your Sales Goals
Using Data Lakes To Sail Through Your Sales GoalsUsing Data Lakes To Sail Through Your Sales Goals
Using Data Lakes To Sail Through Your Sales Goals
 
WHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEM
WHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEMWHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEM
WHAT IS A DATA LAKE? Know DATA LAKES & SALES ECOSYSTEM
 
FirstEigen Brochure- All clouds.pdf
FirstEigen Brochure- All clouds.pdfFirstEigen Brochure- All clouds.pdf
FirstEigen Brochure- All clouds.pdf
 
Building The Agile Database
Building The Agile DatabaseBuilding The Agile Database
Building The Agile Database
 
Improving Data Extraction Performance
Improving Data Extraction PerformanceImproving Data Extraction Performance
Improving Data Extraction Performance
 

Recently uploaded

Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersRaghuram Pandurangan
 
Potential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and InsightsPotential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and InsightsRavi Sanghani
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
Sample pptx for embedding into website for demo
Sample pptx for embedding into website for demoSample pptx for embedding into website for demo
Sample pptx for embedding into website for demoHarshalMandlekar2
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Scott Andery
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Alkin Tezuysal
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesThousandEyes
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesKari Kakkonen
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 

Recently uploaded (20)

Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information Developers
 
Potential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and InsightsPotential of AI (Generative AI) in Business: Learnings and Insights
Potential of AI (Generative AI) in Business: Learnings and Insights
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
Sample pptx for embedding into website for demo
Sample pptx for embedding into website for demoSample pptx for embedding into website for demo
Sample pptx for embedding into website for demo
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
Enhancing User Experience - Exploring the Latest Features of Tallyman Axis Lo...
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examples
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 

DataTools Kleber. Powerful data quality is a single, simple to implement process.

  • 1. What is Kleber? A non technical look into the platform that will change the way you think about data quality Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 2. What is DataTools Kleber? Let’s begin. Capture First things first. If you are a technical person please contact us and we will send you all the architecture, methods calls, functions, services information and otherwise foreign to non technical people jargon you can handle. If you are a business owner, operations manager, marketing guru, non technical PM or otherwise just wanting to get things fixed, read on! Kleber is the culmination of almost 20 years providing data quality software. It binds together all the discrete elements required to deliver true data quality into a single platform. Further still it harmoniously unifies each of the commonly required individual processes into a single process that can be easily implemented from within a website or application. To delver true data quality, the process outlined in this document can be implemented in real-time for on boarding of information and in a batch method to process all the data already residing within your organisation. Create Parse Enhance Verify Match Repair Format Read on to learn about each step in the Kleber process and how these come together as a complete data processing solution. Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 3. “I want quality data. I’ve been told I need a Data Capture solution.” It’s a great place to start. “To assist in fast and accurate capture of contact information” Data Capture solutions make the capture of contact information within your websites and applications quick and easy. It’s also a great place to get into DataTools Kleber. Predictive Search technologies (otherwise known as Type Assist) leverage advanced keystroke reduction capabilities to predict the information being entered and minimise the number of keystrokes required. This makes entering of information not only faster and more accurate but also more pleasurable for the user, reducing abandonment and enhancing the user experience. Because Predictive Search is highly intuitive it requires no user training and can be used by new users with ease. Kleber offers predictive capture for physical addresses, first names, surnames, business names and email addresses. Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 4. “I already use data capture. How is Kleber of any benefit to me?” Handling Exceptions. “Every data capture process creates exceptions that must be handled. This is typically the most complex and costly aspect of any data capture project” Despite advanced logic being used in Predictive Capture technologies not all data will be captured correctly. Even the most sophisticated data capture solutions will have some percentage of error or information that could not be captured. Data Capture solutions have been sold as the golden panacea for data quality issues. The reality is these technologies only handle the “simple fixes”. True data quality requires a mechanism for handling the few percent of truly complicated expectations these technologies create. It is this data that pollutes databases and causes complex and costly exception handling process throughout the information workflow. DataTools Kleber makes handing of these exceptions easy by ensuring even poorly captured data is repaired and cleansed before being stored. Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 5. “OK. So I can fix the exceptions, but that’s still a lot of work, right?” Wrong! One Call does it all. Automated exception handling via “Handling exceptions is as simple as making a single call to Kleber . Kleber will then repair and clean the data turning it into useful information and return it to you” • Parsing to split the captured data into discrete fields All of the individual and complex data processing methods have been combined together in a single function within Kleber. By calling this function the data will be parsed, verified, repaired and formatted. Match keys will also be appended as will various other types of information to provide insight into the data and deliver operational efficiencies. You can call each of these function individually if you want to, but all the hard work has already been done, so why not leave it to us. • Verification of data to ensure its accuracy • Repair of data through results of verification • Formatting of the data to ensure it is suitable for use • Matching to compare data and find duplicates Using Kleber at this point will eliminate the need for complex downstream exception handling. Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 6. “I don’t have Data Capture technologies. Do I need to implement these first?” No. It will provider a better user experience, but you don’t have to use it. Simply send the data as you store it now. “Use your standard data entry forms to capture the data and send it as is or process all the data in your database as a batch.” If you don’t have a data capture solution that’s OK. You can use your existing fields for users to enter data. Once all the data has been entered simply send it to Kleber and get the clean, verified and nicely formatted result. Data Capture technologies simplify the entry of data and make the process faster and more pleasurable for the user. We can provide these too! Kleber can also be used in a Batch fashion whereby an entire list or database can be processed. Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 7. “So how does it work?” It’s complicated – Well it is: but we’ve made it simple. “There is a logical process through which all data can be taken, each step of the process building on the previous on to ultimately deliver true data quality and insight .” Capture Data is captured, whether it be though Predictive technologies or traditional data entry. Parse Parsing breaks the data down into its smallest parts ready for other downstream processes. Copyright DataTools 2014 Verify Parsed data is compared to “sources of truth” to verify the accuracy of the information. Repair Results from the parsing and verification are used to repair the data, making it clean and useful. www.datatools.com.au Format The cleansed data elements can then be put together in a variety of ways to ensure the data meets specific downstream requirements. Match Match Keys are appended to the data in order to compare records against each other and find duplicates. Enhance Other information, is appended to the data in order to deliver true insight. E.g. Geospatial data. Addressing Data Quality
  • 8. “Why do I need to parse my data?” It all starts with parsing. “Parsing is the foundation of quality data. Without quality parsing all other data processes will fail.” Parsing is the process of splitting data apart into its smallest parts in a way that makes sense. It is typically done behind the scenes and, as such, its importance is often underestimated. Proper parsing is often “faked” by many vendors who simply take variations of a record and reference this against a known source of truth until they get a match. This is inefficient and highly dependant on the quality of the source of truth. Kleber provides true parsing. After all, DataTools has been delivering advanced data parsing technologies for almost 20 years! Ex Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 9. “What does verification involve?” Referencing data against a “Source of Truth”. “Compare your data against a catalogue the most accurate and comprehensive datasets available.” Verification, or Validation as it is otherwise referred to, is the process of comparing the data you have against the an authoritative third party dataset. Kleber makes verification of data against third party datasets easy as many of the common data sets are embedded within the Kleber platform. Using the results from a validation process it can be determined what information is correct and what needs to be repaired. Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 10. “But doesn’t verification give the same result as parsing?” Yes & No: For accurate and clean information the results will be similar. For problem data no amount of verification will deliver a clean result. This is where Kleber really shines. Validation is only as good as the parsing technology that underpins it. Remember, parsing is the foundation on which all other data process are built. Properly parsed data can still be repaired and cleaned without being validated, so good parsing is vital. As with the exception handling that Data Capture may create, so too can replying solely on validation technologies. What to you do with data that doesn’t validate? Beware of vendors who promote validation as a means of cleaning data. It is easy to compare data against a structured source of truth. It is entirely more complex to clean and make sense of poorly structured data. This is the domain of a good parser such as that provided within Kleber. Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 11. “OK. So I can use the verified information to repair my data?” Yes. But there is more… “Both verified and un-verified information can be repaired.” After verifying a record, or part thereof, against a known source of truth, any discrepancies between the original record and the reference source can be intelligently examined and the data repaired. E.g. updating a street type from St to Dr. This is particularly useful for address information where reference datasets such as the Australia Post Postal Address File (PAF) can be used to verify and repair data. Remember these processes can happen for a single record during data entry or across an entire database. Imagine repairing all your data in a single process? Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 12. “But what about records that can’t be verified? How do I repair them?” That’s where it comes back to smart parsing. “Remember parsing is the foundation to true data quality.” Traditionally, to repair (i.e. clean) data records that could not be verified a separate manual exception handling process would need to be undertaken. This is a time consuming and costly process and in most cases the result would be inconclusive. The smart parsing capabilities of Kleber enable even unverified data to be cleansed and repaired sufficiently to facilitate further downstream processes such as matching and data appending. This significantly reduces the number of exceptions and time associated in dealing with these. Quite often when an a business is looking for “Data Quality” what they typically mean is “Data Repair”. It is only due to the prolific and highly successful promotion of data validation technologies that businesses look to validation as a source of data quality. Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 13. “So now I have clean data?” Yes. However it may not be in the format required. “Format of data is just as important as its cleanliness and validity.” Formatting of data is vital to it being usable by a business. Different systems, process and communications channels require data to be presented in different formats. Data must be prepared in a way that each system will accept it. After all, you can’t push a round peg into a square hole! Presentation of an address on an invoice is very different to that required for exchange of address data with 3PL providers or with other areas within the same business. Because of Kleber’s advanced parsing capabilities data can be put together in almost any format or structure. Imagine always having a perfectly formatted output for each specific system or communications channel! Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 14. “It’s been parsed, verified, repaired and formatted – Is it clean data now?” Yes, it’s clean. But its not true data quality. “Data Quality is more than just the cleanliness, validity and format of an individual record. Its these things across all the data in its entirety.” Congratulations. You now have a beautifully formatted (and possibly verified) data record – or you might have several of them, or millions, or more! Now how many of these are unique? Kleber includes industry leading data matching technologies that allow you to not only find duplicates in your own database, but also allows you to compare records between databases. Kleber matching technologies can be used on full records to identify unique individuals or on partial records to identify matches at specific levels. Kleber can append a “match key” automatically during data processing if required. Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 15. “Are we there yet?” That depends on how much you want to know! “Appending certain data can provide insights that would otherwise be unknown. Uplifting data will improve the completeness of the data, making it more valuable.” There is a potentially limitless amount of data out there that can be appended to your existing data to deliver insights and value. Geospatial data to plot a point on a map, affluence and behavioural data for marketing and even psychographic and cultural profiling data to predict trends. Missing data, such as an phone number or current address can be added using any one of the various data services providers Kleber makes accessible. The ability to append and uplift data with this other useful information is dependant on having clean, accurate and properly structured data. The Kleber process makes all this possible. Capture Copyright DataTools 2014 Parse Verify Repair www.datatools.com.au Format Match Enhance Addressing Data Quality
  • 16. I think I get it now! Great, but let’s recap. Capture Kleber provides all the necessary (if not vital) components to achieve data quality! Create Parse It provides all of this in ONE simple to implement process (one call) It can be used in both real-time (data entry) and batch (existing data) Quality parsing is vital to all other data processes Enhance Verify Validation is not parsing, but good parsing means better validation Data Quality needn't be difficult, especially if you use Kleber. For further information regarding Kleber and how it can benefit your business please contact one of the friendly staff at DataTools on +61 2 9687 4666. Match Repair Format P.S. We’ve only scratched the surface! Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality
  • 17. Thank You www.datatools.com.au info@datatools.com.au +61 2 9687 4666 Copyright DataTools 2014 www.datatools.com.au Addressing Data Quality