
Matthew Magne describes how SAS Data Quality can help you build a trusted data foundation, one stone at a time.
Matthew Magne describes how SAS Data Quality can help you build a trusted data foundation, one stone at a time.
Jim Harris advocates addressing data quality and governance issues on the way to data lakes and Hadoop.
Clark Bradley explains how SAS can make Hadoop approachable and accessible.
Jim Harris explains why data quality is such a fundamental aspect of master data management.
Lenin und ich sitzen im Publikum und applaudieren heftig: Seine Chefin hat ihren Vortrag beendet über „Datenqualität als Erfolgsfaktor im Internet of Things“. „Kein Datenqualitätsprojekt ohne Hilfe von oben“, raunt Lenin mir zu, "Unterstützung vom Boss ist manchmal wichtiger als tolle Software." Ich will beleidigt darauf hinweisen, dass seine Chefin
In a recent presentation, Jill Dyche, VP of SAS Best Practices gave two great quotes: "Map strategy to data" and "strategy drives analytics drives data." In other words, don't wait for your data to be perfect before you invest in analytics. Don't get me wrong -- I fully understand and
The rise of self-service analytics, and the idea of the ‘citizen data scientist’, has also brought a number of issues to the fore in organizations. In particular, two common areas of discussion are the twin pillars of data quality and data preparation. There is no doubt that good quality, well-prepared
David Loshin extends his exploration of ethical issues surrounding automated systems and event stream processing to encompass data quality and risk considerations.
Lenin hatte gelächelt und von seinen Erfolgen im Internet of Things berichtet; richtig begeistert war er gewesen. – Aber jetzt murrt er: „Das ist alles Müll! Internet of Trash sollte es heißen! Die Daten stimmen nicht, die Leute schimpfen über das Projekt, der Fachbereich und meine Chefin sitzen mir im
.@philsimon advises to be wary of those promising obvious and facile solutions to increasingly challenging governance and privacy issues.
In my last post I described "4 adaptability attributes for analytical success," and in the past I've discussed the strategic role analytics play in helping organizations succeed now and into the future. Now I'd like to discuss three attributes that define a powerful analytics environment: Speed Accuracy Scalability [NOTE: Any
Streaming technologies have been around for years, but as Felix Liao recently blogged, the numbers and types of use cases that can take advantage of these technologies have now increased exponentially. I've blogged about why streaming is the most effective way to handle the volume, variety and velocity of big data. That's
What are the most useful skills a data quality leader can possess? As an editor of an online data quality magazine, I naturally get asked this type of question regularly at events and meetups. My answer may surprise some who are expecting a data-centric response. I firmly believe that sales and
"Two weeks to go," Santa said to himself, with millions of toys stacked up on the shelves. Each year worry hit at the same time – "How do I get the right toy to the right child without losing my mind?" Though Old St. Nick didn't have a computer science degree, deep down
During a data quality assessment, one of my clients discovered that a large chunk of data that ultimately fed into their business analytics engine was sourced externally. After examining the contracts surrounding this data, I found that 100% of it failed to possess service-level agreements (SLAs) for the quality of
Historically, before data was managed it was moved to a central location. For a long time that central location was the staging area for an enterprise data warehouse (EDW). While EDWs and their staging areas are still in use – especially for structured, transactional and internally generated data – big
Data quality initiatives challenge organizations because the discipline encompasses so many issues, approaches and tools. Across the board, there are four main activity areas – or pillars – that underlie any successful data quality initiative. Let’s look at what each pillar means, then consider the benefits SAS Data Management brings
Digitalisierung, Big Data, IoT, Smart Data – die Liste an Ansätzen, die den Klassiker „aus Daten mach Umsatz“ neu definieren wollen, wird länger und länger. An cleveren, schlüssigen und vielversprechenden Methoden mangelt es sicher nicht, ihnen allen gemein ist aber der mahnende Zeigefinger und ewige Spielverderber Datenqualität. Und wie das
Our world is now so awash in data that many organizations have an embarrassment of riches when it comes to available data to support operational, tactical and strategic activities of the enterprise. Such a data-rich environment is highly susceptible to poor-quality data. This is especially true when swimming in data lakes –
We often talk about full customer data visibility and the need for a “golden record” that provides a 360-degree view of the customer to enhance our customer-facing processes. The rationale is that by accumulating all the data about a customer (or, for that matter, any entity of interest) from multiple sources, you
Most enterprises employ multiple analytical models in their business intelligence applications and decision-making processes. These analytical models include descriptive analytics that help the organization understand what has happened and what is happening now, predictive analytics that determine the probability of what will happen next, and prescriptive analytics that focus on
In my prior posts about operational data governance, I've suggested the need to embed data validation as an integral component of any data integration application. In my last post, we looked at an example of using a data quality audit report to ensure fidelity of the data integration processes for
Data governance plays an integral role in many enterprise information initiatives, such as data quality, master data management and analytics. It requires coordinating a complex combination of factors, including executive sponsorship, funding, decision rights, arbitration of conflicting priorities, policy definition, policy implementation, data stewardship and change management. With so much overhead involved in
Data integration teams often find themselves in the middle of discussions where the quality of their data outputs are called into question. Without proper governance procedures in place, though, it's hard to address these accusations in a reasonable way. Here's why.
Data governance has been the topic of many of the recent posts here on the Data Roundtable. And rightfully so, since data governance plays such an integral role in the success of many enterprise information initiatives – such as data quality, master data management and analytics. These posts can help you prepare for discussing
Welcome to the 1st practical step for tackling auto insurance fraud with analytics. It is obvious why our first stop relates with data, the idiom “the devil is in the details” can easily be applied in the insurance fraud sector as “the devil is in the data”. This article analyses
.@philsimon on the need to adopt agile methodologies for data prep and analytics.
Lately I've been binge-watching a lot of police procedural television shows. The standard format for almost every episode is the same. It starts with the commission or discovery of a crime, followed by forensic investigation of the crime scene, analysis of the collected evidence, and interviews or interrogations with potential suspects. It ends
Critical business applications depend on the enterprise creating and maintaining high-quality data. So, whenever new data is received – especially from a new source – it’s great when that source can provide data without defects or other data quality issues. The recent rise in self-service data preparation options has definitely improved the quality of
Hadoop has driven an enormous amount of data analytics activity lately. And this poses a problem for many practitioners coming from the traditional relational database management system (RDBMS) world. Hadoop is well known for having lots of variety in the structure of data it stores and processes. But it's fair to