Provides foundation for secure and intelligent data sharing between business teams across public and private clouds
IBM (NYSE: IBM) announced new offerings to its Watson Data Platform, including data cataloging and data refining, which is designed to make it easier for developers and data scientists to analyze and prepare enterprise data for AI applications, regardless of its structure or where it resides. By improving data visibility and helping to better enforce data security policies, users can now connect and share data across public and private cloud environments.
By 2018, nearly 75 percent of developers will build AI functionality into their apps, according to IDC1. However, they also face the obstacle of making sense of increasingly complex data that lives in different places, and that must be securely and continually ingested to power these apps.
Addressing these challenges, IBM has expanded the functionality of its Watson Data Platform, an integrated set of tools, services and data on the IBM Cloud designed to enable data scientists, developers and business teams to gain intelligence from the data most important to their roles, as well as easily access services like machine learning, AI and analytics.
“We are always looking for new ways to gain a more holistic view of our clients’ campaign data, and design tailored approaches for each ad and marketing tactic,” said Michael Kaushansky, Chief Data Officer at Havas, a global advertising and marketing consultancy. “The Watson Data Platform is helping us do just that by quickly connecting offline and online marketing data. For example, we recently kicked off a test for one of our automotive clients, aiming to connect customer data, advertising information in existing systems, and online engagement metrics to better target the right audiences at the right time.”
Specifically, this expansion includes:
- New Data Catalog and Data Refinery offerings, which bring together datasets that live in different formats on the cloud, in existing systems and in third party sources; as well as apply machine learning to process and cleanse this data so it can be ingested for AI applications;
- The ability to use metadata, pulled from Data Catalog and Data Refinery, to tag and help enforce a client’s data governance policies. This gives teams a foundation to more easily identify risks when sharing sensitive data.
- The general availability of Analytics Engine to separate the storage of data from the information it holds, allowing it to be analyzed and fed into apps at much greater speeds. As a result, developers and data scientists can more easily share and build with large datasets.
More details on the new offerings of the IBM Watson Data Platform may be found at https://www.ibm.com/blogs/bluemix/2017/11/watson-data-platform-announcement/.
“The key to AI starts with a strong data foundation, which turns the volume and velocity of incoming data from a challenge into an asset,” said Derek Schoettle, General Manager, IBM Watson Data Platform. “For companies to innovate and compete with AI, they need a way to grasp and organize data coming in from every source, and to use this complete index of data as the backbone of every decision and initiative.”
To further help companies grasp control of all of their data no matter where it resides, IBM is also announcing a series of new features to its Unified Governance Platform. These bring greater visibility and management of clients’ global data, including new capabilities that help clients as they better prepare for impending data protection regulations such as GDPR.
Built on open source technologies and fueled by IBM Cloud, the Watson Data Platform brings together IBM’s cloud infrastructure, powerful data services and decades of experience helping clients across industries solve their data challenges. Linked closely with the most popular communities among data scientists and developers, including Python and Spark, the Watson Data Platform continues to evolve to build the most open and complete data operating system on the cloud.