from How data quality makes IoT projects more profitable
by Anasia D’mello
Global technology spending on the Internet of Things (IoT) is expected to reach $1.2 trillion (€1 trillion) in 2022, led by industries such as discrete manufacturing $119 billion (€108 billion), process manufacturing $78 billion (€70.8 billion), transportation $71 billion (€64.5 billion) and utilities $61 billion (€55.4 billion).
Indeed, the market for Industry 4.0 products and services is expected to grow significantly over the next few years – and over 60% of manufacturers are expected to be fully connected by that time, utilising a change of technologies such as RFID, wearables and automated systems, says Ramya Ravichandar, VP Products, FogHorn.
Although the industry anticipates positive growth in current and upcoming IoT and IIoT projects, some significant challenges still need to be addressed in order to fully win customer trust and move pilot projects into successful, large-scale IoT productions. While many see connectivity limitations, security risks, and data bias, including data quantity, issues as roadblocks to IoT success, we have found data quality also plays a critical role in delivering effective IoT projects.
What is data quality – and how does it impact deployment success?
Data quality plays a vital role in the increasing adoption of IoT devices in three main ways:
Organisations can only make the right data-driven decisions if the data they use is correct and suitable for the use case at hand.
Poor-quality data is practically useless – and can lead to severe issues, such as inaccurate machine learning models, inaccurate decision-making, or deficient ROI.
Specifically, the classic problems of garbage in/garbage out resurfaced with the increase of artificial intelligence and machine learning applications.
High-quality data feeds, trains, and tunes machine learning (ML) models to empower IoT-enabled factories to make informed data-driven decisions.
For example, the unexpected failure of a steam turbine can create a critical disruption, damage, and economic loss to both the power plant and the downstream power grid. Predictive machine learning models, trained on high-quality data sets, help these industrial organisations maximise the reliability of their equipment by detecting potential failures before significant problems arise.
However, dirty data, including data that is missing, incomplete, or error-prone, leads organisations to make inconvenient, time-consuming, and expensive mistakes. In fact, according to The Data Warehouse Institute (TDWI), dirty data costs U.S. companies around $600 billion (€545 billion) every year. It is a fact that about 80% of a data scientist’s job is focused on data preparation and cleansing to ensure that the ML models provide the right insights.
Looking ahead, organisations must incorporate methodologies to ensure the completeness, validity, consistency, and correctness of its data streams to enhance insight quality, deploy effective IoT projects, and realise optimal ROI.
So, what role does edge computing play in data quality?
Industrial sensors come in many different types and collect high volumes, varieties, and velocities of data, including video, audio, acceleration, vibration, acoustic, and more. If an organisation is able to successfully align, clean, enrich and fuse all these various data streams, it can significantly improve the efficiency, health, and safety of their operations. However, to paint a complete, accurate picture of the factory operations, organisations […]
The post How data quality makes IoT projects more profitable appeared first on IoT Now – How to run an IoT enabled business.