Predictive maintenance (PdM) involves the execution of system checks at predetermined intervals to analyze equipment health. These controls are usually in the form of continual data collection (i.e., temperature, light, pressure, and sound/vibration) from equipment through the use of sensors. The results of these checks determine whether maintenance activities are required.
Aiming to Reduce Unscheduled Downtime
For manufacturers that work on tight margins and even tighter timeframes, unscheduled downtime can be a nightmare. It can cut right to the bottom line—ruining a quarter, a year, or even a company. The calculable costs are clear: ARC Advisory Group reports that the global process industry loses up to $20 billion of its annual production (about $12,500 per hour) due to unscheduled downtime. Hoping for the best and waiting for something to break is costly—it’s somewhere around 50% more expensive to repair an asset that broke in production than if the problem was identified before the failure.
So, with failures potentially harming personnel and the environment, manufacturers need to tame their complex, interdependent operations. Reliability itself can be a significant competitive advantage. But getting there requires a fresh approach.
The Promise of Predictive Maintenance
Predictive maintenance promises to enable the scheduling of corrective maintenance before an issue surfaces. It should also prevent surprise equipment failures. It shows what equipment will need maintenance and when. As a result, companies can allocate the right parts and ensure they can deploy field technicians only when needed. Instead of dealing with an overflowing schedule of unplanned failures that require immediate and time-consuming production stops, predictive maintenance helps companies schedule shorter outages when it makes sense to slow production.
Although it might be confused with preventive maintenance, predictive maintenance is different. Instead of looking at averages or comparable statistics, it looks at the condition of the equipment in real time. As a result, it can make predictions based on the actual conditions, not averages or suppositions.
Just-in-time manufacturing is the goal for most companies. It means that a company doesn’t get stuck with too much inventory and reaps profits faster because it only invests in parts or other components exactly when it needs them. Of course, it requires precise timing—and every element in the value chain needs to be ready when called on. So, a faulty piece of equipment that malfunctions at just the wrong time can cause a company to miss production quotas, lose business or even threaten the safety of a plant.
There are numerous technologies that predictive maintenance employs, including infrared, acoustic, video, and vibration analysis. It can even look at the oil that lubricates a machine to determine if it is functioning to spec.
Looking at Assets Individually and in Groups
Traditional methods monitor single machines or scattered pieces of equipment. They don’t see the entire picture. Predictive maintenance using cognitive machine learning techniques can take all the individual views of thousands of assets to build an integrated view of a factory floor, providing complete visibility and highlighting how assets and their workflows work together—so that if one asset is predicted to go down, it’s easy to understand the broader impact.
One reason that predictive maintenance is a rising trend is that it greatly reduces human errors, which can cause up to 82% of asset failures. As connected assets increase at a dizzying pace due to the IoT, industrial data is overwhelming manufacturers because human beings simply can’t absorb and process all of this data. Without technology to help them, even highly skilled data scientists will almost certainly miss some critical data points. Predictive maintenance that uses data science levels the playing field by applying cognitive techniques for sensor data analysis.
As a result, every enterprise can make automated intelligence available across all levels of decision-makers, ensuring people who need the information stay in the know.
How Progress Helps
The Progress Predictive Maintenance solution automates data science using cognitive analytics, data mining that looks at historical and real-time sensor data, and machine learning to detect and predict anomalies that will cause asset failures. Seeing into the future seems like a fanciful notion, but this is no crystal ball. It’s real and it means companies can avoid unnecessary maintenance costs, shorten disruptions, gain exceptional control over the torrent of sensor data coming from every machine.
The automated approach uses a patented meta-learning solution to detect and predict anomalies, deliver machine health insights, reduce the time required to develop and operationalize models and help data scientists be more effective. Specifically, Progress provides:
• Superior prediction accuracy: Partial data sets and generalized models significantly reduce prediction accuracy. Progress analyzes all machine data to uncover anomalies before they become failures. A built-in digital twin model accounts for environmental, operational and manufacturing factors.
• Highest equipment effectiveness: Existing approaches only detect known, repeated conditions, accounting for just 20% of equipment failures. Progress detects both repeated and random failures, representing the remaining 80%.
• Quick delivery: Traditional solutions take months to implement and tie up valuable data science resources. Progress delivers results in days with minimal data science effort.
• The ability to run complex workloads: Distributed processing lets you run complex workloads at scale while reducing the number of modeling experiments needed to devise highly accurate models.
The Progress Cognitive Predictive Maintenance Solution is powered by DataRPM Anomaly Detection and Prediction, which harnesses the power of machine learning, so asset-intensive organizations gain exceptional control over the torrent of sensor data coming from every machine.
27 en 28 maart 2019 Kom naar de zesde editie van ons jaarlijkse congres met wederom een ijzersterke sprekers line-up. In twee intensieve dagen behandelen wij belangrijke thema’s als Big Data, Agile Datawarehouse Design, Analytics, Machine Learn...
10 april 2019Praktisch seminar waarin Sander Hoogendoorn u laat zien hoe u microservices kunt inzetten in uw softwarearchitectuur.Het nieuwste architectuurprincipe microservices lijkt veelbelovend: verkorte time-to-market, schaalbaarheid, autonomie, ...
17 en 18 april 2019 De wereld van business intelligence en datawarehousing hanteert een unieke terminologie en eigen verzameling technologieën, ontwerptechnieken en producten. Voor nieuwkomers kan dit overweldigend overkomen. Want wat betekenen ...
15 mei 2019Workshop met BPM-specialist Christian Gijsels over business analyse, modelleren en simuleren met de nieuwste release van Sparx Systems' Enterprise Architect, versie 14.Intensieve cursus waarin alle basisfunctionaliteiten van Enterprise Arc...
16 mei 2019 Iedere organisatie heeft te maken met het integreren van systemen en applicaties. Maar welke technologie zet u in bij welke vorm van integratie? Guy Crets bespreekt de verschillende oplossingen voor integratie. Integratie van IT-sys...
5 en 6 juni 2019 Praktische tweedaagse workshop met internationaal gerenommeerde spreker Alec Sharp over herkennen, beschrijven en ontwerpen van business processen. De workshop wordt ondersteund met praktijkvoorbeelden en duidelijke, herbruikbar...
17 t/m 19 juni 2019Driedaagse workshop over requirements management door James Robertson en Adrian Reed. Opstellen, testen en ondubbelzinnig vastleggen van requirements. Unieke driedaagse workshop over requirements management op basis van de Volere m...
20 juni 2019 API Management gaat enerzijds over het promoten van API’s en anderzijds over het actief ondersteunen van ontwikkelaars bij het gebruik ervan. Tegelijkertijd gaat API Management over het gecontroleerd en centraal beveiligd ontsluite...