A blog by Oleg Shilovitsky
Information & Comments about Engineering and Manufacturing Software

How Edge computing can help to manage distributed PLM processes

How Edge computing can help to manage distributed PLM processes
Oleg
Oleg
24 August, 2016 | 4 min for reading

edge-computing

“Single version of truth” is one of the most popular paradigm developed by PLM vendors for the last 10-15 years. The concept is loved by many companies and it has many powerful points. At the same time, the devil is in details and wrong implementation strategy can diminish the value of PLM single version of truth. One of the biggest challenges PLM implementation might have is related to distribution of manufacturing companies across the globe and distribution of product development and manufacturing processes. Replication of data is costly and not always possible. Most of existing PLM architectures are relying on variety of data synchronization cycles. A question how to manage distributed product development processes is one that on the table for many IT and PLM system managers.

Forbes article Will analytics on the edge be the future of big data brings an interesting perspective of “Edge computing”. If you haven’t heard about that, take a look on another article – What is Edge Computing by TechTarget for additional explanations.

In a nutshell, the idea of edge computing is simple – process data near the data source and optimize traffic by splitting data transfer between the data that needs to be synchronized immediately and information that will be transferred later for different purposes such as archiving and record keeping. The following passage can give some examples how edge computing can be applied to IoT domain.

Sometimes known as distributed analytics, it basically means designing systems where analytics is performed at the point where (or very close to where) the data is collected. Often, this is where action based on the insights provided by the data is most needed. Rather than designing centralized systems where all the data is sent back to your data warehouse in a raw state, where it has to be cleaned and analyzed before being of any value, why not do everything at the “edge” of the system?

A simple example would be a massive scale CCTV security system, with perhaps thousands or tens of thousands of cameras covering a large area. It’s likely that 99.9% of the footage captured by the cameras will be of no use for the job it’s supposed to be doing – e.g. detecting intruders. Hours and hours of still footage is likely to be captured for every second of useful video. So what’s the point of all of that data being streamed in real-time across your network, generating expense as well as possible compliance burdens?

Wouldn’t it be better if the images themselves could be analyzed within the cameras at the moment it is captured, and anything found to be useless either discarded or marked as low priority, freeing up centralized resources to work on data of actual value?

It made me think about PLM and distributed processes management. Modern process management for product development, manufacturing and supply chain can apply similar principles to enable data proliferation and distributed data processing. Think about heavy CAD file located on a desktop. Fast data processing can be applied to extract valuable data needed for decision making while rest of data will be delivered later. Similar to that manufacturing process analytic and supply chain optimization can taken faster by recombining fast processed data across the globe. While long synchronization of data can take time, fast processing of data on edges of distributed system can improve decision process and enable processes that took 24 hours cycle before.

What is my conclusion? We are coming to the era of new version of “single point of truth”. The truth is distributed and processed over the time. Companies are going to discover new data processing paradigms relying on availability of powerful computing devices everywhere and network computing. Existing PLM platforms are mostly database driven and relying on centralized data processing. New concepts of parallel data processing can combine edge computing with asynchronous data processing and analysis to deliver better process and decision management to manufacturing companies. Just my thoughts…

Best, Oleg

Want to learn more about PLM? Check out my new PLM Book website.

Disclaimer: I’m co-founder and CEO of openBoM developing cloud based bill of materials and inventory management tool for manufacturing companies, hardware startups and supply chain. My opinion can be unintentionally biased.

Image credit Forbes article 

Recent Posts

Also on BeyondPLM

4 6
8 February, 2025

Saturday is the time to relax, think, and read. It is also a day to plan. Changes are hard and...

13 May, 2015

I’ve been reading a blog Getting PLM right: no one answer by Monica Schnitger. It is a slick writeup started from...

16 February, 2025

Many years ago, I developed applications for AutoCAD to support product development process and analyze data and customer feedback in...

9 June, 2024

The reality of every engineering team or manufacturing enterprise is multiple systems. For the last 20+ years, the question of...

10 April, 2011

I read the following Reuter article: Microsoft Toyota team team up on digital auto network. It made me think about...

18 July, 2021

How big is the future SaaS/cloud PLM market? As much as this number sounds simple, the answer is not easy...

13 March, 2009

As you know, new ideas often come from things already forgotten… Therefore, I like looking at research projects and experimental...

1 February, 2020

PLM for small and medium-sized companies is driving more attention and discussions recently. My article Can Aras PLM Scale Down...

19 April, 2013

Technological predictions are tough and nobody wants to make them. Back in 2010, I came with the following post –...

Blogroll

To the top