30 Aug, 2022 | Strategy

Stream processing at the edge makes “real-time” even faster

It turns out that “real time” is a bit of a misnomer — there’s always some lag between an event and its processing — even if it’s only a few milliseconds. But we can help decrease even that item.

Words by
Mike Rosam, CEO & Co-Founder
Blog 185 feature

If you’re seeking the fastest possible processing speed, “real time” might sound like the finish line. How can a system work quicker than that? It turns out that “real time” is a bit of a misnomer. There’s always some lag between an event and its processing — even if it’s only a few milliseconds.

Why work toward faster processing speeds?

If we can already process data in milliseconds, why do we need to go faster? A gamer might not perceive a difference of a dozen milliseconds. Still, higher latency in their gaming system can mean the difference between dodging an opponent’s death blow and being hit. Game over. It’s not just gaming systems that need to be fast. Self-driving cars, gunshot detection systems, medical alerts, and accident prevention work better when the data is processed faster. Just as gamers sit closer to their routers to pick up a few milliseconds, companies can use edge technology to reduce the distance data travels and increase processing speed.

Check out this simple game made with Quix’s stream processing solution for more on latency in gaming.

Processing data closer to the point of action reduces latency

Reducing the distance between processing and the event taking place reduces latency and increases speed, which has led to new technologies that operate between the cloud and the end user or event. These technologies fall into one of two groups: fog and edge.

Cloud fog and edge graphic

How much faster is processing at the edge?

A group of researchers measured latency from 8,456 users to 6,341 Akamai edge servers and 69 cloud locations. The researchers found that edge servers can provide lower latency than cloud servers, as shown in the table below.

Edge vs cloud latency

What are some examples of edge processing?

Whether you use cloud, fog or edge processing will depend on your use case and tolerance for latency. Edge processing is a good choice for applications that rely on instant responses to data. Let’s look at some examples.

Covid detection and prevention in the workplace

Timing plays a significant role in Covid detection and prevention. The sooner you learn that you had contact with someone contagious, the more quickly you can act. Some companies use edge processing to decrease the risk of catching covid in the office. Bob Violino, a contributing writer at Network World, provides several examples in IoT helps make return-to-work safer.

Innovatus Capital Partners displays real-time, air-quality scores on screens throughout the facility to reassure employees. The system combines technologies from Veea edge-computing and Wynd Technologies, a leader in portable air purifiers. By placing small processing devices called Smart Edge Nodes throughout the building, Innovatus is able to collect, process data, and display data from Wynd IoT sensors in real time. Data is also sent to Wynd Technologies for further analysis.

Nutrien Ltd., a provider of farming products, used proximity-monitoring and contact-tracing technology from Triax Technologies to help enforce social distancing. Wearable trace tags were attached to employees’ hardhats or clothing. The tags flash if employees come within 6 feet of each other, providing an instant warning. Gateways strategically placed around the site collect data from the tags for use in contact tracing, reducing the manual work and helping the company stop potential outbreaks faster.

In both of these examples, edge processing reduces latency and traffic to the cloud. The closed system helps meet data security requirements for sensitive data, like employee locations.

Time-sensitive manufacturing processes

Some processes are more finicky than others, requiring constant monitoring and adjustments to produce the desired results. Pentair, a food and beverage industry filtration expert, already offers solutions that consolidate complex data streams and enable teams to optimize filtration processes. The company’s vision for the future is a self-adapting beer membrane filtration system.

The system, called BrewAssist, will use advanced technologies to autonomously find and implement the best filtration settings throughout the brewing process. This self-adapting system will reduce time spent on data extraction, modeling, and manual responses to optimize beer filtration in real time.

Preventative maintenance

A broken machine can mean workplace accidents, downtime and costly repairs. Preventative maintenance programs predict when parts should be replaced or serviced to keep equipment running optimally. Deploying IIoT sensors and onsite processing takes this one step further to enable companies to detect an imminent failure and shut machines down before an accident occurs. These systems can yield significant cost savings. Texmark Chemicals expects its IIoT-enable predictive maintenance program to reduce maintenance costs by 50 percent.

The hardware and software of edge processing

Tapping into the benefits of edge processing requires both the hardware — IIoT sensors, processing nodes, small data centers, networking infrastructure — and software that can collect, process and transform data streams in real time. The stream processing component is where many companies get stuck. Stream processing isn’t one thing. It’s a sophisticated combination of technologies, including data pipelines and message brokers. Figuring out the components and getting them to work together reliably is hard. Think years, not months, of work. This is where Quix fits in. We’ve built a reliable stream processing infrastructure and made it easy for developers to build their backend on it. With Quix, you can implement the sensors and processors you need and be sure that every bit of data is processed with as few milliseconds of lag as possible.


Talk to a technical expert about your use case if you’re considering using stream processing in your business.

Book a demo
words by
Mike Rosam, CEO & Co-Founder

Mike Rosam is Co-Founder and CEO at Quix, where he works at the intersection of business and technology to pioneer the world's first streaming data development platform. He was previously Head of Innovation at McLaren Applied, where he led the data analytics product line. Mike has a degree in Mechanical Engineering and an MBA from Imperial College London.

Previous Post Next Post

Related content

View all
0182 feature
Strategy | 27 Oct, 2022
Get more from IIoT with streaming data integrations
As the internet of things evolves, so does its application in industrial settings. Find out how to take Industrial IoT from a helpful tool to transformative technology with streaming data integrations.
words by
Mike Rosam, CEO & Co-Founder
Why Io T projects fail 02
Strategy | 11 Oct, 2022
Why IoT projects fail — and what you can do differently to succeed
If your last IoT project wasn’t wholly successful or you’d like to progress along the learning curve faster, we’ve got you covered.
words by
Mike Rosam, CEO & Co-Founder
Blog 191 feature
Strategy | 7 Jun, 2022
To stream, or not to stream? A conversation about when to use (and how we built) stream processing systems
Behind the scenes building McLaren’s F1 stream processing system, and why the world needed a new stream processing client library.
Kiersten Thamm
words by
Kiersten Thamm, Head of Technical Content

The Stream

Updates to your inbox

Get the data stream processing community's newsletter. It's for sharing insights, events and community-driven projects.

Background image