Software Defined Vehicles

Between vehicle and cloud

Edge computing as a building block of modern vehicle architectures

4 min
Abstract blue circuit board graphic with an edge computing chip in the centre
The blueprint specifies the responsibilities of OEMs, network operators, edge and cloud providers as well as service providers, and describes how they should collaborate via standardised interfaces.

Edge computing is moving to the centre of vehicle IT. With new industry blueprints and growing data demands, the architecture of connected vehicles is shifting. An overview of technologies, key players and open architectural questions.

At the end of February 2026, the Automotive Edge Computing Consortium (AECC) sent a signal that is difficult to ignore within the industry. With a new industry blueprint, the consortium aims to define a common framework for the architecture of connected vehicles. Edge, 5G, APIs, cloud services and software-defined vehicles are no longer to be considered in isolation, but as an integrated system capable of supporting connected services at scale.

But what exactly lies behind this ambition? The blueprint outlines a reference architecture with clearly defined layers and roles. It specifies the responsibilities of OEMs, network operators, edge and cloud providers as well as service providers, and describes how they should collaborate via standardised interfaces.

A blueprint for distributed automotive data architectures

Beyond the physical infrastructure, the document also addresses control functions, data flows and security mechanisms. It formulates requirements for connectivity, computing power, data management and security across company and system boundaries, as well as interoperability between different platforms. Using scenarios such as the updating of high-definition maps, the AECC illustrates how sensor data can be pre-processed within the vehicle, aggregated at the edge and further analysed in the cloud. Results from proof-of-concept projects complement the document and are intended to demonstrate that such a distributed architecture is technically feasible and suitable for large-scale mobility services.

Why the industry is moving beyond cloud-only architectures

The weight of this initiative is also reflected in the composition of the consortium. The AECC is not a small think tank, but a coalition of major players from the automotive, telecommunications and IT sectors. Members include Toyota, Intel, Ericsson and NTT, alongside infrastructure and industrial companies such as Mitsubishi Heavy Industries and the Eneos Corporation.

The breadth of participants indicates that edge computing is not about optimising individual electronic control units. Rather, it concerns the strategic design of future mobility and data ecosystems.

Against this backdrop, the AECC argues that purely centralised cloud architectures will no longer be sufficient to meet the requirements of modern mobility services in the long term. As vehicles transition towards software-defined platforms, the demand for large-scale data processing, highly available communication and AI-driven services with extremely low latency continues to grow.

Edge computing in the automotive industry: key facts

What is edge computing? Edge computing is a distributed computing architecture that moves data processing closer to where data is generated instead of relying solely on central cloud data centres.

Why is edge computing relevant for automotive? Connected and software-defined vehicles generate large volumes of data. Processing parts of this data closer to the vehicle can reduce latency and support real-time services.

Who is shaping the architecture? Companies such as Toyota, Intel, Ericsson and NTT collaborate in the Automotive Edge Computing Consortium (AECC) to define common frameworks for automotive edge architectures.

Does edge computing replace the cloud? No. Edge computing complements central cloud infrastructures by distributing workloads between vehicles, regional edge nodes and central data centres.

The blueprint therefore outlines a distributed architecture in which computing resources move closer to the vehicle and edge and cloud resources work together in an orchestrated manner.

Edge computing as a system question

The publication of a consolidated and end-to-end blueprint can be interpreted as a sign of increasing maturity within the consortium. Earlier papers focused on individual use cases or specific technical aspects. Now the attempt is being made to bring challenges, requirements, architectural concepts and implementation experiences together in a coherent overall picture.

For strategists in OEMs and suppliers, this raises a fundamental question: will edge computing become an infrastructural prerequisite for software-defined vehicles, or will it remain merely a complement to cloud computing?

Edge computing is no longer an abstract IT concept. It touches business models, data sovereignty, security architectures and ecosystem partnerships. Companies that aim to scale connected services must decide where data should be processed, who owns the infrastructure and how dependencies on hyperscalers and network operators will evolve.

At this point it becomes clear that the discussion around edge computing is not only an architectural or power question, but first and foremost a technological one. Before evaluating the strategic role edge computing may play in the future vehicle and IT ecosystem, it is worth revisiting the fundamentals: what exactly is edge computing, how does it work technically and how does it differ from classical cloud approaches?

Architectural principle within a distributed infrastructure

In an earlier whitepaper titled “Driving Data to the Edge: The Challenge of Traffic Distribution”, the AECC defines edge computing as a form of distributed computing in which applications, storage and processing power are distributed across multiple geographically dispersed systems in order to meet defined service levels.

Edge computing is therefore not a single product but an architectural principle within a distributed infrastructure. Processing capacity is no longer concentrated solely in central data centres but deliberately moved closer to where data is generated.

The goal is to meet performance requirements reliably while shortening transmission paths. In practice, this means that data does not always have to be transferred over long distances to a central cloud before it can be processed. Instead, initial processing takes place in regional or local computing environments.

The AECC refers to this as processing “in region”. Latency is reduced, network resources are relieved and bottlenecks in the core network can be minimised.

Technically, this model is based on a hierarchical infrastructure in which decentralised instances complement central data centres. Edge servers are positioned between the end system and the cloud and handle processing steps that were previously performed exclusively in central systems.

Applications can run either centrally or on distributed instances depending on their requirements. Data is not forwarded unfiltered but is first processed, aggregated or filtered before being transferred to higher levels of the infrastructure.

Efficient operation of such an architecture requires intelligent traffic management. In traditional network architectures, all traffic passes through predefined exchange points. Edge computing allows data to be redirected early to suitable decentralised computing locations, ensuring that processing takes place where it is technically or economically most efficient.

Potentials and limits of edge computing

For data-intensive systems, edge computing opens new opportunities for scaling. One key advantage lies in decoupling performance from centralisation. Applications no longer have to be orchestrated exclusively through a small number of highly concentrated data centres but can instead be distributed geographically or according to workload.

This can improve system stability under peak loads, enhance regional service quality and enable more flexible operational models. Another advantage is strategic flexibility. Companies gain additional options when designing their infrastructure. Workloads can be shifted between central and decentralised resources depending on performance, cost or availability requirements.

However, edge computing also significantly increases structural complexity. Distributed systems are more difficult to monitor, maintain and configure consistently. Coordination between infrastructure providers, network operators and platform environments requires clear interfaces and robust governance models.

Furthermore, edge computing introduces new security considerations. Attack surfaces are no longer concentrated in a few central locations but distributed across many nodes.

From an economic perspective, edge computing is not automatically efficient. While central cloud models benefit from strong economies of scale, edge infrastructures distribute investments across numerous sites. The deployment and operation of regional nodes requires additional infrastructure, partnerships and integration efforts.

Edge computing is therefore neither a simple extension of existing cloud strategies nor a replacement for them. Rather, it represents a structural transformation of IT architecture that combines technological benefits with organisational, economic and regulatory challenges.

Infrastructure as a strategic decision

With the transition towards software-defined vehicles, backend architecture becomes a strategic issue. OEMs must decide which parts of their value creation they want to centralise and which should operate within distributed infrastructures.

Edge computing shifts this decision from a purely technical optimisation problem to a question of governance and platform strategy. Who controls regional compute nodes? Who orchestrates data flows? And who is responsible for operational performance and security?

For manufacturers increasingly positioning themselves as software companies, differentiation will no longer take place solely at the vehicle platform level. The distributed IT infrastructure behind the vehicle will also become a competitive factor.

In this context, edge computing is less an additional technology than an extension of architectural responsibility.