Your submission was sent successfully! Close

Thank you for contacting us. A member of our team will be in touch shortly. Close

You have successfully unsubscribed! Close

Thank you for signing up for our newsletter!
In these regular emails you will find the latest updates about Ubuntu and upcoming events where you can meet our team.Close

Getting hands-on with AI in automotive

Tags: Automotive , CES , Event

From cloud to edge, hardware-agnostic AI/ML

In January Canonical will reconfirm its presence at CES 2025; here we will be showing a cutting-edge AI/ML demo that showcases how AI models can be trained, deployed, and updated seamlessly across various hardware platforms. This demonstration is a proof of our commitment to building hardware-agnostic solutions, empowering automotive manufacturers to integrate AI into their systems on various vehicle configurations, without being tied to any specific silicon vendor. Whether it’s predictive maintenance, in-car personalization, or even AD/ADAS features, our flexible AI/ML stack is designed to meet the needs for AI in automotive.

Platform-agnostic AI for automotive

With our cars becoming smarter and more connected, the demand for AI capabilities in the automotive industry has grown exponentially. However, one of the main challenges is the wide range of electrical and electronic vehicle architectures that different manufacturers and suppliers use. In order to solve this complex hardware dilemma, it is required to have a truly platform-agnostic AI stack that allows companies to choose the best hardware for their needs—whether it’s NVIDIA’s powerful GPUs, Intel processors, or Qualcomm chipsets—without sacrificing performance or compatibility.

One of the most significant challenges when it comes to deploying AI models in automotive at the edge is hardware fragmentation. OEMs and Tier 1s often work with different types of hardware across their supply chains, making it difficult to maintain a consistent AI experience. Having a hardware-agnostic AI stack solves this problem by running seamlessly across various platforms, reducing the complexity of integrating AI in automotive.

AI fault detection demo, from cloud to edge

One of the highlights of our CES 2025 demo will be a real-world use case demonstrating AI-driven fault detection in automotive manufacturing, from cloud-based model training to real-time deployment at the edge. This demo focuses on how AI can be used as an advantage for quality assurance (QA) tasks, process and analyze data efficiently, ensuring parts meet high-quality standards.

In this demo, we simulate a manufacturing environment where 3D-printed gears are produced for automotive use. Some of these gears are correctly printed, while others are defective due to printing errors. The AI model, trained using object recognition, is tasked with differentiating between well-manufactured and defective parts. The ability to automatically identify defective gears on the production line reduces the need for manual inspection and increases operational efficiency.

The model itself is trained based on a dataset containing images of good and defective gears, in an Ubuntu-based cloud environment running Kubeflow and MLflow on top of MicroK8s. 

Through transfer learning, the model quickly adapts and improves its fault detection capacity. Indeed, if the model misclassifies any parts, this feedback is used to retrain and refine the model. Updates are deployed back to the edge using over-the-air (OTA) updates, ensuring that manufacturers always have the most accurate version of the model running in their production environment. This allows manufacturers to accelerate the AI training process by leveraging pre-trained models tailored to their specific use case.

Once trained, the model is containerized and deployed to edge devices, which would be located on the factory floor. These edge devices run Ubuntu and process real-time data from cameras that scan the gears as they move down the production line. The AI model evaluates each part in real-time, flagging defective parts for removal, while passing those that meet the quality standards. This edge-based deployment is particularly important in the automotive industry, where low latency and immediate decision-making are critical to maintaining production efficiency.

Partnering with industry leaders

Our CES 2025 demo will also highlight our ongoing partnerships with major players in the semiconductor industry, including NVIDIA and Intel. Working closely with NVIDIA, we’ve developed AI/ML solutions that harness the latest GPU technologies, allowing for rapid model training and edge deployment that meets the high standards of AI in automotive.

By supporting multiple architectures, Canonical ensures that manufacturers have the freedom to choose the best tools for their specific use cases. This hardware-agnostic approach is a key theme throughout our AI/ML demo and will be a core focus at our CES 2025 booth.

Other potential use cases

Our demo clearly shows how this flexibility enables developers to seamlessly deploy AI models across a wide range of devices. For example, for object recognition used for advanced driver assistance systems (ADAS), the AI models trained in the cloud can be optimized for edge deployment, ensuring smooth performance in vehicles with limited computing resources. This eliminates the headache of needing to develop AI solutions that are tied to a single hardware vendor, a huge advantage for automotive manufacturers who are working with different semiconductor partners.

AI/ML training: from cloud to edge

As mentioned above, Canonical’s AI/ML stack allows developers to train models in the cloud and optimize them for deployment at the edge, such as inside vehicles. This process ensures that the AI models run efficiently on edge devices, where computing power is often more limited.

With support for Kubernetes, and containerization, our AI/ML solution offers automotive companies the flexibility to train models in different environments. This flexibility not only accelerates development but also ensures that AI models can be easily updated, scaled, or rolled back, as needed. Moreover, our OTA update approach makes it easy to deploy software updates to vehicles securely and efficiently, a critical feature for maintaining up-to-date AI models in the field.

Join us at CES 2025

As AI continues to reshape the automotive industry, Canonical’s hardware-agnostic AI/ML stack is positioned to lead the way. From seamless deployment across diverse hardware platforms to real-world applications in vehicles, our demo at CES 2025 will illustrate how AI/ML is driving the future of automotive. Join us at our booth (#10277 in the North Hall) to experience these innovations in action and learn how our solutions can help you build the next generation of intelligent vehicles.

Don’t miss out on CES 2025—let’s drive the future of AI in automotive together.

To learn more about Canonical and our engagement in automotive: 

Contact Us

Check out our webpage

Watch our webinar with Elektrobit about SDV

Download our whitepaper on V2X (Vehicle-to-Everything)

9outof10financialinstitutions

Why is Ubuntu popular with top financial institutions?

Financial institutions are increasingly pressed for agility and velocity to adapt to changing market conditions, increased customer expectations while satisfying regulatory and compliance requirements.

Learn more about Ubuntu ›

Newsletter signup

Get the latest Ubuntu news and updates in your inbox.

By submitting this form, I confirm that I have read and agree to Canonical's Privacy Policy.

Related posts

Canonical at IAA Transportation 2024 

Book a demo with our team As the automotive industry continues to accelerate towards a more connected, autonomous, and electric future, Canonical is thrilled...

Join Canonical in Brazil at Dell Technologies Forum São Paulo

Canonical is excited to be a part of the Dell Technologies Forum in São Paulo on October 30th. This exclusive event brings together industry leaders to...

Meet our EDU team at EDUCAUSE 2024

We’re excited to announce our participation in EDUCAUSE 2024.