Red Hat Device Edge brings flexible production grade Kubernetes and Linux to resource constrained edge devices

Red Hat has introduced Red Hat Device Edge, a solution for flexibly deploying traditional or containerized workloads on small devices such as robots, IoT gateways, points of sale, public transport and more.

  • 1 year ago Posted in

Red Hat Device Edge delivers an enterprise-ready and supported distribution of the Red Hat-led open source community project MicroShift, a lightweight Kubernetes orchestration solution built from the edge capabilities of Red Hat OpenShift, along with an edge-optimized operating system built from Red Hat Enterprise Linux. This latest product in the Red Hat edge portfolio aims to provide a future-proof platform that allows organizations’ architecture to evolve as their workload strategy changes.

As more and more companies deploy edge computing across a broader range of use cases, many new questions, operational needs and business challenges are poised to arise. In industries like automotive, manufacturing and more, organizations are up against different environmental, security and operational challenges that require an ability to work with small form-factor edge devices in these resource constrained environments. Ultimately, different devices have different requirements in terms of computing power, software compatibility and security footprint.

With Red Hat Device Edge, organizations can have the flexibility to deploy containers at the edge in a small footprint, reducing compute requirements by up to 50% in comparison to traditional Kubernetes edge configurations. It also helps to address many of the emerging questions around large-scale edge computing at the device edge by incorporating:

Kubernetes built for edge deployments, enabling IT teams to use familiar Kubernetes features in a new, smaller, lighter-weight footprint offered by MicroShift. This lowers the barrier of entry for teams building cloud-native applications for edge computing environments and enables them to use existing Kubernetes skills to achieve greater consistency of operations across the entirety of the hybrid cloud, from the datacenter to public clouds to the edge.

An edge-optimized Linux OS built from the world’s leading enterprise Linux platform in Red Hat Enterprise Linux and tailored for small edge devices with intelligent updates that use minimum bandwidth. This helps organizations tackle the challenges of intermittent connectivity while mitigating the impact on edge innovation.

Capabilities for centrally scaling and monitoring edge device fleets with Red Hat Smart Management. IT teams can use zero-touch provisioning, system health visibility and updates with automatic rollbacks to maintain a stronger edge management and application security posture.

Red Hat Device Edge for far-flung, resource constrained use cases across different industries

Red Hat Device Edge was created to help Red Hat customers and partners tackle their most challenging edge environments. For example, Lockheed Martin has been collaborating with Red Hat in the MicroShift project community and is also working to deploy Red Hat Device Edge to modernize and standardize its application delivery and AI workloads in extreme conditions including wildland fire management, contested military environments and space. Additionally, ABB is planning to use Red Hat Device Edge for ABB Ability™ Edgenius™ on resource constrained devices. Edgenius is a comprehensive edge platform for industrial software applications.

Red Hat Device Edge will be aimed at organizations who require small factor edge devices with support for bare metal, virtualized or containerized applications, regardless of industry. Additional use cases include but are not limited to:

Miniature, connected nodes on public transportation where edge devices are often in motion but still need faster processing via AI/ML to analyze data locally in real time (i.e. railways, mining, cars, drones).

Resilient resource nodes at challenging locations like weather monitoring stations where, in spite of the harsh, tough-to-support environments, an edge device will still be capable of taking care of itself with the ability to perform automated software rollbacks, maintain a stronger security posture and better enforce sensitive data controls.

Emerging edge constrained scenarios where thousands of edge devices may be running applications in locations that make weight, temperature and connectivity all major concerns.

Red Hat Device Edge meets organizations wherever they are today in their edge computing journey, as it will run a wide variety of workloads using Podman for edge container management or MicroShift for a Kubernetes API. Customers will even be able to use legacy windows applications within a virtual machine.

Although most remain “unsure how it actually works”, 40% of C-level executives are planning to use AI and the advantages that can be gained through Generative AI (Gen AI) such as ChatGPT to cover critical skills shortages, according to new research by Kaspersky.
Civo has published the results of its research into the challenges faced by Machine Learning (ML) developers in their roles. With more businesses deploying ML, the research highlights the current hurdles faced and the high rate of project failure.
Now Assist in Virtual Agent, flow generation, and Now Assist for Field Service Management are the latest in powerful GenAI solutions to be embedded into the ServiceNow Platform.
Report unveils AI adoption rates for 2024 along with other tech and customer experience predictions.
New innovations in cloud threat detection give SOC teams the edge to pinpoint suspicious activity across their attack surface.
A new report from Appsbroker and CTS shows 86% of organisations have already been impacted by GenAI, with a better understanding of the potentially disruptive impacts of the technology a top priority for 78% of them.
LANL adopts SambaNova Suite and expands existing SambaNova DataScale® deployment to run AI workloads for performing national security, science, technology, and engineering projects.
Collaboration brings together companies’ deep expertise and research in AI and large language models, powered by NVIDIA AI Enterprise software.