Raven: A 100% GPU-driven AI inference framework for real-time video and graphics
ai, multimedia-edge-ai, fluendo-ai-plugins, ravenRedefining real-time AI inference with Raven's GPU-driven architecture.
Powered by our Raven AI Engine, Fluendo AI Plugins offer cross-platform and production-ready solutions optimized for Edge AI that can be used across a wide range of devices and platforms, from desktop PCs to embedded systems.


Easily integrate advanced AI features into existing GStreamer workflows without requiring extensive reconfiguration or disrupting established processes, enabling users to enhance their multimedia content effortlessly.
Run AI solutions efficiently on-device, reducing latency and maximizing resource usage for applications requiring real-time performance in resource-constrained environments like IoT and autonomous systems or offline processing in multimedia applications.
Deploy AI solutions across different hardware environments without worrying about your GPU/CPU/NPU configurations. Our plugins ensure orthogonal and optimal runtime performance in multiple adapters (NVIDIA, AMD, Intel, …)
Easily integrate advanced AI features into existing GStreamer workflows without requiring extensive reconfiguration or disrupting established processes, enabling users to enhance their multimedia content effortlessly.
Run AI solutions efficiently on-device, reducing latency and maximizing resource usage for applications requiring real-time performance in resource-constrained environments like IoT and autonomous systems or offline processing in multimedia applications.
Deploy AI solutions across different hardware environments without worrying about your GPU/CPU/NPU configurations. Our plugins ensure orthogonal and optimal runtime performance in multiple adapters (NVIDIA, AMD, Intel, …)
Each plugin comes pre-configured to handle AI tasks, requiring no AI expertise. The plugins built on Fluendo’s proprietary Raven AI Engine offer seamless integration with GStreamer.
Our system delivers high-performance solutions that work efficiently, even on low-power edge devices. Fluendo AI Plugins empower businesses to transform their video content workflows by improving processing speed, accuracy, and quality across a range of applications.
Each plugin comes pre-configured to handle AI tasks, requiring no AI expertise. The plugins built on Fluendo’s proprietary Raven AI Engine offer seamless integration with GStreamer.
Our system delivers high-performance solutions that work efficiently, even on low-power edge devices. Fluendo AI Plugins empower businesses to transform their video content workflows by improving processing speed, accuracy, and quality across a range of applications.
This plugin enables real-time background subtraction from webcam feeds, seamlessly integrating the resulting video into desktop presentations or virtual environments. It is ideal for video conferencing, live streaming, and content creation.
This plugin enables real-time background subtraction from webcam feeds, seamlessly integrating the resulting video into desktop presentations or virtual environments. It is ideal for video conferencing, live streaming, and content creation.
This plugin applies advanced multi-target detection and tracking algorithms to blur multiple objects, preserving privacy and confidentiality. Particularly valuable for video surveillance, public safety, and video-sharing platforms. Read more about our Anonymizer.
This plugin applies advanced multi-target detection and tracking algorithms to blur multiple objects, preserving privacy and confidentiality. Particularly valuable for video surveillance, public safety, and video-sharing platforms. Read more about our Anonymizer.
This plugin employs a Generative Adversarial Network (GAN) to upscale video and image resolutions by 4x, enhancing the quality of low-resolution content. This is crucial for media restoration, video streaming, and any image clarity applications.
This plugin employs a Generative Adversarial Network (GAN) to upscale video and image resolutions by 4x, enhancing the quality of low-resolution content. This is crucial for media restoration, video streaming, and any image clarity applications.
Our standalone AI plugins are designed to work independently or as part of a larger video processing pipeline, giving you the flexibility to integrate only the features your project requires.
our use cases
These use cases present conceptual examples of how our ideas and technologies could address real-world industry challenges.

Historic sports footage represents a valuable asset for broadcasters, clubs, sports federations, and media archives. However, many classic matches were recorded in low resolutions, analog formats, or early digital standards, resulting in blurry visuals, limited detail, and reduced viewing quality on modern displays.
Manually restoring and enhancing these recordings is complex, time-consuming, and often requires specialized post-production workflows. As demand grows for remastered sports content, documentaries, digital archives, and modern streaming platforms, improving the quality of historic footage has become increasingly important.
With advancements in AI-based superresolution, computer vision models can reconstruct missing details and upscale legacy sports videos to higher resolutions. By processing video frames automatically, such systems can transform historic recordings into clearer, sharper versions suitable for modern screens while preserving the authenticity of the original footage.

Youth and academy sports organizations increasingly record matches and training sessions for performance analysis, coaching review, and player development. However, these recordings often include minors, spectators, and staff members, creating important privacy and regulatory challenges, particularly under frameworks such as GDPR and child protection policies.
Manually anonymizing individuals in sports footage is time-consuming and difficult to scale, especially when clubs, academies, or federations manage large volumes of recorded matches, training sessions, and archived content.
With advances in AI-based person detection, tracking, and segmentation, computer vision systems can automatically detect individuals appearing in sports video and apply anonymization techniques such as face blurring or body masking. At the same time, the system can preserve useful movement and positional metadata, enabling coaches and analysts to study gameplay, tactics, and player behavior without exposing personal identities.

Sports clubs and academies increasingly produce live video streams, interviews, commentary shows, and behind-the-scenes content for digital platforms and social media. These broadcasts often take place in training grounds, stadiums, or mixed media areas, where children, staff members, and spectators may appear in the background.
This creates important privacy and safeguarding challenges, particularly in youth sports environments where minors must not be publicly identifiable without explicit consent.
Manual editing or post-production anonymization is not feasible for live broadcasts or real-time streaming, where video must be processed instantly before distribution.
With advances in AI-based person and face detection, video processing systems can automatically identify individuals appearing in the background of a live stream and apply anonymization techniques such as face blurring or masking in real time. This allows sports organizations to safely broadcast interviews, live shows, and training content while protecting the identity of children and other individuals present in the scene.

Sports clubs, broadcasters, and analytics teams increasingly rely on data-driven insights to understand player performance, tactical behavior, and match dynamics. Traditionally, these metrics are collected using dedicated tracking systems or manual annotation workflows, which can be costly, complex to deploy, and difficult to integrate into existing video infrastructures.
Recent advances in AI-based computer vision enable sports analytics to be derived directly from video. By analyzing match footage frame by frame, AI models can estimate player positioning, movement trajectories, and spatial relationships across the field.
These insights can be transformed into valuable performance metrics such as distance traveled, top speed, positioning patterns, and heatmaps, enabling coaches, analysts, and media teams to better understand player behavior and team dynamics.
When combined with modern multimedia pipelines, these AI-generated insights can be embedded directly into the video stream as structured metadata, enabling downstream systems to access analytics data in real time without requiring separate data pipelines.

The Catalan audiovisual sector currently faces a significant technological gap regarding high-quality, professional-grade synthetic voice solutions. Foundation Text-to-Speech (TTS) technologies work effectively for the majority of languages with large numbers of speakers, such as English, Chinese, and Spanish. Nevertheless, for regional languages like Catalan, these models do not deliver adequate performance and fail to meet the needs of the phonetic nuances and dialectal diversity required for professional dubbing and media production. Furthermore, the rise of generative AI has raised urgent concerns about the privacy of biometric data and the intellectual property rights of voice actors.
Due to these challenges, the industry would require a secure, sovereign, and ethically grounded platform to generate high-fidelity Catalan synthetic speech. It would be essential to establish a system that not only achieves naturalness through advanced Computer Vision and MLOps principles but also ensures total compliance with the EU AI Act and GDPR.
Catalan serves as a proof of concept for sovereign technology in other underrepresented European regional languages.
Bits & Bytes
Explore our blog, one byte at a time. Our team unpack our latest news, industry insights and in-depth articles to connect you with the multimedia world.Redefining real-time AI inference with Raven's GPU-driven architecture.
Efficient bounding box post-processing using GPU-native NMS-Raster.
Real-time AI video anonymization for 4K high-resolution content.
Performance benchmark analysis of real-time 4K AI video anonymization.