Research and technology development are central to Vocapia's activities. As an R&D company, we are involved in a number of collaborative multiparty R&D projects, both at the national and international levels. We are always open to participating in challenging projects. Below is a short description of some of our recent projects.

ALADAN

Vocapia is coordinating the EU EDF project ALADAN. The consortium is comprised of 4 SMEs from 4 EU counties, embarking on a new collaboration with complementary expertise: Crowdee (Germany), Lingea (Czech Republic), Ianus Consulting (Cyprus), and Vocapia (France). The project goals are to design and develop a disruptive framework of development for Artificial Intelligence (AI) based language solutions for defence applications. Four language technologies are being addressed: spoken language identification, speech recognition, spoken term search, and text and speech translation. The proposed novel framework uses open data for model training and only needs small amounts of application-specific data for validation. The ambitious project aims to substantially reduce the costs (time and financial) of developing language technologies for military applications, thereby drastically reducing the barriers for their wider uptake.

KOIOS

Vocapia is a partner of the EU EDF KIOIS project. KOIOS is an interdisciplinary project that develops Frugal Learning methods, i.e., AI techniques to be applied in specific military use cases. The project represents an ambitious investigation initiative that will produce new AI techniques beyond the current state of the Art in different research domains (few shot transfer learning, zero shot leaning, synthetic data for AI, semi or non-supervised learning, domain adaptation, ...).
Vocapia leads the KOIOS activities on applying frugal machine learning techniques to extract information from audio data for further analysis by intelligence analysts. In an ever changing geopolitical situation, the targeted languages, dialects, types and quality of recordings are constantly changing, which directly impacts the requirements for speech technologies with a need for quick development or adaptation. A priority is addressing the robustness of the speech technologies to changes in acoustic conditions and targeted domain(s). For example, to address target data frugality, the acoustic characteristics of the recordings are taken into account to simulate training data.

PRESERVE

The PRESERVE project is developing a trustworthy, transparent, and easy-to-use Hybrid C-UAS C2 platform to support Police Authorities with the prevention, early detection and optimal management of operational response against current and emergent threats in drone technology, increasing the security of citizens in public spaces and contributing significantly to the novel EU Counter-UAS policy. By fusing heterogeneous data from an array of physical sensors and online sources, the project aims to tackle the so far unmanageable threat of weaponised consumer drone swarms through open-source knowledge by non-state actors. Vocapia is customizing their speech technologies to meet the user needs of the pilot use cases. In particular, Vocapia contributes to the creation of a tool for the discovery and continuous monitoring of open data from different external sources, leading activities on automatic speech and speaker recognition in multimedia content. The aim is to detect content of interest in audio and video data from the web, e.g., online training camps. They are adapting their existing AI-based language-dependent solutions to cover highly varied types of speech, accents and acoustic conditions. The structured text produced by the automatic speech recognition component enables further downstream processing to flag radicalized speech across large audio/video collections. The speaker recognition component provides robust intra- and inter-document speaker clustering, and capabilities to train speaker-specific models to identify target speakers.

TILADI

Vocapia is coordinating the DGA TILADI project along with Airbus D&S. The goal of TILADI is to improve the performance and robustness of language technologies for military needs and under a number of very challenging conditions, including dialectal and accentuated speech, under-resourced languages, and very noisy communications. The targeted technologies are language identification, speech transcription and speech translation. For language identification, about 100 languages and 50 dialects are being modeled. An experimentatal platform has been developed integrating all the project results and is available for user trials.