SLG Seminar – Antoine Caubrière – 03/15/2024

11 March 2024

Next SLG meeting will take place on 03/15/2024, from 10 AM to 11 AM. We will host Antoine Caubrière from the company Orange, who will present his recent work Title: Representation of Multilingual Speech through Self-Supervised Learning in an Exclusively Sub-Saharan Context. Abstract: The Orange group operates in over a dozen sub-Saharan African countries with the ambition of offering services tailored to the needs of clients in this region. To provide localized and accessible services to digitally underserved and low-literate individuals, Orange is investing in the development of voice-based conversational agents to inform and assist its clients and employees.The implementation of such a service requires, first and foremost, a technological component for speech recognition and understanding.The strong linguistic diversity of the African continent, coupled with the challenges of limited annotated data, poses one of the challenges in implementing speech processing technology for these languages. One potential solution could be the utilization of self-supervised learning techniques. Leveraging this type of learning enables the training of a speech representation extractor capable of capturing rich features. This approach utilizes a large quantity of unlabeled data for pre-training a model before fine-tuning it for specific tasks. While numerous self-supervised models are shared within the Plus d'infos

Best paper award

19 February 2024

Congratulations to Grace Tessa Masse and Abderrahim Benslimane, for the best paper award they obtained at International Conference on Computing, Networking and Communications (IEEE ICNC 2024) Title: A Secure Hierarchical Federated Learning Using Dirichlet-based Trust Management Abstract—Hierarchical Federated Learning (HFL) is a distributed machine learning training system in which a server works with several clients and edge nodes while maintaining data privacy. Distributed machine learning training systems are also known as Federated Learning, but HFL is a type of Federated Learning that utilizes a hierarchical network architecture to address computational issues when dealing with a high number of clients. However, HFL is vulnerable to attacks such as data poisoning, which may jeopardize the entire training process and result in misclassifications. As system defenders, we have to tackle this issue. Using a label-flipping attack, we investigate the effect of data poisoning attacks on HFL training. We propose a trust management-based strategy to mitigate data poisoning attacks, which assesses client trustworthiness using a Dirichlet distribution. We maintain a record of previous activities, allowing the server to enhance its knowledge based on client reliability. We demonstrate the proposed approach’s effectiveness through improvements in model performance after removing malicious clients, using the MNIST dataset Plus d'infos

SLG Seminar- 15/02/2024

13 February 2024

Thibault Roux will organize a debate on the subject mentioned below: “Recent advances in technology have raised many questions and concerns about their impact on our societies. Many people are concerned about military use, mass surveillance or disinformation. From a more global perspective, Nick Bostrom, a philosopher, theorizes the vulnerable world hypothesis which predicts that science will destroy humanity.In this debate, we will question our own biases as researchers and try to answer the ethical questions raised by this hypothesis. Is science a threat to humanity? Should we stop science? Or more seriously, can we find a solution to prevent ourselves from self-destruction ?”

SLG Seminar – Ryan Whetten – 01/02/2024

25 January 2024

The next SLG meeting will take place in room S5 on Thursday, February 1st, from 12:00 PM to 1:00 PM. Ryan Whetten will present his work, and you can find a brief introduction below. ——————————————————————— Open Implementation and Study of BEST-RQ for Speech Processing Abstract: Self-Supervised Learning (SSL) has proven to be useful in various speech tasks. However, these methods are generally very demanding in terms of data, memory, and computational resources. Recently, Google came out with a model called BEST-RQ (BERT-based Speech pre-Training with Random-projection Quantizer). Despite BEST-RQ’s great performance and simplicity, details are lacking in the original paper and there is no official easy-to-use open-source implementation. Furthermore, BEST-RQ has not been evaluated on other downstream tasks aside from ASR. In this presentation, we will discuss the details of my implementation of BEST-RQ and then see results from our preliminary study on four downstream tasks. Results show that a random projection quantizer can achieve similar downstream performance as wav2vec 2.0 while decreasing training time by over a factor of two.

SLG Seminar – Paul Gauthier Noé – 18/01/2024

10 January 2024

On 18 January from 12 am, we will host a talk from Dr. Paul Gauthier Noé on « Explaining probabilistic predictions … ». The presentation will be hosted on room S6.    More details will follow   Bio: Paul Gauthier Noe just received a PhD in Computer Science in Avignon Université under the supervision of Prof. Jean-François Bonastre and Dr. Driss Matrouf. He was working for the international JST-ANR VoicePersonae project and his main research interests are Speaker verification, Bayesian decision theory, Calibration of probabilities and Privacy in Speech.

SLG Seminar – Fenna Poletiek – 12/01/2024

8 January 2024

On 12 January from 12 am, we will host a virtual talk from Dr. Fenna Poletiek from Institute of Psychology at Leiden University on « Language learning in the lab ».   The presentation will be hosted on room S6.   Abstract: Language learning in the lab Language learning skills have been considered a defining feature of humanness. In this view language cannot be acquired by mere associative or statistical learning processes, only, like many other skills are learned by human and nonhuman primates during development. Indeed, the high (recursive) complexity of human grammars have been shown to make them impossible to learn by exposure to language exemplars only. Some research suggests, however, that at least some statistical learning is recruited in language acquisition (Perruchet & Pacton, 2006). And primates have been shown to mimic complex grammatical patterns after being trained on a sequence of stimulus responses (Rey et al., 2012). We performed series of studies with artificial languages in the lab, to investigate associative and statistical learning processes that support language learning. The results thus far suggest a fine tuned cooperation between three crucial features of the natural language learning process: first, learning proceeds ‘starting small’ with short simple sentences growing in complexity Plus d'infos

PhD defense of Julio Perez-Garcia – 18 December 2023

14 December 2023

Place: University of Avignon, Campus Hannah Arendt, Salle des ThèsesDate: Monday, December 18, 2023 at 14:00. Title: Contribution to security and privacy in the Blockchain-based Internet of Things: Robustness, Reliability, and Scalability. Abstract: The Internet of Things (IoT) is a diverse network of objects or ”things” typically interconnected via the Internet. Given the sensitivity of the information exchanged in IoT applications, it is essential to guarantee security and privacy. This problem is aggravated by the open nature of wireless communications, and the power and computing resource limitations of most IoT devices. At the same time, existing IoT security solutions are based on centralized architectures, which raises scalability issues and the single point of failure problem, making them susceptible to denial-of-service attacks and technical failures. Blockchain has emerged as an attractive solution to IoT security and centralization issues. Blockchains replicate a permanent, append-only record of all transactions occurring on a network across multiple devices, keeping them synchronized through a consensus protocol. Blockchain implementation may involve high computational and energy costs for devices. Consequently, solutions based on Fog/Edge computing have been considered in the integration with IoT. This approach shifts the higher computational load and higher energy consumption to the devices with higher Plus d'infos

ANR PARFAIT Project

14 December 2023

Planning And leaRning For AI-Edge compuTing Partners: Period: 2023-2027

SLG Meeting – St Germes Bengono Obiang – 21/12/2023

12 December 2023

The next SLG meeting will be held in room S1 on Thursday, December 21st, from 12:00 PM to 1:00 PM.    We will have the pleasure of hosting St Germes BENGONO OBIANG, a PhD student in speech processing, focusing on tone recognition in under-resourced languages. He is supervised by Norbert TSOPZE and Paulin MELATAGIA from the University of Yaoundé 1, as well as by Jean-François BONASTRE and Tania JIMENEZ from LIA.   Abstract: Many sub-Saharan African languages are categorized as tone languages and for the most part, they are classified as low resource languages due to the limited resources and tools available to process these languages. Identifying the tone associated with a syllable is therefore a key challenge for speech recognition in these languages. We propose models that automate the recognition of tones in continuous speech that can easily be incorporated into a speech recognition pipeline for these languages. We have investigated different neural architectures as well as several features extraction algorithms in speech (Filter banks, Leaf, Cestrogram, MFCC). In the context of low-resource languages, we also evaluated Wav2vec models for this task. In this work, we use a public speech recognition dataset on Yoruba. As for the results, using the Plus d'infos

ANR EVA Project

1 January 2023

Explicit Voice Attributes Describing a voice in a few words remains a very arbitrary task. We can speak with a “deep”, “breathy”, “bright” or “hoarse” voice, but the full characterization of a voice would require a close set of rigorously defined attributes constituting an ontology. However, such a description grid does not exist. Machine learning applied to speech also suffers the same weakness : in most automatic processing tasks, when a speaker is modeled, abstract global representations are used without making their characteristics explicit. For instance, automatic speaker verification / identification is usually tackled thanks to the x-vectors paradigm, which consists in describing a speaker’s voice by an embedding vector only designed to distinguish speakers. Despite their very good accuracy for speaker identification, x-vectors are usually unsuitable to detect similarities between different voices with common characteristics. The same observations can be made for speech generation. We propose to carry out a comprehensive set of analyses to extract salient, unaddressed voice attributes to enrich structured representations usable for synthesis and voice conversion. Partner list: Project leader: Orange Scientific leader for LIA: Yannick Estève Start date: 01/01/2023 — End date: 31/12/2025 More

1 2