- 데이터 정보
-
오류신고를 진행 하실 데이터 정보를 담은 표입니다. 제목(Main) FAD: A Chinese Dataset for Fake Audio Detection 제목(Sub) 저자 Haoxin Ma;Jiangyan Yi; 제공처 OpenAIRE 리포지터리 OpenAIRE
- 접수 정보
-
오류신고 접수 정보를 담은 표이며, 메일주소, 오류내용을 입력합니다. 아이디 오류신고 접수 정보를 담은 표이며, 메일주소, 오류내용을 입력합니다.오류 구분 - 개인정보 노출방지를 위해 개인정보 내용은 가급적 자제하여 주시기 바랍니다.
- 일방적인 욕설 및 부정적인 내용 작성시 원작자의 판단에 따라 신고자에게 피해가 발생할 수 있습니다. 깨끗하고 청렴한 서비스 문화를 위해 필요한 정보만 기재해주시면 감사하겠습니다.
-
2022
해외
공개
CC-BY-4.0
English
FAD: A Chinese Dataset for Fake Audio Detection
FAD: A Chinese Dataset for Fake Audio Detection Haoxin Ma;Jiangyan Yi;Fake audio detection is a growing concern and some relevant datasets have been designed for research. But there is no standard public Chinese dataset under additive noise conditions. In this paper, we aim to fill in the gap and design a
Chinese fake audio detection dataset (FAD) for studying more generalized detection methods. Twelve mainstream speech generation techniques are used to generate fake audios. To simulate the real-life scenarios, three noise datasets are selected for
noisy adding at five different signal noise ratios. FAD dataset can be used not only for fake audio detection, but also for detecting the algorithms of fake utterances for
audio forensics. Baseline results are presented with analysis. The results that show fake audio detection methods with generalization remain challenging.
The FAD dataset is publicly available. The source code of baselines is available on GitHub https://github.com/ADDchallenge/FAD
The FAD dataset is designed to evaluate the methods of fake audio detection and fake algorithms recognition and other relevant studies. To better study the robustness of the methods under noisy
conditions when applied in real life, we construct the corresponding noisy dataset. The total FAD dataset consists of two versions: clean version and noisy version. Both versions are divided into
disjoint training, development and test sets in the same way. There is no speaker overlap across these three subsets. Each test sets is further divided into seen and unseen test sets. Unseen test sets can
evaluate the generalization of the methods to unknown types. It is worth mentioning that both real audios and fake audios in the unseen test set are unknown to the model.
For the noisy speech part, we select three noise database for simulation. Additive noises are added to each audio in the clean dataset at 5 different SNRs. The additive noises of the unseen test set and the
remaining subsets come from different noise databases. In each version of FAD dataset, there are 138400 utterances in training set, 14400 utterances in development set, 42000 utterances in seen test set, and 21000 utterances in unseen test set. More detailed statistics are demonstrated in the Tabel 2.
Clean Real Audios Collection
From the point of eliminating the interference of irrelevant factors, we collect clean real audios from
two aspects: 5 open resources from OpenSLR platform (http://www.openslr.org/12/) and one self-recording dataset.
Clean Fake Audios Generation
We select 11 representative speech synthesis methods to generate the fake audios and one partially fake audios.
Noisy Audios Simulation
Noisy audios aim to quantify the robustness of the methods under noisy conditions. To simulate the real-life scenarios, we artificially sample the noise signals and add them to clean audios at 5 different
SNRs, which are 0dB, 5dB, 10dB, 15dB and 20dB. Additive noises are selected from three noise databases: PNL 100 Nonspeech Sounds, NOISEX-92, and TAU Urban Acoustic Scenes.
This data set is licensed with a CC BY-NC-ND 4.0 license.
You can cite the data using the following BibTeX entry:
@inproceedings{ma2022fad,
title={FAD: A Chinese Dataset for Fake Audio Detection},
author={Haoxin Ma, Jiangyan Yi, Chenglong Wang, Xunrui Yan, Jianhua Tao, Tao Wang, Shiming Wang, Le Xu, Ruibo Fu},
booktitle={Submitted to the 36th Conference on Neural Information Processing Systems (NeurIPS 2022) Track on Datasets and Benchmarks },
year={2022},
}
데이터 생성 이력정보
관련 과제/논문 정보
- 유발연구데이터가 유발된 과제/논문 정보입니다.
- 관련연구데이터 생산에 참고된 관련 과제/논문 정보입니다.
원문정보
- http://dx.doi.org/10.5281/zenodo.6641573 http://dx.doi.org/10.5281/zenodo.6635521 http://dx.doi.org/10.5281/zenodo.6623227
- 65 조회수
- 1 다운로드수
- 추천수 0
- 공유수 0
- 인용횟수 0
- 제공처
- 리포지터리
-
OpenAIRE
- 라이센스
- CC-BY-4.0
- 공유하기