From:Nexdata Date: 2024-08-14
The rapid development of artificial intelligence is inseparable from the support of high-quality data. Data is not only the fuel that drives the progress of AI model learning, but also the core factor to improve model performance, accuracy and stability. Especially in the field of automatic tasks and intelligent decision-making, deep learning algorithms based on massive data have shown their potential. Therefore, having well-structured and rich datasets has become a top priority for engineers and developers to ensure that AI systems can perform well in a variety of different scenarios.
As more and more voice interactive devices are put into our daily life, the issue of voice enhancement has gradually attracted the attention of scholars from all over the world. They proposed a large number of speech enhancement algorithms, including signal processing-based methods, modeled spectral estimation methods, and supervised learning methods.
Speech enhancement refers to the technology of extracting useful speech signals from background noise to suppress and reduce noise interference when speech signals are interfered and submerged by various noises. However, due to the random nature of interference, it is almost impossible to extract completely pure speech from noisy speech.
Speech enhancement aims to improve the quality and intelligibility of speech by utilizing signal processing algorithms. It mainly includes 1. Speech de-reverberation, reverberation is caused by the reflection of the sound signal by the space environment; 2. Speech noise reduction, the interference mainly comes from various environmental and human noises; 3. Speech separation, the noise mainly comes from Voice signals of other speakers. Improve the quality of speech by removing these noises or human voices. Speech enhancement technology has been used in real life, such as telephones, speech recognition, hearing aids, VoIP, and teleconferencing systems.
As the world's leading data service provider, Nexdata developed noise data, covering multiple application scenes, such as smart home, in-vehicle, and public place, to facilitate the research and development of speech enhancement technology.
1,297 Hours - Scene Noise Data by Voice Recorder
Scene noise data, with a duration of 1,297 hours. The data covers multiple scenarios, including subways, supermarkets, restaurants, roads, etc.; audio is recorded using professional recorders, high sampling rate, dual-channel format collection; time and type of non-noise are annotated. this data set can be used for noise modeling.
531 Hours – In-Car Noise Data by Microphone and Mobile Phone
531 hours of noise data in in-car scene. It contains various vehicle models, road types, vehicle speed and car window close/open condition. Six recording points are placed to record the noise situation at different positions in the vehicle and accurately match the vehicle noise modeling requirements.
20 Hours Microphone Collecting Radio Frequency Noise Data
The data is collected in 66 rooms, 2-4 point locations in each room. According to the relative position of the sound source and the point, 2-5 sets of data are collected for each point. The valid time is 20 hours. The data is recorded in a wide range and can be used for smart home scene product development.
10 Hours - Far-filed Noise Speech Data in Home Environment by Mic-Array
The data consists of multiple sets of products, each with a different type of microphone arrays. Noise data is collected from real home scenes of the indoor residence of ordinary residents. The data set can be used for tasks such as voice enhancement and automatic speech recognition in a home scene.
On the road to intelligent future, data will always be an indispensable driving force. The continuous expanding and optimizing of all kinds of datasets will provide a broader application space for AI algorithms. By constant exploring new data collection and annotation methods, all industries can better handle complex application scenarios. If you have data requirements, please contact Nexdata.ai at [email protected].