{"@context":{"@vocab":"https://cir.nii.ac.jp/schema/1.0/","rdfs":"http://www.w3.org/2000/01/rdf-schema#","dc":"http://purl.org/dc/elements/1.1/","dcterms":"http://purl.org/dc/terms/","foaf":"http://xmlns.com/foaf/0.1/","prism":"http://prismstandard.org/namespaces/basic/2.0/","cinii":"http://ci.nii.ac.jp/ns/1.0/","datacite":"https://schema.datacite.org/meta/kernel-4/","ndl":"http://ndl.go.jp/dcndl/terms/","jpcoar":"https://github.com/JPCOAR/schema/blob/master/2.0/"},"@id":"https://cir.nii.ac.jp/crid/1361699993835302016.json","@type":"Article","productIdentifier":[{"identifier":{"@type":"DOI","@value":"10.1109/icassp.2016.7471631"}},{"identifier":{"@type":"URI","@value":"http://xplorestaging.ieee.org/ielx7/7465907/7471614/07471631.pdf?arnumber=7471631"}}],"dc:title":[{"@value":"Deep clustering: Discriminative embeddings for segmentation and separation"}],"creator":[{"@id":"https://cir.nii.ac.jp/crid/1381699993835302017","@type":"Researcher","foaf:name":[{"@value":"John R. Hershey"}]},{"@id":"https://cir.nii.ac.jp/crid/1381699993835302018","@type":"Researcher","foaf:name":[{"@value":"Jonathan Le Roux"}]},{"@id":"https://cir.nii.ac.jp/crid/1381699993835302019","@type":"Researcher","foaf:name":[{"@value":"Shinji Watanabe"}]},{"@id":"https://cir.nii.ac.jp/crid/1381699993835302016","@type":"Researcher","foaf:name":[{"@value":"Zhuo Chen"}]}],"publication":{"prism:publicationName":[{"@value":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"}],"dc:publisher":[{"@value":"IEEE"}],"prism:publicationDate":"2016-03","prism:startingPage":"31","prism:endingPage":"35"},"reviewed":"false","url":[{"@id":"http://xplorestaging.ieee.org/ielx7/7465907/7471614/07471631.pdf?arnumber=7471631"}],"createdAt":"2016-06-24","modifiedAt":"2022-07-02","relatedProduct":[{"@id":"https://cir.nii.ac.jp/crid/1360009142833550208","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"FastMVAE: A Fast Optimization Algorithm for the Multichannel Variational Autoencoder Method"}]},{"@id":"https://cir.nii.ac.jp/crid/1360009142866227072","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Do We Need Sound for Sound Source Localization?"}]},{"@id":"https://cir.nii.ac.jp/crid/1360013168794878336","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Time-Frequency-Bin-Wise Linear Combination of Beamformers for Distortionless Signal Enhancement"}]},{"@id":"https://cir.nii.ac.jp/crid/1360013168862613504","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Time-Domain Audio Source Separation With Neural Networks Based on Multiresolution Analysis"}]},{"@id":"https://cir.nii.ac.jp/crid/1360021390561283840","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Exploring the Integration of Speech Separation and Recognition with Self-Supervised Learning Representation"}]},{"@id":"https://cir.nii.ac.jp/crid/1360021390747765760","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"DNN-Based Near- and Far-Field Source Separation Using Spherical-Harmonic-Analysis-Based Acoustic Features"}]},{"@id":"https://cir.nii.ac.jp/crid/1360025429416557440","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Deep-learning-based Single-channel Sound Source Separation in Noisy Environments"}]},{"@id":"https://cir.nii.ac.jp/crid/1360025431136416512","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Joint Music Segmentation and Clustering Based on Self-Attentive Contrastive Learning of Multifaceted Self-Similarity Representation"}]},{"@id":"https://cir.nii.ac.jp/crid/1360294643791286528","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Improved Transcription and Speaker Identification System for Concurrent Speech in Bahasa Indonesia Using Recurrent Neural Network"}]},{"@id":"https://cir.nii.ac.jp/crid/1360298757436144000","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Sampling-Frequency-Independent Convolutional Layer and its Application to Audio Source Separation"}]},{"@id":"https://cir.nii.ac.jp/crid/1360302864803354240","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Audio Signal Processing in the 21st Century: The important outcomes of the past 25 years"}]},{"@id":"https://cir.nii.ac.jp/crid/1360572092786971904","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Majorization-Minimization Algorithm for Discriminative Non-Negative Matrix Factorization"}]},{"@id":"https://cir.nii.ac.jp/crid/1360572092882674432","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Harmonic-Temporal Factor Decomposition for Unsupervised Monaural Separation of Harmonic Sounds"}]},{"@id":"https://cir.nii.ac.jp/crid/1360584339756727168","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Acceleration of Sound Source Separation by Time Direction Division-Based Deep Neural Network"}]},{"@id":"https://cir.nii.ac.jp/crid/1360848655242575232","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Environmental sound processing and its applications"}]},{"@id":"https://cir.nii.ac.jp/crid/1360853567706089728","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Perceptual-Similarity-Aware Deep Speaker Representation Learning for Multi-Speaker Generative Modeling"}]},{"@id":"https://cir.nii.ac.jp/crid/1361131420293001600","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"Independent Deeply Learned Matrix Analysis for Determined Audio Source Separation"}]},{"@id":"https://cir.nii.ac.jp/crid/1361694370252824960","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@value":"MultiSpectralNet: Spectral Clustering Using Deep Neural Network for Multi-View Data"}]},{"@id":"https://cir.nii.ac.jp/crid/1390295658299596928","@type":"Article","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Fundamentals and Trends on Sound Source Separation : Overview of Approaches with Probabilistic Model and Deep Learning"},{"@language":"ja","@value":"音源分離技術の基礎と動向"}]},{"@id":"https://cir.nii.ac.jp/crid/1390304029121072512","@type":"Article","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Target sound information extraction: Speech and audio processing with neural networks conditioned on target clues"}]},{"@id":"https://cir.nii.ac.jp/crid/1390573242790286336","@type":"Article","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Speaker-Independent Audio-Visual Speech Separation Based on Transformer in Multi-Talker Environments"}]},{"@id":"https://cir.nii.ac.jp/crid/1390848250124839168","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"A Two-Stage Phase-Aware Approach for Monaural Multi-Talker Speech Separation"}]},{"@id":"https://cir.nii.ac.jp/crid/1390851497212496640","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Single-Channel Multispeaker Separation with Variational Autoencoder Spectrogram Model"}]},{"@id":"https://cir.nii.ac.jp/crid/1390852514694806912","@type":"Article","resourceType":"学術雑誌論文(journal article)","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Real-Time Binaural Target Speech Extraction Using Phase Unwrapping"},{"@language":"ja","@value":"位相アンラッピングを用いた両耳マイクロホンによるリアルタイム目的音源抽出"},{"@language":"ja-Kana","@value":"イソウ アンラッピング オ モチイタ リョウミミ マイクロホン ニ ヨル リアルタイム モクテキオン ゲン チュウシュツ"}]},{"@id":"https://cir.nii.ac.jp/crid/1390855201285323776","@type":"Article","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Improved Speech Separation Performance from Monaural Mixed Speech Based on Deep Embedding Network"}]},{"@id":"https://cir.nii.ac.jp/crid/1390860764354702976","@type":"Article","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"en","@value":"Deep encoder and decoder for time-domain speech separation"}]},{"@id":"https://cir.nii.ac.jp/crid/1390866647407670016","@type":"Article","relationType":["isReferencedBy"],"jpcoar:relatedTitle":[{"@language":"ja","@value":"混ざった声を聞き分ける最新技術：音源分離と目的音声抽出"},{"@language":"en","@value":"Listening to Speech in Mixture: Advances in Source Separation and Target Speech Extraction"}]}],"dataSourceIdentifier":[{"@type":"CROSSREF","@value":"10.1109/icassp.2016.7471631"},{"@type":"CROSSREF","@value":"10.2299/jsp.25.145_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1007/978-3-030-69544-6_8_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/tcss.2019.2926450_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslp.2021.3059114_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslp.2021.3072496_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslp.2021.3126950_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/waspaa58266.2023.10248096_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/iwaenc.2018.8521401_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/icce-taiwan62264.2024.10674154_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslpro.2025.3548449_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1541/ieejeiss.142.643_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1587/essfr.16.4_257_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1299/mej.23-00124_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/access.2021.3077441_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslp.2022.3203907_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/msp.2023.3276171_references_DOI_HzFmuHn6eJ9uRa51evYOq0iexKC"},{"@type":"CROSSREF","@value":"10.1541/ieejeiss.141.1077_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1587/transinf.2021edp7020_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/access.2020.3045791_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/access.2020.3045704_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/gcce59613.2023.10315602_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1250/ast.e24.124_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1587/essfr.18.4_267_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1002/tee.22868_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslp.2019.2925450_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1109/taslp.2020.3037487_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"},{"@type":"CROSSREF","@value":"10.1587/transinf.2019edp7259_references_DOI_5VsAbX12Uym7lYNMRZVbvviDFhV"}]}