Sony would like to extend our heartfelt condolences to those who have died from the COVID-19 virus and their families and pray for the speedy recovery of those currently battling the disease.
As many of you already know, the International Conference on Acoustics, Speech and Signal Processing (ICASSP 2020) is to be held as a fully virtual conference for the first time in its long history.
Sony salutes the IEEE Signal Processing Society for prioritizing the need to protect people's lives with its decision to convert the conference to a virtual environment, and we thank the staff at the organizing committee for their swift preparations.
As one of the sponsors of the event, Sony intended to hold a workshop and host a technology exhibition in the sponsors booth. It was very unfortunate that we couldn't meet you all in Barcelona, but, as an alternative, we intend to introduce several of Sony's latest combined AI and audio technologies on this site, including some still in the development stage. While it may be small, we are keen to contribute to the first virtual conference in any way that we can.
We ICASSP participants readily join the global battle to defeat COVID-19 and, once this global health emergency has passed, we look forward to seeing you all in person next year in Toronto.

ICASSP (International Conference on Acoustics, Speech, and Signal Processing) is the world's largest and most
comprehensive technical conference focused on signal processing and its applications.

AI x Audio:
From Research to Production
Over its 70-year history, Sony has always sought to provide consumers with extraordinary, excellent-quality audio products right from the era of tape recorders and transistor radios.
We are also focused on using our speech processing, image processing, communication, mechatronics & motion control, semiconductor, sensing, and AI technologies to research and develop products, services and entertainment that enable customers to feel Sony's exciting "KANDO" power of emotional connection.
We were scheduled to hold a Sponsor Workshop at ICASSP 2020 to introduce these technologies, so we have reorganized the content into a "virtual workshop" on these site pages.
Here, we focus specifically on introducing technologies using deep learning :1) Open source sound source separation and 2) Music generation
We hope you can take the time to enjoy them.
workshop01
Audio Source Separation and NNabla
In a first part, we will introduce Sony's audio source separation technology and show that many different use cases can be solved with the same tool, which makes it very powerful. In a second part, we will discuss Open-Unmix - an open-source implementation of audio separation - and reveal upcoming features of it. Finally, we will talk about NNabla, Sony's deep learning framework and highlight features which makes NNabla attractive not only for research and development but also for production use cases.
-
Stefan Uhlich
Sony R&D Center Stuttgart Lab.1
Stefan Uhlich received the Dipl.-Ing. and PhD degree in electrical engineering from the University of Stuttgart, Germany, in 2006 and 2012, respectively. From 2007 to 2011 he was a research assistant at the Chair of System Theory and Signal Processing, University of Stuttgart. In this time he worked in the area of statistical signal processing, focusing especially on parameter estimation theory and methods. Since 2011, he is with the Sony Stuttgart Technology Center where he works as a Principal Engineer on problems in music source separation, speech enhancement and deep neural network compactization.
workshop02
The Sound of AI - How Machine Learning Impacts Music Production
New technologies always have the potential to transform the way music is created. AI is a powerful new paradigm and, as such, establishes its place also in the creative workflows of music production. In this talk, we present current solutions of AI-assisted music production and discuss the potentials and challenges, as well as unanswered questions in this paradigm. We also present our research activities at Sony CSL, including some brief details about the underlying machine learning models. For creating relevant tools, we rely on close collaboration with music producers, which can be a challenge on its own. We talk about the gap between the scientific approach and that of artists and the music industry, and we address our efforts to bridge that gap.
-
Stefan Lattner
Sony CSL Paris
Stefan Lattner is a research associate at Sony CSL Paris, where he works on transformation and invariance learning with artificial neural networks. Using this paradigm, he targets rhythm generation (i.e., DrumNet) and is also involved in music information retrieval, audio generation, and recommendation. He obtained his doctorate in the area of music structure modeling from the Johannes Kepler University in Linz, Austria.
-
Cyran Aouameur
Sony CSL Paris
Cyran Aouameur is an assistant researcher at Sony CSL. Graduated from Ircam-organized ATIAM Master's degree, he entered CSL two years ago. Passionate about urban music since he was a child, he has been focusing on developing AI-based solutions for artists to quickly design unique drum sounds and rhythms, which he considers being top-importance elements. He is now partly responsible for the communication with the artists, seeking to get the research and the music industry worlds to understand each other.
Sony's Audio Products &
Service utilizing AI technology
1. 360 Reality Audio

1. 360 Reality Audio
360 Reality Audio is a new music experience that uses Sony's object-based spatial audio technology.
Individual sounds such as vocals, chorus, piano, guitar, bass and even sounds of the live audience can be placed in a 360 spherical sound field, giving artists and creators a new way to express their creativity. Listeners can be immersed in a field of sound exactly as intended by artists and creators.


Optimization by personal ear data uses Sony's original estimation algorithm utilizing machine learning.
We analyze that listener's hearing characteristics by estimating the 3D shape of the ear based a photo of their ear through "Sony | Headphones Connect" app.


2. Noise Canceling Headphones

2. Noise Canceling Headphones
Adaptive Sound Control automatically adjusts to whatever you do.
The Sony | Headphones Connect app offers Adaptive Sound Control, a smart function that automatically detects what you're up to - such as traveling, walking, or waiting - then adjusts ambient sound settings to suit the situation. You can also customize the settings to your preferences.
*As of Dec. 1st 2019. According to research by Sony Corporation, measured using JEITA-compliant guidelines in Truly Wireless style noise cancelling headphones market.


Link about "WF-1000XM3"
Publications
-
Tuesday, 5 May, 17:10 - 17:30   
SS-L1: Active Control of Acoustic Noise over Spatial RegionsArray-geometry-aware spatial active noise control based on direction-of-arrival weighting
Yu Maeno, Yuhta Takida, Naoki Murata, and Yuki Mitsufuji
-
Tuesday, 5 May, 17:10 - 17:30   
AUD-L2: Deep Learning for Source SeparationImproving Voice Separation by Incorporating End-to-end Speech Recognition
Naoya Takahashi, Mayank Singh (IIT-Bonbay), Sakya Basak (IISc), Sriram Ganapathy(IISc), Yuki Mitsufuji
-
Wednesday, 6 May, 11:30 - 13:30   
WE2.PF: Audio ClassificationMetric Learning with Background Noise Class for Few-shot Detection of Rare Sound Events
Kazuki Shimada, Yuichiro Koyama, and Akira Inoue
Recruit Info.
If you are interested in working with us, please click here for more open positions of job and internship.