Inventing a more human conversation with technology

Nuance’s R&D team of talented scientists, linguists and engineers pioneers innovation in human-machine intelligence and communication, resulting in groundbreaking strides in voice, language understanding and AI that redefine how people and technology co-exist.

Research areas

research category image

Voice biometrics enables authentication through processing natural speech patterns

Learn more

research category image
Speech enhancement

Speech enhancement is focused on evolving the technologies used in voice-driven interfaces for an enhanced user experience and greater application flexibility for manufacturers

Learn more

research category image
Speech recognition

Turning speech into text is at the heart of an amazing variety of products and services that enrich peoples’ lives

Learn more

research category image
Natural language & AI

Natural Language Processing and Artificial Intelligence technologies enable users to communicate with the digital technology they use everyday, whether for work or recreation

Learn more

research category image
Text to speech

Text-to-Speech (TTS) is a corner-stone Nuance technology, with a long research history combining ideas integrated from several teams and new hires

Learn more

Recent articles

Textual Inference: getting logic from humans

This paper describes a manual investigation of the SICK corpus, which is the proposed testing set for a new system for natural language inference. The

Read more

CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies

The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which participants train and test their learning systems on the same data

Read more

Universal Dependencies for Portuguese

This paper describes the process of converting the Portuguese Bosque corpus to the Universal Dependencies scheme version 2. The conversion was done by applying to

Read more

Low-Complexity Pitch Estimation Based on Phase Differences Between Low-Resolution Spectra

Detection of voiced speech and estimation of the pitch frequency are important tasks for many speech processing algorithms. Pitch information can be used, e.g., to

Read more

The battle of the giants: a case study of GPU vs FPGA optimisation for real-time image processing

This paper focuses on a thorough comparison of the two main hardware targets for real-time optimization of a computer vision algorithm: GPU and FPGA. Based

Read more

Adieu Features? End-to-end speech emotion recognition using a deep convolutional recurrent network

The automatic recognition of spontaneous emotions from speech is a challenging task. On the one hand, acoustic features need to be robust enough to capture

Read more

Breaking news

Charlie Ortiz featured on CNET’s list of Most Influential Latinos in Tech

Charlie Ortiz, head of Nuance’s Sunnyvale based AI lab, is #15 on CNET’s list of most influentual latinos in the tech industry. He was also featured there last year.

See more news

Upcoming event

- April 20

Nuance is a sponsor of ICASSP 2018 in Calgary.

See more events

Featured researcher

Peter Patel-Schneider
Peter Patel-Schneider

Peter Patel-Schneider is a Researcher in the Natural Language Understanding Lab. His research interest is in the area of Natural Language & AI. Currently, Peter i...

See full bio

Seen on What’s next

Multimodal interaction – How machines learn to understand pointing

As we learn more about the biological world around us, the list of things only humans can do has dwindled – and that’s before computers started to play chess and Go. Counting? Birds can deal with numbers up to twelve. Using tools? Dolphins in Shark Bay, Australia, are using sponges as a tool for hunting. […]

The post Multimodal interaction – How machines learn to understand pointing appeared first on What’s next.