Socket
Socket
Sign inDemoInstall

@untemps/react-vocal

Package Overview
Dependencies
11
Maintainers
1
Versions
57
Alerts
File Explorer

Advanced tools

Install Socket

Detect and block malicious and high-risk dependencies

Install

    @untemps/react-vocal

React component and hook to initiate a SpeechRecognition session


Version published
Weekly downloads
297
decreased by-2.3%
Maintainers
1
Created
Weekly downloads
 

Readme

Source

react-vocal

A React component and hook to initiate a SpeechRecognition session


npm GitHub Workflow Status Codecov

:red_circle: LIVE DEMO :red_circle:

Disclaimer

The Web Speech API is only supported by few browsers so far (see caniuse). If the API is not available, the Vocal component won't display anything.

This component intends to catch a speech result as soon as possible. This can be a good fit for vocal commands or search field filling. For now on it does not support continuous speech (see Roadmap below).
That means either a result is caught and returned or timeout is reached and the recognition is discarded.
The stop function returned by children-as-function mechanism allows to prematurely discard the recognition before timeout elapses.

Special cases

Some browsers supports the SpeechRecognition API but not all the related APIs.
For example, browsers on iOS 14.5, the SpeechGrammar and SpeechGrammarList and Permissions APIs are not supported.

Although the lack of SpeechGrammar and SpeechGrammarList is handled by the underlaying @untemps/vocal library, you need to deal with Permissions by yourself.

Installation

yarn add @untemps/react-vocal

Usage

Vocal component

Basic usage
import Vocal from '@untemps/react-vocal'

const App = () => {
	const [result, setResult] = useState('')

	const _onVocalStart = () => {
		setResult('')
	}

	const _onVocalResult = (result) => {
		setResult(result)
	}

	return (
		<div className="App">
			<span style={{ position: 'relative' }}>
				<Vocal
					onStart={_onVocalStart}
					onResult={_onVocalResult}
					style={{ width: 16, position: 'absolute', right: 10, top: -2 }}
				/>
				<input defaultValue={result} style={{ width: 300, height: 40 }} />
			</span>
		</div>
	)
}

Custom component

By default, Vocal displays an icon with two states:

  • Idle
    Idle state
  • Listening
    Listening state

But you can provide your own component.

  • With a simple React element:
import Vocal from '@untemps/react-vocal'

const App = () => {
	return (
		<Vocal>
			<button>Start</button>
		</Vocal>
	)
}

In this case, a onClick handler is automatically attached to the component to start a recognition session.
Only the first direct descendant of Vocal will receive the onClick handler. If you want to use a more complex hierarchy, use the function syntax below.

  • With a function that returns a React element:
import Vocal from '@untemps/react-vocal'

const Play = () => (
	<div
		style={{
			width: 0,
			height: 0,
			marginLeft: 1,
			borderStyle: 'solid',
			borderWidth: '4px 0 4px 8px',
			borderColor: 'transparent transparent transparent black',
		}}
	/>
)

const Stop = () => (
	<div
		style={{
			width: 8,
			height: 8,
			backgroundColor: 'black',
		}}
	/>
)

const App = () => {
	return (
		<Vocal>
			{(start, stop, isStarted) => (
				<button style={{ padding: 5 }} onClick={isStarted ? stop : start}>
					{isStarted ? <Stop /> : <Play />}
				</button>
			)}
		</Vocal>
	)
}

The following parameters are passed to the function:

ArgumentsTypeDescription
startfuncThe function used to start the recognition
stopfuncThe function used to stop the recognition
isStartedboolA flag that indicates whether the recognition is started or not

Commands

The Vocal component accepts a commands prop to map special recognition results to callbacks.
That means you can define vocal commands to trigger specific functions.

const App = () => {
  return (
    <Vocal commands={{
      'switch border color': () => setBorderColor('red'),
    }}/>
  )
}

commands object is a key/pair model where the key is the command to be caught by the recognition and the value is the callback triggered when the command is detected.

key is not case sensitive.

const commands = {
    submit: () => submitForm(),
    'Change the background color': () => setBackgroundColor('red'), 
    'PLAY MUSIC': play
}

The component utilizes a special hook called useCommands to respond to the commands.
The hook performs a fuzzy search to match approximate commands if needed. This allows to fix accidental typos or approximate recognition results.
To do so the hook uses fuse.js which implements an algorithm to find strings that are approximately equal to a given input. The score precision that distinguishes acceptable command-to-callback mapping from negative matching can be customized in the hook instantiantion.

useCommands(commands, threshold) // threshold is the limit not to exceed to be considered a match

See fuze.js scoring theory for more details.

:warning: The Vocal component doesn't expose that score yet. For now on you have to deal with the default value (0.4)


Vocal component API
PropsTypeDefaultDescription
commandsobjectnullCallbacks to be triggered when specified commands are detected by the recognition
langstring'en-US'Language understood by the recognition BCP 47 language tag
grammarsSpeechGrammarListnullGrammars understood by the recognition JSpeech Grammar Format
timeoutnumber3000Time in ms to wait before discarding the recognition
styleobjectnullStyles of the root element if className is not specified
classNamestringnullClass of the root element
onStartfuncnullHandler called when the recognition starts
onEndfuncnullHandler called when the recognition ends
onSpeechStartfuncnullHandler called when the speech starts
onSpeechEndfuncnullHandler called when the speech ends
onResultfuncnullHandler called when a result is recognized
onErrorfuncnullHandler called when an error occurs
onNoMatchfuncnullHandler called when no result can be recognized

useVocal hook

Basic usage
import React, { useState } from 'react'
import { useVocal } from '@untemps/react-vocal'
import Icon from './Icon'

const App = () => {
	const [isListening, setIsListening] = useState(false)
	const [result, setResult] = useState('')

	const [, { start, subscribe }] = useVocal('fr_FR')

	const _onButtonClick = () => {
		setIsListening(true)

		subscribe('speechstart', _onVocalStart)
		subscribe('result', _onVocalResult)
		subscribe('error', _onVocalError)
		start()
	}

	const _onVocalStart = () => {
		setResult('')
	}

	const _onVocalResult = (result) => {
		setIsListening(false)

		setResult(result)
	}

	const _onVocalError = (e) => {
		console.error(e)
	}

	return (
		<div>
			<span style={{ position: 'relative' }}>
				<div
					role="button"
					aria-label="Vocal"
					tabIndex={0}
					style={{ width: 16, position: 'absolute', right: 10, top: 2 }}
					onClick={_onButtonClick}
				>
					<Icon color={isListening ? 'red' : 'blue'} />
				</div>
				<input defaultValue={result} style={{ width: 300, height: 40 }} />
			</span>
		</div>
	)
}

Signature
useVocal(lang, grammars)
ArgsTypeDefaultDescription
langstring'en-US'Language understood by the recognition BCP 47 language tag
grammarsSpeechGrammarListnullGrammars understood by the recognition JSpeech Grammar Format

Return value
const [ref, { start, stop, abort, subscribe, unsubscribe, clean }]
ArgsTypeDescription
refRefReact ref to the SpeechRecognitionWrapper instance
startfuncFunction to start the recognition
stopfuncFunction to stop the recognition
abortfuncFunction to abort the recognition
subscribefuncFunction to subscribe to recognition events
unsubscribefuncFunction to unsubscribe to recognition events
cleanfuncFunction to clean subscription to recognition events

Browser support flag

Basic usage
import Vocal, { isSupported } from '@untemps/react-vocal'

const App = () => {
	return isSupported ? <Vocal /> : <p>Your browser does not support Web Speech API</p>
}

Events

EventsDescription
audioendFired when the user agent has finished capturing audio for recognition
audiostartFired when the user agent has started to capture audio for recognition
endFired when the recognition service has disconnected
errorFired when a recognition error occurs
nomatchFired when the recognition service returns a final result with no significant recognition
resultFired when the recognition service returns a result
soundendFired when any sound — recognisable or not — has stopped being detected
soundstartFired when any sound — recognisable or not — has been detected
speechendFired when speech recognized by the recognition service has stopped being detected
speechstartFired when sound recognized by the recognition service as speech has been detected
startfired when the recognition service has begun listening to incoming audio

Notes

The process to grant microphone access permissions is automatically managed by the hook (internally used by the Vocal component).

Development

The component can be served for development purpose on http://localhost:10001/ using:

yarn dev

Contributing

Contributions are warmly welcomed:

  • Fork the repository
  • Create a feature branch (preferred name convention: [feature type]_[imperative verb]-[description of the feature])
  • Develop the feature AND write the tests (or write the tests AND develop the feature)
  • Commit your changes using Angular Git Commit Guidelines
  • Submit a Pull Request

Roadmap

  • Add a connector management to plug external speech-to-text services in
  • Support continuous speech

Keywords

FAQs

Last updated on 11 Apr 2024

Did you know?

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc