Voice-verification is a procedure of verifying the identity of a person claiming to be of certain identity by comparing person's voice-sample (utterance) to the enroled voice-print of that identity.
During this procedure, you will send both the enroled voice-print (obtained during the procedure of
voice-enrolment), and current voice-sample/utterance of a person attempting to identify,
together with some options.
As a result, you will receive a response with the similarity score (how much the two faceprints are similar), and content of the
spoken phrase (if you opted for it, by setting this option in the initial call).
MachineSense customer/partner will initially send a voice-utterance BLOB from their
website or mobile app to their own servers (this step is completely independent of MachineSense), and than call the MachineSense
API, including that BLOB and some parameters.
In your call to MachineSense API, you will include both enroled vector and the freshly recorded utterance (end-user) attempting to identify/verify
a person.
In order to help customers start quickly with such client-side (web-based) implementation, MachineSense offers a set of
examples and code, ready for copy/paste into your applications and be customized/modified. Basic operations with capturing the
image, setting up parameters, etc. will be already present in those examples.
Examples are written in vanilla JavaScript, and can be used in any web-based application.
You can find them on our Demo page as well as our
GitHub repository.
Customer creates own client-side page or app, including capturing of the user's selfie image.
Exception to this might be if customer is using MachineSense whitelabel client-side, in which case this is already done for them, or MachineSense
WASM component. The latter process, however is a two-step process and is related to pre-built / ready-to-use modules.
More details about single-step and two-step processes.
(Call from your server to our API.)
POST /voice/v1/enroll_voice
Parameters / body:
{
"audio": "string",
"xvector": [
0
],
"api_key": "string",
"ref": "string",
"method": "string",
"phrase": "string",
"content": {
"include": false,
"language": "string",
"precision": "string"
}
}
Parameters explained:
Code: 200
Default response:
{
"result": "Ok",
"code": 0,
"message": "string",
"data": {
"ref": "string",
"vector": [
0
],
"speech": "string",
"distance": 0.01,
"confidence": 99.99
}
}
Response explained: