[et_pb_section fb_built=”1″ _builder_version=”4.5.1″ _module_preset=”default” background_color=”#000000″ background_image=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/bg-2-orange.jpg” parallax=”on” parallax_method=”off” custom_padding=”120px||43px||false|false” animation_style=”slide” animation_direction=”top” background_last_edited=”on|phone” background_color_gradient_end_tablet=”rgba(0,42,71,0.44)” background_color_gradient_end_phone=”rgba(0,42,71,0.44)” background_color_gradient_overlays_image_tablet=”off” background_color_gradient_overlays_image_phone=”off” locked=”off” global_colors_info=”{}”][et_pb_row use_custom_gutter=”on” gutter_width=”2″ make_equal=”on” module_class=”service-ti-sub” _builder_version=”4.5.1″ _module_preset=”default” width=”90%” width_tablet=”90%” width_phone=”90%” width_last_edited=”on|tablet” max_width=”1210px” max_width_tablet=”90%” max_width_phone=”90%” max_width_last_edited=”on|desktop” custom_padding=”0px||0px||true|” global_colors_info=”{}”][et_pb_column type=”4_4″ _builder_version=”4.7.7″ _module_preset=”default” global_colors_info=”{}”][et_pb_text _builder_version=”4.9.4″ _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#FFFFFF” header_font_size=”60px” header_2_font=”Alata|700|||||||” header_2_text_align=”center” header_2_text_color=”#ffffff” header_2_font_size=”40px” header_2_line_height=”1.5em” header_3_font=”Alata|600|||||||” header_3_text_align=”center” header_3_text_color=”#000000″ header_3_font_size=”54px” header_3_letter_spacing=”-1px” custom_margin=”30px||||false|false” header_font_size_tablet=”” header_font_size_phone=”30px” header_font_size_last_edited=”on|phone” header_3_font_size_tablet=”” header_3_font_size_phone=”35px” header_3_font_size_last_edited=”on|phone” global_colors_info=”{}”]

Sigma’s Audio Annotation Services

[/et_pb_text][et_pb_text _builder_version=”4.9.4″ text_font=”Montserrat||||||||” text_text_color=”#FFFFFF” text_font_size=”18px” text_line_height=”1.1em” link_font=”||||on||||” link_text_color=”#FFFFFF” text_orientation=”center” text_font_size_tablet=”” text_font_size_phone=”20px” text_font_size_last_edited=”on|phone” global_colors_info=”{}”]

Improve your Speech Recognition Models, Text-to-Speech Engines, NLP and IVR systems and Voice Assistants.

We support over 120 languages and dialects and guarantee an accuracy of 98% but can go higher if needed.

Our ML-assisted tools can reduce annotation time and cost significantly.

[/et_pb_text][/et_pb_column][/et_pb_row][/et_pb_section][et_pb_section fb_built=”1″ _builder_version=”4.5.1″ _module_preset=”default” custom_padding=”30px|||||” locked=”off” global_colors_info=”{}”][et_pb_row _builder_version=”4.9.4″ _module_preset=”default” custom_margin=”||30px||false|false” animation_style=”slide” global_colors_info=”{}”][et_pb_column type=”4_4″ _builder_version=”4.9.4″ _module_preset=”default” global_colors_info=”{}”][et_pb_text _builder_version=”4.9.4″ _module_preset=”default” text_text_color=”#000000″ header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”50px” header_2_font=”Alata|700|||||||” header_2_text_align=”center” header_2_text_color=”#ffffff” header_2_font_size=”40px” header_2_line_height=”1.5em” header_3_font=”Alata|600|||||||” header_3_text_align=”center” header_3_text_color=”#000000″ header_3_font_size=”54px” header_3_letter_spacing=”-1px” custom_margin=”30px||||false|false” header_font_size_tablet=”” header_font_size_phone=”30px” header_font_size_last_edited=”on|phone” header_3_font_size_tablet=”” header_3_font_size_phone=”35px” header_3_font_size_last_edited=”on|phone” global_colors_info=”{}”]

What is Audio Annotation?

[/et_pb_text][/et_pb_column][/et_pb_row][et_pb_row column_structure=”2_5,3_5″ use_custom_gutter=”on” gutter_width=”1″ make_equal=”on” module_class=”ser-img-text” _builder_version=”4.9.4″ _module_preset=”default” width=”90%” max_width=”1170px” max_width_tablet=”90%” max_width_phone=”” max_width_last_edited=”on|phone” custom_padding=”0px||0px||true|false” locked=”off” global_colors_info=”{}” width__hover_enabled=”on|desktop” width__hover=”90%”][et_pb_column type=”2_5″ module_class=”ds-vertical-align” _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ee7224″ background_enable_image=”off” background_position=”center_left” animation_style=”slide” animation_direction=”left” global_colors_info=”{}”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/Audio_signal_17659897_m.jpg” title_text=”17659897 – oscilloscope screen showing wave signal” align=”center” _builder_version=”4.9.4″ _module_preset=”default” global_colors_info=”{}”][/et_pb_image][et_pb_button button_url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjIyNjAzIn19@” button_text=”Let’s Talk” button_alignment=”center” _builder_version=”4.9.4″ _dynamic_attributes=”button_url” _module_preset=”default” custom_button=”on” button_text_size=”18px” button_text_color=”#ffffff” button_bg_color=”#000000″ button_border_width=”0px” button_border_color=”#ed7225″ button_border_radius=”0px” button_letter_spacing=”1px” button_font=”Montserrat|600|||||||” button_use_icon=”off” custom_margin=”35px||||false|false” custom_padding=”16px|50px|13px|50px|false|true” animation_style=”slide” locked=”off” global_colors_info=”{}” button_text_color__hover_enabled=”on|hover” button_text_color__hover=”#000000″ button_bg_color__hover_enabled=”on|hover” button_bg_color__hover=”#FFFFFF” button_bg_enable_color__hover=”on” button_border_color__hover_enabled=”on|hover” button_border_color__hover=”#1e2228″ button_letter_spacing__hover_enabled=”on|desktop” button_letter_spacing__hover=”1px”][/et_pb_button][/et_pb_column][et_pb_column type=”3_5″ module_class=”ds-vertical-align” _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ee7224″ background_enable_image=”off” custom_padding=”60px|60px|60px|60px|true|true” custom_padding_tablet=”” custom_padding_phone=”30px|30px|30px|30px|true|true” custom_padding_last_edited=”on|phone” animation_style=”slide” animation_direction=”right” global_colors_info=”{}”][et_pb_text _builder_version=”4.9.4″ text_font=”Montserrat||||||||” text_text_color=”#ffffff” text_font_size=”16px” text_line_height=”1.5em” text_orientation=”center” text_font_size_tablet=”” text_font_size_phone=”18px” text_font_size_last_edited=”on|phone” global_colors_info=”{}”]

It is important to differentiate between transcription and annotation.

Audio or Speech Transcription is the process of converting spoken language into written form. Transcription can be verbatim or non-verbatim. Verbatim transcription includes all conversation filler words, false starts, truncated words or sentences, and pauses. It is a literal conversion of the conversation into written language. Non-verbatim transcription does not include filler words, false starts, truncated words or sentences, and pauses. So, it provides an easy-to-read version of what it has been said in the audio file.

An annotation is any type of additional information that is added to an already existing text, be it a transcription of an audio file or an original text file.

Normally, Audio or Speech Annotation refers to both, the transcription of the audio and the annotation of the resulting text. Annotations add phonological, morphological, syntactic, semantic and discourse information.

It is also usual that Audio or Speech Annotation includes metadata, which is relevant information that refers to the audio file as a whole, rather than individual annotations that provide information about a portion of the data.

[/et_pb_text][/et_pb_column][/et_pb_row][/et_pb_section][et_pb_section fb_built=”1″ module_id=”healthcare” _builder_version=”4.5.1″ _module_preset=”default” background_color=”#f6f7f9″ custom_padding=”50px||100px|||” locked=”off” global_colors_info=”{}”][et_pb_row _builder_version=”4.9.4″ _module_preset=”default” custom_margin=”||30px||false|false” custom_margin_tablet=”” custom_margin_phone=”0px||0px||false|false” custom_margin_last_edited=”on|phone” custom_padding=”0px||||false|false” animation_style=”slide” global_colors_info=”{}”][et_pb_column type=”4_4″ _builder_version=”4.9.4″ _module_preset=”default” global_colors_info=”{}”][et_pb_text _builder_version=”4.9.4″ _module_preset=”default” text_text_color=”#000000″ header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”45px” header_2_font=”Alata|700|||||||” header_2_text_align=”center” header_2_text_color=”#ffffff” header_2_font_size=”40px” header_2_line_height=”1.5em” header_3_font=”Alata|600|||||||” header_3_text_align=”center” header_3_text_color=”#000000″ header_3_font_size=”54px” header_3_letter_spacing=”-1px” custom_margin=”30px||||false|false” header_font_size_tablet=”” header_font_size_phone=”30px” header_font_size_last_edited=”on|phone” header_3_font_size_tablet=”” header_3_font_size_phone=”35px” header_3_font_size_last_edited=”on|phone” global_colors_info=”{}”]

Audio Annotation Services

[/et_pb_text][et_pb_text _builder_version=”4.14.8″ _module_preset=”default” text_text_color=”#000000″ header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”45px” header_2_font=”Alata|700|||||||” header_2_text_align=”center” header_2_text_color=”#ffffff” header_2_font_size=”40px” header_2_line_height=”1.5em” header_3_font=”Alata|600|||||||” header_3_text_align=”center” header_3_text_color=”#000000″ header_3_font_size=”54px” header_3_letter_spacing=”-1px” custom_margin=”30px||||false|false” hover_enabled=”0″ header_font_size_tablet=”” header_font_size_phone=”30px” header_font_size_last_edited=”on|phone” header_3_font_size_tablet=”” header_3_font_size_phone=”35px” header_3_font_size_last_edited=”on|phone” global_colors_info=”{}” sticky_enabled=”0″]

[zb_ziggy_bot]

[/et_pb_text][/et_pb_column][/et_pb_row][et_pb_row column_structure=”1_4,1_4,1_4,1_4″ use_custom_gutter=”on” gutter_width=”2″ make_equal=”on” custom_padding_last_edited=”on|phone” module_class=”gq_r_fijar_botones” _builder_version=”4.9.4″ _module_preset=”default” width=”95%” width_tablet=”90%” width_phone=”90%” width_last_edited=”on|tablet” max_width=”95%” max_width_tablet=”90%” max_width_phone=”90%” max_width_last_edited=”on|desktop” module_alignment=”center” custom_padding=”50px||0px|||” custom_padding_tablet=”” custom_padding_phone=”0px||0px||false|false” global_colors_info=”{}”][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/Speech-annotation.jpg” alt=”Edit audio” title_text=”Speech-annotation” align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Audio Annotation” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” alt=”data” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

It includes audio transcription, annotation and metadata. The type of annotations and metadata are fully tailored to client’s needs. From phonological, morphological, syntactic, semantic and discourse information to audio segmentation, speaker identification, turn taking, emotion, background noise, speech or music. You name it!

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/5-3-Audio-processing1.jpg” alt=”Audio processing” title_text=”5-3-Audio-processing1″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Audio and Video Transcription” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” alt=”Audio processing” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″ title__hover_enabled=”off|desktop”]Be it verbatim or non-verbatim, our team will provide you with best-in-class transcriptions the way you want it, when you need it, in a cost-efficient way.

Sigma offers scalable audio and video transcription services thanks to the optimal combination of our large base of vetted transcribers and our ML-assisted tools.[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/5-3-Audio-processing2.jpg” alt=”Audio processing” title_text=”5-3-Audio-processing2″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Speaker Diarization” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

It consists of partitioning the input audio file into homogeneous audio segments according to their specific sources. These sources include the identity of the speakers whose voice is recorded in the audio file, music, silence, or background noise. This enables automating the process of analyzing any type of conversation, including call center dialogues and debates.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/5-3-Audio-processing4-1.jpg” alt=”Wave audio” title_text=”5-3-Audio-processing4-1″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Phonetic Transcription” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

A phonetic transcription is very similar to a regular transcription, but instead of converting the audio into a sequence of words, it describes the way spoken words are pronounced using phonetic symbols.

The most common alphabetic system of phonetic notation is the International Phonetic Alphabet (IPA).

[/et_pb_blurb][/et_pb_column][/et_pb_row][et_pb_row column_structure=”1_4,1_4,1_4,1_4″ use_custom_gutter=”on” gutter_width=”2″ make_equal=”on” module_class=”gq_r_fijar_botones” _builder_version=”4.9.4″ _module_preset=”default” width=”95%” width_tablet=”90%” width_phone=”90%” width_last_edited=”on|tablet” max_width=”95%” max_width_tablet=”90%” max_width_phone=”90%” max_width_last_edited=”on|desktop” module_alignment=”center” custom_padding=”70px||0px|||” global_colors_info=”{}”][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/3-2-Contact7.jpg” alt=”Edit audio” title_text=”3-2-Contact7″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Emotion Annotation” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” alt=”data” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

Emotion annotation aims to determine feelings such as anger, happiness, sadness, fear, or surprise. It can be performed on text or audio data. Audio emotion analysis is more accurate since audio provides additional clues such as speech rate, pitch, pitch jumps, or voice intensity.

Emotion detection helps improve human-machine communication, analyze call center dialogues, etc.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/3-2-Contact3.jpg” alt=”Edit audio” title_text=”3-2-Contact3″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Sentiment Annotation” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″ title__hover_enabled=”off|desktop”]

It is the process of determining if a segment of speech is perceived as positive, negative or neutral. Audio sentiment analysis is more accurate than text sentiment analysis since audio provides additional information such as the emotional state of the speakers.

It helps gauge customers opinion, monitor brand/product reputation, customer experience and needs, social media, etc.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/audio-classification.jpg” alt=”Classification” title_text=”audio-classification” align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Audio Classification” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” alt=”Classification” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

It consists of listening to the audio recording and classifying it into a series of predetermined categories.

For example, categories that describe the user intent, the background noise, the quality of the recording, the topic, the number or type of speakers, the spoken language or dialect or semantic related information.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/data-relevance.jpg” alt=”Data” title_text=”data-relevance” align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Data Relevance” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” alt=”Data ” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

Data relevance provides information about the quality of data that a system delivers to its users. In particular, it determines to which extend the answer of a search engine or an intelligent assistant provides insight into the question of the user; i.e.: the level of consistency between the content of the data provided and the area of interest of the user.

[/et_pb_blurb][/et_pb_column][/et_pb_row][et_pb_row column_structure=”1_4,1_4,1_4,1_4″ use_custom_gutter=”on” gutter_width=”2″ make_equal=”on” module_class=”gq_r_fijar_botones” _builder_version=”4.9.4″ _module_preset=”default” width=”95%” width_tablet=”90%” width_phone=”90%” width_last_edited=”on|tablet” max_width=”95%” max_width_tablet=”90%” max_width_phone=”90%” max_width_last_edited=”on|desktop” module_alignment=”center” custom_padding=”70px||0px|||” global_colors_info=”{}”][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/reasons-01.jpg” alt=”Edit audio” title_text=”reasons-01″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Speech Annotation Quality Assessment” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” alt=”data” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

It aims to determine the accuracy of the speech annotations, including word error rate (substitutions, insertions and deletions), and label error rate according to the annotation guidelines.

It helps assess the quality of the annotated speech in terms of accuracy and interpretation consistency of the annotation guidelines. It also helps complete the guidelines and resolve its ambiguities.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/reasons-02.jpg” alt=”Edit audio” title_text=”reasons-02″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Speech Database Quality Assessment” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″ title__hover_enabled=”off|desktop”]

This quality assessment service provides information that helps optimize the effort in data collection and annotation. Quality is a multidimensional parameter that depends on factors such as the volume and quality of the audio, the accuracy of the annotations, the data consistency, the domain and customer’s coverage; and the balance.

This helps focus the data collection and annotation effort where is most needed.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/3-2-Contact2.jpg” alt=”Edit audio” title_text=”3-2-Contact2″ align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Intelligent Assistants Assessment” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]

This service measures the performance of the wake word detection, assesses if the pronunciation of the wake word and the subsequent voice commands belong to the same or several users, checks if the voice interactions are in the expected language, or if the answers of the assistant are correct based on the dialogue status and context as well as on the user data and the system knowledge data base, etc.

[/et_pb_blurb][/et_pb_column][et_pb_column type=”1_4″ _builder_version=”4.9.4″ _module_preset=”default” background_color=”#ffffff” custom_padding=”0px|0px|20px|0px|false|true” animation_style=”slide” animation_direction=”bottom” link_option_url=”#” border_radii=”on|NaNpx|NaNpx|NaNpx|NaNpx” box_shadow_style=”preset1″ box_shadow_vertical=”0px” box_shadow_blur=”48px” box_shadow_color=”rgba(0,0,0,0)” global_colors_info=”{}” box_shadow_color__hover_enabled=”on|desktop” box_shadow_color__hover=”rgba(162,162,162,0.5)”][et_pb_image src=”https://sigmaai1.wpengine.com/wp-content/uploads/2021/04/audio-speak.jpg” alt=”Edit audio” title_text=”audio-speak” align=”center” _builder_version=”4.9.4″ _module_preset=”default” locked=”off” global_colors_info=”{}”][/et_pb_image][et_pb_blurb title=”Pronunciation Assessment” url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjUwNiJ9fQ==@” content_max_width=”90%” module_class=”box-card” _builder_version=”4.13.0″ _dynamic_attributes=”url” _module_preset=”default” header_font=”Montserrat|700|||||||” header_text_align=”center” header_text_color=”#000000″ header_font_size=”18px” header_line_height=”1.3em” body_font=”Montserrat||||||||” body_text_align=”left” body_text_color=”#00223f” body_font_size=”16px” body_line_height=”1.5em” module_alignment=”center” custom_padding=”||35px||false|false” animation=”off” border_width_right_image=”8px” border_color_right_image=”rgba(0,0,0,0)” global_colors_info=”{}” header_text_color__hover_enabled=”on|hover” header_text_color__hover=”#ee7224″]The pronunciation assessment aims to determine whether the pronunciation of a word or sentence is correct. The correctness of the pronunciation can be performed comparing it with the standard pronunciation or with the dialect variants.

The pronunciation assessment can be performed on human or synthetic speech.[/et_pb_blurb][/et_pb_column][/et_pb_row][/et_pb_section][et_pb_section fb_built=”1″ _builder_version=”4.9.4″ _module_preset=”default” custom_padding=”0px||0px|||” global_colors_info=”{}”][et_pb_row _builder_version=”4.9.4″ _module_preset=”default” custom_padding=”40px|||||” global_colors_info=”{}”][et_pb_column type=”4_4″ _builder_version=”4.9.4″ _module_preset=”default” global_colors_info=”{}”][et_pb_button button_url=”@ET-DC@eyJkeW5hbWljIjp0cnVlLCJjb250ZW50IjoicG9zdF9saW5rX3VybF9wYWdlIiwic2V0dGluZ3MiOnsicG9zdF9pZCI6IjIyNjAzIn19@” button_text=”Let’s Talk” button_alignment=”center” _builder_version=”4.9.4″ _dynamic_attributes=”button_url” _module_preset=”default” custom_button=”on” button_text_size=”18px” button_text_color=”#ffffff” button_bg_color=”#f27507″ button_border_width=”0px” button_border_color=”#ed7225″ button_border_radius=”0px” button_letter_spacing=”1px” button_font=”Montserrat|600|||||||” button_use_icon=”off” custom_margin=”35px||||false|false” custom_padding=”16px|50px|13px|50px|false|true” animation_style=”slide” locked=”off” global_colors_info=”{}” button_text_color__hover_enabled=”on|hover” button_text_color__hover=”#ffffff” button_bg_color__hover_enabled=”on|desktop” button_bg_color__hover=”#000000″ button_bg_enable_color__hover=”on” button_border_color__hover_enabled=”on|hover” button_border_color__hover=”#1e2228″ button_letter_spacing__hover_enabled=”on|desktop” button_letter_spacing__hover=”1px”][/et_pb_button][/et_pb_column][/et_pb_row][/et_pb_section]

EN