ImageRepositoryAmarnaCloudSpeechSignals

AI Overview😉

  • The potential purpose of this module is to process and analyze audio or video content, specifically speech recognition, to generate transcripts and identify languages. It appears to be a part of Google's content analysis and indexing system.
  • This module could impact search results by providing more accurate and relevant information about audio and video content, allowing users to search for specific phrases or keywords within the content. It may also help Google to better understand the context and meaning of the content, leading to more accurate search rankings and recommendations.
  • A website may change things to be more favorable for this function by providing high-quality, clear, and well-formatted audio and video content, making it easier for the speech recognition module to accurately transcribe and analyze the content. Additionally, providing metadata such as language information and transcripts can help Google's algorithms to better understand the content and improve search results.

Interesting Module? Vote 👇

Voting helps other researchers find interesting modules.

Current Votes: 0

GoogleApi.ContentWarehouse.V1.Model.ImageRepositoryAmarnaCloudSpeechSignals (google_api_content_warehouse v0.4.0)

Next Tag: 10

Attributes

  • duplicateOfYtS3Asr (type: boolean(), default: nil) - If this field is set to true, it means that Youtube already processed the ASR from S3 for the langID. Please find the ASR result from transcript_asr in google3/image/repository/proto/video_search.proto instead.
  • langWithoutLocale (type: String.t, default: nil) - DEPRECATED: Please switch to langid_input. The language id input for creating this ASR without regional info. Same format as in go/ytlangid. This field is populated in Kronos Amarna Cloud Speech operator and passed to Amarna, but it is cleared before stored in Amarna's metadata table.
  • langidInput (type: GoogleApi.ContentWarehouse.V1.Model.ImageRepositoryLanguageIdentificationResult.t, default: nil) - The language identification input used to generate this ASR. This field is populated in Kronos Amarna Cloud Speech operator and passed to Amarna, but cleared before stored in Amarna's metadata table.
  • modelIdentifier (type: String.t, default: nil) -
  • results (type: list(GoogleApi.ContentWarehouse.V1.Model.ImageRepositorySpeechRecognitionResult.t), default: nil) - Raw results from Cloud Speech API
  • s3RecognizerMetadataResponse (type: GoogleApi.ContentWarehouse.V1.Model.ImageRepositoryS3RecognizerMetadataResponse.t, default: nil) - The metadata about the S3 recognizer used.
  • transcriptAsr (type: GoogleApi.ContentWarehouse.V1.Model.PseudoVideoData.t, default: nil) - This field contains full (stitched) transcription, word-level time offset , and word-level byte offset. The value of this field is derived from the SpeechRecognitionResult field above.

Summary

Types

t()

Functions

decode(value, options)

Unwrap a decoded JSON object into its complex fields.

Types

Link to this type

t()

@type t() ::
  %GoogleApi.ContentWarehouse.V1.Model.ImageRepositoryAmarnaCloudSpeechSignals{
    duplicateOfYtS3Asr: boolean() | nil,
    langWithoutLocale: String.t() | nil,
    langidInput:
      GoogleApi.ContentWarehouse.V1.Model.ImageRepositoryLanguageIdentificationResult.t()
      | nil,
    modelIdentifier: String.t() | nil,
    results:
      [
        GoogleApi.ContentWarehouse.V1.Model.ImageRepositorySpeechRecognitionResult.t()
      ]
      | nil,
    s3RecognizerMetadataResponse:
      GoogleApi.ContentWarehouse.V1.Model.ImageRepositoryS3RecognizerMetadataResponse.t()
      | nil,
    transcriptAsr: GoogleApi.ContentWarehouse.V1.Model.PseudoVideoData.t() | nil
  }

Functions

Link to this function

decode(value, options)

@spec decode(struct(), keyword()) :: struct()

Unwrap a decoded JSON object into its complex fields.