Scenes-with-text Detection (v4.3)

About this version

  • Submitter: keighrim
  • Submission Time: 2024-04-11T21:49:21+00:00
  • Prebuilt Container Image: ghcr.io/clamsproject/app-swt-detection:v4.3
  • Release Notes

    This version brings many bug fixes and new models

    • fixed missing NEG label score in the classification property in TimePoint annotations (#87), and app metadata is updated accordingly
    • fixed sampling rate disparity (#90)
    • fixed sinusoidal positional features were not actually used (#47), and newly trained models with the fix are included
    • miscellaneous code clean-up

About this app (See raw metadata.json)

Detects scenes with text, like slates, chyrons and credits.

Inputs

(Note: “*” as a property value means that the property is required but can be any value.)

Configurable Parameters

(Note: Multivalued means the parameter can have one or more values.)

  • startAt: optional, defaults to 0

    • Type: integer
    • Multivalued: False

    Number of milliseconds into the video to start processing

  • stopAt: optional, defaults to 10000000

    • Type: integer
    • Multivalued: False

    Number of milliseconds into the video to stop processing

  • sampleRate: optional, defaults to 1000

    • Type: integer
    • Multivalued: False

    Milliseconds between sampled frames

  • minFrameScore: optional, defaults to 0.01

    • Type: number
    • Multivalued: False

    Minimum score for a still frame to be included in a TimeFrame

  • minTimeframeScore: optional, defaults to 0.5

    • Type: number
    • Multivalued: False

    Minimum score for a TimeFrame

  • minFrameCount: optional, defaults to 2

    • Type: integer
    • Multivalued: False

    Minimum number of sampled frames required for a TimeFrame

  • modelName: optional, defaults to 20240126-180026.convnext_lg.kfold_000

    • Type: string
    • Multivalued: False
    • Choices: 20240126-180026.convnext_lg.kfold_000, 20240212-131937.convnext_tiny.kfold_000, 20240212-132306.convnext_lg.kfold_000

    model name to use for classification

  • useStitcher: optional, defaults to true

    • Type: boolean
    • Multivalued: False
    • Choices: false, true

    Use the stitcher after classifying the TimePoints

  • pretty: optional, defaults to false

    • Type: boolean
    • Multivalued: False
    • Choices: false, true

    The JSON body of the HTTP response will be re-formatted with 2-space indentation

Outputs

(Note: “*” as a property value means that the property is required but can be any value.)

(Note: Not all output annotations are always generated.)