Scenes-with-text Detection (v4.3)
About this version
- Submitter: keighrim
- Submission Time: 2024-04-11T21:49:21+00:00
- Prebuilt Container Image: ghcr.io/clamsproject/app-swt-detection:v4.3
-
Release Notes
This version brings many bug fixes and new models
- fixed missing
NEG
label score in theclassification
property inTimePoint
annotations (#87), and app metadata is updated accordingly - fixed sampling rate disparity (#90)
- fixed sinusoidal positional features were not actually used (#47), and newly trained models with the fix are included
- miscellaneous code clean-up
- fixed missing
About this app (See raw metadata.json)
Detects scenes with text, like slates, chyrons and credits.
- App ID: http://apps.clams.ai/swt-detection/v4.3
- App License: Apache 2.0
- Source Repository: https://github.com/clamsproject/app-swt-detection (source tree of the submitted version)
Inputs
(Note: “*” as a property value means that the property is required but can be any value.)
- http://mmif.clams.ai/vocabulary/VideoDocument/v1 (required) (of any properties)
Configurable Parameters
(Note: Multivalued means the parameter can have one or more values.)
-
startAt
: optional, defaults to0
- Type: integer
- Multivalued: False
Number of milliseconds into the video to start processing
-
stopAt
: optional, defaults to10000000
- Type: integer
- Multivalued: False
Number of milliseconds into the video to stop processing
-
sampleRate
: optional, defaults to1000
- Type: integer
- Multivalued: False
Milliseconds between sampled frames
-
minFrameScore
: optional, defaults to0.01
- Type: number
- Multivalued: False
Minimum score for a still frame to be included in a TimeFrame
-
minTimeframeScore
: optional, defaults to0.5
- Type: number
- Multivalued: False
Minimum score for a TimeFrame
-
minFrameCount
: optional, defaults to2
- Type: integer
- Multivalued: False
Minimum number of sampled frames required for a TimeFrame
-
modelName
: optional, defaults to20240126-180026.convnext_lg.kfold_000
- Type: string
- Multivalued: False
- Choices:
20240126-180026.convnext_lg.kfold_000
,20240212-131937.convnext_tiny.kfold_000
,20240212-132306.convnext_lg.kfold_000
model name to use for classification
-
useStitcher
: optional, defaults totrue
- Type: boolean
- Multivalued: False
- Choices:
false
,true
Use the stitcher after classifying the TimePoints
-
pretty
: optional, defaults tofalse
- Type: boolean
- Multivalued: False
- Choices:
false
,true
The JSON body of the HTTP response will be re-formatted with 2-space indentation
Outputs
(Note: “*” as a property value means that the property is required but can be any value.)
(Note: Not all output annotations are always generated.)
- http://mmif.clams.ai/vocabulary/TimeFrame/v5
- timeUnit = “milliseconds”
- http://mmif.clams.ai/vocabulary/TimePoint/v4
- timeUnit = “milliseconds”
- labelset = a list of [“B”, “S”, “S:H”, “S:C”, “S:D”, “S:B”, “S:G”, “W”, “L”, “O”, “M”, “I”, “N”, “E”, “P”, “Y”, “K”, “G”, “T”, “F”, “C”, “R”]