autoEdit 2 Documentation
1.0.6
1.0.6
  • Introduction
  • Overview
    • Intro
      • High-level overview of the parts
      • from 1.0.5 to 1.0.6
      • Project folder structure
    • Support the project
  • Transcriptions
    • Transcriptions intro
    • Transcription json
    • Transcriber
      • audio to video
      • STT sdks
        • IBM Watson STT
        • Gentle STT
        • Pocketsphinx
    • Read metadata
    • Video preview conversion
    • Hypertranscript
  • Papercuts
    • Papercuts
      • Selections
      • Annotations
      • Tags
  • Paper-edit
    • Paper-edit
      • Paper-edit json
      • search-filter
      • drag-and-drop
      • video-preview
  • Export
    • Export
      • EDL export
      • XML export
  • Appendix
    • Dev configuration
    • Deployment/build
      • Deployment/build for Max OS X
      • Deployment / Build for Linux
      • Deployment / Build for Windows
    • Dependencies
    • Current db setup
    • EDL Format
    • Reusable components
    • Prerequisites
    • JQuery and NWJS Packaging
    • Roadmap
      • Paper-editing Roadmap
      • Extra Features Roadmap
      • Future Roadmap
        • Live video editing
        • Social Media Export
        • Translate transcriptions
        • Web app
          • Multi-user collaboration
        • Searchable Editable Archive
        • NLP insights
        • Slack/Chat bot integration
        • Interactive dev tool
        • Phone mms integration with twillio
        • B-roll computational photography
    • Paper-editing Roadmap
    • Testing
    • Updating automated documentation
    • History of autoEdit versions over time
    • ffmpeg and ffprobe in electron
  • Appendix - Data structures
    • IBM Watson json specs
    • Gentle Json transcription specs
    • Pocketsphinx results
    • autoEdit transcription Json
  • QA List
    • QA Intro
    • QA Launch App
    • QA Transcriptions
    • QA Paperedits
    • QA Export
  • Methods
    • Example: Defining Methods
  • Adobe Panel
    • autoEdit Adobe CEP Panel dev setup
    • autoEdit Adobe CEP Panel integration overview
    • Adobe CEP Jsx functions for autoEdit adobe Panel
Powered by GitBook
On this page
  1. Appendix - Data structures

autoEdit transcription Json

PreviousPocketsphinx resultsNextQA Intro

Last updated 6 years ago

This corresponds to the .

// http://backbonejs.org/#Model
module.exports = Backbone.Model.extend({
  idAttribute: '_id',
  urlRoot: path.join(config.serverUrl, 'transcription'),
  defaults: {
    // title: 'Default Title ',
    // description: 'Default Description',
    // original file path
    // videoUrl: '/',
    // url:'/',
    // sttEngine: 'ibm',
    languageModel: 'en-US_BroadbandModel', // default is american US broadband model
    counterForPaperCuts: 0,
    audioFile: undefined,
    processedAudio: false,
    processedVideo: false,
    // status is marked as false by default and turned to true when transcription has been processed
    // could changed as status marked as null if there's an issue
    // so that can have 3 options. not set yet, gone wrong, success.
    status: false,
    highlights: [],
    // orderedPaperCuts:[],
    videoOgg: undefined,
    //TODO: get date from metadata of video
    metadata: undefined,
    text: undefined,
    //used for error handling when processing transcription
    error: undefined
  },

at the moment v1.0.6 there are some redundant attributes, such as counterForPaperCuts that could be removed.

highlights is used to do selections in transcription show view.

text is the body of the transcription.

Here is an example of the json.

{
    "languageModel": "en-US_BroadbandModel",
    "counterForPaperCuts": 7,
    "audioFile": "/Users/pietropassarelli/Library/Application Support/autoEdit2/media/Jesselyn_Radack-Mobile.mp4.1486205670442.ogg",
    "processedAudio": true,
    "processedVideo": true,
    "status": true,
    "highlights": [
        {
            "id": 0,
            "paperCutOrder": 6,
            "startTime": 580.0699999999999,
            "endTime": 588.0899999999999,
            "reelName": "NA",
            "clipName": "Jesselyn Radack-Mobile.mp4",
            "speaker": "Jesselyn Radack",
            "transcriptionId": 78885461,
            "videoId": "videoId_78885461",
            "videoOgg": "/Users/pietropassarelli/Library/Application Support/autoEdit2/media/Jesselyn_Radack-Mobile.mp4.1486205670442.webm",
            "audioFile": "/Users/pietropassarelli/Library/Application Support/autoEdit2/media/Jesselyn_Radack-Mobile.mp4.1486205670442.ogg",
            "text": "I just said that I wrote more emails and that %HESITATION and I didn't quite know what to do but I knew I didn't want to be a part of this ",
            "offset": "NA",
            "words": [
                {
                    "id": 1116,
                    "text": "I",
                    "startTime": 580.0699999999999,
                    "endTime": 580.24
                },
                {
                    "id": 1117,
                    "text": "just",
                    "startTime": 580.24,
                    "endTime": 580.4300000000001
                },
                ...
            ]
        },

    ],
    "videoOgg": "/Users/pietropassarelli/Library/Application Support/autoEdit2/media/Jesselyn_Radack-Mobile.mp4.1486205670442.webm",
    "metadata": {
        "filePathName": "/Users/pietropassarelli/Downloads/Jesselyn Radack-Mobile.mp4",
        "fileName": "Jesselyn Radack-Mobile.mp4",
        "date": "2016-02-29 12:52:17",
        "reelName": "NA",
        "timecode": "NA",
        "fps": "1/50",
        "duration": 1618.16
    },
    "text": [
        {
            "id": 0,
            "speaker": "Jesselyn Radack",
            "paragraph": [
                {
                    "line": [
                        {
                            "id": 0,
                            "text": "and",
                            "startTime": 12.09,
                            "endTime": 12.36
                        },
                        {
                            "id": 1,
                            "text": "like",
                            "startTime": 12.36,
                            "endTime": 12.53
                        },
                        ...
                    ],
                    "id": 0,
                    "startTime": 12.09,
                    "endTime": 19.92
                },


                .....
            ]
        }
    ],
    "title": "Jesselyn Radack-Mobile",
    "description": "",
    "videoUrl": "/Users/pietropassarelli/Downloads/Jesselyn Radack-Mobile.mp4",
    "sttEngine": "ibm",
    "_id": "78885461",
    "id": "78885461"
}

The part to consider when referring to as autoEdit transcription json is simply the content of text attribute tho:

"text": [
        {
            "id": 0,
            "speaker": "Jesselyn Radack",
            "paragraph": [
                {
                    "line": [
                        {
                            "id": 0,
                            "text": "and",
                            "startTime": 12.09,
                            "endTime": 12.36
                        },
                        {
                            "id": 1,
                            "text": "like",
                            "startTime": 12.36,
                            "endTime": 12.53
                        },
                        ...
                    ],
                    "id": 0,
                    "startTime": 12.09,
                    "endTime": 19.92
                },


                .....
            ]
        }
    ],
backbone transcription model in autoEdit