Skip to main content
Azure AI Video Indexer helps you find important moments in long meeting recordings and other long-form video content. Manually scanning recordings is time-consuming; Video Indexer automates that work by extracting searchable, time-aligned insights — spoken words, speakers, topics, sentiment, faces, on-screen text, and more — so you can jump straight to the moments that matter. For example, a product manager can instantly locate every segment across multiple meetings where a product launch was discussed. In this article we review Video Indexer’s core capabilities, show how to create and connect a resource in the Azure Portal, and outline how to upload, inspect, and programmatically retrieve insights. Key capabilities
  • Automatic transcription with timestamps and speaker diarization.
  • Topic detection, sentiment analysis, and conversation-level signals (questions, decisions).
  • Face detection and optional identification (when you provide reference images).
  • OCR for on-screen text, object and scene detection, and keyframes.
  • Content moderation to flag sensitive or inappropriate content.
These insights help content creators, educators, marketing teams, and enterprises categorize, search, and analyze video content far more efficiently than manual review.
A screenshot of the Microsoft Azure portal showing the "Create storage account" form with fields for Name, Account kind, Performance (Standard/Premium), and Replication (Locally‑redundant storage). The Name field is partially filled in and the page header shows "Azure AI Video Indexer."
Features at a glance
FeatureWhat it extractsTypical use case
Transcription & speaker diarizationTime-aligned text and who spoke whenSearch meetings by keyword or speaker
Topic detectionHigh-level themes and clustersGroup videos by subject or summarize content
Sentiment analysisPositive/negative/neutral per segmentTrack audience or speaker sentiment over time
Face & celebrity recognitionFaces and optional known identitiesTag speakers or highlight mentions of people
OCRText appearing on-screenExtract slide content, captions, or overlays
Scene detection & keyframesScene boundaries and representative framesGenerate thumbnails or chapter markers
Content moderationFlags for sensitive contentEnforce compliance and safety rules
Creating a Video Indexer resource in the Azure Portal
  1. In the Azure Portal search box, type “Video Indexer” and select Azure AI Video Indexer.
  2. Create a new Video Indexer resource. During creation you will:
    • Choose a subscription and resource group.
    • Choose or create a storage account (Video Indexer stores uploaded media and derived artifacts there).
    • Optionally connect other AI resources later for extended capabilities (not required to index videos).
  3. Review and create the resource. After deployment, the resource appears in your portal and can be associated with the Video Indexer web app.
A screenshot of the Microsoft Azure portal on the "Create a Video Indexer resource" review page showing a "Validation passed" message and a summary of resource settings (subscription, resource group, resource name, region, storage account). The browser toolbar and user menu are visible at the top.
Accessing the Video Indexer web app
  • Open the Video Indexer portal at https://videoindexer.ai and sign in with the same account used to create the Azure resource.
  • When you first sign in you may use a trial account; after you provision the Azure resource you can associate that resource and its storage account with your Video Indexer account.
  • Once associated, you can upload videos from your local machine, from the linked storage account, or via a public URL.
Uploading and indexing a video
  • Click Upload and choose a file from your device, select from the connected storage account, or paste a file URL.
  • Choose language, privacy settings, and the indexing preset that fits your needs.
  • Confirm any required consent and click Upload + Index. The service begins ingesting and processing the video.
A screenshot of the Azure AI Video Indexer web app showing an "Upload and index" dialog with a summary overview (video language, indexing preset, privacy) and a checked consent box, with the "Upload + index" button highlighted.
When uploading, be aware that certain advanced insights (for example, custom face recognition or celebrity recognition) may require additional configuration or permissions. Always confirm you have the required rights to process personal data in your region.
Indexing workflow and what runs after upload Once the upload begins, Video Indexer applies multiple models and pipelines, which may include:
  • Audio pre-processing (noise reduction, channel separation)
  • Automatic speech recognition and closed captions
  • Speaker diarization and speaker labeling
  • Object, scene, and keyframe detection
  • OCR for on-screen text
  • Topic and sentiment detection
  • Face detection and optional identification
  • Content moderation checks
Indexing time depends on video length, selected features, and queue load; typical processing ranges from a few minutes for short clips to longer for full-length recordings. Inspecting indexed results
  • After indexing completes open the video in the Video Indexer UI to see a timeline, interactive transcript, detected topics, object/scene tags, faces, and sentiment trends.
  • Use timeline search to jump to keywords, speaker segments, or flagged moments (questions, decisions).
  • Export or download metadata and subtitles, or copy insights into your content workflows.
A man wearing a "KodeKloud" t-shirt sits facing the camera with computer monitors behind him. The image is shown inside a browser window displaying the Azure AI Video Indexer interface and video insights on the right.
Programmatic access and the Video Indexer API Video Indexer provides REST APIs so you can automate uploads, control indexing, poll progress, and fetch insights as JSON for integration into search, analytics, or custom UI experiences. Basic flow
  1. Acquire an access token for your account.
  2. Upload the media or point Video Indexer to a storage URL.
  3. Monitor indexing status until complete.
  4. Retrieve insights (transcript, faces, OCR, topics, sentiment, keyframes).
Example: fetch an account access token (replace placeholders)
curl -X GET "https://api.videoindexer.ai/Auth/{location}/Accounts/{accountId}/AccessToken?allowEdit=true" \
  -H "Ocp-Apim-Subscription-Key: {SUBSCRIPTION_KEY}"
Example: upload a video (basic form upload; use the returned access token)
curl -X POST "https://api.videoindexer.ai/{location}/Accounts/{accountId}/Videos?name=my-video.mp4&accessToken={ACCESS_TOKEN}" \
  -F "file=@/path/to/my-video.mp4"
Note: The API supports many parameters to control language, indexing presets, speaker diarization, and callback webhooks for asynchronous workflows. Store returned JSON (timestamps, object metadata, face tags, OCR results) in your search index or database to enable fast query and rich application experiences. Quick troubleshooting tips
  • If indexing appears stuck, check the job status in the UI or via the API and review the video length/format.
  • Ensure the storage account is correctly linked and that the Video Indexer service has permission to read/write blobs.
  • For custom face recognition, pre-register reference images and ensure you comply with privacy and data protection regulations.
Summary Azure AI Video Indexer converts unstructured video into structured, searchable insights: transcription, speaker identification, face detection, OCR, topics, sentiment, scene segmentation, and content moderation. Use the web app for manual review and exploration, or the REST API to integrate indexing into automated pipelines and build searchable video experiences. Links and references

Watch Video