Narrating over a Video using Multimodal AI

Nodes

24e3b914-15fa-444f-80e3-ca29bdacaf402e3e39e3-0e17-4fae-9ac5-9ee31f9711ed083a55c9-df56-4154-959f-52737890cad03b9f1b78-6b80-4575-aa6c-966987000389818211a2-bc09-427d-bfa4-d8db9fbffb501c04dd23-ae91-4772-b6db-e7ce81fc2f71cc410a33-acd1-4c16-aba9-bd8029d93b47+6

Created by

JiJimleuk

Last edited 39 days ago

This n8n template takes a video and extracts frames from it which are used with a multimodal LLM to generate a script. The script is then passed to the same multimodal LLM to generate a voiceover clip.

This template was inspired by Processing and narrating a video with GPT's visual capabilities and the TTS API

How it works

  • Video is downloaded using the HTTP node.
  • Python code node is used to extract the frames using OpenCV.
  • Loop node is used o batch the frames for the LLM to generate partial scripts.
  • All partial scripts are combined to form the full script which is then sent to OpenAI to generate audio from it.
  • The finished voiceover clip is uploaded to Google Drive.

Sample the finished product here: https://drive.google.com/file/d/1-XCoii0leGB2MffBMPpCZoxboVyeyeIX/view?usp=sharing

Requirements

  • OpenAI for LLM
  • Ideally, a mid-range (16GB RAM) machine for acceptable performance!

Customising this workflow

  • For larger videos, consider splitting into smaller clips for better performance
  • Use a multimodal LLM which supports fully video such as Google's Gemini.

New to n8n?

Need help building new n8n workflows? Process automation for you or your company will save you time and money, and it's completely free!