Skip to content

End-to-end scripting workflow to automatically generate show notes from audio/video transcripts with Whisper.cpp, Ollama, yt-dlp, and Commander.js

License

Notifications You must be signed in to change notification settings

ajcwebdev/autoshow

Repository files navigation

AutoShow

autoshow logo

Outline

Project Overview

AutoShow automates the processing of audio and video content from various sources, including YouTube videos, playlists, podcast RSS feeds, and local media files. It leverages advanced transcription services and language models (LLMs) to perform transcription, summarization, and chapter generation.

Prompts and Content Formats

AutoShow can generate diverse content formats including:

  • Summaries and Chapters:
    • Concise summaries
    • Detailed chapter descriptions
    • Bullet-point summaries
    • Chapter titles with timestamps
  • Social Media Posts:
    • X (Twitter)
    • Facebook
    • LinkedIn
  • Creative Content:
    • Rap songs
    • Rock songs
    • Country songs
  • Educational and Informative Content:
    • Key takeaways
    • Comprehension questions
    • Frequently asked questions (FAQs)
    • Curated quotes
    • Blog outlines and drafts

Key Features

  • Support for multiple input types (YouTube links, local video and audio files)
  • Integration with various:
    • LLMs (ChatGPT, Claude, Gemini)
    • Transcription services (Deepgram, Assembly)
  • Customizable prompts for generating titles, summaries, chapter titles/descriptions, key takeaways, and questions to test comprehension
  • Markdown output with metadata and formatted content

AutoShow Pipeline

The AutoShow workflow includes the following steps that feed sequentially into each other:

  1. The user provides a content input (video URL or local file) and front matter is created based on the content's metadata.
  2. The audio is downloaded (if necessary).
  3. Transcription is performed using the selected transcription service.
  4. A customizable prompt is inserted containing instructions for the show notes or other content forms.
  5. The transcript is processed by the selected LLM service to generate the desired output based on the selected prompts.

Setup

.github/setup.sh checks to ensure a .env file exists and Node dependencies are installed. Run the workflow with the setup script in package.json.

npm run setup

Run AutoShow

Example commands for all available options can be found in docs/README.md.

npm run dev

Open localhost:4321.

Contributors

About

End-to-end scripting workflow to automatically generate show notes from audio/video transcripts with Whisper.cpp, Ollama, yt-dlp, and Commander.js

Topics

Resources

License

Stars

Watchers

Forks

Sponsor this project

 

Packages

No packages published