Autoshow automates the processing of audio and video content from various sources, including YouTube videos, playlists, podcast RSS feeds, and local media files. It performs transcription, summarization, and chapter generation using different language models (LLMs) and transcription services.
The Autoshow workflow includes the following steps:
- The user provides input (video URL, playlist, RSS feed, or local file).
- The system downloads the audio (if necessary).
- Transcription is performed using the selected service.
- The transcript is processed by the chosen LLM to generate a summary and chapters.
- Results are saved in markdown format with front matter.
- Support for multiple input types (YouTube links, RSS feeds, local video and audio files)
- Integration with various:
- LLMs (ChatGPT, Claude, Gemini, Cohere, Mistral, Fireworks, Together, Groq)
- Transcription services (Whisper.cpp, Deepgram, Assembly)
- Local LLM support with Ollama
- Customizable prompts for generating titles, summaries, chapter titles/descriptions, key takeaways, and questions to test comprehension
- Markdown output with metadata and formatted content
- Command-line interface for easy usage
- WIP: Node.js server and React frontend
See docs/roadmap.md
for details about current development work and future potential capabilities.
scripts/setup.sh
checks to ensure a .env
file exists, Node dependencies are installed, and the whisper.cpp
repository is cloned and built. Run the script with the setup
script in package.json
.
npm run setup
Run on a single YouTube video.
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk"
Run on a YouTube playlist.
npm run as -- --playlist "https://www.youtube.com/playlist?list=PLCVnrVv4KhXPz0SoAVu8Rc1emAdGPbSbr"
Run on a list of arbitrary URLs.
npm run as -- --urls "content/example-urls.md"
Run on a local audio or video file.
npm run as -- --file "content/audio.mp3"
Run on a podcast RSS feed.
npm run as -- --rss "https://ajcwebdev.substack.com/feed"
Use local LLM.
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --ollama
Use 3rd party LLM providers.
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --chatgpt GPT_4o_MINI
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --claude CLAUDE_3_5_SONNET
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --gemini GEMINI_1_5_PRO
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --cohere COMMAND_R_PLUS
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --mistral MISTRAL_LARGE
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --fireworks
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --together
npm run as -- --video "https://www.youtube.com/watch?v=MORMZXEaONk" --groq
Example commands for all available CLI options can be found in docs/examples.md
.
-
Main Entry Point (
src/autoshow.ts
)- Defines the command-line interface using Commander.js
- Handles various input options (video, playlist, URLs, file, RSS)
- Manages LLM and transcription options
-
Command Processors (
src/commands
)processVideo.ts
: Handles single YouTube video processingprocessPlaylist.ts
: Processes all videos in a YouTube playlistprocessURLs.ts
: Processes videos from a list of URLs in a fileprocessFile.ts
: Handles local audio/video file processingprocessRSS.ts
: Processes podcast RSS feeds
-
Utility Functions (
src/utils
)downloadAudio.ts
: Downloads audio from YouTube videosrunTranscription.ts
: Manages the transcription processrunLLM.ts
: Handles LLM processing for summarization and chapter generationgenerateMarkdown.ts
: Creates initial markdown files with metadatacleanUpFiles.ts
: Removes temporary files after processing
-
Transcription Services (
src/transcription
)whisper.ts
: Uses Whisper.cpp, openai-whisper, or whisper-diarization for transcriptiondeepgram.ts
: Integrates Deepgram transcription serviceassembly.ts
: Integrates AssemblyAI transcription service
-
Language Models (
src/llms
)chatgpt.ts
: Integrates OpenAI's GPT modelsclaude.ts
: Integrates Anthropic's Claude modelsgemini.ts
: Integrates Google's Gemini modelscohere.ts
: Integrates Cohere's language modelsmistral.ts
: Integrates Mistral AI's language modelsfireworks.ts
: Integrates Fireworks's open source modelstogether.ts
: Integrates Together's open source modelsgroq.ts
: Integrates Groq's open source modelsprompt.ts
: Defines the prompt structure for summarization and chapter generation
-
Web Interface (
web
) and Server (server
)- Web interface built with React and Vite
- Node.js server that handles backend operations for the web interface
- Note: Just a proof of concept, very little functionality built at this point. Expect these to catch up with the CLI starting in Q4 2024