Is llama 2 a good fit considering its small context window?
I presume you want information of some value to you otherwise you wouldn't bother reading an article. Then you feed it to a probabilistic algorithm and so you can not have any idea what the output has to do with the input. Like https://i.imgur.com/n6hFwVv.png you can somewhat decipher what this slop wants to be but what if the summary leaves out or invents or inverts some crucial piece of info?
Although I don't think this particular summarizer works for videos. And I don't think Ollama API supports audio ingestion for transcription. There are some summarizers that work with YouTube specifically (using automatic subtitles).