Fine tuned models for summarisation?
Fine tuned models for summarisation?
I have a db with a lot of data that all need precise summarisation, I would do it myself if it wasn't 20 thousand fields long
It is about 300k tokens, and Gemini 2.5 struggles missing points and making up facts
Separating them into smaller sections is not an option, because even when seperated they can take up 30k tokens, and the info that needs summarisation may span 100k token ranges
I learnt that fine tuning may have better results than general purpose models, and now I'm wondering if there is anything high token count for summarisation.
Any help would be appreciated, even if its to suggest another general purpose model that has better coherency
From my personal experience, I'd say generative AI isn't the best tool for summarization. It also frequently misses the point when I try. Or makes up additional facts which haven't been in the input text. (Or starts going on (wrong) tangents despite the task being to keep it short and concise.) And I'd say all(?) models do that. Even the ones that are supposed to be big and clever.
Edit: Lots of people use ChatGPT etc for summarization, though. So I really don't know who's right here. Maybe my standards are too high, but what I've read as output from small to big models like ChatGPT wasn't great.
There are other approaches in NLP. For example extractive summarization like the BART model from Facebook. That's precise. Some Lemmy bot uses LsaSummarizer, but I don't really know how that works. Or maybe you can re-think what you're trying to do and use RAG instead of summarization.
Looking into BART, thanks.