One things they are really good at today, even in the 2gb 6b param models, is summarization. I want to put together a demo on notedeck where it downloads the model and uses it to summarize threads when you open a large one.

Replies (1)

prompt injection is still a big deal sadly… there’s actually dedicated models for summarization out there. you could fine tune existing summarization models to take additional context input like reply chains via tokens