Dan Ratner’s Post

View profile for Dan Ratner, graphic

Entrepreneur, Writer, Technologist

I've been experimenting with a variety of fine-tuned large language models for writing fiction interactively. LoRA seems to be a very successful technique for capturing tone and turn of phrase, but I keep running into context limits even with models with fairly large windows like Mistral. I made a library for ChatGPT that deals with this by automatically asking the model to summarize the chat history into a dense recap that preserves key information about the chat history while discarding the verbatim chat history to reduce context size. It seems to work pretty well. I'm going to try to adapt this for open source models as well. Has anyone else tried a technique like this? It seems to map to human experience - we don't remember every word of a long conversation, but we do remember key points and unique turns of phrase. For anyone who is interested, here's the demo code: https://lnkd.in/gBtZUbzM

gogpt/chat.go at main · dratner/gogpt

gogpt/chat.go at main · dratner/gogpt

github.com

To view or add a comment, sign in

Explore topics