I have 64 zipped megabytes of AIM conversations I had in high school. how hard would it be to train an LLM to be me from 15 years ago?
I have 64 zipped megabytes of AIM conversations I had in high school. how hard would it be to train an LLM to be me from 15 years ago?
It is called finetuning. I haven't tried it but oobagooba's text-generation-webui has a tab to do it and I believe it is pretty straightforward.
Fine tune a base model on your dataset and then tou will then need to format your prompt in the way your AIM logs are organized. e.g. you will need to add "
<ch00f>
" add the end of your text completion task. It will complete it in the way it learnt it.If you don't have a the GPU for it, many companies offer fine-tuning as a service like Mistral