Hey @sam yes indeed we are happy Discourse customers, and one of the most common pairings with GPT-4 for the exact use case you mentioned ā see the logos + quotes on our homepage. Can we help you with a POC?
Absolutely, I am going to connect you with @Falco and you can discuss.
I think it would be delightful for discourse-ai to ship with a pinecone adapter as well, it reduces enormous amounts of friction for self hosters of our platform.
It seems like youāve done your research on the costs of training, but I wanted to share my understanding based on the OpenAI fine-tuning guide. If I understand https://platform.openai.com/docs/guides/fine-tuning correctly, they recommend using Ada for classification tasks and providing 100 examples of each class. In that case, we would have a total of 200 examples (spam and not spam). Assuming an average example consists of 500 tokens, the total would be 500 * 200 = 100,000 tokens on Ada, which would cost US$ 0.04 to train. If you were to use Davinci instead, the cost would be US$ 3.00.
I guess that the pricing might be for a single step or a single epoch of training, but I couldnāt find any more detailed information on their website. Please let me know if you have any insights or if Iāve misunderstood something.
As I mentioned, those costs were for my use case for my business. My training and usage is Davinci, not Ada, so 75x more expensive there. We also practically max out tokens per request.
I donāt know exactly what Sam/Falco would have in mind for their use caseājust mentioning generally that fine tuning can be expensive at scale!