This is covered in the OP. Yes.
Any chance there could be some settings for this quick chat button. At the moment it appears to always take the user to a personal chat with the bot. Iām not sure I want to encourage lots of hidden personal communication with the bot. Users can go to ChatGPT themselves for that. If we had a setting to either direct user to a group chat channel or to create a new topic in a Category specifically setup for AI Discussions.
Iāve tried setting up a category where users can raise topics and effectively ask the bot questions. Can perhaps we flag categories that perform in this way where the bot will:
- Always respond to the first post of a topic in that category
- Continue to respond to posts if there are no other users in the discussion
- Once other users enter discussion then the bot stops unless specifically replied to or mentioned in the post.
All these are only suggestions / ideas. I really like this plugin!
Oh BTW when trying to enable the summary plugin you mentioned my re-build fails and I have to remove the plugin to get the site up again. (Iāll post this separately on that thread in a bit with details)
Yes you can disable it.
Donāt forget about the quota system.
PR welcome for anything more sophisticated.
Yes, please do, Iām not having the same issue. I just rebuilt a site with it and it went fine.
so sadddddddddddd =(
Well, look on the bright side, you didnāt have to hire me to write the adaptor plugin! I definitely charge more than OpenAI will charge you
Perhaps you can get your community to help pay for it?
I have a suspicion it wonāt be that expensive to run if you keep the user quotas low:
This is what Iāve spent so far building two plugins on OpenAI, well within the free trial quota:
Plus ChatGPT
endpoint model is a 10th the cost of text-davinci-003
Simple math, based on default chatbot settings and gpt-3.5-turbo model.
This means if you have 100 āaverageā high-TL users, who happens to be using their max chatbot requests every week, it would only cost you less than 20 dollars a month.
EDIT: Oops, I forgot to account for input tokens.
The dollar is worth 6 times more than the currency of my country, it is too expensive for me
Do you have any forum for me to see how this gpt chat works? just out of curiosity, I found it so revolutionary, I wanted to test it
Just try ChatGPT:
Or playground if you want to fine-tune. Keep in mind playground doesnāt have turbo-3.5 (yet?)
Pick up a token, itās free for ~3 months and $18 of value. Then you can test out the plugin too.
I think itās best to keep the code as simple as possible, to make better enhancement in the feature without too much complications. Current behavior in topics already is like you describe, except it wonāt reply to the first post automatically. However you can easily achieve this by adding a topic template for your ask-AIBot category mentioning @AIBot (and you can even add a default prompt to fine-tune the responses!!!)
I wonder if it would help to give the LLM a bit more context - perhaps Category description, or the content of a sticky post?
Yeah, Iāve already thought about the Title being important. We could ship that with every request too.
Good suggestion, Iāll roadmap it.
Also, it would be better if it had context of the usernames, it seems to think all the previous messages are mine (in a group chat with multiple participants).
Yeah, I have implemented that in AI Topic Summary (and it can work really well), but not here. Another candidate for a switch, maybe.
The distinction is made by user/assistant, but Iām not sure of the effect of adding in usernames and what format to do that with here ⦠this needs experimentation I suspect.
Yes it will always be bot said this and user said this. The AI is designed for user-bot interactions. Not being in large groups (even though it can work with that). That being said, theoretically, itās possible to feed {user} input with multiple names and messages⦠but would it give better responses?
The AI is as good as the input it gets. Simply adding some usernames is not something that could always improve the responses, it can also generate confused responses. And in general, the username of content is not relevant because the output is based on facts not opinions. So having some doubts on this.
@codergautam Why donāt you experiment on ChatGPT and give it the same input (post content of N posts + adding random usernames above each post?) to see how it will respond in different scenarios. Also Iām curious what responses you are getting that lead to your request?
Just a thought: add the userās bio from their profile as part of the context. Even just āMarcP joined in October 2019 and has read 11 hours recentlyā
That is, put a cast of characters into the preamble.
I hope your wallet is deep Ed. Every time you add data you are spending money. :). There would also be a token count challenge here at present limits.
Well I asked the AI to summarize whatās going on in this chat, and it thought that all the conversation was from me and it proceeded to give me a lecture about how inappropriate I am for doing this.