Experiments with AI based moderation on Discourse Meta

I have not posted in a while, despite daily visiting my little chat window and having it be helpful at least once or twice per day… consistently.

The reason for my delay here was that I had to work through this rather large change.

It provides a subtle, yet critical, improvement to Discourse AI.

I was regularly noticing the moderation bot talk about completely irrelevant images, due to the way we constructed context. The change allows us to present mixed content (containing images and text in a correctly ordered fashion).

This means the LLM no longer gets confused.

What’s next?

  1. We have no way in automation to let it call a rule after post editing has “settled”, llm calls can be expensive, just because people edit typos we don’t want to scan something over and over again. I am not sure if this is required here, but I would like to allow for the possibility of triggering an automation once a post settles into the new shape.
  2. Prompt engineering - the current prompt is OK, but a bit too loud for my liking, it is bugging me a bit too much, I may soften it some
  3. Improved context - one thing that really bugs me is that the automation is now has not awareness of user trust. Some users are far more trusted in a community than others (eg: moderators) I would like to see if we can improve this story.
  4. Ability to run the automation on batches of posts for fast iterations.
  5. I am sure a lot more will pop up.
6 Likes