Optimizing the number of Unicorns and buffer size

faq-material

(ljpp) #4

With these changes the amount of free memory has reduced from 2GB to 1GB, so roughly -0.5 gigs per unicorn added.

Is there a recommendation for the amount of RAM to keep free, thus available for Linux to use as disk cache? I have 1/4 of the RAM (=2GB) allocated to the database buffer.


(Matt Palmer) #5

Increasing the number of unicorn workers to suit your CPU and RAM capacity is perfectly reasonable. The “two unicorns per core” guideline is a starting figure. CPUs differ (wildly) in their performance, and VPSes make that even more complicated (because you can never tell who else is on the box and what they’re doing with the CPU), so you start conservative, and if you find that you’re running out of unicorns before you’re running out of CPU and RAM, then you just keep increasing the unicorns.

As far as disk cache goes, you need as much as you need. As you increase RAM consumption, keep an eye on your IOPS graphs, and particularly the percentage utilisation of the disks (what sysstat refers to as %util). There’s two points you want to be aware of: when your read IOPS start to stay persistently higher (that means that the working set of disk pages no longer fits in memory – you may or may not have hit that already), and when peaks in disk utilisation start to get close to 100%. The former tells you when your RAM consumption is starting to impact performance, and the latter tells you when you’re starting to saturate the disks. You want to consume RAM up to the first point, and you can drive it as far as you’re comfortable towards the second point, depending on your tolerance for performance degradation (keep an eye on your service times!).

Another RAM-related metric to keep an eye on is swap rate. That’s not how much swap space is being used (as long as your swap partition isn’t full, it doesn’t matter), but the number of pages per second being written to and read from swap. Swap writes are fine, but if the system is constantly swapping pages in and out, even only a few per second, you probably want to back off on your RAM usage a bit.

Just to keep you on your toes, swap activity counts towards disk IOPS, too, so your disk utilisation will likely go loco bananas when you start to run out of RAM, due to both extra disk reads (because the cache isn’t big enough) and increased swap activity, because the working set of memory pages doesn’t fit any more. That’s a recipe for performance disaster right there.


How much is Discourse affected by a faster CPU?
How much of a performance booster is 2GB vs 1GB RAM?
(ljpp) #6

Thanks Matt. Crystal clear and informative, as usual.

Here is a some more nerd porn. Our API requests for this year. This is what it looks like when there is a hockey league transfer deadline on the 15th of Feb at 23:59 local time.


#7

What tools are you using for monitoring the api requests, which is shown from your image?


(Joshua Rosenfeld) #8

If I’m not mistaken, that’s Discourse’s built in report visualizer. Go to yoursite.com/admin, then click any of the statistic titles.


(Simeon) #9

Bump.

So are you saying 8 Unicorns is the upper limit on any server? I have a server with 8x CPU and was thinking changing the config to use 16?


(Jeff Atwood) #10

Maybe, it depends how much traffic you have. 16 unicorns sitting around doing nothing and waiting for incoming requests would not get you much.


(Simeon) #11

800k-1.2m pageviews/month. Discourse runs buttery smooth most of the time but then if we ever get 300 concurrent users it grinds to a halt.

Already have db_shared_buffers configured to 25% of the host’s 32gb RAM, looking for more possible fixes.


(Jay Pfaffman) #12

Sounds like it could help. Please report back.


(Simeon) #13

Done. We’ll see, not expecting another traffic spike for awhile.


(ljpp) #14

I am currently using 10 unicorns (IIRC) on my 6-core instance at UpCloud. This gets me above 400+ concurrents. We’ll see in February what it takes to choke this.


(Simeon) #15

Neat. I guess I’m just wondering if there’s a tipping point where having a too high number of unicorns is actually detrimental.


(Sam Saffron) #16

You know, there is only one way to find out … measure.


(Simeon) #17

Bump.

Forum appears to do much better under load but live reloading of posts and updating of notifications is sluggish.


(Rafael dos Santos Silva) #18

Do you use an external reverse proxy or cloudflare?


(Simeon) #19

Cloudflare yes

20 char


(Rafael dos Santos Silva) #20

We had several reports of people using cloudflare proxy feature and having problems with our live updates. Some people claim disabling brotli support helps, but we recommend using Cloudflare in DNS only mode (gray cloud).


(Simeon) #21

Oh right. Well I think I’d rather slow notifications than to lose the security/CDN benefits of Cloudflare.


(Rafael dos Santos Silva) #22


(ljpp) #23

That one needs a wildcard SSL certificate.

@simeon I serve couple of million pageviews per month using CloudFlare + Discourse. Disable Brotli and do not enable advanced acceleration methods (Railgun etc) and you should be good, for now,