Looking for some input/feedback.
I'm running NodeBB on Heroku and RedisToGo. After a large data import (tens of thousands of posts) when I log in as any user and click into "Unread" and "Mark all as read" the memory usage on both the server and Redis instance spikes. In my case I'm still running in test mode with just 512MB in redis (it usually sits at 309 used or so) and 512MB of memory on a basic Heroku dyno.
Basically the dyno grinds to a halt from swapping and the redis sends an alert about being over quota. Once I kill the dyno memory in Redis drops back down, and of course the restarted dyno doesn't resume the operation.
Is this behavior expected? Do I just need to allocate more memory to allow the operation to complete without blowing out my resource quotas? Is there any way to just flag all posts as read by all users as a starting point since I'm guessing it's trying to mark tens of thousands of posts as read all in one operation?