UPDATE: The latest RC version of Lemmy-ui (0.18.2-rc.2) contains fixes for the issue, but if you believe you were vulnerable, you should still rotate your JWT secret after upgrading! Read below for instructions. Removing custom emoji is no longer necessary after upgrading.
Original post follows:
This post is intended as a central place that admins can reference regarding the XSS incident from this morning.
What happened?
A couple of the bigger Lemmy instances had several user accounts compromised through stolen authentication cookies. Some of these cookies belonged to admins, these admin cookies were used to deface instances. Only users that opened pages with malicious content during the incident were vulnerable. The malicious content was possible due to a bug with rendering custom emojis.
Stolen cookies gave attackers access to all private messages and e-mail addresses of affected users.
Am I vulnerable?
If your instance has ANY custom emojis, you are vulnerable. Note that it appears only local custom emojis are affected, so federated content with custom emojis from other instances should be safe.
I had custom emojis on my instance, what should I do?
This should be enough to mitigate now:
- Remove custom emoji
DELETE FROM custom_emoji_keyword;
DELETE FROM custom_emoji;
- Rotate your JWT secret (invalidates all current login sessions)
-- back up your secret first, just in case
SELECT * FROM secret;
-- generate a new secret
UPDATE secret SET jwt_secret = gen_random_uuid();
- Restart Lemmy server
If you need help with any of this, you can reach out to me on Matrix (@sunaurus:matrix.org
) or on Discord (@sunaurus
)
Legal
If your instance was affected, you may have some legal obligations. Please check this comment for more info: https://lemmy.world/comment/1064402
More context:
It seems to me that the scope of this could have been mitigated with a simple privilege separation policy for admin server accounts but I see a lot of (what looks like) server admins using that account as their daily driver.
Also, lemmy-ui should post a security advisory to their github.
Welp, that is some simple lapse of separation-of-duties principles that shouldn’t have been.
I don’t pretend to know the demographics of lemmy server admins, but my gut says it’s predominantly hobbyists and devops types, rather than grizzled system admins.
People in DevOps should know better than reusing their accounts. It doesn’t take a seasoned system admin to know about basic OPSEC and separation amongst profiles.
Agreed. I’m pointing it out in hopes that becomes one of the takeaways of this incident.
(Unrelated, why are you marked at a bot account?)
Wait, why am I marked as a bot account?
This is the first time I’ve heard about this. Whom should I ask for help?
It’s a checkbox in the setting for your account. You should be able to change it.
Hi, could you check now?
Yes and no. Admin accounts often remain logged in as a practical matter. They can’t see incoming reports, registration applications, etc. if the account isn’t logged in. And there is no “middle tier”/sitemod or customizable permissions allowing for anything between community mod and instance admin that would mitigate the need to use admin accounts day to day.
I still wouldn’t use the admin account as my daily driver. Leave it open in another browser/private tab specifically to perform admin actions (as noted) but not for browsing/posting/community modding. I understand how that’s a pain but given how early days we are with this platform and the high probability of more issues surfacing, it’s a necessary pain.
I’d really like to see the devs add some tools to mitigate future risk and further protect admin accounts. The least of which being that admin actions require stronger validation than a browser side cookie, and frequent re-validation to perform admin actions.
I still wouldn’t use the admin account as my daily driver. Leave it open in another browser/private tab specifically to perform admin actions
i’ve thought about switching to a different account for non-admin tasks but I was a normal user here first and I want to keep using this account for normal posting too. I think using a different account is kind of pointless from a security standpoint, because performing mod and admin actions requires looking at user-generated content anyway and there is no way around that (except for maybe moving to a very clipboard-heavy workflow which would make everything take much longer).
There is work underway to fix these XSS bugs in a more comprehensive way (better CSP, and HttpOnly cookies…) which will hopefully land soon.
(i am not a lemmy developer but i’m reading the github and matrix chat…)
I mean. You could still be a normal user and have a separate admin account.
I get your point. But it’s a huge risk.
I get your point. But it’s a huge risk.
Using a separate account doesn’t substantially mitigate the risk. It might reduce the chances of getting randomly exploited, but it’s easy to post things admins need to see to do their job so any attacker wanting to target admins would still be able to even if we used separate accounts.
Systemically fixing the XSS problems is necessary either way :)
edit: actually I guess most common admin activities could be separated from the rarely used ones, at least… or the infrequent actions could simply require re-authenticating. That wouldn’t be a bad idea.
To be clear I’m not talking about separate mod accounts. I’m talking just admin.
that wouldn’t have necessarily stopped this attack I don’t think, but yeah, probably a good idea on multiple levels.
If the separate admin window was open, and a tagged reply or PM was sent to the admin account I think that would render the emote in the notification and trigger the exploit
Fair point. I know that I don’t have the answers. I do think that admin actions need to be more stringently scrutinized. Maybe something like a “sudo” model where your a normal acct 99% of the time and admin actions require a temporary elevation.
This bug also needs to be fixed so that a user logging out / changing password actually correctly invalidates their JWT token: https://github.com/LemmyNet/lemmy/issues/3364
I suspect that last night even if an admin knew their account was compromised, they probably couldn’t do anything about it without DB access.
This bug also needs to be fixed so that a user logging out / changing password actually correctly invalidates their JWT token: https://github.com/LemmyNet/lemmy/issues/3364
Okay it is insane that that issue is 2 weeks old and was not prioritized and because of that the hacker was gifted 2 weeks to prepare an attack.
I love Lemmy and am grateful for the work of the devs. But I see huge issues with the LemmyNet governance.
Previously it was captchas that were removed because one of the two main devs had a strong personal belief that captchas are “useless” and wanted to impose that belief upon everyone, which then led to an enormous wave of bots with the 0.18.0 as captchas were removed. I’m glad he was then convinced that he was wrong to remove them and then took steps to revert that decision, but it took too long and many instances suffered performance problems by being forced to stay with the 0.17.6 version because it had captchas.
Right now the same thing happens again with issues about security being left unanswered for two weeks. I believe right now all feature developments should be paused and a security audit of the whole code base should be the #1 priority.
This is just very bad. Proper governance and prioritization would have avoided exposing minors to lemon party porn and other disgusting content.
I can only imagine how helpless the admin whose account was compromised felt if she didn’t have access to the database to invalidate those tokens, and it could have been prevented if that issue was properly prioritized.
Changing password does invalidate tokens, but the rest of that is accurate.
Thanks for posting. There really should be a button which allows the admins to log everyone out for crisis situations like there I think
Changing the JWT secret does this. So instead of a button, its a line of code, making it less likely to be done by mistake.
True as that is, not every admin has DB access and those with DB access might be AFK.
Surely this exploit proves that it’s best to minimise the number of administrative actions available in the UI?
Destructive actions, sure. This would not be one such actions and it would be self-defeating.
In crisis situations like this, the last thing you want is to rely on some portion of the application UI working.
If the UI doesn’t work, there’s always the DB directly. But it’s good to have that option.
JWT secret keys are not in the DB (speaking typically, maybe for Lemmy they are, but that would be very surprising), that’s typically an environment variable or configuration file sort of thing.
In any case, this isn’t the part that’s broken, it doesn’t need fixed.
Rotate your JWT secret (invalidates all current login sessions)
– back up your secret first, just in case SELECT * FROM secret; – generate a new secret UPDATE secret SET jwt_secret = gen_random_uuid();
It’s in the DB. Check the OP
I am also not suggesting a fix. I am suggesting an improvement
Oof, okay well that’s not how I would’ve done it. The JWT secret in the database itself could be a vulnerability (e.g., someone that gains read only access to the database could then use that as a wedge to create any JWT they wanted). I’m not sure if that’s actually worth bringing up or not (it’s a bit of an odd case).
It used to be in a config file but they moved it to the db so it could be autogenerated on first startup I believe. I guess it’s better than newb instances running around with a jwt secret of “changeme” but still…
Hm… They could’ve edited the config or just
exit(1)
if the credential is the default, but very fair.
afaik to generate those tokens, you configure a secret in an enviroment variable. You cannot generate tokens from looking at valid tokens within the database. Thus storing active tokens in the database is fine since you can always purge all active tokens as this post has also suggested.
This is the secret, not the tokens.
My personal instance has no custom emoji’s, yet the SQL to remove onload= from content removed 16 effected comments. Does that mean federated instances are effected?
This does not mean your instance was affected. You’re just cleaning up comments which may or may not have successfully worked on other instances by abusing the emoji bug there.
Roger okay. Thanks for clearing that up.
I’m curious why the UI update to 0.18.2-rc.1 isn’t mentioned anywhere in this thread. I understood that updating the UI was a fix for this exploit.
My post predates the new versions, but I am editing the post now!
Awesome, thank you!
edit: you may want to specify that it’s
lemmy-ui
that has the update, notlemmy
.
We’ll likely do a lemmy release tomorrow with this fix included.
Stolen cookies gave attackers access to all private messages
Aren’t “private messages” not really private on lemmy anyway?
Under normal circumstances, they are only visible to:
- sender (and whoever has DB access on the sender’s instance)
- recipient (and whoever has DB access on the recipient’s instance)
- in case a private message is reported, all admins of the reporter’s instance
It is still considered a breach of user data if such messages are leaked.
“private” messages are not private anywhere.
unless they’re E2EE
How did the hackers get the cookies in the first place? Compromised devices on the clients?
🅴🆁🅶🆈 1•
How did the hackers get the cookies in the first place? Compromised devices on the clients?
I’m not affiliated with lemmy.world or any other instance but I do software and can explain some of the jargon above. XSS is the abbreviation of one of the most common attacks we see on websites, cross-site scripting. This attack works by having some vulnerable code which arbitrarily executes some javascript on a users browser.
In this case, the attacker seems to have found a vulnerability where a specially crafted character is executed when users read posts or comments containing it. In this case, it was especially bad because of how passwords are stored. When you log in to pretty much any website, passwords are stored in the form of cookies; small pieces of data that are passed back and forth to the web server and the client automatically. Usually, these cookies are set to not be readable by javascript, but in this case, it appears that that flag was not set. This allowed the XSS exploit to be sent back to a computer which was set up to grab these cookies.
One thing to note is that although the cookies were stolen, our passwords wouldn’t have immediately been compromised. Our login cookies store jwt tokens; a cryptographically signed message proving that we provided a password previously. That’s not to say that further escalation isn’t possible, but its hard to say for certain how far the hack went before being noticed.
Thank you for the detailed explanation :)
deleted by creator
Why is lemmy creating html nodes concatenating strings? This is going to happen again and again
I should have expected this judging by the state of the WS bugs
This is not the general way of doing things in Lemmy. It was just a mistake that unfortunately did not get caught in code review. These things happen from time to time (even in projects which are FAR bigger and better funded than Lemmy)
It’s probably worth having someone audit the codebase to ensure that there’s no other places that build HTML using string concatenation. https://github.com/LemmyNet/lemmy-ui/pull/1897 looks like a temporary patch rather than an actual fix. I can try to take a look if I get some free time.
As an example, Google’s own website domain was once bought by someone other than Google after they discovered it on a list of domains for sale. Out of curiosity, they tried to buy it and were (to their surprise) successful. The person owned it for about a minute before Google caught on and reversed the transaction. https://money.cnn.com/2016/01/29/technology/google-domain-purchase/index.html