zlacker

[parent] [thread] 27 comments
1. arter4+(OP)[view] [source] 2023-07-01 19:15:11
This is interesting.

Judging from the screenshot, a huge amount of GET /TweetDetail is generated which triggers some rate limiting, as shown by the 429.

If this is indeed due to the recent decision to enforce authentication for all API calls, it means the curlprit may actually be the API gateway or something similar downstream.

Also, this behavior seem to never stop, which isn't what one would expect from an exponential backoff retry.

I don't claim to be a better engineer than the folks working at Twitter, but it is interesting to see something like this in the wild, all Musk-related considerations aside.

replies(6): >>bheadm+5g >>Quarre+og >>cactus+2h >>romseb+gk >>readyp+ap >>kccqzy+VK
2. bheadm+5g[view] [source] 2023-07-01 20:38:08
>>arter4+(OP)
> If this is indeed due to the recent decision to enforce authentication for all API calls, it means the curlprit may actually be the API gateway or something similar downstream.

The way I understand it, DDoS is not caused by enforced authentication - enforced authentication is just a temporary measure against DDoS.

3. Quarre+og[view] [source] 2023-07-01 20:40:13
>>arter4+(OP)
I would guess the front end was written under the assumption that the back end would still work without auth. Perhaps the backend changes (mandatory auth + rate limiting) were pushed without sufficient testing of the front + back?
4. cactus+2h[view] [source] 2023-07-01 20:43:50
>>arter4+(OP)
Did Elon pay the AWS bill? That seems like a likely culprit. Twitter instances are being forcibly shutdown.
replies(3): >>amluto+1i >>badwol+Jv >>colech+ig2
◧◩
5. amluto+1i[view] [source] [discussion] 2023-07-01 20:50:54
>>cactus+2h
Twitter operates its own datacenters.
replies(3): >>cactus+Ui >>cududa+To >>stefan+LA
◧◩◪
6. cactus+Ui[view] [source] [discussion] 2023-07-01 20:55:45
>>amluto+1i
"Twitter and AWS signed a five-and-a-half-year contract in 2020, which AWS is not willing to renegotiate."

https://gritdaily.com/twitter-owes-aws-millions/

replies(1): >>willia+Vj
◧◩◪◨
7. willia+Vj[view] [source] [discussion] 2023-07-01 21:01:25
>>cactus+Ui
Twitter.com and the associated user-facing services do not run on AWS.
replies(1): >>vGPU+wG
8. romseb+gk[view] [source] 2023-07-01 21:02:57
>>arter4+(OP)
"curlprit" for too many GET's causing a 429 is just the perfect typo.
◧◩◪
9. cududa+To[view] [source] [discussion] 2023-07-01 21:32:24
>>amluto+1i
Yeah but they use GCS for auth, moderation, and caching. They apparently haven’t been paying Google since April and the contract expired June 30th
10. readyp+ap[view] [source] 2023-07-01 21:34:02
>>arter4+(OP)
Could someone report the error at press@twitter.com and see what they think about it?
replies(2): >>vuln+MF >>ineeda+TP
◧◩
11. badwol+Jv[view] [source] [discussion] 2023-07-01 22:16:53
>>cactus+2h
Well, they haven't paid their GCP bill... https://theconversation.com/twitter-is-refusing-to-pay-googl...
replies(1): >>o1y32+FK1
◧◩◪
12. stefan+LA[view] [source] [discussion] 2023-07-01 22:59:25
>>amluto+1i
And yet they also host with AWS, Google Cloud and Oracle. Cloud people take note: this is what lock-in looks like, and it's coming to a place near you.
replies(2): >>SkyPun+sB >>firest+rJ
◧◩◪◨
13. SkyPun+sB[view] [source] [discussion] 2023-07-01 23:04:42
>>stefan+LA
No, this has nothing to do with lock-in. This has done nothing to do with decision making that subverts good engineering.
◧◩
14. vuln+MF[view] [source] [discussion] 2023-07-01 23:43:11
>>readyp+ap
Quite the knee slapper. Thanks. I hope you didn’t spend many brain cycles on it.
replies(3): >>Scalen+ZK >>meepmo+2T >>oneeye+3o1
◧◩◪◨⬒
15. vGPU+wG[view] [source] [discussion] 2023-07-01 23:49:48
>>willia+Vj
While it looks like they never started the move over to AWS, the press release makes it sound like they do use some AWS services.

> In addition, Twitter will continue to use AWS services such as Amazon CloudFront (AWS’s fast content delivery network service that securely delivers data, videos, applications, and APIs with low latency and high transfer speeds to customers globally) and Amazon DynamoDB (AWS’s key-value database that delivers single-digit millisecond performance at any scale).

replies(1): >>willia+sH
◧◩◪◨⬒⬓
16. willia+sH[view] [source] [discussion] 2023-07-01 23:57:39
>>vGPU+wG
I worked there. Services running on GCP are a significant part of the internal service infra (ml platform, etc.) and it's not impossible that the abrupt loss of GCP would cause user-facing problems. The GCP spend was many, many times the AWS spend. Unless things changed since last November, AWS is not a meaningful part of the internal or user-facing infra.

With respect to DynamoDB specifically, Twitter has its own custom distributed key-value store: https://blog.twitter.com/engineering/en_us/a/2014/manhattan-... that twitter.com itself runs on.

replies(1): >>18pfsm+KL
◧◩◪◨
17. firest+rJ[view] [source] [discussion] 2023-07-02 00:12:50
>>stefan+LA
Cloud agnostic is hard
18. kccqzy+VK[view] [source] 2023-07-02 00:27:42
>>arter4+(OP)
While exponential backoff is theoretically optimal, I doubt it's actually used that often in practice. I've seen too many cases where someone decides serving user requests with low latency is so important that they'd rather have a constant randomized backoff than exponential backoff. I've been in many design meetings and seen enough documents where the decision not to use exponential backoff is explicitly made, understanding the tradeoff with overloading and system recovery.
replies(2): >>klabb3+mP >>colech+9g2
◧◩◪
19. Scalen+ZK[view] [source] [discussion] 2023-07-02 00:27:58
>>vuln+MF
Making jokes about Twitter is too easy.
◧◩◪◨⬒⬓⬔
20. 18pfsm+KL[view] [source] [discussion] 2023-07-02 00:34:43
>>willia+sH
Thanks for weighing in with some actual first-hand knowledge. It is appreciated.

The latest on cloud hosting is from a week ago, and I'm guessing you don't have any more recent info than this:

https://www.reuters.com/technology/twitter-resumes-paying-go...

replies(1): >>willia+TN
◧◩◪◨⬒⬓⬔⧯
21. willia+TN[view] [source] [discussion] 2023-07-02 00:56:19
>>18pfsm+KL
Correct, no more recent (or less public) info than that. Like I say, losing GCP could cause problems users notice, but sounds like that’s not going to happen.
◧◩
22. klabb3+mP[view] [source] [discussion] 2023-07-02 01:05:14
>>kccqzy+VK
I wouldn’t be surprised if this has no back off or limit at all. 10 RPS is fast enough that it may simply be sequential.
◧◩
23. ineeda+TP[view] [source] [discussion] 2023-07-02 01:10:00
>>readyp+ap
Whoever is in charge of that account went all Oregon Trail on things and caught dysentery
◧◩◪
24. meepmo+2T[view] [source] [discussion] 2023-07-02 01:43:53
>>vuln+MF
Why do you take people criticizing Elon Musk so personally?
◧◩◪
25. oneeye+3o1[view] [source] [discussion] 2023-07-02 07:46:46
>>vuln+MF
Do you think it's more or less funny than auto-replying to all PR enquiries with poo, when you are an incredibly large company?
◧◩◪
26. o1y32+FK1[view] [source] [discussion] 2023-07-02 12:04:37
>>badwol+Jv
That's slightly outdated information:

https://www.engadget.com/twitter-has-supposedly-started-payi...

◧◩
27. colech+9g2[view] [source] [discussion] 2023-07-02 15:51:45
>>kccqzy+VK
I’ve had to… uhh… eagerly advocate for exponential backoff for weeks of constant uptime issues before someone listened and actually implemented it and solved the problems.

Like several times in different roles.

People do it, exponential backoff is everywhere in your stack, but it doesn’t end up in your application layer until you have enough traffic that you actually have to manage throughout.

◧◩
28. colech+ig2[view] [source] [discussion] 2023-07-02 15:53:08
>>cactus+2h
This is not remotely a likely culprit.
[go to top]