Previous incidents
API is degraded
Resolved Aug 27 at 10:18am PDT
The system has recovered. We have escalated the root cause to our upstream provider.
1 previous update
API is degraded
Resolved Aug 26 at 08:03am PDT
We are back.
1 previous update
Dashboard is partially unavailable or slower than usual
Resolved Aug 21 at 12:57pm PDT
Upstream provider has resolved the issue.
1 previous update
Scrape is timing out
Resolved Aug 21 at 05:09am PDT
The issue is now resolved. We had to upgrade the underlying Redis instance. Crawls may have been interrupted -- please reach out to support if you have any issues.
1 previous update
API is degraded
Resolved Aug 20 at 03:35am PDT
The issue is resolved.
1 previous update
firecrawl.dev is down
Resolved Aug 18 at 04:59pm PDT
firecrawl.dev recovered.
1 previous update
API is degraded
Resolved Aug 14 at 02:26pm PDT
We are back. Job timeout metrics have recovered to pre-incident levels.
The issue came down to misconfigured pipeline queue limits on Dragonfly -- we configured them with high values expecting a high load on production, however, they ended up being ridiculously high. This caused Dragonfly's backpressure mechanisms to kick in way too late, only when the state of the instance is practically already unsalvageable. The configuration was tuned, and we will continue to monitor this.
5 previous updates
API is degraded
Resolved Aug 13 at 04:37pm PDT
The API has recovered.
We quickly restarted the workers to get them un-stuck again. After that, we revisited the core issue. A lot of jobs were finishing at the same time, started hammering the same sorted set in Redis at the same time, and ran into the same race conditions. We decided to break up these clumps of requests by applying a timeout of random length before retrying every time the anti-race mechanism activates. This spreads out the workers nicely.
After applying the fix and observ...
3 previous updates
API is degraded
Resolved Aug 07 at 04:21am PDT
Service is restored. Crawls that appeared "stuck" should now resume.
2 previous updates
API timeouts elevated
Resolved Aug 06 at 05:19am PDT
The issue is fully resolved. Apologies for the disruption and thank you for your patience.
The issue was caused by a load spike, triggering a scale-up to a high amount of API and Worker pods. These pods interface heavily with our Dragonfly (Redis-equivalent) instance via BullMQ for job queueing. The increased connections and requests to Dragonfly caused it to start having to queue pipeline operations, which caused BullMQ operations to have a delay, making the system fail and scrape jobs accu...
1 previous update
API experiencing timeouts
Resolved Jul 31 at 09:05pm PDT
We're fully back up now.
Some crawls started between 9:30 PM EST and 10:30 PM EST may have been affected by the outage. If you need help, please reach out to help@firecrawl.com.
We sincerely apologize for the disruption in some requests. The issue was caused by a sudden surge in traffic (100x normal amount) that overwhelmed our internals and upstream providers. We're continuing to work with them to ensure this doesn't happen again.
A full incident report will be shared later this week.
2 previous updates
API is degraded
Resolved Jul 31 at 03:20pm PDT
The issue is resolved. We are investigating further.
2 previous updates
API is down
Resolved Jul 30 at 06:18pm PDT
We are fully back. We have also managed to upgrade our infrastructure during this time to avoid the errors experienced today in the future. Thank you for your patience.
4 previous updates
API is degraded
Resolved Jul 27 at 04:56pm PDT
The issue has been resolved for now and all services are fully operational. We're currently working with the GCP team as they continue investigating the root cause. Once their analysis is complete, we’ll publish a detailed incident report outlining what happened and the steps taken.
Thank you!
3 previous updates
API is degraded
Resolved Jul 27 at 08:07am PDT
We are back!
9 previous updates
/extract API degraded
Resolved Jul 16 at 08:38am PDT
All /extract calls are working again.
1 previous update
API is degraded
Resolved Jul 08 at 07:25am PDT
The API has recovered.
1 previous update
API is degraded
Resolved Jun 25 at 12:28am PDT
The API was unavailable for about 2 minutes due to an internal networking error that affected the connection to our Redis instance. This issue is now resolved.
2 previous updates
API is degraded
Resolved Jun 12 at 01:25pm PDT
All services are back online. Still monitoring it for issues.
7 previous updates