• Home
  • About Me
  • Blog
  • Dev Journal
  • Resources

HomeAbout MeBlogDev JournalResources
TechExperienceRSSGiving What We Can
GitHubTwitter / XLinkedIn

Web Scrapers Maxing Out My Vercel Fast Data Transfer

The Context

I use Vercel to deploy my website. For small websites, like mine, it’s low cost, reliable, and hassle free.

The Problem

I received an email from Vercel notifying me that I had used over 80% of my Fast Data Transfer allowance on the free tier. This represents outgoing data transfer from Vercel’s CDN to readers of the website.
I checked my analytics dashboard and this represented 100k requests–y site isn’t receiving that much traffic.
After some digging, I found it was Google and OpenAI indexing my site (presumably to train their LLMs on).

The Solution

I tried adding a Crawl-Delay header of 30s to my robots.txt and it seems that the number of incoming requests has dropped significantly.
I’m not sure if this was correlated to the change I made, but I will keep an eye on it going forwards - so far it looks good!
notion image

The Code

// robots.ts import { MetadataRoute } from "next"; import { DOMAIN } from "./constants"; export default function robots(): MetadataRoute.Robots { return { rules: { userAgent: "*", allow: "/", crawlDelay: 30, // tell the robots to leave me alone! }, sitemap: `${DOMAIN}/sitemap.xml`, }; }