Website Cloner

Can you really *really* clone a website?

Landing Page
Nicholas Cage and his clone in Adaptation (2002)

I don’t want to build this website.

I mean, I can build it. But I don’t have the time right now. Plus someone has built it already, rebuilding it feels wasteful.

So I look for a way to copy it. Cleanly. Quietly. Without making a mess.

I just want to copy this site

When I say I want to copy a website, it’s usually for one of a few very specific reasons.

Sometimes it’s about backing something up. I’ve seen too many useful sites quietly disappear or get redesigned into something worse. I don’t want a screenshot or a PDF. I want a copy that actually works, the way people on r/DataHoarder obsess over.

Redditor asking for help in cloning Farmers Almanac
Redditor asking for help in cloning Farmers Almanac

Other times it’s because I’m trying to validate a side project fast. I don’t want to burn weeks polishing a landing page when I don’t even know if the idea is worth polishing. A bad landing page kills conversion, and then I’m stuck guessing whether the idea is bad or the page is. Starting from something proven feels like removing a variable.

There’s also the Framer problem. I’m tired of using design tools like a designer when I’m not one. Dragging boxes, nudging spacing, squinting at typography. I know what good looks like when I see it. I just don’t want to role-play as a designer to get there.

And sometimes it’s much simpler. I can’t find the right template. I’ve looked. But I’ve already found the right website. The structure makes sense. The flow works. I don’t want inspiration from it. I want that, or something close enough that I can start working instead of searching.

In all of these cases, “cloning” isn’t about copying for the sake of copying. It’s about skipping the parts that feel repetitive and fragile, and getting to the part where real work actually starts.

Let me just download it

HTTrack

Why it’s always recommended first

HTTrack shows up because it feels like it was built for this exact job. It has a UI. It says “website mirroring.” It doesn’t ask you to think too hard. Point it at a URL, hit start, wait.

That promise is comforting. Especially when all I want is a copy.

What works

On simple sites, it does what it says. Static pages come down. Images usually resolve. CSS mostly applies. If the site is old, content-heavy, or barely uses JavaScript, the mirror looks close enough to feel useful.

For pure archiving, it’s fine.

What breaks on modern sites

As soon as JavaScript is involved, things unravel. Client-side routing stops working. Pages load without content. Layouts feel half-assembled. The HTML is there, but the site never finishes becoming itself.

It doesn’t fail loudly. It just quietly gives me less than I expected.

How it’s actually used

Once I start reading how it’s used in places like r/DataHoarder, the mismatch makes sense. HTTrack shines when the goal is preservation. Grab everything. Store it. Maybe never touch it again.

That’s a very different goal from reuse.

wget

Why technical users jump to it

After HTTrack, wget feels like the grown-up option. No UI. Just a command and a lot of control. It carries the promise that if I’m precise enough, I can make this work.

What works

wget is great at fetching exactly what the server exposes. HTML, CSS, JS files, images. Recursive crawls. Custom depth. Fine-grained control over what gets downloaded and what doesn’t.

For static resources, it’s reliable and fast.

What breaks on modern sites

It breaks for the same fundamental reason as HTTrack. wget doesn’t execute JavaScript. It never sees what the browser sees after the page loads. Dynamic routes don’t materialize. Data fetched at runtime is missing. Anything assembled client-side simply isn’t there to download.

The difference is that wget gives me more knobs, but none of them solve that core problem.

How it’s actually used

Once I look at how wget shows up in real workflows, the shape of it makes more sense. It lives inside scripts. In cron jobs. In pipelines that run on a schedule and don’t need supervision.

The goal isn’t to inspect what comes back. It’s to get the same thing every time. Pull the files. Sync the directory. Move on.

That’s a very different goal from cloning something I plan to open, edit, and work with.

Maybe AI can just recreate it?

I drop a screenshot into Lovable. Then Same. Just the first fold. Hero, headline, buttons. “Recreate this.”

It thinks for a bit. Longer than I expect. Then it spits out a page that looks… close.

At first glance, it works. Same structure. Same general shape. Enough to fool me for half a second. Then I look closer.

The hero image is different. The font is wrong. Not just different — wrong. Animations are missing. Colors are approximate guesses.

Of course they are. I only gave it a screenshot. It never had the real assets. It never fetched anything. It couldn’t know what font was actually used, or how the easing felt, or why that image worked and another one didn't.

Sure, I could wire all this up myself. Fix the fonts. Swap the assets. Rebuild the animations. At that point, the screenshot wasn’t the shortcut I thought it was.

It didn’t recreate the page. It recreated the silhouette.

I need something I can actually edit

This is the point where my expectations change.

I’m no longer trying to archive anything. I don’t want a frozen copy I’ll never touch again. I’m also done guessing — done with approximations that look right until I try to change one thing and everything falls apart. I need something usable.

Something I can open in an editor. Read. Modify. Move pieces around without fighting the output. Not a screenshot turned into code. Not a pile of downloaded files that technically load. Something that behaves like a starting point, not an artifact.

That’s when it became obvious that none of the existing tools were built for this moment. They were built for preservation, or fetching, or inspiration. Not for reuse.

So I built Cuttly for myself.

I wasn’t trying to make a better cloner. I was trying to close the gap between seeing a site that works and having something I can work from. Something that respects how modern sites are built, and how developers actually want to edit them.

Cuttly exists because I hit this wall too many times.

TL;DR

I wasn’t trying to copy a website for fun or inspiration. I wanted a real starting point — something editable, usable, and close enough that work could actually begin. The usual tools fall apart here. HTTrack and wget are great for archiving, but modern, JavaScript-heavy sites never fully survive the download. AI tools that recreate pages from screenshots get the outline right but miss the soul: the real assets, the fonts, the motion, the small decisions that make the page work. That gap — between seeing a site that works and having something you can work from — is what none of the existing tools handle. That’s the wall I kept hitting, and that’s why Cuttly exists.

SHARE THIS POST