Can AI Write a 100,000-Word Novel? What's the ultimate test for AGI (Artificial General Intelligence)? Calculation? Logic? Or creativity? We tackled the hardest creative challenge: A single AI writing a full-length novel with consistent voice from beginning to end.
๐ Core Innovations
Single Writer System: Not fragmented texts from multiple AIs, but a genuine novel by one author Immediate Critique System: Real-time literary critique and revision for each part 170 Quadrillion Themes: Infinite creative possibilities (4.6 million years at 100 novels/day!) Philosophical Depth: Nobel Prize-level existential exploration and social insight
๐ฒ Infinite Possibilities "The day my father died, I discovered he had another family he'd hidden all his life." One random click generates a powerful opening sentence and a completely new story begins. ๐ Technical Achievements
8,000-word novella auto-generation (approximately 20 minutes) 10 organically structured parts: Perfect narrative arc from introduction to resolution Real-time progress tracking: Session recovery for uninterrupted creation DOCX/TXT export: Korean standard book format (152x225mm) support
๐ Journey Toward AGI This project goes beyond simple text generation. Sustained memory, causal reasoning, emotional nuance, ethical self-censorship, originality - it tests all capabilities required for AGI. Experience it now! Your unique story awaits.
It's been a bit since I took a step back and looked at xet-team progress to migrate Hugging Face from Git LFS to Xet, but every time I do it boggles the mind.
A month ago there were 5,500 users/orgs on Xet with 150K repos and 4PB. Today? ๐ค 700,000 users/orgs ๐ 350,000 repos ๐ 15PB
Meanwhile, our migrations have pushed throughput to numbers that are bonkers. In June, we hit upload speeds of 577Gb/s (crossing 500Gb/s for the first time).
These are hard numbers to put into context, but let's try:
The latest run of the Common Crawl from commoncrawl was 471 TB.
We now have ~32 crawls stored in Xet. At peak upload speed we could move the latest crawl into Xet in about two hours.
We're moving to a new phase in the process, so stay tuned.
This shift in gears means it's also time to roll up our sleeves and look at all the bytes we have and the value we're adding to the community.
I already have some homework from @RichardErkhov to look at the dedupe across their uploads, and I'll be doing the same for other early adopters, big models/datasets, and frequent uploaders (looking at you @bartowski ๐)
Let me know if there's anything you're interested in; happy to dig in!