• 5 Patients Completely Cured! China‘s “Genetic Pencil” Rewrites the History of Thalassemia Treatment

    Imagine being told that your child will need a blood transfusion every two to five weeks for the rest of their life. Now imagine receiving a single treatment, and suddenly, the transfusions stop. Forever.

    That is exactly what happened to five patients in China, and the world of medicine is still trying to process it.

    Recently, a global authority in science, the journal Nature, published the results of a groundbreaking clinical trial coming out of Guangxi, China. The numbers are absolutely stunning. Five patients with transfusion-dependent beta-thalassemia, a severe genetic blood disorder that typically requires a lifetime of blood transfusions, have been completely cured. Every single one of them is now transfusion-free.

    This isn‘t just another incremental step in science. This is a giant leap that positions China at the absolute forefront of the global gene editing race.

    The Heavy Reality of Thalassemia

    Let’s rewind for a second. Beta-thalassemia is not a rare curiosity. It‘s one of the most common monogenic diseases on the planet. Every single year, over 40,000 children are born with this condition. For these children, their bodies simply cannot produce enough healthy hemoglobin, the protein in red blood cells that carries oxygen.

    Without treatment, children with severe beta-thalassemia do not survive past early childhood. With treatment, they face a brutal lifelong regimen. They must visit hospitals every few weeks to receive bags of someone else‘s blood just to stay alive.

    That blood, while lifesaving, comes with a heavy price. Every unit of blood brings iron into the body. Over time, that excess iron acts like a poison, slowly destroying the heart, damaging the liver, and wreaking havoc on the endocrine system. So, on top of the transfusions, these patients also need daily injections of drugs to try to flush the iron out. It is exhausting, expensive, and ultimately, just a stopgap.

    There is a theoretical cure: a bone marrow transplant from a healthy donor. But that relies on finding a perfect genetic match. Siblings have only a 25% chance of being a match. For many patients, that donor never comes. And even if they get a transplant, they face the terrifying risk of graft-versus-host disease, where the donor’s cells attack the patient‘s body. It’s a brutal procedure that is simply out of reach for most of the world.

    The “Genetic Pencil” That Changed Everything

    The therapy making headlines is called CS-101. To understand how it works, forget everything you know about genetic scissors.

    Most people have heard of CRISPR. The old generation of genetic tools, like CRISPR-Cas9, are often described as “molecular scissors.” They cut the DNA double helix at a specific spot. The cell‘s natural repair mechanisms then try to fix the cut. It works, but cutting DNA is invasive. Sometimes the repair goes wrong. The cell might delete the wrong chunk of DNA, shuffle genes around, or even trigger cancer-causing pathways. It’s like trying to fix a typo in a book by cutting the entire page out with a knife and hoping you can tape it back together neatly.

    The Chinese team used something entirely different. They used a base editor. Specifically, a homemade tool called a transformer Base Editor.

    This is not a scissor. It is a pencil, or more accurately, an eraser. Instead of breaking the double helix, this editor chemically changes a single “letter” of DNA into another letter. It’s like going into the blueprint of a skyscraper and changing a single typo without demolishing any walls.

    This precision is a game-changer. Because the DNA is never cut, the risk of dangerous side effects, like large genomic deletions or chromosome rearrangements, drops dramatically. The team essentially designed a “lock and key” system where the editor only activates when it finds the exact right spot in the DNA, making it incredibly safe.

    What the Numbers Actually Mean

    So, how well did the “pencil” work? Let‘s look at the data. The researchers took blood stem cells from five patients, used the base editor to fix the genetic glitch, and then gave the cells back to the patients.

    On average, it took just 16 days for the new, healthy stem cells to start rebuilding a functioning blood system.

    Think about that for a moment. Sixteen days. In less than three weeks, the body was producing its own healthy red blood cells without any help.

    In the following weeks, the patients stopped going to the hospital for blood. In three months, their total hemoglobin, the marker we use to measure if someone is anemic, had risen to near-normal levels. In fifteen months, it was stable at perfectly healthy levels. The first patient who received this treatment in October 2023 has now been free of transfusions for more than 28 months.

    The most remarkable part? Safety. Over a median follow-up of nearly two years, there were no serious adverse events related to the treatment. No cancer. No massive DNA deletions. No scary off-target mutations. In a field where scientists have been terrified of causing secondary cancers with gene editing, this is the data we‘ve been praying for.

    The review panel at Nature even said that CS-101 sets a “new high-water mark” for what we can expect from stem cell gene therapy.

    A Global Perspective: Why This Matters for the West

    Now, this is where the story gets really interesting for international audiences.

    There is already a gene therapy on the market for blood diseases. It’s a CRISPR-based drug from the US and Switzerland approved for sickle cell disease. But there is a catch. It is incredibly expensive, costing upwards of 2 million US dollars per patient.

    Beyond the cost, the technology is different. The existing CRISPR drugs still cut the DNA. They work, but they involve a heavy, aggressive chemotherapy regimen to prepare the body, and there are lingering safety concerns about those DNA cuts.

    The Chinese therapy, CS-101, has several distinct advantages. First, because it doesn‘t cut DNA, it’s potentially safer. Second, the early data shows it works faster. The older drugs take about 35 days on average for patients to stop needing transfusions. The Chinese therapy cut that down to 16 days. Faster recovery means less time in the hospital, fewer complications, and lower overall treatment costs.

    Looking Ahead

    The story doesn‘t end with just five Chinese patients. The company behind this, CorrectSequence Therapeutics, has already used this therapy to treat nearly 20 patients suffering from beta-thalassemia and sickle cell anemia globally. They report a 100% success rate so far. Everyone who has received the treatment has regained the ability to make their own blood.

    And here is the kicker: the trial is going global. Patients from Laos, Malaysia, Pakistan, and Nigeria have proactively applied to join the clinical trials. The company is actively recruiting patients worldwide for the next phases of the trial. They are already talking to pharmaceutical companies in the United States and Europe about partnerships.

    For a long time, the West has viewed China as a manufacturer, not an innovator, in biotech. This perception is now dangerously outdated. Shanghai has turned into a powerhouse of biotech innovation. While US and European giants were locked in patent fights over Cas9, Chinese labs quietly developed better tools like the transformer Base Editor.

    This isn‘t just the story of five people getting their lives back. It’s proof that the next generation of medicine will not be invented solely in Boston or London. It‘s being invented in Shanghai and Guangxi.

    The Human Face of the Data

    Behind all these complex scientific terms like “base editing” and “autologous stem cell transplantation,” there are real human beings. Among the first cohort was a teenager who had been receiving blood transfusions since the age of two. Before treatment, they were getting four units of blood every two weeks. After receiving their own edited cells, they walked out of the hospital within a month and have not needed a single blood transfusion since.

    Imagine the life that child now has. No more skipping school for hospital visits. No more watching the clock for the next injection. No more fear of iron poisoning their heart.

    This is the promise of genetic medicine. It‘s not just about living longer. It’s about living better. It‘s about erasing the disease entirely.

    A New Era of Cures

    We are standing at the beginning of a new era. For decades, doctors could only manage the symptoms of genetic diseases. They could buy time, but they couldn’t fix the root cause. Now, with tools like the genetic pencil, we can go into the human operating system and change the code.

    The CS-101 therapy is currently in the process of seeking market approval. If it gets the green light, it will likely be the first base-editing drug available to the public anywhere in the world.

    For the 30 million carriers in China alone, and the hundreds of millions worldwide, this is the hope they’ve been waiting for. The era of the genetic scissor is ending. The era of the genetic pencil has just begun. And thanks to this breakthrough in China, that pencil is writing a much brighter future for patients everywhere.

  • LLM Prices Have Crashed Through the Floor — But Who’s Actually Making Money?

    Back in 2024, the big tech companies were all flexing their muscles over who had the most parameters and the highest benchmark scores. OpenAI rolled out something with a trillion parameters, Google countered with insane context windows. Every AI conversation started and ended with “Scaling Law.”

    Fast forward to May 2026, and the vibe has done a complete 180.

    In a single week in late April, all the major players dropped new models in a tight cluster — OpenAI launched GPT‑5.5, Anthropic updated Claude Opus, Alibaba released Qwen3.6‑Max, Tencent put out Hunyuan Hy3, and Xiaomi shipped MiMo‑V2.5‑Pro. But this time, nobody was bragging about parameter counts. The whole conversation had shifted. As multiple industry insiders put it, AI competition is moving from “how big is your model” to “how efficient and deployable is it, and what can it actually do in the real world.”

    Let me translate that into plain English: Stop telling me how smart or scholarly your AI is. Just tell me — how much does it cost to get a task done? And once I can afford it, can I actually get the job done?

    Alright, let’s look at just how brutal this price war has become.

    Prices have gone completely insane — some are 100x apart, and one player is even raising prices

    Let me throw out some concrete numbers so you can feel what “prices hitting rock bottom” really means.

    On April 24, DeepSeek officially released its V4 model family. The V4‑Flash version was priced at 2 RMB per million output tokens. The V4‑Pro version was set at 6 RMB, but with a limited‑time 75% discount, it came out even lower. Two days later, DeepSeek doubled down: they announced that the “input cache hit” price across the entire API family would be permanently cut to one‑tenth of its previous rate. V4‑Flash cache hits dropped to 0.02 RMB per million tokens, and V4‑Pro, with the promotional discount stacked on top, went as low as 0.025 RMB per million tokens.

    What does that number actually mean?

    0.02 RMB per million tokens. That’s basically two Chinese cents to process the equivalent of several million words.

    Now let’s look at the other side of the table.

    OpenAI’s GPT‑5.5 Pro: 30permillioninputtokens,30permillioninputtokens,180 per million output tokens.
    Anthropic’s Claude Opus 4.7: 5permillioninputtokens,5permillioninputtokens,25 per million output tokens.
    Google’s Gemini 3.1 Pro: 2permillioninputtokens,2permillioninputtokens,12 per million output tokens.

    The gap is staggering. Every single conversation with GPT‑5.5 Pro costs roughly 32 times more than one with DeepSeek V4. If you specifically look at output cost, the difference is even more extreme — GPT‑5.5 Pro’s output side is nearly 100 times more expensive than V4‑Pro’s.

    One developer put it into a painfully vivid analogy: Saoud Rizwan, the founder of Cline, pointed out that if Uber switched its AI services from Claude to DeepSeek V4, the company’s entire 2026 AI budget could be stretched from running out in 4 months to lasting for 7 years. The ironic kicker? On the exact same day, Uber’s CTO confirmed that their annual AI budget had already been burned through — by April.

    And here’s the really baffling part. At the very moment DeepSeek was slashing prices, OpenAI was hiking them. GPT‑5.5’s three-tier pricing — 5input,5input,30 output, 0.5cachehitisadoublingacrosstheboardcomparedtothepreviousGPT5.4.Gobackeightmonths,andGPT5sinputpricewasjust0.5cachehitisadoublingacrosstheboardcomparedtothepreviousGPT‑5.4.Gobackeightmonths,andGPT‑5’sinputpricewasjust1.25. By April 2026, it had jumped fourfold.

    In the same week, two sets of pricing moved in opposite directions, each by orders of magnitude. As one industry analyst put it, “The phrase ‘price war’ doesn’t even cover it anymore.”

    And this still isn’t the bottom. Buried in DeepSeek’s official pricing notes is a tiny line that says: once the Ascend 950 super‑nodes ship in volume in the second half of the year, the Pro version’s price will be cut significantly again. In other words, 0.025 RMB might not even be the floor.

    This isn’t burning cash to buy market share — real tech is driving the cost down

    A lot of people’s first reaction is: at these prices, is DeepSeek just bleeding money to subsidize usage?

    It genuinely isn’t. The price cuts are backed by real architectural breakthroughs.

    Let me walk you through a core concept: the Mixture of Experts (MoE) architecture.

    A traditional large language model is like a single “all‑knowing brain.” Every time you ask it a question, the entire model has to run, no matter what. It’s like forcing a university professor to solve problems ranging from elementary arithmetic to quantum physics, and every single question requires firing up all their brain cells. That’s obscenely expensive.

    The MoE approach is completely different. You take a giant model and break it into many “expert modules,” each specialized in a different domain. Ask a coding question, and the system activates the coding expert. Ask a legal question, and it activates the legal expert. Only a small subset of experts is used for each task; the vast majority stay “asleep.” That slashes the amount of computation needed.

    DeepSeek V4‑Pro has a total of 1.6 trillion parameters, but for any given task, it only activates 490 billion of them. The parts that aren’t activated consume no compute. You pay only for what you actually use.

    On top of this MoE foundation, DeepSeek V4 adds an even deeper layer of optimization. According to the official technical report, when processing a million‑token long context, V4‑Pro needs only 27% of the compute (FLOPs) per token compared to the previous V3.2. The KV cache used to store conversation context has been compressed to just 10% of V3.2’s size. The more extreme V4‑Flash needs only 10% of the compute and 7% of the cache.

    Let me translate that: the previous generation, handling a million tokens, basically required running the model at near‑full tilt while gobbling up a ton of video memory. V4, by overhauling the attention mechanism — using a hybrid sparse attention design called “CSA + HCA” — achieves a scenario where context length grows 8x, while compute consumption drops by over 70%.

    This is why DeepSeek can afford to push prices down to a couple of cents: not because of subsidies, but because the physical cost of a single API call is genuinely an order of magnitude lower than its competitors’.

    Being a hundred times cheaper than OpenAI isn’t a marketing stunt funded by investor cash; it’s architecture‑level innovation that brings the cost down to a point others simply can’t match.

    Now that it’s dirt cheap, who’s actually going to make money first?

    With prices driven to this extreme, one unavoidable question hangs in the air: who is going to profit from this price war?

    First, a dose of cold water: right now, nobody in this industry is truly making money just from selling API access.

    Reports have pointed out that the LLM sector has swung from “price wars” to “price hikes,” with revenue and net losses both climbing — and not a single company is genuinely profitable yet. OpenAI is raising prices, Anthropic is tweaking its billing model, DeepSeek is slashing prices. They look like opposite strategies, but they’re all symptoms of the same underlying problem: no commercially viable path has been proven for large language models.

    One senior engineer from a major cloud provider explained it simply: “A cache hit means the model ‘remembers’ you’ve asked something similar before and can pull the answer from its memory without doing fresh inference, so it’s cheap. A cache miss means the model is seeing the content for the first time and has to compute it from scratch, so it’s expensive.”

    DeepSeek dares to cut the cache‑hit price to one‑tenth because, first, its architecture is so efficient that the base cost of a single token is already very low, and second, they’ve realized that in long‑running tasks, a huge amount of content is repetitive. “System prompts, role definitions, long documents, and tool descriptions” often make up 80–90% of the input, and all of that can be massively reused through caching.

    So, with prices now “too cheap to matter,” who has the best shot at cashing in first?

    The first group: application companies that can ride this low cost to finally push AI into genuine production environments.
    Before, a single AI Agent calling a large model could easily rack up tens of thousands of dollars in cost, which meant it was only viable for demos — never real scale. Now that per‑task costs have plummeted, processing long documents of millions of tokens, or running complex multi‑step agent tasks, has hit a cost threshold that makes real‑world deployment feasible. You no longer have to gut your features or shrink your scope because of budget. These companies don’t carry the enormous expense of training foundational models; they can just call the best model at a ridiculously low cost and pour all their energy and money into building their actual business scenarios.

    The second group: platforms that can build true differentiation on top of this “cheap intelligence.”
    Tokens themselves are becoming a near‑free commodity. The real moat is no longer the model itself, but everything built on top of it. One founder of a top coding agent startup put it bluntly: V4’s tool‑calling stability and hallucination rate still need to be addressed at the engineering level, and real deployment is impossible without the “scaffolding.” Whoever can weave together the model, agent frameworks, industry‑specific data, and real‑world workflows into a reliable production system will own the pricing power.

    The third group, ironically, is the users — businesses and developers.
    PingCAP co‑founder and CTO Huang Dongxu said it directly: “I’m moving my Hermes workflow from Claude Opus and GPT‑5.4 over to DeepSeek V4. Most day‑to‑day work really doesn’t need that ultra‑strong coding capability.” He’s already switched completely. When comparable capabilities converge across platforms, the user’s value is unlocked — you spend one‑tenth or even one‑hundredth of the money and get the same output.

    The least optimistic position belongs to middle‑layer model companies that have neither genuine architectural breakthroughs nor application‑layer competitive advantages, and whose business is simply reselling API access at a margin. When DeepSeek drives margins close to the bone, the room for these players to survive shrinks dramatically.

    Final thoughts

    This price war isn’t really about “whose model is strongest” anymore. It’s about “who can make AI actually get used” — plugged into customer service systems, embedded in coding workflows, running through contract reviews, dropped into every real business scenario that actually exists.

    There’s a well‑known saying in tech: “For any technology to have a deep impact on human society, it must solve the problems of standardization and cost.” AI has finally reached that stage.

    It’s also worth paying attention to that tiny line in DeepSeek’s pricing notes — “Due to current high‑end compute constraints, the Pro service has very limited throughput right now. We expect that once the Ascend 950 super‑nodes ship in volume in the second half of the year, the Pro price will be significantly cut further.” This means DeepSeek is already doing full‑stack adaptation to domestic chips and is using a “chip‑model co‑design” approach to keep lowering the long‑term cost. Huawei has already announced that the entire Ascend super‑node lineup fully supports V4.

    The price anchor for the whole industry, going forward, might no longer be set in Silicon Valley.

    0.025 RMB per million tokens. For that price, you can have the open‑source model with the largest parameter count in the world process the equivalent of three volumes of The Three‑Body Problem.

    When it’s cheap to this degree, the real question is no longer “Can we afford to use it?” but “Once we start using it, how much value can it actually create?”

    If you’ve already started rolling AI into your business, drop a comment and tell us: now that the cost has come down this much, what’s the first thing you’d want AI to actually do for you?