South Africa withdraws national AI policy after at least 6 of 67 academic citations found to be AI-generated hallucinations



TL;DR

South Africa’s Communications Minister Solly Malatsi withdrew the country’s draft national AI policy after News24 discovered that at least 6 of its 67 academic citations were AI-generated hallucinations, citing fake articles in real journals. The policy had been approved by Cabinet in March and published for public comment. Malatsi called it an “unacceptable lapse” and promised consequence management. The scandal leaves South Africa without an AI governance framework and raises questions about institutional capacity to regulate the technology.

South Africa’s Department of Communications and Digital Technologies spent months drafting a national artificial intelligence policy. It proposed a National AI Commission, an AI Ethics Board, an AI Regulatory Authority, an AI Ombudsperson, a National AI Safety Institute, and an AI Insurance Superfund. It outlined five pillars of AI governance: skills capacity, responsible governance, ethical and inclusive AI, cultural preservation, and human-centred deployment. It adopted a risk-based approach modelled on the EU AI Act. Cabinet approved the draft on 25 March. The Government Gazette published it on 10 April for public comment. And then News24, the South African news outlet, checked the bibliography and discovered that at least six of the document’s 67 academic citations did not exist. The journals were real. The articles were not. The authors credited with foundational research on AI governance had never written the papers attributed to them. Editors at the South African Journal of Philosophy, AI & Society, and the Journal of Ethics and Social Philosophy independently confirmed to News24 that the cited articles had never been published in their pages. The most plausible explanation, according to Communications Minister Solly Malatsi, is that the drafters used a generative AI tool and published the output without verifying a single reference. A government policy designed to govern artificial intelligence was undermined by the artificial intelligence it failed to govern.

The withdrawal

Malatsi announced the withdrawal on 27 April, calling the fictitious citations an “unacceptable lapse” that “compromised the integrity and credibility of the draft policy.” He said consequence management would follow for those responsible for drafting and quality assurance. “This failure is not a mere technical issue,” the minister said. The parliamentary portfolio committee chair offered a more concise assessment, suggesting the department “skip using ChatGPT this time” when redrafting. The document will be revised before being reissued for public comment, but no timeline has been given. South Africa is now without a formal AI governance framework at a time when governments worldwide are grappling with how to regulate AI, and the country’s credibility as a serious participant in that conversation has taken a blow that will outlast the policy revision.

The scandal is not simply that fake citations appeared in a government document. It is that they appeared in a government document about artificial intelligence, written by the department responsible for the country’s digital technology strategy, during the exact period when the world’s most consequential AI governance debates are being fought in Brussels, Washington, and Beijing. The EU AI Act, the most ambitious regulatory framework for artificial intelligence, is grappling with delayed standards and an implementation timeline that has been pushed back to 2027 for high-risk systems. The United States has no federal AI legislation and is watching states legislate independently while the White House attempts to preempt their efforts. China has enacted AI regulations but applies them selectively. Into this landscape, South Africa offered a policy that could not survive a bibliography check.

The pattern

South Africa’s hallucinated citations are an extreme case of a problem that is quietly spreading across institutions that use generative AI for research and drafting. A study published in Nature found that 2.6 per cent of academic papers published in 2025 contained at least one potentially hallucinated citation, up from 0.3 per cent in 2024. If that rate holds across the roughly seven million scholarly publications from 2025, more than 110,000 papers contain invalid references. GPTZero, a Canadian detection startup, analysed more than 4,000 research papers accepted at NeurIPS 2025, one of the world’s premier AI conferences, and found over 100 hallucinated citations across at least 53 papers. In a separate multi-model study, only 26.5 per cent of AI-generated bibliographic references were entirely correct. The problem is structural: large language models generate citations through probabilistic token prediction rather than information retrieval. They do not look up papers. They predict what a citation should look like based on the patterns in their training data, and when the prediction is confident enough, they produce a reference that reads as authoritative but points to nothing.

The South African case is distinctive not because the technology hallucinated, which is a well-documented and inherent limitation of generative AI, but because the hallucinations were published in an official government policy document that passed through Cabinet approval without anyone verifying the references. The drafting process included civil servants, subject matter consultations, and ministerial review. Dumisani Sondlo, the department’s AI policy lead, had previously described the policy development as “an act of acknowledging that we don’t know enough.” That acknowledgment did not extend to acknowledging that the tool being used to help draft the policy was itself unreliable. The six fake citations that News24 identified are the ones that were caught. Whether additional citations in the document’s 67 references are genuine has not been publicly confirmed. The entire bibliography is now under suspicion, and by extension, so is the analytical foundation on which the policy’s proposals were built.

The implications

The immediate consequence is that South Africa’s AI governance timeline has been reset. The draft policy, which was intended to position the country as a leader in responsible AI adoption on the African continent, will need to be redrafted, reconsulted, and resubmitted. The institutional credibility damage extends beyond the policy itself. If the department responsible for governing AI cannot verify whether the sources in its own policy document are real, the question becomes whether it has the capacity to evaluate the AI systems it proposes to regulate. The policy envisioned a multi-regulator model in which AI governance and human oversight would be embedded within existing supervisory frameworks rather than centralised under a single authority. That model requires each participating regulator to have sufficient technical understanding to assess AI systems in their sector. The hallucination scandal does not inspire confidence that the coordinating department meets that threshold.

The broader lesson is not that governments should avoid using AI in policy development. It is that the failure mode of AI is not dramatic. It does not crash. It does not display an error message. It produces fluent, formatted, confident text that looks exactly like the output of a competent researcher. The fake citations in South Africa’s AI policy were not obviously wrong. They were plausible. They cited real journals. They attributed work to real people. They followed the formatting conventions of academic references. The only way to catch them was to check whether each one actually existed, a task that requires exactly the kind of methodical human verification that AI is supposed to make unnecessary. Growing public distrust of AI is not irrational. It is a response to a technology that is simultaneously powerful enough to draft a national policy and unreliable enough to fabricate the evidence that policy rests on. South Africa’s embarrassment is singular, but the underlying failure, using AI without the capacity to verify its output, is not. It is happening in universities, law firms, newsrooms, and government departments around the world. South Africa is simply the first government to publish the receipts. The challenges of implementing AI regulation are real, but they begin with a prerequisite that South Africa’s department did not meet: understanding what the technology does before trying to write the rules for it.



Source link

Leave a Reply

Subscribe to Our Newsletter

Get our latest articles delivered straight to your inbox. No spam, we promise.

Recent Reviews



In short: Accel has raised $5 billion in new capital, comprising a $4 billion Leaders Fund V and a $650 million sidecar, targeting 20-25 late-stage AI investments at an average cheque size of $200 million. The raise follows standout returns from its Anthropic stake (invested at $183B, now valued near $800B) and Cursor (backed at $9.9B, now reportedly around $50B), and lands in a Q1 2026 venture market that deployed a record $297 billion.

Accel, the venture capital firm behind early bets on Facebook, Slack, and more recently Anthropic and Cursor, has raised $5 billion in new capital aimed squarely at AI. The raise, reported by Bloomberg, comprises $4 billion for its fifth Leaders Fund and a $650 million sidecar vehicle, positioning the firm to write average cheques of around $200 million into late-stage AI companies globally.

The fund lands in a venture capital market that has lost any pretence of restraint. Q1 2026 saw $297 billion flow into startups worldwide, 2.5 times the total from Q4 2025 and the most venture funding ever recorded in a three-month period. Andreessen Horowitz has raised $15 billion. Thrive Capital has closed more than $10 billion. Founders Fund is finishing a $6 billion raise. Accel’s $5 billion is substantial but not exceptional in a market where the biggest funds are measured in the tens of billions.

The portfolio that made the pitch

What distinguishes Accel’s fundraise is the portfolio it can point to. The firm invested in Anthropic during its Series G at a $183 billion valuation. Anthropic has since closed a round at $380 billion and is now attracting offers at roughly $800 billion, meaning Accel’s stake has more than quadrupled in value in a matter of months. Anthropic’s annualised revenue has hit $30 billion, a trajectory that no company in history has matched.

The firm’s bet on Cursor has been similarly well-timed. Accel backed the AI code editor in June 2025 at a $9.9 billion valuation. By November, Cursor had raised again at $29.3 billion. By March 2026, the company was reportedly in discussions at a valuation of around $50 billion. For a developer tool that barely existed two years ago, the appreciation is extraordinary.

Accel’s broader AI portfolio extends beyond these two headline positions. The firm has backed Vercel, the frontend deployment platform; n8n, an AI-powered automation tool; Recraft, a professional design platform; and Code Metal, which builds AI development tools for hardware and defence applications. In March 2026, Accel launched an Atoms AI programme in partnership with Google’s AI Futures Fund, selecting five early-stage companies from what it described as a global applicant pool focused on “white space” opportunities in enterprise AI.

The Leaders Fund model

Accel’s Leaders Fund series is designed for later-stage investments, the kind of large cheques that growth-stage AI companies now require. With an average investment size of $200 million and a target of 20 to 25 deals from the new $4 billion fund, the strategy is concentrated: a small number of high-conviction bets on companies that have already demonstrated product-market fit and are scaling revenue.

This is a different game from traditional venture capital. At $200 million per cheque, Accel is competing less with seed and Series A firms and more with the mega-funds, sovereign wealth funds, and corporate investors that have flooded into late-stage AI. The firm’s argument is that its early-stage relationships and technical evaluation capabilities give it an edge in identifying which companies deserve capital at scale, and in securing allocations in rounds that are massively oversubscribed.

Founded in 1983 by Arthur Patterson and Jim Swartz, Accel built its reputation on what the founders called the “prepared mind” approach, a philosophy of deep sector research before investments materialise. The firm’s most famous prepared-mind bet was its 2005 investment of $12.7 million for 10% of Facebook, a stake worth $6.6 billion at the company’s IPO seven years later. The question now is whether Accel’s AI bets will produce returns of comparable magnitude.

What the market is pricing

The sheer volume of capital flowing into AI venture funds reflects a market consensus that artificial intelligence will be the dominant technology platform of the next decade. The numbers are difficult to overstate. OpenAI raised $120 billion in 2026. Anthropic has raised more than $50 billion. xAI closed $20 billion. Waymo secured $16 billion. These are not venture-scale numbers; they are infrastructure-scale capital deployments that would have been unthinkable outside of telecommunications or energy a decade ago.

For limited partners, the investors who commit capital to venture funds, the logic is straightforward: the returns from AI’s winners will be so large that even paying premium valuations will generate exceptional multiples. Accel’s Anthropic position, where a single investment has appreciated several times over in months, is exactly the kind of outcome that makes LPs willing to commit $5 billion to a single firm’s next fund.

The risk is equally visible. Venture capital is a cyclical business, and the current fundraising boom has the characteristics of a cycle peak: record fund sizes, compressed deployment timelines, and a concentration of capital in a single sector. The last time venture capital raised this aggressively, during the 2021 ZIRP era, many of those investments were marked down significantly within two years. AI’s commercial traction is far stronger than the crypto and fintech bets that defined that earlier cycle, but the valuations being paid today leave little margin for error.

The concentration question

Accel’s fund also highlights a structural shift in venture capital. The industry is bifurcating into a small number of mega-firms that can write cheques of $100 million or more and a long tail of smaller funds that compete for earlier-stage deals. The middle ground, the traditional Series B and C investors, is being squeezed by mega-funds moving downstream and by AI companies that skip traditional funding stages entirely, going from seed round to billion-dollar valuations in 18 months.

For a firm like Accel, which operates across offices in Palo Alto, San Francisco, London, and India, the $5 billion raise is a bet that it can maintain its position in the top tier as fund sizes inflate and competition for the best deals intensifies. Its portfolio of 1,199 companies, 107 unicorns, and 46 IPOs provides a track record. But in a market where Anthropic alone could generate returns that justify an entire fund, the temptation to concentrate bets on a handful of AI winners is strong, and the consequences of getting those bets wrong are correspondingly severe.

The broader picture is that AI venture capital has entered a phase where the funds themselves are becoming as large as the companies they once backed. Accel’s $5 billion raise would have made it one of the most valuable startups in Europe just a few years ago. Now it is table stakes for a firm that wants to participate meaningfully in the rounds that matter. Whether this represents rational capital allocation or the peak of a cycle that will eventually correct is the question that every LP writing a cheque today is, implicitly or explicitly, answering in the affirmative.



Source link