
One thing has become increasingly clear amid AI’s growing influence across industries and cultures, and that’s the fact that the arts and humanities will remain indispensable. This isn’t just another passionate claim by educators and scholars. It’s a viewpoint that has been echoed by some of the very architects of the AI innovation. Even Daniela Amodei, who co-founded Anthropic (the AI research company behind the Claude chatbot), recently acknowledged that human-centred thinking and interpretation are what will differentiate humans from machines in the decades ahead.
Fortune Magazine quoted Amodei as saying that studyingart and humanities will be “more important than ever” in the age of AI, because uniquely human qualities such as empathy, critical thinking, communication, and curiosity are what machines can never replicate with authenticity. But that’s not all…
Anthropic Can Relate to this truth on a deeper level
That Anthropic feels compelled to stress the importance of the humanities isn’t just philosophical. Recall that the engines of today’s large language models (LLMs) are trained on vast troves of textbooks, many of which were crafted by human authors and are therefore copyrighted. And the impulse to feed AI as many texts as possible has led to legal and ethical pushback.
Anthropic had its fair share of this legal backlash and, in September 2025, felt compelled to reach a $1.5 billion settlement with a class of book authors who alleged that the company improperly used pirated copies of millions of their books to train its Claude models.
That massive settlement, considered the largest copyright recovery in history, illustrates just how much AI firms need human content to teach their machines.
Anthropic is not alone, by the way. Across the tech landscape, more AI companies are facing growing legal scrutiny for their training data practices. Novelists, including Pulitzer Prize winner Michael Chabon and others, have sued OpenAI and Meta for training models like ChatGPT and LLaMA on their copyrighted works without clear permission.
In some cases, courts have made narrow rulings about “fair use” that protect AI firms from liability; in others, lawsuits continue to press forward. Meanwhile, companies like Salesforce have also been hit with claims that they used hundreds of books without permission to build their own AI models.
AI doesn’t extract ideas in an abstract vacuum; it learns from the breadth of human insight contained in literature, essays, history, poetry, philosophy, and more. This is the biggest takeaway from these lawsuits. And it’s high time everyone acknowledges that.
Why the humanities are essential, not optional, even in the age of AI
The AI era’s most enduring value won’t come from how fast a model can process text or coldly optimise language. It will come from how humans interpret meaning, navigate nuance, and bring empathy and ethical judgment to bear on complex problems. This is precisely what Daniela Amodei highlighted in her commentary when she noted that AI will augment humans, not replace the core human capacities developed through humanities study.
Shae O., a Harvard PhD candidate, celebrated Amodei’s remarks in a now viral LinkedIn post and urged educators and students to see critical thinking and humanistic inquiry as “competitive advantages” in a world where AI can mimic syntax but not substance.
What the AI future should really look like
To be clear, AI will continue to accelerate content creation, streamline workflows, and automate rote analytical tasks. But this doesn’t mean an end to human meaning-makers. In fact, the opposite is true. If anything, the current era is reminding us that art, literature, history, ethics, and philosophy are not distractions from progress, but progress itself.
As AI models crunch numbers and parse sentences, they still rely on the output of human culture (from Shakespeare to Toni Morrison, from Cicero to Chimamanda Ngozi Adichie) to stand any chance of imitating human voice and thought. And creating that culture remains uniquely a human responsibility.
Once again, employers know this, which is why across Silicon Valley and elsewhere, they are increasingly seeking workers who can interpret context, ask better questions than machines, and synthesise diverse perspectives. These are skills at the heart of humanities education, the very competencies that distinguish strategic thinkers from mere code operators.
Get the latest news and insights that are shaping the world. Subscribe to Impact Newswire to stay informed and be part of the global conversation.
Got a story to share? Pitch it to us at info@impactnews-wire.com and reach the right audience worldwide
Discover more from Impact Newswire
Subscribe to get the latest posts sent to your email.



