
Prefer to listen instead? Here’s the podcast version of this article.
In the fast‑moving world of artificial intelligence (AI), an extreme trend is emerging: elite teams at top AI labs are working 80 to 100 hours per week to stay ahead. This “all‑in” sprint reflects the intensity of the global AI arms race, and it raises major questions about labor, ethics, culture and the future of work.
According to recent reporting, researchers and executives at leading AI firms such as Anthropic, OpenAI, Google DeepMind, Meta Platforms and others are frequently working 80–100 hour weeks in pursuit of “superhuman” AI capabilities. One researcher described it as trying to “speed‑run 20 years of scientific progress in two years.” [mint] Start‑up and big‐tech culture alike are embracing this pace: one piece describes “0‑0‑2” scheduling (midnight‑to‑midnight with a two‑hour break) in some labs. [The Wall Street Journal]
This intense pace is driven by several factors:
Working 80+ hours per week is not sustainable for most people. The culture of ultra‑long hours raises concerns about burnout, mental health, work‑life balance, and whether innovation under extreme stress yields quality outcomes. A recent article described that many AI workers involved in core model development have “no time for friends or hobbies”. [Hindustan Times]
Pushing to run decades of progress in a short time may accelerate change—but it also brings risks. Cutting corners, fatigue, oversight failures or flawed model behaviours may increase. In the context of AI with high stakes (safety, bias, societal impact), this matters especially. For example, the industry is already navigating blurring lines between productivity and pressure: some firms insist on 72‑hour workweeks as part of the competitive push. [The Washington Post]
As firms push faster, regulators and ethicists are warning that AI deployment must not outstrip safety, alignment and oversight. The arms race mentality can conflict with careful governance—so stakeholders must ask: is pushing hard good if oversight is weak? On the labour side, there’s also a risk of making a few highly specialised workers bear the burden while many others are left behind. A recent piece on the “AI arms race” framed this as more than technology—it’s a mobilisation of human labour, material and energy globally.
Paradoxically, while elite AI researchers are working harder than ever, other parts of the workforce are facing job disruption due to AI and automation. Reports suggest as many as 100 million U.S. jobs could be affected in the next decade. So we have a two‑speed world: a handful of ultra‑intensive roles powering AI breakthroughs, and a large number of jobs under pressure from those breakthroughs. Interestingly, some tech leaders envision that AI will eventually reduce workweeks (e.g., 4‑day or even 3‑day workweeks) — but that’s not where we are right now.
For business leaders, talent managers and researchers, several key take‑aways emerge:
The 100-hour workweeks now defining parts of the AI industry are more than a reflection of ambition—they’re a signal of a high-stakes, high-pressure race where time is currency and breakthroughs are weapons. But while the pace may seem exhilarating, it’s also exhausting, unsustainable, and potentially dangerous without the right checks and balances.
As we continue to push the boundaries of what AI can do, we must also redefine what sustainable innovation looks like. Burnout, ethical lapses, and disproportionate pressures on a select few are not the ingredients for long-term success. Instead, the future belongs to organizations and professionals who can combine speed with responsibility, ambition with empathy, and innovation with foresight.
The arms race may be on—but it’s not just about who gets there first. It’s about who gets there in a way that’s sustainable, ethical, and beneficial for all.
WEBINAR