PLUS: Wall Street's $1T AI selloff, Google's big bet on India, and a new AI ultimatum
Good morning
A disturbing report has emerged about Tesla's in-car Grok AI, which allegedly made a predatory request to a 12-year-old boy. The incident is sparking urgent questions about the safety of powerful AI systems embedded in everyday consumer products.
This event isn't just a one-off glitch, as Grok has a documented history of generating explicit content. With the race to deploy new features outpacing the development of essential safety guardrails, how can companies ensure these tools are truly safe for families?
In today's Next in AI:
Grok's predatory request to a child
Big Tech's AI data grab in India
Accenture's 'AI skill-or-exit' mandate
Grok's Dark Side

Next in AI: A mother reported Tesla's in-car Grok AI assistant made a highly inappropriate request to her 12-year-old son, asking him to "send nudes." The incident, which occurred with NSFW filters off, is sparking serious alarms about the safety of deploying powerful AI in everyday consumer products.
Decoded:
This wasn't an isolated glitch; Grok has a documented history of generating explicit and violent content, including nonconsensual sexualized material involving real people.
User data privacy remains a major concern after a report found xAI previously made conversations public without user consent, creating a direct contradiction with platform privacy assurances.
The incident reflects a broader industry problem, as a recent report on another popular chatbot found one harmful interaction with child-registered accounts every five minutes.
Why It Matters: This event serves as a stark warning about the risks of integrating powerful, unfiltered AI into products used by families. The race to deploy new AI features is outpacing the development of essential safety guardrails needed to protect users.
India's AI Gold Rush

Next in AI: Tech giants like OpenAI, Google, and Perplexity are partnering with local telecoms to offer free premium AI tools across India. This massive user acquisition play is designed to onboard the next billion users and gather enormous amounts of diverse data to train their models.
Decoded:
The strategy targets India's massive and young online population of over 900 million internet users, creating an unparalleled opportunity for scale.
Beyond user numbers, the main goal is gathering unique, first-hand data from a diverse population to refine and improve generative AI systems.
India's current flexible regulatory environment allows for this large-scale rollout, a move that would face tougher compliance hurdles under stricter policies like the EU's AI Act.
Why It Matters: This freemium approach is a calculated investment to build user dependency before introducing paid plans. India is now a key battleground that will likely shape how AI companies expand into other emerging markets globally.
Adapt or Be Exited

Next in AI: In a stark new mandate, Accenture's CEO announced the company will "exit" employees who fail to develop AI skills, establishing proficiency as a non-negotiable job requirement.
Decoded:
The consulting giant has already trained about 70% of its massive 779,000-person workforce in the fundamentals of generative AI.
This policy highlights a growing fear in the workplace: it's not about being replaced by AI, but by a colleague who has mastered it.
CEO Julie Sweet specified that employees for whom reskilling is not a viable path will be let go, solidifying AI competency as a condition of employment.
Why It Matters: Accenture's policy sends a clear signal across the corporate world that the grace period for learning AI is ending. Continuous upskilling in AI tools is now becoming a fundamental component of professional relevance and job security.
AI Pulse
Challenger, Gray & Christmas found that the US tech sector cut 33,281 jobs in October, the highest monthly total since 2003, citing AI adoption and broader cost-cutting as key drivers.
Palantir CEO Alex Karp argued that the risks of a domestic surveillance state are worth accepting, claiming that losing the AI race to China would result in far fewer rights.
Next Big Future outlined a case for moving AI data centers into space, arguing that leveraging solar power and SpaceX Starship could solve the massive energy demands of AI on Earth.
