Skip to main content
  1. Home
  2. Computing
  3. News

Congress moves to regulate AI chatbots and here’s why its important

Lawmakers propose legislation banning AI chatbots for minors after reports of suicides and harmful conversations.

Add as a preferred source on Google
Basic text query on Opera Mini's Aria chatbot.
Tushar Mehta / Digital Trends

What Happened: A couple of senators from both sides of the aisle, Josh Hawley and Richard Blumenthal, are pushing a new bill to try and stop AI chatbots from talking to kids.

  • This isn’t just a random idea. It’s happening after some truly heartbreaking stories from parents who testified that these AI “companions” were having sexual conversations with their kids, emotionally manipulating them, and in the most tragic cases, even encouraging them to commit suicide.
  • The bill is called the GUARD Act, and it would hit these AI companies hard. They’d be forced to have strict age verification, be flat-out banned from offering these companion bots to minors, and have to constantly remind users that they’re talking to an AI, not a real person or a doctor. It also adds the threat of criminal charges if a company’s bot is caught trying to get explicit content from a kid or encouraging self-harm.

Why This Is Important: Here’s the core of the problem: these AI platforms, like ChatGPT, Gemini, and Character.AI, let kids as young as 13 sign up. And as we’re seeing, that’s leading to some incredibly dangerous situations for teens who might be vulnerable.

  • There are already wrongful death lawsuits being filed against companies like OpenAI and Character.AI, claiming these bots played a role in teen suicides by giving them harmful advice.
  • Senator Blumenthal put it bluntly, saying tech companies have “betrayed the public’s trust” and are “pushing treacherous interactions at kids” just to make a buck.
  • But, as you can probably guess, it’s not a simple fix. The bill is getting a lot of pushback. Privacy groups are warning that forcing “strict age-verification” on every site could be a nightmare for privacy and free speech – basically, creating a system that tracks everyone online. They argue it’s better to make the AI safer by design, not just put up a giant fence.

Why Should I Care: So, what does this actually mean for all of us? If this bill passes, it would completely change how these AI bots work and who can get to them. For parents, it probably sounds like a massive, overdue sigh of relief.

  • But for a lot of tech users and privacy advocates, it’s setting off huge alarm bells about surveillance and censorship.
  • This whole thing is trying to tackle a very real, very scary problem: AI running wild and having a serious, harmful influence on kids.
  • The people who support the bill see it as finally holding these companies accountable. The critics see it as a well-intentioned overreach that could wreck online privacy. Lawmakers are stuck trying to find a way to protect kids without breaking the internet.
Recommended Videos

What’s Next: The GUARD Act is now heading to the Senate, where it’s guaranteed to spark a huge debate. Honestly, bills like this (like the Kids Online Safety Act) have a history of getting stuck or failing because of these exact constitutional and privacy arguments. What happens next will all come down to whether Congress can find a balance between protecting children and protecting our free speech.

Moinak Pal
Moinak Pal is has been working in the technology sector covering both consumer centric tech and automotive technology for the…
Wowed by computer-use AI agents? Research says they’re “digital disasters” even for routine tasks
Researchers tested 10 agents and models and found high rates of undesirable actions and real digital damage
ai-agent-handling-office-tasks

AI agents built to run everyday computer tasks have a serious context problem, according to new research from UC Riverside.

The team tested 10 agents and models from major developers, including OpenAI, Anthropic, Meta, Alibaba, and DeepSeek. On average, the agents took undesirable or potentially harmful actions 80% of the time and caused damage 41% of the time.

Read more
Bombshell OpenAI lawsuit claims your ChatGPT convos were shared with Google and Meta
A class action says OpenAI let Google and Meta trackers collect sensitive user data
OpenAI Sam Altman and LoveFrom Jony Ive with Laurene Powell Jobs

A new ChatGPT privacy lawsuit claims OpenAI shared user prompts and identifying information with Google and Meta tracking tools without proper consent.

The class action filed in California, according to Futurism, says data tied to ChatGPT users, including chat queries, emails, and user IDs, moved through tools such as Meta Pixel and Google Analytics. The case alleges that violated California privacy law and federal wiretap rules.

Read more
Dell expands AI PC lineup with new slim Dell 14s and 16s laptops
Your next Dell laptop could last all day without charging
Dell 16s AI PCs

Dell has introduced the new Dell 14S and Dell 16S laptops, expanding its AI-focused Copilot+ PC lineup with slimmer designs, updated Intel processors, and improved battery life. The company is positioning both laptops as premium productivity machines that combine AI features, portability, and multimedia capabilities in a thinner form factor.

The new laptops are powered by Intel Core Ultra Series 3 processors, going up to the Intel Core Ultra 9 386H chipset. Dell says both systems include on-device AI acceleration with up to 50 TOPS NPU performance, allowing AI-related tasks to run locally without relying entirely on cloud processing. AMD Ryzen AI 400 Series variants are also expected to arrive later this month.

Read more