Wednesday, February 11, 2026

Beware the legal risks of AI meeting agents


AI meeting agents are everywhere. They join Zoom calls, transcribe conversations, summarize action items, and promise to save employees hours of note-taking. From a business perspective, the upside is obvious: better documentation, fewer "I don't remember saying that" disputes, and cleaner follow-up.

But like most shiny tech, AI meeting agents come with real employment law and litigation risk—especially if you don't think through how (and when) you use them.

Start with wiretapping laws. Federal law is one-party consent, but many states are not. CA, CT, FL, IL, MD, MA, MI, MT, NV, NH, PA, and WA require all parties to consent to a recording. An AI agent that silently records or transcribes a meeting can easily violate those statutes. That's not just a technical foul. It can mean statutory damages, attorneys' fees, and—yes—class actions.

Then there's employee relations. If employees learn that every meeting might be recorded, transcribed, and stored indefinitely, candor dies fast. Performance conversations become stilted. Complaints may never get voiced. And if the AI summary is wrong, unreviewed, and uncorrected, congratulations—you've just created inaccurate evidence that will be blown up on a screen in front of a jury.

The real nightmare scenario, though, is forgetting to turn it off.

The meeting ends. People start chatting. Someone vents. Someone jokes. Someone says something they absolutely would not want memorialized in writing. The AI agent is still listening. Now you've captured informal comments that were never meant to be "on the record," and you've handed someone a litigation bombshell.

Bottom line: AI meeting agents can be useful—but only with clear policies, upfront consent, disciplined controls, and training that treats them like recording devices, not harmless assistants. Because when they go wrong, they don't just go wrong. They go off the rails.