The next leap in vehicle automation — so-called “Level 3” or conditional autonomy — promises to let drivers take their hands and eyes off the wheel, yet it raises profound questions about liability and regulation. According to a recent article in The Verge, major automakers such as General Motors are racing to bring eyes-off driving systems to market by 2028. The article points out that under such systems, drivers must “stand ready to take over” when alerted, and failing to do so may land them in legal hot water. At the same time, legal experts at firms like Clifford Law Offices emphasize that U.S. law is still catching up: while the driver has traditionally borne liability for crashes, the presence of automation muddies the waters and leaves manufacturers, drivers, and insurers in a regulatory limbo. A study by the Insurance Institute for Highway Safety flags a key vulnerability: humans out-of-the-loop for extended stretches—even when automation is engaged—may not be able to appropriately retake control in an emergency. Thus we may be headed into a phase where accidents involving semi-autonomous systems trigger uncharted liability issues across manufacturers, regulators and consumers.
Key Takeaways
– As the industry approaches Level 3 automation, the liability framework is murky: who pays when a “hands-off, eyes-off” system fails—the driver, the automaker, or both?
– Law firms and insurers warn that until new statutes or interpretations evolve, drivers remain at risk of being held responsible even when a vehicle is driving itself under certain conditions.
– The human-machine handoff remains the weakest link: research shows that when a driver’s monitoring is reduced, the ability to retake control safely in emergencies declines significantly.
In-Depth
We’re standing on the cusp of a major shift in automotive technology: the industry is steadily moving beyond driver-assist systems that require eyes on the road toward systems that relax that requirement entirely. For example, General Motors recently revealed plans for an “eyes-off” driving mode slated for U.S. release by 2028, beginning with the Cadillac Escalade IQ. Under such Level 3 systems, drivers might watch a movie or check their phone while the car handles itself—until the system prompts them to take back control. But while the technology is advancing fast, the legal and regulatory infrastructure is not keeping pace, and that lag presents risks for consumers and manufacturers alike.
The first major concern is liability: historically, traffic accident law places the burden of fault and damages on the human driver. But with automation, the driver may not be actively driving when the crash occurs. Legal commentary (such as from the Clifford Law Offices) points out that Level 3 autonomy places the vehicle in control under limited conditions while requiring the driver to remain available—but leaves undefined who is ultimately liable when things go wrong: the driver for failing to intervene, the manufacturer for a system failure, or even the fleet operator in commercial cases.
Second, insurers and safety researchers are raising red flags about the “out-of-the-loop” problem. A report by the Insurance Institute for Highway Safety notes that when drivers disengage from active monitoring for periods—relying on the car’s sensors instead—their ability to reacquire situational awareness when the system demands takeover declines. In other words: the system may perform well until one moment of surprise mandates driver intervention—and at that critical moment the driver may not be ready. That scenario presents not only a safety hazard but a liability nightmare: if the driver failed to respond, is the fault theirs? If the system failed to warn appropriately, is the manufacturer liable?
Meanwhile, the regulatory environment remains fragmented. The National Highway Traffic Safety Administration (NHTSA) outlines Level 3 as “conditional automation … the system actively performs driving tasks while the driver remains available to take over,” but also notes that such systems are not yet widely available. At state and federal levels, statutes and regulation struggle to define how traditional liability doctrines—negligence, product liability, and strict liability—apply to this hybrid driver-machine setup.
From a conservative perspective, these developments underscore the need for clarity and caution: innovation is laudable, but when lives are at stake and legal precedent limited, consumers deserve strong protections. Automakers in pursuit of market advantage should not outpace the legal frameworks that allocate responsibility. Insurance companies must prepare for the transition from “driver-driven” accidents to “shared-responsibility” accidents with one foot in the car and one foot in the machine. And drivers considering vehicles with such “hands-off/eyes-off” modes should be aware that the technology may work—but the law might not yet protect them.
In effect, we’re entering a middle era of driving characterized by hybrid control: the car drives, but the human stays—or at least is supposed to stay—ready to act. That halfway state is inherently unstable and ripe for confusion. When a crash happens, juries and courts will have to parse telematics, sensor logs, driver attentiveness, system warnings and handoff timing to decide who was responsible. In many ways, this mirrors other transitional technologies (think aviation autopilot, chemical plant controls) where the human is supervisor rather than pilot. Until legislation, regulation and judicial precedent catch up—and include clear rules for manufacturers, drivers and insurers—those taking the wheel of a Level 3-capable car may well be navigating more than just the road ahead: they’ll be navigating legal ambiguity.

