
Introduction
Some AI systems have become so advanced that they can write code for programmers and even educate them on why they should do the work themselves. Recently, an AI bot for a coding tool named Cursor caused a stir by issuing a false policy statement that misled users.

The Incident
Last Monday, a programmer named BrokenToasterOven encountered a strange issue while using Cursor. He noticed that logging into the tool on one device would immediately log him out on any other device. Initially unsure if he was the only one facing this problem, he posted on Reddit, stating that the user experience had significantly regressed.

To clarify the situation, he emailed Cursor’s customer service. He received a response from a representative named Sam, stating that Cursor’s subscription was designed for use on a single device only, and users needed separate subscriptions for multiple devices. This reply ignited outrage among programmers, as multi-device support is essential for their workflows.

User Reactions
The news prompted many developers to cancel their subscriptions. BrokenToasterOven himself commented that he switched to another service after spending approximately $700 weekly on Cursor. Other users echoed similar sentiments, expressing frustration over the lack of multi-device support and criticizing the decision as foolish.

Cursor’s Response
As discussions intensified, Cursor’s internal team took notice. A developer from Cursor commented that the policy response seemed incorrect and promised to investigate. Three hours later, the Cursor development team clarified that there was no such policy and that users could indeed use Cursor on multiple devices. They apologized for the confusion caused by the AI-generated response and assured users that they were looking into the issue.

Michael Truell, Cursor’s co-founder, also apologized on Hacker News, acknowledging the significant issue and outlining steps being taken to prevent future occurrences. They promised to mark all AI-generated replies clearly and issued a full refund to the affected user.


Broader Implications
This incident highlights a recurring problem with AI systems: when they provide incorrect information, it can lead to serious consequences. A similar incident occurred in February 2024, when a user was misinformed by an airline’s chatbot regarding bereavement fare policies, leading to a legal dispute.

The takeaway is clear: while AI can enhance efficiency, businesses must ensure that users are aware they are interacting with AI and not a human. Clear communication and oversight are essential to prevent misinformation and potential legal issues.

In conclusion, this incident serves as a stark reminder that AI, despite its capabilities, should not replace human oversight in critical areas of communication and customer service.
Comments
Discussion is powered by Giscus (GitHub Discussions). Add
repo,repoID,category, andcategoryIDunder[params.comments.giscus]inhugo.tomlusing the values from the Giscus setup tool.