Skip to content

Copilot exposes private -github pages, some removed by Microsoft

    Screenshot that shows Copilot continues to serve tools Microsoft has taken action to be removed from Github.


    Credit: Lasso

    Lasso ultimately determined that the Microsoft solution was broken down to the public access to a special Bing user interface, once available on cc.bingj.com. However, the solution did not seem to remove the private pages of the cache itself. As a result, the private information was still accessible to Copilot, which in turn would be available for the Copilot user who asked.

    The Lasso researchers explained:

    Although Bing's Cached Link function was eliminated, pages continued to appear in the cache in search results. This indicated that the solution was a temporary patch and although the public access was blocked, the underlying data was not completely deleted.

    When we visited our research into Microsoft Copilot, our suspicions were confirmed: Copilot still had access to the cache data that were no longer available for human users. In short, the solution was only partially, human users were prevented from collecting the cache data, but Copilot still had access to it.

    The post explained simple steps that everyone can take to find and view the same huge series of private repositories that Lasso has identified.

    There is no toothpaste back in the tube

    Developers regularly enclose security vessels, private coding keys and other sensitive information directly into their code, despite best practices that have long since called to be entered such data by safer agent. This potential damage deteriorates when this code is made available in public repositories, another common security that fails. The phenomenon has occurred for more than a decade time and time again.

    When these types of mistakes take place, developers often make the repositories quickly private, hoping to contain the fallout. The findings of Lasso show that simply making the Privé Privé code is not enough. Once exposed, references are irreparably affected. The only story is to rotate all references.

    This advice is still not on the problems that result when other sensitive data is included in repositories that have been switched from the public to private. Microsoft has incurred legal costs to have removed tools from Github after they claimed that they have violated a series of laws, including the Computer Fraud and Abuse Act, the Digital Millennium Copyright Act, the Lanham Act and the Racketeer and Corrupt Organizations Act. Business lawyers prevailed to remove the tools. To date, Copilot continues to undermine this work by making the tools available.

    In an e -mail statement sent after this message, Microsoft wrote: “It is generally understood that large language models are often trained in publicly available information from the internet. If users prefer to avoid their content to train these models, they are encouraged to keep their repositories private at all times.”