Microsoft AI team accidentally leaks 38 terabytes of sensitive data

According to a recent report by cloud security firm Wiz, Microsoft's AI research department inadvertently leaked 38 terabytes of sensitive data while contributing open source AI learning models to a public GitHub repository.

This data breach highlights the complexity of managing and protecting large amounts of data in an open source environment. The Microsoft Research team contributed to the breach by failing to appropriately handle sensitive information contained in the dataset when contributing its open source AI learning models.

Wiz's report states that the 38 terabytes of data leaked may contain a large amount of Microsoft's internal project data, including and not limited to information such as source code, system credentials, and sensitive API keys. This information, if it falls into the hands of malicious hackers, could pose a serious threat to Microsoft's internal systems.

Currently, Microsoft has confirmed the data breach and said that they have taken action to clean up the leaked data. At the same time, Microsoft also emphasized that they will conduct a comprehensive internal review of the incident and take necessary measures to prevent similar incidents from happening again.

This incident reminds us once again that even leading global technology companies can be negligent in data management and protection. For both individuals and organizations, it is important to be fully aware of the importance of data security and take appropriate precautions to reduce the risk of potential data breaches.

Previous:

Next:

Leave a Reply

Please Login to Comment