Microsoft on Monday stated it took steps to appropriate a obtrusive safety gaffe that led to the publicity of 38 terabytes of personal knowledge.
The leak was found on the corporate’s AI GitHub repository and is claimed to have been inadvertently made public when publishing a bucket of open-source coaching knowledge, Wiz stated. It additionally included a disk backup of two former workers’ workstations containing secrets and techniques, keys, passwords, and over 30,000 inside Groups messages.
The repository, named “robust-models-transfer,” is not accessible. Previous to its takedown, it featured supply code and machine studying fashions pertaining to a 2020 analysis paper titled “Do Adversarially Sturdy ImageNet Fashions Switch Higher?”
“The publicity got here as the results of a very permissive SAS token – an Azure characteristic that enables customers to share knowledge in a way that’s each onerous to trace and onerous to revoke,” Wiz stated in a report. The problem was reported to Microsoft on June 22, 2023.

Particularly, the repository’s README.md file instructed builders to obtain the fashions from an Azure Storage URL that by accident additionally granted entry to your complete storage account, thereby exposing further non-public knowledge.
“Along with the overly permissive entry scope, the token was additionally misconfigured to permit “full management” permissions as a substitute of read-only,” Wiz researchers Hillai Ben-Sasson and Ronny Greenberg stated. “Which means, not solely may an attacker view all of the information within the storage account, however they might delete and overwrite current information as properly.”
In response to the findings, Microsoft stated its investigation discovered no proof of unauthorized publicity of buyer knowledge and that “no different inside providers have been put in danger due to this concern.” It additionally emphasised that clients needn’t take any motion on their half.
The Home windows makers additional famous that it revoked the SAS token and blocked all exterior entry to the storage account. The issue was resolved two after accountable disclosure.
To mitigate such dangers going ahead, the corporate has expanded its secret scanning service to incorporate any SAS token that will have overly permissive expirations or privileges. It stated it additionally recognized a bug in its scanning system that flagged the precise SAS URL within the repository as a false constructive.
“As a result of lack of safety and governance over Account SAS tokens, they need to be thought-about as delicate because the account key itself,” the researchers stated. “Due to this fact, it’s extremely really useful to keep away from utilizing Account SAS for exterior sharing. Token creation errors can simply go unnoticed and expose delicate knowledge.”
Identification is the New Endpoint: Mastering SaaS Safety within the Fashionable Age
Dive deep into the way forward for SaaS safety with Maor Bin, CEO of Adaptive Protect. Uncover why id is the brand new endpoint. Safe your spot now.
This isn’t the primary time misconfigured Azure storage accounts have come to mild. In July 2022, JUMPSEC Labs highlighted a state of affairs through which a risk actor may reap the benefits of such accounts to realize entry to an enterprise on-premise setting.
The event is the most recent safety blunder at Microsoft and comes almost two weeks after the corporate revealed that hackers primarily based in China have been in a position to infiltrate the corporate’s programs and steal a extremely delicate signing key by compromising an engineer’s company account and sure accessing an crash dump of the buyer signing system.
“AI unlocks large potential for tech corporations. Nevertheless, as knowledge scientists and engineers race to deliver new AI options to manufacturing, the large quantities of information they deal with require further safety checks and safeguards,” Wiz CTO and co-founder Ami Luttwak stated in a press release.
“This rising know-how requires giant units of information to coach on. With many improvement groups needing to govern large quantities of information, share it with their friends or collaborate on public open-source tasks, circumstances like Microsoft’s are more and more onerous to observe and keep away from.”