In the realm of IT security, the practice known as red teaming -- where a company's security personnel play the attacker to test system defenses -- has always been a challenging and resource-intensive ...
Microsoft has open sourced a key piece of its AI security, offering a toolkit that links data sets to targets and scores results, in the cloud or with small language models. At the heart of ...
Microsoft has introduced a new framework called PyRIT (Python Risk Identification Toolkit for generative AI) for the automation of red teaming processes or finding risks in generative AI systems, ...