How cybersecurity events can combat cyber poverty
What is cyber poverty, and why do cyber inequities affect all organisations and industries? Learn how cybersecurity practitioners can work together to close the cyber poverty gap.
Read MoreCan you stop AI models from using and repurposing your digital content?
The short answer is yes – sort of.
The reality of digital life today is that anything you’ve ever published online has probably been ingested by AI, and perhaps regurgitated by now. LLMs and AI image creators leverage immense volumes of user data from across the internet; and as new and more advanced generative AI systems become commercially available, their appetite for data is only going to grow.
Companies and digital users need to learn how AI uses their content, and what they can do to prevent it from being hoovered up by hungry AI models.
The internet has already been scraped by AI developers. It’s highly likely that anything you’ve posted in the past has already been used in AI training.
Going forward though, it’s worth understanding that while a growing number of AI companies do offer an opt-out option to allow you to say no to your data being used in AI training, many of them make it very difficult to do that.
Tech giants including Meta, Google, and X have included clauses in their privacy policies that say they may use your data to train their AI models. And once data has been used, it’s very difficult to get it removed from the pools of data the AI learns from.
Although their opt-out processes may be unclear, a growing number of companies do enable users to keep your content out of AI training data.
As explained in detail by Wired, those companies include:
As coherent AI governance emerges from this period of rapid growth, it’s likely that more companies will either choose, or be required, to invite users to opt in if they’re OK with their content being used in AI training data – rather than automatically opting them in unless the user actively opts out.
Companies already doing this include Anthropic, with its ‘frontier AI models backed by uncompromising integrity.’ Anthropic leverages user information for training their AI model, Claude, only when the user explicitly grants permission to do so.
Ultimately, the future of data privacy in an AI-powered digital landscape remains uncertain. At this point in time, awareness is crucial: because if you know your content is being used, and how it’s being used, you can make informed decisions to protect it.
Join us at Black Hat MEA 2024 and discover how to improve your organisation’s cyber resilience.
Join the newsletter to receive the latest updates in your inbox.
What is cyber poverty, and why do cyber inequities affect all organisations and industries? Learn how cybersecurity practitioners can work together to close the cyber poverty gap.
Read MoreFind out what retail industry leaders should know about cybersecurity and imminent threats during a surge in B2C sales.
Read MoreCybersecurity offers diverse job opportunities for professionals with a wide range of skills and experience. Discover three specialist roles in cybersecurity.
Read More