Soulcode #2: Want to profit from AI companies using your work? Two legal documents from the UK and the EU point to that direction
The last few weeks in AI (from someone with a soul)
In a world of algorithms and automation, we need a human compass. This newsletter brings you AI and tech news through the lens of ethics and humanity, where code meets consciousness.
We'll explore AI's biggest developments and their impact on our shared future every two weeks. No jargon, no hype – just clear insights with a conscience.
Hey there!
What a couple of weeks! A lot has happened in the AI world, and tech giants' latest releases (from Google, Google again, OpenAI, Amazon, Meta, … etc.) are already littering social media feeds.
I am not gonna cover them for two reasons:
I’m sure you may have already heard of them (if not, I just gave you links to the announcements)
These are incremental improvements. Not incredible or noteworthy novel tools or discoveries.
INSTEAD, two massive movements just dropped in AI regulation that could affect creators’ careers and wallets. Let me break them down in plain English.
🇬🇧 UK's New AI Copyright Plan
The UK government just proposed new rules about how AI companies can use creative work. Here's what it means:
1. You Could Get Paid When AI Uses Your Work
AI companies might need to pay to use your content for training
This includes your writing, photos, code - basically anything you create
You'll be able to say "yes" or "no" to AI using your work
2. More Transparency
AI companies will have to tell us what content they're using
No more mystery about what data trains these AI systems
Why This Matters:
New money-making opportunity: earn from AI companies using your work
Better protection for your creative content
Clearer rules for everyone
🇪🇺 EU's New AI Privacy Rules
The European Data Protection Board (EDPB) just released a much anticipated Opinion on AI and personal data protection. The simple version:
1. Privacy First
Companies can't just say "our AI is anonymous" - they need to prove it
Better protection for your personal information
2. Stricter Rules
AI companies need good reasons to use personal data
If they break the rules, they might have to retrain or even delete their AI
3. Real Consequences
Companies using data illegally could face serious penalties
This means more responsible AI development
What This Means For You:
👩💼 If You Work in Tech:
Clearer rules for building AI
Less legal uncertainty
New opportunities in responsible AI development
✍️ If You Create Content:
Potential new income stream from AI training
More control over how AI uses your work
Better protection for your creations
👥 For Everyone:
Better protection for your personal data
More transparency about how AI works
More control over your content and information
What's Next?
The UK is taking feedback until February 25, 2025. These changes could set the standard for how AI and creativity work together worldwide.
The EDPB's opinion points out that when companies want to use your information because they have a good reason (“legitimate interests,” in legal jargon), they need to pay special attention to the unique risks that AI brings. These risks include unfair treatment of individuals, the potential for generative AI to reuse personal information and broader societal dangers like deepfakes and misinformation. The opinion suggests ways to reduce these risks, which could help data controllers, such as using pseudonymisation (hiding identities), adding filters to outputs, and being transparent with initiatives like model cards (think of them as nutritional labels for AI algorithms) and yearly reports.
This is important for developers: if they don’t comply with these guidelines during the development stage, their AI system could be deemed non-compliant, leading to legal issues and operational problems.
Quick Takeaway:
We're moving toward a future where AI companies must be more transparent and fair to creators and users. Understanding these changes now is a good idea —they might just create new career opportunities.
Stay curious!
—a
P.S. What do you think? Should AI companies pay to use creative work for training? Hit reply - I'd love to hear your thoughts!