I admit that I am biased enough to immediately expect the AI agreement to be exactly what we need right now if this is how Meta reacts to it. Which I know is stupid because I genuinely have no idea what is in it.
If I'd were to guess Meta is going to have a problem with chapter 2 of "AI Code of Practice" because it deals with copyright law, and probably conflicts with their (and others approach) of ripping text out of copyrighted material (is it clear yet if it can be called fair use?)
Edit: from the linked in post, Meta is concerned about the growth of European companies:
"We share concerns raised by these businesses that this over-reach will throttle the development and deployment of frontier AI models in Europe, and stunt European companies looking to build businesses on top of them."
Of course. Skimming over the AI Code of Practice, there is nothing particularly unexpected or qualifying as “overreach”. Of course, to be compliant, model providers can’t be shady which perhaps conflicts with Meta’s general way of work.
The problem is this severely harms the ability to release opens weights models, and only leaves the average person with options that aren't good for privacy.
Nope. This text is embedded in HN and will survive rather better than the prompt or the search result, both of which are non-reproducible. It may bear no relation to reality but at least it won't abruptly disappear.
Not just Meta, 40 EU companies urged EU to postpone roll out of the ai act by two years due to it's unclear nature. This code of practice is voluntary and goes beyond what is in the act itself. EU published it in a way to say that there would be less scrutiny if you voluntarily sign up for this code of practice. Meta would anyway face scrutiny on all ends, so does not seem to a plausible case to sign something voluntary.
One of the key aspects of the act is how a model provider is responsible if the downstream partners misuse it in any way. For open source, it's a very hard requirement[1].
> GPAI model providers need to establish reasonable copyright measures to mitigate the risk that a downstream system or application into which a model is integrated generates copyright-infringing outputs, including through avoiding overfitting of their GPAI model. Where a GPAI model is provided to another entity, providers are encouraged to make the conclusion or validity of the contractual provision of the model dependent upon a promise of that entity to take appropriate measures to avoid the repeated generation of output that is identical or recognisably similar to protected works.
Kaplan's LinkedIn post says absolutely nothing about what is objectionable about the policy. I'm inclined to think "growth-stunting" could mean anything as tame as mandating user opt-in for new features as opposed to the "opt-out" that's popular among US companies.
So then it's something completely worthless in the globally competitive cutthroat business world, that even the companies who signed won't follow, they just signed it for virtue signaling.
If you want companies to actually follow a rule, you make it a law and you send their CEOs to jail when they break it.
"Voluntary codes of conduct" have less value in the business world than toilet paper. Zuck was just tired of this performative bullshit and said the quiet part out loud.
No, it's a voluntary code of conduct so AI providers can start implementing changes before the conduct becomes a legal requirement, and so the code itself can be updated in the face of reality before legislators have to finalize anything. The EU does not have foresight into what reasonable laws should look like, they are nervous about unintended consequences, and they do not want to drive good-faith organizations away, they are trying to do this correctly.
This cynical take seems wise and world-weary but it is just plain ignorant, please read the link.
LMAO. Facebook is not big? Its founder is literally the sleaziest CEO out there. Cambridge Analytica, Myanmar, restrictions on Palestine, etc. Let us not fool ourselves. There are those online who seek to defend a master that could care less about them. Fascinating.
My opinion on this: Europe lags behind in this field, and thus can enact regulations that profit the consumer. We need more of those in the US.
Meta isn't actually an AI company, as much as they'd like you to think they are now. They don't mind if nobody comes out as the big central leader in the space, they even release the weights for their models.
Ask Meta to sign something about voluntarily restricting ad data or something and you'll get your same result there.
About 2 weeks ago OpenAI won a $200 million contract with the Defense Department. That's after partnering with Anduril for quote "national security missions." And all that is after the military enlisted OpenAI's "Chief Product Officer" and sent him straight to Lt. Colonel to work in a collaborative role directly with the military.
And that's the sort of stuff that's not classified. There's, with 100% certainty, plenty that is.
The US, China and others are sprinting and thus spiraling towards the majority of society's destitution unless we force these billionaires hands; figure out how we will eat and sustain our economies where one person is now doing a white or blue (Amazon warehouse robots) collar job that ten use to do.
Just like GDPR, it will tremendously benefit big corporations (even if Meta is resistant) and those who are happy NOT to follow regulations (which is a lot of Chinese startups).
I charge my phone wirelessly. The presence of a port isn't a positive for me. It's just a hole I could do without. The shape of the hole isn't important.
Well Europe haven't enacted policies actually breaking American monopolies until now.
Europeans are still essentially on Google, Meta and Amazon for most of their browsing experiences. So I'm assuming Europe's goal is not to compete or break American moat but to force them to be polite and to preserve national sovereignty on important national security aspects.
A position which is essentially reasonable if not too polite.
> So I'm assuming Europe's goal is not to compete or break American moat but to force them to be polite and to preserve national sovereignty on important national security aspects.
When push comes to shove the US company will always prioritize US interest. If you want to stay under the US umbrella by all means. But honestly it looks very short sighted to me.
You have only one option. Grow alternatives. Fund your own companies. China managed to fund the local market without picking winners. If European countries really care, they need to do the same for tech.
If they don't they will forever stay under the influence of another big brother. It is US today, but it could be China tomorrow.
I’d side with Europe blindly over any corporation.
The European government has at least a passing interest in the well being of human beings while that is not valued by the incentives that corporations live by
Maybe the others have put in a little more effort to understand the regulation before blindly criticising it? Similar to the GDPR, a lot of it is just common sense—if you don’t think that "the market" as represented by global mega-corps will just sort it out, that is.
Our friends in the EU have a long history of well-intentioned but misguided policy and regulations, which has led to stunted growth in their tech sector.
Maybe some think that is a good thing - and perhaps it may be - but I feel it's more likely any regulation regarding AI at this point in time is premature, doomed for failure and unintended consequences.
Yet at the same time, they also have a long history of very successful policy, such as the USB-C issue, but also the GDPR, which has raised the issue of our right to privacy all over the world.
How long can we let AI go without regulation? Just yesterday, there was a report here on Delta using AI to squeeze higher ticket prices from customers. Next up is insurance companies. How long do you want to watch? Until all accountability is gone for good?
Which... has the consequences of stifling innovation. Regulations/policy is two-way street.
Who's to say USB-C is the end-all-be-all connector? We're happy with it today, but Apple's Lightning connector had merit. What if two new, competing connectors come out in a few year's time?
The EU regulation, as-is, simply will not allow a new technically superior connector to enter the market. Fast forward a decade when USB-C is dead, EU will keep it limping along - stifling more innovation along the way.
Standardization like this is difficult to achieve via consensus - but via policy/regulation? These are the same governing bodies that hardly understand technology/internet. Normally standardization is achieved via two (or more) competing standards where one eventually "wins" via adoption.
You mean that thing (or is that another law?) that forces me to find that "I really don't care in the slightest" button about cookies on every single page?
No, the laws that ensures that private individuals have the power to know what is stored about them, change incorrect data, and have it deleted unless legally necessary to hold it - all in a timely manner and financially penalize companies that do not.
that's deflecting responsibility. it's important to care about the actual effects of decisions, not hide behind the best case scenario. especially for governments.
in this case, it is clear that the EU policy resulted in cookie banners
I admit that I am biased enough to immediately expect the AI agreement to be exactly what we need right now if this is how Meta reacts to it. Which I know is stupid because I genuinely have no idea what is in it.
There seem to be 3 chapters of this "AI Code of Practice" https://digital-strategy.ec.europa.eu/en/policies/contents-c... and it's drafting history https://digital-strategy.ec.europa.eu/en/policies/ai-code-pr...
I did not read it yet, only familiar with the previous AI Act https://artificialintelligenceact.eu/ .
If I'd were to guess Meta is going to have a problem with chapter 2 of "AI Code of Practice" because it deals with copyright law, and probably conflicts with their (and others approach) of ripping text out of copyrighted material (is it clear yet if it can be called fair use?)
> is it clear yet if it can be called fair use?
Yes.
https://www.publishersweekly.com/pw/by-topic/digital/copyrig...
Though the EU has its own courts and laws.
District judge pretrial ruling on June 25th, I'd be surprised this doesn't get challenged soon in higher courts.
And acquiring the copyrighted materials is still illegal - this is not a blanket protection for all AI training on copyrighted materials
Presumably it is Meta's growth they have in mind.
Edit: from the linked in post, Meta is concerned about the growth of European companies:
"We share concerns raised by these businesses that this over-reach will throttle the development and deployment of frontier AI models in Europe, and stunt European companies looking to build businesses on top of them."
Of course. Skimming over the AI Code of Practice, there is nothing particularly unexpected or qualifying as “overreach”. Of course, to be compliant, model providers can’t be shady which perhaps conflicts with Meta’s general way of work.
There’s a summary of the guidelines here for anyone who is wondering:
https://artificialintelligenceact.eu/introduction-to-code-of...
It’s certainly onerous. I don’t see how it helps anyone except for big copyright holders, lawyers and bureaucrats.
[flagged]
This all seems fine.
Most of these items should be implemented by major providers…
The problem is this severely harms the ability to release opens weights models, and only leaves the average person with options that aren't good for privacy.
I don't care about your overly verbose, blandly written slop. If I wanted a llm summary, I would ask an llm myself.
This really is the 2025 equivalent to posting links to a google result page, imo.
It is... helpful though. More so than your reply
Touché, I'll grant you that.
More verbose than the source text? And who cares about bland writing when you're summarizing a legal text?
Nope. This text is embedded in HN and will survive rather better than the prompt or the search result, both of which are non-reproducible. It may bear no relation to reality but at least it won't abruptly disappear.
Not just Meta, 40 EU companies urged EU to postpone roll out of the ai act by two years due to it's unclear nature. This code of practice is voluntary and goes beyond what is in the act itself. EU published it in a way to say that there would be less scrutiny if you voluntarily sign up for this code of practice. Meta would anyway face scrutiny on all ends, so does not seem to a plausible case to sign something voluntary.
One of the key aspects of the act is how a model provider is responsible if the downstream partners misuse it in any way. For open source, it's a very hard requirement[1].
> GPAI model providers need to establish reasonable copyright measures to mitigate the risk that a downstream system or application into which a model is integrated generates copyright-infringing outputs, including through avoiding overfitting of their GPAI model. Where a GPAI model is provided to another entity, providers are encouraged to make the conclusion or validity of the contractual provision of the model dependent upon a promise of that entity to take appropriate measures to avoid the repeated generation of output that is identical or recognisably similar to protected works.
[1] https://www.lw.com/en/insights/2024/11/european-commission-r...
Kaplan's LinkedIn post says absolutely nothing about what is objectionable about the policy. I'm inclined to think "growth-stunting" could mean anything as tame as mandating user opt-in for new features as opposed to the "opt-out" that's popular among US companies.
Why does meta need to sign anything? I thought the EU made laws that anyone operating in the EU including meta had to comply to.
It's not a law, it's a voluntary code of conduct given heft by EU endorsement.
> it's a voluntary code of conduct
So then it's something completely worthless in the globally competitive cutthroat business world, that even the companies who signed won't follow, they just signed it for virtue signaling.
If you want companies to actually follow a rule, you make it a law and you send their CEOs to jail when they break it.
"Voluntary codes of conduct" have less value in the business world than toilet paper. Zuck was just tired of this performative bullshit and said the quiet part out loud.
No, it's a voluntary code of conduct so AI providers can start implementing changes before the conduct becomes a legal requirement, and so the code itself can be updated in the face of reality before legislators have to finalize anything. The EU does not have foresight into what reasonable laws should look like, they are nervous about unintended consequences, and they do not want to drive good-faith organizations away, they are trying to do this correctly.
This cynical take seems wise and world-weary but it is just plain ignorant, please read the link.
Interesting because OpenAI committed to signing
https://openai.com/global-affairs/eu-code-of-practice/
The biggest player in the industry welcomes regulation, in hopes it’ll pull the ladder up behind them that much further. A tale as old as red tape.
LMAO. Facebook is not big? Its founder is literally the sleaziest CEO out there. Cambridge Analytica, Myanmar, restrictions on Palestine, etc. Let us not fool ourselves. There are those online who seek to defend a master that could care less about them. Fascinating. My opinion on this: Europe lags behind in this field, and thus can enact regulations that profit the consumer. We need more of those in the US.
> Let us not fool ourselves. There are those online who seek to defend a master that could care less about them. Fascinating.
How could you possibly infer that from what I said?
Fascinating.
Meta isn't actually an AI company, as much as they'd like you to think they are now. They don't mind if nobody comes out as the big central leader in the space, they even release the weights for their models.
Ask Meta to sign something about voluntarily restricting ad data or something and you'll get your same result there.
[dead]
Sam has been very pro-regulation for a while now. Remember his “please regulate me” world tour?
OpenAI does direct business with government bodies. Not sure about Meta.
About 2 weeks ago OpenAI won a $200 million contract with the Defense Department. That's after partnering with Anduril for quote "national security missions." And all that is after the military enlisted OpenAI's "Chief Product Officer" and sent him straight to Lt. Colonel to work in a collaborative role directly with the military.
And that's the sort of stuff that's not classified. There's, with 100% certainty, plenty that is.
The US, China and others are sprinting and thus spiraling towards the majority of society's destitution unless we force these billionaires hands; figure out how we will eat and sustain our economies where one person is now doing a white or blue (Amazon warehouse robots) collar job that ten use to do.
I have a strong aversion to Meta and Zuck but EU is pretty tone-deaf. Everything they do reeks of political and anti-American tech undertone.
They're career regulators
Just like GDPR, it will tremendously benefit big corporations (even if Meta is resistant) and those who are happy NOT to follow regulations (which is a lot of Chinese startups).
And consumers will bear the brunt.
Good. Europe is a regulatory morass. A waste of everyone's time. As their economies become irrelevant they will also stop mattering.
Sent from an iPhone probably having USB-C because of the EU.
I'd deal with the odd lightning cable over having retarded popups on every website I ever visit
Just because they occasionally (and even frequently) do good thing, does not mean that overall their policies don't harm them own economies.
I charge my phone wirelessly. The presence of a port isn't a positive for me. It's just a hole I could do without. The shape of the hole isn't important.
Besides, I posted from my laptop.
I'm surprised that most of the comments here are siding with Europe blindly?
Am I the only one who assumes by default that European regulation will be heavy-handed and ill conceived?
If I've got to side blindly with any entity it is definitely not going to be Meta. That's all there is.
I mean, ideally no one would side blindly at all :D
Well Europe haven't enacted policies actually breaking American monopolies until now.
Europeans are still essentially on Google, Meta and Amazon for most of their browsing experiences. So I'm assuming Europe's goal is not to compete or break American moat but to force them to be polite and to preserve national sovereignty on important national security aspects.
A position which is essentially reasonable if not too polite.
> So I'm assuming Europe's goal is not to compete or break American moat but to force them to be polite and to preserve national sovereignty on important national security aspects.
When push comes to shove the US company will always prioritize US interest. If you want to stay under the US umbrella by all means. But honestly it looks very short sighted to me.
After seeing this news https://observer.co.uk/news/columnists/article/the-networker..., how can you have any faith that they will play nice?
You have only one option. Grow alternatives. Fund your own companies. China managed to fund the local market without picking winners. If European countries really care, they need to do the same for tech.
If they don't they will forever stay under the influence of another big brother. It is US today, but it could be China tomorrow.
I’d side with Europe blindly over any corporation.
The European government has at least a passing interest in the well being of human beings while that is not valued by the incentives that corporations live by
All corporations that exist everywhere make worse decisions than Europe is a weirdly broad statement to make.
So you're surprised that people are siding with Europe blindly, but you're "assuming by default" that you should side with Meta blindly.
Perhaps it's easier to actually look at the points in contention to form your opinion.
I don't remember saying anything about blindly deciding things being a good thing.
Maybe the others have put in a little more effort to understand the regulation before blindly criticising it? Similar to the GDPR, a lot of it is just common sense—if you don’t think that "the market" as represented by global mega-corps will just sort it out, that is.
I'm specifically referring to several comments that say they have not read the regulation at all, but think it must be good if Meta opposes it.
Our friends in the EU have a long history of well-intentioned but misguided policy and regulations, which has led to stunted growth in their tech sector.
Maybe some think that is a good thing - and perhaps it may be - but I feel it's more likely any regulation regarding AI at this point in time is premature, doomed for failure and unintended consequences.
Yet at the same time, they also have a long history of very successful policy, such as the USB-C issue, but also the GDPR, which has raised the issue of our right to privacy all over the world.
How long can we let AI go without regulation? Just yesterday, there was a report here on Delta using AI to squeeze higher ticket prices from customers. Next up is insurance companies. How long do you want to watch? Until all accountability is gone for good?
I mean, getting USB-C to be usable on everything is like a nice-to-have, I wouldn't call it "very successful policy".
It’s just an example. The EU has often, and often successfully, pushed for standardisation to the benefit of end users.
Which... has the consequences of stifling innovation. Regulations/policy is two-way street.
Who's to say USB-C is the end-all-be-all connector? We're happy with it today, but Apple's Lightning connector had merit. What if two new, competing connectors come out in a few year's time?
The EU regulation, as-is, simply will not allow a new technically superior connector to enter the market. Fast forward a decade when USB-C is dead, EU will keep it limping along - stifling more innovation along the way.
Standardization like this is difficult to achieve via consensus - but via policy/regulation? These are the same governing bodies that hardly understand technology/internet. Normally standardization is achieved via two (or more) competing standards where one eventually "wins" via adoption.
Well intentioned, but with negative side-effects.
> GDPR
You mean that thing (or is that another law?) that forces me to find that "I really don't care in the slightest" button about cookies on every single page?
No, the laws that ensures that private individuals have the power to know what is stored about them, change incorrect data, and have it deleted unless legally necessary to hold it - all in a timely manner and financially penalize companies that do not.
That's not the GDPR.
Everything in this thread even remotely anti-EU-regulation is being extreme downvoted
It is fascinating. I assume that the tech world is further to the left, and that interpretation of "left" is very pro-AI regulation.
Yeah it's kinda weird.
Feels like I need to go find a tech site full of people who actually like tech instead of hating it.
Good. As Elon says, the only thing the EU does export is regulation. Same geniuses that make us click 5 cookie pop-ups every webpage
They didn't give us that. Mostly non-compliant websites gave us that.
that's deflecting responsibility. it's important to care about the actual effects of decisions, not hide behind the best case scenario. especially for governments.
in this case, it is clear that the EU policy resulted in cookie banners