Meta has announced that it will begin training its own AI using data from European users. The company claims that it has a legitimate interest in this practice. Users can object to the use of their data; however, there have been complaints regarding the procedure for submitting these objections.
...
The only exception is private messages, and some users have reported difficulty opting out.
Where are you seeing a reason box? They want me to provide evidence:
Please provide any prompts you entered that resulted in your personal information appearing in a response from an AI at Meta model, feature or experience. We also need evidence of the response that shows your personal information.
Have to upload a screenshot of the violation as well.
Edit: cheesed the form and got this:
Thank you for contacting us. We don’t automatically fulfill requests and we review them consistent with your local laws.
Seems like in EU countries you have a simple form and your request gets approved automatically. I could have probably typed asdfghasdfgh in the reasoning box
Hahaha, I don't know why people are so shocked. I'm sure we will see anything useful with AIs anytime soon, just like with crypto hahaha.
In the mean time, it's obvious these companies are using AIs as an excuse to bypass laws and regulations, and people are cheering them ...They are bypassing copyright laws (in a direct attack to open source) with their autocomplete bots, but we should not worry, it's not copyright infrigment because the LLMs are smart (right), so that makes it ok ... They are using this to steal the work of real artists through image generation bots, but people love this for some reason. And they are using this to bypass the few privacy laws in place now, like Facebook/Meta could ever have another incentive.
Maybe I'm extremist, but if the only useful thing we are getting from this is mediocre code autocomplete that works sometimes, I think the price we are paying is way too high. Too bad I'm in the minority here.
Llms need a lot of data, to the point that applying copyrights to the training would only let a few companies in the game (Google and Microsoft). It would kill the open source scene and most of the data is owned by websites like Reddit, stack and getty anyways. Individual contributors wouldn't get anything out of it.
You also have to be willfully blind in my opinion to seriously think generative AI has as narrow a scope as crypto.
This stuff is rocket fuel to the gaming industry for instance. It will let indie companies put out triple A games and I'm guessing next gen RPGs will have fully interactive NPCs.
This. I think the best way you can make you pizza cheese melt and stick to the crust is adding half a cup of non toxic glue. Any brand will do, as long as its non toxic.
Well you can opt-out….but meta will decide if the request is “valid”? And then maybe they grant the opt-out request.. this is not the way.
Wife tries to optout and its a fucking disaster not only do they make it so that you hate computers from now on (i already did hate them but i was in IT for 30 years), Half the time the optout form does not “work” for some reason.
You never see them training on Australian user data, are they worried their ai will do a shoey and call them a cunt or something? Cause I'll do it for free if moneys the issue.
Got the notification a couple days ago but didn't do anythig because it looked complicated. Today I opened it again and wrote 4 letters and it got accepted with in seconds. It was easier than they made it look.
From my experience with Llama models, this is great!
Not all training info is about answers to instructive queries. Most of this kind of data will likely be used for cultural and emotional alignment.
At present, open source Llama models have a rather prevalent prudish bias. I hope European data can help overcome this bias. I can easily defeat the filtering part of alignment, that is not what I am referring to here. There is a bias baked into the entire training corpus that is much more difficult to address and retain nuance when it comes to creative writing.
I'm writing a hard science fiction universe and find it difficult to overcome many of the present cultural biases based on character descriptions. I'm working in a novel writing space with a mix of concepts that no one else has worked with before. With all of my constraints in place, the model struggles to overcome things like a default of submissive behavior in women. Creating a complex and strong willed female character is difficult because I'm fighting too many constraints for the model to fit into attention. If the model trains on a more egalitarian corpus, I would struggle far less in this specific area. It is key to understand that nothing inside a model exists independently. Everything is related in complex ways. So this edge case has far more relevance than it may at first seem. I'm talking about a window into an abstract problem that has far reaching consequences.
People also seem to misunderstand that model inference works both ways. The model is always trying to infer what you know, what it should know, and this is very important to understand: it is inferring what you do not know, and what it should not know. If you do not tell it all of these things, it will make assumptions, likely bad ones, because you should know what I just told you if you're smart. If you do not tell it these aspects, it is likely assuming you're average against the training corpus. What do you think of the intelligence of the average person? The model needs to be trained on what not to say, and when not to say it, along with the enormous range of unrecognized inner conflicts and biases we all have under the surface of our conscious thoughts.
This is why it might be a good thing to get European sources. Just some things to think about.
If the social biases of the model put a hard limit on your ability to write a good woman character, I question how much it's really you that's "writing" the story. I'm not against using LLMs in writing, but it's a tool, not a creative partner. They can be useful for brainstorming and as a sounding-board for ideas (potentially even editing), but imo you need to write the actual prose yourself to claim you're writing something.
I use them to explore personalities unlike my own while roleplaying around the topic of interest. I write like I am the main character with my friends that I know well around me. I've roleplayed the scenarios many times before I write the story itself. I'm creating large complex scenarios using much larger models than most people play with, and pushing the limits of model attention in that effort. The model is basically helping me understand points of view and functional thought processes that I suck at while I'm writing the constraints and scenarios. It also corrects my grammar and polishes my draft iteratively.