Whenever AI is mentioned lots of people in the Linux space immediately react negatively. Creators like TheLinuxExperiment on YouTube always feel the need to add a disclaimer that “some people think AI is problematic” or something along those lines if an AI topic is discussed. I get that AI has many problems but at the same time the potential it has is immense, especially as an assistant on personal computers (just look at what “Apple Intelligence” seems to be capable of.) Gnome and other desktops need to start working on integrating FOSS AI models so that we don’t become obsolete. Using an AI-less desktop may be akin to hand copying books after the printing press revolution. If you think of specific problems it is better to point them out and try think of solutions, not reject the technology as a whole.
TLDR: A lot of ludite sentiments around AI in Linux community.
I won’t rehash the arguments around “AI” that others are best placed to make.
My main issue is AI as a term is basically a marketing one to convince people that these tools do something they don’t and its causing real harm. Its redirecting resources and attention onto a very narrow subset of tools replacing other less intensive tools. There are significant impacts to these tools (during an existential crisis around our use and consumption of energy). There are some really good targeted uses of machine learning techniques but they are being drowned out by a hype train that is determined to make the general public think that we have or are near Data from Star Trek.
Addtionally, as others have said the current state of “AI” has a very anti FOSS ethos. With big firms using and misusing their monopolies to steal, borrow and coopt data that isn’t theirs to build something that contains that’s data but is their copyright. Some of this data is intensely personal and sensitive and the original intent behind the sharing is not for training a model which may in certain circumstances spit out that data verbatim.
Lastly, since you use the term Luddite. Its worth actually engaging with what that movement was about. Whilst its pitched now as generic anti-technology backlash in fact it was a movement of people who saw what the priorities and choices in the new technology meant for them: the people that didn’t own the technology and would get worse living and work conditions as a result. As it turned out they were almost exactly correct in thier predictions. They are indeed worth thinking about as allegory for the moment we find ourselves in. How do ordinary people want this technology to change our lives? Who do we want to control it? Given its implications for our climate needs can we afford to use it now, if so for what purposes?
Personally, I can’t wait for the hype train to pop (or maybe depart?) so we can get back to rational discussions about the best uses of machine learning (and computing in general) for the betterment of all rather than the enrichment of a few.
yeah i see that too. it seems like mostly a reactionary viewpoint. the reaction is understandable to a point since a lot of the “AI” features are half baked and forced on the user. to that point i don’t think GNOME etc should be scrambling to add copies of these features.
what i would love to see is more engagement around additional pieces of software that are supplemental. for example, i would love if i could install a daemon that indexes my notes and allows me to do semantic search. or something similar with my images.
the problems with AI features aren’t within the tech itself but in the surrounding politics. it’s become commonplace for “responsible” AI companies like OpenAI to not even produce papers around their tech (product announcement blogs that are vaguely scientific don’t count), much less source code, weights, and details on training data. and even when Meta releases their weights, they don’t specify their datasets. the rat race to see who can make a decent product with this amazing tech has made the whole industry a bunch of pearl clutching FOMO based tweakers. that sparks a comparison to blockchain, which is fair from the perspective of someone who hasn’t studied the tech or simply hasn’t seen a product that is relevant to them. but even those people will look at something fantastical like ChatGPT as if it’s pedestrian or unimpressive because when i asked it to write an implementation of the HTTP spec in the style of Fetty Wap it didn’t run perfectly the first time.
Reminder that we don’t even have AI yet, just learning machine models, which are not the same thing despite wide misuse of the term AI.
That’s just nitpicking. Everyone here knows what we mean by AI. Yes it refers to LLMs.
Reminds me of Richard Stallman always interjecting to say “actually its gnu/Linux or as I like to say gnu plus Linux”…
Well no Mr Stallman its actually gnu + Linux + Wayland + systemd + chromium and whatever other software you have installed, are you happy now??
So when we actually do have AI, what are we supposed to call it? The current use of the term “AI” is too ambiguous to be of any use.
I dont think the community is generally against AI, there’s plenty of FOSS projects. They just don’t like cashgrabs, enshittification and sending personal data to someone else’s computer.
I don’t see anyone calling for cash grabs or privacy destroying features to be added to gnome or other projects so I don’t see why that would be an issue. 🙂
On device Foss models to help you with various tasks.
You are, if you’re calling for Apple like features.
You might argue that “private cloud” is privacy preserving, but you can only implement that with the cash of Apple. I would also argue that anything leaving my machine, to a bunch of servers I don’t control, without my knowledge is NOT preserving my privacy.
You might argue that “private cloud” is privacy preserving
I don’t know since when “on device” means send it to a server. Come up with more straw men I didn’t mention for you to defeat.
Apple’s “private cloud” is a thing. Not all “Apple Intelligence” features are “on device”, some can and do utilize cloud-based processing power, and this will also be available to app developers.
Apparently this has additional safeguards vs “normal cloud” which is why they are branding it “private cloud”.
But it’s still “someone else’s computer” and apple is not keeping their AI implementation 100% on device.






