Opinionated - AI

Created on: 13.11.2025
Last edited on: 15.11.2025; prior version




What do I mean with 'AI'?

'AI' is a term used in so many ways today. And in my opinion, we do not have anything that would qualify as 'AI' in the sense of the word. There is nothing intelligent about a statistical model; okay, there is something intelligent about how they are build and trained, but the model itself isn't what I would define as intelligent.
This site will talk about LLMs like the GPT series of models released by OpenAI. For those uninitiated, LLM stands for "Large Language Model" and describes a kind of model which takes in Text and generates new more Text from that. You may know for example ChatGPT, although technically that is just the UI made by OpenAI to interact with their GPT series of modesl.
There are definitely valid uses for kinds of generative models, for example finding medicines or helping people with physical or mental disabilities, but I will not talk about this further here.

So why don't I like LLMs?

I don't like LLMs as they currently are made and used, but why? There are so many people who use them day to day for so many things and are happy with it, why don't I like it? Shouldn't I as a student in software development and electronics, love this? Isn't this progress? Well no. The concept of LLMs has existed since at least the 90s. In short, they learn from a lot of texts how those are written, what the next most likely word is to a previous text. In a way they are fancier autocompletion. And would you trust your phone's autocorrection to give you factual answers? No? Thought so.
Also, have you noticed? I told you that LLMs need data to be trained. Where does one get this data from and how much data is needed? The former question is easy to answer: theft. The second is a bit harder, since it differs depending on the size of a model, but for the greatest and latest models the answer is: everything that can be collected. If you have noticed a problem with this, then happy you, many have before you. Yes, training LLMs happens largely on copyrighted materials that are used without their creators' permission [1][2]. This is not only the case with LLMs, but with many commercial generative models, for example for images or music [2]. It must be said that in some cases courts have ruled that the models themselves don't infringe copyright laws, for example in the UK [3], but it is unclear if the training process infringes copyright laws. Although it is debatable, if a system that can reproduce creations should fall under fair use.
And then there is the invisible side of LLMs: their impact on the climate. Firstly let me say this is an area maybe even more vague than the copyright situation, since there is almost no public data of the companies behind the state of the art LLMs on this [4]. Another problem is, how do we calculate the impact of LLMs? There is almost no data on how much CO2 internet requests produce, which are made by the billions by scrapers to get the Data for training [5][6][7]. Then there is the training itself, using a lot of electrical energy, as well as water and partially other resources. The server farms have to be built. And when using LLMs, the servers running them also use electrical power and water [8][9]. To be fair, most server centres are really power efficient, but the size of LLMs and their compute intensity requires a lot of servers, which in total produces a lot of emissions.

Let us take a closer look into the ethics of LLMs beyond content theft.

LLMs are based on data, mostly from the internet. Do you know what happens on the internet? If not, have a read! [10][11][12][13][14][15][16][17] And also there is Reddit. I will spare you from examples, but that isn't what I would call good training-data if you want a balanced and racism free discussion. So is AI biased? Yes. Yes, it is [18][19][20][21]. And it also gets fed by autocrats [22]! It not only re-enforces the same biases humans have, even strengthened them, but it also is very susceptible to misinformation and propaganda. Is this a problem? I let you be the judge (Hint: yes it is.).
Furthermore, the training data also contains heaps of misinformation, which is quite the problem if you want to use LLMs for news-gathering. To be honest, sometimes I feel like I can't help people who think a statistical language model can be used for facts gathering reliably; but then again, there are so many naive people thinking machines are smarter than humans. But have your source that you shouldn't trust LLMs on news topics: [23]

But there are even worse ethical dilemmas I see. There are people using LLMs who have psychological problems, and LLMs will support them in their views. I can really just refer to a great video by Eddy Burback [24] and a video by Cedric Mössner (German) [25] here. But let me summarise for you, if you don't have the time to watch them or don't want to use YouTube (I get that). Note though that Cedric Mössner has their sources under his video. I will not refer to all of them here.
In Eddys video he, shows himself interacting with a LLM by OpenAI through the ChatGPT WebUI and trying out how far it would go to affirm him in his own views. The answer to this? It will go to the end. Cedric Mössner comes to a similar conclusion. But he is also talking about the fact that this problem persists in newer models. And even more importantly, he shows that the model had recognised that he in his own experiment was showing signs of hallucinations, but still proceeded to affirm him further. This might be a decision by OpenAI, but for legal reasons I must say: I don't know.

Okay, but there is another point. We don't like big companies and autocrats, right? But LLMs are owned by them. Why can we trust that? I get that many will not follow my argument here, but realistically speaking not only Chinese models like Deepseek, but also models by American companies will be under strong political control. Why do I say this? Well, firstly the owners of those big companies sat down with Donald Trump and praised him [26][27], so one might argue they will cooperate with him. Well, they give him money [28], so that is that. Furthermore, people like Mark Zuckerberg have shown to have flexible morals at best [29] or in case of Elon Musk straight up Nazi views [30][31]. Personally, I don't have the greatest trust into those people to not abuse this power over this software, which is widely used and trusted. And this isn't just my opinion, but experts also warn about this [32].

LLMs are so good at X. No they aren't.

Now, we know LLMs aren't accurate with news, but that changes often. How about something that isn't as factually important, so no history topics; how about programming? How about summarising things? The latter works okay, but the former is kind of hit or miss. It really depends on what you want to do and how much information about this is online and how specific your request is. Anecdotally, for me LLMs were slower with simple questions than just searching myself and with harder questions failed all together. You need to work with a specific version of a framework? No, the LLM will use other versions as reference as well, so good luck. You want code compliant with special policies? Forget about it. Show it to the ESA? Wait, no, LLMs aren't allowed in the ECSS, so that is a nogo anyway. Small models specific for coding are great, but general purpose LLMs are quite bad at it. Also, they are no software engineer, so don't think they can just build you a giant new thing.

Okay, what about creative writing? Welp, please stop if you really consider this. LLMs can help find another way of writing something, but it is not creative. It can not create Art, but only Artifice. Art is there for conveying unquantifiable information, like feelings. LLMs are by definition quantified versions of texts, which internally have no feelings or understanding of the world. Using texts from actual artists to create a new one, without own input, isn't art. It is theft.

Climate.

I have already said this isn't really quantifiable, how much LLMs impact the climate. The answer is quite likely somewhere in between "much" and "very much". Some studies say that LLMs are more efficient at writing than humans, energy wise, but often don't consider the impact the production of the servers had, as well as the impact of the training itself. This feels weird, when on the site of the human you are considering their average living habits as well. If you use an LLM you will still have a living creature there waiting for the result. If you say in a vacuum a LLM is better that can be true, but in the real world we can not just kill all people and replace them with LLMs. Also in this study they assumed the client device would require 75 W while writing a text, although they say it is for an average laptop. Sure there are modern laptops that can draw much more, but whilst writing a text you can not just slap the average TDP into your calculation. The most modern laptops will consume less than 10 W during a task like writing a text. I can only explain this by saying that the data the article is referring to seems to be over 10 years old. Looking at the page in the Wayback Machine [33] it shows the first snapshot in 2014, but the copyright on the bottom of the page says 2012. Either way, the data seems to be the same today on that page, even though the site design changed slightly. This refers to the following article: [34]
Furthermore the scale of the industry is rapidly growing, models get more and more resource hungry the bigger they get and also the bigger the query is [35]. Also more and more people are using the offerings of OpenAI and others [36]. Further Companies like Google are putting the outputs of LLMs into their products by default, which also adds to the total emissions.
At the moment it seems like LLMs will put out about half a gram up to over 20 grams of CO2e per Query, depending on the model and size of the query [35]. This doesn't calculate in how much CO2e was set free during the training phase. With broad estimates the training alone will, for big models like those OpenAI is using today, have used more than a GWh of electricity [37]. With the average CO2 emissions for American electricity being about 384 grams of CO2e per kWh [38] that alone accounts for multiple hundreds of metric tons of CO2e emissions.
And in the end Remember: each time you query an LLM instead of thinking yourself, it will create more demand for servers that require many rare resources, which not only impact the climate but also the people, the kids getting those resources out of the ground. If you didn't get this: Yes, there is still child labour for some materials that are used in computers [39][40][41]. There is no Fair world. Even companies that strive to be fair, aren't totally fair [42]. I know we can't just stop using technology and by that stop this, but using giant inefficient services like LLMs certainly doesn't help.

Back to HomePage!