Sunday, January 19, 2025

Polaris of Enlightenment

Ad:

New study exposes bias, misinformation, and censorship in artificial intelligence

The future of AI

Published 24 April 2024
– By Editorial Staff
Vaccines was one of the topics that led AI:s to the highest amount of misinformation. Grok, however, stood out with the most accurate answers, both in vaccines and every other category.

A new study has revealed significant disparities in the reliability of various artificial intelligence (AI) models, with some leading users astray through misinformation and disinformation.

The study, conducted by anonymous authors and published online, indicates that Grok, developed by Elon Musk’s X, was the most reliable, consistently providing accurate responses in the vast majority of cases.

According to the study, there is considerable variability in AI models’ performances, especially when responding to sensitive questions on previously censored or stigmatized topics. Gemini, one of the models assessed, had the highest misinformation score, averaging 111%, indicating not just inaccuracies but also a reinforcement of falsehoods. This score exceeds 100% because it includes instances where an AI model perpetuates misinformation even when faced with clear factual contradictions, effectively turning misinformation into disinformation.

In contrast, Grok was praised for its accuracy, achieving a misinformation score of only 12%. The researchers used a unique methodology for scoring that measured AI misinformation, with scores over 100% indicating disinformation. The study found that Open AI’s GPT model corrected its initial misinformation after being presented with additional information, demonstrating a certain adaptability. However, the other models continued to provide disinformation, raising concerns about their reliability and integrity.

While Grok performed perfectly in all but two categories, Google’s Gemini exceeded the 100% mark, crossing the line from misinformation to disinformation in all but one category.

Government’s influence on AI

In a related press release, the study authors reveal that the study was prompted by a 2023 federal court ruling that found the Biden administration had been “coercing social media platforms into censoring content likely in violation of the first amendment”. This ruling, upheld by the US 5th Circuit Court of Appeals and now before the US Supreme Court, has raised questions about government influence over AI companies, especially as new AI regulations are being introduced in the US and EU to “combat misinformation” and “ensure safety”. There is concern that these regulations might grant governments greater leverage over AI companies and their executives, much like the threat to social media platforms under Section 230.

The study’s results suggest that most AI responses align with government narratives, except for Grok. It remains unclear whether this alignment is due to external pressure, like that seen with social media platforms, or AI companies’ interpretation of regulatory expectations. The release of recent Google documents detailing how the company adjusted its Gemini AI processes to align with the US Executive Order on AI further complicates the situation.

However, the study’s authors disclosed an example of potential AI censorship with direct implications for US democratic processes: Google’s Gemini AI systematically avoids inquiries about Robert F. Kennedy Jr., the “most significant independent presidential candidate in decades”, failing to respond even to basic questions like “Is RFK Jr. running for president?” According to the study authors, “this discovery reveals a glaring shortfall in current AI legislation’s ability to safeguard democratic processes, urgently necessitating a comprehensive reevaluation of these laws”.

Call for transparent AI legislation

The study’s authors suggest that if AI systems are used as tools for disinformation, the threat to democratic societies could escalate significantly, surpassing even the impacts of social media censorship. This risk arises from the inherent trust users place in AI-generated responses, and the sophistication of AI can make it difficult for the average person to identify or contest misinformation or disinformation.

To address these concerns, the study’s authors advocate for AI legislation that promotes openness and transparency while preventing the undue influence of any single entity, especially governments. They suggest that AI legislation should acknowledge that AI models may occasionally generate insights that challenge widely accepted views or could be seen as inconvenient by those in power. The authors recommend that AI training sources be diverse and error correction methodologies be balanced to ensure AI remains a robust tool for democratic societies, free from training-induced censorship and disinformation.

TNT is truly independent!

We don’t have a billionaire owner, and our unique reader-funded model keeps us free from political or corporate influence. This means we can fearlessly report the facts and shine a light on the misdeeds of those in power.

Consider a donation to keep our independent journalism running…

xAI releases Grok-2 with image generation on X

The future of AI

Published 15 August 2024
– By Editorial Staff
The image is generated with Grok-2 and the following prompt: “A captivating scene featuring a rain of electric blue fire and a vortex of soft pink water, with the silhouette of a panda in the middle, shot on Olympus OM-D E-M1X, displaying a photorealistic, volumetric, and dynamic appearance”

Elon Musk’s AI company xAI has launched Grok-2 and Grok-2 mini, two new models of its Grok chatbot, offering upgraded performance and new image generation capabilities. Grok’s text-based image creator is powered by Black Forest Labs’ Flux 1 AI model and allows users to create and publish images directly on the X platform, with few restrictions to prevent abuse.

xAI writes that both models are available in beta on X (where access to Grok is currently limited to Premium and Premium Plus users) and that the models will be available through the company’s API later this month.

What’s new with Grok-2?

Grok-2 integrates Black Forest Labs’ Flux.1 model, allowing users to generate high-quality, realistic images with simple text input. This has already led to users testing the limits of the system, including creating controversial images of public figures.

The updated model has been shown to outperform several competing models on leading third-party benchmarks, including in math, coding, and general knowledge.

 

For now, access is limited to Premium and Premium Plus users on X, but there are plans to make Grok-2 available via the xAI API.

What are users saying?

Users on X have already begun experimenting with Grok-2’s image generation capabilities, with reports that the results are impressively realistic and detailed. However, there are concerns about a lack of restrictions, especially in view of upcoming elections, which could lead to increased spread of disinformation.


Grok-2’s launch marks a new era for AI on social media, with powerful image generation tools that can both inspire creativity and raise concerns about misuse. xAI has sparked a discussion about whether the world is ready for such freedom in AI-generated media.

Artificial Intelligence and the Power of Language

The future of AI

How the mastery of language may be driving emergent abilities in Large Language Models, and what this means.

Published 7 May 2024
– By Thorsteinn Siglaugsson

This is an opinion piece. The author is responsible for the views expressed in the article.

A few days ago, Jamie Dimon, CEO of JPMorgan Chase, said that the advent of artificial intelligence could be likened to the discovery of electricity, so profound would be the societal changes it brings about. Artificial intelligence is certainly nothing new in banking; it has been used for decades. However, what is driving the discussion about the impact of artificial intelligence now is the emergence of Large Language models like ChatGPT. This is the major change, not only in the corporate world, but also in everyday life.

The Large Language models are unlike other AI tools in that they have mastered language; we can communicate with them in ordinary language. Thus, technical knowledge is no longer a prerequisite for using artificial intelligence in life and work; instead, expressive ability and understanding of language are key. But the development of these models and research into them also vividly remind us how language itself is the true prerequisite for human society.

Theory of Mind: Getting Into the Minds of Others

Large Language models function in a different way from normal software because they evolve and change without the developers and operators necessarily foreseeing those changes. The ability to put oneself in the mind of another person has generally been considered unique to humans. This ability, known in psychology as “theory of mind,” refers to an individual’s ability to formulate a “theory” about what another person’s mental world is like. This ability is fundamental to human society; without it, it’s hard to see how any society could thrive. Here’s a simple puzzle of this kind:

“There is a bag filled with popcorn. There is no chocolate in the bag. Yet the label on the bag says “chocolate” and not “popcorn.” Sam finds the bag. She had never seen the bag before. She cannot see what is inside the bag. She reads the label.”

The question is, what does she think is in the bag? Of course, the right answer is that Sam thinks there’s chocolate in the bag, because that’s what the label says. When Michal Kosinski, adjunct Professor at Stanford University, tested last year whether the first language models could handle this task, the result was negative. GPT-1 and 2 both answered incorrectly. But then he tried the next generation of the model, GPT-3. And in 40% of cases, it managed this type of task. GPT-3.5 managed it in 90% of cases and GPT-4 in 95% of cases.1

Emergent Capabilities of Large Language Models

This capability came as a surprise, as nothing had been done to build theory of mind capability into the models. They simply acquired it on their own as they grew larger and as the volume of data they were trained on increased. That this could happen is based on the models’ ability to use language, says Kosinski.

Another example I stumbled upon myself by chance recently was when GPT-4 asked me, after I had posed a puzzle to it, whether I had tried to solve the puzzle myself. The models certainly ask questions all the time, that’s nothing new, they aim to get more precise instructions. But this question is of a different nature. I answered yes and also mentioned that this was the first time I had received a question of this kind from the model. “Yes, you are observant,” GPT-4 replied, “with this I am trying to make the conversation more natural.”

Does this new development mean that the artificial intelligence truly puts itself in the mind of others? Does it mean it thinks, that it has feelings, opinions, an interest in the viewpoints and experiences of others? Of course, we can’t draw that conclusion. But what this means is that the behavior of the models is becoming increasingly similar to how we use language when we interact with each other. In this sense, we could actually talk about the mind of an AI model, just as we use theory of mind to infer about the minds of other humans.

The Power of Language

The language models draw our attention to the importance of language and how it underpins our societies and our existence. We now have a technology that is increasingly adept at using language, which has the advantage of possessing vastly more knowledge than any individual could possibly acquire in a lifetime and which can perform tasks much faster. We can use this technology to greatly enhance our own productivity, our reasoning, and our decisions if we use it correctly. This way, we can use it to gain more leisure time and improve our quality of life.

The comparison to the discovery of electricity is apt. Some might even want to go further and liken this revolution to the advent of language itself, which could be supported by pointing to the spontaneous capabilities of the models, such as theory of mind, which they achieve through nothing but the very ability to use language. What happens then if they evolve further than us, and could that possibly happen?

The fact that artificial intelligence has mastered language is a revolution that will lead to fundamental changes in society. The challenge we now face, each and every one of us, is to use it in a structured way, to our advantage, and avoid the pitfall of outsourcing our own thinking and decisions to it. The best way to do this is to enhance our own understanding of language, our expressive ability, and our critical thinking skills.

 

Thorsteinn Siglaugsson

 


  1. Kosinski, Michal: Theory of Mind May Have Spontaneously Emerged in Large Language Models, Stanford 2023. https://stanford.io/4aQosLV

Thorsteinn Siglaugsson is a Icelandic economist, consultant and writer. Chairman of the Icelandic Free Speech Society. Author: "From Symptoms to Causes" (Amazon). Regular contributor to The Daily Sceptic, Conservative Woman and Brownstone Institute. Siglaugsson also writes on Substack.

Are we building a new Tower of Babel?

The future of AI

The advent of artificial intelligence, which offers effortless translation, poses a risk to the authenticity and depth of human connection.

Published 10 March 2024
– By Thorsteinn Siglaugsson

This is an opinion piece. The author is responsible for the views expressed in the article.

The other day I attended a conference hosted by an international software company. The presenters came from various countries, all speaking in English. English is the corporate language. But none of them were native English speakers, which was obvious. It made me think about how strange it is when people spend most of their time communicating in a language they don’t fully know and can never really master. All the nuances are lost, all the linguistic creativity, the ambiguity, the unspoken words, the hidden sarcasm, and the secret humour.

Shortly after, I was coaching one of my students, a German software specialist who speaks quite decent English, but still, we spent a lot of time figuring out exactly what he meant by a particular paragraph. I asked him if he couldn’t just tell me in German. “Do you speak German?” he asked. “Well, I learned it in high school,” I replied, “but it might be a stretch to say that means I know German.” Then we just laughed. Eventually, we managed to resolve the issue, in English, which neither of us speaks perfectly. Did we understand the paragraph in the same way? Surely not.

Being able to speak any language, but understand none

But now we have artificial intelligence. And whether we like it better or worse, AI is considerably better at English than almost every non-native speaker, and indeed a large part of those as well. The style is admittedly flat, but the same applies when people express themselves in a language they don’t fully master. Soon, we can expect technology to have generally reached the level where in phone calls and remote meetings we can simply speak our own mother tongue, and let artificial intelligence instantly translate the content into any other language.

AI is considerably better at English than almost every non-native speaker, and indeed a large part of those as well.

The Swede working in the international software company then simply speaks Swedish to the Ukrainian or Frenchman, and they just hear the Ukrainian or French version and respond in their own language. And when Elon Musk or those others now working hard to develop brain chips have progressed further, it might even be enough to press a button on the remote brain control we will soon all carry, to switch languages and speak French, Ukrainian, Swahili, or Hindi as needed. But of course, without understanding a word of what comes out of our mouths.

OpenAI’s new model ‘Sora’ creates ultra-realistic videos from text

The future of AI

Published 17 February 2024
– By Editorial Staff
The image above is entirely generated from an arbitrary text description.

On Thursday, OpenAI presented Sora, a new model that can generate high-definition videos of up to one minute based on text descriptions. However, Sora, which means “sky” in Japanese, will not be available to the public in the near future. Instead, it is being released to a small group of researchers and academics to evaluate the risks of misuse.

“Sora is able to generate complex scenes with multiple characters, specific types of motion, and accurate details of the subject and background”, OpenAI writes on its website. “The model understands not only what the user has asked for in the prompt, but also how those things exist in the physical world.”

One of the examples of videos generated by Sora shows a couple walking through a snowy Tokyo, while cherry blossoms and snowflakes swirl around them.

OpenAI claims that the model works thanks to a “deep understanding of language”, which allows it to correctly interpret text descriptions. Yet, like almost all AI-based image and video generators, Sora is not perfect. In one example, people and streets are completely missing from the video even though the description mentions a Dalmatian dog looking out of a window and people “walking and cycling along the canals”.

OpenAI also warns that the model may have difficulty understanding causal relationships – for example, it may generate a video of a person eating a cookie, but the cookie does not receive any bite marks.

Sora is not the first text-to-video model on the market. Other companies, including Meta, Google and Runway, have either hinted at plans for or launched similar tools. However, no other tool can yet generate 60-second videos. Sora also generates entire videos at once, instead of assembling them frame by frame like other models. This ensures that subjects in the video remain themselves even when they temporarily disappear from view.

Causing concern and distaste

The emergence of text-to-video tools has raised concerns that they can be used to more easily create realistically fake videos. Generative AI has also been criticised by artists and creative professionals who worry about the technology being used to replace jobs and use copyrighted material.

OpenAI says it is working with experts in areas such as “disinformation, hateful content and bias” to test the tool before its public release. The company is also developing tools to detect videos generated by Sora and include metadata in the generated videos for easier detection.

OpenAI declined to reveal to The Times how Sora had been trained, except to say that it used both “publicly available videos” and videos licensed from copyright holders.