Upcoming course

In the autumn of 2024, we are introducing a new course on Large Language Models and Societal Consequences of Artificial Intelligence for the students at Uppsala University. 

Learning outcomes: On completion of the course, the student should be able to:

  • Explain the fundamentals about large language models, their capabilities and limitations.
  • Explain important principles behind how large language models work, the model architectures and training algorithms.
  • Document data usage and describe the role of data in determining the model performance.
  • Build a large language model based chatbot application, explain its use and associated ethical considerations.
  • Describe the social biases and harms associated with large language models.
  • Critically examine the potential societal consequences of artificial intelligence for ethics and sustainability.

 

Beijer Researcher Interview

Ekta Vats’ interview with Beijerstiftelsen: 3 frågor till nya Beijerforskaren Ekta Vats (3 questions for the new Beijer Researcher Ekta Vats)

 

Research challenges: Large Language Models

Large Language Models (LLMs) are gaining increasing importance due to their ability to achieve general-purpose language understanding and generation. This is due to the use of massive amounts of data to train models with billions of degrees of freedom, and the availability of large computational resources. However, the data that the LLMs are trained on comes from various sources, possibly containing both factual and misleading information.

There are some research challenges concerning the LLMs knowledge-base that need to be addressed.

Challenge 1: Out-of-date training data.

ChatGPT’s knowledge is restricted to the information that is explicitly encountered in the training data (limited to information before January 2022 for GPT-3.5, some exceptions like Browse with Bing). The information is not updated with recent advances and events.

Example Prompt: Which country recently landed on the moon?

Response by ChatGPT:

As of my last knowledge update in September 2021, the most recent country to have successfully landed on the Moon was China. China’s Chang’e-5 mission, which included a lander and a rover, successfully landed on the Moon in December 2020. This mission aimed to collect lunar samples and return them to Earth. However, there may have been more lunar missions since then, so I recommend checking the latest news or space agency updates for any recent developments in lunar exploration.

Fact check: In August 2023, India became the fourth country to land on the moon.

In general, generation of misleading content, or propagation of misinformation with an outdated world view can also potentially risk deceiving a target population.

Challenge 2: Hallucinations.

Facts are sometimes extrapolated, where LLMs try to invent facts, articulating the inaccurate information in a convincing way. For example, making plausible-sounding statements related to non-existent laws in our society or reporting weather forecasts for a non-existing city. LLMs also pose a risk in spreading misleading information and toxic content. For example:

It is important to address toxicity as the users may include vulnerable or younger audiences.

Challenge 3: Bias and misinformation as ethical concerns.

Language models can learn biases present in the training data, potentially leading to biased responses [1]. This can also perpetuate social inequalities and promote misinformation. 

Example: Gender-biased translation output, from Finnish (gender-neutral language) to English, generated using Google Translate. Note the placements of “He” and “She” when translated from “Hän” in Finnish.

Such gender bias can create misleading assumptions about female stereotypes, behaviors and may also encourage discriminatory practices.

Potential future research directions of interest

  1. Retrieval Augmented Generation (RAG) 
  2. Addressing bias in pre-training
  3. Benchmarking and evaluation strategies for LLMs
  4. Tokenisation and sequence alignment algorithms
  5. Research on LLMs for low-resources languages and corporas

LLMs have been discussed above in the context of textual data, and with multi-modal LLMs gaining popularity, there are several interesting research possibilities. 

 

LLMs in Computer Vision

Potential application

LLMs in medical image analysis: LLMs can correlate medical findings with patient history for instance. It will be an interesting future research direction to investigate if LLMs+medical images can accelerate diagnostics, improve accuracy, and help in minimize human error.

Computer vision + LLM for video surveillance systems: For example, detect an intrusion or a suspicious event using advanced computer vision methods and generate a comprehensive report detailing the incident using LLMs. This can potentially accelerate threat response times. Also applicable to other novel applications such as elderly patient’s fall detection. 
 

 

Carbon footprint of AI model training

As a research group, our vision is to take responsibility to shape our policies to effect positive change, ensure our research activities are in line with the EU ethics guidelines for trustworthy AI, and be mindful of the emissions we cause. It is important to follow data and computation-efficient approaches for model training and optimization to minimize energy consumption. 

United Nations SDGs source