XClose

Teaching & Learning

Home
Menu

Introduction to Generative AI

Explore Generative AI (GenAI), how it works and its strengths and weaknesses. Find guidance to help ensure that use of GenAI across is effective, ethical and transparent.

On this page:

What is GenAI (such as ChatGPT)?

GenAIis an Artificial Intelligence (AI) technology that automatically generates content in response to written prompts. The generated content includes texts, software code, images, videos, and music.  

GenAI is trained using datafrom webpages, social media conversations and other online content. It generates its outputs by statistically analysing the distribution of words or pixels or other elements in the data that it has ingested and identifying and repeating common patterns (for example, which words typically follow which words).  

There are many other types of AI applications, that do not involve GenAI, that are having an impact on teaching and learning. These other types of AI (known as ‘teaching and learning with AI’ or ‘AIED’) will be addressed in more detail as these web pages develop. 

𳾱𳾲: 

  • GenAI looks accurate… but it isn’t
  • GenAI looks intelligent… but it isn’t
  • GenAI looks as if it understands… but it doesn’t.

Microsoft CoPilot, also known as Bing Chat Enterprise

staff and students canaccess Microsoft Copilot, which can be used for both text and image generation. With commercial data protection, this is intended as a more secure alternative thanother GenAI services. If you wish to use GenAI, then this is the safestwayto do so.

If you're logged into Microsoft Copilotwith your credentials, what goes in– and what comes out – is not saved or shared, and your data is not used to train the models.

Find out more and how to access Microsoft CoPiloton .

Practical guidance on how educators can use Microsoft CoPilot is available on.


Text GenAI 

In response to a human-written prompt, text GenAI generates text that usually appears as ifa human has written it.

Yet, just like human-written texts, text GenAI outputs can be superficial, inaccurate, untrustworthy, and full of errors. 

"Large language models [which is the technology behind text GenAI] are the ultimate bullshitters because they are designed to be plausible (and therefore convincing) with no regard for the truth.” Associate Professor Carissa Véliz, University of Oxford

Despite appearances, text GenAI does not understand either the prompt written by the human or the text that it generates. 

Every time that we use a text GenAI tool, we need to consider its output from asceptical perspective. 

Examples of text GenAI tools:
  • (Google)
  • (OpenAI) 
  • (Anthropic) 
  • (HuggingFace) 
  • (Meta) 

Please note that does not recommend any of the tools in this list. Microsoft CoPilot is now available for students and staff. Read more about using CoPilot.

Examples of other GenAI tools built on top of GenAI tools:
  • (summarises and answers questions about submitted PDF documents) 
  • (aims to automate parts of researchers’ workflows, identifying relevant papers and summarising key information) 
  • (Google Chrome extension that gives ChatGPT Internet access, to enable more accurate and up-to-date conversations)
  • Microsoft has incorporated ChatGPT into its Bing search engineand is implementing ChatGPT across its Office portfolio.

Please note that does not recommend any of the tools in this list. Microsoft CoPilot is now available for students and staff. Read more about using CoPilot.


Image/video/music GenAI 

Image, video and music GenAI can generate outputs based on human-written prompts. Some can also respond to visual or musical prompts. 

Again, the appearance of image/video and music GenAI outputsmight appear novel. Howeverusually they are only complex combinations of the millions of images/videos/music that they have ingested during their training. 

On the one hand, this is how creativity often works. For example, Rock and Roll music combined ideas from R&B, gospel and country music. 

But importantly,Rock and Roll drew on ideas from the earlier works. Meanwhile, GenAI actually uses the earlier works in its outputs, and without the consent of the original creators.

Another issue raised by image GenAI is how difficult it can be to write an effective prompt. For example, the breakthroughAI image Théâtre D’opéra Spatial, took weeks of prompt writing and fine-tuning hundreds of images.  

Examples of image/video and music GenAI tools:
  • (OpenAI’s image GenAI tool)
  • (Stable Diffusion’s image GenAI tool)
  • (Image GenAI tool) 
  • (Video GenAI tool)
  • (Music GenAI tool)
  • (Music GenAI tool).

Please note that does not recommend any of the tools in this list. Microsoft CoPilot is now available for students and staff. Read more about using CoPilot.

Return to top


How does Generative AI work? 

Both text and image GenAI are based on a set of AI techniques that have been available to researchers for several years and have been built one on top of another.

Text GenAI 

Although the terms listed below are all often used in descriptions of GenAI, it isn’t necessary to understand exactly what they all mean. The most important things to note are the hierarchy of the technologies and their complexity. 

ChatGPT (and other text GenAI) is a type of:

Generative Pre-trained Transformer(GPT: an advanced type of LLM) 

which is a type ofLarge Language Model(LLM: a massive computer-based representation of examples of natural language) 

which is a type ofGeneral-purpose Transformer(an ANN language processor) 

which is a type ofArtificial Neural Network(ANN: an ML approach inspired by how the human brain works, its synaptic connections between neurons) 

which is a type ofMachine Learning(ML: an approach to AI that uses algorithms to automatically improve its performance from data) 

which is a type ofArtificial Intelligence. 


Issues around training a text GPT 

So that a text GenAI can generate text, it first has to be trained. This involves the tool being provided with and processing huge amounts of data scraped from the internet and elsewhere. It is reported, but not confirmed by OpenAI, that the training of GPT4 involved a million gigabytes of data. Processing this data involves identifying patterns, such as which words typically go together (e.g.“Happy” is often followed by “Birthday”). 

Carbon footprint

Training a GPT requires huge amounts of power and indirectly generates huge amounts of carbon, with important consequences for climate change. For example, it is estimated that the training of GPT3 (the GPT used by the first version of ChatGPT made available to the public) consumed 1,287 megawatt hours of electricity and generated 552 tons of carbon dioxide, the equivalent of 123 cars driven for one year. 

Feedback loop

Another concern is that when future GPTs are trained, the data that they ingest are likely to include substantial amounts of text generated by previous versions of GPT. This self-referential loop might contaminate the training data and compromise the capabilities of future GPT models. 

Human costs

Once the text GenAI model is trained but before it is used, it is often checked and refined in a process known as Reinforcement Learning from Human Feedback (RLHF). In RLHF, text GenAI responses are reviewed and validated by human reviewers. These human reviewers ensure that the GenAI responses are appropriate, accurate, and align with the intended purpose. Sometimes the provider of the GenAI then sets up what are known as ‘guardrails’ to prevent the GenAI generating objectionable materials. 

In the development of ChatGPT, the RLHF reviewers mostly were workers in global south countries such as Kenya. Workers were paid less than $3 per hour to review the outputs of ChatGPT and identify any objectionable or nasty materials. This work has had a massive negative impact on many of those who were involved.


How a GPT generates text 

Once the GPT has been trained, generating a text response to a prompt involves the following steps: 

1. The prompt is broken down into smaller units (called tokens) that are input into the GPT. 

2. The GPT uses statistical patterns to predict likely words or phrases that might form a coherent response to the prompt. 

  • The GPT identifies patterns of words and phrases that commonly co-occur in its prebuilt large data model (which comprises text scraped from the Internet and elsewhere). 
  • Using these patterns, the GPT estimates the probability of specific words or phrases appearing in a given context. 
  • Beginning with a random prediction, the GPT uses these estimated probabilities to predict the next likely word or phrase in its response. 

3. The predicted words or phrases are filtered through what are known as ‘guardrails’ to remove any offensive content. 

4. Steps 2 to 3 are repeated until a response is finished. The response is considered finished when it reaches a maximum token limit or meets predefined stopping criteria. 

5. The response is post-processed to improve readability by applying formatting, punctuation, and other enhancements (such as beginning the response with words that a human might use, such as “Sure,” or “Certainly,” or “I’m sorry”). 


Image and Music GenAI 

Image GenAI and music GenAI use a different type of ANN known as Generative Adversarial Networks (GANs) which can also be combined with Variational Autoencoders. Here, we focus on image GANs.  

GANs have two parts (two ‘adversaries’), the ‘generator’ and the ‘discriminator’. The generator creates a random image in response to the human-written prompt, and the discriminator tries to distinguish between this generated image and real images. The generator then uses the result of the discriminator to adjust its parameters, in order to create another image.  

This process is repeated, possibly thousands of times, with the generator making more and more realistic images that the discriminator is increasingly less able to distinguish from real images.

For example, a successful GAN trained on a dataset of thousands of landscape photographs might generate new but unreal images of landscapes that are almost indistinguishable from real photographs.  

Meanwhile, a GAN trained on a dataset of popular music (or even music by a single artist) might generate new pieces of music that are very similar to but still different from the structure and complexity of the original music. 

Return to top


What are the strengths and weaknesses of GenAI?

Some of the benefits of AI and why we shouldcritically evaluate its outputs.

Strengths

GenAI canproduce diverse and seemingly original outputs, creating content that may not have been seen before based on patterns in the data they were trained on.

GenAI can process and interpret human language, allowing them togenerate contextually relevant responses to user prompts.

GenAI can process andgenerate text in multiple languages.

GenAI can be fine-tuned for various tasks and domains, making them widely applicable(e.g., chatbots, content generation, and language translation).

GenAI can learn patterns and representations from vast amounts of data, enabling them tocapture nuances in languageand generate outputs based on the patterns they've seen during training.

GenAI models canremember previous interactions, which results in more coherent and relevant conversation experiences for users.

GenAI cangenerate responses quickly, allowing for rapid interactions and real-time applications. 

Weaknesses

GenAI can generate information that appears factual but is inaccurate.

It’s potentially dangerous that GenAI models appear to understand the content that they use and generate, but inreality they do not understand it. This could lead users to have misplaced trust in the GenAI output.

GenAI output imitates or summarises existing content - mostlywithout the permission of the Intellectual Property owners -but can give the appearance of creativity.

GenAI can produce contentthat is morally and ethically troubling, and its use can raisemoral and ethical issues.

Training and running GenAI models can require significant computational and power resources.

The outputs of GenAI need to be moderated to establish ‘guardrails’ that prevent it generating inappropriate or offensive outputs. For ChatGPT, this was undertaken by poorly paid workers in Kenya, many of whom suffered mental health issues because of the disturbing generated output that they had witnessed.

GenAI can be used to automatically generate fake news and deep fakes.

GenAI is contributing to the digital divide. It relies on huge amounts of data and massive computing power, which is mostly only available to the largest international technology companies and a few economies. This means that the possibility to create and control GenAI is out of reach of mostpeople, especially those in the Global South.

While we understand broadly how GenAI works, because of its complexity it is usually impossible to know why it produces particular outputs.

The output of GenAI is flooding the internet. This poses an interesting recursive risk for future GPT models. These themselves will be trained on online content that earlier GPT models have created (including all its biases and errors).

GenAI tends to output standard answers that replicate the values of the creators of the data used to train the models. This may constrain the development of plural opinions and further marginalize marginalized voices.

Return to top


Further information

Find a broad range of commentaries and resources to inform your own views.Note that including a link on this page does not suggest that supports or endorses the views expressed.

Universities

Arizona State University (March 2023):

Russell Group (July 2023):

University of Cambridge (May 2023):

University of Leeds (undated):

Monash University (updated):

University of Sydney (updated):

Peter Bryant (University of Sydney Associate Dean Education) (January 2023):

Deakin University (March 2023):

Imperial College London (March 2023):

Academic and related organisations

QAA (January 2023):

National Centre for AI in Tertiary Education (JISC) (January 2023):

HEPI (May 2023):

QAA (May 2023):

National Centre for AI (JISC) (May 2023):

Sensemaking, AI, and Learning (SAIL) (May 2023):

Wolfram (February 2023):

National Centre for AI in Tertiary Education (JISC) (March 2023):

UNESCO (2023):

SEDA (March 2023): (recordings)

QAA (updated): .

Media

Business Insider (January 2023):

TechCrunch (February 2023):

Feedback Fruits (December 2022):

Insider (August 2023):

New York Post (July 2023):

BBC (July 2023):

Reuters (April 2023):

Times Higher Education (July 2023):

Wonkhe (June 2023):

Wonkhe (July 2023):

Bounded Regret (June 2023):

FT (May 2023):

Educsause Review (April 2023):

Inside Higher Ed (April 2023):

Wonkhe (April 2023):

Rachel Arthur Writes (April 2023):

Wonkhe (April 2023):

MIT Technology Review (April 2023):

Jim Dickinson (Wonkhe) ():

OpenAI (undated):

Times Higher Education (February 2023):

The Chronicle (March 2023):

#LTHEchat (March 2023):

The Conversation (February 2023): .

ZDNet (February 2023): .

Return to top