Skip to main content

Glossary and FAQ of tems used

This page will be updated from time to time. Last update: 26th May


AI

Artificial intelligence (AI) is a branch of computer science that aims to create machines and systems that can perform tasks that normally require human intelligence and abilities. AI can be applied to various domains and problems, such as speech recognition, computer vision, natural language processing, robotics, gaming, healthcare, and more. AI systems can learn from data and experience, reason and plan, communicate and interact, and perceive and understand their environment.


AI can be classified into different types based on its capabilities and goals. Some common types are:


- Narrow AI: This type of AI is designed to perform a specific task or function, such as playing chess, recognizing faces, or recommending products. Narrow AI systems are often based on machine learning techniques that enable them to learn from data and improve their performance over time. However, they cannot generalise beyond their domain or task and lack common sense or general intelligence.


- General AI: This type of AI is the ultimate goal of AI research, as it aims to create machines that can perform any intellectual task that a human can do. General AI systems would have human-like intelligence and abilities, such as reasoning, learning, creativity, and self-awareness. However, this type of AI is still a theoretical concept and does not exist yet.


- Super AI: (or ASI) This type of AI is an extension of general AI, as it refers to machines that can surpass human intelligence and capabilities in all aspects. Super AI systems would have superior knowledge, skills, speed, memory, and creativity than humans. Some people believe that super AI could pose an existential threat to humanity if not aligned with human values and goals.


Using Consensus provides the definition of AI as:

A complex concept involving algorithms simulating human intelligence, socio-technological apparatuses, and social actor characteristics, with varying emphasis on technical functionality or human-like thinking depending on the context.


Alternatively

It can be argued that AI exists as a machine that can undertake interpolation, and extrapolation tasks, but exhibits no creativity, no consciousness, no dreaming, never has a sense of confidence, purpose or achievement.


Algorithms (for machine learning)


What is AGI?


AGI stands for Artificial General Intelligence, which is a type of artificial intelligence that can perform any intellectual task that a human or an animal can do. Unlike narrow AI, which is designed to solve specific problems, AGI aims to have general cognitive abilities that can adapt to any situation or goal. AGI is a major goal of some artificial intelligence research, but it has not been achieved yet. Some of the challenges of creating AGI include defining and measuring intelligence, replicating human common sense and creativity, and ensuring ethical and safe outcomes.


Cognitive blindness 

refers to a failure to perceive or understand information that is present due to various reasons. The reasons can include anxiety, lack of attention, or insufficient knowledge. There are different types of cognitive blindness, such as metacognitive blindness, which occurs when individuals fail to recognize their own cognitive processes. Cognitive blindness can also occur in relation to specific topics, such as AIDS awareness. Inattentional blindness is another type of cognitive blindness that occurs when people fail to notice unexpected objects when their attention is focused elsewhere. However, it is important to note that cognitive blindness is not always a permanent condition and can be addressed through interventions such as publicizing knowledge and building self-prevention awareness

GPT

Generative Pre-trained Transformer architecture. The model has been trained on a diverse range of internet text, allowing it to generate human-like text in response to prompts given to it. A useful video is below that explains much about training LLMs etc





Neural Network


A neural network is a computational model that simulates the functioning of biological neurons and their connections. Neural networks are composed of artificial neurons, also called nodes, that receive inputs from other nodes or external sources, process them using a mathematical function, and produce an output that can be transmitted to other nodes or used as a final result. Neural networks can learn from data and adjust their weights and biases accordingly, using various learning algorithms. Neural networks are widely used in artificial intelligence applications, such as speech recognition, image analysis, natural language processing, and adaptive control. Neural networks can perform complex tasks that are difficult to solve using conventional programming or statistical methods.



Machine Learning

Machine learning is a field of inquiry that studies how computer systems can learn from data and improve their performance on various tasks without being explicitly programmed. Machine learning is a subfield of artificial intelligence, which aims to create machines that can imitate intelligent human behaviour. Machine learning algorithms use mathematical models and statistical methods to analyse data, identify patterns, and make predictions or decisions. Machine learning algorithms can be classified into different types based on the nature of the data and the learning process, such as supervised learning, unsupervised learning, semi-supervised learning, reinforcement learning, and deep learning. Machine learning has many applications in diverse domains, such as medicine, computer vision, natural language processing, robotics, speech recognition, agriculture, and data mining. Machine learning is also closely related to other fields of study, such as computational statistics, mathematical optimization, and neural networks. Machine learning is an active and growing area of research and innovation that has the potential to transform many aspects of human society.


What is ChatGPT actually doing?

ChatGPT is a conversational AI model developed by OpenAI based on the Generative Pretrained Transformer 3 (GPT-3) architecture. The model has been trained on a diverse range of internet text, allowing it to generate human-like text in response to prompts given to it. ChatGPT can answer questions, converse on a variety of topics, and generate creative writing pieces. ChatGPT is designed to interact in a dialogue format, which makes it possible for it to answer follow-up questions, admit its mistakes, challenge incorrect premises, and reject inappropriate requests. ChatGPT is part of the broader field of artificial intelligence known as natural language processing (NLP), which seeks to teach computers to understand and interpret human language. ChatGPT is built using a deep learning architecture called the Transformer, which enables it to learn patterns in language and generate text that is coherent and human-like. ChatGPT is one of the most advanced language models available today and has the potential to revolutionise the way we interact with computers and digital systems. LLM A large language model (LLM) is a type of artificial intelligence (AI) algorithm that uses deep learning techniques and massively large data sets to understand, summarise, generate and predict new content. LLMs are trained on large quantities of unlabelled text using self-supervised learning, which means they learn from the data itself without human intervention or guidance. LLMs typically have billions of parameters, which are the variables that determine how the model processes the input and produces the output. LLMs use a neural network architecture called transformer, which enables them to capture long-range dependencies and complex relationships among words and sentences.


LLMs emerged around 2018 and have shown remarkable performance on a wide range of natural language processing (NLP) tasks, such as question answering, text summarization, text generation, sentiment analysis, machine translation and more. LLMs are also able to demonstrate general knowledge about the world and memorise facts from the data they are trained on. LLMs are sometimes referred to as foundation models, because they serve as the basis for further optimization and specialisation for specific domains and applications. However, LLMs also pose some challenges and risks, such as ethical, social and environmental implications. For example, LLMs may generate inaccurate, biassed or harmful content that can mislead or harm users or society. LLMs may also consume a lot of energy and resources during training and inference, which can have a negative impact on the environment. Moreover, LLMs may exhibit unpredictable or emergent behaviours that are not intended by the designers or users, such as hallucinations or abilities that were not explicitly programmed into the model. Therefore, LLMs require careful evaluation, monitoring and regulation to ensure their safe and beneficial use. NLP NLP stands for natural language processing, which is a branch of artificial intelligence that deals with the interaction between computers and human languages. NLP aims to enable computers to understand, analyse, generate and manipulate natural language texts or speech. Some of the applications of NLP include machine translation, speech recognition, sentiment analysis, information extraction, text summarization, question answering and chatbots. NLP is a challenging and multidisciplinary field that requires knowledge and skills from linguistics, computer science, mathematics and statistics. NLP involves various tasks and subfields, such as: - Tokenization: splitting a text into smaller units called tokens, such as words or punctuation marks. - Morphology: analysing the structure and formation of words, such as stems, prefixes and suffixes. - Syntax: analysing the grammatical structure and rules of sentences, such as parts of speech and dependency relations. - Semantics: analysing the meaning and logic of words and sentences, such as synonyms, antonyms and entailment. - Pragmatics: analysing the context and purpose of language use, such as speech acts and implicatures. - Discourse: analysing the structure and coherence of longer texts or conversations, such as paragraphs and dialogues. - Phonetics: analysing the sounds and pronunciation of speech, such as vowels and consonants. - Phonology: analysing the patterns and rules of sounds in a language, such as stress and intonation. - Prosody: analysing the rhythm and melody of speech, such as pitch and tone. NLP relies on various techniques and methods to perform these tasks and subfields, such as: - Rule-based systems: using predefined rules and dictionaries to process natural language based on its structure and grammar. - Statistical methods: using mathematical models and algorithms to learn from data and make predictions based on probabilities and frequencies. - Machine learning: using artificial neural networks and other learning algorithms to automatically learn from data and improve performance based on feedback. - Deep learning: using advanced neural networks with multiple layers and complex architectures to perform complex natural language tasks with high accuracy. NLP is a rapidly evolving and expanding field that has many current challenges and future opportunities. Some of the challenges include: - Dealing with ambiguity, variability and diversity of natural language in different domains, genres, styles and dialects. - Handling noisy, incomplete or inconsistent data from various sources and formats, such as web pages, social media posts or speech recordings. - Ensuring robustness, scalability and efficiency of NLP systems in real-world scenarios with large-scale data and limited resources. - Ensuring reliability, validity and fairness of NLP systems in terms of their outputs, outcomes and impacts on users and society.

What is NLG?

Natural language generation (NLG) is a software process that produces natural language output from a data set. NLG is a subfield of artificial intelligence (AI) that aims to create human-like communication from structured or unstructured data. NLG has many applications and benefits for various domains and industries, such as education, business, journalism, healthcare, and entertainment. NLG can help automate tasks that involve writing or speaking, such as generating reports, summaries, headlines, captions, scripts, emails, chatbot responses, and more. NLG can also enhance the accessibility and usability of data by transforming it into natural language that is easy to understand and interact with. How does NLG work? NLG is a complex and multi-stage process that involves several steps and techniques. The exact steps may vary depending on the specific NLG system and its goals, but a general framework can be described as follows: - Content analysis: The first step is to analyse the data source and determine what information should be included in the output. This may involve filtering, selecting, and prioritising the relevant data points based on the purpose and context of the output. - Data understanding: The next step is to interpret the data and identify patterns, trends, insights, and relationships that can be expressed in natural language. This may involve using machine learning algorithms, such as classification, clustering, regression, or sentiment analysis. - Document structuring: The third step is to create a document plan that defines the structure and organisation of the output. This may involve choosing a genre, style, tone, and format for the output, as well as deciding how to segment and order the information into paragraphs, sentences, or bullet points. - Sentence aggregation: The fourth step is to combine the information into coherent sentences that convey the main message and avoid redundancy. This may involve using linguistic techniques, such as conjunctions, anaphora, ellipsis, or paraphrasing. - Grammatical structuring: The fifth step is to apply grammatical rules and conventions to ensure that the sentences are syntactically and semantically correct. This may involve using natural language processing (NLP) tools, such as parsers, taggers, or generators. - Language presentation: The final step is to generate the output in natural language based on a template or a format that suits the target audience and medium. This may involve using natural language understanding (NLU) tools, such as lexicons, ontologies, or dialog systems. What are the challenges of NLG? NLG is a challenging task that requires a high level of linguistic knowledge and computational skills. Some of the main challenges of NLG are: - Data quality: The quality of the output depends on the quality of the input data. If the data is incomplete, inaccurate, inconsistent, or noisy, it may affect the accuracy and reliability of the output. - Data diversity: The diversity of the data sources and formats may pose difficulties for integrating and processing them. For example, different data sources may have different schemas, vocabularies, or representations that need to be aligned and harmonised. - Data complexity: The complexity of the data may make it hard to extract meaningful information and insights from it. For example, some data may be too large, too sparse, too dynamic, or too ambiguous to be easily analysed and interpreted. - Language variability: The variability of natural language may make it hard to generate natural and fluent output that matches the expectations and preferences of the users. For example, different users may have different linguistic backgrounds, dialects, registers, or styles that need to be taken into account. - Language creativity: The creativity of natural language may make it hard to generate original and engaging output that captures the attention and interest of the users. For example, some outputs may require humour, emotion, personality, or storytelling skills that are difficult to emulate by machines. What are the future trends of NLG? NLG is a rapidly evolving field that has seen significant advances in recent years thanks to the development of new technologies and methods. Some of the future trends of NLG are: - Deep learning: Deep learning is a branch of machine learning that uses neural networks to learn from large amounts of data and perform complex tasks. Deep learning has been applied to various aspects of NLG, such as data understanding, sentence aggregation, grammatical structuring and language presentation and has shown promising results in improving the quality and diversity of the output .- Multimodal NLG: Multimodal NLG is an extension of NLG that involves generating output that combines multiple modes of communication such as text image audio or video Multimedia


Comments

Popular posts from this blog

The AI Dilemma and "Gollem-Class" AIs

From the Center for Humane Technology Tristan Harris and Aza Raskin discuss how existing A.I. capabilities already pose catastrophic risks to a functional society, how A.I. companies are caught in a race to deploy as quickly as possible without adequate safety measures, and what it would mean to upgrade our institutions to a post-A.I. world. This presentation is from a private gathering in San Francisco on March 9th with leading technologists and decision-makers with the ability to influence the future of large-language model A.I.s. This presentation was given before the launch of GPT-4. One of the more astute critics of the tech industry, Tristan Harris, who has recently given stark evidence to Congress. It is worth watching both of these videos, as the Congress address gives a context of PR industry and it's regular abuses. "If we understand the mechanisms and motives of the group mind, it is now possible to control and regiment the masses according to our will without their

Beware the Orca, the challenge to ChatGPT and Palm2 is here

  So Google's 'we have no moat' paper was correct. If you train an LLM wisely then it's cost effective and cheap to produce a small LLM that is able to compete or even beat established, costly LLMs, as Microsoft has just found. It's another excellent video from AI Explained, who goes through some of the training procedures, which I won't get into here. Orca, is a model that learns from large foundation models (LFMs) like GPT-4 and ChatGPT by imitating their reasoning process. Orca uses rich signals such as explanations and complex instructions to improve its performance on various tasks. Orca outperforms other instruction-tuned models and achieves similar results to ChatGPT on zero-shot reasoning benchmarks and professional and academic exams. The paper suggests that learning from explanations is a promising way to enhance model skills. Smaller models are often overestimated in their abilities compared to LFMs, and need more rigorous evaluation methods. Explana

What is happening inside of the black box?

  Neel Nanda is involved in Mechanistic Interpretability research at DeepMind, formerly of AnthropicAI, what's fascinating about the research conducted by Nanda is he gets to peer into the Black Box to figure out how different types of AI models work. Anyone concerned with AI should understand how important this is. In this video Nanda discusses some of his findings, including 'induction heads', which turn out to have some vital properties.  Induction heads are a type of attention head that allows a language model to learn long-range dependencies in text. They do this by using a simple algorithm to complete token sequences like [A][B] ... [A] -> [B]. For example, if a model is given the sequence "The cat sat on the mat," it can use induction heads to predict that the word "mat" will be followed by the word "the". Induction heads were first discovered in 2022 by a team of researchers at OpenAI. They found that induction heads were present in