AI Terminologies

Agents

An agent is a computer program or system that is designed to perceive its environment, make decisions and take actions to achieve a specific goal or set of goals. The agent operates autonomously, meaning it is not directly controlled by a human operator.

AGI

Abbreviation: Artificial General Intelligence

Artificial General Intelligence (AGI) refers to AI that can understand, learn, and apply intelligence across various tasks like a human. Unlike narrow AI, which is task-specific, AGI can perform any intellectual task. It includes reasoning, problem-solving, perception, and language understanding. AGI aims to replicate human cognitive abilities.

Annotation

Annotation is the process of labeling data—text, images, or audio—to train AI models. In text-based AI, it involves categorizing sentences, tagging parts of speech, or marking sentiment. Annotated datasets help AI learn patterns, context, and nuances. This process enhances model accuracy and understanding.

ASI

Abbreviation: Artificial Super Intelligence

Artificial Super Intelligence (ASI) refers to a stage of artificial intelligence that surpasses human intelligence across all fields, including creativity, general wisdom, and problem-solving capabilities. Unlike Artificial General Intelligence (AGI), which aims to match human cognitive abilities, ASI represents an AI that is vastly more advanced than the best human brains in practically every field, including scientific understanding, general knowledge, and social skills.

AI

Abbreviation: Artificial Intelligence

AI stands for artificial intelligence, which is the simulation of human intelligence processes by machines or computer systems. AI can mimic human capabilities such as communication, learning, and decision-making.

AI ethics

AI ethics refers to the issues that AI stakeholders such as engineers and government officials must consider to ensure that the technology is developed and used responsibly. This means adopting and implementing systems that support a safe, secure, unbiased, and environmentally friendly approach to artificial intelligence.

Algorithm

An algorithm is a sequence of rules given to an AI machine to perform a task or solve a problem. Common algorithms include classification, regression, and clustering.

API

Abbreviation: Application programming interface

An API, or application programming interface, is a set of protocols that determine how two software applications will interact with each other. APIs tend to be written in programming languages such as C++ or JavaScript.

Anthropomorphism

We use the term anthropomorphism to describe the habit of assigning human-like qualities to AI. While AI systems can imitate human emotions or speech, they don’t possess feelings or consciousness. We might interact with various AI models as if they were colleagues or thought partners, but in reality, they serve as tools for learning and resource development.

AI bias

AI bias can occur when systems reflect and even amplify existing biases in the data trained on, potentially leading to unfair outcomes.

AI governance

AI governance ensures that AI systems operate ethically and transparently, in the best interests of all stakeholders. This often involves developing guidelines or policies around the use and impact of AI.

AIAAS

Abbreviation: AI as a service

AIaaS is the outsourcing of AI capabilities or services through cloud-based platforms, enabling businesses to utilize AI without substantial upfront investment.

AutoML

Abbreviation: Automated machine learning

AutoML refers to the tools and techniques used to automate the application of machine learning. These might include data preprocessing, model selection, and hyperparameter tuning.

Big data

Big data refers to the large data sets that can be studied to reveal patterns and trends to support business decisions. It’s called “big” data because organizations can now gather massive amounts of complex data using data collection tools and systems. Big data can be collected very quickly and stored in a variety of formats.

Bias (in Gen AI)

There are 2 distinct ways of using the term ‘bias’ with regards to Generative AI.

  • It refers to systemic prejudice in AI outputs, often reflecting biases in training data, such as cultural, gender, racial, or socioeconomic biases, leading to unfair or stereotypical content.
  • In neural networks, bias is a parameter that adjusts output independently of inputs, helping models accurately represent complex patterns and improve performance.

Bot

Bot (short for robot) is a software application that automates tasks, from simple actions to complex decision-making. In Generative AI, bots can understand and generate language, respond to queries, or create content. Platforms like ChatGPT and Poe allow users to build custom bots (GPTs).

Chain-of-thought Prompting

Chain-of-thought prompting is when you use a series of intermediate reasoning steps to improve the accuracy and applicability of answers generated by LLMs (Bubeck et al., 2022).

Context window

The context window is the maximum number of tokens (words or parts of words) that an AI model can process and consider simultaneously when generating a response. It is essentially the “memory” capacity of the model during an interaction or task. Models with larger context windows can handle larger attachments/prompts/inputs and sustain “memory” of a conversation for longer (Fogarty, 2023).

Chat Bot

A chatbot is a software application designed to simulate conversation with human users, especially over the internet. It utilizes techniques from the field of natural language processing (NLP) and sometimes machine learning (ML) to understand and respond to user queries. Chatbots can range from simple, rule-based systems that respond to specific keywords or phrases with pre-defined responses, to more sophisticated AI-driven bots capable of handling complex, nuanced, and context-dependent conversations.

ChatGPT

ChatGPT is an AI language model tool, designed to process and generate human-like text, assisting with diverse academic tasks and communication. As of December 2023, there are 2 models – a free version, referred to at Chat GPT-3.5 and a paid-for version referred to as Chat GPT 4.

Neither version of ChatGPT has passed a Privacy Impact Assessment at UBC and as such cannot be required for use within a course.

Completions

Completions are the output produced by AI in response to a given input or prompt. When a user inputs a prompt, the AI model processes it and generates text that logically follows or completes the given input. These completions are based on the patterns, structures, and information the model has learned during its training phase on vast datasets.

Conversational or Chat AI

Conversational AI (Chat AI) enables machines to understand, process, and respond to human language naturally. It powers chatbots and virtual assistants for context-aware interactions. Using NLP, ML, and sometimes speech recognition, it facilitates human-like dialogue.

Cognitive computing

Cognitive computing is essentially the same as AI. It’s a computerized model that focuses on mimicking human thought processes such as pattern recognition and learning. Marketing teams sometimes use this term to eliminate the sci-fi mystique of AI.

Computer vision

Computer vision is an interdisciplinary field of science and technology that focuses on how computers can gain understanding from images and videos. For AI engineers, computer vision allows them to automate activities that the human visual system typically performs.

Data mining

Data mining is the process of sorting through large data sets to identify patterns that can improve models or solve problems.

Data science

Data science is an interdisciplinary field of technology that uses algorithms and processes to gather and analyze large amounts of data to uncover patterns and insights that inform business decisions.

Deep learning

Deep learning is a function of AI that imitates the human brain by learning from how it structures and processes information to make decisions. Instead of relying on an algorithm that can only perform one specific task, this subset of machine learning can learn from unstructured data without supervision.

Data privacy

Data privacy in AI is the practice of ensuring data privacy and protection. This is crucial in AI as models often require large amounts of data for training. The data can include confidential or sensitive information, like health records, financial information, or intellectual property. Breaches of data privacy could occur when not handled correctly. Therefore, strong data privacy practices when collecting, storing, and using data for AI training are essential.

Data augmentation

Data augmentation is a technique to increase the amount of training data by adding slightly modified copies of existing data, thereby improving the performance of the model.

Embedding

Embedding represents words, phrases, or documents as vectors in a high-dimensional space, capturing their meaning and relationships. These vectors help AI understand language by recognizing synonyms, analogies, and context. Generated through algorithms, embeddings improve AI’s ability to process and interpret text.

Emergent behavior

Emergent behavior, also called emergence, is when an AI system shows unpredictable or unintended capabilities.

Explainability/interpretability

Explainability/interpretability is the ability to understand and interpret the decisions made by an AI system. This is essential for trust and transparency, especially in industries like healthcare and finance.

Edge AI

Edge AI refers to running AI algorithms on end devices such as smartphones or Internet of Things devices, allowing data processing at the source and ensuring real-time processing and privacy.

Few-shot learning

Few-shot learning enables AI models to learn from a very small amount of training data. Unlike traditional models that require large datasets, it allows generalization from just a few examples. This is useful in fields with limited data, like rare languages or specialized research.

Fine-tune

Fine-tuning is the process of training a pre-trained AI model on a smaller, specific dataset to optimize it for particular tasks. It helps adapt general models to specialized domains by adjusting parameters while retaining broad knowledge. This enhances accuracy and relevance in targeted applications.

Generative AI

Generative AI refers to artificial intelligence systems that can generate new content—such as texts, images, audio, and video—in response to prompts by a user, after being trained on an earlier set of data. Platforms like Dall-E and NightCafe produce digital art and images that appear like photos, and tools like Synthesia allow users to generate videos from text, using AI avatars. Large Language Models (LLMs), on the other hand, generate text by predicting the next word based on patterns learned from vast amounts of data.

GPT

Abbreviation: Generative Pre-Trained Transformers

Generative Pre-trained Transformers (GPT) are AI models for natural language processing, built on the transformer architecture. Pre-trained on vast text data, they learn language patterns, context, and knowledge. Unlike search engines, GPT generates human-like responses rather than retrieving exact information.

Guardrails

Guardrails refers to restrictions and rules placed on AI systems to make sure that they handle data appropriately and don’t generate unethical content.

Generator

A generator is a component that learns to generate new data resembling its training data.

Generalization

Generalization is the ability of an AI model to apply knowledge learned from training data to unseen data in a relevant and accurate manner.

Hallucinations

Hallucinations in AI refer to incorrect or misleading outputs generated due to insufficient training data, incorrect assumptions, or biases. These errors highlight the need for verifying AI-generated content to prevent misinformation. Critical evaluation is essential for reliable AI use.

Hyperparameter

A hyperparameter is a parameter, or value, that affects the way an AI model learns. It is usually set manually outside of the model.

Inference

Inference is the process where a trained AI model applies its learned knowledge to new, unseen data to make predictions, decisions, or generate content. It is essentially the phase where the AI model, after being trained on a large dataset, is now being used in real-world applications. Unlike the training phase, where the model is learning from examples, during inference, the model is utilizing its learned patterns to perform the specific tasks it was designed for.

Image recognition

Image recognition is the process of identifying an object, person, place, or text in an image or video.

Large Language Model

Large Language Models (LLMs) are AI systems trained on vast text datasets to understand and generate human language. Using deep learning and transformer architectures, they handle tasks like translation, content generation, and sentiment analysis. LLMs excel in processing complex language patterns and contexts.

Limited memory

Limited memory is a type of AI system that receives knowledge from real-time events and stores it in the database to make better predictions.

Language models

LMs are types of AI models that can understand and generate human language. Trained on vast amounts of text data, LMs learn the statistical structure of human language to understand it.

Model

Models are the computational frameworks that enable Generative AI to learn patterns, process data, and generate outputs like text or images. Trained on large datasets, they recognize patterns and make predictions. Their architecture and training data determine their capabilities and accuracy.

Machine learning

Machine learning is a subset of AI that incorporates aspects of computer science, mathematics, and coding. Machine learning focuses on developing algorithms and models that help machines learn from data and predict trends and behaviors, without human assistance.

NLP

Abbreviation: Natural Language Programming

NLP is a field at the intersection of computer science, artificial intelligence, and linguistics, focused on enabling computers to understand, interpret, and generate human language in a way that is both meaningful and useful. It involves the development of algorithms and systems that can analyze, comprehend, and respond to text or voice data in a manner similar to how humans do.

Neural network

A neural network is a deep learning technique designed to resemble the human brain’s structure. Neural networks require large data sets to perform calculations and create outputs, which enables features like speech and vision recognition.

Overfitting

Overfitting occurs in machine learning training when the algorithm can only work on specific examples within the training data. A typical functioning AI model should be able to generalize patterns in the data to tackle new tasks.

Parameters

Parameters are internal variables in AI models that define their behavior and influence how inputs are processed into outputs. In neural networks, they include weights (importance of inputs) and biases (adjustments to outputs). Fine-tuning these parameters improves model accuracy. Larger models, like LLaMA2-70B, have billions of parameters for complex learning.

Prompt

A prompt is the input given to an AI model to initiate or guide its generation process. This input acts as a directive or a set of instructions that the AI uses to produce its output. Prompts are crucial in defining the nature, scope, and specificity of the output generated by the AI system. For instance, in a text-based Generative AI model like GPT (Generative Pre-trained Transformer), a prompt could be a sentence or a question that the model then completes or answers in a coherent and contextually appropriate manner.

Prompt Engineering

Prompt engineering in the context of Generative AI refers to the crafting of input prompts to effectively guide AI models, particularly those like Generative Pre-trained Transformers (GPT), in producing specific and desired outputs. This practice involves formulating and structuring prompts to leverage the AI’s understanding and capabilities, thereby optimizing the relevance, accuracy, and quality of the generated content.

Pattern recognition

Pattern recognition is the method of using computer algorithms to analyze, detect, and label regularities in data. This informs how the data gets classified into different categories.

Predictive analytics

Predictive analytics is a type of analytics that uses technology to predict what will happen in a specific time frame based on historical data and patterns.

Prescriptive analytics

Prescriptive analytics is a type of analytics that uses technology to analyze data for factors such as possible situations and scenarios, past and present performance, and other resources to help organizations make better strategic decisions.

Quantum computing

Quantum computing is the process of using quantum-mechanical phenomena such as entanglement and superposition to perform calculations. Quantum machine learning uses these algorithms on quantum computers to expedite work because it performs much faster than a classic machine learning program and computer.

RL

Abbreviation: Reinforcement Learning

Reinforcement Learning (RL) is a type of learning algorithm where an agent learns to make decisions by performing actions in an environment to achieve a certain goal. The learning process is guided by feedback in the form of rewards or punishments — positive reinforcement for desired actions and negative reinforcement for undesired actions. The agent learns to maximize its cumulative reward through trial and error, gradually improving its strategy or policy over time.

RAG

Abbreviation: Retrieval Augmented Generation

Retrieval Augmented Generation (RAG) enhances AI by combining retrieval-based and generative models. It first retrieves relevant information from a dataset, then uses it to generate informed responses. RAG allows AI to process private or specialized data without exposing it to external models.

Robustness

Robustness refers to the ability of an AI system to continue operating effectively under varying or challenging conditions, including handling new inputs or coping with adversarial attacks.

Semantic Network

Semantic Network is a graphical representation of knowledge, where nodes represent concepts and edges define their relationships (e.g., “is a type of” or “has a property of”). It helps AI understand context, hierarchies, and connections in natural language processing.

Self-attention

Self-attention is a mechanism used in transformer models that allows the model to assign importance (or attention) to different words in an input sequence when generating an output. This lets the model determine which words in a sentence are crucial for understanding the overall context.

Sentiment analysis

Also known as opinion mining, sentiment analysis is the process of using AI to analyze the tone and opinion of a given text.

Structured data

Structured data is data that is defined and searchable. This includes data like phone numbers, dates, and product SKUs.

Supervised learning

Supervised learning is a type of machine learning in which classified output data is used to train the machine and produce the correct algorithms. It is much more common than unsupervised learning.

Temperature

Temperature is a parameter that controls the randomness or creativity in a model‘s responses. When generating text, a higher temperature value leads to more varied and unpredictable outputs, while a lower temperature results in more conservative and expected responses.

Tokens

Tokens are the smallest units of data an AI processes, representing words, sub-words, or characters in NLP. Tokenization breaks text into these units for analysis. AI tools like ChatGPT measure processing costs based on token usage.

Training

Training is the process of teaching an AI model by exposing it to large datasets and adjusting its parameters to minimize errors. The model iterates, comparing predictions to desired outcomes, refining accuracy over time. This enables tasks like language understanding and generation.

Transformer

Transformer is a deep learning architecture that processes entire data sequences simultaneously using self-attention. It captures complex relationships in language, making it ideal for tasks like translation and text generation. The ‘T’ in GPT stands for Transformer.

Transformer models

Transformer models are a type of neural network architecture that have been highly successful in NLP tasks. These models process input data in parallel (as opposed to sequentially), making them more efficient.

Tuning

Tuning adjusts a pre-trained AI model to improve performance on a specific task or dataset. Unlike initial training, it refines parameters for better accuracy and relevance in a targeted application.

Training data

Training data is the information or examples given to an AI system to enable it to learn, find patterns, and create new content.

Transfer learning

Transfer learning is a machine learning system that takes existing, previously learned data and applies it to new tasks and activities.

Turing test

The Turing test was created by computer scientist Alan Turing to evaluate a machine’s ability to exhibit intelligence equal to humans, especially in language and behavior. When facilitating the test, a human evaluator judges conversations between a human and machine. If the evaluator cannot distinguish between responses, then the machine passes the Turing test.

Unstructured data

Unstructured data is data that is undefined and difficult to search. This includes audio, photo, and video content. Most of the data in the world is unstructured.

Unsupervised learning

Unsupervised learning is a type of machine learning in which an algorithm is trained with unclassified and unlabeled data so that it acts without supervision.

Voice recognition

Voice recognition, also called speech recognition, is a method of human-computer interaction in which computers listen and interpret human dictation (speech) and produce written or spoken outputs. Examples include Apple’s Siri and Amazon’s Alexa, devices that enable hands-free requests and tasks.

Zero-Shot Learning

Zero-shot learning enables AI to perform tasks it hasn’t been explicitly trained for by generalizing from prior knowledge. It identifies unseen categories using learned attributes and relationships. This allows AI to make inferences without direct examples.

Sankarnarayan. R
About the Author - Sankarnarayan. R

The founder and mastermind behind ColorWhistle is Sankarnarayan, a professional with over fourteen years of experience and a passion for website design services and digital marketing services. At ColorWhistle, our team has a wide range of skills and expertise and we always put our clients’ satisfaction first. This is what sets us apart from the competition – an eye for detail and the best website development services from the start to the completion of your project.

Leave a Reply

Your email address will not be published. Required fields are marked *

Ready to get started?

Let’s craft your next digital story

Our Expertise Certifications - ColorWhistle
Go to top
Close Popup

Let's Talk

    Sure thing, leave us your details and one of our representatives will be happy to call you back!

    Eg: John Doe

    Eg: United States

    More the details, speeder the process :)