Computer Vision in Nutritional Labeling

Expert-defined terms from the Professional Certificate in AI in Nutrition and Dietetics course at Stanmore School of Business. Free to read, free to share, paired with a globally recognised certification pathway.

Computer Vision in Nutritional Labeling

Artificial Intelligence (AI) #

The simulation of human intelligence processes by machines, especially computer systems. These processes include learning (the acquisition of information and rules for using the information), reasoning (using rules to reach approximate or definite conclusions), and self-correction.

Computer Vision #

A field of AI that focuses on enabling computers to interpret and understand the visual world. Through computer vision, computers can gain high-level understanding from digital images or videos. It involves the development of a theoretical and algorithmic basis to achieve automatic understanding of visual data.

Deep Learning #

A subset of machine learning that is based on artificial neural networks with representation learning. It can process a wide range of data resources, requires less data preprocessing by humans, and can often produce more accurate results than traditional machine learning approaches.

Digital Image #

A digital representation of visual data, in which the information is stored as a grid of pixels, with each pixel storing a value that represents the color and intensity of that point in the image.

Ground Truth #

The true or accepted value or category for a given data point. In the context of computer vision, ground truth refers to the manually annotated or labeled data used to train and evaluate machine learning models.

Image Annotation #

The process of labeling or marking up digital images with information about the objects and regions within the image. Image annotation can include bounding boxes, segmentation masks, or class labels.

Image Classification #

A computer vision task that involves categorizing an input image into one of several predefined classes based on the visual content of the image.

Image Segmentation #

The process of partitioning an image into multiple segments or regions, with the goal of separating different objects or parts of an image from one another. This process can be used for object detection, semantic segmentation, and instance segmentation.

Instance Segmentation #

A computer vision task that involves both object detection and image segmentation, with the goal of identifying and segmenting individual instances of objects within an image.

Machine Learning #

A subset of AI that focuses on the development of algorithms that allow computers to learn and improve from experience without being explicitly programmed.

Neural Network #

A computational model that is inspired by the structure and function of the human brain. Neural networks consist of interconnected nodes or "neurons," and they can be used for a wide range of AI tasks, including computer vision, natural language processing, and time series prediction.

Object Detection #

A computer vision task that involves identifying and locating objects within an image, typically by drawing bounding boxes around the objects.

Optical Character Recognition (OCR) #

A technology that enables the conversion of different types of documents, such as scanned paper documents, PDF files or images captured by a digital camera into editable and searchable data.

Pixel #

The smallest unit of a digital image, representing a single point in the image. Pixels are arranged in a grid, and each pixel stores a value that represents the color and intensity of that point in the image.

Region #

based Convolutional Neural Networks (R-CNN): A family of deep learning models for object detection. R-CNN first generates region proposals, then extracts features from each region using a convolutional neural network, and finally classifies the regions and refines the bounding boxes.

Semantic Segmentation #

A computer vision task that involves partitioning an image into multiple segments or regions, with the goal of assigning a class label to each region.

Transfer Learning #

A technique in machine learning where a pre-trained model is used as the starting point for a new model, typically by fine-tuning the pre-trained model on a new dataset. Transfer learning can help improve the performance of machine learning models, especially when the new dataset is small.

YOLO (You Only Look Once) #

A real-time object detection system that is based on a single convolutional neural network. YOLO treats object detection as a regression problem, and it is able to perform object detection in real-time, making it well-suited for applications such as video surveillance and autonomous vehicles.

Convolutional Neural Networks (CNNs) #

A type of neural network that is commonly used for image classification and object detection tasks. CNNs are designed to take advantage of the spatial structure of images, and they consist of convolutional layers, pooling layers, and fully connected layers.

Faster R #

CNN: An improvement over R-CNN, Faster R-CNN uses a region proposal network (RPN) to generate region proposals, making it faster and more efficient than the original R-CNN.

RetinaNet #

A single-stage object detection model that uses a neural network to predict bounding boxes and class labels simultaneously. RetinaNet is based on a novel loss function called the focal loss, which helps to address the class imbalance problem in object detection.

Single Shot Detector (SSD) #

A single-stage object detection model that is based on a feedforward neural network. SSD is faster than two-stage object detection models such as R-CNN and Faster R-CNN, and it is well-suited for real-time applications.

Data Augmentation #

A technique used to increase the size and diversity of a training dataset by applying various transformations to the existing data, such as rotation, scaling, and flipping. Data augmentation can help improve the performance of machine learning models by reducing overfitting and increasing the robustness of the model.

Batch Normalization #

A technique used in deep learning to normalize the activations of each layer in a neural network. Batch normalization helps to reduce the internal covariate shift, which can improve the convergence speed and stability of the training process.

Dropout #

A regularization technique used in deep learning to prevent overfitting. Dropout randomly sets a fraction of the output units of a layer to zero during training, which helps to prevent the co-adaptation of the units and improves the generalization performance of the model.

Fully Connected Layer #

A type of layer in a neural network where every neuron in the layer is connected to every neuron in the previous layer. Fully connected layers are typically used in the final stages of a neural network for classification tasks.

Pooling Layer #

A type of layer in a convolutional neural network that reduces the spatial dimensions of the input feature maps. Pooling layers help to increase the invariance of the network to small translations and distortions in the input data.

ReLU (Rectified Linear Unit) #

A type of activation function used in neural networks that outputs the input directly if it is positive, and outputs zero otherwise. ReLU helps to introduce non-linearity into the network and is computationally efficient.

Softmax #

A type of activation function used in the output layer of a neural network for multi-class classification tasks. Softmax outputs a probability distribution over the classes, where the sum of all the probabilities is equal to one.

Concatenation #

An operation in deep learning that combines the outputs of two or more layers by concatenating them along a specific dimension. Concatenation helps to increase the representational capacity of the network and can be used for feature fusion.

Max Pooling #

A type of pooling layer in a convolutional neural network that selects the maximum value from a set of neighboring input values. Max pooling helps to increase the invariance of the network to small translations and distortions in the input data.

Average Pooling #

A type of pooling layer in a convolutional neural network that computes the average value from a set of neighboring input values. Average pooling helps to reduce the spatial dimensions of the input feature maps and can help to prevent overfitting.

Global Average Pooling #

A type of pooling layer in a convolutional neural network that computes the average value over the entire spatial extent of the input feature maps. Global average pooling helps to reduce the number of parameters in the network and can help to prevent overfitting.

Region of Interest (RoI) Pooling #

A type of pooling layer in a convolutional neural network that is used in region-based object detection models such as R-CNN and Faster R-CNN. RoI pooling helps to extract fixed-length feature vectors from variable-sized regions of interest, which are then used for object detection

May 2026 cohort · 29 days left
from £99 GBP
Enrol