Glossary

Glossary for streamlined workflow and boosted knowledge.

Action Recognition

Action recognition is the process of identifying and categorizing human actions or movements in videos or images, such as walking, running, or dancing, to enable computer systems to understand and respond to these actions automatically.

Explore more

Expand Arrow

Action Recognition

Action recognition is the process of identifying and categorizing human actions or movements in videos or images, such as walking, running, or dancing, to enable computer systems to understand and respond to these actions automatically.

Explore more

Expand Arrow

Active Learning

Active learning is a machine learning algorithm that gives users the ability to actively to label data points with the intended outputs. The algorithm randomly selects the data points to be labeled next from the unlabeled data pool.

Explore more

Expand Arrow

Active learning

Active learning is a machine learning algorithm that gives users the ability to actively to label data points with the intended outputs. The algorithm randomly selects the data points to be labeled next from the unlabeled data pool.

Explore more

Expand Arrow

Annotation

Annotation is the process of labeling your data to teach your deep learning model the outcome your want to predict. Generally, bounding boxes are used to train for object detection and polygons are used to train for instance segmentation.

Explore more

Expand Arrow

Annotation Format

Annotation format is the specific method to encode the annotation and to describe the bounding box’s size and position (COCO, YOLO, TXT, etc).

Explore more

Expand Arrow

Application Programming Interface (API)

An application programming interface is a mechanism that provides components to convey with other software within databases or applications. Companies can use it to assist digital transformation or an ecosystem. We use REST API to allow users to easily import their models into our platform.

Explore more

Expand Arrow

Attribute/attribute group

An attribute is the item of data that is utilized in machine learning, and the attribute groups define clusters of attributes to create the product’s additional information.

Explore more

Expand Arrow

Augmentation

Augmentations are good for dataset robustness. It allows users to enhance their existing dataset through positional augmentations or color space augmentation. These augmentation techniques enable the model to not lean on specific features while training.‍

Explore more

Expand Arrow

Automated Machine Learning (AutoML)

AutoML leads to automating the tasks to optimize the training models for application to the real world by themselves. It contains the whole process from loading a raw dataset to deploying the ML model.

Explore more

Expand Arrow

Backpropagation

Backpropagation is a two-stage training process of how neural networks improve themselves. It is an ML algorithm that adjusts the parameters from the error calculation of each neuron.

Explore more

Expand Arrow

COCO

COCO is an image dataset stored in the JSON format, gathering to compare different models’ performance and solve common object detection problems.‍

Explore more

Expand Arrow

Classification

Classification is a machine learning task where data is categorized into predefined classes or labels. The goal is to build a model that can predict the correct label for new, unseen data based on patterns and features learned from a training dataset. It's widely used in various applications, such as spam detection or image recognition.

Explore more

Expand Arrow

Clustering

Clustering is an unsupervised technique that groups similar instances according to similarity, and the data points will not include labels.

Explore more

Expand Arrow

Computer Vision

Computer Vision is the science of enabling computers to see and understand images and video. This is accomplished by developing algorithms that can make sense of visual content, for example detecting people or objects in an image or video, or being able to read road signs.‍

Explore more

Expand Arrow

Confusion Matrix

A confusion matrix is a table used in machine learning to evaluate the performance of a classification model. It summarizes the model's predictions by showing the true positive, true negative, false positive, and false negative counts, enabling the assessment of accuracy, precision, recall, and other metrics.

Explore more

Expand Arrow

Explore more

Expand Arrow

Foundation Models

Foundation models are pre-trained convolutional neural networks (CNNs) that have been trained on large image datasets. These models serve as a starting point for various computer vision tasks like object detection, image classification, and segmentation. They provide a foundation of learned features and patterns that can be fine-tuned for specific vision-related applications.

Explore more

Expand Arrow

Generative AI

Generative AI refers to artificial intelligence systems capable of generating data, content, or objects autonomously. These systems, often based on deep learning models like GANs, can produce images, text, audio, or other forms of data, allowing them to create new and original content based on patterns learned from training data.

Explore more

Expand Arrow

Gesture Recognition

Gesture recognition is a technology that interprets human gestures or body movements to control and interact with computers or other devices. It allows users to convey commands, input data, or interact with a system through natural movements, making it valuable in applications like gaming, virtual reality, and user interfaces.

Explore more

Expand Arrow

Instance Segmentation

Instance segmentation is a computer vision task that combines object detection and semantic segmentation. It identifies and delineates individual objects within an image, assigning each pixel to a specific object instance. This provides a detailed understanding of the spatial extent and location of distinct objects in an image.

Explore more

Expand Arrow

Keypoint Detection

Keypoint detection is a computer vision task that identifies and localizes specific points or landmarks in an image. These keypoints represent important features, such as corners or interest points, and are often used for tasks like object tracking, pose estimation, and image alignment.

Explore more

Expand Arrow

Machine Learning Operations (MLOps)

MLOps, short for Machine Learning Operations, is a set of practices and tools that combine machine learning with DevOps to manage the end-to-end machine learning lifecycle. It encompasses model development, deployment, monitoring, and automation, enabling efficient, scalable, and reliable machine learning operations in production environments.

Explore more

Expand Arrow

Model Deployment

Model deployment is the act of making a machine learning model operational and accessible for real-world use. It involves integrating the model into a software application, cloud service, or other systems, so it can make predictions or decisions based on new data in a practical, automated, and scalable manner.

Explore more

Expand Arrow

Object Detection

Object detection is a computer vision task that involves identifying and locating objects within images or videos. It goes beyond image classification by not only classifying objects but also drawing bounding boxes around them, providing information about their positions in the image. It's widely used in applications like autonomous driving and image analysis.

Explore more

Expand Arrow

Object Tracking

Object tracking is a computer vision process that involves monitoring and following the movement of objects within a sequence of images or a video stream over time. It assigns a unique identity to each object and tracks its position and motion as it moves through the frames, enabling applications like video surveillance and autonomous vehicles.

Explore more

Expand Arrow

Pose Estimation

Pose estimation is a computer vision task that identifies and calculates the positions and orientations of key body parts or objects within an image or video, often in the context of human pose analysis. It's used in applications such as motion capture, gesture recognition, and augmented reality.

Explore more

Expand Arrow

Semantic Segmentation

Semantic segmentation is a computer vision task that classifies each pixel in an image to a specific object category or class. It provides a detailed understanding of the objects' spatial layout and enables the delineation of object boundaries in an image, making it useful in applications like image analysis and autonomous driving.

Explore more

Expand Arrow

Training Metrics

Training metrics in machine learning are quantitative measures used to evaluate and assess the performance of a model during its training process. They help in understanding how well the model is learning from the data and can include metrics like accuracy, loss, precision, recall, and F1-score, among others.

Explore more

Expand Arrow

Unstructured Data

Unstructured data refers to information that doesn't have a predefined format or structure, making it harder to organize and analyze using traditional data processing methods. Examples include text, images, audio, and video. Specialized techniques, like natural language processing and computer vision, are used to extract insights from unstructured data.

Explore more

Expand Arrow

Vision Transformers (ViTs)

Vision transformers are a class of neural networks that apply the transformer architecture, originally designed for sequence modelling tasks like language translation, to image processing tasks.

Explore more

Expand Arrow

Vision-Language Models (VLMs)

Vision language models are multi-modal models that can learn simultaneously from images and texts to tackle many tasks, from visual question answering to image captioning.

Explore more

Expand Arrow

Visualization

In computer vision, visualization involves creating graphical representations of visual data, aiding in tasks like image segmentation, object detection, and image classification. It helps interpret and communicate information through visual cues, such as bounding boxes, heatmaps, or feature representations, enhancing the understanding of computer vision algorithms and results.

Explore more

Expand Arrow

YOLO

YOLO is a real-time object detection algorithm in computer vision. It processes an image once to simultaneously predict bounding boxes and class probabilities for multiple objects. YOLO is known for its speed and accuracy, making it suitable for various applications, including autonomous driving and surveillance.

Explore more

Expand Arrow

Zero-Shot / K-Shot Learning

Zero-shot learning is a machine learning approach where a model can recognize and classify objects or concepts it has never seen during training. K-shot learning extends this by allowing the model to learn from a small number (k) of examples for each unseen class, enabling better generalization.

Explore more

Expand Arrow