Skip to content

OpenGradient Glossary

  • DA - Data Availability layer.
  • Enclave nodes - fully isolated virtual machines that are highly protected by a single security perimeter. They have no persistent storage, no external networking, no interactive access.
  • Facilitator - optional services on Base Sepolia that handle payment verification and payment settlement complexity for x402 LLM inference. Facilitators provide payment verification, payment settlement management, payment method abstraction, rate limiting, and receipt generation. Proof settlement and verification happen separately on the OpenGradient network.
  • HACA - Hybrid AI Compute Architecture
  • IModelExecutor - a Solidity interface that contracts must implement to use Model Scheduler. It requires two functions: run() for executing model inference logic and getInferenceResult() for returning the latest model output.
  • Inference - a process of running a specific AI/ML model with a given input
  • Inference Network - OpenGradient's open network of inference nodes.
  • Inference Node - Nodes with specialized hardware, such as GPUs, responsible for model execution on OpenGradient.
  • LLM - Large Language Model - a type of AI model designed to understand and generate human-like text based on natural language prompts.
  • ML - Machine Learning - a subset of artificial intelligence that enables systems to learn and improve from experience without being explicitly programmed.
  • Model Output - the result returned from a model inference execution, containing the model's predictions or generated outputs.
  • Model Scheduler (AlphaSense) - enables scheduling and automating recurring model inference tasks on the OpenGradient blockchain. Contracts implementing IModelExecutor can be registered to execute automatically on a defined schedule.
  • ONNX - Open Neural Network Exchange is an open standard format for ML models.
  • PIPE - Parallelized Inference Pre-Execution Engine - an on-chain inference execution method that allows models to be natively used from smart contracts.
  • Settlement Modes - three modes for x402 LLM inference settlement that determine what data is stored on-chain: SETTLE_INDIVIDUAL (input/output hashes only), SETTLE_BATCH (batch hashes for multiple inferences), and SETTLE_INDIVIDUAL_WITH_METADATA (full model information, complete input/output data, and all inference metadata).
  • TEE - Trusted Execution Environment - hardware or software security architecture that protects sensitive data/code from unauthorized access.
  • Tensor - a data structure that stores N-dimensional data.
  • TPU - Tensor Processing Unit, circuits specifically designed to accelerate ML workloads.
  • VCS - Validation-Computation Separation architecture, where validator nodes are separated from inference nodes. This framework means that running inference on inference nodes does not have to be replicated on all validator nodes.
  • x402 - an open, neutral standard for internet-native payments. It absolves the Internet's original sin by natively making payments possible between clients and servers, creating win-win economies that empower agentic payments at scale. x402 exists to build a more free and fair internet.
  • ZKML - Zero-Knowledge Machine Learning - a combination of ZK and ML techniques, where running inference and training models are designed to generate zk-proofs that validate computation without revealing the input data.