Glossary
Aโ
Aggregation Engineโ
A distributed computation framework built into Tecton that enables efficient
creation and serving of time-windowed aggregation features. It handles complex
calculations over time windows while ensuring online/offline consistency and
optimizing for performance.
Learn more
Attribute Featureโ
A simple feature type in Tecton that represents a direct value from a column in
transformed data. Attribute features are used for storing and serving individual
data points like user properties or metadata.
Learn more
Bโ
Backfillโ
The process of computing historical feature values for a feature view, ensuring
data completeness before serving features in production. This is typically done
in batch to populate past timestamps.
Learn more
Batch Feature Viewโ
A type of Feature View that transforms data from batch sources (like data
warehouses or data lakes) on a schedule. Batch Feature Views pre-compute feature
values and store them in Tecton's feature stores for later retrieval.
Learn more
Batch Sourceโ
A Data Source that specifies how to connect to and read from a batch data
repository like Snowflake, BigQuery, S3, or a Hive table.
Learn more
Batch Scheduleโ
A configuration parameter that defines how frequently a Batch Feature View or
Stream Feature View's offline materialization jobs run, typically specified as a
time interval (e.g., timedelta(days=1)).
Learn more
Cโ
Cacheโ
A temporary storage layer that speeds up feature retrieval by storing recently
accessed data, reducing the need for repeated computations or database queries.
In Tecton, caching improves low-latency access to online features.
Learn more
Clusterโ
A group of computing resources (e.g., Spark, Databricks, or Kubernetes nodes) used to process, store, or serve feature data efficiently.
Compactionโ
A process that optimizes stored feature data by reducing redundancy and merging
smaller data segments, improving performance and reducing storage costs.
Learn more
Compute Engineโ
The processing system responsible for executing feature transformations and
aggregations. This can include batch processing frameworks (e.g., Spark, Rift)
or real-time streaming engines, depending on the feature pipeline
requirements.
Learn more
Control Planeโ
The component of a system responsible for managing configurations,
orchestration, and metadata, ensuring proper coordination of data and compute
operations. In Tecton, the control plane defines feature views, materialization
schedules, and infrastructure settings.
Learn more
Dโ
Data Planeโ
The layer responsible for executing data processing tasks, including feature
transformation, storage, and retrieval. It handles the movement and computation
of data based on configurations from the control plane.
Learn more
Data Sourceโ
An object that defines how Tecton connects to and reads from external data
systems. Data Sources abstract away connection details and ensure consistent
interpretation of data both online and offline.
Learn more
Eโ
Embeddingโ
A vector representation of data (text, images, etc.) that captures semantic
meaning in a form usable by machine learning models. In Tecton, embeddings are
used particularly for GenAI applications to enable semantic search, clustering,
and other similarity-based operations.
Learn more
Entityโ
Represents a business object (e.g., user, product, transaction) that features
describe. Entities serve as primary keys for feature lookups and are used to
join together data from different sources.
Learn more
Environmentโ
A configuration that defines the runtime context for executing feature
transformations, including Python packages and dependencies. Environments can be
Tecton-managed (pre-configured) or custom-defined to support specific feature
computation requirements.
Learn more
Fโ
Featureโ
A measurable property or attribute of data used as input for machine learning
models. Features are derived from raw data and can be transformed, aggregated,
or stored for inference.
Learn more
Feature Server Groupโ
A cluster of live serving nodes capable of autoscaling that return feature
vectors via Tecton's HTTP API. Feature Server Groups provide isolation at the
workspace level, preventing resource contention between different teams'
workloads and enabling granular resource management and cost attribution.
Learn more
Feature Tableโ
A structured storage format that holds computed feature values, indexed by
entity keys and timestamps. Feature tables enable efficient retrieval of feature
data for training and inference.
Learn more
Feature Viewโ
The core building block in Tecton that defines transformations to convert raw
data into features. Feature Views encapsulate logic for computing features
consistently in both online and offline environments.
Learn more
Freshnessโ
The time delay between when raw data is generated and when a corresponding
feature is available for model inference. Lower freshness latency improves
real-time predictions.
Learn more
Mโ
Modelโ
A machine learning algorithm trained on features to make predictions or
classifications based on new input data.
Learn more
Pโ
Pipelineโ
A sequence of processes that transform raw data into features, including
ingestion, transformation, materialization, and serving for model training or
inference.
Learn more
Sโ
Skew/Driftโ
The deviation between training and inference feature distributions (skew) or
gradual changes in data patterns over time (drift), which can degrade model
performance.
Learn more
Streamingโ
A data processing paradigm where data is continuously ingested and transformed
in real-time, enabling low-latency feature updates for online predictions.
Learn more
Tโ
Tileโ
A precomputed, time-bucketed feature aggregation unit used to optimize storage
and retrieval efficiency in Tecton's feature computation framework.
Learn more
Time Travelโ
The ability to query historical feature values at specific timestamps, allowing
models to reconstruct past feature states for training and debugging.
Learn more
Transform Server Groupโ
A group of Transform Server nodes that execute real-time transformations for
Realtime Feature Views. Transform Server Groups support both provisioned and
autoscaling modes, providing isolation between workloads to prevent resource
contention and enable granular resource management.
Learn more
Trainingโ
The process of feeding historical feature data into a machine learning model to
learn patterns and optimize predictive accuracy.
Learn more
TTL (Time-to-Live)โ
A retention policy that defines how long feature data is stored before being
automatically deleted, balancing storage costs and data relevance.
Learn more
Wโ
Workspaceโ
A dedicated environment for managing and deploying Tecton feature pipelines.
Workspaces provide isolation between different stages of development (e.g., dev,
test, prod) or between teams sharing a Tecton deployment.
Learn more