# Copyright 2019 The Feast Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import logging
from datetime import datetime, timedelta
from enum import Enum
from pathlib import Path
from threading import Lock
from typing import Any, Dict, List, Optional
from urllib.parse import urlparse
from google.protobuf.internal.containers import RepeatedCompositeFieldContainer
from proto import Message
from feast import usage
from feast.base_feature_view import BaseFeatureView
from feast.data_source import DataSource
from feast.entity import Entity
from feast.errors import (
ConflictingFeatureViewNames,
DataSourceNotFoundException,
EntityNotFoundException,
FeatureServiceNotFoundException,
FeatureViewNotFoundException,
ValidationReferenceNotFound,
)
from feast.feature_service import FeatureService
from feast.feature_view import FeatureView
from feast.importer import import_class
from feast.infra.infra_object import Infra
from feast.infra.registry import proto_registry_utils
from feast.infra.registry.base_registry import BaseRegistry
from feast.infra.registry.registry_store import NoopRegistryStore
from feast.on_demand_feature_view import OnDemandFeatureView
from feast.project_metadata import ProjectMetadata
from feast.protos.feast.core.Registry_pb2 import Registry as RegistryProto
from feast.repo_config import RegistryConfig
from feast.repo_contents import RepoContents
from feast.request_feature_view import RequestFeatureView
from feast.saved_dataset import SavedDataset, ValidationReference
from feast.stream_feature_view import StreamFeatureView
REGISTRY_SCHEMA_VERSION = "1"
REGISTRY_STORE_CLASS_FOR_TYPE = {
"GCSRegistryStore": "feast.infra.registry.gcs.GCSRegistryStore",
"S3RegistryStore": "feast.infra.registry.s3.S3RegistryStore",
"FileRegistryStore": "feast.infra.registry.file.FileRegistryStore",
"PostgreSQLRegistryStore": "feast.infra.registry.contrib.postgres.postgres_registry_store.PostgreSQLRegistryStore",
"AzureRegistryStore": "feast.infra.registry.contrib.azure.azure_registry_store.AzBlobRegistryStore",
}
REGISTRY_STORE_CLASS_FOR_SCHEME = {
"gs": "GCSRegistryStore",
"s3": "S3RegistryStore",
"file": "FileRegistryStore",
"": "FileRegistryStore",
}
[docs]class FeastObjectType(Enum):
DATA_SOURCE = "data source"
ENTITY = "entity"
FEATURE_VIEW = "feature view"
ON_DEMAND_FEATURE_VIEW = "on demand feature view"
REQUEST_FEATURE_VIEW = "request feature view"
STREAM_FEATURE_VIEW = "stream feature view"
FEATURE_SERVICE = "feature service"
[docs] @staticmethod
def get_objects_from_registry(
registry: "BaseRegistry", project: str
) -> Dict["FeastObjectType", List[Any]]:
return {
FeastObjectType.DATA_SOURCE: registry.list_data_sources(project=project),
FeastObjectType.ENTITY: registry.list_entities(project=project),
FeastObjectType.FEATURE_VIEW: registry.list_feature_views(project=project),
FeastObjectType.ON_DEMAND_FEATURE_VIEW: registry.list_on_demand_feature_views(
project=project
),
FeastObjectType.REQUEST_FEATURE_VIEW: registry.list_request_feature_views(
project=project
),
FeastObjectType.STREAM_FEATURE_VIEW: registry.list_stream_feature_views(
project=project,
),
FeastObjectType.FEATURE_SERVICE: registry.list_feature_services(
project=project
),
}
[docs] @staticmethod
def get_objects_from_repo_contents(
repo_contents: RepoContents,
) -> Dict["FeastObjectType", List[Any]]:
return {
FeastObjectType.DATA_SOURCE: repo_contents.data_sources,
FeastObjectType.ENTITY: repo_contents.entities,
FeastObjectType.FEATURE_VIEW: repo_contents.feature_views,
FeastObjectType.ON_DEMAND_FEATURE_VIEW: repo_contents.on_demand_feature_views,
FeastObjectType.REQUEST_FEATURE_VIEW: repo_contents.request_feature_views,
FeastObjectType.STREAM_FEATURE_VIEW: repo_contents.stream_feature_views,
FeastObjectType.FEATURE_SERVICE: repo_contents.feature_services,
}
FEAST_OBJECT_TYPES = [feast_object_type for feast_object_type in FeastObjectType]
logger = logging.getLogger(__name__)
[docs]def get_registry_store_class_from_type(registry_store_type: str):
if not registry_store_type.endswith("RegistryStore"):
raise Exception('Registry store class name should end with "RegistryStore"')
if registry_store_type in REGISTRY_STORE_CLASS_FOR_TYPE:
registry_store_type = REGISTRY_STORE_CLASS_FOR_TYPE[registry_store_type]
module_name, registry_store_class_name = registry_store_type.rsplit(".", 1)
return import_class(module_name, registry_store_class_name, "RegistryStore")
[docs]def get_registry_store_class_from_scheme(registry_path: str):
uri = urlparse(registry_path)
if uri.scheme not in REGISTRY_STORE_CLASS_FOR_SCHEME:
raise Exception(
f"Registry path {registry_path} has unsupported scheme {uri.scheme}. "
f"Supported schemes are file, s3 and gs."
)
else:
registry_store_type = REGISTRY_STORE_CLASS_FOR_SCHEME[uri.scheme]
return get_registry_store_class_from_type(registry_store_type)
[docs]class Registry(BaseRegistry):
# The cached_registry_proto object is used for both reads and writes. In particular,
# all write operations refresh the cache and modify it in memory; the write must
# then be persisted to the underlying RegistryStore with a call to commit().
cached_registry_proto: Optional[RegistryProto] = None
cached_registry_proto_created: Optional[datetime] = None
cached_registry_proto_ttl: timedelta
def __new__(
cls,
project: str,
registry_config: Optional[RegistryConfig],
repo_path: Optional[Path],
):
# We override __new__ so that we can inspect registry_config and create a SqlRegistry without callers
# needing to make any changes.
if registry_config and registry_config.registry_type == "sql":
from feast.infra.registry.sql import SqlRegistry
return SqlRegistry(registry_config, project, repo_path)
elif registry_config and registry_config.registry_type == "snowflake.registry":
from feast.infra.registry.snowflake import SnowflakeRegistry
return SnowflakeRegistry(registry_config, project, repo_path)
else:
return super(Registry, cls).__new__(cls)
def __init__(
self,
project: str,
registry_config: Optional[RegistryConfig],
repo_path: Optional[Path],
):
"""
Create the Registry object.
Args:
registry_config: RegistryConfig object containing the destination path and cache ttl,
repo_path: Path to the base of the Feast repository
or where it will be created if it does not exist yet.
"""
self._refresh_lock = Lock()
if registry_config:
registry_store_type = registry_config.registry_store_type
registry_path = registry_config.path
if registry_store_type is None:
cls = get_registry_store_class_from_scheme(registry_path)
else:
cls = get_registry_store_class_from_type(str(registry_store_type))
self._registry_store = cls(registry_config, repo_path)
self.cached_registry_proto_ttl = timedelta(
seconds=registry_config.cache_ttl_seconds
if registry_config.cache_ttl_seconds is not None
else 0
)
[docs] def clone(self) -> "Registry":
new_registry = Registry("project", None, None)
new_registry.cached_registry_proto_ttl = timedelta(seconds=0)
new_registry.cached_registry_proto = (
self.cached_registry_proto.__deepcopy__()
if self.cached_registry_proto
else RegistryProto()
)
new_registry.cached_registry_proto_created = datetime.utcnow()
new_registry._registry_store = NoopRegistryStore()
return new_registry
def _initialize_registry(self, project: str):
"""Explicitly initializes the registry with an empty proto if it doesn't exist."""
try:
self._get_registry_proto(project=project)
except FileNotFoundError:
registry_proto = RegistryProto()
registry_proto.registry_schema_version = REGISTRY_SCHEMA_VERSION
proto_registry_utils.init_project_metadata(registry_proto, project)
self._registry_store.update_registry_proto(registry_proto)
[docs] def update_infra(self, infra: Infra, project: str, commit: bool = True):
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
self.cached_registry_proto.infra.CopyFrom(infra.to_proto())
if commit:
self.commit()
[docs] def get_infra(self, project: str, allow_cache: bool = False) -> Infra:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return Infra.from_proto(registry_proto.infra)
[docs] def apply_entity(self, entity: Entity, project: str, commit: bool = True):
entity.is_valid()
now = datetime.utcnow()
if not entity.created_timestamp:
entity.created_timestamp = now
entity.last_updated_timestamp = now
entity_proto = entity.to_proto()
entity_proto.spec.project = project
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, existing_entity_proto in enumerate(
self.cached_registry_proto.entities
):
if (
existing_entity_proto.spec.name == entity_proto.spec.name
and existing_entity_proto.spec.project == project
):
del self.cached_registry_proto.entities[idx]
break
self.cached_registry_proto.entities.append(entity_proto)
if commit:
self.commit()
[docs] def list_entities(self, project: str, allow_cache: bool = False) -> List[Entity]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_entities(registry_proto, project)
[docs] def list_data_sources(
self, project: str, allow_cache: bool = False
) -> List[DataSource]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_data_sources(registry_proto, project)
[docs] def apply_data_source(
self, data_source: DataSource, project: str, commit: bool = True
):
registry = self._prepare_registry_for_changes(project)
for idx, existing_data_source_proto in enumerate(registry.data_sources):
if existing_data_source_proto.name == data_source.name:
del registry.data_sources[idx]
data_source_proto = data_source.to_proto()
data_source_proto.data_source_class_type = (
f"{data_source.__class__.__module__}.{data_source.__class__.__name__}"
)
data_source_proto.project = project
data_source_proto.data_source_class_type = (
f"{data_source.__class__.__module__}.{data_source.__class__.__name__}"
)
registry.data_sources.append(data_source_proto)
if commit:
self.commit()
[docs] def delete_data_source(self, name: str, project: str, commit: bool = True):
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, data_source_proto in enumerate(
self.cached_registry_proto.data_sources
):
if data_source_proto.name == name:
del self.cached_registry_proto.data_sources[idx]
if commit:
self.commit()
return
raise DataSourceNotFoundException(name)
[docs] def apply_feature_service(
self, feature_service: FeatureService, project: str, commit: bool = True
):
now = datetime.utcnow()
if not feature_service.created_timestamp:
feature_service.created_timestamp = now
feature_service.last_updated_timestamp = now
feature_service_proto = feature_service.to_proto()
feature_service_proto.spec.project = project
registry = self._prepare_registry_for_changes(project)
for idx, existing_feature_service_proto in enumerate(registry.feature_services):
if (
existing_feature_service_proto.spec.name
== feature_service_proto.spec.name
and existing_feature_service_proto.spec.project == project
):
del registry.feature_services[idx]
registry.feature_services.append(feature_service_proto)
if commit:
self.commit()
[docs] def list_feature_services(
self, project: str, allow_cache: bool = False
) -> List[FeatureService]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_feature_services(registry_proto, project)
[docs] def get_feature_service(
self, name: str, project: str, allow_cache: bool = False
) -> FeatureService:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_feature_service(registry_proto, name, project)
[docs] def get_entity(self, name: str, project: str, allow_cache: bool = False) -> Entity:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_entity(registry_proto, name, project)
[docs] def apply_feature_view(
self, feature_view: BaseFeatureView, project: str, commit: bool = True
):
feature_view.ensure_valid()
now = datetime.utcnow()
if not feature_view.created_timestamp:
feature_view.created_timestamp = now
feature_view.last_updated_timestamp = now
feature_view_proto = feature_view.to_proto()
feature_view_proto.spec.project = project
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
self._check_conflicting_feature_view_names(feature_view)
existing_feature_views_of_same_type: RepeatedCompositeFieldContainer
if isinstance(feature_view, StreamFeatureView):
existing_feature_views_of_same_type = (
self.cached_registry_proto.stream_feature_views
)
elif isinstance(feature_view, FeatureView):
existing_feature_views_of_same_type = (
self.cached_registry_proto.feature_views
)
elif isinstance(feature_view, OnDemandFeatureView):
existing_feature_views_of_same_type = (
self.cached_registry_proto.on_demand_feature_views
)
elif isinstance(feature_view, RequestFeatureView):
existing_feature_views_of_same_type = (
self.cached_registry_proto.request_feature_views
)
else:
raise ValueError(f"Unexpected feature view type: {type(feature_view)}")
for idx, existing_feature_view_proto in enumerate(
existing_feature_views_of_same_type
):
if (
existing_feature_view_proto.spec.name == feature_view_proto.spec.name
and existing_feature_view_proto.spec.project == project
):
if (
feature_view.__class__.from_proto(existing_feature_view_proto)
== feature_view
):
return
else:
del existing_feature_views_of_same_type[idx]
break
existing_feature_views_of_same_type.append(feature_view_proto)
if commit:
self.commit()
[docs] def list_stream_feature_views(
self, project: str, allow_cache: bool = False
) -> List[StreamFeatureView]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_stream_feature_views(registry_proto, project)
[docs] def list_on_demand_feature_views(
self, project: str, allow_cache: bool = False
) -> List[OnDemandFeatureView]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_on_demand_feature_views(
registry_proto, project
)
[docs] def get_on_demand_feature_view(
self, name: str, project: str, allow_cache: bool = False
) -> OnDemandFeatureView:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_on_demand_feature_view(
registry_proto, name, project
)
[docs] def get_data_source(
self, name: str, project: str, allow_cache: bool = False
) -> DataSource:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_data_source(registry_proto, name, project)
[docs] def apply_materialization(
self,
feature_view: FeatureView,
project: str,
start_date: datetime,
end_date: datetime,
commit: bool = True,
):
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, existing_feature_view_proto in enumerate(
self.cached_registry_proto.feature_views
):
if (
existing_feature_view_proto.spec.name == feature_view.name
and existing_feature_view_proto.spec.project == project
):
existing_feature_view = FeatureView.from_proto(
existing_feature_view_proto
)
existing_feature_view.materialization_intervals.append(
(start_date, end_date)
)
existing_feature_view.last_updated_timestamp = datetime.utcnow()
feature_view_proto = existing_feature_view.to_proto()
feature_view_proto.spec.project = project
del self.cached_registry_proto.feature_views[idx]
self.cached_registry_proto.feature_views.append(feature_view_proto)
if commit:
self.commit()
return
for idx, existing_stream_feature_view_proto in enumerate(
self.cached_registry_proto.stream_feature_views
):
if (
existing_stream_feature_view_proto.spec.name == feature_view.name
and existing_stream_feature_view_proto.spec.project == project
):
existing_stream_feature_view = StreamFeatureView.from_proto(
existing_stream_feature_view_proto
)
existing_stream_feature_view.materialization_intervals.append(
(start_date, end_date)
)
existing_stream_feature_view.last_updated_timestamp = datetime.utcnow()
stream_feature_view_proto = existing_stream_feature_view.to_proto()
stream_feature_view_proto.spec.project = project
del self.cached_registry_proto.stream_feature_views[idx]
self.cached_registry_proto.stream_feature_views.append(
stream_feature_view_proto
)
if commit:
self.commit()
return
raise FeatureViewNotFoundException(feature_view.name, project)
[docs] def list_feature_views(
self, project: str, allow_cache: bool = False
) -> List[FeatureView]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_feature_views(registry_proto, project)
[docs] def get_request_feature_view(self, name: str, project: str):
registry_proto = self._get_registry_proto(project=project, allow_cache=False)
return proto_registry_utils.get_request_feature_view(
registry_proto, name, project
)
[docs] def list_request_feature_views(
self, project: str, allow_cache: bool = False
) -> List[RequestFeatureView]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_request_feature_views(registry_proto, project)
[docs] def get_feature_view(
self, name: str, project: str, allow_cache: bool = False
) -> FeatureView:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_feature_view(registry_proto, name, project)
[docs] def get_stream_feature_view(
self, name: str, project: str, allow_cache: bool = False
) -> StreamFeatureView:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_stream_feature_view(
registry_proto, name, project
)
[docs] def delete_feature_service(self, name: str, project: str, commit: bool = True):
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, feature_service_proto in enumerate(
self.cached_registry_proto.feature_services
):
if (
feature_service_proto.spec.name == name
and feature_service_proto.spec.project == project
):
del self.cached_registry_proto.feature_services[idx]
if commit:
self.commit()
return
raise FeatureServiceNotFoundException(name, project)
[docs] def delete_feature_view(self, name: str, project: str, commit: bool = True):
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, existing_feature_view_proto in enumerate(
self.cached_registry_proto.feature_views
):
if (
existing_feature_view_proto.spec.name == name
and existing_feature_view_proto.spec.project == project
):
del self.cached_registry_proto.feature_views[idx]
if commit:
self.commit()
return
for idx, existing_request_feature_view_proto in enumerate(
self.cached_registry_proto.request_feature_views
):
if (
existing_request_feature_view_proto.spec.name == name
and existing_request_feature_view_proto.spec.project == project
):
del self.cached_registry_proto.request_feature_views[idx]
if commit:
self.commit()
return
for idx, existing_on_demand_feature_view_proto in enumerate(
self.cached_registry_proto.on_demand_feature_views
):
if (
existing_on_demand_feature_view_proto.spec.name == name
and existing_on_demand_feature_view_proto.spec.project == project
):
del self.cached_registry_proto.on_demand_feature_views[idx]
if commit:
self.commit()
return
for idx, existing_stream_feature_view_proto in enumerate(
self.cached_registry_proto.stream_feature_views
):
if (
existing_stream_feature_view_proto.spec.name == name
and existing_stream_feature_view_proto.spec.project == project
):
del self.cached_registry_proto.stream_feature_views[idx]
if commit:
self.commit()
return
raise FeatureViewNotFoundException(name, project)
[docs] def delete_entity(self, name: str, project: str, commit: bool = True):
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, existing_entity_proto in enumerate(
self.cached_registry_proto.entities
):
if (
existing_entity_proto.spec.name == name
and existing_entity_proto.spec.project == project
):
del self.cached_registry_proto.entities[idx]
if commit:
self.commit()
return
raise EntityNotFoundException(name, project)
[docs] def apply_saved_dataset(
self,
saved_dataset: SavedDataset,
project: str,
commit: bool = True,
):
now = datetime.utcnow()
if not saved_dataset.created_timestamp:
saved_dataset.created_timestamp = now
saved_dataset.last_updated_timestamp = now
saved_dataset_proto = saved_dataset.to_proto()
saved_dataset_proto.spec.project = project
self._prepare_registry_for_changes(project)
assert self.cached_registry_proto
for idx, existing_saved_dataset_proto in enumerate(
self.cached_registry_proto.saved_datasets
):
if (
existing_saved_dataset_proto.spec.name == saved_dataset_proto.spec.name
and existing_saved_dataset_proto.spec.project == project
):
del self.cached_registry_proto.saved_datasets[idx]
break
self.cached_registry_proto.saved_datasets.append(saved_dataset_proto)
if commit:
self.commit()
[docs] def get_saved_dataset(
self, name: str, project: str, allow_cache: bool = False
) -> SavedDataset:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_saved_dataset(registry_proto, name, project)
[docs] def list_saved_datasets(
self, project: str, allow_cache: bool = False
) -> List[SavedDataset]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_saved_datasets(registry_proto, project)
[docs] def apply_validation_reference(
self,
validation_reference: ValidationReference,
project: str,
commit: bool = True,
):
validation_reference_proto = validation_reference.to_proto()
validation_reference_proto.project = project
registry_proto = self._prepare_registry_for_changes(project)
for idx, existing_validation_reference in enumerate(
registry_proto.validation_references
):
if (
existing_validation_reference.name == validation_reference_proto.name
and existing_validation_reference.project == project
):
del registry_proto.validation_references[idx]
break
registry_proto.validation_references.append(validation_reference_proto)
if commit:
self.commit()
[docs] def get_validation_reference(
self, name: str, project: str, allow_cache: bool = False
) -> ValidationReference:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.get_validation_reference(
registry_proto, name, project
)
[docs] def list_validation_references(
self, project: str, allow_cache: bool = False
) -> List[ValidationReference]:
registry_proto = self._get_registry_proto(
project=project, allow_cache=allow_cache
)
return proto_registry_utils.list_validation_references(registry_proto, project)
[docs] def delete_validation_reference(self, name: str, project: str, commit: bool = True):
registry_proto = self._prepare_registry_for_changes(project)
for idx, existing_validation_reference in enumerate(
registry_proto.validation_references
):
if (
existing_validation_reference.name == name
and existing_validation_reference.project == project
):
del registry_proto.validation_references[idx]
if commit:
self.commit()
return
raise ValidationReferenceNotFound(name, project=project)
[docs] def commit(self):
"""Commits the state of the registry cache to the remote registry store."""
if self.cached_registry_proto:
self._registry_store.update_registry_proto(self.cached_registry_proto)
[docs] def refresh(self, project: Optional[str] = None):
"""Refreshes the state of the registry cache by fetching the registry state from the remote registry store."""
self._get_registry_proto(project=project, allow_cache=False)
[docs] def teardown(self):
"""Tears down (removes) the registry."""
self._registry_store.teardown()
[docs] def proto(self) -> RegistryProto:
return self.cached_registry_proto or RegistryProto()
def _prepare_registry_for_changes(self, project: str):
"""Prepares the Registry for changes by refreshing the cache if necessary."""
try:
self._get_registry_proto(project=project, allow_cache=True)
if (
proto_registry_utils.get_project_metadata(
self.cached_registry_proto, project
)
is None
):
# Project metadata not initialized yet. Try pulling without cache
self._get_registry_proto(project=project, allow_cache=False)
except FileNotFoundError:
registry_proto = RegistryProto()
registry_proto.registry_schema_version = REGISTRY_SCHEMA_VERSION
self.cached_registry_proto = registry_proto
self.cached_registry_proto_created = datetime.utcnow()
# Initialize project metadata if needed
assert self.cached_registry_proto
if (
proto_registry_utils.get_project_metadata(
self.cached_registry_proto, project
)
is None
):
proto_registry_utils.init_project_metadata(
self.cached_registry_proto, project
)
self.commit()
return self.cached_registry_proto
def _get_registry_proto(
self, project: Optional[str], allow_cache: bool = False
) -> RegistryProto:
"""Returns the cached or remote registry state
Args:
project: Name of the Feast project (optional)
allow_cache: Whether to allow the use of the registry cache when fetching the RegistryProto
Returns: Returns a RegistryProto object which represents the state of the registry
"""
with self._refresh_lock:
expired = (
self.cached_registry_proto is None
or self.cached_registry_proto_created is None
) or (
self.cached_registry_proto_ttl.total_seconds()
> 0 # 0 ttl means infinity
and (
datetime.utcnow()
> (
self.cached_registry_proto_created
+ self.cached_registry_proto_ttl
)
)
)
if project:
old_project_metadata = proto_registry_utils.get_project_metadata(
registry_proto=self.cached_registry_proto, project=project
)
if allow_cache and not expired and old_project_metadata is not None:
assert isinstance(self.cached_registry_proto, RegistryProto)
return self.cached_registry_proto
elif allow_cache and not expired:
assert isinstance(self.cached_registry_proto, RegistryProto)
return self.cached_registry_proto
registry_proto = self._registry_store.get_registry_proto()
self.cached_registry_proto = registry_proto
self.cached_registry_proto_created = datetime.utcnow()
if not project:
return registry_proto
project_metadata = proto_registry_utils.get_project_metadata(
registry_proto=registry_proto, project=project
)
if project_metadata:
usage.set_current_project_uuid(project_metadata.project_uuid)
else:
proto_registry_utils.init_project_metadata(registry_proto, project)
self.commit()
return registry_proto
def _check_conflicting_feature_view_names(self, feature_view: BaseFeatureView):
name_to_fv_protos = self._existing_feature_view_names_to_fvs()
if feature_view.name in name_to_fv_protos:
if not isinstance(
name_to_fv_protos.get(feature_view.name), feature_view.proto_class
):
raise ConflictingFeatureViewNames(feature_view.name)
def _existing_feature_view_names_to_fvs(self) -> Dict[str, Message]:
assert self.cached_registry_proto
odfvs = {
fv.spec.name: fv
for fv in self.cached_registry_proto.on_demand_feature_views
}
fvs = {fv.spec.name: fv for fv in self.cached_registry_proto.feature_views}
request_fvs = {
fv.spec.name: fv for fv in self.cached_registry_proto.request_feature_views
}
sfv = {
fv.spec.name: fv for fv in self.cached_registry_proto.stream_feature_views
}
return {**odfvs, **fvs, **request_fvs, **sfv}