You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
from feast import FileSource, FeatureView, Field
from datetime import timedelta
from feast.types import Float32
source = FileSource(name="source",
path="dataset.parquet",
timestamp_field="event_timestamp")
fv = FeatureView(
name="fv",
schema=[
Field(name="feature", dtype=Float32)
],
source=source,
)
The below snippet should have retuend the latest 1000 records:
from feast import FeatureStore
# lets fetch last 1000 records
entity_df = df[-1000:]
store = FeatureStore('.')
hist = store.get_historical_features(
entity_df = entity_df,
features = ['fv:feature']
)
# this won't work, as it consumes huge RAM in the FileOfflineStore._merge function
hist.to_df()
Current Behavior
It crashes as it generates a huge data frame in the FileOfflineStore._merge function.
Expected Behavior
Assuming the following file:
With the following feature store definition:
The below snippet should have retuend the latest 1000 records:
Current Behavior
It crashes as it generates a huge data frame in the FileOfflineStore._merge function.
Steps to reproduce
https://colab.research.google.com/drive/1yTSQPK6H2zMq3HEswLkGUShlFIuMniS8?usp=sharing
Specifications
Possible Solution
The "if not join_keys" logic in FileOfflineStore._merge, should be reconsidered.
The text was updated successfully, but these errors were encountered: