Skip to content

Commit 384db78

Browse files
committed
Add another subheader in module 0
Signed-off-by: Danny Chiao <danny@tecton.ai>
1 parent 73e72f6 commit 384db78

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

module_0/README.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,9 +26,10 @@ We focus on a specific example (that does not include online features + models):
2626
- [Step 2e (optional): Merge a sample PR in your fork](#step-2e-optional-merge-a-sample-pr-in-your-fork)
2727
- [Other best practices](#other-best-practices)
2828
- [User group 2: ML Engineers](#user-group-2-ml-engineers)
29+
- [Step 0: Understanding `get_historical_features`](#step-0-understanding-get_historical_features)
2930
- [Step 1: Fetch features for batch scoring (method 1)](#step-1-fetch-features-for-batch-scoring-method-1)
3031
- [Step 2: Fetch features for batch scoring (method 2)](#step-2-fetch-features-for-batch-scoring-method-2)
31-
- [Step 3 (optional): Scaling to large datasets](#step-3-optional-scaling-to-large-datasets)
32+
- [Step 3 (optional): Scaling `get_historical_features` to large datasets](#step-3-optional-scaling-get_historical_features-to-large-datasets)
3233
- [User group 3: Data Scientists](#user-group-3-data-scientists)
3334
- [Conclusion](#conclusion)
3435
- [FAQ](#faq)
@@ -381,6 +382,8 @@ Additionally, users will often want to have a dev/staging environment that's sep
381382

382383
Data scientists or ML engineers can use the defined `FeatureService` (corresponding to model versions) and schedule regular jobs that generate batch predictions (or regularly retrain).
383384

385+
### Step 0: Understanding `get_historical_features`
386+
384387
Feast right now requires timestamps in `get_historical_features`, so what you'll need to do is append an event timestamp of `now()`. e.g.
385388

386389
```python
@@ -428,7 +431,7 @@ store = FeatureStore(config=repo_config)
428431
training_df = store.get_historical_features(...).to_df()
429432
```
430433

431-
### Step 3 (optional): Scaling to large datasets
434+
### Step 3 (optional): Scaling `get_historical_features` to large datasets
432435
You may note that the above example uses a `to_df()` method to load the training dataset into memory and may be wondering how this scales if you have very large datasets.
433436

434437
`get_historical_features` actually returns a `RetrievalJob` object that lazily executes the point-in-time join. The `RetrievalJob` class is extended by each offline store to allow flushing results to e.g. the data warehouse or data lakes.

0 commit comments

Comments
 (0)