Skip to content

Commit 9fd061e

Browse files
committed
Fix links
1 parent 36ff8c2 commit 9fd061e

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

48 files changed

+91
-90
lines changed

docs/setup_installation/common/arrow_flight_duckdb.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,8 +12,8 @@ When the service is enabled, clients will automatically use it for the following
1212
- [reading Feature Groups][hsfs.feature_group.FeatureGroup.read]
1313
- [reading Queries][hsfs.constructor.query.Query.read]
1414
- [reading Training Datasets][hsfs.feature_view.FeatureView.get_training_data]
15-
- [creating In-Memory Training Datasets](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#training_data)
16-
- [reading Batch Inference Data](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#get_batch_data)
15+
- [creating In-Memory Training Datasets][hsfs.feature_view.FeatureView.training_data]
16+
- [reading Batch Inference Data][hsfs.feature_view.FeatureView.get_batch_data]
1717
For larger datasets, clients can still make use of the Spark/Hive backend by explicitly setting
1818
`read_options={"use_hive": True}`.
1919

docs/user_guides/fs/compute_engines.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -21,14 +21,14 @@ Hopsworks is aiming to provide functional parity between the computational engin
2121

2222
| Functionality | Method | Spark | Python | Flink | Beam | Java | Comment |
2323
| --- | --- | --- | --- | --- | --- | --- | --- |
24-
| Feature Group Creation from dataframes | [`FeatureGroup.create_feature_group()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#create_feature_group) | :white_check_mark: | :white_check_mark: | - | - | - | Currently Flink/Beam/Java doesn't support registering feature group metadata. Thus it needs to be pre-registered before you can write real time features computed by Flink/Beam. |
25-
| Training Dataset Creation from dataframes | [`TrainingDataset.save()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/training_dataset_api/#save) | :white_check_mark: | - | - | - | - | Functionality was deprecated in version 3.0 |
26-
| Data validation using Great Expectations for streaming dataframes | [`FeatureGroup.validate()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#validate) <br/> [`FeatureGroup.insert_stream()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#insert_stream) | - | - | - | - | - | `insert_stream` does not perform any data validation even when a expectation suite is attached. |
27-
| Stream ingestion | [`FeatureGroup.insert_stream()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#insert_stream) | :white_check_mark: | - | :white_check_mark: | :white_check_mark: | :white_check_mark: | Python/Pandas/Polars has currently no notion of streaming. |
28-
| Reading from Streaming Storage Connectors | [`KafkaConnector.read_stream()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/storage_connector_api/#read_stream) | :white_check_mark: | - | - | - | - | Python/Pandas/Polars has currently no notion of streaming. For Flink/Beam/Java only write operations are supported |
29-
| Reading training data from external storage other than S3 | [`FeatureView.get_training_data()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#get_training_data) | :white_check_mark: | - | - | - | - | Reading training data that was written to external storage using a Storage Connector other than S3 can currently not be read using HSFS APIs, instead you will have to use the storage's native client. |
30-
| Reading External Feature Groups into Dataframe | [`ExternalFeatureGroup.read()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/external_feature_group_api/#read) | :white_check_mark: | - | - | - | - | Reading an External Feature Group directly into a Pandas/Polars Dataframe is not supported, however, you can use the [Query API](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/query_api/) to create Feature Views/Training Data containing External Feature Groups. |
31-
| Read Queries containing External Feature Groups into Dataframe | [`Query.read()`](https://docs.hopsworks.ai/feature-store-api/{{{hopsworks_version}}}/generated/api/query_api/#read) | :white_check_mark: | - | - | - | - | Reading a Query containing an External Feature Group directly into a Pandas/Polars Dataframe is not supported, however, you can use the Query to create Feature Views/Training Data and write the data to a Storage Connector, from where you can read up the data into a Pandas/Polars Dataframe. |
24+
| Feature Group Creation from dataframes | [`FeatureStore.create_feature_group`][hsfs.feature_store.FeatureStore.create_feature_group] | :white_check_mark: | :white_check_mark: | - | - | - | Currently Flink/Beam/Java doesn't support registering feature group metadata. Thus it needs to be pre-registered before you can write real time features computed by Flink/Beam. |
25+
| Training Dataset Creation from dataframes | [`TrainingDataset.save`][hsfs.training_dataset.TrainingDataset.save] | :white_check_mark: | - | - | - | - | Functionality was deprecated in version 3.0 |
26+
| Data validation using Great Expectations for streaming dataframes | [`FeatureGroup.validate`][hsfs.feature_group.FeatureGroup.validate] <br/> [`FeatureGroup.insert_stream`][hsfs.feature_group.FeatureGroup.insert_stream] | - | - | - | - | - | `insert_stream` does not perform any data validation even when a expectation suite is attached. |
27+
| Stream ingestion | [`FeatureGroup.insert_stream`][hsfs.feature_group.FeatureGroup.insert_stream] | :white_check_mark: | - | :white_check_mark: | :white_check_mark: | :white_check_mark: | Python/Pandas/Polars has currently no notion of streaming. |
28+
| Reading from Streaming Storage Connectors | [`KafkaConnector.read_stream`][hsfs.storage_connector.KafkaConnector.read_stream] | :white_check_mark: | - | - | - | - | Python/Pandas/Polars has currently no notion of streaming. For Flink/Beam/Java only write operations are supported |
29+
| Reading training data from external storage other than S3 | [`FeatureView.get_training_data`][hsfs.feature_view.FeatureView.get_training_data] | :white_check_mark: | - | - | - | - | Reading training data that was written to external storage using a Storage Connector other than S3 can currently not be read using HSFS APIs, instead you will have to use the storage's native client. |
30+
| Reading External Feature Groups into Dataframe | [`ExternalFeatureGroup.read`][hsfs.feature_group.ExternalFeatureGroup.read] | :white_check_mark: | - | - | - | - | Reading an External Feature Group directly into a Pandas/Polars Dataframe is not supported, however, you can use the [Query API][hsfs.constructor.query.Query] to create Feature Views/Training Data containing External Feature Groups. |
31+
| Read Queries containing External Feature Groups into Dataframe | [`Query.read`][hsfs.constructor.query.Query.read] | :white_check_mark: | - | - | - | - | Reading a Query containing an External Feature Group directly into a Pandas/Polars Dataframe is not supported, however, you can use the Query to create Feature Views/Training Data and write the data to a Storage Connector, from where you can read up the data into a Pandas/Polars Dataframe. |
3232

3333
## Python
3434

docs/user_guides/fs/data_source/usage.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -160,4 +160,4 @@ For a detailed walkthrough on managing and utilizing training data, refer to the
160160
## Next Steps
161161

162162
We have gone through the basic use cases of a data source.
163-
For more details about the API functionality for any specific connector type, checkout the [API section](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/storage_connector_api/#storage-connector).
163+
For more details about the API functionality for any specific connector type, checkout the [API section][hsfs.storage_connector.StorageConnector].

docs/user_guides/fs/feature_group/create.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ Using the HSFS API you can execute:
3939
)
4040
```
4141

42-
You can read [the full method documentation](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#featuregroup) to get more details.
42+
You can read the full [`FeatureStore.create_feature_group`][hsfs.feature_store.FeatureStore.create_feature_group] documentation to get more details.
4343
If you need to create a feature group with vector similarity search supported, refer to [the vector similarity guide](../vector_similarity_search.md#extending-feature-groups-with-similarity-search).
4444
`name` is the only mandatory parameter of the `create_feature_group` and represents the name of the feature group.
4545

@@ -353,7 +353,7 @@ If a feature group is online enabled, the `insert` method will store the feature
353353

354354
### API Reference
355355

356-
[FeatureGroup](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#featuregroup)
356+
[`FeatureGroup`][hsfs.feature_group.FeatureGroup]
357357

358358
## Create using the UI
359359

docs/user_guides/fs/feature_group/create_external.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,7 @@ Once you have defined the metadata, you can
7777
fg.save()
7878
```
7979

80-
You can read [the full method documentation](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/external_feature_group_api/#externalfeaturegroup) for more details.
80+
You can read the full [`FeatureStore.create_external_feature_group`][hsfs.feature_store.FeatureStore.create_external_feature_group] documentation for more details.
8181
`name` is a mandatory parameter of the `create_external_feature_group` and represents the name of the feature group.
8282

8383
The version number is optional, if you don't specify the version number the APIs will create a new version by default with a version number equals to the highest existing version number plus one.
@@ -126,18 +126,18 @@ For an external feature group to be available online, during the creation of the
126126
```
127127

128128
The `insert()` method takes a DataFrame as parameter and writes it _only_ to the online feature store.
129-
Users can select which subset of the feature group data they want to make available on the online feature store by using the [query APIs](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/query_api/).
129+
Users can select which subset of the feature group data they want to make available on the online feature store by using the [query APIs][hsfs.constructor.query.Query].
130130

131131
### Limitations
132132

133133
Hopsworks Feature Store does not support time-travel queries on external feature groups.
134134

135135
Additionally, support for `.read()` and `.show()` methods when using by the Python engine is limited to external feature groups defined on BigQuery and Snowflake and only when using the [Feature Query Service](../../../setup_installation/common/arrow_flight_duckdb.md).
136-
Nevertheless, external feature groups defined top of any data source can be used to create a training dataset from a Python environment invoking one of the following methods: [create_training_data](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#create_training_data), [create_train_test_split](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#create_train_test_split) or the [create_train_validation_test_split](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#create_train_validation_test_split)
136+
Nevertheless, external feature groups defined top of any data source can be used to create a training dataset from a Python environment invoking one of the following methods: [`FeatureView.create_training_data`][hsfs.feature_view.FeatureView.create_training_data], [`FeatureView.create_train_test_split`][hsfs.feature_view.FeatureView.create_train_test_split] or [`FeatureView.create_train_validation_test_split`][hsfs.feature_view.FeatureView.create_train_validation_test_split].
137137

138138
### API Reference
139139

140-
[External FeatureGroup](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/external_feature_group_api/#externalfeaturegroup)
140+
[`ExternalFeatureGroup`][hsfs.feature_group.ExternalFeatureGroup].
141141

142142
## Create using the UI
143143

docs/user_guides/fs/feature_group/create_spine.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,4 +60,4 @@ You just need to make sure it has the same schema.
6060

6161
### API Reference
6262

63-
[SpineGroup](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/spine_group_api/#spinegroup)
63+
[`SpineGroup`][hsfs.feature_group.SpineGroup].

docs/user_guides/fs/feature_group/data_types.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -105,9 +105,9 @@ When in doubt, use the TEXT type instead, but note that it comes with a potentia
105105

106106
Hopsworks allows users to store complex types (e.g. *ARRAY<INT>*) in the online feature store.
107107
Hopsworks serializes the complex features transparently and stores them as VARBINARY in the online feature store.
108-
The serialization happens when calling the [save()](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#save),
109-
[insert()](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#insert) or [insert_stream()](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#insert_stream) methods.
110-
The deserialization will be executed when calling the [get_serving_vector()](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/training_dataset_api/#get_serving_vector) method to retrieve data from the online feature store.
108+
The serialization happens when calling the [`FeatureGroup.save`][hsfs.feature_group.FeatureGroup.save],
109+
[`FeatureGroup.insert`][hsfs.feature_group.FeatureGroup.insert] or [`FeatureGroup.insert_stream`][hsfs.feature_group.FeatureGroup.insert_stream] methods.
110+
The deserialization will be executed when calling the [`TrainingDataset.get_serving_vector`][hsfs.training_dataset.TrainingDataset.get_serving_vector] method to retrieve data from the online feature store.
111111
If users query directly the online feature store, for instance using the `fs.sql("SELECT ...", online=True)` statement, it will return a binary blob.
112112

113113
On the feature store UI, the online feature type for complex features will be reported as *VARBINARY*.
@@ -221,7 +221,7 @@ The most important validation checks or error messages are mentioned below along
221221
df['text_column'] = df['text_column'].str.slice(0, max_length)
222222
```
223223

224-
- Another option is to simply [create new version of the feature group](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#get_or_create_feature_group) and insert the dataframe.
224+
- Another option is to simply [create new version of the feature group][hsfs.feature_store.FeatureStore.get_or_create_feature_group] and insert the dataframe.
225225

226226
!!!note
227227
The total row size limit should be less than 30kb as per [row size restrictions](#online-restrictions-for-row-size).
@@ -252,7 +252,7 @@ The most important validation checks or error messages are mentioned below along
252252
All timestamp features are stored in Hopsworks in UTC time.
253253
Also, all timestamp-based functions (such as [point-in-time joins](../../../concepts/fs/feature_view/offline_api.md#point-in-time-correct-training-data)) use UTC time.
254254
This ensures consistency of timestamp features across different client timezones and simplifies working with timestamp-based functions in general.
255-
When ingesting timestamp features, the [Feature Store Write API](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#insert) will automatically handle the conversion to UTC, if necessary.
255+
When ingesting timestamp features, the [`FeatureGroup.insert`][hsfs.feature_group.FeatureGroup.insert] will automatically handle the conversion to UTC, if necessary.
256256
The following table summarizes how different timestamp types are handled:
257257

258258
| Data Frame (Data Type) | Environment | Handling |
@@ -261,7 +261,7 @@ The following table summarizes how different timestamp types are handled:
261261
| Pandas DataFrame (datetime64[ns, tz]) | Python-only and PySpark | timezone-sensitive conversion from 'tz' to UTC |
262262
| Spark (TimestampType) | PySpark and Spark | interpreted as UTC, independent of the client's timezone |
263263

264-
Timestamp features retrieved from the Feature Store, e.g., using the [Feature Store Read API](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_group_api/#read), use a timezone-unaware format:
264+
Timestamp features retrieved from the Feature Store, e.g., using the [Feature Store Read API][hsfs.feature_group.FeatureGroup.read], use a timezone-unaware format:
265265

266266
| Data Frame (Data Type) | Environment | Timezone |
267267
|---------------------------------------|-------------------------|------------------------|

docs/user_guides/fs/feature_view/feature-vectors.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ If you need to get more familiar with the concept of feature vectors, you can re
1515

1616
You can get back feature vectors from either python or java client by providing the primary key value(s) for the feature view.
1717
Note that filters defined in feature view and training data will not be applied when feature vectors are returned.
18-
If you need to retrieve a complete value of feature vectors without missing values, the required `entry` are [feature_view.primary_keys](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#primary_keys).
18+
If you need to retrieve a complete value of feature vectors without missing values, the required `entry` are [FeatureView.primary_keys][hsfs.feature_view.FeatureView.primary_keys].
1919
Alternative, you can provide the primary key of the feature groups as the key of the entry.
2020
It is also possible to provide a subset of the entry, which will be discussed [below](#partial-feature-retrieval).
2121

@@ -55,7 +55,7 @@ It is also possible to provide a subset of the entry, which will be discussed [b
5555
Starting from python client v3.4, you can specify different values for the primary key of the same name which exists in multiple feature groups but are not joint by the same name.
5656
The table below summarises the value of `primary_keys` in different settings.
5757
Considering that you are joining 2 feature groups, namely, `left_fg` and `right_fg`, the feature groups have different primary keys, and features (`feature_*`) in each setting.
58-
Also, the 2 feature groups are [joint](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/query_api/#join) on different *join conditions* and *prefix* as `left_fg.join(right_fg, <join conditions>, prefix=<prefix>)`.
58+
Also, the 2 feature groups are [joint][hsfs.constructor.query.Query.join] on different *join conditions* and *prefix* as `left_fg.join(right_fg, <join conditions>, prefix=<prefix>)`.
5959

6060
For java client, and python client before v3.4, the `primary_keys` are the set of primary key of all the feature groups in the query.
6161
Python client is backward compatible.
@@ -285,7 +285,7 @@ This implementation is available as of Hopsworks 3.7.
285285
Initialise the client by calling the `init_serving` method on the Feature View object before starting to fetch feature vectors.
286286
This will initialise the chosen client, test the connection, and initialise the transformation functions registered with the Feature View.
287287
Note to use the REST client in the Hopsworks Cluster python environment you will need to provide an API key explicitly as JWT authentication is not yet supported.
288-
More configuration options can be found in the [API documentation](https://docs.hopsworks.ai/hopsworks-api/{{{hopsworks_version}}}/generated/api/feature_view_api/#init_serving).
288+
More configuration options can be found in the [API documentation][hsfs.feature_view.FeatureView.init_serving].
289289

290290
=== "Python"
291291

0 commit comments

Comments
 (0)