Skip to content

Support for Tensor Caching in Kubeflow Data CacheΒ #3173

@Goku2099

Description

@Goku2099

What you would like to be added?

This issue tracks caching of tokenized datasets to accelerate data loading for fine-tuning use cases. By caching tokenized data, we can avoid re-tokenizing the dataset on every GPU node and significantly improve training speed, especially for hyperparameter optimization once support for a common initializer is available.

The goal is to offload the tokenization step to data-cache CPU nodes, freeing GPU nodes to focus exclusively on training.

Why is this needed?

Tensor caching will boost GPU utilization by enabling reuse of them across training nodes.

Love this feature?

Give it a πŸ‘ We prioritize the features with most πŸ‘

Metadata

Metadata

Type

No type

Projects

No projects

Relationships

None yet

Development

No branches or pull requests

Issue actions