Skip to content

Comments

[Qeff.finetuning] Adding Full document for hf_based finetuning stack#732

Open
tchawada wants to merge 1 commit intoquic:ft_experimentalfrom
tchawada:hf_full_document
Open

[Qeff.finetuning] Adding Full document for hf_based finetuning stack#732
tchawada wants to merge 1 commit intoquic:ft_experimentalfrom
tchawada:hf_full_document

Conversation

@tchawada
Copy link
Contributor

This adds a complete, professionally structured documentation set for the HF-Based QEfficient Finetune Module.
It includes installation steps, dataset preparation workflows, custom dataset support, Hugging Face dataset formatting, prompt template guidance, finetuning instructions, inference usage.

Signed-off-by: Tanisha Chawada <tchawada@qti.qualcomm.com>
fsdp: "full_shard"
fsdp_config: "./configs/accelerate/fsdp_config.yaml"
fsdp_config: "./configs/accelerate/fsdp_tp_parallelism_config.yaml"
```
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We used to support PP+DDP also in earlier stack. Test and verify the same in new stack, and include it


* **tokenizer\_name**: `default = "HuggingFaceTB/SmolLM-135M"` → Matches model name.
* **dataset\_type**: `default = "seq_completion"` → Used for sequence continuation tasks, where the language model learns to generate the correct output (completion) step by step, given an input (prompt).
* **dataset\_name**: `default = "knkarthick/samsum"` → Dataset name for training.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We need to add additional num_samples to slect only a subset of the provided dataset, especially in testing scenario.

## Parallelism

The training script supports multiple parallelism strategies:

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Include the sample commands and steps to run all these different strategies

python finetune.py --config configs/sample_config.yaml
```

**Distributed (TorchRun)**
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We need to add MULTI NODE FT based documentation also.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please refer the existing stack finetune.md

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants