Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Model card issue#1125 #1129

Open
wants to merge 4 commits into
base: main
Choose a base branch
from
Open

Model card issue#1125 #1129

wants to merge 4 commits into from

Conversation

EziOzoani
Copy link
Contributor

Updates to address this issue

  • Addition of training regime in the annotated model card to keep this doc and the template in sync.
  • Defined training_regime, along with examples

Updates to address [this issue](#1125) 
- Addition of training regime in the annotated model card to keep this doc and the template in sync.
- Defined training_regime, along with examples
Sentence rephrasing
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.

@@ -158,6 +158,9 @@ _Write 1-2 sentences on what the training data is. Ideally this links to a Datas

## Training Procedure [optional]

_When you want to know what hardware you'll need to fine-tune a model, consider the following factors: the number of parameters in the model and the training regime you plan to use._

_e.g A model with 3B parameters and fp32 precision format needs at least 48GB of GPU memory, while bf16 requires at least 24GB of memory with Amphere or higher hardware. Mixed pf16 requires at least 54GB of GPU memory._
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
_e.g A model with 3B parameters and fp32 precision format needs at least 48GB of GPU memory, while bf16 requires at least 24GB of memory with Amphere or higher hardware. Mixed pf16 requires at least 54GB of GPU memory._
_e.g A model with 3B parameters and fp32 precision format needs at least 48GB of GPU memory, while bf16 requires at least 24GB of memory with Ampere or higher hardware. Mixed fp16 requires at least 54GB of GPU memory._

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

These numbers sound a bit high to me. In any case, they depend on a number of factors like optimizer choice. Should the recommended optimizer be a part of the training_regime data?

docs/hub/model-card-annotated.md Outdated Show resolved Hide resolved
docs/hub/model-card-annotated.md Outdated Show resolved Hide resolved
@julien-c
Copy link
Member

is this PR still in process, @EziOzoani?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants