Page cover image

Model Configuration

Model Configuration

Field Name
Description

base_model

Specifies the Hugging Face model or its path containing model files (*.pt, *.safetensors, or *.bin).

base_model_ignore_patterns

Allows specifying an ignore pattern for model files in the repository.

base_model_config

Location of the configuration .json file if not included in the base model repository.

model_revision

Specifies a specific model revision from Hugging Face's hub.

tokenizer_config

Optional override for the tokenizer configuration, different from the base model.

model_type

Defines the type of model to load, e.g., AutoModelForCausalLM.

tokenizer_type

Corresponding tokenizer type, e.g., AutoTokenizer.

trust_remote_code

Allows trusting remote code for untrusted sources.

tokenizer_use_fast

Indicates whether to use the use_fast option for tokenizer loading.

tokenizer_legacy

Specifies whether to use the legacy tokenizer setting.

resize_token_embeddings_to_32x

Resizes model embeddings to multiples of 32 when new tokens are added.

is_falcon_derived_model

Boolean flag indicating if the model is derived from Falcon.

is_llama_derived_model

Boolean flag indicating if the model is derived from Llama.

is_mistral_derived_model

Boolean flag indicating if the model is derived from Mistral.

model_config

Optional overrides for the base model configuration, including RoPE Scaling settings.

Last updated

Was this helpful?