Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

Commit 63bd546

Browse files
committed
typos
Signed-off-by: Luke Nezda <lnezda@gmail.com>
1 parent 642e44d commit 63bd546

File tree

3 files changed

+7
-7
lines changed

3 files changed

+7
-7
lines changed

examples/huggingface/pytorch/question-answering/pruning/group_lasso/modeling.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -601,7 +601,7 @@ def forward(self, sequence_output, pooled_output):
601601

602602
class BertPreTrainedModel(nn.Module):
603603
""" An abstract class to handle weights initialization and
604-
a simple interface for dowloading and loading pretrained models.
604+
a simple interface for downloading and loading pretrained models.
605605
"""
606606
def __init__(self, config, *inputs, **kwargs):
607607
super(BertPreTrainedModel, self).__init__()
@@ -663,7 +663,7 @@ def from_pretrained(cls, pretrained_model_name_or_path, state_dict=None, cache_d
663663
. `model.chkpt` a TensorFlow checkpoint
664664
from_tf: should we load the weights from a locally saved TensorFlow checkpoint
665665
cache_dir: an optional path to a folder in which the pre-trained models will be cached.
666-
state_dict: an optional state dictionnary (collections.OrderedDict object) to use instead of Google pre-trained models
666+
state_dict: an optional state dictionary (collections.OrderedDict object) to use instead of Google pre-trained models
667667
*inputs, **kwargs: additional input for the specific Bert class
668668
(ex: num_labels for BertForSequenceClassification)
669669
"""

intel_extension_for_transformers/llm/runtime/graph/__init__.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ def __import_package(self, model_type):
6565
elif model_type == "mistral":
6666
import intel_extension_for_transformers.llm.runtime.graph.mistral_cpp as cpp_model
6767
else:
68-
raise TypeError("Unspported model type {}!".format(model_type))
68+
raise TypeError("Unsupported model type {}!".format(model_type))
6969
self.module = cpp_model
7070

7171
@staticmethod
@@ -212,7 +212,7 @@ def eos_token_id(self):
212212
if self.model_type == 'qwen':
213213
return self.tokenizer.special_tokens['<|endoftext|>']
214214
return self.tokenizer.eos_token_id
215-
215+
216216
def pad_token_id(self):
217217
if self.tokenizer.pad_token_id == None:
218218
if self.batch_size == 1:

intel_extension_for_transformers/llm/runtime/graph/developer_document.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -79,8 +79,8 @@ graph LR;
7979

8080
We need to implement corresponding serialization methods from pytorch format, which is mainly divided into the following three steps.
8181

82-
## 1.1. Hyperparamters
83-
The term **"hyperparamters"** describes a value that is used to configure the behavior of a large language model; this is in contrast to the model's parameters, which are the weight that were derived in the training process that was used to create the model. Each model defines its own hyperparameter structure that defines the hyperparameter values accepted by that model. Valid ITREX graph files must list these values in the correct order, and each value must be represented using the correct data type. Although hyperparameters are different across models, some attributes appear in the hyperparameters for most models:
82+
## 1.1. Hyperparameters
83+
The term **"hyperparameters"** describes a value that is used to configure the behavior of a large language model; this is in contrast to the model's parameters, which are the weight that were derived in the training process that was used to create the model. Each model defines its own hyperparameter structure that defines the hyperparameter values accepted by that model. Valid ITREX graph files must list these values in the correct order, and each value must be represented using the correct data type. Although hyperparameters are different across models, some attributes appear in the hyperparameters for most models:
8484
- n_vocab: the size of the model's vocabulary
8585
- n_embd: the size of the model's " embedding layer", which is used during prompt ingestion.
8686
- n_layer: the number of layers in the model; each layer represents a set of weights.
@@ -328,7 +328,7 @@ Most of our model examples only support single prompt processing. You need to ad
328328
+set_target_properties(${TARGET} PROPERTIES POSITION_INDEPENDENT_CODE ON)
329329
+target_link_libraries(${TARGET} PUBLIC ne_layers jblas::jblas)
330330
```
331-
and and new_model to [models_CMakeLists.txt](https://github.com/intel/intel-extension-for-transformers/blob/main/intel_extension_for_transformers/llm/runtime/graph/models/CMakeLists.txt).
331+
and new_model to [models_CMakeLists.txt](https://github.com/intel/intel-extension-for-transformers/blob/main/intel_extension_for_transformers/llm/runtime/graph/models/CMakeLists.txt).
332332
```diff
333333
add_subdirectory(opt)
334334
add_subdirectory(bloom)

0 commit comments

Comments
 (0)