Skip to content

Commit e3397c2

Browse files
authored
simplify log var (#1368)
Signed-off-by: Qubitium <Qubitium@modelcloud.ai>
1 parent 1ef13b1 commit e3397c2

File tree

121 files changed

+540
-266
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

121 files changed

+540
-266
lines changed

examples/benchmark/generation_speed.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,11 +23,13 @@
2323

2424
import torch
2525
from datasets import Dataset, load_dataset
26-
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig
2726
from logbar import LogBar
2827
from transformers import AutoTokenizer, GenerationConfig
2928
from transformers.generation.logits_process import LogitsProcessor
3029

30+
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig
31+
32+
3133
logger = LogBar.shared()
3234

3335
random.seed(0)

examples/benchmark/ipex.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,7 @@
2020
import torch
2121
from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer
2222

23+
2324
try:
2425
from optimum.intel.utils.modeling_utils import bind_cores_for_best_perf
2526
bind_cores_for_best_perf()
@@ -29,6 +30,7 @@
2930

3031
import argparse
3132

33+
3234
parser = argparse.ArgumentParser(description="Benchmark IPEX vs HF on a pre-trained model.")
3335
parser.add_argument("--model", type=str, required=True, help="Path or name of the pre-trained model.")
3436
parser.add_argument("--cores", type=int, default=8, help="Number of CPU cores to use.")

examples/benchmark/perplexity.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,9 +17,11 @@
1717
import argparse
1818
import os
1919

20-
from gptqmodel.utils import Perplexity
2120
from transformers import AutoTokenizer
2221

22+
from gptqmodel.utils import Perplexity
23+
24+
2325
os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
2426

2527
if __name__ == "__main__":

examples/evaluation/run_language_modeling_task.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,10 +18,12 @@
1818

1919
import datasets
2020
import torch
21+
from transformers import AutoTokenizer
22+
2123
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig
2224
from gptqmodel.eval_tasks import LanguageModelingTask
2325
from gptqmodel.utils.torch import torch_empty_cache
24-
from transformers import AutoTokenizer
26+
2527

2628
DATASET = "tatsu-lab/alpaca"
2729
WITH_INPUT_TEMPLATE = "Instruction:\n{instruction}\n\nInput:\n{input}\n\nOutput:\n"

examples/evaluation/run_sequence_classification_task.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,10 +19,12 @@
1919

2020
import datasets
2121
import torch
22+
from transformers import AutoTokenizer
23+
2224
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig
2325
from gptqmodel.eval_tasks import SequenceClassificationTask
2426
from gptqmodel.utils.torch import torch_empty_cache
25-
from transformers import AutoTokenizer
27+
2628

2729
DATASET = "cardiffnlp/tweet_sentiment_multilingual"
2830
TEMPLATE = "Question:What's the sentiment of the given text? Choices are {labels}.\nText: {text}\nAnswer:"

examples/evaluation/run_text_summarization_task.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,10 +19,12 @@
1919

2020
import datasets
2121
import torch
22+
from transformers import AutoTokenizer, GenerationConfig
23+
2224
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig
2325
from gptqmodel.eval_tasks import TextSummarizationTask
2426
from gptqmodel.utils.torch import torch_empty_cache
25-
from transformers import AutoTokenizer, GenerationConfig
27+
2628

2729
os.system("pip install py7zr")
2830

examples/inference/run_transformers.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,7 @@
1616

1717
from transformers import AutoModelForCausalLM, AutoTokenizer
1818

19+
1920
tokenizer = AutoTokenizer.from_pretrained("TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ")
2021
quantized_model = AutoModelForCausalLM.from_pretrained("TheBloke/TinyLlama-1.1B-Chat-v0.3-GPTQ")
2122
print(tokenizer.decode(quantized_model.generate(**tokenizer("gptqmodel is", return_tensors="pt").to(quantized_model.device))[0]))

examples/inference/run_with_different_backends.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,9 +19,11 @@
1919
import sys
2020
from argparse import ArgumentParser
2121

22-
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig, get_best_device
2322
from transformers import AutoTokenizer
2423

24+
from gptqmodel import BACKEND, GPTQModel, QuantizeConfig, get_best_device
25+
26+
2527
os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
2628
pretrained_model_id = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
2729
quantized_model_id = "./TinyLlama/TinyLlama-1.1B-Chat-v1.0-4bit-128g"

examples/quantization/basic_usage.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,9 +16,11 @@
1616

1717
import os
1818

19-
from gptqmodel import GPTQModel, QuantizeConfig, get_best_device
2019
from transformers import AutoTokenizer
2120

21+
from gptqmodel import GPTQModel, QuantizeConfig, get_best_device
22+
23+
2224
os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
2325

2426
pretrained_model_id = "/monster/data/model/TinyLlama-1.1B-Chat-v1.0" # "TinyLlama/TinyLlama-1.1B-Chat-v1.0"

examples/quantization/basic_usage_autoround.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,9 +15,11 @@
1515
# limitations under the License.
1616

1717
import torch
18+
from transformers import AutoTokenizer
19+
1820
from gptqmodel import GPTQModel
1921
from gptqmodel.quantization.config import AutoRoundQuantizeConfig # noqa: E402
20-
from transformers import AutoTokenizer
22+
2123

2224
pretrained_model_id = "/monster/data/model/TinyLlama-1.1B-Chat-v1.0" # "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
2325
quantized_model_id = "./autoround/TinyLlama-1.1B-Chat-v1.0-4bit-128g"

0 commit comments

Comments
 (0)