Comments (7)
Change "tokenizer_class": "LLaMATokenizer" in llama-13b-hf/tokenizer_config.json into "tokenizer_class": "LlamaTokenizer". It worked for me~
from minigpt-4.
I have ever seen this issue in here, see that if this can help you?
from minigpt-4.
Change "tokenizer_class": "LLaMATokenizer" in llama-13b-hf/tokenizer_config.json into "tokenizer_class": "LlamaTokenizer". It worked for me~
I still met this issue and I chacked it's "Llama"
(in Windows terminal)
from minigpt-4.
'model.layers.18.mlp.gate_proj.weight',` 'model.layers.13.mlp.down_proj.weight', 'model.layers.18.self_attn.q_proj.weight', 'model.layers.39.self_attn.o_proj.weight', 'model.layers.17.mlp.up_proj.weight', 'model.layers.24.self_attn.q_proj.weight', 'model.layers.2.post_attention_layernorm.weight', 'model.layers.17.mlp.down_proj.weight', 'model.layers.27.mlp.down_proj.weight']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
╭─────────────────────────────── Traceback (most recent call last) ────────────────────────────────╮
│ C:\Users\huang.conda\envs\minigpt4\lib\runpy.py:197 in _run_module_as_main │
│ │
│ 194 │ main_globals = sys.modules["main"].dict │
│ 195 │ if alter_argv: │
│ 196 │ │ sys.argv[0] = mod_spec.origin │
│ ❱ 197 │ return _run_code(code, main_globals, None, │
│ 198 │ │ │ │ │ "main", mod_spec) │
│ 199 │
│ 200 def run_module(mod_name, init_globals=None, │
│ │
│ C:\Users\huang.conda\envs\minigpt4\lib\runpy.py:87 in _run_code │
│ │
│ 84 │ │ │ │ │ loader = loader, │
│ 85 │ │ │ │ │ package = pkg_name, │
│ 86 │ │ │ │ │ spec = mod_spec) │
│ ❱ 87 │ exec(code, run_globals) │
│ 88 │ return run_globals │
│ 89 │
│ 90 def _run_module_code(code, init_globals=None, │
│ │
│ C:\Users\huang.conda\envs\minigpt4\lib\site-packages\fastchat\model\apply_delta.py:153 in │
│ │
│ │
│ 150 │ if args.low_cpu_mem: │
│ 151 │ │ apply_delta_low_cpu_mem(args.base_model_path, args.target_model_path, args.delta │
│ 152 │ else: │
│ ❱ 153 │ │ apply_delta(args.base_model_path, args.target_model_path, args.delta_path) │
│ 154 │
│ │
│ C:\Users\huang.conda\envs\minigpt4\lib\site-packages\fastchat\model\apply_delta.py:124 in │
│ apply_delta │
│ │
│ 121 │ print(f"Loading the base model from {base_model_path}") │
│ 122 │ base = AutoModelForCausalLM.from_pretrained( │
│ 123 │ │ base_model_path, torch_dtype=torch.float16, low_cpu_mem_usage=True) │
│ ❱ 124 │ base_tokenizer = AutoTokenizer.from_pretrained( │
│ 125 │ │ base_model_path, use_fast=False) │
│ 126 │ │
│ 127 │ print(f"Loading the delta from {delta_path}") │
│ │
│ C:\Users\huang.conda\envs\minigpt4\lib\site-packages\transformers\models\auto\tokenization_auto │
│ .py:689 in from_pretrained │
│ │
│ 686 │ │ │ │ tokenizer_class = tokenizer_class_from_name(tokenizer_class_candidate) │
│ 687 │ │ │ │
│ 688 │ │ │ if tokenizer_class is None: │
│ ❱ 689 │ │ │ │ raise ValueError( │
│ 690 │ │ │ │ │ f"Tokenizer class {tokenizer_class_candidate} does not exist or is n │
│ 691 │ │ │ │ ) │
│ 692 │ │ │ return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *input │
╰──────────────────────────────────────────────────────────────────────────────────────────────────╯
ValueError: Tokenizer class LLaMATokenizer does not exist or is not currently imported.
(minigpt4) PS I:\chatgpt\minigpt4\MiniGPT-4>
from minigpt-4.
I have observed that a few seconds before the error occurred, the memory usage suddenly spiked to 60GB out of my total 64GB memory. I suspect this issue might be related to the memory consumption. Could you please provide some guidance or suggestions on how to handle this situation? Thank you in advance. @gch8295322
from minigpt-4.
Dear @gch8295322
Thank you for your help earlier. I have prepared the model, but I am still encountering the "TypeError: argument of type 'WindowsPath' is not iterable" issue. I noticed that this problem is also being discussed in #28. I would like to ask if there is a solution to this issue at the moment? Once again, thank you for your assistance.
Best regards
from minigpt-4.
@Wenbobobo
that's wield, I changed the class name and it works, maybe you should just reboot the terminal?
from minigpt-4.
Related Issues (20)
- How to preprocess IconQA for evaluation? HOT 2
- urllib.error.HTTPError: HTTP Error 403: Forbidden
- Is is okay to use CrossEntropyLoss on referring or grounding?
- Hyperparameters for the VQA zero-shot inference
- Question: replacing the Llama 2 7b chat component with an equivalent Code Llama 7b
- cc_sbu_align datasets
- A load persistent id instruction was encountered, but no persistent_load function was specified. HOT 1
- minigpt4-demo
- demo.py 出错 HOT 3
- pip install accelerate仍然不可以 HOT 1
- raise RuntimeError(f"Error(s) in loading state_dict for {model.__class__.__name__}:\n\t{error_msg}") HOT 2
- How to apply MiniGPTv2 to ITM (image-text-matching) tasks
- in PRE_TRAIN(stage1)processing, “Loaded 0 records for train split from the dataset.”
- Siglip
- The same input but different output from the online demo. HOT 1
- What kind of weight do we need to use to load llama2?
- Could not find \'data\' key in response , MiniGpt v2
- Error while finetuning on custom llava conversation dataset
- The download link of train dataset is wrong.
- It seems wrong dataset links in MiniGPT-4/dataset /README_MINIGPTv2_FINETUNE.md HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from minigpt-4.