Push model using huggingface_hub.
Browse files- config.json +1 -1
- generation_config.json +1 -1
- mini_gpt4_llama_v2.py +3 -2
- model-00001-of-00003.safetensors +1 -1
- model-00002-of-00003.safetensors +1 -1
config.json
CHANGED
|
@@ -35,7 +35,7 @@
|
|
| 35 |
"remove_template": false,
|
| 36 |
"token_pooling": true,
|
| 37 |
"torch_dtype": "float32",
|
| 38 |
-
"transformers_version": "4.
|
| 39 |
"use_grad_checkpoint": true,
|
| 40 |
"use_grad_checkpoint_llm": true,
|
| 41 |
"vit_model": "eva_clip_g",
|
|
|
|
| 35 |
"remove_template": false,
|
| 36 |
"token_pooling": true,
|
| 37 |
"torch_dtype": "float32",
|
| 38 |
+
"transformers_version": "4.37.2",
|
| 39 |
"use_grad_checkpoint": true,
|
| 40 |
"use_grad_checkpoint_llm": true,
|
| 41 |
"vit_model": "eva_clip_g",
|
generation_config.json
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
{
|
| 2 |
"_from_model_config": true,
|
| 3 |
-
"transformers_version": "4.
|
| 4 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"_from_model_config": true,
|
| 3 |
+
"transformers_version": "4.37.2"
|
| 4 |
}
|
mini_gpt4_llama_v2.py
CHANGED
|
@@ -644,6 +644,7 @@ class MiniGPT4_Video(Blip2Base, PreTrainedModel):
|
|
| 644 |
repetition_penalty=repetition_penalty,
|
| 645 |
# stopping_criteria=stopping_criteria,
|
| 646 |
# use_fastv=False,
|
|
|
|
| 647 |
)
|
| 648 |
|
| 649 |
answers = []
|
|
@@ -845,11 +846,11 @@ class MiniGPT4_Video(Blip2Base, PreTrainedModel):
|
|
| 845 |
msg = model.load_state_dict(ckpt['model'], strict=False)
|
| 846 |
# push the model to the hub with its metadata and config file
|
| 847 |
model.to('cuda')
|
| 848 |
-
model.push_to_hub("Vision-CAIR/MiniGPT4-video-mistral-hf")
|
| 849 |
video_config = minigpt4_video_config(cfg)
|
| 850 |
# video_config.save_pretrained("minigpt4_video_config")
|
| 851 |
# print("Save Minigpt-4-LLM Config: minigpt4_video_config")
|
| 852 |
-
video_config.push_to_hub("Vision-CAIR/MiniGPT4-video-mistral-hf")
|
| 853 |
return model
|
| 854 |
|
| 855 |
|
|
|
|
| 644 |
repetition_penalty=repetition_penalty,
|
| 645 |
# stopping_criteria=stopping_criteria,
|
| 646 |
# use_fastv=False,
|
| 647 |
+
use_cache=True,
|
| 648 |
)
|
| 649 |
|
| 650 |
answers = []
|
|
|
|
| 846 |
msg = model.load_state_dict(ckpt['model'], strict=False)
|
| 847 |
# push the model to the hub with its metadata and config file
|
| 848 |
model.to('cuda')
|
| 849 |
+
# model.push_to_hub("Vision-CAIR/MiniGPT4-video-mistral-hf")
|
| 850 |
video_config = minigpt4_video_config(cfg)
|
| 851 |
# video_config.save_pretrained("minigpt4_video_config")
|
| 852 |
# print("Save Minigpt-4-LLM Config: minigpt4_video_config")
|
| 853 |
+
# video_config.push_to_hub("Vision-CAIR/MiniGPT4-video-mistral-hf")
|
| 854 |
return model
|
| 855 |
|
| 856 |
|
model-00001-of-00003.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4981849121
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23a0ae99f0ba8f23e67afc7c82027eb2ae2df03505b3b7f4cf2ca3af78cdac47
|
| 3 |
size 4981849121
|
model-00002-of-00003.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4610750095
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d877cf7c7a09380e9a42a11aa951dc1fcc83c48c03f37d14d7b7b8b3f3ae224
|
| 3 |
size 4610750095
|