吳成岳 commited on
Commit
e2f65ae
1 Parent(s): 381e337

update model

Browse files
README.md CHANGED
@@ -30,7 +30,7 @@ Mistral_Pro_8B_v0.1 showcases superior performance on a range of benchmarks. It
30
  | :-: | :-: | :-: | :-: | :-: | :-: | :-: | :-: |
31
  | Gemma-7B | 61.9 | 82.2 | 64.6 | 44.8 | 79.0 | 50.9 | 32.3 |
32
  | Mistral-7B | 60.8 | 83.3 | 62.7 | 42.6 | 78.0 | 39.2 | 28.7 |
33
- | Mistral_Pro_8B_v0.1 | 62.6 | 82.5 | 60.7 | 47.6 | 78.1 | 50.3 | 32.3 |
34
 
35
 
36
  ## Limitations
 
30
  | :-: | :-: | :-: | :-: | :-: | :-: | :-: | :-: |
31
  | Gemma-7B | 61.9 | 82.2 | 64.6 | 44.8 | 79.0 | 50.9 | 32.3 |
32
  | Mistral-7B | 60.8 | 83.3 | 62.7 | 42.6 | 78.0 | 39.2 | 28.7 |
33
+ | Mistral_Pro_8B_v0.1 | 63.2 | 82.6 | 60.6 | 48.3 | 78.1 | 50.6 | 32.9 |
34
 
35
 
36
  ## Limitations
config.json CHANGED
@@ -2,7 +2,6 @@
2
  "architectures": [
3
  "MistralForCausalLM"
4
  ],
5
- "attention_dropout": 0.0,
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
8
  "hidden_act": "silu",
@@ -19,7 +18,7 @@
19
  "sliding_window": 4096,
20
  "tie_word_embeddings": false,
21
  "torch_dtype": "bfloat16",
22
- "transformers_version": "4.37.2",
23
- "use_cache": false,
24
  "vocab_size": 32000
25
  }
 
2
  "architectures": [
3
  "MistralForCausalLM"
4
  ],
 
5
  "bos_token_id": 1,
6
  "eos_token_id": 2,
7
  "hidden_act": "silu",
 
18
  "sliding_window": 4096,
19
  "tie_word_embeddings": false,
20
  "torch_dtype": "bfloat16",
21
+ "transformers_version": "4.34.0.dev0",
22
+ "use_cache": true,
23
  "vocab_size": 32000
24
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.37.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.34.0.dev0"
6
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18bb05a004abb9a8d562fdbc13ecbc071dcad6a4cbd39828c87d0d7c3e3b0625
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f689c8c817b11e536872f82e5cdd4d76376abdb9fa84da83bcb3b51ac1da2e1f
3
  size 4943162336
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ff565f6fe465df47de8674b9afb413fd2b8ed6a5ed6df1b7983b0f0d4905b9a
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:518388960a6dcf8058fcb41e03a7f8d6dbe334263d713432bf21b59cd03dd5b8
3
  size 4999819336
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0e60747d07722721defa773e80bc4d42804468df14b70065dea42e84e595170
3
  size 4915916184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc8ee7b371ff025a73041d01cc1ce050c15c4440c62ce6762705dfd34e7e0532
3
  size 4915916184
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc3710b4a17fae46d1dbfaba133a8be228e0e5bbcc5a5f7703817ff401f37f64
3
  size 3114400560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ca3a0b6960d370bffeea03546c057026eac3dc1dfa881570bd29fa62a04ecb9
3
  size 3114400560
special_tokens_map.json CHANGED
@@ -1,24 +1,5 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "</s>",
17
- "unk_token": {
18
- "content": "<unk>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -33,7 +33,7 @@
33
  "eos_token": "</s>",
34
  "legacy": true,
35
  "model_max_length": 1000000000000000019884624838656,
36
- "pad_token": "</s>",
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",
 
33
  "eos_token": "</s>",
34
  "legacy": true,
35
  "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": null,
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",