Upload folder using huggingface_hub

#2
by cmcmaster - opened
README.md CHANGED
@@ -1,6 +1,11 @@
1
  ---
 
2
  library_name: transformers
3
  tags:
 
 
 
 
4
  - mlx
5
  ---
6
 
 
1
  ---
2
+ base_model: unsloth/gemma-2-2b-it
3
  library_name: transformers
4
  tags:
5
+ - medical
6
+ - unsloth
7
+ - peft
8
+ - qlora
9
  - mlx
10
  ---
11
 
config.json CHANGED
@@ -26,12 +26,11 @@
26
  "pad_token_id": 0,
27
  "query_pre_attn_scalar": 256,
28
  "rms_norm_eps": 1e-06,
29
- "rope_scaling": null,
30
  "rope_theta": 10000.0,
31
  "sliding_window": 4096,
32
- "torch_dtype": "bfloat16",
33
- "transformers_version": "4.45.0.dev0",
34
- "unsloth_version": "2024.8",
35
  "use_cache": true,
36
  "vocab_size": 256000
37
  }
 
26
  "pad_token_id": 0,
27
  "query_pre_attn_scalar": 256,
28
  "rms_norm_eps": 1e-06,
 
29
  "rope_theta": 10000.0,
30
  "sliding_window": 4096,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.44.2",
33
+ "unsloth_version": "2024.9",
34
  "use_cache": true,
35
  "vocab_size": 256000
36
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35c62fb9f10c5a558c1bf438073c53aef81b8f153fe1a49a29d675ad72ab1c3f
3
  size 5228717063
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18dd08296858f1cab8f7ccede39ad31822c78d26ad6e3f27a883cf0b4204d77e
3
  size 5228717063
special_tokens_map.json CHANGED
@@ -1,4 +1,8 @@
1
  {
 
 
 
 
2
  "bos_token": {
3
  "content": "<bos>",
4
  "lstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<start_of_turn>",
4
+ "<end_of_turn>"
5
+ ],
6
  "bos_token": {
7
  "content": "<bos>",
8
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -1995,11 +1995,15 @@
1995
  "special": false
1996
  }
1997
  },
 
 
 
 
1998
  "bos_token": "<bos>",
1999
  "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
2000
  "clean_up_tokenization_spaces": false,
2001
  "eos_token": "<eos>",
2002
- "model_max_length": 8192,
2003
  "pad_token": "<pad>",
2004
  "padding_side": "left",
2005
  "sp_model_kwargs": {},
 
1995
  "special": false
1996
  }
1997
  },
1998
+ "additional_special_tokens": [
1999
+ "<start_of_turn>",
2000
+ "<end_of_turn>"
2001
+ ],
2002
  "bos_token": "<bos>",
2003
  "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
2004
  "clean_up_tokenization_spaces": false,
2005
  "eos_token": "<eos>",
2006
+ "model_max_length": 1000000000000000019884624838656,
2007
  "pad_token": "<pad>",
2008
  "padding_side": "left",
2009
  "sp_model_kwargs": {},