v0.2 (#2)
Browse files- v0.2 (aefbfb1277bc08ec1f2166b03c0659eea6d6bae1)
- Update README.md (a9df87be4af1c824c0a4d245ad25a1c9e16f65ff)
- README.md +7 -3
- config.json +2 -2
- generation_config.json +1 -1
- model-00001-of-00003.safetensors +1 -1
- model-00002-of-00003.safetensors +1 -1
- model-00003-of-00003.safetensors +1 -1
- special_tokens_map.json +2 -2
- training_args.bin +1 -1
README.md
CHANGED
@@ -25,6 +25,11 @@ This model was developed by StarfleetAI with the aim of it being used for autono
|
|
25 |
|
26 |
## Current Features
|
27 |
|
|
|
|
|
|
|
|
|
|
|
28 |
### `v0.1.1`
|
29 |
|
30 |
- Function call response role changed to `tool` to comply with OpenAI
|
@@ -36,7 +41,6 @@ This model was developed by StarfleetAI with the aim of it being used for autono
|
|
36 |
|
37 |
## Work To Do
|
38 |
|
39 |
-
- Needs more training on examples where function call should be performed in the middle/end parts of the conversation, rather than in the beginning
|
40 |
- Needs to eliminate some hallucinations (it, for example, sometimes regressing news articles instead of calling the `get_rss_feed` function, and stuff like that)
|
41 |
- Correct typography (`Answer: 123` instead of `Answer:123`)
|
42 |
- Current date / time / timezone conversational abilities ("Schedule a call tomorrow at 6PM")
|
@@ -91,7 +95,7 @@ For now, the model is only capable of calling one function at a time.
|
|
91 |
### Example Function Call
|
92 |
|
93 |
```
|
94 |
-
<|fn_start|>{"name": "generate_password", "arguments": {"length": 42}}<|fn_end|>
|
95 |
```
|
96 |
|
97 |
## Function Call Response
|
@@ -125,7 +129,7 @@ You are a helpful assistant.<|im_end|>
|
|
125 |
<|im_start|>user
|
126 |
Generate a password, 42 characters long<|im_end|>
|
127 |
<|im_start|>assistant
|
128 |
-
<|fn_start|>{"name": "generate_password", "arguments": {"length": 42}}<|fn_end|><|im_end|>
|
129 |
<|im_start|>tool
|
130 |
{"result": "87cc47fbc865a290d7c7de4be3c893175c51a566b3"}<|im_end|>
|
131 |
<|im_start|>assistant
|
|
|
25 |
|
26 |
## Current Features
|
27 |
|
28 |
+
### `v0.2`
|
29 |
+
|
30 |
+
- Improved function calling in long conversations
|
31 |
+
- Function call arguments became a JSON string to comply with OpenAI
|
32 |
+
|
33 |
### `v0.1.1`
|
34 |
|
35 |
- Function call response role changed to `tool` to comply with OpenAI
|
|
|
41 |
|
42 |
## Work To Do
|
43 |
|
|
|
44 |
- Needs to eliminate some hallucinations (it, for example, sometimes regressing news articles instead of calling the `get_rss_feed` function, and stuff like that)
|
45 |
- Correct typography (`Answer: 123` instead of `Answer:123`)
|
46 |
- Current date / time / timezone conversational abilities ("Schedule a call tomorrow at 6PM")
|
|
|
95 |
### Example Function Call
|
96 |
|
97 |
```
|
98 |
+
<|fn_start|>{"name": "generate_password", "arguments": "{\"length\": 42}"}<|fn_end|>
|
99 |
```
|
100 |
|
101 |
## Function Call Response
|
|
|
129 |
<|im_start|>user
|
130 |
Generate a password, 42 characters long<|im_end|>
|
131 |
<|im_start|>assistant
|
132 |
+
<|fn_start|>{"name": "generate_password", "arguments": "{\"length\": 42}"}<|fn_end|><|im_end|>
|
133 |
<|im_start|>tool
|
134 |
{"result": "87cc47fbc865a290d7c7de4be3c893175c51a566b3"}<|im_end|>
|
135 |
<|im_start|>assistant
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"MistralForCausalLM"
|
5 |
],
|
@@ -21,7 +21,7 @@
|
|
21 |
"sliding_window": 4096,
|
22 |
"tie_word_embeddings": false,
|
23 |
"torch_dtype": "bfloat16",
|
24 |
-
"transformers_version": "4.37.
|
25 |
"use_cache": false,
|
26 |
"vocab_size": 32004
|
27 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "Open-Orca/Mistral-7B-OpenOrca",
|
3 |
"architectures": [
|
4 |
"MistralForCausalLM"
|
5 |
],
|
|
|
21 |
"sliding_window": 4096,
|
22 |
"tie_word_embeddings": false,
|
23 |
"torch_dtype": "bfloat16",
|
24 |
+
"transformers_version": "4.37.2",
|
25 |
"use_cache": false,
|
26 |
"vocab_size": 32004
|
27 |
}
|
generation_config.json
CHANGED
@@ -2,5 +2,5 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 32000,
|
5 |
-
"transformers_version": "4.37.
|
6 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 32000,
|
5 |
+
"transformers_version": "4.37.2"
|
6 |
}
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4943195104
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4a71c46c3c9b876e87b7ec005305429b4844927c6b87e4a0a28bd6bd2dbcabdc
|
3 |
size 4943195104
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999819336
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f9e597c27d779f38d2ee6d892d97ed07cef304c81da2a28be1d7cbd8d893f82
|
3 |
size 4999819336
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4540549112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b083d0b31291502ce1e6bba235fd2d64e727a09d06e2402fe184214ab4139a0e
|
3 |
size 4540549112
|
special_tokens_map.json
CHANGED
@@ -38,9 +38,9 @@
|
|
38 |
},
|
39 |
"unk_token": {
|
40 |
"content": "<unk>",
|
41 |
-
"lstrip":
|
42 |
"normalized": false,
|
43 |
-
"rstrip":
|
44 |
"single_word": false
|
45 |
}
|
46 |
}
|
|
|
38 |
},
|
39 |
"unk_token": {
|
40 |
"content": "<unk>",
|
41 |
+
"lstrip": true,
|
42 |
"normalized": false,
|
43 |
+
"rstrip": true,
|
44 |
"single_word": false
|
45 |
}
|
46 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd4454f7fb0dfa8edc606f22d8d19c5395b34b860407a59e62bcd405072b23e7
|
3 |
size 4664
|