amc-madalin commited on
Commit
527c412
·
verified ·
1 Parent(s): 7e0e80f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +89 -0
README.md CHANGED
@@ -1,3 +1,92 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - vicgalle/alpaca-gpt4
5
+ language:
6
+ - en
7
+ library_name: transformers
8
  ---
9
+ # Model Card for OLMo-1B-instruct-alpaca_amc
10
+
11
+ OLMo-1B-instruct-alpaca_amc is a language model fine-tuned on the Alpaca-GPT4 dataset. This model extends the capabilities of the original OLMo 1B model, by the Allen Institute for AI (AI2), incorporating the ability to follow instructions and perform chatting tasks.
12
+
13
+ ## Model Description
14
+
15
+ OLMo-1B-instruct-alpaca_amc is built upon the OLMo 1B architecture, a Transformer-based autoregressive language model known for its deep understanding and generation of natural language. By fine-tuning OLMo 1B with the Alpaca-GPT4 dataset, we've tailored the model specifically for interactive chatting applications, enabling it to understand and respond to a wide range of conversational prompts and instructions.
16
+
17
+ ### How to Use
18
+
19
+ To use OLMo-1B-instruct-alpaca_amc for generating text or chatting, you can leverage the Hugging Face Transformers library. Here's a quick example in Python:
20
+
21
+ ```python
22
+ import hf_olmo
23
+ from transformers import AutoModelForCausalLM, AutoTokenizer
24
+
25
+ # Model and tokenizer directories
26
+ tokenizer = AutoTokenizer.from_pretrained("amc-madalin/OLMo-1B-instruct-alpaca_amc")
27
+ model = AutoModelForCausalLM.from_pretrained("amc-madalin/OLMo-1B-instruct-alpaca_amc")
28
+
29
+ print("Chat with the model (type 'quit' to stop):")
30
+ while True:
31
+ message = input("You: ")
32
+ if message.lower() == 'quit':
33
+ break
34
+
35
+ inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False, padding=True, truncation=True, max_length=512)
36
+ response = olmo.generate(**inputs, max_length=512, pad_token_id=tokenizer.eos_token_id, eos_token_id=tokenizer.eos_token_id, do_sample=True, top_k=50, top_p=0.95)
37
+ reply = tokenizer.decode(response[0], skip_special_tokens=True)
38
+ print("AI: ", reply)
39
+ ```
40
+
41
+ ### Training Data
42
+
43
+ The model was fine-tuned on the Alpaca-GPT4 dataset, which consists of 52K instruction-following demonstrations generated in the style of self-instruct using GPT-4.
44
+
45
+ ## Applications
46
+
47
+ OLMo-1B-instruct-alpaca_amc is great for research and experimentation with Open LLMs.
48
+
49
+ - Conversational agents
50
+ - Interactive storytelling
51
+ - Educational tool
52
+
53
+ ## Limitations and Biases
54
+
55
+ As with any language model, OLMo-1B-instruct-alpaca_amc inherits biases present in its training data. Users should be aware of these potential biases and limitations.
56
+
57
+ ## How to Contribute
58
+
59
+ I welcome contributions to improve the model and expand its applications!
60
+
61
+ ## Citation
62
+
63
+ If you use OLMo-1B-instruct-alpaca_amc in your research, please cite the original OLMo paper and the dataset used for fine-tuning:
64
+
65
+ ```bibtex
66
+ @article{Groeneveld2023OLMo,
67
+ title={OLMo: Accelerating the Science of Language Models},
68
+ author={Groeneveld, Dirk and Beltagy, Iz and Walsh, Pete and Bhagia, Akshita and Kinney, Rodney and Tafjord, Oyvind and Jha, Ananya Harsh and Ivison, Hamish and Magnusson, Ian and Wang, Yizhong and Arora, Shane and Atkinson, David and Authur, Russell and Chandu, Khyathi and Cohan, Arman and Dumas, Jennifer and Elazar, Yanai and Gu, Yuling and Hessel, Jack and Khot, Tushar and Merrill, William and Morrison, Jacob and Muennighoff, Niklas and Naik, Aakanksha and Nam, Crystal and Peters, Matthew E. and Pyatkin, Valentina and Ravichander, Abhilasha and Schwenk, Dustin and Shah, Saurabh and Smith, Will and Subramani, Nishant and Wortsman, Mitchell and Dasigi, Pradeep and Lambert, Nathan and Richardson, Kyle and Dodge, Jesse and Lo, Kyle and Soldaini, Luca and Smith, Noah A. and Hajishirzi, Hannaneh},
69
+ journal={Preprint},
70
+ year={2024}
71
+ }
72
+ ```
73
+
74
+ ## License
75
+
76
+ This model is open-sourced under the Apache 2.0 license. See the LICENSE file for more details.
77
+
78
+ ## Contact
79
+
80
+ For any questions or support regarding OLMo-1B-instruct-alpaca_amc, please contact me at:
81
+
82
+ [LinkedIn](https://www.linkedin.com/in/alexandru-m-costea/)
83
+
84
+ [e-Mail]([email protected])
85
+
86
+ Or check my work on:
87
+
88
+ [GitHub](https://github.com/amc-madalin)
89
+
90
+ [HuggingFace](https://huggingface.co/amc-madalin)
91
+
92
+ [Website](https://amcostea.com/)