Geraldine commited on
Commit
a4d4d83
·
verified ·
1 Parent(s): aba83ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -0
README.md CHANGED
@@ -5,11 +5,30 @@ pipeline_tag: text-generation
5
  tags:
6
  - openvino
7
  - openvino-export
 
8
  ---
 
9
 
10
  This model was converted to OpenVINO from [`Geraldine/FineLlama-3.2-3B-Instruct-ead`](https://huggingface.co/Geraldine/FineLlama-3.2-3B-Instruct-ead) using [optimum-intel](https://github.com/huggingface/optimum-intel)
11
  via the [export](https://huggingface.co/spaces/echarlaix/openvino-export) space.
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  First make sure you have optimum-intel installed:
14
 
15
  ```bash
@@ -24,3 +43,36 @@ from optimum.intel import OVModelForCausalLM
24
  model_id = "Geraldine/FineLlama-3.2-3B-Instruct-ead-openvino"
25
  model = OVModelForCausalLM.from_pretrained(model_id)
26
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  tags:
6
  - openvino
7
  - openvino-export
8
+ license: llama3.2
9
  ---
10
+ # FineLlama-3.2-3B-Instruct-ead-openvino
11
 
12
  This model was converted to OpenVINO from [`Geraldine/FineLlama-3.2-3B-Instruct-ead`](https://huggingface.co/Geraldine/FineLlama-3.2-3B-Instruct-ead) using [optimum-intel](https://github.com/huggingface/optimum-intel)
13
  via the [export](https://huggingface.co/spaces/echarlaix/openvino-export) space.
14
 
15
+ ## Model Description
16
+
17
+ - **Original Model**: Geraldine/FineLlama-3.2-3B-Instruct-ead
18
+ - **Framework**: OpenVINO
19
+ - **Task**: Text Generation, EAD tag generation
20
+ - **Language**: English
21
+ - **License**: llama3.2
22
+
23
+ ## Features
24
+
25
+ - Optimized for Intel hardware using OpenVINO
26
+ - Supports text generation inference
27
+ - Maintains original model capabilities for EAD tag generation
28
+ - Integration with PyTorch
29
+
30
+ ## Installation
31
+
32
  First make sure you have optimum-intel installed:
33
 
34
  ```bash
 
43
  model_id = "Geraldine/FineLlama-3.2-3B-Instruct-ead-openvino"
44
  model = OVModelForCausalLM.from_pretrained(model_id)
45
  ```
46
+
47
+ ## Technical Specifications
48
+
49
+ ### Supported Features
50
+ - Text Generation
51
+ - Transformers integration
52
+ - PyTorch compatibility
53
+ - OpenVINO export
54
+ - Inference Endpoints
55
+ - Conversational capabilities
56
+
57
+ ### Model Architecture
58
+ - Base: meta-llama/Llama-3.2-3B-Instruct
59
+ - Fine-tuned: Geraldine/FineLlama-3.2-3B-Instruct-ead
60
+ - Final conversion: OpenVINO optimization
61
+
62
+ ## Usage Examples
63
+
64
+ ```python
65
+ from optimum.intel import OVModelForCausalLM
66
+ from transformers import AutoTokenizer
67
+
68
+ # Load model and tokenizer
69
+ model_id = "Geraldine/FineLlama-3.2-3B-Instruct-ead-openvino"
70
+ model = OVModelForCausalLM.from_pretrained(model_id)
71
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
72
+
73
+ # Generate text
74
+ def generate_ead(prompt):
75
+ inputs = tokenizer(prompt, return_tensors="pt")
76
+ outputs = model.generate(**inputs)
77
+ return tokenizer.decode(outputs[0])
78
+ ```