Safetensors
mistral
piuzha commited on
Commit
dc599b9
·
verified ·
1 Parent(s): 63f187d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md CHANGED
@@ -2,6 +2,12 @@
2
  license: apache-2.0
3
  ---
4
 
 
 
 
 
 
 
5
  ## Chat Template
6
 
7
  The chat template is formatted as:
@@ -17,3 +23,34 @@ How are you doing?
17
  <|assistant|>
18
  Thank you for asking! As an AI, I don't have feelings, but I'm functioning normally and ready to assist you. How can I help you today?<|endoftext|>
19
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: apache-2.0
3
  ---
4
 
5
+ <h1 align="center"> Moxin 7B Reasoning </h1>
6
+
7
+ <p align="center"> <a href="https://github.com/moxin-org/Moxin-LLM">Home Page</a> &nbsp&nbsp | &nbsp&nbsp <a href="https://arxiv.org/abs/2412.06845">Technical Report</a> &nbsp&nbsp | &nbsp&nbsp <a href="https://huggingface.co/moxin-org/Moxin-7B-LLM">Base Model</a> &nbsp&nbsp | &nbsp&nbsp <a href="https://huggingface.co/moxin-org/Moxin-7B-Chat">Chat Model</a> &nbsp&nbsp | &nbsp&nbsp <a href="https://huggingface.co/moxin-org/Moxin-7B-Instruct">Instruct Model</a> &nbsp&nbsp | &nbsp&nbsp <a href="https://huggingface.co/moxin-org/Moxin-7B-Reasoning">Reasoning Model</a> &nbsp&nbsp | &nbsp&nbsp <a href="https://huggingface.co/moxin-org/Moxin-7B-VLM">VLM Model</a> </p>
8
+
9
+
10
+
11
  ## Chat Template
12
 
13
  The chat template is formatted as:
 
23
  <|assistant|>
24
  Thank you for asking! As an AI, I don't have feelings, but I'm functioning normally and ready to assist you. How can I help you today?<|endoftext|>
25
  ```
26
+
27
+
28
+ ## Inference
29
+
30
+ You can use the following code to run inference with the model.
31
+
32
+ ```
33
+ import transformers
34
+ import torch
35
+
36
+ model_id = "moxin-org/Moxin-7B-Reasoning"
37
+ pipeline = transformers.pipeline(
38
+ "text-generation",
39
+ model=model_id,
40
+ model_kwargs={"torch_dtype": torch.bfloat16},
41
+ device_map="auto",
42
+ )
43
+
44
+ messages = [
45
+ {"role": "system", "content": "You are a helpful AI assistant!"},
46
+ {"role": "user", "content": "How are you doing?"},
47
+ ]
48
+
49
+ outputs = pipeline(
50
+ messages,
51
+ max_new_tokens=1024,
52
+ )
53
+
54
+ print(outputs[0]["generated_text"][-1])
55
+
56
+ ```