yixinsong commited on
Commit
2cea2c1
·
1 Parent(s): 65104b5
Files changed (1) hide show
  1. app.py +6 -6
app.py CHANGED
@@ -11,7 +11,7 @@ import uuid
11
  import json
12
 
13
  # Constants
14
- SYSTEM_PROMPT = """You are a helpful assistant."""
15
  device = "cuda" if torch.cuda.is_available() else "cpu"
16
  TITLE = "<h1><center>SmallThinker-3B Chat</center></h1>"
17
  MODEL_PATH = "PowerInfer/SmallThinker-3B-Preview"
@@ -64,8 +64,8 @@ logs_file = Path("logs/") / f"data_{uuid.uuid4()}.json"
64
  logs_folder = logs_file.parent
65
 
66
  # scheduler = CommitScheduler(
67
- # repo_id="PowerInfer/SmallThinker-3B-Preview",
68
- # repo_type="model",
69
  # folder_path=logs_folder,
70
  # path_in_repo="data",
71
  # every=5,
@@ -81,7 +81,7 @@ def stream_chat(
81
  max_new_tokens: int = 1024,
82
  top_p: float = 1.0,
83
  top_k: int = 20,
84
- repetition_penalty: float = 1.2,
85
  ):
86
  # Create new history list with current message
87
  new_history = history + [[message, ""]]
@@ -156,8 +156,8 @@ def main():
156
  )
157
 
158
  with gr.Accordion(label="⚙️ Parameters", open=False):
159
- temperature = gr.Slider(minimum=0, maximum=1, step=0.1, value=0.7, label="Temperature")
160
- max_new_tokens = gr.Slider(minimum=128, maximum=32768, step=128, value=16384, label="Max new tokens")
161
  top_p = gr.Slider(minimum=0.0, maximum=1.0, step=0.1, value=0.8, label="Top-p")
162
  top_k = gr.Slider(minimum=1, maximum=100, step=1, value=20, label="Top-k")
163
  repetition_penalty = gr.Slider(minimum=1.0, maximum=2.0, step=0.1, value=1.1, label="Repetition penalty")
 
11
  import json
12
 
13
  # Constants
14
+ SYSTEM_PROMPT = """You are SmallThinker-3B, a helpful AI assistant. You try to follow instructions as much as possible while being accurate and brief."""
15
  device = "cuda" if torch.cuda.is_available() else "cpu"
16
  TITLE = "<h1><center>SmallThinker-3B Chat</center></h1>"
17
  MODEL_PATH = "PowerInfer/SmallThinker-3B-Preview"
 
64
  logs_folder = logs_file.parent
65
 
66
  # scheduler = CommitScheduler(
67
+ # repo_id=logs_id,
68
+ # repo_type="dataset",
69
  # folder_path=logs_folder,
70
  # path_in_repo="data",
71
  # every=5,
 
81
  max_new_tokens: int = 1024,
82
  top_p: float = 1.0,
83
  top_k: int = 20,
84
+ repetition_penalty: float = 1.1,
85
  ):
86
  # Create new history list with current message
87
  new_history = history + [[message, ""]]
 
156
  )
157
 
158
  with gr.Accordion(label="⚙️ Parameters", open=False):
159
+ temperature = gr.Slider(minimum=0, maximum=1, step=0.1, value=0.3, label="Temperature")
160
+ max_new_tokens = gr.Slider(minimum=128, maximum=32768, step=128, value=1024, label="Max new tokens")
161
  top_p = gr.Slider(minimum=0.0, maximum=1.0, step=0.1, value=0.8, label="Top-p")
162
  top_k = gr.Slider(minimum=1, maximum=100, step=1, value=20, label="Top-k")
163
  repetition_penalty = gr.Slider(minimum=1.0, maximum=2.0, step=0.1, value=1.1, label="Repetition penalty")