Hjgugugjhuhjggg commited on
Commit
5fac4ae
·
verified ·
1 Parent(s): 66fb7b1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +38 -1
app.py CHANGED
@@ -26,7 +26,44 @@ global_data = {
26
  'bos': 'bos_token',
27
  'sep': 'sep_token',
28
  'cls': 'cls_token',
29
- 'mask': 'mask_token'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  }
31
  }
32
 
 
26
  'bos': 'bos_token',
27
  'sep': 'sep_token',
28
  'cls': 'cls_token',
29
+ 'mask': 'mask_token',
30
+ 'n_ctx': 'n_ctx_token',
31
+ 'vocab_size': 'vocab_size_token',
32
+ 'n_embd': 'n_embd_token',
33
+ 'n_head': 'n_head_token',
34
+ 'n_layer': 'n_layer_token',
35
+ 'n_vocab': 'n_vocab_token',
36
+ 'block_size': 'block_size_token',
37
+ 'n_rot': 'n_rot_token',
38
+ 'rope_dim': 'rope_dim_token',
39
+ 'rope_scaling': 'rope_scaling_token',
40
+ 'n_positions': 'n_positions_token',
41
+ 'use_cache': 'use_cache_token',
42
+ 'use_parallel_inference': 'use_parallel_inference_token',
43
+ 'parallel_inference_count': 'parallel_inference_count_token',
44
+ 'use_mlock': 'use_mlock_token',
45
+ 'use_mmap': 'use_mmap_token',
46
+ 'use_cpu': 'use_cpu_token',
47
+ 'f16_kv': 'f16_kv_token',
48
+ 'f16_quant': 'f16_quant_token',
49
+ 'f16_output': 'f16_output_token',
50
+ 'use_flash_attn': 'use_flash_attn_token',
51
+ 'max_seq_len': 'max_seq_len_token',
52
+ 'do_sample': 'do_sample_token',
53
+ 'top_k': 'top_k_token',
54
+ 'top_p': 'top_p_token',
55
+ 'temperature': 'temperature_token',
56
+ 'num_return_sequences': 'num_return_sequences_token',
57
+ 'use_repetition_penalty': 'use_repetition_penalty_token',
58
+ 'repetition_penalty': 'repetition_penalty_token',
59
+ 'no_repeat_ngram_size': 'no_repeat_ngram_size_token',
60
+ 'bad_words_ids': 'bad_words_ids_token',
61
+ 'use_token_logging': 'use_token_logging_token',
62
+ 'use_tensor_parallel': 'use_tensor_parallel_token',
63
+ 'tensor_parallel_size': 'tensor_parallel_size_token',
64
+ 'use_gpu_memory_growth': 'use_gpu_memory_growth_token',
65
+ 'use_multi_gpu_inference': 'use_multi_gpu_inference_token',
66
+ 'multi_gpu_inference_count': 'multi_gpu_inference_count_token'
67
  }
68
  }
69