NeMo
English
nvidia
steerlm
llama3
reward model
zhilinw commited on
Commit
2992f66
1 Parent(s): e9440c9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -47,6 +47,7 @@ Llama3-70B-SteerLM-RM is trained with NVIDIA [NeMo-Aligner](https://github.com/N
47
 
48
  | Model | Type of Model| Overall | Chat | Chat-Hard | Safety | Reasoning |
49
  |:-----------------------------|:----------------|:-----|:----------|:-------|:----------|:-----------------------|
 
50
  | ArmoRM-Llama3-8B-v0.1 | Trained with GPT4 Generated Data| 90.8 | 96.9 | 76.8 | 92.2 | 97.3 |
51
  | Cohere May 2024 | Proprietary LLM | 89.5 | 96.4 | 71.3 | 92.7 | 97.7 |
52
  | _**Llama3-70B-SteerLM-RM**_ | Trained with Permissive Licensed Data | 88.8 | 91.3 | 80.3 | **92.8** | 90.7 |
@@ -58,7 +59,7 @@ Llama3-70B-SteerLM-RM is trained with NVIDIA [NeMo-Aligner](https://github.com/N
58
  | Llama3 70B Instruct | Trained with Permissive Licensed Data | 76.0 | 97.6 | 58.9 | 69.2 | 78.5 |
59
 
60
 
61
- Last updated: 1 Jun 2024
62
 
63
  Note that we only consider the first four categories in RewardBench, because the optional fifth category (Prior Sets) is
64
  1. Heavily towards models trained on Anthropic HHH, Anthropic Helpful, OpenAI Summarize and Stanford Human Preferences (constituent datasets for the Prior Sets category) and therefore can be easily gamed (see About page on RewardBench)
 
47
 
48
  | Model | Type of Model| Overall | Chat | Chat-Hard | Safety | Reasoning |
49
  |:-----------------------------|:----------------|:-----|:----------|:-------|:----------|:-----------------------|
50
+ | _**Nemotron-4-340B-RM**_ | Trained with Permissive Licensed Data | **92.0** | 95.8 | **87.1** | 91.5 | 93.7 |
51
  | ArmoRM-Llama3-8B-v0.1 | Trained with GPT4 Generated Data| 90.8 | 96.9 | 76.8 | 92.2 | 97.3 |
52
  | Cohere May 2024 | Proprietary LLM | 89.5 | 96.4 | 71.3 | 92.7 | 97.7 |
53
  | _**Llama3-70B-SteerLM-RM**_ | Trained with Permissive Licensed Data | 88.8 | 91.3 | 80.3 | **92.8** | 90.7 |
 
59
  | Llama3 70B Instruct | Trained with Permissive Licensed Data | 76.0 | 97.6 | 58.9 | 69.2 | 78.5 |
60
 
61
 
62
+ Last updated: 12 Jun 2024
63
 
64
  Note that we only consider the first four categories in RewardBench, because the optional fifth category (Prior Sets) is
65
  1. Heavily towards models trained on Anthropic HHH, Anthropic Helpful, OpenAI Summarize and Stanford Human Preferences (constituent datasets for the Prior Sets category) and therefore can be easily gamed (see About page on RewardBench)