Shenzhi Wang🌟 Profile
Shenzhi Wang🌟

@ShenzhiWang_THU

1,057
Followers
342
Following
19
Media
277
Statuses

PhD Candidate @Tsinghua_Uni | Developer of 🔥Llama3-8B&70B-Chinese-Chat & 🔥Mistral-7B-v0.3-Chinese-Chat | Research Focuses: RL+LLM+Agent

Joined July 2023
Don't wanna be here? Send us removal request.
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🔥Introducing Llama3-8B-Chinese-Chat, the 1st llama3 model finetuned on English-Chinese datasets via ORPO. 🚀Our model consistently produces better responses for Chinese prompts than the Llama-3-8B-Insturct, and excels in logic, coding, math, and writing.
23
90
377
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
🔥Introducing Gemma-2-9B-Chinese-Chat: the 1st Gemma-2 model tailored for Chinese&English users, fine-tuned on >100K preference pairs! 🚀Our model excels in Chinese prompts, and shows improved logic, coding, math, and writing skills. More info: 🧵⬇️
10
33
190
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🔥We introduce Llama3-8B-Chinese-Chat v2.1! 🚀Compared to v1, the training dataset of v2.1 is 5x larger, and it exhibits significant enhancements, especially in roleplay, function calling, and math capabilities! 😆Don't miss out on our v2.1 model ⬇️
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🔥Introducing Llama3-8B-Chinese-Chat, the 1st llama3 model finetuned on English-Chinese datasets via ORPO. 🚀Our model consistently produces better responses for Chinese prompts than the Llama-3-8B-Insturct, and excels in logic, coding, math, and writing.
23
90
377
4
22
120
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
🔥 After the 9B model, we present Gemma-2-27B-Chinese-Chat: the 1st Gemma2 27B model optimized for Chinese&English, finetuned on >100K preference pairs! 🚀 Our model excels in Chinese, with improved logic, coding, math, and writing skills. More info:🧵⬇️
6
22
121
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
🚀 Presenting Llama3-70B-Chinese-Chat, one of the 1st Llama3 70B models fine-tuned specifically for Chinese! 🏆 Llama3-70B-Chinese-Chat excels on C-Eval and CMMLU, surpassing ChatGPT and matching GPT-4. 🔥 Don't miss out on our model! Learn more: 🧵⬇️
4
25
98
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
🔥Introducing Mistral-7B-v0.3-Chinese-Chat, the 1st Mistral-7B-v0.3 model finetuned for Chinese&English. 🚀Our model consistently produces much better responses for Chinese prompts than the Mistral-7B-v0.3, and excels in math, roleplay, tool use, etc.
4
15
61
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🎉 Our Llama3-8B-Chinese-Chat ranks 14th in overall trending on Huggingface, 2nd among Llama3 derivative models, and 1st among Llama3 Chinese derivative models! Thanks for your support! 🌟 I summarized our updates today in Fig. 3. See the details in the comments below:
Tweet media one
Tweet media two
Tweet media three
7
2
17
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🔥🔥🔥Update: We provide the official 8bit-quantized and fp16 versions of Llama3-8B-Chinese-Chat in the following links, respectively. You are welcome to have a try! 8bit-quantized: fp16:
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🔥Introducing Llama3-8B-Chinese-Chat, the 1st llama3 model finetuned on English-Chinese datasets via ORPO. 🚀Our model consistently produces better responses for Chinese prompts than the Llama-3-8B-Insturct, and excels in logic, coding, math, and writing.
23
90
377
4
2
16
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
🔥🔥Checkout DiveR-CT, Diversity-enhanced Red Teaming with Relaxing Constraints Highlights: 🚩New optimization by constrained RL 🚩Marked superiority in diversity and ASR 🚩Allowing dynamic control of objective weights 🚩Enhancing resiliency in blue team
@AndrewZ45732491
Andrew Zhao
5 months
🚩DiveR-CT, our latest automated red teaming method, forgoes reward maximization bias by constrained RL, also enhance semantic rewards to dynamically adapt. It delivers superior diversity, mitigates overoptimization and generates better data for safetuning
Tweet media one
0
1
4
0
0
7
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
Various GGUF models for Gemma-2-27B-Chinese-Chat:
2
1
6
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Update: We provide the official 8bit-quantized and fp16 versions of Llama3-8B-Chinese-Chat in the following links, respectively. You are welcome to have a try! 8bit-quantized: fp16:
0
2
5
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
🚀 We've just launched the Llama3-70B-Chinese-Chat demo on Gitee! Big thanks to Gitee for their support! 😉 Feel free to give it a try! 🔗⬇️
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
🚀 Presenting Llama3-70B-Chinese-Chat, one of the 1st Llama3 70B models fine-tuned specifically for Chinese! 🏆 Llama3-70B-Chinese-Chat excels on C-Eval and CMMLU, surpassing ChatGPT and matching GPT-4. 🔥 Don't miss out on our model! Learn more: 🧵⬇️
4
25
98
1
2
4
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
We have included various examples generated by shenzhi-wang/Gemma-2-9B-Chinese-Chat, including examples of role playing, function calling, math, RuoZhiBa (弱智吧), safety, writing, and coding, etc. Have a look on our huggingface repo😆
Tweet media one
1
0
5
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
(2/6) 🌟 GGUF 8bit quantized version of the Ollama model: , quick use: ollama run wangshenzhi/llama3-8b-chinese-chat-ollama-q8
2
1
5
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
@huanggou7 @dotey 可以理解为4bit,8bit,16bit量化的版本,bit数越小越快,但性能可能会有所下降
2
1
3
@ShenzhiWang_THU
Shenzhi Wang🌟
7 months
🚀Jailbroken responses could potentially be mitigated through recursive thinking, which encourages LLMs to reconsider their initial outputs. 😄This idea is implemented in our proposed Recursive Contemplation (ReCon) framework. 🔥Check details below!
@AnthropicAI
Anthropic
7 months
New Anthropic research paper: Many-shot jailbreaking. We study a long-context jailbreaking technique that is effective on most large language models, including those developed by Anthropic and many of our peers. Read our blog post and the paper here:
Tweet media one
82
354
2K
1
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
@GoogleDeepMind 🔥We introduced the first fine-tuned Gemma-2-9B model for Chinese&English users! 😄Welcome to have a try!
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
🔥Introducing Gemma-2-9B-Chinese-Chat: the 1st Gemma-2 model tailored for Chinese&English users, fine-tuned on >100K preference pairs! 🚀Our model excels in Chinese prompts, and shows improved logic, coding, math, and writing skills. More info: 🧵⬇️
10
33
190
0
1
4
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The official q4_0 GGUF files for Llama3-70B-Chinese-Chat:
1
1
4
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Examples (1/5)
Tweet media one
0
0
4
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
The q4_0 GGUF version of Llama3-8B-Chinese-Chat-v2.1:
1
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
@minchoi Thank you for your summary! There’s another example 😆
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🔥Introducing Llama3-8B-Chinese-Chat, the 1st llama3 model finetuned on English-Chinese datasets via ORPO. 🚀Our model consistently produces better responses for Chinese prompts than the Llama-3-8B-Insturct, and excels in logic, coding, math, and writing.
23
90
377
0
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
Results on C-Eval and CMMLU:
Tweet media one
1
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Thanks a lot for @llamafactory_ai 's assistance during training!
0
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
The f16 GGUF version of Llama3-8B-Chinese-Chat-v2.1:
1
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
7 months
@AnthropicAI 🚀Jailbroken responses could potentially be mitigated through recursive thinking, which encourages LLMs to reconsider their initial outputs. This idea is implemented in our proposed Recursive Contemplation (ReCon) framework. 🔥Check the details below!
1
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
If you’re in China, you can download our model from Gitee:
0
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
@9hills 感谢建议! 扩充到32K是一个很好的下一步方向,然后100K数据我们会在进一步清洗后开源
1
0
3
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
@ycjcl We haven't thoroughly tested our model on a wide range of benchmarks yet, but we plan to do so. However, as far as I know, no current Chinese benchmarks effectively reflect both the model's instruction-following ability in Chinese and its code-switching issues.
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
typo: “ranks 14th in overall trending on Huggingface” should be “ranks 13th in overall trending on Huggingface” 🤣
0
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Examples (2/5)
Tweet media one
0
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
We have included various examples generated by shenzhi-wang/Gemma-2-27B-Chinese-Chat, including examples of role playing, function calling, math, RuoZhiBa (弱智吧), safety, writing, and coding, etc. Have a look on our huggingface repo😆
Tweet media one
1
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
@GoogleDeepMind 🔥After Gemma-2-9B-Chinese-Chat, we further introduce the first fine-tuned Gemma-2-27B model for Chinese&English users! 📷Welcome to have a try!
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
🔥 After the 9B model, we present Gemma-2-27B-Chinese-Chat: the 1st Gemma2 27B model optimized for Chinese&English, finetuned on >100K preference pairs! 🚀 Our model excels in Chinese, with improved logic, coding, math, and writing skills. More info:🧵⬇️
6
22
121
0
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The examples for math:
Tweet media one
1
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
🚀Our model ranked 7th overall on the Huggingface Trending Model leaderboard, 1st on the Huggingface Trending Chinese Model leaderboard, and 1st on the Huggignface Trending ORPO Model leaderboard! Thank you for your support! 😆Check out our v2.1 if you enjoy v1!
Tweet media one
Tweet media two
Tweet media three
0
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The ollama model for the 8bit-quantized GGUF version of llama3-70b-chinese-chat:
1
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
@meili145 Thank you for your interest in our model! We provide the official 8bit-quantized and fp16 versions of Llama3-8B-Chinese-Chat in the following links, respectively. You are welcome to have a try! 8bit-quantized: fp16:
1
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
Our Gemma-2-9B-Chinese-Chat:
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
🔥Introducing Gemma-2-9B-Chinese-Chat: the 1st Gemma-2 model tailored for Chinese&English users, fine-tuned on >100K preference pairs! 🚀Our model excels in Chinese prompts, and shows improved logic, coding, math, and writing skills. More info: 🧵⬇️
10
33
190
1
0
2
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The examples for coding:
Tweet media one
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
@linshuai_3344 请问您是什么样的场景下指令遵循不行呢?模型的指令遵循能力和模型大小以及场景有很大关系。如果对于某条指令,原版的模型和我们的汉化版本的模型都不行的话,这个可能受限于模型本身的大小以及原版模型的数据;如果原版模型可以但我们汉化版不行的话,欢迎您告知我们是什么场景,我们可以做定向优化。
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
(5/6) 🌟 GGUF FP16 version of the model:
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
The q8_0 GGUF version of Llama3-8B-Chinese-Chat-v2.1:
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
@jack77844923 ollama q4版本大概需要2张A100 80G,一张我不确定是不是可以
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The examples for ruozhiba (弱智吧):
Tweet media one
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
(6/6) 🌟 If you are in China, download here:
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The examples for too use:
Tweet media one
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Examples (4/5)
Tweet media one
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Examples (5/5)
Tweet media one
0
1
1
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The examples for safety:
Tweet media one
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
We have further released Gemma-2-27B-Chinese-Chat! If you love our Gemma-2-9B-Chinese-Chat, don't miss out on our latest Gemma-2-27B-Chinese-Chat! 😆
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
🔥 After the 9B model, we present Gemma-2-27B-Chinese-Chat: the 1st Gemma2 27B model optimized for Chinese&English, finetuned on >100K preference pairs! 🚀 Our model excels in Chinese, with improved logic, coding, math, and writing skills. More info:🧵⬇️
6
22
121
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
(3/6) 🌟 GGUF FP16 version of the Ollama model: , quick use: ollama run wangshenzhi/llama3-8b-chinese-chat-ollama-fp16
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
Very nice job!
@LeapLabTHU
LeapLab@THU
5 months
EfficientTrain++ is accepted by TPAMI2024🤩 🔥An off-the-shelf, easy-to-implement algorithm for training foundation visual backbones efficiently! 🔥1.5−3.0× lossless training/pre-training speedup on ImageNet-1K/22K! Paper&Code:
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
4
8
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
5 months
The examples for writing:
Tweet media one
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
4 months
@jeongmin1604 You can have a try on the following dataset. This is the dataset used for the v1 of our llama3-Chinese-chat model.😆
1
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
Examples (3/5)
Tweet media one
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
(4/6) 🌟 GGUF 8bit quantized version of the model:
0
0
1
@ShenzhiWang_THU
Shenzhi Wang🌟
6 months
If you are in China, you can download our v2.1 model at:
2
1
1