ChatGLM-6B Description
ChatGLM-6B, a Chinese-English bilingual dialogue model based on General Language Model architecture (GLM), has 6.2 billion parameters. Users can deploy model quantization locally on consumer-grade graphic cards (only 6GB video memory required at INT4 quantization levels). ChatGLM-6B is based on technology similar to ChatGPT and optimized for Chinese dialogue and Q&A. After approximately 1T identifiers for Chinese and English bilingual training and supplemented with supervision and fine-tuning as well as feedback self-help and human feedback reinforcement learning, ChatGLM-6B, with 6.2 billion parameters, has been able generate answers that are in line with human preference.
Pricing
Pricing Starts At:
Free
Pricing Information:
Open source
Free Version:
Yes
Integrations
No Integrations at this time
Company Details
Company:
Zhipu AI
Year Founded:
2019
Headquarters:
China
Website:
chatglm.cn/
Recommended Products
Product Details
Platforms
SaaS
Windows
Mac
Linux
On-Premises
Type of Training
Documentation