Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
DeepSeek has launched DeepSeek-V3.1-Terminus, an upgrade to the V3.1 architecture that integrates user suggestions to enhance output stability, consistency, and overall agent performance. This new version significantly decreases the occurrences of mixed Chinese and English characters as well as unintended distortions, leading to a cleaner and more uniform language generation experience. Additionally, the update revamps both the code agent and search agent subsystems to deliver improved and more dependable performance across various benchmarks. DeepSeek-V3.1-Terminus is available as an open-source model, with its weights accessible on Hugging Face, making it easier for the community to leverage its capabilities. The structure of the model remains consistent with DeepSeek-V3, ensuring it is compatible with existing deployment strategies, and updated inference demonstrations are provided for users to explore. Notably, the model operates at a substantial scale of 685B parameters and supports multiple tensor formats, including FP8, BF16, and F32, providing adaptability in different environments. This flexibility allows developers to choose the most suitable format based on their specific needs and resource constraints.
Description
GLM-4.7 Flash serves as a streamlined version of Z.ai's premier large language model, GLM-4.7, which excels in advanced coding, logical reasoning, and executing multi-step tasks with exceptional agentic capabilities and an extensive context window. This model, rooted in a mixture of experts (MoE) architecture, is fine-tuned for efficient inference, striking a balance between high performance and optimized resource utilization, thus making it suitable for deployment on local systems that require only moderate memory while still showcasing advanced reasoning, programming, and agent-like task handling. Building upon the advancements of its predecessor, GLM-4.7 brings forth enhanced capabilities in programming, reliable multi-step reasoning, context retention throughout interactions, and superior workflows for tool usage, while also accommodating lengthy context inputs, with support for up to approximately 200,000 tokens. The Flash variant successfully maintains many of these features within a more compact design, achieving competitive results on benchmarks for coding and reasoning tasks among similarly-sized models. Ultimately, this makes GLM-4.7 Flash an appealing choice for users seeking powerful language processing capabilities without the need for extensive computational resources.
API Access
Has API
API Access
Has API
Pricing Details
Free
Free Trial
Free Version
Pricing Details
Free
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
DeepSeek
Founded
2023
Country
China
Website
api-docs.deepseek.com/news/news250922
Vendor Details
Company Name
Z.ai
Founded
2019
Country
China
Website
docs.z.ai/guides/llm/glm-4.7#glm-4-7-flash