Anthropic Looks To Fund a New, More Comprehensive Generation of AI Benchmarks 8
AI firm Anthropic launched a funding program Monday to develop new benchmarks for evaluating AI models, including its chatbot Claude. The initiative will pay third-party organizations to create metrics for assessing advanced AI capabilities. Anthropic aims to "elevate the entire field of AI safety" with this investment, according to its blog. TechCrunch adds: As we've highlighted before, AI has a benchmarking problem. The most commonly cited benchmarks for AI today do a poor job of capturing how the average person actually uses the systems being tested. There are also questions as to whether some benchmarks, particularly those released before the dawn of modern generative AI, even measure what they purport to measure, given their age.
The very-high-level, harder-than-it-sounds solution Anthropic is proposing is creating challenging benchmarks with a focus on AI security and societal implications via new tools, infrastructure and methods.
The very-high-level, harder-than-it-sounds solution Anthropic is proposing is creating challenging benchmarks with a focus on AI security and societal implications via new tools, infrastructure and methods.
Benchmarks (Score:2)
Re: (Score:2)
The HuggingFace leaderboard doesn't include closed models, such as GPT-4, Claude, Gemini, etc.
Re: (Score:3)
Benchmarks are there for one reason only - to promote the sales of their snake oil.
Re: (Score:2)
If Anthropic wants to fund proper benchmarks, they should pay more taxes and lobby the government to get NIST involved.Then let those guys do the Real Science.
Private money to develop new private benchmarks is a great way to give money to your mates who are itching to do a startup with guaranteed funding.
Hopefully they do like the SEAL Leaderboard... (Score:3)
... and have the 3rd party org keep the benchmark questions private, so that nobody can be accused of training to the test (IMHO more of a problem with third parties doing finetunes than the sort of companies making foundational models, as it's such an easy hack to gain prestige)
AI all the way down (Score:2)
Surely these benchmarks are AI-based.
Why have a human evaluate the AI, when an AI can do it?
Besides, there's nothing ChatGPT can do better than talk about ChatGPT.
censorship benchmark (Score:1)