Theme
AI Resources
ClawMark
ClawMark is a living-world benchmark positioned around multi-day, multimodal coworker agents, with 100 tasks spread across professional domains and tool-backed work environments.
The official repository presents ClawMark as a benchmark for evaluating coworker-style agents across timelines, multimodal evidence, and multiple external systems. This page is for general reference, not a recommendation. Check the original source before relying on the resource.
What it is
A benchmark for coworker agents
ClawMark is framed as an evaluation benchmark rather than an agent product, with the repository centered on long-running work tasks that unfold across several days and several tool environments.
Why it stands out
Multi-day and multimodal pressure
It brings together timelines, real tool backends, multimodal evidence, and deterministic scoring instead of a single-turn prompt or a lighter toy environment.
Availability
Repository with tasks and scoring framework
Public materials are available through a GitHub repository containing the benchmark framework, task layouts, scoring setup, and environment instructions for readers who want to inspect how the evaluation is structured.
Why it matters
Why readers may notice it
ClawMark matters because many agent benchmarks still feel too narrow or too short. A benchmark built around multi-day work, changing state, and multimodal evidence gives readers a more demanding lens for judging agent capability.
What readers may want to know
Where it fits
This project fits in the benchmark and dataset layer rather than the model or app layer. It is more relevant to readers evaluating coworker-style agents and agent robustness than to readers looking for a standalone assistant product.
Reporting note
What appears notable
Based on the official materials, what readers may want to notice is the benchmark's combination of 100 tasks, 13 professional domains, multimodal raw evidence, cross-environment tool coordination, and deterministic rule-based scoring.
Before using
What readers may want to review
Which task domains and external environments are included in the current benchmark release.
What local setup, credentials, and Docker requirements are needed to run the full evaluation stack.
How the reported scores, runs, and token counts were defined before comparing models too directly.
Best fit
Who may find it relevant
Readers following agent benchmarks and coworker-agent evaluation work.
Builders who want a harder benchmark for multi-step, multimodal, tool-using agents.
Less relevant for readers focused only on consumer chat products or single-model demos.
Editorial note
Why it is included here
ClawMark is included because its repository materials show a benchmark style built around longer timelines, multimodal evidence, and changing work environments, making it useful for readers comparing agent evaluation approaches.
Source links
Original materials
Reader note
Before relying on this entry
LifeHubber lists entries for general reader reference only, and this should not be treated as advice. We do not verify every entry in depth, and a listing should not be treated as an endorsement, safety review, professional advice, or confirmation that anything listed is suitable for any specific use, including medical, legal, financial, security, compliance, research, or operational uses. Before relying on anything listed, review the original materials, terms, privacy practices, limitations, and any risks that matter for your own situation.
Get occasional updates when new AI resources are added
More in Datasets
Keep browsing this category
A few more places to continue in datasets.
General365
meituan-longcat/General365
A manually curated benchmark for general reasoning in LLMs, designed around high difficulty, broad task diversity, K-12-scope knowledge, and hybrid scoring.
LARYBench
meituan-longcat/LARYBench
A benchmark for evaluating latent action representations, with pipelines for action semantics, robotic control regression, and broader vision-to-action alignment.
Monitorability Evals
openai/monitorability-evals
An OpenAI evaluation-data release for studying monitorability, with public eval splits, prompt templates, dataset mappings, and metric code from the Monitoring Monitorability paper.
Related in LifeHubber
Continue browsing
Keep browsing across AI, including AI Resources for more tools and projects to explore, AI Ballot for a clearer view of what readers are leaning toward, and AI Guides for help with choosing and using AI tools well.