Codesota · Models · DAT-DETWan et al. (Baidu)3 results · 1 benchmarks
Model card

DAT-DET.

Wan et al. (Baidu)open-sourceUnknown paramsInteractive attention transformer for multi-granularity text detection

Detection head of DAT (Dual-granularity Attention Transformer). Unified model for text at stroke, word, line, paragraph levels. ICML 2024. arxiv:2405.19765

§ 01 · Benchmarks

Every benchmark DAT-DET has a recorded score for.

#BenchmarkArea · TaskMetricValueRankDateSource
01Total-TextComputer Vision · Scene Text Detectionf-measure91.0%#2/332024-05-30source ↗
02Total-TextComputer Vision · Scene Text Detectionprecision94.0%#2/302024-05-30source ↗
03Total-TextComputer Vision · Scene Text Detectionrecall88.2%#3/302024-05-30source ↗
Rank column shows this model’s position vs all other models scored on the same benchmark + metric (competitors after the slash). #1 in red means current SOTA. Sorted by rank, then newest result.
§ 02 · Strengths by area

Where DAT-DET actually performs.

Computer Vision
1
benchmark
avg rank #2.3
§ 03 · Papers

1 paper with results for DAT-DET.

  1. 2024-05-30· Computer Vision· 3 results

    Towards Unified Multi-granularity Text Detection with Interactive Attention

§ 04 · Related models

Other Wan et al. (Baidu) models scored on Codesota.

DAT-SEG
Unknown params · 0 results
§ 05 · Sources & freshness

Where these numbers come from.

arxiv
3
results
3 of 3 rows marked verified.