--- license: cc-by-3.0 tags: - agent - workflow - multimodal - spreadsheet - pdf - image - code - finance - accouning modalities: - text - spreadsheet - pdf - image - code configs: - config_name: Finch_Dataset_All data_files: - split: test path: - finch_workflows_test.jsonl --- ![Finch cover figure](figs/finch_workflow.jpeg) # Finch: Benchmarking Finance & Accounting across Spreadsheet-Centric Enterprise Workflows This repository contains the dataset for **Finch**, an enterprise-level benchmark for evaluating an agent’s ability to act like a skilled finance & accounting expert on real-world workflows. * **Paper**: _to be added_ * **Project Page**: https://huggingface.co/datasets/FinWorkBench/Finch * **Code**: https://github.com/FinWorkBench --- ## Dataset Description Finch focuses on **messy and long-horizon finance & accounting workflows** that span: > data entry/import, structuring/formatting, web search, cross-sheet/file retrieval, calculation, financial modeling, validation, translation, visualization, and reporting. The workflows are derived from **real-world enterprise workspaces** (primarily Enron, as well as World Bank, Canadian/British government agencies, and other corporations), including: - Enterprise **email threads** where collaborators naturally describe, discuss, and track workflows - Large and messy **spreadsheets** with multimodal artifacts including text, tables, formulas, charts, pivots, images, etc - Interlinked **PDFs and documents** that provide additional business context We adopt a three-step workflow labeling process: 1. **Inducing workflow types** from real collaborative context in enterprise email threads. 2. **Deriving concrete workflow instances** by analyzing changes across spreadsheet versions. 3. **Conductin meticulous expert annotation** of task instructions, input files, and reference outputs, involving hundreds of hours of expert work. This process yields **172 enterprise-grade workflows—primarily multi-task composite** — each with carefully written instructions and aligned input/reference files, capturing the intrinsic **compositional, messy, multimodal, and collaborative nature** of real-world finance & accounting work. In this release, we provide full annotations for the first 72 workflows, with the remaining 100 to be released in a subsequent update. Experiment results show that even frontier agents solve fewer than 30% of the workflows, revealing a substantial performance gap for real-world enterprise scenarios. --- ## 📁 Dataset Structure The instruction-tuning corpus is released in **JSONL** format. Each line corresponds to one **workflow-centric example**: ```json { "id": "", "instruction_en": "", "source_files": ["", "..."], "source_files_urls": ["", "..."], "reference_outputs": { "files": [""], "text": "" }, "reference_file_urls": [""], "task_type": "", "business_type": "" } ``` --- ## 📣 Feedback & Issues If you find any issues with the dataset or have suggestions, please open a discussion in the **Community** tab — we value your feedback!