Chart captions are crucial for helping readers understand and retain data, especially for those with visual disabilities. However, writing effective captions is a time-consuming task. To address this issue, MIT researchers have developed a dataset called VisText to improve automatic captioning systems. By using this dataset, machine-learning models can be trained to generate precise and semantically rich captions that describe data trends and complex patterns. The researchers found that their models outperformed other autocaptioning systems in effectively captioning charts. The goal is to provide VisText as a tool for researchers working on chart autocaptioning, with the aim of improving accessibility for people with visual disabilities. The dataset incorporates human values to ensure that the generated captions meet users’ needs. The study will be presented at the Annual Meeting of the Association for Computational Linguistics.