The CADReview dataset is introduced to support the CAD review task—automatically detecting and correcting errors in CAD programs based on reference images. This task is fundamentally challenging for current multimodal large language models (MLLMs) due to the intricate alignment required between code and 3D geometry. The major challenges include:
- Component-level reasoning: 3D objects consist of multiple geometric components, each mapped to specific code blocks that must be interpreted accurately.
- Internal structure analysis: Some errors occur in visually hidden internal components, demanding programmatic analysis rather than visual inspection alone.
- Precise code transformation: Correcting errors requires geometric reasoning and spatial transformations reflected directly in the code (e.g., translation, rotation).
To facilitate progress in this domain, we construct the CADReview dataset with the following characteristics:
- Large-scale and diverse: Comprising 20K program-image pairs, the dataset includes both human-made and machine-generated CAD programs with realistic and diverse 3D object structures.
- Error-rich examples: Each program may contain one or more of eight distinct error types (e.g., logic, primitive, position, constant), covering real-world failure modes in CAD design workflows.
- High geometric complexity: Human-made programs feature advanced constructs such as macros, control flows, and boolean operations.
- Multiview reference images: Every sample is paired with three rendered design drawings from varied viewpoints to enhance visual alignment and support complex error detection.
Overall, CADReview serves as a comprehensive benchmark for evaluating AI systems on real-world CAD program review tasks, emphasizing both accuracy and spatial reasoning.