Product Image

Manufacturing GraphRAG Benchmark Dataset

Benchmark dataset of 10K expert-level questions testing RAG models on sustainability, advanced manufacturing, and AI policy reasoning using NIST AMS and Circular Economy sources.

    • Labeling Type: answer
    • Data Format: Tabular
    • Data Type: Benchmark

Main Product

Data Quantity (Samples)

Total Price

$ 29,000

(VAT Included)

Looking for custom-made dataset or researcher-accessible data? Please contact us for inquiries.

Data Information

About
Test
Model
Meta Data
Data Samples
Benchmark Path
01.Data Introduction
This dataset evaluates retrieval-augmented generation (RAG) models on complex QA tasks drawn from advanced manufacturing, sustainability, and AI integration domains. It incorporates references to NIST Advanced Manufacturing Series (AMS) reports, Circular Economy frameworks, additive manufacturing, investment analysis (NPV, IRR), and workforce development initiatives supported by U.S. federal agencies such as NIST and NSF.
02.Data Utilization
Designed for benchmarking Naive RAG vs. Graph RAG models in industrial, sustainability, and policy-oriented question answering. Supports evaluation of reasoning across technical standards (ISO, ASTM, IEC), government initiatives (NIST AMS 100-47, AMS 500-1, AMS 100-63), and sustainability transitions (Circular Economy, Bioeconomy, Net Zero Manufacturing).
03.Key Features
Icon 1
Icon 2
Icon 3
Icon 4
Icon 5
Icon 6
Icon 7
Icon 8
Icon 9
Icon 10
10K Manufacturing, Sustainability, and AI Policy QA Pairs
Covers interconnected topics such as Circular Economy transitions, NIST-led AI in manufacturing (AMS 100-47), additive manufacturing process chains, net present value (NPV) in supply chain investment, smart manufacturing reference architectures (SMS/SMMS), and U.S. STEM workforce programs. Tests multi-hop reasoning across sustainability policy, engineering innovation, and financial analysis contexts.

This dataset was created by ⓒ CUBIG, based in part on publicly available data.

Unauthorized reproduction, redistribution, or resale of the dataset is prohibited.

Sources

    Cubig Logo
    Test RAG Architectures
    with a Trusted Benchmark Dataset
    Pick a question to see how Naive RAG and Graph RAG perform on this benchmark.
    Start Exploring
    01.Naive RAG
    • Vector DB: FAISS
    • Embedding model: text-embedding-3-small
    • LLM model: GPT-4.1
    Naive RAG Pipeline
    #1.
    Question
    #2.
    Vector Store
    #3.
    Retrieved Docs
    #4.
    Prompt
    #5.
    LLM Generation
    #6.
    Response
    02.Graph RAG
    • Graph Construction: NetworkX
    • Embedding model: text-embedding-3-small
    • LLM model: GPT-4.1
    Graph RAG Pipeline
    #1.
    Question
    #2.
    Entity Extraction
    #3.
    Subgraph
    #4.
    Combined Context
    #5.
    LLM Generation
    #6.
    Response
    03.Naive RAG vs Graph RAG

    Out of 10000 benchmark questions,

    • Graph RAG achieved 100% accuracy (10000 correct answers),
    • Naive RAG reached only 28% accuracy (2789 correct answers).

    Accuracy Comparison

    Graph RAG
    Naive RAG
    050%100%
    Domain
    etc
    Zoodata Volume
    1000 items
    Zoodata Type
    Benchmark data
    Labeling Type
    answer
    Zoodata Formats
    Tabular
    Registration Date
    2025-10-27
    Existence of labeling
    Exist
    Labeling Formats
    JSON
    Sample Data
    Sample data can be viewed after logging in.

    See how this benchmark
    could work for your model.