Product Image

Egyptian Plant Leaf Image (EPLID) Dataset

A comprehensive collection of plant leaf images, each representing a distinct class with unique characteristics, making it well-suited for the development of AI models for plant recognition and disease detection.

    • Labeling Type: leaf
    • Data Format: Image
    • Data Type: Synthetic Data

Main Product

Data Quantity (Samples)

Total Price

$ 6,400

(VAT Included)

Looking for custom-made dataset or researcher-accessible data? Please contact us for inquiries.

About Dataset

1) Data Introduction

• The Egyptian Plant Leaf Image Dataset (EPLID) is a plant leaf image-based computer vision dataset developed for the purposes of monitoring plant health, detecting diseases, and identifying plant species. It consists of eight distinct classes — Apple, Berry, Fig, Guava, Orange, Palm, Persimmon, and Tomato — each organized into separate folders for convenient labeling and model training.

2) Data Utilization

(1) Characteristics of the Egyptian Plant Leaf Image Dataset (EPLID): • The dataset contains images captured in real-world conditions, making it well-suited for the development of AI models that can be applied in practical agricultural environments. • Each image clearly presents the leaf's texture, color, and venation, enabling high-precision applications in plant recognition and disease detection tasks. (2) Applications of the Egyptian Plant Leaf Image Dataset (EPLID): • Development of AI models for plant disease detection: The dataset can be used to train deep learning models that automatically identify plant diseases by learning abnormal leaf patterns such as spots, discoloration, and surface damage. • Construction of crop classification and cultivar identification systems: The dataset can support the development of models that classify different crop types and identify plant varieties based on their leaf characteristics.

Meta Data

DomainetcZoodata formatsImage
Zoodata volume1000 itemsRegistration Date2025.06.16
Zoodata typeSynthetic dataExistence of labelingExist
Labeling typeleafLabeling formatsJSON

Normal

Performance 1
Very Good

Outstanding

Performance 2
100

Data Samples 5

Data sample
Data sample
Data sample
Data sample

Utility

Downstream Classification (▲)KID (▼)One Class Classification (▼)
Total000
SuitabilityOKOKOK

The higher the value, the better (▲)

Model Performance

Downstream classification accuracy is an indicator used to evaluate the usefulness of synthetic data. It measures whether synthetic data performs similarly to real data. The method involves training the same model separately on real data and synthetic data, and then comparing the accuracies of the two models. Interpretation: A high accuracy rate means that the model trained on synthetic data performs similarly to the one trained on real data, indicating that the synthetic data is of high quality and well represents the real data.

The closer to zero or the lower the value, the better (▼)

Quality

KID (Kernel Inception Distance) is a metric used to evaluate the similarity between generated images and real images. It compares the differences between the two sample distributions using Kernel Mean Embedding, without assuming a normal distribution. Interpretation: A lower KID score suggests that generated images are more similar to real images, with a score close to 0 being ideal. Specifically, a score below 0.01 indicates very high similarity.


Privacy

LPIPS (▲)SSIM (▼)
Total00
SuitabilityOKOK

The higher the value, the better (▲)

Perceptual Similarity

Learned Perceptual Image Patch Similarity (LPIPS) is a metric used to measure the visual similarity between two images by utilizing neural networks to extract key features and calculate the distance between them. High LPIPS value: Indicates high similarity between images, raising the risk of information leakage. Low LPIPS value: Suggests that synthetic images are perceptually different from real images, indicating a lower risk of sensitive information leakage.

The closer to zero or the lower the value, the better (▼)

Structural Similarity

The Structural Similarity Index Measure (SSIM) is a metric used to assess the similarity between two images. It is primarily used to compare the quality of a restored or compressed image with the original image. SSIM measures visual similarity by considering brightness, contrast, and structure. High SSIM value (0.9 or above): Indicates that the synthetic image is very similar to the real image, which may increase the risk of information leakage.
Low SSIM value (0.6 or below): Indicates low similarity and reduced risk of leakage.

Premium Report Information

If you purchase the premium report product, you will be able to view the analysis results of a more detailed dataset.
select premium data

Premium dataset sample