R-TF-028-005 AI Development Report
Table of contents
- Introduction
- Data Management
- Model Development and Validation
- ICD Category Distribution and Binary Indicators
- Erythema Intensity Quantification
- Desquamation Intensity Quantification
- Induration Intensity Quantification
- Pustule Intensity Quantification
- Crusting Intensity Quantification
- Xerosis Intensity Quantification
- Swelling Intensity Quantification
- Oozing Intensity Quantification
- Excoriation Intensity Quantification
- Lichenification Intensity Quantification
- Wound Characteristic Assessment
- Inflammatory Nodular Lesion Quantification
- Acneiform Lesion Type Quantification
- Acneiform Inflammatory Lesion Quantification
- Hive Lesion Quantification
- Body Surface Segmentation
- Wound Surface Quantification
- Hair Loss Surface Quantification
- Nail Lesion Surface Quantification
- Hypopigmentation/Depigmentation Surface Quantification
- Skin Surface Segmentation
- Surface Area Quantification
- Acneiform Inflammatory Pattern Identification
- Follicular and Inflammatory Pattern Identification
- Inflammatory Pattern Identification
- Inflammatory Pattern Indicator
- Dermatology Image Quality Assessment (DIQA)
- Fitzpatrick Skin Type Identification
- Domain Validation
- Body Site Identification
- Summary and Conclusion
- State of the Art Compliance and Development Lifecycle
- AI Risks Assessment Report
- Related Documents
Introduction
Context
This report documents the development, verification, and validation of the AI algorithm package for the Legit.Health Plus medical device. The development process was conducted in accordance with the procedures outlined in GP-028 AI Development and followed the methodologies specified in the R-TF-028-002 AI Development Plan.
The algorithms are designed as offline (static) models. They were trained on a fixed dataset prior to release and do not adapt or learn from new data after deployment. This ensures predictable and consistent performance in the clinical environment.
Algorithms Description
The Legit.Health Plus device incorporates 31 AI models that work together to fulfill the device's intended purpose. A comprehensive description of all models, their clinical objectives, and performance specifications is provided in R-TF-028-001 AI/ML Description.
The AI algorithm package includes:
Clinical Models (directly fulfilling the intended purpose):
- ICD Category Distribution and Binary Indicators (1 model): Provides interpretative distribution of ICD-11 categories and binary risk indicators (malignancy, pre-malignant, associated with malignancy, pigmented lesion, urgent referral, high-priority referral).
- Visual Sign Intensity Quantification Models (10 models): Quantify the intensity of clinical signs including erythema, desquamation, induration, pustule, crusting, xerosis, swelling, oozing, excoriation, and lichenification.
- Wound Characteristic Assessment (1 model): Evaluates wound tissue types and characteristics.
- Lesion Quantification Models (4 models):
- Inflammatory Nodular Lesion Quantification
- Acneiform Lesion Type Quantification (multi-class detection of papules, pustules, comedones, nodules, cysts)
- Inflammatory Lesion Quantification
- Hive Lesion Quantification
- Surface Area Quantification Models (6 models):
- Body Surface Segmentation
- Wound Surface Quantification
- Hair Loss Surface Quantification
- Nail Lesion Surface Quantification
- Hypopigmentation/Depigmentation Surface Quantification
- Surface Area Quantification (generic measurement model)
- Pattern Identification Models (4 models):
- Acneiform Inflammatory Pattern Identification
- Follicular and Inflammatory Pattern Identification
- Inflammatory Pattern Identification
- Inflammatory Pattern Indicator
Non-Clinical Models (supporting proper functioning - 5 models):
- Dermatology Image Quality Assessment (DIQA): Ensures image quality is suitable for analysis.
- Fitzpatrick Skin Type Identification: Identifies skin phototype to support equity and bias mitigation.
- Domain Validation: Verifies images are within the validated domain.
- Skin Surface Segmentation: Identifies skin regions for analysis.
- Body Site Identification: Determines anatomical location.
Total: 26 Clinical Models + 5 Non-Clinical Models = 31 Models
This report focuses on the development methodology, data management processes, and validation results for all models. Each model shares a common data foundation but may require specific annotation procedures as detailed in the respective data annotation instructions.
AI Standalone Evaluation Objectives
The standalone validation aimed to confirm that all AI models meet their predefined performance criteria as outlined in R-TF-028-001 AI/ML Description.
Performance specifications and success criteria vary by model type and are detailed in the individual model sections of this report. All models were evaluated on independent, held-out test sets that were not used during training or model selection.
Data Management
Overview
The development of all AI models in the Legit.Health Plus device relies on a comprehensive dataset compiled from multiple sources and annotated through a multi-stage process. This section describes the general data management workflow that applies to all models, including collection, foundational annotation (ICD-11 mapping), and partitioning. Model-specific annotation procedures are detailed in the individual model sections.
Data Collection
The dataset was compiled from multiple distinct sources as detailed in R-TF-028-003 Data Collection Instructions - Custom Gathered Data and R-TF-028-003 Data Collection Instructions - Archive Data:
- Archive Data: Images sourced from reputable online sources and private institutions.
- Custom Gathered DAta: Images collected under formal protocols at clinical sites.
This combined approach resulted in a comprehensive dataset covering diverse demographic characteristics (age, sex, Fitzpatrick skin types I-VI), anatomical sites, imaging conditions, and pathological conditions.
Dataset summary:
- Total images: [NUMBER OF IMAGES] (to be completed)
- Sources: 17
- ICD-11 categories: [NUMBER OF CATEGORIES] (to be completed)
- Demographic diversity: Ages [AGE RANGE], Fitzpatrick types I-VI, global geographic representation
| ID | Dataset Name | Type | Description | ICD-11 Mapping | Crops | Diff. Dx | Sex | Age |
|---|---|---|---|---|---|---|---|---|
| 1 | Torrejon-HCP-diverse-conditions | Multiple | Dataset of skin images by physicians with good photographic skills | ✓ Yes | Varies | ✓ | ✓ | ✓ |
| 2 | Abdominal-skin | Archive | Small dataset of abdominal pictures with segmentation masks for `Non-specific lesion` class | ✗ No | Yes (programmatic) | — | — | — |
| 3 | Basurto-Cruces-Melanoma | Custom gathered | Clinical validation study dataset (`MC EVCDAO 2019`) | ✓ Yes | Yes (in-house crops) | — | ✓ | ✓ |
| 4 | BI-GPP (batch 1) | Archive | Small set of GPP images from Boehringer Ingelheim (first batch) | ✓ Yes | No | — | — | — |
| 5 | BI-GPP (batch 2) | Archive | Large dataset of GPP images from Boehringer Ingelheim (second batch) | ✓ Yes | Yes (programmatic) | — | ✓ | ✓ |
| 6 | Chiesa-dataset | Archive | Sample of head and neck lesions (Medela et al., 2024) | ✓ Yes | Yes (in-house crops) | — | ◐ | ◐ |
| 7 | Figaro 1K | Archive | Hair style classification and segmentation dataset, repurposed for `Non-specific finding` | ✗ No | Yes (in-house crops) | — | — | — |
| 8 | Hand Gesture Recognition (HGR) | Archive | Small dataset of hands repurposed for non-specific images | ✗ No | Yes (programmatic) | — | — | — |
| 9 | IDEI 2024 (pigmented) | Archive | Prospective and retrospective studies at IDEI (DERMATIA project), pigmented lesions only | ✓ Yes | Yes (programmatic) | — | ✓ | ◐ |
| 10 | Manises-HS | Archive | Large collection of hidradenitis suppurativa images | ✗ No | Not yet | — | ✓ | ✓ |
| 11 | Nails segmentation | Archive | Small nail segmentation dataset repurposed for `non-specific lesion` | ✗ No | Yes (programmatic) | — | — | — |
| 12 | Non-specific lesion V2 | Archive | Small representative collection repurposed for `non-specific lesion` | ✗ No | Yes (programmatic) | — | — | — |
| 13 | Osakidetza-derivation | Archive | Clinical validation study dataset (`DAO Derivación O 2022`) | ✓ Yes | Yes (in-house crops) | ◐ | ✓ | ✓ |
| 14 | Ribera ulcers | Archive | Collection of ulcer images from Ribera Salud | ✗ No | Yes (from wound masks, not all) | — | — | — |
| 15 | Transient Biometrics Nails V1 | Archive | Biometric dataset of nail images | ✗ No | Yes (programmatic) | — | — | — |
| 16 | Transient Biometrics Nails V2 | Archive | Biometric dataset of nail images | ✗ No | No (close-ups) | — | — | — |
| 17 | WoundsDB | Archive | Small chronic wounds database | ✓ Yes | No | — | ✓ | ◐ |
Total datasets: 51 | With ICD-11 mapping: 37
Legend: ✓ = Yes | ◐ = Partial/Pending | — = No
Foundational Annotation: ICD-11 Mapping
Before any model-specific training could begin, all diagnostic labels across all data sources were standardized to the ICD-11 classification system. This foundational annotation step is required for all models and is detailed in R-TF-028-004 Data Annotation Instructions - ICD-11 Mapping.
The ICD-11 mapping process involved:
- Label Extraction: Extracting all unique diagnostic labels from each data source
- Standardization: Mapping source-specific labels (abbreviations, alternative spellings, legacy coding systems) to standardized ICD-11 categories
- Clinical Validation: Expert dermatologist review and validation of all mappings
- Visible Category Consolidation: Grouping ICD-11 codes that cannot be reliably distinguished based on visual features alone into unified "Visible ICD-11" categories
This standardization ensures:
- Consistent diagnostic ground truth across all data sources
- Clinical validity and regulatory compliance (ICD-11 is the WHO standard)
- Proper handling of visually similar conditions that require additional clinical information for differentiation
- A unified diagnostic vocabulary for the ICD Category Distribution model and all other clinical models
Key outputs:
- Master ICD-11 mapping matrix linking all source labels to standardized categories
- Documentation of clinical rationale for category consolidation decisions
- Version-controlled ground truth diagnostic classification for the entire dataset
(to be completed)
(include the csv files detailed in the R-TF-028-004 Data Annotation Instructions - ICD-11 Mapping)
Model Development and Validation
This section details the development, training, and validation of all AI models in the Legit.Health Plus device. Each model subsection includes:
- Model-specific data annotation requirements
- Training methodology and architecture
- Performance evaluation results
- Bias analysis and fairness considerations
ICD Category Distribution and Binary Indicators
Model Overview
Reference: R-TF-028-001 AI/ML Description - ICD Category Distribution and Binary Indicators section
The ICD Category Distribution model is a deep learning classifier that outputs a probability distribution across ICD-11 disease categories. The Binary Indicators are derived from this distribution using an expert-curated mapping matrix.
Models included:
- ICD Category Distribution (outputs top-5 diagnoses with probabilities)
- Binary Indicators (6 derived indicators):
- Malignant
- Pre-malignant
- Associated with malignancy
- Pigmented lesion
- Urgent referral (≤48h)
- High-priority referral (≤2 weeks)
Data Requirements and Annotation
Foundational annotation: ICD-11 mapping (completed via R-TF-028-004 Data Annotation Instructions - ICD-11 Mapping)
All images in the training, validation, and test sets were annotated with standardized ICD-11 diagnostic labels following the comprehensive mapping process described in the Data Management section.
Binary Indicator Mapping: A dermatologist-validated mapping matrix was created to link each ICD-11 category to the six binary indicators. This mapping defines which disease categories contribute to each indicator (e.g., melanoma, squamous cell carcinoma, and basal cell carcinoma all contribute to the "Malignant" indicator).
Dataset statistics:
- Total images with ICD-11 labels: [NUMBER] (to be completed)
- Number of ICD-11 categories: [NUMBER] (to be completed)
- Training set: [NUMBER] images
- Validation set: [NUMBER] images
- Test set: [NUMBER] images
Training Methodology
Pre-processing:
- Input images resized to model-required dimensions
- Data augmentation during training: random cropping (guided by bounding boxes where available), rotations, color jittering, histogram equalization
- No augmentation applied to test data
Architecture: [VIT or EfficientNet - to be determined]
- Vision Transformer (ViT) or EfficientNet architecture
- Transfer learning from large-scale pre-trained weights
Training:
- Optimizer: Adam
- Loss function: Cross-entropy
- Learning rate policy: One-cycle policy for super-convergence
- Early stopping based on validation set performance
- Training duration: [NUMBER] epochs
Post-processing:
- Temperature scaling for probability calibration
- Test-time augmentation (TTA) for robust predictions
Performance Results
ICD Category Distribution Performance:
| Metric | Result | Success Criterion | Outcome |
|---|---|---|---|
| Top-1 Accuracy | [TO FILL] | ≥ 55% | [PENDING] |
| Top-3 Accuracy | [TO FILL] | ≥ 70% | [PENDING] |
| Top-5 Accuracy | [TO FILL] | ≥ 80% | [PENDING] |
Binary Indicator Performance:
| Indicator | Result (AUC) | Success Criterion | Outcome |
|---|---|---|---|
| Malignant | [TO FILL] | ≥ 0.80 | [PENDING] |
| Pre-malignant | [TO FILL] | ≥ 0.80 | [PENDING] |
| Associated with malignancy | [TO FILL] | ≥ 0.80 | [PENDING] |
| Pigmented lesion | [TO FILL] | ≥ 0.80 | [PENDING] |
| Urgent referral | [TO FILL] | ≥ 0.80 | [PENDING] |
| High-priority referral | [TO FILL] | ≥ 0.80 | [PENDING] |
Verification and Validation Protocol
Test Design:
- Held-out test set sequestered from training and validation
- Stratified sampling to ensure representation across ICD-11 categories
- Independent evaluation on external datasets (DDI, clinical study data)
Complete Test Protocol:
- Input: RGB images from test set
- Processing: Model inference with TTA
- Output: ICD-11 probability distribution and binary indicator scores
- Ground truth comparison: Expert-labeled ICD-11 categories and binary mappings
- Statistical analysis: Top-k accuracy, AUC-ROC with 95% confidence intervals
Data Analysis Methods:
- Top-k accuracy calculation with bootstrapping for confidence intervals
- ROC curve analysis and AUC calculation for binary indicators
- Confusion matrix analysis for error pattern identification
- Statistical significance testing (DeLong test for AUC comparisons)
Test Conclusions: (to be completed after validation)
Bias Analysis and Fairness Evaluation
Objective: Evaluate model performance across demographic subpopulations to identify and mitigate potential biases that could affect clinical safety and effectiveness.
Subpopulation Analysis Protocol:
1. Fitzpatrick Skin Type Analysis:
- Performance metrics (Top-k accuracy, AUC) disaggregated by Fitzpatrick types I-VI
- Datasets: DDI dataset, internal test set with Fitzpatrick annotations
- Statistical comparison: Chi-square test for performance differences across groups
- Success criterion: No statistically significant performance degradation (p < 0.05) in any Fitzpatrick type below overall acceptance thresholds
2. Age Group Analysis:
- Stratification: Pediatric (under 18 years), Adult (18-65 years), Elderly (over 65 years)
- Metrics: Top-k accuracy and AUC per age group
- Data sources: Clinical study datasets with age metadata
- Success criterion: Performance within ±10% across age groups
3. Anatomical Site Analysis:
- Site categories: Face, trunk, extremities, intertriginous areas, acral sites
- Evaluation: Top-k accuracy per anatomical location
- Success criterion: No anatomical site with performance below acceptance threshold
4. Sex/Gender Analysis:
- Performance comparison between male and female subgroups
- Statistical testing for significant differences
- Success criterion: No gender-based performance disparity >5%
5. Image Quality Impact:
- Analysis of performance degradation with varying image quality (DIQA scores)
- Identification of quality thresholds for reliable predictions
- Mitigation: DIQA-based rejection criteria for low-quality images
6. Rare Condition Representation:
- Analysis of performance on rare vs. common ICD-11 categories
- Class-wise sensitivity and specificity reporting
- Mitigation strategies for underrepresented conditions
Bias Mitigation Strategies:
- Multi-source data collection ensuring demographic diversity
- Fitzpatrick type identification for bias monitoring
- Data augmentation targeting underrepresented subgroups
- Threshold optimization per subpopulation if necessary
- Clinical validation with diverse patient populations
Results Summary: (to be completed after bias analysis)
| Subpopulation | Metric | Result | Comparison to Overall | Assessment |
|---|---|---|---|---|
| Fitzpatrick I-II | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Fitzpatrick III-IV | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Fitzpatrick V-VI | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Age: Pediatric | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Age: Adult | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Age: Elderly | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Sex: Male | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Sex: Female | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Site: Face | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Site: Trunk | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Site: Extremities | Top-5 Acc. | [TO FILL] | [TO FILL] | [PASS/FAIL] |
Bias Analysis Conclusion: (to be completed)
Erythema Intensity Quantification
Model Overview
Reference: R-TF-028-001 AI/ML Description - Erythema Intensity Quantification section
This model quantifies erythema (redness) intensity on an ordinal scale (0-9), outputting a probability distribution that is converted to a continuous severity score via weighted expected value calculation.
Clinical Significance: Erythema is a cardinal sign of inflammation in numerous dermatological conditions including psoriasis, atopic dermatitis, and other inflammatory dermatoses.
Data Requirements and Annotation
Foundational annotation: ICD-11 mapping (completed)
Model-specific annotation: Erythema intensity scoring (R-TF-028-004 Data Annotation Instructions - Visual Signs)
Medical experts (dermatologists) annotated images with erythema intensity scores following standardized clinical scoring protocols (e.g., Clinician's Erythema Assessment scale). Annotations include:
- Ordinal intensity scores (0-9): 0 = none, 9 = maximum erythema
- Multi-annotator consensus for ground truth establishment (minimum 2-3 dermatologists per image)
- Quality control and senior dermatologist review for ambiguous cases
Dataset statistics:
- Images with erythema annotations: [NUMBER] (to be completed)
- Training set: [NUMBER] images
- Validation set: [NUMBER] images
- Test set: [NUMBER] images
- Average inter-annotator agreement (ICC): [VALUE] (to be completed)
- Conditions represented: Psoriasis, atopic dermatitis, rosacea, contact dermatitis, etc.
Training Methodology
Architecture: [CNN-based or ViT-based - to be determined]
- Deep learning model tailored for ordinal regression
- Transfer learning from pre-trained weights (ImageNet or domain-specific)
- Input size: [SIZE] pixels
Training approach:
- Loss function: Ordinal cross-entropy or weighted expected value optimization
- Optimizer: Adam with learning rate [LR]
- Data augmentation: Rotations, color jittering (carefully controlled to preserve erythema characteristics), cropping
- Regularization: Dropout, weight decay
- Training duration: [NUMBER] epochs with early stopping
Post-processing:
- Weighted expected value calculation for continuous score
- Probability calibration if needed
- Output range: 0-9 continuous scale
Performance Results
Performance evaluated using Relative Mean Absolute Error (RMAE) compared to expert consensus.
Success criterion: RMAE ≤ 20% (performance superior to inter-observer variability)
| Metric | Result | Success Criterion | Outcome |
|---|---|---|---|
| RMAE (Overall) | [TO FILL] | ≤ 20% | [PENDING] |
| Pearson Correlation | [TO FILL] | ≥ 0.85 | [PENDING] |
| Expert Inter-observer ICC | [TO FILL] | Reference | N/A |
| Model vs. Expert ICC | [TO FILL] | ≥ Expert ICC | [PENDING] |
Verification and Validation Protocol
Test Design:
- Independent test set with multi-annotator ground truth (minimum 3 dermatologists per image)
- Comparison against expert consensus (mean of expert scores)
- Evaluation across diverse conditions (psoriasis, eczema, rosacea), Fitzpatrick skin types, and anatomical sites
Complete Test Protocol:
- Input: RGB images from test set with expert erythema intensity annotations
- Processing: Model inference with probability distribution output
- Output: Continuous erythema severity score (0-9) via weighted expected value
- Ground truth: Consensus intensity score from multiple expert dermatologists
- Statistical analysis: RMAE, ICC, Pearson/Spearman correlation, Bland-Altman analysis
Data Analysis Methods:
- RMAE calculation: Relative Mean Absolute Error comparing model predictions to expert consensus
- Inter-observer variability measurement (ICC among experts as benchmark)
- Correlation analysis: Pearson and Spearman correlation coefficients
- Bland-Altman plots for agreement assessment
- Bootstrap resampling (1000 iterations) for 95% confidence intervals
- Subgroup analysis for bias detection
Test Conclusions: (to be completed after validation)
Bias Analysis and Fairness Evaluation
Objective: Ensure erythema quantification performs consistently across demographic subpopulations, with special attention to Fitzpatrick skin types where erythema visualization varies.
Subpopulation Analysis Protocol:
1. Fitzpatrick Skin Type Analysis (Critical for erythema):
- RMAE calculation per Fitzpatrick type (I-II, III-IV, V-VI)
- Recognition that erythema contrast decreases with increasing melanin content
- Comparison of model performance vs. expert inter-observer variability per skin type
- Success criterion: RMAE ≤ 20% maintained across all skin types
2. Disease Condition Analysis:
- Performance per condition: Psoriasis, atopic dermatitis, rosacea, contact dermatitis, cellulitis
- Disease-specific annotation challenges and inter-observer variability
- Success criterion: Model performance better than or equal to expert variability for each condition
3. Anatomical Site Analysis:
- Site-specific performance: Face, trunk, extremities, intertriginous areas
- Recognition of site-specific visualization challenges (shadows, curvature)
- Success criterion: No site with RMAE > 25%
4. Severity Range Analysis:
- Performance stratified by severity: Mild (0-3), Moderate (4-6), Severe (7-9)
- Detection of ceiling or floor effects
- Success criterion: Consistent RMAE across severity levels
5. Image Quality Impact:
- RMAE correlation with DIQA scores
- Performance degradation with poor lighting/focus
- Mitigation: DIQA-based quality filtering
6. Age Group Analysis:
- Performance in pediatric, adult, elderly populations
- Age-related skin changes (thinner skin, vascular changes)
- Success criterion: No age group with significantly degraded performance
Bias Mitigation Strategies:
- Training data balanced across Fitzpatrick types (minimum 20% representation of types V-VI)
- Fitzpatrick-specific data augmentation
- Potential Fitzpatrick-conditional model calibration
- Collaborative training with other chromatic intensity models (desquamation, induration)
Results Summary: (to be completed after bias analysis)
| Subpopulation | RMAE | Expert ICC | Model vs Expert | Assessment |
|---|---|---|---|---|
| Fitzpatrick I-II | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Fitzpatrick III-IV | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Fitzpatrick V-VI | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Psoriasis | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Atopic Dermatitis | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Rosacea | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Mild Severity (0-3) | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Moderate Severity (4-6) | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
| Severe Severity (7-9) | [TO FILL] | [TO FILL] | [TO FILL] | [PASS/FAIL] |
Bias Analysis Conclusion: (to be completed)
Desquamation Intensity Quantification
Model Overview
Reference: R-TF-028-001 AI/ML Description - Desquamation Intensity Quantification section
This model quantifies desquamation (scaling/peeling) intensity on an ordinal scale (0-9), critical for assessment of psoriasis, seborrheic dermatitis, and other scaling conditions.
Clinical Significance: Desquamation is one of the three cardinal signs in PASI scoring for psoriasis and a key indicator in many inflammatory dermatoses.
Data Requirements and Annotation
Foundational annotation: ICD-11 mapping (completed)
Model-specific annotation: Desquamation intensity scoring (R-TF-028-004 Data Annotation Instructions - Visual Signs)
Dataset statistics: (to be completed)
Training Methodology
Architecture: (to be determined)
Performance Results
| Metric | Result | Success Criterion | Outcome |
|---|---|---|---|
| RMAE (Overall) | [TO FILL] | ≤ 20% | [PENDING] |
| Pearson Correlation | [TO FILL] | ≥ 0.85 | [PENDING] |
Verification and Validation Protocol
(Follow same comprehensive protocol as Erythema model)
Bias Analysis and Fairness Evaluation
Subpopulation Analysis: Fitzpatrick types, disease conditions (psoriasis, eczema, seborrheic dermatitis), anatomical sites, severity ranges.
Results Summary: (to be completed)
Induration Intensity Quantification
Model Overview
Reference: R-TF-028-001 AI/ML Description - Induration Intensity Quantification section
This model quantifies induration (plaque thickness/elevation) on an ordinal scale (0-9), essential for psoriasis PASI scoring and assessment of infiltrative conditions.
Clinical Significance: Induration reflects tissue infiltration and is a key component of psoriasis severity assessment.
Data Requirements and Annotation
Foundational annotation: ICD-11 mapping (completed)
Model-specific annotation: Induration intensity scoring (R-TF-028-004 Data Annotation Instructions - Visual Signs)
Dataset statistics: (to be completed)