Skip to content

Conversation

harshach
Copy link
Collaborator

@harshach harshach commented Oct 13, 2025

Describe your changes:

Fixes #23853

I worked on ... because ...

Type of change:

  • Bug fix
  • Improvement
  • New feature
  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • Documentation

Checklist:

  • I have read the CONTRIBUTING document.
  • My PR title is Fixes <issue-number>: <short explanation>
  • I have commented on my code, particularly in hard-to-understand areas.
  • For JSON Schema changes: I updated the migration scripts or explained why it is not needed.

Copy link
Contributor

TypeScript types have been updated based on the JSON schema changes in the PR

@github-actions github-actions bot requested a review from a team as a code owner October 13, 2025 04:11
"resourceType": {"type": "string"}
}
},
"hyperparameters": {
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should we keep this as an open KV pair for people to define? similar to how we have it in mlmodel https://github.com/open-metadata/OpenMetadata/blob/main/openmetadata-spec/src/main/resources/json/schema/entity/data/mlmodel.json#L190

might be hard to keep tabs on every param

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  • These 6 hyperparameters are the most common ones used in LLM training (universal across PyTorch, TensorFlow, JAX)
    • Having them as structured fields provides type safety (number vs integer) and validation
    • Makes UI forms easier - we can render proper input fields with appropriate controls

"weightDecay": {"type": "number"}
}
},
"trainingMetrics": {
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

same here, I guess depending on the model the metrics teams can track could be different

},
"required": ["name", "ruleType", "action"]
},
"biasThreshold": {
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Would it make sense to define Bias Metrics with our Metrics definition, link those metrics as entityRefs to the llmModel, and then have the threshold here as an EntityReference + Threshold number? This way we're sure we are always going to be following the same definitions, and users might be able to go and add further pieces without them being coupled to the schema

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Bias metrics are generic no? teams do not exactly what they want to track here. We need to be more prescriptive than having a generic metric

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

  • User Experience: These are "snapshot" metrics that represent the current state. Forcing users to create separate Metric entities for every AI application's bias score adds
    significant overhead
    • Use Case: biasMetrics, performanceMetrics, etc. are NOT time-series metrics - they're current evaluation snapshots that belong to the entity
    • Governance: Having bias metrics inline makes governance checks easier - you can immediately see if an application has bias without needing to follow references
    • OpenMetadata Metric entities: These are designed for database/table metrics (like row count, column distributions), not AI evaluation metrics
    • Different semantics: EntityReference to Metrics implies these are reusable metric definitions. But "overall bias score for Application X" is specific to that application, not
      reusable

}
}
},
"biasMetrics": {
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

same, looks like we're redefining bias with possibly different keys/props

Copy link
Contributor

Jest test Coverage

UI tests summary

Lines Statements Branches Functions
Coverage: 63%
63.46% (46286/72940) 40.71% (21864/53712) 43.74% (6958/15907)

Copy link

Copy link

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

backend safe to test Add this label to run secure Github workflows on PRs

Projects

None yet

Development

Successfully merging this pull request may close these issues.

AI Governance and Compliance Framework for AI Applications

2 participants