Amazon Bedrock Performance Analysis Model Evaluation
This new feature allows you to choose the foundation model that produces the best results for your specific use case, which makes it easier for you to integrate generative AI into your application
If you choose Text classification, for instance, you can assess robustness and/or accuracy in relation to either an internal or your own dataset
For all human assessment settings and for some combinations of job types and metrics for automatic evaluation, the reference response is optional
These datasets can be selected as needed because they are made to evaluate particular kinds of activities and metrics
The status of every model evaluation job that you have is accessible through the console and the newly added GetEvaluationJob API function
Obtain and Examine the Evaluation Report Obtain the report and evaluate the model’s performance using the previously chosen metrics
Using the console or the recently introduced model evaluation API, you may now halt a running job