POST https://api.zatomic.ai/v1/prompts/scoring
Zatomic API
- Introduction
- Versioning
- Authentication
- Workspaces
- Status codes and errors
- Token usage
- Expanding objects
- OpenAPI spec
- Prompts
- The Prompt object
- Creating a prompt
- Updating a prompt
- Deleting a prompt
- Retrieving a prompt
- Retrieving all prompts
- Generating a prompt
- Versions
- The Version object
- Creating a version
- Updating a version
- Deleting a version
- Retrieving a version
- Retrieving all versions
- Retrieving a version score
- Calculating a version score
- Retrieving a version risk
- Analyzing a version risk
- Retrieving a version balance
- Analyzing a version balance
- Retrieving a version heatmap
- Generating a version heatmap
- Improving a version
- Scoring Criteria
- The Scoring Criteria object
- The Scoring Criterion object
- Creating scoring criteria
- Updating a scoring criteria
- Deleting a scoring criteria
- Retrieving a scoring criteria
- Retrieving all scoring criteria
- Generating scoring criteria
- Creating a scoring criterion
- Updating a scoring criterion
- Deleting a scoring criterion
- Retrieving a scoring criterion
- Scoring Criteria Results
- The Scoring Criteria Results object
- Scoring
- The Scoring object
- Calculating a prompt score
- Risk
- The Risk object
- Analyzing prompt risk
- Balance
- The Balance object
- Analyzing prompt balance
- Heatmaps
- The Heatmap object
- Generating a prompt heatmap
Scoring
Prompt scoring uses various criteria to analyze prompts and assign them a score and rating, with higher scores leading to better prompt performance.
Scoring can be performed and retrieved on individual prompt versions using their specific scoring endpoints. You can also score prompts without a version stored in the system by using the non-version specific endpoint.
Prompts are scored in the following ranges:
Scoring Range | Prompt Rating |
---|---|
0 - 49% | Poor |
50 - 74% | Fair |
75 - 89% | Good |
90 - 100% | Excellent |
Endpoints