Cloud Provider
Service Name
Inefficiency Type
Clear filters
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Showing
1234
out of
1234
inefficiencis
Filter
:
Filter
x
Using High-Cost Models for Low-Complexity Tasks
AI
Cloud Provider
Azure
Service Name
Azure Cognitive Services
Inefficiency Type
Overpowered Model Selection

Some workloads — such as text classification, keyword extraction, intent detection, routing, or lightweight summarization — do not require the capabilities of the most advanced model families. When high-cost models are used for these simple tasks, organizations pay elevated token rates for work that could be handled effectively by more efficient, lower-cost models. This mismatch typically arises from defaulting to a single model for all tasks or not periodically reviewing model usage patterns across applications.

Provisioned Throughput OpenAI Deployment in Non-Production Environments
AI
Cloud Provider
Azure
Service Name
Azure Cognitive Services
Inefficiency Type
Overprovisioned Deployment Model

PTU deployments guarantee dedicated throughput and low latency, but they also require paying for reserved capacity at all times. In non-production environments—such as dev, test, QA, or experimentation—usage patterns are typically sporadic and unpredictable. Deploying PTUs in these environments leads to consistent baseline spend without corresponding value. On-demand deployments scale usage cost with actual consumption, making them more cost-efficient for variable workloads.

There are no inefficiency matches the current filters.