

Transforming unstructured English tasks into scientifically sound, quantified value metrics using Python & NLP.
Raw English text and tasks describing work performed (e.g., “Expediting delivery speed…”).
Extract value phrases, calculate confidence scores, and map to semantic concepts.
Standardized tags (e.g., “Agility”) backed by numerical confidence metrics.
The core engine follows a strict five-step scientific process to ensure data integrity and defensibility.
Identify phrases representing benefits (e.g., “reduce latency”).
Assign a numeric probability score to the extraction accuracy.
Map phrases to the 10 defined concepts via vector similarity.
Filter and select the top 1-2 tags with highest relevance.
Deliver structured, defensible JSON data for analysis.
The system maps extracted text to a fixed ontology of 10 business value drivers. This standardization allows for aggregation across thousands of unique tasks.
Fig 1. The Ontology of Value Tags
How does the AI decide? It plots the Confidence Score of the extraction against the Semantic Similarity to the tag. Only results in the “High Confidence / High Similarity” quadrant (top right) are automatically accepted.
By processing a batch of English tasks through the system, we generate a frequency distribution of organizational focus.
In this dataset, Efficiency and Agility are the dominant value drivers, suggesting the team is currently focused on optimization and speed rather than new features (Innovation) or defensive measures (Security).