Discussion about this post

User's avatar
john boddie's avatar

Another question about data that is seldom asked is, "How much quality is 'good enough?'" There are many existing data stores that can never be fully accurate. An example is the number of active service lines managed by a phone company. Several years ago, a project related to assigning equipment for phone service identified six sources, each of which was treated as accurate by individual business processes - billing, service, provisioning (assigning equipment), engineering, finance, and sales/marketing. Each of these had a different number of active lines and the range exceeded 24,000. Could these differences be reduced? Maybe, but in the time required to gather detailed information in the company's central offices (where the network switches live), the number will have changed as new customers are added and existing customers leave. In addition, it's not clear that any connected circuit is really active (it may be connected but it's not used or billed). How much is the gap costing the company? Clearly, it's in the company's interest to have some number of inactive service lines in order to provide service quickly to customers. In situations like this, the cost to achieve a defined level of data quality may be greater than the benefit gained through improved management of idle equipment. So how do we figure out if the quality of data is "good enough?"

Expand full comment
1 more comment...

No posts