mindmeld.models.evaluation module¶
This module contains base classes for models defined in the models subpackage.
-
class
mindmeld.models.evaluation.EntityModelEvaluation(config, results)[source]¶ Bases:
mindmeld.models.evaluation.SequenceModelEvaluationGenerates some statistics specific to entity recognition
-
class
mindmeld.models.evaluation.EvaluatedExample[source]¶ Bases:
mindmeld.models.evaluation.EvaluatedExampleRepresents the evaluation of a single example
-
example¶ The example being evaluated
-
expected¶ The expected label for the example
-
predicted¶ The predicted label for the example
-
proba¶ dict -- Maps labels to their predicted probabilities
-
label_type¶ str -- One of CLASS_LABEL_TYPE or ENTITIES_LABEL_TYPE
-
is_correct¶
-
-
class
mindmeld.models.evaluation.ModelEvaluation(config, results)[source]¶ Bases:
mindmeld.models.evaluation.ModelEvaluationRepresents the evaluation of a model at a specific configuration using a collection of examples and labels.
-
config¶ ModelConfig -- The model config used during evaluation.
-
results¶ list of EvaluatedExample -- A list of the evaluated examples.
-
correct_results()[source]¶ Returns: Collection of the examples which were correct Return type: iterable
-
get_accuracy()[source]¶ The accuracy represents the share of examples whose predicted labels exactly matched their expected labels.
Returns: The accuracy of the model. Return type: float
-
get_stats()[source]¶ Returns a structured stats object for evaluation.
Returns: Structured dict containing evaluation statistics. Contains precision, recall, f scores, support, etc. Return type: dict
-
incorrect_results()[source]¶ Returns: Collection of the examples which were incorrect Return type: iterable
-
print_stats()[source]¶ Prints a useful stats table for evaluation.
Returns: Structured dict containing evaluation statistics. Contains precision, recall, f scores, support, etc. Return type: dict
-
raw_results()[source]¶ Exposes raw vectors of expected and predicted for data scientists to use for any additional evaluation metrics or to generate graphs of their choice.
Returns: tuple containing: - NamedTuple: RawResults named tuple containing
- expected: vector of predicted classes (numeric value)
- predicted: vector of gold classes (numeric value)
- text_labels: a list of all the text label values, the index of the text label in
- this array is the numeric label
Return type: (tuple)
-
-
class
mindmeld.models.evaluation.RawResults(predicted, expected, text_labels, predicted_flat=None, expected_flat=None)[source]¶ Bases:
objectRepresents the raw results of a set of evaluated examples. Useful for generating stats and graphs.
-
predicted¶ list -- A list of predictions. For sequences this is a list of lists, and for standard classifieris this is a 1d array. All classes are in their numeric representations for ease of use with evaluation libraries and graphing.
-
expected¶ list -- Same as predicted but contains the true or gold values.
-
text_labels¶ list -- A list of all the text label values, the index of the text label in this array is the numeric label
-
predicted_flat¶ list -- (Optional): For sequence models this is a flattened list of all predicted tags (1d array)
-
expected_flat¶ list -- (Optional): For sequence models this is a flattened list of all gold tags
-