Custom Metric Definition Language

In Aporia, custom metrics are defined using syntax that is similar to python's.

There are three building blocks which can be used in order to create a custom metric expression:

  • Constants - a numeric value (e.g. 2, 0.5, ..)

  • Functions - out of the builtin function collection you can find below (e.g. sum, count, ...). All those functions return a numeric value.

  • Binary operation - +, -, *, /, **. Operands can be both constants or function calls.

Builtin Functions

Before we dive into each of the supported function, there are two general concepts you should be familiar with regarding all functions - field expressions and data segment filters.

Field Expressions

A field expression can be described in the following format:

<field_category>.<field_name>[<segment filter>]

Field category is one of the following: features / raw_inputs / predictions / actuals. Note that you can only use categories which you defined in you schema while creating your model version. In addition, don't forget that predictions and actuals categories have the same field names.

The segment filter is optional, for further information about the filters read the section below.

Data Segment Filters

Data segment filters are boolean expressions, designed to reduce to a specific data segment the field on which we perform the function.

Each boolean condition in a segment filter is a comparison between a field and a constant value. For example:

[features.Driving_License == True] // will filter out records in which Driving_License != True
[raw_inputs.Age <= 35]             // will only include records in which Age <= 35

Conditions can be combined using and / or and all fields can be checked for missing values using is None / is not None.

The following describe the supported combinations:

Type / Operation
==
!=
<
>
>=
<=

Boolean

True/False

True/False

βœ–οΈ

βœ–οΈ

βœ–οΈ

βœ–οΈ

Categorical

numeric constants

numeric constants

βœ–οΈ

βœ–οΈ

βœ–οΈ

βœ–οΈ

String

numeric constants

numeric constants

βœ–οΈ

βœ–οΈ

βœ–οΈ

βœ–οΈ

Numeric

numeric constants

numeric constants

numeric constants

numeric constants

numeric constants

numeric constants

The table cells indicates the type we can compare to.

Examples

Supported functions

chevron-rightaccuracyhashtag

Parameters

  • prediction: prediction field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightactuals_counthashtag

Parameters

No parameters needed, cannot apply filters on this metric.

chevron-rightactuals_ratiohashtag

Parameters

No parameters needed, cannot apply filters on this metric.

chevron-rightauc_rochashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightcounthashtag

Parameters

No parameters needed, cannot apply filters on this metric.

chevron-rightf1_scorehashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • average: the average strategy (micro / macro / weighted)

  • top_k: consider only top-k items.

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightfn_counthashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightfp_counthashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightfp_ratehashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightloglosshashtag

Parameters

  • prediction: prediction field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightmaehashtag

Parameters

  • prediction: prediction field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightmapehashtag

Parameters

  • prediction: prediction field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightmaxhashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightmeanhashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightmedianhashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightminhashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightmiss_ratehashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightmissing_counthashtag

Parameters

  • field: the field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightmissing_ratiohashtag

Parameters

  • field: the field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightmsehashtag

Parameters

  • prediction: prediction field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightndcghashtag

Parameters

  • prediction: prediction field

  • label: label field

  • rank: the rank position

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightnot_missing_counthashtag

Parameters

  • field: the field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightprecision_scorehashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • average: the average strategy (micro / macro / weighted)

  • top_k: consider only top-k items.

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightrecall_scorehashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • average: the average strategy (micro / macro / weighted)

  • top_k: consider only top-k items.

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightrmsehashtag

Parameters

  • prediction: prediction field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightspecificityhashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightstdhashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict

chevron-rightsumhashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict

chevron-righttn_counthashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-righttp_counthashtag

Parameters

  • prediction: prediction probability field

  • label: label field

  • threshold: numeric. Probability threshold according to which we decide the if a class is positive

  • filter: the filter we want to apply on the records before calculating the metric

chevron-rightunique_counthashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightvalue_counthashtag

Parameters

  • field: the field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • value: the value we want to count

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightvalue_percentagehashtag

Parameters

  • field: the field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • value: the value we want to count

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightvariancehashtag

Parameters

  • field: numeric or dict. The field for which the metric will be computed. Can be of any category (feature / raw_input / prediction / actual)

  • filter: the filter we want to apply on the records before calculating the metric

  • keys: keys to filter in when field type is dict.

chevron-rightwapehashtag

Parameters

  • prediction: prediction field

  • label: label field

  • filter: the filter we want to apply on the records before calculating the metric

Last updated