abacusai.api_class

Submodules

Package Contents

Classes

ApiClass

Helper class that provides a standard way to create an ABC using

_ApiClassFactory

Helper class that provides a standard way to create an ABC using

BatchPredictionArgs

Helper class that provides a standard way to create an ABC using

AnomalyDetectionBatchPredictionArgs

Batch Prediction Config for the ANOMALY_DETECTION problem type

AnomalyOutliersBatchPredictionArgs

Batch Prediction Config for the ANOMALY_OUTLIERS problem type

ForecastingBatchPredictionArgs

Batch Prediction Config for the FORECASTING problem type

NamedEntityExtractionBatchPredictionArgs

Batch Prediction Config for the NAMED_ENTITY_EXTRACTION problem type

PersonalizationBatchPredictionArgs

Batch Prediction Config for the PERSONALIZATION problem type

PredictiveModelingBatchPredictionArgs

Batch Prediction Config for the PREDICTIVE_MODELING problem type

PretrainedModelsBatchPredictionArgs

Batch Prediction Config for the PRETRAINED_MODELS problem type

SentenceBoundaryDetectionBatchPredictionArgs

Batch Prediction Config for the SENTENCE_BOUNDARY_DETECTION problem type

ThemeAnalysisBatchPredictionArgs

Batch Prediction Config for the THEME_ANALYSIS problem type

ChatLLMBatchPredictionArgs

Batch Prediction Config for the ChatLLM problem type

_BatchPredictionArgsFactory

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

BlobInput

Binary large object input data.

ApiClass

Helper class that provides a standard way to create an ABC using

ParsingConfig

Helper class that provides a standard way to create an ABC using

DocumentProcessingConfig

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

VectorStoreTextEncoder

Generic enumeration.

VectorStoreConfig

Configs for vector store indexing.

DocumentRetrieverConfig

Configs for document retriever.

ApiEnum

Generic enumeration.

ProblemType

Generic enumeration.

RegressionObjective

Generic enumeration.

RegressionTreeHPOMode

Generic enumeration.

RegressionAugmentationStrategy

Generic enumeration.

RegressionTargetTransform

Generic enumeration.

RegressionTypeOfSplit

Generic enumeration.

RegressionTimeSplitMethod

Generic enumeration.

RegressionLossFunction

Generic enumeration.

ExplainerType

Generic enumeration.

SamplingMethodType

Generic enumeration.

MergeMode

Generic enumeration.

FillLogic

Generic enumeration.

BatchSize

Generic enumeration.

HolidayCalendars

Generic enumeration.

ExperimentationMode

Generic enumeration.

PersonalizationTrainingMode

Generic enumeration.

PersonalizationObjective

Generic enumeration.

ForecastingObjective

Generic enumeration.

ForecastingFrequency

Generic enumeration.

ForecastingDataSplitType

Generic enumeration.

ForecastingLossFunction

Generic enumeration.

ForecastingLocalScaling

Generic enumeration.

ForecastingFillMethod

Generic enumeration.

ForecastingQuanitlesExtensionMethod

Generic enumeration.

NERObjective

Generic enumeration.

NERModelType

Generic enumeration.

NLPDocumentFormat

Generic enumeration.

SentimentType

Generic enumeration.

ClusteringImputationMethod

Generic enumeration.

ConnectorType

Generic enumeration.

PythonFunctionArgumentType

Generic enumeration.

PythonFunctionOutputArgumentType

Generic enumeration.

VectorStoreTextEncoder

Generic enumeration.

LLMName

Generic enumeration.

MonitorAlertType

Generic enumeration.

FeatureDriftType

Generic enumeration.

DataIntegrityViolationType

Generic enumeration.

BiasType

Generic enumeration.

AlertActionType

Generic enumeration.

ApiClass

Helper class that provides a standard way to create an ABC using

_ApiClassFactory

Helper class that provides a standard way to create an ABC using

SamplingConfig

An abstract class for the sampling config of a feature group

NSamplingConfig

The number of distinct values of the key columns to include in the sample, or number of rows if key columns not specified.

PercentSamplingConfig

The fraction of distinct values of the feature group to include in the sample.

_SamplingConfigFactory

Helper class that provides a standard way to create an ABC using

MergeConfig

An abstract class for the merge config of a feature group

LastNMergeConfig

Merge LAST N chunks/versions of an incremental dataset.

TimeWindowMergeConfig

Merge rows within a given timewindow of the most recent timestamp

_MergeConfigFactory

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

_ApiClassFactory

Helper class that provides a standard way to create an ABC using

TrainingConfig

Helper class that provides a standard way to create an ABC using

PersonalizationTrainingConfig

Training config for the PERSONALIZATION problem type

RegressionTrainingConfig

Training config for the PREDICTIVE_MODELING problem type

ForecastingTrainingConfig

Training config for the FORECASTING problem type

NamedEntityExtractionTrainingConfig

Training config for the NAMED_ENTITY_EXTRACTION problem type

NaturalLanguageSearchTrainingConfig

Training config for the NATURAL_LANGUAGE_SEARCH problem type

ChatLLMTrainingConfig

Training config for the CHAT_LLM problem type

SentenceBoundaryDetectionTrainingConfig

Training config for the SENTENCE_BOUNDARY_DETECTION problem type

SentimentDetectionTrainingConfig

Training config for the SENTIMENT_DETECTION problem type

DocumentClassificationTrainingConfig

Training config for the DOCUMENT_CLASSIFICATION problem type

DocumentSummarizationTrainingConfig

Training config for the DOCUMENT_SUMMARIZATION problem type

DocumentVisualizationTrainingConfig

Training config for the DOCUMENT_VISUALIZATION problem type

ClusteringTrainingConfig

Training config for the CLUSTERING problem type

ClusteringTimeseriesTrainingConfig

Training config for the CLUSTERING_TIMESERIES problem type

EventAnomalyTrainingConfig

Training config for the EVENT_ANOMALY problem type

CumulativeForecastingTrainingConfig

Training config for the CUMULATIVE_FORECASTING problem type

AnomalyDetectionTrainingConfig

Training config for the ANOMALY_DETECTION problem type

ThemeAnalysisTrainingConfig

Training config for the THEME ANALYSIS problem type

AIAgentTrainingConfig

Training config for the AI_AGENT problem type

CustomTrainedModelTrainingConfig

Training config for the CUSTOM_TRAINED_MODEL problem type

CustomAlgorithmTrainingConfig

Training config for the CUSTOM_ALGORITHM problem type

OptimizationTrainingConfig

Training config for the OPTIMIZATION problem type

_TrainingConfigFactory

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

ForecastingMonitorConfig

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

_ApiClassFactory

Helper class that provides a standard way to create an ABC using

AlertConditionConfig

Helper class that provides a standard way to create an ABC using

AccuracyBelowThresholdConditionConfig

Accuracy Below Threshold Condition Config for Monitor Alerts

FeatureDriftConditionConfig

Feature Drift Condition Config for Monitor Alerts

DataIntegrityViolationConditionConfig

Data Integrity Violation Condition Config for Monitor Alerts

BiasViolationConditionConfig

Bias Violation Condition Config for Monitor Alerts

_AlertConditionConfigFactory

Helper class that provides a standard way to create an ABC using

AlertActionConfig

Helper class that provides a standard way to create an ABC using

EmailActionConfig

Email Action Config for Monitor Alerts

_AlertActionConfigFactory

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

FeatureMappingConfig

Helper class that provides a standard way to create an ABC using

ProjectFeatureGroupTypeMappingsConfig

Helper class that provides a standard way to create an ABC using

ApiClass

Helper class that provides a standard way to create an ABC using

PythonFunctionArgument

A config class for python function arguments

OutputVariableMapping

A config class for python function arguments

ApiClass

Helper class that provides a standard way to create an ABC using

_ApiClassFactory

Helper class that provides a standard way to create an ABC using

FeatureGroupExportConfig

Helper class that provides a standard way to create an ABC using

FileConnectorExportConfig

Helper class that provides a standard way to create an ABC using

DatabaseConnectorExportConfig

Helper class that provides a standard way to create an ABC using

_FeatureGroupExportConfigFactory

Helper class that provides a standard way to create an ABC using

class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class._ApiClassFactory

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key
config_class_map
classmethod from_dict(config)
Parameters:

config (dict) –

Return type:

ApiClass

class abacusai.api_class.BatchPredictionArgs

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

_support_kwargs: bool
kwargs: dict
problem_type: abacusai.api_class.enums.ProblemType
classmethod _get_builder()
class abacusai.api_class.AnomalyDetectionBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the ANOMALY_DETECTION problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param prediction_time_endpoint: The end point for predictions. :type prediction_time_endpoint: str :param prediction_time_range: Over what period of time should we make predictions (in seconds). :type prediction_time_range: int :param minimum_anomaly_score: Exclude results with an anomaly score (1 in x event) below this threshold. Range: [1, 1_000_000_000_000]. :type minimum_anomaly_score: int :param summary_mode: Only show top anomalies per ID. :type summary_mode: bool :param attach_raw_data: Return raw data along with anomalies. :type attach_raw_data: bool :param small_batch: Size of batch data guaranteed to be small. :type small_batch: bool

for_eval: bool
prediction_time_endpoint: str
prediction_time_range: int
minimum_anomaly_score: int
summary_mode: bool
attach_raw_data: bool
small_batch: bool
__post_init__()
class abacusai.api_class.AnomalyOutliersBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the ANOMALY_OUTLIERS problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param threshold: The threshold for detecting an anomaly. Range: [0.8, 0.99] :type threshold: float

for_eval: bool
threshold: float
__post_init__()
class abacusai.api_class.ForecastingBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the FORECASTING problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation :type for_eval: bool :param predictions_start_date: The start date for predictions. :type predictions_start_date: str :param use_prediction_offset: If True, use prediction offset. :type use_prediction_offset: bool :param start_date_offset: Sets prediction start date as this offset relative to the prediction start date. :type start_date_offset: int :param forecasting_horizon: The number of timestamps to predict in the future. Range: [1, 1000]. :type forecasting_horizon: int :param item_attributes_to_include_in_the_result: List of columns to include in the prediction output. :type item_attributes_to_include_in_the_result: list :param explain_predictions: If True, explain predictions for the forecast. :type explain_predictions: bool

for_eval: bool
predictions_start_date: str
use_prediction_offset: bool
start_date_offset: int
forecasting_horizon: int
item_attributes_to_include_in_the_result: list
explain_predictions: bool
__post_init__()
class abacusai.api_class.NamedEntityExtractionBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the NAMED_ENTITY_EXTRACTION problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param verbose_predictions: Return prediction inputs, predicted annotations and token label probabilities. :type verbose_predictions: bool

for_eval: bool
verbose_predictions: bool
__post_init__()
class abacusai.api_class.PersonalizationBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the PERSONALIZATION problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param number_of_items: Number of items to recommend. :type number_of_items: int :param result_columns: List of columns to include in the prediction output. :type result_columns: list :param score_field: If specified, relative item scores will be returned using a field with this name :type score_field: str

for_eval: bool
number_of_items: int
item_attributes_to_include_in_the_result: list
score_field: str
__post_init__()
class abacusai.api_class.PredictiveModelingBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the PREDICTIVE_MODELING problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param explainer_type: The type of explainer to use to generate explanations on the batch prediction. :type explainer_type: enums.ExplainerType :param number_of_samples_to_use_for_explainer: Number Of Samples To Use For Kernel Explainer. :type number_of_samples_to_use_for_explainer: int :param include_multi_class_explanations: If True, Includes explanations for all classes in multi-class classification. :type include_multi_class_explanations: bool :param features_considered_constant_for_explanations: Comma separate list of fields to treat as constant in SHAP explanations. :type features_considered_constant_for_explanations: str :param importance_of_records_in_nested_columns: Returns importance of each index in the specified nested column instead of SHAP column explanations. :type importance_of_records_in_nested_columns: str :param explanation_filter_lower_bound: If set explanations will be limited to predictions above this value, Range: [0, 1]. :type explanation_filter_lower_bound: float :param explanation_filter_upper_bound: If set explanations will be limited to predictions below this value, Range: [0, 1]. :type explanation_filter_upper_bound: float :param bound_label: For classification problems specifies the label to which the explanation bounds are applied. :type bound_label: str :param output_columns: A list of column names to include in the prediction result. :type output_columns: list

for_eval: bool
explainer_type: abacusai.api_class.enums.ExplainerType
number_of_samples_to_use_for_explainer: int
include_multi_class_explanations: bool
features_considered_constant_for_explanations: str
importance_of_records_in_nested_columns: str
explanation_filter_lower_bound: float
explanation_filter_upper_bound: float
explanation_filter_label: str
output_columns: list
__post_init__()
class abacusai.api_class.PretrainedModelsBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the PRETRAINED_MODELS problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param files_output_location_prefix: The output location prefix for the files. :type files_output_location_prefix: str :param channel_id_to_label_map: JSON string for the map from channel ids to their labels. :type channel_id_to_label_map: str

for_eval: bool
files_output_location_prefix: str
channel_id_to_label_map: str
__post_init__()
class abacusai.api_class.SentenceBoundaryDetectionBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the SENTENCE_BOUNDARY_DETECTION problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation :type for_eval: bool :param explode_output: Explode data so there is one sentence per row. :type explode_output: bool

for_eval: bool
explode_output: bool
__post_init__()
class abacusai.api_class.ThemeAnalysisBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the THEME_ANALYSIS problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param analysis_frequency: The length of each analysis interval. :type analysis_frequency: str :param start_date: The end point for predictions. :type start_date: str :param analysis_days: How many days to analyze. :type analysis_days: int

for_eval: bool
analysis_frequency: str
start_date: str
analysis_days: int
__post_init__()
class abacusai.api_class.ChatLLMBatchPredictionArgs

Bases: BatchPredictionArgs

Batch Prediction Config for the ChatLLM problem type :param for_eval: If True, the test fold which was created during training and used for metrics calculation will be used as input data. These predictions are hence, used for model evaluation. :type for_eval: bool :param product: Generate a response for every question and chunk combination :type product: bool

for_eval: bool
product: bool
__post_init__()
class abacusai.api_class._BatchPredictionArgsFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key = 'problemType'
config_class_map
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.BlobInput

Bases: abacusai.api_class.abstract.ApiClass

Binary large object input data.

Parameters:
  • filename (str) – The original filename of the blob.

  • contents (bytes) – The binary contents of the blob.

  • mime_type (str) – The mime type of the blob.

  • size (int) – The size of the blob in bytes.

filename: str
contents: bytes
mime_type: str
size: int
classmethod from_local_file(file_path)
Parameters:

file_path (str) –

Return type:

BlobInput

classmethod from_contents(contents, filename=None, mime_type=None)
Parameters:
  • contents (bytes) –

  • filename (str) –

  • mime_type (str) –

Return type:

BlobInput

class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.ParsingConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

escape: str
csv_delimiter: str
file_path_with_schema: str
class abacusai.api_class.DocumentProcessingConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

extract_bounding_boxes: bool = False
convert_to_markdown: bool = False
use_doctr: bool = False
remove_watermarks: bool
use_full_ocr: bool
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.VectorStoreTextEncoder

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

E5 = 'E5'
OPENAI = 'OPENAI'
SENTENCE_BERT = 'SENTENCE_BERT'
E5_SMALL = 'E5_SMALL'
class abacusai.api_class.VectorStoreConfig

Bases: abacusai.api_class.abstract.ApiClass

Configs for vector store indexing.

Parameters:
  • chunk_size (int) – The size of text chunks in the vector store.

  • chunk_overlap_fraction (float) – The fraction of overlap between chunks.

  • text_encoder (VectorStoreTextEncoder) – Encoder used to index texts from the documents.

chunk_size: int
chunk_overlap_fraction: float
text_encoder: abacusai.api_class.enums.VectorStoreTextEncoder
class abacusai.api_class.DocumentRetrieverConfig

Bases: VectorStoreConfig

Configs for document retriever.

class abacusai.api_class.ApiEnum

Bases: enum.Enum

Generic enumeration.

Derive from this class to define new enumerations.

__eq__(other)

Return self==value.

__hash__()

Return hash(self).

class abacusai.api_class.ProblemType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AI_AGENT = 'ai_agent'
ANOMALY_DETECTION = 'anomaly_new'
ANOMALY_OUTLIERS = 'anomaly'
EVENT_ANOMALY = 'event_anomaly'
CLUSTERING = 'clustering'
CLUSTERING_TIMESERIES = 'clustering_timeseries'
CUMULATIVE_FORECASTING = 'cumulative_forecasting'
NAMED_ENTITY_EXTRACTION = 'nlp_ner'
CHAT_LLM = 'chat_llm'
SENTENCE_BOUNDARY_DETECTION = 'nlp_sentence_boundary_detection'
SENTIMENT_DETECTION = 'nlp_sentiment'
DOCUMENT_CLASSIFICATION = 'nlp_classification'
DOCUMENT_SUMMARIZATION = 'nlp_summarization'
DOCUMENT_VISUALIZATION = 'nlp_document_visualization'
PERSONALIZATION = 'personalization'
PREDICTIVE_MODELING = 'regression'
FORECASTING = 'forecasting'
CUSTOM_TRAINED_MODEL = 'plug_and_play'
CUSTOM_ALGORITHM = 'trainable_plug_and_play'
FEATURE_STORE = 'feature_store'
IMAGE_CLASSIFICATION = 'vision_classification'
OBJECT_DETECTION = 'vision_object_detection'
IMAGE_VALUE_PREDICTION = 'vision_regression'
MODEL_MONITORING = 'model_monitoring'
LANGUAGE_DETECTION = 'language_detection'
OPTIMIZATION = 'optimization'
PRETRAINED_MODELS = 'pretrained'
THEME_ANALYSIS = 'theme_analysis'
class abacusai.api_class.RegressionObjective

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AUC = 'auc'
ACCURACY = 'acc'
LOG_LOSS = 'log_loss'
PRECISION = 'precision'
RECALL = 'recall'
F1_SCORE = 'fscore'
MAE = 'mae'
MAPE = 'mape'
WAPE = 'wape'
RMSE = 'rmse'
R_SQUARED_COEFFICIENT_OF_DETERMINATION = 'r^2'
class abacusai.api_class.RegressionTreeHPOMode

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

RAPID = ('rapid',)
THOROUGH = 'thorough'
class abacusai.api_class.RegressionAugmentationStrategy

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

SMOTE = 'smote'
RESAMPLE = 'resample'
class abacusai.api_class.RegressionTargetTransform

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

LOG = 'log'
QUANTILE = 'quantile'
YEO_JOHNSON = 'yeo-johnson'
BOX_COX = 'box-cox'
class abacusai.api_class.RegressionTypeOfSplit

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

RANDOM = 'Random Sampling'
TIMESTAMP_BASED = 'Timestamp Based'
ROW_INDICATOR_BASED = 'Row Indicator Based'
class abacusai.api_class.RegressionTimeSplitMethod

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

TEST_SPLIT_PERCENTAGE_BASED = 'Test Split Percentage Based'
TEST_START_TIMESTAMP_BASED = 'Test Start Timestamp Based'
class abacusai.api_class.RegressionLossFunction

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

HUBER = 'Huber'
MSE = 'Mean Squared Error'
MAE = 'Mean Absolute Error'
MAPE = 'Mean Absolute Percentage Error'
MSLE = 'Mean Squared Logarithmic Error'
TWEEDIE = 'Tweedie'
CROSS_ENTROPY = 'Cross Entropy'
FOCAL_CROSS_ENTROPY = 'Focal Cross Entropy'
AUTOMATIC = 'Automatic'
CUSTOM = 'Custom'
class abacusai.api_class.ExplainerType

Bases: enum.Enum

Generic enumeration.

Derive from this class to define new enumerations.

KERNEL_EXPLAINER = 'KERNEL_EXPLAINER'
LIME_EXPLAINER = 'LIME_EXPLAINER'
TREE_EXPLAINER = 'TREE_EXPLAINER'
EBM_EXPLAINER = 'EBM_EXPLAINER'
class abacusai.api_class.SamplingMethodType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

N_SAMPLING = 'N_SAMPLING'
PERCENT_SAMPLING = 'PERCENT_SAMPLING'
class abacusai.api_class.MergeMode

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

LAST_N = 'LAST_N'
TIME_WINDOW = 'TIME_WINDOW'
class abacusai.api_class.FillLogic

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AVERAGE = 'average'
MAX = 'max'
MEDIAN = 'median'
MIN = 'min'
CUSTOM = 'custom'
BACKFILL = 'bfill'
FORWARDFILL = 'ffill'
LINEAR = 'linear'
NEAREST = 'nearest'
class abacusai.api_class.BatchSize

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

BATCH_8 = 8
BATCH_16 = 16
BATCH_32 = 32
BATCH_64 = 64
BATCH_128 = 128
BATCH_256 = 256
BATCH_384 = 384
BATCH_512 = 512
BATCH_740 = 740
BATCH_1024 = 1024
class abacusai.api_class.HolidayCalendars

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AU = 'AU'
UK = 'UK'
US = 'US'
class abacusai.api_class.ExperimentationMode

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

RAPID = 'rapid'
THOROUGH = 'thorough'
class abacusai.api_class.PersonalizationTrainingMode

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

EXPERIMENTAL = 'EXP'
PRODUCTION = 'PROD'
class abacusai.api_class.PersonalizationObjective

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

NDCG = 'ndcg'
NDCG_5 = 'ndcg@5'
NDCG_10 = 'ndcg@10'
MAP = 'map'
MAP_5 = 'map@5'
MAP_10 = 'map@10'
MRR = 'mrr'
PERSONALIZATION = 'personalization@10'
COVERAGE = 'coverage'
class abacusai.api_class.ForecastingObjective

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

ACCURACY = 'w_c_accuracy'
WAPE = 'wape'
MAPE = 'mape'
CMAPE = 'cmape'
RMSE = 'rmse'
CV = 'coefficient_of_variation'
BIAS = 'bias'
SRMSE = 'srmse'
class abacusai.api_class.ForecastingFrequency

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

HOURLY = '1H'
DAILY = '1D'
WEEKLY_SUNDAY_START = '1W'
WEEKLY_MONDAY_START = 'W-MON'
WEEKLY_SATURDAY_START = 'W-SAT'
MONTH_START = 'MS'
MONTH_END = '1M'
QUARTER_START = 'QS'
QUARTER_END = '1Q'
YEARLY = '1Y'
class abacusai.api_class.ForecastingDataSplitType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AUTO = 'Automatic Time Based'
TIMESTAMP = 'Timestamp Based'
ITEM = 'Item Based'
PREDICTION_LENGTH = 'Force Prediction Length'
class abacusai.api_class.ForecastingLossFunction

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

CUSTOM = 'Custom'
MEAN_ABSOLUTE_ERROR = 'mae'
NORMALIZED_MEAN_ABSOLUTE_ERROR = 'nmae'
PEAKS_MEAN_ABSOLUTE_ERROR = 'peaks_mae'
MEAN_ABSOLUTE_PERCENTAGE_ERROR = 'stable_mape'
POINTWISE_ACCURACY = 'accuracy'
ROOT_MEAN_SQUARE_ERROR = 'rmse'
NORMALIZED_ROOT_MEAN_SQUARE_ERROR = 'nrmse'
ASYMMETRIC_MEAN_ABSOLUTE_PERCENTAGE_ERROR = 'asymmetric_mape'
STABLE_STANDARDIZED_MEAN_ABSOLUTE_PERCENTAGE_ERROR = 'stable_standardized_mape_with_cmape'
GAUSSIAN = 'mle_gaussian_local'
GAUSSIAN_FULL_COVARIANCE = 'mle_gaussfullcov'
GUASSIAN_EXPONENTIAL = 'mle_gaussexp'
MIX_GAUSSIANS = 'mle_gaussmix'
WEIBULL = 'mle_weibull'
NEGATIVE_BINOMIAL = 'mle_negbinom'
LOG_ROOT_MEAN_SQUARE_ERROR = 'log_rmse'
class abacusai.api_class.ForecastingLocalScaling

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

ZSCORE = 'zscore'
SLIDING_ZSCORE = 'sliding_zscore'
LAST_POINT = 'lastpoint'
MIN_MAX = 'minmax'
MIN_STD = 'minstd'
ROBUST = 'robust'
ITEM = 'item'
class abacusai.api_class.ForecastingFillMethod

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

BACK = 'BACK'
MIDDLE = 'MIDDLE'
FUTURE = 'FUTURE'
class abacusai.api_class.ForecastingQuanitlesExtensionMethod

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

DIRECT = 'direct'
QUADRATIC = 'quadratic'
ANCESTRAL_SIMULATION = 'simulation'
class abacusai.api_class.NERObjective

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

LOG_LOSS = 'log_loss'
AUC = 'auc'
PRECISION = 'precision'
RECALL = 'recall'
ANNOTATIONS_PRECISION = 'annotations_precision'
ANNOTATIONS_RECALL = 'annotations_recall'
class abacusai.api_class.NERModelType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

PRETRAINED_BERT = 'pretrained_bert'
PRETRAINED_ROBERTA_27 = 'pretrained_roberta_27'
PRETRAINED_ROBERTA_43 = 'pretrained_roberta_43'
PRETRAINED_MULTILINGUAL = 'pretrained_multilingual'
LEARNED = 'learned'
class abacusai.api_class.NLPDocumentFormat

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AUTO = 'auto'
TEXT = 'text'
DOC = 'doc'
TOKENS = 'tokens'
class abacusai.api_class.SentimentType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

VALENCE = 'valence'
EMOTION = 'emotion'
class abacusai.api_class.ClusteringImputationMethod

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

AUTOMATIC = 'Automatic'
ZEROS = 'Zeros'
INTERPOLATE = 'Interpolate'
class abacusai.api_class.ConnectorType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

FILE = 'FILE'
DATABASE = 'DATABASE'
STREAMING = 'STREAMING'
APPLICATION = 'APPLICATION'
class abacusai.api_class.PythonFunctionArgumentType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

FEATURE_GROUP = 'FEATURE_GROUP'
INTEGER = 'INTEGER'
STRING = 'STRING'
BOOLEAN = 'BOOLEAN'
FLOAT = 'FLOAT'
JSON = 'JSON'
LIST = 'LIST'
DATASET_ID = 'DATASET_ID'
MODEL_ID = 'MODEL_ID'
FEATURE_GROUP_ID = 'FEATURE_GROUP_ID'
MONITOR_ID = 'MONITOR_ID'
BATCH_PREDICTION_ID = 'BATCH_PREDICTION_ID'
DEPLOYMENT_ID = 'DEPLOYMENT_ID'
class abacusai.api_class.PythonFunctionOutputArgumentType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

NTEGER = 'INTEGER'
STRING = 'STRING'
BOOLEAN = 'BOOLEAN'
FLOAT = 'FLOAT'
JSON = 'JSON'
LIST = 'LIST'
DATASET_ID = 'DATASET_ID'
MODEL_ID = 'MODEL_ID'
FEATURE_GROUP_ID = 'FEATURE_GROUP_ID'
MONITOR_ID = 'MONITOR_ID'
BATCH_PREDICTION_ID = 'BATCH_PREDICTION_ID'
DEPLOYMENT_ID = 'DEPLOYMENT_ID'
ANY = 'ANY'
class abacusai.api_class.VectorStoreTextEncoder

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

E5 = 'E5'
OPENAI = 'OPENAI'
SENTENCE_BERT = 'SENTENCE_BERT'
E5_SMALL = 'E5_SMALL'
class abacusai.api_class.LLMName

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

OPENAI_GPT4 = 'OPENAI_GPT4'
OPENAI_GPT3_5 = 'OPENAI_GPT3_5'
OPENAI_GPT3_5_SHORT = 'OPENAI_GPT3_5_SHORT'
CLAUDE_V2 = 'CLAUDE_V2'
ABACUS_GIRAFFE = 'ABACUS_GIRAFFE'
ABACUS_LLAMA2_QA = 'ABACUS_LLAMA2_QA'
ABACUS_LLAMA2_CODE = 'ABACUS_LLAMA2_CODE'
LLAMA2_CHAT = 'LLAMA2_CHAT'
PALM = 'PALM'
PALM_TEXT = 'PALM_TEXT'
class abacusai.api_class.MonitorAlertType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

ACCURACY_BELOW_THRESHOLD = 'AccuracyBelowThreshold'
FEATURE_DRIFT = 'FeatureDrift'
DATA_INTEGRITY_VIOLATIONS = 'DataIntegrityViolations'
BIAS_VIOLATIONS = 'BiasViolations'
class abacusai.api_class.FeatureDriftType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

KL = 'kl'
KS = 'ks'
WS = 'ws'
JS = 'js'
class abacusai.api_class.DataIntegrityViolationType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

NULL_VIOLATIONS = 'null_violations'
TYPE_MISMATCH_VIOLATIONS = 'type_mismatch_violations'
RANGE_VIOLATIONS = 'range_violations'
CATEGORICAL_RANGE_VIOLATION = 'categorical_range_violations'
TOTAL_VIOLATIONS = 'total_violations'
class abacusai.api_class.BiasType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

DEMOGRAPHIC_PARITY = 'demographic_parity'
EQUAL_OPPORTUNITY = 'equal_opportunity'
GROUP_BENEFIT_EQUALITY = 'group_benefit'
TOTAL = 'total'
class abacusai.api_class.AlertActionType

Bases: ApiEnum

Generic enumeration.

Derive from this class to define new enumerations.

EMAIL = 'Email'
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class._ApiClassFactory

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key
config_class_map
classmethod from_dict(config)
Parameters:

config (dict) –

Return type:

ApiClass

class abacusai.api_class.SamplingConfig

Bases: abacusai.api_class.abstract.ApiClass

An abstract class for the sampling config of a feature group

classmethod _get_builder()
__post_init__()
class abacusai.api_class.NSamplingConfig

Bases: SamplingConfig

The number of distinct values of the key columns to include in the sample, or number of rows if key columns not specified.

Parameters:
  • sampling_method (SamplingMethodType) – N_SAMPLING

  • sample_count (int) – The number of rows to include in the sample

  • key_columns (List[str]) – The feature(s) to use as the key(s) when sampling

sample_count: int
key_columns: List[str]
sampling_method: abacusai.api_class.enums.SamplingMethodType
class abacusai.api_class.PercentSamplingConfig

Bases: SamplingConfig

The fraction of distinct values of the feature group to include in the sample.

Parameters:
  • sampling_method (SamplingMethodType) – PERCENT_SAMPLING

  • sample_percent (float) – The percentage of the rows to sample

  • key_columns (List[str]) – The feature(s) to use as the key(s) when sampling

sample_percent: float
key_columns: List[str]
sampling_method: abacusai.api_class.enums.SamplingMethodType
class abacusai.api_class._SamplingConfigFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_class_key = 'sampling_method'
config_class_map
class abacusai.api_class.MergeConfig

Bases: abacusai.api_class.abstract.ApiClass

An abstract class for the merge config of a feature group

classmethod _get_builder()
__post_init__()
class abacusai.api_class.LastNMergeConfig

Bases: MergeConfig

Merge LAST N chunks/versions of an incremental dataset.

Parameters:
  • merge_mode (MergeMode) – LAST_N

  • num_versions (int) – The number of versions to merge. num_versions == 0 means merge all versions.

num_versions: int
merge_mode: abacusai.api_class.enums.MergeMode
class abacusai.api_class.TimeWindowMergeConfig

Bases: MergeConfig

Merge rows within a given timewindow of the most recent timestamp

Parameters:
  • merge_mode (MergeMode) – TIME_WINDOW

  • feature_name (str) – Time based column to index on

  • time_window_size_ms (int) – Range of merged rows will be [MAX_TIME - time_window_size_ms, MAX_TIME]

feature_name: str
time_window_size_ms: int
merge_mode: abacusai.api_class.enums.MergeMode
class abacusai.api_class._MergeConfigFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_class_key = 'merge_mode'
config_class_map
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class._ApiClassFactory

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key
config_class_map
classmethod from_dict(config)
Parameters:

config (dict) –

Return type:

ApiClass

class abacusai.api_class.TrainingConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
kwargs: dict
problem_type: abacusai.api_class.enums.ProblemType
algorithm: str
classmethod _get_builder()
class abacusai.api_class.PersonalizationTrainingConfig

Bases: TrainingConfig

Training config for the PERSONALIZATION problem type :param objective: Ranking scheme used to select final best model. :type objective: PersonalizationObjective :param sort_objective: Ranking scheme used to sort models on the metrics page. :type sort_objective: PersonalizationObjective :param training_mode: whether to train in production or experimental mode. Defaults to EXP. :type training_mode: PersonalizationTrainingMode :param target_action_types: List of action types to use as targets for training. :type target_action_types: List[str] :param target_action_weights: Dictionary of action types to weights for training. :type target_action_weights: Dict[str, float] :param session_event_types: List of event types to treat as occurrences of sessions. :type session_event_types: List[str] :param test_split: Percent of dataset to use for test data. We support using a range between 6% to 20% of your dataset to use as test data. :type test_split: int :param recent_days_for_training: Limit training data to a certain latest number of days. :type recent_days_for_training: int :param training_start_date: Only consider training interaction data after this date. Specified in the timezone of the dataset. :type training_start_date: str :param test_on_user_split: Use user splits instead of using time splits, when validating and testing the model. :type test_on_user_split: bool :param test_split_on_last_k_items: Use last k items instead of global timestamp splits, when validating and testing the model. :type test_split_on_last_k_items: bool :param test_last_items_length: Number of items to leave out for each user when using leave k out folds. :type test_last_items_length: int :param test_window_length_hours: Duration (in hours) of most recent time window to use when validating and testing the model. :type test_window_length_hours: int :param explicit_time_split: Sets an explicit time-based test boundary. :type explicit_time_split: bool :param test_row_indicator: Column indicating which rows to use for training (TRAIN), validation (VAL) and testing (TEST). :type test_row_indicator: str :param full_data_retraining: Train models separately with all the data. :type full_data_retraining: bool :param sequential_training: Train a mode sequentially through time. :type sequential_training: bool :param data_split_feature_group_table_name: Specify the table name of the feature group to export training data with the fold column. :type data_split_feature_group_table_name: str :param optimized_event_type: The final event type to optimize for and compute metrics on. :type optimized_event_type: str :param dropout_rate: Dropout rate for neural network. :type dropout_rate: int :param batch_size: Batch size for neural network. :type batch_size: BatchSize :param disable_transformer: Disable training the transformer algorithm. :type disable_transformer: bool :param disable_gpu: Disable training on GPU. :type disable_gpu: boo :param filter_history: Do not recommend items the user has already interacted with. :type filter_history: bool :param max_history_length: Maximum length of user-item history to include user in training examples. :type max_history_length: int :param compute_rerank_metrics: Compute metrics based on rerank results. :type compute_rerank_metrics: bool :param add_time_features: Include interaction time as a feature. :type add_time_features: bool :param disable_timestamp_scalar_features: Exclude timestamp scalar features. :type disable_timestamp_scalar_features: bool :param compute_session_metrics: Evaluate models based on how well they are able to predict the next session of interactions. :type compute_session_metrics: bool :param max_user_history_len_percentile: Filter out users with history length above this percentile. :type max_user_history_len_percentile: int :param downsample_item_popularity_percentile: Downsample items more popular than this percentile. :type downsample_item_popularity_percentile: float

objective: abacusai.api_class.enums.PersonalizationObjective
sort_objective: abacusai.api_class.enums.PersonalizationObjective
training_mode: abacusai.api_class.enums.PersonalizationTrainingMode
target_action_types: List[str]
target_action_weights: Dict[str, float]
session_event_types: List[str]
test_split: int
recent_days_for_training: int
training_start_date: str
test_on_user_split: bool
test_split_on_last_k_items: bool
test_last_items_length: int
test_window_length_hours: int
explicit_time_split: bool
test_row_indicator: str
full_data_retraining: bool
sequential_training: bool
data_split_feature_group_table_name: str
optimized_event_type: str
dropout_rate: int
batch_size: abacusai.api_class.enums.BatchSize
disable_transformer: bool
disable_gpu: bool
filter_history: bool
max_history_length: int
compute_rerank_metrics: bool
add_time_features: bool
disable_timestamp_scalar_features: bool
compute_session_metrics: bool
query_column: str
item_query_column: str
max_user_history_len_percentile: int
downsample_item_popularity_percentile: float
__post_init__()
class abacusai.api_class.RegressionTrainingConfig

Bases: TrainingConfig

Training config for the PREDICTIVE_MODELING problem type :param objective: Ranking scheme used to select final best model. :type objective: RegressionObjective :param sort_objective: Ranking scheme used to sort models on the metrics page. :type sort_objective: RegressionObjective :param tree_hpo_mode: (RegressionTreeHPOMode): Turning off Rapid Experimentation will take longer to train. :param type_of_split: Type of data splitting into train/test (validation also). :type type_of_split: RegressionTypeOfSplit :param test_split: Percent of dataset to use for test data. We support using a range between 5% to 20% of your dataset to use as test data. :type test_split: int :param disable_test_val_fold: Do not create a TEST_VAL set. All records which would be part of the TEST_VAL fold otherwise, remain in the TEST fold. :type disable_test_val_fold: bool :param k_fold_cross_validation: Use this to force k-fold cross validation bagging on or off. :type k_fold_cross_validation: bool :param num_cv_folds: Specify the value of k in k-fold cross validation. :type num_cv_folds: int :param timestamp_based_splitting_column: Timestamp column selected for splitting into test and train. :type timestamp_based_splitting_column: str :param timestamp_based_splitting_method: Method of selecting TEST set, top percentile wise or after a given timestamp. :type timestamp_based_splitting_method: RegressionTimeSplitMethod :param test_splitting_timestamp: Rows with timestamp greater than this will be considered to be in the test set. :type test_splitting_timestamp: str :param sampling_unit_keys: Constrain train/test separation to partition a column. :type sampling_unit_keys: List[str] :param test_row_indicator: Column indicating which rows to use for training (TRAIN) and testing (TEST). Validation (VAL) can also be specified. :type test_row_indicator: str :param full_data_retraining: Train models separately with all the data. :type full_data_retraining: bool :param rebalance_classes: Class weights are computed as the inverse of the class frequency from the training dataset when this option is selected as “Yes”. It is useful when the classes in the dataset are unbalanced.

Re-balancing classes generally boosts recall at the cost of precision on rare classes.

Parameters:
  • rare_class_augmentation_threshold (float) – Augments any rare class whose relative frequency with respect to the most frequent class is less than this threshold. Default = 0.1 for classification problems with rare classes.

  • augmentation_strategy (RegressionAugmentationStrategy) – Strategy to deal with class imbalance and data augmentation.

  • training_rows_downsample_ratio (float) – Uses this ratio to train on a sample of the dataset provided.

  • active_labels_column (str) – Specify a column to use as the active columns in a multi label setting.

  • min_categorical_count (int) – Minimum threshold to consider a value different from the unknown placeholder.

  • sample_weight (str) – Specify a column to use as the weight of a sample for training and eval.

  • numeric_clipping_percentile (float) – Uses this option to clip the top and bottom x percentile of numeric feature columns where x is the value of this option.

  • target_transform (RegressionTargetTransform) – Specify a transform (e.g. log, quantile) to apply to the target variable.

  • ignore_datetime_features (bool) – Remove all datetime features from the model. Useful while generalizing to different time periods.

  • max_text_words (int) – Maximum number of words to use from text fields.

  • perform_feature_selection (bool) – If enabled, additional algorithms which support feature selection as a pretraining step will be trained separately with the selected subset of features. The details about their selected features can be found in their respective logs.

  • feature_selection_intensity (int) – This determines the strictness with which features will be filtered out. 1 being very lenient (more features kept), 100 being very strict.

  • batch_size (BatchSize) – Batch size.

  • dropout_rate (int) – Dropout percentage rate.

  • pretrained_model_name (str) – Enable algorithms which process text using pretrained multilingual NLP models.

  • is_multilingual (bool) – Enable algorithms which process text using pretrained multilingual NLP models.

  • loss_function (RegressionLossFunction) – Loss function to be used as objective for model training.

  • loss_parameters (str) – Loss function params in format <key>=<value>;<key>=<value>;…..

  • target_encode_categoricals (bool) – Use this to turn target encoding on categorical features on or off.

  • drop_original_categoricals (bool) – This option helps us choose whether to also feed the original label encoded categorical columns to the mdoels along with their target encoded versions.

  • monotonically_increasing_features (List[str]) – Constrain the model such that it behaves as if the target feature is monotonically increasing with the selected features

  • monotonically_decreasing_features (List[str]) – Constrain the model such that it behaves as if the target feature is monotonically decreasing with the selected features

  • data_split_feature_group_table_name (str) – Specify the table name of the feature group to export training data with the fold column.

  • custom_loss_functions (List[str]) – Registered custom losses available for selection.

  • custom_metrics (List[str]) – Registered custom metrics available for selection.

objective: abacusai.api_class.enums.RegressionObjective
sort_objective: abacusai.api_class.enums.RegressionObjective
tree_hpo_mode: abacusai.api_class.enums.RegressionTreeHPOMode
type_of_split: abacusai.api_class.enums.RegressionTypeOfSplit
test_split: int
disable_test_val_fold: bool
k_fold_cross_validation: bool
num_cv_folds: int
timestamp_based_splitting_column: str
timestamp_based_splitting_method: abacusai.api_class.enums.RegressionTimeSplitMethod
test_splitting_timestamp: str
sampling_unit_keys: List[str]
test_row_indicator: str
full_data_retraining: bool
rebalance_classes: bool
rare_class_augmentation_threshold: float
augmentation_strategy: abacusai.api_class.enums.RegressionAugmentationStrategy
training_rows_downsample_ratio: float
active_labels_column: str
min_categorical_count: int
sample_weight: str
numeric_clipping_percentile: float
target_transform: abacusai.api_class.enums.RegressionTargetTransform
ignore_datetime_features: bool
max_text_words: int
perform_feature_selection: bool
feature_selection_intensity: int
batch_size: abacusai.api_class.enums.BatchSize
dropout_rate: int
pretrained_model_name: str
is_multilingual: bool
loss_function: abacusai.api_class.enums.RegressionLossFunction
loss_parameters: str
target_encode_categoricals: bool
drop_original_categoricals: bool
monotonically_increasing_features: List[str]
monotonically_decreasing_features: List[str]
data_split_feature_group_table_name: str
custom_loss_functions: List[str]
custom_metrics: List[str]
__post_init__()
class abacusai.api_class.ForecastingTrainingConfig

Bases: TrainingConfig

Training config for the FORECASTING problem type :param prediction_length: How many timesteps in the future to predict. :type prediction_length: int :param objective: Ranking scheme used to select final best model. :type objective: ForecastingObjective :param sort_objective: Ranking scheme used to sort models on the metrics page. :type sort_objective: ForecastingObjective :param forecast_frequency: Forecast frequency. :type forecast_frequency: ForecastingFrequency :param probability_quantiles: Prediction quantiles. :type probability_quantiles: List[float] :param force_prediction_length: Force length of test window to be the same as prediction length. :type force_prediction_length: int :param filter_items: Filter items with small history and volume. :type filter_items: bool :param enable_feature_selection: Enable feature selection. :type enable_feature_selection: bool :param enable_padding: Pad series to the max_date of the dataset :type enable_padding: bool :param enable_cold_start: Enable cold start forecasting by training/predicting for zero history items. :type enable_cold_start: bool :param enable_multiple_backtests: Whether to enable multiple backtesting or not. :type enable_multiple_backtests: bool :param num_backtesting_windows: Total backtesting windows to use for the training. :type num_backtesting_windows: int :param backtesting_window_step_size: Use this step size to shift backtesting windows for model training. :type backtesting_window_step_size: int :param full_data_retraining: Train models separately with all the data. :type full_data_retraining: bool :param additional_forecast_keys: List[str]: List of categoricals in timeseries that can act as multi-identifier. :param experimentation_mode: Selecting Thorough Experimentation will take longer to train. :type experimentation_mode: ExperimentationMode :param type_of_split: Type of data splitting into train/test. :type type_of_split: ForecastingDataSplitType :param test_by_item: Partition train/test data by item rather than time if true. :type test_by_item: bool :param test_start: Limit training data to dates before the given test start. :type test_start: str :param test_split: Percent of dataset to use for test data. We support using a range between 5% to 20% of your dataset to use as test data. :type test_split: int :param loss_function: Loss function for training neural network. :type loss_function: ForecastingLossFunction :param underprediction_weight: Weight for underpredictions :type underprediction_weight: float :param disable_networks_without_analytic_quantiles: Disable neural networks, which quantile functions do not have analytic expressions (e.g, mixture models) :type disable_networks_without_analytic_quantiles: bool :param initial_learning_rate: Initial learning rate. :type initial_learning_rate: float :param l2_regularization_factor: L2 regularization factor. :type l2_regularization_factor: float :param dropout_rate: Dropout percentage rate. :type dropout_rate: int :param recurrent_layers: Number of recurrent layers to stack in network. :type recurrent_layers: int :param recurrent_units: Number of units in each recurrent layer. :type recurrent_units: int :param convolutional_layers: Number of convolutional layers to stack on top of recurrent layers in network. :type convolutional_layers: int :param convolution_filters: Number of filters in each convolution. :type convolution_filters: int :param local_scaling_mode: Options to make NN inputs stationary in high dynamic range datasets. :type local_scaling_mode: ForecastingLocalScaling :param zero_predictor: Include subnetwork to classify points where target equals zero. :type zero_predictor: bool :param skip_missing: Make the RNN ignore missing entries rather instead of processing them. :type skip_missing: bool :param batch_size: Batch size. :type batch_size: ForecastingBatchSize :param batch_renormalization: Enable batch renormalization between layers. :type batch_renormalization: bool :param history_length: While training, how much history to consider. :type history_length: int :param prediction_step_size: Number of future periods to include in objective for each training sample. :type prediction_step_size: int :param training_point_overlap: Amount of overlap to allow between training samples. :type training_point_overlap: float :param max_scale_context: Maximum context to use for local scaling. :type max_scale_context: int :param quantiles_extension_method: Quantile extension method :type quantiles_extension_method: ForecastingQuanitlesExtensionMethod :param number_of_samples: Number of samples for ancestral simulation :type number_of_samples: int :param symmetrize_quantiles: Force symmetric quantiles (like in Gaussian distribution) :type symmetrize_quantiles: bool :param use_log_transforms: Apply logarithmic transformations to input data. :type use_log_transforms: bool :param smooth_history: Smooth (low pass filter) the timeseries. :type smooth_history: float :param local_scale_target: Using per training/prediction window target scaling. :type local_scale_target: bool :param timeseries_weight_column: If set, we use the values in this column from timeseries data to assign time dependent item weights during training and evaluation. :type timeseries_weight_column: str :param item_attributes_weight_column: If set, we use the values in this column from item attributes data to assign weights to items during training and evaluation. :type item_attributes_weight_column: str :param use_timeseries_weights_in_objective: If True, we include weights from column set as “TIMESERIES WEIGHT COLUMN” in objective functions. :type use_timeseries_weights_in_objective: bool :param use_item_weights_in_objective: If True, we include weights from column set as “ITEM ATTRIBUTES WEIGHT COLUMN” in objective functions. :type use_item_weights_in_objective: bool :param skip_timeseries_weight_scaling: If True, we will avoid normalizing the weights. :type skip_timeseries_weight_scaling: bool :param timeseries_loss_weight_column: Use value in this column to weight the loss while training. :type timeseries_loss_weight_column: str :param use_item_id: Include a feature to indicate the item being forecast. :type use_item_id: bool :param use_all_item_totals: Include as input total target across items. :type use_all_item_totals: bool :param handle_zeros_as_missing_values: If True, handle zero values in demand as missing data. :type handle_zeros_as_missing_values: bool :param datetime_holiday_calendars: Holiday calendars to augment training with. :type datetime_holiday_calendars: List[HolidayCalendars] :param fill_missing_values: Strategy for filling in missing values. :type fill_missing_values: List[dict] :param enable_clustering: Enable clustering in forecasting. :type enable_clustering: bool :param data_split_feature_group_table_name: Specify the table name of the feature group to export training data with the fold column. :type data_split_feature_group_table_name: str :param custom_loss_functions: Registered custom losses available for selection. :type custom_loss_functions: List[str] :param custom_metrics: Registered custom metrics available for selection. :type custom_metrics: List[str]

prediction_length: int
objective: abacusai.api_class.enums.ForecastingObjective
sort_objective: abacusai.api_class.enums.ForecastingObjective
forecast_frequency: abacusai.api_class.enums.ForecastingFrequency
probability_quantiles: List[float]
force_prediction_length: bool
filter_items: bool
enable_feature_selection: bool
enable_padding: bool
enable_cold_start: bool
enable_multiple_backtests: bool
num_backtesting_windows: int
backtesting_window_step_size: int
full_data_retraining: bool
additional_forecast_keys: List[str]
experimentation_mode: abacusai.api_class.enums.ExperimentationMode
type_of_split: abacusai.api_class.enums.ForecastingDataSplitType
test_by_item: bool
test_start: str
test_split: int
loss_function: abacusai.api_class.enums.ForecastingLossFunction
underprediction_weight: float
disable_networks_without_analytic_quantiles: bool
initial_learning_rate: float
l2_regularization_factor: float
dropout_rate: int
recurrent_layers: int
recurrent_units: int
convolutional_layers: int
convolution_filters: int
local_scaling_mode: abacusai.api_class.enums.ForecastingLocalScaling
zero_predictor: bool
skip_missing: bool
batch_size: abacusai.api_class.enums.BatchSize
batch_renormalization: bool
history_length: int
prediction_step_size: int
training_point_overlap: float
max_scale_context: int
quantiles_extension_method: abacusai.api_class.enums.ForecastingQuanitlesExtensionMethod
number_of_samples: int
symmetrize_quantiles: bool
use_log_transforms: bool
smooth_history: float
local_scale_target: bool
timeseries_weight_column: str
item_attributes_weight_column: str
use_timeseries_weights_in_objective: bool
use_item_weights_in_objective: bool
skip_timeseries_weight_scaling: bool
timeseries_loss_weight_column: str
use_item_id: bool
use_all_item_totals: bool
handle_zeros_as_missing_values: bool
datetime_holiday_calendars: List[abacusai.api_class.enums.HolidayCalendars]
fill_missing_values: List[dict]
enable_clustering: bool
data_split_feature_group_table_name: str
custom_loss_functions: List[str]
custom_metrics: List[str]
__post_init__()
class abacusai.api_class.NamedEntityExtractionTrainingConfig

Bases: TrainingConfig

Training config for the NAMED_ENTITY_EXTRACTION problem type :param objective: Ranking scheme used to select final best model. :type objective: NERObjective :param sort_objective: Ranking scheme used to sort models on the metrics page. :type sort_objective: NERObjective :param ner_model_type: Type of NER model to use. :type ner_model_type: NERModelType :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param test_row_indicator: Column indicating which rows to use for training (TRAIN) and testing (TEST). :type test_row_indicator: str :param dropout_rate: Dropout rate for neural network. :type dropout_rate: float :param batch_size: Batch size for neural network. :type batch_size: BatchSize :param active_labels_column: Entities that have been marked in a particular text :type active_labels_column: str :param document_format: Format of the input documents. :type document_format: NLPDocumentFormat :param include_longformer: Whether to include the longformer model. :type include_longformer: bool

objective: abacusai.api_class.enums.NERObjective
sort_objective: abacusai.api_class.enums.NERObjective
ner_model_type: abacusai.api_class.enums.NERModelType
test_split: int
test_row_indicator: str
dropout_rate: float
batch_size: abacusai.api_class.enums.BatchSize
active_labels_column: str
document_format: abacusai.api_class.enums.NLPDocumentFormat
include_longformer: bool
__post_init__()
class abacusai.api_class.NaturalLanguageSearchTrainingConfig

Bases: TrainingConfig

Training config for the NATURAL_LANGUAGE_SEARCH problem type :param abacus_internal_model: Use a Abacus.AI LLM to answer questions about your data without using any external APIs :type abacus_internal_model: bool :param num_completion_tokens: Default for maximum number of tokens for chat answers. Reducing this will get faster responses which are more succinct :type num_completion_tokens: int :param larger_embeddings: Use a higher dimension embedding model. :type larger_embeddings: bool :param search_chunk_size: Chunk size for indexing the documents. :type search_chunk_size: int :param chunk_overlap_fraction: Overlap in chunks while indexing the documents. :type chunk_overlap_fraction: float :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int

abacus_internal_model: bool
num_completion_tokens: int
larger_embeddings: bool
search_chunk_size: int
index_fraction: float
chunk_overlap_fraction: float
__post_init__()
class abacusai.api_class.ChatLLMTrainingConfig

Bases: TrainingConfig

Training config for the CHAT_LLM problem type :param document_retrievers: List of document retriever names to use for the feature stores this model was trained with. :type document_retrievers: List[str] :param num_completion_tokens: Default for maximum number of tokens for chat answers. Reducing this will get faster responses which are more succinct :type num_completion_tokens: int :param system_message: The generative LLM system message :type system_message: str :param temperature: The generative LLM temperature :type temperature: float :param metadata_columns: Include the metadata column values in the retrieved search results. :type metadata_columns: list

document_retrievers: List[str]
num_completion_tokens: int
system_message: str
temperature: float
metadata_columns: list
__post_init__()
class abacusai.api_class.SentenceBoundaryDetectionTrainingConfig

Bases: TrainingConfig

Training config for the SENTENCE_BOUNDARY_DETECTION problem type :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param dropout_rate: Dropout rate for neural network. :type dropout_rate: float :param batch_size: Batch size for neural network. :type batch_size: BatchSize

test_split: int
dropout_rate: float
batch_size: abacusai.api_class.enums.BatchSize
__post_init__()
class abacusai.api_class.SentimentDetectionTrainingConfig

Bases: TrainingConfig

Training config for the SENTIMENT_DETECTION problem type :param sentiment_type: Type of sentiment to detect. :type sentiment_type: SentimentType :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param dropout_rate: Dropout rate for neural network. :type dropout_rate: float :param batch_size: Batch size for neural network. :type batch_size: BatchSize :param compute_metrics: Whether to compute metrics. :type compute_metrics: bool

sentiment_type: abacusai.api_class.enums.SentimentType
test_split: int
dropout_rate: float
batch_size: abacusai.api_class.enums.BatchSize
compute_metrics: bool
__post_init__()
class abacusai.api_class.DocumentClassificationTrainingConfig

Bases: TrainingConfig

Training config for the DOCUMENT_CLASSIFICATION problem type :param zero_shot_hypotheses: Zero shot hypotheses. Example text: ‘This text is about pricing’. :type zero_shot_hypotheses: List[str] :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param dropout_rate: Dropout rate for neural network. :type dropout_rate: float :param batch_size: Batch size for neural network. :type batch_size: BatchSize

zero_shot_hypotheses: List[str]
test_split: int
dropout_rate: float
batch_size: abacusai.api_class.enums.BatchSize
__post_init__()
class abacusai.api_class.DocumentSummarizationTrainingConfig

Bases: TrainingConfig

Training config for the DOCUMENT_SUMMARIZATION problem type :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param dropout_rate: Dropout rate for neural network. :type dropout_rate: float :param batch_size: Batch size for neural network. :type batch_size: BatchSize

test_split: int
dropout_rate: float
batch_size: abacusai.api_class.enums.BatchSize
__post_init__()
class abacusai.api_class.DocumentVisualizationTrainingConfig

Bases: TrainingConfig

Training config for the DOCUMENT_VISUALIZATION problem type :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param dropout_rate: Dropout rate for neural network. :type dropout_rate: float :param batch_size: Batch size for neural network. :type batch_size: BatchSize

test_split: int
dropout_rate: float
batch_size: abacusai.api_class.enums.BatchSize
__post_init__()
class abacusai.api_class.ClusteringTrainingConfig

Bases: TrainingConfig

Training config for the CLUSTERING problem type :param num_clusters_selection: Number of clusters. If None, will be selected automatically. :type num_clusters_selection: int

num_clusters_selection: int
__post_init__()
class abacusai.api_class.ClusteringTimeseriesTrainingConfig

Bases: TrainingConfig

Training config for the CLUSTERING_TIMESERIES problem type :param num_clusters_selection: Number of clusters. If None, will be selected automatically. :type num_clusters_selection: int :param imputation: Imputation method for missing values. :type imputation: ClusteringImputationMethod

num_clusters_selection: int
imputation: abacusai.api_class.enums.ClusteringImputationMethod
__post_init__()
class abacusai.api_class.EventAnomalyTrainingConfig

Bases: TrainingConfig

Training config for the EVENT_ANOMALY problem type :param anomaly_fraction: The fraction of the dataset to classify as anomalous, between 0 and 0.5 :type anomaly_fraction: float

anomaly_fraction: float
__post_init__()
class abacusai.api_class.CumulativeForecastingTrainingConfig

Bases: TrainingConfig

Training config for the CUMULATIVE_FORECASTING problem type :param test_split: Percent of dataset to use for test data. We support using a range between 5 ( i.e. 5% ) to 20 ( i.e. 20% ) of your dataset. :type test_split: int :param historical_frequency: Forecast frequency :type historical_frequency: str :param cumulative_prediction_lengths: List of Cumulative Prediction Frequencies. Each prediction length must be between 1 and 365. :type cumulative_prediction_lengths: List[int] :param skip_input_transform: Avoid doing numeric scaling transformations on the input. :type skip_input_transform: bool :param skip_target_transform: Avoid doing numeric scaling transformations on the target. :type skip_target_transform: bool :param predict_residuals: Predict residuals instead of totals at each prediction step. :type predict_residuals: bool

test_split: int
historical_frequency: str
cumulative_prediction_lengths: List[int]
skip_input_transform: bool
skip_target_transform: bool
predict_residuals: bool
__post_init__()
class abacusai.api_class.AnomalyDetectionTrainingConfig

Bases: TrainingConfig

Training config for the ANOMALY_DETECTION problem type :param test_split: Percent of dataset to use for test data. We support using a range between 5 (i.e. 5%) to 20 (i.e. 20%) of your dataset to use as test data. :type test_split: int :param value_high: Detect unusually high values. :type value_high: bool :param mixture_of_gaussians: Detect unusual combinations of values using mixture of Gaussians. :type mixture_of_gaussians: bool :param variational_autoencoder: Use variational autoencoder for anomaly detection. :type variational_autoencoder: bool :param spike_up: Detect outliers with a high value. :type spike_up: bool :param spike_down: Detect outliers with a low value. :type spike_down: bool :param trend_change: Detect changes to the trend. :type trend_change: bool

test_split: int
value_high: bool
mixture_of_gaussians: bool
variational_autoencoder: bool
spike_up: bool
spike_down: bool
trend_change: bool
__post_init__()
class abacusai.api_class.ThemeAnalysisTrainingConfig

Bases: TrainingConfig

Training config for the THEME ANALYSIS problem type

__post_init__()
class abacusai.api_class.AIAgentTrainingConfig

Bases: TrainingConfig

Training config for the AI_AGENT problem type :param description: Description of the agent function. :type description: str :param enable_binary_input: If True, the agent will be able to accept binary data as inputs. :type enable_binary_input: bool

description: str
enable_binary_input: bool
__post_init__()
class abacusai.api_class.CustomTrainedModelTrainingConfig

Bases: TrainingConfig

Training config for the CUSTOM_TRAINED_MODEL problem type :param max_catalog_size: Maximum expected catalog size. :type max_catalog_size: int :param max_dimension: Maximum expected dimension of the catalog. :type max_dimension: int :param index_output_path: Fully qualified cloud location (GCS, S3, etc) to export snapshots of the embedding to. :type index_output_path: str :param docker_image_uri: Docker image URI. :type docker_image_uri: str :param service_port: Service port. :type service_port: int

max_catalog_size: int
max_dimension: int
index_output_path: str
docker_image_uri: str
service_port: int
__post_init__()
class abacusai.api_class.CustomAlgorithmTrainingConfig

Bases: TrainingConfig

Training config for the CUSTOM_ALGORITHM problem type :param train_function_name: The name of the train function. :type train_function_name: str :param predict_many_function_name: The name of the predict many function. :type predict_many_function_name: str :param training_input_tables: List of tables to use for training. :type training_input_tables: List[str] :param predict_function_name: Optional name of the predict function if the predict many function is not given. :type predict_function_name: str :param train_module_name: The name of the train module - only relevant if model is being uploaded from a zip file or github repositoty. :type train_module_name: str :param predict_module_name: The name of the predict module - only relevant if model is being uploaded from a zip file or github repositoty. :type predict_module_name: str :param test_split: Percent of dataset to use for test data. We support using a range between 6% to 20% of your dataset to use as test data. :type test_split: int

train_function_name: str
predict_many_function_name: str
training_input_tables: List[str]
predict_function_name: str
train_module_name: str
predict_module_name: str
test_split: int
__post_init__()
class abacusai.api_class.OptimizationTrainingConfig

Bases: TrainingConfig

Training config for the OPTIMIZATION problem type :param solve_time_limit: The maximum time in seconds to spend solving the problem. Accepts values between 0 and 86400. :type solve_time_limit: float

solve_time_limit: float
__post_init__()
class abacusai.api_class._TrainingConfigFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key = 'problem_type'
config_class_map
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.ForecastingMonitorConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

id_column: str
timestamp_column: str
target_column: str
start_time: str
end_time: str
to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class._ApiClassFactory

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key
config_class_map
classmethod from_dict(config)
Parameters:

config (dict) –

Return type:

ApiClass

class abacusai.api_class.AlertConditionConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

alert_type: abacusai.api_class.enums.MonitorAlertType
classmethod _get_builder()
class abacusai.api_class.AccuracyBelowThresholdConditionConfig

Bases: AlertConditionConfig

Accuracy Below Threshold Condition Config for Monitor Alerts :param threshold: Threshold for when to consider a column to be in violation. The alert will only fire when the drift value is strictly greater than the threshold. :type threshold: float

threshold: float
class abacusai.api_class.FeatureDriftConditionConfig

Bases: AlertConditionConfig

Feature Drift Condition Config for Monitor Alerts :param feature_drift_type: Feature drift type to apply the threshold on to determine whether a column has drifted significantly enough to be a violation. :type feature_drift_type: str :param threshold: Threshold for when to consider a column to be in violation. The alert will only fire when the drift value is strictly greater than the threshold. :type threshold: float :param minimum_violations: Number of columns that must exceed the specified threshold to trigger an alert. :type minimum_violations: int

feature_drift_type: abacusai.api_class.enums.FeatureDriftType
threshold: float
minimum_violations: int
class abacusai.api_class.DataIntegrityViolationConditionConfig

Bases: AlertConditionConfig

Data Integrity Violation Condition Config for Monitor Alerts :param data_integrity_type: This option selects the data integrity violations to monitor for this alert. :type data_integrity_type: enums.DataIntegrityViolationType :param minimum_violations: Number of columns that must exceed the specified threshold to trigger an alert. :type minimum_violations: int

data_integrity_type: abacusai.api_class.enums.DataIntegrityViolationType
minimum_violations: int
class abacusai.api_class.BiasViolationConditionConfig

Bases: AlertConditionConfig

Bias Violation Condition Config for Monitor Alerts :param bias_type: This option selects the bias metric to monitor for this alert. :type bias_type: enums.BiasType :param threshold: Threshold for when to consider a column to be in violation. The alert will only fire when the drift value is strictly greater than the threshold. :type threshold: float :param minimum_violations: Number of columns that must exceed the specified threshold to trigger an alert. :type minimum_violations: int

bias_type: abacusai.api_class.enums.BiasType
threshold: float
minimum_violations: int
class abacusai.api_class._AlertConditionConfigFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key = 'alert_type'
config_class_key_value_camel_case = True
config_class_map
class abacusai.api_class.AlertActionConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

action_type: abacusai.api_class.enums.AlertActionType
classmethod _get_builder()
class abacusai.api_class.EmailActionConfig

Bases: AlertActionConfig

Email Action Config for Monitor Alerts :param email_recipients: List of email addresses to send the alert to. :type email_recipients: List[str] :param email_body: Body of the email to send. :type email_body: str

email_recipients: List[str]
email_body: str
__post_init__()
class abacusai.api_class._AlertActionConfigFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key = 'action_type'
config_class_map
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.FeatureMappingConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

feature_name: str
feature_mapping: str
nested_feature_name: str
class abacusai.api_class.ProjectFeatureGroupTypeMappingsConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

feature_group_id: str
feature_group_type: str
feature_mappings: List[FeatureMappingConfig]
classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class.PythonFunctionArgument

Bases: abacusai.api_class.abstract.ApiClass

A config class for python function arguments

Parameters:
  • variable_type (PythonFunctionArgumentType) – The type of the python function argument

  • name (str) – The name of the python function variable

  • is_required (bool) – Whether the argument is required

  • value (Any) – The value of the argument

  • pipeline_variable (str) – The name of the pipeline variable to use as the value

variable_type: abacusai.api_class.enums.PythonFunctionArgumentType
name: str
is_required: bool
value: Any
pipeline_variable: str
class abacusai.api_class.OutputVariableMapping

Bases: abacusai.api_class.abstract.ApiClass

A config class for python function arguments

Parameters:
variable_type: abacusai.api_class.enums.PythonFunctionOutputArgumentType
name: str
class abacusai.api_class.ApiClass

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

_upper_snake_case_keys: bool
_support_kwargs: bool
__post_init__()
classmethod _get_builder()
__str__()

Return str(self).

_repr_html_()
__getitem__(item)
Parameters:

item (str) –

__setitem__(item, value)
Parameters:
  • item (str) –

  • value (Any) –

_unset_item(item)
Parameters:

item (str) –

get(item, default=None)
Parameters:
  • item (str) –

  • default (Any) –

pop(item, default=NotImplemented)
Parameters:
  • item (str) –

  • default (Any) –

to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

classmethod from_dict(input_dict)
Parameters:

input_dict (dict) –

class abacusai.api_class._ApiClassFactory

Bases: abc.ABC

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key
config_class_map
classmethod from_dict(config)
Parameters:

config (dict) –

Return type:

ApiClass

class abacusai.api_class.FeatureGroupExportConfig

Bases: abacusai.api_class.abstract.ApiClass

Helper class that provides a standard way to create an ABC using inheritance.

connector_type: abacusai.api_class.enums.ConnectorType
classmethod _get_builder()
class abacusai.api_class.FileConnectorExportConfig

Bases: FeatureGroupExportConfig

Helper class that provides a standard way to create an ABC using inheritance.

connector_type: abacusai.api_class.enums.ConnectorType
location: str
export_file_format: str
to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

class abacusai.api_class.DatabaseConnectorExportConfig

Bases: FeatureGroupExportConfig

Helper class that provides a standard way to create an ABC using inheritance.

connector_type: abacusai.api_class.enums.ConnectorType
database_connector_id: str
mode: str
object_name: str
id_column: str
additional_id_columns: List[str]
data_columns: Dict[str, str]
to_dict()

Standardizes converting an ApiClass to dictionary. Keys of response dictionary are converted to camel case. This also validates the fields ( type, value, etc ) received in the dictionary.

class abacusai.api_class._FeatureGroupExportConfigFactory

Bases: abacusai.api_class.abstract._ApiClassFactory

Helper class that provides a standard way to create an ABC using inheritance.

config_abstract_class
config_class_key = 'connectorType'
config_class_map