Source code for pyabsa.tasks.TextAdversarialDefense.configuration.tad_configuration

# -*- coding: utf-8 -*-
# file: tad_configuration.py
# time: 02/11/2022 19:56
# author: YANG, HENG <hy345@exeter.ac.uk> (杨恒)
# github: https://github.com/yangheng95
# GScholar: https://scholar.google.com/citations?user=NPq5a_0AAAAJ&hl=en
# ResearchGate: https://www.researchgate.net/profile/Heng-Yang-17/research
# Copyright (C) 2022. All Rights Reserved.

import copy

# if you find the optimal param set of some situation, e.g., some model on some datasets
# please share the main use template main
from pyabsa.framework.configuration_class.configuration_template import ConfigManager
from ..models.__classic__.tad_lstm import TADLSTM
from ..models.__plm__.tad_bert import TADBERT

[docs] _tad_config_template = { "model": TADBERT, "optimizer": "adamw", "learning_rate": 0.00002, "patience": 99999, "cache_dataset": True, "warmup_step": -1, "show_metric": False, "max_seq_len": 80, "dropout": 0, "l2reg": 0.000001, "num_epoch": 10, "batch_size": 16, "initializer": "xavier_uniform_", "seed": 52, "output_dim": 3, "log_step": 10, "evaluate_begin": 0, "cross_validate_fold": -1, "use_amp": False, "overwrite_cache": False, # split train and test datasets into 5 folds and repeat 3 trainer }
[docs] _tad_config_base = { "model": TADBERT, "optimizer": "adamw", "learning_rate": 0.00002, "pretrained_bert": "microsoft/deberta-v3-base", "cache_dataset": True, "warmup_step": -1, "show_metric": False, "max_seq_len": 80, "patience": 99999, "dropout": 0, "l2reg": 0.000001, "num_epoch": 10, "batch_size": 16, "initializer": "xavier_uniform_", "seed": 52, "output_dim": 3, "log_step": 10, "evaluate_begin": 0, "cross_validate_fold": -1, "use_amp": False, # split train and test datasets into 5 folds and repeat 3 trainer }
[docs] _tad_config_english = { "model": TADBERT, "optimizer": "adamw", "learning_rate": 0.00002, "patience": 99999, "pretrained_bert": "microsoft/deberta-v3-base", "cache_dataset": True, "warmup_step": -1, "show_metric": False, "max_seq_len": 80, "dropout": 0, "l2reg": 0.000001, "num_epoch": 10, "batch_size": 16, "initializer": "xavier_uniform_", "seed": 52, "output_dim": 3, "log_step": 10, "evaluate_begin": 0, "cross_validate_fold": -1, "use_amp": False, # split train and test datasets into 5 folds and repeat 3 trainer }
[docs] _tad_config_multilingual = { "model": TADBERT, "optimizer": "adamw", "learning_rate": 0.00002, "patience": 99999, "pretrained_bert": "microsoft/mdeberta-v3-base", "cache_dataset": True, "warmup_step": -1, "show_metric": False, "max_seq_len": 80, "dropout": 0, "l2reg": 0.000001, "num_epoch": 10, "batch_size": 16, "initializer": "xavier_uniform_", "seed": 52, "output_dim": 3, "log_step": 10, "evaluate_begin": 0, "cross_validate_fold": -1, "use_amp": False, # split train and test datasets into 5 folds and repeat 3 trainer }
[docs] _tad_config_chinese = { "model": TADBERT, "optimizer": "adamw", "learning_rate": 0.00002, "patience": 99999, "cache_dataset": True, "warmup_step": -1, "show_metric": False, "pretrained_bert": "bert-base-chinese", "max_seq_len": 80, "dropout": 0, "l2reg": 0.000001, "num_epoch": 10, "batch_size": 16, "initializer": "xavier_uniform_", "seed": 52, "output_dim": 3, "log_step": 10, "evaluate_begin": 0, "cross_validate_fold": -1, "use_amp": False, # split train and test datasets into 5 folds and repeat 3 trainer }
[docs] _tad_config_glove = { "model": TADLSTM, "optimizer": "adamw", "learning_rate": 0.001, "cache_dataset": True, "warmup_step": -1, "show_metric": False, "max_seq_len": 100, "patience": 20, "dropout": 0.1, "l2reg": 0.000001, "num_epoch": 100, "batch_size": 64, "initializer": "xavier_uniform_", "seed": 52, "embed_dim": 300, "hidden_dim": 300, "output_dim": 3, "log_step": 5, "hops": 3, # valid in MemNet and RAM only "evaluate_begin": 0, "cross_validate_fold": -1, "use_amp": False, }
[docs] class TADConfigManager(ConfigManager): def __init__(self, args, **kwargs): """ Available Params: {'model': BERT, 'optimizer': "adamw", 'learning_rate': 0.00002, 'pretrained_bert': "roberta-base", 'cache_dataset': True, 'warmup_step': -1, 'show_metric': False, 'max_seq_len': 80, 'patience': 99999, 'dropout': 0, 'l2reg': 0.000001, 'num_epoch': 10, 'batch_size': 16, 'initializer': 'xavier_uniform_', 'seed': {52, 25} 'embed_dim': 768, 'hidden_dim': 768, 'output_dim': 3, 'log_step': 10, 'evaluate_begin': 0, 'cross_validate_fold': -1 # split train and test datasets into 5 folds and repeat 3 trainer } :param args: :param kwargs: """ super().__init__(args, **kwargs) @staticmethod
[docs] def set_tad_config(configType: str, newitem: dict): if isinstance(newitem, dict): if configType == "template": _tad_config_template.update(newitem) elif configType == "base": _tad_config_base.update(newitem) elif configType == "english": _tad_config_english.update(newitem) elif configType == "chinese": _tad_config_chinese.update(newitem) elif configType == "multilingual": _tad_config_multilingual.update(newitem) elif configType == "glove": _tad_config_glove.update(newitem) else: raise ValueError( "Wrong value of configuration_class type supplied, please use one from following type: template, base, english, chinese, multilingual, glove" ) else: raise TypeError( "Wrong type of new configuration_class item supplied, please use dict e.g.{'NewConfig': NewValue}" )
@staticmethod
[docs] def set_tad_config_template(newitem): TADConfigManager.set_tad_config("template", newitem)
@staticmethod
[docs] def set_tad_config_base(newitem): TADConfigManager.set_tad_config("base", newitem)
@staticmethod
[docs] def set_tad_config_english(newitem): TADConfigManager.set_tad_config("english", newitem)
@staticmethod
[docs] def set_tad_config_chinese(newitem): TADConfigManager.set_tad_config("chinese", newitem)
@staticmethod
[docs] def set_tad_config_multilingual(newitem): TADConfigManager.set_tad_config("multilingual", newitem)
@staticmethod
[docs] def set_tad_config_glove(newitem): TADConfigManager.set_tad_config("glove", newitem)
@staticmethod
[docs] def get_tad_config_template(): _tad_config_template.update(_tad_config_template) return TADConfigManager(copy.deepcopy(_tad_config_template))
@staticmethod
[docs] def get_tad_config_base(): _tad_config_template.update(_tad_config_base) return TADConfigManager(copy.deepcopy(_tad_config_template))
@staticmethod
[docs] def get_tad_config_english(): _tad_config_template.update(_tad_config_english) return TADConfigManager(copy.deepcopy(_tad_config_template))
@staticmethod
[docs] def get_tad_config_chinese(): _tad_config_template.update(_tad_config_chinese) return TADConfigManager(copy.deepcopy(_tad_config_template))
@staticmethod
[docs] def get_tad_config_multilingual(): _tad_config_template.update(_tad_config_multilingual) return TADConfigManager(copy.deepcopy(_tad_config_template))
@staticmethod
[docs] def get_tad_config_glove(): _tad_config_template.update(_tad_config_glove) return TADConfigManager(copy.deepcopy(_tad_config_template))