【问题标题】:how to use cross-validation with ktrain?如何使用 ktrain 进行交叉验证?
【发布时间】:2021-09-30 17:23:24
【问题描述】:

我正在使用ktrain 包来执行多类文本分类。官方ktrain 网站上的示例效果很好(https://github.com/amaiya/ktrain

categories = ['alt.atheism', 'soc.religion.christian','comp.graphics', 'sci.med']
from sklearn.datasets import fetch_20newsgroups
train_b = fetch_20newsgroups(subset='train', categories=categories, shuffle=True)
test_b = fetch_20newsgroups(subset='test',categories=categories, shuffle=True)
(x_train, y_train) = (train_b.data, train_b.target)
(x_test, y_test) = (test_b.data, test_b.target)

# build, train, and validate model (Transformer is wrapper around transformers library)
import ktrain
from ktrain import text
MODEL_NAME = 'distilbert-base-uncased'
t = text.Transformer(MODEL_NAME, maxlen=500, class_names=train_b.target_names)
trn = t.preprocess_train(x_train, y_train)
val = t.preprocess_test(x_test, y_test)
model = t.get_classifier()
learner = ktrain.get_learner(model, train_data=trn, val_data=val, batch_size=6)
learner.fit_onecycle(5e-5, 4)
learner.validate(class_names=t.get_classes())

准确度很高。

但是,我将此模型与其他使用 scikit-learn 训练的模型进行比较,特别是其他模型的准确性是使用交叉验证评估的

cross_val_score(sgd_clf, X_train, y_train, cv=3, scoring="accuracy")

如何调整上面的代码以确保与 ktrain 一起使用的转换器模型也使用相同的交叉验证方法进行评估?

【问题讨论】:

    标签: python tensorflow scikit-learn deep-learning ktrain


    【解决方案1】:

    你可以试试这样的:

    from ktrain import text
    import ktrain
    import pandas as pd
    from sklearn.model_selection import train_test_split,KFold
    from sklearn.metrics import accuracy_score
    from sklearn.datasets import fetch_20newsgroups
    
    # load text data
    categories = ['alt.atheism', 'soc.religion.christian','comp.graphics', 'sci.med']
    train_b = fetch_20newsgroups(subset='train', categories=categories, shuffle=True)
    test_b = fetch_20newsgroups(subset='test',categories=categories, shuffle=True)
    (x_train, y_train) = (train_b.data, train_b.target)
    (x_test, y_test) = (test_b.data, test_b.target)
    df = pd.DataFrame({'text':x_train, 'target': [train_b.target_names[y] for y in y_train]})
    
    # CV with transformers
    N_FOLDS = 2
    EPOCHS = 3
    LR = 5e-5
    def transformer_cv(MODEL_NAME):
        predictions,accs=[],[]
        data = df[['text', 'target']]
        for train_index, val_index in KFold(N_FOLDS).split(data):
            preproc  = text.Transformer(MODEL_NAME, maxlen=500)
            train,val=data.iloc[train_index],data.iloc[val_index]
            x_train=train.text.values
            x_val=val.text.values
    
            y_train=train.target.values
            y_val=val.target.values
    
            trn = preproc.preprocess_train(x_train, y_train)
            model = preproc.get_classifier()
            learner = ktrain.get_learner(model, train_data=trn, batch_size=16)
            learner.fit_onecycle(LR, EPOCHS)
            predictor = ktrain.get_predictor(learner.model, preproc)
            pred=predictor.predict(x_val)
            acc=accuracy_score(y_val,pred)
            print('acc',acc)
            accs.append(acc)
        return accs
    print( transformer_cv('distilbert-base-uncased') )
    
    # output:
    # [0.9627989371124889, 0.9689716312056738]
    

    参考:有关回归示例,请参阅 this Kaggle notebook

    【讨论】:

      猜你喜欢
      • 2022-01-25
      • 2020-11-30
      • 2017-06-29
      • 2015-06-11
      • 2020-07-13
      • 2019-09-06
      • 2017-04-21
      • 1970-01-01
      • 1970-01-01
      相关资源
      最近更新 更多