was successfully added to your cart.

    Benchmarking information for NER and Classification Models in Annotation Lab 4.4.0

    Avatar photo
    Ph.D. in Computer Science – Head of Product

    A new generation of the NLP Lab is now available: the Generative AI Lab. Check details here https://www.johnsnowlabs.com/nlp-lab/

    In the artificial intelligence (AI) domain, benchmarking a model is the process of evaluating the performance (accuracy, precision, recall, F1 score, etc.) of that model by comparing it to a standard or baseline. This can be done by using established datasets and metrics to measure the accuracy, speed, and other characteristics of the model. The results of benchmarking can be used to compare different models and to identify the strengths and weaknesses of each one. It can also be used to identify areas where a model may need further optimization or improvement.

    There are many dimensions that can be used to evaluate a deep learning (DL) model. Some common ones include:

    1. Accuracy: This is the most common metric used to evaluate a DL model. It measures the percentage of predictions made by the model that are correct.
    2. Precision: This measures the proportion of positive predictions that are actually correct.
    3. Recall: This measures the proportion of actual positive cases that are correctly predicted by the model.
    4. F1 score: This is a combination of precision and recall, and is calculated as the harmonic mean of the two.

    Annotation Lab 4.4.0 supports model benchmarking in two ways:

    • First of all, it is possible to quickly view benchmarking data obtained for the pre-trained models by John Snow Labs team of data scientists and for the models trained within the Annotation Lab. This information is available on the Models page of the Hub, by navigating to the model of interest and by clicking on the benchmarking icon. Read more here.
    • Second, it is possible to test the pre-trained models on your custom data by creating a project, adding some tasks, annotating them, and tagging them as Test. Then navigate to the new project, and from the Train page, click on the Test Configuration button. Read more here.

    Training logs with confusion matrices for NER models were already available in Annotation Lab since release 4.2.0, which made it easier to understand the performance of the model and judge whether the model is underfitting or overfitting. Release 4.4.0 adds the same features for classification models. Those can be accessed on the training logs as well as by clicking on the benchmarking icon available on the model’s tiles in the Models tab of the Hub page.

    Dashboard of Annotation Lab presenting the Annotation Lab Project

    Get & Install It Here.

    Full Feature Set Here.

    How useful was this post?

    Try The Generative AI Lab - No-Code Platform For Model Tuning & Validation

    See in action
    Avatar photo
    Ph.D. in Computer Science – Head of Product
    Our additional expert:
    Dia Trambitas is a computer scientist with a rich background in Natural Language Processing. She has a Ph.D. in Semantic Web from the University of Grenoble, France, where she worked on ways of describing spatial and temporal data using OWL ontologies and reasoning based on semantic annotations. She then changed her interest to text processing and data extraction from unstructured documents, a subject she has been working on for the last 10 years. She has a rich experience working with different annotation tools and leading document classification and NER extraction projects in verticals such as Finance, Investment, Banking, and Healthcare.

    No-Code Visual Entity Recognition in the Annotation Lab

    A new generation of the NLP Lab is now available: the Generative AI Lab. Check details here https://www.johnsnowlabs.com/nlp-lab/ Annotation Lab has improved...
    preloader