We are happy to announce the release of Generative AI Lab, marking the transition from the previous NLP Lab to a state-of-the-art No-Code platform that enables domain experts to train task-specific AI models using large language models (LLMs). This new product introduces a suite of advanced features and functionalities designed to facilitate the creation, management, and deployment of AI-driven solutions efficiently and accurately. With robust integration capabilities, comprehensive model training tools, and enhanced security measures tailored for high-compliance sectors such as healthcare, Generative AI Lab sets a new standard in the generative AI platform landscape. John Snow Labs is committed to democratizing AI tool interaction and making it more accessible, especially within the healthcare sector. We aim to equip healthcare professionals, from clinicians to researchers, with the tools to construct bespoke AI models. These models are pivotal in analyzing extensive datasets, discerning patterns, aiding in diagnosis, and enhancing patient care, all achievable without in-depth coding expertise. This opens new avenues in personalized medicine, accelerates research, and improves patient outcomes, revolutionizing the healthcare landscape.
In this release, Generative AI Lab enhances its capabilities by integrating seamlessly with the OpenAI API, enabling the effortless training, testing, and refinement of task-specific models tailored to the distinct needs of your domain and use case. Whether your focus is refining document processing, orchestrating AI assets, or fortifying regulatory compliance, Generative AI Lab delivers a comprehensive, secure, and effective framework to transform your AI ambitions into tangible outcomes.
Generative AI Lab marks a significant leap forward in hardware architecture and performance optimization for model training and Visual Document understanding. Now available on the AWS marketplace with a GPU-enabled template, this release simplifies license provisioning and unlocks immediate access to all features, including prompts, Rules, pretrained medical models, medical resolvers, medical model training, and visual document understanding. This enhancement not only accelerates performance but also broadens your access to a plethora of AI capabilities, reaffirming our commitment to advancing the NLP community’s capabilities.
Use LLMs to bootstrap task-specific models.
Generative AI Lab facilitates seamless integration with the OpenAI API, empowering domain experts to easily define prompts for classification or entity extraction. This integration allows Generative AI Lab to process the LLM’s responses, adjust the indexes of the extracted segments, and overlay pre-annotation results directly onto the original documents.
Once pre-annotations are generated, domain experts can step in and review these results through a user-friendly interface, offering their expertise in the form of adjustments or corrections. This refined data can then be employed to develop smaller, more specialized models that are optimized for processing the document of interest.
Furthermore, Generative AI Lab supports comprehensive training experiments and provides access to benchmarking data to evaluate the performance during the model training process. Users can augment the training dataset with additional examples and reinitiate the training process for continuous model enhancement, ensuring sustained improvement and adaptation.
The final step is iterative refinement. Here, users can assess the model’s performance metrics and introduce more data as needed. This process ensures that the model can be adapted and improved as a response to new information and evolving requirements in the healthcare domain.
Private, on-premise, high-compliance prompt engineering
Protecting Personal Health Information (PHI) is paramount in the healthcare sector. To this end, Generative AI Lab supports Zero-Shot models that can process PHI directly within your infrastructure, thus ensuring privacy and compliance. Mirroring the workflow used for integrating LLMs via external APIs, you can now utilize Zero-Shot Learning for the pre-annotation of your documents. This functionality is available directly within the Generative AI Lab, eliminating the need for external API calls to LLMs.
By adopting this approach, you can ensure adherence to stringent healthcare regulations, providing reassurance and peace of mind.
Organize and share models, prompts, and rules within one private enterprise hub.
The Models Hub acts as a centralized platform where users from your organization can easily manage their AI development lifecycle. It supports operations like the secure sharing, searching, filtering, testing, publishing, importing, and exporting of AI models, prompts, and rules. This functionality simplifies the management of proprietary AI assets, enabling teams to efficiently collaborate and leverage these assets for their projects.
The Models Hub implements role-based access control (RBAC), allowing you to define who in your organization has access to your assets, who can experiment with prompts or rules, or who can export your models. Versioning and backup features are available to keep a record of changes made to your assets, ensuring that you can always revert to previous versions if needed.
Finally, the playground allows easy editing and testing of prompts, rules, or models without coding.
Generative AI Lab is integrated with the NLP Models Hub, which gives access to an extensive library of over 40,000 models and pipelines ready to be integrated into your projects. This integration not only enhances your capabilities but also provides easy access to model benchmarking data, model documentation, and one-click downloads.
Models:
Within the Models page, you’ll find a private repository tailored to your organization’s needs, including models you’ve trained, uploaded, or downloaded from the NLP Models Hub. This centralized management system ensures your AI assets are organized and readily available.
Rules:
The Rules page offers a dedicated space for creating and managing the rules you can define and use in your projects. With an intuitive editing interface and practical examples, crafting custom rules becomes a straightforward process.
Prompts:
Lastly, the Prompts page allows you to curate a collection of prompts, essential for preannotating your documents and training your AI models. Through an easy-to-use editing and testing interface, you can ensure your prompts are effective and achieve the intended responses.
Deployment
AWS Marketplace
Generative AI lab is available on AWS Marketplace as a one-click deployment within your security parameter. This subscription offers immediate access to Visual document understanding features, including tools for Optical Character Recognition, PDF preannotations, or Visual Model Training. For healthcare professionals, the platform offers specialized resources such as embeddings and models designed and tuned for healthcare data, covering tasks like entity recognition, assertion status detection, relation extraction, or entity resolution.
And, you’re never alone in this process; professional support is always at your fingertips to assist with any questions or integrations.
On-premise:
Generative AI lab is also available for on-premise deployments on both GPU and CPU servers. To utilize GPU resources, add the case-insensitive ‘gpu’ parameter when running the installation script, which is available here. This parameter acts as a toggle: if specified, the script will activate GPU support, assuming the host server has the necessary GPU resources. If omitted, the script will proceed without configuring GPU-related settings.
Use gpu
(case-insensitive) optional parameter with annotationlab-installer.sh script to enable usage of GPU resources. This will only work if your host has GPU resources. This parameter is used as a flag, enabling GPU resources when used, otherwise, the installer will ignore anything related to GPU.
$ ./annotationlab-installer.sh gpu
Please note that you can disable GPU usage anytime by modifying the install script and setting the useGPU
variable to false. This action will stop the application from utilizing GPU resources but will not uninstall any previously installed Nvidia drivers or plugins. With this update, a GPU-based AMI is now available, allowing users to leverage the increased performance and capabilities that come with GPU acceleration.
Migrate your NLP Lab Backup to Generative AI Lab
Migrating to the new version is easy! Users using the NLP Lab can migrate their annotated data and configured settings to Generative AI Lab through our Backup and Restore feature. This process enables users to back up their projects (including data and files) from an NLP Lab server to Azure Blob or AWS S3 and then restore the configurations to a Generative AI server. For this, the following steps need to be taken:
Follow these steps to migrate your data:
- Backup Data:
- Navigate to the Backup page of your Generative AI Lab instance.
- Enter backup details.
- Schedule an immediate backup via backend modification:
kubectl edit cronjob
- Monitor the backup pod status:
kubectl get pods
- Verify Backup:
- Upon completion, your backed-up database and files will be visible in cloud storage.
- Restore Data:
- Access the backend of your target Generative AI Lab instance.
- Transfer backed-up data from cloud storage to artifacts/restore/database.
- Perform database restoration:
sudo ./restore_all_databases.sh
- Copy backed-up files from cloud storage to artifacts/restore/files.
- Execute file restoration:
sudo ./restore_files.sh
- Verify Restoration:
- Access the UI; all data and files should be restored.
Resource configuration
GPU Resource Availability
If the Generative AI Lab is equipped with a GPU, the following message will be displayed on the infrastructure page:
Visual NER Training with GPU
The Training & Active Learning page now includes a new option, “Use available GPU” for Visual NER projects. Selecting this option enables Visual NER model training using GPU.
Note: Find CPU vs GPU Benchmarks for Visual NER model training here.
Using Healthcare and Visual Document Understanding Features
The Generative AI Lab supports the PAYG (Pay-As-You-Go) license option offered by John Snow Labs for the use of pre-trained medical and visual models. This option is added to the support for floating licenses and airgap licenses and was added for enhanced flexibility, cost reduction, and the mechanism of paying only for the utilized resources.
PAYG License Features
- PAYG License Included in AMI Installation: The Generative AI Lab AWS product includes a PAYG license key generated at subscription time and readily available on the License page within the AMI environment. Users do not need to worry about manually adding the license. Therefore, concerns regarding expiration or accidental deletion are eliminated.
- BYOL for on-premise deployments: For on-premise deployments of the Generative AI Lab, users can buy a PAYG license from my.johnsnowlabs.com, download it, and import it to Generative AI Lab via the License page. Note that the deployment server needs to allow license heartbeat to be sent to johnsnowlabs services to validate license usage.
- Flexible Billing: With the PAYG license, users are billed based on only the resources they use, offering a more tailored and cost-effective pricing model.
- Support for Multiple Servers: The PAYG license also supports running multiple training and pre-annotation servers in parallel. This license enables users to deploy and utilize multiple pre-annotation servers and training instances in parallel, boosting workflow efficiency and productivity, allowing the execution of tasks simultaneously and accelerating project completion.
Cost Awareness Banner for PAYG License
With the introduction of PAYG license support, proactive measures have been taken to inform users about the potential costs associated with using licensed features. Users will now be presented with a noticeable message banner at the top of the page stating: “Continuous Server Usage Incurs Costs! Please check the deployed server.” The message is always shown even if no server is deployed on the cluster page. It helps users be aware that they are billed based on application and resource usage.
By presenting this message, users are reminded to monitor their server usage and associated costs, promoting cost-conscious behavior. This feature enhances user awareness and ensures transparency regarding the cost implications of utilizing the PAYG license within Generative AI Lab.
Bug Fixes and Improvements
Support for multiple spaces and tabs in Label Metadata
Previously, within the label metadata, multiple spaces were considered a single space, and tabs were not supported. Pressing the tab key would render the text area inactive. Now, users can utilize multiple spaces and tabs within label metadata, both of which are preserved when adding metadata to labeled texts.
Pre-annotation using the Text Classification model with an HC license is not working.
Previously, healthcare classification models such as “classifierml_ade” were not deployed in pre-annotation, even with a healthcare license. However, this issue has now been rectified.
The Zoom Feature is not working in predictions and submitted completions in the Visual NER Project.
In previous versions, the zoom-in and zoom-out functionalities for submitted completions and predictions in the Visual NER project were not functioning properly. This issue has been resolved.
Getting Started is Easy
Generative AI Lab is a text annotation tool that can be deployed in a couple of clicks on the AWS, or installed on-premise with a one-line Kubernetes script.
Get started here: https://nlp.johnsnowlabs.com/docs/en/alab/install
Start your journey with Generative AI Lab and experience the future of data analysis and model training today!
Try The Generative AI Lab - No-Code Platform For Model Tuning & Validation
See in action