Building Production-Grade spaCy Text Classification Pipelines for Business Data
Unlock the full potential of spaCy with this guide to building production-grade text classification pipelines for business data.
Biology is a vast subject with dozens of subdisciplines. Some of them like computational biology, bioinformatics, and systems biology have pioneered the processing of biological data using machine learning approaches, data analysis methods, large-scale data mining, predictive models, Bayesian models, and computer science algorithms like genetic algorithms.
Let's talk about six cutting-edge uses of machine learning in biology.
Scientists saved millions of lives by creating novel vaccines and drugs in record time. Machine learning played a prominent role in many aspects of the COVID-19 response — vaccine design, drug discovery, and disease modeling.
Scientists had to find the part of the virus that was most likely to stimulate an immune response. The virus contained RNA genetic material surrounded by a sheath of spike proteins. The spike proteins let the virus bind to receptors on our cells. So, most vaccines targeted them. If the vaccine created antibodies that neutralized the spike proteins, the virus could not attack cells.
A Stanford University research team was one of the first to study vulnerable spots on the virus. They used two machine learning models named NetMHCpan and MARIA.
NetMHCpan was a simple feed-forward neural network. It had a single hidden layer that predicted receptor and protein interactions. Its inputs were the spike protein amino acid sequences and cell surface MHC protein sequences. Its outputs were the probability that they'd bind.
MARIA was a recurrent neural network model for predicting cancer immunity responses. It was repurposed for COVID-19.
Meanwhile, a DeepMind team used AlphaFold — a protein-folding neural network that we’ll meet again later — to predict the 3D structures of the SARS-CoV-2 virus’s proteins from its genetic sequence. Understanding protein structures was another key step in designing vaccines.
Most smartphones today support hardware-accelerated neural network training and inference. Machine learning software like TensorFlow and Core ML can run neural networks on smartphones. Model quantization has enabled complex models to fit a smartphone's memory. Artificial intelligence at the edge is finally a reality.
Companies like Butterfly Network and Livongo are running deep neural networks on smartphones for innovative medical diagnostics.
For example, Butterfly sells iQ+, a smart portable ultrasound device for consumers. It connects to a person's smartphone to display ultrasound imagery in real-time. They can monitor their womb, heart, lungs, bladder, and other organs. The app measures and diagnoses various health conditions using neural networks.
Real-time images or raw data from a portable device or camera are processed by one or more deep neural networks. A convolutional neural network (CNN) is particularly good at detecting the visual features that are characteristic of each health condition. What happens next depends on the purpose of the network. A classifier network tells the person the nature of the condition it found. A detection or segmentation network reports the location, shape, volume, or pixels of a condition.
Sometimes, a condition can only be detected by analyzing the changes in images or other measurements over time. A recurrent neural network or transformer network is good at detecting such sequence changes.
But machine learning and deep learning are capable of a lot more, such as correlating across different kinds of data. For example, a deep learning model can be trained to associate ultrasound imagery features with clinical information extracted from electronic or paper health records using intelligent document processing.
Such a model can be deployed to a patient's or doctor's smartphone to assist them by suggesting clinical diagnoses based on ultrasound images.Conversely, if the model notices a diagnosis in a new patient's health records, it verifies that their ultrasound matches the expected image and alerts the doctor if there's a mismatch.
If you are a biotech scientist or professional with an innovative device idea, this is a great time to commercialize it. Machine learning has enabled device-side intelligent processing of images, videos, text, and speech. A problem that you assumed is too difficult may actually be easily solved now by hiring the right machine learning expert.
Drug discovery used to be a resource-intensive, time-consuming, trial-and-error process. Biochemists had to manually apply thousands of chemicals to infected tissue samples under controlled conditions. After a few days, the cultures had to be examined one by one manually for desired effects. Automated examination was a difficult problem due to the variety of images.
Advances in computer vision and deep learning over the last decade have intelligently automated, accelerated, and scaled this process by an order of magnitude.
For example, consider a single gene disorder like sickle cell anemia. It's the effect of an abnormal protein due to a mutated gene. Proteins are essential to the normal functioning of our organs and cells. They function as enzymes, hormones, antibodies, and transport at the cellular level.
Genes, DNA, and proteins are closely related.
The nucleic acids, DNA and RNA, are long polymers made up of nucleotides. Each nucleotide consists of a nucleobase molecule represented by its first letter A, G, C, T, or U. This is why we describe DNA using DNA sequences of these letters. Human DNA consists of around 6 billion nucleotide pairs in total packed into 46 chromosomes of different DNA lengths.
Unlike chromosomes and DNA, a gene is not another physical thing. It’s merely an abstract label for a particular segment of DNA. A gene refers to a DNA segment that has been experimentally determined to encode for a protein. Each protein is a long molecule made up of some combination of the 20 amino acids. Each group of three nucleotides in a gene represents one of these 20 amino acids.
In this way, a gene acts as a template to build a particular protein. The process — involving RNA transcription and translation — is called gene expression. A mutated gene means its nucleotide sequence is abnormal. That results in an abnormal amino acid sequence and hence an abnormal protein.
Drug discovery aims to find a chemical that disables the production of an abnormal protein by disabling the expression of its mutated gene. It can be scaled up by using robots to simultaneously apply thousands of drugs to samples. The samples are monitored using automated fluorescent microscopy.
The microscope images of these thousands of samples are analyzed by convolutional neural networks trained on large datasets of such images. They look for minor visual changes in shape, morphology, area, counts, and other such traits collectively called the sample's phenotype. The neural networks classify the samples as diseased or treated or segment out desired components for further processing using segmentation neural networks like U-net.
In this way, companies like Recursion are using machine learning and deep learning to speed up drug discovery by a factor of at least 10x. Classical computer vision techniques like thresholding and watershed segmentation that involved manually adjusted parameter values for each type of cell are no longer necessary. Moreover, deep learning does not suffer from the manual feature selection and accuracy problems that hobbled classical pattern recognition methods like support vector machines, decision trees, and random forests.
Cancer therapy is often a combination of approaches involving chemotherapy, immunotherapy, and others. However, individual responses to therapies can differ widely. A treatment approach that worked for one person or one type of cancer may not work for another person or a variant of the same cancer. Such data is said to have high variability.
Machine learning is helping design personalized cancer therapies instead.
Notable Labs is one such company designing them using automated drug sensitivity screening. They receive blood and bone marrow samples from partner hospitals and apply a large number of FDA-approved drugs and drug combinations to the samples. The samples are observed using high-throughput flow cytometry that involves illuminating cells with multiple lasers.
White blood cell counts and phenotype traits like cell surface, cell membrane integrity, and morphology are measured by convolutional neural networks. The observed cell changes are correlated against the drugs and clinical observations using hierarchical clustering, principal component analysis, and analysis of variance.
They enable the company to detect drugs that showed maximum sensitivity and efficacy on that particular patient's cancer cells. The patient's treatment is then started using that drug or drug combination. Their outcome analysis showed regression of cancer in many patients and proved the effectiveness of the approach.
Reverie Labs is another company working on automated drug discovery for cancers. Their platform is capable of generating potential drug molecules based on desired molecular properties. They use generative techniques like genetic algorithms and generative neural networks to come up with new drug ideas.
Drugs are large molecules consisting of a number of smaller molecules held together by chemical bonds. Graphs are a natural way of modeling them — atoms are nodes and bonds are edges. Various graph neural networks can then be used to model and predict their molecular properties and behaviors.
Graph convolutional networks were popular for this kind of molecular graph processing till 2019. However, just like regular convolutional networks, they are good at detecting and aggregating local features but not higher-order graph connections and properties. So, graph transformer networks are becoming popular in the field. Thanks to their self-attention mechanism, they can detect relationships over larger areas of the graph while paying attention to the surrounding chemical context of each atom.
These techniques are applicable to any disease or disorder. If you are a biotech scientist or a medical specialist who has novel therapeutic ideas, we urge you to discuss your ideas with a machine learning expert. You may be surprised just how many difficult problems have become much easier now.
A genome is the set of all genes. Genomics is the study of health and other impacts of the genome.
Gene expression is the process of creating proteins from genetic instructions. Gene expression results in phenotype traits that affect everything from cells to body structure.
Thanks to quicker and cheaper next-generation sequencing methods, genome datasets and genetic signatures are increasingly being used in clinical decisions.
However, gene expression data structure cannot yet be exploited by neural networks. Worse, it's very high-dimensional data which makes any kind of data analysis noisy. So, the focus has been on reducing its dimensionality.
Machine learning techniques like autoencoders and variational autoencoders are used to reduce dimensionality. These unsupervised learning approaches have middle layers that create low-dimensional representations of gene expression data. These representations — also known as embeddings — can be used for prediction and classification tasks more efficiently.
Proteins perform a critical role at the cellular and higher levels. They act as enzymes, antibodies, hormones, and much more. The set of all proteins expressed in a body is its proteome. Proteomics is the study of the proteome.
Proteins are large polymers made up of long chains of the 20 amino acids. But it's not merely the protein sequence that affects its behavior. The entire 3D structure and its orientation also affect its behavior. Protein folding is the process where an amino acid chain attains a natural folded 3D structure to become a protein capable of performing its biological functions correctly.
Since proteins function as antibodies and enzymes, it's often necessary to engineer them to fight diseases and disorders.
However, predicting how an amino acid chain folds had been an unsolved problem for many years until 2020 when DeepMind announced their AlphaFold neural network architecture. It was able to predict protein folding with high accuracy.
AlphaFold uses two concepts to achieve this. First, contact between amino acid residues — the part of an amino acid molecule that makes it unique from other amino acids — is a good predictor of the protein structure. Second, defining a potential energy metric for this contact and optimizing it using machine learning methods predicts the final structure accurately.
AlphaFold uses a convolutional neural network to predict the distance between pairs of amino acid residues of an input sequence. These distances are used to formulate a potential function. Minimizing this potential function using gradient descent results in a protein structure that has the lowest energy and matches real-life protein structures.
Since graph convolutional networks and graph transformer networks have proved effective in other bioinformatics and modeling tasks, we can perhaps anticipate their use in improved versions of AlphaFold in the future.
For decades, the processing of images, videos, audio, and text were considered difficult barriers. Many innovations in biology and life sciences were never undertaken because the scientists and researchers who thought them up did not know how to overcome these problems.
But deep learning running on inexpensive cloud and smartphone hardware has dismantled these barriers. The data engineering aspects of biological innovations are a solved problem today. You scientists, researchers, and founders can focus on biology while letting data experts like us help you out with the rest. Contact us with your innovation idea and let's bring it to life together.
Unlock the full potential of spaCy with this guide to building production-grade text classification pipelines for business data.
We compare 12 AI text summarization models through a series of tests to see how BART text summarization holds up against GPT-3, PEGASUS, and more.
Let’s take a look at what intent classification is in conversational ai and how you can build a GPT-3 intent classification model for conversational ai and chatbot pipelines.
Discover the capabilities of zero-shot object detection, which enables anyone to use a model out-of-the-box without any training and generate production-grade results.
What is facial expression recognition and what SOTA models are being used today in production
Get a simple TensorFlow facial recognition model up & running quickly with this tutorial aimed at using it in your personal spaces on smartphones & IoT devices.
Explore accurate classification algorithms using the latest innovations in deep learning, computer vision, and natural language processing.
Learn what human activity recognition means, how it works, and how it’s implemented in various industries using the latest advances in artificial intelligence.
What is the the SetFit architecture and how does it outperform GPT-3 and other few shot large language models
What is image classification and how we build production level TensorFlow image classification systems for recognizing various products on a retail shelf.
Explore the application of intelligent document processing (IDP) in different industries and dive in-depth on intelligent document pipelines.
How to build an image classification model in PyTorch with a real world use case. How you can perform product recognition with image classification
Let's build a custom CTA generator that you'll actually want to use for your website copy
We’re going to look at how we built a state of the art NLP pipeline for blended summarization and NER to process master service agreements (MDAs) that vary the outputs based on the input document and what is deemed important information.
Get a comprehensive overview of a purchase order vs. invoice, including when businesses use each, what information goes in them, and more.
Learn what Google Shopping categories are used for and how you can automate fitting products to this taxonomy using ai.
Automatically categorize your Shopify store products to the Shopify Product Taxonomy instantly with ai based PIM software
Dive deep into 3-way invoice matching, including how it works, eight benefits for your business, and the problems with doing it manually.
Smart farming using computer vision and deep learning provides the most promising path forward in the slow-moving industry of agriculture.
How we leveraged large language models to build a legal clause rewriting pipeline that generates stronger language and more clarity in legal clauses
Using ai for document information extraction to automate various parts of the loan process.
Apply AI to your favorite sport with this guide. Learn how automated ball tracking can change the game for coaches and players.
Categorize your ecommerce products to the 2021 google product taxonomy tree instantly with our Ai software
Surveying the current landscape of ecommerce automation and how you can use ai to automate huge chunks of your product management.
Classify your product data against an existing product category database or generate categories and tags in seconds using artificial intelligence
Warehouse automation plays a crucial role across your supply chain. Learn about how machine learning and ai software can be integrated into your warehouse automation stack.
4 different NLP methods of summarizing longer input text into different methods such as extractive, abstractive, and blended summarization
iscover an invoice OCR tool that will revolutionize the way you handle invoices. There’s no human intervention needed & a dramatically lower per-invoice cost.
Instead of invoice matching taking upwards of a week, it could take mere seconds with the proper automation solution. Learn more here.
Manual and template-based invoicing are riddled with low accuracy and required human intervention. Learn how to systematically eliminate these issues with the right invoice data capture software.
A complete walkthrough guide on how to use visual search in ecommerce stores to create more sales and real examples of companies already using it.
Automating the extraction of data from invoices can reduce the stress of your accountants by finding inaccuracies, digitizing paper invoices, and more.
How you can use machine learning based data matching to compare data features in a scalable architecture for deduping, record merging, and operational efficiency
Learn how lifetime value or LTV prediction can improve your marketing strategies. Then, discover the best statistical & machine learning models for your predictions.
A deep understanding of how we use gpt-3 and other NLP processes to build flexible chatbot architectures that can handle negotiation, multiple conversation turns, and multiple sales tactics to increase conversions.
The popular HR company O.C. Tanner, which has been in business since 1927 and has over 1500 employees, was looking to research and design two GPT-3 software products to be used as internal tools with their clients. GPT-3 based products can be difficult to outline and design given the sheer lack of publicly available information around optimizing and improving these systems to a production level.
We’ll compare Tableau vs QlikView in terms of popularity, integrations, ease of use, performance, security, customization, and more.
With a context-aware recommender system, you can plan ways to recreate some of the contextual conditions that persuade them to buy more from you.
We’re going to walk through building a production level twitter sentiment analysis classifier using GPT-3 with the popular tweet dataset Sentiment140.
Find out how machine learning in medical imaging is transforming the healthcare world and making it more efficient with three use cases.
Discover ways that machine learning in health care informatics has become indispensable. Review the results of two case studies and consider two key challenges.
Accelerate your growth by pivoting key areas of your business to AI. Your business outcomes will be achieved quicker & you’ll see benefits you didn’t plan for.
We built a GPT-3 based software solution to automate raw data processing and data classification. Our model handles keyword extraction, named entity recognition, text classification | Case Study
We built a custom GPT-3 pipeline for key topic extraction for an asset management company that can be used across the financial domain | Case Study
How you can use GPT-3 to create higher order product categorization and product tagging from your ecommerce listings, and how you can create a powerful product taxonomy system with ai.
5 ways you can use product matching software in ecommerce to create real value that raises your sales metrics and improves your workflow operations.
Data mining and machine learning in cybersecurity enable businesses to ensure an acceptable level of data security 24/7 in highly dynamic IT environments. Learn how data security is getting increasingly automated.
Product recognition software has tremendous potential to improve your profits and slash your costs in your retail business. Find out just how useful it is.
Big data has evolved from hype to a crucial part of scaling your organization in every modern industry. Learn more about how big data is transforming organizations and providing business impacts.
Learn how natural language processing can benefit everybody involved in education from individual students and teachers to entire universities and mass testing agencies.
Here’s how automated data capture systems can benefit your business in some key ways and some real-life examples of what it looks like in practice.
Use these power ai and machine learning tools to create business intelligence in your marketing that pushes your business understanding and analytics past your competition.
We built a custom ML pipeline to automate information extraction and fine tuned it for the legal document domain.
In this practical guide, you'll get to know the principles, architectures, and technologies used for building a data lake implementation.
An enterprise data warehouse (EDW) is a repository of big data for an enterprise. It’s almost exclusive to business and houses a very specific type of data.
Dlib is a versatile and well-diffused facial recognition library, with perhaps an ideal balance of resource usage, accuracy and latency, suited for real-time face recognition in mobile app development. It's becoming a common and possibly even essential library in the facial recognition landscape, and, even in the face of more recent contenders, is a strong candidate for your computer vision and facial recognition or detection framework.
Learn how to utilize machine learning to get a higher customer retention rate with this step-by-step guide to a churn prediction model.
Machine learning algorithms are helping the oil and gas industry cut costs and improve efficiency. We'll show you how.
We’ll show you the difference between machine learning vs. data mining so you know how to implement them in your organization.
Here’s why you should use deep learning algorithms in your business, along with some real-world examples to help you see the potential.
Beam search is an algorithm used in many NLP and speech recognition models as a final decision making layer to choose the best output given target variables like maximum probability or next output character.
Best Place For was looking for an image recognition based software solution that could be used to detect and identify different food dishes, drinks, and menu items in images sourced from blogs and Instagram. The images would be pulled from restaurant locations on Instagram and different menu items would be identified in the images. This software solution has to be able to handle high and low quality images and still perform at the highest production level, while accounting for runtime as well as accuracy.
Deep learning recommendation system architectures make use of multiple simpler approaches in order to remediate the shortcomings of any single approach to extracting, transforming and vectorizing a large corpus of data into a useful recommendation for an end user.
Let's take a look at the architecture used to build neural collaborative filtering algorithms for recommendation systems
GPT-3 is one of the most versatile and transformative components that you can include in your framework, application or service. However, sensational headlines have obscured its wide range of capabilities since its launch. Let’s take a look at the ways that companies and researchers are achieving real-world results with GPT-3, and examine the untapped potential of this 'celebrity AI'.
How to get started with machine learning based dynamic pricing algorithms for price optimization and revenue management
Let's take a look at how you can use spaCy, a state of the art natural language processing tool, to build custom software tools for your business that increase ROI and give you data insights your competitors wish they had.
The landscape for AI in ecommerce has changed a lot recently. Some of the most popular products and approaches have been compromised or undermined in a very short time by a new global impetus for privacy reform, and by the way that the COVID-19 pandemic has transformed the nature of retail.
Extremely High ROI Computer Vision Applications Examples Across Different Industries
Building Data Capture Services To Collect High ROI Business Data With Machine Learning and AI
Software packages and Inventory Data tools that you definitely need for all automated warehouse solutions
Inventory automation with computer vision - how to use computer vision in online retail to automate backend inventory processes