Hi Hacker News! We’re Nurlybek and Michael, the cofounders of Biodock (
http://www.biodock.ai/). We help scientists expedite microscopy image analysis.
Michael and I built Biodock due to the challenges we experienced in microscopy image analysis while we were at Stanford. As a Ph.D. student, I spent hours manually counting through lipid droplets in microscope images of embryonic tissues. The incredible frustration I felt led me to try all kinds of software. Eventually, I went out to seek help from other scientists. Michael, a computer science student, was working in a lab just across from mine when he got my email asking for help. We got to chatting in a med school cafe and realized that we were both tackling the same issues with microscopy images.
Microscopy images are one of the most fundamental forms of data in biomedical research, from discovery all the way to clinical trials. They can be used to show the expression of genes, the progression of the disease, and the efficacy of treatments.
However, images are also very frustrating, and we think a lot of that has to do with the current tools available. To analyze their images, many scientists at top research institutions use software techniques invented 50 years ago, like thresholding and filtering. Some even spend their days manually drawing regions around cells or regions. Not only is this extremely frustrating, but it slows down the research cycle, meaning that it takes a lot more time and money to create potentially lifesaving cures. Contrast these tools to the incredible recent headway into deep learning - where applications like AlphaFold have led to incredible gains in what was previously possible.
Our goal is to bring these performance gains to research scientists. The current core module in Biodock is AI cell segmentation for fluorescent cells, based mostly on Mask R-CNN and U-Net architectures, and trained on thousands of cell images. Essentially, it identifies where each cell is and calculates important features like location, size, and fluorescent expression for each cell. This module performs around 40% more accurately than other software.
So how is this different from training deep learning models yourself? First, our pretrained modules are trained on a huge amount of data, which allows for great performance for all scientists without needing to label data or optimize training. Secondly, we’ve spent time carefully building our cloud architecture and algorithms for production, including a large cluster of GPUs. We even slice images into crops, process them in parallel, and stitch them together. We also have storage, data integrations, and visualizations built into the platform.
We know that AI cell segmentation addresses only a small fraction of microscopy analysis in the biomedical space, and we are launching several more modules soon, tackling some of the most difficult images in the space. So far, we’ve been able to generate different custom AI modules for diverse tissues and imaging modalities (fluorescence, brightfield, electron microscopy, histology). Eventually, we want to link other biological data analyses into the cloud including DNA sequences, proteomics, and flow cytometry, to power the 500K scientists and 3K companies in the US biotech and pharma space.
We would love to hear from you and get your feedback—especially if you've ever spent hours on image analysis!
I do wonder, though, about the wisdom of doing that sort of analysis in the cloud. Our projects routinely use several terabytes of images (we have about 150TB stored right now, most of which is full-slide images), and uploading them somewhere isn't just a simple fire-and-forget procedure. Cool analysis algorithms might not be enough to make up for the headache of having to wait for days on end for the uploads to reach the cloud.