Data benchmark
Overview
After defining your assays, uploading data, and setting objectives, you can run a benchmark to evaluate how well your data fits Cradle's models. This process provides insights into your data quantity, quality and how well Cradle's models fit.
What is a data benchmark?
The data benchmark results in a report with insightful metrics and graphs that indicate:
Whether the data quantity and quality is sufficient for machine learning.
The performance of Cradle's machine learning algorithms in learning from your data.
This process helps both you and Cradle determine how well Cradle's models learn from your data and whether the project is worth pursuing.
How to run a benchmark
After your data is uploaded, click on Start benchmark.
The platform will show you how much time it will take to run the benchmark.
Report & results
Once the benchmark report is ready you will receive a notification from Cradle. Log in to the platform to view your report.
Note: the benchmark report is different for each customer and tailored to your experimental data
The benchmark helps indicate the following possible results and next steps:
Good fit: The quantity and quality of the data are good enough to enable strong performance of machine learning models. We suggest moving forward with generating & testing sequences for this project.
Discussion needed: The quantity or quality of the data does not enable us to train a machine learning model that has good performance. We either need to find more data, use the data in a different way, or start the project with a zero-shot round instead (where sequences are generated from a large model of the generic protein landscape). The Customer Success and Machine Learning teams will have an in-depth discussion with you to make a decision.
Next steps
After you have completed the benchmark and want to use Cradle for your project, configure Rounds and proceed to Sequence generation.
Last updated
Was this helpful?