Breakthroughs do not pay salaries until someone budgets for them.

AI for science is one of the most serious founder openings in Europe right now, but it is also a perfect place to hide from customers behind beautiful lab slides, impressive papers and grant language.

TL;DR: AI for science uses machine learning, foundation models, robotics, simulation, lab automation, self-driving labs and scientific data systems to shorten the path from question to experiment to verified result. The founder trap is treating the breakthrough as the business. A startup still needs a buyer with a budget, a narrow workflow, reliable lab data, safety evidence, human review, IP thinking and a path from research proof to paid use.

I am Violetta Bonenkamp, founder of Mean CEO, CADChain, and F/MS Startup Game. CADChain sits close to hard technology: CAD data, intellectual property, machine learning, R&D and public funding. That has made me impatient with startup advice that treats every hard problem as if it were a simple SaaS landing page.

AI for science is not a toy category.

It touches biology, chemistry, materials, robotics, energy, climate, medicine, manufacturing, lab data, compute and public research funding.

That also means founders need more discipline, not less.

1 · Definition

What AI For Science Means

AI for science means using AI systems to help scientists ask better questions, design experiments, search literature, predict structures, simulate processes, run robotic lab workflows, analyze data and choose the next test.

It can show up as:

Founder checklist
Founder checks worth seeing together
  • Protein structure prediction.
  • Protein and molecule design.
  • Materials discovery.
  • Climate and weather models.
  • Earthquake and hazard prediction.
  • Lab robotics.
  • Automated microscopy.
  • Synthetic biology support.
  • Data cleaning for research records.
  • Scientific literature review.
  • Experiment planning.
  • Autonomous lab workflows.

The Royal Society report on science in the age of AI says AI is changing the methods and nature of scientific research across fields. It also warns that research integrity, skills and ethics change when AI becomes part of the scientific process.

That is the right frame.

AI for science is not one product category. It is a change in how research work gets done.

For founders, the useful question is narrower:

Which paid scientific workflow can I make faster, cheaper, safer or easier to repeat?

2 · Market signal

Why The Lab Is The Hard Part

Scientific discovery has a brutal bottleneck: the physical world.

A model can suggest a molecule in seconds. A lab may still need weeks or months to synthesize, test, repeat, document and interpret it. A materials model can rank candidates. A research team still has to make samples, run instruments, inspect failures and store data cleanly enough that the next person can trust it.

This is why lab robotics and autonomous experimentation platforms are part of the same market. AI can propose. Robots can execute. Data systems can remember. Humans still decide what matters.

The Berkeley Lab A-Lab is a useful signal. It combines robots and AI to speed materials work, with a closed-loop setup that can choose promising next tests. Berkeley Lab says A-Lab can process 50 to 100 times as many samples as a human every day.

That is not magic.

It is a stack:

Founder checklist
Founder checks worth seeing together
  • Computation.
  • Historical data.
  • Robotic execution.
  • Instrument data.
  • Human scientific goals.
  • Repeatable records.

Most startups will not build the whole stack.

They should not try.

The wedge is usually one painful layer inside it.

3 · Europe lens

Europe Is Taking AI For Science Seriously

Europe has noticed that AI for science is not a side topic.

On October 8, 2025, the European Commission said the AI in Science Strategy would put the EU at the front of AI-led research, with RAISE, the Resource for AI Science in Europe, as a virtual institute that pools AI resources for science. The Commission also described plans for EUR600 million from Horizon Europe for compute access for science and plans to double annual Horizon Europe AI spending to over EUR3 billion.

On November 3, 2025, the Commission launched the RAISE pilot for AI science in Europe, with EUR107 million under Horizon Europe.

That matters for founders because public money, compute access and research networks can create demand signals.

But do not confuse a policy signal with a customer.

Europe’s deep tech shift from software to science explains the wider move: science-based companies are back in the room because Europe needs serious technology in health, energy, compute, climate, defense, manufacturing and security.

The founder job is to turn that attention into a buyer conversation.

4 · Decision filter

The AI For Science Founder Table

Use this table before you build, pitch, or apply for funding.

Startup map
The AI For Science Founder Table
Literature and prior-art scan
Likely buyer

R&D lead, IP team, lab head

What to automate

Paper search, claim mapping, source checks

Proof before scale

Fewer missed sources and faster review

Experiment design assistant
Likely buyer

Principal investigator, lab manager

What to automate

Protocol drafts, reagent plans, run order

Proof before scale

Human accepted plans and fewer failed runs

Robotic sample handling
Likely buyer

Lab operations team, CRO, biotech lab

What to automate

Pipetting, plate work, sorting, assay setup

Proof before scale

Lower error count and better repeatability

Self-driving materials workflow
Likely buyer

Materials lab, energy team, industrial R&D

What to automate

Candidate selection, synthesis plan, test loop

Proof before scale

More usable candidates per month

Protein design support
Likely buyer

Biotech team, pharma R&D, university spinout

What to automate

Sequence ideas, structure checks, screen planning

Proof before scale

Wet-lab validation and IP review

Lab data layer
Likely buyer

Research facility, university lab, CRO

What to automate

Metadata, instrument logs, experiment records

Proof before scale

Searchable records and easier repeat runs

Safety and access layer
Likely buyer

Regulated lab, pharma, industrial buyer

What to automate

Approval logs, user rights, change history

Proof before scale

Reviewed incidents and cleaner audit trail

Grant-to-pilot evidence pack
Likely buyer

Deep tech founder, research spinout

What to automate

Proof folder, buyer notes, budget logic

Proof before scale

Paid pilot path before the next application

This is not a list of "cool ideas."

It is a buyer map.

If you cannot name the buyer, you do not have a product yet.

5 · Key idea

Protein AI Proved The Category, But Wet Labs Still Decide

Protein AI made AI for science visible to the public.

The 2024 Nobel Prize in Chemistry was split between David Baker for computational protein design and Demis Hassabis and John Jumper for protein structure prediction. The Nobel press release says AlphaFold2 has been used by more than two million people from 190 countries.

That is a real scientific shift.

It still does not remove the founder problem.

If you build around proteins, drugs or materials, the model’s suggestion is only part of the path. Someone still has to handle wet-lab testing, data rights, safety, manufacturability, regulation, IP, reimbursement, partnerships and sales.

That is why AI-designed drugs, proteins and materials should be treated as a commercial system, not a headline category.

A beautiful molecule is not a company.

A validated workflow with a buyer, a budget and repeatable proof has a chance.

6 · Capital lens

Self-Driving Labs Are Not A Shortcut Around Science

Self-driving labs combine AI and lab automation in a loop: choose an experiment, run it, analyze the result, update the plan and choose the next experiment.

A Royal Society Open Science review of self-driving laboratories describes systems that combine AI and lab automation for chemistry, materials science and biology. It also raises issues around safety, cybersecurity, IP and human accountability.

Nature’s 2026 feature on the self-driving lab revolution shows the same tension. Robotic tools can take over tasks that humans used to do, but the question becomes what scientists should control, verify and own.

For a founder, the lesson is plain:

Automation does not remove scientific responsibility.

It moves responsibility into:

  • Experiment selection.
  • Data quality.
  • Instrument calibration.
  • Human approval.
  • Safety limits.
  • Error logging.
  • IP records.
  • Repeat testing.
  • Buyer proof.

That is why the boring layers can become good businesses.

The startup that tracks the experiment trail may make money before the startup trying to own the whole autonomous lab.

7 · Risk filter

Where Bootstrapped Founders Can Actually Enter

Most bootstrapped founders cannot start by building a full robotic laboratory.

That is fine.

A tiny team can enter through a wedge:

  • A lab data cleanup service that turns messy instrument files into searchable records.
  • A literature and patent review workflow for one scientific niche.
  • A protocol assistant for one repeated assay.
  • A safety checklist and approval log for AI-planned experiments.
  • A grant-to-pilot evidence pack for research spinouts.
  • A procurement-ready proof file for lab automation vendors.
  • A model evaluation layer for one scientific task.
  • A low-cost scheduling and sample tracking layer for small labs.

The F/MS AI workflow workshop makes a point I like: automation should start from work you have tested yourself, with clear inputs, style, checks and cost logic. The same applies to lab startups, except the tolerance for fantasy is lower because lab errors can waste money, damage samples or create safety risk.

The F/MS Startup Game teaches founders to move from problem to first customer through practical proof. AI for science founders need that discipline twice as much because science can seduce smart people into building for elegance instead of buyers.

8 · Opportunity map

Grants Can Help, But The Grant Is Not The Market

AI for science often needs public money because lab work, compute, data access and technical proof can be expensive.

I am not anti-grant.

I am anti-founder-amnesia.

AI for science may need grants, university partners, public programs, corporate pilots and private money in the same stack. Use public-private funding for European deep tech to keep public money tied to technical proof, buyer proof, and commercial progress.

Use public money for:

  • Data access.
  • Lab validation.
  • Safety testing.
  • Instrument links.
  • IP support.
  • Pilot setup.
  • Specialist scientific advice.
  • Compute access.

Do not use it for:

  • Avoiding buyer calls.
  • Building endless features.
  • Hiring before cash arrives.
  • Serving proposal text instead of lab users.
  • Treating a paper as sales proof.
  • Entering a consortium with no buyer path.

Write this sentence before applying:

This funding will help us prove one scientific workflow for one buyer type by one date.

If you cannot fill that sentence, you are probably about to donate months of your life to paperwork.

9 · Buyer lens

What Buyers Actually Pay For

AI for science buyers do not pay for "AI."

They pay for a painful job to improve.

Buyers may pay when your product:

  • Reduces failed experiment runs.
  • Cuts time from idea to first lab result.
  • Makes lab data easier to find and reuse.
  • Lowers manual work around sample prep.
  • Helps scientists choose better next tests.
  • Creates an audit trail for regulated work.
  • Makes IP review faster.
  • Lets a small lab use tools once reserved for big teams.
  • Helps a spinout show proof to partners or investors.
  • Turns a grant project into a paid pilot path.

Notice the pattern.

The buyer pays for a business or research result she already wants.

AI is the method.

The budget belongs to the pain.

10 · Market signal

The Data Problem Nobody Wants To Admit

AI for science runs on data, and scientific data is often messy.

Lab notebooks are incomplete.

Instrument files use different formats.

Metadata is missing.

Negative results disappear.

Protocol changes live in someone’s memory.

Samples get renamed.

Old experiments cannot be repeated because the context is gone.

This is why a founder should not rush past data operations. In many labs, the first AI product should not be a model. It should be the data layer that makes model use possible.

CADChain taught me a similar lesson in engineering workflows. CAD data is not useful if rights, access, versions and ownership cannot be traced. Scientific data has the same harsh logic. If nobody can trust the record, nobody should trust the model.

A startup can win by making lab records:

  • Searchable.
  • Versioned.
  • Linked to instruments.
  • Linked to people.
  • Linked to samples.
  • Linked to approvals.
  • Easy to export.
  • Easy to audit.

That does not sound glamorous.

Good.

Glamour rarely pays invoices in deep tech.

11 · Risk filter

Safety, IP And Governance Cannot Wait

AI for science can touch dangerous materials, health data, animal studies, patient records, dual-use biology, toxic chemicals, manufacturing secrets and patentable inventions.

So the startup needs guardrails early.

Useful guardrails include:

  • Human approval for high-risk experiment plans.
  • Restricted access to sensitive datasets.
  • Logs for model suggestions and human edits.
  • Clear rules on who owns model-generated hypotheses.
  • IP review before public sharing.
  • Safety blocks for restricted protocols.
  • Supplier and instrument records.
  • Versioned datasets.
  • Incident notes.

An AI governance trail fits naturally here, even if the lab startup is tiny. If an AI system suggests experiments, changes parameters, routes data or influences a research decision, the team needs a record.

Not for theatre.

For memory, safety, buyers and future diligence.

12 · Key idea

A Founder-Friendly Offer To Sell First

Do not start with "autonomous science platform."

Start with one paid offer:

AI lab workflow audit and pilot setup

Best for: a small biotech lab, materials lab, university spinout, CRO, climate lab or industrial R&D team with one repeated scientific workflow that wastes time or samples.

Scope:

  • Map one workflow.
  • Identify repeated manual steps.
  • List data sources and missing metadata.
  • Choose one AI or automation step.
  • Set human review points.
  • Define safety limits.
  • Create one experiment log.
  • Run one controlled pilot.
  • Compare time, errors, repeatability and buyer value.

Buyer promise:

"We will find one lab workflow where AI or automation can save time without breaking scientific trust."

That sentence is not flashy.

It is buyable.

13 · Action plan

What To Do This Week

If you want to build in AI for science, do this before touching a pitch deck:

  1. Pick one scientific field.
  2. Pick one repeated lab or data workflow.
  3. Interview five buyers who own that workflow.
  4. Ask what fails, what costs time and what blocks budget.
  5. Collect three real samples of messy data, if allowed.
  6. Write the human review rule before writing the AI prompt.
  7. Define the smallest paid pilot.
  8. Decide what evidence proves the pilot worked.
  9. Check IP and safety boundaries.
  10. Ask whether the buyer can pay this quarter.

If the buyer cannot pay, keep learning.

If the buyer can pay, build the ugly version that proves the job.

14 · Verdict

The Bottom Line

AI for science will create serious companies.

It will also create elegant science projects with no buyer.

The difference is not intelligence. It is commercial discipline.

Founders should stop asking whether AI can speed discovery.

It can.

Ask instead:

  • Who pays for the speed?
  • Which workflow hurts now?
  • Which lab result proves value?
  • Which human remains accountable?
  • Which record survives diligence?
  • Which safety limit blocks dumb automation?
  • Which buyer has budget this quarter?

Breakthroughs matter.

But customers pay salaries.

What is AI for science?

AI for science is the use of AI systems in scientific research. It can support literature review, data analysis, experiment planning, simulation, protein design, materials discovery, lab robotics and self-driving lab workflows.

The practical founder version is narrower: AI for science should help one research team get from a scientific question to a trusted result faster, cheaper or with less waste.

How does AI for science automate research labs?

AI for science can automate research labs by helping choose experiment plans, running robotic workflows, analyzing instrument data, adjusting the next test and keeping records of what happened.

Automation can handle repeated tasks, but a serious lab still needs human scientific judgment, safety boundaries, clean data, instrument checks and review points. A founder should never sell "hands-off science" when the buyer still carries the risk.

What is a self-driving lab?

A self-driving lab is a research setup where AI and lab automation form a loop. The system proposes an experiment, robotic tools run it, data is analyzed, and the next experiment is chosen from the result.

Self-driving labs are most visible in chemistry, materials science and biology. They can reduce manual work and speed up search, but they also raise questions around safety, IP, cybersecurity and human accountability.

Why should founders care about AI for science?

Founders should care because scientific workflows are full of expensive delays: failed experiments, messy data, slow sample prep, repeated manual work, hard literature searches and weak records.

These delays are commercial openings. A small founder does not need to own the whole lab. She can solve one painful layer that a lab, spinout, pharma team, CRO or industrial buyer will pay to fix.

Which buyers pay for AI for science startups?

Likely buyers include biotech teams, pharma R&D groups, contract research organizations, materials labs, climate and energy research teams, university spinouts, industrial R&D teams, grant-backed consortia and lab automation vendors.

The buyer is usually the person who owns the workflow pain: a lab manager, R&D lead, principal investigator, operations lead, IP lead, safety lead or founder of a research spinout.

How do lab robotics and AI fit together?

Lab robotics executes physical tasks such as pipetting, sample handling, plate work, synthesis, imaging or instrument loading. AI helps choose, interpret or adjust the work.

Together, they can form a research loop. The robot runs the experiment. The AI helps analyze the result. A human scientist checks the meaning, safety and next step.

What proof does an AI for science startup need?

An AI for science startup needs proof that the product improves a real workflow. That can mean fewer failed runs, faster review, cleaner data, better repeatability, fewer manual steps, more usable candidates per month, or a paid pilot that leads to a larger contract.

Scientific proof and business proof are different. A model can be scientifically impressive and still fail as a startup if no buyer budgets for it.

Can bootstrapped founders build in AI for science?

Yes, but they should enter through a narrow wedge. A bootstrapped founder can build a data layer, review workflow, protocol assistant, safety log, pilot evidence pack, literature workflow or sample tracking tool before touching expensive lab hardware.

The trick is to sell a paid service or pilot first, then automate the repeated parts. That keeps the company close to buyer pain and reduces fantasy spending.

What risks should AI for science founders handle early?

AI for science founders should handle safety, IP, data rights, research integrity, access control, audit trails, model limits and human accountability early.

If the product touches health data, chemicals, biological materials, regulated research or patentable inventions, weak governance can kill buyer trust. A small record system beats a big apology later.

How should a founder price an AI for science tool?

Price against the buyer’s pain, not the AI feature. A buyer may pay to reduce failed lab runs, save scientist hours, prepare a grant-to-pilot evidence file, improve repeatability, shorten review time or make data easier to reuse.

For a first offer, use a fixed-scope pilot around one workflow. Name the workflow, timeline, human review point, proof target and buyer decision at the end. That makes the purchase easier to approve.