Please Google/Demis/Sergei, just release the darn weights. This thing ain't gonna be curing cancer sitting behind an API and it's not gonna generate that much GCloud revenue when the model is this tiny.
You can state as a philosophical ideal that you prefer open source or open weights, but that's not something deepmind has prioritized ever.
I think it's worth discussing:
* What are the advantages or disadvantages of bestowing a select few with access?
* What about having an API that can be called by anyone (although they may ban you)?
* Vs finally releasing the weights
But I think "behind locked down API where they can monitor usage" makes sense from many perspectives. It gives them more insight into how people use it (are there things people want to do that it fails at?), and it potentially gives them additional training data
But the submission blog post writes:
> To advance scientific research, we’re making AlphaGenome available in preview via our AlphaGenome API for non-commercial research, and planning to release the model in the future. We believe AlphaGenome can be a valuable resource for the scientific community, helping scientists better understand genome function, disease biology, and ultimately, drive new biological discoveries and the development of new treatments.
And at that point, they're painting this release as something they did in order to "advance scientific research" and because they believe "AlphaGenome can be a valuable resource".
So now they're at a cross-point, is this release actually for advancing scientific research and if so, why aren't they doing it in a way so it actually maximizes advancing scientific research, which I think is the point parent's comment.
Even the most basic principle for doing research, being able to reproduce something, goes out the window when you put it behind an API, so personally I doubt their ultimate goal here is to serve the scientific community.
Edit: Reading further comments it seems like they've at least claimed they want to do a model+weights release of this though (from the paper: "The model source code and weights will also be provided upon final publication.") so remains to be seen if they'll go through with it or not.
The precedent I'm going with is specifically in the gene regulatory realm.
Furthermore, a weight release would allow others to finetune the model on different datasets and/or organisms.
This is a real tradeoff of freedom vs _. I agree that I'm not always a fan of Google being the one in control, but I'm much happier that they are even releasing an API. That's not something they did for go! (Of course there was a book written so someone got access)
And if they don't, I'm not sure how this will gain adoption. There are tons of well-maintained and established workflows out there in the cloud and on-prem that do all of these things AlphaGenome claim to do very well - many that Google promotes on their own platform (e.g., GATK on GCP).
(People in tech think people in science are like people in tech just jump on the latest fads from BigTech marketing - when it's quite opposite it's all about whether your results/methods will please the reviewers in your niche community)
Page 59 from the preprint[1]
Seems like they do intend to publish the weights actually
[1]: https://storage.googleapis.com/deepmind-media/papers/alphage...
This is in the press release, so they are going to release the weights.
What you’re describing is more like whole cell simulation. Whole cells are thousands of times larger than a protein and cellular processes can take days to finish. Cells contain millions of individual proteins.
So that means that we just can’t simulate all the individual proteins, it’s way too costly and might permanently remain that way.
The problem is that biology is insanely tightly coupled across scales. Cancer is the prototypical example. A single mutated letter in DNA in a single cell can cause a tumor that kills a blue whale. And it works the other way too. Big changes like changing your diet gets funneled down to epigenetic molecular changes to your DNA.
Basically, we have to at least consider molecular detail when simulating things as large as a whole cell. With machine learning tools and enough data we can learn some common patterns, but I think both physical and machine learned models are always going to smooth over interesting emergent behavior.
Also you’re absolutely correct about not being able to “see” inside cells. But, the models can only really see as far as the data lets them. So better microscopes and sequencing methods are going to drive better models as much as (or more than) better algorithms or more GPUs.
Side note: whales rarely get cancer.
Personally, I think arc's approach is more likely to produce usable scientific results in a reasonable amount of time. You would have to make a very coarse model of the cell to get any reasonable amount of sampling and you would probably spend huge amounts of time computing things which are not relevant to the properties you care amount. An embedding and graphical model seems well-suited to problems like this, as long as the underlying data is representative and comprehensive.
Can't emphasize enough about how DNA requires human data curation to make things work, even from day one alignments models were driven based on biological observations. Glad to see UBERON, which represents a massive amount of human insight and data curation of what is for all intents and purposes a semantic-web product (OWL based RDF at the heart) playing a significant role.
I’d pitch this paper as a very solid demonstration of the approach, and im sure it will lead to some pretty rapid developments (similar to what Rosettafold/alphafold did)
For instance, Evo2 by the Arc Institute is a DNA Foundation Model that can do some really remarkable things to understand/interpret/design DNA sequences, and there are now multiple open weight models for working with biomolecules at a structural level that are equivalent to AlphaFold 3.
To a man with a hammer…
There are technologies applicable broadly, across all business segments. Heat engines. Electricity. Liquid fuels. Gears. Glass. Plastics. Digital computers. And yes, transformers.
I parted ways with Google a while ago (sundar is a really uninspiring leader), and was never able to transfer into DeepMind, but I have to say that they are executing on my goals far better than I ever could have. It's nice to see ideas that I had germinating for decades finally playing out, and I hope these advances lead to great discoveries in biology.
It will take some time for the community to absorb this most recent work. I skimmed the paper and it's a monster, there's just so much going on.
I understand, but he made google a cash machine. Last quarter BEFORE he was CEO in 2015, google made a quarterly profit of around 3B. Q1 2025 was 35B. a 10x profit growth at this scale well, its unprecedented, the numbers are inspiring themselves, that's his job. He made mistakes sure, but he stuck to google's big gun, ads, and it paid off. The transition to AI started late but gemini is super competitive overall. Deepmind has been doing great as well.
Sundar is not a hypeman like Sam or Cook, but he delivers. He is very underrated imo.
Satya looked like a genius last year with OpenAI partnership, but it is becoming increasingly clear that MS has no strategy. Nobody is using Github Copilot (pioneer) or MS Copilot (a joke). They dont have any foundational models, nor a consumer product. Bing is still.. bing, and has barely gained any market share.
Their strategy and execution was insanely good, and I doubt we'll ever see anything so comprehensive ever again.
1. Clear mission statement: A PC in very house.
2. A nationwide training + certification program for software engineers and system admins across all of Microsoft's tooling
3. Programming lessons in schools and community centers across the country to ensure kids got started using MS tooling first
4. Their developer operations divisions was an insane powerhouse, they had an army of in house technical writers creating some of the best documentation that has ever existed. Microsoft contracted out to real software engineering companies to create fully fledged demo apps to show off new technologies, these weren't hello world sample apps, they were real applications that had months of effort and testing put into them.
5. Because the internet wasn't a distribution platform yet, Microsoft mailed out huge binders of physical CDs with sample code, documentation, and dev editions of all their software.
6. Microsoft hired the top technical writers to write books on the top MS software stacks and SDKs.
7. Their internal test labs had thousands upon thousands of manual testers whose job was to run through manual tests of all the most popular software, dating back a decade+, ensuring it kept working with each new build of Windows.
8. Microsoft pressed PC OEMs to lower prices again and again. MS also put their weight behind standards like AC'97 to further drop costs.
9. Microsoft innovated relentlessly, from online gaming to smart TVs to tablets. Microsoft was an early entrant in a ton of fields. The first Windows tablet PC was in 1991! Microsoft tried to make smart TVs a thing before there was any content, or even wide spread internet adoption (oops). They created some of the first e-readers, the first multimedia PDAs, the first smart infotainment systems, and so on and so forth.
And they did all this with a far leaner team than what they have now!
(IIRC the Windows CE kernel team was less than a dozen people!)
They also leveraged their relationship with Intel to the max - Wintel was a phrase for a reason. Companies like Apple faltered, in part, in the 90's because of hardware disadvantages.
Often their competitors had superior products - but MS still won through - in part helped by their ruthlessly leveraging of synergies across their platforms. ( though as new platforms emerged the desire to maximise synergies across platforms eventually held them back).
That aggressive, Windows everywhere behaviour, is what united it's competitors around things like Java, then Linux and open source in general which stopped MS's march into the data centre, and got regulators involved when they tried to strangle the web.
It showed
CE was a dog and probably a big part of the reason Windows Phone failed. Migrating off of it was a huge distraction and prevented the app platform from being good for a long time. I was at Microsoft and worked on Silverlight for a bit back then.
IMHO the reason for Microsoft's failed phone venture was moving onto the windows kernel and 2xing system requirements.
You have got to be kidding. The 90s was my heyday, and Microsoft documentation was extravagantly unhelpful, always.
One of my internships was at a company writing an example app for SQL server offline replication. Taking a DB that had changed while offline and syncing them to a master DB when reconnection happened. (Back in 2004 or so, now days this is an easier thing).
The company I interned at was hired by MSFT to write a sample app for Fabrikam Fine Furniture that did the following:
1. Sales people on the floor could draw a floorplan on a tablet PC of a desired sectional couch layout and the pieces would be identified and the order automatically made up .
2. Customer enters their delivery info on the tablet.
3. DB replicated down to the delivery driver's tablet PC when the driver next pulls into the loading bay with all the order info.
4. After the delivery is finished and signed for on the tablet PC, the customer's signature is digitally signed so it cannot be tampered with later.
5. When the delivery truck pulls back into the depot, SQL server replication happens again, syncing state changes from the driver back to the master DB.
That is an insane sample app, just one of countless thousands that Microsoft shipped out. Compare that to the bare bones hello world samples you get now days.
I am going to have to disagree with this. Azure is number 2, because MS is number 1 in business software. Cloud is a very natural expansion for that market. They just had to build something that isn't horrible and the customers would have come crawling to MS.
- Created the windows server product
- Created the "rent a server" business line
- Identified the need for a VM kernel and hired the right people
- Oversaw MSFT's build out of web services (MSN, Xbox Live, Bing) which gave them the distributed systems and uptime know-how
- Picked Satya to take over Azure, and then to succeed him
Google is not behind capability wise, they are in front of MSFT actually. The customer relationships matter a whole lot more.
I dont disagree with anything you said because turning a ship around is hard. But hand-to-heart, what big tech company is truly innovating to the future. Lets look at each company.
Apple - bets are on VR/AR. Apple Car is dead. So it is just Vision Pro
Amazon - No new bets. AWS is printing money, but nothing for the future.
Microsoft - No new bets. They fumbled their early lead in AI.
Google - Gemini, Waymo ..
I think Satya gets a lot more coverage than his peer at Google.
IMO Google should have invested more in Waymo and scaled sooner. Instead they partnered with traditional automakers and rideshare companies, sought outside investment, and prioritized a prestige launch in SF over expanding as fast as possible in easier markets.
In other areas they utterly wasted huge initial investments in AR/VR and robotics, remain behind in cloud, and Google X has been a parade of boondoggles (excluding Waymo which, again, predates Sundar and even X itself).
You could also argue that they fumbled AI, literally inventing the transformer architecture but failing at building products. Gemini 2.5 Pro is good, but they started out many years ahead and lost their lead.
Microsoft - No new bets. Really? Their OpenAI deal and integrating that tech into core products?
Amazon - No new bets? It's still trying drone delivery, and it's also got project Kuiper - moving beyond data centres to providing the network
This is all the 1st step of embrace and extinguish.
People like Scott Guthrie who was a key person behind dot.net, and went on to be the driving force behind Azure. Anyone who did any dot.net work 10+ years ago would know the ScottGu blog and his red shirt.
Google similarly bet on Demis, and the results also show. For someone who got his start doing level design on Syndicate (still one of my all-time favourite games) he's come a long way.
Managing to keep the MS Office grift going and even expand it with MS Teams is something
100% it's Demis.
A Demis vs. Satya setup would be one for the ages.
He's also happens to be a really nice guy in person.
"Somethings are because of CEO, and some things are in spite of CEO"
And it was "willy nilly" attributed that enshittification was because of CEO (how do we know? maybe it was CFO, or board) and Gemini because of Demis (how do we know? maybe it was CEO, or CFO, or Demis himself).
I see somebody saying something on here, I tend to assume that they have a reason for believing it.
If your opinions differ from theirs, you could talk about what you believe, instead of incorrectly saying that a CEO can only be responsible for everything or nothing that a company does.
Google's revenue in 2014 was $75B and in 2024 it was $348B, that's 4.64 times growth in 10 years or 3.1 times if corrected for the inflation.
And during this time, Google failed to launch any significant new revenue source.
The question will be, when and how will the LLM's be attacked with product placements.
Open marked advertisement in premium models and integrated ads in free tier ones?
I still hope for a mostly adfree world, but in reality google seems in a good position now for the transition towards AI (with ads).
Haven't you been watching the headlines here on HN? The volume of major high-quality Google AI releases has been almost shocking.
And, they've got the best data.
If by competitive you mean "We spent $75 Billion dollars and now have a middle of the pack model somewhere between Anthropic and Chinese startup", that's a generous way to put it.
I’m no Google lover — in fact I’m usually a detractor due to the overall enshittification of their products — but denying that Gemini tops the pile right now is pure ignorance.
I'm sure you're a smart person, and probably had super novel ideas but your reply comes across as super arrogant / pretentious. Most of us have ideas, even impressive ones (here's an example - lets use LLMs to solve world hunger & poverty, and loneliness & fix capitalism), but it'd be odd to go and say "Finally! My ideas are finally getting the attention".
perhaps this is the appropriate forum to reference pg
Think of all the tiresome Twitter discussions that went like "I like bagels -> oh, so you hate croissants?".
What makes you think that LLMs can do it?
[1] relapsed capitalist, at best, check the recent Doomscroll interview
I have incredibly mixed feelings on Sundar. Where I can give him credit is really investing in AI early on, even if they were late to productize it, they were not late to invest in the infra and tooling to capitalize on it.
I also think people are giving maybe a little too much credit to Demis and not enough to Jeff Dean for the massive amount of AI progress they've made.
One interesting example of such a problem and why it is important to solve it was recently published in Nature and has led to interesting drug candidates for modulating macrophage function in autoimmunity: https://www.nature.com/articles/s41586-024-07501-1
There is a concerning gap between prediction and causality. In problems, like this one, where lots of variables are highly correlated, prediction methods that only have an implicit notion of causality don't perform well.
Right now, SOTA seems to use huge population data to infer causality within each linkage block of interest in the genome. These types of methods are quite close to Pearl's notion of causal graphs.
This has existed for at least a decade, maybe two.
> There is a concerning gap between prediction and causality.
Which can be bridged with protein prediction (alphafold) and non-coding regulatory predictions (alphagenome) amongst all the other tools that exist.
What is it that does not exist that you "found it disappointing that they ignored"?
Methods have evolved a lot in a decade.
Note how AlphaGenome prediction at 1 bp resolution for CAGE is poor. Just Pearson r = 0.49. CAGE is very often used to pinpoint causal regulatory variants.
So out of my own frustration, I drew this. It's a cross-section of a single base pair, as if you are looking straight down the double helix.
Aka, picture a double-strand of DNA as an earthworm. If one of the earthworms segments is a base-pair, and you cut the earthworm in half, and turn it 90 degrees, and look into the body of the worm, you'd see this cross-sectional perspective.
Apologies for overly detailed explanation; it's for non-bio and non-chem people. :)
https://www.instagram.com/p/CWSH5qslm27/
Anyway, I think the way base pairs bond forces this major and minor grove structure observed in B-DNA.
My graduate thesis was basically simulating RNA and DNA duplexes in boxes of water for long periods of time (if you can call 10 nanoseconds "long") and RNA could get stuck for very long periods of time in the "wrong" (IE, not what we see in reality) conformation, due to phosphate/ 2' sugar hydroxyl interactions.
At least they got the handedness right.
> AlphaGenome will be available for non-commercial use via an online API at http://deepmind.google.com/science/alphagenome
So, essentially the paper is a sales pitch for a new Google service.