There have been a lot of inaccuracies reported about me and Stability AI, today I set the record straight with our team with the message below.
We are going to start being more assertive about what we do and I have some very interesting stories to share about some of the past elements touched upon here.
We have some very interesting things coming up.
This also is very intriguing with regards to the future of media and trust - something that will be essential in the coming years with generative AI and that we have a lot to contribute to.
To: Stability AI Team
Subject: Recent Forbes Article
Team,
When we launched Stability AI, we did so with the goal of using AI to help activate humanity’s potential. We saw a chance to enhance scientific research, education and the arts, and positively impact industries in ways that are transformative. That incredible mission has been guiding our team’s work since day one.
In a race to achieve this, we have put the bulk of our focus into our core work: creating innovative products and supporting research implementing AI in the fields of medicine, healthcare, language, imaging, music and more. In just over 18 months, we have grown from 20 employees to more than 170, while building a young, dynamic company that is a leader in the AI space, both through our core teams’ innovation and supporting others’.
It is not surprising to me that as we have grown in prominence and impact, we have also increasingly become the target of several attacks and misrepresentations in the media. This latest article attacking our company by Forbes is no exception. Despite our team spending weeks and weeks going back and forth with Forbes to correct the record, they have clearly chosen to ignore the truth on many of these issues.
Throughout my career, I have always been quick to praise and attribute the work of collaborators. I have apologized for any errors I make and have always strived to be constantly improving. I have categorically denied the spurious claims and characterizations throughout this story, but I want to set the record straight for you, my valued team, here today. These were all clearly communicated to Forbes ahead of release.
There are countless false accusations and misrepresentations in this Forbes story which I will clarify here, in particular:
● My degree from Oxford:
● My prior hedge fund work:
● Our work with the United Nations around COVID-19:
● Stable Diffusion:
● Our partnership with Amazon:
● Our payment of wages and payroll taxes:
● Our efforts to strengthen and improve our HR processes:
● Our relationship with MidJourney:
“MidJourney is an independent organisation we support that implements and makes usable the image models we create. It currently has almost 57,000 members, 14,000 of whom are active and after our support for the beta has been scaling organically through a subscription that covers their compute costs. We would like to help them scale this aggressively, introducing aesthetic scoring and preference learning so the system improves for each user using the inputs of all users. We provide strategic and model support for MidJourney whose focus is not scaling for the sake of it, but improving access to these new models for as many people as possible.”
● Zehra’s support for our company:
● Our continued fundraising efforts:
● My autism research etc:
It is in a way encouraging that after a huge number of interviews directly targeting those who have axes to grind against Stability AI this is the sum of what they could attack us with.
The company has moved from a pre-seed startup last year to a scale up this year and is well on its way to becoming a multinational with huge interest and increasing support in the most amazing transformative technology we have ever seen.
We are a diverse team coming together to build the foundation to activate humanity’s potential and there will be significant scrutiny rightly on everything we do.
We must focus on being open at our core, admitting our faults and constantly improving so we can achieve our mission and make the world a happier place.
The alternative is a closed panopticon which none of us want and we are fortunate to have an amazing and growing team that can deliver on this.
Let’s ship so they can’t ignore the good we do.
Thank you,
Emad
]]>
Today Google announced/released applications based on PaLM 2, their new smol Language Model.
While the trend had been larger models until a year ago, something reflected in our calling them LLMs, Large Language Models, since these have become useful we have actually seen a move to flip from research to engineering to get them to a stage they can go to billions of people without using every GPU on Earth.
We have also seen that many of the parameters in these models and underlying data aren't actually required, we have been feeding them junk food.
The accompanying technical document is short on details on the final architecture, datasets etc but does show that models similar from 400m to 14b parameters achieve great performance, something we have see with the open source explosion around LLaMa and other base models recently as folk realised you can get large models to teach smaller ones.
This comes a few days after an article that was sent to me approximately 127387129837 times,
Google "We Have No Moat, And Neither Does OpenAI"
To the surprise of many I didn't actually agree with this piece on a number of things, even though I am one of the bigger proponents for open source AI.
While the open language model acceleration has been wonderful to see, the truth is that a lot of the work around this has been by a handful of folk and mostly fine tunes and some adjustments to the base model.
Some of the details in the piece were also a bit odd, does anyone at Google actually use LoRA?
These techniques are nothing new and distillation and student teacher approaches are well known and actually the former pioneered at Google.
There have been a number of internal issues and walls that have prevented information sharing and even process optimisation - why we saw Chinchilla outperform the original PaLM yet the original still used for example.
Google started with innovation but innovation is a really difficult thing to build a good company around - the innovation is the catalyst but the ingredients to a good company are turning that research and innovation into engineering and operation (sound familiar?).
I wrote more on that here:
While this article fits with much of our thesis I think it has a misunderstanding of what moats actually are
— Emad (@EMostaque) May 4, 2023
It is v difficult to build a business with innovation as a moat, base requirement is too high
Data, distribution, great product are moatshttps://t.co/Qjz3Syfzxg
The narrative will likely turn as Google integrates Palm 2 and then Gemini into its various products to drive real customer value - maximising function and flow and reducing annoying frustration.
gmail.ai wen
Similarly while the models before they start slimming are already good enough, fast enough and cheap enough to make an impact, we will see this really take off into 2024 as we start to explore some of the limits of optimisation and engineering, something Google is good at (Amazon too actually).
This aligns with my call in January saying not to write off Google as they have all the right ingredients to be one of the couple of main players in this space, something that will become more and more apparent.
Stories about how chatGPT will kill Google are a bit silly.
— Emad (@EMostaque) January 4, 2023
Google have the best full stack LLM team and infra with custom chips (PaLM, LaMBDA, Chinchilla, MUM, TPUs etc)
Nobody can bet them on innovation, cost or go to market.
Institutional inertia is only limiting factor.
The dynamic this sets is interesting though - if the models are this good, this fast and getting better, where is the space for other proprietary players?
(Also why did they use Adobe Firefly in Bard instead of Imagen or Muse?)
I think folk need to consider what would be the impact if it turned out you could get GPT-4 level performance in 1 billion parameters or less, something that could work offline on a phone (maybe the cool Pixel Fold)?
Looking at how things are going I don't think this is inconceivable from proprietary models and we will continue to see massive innovation in this space at a pace few of us expect.
I think Google, OpenAI and a few others will lead on this and it will be really hard to compete on features, meaning focusing on the unchanging needs for a business around this will be critical.
We shall see, but for now Sundar came to dance.
edit: heh AI is awesome
Freedom and Alignment are 100% orthogonal concepts and should NOT be conflated.
— JJ (@JosephJacks_) February 25, 2023
We risk the most catastrophic outcomes by societally conflating these things.
Terrorists use open source to murder people every day… that’s reality. And we cannot change that but must strive to…
"A purely peer-to-peer version of electronic cash would allow online payments to be sent directly from one party to another without going through a financial institution."
The intelligent internet.
— Emad (@EMostaque) August 25, 2022
Personalised generative search models for everyone that compress knowledge.
Image, audio, text & more.
Distributed & dynamic - a protocol for the next generation.
This is how we build the foundation to activate human potential @StabilityAI
The spreading of model weights and customisations of those weights that are even smaller is something with similar properties and combined with identity & value transfer rails can lead to some very interesting outcomes for the benefit of society."It takes advantage of the nature of information being easy to spread but hard to stifle."
Over the years we build up all sorts of mental models of how the world is.
One of the ones I have been using recently is that it is relatively easy to speak face to face, notwithstanding shyness, social anxiety and more.
Writing is harder, much harder, especially concise writing.
Blaise Pascal in 1657 wrote:
Je n’ai fait celle-ci plus longue que parce que je n’ai pas eu le loisir de la faire plus courte.
"I have made this longer than usual because I have not had time to make it shorter."
In a presentation (hoho) a while ago I did this slide to show how it is:
Now with generative AI it is going to be more like this:
Maybe. Friends like Tome and others are working hard to make this a reality.
One of the main things here is that it is good to create and get stuff out.
As the CEO of one of the more differentiated companies in one of fastest growing and likely world-changing sectors we have ever seen its really not easy.
I've had loads of failures despite doing a bunch of interesting things, but never lived up to my own expectations of myself.
With Asperger's & ADHD it's been quite difficult and it's only recently I've been getting to grips with my own mind.
Aphantasia and a lack of an internal voice also make the world strange versus others.
It's a privileged position, although a bit lonely, that I find myself so I thought it might be nice to take my own advice and get things out, something I think may be interesting to folk out there.
This is part of a broader shift I want to bring at Stability AI where we will start building language and other models in the open - lots of folk want to help and its good to get things out there.
I am also going to be experimenting with different AI tools to see how the communication process can be eased - this post has no AI but future ones will.
Even while writing this out right now I am feeling quite nervous - what if folk judge what I write, I say stupid things, I bore them etc
I think this is a very human thing and I am going to try not to let it stop me.
I'll write about whatever I find interesting and maybe you will too.
Away we go.
]]>