Newsletter Subject

Scaling AI – Big and Small

From

brownstoneresearch.com

Email Address

feedback@e.brownstoneresearch.com

Sent On

Tue, Aug 6, 2024 08:31 PM

Email Preheader Text

Scaling AI – Big and Small By Jeff Brown, Editor, The Bleeding Edge ---------------------------

[The Bleeding Edge]( Scaling AI – Big and Small By Jeff Brown, Editor, The Bleeding Edge --------------------------------------------------------------- It’s the greatest prize in history… The creation of an artificial general intelligence (AGI). Yesterday, we spent some time studying the field of competition. The major players are Microsoft, OpenAI, Meta, Alphabet (Google), xAI (Elon Musk), Apple, Anthropic, and I’d throw in a little-known but no longer small competitor, Mistral AI. For anyone who missed that discussion, you can catch up here in [The Bleeding Edge – The Field for Large Language Models](. This competition is not for the faint of heart. The ticket to enter the race starts at 100,000 high-powered GPUs to train these foundational models and $1 billion in costs to do so. It only goes up from there. And yes, more capital means more computing… and a chance to beat the competition in this all-out arms race. It’s not just about making a ton of money from their large-language product offerings (in the form of subscription, API fees, or licensing fees), either. While the economic incentives are what justifies the capital spend and investment, the race is also highly ideological. The Incredible Power of Rewriting Reality The winner, or winners, will have incredible power. The power to rewrite history, science, and biology in whatever way they prefer it to be… not necessarily as it actually is. After all, a large language model (LLM) like Google’s Gemini 1.5, Meta’s Llama 3.1 405b, or OpenAI’s GPT-4o feels like an all-knowing magical black box to most of us. We tell it what we need… and seconds later, sometimes milliseconds, the answer just appears. We got a taste of how this could work this past February when Google announced to the world the highly anticipated release of its foundational LLM, Gemini, version 1.5. The historical likeness of the Vikings – the seafaring Norse people from Scandinavia who raided, traded, and settled in various parts of Europe from the 8th to 11th centuries – is not represented in these images. Does anybody really care? It will be quite hard for the general population to even know that the answers are heavily biased. In fact, they have been programmed by humans to be so. After all, garbage in, garbage out. That’s why this competition is so important. It’s not just about creating the largest productivity-enhancing technology in human history. It’s also about whether or not our foundation of knowledge will be real and factual… or made up. The implications are staggering. Every month, we see a new model released… Each new model makes major improvements, a step ahead of the pack, and a new approach that takes us one step closer to the prize. It’s this tangible, material progress that drives billions more capital into the field. An entire industry and complete ecosystem is being built from the dirt up, in real time. The sprouts came out of the ground so quickly… and they are maturing before our eyes as if it’s all happening in a single growing season. But I don’t think most people understand what’s at stake – aside from a new suite of cool tools to help us generate on the spot a college essay, an email, an original photo, or some computer code. Today’s LLMs are already writing legal agreements, conducting patent infringement searches, writing complex software code, developing websites, interacting with the internet, and even discovering how life’s 200 million plus proteins fold. There’s so much progress. And it’s worth a fortune. It’s like a massive greenfield of opportunity. And any legitimate team with expertise in AI can get funding. And the talent is in flux. Just yesterday, OpenAI’s president and two other key employees left OpenAI, arguably the leader in the race to AGI right now. Not the kind of horse we’d expect the talent to jump off of. One went to competitor Anthropic, and I’m confident the other two will pop up somewhere soon. What does that mean for the future of OpenAI? Will it be able to maintain its lead? And amid this frenetic competition to advance large language models, there is something small happening. More specifically, small large-language models. An Answer to Scaling AI? I know this sounds counterintuitive. After all, the game in artificial intelligence (AI) is “go big or go home,” isn’t it? But something we touched on yesterday was a hint. We can have the greatest, highest performing large language model (LLM) like GPT-4o, but if it is very expensive to operate, it just won’t scale to be used by billions of people. That’s where smaller LLMs come into play. After all, most of us don’t need an all-knowing, all-capable AI for a small range of tasks we regularly need to achieve. It doesn’t matter if it’s for a work task or one at home, we usually don’t need an advanced AI that does just about everything. We need one that does a few things well. And that’s precisely why the same companies that are building large LLMs with hundreds of billions, even trillions of parameters, have also been in a race to build small LLMs. These models are typically referred to as “cost-efficient” models, designed with both performance and cost of operation in mind. And that means scale. The most recent industry development for small LLMs was the launch by OpenAI of GPT-4o-mini. The name says it all. Source: Artificial Analysis As we can see above, we have a field of small LLMs, and their performance is graphed against their cost to operate. With most charts, the “best” technology or company is in the upper right quadrant. But in the above chart, the ideal place is in the upper left. This equates to high performance and low cost. OpenAI’s release of GPT-4o-mini was a big deal because it stands out as the highest-performing small LLM and also one of the cheapest models. GPT-4o-mini even outperformed GPT-4 (OpenAI’s previous LLM) on some benchmarks. GPT-4o-mini scores 82% on the massive multitask language understanding ([MMLU]( benchmark, which covers a wide range of tasks and knowledge. It is priced at less than half Anthropic’s Claude 3 Haiku and Google’s Gemini 1.5 Flash models, and just a fraction of the cost compared to Meta’s Llama 3. And compared to past frontier large language models, it is a magnitude cheaper. Our instincts probably suggest that we’d be giving up a lot of functionality with these new small models. But that’s not the case. GPT-4o-mini is capable of supporting text and real-time computer vision as inputs, with support for audio, image, and video soon. That means, for fractions of a penny, we’ll be able to speak with our AIs, show them our world using our smartphone’s camera, and provide it images and video. And it will be able to speak with us, write to us, and provides us images and even video. Some critics have suggested that this race for cost efficiency is a race to the bottom. They’re completely wrong. And they miss the point entirely. The Industry Shift Towards Inference The industry is proactively getting ahead of what the market needs. AI isn’t a one-size-fits-all technology. There will be smaller models that specialize in scheduling, design, history, process automation, construction, materials… you name it. They will be designed and optimized for specific tasks while maintaining powerful natural language processing capabilities. This isn’t a race to the bottom, it’s a race to scale. AI isn’t something that’s being developed just for the top 10% of the global population, it’s being developed to reach the world’s connected population… which is more than 5 billion. And this isn’t just about developing cost-efficient software. When it comes to running these models, there is a new breed of AI-specific semiconductor companies that are designed to run these models – this is called inference. One of those companies has just recently confidentially filed for its IPO. It’s a company that I’ve been following for years. And in a few days, I’ll be sharing more details about how investors can get ready for that event – [you can get ready by going here](. Counterintuitive or not, something big and small is happening in the world of artificial intelligence. Does that mean that we’re nearing the end or closer to the beginning? I’ll let you decide. But I’ll leave you with this… Microsoft just spent $19 billion last quarter on capital expenditures (CapEx). Meta increased its CapEx forecast for the year to between $37 and $40 billion. Amazon has already spent $30.5 billion in the first half of this year and forecasted it will spend more in the second half. Current forecasts for the data center CapEx spend in 2028 are almost $200 billion… just for the top five U.S. hyperscale tech companies. I’ve never seen growth like this… And AI is a foundational technology, a fabric, that is being woven into every industry. It will accelerate innovation, efficiencies, breakthroughs, and economic growth in all sectors. So who cares if a bunch of institutions and hedge funds get their hands stuck in the cookie jar on the yen carry trade? They used too much leverage, and they got caught. It’s just a blip of volatility that won’t get in the way of either big or small breakthroughs in artificial intelligence. --------------------------------------------------------------- Like what you’re reading? Send your thoughts to feedback@brownstoneresearch.com. [Brownstone Research]( Brownstone Research 55 NE 5th Avenue, Delray Beach, FL 33483 [www.brownstoneresearch.com]( To ensure our emails continue reaching your inbox, please [add our email address]( to your address book. This editorial email containing advertisements was sent to {EMAIL} because you subscribed to this service. To stop receiving these emails, click [here](. Brownstone Research welcomes your feedback and questions. But please note: The law prohibits us from giving personalized advice. To contact Customer Service, call toll free Domestic/International: 1-888-512-0726, Mon–Fri, 9am–7pm ET, or email us [here](mailto:memberservices@brownstoneresearch.com). © 2024 Brownstone Research. All rights reserved. Any reproduction, copying, or redistribution of our content, in whole or in part, is prohibited without written permission from Brownstone Research. [Privacy Policy]( | [Terms of Use](

EDM Keywords (225)

yesterday yes years year woven worth world winners winner whole whether way volatility vikings video usually used use us two train touched ton ticket throw thoughts think tell technology taste tasks talent support suggested subscribed stands spot spent spend specialize speak something smartphone small sharing settled service sent see sectors scandinavia scale running run represented release redistribution real reach race quickly questions provide programmed prize priced president prefer precisely power pop play performance people penny part parameters pack optimized opportunity operation operate openai one need necessarily nearing name money models missed miss mind microsoft meta means mean maturing matter making maintain made lot llms like life let less leave leader lead launch knowledge knowing know kind justifies jump ipo investors investment internet institutions inputs industry important implications images hundreds humans horse home history hint heart happening ground graphed got google going goes giving get garbage game future functionality fractions fraction foundation fortune form forecasted following flux field feedback faint factual fact fabric eyes expertise expensive expect everything event europe equates enter ensure end email discussion dirt developed details designed decide days critics creation creating covers counterintuitive costs cost content confident computing competition compared company companies comes closer charts chart chance catch cares capital capable camera bunch built bottom blip biology billions beginning beat appears anyone answers answer amid also ai actually achieve able 8th 37 2028

Marketing emails from brownstoneresearch.com

View More
Sent On

17/10/2024

Sent On

16/10/2024

Sent On

15/10/2024

Sent On

15/10/2024

Sent On

14/10/2024

Sent On

11/10/2024

Email Content Statistics

Subscribe Now

Subject Line Length

Data shows that subject lines with 6 to 10 words generated 21 percent higher open rate.

Subscribe Now

Average in this category

Subscribe Now

Number of Words

The more words in the content, the more time the user will need to spend reading. Get straight to the point with catchy short phrases and interesting photos and graphics.

Subscribe Now

Average in this category

Subscribe Now

Number of Images

More images or large images might cause the email to load slower. Aim for a balance of words and images.

Subscribe Now

Average in this category

Subscribe Now

Time to Read

Longer reading time requires more attention and patience from users. Aim for short phrases and catchy keywords.

Subscribe Now

Average in this category

Subscribe Now

Predicted open rate

Subscribe Now

Spam Score

Spam score is determined by a large number of checks performed on the content of the email. For the best delivery results, it is advised to lower your spam score as much as possible.

Subscribe Now

Flesch reading score

Flesch reading score measures how complex a text is. The lower the score, the more difficult the text is to read. The Flesch readability score uses the average length of your sentences (measured by the number of words) and the average number of syllables per word in an equation to calculate the reading ease. Text with a very high Flesch reading ease score (about 100) is straightforward and easy to read, with short sentences and no words of more than two syllables. Usually, a reading ease score of 60-70 is considered acceptable/normal for web copy.

Subscribe Now

Technologies

What powers this email? Every email we receive is parsed to determine the sending ESP and any additional email technologies used.

Subscribe Now

Email Size (not include images)

Font Used

No. Font Name
Subscribe Now

Copyright © 2019–2024 SimilarMail.