As Columbia Business School professor Rita McGrath points out,
it’s about identifying “the strategic inflection points” at which the
cost of waiting exceeds to cost acting — in other words, identifying the
most strategic point to enter a market or adopt a technology, balancing
the risks and opportunities based on market readiness, technological
maturity and organizational capacity.
This speaks to the growing adoption of agile, “act, learn, build”
approaches over traditional “prove-plan-execute” orientations. The
popularity of techniques like discovery-driven
planning, the
lean startup, and other agile approaches and propagated this
philosophy in which, rather than building bullet-proof business cases,
one makes small steps, learning from them, and deciding whether to
invest further.
― “6 Strategic Concepts That Set High-Performing Companies
Apart”, Harvard Business Review [Article]
It’s a very good read. It provided real world business examples such
as Nvidia’s partnership with ARM Holdings and Amazon’s Alexa offering
for the strategic concept “borrow someone’s road”, Microsoft’s decision
to make Office available on Apple’s iOS devices in 2014 and Microsoft’s
partnership with Adobe, Salesforce, and Google for the strategic concept
“Parter with a third party”, Deere & Co’s decision on openly
investing in precision agriculture technologies for the strategic
concept “reveal your strategy”, Mastercard’s “Beyond Cash” initiative in
2012 for the strategic concept “be good”, Ferrari’s strategic entry into
the luxury SUV market for the strategic concept “let the competition
go”, and Tesla’s modular approach to battery manufacturing for the
strategic concept “adopt small scale attacks”.
Gig work is structured in a way that strengthens the alignment
between customers and companies and deepens the divide between customers
and workers, leading to systemic imbalances in its service
triangle.
Bridging the customer-worker divide can result in higher customer
trust and platform commitment, both by the customer and the
worker.
To start, platforms need to increase transparency, reduce
information asymmetry, and price their services clearly, allowing
customers to better understand what they are paying for rather than only
seeing an aggregated total at the end of the transaction. This, in turn,
can help customers get used to the idea that if workers are to be paid
fairly, gig work cannot be a free or low-cost service.
Gig workers might be working through an app, but they are not
robots, and they deserve to be treated respectfully and thoughtfully. So
tip well, rate appropriately, and work together to make the experience
as smooth as possible both for yourself and for workers.
― “How Gig Work Pits Customers Against Workers”, Harvard
Business Review [Article]
This is a good article for better understanding how gig work
structured differently than other business model, and what the key
points are for better business performance and triangle
relationships.
TCP/IP unlocked new economic value by dramatically lowering the
cost of connections. Similarly, blockchain could dramatically reduce the
cost of transactions. It has the potential to become the system of
record for all transactions. If that happens, the economy will once
again undergo a radical shift, as new, blockchain-based sources of
influence and control emerge.
“Smart contracts” may be the most transformative blockchain
application at the moment. These automate payments and the transfer of
currency or other assets as negotiated conditions are met. For example,
a smart contract might send a payment to a supplier as soon as a
shipment is delivered. A firm could signal via blockchain that a
particular good has been receivedor the product could have GPS
functionality, which would automatically log a location update that, in
turn, triggered a payment. We’ve already seen a few early experiments
with such self-executing contracts in the areas of venture funding,
banking, and digital rights management.
The implications are fascinating. Firms are built on contracts,
from incorporation to buyer-supplier relationships to employee
relations. If contracts are automated, then what will happen to
traditional firm structures, processes, and intermediaries like lawyers
and accountants? And what about managers? Their roles would all
radically change. Before we get too excited here, though, let’s remember
that we are decades away from the widespread adoption of smart
contracts. They cannot be effective, for instance, without institutional
buy-in. A tremendous degree of coordination and clarity on how smart
contracts are designed, verified, implemented, and enforced will be
required. We believe the institutions responsible for those daunting
tasks will take a long time to evolve, And the technology challenges
especially security are daunting.
― “The Truth About Blockchain”, Harvard Business
Review [Article]
This is the second Blockchain related article I have read from
Harvard Business Review. Different authors have different perspectives.
Unlike the previous article with a lot of concerns and cautions about
Web3, this article seems more optimistic. It proposed a framework for
adopting blockchain to revolutionize modern business, and a guidance to
Blockchain investment. It points out that Blockchain has great
potentials in boosting the efficiency and reducing the cost for all
transactions and then explained the reason why the adoption of
Blockchain would be slow by making a comparison with TCP/IP, which took
more than 30 years to reshape the economy by dramatically lowering the
cost of connections. This is an interesting comparison: e-mail enabled
bilateral messaging as the first application of TCP/IP, while bitcoin
enables bilateral financial transactions as the first application of
Blockchain. It reminds me about what people (Jun Lei, Huateng Ma, Lei
Ding, etc) were thinking about internet mindset and business model back
in 2000s.
In the end, the authors proposed a four-quadrant framework for
adopting Blockchain step by step. The four quadrants are created by two
dimensions: novelty (equivalent to the amount of efforts required to
ensure users understand the problem) and complexity (equivalent to the
amount of coordination and collaboration required to produce values).
With the increase of both dimensions, the adoption will require more
institutional change. An example of “low novelty and low complexity” is
simply adding bitcoin as an alternative transaction method. An example
of “low novelty and high complexity” is building a new, fully formed
cryptocurrency system which requires wide adoption from every monetary
transaction party and consumers’ complete understanding of
cryptocurrency. An example of “high novelty and low complexity” is
building a local private network on which multiple organizations are
connected via a distributed ledger. An example of “high novelty and high
complexity” is building “smart contracts”.
News
Does Amazon’s cashless Just Walk Out technology rely on 1,000
workers in India? [Link]
Amazon insists Just Walk Out isn’t secretly run by workers
watching you shop [Link]
An update on Amazon’s plans for Just Walk Out and
checkout-free technology [Link]
It’s been reported that there are over 1000 Indian workers behind the
cameras of Just Walk Out. It sounds dystopian and reminds me of
“Snowpiercer” movie in 2013. In 2022, about 700 of every 1000 Just Walk
Out sales had to be reviewed by Amazon’s team in India, according to The
Information. Amazon spokesperson explained that the technology is made
by AI (computer vision and deep learning) while it does rely on human
moderators and data labelers. Amazon clarified that it’s not true that
Just Walk Out relies on human reviewers. They said object detection and
receipt generation are completely AI powered, so no human watching live
videos. But human are responsible for labeling and annotation for data
preparation, which also requires watching videos.
I guess the technology was not able to complete the task end-to-end
by itself without supervision or it’s still on the developing stage? I
believe it could be Amazon’s strategy to build and test Just Walk Out,
Amazon Dash Cart, and Amazon One at the same time while improving AI
system, since they are “just getting started”. As Amazon found out that
customers prefer Dash Cart in large stores, it has already expanded Dash
Cart to all Amazon Fresh stores as well as third-party grocers. And
customers prefer Just Walk Out in small stores, so it’s available now in
140+ thrid-party locations. Customers love Amazon One’s security and
convenience regardless the scale of stores, so it’s now available at
500+ Whole Foods Market stores, some Amazon stores, and 150+ third-party
locations.
Data centres consume water directly to prevent information
technology equipment from overheating. They also consume water
indirectly from coal-powered electricity generation.
The report said that if 100 million users had a conversation with
ChatGPT, the chatbot “would consume 50,000 cubic metres of water – the
same as 20 Olympic-sized swimming pools – whereas the equivalent in
Google searches would only consume one swimming pool”.
― China’s thirsty data centres, AI industy could use more
water than size of South Korea’s population by 2030: report
warns [Link]
The rapid growth of AI could dramatically increase demand on water
resources. AI eats tokens, consumes compute, and drinks water.
15 Graphs That Explain the State of AI in 2024 [Link]
Stanford Institute for Human-Centered Artificial Intelligence (HAI)
published 2024’s AI Index report [Link].
502-page reading journey started.
ins_emails1ins_emails2ins_emails3ins_emails4
― Leaked emails reveal why Mark Zuckerberg bought
Instagram [Link]
Zuckerberg’s discussion of Instagram acquisition back in 2012 proved
his corporate strategic foresights. He was aiming to buy the time and
network effect, rather than simply neutralizing competitors or improving
products. He bought Instagram for \(\$1\)B, today it is worth \(\$500\)B. It’s very impressive.
Introducing Meta Llama 3: The most capable openly available
LLM to date [Link]
Meta released early versions of Llama 3. Pretrained and
instruction-fine-tuned Llama3 with 8B and 70B parameters are now
open-source. Its 405B version is still training.
Llama 3 introduces Grouped Query Attention (GQA), which reduces the
computational complexity of processing large sequences by grouping
attention queries. Llama 3 also had extensive pre-training involving
over 15 trillion tokens, including a significant amount of content in
different languages, enhancing its applicability across diverse
linguistic contexts. Post-training techniques include finetuning and
rejection sampling which refine the model’s ability to follow
instructions and minimize error.
Cheaper, Better, Faster, Stronger - Continuing to push the
frontier of AI and making it accessible to all. [Link]
Mistral AI’s Mixtral 8x22B has a Sparse Mixture-of-Experts (SMoE)
architecture, which maximize efficiency by activating only 44B out of
176B parameters. The model’s architecture ensures that only the most
relevant “experts” are activated during specific tasks. The experts are
individual neural networks as apart of SMoE model. They are trained to
become proficient at particular sub-tasks out of the overall task. Since
only a few experts are engaged for any given input, this design reduces
computational complexity.
GPT-4-Turbo has significantly enhanced its multimodal capabilities by
incorporating AI vision technology. This model is able to analyze
videos, images, and audios. Its tokenizer now has a larger 128000 token
context window, which maximizes its memory.
The race to lead A.I. has become a desperate hunt for the digital
data needed to advance the technology. To obtain that data, tech
companies including OpenAI, Google and Meta have cut corners, ignored
corporate policies and debated bending the law, according to an
examination by The New York Times.
Tech companies are so hungry for new data that some are
developing “synthetic” information. This is not organic data created by
humans, but text, images and code that A.I. models produce — in other
words, the systems learn from what they themselves generate.
― How Tech Giants Cut Corners to Harvest Data for
A.I. [Link]
OpenAI developed a speech recognition tool ‘Whisper’ to transcribe
the audio from YouTube videos, generating text data for AI system.
Google employees know OpenAI had harvested YouTube videos for data but
they didn’t stop OpenAI because Google had also used transcripts of
YouTube videos for training AI models. Google’s rules about the legal
usage of YouTube videos is vague and OpenAI’s employee were wading into
a legal gray area.
As many tech companies such as Meta and OpenAI reached the stage of
data shortage, OpenAI started to train AI models by using synthetic data
synthesized by two different AI models, one produces the data, the other
judges the information.
Musk released the preview of first multimodal model Grok-1.5V. It is
able to understand both textual and visual information. One unique
feature is that it adopts Rust, JAX, and Kubernetes to construct its
distributed training architecture.
One page of the Microsoft presentation highlights a variety of
“common” federal uses for OpenAI, including for defense. One bullet
point under “Advanced Computer Vision Training” reads: “Battle
Management Systems: Using the DALL-E models to create images to train
battle management systems.” Just as it sounds, a battle management
system is a command-and-control software suite that provides military
leaders with a situational overview of a combat scenario, allowing them
to coordinate things like artillery fire, airstrike target
identification, and troop movements. The reference to computer vision
training suggests artificial images conjured by DALL-E could help
Pentagon computers better “see” conditions on the battlefield, a
particular boon for finding — and annihilating — targets.
OpenAI spokesperson Liz Bourgeous said OpenAI was not involved in
the Microsoft pitch and that it had not sold any tools to the Department
of Defense. “OpenAI’s policies prohibit the use of our tools to develop
or use weapons, injure others or destroy property,” she wrote. “We were
not involved in this presentation and have not had conversations with
U.S. defense agencies regarding the hypothetical use cases it
describes.”
Microsoft told The Intercept that if the Pentagon used DALL-E or
any other OpenAI tool through a contract with Microsoft, it would be
subject to the usage policies of the latter company. Still, any use of
OpenAI technology to help the Pentagon more effectively kill and destroy
would be a dramatic turnaround for the company, which describes its
mission as developing safety-focused artificial intelligence that can
benefit all of humanity.
― Microsoft Pitched OpenAI’s DALL-E as Battlefield Tool for
U.S. Military [Link]
Other than what has mentioned in the news, by cooperating with
Department of Defense, AI can understand how human battle and defense,
which is hard to learn from current textual and visual information from
the internet. So it’s possible that this is the first step of AI
troop.
Microsoft scientists developed what they call a qubit
virtualization system. This combines quantum error-correction techniques
with strategies to determine which errors need to be fixed and the best
way to fix them.
The company also developed a way to diagnose and correct qubit
errors without disrupting them, a technique it calls “active syndrome
extraction.” The act of measuring a quantum state such as superposition
typically destroys it. To avoid this, active syndrome extraction instead
learns details about the qubits that are related to noise, as opposed to
their quantum states, Svore explains. The ability to account for this
noise can permit longer and more complex quantum computations to proceed
without failure, all without destroying the logical qubits.
― Microsoft Tests New Path to Reliable Quantum Computers
1,000 physical qubits for each logical one? Try a dozen, says
Redmond [Link]
Think about it in the sense of another broad, diverse category
like cars. When they were first invented, you just bought “a car.” Then
a little later, you could choose between a big car, a small car, and a
tractor. Nowadays, there are hundreds of cars released every year, but
you probably don’t need to be aware of even one in ten of them, because
nine out of ten are not a car you need or even a car as you understand
the term. Similarly, we’re moving from the big/small/tractor era of AI
toward the proliferation era, and even AI specialists can’t keep up with
and test all the models coming out.
This week, the speed of releasing LLMs becomes about 10 per week.
This article provides good explanation about why we don’t need to keep
up with it or test all released models. Car is a good analogy to AI
model nowadays. There are all kinds of brands and sizes, and designed
for different purposes. Hundreds of cars are released every year, but
you don’t need to know them. Majority of the models are not
groundbreaking but whenever there is big step, you will be aware of
it.
Although not necessary to catch up all the news, we at least need to
be aware of the main future model features - where modern and future
LLMs are heading to: 1) multimodality 2) recall capability 3)
reasoning.
ByteDance Exploring Scenarios for Selling TikTok Without
Algorithm [Link]
ByteDance is internally exploring scenarios for selling TikTok’s US
business to non-tech industry without the algorithm if they exhausted
all legal options to fight legislation of the ban. Can’t imagine who
without car expertise is going to buy a car without engine.
Developers and creators can take advantage of all these
technologies to create mixed reality experiences. And they can reach
their audiences and grow their businesses through the content discovery
and monetization platforms built into Meta Horizon OS, including the
Meta Quest Store, which we’ll rename the Meta Horizon Store.
Introducing Our Open Mixed Reality Ecosystem [Link]
Everyone knows how smart Zuck is in the idea of open-source.
Other news:
Elon Musk says Tesla will reveal its robotaxi on August
8th [Link]
SpaceX launches Starlink satellites on record 20th reflight
of a Falcon 9 rocket first stage [Link]
Deploy your Chatbot on Databricks AI with RAG, DBRX Instruct,
Vector Search & Databricks Foundation Models [Link]
Adobe’s ‘Ethical’ Firefly AI Was Trained on Midjourney
Images [Link]
Exclusive: Microsoft’s OpenAI partnership could face EU
antitrust probe, sources say [Link]
Anthropic-cookbook: a collection of notebooks / recipes
showcasing some fun and effective ways of using Claude [Link]
Amazon deploys 750,000+ robots to unlock AI
opportunities [Link]
Apple’s four new open-source models could help make future AI
more accurate [Link]
The Mystery of ‘Jia Tan,’ the XZ Backdoor Mastermind
[Link]
YouTube
If someone whom you don’t trust or an adversary gets something
more powerful, then I think that that could be an issue. Probably the
best way to mitigate that is to have good open source AI that becomes
the standard and in a lot of ways can become the leader. It just ensures
that it’s a much more even and balanced playing field.
― Mark Zuckerberg - Llama 3, $10B Models, Caesar Augustus,
& 1 GW Datacenters [Link]
What I learned from this interview: The future of Meta AI would be a
kind of AI general assistant product where you give it complicated tasks
and then it goes away and does them. Meta will probably build bigger
clusters. No one has built 1GW data center yet but building it could
just be a matter of time. Open source can be both bad and good. People
can use LLM to do harmful things, while what Mask worries more about is
the concentration of AI, where there is an untrustworthy actor having
the super strong AI. Open source software can make AI not getting stuck
in one company but can be broadly deployed to a lot of different
systems. People can set standards on how it works and AI can get checked
and upgraded together.
It is clear that inference was going to be a scaled problem.
Everyone else had been looking at inference as you take one chip, you
run a model on it, it runs whatever. But what happened with AlphaGo was
we ported the software over, and even though we had 170 GPUs vs 48 TPUs,
the 48 TPUs won 99 out of 100 games with the exact same software. What
that meant was compute was going to result in better performance. And so
the insight was - let’s build scaled inference.
(Nvidia) They have the ecosystem. It’s a double-sided market. If
they have a kernel-based approach they already won. There’s no catching
up. The other way that they are very good is vertical integration and
forward integration. What happens is Nvidia over and over again decides
that they want to move up the stack, and whatever the customers are
doing, they start doing it.
Nvidia is incredible at training. And I think the design decision
that they made including things like HBM, were really oriented around
the world back then, which was everything is about training. There
weren’t any real world application. None of you guys were really
building anything in the wild where you needed super fast
inference.
What we saw over and over again was you would spend 100% of your
compute on training, you would get something that would work well enough
to go into production, and then it would flip to about 5%-10% training
and 90%-95% inference. But the amount of training would stay the same,
the inference would grow massively. And so every time we would have a
success at Google, all of a sudden, we would have a disaster, we called
it the success disaster, where we can’t afford to get enough compute for
inference.
HBM is this High Bandwidth Memory which is required to get
performance, because the speed at which you can run these applications
depends on how quickly you can read that into memory. There’s a finite
supply, it’s only for data centers, so they can’t reach into the supply
for mobile or other things, like you can with other parts. Also Nvidia
is the largest buyer of super caps in the world and all sorts of other
components. The 400 gigabit cables, they’ve bought them all out. So if
you want to compete, it doesn’t matter how good of a product you design,
they’ve bought out the entire supply chain for years.
The biggest difference between training and inference is when you
are training, the number of tokens that you are training on is measured
in month, like how many tokens can we train on this month. In inference,
what matters is how many tokens you can generate per millisecond or a
couple milliseconds.
It’s fair to say that Nvida is the exemplar in training but
really isn’t yet the equivalent scaled winner in inference.
In order to get the latency down, we had to design a completely
new chip architecture, we had to design a completely new networking
architecture, an entirely new system, an entirely new runtime, an
entirely new compiler, and entirely new orchestration layer. We had to
throw everything away and it had to be compatible with PyTorch and what
other people actually developing in.
I think Facebook announced that by the end of this year, they are
going to have the equivalent of 650000 H100s. By the end of this year,
Grok will have deployed 100000 of our LPUs which do outperform the H100s
on a throughput and on a latency basis. So we will probably get pretty
close to the equivalent of Meta ourselves. By the end of next year, we
are going to deploy 1.5M LPUs, for comparison, last year Nvidia deployed
a total of 500000 H100s. So 1.5M means Grok will probably have more
inference GenAI capacity than all of the hyperscalers and clouds service
providers combined. So probably about 50% of the inference compute in
the world.
I get asked a lot should we be afraid of AI and my answer to that
is, if you think back to Galileo, someone who got in a lot of trouble.
The reason he got in trouble was he invented the telescope, popularized
it, and made some claims that we were much smaller than everyone wanted
to believe. The better the telescope got the more obvious it became that
we were small. In a large sense, LLMs are the telescope for the mind,
it’s become clear that intelligence is larger than we are and it makes
us feel really really small and it’s scary. But what happened over time
was as we realized the universe was larger than we thought and we got
used to that, we started to realize how beautiful it was and our place
in the universe. And I think that’s what’s going to happen. We’re going
to realize intelligence is more vast than we ever imagined. And we are
going to understand our place in it, and we are not going to be afraid
of it.
This is a very insightful conversation especially in the part of
comparison of training and inference. The answer to the final question
is fascinating to end the conversation. A great takeaway that
“intelligence is a telescope for the mind, in that we realize that we
are small, while then also opportunity to see intelligence is vast and
to not be afraid of it.”.
Meta Announces Llama 3 at Weights & Biases’ Conference -
Weights & Biases [Link]
Economic value is getting disintegrated, there no value in
foundational models economically. So then the question is who can build
on top of them the fastest. Llama was announced last Thursday, 14 hours
later Groq actually had that model deployed in the Groq Cloud, so that
100K+ developers could start building on it. That’s why that model is so
popular. It puts the closed models on their heels. Because if you can’t
both train and deploy iteratively and quickly enough, these open source
alternatives will win, and as a result the economic potential that you
have to monetize those models will not be there. - Chamath
Palihapitiya
By open-sourcing these models they limit competition because VCs
are no longer going to plow half a billion dollars into a foundational
model development company, so you limit the commercial interest and the
commercial value of foundational models. - David Friedberg
AI is really two markets - training and inference. And inference
is going to be 100 times bigger than training. And Nvidia is really good
at training and very miscast at inference. The problem is that right now
we need to see a capex build cycle for inference, and there are so many
cheap and effective solutions, Groq being one of them but there are many
others. And I think why the market reacted very negatively was that it
did not seem that Facebook understood that distinction, that they were
way overspending and trying to allocate a bunch of GPU capacity towards
inference that didn’t make sense. - Chamath Palihapitiya
You want to find real durable moats not these like legal
arrangement that try to protect your business through these types of
contracts. One of the reasons why the industry moves so fast is best
practices get shared very quickly, and one of the ways that happens is
that everybody is moving around to different companies (average term of
employment is 18-36 months). There are people who violate those rules
(taking code to the new company etc), and that is definitely breaking
the rules, but you are allowed to take with you anything in your head,
and it is one of the ways that best practices sort of become more
common. - David Sacks
Are LLMs Hitting A Wall, Microsoft & Alphabet Save The
Market, TikTok Ban - Big Technology Podcast [Link]
Papers and Reports
ReALM: Reference Resolution As Language ModelingLink]
Apple proposed the ReALM model with 80M, 250M, 1B, and 3B parameters.
It can be used on mobile devices and laptops. The task of ReALM is
“Given relevant entities and a task the user wants to perform, we wish
to extract the entity (or entities) that are pertinent to the current
user query. “. The relevant entities can be on-screen entities,
conversational entities, and background entities. The analysis shows
that ReALM beats MARRs and has similar performance with GPT-4.
Bigger is not Always Better: Scaling Properties of Latent
Diffusion Models [Link]
Leave No Context Behind: Efficient Infinite Context
Transformers with Infini-attention [Link]
Google introduced the next generation transformer -
infini-transformer. It’s able to take infinite length of input without
the requirement of more memory or computation. Unlike vanilla attention
mechanism in traditional transformer which reset their attention memory
after each context window to manage new data, infini-attention retains a
compressive memory and builds in both masked local attention and
long-term linear attention mechanisms. The model compresses and reuses
key-value states across all segments, allowing it to pull relevant
information from any part of the document.
AI agents are starting to transcend their digital origins and
enter the physical world through devices like smartphones, smart
glasses, and robots. These technologies are typically used by
individuals who are not AI experts. To effectively assist them, Embodied
AI (EAI) agents must possess a natural language interface and a type of
“common sense” rooted in human-like perception and understanding of the
world.
OpenEQA: Embodies Question Answering in the Era of Foundation
Models [Link] [Link]
The OpenEQA introduced by Meta is the first open vocab benchmark
dataset for the formulation of Embodied Question Answering (EQA) task of
understanding environment either by memory or by active exploration,
well enough to answer questions in natural language. Meta also provided
an automatic LLM-powered evaluation protocol to evaluate the performance
of SOTA models like GPT-4V and see whether it’s close to human-level
performance.
OpenEQA looks like the very first step towards “world model” and I’m
excited that it’s coming. The dataset contains over 1600 high-quality
human generated questions drawn from over 180 real-world environments.
If the future AI agent can answer N questions over N real-world
environments, where N is approximately infinity, we can call it God
intelligence. But we are probably not able to achieve that “world model”
at least with my limited imagination, because it requires un-infinite
compute resources and there can be ethical issues. However, if we take
one step back, instead of creating “world model”, a “human society
model” or “transformation model”, etc, sounds more possible. Limiting
question to a specific pain point problem and limiting environment
according to it would both save resources and contribute AI’s value to
human society.
OpenELM: An Efficient Language Model Family with Open-source
Training and Inference Framework [Link] [Link]
OpenELM is a small language model (SLM) tailored for on-device
applications. The models range from 270M to 3B parameters, which are
suitable for deployment on mobile devices and PCs. The key innovation is
called “layer-wise scaling architecture”. It allocates fewer parameters
to the initial transformer layers and gradually increases the number of
parameters towards the final layers. This approach optimizes compute
resources while remaining high accuracy. Inference of OpenELM can be run
on Intel i9 workstation with RTX 4090 GPU and an M2 Max MacBook Pro.
Phi-3 Technical Report: A Highly Capable Language Model
Locally on Your Phone [Link] [Link]
Microsoft launched Phi-3 family including mini (3.8B), small (7B),
and medium (14B). These models are designed to run efficiently on both
mobile devices and PCs. All models leverage a transformer decoder
architecture. The performance is comparable to larger models such as
Mixtral 8x7B and GPT3.5. It supports a default of 4K context length but
is expandable to 128K through LongRope technology. The models are
trained on web data and synthetic data, using two-phase approach which
enhances both general knowledge and specialized skills (e.g. logical
reasoning), and fine tuned in specific domains. Mini (3.8B) is
especially optimized for mobile usage, requiring 1.8GB memory when
compressed to 4-bits and processing 12+ tokens per second on mobile
devices such as iPhone 14.
VASA-1: Lifelike Audio-Driven Talking Faces Generated in Real
Time [Link] [Link]
2024 Generative AI Prediction Report from CB
insights [Link]
You get paid based on how hard you are to replace.
You get paid based on how much value you deliver.
Focus on being able to produce value and money will
follow.
― Andrew Lokenauth
What he’s saying is so true - Don’t work so hard and end up losing
yourself.
There is a popular saying on Wall Street. While IPO means Initial
Public Offering, it also means “It’s Probably Overpriced” (coined by Ken
Fisher).
I don’t invest in brand-new IPOs during the first six months.
Why? Shares tend to underperform out of the gate for new public
companies and often bottom around the tail end of the lock-up period,
with anticipation of selling pressure from insiders. It’s also critical
to gain insights from the first few quarters to form an opinion about
the management team.
Do they forecast conservatively?
Do they consistently beat their guidance?
If not, it might be a sign that they are running out of steam and
may have embellished their prospects in the S-1. But we need several
quarters to understand the dynamic at play.
An analysis of Rubrik, a Microsoft-backed cybersecurity company going
public. I’ve got some opinions from the author in terms of company
performance and strategic investment.
Intel Unleashes Enterprise AI with Gaudi 3 - AI
Supremacy [Link]
Intel is a huge beneficiary of Biden’s CHIPS Act. In late March 2024,
Intel will receive up to \(\$8.5\)
billion in grants and \(\$11\) billion
in loans from the US government to produce cutting-edge
semiconductors.
US Banks: Uncertain Year - App Economy Insights [Link]
Formula 1’s recent surge in popularity and revenue isn’t simply a
product of fast cars and daring drivers. The Netflix docuseries Drive to
Survive, which premiered in March 2019 and is already in its sixth
season, has played a transformative role in igniting global interest and
fueling unprecedented growth for the sport.
The docuseries effectively humanized the sport, attracting new
fans drawn to the high-stakes competition, team rivalries, and
compelling personal narratives.
― Formula 1 Economics - App Economy Insights [Link]
Recent business highlights: 1) focus on drama and storylines around
sports, 2) subscribers can download exclusive games on the App Store for
free, since Nov 2021, and Netflix is exploring game monetization through
in-app purchases or ads, 3) for Premium Subscription Video on Demand,
churn plummets YoY, 4) the \(\$6.99\)/month ad-supported plan was
launched in Nov 2023, memberships grew 65% QoQ and monetization is still
lagging, 5) started limiting password sharing within one household.
Boeing 737 MAX’s two fatal crashes due to faulty software have eroded
public trust. In addition to quality issues, Boeing is facing severe
production delays. Airbus on the other hand has captured significant
market share from Boeing. Airbus is heavily investing in technologies
such as hydrogen-powered aircraft and sustainable aviation fuels. Airbus
is also investing in the A321XLR and potential new widebody
aircraft.
We disagree on what open-source AI should mean -
Interconnects [Link]
This is a general trend we have observed a couple of years ago.
We called is Mosaic’s Law where a model of a certain
capability will require 1/4th the dollars every year from hw/sw/algo
advances. This means something that is \(\$100\)m today -> \(\$25\)m next year -> \(\$6\)m in 2 yrs -> \(\$1.5\)m in 3 yrs.― Naveen Rao on
X [Link]
DBRX: The new best open model and Databricks’ ML strategy -
Interconnects [Link]
In the test of refusals, it shows that the inference system seems to
contain an added filtering in the loop to refuse illegal requests.
Llama 3: Scaling open LLMs to AGI - Interconnects
[Link]
Q1 FY24 is bad and probably the worst. This means Tesla is going to
get better in the rest of the year. It sounds that Elon is more clear
and focused on his plan. And promises are met though there are some
delays [Master Plan,
Part Deux].
Recent business highlights: 1) cancelling Model 2 and focusing on
Robotaxis and next-gen platform (Redwood), 2) laying off 10%+, 3) FSD
price cuts and EV (Model 3 and Model Y) price cuts, 4) Recall Cybertruck
due to safety issues, 5) North American Charging Standard (NACS) is
increasingly adopted by major automakers, 6) reached ~1.2B miles driven
by FSD beta, 7) energy storage deployment increased sequentially.
What is competitive in the market: 1) competitive pressure from BYD,
2) OpenAI’s Figure 01 robot and Boston Dynamics’s next-gen Atlas are
competing with Optimus.
With its open-source AI model Llama, Meta learned that the
company doesn’t have to have the best models — but they need a lot of
them. The content creation potential benefits Meta’s platforms, even if
the models aren’t exclusively theirs.
Like Google with Android, Meta aims to build a platform to avoid
being at the mercy of Apple or Google’s ecosystems. It’s a defensive
strategy to protect their advertising business. The shift to a new
vision-based computing experience is an opportunity to do so.
Meta has a head start in the VR developer community compared to
Apple. A more open app model could solidify this advantage.
By now, Meta has a clear playbook for new products:
Release an early version to a limited audience.
Gather feedback and start improving it.
Make it available to more people.
Scale and refine.
Monetize.
He also shared some interesting nuggets:
Roughly 30% of Facebook posts are AI-recommended.
Over 50% of Instagram content is AI-recommended.
― Meta: The Anti-Apple - App Economy Insights [Link]
Recent business highlights: 1) announced an open model for Horizon OS
- which powers its VR headsets, 2) Meta AI is now powered by Llama 3, 3)
whether not TikTok will still exist in US does not matter since the
algorithm will not be sold, then it will benefit any competitor company
such as Meta.
Math is good at optimizing a known system; humans are good at
finding a new one. Put another way, change favors local maxima;
innovation favors global disruption.
― “Lean Analytics, Use Data to Build a Better Startup
Faster”
Sometimes people who are doing hard data work may forget to step back
and look at the big picture. This is a common mistake because we can
definitely go from scientific data analysis to actionable insight for
making better business decision. But we need to have some additional
thoughts about whether the decision is a global optima or it’s just
local due to the limited sample, restricted project goal, or restricted
team scope.
Articles
When I decided to end my voluntary immersion in the driver
community, I could not shake the feeling that the depersonalization of
app workers is a feature, not a bug, of an economic model born of and
emboldened by transformations that are underway across the global
economy. This includes increasingly prevalent work arrangements
characterized by weak employer-worker relations (independent
contracting), strong reliance on technology (algorithmic management,
platform-mediated communication), and social isolation (no coworkers and
limited customer interactions).
As forces continue to erode traditional forms of identity
support, meaningful selfdefinition at work will increasingly rely on how
we collectively use and misuse innovative technologies and business
models. For example, how can companies deploy algorithmic management in a
way that doesn’t threaten and depersonalize workers? How can focusing on
the narratives that underlie and animate identities help workers
reimagine what they really want and deserve out of a career coming out
of the pandemic and the Great Resignation? Will increasingly immersive
and realistic digital environments like the metaverse function as
identity playgrounds for workers in the future? How will Web3 broadly,
and the emergence of novel forms of organizing specifically (e.g.,
decentralized autonomous organizations or DAOs), affect the careers,
connections, and causes that are so important to workers? What role can
social media platforms, online discussion forums, and other types of
virtual water coolers play in helping independent workers craft and
sustain a desirable work identity? In short, how can we retain the human
element in the face of increasingly shrewd resource management
tactics?”
― “Dehumanization is a Feature of Gig Work, Not a Bug”,
Harvard Business Review, The Year in Tech 2024 [Article]
This reminds me that last year when I was on vacation in LA, I’ve
talked to a driver worked for both Lyft and Uber in LA. He complained
Lyft’s route recommendation algorithm is shitty, not helpful at all, a
waste of time, while Uber is better in comparison. At that time I
realized how important it is to strengthen employer worker relations and
gather feedback from workers or clients on the product improvement. This
is a great article where the author raises his concerns of
dehumanization of workers in the future. While technology is advancing
and economy is transforming, we don’t expect people to forget who they
are in their daily basis work.
Bringing a new technology to market presents a chicken-or-egg
problem: The product needs a supply of complementary offerings, but the
suppliers and complementors don’t exist yet, and no entrepreneur wants
to throw their lot in with a technology that isn’t on the market
yet.
There are two ways of “solving” this problem. First, you can time
the market, and wait until the ecosystem matures— though you risk
waiting a long time. Second, you can drive the market, or supply all the
necessary inputs and complements yourself.
― “Does Elon Musk Have a Strategy?”, Harvard Business Review,
The Year in Tech 2024 [Article]
Here are two examples. To drive the market, Musk supplies both
electric vehicles and charging stations. Zuckerberg proposed the concept
of metaverse and changed his company’s name.
This is where Musk’s Wall Street critics might say he’s weakest.
Many of his businesses don’t articulate a clear logic, which is
demonstrated by the unpredictable way these businesses ultimately reach
solutions or products.
Musk has spelled out some of his prior logic in a set of “Master
Plans,” but most of the logical basis for exactly how he will succeed
remains ambiguous. But this isn’t necessarily Musk’s fault or due to any
negligence per se: When pursuing new technologies, particularly ones
that open up a new market, there is no one who can anticipate the full
set of possibilities of what that technology will be able to do (and
what it will not be able to do).
― “Does Elon Musk Have a Strategy?”, Harvard Business Review,
The Year in Tech 2024 [Article]
Elon is interesting, but I have to say that we human need this type
of person to leap to the future.
What could he possibly want with Twitter? The thing is, over the
last decade, the technological landscape has changed, and how and when
to moderate speech has become a critical problem-and an existential
problem for social media companies. In other words, moderating speech
has looked more and more like the kind of big, complex strategic problem
that captures Musk’s interest.
― “Does Elon Musk Have a Strategy?”, Harvard Business Review,
The Year in Tech 2024 [Article]
This is a great article which profiles Elon Musk specifically in his
strategies and vision. It answered my question confusing me for two
years: what was Musk thinking on buying Twitter? The answer is: Musk’s
vision is not in pursuit of a specific type of solution but is in
pursuit of a specific type of problem. If we go back to 2016, Igor as a
CEO of Disney decided not to buy Twitter because he looked at Twitter as
the solution: a global distribution platform, while concerned the
quality of speech on it is a problem. Musk was looking for challenges
and complexities while Igor was preventing them and looking for
solutions.
YouTube
“One of the things that I think OpenAI is doing that is the most
important of everything that we are doing is putting powerful technology
in the hands of people for free as a public good. We don’t run ads on
our free version. We don’t monetize it in other ways. I think that kind
of ‘open’ is very important, and is a huge deal for how we fulfill the
mission. “― Sam
“For active learning, the thing is it truly needs a problem. It
needs a problem that requires it. It is very hard to do research about
the capability of active learning if you don’t have a task. You will
come up with an artificial task, get good results, but not really
convince anyone. “, “Active learning will actually arrive with the
problem that requires it to pop up.”― Ilya
“To build an AGI, I think it’s going to be Deep Learning plus
some ideas, and self-play would be one of these ideas. Self-play has
such properties that can surprise us in truly novel ways. Almost all
self-play systems produce surprising behaviors that we didn’t expect.
They are creating solutions to problems.”, “Not just random surprise but
to find the surprising solution to a problem.”― Ilya
“Transferring from simulation to the real world is definitely
possible and it’s been exhibited many times by many different groups.
It’s been especially successful in vision. Also OpenAI in the summer has
demonstrated robot hand which was trained entirely in simulation. “,
“The policy that it learned in simulation was trained to be very
adaptive. So adaptive that when you transfer if could very quickly adapt
to the physical world.”― Ilya
“The real world that I would imagine is one where humanity are
like the board members of a company where the AGI is the CEO. The
picture I would imagine is you have some kind of different entities,
countries or cities, and the people who live there vote for what the AGI
that represents them should do. You could have multiple AGIs, you would
have an AGI for a city, for a country, and it would be trying to in
effects take the democratic process to the next level.” “(And the board
can always fire the CEO), press the reset button, re-randomize the
parameters.”― Ilya
“It’s definitely possible to build AI system which will want to
be controlled by their humans.”, “It will be possible to program an AGI
to design it in such a way that it will have a similar deep drive that
it will be delighted to fulfill, and the drive will be to help humans
flourish.”― Ilya
“I don’t know if most people are good. I think that when it
really counts, people can be better than we think.”― Ilya
Sam Altman: OpenAI, GPT-5, Sora, Board Saga, Elon Musk, Ilya,
Power & AGI | Lex Fridman Podcast [Sam]
Ilya Sutskever: Deep Learning | Lex Fridman Podcast
[Ilya]
I watched Lex’s interview with Sam Altman uploaded on March 18, 2024,
and an older interview with Ilya Sutskever happened 3 years ago. Elon’s
lawsuit against OpenAI frustrated Sam but Sam is optimistic about the
future and everything he is going to release in the next few months. Sam
answered the questions “what does open mean in OpenAI” that ‘open’
mainly means putting powerful tech in the hands of people for free as a
public good, but not necessarily mean open-source. He said there can be
open-source models or closed-source models. About the transition between
non-profit to capped for-profit, Sam said OpenAI is not setting a
precedent for startup to mimic it but he suggested most startups should
go for for-profit directly if they pursue profitability in the
beginning.
Ilya’s interview is more interesting to me because he talked a lot
about vision, tech, philosophy in Deep Learning. It’s impressive that he
had such thoughts 3 years ago.
News
Speech is one kind of liability for companies using generative
AI. The design of these systems can create other kinds of harms—by, say,
introducing bias in hiring, giving bad advice, or simply making up
information that might lead to financial damages for a person who trusts
these systems.
Because AIs can be used in so many ways, in so many industries,
it may take time to understand what their harms are in a variety of
contexts, and how best to regulate them, says Schultz.
― The AI Industry Is Steaming Toward A Legal Iceberg
[Link]
The Section 230 of Communications Decency Act of 1996 has protected
internet platforms from being held liable for the things we say on them,
but it doesn’t cover speech that a company’s AI generates. It’s likely
that in the future companies use AI will be liable for whatever it does.
It could be a driver of pushing companies to take effort to avoid
problematic AI output, and reduce “hallucinations” (when GenAI makes
stuff up).
A very obvious downside is people could use AI to do harmful things,
but I think people are good to work together and prevent that from
different angles such as legal aspect or open source software. I worry
more about things that are potential or cannot be seen at least in these
years when AI is still immature - which is being too early to rely on AI
and deviating from truth. For example, it is too soon for people to lose
faith in jobs like teachers, historians, journalists, writers, etc, but
I’m concerning people are already losing faith in those jobs because of
the development of AI and some violations of copyrighted work. As we
have seen that AI could have wrong understanding of facts, have biased
opinions, and make things up that don’t exist, we lives could fight for
the truth but the dead cannot talk.
China’s latest EV is a ‘connected’ car from smart phone and
electronics maker Xiaomi [Link]
Xiaomi started EV manufacturing since 2021 and launched its first EV
“SU7” on March 28th 2024. It has the following reasons of success: 1)
efficient technology manufacturing in a large scale. Though Xiaomi has
no experience in auto field, it is a supply chain master, and has
perfect partnership with various suppliers. 2) affordable price. SU7’s
start price is 215900 yuan while Tesla’a model 3 is 245900 yuan. 3)
customer experience oriented innovation. SU7 model can link to over 1000
Xiaomi devices as well as Apple’s devices. In addition, Xiaomi aims to
connect its cars with its phones and home appliances in a “Human x Car x
Home” ecosystem.
“The world is just now realizing how important high-speed
inference is to generative Al,” said Madra. “At Groa, we’re giving
developers the speed, low latency, and efficiency they need to deliver
on the generative Al promise. I have been a big fan of Groq since I
first met Jonathan in 2016 and I am thrilled to join him and the Groq
team in their quest to bring the fastest inference engine to the
world.”
“Separating GroqCloud and Groq Systems into two business units
will enable Groq to continue to innovate at a rapid clip, accelerate
inference, and lead the Al chip race, while the legacy providers and
other big names in Al are still trying to build a chip that can compete
with our LPU,” added Ross.
― Groq® Acquires Definitive Intelligence to Launch
GroqCloud [Link]
Al chip startup Groq acquired Definitive Intelligence to launch
GroqCloud business unit led by Definitive Intelligence’s CEO Sunny
Madra. Groq is also forming a Groq Systems business unit by infusing
engineering resources from Definitive Intelligence, which aims to
greatly expanding its customer and developer ecosystem.
Groq’s founder Janathan Ross is the inventor of the Google Tensor
Processing Unit (TPU), Google’s custom Al accelerator chip used to run
models. Groq is creating a Language Processing Unit (LPU) inference
engine, which is claimed to be able to run LLM at 10x speed. Now
GroqCloud provides customers the Groq LPU inference engine via the
self-serve playground.
The House voted to advance a bill that could get TikTok banned in
the U.S. on Wednesday. In a 352-65 vote, representatives passed the
bipartisan bill that would force ByteDance, the parent company of
TikTok, to either sell the video-sharing platform or prohibit it from
becoming available in the U.S.
― What to Know About the Bill That Could Get TikTok Banned in
the U.S. [Link]
TikTok is considered as critical threats to US national security
because it is owned by ByteDance and required to collaborate with the
Chinese Communist Party (CCP). If the bill is passed then ByteDance has
to either sell the platform within 180 days or face a ban. TikTok
informed users that Congress is planning a total ban of TikTok and
encouraged users to speak out against the ban. Shou Zi Chew said the ban
would put more than 300000 American jobs at risk.
San Francisco-based Anthropic introduced three new AI models —
Claude 3 Opus, Sonnet and Haiku. The literary names hint at the
capabilities of each model, with Opus being the most powerful and Haiku
the lightest and quickest. Opus and Sonnet are available to developers
now, while Haiku will arrive in the coming weeks, the company said on
Monday.
― AI Startup Anthropic Launches New Models for Chatbot
Claude [Link]
Waymo’s progress in California comes after General Motors-owned
Cruise and Apple bowed out of the autonomous vehicle business in
California, while Elon Musk’s Tesla has yet to develop an autonomous
vehicle that can safely operate without a human driver at the
controls.
― Waymo approved by regulator to expand robotaxi service in
Los Angeles, San Francisco Peninsula [Link]
Elon Musk requires “FSD” demo for every prospective Tesla
buyer in North America [Link]
Full Self Driving era seems to start, but Tesla’s FSD system does not
turn cars into autonomous vehicles, so drivers still need to be
attentive to the road and ready to steer or brake at any time while
using FSD or FSD Beta. Will FSD help with Tesla’s stock?
SpaceX Starship disintegrates after completing most of third
test flight [Link]
SpaceX’s Starship rocket successfully completed a repeat of stage
separation during initial ascent, open and close its payload door in
orbit, the transfer of super-cooled rocket propellant from one tank to
another during spaceflight. But it skipped Raptor engine re-ignition
test, failed re-entry to the atmosphere, and flying the rocked back to
Earth. Overall, completion of many of the objectives represented
progress in the development of spacecraft for the business and SpaceX
and NASA’s moon program.
Musk founded xAI in March 2023 aiming to “understand the true nature
of the universe”. It released the weights and network architecture of
314B Grok-1 on March 17, 2024. It’s under the Apache 2.0 license meaning
it allows for commercial use. The model can be found in Github.
GB200 has a somewhat more modest seven times the performance of
an H100, and Nvidia says it offers four times the training
speed.
Nvidia is counting on companies to buy large quantities of these
GPUs, of course, and is packaging them in larger designs, like the GB200
NVL72, which plugs 36 CPUs and 72 GPUs into a single liquid-cooled rack
for a total of 720 petaflops of AI training performance or 1,440
petaflops (aka 1.4 exaflops) of inference. It has nearly two miles of
cables inside, with 5,000 individual cables.
And of course, Nvidia is happy to offer companies the rest of the
solution, too. Here’s the DGX Superpod for DGX GB200, which combines
eight systems in one for a total of 288 CPUs, 576 GPUs, 240TB of memory,
and 11.5 exaflops of FP4 computing.
Nvidia says its systems can scale to tens of thousands of the
GB200 superchips, connected together with 800Gbps networking with its
new Quantum-X800 InfiniBand (for up to 144 connections) or Spectrum-X800
ethernet (for up to 64 connections).
― Nvidia reveals Blackwell B200 GPU, the ‘world’s most
powerful chip’ for AI [Link]
[keynote]
Two B200 GPUs combined with one Grace CPU is a GB200 Blackwell
Superchip. Two GB200 superchip is one Blackwell compute node. 18
Blackwell compute notes contain 36 CPU + 72 GPUs, becoming one larger
virtual GPU - GB200 NVL72.
Nvidia also offers packages for companies such as DGX Superpod for
DGX GB200 which combines 8 such GB200 NVL72. 8 GB200 NVL72 combined with
xx becomes one GB200 NVL72 compute rack. And the AI factory or full data
center in the future would consists about 56 GB200 NVL72 compute racks,
which is in total around 32000 GPUs.
The Blackwell superchip will be 4 times faster and 25 times energy
efficient than H100.
OpenAI is expected to release a ‘materially better’ GPT-5 for
its chatbot mid-year, sources say[Link]
On March 14 (local time), during a meeting with the Korean Silicon
Valley correspondent group, CEO Altman mentioned, “I am not sure when
GPT-5 will be released, but it will make significant progress as a model
taking a leap forward in advanced reasoning capabilities. There are many
questions about whether there are any limits to GPT, but I can
confidently say ‘no’.” He expressed confidence that if sufficient
computing resources are invested, building AGI that surpasses human
capabilities is entirely feasible.
Other news:
Elon Musk sues OpenAI for abandoning its mission to benefit
humanity [Link]
A major AT&T data leak posted to the dark web included
passcodes, Social Security numbers [Link]
Apple accused of monopolizing smartphone markets in US
antitrust lawsuit [Link]
Amazon Invests $2.75 Billion in AI Startup Anthropic
[Link]
Adam Neumann looks to buy back WeWork for more than $500M:
sources [Link]
NVIDIA Powers Japan’s ABCI-Q Supercomputer for Quantum
Research [Link]
Lilac Joins Databricks to Simplify Unstructured Data
Evaluation for Generative AI [Link]
Papers and Reports
Scaling Instructable Agents Across Many Simulated
Worlds [Link]
Google DeepMind SIMA Team is working on the Scalable, Instructable,
Multiworld Agent (SIMA) project. The goal is to develop an agent that
follows instructions to complete tasks in any 3D environments. So far
they are making progress on making AI agent understand the environment
from computer screen, and use keyboard-and-mouse controls to interact
with environment, follow language instructions, and play the video game
to maximize the win-rate.
OpenAI has similar work called OpenAI Universe, which
aims to train and validate AI agent on performing real world tasks. They
started from video game environment as well. Although the goals of these
two project sound similar, the minor difference is that OpenAI Universe
intended to develop a platform where AI is able to interact with games,
websites, and applications, while SIMA aims to develop an AI agent or
maybe a robot to interact with the real world.
Announcing HyperGAI: a New Chapter in Multimodal Gen
AI [Link]
Introducing HPT: A Groundbreaking Family of Leading
Multimodal LLMs [Link]
The startup HyperGAI aims to develop models for multimodal
understanding and multimodal generation. They released HPT air and HPT
pro. HPT pro outperforms GPT-4V and Gemini Pro on the MMbench and
SEED-Image benchmark.
Mora: Enabling Generalist Video Generation via A Multi-Agent
Framework [Link]
Sora is the first video generation model, however it is not
open-source. Lehigh University and Microsoft Research developed Mora to
address the gap of no other video generation models to parallel with
Sora in performance. Mora introduces an innovative multi-agent
framework. As a result, Mora marks a considerable advancement in video
generation from text prompts. The evaluation shows that Mora competes
with Sora on most of the tasks, but not as refined as Sora in tasks such
as changes in the video content, and video connectivity.
CrowdStrike has repeated in its investor presentations how it
wants to be the leading ‘Security Cloud’ and emulate other
category-defining cloud platforms:
Workday (HR Cloud).
Salesforce (CRM Cloud).
ServiceNow (Service Management Cloud).
Public cloud software companies are overwhelmingly unprofitable
businesses. However, in FY24, Salesforce (CRM) demonstrated that margins
can expand quickly once the focus turns to the bottom line (see visual).
And when the entire business is driven by recurring subscription revenue
and highly predictable unit economics, you are looking at a finely-tuned
cash flow machine.
Oracle services for enterprise software and cloud solutions: 1) cloud
suite (cloud applications and services), 2) data analytics, 3)
autonomous database, 4) enterprise resource planning (ERP) to improve
operational efficiencies and integrated solutions to streamline complex
business functions.
Key news highlights: 1) Oracle acquired Cerner in June 2022 which is
a leading provider of electronic health records (EHR) and other
healthcare IT solutions used by hospitals and health systems. Oracle is
expanding cloud services including the upcoming launch of Ambulatory
Clinic Cloud Application Suite for Cerner customers. 2) The adoption of
Oracle Cloud Infrastructure (OCI) are across different segments: cloud
natives customers such as Zoom, Uber, ByteDance looking for high price
performance and integrated security and privacy, AI/ML customers looking
for key differentiation, compute performance, and networking design,
generative AI customers looking for control, data security, privacy, and
governance. 3) TikTok is probably an essential component of the growth
of OCI Gen2 infrastructure cloud services. 4) Oracle signed big
Generation 2 Cloud infrastructure contract with Nvidia. 5) Oracle is a
critical customer in Sovereign AI. It’s starting to win business per
country for sovereign cloud, especially the cloud companies in
Japan.
Duolingo launched the Duolingo Max subscription tier ($168/year),
with Gen AI features enabling a more conversational and listening
approach. Duolingo has leveraged AI in two areas: 1) using AI to create
content, which allows it to experiment faster, 2) using AI to power
spoken conversation with characters.
What is coming: Duolingo launched Math and Music courses into its app
in 2023.
Read arguments between OpenAI and Elon. Learned that Elon once
believed there is 0% probability for OpenAI to succeed and wanted OpenAI
to become for-profit so it can be merged to Tesla and being controlled
by Elon himself.
We know from our past experiences that big things start small.
The biggest oak starts from an acorn. If you want to do anything new,
you’ve got to be willing to let that acorn grow into a little sapling
and then into a small tree and maybe one day it will be a big business
on its own.
He was a free thinker whose ideas would often run against the
conventional wisdom of any community in which he operated.
I’ve always actually found something to be very true, which is
most people don’t get those experiences because they never ask. I have
never found anybody who didn’t want to help me when I’ve asked them for
help. I have never found anyone who said no or hung up the phone when I
called. I just asked. And when people ask me, I try to be as responsive,
to pay back that debt of gratitude. Most people never pick up the phone
and call. Most people never ask. That is what separates the people that
do things from the people that just dream about them. You’ve got to act
and you’ve got to be willing to fail. You’ve got to be ready to crash
and burn with people on the phone, with starting a company, with
whatever. If you’re afraid of failing, you won’t get very far.
His company and its computer into something aspirational. He
links this machine made a few months earlier, a few months ago by some
disheveled California misfits to Rolls Royce, the 73 year old paragon of
sophisticated industrial manufacturing and elite consumer taste. He even
calls Apple a world leader, an absolutely unprovable claim that rockets
the little company into the same league as IBM, which was then the
industry’s giant. He was an extraordinary speaker and he wielded that
tool to great effect.
People that are learning machines and they refuse to quit are
incredibly hard to beat.
When you have something that’s working, you do not talk about it.
You shut up because the more you talk about it, the more broadcasting
you do about it, the more it encourages competition.
The only purpose for me in building a company is so that the
company can make products. One is a means to the other. Over a period of
time, you realize that building a very strong company and a very strong
foundation of talent and culture in a company is essential to keep
making great products. The company’s one of the most amazing inventions
of humans, this abstract construct that’s incredibly powerful. Even so,
for me, it’s about the products. It’s about working together with really
fun, smart, creative people and making wonderful things. It is not about
the money. What a company is, then, is a group of people who can make
more than just the next big thing. It is a talent. It is a capability.
It is a culture. It is a point of view. And it is a way of working
together to make the next thing and the next one and the next
one.
In that case, Steve would check it out, and the information he’d
glean would go into the learning machine that was his brain. Sometimes,
that’s where it would sit and nothing would happen. Sometimes, on the
other hand, he’d concoct a way to combine it with something else that
he’d seen or perhaps to twist it in a way to benefit an entirely
different project altogether. This was one of his great talents, the
ability to synthesize separate developments and technologies into
something previously unimaginable.
I felt I had let the previous generations of entrepreneurs down,
that I had dropped the baton as it was being passed to me. I met with
David Packard and Bob Noyce, and tried to apologize for screwing up so
badly. I was a very public failure and even thought about running away
from the Valley. But something slowly began to dawn on me. I still love
what I did. The turn of events at Apple had not changed that one bit. I
had been rejected, but I was still in love, and so I decided to start
over.
― Founders #265 Becoming Steve Jobs: The Evolution of a
Reckless Upstart into a Visionary Leader [Link]
It helps if you can be satisfied with an inner scorecard, I would
also say it’s probably the only – the single only way to have a happy
life.
I wanted money. It could make me independent then I could do with
what I wanted to do with my life. And the biggest thing I wanted to do
was work for myself. I didn’t want other people directing me. The idea
of doing what I wanted to do every day was very important to
me.
I like to work by myself where I could spend my time thinking
about things I wanted to think about. Washington was upsetting at first,
but I was in my own world all the time. I could be sitting in a room
thinking or could be writing around flinging things and
thinking.
Walt Disney seldom dabbled. Everyone who knew him remarked on his
intensity. When something intrigued him, he focused himself entirely on
it as if it were the only thing that mattered.
Intensity is the price of excellence.
People ask me where they should go to work, and I always tell
them to go work for whom they most admire.
That’s like saving sex for your old age, do what you love and
work for whom you admire the most, and you’ve given yourself the best
chance in life you can.
You’ll get very rich if you thought of yourself as having a card
with only 20 punches in a lifetime, and every financial decision used up
one punch. You will resist the temptation to dabble. You make more good
decisions, and you would make more big decisions.
Instead, he said, basically, when you get to my age, you’ll
really measure your success in life by how many of the people you want
to have love you actually do love you. I know people who have a lot of
money, and they get testimonial dinners and they get hospital wings
named after them. But the truth is that nobody in the world loves them.
If you get to my age and life and nobody thinks well of you, I don’t
care how big your bank account is.
Your life is a disaster. That’s the ultimate test of how you’ve
lived your life. The trouble with love is you can’t buy it. You can buy
sex. You can buy testimonial dinners. You can buy pamphlets that say how
wonderful you are, but the only way to get love is to be lovable. It is
very irritating if you have a lot of money. You’d like to think you
could write a check. I’ll buy $1 million worth of love, please, but it
doesn’t work that way.
The biggest threat to dynastic family continuity was enrichment
and success.
Almost all of the dynasties started as outsiders.
Those on the margins often come to control the center.
Great industrial leaders are always fanatically committed to
their jobs. They are not lazy or amateurs.
A man always has two reasons for the things he does, a good one
and the real one.
Do it yourself, insist on quality, make something that will
benefit society, and pick a mission that is bigger than
yourself.
It is impossible to create an innovative product, unless you do
it yourself, pay attention to every detail, and then to test it
exhaustively. Never entrust your creation of a product to others, for
that will inevitably lead to failure and cause you deep regret.
― Founders #307 The World’s Great Family Dynasties:
Rockefeller, Rothschild, Morgan, & Toyada [Link]
Amazon’s single-threaded leadership: “The basic premise is that
for each project, there is a single leader whose focus is that project
and that project alone. And that leader oversees teams of people whose
attention is similarly focused on that one project.”
Similar idea in Peter Thiel’s book Zero to One: “The best thing I
did as a manager at PayPal was to make every person in the company
responsible for doing just one thing. Every employee’s one thing was
unique, and everyone knew I would evaluate him only on that one thing. I
had started doing this just to simplify the task of managing people, but
then I noticed a deeper result. Defining roles reduced
conflict.”
“When your dependencies keep growing, it’s only natural to try
speeding things up by improving your communication. We finally realize
that all of this cross-team communication didn’t really need refinement
at all. It needed to be eliminated. It wasn’t just that we had the wrong
solution in mind. Rather, we’ve been trying to solve the wrong problem
altogether.”
Jeff’s vision was that we needed to focus on loosely coupled
interaction via machines through well-defined APIs rather than via
humans through e-mails and meetings. This would free each team to act
autonomously and move faster.
From his 2016 shareholder letter, Jeff suggested that most
decisions should probably be made with somewhere around 70% of the
information you wish you had. If you wait for 90%, in most cases, you’re
probably being slow. Plus, either way, you need to be good at quickly
recognizing and correcting bad decisions. If you’re good at course
correcting, being wrong, may be less costly than you think, whereas
being slow is going to be expensive for sure.
“The best way to fail and inventing something is by making it
somebody’s part-time job. And so the problem that they were trying to
solve and the vision they had was how to move faster and remove
dependencies, but what they also realized once this was in place, that
ownership and accountability are much easier to establish under the
single-threaded leader model.”
“Most large organizations embrace the idea of invention but are
not willing to suffer the string of failed experiments necessary to get
there.” “Long-term thinking levers are existing abilities and lets us do
new things we couldn’t otherwise contemplate. Long-term orientation
interacts well with customer obsession. If we can identify a customer
need and if we can further develop conviction that the need is
meaningful and durable, our approach permits us to work patiently for
multiple years to deliver a solution.”
Invention works well where differentiation matters.
Differentiation with customers is often one of the key reasons to
invent.
Working backwards exposes skill sets that your company needs but
does not have yet. So the longer that your company works backwards, the
more skills it develops and the more skills it develops, the more
valuable it becomes over time.
Founders force the issue. Not outsourcing means it’s going to be
more expensive, going to spend a lot of or money. It’s going to take
longer to get a product out there. But at the end of that, if we are
successful, we have a set of skills that we lacked beforehand, then we
can go out and do this over and over again.
“My passion has been to build an enduring company where people
were motivated to make great products. Everything else was secondary.
Sure, it was great to make a profit because that’s what allowed you to
make great products. But the products, not the profits, were the
motivation. Sculley flipped these priorities to where the goal was to
make money.”
“It’s a subtle difference, but it ends up meaning everything, the
people you hire, who gets promoted, what you discuss in meetings. Some
people say, give the customer what they want, but that’s not my
approach. Our job is to figure out what they’re going to want before
they do. I think Henry Ford once said, ‘If I asked customers what they
wanted, they would have told me, a faster horse.’ People don’t know what
they want until you show it to them. That’s why I never rely on market
research. Our task is to read things that are not yet on the page. Edwin
Land of Polaroid talked about the intersection of the humanities and
science. I like that intersection. There’s something magical about that
place.”
“There are a lot of people innovating, and that’s not the main
distinction of my career. The reason Apple resonates with people is that
there’s a deep current of humanity in our innovation. I think great
artists and great engineers are similar in that they both have a desire
to express themselves. In fact, some of the best people working on the
original Mac were poets and musicians on the side.”
“In the ‘70s, computers became a way for people to express their
creativity. Great artists like Leonardo da Vinci and Michelangelo were
also great at science. Michelangelo knew a lot about how to quarry
stone, not just how to be a sculptor. At different times in the past,
there were companies that exemplified Silicon Valley. It was
Hewlett-Packard for a long time. Then in the semiconductor era, it was
Fairchild and Intel. I think that it was Apple for a while, and then
that faded. And then today, I think it’s Apple and Google and a little
more so Apple. I think Apple has stood the test of time. It’s been
around for a while, but it’s still at the cutting edge of what’s going
on.”
“It’s easy to throw stones at Microsoft, and yet I appreciate
what they did and how hard it was. They were very good at the business
side of things. They were never as ambitious product-wise as they should
have been. Bill likes to portray himself as a man of the product, but
he’s really not. He’s a businessperson. Winning business was more
important than making great products. He ended up the wealthiest guy
around. And if that was his goal, then he achieved it. But it’s never
been my goal. And I wonder in the end if it was his goal.”
“I admire him for the company he built. It’s impressive, and I
enjoyed working with him. He’s bright and actually has a good sense of
humor. But Microsoft never had the humanities and liberal arts in its
DNA. Even when they saw the Mac, they couldn’t copy it well. They
totally didn’t get it. I have my own theory about why decline happens at
companies. The company does a great job, innovates and becomes a
monopoly or close to it in some field. And then the quality of the
product becomes less important. The company starts valuing great
salesmen because they’re the ones who can move the needle on revenues,
not the product engineers and designers.”
“So the salespeople end up running the company. When the sales
guys run the company, the product guys don’t matter so much, and a lot
of them just turn off. It happened at Apple when Sculley came in, which
was my fault. Apple was lucky, and it rebounded. I hate it when people
call themselves entrepreneurs when what they’re really trying to do is
launch a startup and then sell or go public so they can cash in and move
on. They’re unwilling to do the work it takes to build a real company,
which is the hardest work in business. That is how you really make a
contribution and add to the legacy of those who went before.”
“You build a company that will stand for something a generation
or two from now. That’s what Walt Disney did and Hewlett and Packard and
the people who built Intel. They created a company to last, not just to
make money. That’s what I want Apple to be. I don’t think I run
roughshod over people. But if something sucks, I tell people to their
face. It is my job to be honest. I know what I’m talking about, and I
usually turn out to be right. That’s the culture I try to create. We are
brutally honest with each other, and anyone can tell me they think I’m
full of s***, and I can tell them the same.”
“And we’ve had some rip-roaring arguments where we were yelling
at each other and it’s some of the best times I’ve ever had. I feel
totally comfortable saying, ‘Ron, that story looks like s,’ in front of
everyone else. Or I might say, ‘God, we really fed up the engineering on
this,’ in front of the person that’s responsible. That’s the ante for
being in the room. You’ve got to be able to be super honest. Maybe
there’s a better way, a gentlemen’s club, where we all wear ties and
speak in soft language and velvet code words. But I don’t know that way
because I’m middle class from California.”
“I was hard on people sometimes, probably harder than I needed to
be. I remember the time when my son was six years old, coming home, I
had just fired somebody that day. And I imagined what it was like for
that person to tell his family and his young son that he had lost his
job. It was hard, but somebody has got to do it. I figured that it was
always my job to make sure that the team was excellent. And if I didn’t
do it, nobody was going to do it. You always have to keep pushing to
innovate.”
“Bob Dylan could have sung protest songs forever and probably
made a lot of money, but he didn’t. He had to move on. And when he did,
by going electric in 1965, he alienated a lot of people. His 1966 Europe
tour was his greatest. He would come on and do a set of acoustic guitars
and the audience loved him. Then he would do an electric set and the
audience booed. There was one point where he was about to sing Like a
Rolling Stone and someone from the audience yells, “Judas,” and Dylan
says, ‘Play it f***ing loud,’ and they did. The Beatles were the same
way. They kept evolving, moving, refining their art. That is what I’ve
always tried to do. Keep moving. Otherwise, as Dylan says, ‘If you’re
not busy being born, you’re busy dying.’”
“What drove me? I think most creative people want to express
appreciation for being able to take advantage of the work that’s been
done by others before us. I didn’t invent the language or mathematics I
use. I make little of my own food, none of my own clothes. Everything I
do depends on other members of our species and the shoulders that we
stand on. And a lot of us want to contribute something back to our
species and to add something to that flow. It’s about trying to express
something in the only way that most of us know how. We try to use the
talents we do have to express our deep feelings, to show our
appreciation of all the contributions that came before us, and to add
something to that flow. That is what has driven me.”
“He was not a model boss or human being, tightly packaged for
emulation. Driven by demons, he would drive those around him to fury and
despair. But his personality and passions and products were all
interrelated. His tale is thus both instructive and cautionary, filled
with lessons about innovation, character, leadership, and
values.”
“I don’t focus too much on being pragmatic. Logical thinking has
its place but really go on intuition and emotion. I began to realize
that an intuitive understanding and consciousness was far more
significant than abstract thinking and intellectual logical
analysis.”
“Whatever he was interested in, he would generally carry to an
irrational extreme.”
Charlie Munger says, “In business, we often find that the winning
system goes almost ridiculously far in maximizing or minimizing one or a
few variables.”
“He made me do something I didn’t think I could do. It was the
brighter side of what would become known as his reality distortion
field. If you trust him, you can do things,” Holmes said. “If he decided
that something should happen, then he’s just going to make it
happen.”
“I taught him that if you act like you can do something, then it
will work. I told him, pretend to be completely in control and people
will assume that you are.”
“Jobs had a bravado that helped him get things done, occasionally
by manipulating people. He could be charismatic, even mesmerizing, but
also cold and brutal. Jobs was awed by Wozniak’s engineering wizardry
and Wozniak was awed by Jobs’ business strive. I never wanted to deal
with people and step on toes. But Steve could call up people he didn’t
know and make them do things.”
“In order to do a good job of those things that we decide to do,
we must eliminate all the unimportant opportunities.”
“The world is a very malleable place. If you know what you want
and you go forward with maximum energy and drive and passion, the world
will often reconfigure itself around you much more quickly and easily
than you would think.”
The reality distortion field was a confounding combination of a
charismatic rhetorical style, indomitable will and an eagerness to bend
any fact to fit the purpose at hand.
“Jobs is a strong world elitist artist, who doesn’t want his
creations mutated inauspiciously by unworthy programmers. It would be as
if someone off the street added some brush strokes to a Picasso painting
or changed the lyrics to a Bob Dylan song.”
“If you want to live your life in a creative way, you have to not
look back too much. You have to be willing to take whatever you’ve done
and whoever you were and throw them away. The more the outside world
tries to reinforce an image of you, the harder it is to continue to be
an artist, which is why a lot of times artists have to say, ‘Bye, I have
to go. I’m going crazy, and I’m getting out of here.’ And then they go
hybrid somewhere. Maybe later, they reemerge a little
differently.”
― Founders #214 Steve Jobs: The Exclusive Biography
[Link]
Articles
The qubit in superposition has some probability of being 1 or 0,
but it represents neither state, just like our quarter flipping into the
air is neither heads nor tails, but some probability of both. A quantum
computer can use a collection of qubits in superpositions to play with
different possible paths through a calculation. If done correctly, the
pointers to incorrect paths cancel out, leaving the correct answer when
the qubits are read out as Os and 1s.
Grover’s algorithm, a famous quantum search algorithm, could find
you in a phone book of 100 million names with just 10,000 operations. If
a classical search algorithm just spooled through all the listings to
find you, it would require 50 million operations, on average.
Qubits have to be carefully shielded, and operated at very cold
temperatures-sometimes only fractions of a degree above absolute zero. A
major area of research involves developing algorithms for a quantum
computer to correct its own errors, caused by glitching qubits.
Some researchers, most notably at Microsoft, hope to sidestep
this challenge by developing a type of qubit out of clusters of
electrons known as a topological qubit. Physicists predict topological
qubits to be more robust to environmental noise and thus less
error-prone, but so far they’ve struggled to make even one.
Teams in both the public and private sector are betting so, as
Google, IBM, Intel, and Microsoft have all expanded their teams working
on the technology, with a growing swarm of startups such as Xanadu and
QuEra in hot pursuit. The US, China, and the European Union each have
new programs measured in the billions of dollars to stimulate quantum
R&D. Some startups, such as Rigetti and lonQ, have even begun
trading publicly on the stock market by merging with a so-called
special-purpose acquisition company, or SPAC-a trick to quickly gain
access to cash.
Chemistry simulations may be the first practical use for these
prototype machines, as researchers are figuring out how to make their
qubits interact like electrons in a molecule. Daimler and Volkswagen
have both started investigating quantum computing as a way to improve
battery chemistry for electric vehicles. Microsoft says other uses could
include designing new catalysts to make industrial processes less energy
intensive, or even pulling carbon dioxide out of the atmosphere to
mitigate climate change. Tech companies like Google are also betting
that quantum computers can make artificial intelligence more
powerful.
Big Tech companies argue that programmers need to get ready now.
Google, IBM, and Microsoft have all released open source tools to help
coders familiarize themselves with writing programs for quantum
hardware. IBM offers online access to some of its quantum processors, so
anyone can experiment with them. Launched in 2019, Amazon Web Services
offers a service that connects users to startup-built quantum computers
made of various qubit types over the cloud. In 2020, the US government
launched an initiative to develop a K-12 curriculum relating to quantum
computing. That same year, the University of New South Wales in
Australia offered the world’s first bachelor’s degree in quantum
engineering.
This article is pretty comprehensive in describing quantum computing
mechanism and techniques. One interesting fact is that quantum computers
are on the verge of breaking into bank accounts and breaking encryption
and cryptography. Shor’s algorithm has been proven mathematically that
if you had a large enough quantum computer, you could find the prime
factor of large numbers - the basis of RSA encryption, the most commonly
used thing on the internet. Although we are far away from being able to
have a quantum computer big enough to execute Shor’s algorithm on that
scale, cryptography research has already been preparing for quantum
computers’ code-breaking capabilities.
News
Neuralink’s brain-computer interface, or BCI, would allow people
to control a computer or mobile device wirelessly “just by thinking
about it,” according to the company’s website.
The goal of the new technology is to allow paralyzed people the
ability to control a computer cursor or keyboard using just their
thoughts.
Beyond helping paralyzed patients regain some mobility and
communicate without typing, Neuralink’s longer-term goals include
helping restore full mobility and sight.
― First human to receive Neuralink brain implant is
‘recovering well,’ Elon Musk says [Link]
Biderman notes that the leak is likely harmful in terms of
reducing trust between companies like Meta and the academics they share
their research with. “If we don’t respect people’s good faith attempts
to disseminate technology in ways that are consistent with their legal
and ethical obligations, that’s only going to create a more adversarial
relationship between the public and researchers and make it harder for
people to release things,” she notes.
― Meta’s powerful AI language model has leaked online — what
happens now? [Link]
Meta is taking the lead of open-source LLM by releasing the AI
language model LLaMA. Some say open source is necessary to ensure AI
safety and faster LLM progress. Others argue that there will be more
personalized spam and phishing due to the fact of the model has already
leaked on 4chan, and a wave of malicious use of AI. There are pros and
cons of open sourcing LLM, just like last year OpenAI open sourced
Stable Diffusion which has a lot of bad potential influences. But while
every is making AI models private, there has to be someone who makes it
public, then everyone goes public. The good and necessary thing is that
open source software can help decentralize AI power.
The OpenAI chief executive officer is in talks with investors
including the United Arab Emirates government to raise funds for a
wildly ambitious tech initiative that would boost the world’s
chip-building capacity, expand its ability to power AI, among other
things, and cost several trillion dollars, according to people familiar
with the matter. The project could require raising as much as $5
trillion to $7 trillion, one of the people said.
― Sam Altman Seeks Trillions of Dollars to Reshape Business
of Chips and AI [Link]
“Sora has a deep understanding of language, enabling it to
accurately interpret prompts and generate compelling characters that
express vibrant emotions,” OpenAI writes in a blog post. “The model
understands not only what the user has asked for in the prompt, but also
how those things exist in the physical world.”
“[Sora] may struggle with accurately simulating the physics of a
complex scene, and may not understand specific instances of cause and
effect. For example, a person might take a bite out of a cookie, but
afterward, the cookie may not have a bite mark. The model may also
confuse spatial details of a prompt, for example, mixing up left and
right, and may struggle with precise descriptions of events that take
place over time, like following a specific camera trajectory.”
― OpenAI’s newest model Sora can generate videos — and they
look decent [Link]
The predictor in this Joint Embedding Predictive Architecture
serves as an early physical world model: You don’t have to see
everything that’s happening in the frame, and it can tell you
conceptually what’s happening there.
― V-JEPA: The next step toward Yann LeCun’s vision of
advanced machine intelligence (AMI)Link]
OpenAI released amazing technology again! Compared to other release
language models, Sora seems to start to have the capability of
understanding physical world, but OpenAI acknowledged that that might
not be true. In the meantime, Meta developed V-JEPA, which is not
focusing on linking language to videos, but learning the cause and
effect from videos and gaining the capability of understand and reason
the object-object interactions in the physical world.
Google’s Gemini 1.5 Pro employs a Mixture-of-Experts (MoE)
architecture which helps the model to process large datasets by
activating relevant neural network segments. It’s capable of managing up
to 1M tokens - equivalent to 700000 words, one hour of video, or 11
hours of audio. What’s exciting is that it leverages a transformer-based
architecture with a specifically designed long context window, which
allows it to remember and process vast amounts of information. It’s able
to achieve tasks like summarizing lectures from lengthy videos. It’s
really able to retrieve ‘needles’ from a ‘haystack’ of millions of
tokens across different structures of data sources with accuracy of
99%.
Other news:
Nvidia Is Now More Valuable Than Amazon And Google
[Link]
Nvidia Hits $2 Trillion Valuation on Insatiable AI Chip
Demand [Link]
Elon Musk Says Neuralink’s First Brain Chip Patient Can
Control Computer Mouse By Thought [Link]
Capital One to Acquire Discover, Creating a Consumer Lending
Colossus [Link]
White House touts $11 billion US semiconductor R&D
program [Link]
Meta to deploy custom-designed Artemis AI processor alongside
commercial GPUs [Link]
Google Cloud (GCP and Workspace) revenue growth reaccelerated by 4
percentage points, while AWS and Azure show softer momentum. Key
business highlights: 1) Gemini in search for faster Search Generative
Experience (SGE), 2) Conversational AI tool Bard now powered by Gemini
Pro and will be powered by Gemini Ultra, 3) YouTube now has over 100M
subscribers across Music and Premium, 4) Cloud driven by AI - Vertex AI
platform and Duet AI agents, leads to expand relationships with many
leading brands (e.g. Hugging Face, McDonald’s, Motorola Mobility,
Verizon. ), 5) Waymo reached over 1M fully autonomous ride-hailing
trips, 6) Isomorphic Labs partnered with Eli Lilly and Novartis to apply
AI to treat diseases.
AI specific business highlights: 1) Google is transforming searching
behavior of customers: Search Generative Experience (SGE) is introducing
a dynamic AI enhanced search experience, 2) Gemini includes Gemini Nano,
Gemini Pro, and Gemini Ultra. Gemini Nano is optimized for on-device
tasks and already available on Pixel 8 phone. Gemini Pro is currently in
early preview through Cloud and specific apps. Gemini Ultra will be
released later in 2024, 3) the conversational AI - Bard - might be
exclusive to Tensor-powered Pixel phones and will be accessible through
voice commands or double-tapping device side buttons. Bard will also be
integrated with apps (e.g. Gmail, Maps, Drive) and Camera on Android
phones.
Amazon: Ads Take the Cake - App Economy Insights [Link]
Key updates on Amazon business: 1) Infrastructure: Amazon has
developed customized ML chips e.g. Trainium for training and Inferentia
for inference. Additionally, it offers Graviton for generalized CPU
chips, and launched Trainium2 with four times training performance. 2)
Model: Bedrock is the LLM as a Service, allowing customers to run
foundational models, customize them and create agents for automated
tasks and workflows. 3) Apps: Amazon Q is a workplace-focused generative
AI chatbot. It’s designed for business to assist with summarizing docs
and answering internal questions. It’s built with high security and
privacy, and integrated with Slack, Gmail, etc.
What else to watch: 1) Cloud: Gen AI benefits Amazon (AWS) as well as
existing market leaders in cloud infrastructure. 2) Project Kuiper is an
initiative to increase global broadband access through a constellation
of 3,236 satellites in low Earth orbit (LEO). Amazon is on track of
launching it in the first half of 2024 and will start beta testing in
the second half of the year. 3) Prime Video (with ads) remains a large
and profitable business. 4) Investment in live sports as a critical
customer acquisition strategy.
What is coming: 1) Rufus - a Gen AI-powered shopping assistant with
conversational AI capabilities 2) amazon’s advertising revenue is
catching up Meta and Google, with Prime Video a probable
accelerator.
Meta: The Zuck ‘Playbook’ - App Economy Insights [Link]
The Zuck Playbook: 1) Massive compute investment, 2) open-source
strategy, 3) future-focused research, 4) data and feedback utilization,
5) experimentation culture, 6) growth before monetization.
Meta’s business segments: 1) Family of Apps (Facebook, Instagram,
Messenger, and WhatsApp), 2) Reality Labs (virtual reality hardware and
supporting software).
Key business highlights: 1) 1B+ revenue in Q4 2023 for the first time
with Quest, and Quest 3 is off to a strong start, 2) established the
right feedback loops with Stories and Reels to test new features and
products, 3) Ray-Ban Meta smart glasses is off to a strong start. 4)
Reels and Threads are growing. 5) Llama 3 and AGI. Zuck is aiming to
position Meta as a leader in the field of AI without necessarily
monopolizing control over it.
Microsoft: AI at Scale - App Economy Insights [Link]
Key business highlights: 1) AI’s impact on Azure’s growth: most
essential revenue growth drivers are Azure OpenAI and OpenAI APIs, 2)
small language models: Orca 2 leverages Meta’s Llama 2 base models, fine
tuned with synthetic data, and Phi 2 is a transformer-based SLM designed
for cloud and edge deployment, 3) new custom AI chips: Microsoft’s first
custom chips - Maia and Cobalt. Maia 100 GPU is tailored for AI
workloads, Cobalt 100 powers general cloud services, 4) rebranding Bing
Chat as Copilot. 5) introducing a new key (Copilot key) on keyboard to
Windows 11 PCs.
What’s else in Microsoft’s portfolio: 1) Azure AI services gain more
new customers, 2) Github Copilot revenue accelerated, 3) Microsoft 365
Copilot show faster customer adoption, 4) LinkedIn, 5) Search - not
gaining market share in Search, 6) Gaming: with acquisition of
Activision Blizzard, hundreds of millions of gamers are added in to the
ecosystem. Innovation of cloud gaming improves player experience. 7)
Paid Office 365 commercial seats.
The Digital Markets Act (DMA) is a European Union regulation
designed to promote fair competition and innovation in the digital
sector by preventing large tech companies from monopolizing the market.
It aims to ensure consumers have more choices and access to diverse
digital services by regulating the practices of platforms acting as
digital “gatekeepers.”
― Apple App Store Shenanigans - App Economy Insights
[Link]
Recent news highlights: 1) as DMA compliance, Apple will allow for
third-party stores and payment systems to App Store in Europe, so the
developers can avoid 30% fee from Apple, 2) EU fines Apple €500M for
unfairly competing with Spotify by restricting it from linking out to
its own website for subscriptions. These anticompetitive practices by
favoring its services over rivals have a bad impact on Apple’s
reputation. 3) revenue from China is continuously declining.
What is coming: 1) Vision Pro has 600+ native apps and games, and is
supported by mixed streaming (Disney+, Prime Video). But Netflix and
YouTube have held back. TikTok launched a native Vision Pro App tailored
for an immersive viewing experience. 2) AI.
In a context where the crux of the thesis is the durability of
the demand for NVIDIA’s AI solutions, inference will likely become more
crucial to future-proof the business.
Jensen Huang previously described a ‘generative AI wave’ from one
category to the next:
→ Startups and CSPs.
→ Consumer Internet.
→ Software platforms.
→ Enterprise and government.
Huang continues to see three massive tailwinds:
Transition from general-purpose to accelerated
computing
Generative AI.
A whole new industry (think ChatGPT, Midjourney, or
Gemini).
History tells us that highly profitable industries tend to
attract more competition, leading to mean reversion for the best
performers.
― Nvidia at Tipping Point - App Economy Insights [Link]
AI system operate through two core stages: training and inference.
Nvidia dominates the Training segment with its robust GPU but faces
stiffer competition in the Inference segment with Intel, Qualcomm,
etc.
Nvidia’s equity portfolio: Arm Holdings (ARM), Recursion
Pharmaceuticals (RXRX), SoundHound AI (SOUN), TuSimple (TSPH), Nano-X
Imaging (NNOX), showing Nvidia’s expansive approach to AI.
Microsoft developed custom AI chips Maia and Cobalt to lessen
reliance on Nvidia and benefit OpenAI. This shows a desire for
self-reliance across Nvidia’s largest customers, which could challenge
the company’s dominance in AI accelerators.
Key business highlights: 1) Nvidia has three major customer
categories: Cloud Service Providers (CSPs) for all hyperscalers (Amazon,
Microsoft, Google), consumer internet companies such as Meta who
invested in 350000 H100s from Nvidia, and enterprise such as Adobe,
Databricks, and Snowflake who are adding AI copilots to their platforms.
2) Sovereign AI.
Apple Vision Pro Vs. Meta Quest: The Ultimate
Showdown [Link]
Apple Vision Pro review: magic, until it’s not [Link] [Link]
Apple Vision Pro launched on Feb 2, 2024 at $3,499. It’s interesting
that Meta and Apple are starting on opposite ends of the spectrum. Meta
quest has the right price and will try to improve the technology
overtime. Apple Vision Pro has the right technology and will try to
lower the price overtime. With a price of 3499, Apple is targeting the
high end of the market, not aiming for a mass market product in the
first iteration. Instead their sights are set on the early adaptors.
It’s a pattern that most Apple products take several years to achieve
the mass production. The first iteration of iPhone in 2007 was a soft
launch. iPhone didn’t crack 10M units per quarter until the iPhone 4 in
late 2010. Now Apple sells about 200M iPhones every year. So it’s highly
possible that mass adoption of technologically improved mixed-reality
headsets with more affordable pricing is coming in a decade.
Microsoft Game Day Commercial | Copilot: Your everyday AI
companion [Link]
Microsoft’s first Super Bowl Commercial to highlight its
transformation into an AI-centric company, with a focus on Copilot’s
ability of simplifying coding and digital art creation, etc.
Never ever think about something else when you should be thinking
about the power of incentives.
Fanaticism and scale combined can be very powerful.
Invert always invert, you could innovate by doing the exact
opposite of your competitors.
Once you get on the ball, stay on the ball. And once you start
down, it is mighty hard to turn around.
Success in my mind comes from having a successful business, one
that is a good place to work, one that offers opportunity for people,
and one that you could be proud of to own.
Whatever you do, you must do it with gusto, you must do it in
volume. It is a case of repeat, repeat, repeat.
Extreme success is likely to be caused by some combination of the
following factors. Number one, extreme maximization or minimization of
one or two variables. Number two, adding success factors so that a
bigger combination drives success, often in a nonlinear fashion. Number
three, an extreme of good performance over many factors. And finally,
four, catching and riding some sort of wave.
Learning is not memorizing information, learning is changing your
behavior.
Troubles from time to time should be expected. They are an
inescapable part in life. So why let them bother you, just handle them
and then move on.
You are not changing human nature things will just keep repeating
forever.
You need to do your best to avoid problems and the way you do
that is you go for great. It’s hard to do but it makes your life easier
if you go for great. Great businesses are rare, great people are rare.
But it worth the time to find. Great businesses threw off way less
problems than average or low quality businesses, just like great people
cause way less problems in life than average or low quality
people.
Opportunity is a strange beast. It frequently appears after a
loss.
When you read biographies of people who’ve done great work, it is
remarkable how much luck is involved. They discover what to work on as a
result of chance meeting, or by reading a book, they happen to pick up.
So you need to make yourself a big target for luck. And the way to do
that is to be curious.
It’s the impossibility of making something new out of something
old. In a trade where novelty is all important, I decided that I was not
meant by nature to raise corpses from the dead.
I think of my work as a femoral architecture dedicated to the
beauty of the female body.
The entrepreneur only ever experiences two states, that of
euphoria and terror.
My life, in fact, revolves around the preparation of a collection
with its torments and happiness. I know that in spite of all the
delights of a vacation, it will seem an intolerable gap. My thoughts
stay with my dresses. It is now that I like to sit down in front of my
dresses, gaze at them a last time altogether and thank them from the
bottom of my heart.
The goal is to not have the longest train, but to arrive at the
station first using the least fuel.
Find your edge, don’t diversify, and never repeat what
works.
The formula that allowed Murphy to overtake Paley was deceptively
simple: number one, focus on industries with attractive economic
characteristics; number two, selectively use leverage to buy occasional
large properties; number three, improve operations; number four, pay
down debt; and number five, repeat this loop.
The behavior of peer companies will be mindlessly
imitated.
The business of business is a lot of little decisions every day
mixed up with a very few big decisions.
Stay in the game long enough to get lucky.
The outsider CEOs shared an unconventional approach, one that
emphasized flat organizations and dehydrated corporate staffs.
Decentralization is the cornerstone of our philosophy. Our goal
is to hire the best people we can and give them the responsibility and
authority they need to perform their jobs. We expect our managers to be
forever cost conscious and to recognize and exploit sales
potential.
Headquarters staff was anorexic. No vice presidents in functional
areas like marketing, strategic planning, or human resources. No
corporate counsel and no public relations department either. In the
Capital Cities culture, the publishers and station managers had the
power and the prestige internally, and they almost never heard from New
York if they were hitting their numbers.
The company’s guiding human resource philosophy: Hire the best
people you can and leave them alone.
Murphy delegates to the point of Anarchy. Frugality was also
central to the Ethos.
Murphy and Burke realized early on that while you couldn’t
control your revenues, you can control your costs. They believed that
the best defense against the revenue lumpiness inherent in
advertising-supported businesses was a constant vigilance on costs,
which became deeply embedded in the company culture.
Life is like a big lake. All the boys get in the water at one end
and start swimming. Not all of them will swim across. But one of them I
assure will and that is Gróf.
The money will come as a byproduct of building great products and
building a great organization, but you absolutely cannot put that first
or you’re dooming yourself.
I was worth about \(\$1\)
million when I was 23. I was worth \(\$10\) million when I was 24, and I was
worth over $100 million when I was 25. And it wasn’t that important
because I never did it for the money.
I’m looking for a fixer-upper with a solid foundation. I am
willing to tear down walls, build bridges, and light fires. I have great
experience, lots of energy, a bit of that vision thing, and I’m not
afraid to start from the beginning.
Apple is about people who think outside the box, people who want
to use computers to help them change the world, to help them create
things that make a difference and not just get a job done.
As technology becomes more and more complex, Apple’s core
strength of knowing how to make very sophisticated technology,
comprehensible to mere mortals is in even greater demand.
Be a yardstick of quality. Some people are not used to an
environment where excellence is expected.
Design is a funny word. Some people think design means how it
looks. But of course, if you dig deeper, it’s really how it works. The
design of the Mac wasn’t what it looked like, although, that was part of
it. Primarily, it was how it worked. To design something really well,
you have to get it. You have to really grok what it’s all about. It
takes a passionate commitment to really, thoroughly understand
something, to chew it up, not just quickly swallow it.
Simplicity is complexity resolved. Once you get into the problem,
you see that it’s complicated, and then you come up with all these
convoluted solutions. That’s where most people stop. And the solutions
tend to work for a while, but the really great person will keep
going.
I always considered a part of my job was to keep the quality
level of people in the organizations I work with very high. That’s what
I consider one of the few things that actually can contribute
individually to, to really try to instill in the organization the goal
of only having A players.
The people who are doing the work are the moving force behind the
Macintosh. My job is to create a space for them to clear out the rest of
the organization and keep it at bay.
I’ve always been attracted to the more revolutionary changes. I
don’t know why. Because they’re harder. They’re just much more stressful
emotionally. And you usually go through a period where everyone tells
you that you’ve completely failed.
“I could see what the Polaroid camera should be. It was just as
real to me as it was – as if it was sitting in front of me before I had
ever built one.”
And Steve said, “Yes. That’s exactly the way I saw the Macintosh.
If I ask someone who had only use a personal calculator, what a
Macintosh should be, they couldn’t have told me. There was no way to do
consumer research on it. I had to go and create it and then show it to
the people and say, now what do you think?”
Both of them had this ability to, well, not invent products, but
discover products. Both of” – this is wild, man. “Both of them said
these products had always existed. It’s just that no one has ever seen
them before. We were the ones who discovered them. The polaroid camera
had always existed, and the Macintosh had always existed. It was a
matter of discovery. Steve had huge admiration for Dr. Land. He was
fascinated by him.
“Jobs had said several times that he thinks technological
creativity and artistic creativity are two sides of the same coin. When
asked about the differences between art and technology, he said, ‘I’ve
never believed that they’re separate.’ Leonard da Vinci was a great
artist and a great scientist. Michelangelo knew a tremendous amount
about how to cut stones at a quarry, not just how to make a sculpture,
right?”
“I don’t believe that the best people in any of these fields see
themselves as one branch of a forked tree. I just don’t see that. People
bring these ideas together a lot. Dr. Land at Polaroid said, ‘I want
Polaroid to stand at the intersection of art and science, and I’ve never
forgotten that.’”
In 30 years since founding Apple, Jobs has remained remarkably
consistent. The demand for excellence, the pursuit of great design, the
instinct for marketing, the insistence on each – on ease of use and
compatibility, all have been there from the get-go.
“The things that Jobs cares about, design, ease of use, good
advertising, are right in the sweet spot of the new computer industry.
Apple is the only company left in this industry that designs the whole
thing,” Jobs said.
“Hardware, software, developer relations, marketing. It turns out
that, in my opinion, that is Apple’s greatest strategieec advantage. It
is Apple’s core strategic advantage. If you believe that there’s still
room for innovation in this industry, which I do, because Apple can then
innovate faster than anyone else. The great thing is that Apple’s DNA
hasn’t changed,” Jobs said. “The place where Apple has been standing for
the last two decades is exactly where computer technology and the
consumer electronics markets are converging. So it’s not like we’re
having to cross the river to go somewhere else. The other side of the
river is coming to us.”
A topological qubit is a system that encodes data into the
properties of pairs of non-Abelian anyons by physically swapping them
around with each other in space. Non-Abelian anyons is desirable
component for holding and manipulating information in quantum computers
because of its resilience which is rooted in math from the field of
topology - the study of spatial relationships and geometry that hold
true even when shapes are distorted.
― The Holy Grail of Quantum Computing, Weird [Link]
Researcher from Google and Quantinuum demonstrated a mechanism needed
for a component called a topological qubit, which should promise a means
to maintain and manipulate information encoded into quantum states more
robustly than existing hardware designs. Topological qubits store and
work with digital information in non-Abelian anyons, which retain a sort
of “memory” of their past movement that enables the representation of
binary data. While physicists previously proved the existence of
non-Abelian anyons, Quantinuum’s and Google’s work is the first to
demonstrate their signature feature, memory of movement. However, people
disagreed that topological qubits had been created because the object
was too fragile for practical use and could not reliably manipulate
information to achieve practical quantum computing. Delivering a
practical topological qubit will require all kinds of studies of
non-Abelian anyons and the math underpinning their quirky behavior.
Technological breakthroughs can be expected after incremental
progress.
Google Quantum Al’s paper published in May 2023 is here and
no I’m not going to read this theoretical physics paper :)
Apple’s rivals, such as Samsung, are gearing up to launch a new
kind of “AI smartphone” next year. Counterpoint estimated more than 100
million AI-focused smartphones would be shipped in 2024, with 40 percent
of new devices offering such capabilities by 2027.
Optimizing LLMs to run on battery-powered devices has been a
growing focus for AI researchers. Academic papers are not a direct
indicator of how Apple intends to add new features to its products, but
they offer a rare glimpse into its secretive research labs and the
company’s latest technical breakthroughs.
“Our work not only provides a solution to a current computational
bottleneck but also sets a precedent for future research,” wrote Apple’s
researchers in the conclusion to their paper. “We believe as LLMs
continue to grow in size and complexity, approaches like this work will
be essential for harnessing their full potential in a wide range of
devices and applications.”
― Apple wants AI to run directly on it hardware instead of in
the cloud [Link]
Apple is developing solutions to running LLM or other AI models
directly on a customer’s iPhone. The published paper is titled “LLM in a Flash“. Apple’s
focus is different from Microsoft’s and Google’s focus of developing
Chatbots and GenAI over the Internet from cloud computing platform. I
think the smartphone market would be revived and customer experience
would be largely changed in the future with this potential vision of
personalized and mobile AI agents / assistants. If this personalized
little AI agent can learn everything about a person from childhood to
adulthood, can we eventually get a perfect adult mind? It reminds me of
what Alan Turing said about AI:
“Instead of trying to produce a program to simulate the adult
mind, why not rather try to produce one which simulates the child’s? If
this were then subjected to an appropriate course of education one would
obtain the adult brain.”
Since the beginning of 2017, China has chalked up more than 18
million EV sales, nearly half the world’s total and over four times more
than the US, according to BloombergNEF data. By 2026, the research group
projects that over 50% of all new passenger vehicle sales in China will
be electric, compared to a little over a quarter in the US.
The growth of the network was both a result of state planning and
private enterprise. Giant state-owned companies like State Grid Corp. of
China were given mandates to roll out chargers, while private companies
like Qingdao TGOOD Electric Co. jumped at the chance to build charging
posts—in part to lay early claim to the best locations. Baidu’s mapping
software—the Chinese equivalent of Google Maps—has them all integrated,
delivering constant reminders of where to go. Payment is typically via
an app or the ubiquitous WeChat platform.
Demand for new lithium-ion batteries is expected to increase
about five-fold between 2023 and 2033, according to Julia Harty, an
energy transition analyst at FastMarkets. Meeting that will require
recycling as well as mining.
― Electric Cars Are Driving China Toward the End of the Age
of Oil [Link]
I haven’t been back to China for 3 years, and last time (2021) there
are just a few EVs. Maybe a few years later there are only a few
gasoline cars left on the roads. The transition from gasoline cars to
electrical vehicles won’t be fast but the peak of sales for gasoline
cards is coming soon in China.
The
lawsuit, filed in Federal District Court in Manhattan, contends that
millions of articles published by The Times were used to train automated
chatbots that now compete with the news outlet as a source of reliable
information.
In its complaint, The Times said it approached Microsoft and
OpenAI in April to raise concerns about the use of its intellectual
property and explore “an amicable resolution,” possibly involving a
commercial agreement and “technological guardrails” around generative
A.I. products.
The lawsuit could test the emerging legal contours of generative
A.I. technologies — so called for the text, images and other content
they can create after learning from large data sets — and could carry
major implications for the news industry.
“If The Times and other news organizations cannot produce and
protect their independent journalism, there will be a vacuum that no
computer or artificial intelligence can fill,” the complaint reads. It
adds, “Less journalism will be produced, and the cost to society will be
enormous.”
― The Times Sues OpenAI and Microsoft Over A.I. Use of
Copyrighted Work [Link]
Microsoft, OpenAI hit with new lawsuit by authors over AI
training [Link]
There is always a group of people who are positively working on
changing the world while another group of people who are suspicious and
concerned. Every lawsuit in AI field allows us to hold on and reflect on
whether we are doing the right things and how to fix the problems along
the way of innovation and development. It is a good point that if the
intellectual property of journalism is not well-protect while AI is
still in its immature stage with a lot of mistakes in text learning and
generation, then less journalism will be produced, and less truths will
be revealed and documented. The damage to the society is enormous.
DOJ close to filing massive antitrust suit against Apple over
iPhone dominance: report [Link]
Apple is under extensive investigation of DOJ for potential antitrust
violations. Google currently faces multiple antitrust cases as well.
Per the IRS, for-profit entities and not-for-profit entities are
fundamentally at odds with each other, so in order to combine the two
competing concepts, OpenAl came up with a novel structure which allowed
the non-profit to control the direction of a for-profit entity while
providing the investors a “capped” upside of 100x. This culminated in a
$1Bn investment from Microsoft, marking the beginning of a key strategic
relationship, but complicating the company’s organizational structure
and incentives.
“I deeply regret my participation in the board’s actions,”
Sutskever, a longtime Al researcher and cofounder of OpenAl, posted on
X. “I never intended to harm OpenAl. I love everything we’ve built
together and l will do everything I can to reunite the
company.”
― 4 days from fired to re-hired: A timeline of Sam Altman’s
ouster from OpenAI [Link]
This article (Quick Essay: A Short History of OpenAI) reviewed the
history of OpenAl from 2015 when it was founded to 2023 when Sam was
once fired. A crucial development happened in 2018 when the company
first introduced the foundational architecture of GPT in a paper
“Improving Language Understanding by Generative Pre-Training”. This
leads to the flagship product of the company - ChatGPT, in Nov 2022. In
2019 OpenAl transitioned from nonprofit to a “capped-profit” model. This
novel convoluted corporate structure led to conflicting motivations and
incentives within the company and it latter raised the board’s concern
about the company’s commitment to safety.
Sam Altman was fired on Nov 17, 2023. On Nov 19, 2023, OpenAl hired
former Twitch CEO Emmett Shear as its interim CEO. Meanwhile Microsoft
CEO Satya Nadella announced that they would hired Sam to lead a new Al
department. On Nov 20, 2023, Nearly all 800 OpenAl employees signed a
letter calling for the resignation of the company’s board and the return
of Altman as CEO. On Nov 21, 2023, Sam returned as CEO of OpenAl.
One story in 2020: Two of the lead Al developers Amodei and his
sister Daniela left OpenAl in late 2020 to launch Anthropic over
concerns the company was moving too quickly to commercialize its
technology. Anthropic was founded aiming to develop more safer and
trustworthy model and it has billions invested from Google, Amazon,
Salesforce, Zoom, etc.
Since Sam was rehired, the questions about neglecting Al safety has
been quieted, and new board members appear to be more focused on
profitability. There is no doubt of OpenAl’s capability of profitably
scaling ChatGPT, but it should raise doubts about whether OpenAl is
still committing to its purpose in the future.
Sutskever was recruited to OpenAl from Google in 2015 by Elon Musk,
who describes him as “the linchpin for OpenAl being successful”. A tweet
from Greg Brockman confirms that Ilya was a key figure in Altman’s
removal. But Sutskever also a signee calling Sam back to CEO. He later
said he deeply regret his participation in the board’s actions.
National Artificial Intelligence Research Resource
Pilot [Link]
Training AI models costs huge amount of money. There is growing
divide between industry and academia in AI. Thanks to this pilot
programs stepping towards democratizing AI access.
Altman Seeks to Raise Billions for Network of AI Chip
Factories [Link]
SubStack
Tesla: AI & Robotics - App Economy Insights [Link]
What Tesla is experiencing: 1) price cuts, 2) prioritizing volume and
fleet growth, 3) continued improvement in the cost of goods sold per
vehicle. What negatively impact gross margin are 1) price cuts, 2)
Cybertruck production ramp, 3) AI, 4) other product expenses. What
positively impact gross margin are 1) lower cost per vehicle, 2)
delivery growth, 3) gross margin improvement for non-auto segments. What
to watch of Tesla’s business: 1) Model Y Triumph, 2) supercharging the
EV market (North American Charging Standard (NACS)), 3) market share hit
4% in North America, 4) Autopilot and Full Self-Driving (FSD) beta
software (V12), 5) energy storage deployments (15-gigawatt hours of
batteries delivered), 5) Optimus humanoid robot, 6) next-generation
platform “Redwood”, 7) Dojo supercomputer.
The Netherlands has restricted the export of ASML’s cutting-edge
extreme ultraviolet (EUV) lithography machines to China, a decision
influenced by international diplomatic pressures, notably from the
US.
ASML is the sole producer of advanced EUV lithography machines at the
center of the global chip supply chain. Since the restriction of export
of EUV lithography machines to China, it is now also at the mercy of the
US-China trade war. There is a list of risk ASML is facing: IP theft and
security breaches, cybersecurity costs, China’s ambitions of developing
its semiconductor sector, semiconductor industry volatility (shortage
and gluts), etc.
Paper and Reports
The economic potential of generative AI: The next
productivity frontier [Link]
Thanks to my favorite quote and my motto in life for pulling me back
from the dark over and over and over again.
“It is not the critic who counts; not the man who points out how the
strong man stumbles, or where the doer of deeds could have done them
better. The credit belongs to the man who is actually in the arena,
whose face is marred by dust and sweat and blood; who strives valiantly;
who errs, who comes short again and again... who at the best knows in
the end the triumph of high achievement, and who at the worst, if he
fails, at least fails while daring greatly.”