CoreWeave's founders, with backgrounds in institutional commodity trading, saw cryptocurrency mining as an arbitrage opportunity due to predictable power costs and efficient revenue modeling.
They focused on GPU-oriented mining for Ethereum because GPUs could also run AI workloads, providing more versatility.
CoreWeave transitioned into the cloud infrastructure market in 2019, building a cloud specifically designed for running AI and highly parallelizable workloads.
CoreWeave's cloud infrastructure outperforms competitors due to engineering decisions made for AI workloads rather than website hosting and data storage.
CoreWeave specializes in providing AI infrastructure at scale, with clients using 10,000 GPUs simultaneously in a single contiguous fabric, creating supercomputers.
GPUs consume a significant amount of energy, accounting for 10% of the total cost of delivering GPU infrastructure.
Data center capacity, especially data centers with sufficient power, is the bottleneck in the cloud infrastructure market.
GPUs are more power-efficient than CPUs on a workload basis but consume more power on a density basis.
The immense power consumption of GPUs is justified by their ability to unlock value from data.
There is a need for more baseload power, such as nuclear power, to support the consistent demand growth of data centers and electric vehicles.
CoreWeave's Brannin McBee discusses the future of AI infrastructure, GPU economics, and data centers.
McBee emphasizes the importance of contracting a substantial amount of capacity to ensure business growth and warns that the lack of capacity will hinder other market participants.
McBee also mentions the challenge of managing the heat generated by these systems.
Data centers consume a significant amount of energy, with cooling infrastructure requiring 2 to 3 times more energy than the infrastructure itself.
The industry is transitioning from forced air cooling to liquid cooling for improved energy efficiency.
CoreWeave focuses on direct-to-chip liquid cooling for its operational efficiency, involving pipes running to the chips for easier servicing.
Data centers are highly controlled environments with strict measures to prevent accidents.
GPU economics are changing due to the rising demand for AI infrastructure, leading to increased GPU costs and the need for data center reconfiguration to accommodate power and cooling requirements.
GPUs are expensive, but inference engines and LPUs (Language Processing Units) are emerging as purpose-built hardware for inference, potentially lowering costs.
Different models will require different types of infrastructure for optimal efficiency.
Companies like Microsoft and Meta are building their own silicon to solve for specific models they run internally, not to replace GPUs.
GPUs will continue to dominate in training the most demanding and complex workloads, especially latest generation models.
Inference may see various levels of infrastructure solutions, but models trained on A100s will likely run best on A100s due to software compatibility.
Nvidia's open-source driver solution (CUDA) has become the default across the market, giving them a strong advantage in the AI infrastructure sector.
Brannin McBee, CEO of CoreWeave, discusses the future of AI infrastructure, GPU economics, and data centers.
AI is rapidly evolving and becoming more accessible, leading to increased demand for AI infrastructure.
Traditional data centers are not well-suited for AI workloads, which require high-performance computing and specialized hardware.
CoreWeave is developing new AI infrastructure solutions that are optimized for AI workloads.
GPUs are essential for AI workloads, but they are also expensive.
The cost of GPUs is a major barrier to entry for many organizations that want to use AI.
CoreWeave is working to reduce the cost of GPUs by developing new GPU architectures and optimizing GPU utilization.
Data centers are critical for AI infrastructure, as they provide the compute power and storage capacity needed for AI workloads.
CoreWeave is building new data centers that are optimized for AI workloads.
These data centers will provide the infrastructure needed to support the growth of AI.
Brannin McBee concludes by discussing the importance of AI infrastructure and the role that CoreWeave is playing in developing new AI infrastructure solutions.
CoreWeave's Brannin McBee discusses the challenges and solutions in building AI infrastructure, particularly for large language models.
A key challenge is the throughput of data movement between GPUs, which can be a bottleneck. Non-blocking Infiniband fabric is crucial for ensuring high performance and efficiency without bottlenecks.
Building a 16,000 GPU fabric involves complex physical engineering, with 48,000 discrete connections and 500 miles of fiber optic cabling.
CoreWeave provides a software solution and platform specifically designed for these types of workloads, catering to clients who require the best engineering solutions.
Large companies like Microsoft, Meta, and Google are building infrastructure as a means to utilize their capital, creating potential jobs and opportunities for innovation, as they are unable to deploy it through M&A due to regulatory restrictions.
Promising use cases for AI infrastructure include integrating AI into existing products seamlessly, without requiring users to learn new interfaces or applications.
The growth of AI integration into existing user processes is limited by the availability of cloud infrastructure.
Cloud infrastructure limitations can delay product launches and hinder the growth of AI-powered applications, especially for those that require extensive GPU usage, such as co-pilot products and search engines.
The high cost of processing AI queries using GPUs is a barrier to widespread adoption.
Integrating AI into software products may become mandatory, leading to increased competition and potential market share loss for companies that don't adopt AI.
Companies like Microsoft are investing heavily in AI infrastructure to gain a strategic advantage.
Generative AI, particularly in advertising, will drive significant demand for infrastructure, especially GPU-based infrastructure.
The demand for AI infrastructure exceeds the capabilities of current cloud systems, necessitating a fundamental shift in infrastructure design.
Supply and demand for AI infrastructure may normalize by the end of this decade.
Infrastructure growth is on a heavy trajectory, but supply may catch up to demand by then.
CoreWeave's background in commodity trading helps them assess supply and demand in the AI infrastructure market.
Unlike fungible cloud infrastructure for hosting websites, the lack of fungibility in AI infrastructure is being de-commoditized through software and infrastructure disruption.
CoreWeave is hiring 20 people a week to configure and manage their infrastructure.
They have hundreds of people unboxing and racking equipment, and semi-trucks arriving at their 28 data centers across the US.