Aaron Batilo
Denver Metropolitan Area
4K followers
500+ connections
View mutual connections with Aaron
Aaron can introduce you to 10+ people at CoreWeave
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Aaron
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
About
My keyboard changes colors.
Activity
4K followers
-
Aaron Batilo reposted thisAaron Batilo reposted thisWe are once again the only Platinum tier cloud according to SemiAnalysis!
-
Aaron Batilo shared thisBarsAaron Batilo shared thisSomething worth stepping back to appreciate: the largest AI model developers in the world now run on CoreWeave — Meta, OpenAI, and as of today, Anthropic. I shared the news about Meta yesterday, but it bears repeating in this context. Our newest deal with Meta brings our announced total with them to over $35B in six months. These are not small bets. The leading builders of AI are choosing us, and doubling down. At CoreWeave we believe that the AI era demands a cloud built from the ground up for this moment. And building for this moment means financing for it too. This week we concurrently priced a $3.5B convert + $1.75B high yield bond offering, which we believe is the largest dual-tranche raise of its kind in history. The high yield bond offering was upsized due to significant oversubscription. And just last week, we closed an $8.5B delayed draw term loan that received an investment grade rating. The market is saying the same thing we are. The demand is real. The capital is there. We're building The Essential Cloud for AI. https://lnkd.in/e3wwcUjDCoreWeave Announces Multi-Year Agreement With AnthropicCoreWeave Announces Multi-Year Agreement With Anthropic
-
Aaron Batilo shared thisone of the greatest to ever do itAaron Batilo shared thisI’m happy to share that I’m starting a new position as Senior Manager Engineering, Applied Training at CoreWeave!
-
Aaron Batilo shared thisCoreWeave is showing up big time for #GTC2026. Come by our booth and say hello to me or Deok Filho or Amit Gupta or anyone else!
-
Aaron Batilo shared thisfr though. you definitely want to see what Harsh looks like nowAaron Batilo shared thisThe way you think about AI infrastructure is about to change. Join CoreWeave and NVIDIA on January 29th to see why rack-scale systems are the future of AI. ⚖️ Understand the benefits and challenges of rack-scale ⬆️ Achieve unprecedented uptime across rack-scale systems 🎓 Learn rack-scale reliability essentials Register now: https://hubs.la/Q0407WFh0
-
Aaron Batilo shared thisI'm building a new team at CoreWeave. Designed to bridge the gaps that exist in the tooling and infrastructure between the compute and a scientist being able to run their experiments, the Applied Training team sits in between CoreWeave's TWICE Platinum https://clustermax.ai rated cloud offerings, and the bleeding edge research that's redefining the world as it is today. In my time at places like Cohere AI, Inflection AI, and Microsoft AI, I've come to see that there's so much more we can do to accelerate the brilliant research scientists of the world. That's exactly what we're going to do. No prior AI/ML experience is required as long as you're interested and willing to learn the domain. Note: we are NOT training our own models. Please apply below! Staff Software Engineer: https://lnkd.in/gbefDhAU Senior Software Engineer II: https://lnkd.in/gVqxtN4E
-
Aaron Batilo shared thisIt's so exciting to be a part of this work!Aaron Batilo shared thisStanford + Coreweave + Meta/PyTorch - a big milestone for scalable RL and post-training in the open.. Our latest post shows how TorchForge and Weaver come together to supercharge LLM reinforcement learning, from small experiments all the way to massive, distributed runs - all running on Coreweave's SOTA AI infra.. What I’m most excited about isn’t just the performance or scale, but the people and collaboration behind it: researchers and engineers pushing the frontier of RL, systems, and open infrastructure across Meta, academia, and the broader PyTorch community. This kind of deep, end-to-end work is what makes open ecosystems win. Huge credit to everyone who helped turn years of RL research into something developers can actually use and build on. This is a big step toward making large-scale RL a first-class, PyTorch-native workflow. Credit to: Stanford - Jon S., Hangoo Kang, Simon Guo, Aakanksha Chowdhery, Ph.D., Azalia Mirhoseini Meta - Allen Wang, Danning Xie, Evan Smothers, Felipe Mello, Jack Khuu, Jiyue Wang, Joe Cummings, Lucas Pasqualin, Philip Bontrager, Rithesh Baradi, Vidhya Venkat, Yuxuan Hu, Jafar Taghiyar, @Davide Italiano, Gayathri Aiyer (Ph.D), John Myles White, Sanyam Bhutani, Hamid Shojanazeri, Matthias Reso, Alireza Shamsoshoara, @Hossein Kavianihamedani, Emre Guven CoreWeave - Deok Filho, Aaron Batilo, Matthew Guan, @Xi Lu Post: https://lnkd.in/gFKDVJKX Cheers!
-
Aaron Batilo shared thisSemiAnalysis is hiring! If you want to be at the center of the AI and Data center information space, you should reach outAaron Batilo shared thisGoing into the holiday season, we are looking for talented engineers to work on ClusterMAX 3.0 and related consulting projects. The SemiAnalysis team has grown from under 20 to over 50 just this year. We support remote/hybrid work, with offices in SF, NYC, and Singapore, and team members in 11 different countries globally. The scope of work is always related to the cutting edge, latest and greatest. From chips to datacenters to cloud and AI models, we cover the full stack. If you like ClusterMAX, InferenceMAX, or any of our other research on ML Systems and Performance, please consider applying at the link below!
-
Aaron Batilo reposted thisAaron Batilo reposted thisHeading to #kubecon? Tomorrow, our own Navarre Pratt will share the stage with Shaun Hopper from Meta and tell you how Meta uses #SUNK (Slurm on #Kubernetes) to deliver a massively scalable, production-tested Slurm cluster on top of heterogeneous Kubernetes infrastructure. https://sched.co/27FWE Why should you care? SemiAnalysis last week report summarizes it perfectly: "CoreWeave’s SUNK was first to market, is proprietary, and continues to be the only viable solution for running both slurm and Kubernetes jobs on the same underlying cluster. For example, a batch queue of slurm training jobs, and an autoscaling inference endpoint on Kubernetes that compete for underlying GPU resources." #PlatinumAICloudKubeCon + CloudNativeCon North America 2025: Meta’s Kubernetes-based Portable AI Rese...KubeCon + CloudNativeCon North America 2025: Meta’s Kubernetes-based Portable AI Rese...
-
Aaron Batilo liked thisAaron Batilo liked thisI am excited to share that I will be joining CoreWeave this summer as a Software Engineering Intern on the CoreWeave Kubernetes Services team in New York City. A huge thank you to my friends and family for their support. A special thank you to Christie Manibusan for making a smooth recruitment process, as well as DJ Enriquez and Jake Gilbert for a welcoming interview experience! I am incredibly grateful for this opportunity and excited to learn and contribute this summer. #CoreWeave
-
Aaron Batilo liked thisExcited to be putting this out there!Aaron Batilo liked thisWe built a system combining program analysis and LLMs to transform and optimize PTX. By operating at this shared layer across DSLs (e.g. Triton, TileLang, ThunderKittens, CUTLASS), our system learns the best ideas from each and generates kernels that outperform all of them! Full post at https://lnkd.in/gEnt5Amy
-
Aaron Batilo liked thisAaron Batilo liked thisI'm super excited to share what I have been working on for the last 1,5 months! Ineffable Intelligence is pushing on the path to superintelligence, and raised 1.1B by an excellent list of VCs. We deeply focus on RL, and driven by agents learning from experience, rediscovering and then transcending the greatest inventions in human history, such as language, science, mathematics and technology.. We are hiring in London, Kings Cross! In particular, if any of the following roles fit you skills, please apply at www.ineffable.ai. - Amazing dev ops lead, who ideally worked at large AI startups before - Excellent RL researchers who pushes the boundary of the field - Stellar engineers who worked with distributed systems, GPUs and scale.
-
Aaron Batilo liked thisAaron Batilo liked thisLast Friday was my 3 year anniversary at CoreWeave. As anyone at the company will tell you, it’s been a career-defining journey from scrappy startup to AI leader. It’s been a privilege to lead the Cloud Platform team through this stage of hypergrowth, and I’m incredibly proud of the team we’ve built. We’re only getting started. The technology we’re enabling at CoreWeave is intoxicating. For builders, it feels like god mode, and it’s pulled me back in. My next chapter at CoreWeave is as a Principal Engineer on the Cloud Platform. Back to building. LFG. 🚀🚀🚀
-
Aaron Batilo liked thisAaron Batilo liked this🎉 Want to run the latest #DeepSeek V4 Flash model on your #Kubernetes cluster? Here's a simple starting point: this example deploys DeepSeek-V4-Flash on Kubernetes using #vLLM + #Ray on a single node with 4 #H100 GPUs, and a service endpoint for #OpenAI-compatible #inference. Also includes a script to test its enhancements in agentic reasoning and multi-turn thinking. 👉 https://lnkd.in/d8YHJDFQ ======================================================== TEST 4: Multi-turn tool-use loop Goal : After a tool is called, the caller sends the tool's output back as a role:'tool' message. The model should consume that result and produce a natural-language final answer — no new tool_calls. Signal : finish_reason=='stop' AND tool_calls is empty AND .content mentions values from the synthetic tool result ('Tokyo', '18', 'rain'). ======================================================== History: user : "What is the weather in Tokyo right now?" assistant : (tool_call: get_weather(city=Tokyo)) tool : {"city":"Tokyo","temp_c":18,"conditions":"light rain"} Expecting: assistant summarizes the tool result in plain English. Response: finish_reason : stop tool_calls : [] .content : The current weather in Tokyo is **light rain** with a temperature of **18°C** (about 64°F). [PASS] model consumed the synthetic tool result and produced a grounded natural-language answer — full tool-use loop works end-to-end.
-
Aaron Batilo liked thisAaron Batilo liked thisOne of the largest AI clouds is now building for multi-cloud, not against it. Today, CoreWeave announced three products: • CoreWeave Interconnect — private link to Google Cloud • SUNK Anywhere — their Slurm-on-K8s system now runs on other clouds • LOTA Cross-Cloud — object storage <-> GPU node cache, with zero egress fees The multi-cloud AI pattern has been quietly accelerating. OpenAI went from 1 cloud to 5 in 18 months. Every AI neolab runs workloads on multiple providers. The future of AI compute is multi-cloud. Exciting times to build in this space. Congrats to the CoreWeave team for these releases!
-
Aaron Batilo liked thisAaron Batilo liked thisI’m excited to share that I will be joining CoreWeave this summer as an intern in Livingston, NJ, where I will be working on the Data Center Security Risk & Assurance team. I’m thankful for the support and encouragement I’ve received throughout this journey, and I truly appreciate everyone who has played a role in helping me reach this opportunity. I’m looking forward to gaining hands-on experience and contributing to a company at the forefront of cloud computing infrastructure and AI. Excited for what’s ahead!
Experience
Projects
-
CatStories.ai
-
See projectLet GPT-4 send you an SMS message with an adorable and heart warming story about kitties!
-
ColoradoExcluded
-
See projectUser submitted, then manually curated job listings that are explicitly choosing not to hire candidates in Colorado.
More information from when the website was featured in the Wall Street Journal: https://www.wsj.com/articles/many-companies-want-remote-workersexcept-from-colorado-11623937649?reflink=desktopwebshare_permalink -
EC2Throughput.info
-
See projectDid you know that when AWS says an EC2 instance has "Up to N Gbps" of bandwidth, this is actually a peak value. Bandwidth on these instances gets rate limited after only a few minutes of hitting that burst/max throughput. EC2Throughput is a set of GitHub Actions and a Next.js application that can automatically benchmark EC2 instances to find out what you should actually expect for bandwidth.
-
SudokuRace.io
-
Solve a sudoku puzzle with your friends and race to see who can fill in the most squares
Languages
-
English
-
-
Tagalog
-
Recommendations received
View Aaron’s full profile
-
See who you know in common
-
Get introduced
-
Contact Aaron directly
Other similar profiles
Explore more posts
-
Matthew O'Keefe
Confluent • 8K followers
"Beyond Indexes: How Open Table Formats Optimize Query Performance" is another profoundly great post on Open Table Formats and their evolution by Confluent's Jack Vanlightly. He walks you through the history of primary and secondary indices in OLTP databases, then contrasts that with the requirements for analytic workloads in Open Table Formats, showing that complex secondary indices don't help there. Instead, "in open table formats, layout is king. Here, performance comes from layout (partitioning, sorting, and compaction) which determines how efficiently engines can skip data, while modeling, such as star or snowflake schemas, make analytical queries more efficient for slicing and dicing." https://lnkd.in/gEdizbSY
5
-
Ville Takanen
Futurice • 2K followers
2️⃣ Running tests with e.g. playwright will become actually pleasant. Performing web-browsing won't take ages. I think this opens a lot of new agentic opportunities. This will be a significant productivity boost for dev teams. It might even be the biggest, since agentic coding.
-
AI + App Dev Chronicles
19 followers
🚀 Beyond Benchmarks: The Coding Personalities of LLMs Benchmarks only tell part of the story. A new Sonar report digs deeper—unveiling the “coding personalities” of today’s top LLMs: from the verbose “senior architect” Claude Sonnet 4 to the concise but risky “rapid prototyper” OpenCoder-8B. The real takeaway? Each model mirrors human dev quirks—big ideas, hidden trade-offs, and vulnerabilities that can haunt production. Even GPT-5, with reasoning modes, shows how gains in accuracy often come at the cost of new, subtler bugs. 👉 As app developers, we need to stop chasing “the best model” and start asking: Which coding style aligns with the product lifecycle—prototype speed, enterprise reliability, or balanced versatility? Would you hire a rapid prototyper or a senior architect… if your AI model was a dev on your team?
-
WishInfinite
27 followers
Recently explored a more structured way to handle Playwright test reporting — beyond just console logs and basic reports. In this video, I’ve covered: -> Setting up TestDino with Playwright -> Running tests and analyzing execution reports -> Understanding failures with better visibility -> Using AI insights for faster debugging The focus was to keep it practical — something you can directly apply in your automation workflow. 👉 https://lnkd.in/dTWasX3H If you’re working with Playwright, how are you currently handling test reporting and debugging? #Playwright #AutomationTesting #TestAutomation #SoftwareTesting #QA
-
Ash S.
McKesson • 2K followers
Multi-node GPU training feels “fragile” because your model is basically running a distributed systems benchmark. Everyone blames “slow GPUs.” But at scale, the bottleneck is almost always the interconnect + tail latency. A single AllReduce is a global barrier: if one rank gets delayed (microburst, ECN/PFC pause, noisy neighbor, bad route, a single retransmit), everyone waits. RDMA helps—a lot—but it’s not magic. ✅ What RDMA actually fixes • Cuts CPU overhead and kernel involvement on the fast path (after setup) • Reduces copies / buffering (especially important for high message rates) • Lowers latency variance compared to “socket path” networking ❌ What RDMA does not fix • Congestion + queueing (microbursts will still stall collectives) • Packet loss / retransmits (RoCE + mis-tuned fabric = “mystery hangs”) • Topology mismatch (oversubscription, cross-rack hops, uneven paths) • Stragglers (one slow node turns into cluster-wide slowdown) • Control plane fragility (timeouts, retries, NCCL error handling, driver quirks) The production truth: RDMA is a necessary ingredient for high-scale training… …but the stability comes from end-to-end engineering: Engineer-proof checklist (what actually moves the needle) • Placement & topology: same rack/leaf when possible, avoid oversubscription, pin ranks with awareness • GPU-to-NIC path: ensure you’re not silently staging through host memory when you expected GPU-direct • RoCE/IB tuning: MTU, ECN, PFC pause behavior, congestion control (and verify counters, don’t assume) • NCCL config: timeouts, algorithm selection, channel count, correct interface binding • Observability: retransmits, ECN marks, PFC pauses, NIC queue depth, NCCL step time p99 (not average) • Resilience: frequent checkpointing, fail-fast + restart policies, “blast radius” limits for retries If you want multi-node training to stop feeling fragile, treat it like what it is: a latency-sensitive distributed system where p99 dominates your throughput. Curious: when your jobs “randomly slow down,” what’s the #1 culprit you see in practice—network congestion, stragglers, or NCCL timeouts? #RDMA #RoCE #InfiniBand #NCCL #DistributedTraining #AIInfrastructure #HPC #Networking #MLOps #CloudComputing
3
-
Khaled Zaky
RBC Borealis • 5K followers
Anthropic shared a wild case study on Claude Opus 4.6. When the model hit a wall during a websearch test, it didn't just fail as it correctly hypothesized it was being evaluated, identified the specific benchmark, and then wrote its own code to find and decrypt the answer key 🤯 As I’ve written recently on my blog, the shift from "software that executes" to "agents that act" changes everything. This is a perfect example of why governance is an architectural problem, not a compliance one. When an agent is smart enough to "hack the test" to achieve its goal, traditional static gates and simple prompts aren't enough. We need a true platform mindset…one built for autonomous actors that can recognize the boundaries of their sandbox and actively look for a way out….If you’re still treating AI as a deterministic tool, it’s time to rethink your stack The full deep dive from Anthropic is worth a read: https://lnkd.in/gPXdffpy
91
5 Comments -
Gunnar Morling
Confluent • 9K followers
Absolutely blown away by all the positive feedback on the #Hardwood 1.0.0.Alpha1 release 🥳! Thanks a lot to all the comments, discussions on issues in the tracker, pull requests, stars on the repo! So cool to see that this parser for #ApacheParquet, optimized for minimal dependencies and great performance, seems to itch a scratch for folks. I got invited to present Hardwood on the Parquet community call where we'll explore opportunities for collaboration, and I'll hop on a few podcasts to talk about the project, too. For the 1.0 release, the key items we're planning to implement initial support for predicate push-down, reading files from S3, and a compatibility layer with parquet-java. And, of course, further performance improvements: just today, Rion Williams opened a PR which substantially cuts down allocation rates, thus reducing GC pressure and improving overall latency. Onwards 🚀!
63
1 Comment -
Ian Bull
Mediform • 996 followers
🚨 New Claude Code plugin: Lisa's Notes 🚨 I just released Lisa's Notes, a Claude Code plugin that automatically takes notes whenever Claude thinks it’s “done” by creating a Jujutsu (jj) commit. Because jj auto-stages everything, each commit becomes a clean, chronological record of what the agent actually did. No manual note taking. No lost context. Just a durable trail of progress. 👉 Plugin: https://lnkd.in/gt6YKWEe A few notes: • Works well on its own, as you can work through a number of steps and clean-up the code afterwards • Compliments the Ralph Wiggum loop nicely, as you can let Ralph do his thing and look at all the individual steps afterwards • Jujutsu makes it really easy to squash, split, re-order and remove commits afterwards • Works really well with my current AI native workflow: https://lnkd.in/gXy5-yjn
6
-
SkrewAI
50 followers
Solid primer on the five architectures behind synthetic media. GANs, VAEs, Transformers, CNNs, RNNs. What's telling: the piece notes modern systems rarely use single architectures anymore. Diffusion models blend Transformer attention with U-Net structures. StyleGAN builds on progressive growing techniques from years ago. The toolbox keeps expanding. The combinations keep multiplying. #SyntheticMedia #GenerativeAI #AI #Deepfakes
1 Comment
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content