Weights & Biases’ cover photo
Weights & Biases

Weights & Biases

Software Development

San Francisco, California 92,071 followers

The AI developer platform.

About us

Weights & Biases: the AI developer platform. Build better models faster, fine-tune LLMs, develop GenAI applications with confidence, all in one system of record developers are excited to use. W&B Models is the MLOps solution used by foundation model builders and enterprises who are training, fine-tuning, and deploying models into production. W&B Weave is the LLMOps solution for software developers who want a lightweight but powerful toolset to help them track and evaluate LLM applications. Weights & Biases is trusted by over a 1,000 companies to productionize AI at scale including teams at OpenAI, Meta, NVIDIA, Cohere, Toyota, Square, Salesforce, and Microsoft. Sign up for a 30-day free trial today at http://wandb.me/trial.

Website
https://wandb.ai/site
Industry
Software Development
Company size
201-500 employees
Headquarters
San Francisco, California
Type
Privately Held
Founded
2017
Specialties
deep learning, developer tools, machine learning, MLOps, GenAI, LLMOps, large language models, llms, Generative AI, Developer Tools, Experiment Tracking, AI Governance, Model Monitoring, Inference, Open Source AI, Model Comparison, Evals & Scorers, Data Quality, Generative AI, AI Observability, Agentic Workflows, RAG (Retrieval-Augmented Generation), Prompt Engineering, Hyperparameter Tuning, Benchmarking, Large Language Models (LLMs), Reproducibility, Dataset Versioning, and Tracing

Products

Locations

Employees at Weights & Biases

Updates

  • Weights & Biases reposted this

    Europe doesn’t need more people saying we should build AI companies. It needs more people helping the builders actually do it. We just got back from a live Founders and Builders podcast episode at Project Europe’s Barcelona hackathon, with 90 of Europe’s best engineers/builders hacking through the night. CoreWeave were there too, after signing Anthropic and Jane Street as clients that same week 👀 So I wanted to sit down with Ben Richardson, VP Strategy, to understand why the fastest growing AI Cloud in the world had flown a team out to Barca to work with these founders, helping them think, build, and access the infrastructure they needed. We got into: • Why AI needed a purpose-built cloud <- this was fascinating and not obvious to me • What founders should understand about infrastructure as they scale • CoreWeave's approach to M&A and what most founders usually miss • How you can get involved in the next AI hackathon My favourite idea from the conversation was that the value of a hackathon isn’t just the projects built overnight. It’s the co-founder conversations, the relationships, the energy, and the ideas that might become companies six months from now. That’s what Project Europe and CoreWeave are creating. Thanks again to all the partners that came down for the hack: THEKER Robotics, Clay, Preply, Biorce, Cala, 20VC and we're super happy to have CoreWeave as a partner for Founders & Builders.

  • Weights & Biases reposted this

    Training robotics AI is complex by default because every modality adds another layer of uncertainty. These systems operate in the physical world, generating rich, multimodal outputs like videos, trajectories, segmentation maps, and simulation results across countless experiments. Comparing performance has traditionally been slow, manual, and error prone. As workflows scale, it becomes easier to miss subtle regressions or overlook key insights, slowing progress when speed matters most. To move faster, robotics teams need better ways to evaluate and compare experiments, with advanced visualization and a unified platform to track and analyze model behavior without switching between tools. Weights & Biases brings AI and high fidelity simulation into a single development view, enabling faster iteration, deeper insights, and more rigorous evaluation for physical AI teams. https://loom.ly/f5c_0ro #icra #robotics #WeightsAndBiases

    • No alternative text description for this image
  • Weights & Biases reposted this

    𝗪𝗮𝘆𝘀 𝘁𝗼 𝘀𝗰𝗮𝗹𝗲 𝗮𝘂𝘁𝗼𝗻𝗼𝗺𝘆 Autonomous driving always seemed to be "10 years away". Today: Waymo is expanding to 10+ cities, Waabi running driverless freight and expanding into robotaxies, Wayve driving in cities it has never seen before. Now the question is: can these players scale beyond largest metros with good weather and how to get there. Some are going all-in on the full stack: owning the vehicle, software, operations, and customer experience. Others are doing the opposite: focusing on one layer and partnering across the rest of the ecosystem. Both approaches have benefits and tradeoffs.  The winners in the space will be defined not only by tech, but also their ability to scale. Because in AV, scale isn’t just distribution: it’s learning, safety, economics, and credibility with regulators, insurers and consumers. I unpack the different paths to scale and what might win here 👇

  • Weights & Biases reposted this

    Physical AI pipelines aren’t shaped like the ML training you’re used to. Standard fine-tuning is one shape. Data goes in, model comes out and an eval at the end. A production VLA pipeline is three shapes at once. Data loading runs on CPUs. Training runs on one class of GPU. Stochastic sim-eval runs on another — and a single rollout tells you almost nothing. You need hundreds, with domain randomization, before you can trust that the policy will transfer to real hardware. Each stage wants different infrastructure. Each one scales independently. While all of it has to roll up to a single training decision. On May 12 at 10am PT, our own Anushrav V. from Weights & Biases and Ian D. Jordan, PhD from Anyscale are walking through the actual blueprint: → Orchestrating heterogeneous training + sim-eval workloads with Ray → Aggregating thousands of stochastic rollouts into something you can act on → Tracking video rollouts, trajectories, and task success alongside training metrics in W&B Models → Closing the gap between simulation and real-world deployment If you’re building anything in robotics or physical AI, this one’s worth the hour. Save your spot → https://lnkd.in/gHAHDTgn

    • No alternative text description for this image
  • Physical AI pipelines aren’t shaped like the ML training you’re used to. Standard fine-tuning is one shape. Data goes in, model comes out and an eval at the end. A production VLA pipeline is three shapes at once. Data loading runs on CPUs. Training runs on one class of GPU. Stochastic sim-eval runs on another — and a single rollout tells you almost nothing. You need hundreds, with domain randomization, before you can trust that the policy will transfer to real hardware. Each stage wants different infrastructure. Each one scales independently. While all of it has to roll up to a single training decision. On May 12 at 10am PT, our own Anushrav V. from Weights & Biases and Ian D. Jordan, PhD from Anyscale are walking through the actual blueprint: → Orchestrating heterogeneous training + sim-eval workloads with Ray → Aggregating thousands of stochastic rollouts into something you can act on → Tracking video rollouts, trajectories, and task success alongside training metrics in W&B Models → Closing the gap between simulation and real-world deployment If you’re building anything in robotics or physical AI, this one’s worth the hour. Save your spot → https://lnkd.in/gHAHDTgn

    • No alternative text description for this image
  • View organization page for Weights &amp; Biases

    92,071 followers

    Our colleague Russell Ratshin broke a lamp while moving a table to vacuum. 💥 Around the same time, the family decided their living room needed an upgrade — clearing out the kid clutter and turning it into something more grown-up, record player included. You can measure. You can squint and try to picture it. But you don't really know how a sofa, a lamp, or a set of chairs fits your living room until it's there. So rather than keep guessing, Russ built a small AI agent to close that gap — drop in a photo of the room and a photo of the piece, get back a staged image of both together. Then he wrote it up as the inaugural post of our new AI Builders series. The whole thing comes together with a surprisingly small stack: → marimo — the AI-native notebook environment where he prototyped → Google Gemini 3.1 Flash Image Preview, aka Nano Banana 2 — handles the image generation → W&B Weave — wraps the agent with observability, capturing traces, generated images, cost, and latency in one place The post walks through the use case, the setup, the agent code (~50 lines wrapped in a weave.Model class), and why observability and evaluation matter the moment you put an LLM-powered feature in front of real users. 🐈 There's also a brief appearance from Max, Russ's orange cat. If you're building agents — for interior design or anything else — the patterns translate directly. Read the full post: https://lnkd.in/dd6AJgX5 Clone the GitHub Repo: https://lnkd.in/d_eysjr9

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • Every ML team has the same 3 or 4 links they open on every run. The Grafana dashboard. The Datadog view. The eval results page. The dataset card. Useful, but in W&B they lived inside the config blob. Scroll, find, click. You can now pin selected config keys as References on the run Overview tab. What that means: ➙ Pin one key or many at once ➙ References get their own section at the top of the page, above Notes and Tags ➙ Values render as markdown, so a string like "[Grafana](https://...) | [Datadog](https://...)" displays as a clean row of clickable links ➙ Order is preserved, so you control the layout Small change, real quality of life upgrade if you live in the Overview tab during debugging or reviews. Super easy to get started, see the docs below! #MLOps #WeightsAndBiases #MachineLearning

    • No alternative text description for this image

Similar pages

Browse jobs

Funding