Lezhi Li
San Francisco Bay Area
2K followers
500+ connections
View mutual connections with Lezhi
Lezhi can introduce you to 10+ people at Apple
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with Lezhi
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
Activity
2K followers
-
Lezhi Li shared thisLast weekend, I dived into this book on training large-scale foundation models, and was instantly impressed by its quality. Having been working on foundation model training at scale, the content of this book feels very close to heart. One key takeaway: distributed training is transitioning from an auxiliary skill into a core of machine learning engineering as enter the LLM era; and I can’t agree more. If today’s system design interviews are shaped by the internet era, it’s only a matter of time before LLM scaling becomes the new standard in AI system design. Would recommend this book as a must-read for anyone working on foundation model training!
-
Lezhi Li shared thisGlad to share that I will be attending the #kdd2023 conference next week! Come say hi if you’re around. You are also welcome to come visit the Apple booth and learn about our latest research publications and career opportunities in AI and machining learning. Looking forward to seeing you there!
-
Lezhi Li shared thisWe have open-sourced manifold at https://lnkd.in/geGxRic. Read more about it on Uber eng blog at https://lnkd.in/gKeTNDB #ML #dataviz Lezhi Li shared thisWe built Manifold, a model-agnostic visualization tool for machine learning performance diagnosis and model debugging, to facilitate a more informed and actionable model iteration process. https://ubere.ng/2Hac0O8 via Lezhi Li and Yang Wang #MachineLearning #DataVisualization #DataModeling
-
Lezhi Li shared thisMy talk on how we built Manifold - a visual debugging tool for machine learning, presented at Tensorflow Dev Summit: https://lnkd.in/gTbzqpK #machinelearning #visualizationBuilding a Visual Debugging Tool for ML - TF.js in Interactive Visual Analytics (TF Dev Summit '19)Building a Visual Debugging Tool for ML - TF.js in Interactive Visual Analytics (TF Dev Summit '19)
-
Lezhi Li shared thisOur work in #MachineLearning #Visualization at Uber: Manifold - model-agnostic visual debugging tool for ML https://lnkd.in/gKeTNDB. Great collaborative work between the Visualization team, the ML Platform team, and Data Science Product teams to make this happen! #datascience #ml #visualisation
-
Lezhi Li shared thisHad fun teaching a class on Data Visualization for Data Science for UCDavis MSBA program #datascience #datavisualization
-
Lezhi Li shared thisLezhi Li shared thisJoin the Women in Data group at Uber on October 2 for their "Moving the World With Data" meetup. Let us know you're coming—RSVP here.
-
Lezhi Li shared thisLezhi Li shared thisInterested in working on Uber's open data platform, Movement, to help shape the future of cities? Our team is looking for people with backend and data engineering skills–get in touch!
-
Lezhi Li shared thisUber's ML platform. Proud to see screenshots of some of my machine learning visualization work :)Lezhi Li shared thisCheck out the blog post on Uber's ML Platform
-
Lezhi Li liked thisLezhi Li liked thisExcited to share recent research from our team on 𝗮𝗴𝗲𝗻𝘁𝗶𝗰 𝗥𝗟 and 𝗺𝘂𝗹𝘁𝗶-𝗮𝗴𝗲𝗻𝘁 𝘀𝘆𝘀𝘁𝗲𝗺𝘀 𝗳𝗼𝗿 𝗰𝗼𝗱𝗲 𝗼𝗽𝘁𝗶𝗺𝗶𝘇𝗮𝘁𝗶𝗼𝗻, with more to come! 𝟭. 𝗦𝗔𝗚𝗘: 𝗥𝗲𝗶𝗻𝗳𝗼𝗿𝗰𝗲𝗺𝗲𝗻𝘁 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴 𝗳𝗼𝗿 𝗦𝗲𝗹𝗳-𝗜𝗺𝗽𝗿𝗼𝘃𝗶𝗻𝗴 𝗔𝗴𝗲𝗻𝘁 𝘄𝗶𝘁𝗵 𝗦𝗸𝗶𝗹𝗹 𝗟𝗶𝗯𝗿𝗮𝗿𝘆 [ACL 2026] SAGE enables LLM agents to autonomously build and reuse skill libraries through RL, moving beyond static prompting. On AppWorld, it lifts Qwen2.5-32B from base ReAct performance of 𝟯𝟵.𝟮% 𝗧𝗚𝗖 / 𝟭𝟴.𝟲% 𝗦𝗚𝗖 to 𝟳𝟮.𝟬% 𝗧𝗚𝗖 / 𝟲𝟬.𝟳% 𝗦𝗚𝗖 (TGC = Task Goal Completion; SGC = Scenario Goal Completion, where all tasks in a scenario must succeed) — while using 𝟮𝟲% 𝗳𝗲𝘄𝗲𝗿 𝘀𝘁𝗲𝗽𝘀 and 𝟱𝟵% 𝗳𝗲𝘄𝗲𝗿 𝘁𝗼𝗸𝗲𝗻𝘀, outperforming proprietary models. 📄 https://lnkd.in/gYCxYKmE 𝟮. 𝗦𝗔𝗟𝗧: 𝗦𝘁𝗲𝗽-𝗹𝗲𝘃𝗲𝗹 𝗔𝗱𝘃𝗮𝗻𝘁𝗮𝗴𝗲 𝗔𝘀𝘀𝗶𝗴𝗻𝗺𝗲𝗻𝘁 𝗳𝗼𝗿 𝗟𝗼𝗻𝗴-𝗵𝗼𝗿𝗶𝘇𝗼𝗻 𝗔𝗴𝗲𝗻𝘁𝘀 𝘃𝗶𝗮 𝗧𝗿𝗮𝗷𝗲𝗰𝘁𝗼𝗿𝘆 𝗚𝗿𝗮𝗽𝗵 [EACL 2026] SALT addresses the coarse credit assignment problem in multi-turn agent RL by constructing trajectory graphs to assign fine-grained, step-level advantages. As a lightweight plug-and-play module, it consistently boosts existing RL methods — e.g., lifting both GRPO & RLOO baselines from 𝟴𝟭.𝟴% → 𝟴𝟱.𝟮% on ALFWorld and improving AppWorld TGC by +𝟰.𝟳𝗽𝗽 (Test-N) and +𝟲.𝟳𝗽𝗽 (Test-C). 📄 https://lnkd.in/gcN4aMED 𝟯. 𝗔𝗴𝗲𝗻𝘁𝗶𝗰 𝗖𝗼𝗱𝗲 𝗢𝗽𝘁𝗶𝗺𝗶𝘇𝗮𝘁𝗶𝗼𝗻 𝘃𝗶𝗮 𝗖𝗼𝗺𝗽𝗶𝗹𝗲𝗿-𝗟𝗟𝗠 𝗖𝗼𝗼𝗽𝗲𝗿𝗮𝘁𝗶𝗼𝗻 A multi-agent framework that pairs LLM creativity with compiler reliability for C++ code optimization. By coordinating LLM agents at multiple abstraction levels (C, IR, Assembly) with compiler tools, it achieves speedups of 𝘂𝗽 𝘁𝗼 𝟭.𝟮𝟱𝘅 over both compiler-only (clang -O3) and LLM-only baselines while maintaining correctness. 📄 https://lnkd.in/ghfk6Anz Shout out to our amazing team and interns for the great work: Jiongxiao Wang, Jiazheng Li, Benjamin Mikek, Yawei Wang, Qiaojing (David) Yan, Yijun Tian, Soumya Smruti Mishra, Zhichao Xu, Megha Gandhi, Huan Song, Danylo Vashchilenko, Bryan Lu, Lin Lee Cheong #NLP #ReinforcementLearning #LLMAgents #ACL2026 #EACL2026 #AI #MachineLearning #AgenticAI #CodeOptimization
-
Lezhi Li liked thisLezhi Li liked thisI gave Claude Code two GPUs and a week. Its team of agents ran ~500 experiments and built a state-of-the-art recommender model. Read my experience here, and follow my project at Github (link in the post below).I gave Claude Code two GPUs and a week. Its team of agents ran ~500 experiments and built a state-of-the-art recommender model.I gave Claude Code two GPUs and a week. Its team of agents ran ~500 experiments and built a state-of-the-art recommender model.Xu Ning
-
Lezhi Li liked thisLezhi Li liked thisA few weeks ago, I posted about knowledge management via Obsidian and Cursor to manage the many threads I had to juggle at work. While definitely a useful project, it felt incomplete and too small in scope. I wanted to build something that could help me keep track of all the written (and to some extent the spoken) content I consume - to build a personal knowledge base, a “Second Brain” if you will. I built project "Second Brain" in under a month - an LLM-powered system that ingests everything I read (papers, articles, books, podcasts, code repos), stores it in a knowledge graph, and then helps me actually learn from it. Built on top of the knowledge graph is an active learning system that is grounded in cognitive science: spaced repetition, active recall, AI-generated exercises, and adaptive difficulty. The system optimizes my learning and elevates my learning methods beyond just highlighting and re-reading (which are some of the least effective study techniques). Designing and building project “Second Brain” was a lesson in itself. The project was implemented almost entirely by Cursor / Claude Code - my key contribution wasn't the code but the spec. I spent most of my time iterating on design docs and implementation plans before a single line was written. When something broke, I fixed the spec first, not the code. In the age of powerful AI assistants, we're no longer constrained by execution speed. We're constrained by the clarity of our ideas. Full writeup: https://lnkd.in/dPUsqmgK GitHub: https://lnkd.in/dmmhijcM
-
Lezhi Li liked thisLezhi Li liked thisLudwig 0.11 is out and it's a foundation reset. After a long period of limited activity and some neglect, I went back into the Ludwig codebase and did something I think many open-source projects eventually need: I cleaned it up. Thoroughly. It was cathartic. Almost a spiritual experience. There was nostalgia in seeing the names of past contributors again, like stepping back into your childhood room and recognizing the layers of who you used to be. Over the past weeks I: - Addressed all open issues and PRs - Modernized dependencies (PyTorch, Ray, transformers, NumPy, MLflow and more) - Dropped older Python versions (3.8/3.9) - Removed legacy backends like Horovod, Neuropod, and LightGBM to simplify the architecture - Migrated configuration schemas from marshmallow to pydantic v2 - Fixed distributed training, Ray Tune checkpoints, and LoRA save/load edge cases - Reduced CI runtime by ~60–70% and significantly sped up the test suite - Refreshed documentation at https://ludwig.ai/ This shipped as: 0.11, the modernization release with two quick follow-up 0.11.1 and 0.11.2 releases with fixes and improvements additional stability, performance, and CI optimizations The project really needed a reset like this. The goal was simple: bring Ludwig back into a state where people can confidently use it, extend it, and learn from it today. If you have used Ludwig before, or were curious but hesitant because of ecosystem drift, now is a good time to take another look: https://lnkd.in/gA_SRdft Thanks to everyone who stuck around. #opensource #machinelearning #AI #python #deeplearning
-
Lezhi Li liked thisLezhi Li liked thisI started at Apple as a physicist trying to pivot into AI. I leave as a multimodal foundation model researcher, building models that generate images, video, and language at scale. Here’s a reflection on the journey and what comes next.From Cambridge Physics to Multimodal Foundation Models. Thank you, Apple.From Cambridge Physics to Multimodal Foundation Models. Thank you, Apple.Jesse Allardice
-
Lezhi Li liked thisLezhi Li liked thisA flywheel that needs humans to spin it is a treadmill. In 2026, the real turning point is rebuilding RL infra so coding agents can run the loop end-to-end: implement → debug → run → attribute → report. Shorter cycle time + cleaner signal = compounding capability. Full post: https://lnkd.in/gqkNRw84
-
Lezhi Li liked thisSharing a technical report on the Parallel Track (PT) transformer architecture (https://lnkd.in/g7j85Cit). This core design underpins PT-MoE and enables strong model quality with competitive cost efficiency on Apple’s Private Cloud Compute (PCC) platform. It was first described in the Apple Foundation Model Report 2025 (https://lnkd.in/g-U5K_WF). Joint work with Nan Du, Tom Gunter, Tao Lei, Kulin Seth, Senyu Tong, Jianyu Wang, Guoli Yin, Xiyou Z., Xuan Kelvin Zou and Ruoming Pang. [Special credit goes to Ruoming for sketching out the original idea.]
-
Lezhi Li liked thisLezhi Li liked thisCONFUSED ABOUT LIFE? UNSURE OF WHAT YOU'RE DOING? I began 2025 by sending an email with this exact subject line. Within 20 minutes, so many students responded that my weekends became completely booked with "Confusion Coffees." Over lattes and teas, I spent my final college semester listening to heartwrenching stories that often led to the same realization: sometimes growth isn't about improving who you are. Sometimes it's about shifting how you pay attention, what you consider to be true. Sometimes, growth is transformation: changing who you want to become. Last May, I graduated from Harvard (cum laude, high honors, Phi Beta Kappa!). Looking back, it's remarkable how much I've changed—majors, perspectives, directions, postgrad plans. It was, as Dean Rakesh Khurana would put it, a transformative experience. Some highlights from the past year: -designing immersive theater experiences of dating in 2040 with The Rithm Project and OddKnock Productions (joint datasets rather than joint bank accounts, anyone?) -performing slam poetry on AIxEducation at the LearnerStudio summit -dreaming up FutureShock with The Reinvention Lab at Teach For America -singing the Class Ode with Fahim Ahmed, Ari Cheriyan, Roseanne Strategos, and Henry Wu at commencement after years of waking up at 8am every day to sing Morning Prayers together -presenting Conflux Collective at CHI 2025 in Japan, with Sofia Chen, who took over the organization with so much grace -learning across the civil-military divide at Hertog Foundation and Institute for the Study of War as a War Studies Fellow -writing a chapter for the forthcoming Handbook of Motivation and Social Psychology with Hal Hershfield -launching the public version of Future You with Pat Pataranutaporn and Pattie Maes, presenting it at South By Southwest with Kate Gardiner and Christopher Bryan, then being named a 2025 Fast Company World Changing Idea, then showcasing Possible You (the multiverse version) at NeurIPS this December -making space for wholehearted perspective exchanges—between artists and engineers (shoutout Conflux Collective), community organizers and technologists (shoutout Engineering Hope), startup founders and mission-driven students (shoutout startups at harvard ) None of this would have been able to happen without the support of my family, friends, mentors, professors, and advisors during the many times of confusion, uncertainty, and change; to those of you reading this post: thank you for expanding my conception of the future and challenging me to create my own place within it—for encouraging me to transform into the person I am today. Now, I'm pursuing my PhD in psychology at Stanford, advised by Carol Dweck, where I am investigating the mechanisms underpinning how we grow, transform, become… (Thank you Vincent Po for the photos!)
Experience
Education
Licenses & Certifications
Courses
-
Computational Design
-
-
Computer Vision
6.869
-
Data Science
cs 109
-
Data Visualization
cs171
-
Functional Programming
cs51
Languages
-
English
Full professional proficiency
-
Chinese
Native or bilingual proficiency
-
German
Elementary proficiency
View Lezhi’s full profile
-
See who you know in common
-
Get introduced
-
Contact Lezhi directly
Other similar profiles
Explore more posts
-
Yizhe Zhang
Apple • 4K followers
We (w/ Shansan Gong, Ruixiang ZHANG, Huangjie Zheng, Jiatao Gu, Navdeep Jaitly, Lingpeng Kong) released a family of 7B diffusion language models, DiffuCoder, that specializes on code generation, with a focus on understanding and improving masked diffusion models. A core analysis of DiffuCoder is the autoregressiveness (AR-ness) score, a novel metric that quantifies the causal patterns in decoding, revealing how diffusion models break from strict left-to-right generation for more flexible, non-linear code planning. Recent advances in autoregressive (AR) models dominate code generation, but diffusion-based LLMs (dLLMs) like DiffuCoder offer a promising alternative, especially for complex programming tasks. DiffuCoder explores how these models decode differently—showing less global AR-ness in code tasks compared to math—and how temperature affects both token selection and generation order, unlike traditional AR models. We also introduce coupled-GRPO, a post-training RL method with a coupled-sampling scheme, to reduce performance drops during accelerated decoding, boosting parallelism and efficiency. We use a self-improvement pipeline that leverages AR-ness analysis, coupled-GRPO optimization, and evaluation on benchmarks like AceCode-89k to refine decoding strategies. This approach enables DiffuCoder to navigate diverse code generation pathways and enhance performance with modest computational overhead. Looking ahead, we aim to further leverage Reinforcement Learning to steer code generation through these decoding patterns, with the discrete nature of AR-ness scores providing a foundation for search-based strategies—ideal for the sparse rewards of optimizing complex code structures. Check out our full paper and code for a deeper dive! Paper: https://lnkd.in/gVWU3BDJ Code: https://lnkd.in/gmXTZ_6n Models: https://lnkd.in/gTcKCDr9 #MachineLearning #AI #CodeGeneration #DiffusionModels #NLP
220
5 Comments -
Nina Peñaflor
LLM Arena • 1K followers
👉 My Key Takeaways from Chip Huyen's Recent Interview on Lenny's Podcast Chip Huyen is the author of the widely recognized "AI Engineering: Building Applications with Foundation Models". Link to the podcast: https://lnkd.in/dxZ-tFWX 💡 Importance of post-training. Pre-training gives you raw capabilities (next token prediction on massive data), but post-training is what makes the model actually usable. SFT on high-quality examples + RLHF. Fine-tuning should be your last resort, not first. Most problems can be solved with better prompts, better data, or RAG. 💡 Evals. You can't improve what you can't measure. Need multiple types: unit tests (does this specific prompt work?), integration tests (does the whole pipeline work?), regression tests (did we break something?), and user feedback loops. The hardest part isn't writing evals; it's maintaining them as your product evolves. 💡AI products. Reliability and UX matter more than models. Most AI product failures aren't about bad models: they're about reliability (API limits, latency spikes, poor monitoring) and UX (users don't understand how to use it, doesn't fit workflow). Building reliable platforms and talking to users constantly beats chasing SOTA models. Most insights come from watching users, not from benchmarks. 💡How to improve AI-powered apps. What people think improves apps: staying current on AI news, chasing newest agentic framework, obsessing over vector database choice, constantly evaluating model benchmarks, fine-tuning models. What actually improves apps: talking to users, building reliable platforms, preparing better data, optimizing end-to-end workflows, writing better prompts. Better prompt engineering beats switching models 90% of the time. A well-crafted system prompt, clear instructions, good examples (few-shot), and proper output formatting can transform a mediocre experience into a great one. 💡 Advice for builders. Start with user problem, not with cool AI technique. Use the simplest solution that works (often that's a good prompt, not a fine-tuned model). Build evals early. Focus on end-to-end experience. Don't fine-tune unless you've exhausted everything else. Don't treat AI as deterministic (it's not, you need to handle variability). Don't ignore data quality (garbage in, garbage out).
1
1 Comment -
Julius Kusuma
Meta • 3K followers
We developed an open-source AI tool to design concrete mixes that are stronger, more sustainable, and ready to build with faster—speeding up construction while reducing environmental impact. https://lnkd.in/gPCk8tCM But the impact of this AI tool is not just hypothetical! Amrize used Meta’s AI-based technologies to design a new low-carbon mix, and successfully deployed it in an at-scale slab-on-grade application at Meta's new data center in Rosemont, MN. Compared to the legacy mix, this new AI-designed mix is: 🦁 Stronger ⏱ Faster 🍃 Lower carbon ⏱ ️The ideal set time All this was achieved without needing any new materials, nor special equipment. Best of all, the AI is open-sourced. https://lnkd.in/g2KA7KZW This work was featured in a Meta engineering blog article published today! https://lnkd.in/gBU9HY8H
98
9 Comments -
Tieying Zhang
ByteDance • 8K followers
Our team ByteBrain, in collaboration with Tsinghua University, proposed ChatTS, a time-series multimodal LLM, accepted to VLDB 2025. Here is a brief introduction: Understanding time series is crucial for its application in real-world scenarios. Recently, large language models (LLMs) have been increasingly applied to time series tasks, leveraging their strong language capabilities to enhance various applications. However, research on multimodal LLMs (MLLMs) for time series understanding and reasoning remains limited, primarily due to the scarcity of high-quality datasets that align time series with textual information. This paper introduces ChatTS, a novel MLLM designed for time series analysis. ChatTS treats time series as a modality, similar to how vision MLLMs process images, enabling it to perform both understanding and reasoning with time series. To address the scarcity of training data, we propose an attribute-based method for generating synthetic time series with detailed attribute descriptions. We further introduce Time Series Evol-Instruct, a novel approach that generates diverse time series Q&As, enhancing the model’s reasoning capabilities. To the best of our knowledge, ChatTS is the first MLLM that takes multivariate time series as input for understanding and reasoning, which is fine-tuned exclusively on synthetic datasets. We evaluate its performance using benchmark datasets with real-world data, including six alignment tasks and four reasoning tasks. Our results show that ChatTS significantly outperforms existing vision-based MLLMs (e.g., GPT-4o) and text/agent-based LLMs, achieving a 46.0% improvement in alignment tasks and a 25.8% improvement in reasoning tasks. We have open-sourced the source code, model checkpoint and datasets at https://lnkd.in/gYDJ9S4j Paper: https://lnkd.in/gqXW9tPY About ByteBrain Team: ByteBrain is a platform of AI for Infra (AI for System) at ByteDance. The goal is to leverage AI to automatically optimize system performance and reduce labor and resource costs. ByteBrain integrates machine learning, LLM and OR techniques to support both internal infrastructure systems at ByteDance as well as the Volcano Engine Cloud.
153
3 Comments -
Hemanth HM
PayPal • 5K followers
black-forest-labs/FLUX.2-klein-9B is crazy fun! Under FLUX Non-Commercial License, it bring: - Real-time AI agents with visual reasoning. - Instant, iterative design workflows. - Low-latency production pipelines on consumer-grade hardware. - Sub-second speed - Text-to-Image + Native Editing in one pipeline. - Fits in ~20GB VRAM (RTX 4090 ready). The 4-step distilled version is a game-changer for anyone looking to deploy high-quality vision models without the massive VRAM overhead. #GenerativeAI #MachineLearning #FLUX2 #OpenSource #AIEngineering
6
3 Comments -
Arjuna Anand
AmpleCase • 7K followers
🚨 breaking - ai apps & models launched in last 48 hours 1. hunyuan-a13b model open-sourced by tencent - compact yet powerful, moe architecture, 13b active, 80b total - hybrid model with flexibility to choose fast or slow thinking modes - 256k context window - enhanced agent task capabilities - efficient inference using gqa & multiple quantization methods 2. pangu models open-sourced by huawei - 7b & moe based pro 72b model - optimized for inference on huawei's ascend chips 3. ernie 4.5 models open-sourced by baidu - moe architecture with 10 variants - 0.3b dense model, 3b, 47b, 424b models - multimodal models 4. qwen-vlo by alibaba (already open-sourced) - upgrade from qwen2.5-vl - supports text to image & image to image generations - unique progressive generation feature where user can see steps of image generation
2
-
QiGoal Capital | 契阔集团
310 followers
AI's Next Chapter: Key Trends from China's NetEase Future Conference: At the 2025 NetEase Future Conference, Academician Pan Yunhe outlined a pivotal AI shift: a "two-wheel drive" of thought simulation (e.g., LLMs) and action simulation (e.g., robotics). China's new "AI Plus" action plan focuses on deep industrial integration. Here are three critical trends he identified: 𝐓𝐫𝐞𝐧𝐝 𝟏: 𝐓𝐡𝐞 𝐑𝐢𝐬𝐞 𝐨𝐟 𝐒𝐩𝐞𝐜𝐢𝐚𝐥𝐢𝐳𝐞𝐝 𝐋𝐚𝐫𝐠𝐞 𝐌𝐨𝐝𝐞𝐥𝐬 The "hallucination" problem becomes critical as AI Agents take action. • Solution: Use high-quality, "textbook-level" data from academia, industry, and real-world applications. • New Path: Proposes a dual approach: continue general models while also training specialized models first, then federating them into a general intelligence. 𝐓𝐫𝐞𝐧𝐝 𝟐: 𝐄𝐦𝐛𝐨𝐝𝐢𝐞𝐝 𝐈𝐧𝐭𝐞𝐥𝐥𝐢𝐠𝐞𝐧𝐜𝐞 𝐄𝐱𝐩𝐚𝐧𝐝𝐬 The concept is evolving beyond humanoid robots. • Broader Scope: Now includes drones, smart vehicles, and intelligent machinery. • Deeper Integration: Future systems will combine LLMs with visual, action, and mechanical models (e.g., sensors). Lightweight exoskeletons are a leading example. 𝐓𝐫𝐞𝐧𝐝 𝟑: 𝐀𝐈 𝐃𝐫𝐢𝐯𝐞𝐬 𝐭𝐡𝐞 "𝐏𝐥𝐚𝐭𝐟𝐨𝐫𝐦 𝐄𝐜𝐨𝐧𝐨𝐦𝐲 2.0" The next wave moves beyond life-service platforms (e-commerce) to industrial-tech platforms. • The New Frontier: Platforms for scientific research and industrial design, like Group Core Technology's interior design platform that automates processes for a network of 50,000 enterprises. These trends highlight AI's maturation towards reliable, industrial-scale application. #ArtificialIntelligence #AI #ChinaTech #Innovation #LargeLanguageModels #EmbodiedAI #PlatformEconomy #IndustrialAI #QiGoalCapital #QiGoal
1
-
Amit Yadav
Fern • 11K followers
Last few months, I have invested some of my personal time to guide AI startups and graduate students across several universities on projects related to GenerativeAI and Speech Processing. ✨ Excited to share promising findings from one of such collaborations led by awesome graduate students Sagnik and Abhiram from Stanford University. ⁉️ Challenge: Generative models for discrete data face two critical limitations: 1️⃣ They rely on continuous embeddings for inherently discrete distributions 2️⃣ They optimize variational bounds instead of exact likelihood, leading to inaccurate probability estimates 🎯 Our Solution: Information-Theoretic Discrete Poisson Diffusion Model (ItDPDM) - the first approach to tackle both issues simultaneously 📊 Results: Our theoretical proofs and exhaustive empirical validation demonstrate that ItDPDM achieves better likelihood estimates compared to existing generative methods for discrete data Super proud of Sagnik, Abhiram for extensive experiments to validate superiority of information theory driven discrete state modeling of discrete data. Please refer pre-print of the manuscript in the comment. Looking forward to how community use this work, feel free to DM to discuss the work. #generativeAI #diffusionmodel #AI #deeplearning #informationtheory
95
2 Comments
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content