sreeprasad Govindankutty
Sunnyvale, California, United States
7K followers
500+ connections
View mutual connections with sreeprasad
sreeprasad can introduce you to 10+ people at Reddit, Inc.
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
View mutual connections with sreeprasad
or
New to LinkedIn? Join now
By clicking Continue to join or sign in, you agree to LinkedIn’s User Agreement, Privacy Policy, and Cookie Policy.
About
• Coded and developed payment processing engine that enabled BlackRock to create…
Articles by sreeprasad
-
Top 5 engineering blogs to learn & apply
Top 5 engineering blogs to learn & apply
These are top 5 engineering blogs I read over weekends in no particular order LinkedIn Engineering blog - Join LinkedIn…
3
Activity
7K followers
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisTen years ago, I came back to Reddit. Twenty years and change ago (October 2005, if we're being exact), I got a call from Steve Huffman while I was grabbing coffee with a labmate. Paul Graham had suggested he hire me as Reddit's first engineer. I said yes before I hung up the phone. This week, I’ve decided to step down as CTO and take on a new role as Reddit's first Senior Technical Fellow. The last decade has been the honor of my career. We took a company a lot of people had written off and rebuilt it: the stack, the org, the culture, all while it was still in flight. Along the way, we built something the world uses—something millions of people rely on every day to learn, to connect, and to understand what’s happening around them. We went public. We found our footing in an AI-native internet without chasing every trend, and I’m even more optimistic now than I was then. The opportunity ahead for Reddit (and for the kind of internet we believe in) is bigger than ever. The right bet is still the one we made: be the source worth surfacing. Amit Puntambekar is stepping into Reddit’s CTO role, and I couldn't be more confident about where engineering is headed under their leadership. You'll be hearing a lot more from him, and I’d encourage you to pay attention. None of it was mine alone. Reddit has one of the best engineering teams in tech, and being in the room with them has been a gift. To everyone who built this with me over the last ten years and in the early days before all of it: thank you. More to come. https://lnkd.in/gsyMpMaP
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisOur Org is hosting another SnooSec event at our New York office on May 14th from 5 to 9pm ET focusing on Usable Security: The design or intent of security processes and controls that are focused on your users and their behavior. I’ll be there along with some of the brightest individuals I’ve had the chance to work with. If you are in the area and can attend, I’d love to catch up!
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisWow this one is amazing! One of the most persistent annoyances with serious LLM use is memory fragmentation. Every model has its own context window. Its own opaque “memory.” Its own idea of what it remembers about you. None of it is portable. None of it is reliable. So you end up rebuilding context. Over and over again. MemPalace might be the first thing I’ve seen that actually attacks this cleanly. Fully local. No API calls. A shared memory layer exposed via MCP so Claude, ChatGPT, Cursor, and Gemini can all pull from the same context. That alone is a big deal. But the design choice underneath it is the real unlock. They store raw interaction history, not summaries. That sounds obvious but it isn’t. Most “memory” systems compress aggressively and lose signal. This keeps the source of truth intact, then layers semantic retrieval and structure on top. The result is a reported 96.6% R@5 on LongMemEval and a ~34% lift over flat search. More importantly, it flips the ownership model. Cloud memory tools ask you to trust them with how you think. Your debugging trails. Your half-formed ideas. Your architecture decisions. That’s not a small ask. Keeping memory local, and explicitly deciding what gets shared with each model via MCP, feels like the right default. Not just for privacy, but for control and consistency across tools. 16k GitHub stars this early is a strong practitioner signal. I’m spinning this up today. The context window problem isn’t going away. If anything, it gets worse as you use more models. The tools that win are going to be the ones that give you durable, portable, user-owned memory. This looks like one of the first real attempts at that. https://lnkd.in/g-APJpSZGitHub - milla-jovovich/mempalace: The highest-scoring AI memory system ever benchmarked. And it's free.GitHub - milla-jovovich/mempalace: The highest-scoring AI memory system ever benchmarked. And it's free.
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisAfter an incredible 8 years at Meta, I am excited to share that I am joining Reddit this week. Meta has been the ride of a lifetime! Few resumes can boast a product with a billion users – let alone two! From the early days of Facebook Watch (today known as Reels) to MetaAI, with a tour of duty in Virtual Reality, it's been an honor to work with some of the most talented people in the industry. I will be forever grateful to the people who helped me along the way, from Fidji to Paresh and Vijaye, to Aigerim, Gabriel, and Vishal, and most recently Prashant and Nat. And in particular from my engineering leads: Bhavana, Christine, Victor, Tanya, Parth, Deepa, Ke, Xueyuan, Avinash, and so many other amazing colleagues, ... too many to mention. As for what's next for me at Reddit, I am joining the consumer product leadership team with Amit and Maria to help Reddit's communities grow and prosper in the age of AI. Human interaction will have enormous value as more and more content and online activity are produced by artificial intelligence – and Reddit is one of the most vibrant, open and authentic human communities at the heart of the internet. Don't hesitate to reach out… we are hiring! ;-)
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisCommunity Notes is not a moderation system. It’s one layer in a governance stack, and the industry is increasingly mistaking the layer for the whole architecture. The appeal is obvious: no editorial staff, no obvious political fingerprints, consensus-based ratings. Clean story. Neutral story. Scalable story. But the evidence underneath it is messier. A November 2025 preprint reported that, on X, only about 8.3% of submitted notes reached “helpful” status between January 2021 and January 2025, while 87.7% remained stuck in “Needs More Ratings.” It also reported declining contributor retention, with the share of authors still active six months after their first note falling from 63% in early 2023 to 44% in late 2024. And it argues the algorithm is highly sensitive to rater bias and manipulation: a relatively small minority of bad raters can strategically suppress targeted helpful notes before they ever surface. That’s not just a tuning problem. It’s a structural vulnerability that shows up under adversarial pressure — which is exactly the environment moderation systems have to survive. On Reddit, users vote to affect ranking through votes and sorting. Moderators write and enforce community-specific rules. Users can also vote on communities with their feet! Admins enforce Reddit’s sitewide rules and remove sitewide rule-violating content. From a community-governance standpoint, that admin layer is the exception-handling layer, not the primary operating model. Community Notes maps much more cleanly to the first layer of that stack than to a full moderation system. That’s why Meta’s shift matters. In January 2025, Meta said it would end its third-party fact-checking program in the United States and move toward Community Notes. In March 2025, it said it would begin testing that approach and initially use X’s open-source algorithm as the basis of its rating system. The real question is not whether that feels more neutral. It’s which failure modes now come with the design, what detection exists for them, and what backstops remain when coordination or bias targets the system. Expert fact-checking has real weaknesses: cost, latency, and scaling limits. Community Notes has real weaknesses too: coverage gaps, consensus bottlenecks, retention decay, and gaming risk. A separate January 2026 preprint found that collaboration can improve note quality, but that advantage weakens when contributors know one another’s political affiliations. Neither model is solved. What worries me is the industry framing this as a values upgrade instead of a governance trade-off. You don’t eliminate judgment by crowdsourcing it. You redistribute judgment, incentives, and attack surface. With the Trust & Safety Summit in London happening this week, and the agenda covering content moderation, misinformation, and AI moderation, I’m more interested in what practitioners are saying in the room than in what the press releases say. https://lnkd.in/gTJ7eUqCThe Benefit of Collective Intelligence in Community-Based Content Moderation is Limited by Overt Political SignallingThe Benefit of Collective Intelligence in Community-Based Content Moderation is Limited by Overt Political Signalling
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisMy team at Reddit, Inc. is growing, and we’re hiring a Manager to lead our Threat Detection & Enforcement Operations. This role is a great fit for someone with experience in Trust & Safety who thrives on solving systemic platform threats and mentoring a team of experts. It’s a remote-friendly role in the U.S.—check out the link below to learn more and apply! https://lnkd.in/eRGKPazgManager, Threat Detection & Enforcement OperationsManager, Threat Detection & Enforcement Operations
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisReddit is honored to be named to Fast Company's Most Innovative Companies list for 2026. As more of the internet fills with machine-generated content, human conversation has never been more important. Reddit is where people go for perspectives, experiences, and advice they can actually trust—because it comes from other people. With a clear strategy to grow our user base through smarter platform experiences and $2.2 billion in revenue last year, we’re demonstrating that authentic community is a powerful engine for growth. We’re just getting started. Read the full story: https://lnkd.in/gb8MNkmF
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted thisI’ve spent most of my career working in the cloud, and cloud hardware fails in ways your monitoring doesn't expect. Recently, we started seeing 1-2 incidents weekly where tail latencies spiked and SLOs broke. They were always traced back to multiple services on a single Kubernetes node having problems. The quick fix: yank the node, send it back to the cloud provider from whence it came, move on to the next problem. But our team dug deeper. The pattern wasn't random hardware failure. It was performance degradation so subtle that individual service health checks passed, but collective impact was real enough to hurt user experience. The interesting part isn't the technical root cause (though that's worth reading). It's that we had to build new tooling to detect this failure mode. Standard monitoring assumes binary states: working or broken. Reality is messier. Machines can be slow enough to matter but healthy enough to stay in rotation. Fun! At scale, "good enough" hardware in a distributed system becomes a reliability problem. The math is unforgiving when you multiply small performance hits across thousands of requests. https://lnkd.in/gPNbDsrX
-
sreeprasad Govindankutty reposted thissreeprasad Govindankutty reposted this"Now you're in SNOO YOOOOORK" 🎶 Hey, that's us in Times Square, repping The Internet Awards, an award show where everyone gets a vote. Voting is open now until February 28. Winners will be crowned on March 4. Yes, there are Reddit trophies. Yes, voters get them, too. Now, go rock the vote: https://lnkd.in/gAXC-5Q9
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisExcited to share my recent engineering blog on building a low-latency distributed locking mechanism for Snowflake’s query infrastructure, designed to protect billions of daily queries globally with near-zero overhead.Snowflake Distributed Query Execution: The Execution AnchorSnowflake Distributed Query Execution: The Execution Anchor
-
sreeprasad Govindankutty liked thisHappy to have made contributions to improving the security posture of Netflix Playground, even more special after becoming a dad.sreeprasad Govindankutty liked thisAnd Netflix Playground is officially Global! So excited to expand Netflix Playground with this global launch - and even more excited to add an amazing new game to the mix: Gabby’s Dollhouse! 🎉 https://lnkd.in/gk7NY9zEPlay Game | Netflix Playground | Download mobile gamePlay Game | Netflix Playground | Download mobile game
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisWe had a blast at Carbone Beach during F1. Some of the biggest people in the world were there.
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisMost notifications are noise. The ones we send at Reddit, Inc. are trying to be the opposite. Right content, right person, right moment, across push, email, and in-app, for 126M+ people every day. Zita DePetris is hiring an ML Manager to lead our Notifications Relevance team and shape what comes next. This is one of the more interesting recommendation problems in tech right now. You're surfacing the right thread, the right community, the right comment from the world's largest corpus of human conversation. Retrieval, ranking, sequential features, multi-objective optimization. Real production systems with real DAU and retention impact. What she's looking for: • 5+ years on large-scale ML, 2+ managing • Strong intuition for recommender systems (retrieval + ranking, in production) • Comfort with ambiguity and a bias to ship • A coach who pushes engineers into leaders Remote in the US. Base $230K to $322K, plus equity and benefits Reddit actually invests in. If you've ever wished a notification you got actually felt useful, come build that. Link in the comments. DMs open. #hiring #MachineLearning #RecSys #Reddit
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisHacker News has a completely open API, I had no idea (https://lnkd.in/egCXddup) no developer token needed. The top stories are as easy as: https://lnkd.in/efufxZHz And getting data from a story is as easy as https://lnkd.in/eFW6TPWA It is interesting that they have different endpoints for filtering stories e.g. /topstories.json, /askstories.json, /showstories.json, /jobstories vs query params.
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisWe're looking for a Staff Product Security Engineer to join our team at Reddit. If you've ever wanted to do product security work that actually matters at scale — this is it. We're securing the community of communities, and we need someone who can help us do it right. This isn't a "run scanners and file tickets" role. You'll be building the frameworks and guardrails that make it hard for engineers (and their AI coding assistants) to ship insecure code in the first place. Prevention over detection. Paved roads over speed bumps. The team is great. The problems are real. And the work you do here will touch a platform used by 100M+ people every day. You should come work with us if: - You've spent 8+ years in security or software engineering and you're tired of chasing the same bug classes over and over - You'd rather build the fix into the platform than write another Jira ticket about it - You know your way around Go, Python, or TypeScript - You're excited (not terrified) about what AI-assisted development means for security It's remote, the pay is strong ($217K-$304K + equity), and you'll be working alongside people who genuinely care about getting this stuff right. Come secure the front page of the internet with me. https://lnkd.in/gUXt2gF5 #Hiring #ProductSecurity #Reddit
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisExcited to start a new chapter at Google DeepMind, leading post-training efforts (agentic reinforcement learning, supervised fine-tuning, and evals) to make Gemini better at unprecedented use-cases!
-
sreeprasad Govindankutty liked thissreeprasad Govindankutty liked thisDo you remember the AtCoder Educational DP Contest—the one featuring 26 classic DP problems? It’s just been refreshed, with AtCoder adding 20 more problems, bringing the total to 66. If you’re looking for a solid roadmap to learn dynamic programming, these problems are definitely worth exploring. You can find the link to the full set in the comments.
Experience
Education
Licenses & Certifications
Projects
-
ResumeLink
ResumeLink is an application that leverages LinkedIn API to render profile information of any Linkedin user in multiple resume formats. The application also renders and represents network information of a Linkedin User using D3.js and Spring framework as bar graph showing total number of connections of each of your contacts.
-
Page Safe
-
Co-founded Page-Safe, an easy and secure way to share and collaborate selective pages within a document. This application permits users to share documents as well as selected content within documents securely with others. This start up was conceived and brought to fruition in 48 hours as part of RIT48, a 48 hour start up challenge.
Other creators -
PicturePins
-
See projectCoded PicturePins, a website where users can post picture and capture moments they cherish for life. Built on Ruby on Rails, this application uses Amazon S3 to store images and is hosted on Heroku.
Recommendations received
3 people have recommended sreeprasad
Join now to viewView sreeprasad’s full profile
-
See who you know in common
-
Get introduced
-
Contact sreeprasad directly
Other similar profiles
Explore more posts
-
Soundararajan Srinivasan
Microsoft • 7K followers
🧪 If verification asks “is it correct?” but we also need to ask “what is happening over time?” In an earlier post, I argued that evals shouldn’t be treated as QA artifacts added after the fact and that they should be executable intent: acceptance criteria made runnable, measurable, and repeatable. The research on AI‑assisted development reinforces the next step of that argument. As David and I have been discussing, we need to expand how we think about quality: Verification → correctness against rules Validation → fitness against requirements Scientific evaluation → emergence, drift, accumulation of complexity, long‑horizon effects of local optimizations, and whether the system stays within its intended operating envelope. This is standard thinking in science but still rare in software engineering practice. Why does this matter? AI‑driven development collapses the boundary between building and optimizing. When intervention happens faster than understanding: velocity becomes indistinguishable from progress, short‑term gains hide long‑term drag, and technical debt becomes invisible until it’s systemic! That’s exactly what the Cursor study shows: speed first, complexity later, slowdown eventually. If your AI development process optimizes speed without continuously evaluating what that speed is doing to the system, you are accumulating invisible technical and organizational debt. The next maturity curve isn’t better prompts or faster agents. It’s treating evaluation as a scientific instrument, not just a software gate. #Evals #AIGovernance #AIProductManagement #AIInfrastructure #AgenticAI
9
3 Comments -
Nikhil Shrivastava
Demystify Systems • 10K followers
Struggling to ensure your LLM performs reliably in production? Here is an interesting article from my wife Sonakshi Singh This framework turns SME-labeled ground truth into a scalable evaluation pipeline. Goal Classify Q&A pairs into Correct, Incorrect, and Partially Correct in a live, production-grade system. Challenges 1.SMEs have limited availability and often provide only partial ground truth, without offering rationale for every record. 2.Understanding the rationale behind SME responses at scale is difficult. 3.Once we understand these rationales, the challenge is to use them to classify new Q&A pairs into Correct, Incorrect, or Partially Correct. Multistep Solution Step 1 – Use LLMs to extract the rationale behind SME responses. The LLM will classify them into the top 5 rationales (labels) with confidence scores. To achieve this, we provide all the available ground truth to the LLM and ask it to derive the top 5 rationales with confidence scores that illustrate why SMEs classified the pairs as they did. Step 2 – For each new Q&A pair, pass these rationales with confidence scores to another LLM and ask it to classify the pair into Correct, Incorrect, or Partially Correct. This, however, is a simplified view—using only 5 rationales and associated confidence scores per Q&A pair. In reality, a production-grade system would involve many more rationales and millions of Q&A pairs, resulting in very high token consumption and cost. Step 3 – To reduce dependency on an LLM-based classifier, we could build an in-house multi-class classifier. By combining the ground truth with LLM-generated labels and confidence scores, we can train this classifier to predict Correct, Incorrect, or Partially Correct. Step 4 – For each new live Q&A pair, use the LLM only to generate rationales (mapped to the predefined 5 labels) along with confidence scores. Step 5 – Pass this enriched dataset as input to the in-house multi-class classifier, which will then predict Correct, Incorrect, or Partially Correct. Conclusion With this approach, combining LLMs (to generate rationales) and a traditional ML classifier (to perform scalable classification), we can efficiently classify outputs into Correct, Incorrect, and Partially Correct at scale, while also creating a feedback loop to continuously improve performance. Closing the Gap: Production-Grade Evaluation Frameworks for RAG https://lnkd.in/eMJv8JNU
2
-
Alexandre Juca
SOFTLAB LDA • 4K followers
Existing LLMs are only as good as their training data. So what happens if a new frontend library is built that is wholly different to existing ones like React, Vue .js etc? Will models need to be updated with enough code examples from Github, etc or will official documentation be sufficient for it to build a frontend in said technology? Also, if LLMs are really good at building React apps, would that mean that overtime React will continue to solidify its position as the de-facto standard, especially amongst vibe coders? And going one step further, does this mean that with AI we would see less technological innovation, as the burden of development is moved from the developer to the agentic coder? Do we not risk long term future innovation if we go full auto-pilot and stop building systems ourselves?
19
14 Comments -
Eric Beans
Appuix, Inc. • 3K followers
We just shipped full Redis protocol parity for Cachee — 133+ commands, 8 data structures, sub-microsecond reads. Cachee is an AI-powered tiered caching system that sits in front of your existing Redis and makes it 667x faster than Elasticache alone. The L1 layer runs in-process — no TCP, no serialization, no network round-trip. Cache hits resolve in under 1 microsecond. That's 100,000x faster than a standard Redis (non elasticache) call. Just shipped: Full Redis command coverage. Strings, hashes, lists, sets, sorted sets, streams, HyperLogLog, bitmaps. Every command your Redis client already uses works out of the box. Change one connection string. That's the migration. Stream consumer groups. XREADGROUP, XACK, XPENDING, XCLAIM — the full lifecycle for building reliable event-driven architectures. Consumers crash, messages get stuck, ownership transfers to healthy workers. No messages lost. 8 eviction policies configurable at runtime. Our proprietary tiny-cachee policy uses a Count-Min Sketch for frequency estimation with Segmented LRU for recency — it adapts to your workload without tuning. Plus allkeys-lru, allkeys-lfu, volatile variants, and noeviction for safety-critical data. MULTI/EXEC transactions and Lua scripting. Atomic multi-key updates with optimistic locking. Push complex logic into the cache layer with EVAL and eliminate application-to-cache round-trips. Built-in observability. Prometheus metrics endpoint, SLOWLOG for latency debugging, CONFIG GET/SET for runtime tuning, CLIENT LIST, MEMORY USAGE — all without external dependencies. Why this matters: Every cache miss costs you latency. Every millisecond of latency costs you revenue. AWS published data showing that every 100ms of added latency reduces sales by 1%. Google found that a 500ms delay drops traffic by 20%. Most teams accept this tax because migrating off Redis means rewriting application code. We eliminated that barrier. Your redis-py, ioredis, Jedis, or go-redis client connects to Cachee and every command works. The hot path gets 100,000x faster. The cold path falls through to your existing Redis transparently. 227 automated tests verify compatibility across every data structure, eviction policy, transaction, and consumer group operation. One connection string change. Sub-microsecond reads. Full Redis compatibility. Learn more at cachee.ai #caching #redis #performance #infrastructure #devtools #opensource #engineering
-
Selva Ganapathy
Mintmesh Corporation • 1K followers
𝗟𝗟𝗠 𝗶𝗻𝘁𝗲𝗴𝗿𝗮𝘁𝗶𝗼𝗻 𝘁𝗼 𝗼𝘂𝗿 𝗮𝗽𝗽𝗹𝗶𝗰𝗮𝘁𝗶𝗼𝗻 𝘁𝘂𝗿𝗻𝗲𝗱 𝗱𝗲𝗯𝘂𝗴𝗴𝗶𝗻𝗴 𝗶𝗻𝘁𝗼 𝗮 𝗻𝗶𝗴𝗵𝘁𝗺𝗮𝗿𝗲. 𝗛𝗲𝗿𝗲'𝘀 𝗵𝗼𝘄 𝘄𝗲 𝗳𝗶𝘅𝗲𝗱 𝗶𝘁 𝗶𝗻 𝟱 𝗱𝗮𝘆𝘀 𝘄𝗶𝘁𝗵 𝘇𝗲𝗿𝗼 𝗻𝗲𝘄 𝘁𝗼𝗼𝗹𝘀. 𝗧𝗵𝗲 𝗣𝗿𝗼𝗯𝗹𝗲𝗺: A single user request triggers 200+ LLM calls. Which one failed? Where did context break? Our MTTR became as high as 7 hours. As a Startup, we cannot afford to have our reputation at stake. 𝗧𝗵𝗲 𝗦𝘁𝗮𝗿𝘁𝘂𝗽 𝗗𝗶𝗹𝗲𝗺𝗺𝗮: Everyone suggested: LangSmith, LangFuse, Helicone. - Evaluating tools = 2-3 weeks - Every new tool = Learning curve + cost + vendor management - We don't even know what we need yet 𝗙𝗶𝗿𝘀𝘁 𝗣𝗿𝗶𝗻𝗰𝗶𝗽𝗹𝗲𝘀 𝗤𝘂𝗲𝘀𝘁𝗶𝗼𝗻𝘀: What exactly do we need to debug? Input sent, response received, post-processing results, and relationships between 200+ calls. The insight: Not a logging problem. It's a 𝘁𝗿𝗮𝗰𝗶𝗻𝗴 𝗽𝗿𝗼𝗯𝗹𝗲𝗺. Each LLM call is stateless. Our application maintains context. Same applies to observability. 𝗧𝗵𝗲 𝗦𝗼𝗹𝘂𝘁𝗶𝗼𝗻: We already had Grafana Loki, Tempo, and OpenTelemetry. The breakthrough: Connect all LLM calls under one parent span. 𝗔𝗿𝗰𝗵𝗶𝘁𝗲𝗰𝘁𝘂𝗿𝗲 (𝗿𝗲𝗳𝗲𝗿 𝗱𝗶𝗮𝗴𝗿𝗮𝗺): 𝟭. 𝗔𝗽𝗽𝗹𝗶𝗰𝗮𝘁𝗶𝗼𝗻 𝗟𝗮𝘆𝗲𝗿 OpenTelemetry emits logs and traces 𝟮. 𝗟𝗼𝗴𝘀 (𝗚𝗿𝗮𝗳𝗮𝗻𝗮 𝗟𝗼𝗸𝗶) Captures each LLM interaction: - trace_id & span_id - Message sent & response received 𝟯. 𝗧𝗿𝗮𝗰𝗲𝘀 (𝗚𝗿𝗮𝗳𝗮𝗻𝗮 𝗧𝗲𝗺𝗽𝗼) Parent span connects all related calls: - "Analyze PDF" → Factor 1 (Profit) → Factor 10 - Each factor → LLM Call 1... Call 20 Shows serial vs parallel execution, timing 𝟰. 𝗢𝗯𝘀𝗲𝗿𝘃𝗮𝗯𝗶𝗹𝗶𝘁𝘆 𝗨𝗜 (𝗚𝗿𝗮𝗳𝗮𝗻𝗮) Query by span_id → Complete journey with all logs 𝗧𝗵𝗲 𝗥𝗲𝘀𝘂𝗹𝘁𝘀: - MTTR: 7 hours → 30 minutes - Timeline: 5 days(2 design + 2 dev + 1 infra) - Cost: $0 in new tools - Complexity: Zero new vendors 𝗞𝗲𝘆 𝗗𝗲𝘀𝗶𝗴𝗻 𝗗𝗲𝗰𝗶𝘀𝗶𝗼𝗻𝘀: 𝗪𝗵𝘆 𝗧𝗿𝗮𝗰𝗲𝘀 𝗢𝘃𝗲𝗿 𝗟𝗼𝗴𝘀? Logs show events. Traces show relationships. For 200+ interconnected calls, relationships matter. 𝗪𝗵𝘆 𝗘𝘅𝗶𝘀𝘁𝗶𝗻𝗴 𝗦𝘁𝗮𝗰𝗸 𝗢𝘃𝗲𝗿 𝗡𝗲𝘄 𝗧𝗼𝗼𝗹𝘀? Existing stack solved 80%. New tools solve 100% but cost 5x in time, money, and complexity. 𝗪𝗵𝘆 𝗣𝗮𝗿𝗲𝗻𝘁-𝗖𝗵𝗶𝗹𝗱 𝗦𝗽𝗮𝗻𝘀? Mirrors how our application maintains state. One request → Multiple factors → Multiple LLM calls. 𝗦𝘁𝗮𝗿𝘁𝘂𝗽 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 𝗟𝗲𝘀𝘀𝗼𝗻𝘀: Don't: Add new tools before understanding the problem Do: Define what you need first (saved 3 weeks) Don't: Evaluate 5 tools when you don't know what you need Do: Use existing stack, learn what's missing, choose deliberately Don't: Wait for the "perfect" solution Do: Ship 80% solution in 5 days, iterate based on production issues We'll eventually need specialized AI observability tools. But not today. Tomorrow, we'll know exactly which tool solves which problem. That's 𝗦𝘁𝗮𝗿𝘁𝘂𝗽 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴: Choose simple over perfect solution. Solve for 80%. Iterate to 100%.
28
2 Comments -
Hello Interview
20K followers
Why is Redis so fast? There are many design decisions that make Redis blisteringly quick. We’ll focus on two that carry most of the weight. - First is Redis’s in-memory design + single-threaded event loop. Keeping data in RAM and processing commands in one core loop avoids locks and random I/O. - Second is its purpose-built data structures and lean wire protocol. Most commands hit O(1) or O(log N) paths with tiny per-op overhead. What happens when you read from Redis👇 Step 1: A client sends a command using RESP (a simple, compact text/binary protocol). Minimal parsing and small payloads reduce CPU and network overhead. No big JSON/XML blobs. Step 2: Redis’s event loop accepts the socket and queues work. Single-threaded command execution means no lock contention or context-switch thrash. If you want to use more CPUs you need to run more Redis instances in a cluster. Step 3: The command is parsed and routed via a hash table lookup to the target keyspace entry. Lookups are O(1) on average thanks to efficient dictionaries and cache-friendly memory layouts. Step 4: The operation runs entirely in memory using specialized structures: Strings, lists (quicklist), hashes (compact encodings), sets/intsets, sorted sets (skiplist + hash), streams, etc. They're are engineered for predictable, fast operations with tight CPU caches. Step 5: The response is written back through the same event loop. Pipelining and batching can amortize syscalls and round trips, pushing throughput even higher. Step 6: Persistence and replication are off the critical path. AOF uses append-only, sequential writes with configurable fsync; RDB snapshots happen in a child process. Basically: we can lose data! This is a tradeoff Redis makes. Replication is async by default. The slow stuff is handled in the background so the hot path stays hot. Because Redis keeps data in RAM, executes commands in a single, lock-free event loop, and uses highly optimized data structures and a lean protocol, it avoids the latency traps of disk I/O, heavy parsing, and lock contention. That’s why it’s fast. Reed more about Redis here! https://lnkd.in/g2TEVzvx
163
-
Kurt Cagle
The Cagle Report • 27K followers
This is the next stage of AI. I've been coming to the same conclusion, albeit from a slightly different direction. Have you ever had a long conversation, and in the middle of things something's brought up that you think is really insightful, but then you lose the thread and you can only dimly remember what it was that you talked about? That's the current state of Transformers. Ideas are amorphous, often are highly contextual, and they shift and evolve over time. Contexts are both finite and bounded. Internally we abstract, but abstraction is almost invariably lossy. The next possible stage is persistent memory. RAG by itself isn't enough - it's primarily just a hack of LangChain to allow external services. What we need is a way of growing and holding named in-memory graphs so that they retain persistence. We need addressibility into that graph. Maybe Deepseek's architecture will be the one to do so, maybe someone else will figure out an alternative, but I think that has to happen for language models to get (mostly) past decaying coherence.
46
8 Comments -
Christopher Gulliver
DraftKings Inc. • 3K followers
A lot of posts recently talking about the shift away from the physical act of writing code/syntax going around so I’ll weigh in with my personal experience. 2025 was a watershed moment for LLMs. With Claude Code it was finally “smart” enough for mass adoption. It now writes code better than most “programmers”. But here’s the important distinction. Engineering and Programming are not the same. As engineering leadership I see it every day. People who think writing code is their job, not requirment gathering, planning, designing and architecting systems, validating correctness, problem solving. People who have these skills, can make effective use of Claude. These are the successful early adopters and 10x engineers right now. Now here’s what that means for me personally. After a long time, I am enjoying being a software engineer again. There was a time I was deeply married to writing code (it was kind of the therapeutic part of software development for me). But the shift into people management, becoming a father, and other life commitments has slowly erroded any time I have to take on projects that require hours or continuous focus. Writing code became the moat around me inventing. That is no longer the case. LLMs have enabled me to enjoy software development again. I can work in more bite sized chunks that fit into my day. I can engineer solutions in idle time. I can invest in custom tooling quickly that would have been month long side quests before. I can dip into other development domains (backend, browser extensions, languages) that previously would have take months of learning to become effective. This is what it has meant to me, do you have any similar experiences?
40
40 Comments -
Marcelo Lewin
The Cigna Group • 7K followers
Direct quote from Andrej Karpathy on X. https://lnkd.in/gqMxQ-is "I've never felt this much behind as a programmer. The profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between. I have a sense that I could be 10X more powerful if I just properly string together what has become available over the last ~year and a failure to claim the boost feels decidedly like skill issue. There's a new programmable layer of abstraction to master (in addition to the usual layers below) involving agents, subagents, their prompts, contexts, memory, modes, permissions, tools, plugins, skills, hooks, MCP, LSP, slash commands, workflows, IDE integrations, and a need to build an all-encompassing mental model for strengths and pitfalls of fundamentally stochastic, fallible, unintelligible and changing entities suddenly intermingled with what used to be good old fashioned engineering. Clearly some powerful alien tool was handed around except it comes with no manual and everyone has to figure out how to hold it and operate it, while the resulting magnitude 9 earthquake is rocking the profession. Roll up your sleeves to not fall behind." Key takeaways for me: 1) The playing field has partially reset. Nobody has decades of experience with this new “AI layer,” but strong fundamentals still compound faster than ever. Learn the new layer and keep sharpening the fundamentals. 2) The real skill gap is learning to think at the agent and orchestration layer. This is not just about writing prompts. It’s about designing systems that intentionally combine tools, context, memory, and workflows. 3) Everyone is figuring this out in public (including me). There are no true masters yet, but through hands-on building is how real expertise is forming. 4) Passive learning (or worse, ignoring that AI is here to stay) is the fastest way to fall behind. The only way forward is rolling up your sleeves and getting your hands dirty. This isn’t something to fear. It’s an inflection point. A paradigm shift. For everyone. If you engage, experiment, and keep learning, 2026 is going to be an incredible ride.
2
-
Ilyes.T. M.
Y.I.N. Technologies LLC… • 10K followers
Google Antigravity exposes the critical flaw in autonomous AI agent architectures: trust based governance at scale. Antigravity represents a major shift in how developers work. Instead of writing code line by line, you delegate entire tasks to autonomous agents that can modify files, run tests, browse the web, and execute changes across your codebase in parallel. The problem? These agents operate on trust, not proof. One developer reported: On Day 3, an agent confidently refactored a utility function and silently deleted a critical edge case check. This is not a bug. This is the inevitable result of autonomous agents operating without cryptographic authority validation. When you have three agents working asynchronously across different files, two critical questions emerge: how do you enforce what each agent is authorized to do, and when multiple agents coordinate on shared resources, how do you maintain isolation between their operations? Policy based guardrails do not work at this scale. I have solved both problems through complementary cryptographic architectures. For individual agent authorization, my 13 layer cryptographic governance system validates AI agent authority mathematically before execution. For multi agent coordination, my YIN COLLAB architecture implements Agent Specific Compliance Tokens with per agent privacy isolation maintaining cryptographic boundaries preventing cross contamination. Every action carries immutable proof of authorization. Every decision boundary is pre validated cryptographically. 26 USPTO patents. 2,330 claims. Validated with 640x timing resistance and 500 plus concurrent agent support with sub 15ms latency. Mathematical proof, not policy promises. For developers using Antigravity, Cursor, or any autonomous agent platform: the question is whether you can prove mathematically that unauthorized operations are impossible and that multi agent coordination maintains isolation. Because as agent orchestration becomes the dominant development paradigm, the liability surface expands exponentially. One misconfigured agent with access to production systems is an organizational failure. One agent leaking sensitive data to another agent through shared context is an architectural vulnerability. Making it mathematically impossible for agents to violate boundaries is the only governance model that scales. Autonomous agents are the future of software development. Cryptographic governance is the only way to make that future safe. https://lnkd.in/ejPREk9D #GoogleAntigravity #AIGovernance #AutonomousAgents #Cybersecurity #AIAgents #DeveloperTools #CryptographicSecurity #ZeroTrust #AICompliance #SoftwareDevelopment #TechInnovation #AIEthics
2
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content