Register allocation in the Go compiler (2024)
11 by elvis70 | 1 comments on Hacker News.
Tuesday, September 30, 2025
Monday, September 29, 2025
Sunday, September 28, 2025
Saturday, September 27, 2025
Friday, September 26, 2025
Thursday, September 25, 2025
Wednesday, September 24, 2025
Tuesday, September 23, 2025
Monday, September 22, 2025
Sunday, September 21, 2025
Saturday, September 20, 2025
Friday, September 19, 2025
Thursday, September 18, 2025
New top story on Hacker News: Learn Your Way: Reimagining Textbooks with Generative AI
Learn Your Way: Reimagining Textbooks with Generative AI
46 by FromTheArchives | 12 comments on Hacker News.
46 by FromTheArchives | 12 comments on Hacker News.
Wednesday, September 17, 2025
Tuesday, September 16, 2025
Monday, September 15, 2025
New top story on Hacker News: Show HN: MCP Server Installation Instructions Generator
Show HN: MCP Server Installation Instructions Generator
7 by pmig | 0 comments on Hacker News.
Hey HN, we’ve been experimenting a lot with MCP servers lately, and one of the most time-consuming challenges has been connecting MCP clients to remote MCP servers. To solve this, we built a library that generates them on the fly, enabling 1-click installation buttons and links for most clients out there. Feel free to try out the generator and use it to improve the README of your remote MCP server with the generated markdown. You can even configure the library to return HTML instructions if someone accesses your remote MCP server via the web.
7 by pmig | 0 comments on Hacker News.
Hey HN, we’ve been experimenting a lot with MCP servers lately, and one of the most time-consuming challenges has been connecting MCP clients to remote MCP servers. To solve this, we built a library that generates them on the fly, enabling 1-click installation buttons and links for most clients out there. Feel free to try out the generator and use it to improve the README of your remote MCP server with the generated markdown. You can even configure the library to return HTML instructions if someone accesses your remote MCP server via the web.
New top story on Hacker News: Show HN: AI-powered web service combining FastAPI, Pydantic-AI, and MCP servers
Show HN: AI-powered web service combining FastAPI, Pydantic-AI, and MCP servers
10 by Aherontas | 2 comments on Hacker News.
Hey all! I recently gave a workshop talk at PyCon Greece 2025 about building production-ready agent systems. To check the workshop, I put together a demo repo: (I will add the slides too soon in my blog: https://ift.tt/i4P5nZh ) https://ift.tt/KICv5HM... The idea was to show how multiple AI agents can collaborate using FastAPI + Pydantic-AI, with protocols like MCP (Model Context Protocol) and A2A (Agent-to-Agent) for safe communication and orchestration. Features: - Multiple agents running in containers - MCP servers (Brave search, GitHub, filesystem, etc.) as tools - A2A communication between services - Minimal UI for experimentation for Tech Trend - repo analysis I built this repo because most agent frameworks look great in isolated demos, but fall apart when you try to glue agents together into a real application. My goal was to help people experiment with these patterns and move closer to real-world use cases. It’s not production-grade, but would love feedback, criticism, or war stories from anyone who’s tried building actual multi-agent systems. Big questions: Do you think agent-to-agent protocols like MCP/A2A will stick? Or will the future be mostly single powerful LLMs with plugin stacks? Thanks — excited to hear what the HN crowd thinks!
10 by Aherontas | 2 comments on Hacker News.
Hey all! I recently gave a workshop talk at PyCon Greece 2025 about building production-ready agent systems. To check the workshop, I put together a demo repo: (I will add the slides too soon in my blog: https://ift.tt/i4P5nZh ) https://ift.tt/KICv5HM... The idea was to show how multiple AI agents can collaborate using FastAPI + Pydantic-AI, with protocols like MCP (Model Context Protocol) and A2A (Agent-to-Agent) for safe communication and orchestration. Features: - Multiple agents running in containers - MCP servers (Brave search, GitHub, filesystem, etc.) as tools - A2A communication between services - Minimal UI for experimentation for Tech Trend - repo analysis I built this repo because most agent frameworks look great in isolated demos, but fall apart when you try to glue agents together into a real application. My goal was to help people experiment with these patterns and move closer to real-world use cases. It’s not production-grade, but would love feedback, criticism, or war stories from anyone who’s tried building actual multi-agent systems. Big questions: Do you think agent-to-agent protocols like MCP/A2A will stick? Or will the future be mostly single powerful LLMs with plugin stacks? Thanks — excited to hear what the HN crowd thinks!
Sunday, September 14, 2025
Saturday, September 13, 2025
Friday, September 12, 2025
Thursday, September 11, 2025
Wednesday, September 10, 2025
New top story on Hacker News: Launch HN: Recall.ai (YC W20) – API for meeting recordings and transcripts
Launch HN: Recall.ai (YC W20) – API for meeting recordings and transcripts
13 by davidgu | 3 comments on Hacker News.
Hey HN, we're David and Amanda from Recall.ai ( https://www.recall.ai ). Today we’re launching our Desktop Recording SDK, a way to get meeting data without a bot in the meeting: https://ift.tt/r69hZCB . It’s our biggest release in quite a while so we thought we’d finally do our Launch HN :) Here’s a demo that shows it producing a transcript from a meeting, followed by examples in code: https://www.youtube.com/watch?v=4croAGGiKTA . API docs are at https://docs.recall.ai/ . Back in W20, our first product was an API that lets you send a bot participant into a meeting. This gives developers access to audio/video streams and other data in the meeting. Today, this API powers most of the meeting recording products on the market. Recently, meeting recording through a desktop form factor instead of a bot has become popular. Many products like Notion and ChatGPT have added desktop recording functionality, and LLMs have made it easier to work with unstructured transcripts. But it’s actually hard to reliably record meetings at scale with a desktop app, and most developers who want to add recording functionality don’t want to build all this infrastructure. Doing a basic recording with just the microphone and system audio is fairly straightforward since you can just use the system APIs. But it gets a lot harder when you want to capture speaker names, produce a video recording, get real-time data, or run this in production at large scale: - Capturing speaker names involves using accessibility APIs to screen-scrape the video conference window to monitor who is speaking at what time. When video conferencing platforms change their UI, we must ship a change immediately, so this keeps working. - Producing a video recording that is clean, and doesn’t capture the video conferencing platform UI involves detecting the participant tiles, cropping them out, and compositing them together into a clean video recording. - Because the desktop recording code runs on end-user machines, we need to make it as efficient as possible. This means writing highly platform-optimized code, taking advantage of hardware encoders when available, and spending a lot of time doing profiling and performance testing. Meeting recording has zero margin for failure because if anything breaks, you lose the data forever. Reliability is especially important, which dramatically increases the amount of engineering effort required. Our Desktop Recording SDK takes care of all this and lets developers build meeting recording features into their desktop apps, so they can record both video conferences and in-person meetings without a bot. We built Recall.ai because we experienced this problem ourselves. At our first startup, we built a tool for product managers that included a meeting recording feature. 70% of our engineering time was taken up by just this feature! We ended up starting Recall.ai to solve this instead. Since then, over 2000 companies use us to power their recording features, e.g. Hubspot for sales call recording, Clickup for their AI note taker. Our users are engineering teams building commercial products for financial services, telehealth, incident management, sales, interviewing, and more. We also power internal tooling for large enterprises. Running this sort of infrastructure has led to unexpected technical challenges! For example, we had to debug a 1 in 36 million segfault in our audio encoder ( https://ift.tt/iWZcm7P... ), we encountered a Postgres lock-up that only occurs when you have tens of thousands of concurrent writers ( https://ift.tt/V7pQABN ), and we saved over $1M a year on AWS by optimizing the way we shuffle data around between our processes ( https://ift.tt/rjna9Gi ). You can try it here: https://www.recall.ai . It's self-serve with $5 of free credits. Pricing starts at $0.70 for every hour of recording, prorated to the second. We offer volume discounts with scale. All data recorded through Recall.ai is the property of our customers, we support 0-day retention, and we don’t train models on customer data. We would love your feedback!
13 by davidgu | 3 comments on Hacker News.
Hey HN, we're David and Amanda from Recall.ai ( https://www.recall.ai ). Today we’re launching our Desktop Recording SDK, a way to get meeting data without a bot in the meeting: https://ift.tt/r69hZCB . It’s our biggest release in quite a while so we thought we’d finally do our Launch HN :) Here’s a demo that shows it producing a transcript from a meeting, followed by examples in code: https://www.youtube.com/watch?v=4croAGGiKTA . API docs are at https://docs.recall.ai/ . Back in W20, our first product was an API that lets you send a bot participant into a meeting. This gives developers access to audio/video streams and other data in the meeting. Today, this API powers most of the meeting recording products on the market. Recently, meeting recording through a desktop form factor instead of a bot has become popular. Many products like Notion and ChatGPT have added desktop recording functionality, and LLMs have made it easier to work with unstructured transcripts. But it’s actually hard to reliably record meetings at scale with a desktop app, and most developers who want to add recording functionality don’t want to build all this infrastructure. Doing a basic recording with just the microphone and system audio is fairly straightforward since you can just use the system APIs. But it gets a lot harder when you want to capture speaker names, produce a video recording, get real-time data, or run this in production at large scale: - Capturing speaker names involves using accessibility APIs to screen-scrape the video conference window to monitor who is speaking at what time. When video conferencing platforms change their UI, we must ship a change immediately, so this keeps working. - Producing a video recording that is clean, and doesn’t capture the video conferencing platform UI involves detecting the participant tiles, cropping them out, and compositing them together into a clean video recording. - Because the desktop recording code runs on end-user machines, we need to make it as efficient as possible. This means writing highly platform-optimized code, taking advantage of hardware encoders when available, and spending a lot of time doing profiling and performance testing. Meeting recording has zero margin for failure because if anything breaks, you lose the data forever. Reliability is especially important, which dramatically increases the amount of engineering effort required. Our Desktop Recording SDK takes care of all this and lets developers build meeting recording features into their desktop apps, so they can record both video conferences and in-person meetings without a bot. We built Recall.ai because we experienced this problem ourselves. At our first startup, we built a tool for product managers that included a meeting recording feature. 70% of our engineering time was taken up by just this feature! We ended up starting Recall.ai to solve this instead. Since then, over 2000 companies use us to power their recording features, e.g. Hubspot for sales call recording, Clickup for their AI note taker. Our users are engineering teams building commercial products for financial services, telehealth, incident management, sales, interviewing, and more. We also power internal tooling for large enterprises. Running this sort of infrastructure has led to unexpected technical challenges! For example, we had to debug a 1 in 36 million segfault in our audio encoder ( https://ift.tt/iWZcm7P... ), we encountered a Postgres lock-up that only occurs when you have tens of thousands of concurrent writers ( https://ift.tt/V7pQABN ), and we saved over $1M a year on AWS by optimizing the way we shuffle data around between our processes ( https://ift.tt/rjna9Gi ). You can try it here: https://www.recall.ai . It's self-serve with $5 of free credits. Pricing starts at $0.70 for every hour of recording, prorated to the second. We offer volume discounts with scale. All data recorded through Recall.ai is the property of our customers, we support 0-day retention, and we don’t train models on customer data. We would love your feedback!
Tuesday, September 9, 2025
Monday, September 8, 2025
Sunday, September 7, 2025
Saturday, September 6, 2025
Friday, September 5, 2025
Thursday, September 4, 2025
New top story on Hacker News: A high schooler writes about AI tools in the classroom
A high schooler writes about AI tools in the classroom
88 by dougb5 | 74 comments on Hacker News.
https://ift.tt/klip70W
88 by dougb5 | 74 comments on Hacker News.
https://ift.tt/klip70W
Wednesday, September 3, 2025
New top story on Hacker News: Show HN: Entropy-Guided Loop – How to make small models reason
Show HN: Entropy-Guided Loop – How to make small models reason
11 by andrewmonostate | 0 comments on Hacker News.
TLDR: A small, vendor-agnostic inference loop that turns token logprobs/perplexity/entropy into an extra pass and reasoning for LLMs. - Captures logprobs/top-k during generation, computes perplexity and token-level entropy. - Triggers at most one refine when simple thresholds fire; passes a compact “uncertainty report” (uncertain tokens + top-k alts + local context) back to the model. - In our tests on technical Q&A / math / code, a small model recovered much of “reasoning” quality at ~⅓ the cost while refining ~⅓ of outputs. I kept seeing “reasoning” models behave like expensive black boxes. Meanwhile, standard inference already computes useful signals both before softmax normalization and after it(logprobs), which we usually throw away. This loop tries the simplest thing that you could think of: use those signals to decide when (and where) to think again. GitHub (notebook + minimal code): https://ift.tt/v2pLKoZ Paper (short & engineer made): https://ift.tt/ru7BVew Blog (more context): https://ift.tt/iPx1kMp Requirements: Python, API that exposes logprobs (tested with OpenAI non reasoning 4.1). OPENAI_API_KEY and WEAVE for observability. Run the notebook; it prints metrics and shows which tokens triggered refinement. - Python, simple loop (no retraining). - Uses Responses API logprobs/top-k; metrics: perplexity, max token entropy, low-confidence counts. - Weave for lightweight logging/observability (optional). - Passing alternatives (not just “this looks uncertain”) prevents over-correction. - A simple OR rule (ppl / max-entropy / low-confidence count) catches complementary failure modes. - Numbers drift across vendors; keeping the method vendor-agnostic is better than chasing fragile pairings. - Needs APIs that expose logprobs/top-k. - Results are indicative—not a leaderboard; focus is on within-model gains (single-pass vs +loop). - Thresholds might need light tuning per domain. - One pass only; not a chain-of-thought replacement. - Run it on your models and ideas (e.g., 4o-mini, v3, Llama variants with logprobs) and share logs in a PR for our README in GitHub if you'd like, PRs welcome - I’ll credit and link. Overall let me know if you find making small models reason like this useful!
11 by andrewmonostate | 0 comments on Hacker News.
TLDR: A small, vendor-agnostic inference loop that turns token logprobs/perplexity/entropy into an extra pass and reasoning for LLMs. - Captures logprobs/top-k during generation, computes perplexity and token-level entropy. - Triggers at most one refine when simple thresholds fire; passes a compact “uncertainty report” (uncertain tokens + top-k alts + local context) back to the model. - In our tests on technical Q&A / math / code, a small model recovered much of “reasoning” quality at ~⅓ the cost while refining ~⅓ of outputs. I kept seeing “reasoning” models behave like expensive black boxes. Meanwhile, standard inference already computes useful signals both before softmax normalization and after it(logprobs), which we usually throw away. This loop tries the simplest thing that you could think of: use those signals to decide when (and where) to think again. GitHub (notebook + minimal code): https://ift.tt/v2pLKoZ Paper (short & engineer made): https://ift.tt/ru7BVew Blog (more context): https://ift.tt/iPx1kMp Requirements: Python, API that exposes logprobs (tested with OpenAI non reasoning 4.1). OPENAI_API_KEY and WEAVE for observability. Run the notebook; it prints metrics and shows which tokens triggered refinement. - Python, simple loop (no retraining). - Uses Responses API logprobs/top-k; metrics: perplexity, max token entropy, low-confidence counts. - Weave for lightweight logging/observability (optional). - Passing alternatives (not just “this looks uncertain”) prevents over-correction. - A simple OR rule (ppl / max-entropy / low-confidence count) catches complementary failure modes. - Numbers drift across vendors; keeping the method vendor-agnostic is better than chasing fragile pairings. - Needs APIs that expose logprobs/top-k. - Results are indicative—not a leaderboard; focus is on within-model gains (single-pass vs +loop). - Thresholds might need light tuning per domain. - One pass only; not a chain-of-thought replacement. - Run it on your models and ideas (e.g., 4o-mini, v3, Llama variants with logprobs) and share logs in a PR for our README in GitHub if you'd like, PRs welcome - I’ll credit and link. Overall let me know if you find making small models reason like this useful!
New top story on Hacker News: Vector search on our codebase transformed our SDLC automation
Vector search on our codebase transformed our SDLC automation
17 by antonybrahin | 2 comments on Hacker News.
Hey HN, In software development, the process of turning a user story into detailed documentation and actionable tasks is critical. However, this manual process can often be a source of inconsistency and a significant time investment. I was driven to see if I could streamline and elevate it. I know this is a hot space, with big players like GitHub and Atlassian building integrated AI, and startups offering specialized platforms. My goal wasn't to compete with them, but to see what was possible by building a custom, "glass box" solution using the best tools for each part of the job, without being locked into a single ecosystem. What makes this approach different is the flexibility and full control. Instead of a pre-packaged product, this is a resilient workflow built on Power Automate, which acts as the orchestrator for a sequence of API calls: Five calls to the Gemini API for the core generation steps (requirements, tech spec, test strategy, etc.). One call to an Azure OpenAI model to create vector embeddings of our codebase. One call to Azure AI Search to perform the Retrieval-Augmented Generation (RAG). This was the key to getting context-aware, non-generic outputs. It reads our actual code to inform the technical spec and tasks. A bunch of direct calls to the Azure DevOps REST API (using a PAT) to create the wiki pages and work items, since the standard connectors were a bit limited. The biggest challenge was moving beyond simple prompts and engineering a resilient system. Forcing the final output into a rigid JSON schema instead of parsing text was a game-changer for reliability. The result is a system that saves us hours on every story and produces remarkably consistent, high-quality documentation and tasks. The full write-up with all the challenges, final prompts, and screenshots is in the linked blog post. I’m here to answer any questions. Would love to hear your feedback and ideas!
17 by antonybrahin | 2 comments on Hacker News.
Hey HN, In software development, the process of turning a user story into detailed documentation and actionable tasks is critical. However, this manual process can often be a source of inconsistency and a significant time investment. I was driven to see if I could streamline and elevate it. I know this is a hot space, with big players like GitHub and Atlassian building integrated AI, and startups offering specialized platforms. My goal wasn't to compete with them, but to see what was possible by building a custom, "glass box" solution using the best tools for each part of the job, without being locked into a single ecosystem. What makes this approach different is the flexibility and full control. Instead of a pre-packaged product, this is a resilient workflow built on Power Automate, which acts as the orchestrator for a sequence of API calls: Five calls to the Gemini API for the core generation steps (requirements, tech spec, test strategy, etc.). One call to an Azure OpenAI model to create vector embeddings of our codebase. One call to Azure AI Search to perform the Retrieval-Augmented Generation (RAG). This was the key to getting context-aware, non-generic outputs. It reads our actual code to inform the technical spec and tasks. A bunch of direct calls to the Azure DevOps REST API (using a PAT) to create the wiki pages and work items, since the standard connectors were a bit limited. The biggest challenge was moving beyond simple prompts and engineering a resilient system. Forcing the final output into a rigid JSON schema instead of parsing text was a game-changer for reliability. The result is a system that saves us hours on every story and produces remarkably consistent, high-quality documentation and tasks. The full write-up with all the challenges, final prompts, and screenshots is in the linked blog post. I’m here to answer any questions. Would love to hear your feedback and ideas!
Tuesday, September 2, 2025
Monday, September 1, 2025
Subscribe to:
Comments (Atom)