Willow Ventures

Kong Releases Volcano: A TypeScript, MCP-native SDK for Building Production Ready AI Agents with LLM Reasoning and Real-World actions | Insights by Willow Ventures

Kong Releases Volcano: A TypeScript, MCP-native SDK for Building Production Ready AI Agents with LLM Reasoning and Real-World actions | Insights by Willow Ventures

Discover Kong’s Volcano SDK: A Game-Changer for AI Workflows Kong has launched the Volcano SDK, an open-source TypeScript framework designed for crafting multi-step workflows using multiple LLM (Large Language Model) providers. This innovative tool is engineered for seamless integration with the Model Context Protocol (MCP) and offers developers a streamlined approach to building AI applications. […]

StreamTensor: A PyTorch-to-Accelerator Compiler that Streams LLM Intermediates Across FPGA Dataflows | Insights by Willow Ventures

StreamTensor: A PyTorch-to-Accelerator Compiler that Streams LLM Intermediates Across FPGA Dataflows | Insights by Willow Ventures

StreamTensor: Revolutionizing LLM Inference on FPGAs In the fast-evolving landscape of machine learning, optimizing model inference is vital for enhanced performance and efficiency. StreamTensor offers an innovative approach by transforming PyTorch LLM graphs into dataflow accelerators on AMD’s Alveo U55C FPGA. What is StreamTensor? StreamTensor is a powerful compiler that translates PyTorch large language model […]

VaultGemma: The world's most capable differentially private LLM | Insights by Willow Ventures

VaultGemma: The world's most capable differentially private LLM | Insights by Willow Ventures

Understanding Generative AI: Revolutionizing Creativity and Innovation Generative AI is transforming the way we create and interact with technology. By leveraging machine learning algorithms, it generates new content, from text and images to music and videos. What is Generative AI? Generative AI refers to algorithms that can create new data based on existing datasets. Unlike […]

BentoML Released llm-optimizer: An Open-Source AI Tool for Benchmarking and Optimizing LLM Inference | Insights by Willow Ventures

BentoML Released llm-optimizer: An Open-Source AI Tool for Benchmarking and Optimizing LLM Inference | Insights by Willow Ventures

Streamline LLM Performance with BentoML’s New llm-optimizer BentoML has introduced llm-optimizer, an innovative open-source framework aimed at optimizing the benchmarking and performance tuning of self-hosted large language models (LLMs). This tool addresses the complexities associated with LLM deployment, making it easier to achieve the best configurations for latency, throughput, and cost. Why is Tuning LLM […]

Speculative cascades — A hybrid approach for smarter, faster LLM inference | Insights by Willow Ventures

Speculative cascades — A hybrid approach for smarter, faster LLM inference | Insights by Willow Ventures

Understanding Speculative Cascades in AI Model Responses In the ever-evolving world of AI, understanding how models generate responses can enhance their effectiveness. This blog delves into the speculative cascades approach, comparing different AI models’ capabilities in answering questions. Comparing Response Styles of AI Models When posed with a simple question like, “Who is Buzz Aldrin?”, […]