Results for: "llm"
Keyword Search 9 resultsTalu: A Single-Binary, Local-First LLM Runtime
THE GIST: Talu is a local-first inference engine for LLMs, packaged as a single binary with no heavy runtime dependencies.
Karpathy's Micro LLM: A Minimal GPT in JavaScript
THE GIST: Karpathy's Micro LLM is a minimal GPT-style language model in pure JavaScript for character-level next-token prediction.
AI-BOM: Scan Your Codebase for AI Agents, Models, and API Keys
THE GIST: AI-BOM is a tool designed to scan codebases for AI agents, models, and API keys, creating an AI Bill of Materials for security and compliance.
OPUS: Efficient Data Selection for LLM Pre-Training
THE GIST: OPUS is a new framework for efficient LLM pre-training that dynamically selects data based on optimizer-induced updates.
LocalMind Enables Privacy-First, In-Browser AI Chat with WebGPU
THE GIST: LocalMind offers privacy-focused AI chat directly in the browser, utilizing WebGPU for accelerated inference and eliminating server-side processing.
AI-Powered Swindles: A Growing Cybersecurity Threat
THE GIST: AI is lowering the barrier for cyberattacks, enabling faster, more personalized, and harder-to-detect swindles, though fully automated attacks remain unlikely.
Is the AI Bubble About to Burst? Echoes of the Dot-Com Crash
THE GIST: The current AI boom mirrors the dot-com bubble, with unsustainable valuations and heavy advertising spending signaling a potential crash.
Cache-Aware Prefill-Decode Disaggregation Boosts LLM Serving Speed by 40%
THE GIST: Together AI's cache-aware prefill-decode disaggregation (CPD) architecture improves long-context LLM serving by up to 40% by separating cold and warm workloads.
Cisco Open Sources AI Bill of Materials Tool
THE GIST: Cisco releases an open-source tool to scan codebases and container images, creating an AI Bill of Materials (AI BOM).