You could build an entire data center around a new GPU with elaborate scale-up networking, exotic chiplet architectures, and advance liquid cooling. Or if you’re AMD, you could release a […]
The post AMD Unveils PCIe GPU Card for AI Inference appeared first on AIwire.
The coalition's standardization efforts could accelerate AI infrastructure development, reduce costs, and prevent vendor lock-in, fostering innovation.
The post 3M joins AMD, Meta, Oracle, Cisco, Arista to standardize expanded beam optical connectivity for AI data centers appeared first on Crypto Briefing.
In the technology world, we often talk about “inflection points” as if they are common occurrences. They aren’t. Most companies spend decades searching for one, only to miss the turn when it finally arrives. But today, AMD didn’t just meet […]
The post The AMD Inflection and How Execution and AI Strategy Are Redefining the Semiconductor Hierarchy appeared first on TechSpective.
Lemonade, created by AMD, is a server application plus GUI for running local AI models, similar to projects like LM Studio (or, more distantly, ComfyUI). What it lacks in configurability, it tries to make up for in broader integration with third-party apps that use standard APIs, and with support for non-NVIDIA runtimes.
Lemonade works with a variety of runtimes and back-end engines. It supports AMD GPUs, Ryzen NPUs, Vulkan, and CPU execution (although not for all tasks), along with the llamacpp, whispercpp, sd-cpp, kokoro, ryzenai-llm and flm back ends. In addition to providing its own set of APIs, Lemonade interoperates with a broad set of industry standards including OpenAI, Ollama, Anthropic, and llama.cpp. Both GGUF and ONNX models are supported.
The biggest omission is NVIDIA-specific GPU support. Only Vulkan (generic GPU) and AMD (ROCm) GPUs are supported. So, if you plan on using StableDiffusion models with NVIDIA hardware, you may want to look elsewhere for now. (StableDiffusi
REDWOOD CITY, Calif., May 12, 2026 — MinIO today announced MemKV, a context memory store that delivers microsecond context retrieval at petabyte scale for agentic AI inference workloads. MemKV joins […]
The post MinIO Announces MemKV, Purpose-Built Context Memory Store for AI Inference appeared first on AIwire.
The post NVIDIA Launches Fleet Intelligence for GPU Monitoring appeared on BitcoinEthereumNews.com.
Felix Pinkston
May 11, 2026 20:27
NVIDIA’s new Fleet Intelligence service offers real-time GPU fleet monitoring, improving efficiency and reliability for data centers.
NVIDIA has announced the general availability of Fleet Intelligence, a managed service aimed at providing real-time monitoring for GPU fleets. Designed for data center operators and enterprises scaling NVIDIA GPUs, this service tackles the complexities of managing heterogeneous hardware, fast-evolving software stacks, and variable workloads. The goal is clear: optimize performance, reduce downtime, and maximize return on investment (ROI). Fleet Intelligence employs a lightweight, host-based agent to stream telemetry data to a cloud-based platform. This enables precise insights into key operational metrics, including power consumption, temperature, performance, health, and configuration consistency. NVIDIA
The post IREN Nvidia deal worth $3.4B over five years appeared on BitcoinEthereumNews.com.
IREN has signed a $3.4 billion IREN Nvidia deal to deploy up to 5 gigawatts of AI infrastructure over five years. Summary IREN will provide Nvidia with managed GPU cloud services worth $3.4 billion over five years for the chipmaker’s internal AI and research workloads. Nvidia received a five-year warrant to purchase up to 30 million IREN shares at $70 each, representing a potential $2.1 billion equity stake. The partnership builds on IREN’s prior $9.7 billion agreement with Microsoft, pushing the company’s total committed revenue past $15 billion. Bitcoin miner turned AI infrastructure provider IREN has announced a five-year, $3.4 billion AI cloud contract with Nvidia, alongside a broader strategic partnership to build out 5 gigawatts of next-generation infrastructure. The deal was disclosed alongside IREN’s third-quarter FY2026 earnings on May 7. IREN will provide Nvidia with managed GPU cloud s
The post Teradyne sees growth potential from AI networking and GPU expansion appeared on BitcoinEthereumNews.com.
Teradyne reported Q1 2026 revenue of $1.282 billion, an 87% increase year-over-year, driven almost entirely by the insatiable appetite for AI-related chip testing. Non-GAAP earnings per share came in at $2.56, and the company’s free cash flow is projected to reach $950 million for the full year. The AI testing gold rush The automated test equipment market stood at $9 billion in 2025 and is forecasted to grow significantly from there, fueled by AI-driven demands for compute, networking, and memory chip testing. The company secured multiple production test orders for merchant GPUs during Q1 2026. Merchant GPUs, the kind sold on the open market rather than kept in-house by hyperscalers, represent a massive addressable market that Teradyne has been working to crack for years. Full-year outlook and what the numbers say Analysts are projecting Teradyne’s full-year 2026 revenue gr
Teradyne's growth in AI chip testing could reshape tech markets, impacting GPU availability and potentially squeezing crypto mining operations.
The post Teradyne sees growth potential from AI networking and GPU expansion appeared first on Crypto Briefing.