Oriole Networks Raises $22M for Photonics to Cut LLM Energy Use

London, 21st October: Oriole Networks – a company using light to train Large Language Models with low energy consumption – has raised an additional $22 million from investors to scale its “super-brain” solution.  The round was led by Plural with all existing investors – UCL Technology Fund, XTX Ventures, Clean Growth Fund, and Dorilton Ventures – reinvesting. Oriole Networks addresses […]

HPC News Bytes 20241014: AMD Rollout, Foxconn’s Massive AI HPC, AI Drives Nobels, Are LLM’s Intelligent?

A good mid-October morn to you! Here’s a brief (6:30) run-through of developments from the world of HPC-AI, including: AMD’s products rollout, Foxconn’s big Blackwell AI HPC in Taiwan, AI for science drives Nobel Prizes, Meta AI guru’s AGI skepticism

Cerebras Claims Fastest AI Inference

AI compute company Cerebras Systems today announced what it said is the fastest AI inference solution. Cerebras Inference delivers 1,800 tokens per second for Llama3.1 8B and 450 tokens per second for Llama3.1 70B, according to the company, making it 20 times faster than GPU-based solutions in hyperscale clouds.

NVIDIA and Google DeepMind Collaborate on LLMs

Intended to make it easier for developers to create AI-powered applications with world-class performance, NVIDIA and Google today announced three new collaborations at Google I/O ’24. Using TensorRT-LLM, NVIDIA is working with Google to optimize two new models it introduced at the event: Gemma 2 and PaliGemma. These models are built from the same research and […]

Amazon Adds $2.75B to Stake in GenAI Startup Anthropic

Amazon announced it has made its biggest-ever investment, $2.75 billion, in OpenAI/Chat-GPT competitor Anthropic, another indication that the generative AI phenomenon continues to heat up. Today’s news follows Amazon and Anthropic announcing an earlier $1.25 billion investment last September – the announcement today brings the total investment to $4 billion. “We have a notable history with […]

Oriole Networks Raises £10m for Faster LLM Training

London, 27 March 2024: Oriole Networks – a startup using light to train LLMs faster with less power – has raised £10 million in seed funding to improve AI performance and adoption, and solve AI’s energy problem. The round, which the company said is one of the UK’s largest seed raises in recent years, was co-led […]

Accelerated HPC for Energy Efficiency with AWS and NVIDIA

Many industries are starting to run HPC in the cloud. Find out how GPU-accelerated compute, from AWS and NVIDIA, is helping organizations run HPC workloads and AI/ML jobs faster, in a more energy-efficient way.

Datasaur Launches LLM Lab for ChatGPT and Similar Models

Oct. 27, 2023 — Datasaur, a natural language processing (NLP) data-labeling platform, today launched LLM Lab, an interface designed for data scientists and engineers to build and train custom LLM models like ChatGPT. The product will provide a wide range of features for users to test different foundation models, connect to their own internal documents, […]

Dell and Meta in GenAI Pact with Llama 2 LLMs

Dell Technologies (NYSE: DELL) is collaborating with Meta in a partnership designed to make it easy to deploy Meta’s Llama 2 large language models on premises with Dell’s generative AI portfolio of IT infrastructure and client devices. Dell said the collaboration simplifies the on-prem AI environment by combining Dell’s infrastructure portfolio and Llama 2 AI models….

Federated GPU Infrastructure for AI Workflows

[Sponsored Guest Article] With the explosion of use cases such as Generative AI and ML Ops driving tremendous demand for the most advanced GPUs and accelerated computing platforms, there’s never been a better time to explore the “as-a-service” model to help get started quickly.  What could take months of shipping delays and massive CapEx investments can be yours on demand….