Enterprise

AWS brings prompt routing and caching to its Bedrock LLM service

As businesses move from trying out generative AI in limited prototypes to putting them into production, they are becoming increasingly price conscious. Using large language models isn’t cheap, after all. One way to reduce cost is to go back to an old concept: caching. Another is to route simpler queries to smaller, more cost-efficient models. […]

AWS brings prompt routing and caching to its Bedrock LLM service Read More »

AWS makes its SageMaker HyperPod AI platform more efficient for training LLMs

At last year’s AWS re:Invent conference, Amazon’s cloud computing unit launched SageMaker HyperPod, a platform for building foundation models. It’s no surprise, then, that at this year’s re:Invent, the company is announcing a number of updates to the platform, with a focus on making model training and fine-tuning on HyperPod more efficient and cost-effective for

AWS makes its SageMaker HyperPod AI platform more efficient for training LLMs Read More »

Linux Foundation report highlights the true state of open-source libraries in production apps

There are many metrics to track the prevalence of open-source components, such as GitHub stars and downloads, but they don’t paint the full picture of how they’re being used in production codebases. Census III of Free and Open Source Software: Application Libraries leans on more than 12 million data points from software composition analysis (SCA)

Linux Foundation report highlights the true state of open-source libraries in production apps Read More »

Amazon teams up with Orbital to remove CO2 from the air at one of its datacenters 

AI’s surging power demand has put several big tech firms at risk of blowing through their climate commitments. But Amazon has partnered with Orbital, an AI startup, to test a new material that removes carbon dioxide from the atmosphere — and they’re using an AWS datacenter as a first site.  One of carbon capture’s biggest

Amazon teams up with Orbital to remove CO2 from the air at one of its datacenters  Read More »

AWS wants Amazon Q to become your buddy for the entire software development lifecycle

At its re:Invent conference, AWS today announced a series of updates to Q Developer, its coding assistant platform that competes with the likes of GitHub Copilot. The focus here is on going beyond code completion and to help developers with a wider range of routine tasks involved in the end-to-end software lifecycle. The service, which

AWS wants Amazon Q to become your buddy for the entire software development lifecycle Read More »

AWS and GitLab team up to bring Amazon Q agents to GitLab’s Duo Assistant

GitLab, the popular developer and security platform, and AWS, the popular cloud computing and AI service, today announced that they have teamed up to combine GitLab’s Duo AI assistant with Amazon’s Q autonomous agents. The goal here, the two companies say, is to accelerate software innovation and developer productivity, and unlike so many partnerships in

AWS and GitLab team up to bring Amazon Q agents to GitLab’s Duo Assistant Read More »

AWS announces Aurora DSQL, a new distributed SQL database that promises virtually unlimited scalability

At its re:Invent conference, Amazon’s AWS cloud computing unit today announced Amazon Aurora DSQL, a new serverless, distributed SQL database that promises high availability (99.999% for multi-region availability), strong consistency, PostgreSQL compatibility, and, the company says, “4x faster reads and writes compared to other popular distributed SQL databases.” AWS argues that Aurora DSQL will offer

AWS announces Aurora DSQL, a new distributed SQL database that promises virtually unlimited scalability Read More »

AWS’ Trainium2 chips for building LLMs are now generally available, with Trainium3 coming in late 2025

At its re:Invent conference, AWS today announced the general availably of its Trainium2 (T2) chips for training and deploying large language models (LLMs). These chips, which AWS first announced a year ago, will be four times as fast as their predecessors, with a single Trainium2-powered EC2 instance with 16 T2 chips providing up to 20.8

AWS’ Trainium2 chips for building LLMs are now generally available, with Trainium3 coming in late 2025 Read More »

Pat Gelsinger set to net more than $10M severance from Intel

Intel’s newly departed ex-CEO Pat Gelsinger could be walking away with more than $10 million in severance pay. As per a filing with the Securities and Exchange Commission (SEC) today, Intel and Gelsinger entered into a “retirement and separation agreement” which will entitle the former CEO to a payment equal to 18 months of his

Pat Gelsinger set to net more than $10M severance from Intel Read More »