Articles for category: AI News

Beyond the Cloud: Exploring the Benefits and Challenges of On-Premises AI Deployment

When you mention AI, both to a layman and an AI engineer, the cloud is probably the first thing that comes to mind. But why, exactly? For the most part, it’s because Google, OpenAI and Anthropic lead the charge, but they don’t open-source their models nor do they offer local options.  Of course, they do have enterprise solutions, but think about it—do you really want to trust third parties with your data? If not, on-premises AI is by far the best solution, and what we’re tackling today. So, let’s tackle the nitty gritty of combining the efficiency of automation with

How DeepMind’s AlphaGeometry2 Achieved Gold-Medalist Status in the International Math Olympiad

Created using Midjourney DeepMind’s journey toward mathematical AI dominance took a major leap last year when AlphaProof and AlphaGeometry nearly clinched gold at the International Math Olympiad (IMO). Now, with the latest upgrade, AlphaGeometry2 (AG2) has officially surpassed top human competitors in geometry, marking a milestone in AI-driven mathematical reasoning. The general consensus among IMO competitors is that geometry problems are among the toughest in each day of the Olympiad. AG2 represents a significant advancement in AI-driven mathematical reasoning, particularly in solving Olympiad geometry problems. Building on its predecessor, AlphaGeometry, AG2 surpasses the performance of an average gold medalist in

Mistral cracks AI document analysis

PLUS: China’s Manus demos ‘world’s first fully autonomous’ AI agent Good morning, AI enthusiasts. French startup Mistral just turned AI document processing on its head — with a new model that makes complex data extraction as simple as an API call. With speeds of up to 2000 pages per minute and the ability to handle multilingual texts, images, charts, and more, is this the tech that finally converts static archives into the AI-powered gold mines of tomorrow? P.S. Our next workshop is today at 3:30 PM EST! Join to learn how to use the latest AI tools to take your

The Human Toll of Waiting for AI to Take Over

Hanna Barakat + AIxDESIGN & Archival Images of AI / Better Images of AI / Data Mining 1 / CC-BY 4.0 So much time thinking about AI has changed how I see people. I notice the workers I used to overlook, doing the jobs no one wants to think about. Living in Madrid, that doesn’t mean miners or fishermen, but warehouse staff, delivery drivers, supermarket stockers, and cleaners. Like the worn landscapes of a city you’ve walked a thousand times, they fade into the background, part of the scenery. They blend in like mundane noise we drown out with premium

LWiAI Podcast #201 – GPT 4.5, Sonnet 3.7, Grok 3, Phi 4

Our 201st episode with a summary and discussion of last week’s big AI news!Recorded on 03/02/2025 Join our brand new Discord here! https://discord.gg/nTyezGSKwP Hosted by Andrey Kurenkov and guest host Sharon ZhouFeel free to email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai In this episode: – The release of GPT-4.5 from OpenAI, Anthropic’s Claude 3.7, and Grok 3 from XAI, comparing their features, costs, and capabilities. – Discussion on new tools and applications including Sesame’s new voice assistant and Google’s AI coding assistant, Gemini Code Assist, highlighting their unique benefits. – OpenAI’s continued user growth despite competition, pricing

Advance Trustworthy AI and ML, and Identify Best Practices for Scaling AI 

By John P. Desmond, AI Trends Editor   Advancing trustworthy AI and machine learning to mitigate agency risk is a priority for the US Department of Energy (DOE), and identifying best practices for implementing AI at scale is a priority for the US General Services Administration (GSA).   That’s what attendees learned in two sessions at the AI World Government live and virtual event held in Alexandria, Va. last week.    Pamela Isom, Director of the AI and Technology Office, DOE Pamela Isom, Director of the AI and Technology Office at the DOE, who spoke on Advancing Trustworthy AI and ML Techniques for Mitigating Agency

Microsoft readies in-house AI models to rival OpenAI and Anthropic, plans API access in 2025

Summary Microsoft is reportedly developing its own family of language models under “CEO of AI” Mustafa Suleyman, aiming to match capabilities offered by OpenAI and Anthropic. The company is said to be planning to release these models through an API later this year. The Information reports that Microsoft’s AI team reached a key milestone with their internal model family, known as MAI. Testing indicates the models perform at nearly the same level as leading offerings from OpenAI and Anthropic on standard benchmarks. The family includes a reasoning model specifically designed to match OpenAI’s o1 capabilities. These new models are much

Llama 3 Meets MoE: Pioneering Low-Cost High-Performance AI

The transformative impact of Transformers on natural language processing (NLP) and computer vision (CV) is undeniable. Their scalability and effectiveness have propelled advancements across these fields, but the rising complexity of these models has led to soaring computational costs. Addressing this challenge has become a priority, prompting exploration into alternative approaches like Mixture-of-Experts (MoE) architectures, which aim to boost model capacity without proportional increases in computation. However, training MoE models from scratch is fraught with difficulties, including overfitting and instability in routing mechanisms. To tackle these issues, researchers from the University of Texas at Austin and NVIDIA have introduced a

New method significantly reduces AI energy consumption

The SuperMUC-NG at the Leibniz Supercomputing Centre is the eighth fastest computer in the world. Credit: Veronika Hohenegger, LRZ AI applications such as large language models (LLMs) have become an integral part of our everyday lives. The required computing, storage and transmission capacities are provided by data centers that consume vast amounts of energy. In Germany alone, this amounted to about 16 billion kWh in 2020, or around 1% of the country’s total energy consumption. For 2025, this figure is expected to increase to 22 billion kWh. The arrival of more complex AI applications in the coming years will substantially

Andrew Barto and Richard Sutton win 2024 Turing Award

Andrew Barto and Richard Sutton. Image credit: Association for Computing Machinery. The Association for Computing Machinery, has named Andrew Barto and Richard Sutton as the recipients of the 2024 ACM A.M. Turing Award. The pair have received the honour for “developing the conceptual and algorithmic foundations of reinforcement learning”. In a series of papers beginning in the 1980s, Barto and Sutton introduced the main ideas, constructed the mathematical foundations, and developed important algorithms for reinforcement learning. The Turing Award comes with a $1 million prize, to be split between the recipients. Since its inception in 1966, the award has honoured