Articles for category: AI Tools

Challenges & Solutions For Monitoring at Hyperscale

“What is not measured, cannot be improved.” This quote has become a guiding principle for teams training foundation models. When you’re dealing with complex, large-scale AI systems, things can spiral quickly without the right oversight. Operating at hyperscale poses significant challenges for teams, from the large volume of data generated to the unpredictability of hardware failures and the need for efficient resource management. These issues require strategic solutions, that’s why monitoring isn’t just a nice-to-have—it’s the backbone of transparency, reproducibility, and efficiency. During my talk at NeurIPS,  I broke down five key lessons learned from teams facing large-scale model training

What’s New in AI/BI – Feb ‘25 Roundup

Introduction AI/BI Dashboards and Genie are evolving at a breakneck pace. In this roundup, we’ll highlight the most impactful updates from the past three months that make AI/BI more powerful, easier to use, and smarter than ever. For those unfamiliar, AI/BI is a suite of Business Intelligence (BI) capabilities that are included with the Databricks SQL product. Now is the perfect time to start if you haven’t explored it yet. With Databricks AI/BI, you can quickly and easily unlock and share insights from your data—without the need for a separate BI system. Let’s take a closer look at the latest

Prozessvisualisierung mit generativer KI im Praxistest

Prozessvisualisierung mit generativer KI im Praxistest Auswahl des passenden Modells Auf Sprachmodelle zugreifen Bildprompts nutzen Mit Feedback arbeiten Weitere Beispiele Fazit Artikel in iX 3/2025 lesen Wer sich auf die Suche nach intelligenten Helferlein für den Büro- oder auch Privatalltag begibt, wird bereits bei OpenAI fündig: Die GPTs, spezialisierte, von ChatGPT abgeleitete Chatbots, stellen allerlei nützliche Handreichungen in Aussicht – von der Wikipedia-Recherche bis zur Wireframe-Gestaltung scheint vieles abgedeckt. Beim Blick auf das tatsächliche Vorgehen werden jedoch Schwächen deutlich: Bei Aktivierung eines GPT übernimmt ChatGPT eine andere Persona und verwendet bestimmte Tools und Hintergrundinformationen, die vom Ersteller des GPT konfiguriert

AI video is having its Stable Diffusion moment

AI video used to not be very good: Will Smith eating spaghetti, u/chaindrop, March 2023 Then, 10 months later, OpenAI announced Sora: Creating video from text, OpenAI, February 2024 Sora reset expectations about what a video model could be. The output was high resolution, smooth, and coherent. The examples looked like real video. It felt like we’d jumped into the future. The problem was, nobody could use it! It was just a preview. This was like when OpenAI announced the DALL-E image generation model back in 2021. It was one of the most extraordinary pieces of software that had been

Optimizing AI Models with Quanto

The transformer-based diffusion models are improving day by day and have proven to revolutionize the text-to-image generation model. The capabilities of transformers enhance the scalability and performance of any model, thereby increasing the model’s complexity. “With great power comes great responsibility” In this case, with great model complexities comes great power and memory consumption. For instance, running inference with models like Stable Diffusion 3 requires a huge GPU memory, due to the involvement of components—text encoders, diffusion backbones, and image decoders. This high memory requirement causes set back for those using consumer-grade GPUs, which hampers both accessibility and experimentation. Enter

Quickly Start Evaluating LLMs With OpenEvals

Evaluations (evals) are important for bringing reliable LLM powered applications or agents to production, but it can be hard to know where to start when building evaluations from scratch. Our new packages—openevals and agentevals—provide a set of evaluators and a common framework that you can easily get started with. What are evals? Evals provide systematic ways to judge LLM output quality based on criteria that’s important for your application. There are two components of evals: the data that you’re evaluating over and the metric that you’re evaluating on. The quality and diversity of the data you’re evaluating over directly influences

5 New Thinking Styles for Working With Thinking Machines

It’s the last day of Every’s think week—our quarterly time to dream up new ideas and products that can help us improve how we do our work and, more importantly, your experience as a member of our community. In lieu of publishing new stories, we’ve been re-upping pieces by Dan Shipper (who’s been on hiatus from writing his regular Chain of Thought column to work on a longer piece) that cover basic, powerful questions about AI. Last up is his piece about how humans should think in a world with thinking machines. We’ll be back with a new piece on

TSMC N2 + Next-Gen SoIC, Intel EMIB-T, Meta 3D Stacked Memory, CFET, 2D Materials, and More – SemiAnalysis

The semiconductor industry isn’t built on overnight breakthroughs. It’s built on large leaps of progress, compounding year after year, at a higher rate than probably any other industry in history. The International Electron Device Manufacturing conference IEDM is one of the key venues where chipmakers can show off this progress. Paper topics range from commercially relevant, those that could eventually be, and others that probably won’t be but are interesting technology anyways. Semiconductors: incremental gains compounded over 50+ years. Source: AMD For logic: TSMC’s N2 process, 2D materials including Samsung and others, CFET advancements, and Intel scaling silicon channels beyond

llm-ollama 0.9.0

llm-ollama 0.9.0. This release of the llm-ollama plugin adds support for schemas, thanks to a PR by Adam Compton. Ollama provides very robust support for this pattern thanks to their structured outputs feature, which works across all of the models that they support by intercepting the logic that outputs the next token and restricting it to only tokens that would be valid in the context of the provided schema. With Ollama and llm-ollama installed you can run even run structured schemas against vision prompts for local models. Here’s one against Ollama’s llama3.2-vision: llm -m llama3.2-vision:latest \ 'describe images' \ --schema

Why partitioned tables are powerful

“You don’t have to be an engineer to be a racing driver, but you do have to have Mechanical Sympathy.” – Jackie Stewart, racing driver This simple quote has deep meaning when it comes to many facets of life. Its value cannot be understated when it comes to software. In this article, we’ll touch upon an important concept in Deephaven – partitioned tables. Like with auto racing, having a grasp of the mechanics behind a system will enable you to maximize its potential. So, let’s take a closer look at partitioned tables and how they can help you get the