Articles for category: AI Tools

Fine-tune FLUX.1 with your own images

FLUX.1 is a family of text-to-image models released by Black Forest Labs this summer. The FLUX.1 models set a new standard for open-source image models: they can generate realistic hands, legible text, and even the strangely hard task of funny memes. You can now fine-tune FLUX.1 [dev] with Ostris’s AI Toolkit on Replicate. Teach the model to recognize and generate new concepts by showing it a small set of example images, allowing you to customize the model’s output for specific styles, characters, or objects. Ostris’s toolkit uses the LoRA technique for fast, lightweight trainings. People have already made some amazing

A Fun and Easy Guide to run LLMs via React Native on your Phone!

As LLMs continue to evolve, they are becoming smaller and smarter, enabling them to run directly on your phone. Take, for instance, the DeepSeek R1 Distil Qwen 2.5 with 1.5 billion parameters, this model really shows how advanced AI can now fit into the palm of your hand! In this blog, we will guide you through creating a mobile app that allows you to chat with these powerful models locally. The complete code for this tutorial is available in our EdgeLLM repository. If you’ve ever felt overwhelmed by the complexity of open-source projects, fear not! Inspired by the Pocket Pal

NVIDIA Reveals Neural Rendering, AI Advancements at GDC 2025

AI is leveling up the world’s most beloved games, as the latest advancements in neural rendering, NVIDIA RTX and digital human technologies equip game developers to take innovative leaps in their work. At this year’s GDC conference, running March 17-21 in San Francisco, NVIDIA is revealing new AI tools and technologies to supercharge the next era of graphics in games. Key announcements include new neural rendering advancements with Unreal Engine 5 and Microsoft DirectX; NVIDIA DLSS 4 now available in over 100 games and apps, making it the most rapidly adopted NVIDIA game technology of all time; and a Half-Life

How GoDaddy built a category generation system at scale with batch inference for Amazon Bedrock

This post was co-written with Vishal Singh, Data Engineering Leader at Data & Analytics team of GoDaddy Generative AI solutions have the potential to transform businesses by boosting productivity and improving customer experiences, and using large language models (LLMs) in these solutions has become increasingly popular. However, inference of LLMs as single model invocations or API calls doesn’t scale well with many applications in production. With batch inference, you can run multiple inference requests asynchronously to process a large number of requests efficiently. You can also use batch inference to improve the performance of model inference on large datasets. This

Mixed-input matrix multiplication performance optimizations

AI-driven technologies are weaving themselves into the fabric of our daily routines, with the potential to enhance our access to knowledge and boost our overall productivity. The backbone of these applications lies in large language models (LLMs). LLMs are memory-intensive and typically require specialized hardware accelerators to efficiently deliver tens of exaflops of computing power. This blog post shows how we can start addressing the computational challenges by utilizing memory more effectively. The bulk of an LLM’s memory and compute are consumed by weights in matrix multiplication operations. Using narrower data types reduces memory consumption. For example, storing weights in

Growing Neural Cellular Automata

Contents This article is part of the Differentiable Self-organizing Systems Thread, an experimental format collecting invited short articles delving into differentiable self-organizing systems, interspersed with critical commentary from several experts in adjacent fields. Differentiable Self-organizing Systems Thread Self-classifying MNIST Digits Most multicellular organisms begin their life as a single egg cell – a single cell whose progeny reliably self-assemble into highly complex anatomies with many organs and tissues in precisely the same arrangement each time. The ability to build their own bodies is probably the most fundamental skill every living creature possesses. Morphogenesis (the process of an organism’s shape development)

The Illustrated Stable Diffusion – Jay Alammar – Visualizing machine learning one concept at a time.

Translations: Chinese, Vietnamese. (V2 Nov 2022: Updated images for more precise description of forward diffusion. A few more images in this version) AI image generation is the most recent AI capability blowing people’s minds (mine included). The ability to create striking visuals from text descriptions has a magical quality to it and points clearly to a shift in how humans create art. The release of Stable Diffusion is a clear milestone in this development because it made a high-performance model available to the masses (performance in terms of image quality, as well as speed and relatively low resource/memory requirements). After

Ascending Levels of Nerd – O’Reilly

In developing the content for our May 8 virtual conference Coding with AI: The End of Software Development as We Know It, we couldn’t help but want to feature Harper Reed, whose recent post “My LLM Codegen Workflow ATM” so perfectly encapsulates the kind of experimentation that developers are going through as they come to grips with the transformation that AI is bringing to how they work, what they can accomplish, and which tools they should be adopting. Harper lays out his current workflows and tools with detailed examples for both greenfield code and legacy code that make it easy

Democratizing Machine Learning with Simplicity, Power, and Innovation – The Official Blog of BigML.com

As 2024 draws to a close, it’s time to reflect on how the BigML team has been working to enhance our platform, solidifying its place as a leader in the machine learning space. In a world where artificial intelligence is reshaping industries, BigML remains a game-changer with its no nonsense approach to bringing more self-directed users on board the ML/AI boat. Since our inception in 2011, our mission has been clear: to democratize machine learning by making it easy, accessible, transparent, traceable, interpretable, scalable, and user-friendly for everyone, regardless of technical expertise. What truly sets BigML apart is its ability