Ultimate Guide to Optimizing Stable Diffusion XL

Ultimate Guide to Optimizing Stable Diffusion XL

How to achieve the best quality and performance of SDXL on your own graphics card, and how to choose the appropriate optimization methods and tools, has been a confusing question for GenAI users, as there has been no clear and detailed evaluation report available in the industry for reference.Until full-stack developer Félix San stepped in. … Read more

Build Your Large Model Assistant Without GPU Using OpenRouter

Build Your Large Model Assistant Without GPU Using OpenRouter

1. Build Your Own Large Model Assistant The emergence of large models has brought revolutionary changes to many fields, from natural language processing to computer vision, and even in medicine and finance. However, for many developers, experimenting with and applying some open-source models can be a challenge, as they often require expensive hardware resources to … Read more

Collection of PyTorch Tricks

Collection of PyTorch Tricks

↑ ClickBlue Text Follow the Jishi platform Author丨z.defying@Zhihu (Authorized) Source丨https://zhuanlan.zhihu.com/p/76459295 Editor丨Jishi Platform Jishi Guide This article organizes 13 tips for using PyTorch, including specifying GPU IDs, gradient clipping, and expanding the dimensions of a single image, which can help workers complete tasks more efficiently. >>CV Developer 2021 New Year Exclusive Red Packet Cover has been … Read more

Lightning Attention-2: Next-Gen Attention Mechanism for Long Sequences

Lightning Attention-2: Next-Gen Attention Mechanism for Long Sequences

Machine Heart Column Machine Heart Editorial Team Lightning Attention-2 is a new type of linear attention mechanism that aligns the training and inference costs of long sequences with those of a 1K sequence length. The limitation on sequence length in large language models significantly restricts their applications in artificial intelligence, such as multi-turn dialogue, long … Read more

Speeding Up PyTorch by Four Times: Enhancing DALI Utilization and Creating CPU-Based Pipelines

Speeding Up PyTorch by Four Times: Enhancing DALI Utilization and Creating CPU-Based Pipelines

Big Data Digest Production Source:Medium Compiled by:Zhao Jike In recent years, there have been significant advancements in deep learning hardware, with Nvidia’s latest products, the Tesla V100 and Geforce RTX series, featuring dedicated tensor cores designed to accelerate common operations in neural networks. Notably, the V100 has sufficient capability to train neural networks at thousands … Read more

Building and Training Deep Learning Models with PyTorch

Building and Training Deep Learning Models with PyTorch

PyTorch occupies an important position in the field of deep learning. In real life, it is widely used in various areas such as image recognition and natural language processing. For example, in medical image diagnosis, models built with PyTorch can quickly and accurately identify lesions in X-ray and CT images; in intelligent customer service systems, … Read more

17 Methods To Speed Up PyTorch Training

17 Methods To Speed Up PyTorch Training

Selected from efficientdl.com Author: LORENZ KUHN Translated by: Machine Heart Editor: Chen Ping Master these 17 methods to accelerate your PyTorch deep learning training with minimal effort. Recently, a post on Reddit gained immense popularity. The topic was about how to speed up PyTorch training. The original author is LORENZ KUHN, a master’s student in … Read more

Exploring NVIDIA Blackwell GPU Features Beyond Neural Rendering

Exploring NVIDIA Blackwell GPU Features Beyond Neural Rendering

During CES 2025, NVIDIA unveiled the GPU based on the Blackwell architecture and showcased the performance and features of NVIDIA RTX AI technology at its Editor’s Day event. Subsequently, NVIDIA held a further communication sharing session in Shenzhen, detailing the Blackwell architecture GPU and its functionalities. So, what other aspects are worth our in-depth exploration? … Read more

Colossal-AI: Reducing AIGC Training Costs Significantly

Colossal-AI: Reducing AIGC Training Costs Significantly

Machine Heart released Machine Heart Editorial Team How to better, faster, and cheaper achieve training and fine-tuning of AIGC models has become the biggest pain point for the commercialization and explosive application of AIGC. Colossal-AI, based on its professional technical accumulation in democratizing large models,open-sourced a complete Stable Diffusion pre-training and personalization fine-tuning solution, accelerating … Read more