Saturday, January 3, 2026

Keep Your Ubuntu-based VPN Server Up to Date

Enterprise-Grade Security for Small Businesses with Linux and Open Source

Ethics for Ephemeral Signals – A Manifesto

When Regex Falls Short – Auditing Discord Bots with AI Reasoning Models

Cisco Live 2025: Bridging the Gap in the Digital Workplace to Achieve ‘Distance Zero’

Agentforce London: Salesforce Reports 78% of UK Companies Embrace Agentic AI

WhatsApp Aims to Collaborate with Apple on Legal Challenge Against Home Office Encryption Directives

AI and the Creative Industries: A Misguided Decision by the UK Government

CityFibre Expands Business Ethernet Access Threefold

Nvidia discusses the effects of AI workloads on data storage in an interview

AI workloads are unlike those seen in the enterprise before, with varying impacts on storage throughout different AI phases. After intense training, AI is put to work inferencing based on what it has learned, taking into account factors such as AI frameworks, storage demands of retrieval-augmented generation (RAG), and checkpointing.

Nvidia’s vice-president and general manager of DGX Systems, Charlie Boyle, discussed these challenges and practical tips for customers embarking on AI projects at the recent Pure Storage Accelerate event in Las Vegas. Understanding good and bad data is crucial for AI success, with the need to differentiate between data that adds value and outdated information.

To start with AI, Boyle recommends beginning with existing models that can be fine-tuned for specific needs, rather than building foundational models from scratch. For customers looking to put AI to work, leveraging ready-made applications and data to customize them for their own needs is key, with no need for extensive coding or advanced degrees in AI.

In terms of I/O profile, the demands of AI workloads such as training, fine-tuning, inference, RAG, and checkpointing vary. Fast storage is necessary for training large models from scratch, with checkpointing being I/O-intensive and critical for minimizing data loss in case of system failures during training runs. Checkpointing involves all compute stopping for writes, with the timing of checkpoints crucial for ensuring data integrity.