The storage question for AI home labs gets answered wrong more often than almost any other component decision. People either buy an expensive NAS because it looks professional, or they skip shared storage entirely and end up with datasets scattered across three machines with no coherent access pattern. Neither is optimal, and the right answer depends on specifics that most guides gloss over.
I spent six years as a storage systems architect before moving into tech journalism, and the NAS vs DAS question comes up constantly in AI infrastructure discussions — at home lab scale and at enterprise scale. The physics of the trade-off are the same regardless of budget. What changes is the cost of getting it wrong.
Defining the Terms Precisely
Direct-Attached Storage (DAS) means storage connected directly to a single host — internal drives, an external enclosure connected via USB or Thunderbolt, or a PCIe NVMe expansion card. The storage is local to one machine and not directly accessible by others without going through that machine’s network stack.
Network-Attached Storage (NAS) means storage accessible over a network — either a dedicated NAS appliance or a machine running NAS software like TrueNAS or OpenMediaVault. Multiple clients access the storage simultaneously via protocols like NFS, SMB, or iSCSI. The NAS device handles the storage, the network handles the transport, and clients treat it as a remote filesystem.
The distinction matters because these two architectures have fundamentally different performance characteristics, failure modes, and operational trade-offs — and the right choice depends on which of those factors dominates your workload.
The Throughput Reality for AI Dataset Access
Sequential read throughput is what matters for AI training dataset loading — you’re reading large files in order, not bouncing around randomly. Here’s what the different options actually deliver in practice.
The numbers tell an immediate story. Internal NVMe DAS is six to eight times faster than a 10GbE NAS — the network is the ceiling, not the storage. But 10GbE NAS at 900–1,100 MB/s is comparable to Thunderbolt-attached external NVMe and far faster than any spinning disk configuration. The 1GbE NAS number — 90–115 MB/s — is the one that causes pain in practice, because it’s only marginally faster than a single internal HDD and substantially slower than anything with real AI workload requirements.
A NAS on 1GbE is not a viable AI storage solution for active training workloads. At 115 MB/s peak, loading a 100GB dataset takes over fourteen minutes. The same dataset loads in under two minutes over 10GbE NAS, and under thirty seconds from local NVMe. If you’re considering NAS for AI work, 10GbE is a hard requirement, not an upgrade.
Where DAS Wins Unambiguously
For single-machine AI workloads — one workstation running training, inference, or fine-tuning — DAS is the correct choice for active working storage. The latency of local NVMe is sub-millisecond. There’s no network protocol overhead, no NIC processing, no NAS CPU mediating every I/O operation. Model files load faster, checkpoints save faster, and the CPU cycles that would be spent on network I/O are free for actual compute work.
The practical DAS configuration for a home AI workstation is straightforward: a 2TB NVMe Gen4 drive for the OS, model weights, and active datasets, plus a 4–8TB internal HDD for archival datasets not currently in active training rotation. This covers the vast majority of home lab AI workloads at minimum cost and maximum simplicity.
DAS also wins on reliability for single-machine setups. There’s no network dependency, no NAS appliance to fail, no cable between the storage and the compute. For a single-user home lab, the added resilience of NAS architecture — protecting against drive failure via RAID — is real but often oversold relative to the operational complexity it adds. A good backup strategy provides equivalent protection at lower complexity.
Where NAS Wins Unambiguously
The moment you have more than one machine that needs access to the same data, NAS becomes compelling. Maintaining synchronized copies of large datasets across multiple machines manually is an operational tax that compounds quickly. A 500GB image dataset kept in sync between three machines via rsync is manageable. A 2TB multimodal dataset updated regularly is a full-time job to keep synchronized without shared storage.
NAS also wins for dataset archiving and management at scale. A four-bay NAS with 20TB of usable HDD storage costs $500–$800 all-in and provides a single organized repository for all training data — accessible from any machine on the network, browsable via a web interface, protected by RAID against drive failure, and trivially expandable by swapping drives for higher-capacity ones. That’s operationally cleaner than distributing datasets across internal drives on multiple machines.
Model serving is another NAS sweet spot. If you’re running an inference server that serves requests to multiple client machines, hosting model weights on a 10GbE NAS means the inference server loads models from shared storage without manual file distribution. When you update a model, every client sees the update immediately. For a multi-machine home lab running local AI services, this operational simplicity is genuinely valuable.
The Latency Question for Training Workloads
Storage latency matters differently for different AI workload patterns. For training with a standard PyTorch DataLoader, data is loaded into RAM at the start of each epoch — or prefetched into a RAM buffer asynchronously. Once the data is in RAM, storage latency becomes irrelevant for training speed. The question is how long the initial load takes and whether the prefetch buffer can keep the GPU fed between loads.
With a well-configured DataLoader using multiple workers and prefetching, a 10GbE NAS at 900 MB/s is rarely the bottleneck for training throughput. The GPU’s data consumption rate during forward and backward passes is typically slower than the NAS can supply, especially for image datasets with augmentation in the preprocessing pipeline. The preprocessing CPU work is often the actual constraint, not storage bandwidth.
Where NAS latency does matter is checkpoint frequency. If your training loop saves a checkpoint every hundred steps and each checkpoint is 5GB, a 10GbE NAS saves that checkpoint in five to six seconds. From local NVMe, the same save takes under one second. If your training pipeline blocks on checkpoint saves — rather than saving asynchronously — this adds up over a long training run. The fix is asynchronous checkpoint saving, not switching from NAS to DAS.
iSCSI: The Middle Path Worth Knowing About
Most home lab NAS deployments use NFS or SMB — file-level protocols that present the NAS as a remote filesystem. iSCSI is a block-level protocol that presents a NAS volume as a raw block device to a single client, which the client then formats and mounts as a local disk. From the client’s perspective, an iSCSI volume behaves identically to a local drive.
iSCSI over 10GbE delivers throughput close to the network ceiling — 900–1,050 MB/s — and with lower protocol overhead than NFS for random I/O workloads. For a single-client workload where you want NAS flexibility (centralized storage, RAID protection, easy expansion) with near-DAS performance, iSCSI is worth considering. The trade-off is that iSCSI volumes are exclusive to one client at a time — you lose the multi-client simultaneous access that makes NFS and SMB useful.
The Hybrid Architecture: What Most Multi-Machine Labs Should Use
For a home lab with two to four machines doing AI work, the architecture that works best in practice is a hybrid: local NVMe DAS on each workstation for active training data and model weights, plus a NAS for dataset archiving, model repositories, and shared access to data that isn’t actively being trained on.
The workflow this enables is clean: pull datasets from the NAS to local NVMe when starting a new training run, train locally at full NVMe speed, push checkpoints and results back to the NAS for storage and sharing. The NAS handles the organization and redundancy problem. The local NVMe handles the performance problem. Neither compromises the other.
This is how most professional ML teams structure local infrastructure — not because it’s clever, but because it matches the access pattern of AI workloads naturally. Data that’s actively being trained on needs to be fast and local. Data that’s being stored or shared needs to be organized and accessible. These are different requirements that benefit from different solutions.
Practical NAS Recommendations for AI Home Labs
A four-bay NAS running TrueNAS Scale or Synology DSM with four 8TB HDDs in RAID-Z1 (TrueNAS) or SHR (Synology) gives you roughly 20–22TB of usable capacity with single-drive failure protection. Add an NVMe SSD as an L2ARC read cache if you’re frequently accessing the same datasets — the SSD accelerates repeated reads dramatically without the cost of an all-flash configuration.
The 10GbE connection is non-negotiable for AI workloads. A NAS on 1GbE is a dataset archive with a painful access tax, not a working storage solution. If your NAS doesn’t have an onboard 10GbE port, add a 10GbE NIC — most NAS appliances from Synology, QNAP, and similar vendors support PCIe NIC expansion.
For all-flash NAS that serves as primary training storage rather than archive, an NVMe-based NAS over 10GbE saturates the network interface. At that point you’re paying significant cost per terabyte for performance that’s capped by the network anyway. All-flash NAS makes sense for latency-sensitive inference workloads where models are frequently swapped — not for sequential training data loading where HDD RAID with an SSD cache is operationally equivalent at a fraction of the cost.
For detailed NAS recommendations at various budget points, see our Best NAS Drives 2026 guide, which covers both appliances and DIY builds across capacity tiers.
Frequently Asked Questions
Can I use a NAS as primary storage for AI training without a local NVMe?
Yes, if you have 10GbE connectivity and your dataset loading is asynchronous. A 10GbE NAS at 900+ MB/s is sufficient for most training workloads where the DataLoader prefetch buffer keeps the GPU fed. You’ll feel the difference compared to local NVMe during cold dataset loads and checkpoint saves, but during steady-state training with proper prefetching, the GPU is rarely waiting on storage. On 1GbE, this approach fails — the bandwidth is too low to keep even modest GPU utilization fed.
What RAID level should I use on a home lab NAS for AI datasets?
RAID-Z1 on TrueNAS (equivalent to RAID 5) or SHR on Synology for a four-bay configuration. This gives you single-drive failure protection with good read performance and reasonable write performance. RAID-Z2 (RAID 6 equivalent) provides two-drive failure protection at the cost of one additional drive’s worth of capacity — worthwhile for critical data, but the extra protection comes at meaningful cost per usable terabyte. RAID 10 gives excellent read and write performance with one-drive failure protection per mirrored pair, but sacrifices 50% of raw capacity. For large, relatively static AI datasets, RAID-Z1 is the practical sweet spot.
Is Thunderbolt DAS a good alternative to NAS for multi-machine access?
No, not in practice. Thunderbolt DAS connects to a single host. Sharing it to other machines requires that host to act as a file server over the network, at which point you’re limited by the host’s network throughput anyway — and you’ve added latency and a single point of failure (the host machine must be running). For single-machine use, Thunderbolt DAS is excellent. For multi-machine access, it’s an awkward workaround compared to purpose-built NAS.
How much RAM does a NAS need for AI workloads?
For a NAS serving AI datasets via NFS or SMB, RAM functions primarily as read cache — frequently accessed data is cached in RAM and served at memory speeds rather than disk speeds. Eight gigabytes of RAM is the minimum for a functioning NAS OS with any caching headroom. Sixteen gigabytes is comfortable for a four to eight-bay NAS serving two to four clients. Thirty-two gigabytes makes sense if you’re running ZFS with significant ARC cache on TrueNAS — ZFS uses RAM aggressively for caching and benefits from more. For Synology DSM, 8–16GB is typically sufficient.
Should I use NFS or SMB for Linux AI workstations?
NFS for Linux-to-Linux connections. SMB is the correct choice when you have Windows clients, and it works fine on Linux via the cifs kernel module — but NFS has lower protocol overhead for Linux-native workloads and better performance for the large sequential reads typical of dataset loading. Configure NFSv4 with async read and appropriate rsize/wsize values (typically 1MB) for AI dataset access patterns. The default NFS mount options are not optimized for large sequential reads and can leave significant throughput on the table.
WRITTEN BY

Sarah Lin
Storage & NAS Editor
6 years as a storage systems architect before moving into tech journalism. Has designed storage architectures for AI training environments ranging from single-node home labs to multi-petabyte enterprise datasets.
Specialties: NAS setup & configuration · SSD benchmarking · RAID configurations · ZFS & TrueNAS · AI dataset storage
Stay up to date with the latest AI hardware reviews, buying guides, and analysis at AiGigabit.com. Bookmark us for daily updates.


