I just built a new nas4free to replace an aging 8x500gb zfs2. No problems with any of the disks just running out of room. Upgraded to 6x3TB ZFS2. When transferring data to the new box I got very slow network speeds of 30-40 MB/s. I was using an Intel PCI gigabit network card in the new machinne. I then discovered that I also had a spare Intel PCI-e gigabit network card so I stopped the file transfer, swapped in the PCI-e card and was gratified to see 110 MB/Sec network transfers. This was out of the box with no system tweaks other than the native tuning options (but not zfs kernel tune).
Memory used was between 25 and 30% of 8GB during the transfer of 2.81TB of data. Processors were a Q6600 Quad core and 2500. The Q6600 showed above 85% utilisation at some points during the transfer so I'm happy that the new box has the 2500.
File transfer was done using rsync.
Just posted for information rather than to start a thread of comments.
This is the old XigmaNAS forum in read only mode,
it will taken offline by the end of march 2021!
I like to aks Users and Admins to rewrite/take over important post from here into the new fresh main forum!
Its not possible for us to export from here and import it to the main forum!
it will taken offline by the end of march 2021!
I like to aks Users and Admins to rewrite/take over important post from here into the new fresh main forum!
Its not possible for us to export from here and import it to the main forum!
Gigabit maxed out
-
ku-gew
- Advanced User

- Posts: 172
- Joined: 29 Nov 2012 09:02
- Location: Den Haag, The Netherlands
- Status: Offline
Re: Gigabit maxed out
What models are the cards? strange that two Intel have so different performances.
HP Microserver N40L, 8 GB ECC, 2x 3TB WD Red, 2x 4TB WD Red
XigmaNAS stable branch, always latest version
SMB, rsync
XigmaNAS stable branch, always latest version
SMB, rsync
- ChriZathens
- Forum Moderator

- Posts: 758
- Joined: 23 Jun 2012 09:14
- Location: Athens, Greece
- Contact:
- Status: Offline
Re: Gigabit maxed out
I believe his first card was a PCI card, hence the slower speed
My Nas
Backup Nas: U-NAS NSC-400, Gigabyte MB10-DS4 (4x4TB Seagate Exos disks in RaidZ configuration - 32GB RAM)
- Case: Fractal Design Define R2
- M/B: Supermicro x9scl-f
- CPU: Intel Celeron G1620
- RAM: 16GB DDR3 ECC (2 x Kingston KVR1333D3E9S/8G)
- PSU: Chieftec 850w 80+ modular
- Storage: 8x2TB HDDs in a RaidZ2 array ~ 10.1 TB usable disk space
- O/S: XigmaNAS 11.2.0.4.6625 -amd64 embedded
- Extra H/W: Dell Perc H310 SAS controller, crosflashed to LSI 9211-8i IT mode, 8GB Innodisk D150SV SATADOM for O/S
Backup Nas: U-NAS NSC-400, Gigabyte MB10-DS4 (4x4TB Seagate Exos disks in RaidZ configuration - 32GB RAM)
-
ku-gew
- Advanced User

- Posts: 172
- Joined: 29 Nov 2012 09:02
- Location: Den Haag, The Netherlands
- Status: Offline
Re: Gigabit maxed out
Oh I see. Interesting! Standard PCI should have 3x that bandwidth.
HP Microserver N40L, 8 GB ECC, 2x 3TB WD Red, 2x 4TB WD Red
XigmaNAS stable branch, always latest version
SMB, rsync
XigmaNAS stable branch, always latest version
SMB, rsync
- b0ssman
- Forum Moderator

- Posts: 2438
- Joined: 14 Feb 2013 08:34
- Location: Munich, Germany
- Status: Offline
Re: Gigabit maxed out
yes buts its shared between all pci devices plus overhead
Nas4Free 11.1.0.4.4517. Supermicro X10SLL-F, 16gb ECC, i3 4130, IBM M1015 with IT firmware. 4x 3tb WD Red, 4x 2TB Samsung F4, both GEOM AES 256 encrypted.
