Afficher la pageAnciennes révisionsExporter en PDFTout plier/déplierHaut de page Cette page est en lecture seule. Vous pouvez afficher le texte source, mais ne pourrez pas le modifier. Contactez votre administrateur si vous pensez qu'il s'agit d'une erreur. ===== ZFS snapshots ===== * Voir espace utilisé : <code> [root@nas /ZP_dataM1/ovh]# zfs list -o space </code> * Cleaner des snapshots <code> [root@nas /ZP_dataM1/ovh]# zfs list -t snap |grep ovh |tail -4 ZP_dataM2/ovh@zfs-auto-snap_weekly-2015-11-28-10h30 16.2M - 129G - ZP_dataM2/ovh@zfs-auto-snap_weekly-2015-11-29-10h30 644M - 128G - ZP_dataM2/ovh@zfs-auto-snap_daily-2015-11-30-08h30 0 - 128G - ZP_dataM2/ovh@zfs-auto-snap_weekly-2015-11-30-10h30 0 - 128G - </code> <code> [root@nas /ZP_dataM1/ovh]# zfs destroy ZP_dataM2/ovh@zfs-auto-snap_weekly-2015-11-30-10h30 </code> ===== Infos disques + perfs ===== <code> [root@freebsdVM ~]# diskinfo -ctv da2 da2 512 # sectorsize 8589934592 # mediasize in bytes (8.0G) 16777216 # mediasize in sectors 0 # stripesize 0 # stripeoffset 1044 # Cylinders according to firmware. 255 # Heads according to firmware. 63 # Sectors according to firmware. # Disk ident. I/O command overhead: time to read 10MB block 0.089807 sec = 0.004 msec/sector time to read 20480 sectors 6.371639 sec = 0.311 msec/sector calculated command overhead = 0.307 msec/sector Seek times: Full stroke: 250 iter in 2.603755 sec = 10.415 msec Half stroke: 250 iter in 5.275366 sec = 21.101 msec Quarter stroke: 500 iter in 7.446248 sec = 14.892 msec Short forward: 400 iter in 3.744817 sec = 9.362 msec Short backward: 400 iter in 3.695824 sec = 9.240 msec Seq outer: 2048 iter in 0.680164 sec = 0.332 msec Seq inner: 2048 iter in 0.904928 sec = 0.442 msec Transfer rates: outside: 102400 kbytes in 0.844724 sec = 121223 kbytes/sec middle: 102400 kbytes in 0.892766 sec = 114700 kbytes/sec inside: 102400 kbytes in 1.150101 sec = 89036 kbytes/sec </code> ===== Monter un snap ZFS ===== <code> zfs list -t snap -r ZP_dataM1/mp3 mount -t zfs ZP_dataM1/mp3@zfs-auto-snap_daily-2015-11-21-08h30 /mnt </code> ===== gpart (fdisk) ===== gpart show -l da0 gpart show da0 ===== zpool dégradé ===== <code> camcontrol rescan all zpool online system /dev/gpt/system1 </code> ===== Install avec zpool mirroré (RAID 1) ===== => choisir "Shell" lors du step de partitionnement * Lister les disques camcontrol devlist * Création de la table de partition, sur chaque disque : <code> # gpart create -s gpt da0 # gpart add -b 34 -s 512k -t freebsd-boot -l boot0 da0 # gpart add -s 2G -t freebsd-swap -l swap0 da0 # gpart add -s 10G -t freebsd-zfs -l system0 da0 </code> <code> # gpart create -s gpt da1 # gpart add -b 34 -s 512k -t freebsd-boot -l boot1 ada1 # gpart add -s 2G -t freebsd-swap -l swap1 ada1 # gpart add -s 10G -t freebsd-zfs -l system1 ada1 </code> * Install du bootcode <code> # gpart bootcode -b /boot/pmbr -p /boot/gptzfsboot -i 1 ada0 # gpart bootcode -b /boot/pmbr -p /boot/gptzfsboot -i 1 ada1 </code> * Création du zpool <code> # zpool create -m none -o altroot=/mnt -o cachefile=/var/tmp/zpool.cache \ system mirror /dev/gpt/system0 /dev/gpt/system1 # zfs set mountpoint=/ system </code> * Création des FS <code> # zfs create -o compression=on -o setuid=off system/tmp # chmod 1777 /mnt/tmp # zfs create system/usr # zfs create system/usr/home # cd /mnt # ln -s usr/home home # zfs create system/usr/local # zfs create -o compression=on -o setuid=off system/usr/ports # zfs create -o exec=off -o setuid=off system/usr/ports/distfiles # zfs create -o exec=off -o setuid=off system/usr/ports/packages # zfs create system/usr/obj # zfs create -o compression=on -o exec=off -o setuid=off system/usr/src # zfs create system/var # zfs create -o exec=off -o setuid=off system/var/backups # zfs create -o compression=on -o exec=off -o setuid=off system/var/crash # zfs create -o exec=off -o setuid=off system/var/db # zfs create -o exec=on -o compression=on -o setuid=off system/var/db/pkg # zfs create -o exec=off -o setuid=off system/var/empty # zfs create -o compression=on -o exec=off -o setuid=off system/var/log # zfs create -o compression=on -o exec=off -o setuid=off system/var/mail # zfs create -o exec=off -o setuid=off system/var/run # zfs create -o compression=on -o setuid=off system/var/tmp # chmod 1777 /mnt/var/tmp # zpool set bootfs=system system # mkdir -p /mnt/boot/zfs # cp /var/tmp/zpool.cache /mnt/boot/zfs/zpool.cache </code> => Continuer l'install, ouvrir un shell à la fin de l'install <code> # echo 'zfs_load="YES"' >> /boot/loader.conf # echo 'vfs.root.mountfrom="zfs:system"' >> /boot/loader.conf # echo 'zfs_enable="YES"' >> /etc/rc.conf # cd /media # mkdir cdrom flash </code> * /etc/fstab <code> # Device Mountpoint FStype Options Dump Pass# /dev/gpt/swap0 none swap sw 0 0 /dev/gpt/swap1 none swap sw 0 0 /dev/cd0 /media/cdrom cd9660 ro,noauto 0 0 </code> * Après premier boot <code> # zfs set readonly=on system/var/empty # rm /etc/motd </code> ===== FreeBSD ports ===== * Mise à jour <code> # portsnap fetch extract # portsnap fetch update </code> * Install <code> cd /usr/ports/net/samba41 make install clean </code> ===== FreeBSD packages ===== <code> pkg update pkg search xxxxxxx pkg install xxxxxxx pkg info -D -x subsonic-jetty-5.2.1 </code> ===== Empêcher routes dynamiques ===== * **/etc/sysctl.conf** <code> net.inet.ip.redirect=0 net.inet.icmp.drop_redirect=1 net.inet.icmp.log_redirect=0 </code> ===== Monter image ISO ===== <code> mdconfig -a -t vnode -f /path/to/image.iso -u 1 mount -t cd9660 /dev/md1 /mnt/cdrom </code> ===== Démonter image ISO ===== <code> mount -u /mnt/cdrom mdconfig -d -u 1 </code> ===== Augmenter la taille d'un zpool mirroré ===== On a un zpool dataZP de 4 Go et on veut le faire passer à 8 Go. On retire un premier disque de 4 Go qu'on remplace par un disque de 8 Go. Puis on remplace le disque restant de 4 Go par un nouveau disque de 8 Go. * Soit la config ci-dessous : <code> [root@freebsdVM ~]# egrep 'da[0-9]' /var/run/dmesg.boot|grep MB|grep -v trans da0: 8192MB (16777216 512 byte sectors: 255H 63S/T 1044C) da1: 8192MB (16777216 512 byte sectors: 255H 63S/T 1044C) da2: 4096MB (8388608 512 byte sectors: 255H 63S/T 522C) da3: 4096MB (8388608 512 byte sectors: 255H 63S/T 522C) </code> <code> [root@freebsdVM ~]# zpool status dataZP pool: dataZP state: ONLINE scan: scrub repaired 0 in 0h0m with 0 errors on Fri Nov 6 11:33:06 2015 config: NAME STATE READ WRITE CKSUM dataZP ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 da3 ONLINE 0 0 0 da2 ONLINE 0 0 0 errors: No known data errors </code> <code> [root@freebsdVM ~]# zpool list NAME SIZE ALLOC FREE FRAG EXPANDSZ CAP DEDUP HEALTH ALTROOT dataZP 3.98G 164K 3.98G 0% - 0% 1.00x ONLINE - zroot 5.97G 1.06G 4.91G 10% - 17% 1.00x ONLINE - </code> * On retire un disque de 4 Go : <code> [root@freebsdVM ~]# zpool status dataZP pool: dataZP state: DEGRADED status: One or more devices has been removed by the administrator. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Online the device using 'zpool online' or replace the device with 'zpool replace'. scan: scrub repaired 0 in 0h0m with 0 errors on Fri Nov 6 11:33:06 2015 config: NAME STATE READ WRITE CKSUM dataZP DEGRADED 0 0 0 mirror-0 DEGRADED 0 0 0 15131538193711764791 REMOVED 0 0 0 was /dev/da3 da2 ONLINE 0 0 0 errors: No known data errors </code> * On ajoute un disque de 8 Go : <code> [root@freebsdVM ~]# diskinfo -v da3 |grep bytes 8589934592 # mediasize in bytes (8.0G) </code> * On l'ajoute au pool : <code> zpool online dataZP da3 </code> * On retire le disque de 4 Go restant et on ajoute un disque de 8 Go : <code> [root@freebsdVM ~]# zpool online dataZP da2 </code> * On étend le ZP : => cf. EXPANDSZ = 4G <code> [root@freebsdVM ~]# zpool list NAME SIZE ALLOC FREE FRAG EXPANDSZ CAP DEDUP HEALTH ALTROOT dataZP 3.98G 172K 3.98G 0% 4G 0% 1.00x ONLINE - zroot 5.97G 1.06G 4.91G 10% - 17% 1.00x ONLINE - </code> <code> zpool online -e dataZP da2 zpool online -e dataZP da3 </code> <code> NAME SIZE ALLOC FREE FRAG EXPANDSZ CAP DEDUP HEALTH ALTROOT dataZP 7.98G 310K 7.98G 0% - 0% 1.00x ONLINE - zroot 5.97G 1.06G 4.91G 10% - 17% 1.00x ONLINE - </code> ===== Exclure package - upgrade ===== <code> [root@nas /var/log]# pkg lock nut-2.7.3_3 nut-2.7.3_3: lock this package? [y/N]: y Locking nut-2.7.3_3 </code> informatique/bsd.txt Dernière modification : 2018/12/11 08:39de ben