diff --git a/README.md b/README.md
index e45e1ef..4634e81 100644
--- a/README.md
+++ b/README.md
@@ -48,14 +48,14 @@ It has the following features:
 
 ## Installation
 
-```shell session
+```console
 sadfas
 
 ```
 
 ## Usage
 
-```shell session
+```console
 usage: zfs_autobackup [-h] [--ssh-source SSH_SOURCE] [--ssh-target SSH_TARGET]
                       [--keep-source KEEP_SOURCE] [--keep-target KEEP_TARGET]
                       [--no-snapshot] [--allow-empty] [--ignore-replicated]
@@ -156,7 +156,7 @@ Its important to choose a unique and consistent backup name. In this case we nam
 
 On the source zfs system set the ```autobackup:offsite``` zfs property to true:
 
-```shell session
+```console
 [root@pve ~]# zfs set autobackup:offsite1=true rpool
 [root@pve ~]# zfs get -t filesystem,volume autobackup:offsite1
 NAME                                    PROPERTY             VALUE                SOURCE
@@ -171,7 +171,7 @@ rpool/swap                              autobackup:offsite1  true
 
 Because we dont want to backup everything, we can exclude certain filesystem by setting the property to false:
 
-```shell session
+```console
 [root@pve ~]# zfs set autobackup:offsite1=false rpool/swap
 [root@pve ~]# zfs get -t filesystem,volume autobackup:offsite1
 NAME                                    PROPERTY             VALUE                SOURCE
@@ -192,7 +192,7 @@ First install the ssh-key on the server that you specify with --ssh-source or --
 
 #### Method 1: Run the script on the backup server and pull the data from the server specfied by --ssh-source. This is usually the preferred way and prevents a hacked server from accesing the backup-data
 
-```shell session
+```console
 [root@backup ~]# zfs_autobackup --ssh-source pve.server.com offsite1 backup/pve --progress --verbose --resume
 
   #### Settings summary
@@ -238,7 +238,7 @@ First install the ssh-key on the server that you specify with --ssh-source or --
 
 Method 2: Run the script on the server and push the data to the backup server specified by --ssh-target:
 
-```shell session
+```console
 [root@pve ~]# zfs_autobackup --ssh-target backup.server.com offsite1 backup/pve --progress --verbose --resume
 
   #### Settings summary
@@ -272,7 +272,7 @@ Method 2: Run the script on the server and push the data to the backup server sp
 
 Add this to your ~/.ssh/config:
 
-```shell session
+```console
 Host *
     ControlPath ~/.ssh/control-master-%r@%h:%p
     ControlMaster auto
@@ -287,7 +287,7 @@ Thanks @mariusvw :)
 
 The correct way to do this is by creating ~/.ssh/config:
 
-```shell session
+```console
 Host smartos04
     Hostname 1.2.3.4
     Port 1234
@@ -320,7 +320,7 @@ Try using something like: --filter-properties xattr
 
 Restoring can be done with simple zfs commands. For example, use this to restore a specific SmartOS disk image to a temporary restore location:
 
-```shell session
+```console
 root@fs1:/home/psy#  zfs send fs1/zones/backup/zfsbackups/smartos01.server.com/zones/a3abd6c8-24c6-4125-9e35-192e2eca5908-disk0@smartos01_fs1-20160110000003 | ssh root@2.2.2.2 "zfs recv zones/restore"
 ```
 
@@ -332,7 +332,7 @@ You can monitor backups by using my zabbix-jobs script. (<https://github.com/psy
 
 Put this command directly after the zfs_backup command in your cronjob:
 
-```shell session
+```console
 zabbix-job-status backup_smartos01_fs1 daily $?
 ```
 
@@ -352,7 +352,7 @@ The backup will go to a machine named smartos03.
 
 On each node select the filesystems as following:
 
-```shell session
+```console
 root@h4:~# zfs set autobackup:h4_smartos03=true rpool
 root@h4:~# zfs set autobackup:h4_smartos03=false rpool/data
 root@h4:~# zfs set autobackup:data_smartos03=child rpool/data
@@ -371,7 +371,7 @@ Extra options needed for proxmox with HA:
 
 I use the following backup script on the backup server:
 
-```shell session
+```console
 for H in h4 h5 h6; do
   echo "################################### DATA $H"
   #backup data filesystems to a common place