diff --git a/docs.it4i/anselm/storage.md b/docs.it4i/anselm/storage.md
index 3b0e0ae28ed89efa99f06d25be3d9347b7125553..2fa8128b41d1f13992429a35319281a74c8a0759 100644
--- a/docs.it4i/anselm/storage.md
+++ b/docs.it4i/anselm/storage.md
@@ -120,7 +120,8 @@ Default stripe size is 1MB, stripe count is 1. There are 22 OSTs dedicated for t
 | Mountpoint           | /home  |
 | Capacity             | 320 TB |
 | Throughput           | 2 GB/s |
-| User quota           | 250 GB |
+| User space quota     | 250 GB |
+| User inodes quota    | 500 k  |
 | Default stripe size  | 1 MB   |
 | Default stripe count | 1      |
 | Number of OSTs       | 22     |
@@ -145,10 +146,11 @@ The SCRATCH filesystem is realized as Lustre parallel filesystem and is availabl
 | SCRATCH filesystem   |          |
 | -------------------- | -------- |
 | Mountpoint           | /scratch |
-| Capacity             | 146TB    |
-| Throughput           | 6GB/s    |
-| User quota           | 100TB    |
-| Default stripe size  | 1MB      |
+| Capacity             | 146 TB   |
+| Throughput           | 6 GB/s   |
+| User quota           | 100 TB   |
+| User inodes quota    | 10 M     |
+| Default stripe size  | 1 MB     |
 | Default stripe count | 1        |
 | Number of OSTs       | 10       |
 
@@ -178,7 +180,7 @@ Filesystem:    /scratch
 Space used:    0
 Space limit:   93T
 Entries:       0
-Entries limit: 0
+Entries limit: 10m
 ```
 
 In this example, we view current size limits and space occupied on the /home and /scratch filesystem, for a particular user executing the command.
@@ -269,8 +271,8 @@ The local scratch filesystem is intended for temporary scratch data generated du
 | ------------------------ | -------------------- |
 | Mountpoint               | /lscratch            |
 | Accesspoint              | /lscratch/$PBS_JOBID |
-| Capacity                 | 330GB                |
-| Throughput               | 100MB/s              |
+| Capacity                 | 330 GB               |
+| Throughput               | 100 MB/s             |
 | User quota               | none                 |
 
 ### RAM Disk
@@ -287,13 +289,13 @@ The local RAM disk filesystem is intended for temporary scratch data generated d
 !!! note
     The local RAM disk directory /ramdisk/$PBS_JOBID will be deleted immediately after the calculation end. Users should take care to save the output data from within the jobscript.
 
-| RAM disk    |                                                                                                         |
-| ----------- | ------------------------------------------------------------------------------------------------------- |
-| Mountpoint  | /ramdisk                                                                                                |
-| Accesspoint | /ramdisk/$PBS_JOBID                                                                                     |
-| Capacity    | 60GB at compute nodes without accelerator, 90GB at compute nodes with accelerator, 500GB at fat nodes   |
-| Throughput  | over 1.5 GB/s write, over 5 GB/s read, single thread, over 10 GB/s write, over 50 GB/s read, 16 threads |
-| User quota  | none                                                                                                    |
+| RAM disk    |                                                                                                          |
+| ----------- | -------------------------------------------------------------------------------------------------------- |
+| Mountpoint  | /ramdisk                                                                                                 |
+| Accesspoint | /ramdisk/$PBS_JOBID                                                                                      |
+| Capacity    | 60 GB at compute nodes without accelerator, 90 GB at compute nodes with accelerator, 500 GB at fat nodes |
+| Throughput  | over 1.5 GB/s write, over 5 GB/s read, single thread, over 10 GB/s write, over 50 GB/s read, 16 threads  |
+| User quota  | none                                                                                                     |
 
 ### Tmp
 
@@ -301,13 +303,13 @@ Each node is equipped with local /tmp directory of few GB capacity. The /tmp dir
 
 ## Summary
 
-| Mountpoint | Usage                     | Protocol | Net Capacity   | Throughput | Limitations | Access                  | Services                    |        |
-| ---------- | ------------------------- | -------- | -------------- | ---------- | ----------- | ----------------------- | --------------------------- | ------ |
-| /home      | home directory            | Lustre   | 320 TiB        | 2 GB/s     | Quota 250GB | Compute and login nodes | backed up                   |        |
-| /scratch   | cluster shared jobs' data | Lustre   | 146 TiB        | 6 GB/s     | Quota 100TB | Compute and login nodes | files older 90 days removed |        |
-| /lscratch  | node local jobs' data     | local    | 330 GB         | 100 MB/s   | none        | Compute nodes           | purged after job ends       |        |
-| /ramdisk   | node local jobs' data     | local    | 60, 90, 500 GB | 5-50 GB/s  | none        | Compute nodes           | purged after job ends       |        |
-| /tmp       | local temporary files     | local    | 9.5 GB         | 100 MB/s   | none        | Compute and login nodes | auto                        | purged |
+| Mountpoint | Usage                     | Protocol | Net Capacity   | Throughput | Space/Inodes quota | Access                  | Services                    |        |
+| ---------- | ------------------------- | -------- | -------------- | ---------- | ------------------ | ----------------------- | --------------------------- | ------ |
+| /home      | home directory            | Lustre   | 320 TiB        | 2 GB/s     | 250 GB / 500 k     | Compute and login nodes | backed up                   |        |
+| /scratch   | cluster shared jobs' data | Lustre   | 146 TiB        | 6 GB/s     | 100 TB / 10 M      | Compute and login nodes | files older 90 days removed |        |
+| /lscratch  | node local jobs' data     | local    | 330 GB         | 100 MB/s   | none / none        | Compute nodes           | purged after job ends       |        |
+| /ramdisk   | node local jobs' data     | local    | 60, 90, 500 GB | 5-50 GB/s  | none / none        | Compute nodes           | purged after job ends       |        |
+| /tmp       | local temporary files     | local    | 9.5 GB         | 100 MB/s   | none / none        | Compute and login nodes | auto                        | purged |
 
 ## CESNET Data Storage