Usually yes, that is a limit you would mostly not want to exceed. It would slow down producers (not ack newer writes until older backup copies are flushed to disk).
The backup copies of the data is stored on a configurable amout of _other_ systems in the cluster. They receive it to memory, and then the write is acked to the client. Then the backup is written asyncronous to disk on those systems.
You could configure that backups are not written to disk at all, if you have so much memory in the cluster (and do not care for that additional level of durability). Or the size of the buffer of unflushed data which would cover some longer write spikes.
The point is more that the backup copies become "free" once they are written to disk. Otherwise the cluster could only store 25% of it's memory size if you want 3 backup copies.