Hi Junsen,
This is a good question. For the Julia version of ITensor, we are soon planning to implement removeqns for the ITensor level but haven't just yet.
But if what you want for now is to map block sparse ITensors to dense ITensors, you can call the function dense on an ITensor and it will do the behavior that you are wanting.
The plan for removeqns is that it will keep other sparsity (such as diagonal sparsity) while removing QN information. This is in contrast to the dense functiono which removes all sparsity, whether associated with QNs or not.
So in short, please try dense on your ITensors. I'm implementing a version of that you can call directly on an MPS too right now, but with the current version you'll need to call it on each tensor yourself.
Best,
Miles