I stream my desktop, with a few virtual screens, using ImmersedVR [1], but other options are available, like VirtualDesktop [2] (over local/remote network), which can be used with something like ShadowPC for full cloud. Others I know code completely in the native browser with the multi-screen interface you see in Meta commercials. Native is still a work in progress, but you can sideload android apps to put them into their own windows.
For desktop streaming, you can use your desktop keyboard/mouse, or a bluetooth keyboard/mouse connected to the headset. For native apps, you use a bluetooth keyboard/mouse connected to the headset. The headset sees/tracks the (supported) keyboard with the passthrough cameras and shows a passthrough window/overlay, where it is.
I have no doubt that virtual displays are the future, but probably not until form factors are like the nreal air [3], which I think is still too limited for resolution. Next gen, I'll probably switch over. People will probably be happy enough once PPD triples.
Out of curiosity, how does that work? What's the flow, what software do you use, do you still use a keyboard to type?