Bug in java.net.http.WebSocket (?)
Simon Fischer
simon.fischer at ipp.mpg.de
Wed Mar 5 22:05:56 UTC 2025
Hi Daniel,
thanks for replying!
IMO, that reading of network byte order is a common misconception. The
rfc for IP/UDP/TCP defined that the fields of the IP headers would be
encoded in big endian, which was a very common native byte order at the
time. Afaik, there has never been any kind of standard defining that
network protocols should be big endian - that is just an interpretation
of the term "network byte order".
Any application layer protocol in the ISO/OSI stack can define byte
orders as the author likes. And in times where likely >95% of all
compute resources natively speak little endian, the sensible choice is
pretty obvious in my view.
But back to the actual problem here: Websockets, as any layered network
protocol, need to be transparent. I transfer a binary message (which is
a byte blob, in case of the WebSocket API its a ByteBuffer) to the
websocket layer, which uses the WS protocol on top of TCP on top of IP
on top of ethernet/wifi/5G/whatever to transport it to the reciever. And
on the other and, another websocket layer removes all the things around
and again just delivers a byte blob to the application. That final byte
blob must have the exact same content as the original one, or the
protocol is unusable. And this expectation is what the bug violates. The
byte order of the source buffer is just a hint for extracting multi-byte
values from it. It cannot influence the byte blob recieved by the
communication partner.
It is very likely that very few people have stumbled upon that problem,
because (for historic reasons, but unreasonably so) you have to actually
work hard in Java in order to work with little endian ByteBuffers
(although the JVM internally works with native byte order, which again
is 95% little endian). But to be frank: if any developer has ever built
a system based on the Java Websocket API, using little endian
ByteBuffers in the client and hacking the server in order to work with
the unintentional re-ordering of bytes at the API boundary (which is
actually hard, because that has to be applied selectively, when
fragmentation and alignments come into play), they surely deserve the
pain of a malfunctioning system when some undocumented misbehaviour
present since Java 11 is fixed...
Ok, rant over, sorry :-)
have a good evening
Simon Fischer
__
_
_
Simon Fischer
Developer - CoDaC
Department Operation
Max Planck Institute for Plasmaphysics
Wendelsteinstrasse 1
17491 Greifswald, Germany
Phone: +49(0)3834 88 1215
On March 5, 2025 19:50:33 Daniel Fuchs <daniel.fuchs at oracle.com> wrote:
> Hi Simon,
>
> Thank you for the report.
>
> I am not too familiar with WebSocket. But since this is
> a networking protocol I would expect binary data to be
> transferred on the wire in network byte order.
>
> So the expectation from WebSocket::sendBinary that the
> byte buffer is in network byte order (big endian) does
> not seem unrealistic to me.
>
> More concretely, since the API is there since java 11
> (at least in its standard form), it would be difficult
> to change this long standing behavior.
>
> That said - a deeper analysis of the issue and possible
> options is probably warranted. At least the expectations
> should be documented.
>
> best regards,
>
> -- daniel
>
> On 05/03/2025 17:53, Fischer, Simon wrote:
>> Hi all,
>>
>> no idea if I am in the right place here, but I have no account to create
>> a tracker issue and also could not find out how to get one…
>>
>> I was just using the java.net.html.WebSocket (jdk17 specifically, but
>> the issue should still be there at least in 21 from a quick look into
>> the code) for testing purposes and stumbled upon what I think is a bug:
>>
>> When I supply a _/little endian/_ ByteBuffer to WebSocket.sendBinary,
>> the payload will be scrambled on the way to the server.
>>
>> The actual issue is in jdk.internal.net.http.websocket.Frame. The
>> loop()-portion of the transferMasking algorithm uses
>> ByteBuffer.putLong(ByteBuffer.getLong) (“dst.putLong(j, src.getLong(i) ^
>> maskLong);”) to try to optimize the masking and data transfer. Problem
>> is that src is a ByteBuffer provided by the user, with the endianness
>> set by the user, while dst is an internally allocated ByteBuffer with
>> the default byte order. This obviously can lead to 8-byte-blocks of the
>> original message being re-ordered on the way to the client.
>>
>> The solution IMO would be to ensure that both buffers are set to the
>> same endianness. And it should probably be _/native endian/_, as the use
>> of a specific endianness would likely render the
>> long-vectorization/optimization useless on a machine which does not
>> support that endianness natively (getLong would reverse the byte order
>> when loading into native CPU register and putLong would reorder again).
>> In that case, actually any case, care must also be taken with regard to
>> the right encoding of the “maskLong” 64bit integer.
>>
>> Alternatively, one could just adopt the documentation to require the
>> ByteBuffer provided to WebSocket.sendBinary() to be default(/big)-endian
>> encoded. Semi-solution IMO.
>>
>> Would be interested in feedback and hope this finds somebody who can
>> make use of it J
>>
>> Best regards
>>
>> Simon Fischer
>>
>> --
>>
>> Simon Fischer
>>
>> Developer
>>
>> E5-CoDaC
>>
>> Max Planck Institut for Plasmaphysics
>>
>> Wendelsteinstrasse 1
>>
>> 17491 Greifswald, Germany
>>
>> Phone: +49(0)3834 88 1215
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mail.openjdk.org/pipermail/net-dev/attachments/20250305/8551918d/attachment.htm>
More information about the net-dev
mailing list