Surely "1337" is the same as 1337, right?

https://sopuli.xyz/post/14326368

Surely "1337" is the same as 1337, right? - Sopuli

Meme transcription: Panel 1: Bilbo Baggins ponders, “After all… why should I care about the difference between int and String? Panel 2: Bilbo Baggins is revealed to be an API developer. He continues, “JSON is always String, anyways…”

To whoever does that, I hope that there is a special place in hell where they for you to do type safe API bindings for a JSON API, and every time you use the wrong type for a value, they cave your skull in.
The worst thing is: you can’t even put an int in a json file. Only doubles. For most people that is fine, since a double can function as a 32 bit int. But not when you are using 64 bit identifiers or timestamps.
That’s an artifact of JavaScript, not JSON. The JSON spec states that numbers are a sequence of digits with up to one decimal point. Implementations are not obligated to decode numbers as floating point. Go will happily decode into a 64-bit int, or into an arbitrary precision number.
What that means is that you cannot rely on numbers in JSON. Just use strings.
Unless you’re dealing with some insanely flexible schema, you should be able to know what kind of number (int, double, and so on) a field should contain when deserializing a number field in JSON. Using a string does not provide any benefits here unless there’s some big in your deserialzation process.
What’s the point of your schema if the receiving end is JavaScript, for example? You can convert a string to BigNumber, but you’ll get wrong data if you’re sending a number.
I’m not following your point so I think I might be misunderstanding it. If the types of numbers you want to express are literally incapable of being expressed using JSON numbers then yes, you should absolutely use string (or maybe even an object of multiple fields).
The point is that everything is expressable as JSON numbers, it’s when those numbers are read by JS there’s an issue
Can you give a specific example? Might help me understand your point.
I am not sure what could be the example, my point was that the spec and the RFC are very abstract and never mention any limitations on the number content. Of course the implementations in the language will be more limited than that, and if limitations are different, it will create dissimilar experience for the user, like this: Why does JSON.parse corrupt large numbers and how to solve this
JSON

This is what I was getting at here programming.dev/comment/10849419 (although I had a typo and said big instead of bug). The problem is with the parser in those circumstances, not the serialization format or language.
Surely "1337" is the same as 1337, right? - programming.dev

Meme transcription: Panel 1: Bilbo Baggins ponders, “After all… why should I care about the difference between int and String? Panel 2: Bilbo Baggins is revealed to be an API developer. He continues, “JSON is always String, anyways…”

I disagree a bit in that the schema often doesn’t specify limits and operates in JSON standard’s terms, it will say that you should get/send a number, but will not usually say at what point will it break.

This is the opposite of what C language does, being so specific that it is not even turing complete (in a theoretical sense, it is practically)

Then the problem is the schema being under specified. Take the classic pet store example. It says that the I’d is int64. petstore3.swagger.io/#/store/placeOrder

If some API is so underspecified that it just says “number” then I’d say the schema is wrong. If your JSON parser has no way of passing numbers as arbitrary length number types (like BigDecimal in Java) then that’s a problem with your parser.

I don’t think the truly truly extreme edge case of things like C not technically being able to simulate a truly infinite tape in a Turing machine is the sort of thing we need to worry about. I’m sure if the JSON object you’re parsing is some astronomically large series of nested objects that specifications might begin to fall apart too (things like the maximum amount of memory any specific processor can have being a finite amount), but that doesn’t mean the format is wrong.

Swagger UI

Underspecified schema is indeed a problem, but I find it too common to just shrug it off

Also, you’re very right that just using strings will not improve the situation 🤝

What makes you think so?

const bigJSON = '{"gross_gdp": 12345678901234567890}'; JSON.parse(bigJSON, (key, value, context) => { if (key === "gross_gdp") { // Ignore the value because it has already lost precision return BigInt(context.source); } return value; }); > {gross_gdp: 12345678901234567890n}

developer.mozilla.org/en-US/docs/Web/…/parse

JSON.parse() - JavaScript | MDN

The JSON.parse() static method parses a JSON string, constructing the JavaScript value or object described by the string. An optional reviver function can be provided to perform a transformation on the resulting object before it is returned.

MDN Web Docs
Because no one is using JSON.parse directly. Do you guys even code?
It’s neither JSON’s nor JavaScript’s fault that you don’t want to make a simple function call to properly deserialize the data.
It’s not up to me. Or you.