this post was submitted on 14 May 2024
312 points (91.5% liked)

Programmer Humor

32562 readers
398 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
 

Explanation: Python is a programming language. Numpy is a library for python that makes it possible to run large computations much faster than in native python. In order to make that possible, it needs to keep its own set of data types that are different from python's native datatypes, which means you now have two different bool types and two different sets of True and False. Lovely.

Mypy is a type checker for python (python supports static typing, but doesn't actually enforce it). Mypy treats numpy's bool_ and python's native bool as incompatible types, leading to the asinine error message above. Mypy is "technically" correct, since they are two completely different classes. But in practice, there is little functional difference between bool and bool_. So you have to do dumb workarounds like declaring every bool values as bool | np.bool_ or casting bool_ down to bool. Ugh. Both numpy and mypy declared this issue a WONTFIX. Lovely.

you are viewing a single comment's thread
view the rest of the comments
[–] Ephera@lemmy.ml 12 points 6 months ago (4 children)

So many people here explaining why Python works that way, but what's the reason for numpy to introduce its own boolean? Is the Python boolean somehow insufficient?

[–] baod_rate@programming.dev 22 points 6 months ago

From numpy's docs:

The bool_ data type is very similar to the Python bool but does not inherit from it because Python’s bool does not allow itself to be inherited from, and on the C-level the size of the actual bool data is not the same as a Python Boolean scalar.

and likewise:

The int_ type does not inherit from the int built-in under Python 3, because type int is no longer a fixed-width integer type.

[–] breadsmasher@lemmy.world 8 points 6 months ago* (last edited 6 months ago)

here’s a good question answer on this topic

https://stackoverflow.com/questions/18922407/boolean-and-type-checking-in-python-vs-numpy

plus this is kinda the tools doing their jobs.

bool_ exists for whatever reason. its not a bool but functionally equivalent.

the static type checker mpy, correctly, states bool_ and bool aren’t compatible. in the same way other type different types aren’t compatible

[–] mynachmadarch@kbin.social 6 points 6 months ago (2 children)

Technically the Python bool is fine, but it's part of what makes numpy special. Under the hood numpy uses c type data structures, (can look into cython if you want to learn more).

It's part of where the speed comes from for numpy, these more optimized c structures, this means if you want to compare things (say an array of booleans to find if any are false) you either need to slow back down and mix back in Python's frameworks, or as numpy did, keep everything cython, make your own data type, and keep on trucking knowing everything is compatible.

There's probably more reasons, but that's the main one I see. If they depend on any specific logic (say treating it as an actual boolean and not letting you adding two True values together and getting an int like you do in base Python) then having their own also ensures that logic.

[–] Ephera@lemmy.ml 2 points 6 months ago

You know, at some point in my career I thought, it was kind of silly that so many programming languages optimize speed so much.

But I guess, that's what you get for not doing it. People having to leave your ecosystem behind and spreading across Numpy/Polars, Cython, plain C/Rust and probably others. 🫠

[–] rwhitisissle@lemmy.ml 2 points 6 months ago

This is the only actual explanation I've found for why numpy leverages its own implementation of what is in most languages a primitive data type, or a derivative of an integer.

[–] palordrolap@kbin.social 5 points 6 months ago* (last edited 6 months ago)

Someone else points out that Python's native bool is a subtype of int, so adding a bool to an int (or performing other mixed operations) is not an error, which might then go on to cause a hard-to-catch semantic/mathematical error.

I am assuming that trying to add a NumPy bool_ to an int causes a compilation error at best and a run-time warning, or traceable program crash at worst.