unsigned long long to_ullong() const | (since C++11) (until C++23) | |
constexpr unsigned long long to_ullong() const | (since C++23) |
Converts the contents of the bitset to an unsigned long long
integer.
The first bit of the bitset corresponds to the least significant digit of the number and the last bit corresponds to the most significant digit.
(none).
the converted integer.
std::overflow_error
if the value can not be represented in unsigned long long
.
#include <iostream> #include <bitset> #include <limits> int main() { std::bitset<std::numeric_limits<unsigned long long>::digits> b( 0x123456789abcdef0LL ); std::cout << b << " " << std::hex << b.to_ullong() << '\n'; b.flip(); std::cout << b << " " << b.to_ullong() << '\n'; }
Output:
0001001000110100010101100111100010011010101111001101111011110000 123456789abcdef0 1110110111001011101010011000011101100101010000110010000100001111 edcba9876543210f
returns a string representation of the data (public member function) |
|
returns an unsigned long integer representation of the data (public member function) |
© cppreference.com
Licensed under the Creative Commons Attribution-ShareAlike Unported License v3.0.
https://en.cppreference.com/w/cpp/utility/bitset/to_ullong