I know a google engineer who was saying they’re having to update their code bases to handle > 16 exabytes of storage, if you can imagine. But yeah, that’s storage, not RAM.
I know a google engineer who was saying they’re having to update their code bases to handle > 16 exabytes of storage, if you can imagine. But yeah, that’s storage, not RAM.
I watched a documentary on this awhile back. The municipality asked the public if it would be enough for them to dump treated water into a lake and then draw from that lake? And then someone with expertise in the matter commented that this would necessitate another treatment phase, since any wild animal could take a dump in the lake. So he seemed to think closing the loop made the most sense from a practical standpoint.
I think people here tend to question and fact-check posts and comments a lot, which is a healthy thing. Now some say reality skews left, in which case could it be that the right have left because the left is right?
Most recently? My wife was wfh and out of the kindness of my heart I brought her a coke. She was on a zoom call with her entire team. I was pantsless.
literally good for you
I actually asked my family doctor at one point about the health effects of masturbation. She said that as a guy, if you are not otherwise sexually active, it’s good for the prostate to keep the plumbing working down there.
I started in C and switch to C++. It’s easy to think that the latter sort of picked up where the former left off, and that since the advent of C++11, it’s unfathomably further ahead. But C continues to develop and occasionally gets some new feature of its own. One example I can think of is the restrict
key word that allows for certain optimizations. Afaik it’s not included in the C++ standard to date, though most compilers support it some non-standard way because of its usefulness. (With Rust, the language design itself obviates the need for such a key word, which is pretty cool.)
Another feature added to C was the ability to initialize a struct
with something like FooBar fb = {.foo=1, .bar=2};
. I’ve seen modern C code that gives you something close to key word args like in Python using structs. As of C++20, they sort of added this but with the restriction that the named fields have to come in the same order as they were originally defined in the struct, which is a bit annoying.
Over all though, C++ is way ahead of C in almost every respect.
If you want to see something really trippy, though, have a look at all the crazy stuff that’s happened to FORTRAN. Yes, it’s still around and had a major revision in 2018.
So the next captcha will be a list of AI-generated statements and you have to decide which are bat shit crazy?
“Recall uses Copilot+ PC advanced processing capabilities to take images of your active screen every few seconds,”
Seems like a lot of extra disk thrashing that would shorten the life expectancy of an SSD? Like it would be considerably more than your usual background chatter of daemons writing to log files and what not. Unless I’m misunderstanding this?
I think I could get very nervous coding for the military, depending on what sort of application I was working on. If it were some sort of administrative database, that doesn’t sound so bad. If it were a missile guidance system, on man! A single bug and there goes a village full of civilians. Even something without direct human casualties could be nerve-wracking. Like if it were your code which bricked a billion-dollar military satellite.
Speaking of missile guidance systems, I once met someone who worked a stint for a military contractor. He told me a story about a junior dev who discovered an egregious memory leak in a cruise missile’s software. The senior dev then told him “Yeah, I know about that one. But the memory leak would take an hour before it brings the system down and the missile’s maximum flight time is less than that, so no problem!” I think coding like that would just drive me into some OCD hell.
I have only written potentially life-threatening code once in my life. It had to do with voltage/current regulation in the firmware of a high-powered instrument used by field workers at the company where I work. It was a white-knuckled week I spent on just a single page of code, checking and re-checking it countless times and unit testing it in every conceivable way I could imagine.
Several years ago, I went under the knife and the whole day from the point they put me under is a total blank. It’s unsettling because I am told I carried on conversations with the doctor, family members, etc. after initially coming to from anaesthesia, but it’s only starting the following morning when I woke up in a regular hospital bed that I could start remembering again.
I totally get where people are going with eliminating dictators and what not, but knowing myself as well as I do… yeah, you’d probably find me down at the Chinese buffet.
Fair enough. I’m just looking for some independent confirmation as this is pretty big news.
Is this official though, or wishful thinking on the part of Cameron?
I remember ads claiming it was cutting edge nanotechnology! And I thought oh cool, you mean like there are tiny robots running around in the shampoo? But no, it was microplastics.
This happens naturally in the form of meteors streaking through the sky. Each one of those is adding a tiny amount of mass to the planet.
But you’ve got me wondering about something now. When a large asteroid hits the planet, it obviously adds its own mass, but it also kicks up a lot of debris into space. Some percentage of that will reach orbital escape velocity and never come back. But I honestly don’t know if there is a net mass increase or decrease after such an event? We’re generally concerned about other more pressing matters in such a scenario!
From what other posters are saying, it may be the other way around? That is, most mammals cannot see green, so it doesn’t matter from a camouflage perspective among mammals. Humans (and primates in general) are an outlier in this repect.
Bird of prey can, though, so there’s that.
Wow, that is fascinating!
Makes me wonder about the other direction, going into the near infrared as opposed to UV. I remember from a class in remote sensing that many plants are actually most reflective in that band (more so than in green, even). NIR air photos are often used by biologists to get an indication of the health of a forest. But I have no idea whether animals also reflect NIR? It may be that most animals cannot see in that band in the first place, so it would not offer any camouflage advantage.
Great read! That explains a lot.
I’ve been deep diving a bit myself and found this article that explains another thing that’s puzzled me over the years. Some birds have crazy vibrant coloration that almost glistens, like peacock feathers. Outside of the zoo, I’ve noticed it a bit in common grackles. They look black on first glance, but when you study them closely, they have this kind of purple sheen around their heads. Apparently, it’s still melanin at work here, but it’s structured in a very special way.
You can always combine integer operations in smaller chunks to simulate something that’s too big to fit in a register. Python even does this transparently for you, so your integers can be as big as you want.
The fundamental problem that led to requiring 64-bit was when we needed to start addressing more than 4 GB of RAM. It’s kind of similar to the problem of the Internet, where 4 billion unique IP addresses falls rather short of what we need. IPv6 has a host of improvements, but the massively improved address space is what gets talked about the most since that’s what is desperately needed.
Going back to RAM though, it’s sort of interesting that at the lowest levels of accessing memory, it is done in chunks that are larger than 8 bits, and that’s been the case for a long time now. CPUs have to provide the illusion that an 8-bit byte is the smallest addressible unit of memory since software would break badly were this not the case, but it’s somewhat amusing to me that we still shouldn’t really need more than 32 bits to address RAM at the lowest levels even with the 16 GB I have in my laptop right now. I’ve worked with 32-bit microcontrollers where the byte size is > 8 bits, and yeah, you can have plenty of addressible memory in there if you wanted.