Poik

joined 1 year ago
[–] [email protected] 30 points 2 months ago

Not necessarily the case, but if it's affecting your life so strongly, you might want to get checked by a medical professional.

Long COVID can destroy your life. Depression can destroy your life. Iron deficiency can ruin your life. A lot of things you might just think is just being tired may actually have a cause. Especially if simple fixes like "touch grass" style clichés do nothing for you.

It's not always the answer, but it's good to rule out in that case.

[–] [email protected] 5 points 2 months ago

I was told in 2009 "Why optimize? Hardware upgrades will make your efforts obsolete anyway." So... I devoted my time to optimization, because fuck that. I ended up doing algorithm optimization in my first full time job, and loved... That part of the job at least.

Indie games and co-op games are my jam. I feel for all of this comment.

[–] [email protected] 7 points 2 months ago

He got free food and a bed? Jealous.

[–] [email protected] 5 points 2 months ago

The master branch in git isn't the same though. It's closer related to the word "remaster." Master used to mean the original document is still used everywhere in tech and outside of it.

Main makes more sense since a master copy should be something that doesn't change in my opinion. But that's semantics...

[–] [email protected] 2 points 3 months ago (1 children)

After you hit full time student, the rest of the classes are free, so I filled my schedule way too full. All my favorite teachers were in the math department too. No regrets. That and I ended up using my math skills when I switched to machine learning engineering.

That must have been cathartic as hell.

[–] [email protected] 17 points 3 months ago (7 children)

As someone with a degree in math and a degree in engineering... One of those degrees got me a job.

That being said, that's the way engineers look at managers, as generally they want to build something that works and is safe, but all managers care about is getting it done quickly and under budget, which means any micromanaging gets pushed down to the technician to have to deal with... And trying to argue gets you fired.

[–] [email protected] 2 points 3 months ago

Our cats use it to beg for treats. Very rarely do I see them on it and not meowing for attention.

[–] [email protected] 1 points 3 months ago

It's valid, and it sucks. If you can even do $5, it's worth it. But the world is absolutely against you right now. A lot of older folk don't quite get how bad it's gotten.

However, saving a dollar today is worth more than saving two dollars ten years from now. And having an emergency fund might actually save your life.

Hopefully something happens to shake up housing. These prices are absolutely criminal.

[–] [email protected] 8 points 4 months ago (2 children)

That's LLM bull. The model already knows hangman; it's in the training data. It can introduce variations on the data, especially in response to your stimuli, but it doesn't reinvent that way. If you want to see how it can go astray ask it about stuff you know very well, and watch how it's responses devolve. Better yet, gaslight it. It's very easy to convince LLMs that they're wrong because they're usually trained for yes-manning and non confrontation.

Now don't get me wrong, LLMs are wicked neat, but they don't come up with new ideas, but they can be pushed towards new concepts, even when they don't grasp them. They're really good at sounding sure of themselves, and can easily get people to "learn" new "facts" from them, even when completely wrong. Always look up their sources, (which Bard (Google's) can natively get for you in its UI) but enjoy their new ideas for the sake of inspiration. They're neat toys, which can be used to provide natural language interfaces to expert systems. They aren't expert systems.

But also, and more importantly, that's not zero-shot learning. Neat little anecdote from a conversation with them though. Which model are you using?

[–] [email protected] 3 points 4 months ago

No. AI and, what you're more likely to be referring to, machine learning has had applications for decades. Basic work was used back into the '60s, mostly for quick things, and 1D data analysis was useful long before images (voice and stuff like biometrics). But there are many more types of AI. Bayesian networks (still in the learned category) were huge breakthroughs and still see a lot of use today. Decision trees, Markov chains, and first order logic are the most common video games AI and usually rely on expert tuning rather than learned results.

AI is a huge field that's been around longer than you expected, and permeates a lot of tech. Image stuff is just the hot application since it's deep learning based buff that started around 2009 with a bunch of papers that helped get actual beneficial learning in deeper models (I always thought it started roughly with Deep Boltzmann Machines, but there's a lot of work in that era that chipped away at the problem). The real revolution was general purpose GPU programming getting to a state where these breakthroughs weren't just theoretical.

Before that, we already used a lot of computer vision, and other techniques, learned and unlearned, for a lot of applications. Most of them would probably bore you, but there are a lot of safety critical anomaly detectors.

[–] [email protected] 7 points 4 months ago (4 children)

This actually is a symptom from the sort of "beneficial" overfit in Deep Learning. As someone whose research is in low data, long tails, and few shot learning, there's a few things that smaller networks did better in generalization, and one thing they particularly did better (without explicit training for it) is gauging uncertainty. This uncertainty is sometimes referred to as calibration. Calibrating deep networks can yield decent probabilities that can be used to show uncertainty.

There are other tricks for this. My favorite strategies prep the network for learning new things. Large margin training and the like are a good thing to look into. Having space in the output semantic space (the layer immediately before the output or earlier for encoder decoder style networks) allows for larger regions for distinct unknown values to be separated from the known ones, which helps inherently calibrate the network.

[–] [email protected] 2 points 4 months ago

Which end? The main story is just a narrative device, in fact you shouldn't really obey the narrator at all. Calling any end "The End" doesn't make sense in the context of the game, really. Unless you just broke out of the mind control facility three times then called it quits? That end is supposed to be non enticing so that you try literally anything else before putting it down. I think the going insane end sticks with me the most. Although the game dev commentary in the recent release is fun.

view more: ‹ prev next ›