Godot does have a special thing for mesh instancing, I think variations were possible as well like different colored triangles maybe? https://docs.godotengine.org/en/stable/tutorials/performance/vertex_animation/animating_thousands_of_fish.html
Godot does have a special thing for mesh instancing, I think variations were possible as well like different colored triangles maybe? https://docs.godotengine.org/en/stable/tutorials/performance/vertex_animation/animating_thousands_of_fish.html
The way I understand the users didn’t necessarily realize McAfee is responsible, just that a bunch of sqlite files appeared in temp so they might not connect the dots here anyway. Or even know McAfee is installed considering their shady practices.
I do think we’re machines, I said so previously, I don’t think there is much more to it than physical attributes, but those attributes let us have this discussion. Remarkable in its own right, I don’t see why it needs to be more, but again, all personal opinion.
I read this question a couple times, initially assuming bad faith, even considered ignoring it. The ability to change, would be my answer. I don’t know what you actually mean.
Personally my threshold for intelligence versus consciousness is determinism(not in the physics sense… That’s a whole other kettle of fish). Id consider all “thinking things” as machines, but if a machine responds to input in always the same way, then it is non-sentient, where if it incurs an irreversible change on receiving any input that can affect it’s future responses, then it has potential for sentience. LLMs can do continuous learning for sure which may give the impression of sentience(whispers which we are longing to find and want to believe, as you say), but the actual machine you interact with is frozen, hence it is purely an artifact of sentience. I consider books and other works in the same category.
I’m still working on this definition, again just a personal viewpoint.
Not sure if this is the right answer I’m not familiar with that ecosystem: They have comparisons on their site
Its a thing. https://en.m.wikipedia.org/wiki/Busy_waiting
Not an answer to the question, but in case performance is the goal, Torchaudio has it here
Ah, even then it could just be a consequence of training samples usually being chronological(most often the expected resolution for conflicting instructions is “whatever you heard last”, with some exceptions when explicitly stated) so it learns to think that way. I did find the pattern also applies to GPT trained on long articles where you’d expect it not to, so wanted to just explain why that might be.
Or I should explain better: most training samples will be cut off at the top, so the network sort of learns to ignore it a bit.
Yes, that’s by design, the networks work on transcripts per input, it does genuinely get cut off eventually, usually it purges an entire older line when the tokens exceed a limit.
I was a curious child, and things spiralled out of control from there…
Anybody have a link to the paper? The article strikes me as a used car salemans trying to sell me a journal. Mostly what I’m getting is new reinforcement learning technique catered to language? But what model architecture? Is it new? I’d like to know
Maybe a bit late, but I’ve worked on this kind of functionality. I did not work on the algorithm, but the guys who did say:
This is of course based on trust, but these are the claims.
Ah, that makes sense. Most cloud providers have the full nine yards with online hardware provisioning and imaging I forgot you could still just rent a real machine.
Hmm, wonder if there was some reason they didnt just extract the original certificates from the VPS if it was actually the hosting provider, I mean even with mitigation it should be sitting in a temp folder somewhere, surely they could? Issuing new ones seems like a surefire way to alert the operators, unless they already used Let’s Encrypt of course.
They previously did not use APEX but that seems to have changed recently: https://github.com/GrapheneOS/grapheneos.org/commit/7bf9b2671667828d1553c92bf4f64cc749b74d0b Regardless it will need the verified boot keys it seems so Google can’t update them, likely the devs will take responsibility to update the CAs. No idea if they will restore the user control though.
I feel like this is just describing the future of business processing consultants. Like there’s already a role for this, unless I’m missing something?
It would be luck based for pure LLMs, but now I wonder if the models that can use Python notebooks might be able to code a script to count it. Like its actually possible for an AI to get this answer consistently correct these days.