Yeah they’ve put them in a couple places, It’s pretty bad. Had to work out how to create a custom uBlock Origin rule to block them.
Yeah they’ve put them in a couple places, It’s pretty bad. Had to work out how to create a custom uBlock Origin rule to block them.
I think calling it just like a database of likely responses is too much of a simplification and downplays what it is capable of.
I also don’t really see why the way it works is relevant to it being “smart” or not. It depends how you define “smart”, but I don’t see any proof of the assumptions people seem to make about the limitations of what an LLM could be capable of (with a larger model, better dataset, better training, etc).
I’m definitely not saying I can tell what LLMs could be capable of, but I think saying “people think ChatGPT is smart but it actually isn’t because <simplification of what an LLM is>” is missing a vital step to make it a valid logical argument.
The argument is relying on incorrect intuition people have. Before seeing ChatGPT I reckon if you’d told people how an LLM worked they wouldn’t have expected it to be able to do things it can do (for example if you ask it to write a rhyming poem about a niche subject it wouldn’t have a comparable poem about in its dataset).
A better argument would be to pick something that LLMs can’t currently do that it should be able to do if it’s “smart”, and explain the inherent limitation of an LLM which prevents it from doing that. This isn’t something I’ve really seen, I guess because it’s not easy to do. The closest I’ve seen is an explanation of why LLMs are bad at e.g. maths (like adding large numbers), but I’ve still not seen anything to convince me that this is an inherent limitation of LLMs.
Thanks for the info on crossposting! I thought I’d seen someone mention a cross posting feature but couldn’t see any button to do it. I’m using the Jerboa app on Android which I guess doesn’t have that button, but I see it on the website now as you say.
It’s also good to know that linking to the original URL is generally better and the rest can be handled by the UI - that does seem nicer.
I’d be happy if we’d just accepted “referer” as the correct spelling for everything, but instead we have the “Referrer-Policy” header, so now I need to check the correct spelling for anything involving referring…
I do sort of like the idea that because we want to keep backwards compatibility on software we just change the language instead since that’s easier.
What sort of features 🤔
I use VSCode with config options to disable telemetry. Probably not perfect but good enough for me, I’m very happy using VSCode
When it happens docker+wsl become completely unresponsive anyway though. Stopping containers fails, after closing docker desktop wsl.exe --shutdown
still doesn’t work, only thing I’ve managed to stop the CPU usage is killing a bunch of things through task manager. (IIRC I tried setting a cap while trying the hyper-v backend to see if it was a wsl specific problem, but it didn’t help, can’t fully remember though).
This is the issue that I think was closest to what I was seeing https://github.com/docker/for-win/issues/12968
My workaround has been to start using GitHub codespaces for most dev stuff, it’s worked quite nicely for the things I’m working on at the moment.
My experience using docker on windows has been pretty awful, it would randomly become completely unresponsive, sometimes taking 100% CPU in the process. Couldn’t stop it without restarting my computer. Tried reinstalling and various things, still no help. Only found a GitHub issue with hundreds of comments but no working workarounds/solutions.
When it does work it still manages to feel… fragile, although maybe that’s just because of my experience with it breaking.
9 minutes 34 of audio bible anyone?? https://youtube.com/watch?v=mdEttg1lN0I
deleted by creator
Assuming x and y are totally ordered 🤮
(for anyone wondering, a monad is just a monoid in the category of endofunctors)
Ah, that’s too boring. I have a range of responses to pick from to keep things interesting:
For me, no text means “I haven’t really reviewed this properly so don’t want to write anything that could be used against me if (when?) this breaks something in prod”
Agreed, and the questions I have that MDN doesn’t answer would probably be ones even less likely for the AI explain to get right.
Confusable characters get a little yellow box which is different from the squiggly underlines most linters and stuff use which at least makes it a bit more recogniseable.
Personally I can’t stand having underlines all over my code, so I’ll usually just “fix” the non-issue if possible, or otherwise just disable whatever the warning is entirely.
To be fair, it’s no worse than articles some people write on those nonsense websites.
It probably really depends on the project, though I’d probably try and start with the tests that are easiest/nicest to write and those which will be most useful. Look for complex logic that is also quite self-contained.
That will probably help to convince others of the value of tests if they aren’t onboard already.