TLS doesn’t encrypt the host name of the urls you are visiting and DNS traffic is insanely easy to sniff even if you aren’t using your ISPs service.
TLS doesn’t encrypt the host name of the urls you are visiting and DNS traffic is insanely easy to sniff even if you aren’t using your ISPs service.
babe wake up, a new bone-apple-tea just dropped
the hostname of a website is explicitly not encrypted when using TLS. the Encrypted Client Hello extension fixes this but requires DNS over HTTPS and is still relatively new.
just a guess, but in order for an LLM to generate or draw anything it needs source material in the form of training data. For copyrighted characters this would mean OpenAI would be willingly feeding their LLM copyrighted images which would likely open them up to legal action.
open source software getting backdoored by nefarious committers is not an indictment on closed source software in any way. this was discovered by a microsoft employee due to its effect on cpu usage and its introduction of faults in valgrind, neither of which required the source to discover.
the only thing this proves is that you should never fully trust any external dependencies.
yeah silly me for supporting artists with my money but also downloading drm-free copies of things so I can actually exercise a semblance of ownership. but sure, keelhaul me so you can keep your sense of smug superiority.
AI is a tool that is fundamentally based on the concept of theft and plagiarism. The LLM training data comes from artists and creators that did not consent to their work being plagiarized by a hallucinating machine.
a decentralized community that correctly prioritizes security would absolutely be using signed commits and other web-of-trust security practices to prevent this sort of problem
you understand there’s more than one way to have an economy right? that there’s more than one way for labor to be rewarded for its output?
saying “our economic system needs to end” has nothing to do with what you wrote
faster can still lead to battery life improvements. if the CPU is able to complete tasks in less time, it can then enter a lower power state sooner which will result in less battery usage overall
assuming you have a GNU toolchain you can use the find
command like so:
find . -type f -executable -exec sh -c '
case $( file "$1" ) in (*Bourne-Again*) exit 0; esac
exit 1' sh {} \; -print0 | xargs -0 -I{} cp {} target/
This first finds all executable files in the current directory (change the “.” arg in find to search other dirs), uses the file
command to test if it’s a bash file, and if it is, pipes the file name to xargs
which calls cp
on each file.
note: if “target” is inside the search directory you’ll get output from cp
that it skipped copying identical files. this is because find
will find them a free you copy them so be careful!
note 2: this doesn’t preserve the directory structure of the files, so if your scripts are nested and might have duplicate names, you’ll get errors.
why use docker here? you’re just adding layers of abstraction in an environment that can’t seem to really support them.
that said, switching to 32bit linux, if the VPS supports it, will save you memory.
I realize I’m late to this thread, but if you’re serious about archiving a VHS in the best manner possible, you have to go the RF capture route: https://github.com/oyvindln/vhs-decode
This method effectively captures the “raw” signal stored on the tape, allowing you to convert it after you’ve captured it however you see fit. You don’t have to worry about cheap digitizers/capture cards/etc distorting the signal.
you’re so close, just why exactly do you think people are using it for these things it’s not meant for?
because every company, every CEO, every VP, is pushing every sector of their companies to adopt AI no matter what.
most actual people understand the limitations you list, but it’s the capitalists at the table that are making AI show up where it’s not wanted