Yeah, this is actually a pretty great application for AI. It’s local, privacy-preserving and genuinely useful for an underserved demographic.
One of the most wholesome and actually useful applications for LLMs/CLIP that I’ve seen.
Yeah, this is actually a pretty great application for AI. It’s local, privacy-preserving and genuinely useful for an underserved demographic.
One of the most wholesome and actually useful applications for LLMs/CLIP that I’ve seen.
It would be better to have this as a FUSE filesystem though - you mount it on an empty directory, point the tool at your unorganised data and let it run its indexing and LLM categorisation/labelling, and your files are resurfaced under the mountpoint without any potentially damaging changes to the original data.
The other option would be just generating a bunch of symlinks, but I personally feel a FUSE implementation would be cleaner.
It’s pretty clear that actually renaming the original files based on the output of an LLM is a bad idea though.
(6.9-4.2)/(2024-2018) = 0.45 “version increments” per year.
4.2/(2018-1991) = 0.15 “version increments” per year.
So, the pace of version increases in the past 6 years has been around triple the average from the previous 27 years, since Linux’ first release.
I guess I can see why 6.9 would seem pretty dramatic for long-time Linux users.
I wonder whether development has actually accelerated, or if this is just a change in the approach to the release/versioning process.
At least in some circumstances, the risks of sharing your DNA include having children…
Tbf 500ms latency on - IIRC - a loopback network connection in a test environment is a lot. It’s not hugely surprisingly that a curious engineer dug into that.
Zsh is a nice balance of modern features and backwards compatibility with bash.
Crostini is an official feature built by Google that allows you to run Linux on a tightly integrated hypervisor inside Chrome OS. You keep a lot of Chrome OS’ security benefits while getting a Linux machine to play with.
That said, no, it’s not illegal to install a different operating system on your Chromebook hardware. They are just PCs, under the hood. You might lose some hardware security features though, e.g. the capabilities provided by integration of the Titan silicon.
If you had a job at Google, corporate IT would definitely not be happy if you wiped the company-managed OS and installed an unmanaged Linux distro :)
The reddest of red flags.
Open source vulnerabilities typically stem from poorly written code
Yeah, because paid programmers never write bad closed-source code…
Tl;dr: TPMs are very unlikely to make your privacy better or worse, but they could definitely be abused by a company like MS to make end users’ experiences worse. They could also be used for significant security and privacy gains… they’re a tool.
The TPM can be used to provide a cryptographic binding between aspects of your system’s configuration and a unique key which is resident within the TPM (a process called “attestation”). It can also generate secondary keys that are associated with the base key, and use those to do cryptographic operations like encryption/decryption and authentication.
Telemetry wise, the TPM’s only utility might be to “prove” that the data sent from your PC wasn’t tampered with. That said, I don’t think MS is actually doing that, and they don’t need to in order to be incredibly invasive in their telemetry.
The (imo) worst way in which a TPM might be abused in a user-hostile sense is to detect if the OS has been modified by the user, or if an installation isn’t legitimate, etc. That could be used to disable certain features if you try to install unauthorised software, dual boot Linux or whatever. This would be similar to the smartphones of today, which can for example disable access to banking apps if jailbroken/rooted.
TPMs (>2.0 at least) otherwise have the potential to realise a significant improvement in security and privacy for users, if used correctly. They can be used for encryption and credentials that are bound in hardware and therefore practically impossible to steal. And can detect hardware tampering and potentially foil Evil Maid attacks. Imagine if your login sessions for various websites were bound to your hardware, such that a dodgy extension could never steal your cookies.
I found it much more barebones in my tinkering. It doesn’t seem to support pulling via SSH (and definitely doesn’t support signing commits). Configuration options appear extremely limited, both in documentation and the UI.
It looks nice, but I don’t really see the point to it when Gitea Actions is now a thing. Gitea is a more mature product, and is similarly fast and lightweight.
This is why self hosted to me means actually running it on my own hardware in a location I have at least some control of physical access.
That said, an ISP could perform the same attack on a server hosted in your home using the HTTP-01 ACME challenge, so really no one is safe.
HSTS+certificate pinning, and monitoring new certificates issued for your domains using Certificate Transparency (crt.sh can be used to view these logs) is probably the only way to catch this kind of thing.
Of course they do, but it isn’t the ISP’s job to do so. I believe that is the point that the EFF is making here.
Censorship sometimes needs to happen to protect people, but it should be conducted by website owners/platforms and government authorities – on each end of the information transaction, not in transit by an ISP.
Njalla is mine. I like the privacy protections they offer.
Are CloudFlare, Amazon or Microsoft any better? Google at least take security (if not privacy) very seriously.
In general it seems bad to have any huge profit-driven organisation exercising significant control over open standards, but I do think that Google is lesser than many of the other evils.
Looks like a very cool project, thanks for building it and sharing!
Based on the formula you mentioned here, it sounds like an instance with one user who has posted at least one comment will have a maximum score of 1. Presumably the threshold would usually be set to greater than 1, to catch instances with lots of accounts that have never commented.
This has given me another thought though: could spammers not just create one instance per spam account? If you own something like blah.xyz, you could in theory create ephemeral spam instances on subdomains and blast content out using those (e.g. spamuser@esgdf.blah.xyz, spamuser@ttraf.blah.xyz, etc.)
Spam management on the Fediverse is sure to become an interesting issue. I wonder how practical the instance blocking approach will be - I think eventually we’ll need some kind of portable “user trustedness” score.
Maybe I’m being stupid, but how does this service actually determine suspicious-ness of instances?
If I self-host an instance, what are my chances of getting listed on here and then unilaterally blocked simply because I have a low active user count or something?
Power management is going to be a huge emerging issue with the deployment of transformer model inference to the edge.
I foresee some backpedaling from this idea that “one model can do everything”. LLMs have their place, but sometimes a good old LSTM or CNN is a better choice.