Seems quite absurd that they would shut down the only system that could tell journalists what was actually happening in the criminal courts under the pretext that they sent information to a third-party AI company (who doesn’t these days). Here’s a rebuttal by one of the founders i believe: https://endaleahy.substack.com/p/what-the-minister-said
Absolutely fucking crazy that you typed this out as a legitimate defense of allowing extremely sensitive personal information to be scraped.
> only system that could tell journalists what was actually happening in the criminal courts
Who cares? Journalism is a dead profession and the people who have inherited the title only care about how they can mislead the public in order to maximize profit to themselves. Famously, "journalists" drove a world-renowned musician to his death by overdose with their self-interest-motivated coverage of his trial[1]. It seems to me that cutting the media circus out of criminal trials would actually be massively beneficial to society, not detrimental.
Absolutely fucking crazy that you call accurately describing the reality of AI scraping "absolutely fucking crazy" while at the same time going "who cares?" on attacks against journalism and free speech.
>Oh no, some musician died, PASS THE NATIONAL SECURITY ACT, LOCK DOWN ALL INFORMATION ABOUT CRIMINALS, JAIL JOURNALISTS!!!!
"Free speech" is some kind of terminal brain worm that begs itself to be invoked to browbeat most anybody into submission, I suppose. Now we're apparently extending "free speech" to mean "the government must publish sensitive private information about every citizen in an easily-scraped manner". Well, I don't buy your cheap rhetorical trick. I support freedom of speech in exactly the scope it was originally intended, that is, the freedom to express ideas without facing government censorship or retaliation. Trying to associate your completely unrelated argument with something that everybody is expected to agree with by default is weak.
Nor is it "an attack on journalism". If a real journalist still exists in the current year, they can do investigative work to obtain information that is relevant to the public's interests and then publish it freely. Nobody is stopping them.
> LOCK DOWN ALL INFORMATION ABOUT CRIMINALS
Notably, people who aren't criminals may find themselves in court to determine whether or not they aren't. Unfortunately, people like yourself and the entirety of the broadcast and print media then go on to presume every person who goes to court are criminals and do everything in their power to ruin their lives. Far from just "some musician", most people who are arrested on serious charges get a black mark on their record that effectively destroys their careers and denies them the ability to rent property outside of a ghetto, as employers and landlords discriminate against them baselessly even after they are acquitted of all charges against them.
next time you should read what the thread is about before commenting, this is not about "wanting the government to publish X", it's about the government ordering a private company to delete, from their own servers, data that the government had published
it's a clear cut free speech issue, you just don't want to admit it since certain data being available and spread to other people doesn't suit your political ideology
When there is a risk of feeding sensitive data to the AI giants the first reaction should be to pull the plug. I'm impressed the government acted quickly and decisively for once. Maybe the company involved will think twice before entering an agreement with an AI company. Notice in the whole rant it is never mentioned which AI giant they were feeding.
The government provided data to a private company. The private company sold resold access to a third party for AI ingestion. it's a plain case of tough titties to the private company.
That said I don't know why the hell the service concerned isn't provided by the government itself.
Perhaps that is true, but the response linked by GP claims exactly the opposite:
"We hired a specialist firm to build, in a secure sandbox, a safety tool for journalists. They are experts in building privacy-preserving AI solutions - for people like law firms or anyone deeply concerned with how data is held, processed, and protected. That’s why we chose them. Their founders are not only respected academics in addition to being professionals, they have passed government security clearance and DBS checks in the past, and have worked on data systems for the National Archives, the Treasury, and other public agencies. They’ve published academic papers on data compliance for machine learning.
"The Minister says we ‘shared data with an AI company”... as if we were pouring this critically sensitive information into OpenAI or some evil aggregator of data. This is simply ridiculous when you look at what we do and how we did it.
"We didn’t “share” data with them. We hired them as our technical contractor to build a secure sandbox to test an idea, like any company using a cloud provider or an email service. They worked under a formal sub-processor agreement, which means under data protection law they’re not even classified as a “third party.” That’s not our interpretation. It’s the legal definition in the UK GDPR itself. ...
"And “for commercial purposes”? The opposite is true. We paid them £45,000 a year. They didn’t pay us a penny. The money flowed from us to them. They were prohibited, in writing, from selling, sharing, licensing, or doing anything at all with the data other than providing the service we hired them for.. and they operated under our supervision at all times. They didn’t care what was in the data - we reviewed, with journalists, the outputs to make sure it worked."
If this is true, it does seem that the government has mischaracterized what happened.
It sounds very reasonable. But it's also directly contradicted by the government information about this case, which was very specific even about the number of breaches:
> Our understanding is that some 700 individual cases, at least, were shared with the AI company. We have sought to understand what more may have been shared and who else may have been put at risk, but the mere fact that the agreement was breached in that way is incredibly serious.
> ... the original agreement that was reached between Courtsdesk and the previous Government made it clear that there should not be further sharing of the data with additional parties. It is one thing to share the data with accredited journalists who are subject to their own codes and who are expected to adhere to reporting restrictions, but Courtsdesk breached that agreement by sharing the information with an AI company.
But when the conspiracy involves lack of prosecution or inconsistent sentencing at scale and then the Ministry of Justice issues a blanket order to delete one of the best resources to look into those claims...? Significantly increases the legitimacy of the claims.
I assumed it was the usual conspiracy stuff up until this order.
This seems like an interesting approach and automating visitor qualification is definitely useful. But quick question though, once you drop in that “single line” of code, how soon can a team expect to see their first qualified leads?
You’ll see results as soon as your first qualified visitors arrive. Our agent evaluates every unique visitor, researches and qualifies B2B prospects based on your custom criteria, and starts showing you qualified leads in real time.
From personal experience (London based), I’ve seen a least 2 meta x rayban shops in high traffic areas, which never seem to be empty. Adding to that, recently on social media someone expressed a similar sentiment until they saw what the glasses themselves look like, then realized they looked like ordinary glasses and are undistinguishable if the recording light is covered by black masking tape
Unfortunately I’ve seen some use it for rage baiting purposes then catching peoples reaction without the knowledge that they are being filmed, but on the bright side I guess it’s useful for travelling without looking like an obnoxious tourist/influencer with their phones out constantly
Good post, bit too “mathy” but makes me think of “Asynchronous computing @Facebook: Driving efficiency and developer productivity at Facebook scale”. Where they touch on capacity optimization (queuing + time shifting), capacity regulation along with user delay tolerance (bc not all jobs, even at the same priority level, are equal)
I think the issue with the math is that it doesn't read well.
For example, the paragraphs around the paragraph with "compute the exact Poisson tail (or use a Chernoff bound)" and that paragraph itself could be better illustrated with lines of math instead of mostly language.
I think you do need some math if you want to approach this probabilistically, but I agree that might not be the most accessible approach, and a hard threshold calculation is more accessible and maybe just as good.
For something like this, annotated graphs and examples (IMO) work a lot better than formulas in explaining the problem and solution.
Particularly because distributed computer systems aren't pure math problems to be solved. Load often comes from usage which is often closer to random inputs rather than predicable variables. Further, how load is processed depends on a bunch of things from the OS scheduler to the current load on the network.
It can be hard to really intuitively understand that a bottlenecked system processes the same load slower than an unbound system.
> Asynchronous computing @Facebook: Driving efficiency and developer productivity ... optimization (queuing + time shifting), capacity regulation along with user delay tolerance ...
We can infer a more detailed priority by understanding how long each of these asynchronous requests can be delayed ... For each job to be executed, we try to execute it as close as possible to its delay tolerance.
... we defer jobs with a long delay tolerance so that the workload is spread over a longer time window. Queueing plays an important role in selecting the most urgent job to execute first.
... Time shifting ... optimize capacity in Async:
1. Predictive compute collects the data people used yesterday. Predicated on which data people may need, it precomputes before peak hours and stores the data in cache ... This moves the computing lift from peak hours to off-peak hours and trades a little cache miss for better efficiency.
2. Deferred compute ... schedules a job as part of user request handling but runs at much later time. For instance, the "people you may know" list is processed during off-peak hours, then shown when people are online (generally during peak hours).
I feel tha I'm missing something obvious. Isn't this doc reinventing the wheel in terms of what very basic task queue systems do? It describes task queues and task prioritization, and how it supports tasks that cache user data. What am I missing?
Yes, what is described in the blog post is (by now) a fairly established way to deal with task queues at scale (if you're familiar with those kinds of systems, but not everyone is).
Really great introductory article on color space in general, I really appreciate that they touched on perceptual uniformity and how we all perceive colors differently. It’s great to find out that applications like Oklab came out recently to fix this by manipulating the distances between colors to try compensate for human perception while also being more straightforward to calculate so that it can be used in real-time applications. Also the UI of this blog post was so aesthetically pleasing, that it was worth burning my retinas with the light mode
Not that surprising that things have gotten to this point, just a few days ago schools in Florida were testing a new drone defense system against shootings. Between see-through backpacks, armed teachers, metal detectors, and other things you’d think it would be more easier to severely restrict firearm access to under 21 year olds and make the parents criminally liable if they are found to have facilitated access in any way in the wake of a shooting. But yeah i guess dipping into online conversations and immediately notifying both school officials and law enforcement is a good solution (/j)
‘Berény is likely to have painted Sleeping Lady with Black Vase from 1927 to 1928, by which time he had returned to Budapest, in his native Hungary, from Berlin where he had fled following World War I.The painting was last seen in public in 1928 when it exhibited at the Ernst Museum before being sold.The buyer is likely to have been Jewish and left the country in the lead up to or during World War II. Hungary was turbulent during this period and the painting was considered lost’. Sad to see a painting with such history ending up in a private collection, also considering that its some of Bereny’s best work too
The quantity & quality of a child's diet determines how fast they age & their health later in life, its just crazy that US children now get 2/3rds of their energy from ultra-processed foods that are designed to be hyper-palatable & maintain hunger. Unfortunately, there’s just a lot of misinformation out there as it relates to diet (food pyramid being one of the most egregious examples). For one, although most ultra processed foods are bad, some can be actually beneficial for those with medical issues.
‘Considering the pool of available IPv4 addresses has been exhausted for quite a while now, and was running out for public use years ago’ I thought it was logical that most systems that have adopted IPv6. Crazy to think that it turns out it wasn’t, but shout out to apple and their stringent dev requirements bc they require support IPv6-only networks.
The pool is not exhausted, the IPv6 cabal realized after 20 years that actively fighting against freeing up more space was the only way they could drive adoption.
If we just stop listing 240/4 as a bogon we could be allocating from it in a few years.
This seems great! Would love to join however I can only seem to find the 2008 and 2012 pdf of The Art of Multiprocessor Programming for free, is there a link for the 2020 version?