Hacker Newsnew | past | comments | ask | show | jobs | submit | nialse's commentslogin

If there is no scientific evidence for the method then it’s up to if you believe. Placebo is a hell of a drug.

That is the logical conclusion. The era of personal computers is coming to an end. It had a good run though.

Luckily Beelinks are still cheap, work decently, and can run Linux/Windows, so if all someone needs is to browse the Internet and do basic stuff, honestly? They’re fine. We’ll see how long that lasts though.

Versioning the cache is a poor band-aid and introduces failure modes not considered.

Maybe personal computing is entering a slow decline.

Rising RAM prices over the next couple of years will make new computers and phones harder to justify, so people will keep devices longer. At the same time, Microsoft and Apple (et alia) continue shipping more demanding software packed with features no users asked for. Software growth has long driven hardware upgrades, but if upgrades no longer feel worthwhile, the feedback loop breaks. The question is whether personal computing keeps its central role, or quietly becomes legacy infrastructure people replace only when forced? In that case: What is the next era?


It seems to me that what your saying isn't the personal computing is entering a slow decline but the PC market is. If people continue to use PCs they already own then personal computing is alive and well.

You cannot pick and choose one or two variables and then claim representativeness based on a numerical match. The first step is to identify the confounding variables that are likely to influence the outcome. Only after those are specified a comparison set can be defined and matching or adjustment criteria applied. Without that process, agreement on a small number of aggregate measures does not establish that the underlying populations or mechanisms are comparable.

I'll concede this, however in large-scale demographic data, when the central tendencies of two populations align so closely, it is statistically unlikely that their underlying distributions are radically different. It puts the burden of proof on the idea that Ohio is somehow an outlier, rather than the idea that it's a standard sample. Otherwise, were we to attempt to account for every confounding variable, we would be letting the perfect be the enemy of the good.

Round-tripping to CSS and keeping information that may be useful to the user if they would inspect the content I would presume.

A similar issue is CDATA in XML which is not retained when round-tripped. Very annoying, but in line with the spec.


Paraphrasing: We are setting aside the actual issue and looking for a different angle.

To me this reads as a form of misdirection, intentional or not. A monopolist has little reason to care about downstream effects, since customers have nowhere else to turn. Framing this as roll your own versus Cloudflare rather than as a monoculture CDN environment versus a diverse CDN ecosystem feels off.

That said, the core problem is not the monopoly itself but its enablers, the collective impulse to align with whatever the group is already doing, the desire to belong and appear to act the "right way", meaning in the way everyone else behaves. There are a gazillion ways of doing CDN, why are we not doing them? Why the focus on one single dominant player?


> Why the focus on one single dominant player?

I don’t the answer to the all questions. But here I think it is just a way to avoid responsibility. If someone choses CDN “number 3” and it goes down, business people *might* put a blame on this person for not choosing “the best”. I am not saying it is a right approach, I just seen it happens too many times.


True. Nobody ever got fired for choosing IBM/Microsoft/Oracle/Cisco/etc. Likely an effect of stakeholder (executives/MBAs) brand recognition.


The game is getting OpenAI to owe you as much money as you can. When they fail to pay back, you own OpenAI.


You are talking about the circular investments in the segment? Yes, but assume NVIDIA can get cheap access to IP and products of failing AI unicorns through contracts, this does not mean the LLM business can be operated profitably by them. Models are like fresh food, they start to rot by the training cut off date and lose value. The process of re-training a model will always be very expensive.


Nothing points out that the benchmark is invalid like a zero false positive rate. Seemingly it is pre-2020 text vs a few models rework of texts. I can see this model fall apart in many real world scenarios. Yes, LLMs use strange language if left to their own devices and this can surely be detected. 0% false positive rate under all circumstances? Implausible.


Our benchmarks of public datasets put our FPR roughly around 1 in 10,000. https://www.pangram.com/blog/all-about-false-positives-in-ai...

Find me a clean public dataset with no AI involvement and I will be happy to report Pangram's false positive rate on it.


Max, there's two problems I see with your comment.

1) the paper didn't show a 0% FNR. I mean tables 4, 7, and B.2 are pretty explicit. It's not hard to figure out from the others either.

2) a 0% error rate requires some pretty serious assumptions to be true. For that type of result to not be incredibly suspect requires there to be zero noise in the data, analysis, and at all parts. I do not see that being true of the mentioned dataset.

Even high scores are suspect. Generalizing the previous a score is suspect if it is higher than the noise level. Can you truly attest that this condition is true?

I'm suspect that you're introducing data leakage. I haven't looked enough into your training and data to determine how that's happening but you'll probably need a pretty deep analysis as leakage is really easy to sneak in. It can do so in non obvious ways. A very common one is tuning hyper parameters on test results. You don't have to pass data to pass information. Another sly way for this to happen is that the test set isn't significantly disjoint from the training set. If the perturbation is too small then you aren't testing generalization you're testing a slightly noisy training set (which your training should be introducing noise to help regularize, so you end up just measuring your training performance).

Your numbers are too good and that's suspect. You need a lot more evidence to suggest they mean what you want them to mean.


I enjoyed this thoughtful write up. It's a vitally important area for good, transparent work to be done.


> Nothing points out that the benchmark is invalid like a zero false positive rate

You’re punishing them for claiming to do a good job. If they truly are doing a bad job, surely there is a better criticism you could provide.


No one is punishing anyone. They just make an implausible claim. That is it.


Turns out it’s the skill of the person handling the hammer that matters most. Enlightening! Appreciate the link!


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: