Hacker Newsnew | past | comments | ask | show | jobs | submit | KomoD's commentslogin

I like the idea (I've discovered tons of movies and shows from youtube shorts) but sorry, no way I'll auth my Google account to this

Bad data. A bunch of links just lead to 404, the filters aren't correct (since when is organic pumpkin seed protein powder nut flavored?)

$239.95 for 9068g with 16g serving size (60 calories, 10g of protein) is not 4 servings and is not 0.2g protein/$1, 1 cal/$1?

And for another product your site says 18 servings, the listing says 16 servings


Thanks for checking it out / the feedback.

I haven’t noticed the 404s, but I’ll check.

On filters: Agreed, I see a few pumpkin seed protein powders tagged as nut flavored that should be unflavored. Will address.

On the “4 servings” issue: I wasn’t handling multi-packs properly, planning on adding that. In addition that products nutrition label did not have serving count, so looks like the llm got confused and took pack count. Im skeptical about inferring serving count from total weight / serving size, as I’ve seen many examples of incorrect total weight information on products.

On 18 vs 16 servings: that’s a LLM misread from the nutrition panel. Thinking reprocessing with larger model may address it.


This is such a non-issue, it's just a name.

If someone asks about it "It's a tool to view docx files", end of conversation


We've got `git` (an insult), `kill` (violent), `slack` (not doing work) and `fsck` (looks like fuck). Doxx seems ok to me too.


I've seen the `itsdangerous` [1] package (which is a dependency for lot of Python projects) raise some eyebrows several times.

[1] https://itsdangerous.palletsprojects.com/en/stable/


I get the sense you've never worked under the oppresive thumb of dashboard-driven enterprise IT, heh


> No, he (Matthew) opted everyone in by default

Now you're just lying.

I checked several of my Cloudflare sites and none have it enabled by default:

"No robots.txt file found. Consider enabling Cloudflare managed robots.txt or generate one for your website"

"A robots.txt was found and is not managed by Cloudflare"

"Instruct AI bot traffic with robots.txt" disabled


I think lying is a bit strong, I think they're potentially incorrect at worst.

The Cloudflare blog post where they announced this a few weeks ago stated "Cloudflare, Inc. (NYSE: NET), the leading connectivity cloud company, today announced it is now the first Internet infrastructure provider to block AI crawlers accessing content without permission or compensation, by default." [1]

I was also a bit confused by this wording and took it to mean Cloudflare was blocking AI traffic by default. What does it mean exactly?

Third party folks seemingly also interpreted it in the same way, eg The Verge reporting it with the title "Cloudflare will now block AI crawlers by default" [2]

I think what it actually means is that they'll offer new folks a default-enabled option to block ai traffic, so existing folks won't see any change. That aligns with text deeper in their blog post:

> Upon sign-up with Cloudflare, every new domain will now be asked if they want to allow AI crawlers, giving customers the choice upfront to explicitly allow or deny AI crawlers access. This significant shift means that every new domain starts with the default of control, and eliminates the need for webpage owners to manually configure their settings to opt out. Customers can easily check their settings and enable crawling at any time if they want their content to be freely accessed.

Not sure what this looks like in practice, or whether existing customers will be notified of the new option or something. But I also wouldn't fault someone for misinterpreting the headlines; they were a bit misleading.

[1]: https://www.cloudflare.com/en-ca/press-releases/2025/cloudfl...

[2]: https://www.theverge.com/news/695501/cloudflare-block-ai-cra...


> I think lying is a bit strong, I think they're potentially incorrect at worst.

I understand that you're trying to be generous, but the claim that "Matthew opted everyone in by default" is flat out incorrect.


We are in agreement; I just think saying "lying" implies a level malintent that isn't present -- it's rather overly ungenerous. The poster is at worst incorrect. And their misconception is understandable given the company's own confusing marketing.


I recently created a new Cloudflare account for a project I’m working on and moved two domains into it, and the settings were both on by default without asking me about it at all. The original press release specifically mentioned enabling it by default.

> Cloudflare, Inc. (NYSE: NET), the leading connectivity cloud company, today announced it is now the first Internet infrastructure provider to block AI crawlers accessing content without permission or compensation, *by default*.


That's not Cloudflare's fault, that's the website owner's fault.

If they want the RSS feeds to be accessible then they should configure it to allow those requests.


No... it doesn't run locally, just open devtools and see for yourself?

Do you really think your Toshiba Satellite can run Claude's models?


wat? model run the cloud


Yes, I block DO on all my servers.


> Malicious actors can easily hop onto residential proxies.

They can, but most don't. It's a lot more expensive than spinning up a $5 droplet


$4 for 1GB, which is more than enough: https://oxylabs.io/pricing/residential-proxy-pool


> I wouldn't say they’re “obviously” AI fakes.

I'd say they're obviously AI fakes, just trying a few: B249AL (it made her bald), SA487AB (different shape, hair color and hair), TN248DF (it grew his hair back), HA26ND (bald, again) and NG166QE (I don't even need to explain)...


Nope.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: