Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Assuming you don't store the salts, this produces a value that is useless for anything but counting something like DAU. Which you could equally just do by counting them all and deleting all the data at the end of the day, or using a cardinality estimator like HLL.


DAU in regards to a given page.

Have you read the article? That is what the author's goal seems to be.

He wants to prevent multiple requests to the same page by the same IP counted multiple times.


Is that more efficiently done with an appropriate caching header on the page as it is served?

Cache-Control: private, max-age=86400

This prevents repeat requests for normal browsers from hitting the server.


That same uselessness for long-term identification of users is what makes this approach compliant with laws regulating use of PII, since what you have after a small time window isn't actually PII (unless correlated with another dataset, but that's always the case).


That's precisely all that OP is storing in the original article.

They're just getting a list of hashes per day, and associated client info. They have no idea if the same user visit them on multiple days, because the hashes will be different.




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: