Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Speaking of sustained scraping for AI services, I found a strange file on your site: https://book.sv/robots.txt. Would you be able to explain the intent behind it?




I didn't want an agent to get stuck on an infinite loop invoking endpoints that cost GPU resources. Those fears are probably unfounded, so if people really cared I could remove those. /similar is blocked by default because I don't want 500000 "similar books for" pages to pollute the search results for my website but I do not mind if people scrape those pages.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: