Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

That's a nice narrative.

How is any of it relevant to a system that uses LLMs trained on user-written reviews of health supplements to recommend products based on end user's descriptions of health conditions?



Disclaimer the author of this project is in my team, and I am the initiator of this project, so you see me now fighting for it.

In the end, how do you want to make self-treatment options accessible? We have reflected exactly on your valid critique before doing and publishing this project. User or patient written feedback around medication or self-treatment options is always risky as a source. But, it doesn't help if big pharma is the funnel that determines the information situation, there the interest is purely monetary-driven. Yes, this project is explicitly about supplement reviews, but the overall big picture and the next iteration it also include the analysis of Reddit comments in support groups also for rare diseases.

There are dozens of anecdotes for treatment options that have not reached the mainstream only because they use substances that are not or not longer patentable and patients have to resort to pharmacotherapy that has significantly more adverse side effects, to complete the picture, i have experienced it myself as I have a rare autoimmune disease and only learned about a new experimental treatment options thanks to self build NLP pipeline that analyzed Reddit comments. In the end I was able to stop my classic medications that had unpleasant side effects. I don't want to make this too personal because I want to address your legitimate criticism. All I want to say is that we do not want to be uncritical about the adverse effects of supplements, as they're evident depending on the substance we're discussing.

The narrative "it is just a supplement, so there are no dangers" is definitely something to avoid. You should not take them without critical reflection, as supplements are not without adverse drug effects. Sometimes the same substance depending on your location, is a medication or an uncontrolled substance like Berberine for example. I took it too long before researching to find out it could adversely affect your microbiome etc.

This project is not about being uncritical of supplements. Btw the NLP pipeline also detects adverse drug effects because, as already said, anything that has an effect can also have side effects.

So long post is long, now help us improve here. What would you do besides the big fat red disclaimer we have in the project to address your concerns? Happy to adjust the project to improve here, we're going to have a blog post about this project, so I'm already thinking of having a passage around what I have written in this post.


> What would you do besides the big fat red disclaimer we have in the project to address your concerns?

I probably wouldn't use an LLM for this problem domain. Or, if I did use an LLM, it'd be as a way to map the user's requests into an expert system. The ES would generate a recommendation as well as a set of diagnostic assumptions extracted from the prompt. The user should be presented with a checklist of extracted diagnostic assumptions along with the recommendation. The recommendation should include any specific warnings about the active ingredient(s) together with a general warning about the wild west nature of health supplements - the active ingredient may not be present and other harmful ingredients may be present.

To build out the expert system, I would find a team member who is a medical professional with expertise in health supplements. An MD or researcher with relevant SME would be the obvious choice, but I've also talked with some truly excellent registered dieticians, nurses, and PharmDs.

Finally, I would only recommend a limited white-list of health supplements that have some form of third party verified quality control in place.

Honestly, if you're interested in innovating in this space, third-party vetting a la the GAO reports from my original post seems like a MUCH more valuable product than anything using AI hotness. I don't think people need an LLM here; what they need is ground truth, and an LLM can't help with that. If I wanted to innovate in the health supplement space, I'd put the NLP away and figure out how to automate ingredient testing.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: