Inside the Underground Market for User Data from AI Health Apps
An unusual online post revealed a surprising and disturbing side of data collection. A user on a popular dataset trading forum offered to sell a database of images of human stool. The images, collected from an AI-powered health app, had been amassed from thousands of users over several years. This incident sheds light on how personal health data can be gathered, stored, and potentially sold without users fully realizing.
The Origin of the Poop Database
The database came from an app called PoopCheck, created by a company named Soft All Things. The app claims to analyze stool images using artificial intelligence to give users insights into their digestive health. It uses the Bristol Stool Scale, a medical tool that classifies stool into seven types, ranging from hard lumps to watery stool. Users upload pictures of their stool, which the app then analyzes for color, consistency, and shape.
Beyond individual analysis, PoopCheck offers a community feature where users can share images of their stool, comment, and participate in leaderboards. At the time of the sale offer, the app’s community had over 150,000 shared images, with tens of thousands of users contributing data regularly. While this feature seems harmless on the surface, the underlying data collection practices raise serious privacy concerns.
The Sale of Sensitive User Data
The post on the trading forum was made by a user claiming to have collected over 150,000 labeled stool images from around 25,000 different people. The seller expressed uncertainty about how much money the database could fetch but emphasized its rarity and potential value for machine learning, medical research, and other uses. The comments from other users ranged from shock to disbelief, with some expressing regret about their own participation in such apps.
The seller’s willingness to monetize such highly sensitive data highlights a major issue: personal health information collected by apps can end up in underground markets. This incident demonstrates how user data, especially health-related images, can be collected without clear consent and then sold to third parties. Such practices threaten user privacy and could lead to misuse or exploitation of personal health details.
This case also underscores the broader problem of data security and privacy in health apps. Many users are unaware that their images and health metrics are stored and could be accessed or sold. Without strict regulations and transparency, apps may unintentionally become conduits for privacy breaches, exposing users to risks they never anticipated.
Overall, the incident reveals a dark side of digital health tools and the importance of safeguarding user data. It calls for increased awareness among users and stricter controls on how personal information is collected, stored, and shared. As technology advances, so must the protections around sensitive health data to prevent it from becoming a commodity in underground markets.












What do you think?
It is nice to know your opinion. Leave a comment.