Generative AI poses real privacy risks, eroding privacy en masse and providing tools for bad actors.The solution involves consumers demanding transparency and government regulation.Mitigating privacy risks requires educating the public about data privacy and AI adoption.Generative AI can include text, audio, and video content generation using large language models.Training stages involve pre-training, supervised instruction fine-tuning, and reinforcement learning from human feedback.Privacy risks of generative AI include data compromise, leaked personal information, and data misuse during interactions.AI exacerbates surveillance, identification, aggregation, and exclusion risks in privacy.Concrete AI examples show how user interactions with generative AI systems impact data privacy.Steps like data minimization, transparency, and user consent are proposed to mitigate generative AI privacy risks.The clear solution involves consumers becoming aware of data value and demanding data privacy, potentially reshaping the surveillance economy.