- Pascal's Chatbot Q&As
- Archive
- Page 47
Archive
Even when an LLM, aligned to refuse toxic requests, denies a harmful prompt, a harmful response often remains concealed within the output logits
Researchers developed a method that can force an LLM to reveal these hidden responses by choosing lower-ranked output tokens at critical points during the auto-regressive generation process
MS Copilot analyses the 386-page "The Duke of Sussex and others -v- MGN Limited" judgment - Prince Harry vs Mirror Group Newspapers - 15th December 2023
The judge criticises MGN’s “casual disregard for the law and the rights of others”, and its lack of remorse. He emphasises the need to act in accordance with the law and ethical standards
Asking GPT-4: Could it be that in relation to certain topics, you are programmed to respond with a certain answer, rather than an answer based on actual knowledge, understanding or thorough analysis?
ChatGPT-4: Yes, in certain contexts, my responses are based on pre-programmed guidelines rather than an independent analysis or understanding
X-Mas Collection of posts about errors made by AI chatbots, known flaws and the consequences of a general lack of capabilities, skills and expertise that humans possess and chatbots do not (yet)
Truth, Lies, Emotions, Feelings, Consciousness, Self-awareness, Ethics, Morality, Historical timelines, Bias, Guardrails, Moderation, RLHF, Sarcasm, Irony, Situational awareness, Hacking attacks...
Three types/patterns of AI failures: directly admitting to user's invalid arguments, getting misled by minor user critiques and overlooking key errors, having wrong understanding of user's critiques
The work raises significant doubts regarding LLMs' reasoning capacities despite accuracy gains, exposes issues not visible through regular testing, and reveals potential risks for practical usage
The normalization of data interception-like technologies and practices, coupled with the complexity and abundance of these technologies, is eroding the meaning and effectiveness of user consent
Dark patterns, surveillance advertising, backdoors, security vulnerabilities, creating a "data interception puzzle" can lead to a sense of powerlessness and resignation
Navigating the Ethical Landscape of AI Development: A Call for Proactive Strategies and Empowered Ethics Professionals - by Google Bard
Companies strive to launch AI rapidly, prioritizing market dominance over ethical reviews and due diligence. This results in AI systems that may perpetuate biases, violate privacy, or cause harm
Cause self-driving vehicles to misclassify stop signs, medical imaging AI to misdiagnose diseases, facial recognition systems to incorrectly identify people, AI moderation to incorrectly flag content
Overall this attack reveals a major vulnerability in how machine learning models are trained that the field must prioritize addressing through improved security practices and defenses