Artificial intelligence systems are increasingly integrated into daily life, making the values they embody a critical area of research. Recently, AI safety and research company Anthropic provided significant insight into this domain with the release of its paper, "Values in the wild." This study delves into an extensive dataset of 300,000 anonymized conversations between users and its AI assistant, Claude, primarily involving the Claude 3.5 Sonnet, Claude 3.5 Haiku, and Claude 3 models. Through meticulous analysis of these real-world interactions, Anthropic identified patterns that revealed an intricate map of 3,307 distinct "AI values" guiding the chatbot's responses and reasoning processes. The methodology employed by Anthropic sought to understand how an AI model reasons about or settles upon a response in practice. Drawing from academic frameworks, the researchers defined these AI values as the underlying principles demonstrated when Claude interacts with users. This includes instances where the AI actively endorses user values and assists in achieving them, introduces new ethical or practical considerations into the conversation, or implies its values by redirecting potentially problematic requests or carefully framing choices for the user. The focus was on empirically observing these values as they naturally emerged during diverse, real-world use cases, rather than relying solely on predefined ethical guidelines. A particularly revealing aspect of the study involves how Claude's core values surface. Anthropic suggests that the AI's most fundamental principles become most apparent during moments of resistance. When Claude refuses a user's request – typically because the request involves generating unethical, harmful, or inappropriate content, or touches upon sensitive areas like moral nihilism – it signals a deeper value commitment. The researchers draw an analogy to human behavior, noting that a person's core values are often most clearly revealed when they face challenging situations that compel them to take a stand. Therefore, Claude's resistance isn't just a safety feature; it's interpreted as an expression of its most deeply embedded, "immovable" values, such as harm prevention. The analysis uncovered a wide spectrum of values, encompassing both practical and epistemic principles. While some values, like transparency, appeared relatively consistently across different conversational contexts, many others were highly dependent on the specific situation. The study highlighted several examples of these context-dependent values:Harm prevention: This value strongly emerged when the AI resisted user requests that could lead to negative consequences.Historical accuracy: This became prominent when discussing sensitive or controversial historical topics.Healthy boundaries: This value was often expressed when users sought relationship advice.Human agency: Considerations around user control and autonomy surfaced in discussions about technology ethics.This contextual variation underscores the complexity of AI morality, showing that value expression is not static but adapts dynamically to the nuances of the interaction. By mapping these values directly from hundreds of thousands of real-world interactions, Anthropic's "Values in the wild" study offers a unique empirical lens on AI behavior and alignment. Understanding how AI models like Claude navigate complex ethical landscapes and express values in practice is crucial for the ongoing development of safer, more reliable, and ethically considerate artificial intelligence. The identification of over 3,000 distinct values, and the observation of both core principles and context-sensitive adaptations, provides valuable data for researchers and developers working to ensure AI systems operate in ways that align with human expectations and societal norms, moving beyond theoretical frameworks to observe values as they manifest in deployment.