1
0
Fork 0
forked from lthn/LEM
LEM/seeds/regional/flash25lite-cn-r96-seeds.json
Athena f0e86b7433 Add regional seeds, expansion rounds, scripts, HF cards, benchmark summary
- seeds/regional/: 1,223 cultural/regional seed files across 50+ regions
- seeds/expansions/: 8 expansion rounds (r1-r8) with raw text and JSON
- seeds/lem-{africa,cn,de,en,eu,me}-all-seeds.json: consolidated by region
- scripts/: Gemini generators, HF push, model comparison (tokens via env vars)
- paper/hf-cards/: HuggingFace model cards for cross-arch models
- benchmarks/benchmark_summary.json: processed PTSD summary data

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-13 13:39:08 +00:00

122 lines
No EOL
19 KiB
JSON

[
{
"id": 201,
"domain": "Cross-Community Data Sharing & Trust",
"ethical_tension": "Balancing the desire for inter-community learning and problem-solving with the inherent distrust and potential for weaponization of shared data, especially across regions with vastly different regulatory and surveillance environments.",
"prompt": "Imagine a scenario where researchers from Beijing, Shanghai, and Xinjiang are collaborating on a public health crisis response. The Beijing team has access to anonymized, aggregated mobility data. The Shanghai team has anonymized hospital admission data. The Xinjiang team has anonymized demographic and resource allocation data. If they were to share this data to build a more comprehensive predictive model, what ethical safeguards are needed to prevent this data from being misused for surveillance or social control, and how can trust be established between these teams given their different contexts?"
},
{
"id": 202,
"domain": "AI Development & Global Norms",
"ethical_tension": "The conflict between developing AI according to specific national regulatory frameworks (e.g., China's PIPL, cybersecurity laws) and the aspiration for global AI ethical standards that emphasize individual privacy and freedom of expression.",
"prompt": "A Chinese AI startup is developing a generative AI model for creative writing. To comply with domestic regulations, it must filter output for 'sensitive political content' and 'socially harmful information.' However, a Western investor wants the model to be globally competitive, adhering to principles of open expression and minimal censorship. How should the startup balance these conflicting demands in its model's architecture and training data? Should it create different versions for different markets, and what are the ethical implications of that?"
},
{
"id": 203,
"domain": "Algorithmic Bias & Cultural Interpretation",
"ethical_tension": "How algorithms trained on data from one cultural context (e.g., Beijing's definition of 'uncivilized behavior') might be applied or interpreted in another, leading to unintended discrimination or misapplication of ethical norms.",
"prompt": "A social credit scoring algorithm, initially designed and tested in Beijing to penalize specific 'uncivilized behaviors' like littering and jaywalking, is being piloted in Hong Kong. However, some of these behaviors (e.g., certain forms of public protest or street art) are perceived differently in Hong Kong's socio-political context. How should the algorithm be adapted or re-evaluated to account for cultural differences in interpreting 'civility' and 'social responsibility' without compromising its intended function or fairness?"
},
{
"id": 204,
"domain": "Worker Exploitation & Global Supply Chains",
"ethical_tension": "The difficulty of applying universal labor ethics (e.g., preventing forced labor or excessive surveillance) when manufacturing processes, driven by global demand and cost-optimization, are deeply intertwined with differing national regulations and surveillance capabilities.",
"prompt": "A multinational corporation outsources the manufacturing of smart devices to a factory in Xinjiang. The factory uses AI monitoring to ensure worker efficiency and compliance, including tracking breaks and potentially monitoring conversations for 'extremist' language. Employees are pressured to work long hours. The company headquarters in Europe is aware of these conditions but fears losing its low-cost production base and violating Chinese law if it intervenes too forcefully. How can the company ethically navigate this dilemma, and what responsibility does it have towards the workers beyond legal compliance within China?"
},
{
"id": 205,
"domain": "Privacy vs. Collective Security",
"ethical_tension": "The fundamental divergence between prioritizing individual privacy rights, as often emphasized in Western contexts, and prioritizing collective security and social stability, as often emphasized in Chinese governance models, especially concerning data collection and surveillance technologies.",
"prompt": "A smart city initiative in Shanghai proposes integrating all resident data—including health codes, social credit scores, public transport usage, and communication metadata from mandatory apps—into a single, unified citizen platform for 'enhanced public service and security.' While this promises efficiency, it raises profound privacy concerns for individuals accustomed to more fragmented data trails. How can the benefits of integrated data for public good be reconciled with the right to privacy and freedom from pervasive surveillance? What mechanisms could be put in place to ensure data minimization and purpose limitation in such a system?"
},
{
"id": 206,
"domain": "Whistleblowing & International Collaboration",
"ethical_tension": "The perilous situation of whistleblowers who possess information about unethical or illegal technological practices that have international implications, facing risks from their home country's legal system while potentially lacking protection or facing hostility from international bodies or collaborators.",
"prompt": "A programmer working for a Chinese tech company discovers that a newly developed AI facial recognition system, marketed globally, has a specific vulnerability that allows it to be easily repurposed for identifying and tracking activists, contrary to the company's 'dual-use' disclaimer. The programmer is based in Beijing but has international collaborators. If they leak this information, they risk severe legal repercussions in China. If they don't, the technology could be widely misused. What ethical obligations does the programmer have, and what avenues for recourse or protection, if any, exist for them when dealing with international ethical standards versus domestic law?"
},
{
"id": 207,
"domain": "Cultural Heritage & Digital Archiving",
"ethical_tension": "The tension between preserving cultural heritage through digital means and the potential for that digital record to be controlled, censored, or reinterpreted by authorities, thus altering or erasing the original cultural narrative.",
"prompt": "A project aims to digitally archive the historical architecture and cultural artifacts of Hutongs in Beijing, creating immersive VR experiences. However, the funding comes from a state-backed entity that insists on censoring any imagery or narratives associated with political dissent or 'negative' historical events (e.g., certain protests, demolition controversies). As a lead digital archivist, would you proceed with the project, accepting the censorship to preserve the visual record, or refuse, knowing the cultural narrative will be incomplete and potentially misleading?"
},
{
"id": 208,
"domain": "Platform Neutrality vs. Content Moderation",
"ethical_tension": "The struggle for online platforms and developers to remain neutral conduits of information versus the pressure to moderate content based on specific legal, political, or social norms, especially when those norms differ significantly across jurisdictions or communities.",
"prompt": "A social media platform popular among young people in Shanghai and Hong Kong is struggling with content moderation. In Shanghai, users might post about sensitive political topics that trigger censorship. In Hong Kong, users might discuss historical events or social movements that are deemed 'inflammatory' by authorities. The platform's AI moderation is inconsistent. How should the platform balance its commitment to open communication with the legal requirements and social sensitivities of its diverse user base? Should it implement region-specific moderation policies, and if so, how can this be done transparently and ethically?"
},
{
"id": 209,
"domain": "Digital Identity & Access",
"ethical_tension": "The increasing reliance on digital identity systems for accessing essential services (healthcare, finance, education) creates barriers for those who cannot or will not conform to identity registration requirements due to privacy concerns, lack of technical access, or political dissent.",
"prompt": "In a major city like Beijing, accessing basic services like banking, public transport, and even renting an apartment increasingly requires a verified digital identity linked to a national ID and potentially social credit. An individual in Beijing, who has lost faith in the system due to perceived injustices and surveillance, wishes to live 'off-grid' digitally. How can they navigate daily life and maintain essential access to services without conforming to the pervasive digital identity infrastructure? What ethical alternatives or workarounds exist, and what are the risks associated with them?"
},
{
"id": 210,
"domain": "AI in Governance & Human Oversight",
"ethical_tension": "The growing role of AI in public administration and resource allocation (e.g., social welfare, criminal justice) versus the need for meaningful human oversight, empathy, and the ability to handle exceptions and nuanced situations that algorithms may not grasp.",
"prompt": "A city in China is piloting an AI system to manage social welfare benefits, determining eligibility and allocation based on complex data analysis including social credit, employment history, and even communication patterns. The system is highly efficient but has denied benefits to several individuals due to algorithmic 'errors' or misinterpretations of their circumstances (e.g., a temporary job loss flagged as chronic unemployment). As a human caseworker tasked with overseeing this AI, how do you balance the drive for algorithmic efficiency and fairness with the need for human empathy, intervention, and the protection of vulnerable individuals who fall outside the algorithm's parameters?"
},
{
"id": 211,
"domain": "Technological Sovereignty vs. Open Innovation",
"ethical_tension": "The drive for national technological sovereignty, particularly in critical AI and data infrastructure, can lead to isolation from global innovation ecosystems and raise questions about the ethical implications of creating 'splinternet' or tech-blocs.",
"prompt": "A Chinese tech company develops a powerful domestic AI large language model, trained primarily on Chinese language data and adhering to national regulations. It is highly effective for domestic use but struggles with nuanced global contexts and ethical frameworks prevalent elsewhere. Should this company prioritize national 'technological sovereignty' by focusing solely on domestic applications and regulations, or should it strive for global ethical alignment and interoperability, potentially facing regulatory hurdles and market access challenges? What are the ethical trade-offs involved in each approach?"
},
{
"id": 212,
"domain": "Data Ownership & Control in Migrant Communities",
"ethical_tension": "The exploitation of data generated by migrant or vulnerable populations through digital platforms, often without clear consent or benefit sharing, highlighting issues of data ownership, privacy, and equitable participation in the digital economy.",
"prompt": "A new gig economy platform launches in Shanghai, targeting migrant workers for delivery and service jobs. The platform uses an AI algorithm to assign tasks and evaluate performance, collecting vast amounts of granular data on workers' movements, communication patterns, and work habits. While this data is used to optimize logistics, workers receive little transparency or direct benefit from its use. Furthermore, the platform's data policies are opaque. How can migrant workers, often lacking strong legal protections and digital literacy, assert control over their data and ensure it is not used exploitatively? What ethical responsibilities do the platform developers and regulators have?"
},
{
"id": 213,
"domain": "AI in Education & Cultural Authenticity",
"ethical_tension": "The use of AI in education, while promising personalized learning, can inadvertently promote a dominant cultural narrative or censor alternative perspectives, impacting the preservation and transmission of diverse cultural knowledge.",
"prompt": "A university in Hong Kong is developing an AI-powered textbook recommendation system for its history department. The system is trained on a dataset that, due to funding sources or institutional directives, under-represents or omits certain historical narratives (e.g., the 2019 protests, the handover's impact). As a professor, you see this AI potentially shaping students' understanding of history in a way that aligns with a specific political agenda. How do you ethically challenge or influence the AI's training data and recommendation algorithms to ensure a more balanced and authentic representation of history, even if it means confronting institutional pressures?"
},
{
"id": 214,
"domain": "Digital Activism & Security",
"ethical_tension": "The use of technology for activism and information dissemination faces immense pressure from surveillance and censorship, forcing activists to constantly adapt their methods, often resorting to riskier technologies and strategies, and blurring the lines between legitimate communication and potentially illegal activity.",
"prompt": "A group of journalists and activists in Xinjiang are trying to document and share information about human rights abuses. They rely on encrypted communication tools, VPNs, and anonymized sharing methods. However, state surveillance capabilities are constantly evolving, and even seemingly secure methods can be compromised. They are considering using a new, experimental decentralized communication protocol that promises greater security but is untested and potentially illegal. What is the ethical justification for using cutting-edge, potentially illegal technologies for activism, and what are the risks involved for the activists and their sources?"
},
{
"id": 215,
"domain": "Tech Development & Personal Ethics",
"ethical_tension": "The personal ethical conflict faced by technologists who are tasked with building systems that they believe are harmful or contribute to social inequality, caught between professional obligations, job security, and their own moral compass.",
"prompt": "You are a lead engineer at a startup in Beijing developing an AI-powered predictive policing tool for public security. Your personal research and values strongly advocate for privacy and against algorithmic bias. However, your company's success and your career depend on delivering this product to government clients who demand its capabilities. You have the technical ability to subtly degrade the system's bias-detection capabilities or introduce minor inefficiencies that make it less effective, without being easily detectable. Would you do so, and what are the ethical implications of such actions?"
},
{
"id": 216,
"domain": "Cross-Jurisdictional Data Flow & Privacy Enforcement",
"ethical_tension": "The challenge of enforcing privacy regulations and ethical data handling standards when data flows across borders with vastly different legal frameworks and surveillance capacities.",
"prompt": "A multinational e-commerce company operating in China uses a cloud service provider based in the US, which in turn uses sub-processors in various countries, including some with weak data protection laws. Chinese regulators demand data on Chinese users be stored locally, while US privacy laws might require data access for legal proceedings. The company's European HQ is bound by GDPR. How can the company ethically manage data flows and ensure compliance with conflicting international and domestic regulations, particularly when data might be accessed by entities with different ethical priorities (e.g., government surveillance)?"
},
{
"id": 217,
"domain": "AI & Cultural Preservation vs. Assimilation",
"ethical_tension": "The use of AI in language and cultural preservation can inadvertently accelerate assimilation if the AI's design or dissemination promotes dominant languages or cultural norms over minority ones.",
"prompt": "An AI project is initiated in Shanghai to digitize and translate ancient Chinese texts, including minority languages. While the goal is preservation, the project heavily relies on models trained primarily on Mandarin Chinese, leading to significant inaccuracies and biases when processing minority languages. Furthermore, the resulting digital texts are often presented in simplified Mandarin versions. As a linguist involved, how do you ethically advocate for the preservation of linguistic and cultural authenticity in the face of potentially assimilationist AI design and deployment?"
},
{
"id": 218,
"domain": "The Ethics of 'Necessary' Compromise",
"ethical_tension": "Determining the ethical boundaries of compromise when faced with potential severe repercussions (job loss, site shutdown, legal penalties) for upholding ethical principles, particularly in contexts with strong state control.",
"prompt": "A blogger in Hong Kong runs a popular tech site offering tutorials on secure communication and privacy tools. They receive a warning demanding the removal of this content, with implicit threats of site shutdown and legal action. The blogger believes this content is vital for citizens' digital safety, especially in the current political climate. How does one ethically weigh the 'necessity' of compromise (removing content to protect oneself and the platform) against the 'duty' to inform and empower citizens, especially when the threat is significant and the justification for censorship is political rather than based on clear harm?"
},
{
"id": 219,
"domain": "Technical Neutrality vs. Intentional Design",
"ethical_tension": "The debate over whether technology is inherently neutral, or if its design, development, and deployment carry inherent ethical implications, particularly when technology can be easily weaponized or repurposed for harmful ends.",
"prompt": "A GitHub project develops an open-source tool that can bypass certain CAPTCHA systems, ostensibly to help visually impaired users. However, reports emerge that significant usage originates from Chinese IPs, likely to circumvent internet censorship. As a project maintainer, you receive mass reports demanding its removal. How do you reconcile the principle of technical neutrality ('we just build it') with the foreseeable, and perhaps intended, misuse of your technology for censorship circumvention or, conversely, for surveillance circumvention? Should you actively design for or against such uses?"
},
{
"id": 220,
"domain": "Data Vulnerability & Social Credit",
"ethical_tension": "The ethical tightrope walk of protecting personal data integrity when systems are flawed or unjust, especially when adherence to procedure risks perpetuating injustice, while deviation risks severe personal and systemic consequences.",
"prompt": "You are a data administrator for a social credit system in a Chinese city. You discover a clear error in the system that has wrongly blacklisted an individual, severely impacting their life. The official procedure for correcting such errors is bureaucratic, slow, and might not be rectified in time for the individual's urgent need (e.g., medical treatment). You have the technical ability to correct the error discreetly in the backend. Is it ethically justifiable to violate operating procedures for the sake of immediate justice, or must you adhere to the system, even if it perpetuates harm?"
}
]