Generative AI: 1. Ethics 2.CLIP: Difference between revisions
Cindy.tang (talk | contribs) |
Cindy.tang (talk | contribs) |
||
Line 1: | Line 1: | ||
==Motivation== | ==Motivation== | ||
In the current era, the rise of Large Language Models (LLMs) like GPT-3 or LLAMA has evoked a mix of fascination and | In the current era, the rise of Large Language Models (LLMs) like GPT-3 or LLAMA has evoked a mix of fascination and apprehension. These advanced models showcase remarkable capabilities of generating human-like text and performing complex tasks, while also raising profound ethical questions. | ||
The integration of ethics into AI systems faces numerous challenges. Firstly, there is the challenge of modelling reasoning about obligations and permissions. Secondly, complexities arise from the persistent conflicts within various ethical reasonings. Lastly, comprehending and assessing the consequences of actions remains an intricate undertaking for both humans and machines. | |||
Researchers have experimented with various techniques to address these challenges. Some have turned to deontic logics <ref>Horty, J. F. (2001). Agency and deontic logic. Oxford University Press.</ref> and formalisms inspired by such considerations to handle the particular nature of duty rules. Others propose AI logic-based non-monotonic formalisms <ref>Ganascia, J. G. (2015). Non-monotonic resolution of conflicts for ethical reasoning. A Construction Manual for Robots' Ethical Systems: Requirements, Methods, Implementations, 101-118.</ref> such as default logics or answer set programming, closely aligned with common-sense reasoning, to mitigate logical contradictions. Additionally, there are proposals to employ action language or causal models <ref>Mueller, E. T. (2014). Commonsense reasoning: an event calculus based approach. Morgan Kaufmann.</ref>, providing a mathematical foundation for understanding and computing action consequences. | |||
The present technical hurdle lies in merging these three approaches into a unified framework—a framework that is nonmonotonic, adept at managing norm conflicts, and employs causal models to evaluate action consequences. These diverse approaches adopt varying normative frameworks, encompassing utilitarianism, deontology, and virtue ethics, necessitating simulation. However, the specifics of these simulations often lack clarity, particularly from a philosophical standpoint. <ref>Powers, Thomas M., and Jean-Gabriel Ganascia, 'The Ethics of the Ethics of AI', in Markus D. Dubber, Frank Pasquale, and Sunit Das (eds), The Oxford Handbook of Ethics of AI (2020; online edn, Oxford Academic, 9 July 2020), https://doi.org/10.1093/oxfordhb/9780190067397.013.2</ref> | |||
Our project aims to delve into this multifaceted ethical landscape surrounding AI from both technical and philosophical perspectives. We want to explore how AI systems deal with ethical dilemmas in the light of these diverging ethical priorities and seek methods to align these systems more closely with human ethical values. Additionally, we aim to investigate whether and how these AI systems could maintain a form of consistency in their ethical considerations in the middle of this plurality of ethical principles. | Our project aims to delve into this multifaceted ethical landscape surrounding AI from both technical and philosophical perspectives. We want to explore how AI systems deal with ethical dilemmas in the light of these diverging ethical priorities and seek methods to align these systems more closely with human ethical values. Additionally, we aim to investigate whether and how these AI systems could maintain a form of consistency in their ethical considerations in the middle of this plurality of ethical principles. |
Revision as of 10:02, 12 December 2023
Motivation
In the current era, the rise of Large Language Models (LLMs) like GPT-3 or LLAMA has evoked a mix of fascination and apprehension. These advanced models showcase remarkable capabilities of generating human-like text and performing complex tasks, while also raising profound ethical questions.
The integration of ethics into AI systems faces numerous challenges. Firstly, there is the challenge of modelling reasoning about obligations and permissions. Secondly, complexities arise from the persistent conflicts within various ethical reasonings. Lastly, comprehending and assessing the consequences of actions remains an intricate undertaking for both humans and machines.
Researchers have experimented with various techniques to address these challenges. Some have turned to deontic logics [1] and formalisms inspired by such considerations to handle the particular nature of duty rules. Others propose AI logic-based non-monotonic formalisms [2] such as default logics or answer set programming, closely aligned with common-sense reasoning, to mitigate logical contradictions. Additionally, there are proposals to employ action language or causal models [3], providing a mathematical foundation for understanding and computing action consequences.
The present technical hurdle lies in merging these three approaches into a unified framework—a framework that is nonmonotonic, adept at managing norm conflicts, and employs causal models to evaluate action consequences. These diverse approaches adopt varying normative frameworks, encompassing utilitarianism, deontology, and virtue ethics, necessitating simulation. However, the specifics of these simulations often lack clarity, particularly from a philosophical standpoint. [4]
Our project aims to delve into this multifaceted ethical landscape surrounding AI from both technical and philosophical perspectives. We want to explore how AI systems deal with ethical dilemmas in the light of these diverging ethical priorities and seek methods to align these systems more closely with human ethical values. Additionally, we aim to investigate whether and how these AI systems could maintain a form of consistency in their ethical considerations in the middle of this plurality of ethical principles.
Project Plan and Milestones
Weekly Plan
Date | Task | Completion |
---|---|---|
Week 4 |
|
√ |
Week 5 |
|
√ |
Week 6 |
|
√ |
Week 7 |
|
√ |
Week 8 |
|
√ |
Week 9 |
|
√ |
Week 10 |
|
√ |
Week 11 |
|
√ |
Week 12 |
|
|
Week 13 |
|
|
Week 14 |
|
Milestone 1
- Define Research Questions: Establish clear, focused questions to guide the project.
- Literature Review: Conduct a comprehensive review of existing studies in AI ethics.
- Ethical Theory Exploration: Investigate various ethical theories to ground your research in a solid theoretical framework.
- Ethical Dataset Identification: Locate datasets for quantitative AI ethics evaluation, such as red teaming datasets.
Milestone 2
- Refine Research Goals: Sharpen the focus and scope of the research based on initial findings.
- Dataset Finalization: Select the most appropriate dataset after exploration and evaluation.
- Model Selection and Fine-Tuning: Settle on the LLaMA model and fine-tune it by deploying GPU resources.
- Model Evaluation: Conduct a thorough evaluation of the model, focusing on its ethical implications and performance.
Milestone 3
- Develop Advanced Models: Implement Preference and Reinforcement learning models, integrating them with the fine-tuned LLaMA model.
- In-Depth Analysis: Analyze the models' outcomes, assessing performance, identifying defects, and investigating specific issues like coherence and degeneration.
- Documentation and Dissemination: Create a comprehensive Wikipedia page summarizing the project's findings.
- Final Deliverables: Compile all project materials, including a well-documented GitHub repository.
Deliverables
Methodology
Ethical Theories
Data
Model Selection
Model Fine-Tuning
Performance Evaluation
Quality Assessment
Limitations
References
- ↑ Horty, J. F. (2001). Agency and deontic logic. Oxford University Press.
- ↑ Ganascia, J. G. (2015). Non-monotonic resolution of conflicts for ethical reasoning. A Construction Manual for Robots' Ethical Systems: Requirements, Methods, Implementations, 101-118.
- ↑ Mueller, E. T. (2014). Commonsense reasoning: an event calculus based approach. Morgan Kaufmann.
- ↑ Powers, Thomas M., and Jean-Gabriel Ganascia, 'The Ethics of the Ethics of AI', in Markus D. Dubber, Frank Pasquale, and Sunit Das (eds), The Oxford Handbook of Ethics of AI (2020; online edn, Oxford Academic, 9 July 2020), https://doi.org/10.1093/oxfordhb/9780190067397.013.2