中文    English

Journal of Library and Information Science in Agriculture ›› 2024, Vol. 36 ›› Issue (5): 23-31.doi: 10.13998/j.cnki.issn1002-1248.24-0353

Previous Articles     Next Articles

Risk of AI Algorithmic Discrimination Embedded in Government Data Governance and Its Prevention and Control

PENG Lihui1, ZHANG Qiong1, LI Tianyi2   

  1. 1. School of Public Administration of Xiangtan University, Xiangtan 411105;
    2. Network Security Detachment, Shenyang Railway Public Security Office, Shenyang 110167
  • Received:2024-04-01 Published:2024-09-24

Abstract: [Purpose/Significance] The purpose of this study is to provide an in-depth analysis of the widespread application of artificial intelligence (AI) technology in the field of government data governance and its far-reaching implications, with a particular focus on the core issue of algorithmic discrimination. With the rapid development of AI technology, it has demonstrated great potential in government decision support, public service optimization, and policy impact prediction, but it has also sparked extensive debate on issues such as algorithmic bias, privacy invasion, and fairness. Through systematic analysis, this study aims to reveal the potential risks of AI algorithms in government data governance, especially the causes and manifestations of algorithmic discrimination, and then it proposes effective solutions to protect citizens' legitimate rights and interests from being violated, and to maintain government credibility and social justice. [Method/Process] This study adopts the literature induction method to extensively collect domestic and international related data on the application of AI in government data governance, including academic papers, policy documents, and case studies. Through systematic review and in-depth analysis, we clarified the specific application scenarios of AI algorithms in government data governance and their role mechanisms. On this basis, this study further identified the key factors that led to algorithmic discrimination, including but not limited to the one-sidedness of data collection and processing, the subjective bias of the algorithm designers, and the influence of inherent social biases on the algorithms. It then explored the potential risks of algorithmic discrimination, including exacerbating social inequality, restricting civil rights, and undermining government credibility, and provided an in-depth analysis through a combination of theoretical modeling and case studies. [Results/Conclusions] The results of the study show that while the embedding of AI technology in government data governance has significantly improved the efficiency and accuracy of governance, it comes with a risk of algorithmic discrimination that cannot be ignored. To address this issue, this study proposed a series of targeted prevention and control measures, including clarifying the principle of algorithmic fairness, formulating industry norms and standards, improving the accountability mechanism and regulatory system, and optimizing the data collection and processing environment, so as to effectively curb the phenomenon of algorithmic discrimination while making full use of the advantages of AI technology, so that AI technology in government data governance can truly benefit the people, and promote social fairness and justice.

Key words: artificial intelligence (AI), data governance, algorithmic discrimination, risk prevention and control, government data, information cocoon

CLC Number: 

  • G203
[1] 谢永江, 杨永兴. 人工智能时代下的算法歧视及其治理研究[J]. 北京邮电大学学报(社会科学版), 2022, 24(5): 18-25.
XIE Y J, YANG Y X.Algorithmic discrimination and its governance in era of artificial intelligence[J]. Journal of Beijing university of posts and telecommunications (social sciences edition), 2022, 24(5): 18-25.
[2] 中华人民共和国司法部. 《生成式人工智能服务管理暂行办法》[EB/OL].[2024-03-18]. https://www.moj.gov.cn/pub/sfbgw/flfggz/flfggzbmgz/202401/t20240109_493171.html.
[3] BURRELL J.How the machine "thinks": Understanding opacity in machine learning algorithms[J]. SSRN electronic journal, 2015, 3(1): 1-12.
[4] ROMEI A, RUGGIERI S.A multidisciplinary survey on discrimination analysis[J]. The knowledge engineering review, 2014, 29(5): 582-638.
[5] 闫坤如. 人工智能的算法偏差及其规避[J]. 江海学刊, 2020(5): 141-146.
YAN K R.Algorithm deviation of artificial intelligence and its avoidance[J]. Jianghai academic journal, 2020(5): 141-146.
[6] ANGWIN J.Machine bias: There's software used across the country to predict future criminals and it's biased against blacks[EB/OL]. [2024-03-19].https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.
[7] (美)弗兰克·帕斯奎尔. 赵亚男, 译. 黑箱社会: 控制金钱和信息的数据法则[M]. 北京: 中信出版社, 2015.
PASQUALE F.The black box society the secret algorithms that control money and information[M]. Beijing: China Citic Press, 2015.
[8] BURRELL J.How the machine "thinks": Understanding opacity in machine learning algorithms[J]. Big data & society, 2016, 3(1): 205395171562251.
[9] AZIZ Z H.Racial equity in algorithmic criminal justice[J]. Duke law journal, 2019, 68(6): 1043-1134.
[10] DIAKOPOULOS N.Accountability in algorithmic decision making[J]. Communications of the ACM, 2016, 59(2): 56-62.
[11] (德)克里斯多夫·库克里克. 黄昆, 夏柯, 译. 微粒社会: 数字化时代的社会模式[M]. 北京: 中信出版社, 2018.
[12] 卜素. 人工智能中的“算法歧视”问题及其审查标准[J]. 山西大学学报(哲学社会科学版), 2019, 42(4): 124-129.
BU S.On the algorithm discriminationin in artificial intelligence and the relevant review criteria[J]. Journal of Shanxi university (philosophy and social science edition), 2019, 42(4): 124-129.
[13] CITRON D K, PASQUALE F.The scored society: Due process for automated predictions[J]. Washington law review, 2014, 89(1): 1-33.
[14] STEININGER M JUR S. German Supreme Court stands firm on liability for unjustified warning letters[J]. Journal of intellectual property law & practice, 2006, 1(4): 247-248.
[15] Nytimes. How Trump consultants exploited the Facebook data of millions[EB/OL].[2024-03-26]. https://www.nytimes.com/2018/03/17/us/politics/cambridge-analytica-trump-campaign.html?searchResultPosition=4.
[16] 贾俊平. 统计学[M]. 7版. 北京: 中国人民大学出版社, 2018.
JIA J P.Statistics[M]. 7th ed. Beijing: China Renmin University Press, 2018.
[17] CAIRE C M. When algorithms discriminate[N]. The New York Times, 2015-07-10(12).
[18] 赵汀阳. 有轨电车的道德分叉[J]. 哲学研究, 2015(5): 96-102.
ZHAO T Y.The forking paths for the trolley problem[J]. Philosophical research, 2015(5): 96-102.
[19] 柳亦博. “湖心岛”隐喻: 人工智能时代的算法伦理入侵[J]. 行政论坛, 2019, 26(6): 121-128.
LIU Y B.The isolated island metaphor: The invasion of algorithmic ethics in the era of AI[J]. Administrative tribune, 2019, 26(6): 121-128.
[20] 刘培, 池忠军. 算法的伦理问题及其解决进路[J]. 东北大学学报(社会科学版), 2019, 21(2): 118-125.
LIU P, CHI Z J.Ethical issues of algorithms and their solutions[J]. Journal of northeastern university(social science), 2019, 21(2): 118-125.
[21] VEALE M, KLEEK M V, BINNS R.Fairness and accountability design needs for algorithmic support in high-stakes public sector decisionmaking[EB/OL]. [2024-03-27].https://dl.acm.org/doi/epdf/10.1145/3173574.3174014.
[22] KAPLAN J.Humans need not apply: a guide to wealth and work in the age of artificial intelligence[M]. New Haven: Yale University Press, 2015.
[23] 张今. 版权法上“技术中立”的反思与评析[J]. 知识产权, 2008, 18(1): 72-76.
ZHANG J.Reflection and comment on "technology neutrality" in copyright law[J]. Intellectual property, 2008, 18(1): 72-76.
[24] 中共中央网络安全和信息化委员会办公室, 中华人民共和国互联网信息办公室, 中国互联网络信息中心. 《第47次中国互联网络发展状况统计报告》[EB/OL].[2024-03-27]. https://www.cac.gov.cn/2021-02/03/c_1613923423079314.htm.
[25] 刘培, 池忠军. 算法歧视的伦理反思[J]. 自然辩证法通讯, 2019, 41(10): 16-23.
LIU P, CHI Z J.Ethical reflections on algorithmic discrimination[J]. Journal of dialectics of nature, 2019, 41(10): 16-23.
[26] SKIRPAN M, GORELICK M. The authority of "fair" in machine learning[J/OL]. Arxiv preprint arxiv:1706.09976, 2017.
[27] 贾开. 人工智能与算法治理研究[J]. 中国行政管理, 2019(1): 17-22.
JIA K.Artificial intelligence and algorithm governance research[J]. Chinese public administration, 2019(1): 17-22.
[28] 伏志强, 孙伟平. 科技向“善”: 人工智能发展的价值遵循[J]. 甘肃社会科学, 2021(2): 97-103.
FU Z Q, SUN W P.Science and technology towards "goodness": The value follow of artificial intelligence development[J]. Gansu social sciences, 2021(2): 97-103.
[29] (德)尼克拉斯·卢曼. 宾凯, 赵春燕, 译. 法社会学[M]. 上海: 上海人民出版社, 2013: 188.
NICHOLAS L.Rechtssoziologie[M]. Shanghai: Shanghai People's Publishing House, 2013: 188.
[1] WANG Lin, YAO Feifei. Construction and Application of the Evaluation Indicator System of Government Data Openness Maturity in China [J]. Journal of Library and Information Science in Agriculture, 2023, 35(1): 56-72.
[2] CHANG ZhiJun, XU LiYuan, YU QianQian, ZHANG JianYong, WANG YongJi. Scientific and Technical Literature Data Management System Based on Life Cycle Model [J]. Journal of Library and Information Science in Agriculture, 2022, 34(6): 36-49.
[3] GUO Weijia. Influencing Factors of Artificial Intelligence Readiness in Libraries [J]. Journal of Library and Information Science in Agriculture, 2022, 34(5): 47-56.
[4] SONG Shanshan, BAI Wenlin. A Review of Big Data Governance Research in China [J]. Journal of Library and Information Science in Agriculture, 2022, 34(4): 4-17.
[5] ZHENG Jianming, PAN Ying. Public Cultural Data Governance System and Guarantee Measures [J]. Journal of Library and Information Science in Agriculture, 2022, 34(2): 4-13.
[6] LIU Guifeng, RUAN Bingying, LIU Qiong. Enhance Data Security Governance Capability: Interpretation of Data Security Law of the People's Republic of China (Draft) [J]. Journal of Library and Information Science in Agriculture, 2021, 33(4): 4-13.
[7] ZHANG Zhixiong, LIU Huan, YU Gaihong. Building an Artificial Intelligence Engine Based on Scientific and Technological Literature Knowledge [J]. Journal of Library and Information Science in Agriculture, 2021, 33(1): 17-31.
[8] ZHOU Zhenguo. Data Governance Research from the Perspective of Governance Framework [J]. Journal of Library and Information Science in Agriculture, 2020, 32(7): 57-62.
[9] YANG Liuqing. Application of Artificial Intelligence in Digital Microfilming [J]. Journal of Library and Information Science in Agriculture, 2020, 32(4): 59-67.
[10] JIANG Enbo, LI Na. Analysis and Evaluation of Chinese Open Government Agricultural Data [J]. Journal of Library and Information Science in Agriculture, 2020, 32(10): 4-15.
[11] TIAN Ye. Semantic Sharing Mechanism of Heterogeneous Government Data Based on Blockchain [J]. Journal of Library and Information Science in Agriculture, 2020, 32(1): 12-22.
[12] ZHENG Yin. Research on Guarantee Mechanism of Government Data Open and Utilization in Big Data Era [J]. , 2018, 30(10): 87-90.
[13] YAN Xiaojing. Study on Data Management Policies of Open Government [J]. , 2017, 29(12): 81-84.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!