Human–AI Collaboration as Critical Digital Infrastructure: Hybrid Impact on Enterprise Operations and Quality Engineering
Keywords:
Human-AI Collaboration, Human-in-the-Loop, Mixed-Initiative Systems, AIOps, AgentOps, Retrieval-Augmented Generation, Quality Engineering, AI Governance, Trustworthy AIAbstract
Collaboration with AI has crossed the divide from curiosity in the lab to a core expectation of enterprise automation, especially in regulated and high-reliability systems. AI assistants based on LLMs, RAG, and agentic AIOps are all indispensable tools for engineers and support professionals when it comes to knowledge retrieval, drafting, incident triage, root cause analysis, and workflow automation. However, quality engineering teams can rely upon human-in-the-loop (HITL) collaboration to accelerate test design, understand defects, and assess safety for release. This can lead to operational hazards, including hallucinations (confabulation), automation bias, model drift, exposure of private training data, and governance issues, which can erode trust in AI outputs when they're assumed to be correct. This summarizes the state of the art in human-AI interaction models, including the HITL pipeline, mixed-initiative control sharing, and symbiotic teaming, and their supporting toolchains, benefits, and challenges. To inform future governance directions, moving on to discussing NIST AI RMF 1.0 and Generative AI Profile (NIST AI 600-1), ISO/IEC 23894, and ISO/IEC 42001. A hybrid governance approach is proposed. It introduces principles for evidence-based grounding, risk-based autonomy, and traceable decision-making. Lastly, it introduces a vision of collaborative adaptation where AI initiatives based on confidence, impact, and policy constraints maintain human accountability while achieving scalability of productivity and reliability.
Downloads
References
NIST, "AI Risk Management Framework (AI RMF 1.0)," Jan. 2023. [Online]. Available: https://www.nist.gov/itl/ai-risk-management-framework
NIST, "Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile (NIST AI 600-1)," Jul. 2024. [Online]. Available: https://doi.org/10.6028/NIST.AI.600-1
ISO, "ISO/IEC 23894:2023—Information technology—Artificial intelligence—Guidance on risk management," Feb. 2023. [Online]. Available: https://cdn.standards.iteh.ai/samples/77304/cb803ee4e9624430a5db177459158b24/ISO-IEC-23894-2023.pdf
Microsoft, "ISO/IEC 42001:2023 — Information technology—Artificial intelligence—Management system," Dec. 2023. [Online]. Available: https://learn.microsoft.com/en-us/compliance/regulatory/offering-iso-42001
Official Journal of the European Union, “Regulation (EU) 2024/1689 (Artificial Intelligence Act),” Official Journal of the European Union, Jul. 2024. [Online]. Available: https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=OJ:L_202401689
Saleema Amersh et al., "Guidelines for Human-AI Interaction," Proc. CHI, 2019. [Online]. Available: https://www.microsoft.com/en-us/research/publication/guidelines-for-human-ai-interaction/
Microsoft Research, "The Human-AI eXperience (HAX) Toolkit Project," 2021–2025. [Online]. Available: https://www.microsoft.com/en-us/research/project/hax-toolkit/
Xingjiao Wu et al., "A Survey of Human-in-the-Loop for Machine Learning," Future Generation Computer Systems, 2022 (arXiv:2108.00941). [Online]. Available: https://arxiv.org/pdf/2108.00941
Eduardo Mosqueira-Rey et al., "Human-in-the-loop machine learning: a state of the art," Artificial Intelligence Review, 2022/2023. [Online]. Available: https://link.springer.com/article/10.1007/s10462-022-10246-w
Steffen Holterr and Mennatallah El-Assady, "Deconstructing Human-AI Collaboration: Agency, Interaction, and Adaptation," Computer Graphics Forum (EuroVis), 2024. [Online]. Available: https://onlinelibrary.wiley.com/doi/pdf/10.1111/cgf.15107
George Fragiadakis et al., "Evaluating Human-AI Collaboration: A Review and Methodological Framework," arXiv:2407.19098, 2024. [Online]. Available: https://arxiv.org/html/2407.19098v1
Leila Methnani et al., "The Impact of Mixed-Initiative on Collaboration in Hybrid AI," 2024. [Online]. Available: https://umu.diva-portal.org/smash/get/diva2:1885275/FULLTEXT01.pdf
National Academies of Sciences, Engineering, and Medicine, "Human-AI Teaming: State-of-the-Art and Research Needs," 2021. [Online]. Available: https://www.sintef.no/globalassets/project/hfc/documents/2021-human-ai-interaction-26355.pdf
Sophie Berretta et al., "Defining human-AI teaming the human-centered way: a scoping review and network analysis," Frontiers in Artificial Intelligence, 2023. [Online]. Available: https://doi.org/10.3389/frai.2023.1250725
Youcef Remil et al., "AIOps Solutions for Incident Management: Technical Guidelines and a Comprehensive Literature Review," arXiv:2404.01363, 2024. [Online]. Available: https://arxiv.org/html/2404.01363v1
Yinfang Chen et al., "AIOpsLab: A Holistic Framework for Evaluating AI Agents for Enabling Autonomous Cloud (AgentOps)," Microsoft Research, 2024. [Online]. Available: https://www.microsoft.com/en-us/research/wp-content/uploads/2024/10/AIOpsLab-6705feab5dcdb.pdf
Adam Crume, "Site Reliability Engineering: Incident Management Guide," Google, 2023. [Online]. Available: https://static.googleusercontent.com/media/sre.google/en//static/pdf/IncidentManagementGuide.pdf
Betsy Beyer et al., "Site Reliability Engineering: How Google Runs Production Systems," O'Reilly, 2016. [Online]. Available: http://repo.darmajaya.ac.id/4636/1/Site%20Reliability%20Engineering_%20How%20Google%20Runs%20Production%20Systems%20%28%20PDFDrive%20%29.pdf
Ben Wiseman, "Work Trend Index Special Report: What Can AI Assistant's Earliest Users Teach Us About Generative AI at Work?" Nov. 2023. [Online]. Available: https://assets-c4akfrf5b4d3f4b7.z01.azurefd.net/assets/2023/11/Microsoft_Work_Trend_Index_Special_Report_2023_Full_Report.pdf
Najeeb Abdulhamid, et al., "Microsoft New Future of Work Report 2023," 2023. [Online]. Available: https://www.microsoft.com/en-us/research/wp-content/uploads/2023/12/NFWReport2023_v5.pdf
OpenAI, "GPT-4 Technical Report," arXiv:2303.08774, 2024. [Online]. Available: https://arxiv.org/pdf/2303.08774
Patrick Lewis et al., "Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks," NeurIPS, 2021. [Online]. Available: https://arxiv.org/pdf/2005.11401
Jason Wei et al., "Chain-of-Thought Prompting Elicits Reasoning in Large Language Models," NeurIPS, 2023. [Online]. Available: https://arxiv.org/abs/2201.11903
Shunyu Yaol., "ReAct: Synergizing Reasoning and Acting in Language Models," arXiv, 2023. [Online]. Available: https://arxiv.org/abs/2210.03629
Timo Schick et al., "Toolformer: Language Models Can Teach Themselves to Use Tools," arXiv:2302.04761, 2023. [Online]. Available: https://arxiv.org/abs/2302.04761
Xiao Liu et al., "AgentBench: Evaluating LLMs as Agents," ICLR, 2025. [Online]. Available: https://arxiv.org/abs/2308.03688
Chang Ma et al., "AgentBoard: An Analytical Evaluation Board of Multi-turn LLM Agents," NeurIPS Datasets and Benchmarks, 2024. [Online]. Available: https://proceedings.neurips.cc/paper_files/paper/2024/file/877b40688e330a0e2a3fc24084208dfa-Paper-Datasets_and_Benchmarks_Track.pdf
LangChain, "Build a RAG agent with LangChain," Documentation, 2024. [Online]. Available: https://docs.langchain.com/oss/python/langchain/rag
LangChain, "LangGraph overview," Documentation, 2024. [Online]. Available: https://docs.langchain.com/oss/javascript/langgraph/overview
Abdul Javid and Amber Welch, "AI lifecycle risk management: ISO/IEC 42001:2023 for AI governance," AWS Security Blog, May 2025. [Online]. Available: https://aws.amazon.com/blogs/security/ai-lifecycle-risk-management-iso-iec-420012023-for-ai-governance/
Downloads
Published
How to Cite
Issue
Section
License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.


