The ethical AI playbook 2.0

Part 3: Data traceability

The ethical AI playbook 2.0

An ethical playbook for artificial intelligence for the real estate and construction sector compiled by the Building Information Foundation RTS and A-INS Group. The purpose of the updated playbook is to support actors in the built environment in utilizing artificial intelligence, and to promote sustainable new technologies in the sector.

“What kind of data do we produce, and how do we use it?”

Accuracy, traceability, reliability, and transparency of AI-generated data

According to Liang et al. (2024), improving trust and acceptance of AI within organizations in the AECO sector requires a focus on transparency, ethics, and the involvement of internal stakeholders. In the AECO sector, ensuring the accuracy and traceability of AI-generated data is challenging, because there are still few technical mechanisms available for this. Transparency is central in AI development, as users must understand how AI systems function, make decisions, and what data they rely on. For trust to grow, users must have the opportunity to be involved in AI-supported decision-making (Human-in-the-loop) and to evaluate the AI’s suggestions.

This section of the AI playbook for the AECO sector examines the accuracy, traceability, reliability, and transparency of AI-generated data. We review different challenges related to AI-generated outputs, such as bias, and consider how reliability can be improved in AI applications within the sector.

Current situation

The European Commission has published the AI Act (2024), one of whose objectives is to increase transparency, reliability, and user control over rights related to AI-generated data. The traceability of copyrights is a challenge, and therefore the AI Act imposes obligations on AI model providers to ensure compliance with copyright rules and requires them to produce a summary of the training data used to train the AI system. The EU AI Act defines several roles. Depending on your relationship to the AI system, you may be a provider, deployer, distributor, importer, authorised representative, or product manufacturer.

A key reliability issue for AI is the traceability and biases of the information it produces.  Generative AI offers good starting points for users to innovate ideas and explore alternatives, but it does not provide reliable, precise information for design or construction-related questions. The first commercial AI assistant in the AECO sector that maintains information traceability is Alvar, developed by Rakennustieto Oy (link). It searches for answers to user queries exclusively from RT cards, which are the AECO sector’s jointly agreed best practices for construction. The AI assistant cites the content of the RT card, and the user can verify the correct information from an approved source.

The broader operating environment of AI also includes contradictions. For example, the EU’s data policy invests in the creation of data markets and the collection of large datasets to accelerate business and public sector development. At the same time, data protection and the requirement to specify the purpose of data use aim to restrict the usability of data and ensure the protection of individuals. AI developers in the AECO sector must take these perspectives into account and balance between safeguarding individual rights and the need for extensive datasets.

Roles under the EU AI Act:

Provider” means any natural or legal person, public authority, agency or other body that develops or has developed an AI system or a general-purpose AI model and places it on the market or puts an AI system into service under its own name or trademark, whether for payment or free of charge;

Deployer” means any natural or legal person, public authority, agency or other body using an AI system under its authority, except where the AI system is used in a personal capacity for non-professional activities;

Distributor” means any natural or legal person in the supply chain, other than the provider or importer, that makes an AI system available on the Union market;

Importer” means any natural or legal person located or established in the Union who places on the market an AI system bearing the name or trademark of a natural or legal person established in a third country;

Authorised representative” means any natural or legal person located or established in the Union who has received and accepted a written mandate from the provider of an AI system or a general-purpose AI model to perform and carry out on the provider’s behalf the obligations and procedures set out in this Regulation;

Product manufacturer” places on the market or puts into service an AI system together with its product under its own name or trademark.

 

Bias and Fairness

Two of the most significant factors challenging the reliability of current AI systems are bias and fairness. According to Ferrara (2023), biases in generative AI emerge when models produce content that reflects prejudiced or unjust perspectives, often stemming from biases in the data used to train the model. Fairness, in turn, means that AI systems treat all users and situations equitably, without discrimination or favoritism. Bias in AI can lead to unfair treatment of individuals or groups and reinforce existing inequalities.

Explainability and Interpretability of AI-Generated Information

Explainability and interpretability are critical for ensuring trust, transparency, and ethical use when applying AI systems. Explainability refers to an AI system’s ability to provide reasons for its decisions or actions. This means that system should indicate to user how and why the system arrived at a particular outcome. Interpretability refers to how understandable the internal process of an AI model is. The difference between these two can be summarized as follows: interpretability relates to the model’s internal structure and how easily it can be directly understood, while explainability focuses on how well the model can offer meaningful and useful explanations of its outputs, even when the model itself is not inherently easy to understand.

Monitoring and Traceability

Establishing clear monitoring mechanisms for AI-based processes is essential to ensure that all decisions and actions taken by AI systems can be traced. Maintaining logs of AI operations and decisions enables oversight and helps ensure compliance with ethical standards. This transparency supports the identification and correction of potential errors or biases in AI-generated outputs.

Data Transparency

Data transparency means that users of an AI system must be able to easily understand how the AI arrived at its conclusion and what data it used in its decision-making process. Many AI algorithms, especially deep learning models, are opaque, and this “black box” characteristic highlighted by Liang et al. (2024) can weaken user trust and the accountability of the AI system’s owner. Filipsson (2024) suggests emerging methods in explainable AI (XAI) as a solution. These methods aim to enable AI systems to clarify how they reached a given outcome and to verify the data sources used.

Transparency is needed in AI development within the industry

Weber-Lewerenz (2021) emphasizes that the reasoning behind AI-driven design decisions in the AECO sector should be transparent and understandable to both designers and clients in order to maintain trust and enable informed adjustments. AI should complement, not replace, human judgment. Recommendations produced by AI should be evaluated critically by experts, especially in AECO sector, before they are adopted, since human safety as well as economic and environmental sustainability are at stake.

An organization in the AECO sector can demonstrate transparency by clearly communicating how the developed AI systems operate, including information about the data they use and their decision-making processes. Documenting decision-making processes through logs and audit trails is also essential to maintain and remember when developing AI systems. The adoption of explainable AI methods can help end users understand how the AI arrives at its results. Highlighting the limitations of AI systems and involving end users in their design process further enhances transparency.

The EU AI Act (2024) emphasizes increasing openness, developer accountability, and transparency. Under the Act, the data sources or datasets used to train general-purpose AI models, such as large private or public databases or archives, must be described in a summary. In addition, the Act stresses the responsibility of developers of high-risk AI systems to ensure transparency. High-risk AI systems must undergo conformity assessments and provide clear, transparent descriptions of how the systems operate.

High-risk AI systems are those that are either integrated into products subject to product-safety regulations or that may pose significant risks to human health, safety, or fundamental rights. Product-safety–regulated items include, for example, machinery that moves people, health technology, and toys. If a product already requires CE marking, any AI integrated into it is considered high-risk. Other high-risk use cases include student admissions, employment decisions, access to essential public services, elections, insurance, the operation of critical infrastructure, and other situations in which AI may make decisions that restrict human rights or endanger safety. In the AECO sector, there is still limited information available on high-risk AI systems, but likely examples include AI-controlled construction machinery and AI-based control of critical infrastructure.

According to Emaminejad et al. (2022), explainability and interpretability in AI have not been widely studied in the AECO sector. However, integrating blockchain technology with AI algorithms can make it easier to track decision-making chains, thereby improving AI explainability. This development, however, requires significant research investment before practical deployment.

Various ethical perspectives

Liang et al. (2024) highlight, in addition to transparency in AI within the sector, several ethical concerns, particularly the reduction of jobs, the importance of data privacy, information security, and conflicts in decision-making. For example, concerns about decision-making conflicts arise from the fact that AI used in design processes may not adequately consider factors that influence sustainability and user comfort, such as natural light availability in floor plans or the responsible use of materials. With current data availability, these aspects are still difficult to measure, and experiential knowledge is challenging to translate into AI algorithm behaviour. In such situations, human and AI-driven decisions may conflict, making it essential to ensure that the reasoning behind AI-assisted decisions is transparent and understandable to both designers and clients. This transparency increases trust and clarifies the conscious design choices made by professionals.

Future Outlook

Ensuring the reliability of AI is an ongoing process

Deodhar et al. (2024) emphasize that the ethical principles of AI (just like its technological components) are in constant evolution. Organisations should therefore never consider themselves “finished” when it comes to ensuring the reliability and ethical governance of AI-generated information. In the AI era, ensuring reliability is a continuous process that requires regularly updated practices within companies in the AECO sector.

Strong attention must be paid to AI security and especially to individual data privacy, as the number of cybercriminals continues to grow. From the perspective of cybercrime, the AECO sector is attractive due to the large capital flows involved in projects.

New methods will be needed in the future to ensure reliability, such as managing data biases in algorithms and introducing automated monitoring of AI-generated material. Emaminejad et al. (2022) stress that improving AI reliability requires clearer definitions of human–machine collaboration and designing AI solutions so that they integrate directly into traditional operational processes of the AECO sector.

Consider at least these!

  • Develop ethical guidelines and frameworks tailored to your own needs

Organisations in the AECO sector should develop their own ethical guidelines for the use of AI, tailored to the specific risks and challenges of their operations (for example, Yleisradio’s (The Finnish Broadcasting Company) principles for responsible AI). Ethical principles may include additional perspectives, such as AI-assisted design principles or rigorous data management practices. Data management practices may include descriptions of how data sources are updated and validated, or what the responsibility mechanisms are, or who is accountable if AI systems fail or cause harm. Ethical guidelines should be updated regularly and should address the consequences of improper AI use.

  • Ensure the reliability and transparency of data as a foundational principle in AI system design — remember the roles defined in the EU AI Act

Transparency is an essential part of AI system development from the outset, particularly regarding clear documentation of model development, data used, and decision-making processes. Stakeholders must be able to understand and, if necessary, challenge AI decisions. In AI applications used in the AECO sector, it is crucial that algorithmic decision-making is transparent and assessable, especially in safety-critical applications. Remember the obligations brought by the EU AI Act and your role within it.

  • Remember clear accountability mechanisms for handling AI-generated information

AI systems must have accountability mechanisms that define how the system is expected to operate by default. In addition, responsibilities must be clear in cases where the AI system fails or causes harm. Governance models can be used to distribute responsibilities among developers, users, and operators.


Read more

Reliability: The reliability of artificial intelligence refers in particular to the ability of systems to operate consistently and to produce dependable results. This capability is closely linked to data integrity and quality. Reliable AI systems are especially important in situations where they are used to support decision-making related to safety and project success.

Tracking: Refers to mechanisms used to monitor and record the operation and decisions of AI systems. Tracking is essential, for example, to maintain auditability and to identify potential errors or biases.

Traceability: Describes the ability to trace the entire path of an AI-driven process, including the data used, the decisions made, and the resulting actions. Maintaining traceability ensures that all AI activities are documented and can be reviewed in accordance with ethical standards.

Updatability: AI systems or models must be capable of being updated as information changes, so that they provide users with up-to-date information.

Safety: AI must be used in a way that does not compromise the safety of employees or any stakeholders. AI recommendations, especially in safety-critical situations, must be carefully assessed by human experts. Maintaining reliable data and implementing transparent processes are particularly important from a safety perspective.

Data Integrity & Accuracy / Quality: Refers to the overall correctness, consistency, and reliability of the data used by AI systems. High-quality data is essential for AI to function effectively, especially in situations where data guides critical decisions. Data validation, cleaning, and updating are key practices for maintaining data integrity.

Explainability and Interpretability: Describe the ability of AI systems to make their decision-making processes understandable to users and stakeholders. This contrasts with “black box” AI, where the logic behind decisions is opaque.

Transparency: Users and stakeholders can clearly understand how AI algorithms make decisions. This requires open communication about AI methods, data, and the logic underlying their outputs. Transparency is key to building trust in AI systems.

Fairness/Bias: Refers to the ethical obligation of AI systems to avoid favoring certain groups at the expense of others. Bias may arise from the data used, the algorithm itself, or human influence, potentially leading to discrimination. Fairness requires that AI systems are trained on diverse and representative datasets and are regularly audited for bias.

Accountability: Accountability refers to clear mechanisms that define responsibility for the actions and outcomes of AI systems. If an AI system fails or causes harm, it must be clear who is responsible. Accountability frameworks are essential to ensure that misuse of AI has consequences.

Trust: Building trust in AI technology is critically important in the construction sector. Trust requires ensuring that AI systems are reliable, accurate, safe, and aligned with user expectations. Transparency and clear communication between AI systems and their users are key to strengthening trust.

Philipsson Fredrik, 2024, Building Trust in AI Ethics, link:  https://redresscompliance.com/building-trust-in-ai-a-guide-to-ethical-considerations/, used 10.12.2024

European commission, Artificial Intelligence Act 2024/1689, link: https://eur-lex.europa.eu/legal-content/EN/TXT/HTML/?uri=OJ:L_202401689, used 10.12.2024

Ferrara, Emilio, Fairness And Bias in Artificial Intelligence: A Brief Survey of Sources, Impacts, And Mitigation Strategies, (2023), link: https://www.researchgate.net/publication/370071122_Fairness_And_Bias_in_Artificial_Intelligence_A_Brief_Survey_of_Sources_Impacts_And_Mitigation_Strategies

Ci-Jyun Liang, Thai-Hoa Le, Youngjib Ham, Bharadwaj R.K. Mantha, Marvin H. Cheng, Jacob J. Lin, Ethics of artificial intelligence and robotics in the architecture, engineering, and construction industry, Automation in Construction, Volume 162, 2024, 105369, ISSN 0926-5805, link: https://www.sciencedirect.com/science/article/pii/S0926580524001055

Weber-Lewerenz, B., Corporate digital responsibility (CDR) in construction engineering—ethical guidelines for the application of digital transformation and artificial intelligence (AI) in user practice. SN Appl. Sci. 3, 801 (2021). Link: https://doi.org/10.1007/s42452-021-04776-1

European commission, AI Act fact pages (2024), link: https://www.euaiact.com/key-issue/5, used 10.12.2024

In Finnish: Yleisradio, vastuullisen tekoälyn periaatteet, link https://yle.fi/aihe/a/20-10005659, used 10.12.2024

In Finnish: Poikola Antti, Markkanen Jouni, Parkkila Janne, Tietopolitiikan käsikirja (2024), link https://tietopolitiikka.fi/2024/11/14/kasikirja-tekoalysta-paatoksentekijoille/ , used 10.12.2024

Emaminejad, Newsha & North, Alexa & Akhavian, Reza. (2022). Trust in AI and Implications for AEC Research: A Literature Analysis. 295-303. 10.1061/9780784483893.037. Link: https://www.researchgate.net/publication/360826085_Trust_in_AI_and_Implications_for_AEC_Research_A_Literature_Analysis,

Swanand Deodhar, Favour Borokini, Ben Waber. How Companies Can Take a Global Approach to AI Ethics, Harward Business Review, 2024, link: https://hbr.org/2024/08/how-companies-can-take-a-global-approach-to-ai-ethics, used 10.12.2024