TABLE OF CONTENTS
- An Introduction to the Safe Use of AI
- PART I Analyzing Top AI Services
- PART II Evaluation of Artificial Intelligence Solutions within the University of Akureyri
- HLUTI III Updated Guidelines and Policies for the University
- PART IV Appendices and Summary Tables
- Final words: A responsible framework for safe innovation
A New Landscape of Risk: An Introduction to the Safe Use of AI
Purpose and context
This document is an appendix and a detailed elaboration of the principles set out in the "Guiding Policy on the Responsible and Ethical Use of Artificial Intelligence at the University of Akureyri" (hereinafter referred to as the "Principal Policy").
The main policy allows the use of AI tools if strict conditions are met on privacy, secure data handling, copyright, and academic integrity.
The purpose of this document is to provide the in-depth, technical and legal analysis necessary to interpret and enforce these conditions in an ever-changing technological environment. The document is intended for the entire academic community, students, teachers, researchers and administrative staff, as a fundamental guide to the safe and responsible use of this revolutionary technology. The document will be updated as needed, but is intended to reflect and ensure that users are aware and informed about the requirements for the use of AI and how different services handle the user's data.
An ever-changing risk landscape
An analysis of the top AI solutions demonstrates the fundamental risks associated with the use of large language models (LLMs). That risk lies in the fact that users' data is sent for processing on service providers' servers, which is a fundamental difference from traditional, local software. The pace of development in this field is unprecedented, and new models regularly appear on the scene, in addition to which the University of Akureyri (UNAK) has made certain tools available to the entire academic community. These developments call for a deeper and more comprehensive analysis that shifts the debate from general principles to a specific assessment supported by evidence at individual service providers and service levels.
Two cornerstones of true data security
The analysis in this document reveals a crucial conclusion: true data security in the world of generative artificial intelligence is not determined by branding, model performance, or whether the service is paid for. It is based on two inalienable pillars:
- Legally binding contract terms: Enterprise solutions that ensure, in a clear and legally binding manner, that customer data is never used to train or improve the service provider's models.
- Jurisdiction and data sovereignty: In which country and under which legislation the service provider operates and where users' data is stored. For a European university subject to the EU General Data Protection Regulation (GDPR), this is essential.
These two factors will serve as a common thread throughout the analysis and form the basis for the risk categories and practical recommendations presented.
PART I
Analyzing Top AI Services
1. The Fundamental Difference: Service Level and Data as a Commercial Product
To understand the risks associated with the use of AI, it is necessary to understand the fundamental differences that exist between the business models of service providers. These differences are reflected in three distinct levels of service: free consumer editions, paid individual subscriptions, and enterprise solutions.
Free consumer versions:
In this model, the user is not the customer, but rather is part of the product. The data that a user enters – questions, snippets, ideas – is the currency that is paid for the service. Service providers such as OpenAI (for free ChatGPT) and Google (for Gemini) reserve the right to use this data to train, refine, and develop their models. The default setting is almost without exception that data collection for training is enabled. The user must show initiative, find the right settings and opt out (e. opt-out) of this use.
Paid Individual Subscriptions (Plus/Pro):
This often creates serious misunderstandings and a false sense of security. When a user pays for a subscription like ChatGPT Plus or Gemini Advanced, they are buying performance and access to newer and more powerful models. He is not buying increased privacy. In the vast majority of cases, the same privacy terms apply to paid individual subscriptions as they do to free versions. The user still needs to opt out of data usage for training manually.
Enterprise Solutions (Enterprise/Business/API):
At this stage, the business relationship changes dramatically. Here, the customer is a company or organization that pays for services where data security is a fundamental prerequisite. The customer's data is no longer a currency but a valuable asset that should be protected. All major service providers offer legally binding contracts that ensure that a customer's data is never used to train their public models. This is not a setting that needs to be opted out; This is the core of the service.
This clear distinction is key to any risk assessment: in consumer editions, your data is assumed to be part of the development process by default; In enterprise solutions, it is assumed by default that your data is entirely your property and treated as confidential.
2. Authorized Western entities: OpenAI, Google, Anthropic, Meta, and Microsoft
Looking at the main service providers headquartered in North America and Europe (hereinafter referred to as "Western entities"), it becomes clear that they all operate according to the tripartite framework described above. Despite the competition in the performance and efficacy of models, there is consistency in their approach to data security at the enterprise level, which has created a certain gold standard in the market.
At the enterprise level, the picture is harmonized. All of these entities – OpenAI (with API, Teams, and Enterprise), Google (with Workspace and Google Cloud API), Anthropic (with API and For Work), Meta (with the Llama models through cloud services), and Microsoft (with Copilot for Microsoft 365 and Azure AI services) – offer enterprise solutions based on the following core assurances:
Contractual commitment to no training:
A clear and legally binding guarantee that a customer's data (both input and output) will not be used to train their generic models.
Data Isolation and Security Certifications:
The services are designed to isolate customer data and meet stringent global security standards, such as SOC 2 and ISO/IEC 27001.
Regulatory compliance:
The service providers commit to comply with the requirements of major data protection regulations, including the GDPR in Europe.
Powerful management tools:
Enterprise customers get access to dashboards that allow them to control employee access and monitor usage.
This coordinated approach at the company level creates a clear framework of reference and minimum requirements for all data processing for which the University of Akureyri is responsible.
3. Other Services and Jurisdictional Risks
With the advent of a number of new large-language models, especially from companies outside the traditional Western technology sector, a new dimension has been added to the risk assessment: jurisdictional risk. For an organization within the European Economic Area that is subject to the strict requirements of the GDPR, there is a legal obligation to ensure that personal data is only transferred to countries that provide comparable protection. The legal environment of the service provider's home country will therefore be a key factor in all risk assessments.
Many services, especially those operated outside of Western legal systems (e.g. in China), have terms of use that can be difficult for users within academia to accept. These terms may include:
- Broad permissions for data use: A clear right to use all input data to train and improve the models, often without the possibility of opting out.
- Data storage in insecure jurisdictions: An obligation to save data in countries where data protection laws are weaker than those required by the GDPR and where authorities may require access to data on short notice.
- Lack of enterprise solutions: There is no enterprise solution available that provides contractual assurance against the use of data for training.
Conclusion: Although curiosity about new technology is natural, the use of service providers where data storage cannot be guaranteed within the EU/EEA (or countries with an adequate data protection ruling) and where the terms and conditions allow the use of data for training, is incompatible with the security requirements of the University of Akureyri. Such tools fall into the riskiest category, as their use is entirely the responsibility of the user himself and is prohibited in all university activities.
PART II
Evaluation of Artificial Intelligence Solutions within the University of Akureyri
4. Internal Chatbot (borg.unak.is): Security Ensured by OpenAI Teams API
The University of Akureyri has developed and implemented its own chatbots that are accessible to staff and students through the University's intranet, Borg (borg.unak.is).
Verification of technical foundation
The university's borg.unak.is chatbots are based on OpenAI's APIs, specifically the services covered by OpenAI's General Business and Corporate Terms and Conditions (formerly known as Teams). This means that all interactions with these chatbots are subject to OpenAI's strictest data protection terms.
Analysis of API protection measures
Using OpenAI's business API, the university ensures the following basic protections:
- No use for training: OpenAI guarantees by legally binding means that data sent through APIs is not used to train or improve their public models.
- Data ownership: The University of Akureyri retains full ownership of the data submitted to the system and the results obtained from it.
- Strict security requirements: All data transmission is encrypted and the service is certified according to international security standards such as SOC 2 Type 2 and ISO/IEC 27001.
- Data storage and deletion: For API services, OpenAI stores data for a maximum of 30 days for the sole purpose of monitoring abuse, after which it is deleted.
Bottom Line
The University's internal chatbots on borg.unak.is are formally classified as a secure enterprise solution. They meet all the basic requirements for data protection and security. However, caution should always be exercised and data that is subject to special statutory restrictions (e.g. sensitive personal data) should be placed in any cloud system unless absolutely necessary and following a special risk assessment.
5. Specialized research tools: Scite.ai
The University of Akureyri provides all students and staff with access to Scite.ai, a powerful research tool that utilizes artificial intelligence to analyze scientific literature.
Dual functionality, unified security:
Scite.ai offers two main features:
- Citation Analysis (Smart Citations): This is the core functionality of Scite.ai. The system analyzes how published scientific journals cite other published publications. Use of this feature includes searching and analyzing publicly available, published data. The risks associated with this use are therefore negligible.
- Scite Assistant: This is the generative AI component of the service where users can submit questions or upload their own text for analysis. Here, the user submits his/her own, possibly unpublished, and sensitive data for processing.
Technical implementation analysis and data security
Scite Assistant is not a self-contained language model. The service is an intermediary that connects securely to enterprise solutions (Enterprise API) from authorized service providers such as OpenAI and Anthropic. This technical implementation is key. This means that all data that a user enters into Scite Assistant is subject to the strictest data protection terms of these backend services. The main point is that the data is not used to train OpenAI 's or Anthropic's public models, in full compliance with their terms for APIs and enterprise solutions. The data protection is therefore comparable to that guaranteed in the University's internal chatbots.
Conclusion and recommendations for the use of Scite.ai
Based on this analysis, both parts are Scite.ai considered safe for use within the academic community.
- Smart Citations: The use of this section Scite.ai is safe and strongly encouraged.
- Scite Assistant: The use of this section is safe for working with sensitive research data. As the service uses secure corporate connections, it is suitable for working with, analysing and compiling unpublished research, grant applications and other confidential data.
Since the University of Akureyri provides formal access to Scite.ai, the tool as a whole is classified as a secure and accepted solution for the academic community.
6. Leganto AI Syllabus Assistant: Data Security and Activity Analysis
Connection to the University and purpose: The Information Services and Library of the University of Akureyri is implementing a new tool, AI Syllabus Assistant, within the electronic reading list system Leganto. The system is adopted in collaboration with the University of Akureyri Centre for Teaching and Information Technology, the National Library System and Ex Libris, the vendor of Leganto. The goal is to coordinate and simplify students' access to all course materials in one place.
Activity analysis
The tool uses collections of large-scale language models to analyze lesson plans and automatically create draft e-reading lists. Teachers upload a document (e.g., Word, PDF) and the AI identifies references to course materials, extracts information, and creates an organized reading list. The system then tries to link each source to the library's collections.
Data processing and verification of security
Data security is key. Ex Libris (which is part of Clarivate) has made a clear statement about the handling of data in this process. In the frequently asked questions (FAQs) about the tool, the following is stated:
"Are you using my data to train your AI?" "No. Your data is not used to directly or indirectly train LLMs. This feature utilizes pre-trained large language models (LLM) to process your content and create a list of resources. Your input is not stored by the large language model or used for any other purpose than to build the list."
This means that HA's data is not used to train AI models. Lesson plans uploaded are sent for processing only to create the reading list and are not saved by the language models. This policy is fully compliant with the University's security requirements.
Risk factors and limitations
Despite secure data processing, there are certain risk factors that concern functionality:
- Limited accuracy: Ex Libris verifies that the tool is not 100% accurate. All lists require manual review.
- Icelandic content: The AI has primarily been tested in English. Activity with Icelandic lesson plans is the single biggest risk factor and requires thorough testing.
- Unexplained errors: The vendor acknowledges that some plans don't work without a known explanation.
Conclusion and recommendations
The Leganto AI Syllabus Assistant is a secure tool from a data protection standpoint and falls under Category 1. However, its implementation requires caution. It is recommended that:
- Define a clear procedure for all auto-generated lists to be manually reviewed.
- Perform a thorough testing phase with Icelandic material before the tool goes into general use.
- Prepare educational material that explains the advantages and limitations.
7. Other Services and Users' Responsibilities
The market for AI is full of new solutions and startups. It is important to understand that many of these solutions are not standalone large-scale language models, but are based on connecting to basic models from large Western sources such as OpenAI, Google, Meta (Facebook) or Anthropic.
Scite.ai and the Leganto AI Syllabus Assistant are examples of services that do this responsibly by using secure enterprise connections (APIs) that ensure data protection. However, there are many innovative companies in the market that do not meet these requirements. Such services may use insecure connections or have ambiguous terms that allow the use of data for training purposes.
It is therefore important that staff and students exercise caution and carefully examine the privacy terms and data security of each new tool before using it for data related to the University.
HLUTI III
Updated Guidelines and Policies for the University
8. Risk Classification System for AI Tools
To enable users to make informed and confident decisions, a four-tier risk classification system is proposed. The system is based on (1) a contractual guarantee against the use of data for training purposes and (2) the jurisdiction of the service provider.
Risk Classification System for AI Tools:
| Risk category |
Description |
Example |
Authorized data |
Features/Requirements |
| Category 1: Approved |
Centrally managed service with a corporate agreement that ensures data protection. |
borg.unak.is, Copilot for M365, Scite.ai, Leganto AI |
Any data that is not subject to specific statutory restrictions on processing or specific requirements or rules apply. |
No special action required on the part of the user. |
| Category 2: Permitted |
A business solution purchased from a department/individual. Requires confirmation that security terms are met. |
ChatGPT Teams, Claude for Work |
Professional and research data. |
The user is responsible for making sure that the terms and conditions are in order. |
| Category 3: Limited |
Consumer versions (both free and paid). Data is used for training by default and is compromised. |
Free ChatGPT, Gemini, Perplexity, etc. |
Only public, non-personally identifiable and completely non-sensitive data. |
Use is the sole responsibility of the user. It is not recommended to use such services for intellectual property and research data, unless the user has verified that the service guarantees that the data will not be used for training purposes. |
| Category 4: Unsupported / At Your Own Risk |
Services that UNAK does not endorse due to significant jurisdictional or security risks. |
Deepseek, Qwen (consumer publishing) and other services outside the EU/US. |
Only public data, user-owned data, or other non-sensitive data. |
All use is the sole responsibility of the user. The University does not support these tools and recommends against the use of a University email address (@unak.is) for registration. |
9. Practical tips for the university community
On the basis of the risk classification system, the following practical guidelines are presented.
For students:
- Use the university's tools: For all learning activities, it is best to use the tools provided by the university (Category 1), such as Copilot and Scite.ai. They're safe for you and your data.
- Protect your personal information: Remember that the use of Category 3 and 4 tools is your responsibility. Be careful never to put personally identifiable or sensitive information on such services.
- Protect your study work: Consider unfinished assignments and essays as confidential. Use of Category 3 or 4 tools for such data is your sole responsibility and is strongly discouraged.
For researchers and teachers:
- Protect intellectual property: All unpublished research papers, manuscripts, and grant applications are valuable intellectual property. The use of tools in Categories 3 and 4 to work with such data is the sole responsibility of the user and in violation of the University's policy on the protection of intellectual property.
- Choose the right category: Use Category 1 tools (e.g. borg.unak.is, Copilot, Scite.ai) for daily work. If more powerful tools are needed, ensure that you are using a Tier 2 solution (e.g., ChatGPT Team with an enterprise agreement).
- Mentor students: Teachers are responsible for educating their students about these principles and directing them to safe tools (Category 1).
For administrative staff:
- Only approved systems for sensitive data: All processing of student or staff personal data, financial information or other administrative data shall only take place in systems in Category 1.
- Responsibility for the handling of administrative data: The use of tools in Categories 3 and 4 to work with administrative data is the sole responsibility of the user. Such use constitutes a serious deviation from the University's privacy and security policy and may constitute a violation of it.
PART IV
Appendices and Summary Tables
10. Comparison Table – Data Security at a Glance
Table 1: An In-Depth Comparison of Data Security AI Services
| Líkan/Þjónusta |
Þjónustustig |
Gögn notuð til þjálfunar? (Sjálfgefið) |
Gagnaeinangrun (Fyrirtækjastig) |
Geymslustaðsetning gagna (Lögsaga) |
Helstu vottanir (Fyrirtækjastig) |
| ChatGPT (OpenAI) |
Consumer (Free/Plus) |
Yes (Can be opted out) |
No |
International (mainly US) |
N/A |
| |
Company (API/Teams/Ent.) |
No (Default) |
Yes |
US/EU (optional) |
SOC 2, ISO 27001, GDPR, HIPAA |
| Gemini (Google) |
Consumer (Free/Adv.) |
Yes (Can be opted out) |
No |
International (mainly US) |
N/A |
| |
Enterprise (Workspace/Cloud) |
No (Default) |
Yes |
International (manageable) |
SOC 2/3, ISO 27001, GDPR, HIPAA |
| Copilot (Microsoft) |
Consumer (Free/Pro) |
No (Requires approval) |
No |
International (mainly US) |
N/A |
| |
Enterprise (M365/Azure) |
No (Default) |
Yes |
International (manageable) |
Comprehensive Microsoft 365 compliance |
| Scite.ai |
All Users (Assistant) |
No (Uses OpenAI/Anthropic's secure API) |
Yes (via backend) |
United States |
Refers to backend certifications |
| Leganto AI (Ex Libris/Clarivate) |
Company (University subscription) |
No (According to terms and conditions) |
Yes |
International |
Clarivate Certifications |
| Services outside Western legal systems |
Consumer (Often a single version) |
Most of the time Yes (Cannot be opted out) |
No |
Often outside the EU/US |
Few or none |
Final words: A responsible framework for safe innovation
The analysis in this appendix conclusively confirms that there is a deep gulf between consumer publications and enterprise solutions in the world of artificial intelligence. In consumer versions, the user's data is the currency. In enterprise solutions, the user's data is a valuable asset that must be protected.
To this basic principle is added jurisdictional risks. For the University of Akureyri, which operates under European data protection legislation, this factor is an exclusion criterion. Services that cannot guarantee the handling and storage of data in accordance with the requirements of the GDPR are not considered for the processing of sensitive data.
The purpose of this document is not to curb the use of AI, but to create a safe and predictable environment where academia can take full advantage of these powerful technologies – to advance research, improve teaching, and increase efficiency – without sacrificing the fundamental values of privacy, data security, and the protection of intellectual property.
By providing clear categorization and practical guidance, the university community is better equipped to make informed decisions. The responsibility for data protection ultimately rests with each and every user. This document provides the knowledge and tools necessary to assume that responsibility professionally and determinedly.