How Large Language Models Are Changing University-Level Research
Generative AI, especially large language models (LLMs) like ChatGPT, Gemini, Claude, and others, are reshaping academic research in higher education. These tools offer new opportunities but also bring challenges that students, researchers, and educators must address. Developing AI literacy and following discipline-specific guidelines are essential steps for responsible use in research.
AI in the Research Process
Recent studies show that AI-generated content is becoming common in academic writing—for example, over 13% of biomedical abstracts last year indicated AI involvement. LLMs assist with a wide range of research tasks, including:
- Brainstorming and refining research ideas
- Designing experiments and conducting literature reviews
- Writing and debugging code
- Analyzing and visualizing data
- Developing interdisciplinary frameworks
- Suggesting sources, summarizing texts, and drafting abstracts
- Presenting research findings in accessible formats
Despite these benefits, ethical and quality concerns remain. Risks such as data misrepresentation, replication difficulties, biases, privacy issues, and citation inaccuracies require careful oversight.
AI Research Assistants and Deep Research Agents
Two main categories of AI tools are emerging to support academic research:
- AI research assistants: These tools help with concept mapping (e.g., Kumu, MindMeister), literature reviews (e.g., Elicit, NotebookLM), literature search (e.g., ResearchRabbit, Scite), summarization (e.g., Scholarcy), and trend analysis (e.g., Scinapse).
- Deep research AI agents: These advanced platforms combine LLMs with retrieval-augmented generation and reasoning to conduct detailed multi-step analyses. They generate comprehensive reports with citations by synthesizing information from scholarly and online sources. Recent offerings from companies like Google Gemini and Perplexity.ai showcase these capabilities.
Open-access tools like Ai2 ScholarQA are also being developed to streamline literature reviews and improve research efficiency.
Guidelines for Responsible Use
Several guidelines encourage ethical AI use in academia:
- The Government of Canada's Guide on Generative AI advises federal institutions and academics on responsible AI use, emphasizing transparency and communication.
- Canada's Tri-Council agencies provide discipline-specific guidance for researchers receiving federal grants.
- The Observatory in AI Policies in Canadian Post-Secondary Education tracks AI policies from over 30 Canadian universities.
Supporting Interdisciplinary Research
LLMs help bridge gaps between disciplines by combining data and methods across fields. They automate data collection and analysis, facilitating collaboration between areas like biology and engineering or social sciences and climate studies. AI-powered “expert finder” platforms map researcher expertise and uncover new interdisciplinary partnerships.
Building AI Literacy in Research
Universities and research organizations are expanding AI literacy programs to equip students and faculty with the skills needed to use generative AI tools effectively. For example, the Alberta Machine Intelligence Institute offers AI education from K-12 through higher education.
Developing tailored AI literacy training focused on research applications is increasingly urgent. This includes understanding both the strengths and limitations of LLMs at each stage of the research and writing process.
For educators and researchers looking to deepen their AI knowledge and skills, exploring practical AI training courses can be valuable. Consider checking resources such as Complete AI Training’s latest courses to stay current with AI tools relevant to academic research.
Your membership also unlocks: