To successfully collaborate on survey results for a literature review, your team must establish a centralized reference database, define clear inclusion criteria, and use shared tracking tools to systematically extract and synthesize data.
When conducting a systematic review or a comprehensive literature survey, working with co-authors helps reduce bias and speeds up the research process. However, without a clear workflow, you risk duplicating efforts, encountering version control issues, or losing track of critical references. Here is a practical approach to managing and sharing your collaborative literature review.
1. Define Clear Inclusion and Exclusion Criteria
Before diving into the results of your literature search, agree on a strict screening protocol. Create a shared document detailing exactly what types of studies, publication dates, geographic regions, and methodologies qualify for your review. This ensures all collaborators evaluate the surveyed papers through the same objective lens and prevents scope creep.
2. Centralize Your Reference Management
You need a single source of truth for the papers your team is reviewing. Instead of emailing messy folders of PDFs back and forth, use a cloud-based reference manager. For example, using WisPaper's My Library allows you to organize your surveyed papers in a Zotero-style manager and use AI to chat directly with your uploaded documents, making it much easier to extract key data alongside your co-authors.
3. Build a Shared Data Extraction Matrix
Once your team has screened the initial literature survey results, you need a structured way to analyze the content. Track the extracted data in a shared spreadsheet or synthesis matrix. Create standardized columns for essential metadata such as the authors, publication year, methodology, sample size, key findings, and study limitations. Cloud platforms allow multiple researchers to input data and code qualitative themes simultaneously, keeping the synthesis phase highly organized.
4. Implement Independent Screening
To maintain the rigor of your review, have at least two researchers independently screen the titles, abstracts, and full texts of the surveyed literature. Track these individual decisions—typically categorized as "include," "exclude," or "maybe"—within your shared database before comparing notes. This minimizes selection bias and ensures a high-quality final paper.
5. Resolve Conflicts Through Consensus Meetings
Disagreements on whether a paper belongs in the final review are a normal part of academic collaboration. Schedule regular check-ins to review conflicting survey results. Having a third senior researcher act as a tie-breaker can quickly resolve disputes, align the team's understanding of the research topic, and keep the writing process moving forward efficiently.

