Skip to main content

Table 5 Summary of STARDIT Beta Version data fields

From: Standardised data on initiatives—STARDIT: Beta version

Section Data category Data field
Core: Initiative context—This information locates the initiative within a clear context Identifying information   Initiative name*
    Geographic location(s)*
    Purpose of the initiative (aims, objectives, goals)*
    Organisations or other initiatives involved (list all if multi-centre)*
    Relevant publicly accessible URLs/URIs
    Other identifiers (e.g. RAiD [166], clinical trial ID [167, 168])
    Keywords or metatags—including relevant search headings (e.g. MeSH [169])
    Other relevant information (free text)
  Status of initiative   What is the current state of the initiative?*
Select from:
1. Prospective—this report is prospective or describes planned activity
2. Ongoing—the initiative is still taking place
3. Completed—the initiative has finished (evaluation and impact assessment may be ongoing)
    Date range (start and end dates of initiative)
  Methods and paradigms   Methods of the initiative (what is planned to be done, or is being reported as done). Include information about any populations or eco-systems being studied, any ‘interventions’, comparators and outcome measures (qualitative or quantitative)*
If appropriate, include a link to a publicly accessible document (such as a research protocol or project plan)
    Include any information about theoretical or conceptual models or relevant ‘values’ of people involved with this initiative, including any rationale for why certain methods were chosen
Report authorship—Information about who completed the report and how
Please note this section can be completed multiple times if there are multiple authors
Identifying information for each author (authors can be anonymised in the public report but at least one verified identity will need to be sent to STARDIT Editors to attempt to prevent falsified reports)   Name*
    Publicly accessible profiles, institutional pages*
    Open Researcher and Contributor ID (orcid.org)*
    Tasks in report completion
    Other information
  Accountability   Key contact at initiative for confirming report content (include institutional email address)*
  Date   Date of report submission (automatically generated)
Input: Ethics assessment Ethics approval information (if applicable)   Assessing organisation or group*
    Approval date and approval ID—include any relevant URL
Input: Human involvement in initiative
Who is involved in this initiative and how?
Editors assessing involvement may need to use the STARDIT ‘Indicators of involvement’ tool
Details about how each group or individual was involved in the initiative   Who was involved or how would you label those involved (select from group labels or submit new group label name in free-text)*
You can name individuals or use ‘labels’ to describe groups of people such as ‘professional researchers’, ‘service users’ or ‘research participants’. Additional ‘labels’ or ‘meta-tags’ to describe people may be added if appropriate
    How many people were in each grouping label?
    Tasks of this person or group (list as many as possible)*—including any information about why certain people were included or excluded in certain tasks (such as data analysis)
    Method of doing task? How did these people complete these tasks? (what methods were used)—for example ‘group discussion’ or ‘reviewing documents’
    Communication modes? What modes of communication were used—for example, ‘group video calls’, ‘telephone interviews’ or ‘postal survey’
    How were people recruited, contacted or informed about these tasks?
  Involvement appraisal   Methods of appraising and analysing involvement (assessing rigour, deciding outcome measures, data collection and analysis)
    Enablers of involvement (what do you expect will help these people get involved—or what helped them get involved)
Examples of enablers
    Barriers of involvement (what do you expect will inhibit these people from getting involved—or what inhibited them from getting involved). Are there any known equity issues which may contribute?
Examples of barriers, and any attempts to overcome them
    How did the initiative change as a result of involving people? For example, did the initiative design or evaluation plan change?
Note: this can be answered separately for different individuals or groupings of people
  Involvement outcomes, impacts or outputs   Were there any outcomes, impacts or outputs from people being involved?* When describing these, attempt to label which groupings were affected and how. These can include impacts on people, organisations, processes or other kinds of impacts
  Learning points from involving people   What worked well, what could have been improved? Was anything learned from the process of involving these people?
  Stage   Which stage of the initiative were these people involved? (please provide information about any distinct stages of this initiative, noting some may overlap)
  Financial or other interests (including personal or professional interests)   Describe any interests (financial or otherwise), conflicting or competing interests, or how anyone involved may be personally, financially or professionally affected by the outcome of the initiative* Including any relevant information about authors of this report
Input: Material involvement in initiative
Mapping financial or other ‘interests’
Financial   What was the estimated financial cost for the initiative
    Funding information (link to publicly accessible URL if possible)—this may include the project funder, funding agreements, grants, donations, public ledgers, transaction data or relevant block(s) in a blockchain
  Time   How much time was spent on this project
Note: this can be answered separately for different individuals or groupings of people
  Other   Describe any costs or resources that cannot be measured financially or quantitatively—this may include expertise, traditional or Indigenous knowledge, volunteer time or donated resources
Outputs: Data
including code, hardware designs or other relevant information
Sensitive data Secure criteria Data adheres to relevant industry/discipline data security requirements
   Repository How is data entered, changed or removed within a repository?
   Usage Who is the data from this initiative shared with?
    Who has access to sensitive data and how is this decided?
   Safety Is data encrypted? Is it anonymised or de-identified? What methods are used for re-identification? What is the risk of unauthorised re-identification?
  Open data FAIR criteria Data adheres to FAIR criteria [170]
   Findable Describe relevant metadata, how the data is machine readable and other relevant information
   Accessible How can data be accessed—include any information about authentication and authorisation
   Interoperable How is data interoperable or integrated with other data? Include information about applications or workflows for analysis, storage, and processing, and resulting file formats or other outputs
   Reusable How can data be replicated and/or combined?
  Indigenous data CARE principles Data adheres to CARE principles [171, 172]
   Collective Benefit How will Indigenous Peoples derive benefit from the data
   Authority to Control How will Indigenous Peoples and their governing bodies determine how relevant data are represented and identified
   Responsibility How will those using the data provide evidence of these efforts and the benefits accruing to Indigenous Peoples
   Ethics How have Indigenous Peoples’ rights and wellbeing been centred during the data life cycle
  All data Hosting Where is it data stored and hosted -share any location data if appropriate
   Owner Who ‘owns’ the data or claims any kind of copyright, patent(s), or other specific types of intellectual property—include relevant open licensing information
   Analysis methods Describe methods used to analyse the data (including a link to any relevant code and information about validity)
   Usage How can data be used? Include information about license and attribution
   Dissemination How is information about this data disseminated? For example, how are results from analysis shared?
   Impact impact/effect of the output
   Data control Who controls access to the data? How are decisions about data access made? Is data anonymised or de-identified? What methods are used for re-identification? What is the risk of unauthorised re-identification? How is this risk managed?
   Management and quality Which person (or organisation) is responsible for managing (or ‘curating’) the data?
    Who is accountable for ensuring the quality and integrity of the data? (this may be an individual or organisation)
Impacts and outputs: Publications, events, changes, learning items etc What was learned   What new knowledge has been generated? (if appropriate, include effect size, relevant statistics and level or evidence)*
  Knowledge translation   Describe how the learning or knowledge generated from this initiative has or will be used
  Impacts   Have there been any outcomes, or has anything changed or happened as a result of this initiative that isn’t captured in previous answers?*
  Measurement and evaluation   How has or how will this be measured or evaluated?
    Who is involved in measuring or evaluating this?
    Who was or is involved in deciding on the outcomes used to evaluate any impacts or outcomes? How were they involved?
Information completed by Editors    
STARDIT report version number (assigned)    Report number assigned to distinguish it from any future updated reports
Indicators completed by Editors and/or peer reviewers
Editors and peer reviewers assessing the report will need to look for indicators in the following categories on publicly accessible URLs*
Indicators of involvement   Use the STARDIT ‘Indicators of involvement’ tool
  Indicators of data practice compliance   Use the relevant criteria
  Indicators of translation and impact   
  Other indicators