Skip to main content

Main menu

  • HOME
  • ARTICLES
    • Current Issue
    • Ahead of Print
    • Archives
    • Abstracts In Press
    • Special Issue Archive
    • Subject Collections
  • INFO FOR
    • Authors
    • Reviewers
    • Call For Papers
    • Subscribers
    • Advertisers
  • SUBMIT
    • Manuscript
    • Peer Review
  • ABOUT
    • The JABFM
    • The Editing Fellowship
    • Editorial Board
    • Indexing
    • Editors' Blog
  • CLASSIFIEDS
  • Other Publications
    • abfm

User menu

Search

  • Advanced search
American Board of Family Medicine
  • Other Publications
    • abfm
American Board of Family Medicine

American Board of Family Medicine

Advanced Search

  • HOME
  • ARTICLES
    • Current Issue
    • Ahead of Print
    • Archives
    • Abstracts In Press
    • Special Issue Archive
    • Subject Collections
  • INFO FOR
    • Authors
    • Reviewers
    • Call For Papers
    • Subscribers
    • Advertisers
  • SUBMIT
    • Manuscript
    • Peer Review
  • ABOUT
    • The JABFM
    • The Editing Fellowship
    • Editorial Board
    • Indexing
    • Editors' Blog
  • CLASSIFIEDS
  • JABFM on Bluesky
  • JABFM On Facebook
  • JABFM On Twitter
  • JABFM On YouTube
OtherAbout Practice-Based Research Networks

Evaluating Primary Care Research Networks: A Review of Currently Available Tools

Joan M. C. Bleeker, Wim A. B. Stalman and Henriëtte E. van der Horst
The Journal of the American Board of Family Medicine July 2010, 23 (4) 465-475; DOI: https://doi.org/10.3122/jabfm.2010.04.090297
Joan M. C. Bleeker
MSc
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Wim A. B. Stalman
MD, PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Henriëtte E. van der Horst
MD, PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • References
  • Info & Metrics
  • PDF
Loading

Abstract

Background: An increasing number of primary care research networks (PCRNs) are being developed around the world. Despite the fact that they have existed for a long time in some countries, little is known about what they have actually achieved. There is an ongoing debate in the literature about the appropriate framework for the evaluation of PCRNs. Here, we aim to provide an overview of the tools that are currently available for measuring the performance of PCRNs and practices involved in PCRNs or research.

Methods: We performed electronic searches in bibliographic databases and several additional searches. We composed a checklist to evaluate the design, content, and methodological quality of the tools.

Results: We identified 4 tools for the evaluation of PCRNs or the measurement of primary care practices involved in PCRNs or research.

Conclusions: The results of our study showed that various methods, areas of interest, dimensions, and indicators for the evaluation of PCRNs have been proposed. However, no generic and validated tool that enables meaningful comparison between different network models has been developed. It is, therefore, time to reflect on the appropriateness and effectiveness of PCRNs and determine the desired outcomes (ends) of PCRNs and how we can best achieve them in the future (means). To open up the “black box” of the effectiveness of the PCRNs, it may be relevant to observe the effects of network and research participation on those involved in networks.

  • Practice-based Research
  • PBRN
  • Evaluation

Increasing numbers of primary care research networks (PCRNs) are being developed around the world.1–4 In some countries their number is steadily increasing. In the United States, approximately 100 PCRNs are currently active,5,6 whereas a decade earlier only 28 PCRNs could be identified.6 PCRNs initially were developed to provide an infrastructure for primary care research, to strengthen and develop the research base of primary care, and to increase research capacity and research activity among primary care providers.7–11

Current networks operate in different ways, depending on aspects such as contexts, baseline of experience, and resources. Therefore, PCRNs represent a great diversity of network models, objectives, and configurations.12,13 In general, PCRNs aim to conduct research close to practice, identify and respond to current primary care research needs, and to engage more primary care professionals with research. In addition to these research-based objectives, it has been suggested that PCRNs can also play an important role in the translation and diffusion of research findings into practice14–16 and other activities that are relevant in the field of primary care, like quality improvement, training, and education.17–22

Internationally, both policy advisers and researchers have acknowledged the importance of PCRNs and have recognized the challenging opportunities they provide for the improvement of primary care. Yet, despite their long existence in some countries like The Netherlands and the United Kingdom, little is known about the effectiveness of PCRNs23–26 or their added value compared with other research capacity–building initiatives in primary care. Another issue that has not yet been studied is the impact of network and research involvement in daily practice in terms of increased use of research evidence and quality improvement. There seems to be a fundamental lack of evaluative studies about the progress, impact, and output of PCRNs, which prevents a validated comparison of different network models and thus insight into what type of model is optimal for achieving specific goals.

Valid and agreed methods of evaluation are urgently needed to provide stakeholders and policy advisers with comprehensive and meaningful data about these issues. Therefore, we reviewed the literature on PCRNs to identify and describe currently available frameworks and tools for evaluating PCRNs and practices involved in PCRNs or research. The design, content, and methodological quality of the identified frameworks and tools were also reviewed to provide more detailed information about their usefulness.

Methods

Search Strategy

We used the search strategy described in Table 1 to search for relevant publications in the computerized bibliographic database of Medline (PubMed). Subsequently, we searched Embase and Picarta for additional publications. We also screened the references of relevant articles, contacted experts and network organizations, and searched the Internet (Google) for additional publications. Initially, we conducted a manual search in the Dutch scientific journal of general practice (Huisarts en Wetenschap) to identify additional publications concerning Dutch PCRNs.

View this table:
  • View inline
  • View popup
Table 1.

Key Words and Search Strategy

Inclusion Criteria

Two authors (JB and HvdH) independently screened the titles and abstracts of all citations and selected articles for full-text reading. Each author selected articles for the preliminary selection by using the following inclusion criteria: (1) publications that discussed the evaluation of PCRNs; (2) publications that presented or discussed a (conceptual) framework or tool for measuring the performance of PCRNs or practices involved in PCRNs or research; and (3) studies that reported the results of a (comparative) PCRN evaluation study. This preliminary selection enabled us to review any method, framework, or tool that had been previously proposed, developed, or used. In the final selection we included only those articles that provided information regarding a (conceptual) framework or tool used for measuring the performance of PCRNs or practices involved in PCRNs or research. After each selection both reviewers met to discuss the differences in their article selection and establish consensus. We excluded editorials, letters, and comments.

Data Extraction and Quality Assessment

We composed a provisional checklist (see Table 2) to review the identified frameworks and tools and to detect shortcomings and gaps in knowledge of their methodological quality. To compose this checklist we contacted experts in psychometric research who had been previously involved in the development of a checklist for assessing the methodological quality of studies about measurement properties of health status measurement instruments.27,28

View this table:
  • View inline
  • View popup
Table 2.

Checklist to Evaluate Methodological Quality of Primary Care Research Networks

Two reviewers (JB and HvdH) independently rated the information that was provided for each framework and tool based on the checklist. Disagreements between the reviewers were discussed and resolved during a consensus meeting.

Description of the Checklist Dimensions and Items

Content Validity

When a new tool is constructed to measure a particular concept, such as the “productivity” or “effectiveness” of PCRNs, it is important that its measurement areas and items capture all relevant and meaningful aspects of that concept. In psychometric research this is called “establishing content validity.”27–30 Content validity can be established by appropriate literature support, involvement of experts, and agreement among subject-matter experts about the relevance of the proposed areas and items to be measured.29,30 To judge content validity we determined whether the authors had reviewed relevant literature, involved experts in the development and item selection/reduction process, and/or reported outcomes of a content validity study, such as the level of agreement between subject-matter experts.

We considered content validity as one of the most important aspects of methodological quality. Only if the content validity of a tool is adequate will one consider using the tool, and evaluation of the other aspects of methodological quality is useful.

Reliability

A prerequisite for the establishment of reliability is that the tool itself and the measurement methods are adequately described and standardized. Reliability can further be established by evaluating the degree of agreement between different observers (interobserver reliability) or the agreement between observations made by the same observer/rater on 2 different occasions (intra-observer reliability).29,30

To assess reliability we determined whether the tools and measurement methods were described in sufficient detail to permit replication. For the tools that were tested or applied, we determined whether the authors reported on interobserver and/or intra-observer reliability.

Feasibility

If a new tool is considered to measure all meaningful aspects of the concept of interest, the feasibility of the tool has to be established.29 We therefore determined whether the tools were tested in a pilot study or had been otherwise applied, for instance for a (comparative) PCRN evaluation survey.

Practice Investment

Because the usefulness of a tool is also determined by practical and financial aspects, we also evaluated the information provided concerning time investment29 and costs.

Description of the Tools

Descriptive data extracted from the publications included the level of measurement (network/practice), the concept of measurement, and the content of the tools in terms of domains of interest/proposed measurement areas and the number of proposed items.

Results

We performed 2 searches in Medline. The first search, which was conducted in July 2005, resulted in 631 hits. We selected 69 articles for full-text reading (62 unanimous, 7 based on consensus); 5 articles from additional searches were included. For the preliminary selection we included 12 articles (8 unanimous, 4 based on consensus) and for the final selection 5 articles (unanimous decision).26,31–34 An update of our search in Medline was conducted in May 2009, which this resulted in 1 additional article for the final selection.35

From the 6 articles we included in the final selection, 2 described a tool kit for the evaluation of PCRNs, developed by Fenton et al,34,35 and 1 article concerned a framework for the evaluation of PCRNs, developed by Clement et al.32 Furthermore, we identified 2 articles that described a tool, developed by Doorn et al,26,31 for measuring practice performance/output in academic general practice networks, and 1 article that described an accreditation scheme for primary care practices involved in research, developed by Carter et al.33

Identified Frameworks/Tools for Evaluating PCRNs

  1. In 2001, Clements et al32 developed a framework to evaluate the effectiveness of PCRNs. They considered PCRNs to be effective if they were capable of achieving predefined and agreed network objectives. They proposed 7 relevant and agreed-upon objectives (see Table 3) for the evaluation of PCRNs, and for each objective a set of example indicators for measurement; 4 objectives in the framework referred to the research function of PCRNs, 1 to a better use of research findings, and 2 to network building (see Table 3).

  2. In 2007, Fenton et al35 presented a tool kit to evaluate the organizational design, management, and appropriateness of PCRNs. The tool kit was generated from the evaluation of 5 PCRNs in the United Kingdom.34 Fenton et al35 identified 8 organizational dimensions and 28 subdimensions (see Table 3) that they considered to be relevant for PCRNs and their potential productivity (defined as “creating ideas and knowledge, or intellectual capital”). The focus of the tool kit is on the structures and processes within PCRNs and not on measurement of output. In the evaluation study the (sub)dimensions were rated on a 5-point scale, varying from 0 = not present (the dimension is not applicable) to 4 = very strong (the dimension is very strongly present).35

View this table:
  • View inline
  • View popup
Table 3.

Description of the Identified Tools for the Evaluation of Primary Care Research Networks

Identified Frameworks/Tools for Practices Involved in PCRNs or Research

  1. In 1999, Doorn et al26,31 developed a tool with which to measure the extent of academic activity and the quality of health care in general practices involved in a university-linked general practice network: Huisartsgeneeskundige Academiserings Lineaal Maastricht (HALMA). They identified 3 practice performance/output areas that they considered to be relevant for the effectiveness of PCRNs and proposed 70 expert-validated indicators for evaluating practice performances/output in networks (see Table 4). The tool includes indicators for measuring practice involvement in and contributions to research and registration, the quality of care provided by participating practices, and their involvement in and contributions to academic medical teaching and education.

  2. In 2002, the Royal College of General Practitioners (RCGP) in the United Kingdom introduced the Primary Care Research Team Assessment scheme (PCRTA).33 This scheme provides both primary care teams and stakeholders (such as funding bodies) with a mechanism for assessing the research infrastructure within a practice against professionally developed and tested standards. The scheme presents 7 areas of interest for assessing the quality of the research infrastructure in practices involved in research (see Table 4) and distinguishes 2 levels of accreditation: collaborator research practices and investigator-led research practices. For both levels the scheme describes essential and desirable quality indicators. In addition to focusing on aspects such as strategic planning, research resources, and infrastructure, the scheme also addresses aspects such as education of other primary care professionals and the dissemination of research findings. A further description of the identified tools, in terms of concept of interest, proposed dimensions, and number of items, is provided in Table 3 and 4.

View this table:
  • View inline
  • View popup
Table 4.

Description of the Identified Tools for Practices Involved in Primary Care Research Networks or Research

Methodological Assessment of the Identified Frameworks/Tools

Content Validity

All authors reported that they had reviewed PCRN literature to identify earlier research or theoretical concepts. However, both Clement and Fenton had already stated literature to support the development of methods for the evaluation of PCRNs, and their content validity is scarce.32,34 Clement's literature search failed to identify any relevant studies focusing on the evaluation of PCRNs. The set of objectives in Clements’ framework was partially based on the findings of a national survey among 22 network coordinators36 in the United Kingdom. Harvey, Fenton, and Sturt34,35 performed a review of the organizational science literature on networks to identify those network dimensions that create social and intellectual capital for networks.

The involvement of professionals/stakeholders in the development of the tools was reported by Carter (PCRTA),33 Doorn (HALMA),26,31 and Fenton (tool kit).34,35 National key figures and stakeholders were involved in the PCRTA development and evaluation process. General practitioners who were participating in a Dutch academic general practice network validated the proposed HALMA indicators. For the latter, Doorn at al31 also published verifiable outcomes of the level of agreement for each item. Participants in research networks were involved in the development and refining of Fenton's tool kit.34,35

Reliability

For 2 tools only—the PCRTA (Carter) and the tool kit (Fenton)—the items and measurement methods were described in sufficient detail to permit replication, and these 2 tools met our requirements for standardization.33–35 Carter and colleagues33 reported that the reliability of the PCRTA assessment was (qualitatively) evaluated by an independent researcher after the pilot study. Outcomes of intrareliability or inter-reliability tests were not reported.

Feasibility

Two tools were tested in a pilot study (PCRTA and HALMA), and 1 tool (Fenton's tool kit) was used once for a comparative case study of 5 PCRNs in the United Kingdom. One tool (Clements’ framework) has neither been tested nor applied. Although Doorn et al performed a pilot study the information they provided (an unpublished article about the results of this feasibility study) was not sufficient to enable us to complete our assessment of the methodological quality of HALMA.

Practice Investment

We could only evaluate time investment and costs for the PCRTA. A sample timetable for the assessment visit was presented whereas time investment in the preparation of documentary evidence for leading general practitioners and administrative support were evaluated for practices that participated in the pilot study. The costs for accreditation could be derived from the website of the RCGP in the United Kingdom. A summary of our qualitative assessment of the checklist items is provided in Table 5.

View this table:
  • View inline
  • View popup
Table 5.

Qualitative Assessment of Information Provided for the Checklist Items

Discussion

We identified 4 tools for the evaluation of PCRNs or measurement of primary care practices involved in PCRNs or research. For evaluating the “overall” performance of PCRNs we identified 2 tools: Clement's32 objective-based framework and Fenton et al's34,35 PCRNs tool kit. The content validity of the proposed objectives and indicators in Clement's objective-based tool seems to be questionable when set against the criteria of our checklist, and because the tool has not been tested or otherwise applied, the feasibility of the indicators is also questionable. However, incorporating the objectives of the network in the framework is a strength of Clement's tool. Information about objectives is needed to identify appropriate dimensions and input and output indicators for measurement and evaluation. The second tool we identified for the evaluation of PCRNs, Fenton's tool kit, was generated from a contextualized case study of 5 PCRNs in the United Kingdom and subsequently refined. The data were collected at horizontal (time) and vertical levels (individual, group, organization) of analysis. Fenton et al also incorporated exogenous factors in the evaluation (like institutional and geographical data). This case study provides insight into the early stages of network development because Fenton generated a basic set of generic input indicators of network configuration and management. This set enables us to make meaningful comparisons between different network models from the perspective of congruency between objectives, organization, and management inputs. The underlying assumption is that if these 3 are well honed, productive outputs will follow.34,35 However, the data collection and the analyses seem to be rather time consuming. To evaluate the practice investment, more information on the costs and time investment is needed.

For the measurement of the development and performance of practices involved in PCRNs or research we identified 2 tools: HALMA (Doorn)26,31 and the PCRTA (Carter).33 In addition to providing research-based indicators, both tools also provide a set of indicators to measure practice outputs in the education of professionals (PCRTA)33 and academic medical education/teaching (HALMA).26,31 HALMA also provides a set of indicators to measure practice improvement and quality of care over time.

For the PCRTA33 we were able to make the most complete qualitative assessment of methodologic quality. The information provided about the development process was sufficient to evaluate the content validity of the tool, the description of the scheme and the measurement methods met our criteria for standardization, a pilot study was performed and reported, and we could also evaluate the practice investment in terms of time investment and current costs of accreditation. The scheme provides a dataset of qualitative and quantitative indicators that can be used by PCRNs to observe the (possible) effects of network participation on practice/professional development and research performance over time. It can also be used as a benchmark for the quality of the research infrastructure of participating practices. The RCGP in the United Kingdom has acknowledged the scheme and introduced it in 2001. However, despite the low costs of accreditation for primary care practices in England and Wales, the number of practices to date that have gained accreditation as a collaborator practice (9 practices) or an investigator research practice (20 practices) is still relatively small.37 This may indicate that the accreditation process is more time consuming or difficult for practices than the results of the pilot study suggested, or, alternatively, that the benefits gained from accreditation are too small.

For HALMA we could only assess its content validity. Doorn et al31 did not provide sufficient information to make it possible to evaluate other aspects of its methodologic quality. We found the number of indicators for measuring the quality of care rather low, and most of these indicators differ from the indicators that are currently used to measure the quality of care (in The Netherlands). HALMA provides a basic set of expert-validated indicators for assessing the extent of research and academic medical teaching/education in PCRNs. The Dutch PCRNs explicitly aim to map research and teaching activity in everyday practice. However, for PCRNs that do not have this focus, it might also be interesting to assess the involvement in education/teaching of practices/professionals involved in PCRNs. An earlier study, conducted by Gray et al,38 showed that general practices in the United Kingdom with a leading role in research (as can be expected from practices involved in PCRNs) are more involved in teaching than in teaching general practices involved in research. PCRNs might well also prove to be an effective tool for increasing teaching capacity in primary care.

In the literature about PCRNS there is an ongoing debate about the appropriate framework to measure the effectiveness and added value of PCRNs. Evaluation involves assessing the success of an intervention against a set of indicators or a body of criteria based on the targets and underlying principles in relation to the initiative. Some of the targets and underlying principles of PCRNs are reflected in the design of the identified frameworks and tools. However, most of the identified frameworks and tools were proposed and developed in a relative early state of network development, and ideas about the effectiveness, targets, and principles of PCRNs may have evolved over time. We think it is necessary to (1) again discuss the goals, such as desired impact and outcomes, of PCRNs and the best ways to achieve these goals in the future; and (2) to reflect on the appropriateness, effectiveness, and added value of PCRNs. The questions are, Why should we invest in PCRNs? In what ways are PCRNs more effective and efficient than other research capacity building initiatives? Can PCRNs address certain issues that are relevant for the future of primary care better than through other approaches? Any framework or tools for internal or comparative evaluation should also consider this perspective.

Our goal was to provide an overview of the tools that are currently available for the evaluation of PCRNs and practices involved in PCRNs or research. We studied both the methodologic quality of the tools and their areas of interest. The results of our study show that various methods, areas of interest, structures, processes, output dimensions, and indicators for the evaluation of PCRNs have been proposed and identified. However, no generic and validated tool that enables meaningful comparison between different network models has yet been developed. In our opinion, Fenton's tool kit and contextualized case study of PCRNs might provide a basis for a generic framework/tool. To open the “black box” of the effectiveness of PCRNs it may also be relevant to monitor and evaluate the development and performances of practices and professionals over time, as suggested by Doorn.26 The PCRTA and HALMA might provide a basis for a generic framework/tool to observe the effects of network and research participation on those involved in networks.

Strengths and Limitations of Our Study

We conducted a very extensive search of several databases, contacted experts in the relevant fields, and also screened the so-called “gray literature.” We are confident that we did not miss any relevant frameworks or tools for the evaluation of PCRNs or practices involved in PCRNs or research.

Regrettably, there is no accepted checklist that can be used to review the methodological quality of such complex tools. Therefore, we contacted experts in this domain to discuss relevant quality criteria to review the identified frameworks/tools and to detect shortcomings and gaps in knowledge of their methodologic quality. We have provided comprehensive information about all quality items so that readers can make their own assessment of the methodologic quality and state of development of the identified frameworks and tools. We acknowledge that our checklist might not be comprehensive and some of the quality criteria (like inter-reliability) might be arbitrary for this type of tools. Further discussion about the relevant quality criteria for evaluating the methodological quality of this type of tools is needed.

Acknowledgments

We like to thank our colleagues at the EMGO Institute and specifically Henrica C.W. de Vet and Sandra D.M. Bot for their time and support in the development of the provisional checklist we used to review the identified framework and tools.

Notes

  • This article was externally peer reviewed.

  • Funding: none.

  • Conflict of interest: none declared.

  • See Related Commentary on Page 440.

  • Received for publication December 25, 2009.
  • Revision received April 14, 2010.
  • Accepted for publication April 19, 2010.

References

  1. ↵
    van Weel C, Smith H, Beasley JW. Family practice research networks. Experiences from 3 countries. J Fam Pract 2000; 49: 938–43.
    OpenUrlPubMed
  2. Laurence COM, Beilby JJ, Marley JE, Newbury J, Wilkinson D, Symon B. Establishing a practice-based primary care research network. The University Family Practice Network in South Australia. Aust Fam Physician 2001; 30: 508–12.
    OpenUrlPubMed
  3. Thomas P, Griffiths F, Kai J, O'Dwyer A. Networks for research in primary health care. BMJ 2001; 322: 588–90.
    OpenUrlFREE Full Text
  4. Abyad A. Research networks. MEJFM 2007; 5: 55–7.
    OpenUrl
  5. ↵
    Tierny WM, Oppenheimer CC, Hudson BL, et al. A national survey of primary care practice-based research networks. Ann Fam Med 2007; 5: 242–50.
    OpenUrlAbstract/FREE Full Text
  6. ↵
    Green LA. A short history of primary care practice-based research networks: from concept to essential research laboratories. J Am Board Fam Med 2006; 19: 1–10.
    OpenUrlFREE Full Text
  7. ↵
    Nutting PA, Green LA. Practice-based research networks: reuniting practice and research around the problems most of the people have most of the time. J Fam Pract 1994; 38: 335–6.
    OpenUrlPubMed
  8. Nutting PA. Practice-based research networks: building the infrastructure of primary care research. J Fam Pract 1996; 42: 199–203.
    OpenUrlPubMed
  9. Nutting PA, Beasley JW, Werner, JJ. Practice-based research networks answer primary care questions. JAMA 1999; 281: 686–8.
    OpenUrlCrossRefPubMed
  10. van Weel C. General Practice research networks: gateway to primary care evidence. Med J Aust 2002; 177: 62–3.
    OpenUrlPubMed
  11. Lindbloom EJ, Ewigman BG, Hickner JM. Practice-based research networks—the laboratories in primary care research. Med Care 2004; 42: 45–9.
    OpenUrl
  12. ↵
    Griffiths F, Wild A, Harvey J, et al. The productivity of primary care research networks. Br J Gen Pract 2000; 50: 913–5.
    OpenUrlAbstract/FREE Full Text
  13. ↵
    Fenton E, Harvey J, Griffiths F, Wild A, Sturt J. Reflections from organizational science on the development of primary health care research networks. Fam Pract 2001; 18: 540–4.
    OpenUrlAbstract/FREE Full Text
  14. ↵
    Genel M, Dobs A. Translating clinical research into practice: practice-based research networks—a promising solution. J Investg Med 2003; 51: 64–71.
    OpenUrl
  15. Cohen DJ, Crabtree BF, Etz RS, et al. Fidelity versus flexibility: translating evidence-based research into practice. Am J Prev Med 2008; 35(5 Suppl): S381–9.
    OpenUrl
  16. Thomas P, While A. West London Research Network. Increasing research capacity and changing the culture of primary care towards reflexive inquiring practice. J Interprol Care 2001; 15: 133–9.
    OpenUrl
  17. ↵
    Metsemakers JF, Knottnerus JA, van Schendel GJ, Kocken RJ, Limonard CB. Unlocking patients’ records in general practice for research, medical education and quality assurance: the Registration Network Family Practices. Int J Biomed Comput 1996; 42: 43–50.
    OpenUrlCrossRefPubMed
  18. Interfaculty Council of General Practice. Report on the results of the Dutch academic practices networks 1992–1997 [Dutch]. Nijmegen: Department of General Practice, University of Nijmegen; 1998.
  19. Thomas P, Kai J, O'Dwyer A, Griffiths F. Primary care groups and research networks: opportunities for R&D in context. Br J Gen Pract 2000; 50: 91–2.
    OpenUrlFREE Full Text
  20. Edwards JB, Oppewal S, Logan CL. Nurse-managed primary care: outcomes of a faculty practice network. J Am Ac of Nurse Pract 2003; 15: 563–9.
    OpenUrl
  21. Mold JW, Peterson KA. Primary care practice-based research networks: working at the interface between research and quality improvement. Ann Fam Med 2005; 3(Suppl 1): S12–S20.
    OpenUrl
  22. Thomas P, Graffy J, Wallace P, Kirby M. How primary care networks can help integrate academic and service initiatives in primary care. Ann Fam Med 2006; 4: 235–9.
    OpenUrlAbstract/FREE Full Text
  23. ↵
    Carter YH, Shaw S, Sibbald B. Primary care research networks: an evolving model meriting national evaluation. Br J Gen Pract 2000; 50: 859–60.
    OpenUrlFREE Full Text
  24. Gunn JM. Should Australia develop primary care research networks? Med J Aust 2002; 177: 63–6.
    OpenUrlPubMed
  25. Stock N, Smith H. Primary care research networks for Australia? Austr Fam Phys 2002; 31: 95–6.
    OpenUrl
  26. ↵
    Doorn B, Kocken RJJ, Crebolder HFJM, Dinant GJ, Knottnerus JA. The development of an tool for academic general practices [Dutch]. Huisarts Wet 1999; 42: 57–60.
    OpenUrl
  27. ↵
    Terwee CB, Bot SDM, de Boer MR, et al. Quality criteria were proposed for measurement properties of health status questionnaires. J Clin Epidemiol 2007; 60: 34–42.
    OpenUrlCrossRefPubMed
  28. ↵
    Mokkink LB, Terwee CB, Patric DL, et al. The COSMIN checklist for assessing the methodological quality of studies on measurement properties of health status measurement instruments: an international Delphi study. Qual Life Res 2010; 19: 539–49.
    OpenUrlCrossRefPubMed
  29. ↵
    Bot SDM, Terwee CB, van der Windt DAW, Bouter LM, Dekker J, de Vet HCW. Clinimetric evaluation of shoulder disability questionnaires: a systematic review of the literature. Ann Rheum Dis 2004; 63: 335–41.
    OpenUrlAbstract/FREE Full Text
  30. ↵
    Streiner DL, Norman GR. Health measurement scales: a practical guide to their development and use. New York, NY: Oxford University Press; 2003.
  31. ↵
    Doorn B, Kocken RJJ, Crebolder HFJM, Dinant GJ, Knottnerus JA. HALMA: validating a tool for academic general practices [Dutch]. Huisarts Wet 1999; 42: 267–70.
    OpenUrl
  32. ↵
    Clement S, Pickering A, Rowland G, Thiry K, Candy B, de Lusignan S. Towards a conceptual framework for evaluating primary care research networks. Br J Gen Pract 2000; 50: 651–2.
    OpenUrlAbstract/FREE Full Text
  33. ↵
    Carter YH, Shaw S, Macfarlane F. Primary Care Research Team Assessment (PCRTA): development and evaluation. Occas Pap R Coll Gen Pract 2002; (81): iii-vi, 1–72.
    OpenUrl
  34. ↵
    Harvey J, Fenton E, Sturt J. Evaluation of primary care R&D networks in North Thames region. Final report. Coventry, West Midlands, UK: Centre for Creativity, Strategy and Change, Warwick Business School. University of Warwick; 2002.
  35. ↵
    Fenton E, Harvey J, Sturt J. Evaluating primary care research networks. Health Serv Manage Res 2007; 20: 162–73.
    OpenUrlAbstract/FREE Full Text
  36. ↵
    National Health Services. Executive R&D in primary care: National Working Group report (Mant Report). London: Department of Health; 1997.
  37. ↵
    Royal College of General Practitioners. Royal College of General Practicioners –investigator led accredited practices. Available at http://www.rcgp.org.uk/clinical_and_research/circ/research/primary_care_research_team_ass/investigator_led_accredited_pr.aspx. Accessed 5 November 2009.
  38. ↵
    Gray S. Mapping teaching and research activity in general practice. Med Teach 2000; 22: 64–9.
    OpenUrl
PreviousNext
Back to top

In this issue

The Journal of the American Board of Family Medicine: 23 (4)
The Journal of the American Board of Family Medicine
Vol. 23, Issue 4
July-August 2010
  • Table of Contents
  • Table of Contents (PDF)
  • Cover (PDF)
  • Index by author
  • Back Matter (PDF)
  • Front Matter (PDF)
Print
Download PDF
Article Alerts
Sign In to Email Alerts with your Email Address
Email Article

Thank you for your interest in spreading the word on American Board of Family Medicine.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Evaluating Primary Care Research Networks: A Review of Currently Available Tools
(Your Name) has sent you a message from American Board of Family Medicine
(Your Name) thought you would like to see the American Board of Family Medicine web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
4 + 2 =
Solve this simple math problem and enter the result. E.g. for 1+3, enter 4.
Citation Tools
Evaluating Primary Care Research Networks: A Review of Currently Available Tools
Joan M. C. Bleeker, Wim A. B. Stalman, Henriëtte E. van der Horst
The Journal of the American Board of Family Medicine Jul 2010, 23 (4) 465-475; DOI: 10.3122/jabfm.2010.04.090297

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Share
Evaluating Primary Care Research Networks: A Review of Currently Available Tools
Joan M. C. Bleeker, Wim A. B. Stalman, Henriëtte E. van der Horst
The Journal of the American Board of Family Medicine Jul 2010, 23 (4) 465-475; DOI: 10.3122/jabfm.2010.04.090297
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Methods
    • Results
    • Discussion
    • Acknowledgments
    • Notes
    • References
  • Figures & Data
  • References
  • Info & Metrics
  • PDF

Related Articles

  • What Do We Get From Participating in Practice-based Research Networks?
  • PubMed
  • Google Scholar

Cited By...

  • What Do We Get From Participating in Practice-based Research Networks?
  • Google Scholar

More in this TOC Section

  • Practice-based Research Networks (PBRNs) Bridging the Gaps between Communities, Funders, and Policymakers
  • Lessons Learned from Developing a Patient Engagement Panel: An OCHIN Report
  • Lessons from Initiating the First Veterans Health Administration (VA) Women's Health Practice-based Research Network (WH-PBRN) Study
Show more About Practice-Based Research Networks

Similar Articles

Navigate

  • Home
  • Current Issue
  • Past Issues

Authors & Reviewers

  • Info For Authors
  • Info For Reviewers
  • Submit A Manuscript/Review

Other Services

  • Get Email Alerts
  • Classifieds
  • Reprints and Permissions

Other Resources

  • Forms
  • Contact Us
  • ABFM News

© 2025 American Board of Family Medicine

Powered by HighWire