Abstract
Using a large sample of faculty members of Canadian business schools, this article attempts to shed new light on the efficiency of academic research as measured, at the researcher’s level, by the peer-reviewed article counts and citations. Metrics on outputs from the Web of Science and from the Google Scholar databases, augmented by a survey data on factors explaining the productivity and impact performances of these faculty members, are used to assess their academic research efficiency and to perform an empirical investigation of the determinants of researchers’ efficiency, using the two-stage Bootstrap DEA approach. Results reveal that there is substantial room for improvements of technical efficiency, both across the eight fields considered in this study, and within each field. The analyses also enabled to identify determinants that might explain the academic efficiency gap between scholars across the eight research fields considered in this study, notably certification from independent agencies, seniority, sources of funding, affiliation to a business school with a doctoral program, and prestige and reputation of university of affiliation.


Similar content being viewed by others
Notes
A journal indexed in the Web of Science is used to be called an ISI journal. ISI means the Institute for Scientific Information, which developed and produced the Science Citation Index (SCI), Social Sciences Citation Index (SSCI), and the Arts and Humanities Citation Index (AHCI).
Bowen and Schuster (1986) describe the faculty role as encompassing instruction, research, public service, and institutional governance and operation (e.g., administration).
Detailed steps to obtain unbiased efficiency scores together with confidence intervals can be found in Simar and Wilson (2000) p. 788–791. In our analysis, the computation of efficiency scores was performed with Wilson’s FEAR 2.0 software (2008), and the truncated regression models were performed in R.
We employed algorithm 2 from Simar and Wilson (2007), pp. 42–43.
Confidence intervals are obtained from 1000 bootstrapping replications. As a robustness check, we also tried 2000 replications. The change of the number of bootstrap replications did not have a substantive impact on the results.
References
Abbott, M., & Doucouliagos, C. (2009). Competition and efficiency: Overseas students and technical efficiency in Australian and New Zealand universities. Education Economics, 17(1), 31–57.
Abramo, G., Cicero, T., & D’Angelo, C. A. (2011). A field-standardized application of DEA to national-scale research assessment of universities. Journal of Informetrics, 5(4), 618–628.
Abramo, G., Cicero, T., & D’Angelo, C. A. (2013). Individual research performance: A proposal for comparing apples to oranges. Journal of Informetrics, 7(2), 528–539.
Abramo, G., D’Angelo, C. A., & Costa, F. (2012). Identifying interdisciplinarity through the disciplinary classification of coauthors of scientific publications. Journal of the Association for Information Science and Technology, 63(11), 2206–2222.
Abramo, G., D’Angelo, C. A., Costa, F. D., & Solazzi, M. (2009). University–industry collaboration in Italy: A bibliometric examination. Technovation, 29(6–7), 498–507.
Abramo, G., D’Angelo, C. A., & Murgia, G. (2016). The combined effects of age and seniority on research performance of full professors. Science and Public Policy, 43(3), 301–319.
Abramo, G., D’Angelo, C. A., & Murgia, G. (2017). The relationship among research productivity, research collaboration, and their determinants. Journal of Informetrics, 11(4), 1016–1030.
Abramo, G., D’Angelo, C. A., & Pugini, F. (2008). The measurement of Italian universities research productivity by a non-parametric bibliometric methodology. Scientometrics, 76(2), 225–244.
Abramo, G., D’Angelo, C. A., & Reale, E. (2019). Peer review versus bibliometrics: Which method better predicts the scholarly impact of publications? Scientometrics, 121(1), 537–554.
Adams, J., & Griliches, Z. (1996). Measuring science: An exploration. Proceedings of the National Academy of Sciences, 93(23), 12664–12670.
Adler, N. J., & Harzing, A. W. (2009). When knowledge wins: Transcending the sense of and nonsense of academic rankings. Academy of Management Learning and Education, 8(1), 72–95.
Adrienne, C., & Fisher, D. (2008). The exchange university: Corporatization of academic culture. Vancouver: UBC Press.
Agasisti, T., Arnaboldi, M., & Azzone, G. (2008). Strategic management accounting in universities: The Italian experience. Higher Education, 55(1), 1–15.
Agasisti, T., Catalano, G., Landoni, P., & Verganti, R. (2012). Evaluating the performance of academic departments: An analysis of research-related output efficiency. Research Evaluation, 21(1), 2–14.
Agasisti, T., & Johnes, G. (2009). Beyond frontiers: Comparing the efficiency of higher education decision-making units across more than one country. Educations Economics, 17(1), 59–79.
Agasisti, T., & Johnes, G. (2010). Heterogeneity and the evaluation of efficiency: The case of Italian universities. Applied Economics, 42(11), 1365–1375.
Agasisti, T., Munda, G., & Hippe, R. (2019). Measuring the efficiency of European education systems by combining data envelopment analysis and multiple-criteria evaluation. Journal of Productivity Analysis, 51(2–3), 105–124.
Agasisti, T., & Pérez-Esparrells, C. (2010). Comparing efficiency in a cross-country perspective: The case of Italian and Spanish state universities. Higher Education, 59(1), 85–103.
Agasisti, T., & Pohl, C. (2012). Comparing German and Italian public universities: Convergence or divergence in the higher education landscape? Managerial and Decision Economics, 33(2), 71–85.
Agasisti, T., & Salerno, C. (2007). Assessing the cost efficiency of Italian universities. Education Economics, 15(4), 455–471.
Agasisti, T., & Wolszczak-Derlacz, J. (2016). Exploring efficiency differentials between Italian and Polish universities, 2001–11. Science and Public Policy, 43(1), 128–142.
Aksnes, D. W., Langfeldt, L., & Wouters, P. (2019). Citations, citation indicators, and research quality: An overview of basic concepts and theories. SAGE Open. https://doi.org/10.1177/2158244019829575.
Allison, P. D., & Long, J. S. (1990). Departmental effects on scientific productivity. American Sociological Review, 55, 469–478.
Altbach, P. (2006). The dilemmas of ranking. International Higher Education, 42, 2–3.
Amara, N., & Landry, R. (2012). Counting citations in the field of business and management: Why use Google Scholar rather than the Web of Science. Scientometrics, 93(3), 553–581.
Amara, N., Landry, R., & Halilem, N. (2013). Faculty consulting: Between formal and informal knowledge transfer. Higher Education, 65(3), 359–384.
Amara, N., Landry, R., & Halilem, N. (2015). What can university administrators do to increase the publication and citation scores of their faculty members? Scientometrics, 103(2), 489–530.
Amara, N., Rhaiem, M., & Halilem, N. (2019). Assessing research efficiency of Canadian scholars in the management field: Evidence from DEA and fsQCA analyses. Journal of Business Research. https://doi.org/10.1016/j.jbusres.2019.10.059.
Arza, V. (2010). Channels, benefits and risks of public–private interactions for knowledge transfer: Conceptual framework inspired by Latin America. Science and Public Policy, 37(7), 473–484.
Assaf, A., & Matawie, K. M. (2010). Improving the accuracy of DEA efficiency analysis: A bootstrap application to the health care foodservice industry. Applied Economics, 42(27), 3547–3558.
Avkiran, N. K. (2001). Investigating technical and scale efficiencies of Australian universities through data envelopment analysis. Socio-Economic Planning Sciences, 35(1), 57–80.
Baccini, A., Banfi, A., De Nicolao, G., & Galimberti, P. (2015). University ranking methodologies. An interview with Ben Sowter about the Quacquarelli Symonds World University Ranking. RT. A Journal on Research Policy and Evaluation, 3(1), 1–8.
Baldini, N., Grimaldi, R., & Sobrero, M. (2006). Institutional changes and the commercialization of academic knowledge: A study of Italian universities’ patenting activities between 1965 and 2002. Research Policy, 35(4), 518–532.
Baldini, N., Grimaldi, R., & Sobrero, M. (2007). To patent or not to patent? A survey of Italian inventors on motivations, incentives, and obstacles to university patenting. Scientometrics, 70(2), 333–354.
Banker, R. D., Charnes, A., Cooper, W. W., & Maindiratta, A. (1988). A comparison of DEA and translog estimates of production frontiers using simulated observations from a known technology. In Applications of modern production theory: Efficiency and productivity (pp. 33–55). Dordrecht: Springer.
Bellas, M. L., & Toutkoushian, R. K. (1999). Faculty time allocations and research productivity: Gender, race, and family effects. Review of Higher Education, 22(4), 367–390.
Bennis, W. G., & O’Toole, J. (2005). How business schools lost their way. Harvard Business Review, 83(5), 96–104.
Bercovitz, J., & Feldman, M. (2004). Academic entrepreneurs: Social learning and participation in university technology transfer. Minneapolis: Hubert H. Humphrey Institute of Public Affairs, University of Minnesota.
Bercovitz, J., & Feldman, M. (2008). Academic entrepreneurs: Organizational change at the individual Level. Organisation Science, 19(1), 69–89.
Biscaro, C., & Giupponi, C. (2014). Co-authorship and bibliographic coupling network effects on citations. PLoS ONE, 9(6), e99502.
Blumenthal, D., Campbell, E., Anderson, M., Causino, N., & Seashore-Louis, K. (1996). Participation of life-science faculty in research relationships with industry. New England Journal of Medicine, 335(23), 1734–1739.
Bogetoft, P., & Otto, L. (2010). Benchmarking with DEA, SFA, and R (Vol. 157). Berlin: Springer.
Bolli, T., Olivares, M., Bonaccorsi, A., Daraio, C., Aracil, A. G., & Lepori, B. (2016). The differential effects of competitive funding on the production frontier and the efficiency of universities. Economics of Education Review, 52, 91–104.
Bonaccorsi, A., & Cicero, T. (2016). Nondeterministic ranking of university departments. Journal of Informetrics, 10(1), 224–237.
Bonaccorsi, A., & Daraio, C. (2003). A robust nonparametric approach to the analysis of scientific productivity. Research Evaluation, 12(1), 47–69.
Bonaccorsi, A., Daraio, C., & Simar, L. (2006). Advanced indicators of productivity of universities. An application of robust nonparametric methods to Italian data. Scientometrics, 66(2), 389–410.
Bonaccorsi, A., Daraio, C., & Simar, L. (2014). Efficiency and economies of scale and scope in European universities. A directional distance approach. Technical Paper, 8, 1–31.
Bowen, H. R., & Schuster, J. H. (1986). American professors: A national resource imperiled. Oxford University Press.
Bozeman, B., & Corley, E. (2004). Scientists’ collaboration strategies: Implications for scientific and technical human capital. Research Policy, 33(4), 599–616.
Bozeman, B., Dietz, J. S., & Gaughan, M. (2001). Scientific and technical human capital: An alternative approach to R&D evaluation. International Journal of Technology Management, 22(8), 716–740.
Bozeman, B., & Gaughan, N. (2007). Impacts of grants and contracts on academic researchers’ interactions with industry. Research Policy, 36(5), 694–707.
Burt, R. S. (1997). The contingent value of social capital. Administrative Science Quarterly, 42, 339–365.
Butler, L. (2003). Explaining Australia’s increased share of ISI publications—the effects of a funding formula based on publication counts. Research Policy, 32(1), 143–155.
Butler, L., & Visser, M. S. (2006). Extending citation analysis to non-source items. Scientometrics, 66(2), 327–343.
Buzzigoli, L., Giusti, A., & Viviani, A. (2009). The Evaluation of university departments: A case study for Firenze. International Advances in Economic Research, 16(1), 24–38.
Carayol, N., & Matt, M. (2004a). Does research organization influence academic production? Laboratory level influence from a large European University. Research Policy, 33(8), 1081–1102.
Carayol, N., & Matt, M. (2004b). The exploitation of complementarities in scientific production process at the laboratory level. Technovation, 24(6), 455–465.
Carayol, N., & Matt, M. (2006). Individual and collective determinants of academic scientists’ productivity. Information Economics and Policy, 18(1), 55–72.
Cattaneo, M., Meoli, M., & Signori, A. (2016). Performance-based funding and university research productivity: The moderating effect of university legitimacy. The Journal of Technology Transfer, 41(1), 85–104.
Chan, K. C., Chen, C. R., & Cheng, L. T. (2006). A ranking of accounting research output in the European region. Accounting and Business Research, 36(1), 3–17.
Chang, P. L., & Hsieh, P. N. (2008). Bibliometric overview of operations research/management science research in Asia. Asia-Pacific Journal of Operational Research, 25(2), 217–241.
Charnes, A., Cooper, W. W., & Rhodes, E. (1978). Measuring the efficiency of decision-making units. European Journal of Operational Research, 2(6), 429–444.
Cheng, S. (2015). A comparison of compliance and aspirational accreditation models: Recounting a university’s experience with both a Taiwanese and an American accreditation body. Higher Education, 70(6), 1017–1032.
Cherchye, L., & Abeele, P. V. (2005). On research efficiency: A micro-analysis of Dutch university research in economics and business management. Research Policy, 34(4), 495–516.
Clark, M. J., Hartnett, R. T., & Baird, L. J. (1976). Assessing dimensions of quality in graduate education. Princeton, NJ: Educational Testing Service.
Clermont, M. (2016). Effectiveness and efficiency of research in Germany over time: An analysis of German business schools between 2001 and 2009. Scientometrics, 108(3), 1347–1381.
Cohen, L., Duberley, J., & McAuley, J. (1999). Fueling discovery or monitoring productivity: Research scientists’ changing perceptions of management. Organization, 6(3), 473–497.
Coleman, J. C. (1990). Foundations of social theory. Cambridge: The Belknap Press of Harvard University Press.
Cooper, W. W., Seiford, L. M., & Zhu, J. (2011). Handbook on data envelopment analysis (2nd ed.). New York: Springer.
Copper, W. W., Seiford, L. M., & Tone, K. (2000). Data envelopment analysis: A comprehensive text with models, applications, references and DEA-solver software. Boston: Kluwer Academic Publishers.
Crespi, G., & Guena, A. (2008). An empirical study of scientific production: A cross-country analysis, 1981–2002. Research Policy, 37(4), 565–579.
Daraio, C., Bonaccorsi, A., & Simar, L. (2015). Rankings and university performance: A conditional multidimensional approach. European Journal of Operational Research, 244(3), 918–930.
de Janasz, S. C., & Forret, M. L. (2008). Learning the art of networking: A critical skill for enhancing social capital and career success. Journal of Management Education, 32(5), 629–650.
De Winter, J. C. F. (2015). The relationship between tweets, citations, and article views for PLOS ONE articles. Scientometrics, 102(2), 1773–1779.
De Witte, K., Rogge, N., Cherchye, L., & Van Puyenbroeck, T. (2013). Accounting for economies of scope in performance evaluations of university professors. Journal of the Operational Research Society, 64(11), 1595–1606.
Deem, R. (2001). Globalisation, New Managerialism, Academic Capitalism and Entrepreneurialism in Universities: Is the local dimension still important? Comparative education, 37(1), 7–20.
Dehon, C., McCathie, A., & Verardi, V. (2010). Uncovering excellence in academic rankings: A closer look at the Shanghai ranking. Scientometrics, 83(2), 515–524.
D’Este, P., & Patel, P. (2007). University–industry linkages in the UK: What are the factors underlying the variety of interactions with industry? Research Policy, 36(9), 1295–1313.
D’Este, P., & Perkmann, M. (2011). Why do academics engage with industry? The entrepreneurial university and individual motivations. The Journal of Technology Transfer, 36(3), 16–339.
Diamond, A. M. (1984). An economic-model of the life-cycle research productivity of scientists. Scientometrics, 6(3), 189–196.
Dietz, J. S., & Bozeman, B. (2005). Academic careers, patents, and productivity: Industry experience as scientific and technical human capital. Research Policy, 34(3), 349–367.
Dillman, D. A. (2000). Mail and internet surveys: The tailored design methods (2nd ed.). New York: Wiley.
Dillman, D. A., & Bowker, D. K. (2001). The web questionnaire challenge to survey methodologists. Online social sciences, 53–71.
Docampo, D. (2011). On using the Shanghai ranking to assess the research performance of university systems. Scientometrics, 86(1), 77–92.
Docampo, D. (2013). Reproducibility of the Shanghai academic ranking of world universities results. Scientometrics, 94(2), 567–587.
Duan, S. X. (2019). Measuring university efficiency: An application of data envelopment analysis and strategic group analysis to Australian universities. Benchmarking: An International Journal, 26(4), 1161–1173.
Duque, R. B., Ynalvez, M., Sooryamoorthy, R., Mbatia, P., Dzorgbo, D. B. S., & Shrum, W. (2005). Collaboration paradox: Scientific productivity, the Internet, and problems of research in developing areas. Social Studies of Science, 35(5), 755–785.
Durand, R., & Mcguire, J. (2005). Legitimating agencies in the face of selection: The case of AACSB. Organization Studies, 26(3), 165–196.
Engwall, L. (2007). The anatomy of management education. Scandinavian Journal of Management, 23(1), 4–35.
Etzkowitz, H. (1983). Entrepreneurial scientists and entrepreneurial universities in American academic science. Minerva, 21, 198–233.
Fairweather, J. S. (1993). Faculty rewards reconsidered: The nature of tradeoffs. Change: The Magazine of Higher Learning, 25(4), 44–47.
Farshad, M., Sidler, C., & Gerber, C. (2013). Association of scientific and nonscientific factors to citation rates of articles of renowned orthopedic journals. European Orthopedics and Traumatology, 4(3), 125–130.
Fedderke, J. W. (2013). The objectivity of national research foundation peer review in South Africa assessed against bibliometric indexes. Scientometrics, 97(2), 177–206.
Fedderke, J. W., & Goldschmidt, M. (2015). Does massive funding support of researchers work? Evaluating the impact of the South African research chair funding initiative. Research Policy, 44(2), 467–482.
Feldman, M. P., Feller, I., Bercovitz, J. E., & Burton, R. M. (2002). University technology transfer and the system of innovation. In Institutions and systems in the geography of innovation (pp. 55–77). Boston, MA: Springer.
Field, A. (2009). Discovering statistics using SPSS (3rd ed.). Thousand Oaks: SAGE.
Finkelstein, M. J., Walker, E., & Chen, R. (2013). The American faculty in an age of globalization: Predictors of internationalization of research content and professional networks. Higher Education, 66(3), 325–340.
Fox, M. F. (1992). Research, teaching and publication productivity: Mutuality versus competition in academia. Sociology of Education, 65, 293–305.
Franceschet, M. (2010). A comparison of bibliometric indicators for computer science scholars and journals on Web of Science and Google Scholar. Scientometrics, 83(1), 243–258.
Gaddis, S. E. (1998). How to design online surveys. Training and Development, 52(6), 67–72.
Ganley, A., & Cubbin, J. S. (1992). Public sector efficiency measurement: Application of data envelopment analysis. Amsterdam: Elsevier.
Gaughan, M., & Bozeman, B. (2002). Using curriculum vitae to compare some impacts of NSF research grants with research center funding. Research Evaluation, 11(1), 17–26.
Geuna, A., & Nesta, L. (2006). University patenting and its effects on academic research: The emerging European evidence. Research Policy, 35(6), 790–807.
Geuna, A., & Piolatto, M. (2016). Research assessment in the UK and Italy: Costly and difficult, but probably worth it (at least for a while). Research Policy, 45(1), 260–271.
Gibbons, M., Limoges, C., Nowotny, H., Schwartzman, S., Scott, P., & Trow, M. (1994). The new production of knowledge: The dynamics of science and research in contemporary societies. London: Sage.
Gralka, S., Wohlrabe, K., & Bornmann, L. (2019). How to measure research efficiency in higher education? Research grants vs. publication output. Journal of Higher Education Policy and Management, 41(3), 322–341.
Grimpe, C. (2012). Extramural research grants and scientists’ funding strategies: Beggars cannot be choosers? Research Policy, 41(8), 448–1460.
Groot, T., & Garcia-Valderrama, T. (2006). Research quality and efficiency—An analysis of assessments and management issues in Dutch economics and business research programs. Research Policy, 35(9), 1362–1376.
Halilem, N., Amara, N., & Landry, R. (2011). Is the academic Ivory Tower becoming a managed structure? A nested analysis of the variance in activities of researchers from natural sciences and engineering in Canada. Scientometrics, 86(2), 431–448.
Halkos, G. E., Tzeremes, N. G., & Kourtzidis, S. A. (2012). Measuring public owned university departments’ efficiency: A bootstrapped DEA approach. Journal of Economics and Econometrics, 55(2), 1–24.
Hammarfelt, B., & De Rijcke, S. (2015). Accountability in context: Effects of research evaluation systems on publication practices, disciplinary norms, and individual working routines in the faculty of Arts at Uppsala University. Research Evaluation, 24(1), 63–77.
Harley, S. (2002). The impact of research selectivity on academic work and identity in UK universities. Studies in Higher Education, 27(2), 187–205.
Harzing, A. W. (2007). Publish or Perish, available from http://www.harzing.com/pop.htm.
Harzing, A. W., & Alakangas, S. (2016). Google Scholar, Scopus and the Web of Science: A longitudinal and cross-disciplinary comparison. Scientometrics, 106(2), 787–804.
Hazelkorn, E. (2011). Rankings and the reshaping of higher education. The battle for world-class Excellence. London: Palgrave MacMillan.
Hazeltorn, E. (2008). Learning to live with league tables and ranking: The experience of institutional leaders. Higher Education Policy, 21(1), 193–215.
Hedrick, D. W., Henson, S. E., Krieg, J. M., & Wassell, C. S., Jr. (2010). The effects of AACSB accreditation on faculty salaries and productivity. Journal of Education for Business, 85(5), 284–291.
Hicks, D. (1999). The difficulty of achieving full coverage of international social science literature and the bibliometric consequences. Scientometrics, 44(2), 193–215.
Hicks, D., Wouters, P., Waltman, L., de Rijcke, S., & Rafols, I. (2015). The Leiden Manifesto for research metrics. Nature, 520(7548), 429–431.
Hopkins, K. D., Gollogly, L., Ogden, S., & Horton, R. (2002). Strange results mean it’s worth checking ISI data. Nature, 415(6873), 732.
Iorio, R., Labory, S., & Rentocchini, F. (2017). The importance of pro-social behaviour for the breadth and depth of knowledge transfer activities: An analysis of Italian academic scientists. Research Policy, 46(2), 497–509.
Jacob, B. A., & Lefgren, L. (2011). The impact of research grant funding on scientific productivity. Journal of Public Economics, 95(9–10), 1168–1177.
Jacsó, P. (2008). The plausibility of computing the h-index of scholarly productivity and impact using reference-enhanced databases. Online Information Review, 32(2), 266–283.
Johnes, G., & Johnes, J. (2016). Costs, efficiency, and economies of scale and scope in the English higher education sector. Oxford Review of Economic Policy, 32(4), 596–614.
Johnes, J. (2004). Efficiency measurement. In G. Johnes & J. Johnes (Eds.), International handbook on the economics of education (pp. 613–743). Cheltenham: Edward Elgar.
Johnes, J. (2006). Data envelopment analysis and its application to the measurement of efficiency in higher education. Economics of Education Review, 25(3), 273–288.
Judge, W. Q., Weber, T., & Muller-Kahle, M. I. (2012). What are the correlates of interdisciplinary research impact? The case of corporate governance research. Academy of Management Learning and Education, 11(1), 82–98.
Julian, S. D., & Ofori-Dankwa, J. C. (2005). Is accreditation good for the strategic decision making of traditional business schools. Academy of Management Learning and Education, 5(2), 225–233.
Kademani, B. S., Kumar, V., Surwase, G., Sagar, A., Mohan, L., Kumar, A., et al. (2007). Research and citation impact of publications by the chemistry division at Bhabha atomic research centre. Scientometrics, 71(1), 25–57.
Katharaki, M., & Katharakis, G. (2010). A comparative assessment of Greek universities efficiency using quantitative analysis. International Journal of Educational Research, 49(4–5), 115–128.
Keith, B., & Babchuk, N. (1998). The quest for institutional recognition: A longitudinal analysis of scholarly productivity and academic prestige among sociology departments. Social Forces, 76(4), 1495–1533.
Kempkes, G., & Pohl, C. (2010). The efficiency of German universities—Some evidence from nonparametric and parametric methods. Applied Economics, 42(16), 2063–2079.
Khatri, N., Ojha, A. K., Budhwar, O., Srinivasan, V., & Varma, A. (2012). Management research in India: Current state and future directions. IIMB Management Review, 24(2), 104–115.
Khurana, R. (2007). From higher aims to hired hands. Princeton, NJ: Princeton University Press.
Kieser, A., Nicolai, A., & Seidl, D. (2015). The practical relevance of management research: Turning the debate on relevance into a rigorous scientific research program. The Academy of Management Annals, 9(1), 143–233.
Kincl, T., Novák, M., & Štrach, P. (2013). A cross-cultural study of online marketing in international higher education—A keyword analysis. New Educational Review, 32(2), 49–65.
Korhonen, P., Tainio, R., & Wallenius, J. (2001). Value efficiency analysis of academic research. European Journal of Operational Research, 130(1), 121–132.
Kou, M., Zhang, Y., Zhang, Y., Chen, K., Guan, J., & Xia, S. (2020). Does gender structure influence R&D efficiency? A regional perspective. Scientometrics, 122(1), 477–501.
Kounetas, K., Anastasiou, A., Mitropoulos, P., & Mitropoulos, I. (2011). Departmental efficiency differences within a Greek university: An application of a DEA and Tobit analysis. International Transactions in Operational Research, 18(5), 545–559.
Kuo, J. S., & Ho, Y. S. (2008). The cost efficiency impact of the university operation fund on public universities in Taiwan. Economics of Education Review, 27(5), 603–612.
Kyvik, S. (1989). Productivity differences fields of learning, and Lotka’s law. Scientometrics, 15(3–4), 205–214.
Kyvik, S. (1991). Productivity in academia. Oslo: Scientific Publishing at Norwegian Universities Universitetsforlaget.
Lahiri, S. (2011). India-focused publications in leading international business journals. Asia Pacific Journal of Management, 28(2), 427–447.
Lahiri, S., & Kumar, V. (2012). Ranking international business institutions and faculty members using research publication as the measure. Management International Review, 52(3), 317–340.
Landry, R., Saihi, M., Amara, N., & Ouimet, M. (2010). Evidence on how academics manage their portfolio of knowledge transfer activities. Research Policy, 39(10), 1387–1403.
Lariviere, V., Macaluso, B., Archambault, E., & Gingras, Y. (2010). Which scientific elites? On the concentration of research funds, publications and citations. Research Evaluation, 19(1), 45–53.
Lee, S., & Bozeman, B. (2005). The impact of research collaboration on scientific productivity. Social Studies of Science, 35(5), 673–702.
Lehmann, S., Jackson, A., & Lautrup, B. (2008). A quantitative analysis of indicators of scientific performance. Scientometrics, 76(2), 369–390.
Levin, S. G., & Stephan, P. E. (1991). Research productivity over the life cycle: Evidence for academic scientists. American Economic Review, 81, 114–132.
Leydesdorff, L., & Shin, J. C. (2011). How to evaluate universities in terms of their relative citation impacts: Fractional counting of citations and the normalization of differences among disciplines. Journal of the American Society for Information Science and Technology, 62(6), 1146–1155.
Li, Y., Chen, Y., Liang, L., & Xie, J. (2012). DEA models for extended two-stage network structures. Omega, 40(5), 611–618.
Littell, J. H., Corcoran, J., & Pillai, V. (2018). Systematic reviews and meta-analysis. New York: Oxford University Press.
Louis, K. S., Blumenthal, D., Gluck, M., & Stoto, M. A. (1989). Entrepreneurs in academic: An exploration of behaviors among life scientists. Administrative Science Quarterly, 34(1), 110–131.
Lowe, R. A., & Gonzalez-Brambila, C. (2007). Faculty entrepreneurs and research productivity. Journal of Technology Transfer, 32(3), 173–194.
Lu, W. M. (2012). Intellectual capital and university performance in Taiwan. Economic Modelling, 29(4), 1081–1089.
Lukman, R., Krajnc, D., & Glavic, P. (2010). University ranking using research, educational and environmental indicators. Journal of Cleaner Production, 18(7), 619–628.
Macilwain, C. (2013). Halt to avalanche of performance metrics. Nature, 500(7462), 255.
Mangematin, V., & Baden-Fuller, C. (2008). Global contests in the production of business knowledge: Regional centres and individual business schools. Long Range Planning, 41(1), 117–139.
Marginson, S. (2005). There must be some way out of here. Tertiary Educ. Management Conference, Keynote address, Perth, Australia.
Martin, E. (2006). Efficiency and quality in the current higher education context in Europe: An application of the data envelopment analysis methodology to performance assessment of departments within the University of Zaragoza. Quality in Higher Education, 12(1), 57–79.
McDowell, J. M. (1982). Obsolescence of knowledge and career publication profiles: Some evidence of differences among fields in costs of interrupted careers. American Economic Review, 72, 752–768.
McMillan, M. L., & Datta, D. (1998). The Relative efficiencies of Canadian universities: A DEA perspective. Canadian Public Policy, 24(4), 485–511.
Medin, E., Anthun, K. S., Häkkinen, U., Kittelsen, S. A., Linna, M., Magnussen, J., et al. (2011). Cost efficiency of university hospitals in the Nordic countries: A cross-country analysis. The European Journal of Health Economics, 12(6), 509–519.
Menard, S. (1995). Applied logistic regression analysis. In Sage university paper series on quantitative application in the social sciences, series no. 106 (2nd ed.). Thousand Oaks, CA: Sage.
Merigó-Lindahl, J. M. (2012). Bibliometric analysis of business and economics in the Web of Science. Studies in Fuzziness and Soft Computing, 287, 3–18.
Mingers, J., & Lipitakis, E. (2010). Counting the citations: A comparison of Web of Science and Google Scholar in the field of management. Scientometrics, 85(2), 613–625.
Mingers, J. C., & Lipitakis, E. A. (2014). A bibliometric comparison of the research of three UK business schools. In Proceedings of the international multiconference of engineers and computer scientists (Vol. II), IMECS 2014, March 12–14, 2014, Hong Kong.
Mishra, V., & Smyth, R. (2013). Are more senior academics really more research productive than junior academics? Evidence from Australian law schools. Scientometrics, 96(2), 411–425.
Moed, H. F. (2002). The impact factors debate: The ISI’s uses and limits. Nature, 415(6873), 731–732.
Moore, W. J., Newman R. J., Sloane, P. J., & Steely, J. D. (2002). Productivity effects of research assessment exercises. Discussion Paper 2002–02, Centre for European Labour Market Research, University of Aberdeen.
Mugabushaka, A. M., Kyriakou, A., & Papazoglou, T. (2016). Bibliometric indicators of interdisciplinarity: the potential of the Leinster–Cobbold diversity indices to study disciplinary diversity. Scientometrics, 107(2), 593–607.
Najman, J. M., & Hewitt, B. (2003). The validity of publication and citation counts for sociology and other selected disciplines. Journal of Sociology, 39(1), 63–81.
Nelson, R. R. (2001). Observations on the post-bayh-dole rise of patenting at American universities. Journal of Technology Transfer, 26(1–2), 13–19.
O’Connell, C. (2013). Research discourses surrounding global university rankings: Exploring the relationship with policy and practice recommendations. Higher Education, 65(6), 709–723.
Olivares, M., & Wetzel, H. (2011). Competing in the higher education market: Empirical evidence for economies of scale and scope in German higher education institutions. CESifo Economic Studies, 60(4), 653–680.
Parteka, A., & Wolszczak-Derlacz, J. (2013). Dynamics of productivity in higher education: cross-European evidence based on bootstrapped Malmquist indices. Journal of Productivity Analysis, 40(1), 67–82.
Pastor, J. M., & Serrano, L. (2016). The determinants of the research output of universities: Specialization, quality and inefficiencies. Scientometrics, 109(2), 1255–1281.
Pezzoni, M., Sterzi, V., & Lissoni, F. (2012). Career progress in centralized academic systems: Social capital and institutions in France and Italy. Research Policy, 41(4), 704–719.
Pfeffer, J., & Langton, N. (1993). The effect of wage dispersion on satisfaction, productivity, and working collaboratively: Evidence from college and university faculty. Administrative Science Quarterly, 38(3), 382–407.
Picard-Aitken, M., Foster, T., & Labrosse, I. (2010). Tenth-year evaluation of the Canada Research Chairs Program: Final evaluation report. Natural Sciences and Engineering Research Council of Canada.
Pina, D. G., Barać, L., Buljan, I., Grimaldo, F., & Marušić, A. (2019). Effects of seniority, gender and geography on the bibliometric output and collaboration networks of European Research Council (ERC) grant recipients. PLoS ONE, 14(2), e0212286.
Piro, F. N., & Sivertsen, G. (2016). How can differences in international university rankings be explained? Scientometrics, 109(3), 2263–2278.
Porac, J. F., Wade, J. B., Fischer, H. M., Brown, J., Kanfer, A., & Bowker, G. (2004). Human capital heterogeneity, collaborative relationships, and publication patterns in a multidisciplinary scientific alliance: A comparative case study of two scientific teams. Research Policy, 33(4), 661–678.
Prasad, A., Segarra, P., & Villanueva, C. E. (2019). Academic life under institutional pressures for AACSB accreditation: Insights from faculty members in Mexican business schools. Studies in Higher Education, 44(9), 1605–1618.
Pratt, M., Margaritis, D., & Coy, D. (1999). Developing a research culture in a university faculty. Journal of Higher Education Policy and Management, 21(1), 43–56.
Putnam, R. (2001). Social capital: Measurement and consequences. Canadian Journal of Policy Research, 2(1), 41–51.
R Core Team. (2013). R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing.
Radicchi, F., & Castellano, C. (2012). A reverse engineering approach to the suppression of citation biases reveals universal properties of citation distributions. PLoS ONE, 7(3), e33833.
Ramos-Vielba, I., Sánchez-Barrioluengo, M., & Woolley, R. (2016). Scientific research groups’ cooperation with firms and government agencies: Motivations and barriers. The Journal of Technology Transfer, 41(3), 558–585.
Reis, N. R., Ferreira, M. P., & Santos, J. C. (2011). The cultural models in international business research: A bibliometric study of IB journals (pp. 1–31). Glob Advantage, Center of Research in International Business & Strategy, Leiria, Portugal.
Rhaiem, M. (2017). Measurement and determinants of academic research efficiency: A systematic review of the evidence. Scientometrics, 110(2), 581–615.
Rhaiem, M., & Bornmann, L. (2018). Reference Publication Year Spectroscopy (RPYS) with publications in the area of academic efficiency studies: What are the historical roots of this research topic? Applied Economics, 50(13), 1442–1453.
Rinia, E. J., Leeuwen, T. N., & Van Raan, A. F. J. (2002). Impact measures of interdisciplinary research in physics. Scientometrics, 53(2), 241–248.
Robst, J. (2000). Do state appropriations influence cost efficiency in public higher education? Applied Economics Letters, 7(11), 715–719.
Roller, R. H., Andrews, B. K., & Bovee, S. L. (2003). Specialized accreditation of business schools: A comparison of alternative costs, benefits, and motivations. Journal of Education for Business, 78(4), 197–204.
Rousseau, S., & Rousseau, R. (1997). Data envelopment analysis as a tool for constructing scientometric indicators. Scientometrics, 40(1), 45–56.
Ruiz-Castillo, J., & Costas, R. (2014). The skewness of scientific productivity. Journal of Informetrics, 8(4), 917–934.
Ryazanova, O., & McNamara, P. (2016). Socialization and proactive behavior: Multilevel exploration of research productivity drivers in US business schools. Academy of Management Learning and Education, 15(3), 525–548.
Saad, G. (2006). Exploring the h-index at the author and journal levels using bibliometric data of productive consumer scholars and business-related journals respectively. Scientometrics, 69(1), 117–120.
Sabharwal, M. (2013). Productivity and leadership patterns of female faculty members in public administration. The Journal of Public Affairs Education, 19(1), 73–96.
Sagarra, M., Mar-Molinero, C., & Agasisti, T. (2017). Exploring the efficiency of Mexican universities: Integrating data envelopment analysis and multidimensional scaling. Omega, 67, 123–133.
Saisana, M., d’Hombres, B., & Saltelli, A. (2011). Rickety numbers: Volatility of university rankings and policy implications. Research Policy, 40(1), 165–177.
Salas-Velasco, M. (2020). Measuring and explaining the production efficiency of Spanish universities using a non-parametric approach and a bootstrapped-truncated regression. Scientometrics, 122(2), 825–846.
Sav, T. G. (2012). Stochastic cost inefficiency estimates and rankings of public and private research and doctoral granting universities. Journal of Knowledge Management, Economics and Information Technology, 4(3), 11–29.
Sellers-Rubio, R., Mas-Ruiz, F. J., & Casado-Diaz, A. B. (2010). University efficiency: Complementariness versus trade-off between teaching, research and administrative activities. Higher Education, 64(4), 373–391.
Senter, R. (1986). A causal model of productivity in a research facility. Scientometrics, 10(5–6), 307–328.
Shapiro, D. L., Kirkman, B. L., & Courtney, H. G. (2007). Perceived causes and solutions of the translation problem in management research. Academy of Management Journal, 50(2), 249–266.
Shin, J. C., Toutkoushian, R. K., & Teichler, U. (Eds.). (2011). University rankings theoretical basis, methodology and impacts on global higher education. Berlin: Springer.
Sīle, L., & Vanderstraeten, R. (2019). Measuring changes in publication patterns in a context of performance-based research funding systems: the case of educational research in the University of Gothenburg (2005–2014). Scientometrics, 118(1), 71–91.
Silkman, R. H. (1986). Measuring efficiency: An assessment of data envelopment analysis. San Francisco: USA, Jossey-Bass.
Silva, T. H. P., Moro, M. M., & Silva, A. P. C. (2015). Authorship contribution dynamics on publication venues in computer science: an aggregated quality analysis. In Proceedings of the ACM symposium on applied computing (pp. 1142–1147). Salamanca, Spain.
Simar, L., & Wilson, P. W. (2007). Estimation and inference in two-stage, semi parametric models of production processes. Journal of Econometrics, 136(1), 31–64.
Simar, L., & Wilson, W. P. (2000). Statistical inference in nonparametric frontier models: The state of the art. Journal of Productivity Analysis, 13(1), 49–78.
Slaughter, S., & Leslie, L. L. (1997). Academic capitalism: Politics, policies and the entrepreneurial university. Baltimore, MD: Johns Hopkins University Press.
Slaughter, S., & Rhoades, G. (2004). Academic capitalism and the new economy: Markets, state, and higher education. Baltimore, MD: The Johns Hopkins University Press.
Smith, T. E., Jacobs, K. S., Osteen, P. J., & Carter, T. E. (2018). Comparing the research productivity of social work doctoral programs using the h-Index. Scientometrics, 116(3), 1513–1530.
Stella, A., & Woodhouse, D. (2006). Ranking of higher education institutions. AUQA Occasional Publication no. 6, August. Melbourne: Australian Universities Quality Agency. Available online from:http://www.auqa.edu.au/files/publications/ranking_of_higher_education_institutionsfinal.pdf. Accessed 6 January 2014.
Stephan, P. E., & Levin, S. G. (1992). Striking the mother lode in science: The importance of age, place and time. Oxford: Oxford University Press.
Stevens, P. A. (2005). A stochastic frontier analysis of English and Welsh Universities. Education Economics, 13(4), 355–374.
Sun, Y., Zhang, C., & Kok, R. A. (2019). The role of research outcome quality in the relationship between university research collaboration and technology transfer: empirical results from China. Scientometrics, 112(2), 1003–1026.
Talukdar, D. (2011). Patterns of research productivity in the business ethics literature: Insights from analyses of bibliometric distributions. Journal of Business Ethics, 98(1), 137–151.
Taylor, P., & Braddock, R. (2007). International university ranking systems and the idea of university excellence. Journal of Higher Education Policy and Management, 29(3), 245–260.
Thanassoulis, E. (2001). Introduction to the theory and application of data envelopment analysis: A foundation text with integrated software. Norwell, MA: Kluwer Academic Publishers.
Thanassoulis, E., Kortelainen, M., Johnes, G., & Johnes, J. (2011). Costs and efficiency of higher education institutions in England: A DEA analysis. Journal of the operational research society, 62(7), 1282–1297.
The Council of Canadian Academies. (2009). Better research for better business. The Expert Panel on Management, Business, and Finance Research. Council of Canadian Academies, 45p. http://marcelcoupart.tk/download/sMKGH9urPiYC-better-research-for-better-business. Retrieved 10 January 2020.
Thomson Reuters. (2009). Top 20 countries in economics and business. http://sciencewatch.com/dr/cou/2009/09febECO/. Retrieved 16 December 2019.
Thursby, J. G. (2000). What do we say about ourselves and what does it mean? Yet another look at economics department research. Journal of Economic Literature, 38(2), 383–404.
Tyagi, P., Yadav, S. P., & Singh, S. P. (2009). Relative performance of academic departments using DEA with sensitivity analysis. Evaluation and Program Planning, 32(2), 168–177.
Van der Stocken, T., Hugé, J., Deboelpaep, E., Vanhove, M. P., De Bisthoven, L. J., & Koedam, N. (2016). Academic capacity building: holding up a mirror. Scientometrics, 106(3), 1277–1280.
Van Raan, A. F. J. (2005). Fatal attraction: Conceptual and methodological problems in the ranking of universities by bibliometric methods. Scientometrics, 62(1), 133–143.
Vieira, E. S., & Gomes, J. A. (2010). Citations to scientific articles: Its distribution and dependence on the article features. Journal of Informetrics, 4(1), 1–13.
Vieira, P. C., & Teixeira, A. C. (2010). Are finance, management, and marketing autonomous fields of scientific research? An analysis based on journal citations. Scientometrics, 85(3), 627–646.
Von Tunzelmann, N., & Kraemer Mbula, E. (2003). Changes in research assessment practices in other countries since 1999: Final report (pp. 46). Report to the Higher Education Funding Council for England.
Wang, X., Zhao, Y., Liu, R., & Zhang, J. (2013). Knowledge-transfer analysis based on co-citation clustering. Scientometrics, 97(3), 859–869.
Warning, S. (2004). Performance differences in German higher education: Empirical analysis of strategic groups. Review of Industrial Organization, 24(4), 393–408.
Weinberg, B. A., Owen-Smith, J., Rosen, R. F., Schwarz, L., Allen, B. M., Weiss, R. E., et al. (2014). Science funding and short-term economic activity. Science, 344(6179), 41–43.
Weingart, P. (2005). Impact of bibliometrics upon the science system: Inadvertent consequences? Scientometrics, 62(1), 117–131.
Weiss, Y., & Lillard, L. A. (1982). Output variability, academic labor contracts, and waiting times for promotion. Research in Labor Economics, 5, 157–188.
Wellen, R. (2009). Corporatization and commercialization, governance, research and innovation, universities and the academic mission. Academic Matters, OCUFA’s Journal of Higher Education. Available on: https://academicmatters.ca/grappling-with-academic-capitalism-in-canadian-universities/.
Wilsdon, J. (2015). We need a measured approach to metrics. Nature, 523(7559), 129.
Wohlrabe, K., de Moya Anegon, F., & Bornmann, L. (2019). How efficiently do elite US universities produce highly cited papers? Publications, 7(1), 4.
Woiceshyn, J., & Eriksson, P. (2014). Academic engagement at Canadian and finnish business schools. In Academy of management proceedings (Vol. 2014, No. 1, pp. 14015). Briarcliff Manor, NY 10510: Academy of Management.
Wolszczak-Derlacz, J. (2017). An evaluation and explanation of (in) efficiency in higher education institutions in Europe and the US with the application of two-stage semi-parametric DEA. Research Policy, 46(9), 1595–1605.
Wolszczak-Derlacz, J., & Parteka, A. (2011). Efficiency of European public higher education institutions: A two-stage multicountry approach. Scientometrics, 89(3), 887–917.
Worthington, A. (2004). Frontier efficiency measurement in healthcare: A review of empirical techniques and selected applications. Medical Care Research and Review, 61(2), 1–36.
Worthington, A. C., & Higgs, H. (2011). Economies of scale and scope in Australian higher education. Higher Education, 61(4), 387–414.
Worthington, A. C., & Lee, B. L. (2008). Efficiency, technology and productivity change in Australian universities, 1998–2003. Economics of Education Review, 27(3), 285–298.
Wu, D., Li, M., Zhu, X., Song, H., & Li, J. (2015). Ranking the research productivity of business and management institutions in Asia-Pacific region: empirical research in leading ABS journals. Scientometrics, 105(2), 1253–1257.
Xu, F., Liu, W. B., & Mingers, J. (2015). New journal classification methods based on the global h-index. Information Processing and Management, 51(2), 50–61.
Yang, G. L., Fukuyama, H., & Song, Y. Y. (2018). Measuring the inefficiency of Chinese research universities based on a two-stage network DEA model. Journal of Informetrics, 12(1), 10–30.
Ylijoki, O. H. (2003). Entangled in academic capitalism? A case-study on changing ideals and practices of university research. Higher Education, 45(3), 307–335.
Zamojcin, K. A., & Bernardi, R. A. (2013). Ranking North American accounting scholars publishing education papers: 1966 through 2011. Journal of Accounting Education, 31(2), 194–212.
Zamutto, R. F. (2008). Accreditation and the globalization of business. Academy of Management Learning and Education, 7(2), 256–268.
Ziman, J. (1996). “Post-academic science”: Constructing knowledge with networks and norms. Science Studies, 9(1), 67–80.
Zucker, L. G., Darby, M. R., & Armstrong, J. S. (2002). Commercializing knowledge: University science, knowledge capture, and firm performance in biotechnology. Management Science, 48(1), 138–153.
Zuckerman, H. A., & Merton R. K. (1972). Age, aging, and age structure in science. In M. R. Riley, M. Johnson, & A. Foner (Eds.), A Sociology of Age Stratification: Aging and Society, vol. 3 (pp. 497–559., New York: Russel Sage foundation. Reprinted in: The Sociology of Science: Collected Papers of R.K. Merton N.W. Storer (Ed.), 1973. Chicago University: Chicago Press.
Zuo, K., & Guan, J. (2017). Measuring the R&D efficiency of regions by a parallel DEA game model. Scientometrics, 112(1), 175–194.
Acknowledgments
The authors would like to acknowledge financial assistance provided by The Social Sciences and Humanities Research Council of Canada, and by The Fonds de recherche du Québec—Société et culture. We also would like to thank all the faculty members of Canadian business schools who participated in our survey.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix 1. Definitions of explanatory variables and descriptive statistics
Measure | Sub-items | Mean (SD) | Percentage (number) | |
---|---|---|---|---|
Inputs | ||||
Time dedicated to research activities | Measured as the percentage of scholar’s time dedicated to research activities | 35.30% (16.02%) | ||
Research funds | The total research funding (for research projects and infrastructure) of all scholar’s research projects during the past 12 months | 61263.33 137128.3 | ||
Outputs | ||||
Number of publications | Measured as the lifetime number of scholar’s scientific contributions as compiled by | |||
Web of Science database | 7.51 (10.359) | |||
Google Scholar database | 22.531 (31.917) | |||
Number of citations | Measured as the lifetime number of scholar’s citations as compiled by | |||
Web of Science database | 79.04 (225.513) | |||
Google Scholar database | 284.198 (1017.651) | |||
Factors driving or hampering academic research efficiency | ||||
Seniority [SENIOR] | The level of seniority in the academic ranks was measured as follows | |||
Assistant professor [ASSIST] is a binary variable coded 1 if the scholar is an assistant professor, and coded 0 otherwise | 25.2% (203) | |||
Associate professor [ASSOC] is a binary variable coded 1 if the scholar is an associate professor, and coded 0 otherwise | 39.0% (315) | |||
Full professor [FULL] is a binary variable coded 1 if the scholar is a full professor, and coded 0 otherwise. This last category of scholars was used as the reference category | 35.8% (289) | Benchmark | ||
Public sources of research funding [PUBLIC] | The level of scholar’s total research budget funded during the past 12 months by provincial and federal research councils was measured by the three following binary variables | |||
Non-funded by research Councils [NO_PFUND] is a binary variable coded 1 if the scholar was not funded over the past 12 months by provincial nor federal research councils, and coded 0 otherwise | 50.8% (410) | |||
Partially funded by research Councils [PAR_PFUND] is a binary variable coded 1 if the percentage of scholar’s total research funding over the past 12 months funded by provincial and federal research councils ranges between 1 and 99%, and coded 0 otherwise | 27.1% (219) | |||
Totally funded by research Councils [TOT_PFUND] is a binary variable coded 1 if, over the past 12 months, the funding from provincial and federal research councils represented 100% of scholar’s total research funding, and coded 0 otherwise. This last category was used as the reference category | 22.1% (178) | Benchmark | ||
Private sources of research funding [PRIVATE] | The level of scholar’s total research budget funded during the past 12 months by industry grants (contracts by third parties) was measured by the dichotomous variable Coded 1 if the scholar over the past 12 months was funded by industry, and coded 0 otherwise | |||
Funded by industry | 73.0% (589) | |||
Not funded by industry | 27.0% (218) | |||
World ranking of scholars’ universities of affiliation [PREST] | The academic ranking of scholars’ universities of affiliation is based on the 2010 Academic Ranking of World Universities (ARWU). ARWU is one of most popular and employed ranking tables (Lukman et al. 2010). Three types of universities were distinguished by the three following binary variables | |||
Third tier Universities [OUT_LIST] is a binary variable coded 1 if the scholar’s university of affiliation was not in the ARWU top-500 ranking for the year 2010, and coded 0 if his university was in the top-500 ranking for the year 2010 | 22,8% (184) | |||
Second tier Universities [IN_LIST] is a binary variable coded 1 if the scholar’s university of affiliation was in the ARWU top-500 ranking but not in the Canadian top-5 in this list for the year 2010, and coded 0 otherwise | 66,4% (536) | |||
Top-5 Universities [TOP_5] is a binary variable coded 1 if the scholar’s university of affiliation was in the Canadian top-5 ARWU ranking for the year 2010, and coded 0 otherwise This last category was used as the reference category | 10.8% (87) | Benchmark | ||
Factors driving or hampering academic research efficiency | ||||
Accreditation [REPUT] | Dichotomous variable Coded ‘1’ (with AASCB), if the scholar’s university of affiliation was accredited through The Association to Advance Collegiate Schools of Business (AACSB), and coded 0 otherwise (without accreditation) | |||
With certification | 75.3 (608) | |||
Without certification | 24.7% (199) | |||
Size effects [DOCPROG] | Dichotomous variable Coded ‘1’ (with doctoral program), if the scholar’s university of affiliation has a doctoral program, and coded 0 otherwise (without doctoral program) | |||
With doctoral program | 75.7% (611) | |||
Without doctoral program | 24.4% (196) | |||
Strength of ties with companies [TIES] | Dichotomous variable Coded ‘1’ (Strong ties), if the scholar described his working relationship with managers/employees in companies in the past 3 years as very close (practically like being in the same work group), or somewhat close (like discussing and solving issues together), and 0 otherwise (weak ties) (somewhat distant, like with people that you do not know well; distant, like a working group with which you can only have a quick exchange of information; or very distant, practically like with people that you do not know at all) | |||
Strong ties | 59.9% (483) | |||
Weak ties | 40.1% (324) | |||
Frequency of contacts with companies [CONTACT] | Dichotomous variable coded ‘1’, (Frequent contact) if the scholar has had Very often or Often person-to-person contact with managers and/or employees in companies in the past 3 years, and 0 otherwise (Never, Rarely, or Sometimes) | |||
Very often and Often | 38.7% (312) | |||
Never, rarely, and sometimes | 61.3% (495) | |||
Control variable | ||||
Business disciplines | A series of eight dichotomous variables indicating the scholars’ business disciplines: | |||
Human resources management [HRM] | 14.4% (116) | |||
Finance [FINAN] | 9.7% (78) | |||
Marketing [MARK] | 14.7% (119) | |||
Information management [INFOR] | 7.9% (64) | |||
Accounting [ACCOUNT] | 12.8% (103) | |||
Operational Research [OPER] | 5.9% (48) | |||
Economics [ECON] | 7.2% (58) | |||
Management [MNG] | 27.4% (221) |
Appendix 2. Comparison of means of total number of papers published between faculty members in the FS and those in the ROP sample (Independent-samples T test on ranked data)
Total number of papers published according to WoS | FS | ROP | T test for equality of means†† |
---|---|---|---|
Number of cases | 807 | 2327 | |
Means | 1596.2 | 1557.5 | 1.089 |
Standard deviation | 848.6 | 875.9 | |
P value for the Levene test of equality of variances | 0.039** |
Appendix 3. Comparison of means of total number of citations between faculty members in the FS and in the ROP sample (Independent-samples T test on ranked data)
Total number of citations according to WoS | FS | ROP | T test for equality of means†† |
---|---|---|---|
Number of cases | 807 | 2327 | |
Means | 1573.7 | 1565.3 | 0.244 |
Standard deviation | 848.6 | 875.9 | |
P value for the Levene test of equality of variances | 0.334 |
Appendix 4. Distribution of samples (FS vs. ROP) of faculty members according to academic rank (Chi square test)
Academic rank | All faculty members | FS | ROP | Pearson Chi square†† | |||
---|---|---|---|---|---|---|---|
Number | % | Number | % | Number | % | ||
Full professor | 1169 | 37.3 | 289 | 35.8 | 880 | 37.8 | 1.629 |
Associate professor | 1167 | 37.2 | 315 | 39.0 | 852 | 36.6 | |
Assistant professor | 798 | 25.5 | 203 | 25.2 | 595 | 25.6 | |
Total | 3134 | 100.0 | 807 | 100.0 | 2327 | 100.0 |
Rights and permissions
About this article
Cite this article
Rhaiem, M., Amara, N. Determinants of research efficiency in Canadian business schools: evidence from scholar-level data. Scientometrics 125, 53–99 (2020). https://doi.org/10.1007/s11192-020-03633-z
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11192-020-03633-z