Thursday, September 24, 2020

Cэтгүүлийн нөлөөллийн индекс (Journal Impact Factor), тооцоолол, утга учир, үнэ цэнэ, шүүмж

 Академик сэтгүүлийн импакт фактор, түүний тооцоолол,  үнэ цэнэ, шүүмж


Сэтгүүлийн импакт фактор (нөлөөллийн хүчин зүйл)  нь академик сэтгүүлийн шинжлэх ухаанч байдлыг илтгэх индекс юм. Энэ нь уг сэтгүүлийн өмнөх хоёр дугаарт хэвлэгдсэн өгүүллүүд тухайн жилд дундажаар хэдэн удаа ишлэгдснийг илтгэнэ.


Импакт факторыг тооцоолох

 Сэтгүүлийн тухайн жилийн нөлөөллийн индекс гэдэг уг сэтгүүлийн өмнөх хоёр жилд хэвлэгдсэн хэвлэлээс (publications )-аас тухайн жилд авсан ишлэлийг тоог тэрхүү хоёр жилд уг сэтгүүлд хэвлэгдсэн  ишлэгдэж болохуйц нийт зүйлийн (citable items ) тоонд харьцуулсан харьцаа хэлнэ. 


 

'Nature' сэтгүүлийн 2017 оны нөлөөллийн индексийг хэрхэн бодсныг жишээ болгон дор харуулав.


 Энэхүү сэтгүүлийн 2015, 2016 оны дугаарт хэвлэгдсэн ишлэгдэж болохуйц зүйл нь нийтдээ 1782 байсан бол тэдгээрээс 2017 онд авсан ишлэлийн тоо нь 74090 байсан байна. Иймд энэ сэтгүүлийн нөлөөллийн индекс нь 41,577 . 

"Энэ нь, дундажаар,  уг сэтгүүлийн 2015,2016 оны дугаарт хэвлэгдсэн өгүүлэл (papers) бүр 2017 онд ойролцоогоор 42 удаа  ишлэгдсэн байна" гэсэн утгыг илтгэнэ. Өөрөөр хэлбэл, Энэ нь уг чанараараа, сэтгүүлд хэвлэгдсэн өгүүлэл мэргэжлийн бүлийн хувьд ч, танин мэдэхүйн хувьд харьцангуй үнэт цэнэтэй байсан байна гэсэн утгыг илтгэнэ. 

Шүүмжлэлт санаа: Импакт фактор нь эцэстээ ишлэлийн тоо тул  академик сэтгүүлийн шинжлэх ухаанч  байдал, үнэ цэнийг түүгээр тогтоох нь  учир дутагдалтай. Гэвч одоогоор академик сэтгүүлийн үнэ цэнийг тогтоох түүнээс өөр сайн аргыг хүн төрөлхтөн олоогүй байна. 


Ц. Лувсандорж 

2020 оны 9 дүгээр сарын 24-ний өдөр

Жич: Энэ талаар доор бичвэрээс дэлгэрүүлж үзэж болно. 


Эх үүсвэр:wikipedia.com

This means that, on average, its papers published in 2015 and 2016 received roughly 42 citations each in 2017. Note that 2017 impact factors are reported in 2018; they cannot be calculated until all of the 2017 publications have been processed by the indexing agency.

The value of impact factor depends on how to define "citations" and "publications"; the latter are often referred to as "citable items". In current practice, both "citations" and "publications" are defined exclusively by ISI as follows. "Publications" are items that are classed as "article", "review" or "proceedings paper"[7] in the Web of Science (WoS) database; other items like editorials, corrections, notes, retractions and discussions are excluded. WoS is accessible to all registered users, who can independently verify the number of citable items for a given journal. In contrast, the number of citations is extracted not from the WoS database, but from a dedicated JCR database, which is not accessible to general readers. Hence, the commonly used "JCR Impact Factor" is a proprietary value, which is defined and calculated by ISI and can not be verified by external users.[8]

New journals, which are indexed from their first published issue, will receive an impact factor after two years of indexing; in this case, the citations to the year prior to Volume 1, and the number of articles published in the year prior to Volume 1, are known zero values. Journals that are indexed starting with a volume other than the first volume will not get an impact factor until they have been indexed for three years. Occasionally, Journal Citation Reports assigns an impact factor to new journals with less than two years of indexing, based on partial citation data.[9][10] The calculation always uses two complete and known years of item counts, but for new titles one of the known counts is zero. Annuals and other irregular publications sometimes publish no items in a particular year, affecting the count. The impact factor relates to a specific time period; it is possible to calculate it for any desired period. For example, the JCR also includes a five-year impact factor, which is calculated by dividing the number of citations to the journal in a given year by the number of articles published in that journal in the previous five years.[11][12]

Use

The impact factor is used to compare different journals within a certain field. The Web of Science indexes more than 11,500 science and social science journals.[13]

Journal impact factors are often used to evaluate the merit of individual articles and individual researchers.[14] This use of impact factors was summarised by Hoeffel:[15]

Impact Factor is not a perfect tool to measure the quality of articles but there is nothing better and it has the advantage of already being in existence and is, therefore, a good technique for scientific evaluation. Experience has shown that in each specialty the best journals are those in which it is most difficult to have an article accepted, and these are the journals that have a high impact factor. Most of these journals existed long before the impact factor was devised. The use of impact factor as a measure of quality is widespread because it fits well with the opinion we have in each field of the best journals in our specialty....In conclusion, prestigious journals publish papers of high level. Therefore, their impact factor is high, and not the contrary.

As impact factors are a journal-level metric, rather than an article- or individual-level metric, this use is controversial. Garfield agrees with Hoeffel,[16] but warns about the "misuse in evaluating individuals" because there is "a wide variation [of citations] from article to article within a single journal".[17]

Criticism

Numerous critiques have been made regarding the use of impact factors.[18][19][20] A 2007 study noted that the most fundamental flaw is that impact factors present the mean of data that is not normally distributed, and suggested that it would be more appropriate to present the median of these data.[21] There is also a more general debate on the validity of the impact factor as a measure of journal importance and the effect of policies that editors may adopt to boost their impact factor (perhaps to the detriment of readers and writers). Other criticism focuses on the effect of the impact factor on behavior of scholars, editors and other stakeholders.[22] Others have made more general criticisms, arguing that emphasis on impact factor results from negative influence of neoliberal policies on academia claiming that what is needed is not just replacement of the impact factor with more sophisticated metrics for science publications but also discussion on the social value of research assessment and the growing precariousness of scientific careers in higher education.[23][24]

Validity as a measure of importance

It has been stated that impact factors and citation analysis in general are affected by field-dependent factors[25] which may invalidate comparisons not only across disciplines but even within different fields of research of one discipline.[26] The percentage of total citations occurring in the first two years after publication also varies highly among disciplines from 1–3% in the mathematical and physical sciences to 5–8% in the biological sciences.[27] Thus impact factors cannot be used to compare journals across disciplines.

Impact factors are sometimes used to evaluate not only the journals but the papers therein, thereby devaluing papers in certain subjects.[28] The Higher Education Funding Council for England was urged by the House of Commons Science and Technology Select Committee to remind Research Assessment Exercise panels that they are obliged to assess the quality of the content of individual articles, not the reputation of the journal in which they are published.[29] The effect of outliers can be seen in the case of the article "A short history of SHELX", which included this sentence: "This paper could serve as a general literature citation when one or more of the open-source SHELX programs (and the Bruker AXS version SHELXTL) are employed in the course of a crystal-structure determination". This article received more than 6,600 citations. As a consequence, the impact factor of the journal Acta Crystallographica Section A rose from 2.051 in 2008 to 49.926 in 2009, more than Nature (at 31.434) and Science (at 28.103).[30] The second-most cited article in Acta Crystallographica Section A in 2008 only had 28 citations.[31] Additionally, impact factor is a journal metric and should not be used to assess individual researchers or institutions.[32][33]

Journal rankings constructed based solely on impact factors only moderately correlate with those compiled from the results of expert surveys.[34]

A.E. Cawkell, former Director of Research at the Institute for Scientific Information remarked that the Science Citation Index (SCI), on which the impact factor is based, "would work perfectly if every author meticulously cited only the earlier work related to his theme; if it covered every scientific journal published anywhere in the world; and if it were free from economic constraints."[35]

Editorial policies that affect the impact factor

A journal can adopt editorial policies to increase its impact factor.[36][37] For example, journals may publish a larger percentage of review articles which generally are cited more than research reports.[5]

Journals may also attempt to limit the number of "citable items"—i.e., the denominator of the impact factor equation—either by declining to publish articles that are unlikely to be cited (such as case reports in medical journals) or by altering articles (e.g., by not allowing an abstract or bibliography in hopes that Journal Citation Reports will not deem it a "citable item"). As a result of negotiations over whether items are "citable", impact factor variations of more than 300% have been observed.[38] Items considered to be uncitable—and thus are not incorporated in impact factor calculations—can, if cited, still enter into the numerator part of the equation despite the ease with which such citations could be excluded. This effect is hard to evaluate, for the distinction between editorial comment and short original articles is not always obvious. For example, letters to the editor may refer to either class.

Another less insidious tactic journals employ is to publish a large portion of its papers, or at least the papers expected to be highly cited, early in the calendar year. This gives those papers more time to gather citations. Several methods, not necessarily with nefarious intent, exist for a journal to cite articles in the same journal which will increase the journal's impact factor.[39][40]

Beyond editorial policies that may skew the impact factor, journals can take overt steps to game the system. For example, in 2007, the specialist journal Folia Phoniatrica et Logopaedica, with an impact factor of 0.66, published an editorial that cited all its articles from 2005 to 2006 in a protest against the "absurd scientific situation in some countries" related to use of the impact factor.[41] The large number of citations meant that the impact factor for that journal increased to 1.44. As a result of the increase, the journal was not included in the 2008 and 2009 Journal Citation Reports.[42]

Coercive citation is a practice in which an editor forces an author to add extraneous citations to an article before the journal will agree to publish it, in order to inflate the journal's impact factor. A survey published in 2012 indicates that coercive citation has been experienced by one in five researchers working in economics, sociology, psychology, and multiple business disciplines, and it is more common in business and in journals with a lower impact factor.[43] However, cases of coercive citation have occasionally been reported for other disciplines.[44]

Correlation between impact factor and quality

The journal impact factor (JIF) was originally designed by Eugene Garfield as a metric to help librarians make decisions about which journals were worth subscribing to, as the JIF aggregates the number of citations to articles published in each journal. Since then, the JIF has become associated as a mark of journal "quality", and gained widespread use for evaluation of research and researchers instead, even at the institutional level. It thus has significant impact on steering research practices and behaviours.[45][46]

Already around 2010, national and international research funding institutions have pointed out that numerical indicators such as the JIF should not be referred to as a measure of quality.[note 1] In fact, the JIF is a highly manipulated metric,[47][48][49] and the justification for its continued widespread use beyond its original narrow purpose seems due to its simplicity (easily calculable and comparable number), rather than any actual relationship to research quality.[50][51][52]

Empirical evidence shows that the misuse of the JIF – and journal ranking metrics in general – has a number of negative consequences for the scholarly communication system. These include confusion between outreach of a journal and the quality of individual papers and insufficient coverage of social sciences and humanities as well as research outputs from across Latin America, Africa, and South-East Asia.[53] Additional drawbacks include the marginalization of research in vernacular languages and on locally relevant topics, inducement to unethical authorship and citation practices as well as more generally fostering of a reputation economy in academia based on publishers" prestige rather than actual research qualities such as rigorous methods, replicability and social impact. Using journal prestige and the JIF to cultivate a competition regime in academia has been shown to have deleterious effects on research quality.[54]

JIFs are still regularly used to evaluate research in many countries which is a problem since a number of outstanding issues remain around the opacity of the metric and the fact that it is often negotiated by publishers.[55][56][57] However, these integrity problems appear to have done little to curb its widespread mis-use.

A number of regional focal points and initiatives are now providing and suggesting alternative research assessment systems, including key documents such as the Leiden Manifesto[note 2] and the San Francisco Declaration on Research Assessment (DORA). Recent developments around 'Plan S' call on a broader adoption and implementation of such initiatives alongside fundamental changes in the scholarly communication system.[note 3] Thus, there is little basis for the popular simplification which connects JIFs with any measure of quality, and the ongoing inappropriate association of the two will continue to have deleterious effects. As appropriate measures of quality for authors and research, concepts of research excellence should be remodelled around transparent workflows and accessible research results.[58][59][60]

Negotiated values

The method of calculation of the impact factor by Clarivate is not generally known and the results are therefore not predictable nor reproducible. In particular, the result can change dramatically depending on which items are considered as "citable" and therefore included in the denominator.[61] Publishers routinely discuss with Clarivate how to improve the "accuracy" of their journals' impact factor and therefore get higher scores.[62]

Such discussions routinely produce "negotiated values" which result in dramatic changes in the observed scores for dozens of journals, sometimes after unrelated events like the purchase by one of the big five publishers.[63]

Distribution skewness

Journal impact factors are influenced heavily by a small number of highly cited papers. In general, most papers published in 2013–14 received many fewer citations than indicated by the impact factor. Two journals (Nature [blue], PLOS ONE [orange]) are shown to represent a highly cited and less cited journal, respectively. Note that the high citation impact of Nature is derived from relatively few highly cited papers. Modified after Callaway 2016.[64]

Because citation counts have highly skewed distributions,[65] the mean number of citations is potentially misleading if used to gauge the typical impact of articles in the journal rather than the overall impact of the journal itself.[66] For example, about 90% of Nature's 2004 impact factor was based on only a quarter of its publications, and thus the actual number of citations for a single article in the journal is in most cases much lower than the mean number of citations across articles.[67] Furthermore, the strength of the relationship between impact factors of journals and the citation rates of the papers therein has been steadily decreasing since articles began to be available digitally.[68]

Critics of the JIF state that use of the arithmetic mean in its calculation is problematic because the pattern of citation distribution is skewed. Citation distributions for eight selected journals in,[69] along with their JIFs and the percentage of citable items below the JIF shows that the distributions are clearly skewed, making the arithmetic mean an inappropriate statistic to use to say anything about individual papers within the citation distributions. More informative and readily available article-level metrics can be used instead, such as citation counts or "altmetrics', along with other qualitative and quantitative measures of research "impact'.[60][70]

Université de Montréal, Imperial College London, PLOS, eLife, EMBO Journal, The Royal Society, Nature and Science proposed citation distributions metrics as alternative to impact factors.[71][72][73]

Responses

Because "the impact factor is not always a reliable instrument", in November 2007 the European Association of Science Editors (EASE) issued an official statement recommending "that journal impact factors are used only—and cautiously—for measuring and comparing the influence of entire journals, but not for the assessment of single papers, and certainly not for the assessment of researchers or research programmes".[19]

In July 2008, the International Council for Science (ICSU) Committee on Freedom and Responsibility in the Conduct of Science (CFRS) issued a "statement on publication practices and indices and the role of peer review in research assessment", suggesting many possible solutions—e.g., considering a limit number of publications per year to be taken into consideration for each scientist, or even penalising scientists for an excessive number of publications per year—e.g., more than 20.[74]

In February 2010, the Deutsche Forschungsgemeinschaft (German Research Foundation) published new guidelines to evaluate only articles and no bibliometric information on candidates to be evaluated in all decisions concerning "performance-based funding allocations, postdoctoral qualifications, appointments, or reviewing funding proposals, [where] increasing importance has been given to numerical indicators such as the h-index and the impact factor".[75] This decision follows similar ones of the National Science Foundation (US) and the Research Assessment Exercise (UK).[citation needed]

In response to growing concerns over the inappropriate use of journal impact factors in evaluating scientific outputs and scientists themselves, the American Society for Cell Biology together with a group of editors and publishers of scholarly journals created the San Francisco Declaration on Research Assessment (DORA). Released in May 2013, DORA has garnered support from thousands of individuals and hundreds of institutions,[24] including in March 2015 the League of European Research Universities (a consortium of 21 of the most renowned research universities in Europe),[76] who have endorsed the document on the DORA website.

Closely related indices

Some related values, also calculated and published by the same organization, include:

  • Cited half-life: the median age of the articles that were cited in Journal Citation Reports each year. For example, if a journal's half-life in 2005 is 5, that means the citations from 2001–2005 are half of all the citations from that journal in 2005, and the other half of the citations precede 2001.[77]
  • Aggregate impact factor for a subject category: it is calculated taking into account the number of citations to all journals in the subject category and the number of articles from all the journals in the subject category.
  • Immediacy index: the number of citations the articles in a journal receive in a given year divided by the number of articles published.

As with the impact factor, there are some nuances to this: for example, ISI excludes certain article types (such as news items, correspondence, and errata) from the denominator.[78][79][80] [81]

Other measures of impact

Additional journal-level metrics are available from other organizations. For example, CiteScore: is a metric for serial titles in Scopus launched in December 2016 by Elsevier.[82][83] While these metrics apply only to journals, there are also author-level metrics, such as the H-index, that apply to individual researchers. In addition, article-level metrics measure impact at an article level instead of journal level. Other more general alternative metrics, or "altmetrics", may include article views, downloads, or mentions in social media.

Counterfeit

Fake impact factors or bogus impact factors are produced by companies or individuals not affiliated with the Journal Citation Reports (JCR).[84] According to an article published in the United States National Library of Medicine, these include Global Impact Factor (GIF), Citefactor, and Universal Impact Factor (UIF).[85] Jeffrey Beall maintained a list of such misleading metrics.[86][87] Another deceitful practice is reporting "alternative impact factors", calculated as the average number of citations per article using citation indices other than JCR, even if based on reputable sources such as Google Scholar (e.g., "Google-based Journal Impact Factor").[88]

False impact factors are often used by predatory publishers.[89] Consulting Journal Citation Reports' master journal list can confirm if a publication is indexed by Journal Citation Reports.[90] The use of fake impact metrics is considered a "red flag".[91]

See also

Tuesday, September 8, 2020

Key terms and challenges for education marketing

  

Ideas and points in the presentation on the education marketing

Key terms for education marketing: 

-Education marketing

-4Ps of marketing

-Education marketing strategies

 Key challenges for higher education 

-Sustainability 

-Internalization  

Thoughts on furthering the themes related to education marketing

Education marketing as an academic subject 

Education marketing communities and institutions 

Education marketing as a occupation 

Education marketing as a an elective course in teacher education  

Journal of Education marketing

 Problems:

 How to sustain our  higher education institutions, colleges and universities?

 How to  conceptualize the key understandings and conceptualization into our particular contexts and cases?

How to institutionalize education marketing?

 prof Luvsandorj

Department of Didactics,SMS, MSUE

2020.9.8

Thursday, August 27, 2020

МБУС-ийн Дидактикийн тэнхмийн их семинар (2020-2021 оны хичээлийн жилийн 1 дүгээр улирал)

МБУС-ийн Дидактикийн тэнхмийн 2020-2021 оны хичээлийн жилийн 1 дүгээр улирлын их семинарын хуваарь 

Сэдэв: Боловсролын зах зээлийн тулгамдсан асуудал, шийдэл, арга зам 



 

Tuesday, February 19, 2019

Боловсролын эдийн засгийн тулгамдсан асуудал, шийдэл, арга зам



 Боловсролын эдийн засгийн тулгамдсан асуудал, шийдэл, арга зам


Боловсролын эдийн засгийн ухаан[1] нь боловсролтой холбоотой эдийн засгийн асуудлыг[2] судалдаг ухаан юм. Энэ утгаараа, энэ нь нэг талаасаа, боловсрол судлалын салбар, нөгөө талаасаа, эдийн засгийн ухааны хэсэг болдог. Ийм учраас боловсролын эдийн засгийн ухааны үндсэн асуудлыг дэвшүүлэхэд эдийн засгийн үндсэн асуудлыг боловсролд буулгаж тодорхойлох; мөн боловсролын үндсэн асуудлыг эдийн засгийн ухаанаар харж тогтоох гэсэн хоёр янзын хандлага, арга зам бий.
Боловсролын хязгаарлагдмал нөөцөөр түүний хязгааргүй хэрэгцээг хэрхэн хангах вэ?  Энэ нь боловсролын эдийн засгийн ухааны ч, боловсролын эдийн засгийн ч үндсэн асуулт мөн. Энэ асуултын асуудал нь “хязгаарлагдмал нөөц”, “хязгааргүй хэрэгцээ” гэсэн үгийн утганд байдаг. Хязгаарлагдмал нөөцөөр хязгааргүй хэрэгцээ хангах гэдэг нь математик талаасаа, хамгийн их, хамгийн бага утга, хил хязгаарыг тогтоох бодлогыг тавьж шийдэхэд хүргэдэг бол  арвилах, хямагдах гэсэн амьдралын утгаасаа, энэ нь боловсролын аливаа үйл, үйлдэл, ажиллагаа, ажил үйлсээс үр дүн[3], үр ашиг[4], үр өгөөж[5] гаргахад хөтөлдөг.
Боловсрол, нэг талаасаа, хөрөнгө оруулалт, нөгөө талаасаа, үйлчилгээ болно. Ийм учраас боловсролд оруулах хөрөнгө оруулалт, үйлчилгээний үр дүн, үр ашиг, үр өгөөжийг тооцож тогтоох нь боловсролын эдийн засгийн ухааны ч, боловсролын эдийн засгийн ч үндсэн асуудал болдог.Боловсролын хөрөнгө оруулалт, үйлчилгээний түгээмэл хэлбэр бүрийг үр дүн, үр ашиг, үр өгөөжтэй байхыг шалгуур болгон хэмжин үнэлж, дүгнэлт хийх замаар онол, практикийн олон асуудлыг олж дэвшүүлж болно. Үүнийг хийх ерөнхий аргачлалыг дор хүснэгтлэн үзүүлэв (Хүснэгт).

   Хүснэгт.Боловсролын эдийн засгийн тулгамдсан асуудлын томьёоллын бүдүүвч
Боловсрол

Мөн чанар

Шинж хэлбэр
Шалгуур
Үр дүн
Үр ашиг
Үр өгөөж


Хөрөнгө оруулалт
Нэг хүүхдийн хувьсах зардал



Оюутны сургалтын төлбөр



Төсөл, хөтөлбөрийн  гадаад, дотоод эх үүсвэр, санхүүжилт



Зээл, тусламж,



Хандив



Төсөв













Үйлчилгээ
Хичээл



Сургалт



Шалгалт



Үнэлгээ



Судалгаа



Зөвлөгөө[6]



Зөвлөмж, дүгнэлт[7]



Сургалтыг удирдах тогтолцооны программ[8]



Бодлого



Хөтөлбөр



Хууль








Дээр харуулсан бүдүүвчийн дагуу боловсролын эдийн засагт тулгарч буй асуудлыг боловсролын түвшнээр нь дараахь байдлаар нарийвчлан томьёолж болно. Үүнд:
  Нэг.  Ерөнхий боловсролын түвшинд:
-Ерөнхий боловсролын сургууль, цэцэрлэгт суралцах нэг хүүхдийн хувьсах зардлыг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар үйл өртөгжүүлэх аргачлалаар яаж үнэн бодитой тогтоох вэ?
-Ерөнхий боловсролын сургууль, цэцэрлэгийн нэг бүлэгт байх хүүхдийг тоог хөрөнгө оруулалт, үйлчилгээ үр дүн, үр өгөөж, үр ашигтай байхаар хэрхэн оновчтой зөв тогтоох вэ?
-Ерөнхий боловсролын нэг сургууль, цэцэрлэгт байх хүүхдийн тоо, анги бүлгийн тоог хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой тогтоох вэ?
-Ерөнхий боловсролын сургууль, цэцэрлэгийн байршил, түвшинг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой хийх вэ?
-Ерөнхий боловсролын сургууль, цэцэрлэгийн бүтэц, бүрэлдэхүүн, ажил үүргийн хуваарь, цалинг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой тогтоох вэ?
-Ерөнхий боловсролын сургалтын төлөвлөгөөний хичээлийн тоо, үзэх цагийн хэмжээг хөрөнгө оруулалт, үйчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн зөв тогтоох вэ?
-Ерөнхий боловсролын сургууль, цэцэрлэгийн багшийн нэг цагийн ажлын багтаамж, бүрдэл, багшийн тоог хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар үйл ажиллагааг өртөгжүүлэх аргачлалаар яаж бодитой өртөгжүүлж, цэнэжүүлэх вэ?
-Ерөнхий боловсролын үйлчилгээний чанар, хөрөнгө оруулалтын үр дүн, үр ашиг, үр өгөөжийг нэмэгдүүлэхэд нөлөөлж буй хүчин зүйлийг гүйцэт тогтоож ерөнхий боловсролын эрэлт, нийлүүлэлтийн хийгээд үйлдвэрлэлийн функцийг хэрхэн бичих вэ?
-Ерөнхий боловсролын салбарын төсөл, хөтөлбөрийг хөрөнгө оруулалт үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн бодитой төлөвлөж хэрэгжүүлэх вэ?
-Ерөнхий боловсролын бодлогыг үйлчилгээ, хөрөнгө оруулалт үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой зөв боловсруулж хэрэгжүүлэх вэ?
  Хоёр. Дээд боловсролын түвшинд:
-Их сургуулийн нэг кредит цагийн агууламж, бүрдэл бүхий бүхэл, багц ажлын үнэ цэнийг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар үйл өртөгжүүлэх аргачлалаар  яаж үнэн бодитой тогтоож болох вэ?
-Их сургуулийн сургалтын төлөвлөгөө, хөтөлбөрийн агуулгын үнэ цэнэ, үзэх хичээл, судлах цагийн тоо хэмжээг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой зөв тогтоох вэ?
-Их сургуулийн хичээлд суух оюутны тоог хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой тогтоох вэ?
-Манай улсад байвал зохих их дээд сургууль, коллеж, мэргэжлийн сургалт, үйлдвэрлэлийн төвийн төрөл, хэлбэр, тоо,  байршил, түвшинг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой тогтоох вэ?
-Их дээд сургууль, коллеж, мэргэжлийн сургалт, үйлдвэрлэлийн төвийн бүтэц, бүрэлдэхүүн, удирдлага, зохион байгуулалт, ажил үүргийн хуваарийг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой тогтоох вэ?
-Их сургуулийн профессорын нэг цагийн ажлын багтаамж, бүрдлийг үйл ажиллагааг өртөгжүүлэх аргачлалаар хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар яаж бодитой өртөгжүүлж, цэнэжүүлэх вэ?
-Их сургуулийн багшийн ангилал, түвшинг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар яаж бодитой өртөгжүүлж, цэнэжүүлэх вэ?
-Дээд боловсролын үйлчилгээний чанар, хөрөнгө оруулалтын үр дүн, өгөөжийг нэмэгдүүлэхэд нөлөөлж буй хүчин зүйлийг гүйцэт тогтоож мэргэжлийн болон дээд боловсролын эрэлт, нийлүүлэлтийн хийгээд үйлдвэрлэлийн функцийг хэрхэн бичих вэ?
-Дээд боловсролын бодлого, төлөвлөлтийг хөрөнгө оруулалт, үйлчилгээ үр дүн, үр ашиг, үр өгөөжтэй байхаар хэрхэн оновчтой зөв боловсруулж хэрэгжүүлэх вэ?
           Манай улсын боловсролын үйлчилгээ, түүний өртөг нь олон улсын жишиг үнэлгээнээс харьцангуй их доогуур байдаг. Мөн боловсролын чанар, хүний хөгжлийн индекс, дотоодын нийт бүтээгдэхүүний цэнээр ч бас л олон улсын дундажаас нилээд доор эрэмбэлэгддэг. Энэ хоёр хувьсагчийн хооронд хамаарал бий.  Учир нь бодит өртөг, зардлаар ажил бодитой хийгддэг. Хуурмаг нь харин хуурмагийг л хийдэг.
Манай улсын боловсролын чанарыг нэмэгдүүлж түүний хувь хүний амьдралд нөлөөлөх хувь нэмрийг өсгөх, нийгмийн сайн сайханд нэмэх үнэ цэнийг өсгөх,  дотоод нийт бүтээгдэхүүнд эзлэх жинг нэмэх, хамгийн чухал нь, боловсролоор эх орондоо мэдлэгийн даяар зах зээлээс хөрөнгө босгох, хөрөнгө оруулах, “мэдлэгийн хүчийг хөгжлийн хүч” болгохын тулд боловсролын эдийн засгийн тулгамдсан асуудлыг бүхэл, цул байдлаар шинжилж, төв, зөв хандлагаар харж үе шаттайгаар үргэлжлүүлж шийдвэрлэх хэрэгтэй. Үүнийг хийх арга  зам нь  боловсролын салбарын үйл бүрийг үнэн өртөгжүүлж цэнэжүүлэх явдал мөн. Бас үнэн бодитой өртөгтэй үйл бүрээс үр дүн, үр ашиг, үр өгөөж нэхэх соёл юм. Ингэж өртөгжүүлж, үр өгөөж нэхэхэд бидэнд боловсролын эдийн засгийн ухаан, бас боловсролын эдийн засаг хэрэгтэй байна. Боловсролын эдийн засгийн ухааныг ийм учраас энэ улсад бодлоготой хөгжүүлж, үндэсний хүний нөөцийг бүтээх бодит шаардлага байна. Авах арга хэмжээ, шийдвэрлэх арга замын хэсэг болсон энэ шаардлага эргээд  өөрөө бас л боловсролын эдийн засгийн бодлогын түвшний бас л нэг асуудал болж байна.


[1] Education economics or  the economics of education
[2] Economy of Education
[3] Productiveness, productivity, effectiveness
[4] Efficiency
[5] Return
[6] Consulting
[7] Expertizing
[8] LMS, learning management systems’ software