This event was held Friday October 27, 2017
Kate McCain - Undercounting the gift givers: issues when tallying acknowledgements in life sciences research
ongoing research effort - she originally worked on this 20 years ago but has come back to it recently. Background - model organisms - useful to organize research around. Community databases, stock centers, community databases, community ethos wrt sharing.
Ways to focus research - by journal is often done, but she uses this model organism. She is looking at 1980-2004 during growth phase when there is more sharing because nascent research area. And she is looking at acknowledgements.
Compared to citations - acknowledged most likely to be alive.
Personal ack vs. funding - she's interested in personal ackn. "peer interactive communication"
May be lots of different places: end note, methods section, end of text with no section label, ... No control or standardization of how people are named, what granularity they are thanked for, etc.
WoS mostly gets funding ack, and only secondarily sweeps up some personal ack (if they are in the same block, which is not always the case).
Undercounting big deal: text extraction relying on formal ack section. personal name disambiguation. Sampling or single year studies.
Check her slides to see what she found where. She also categorized types of ack - animals, software, data, editing, etc.
Top 15 individuals listed - first few time periods dominated by University of Oregon - founders and suppliers of fish early on.
She then went through profiles of some individuals with the diversity of how they appeared.
Trends - fewer examples of thanking for research materials - have their own, get from repository, or get from stock center
questions: manually - yes? learn things to help automate - yes, but lots and lots and lots of ways to trip up. Also just picking up surnames is not enough because then get some citations mixed in, named equations/methods, etc.
Reminds me of: http://www.npr.org/2017/03/30/521931310/-thanksfortyping-spotlights-unnamed-women-in-literary-acknowledgements
questions: in the lab outside of the lab. also tracking people who are frequently acknowledged and not often co-authors/cited
questions: comment - collaboration - set up something from PMC data (already coded in XML), but only using ack section and not the Materials & Methods (M&M) section.
Isabelle Dorsch - Relative Visibility
How well known. She's comparing personal publication list and information services (like WoS).
Relative visibility (IS) = (d/r)*100
d= in information services, r=publication list
Previous work - Cronin & Stock, and ISSI board study
Issues - finding the personal list, is it up to date and complete, is it structured to be used at all, what types of documents to keep (novels? newspaper articles?), keep in press?
(*discussion of this on SIGMETRICS really found that a combined edited list is probably best, but these aren't universally available - list maintained by information service but updated by author)
Which information service matters (of course) - visibility to one field when author publishes in multiple. Conference paper coverage, book coverage, etc.
questions: new author - only two publications - 100% (they only looked at established authors). Very dependent on the database
Judit Bar-Ilan - CiteScore vs JIF and Other Journal Indicators
Criticisms of JIF but still heavily used. Standard definition. Criticisms like lack of transparency. Things in the numerator not included as "citable items" in the denominator. Also now offer a 5year JIF
Citescore - publication window 3 years. They count all items so no numerator/denominator coverage mismatch. Transparent - can see all the citations that are covered. Freely available. Some criticism that covers too many different document types
EigenFactor, SJR, pagerank type indicators - more weight to more influential sources
Article Influence - normalized - average journal is 1.
She looked at correlations - for those sources that appear on most sources.
Quite high - CS-JIF 0.94,
HOWEVER - Lancet is 5 in JIF, 314 in CS - so huge differences and she suspects due to notes, editorials, etc.
Top 20 by CS are almost all review journals (Annual Review of... , Progress in... )
Eigenfactor doesn't include journal self-citation, and doesn't correlate as well with others.
Note also that even though high correlation, there are these big differences.
question: comment - real correlation between size of journal and JIF, Eigenfactor is the only one that corrects for this.
Zhao, Mao, & Kun Lu (speaking, not student) - An Exploratory Study on Co-word Network Simulation
Network centrality and other network measures for co-word network. Are they correlated. Are there differences in disciplines in these measures. Looking at generative process of a co-word network.
Q: co-word can mean 3 different things: words that appear in the text, co-descriptor - uses carefully assigned things, keywords plus - is another thing separately (not controlled, but titles of articles cited). Are you simulating second hand natural language assigned things.
Antoine Archambault, Philippe Mongeon (speaking), Vincent Larivière - The concentration of journal use in Canadian universities
As Canadian universities have to cut big packages due to budgetary issues.
Evaluating - downloads statistics from the 28 universities (~300 Excel files, 5M lines), references (articles written by authors at these universities citing these journals) perceived importance of journals (what journals do you perceive important to your research, your teaching) 23 of 28 universities, 5,500 participants (of which 3k from their university so actually disappointing response)
Cleaning important journals - title disambiguation, manual validation, classification by major disciplinary area (AH, SS, BM, NSE) - WoS, NSF, Ulrich's, Google, also verified research journal and not newsletter, etc.
47k unique journals.
Priority journals - 80/20 rule - anything in top of 80% downloads, references, mentions (10% of subscriptions account for 80% of any of these measures)
66% of the 47k journals are not in the top anywhere.
Broke out by publishers - Springer 80% of publications were not in anyone's top. Sage only 22% were not in anyone's
Only 41.6% overlap of core journals between universities
Correlation of cites, downloads, mentions (cites are super lengthy for institutions to do themselves can they just use downloads?) - answer is no. Have to use the 3 measures, not completely correlated.
Q: can you some sort of demand driven acquisition
Q: are there libraries of record - keep even if don't use locally
Q: combining visibility presentation earlier with this.
Christine Meschede Cross-Metric Compatibility of Altmetrics: Consistency of the Metrics from PlumX and Altmetric.com
(posting before notes - battery going right now - will try to update)