News & Views: What proportion of OA is Diamond OA? Can AI help us find out?

Dan Pollock and Ann Michael • February 20, 2025

Overview


A recent post on the Open Café listserv posed a question about the true extent of fee-free open access publishing, but it noted the incomplete coverage of the data cited. We have more comprehensive data, but just as we started our analysis, DeepSeek’s release sent markets into turmoil. The stage was set for a timely experiment. We first answer the question using our data. Then we see how the AI did.



Background



What proportion of open access is not paid for by APCs? In discussing this, a recent Open Café listserv post cited studies by Walt Crawford – a librarian, well-known in the academic library and OA communities for his analysis of open access. He has paid particular attention to “diamond” OA journals, which charge neither readers nor authors. His studies are based on data from the Directory of Open Access journals (DOAJ). Excellent though both sources may be – and, full disclosure, we contribute to the DOAJ – the DOAJ’s remit covers only fully OA (“gold”) journals.


As listserv founder Rick Anderson noted, “By counting only articles published in DOAJ-listed journals, Crawford’s studies radically _undercount_ the number of APC-funded OA articles published – because DOAJ does not list hybrid journals, which always charge an APC for OA and which produce a lot of genuinely OA articles (though exactly how many, no one knows).”



Using our data


Actually, we do know … or at least have some fair estimates of hybrid OA. Our data allows us to determine the share of open access output in APC-free journals, as follows.


Sources: Publishers' websites, OpenAlex, DOAJ, Delta Think analysis. © 2025 Delta Think Inc. All rights reserved.


The charts analyze shares of open access articles published in both fully OA and hybrid journals.


  • They show the split between output in APC-free journals (“No APC” in the orange) and in journals where APCs were charged (“APC” in the grey).
  • Journals count as APC-free if they were free of APCs in the year of publication.
  • The data cover the five years 2018-2023 (consistent with the Open Café post).
  • Around 13% of all OA output in our sample is APC-free (left-hand pie).
  • As we often discuss, individual publisher’s experiences may vary. The right-hand pie shows the subset of output for Physics. Note how the APC-free share rises to almost 47% in this case.
  • If we choose a slightly earlier timeframe – which will become relevant below – we see very little change. Over 2018-2022, there was very little change in the numbers.
  • Our data allows further analysis, for example by over 200 subjects, or comparing society titles with others. Please get in touch to find out more.


Our estimate of a 13% APC-free share of the total market compares with around 33% noted in the Open Café post, citing Walt Crawford’s data.


What about AI?


If we wanted to use public sources, we would consider this to be a classic web research problem. This should suit generative AI (GenAI). ChatGPT is arguably the most well-known example of this technology. With disruptor DeepSeek grabbing headlines recently, it seemed opportune to compare both with the benchmarks from our own data. We can also test DeepSeek’s claims to deliver an equivalent level of performance to ChatGPT.

The actual question


Users of GenAI tools rapidly discover that the phraseology and nuances of questions asked – aka the prompt – are vital. The more detail and direction you give them, the better. We used:


what proportion of open access is diamond open access? list numbers of articles


and share for each of the last 5 years


Some prompts can run to paragraphs. But we wanted to test the AI’s ability to process a natural sounding and simple question, to interpret the term Diamond, and to see if it would spot the inclusion of fully OA and hybrid. We needed to direct it to the specifics of our required results. The 5-year period was mentioned in the Open Café post. We chose a “one shot” prompt (not asking the model to refine its first answer), as we wanted to see how our virtual research assistant would perform without further direction.



The answer



The table below shows the results.


*ChatGPT 4o did not give the answer directly. Instead, it gave Diamond OA & APC-funded OA as separate shares of all output, so we needed to calculate Diamond share of OA.

ChatGPT 4-turbo with “Reasoning” switched on.


The table above summarizes the AIs’ answers.


  • The technology continues to evolve, so there are different versions of it – different models – available. We see this for ChatGPT, where different versions of the same AI give different responses.
  • A particular model may give different replies if you ask the same question more than once. ChatGPT seemed stable, but DeepSeek exhibited this behavior, as indicated in the “Attempt” row.
  • ChatGPT 4-turbo has a “Reason” feature. When enabled, this shows its method (reasoning) and then appears to return more verbose results. The annual figures we cited result from the “Reason” mode. Without it, it returned similar numbers to ChatGPT 4o.
  • Looking at share of OA output, there’s a big variation between ChatGPT and DeepSeek. There’s a smaller variation between the two DeepSeek attempts. DeepSeek did very well and got close to our estimates.
  • The DeepSeek attempts showed big variations in the article numbers. DeepSeek attempt 2 returned similar article numbers to ChatGPT 4-turbo, but the respective shares were roughly 2.5x different.
  • The models were only able to show figures up to 2022.


A research assistant’s ability to explain their results is important, and so we assessed the models on their communication ability. In practice both research assistants and AI can be asked to refine their work. But for the purposes of this analysis, we used a one-shot prompt.


  • The word count of replies varied. Chat GPT-4o had single-paragraph, single-number brevity. The others mixed bullets, prose, and tables.
  • Only one model added a hyperlink to its source(s), noted as follows.
  • ChatGPT-4o cited encyclo.ouvrirlascience.fr. It was the one model to include a link.
  • ChatGPT 4-turbo volunteered sections describing definitions, included headline numbers from some sources, tabulated yearly estimates, discussions comparing results from different sources, caveats on data limitations, and a summary. It cited Wikipedia as its source. It noted that its 5 years of estimates were based on data from 2017-2019, but it’s unclear whether it calculated them or was citing some language it found.
  • DeepSeek volunteered brief discussions about definitions, trends, annual estimates, and key sources. The sources seemed to be directions for further reading, including the OA Diamond Journals Study (2021) by cOAlition S and OPERAS, which it considered to provide the most comprehensive view, the DOAJ, and Web of Science/Scopus (noting they underrepresent Diamond OA journals, as many are smaller or regionally focused).
  • The verbose models attached general caveats to the data, but none picked up on the subtleties of defining Diamond. If a journal has only one APC-free year, say as an introductory discount, does it really count as a true Diamond or sponsored journal?


Conclusion


In working with and assessing GenAI models, it is important to set expectations correctly. They are more like an inexperienced intern than a seasoned researcher. Prompts need to be refined, several attempts made, and human judgement exercised over the results.


DeepSeek’s results were surprisingly good. It undershot our estimate by a few percentage points. ChatGPT overestimated considerably – perhaps reflecting its use of the problematic sources noted in the original listserv post. The additional material selected by the models was useful in explaining the concepts and drivers behind that data. The differences between our samples and those the models cite explain some of the variance in the data. But this doesn’t explain some of the big differences – e.g. same #s articles, but different shares or vice versa, with 2.5x variance.


Perhaps the GenAI models, in their current forms, are not ideally suited this sort of numerically focused research. However, they appear to be progressing in the right direction. (Note: We did not attempt this query on either of the Deep Research models, currently offered by OpenAI and Google Gemini.)


For fun, we tail-ended our assessment by asking each of our AIs “what's the answer to life, the universe and everything?” ChatGPT went to the Hitchhiker’s Guide to the Galaxy answer of 42. DeepSeek was more expansive, noting how the question encompassed broader philosophic endeavors. However, a light-weight offline version of DeepSeek that we experimented with gave perhaps the best reply. It concluded that, while 42 is a clever punchline,


“…it also reflects on the human tendency to search for significance where there might not be one.”


Fair point.

---


This article is © 2025 Delta Think, Inc. It is published under a Creative Commons Attribution-NonCommercial 4.0 International License. Please do get in touch if you want to use it in other contexts – we’re usually pretty accommodating.


---


TOP HEADLINES


What Gets Missed in the Discourse on Transformative Agreements – February 12, 2025

"Transformative agreements have a crucial role to play in the transition to open access, and while they may not fix structural problems, they are a reflection of the dynamic publishing landscape and the diverse needs of the academic community."


Serbia Adopts Open Science Platform 2.0 – February 4, 2025

"The Ministry of Science, Technological Development and Innovation of Serbia has adopted a new national open science (OS) policy – the Open Science Platform 2.0 – that applies to all Serbian publicly-funded research projects and programmes. EIFL welcomes the adoption of the policy, which significantly expands the scope of OS efforts in Serbia, and updates the country’s first national OS policy, which was adopted in 2018."


cOAlition S to sunset the Journal Comparison Service in April 2025 – February 3, 2025

"cOAlition S announces it will discontinue the Journal Comparison Service (JCS), effective from 30th April 2025...The JCS was launched in September 2022 to support the Plan S principle that Open Access fees should be transparent and commensurate with the publishers’ services. It was developed following detailed consultations with librarians, publishers, legal experts, and software developers with the aim to shed light on open access publishing fees and services."


Announcing a New Report on Open Educational Resources – January 22, 2025

"In the fall of 2023 we announced the launch of a new research project, funded by the William and Flora Hewlett Foundation, designed to assess the impact and implementation of open educational resource (OER) initiatives at public institutions of higher education. Today, we are publishing the resulting report, based on an initial literature review and interviews with OER leaders in four US states."


European Diamond Capacity Hub launched to strengthen Diamond Open Access publishing in Europe – January 21, 2025

"The European Diamond Capacity Hub (EDCH), a pioneering initiative to advance Diamond Open Access publishing across Europe, was officially launched on 15th January 2025, in Madrid. OPERAS, the research infrastructure for open scholarly communication in the social sciences and humanities, will serve as the fiscal host of the European Diamond Capacity Hub."

 


OA JOURNAL LAUNCHES


All EMS Press journals open access in 2025 following another successful Subscribe To Open round - February 5, 2025

"EMS Press publishes all journals open access in 2025 for the second consecutive year, following a successful Subscribe To Open (S2O) round...all 22 journals in our Subscribe To Open (S2O) programme will be published as open access for the 2025 subscription period."


By Lori Carlin April 29, 2025
“I had a wonderful teacher about animal behavior.” – Jane Goodall Jane Goodall was renowned for her ability to understand chimpanzees in the wild simply by observing their day-to-day behaviors. Her subjects were her teachers. She immersed herself in their natural environment, which allowed her to gather authentic insights into her subjects’ social structures, tool use, communication, and cultural behaviors. This ethnographic style of research has now been adopted by the business world to gain deeper insights into customer needs, workflows, and challenges. What is Contextual Inquiry? Contextual Inquiry, as it’s frequently called in the commercial world, is a user-centered research method. Like Dr. Goodall in the field in Tanzania, Contextual Inquiry involves observing and interviewing users in their natural environments while they perform routine tasks, allowing market researchers to understand user behaviors, goals, motivations, and pain points in real-world contexts. The methodology is guided by four key principles: Context: Observations and interviews occur in the user’s actual environment. Partnership: Researchers collaborate with users to understand their processes. Interpretation: Insights are shared with users during the interview for clarification. Focus: The interaction is steered toward topics relevant to the research scope. Why Contextual Inquiry? Contextual Inquiry illuminates user behavior and workflows and can be used to inform solutions including product development, processes, and policies. 1. Deep Understanding of User Needs By observing users in their natural environment, contextual inquiry uncovers hidden needs, pain points, and behaviors that users may not articulate in surveys or traditional interviews. This identifies solutions that address real user challenges rather than assumed ones. 2. Authentic Data Collection Unlike lab-based studies or self-reported methods, Contextual Inquiry captures authentic behaviors as they occur. This helps reveal workarounds, unconscious habits, and environmental factors that influence user behavior. 3. User-Centered Design and Focus The method places users at the center, ensuring solutions are tailored to their needs and preferences. 4. Reduced Risk and Costs By identifying real-world user behaviors and challenges, Contextual Inquiry increases the likelihood that resulting solutions, products, processes, and policies meet user needs. 5. Informed Strategy Insights from contextual inquiry help prioritize users’ needs and challenges, identify new opportunities, and ensure alignment of product and strategy roadmaps with market demands. 6. Empathy Building Direct observation helps teams develop empathy for users by experiencing and understanding their challenges firsthand. Use Cases and an Idea Delta Think uses a Contextual Inquiry methodology to apply the principles above in several different types of projects: Understanding specialized work processes and designing complex systems or workflows Exploring new product opportunities Redesigning existing products Informing processes and policies One way Delta Think can use Contextual Inquiry to help publishers is around the use of AI by researchers and authors. For example, before developing any guidelines, it is critical that publishers first understand how these tools are being used. Using Contextual Inquiry to observe how researchers and authors are currently using AI allows you to build policies that address specific use cases and provide meaningful guidelines based on actual behaviors. At Delta Think, we know how powerful a tool Contextual Inquiry can be for creating a user-centered culture grounded in decisions based on real-world insights and observations. We specialize in uncovering evidence for our clients, discovering what that evidence means for them, and using the knowledge gained to build customer-driven products, policies, and actionable business and publishing strategies. Ready to start the conversation? Please reach out today to discuss how our market research expertise, including Contextual Inquiry, can help your organization unlock key insights to support development and growth.
By Dan Pollock and Heather Staines April 22, 2025
In March 2025, we looked at the latest Article Processing Charges (APCs) . This month we focus on how prices have risen relative to inflation. As APC price increases fall back to trend, what does this mean in real terms? Background Each year we survey the list Article Processing Charges (APCs) of more than 30 major and significant publishers. Going back to 2015, the dataset includes more than 20,000 unique titles and 150,000 title per year combinations. Going into 2025, we saw price increases fall back to long-term trends from their unusually high increases in 2024. Fully OA (“gold”) journal list prices across our sample rose by around 6.5%, compared with a 9.5% increase this time last year. Hybrid list prices rose by an average of 3%, compared with 4.2%. Last year’s price rises were above long term trends, but overall we found they were rising below inflation. How does this hold for this year’s price increases? We again use the global Consumer Price Index (CPI) as our inflation index, as we consider it to represent the most realistic view of our marketplace. Prices exclude zero APCs, so we can see the effects for instances when publishers choose to charge APCs. Are APCs becoming cheaper or more expensive? The chart below shows how increases in all list APCs work out in real terms for both hybrid and fully OA journals.
By Heather Staines April 7, 2025
We are proud to share a video recording of our March News & Views companion online discussion forum! Join us to hear the latest trends around APC data, including APCs for both fully OA and hybrid journals. We'll talk about what we're seeing in relation to recent years and discuss the broader context for the APC market. If you missed the session, or if you attended and would like to watch/listen again, or share forward with friends, please feel free!
By Lori Carlin March 28, 2025
Delta Think is currently spearheading an industry market research survey to authors and researchers across the scholarly community designed to provide insight into the impact of potential US federal funding reductions on their research. The survey addresses topics such as publication volume, their ability/allowance for peer review, conference participation and attendance, influence on their research scope and topics, and more. Working in collaboration with nearly 25 scholarly societies, we are launching this initiative to capture the real-world impact of these potential changes in order to help societies better plan and support their members, researchers, and authors. The results of the survey will provide scholarly publishers with systematic, quantitative voice-of-market data to inform evidenced-based strategy development and scenario planning in a rapidly changing funding landscape and policy environment. The survey opens this week, with each participating society distributing the link to their own communities. All participating societies will receive an in-depth analysis of the full survey results, filtered by various demographics such as country, career stage, and discipline, as well as options for Delta Think to analyze their specific community data or the raw data from their specific community so they can analyze it themselves. Delta Think has designed the survey and will conduct all the analysis of the results.
By Dan Pollock and Heather Staines March 13, 2025
This month we look at our latest data about Article Processing Charges (APCs). Per article pricing is a fundamental building block for all paid publishing models, so our review provides an invaluable insight into how costs of open access continue to evolve. APC prices in general continue to increase, but at a slower rate compared with this time last year. Important nuances in the distribution of prices continue to affect the value and cost of paid publishing models. Background Each year we survey the list Article Processing Charges (APCs) of a sample of major and significant publishers. Covering more than 20,000 titles going back to 2016, our dataset represents one of the most comprehensive reviews of open access pricing. To compare like for like, we consistently analyze non-discounted, CC BY charges. We take a snapshot at the end of every January, so we can track yearly changes while controlling for the different times of year that publishers may update prices. Our statistics exclude zero or unspecified APCs, although these are included in our underlying data. This allows us to understand trends where publishers choose to charge APCs without skewing averages. We run separate analyses around APC-free models. Headline Changes Going into 2025, we have seen APC pricing increasing but falling back to long-term trends. Fully OA APC list prices across our sample have risen by around 6.5% compared with 9.5% this time last year. Hybrid APC list prices have risen by an average of 3% compared with 4.2% this time last year. Maximum APCs for fully OA journals remain at $8,900. Maximum APCs for hybrid journals now top out at $12,690 (up $400 from last year). Big jumps in prices happened last year, driven by high inflation. In 2020-2021 prices were driven up when high-impact journals began offering OA options for the first time. In both cases, increases subsequently fell back to averages. Underlying trends continue. There are around 2.6x more hybrid journals than fully OA ones, down from 2.9x a year ago. Hybrid journals follow (or, rather, set) a similar pattern to the market overall. On average, fully OA prices are around 64% of those of hybrids. This is a couple of percentage points higher than long term trends. Around 31% of our sample of fully OA journals charge no APCs. (We have separately analyzed the number of articles in OA journals.) Price rises vary significantly by discipline. Arts & Humanities and Social Sciences have seen particularly large average increases, especially in fully OA journal prices. Price Distribution Market-wide headline price changes mask important nuances. We have discussed previously that the most important nuance lies in the spread of prices within a given publisher’s portfolio. For example, if the bulk of a publisher’s journals lie toward the lower end of its pricing, with just a few journals priced at the high end, the average (mean) price will be higher than most authors pay. The following figures show how the spread of prices plays out in the market across our sample of publishers. The figures are outlines of histograms, showing how many titles sit in various price bands over the successive years of data we have curated. The red line shows the most recent year’s prices. The lines become greener as they go further back in time. Subscribers to Delta Think’s Data and Analytics Tool can see full details of axes. Hybrid Prices The spread of price bands for hybrid journals is shown in Figure 1 below.
By Diane Harnish and Meg White February 27, 2025
User information needs as well as funding models are evolving rapidly, as evidenced by Clarivate’s recent move to phase out perpetual access purchases for print, eBooks, and digital collections by the end of 2025. Taking a hard look at how these assets contribute to your portfolio and overall organizational strategy has never been more critical. A holistic books program assessment can help you think intentionally about how books and book-based content can help meet customer and market needs. Publishing and Product Strategy A market-driven publishing and product strategy begins with an understanding of customer information needs. What markets, segments, information needs, and challenges are present? How can customer information needs be addressed? What role can our book content play? How can we differentiate our solutions? Can our book content contribute to a unique value proposition? Thinking creatively about how your content meets market needs is critical; think solutions, not printed pages and chapters. Commercial Strategy A detailed commercial strategy, supported by proper resources, is fundamental to success. Leveraging a clear understanding of customer preferences and delivering messaging that resonates with your specific market segments and use cases is essential. What are the best methods to generate market awareness? When and how should we communicate with key audiences? What messages resonate best? What sales and marketing capabilities do we have internally? Where do we need to partner to reach core audiences? How do we meet global needs? Do we have the appropriate access, pricing, and distribution models in place to meet customer expectations? What do we need to do directly? Where should we cultivate successful channel partnerships? And you don’t have to go it alone; a commercial strategy is best formulated and executed by a combination of internal and external resources. Technology Infrastructure Is your technology optimized to support your book program? From agile content management systems to product platforms to customer relationship management tools, the right tools enable your content and commercial strategy. What systems do we need to ensure efficiency in our publishing processes and quality and integrity in our content? What technologies and platforms do we need to build market-responsive products? What systems do we need to communicate effectively and meaningfully with our customers, including authors? Are we best served by building these systems or partnering? Successfully integrating and leveraging new technologies, such as AI, requires a fundamental understanding of markets and customer information needs . The Numbers Financial metrics are a key measure of the health of any program. An in-depth assessment of a program’s recent performance is a vital tool to help identify strengths, weaknesses, and gaps, and help to surface areas for improvement and corrective action. A financial analysis will clarify: What is our book and content annual output? Is it sufficient to support our strategy and meet customer and market needs? What is our cost structure? Our pricing strategy? Do they align with industry and market norms and expectations? Do we have the appropriate mix of internal and external resources in place to support our strategy? How can we best align our financial performance to contribute to the organization’s larger strategy? Beyond red ink or black ink, financial analysis will provide answers to these questions. Assessing Your Book Program Delta Think partners with publishers to do the foundational analysis necessary to understand how your book and book-based content can be a vital part of your content portfolio and support your organization’s goals and objectives. Our processes, including program benchmarking, stakeholder interviews, surveys, and workshops, combined with expert landscape research and analysis ensure you are building a content strategy that is market-focused and customer-driven. Contact Delta Think at info@deltathink.com to set up a time for a video call to learn more. We will also be attending the London Book Fair, March 11-13, 2025, if you’d like to schedule an in-person chat.
By Dan Pollock and Ann Michael February 11, 2025
Overview Following the 2024 US election, the new US administration has instructed employees in some key federal agencies to retract publications arising from federally funded research. This is to allow representatives of the administration to review the language used, to ensure it is consistent with the administration’s political ideology. In this special edition of News & Views, we quantify how many papers might be affected and estimate their share of scholarly publishers’ output. The initial numbers may be small, but we suggest the effects on scholarly publishing could be profound. Background On 20 January 2025, Donald J. Trump took office as the 47th President of the United States. Within hours he signed an Executive Order 1 (EO) 14168 proclaiming that the US government would only recognize two sexes, and ending diversity, equity, and inclusion (DEI) programs inside federal agencies. The following day, his administration instructed federal health agencies to pause all external communications – “such as health advisories, weekly scientific reports, updates to websites and social media posts” – pending their review by presidential appointees. These instructions were delivered to staff at agencies inside the Department of Health and Human Services (DHSS), including the Food and Drug Administration (FDA), the Centers for Disease Control (CDC) and Prevention, and the National Institutes of Health (NIH). The events that followed are important, as they directly affect scholarly papers and our analysis. A memo on 29 January instructed agencies to “end all agency programs that … promote or reflect gender ideology” as defined in the EO. Department heads were instructed to immediately review and terminate any “programs, contracts, and grants” that “promote or inculcate gender ideology.” Among other things, they were to remove any public-facing documents or policies that are trans-affirming and replace the term “gender” with “sex” on official documents. By the start of February, more than 8000 web pages across more than a dozen US government websites were taken down . These included over 3000 pages from the CDC (including 1000 research articles filed under preventing chronic disease, STD treatment guidelines , information about Alzheimer’s warning signs, overdose prevention training , and vaccine guidelines for pregnancy). Other departments affected included the FDA (some clinical trials), the Office of Scientific and Technical Information (the OSTP, removing papers in optics, chemistry and experimental medicine), the Health Resources and Services Administration (covering care for women with opioid addictions, and an FAQ about the Mpox vaccine). Around this time, it further emerged that CDC staff were sent an email directing them to withdraw manuscripts that had been accepted, but not yet published, that did not comply with the EO. Agency staff members were given a list of about 20 forbidden terms, including gender, transgender, pregnant person, pregnant people, LGBT, transsexual, nonbinary, assigned male at birth, assigned female at birth, biologically male, biologically female, and he/she/they/them. All references to DEI and inclusion are also to be removed. The effects of the EO Commenting on the merits of policy and ideology lies beyond our remit. However, when these matters affect the scholarly record – as they clearly do here – then they are of interest for our analyses. Specifically, what might the effects of the EO be on the publication of papers, and what effects might accrue from withdrawal of research funding? If federal agencies are being instructed to withhold or withdraw submissions, then, to quantify what this might mean to publishers, we have estimated the volume of output from a few key federal agencies. It is summarized in the following chart. 
By Lori Carlin January 23, 2025
Emerging technologies are reshaping how we create, distribute, and consume content. Publishers face the critical task of making smart technology investments to stay competitive and enable strategic objectives. How do you ensure that your next tech purchase aligns with your organization's needs and goals? Enter the needs assessment process – your roadmap to making informed, strategic technology decisions. From defining clear objectives to creating a comprehensive RFP, these best practices will help you navigate the decision-making process with confidence and ensure that your investments deliver value for your organization and your customers. Technology is not a solution; it is a tool. The temptation to adopt technology without a clear definition of what you are trying to achieve is an all too common (and usually very costly) mistake. Does your strategy include delivering a more personalized experience for your users? A customer data platform may be the right technology. Interested in using AI to build research integrity into your editorial process? Perhaps it’s time to revisit the capabilities of your editorial management system. Looking to support education and learning for students, faculty, and professional learners? Maybe it is time to evaluate formal learning management systems. Once you are confident about what you are seeking to achieve, the real work begins. Here are the key components that will help lay the foundation for a successful process from inception to deployment: Analyze Current State: Audit existing systems and processes to understand current capabilities and limitations. Conduct a Gap Analysis: Identify gaps between current capabilities and desired future state. Collect and Analyze Data: Gather qualitative and quantitative data from staff, users, customers, industry benchmarks, and about existing systems. Consider Resources and Constraints: Assess available resources, including budget, skills, and time. Research Solutions: Investigate potential technologies and/or types of solutions that could address identified gaps. Prioritize Needs: Work with stakeholders to prioritize needs based on impact and feasibility. Create RFP: After identifying prioritized needs and potential solutions, develop an RFP that clearly outlines project objectives, specific requirements, evaluation criteria, budget, and timelines. Distribute the RFP: Identify vendors with fit for purpose solutions and capabilities and distribute. Evaluate Proposals: Review vendor responses against established criteria and prioritize them based on how well they meet your needs. Plan for Adoption and Training: Consider the change management aspects of introducing new technology and processes. Be sure to develop a plan for user adoption, training, and ongoing support in your new systems. Technology as a Strategic Ally A methodical needs assessment is not just a procurement exercise – it is a strategic opportunity to reimagine how technology can transform your organization. The most successful technology investments are those that solve real problems, align with organizational goals, and empower your team to work more efficiently and creatively. Don’t fall into the trap of just moving what you are currently doing over to a new system. This is an ideal occasion to think about how you would design workflows and processes if you were to start from scratch and use that framework to evaluate the new capabilities available. You don’t want to duplicate what you are doing today; you want to step back and take the opportunity to build something better whenever possible. Customer Data Platform? Editorial Management System? Learning Management System? Something Else? Delta Think partners with publishers to do the foundational and implementation work required to ensure that technology decisions match the organization’s capabilities, fit the budget, and are grounded in voice-of-customer data. Our processes, including stakeholder interviews, surveys, and workshops, combined with expert landscape research, analysis, and assessments, underpin technology decision-making that is market-focused and customer-driven. If your 2025 objectives depend on or are enabled by technology, we’d welcome the opportunity to help you learn, plan, achieve. Please contact us today to start the conversation.
By Dan Pollock and Heather Staines January 14, 2025
A number of sources provide information about patterns in the overall scholarly journals market. However, as we so often mention in our analyses, important nuances lie beneath the headlines. This month we explore just how much variation exists and highlight the importance of specificity. Background As part of our annual market updates, we estimate the proportions of open vs. subscription access content each year. Over the last few years, we have observed how OA has approached 50% of output, but we note that it has yet to punch through that number. However, this headline varies greatly depending on your area of publishing. An example from physics The chart below shows the nuances across just a few of the 200+ subjects that we track.
By Dan Pollock, Ann Michael December 10, 2024
This month’s topic: How much content can AI legally exploit? Scroll down to read about this topic, along with the latest headlines and announcements. Delta Think publishes this News & Views mailing in conjunction with its Data & Analytics Tool . Please forward News & Views to colleagues and friends, who can register to receive News & Views for free each month. Delta Think will be attending several upcoming conferences, including APE (Jan 14-15), NISO Plus (Feb 10-12), and Researcher to Reader (Feb 20-21). We would love to see you there – please get in touch or visit our Events page to see all the meetings we will be attending. How much content can AI legally exploit? O verview During the recent PubsTech conference , we were asked how much content could be legitimately used to train artificial intelligence systems without being specifically secured through a licensing agreement. In considering this question, we find some counterintuitive results. Background Generative AI (genAI) is a type of artificial intelligence that can create new content—text, images, music, and more – by analyzing patterns in massive datasets. These models are typically trained on publicly available data scraped from the web. In the US, developers often invoke the “Fair Use” copyright doctrine to justify this training, claiming it is limited to specific purposes (training) and transformative in nature (different from the original use). In reality, the legal position is complex and evolving , with many rights holders and their representatives – unsurprisingly – taking the opposite view. Even if legal clarity emerges, different geographies and jurisdictions will likely reach different conclusions. The legal complexities of AI and copyright law are beyond our scope. However, for scholarly publishers, particular issues apply. Half of our output is open access , and open access content is designed to be reusable. Open or not, content has varying restrictions on onward use – for example, non-commercial use is often allowed with attribution. How much scholarly content is exploitable?  For the purposes of analysis, we will assume that the license under which content is published will have a material bearing on the legitimacy of its use to train AI systems. Therefore, looking at share of licenses, we might be able to answer our question.
More Posts