10 tips when unlocking AI for R&D
30. November 2023
Von Ann-Marie Roche

© istock.com/metamorworks
The first edition of our new webinar series explores the perils and pitfalls of generative AI for R&D. Four experts talk about how to take advantage of new technology without losing sight of the big picture.
Elsevierâs four-part webinar series AI in innovation: Unlocking R&D with data-driven AI outlines the issues that can derail your AI projects and how to prepare yourself for these innovations â and those around the corner. In the first edition, a panel of AI and data experts explore the perils, pitfalls and promise of generative AI for R&D Wird in neuem Tab/Fenster geöffnet.
Moderated by Elsevierâs Commercial Director for Corporate Markets Zen Jelenje Wird in neuem Tab/Fenster geöffnet, the panel consisted of Elsevierâs VP of Data Science Life Sciences Mark Sheehan Wird in neuem Tab/Fenster geöffnet and two experts from Elsevierâs SciBite: Director of Data Science & Professional Services Dr Joe Mullen Wird in neuem Tab/Fenster geöffnet and Head of Ontologies Dr Jane Lomax Wird in neuem Tab/Fenster geöffnet.Â
With Elsevierâs history of providing enriched and curated scientific data in AI-driven solutions such as Reaxys and Embase, this episode focuses on the questions our scientists, data scientists and computational chemists get from customers about AI and, more recently, large language models (LLMs).
As Zen explains: âThese arenât simple questions, and we definitely donât have all the answers. But today, we have a diverse team from Elsevier and SciBite to explore some of these topics.â

Zen Jelenje
Do watch the whole episode Wird in neuem Tab/Fenster geöffnet â a lot was covered. Meanwhile, here are some tips from the panel for navigating these changing times.
Tip #1: Get your data in order.
Itâs easy to get distracted by all the noise and hype around LLMs, particularly ChapGPT. But to take advantage of any AI technology, you need to start with your data.
âYour data need to be well organized, well-structured and FAIR Wird in neuem Tab/Fenster geöffnet â meeting the principles of Findability, Accessibility, Interoperability and Reusability,â Joe says. âOnly then will you be ready and flexible enough to quickly and seamlessly latch onto the best solution for the problem you want to solve.â (see Tip #2).

Joe Mullen, PhD
Tip #2: Donât rush to a âsolution.â Start by asking, âWhatâs the specific problem I want to solve?â
âYou've got to remain focused on identifying what the problems are, and only then look at the ever-evolving solutions to solve those problems,â Joe says.Â
âInstead of thinking of it as whether to invest in AI,â adds Zen, "you need to ask the question, âHow does this improve my research?ââ
Tip #3: Donât consider LLMs as an all-in solution â especially for life sciences. (However, LLMs can still be part of the solution.)
At the end of the day, scientific progress is built on providence, transparency and reproducibility. And LLMs like ChatGPT are simply not built for that â for now anyway. Currently, much of Elsevierâs work is built on ontologies. âThese use language to create a model of a domain,â Jane explains. âIt's a codification of what humans understand about a particular domain â facts as we now understand them. And I think that's always going to be something that's necessary and useful.
âLLMs, on the other hand, are probabilistic models that are really powerful at generating and understanding human language,â Jane adds. âTheyâre amazing, and we use them internally.â But unfortunately, LLMs also hallucinate, and the information is not properly sourced. So in the longer-term, many hope âto have an LLM with an ontology-based factual backbone â and then youâll have something truly powerful,â she says.
âI also think that LLMs can bring value to one of our main aims at SciBite,â says Joe. âAnd thatâs supporting data democratization â improving the access and interpretation of data. But LLMs wonât be able to supply this by themselves due to their limitations.âÂ

Jane Lomax, PhD
Tip #4: Donât underestimate scaling.
âOne piece of advice: donât underestimate the difficulty in being able to scale these types of technologies to production,â says Jane. âWhen we started with this three years ago, we ended up having to take a step back and first build the infrastructure and invest in the skills. We learned a lot through that process, but it was quite a learning curve. So, if you're investing in this, donât overlook this. Come chat with us.âÂ
Tip #5: Think operationalization.
âNew technology brings new holistic cost considerations,â Joe says. âThere are costs associated with rolling out some of these larger models: monetary costs, time costs, disk and carbon footprint costs, and so on and so forth.âÂ
Tip #6: Get your hands dirty (while failing fast, learning fast and moving on).
âI read a McKinsey report the other day about whether you want to be a taker, a shaper or a maker in the AI space,â Mark says. âAre you going to wait until itâs fully cooked? Nothing wrong with that. And it can depend on the industry or your companyâs appetite for risk and investment.â But for Elsevier, the road was clear: jump in now.Â
âAnd definitely having the right team in place is important,â adds Zen. âAnd since some of the questions we try to address in scientific research are really specific to the domain, it's also harder to wait for somebody else to do it for us.â Â
âItâs actually very fulfilling to bring the team together to work on new innovations using the latest technologies,â Mark says. âBut itâs important to acknowledge there will be bumps on the road on that digital transformation journey. There will be mistakes and there will be failures. But itâs also incredibly rewarding when you get it right. You need to learn from your mistakes, pick yourself up and move forward.â
âAnd this is what weâve been doing for the last 12 to 18 months in terms of GenAI, specifically LLMs. Weâre getting our subject matter experts, our data scientists and our data analysts together to really get their hands dirty and ask, âWhat can I do now that I couldn't do yesterday?â It's like you're building your muscles up in this space. Youâre learning as you go.âÂ

Mark Sheehan
Tip #7: Think modularity.
âOur enrichment pipelines continue to become more automated and feature more of the latest AI technologies as we iterate,â says Mark. âAnd certainly, it's not the case that as soon as a new technology comes in, we throw out what we had before. It works well that we have a mix of rule-based technologies and machine learning technologies. And now we're exploring the latest Gen AI technologies. These can all be complementary.â
âWe always try to find a way to integrate all these different pipelines, datasets and capabilities into what my team calls a Lego set,â Mark adds. âIt's a great way to approach things in a modular and flexible way without getting too obsessed about the latest or greatest technologies.â
Tip #8: Stay on top of whatâs happening.
It might be simpler to wait for others to fail and then adopt. But here you risk being left behind â and losing any competitive edge. As Joe points out: âAround 10 years ago, AI was beating humans at Space Invaders Wird in neuem Tab/Fenster geöffnet. Around five years ago, AI got better at Go Wird in neuem Tab/Fenster geöffnet. Just a few weeks ago, AI started beating humans in real-time drone racing Wird in neuem Tab/Fenster geöffnet. AI is evolving at such a pace, you need to keep yourself skilled up and aware of what's going on around you.
âAnd again, this is about getting your hands dirty. Reading a few articles and blogs isnât enough. But itâs a difficult balance: keeping on top of things without getting sucked in, while just trying to identify those problems you want to solve.â
Tip #9: Keep humans in the loop.Â
As the panel discussed, Subject Matter Experts (SMEs) remain essential to validate the output of any AI algorithm â and more so when it comes to LLMs. For instance, these SMEs can be deployed as prompt engineers to ask the right questions to the LLMs so the resulting output is easier to validate. Â
âPrompt engineering is actually a skill that we should all have some appreciation and understanding of,â Joe says. âIt's not as straightforward as some people might expect. You need to be able to relay your understanding of the world to an LLM ⊠and this comes back again to the real importance of SMEs when applying it to scientific domains that really require some expertise.â
Tip #10: While waiting on regulatory decisions, aim to be responsible.
âIf you ask me about the regulatory environment today, this webinar would be out of date in a month or so,â Mark says. And indeed: watch this space. But meanwhile, you should aim to be responsible. âRegulations are all about governments coming in saying we need to manage this space because weâre concerned about the future. But it could start with responsible AI where the actual practitioners go âHow can we be responsible and ethical about how we approach this?â And at Elsevier, weâve really tried to bake this into our daily work from the start with our Responsible AI principles Wird in neuem Tab/Fenster geöffnet.â
For the full iceberg of insight, watch the webinar. And in the meantime, donât forget to get your data in order (see tip #1)!
Mitwirkende*r

AR
Ann-Marie Roche
Senior Director of Customer Engagement Marketing
Elsevier
Weiterlesen ĂŒber Ann-Marie Roche