Wednesday, 21 June 2023

How to Use ChatGPT to Write Scientific Research Paper?

 Source: https://drasmajabeen.com/chatgpt-for-scientific-research-paper-writing/

How to Use ChatGPT for Scientific Research Paper writing?


How to Use ChatGPT to Write Scientific Research Paper?

ChatGPT is an AI language model that generates text using input provided by the user. ChatGPT can be used as a tool to assist in the writing process for scientific research paper. Writing a scientific research paper requires not only knowledge of the subject but also skills like critical thinking, problem-solving, analysis, and data interpretation. Therefore, it is essential to use ChatGPT in combination with your own expertise, knowledge and skills.

In this article lets discuss 9 important steps to use ChatGPT for Scientific research paper writing.

1.  Compile Research Material

Before using ChatGPT collect and organize all the research materials you need for your scientific research paper. This includes articles, books, journals, and any other sources that you plan to use in your research.

2. Brainstorming session with ChatGPT

Brainstorm the research topic by using different variations of the research topic in ChatGPT and compile finalized version accordingly.

3. Define your research question

Move to the second step by identifying the research question or research hypothesis to be addressed in research paper.

4. Conduct a literature review

Further, Use ChatGPT to search for relevant scientific literature relevant to your research topic. Highlight and specify literature relevant to the research question and research hypothesis.

5. Summarize Important Research Articles

Get your highlighted research articles summarized by ChatGPT.

6. Gaps in the literature

Try to brainstom ChatGPT for identification of research gaps in the literature.

7. Analyze data

Analyze data Collected through surveys/other means Then use ChatGPT to help analyze and interpret your data, as well as generate visualizations to support your findings.

8. Create an outline of the Research paper

Using ChatGPT organize your research paper by creating an outline and structuring your arguments logically. Use ChatGPT to generate sections of your paper, such as the introduction, methods, results, and discussion sections by using subject-specific commands and data inputs for each section.  However, ensure to strongly review content generated by ChatGPT.

9. Edit and proofread your paper

Use ChatGPT for editing and proofreading your paper for grammar, punctuation, and spelling errors. But don’t forget to carefully cross-check and review your paper and make all necessary revisions of ChatGPT generated content to ensure flow accuracy and clarity and authenticity of research.

So in this way by following above mentioned 9 steps you can write your research paper using ChatGPT.

4 Facts about ChatGPT in scientific research paper writing

  1. ChatGpt cannot write a scientific research paper entirely on its own. Scientific knowledge human expertise, critical thinking and analytics are essential for research paper writing. ChatGPT can be used as an assistant in generating content for a scientific research paper.
  2. ChatGPT is a tool that can assist in research but it cannot replace human researchers.
  3. ChatGPT can be used in analysis and data interpretation but still human input is required for data analysis and accuracy.
  4. ChatGPT-generated content may contain errors, so it is important to review content generated by ChatGPT for accuracy.

Sunday, 18 June 2023

Research Tools for Literature Search, Paper Writing, and Journal Selection

 Source: https://doi.org/10.6084/m9.figshare.23540007.v1

📚🖋️ Excited to share valuable insights and resources. #ResearchTools #AcademicWriting #LiteratureSearch Get it here: https://doi.org/10.6084/m9.figshare.23540007.v1 

Sunday, 4 June 2023

How To Find Best Journal For My Article

 Source: https://phdnotpad.com/2023/01/08/how-to-find-best-journal-for-my-article/

How To Find Best Journal For My Article

Unpublished research is essentially a waste of time and effort in academia. Unless it is published in an approved and peer-reviewed publication, scientific research will have no significance.

Publication in scholarly journals, historically founded on the concept of peer review, is the foundation of scientific research. Therefore, you must choose the appropriate scientific journal to ensure that your scientific work reaches the target scientific community.

Unpublished research is essentially a waste of time and effort in academia. Unless it is published in an approved and peer-reviewed publication, scientific research will have no significance.

Publication in scholarly journals, historically founded on the concept of peer review, is the foundation of scientific research. Therefore, you must choose the appropriate scientific journal to ensure that your scientific work reaches the target scientific community.

How to find best scientific journals?

and understand its categorization by using the steps I’ve provided below.

  1. Be aware of the strengths of your research in relation to publications in your field, including its originality, relevance, and volume. It makes no sense to go to the most important publication in your field with a little study. It is unfair to publish an important and distinguished paper in a journal with a low impact factor. To determine the intensity of the research, choose from your area of expertise or consult colleagues: B-medium, C-low, and A-high. Later, we will return to this classification.
  1. Choose at least five publications relevant to your field of study and research based on their titles only. Here, we will access the Journal Citation Reports section of the ISI Web of Science database (you need to log in with your university or Saudi Digital Library account). To access and search for journal rankings, as well as to see the impact factor and ranking of each journal, see the section below.
  2. If you decide that your research has high power, choose journals that are ranked in your field of study in either the first or second quarter (Quarter I or Quarter II); However, if you decide your research is average, choose journals that are ranked in the second or third quarter (Q2 or Q3). And if you determine your search is low, choose a journal from Q3 or Q4.
  3. Start researching topics related to your studies in each journal now, to ensure that your paper will be accepted by the publication’s editors as well as work related to your main area of interest. Journals that are not relevant to your field of study should be excluded.
  4. You should only have one, two, or three magazines at this point, but you don’t know which one to choose! To make your final decision, you can now add the following criteria:

what is cretiria to find the best journal?

what is cretiria to find the best journal?
what is cretiria to find the best journal?

A: If you want to publish your work as soon as possible, pick the journal with the fastest turnaround time for examining your study and responding to you. A busy magazine with a renowned editorial team is one that publishes its issues in between three weeks and a month. A journal is seen as poor if it takes more than two months. Don’t even consider sending her your study, which takes more than six months. The average magazine speed can be determined in a number of methods.

Elsevier Insights’ Speed field will tell you the journal’s speed if it is published by an Elsevier publisher. (In this industry, several periodicals do not disclose their figures.)

Open the journal’s publisher page if Springer is the publisher, then click on Journal Metrics.
The author of this piece has compiled meta data from tens of thousands of research publications; find your journal there.
the website of the first magazine.

B. From the remaining magazines you were unsure of, pick the one with the most influence.

C. Opt for the journal that publishes the most research each year since those journals have a higher acceptance rate than the others. Databases from Scopus or ISI Web of Knowledge can be used to confirm the figure.

D. If you don’t have a research funding to finance it, omit the journal with the hefty publication fees.

  • Final advice.
  • Don’t send your manuscript to a journal without an impact factor or that isn’t accessible through Scopus or Web of Knowledge. Such a publication is useless.
  • Squeeze if the magazine’s website appears to be of poor quality.
  • Check their nature by reading the Author rules or the Author’s Guide on the journal’s website.
  • Avoid submitting your work to more than one journal at once since doing so is unethical.
  • Remember that it is quite uncommon for your study to receive final approval in its first form. Arbitrators frequently demand a number of changes before approving it.
  • Predatory periodicals, which are frequently open access, should be avoided.
  • Consider it a danger indication rather than a benefit if publishing in a journal appears too simple or the content looks too “squeezed.”
  • Of course, there are exceptions to the aforementioned search strategy.
  • For instance, you go immediately to a research article that is appropriate for a publication with a small readership.
  • Every sober study with a rigorous approach deserves a beautiful conclusion in the form of a scientific article.
  • As a result, give your research project careful consideration before submitting it for publication.
  • One of the main causes of a submission being rejected or being published in the incorrect location is the poor selection of a certain magazine.
  • I do not recommend the use of automatic journal suggestion tools. At least I didn’t find it useful personally. But if you like to try it, this is the most popular one:
  1. Springer Journal Suggester
  2. Elsevier Journal Finder
  3. Wiley Journal finder
  4. IEEE publication recommender 
  5. Journal Guide
  6. Co-factor journal selector

Friday, 2 June 2023

Is AI-generated content actually detectable?

 Source: 

Is AI-generated content actually detectable?

AI

Credit: Pixabay/CC0 Public Domain

In recent years, artificial intelligence (AI) has made tremendous strides thanks to advances in machine learning and growing pools of data to learn from. Large language models (LLMs) and their derivatives, such as OpenAI's ChatGPT and Google's BERT, can now generate material that is increasingly similar to content created by humans. As a result, LLMs have become popular tools for creating high-quality, relevant and coherent text for a range of purposes, from composing social media posts to drafting academic papers.

Despite the wide variety of potential applications, LLMs face increasing scrutiny. Critics, especially educators and original content creators, view LLMs as a means for plagiarism, cheating, deception and manipulative social engineering.

In response to these concerns, researchers have developed novel methods to help distinguish between human-made content and machine-generated texts. The hope is that the ability to identify automated content will limit LLM abuse and its consequences.

But University of Maryland computer scientists are working to answer an important question: can these detectors accurately identify AI-generated content?

The short answer: No—at least, not now

"Current detectors of AI aren't reliable in practical scenarios," said Soheil Feizi, an assistant professor of computer science at UMD. "There are a lot of shortcomings that limit how effective they are at detecting. For example, we can use a paraphraser and the accuracy of even the best detector we have drops from 100% to the randomness of a coin flip. If we simply paraphrase something that was generated by an LLM, we can often outwit a range of detecting techniques."

In a recent paper, Feizi described two types of errors that impact an AI text detector's reliability: type I (when human text is detected as AI-generated) and type II (when AI-generated text is simply not detected).

"Using a paraphraser, which is now a fairly common tool available online, can cause the second type of error," explained Feizi, who also holds a joint appointment in the University of Maryland Institute for Advanced Computer Studies. "There was also a recent example of the first type of error that went viral. Someone used AI detection software on the U.S. Constitution and it was flagged as AI-generated, which is obviously very wrong."

According to Feizi, such mistakes made by AI detectors can be extremely damaging and often impossible to dispute when authorities like educators and publishers accuse students and other content creators of using AI. When and if such accusations are proven false, the companies and individuals responsible for developing the faulty AI detectors could also suffer reputational loss.

In addition, even LLMs protected by watermarking schemes remain vulnerable against spoofing attacks where adversarial humans can infer hidden watermarks and add them to non-AI text so that it's detected to be AI-generated. Reputations and may be irreversibly tainted after faulty results—a major reason why Feizi calls for caution when it comes to relying solely on AI detectors to authenticate human-created content.

"Let's say you're given a random sentence," Feizi said. "Theoretically, you can never reliably say that this sentence was written by a human or some kind of AI because the distribution between the two types of content is so close to each other. It's especially true when you think about how sophisticated LLMs and LLM-attackers like paraphrasers or spoofing are becoming."

"The line between what's considered human and artificial becomes even thinner because of all these variables," he added. "There is an upper bound on our detectors that fundamentally limits them, so it's very unlikely that we'll be able to develop detectors that will reliably identify AI-generated content."

Another view: More data could lead to better detection

UMD Assistant Professor of Computer Science Furong Huang has a more optimistic outlook on the future of AI detection.

Although she agrees with her colleague Feizi that current detectors are imperfect, Huang believes that it is possible to point out artificially generated content—as long as there are enough examples of what constitutes human-created content available. In other words, when it comes to AI analysis, more is better.

"LLMs are trained on massive amounts of text. The more information we feed to them, the better and more human-like their outputs," explained Huang, who also holds a joint appointment in the University of Maryland Institute for Advanced Computer Studies. "If we do the same with detectors—that is, provide them more samples to learn from—then the detectors will also grow more sophisticated. They'll be better at spotting AI-generated text."

Huang's recent paper on this topic examined the possibility of designing superior AI detectors, as well as determining how much data would be required to improve its detection capabilities.

"Mathematically speaking, we'll always be able to collect more data and samples for detectors to learn from," said UMD computer science Ph.D. student Souradip Chakraborty, who is a co-author of the paper. "For example, there are numerous bots on social media platforms like Twitter. If we collect more bots and the data they have, we'll be better at discerning what's spam and what's human text on the platform."

Huang's team suggests that detectors should take a more holistic approach and look at bigger samples to try to identify this AI-generated "spam."

"Instead of focusing on a single phrase or sentence for detection, we suggest using entire paragraphs or documents," added Amrit Singh Bedi, a research scientist at the Maryland Robotics Center who is also a co-author of Huang's paper. "Multiple sentence analysis would increase accuracy in AI detection because there is more for the system to learn from than just an individual sentence."

Huang's group also believes that the innate diversity within the human population makes it difficult for LLMs to create content that mimics human-produced text. Distinctly human characteristics such as certain grammatical patterns and word choices could help identify text that was written by a person rather than a machine.

"It'll be like a constant arms race between generative AI and detectors," Huang said. "But we hope that this dynamic relationship actually improves how we approach creating both the generative LLMs and their detectors in the first place."

What's next for AI and AI detection

Although Feizi and Huang have differing opinions on the future of LLM detection, they do share several important conclusions that they hope the public will consider moving forward.

"One thing's for sure—banning LLMs and apps like ChatGPT is not the answer," Feizi said. "We have to accept that these tools now exist and that they're here to stay. There's so much potential in them for fields like education, for example, and we should properly integrate these tools into systems where they can do good."

Feizi suggests in his research that security methods used to counter generative LLMs, including detectors, don't need to be 100% foolproof—they just need to be more difficult for attackers to break, starting with closing the loopholes that researchers already know about. Huang agrees.

"We can't just give up if the detector makes one mistake in one instance," Huang said. "There has to be an active effort to protect the public from the consequences of LLM abuse, particularly members of our society who identify as minorities and are already encountering social biases in their lives."

Both researchers also believe that multimodality (the use of text in conjunction with images, videos and other forms of media) will also be key to improved AI detection in the future. Feizi cites the use of secondary verification tools already in practice, such as authenticating phone numbers linked to social media accounts or observing behavioral patterns in content submissions, as additional safeguards to prevent false AI detection and bias.

"We want to encourage open and honest discussion about ethical and trustworthy applications of generative LLMs," Feizi said. "There are so many ways we can use these AI tools to improve our society, especially for student learning or preventing the spread of misinformation."

As AI-generated texts become more pervasive, researchers like Feizi and Huang recognize that it's important to develop more proactive stances in how the public approaches LLMs and similar forms of AI.

"We have to start from the top," Huang said. "Stakeholders need to start having a discussion about these LLMs and talk to policymakers about setting ground rules through regulation. There needs to be oversight on how LLMs progress while researchers like us develop better , watermarks or other approaches to handling AI abuse."

Both papers are published on the arXiv preprint server. 

More information: Vinu Sankar Sadasivan et al, Can AI-Generated Text be Reliably Detected?, arXiv (2023). DOI: 10.48550/arxiv.2303.11156

Souradip Chakraborty et al, On the Possibilities of AI-Generated Text Detection, arXiv (2023). DOI: 10.48550/arxiv.2304.04736

Journal information: arXiv


Explore further

Why GPT detectors aren't a solution to the AI cheating problem

 

Thursday, 1 June 2023

Prompt engineering - Something for librarians here?

 Source: http://musingsaboutlibrarianship.blogspot.com/2023/06/prompt-engineering-something-for.html


GPT4 defines prompt engineering as 

the process of creating, designing, and refining prompts or questions that are used to generate responses or guide the behavior of an AI model, such as a chatbot or a natural language processing system. This involves crafting the prompts in a way that effectively communicates the desired information, while also considering the AI's capabilities and limitations to produce accurate, coherent, and relevant responses. The goal of prompt engineering is to improve the performance and user experience of an AI system by optimizing the way it receives instructions and delivers its output.

Or as Andrej Karpathy a Computer Scientist at OpenAI puts it 

The hottest new programming language is English.

If you are a librarian reading this, it is likely you have wondered, isn't this a little like what librarians do? When a user approaches us at a reference desk, we are trained to do reference interviews to probe for what our users really want to know. In a similar vein, evidence synthesis librarians help assist with Systematic reviews in developing protocols, which include problem formulation as well as crafting of search strategies. Lastly from the information literacy front, we teach users how to search.

In other words, Are librarians the next prompt engineers?

As I write this, there is even a published article entitled - "The CLEAR path: A framework for enhancing information literacy through prompt engineering", though this doesn't really teach one how to do prompt engineering as typically defined.

My first thought on prompt engineering was skepticism.  

This post will cover some of the following

  • While there is some evidence that better prompts can elicit better outputs, is there truly a "science of prompting" or is it mostly snake oil?
  • If there is something that is useful to teach, is it something there that librarians are potentially qualified and capable to teach without doing a lot of upskilling? Or is it something that is out of reach of the typical librarian?
At points, I try to draw parallels with the now well-established task in teaching how to search and see how well they hold up. 

To anticipate the conclusion, my current view is that on the whole we probably might want to be involved here, particularly if nobody else steps up. This is despite us being hazy on how effective prompt engineering can be.
 

Is there a science of prompting?

As librarians, we are always eager to jump in and help our users in every way we can. But I think we should be careful not to fall for hype and jump on bandwagons at the drop of a hat. The reputation of the library is at stake, and we do want librarians to teach something that was mostly ineffective.

I admit my initial impression of "prompt engineering" was negative because it seemed to me there was too much hype around it. People were going around sharing super long complicated prompts that were supposedly guaranteed to give you magical results by just editing one part of the prompt. 

As Hashimoto notes a lot of this is what he calls "blind prompting" 

"Blind Prompting" is a term I am using to describe the method of creating prompts with a crude trial-and-error approach paired with minimal or no testing and a very surface level knowledge of prompting. Blind prompting is not prompt engineering.

These types of "blind prompts" often feel to me more like magic incarnations, that someone found to work once, and you just copy them blindly without understanding why they work (if they do at all). Given how much of a black box neutral nets (which transformer-based language models are), how likely is it that we are sure a certain crafted prompt works better when we don't even understand why it might?

Another reason to be skeptical of the power of such long magical prompts is from Ethan Mollick, a professor at Wharton who has been at the forefront of using ChatGPT for teaching in class. 

In an interesting experiment, he found that students who adopted a strategy of going back and forth with ChatGPT in a coediting or iterative manner got far better results when trying to write an essay, than those who adopted simple prompts or those who did long complicated prompts at one go. 

This makes sense and parallels experiences in teaching searching. In general, slowly building up your search query iteratively will usually beat putting all the search keywords in one go particularly if this is an area you do not know.

This isn't evidence against the utility of prompt engineering, just a caution about believing in long magical prompts without iterative testing or strong evidence.


Looking at Prompt engineering courses and guides

To give prompt engineering a fair chance, I looked at some of the prompt guides listed by OpenAI themselves at OpenAI cookbook.

These guides included

I also looked through courses on Udemy and Linkedin Learning. The courses on prompt engineering for the former looked more substantial, so I looked at

Here's what I found. To set the context, I have immediate level knowledge on how Large Language Models work and have been reading research papers on them (with varying amounts of comprehension) for over a year, as such I didn't really learn much in these content that I didn't already know. (This isn't necessarily a terrible thing!)

The good thing is all the prompt guides and the courses I looked at covered remarkably similar ground. This is what you expect if you have a solid body of theory. So far so good.

Most of them would cover things like


Reading some of these prompt guides gave me remarkably similar vibes as guides for doing evidence synthesis for example - Cochrane Handbook for Systematic Reviews of Interventions where the best practice for prompting is often followed by a academic citation (see below for an example).



A lot of the advice given for prompt engineering tends to be common sense, be specific, give it context, specify how you want the output to be etc. Unless the language model is a mind reader it won't know what you want!

But is there all to prompt engineering? Are there specific prompting techniques that are particular to the use of large language models like ChatGPT and have evidence behind them to suggest they improve performance? Ideally, we should even be able to understand why they work based on our understanding of the transformer architecture. 

To take an analogy from teaching Boolean searching, we teach users to drop stop words when constructing Boolean search strategies based on our understanding of how traditional search engines work (ironically, modern search engines might not even be optimised for strict Boolean because of how rarely it is used and might be counterproductive for modern search engines that use semantic search).

Similarly, we know from decades of practice and research, a strategy of

a. Breaking a search intent into key concepts
b. Combining synonyms of each concept with an OR operator
c. Chaining them all together with a AND operator

is likely to increase recall, particularly if further covered with citation searching.  

Do we have something similar for prompt engineering?

From what I can tell, exploiting in-context learning seems to be one commonly used technique.

The basic technique - In-context learning

The paper introducing GPT-3 in 2020 (Brown 2020) was entitled - Language Models are Few-Shot Learners and was a revelation. Before that, when you wanted to improve the performance of a neural net model, you had to fine-tune it with additional specialized data. This was of course slow and not easy if you were not comfortable with deep learning.
As of time of writing, you can fine-tune GPT-3 base models but not the newer ChatGPT-3.5 turbo and GPT 4 models



However, GPT-3 showed that when the models got big enough, one could improve results for some tasks by typing examples in the prompt in plain text and it would "learn" and do better, compared to not giving examples. This is known as in-context learning.  

As shown in the example above, for a translation task you can type in one example in the prompt (one shot) or many examples (many shot). Of course, one could give no examples, and this would be known as zero shot prompting.

This is what makes prompt engineering possible, since you just type in prompts, there is no machine learning training at all! Of course, unlike fine tuning which changes the weights of the models leading to a permanent alteration, prompt engineering with examples will only be temporarily "learnt" and the change will not persist once the session ends.

Best way to give examples in prompts

Besides the observation that few-show learning works, are there any best practices in the way we craft the prompts when giving examples?

In other words, what is the best way to give these examples in the prompt to get the best results ? There is a surprising amount of guidance. for example on how to provide examples, what examples to give and the order they are given. 


keep the selection of examples diverse, relevant to the test sample and in random order to avoid majority label bias and recency bias.

This is to say if say you are doing a sentiment analysis task on comments and you are doing few-shots prompting, you might not just want to give examples that cover all three categories of values say "Positive", "Negative", "Neutral" but also the numbers of the examples reflect the expected distribution. E.g. If you expect more "positive" comments, your examples in the few-shot learning should reflect that.

The example should be also given in random order, according to another paper.


More advanced prompt technique - Chain of thought Prompting

Another famous prompt engineering technique most people would have heard of is Chain-of-Thought Prompting.


The chain of thought prompt is based on the observation that if you provided examples of how to reason in the few shot prompts the Language Model would do better at many reasoning tasks as opposed to giving the answers directly in the examples.






In fact, it was later found you could even skip the example and just prompt it to "Let's think step bv step" and the results would improve!

I call these tips observations, but they are in fact results from papers (typically preprints on Arxiv).

The reasoning why this works is roughly, if you guide the LLM to try to reason its way to an answer it might do better than if it tried to get to the answer immediately. I've seen it explained as "you need to give the LLMs more tokens to reason". 

This makes more sense if you understand that the popular GPT type language models are auto-regressive decoder only models and they generate tokens one by one and unlike a human they cannot "go backwards" once they have generated a token. The prompt "Let's think step by step" is meant to "encourage" it to try solving the solutions in small steps rather than try to jump to the answer in one step.

This is also why changing the order by prompting the model to answer the question and then give reasons why is unlikely to help.

For example, the following prompt is bad or at least will not get you the advantage of Chain of Thought

You are a financial expert with stock recommendation experience. Answer “YES” if good news, “NO” if bad news, or “UNKNOWN” if uncertain in the first line. Then elaborate with one short and concise sentence on the next line. Is this headline good or bad for the stock price of company name in the term term?


Skeptical of the science of prompts?

While it is good that these prompt guides are mostly quoting advice from research papers, as an academic librarian, you are of course aware that any individual finding from a paper even a peer reviewed one should be treated with caution (never mind the ones being quoted in this field are often just preprints). 

While I have no doubt the general practice of COT and few-shot learning works (there is far too much follow up and reproduced research), many of the specific techniques (e.g. the way you specify examples) might be on shaky ground given they tend to have much less support for their findings.

Take for example, the advice you should do role prompting, e.g. "You are a brilliant Lawyer". I have even seen people try "You have a IQ of 1000" or other similar prompts. Do they really lead to better results? In some sense, such prompts do give the model more context, but it is hard to believe telling it to "Act like a super genius" will really make it super brilliant. :)

To make things worse, while we can understand why things like "Chain of Thought" and variant works, a lot of the advice quoted from papers have findings that as far as I can tell are purely empirical and we do not understand why they work. 

For example, as far as I can tell, tips for Example Selection and Example Ordering are just empirical results found, knowledge of how LLM works makes you none the wiser.

This is owing to the black box nature of Transformers which are neural nets.  



Do the findings even apply for the latest models?

Another problem with such prompt engineering citations is they often are tested on older models than what is state of art because of a time lag. For example, some papers cited are using ChatGPT (GPT3.5) or even GPT3.0 models rather than GPT4 which is the latest at the time of writing of this post.

As such are we sure the findings generalize across LLMs as they improve? 

For sure, a lot of the motivation for these invented special prompt techniques was to solve hard problems that earlier models like GPT-3, or ChatGPT-3.5 could not solve without prompt engineering. Many of these problems, work in GPT-4 without special prompts. It might be also some of the suggested prompt engineering techniques might even hurt GPT-4.

For example, there is some suggestions that role prompting no longer works for new models.



We also assume everyone is using OpenAI's GPT family for both researching of prompt engineering and for use and there is a greater likelihood the findings are stable even for newer versions. In fact, people might be using Opensource LLMA models or even "Retriever augmented Language Models" like Bing Chat or ChatGPT with browsing plugin.

Given how most of the advice you see is based on research assuming you are using ChatGPT without search, this can be problematic if one started to shift to say using Bing Chat, or even ChatGPT with browsing plugin. Take the simplest question, does COT prompts help with Bing Chat? I am not sure there is research on this.

A rebuttal to this is that when you look at evidence synthesis guides like Cochrane Handbook for Systematic Reviews of Interventions it cites even older evidence and search engines have also changed a lot compared to say 10 years ago. (Again, think how some modern search engines now are often not optimized for strict Boolean or even use Semantic Search and do not ignore stop words)

One can rebut that rebuttal and say even if search engines change, they are a mature class of product and unlikely to change as much as Language Models which are new technology.

In conclusion, I don't think we have enough evidence at this point to say how much prompt engineering helps, though I think most people would say in many situations it can help get better quality answers. The fact neural net based transformers lack explain-ability, making it even harder to be sure.

If there is a "science of prompting", it is very new and developing, though it is still possible to teach prompt engineering if we ensure we send out the right message and not overhype prompt engineering.


How hard is it for librarians to teach prompt engineering?

Understanding the limitations of the current technology - Transformer based, Decoder only language model will help with prompt engineering. I already gave an example earlier on why understanding the autoregressive nature of GPT models helps you understand why Chain of Thought prompting works.

But how deep does this knowledge go to effectively do prompt engineering? If it is too deep, librarians without the necessary background would struggle.

Do you need to learn the ins and outs of the transformer architecture, master and understand concepts /modules like Masked Multi-Head Attention, positional encoding or even understand the actual code?

Or do you just need a somewhat abstract level understanding of embeddings, idea of language models, and how GPT models are pretrained/self-supervised learning with large amount of text on the prediction next token task. 

You might even understand how GPT3.5 models are further aligned using RLHF (Reinforcement learning with human feedback), the difference between encoder only models (e.g. BERT), decoder only (e.g. GPT3) and encoder-decoder models (e.g. T5).

Or perhaps, you don't even need to know all that. 

Most prompt engineering guides and courses try to give the learner a very high-level view of how LLMs work. Compared to what I have been studying on LLMs, this level of understanding they try to impart is quite simple and I wager most librarians who are willing to study this should be capable of reaching the necessary level of expertise.

Why don't you need to know very deep knowledge of LLMs to do prompt engineering? Part of it is the same reason you can teach effective Boolean searching without really needing to know the details of how search engines work. How many librarians teaching searching really understand how Elastic search, inverted indexes etc. work? Many would not even remember how TF-IDF (or the more common BM25) works even if they were exposed to it once in Library School class years ago. I certainly didn't.

It also helps that LLMs are very much black boxes.

While in some cases general understanding of how decoder only Transformer based models works helps you understand why some techniques work (see above on general reasoning for COT prompting), some techniques and advice given - if there is a reason, looks opaque to me.

For example, as far as I can tell, tips for Example Selection and Example Ordering are just empirical results found, knowledge of how LLM works makes you none the wiser.


Advanced prompt engineering techniques need basic coding

That said when you look at the Prompt engineering techniques it splits into two types. The first type is what I already mentioned, simple prompts you can manually key into the prompt and use.

However advanced prompt found requires you to make many multiple prompts that are often unrealistic to do by hand. For example, Self-Consistency prompting requries you to prompt the system multiple times with the same prompt and take the majority answer.

Other even more advanced prompts like ReAct1(reason, act), are even more complicated, requiring you to type in multiple prompts including getting answers from external systems that will be too unwieldly to do manually.

https://learnprompting.org/docs/advanced_applications/react

Realistically these types of techniques involve the use of OpenAI's APIs in scripts and/or use of LLM frameworks like Langchain to automate and chain together multiple prompts and tools/plugins.

They are mostly quite easy to do if you have beginner level understanding of coding (often all you need is to run a Jupyter Notebook) but not all librarians are capable of that.


How much is there to teach? Will users benefit?

Assuming we keep to standard prompt engineering with prompts that are meant to be manually typed in the web interface, how much content is there really to teach? Is prompt engineering so intuitive and obvious there is no pointing teaching it?

My opinion is that one could cover the main things on prompt engineering fairly quickly say in a one-shot session. Could users learn prompt engineering themselves from reading webpages? Of course, they could! But librarians teach Boolean operators which could also be self-taught easily, so this isn't a reason not to teach.

I also think there might be an argument on why we should teach prompt engineering basically because it is so different from normal search.

While trying out formal prompt engineering, I found my instincts warring with myself when I was typing out the long prompts based on various elements that prompt engineering claims will give better results. 

Why is it so? I suspect it is ingrained in me when doing search queries to remove stop words and focus on the key concepts, so it feels odd to type out such long prompts. If it feels as unnatural to me as to most users who are also used to search engines, this again points to the need for librarians to educate.

Lastly, while there is currently hype about how Prompt engineering could give you 6 figure jobs  , Ethan Mollick disagrees and thinks there is no magic prompt and prompt engineering is going to be a temporary phase and will be less important as the technology improves.

He even quotes OpenAI staff.

I agree and one way to think about it is this. Will knowledge of Prompt engineering be more like knowing how to use email/Office which everyone knows (but with different degrees of skill) or will it be something like Python coding not known to most people at all.

I think it's the former and there's some utility for librarians to learn and teach prompt engineering in short workshops in the short term.

 

Conclusion

My current tentative conclusion is that there might indeed be something there in prompt engineering for librarians.

This is particularly so since many institutions are turning to librarians anyway with the idea that we are capable of guiding users here.

The main issue with librarians claiming expertise here is that this is a very new area, and I am personally not comfortable claiming I am an expert in prompt engineering and my users should listen to me. I believe if librarians are to do this, we should ensure users are aware this is as new to us as to them and we also learn together with them. 
 
I also worry we will be asked to show prompt engineering in domains we are not familiar with and hence be unable to easily evaluate the quality of output. 
 
0

Add a comment

 Source: https://filmdaily.co/news/how-to-do-seo-for-video-production/

TL;DR:

📽️ Want to boost your video production's online visibility? Check out this comprehensive guide on how to do SEO for video production! 🚀🎥 Learn the best practices and techniques to optimize your videos for search engines. Don't miss out on reaching your target audience effectively! #SEO #VideoProduction

Link: https://filmdaily.co/news/how-to-do-seo-for-video-production/

How to do SEO for Video Production

How to do SEO for Video Production

In today’s digital age, video has become a dominant force in content creation and consumption. It’s no secret that videos are captivating, engaging, and have the potential to reach a massive audience. However, without proper optimization, your video may go unnoticed in the vast sea of online content. That’s where search engine optimization (SEO) comes into play. By implementing effective SEO strategies, you can maximize the visibility and impact of your video productions. In this blog post, we will explore the essential steps to master SEO for video production.

Conduct Thorough Keyword Research

  1. Keywords are the foundation of SEO Company. Start by identifying relevant keywords and phrases that align with your video’s content and target audience. Use keyword research tools to discover popular and trending keywords in your industry. Look for long-tail keywords that have lower competition but high search volume. Incorporate these keywords naturally in your video title, description, tags, and transcript to enhance its visibility in search results.

Create Compelling Video Titles and Descriptions

  1. Your video’s title and description play a crucial role in attracting viewers and search engines. Craft a concise and compelling title that accurately reflects your video’s content and includes your primary keyword. Use descriptive and keyword-rich phrases in the video description to provide additional context and entice viewers to click. Remember to optimize your descriptions for both users and search engines by including relevant information and links to your website or social media channels.

Optimize Video Metadata and Thumbnails

  1. Metadata refers to the technical details associated with your video, such as file name, format, size, and duration. Ensure that your video’s metadata is optimized for SEO by including relevant keywords in the file name and format. Additionally, customize your video thumbnail to make it visually appealing and representative of your content. A compelling thumbnail can significantly increase the click-through rate and visibility of your video in search results.

Provide Accurate Transcripts and Closed Captions

  1. Transcripts and closed captions are not only beneficial for accessibility but also for SEO. Search engines can crawl and index the text within transcripts and captions, making your video more discoverable. Include accurate transcripts that capture the spoken content of your video. This not only improves SEO but also enables viewers to find specific information within your video more easily.

Optimize Video Loading Speed and Mobile Experience

  1. Page loading speed and mobile-friendliness are critical factors in SEO. Optimize your video’s loading speed by compressing the file size without compromising quality. Consider using a content delivery network (CDN) to ensure faster delivery of your videos. Additionally, make sure your video player and website are mobile-responsive, as an increasing number of users access content on mobile devices. A seamless mobile experience will positively impact user engagement and search rankings.

Leverage Video Schema Markup

  1. Schema markup provides additional context and structured data to search engines about your video content. Implement video schema markup on your website to enhance the visibility and appearance of your videos in search results. This markup can include details such as the video’s title, description, duration, thumbnail URL, upload date, and more. By leveraging schema markup, you can improve your video’s chances of appearing as a rich snippet in search engine results pages.

Promote and Share Your Videos

  1. SEO is not limited to on-page optimization. Actively promote and share your videos across various platforms, such as YouTube, social media, and your website. Encourage viewers to like, comment, and share your videos, as user engagement signals are considered by search engines when determining rankings. Embed your videos in relevant blog posts, email newsletters, and guest articles to expand their reach and attract more organic traffic.

In conclusion, mastering SEO for video production is vital to maximize the visibility and impact of your digital content.