cover of episode 2024 Recap: The Art of Prompt Engineering with Alex Banks, Founder and Educator, Sunday Signal

2024 Recap: The Art of Prompt Engineering with Alex Banks, Founder and Educator, Sunday Signal

2024/12/23
logo of podcast DataFramed

DataFramed

People
A
Alex Banks
Topics
Alex Banks: 提示工程是一项基础技能,每个人都应该学习,以更好地利用语言模型,提高工作效率。有效的提示取决于输入质量,清晰具体的提示能获得更精准的回应。编写有效提示的策略包括:清晰性、具体性以及开放式提问。可以使用角色-问题-解决方案框架来提高提示的有效性,并引导模型进行更深入的思考。一些提示技巧(如“深呼吸”或“我会付你100美元”)可以为模型提供更多思考空间,从而改善输出结果,但其背后的机制尚不明确。评估 ChatGPT 响应的有效性可以使用 LARF 框架:逻辑一致性、准确性、相关性 和事实正确性。可以通过提示工程技术来减少大型语言模型输出中的偏差,例如使用能够访问最新信息和可靠数据的工具。一些非显而易见的工具(例如 Scholar AI)可以提高提示质量,例如通过访问同行评审文章来确保事实的准确性。链式思维提示是一种高级技术,通过提供逐步推理的路线图来引导模型得出答案。ChatGPT 的输出质量最近有所下降,这可能是由于计算限制或其他因素造成的。可以通过改进提示工程技术来缓解这个问题。提示工程将成为一项重要的基础技能,而不是一个独立的职业。为 AI 应用程序编写提示与使用 ChatGPT 等工具编写提示之间存在差异,前者需要更长、更复杂的提示,并需要使用分隔符等技术。除了提示工程外,专业人士还需要掌握其他AI技能,例如制定AI战略和了解各种AI工具。下一代大型语言模型将朝着更通用化和个性化的方向发展,并在边缘计算领域发挥更大的作用。 Adel: 与Alex Banks讨论了提示工程的技巧、评估方法、以及在不同情境下的应用,并探讨了提示工程作为一项技能和职业发展路径的可能性。 Adel: 与Alex Banks就提示工程的有效性、评估方法、以及应对大型语言模型局限性(如幻觉和偏差)的策略进行了深入探讨,并对提示工程的未来发展趋势和对专业人士技能需求的影响进行了展望。

Deep Dive

Key Insights

Why is prompt engineering considered a foundational skill for the future?

Prompt engineering is essential because it allows users to effectively harness the power of language models, ensuring they get the best possible outputs. It helps individuals and organizations supercharge their productivity by leveraging AI tools more effectively.

What are some strategies for creating effective prompts in ChatGPT?

Effective prompts require clarity, specificity, and sometimes open-endedness. For example, providing context (e.g., 'I am an eighth-grade math teacher') and asking for creative examples can yield better results. Open-ended prompts allow the model to think outside the box and generate unexpected but valuable insights.

How can prompt engineering help reduce bias in AI outputs?

Bias in AI outputs often stems from the one-dimensional nature of the model's knowledge and the data it was trained on. Using tools like web browsing or referencing peer-reviewed papers can help fact-check and reduce hallucinations, ensuring more accurate and less biased responses.

What is chain-of-thought prompting, and why is it effective?

Chain-of-thought prompting involves providing a roadmap for the model to follow, guiding it step by step to arrive at the desired answer. This technique is particularly useful for complex tasks, as it allows the model to reason through problems more effectively, using examples and structured thinking to shape the output.

What are the key differences between prompt engineering for developers and general users?

Developers often write longer, more complex prompts that require detailed system-level instructions, using delimiters and structured formats. General users, on the other hand, focus on simpler prompts for tasks like writing or generating ideas. Both, however, rely on clarity, specificity, and examples to get the best results.

What are the must-have AI skills for professionals in the future?

AI literacy starts with understanding prompt engineering to maximize the potential of current AI tools. Beyond that, professionals should be proactive in identifying AI tools that can solve specific business problems, such as using generative AI for video creation or voiceovers. Being able to leverage AI to enhance productivity will be crucial.

What does the future hold for large language models like GPT-5?

The future of large language models is likely to trend toward artificial general intelligence, with models becoming more capable of handling complex tasks in users' voices. Additionally, there will be a proliferation of smaller, hyper-specific models tailored to individual or organizational needs, offering faster and more personalized solutions.

Chapters
This chapter introduces prompt engineering as a crucial skill in harnessing the power of language models. It explores the potential of AI and how prompt engineering can enhance productivity and transform daily life. Alex Banks' journey into AI and the creation of Sunday Signal are highlighted.
  • Prompt engineering is a foundational skill.
  • AI is a platform with potential to diffuse quickly.
  • AI can augment human potential and allow anyone to become a storyteller.

Shownotes Transcript

Translations:
中文

I see prompt engineering as a foundational skill that everyone should take time to learn so that they become discerning users of language models that can be super, super powerful and ultimately supercharge the work that they're doing today.

Hello everyone, I'm Adel, data evangelist and educator at DataCamp. And if you're new here, DataFramed is a weekly podcast in which we explore how individuals and organizations can succeed with data and AI.

Since the launch of ChatGPT, probably one of the single most trending terms in the generative AI space outside of ChatGPT has been prompt engineering. The act of tuning your instructions to get the best possible response from ChatGPT is treated like alchemy by some and science by others. So what makes the most effective prompt for ChatGPT?

Enter Alex Banks. Alex Banks has been building and scaling AI products since 2021. He writes Sunday Signal, a newsletter offering a blend of AI advancements and broader thought-provoking insights, which we have linked below. His latest course on DataCamp is on Understanding Prompt Engineering, which we have also linked below. And he consistently shares his expertise on LinkedIn and X.

Throughout the episode, we spoke about strategies for building more effective ChatshipD prompts, why certain prompts fail and others succeed, how to best approach measuring the effectiveness of a prompt, what he thinks the future of prompt engineering will look like, and a lot more. If you enjoyed this episode, make sure to let us know in the comments, on social, or more. And now, on to today's episode. ♪

Alex Banks, great to have you on DataFramed. Adil, it's a real pleasure. Thank you so much for coming on. So you teach the Understanding Prompt Engineering course on Datacamp and you write the Sunday Signal, which is a great AI newsletter that I highly recommend everyone to subscribe to. And you're also super active in the AI community. So maybe before we get into discussing prompt engineering, walk me through how you got into AI and what led you to create the Sunday Signal.

Sure. Well, thank you very much for having me, Adil. And to give you a bit of an idea of my background and how I got involved in the AI space, I started creating content at the beginning of January 2022, and I was immediately taken back by the opportunity of AI, where it is essentially on-demand intelligence. And what I've

What I mean by that, it is way cheaper and way faster than what you can get from humans. And that fundamentally sparked my curiosity. And what that led me to realize was that if you see the explosion of stuff people are building right now, it ultimately seems like AI is the platform that everyone's been waiting for. And I'm a firm believer that there will be

never be a technological advancement in our lifetimes that will diffuse as fast as AI. And the beauty now is we have tools that can augment human potential and allow anyone to become a storyteller. And for me, that storytelling component, Adil, is just such a foundational component from anything from business leadership,

writing. I'm a big believer in that the most important trait of an entrepreneur is someone who can tell a story. And many people don't think, you know, well, at least for me, I didn't think I could be a good storyteller. I was always quite reserved as a child.

But AI just allowed me to realize that potential to the nth degree, whereby there are now tools that you, I, anyone else can use, like ChatGPT, video tools such as RunwayML, image tools like MidJourney, and coding tools like Cursor that dramatically reduce the barriers between

to create something meaningful. And that really excites me. And how that led me to get started writing Sunday Signal, Adol, was I was at the beginning of writing on, it was previously Twitter, now X. I'm still trying to get past that naming friction point. And I was starting to look for nuggets of wisdom that I could infuse into my writing. And luckily, Twitter's great in that it

forces you to put your writing into 280 character chunks to really distill the meaning in essence of what it is you want to convey. And what I realized was that my Twitter feed was just getting more and more noisy

in this sea of chatter. And it always felt difficult to cut through the noise. And one way I overcame that at all was I was creating these things called Twitter lists, which is a great way to create this curated feed that I could digest a lot easier and take my learnings from. And the questions that helped inform that were, look, who do I trust? Who do I respect? And

How do I deploy this in my own writing so that I'm not only telling stories, but also delivering insights at the same time?

And what that led me to realize was that, look, if I'm going to write something meaningful, I may as well do exactly what it says on the tin. And Sunday Signal is exactly that. It gives you signal in a sea of noise every Sunday with my favorite AI highlights, one article, one idea, one quote, and one question to ponder every Sunday. It's straight in your inbox. And what I love about that is it infuses really

Both my curiosities, number one, be it the cutting edge of AI. And secondly, you know, you've got the timely and the timeless ideas. The timely be it the AI highlights and the timeless ideas be an article by Paul Graham or an idea that can stand the test of time for centuries to come.

And I think for me, that is such a beautiful infusion following the barbell approach of ideas. And I think where that leads me to now is really a position where I'm using these AI tools. At least I'm trying to be a more discerning user of these tools as I keep on infusing what I learn, whether that be off Twitter, a subreddit into my writing and ultimately distilling the signal from the noise.

Yeah, and definitely agree there on the signal from the noise. I highly recommend, as I mentioned, Sunday Signal, jam-packed with information and tons of value amidst a lot of noise out there in the chatter. And you mentioned on generative AI tools being able to reduce the barrier to entry to a lot of many tasks that we thought previously required a lot of skills or had a high barrier to entry, as you mentioned, coding, writing, creating images, creating videos. And the

key to creating these really high quality outputs from AI systems rests with effective prompt engineering, right? Which is kind of the meat of today's conversation. So maybe to kind of deep dive into what makes an effective prompt in ChatGPT, can you give me an example or a trait of what makes a good prompt in ChatGPT?

Yeah, look, Adele, the number one piece of advice I've picked up when writing an effective prompt in JackGBT is that your output can only be as good as your input. And I think that resonates deeply with me, where if I were to ask, write me an essay on Formula One, it's going to produce something pretty vague. It's going to be okay, but it's going to be super, super vague. But when I start to highlight my interests, ideas, and preferences,

a really clever thing starts to happen whereby you and the language model start to make new connections across domains and it learns from you iteratively to get your desired response. And perhaps if I could distill that into some of my favorite strategies that I like to use to ultimately get the best results from prompting ChatGPT, I think it would perhaps be useful to highlight some of these. So straight off the bat,

Clarity, number one. It's absolutely vital to include the relevant context to reduce ambiguity. For example, if I'm a teacher and I'm maybe wanting to create a lesson plan, I can quite clearly state, look, I am an eighth grade math teacher preparing to teach trigonometry. And what that does is it immediately gives a clear picture to the language model of who you are and what it is you're wanting to achieve.

Following that specificity, so the more specific you are, the closer you get to your desired answer. For example, following that trigonometry teaching, well, can you also recommend some creative real-world examples to make this lesson on trigonometry more engaging for my students? So,

All of a sudden, it has the context of who you are. It has some specificity of the length of time the lesson is going to be going on for, and also the number of students. Number three, and this is something that I think is also quite non-obvious, is sometimes keeping the prompt open-ended, Adolphe. So allowing ChatGPT to think outside of the box can often yield results

rich results that were non-obvious from the outset. So look, can you also recommend some creative real world examples so that when I'm teaching this lesson on trigonometry, it can be more engaging for my students? And that's really wonderful because it's all about sparring back and forth with the language model and uncovering ideas that you would have never otherwise have set foot on.

That's great. And then, you know, you mentioned a couple of things here on, especially on having an open-ended prompt. Maybe walk me through that concept in a bit more detail. I've never seen someone conceptualize the open-endedness of a prompt. Walk me through examples of how that looks like in a bit more detail. I think perhaps the best example here, Adil, would be to use one of the frameworks that I really enjoy and actually use quite frequently. And it really touches on that open-endedness concept.

idea quite nicely. And the framework that I like to use is the persona problem solution framework. And to give you a bit of context for this, Dave Klein, who is also a prolific creator, he runs the MGMT Accelerator. He teaches management and leadership to a whole host of individuals, also an ex-Bridgewater colleague. So under the reins of Ray Dalio, the prolific Wall Street investor.

Anyway, he asked me, Alex, look, I'm wanting to create a great prompt for my leaders. They've got a host of leadership problems they want help solving. What prompt can I use to help them answer their questions? And I spent some time scratching my head and I thought, why not allow the language model, ChatGPT, take the shoes of the prolific investor himself, Ray Dalio. So I go, look,

ChatGPT, you are an advisor to Ray Dalio. You're an expert problem solver for leadership tasks and a renowned detailed prompt writer for large language models. And here's this list of problems that I'm wanting to solve. And I go, here are all the problems. And then the really neat bit is this, Adol, where at the end, so I've defined the persona, I've specified the problems that I'm wanting to solve. And now I come in for the third and final part, which is the solution.

And I go, look, for each of these X problems, please provide number one, three relevant online resources to solve the problem.

a mental model or framework to deal with the problem at hand, and also include an explanation of how the problem is solved using the mental model or framework. And then also a reference to Ray Dalio's book Principles with respect to the problem at hand. And then the fourth and final part for this, which highlights that open-endedness that you mentioned earlier, Adil, provide a detailed prompt for a large language model to solve the problem at hand. And what we're doing here is we are

using ChatGPT to essentially prompt itself and to think, okay, what effective technique can I use here to get the most out of my auto-aggressive nature of just next word prediction? And this is me, in essence, trying to bridge the chasm between the simplistic nature of how these large language models think to something a little bit more

a little bit more involved and a little bit more beautiful, given the current constraints of the data and how these models compute right now. And what I really like about that is we get to define exactly how the solution looks like. And

From what I specified earlier, you know, write me an essay on Formula One. We've gone to this almost opposite end of the spectrum where we're getting the model to self-reflect, look inside of itself. And by using, by being inherently specific, we're really making the most and getting the most out of ChatGPT's capability. Yeah, it's interesting you mentioned here that self-reflection part because that's what I wanted to kind of touch upon.

You know, there's also a few techniques that you share in the course that we're going to expand upon throughout our conversation that lead to that kind of self-reflection. And it's interesting seeing how that provides better results, like even stuff such as take a deep breath or think step by step about your actions creates that interesting dynamic within the language model that lets itself reflect and produce better output. So maybe walk me through the importance of that self-reflection element and what you've seen while prompting tools like Chachapati.

Yeah, it's really interesting. And you've probably seen some examples pop up, some viral examples. You know, number one, take a deep breath. Or number two, I'll pay you $100 if you create a great output. And the list goes on and on, Adil. And it's really, really interesting to see and something that I think is still very hard to find a clear-cut solution to. All we're doing right now is simply learning how these models behave more and more.

And as a result, you get more information to determine how we can prompt the best output. Now, tools like this are asking or at least telling, oh, I'll tip you or take a deep breath. They are giving the language model more space and more breathing room. And from the outset, that seems quite obvious. But when you really think about it, why would I as...

a next word prediction tool, think about creating an answer that is better than if I wasn't getting $100 tip after this. And right now, that is the non-obvious bit. And we still don't have an answer for that, quite honestly, Adil. But we are gathering more and more information, more and more data as to what makes these systems perform well. And that part really excites me.

Yeah, definitely. And, you know, we're talking here about a few different ways to optimize your prompts. You mentioned that persona actions kind of solutions or like persona problem solutions framework. You also mentioned the open-endedness clarity, reducing ambiguity. I think one big challenge for me when prompting tools like ChatGP is I don't know how to evaluate whether a prompt is effective or not.

What are ways that you can evaluate the responses of ChatGPT and kind of evaluate the effectiveness as a prompt, even systematically and from a more scientific perspective than just looking at a comparison of how the output is changing? Yeah, I think that's such a fantastic question, Adil. And I think there's a really simple acronym here that I like to use.

to effectively evaluate responses from ChatGPT, and that is LARF. Now, it isn't as humorous as it actually sounds. The acronym is L-A-R-F. And whilst I go into a lot more detail on the course with understanding prompt engineering, I think it would be useful to give a high-level overview of how this can make you a more discerning user of ChatGPT to effectively evaluate your responses. So starting with L,

which stands for logical consistency. Why I like to start it here is if I'm asking chat GPT, look, what are the

and drawbacks of the drag reduction system on a Formula One car. And as you can see, we've got Formula One as quite a recurring theme throughout this podcast, which happens to be one of my favourite pastimes. Anyway, okay. And it states this list and it goes, oh, it introduces the, oh, it makes the car's top speed higher. But then also on the drawbacks, it says, oh, it makes the car's top speed higher. All of a sudden, you've got this contradicting statement saying that it is also a

a benefit, but also a drawback. And that's, whilst it's quite a simple example, it highlights that these models are fallible. They do make mistakes and using the discerning human eye to review the output and check for that coherence, I think is super, super valuable. Moving on, you have accuracy and this tendency for models to hallucinate

And what I mean by hallucination is ChachiBT can often state an answer. It can often confidently state an incorrect answer. So if you go, look, who was the first person to walk on the moon? And it goes, oh, it was Buzz Aldrin. Obviously, the correct answer is Neil Armstrong with Buzz Aldrin being the

the second person. So it's super useful to cross-reference these answers with alternate resources. For example, you can add things such as the browsing capability or even use plugins or GPTs that can reference papers and resources that are ultimately infusing the output with factual data that can ultimately lead to a better response.

R stands for relevance. So this is essentially meeting the context. And what I mean by that, Adil, is you're essentially ensuring that the response aligns with the context and actually what you wanted to get out of the answer when you were writing the prompt. So if you're asking for a list of great restaurant recommendations in London, and it says, oh, here's this recommendation, but it's in London,

New York City, all of a sudden, you aren't meeting the context of what it is that you wanted to achieve. Now, I think similarly, tools can be a fantastic way of overcoming these limitations. And I'm sure we'll get onto those a little bit later. And then the final part of this acronym, F, factual correctness. So as we're all aware, these models have a cutoff date. And when you ask a question without the context of online browsing,

it's unable to tell you what happened in January of 2024. Now, it might do, but it would be hallucinating, which we talked about earlier, where it's confidently stating this incorrect answer. For example, who won the World Cup or other great sporting events or happenings which took place past this cutoff date. Now, why I think it's really important to understand these is...

It equips you with where Chatsheput's strengths, but also limitations lie. And by understanding both, it really allows you to get the most out of the model and enables you to achieve whatever task or action it is that you set out to do.

And, you know, one thing you mentioned that I'd like to kind of latch on to and expand a bit more on is the aspect of hallucination with large language models. We've definitely seen that both pretty funny high profile use cases of large language models hallucinating in public, but also, you know, a bit of a dark side of the aspect of hallucination was that

models like Chachapati, especially image generation models and video generation models, tend to have biased outputs. Maybe that's been fixed now, but a year ago, if you had put in Midjourney a picture of five doctors, these five doctors will most likely be of a particular demographic versus another. Maybe walk me through ways that you can leverage prompt engineering to minimize bias and this type of harmful impacts or aspects of the outputs of large language models and AI generation tools in general.

Yeah, I think it would be perhaps useful to tie this to an example, Adol, and ways you can overcome that. And the most prolific that I've seen of recent is the reversal curse, which was first highlighted to me in Andre Kapathy's introduction to large language models video. And the way this reversal curse works is if you ask ChatGPT who is

Tom Cruise's mother, it will respond with, it's Mary DeFifer, which is correct. But if you ask who is Mary DeFifer's son, Jack G.P.T. will respond, I don't know, as a large language girl, and that usual spiel that it provides. And that's really interesting because it knows who the mother is, but it doesn't know who the son is. And what does that show?

So it shows that ChatGPT's knowledge is very one-dimensional. And what I mean by that is you have to ask questions from certain angles, certain ways to peer in and find the answer. And this is very unlike other feats of engineering, both from a software and hardware lens, because we still don't know exactly how these models work. And what that shows is...

number one, an inherent flaw in its knowledge understanding, but number two, how that leads into biases, which are so often a mirror of society and a mirror of the quality of the data set that these models have been trained on. So given that it is ingesting an enormous quantity of information that is typically web docs, texts, et cetera,

When you ingest that, it actually amplifies these biases that are present inside of the data. And these include stereotypes, misinformation, asking a question as simple as who typically cooks in the household and it responds in a gendered answer. It quite clearly represents a bias that has been

It might have been absorbed from historical or cultural data. And the best way I've found to overcome this is to use tools that can include

or at least it can overcome this factual correctness idea. So I like to use web browsing because I like to get up-to-date information. I like to use tools that reference an archive of papers because then I get to factually correct the answers that I'm receiving. And what I like about that is it allows me to overcome these shortcomings

and avoid any imagination of details or facts that can so often lead to incorrect outputs that if used in sensitive settings can lead to often results that can ultimately be quite devastating in a sensitive time. So it's really important that we use the appropriate tools in the appropriate setting to get the answers that we want.

Yeah, and you mentioned tools here, specifically, you know, using chat GPTs, browsing capabilities, also maybe using other GPTs if actually correct. What are some non-obvious tools that you've worked with that helped you improve the quality of your prompting? Yeah, such a great question, Adil. And I think, you know, there are some that come to mind. And I think, you know,

If I'm looking at, say, the GPT store, tools, say, Scholar AI, for example, is one where you get to, it essentially acts as an AI scientist, where it searches across 200 million plus peer-reviewed articles, and you're able to create, save, and summarize citations. And the beauty about doing that is you get to extract figures and tables from articles

lots of peer-reviewed articles. And why is that good? It's good because JCPT is famous for hallucinating, and that's a serious problem when you're wanting to write and develop source material. So by being able to directly query relevant peer-reviewed studies that can link directly to the data, all of a sudden we start to enter this new era of value creation with AI where

Now, all of a sudden, I can create something really meaningful that is backed by data and factual correctness by being able to tap into this wonderful resource of human knowledge and intuition on demand.

And I take a bit of a step forward here on our prompting journey, right? Because one aspect of the course that is kind of more advanced prompt engineering techniques is chain of thought prompting. And I've seen you write about this and I've seen the community write about this quite a lot. Maybe walk us through what chain of prompting is in a bit more detail and share some examples over why chain of thought prompting is so effective at building up effective outputs from tools like ChatGPT.

Yeah, absolutely. It's such a great tool that you can equip to your arsenal when you're wanting to create great outputs using tools like ChatGPT, Erdal. And so what is chain of thought prompting? Chain of thought prompting is a fairly advanced technique that ultimately takes training a step further, whereby you're not just...

giving chat GPT examples, but you're actually providing a roadmap of how to arrive at the answer. And what I love is that you get to really be almost this guiding hand to the model for exactly where you want to go. So if I'm solving a homework problem, or if I'm

traveling or if I'm doing something that is unique and specialized that is typically outside of ChachiBT's knowledge base, I can allow the model to think step by step. So there are different ways we can think about this and we can think about breaking it down, Adil. So you've got a zero-shot chain of thought where you're going, look,

Here's this scenario, just think step by step and go for it. You're almost throwing Jack GPT in the deep end where you're getting it to reason, think through, but you don't have this predefined set of thoughts for it to reason through. Now that can be useful. You get to peer inside of the model's thought process, verify and trust its conclusions.

But I feel there are better ways of going about this, which involve one shot and a few shot chain of thought prompting. So one shot is really just, it means one example. So how can I provide an example of

solving this problem that ChatGPT can learn from and ultimately inform the output that it's going to respond by. Few shot just means a few examples. I'm going to give a few different scenarios with some nuances and subtleties for how I can reason through this answer. And training techniques are really, really great because they ultimately help shape the answer that you want to achieve. You're very much

as I said earlier, handholding the model and directing ChatGPT to get exactly, or at least as close to the answer as you like. And as I'm sure you'll explore inside the course that is understanding prompt engineering, for example, if we use one of my favorites, you know, I'm

I'm an astronaut in space. I met some aliens. I avoided two. I met three. I said goodbye to four. How many aliens did I meet? And then the answer would be, okay, well, does avoiding an alien mean that I met them? No. So therefore, don't include it in my answer. Little subtleties that the model may or may not have assumed.

that acted as an interaction can now be easily verified and confirmed that it is or isn't the right thing to include when reasoning through this problem. And that's the beauty of chain of thought.

Yeah, and I experienced that beauty as well, you know, that you mentioned here, like, especially for writing tasks, because you're able to show the model examples of your own writing, right? And it's able to emulate your voice and tone, your structure to phrasing, right? So being able to provide examples is very useful for these types of use cases when working with ChatGPT. Yeah, examples can be super, super helpful. And I often use it a lot when I'm crafting emails. So

I like to write emails in my own voice, in my own tone. And if I'm wanting to provide examples and ultimately get a response for something quite complex, I can put in some unstructured thoughts, some unstructured ideas, and present some examples of how I typically write my emails. And all of a sudden, ChatGPT will sound quite close to how I like to respond, which is a really beautiful thing.

Now, there's one thing that I wanted to kind of also, you know, I'd be remiss not to ask you about given that we're having a discussion here is that something that I and quite a few people in the Chachapati and AI communities have picked up on in the past couple of months, which is Chachapati is becoming more and more lazy.

I'm not sure if you've noticed that as well, where ChatGPT's outputs are becoming more laconic, more terse, like a bit. It's not as useful or helpful as it was before. I've even seen the OpenAI leadership team address this in the past. One, have you seen this as something that's been in your experience of using ChatGPT? And maybe if so, what are ways that we can leverage prompt engineering to kind of reduce some of that downside effect?

It's really interesting. And I definitely saw a bit of a drop off from when GPT-4 was released to some of the more recent outputs that I was receiving. Now, whether or not that is a compute constraint for

the hundreds of millions of users that are querying these models to then use different or perhaps scaled down versions to provide these answers, that remains a question to be asked. But what we can think about here in terms of getting better responses is to use some of these techniques that we highlighted previously. So using chain of thought, using better examples, because ultimately that

that final stage of training these models before they are put in the hands of consumer, which is reinforcement learning from human feedback.

That's really the best way, choosing the best responses to these answers and getting the model to ultimately understand what constitutes a great answer and how can I act as humanly as possible. So using examples can be such a great way to help inform and help steer the model in a direction that

it can so often go off on a tangent and provide irrelevant and otherwise generalized responses that aren't specific to the problems that you're wanting to solve. Yeah, brilliant. So maybe let's switch gears here a bit, Alex. And instead of talking about best practices in prompt engineering, let's take a higher level look and look at prompt engineering as a skill set or maybe as a career path.

Big part of the conversation last year is that prompt engineering will become more and more of a career path. And, you know, we started seeing at certain points in time roles like prompt engineer pop up in the lexicon, but also in certain job posting. Do you think that prompt engineering will become a viable career path in the future or a skill that just everyone needs to learn, like Googling, for example? So I'd love to see here where you sit on that side of that debate.

I see prompt engineering as a foundational skill that everyone should take time to learn so that they become discerning users of language models that can be super, super powerful and ultimately supercharge the work that they're doing today.

We see this through the job opportunities that are advertised today, Adol, where if you look at companies such as Hebbia, which is a company for analyzing and searching across your documents, they're paying up to 250K for prompt engineers to join their organization and write these great system prompts that can ultimately help steer the model to be really effective in providing great outputs to the thoughts, the reasoning, and the tasks that it is programmed to do. And

Clearly, that has a lot of value attributed to it right now. Paying a quarter of a million to be a great writer, specifically for language models, I think is fantastic. It really highlights the value of the opportunity that is attributed to this right now. I think over the next

half a decade. It is going to be such a core competency that everyone must pay attention to. If you don't, you will very much likely fall behind. You will be unable to extract the best from these models.

ultimately retrieve and receive outputs that are far greater than that of your peers. So for me, it is absolutely a core skill that I'm paying a lot of attention to. But at least the conversations that I have with my peers as well, Adil, being a great prompt engineer is something that is definitely worth your time.

And do you think we'll actually see more jobs open up like prompt engineer where the core function or core responsibility of that job is writing effective prompts? Or do you think it will disappear into the background of almost most roles today? Yeah. So in terms of building products on the application layer, ultimately the system prompt is really important.

the core piece of value add that you provide to your users, right? Where you're taking this generalized LLM and you're steering it to hyper-specific use cases. And so for anything operating on that layer, I see it as absolutely not only a core competency, but also a core role that is optimized for. As we think about thinking of future states and extrapolating that across future states, Adil,

With the emergence of tools like GPT-5, and as we trend closer and closer towards AGI, thinking about prompt engineering,

starts to take a little bit of a different shape where all of a sudden these systems are getting more and more intelligent. And as we highlighted earlier, being able to prompt themselves all of a sudden starts to become a very scary but very real situation. And so I think that will then get generalized into more and more of a competency that it would be as if you were to use Microsoft Excel.

It's a tool that has now so many different spinoffs and runoffs that it just seems second nature. And that's how I see property engineering going, where it is such an acute skill to learn right now. But as we learn and as we grow and as these systems scale, it will definitely become more of a common skill.

and well-understood practice that isn't as highly attributable to where it is right now. And, you know, when you mentioned something on system prompts for building applications, right, which I think kind of segues to my next question really well, which is kind of, now there are mainly two profiles, I think, that need to learn prompt engineering. Like you have developers working with AI models, building applications that need to write

system level prompts and then you have everyone else that needs to learn like the prompt engineering for working with consumer tools like ChatGPT. Maybe focus on the developer persona a bit more deeply. What are the nuances between prompt engineering for building AI applications and prompt engineering for using tools like ChatGPT? Yeah, it's such a great question. I think there's a lot to consider here, Adil. So when you think about developers, you know,

these guys are writing super, super long prompts. Often the context window is significantly longer than write me a poem on racing. And you're working through a very long series of persona setting and at least being able to determine a great, we'll use system prompt as an example, using delimiters, separating your prompt into easily digestible sections that the language model can use and ultimately infuse into the output that it is creating. And so

When we think about comparing and contrasting those two, the former definitely takes a greater depth, but definitely there is more rigor, but also more complexity that is added to prompts of this nature versus getting started on prompt engineering. You definitely don't have to be an expert. And prompt engineering for working as a developer is really just an extension of

asking Jack GPT, you know, to think about drafting an outline for my next essay that I'm going to write. It all starts from the simple, which then becomes the complex. And when you start to understand that and start to see, okay, you know, all complex systems start as simple systems, then it becomes a lot clearer for what actually you're writing.

And so when we do start contrasting these together, it's simply an extension of what a really simple prompt used to be. And all of a sudden, you're using all those techniques that we highlighted previously, Adel, such as providing examples, the output format, the context, the style, the audience, the length, being specific, being clear. You're just bundling all of those different formulas together to create something that is significantly more powerful

than writing something quite generalized. Both examples, I think, are really great because everyone starts from ground zero, especially when I look back to when I started using ChanchiPT. As I'm sure you were, Adol, we were bewildered and shocked at how capable these systems were, especially in an interface that is so native to us as humans that is Chat. So when you go from that to all of a sudden

Using tools, using techniques that can be used to craft detailed exceptional prompts, that's where these great developer prompts lie. And I think that's a really beautiful thing to understand.

Yeah, definitely. And maybe a couple of final questions from my side before we wrap up today's episode. If you were to take a bit of a step back looking at, we've talked about prompt engineering quite a lot as a skill set, but if we want to take a step back and look at your average organization and the different profiles it has from developers, non-developers, what do you think are going to be must-have AI skills professionals need to have outside of prompt engineering? And

If you were to define what general AI literacy looks like, what would the skills that can make it up look like? I think a base AI literacy fundamentally starts with getting the most out of the current systems on the market. And there's no better place to start than understanding prompt engineering. Getting the most out of the systems, your input is only as good as your output. As we started, Adele, that is the foundational truth that must be built up from.

The next, I think, is a more proactive approach, which is recognizing the current landscape of generalized tools such as ChatGPT, image tools such as Midjourney, and audio tools such as, say, Eleven Labs. Being proactive and playing a core part to help build and determine your own organization's AI strategy and AI roadmap can be a really useful thing.

to not only get ahead, but stay ahead in this wonderful time that we are going through today. What I mean by that is thinking about it from a specific use case lens. What key problems are my business facing right now? Or am I facing as an individual right now? And what tools can I think of that could help me go from A to B, A, far,

quicker and more effective fashion than if I were to go about it alone. And a simple Google search query or even asking ChatGPT can often uncover non-obvious insights and tools, techniques that you can use to infuse inside of you or your organization's strategy to ultimately go and achieve something great. And that could be, oh, okay, I want to create a

promotional video of my tool. I might use Synthesia to create a generative AI video or an instructional video for my users. Or it could be, okay, I'm a voiceover artist, but I'm constrained by the number of voiceovers I can do in one day. Why don't I just train 11 labs to understand my voice really, really well so that I can create 100 voiceovers in an hour? And all of a sudden you're leveraging not only yourself, but your time. And that gets super, super powerful because

Because when you look at the other side of the fence, where individuals are still operating on and haven't heard of chat GPT or haven't heard of generative AI, all of a sudden you feel superhuman. And that is a really exciting thought.

Definitely, that is really exciting. And as we close out our episode, Alex, I'd remiss not to ask you, this year is going to be pivotal for the generative AI space, right? GPT-5 is most likely going to be released. Mark Zuckerberg mentioned that Lama 3 is being trained. Maybe what do you think the next generation of large language models hold for us, right? And what type of use cases will they unlock?

Yeah, I think we're definitely trending to a highly generalist intelligence system as we get closer and closer to artificial general intelligence, which is the ability to access a media knowledge work on demand. So, okay, Alex, number two, I'm wanting you to write these reports, respond to these emails, go.

and all of a sudden it can do it in my voice, in my tone quicker than me because it doesn't have to sleep, it doesn't have to pay sick leave, it doesn't have to do anything like that. And what does that mean? And where does that take us? Well, I think there is a lot of opportunity there with respect to intelligence at the edge. So I see as we are scaling towards AGI,

the proliferation of smaller models at the edge that are by an organizational individual lens that are hyper-specific and hyper-tailored to you, the way you work, the way you write, internal knowledge, organizational understanding. That is where a lot of value and a lot of alpha is yet to be exploited. That part I'm really, really excited about, which I don't think a lot of people are paying too much attention to. But I think

specific, refined models tailored to you or your organization that are super quick, super fast. And a game changer. I agree. Yeah. Will be absolutely phenomenal. So that's something that I'm really looking forward to. Okay. That is awesome. Now, Alex, as we wrap up today's episode, do you have any final notes or call to action before we end our chat?

Other than feel free to go check out the course on understanding problem engineering, nothing else to add, Adele. It's been an absolute pleasure speaking to you today. If there's one thing that I can stress, it would be if there's anything that you think about and wanting to get your feet wet inside of AI or generative AI, you will have a far easier time getting the output you desire.

by understanding to control the inputs and write a great input. And the way you do that is by understanding prompt engineering, which I believe is a foundational skill for the next five, 10 years. So pay a lot of attention to it, respect it, and go and move and move fast. And whoever's listening, make sure to subscribe to the Sunday Signal. And with that, thank you so much, Alex, for coming on Data Friend.