Slide 1

Slide 1 text

1 1

Slide 2

Slide 2 text

2 2 Download this deck: https://speakerdeck.com/ipullrank

Slide 3

Slide 3 text

3 Salutations! I’m Mike King (@iPullRank)

Slide 4

Slide 4 text

4

Slide 5

Slide 5 text

5 The Mechanics of Modern Search

Slide 6

Slide 6 text

6 6 Real quick. Let’s talk about how search engines work.

Slide 7

Slide 7 text

7 7 Search Engines Work based on the Vector Space Model Documents and queries are plotted in multidimensional vector space. The closer a document vector is to a query vector, the more relevant it is.

Slide 8

Slide 8 text

8 8 TF-IDF Vectors The vectors in the vector space model were built from TF-IDF. These were simplistic based on the Bag-of-Words model and they did not do much to encapsulate meaning.

Slide 9

Slide 9 text

9 9 Relevance is a Function of Cosine Similarity When we talk about relevance, it’s the question of similar is determined by how similar the vectors are between documents and queries. This is a quantitative measure, not the qualitative idea of how we typically think of relevance.

Slide 10

Slide 10 text

10 10 In SEO we still think Google is Here

Slide 11

Slide 11 text

11 11 The lexical model counts the presence and distribution of words. Whereas the semantic model captures meaning. This was the huge quantum leap behind Google’s Hummingbird update and most SEO software has been behind for over a decade. Google Shifted from Lexical to Semantic a Decade Ago

Slide 12

Slide 12 text

12 Word2Vec Gave Us Embeddings Word2Vec was an innovation led by Tomas Mikolov and Jeff Dean that yielded an improvement in natural language understanding by using neural networks to compute word vectors. These were better at capturing meaning. Many follow-on innovations like Sentence2Vec and Doc2Vec would follow.

Slide 13

Slide 13 text

13 13 Words are Converted to Multi-dimensional Coordinates in Vector Space

Slide 14

Slide 14 text

14 14 Tomas Mikolov Led the Word2Vec Research Tomas is a Czech computer scientist behind many of these natural language understanding innovations.

Slide 15

Slide 15 text

15 He was accompanied by the Chuck Norris of Computer Science Jeff Dean Jeff Dean has been a part of nearly every major innovation that has powered Google Search.

Slide 16

Slide 16 text

16 16 Word2Vec Gave Us Semantic and Syntactic Relationships Between Words

Slide 17

Slide 17 text

17 17 This Allows for Mathematical Operations Comparisons of content and keywords become linear algebraic operations.

Slide 18

Slide 18 text

18 18 Word2Vec Gave Us Hummingbird

Slide 19

Slide 19 text

19 19 Google Has Been More Like This Since Hummingbird

Slide 20

Slide 20 text

20 20 We Went from Sparse Representations to Dense Representations

Slide 21

Slide 21 text

21 21 This is a huge problem because most SEO software still operates on the lexical model.

Slide 22

Slide 22 text

22 22 What’s interesting is that Generative AI comes from Search technology

Slide 23

Slide 23 text

23 23 8 Google Employees Are Responsible for Generative AI https://www.wired.com/story/eight-google-employees-invented-modern-ai-transformers-paper/

Slide 24

Slide 24 text

24 24 The Transformer The transformer is a deep learning model used in natural language processing (NLP) that relies on self- attention mechanisms to process sequences of data simultaneously, improving efficiency and understanding in tasks like translation and text generation. Its architecture enables it to capture complex relationships within the text, making it a foundational model for many state-of-the-art NLP applications.

Slide 25

Slide 25 text

25 25 The Transformer Gave us BERT

Slide 26

Slide 26 text

26 26 Word2Vec Captured Relationship, but Not Context – BERT Captures Context

Slide 27

Slide 27 text

27 27 BERT Yields Embeddings with Higher Dimensionality and Information Capture

Slide 28

Slide 28 text

28 28 Since the Introduction of BERT, Google Has Looked More Like This

Slide 29

Slide 29 text

29 29 Under the SGE/AIO Model, Google is Structured Liked This

Slide 30

Slide 30 text

30 30 Embeddings are one of the most fascinating things we can leverage as SEOs to catch up to what Google is doing.

Slide 31

Slide 31 text

31 Site Embeddings Are Used to Measure How On Topic a Page is Google is specifically vectorizing pages and sites and comparing the page embeddings to the site embeddings to see how off-topic the page is. Learn more about embeddings: https://ipullrank.com/content-relevance

Slide 32

Slide 32 text

32 Content needs to be more focused We’ve learned definitively that Google uses vector embeddings to determine how far off given a page is from the rest of what you talk about. This indicates that it will be challenging to go far into upper funnel content successfully without a structured expansion or without authors who have demonstrated expertise in that subject area. Encourage your authors to cultivate expertise in what they publish across the web and treat their bylines like the gold standard that it is.

Slide 33

Slide 33 text

33 Build Topic Clusters Well defined topic clusters can position your website and brand into an authority in your space and strengthen your entities in the eyes of Google. A site that focuses on a series of topics that are relevant to each other are going to benefit in rankings. Here are a few tools that can help you design and build your topic clusters systematically. Thruuu https://thruuu.com/keyword-clustering-tool Keyword Insights https://www.keywordinsights.ai/features/keyword-clustering/

Slide 34

Slide 34 text

34 Let Screaming Frog Do the Heavy Lifting Generate embeddings while you crawl using Screaming Frog SEO Spider. Take the file to Colab and do the following things: Keyword - Landing Page Relevance Scoring Keyword Mapping Link Building Target Identification Redirect Mapping Internal Link Mapping https://ipullrank.com/vector- embeddings-is-all-you-need You can also work with your language model to combine crawl data with SERP data and do things like information gain calculations.

Slide 35

Slide 35 text

35 35 Leveraging generative AI is a combination of content strategy, your unique creative angles, and deep understanding of the technical nuances of a channel.

Slide 36

Slide 36 text

36 36 This is our opportunity to get up from the kids table. This is our opportunity to get up from the kids table

Slide 37

Slide 37 text

37 People are Searching Differently?

Slide 38

Slide 38 text

38 38 In the States: “Search It Up” is the new “Google it”

Slide 39

Slide 39 text

39 39

Slide 40

Slide 40 text

40 40 What does it mean when a brand ceases to be a verb?

Slide 41

Slide 41 text

41 There’s A Lot of Discussion of ChatGPT Replacing Google The same is true for Tiktok, Perplexity, Bing’s CoPilot, and [insert new genAI search tool here].

Slide 42

Slide 42 text

42 42 40% of People Leaving ChatGPT Go to Google My assumption is that many of these people are fact-checking. That is a bad behavior to establish for nearly half of your users. This is also an indication that people are deeply aware of the issues related to hallucinations. In other words, people don’t trust the product.

Slide 43

Slide 43 text

43 43 Yes, TikTok is a (Nascent) Search Engine 41% of Tiktok users perform searches, but the search volume around a series of broad and meaningful queries is not there to make it more than a small supplement to Google Search.

Slide 44

Slide 44 text

44 44 21% of People Going to Tiktok Come from Google 24.5% of People Leaving Tiktok Go to Google

Slide 45

Slide 45 text

45 45 The Last Time the Major Search Engine Died it Looked Like This

Slide 46

Slide 46 text

46 46 Google Search Still Dwarfs Everything, but More People are Using More Channels

Slide 47

Slide 47 text

47 47 It means there is fragmentation in how information needs are being met.

Slide 48

Slide 48 text

48 48 I wish OpenAI the best with this. It will be very difficult to supplant Google as the search engine of record.

Slide 49

Slide 49 text

49 49 Users Have Very High Expectations of a Search Box

Slide 50

Slide 50 text

50 50 …but search outside of major search engines is still really bad

Slide 51

Slide 51 text

51 51 Google is still the main event, but we are going back into a world where we need to optimize for multiple search engines across a series of channels.

Slide 52

Slide 52 text

52 So, It Turns Out Google is All About the Clicks

Slide 53

Slide 53 text

53 53 Google’s Algorithms Inner Workings Have Been Put on Full Display Lately Through a combination of what’s come out of Google’s DOJ antitrust trial and the Google API documentation leak, we have a much clearer picture of how Google actually functions.

Slide 54

Slide 54 text

54 54 I was the First to Publish on the Google Leak, but it was a Team Effort

Slide 55

Slide 55 text

55 55 We Now Have a Much Stronger Understanding of the Architecture https://searchengineland.com/how-google-search-ranking-works-445141

Slide 56

Slide 56 text

56 56 These specifics help us zero in on what really matters.

Slide 57

Slide 57 text

57 The Primary Takeaway is the Value of User Behavior in Organic Search Google’s Navboost system keeps track of user click behavior and uses that to inform what should rank in various contexts.

Slide 58

Slide 58 text

58 58 Google Has Denied this Many Times

Slide 59

Slide 59 text

59 59 Many times…

Slide 60

Slide 60 text

60 60 But it’s Been Confirmed in Pandu Nayak’s Testimony

Slide 61

Slide 61 text

61 61 And in Google’s documentation for their Cloud Search services

Slide 62

Slide 62 text

62 It Looks Like Google is Using Clicks After All…

Slide 63

Slide 63 text

63 63 I remain adamant that both Google and the SEO community owes @randfish an apology.

Slide 64

Slide 64 text

64 64

Slide 65

Slide 65 text

65 65 That’s Also Why Google Was So Mad when @DejanSEO Did This…

Slide 66

Slide 66 text

66 66 User Click Data is What Makes Google More Powerful Than Any Other Search Engine The court opinion in the DoJ Antitrust trial, Google’s leaked documents, and Google’s own internal documentation all support the fact that click behavior is what makes Google perform the way that it does.

Slide 67

Slide 67 text

67 67 13 Months of Google Data = 17 Years of Bing Data

Slide 68

Slide 68 text

68 68 You Need to Keep People on your Website to Win at Organic Search

Slide 69

Slide 69 text

69 Modern SEO Needs UX Baked-in Google has expectations of performance for every position in the SERP. The user behavior signals collected reinforce what should rank and demote what doesn’t perform just like a social media channel. The best way to scale this is by generating highly-relevant content with a strong user experience.

Slide 70

Slide 70 text

70 How Does PPC Fit Into This?

Slide 71

Slide 71 text

71 71 Navboost Makes it Clear that Paid and Organic both operate from the concept of expected CTR

Slide 72

Slide 72 text

72 72 Larry Kim Has Talked About Expected CTR and Quality Score Being a Normalized CTR Effectively, what Larry described many years ago is very similar to how NavBoost works in Organic Search. https://www.marketingprofs.com/articles/2014/25432/four-adwords-mistakes-that-drag-your-ctr-down Google has taken matters into its own hands and is using expected CTR to drive everything via Performance Max.

Slide 73

Slide 73 text

73 73 Google wants us to spend more and the way we do that is by getting marketers higher ROI, so Google is using generative AI to help you improve.

Slide 74

Slide 74 text

74 74 Marketers are Expecting Disruption of Search Marketing from GenAI https://www.emarketer.com/content/google-streamlines-ad-creation-with-new-ai- features-performance-max

Slide 75

Slide 75 text

75 75 Google is pushing everyone back to being strategic marketers rather than channel managers.

Slide 76

Slide 76 text

76 76 Google is pumping Generative AI all over PPC https://support.google.com/google-ads/answer/14150986?hl=en

Slide 77

Slide 77 text

77 77 Google is pumping Generative AI all over PPC - Creative Generation https://blog.google/products/ads-commerce/ai-powered-ads-google-marketing-live/

Slide 78

Slide 78 text

78 78 Performance Max - Search Themes Search themes in Performance Max campaigns will have the same prioritization as your phrase match and broad match keywords in Search campaigns. Exact match keywords that are identical to the search queries will continue to be prioritized over search themes and other keywords. Keep in mind that search themes are optional. You’ll also have access to tools like brand exclusions to help control the types of search traffic that Performance Max serves on. https://support.google.com/google-ads/answer/14179631?hl=en

Slide 79

Slide 79 text

79 79 PPC Specialists Hate pMax Source: Reddit

Slide 80

Slide 80 text

80 Google AI Shopping Features - Try On https://blog.google/products/sho pping/ai-virtual-try-on-google- shopping/

Slide 81

Slide 81 text

81 81 Google Has Ad Units for AI Overviews, But No One Is Buying Them! Google has been offering these for a few months, but brands are not interested likely due to the brand safety issue that they represent.

Slide 82

Slide 82 text

82 82 Google Merchant Center Next https://blog.google/products/shopping/helping-retailers-prep-for-holiday/

Slide 83

Slide 83 text

83 83 Baidu’s AiAds Study Suggests This Concept is Here to Stay https://arxiv.org/pdf/1907.12118

Slide 84

Slide 84 text

84 Researchers Warn How it Can Go Off the Rails The study highlights how a lack of data transparency and algorithmic missteps, such as flawed conversion tracking, can lead to long-term negative impacts on ad performance. It shows how AI algorithms can misinterpret data errors, resulting in decreased impressions, clicks, and conversions, ultimately harming advertisers' ROI​. https://asistdl.onlinelibrary.wiley .com/doi/full/10.1002/asi.24798

Slide 85

Slide 85 text

85 85 Google’s own tools will be the most powerful because only they have the data on what users do within their ecosystem.

Slide 86

Slide 86 text

86 UX That Yields Better NavBoost Performance

Slide 87

Slide 87 text

87 NavBoost Performance Starts at the SERP Itself Continually testing your metadata is a must. Check out the SearchPilot team’s case studies for ideas: https://www.searchpilot.com/resources /case-studies/tag/meta-data

Slide 88

Slide 88 text

88 88 Design Content for the Human Condition Design your content so it is easier to consume and it will yield better performance metrics. https://moz.com/blog/10-super-easy-seo- copywriting-tips-for-link-building

Slide 89

Slide 89 text

89 89 Google is Using Passage Indexing to Try to Drop the User Into the Right Spot

Slide 90

Slide 90 text

90 90 Use Logical Chunking To Get Users to the Information Faster https://www.nngroup.com/articles/in-page-links-content-navigation/

Slide 91

Slide 91 text

91 91 Build Your Jump Links

Slide 92

Slide 92 text

92 92 Make Your Content Scannable

Slide 93

Slide 93 text

93 93 Jump Links in the SERP are a Good Sign

Slide 94

Slide 94 text

94 94 Build Pages that Are Easy to Parse Create semantically relevant content Build a table of contents Drop anchor links throughout the page to help Google understand where the user is meant to go.

Slide 95

Slide 95 text

95 Stop Leading With This I came for a recipe. Not your Grandma’s life story!

Slide 96

Slide 96 text

96 Less is More, More or Less It’s time to cut out the content madness

Slide 97

Slide 97 text

97 97 You Don’t Need Link Volume, You Need Link Quality Indexing Tier Impacts Link Value A metric called sourceType that shows a loose relationship between the where a page is indexed and how valuable it is. For quick background, Google’s index is stratified into tiers where the most important, regularly updated, and accessed content is stored in flash memory. Less important content is stored on solid state drives, and irregularly updated content is stored on standard hard drives. The higher the tier, the more valuable the link. Pages that are considered “fresh” are also considered high quality. Suffice it to say, you want your links to come from pages that either fresh or are otherwise featured in the top tier. Get links from pages that live in the higher tier by modeling a composite score based on data that is available.

Slide 98

Slide 98 text

98 98 No one is actually improving content when they do the Skyscraper Technique

Slide 99

Slide 99 text

99 99 Google Stores Your Content Like the Wayback Machine and Uses the Change History Google’s file system is capable of storing versions of pages over time similar to the Wayback Machine. My understanding of this is that Google keeps what it has indexed forever. This is one of the reasons you can’t simply redirect a page to an irrelevant target and expect the link equity to flow. The docs reinforce this idea implying that they keep all the changes they’ve ever seen for the page. You’re not going to get away with things by simply changing your pages once.

Slide 100

Slide 100 text

10 0 Indexing is Also Harder It’s not being talked about as much, but indexing has gotten a lot harder since the Helpful Content update. You’ll see a lot more pages in the “Discovered - currently not indexed” and “Crawled - currently not indexed” than you did previously because the bar is higher for what Google deems worth capturing from the web.

Slide 101

Slide 101 text

10 1 10 1 Google Wants to Crawl Even Less Gary Illyes has indicated that he wants to have Google crawl less. Search quality certainly cannot suffer, so crawlin has to get increasingly intelligent.

Slide 102

Slide 102 text

10 2 10 2 I Believe This is a Function of Information Gain Conceptually, as it relates to search engines, Information Gain is the measure of how much unique information a given document adds to the ranking set of documents. In other words, what are you talking about that your competitors are not?

Slide 103

Slide 103 text

10 3 103 That’s not all though…

Slide 104

Slide 104 text

104 10 4 There are Gold Standard Documents There is no indication of what this means, but the description mentions “human-labeled documents” versus “automatically labeled annotations.” I wonder if this is a function of quality ratings, but Google says quality ratings don’t impact rankings. So, we may never know.

Slide 105

Slide 105 text

105 Measure Your Content Against the Quality Rater Guidelines Elias Dabbas created a python script and tool that uses the Helpful Content Recommendations to show a proof of concept way to analyze your articles. We’d use the Search Quality Rater Guidelines which serve as the Golden Document standard. Code: https://blog.adver.tools/posts/llm-content-evaluation/ Tools: https://adver.tools/llm-content-evaluation/

Slide 106

Slide 106 text

10 6 106 In conclusion: “More content” is no longer inherently the most effective approach because there’s no guarantee of traffic from Google. Google’s sophistication won’t allow it.

Slide 107

Slide 107 text

10 7 10 7 The only content you should be making

Slide 108

Slide 108 text

10 8 10 8 I’m Leaving Y’all with Three Actions Today 1. How to Prune Your Content 2. How to Use LLMs to Generate Valuable Content 3. AI Tools to Use for SEO

Slide 109

Slide 109 text

10 9 The Content Pruning Process

Slide 110

Slide 110 text

11 0 11 0 Pruning and Optimization Work Quite Well Together

Slide 111

Slide 111 text

11 1 Aleyda Has a Process Aleyda’s workflow is a great place to work through whether your content should be pruned or not. https://www.aleydasolis.com/en/crawli ng-mondays/how-to-prune-your-website- content-in-an-seo-process- crawlingmondays-16th-episode/

Slide 112

Slide 112 text

11 2 11 2 We like automate to get to a Keep. Revise. Kill. (Review.)

Slide 113

Slide 113 text

11 3 11 3 Content Decay The web is a rapidly changing organism. Google always wants the most relevant content, with the best user experience, and most authority. Unless you stay on top of these measures, you will see traffic fall off over time. Measuring this content decay is as simple comparing page performance period over period in analytics or GSC. Just knowing content has decayed is not enough to be strategic.

Slide 114

Slide 114 text

11 4 11 4 It’s not enough to know that the page has lost traffic.

Slide 115

Slide 115 text

11 5 11 5

Slide 116

Slide 116 text

11 6 11 6 The Content Potential Rating (CPR).

Slide 117

Slide 117 text

11 7 11 7 Content Potential Score

Slide 118

Slide 118 text

11 8 11 8 Interpreting the Content Potential Rating 80 - 100: High Priority for Optimization 60 - 79: Moderate Priority for Optimization 40 - 59: Selective Optimization 20 - 39: Low Priority for Optimization 0 - 19: Minimal Benefit from Optimization If you want quick and dirty, you can prune everything below a 40 that is not driving significant traffic.

Slide 119

Slide 119 text

11 9 11 9 Combining CPR with pages that lost traffic helps you understand if it’s worth it to optimize.

Slide 120

Slide 120 text

12 0 12 0 Step 1. Pull the Rankings Data from Semrush Organic Research > Positions > Export

Slide 121

Slide 121 text

12 1 12 1 Step 2: Pull the Decaying Content from GSC Google Search Console is a great source to spot Content Decay by comparing the last three months year over year. Filter for those pages where the Click Difference is negative (smaller than 0) then export.

Slide 122

Slide 122 text

12 2 12 2 Step 3: Drop them in the Spreadsheet and Press the Magic Button

Slide 123

Slide 123 text

12 3 The Output is List of URLs Prioritized by Action Each URL is marked as Keep, Revise, Kill or Review based on the keyword opportunities available and the effort required to capitalize on them. Sorting the URLs marked as “Revise” by Aggregated SV and CPR will give you the best opportunities first.

Slide 124

Slide 124 text

12 4 12 4 Get your copy of the Content Pruning Workbook : https://ipullrank.com/cpr-sheet

Slide 125

Slide 125 text

12 5 How to Kill Content Content may be valuable for channels outside of Organic Search. So, killing it is about changing Google’s experience of your website to improve its relevance and reinforce its topical clusters. The best approach is to noindex the pages themselves, nofollow the links pointing to them, and submit an XML sitemap of all the pages that have changed. This will yield the quickest recrawling and reconsideration of the content.

Slide 126

Slide 126 text

12 6 12 6 How to Revise Content Review content across the topic cluster Use co-occurring keywords and entities in your content Add unique perspectives that can’t be found on other ranking pages Answer common questions Answer the People Also Ask Questions Restructure your content using headings relevant to the above Add relevant Structured markup Expand on previous explanations Add authorship Update the dates Make sure the needs of your audiences are accounted for Add to an XML sitemap of only updated pages

Slide 127

Slide 127 text

12 7 How to Review Content The sheet marks content that has a low content potential rating and a minimum of 500 in monthly search volume as “Review” because they may be long tail opportunities that are valuable to the business. You should take a look at the content you have for that landing page and determine if you think the effort is worthwhile.

Slide 128

Slide 128 text

12 8 What is Retrieval Augmented Generation (RAG)?

Slide 129

Slide 129 text

12 9 12 9 Combining a Search Engine with a Language Model is called “Retrieval Augmented Generation” Neeva (RIP), Bing, and now Google’s Search Generative Experience all use pull documents based on search queries and feed them to a language model to generate a response. This concept was developed by the Facebook AI Research (FAIR) team.

Slide 130

Slide 130 text

13 0 13 0 Google’s Initial Version of this is called Retrieval-Augmented Language Model Pre-Training (REALM) from 2021 REALM identifies full documents, finds the most relevant passages in each, and returns the single most relevant one for information extraction.

Slide 131

Slide 131 text

13 1 13 1 DeepMind followed up with Retrieval-Enhanced Transformer (RETRO) DeepMind's RETRO (Retrieval-Enhanced Transformer) is a language model that combines a large text database with a transformer architecture to improve performance and reduce the number of parameters required. RETRO is able to achieve comparable performance to state-of-the-art language models such as GPT-3 and Jurassic-1, while using 25x fewer parameters.

Slide 132

Slide 132 text

13 2 Google’s Later Innovation Retrofit Attribution using Research and Revision (RARR) RARR does not generate text from scratch. Instead, it retrieves a set of candidate passages from a corpus and then reranks them to select the best passage for the given task.

Slide 133

Slide 133 text

13 3 13 3 AIO is built from REALM/RETRO/RARR + PaLM 2 and MUM MUM is the Multitask Unified Model that Google announced in 2021 as way to do retrieval augmented generation. PaLM 2 is their latest (released) state of the art large language model. The functionality from REALM, RETRO, and RARR is also rolled into this.

Slide 134

Slide 134 text

13 4 13 4 Sounds cool, but how does it work?

Slide 135

Slide 135 text

13 5 13 5 Documents are Broken into Chunks and the Most Relevant Chunks are Fed to the Language Model to Generate a Response

Slide 136

Slide 136 text

13 6 How to Appear in LLMs

Slide 137

Slide 137 text

13 7 13 7 Blocking LLMs is a Mistake. Appearing in these places will be recognized as brand awareness opportunities very soon.

Slide 138

Slide 138 text

Embrace Structured Data There are three models gaining popularity: 1. KG-enhanced LLMs - Language Model uses KG during pre-training and inference 2. LLM-augmented KGs - LLMs do reasoning and completion on KG data 3. Synergized LLMs + KGs - Multilayer system using both at the same time https://arxiv.org/pdf/2306.08302.pdf Source: Unifying Large Language Models and Knowledge Graphs: A Roadmap

Slide 139

Slide 139 text

13 9 13 9 What is Mitigation for AIO? 1. Manage expectations on the impact 2. Understand the keywords under threat 1. Re-prioritize your focus to keywords that are not under threat 1. Optimize the passages for the keywords you want to save

Slide 140

Slide 140 text

14 0 14 0

Slide 141

Slide 141 text

14 1 14 1 We Can Also Show You Per Keyword How You Show Up

Slide 142

Slide 142 text

14 2 14 2 It’s all about the Fraggles. (Fragment + Handle)

Slide 143

Slide 143 text

14 3 14 3

Slide 144

Slide 144 text

14 4 14 4 The Fraggles Show What AIO Used for the AI Snapshot

Slide 145

Slide 145 text

14 5 14 5 Scroll to Text

Slide 146

Slide 146 text

14 6 14 6 Fraggles Relevance Relevance against the chunks to keyword: Relevance against AI Snapshot:

Slide 147

Slide 147 text

Check out MarketBrew’s Free Tool to Help

Slide 148

Slide 148 text

No content

Slide 149

Slide 149 text

No content

Slide 150

Slide 150 text

No content

Slide 151

Slide 151 text

15 1 The GEO team shared their ChatGPT prompts The GEO team also shared the ChatGPT prompts that help them improve their visibility. You can augment them and put them to work right away. https://github.com/GEO- optim/GEO/blob/main/src/geo_function s.py

Slide 152

Slide 152 text

Check out @GarrettSussman’s post on how to optimize for AI Overviews: https://ipullrank.com/optimize-content-for-sge

Slide 153

Slide 153 text

15 3 Using AI for your SEO

Slide 154

Slide 154 text

15 4 With AI, I’m giving y’all legos. What you build is up to you, but I’m going to show things to consider.

Slide 155

Slide 155 text

15 5 Setting Up LLMs Locally You don’t need ChatGPT anymore

Slide 156

Slide 156 text

15 6 15 6

Slide 157

Slide 157 text

15 7 LLaMa 3.1 was SOTA like 3 weeks ago Facebook’s open source model is outperforming the best closed-source models on a variety of different evaluation metrics. New open source models pop up weekly that continue to shift the state of the art.

Slide 158

Slide 158 text

15 8 15 8

Slide 159

Slide 159 text

15 9 15 9

Slide 160

Slide 160 text

16 0 16 0

Slide 161

Slide 161 text

16 1 16 1

Slide 162

Slide 162 text

16 2 16 2

Slide 163

Slide 163 text

16 3 163 You can now unlock state of the art generative AI use cases from your laptop for free.

Slide 164

Slide 164 text

16 4 Make Sure You Hook It Up To Your GPU On a Windows machine you’ll need to go to the NVIDIA Control Panel and add the Ollama server application under Manage 3D Settings.

Slide 165

Slide 165 text

16 5 How to Use LLMs for Content and SEO

Slide 166

Slide 166 text

16 6 16 6 The Three Laws of Generative AI content 1. Generative AI is not the end-all-be-all solution. It is not the replacement for a content strategy or your content team. 2. Generative AI for content creation should be a force multiplier to be utilized to improve workflow and augment strategy. 1. You should consider generative AI content for awareness efforts, but continue to leverage subject matter experts for lower funnel content.

Slide 167

Slide 167 text

16 7 16 7 Think back to 7 Minutes Ago - Retrieval Augmented Generation

Slide 168

Slide 168 text

16 8 16 8 It’s Not Difficult to Build with Llama Index sitemap_url = "[SITEMAP URL]" sitemap = adv.sitemap_to_df(sitemap_url) urls_to_crawl = sitemap['loc'].tolist() ... # Make an index from your documents index = VectorStoreIndex.from_documents(documents) # Setup your index for citations query_engine = CitationQueryEngine.from_args( index, # indicate how many document chunks it should return similarity_top_k=5, # here we can control how granular citation sources are, the default is 512 citation_chunk_size=155, ) response = query_engine.query("YOUR PROMPT HERE")

Slide 169

Slide 169 text

16 9 169 Everyone can code now.

Slide 170

Slide 170 text

17 0

Slide 171

Slide 171 text

17 1 17 1

Slide 172

Slide 172 text

172 PAGE GENERATIVE AI PRODUCTIVITY USE CASES RAG opens up a series of generative AI use cases that work well for your situation. Briefing & Business Cases Content Analysis First-pass Brand Review First-pass Legal Review Content First Draft Keyword Insertion Structured Data Generation Link Identification & Insertion Generating Voiceovers Generating Images Generating Videos Writing Code

Slide 173

Slide 173 text

17 3 @BritneyMuller’s Guide to Using Colab Britney talked about how easy it is to use Colab with Python. Now it’s even easier to using LLMs. https://github.com/BritneyMuller/colab- notebooks?tab=readme-ov-file

Slide 174

Slide 174 text

17 4 Just describe what you want You can tell your language model what you want the code to do and it will handle the rest. If it doesn’t work, just describe what went wrong or paste the error and it will fix it for you. In this example my prompt is: {write python code for colab that takes a csv file of keywords and using bertopic with the chatgpt to compute the natural language topics for each row.}

Slide 175

Slide 175 text

17 5 17 5 Colab + Gemini

Slide 176

Slide 176 text

17 6 17 6 Llama Index - RAG - https://www.llamaindex.ai/

Slide 177

Slide 177 text

17 7 17 7 LangChain - Build Agents - https://www.langchain.com/

Slide 178

Slide 178 text

17 8 17 8 LangFuse - Prompt Management & Observability tool - https://langfuse.com/

Slide 179

Slide 179 text

17 9 17 9 Literal AI

Slide 180

Slide 180 text

18 0 180 You don’t need to code for any of this.

Slide 181

Slide 181 text

18 1 18 1

Slide 182

Slide 182 text

18 2 18 2

Slide 183

Slide 183 text

18 3 18 3

Slide 184

Slide 184 text

18 4 184

Slide 185

Slide 185 text

18 5 185

Slide 186

Slide 186 text

18 6 186

Slide 187

Slide 187 text

18 7 18 7 Bubble - No Code Apps - https://bubble.io/

Slide 188

Slide 188 text

18 8 Integrate Promptitude with Zapier or Make

Slide 189

Slide 189 text

18 9 Prompts You Need To Write ChatGPT is very effective at doing the following SEO related tasks: Page Title writing Meta Description writing Keyword Insertion Link Insertion You should use your own prompts for these though so you don’t copy other people’s patterns.

Slide 190

Slide 190 text

19 0 Page Titles Feature Token Count features Hypothesis There’s no hard max page title length indicated in the attributes so we can test lengths longer than the 60-70 characters to determine impact.

Slide 191

Slide 191 text

19 1 Page Title Test Hypothesis: A page title that’s longer than the standard best practice will negative impact rankings for primary keyword target. Variables: Control Short page title Long page title Metrics: Ranking Increase

Slide 192

Slide 192 text

19 2 lastSignificantUpdate lastSignificantUpdate - The date of the last time Google encountered the page as materially updated. Feature Hypothesis Making substantial updates to pages regularly yields improved crawl activity and more opportunities to rank better.

Slide 193

Slide 193 text

19 3 19 3 TL;DR Generation for Significant Updates

Slide 194

Slide 194 text

19 4 Test Structure The goal of this test is to determine how much content is considered a “significant update” that yields crawl activity. Create control and variants pages testing the length of added content: We measure the impact on organic traffic in order to capture changes to rankings and/or changes to clickthrough rate.

Slide 195

Slide 195 text

19 5 19 5 Brief Writing

Slide 196

Slide 196 text

19 6 19 6 Persona Modeling based on Desk Research Data

Slide 197

Slide 197 text

19 7 19 7 Persona Writing with SparkToro

Slide 198

Slide 198 text

19 8 19 8 Brand Voice and Tone Review

Slide 199

Slide 199 text

19 9 19 9 Schema Markup Generator https://chatgpt.com/g/g-MvH0WHO3e-schema-markup-generator-gpt

Slide 200

Slide 200 text

20 0 20 0 Taskade - AI All-in-One - https://www.taskade.com/

Slide 201

Slide 201 text

20 1 20 1 AIPower - All-in-One for WP - https://aipower.org/

Slide 202

Slide 202 text

20 2 20 2 Thunderbit - Build No Code AI Automation Tools - https://thunderbit.com/

Slide 203

Slide 203 text

20 3 20 3 Keyword Insights - Keyword Clustering Tool

Slide 204

Slide 204 text

20 4 20 4 Octoparse - Combine a scraper with Generative AI - https://www.octoparse.ai/

Slide 205

Slide 205 text

20 5 20 5 DejanSEO’s LinkBERT - https://www.linkbert.com https://dejanmarketing.com/tools/li nkbert/

Slide 206

Slide 206 text

20 6 20 6 InLinks - Automated Internal Links - https://inlinks.com/

Slide 207

Slide 207 text

20 7 20 7 Respona - AI-enabed Link Building - https://www.respona.com

Slide 208

Slide 208 text

20 8 20 8 UX Sniff - An AI-enabled HotJar https://uxsniff.com/

Slide 209

Slide 209 text

20 9 Roll the Credits

Slide 210

Slide 210 text

21 0 21 0 What you should know and do to win Google is still the primary show in town Relevance is a quantitative measure GenAI works on the same math as search engines Focus on making your chunks for relevant to rank in GenAI Search Improve UX to drive more long clicks PPC and SEO are both operating on an threshold of expected CTR Focus on content your audience wants, prune what they don’t Use RAG to generate content with AI Embrace AI tools to improve your workflows and your ability to test

Slide 211

Slide 211 text

21 1

Slide 212

Slide 212 text

Contact me if you want to get better results from your SEO: [email protected] Thank You | Q&A Award Winning, #GirlDad Featured by Content Pruning worksheet: https://ipullrank.com/cpr-sheet/ Download the Slides: https://speakerdeck.com/ipullrank Mike King Chief Executive Officer @iPullRank