Majestic

  • サイトエクスプローラ
    • Majestic
    • サマリ
    • 参照元ドメイン
    • 被リンク
    • * 新規
    • * リンク切れ
    • コンテキスト
    • アンカーテキスト
    • ページ
    • トピック
    • Link Graph
    • 関連サイト
    • 高度なツール
    • Author ExplorerBeta
    • Summary
    • Similar Profiles
    • Profile Backlinks
    • Attributions
  • 比較
    • サマリ
    • 被リンク履歴
    • Flow Metric 履歴
    • トピック
    • クリークハンター
  • リンクツール
    • マイ Majestic
    • 最近のアクティビティ
    • レポート
    • キャンペーン
    • 確認済みドメイン
    • OpenApps
    • API キー
    • キーワード
    • キーワードジェネレータ
    • キーワードチェッカー
    • 検索エクスプローラ
    • リンクツール
    • 一括被リンク
    • 近隣サイトチェッカー
    • URL を送信
    • テスト中
    • Index Merger
    • Link Profile ファイト
    • 相互リンク
    • Solo Links
    • PDF レポート
    • Typo Domain
  • Free SEO Tools
    • スタートアップ
    • 被リンクチェッカー
    • Majestic Million
    • Majestic プラグイン
    • Google Sheets
    • Post Popularity
    • Social Explorer
  • サポート
    • ブログ 外部リンク
    • サポート
    • スタートアップ
    • ツール
    • 会員プランとお支払い
    • FAQ(よくある質問)
    • 用語集
    • スタイルガイド
    • チュートリアル動画
    • API リファレンスガイド 外部リンク
    • お問い合わせ
    • 被リンクとSEOについて
    • SEO in 2025
    • The Majestic SEO Podcast
    • All Podcasts
    • What is Trust Flow?
    • リンク構築ガイド
  • 新規登録無料
  • プランと料金
  • ログイン
  • Language flag icon
    • English
    • Deutsch
    • Español
    • Français
    • Italiano
    • 日本語
    • Nederlands
    • Polski
    • Português
    • 中文
  • スタートアップ
  • ログイン
  • プランと料金
  • 新規登録無料
    • サマリ
    • 参照元ドメイン
    • マップ
    • 被リンク
    • 新規
    • リンク切れ
    • コンテキスト
    • アンカーテキスト
    • ページ
    • トピック
    • Link Graph
    • 関連サイト
    • 高度なツール
    • サマリ
      Pro
    • 被リンク履歴
      Pro
    • Flow Metric 履歴
      Pro
    • トピック
      Pro
    • クリークハンター
      Pro
  • 一括被リンク
    • キーワードジェネレータ
    • キーワードチェッカー
    • 検索エクスプローラ
      Pro
  • 近隣サイトチェッカー
    Pro
    • Index Merger
      Pro
    • Link Profile ファイト
      Pro
    • 相互リンク
      Pro
    • Solo Links
      Pro
    • PDF レポート
      Pro
    • Typo Domain
      Pro
  • URL を送信
    • Summary
      Pro
    • Similar Profiles
      Pro
    • Profile Backlinks
      Pro
    • Attributions
      Pro
  • カスタムレポート
    Pro
    • スタートアップ
    • 被リンクチェッカー
    • Majestic Million
    • Majestic プラグイン
    • Google Sheets
    • Post Popularity
    • Social Explorer
    • スタートアップ
    • ツール
    • 会員プランとお支払い
    • FAQ(よくある質問)
    • 用語集
    • チュートリアル動画
    • API リファレンスガイド 外部リンク
    • お問い合わせ
    • メッセージ
    • The Company
    • スタイルガイド
    • 利用規約
    • プライバシーポリシー
    • GDPR
    • お問い合わせ
    • SEO in 2025
    • The Majestic SEO Podcast
    • All Podcasts
    • What is Trust Flow?
    • リンク構築ガイド
  • ブログ 外部リンク
    • English
    • Deutsch
    • Español
    • Français
    • Italiano
    • 日本語
    • Nederlands
    • Polski
    • Português
    • 中文

Familiarise yourself with the capabilities and limitations of large language models

Bastian Grimm

Just as we need to get into the nuts and bolts of SEO, we also need to take the time to understand the technology that powers AI, as Bastian Grimm from PeakAce explains.

@basgr    
Bastian Grimm 2024 podcast cover with logo
More SEO in 2024 YouTube Podcast Playlist Link Spotify Podcast Playlist Link Audible Podcast Playlist Link Apple Podcast Playlist Link

Familiarise yourself with the capabilities and limitations of large language models

Bastian says: “Get familiarised with large language models.

This is one of the foundational elements of AI. Marketers need to understand how they work and what capabilities they have, but also what limitations they have, so they can become more efficient and utilise them in the best way possible.”

How would you summarise the way that large language models work?

“Essentially, it’s a huge amount of data that has been put into a specific type of database, and with that data, you can fulfil certain types of tasks. For example, ChatGPT is an interface for a large language model provided by OpenAI.

It’s important to understand that these models don’t write anything. They have training data and, based on that training data, they predict the likelihood of what’s supposed to happen next. This is how they generate text.

If you give the model a task through prompting (formulating something via the interface), there is no creative magic happening. Based on the data that has been ingested in the training process, the model simply creates a new output.”

What should and shouldn’t LLMs be used for?

“There’s a whole lot of new tech around AI, including on the generative AI side, and it’s moving extremely fast. The biggest issue is that a lot of the stuff that’s currently being shown off is predominantly in a beta state. Most of it is not really production-ready.

People are fascinated by the topic (myself included) and there are a lot of gains to be found already. On the other hand, we need to be cautious. Things are changing fast and what is currently there has limitations.

Say you task the model with creating a piece of text around the Audi A3. The Audi A3 has different features in different countries. In Germany, it might have different dimensions, or it might come with a different default package than it does in the UK or France. If you just task the model with producing a piece of content around that, then you might end up with mixed facts. The dimensions might be wrong, or the default pricing might be wrong, for example.

It’s extremely important to fact-check, double-check, and triple-check the output. It can be used for research, and it can be used for drafting things but it’s by no means meant to produce content without double-checking. That is one of the biggest misconceptions with ChatGPT, Jasper, or any other model right now.”

If your prompts are higher quality, do you need to fact-check less?

“You can increase the quality, but errors will still happen. The model is predicting things and predictions can be wrong.

Using the Google search generative experience (the AI-based snippet on top of the search results), I was inputting queries about tech companies. I asked, ‘How many unicorns are there in France?’, and it said, ‘27’. When I gave it a slightly different prompt for the same question it said, ‘36’. This is what we call a hallucination.

The issue is that the model doesn’t know. There is no exact number and there is no precise data source that says exactly how many unicorn companies are in France. Therefore, it’s trying to ‘guesstimate’ what the right answer could be. That is very dangerous. That was a soft example, but a lot of important research relies on information being accurate, especially from big companies like Google. You have to be mindful and cautious about blindly trusting that info.

On the other hand, they’re fantastic for speeding up certain processes and tasks that you’re doing in your day-to-day. If you are doing research, you could now use a large language model where before you would have needed to go to 10/20/30 different sources. You still have to check it, but you would have needed to check your 20 sources as well. In that regard, you’re becoming more efficient.

We have a huge client that does corporate training within their organisation, and they previously had to do recordings in person, in a studio. It’s expensive, it takes time, the person needs to be there, etc. Now, they’re doing it with an AI avatar. They recorded the speaker once and recorded their voice, so they have voice samples from previous trainings. Now they only need to produce scripts and the AI avatar is going to be in the video.

It’s even cooler because they’re a multilingual organisation. With AI, they can scale that because they just use a different avatar or voiceover and they have it in French, Spanish, etc. Some people are not as fluent in English as others, and this helps to ensure that their experience is included and onboarded. There are already massive gains to be found in terms of efficiency and even satisfaction.”

Can using a tool like Jasper be more effective than using ChatGPT natively?

“It depends on the use case.

It’s also worth noting that there are different large language models as well – like Neuroflash which is popular in the German region – and they all have different training and different types of feedback mechanisms. Some are better suited to certain types of tasks than others. Med-PaLM is a large language model specifically for pharmaceutical and medical information. It is much more accurate in those areas, but it has only been trained on that kind of content. It couldn’t give you an answer about the best chess strategy, for example.

Having specific models for specific tasks does make sense but we’re currently using large language models for a lot of things they are not meant to be used for. In the future, those models may start to rely on third-party tools through APIs or a different type of architecture so that they can then pull data from other sources when they’re not confident enough to properly answer a question.

There are a ton of smaller tools right now, which is mainly because they are trying to solve certain specific tasks. As an SEO, this is great, especially if you’re on the freelance or consulting side. You can use these tools without having to spend tons of time and/or money on implementation, which is a massive efficiency gain. However, there’s a high chance that 80% of those special interest tools will die out in the next 18/24 months and be absorbed by the bigger models.”

Are there any specific SEO tasks that you would definitely use LLMs for at the moment?

“It can already really help with things like meta descriptions and page titles when it comes to working at scale. I would still have someone handcrafting the meta description in a more manual way for the homepage or the meta description template for a certain specific type of content.

However, if you’re talking about thousands or millions of pages, then the next best thing after templating it would be to use an LLM. You can dump in a URL (or even scrape that through a ChatGPT plugin), and then it’s going to pull the data and create meta descriptions within certain limitations. For that kind of content generation, and suggestions at scale, it’s great.

It can also produce more lightweight content. An article about a certain island in Greece might be relatively easy to do compared to a very specific guideline for XYZ because that’s not based on common knowledge. It’s also great for classification. If you need a new hierarchy for internal linking, you can map out and classify pages through an LLM. It’s much more efficient and it can ingest much more data than you would ever be able to do manually.

It is going to get crazier in the next few years, from a variety of standpoints. Google has the SGE snippet on top, and that’s obviously going to change search results quite significantly. Not only is something pushed to the top, but that may have an impact on metrics, and we need to consider how to ensure our clients or brands appear in the snippets.

There are also going to be huge advancements on the tooling side. Running your own LLM is not rocket science anymore. Google’s Vertex and some others have created relatively simple, ready-made solutions already. With a bit more time, a lot of the tool providers will move towards that direction.”

How are LLMs likely to change and how can SEOs prepare for those changes?

“Sadly, there’s no one answer to that. Most LLMs are transformer-based right now. That’s the architecture and the underlying foundational technology. They are great for what we’re currently using them for, but they also have limitations.

For example, to input more knowledge into an LLM, you need more training data. That means the model will grow in size, which also means that you need more hardware. If you query a transformer-based LLM currently, from a tech standpoint, you need to go through the entire model. That obviously doesn’t scale endlessly. It scales to the extent that you can throw more and more hardware at the problem (you’ve probably seen the insane numbers that NVIDIA released recently and the stock market’s reaction to it), however, it can’t scale endlessly.

Therefore, we will probably start seeing different approaches, and some projects are currently researching different directions that could be taken. From an implementation standpoint, as an organisation, you need to build some middleware.

You might currently be in the position to use an LLM for your SEO, which is fantastic. You can take all your knowledge and your content, put it in an LLM and then run certain types of models on your own. However, the LLM model itself might change. If you build a middleware API layout then, moving forward, it will be much easier to adapt to a new model. You can just move over the data and then instantly benefit from the work that you have already done.

We know that the models might change. I can’t say how this might look in practice (I’m coming from a practitioner’s standpoint, not a researcher’s) but this would be a logical approach to take. We’re seeing a lot of large organisations do this because they want to reap the benefits right now. There are fantastic gains to be had. You can make customer service more efficient and share knowledge internally more efficiently if you can collect all your knowledge from different sources and put it into these types of models.

Don’t just sit on it and wait two years. Taking a hybrid approach, and being prepared to shift over, is a much better way to do it.”

How should SEOs use LLMs to enhance their knowledge and stay up-to-date?

“The big difference is prompting and how you actually interact with the LLMs. This is going to change too because we’re moving towards a more multimodal world, where we will not only have the ability to prompt using text but also combine different types of inputs. BARD now allows you to upload a photo and ask, ‘What’s in this image?’, and it will give you a description of the photo. It’s like a combination of Lens and BARD. We’re also seeing multimodality with things like gestures.

You need to understand how you can prompt the model to ensure that you get what you want out of it, keeping in mind that there are certain limitations such as hallucination. You might use a plugin to bring in external data or do certain calculations. Even if you’re not using it in practice, I would strongly recommend at least having a play.

Also, don’t forget about the code interpreter in ChatGPT Plus. As SEOs, we do a ton of analysis, looking at numbers and trying to figure out what’s happening. From an agency perspective, if you look at a new website that you’re not familiar with and you’re seeing ups and downs and dips, you can overlay that with different types of data sources. You can ask a model like this to generate ideas about what might have happened. That is a really cool way to use the LLM as a research assistant before you do it yourself to double-check. Again, it’s about the efficiency that you gain from utilising it.

Experimenting and having a play will help you to understand what is happening, even though it’s hard to say that you understand how AI really works. Even engineers don’t understand it completely, in a way, because the machines are reiterating and learning by themselves. However, you can understand how to utilise it, play with it, and figure it out.

That has always been part of SEO. When I started out, we learnt by tinkering and playing, and this has never really changed. If you’re not naturally curious about things, then you will have a very hard time. That is going to be even more true in the future.”

If an SEO is struggling for time, what should they stop doing right now so they can spend more time doing what you suggest in 2024?

“Stop worrying so much about all the possible ranking factors. People spend so much time trying to figure out isolated, tiny things. Changing a title here or a meta description there is not going to make a huge difference. It might make a difference for that page specifically but, in the grand scheme of things, it’s minor.

Move away from the small nitty-gritty stuff. Don’t buy cheap links and don’t buy cheap content. That’s a massive waste of time and money. Instead, zoom out. Try to understand what Google actually wants to rank, because that’s a big question for the future.

This is going to be a very hard realisation for some but, if you look at the current Google SGE implementation, it’s becoming much harder to acquire certain types of organic traffic. In the past, a lot of people relied on churning out generic content, but that’s essentially common knowledge. There’s no reason for Google to rank an article that is just a rephrasing of something that has already been put out there 100,000 times. They will just do that themselves and keep the traffic.

Moving forward, you need to figure out how to become more of an expert and much more well-versed in a certain topic. Just saying, ‘Here’s how to do XYZ’, is not going to cut it in 2024.”

Bastian Grimm is CEO at Peak Ace, and you can find him over at PeakAce.Agency.

@basgr    

Also with Bastian Grimm

Bastian Grimm 2025 podcast cover with logo
SEO in 2025
Manage your crawl budget effectively

Bastian Grimm from Peak Ace AG has noticed a significant but overlooked issue arising from the proliferation of AI-generated content: Google struggling to find the budget to crawl and index your site.

Bastian Grimm 2023 podcast cover with logo
SEO in 2023
Get your head around AI

Bastian Grimm informs SEOs that you can no longer be ignorant about the way AI and machine learning are impacting the online environment in 2023. You need to start understanding now if you want to stay in front of the pack.

Bastian Grimm 2022 podcast cover with logo
SEO in 2022
Are you prepared for Platform-less, server-less SEO on the edge?

Bastian's tip follows on from Nick's, highlighting that Platform-less, server-less SEO has the potential to make several SEO operations much more efficient.

Choose Your Own Learning Style

Webinar iconVideo

If you like to get up-close with your favourite SEO experts, these one-to-one interviews might just be for you.

Watch all of our episodes, FREE, on our dedicated SEO in 2024 playlist.

youtube Playlist Icon

Podcast iconPodcast

Maybe you are more of a listener than a watcher, or prefer to learn while you commute.

SEO in 2024 is available now via all the usual podcast platforms

Spotify Apple Podcasts Audible

Book iconBook

This is our favourite. Sometimes it's better to sit and relax with a nice book.

The best of our range of interviews is available right now as a physical copy and eBook.

Amazon US Amazon UK

Don't miss out

Opt-in to receive email updates.

It's the fastest way to find out more about SEO in 2025.


当社ではページの改善に取り組んでいます。皆様のご意見をお寄せください

フレッシュインデックス

クロールした固有 URL 164,777,211,282
検出した固有 URL 640,941,217,070
日付範囲 17 Jul 2025 ~ 14 Nov 2025
最終更新日 1 時間 45 分前

ヒストリックインデックス

クロールした固有 URL 4,502,566,935,407
検出した固有 URL 21,743,308,221,308
日付範囲 06 Jun 2006 ~ 26 Mar 2024
最終更新日 03 May 2024

ソーシャル

  • LinkedIn
  • YouTube
  • Facebook
  • Bluesky
  • Twitter
  • ブログ 外部リンク

会社

  • Flow Metric スコア
  • 会社案内
  • 利用規約
  • プライバシーポリシー
  • GDPR
  • お問い合わせ

ツール

  • プランと料金
  • サイトエクスプローラ
  • ドメインを比較
  • 一括被リンク
  • 検索エクスプローラ
  • デベロッパー API 外部リンク

...用 MAJESTIC

  • Link Context
  • 被リンクチェッカー
  • SEO プロ用
  • メディアアナリスト向け
  • インフルエンサーの発見
  • エンタープライズ 外部リンク

PODCASTS & PUBLICATIONS

  • The Majestic SEO Podcast
  • SEO in 2025
  • 2024年のSEO
  • 2023年のSEO
  • 2022年のSEO
  • All Podcasts
top ^