Screaming Frog SEO Spider Update – Version 4.0

Dan Sharp

Posted 7 July, 2015 by in Screaming Frog SEO Spider

Screaming Frog SEO Spider Update – Version 4.0

I’m really pleased to announce version 4.0 of the Screaming Frog SEO Spider, codenamed internally as ‘Ella’.

We have been busy in development working on some significant improvements to the SEO Spider due for release later this year, which include a number of powerful new features we’ve wanted to release for a very longtime. Rather than wait, we decided to release some of these features now, with much more on the way.

Therefore, version 4.0 has two new big features in the release, and here’s the full details –

1) Google Analytics Integration

You can now connect to the Google Analytics API and pull in data directly during a crawl.

Google Analytics Integration

To get a better understanding of a website’s organic performance, it’s often useful to map on-page elements with user data and SEOs have for a long-time combined crawl data with Google Analytics in Excel, particularly for Panda and content audits. GA data is seamlessly fetched and matched to URLs in real time as you crawl, so you often see data start appearing immediately, which we hope makes the process more efficient.

The SEO Spider not only fetches user and session metrics, but it can also collect goal conversions and ecommerce (transactions and revenue) data for landing pages, so you can view your top performing pages when performing a technical or content audit.

If you’re running an Adwords campaign, you can also pull in impressions, clicks, cost and conversion data and we will match your destination URLs against the site crawl, too. You can also collect other metrics of interest, such as Adsense data (Ad impressions, clicks revenue etc), site speed or social activity and interactions.

To set this up, start the SEO Spider and go to ‘Configuration > API Access > Google Analytics’.

google analytics configuration

Then you just need to connect to a Google account (which has access to the Analytics account you wish to query) by granting the ‘Screaming Frog SEO Spider’ app permission to access your account to retreive the data. Google APIs use the OAuth 2.0 protocol for authentication and authorisation.

Google Analytics Set-up

Once you have connected, you can choose the relevant Analytics account, property, view, segment and date range!

Google analytics view

Then simply select the metrics that you wish to fetch. The SEO Spider currently allow you to select up to 20, which we might extend further. If you keep the number of metrics to 10 or below with a single dimension (as a rough guide), then it will generally be a single API query per 10k URLs, which makes it super quick –

Google analytics metrics

You can also set the dimension of each individual metric, as you may wish to collect data against page path and, or landing page for example.

Google analytics Dimensions

There are circumstances where URLs in Google Analytics might not match URLs in a crawl, so we have a couple of common scenarios covered in our configuration, such as matching trailing and non-trailing slash URLs and case sensitivity (upper and lowercase characters in URLs).

If you have millions of URLs in GA, you can also choose to limit the number of URLs to query, which is by default ordered by sessions to return the top performing page data.

Google analytics settings

When you hit ‘start’ to crawl, the Google Analytics data will then be fetched and display in respective columns within the ‘Internal’ and ‘Analytics’ tabs. There’s a separate ‘Analytics’ progress bar in the top right and when this has reached 100%, crawl data will start appearing against URLs. Fetching the data from the API is independent of the crawl, and it doesn’t slow down crawl speed itself.

Google Analytics data in full

There are 3 filters currently under the ‘Analytics’ tab, which allow you to filter by ‘Sessions Above 0’, ‘Bounce Rate Above 70%’ and ‘No GA Data’. ‘No GA Data’ means that for the metrics and dimensions queried, the Google API didn’t return any data for the URLs in the crawl. So the URLs either didn’t receive any visits (sorry, ‘sessions’), or perhaps the URLs in the crawl are just different to those in GA for some reason.

No GA Data

For our site, we can see there is ‘no GA data’ for blog category pages and a few old blog posts, as you would expect really (the query was landing page, rather than page). Remember, you may see pages appear here which are ‘noindex’ or ‘canonicalised’, unless you have ‘respect noindex‘ and ‘respect canonicals‘ ticked in the advanced configuration tab.

Please note – If GA data does not get pulled into the SEO Spider as you expected, then analyse the URLs in GA under ‘Behaviour > Site Content > All Pages’ and ‘Behaviour > Site Content > Landing Pages’ depending on your query.

If they don’t match URLs in the crawl, then GA data won’t able to be matched up and appear in the SEO Spider. We recommend checking your default Google Analytics view settings (such as ‘default page’) and filters which all impact how URLs are displayed and hence matched against a crawl. If you want URLs to match up, you can often make the required amends within Google Analytics.

This is just our first iteration and we have some more advanced crawling, matching, canonicalisation and aggregation planned which will help in more complicated scenarios and provide further insights.


2) Custom Extraction

The new ‘custom extraction’ feature allows you to collect any data from the HTML of a URL. As many of you will know, our original intention was always to extend the existing ‘custom search’ feature, into ‘custom extraction’, which has been one of the most popular requests we have received.

If you’re familiar with scraping using import XML and Xpath, SeoTools for Excel, Scraper for Chrome or our friends at URL Profiler, then you’ll be at home using Xpath & CSS Path selectors and using the custom extraction feature.

You’ll find the new feature under ‘Configuration > Custom’.

custom extraction config

‘Search’ is of course the usual custom source code search feature you should be familiar with –

custom search

‘Extraction’ is similar to the custom search feature, you have 10 fields which allow you to extract anything from the HTML of a web page by using either Xpath, CSS Path or failing those, regex. You can include the attribute as usual in XPath and for CSS Path an optional attribute field will appear after selection.

custom extraction

When using XPath or CSS Path to collect HTML, you can choose what to extract:

  1. Extract HTML Element: The selected element in full and the HTML content inside.
  2. Extract Inner HTML: The HTML content inside of the selected element. If the selected element contains other HTML elements, they will be included.
  3. Extract Text: The text content of the selected element, and the text content of any sub elements (essentially the HTML stripped entirely!).

You’ll get a lovely tick next to your regex, Xpath or CSS Path if it’s valid syntax. If you’ve made a mistake, a red cross will remain in place!

This will allow you to collect any data that we don’t currently as default or anything unique and specific to the work you’re performing. For example, Google Analytics IDs, schema, social meta tags (Open Graph Tags & Twitter cards), mobile annotations, hreflang values, as well as simple things like price of products, discount rates, stock availability etc. Let’s have a look at collecting some of these, with specific examples below.

Authors & Comments

Lets say I wanted to know the authors of every blog post on the site and the number of comments each have received. All I need to do is open up a blog post in Chrome, right click and ‘inspect element’ on the data I want to collect and right click again on the relevant HTML and copy the relevant CSS path or XPath. If you use Firebug in Firefox, then you can do the same there, too.

author CSS Path

I can also name the ‘extractors’, which correspond to the column names in the SEO Spider. In this example, I’ve used a CSS Path and XPath for the fun of it.

author and comments extraction

The author names and number of comments extracted then shows up under the ‘extraction’ filter in the ‘custom’ tab, as well as the ‘internal’ tab allowing you to export everything collected all together into Excel.

author & comments extracted

Google Analytics ID

Traditionally the custom search feature has been really useful to ensure tracking tags are present on a page, but perhaps sometimes you may wish to pull the specific UA ID. Let’s use regex for this one, it would be –


Google Analytics ID Extraction

And the data extracted is as you would expect –


Additional Headings

As default, the SEO Spider only collects h1s and h2s. However, perhaps you would like to collect h3s as well. Regex should generally be your last resort, for collecting items which CSS Path or XPath can’t – such as HTML comments or inline JavaScript . However, as an example, a quick regex for this one might be –


h3 custom extraction

The Xpath would just be –


Alongside the ‘Extract Text’ option. The first h3s returned are as follows –

h3 extracted

Mobile Annotations

If you wanted to pull mobile annotations from a website, you might use an Xpath such as –

//link[contains(@media, '640') and @href]/@href

mobile annotations extraction

Which for the Huffington Post would return –

mobile annotations extracted


We are working on a more robust report for hreflang as there can be so many types of errors and problems with the set-up, but in the meantime you can collect them using the new custom extraction feature as well. You might need to know how many hreflang there are on a page first –



hreflang extraction

The above will collect the entire HTML element, with the link and hreflang value.

hreflang extracted

So, perhaps you wanted just the hreflang values, you could specify the attribute using @hreflang –

hreflang custom extraction of values

This would just collect the language values –

hreflang language values extracted

Social Meta Tags

You may wish to extract social meta tags, such as Facebook Open Graph tags or Twitter Cards. Your set-up might be something like –

//meta[starts-with(@property, 'og:title')][1]/@content
//meta[starts-with(@property, 'og:description')][1]/@content
//meta[starts-with(@property, 'og:type')][1]/@content


social meta tags extraction

Which on Moz, would collect some lovely social meta tag data –

social meta tags extracted


You may wish to collect the types of various Schema on a page, so the set-up might be –



schema extraction

And the extracted data is –

schema extracted

Email Addresses

Perhaps you wanted to collect email addresses from your website or websites, the Xpath might be something like –

//a[starts-with(@href, 'mailto')][1]
//a[starts-with(@href, 'mailto')][2]


email address extraction

From our website, this would return the two email addresses we have in the footer on every page –

email address extracted

That’s enough examples for now, and I am sure you will all have plenty of other smart ways this feature can be used. My Xpath is probably fairly poor, so feel free to improve! You can read our web scraping guide with more examples.

To help us avoid a sudden influx of queries about Xpath, CSS Path and regex syntax, please do read some of the really useful guides out there on each, before submitting a support query. If you have any guides you’d like to share in the comments, they are more than welcome too!

Other Updates

Some of you may notice there is an ‘address’ and ‘URL encoded Address’ field within the ‘URL Info’ tab in the lower window pane. This is because we have carried out a lot of internal research around URL encoding and how Google crawl and index them. While this feature isn’t as exiciting as those above, it’s a fairly significant improvement.

When loading in crawl files saved from previous releases the new ‘URL Encoded Address’ and ‘Address’ are not updated to the new behaviour implemented in version 4.0. If you want this information to be 100% accurate, you will need to re-crawl the site.

We have also performed other bug fixes and smaller updates in version 4.0 of the Screaming Frog SEO Spider, which include the following –

  • Improved performance for users using large regex’s in the custom filter & fixed a bug not being able to resume crawls with these quickly.
  • Fixed an issue reported by Kev Strong, where the SEO Spider was unable to crawl urls with an underscore in the hostname.
  • Fixed X-Robots-Tags header to be case insensitive, as reported by Merlinox.
  • Fixed a URL encoding bug.
  • Fixed an bug where the SEO Spider didn’t recognise text/javascript as JavaScript.
  • Fixed a bug with displaying HTML content length as string length, rather than length in bytes.
  • Fixed a bug where manual entry in list mode doesn’t work if a file upload has happened previously.
  • Fixed a crash when opening the SEO Spider in SERP mode and hovering over bar graph which should then display a tooltip.

Small Update – Version 4.1 Released 16th July 2015

We have just released a small update to version 4.1 of the Screaming Frog SEO Spider. There’s a new ‘GA Not Matched’ report in this release, as well as some bug fixes. This release includes the following –

GA Not Matched Report

We have released a new ‘GA Not Matched’ report, which you can find under the ‘reports’ menu in the top level navigation.

Google Analytics Not Matched Report

Data within this report is only available when you’ve connected to the Google Analytics API and collected data for a crawl. It essentially provides a list of all URLs collected from the GA API, that were not matched against the URLs discovered within the crawl.

This report can include anything that GA returns, such as pages in a shopping cart, or logged in areas. Hence, often the most useful data for SEOs is returned by querying the landing page path dimension and ‘organic traffic’ segment. This can then help identify –

  1. Orphan Pages – These are pages that are not linked to internally on the website, but do exist. These might just be old pages, those missed in an old site migration or pages just found externally (via external links, or referring sites). This report allows you to browse through the list and see which are relevant and potentially upload via list mode.
  2. Errors – The report can include 404 errors, which sometimes include the referring website within the URL as well (you will need the ‘all traffic’ segment for these). This can be useful for chasing up websites to correct external links, or just 301 redirecting the URL which errors, to the correct page! This report can also include URLs which might be canonicalised or blocked by robots.txt, but are actually still indexed and delivering some traffic.
  3. GA URL Matching Problems – If data isn’t matching against URLs in a crawl, you can check to see what URLs are being returned via the GA API. This might highlight any issues with the particular Google Analytics view, such as filters on URLs, such as ‘extended URL’ hacks etc. For the SEO Spider to return data against URLs in the crawl, the URLs need to match up. So changing to a ‘raw’ GA view, which hasn’t been touched in anyway, might help.

Other bug fixes in this release include the following –

  • Fixed a couple of crashes in the custom extraction feature.
  • Fixed an issue where GA requests weren’t going through the configured proxy.
  • Fixed a bug with URL length, which was being incorrectly reported.
  • We changed the GA range to be 30 days back from yesterday to match GA by default.

I believe that’s everything for now and please do let us know if you have any problems or spot any bugs via our support. Thanks again to everyone for all their support and feedback as usual.

Now go and download version 4.1 of the SEO Spider!

Dan Sharp is founder & Director of Screaming Frog. He has developed search strategies for a variety of clients from international brands to small and medium-sized businesses and designed and managed the build of the innovative SEO Spider software.


  • Stephen 9 years ago

    With these 2 updates you guys rock the world. Both are awesome features, and must have for any seo tool.

  • Great tool, thanks for keeping up with new features.

  • Luke Fitzgerald 9 years ago

    Love the Analytics functionality guys, thanks for this great update.

    Was just wondering, where’s the ‘too short’ option gone from the H1 and h2 dropdown fields? That’d be nice to have back again!

    • screamingfrog 9 years ago

      Hey Luke,

      Good to hear it’s helping!

      We actually removed those filters back in November 2012. So maybe you were using a really old version?

      We felt they were not particularly useful, as headings can be pretty short!

      Cheers for the feedback!


  • Luke Fitzgerald 9 years ago

    Cheers for clarifying Dan,

    My bad, I was actually looking for same on Page Title tab, not H1, and they’re in there alright! Just a bit lost in the new nav, great update nonetheless!



  • Ross Allen 9 years ago

    Great update! I’ve already lost an hour just exploring the new Analytics functionality!

  • Frederik Nørgaard 9 years ago

    Nice that you bundle the updates so you dont have to update every week.

    Looking forward to testing the GA funktion.

  • Pozycjonowanie 9 years ago

    I was waiting for this integration. Good Job!!!

  • Jon 9 years ago

    I’ve just bought a license as I’m trying to extract all anchor elements that have an ID that starts with ‘Interstitial’ but so-far having no joy. I’m using Configuration > Custom > Extraction. I’m using this Xpath : //a[starts-with(@id, ‘Interstitial’)] but it only seems to get one hit per-page – I need it to get all of them as there is lots – can you help with this?

  • Great news :) Now all the necessary data will be in one tool :)

  • Stephen Watts 9 years ago

    Hi Dan, awesome to include the extract feature, this is truly great and saves a great deal of time switching between Screaming Frog & SEO Tools for Excel.

    Question — for a regex match like <– Screaming Frog is pulling in multiple lines when I'd expect it to stop at the highlighted query. Any ideas?

  • Andrew 9 years ago

    What a huge update! Really excited to use the new custom extract feature, so many possible uses. There were a handful of times I needed a specific item that wasn’t available in the normal crawl.

  • Jack Norell 9 years ago

    Superb, this was the update I needed to stop messing about with 80 Legs for most of my crawls!

  • Officina Seo 9 years ago

    Any update increase my statisfaction, I discovered you some years ago!

  • Andrea Moro 9 years ago

    Interesting update, though I can see some usabilities problems on the ga use, especially when it covers to the date selection.

    That would be more useful to be asked prior a crawl start rather than being confined into a relatively hidden configuration tab.

    Hope it makes sense.

    • screamingfrog 9 years ago

      Hi Andrea,

      We had an internal debate whether to include ‘date’ within the initial account tab.

      However, the argument was that this is very much an ‘account’ tab and ‘date’ in theory has no place there. It was a difficult one, we had it the other way around to start with!

      Having a separate tab for ‘date’ (which I’d argue is pretty clear to see still) also allows us to add additional presets in time to cover different date ranges etc as well.

      Cheers for the feedback.


  • Wrel 9 years ago

    Wow. thanks for the GA Integration.

  • John 9 years ago

    Holy Sh*t YES YES YES xpath scraping has been the one missing feature on this tool!

    Amazing work!

  • Alex 9 years ago

    What an awesome job you guys do!!!
    I can’t nearly imagine any SEO person without Screaming Frog now!

  • Michael Field 9 years ago

    This is some of the best news I’ve had all year. Goodbye vlookup, hello GA Integration!

  • Watt 9 years ago

    Wawww it’s a great update ! Good job

  • Daniel Herndler 9 years ago

    Very well done! I’m already testing the new features for my tasks and projects and it really makes work easier. ;)

    Thanks and all the best!

  • Tom 9 years ago

    Great new features guys, Christmas come early! Really looking forward to trying some of these out :)

  • Kristian Mølholt 9 years ago

    Very nice update – thanks guys.
    I just love the screaming frog.
    It makes my job soooo much easier.
    Thanks for Analytics and xPath additions.

  • Gcod 9 years ago

    Finally! Scraping! Thanks!

  • Tamas Varga 9 years ago

    These are really powerful updates, turning points in in how to integrate and analize on-page factors within one tool. Congrats!

    As we may use the Analytics feature addictively it would be great to have a timestamp option (analytics_sessions_above_0_today’s_date) when exporting, so that you do not accidentally overwrite previous exports. Thanks

  • Dimitar 9 years ago

    I wasn’t using this tool for a long time, but with these new features i’m gonna give him a second chance. It seems a lot more useful now. Good job guys :)

  • SEO Plus 9 years ago

    holy you-know-what!

    really awesome update – can’t wait to see what comes from this!

    hope to see some video training sessions on how people are using this.

    Wish there was a Master Class on this somewhere – please let us know if you start to make or see them.

    There’s something new to learn every day!

    keep up the really awesome work!

  • Walk to Remember 9 years ago

    Fantastic! I’m using these new features as I type this. Great work.

  • Michal Krutiš 9 years ago

    Hi. Both features are perfect. I will use them daily.
    I had to for search using Open/Google Refine and pairing pages with GA data via Excel.
    Is possible to make feature that will find hidden content – cloaking?

  • Alex Hernandez 9 years ago

    I can’t get the schema or open graph custom extractors to work.

    • Alex Hernandez 9 years ago

      I got it!

  • Niall Flynn 9 years ago

    These look beast! Good Work SF, I tip my hat to you all.

  • Alex Hernandez 9 years ago

    I’m not the greatest with xpath, is there any way I could pull a whole group of specific schema elements, kind of like how you pull the OG and Twitter elements?

  • Phillip 9 years ago

    excellent- thanks for the update!

  • Fabrizio ballarini 9 years ago

    Any tips on how to extract multiple instances of the same code using regex? I can extract the first match but not the multiple matches on the same page.

  • Quentin 9 years ago

    Best update of the year, and for sure !
    Orphan Pages identification is a must !
    Add your http log files to this and your are the king of the world :)

  • sawtnass 9 years ago

    Fantastic! I’m using these new features as I type this. Great work

  • Kingbo 9 years ago

    Thanks for this great tool, thanks for keeping up with new features.

  • Gabe 9 years ago

    Great new features, but it would really be useful if you could add this scraping feature to XML sitemaps as well. For example, I’d like to be able to extract hreflang and mobile annotations for each URL in an XML sitemap and as far as I can tell, this is currently not possible with the current version of your software.

    • screamingfrog 9 years ago

      Hi Gabe,

      We plan on developing these as proper features, rather than developing the ability to scrape an XML sitemap.



      • Szymon Szperna 9 years ago

        Were you planning to refresh the graphic design of your program in the future?
        I think your application is excellent – very intuitive and useful but the interface is a little bit archaic :)

        • screamingfrog 9 years ago

          Heheheh! Thanks Szymon.

          A face only a mother could love ;-) I won’t argue with that, but we’ve heard that’s also kinda part of the Spider’s appeal!

          We may update in the future though, you can switch to a Windows look and feel if you’d prefer (under Configuration > User Interface).


  • Roderick 9 years ago

    Just had a quick look but it looks great again. Love the Analytics integration.

  • Love the Analytics functionality guys, thanks for this great update. I’ve already lost an hour just exploring the new Analytics functionality! The best tool.

  • Carlos 9 years ago

    We are waited for this update for a long time. Our customers will certainly appreciate it – reports will be more valuable :)

  • Nadeer 9 years ago


    Really great addition.

    Can you please tell me to what value Sessions corresponds to in GA?

    I mean, I’m not seeing the same numbers for GA sessions in SF vs GA.
    Does that correspond to Entrance, Unique Pageviews?

    Thatt’s what I see if I look at Landing pages


    • screamingfrog 9 years ago

      Hi Nadeer,


      It looks like you’re looking under ‘behaviour > site content > all pages’ in GA, when the default dimension in the SEO Spider is for landing pages, which you can see under ‘behaviour > site content > landing pages’. For ‘pages’ entrances, is the same thing as LP sessions.

      There are some reasons they might not match 100% (and are actually more accurate than in GA) discussed here –



      • Nadeer 9 years ago

        Hey Dan,

        Oh….I see, I should have noticed that ahah, now it makes more sense, thanks!

  • Jeremy 9 years ago

    Very good update! I am already 2 hours in trying to comprehend it all!

  • vseo 9 years ago

    Due to a custom GTM config, my url on analytics are slightly different than the URL on screaming frog.
    So, no matches.
    I have tried to with URL rewriting but no way… any suggestion?

  • Lasse 9 years ago

    Very nice update. Great work!

  • Guillermo 9 years ago

    Hi Dan,

    Very cool GA integration.
    This update add lot of value at the crawling process
    You have become an indispensable tool

    Many thx ;) from Spain

  • Per Pettersson 9 years ago

    Seriously good updates, guys. Seriously good.

  • Benoît Quimper 9 years ago

    Great to see an integration with Search Console! It would be great to have the top 10 keywords per URL like in URL Profiler. ;-)

    • screamingfrog 9 years ago

      Thanks Benoît!

      Think you left this on the wrong post (should have been 5.0?), but appreciate the suggestion.



  • Raul 9 years ago

    Hi Screaming Frog Team,
    I really like this new version , thanks.
    Do you know if we would be able to extract the Twitter user name from a list of url ? What would be the regex ?
    We indeed need to get the Twitter names, in addition to emails, from a list of urls.
    For instance, let’s say that on this url,, I want to extract the Twitter account that is published on it. How could I do that ?
    Thanks a lot for help

    • screamingfrog 9 years ago

      Hi Raul,

      A regex like this should work –*?)[‘”]


      Depending on what you want to collect! ;-)



      • Raul 9 years ago

        Many thanks for the answer. I tried with no success on this page:

        You will see at teh bottome of the page twitter name

        <a href="……

        Do you think we can extract ?

        Thanks again

  • Raul 9 years ago

    Thanks for the advice.
    I tried the code with no success, end the field Extractor1 remains empty. Maybe you should try with the url I sent you.

  • Luciano Bellacci 9 years ago

    Great tool. Thanks guys for keeping up the good work :)

  • Pozycjonowanie 8 years ago

    I was waiting for this integration. Great tool!!!

  • John 8 years ago

    Superb. Good update, guys. I love your tool. Use it every day :)

  • Victor 8 years ago

    Tengo una pregunta,

    Como podría extraer la Url + etiqueta titulo + Alt de una imagen “”

    Utilizo Screamingflog –> Custom

    y me gustaría poder sacar esa información utilizando Xpath o Csspath …
    seria de mucha ayuda!

  • wp 8 years ago

    Hervorragend. Gute Update , Jungs. Ich liebe dein Werkzeug . Verwenden Sie es jeden Tag :)

  • Thomas intoto 8 years ago

    You do indeed live up to the name Screaming Frog. Well done!
    Also nice to have a tool that support Mac.

  • Mera Status 8 years ago

    The tools works perfectly for small and medium size sites. I’ve been able to get all urls within hours.

    But the only issue i face is with large sites with atleast, 100k total links.

    Though i’ve extended the ram till 4 Gb out of 8 Gb (i’ve had 4 gb ram before) but still i face hanging problem in the system due to the tool.

    4 gb takes the tools and rest processes face hard time to process like excel or anything.

    You guys working towards this ?

    • screamingfrog 8 years ago

      Hey Mera,

      Thanks for the comment. If you have 8GB of RAM in a 64bit machine, I recommend allocating it all –

      You should then be able to crawl fairly sizeable websites. The amount of memory required does depend on the website, but it should be good for 100k in most situations!

      I have about 8GB allocated in a desktop at home, and have crawled a lot more.

      We are working towards more scalable solutions as well anyway.



      • Mera Status 8 years ago

        Thanks Dan for the prompt reply.

        Will do the trick!

        I wish you guys all best, your tool is the best SEO tool so far (My all time favorite too).


  • Frank Frederiksen 8 years ago

    Great update, looking forward to it.

  • We use this tool on daily basis for our clients. Didnt even knew that it was updated, but hey. Thanks

  • Where can i report a bug? We use this weekly, but recently i had a bug everytime i used the tool. The same bug.

    Still at great tool.

  • m88 indonesia 7 years ago

    Why do I download urlprofiler on and my antivirus software says there is a virus. Is it safe?

  • Will Richardson 6 years ago

    We’re migrating an old GA (hardcoded) set up to a GTM verison – we’re seeing some discrepancy as you;d expect, but its a bit too far. We want to verify that Tag Manager is across all the page / same pages as the old set up to rule out issues. How would I do the same Search / Extraction funciton for the UA code but for our GTM account?


Leave A Comment.

Back to top