10

Me: Hey SEO guy. I am updating our online store from Flask/jQuery in ReactJS.

SEO guy: That is amazing. Google LOVES ReactJS and it will crawl the site very fast.

*fast forward*

SEO guy: Hey, did you change anything in the site because the site is not ranking anymore on Google. The URLs are dynamically generated in front end. Google does not like that.

ME: But you said that Google loves React. It took me nearly 1 month to migrate the code in React.

Fucking hell.

Comments
  • 5
    As long as you can link direct to the pages it should work with google

    But if the only way to view a page is through actively clicking, yes that will be a problem.

    Links need to be links for google to index them.
  • 1
    @Voxera with SPA frameworks (unless you've using server-side rendering) the client is only getting an empty template HTML without any content, which is then filled by the framework in JS. No content means no links either and crawlers don't execute JS. You can add a sitemap, but all of these links will only contain that empty template. SPAs are useless for bots unless they use SSR.
  • 1
    @24th-Dragon I was looking into that. And also GatsbyJS. Next.js might take some time until I get familiar with documentation.

    I'll get back to Flask with AngularJS. Fuck me.
  • 0
    @halfflat It has its advantages. SPAs feel more responsive, load once and then only fetch data, they can offer richer, more user-friendly UI. I believe with SSR one could solve the problem of heavy reliance on JS completely, I'm not sure if such implementations exist though.
  • 0
    @24th-Dragon That sounds cool!
  • 0
    @24th-Dragon Wait a minute. Vue is server-sided?
  • 0
    @ColdFore It isn't.
  • 0
    @gronostaj Google executes JavaScript for it's index, because of the many SPAs. However having a site which effectively only works with JS is probably still not the best variant.
  • 2
    @gronostaj You are factually wrong though

    https://developers.google.com/searc...

    > Googlebot queues all pages for rendering, unless a robots meta tag or header tells Googlebot not to index the page. The page may stay on this queue for a few seconds, but it can take longer than that. Once Googlebot's resources allow, a headless Chromium renders the page and executes the JavaScript. Googlebot parses the rendered HTML for links again and queues the URLs it finds for crawling. Googlebot also uses the rendered HTML to index the page.

    Googles bot will open a page, find all the links on that page, open those pages, repeat until it has found all the pages, then it will render the found pages with javascript, find all the links on that page and repeat the first process.
  • 0
    @gronostaj google today actually runs js on pages they crawl.

    So if you use react router and have backend return the page for all relevant paths google still should be able to crawl it.

    At least that is what I read in a paper they released a couple of years ago.

    So you should not need server side rendering.

    https://google.se/amp/s/...

    https://google.se/amp/s/...
  • 0
    Have you looked into nfl-helmet? It creates the necessary title and meta description tags for pages to be crawled in React.
Add Comment