Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
SEO effect of URL with subfolder versus parameters?
-
I'll make this quick and simple. Let's say you have a business located in several cities. You've built individual pages for each city (linked to from a master list of your locations).
For SEO purposes is it better to have the URL be a subfolder, or a parameter off of the home page URL:
https://www.mysite.com/dallas which is essentially https://www.mysite.com/dallas/index.php
or
http://www.mysite.com/?city=dallas which is essentially https://www.mysite.com/index.php?city=dallas
-
Thanks Miriam, This is very helpful and makes a lot of sense. What do you think of towns and villages, or boroughs of a large city. Do you think the close proximity is dangerous territory re: keyword permutations?
I take your point about unique content tailored to the people of the city - it makes a lot of sense. But what about locations that are closer to each other?
I know it's a tricky question but any insight would be most welcome.
-
That's a good question, Andrew. It's true that it's no longer a best practice to build out a set of pages featuring slightly different permutations of a keyword (car repair, auto repair, repairing cars, fixing cars, etc.). That approach is now quite dated. Honestly, it never made any sense beyond the fact that when Google wasn't quite so sophisticated, you could trick your way into some additional rankings with this type of redundant content.
The development of location landing pages is different. These are of fundamental use to consumers, and the ideal is to create each city's landing page in a way that is uniquely helpful to a specific audience. So, for example, your store in Detroit is now having a special on winter clothing right now, because it's still snowing there. Meanwhile, your store in Palm Beach is already stocking swim trunks. For a large, multi-location Enterprise, location landing pages can feature highly differentiated content, including highlights of regional-appropriate inventory and specials, as well as unique NAP, driving directions, reviews from local customers, and so much more.
The key to avoiding the trap of simply publishing a large quantity of near-duplicate pages is to put in the effort to research the communities involved and customize these location pages to best fit local needs.
-
Hi Searchout,
Good for you for creating a unique page for each of your locations. I like to keep URLs as simple as possible, for users, so I'd go with:
etc.
From an SEO perspective, I don't think there's a big difference between root URLs and subfolders. If you're using one structure, I doubt you'd see any difference from doing it differently (unless you were using subdomains, which is a different conversation).
-
Of course that cities will be counted.
That´s why im always reinforcing the idea of creating UNIQUE and Special pages for each keyword.
Google is getting smarter and smarter, so simple variations in a few words are easly detected.Hope it helps.
Best luck.
GR. -
Hi
Thanks for your response I'm interested in this too. I've been targeting cities with their own pages but I head recently that google are going to be clamping down on multiple keyword permutations. Do you think cities will be counted in this?
-
Hi there!
In my opinion, for SEO purposes it is correct to have a unique page (really different from other, not just changing the city name and location) por each big city you are optimizing.
Thus said, a subfolder is useful in order to show google the name of the city in the URL. It is common that google considers parameters different than folders.Also, remember to avoid duplicate content. /dallas/ and /dallas/index.php should not be accesible and indexable for google. Redirect one to the other or canonicalize one to the other. Same with www, non-www, http and https versions.
Hope it helps.
Best luck.
GR.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How does Infinite Scrolling work with unique URLS as users scroll down? And is this SEO friendly?
I was on a site today and as i scrolled down and viewed the other posts that were below the top one i read, i noticed that each post below the top one had its own unique URL. I have not seen this and was curious if this method of infinite scrolling is SEO friendly. Will Google's spiders scroll down and index these posts below the top one and index them? The URLs of these lower posts by the way were the same URLs that would be seen if i clicked on each of these posts. Looking at Google's preferred method for Infinite scrolling they recommend something different - https://webmasters.googleblog.com/2014/02/infinite-scroll-search-friendly.html . Welcome all insight. Thanks! Christian
Intermediate & Advanced SEO | | Sundance_Kidd0 -
Attack of the dummy urls -- what to do?
It occurs to me that a malicious program could set up thousands of links to dummy pages on a website: www.mysite.com/dynamicpage/dummy123 www.mysite.com/dynamicpage/dummy456 etc.. How is this normally handled? Does a developer have to look at all the parameters to see if they are valid and if not, automatically create a 301 redirect or 404 not found? This requires a table lookup of acceptable url parameters for all new visitors. I was thinking that bad url names would be rare so it would be ok to just stop the program with a message, until I realized someone could intentionally set up links to non existent pages on a site.
Intermediate & Advanced SEO | | friendoffood1 -
Is .ME domain is effective in SEO ?
I am always listening about TLD. com. org .net but what about the .me domain. Can this will be effective in SEO. Can i able to beat down my competitors, if i choose .me . I also have a .com or other TLD option but if i am making my name than .me is for me but i need your suggestion for the seo purpose. Is there really domain affective in term of SEO.
Intermediate & Advanced SEO | | pnb5670 -
Domain Alias SEO
We have 5 domain alias of our existing sites
Intermediate & Advanced SEO | | unibiz
All 5 domain alias are domain alias of our main site. It means, all domain alias will have exactly same site and contents
Like Main domain: www.mywebsite.com
DomainAlias: www.myproduct.com, www.myproduct2.com, www.myproduc3.com
And if anybody will open our site www.myproduct.com, it will open same website which I have in primary site what can i do to rank all website without any penalty....i s there any way? This is domain alias of in hosting industry Thanks0 -
What should I cover in a SEO proposal ?
What should I cover in a SEO proposal? Is there any sample SEO Proposal template in SEOMoz?
Intermediate & Advanced SEO | | kashyaplakkad1 -
URL Shorteners. Are they SEO Friendly?
Do URL shortener services like bit.ly act as 301 redirects? I was thinking about utilizing one for longer query based URLs and didn't want to risk losing link juice. Thanks for the insight! Regards - Kyle
Intermediate & Advanced SEO | | kchandler0 -
Is it safe to redirect multiple URLs to a single URL?
Hi, I have an old Wordress website with about 300-400 original pages of content on it. All relating to my company's industry: travel in Africa. It's a legitimate site with travel stories, photos, advice etc. Nothing spammy about. No adverts on it. No affiliates. The site hasn't been updated for a couple of years and we no longer have a need for it. Many of the stories on it are quite out of date. The site has built up a modest Mozrank value over the last 5 years, and has a few hundreds organically achieved inbound links. Recently I set up a swanky new branded website on ExpressionEngine on a new domain. My intention is to: Shut down the old site Focus all attention on building up content on the new website Ask the people linking to the old site to my new site instead (I wonder how many will actually do so...) Where possible, setup a 301 redirect from pages on the old site to their closest match on the new site Setup a 301 redirect from the old site's home page to new site's homepage Sounds good, right? But there is one issue I need some advice on... The old site has about 100 pages that do not have a good match on the new site. These pages are outdated or inferior quality, so it doesn't really make sense to rewrite them and put them on the new site. I call these my "black sheep pages". So... for these "black sheep pages" should I (A) redirect the urls to the new site's homepage (B) redirect the urls the old site's home page (which in turn, redirects to the new site's homepage, or (C) not redirect the urls, and let them die a lonely 404 death? OPTION A: oldsite.com/page1.php -> newsite.com
Intermediate & Advanced SEO | | AndreVanKets
oldsite.com/page2.php -> newsite.com
oldsite.com/page3.php -> newsite.com
oldsite.com/page4.php -> newsite.com
oldsite.com/page5.php -> newsite.com
oldsite.com -> newsite.com OPTION B: oldsite.com/page1.php -> oldsite.com
oldsite.com/page2.php -> oldsite.com
oldsite.com/page3.php -> oldsite.com
oldsite.com/page4.php -> oldsite.com
oldsite.com/page5.php -> oldsite.com
oldsite.com -> newsite.com OPTION 😄 oldsite.com/page1.php : do not redirect, let page 404 and disappear forever
oldsite.com/page2.php : do not redirect, let page 404 and disappear forever
oldsite.com/page3.php : do not redirect, let page 404 and disappear forever
oldsite.com/page4.php : do not redirect, let page 404 and disappear forever
oldsite.com/page5.php : do not redirect, let page 404 and disappear forever
oldsite.com -> newsite.com My intuition tells me that Option A would pass the most "link juice" to my new site, but I am concerned that it could also be seen by Google as a spammy redirect technique. What would you do? Help 😐1 -
How do you implement dynamic SEO-friendly URLs using Ajax without using hashbangs?
We're building a new website platform and are using Ajax as the method for allowing users to select from filters. We want to dynamically insert elements into the URL as the filters are selected so that search engines will index multiple combinations of filters. We're struggling to see how this is possible using symfony framework. We've used www.gizmodo.com as an example of how to achieve SEO and user-friendly URLs but this is only an example of achieving this for static content. We would prefer to go down a route that didn't involve hashbangs if possible. Does anyone have any experience using hashbangs and how it affected their site? Any advice on the above would be gratefully received.
Intermediate & Advanced SEO | | Sayers1