google ignoring a page

  • overload
  • Newbie
  • Newbie
  • User avatar
  • Posts: 12
  • Loc: Craiova, Romania

Post 3+ Months Ago

I have a product page (php script?id=xxx) that is not indexed by google. how can this happen since I have links to lots of products in all other over 100 pages that are indexed?

is this possible that google penalize only one page?

is it ok if I rename it?
  • Anonymous
  • Bot
  • No Avatar
  • Posts: ?
  • Loc: Ozzuland
  • Status: Online

Post 3+ Months Ago

  • Axe
  • Genius
  • Genius
  • User avatar
  • Posts: 5739
  • Loc: Sub-level 28

Post 3+ Months Ago

It could be that Google is not indexing it because of PHP sessions (I had this problem on my site for the first 6 months before figuring out what it was)

Best thing to do... Check out The Poodle Predictor and pop in your URL.

Then, put your mouse over some of the links... if you're seeing things like...

Code: [ Select ]
&PHPSESSID=123efh38r93a29r82y389a2

At the end of your URLs, try adding this line to a .htaccess in the site's root directory..

Code: [ Select ]
php_flag session.use_trans_sid off

Then run your URL through the Poodle predictor again. If those sessionID parts disappear from the URL, Google should start seeing you good :)
  • overload
  • Newbie
  • Newbie
  • User avatar
  • Posts: 12
  • Loc: Craiova, Romania

Post 3+ Months Ago

Thanks for the tip Axe, I didn't know about this poodle tool,
it looks nice.

I knew about php session, I read on some other post about
this and I was testing for googlebot and I was not registering
sessions. I can not use use_trans_sid off because I use the
session since this an e-commerce site with user sections.

When you had this problem, you did not get indexed any page
at all, or only some? I also see my search results page is not
indexed. This really sucks, either google is stopide or I am
somehow penalized for God knows what reason because links
are ok, I even checked my page with tidy to see if I have some
messed-up tags.
  • Axe
  • Genius
  • Genius
  • User avatar
  • Posts: 5739
  • Loc: Sub-level 28

Post 3+ Months Ago

Don't worry, setting use_trans_sid off still works for regular people browsing. I've got this setup on PostNuke, and osCommerce, and having zero problems with knowing who users are.

The only pages I had indexed, were the pages that had straight links, without the $POSTNUKESID= bit on the end..

Which was people linking straight to the domain, to /index.php, or to specific articles on the site from elsewhere on the web.

Those were the only ones that Google would see. Anything that was dynamically generated by the site (with the session id on the end), it wouldn't see.

Try setting that in your .htaccess, then login, it should still remember who you are (if you notice, browsing those types of sites in IE doesn't put the session ID on the end of the URL anyway). :)
  • overload
  • Newbie
  • Newbie
  • User avatar
  • Posts: 12
  • Loc: Craiova, Romania

Post 3+ Months Ago

I just dug a bit on some other forums and found this: my page is not
google friendly. crap. how can a page be unfriendly? it is not porn it is
not same content, just products. yes, there is little change (like product
name and some description, but still) so, the solution is to use .htaccess
to rewrite paths so that products/foo would go to product.php?id=foo
pretty easy using htaccess and regex, shame on me not having known
about it before, also shame on you guys that knew about it and let me
search dozens of google forums to find a pretty simple answer.
  • Axe
  • Genius
  • Genius
  • User avatar
  • Posts: 5739
  • Loc: Sub-level 28

Post 3+ Months Ago

Google will index full URLs with ?'s and &'s...

See, Google sees my site's chat page both on its real URL and the fake URL - it does, however, rate the fake .html (PR5/10) page higher than the real one (PR4/10).

But, there have been a few other posts about mod_rewrite here recently (just head up to the search here), which is what you'll need to have on the system to produce the fake URLs in the .htaccess file.
  • SEO Guy
  • Beginner
  • Beginner
  • User avatar
  • Posts: 47
  • Loc: Canada

Post 3+ Months Ago

Google can parse one or two non alpha numeric ID Chars such as & and = but past that it will consider such chars as stop text as it doenst want the spider getting caught in a never ending loop

Using mod_rewrite allows for complete elimination of both sids and non alpha numeric chars so its problem solved
  • natural_angel33
  • Graduate
  • Graduate
  • User avatar
  • Posts: 127
  • Loc: Las Vegas, NV (blech!)

Post 3+ Months Ago

axe,

do you know of a modification to that line you posted earlier for cgi pages rather than php pages?

Keri
  • Axe
  • Genius
  • Genius
  • User avatar
  • Posts: 5739
  • Loc: Sub-level 28

Post 3+ Months Ago

I'm honestly not sure exactly how they're handled with CGI, I haven't played around with Perl in a few years (not since I found PHP actually, heh).

Your best bet would first be looking at how Perl handles sessions, and if they do indeed pose a problem to search engines with messed up URLs. The best way to find that out is just type your URL into the Poodle Predictor (the link's right up there at the top of this thread). If the links look the same in the poodle predictor as they do in your browser then all is good in that respect.

Then you can just go ahead and use the mod_rewrite to alter the regular URL.
  • overload
  • Newbie
  • Newbie
  • User avatar
  • Posts: 12
  • Loc: Craiova, Romania

Post 3+ Months Ago

I've done this change and also renamed file but still no change.

google bot has not hit the page yet (maybe once I think) and
still it doesn't get indexed. what could still be the problem?

I've check on some articles that google does not like multi-domain sites.
I have 4 addresses pointed to the same site, is this a problem?

how come similar pages to this (same script with more id's / dynamic content) gets indexed?
  • Axe
  • Genius
  • Genius
  • User avatar
  • Posts: 5739
  • Loc: Sub-level 28

Post 3+ Months Ago

It can take a while, Google still hasn't given my new site a PR that I put up almost 4 weeks ago - and there's at least 30,000 pages out there on the web across various sites linking to it - and these are 30,000 pages that Google has already had indexed for months. There's still at least that many pages linking to my new site that Google hasn't indexed yet, plus about 4,000 pages internally on the site that link to the mine index page of the site. So, sometimes it can take a while.

Post Information

  • Total Posts in this topic: 11 posts
  • Users browsing this forum: No registered users and 12 guests
  • You cannot post new topics in this forum
  • You cannot reply to topics in this forum
  • You cannot edit your posts in this forum
  • You cannot delete your posts in this forum
  • You cannot post attachments in this forum
 
 

© 1998-2014. Ozzu® is a registered trademark of Unmelted, LLC.