Stopping PDFs Theifs! Filetype:pdf Google?

by entry
7 replies
When you enter this in google, Filetype : pdf for say "quit smoking" it pulls the pdf files across the internet which are downloadable, and are in google search format. along sidfe their urls.

How can you prevent your pdf from being indexed? so that nobody can find it in a google search, so the seo spiders cannot get to it.

so when this gets entered
Filetype : pdf "entry"

my PDFs which are hosted Dont get listed?


{what is the solution, using no follow for example, in the folder ?}

but the downloaders from my Optin form should be able to download the ebook with No restrictions.
#filetypepdf #google #pdfs #stopping #theifs
  • Profile picture of the author aaron_nimocks
    Create a robot.txt file in your main directory and add

    PHP Code:
    User-agent: *
    Disallow: /filename.pdf 
    Signature

    My free PSD logs can be downloaded at PSD Bum. Enjoy!

    {{ DiscussionBoard.errors[3298837].message }}
    • Profile picture of the author CDarklock
      Originally Posted by aaron_nimocks View Post

      Create a robot.txt file in your main directory and add

      PHP Code:
      User-agent: *
      Disallow: /filename.pdf 
      Not exactly. Basic setup:

      1. Create a folder on your web host named "secure"
      2. Stick a 0-byte index.html in that folder
      3. Disallow the "secure" folder in your robots.txt
      4. Create a new folder underneath it with your files in it

      Advanced setup:

      5. Set a script as your 404 error handler for the "secure" folder
      6. Look up the end of the URI in your MySQL database
      7. If it's found, redirect to your download page
      8. A second script needs to be updating your MySQL database with unique download IDs for customers, and sending the proper URIs to those customers.

      The basic setup actively prevents spiders from indexing your download folder, but does not stop your customers from sharing their download links. The advanced setup at least lets you track which customer's download link is being used.
      Signature
      "The Golden Town is the Golden Town no longer. They have sold their pillars for brass and their temples for money, they have made coins out of their golden doors. It is become a dark town full of trouble, there is no ease in its streets, beauty has left it and the old songs are gone." - Lord Dunsany, The Messengers
      {{ DiscussionBoard.errors[3323589].message }}
      • Profile picture of the author entry
        Originally Posted by CDarklock View Post

        Not exactly. Basic setup:

        1. Create a folder on your web host named "secure"
        2. Stick a 0-byte index.html in that folder
        3. Disallow the "secure" folder in your robots.txt
        4. Create a new folder underneath it with your files in it

        Advanced setup:

        5. Set a script as your 404 error handler for the "secure" folder
        6. Look up the end of the URI in your MySQL database
        7. If it's found, redirect to your download page
        8. A second script needs to be updating your MySQL database with unique download IDs for customers, and sending the proper URIs to those customers.

        The basic setup actively prevents spiders from indexing your download folder, but does not stop your customers from sharing their download links. The advanced setup at least lets you track which customer's download link is being used.

        Cheers for your help! I will go and copycat that method

        Signature
        I Have to say a Massive...THANK YOU to every Warrior who has helped me, and thanks to every warrior who helps me in the future...
        {{ DiscussionBoard.errors[3323728].message }}
  • Profile picture of the author Shaun OReilly
    I use download protection systems such as DLGuard
    and RAP (Rapid Action Profits).

    These allow me to place my files in a non-web-accessible
    folder that cannot be indexed in the search engines.

    So instead of placing the file in a location such as...

    example.com/file.pdf or example.com/folder/file.pdf

    I place it in a file above the public_html folder as then
    it can't be indexed or even accessed via the web - unless
    the person really purchased via DLGuard, RAP or whatever.

    Dedicated to mutual success,

    Shaun
    Signature

    .

    {{ DiscussionBoard.errors[3298862].message }}
    • Profile picture of the author entry
      Can we use robots.txt to tell crawlers to not to index pdf files.

      What are the steps for that ?

      and what would I name the file?


      and code what would I include in the file ?
      Signature
      I Have to say a Massive...THANK YOU to every Warrior who has helped me, and thanks to every warrior who helps me in the future...
      {{ DiscussionBoard.errors[3323100].message }}
  • Profile picture of the author Tony Dean
    A simple way to stop Googlebot from indexing it in the first place is to zip the pdf and call it some anonymous number like 0673451.zip or somesuch.
    Signature

    {{ DiscussionBoard.errors[3323492].message }}
  • Profile picture of the author Tony Dean
    I just used your code in Google and put in "How To Make A Million Dollars In Seven Days"

    I did not expect to find anything so flamboyant an ebook, but I found several!
    Signature

    {{ DiscussionBoard.errors[3323599].message }}

Trending Topics