I've seen news sites (CNN, Fox News, etc.) use HTML files as their post content. For my blog, I currently use dynamic pages (e.g. www.example.com/post/?id=3).
I'm wondering if this is the correct way to go, mostly because AdSense won't accept /post/ for ads. Is this because it's just pulling up /post/ & not the id?
So basically, which way do you recommend? Thanks
It depends on the contend of your page. But basically the good way is to create easy to read links like:
http://example.com/drive-to-norway
It's because it's easy to read for people and before clicking the user knowns what it could be (instead for example: http://example.com/id=3)
Some bigger pages do not use that convention because they for example sell a lot of similar items and having named, unique links without any numbering isn't possible/easy for them. Like I wrote at the beginning - it depends on content.
When you get a list of files through the Box API, you also get the tags assigned to those files in string format. These tags include your own Box tags as well as shared tags (My Tags vs All Tags).
However, if you want to verify in a later phase if some of these tags have been deleted for instance, that seems not possible. The 'export tags' API method only returns your own tags.
This makes it really difficult to keep tags in sync, and I don't see a good reason why you would be able to see tags in the Box UI but not in the API. In the end it's just another representation of the same data, or that's what you would expect.
Anyone knows a possible workaround for this?
There isn't a known workaround for this at this time, but we're in the process of building out tags support in our V2 API and will take this use case into consideration.
I've googled a lot but it seems to be rather difficult to find complete references on the web that is easily digestable,
anyhow,
I've got a fully functioning html5 website that allows people to list their property online and run it in an ebay fashion, although I'd like to optimise my code with schema html tags, from what I gather I can use the schema.org/thing item but I was wondering if there is anything dedicated more to property and also I wanted to ask if anybody managed to get the userinteraction tags working, for stuff like PageLikes etc
Depending on the property, you could use http://schema.org/SingleFamilyResidence or any other kind of http://schema.org/Residence. Also there are subtypes of http://schema.org/Store if you're allowing those kinds of properties. Really the list is pretty extensive of the types of properties you could have. See http://schema.org/docs/full.html for the entire list of schema.org types.
I understand that no matter what I do, someone will be able to copy it. However I can still make them work hard for it. What are some good ways of making data not easily copied using php compatible coding.
--- Added ----
The data is a listing of results for certain local sports events. We send people out to collect the information, post the information, make corrections and such. However a competing website takes our results (I know they are directly copying them) and never updates them which causes people to call our office and complain.
---- Answer for my Use ----
I picked one of them, however I am going to use multiple of your answers. I am going to add my link in a using the copy pasta trick. I am going to put fake hidden text into it. I am also going to do the fake hidden text trick with different versions of the div tag that are fake (making it even harder to scrape or to do something like copy to textpad and replace it real easily), and I am going to talk to a lawyer as well about legal recourse and what I can do to make it illegal for them to copy the data (such as creative bios or something cool like that). Thanks for your help.
Joe, you can't really make them work really hard to get your data. It's essentially just a single request to any of your pages. Your best option is to explicitly state that you own the rights to all of your content, and that any infringement on that ownership will lead to legal ramifications*.
* Not a lawyer
Your data will be copied to every computer that requests the page and it will stay there until the person clears their cache. To answer your question, you can't.
What you can do is create a CSS style such as:
.copy-pasta { display: none; }
And then throughout your content, add something like this:
<p class="copy-pasta">Content provided via [your website here]</p>
This will increase your page rank when copy-pasters blatantly steal your content, meaning you will show up first in search results.
Place some <div style="display: inline; position: absolute; overflow: hidden; width: 0px">useless words</div> in the text. It won't display for reading, but if someone copy and paste... "WOW where it came from WTF!! *CRY*"
How about putting links to your site in with the displayed data? No big fanfare, but just suggest that the for the most up to date figures, they can go to the real website that publishes them.
Most of what you try will only work for a time. Until you exceed their laziness factor. (What they're doing suggests a high laziness factor.)
Laws don't protect publicly available data, but you may be able to protect the packaging and presentation.
Programs used to copy out data look for the data using pattern-matching. You could 'decorate' your data with randomly-chosen tags (like one row would have a span tag surrounding it, the next row a div, etc...). Just a thought.
Clarification:
With screen-scraper at least, the user of the program specifies what HTML comes before the data they want, and what HTML comes after it. You can make it more difficult for them to automatically retrieve the data.
Why are people calling your office to complain if the data is on a competing website? If they have a domain name that is similar enough to yours that people are confusing the two of you or if they've put something on their site that makes it look like you've endorsed them, then you've got them for trademark infringement.
Disable the context menu is a start.
$(document).bind('contextmenu', function(e)
{
return false;
});
Or
<body oncontextmenu="return false;">
Forbidding people to get data is almost impossible. You can mess up your tags and make the code really dirty and hard to parse... but it's not really enough. You could also generate a big image with the data in it, this would be painful to parse! ... but you don't want to do that.
Because you said...
However a competing website takes our
results (I know they are directly
copying them) and never updates them
which causes people to call our office
and complain.
... my call would be to take this the other way and create an API allowing people to get your content in a way that YOU designed.
Also if they are just shamelessly stealing your data and they don't have the right to do it, consider a legal option.
Another option is to use PHP code to generate images from the site's HTML. You would use the images to display the content, instead of HTML which can be easily copied out. Example code is here, and I bet you could find more code to do this by Googling:
http://www.acasystems.com/en/web-thumb-activex/faq-php-convert-html-to-image.htm
Try Copyscape it wont prevent your content from being copied, but it will make finding the copies very easy.
You may encrypt the data on the page, and have javascript obfuscated decoding routine that will decode it for you viewers. You may switch keys and encryption algorithms from time to time. Same javascript should disable ability to select text and/or copy it to prevent manual copy-pasting.
They won't be able to copy manually and their scraper would have to be able to run javascript to get the data.
Caveat is that the data won't be visible for Google, but if data is rather numeric it might not be such a big harm.
If they scrape automatically and very often you may also try to pinpoint their IP by observing most active IP-s on your site and serve them fake data.
Please don't use lawyers, that's hitting below the belt.
use swf to display your data, just like other online books
What is the actual procedure for attaching multiple tags to a particular content in a project development. What is this tagging all about???
I need to create a tag cloud for my project in .NET using c#. Help me out as a beginner for basic tagging concept.
Tags are key words add information about the item being tagged. Tags add semantic information about something in an effort to further it.
For instance, A picture of your father on his birthday could be tagged 'dad','family','event','birthday' etc...
By adding tags to the picture you add context and make the image more easily indexable, sortable and searchable.
Tags are purposely generic and flexible because different people can place different personal meaning to the same artifact, or the same person may apply different meaning in different contexts - like adding the same picture to a stock photography web site or checking it into a source code repository as part of a project.
Generally the procedure is to ask the owner of the item to add a list of tags in a text field. Some sites like stackoverflow constrain (most) users to use existing tags, others like delicious make the tags up to the user.
A tag in the software context typically means a meaningful name or attribute being assigned to that software. In version control scenarios a tag is a meaningful name given to a particular state of the files represented by that name. For example the tag 20090401 might be assigned to the source code as it looked on April 1, 2009. Tagging something can also mean describing it or categorizing it. For example software such as IE8, Chrome, or Firefox might all be tagged "Browsers" to categorize them on a download page. Allowing users to create tags and use existing tags is a powerful method to categorize content and help people zero in on items of interest. A tag is simply an extra tidbit of information a person can gain insight into data with.
Multiple tagging is useful for many reasons in software development. For example in my git repository I have a habit of creating tags based on date which can easily be ordered and parsed by a computer. I can also give changes a more human consumable name such as the tag "Deleted_Duplicates", or "RC1", or "V1_Delievered_To_Michigan". This allows for an understanding while also allowing for machine processing.