I am onto a photo project. Dynamically added a watermark in the large picture preview. But someone with general idea can manipulate the link and get the original picture. Disabling Right button seems to be useless. In this case what can be the appropriate solution?
Disable any access to original pictures. Make server script that receives picture's ID (or name) and sets watermark before picture will be shown. And make your application such way that this script only can access images by users. If manipulation with link allows users to have unauthorized access, then it looks like you have problems with architecture of project.
Related
I'm attempting to get an array of pixels of the screen (web page) but i know of no way of doing that without using canvas (either straight-up or converting HTML dom elements into canvas, first). I need to capture every pixel on the screen and i don't know what operating system is going to be used so i can't request the display from the O/S, either. Is there a third-party tool, possibly, or a way to do this from the window object in the DOM?
I have only one idea. Maybe you should try to move this functionality to the server. You can use WkHtmlToPDF(http://wkhtmltopdf.org/) for saving websites as PDF, pdf file you can convert to an image and read pixels array.
As web developers with no control of the client machine, there's two approaches to getting a screenshot of a webpage:
Open the webpage in a headless browser on the server and make the screenshot there. phantomjs is a popular one.
(I'm including this for completeness, though you said you don't want to take that route): Use the canvas element on the client. html2canvas is an interesting project that re-renders an entire HTML document into a canvas element so a screenshot can be made.
If your use case allows it, you could of course instruct your users to take a screenshot and paste it in an upload form that can handle images from the clipboard. On Windows, that's a matter of hitting "Print Screen" and CTRL+V.
Here is an api to generate images from online web pages: http://www.page2images.com/Create-Website-Screenshot-with-Javascript-API
So I'm looking for a way to systematically access data from a website. This data is updated every 15 minutes or so, and is generated through a datamart system that makes custom reports following several input parameters: the desired date interval, the specific dataset.
All these parameters require me to click on some specific buttons; I was wondering if it would be possible to automate these click inputs using Matlab (or something else if need be), to retrieve the data and treat it automatically.
Thanks in advance!
I suggest you take a look at http://www.autohotkey.com/. This is a great tool which allows for the automatic clicking on any window (including a browser page) under Windows. It even will let you "search" your screen for pixel images and then click on those images. This would allow you to make a very small bmp file of the link you would like to click on, and then your script can search your page and click directly on the link.
As far as getting data into matlab I'm not exactly sure of the best way to do this, but you might consider saving the html of the page, and then parsing that from matlab.
To preface this I'm not an ExactTarget expert.... We use ExactTarget and are interested in more archiving our email newsletters so that our subscribers can later use our articles as a reference. We have an archive folder set up in the exact target system, but that is not customer-facing and contains things like extra code.
As I understand it, the current process for archiving involves taking screenshots of the email that is sent out which is stitched together as a PDF. The PDF's are then stuck behind a password protected folder on a website.
Obviously, this is a hacky/low-tech way to do this.
Are there any ways to export these mails for archival purposes that people are using?
I am guessing from the above that you are using the premade templates and creating content using the Exact Target WYSIWYG editor. From there I am assuming also that these articles are in content boxes and HOPEFULLY each 'article' is in a single content box.
If so, then your best bet is to go to the HTML tab on this box and copy that code there and have this inserted onto your company blog (or website if you do not have a blog). Your network admin or website designer will likely know the best way to insert these snippets of code to fit the layout and navigation of the site.
If your message is not personalized, or if so you mdke a generic recipient in your list, you may be able to use the 'view as webpage' link to get the source code. Not sure if those links expire after X days or something though.I have never tried this or tested, but on most email sends there is a "view as webpage" link, maybe try taking one of these links and either hosting on your blog/site or using that link as a reference for your customers.
Using a PDF or screen shots is far from optimal as the content becomes an image and loses a lot of value.
Hope this helps.
To make it even easier to work with user content, we enable image post-processing. This way, regardless of what type of file a user uploads from the Cloud or their local device, you can be sure it's in exactly the right size. To convert an image, take the filepicker url and append /convert, along with query parameters specifying what you want to change. See the DocsĀ»
filepicker.io shows an example of a cropping tool on their front page. Could that be built into the picker itself?
We've discussed building it into the upload experience a la iOS, but think that the functionality is best done as a step after the upload. The demo on our front page is done using JCrop, and at some point we'll open-source the demo as a jquery plugin or similar.
We are implementing a crop feature on our image uploads.
Here are the sequence of events.
1a)We call Filepicker.pick method to upload file to filepicker,
1b)Then filepicker uploads it to s3.
2a)Then we call filepicker.stat method to get dimensions.
2b)Then we call filepicker.convert to get small preview image and set that in crop dialog on the user end.
3a)When user adjusts the cropping area using jquery, we capture the dimensions.
3b)Then when user presses the crop button, we call filepicker.crop to get the cropped image.
4a)After that we call filepicker.convet to get 3 sizes. Thumb, small, medium.
5a) Then we move these images to S3 upon save on the user end.
Our end goal is that our user experience while uploading/cropping a file goes smoothly and as fast as possible. Currently this is a challenge, and the whole process is slow. So we are trying to eliminiate step 1b, as it is really not needed.
Does our overall process make sense in your experience of seeing feature implementation for this?
Thanks a lot.
In order to do the necessary processing, the file has to exist online in some form. If the image is being pulled from facebook, dropbox, etc. then it will not be uploaded to S3. However, if the file is local, in order to be processed it must first be uploaded to S3, so there's no way around this step.