Is it possible to run perl in google drive - perl

I have uploaded my entire perl directory to google drive, including perl.exe, /lib, perl scripts, and data files.
Is it possible to run perl.exe on perl scripts using the data files, within the google drive?
If so, where can I find out how to do it?

Google drive is a File storge system.
It is not a server that can run any applications.

Related

Unzip a file from several large compressed folder from command line

I have several large zipped folders in my cloud storage drive. I want to transfer a specific file from each of the zipped folders to my local hard drive (I cant copy all of them since i dont have enough space). Is there a way to do this using command line/cmd or powershell. I am using Windows 10 (Build 18362).
The file name is the same so i was hoping if i can write a loop to do this.

unable to run php cli from the command line when xampp portable is installed in another drive

I have installed a xampp portable package on my drive D, added the php folder path inside it on the system environment so I would be able to run php basically anywhere from the command line. but when i try to check if its working by executing the "php -v" command, it runs but i get this following error/warning messages.
"Warning: PHP Startup: Unable to load dynamic library 'xampp-portable\php\ext[:any].dll - The specified module could not be found."
*the [:any] there just means some dll filename.
and i get a lot of that warnings with just varying dll filenames,
anyway, here's the catch. when i do change my current path in the command line to somewhere in drive D, and then run the same command "php -v" it runs smoothly.
so what configuration should i do to fix this problem?
by the way, i do the same with mysql, putting its bin directory path on the system enviroment, run it in both inside drive C and drive D. and it runs smoothly. so i guess this problem is only on php.
I've posted the same question in xampp forums and this is the answer that was given to me. I'll be quoting the answer as it was and give credit to the one who answere it.
XAMPP is very proud that it don't have to set any environment
variables or registry values.
What you want to do is not possible with XAMPP portable without
mapping to a specific Drive letter. All configurations in xampp
portable have relative paths, so if you want to use cli from a
different drive letter, php can be executed, but all extensions of
them throw errors.
I would install the full version of XAMPP, map it to the current drive
letter with help of the setup_xampp.bat file and use the XAMPP control
panel Shell instead of default command line (advantage: The XAMPP
control panel shell sets local include paths to the needed XAMPP
folders).
Works for me (i can even use pear, git, composer, and other cli
scripts with the Xampp Shell if correct integrated).
best wishes, Altrea
credits to Altrea
Open your php.ini file, replace all relative paths to absolute paths. Eg: \xampp\ to C:\xampp\

Copy Directories Recursively From Ftp Server Using Perl

I need to write a perl script which has to log in to an FTP server and download all the sub-directories and contents on sub-directories to local machine. The version of Perl on the FTP server is 5.8.8, i can't upgrade it. One method is to create directories on local machine and then copy each file. I was wondering if there is any command to copy a directory and its content. Is it possible to "tar" the directory to save space?
Thanks,
Amit.
There is Net::FTP::Recursive. I haven't tried it but it seems to fit your requirements.

suitable packages in perl for with features mentioned in the subject

I need the package having below features:
needs client based secure file transfer machanism for getting files from multiple directories at the remote machine.
Must have features to do ls at remote machine.
Must have functions to know the file permissions at remote machine.
It have to use single connection for all file transfers.
That package have to use less number of resources and it have to do transfer at fast.
zipping the files at the remote machine.
Is this homework? And, have you tried searching CPAN? There are several modules that do what you need, such as Net::SCP and IO::Compress::Gunzip.
You can get most of this done with Net::SFTP. How well the remote server complies with the compression part is mainly a function of that machine. If it supports it, you can probably issue a site command to do it.

Why is my Perl source code being displayed in the browser?

What does it mean when I access my Perl script via URL, but when I do, the source code is printed on the screen?
Does this mean that Perl isn't properly set up? I'm using Apache on Fedora.
This means your webserver isn't set up to execute the script at that url. What server are you using?
It could also mean you are putting the Perl script in the wrong folder. The cgi-bin folder is still widely used as the folder where CGI scripts should be stored and run from. Other folders may just open the file and read it as text, similar to an HTML document, instead of running the document as code. But it can also mean the server is just not setup properly to run Perl scripts or other types of server-side scripts.
This means that you probably aren't doing what your server is expecting. Usually servers expect either that every file in a CGI directory is an executable, or that files with a certain extension are executable and it can serve any other file as its content.
Figure out which file extension your server expects your CGI program to use.
Ovid's CGI Course may help you (if you use CGI of course)