crawl-300d-2M-subword.zip corrupted or cannot be downloaded - fasttext

I am trying to use this fasttext model crawl-300d-2M-subword.zip from the official page onI my Windows machine, but the download fails by the last few Kb.
I managed to successfully download the zip file into my ubuntu server using wget, but the zipped file is corrupted whenever I try to unzip it. Example of what I am getting:
unzip crawl-300d-2M-subword.zip
Archive: crawl-300d-2M-subword.zip
inflating: crawl-300d-2M-subword.vec
inflating: crawl-300d-2M-subword.bin bad CRC ff925bde (should be e9be08f7)
It is always the file crawl-300d-2M-subword.bin, which I am interested in, that has problems in te unzipping.
I tried the two ways many times but with no success. it seems to me no one had this issue before

I've just downloaded & unzipped that file with no errors, so the problem is likely unique to your system's configuration, tools, or its network-path to the download servers.
One common problem that's sometimes not prominently reported by a tool like wget is a download that keeps ending early, resulting in a truncated local file.
Is the zip file you received exactly 681,808,098 bytes long? (That's what I get.)
What if you try another download tool instead, like curl? (Such a relay between different endpoints might not trigger the same problems.)
Sometimes if repeated downloads keep failing in the same way, it's due to subtle misconfiguration bugs/corruption unique to the network path from your machine to the peer (download origin) machine.
Can you do a successful download of the zip file (of full size per above) to anywhere else?
Then, transfer from that secondary location to where you really want it?
If you're having problems on both a Windows machine, and a Ubuntu server, are they both on the same local network, perhaps subject to the same network issues – either bugs, or policies that cut a particular long download short?

Related

Antivirus flagged dart.exe as ransomware

I was trying to setup flutter when my antivirus detected dart.exe as ransomware... I downloaded the package from the git repository using the documentation command line. Should I give it the permission or delete and download the zip?
I had the same thing happen with my generated build. I just ignored it.
You should check the MD5 or SHA-1 checksum of what you've downloaded with the values they provide. Then you can know for sure that they came from the correct source. (IF they provide one)
See this answer for info of how to verify
https://security.stackexchange.com/questions/189000/how-to-verify-the-checksum-of-a-downloaded-file-pgp-sha-etc
In general, just double check you're downloading from the correct domain. (Are you sure you're on github.com, is your connection secured? etc.)
General antivirus programs are cautious about downloading .exe files from the internet. (as they should be)

Downloaded Eclipse Mars 1 checksums don't match

I downloaded Eclipse Mars 1 from a mirror chosen for me on the Eclipse website. But the md5sum on the downloaded file is 929f821dc83eaac83fc6320b291dcb7f while on the Eclipse website it's been given as 72a722a59a43e8ed6c47ae279fb3d355. This struck me as odd because it's the first time I'm experiencing something like this -- I've made like ~20 such downloads, different files, different servers, since I started learning how to code -- and all of them matched the checksums on their respective servers.
So I decided to get a fresh download from a different mirror. I chose https://spring.io/tools/eclipse. And the thing is, I can't find any checksums around there but the md5sum for their file is a562f87ddf353dd8519edfc072d4c67d. I'm confused. I'm under the impression that irregardless of the mirror, the file hash should match the hash given on the Eclipse website -- 72a722a59a43e8ed6c47ae279fb3d355
Either the md5sum was not updated correctly on the website or you had a hiccup in your Internet connection that made the file change, resulting in the difference you're having.
Another rare case might be that you made a mistake in choosing the checking
option: for example I once copied the sha1 code instead of md5.

Samba read speeds very slow through Explorer, but OK through Firefox

I have a file server running Ubuntu 12.04 and Samba 3.6.3. A Samba share is mapped to a drive on a Windows 8 machine.
When copying a test file to a local drive (which is an SSD and not a bottleneck here), it is very slow when doing so through Explorer. It is similarly slow when downloading the file through Internet Explorer. When downloading through Firefox (by entering the file URI), however, it is more than 10x as fast, as the image below shows.
What's going on here? I know that Samba is not fast, but I thought that's generally the case when dealing with lots of small files, when its request logic is very inefficient. The test file was 826 MB.
Removing custom "socket options" line in smb.conf (the Samba configuration file) solved it for me.
It seems that it's best to leave that option blank nowadays, since it will calculate optimal values itself. Firefox seemed to be either using its own SMB protocol settings, or ignoring those set by the Samba server.

online space to store files using commandline

I require a small space online (free) where I can
upload/download few files automatically using a script.
Space requirement is around 50 MB.
This should be such that it could be automated so I can set
it to run without manual interaction i.e. No GUI
I have a dynamic ip & have no tech on setting up a server.
Any help would be appreciated. Thanks.
A number of online storage services provide 1-2 GB space for free. Several of those have command-line clients. E.g. SpiderOak that I use has a client that can run in a headless (non-GUI) mode to upload files, and there's even a way to download files from it by wget or curl.
You just set up things in GUI mode, then put files into the configured directory and run SpiderOak with right options; files get uploaded. Then you either download ('restore') all or some of the files via another SpiderOak call or get them via HTTP.
About the same applies to Dropbox, but I have no experience with that.
www.bshellz.net gives you a free shell running Linux. I think everyone gets 50mb so you're in luck!

Is there a way to make Netbeans file uploads not fail on a single error?

Is there a way to make Netbeans's FTP uploads less error sensitive(reconnect and retry on error)?
It is simply unable to handle large amounts of files. If I upload 1000 files and there is some sort of failure at the 400th file, it will fail for the rest of the files, and it will not record the files that failed. So I will have to upload ALL the files again, even the ones that uploaded successfully on the previous attempt, but it will fail agan again and again.
This post describes changing a NetBeans FTP setting to Passive. Doing so reduced many of my transfer errors.