PgAdmin4 download files - postgresql

Is there any way to download the backup files saved on the server's file system by pgadmin4 server. The best way I thought was to make the files available through Apache, but this solution has problems with authentication, or leaving the files public or requiring a new password.

You can use Store Manager to download backup files from pgadmin.
You can access it from Menu -> Tools -> Store Manager.
From the docs,
Storage Manager is a feature that helps you manage your systems
storage device. You can use Storage Manager to:
Download, upload, or manage operating system files. To use this feature, pgAdmin must be running in Server Mode on your client
machine.
Download backup or export files (custom, tar and plain text format) on a client machine.
Download export dump files of tables.

It was implemented in version 4.28.

Related

Sharing vscode-server for multiple users to save disk space

We have a Linux server (Ubuntu 20.04) with ~100 users and very limited disk space. The disk quota for each user should be limited to 100MB. The access to the server is via SSH.
Several users want to use VSCode for remote development of source code (not shared) that resides on the server. They can install the remote server for VSCode, but it results in a rather large directory .vscode-server from 300MB and up for each user. This would fill up the disk with just a few users activating and using it.
Ideally the users could share the VSCode-server part just as they share all the other software development tools on this server. Is this possible, and how to do it?
Upgrading the server hardware/disk is not possible for the scope of this question.

Move website to another server without FTP or SSH

I have a problem with an existing TYPO3 website. (9.5.13)
Unfortunately, I no longer have FTP or SSH access to this server. (Provider can no longer be contacted) I would therefore have to move this website to another server. I only have admin access to Typo3 website.
Unfortunately I can't find a way to do this without copying the files and the database to the new server.
Is there a possibility? If so, how can I do this?
There is no real good way if you don't have a proper access to the server.
For the database you could use the extension phpmyadmin or t3adminer
but there is IMO no extension to download the content of the fileadmin (which will be a native feature of TYPO3 11 LTS).
However you could download at least all extensions of course.
In former times, there were two extensions trying to implement a FTP client as backend-module. The one was "Typo3 quixplorer", the other was "Typo3 eXtplorer"... Quixplorer is outdated (last update of extension in 2010), eXtplorer had its last update 2008...
Both extensions have security issues! Be careful and remove them immediately after usage!
Maybe you can find similar projects running in current web environments...
You may build some php script to archive the whole fileadmin-folder, and make a database dump. which you can download from the server.
Or you use a script to contact the new server to transfer the data immideatly.
It's just a question how you can upload these scripts as TYPO3 does not allow the direct upload of executables.
This is a really unfortunate situation, but you can copy the database and the files by using this extensions. Be aware: I haven't tested this extensions or took a look into the code and picked them by their description and README files. Use them carefully and on your own risk!
Note: In general the TYPO3 Security Team recommends to not use any
extension that bundles database management tools on production TYPO3
websites.
Database:
Use EXT:adminer or EXT:phpmyadmin to download your database tables.
Files:
Install EXT:commands or EXT:additional_scheduler to zip & download your files & folders.
All in one solution:
While researching I've found the extension EXT:ns_backup. Maybe you can us it to transfer the backup to cloud services like dropbox.

Local Database File Associated with .dwg File

I am planning on building an AutoCAD extension application that will require custom data be stored outside of the standard .dwg file for AutoCAD drawings. I would like there to be a local file that this custom data is stored in so that the data can be read into AutoCAD or saved from AutoCAD while offline. I have been imagining that each .dwg file would have it’s own separate database file associated with it, but I am also open to the idea of having a single data file locally stored in order to allow for offline reading/writing of my custom data. Does MongoDB support this type of local data storage? There will be a cloud-based database where the data can be read from/written to, but I want there to be a local storage system to allow for offline read/write and also improved speed. I am just a bit confused about this because most resources online seem to address cloud storage and I am having a hard time understanding how to use MongoDB to implement a reliable local storage system.
It's possible to install the MongoDB Community Server edition locally on your machine.
You can download the installer here.
Installation instructions can be found here.
This post addresses where the data is stored. Basically it's one storage location per machine (where you can put all your databases).
You may need a GUI interface to browse all your databases. The community edition installer will prompt you to install Compass. I'm using a different software called Robo 3T for that.
Something like nedb-promises may be of interest for creating a database local to the application.
(I've also been looking into how to use MongoDB locally, so the above is a summary of what I've found so far.)

Auto upload remote files into Google cloud storage via FTP?

I download a lot of csv files via ftp from different sources on a daily basis. I then upload these files into Google Cloud Storage.
Are there any programs/api/tools to automate this?
Looking for a best way, if possible, to load these files directly into Google Cloud Storage without having to locally download them. Something that I can deploy on Google Compute, so I don't need to run a local programs like Filezilla/CrossFTP. The program/tool will keep checking the remote location on a regular basis and load new files into Google Cloud Storage; ensuring a checksum match.
I apologize in advance if this is too vague/generic question.
Sorry, no. Automatically importing objects from a remote FTP server is not currently a feature of GCS.

hidden (?) files

I'm writing an app that includes a folder/file explorer. That part is working. However, another app (Astro File Manager) has the ability to backup all of the installed apps to a backup location on the SD card. I cannot find or see the folder(s) where installed apps are stored. Supposedly it is in "/data/apps". My phone is NOT rooted and Astro works fine. Does anyone know what the trick is to seeing these folders/files? Is it possible to do this in B4A? Thanks!
Android doesn't allow you to access other programs files. Astro probably uses the backup API that allows you to backup other programs. However the backup API doesn't give you access to the files.