How do I exclude a sub folder/directory from the Azure Backup for an App Service? - powershell

How do I exclude a sub folder/directory from the Azure Backup for an App Service?
Our backup system seems to fail because this folder makes the site exceed the backup limit. So I'd like to exclude only that folder.
The website + database size exceeds the 10 GB limit for backups. Your
content size is 10 GB.

#AjayKumar-MSFT's link to Partial Backups works, but here's the details:
create a file called _backup.filter
list one directory or file per line
Upload _backup.filter file to the D:\home\site\wwwroot\ directory of your site
Example:
\site\wwwroot\Images\brand.png
\site\wwwroot\Images\2014
\site\wwwroot\Images\2013

You may filter out unnecessary files from the backup by Configuring Partial Backups. You can exclude static content that does not often change such as videos/images or unneeded log files (directories).
Partial backups allow you to choose exactly which files you want to backup.

Firstly I suggest you could check your app usage(Make sure your web app has already use 10GB). You can determine folder content size using the Azure Web Apps Disk Usage Site Extension.
You could follow this steps to see the disk usage:
Browse the Kudu site for your web app: https://sitename.scm.azurewebsites.net Click on Site extensions and then click on Gallery.
Search for Azure Web Apps Disk Usage site extension. Click on + icon to install the site extension.
Click run to start the Disk Usage Site Extension
If your web site is not exceed the 10GB. I suggest you could create a new service plan and move the your web app to the new service plan and test again. Maybe something wrong with the web app server.
If this still doesn't solve your issue. You don't exceed the 10GB but still show this error. I suggest you could create an Azure support request.
How do I exclude a sub folder/directory from the Azure Backup for an App Service?
As Ajay says, you could use Partial backups.

Related

Move website to another server without FTP or SSH

I have a problem with an existing TYPO3 website. (9.5.13)
Unfortunately, I no longer have FTP or SSH access to this server. (Provider can no longer be contacted) I would therefore have to move this website to another server. I only have admin access to Typo3 website.
Unfortunately I can't find a way to do this without copying the files and the database to the new server.
Is there a possibility? If so, how can I do this?
There is no real good way if you don't have a proper access to the server.
For the database you could use the extension phpmyadmin or t3adminer
but there is IMO no extension to download the content of the fileadmin (which will be a native feature of TYPO3 11 LTS).
However you could download at least all extensions of course.
In former times, there were two extensions trying to implement a FTP client as backend-module. The one was "Typo3 quixplorer", the other was "Typo3 eXtplorer"... Quixplorer is outdated (last update of extension in 2010), eXtplorer had its last update 2008...
Both extensions have security issues! Be careful and remove them immediately after usage!
Maybe you can find similar projects running in current web environments...
You may build some php script to archive the whole fileadmin-folder, and make a database dump. which you can download from the server.
Or you use a script to contact the new server to transfer the data immideatly.
It's just a question how you can upload these scripts as TYPO3 does not allow the direct upload of executables.
This is a really unfortunate situation, but you can copy the database and the files by using this extensions. Be aware: I haven't tested this extensions or took a look into the code and picked them by their description and README files. Use them carefully and on your own risk!
Note: In general the TYPO3 Security Team recommends to not use any
extension that bundles database management tools on production TYPO3
websites.
Database:
Use EXT:adminer or EXT:phpmyadmin to download your database tables.
Files:
Install EXT:commands or EXT:additional_scheduler to zip & download your files & folders.
All in one solution:
While researching I've found the extension EXT:ns_backup. Maybe you can us it to transfer the backup to cloud services like dropbox.

Appservice deployment issue

Hi iam deploying my war file using azure appservice from azuredevops. deployment is getting completed successfully but when i hit the url Microsoft azure welcome page is displayed instead of application. can you please suggest a work around
In Azure Web Apps, the default document is the web page that is displayed at the root URL for a website. The first matching file in the list is used. Refer File structure on azure to know more about the files and directories on Azure Web App.
You may check whether the deployed files are available or not using Kudu Console.
Also, make sure that your startup file (For ex: index.htm) is added to the default documents section as mentioned above. You can delete the remaining files if not required and see if that works.
For details , you can refer to this case.

Backup configuration page not loading

I upgraded my app service to Standard a few days ago. However, when I go to the backup configuration page, it never finishes loading. I just get the dots across the top of the screen forever.
According to your description, I have checked this issue. Per my test, I found I could not configure my backups for my mobile app on Azure Portal as follows:
While for web app, I could configure the backups via Azure Portal as follows:
Then I Use REST to back up and restore App Service apps and Use PowerShell to back up and restore App Service apps to check this issue. After configure the backup for my mobile app, I could see the Configure button on Azure Portal, but when I try to change the backup configuration, I encountered the same issue as you mentioned.
Per my understanding, you could leverage the powershell command Edit-AzureRmWebAppBackupConfiguration for a simple way to edit your Backup Configuration, you could refer to here for more details.

Auto upload remote files into Google cloud storage via FTP?

I download a lot of csv files via ftp from different sources on a daily basis. I then upload these files into Google Cloud Storage.
Are there any programs/api/tools to automate this?
Looking for a best way, if possible, to load these files directly into Google Cloud Storage without having to locally download them. Something that I can deploy on Google Compute, so I don't need to run a local programs like Filezilla/CrossFTP. The program/tool will keep checking the remote location on a regular basis and load new files into Google Cloud Storage; ensuring a checksum match.
I apologize in advance if this is too vague/generic question.
Sorry, no. Automatically importing objects from a remote FTP server is not currently a feature of GCS.

hidden (?) files

I'm writing an app that includes a folder/file explorer. That part is working. However, another app (Astro File Manager) has the ability to backup all of the installed apps to a backup location on the SD card. I cannot find or see the folder(s) where installed apps are stored. Supposedly it is in "/data/apps". My phone is NOT rooted and Astro works fine. Does anyone know what the trick is to seeing these folders/files? Is it possible to do this in B4A? Thanks!
Android doesn't allow you to access other programs files. Astro probably uses the backup API that allows you to backup other programs. However the backup API doesn't give you access to the files.