Postgres on HDD Partition guide - postgresql

I am trying to reproduce the following
/
/var
/usr
/tmp
/home
[PGDATA]
[PGDATA]pg-xlog
[PGDATA]base
[PGDATA] / [tablespace]
[PGDATA] xlog-archive
but i'm a bit confuse with [PGDATA], i'm using Virtual host Ubuntu server 14.04 and i'm stuck in partition HDD stage. any references to read about this?
Thanks in advance

Related

external hard disk not detected in my centos 8 server?

My Centos 8 server not detected WD external hard drives which are mount in / with different folder like PHD10, PHD11 etc.
It's faults when running machine.
In /etc/fstab UUID show all disk & file system is ext4.
but in lsblk or fdisk -l not show those disk.
Can someone help?

How to change the path of /tmp folder in ubuntu 16

I am using Ubuntu 16.04 server. I have a couple of apps that are running and therefore can't reboot occasionally. Because of this my /tmp folder is filling up pretty fast and sometimes I have problem with logging in to the server
I want help with redirecting /tmp to another path like /mnt so I will be able to clear it periodically.

Kickstart: create lvm volume group without partition

I am trying to setup a redhat 7 kickstart for a server with 2 disks.
On the second disk I want to use the full disk without partitioning in lvm.
Once the system is installed, the config works:
pvcreate /dev/sdb
vgcreate data /dev/sdb
lvcreate -l +100%FREE -n data data
mkfs.xfs /dev/mapper/data-data
echo -e "/dev/mapper/data-data\t/data\txfs\tdefaults\t0 1" >> /etc/fstab
mount /data
But I cannot manage to have the following partitioning to work as expected
The partitioning system on kickstart as I know it will only create a partition on /dev/sdb and the volume is create on /dev/sdb1 at the end.
I managed to workaround the issue by using a postscript, but I compiled packages to install in this folder, so I would need the formating to be done before at least in a pre script if the partitioning is not possible.

How to Mount Disk for Google Cloud Compute Engine to use with /home?

I have a VM Instance with a small 10GB boot disk running CentOS 7 and would like to mount a larger 200GB Persistent Disk to contain data relating to the /home directory from a previous dedicated server (likely via scp).
Here's what I tried:
Attempt #1, Symlinks Might work, but some questions.
mounted the disk to /mnt/disks/my-persistent-disk
created folders on the persistent disk that mirror the folders in the old server's /home directory.
created a symlink in the /home directory for each folder, pointing to the persistent disk.
scp from old server to the VM /home/example_account for the first account. Realized scp does not follow symlinks (oops) and therefore the files went to the boot drive instead of the disk.
I suppose I could scp to /mnt/disks/my-persistent-disk and manage the symlinks and folders. Would this pose a problem? Would making an image of the VM with this configuration carry over to new instances (with autoscaling etc)?
Attempt #2, Mounting into /home.
Looking for a more 'natural' configuration that works with ftp, scp etc, I mounted the disk in /home/example_account
$ sudo mkdir -p /home/example_account
$ sudo mount -o discard,defaults /dev/sdc /home/example_account
$ sudo chmod a+w /home/example_account
#set the UUID for mounting at startup
$ sudo blkid /dev/sdc
$ sudo nano /etc/fstab
$ lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 10G 0 disk
└─sda1 8:1 0 10G 0 part /
sdc 8:32 0 200G 0 disk /home/example_account
scp from old server to the VM in the /home/example_account works fine. Yay. However, I would like to have more than just 1 folder in the /home directory. I suppose I could partition the disk but this feels a bit cumbersome and I'm not exactly sure how many accounts I will use in the future.
Attempt #3, Mount as /home
I felt the best solution was to have the persistent disk mount as the /home directory. This would allow for easily adding new accounts within /home without symlinks or disk partitions.
Attempted to move /home directory to /home.old but realized the Google Cloud Compute Engine would not allow it since I was logged into the system.
Changed to root user, but still said myusername#instance was logged in and using the /home directory. As root, I issued pkill -KILL -u myusername and the SSH terminated - apparently how the Google Cloud Compute Engine works with their SSH windows.
As I cannot change the /home directory, this method does not seem viable unless there is a workaround.
My thoughts:
Ideally, I think #3 is the best solution but perhaps there is something I'm missing (#4 solution) or one of the above situations is the preferable idea but perhaps with better execution.
My question:
In short, how to I move an old server's data to a Google Cloud VM with a persistent disk?

in qemu/kvm, can I mount a host partition as a virtual disk drive?

I have a new laptop with a Windows partition and a Linux partition. I want to run qemu/kvm under Linux, and run an instance of Windows XP on that virtual machine. The desired setup is to have a virtual disk (in a Linux host file) as the C: drive on the virtual machine, and the underlying host Windows partition as the D: drive on the virtual machine. This is the setup I have been using for a while under VirtualBox on another laptop. Is it possible to do this with qemu/kvm?
One note - I don't have the Windows partition mounted under Linux, so the virtual machine instance would have the Windows partition all to itself. I would prefer the Windows partition to be read/write, but that is not absolutely necessary.
Thanks in advance,
Greg Johnson
I am pretty sure what you want to do can be done. You would need to ...
Create a file for the master boot record to be used by qemu/kvm
Mount it as a loop device
Prepend it to your Windows partition using software RAID
Then start QEMU using the RAID device as "-hdc" for drive C: or "-hdd" for drive D:. The drive would be read-write.
You can find detailed instructions at the Arch Linux QEMU wiki page.