Response
stringlengths 8
2k
| Instruction
stringlengths 18
2k
| Prompt
stringlengths 14
160
|
---|---|---|
you can backup with dump and restore it with restore
please check these 2 commands:
man restore
man restore
FYI:
33.6. Making backups with dump http://www.tldp.org/LDP/solrhe/Securing-Optimizing-Linux-RH-Edition-v1.3/chap29sec309.html
33.7. Restoring files with dump http://www.tldp.org/LDP/solrhe/Securing-Optimizing-Linux-RH-Edition-v1.3/chap29sec310.html
|
I have installed RedHat Enterprise Linux 7.2 on two systems, one of the systems was issued to a developer and after a few days a problem occurred, which we couldn't resolve so we decided to rebuild the system which took quite a while. However now that the system is back up and running and ready to go, I would like to take a backup of the state it is in now, so that if another fault occurs we can restore the image from the backup.
But how? I've searched around and can find nothing for RHEL versions newer than v6.
I would like to create an image of the system to DVD and be able to restore from DVD.
|
RHEL 7.2 How to backup install
|
0
If you want to use replica of VM as a solution to start replicated VM on the second site during a Disaster Recovery Plan, then this is possible, available in Windows Server (2012 or later) with Hyper-V Replica. This feature has no additional costs, it's included in Windows Server Licence.
This feature allows to replicate VM with a RPO of 30 seconds / 5 minutes or 15 minutes (depending on what you need, your network speed..). VM on recovery site (site 2) are shutdown until you want to use Site 2.
More information on Hyper-V Replica : https://technet.microsoft.com/en-us/library/jj134172%28v=ws.11%29.aspx
More over, in Windows Server 2016 (that will be release in H2 CY16) there is also a feature to replicate a Windows Volume between 2 serveurs (or 2 clusters) : Storage Replica
More information on Storage Replica : https://technet.microsoft.com/en-us/library/mt126104.aspx
Share
Improve this answer
Follow
answered May 9, 2016 at 12:57
Stanislas Quastana ProStanislas Quastana Pro
10133 bronze badges
Add a comment
|
|
We are in position to decide to have 2 sites so when our Main Systems (site1 -- which is Primary location for our businesses) is down, so we have some virtual machines, file servers/file shares, SQL and Exchange in standby on other site (site2 -- secondary location). So we have some sort of backup so we could possibly have whole company up and running so fast we can.
What i want to ask you guys is about "live backup" Servers/file shares.
Do Windows Server have some tools so we can create exactly same copy of file shares on site2. Like fail-over cluster or something? We want that site1 and site2 file shares will communicate and have some sort of contact so when user copy some pictures to Primary file share (//fileshare1), then the secondary fileshare (//fileshare2) or Server, will now that there was been some changes in primary server/fileshare, and it'll copy that picture to site2. Some sort of "live backup" or mirroring.
Do Windows Server have some options like this?
Thanks for all your help!
|
Live "Backup" or Disk Mirroring. Windows Server 2012
|
Suppose you have BackUp method in HomeController, then execute the follwoing store procedure.
_dbcontext.Database.ExecuteSqlCommand(TransactionalBehavior.DoNotEnsureTransaction, @"EXEC [dbo].[spBackUpDatabases]");
and spBackUpDatabases store procedure is like
USE [TestAppDbContext]
GO
SET ANSI_NULLS ON
GO
SET QUOTED_IDENTIFIER ON
GO
ALTER PROCEDURE [dbo].[spBackUpDatabases]
AS
BEGIN
SET NOCOUNT ON;
-- backup database TestAppDbContext to disk='D:/TestAppDb/test.bak'
exec( 'backup database TestAppDbContext to disk = N''D:/TestAppDb/test.bak'' with format, init;' );
END
I think this help u.
|
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 7 years ago.
Improve this question
Please help me I am using SQL Server. I want to create a backup to my device after action EXECUTION text execution in mvc C#.
|
SQL Server database backup from C#, using stored procedure [closed]
|
0
For first iteration I get something like this
C:\IBM\InformationServer\Clients\Classic\dscmdexport /U=alato /P=rt9jTT123 %A C:\Users\alato.EXORIG
O\Desktop\export\Fri 03/11/2016\%A.dsx
They schould be like this:
C:\IBM\InformationServer\Clients\Classic\dscmdexport /U=alato /P=rt9jTT123 VALUE_OF_VARIABLE C:\Users\alato.EXORIGO\Desktop\export\Fri 03/11/2016\VALUE_OF_VARIABLE.dsx
Share
Improve this answer
Follow
answered Mar 11, 2016 at 11:52
Paweł SorkPaweł Sork
111 bronze badge
1
This is not an answer. You should put this information in your question (edit your question)
– dbenham
Mar 11, 2016 at 23:05
Add a comment
|
|
I try to run this simple script on Windows Serwer 2012, which works fine on Windows 7. For loop doesn't work, because parameter(%%A line of txt file) is not recognized in the statment. How I Can do this loop to works on Windows Server 2012?
@Echo off
mkdir E:\datastage_backup\%date:~-4,4%%date:~-10,2%%date:~7,2%
REM ***Nazwa serwera***
SET H = INFOSRV
REM ***Ip:Port***
SET D = 111.11.11.11:1111
REM ***Nazwa katalogu z exportem***
SET EXPORT_CMD = C:\Users\alato.EXORIGO\Desktop\export\
C:\IBM\InformationServer\Clients\Classic\dsjob -server SMYK-INFOSRV -user alato -password password -domain 111.11.11.11:1111 -lprojects > %EXPORT_CMD%list.txt
@echo on
for /F %%A in (list.txt) do (
C:\IBM\InformationServer\Clients\Classic\dscmdexport %D% %H% /U=alato /P=password %%A C:\Users\alato.EXORIGO\Desktop\export\%DATE%\%%A.dsx
)
|
How to make this simple .bat script on Windows Server 2012?
|
0
you can use xcopy command, it has several parameters, for example, you can specify to copy files created after a certain date, or copy folders and subfolders, etc, take a look here: http://ss64.com/nt/xcopy.html
you can backup your files using this script, copy and paste on a plain text file (.txt) modify as needed and replace the extension .txt to .bat, here's the script:
@echo off
echo
xcopy "enter your file path to be copied here, eg C:\Program Files\Microsoft Games and remove quotes, if folder name has spaces, keep quotes" "enter your backup destination here" "enter parameters here, such as /m /d, etc"
so the syntax is: XCOPY [source] [destination] [options]
here's an example:
xcopy "C:\Users\Myusername\Documents\My Received Files" E:\Backup /w /E /-Y /V /I
Share
Improve this answer
Follow
answered Feb 11, 2016 at 19:42
WolfWolf
11533 silver badges1414 bronze badges
Add a comment
|
|
I need to backup batch files from pc to usb flash drive (located in F drive). How do I do so?
|
Backup batch files to usb flash drive from computer
|
0
The app that you want to create as backup use content provider in that app for insert,add, create, delete the data.That app work like a database for you.
For second app fetch the data from the first app.
use this link.
http://developer.android.com/guide/topics/providers/content-provider-creating.html
Share
Improve this answer
Follow
answered Jan 21, 2016 at 11:11
Sundeep BadhotiyaSundeep Badhotiya
84022 gold badges99 silver badges1414 bronze badges
Add a comment
|
|
I'm very new to android developing, and I'm making a project in my university for an app that backs up every other application's data to cloud. So in order to do this, I have to access it. Is it possible? And is there a way to do this without rooting the device?
|
How to access other apps data for a backup application?
|
0
folderBrowserDialog1.ShowDialog();
string path1 = folderBrowserDialog1.SelectedPath;
//MessageBox.Show(path1);
DateTime dmiladi = DateTime.Now;
PersianCalendar p = new PersianCalendar();
string strdate = p.GetYear(dmiladi).ToString() + "/" + p.GetMonth(dmiladi).ToString() + "/" + p.GetDayOfMonth(dmiladi).ToString();
string BackUpLocation = path1;
string BackUpFileName = "backup_sch";
string DatabaseName = Application.StartupPath + @"\DataDirectory\sch.mdf";
string ServerName = "(LocalDB)\\v11.0";
DatabaseName = "[" + DatabaseName + "]";
string fileUNQ = p.GetYear(dmiladi).ToString() + "_" + p.GetMonth(dmiladi).ToString() + "_" + p.GetDayOfMonth(dmiladi).ToString() + "_time" + DateTime.Now.Hour.ToString() + "_" + DateTime.Now.Minute.ToString() + "_" + DateTime.Now.Second.ToString();
BackUpFileName = BackUpFileName + fileUNQ + ".bak";
string SQLBackUp = @"BACKUP DATABASE " + DatabaseName + " TO DISK = N'" + BackUpLocation + @"\" + BackUpFileName + @"'";
string svr = "Server=" + ServerName + ";Database=master;Integrated Security=True";
SqlConnection cnBk;
SqlCommand cmdBkUp;
cnBk = new SqlConnection(svr);
cmdBkUp = new SqlCommand(SQLBackUp, cnBk);
cnBk.Open();
cmdBkUp.ExecuteNonQuery();
Share
Improve this answer
Follow
answered Jan 10, 2016 at 14:43
KeramatKeramat
6333 bronze badges
Add a comment
|
|
My connection string is :
@"Data Source=(LocalDB)\v11.0;AttachDbFilename=" + Application.StartupPath + @"\DataDirectory\sch.mdf;Integrated Security=True"
How to backup this database in a C# Windows Forms application?
|
How to backup database with this connection string
|
0
Try logging into the older portal (https://manage.windowsazure.com) and running the backup/export manually there.
We had an automated backup that started failing that we were able to resolve the issue by running the export manually from the older portal and viewing the very detailed error message from the notifications tab on the bottom of the portal. You can read more details about the exact issue here: Automated azure sql export fails
Share
Improve this answer
Follow
edited May 23, 2017 at 12:14
CommunityBot
111 silver badge
answered Sep 5, 2015 at 17:49
Casey RCasey R
2122 bronze badges
Add a comment
|
|
Like the posting I reference below, we have also experienced failure in our automated Azure SQL DB backup. The automated backup has been running for months, but failed this weekend. In our case, this is for a Database running V2. Note that while the entry for AutoExportToStorageAccount succeeded, there is no .bacpac file to be found.
Has anybody else seen this?
Automated Azure SQL DB export fails - indexing for document type .xml is not supported on MS Azure SQL DB v12
|
Automated Azure SQL DB export fails
|
0
There is a tool for repairing stores that should work with community versions : https://github.com/jexp/store-utils
For your last question, I don't think Neo4j differ from other databases, if you need support then you'll need a license.
An unclean shutdown is caused by something and would have probably the same effect on enterprise version.
Is it safe to rely on strict community version of neo4j: yes as it is safe to rely on mysql community version.
Share
Improve this answer
Follow
answered Aug 3, 2015 at 22:58
Christophe WillemsenChristophe Willemsen
19.8k22 gold badges2929 silver badges3737 bronze badges
Add a comment
|
|
Is neo4j community version usable to store any valuable data ? I started small project to store data in this db and I was surprised how easy is to implement custom ORM and work with transactions. I gather data for some months, but I had unclean shutdown of my db. Some relations got corrupted and then I realized that community version is missing essential tools like online backup, restore and recover db.
I think that Neo guys are trying too hard to push toward enterprise. So anybody, would you recommend Neo db (strictly community) for business project ?
|
Neo4j comunity backup and recovery from corrupted db
|
0
There are 2 regions East US, make sure you have selected the correct one. You won't find VMs in a region that's in different from the vault.
There's no Microsoft solution atm to backup Linux VMs other than Recovery Services.
There's a few third party solutions. A popular one and homologated by Microsoft is Commvault:
https://www.microsoft.com/enterprise/partner/commvault/default.aspx
Share
Improve this answer
Follow
answered Jul 24, 2015 at 19:13
Bruno FariaBruno Faria
5,22233 gold badges2525 silver badges2727 bronze badges
0
Add a comment
|
|
I am unable to add any Linux VM's to a vault for backup. I created a vault in the same region as my servers (East US) but when I do a discovery it does not find anything. I tried adding a vault to another subscription and it found only the server in that region. How do I make it see all my VM's? Or, anyone have another solution to backup Linux (Ubuntu) VM's? I've spent hours searching.
Thanks!
|
Add Azure Linux VM to Vault
|
0
Needed to run rsync under sudo. Rookie mistake.
Share
Improve this answer
Follow
answered Jul 17, 2015 at 17:50
Elcid_91Elcid_91
1,60744 gold badges2727 silver badges5151 bronze badges
Add a comment
|
|
I have an Ubuntu 15.04 Server set up as a AD controller. The server also has shares for which I have assigned the proper permissions for our staff. My problem is that I need to backup the shares using rsync on the server but the server user (svradmin) (not being a part of the AD domain) does have SUDO rights but not access to the shares. How do I get around this? I am not sure how to give the server user access to backup the shares.
|
Ubuntu give server user access to Samba AD Shares
|
0
For unix-like OSes:
rm -rf folderA
cp -pR folderB folderA
Share
Improve this answer
Follow
answered Jul 15, 2015 at 23:19
AlpAlp
3,04511 gold badge1313 silver badges2828 bronze badges
Add a comment
|
|
I'm new in Google Apps Script, and I have been searching and trying to do a script that deletes the content of a folder (folder a) without going to trash, then copy another folder with content (folder b), and move the content of the copied folder (copy of folder b) to the first folder (folder a).
I know that this can be confusing but I can´t explain another way :(
Thanks :D
|
Backup Files in Google Drive with Google Apps Script
|
0
Yes, you can do this with dd. I would probably do it like this:
# dd if=/dev/sda2 of=/media/usb/debian-backup.img
where /media/usb/ is the mount point of your external hard drive.
If you don't resize any of your partitions when doing the reinstall, you won't need to touch the MBR.
Share
Improve this answer
Follow
answered Jun 29, 2015 at 12:57
rulrul
78444 silver badges1010 bronze badges
1
Thank you! This was the final push for me. I tried it yesterday, and it worked fine. Now I'm off to restoring all my data and such.
– Mats Punt
Jul 3, 2015 at 9:54
Add a comment
|
|
I want to re-install my Debian install, for various reasons. Before doing this, I would like to backup my Debian partition, if something goes wrong(highly unlikely). I have backupped entire disks before, but not single partitions. My partition layout:
sda 8:0 0 232,9G 0 disk
├─sda1 8:1 0 35G 0 part
├─sda2 8:2 0 194G 0 part /
└─sda3 8:3 0 4G 0 part
sda1: Windows 8.1
sda2: Debian Jessie
sda3: swap
Can I just bootup a portable Linux drive using a USB Key, and use "dd" to copy "sda2" to my external hard drive?
And since it is the second partition, do I have to do anything with the MBR, if I wish to restore my backupped partition?
Sorry if these questions sound noobish(or if it's a duplicate), but I researched on how to do it, but wanted to be 100% sure before proceeding.
Thank you.
|
Debian install: backup partition
|
0
Deja-Dup is based on duplicity which in turn does encrypted backup of local files/folders to local or remote storages.
so, No, you cannot. however, it has been reported that duplicity worked flawlessly with fuse based or similar ways that mount remote file systems locally, so that local software can access it "as if" it were located on your harddrives.
..ede/duply.net
Share
Improve this answer
Follow
answered Apr 23, 2015 at 9:59
ede-duply.netede-duply.net
51822 silver badges55 bronze badges
1
That's exactly what I was saying, but what if the connection gets lost? There will be an empty incremental backup on one day and on the next there will be an incremental backup that will backup everything again.
– ivanpop
Apr 23, 2015 at 18:48
Add a comment
|
|
I know that it's possible to save backups on a remote location, but can I backup a remote location?
I have a server with data and a pc, that will keep the backups, on the same network. I can mount the data from the server on the pc, using sshfs and backup the mounted folder. I want to make backups daily and keep backups for at least a month. The question is, what will happen if the server stops or the connection is lost? Will Deja Dup make backup of an empty folder and on the next day, when all the files are reachable, make an incremental backup, which will do a backup of everything again? The server has about 300gb of data, that has to be backed up. On the first day, a 300gb backup will be made. Incremental backups will be small, about 50-80mb a day, but if the connection gets lost there will be an empty backup and on the next day a 300gb incremental backup. If the connection gets lost 5 times a month, that means that there will be 1.5TB of useless incremental backups.
|
Deja Dup: Backup folder from the network
|
0
You can try a few things with filezilla. First, try reconnecting to the server you are pushing your file to. If that doesn't work, see if you are able to move other files between the two servers. If you end up needing to, you may want to use another program such as Dropbox, Google Drive, or Github to back up your work.
Share
Improve this answer
Follow
answered Feb 23, 2015 at 21:56
Alexander IassoAlexander Iasso
1566 bronze badges
1
i tried reconnecting but still no luck. I can drop files into my ftp but I cant drag things out. this is strange because I remember dragging files out before.
– Bearsaurus
Feb 24, 2015 at 21:19
Add a comment
|
|
I made changes to a wp template and I need to export it. I can't seem to drag and drop my template folder through filezilla. I tried allowing permissions for every thing but nothing seems to work. Am I missing a step or is there another way?
|
How do I backup my wordpress template through Filezilla?
|
0
nope
free tools for recovering your files http://www.techradar.com/news/software/applications/best-free-recovery-software-1141256
Share
Improve this answer
Follow
answered Feb 16, 2015 at 1:04
Adrian RaycoAdrian Rayco
2933 bronze badges
Add a comment
|
|
I deleted the file project of my website, but I have only my ftp file, so now my site works, but I can't edit it anymore. Can I create a new file project from my file on the ftp?
Ps
My abobe CC account had not created a backup
|
Recover adobe muse file project
|
I still do not know, why this does not work, but i solved it with --include-globbing-filelist and a filelist with "+ " and "- " entries for includes/excludes and now it works.
|
I want to backup all homes of my lxc containers with duplicity and i use (commandline simplified to the problem):
duplicity --include '/data/lxc/**/rootfs/home' --exclude '/data/lxc'
which does not match the homes, while
duplicity --include '/data/lxc/oneofthecontainers/rootfs/home' --exclude '/data/lxc'
works.
further testing shows, that
duplicity --include '/data/lxc/oneofthecontainers/rootfs/home/**' --exclude '/data/lxc'
does not work either. The manpage of duplicity tells me first match wins and * and ** are allowed as wildcards, where ** matches everything and * only one path component.
|
Duplicity include with wildcards
|
0
You could use an IF ELSE statement to handle backup versions.
IF EXIST C:\mybackup\OLD\1.bak (ren C:\mybackup\OLD\1.bak.old) ELSE goto backup
:backup
MOVE /Y C:\mybackup\*.bak C:\mybackup\OLD
scBackup64.exe -b C:\mybackup\
ALso, you can include two commands for the IF statement. IF EXIST C:\mybackup\OLD\1.bak (ren C:\mybackup\OLD\1.bak.old && copy C:\mybackup\OLD\1.bak.old C:\mybackup\OLD2\1.bak.old) ELSE goto backup
Share
Improve this answer
Follow
answered Dec 16, 2014 at 23:25
user4317867user4317867
2,41544 gold badges3232 silver badges5858 bronze badges
Add a comment
|
|
We have this printer management software that comes with built-in backup/restore command to backup its database.
And I need to create a batch file that will:
1a. check if folder name OLD exist on C:\mybackup\
1b. if it does then remove the folder and the contents, then
2a. check if C:\mybackup\*.bak exist
2b. if it does then create a folder called OLD in C:\mybackup\
2c. then move C:\mybackup\*.bak to C:\mybackup\OLD
then run backup command which creates multiple .bak files on C:\mybackup\
scBackup64.exe -b C:\mybackup\ exit
I can only think of the commands, I wonder how to wrap it up with C:\mybackup\0 or C:\mybackup\1
C:\mybackup\2
|
Batch file help please with if exist
|
0
If you are using a SP, you can do something like:
CREATE PROC sp_createATable
@name VARCHAR(20) AS
CREATE TABLE @name
...
do your insert
Or, if you want to, w/o SP:
DECLARE @name varchar(20)
SET @name = 'tblName' + SELECT CONVERT(VARCHAR(8), GETDATE(), 112) AS [YYYYMMDD]
CREATE TABLE @name
...
do your insert
Share
Improve this answer
Follow
answered Dec 7, 2014 at 0:28
xcoderxcoder
1,39622 gold badges1919 silver badges4747 bronze badges
Add a comment
|
|
I have a SQL Server table called "tblProducts".
Sometimes I backup this table by making a copy of it with this simple query:
SELECT *
INTO [test01].[dbo].[tblProducts_20141206]
FROM [test01].[dbo].[tblProducts]
Every time when making a backup, the date is included in the table name.
I would like to create a SQL Job that runs this kind of query once every week.
Is it possible to maybe in a stored procedure or declaring a variable to achieve this that allows the backed-up table name to be named like [tblProducts_todaysDate]?
Thanks.
|
How to SELECT * into a SQL table incremntally by date?
|
0
Using Password 'YES' means, your password is incorrect. Try logging in from root user and reset the password of user1 and try again with new password.
EDIT: This error can also occur due to database privilege issue. Grant access to user1 on database that you are trying to connect to.
Share
Improve this answer
Follow
answered Dec 4, 2014 at 7:30
Haris MehmoodHaris Mehmood
88244 gold badges1515 silver badges2626 bronze badges
3
Thank you @Haris but all pirivileges were given to the user.
– JOHN
Dec 4, 2014 at 7:44
Did you try to debug using first part of my answer? Checking the password!
– Haris Mehmood
Dec 4, 2014 at 7:47
yep bro. but still like that
– JOHN
Dec 4, 2014 at 7:50
Add a comment
|
|
This question already has answers here:
MySQL: Access denied for user 'userName'@'localhost'
(4 answers)
Closed 9 years ago.
I hosted my database to a server. now i want to get backup for a table call product_tb. so i used following code. and i got this error msg. all pirivilages are given to the user. data base connection is working. because my site is working well. and i use this on my local xampp. its working on that
Access denied for user 'user1'@'localhost' (using password: YES)
this is my code..
include('db.php');
$table_name = "product_tb";
$backup_file = "product.sql";
$sql = "SELECT * INTO OUTFILE '$backup_file' FROM $table_name";
$retval = mysql_query( $sql );
if(! $retval )
{
die('Could not take data backup: ' . mysql_error());
}
echo "Backedup data successfully\n";
Help me to resolve this problem guys
|
how to solve Access denied for user 'user1'@'localhost' (using password: YES) [duplicate]
|
I recommend using the WordPress Duplicator Plugin. It's easy, and works almost every time.
When it doesn't work, you'll have to revert to the official method of doing migrations. WordPress has a great overview of Moving WordPress in the Codex, so I won't repeat it all here. In a nutshell, you'll need to:
Save your existing local site files.
Export your database - go in to mySQL and export the database.
Move the backed up files and database into a new folder - somewhere safe - this is your site backup.
Log in to the site you want to move (on your localhost) and go to Settings > General, then change the URLs. (ie from http://localhost/ to
http://example.net ) - save the settings and expect to see a 404 page.
Save your local site files again.
Export the database again.
Edit wp-config.php with the new server's mySQL database name, user and password.
Upload the files.
Import the database on the new server.
|
Closed. This question is seeking recommendations for software libraries, tutorials, tools, books, or other off-site resources. It does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for software libraries, tutorials, tools, books, or other off-site resources. You can edit the question so it can be answered with facts and citations.
Closed 9 years ago.
Improve this question
I need some help with Wordpress. I have developed a test website on localhost which I now need to transfer to a hosting server. I tried using akeeba backup but it is taking long time. Before that I also tried transferring files directly to the htaccess folder on the hosting website. It is taking lot of time. So please suggest some software or any other way I can fulfil my purpose. thanks in advance.
|
Transferring Wordpress Test Site from localhost to server [closed]
|
To backup your site's files...
tar czf opencart.tgz /path/to/opencart/
Also adding database backup here for good measure...
mysqldump -uUSERNAME -pPASSWORD DATABASENAME > backup.sql
(replacing the words in capitals... obviously!)
|
Anyone know how to do an opencart tar backup using ssh/command line on a dedicated server?
I have tried searching everything and cannot find any information for this.
Thank you
|
opencart tar backup via comand line on 1and1 dedicated server
|
0
Have a look at the BACKUP DATABASE documentation provided by Microsoft. This page should give you sufficient information to create the queries required to take full and differential database backups.
You may also want to look into Maintenance Plans, a feature that ships with SQL Server. This provides you with a GUI to easily create and schedule backups as well as other common maintenance tasks.
Share
Improve this answer
Follow
answered Sep 5, 2014 at 3:10
nabrondnabrond
1,37888 silver badges1717 bronze badges
3
I know about GUI, but i need Transact-SQL Query.
– Tomek Małecki
Sep 5, 2014 at 9:44
@TomekMałecki Were you able to work out the T-SQL necessary to create your backups?
– nabrond
Sep 8, 2014 at 16:52
That was a stupid task from university teacher. That is not make sense when you have GUI.
– Tomek Małecki
Jul 23, 2017 at 19:01
Add a comment
|
|
I'm running SQL Server Management Studio 2012.
I need to create full and differential backups.
Full - in monday
Differential - in wednesday
I can't find sql query to do this. I know that is possible to schedule job in software, but i need query .
Is that possible too?
|
SQL query to full / differential backup in schedule.
|
0
How to make backup of live webpage:
1) Copy all files from FTP to your local machine, write them to some DVD or place somewhere safe.
2) Go to PhpMyAdmin > Export > press Go
If database is big, select Custom export and select Compression under Output
Now you have backup of your webpage AND still running webpage itself. There is no such thing as Cold/Hot backup as you described.
Share
Improve this answer
Follow
answered Aug 11, 2014 at 13:41
JustinasJustinas
42.3k55 gold badges6868 silver badges9999 bronze badges
4
tutorialized.com/tutorial/… this is one of many links that says there is!
– user1834091
Aug 11, 2014 at 13:45
@user1834091 I don't think you will ever stop your webpage to do backup. To update database with big changes - yes, there might be advised to make webpage inaccessible
– Justinas
Aug 11, 2014 at 13:49
I understand that .. I am asking how would you go about backing up the files stored outside the database with only their paths inside of it while the users are actually still accessing that database (it doesn't make sense to just leave those files not backed up as the users actually use them and if said files somehow are lost or inaccessible it is a huge inconvenience for the user)
– user1834091
Aug 11, 2014 at 14:02
@user1834091 You can't use MySql to access files, it's impossible. Use FTP (as separate step of backup) to access files and download them to your computer. Or set up some Cron Job to do copies of files regularly
– Justinas
Aug 11, 2014 at 14:09
Add a comment
|
|
so I am learning about databases and still fairly new
but from what I learned so far I gathered that if the database is large then hot backup should be used not cold backup. (from my understanding hot backup is when you backup database while users are still using it , cold is when you need to have a downtime to do the backup- users can't use it)
but when you have somewhat large files (eg. PDFs of size 20M?) stored in some directory and you store the paths for the files inside the database ... if you need to do a hot backup how would you go about backing up those files??
what approaches should be used and do they have down sides?
is it possible to do a hot backup for them? if not then why!
also does it really matter what type of database is used?? (MySQL vs sqlserver ,...)
or is it a general approach for any type?
I have already googled this and got no answers! (maybe I am using the wrong terms? please point out the right ones!!!!)
if you think my question is too general please point me out in the right direction
please excuse my English as it is not my first language
I appreciate any help I can get
|
how to do database hot bbackup?
|
Edit: here is the "ThisWorkbook" module, which is where you should add this code:
Original response: Add the following into the "ThisWorkbook" module:
Private Sub Workbook_BeforeClose(Cancel As Boolean)
Dim awb As Workbook
Dim BackupFileName As String
Dim i As Long
Dim OK As Boolean
Dim SameFileFormat As XlFileFormat
If TypeName(ActiveWorkbook) = "Nothing" Then Exit Sub
SameFileFormat = ThisWorkbook.FileFormat '<~ grab the file format
Set awb = ActiveWorkbook
If awb.Path = "" Then
Application.Dialogs(xlDialogSaveAs).Show
Else
BackupFileName = awb.FullName
i = 0
While InStr(i + 1, BackupFileName, ".") > 0
i = InStr(i + 1, BackupFileName, ".")
Wend
If i > 0 Then BackupFileName = Left(BackupFileName, i - 1)
BackupFileName = BackupFileName & "-bak" '<~ add "-bak" to the end of the filename
OK = False
On Error GoTo NotAbleToSave
With awb
Application.StatusBar = "Saving the workbook backup"
Application.DisplayAlerts = False
.SaveAs Filename:=BackupFileName, FileFormat:=SameFileFormat '<~ save occurs here
OK = True
Application.DisplayAlerts = True
Application.StatusBar = "Backup saved!"
Application.StatusBar = False
End With
End If
NotAbleToSave:
Set awb = Nothing
Application.StatusBar = False
If Not OK Then
MsgBox "Backup Copy Not Saved!", vbExclamation, ThisWorkbook.Name
End If
End Sub
|
This question already has an answer here:
Backup on File Close Excel VBA
(1 answer)
Closed 9 years ago.
I want Excel to automatically backup a workbook on file close without prompts to the user. I found the excellent code below online (forgot source) but the backup FileType is changing to a BAK File that I cannot open. How do I fix this problem. Both files will be in the same folder & the backup should have same file name & "-bak" or ".bak".
Sub SaveWorkbookBackup()
Dim awb As Workbook, BackupFileName As String, i As Integer, OK As Boolean
If TypeName(ActiveWorkbook) = "Nothing" Then Exit Sub
Set awb = ActiveWorkbook
If awb.Path = "" Then
Application.Dialogs(xlDialogSaveAs).Show
Else
BackupFileName = awb.FullName
i = 0
While InStr(i + 1, BackupFileName, ".") > 0
i = InStr(i + 1, BackupFileName, ".")
Wend
If i > 0 Then BackupFileName = Left(BackupFileName, i - 1)
BackupFileName = BackupFileName & ".bak"
OK = False
On Error GoTo NotAbleToSave
With awb
Application.StatusBar = "Saving this workbook..."
.Save
Application.StatusBar = "Saving this workbook backup..."
.SaveCopyAs BackupFileName
OK = True
End With
End If
NotAbleToSave:
Set awb = Nothing
Application.StatusBar = False
If Not OK Then
MsgBox "Backup Copy Not Saved!", vbExclamation, ThisWorkbook.Name
End If
End Sub
|
Backup When Closing File Excel 2010 VBA [duplicate]
|
0
Personally I think that things have improved significantly with SQL Server 2014, in that true encryption of backups is now supported.
I'd recommend taking a look at http://msdn.microsoft.com/en-us/library/dn449489(v=sql.120).aspx for details of how to encrypt a backup using a certificate.
Share
Improve this answer
Follow
answered Apr 17, 2014 at 18:33
Paul McLoughlinPaul McLoughlin
2,28311 gold badge1515 silver badges1515 bronze badges
Add a comment
|
|
The mediapassword option that was available when backing up a SQL Server database was discontinued in SQL Server 2012.
Is there a suggested replacement procedure to follow now? I want to protect a database backup so that it can only be reattached to a SQL instance if one knows a secret (like a password). This secret can be generated at the time of the back or restore.
The only people who should be reattaching the database are the clients who own it, or support staff if they need to examine data within it. Because of this, server or SQL instance specific encryption is not an option. The backup and restore is done programmatically so the complexity of the process is not an issue.
What are the available options? Will I have to implement a secondary operation after the backup (like creating a passworded zip of the backup file, or manually encrypt the backup)?
|
"Protecting" a SQL backup
|
0
You don't need to set up an rsync server (rsyncd) - you can just use ssh (which is used by default for rsync) and rsync will be taken care of automatically. Create an account on your server for each user and then they can just backup as as, e.g.
$ rsync -av /path/to/local/files username@your_server:path/to/backups/
So all you need to do other than creating user accounts is to open port 22 for incoming ssh traffic.
Share
Improve this answer
Follow
answered Mar 28, 2014 at 11:14
Paul RPaul R
211k3737 gold badges398398 silver badges565565 bronze badges
3
But once they have a username/password, they can also use ssh to log into the machine, start commands and go into other directories. How do I prevent that?
– Pascal Rottier
Mar 28, 2014 at 11:19
So long as you have permissions set up correctly then they won't be able to do any harm, even if they did decide to log in and run some commands. However it's probably easy enough to limit their access, but this question is more the scope of superuser.com or serverfault.com, since it's not a programming question.
– Paul R
Mar 28, 2014 at 11:39
Paul, you are absolutely right. I did post this in the wrong domain.
– Pascal Rottier
Mar 28, 2014 at 12:15
Add a comment
|
|
I want to offer a backup storage service for some of my friends. I have a QNAP nas and want to make it accessable across the internet so my friends can backup their files on it.
I think rsync is the best protocol for this. But I want to know how to make it secure. I can start the rsync server and configure my router to forward the port, but then the data goes across the net unencrypted. I can use ssh instead. But how do I set things up so that they cannot login to the machine, or at least, not be able to see the files that others have stored on there? I basically want to sandbox them.
I've been searching the net a lot and have found plenty of information about setting up your personal rsync server to backup your personal stuff. But I have not been able to find anything about the usecase I described above.
|
How do I set up an rsync server for friends
|
0
I am completely guessing here however lets start with this, does it work if you try and connect to the share using IP?
A few things to consider in the mean time? What O.S is it?
-> Is network discovery off?
-> Have any firewalls been accidentally turned on
-> We had a similar sort of problem when the server lost it's trust relationship with AD (required a reboot I am afraid).
Unfortunately this error can relate to a range of problems including network devices, anti-virus, firewalls, shares, user accounts etc etc.
Share
Improve this answer
Follow
answered Mar 26, 2014 at 12:39
n34_pandan34_panda
2,60755 gold badges2424 silver badges4040 bronze badges
1
OS: Windows Server 2008 R2 Connecting using the IP does not work. Network discovery is on. All firewalls are off. I figured we would need a reset but I was hoping to avoid that. I understand that is not always an option though
– JBrook
Mar 26, 2014 at 12:54
Add a comment
|
|
Currently we have a system in place where multiple server backup to a server in house. There are a total of 11 different servers backing up to this one storage server. Without any change(any that we are aware of) one of the servers stopped being able to connect to the storage server. It's weird too because the one that can't connect is actually our DNS server. It can ping the storage server and nslookup returns the appropriate value. However when I tried to browse to the server in windows explore via network I get the following message:
"Check the spelling of the name. Otherwise, there might be a problem with your network. To try to identify and resolve network problems, click Diagnose." - Error Code: 0x800004005 Unspecified error.
If at all possible I would like the solution to not have to restart the server(obviously that's a big request) but we run 24/7 and can't have the DNS server down for the next few weeks.
Thanks in advance!
|
Can't connect to local server
|
0
look for a windows version of DD.
I use one of them to make images of a robot WIndows XP sd disk.
You can make a 'bkp' copy to a file of anything.
Share
Improve this answer
Follow
answered Feb 8, 2014 at 19:26
Spiked3Spiked3
33766 silver badges1818 bronze badges
Add a comment
|
|
i want to backup my computer by making an image of it.
i don't have an unused hard disk that i can use for that purpose.
my computer have one 1tb hard disk which is divided into two partitions C: and D:
is it possible to save an image of my computer into D drive, WITH OUT FORMATTING IT?? i have some really important staff over there...
i'm using windows 8.1
*i looked all over the net for an answer for my question. again, i don't want my D: drive to get formatted! i only want to save an image or something to restore from as a file into my D: drive
thank a lot.
Asaf
|
saving an image of my windows 8.1
|
0
sync_first will resolve your problem. In fact, it should be on by default because having it defaulted to off makes little sense.
Share
Improve this answer
Follow
answered Oct 11, 2013 at 19:50
InternetSeriousBusinessInternetSeriousBusiness
2,6151616 silver badges1717 bronze badges
1
mhh, I´m not sure. See: If the daily backup is running (rotation) and the hourly backup ends and starts the "rsync hourly", does this wait until the rotation of the daily backup is finished ? Or how long does a rotation of 20gig backups with thousends of small files take ?
– user2693017
Oct 11, 2013 at 20:15
Add a comment
|
|
I´m looking for a way to make non corrupted backups.
I tried rsnapshot but there is one problem. I can´t run the hourly backups via. a cronjob and so I can´t ensure that the daily backup is finished before the next hourly starts and rsnapshot doesn´t put the hourly into a queue. So if the hourly starts during the daily is running it changes the source of the daily, which corrupts the backup.
I have to run the hourly backups via. a java application at different times and the daily weekly... via. a cronjob.
Is there a way to prevent that the hourly backup changes the source of the daily backup, while the daily backup is running ?
Or how else could I solve this ?
edit:
I just read the man again and found sync_first 1, but I´m not sure if this would solve the problem, what do you think ? This is hard to try out.
|
Backup Solution for a difficult situation on debian
|
You shouldn't be inserting Java into this in the first place. MySQL already comes with backup functions. Use them. Don't roll your own; don't add unreliability; don't add any code at all. Just define a backup schedule.
|
Am backing up MySQL database using java. I will like to know if there is a way to monitor the backup progress and then display in percent on JProgressBar?
Thanks.
|
Mysql Backup Progress Monitor in java
|
0
Find a way to let the readers know that the writer is writing, so they can stop reading. (Or alternatively make the writer wait for the readers to finish)
Make the writer write the data to a new file (file_new.bin)
Copy/move the old file (file.bin) to a new location (file_old.bin)
Replace the old file (file.bin) with the new file (file_new.bin)
You might consider keeping the old version (file_old.bin) for a simple backup (or alternatively send it to a separate folder(or server) with datestamp etc. for more advanced backup) or simply delete it to save storage
Share
Improve this answer
Follow
answered Aug 5, 2013 at 11:50
stan0stan0
11.7k66 gold badges4343 silver badges6060 bronze badges
Add a comment
|
|
One of our processes(Writer) will be creating a file routinely. Given that Other processes(let's call them readers) will be reading this file at different times, the following questions arise.
a) when the Writer is writing to the file, at the same time the Reader(independent of Writer) doesn't read an incomplete version of the same file. ?
b) should we create a backup file (file.bin~) ? What happens at the instant we are renaming our old file(file.bin) to the backup file (file.bin~) and creating new file (file.bin)?
This is somewhat similar to a backup program that backs up a file while someone is saving a file with an editor.
SUMMARY : while all the file saving and backup file creation are going on, How do you ensure that the Reader program never gets an incomplete file (otherway put, how do you ensure that the Reader program always gets a complete file ?)
Thank you,
|
What is a good algorithm for saving a file periodically and ensuring there's a backup file as well
|
Typically I'd create a file with a name based on the original file's name:
my $file = 'input.txt';
my $new_file = "$file.new";
Start reading lines from the input file, and manipulate them as necessary before writing them to the new file.
When you reach the end-of-file of the input file, close them both. Rename the input file to "$file.old", then rename the new file to the old name $file.
You want to keep the original file intact as long as possible so it remains available in case something fails during processing.
If you have to roll back, reverse the rename process if processing completed. If processing didn't complete just delete the new file.
|
I am trying to create a script that manipulates a configuration file . So I need to take back up of the existing configuration file , in case there is any problem during manipulation the contents of the back up file should replace the contents of the configuration file . also when rollback is given as argument to the script the contents of the backup file should replace configuration file .
|
What is the ideal way to take backup of a configuration file using perl script, which you are going to edit
|
0
First you have to answer the question of which user's data should be copied when no one is logged in. Remember that Windows is a multi-user operating system, even if you've only created one account.
The reason everything works fine when you run the batch file as a particular user is because it just copies the data for that user.
To make it run when no one is logged in, you can set the batch file up as a Scheduled Task and specify a particular user account under which it should run.
Of course, this will only back up data for one user. If you want to backup data for all users, you might be better off copying the entire Users subdirectory.
Share
Improve this answer
Follow
answered Jul 10, 2013 at 21:00
Cody Gray - on strike♦Cody Gray - on strike
242k5050 gold badges496496 silver badges575575 bronze badges
2
Thanks for your feedback. I did think about that, and I have a few ideas on how to fix it. First, is there a way to check to see who the last person to log in was? Each PC that I need to back up has one primary user, so the last user to log in should be the profile that I want to backup but I am not sure how to get that value in a batch file. My second option would be to backup the whole user directory, however this is wasteful because off all the extra data that would be backed up.
– user2570202
Jul 11, 2013 at 17:10
I don't think there's a programmatic way to find the last user logged in, but you might start with the computer's name as a basis. You could build a simple "IF" tree to determine who the user is expected to be... (Not valid code... just a bad example to demonstrate!) if %COMPUTERNAME%="Sales1" copy c:\users\salesman\*.* c:\backup\ /r
– Drakkim
Nov 16, 2013 at 6:42
Add a comment
|
|
I am trying to make a batch file that backs up user data when no one is logged in. I created one that works fine when a user is logged in using the %USERNAME% variable, but I would like the backup to run at night when no one is logged in. Ideally the backup would copy a users Documents, Desktop, and Outlook Data File to a shared network drive or USB drive. Thanks in advance for any advice!
|
Batch File that backs up user's data when no one is logged in
|
0
If there are many projects - You could easily move all of your projects into one folder and then zip the folder so that it doesn't take up too much space (also helps the moving process to become quicker).
To zip you could use applications such as winRAR which will be helpful to you.
Ofcourse you will need to then unzip this folder on the other pc.
You are probably thinking into it way too much, it is literally a case of copy and paste, no strings attached.
Share
Improve this answer
Follow
answered Jun 24, 2013 at 11:47
Philip GullickPhilip Gullick
99577 silver badges2222 bronze badges
Add a comment
|
|
I have visual studio 2010 professional edition and I want to backup all my projects so I could transfer them onto a different computer is there any way I can do that.
|
Visual Studio 2010 backup
|
0
Just wrap around mysqldump:
<?=shell_exec("mysqldump -h $server -u $dbusername -p$dbpassword $db_name");?>
Note that you should really escape the variable terms of the command with escapeshellarg(), but I ommitted it for brevity.
Share
Improve this answer
Follow
answered Jun 14, 2013 at 10:03
eggyaleggyal
124k1818 gold badges215215 silver badges239239 bronze badges
1
i am on shared hosting and shell_exec() or exce() is disabled.
– user1766734
Jun 14, 2013 at 10:11
Add a comment
|
|
I have the below script which is not working as it should:
The script should create a sql file for my database and output the file directly instead of saving it but instead i am getting an empty file!!
please help!
// Connect to database
$connection = @mysql_connect($server, $dbusername, $dbpassword) or die(mysql_error());
$db = @mysql_select_db($db_name,$connection) or die(mysql_error());
//get all of the tables
$tables = array();
$result = mysql_query('SHOW TABLES');
while($row = mysql_fetch_row($result))
{
$tables[] = $row[0];
}
//cycle through
foreach($tables as $table)
{
$result = mysql_query('SELECT * FROM '.$table);
$num_fields = mysql_num_fields($result);
$return.= 'DROP TABLE '.$table.';';
$row2 = mysql_fetch_row(mysql_query('SHOW CREATE TABLE '.$table));
$return.= "\n\n".$row2[1].";\n\n";
for ($i = 0; $i < $num_fields; $i++)
{
while($row = mysql_fetch_row($result))
{
$return.= 'INSERT INTO '.$table.' VALUES(';
for($j=0; $j<$num_fields; $j++)
{
$row[$j] = addslashes($row[$j]);
$row[$j] = str_replace("\n","\\n",$row[$j]);
if (isset($row[$j])) { $return.= '"'.$row[$j].'"' ; } else { $return.= '""'; }
if ($j<($num_fields-1)) { $return.= ','; }
}
$return.= ");\n";
}
}
$return.="\n\n\n";
}
$FileName = $db_name . '_' . date("d-m-y") . '.sql';
header('Content-Type: application/sql');
header("Content-length: " . filesize($NewFile));
header('Content-Disposition: attachment; filename="' . $FileName . '"');
echo $return;
exit();
Update #1: I cannot use mysqldump as i am on shared hosting and shell exec() is diabled
|
how to output a MYSQL Database Backup file without saving the file first?
|
If the sql is in a loop then you are overwriting the file each time. Also placing variables inside string is not advisable as arrays don't evaluate unless you use curly brackets too!
<?php
for($i=0;$i<sizeof($tables);$i++)
{
$query = "SELECT * INTO OUTFILE '".$tables[$i]."_Out.txt' FIELDS TERMINATED BY ',' ENCLOSED BY '\"' LINES TERMINATED BY '#' FROM ".$tables[$i];
mysqli_query($query);
}
?>
|
I am trying to save a array of tables into separate file. Why doesn't this code work?
<?php
$query = "SELECT * INTO OUTFILE 'pessoa_Out.txt' FIELDS TERMINATED BY ',' ENCLOSED BY '\"' LINES TERMINATED BY '#' FROM $tables[$i]";
?>
I have already tried to save a single table and I was successful. I can also list the array values, so the problem must be on this line.
|
Export a PHP/MYSQL database, table by table
|
Could it be that you have performed the scp both with the external drive mounted and without the external drive mounted? With the external drive mounted the scp would write to /media/MyDisk with the external drive mounted at this point. If then the external drive was removed and the scp was run again the files would be written to /media/MyDisk as part of the root file system.
|
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 11 years ago.
Improve this question
I have an external hard drive, called MyDisk - I also run a command to scp a folder from another server to my external hard drive. The problem is - my scp downloads the backups to a folder /media/MyDisk but this takes up the hard disk space on the main hard drive.
I need to be able to write these straight to the hard drive without download to /media/MyDisk (as this is essentially like downloading the files to /home/myuser or another local folder)
When I look directly on MyDisk I see my backups, but even when I eject the drive and look in /media/MyDisk - the backups are there.
Is it possible to just download them directly to MyDisk without involving /media/ folder? I notice that when you want to change directory in the terminal you go through cd /media/MyDisk but is there a way to avoid using this "local" folder?
Really confused, any help would be great!
|
Ubuntu scp to external hard drive [closed]
|
0
On Linux, UNIX, and Windows platforms, the HADR feature has been a popular approach for mirroring a DB2 database to an identical standby database since it was first introduced many years ago in v8.2. Starting with DB2 v10.1, HADR can mirror the primary database to up to three standby databases instead of just one. HADR synchronizes the primary and secondary database(s) through transaction shipping over TCP/IP rather than a split mirror at the SAN layer.
In a DB2 HADR implementation (and in a DB2 pureScale cluster), the role of the witness is handled by a bundled and integrated installation of IBM Tivoli System Automation for Multiplatforms (SA MP), which monitors a variety of resources inside the DB2 engine and throughout the operating system. The monitoring policies, heartbeat detection, and automated failover scripts are provided by IBM as part of the bundle, along with db2haicu, a configuration utility. SA MP software is running on each DB2 server involved with HADR, and is not isolated to a separate, dedicated witness server (as done in MS SQL Server).
Share
Improve this answer
Follow
answered Nov 16, 2012 at 2:08
Fred SobotkaFred Sobotka
5,2722323 silver badges3232 bronze badges
Add a comment
|
|
I'm seeking for some guidance in performing mirroring with a witness in DB2. I've been searching for hours... I've found some information about split mirroring, but I've found nothing about how to configure a witness in DB2. Please help!
Regards
|
database mirroring with a witness db2
|
0
Did you try:
rsync -avz --delete
Share
Improve this answer
Follow
answered Aug 24, 2012 at 15:21
SatishSatish
16.9k3030 gold badges9999 silver badges155155 bronze badges
Add a comment
|
|
I have a script which I run each night on my web server to copy all the contents from the live store to a back store. I'm currently testing but its doesn't to be working in the way I need it - basically I need a complete copy of the live store to the back up store so any changes in live are shown in the backup store. Whats not working is if someone makes a change in the backup store these should be over written with the live data?
It doesn't seem to be working - have I done something wrong with the rsync switches? Just so you know its abouts 300G of data being copied
Thanks
#!/bin/bash
#Nightly RYSNC backup script
RSYNC=`which rsync`
SOURCE="/mnt/sdb1"
TARGET="/mnt/sdc1"
MAIL_ADDR="[email protected]"
OPTS="-vaz --exclude=/sdb1/lost+found --exclude=/sdb1/home/temp --exclude=/sdb1/temp/_tmp"
cat /dev/null > /tmp/rsync-errors.txt
nohup $RSYNC $OPTS $SOURCE $TARGET >> /var/log/rsync-raw-out.txt 2>> /tmp/rsync-errors.txt
if [ -s '/tmp/rsync-errors.txt' ]; then
mail -s "Errors on $HOSTNAME" < /tmp/rsync-errors.txt $MAIL_ADDR
else
echo "Successful, no errors reported." | mail -s "Nightly Report for $HOSTNAME" $MAIL_ADDR
fi
|
Rsync - Changes dont seem to be carrying across
|
0
Not without having a second hard disk, or a second computer to mount it to.
Also you would need to have data recovery software to do a complete recovery on something like this.
Basic instructions.
shutdown the computer, secure the drive.
Install another hard disk, format, install windows.
Lately I've been recommending getdataback by runtime.org, it's about $99. You most likely need the ntfs version. Install that, (You can test it without purchasing it but you cannot complete the recovery without installing a license)
shutdown the computer.
slave the old hard disk
change the bios settings, if the drive is not identified.
boot into windows run, get data back for ntfs.
choose formatted media
let it run for a few hours.
select the data you want back, save the recovery, install the license, complete the recovery.
it's as easy as that.
More importantly, stop using the computer as any download or use could destroy the data still sitting on the drive by overwriting the data.
Share
Improve this answer
Follow
answered Apr 12, 2013 at 0:03
CoReCoRe
11277 bronze badges
Add a comment
|
|
I Formatted My D: dard drive partition and I installed Windows on it by mistake. I would have to Install window to C: drive and that was my mistake.
Is there anyway to get my old data of that was stored in D: drive before format.
Please Help me to do this..
|
Backup Formatted Data
|
Common use case "to maintain separate db connections" is to have 1 web.config with multiple configuration folders in your service hosting project.
Basically, your web.config should have connectionStrings section to reference each configuration environment (Dev/Local/QA/Prod):
<connectionStrings configSource=".\Config\Dev\connectionStrings.config" />
|
I have a problem. My school's databases are only accessible locally. So when I work from home or without an internet connection, I need to switch out the connectionString in my web.config to reflect a local database.
Is there a way to have two seperate connectionstrings, where one is used if the other is unavailable? Also, can I automatically mirror the first one to my local DB if it's available?
I should mention; this is a Web Forms project using .NET 4.0
|
Backup database connection in ASP.NET
|
0
Version control is the best option for software Maintenance. I personally prefer GIT.
http://git-scm.com/
Share
Improve this answer
Follow
answered May 16, 2012 at 11:23
Mohamed JameelMohamed Jameel
60255 silver badges2121 bronze badges
2
How I can use version control. Do you know any software that make me able to create lot's of copy on my computer or sync them online when I made changes.
– user1377045
May 17, 2012 at 13:21
creating multiple copies r not the right approach. Because if u edit 1 copy u have to edit all ur copies. so only v use version control.
– Mohamed Jameel
May 17, 2012 at 13:27
Add a comment
|
|
I have worked from 2-3 days and now I find css file got corrupted. Now I am frustrated because I have not made any backup. I have made a upload to testing server when I have started a little bit.
Can someone let me know how you like to make backup of code.
|
Is someone have idea to recover file from Dreamweaver cs6 in windows?
|
0
In order to make incremental baclups in rsync, you should put the --update or -u option. The only situation in which the a file existing in the receiver will be updated is when the archive exists and has the same timestamp in both ends but the size differs.
About the partial updates, all the temporary uploads are stored in a temporary archive and then moved to the dest directory when uploaded. you can use the --partial in case of a rsync or network problem, this will resume the partial updates next time you execute the sync again.
You can check the whole options from this man page.
Share
Improve this answer
Follow
edited Apr 24, 2012 at 6:52
Zsolt Botykai
50.8k1414 gold badges8686 silver badges111111 bronze badges
answered Apr 23, 2012 at 21:18
GabrielGabriel
111 bronze badge
Add a comment
|
|
I have a server (Machine A) that receives uploads throughout the day from other machines. I have a script running on another internal server (running as cron - Machine B) that uses rsync to pull these files onto itself and remove the originals on Machine A. Some of these uploads last an hour or more.
How do I use rsync so that it won't attempt to copy files that are currently uploaded (being written to)? I don't want it to pull partial uploads and then attempt to process them.
I'm using Ubuntu 10.04 64-bit on both machine A & B.
|
Use rsync without copying files that are in use
|
0
After browsing the backup gem docs, I've noticed the minimum version of ruby supported is 1.8.7.
Share
Improve this answer
Follow
answered Mar 15, 2012 at 18:02
ericalliericalli
1,2231515 silver badges2626 bronze badges
Add a comment
|
|
I just installed the gem backup. when I try something im getting this erroe. can I know what this is?
$ backup
/usr/lib/ruby/gems/1.8/gems/backup-3.0.23/lib/backup/cli/utility.rb:240: warning: parenthesize argument(s) for future version
/usr/lib/ruby/gems/1.8/gems/backup-3.0.23/lib/backup/cli/utility.rb:104: wrong argument type Symbol (expected Proc) (TypeError)
from /usr/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:29:in `gem_original_require'
from /usr/lib/ruby/site_ruby/1.8/rubygems/custom_require.rb:29:in `require'
from /usr/lib/ruby/gems/1.8/gems/backup-3.0.23/bin/backup:8
from /usr/bin/backup:19:in `load'
from /usr/bin/backup:19.
$ gem -v
1.4.0
gem list backup
*** LOCAL GEMS ***
backup (3.0.23)
ruby -v
ruby 1.8.6 (2007-09-24 patchlevel 111) [i686-linux]
|
Issue with backup gem
|
exchange documentation node in All Programs in Start menu
|
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 9 years ago.
Improve this question
I was wondering how to use the new Exchange Backup and Restore SDK from MSDN.
Check it out here, http://www.microsoft.com/download/en/details.aspx?displaylang=en&id=4979
The documentation says it contains example, but I don't find any ways to use it though.
Or does it contain only documentation?
Any help is much appreciated.
|
Using Exchange Backup and Restore SDK from MSDN [closed]
|
Try:
gunzip < backup.dump.gz | mysql -u root -p
And enter your password when prompted.
|
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 12 years ago.
Improve this question
I have created a backup of my database and now I want to dump it in to my local machine. My machine is linux machine and I have th eroot password for machine. I tried to dump the in to my local backup database using root password on ssh, but i keep getting an errror . "ERROR 1045 (28000): Access denied for user 'root'@'localhost' (using password: NO)". Please suggest me what to do
This is the command I am using " mysql < backup.dump.gz"
|
Can't dump mysql backup file [closed]
|
0
Simlpy create a back up for you data base. I should not be a big problem to triger that proces from your code.
check this sites:
How to Create Full Database Backup on MS SQL Server for a Database using T-SQL Backup Database command and SqlCmd Utility
How to: Create a Full Database Backup
Create a new store procedure similar to the content below and call it from your code.
Copy Code USE AdventureWorks2008R2;
GO
BACKUP DATABASE AdventureWorks2008R2
TO DISK = 'Z:\SQLServerBackups\AdventureWorks2008R2.Bak'
WITH FORMAT,
MEDIANAME = 'Z_SQLServerBackups',
NAME = 'Full Backup of AdventureWorks2008R2';
GO
Share
Improve this answer
Follow
edited May 3, 2011 at 8:23
answered May 3, 2011 at 8:14
SebastjanSebastjan
14211 gold badge55 silver badges1313 bronze badges
1
How to create the back up thing?
– Shamal
May 3, 2011 at 8:20
Add a comment
|
|
I'm designing a Windows application using C# in Visual Studio. I need to create back up button or something that would back up my data. How do I do that?
|
Backup C# Windows application data
|
0
HTTP-Track website mirroring utility.
Wget and scripts
RSync and FTP login (or SFTP for security)
Git can be used for backup and has security features and networking ability.
7Zip can be called from the command line to create a zip file.
In any case you will need to implement either secure FTP (SSH secured) OR a password-secured upload form. If you feel clever you might use WebDAV.
Share
Improve this answer
Follow
answered Feb 2, 2010 at 4:10
BobMcGeeBobMcGee
19.9k1010 gold badges4646 silver badges5757 bronze badges
2
Thanks! We are trying to avoid having to put something together ourselves... Something tired and tested, that may only need to be modified would be better, if its out there lol
– Indigo
Feb 2, 2010 at 4:30
Any service you can find will need significant customization for you to use it in your specific case. The above options all take a lot of the heavy lifting out of the work, and can be easily scripted and automated.
– BobMcGee
Feb 2, 2010 at 4:43
Add a comment
|
|
Does anyone know of a script or program that can be used for backing up multiple websites?
Ideally, I would like the have it setup on a server where the backups will be stored.
I would like to be able to add the website login info, and it connects and creates a zip file or similar that it would then be sent back to the remote server to be saved as a backup etc...
But it would also need to be able to be set up as a cron so it backed up everyday at least?
I can find PC to Server backups that a similar, but no server to server remote backup scripts etc...
It would be heavily used, and needs to be a gui so the less techy can use it too?
Does anyone know of anything similar to what we need?
|
Multiple Website Backup
|
Deleting the file will mean you will not be able to do point-in-time restoration - you will only be able to restore to the last database backup you made.
If point-in-time restores are unimportant to you, you can safely delete the backup - otherwise store it somewhere safe (like a tape, or other removable storage).
If you're not using anything that requires a transaction log (e.g. database mirroring) you could also consider switching the database into "simple recovery mode" which avoids transaction log use anyway.
|
I wanted to reduce the size of my log file in SQL SERVER 2005, which grown to 16 gigs, so I created a backup and used the dbcc shrinkfile command to shrink it. All that is set. Now what should I do with the backup file that is created - should I delete it? What impact will deletion have?
|
SQL Server 2005 Backup File Deletion
|
0
Change the settings in the task scheduler for Online Backups. See the snippet below.
Share
Improve this answer
Follow
edited Jul 14, 2020 at 18:18
Dharman♦
31.9k2525 gold badges9191 silver badges139139 bronze badges
answered Jul 14, 2020 at 18:12
SadiqhAhmed-MSFTSadiqhAhmed-MSFT
17144 bronze badges
1
Hi, Thank for coming back to me. I have tried it and it is still disabling! Does anybody have any other ideas?
– ChampionIS
Jul 17, 2020 at 9:04
Add a comment
|
|
I have installed Azure Recovery Services (MARS) onto a 2019 server. I can fully configure it using the GUI, but the scheduled backups just don't run.
I can run the back manually and it runs perfectly and completes quickly; however, when I try to use the scheduler, it doesn't run.
I have checked the Task Scheduler and the job keeps switching to disabled with the notification:
User "System" disabled Task Scheduler task "\Microsoft\OnlineBackup\Microsoft-OnlineBackup"
When I installed the application, I changed the default path to C:\Domain Services to keep them separate, is this where it went wrong?
I have other servers on the backup platform which are not having any issues at all, I have also tried the steps in:
https://learn.microsoft.com/en-us/azure/backup/backup-azure-mars-troubleshoot#backups-dont-run-according-to-schedule
And also
https://dirteam.com/bas/2019/01/09/the-mysterious-case-of-azure-backup-agent-not-running-its-schedule/
But it is not fixing the issue.
I am completely out of ideas, hoping that somebody can help me!
|
Azure Recovery services scheduled tasks keep going disabled
|
-1
There is detailed guidance to move tfs into another server: Move or clone from one hardware to another for Azure DevOps on-premises.
The major steps:
Backup Reporting Service Databases, Reporting Service Encryption Key, TFS Databases (your Tfs_Configuration.bak and Tfs_DefaultCollection.bak)
Install new sql,tfs into new hardware
Restore items from step 1 into new hardware
Reconfigure server IDs and remap databases
Configure tfs app
Share
Improve this answer
Follow
answered Jan 21, 2020 at 13:43
Shamrai AleksanderShamrai Aleksander
14.6k33 gold badges2424 silver badges3333 bronze badges
Add a comment
|
|
I need a quick help, Request everyone to pls. pls. look on the same & let me know the response.
I want to restore & Configure the TFS on some other server as my original TFS server is not working.
So to restore the/ Configure the TFS on another server,I have the .bak file for the Tfs_Configuration.bak and Tfs_DefaultCollection.bak and one folder named as TFSData which is having the cache files. Screenshots are there. These 3 things are there.. So are these 3 sufficient to configure the tfs on some new server? Or anything else is needed?
Also pls. let me know where the TFS Data is actually saved in server? I am using git repository in TFS.
Pls. suggest
|
TFS Restore to another server
|
0
first of all you need to generate the full backup of your cpanel then you can save it in your local disc via ssh command.
like this
scp -P 7822 /c/users/mh/desktop/Email/Migration/xyz.zip [email protected]:/home/xyz/xyz.com.pk/
Share
Improve this answer
Follow
edited Sep 22, 2019 at 9:28
pissall
7,27922 gold badges2626 silver badges4646 bronze badges
answered Sep 22, 2019 at 9:05
Softsite TechnologiesSoftsite Technologies
111 bronze badge
Add a comment
|
|
it is possibile to save a FULL backup of cpanel, while logged via SSH as cpanel user (not root user)
via cpanel api it is possibile, but only with root access
is there another way? it should be possibile, as it is for the user logged in cpanel normally...
|
cpanel full backup via ssh, but logged as cpanel user
|
0
I agree with @RobertFridzema that the whole server becomes unresponsive when using ZIP functionality from spatie package.
Had the same situation with a customer project. My suggestion is to keep the source code files within version control. Just backup the dynamic/changing files with rsync (incremental works best and fast) and create a separate database backup strategy. For example with MySQL/Mariadb: mysqldump, encrypt the resulting file and move it to an external storage as well.
If ZIP creation still is a problem, I would maybe use a storage which is already set up with raid functionality or if that is not possible, I would definitly not use the ZIP functionality on the live server. rsync incremental to another server and do the backup strategy there.
Share
Improve this answer
Follow
answered Feb 20, 2022 at 5:46
René PardonRené Pardon
30755 silver badges1414 bronze badges
Add a comment
|
|
I am searching for a backup strategy for my web application files.
I am hosting my (laravel) application at an ubuntu (18.04) server in the cloud and currently have around 80GB of storage that needs to be backed up (this grows fast). The biggest files are around ~30mb, the rest of it are small jpg/txt/pdf files.
I want to make at least 2 times a day a full backup of the storage directory and store it as a zip file on a local server. I have 2 reasons for this: independence from cloud providers, and for archiving.
My first backup strategy was to zip all the contents of the storage folder en rsync the zip, this goes well until a couple of gigabytes then the server is completely stuck on cpu usage.
My second approach is with rsync, but this i can't track when a file is deleted / added.
I am looking for a good backup strategy that preferable generate zips before or after backup and stores them so we can browse and examine back in time.
Strange enough i could not find anything that suits me, i hope anyone can help me out.
|
Backup strategy ubuntu laravel
|
-1
Follow this link and adapt it to your convenience. https://wiki.postgresql.org/wiki/Automated_Backup_on_Windows
Share
Improve this answer
Follow
answered May 24, 2018 at 9:41
VASHIAVASHIA
2611 silver badge77 bronze badges
Add a comment
|
|
I'm trying to backup postgresql on Windows, but I do not know the correct setting for pgAdmin.so. How do I enable postgresql Automatic backup on windows?
|
How to enable postgresql Automatic backup on windows?
|
Currently, you can’t delete individual recovery point/Backed-up data from the recovery services vault; recovery point/Backed-up data deletes automatically if it reaches the retention period. if you choose to delete your backup data while stopping the protection, it will delete all the recovery points/Backed-up data associated with the item.
Refer the below article to exclude file/folder from Backup:
Manage exclusion settings
|
I have Azure backup installed on my PCs and by mistake selected the wrong folder for Backup which is very large (around 50GB).
How to delete that data from the Azure backup (Stored in the cloud).
I can not find a way to delete that files from the cloud.
Backup for File and Folders on Win 10 PC. Yes in Recovery Services Vault. I do not want to delete entire vault, but one folder only.
|
How to delete files from Backup data On azure Backup
|
-1
As of Windows 8.1 "Windows Easy Transfer" only transfers one way and can no longer be used to backup your data to move to a new system or restoring your data after a factory reset.
As of Windows 7 and up you can setup "File History" to an external drive or network drive. This can be used to backup your files and restore them to your new system. After making a backup of your files, which could take a while, you can verify the content of your backup folder
(Drive Letter):\(User)\(PC Name)\Data\C\Users
After you have verified the contents you can safely format your system. After your system is back up simply go to your "File History" settings and select the drive you stored your backup on. The system will see that a previous backup was made and ask you if you want to restore the files before making a backup.
Share
Improve this answer
Follow
answered Jan 28, 2015 at 21:14
Nicholas AguirreNicholas Aguirre
2044 bronze badges
Add a comment
|
|
It would appear some major changes have occurred in Windows 8.1, specifically how files can be backed up for a factory reset. Normally I would make a backup of all personal files with "Windows Easy Transfer" then format the OS to a base Windows install. At which point I could run "Windows Easy Transfer" and restore all my files to my freshly formatted PC without issue.
Now in Windows 8.1 it would appear you can no longer make a backup of your current PC and Microsoft has stated that they are no longer supporting easy transfer. So what is the best way to make a backup of your personal files before doing a reset in Windows 8.1. The current OS has some nasty viruses and a refresh doesn't seem appropriate do to the current state of the OS. So I would like to perform a reset (factory reset) and restore the files onto my fresh install.
I have been looking at "File History" which appears to be Microsoft's new way of backing up files but I can't find any information about how or if you can recover all your files from "File History" on a new OS. All the information I see points to recovering a file on the current OS if the file is deleted or damaged.
I like many others really liked the safety and catch all of using easy transfer. I have made backups of files only to find I missed a folder here or there and can never get said files back. That issue was eliminated with "Windows Easy Transfer" and now with that being gone my goal is to find the safest/best solutions.
|
Windows 8.1 File Backup Before Factory Reset
|
Already solved. I built my own plugin for it.
|
I was wondering if someone could explain me how to backup database from my application. I build this app using Phonegap.
Thank you
|
How to backup database phonegap / cordova?
|
0
There is a possibility, that you are getting the error because the file doesn't exist. I could be wrong, but I looked into the source and it doesn't check whether the file already exists.
Share
Improve this answer
Follow
answered Feb 19, 2016 at 20:59
Ferhat SayanFerhat Sayan
21611 gold badge44 silver badges1919 bronze badges
1
i'm creating a new file which is the backup of the database.
– DOODpls
Feb 20, 2016 at 15:08
Add a comment
|
|
I'm logged in as an admin on my PC, but this shows up
here's my code:
string file = "C:\\Users\\owner\\Documents\\backup.sql";
using (MySqlConnection conn = new MySqlConnection(constring))
{
using (MySqlCommand cmd = new MySqlCommand())
{
using (MySqlBackup mb = new MySqlBackup(cmd))
{
try
{
cmd.Connection = conn;
conn.Open();
mb.ExportToFile(file);
conn.Close();
}catch(Exception ex){
MessageBox.Show(ex.Message);
}
}
}
}
PS: I already have a connection string at the very top of my code.
|
access to the path """"" is denied C#
|
Here is a script I used on my Linux box to perform the same type of task. Basically I just a raw image of each disk to a folder. I had another script I ran later that mounted each and dumped a directory listing into a file.
#!/bin/bash
floppydev='/dev/sdb'
savepath='/srv/floppy_imgs'
while true
do
echo "Press a key to create an image of the next floppy"
read -n 1
dd if=$floppydev of=/dev/null count=1 2> /dev/null
errlvl=$?
#if the disk isn't in the drive then wait
while [ $errlvl -ne 0 ]
do
sleep 1
dd if=$floppydev of=/dev/null count=1 2> /dev/null
errlvl=$?
done
filename=$(date +'img-%Y%m%d-%H%M%S.flp')
if [ ! -f $savepath/$filename ]
then
echo "creating image as $filename"
dd if=$floppydev of=$savepath/$filename
errlvl=$?
if [ $errlvl -ne 0 ]
then
echo 'the image copy failed!'
rm -i $savepath/$filename
else
mlabel -s -i $savepath/$filename ::
md5sum $savepath/$filename > $savepath/$filename.md5
echo "copy complete"
echo " "
fi
fi
done
|
I still have a large number of floppies. On some of them there probably is source code I don't want to lose. I also don't want to take look at each one individually, as that's going to take a lot of time. What software would be best for copying all data to a hard disk, preferably while creating an index at the same time?
I would also be interested in imaging mac floppies, but it doesn't have to be on the same machine.
[responses]
The goal is to finally get rid of all those boxes with floppies. I was asking about images as xcopy doesn't copy all (hidden?) sectors, does it? Is xxcopy better?
I don't want to type a name for each floppy.
Disk Utility on the mac probably needs a bit too much keyboard or mouse action, but might be appescriptable
|
What software do I use to put floppies as images on a hard disk?
|
4
I'm uninstalling the app.
Will the camera permission remain?
NO
Will it ask me for the camera permission again when I'm reinstalling it?
YES
Share
Improve this answer
Follow
answered Jul 25, 2018 at 6:49
AskNileshAskNilesh
68.6k1616 gold badges128128 silver badges164164 bronze badges
Add a comment
|
|
I'm installing an app from the Google play store.
It asks me for the camera access and I'm allowing it.
Then I'm uninstalling the app.
Will the camera permission remain?
Will it ask me for the camera permission again when I'm reinstalling it?
|
Can a permission of the app remains after the app was uninstalled?
|
As far as I know there is no official tool to encrypt your mongodb data as an export or as a database on disk. Right now, if someone gets full access to your files on disk, you are compromised. It is trivial attach a mongod instance to a db folder. A common solution to this problem is to use your application to encrypt/decrypt the fields of a collection you want to keep secret, only storing the encrypted version in the database. (But also please note that mongodb does support SSL for encrypted connections between machines)
If you wanted to make secure backups, one avenue would be to use the mongodump utility to write a dump of your database to a TrueCrypt drive, or some other encrypted filesystem/folder.
Additionally, there are some third-party tools in place for dealing with encryption in mongodb. A quick search reveals that http://www.gazzang.com/products/zncrypt/mongodb seems to be a popular solution.
|
Closed. This question is seeking recommendations for software libraries, tutorials, tools, books, or other off-site resources. It does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for software libraries, tutorials, tools, books, or other off-site resources. You can edit the question so it can be answered with facts and citations.
Closed 3 years ago.
Improve this question
Is there any tool to take mongodb backup with password protection / encryption? If mongodb database files fall into wrong hands, can they attach to a mongod instance and get access to data (even if I have user / password protection for db)
Thanks!
|
mongodb secure backups and file protection [closed]
|
3
Well the best practice for working on a live server is: Don't do it :-)
Usually one does have a testing environment which is identical to the production environment. If your build is working on the testing environment, you push it to the production systems.
If you really want to work on the live system, you should use git or svn to keep track of all the changes you do to any file. If one of your edits fail, you can use these systems to go back in time and back to a consistent state of your sources. version control on Wikipedia
A complete testing environment first seems like a big effort. But a broken production site might have a serious impact on your business. And if there are some APIs or services which can not be replicated on the testing site, use mocking to simulate those APIs. mock objects on Wikipedia
Share
Improve this answer
Follow
answered Jan 6, 2017 at 23:45
Markus MüllerMarkus Müller
2,61111 gold badge1818 silver badges2727 bronze badges
Add a comment
|
|
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I am just learning PHP because I paid someone to develop a fairly large data mining site which uses two dedicated servers. They have abandoned the project, now I have the working version on my servers which can't be edited locally due to many cron jobs, two dedicated servers and several api's.
What are the best practices for working on a live server? What is the correct way to backup on the fly? Any software that automates this?
I am used to learning on localhost and if something goes wrong I wan't to be able to restore the live site to normal. I understand I can manually create a backup of the file I am working on every time, but if you have any shortcuts that would be great. Thanks, Max.
|
Tips for developing on remote (live) server [closed]
|
2
Here at my company I do the following.
At the end of the day I take a full back-up, using a SQL Server job. The syntax for this is well documented in BOL however here is an example:
BACKUP DATABASE [DATABASE]
TO DISK = '[PATH]\[DATABASE].BAK'
WITH INIT;
I then execute a .bat file that runs a compression process using a free Zip utility to make the files smaller before been shipped off.
During the day I have 4 transaction log backup jobs that run every 10-12 minutes. Each backing up to a separate file.
Then Every hour I run a differential back-up (this will back-up only the differences since the last back-up) I then destroy the transaction files and start again.
Hope that gives you a little info on how it can be done in SQL Server without any 3rd party costly tools.
Share
Improve this answer
Follow
edited Feb 16, 2009 at 9:42
answered Feb 16, 2009 at 9:34
MrEdmundoMrEdmundo
5,1551313 gold badges4747 silver badges5959 bronze badges
3
Which free zip utility do you use? I've played around with PowerArchiver for doing this as they have a scheduled zip function built in. This was very unreliable though so I stopped using it.
– Craig Bovis
Feb 16, 2009 at 10:18
Hi, we use something called 7-zip. The previous guy before me had it installed and I've never had reason to upgrade or remove it. No reliability problems at all. Runs with a simple source - destination command call.
– MrEdmundo
Feb 16, 2009 at 10:39
Thanks, sounds like the one tool could handle both my SQL compression plus my file backups as well.
– Craig Bovis
Feb 16, 2009 at 12:15
Add a comment
|
|
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 11 years ago.
What tools do you guys recommend for backing up files and SQL Server databases on Windows 2003?
I'm looking for tools to run daily backups of both files and SQL Server databases. Compress them up and stick them on NAS drives.
I'm looking for cheap / FREE options if possible.
|
What are your recommended Windows Server 2003 Backup Tools? [closed]
|
2
Sql Server 2005 - 2008R2
If you are on Sql 2005 to 2008 R2 edition you have following options to have a reporting server.
Replication.
Mirroring.
Log Shipping.
Replication
Will allow you have a live almost Real time data on a separate server to pull your reports. And the replicated server(Report Server to be) is available for reads and writes.
Mirroring
Will also give you an almost real time copy of data from Primary Server, Mirror Database(Report Server to be) will always be in a restoring state as log are constantly being restored on the mirror database from the primary database. Therefore database is not available for direct Reads/Writes. You will need to create Snapshots of that mirror database and direct calls from your reporting application to them snapshots.
Log Shipping
Will have the secondary database in Standby mode i.e you can read from it but no writes. It will disconnect users when ever logs are restored on it. Or other option is to wait for until no users are connected to the database which may leave you a well out dated data.
Sql Server 2012-2014
Always on Availability groups
Always on Availability groups . A new feature introduced in Sql server 2012, requires you have Sql Server Cluster and you can have multiple copies of that primary database. Read here for more details Overview of AlwaysOn Availability Groups (SQL Server)
Share
Improve this answer
Follow
answered May 13, 2014 at 19:05
M.AliM.Ali
68.7k1515 gold badges105105 silver badges129129 bronze badges
Add a comment
|
|
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
Can someone advise me on the best way to accomplish the following: transfer all data from around 10 tables - varying from around 200 to 2,000,000 rows - to another server. I've looked into Log Shipping, backing up and restoring the database (although I'm not sure if I can back up specific tables only?). This will need to be done on a daily basis.
Would it be possible/advisable to set them up as Linked Servers, and create a job the selects all data from the source table to the destination tables?
The destination server is only used for reporting purposes, i.e. there is no requirement for it to be a backup for the primary server.
Any advice is appreciated.
Thanks
|
SQL Server - Schedule table transfer to another server [closed]
|
2
Getting a MySQL backup need not have anything to do with Java.
Use the MySQL backup scripts. Run them using a cron job if you wish.
Share
Improve this answer
Follow
answered Sep 21, 2011 at 13:31
duffymoduffymo
307k4646 gold badges372372 silver badges562562 bronze badges
3
can u post any backup script because i am not aware of backup scripts
– Vasu
Sep 21, 2011 at 13:39
my requirement is i should not use any mysqldump commands. so how can i do this.
– Vasu
Sep 21, 2011 at 13:56
Hi duffymo,if i use mysqldump i think i should have previlizes to ruu the dump command and if the mysql server is running in another machine they should open the remote connections we can't ask client to do all this things without doing this can u suggest any option.sorry for my english,thanks in advance
– Vasu
Sep 22, 2011 at 6:21
Add a comment
|
|
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 12 years ago.
I want to know how to get a mysql database backup without using runtime commands from Java code.
|
How to get a database backup without using runtime commands from Java [closed]
|
1
A simple way to backup your database is to use mysqldump. This is a tool completely independant of Django. It produces a file with all your database, tables and data, in SQL format. It is very popular and you can find many ways to use it on the internet.
You can use it directly with a scheduled task on your server for example (with crontab on Linux)
You can also integrates it in Django with a management command (for using it on the server). You may also run it inside a view in order to trigger it from your browser.
Like every external program, you can use the subprocess module to run it from Django.
Finally, i would advice to use mysqlcheck after mysqldump for optimizing the database regularly.
Share
Improve this answer
Follow
answered Sep 12, 2023 at 13:24
0urz4g0urz4g
40233 silver badges1010 bronze badges
1
Please, if my answer helped you, can you consider marking it as accepted ? Thank you
– 0urz4g
Oct 24, 2023 at 6:26
Add a comment
|
|
I have project in Django and Database in MySQL and I want Backup of DB every week so that If anything goes wrong with server or database so I can use that DB, but unfortunalty I didn't find any solution that can help me.
I'm trying to get backup of database every week, I have MySQL database and I have also tried using 'django-dbbackup' but it's not giveing output as expected in '.sql' format.
|
Django MySQL Backup in sql file format
|
If your concern really is the load and not crashing the production, then maybe you should take a look at this post : How can I slow down a MySQL dump as to not affect current load on the server?
about how to backup large production databases, using the right mysqldump args.
Slicing a production database may end up more dangerous in the end.
Also I don't know how often entries get updated in the db, but slicing the export would give you an inconsistent dump regarding the data, having slices of the same table, from different times
|
I have Linux server and a huge mysql table that I need to dump. The thing is, the sever is production and I don’t want to crash it by dumping all at once. Also I intend to pipe it over ssh to another server. Because I don’t want to fill up the disk space. I know about the mysqldump —where clause but I don’t want to script those IDs. Is there any native functionality in mysql that allows dumping in parts? It doesn’t have to be a mysqldump but it needs to be in parts so I don’t crash the server and I’ll need to pipe this over ssh.
Additional info: records are never updated in this table. They are only added
MySQL documentation: as outlined in their docs, mysqldump in not suited for large databases. They suggest to backup raw data files.
|
How can I dump mysql table in parts?
|
1
This should help.
VM is not in a state that allows backups.
Check if VM is in a transient state between Running and Shut down. If it is, wait for the VM state to be one of them and trigger backup again.
If the VM is a Linux VM and uses [Security Enhanced Linux] kernel module, you need to exclude the Linux Agent path(/var/lib/waagent) from security policy to make sure backup extension gets installed.
https://learn.microsoft.com/en-us/azure/backup/backup-azure-vms-troubleshoot
Share
Improve this answer
Follow
answered Feb 24, 2018 at 14:51
HannelHannel
1,65633 gold badges1111 silver badges1717 bronze badges
6
As said this is already checked.Please suggest apart from docs available in google.
– avinash mitta
Feb 25, 2018 at 7:37
If you are looking for a solution or cause, no one can provide that for you. The error indicates that the VMs were not in a running state at the time backup was initiated. You can confirm the state of a VM by running: (Get-AzureRmVM -ResourceGroupName <rgname> -Name <vmname> -Status).Statuses If you have confirmed that a VM is in a running state you can manually try backup. You will have to investigate why VM was not in a running state.
– Hannel
Feb 25, 2018 at 14:45
Vm was in a runnning state .We can able to login to VM. tried manual backup as well. Still we are facing same issue.
– avinash mitta
Feb 26, 2018 at 4:48
If the status is running and the VM is not Linux VM using SELinux or already has an exception for waagent then you need to engage Microsoft Azure Recovery Service team for assistance as the issue might not be on your side.
– Hannel
Feb 26, 2018 at 6:13
Oh Sure ..! Thank you .Will do it.
– avinash mitta
Feb 26, 2018 at 8:05
|
Show 1 more comment
|
Azure Backup is failing with "VM is not in a state that allows backups." for multiple VMs (almost 100)
Can anyone give a clue about the error and fix for that?
|
Azure Backup is failing with "VM is not in a state that allows backups." for Multiple Vms
|
1
Have you tried to specify the user who executes the command? Like this:
# Example of job definition:
# .---------------- minute (0 - 59)
# | .------------- hour (0 - 23)
# | | .---------- day of month (1 - 31)
# | | | .------- month (1 - 12) OR jan,feb,mar,apr ...
# | | | | .---- day of week (0 - 6) (Sunday=0 or 7) OR sun,mon,tue,wed,thu,fri,sat
# | | | | |
# * * * * * user-name command to be executed
30 7 * * * root /home/centos/backupfile/runbackup
Share
Improve this answer
Follow
answered Dec 6, 2017 at 14:14
Tobias ScheckTobias Scheck
63388 silver badges1616 bronze badges
Add a comment
|
|
I am using limux CentOS7 to run a backup script that will create a .tar file and then send the file to an S3 bucket.
However when the cornjob is executed, it only runs half of the script and creates the .tar file but does not upload it to S3.
The backup script;
TIME=`date +%b-%d-%y` # This Command will add date in Backup File Nam e.
FILENAME=backup-$TIME.tar.gz #define Backup file name format.
SRCDIR=/opt # Location of Important Data Directory (Source of backup).
DESDIR=/home/centos/backupfile # Destination of backup file.
tar -cvf $DESDIR/$FILENAME $SRCDIR
aws s3 cp /home/centos/backupfile/$FILENAME s3://backup/
The cronjob is;
30 7 * * * /home/centos/backupfile/runbackup
When I run the back up file manually by running ./runbackup it works fine and sends the file to an s3 bucket.
I have checked permissions of the .tar file and it all looks fine, not too sure why it wont work. I have tried moving the file locations, changing the cronjob settings but no luck.
Any ideas?
|
Linux backup script won't run
|
You will probably have to add your database to the listener.ora in ORACLE_HOME\network\admin
SID_LIST_LISTENER =
(SID_LIST =
....
(SID_DESC =
(SID_NAME = YOURDATABASE)
(ORACLE_HOME = ORACLE_HOME)
(GLOBAL_DBNAME = YOURDATABASE)
)
)
After you have done that restart the listener.
|
I need to transfer my oracle 11g databases to a new machine. Both machines are using windows 7.
Already found a lot of descriptions how to do that but its really hard if you are not an oracle expert because there is always information missing.
What I tried is this:
http://www.orafaq.com/forum/t/82720/
But I get stuck at the point where to start the database again in "mounted" mode. After trying "startup mount" I always get "ORA-12514".
So, it is not possible for me, to get the database in a state where I can run the backup operation in RMAN.
Also searched for that problem an found answers what to do but not how do do that.
Thank you very much in advance for any help.
|
Problems performing backup operation using RMAN and Oracle 11g XE
|
I think you may be confusing physical, on-disk filesystems with network filesystems.
HFS+ and NTFS are physical, on-disk layouts.
Samba/NFS (Network File System)/AFP (Apple Filing Protocol) are network filesystems.
There is nothing to stop you sharing an HFS+ physical filesystem via Samba (network filesystem) with Windows clients. Likewise, you could theoretically, share an NTFS filesystem with an OSX client via AFP.
You can just share a directory (folder) from your Mac by going to:
Apple menu (top left of screen) -> Preferences -> Sharing
Then set up like in red:
Another, brilliant option which I use for serving all my music to a SONOS system, is to get a little Raspberry Pi, that uses almost zero power, and add a 256MB USB memory stick (or maybe 4 off 64GB memory sticks as that can be cheaper) and RAID them together and make that available via Samba. It is silent and uses no power!
|
A few years ago I switch from PC to Mac. I didn't do this because I preferred to use a Mac, but because I desired experience working with both systems. Now, I see the pros and cons of both sides, and I use them both regularly. In fact, my job requires it.
Now though, I would like to create a central repository of all my PC / Mac data. Unfortunately there is a language barrier between NTFS and HFS+.
Is there any way I can create an efficient and reliable central repository for all my data? I prefer not to use 3rd party drivers as I've found them to be complex and often unreliable.
|
How can I create a central storage location for my Mac and Windows machines to share?
|
Check out SSMS ToolsPack. It wont stop you from updating/deleting some rows. But will give you a warning if you are running a update/delete without a where clause. Or if you run a TRUNCATE or DROP TABLE statement. See config for settings. Also has color coding for windows.
Also has Window connection coloring - you can color the query windows based on the server/database name
|
Next week I will create some simple select queries for PowerBI for a new customer who wants to have more insight in his business.
Until now I have only done this for our own company. I am afraid that by installing SQL Server Management Studio and building some queries in Management Studio I might (in a freak accident scenario) damage his database. I know this is unlikely.
However I do not really want to mess with his configuration. I also do not want to give him any ground to argument against me if anything unrelated does not work afterwards.
What would be a reasonable way to get my queries without really touching his database ? I thought of using a 3rd party frontend like Heidi SQL or FlySpeed SQL (even better because you cannot do admin tasks with it). I cannot just start with PowerBI because I need to analyze his DB first (scroll through tables etc).
Also I thought of making a backup of his DB first but that involves playing around with Management Studio.
Thanks in advance for any suggestions!
|
How not to mess up a customers SQL database?
|
1
Assuming that you are referring to a non-jail broken device then in a word, no. An app cannot access data outside of its own sandbox
Share
Improve this answer
Follow
answered Oct 15, 2014 at 11:11
Paulw11Paulw11
111k1414 gold badges165165 silver badges195195 bronze badges
Add a comment
|
|
My app need to take backup of user's iPhone's almost entire data to iCloud.
Can we access the call log, SMS, MMS etc. databases of iPhone just as iCloud does internally while backing up programmatically?
Is there some means to give user option list for data he wants to backup to iCloud in our app and backup only that data instead of all data programmatically?
Thanks
|
Can we take over the air backup of Applications, Call Log, SMS, MMS, Bookmarks, Alarms, Dictionary on an iPhone?
|
Check if you have any query tab connected to this specific database. If yes, you may choose another database from the dropdown instead of closing it.
|
Why I always got this Error when Restoring a backup
"Exclusive access could not be obtained because the database is in use"
I try this but did'nt work:
use Master
ALTER DATABASE yourdatabasename SET MULTI_USER WITH ROLLBACK IMMEDIATE;
I always have to close the ssms and open it again to restore backup. How I will avoid this error especially when I will use the backup/restore in my c# application? Do I have to change some properties of my database? I need this issue to resolve to avoid encountering it with the application which I am developing.
|
Error when Restoring a backup
|
You can build your own service on top of Amazon's cloud infrastructure. See this: http://aws.amazon.com/backup-storage/
|
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I want to start my own file storage/backup service like Dropbox, Skydrive, Box etc. I have 60000 customers. I want to give each customer 1GB free space (lifetime). That means I need 60TB of storage. I will increase space for more users time to time. I want my file backup service to be cloud based, redundancy. Can you please tell me how many servers & how powerful servers it's needed? Also, what equipment, accessories it's needed to make a powerful service? From where I can buy them with best price? Please tell me. Also, if you have other information regarding file backup/storage service, it will be helpful for me please. Also feel free to ask me :)
|
How many servers & what equipment needed to start a file storage/Backup service? [closed]
|
Use below script to create a login with sysadmin permission for all databases. Replace 'test' with your username.
USE [master]
GO
CREATE LOGIN [test] WITH PASSWORD=N'test', DEFAULT_DATABASE=[master], CHECK_EXPIRATION=OFF, CHECK_POLICY=OFF
GO
EXEC master..sp_addsrvrolemember @loginame = N'test', @rolename = N'sysadmin'
GO
Then create your backup..
|
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 11 years ago.
Improve this question
I am unable to create a database backup, as the user which I login is not a superuser and it doesn't have all privileges.
How can I create a user that has all permissions over the databases?
|
Unable to create backup for database [closed]
|
1
Are you looking for something like mysqldump?
Or u can just use export functionality in phpMyAdmin or smth like.
Lol, then just
mysqldump -uUSER_NAME -hHOST -p -B database1 database2 database3 > databases.sql
Then just enter the password and that's it. The dump is in databases.sql
Share
Improve this answer
Follow
edited Dec 20, 2011 at 10:35
answered Dec 20, 2011 at 10:17
XimikXimik
2,47533 gold badges2727 silver badges5454 bronze badges
2
yes like mysqldump.But i dnt know how to use it?
– Kichu
Dec 20, 2011 at 10:18
1
@Kichu then Google the manual. Seriously.
– Pekka
Dec 20, 2011 at 10:18
Add a comment
|
|
This question already has answers here:
mysqldump on remote server
(5 answers)
Closed 9 years ago.
I want to get the databse backup using mysql query.
If there is any query to get database backup in mysql?
|
backuping Database Using Mysql query [duplicate]
|
1
You can specify a "where" condition on the command line.
See the manual: --where
Share
Improve this answer
Follow
answered Jan 12, 2011 at 0:06
PekkaPekka
446k146146 gold badges977977 silver badges1.1k1.1k bronze badges
Add a comment
|
|
If I have a large database with hundreds of tables, and I only want to mysqldump the data from the first month, how would I do that?
Also acceptable would be the first X entries in the DB.
Thanks
|
Mysqldump by date range or first X entries?
|
0
Don't know how you are trying to extract the messages, but if you are getting "JMSObjectMessage incompatible with javax.jms.TextMessage", this implies an application using JMS 2.0.
Most likely the messages are not TextMessages. Logic that looks like:
public Inspector showMessageType() {
if (null != message) {
logger.info("Message is of type : " + message.getClass());
if (message instanceof TextMessage) {
logger.info("Message matches TextMessage");
} else if (message instanceof BytesMessage) {
logger.info("Message matches BytesMessage");
} else if (message instanceof MapMessage) {
logger.info("Message matches MapMessage");
} else if (message instanceof StreamMessage) {
logger.info("Message matches StreamMessage");
} else if (message instanceof ObjectMessage) {
logger.info("Message matches ObjectMessage");
}
}
return this;
}
might help you determine the actual message type.
Share
Improve this answer
Follow
answered Mar 7 at 16:21
chughtschughts
4,33533 gold badges1515 silver badges2929 bronze badges
Add a comment
|
|
Closed. This question needs debugging details. It is not currently accepting answers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Closed 4 days ago.
Improve this question
I want to backup the error queue data
but the messages are unreadable
Any idea why?
Screeshot
We are trying to backup the data of the error Queue and clear it after , but the execution team cant extract it with error
JMSObjectMessage incompatible with javax.jms.TextMessage
Message
View
|
IBMMQ unreadable data [closed]
|
0
You need to create a script or use a tool to automate the migration of data from .sql files in MySQL to SQL Server or Azure SQL. This script should be scheduled to run daily after the .sql file is created, ensuring the automation of the data restoration process.
Share
Improve this answer
Follow
answered Nov 2, 2023 at 6:31
GlebMiller1993GlebMiller1993
1
Add a comment
|
|
This question already has answers here:
Import MySQL database into a SQL Server
(9 answers)
Closed 4 months ago.
We have a daily backup restore from MYSQL in the form of a .sql file. I want to move that to the SQL Server or to the Azure SQL database.
please suggest the way.
Thanks
|
Restore daily MYSQL dump to new SQL Server Database [duplicate]
|
0
is there any way I can restore the old messages into my andorid phone?
In short yes.
It could be accomplished in a number of ways.
The file could be copied into the assets folder and copied from the assets folder to a suitable location and opened via the SQLite api (typically as an SQliteDatabase and typically via )
this assumes that if the database is using WAL (Write-Ahead Logging) that the database has been fully checkpointed, otherwise there could be 3 required files the database file, the -wal file (same file name but suffixed with -wal) and the -shm file (suffixed with -shm)).
The file could then be included as an asset, the asset copied and then the database opened via the SQLite API or possibly attached to the database and the data extracted.
note that as an asset it will be distributed with the app.
You could alternately copy the database file from an online location.
You could use an SQLite tool to extract the appropriate data into a file that could included as an asset or accessed online.
Share
Improve this answer
Follow
answered Aug 4, 2023 at 7:58
MikeTMikeT
54.2k1616 gold badges5252 silver badges7272 bronze badges
Add a comment
|
|
I managed to recevover from my old broken iphone a chatstorage.sqlite backup.
I have now an android phone with whatsapp but with none of my old iphone chats.
is there any way I can restore the old messages into my andorid phone?
I dont have any idea what to do
|
Integrating a chatstorage.sqlite backup file of whatsapp into andorid phone
|
0
As the error message states there is a network error and that can always happen.
The only option usually is to retry it again after some time.
Another reason might be that it's just the QNAP server which is not answering, i.e. because it's busy with some stuff. Then the problem might be that the server should get the running software reduced, down to the important things. I know QNAP offers a bunch of things to run on the servers and that can even be increased by some software packages. In this case it might be advisable to uninstall some things.
Also increasing the swap partition and increasing memory might help.
Share
Improve this answer
Follow
answered Jul 11, 2022 at 7:05
DavidDavid
5,97633 gold badges3434 silver badges4646 bronze badges
Add a comment
|
|
We have a backup system in our company, those backups go to 3 different locations. For some reason one location called "Hades" has an error sometimes.
I would be happy if anybody knows what could be the problem
Error while backuping (name of the program that we want to make a backup for) on Qnap Hades: System.IO.IOException: The network is busy. at System.IO.__Error.WinIOError(Int32 errorCode, String maybeFullPath) at System.IO.File.InternalCopy(String sourceFileName, String destFileName, Boolean overwrite, Boolean checkHost) at System.IO.File.Copy(String sourceFileName, String destFileName, Boolean overwrite) at ERPBackUp.Program.<>c__DisplayClass5_0.b__1()
|
Backup on NAS - The network is busy ERROR
|
0
There are multiple choices for that.
You may use your OS task scheduler (at, cron).
Db2 has its own task scheduler called Admin Task Scheduler (ATS).
You may use the ADMIN_TASK_ADD routine to schedule online backup like described in the Example 1 after setting up Db2 ATS.
You may set up Automatic database backup (both offline and online) as well.
Share
Improve this answer
Follow
answered Mar 14, 2022 at 14:18
Mark BarinsteinMark Barinstein
12k22 gold badges99 silver badges1818 bronze badges
Add a comment
|
|
I want to schedule datebase backup in IBM DB2.
Is it possible?
I found some information about it from documentation: https://www.ibm.com/docs/en/db2/11.1?topic=commands-backup-database
I would like to schedule it to happen every 3 months and saved in a specific location.
Thanks for the answers and best regards.
|
How to schedule datebase backup in IBM DB2
|
0
You can imagine to set up, on each workstation a planned task that performs a gcloud rsync regularly in dedicated folder on Google Cloud Storage (and that get the correct folder from the local workstation)
Share
Improve this answer
Follow
answered May 13, 2021 at 19:00
guillaume blaquiereguillaume blaquiere
71.1k33 gold badges5151 silver badges8686 bronze badges
2
Thanks for the help Guillaume. I've been researching about rsync, but I have some doubts about how to use rsync on windows, should it be used with any app or can it be configured by the windows taskbar? and also on how to have a method to ensure that the backup was actually made without errors.
– Rodrigo Mateus Finger
May 17, 2021 at 12:28
rsync is simply a tool that sync the local directory with the remote directory. No more. It's not a backup tool for windows, no error checks, no scheduling, no auto-detection on file change. Run rsync to send the changed/new file to the cloud on a scheduled done by task scheduler on windows. If you fear to erase valid file on GCS with corrupted new file (with rsync), you can activate the versioning on GCS (and set a lifecycle to purge the old versions)
– guillaume blaquiere
May 17, 2021 at 14:50
Add a comment
|
|
I have 4 workstations (windows) that I need to backup and save to gloud storage, I would like it to be automatically. it is possible?
|
how can i create automatic file backup to save to cloud storage?
|
In the meantime I found the solution I was looking for:
Virtual Volume View:
http://vvvapp.sourceforge.net/
Works with Linux, MacOS and Windows!
|
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 months ago.
Improve this question
My problem is the following:
I usually backup files (e.g. pictures) on external harddisc drives and the store them away in safe places. In the meantime also on NAS. But I don't want to have them connected and online all the time, for power and security reasons.
If I'm now looking for an old file (e.g. a special jpg from the holiday in April 2004) I would have to connect a few discs and search them for the needed file.
To overcome this problem I usually create a recursive dir-dump into a textfile for the whole disc after backup.
This way I can search the filename in the text-file.
But there still is a problem if I don't exactly know the file name that I am looking for. I know the Year and month and maybe the camera I was using then, but there must be hundreds of files in this month.
Therefore I would like to create a "dummy"-backup-filesystem with all the filesnames on the harddisc but without the actual data behind it. This way I could click through the folders and see the foldernames and filenames and easily find the respective file.
The question is: How do I create such a filesystem copy with the complete folderstructures but only the filenames and not the data?
I'm working on Linux, Opensuse, but I guess this is not a linux specific question.
|
How do I create a searchable backup of all filesnames of a remote drive which is not always connected? [closed]
|
0
Turned out it was a typo in the ip field in the database.
Share
Improve this answer
Follow
answered Jun 17, 2020 at 14:25
ThurstonLeviThurstonLevi
7271515 silver badges4141 bronze badges
Add a comment
|
|
We have an inhouse backup server (ubuntu) the inhouse server calls numerous remote servers using rsync. In order to set this up with a new website i need to ssh into the remote server and add my key to the authorized_keys file. Once i can login to the remote site via ssh from the backup server the rsync is then ran manually to build the structure (no reason for this but to confirm and to speed the backup up).
Today however I'm trying to add our newest website to the backup but the rsync command gives a 255 error and fails to connect due to a connection refused issue.
To confirm:
The remote server is lightsail with lampstack
We have multiple sites being backed up with lightsail and we use other servers too
Yes I can ssh into the remote site from on the backup server so key is correct and matches whats used in the rsync command
The rsync is generated and copied and pasted and has worked before
The .ssh folder on remote is 0700 and the authorised_keys is 600 and owner is bitnami
The pem file is in the correct folder /var/www/.ssh on backup server
The user I'm logged in as on the remote server when i run this is www-data (for ssh and rsync)
simplified rsyn command is:
rsync -rLDvvvcs -e "ssh -i /var/www/.ssh/LightsailKey.pem -p 22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null" --progress --exclude-from '/path/to/exclude.txt' --delete --backup --backup-dir=/deleted_files/project-name/ --chmod=Du=rwx,Dgo=rx,Fu=rw,Fgo=r bitnami@{ip}:/home/bitnami/live/my-website/htdocs/ /mnt/incs/project-name/htdocs
Error from running this is
ssh: connect to host {ip} port 22: Connection refused
rsync: connection unexpectedly closed (0 bytes received so far) [Receiver]
rsync error: unexplained error (code 255) at io.c(235) [Receiver=3.1.2]
[Receiver] _exit_cleanup(code=12, file=io.c, line=235): about to call exit(255)
What am i missing with this?
thanks
|
Runing rsync to new website gives connection refused
|
0
try this script with some changes:
#!/bin/bash
origin_dir="/var/opt/backup"
dest_dir=$(date +"/data/bkp_%F" -u)
log=$(date +"/data/bkp_%F.log" -u)
#send every output to log file
exec >> $log 2>&1
#copy the files do newer timestamped dir
cp -v $origin_dir/*.jar $dest_dir/
#zip files
bzip2 -v --compress --best $dest_dir/*.jar
#delete older dir with more than 90 days
find /data/ -type d -name 'bkp_*' -mtime +90 -exec rm -r -v {} \;
Share
Improve this answer
Follow
answered May 22, 2019 at 12:21
fauezfauez
1922 bronze badges
Add a comment
|
|
I am trying to automate a backup process, so in my current ec2 Linux instance in /var/opt/.../backup-daily folder they will be jar file, so I want them to be copied into a newly created folder with current timestamp, and then zip the file, and copy the zip file into s3 bucket, and later on I want to delete the folder, which is 90days old, pls provide me a Linux script
I am manually updating this process, but I need a script that runs daily at 10:00 am every day., my boss asked me to write a script to automate it
retentionDate=(datetime.now() - timedelta(days=90)).date()
sourceDirectory="give the directory where folder will be created with dateandtime"
I couldn't automate the process, all I need is to create a new folder with time stamp and copy all the jar files from the backup folder to the timestamped folder and then, the folder should be zip the files and then upload into an s3 bucket, and delete the files after 90days
|
i want to created a folder with current date and copy files local ec2 instance to newly created folder , i want to automated this for every day
|
Includes without any excludes are meaningless.
--exclude='*' would exclude everything not explicitly included, from every subfolder, even the included ones.
--exclude='*/' would exclude every directory not explicitly included, but allow copying files within included directories (and the root).
--exclude='/*' would exclude all root directories and files not explicitly included, but allow all directories and files within included directories. You probably want this one.
You should add your exclude rule after your include rules. The rule is that, for each directory and file, it's the first matching include/exclude rule that matters, and the default (when no rule matches) is to include.
By "root" I mean the root of the copied directory, not the root of the whole file system.
P.S. Your command also has the destination directory inside the source directory; you probably want an exclude rule for that!
|
I am trying to backup some of the essential folders in the / in my ubuntu system. I am using
sudo rsync -aAXv --delete --include="/etc" --include="/home" --include="/usr/local" // /home/$USER/Desktop/bkup/
This command should only copy /etc, /home, /usr/local dirs and leave the rest of the files. But, when I run this command this copies every dir and every file in the / dir.
I am not sure what wrong I am doing here.
|
rsync backs up everything
|
0
rsync -r fits in your use case and it's a single line command.
rsync -r source destination
or the options you need according to your specific case.
So, you don't need a python script for that, but you can still write it and let it use the command above.
Moreover, if you want the Python script to do it in an automatic way, you may check the event scheduler module.
Share
Improve this answer
Follow
edited Sep 7, 2018 at 7:43
answered Sep 7, 2018 at 7:37
abcabc
11.7k22 gold badges2727 silver badges5454 bronze badges
2
Does it automatically sync files (once run; no need to run again) or Do we have to run each time to sync local and remote ?
– Mohammad Zain Abbas
Sep 7, 2018 at 7:48
1
Each time you want to sync files you run it, but you could schedule the sync every certain amount of time (e.g., every hour) scheduling it, or by using cron.
– abc
Sep 7, 2018 at 7:52
Add a comment
|
|
I have an application running on linux server and I need to create a local backup of it's data.
However, new data is being added to the application after every hour and I want to sync my local backup data with server's data.
I want to write a script (shell or python) that can automatically download new added data from the linux server to my local machine backup. But I am newbie to the linux envoirnment and don't know how to write shell script to achieve this.
What is the better way of achieving this ? And what would be the script to do so ?
|
Script that can automatically download new data from the server to my local backup
|
0
mysqldump -u username -ppassword db_name | gzip -c | ssh user@host 'cat > /backups/dump.sql.gz'
it will send compressed zip file to the server
if u dont want it to be compressed then
mysqldump -u username -ppassword db_name | ssh user@host 'cat > /backups/dump.sql'
Share
Improve this answer
Follow
answered May 10, 2018 at 6:09
Ganesh selvamGanesh selvam
333 bronze badges
2
I have only 20GB space on my server while the db is 120GB, it will directly send it to the new server or it will be saved first because its not enough space, somehow I need to send it it directly. I'm not able to downloaded directly in destination server because IP its not allowed.
– mario
May 10, 2018 at 6:16
it will not save ,it will send directly to that server and save there for that u need ip access atleast hostname access
– Ganesh selvam
May 11, 2018 at 5:00
Add a comment
|
|
I have access to remote mysql only from my IP (CentOS OS) and I need to created a backup of database to a different location because its huge around 120GB. So right now I need some command to do that.
Will this commands work ? If so what should I add after backup-file.sql to actually backup the file directly to a different CentOS server because my server doesn't have enough space. Thank you
mysql> use db_name;
mysql> source backup-file.sql;
|
Backup mysql database on a different location
|
0
I suggest you to use Azure IaaS VM backup to backup and restore your VMs. Azure backups can be created through the Azure portal. This method provides a browser-based user interface to create and configure Azure backups and all related resources. You can protect your data by taking backups at regular intervals. Azure Backup creates recovery points that can be stored in geo-redundant recovery vaults. This article details how to back up a virtual machine (VM) with the Azure portal.
Share
Improve this answer
Follow
answered Apr 28, 2018 at 14:28
Vikranth SVikranth S
48155 silver badges1010 bronze badges
1
Thanks Vikranth. We currently use Azure snapshots. However it takes several hours to restore a snapshot. IF we are recovering from a crashed server that is too long to wait. So I was trying to find some backup solution with faster recovery times
– Daryl Rinaldi
Apr 30, 2018 at 10:50
Add a comment
|
|
I currently use Azure snapshots to backup my Azure-hosted Windows servers. The problem is that if my Azure-hosted Windows VM fails restoring a snapshot can take hours. That is way too much downtime. Is there a solution that will let me backup an Azure VM and restore it directly to Azure that is faster and/or better than the built-in Azure snapshotting?
|
Azure VM Backup strategy
|
I saw many people facing problems in using the Key/value backup API provided by Google. In my case, due to following reasons my backup/restore functionality was not working.
when you request backup using BackupManager.dataChanged() it doesn't backup your data instantly. I waited for 1 day,but my phone wasn't connected to Internet for long.so my backup was scheduled but never occurred.
while restoring data do remember that even after restore is finished, your shared preferences are not refreshed i.e sharedPref.contains("mypref") will return false.I don't know if its a bug or not, but when the activity is killed and restarted again shared preferences are set.
for Testing purpose you can view this link to have instant backups and restores
This backup API does not require any special permission. just
<application
android:allowBackup="true"
android:backupAgent="CreateBackup"/> is enough.
|
I'm using the default backup manager ,BackupAgentHelper for shared prefrences
can someone explain these errors:
1.W/PackageManager: checkUidPermission(): android.permission.BACKUP of 10129 is denied.
2.W/RestoreSession: No data available for this package; not restoring
what are the possible cases in which backup does not occur.
note:I have already added android:allowBackup="true" attribute.
note2: i know the backup process doesn't create instant backup. i have tried waiting for 1 day.
|
key/value shared preferences backup issue
|
0
"BackupBuddy" or "updraftplus" are the best plugin to take backup and restore. Also In the BackupBuddy, you can set schedule backup.
Share
Improve this answer
Follow
answered Sep 25, 2017 at 14:01
user8357469user8357469
Add a comment
|
|
I want to make a backup of my production website to my local server xampp on my desktop. I try with Akeeba Backup but It's don't work correctly I have some errors during the proceed.
Do you have some tips for me ?
Best regard,
Lordaker.
|
Wordpress: How to make a backup of a production production website to a local server like xampp
|
0
You can use this for that problem
0 0 1 * * crontab -l | mail -s "Backup Cron" [email protected]
and try on https://crontab.guru/every-month if you make cron schedule
Share
Improve this answer
Follow
edited Jul 17, 2017 at 5:45
answered Jul 17, 2017 at 5:36
Komang SuryadanaKomang Suryadana
68711 gold badge77 silver badges1717 bronze badges
Add a comment
|
|
is there a simple way to send the content of the cron as an email by cron? Like: I would to receive a "backup" of my crontabs every month via email.
Thnx for your kind help,
Dirk.
|
Sending content of cron with cron via email
|
0
If using Python, use the is_file() method of the io module to test if the file exists. Then if it doesn't, have your code fetch the file, maybe using wget if your webserver is running on linux. See:
How do I check whether a file exists using Python?
Share
Improve this answer
Follow
answered Jun 28, 2017 at 19:32
JasonGJasonG
12711 silver badge88 bronze badges
2
Yes i could check for the file id but if the file is missing for e.g how to associate it with the backup file stored locally, there is no file id, only the file it self.
– romu
Jun 28, 2017 at 19:41
Did you misread my answer? If you already have the file id, and the file name, you can check the filesystem to see if the file exists. If not, fetch the file from another location to replace the once deleted file. Or am I misunderstanding your question?
– JasonG
Jun 28, 2017 at 19:44
Add a comment
|
|
i need some help on the following topic. I recently created a website streaming media files hosted at Openload.co.
API:
https://openload.co/api
On the website media files are available through the embedded video player code provided by Openload which are stored inside a MySQL database table
db.media.embed (Of type text)
<iframe src="https://openload.co/embed/hpPbM8l8rV0/video.mp4" scrolling="no" frameborder="0" width="700" height="430" allowfullscreen="true" webkitallowfullscreen="true" mozallowfullscreen="true"></iframe>
Now the following question:
How to check the files uploaded to my Openload.co account with the embedded code links inside my db´s table and if missing replace the file with the same file reuploaded from a external/local FTP backup source or similar.
I don´t have any existing approach on how to do this so far in code. So i would like to ask here how to deal with it best?
|
Check deleted files on filehost, reupload and replace links on website?
|
0
In this order:
Insert tape and mount the file system
Next, make sure the drive's system restore option is disabled for this drive.
Then, regedit then search for HKEY_LOCAL_MACHINE\SYSTEM\MountedDevices and rename the corresponding /DosDevice/H: entry to /DosDevice/L:
Windows system restore will reserve the hardware the drive letter, Disabling it on the drive will cause the system to read the /DosDevice entry. With some group policy option settings, you must disable system restore on the whole system, then reserve the drive letter in disk management.
Share
Improve this answer
Follow
edited Jun 9, 2017 at 17:32
answered Jun 9, 2017 at 17:24
AC TheGypsyAC TheGypsy
4122 bronze badges
2
How do I disable the drive's system restore option?
– War-Pig
Jun 12, 2017 at 14:57
in windows 2012 server they call it windows server backup. technet.microsoft.com/en-us/library/cc770757.aspx
– AC TheGypsy
Jun 12, 2017 at 16:01
Add a comment
|
|
We require to take backup for client regularly on L drive.We assigned the drive letter 'L' from disk management but when the next tape is inserted for next day backup it goes back to H again. We keep changing the drive letter to 'L' but drive letter automatically becomes H every time a new tape is inserted.
Didn't know how to fix it so we rebooted the server but that didn't do any good.
Any help would be appreciated.
|
Drive letter keeps changing on Windows Server 2012
|
0
This is not really the responsibility of the RTOS but something which needs to be designed in to the system if required, whether or not using a RTOS. Issues can occur if a power failure occurs during writing to the non-volatile memory. There are numerous methods of managing this, such as having the non-volatile memory mirrored and checked with a CRC, so that if there is a power failure only one copy is corrupted.
Share
Improve this answer
Follow
answered Jun 5, 2017 at 7:02
Realtime RikRealtime Rik
1,6541212 silver badges2222 bronze badges
Add a comment
|
|
I am studying RTOS and its backup policy.
Is there any RTOS which has backup policy like save data from SRAM to non-volatile memory before power failure?
It does not have to be an RTOS. any OS is ok :)
Does anybody have an documentation about it?
Best,
Mirae
|
I would like to know about backup policy in any RTOS
|
0
conv=noerror makes dd(1) continue after a reading error, and this is not what you want. Also conv=sync fills incomplete blocks (mainly last block) with zeros up to fill a complete block, so probably this appending zeros to your last block is what is making your file greater than the actual disk size.
You don't need to use any of the conv options you used. No conversion is going to be made, and dd(1) will write the incomplete last block in case of the image doesn't have a full block size (which is the case)
Just retry your command with:
dd if=/dev/sda of=yourfile.img
and then
dd if=yourfile.img of=/dev/sdb
If you plan to use some greater buffer size (not needed, as you are using a block device and the kernel doesn't impose a blocksize for reading block devices) just use a multiple of the sector size that is a divisor of the whole disk size (something like one full track ---absurd, as today's disks' tracks are completely logical and don't have any relationship with actual disk geometry)
Share
Improve this answer
Follow
answered May 17, 2016 at 14:02
Luis ColoradoLuis Colorado
11.8k11 gold badge1616 silver badges3434 bronze badges
1
Can you help me of I've create .img file( in which ubuntu os) and I've restore to disk but not booting ubuntu? have a any idea about it?
– Nullpointer
Aug 26, 2016 at 10:20
Add a comment
|
|
I created an image file using dd on my disk /dev/sda which fdisk says it is 500107862016 bytes in size. The resulting image file is 500108886016 bytes which is exactly 1024000 bytes larger.
Why is the image file 1MB larger than my source disk? Is there something related to the fact that I specified bs=1M in my dd command?
When I restore the image file onto another identical disk, I get "dd: error writing ‘/dev/sda’: No space left on device" error. Is this a problem? Will my new disk be corrupted?
|
DD Image larger than source
|
At the end, it was an issue related to the batch file I was using. I needed to run the program with administrator privileges. This made the batch script be on hold (status as "running") when trying to 7-zip my files, hence not finalizing.
|
Closed. This question needs debugging details. It is not currently accepting answers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Closed 7 years ago.
Improve this question
I am trying to setup a Task in my server that is running Windows Server 2012. The task's action is to execute a batch file, which backs up all my mysql databases. The task is set to run every 1 minute for 1 day, like so:
Unfortunately, the script runs once, and it doesn't execute again. I can tell, because in the Status Column in the Task Scheduler window it says running, but it does not either create or override the old mysql dumps.
Could someone explain why this is happening?
Thank you for your help in advance!
Cheers.
NOTE:
The database size is roughly 10Mb, therefore I am 100% sure that I am not running with any performance issues.
|
Windows Server 2012 Task is not working. Batch file [closed]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.