I have website that create from laravel. I want to do some customization to that. Project files does not have in my local machine. I have files that deployed to my hosting. I need to help for edit it. How can i do it
Simply download your files from remote host, edit anything you want and re-upload them.
It's good to remember that you don't need to download the vendor folder,
since you can use composer install when you have the files in your local machine.
After finishing the update, upload the files back to the server, and, if you have ssh access, run composer update, to update your dependencies. If you can't use ssh, just overwrite the vendor folder on remote host.
Hope it helps.
I will assume that you are using a IaaS
First:
You can do it by accessing via ssh with PuTTY, if you're completely sure what to change. Anyways it is the easy way but not the best one.
Second: (What I suggest)
You can use a version-control system, via Github. You can use to upload your local changes and pull them in your deployed server.
Third
You can use filezilla to tranfer your files to the server fia FTP.
Related
I created a website using laravel.
I uploaded the site to my shared server, and site was running fine.
After that I had to add a library called Intervention. In local, I used this command and everything is working fine.
php composer.phar require intervention/image
I made changes in files and everything is working fine in local.
Now, I need to upload the updated files. So rather than deleting whole online folder, and re-uploading, I just want to upload changed files (I know the controller, view files that were changed).
But, as it is a shared server there is no way or place to execute composer update command or get that one library (thus now I am getting intervention Image class not found error).
So, what would be ideal thing to do in this case?
The easiest one maybe delete the whole folder and upload it again but the files are almost 500mb so do not want that.
Any suggestions to handle this situation? (updating composer libraries after deployment)
Not really a php or laravel question, but if you're using ftp to upload there's usually an option to only upload changed files.
For example in filezilla you can set the over-ride option here:
In the past, with Eclipse and a PHP Server/system, I had it setup so that when I commited changes to the CVS repository, it also saved the actual php files on the server. I had this functionality on a another computer in the past (I can't check this computer). The files for the repository seemed to have been saved in a different folder. So the cvs is in a folder stucture like var/cvs and my system files/PHP files facing clients are in something like var/www/html/. How would one go about setting something like this up? I use sftp to change files right now with Filezilla. It was very convenient before being able to commit the changes and check the web to make sure that changes worked. Right now I have to commit the changes then save the file with ftp to see the changes. Would love to be able to get rid of the sftp with Filezilla step if at all possible...
It sounds to me that you are testing your latest changes on the live website, which is bad idea, because if you inadvertently edit some error in the files, your website may expose that to the public.
My current work-flow is as follows:
I use Netbeans on a local project, which is the SVN checkout too. On most projects I use the Netbeans option "Copy files from source folder to another location" to copy the edited files "on save" to the local test webserver directory. If the changes work on the local webserver, I'll commit them to the SVN repository and login to the live-webserver via SSH and checkout the latest revision from the SVN.
So in fact I have four copies of each file:
The working copy (a Netbeans project and SVN checkout)
/home/feeela/projects/xyz/ (editing here only)
The test-server copy; Netbeans stores a copy there on each save;
/var/www/vhosts/xyz/ (127.0.0.1/xyz/)
The SVN repository; I'll manually commit files to it after testing on the local webserver;
/var/svn/xyz/ (svn commit -m "my last change")
The SVN checkout on the live-server, which is the actual website;
/var/www/vhosts/xyz/ (svn update # xyz.com/)
I don't have a clue, how setup the "local copy" feature (which can also refer to some other machine) with Eclipse. If someone knows a way to reproduce the above workflow using Eclipse and not need to manually sync the files to the test-server, I#ll be glad to read it hereā¦
You could use a post-commit hook script on the CVS server to update (refresh) a working copy on var/www/html/. Every time you commit, the hook script would thus get the latest version of the files on the server and put them in var/www/html/.
I have the project in the remote server and i have only ssh access. How i can work with this project? I am looking for information for either PhpStorm or NetBeans.
You have several options, depending on your system.
You can rsync or scp up your files when you want to publish.
You can use version control (cvs, svn, git, mercurial, ...) and via ssh perform an "update" when you wish your code to go live.
You can set up a sshfs filesystem locally.
I don't know about NetBeans, but PhpStorm has a feature that allows uploading files as soon as they change to a server (via FTP or SSH).
I don't have a lot of experience with this feature but it does seem to work well. PhpStorm can also be configured to deploy changes that occurred outside (it seems to do so as soon as it regains focus). It also seems to properly delete files that have been deleted from the project, so it's a real sync, not just "upload".
See how to configure it under "To have PhpStorm upload changed files automatically" here: http://www.jetbrains.com/phpstorm/webhelp/uploading-and-downloading-files.html
I am currently using Aptana to edit files and upload them directly to the webserver whenever I save them. But it's a big project I am working on and I regularly need to search for a specific line, or piece of code, in multiple files. it's literally 100's of files, and Aptana's search function does not work directly on FTP.
Now my question is, is there a way to store all the files locally AND sync them through FTP whenever I save them?
If it is a big project you should really consider not editing your files live on your server, but edit them locally and use some kind of revisioning system (cvs, mercury, svn, git, clearcase, perforce, ...). Have you considered your options on using an external tool to sync your files from your local copy? Such as rsync or scp.
Do you work in a Windows or Unix environment?
I just installed Zend Studio and it seems pretty great. But I would like to be able to copy my files from my dev folder to my local server's htdocs folder whenever I debug or run the solution.
I am already able to copy files to my remote server automatically on every save. I would imagine I should be able to do this automatically too for my local server?
Please let me know if this is possible.
Thanks.
I highly recommend you don't try to save copies to multiple locations. Check out the Subclipse plugin and use SVN to deploy a copy of your code to your dev server instead of copying it in between servers.