Moodle Data size freeup [closed] - php

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 2 years ago.
Improve this question
I would like to reduce my moodle data size as the size is more than 115GB, i have installed moodle 2.9.1. Please help me with the methods to reduce the size of the moodle data folder by deleting old data or unwanted data from it without affecting the working application. Also please let me know if any moodle plugins available for this. Thanks in advance

I'm assuming you are using Moodle 2.0 or above (you don't specify in your question).
You can probably safely remove files from the "temp" subdirectory.
It is likely, however, that the vast majority of files will be found in the "filedir" subdirectory. There is no safe way to manually remove files from here - they must be deleted via the user interface or by writing code to use the Moodle files API to delete unwanted files.
Deleting files directly from the "filedir" without allowing Moodle to also update the relevant entries in the mdl_files table will result in fatal errors if the file is accessed via the Moodle code.
I suggest you start by looking to see if there are old, unused courses that can be deleted via the Moodle interface.

Related

Can a web page be created to upload files to a specific OneDrive, of the user? [closed]

Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 9 days ago.
Improve this question
I'm trying to create a web page which will store uploaded files in my account's OneDrive, regardless of which user uploads a file to the page. The page is hosted using XAMPP. I'd attempted to set this up using Google Drive, but found that I need to have a paid Workspace account to do so.
To avoid more wasted time I want to know whether this can be set up with OneDrive, as some of the questions I've seen are ambiguous about whether the files are stored to a specified OneDrive or that of the user interacting with the page. I have a Microsoft Account with Office 365; some pages I've seen have said this is neccessary.

Drupal website hacked, but cannot find source? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 7 years ago.
Improve this question
I got mail from Google Webmaster tools that strange URLs where indexed. URLs like mywebsite.com/cheap-medicine/, etc.
I have a Drupal website and I can see those URLs are indexed. And using proxy I can see the page myself. However, I cannot find the source.
I have looked into a bunch of files but they are unchanged.
Also I searched my entire database and of course looked into Drupal backend for strange content.
I even searched my entire server using Linux grep, also no result for words on the page. The database URL / routing tables also show no strange URLs.
I did of course also check .htaccess files
How are these URLs accessible if I cannot find them anywhere?
Look into your .htaccess file, it contains a lot of power. It can make these strange URIs mask themselves. Try to check the validity of that file. This might be where this is coming from.
If your .htaccess file, or any .htaccess file inside any subdirectory of the site weren't hacked on then you probably want to reinstall the Drupal core. If you followed proper development practice by never editing third party core files, then you will not lose any work or time, because it will be a fresh default copy of what you installed the first time.
After this, make sure core runs correctly in a default state, and that the problem is gone. Then you can copy back in your source files to your Drupal framework and reconfigure and resume.
If the problem comes back after you put your source files back, then the problem is in your sources.
You can also try grepping for the terms individually i.e. grep -rin "medicine" ./* on a GNU/Linux box to see if these terms show up.

PHP Session/Cookies Not Working - Multiple Servers [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 years ago.
Improve this question
I have an Nginx load balancer load balancing two nginx servers serving PHP content (Moodle) up. They share the same database and use the same sessions directory using an NFS share on a separate server.
I am running PHP 5.4.1 and the latest version of Moodle.
Right now the load balancing works just fine, and I am able to access both worker nodes using the proxy. However, when logging in to Moodle, I get an error saying that cookies are disabled. They obviously are not, and logging into one of the worker nodes works just fine. When accessing the nodes individually, the MoodleSession cookie gets set, but when accessing it through the load balancer, no cookie is set.
I have tried changing the cookie mode to use the MySQL database, but this does not work, either.
What can I do get multiple worker nodes to set cookies that the server is storing in a common directory (NFS)?
Try checking for spaces or unknown characters in config.php. This may be preventing cookies from being set and return that warning.

Avoid user to go ../../ [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 years ago.
Improve this question
I have LAMP installed in my server and I use virtualhosts to map domains to subdirectories. I need to allow my customers to upload files (including php) to their server using FTP.
The problem is that a customer using a domain xxx.com.br uploaded a file test.php and executed it like:
xxx.com.br/test.php
The content of test.php if file_put_contents("../../xxx.txt","teste") and it worked! The file xxx.txt was created 2 levels above his domain folder! How do I prevent this from happening?
Don't give the PHP process access to directories it isn't meant to reach.
That's kind of the point of the whole permission system.
In Linux, PHP will generally run as its own user, just make sure that user doesn't have read or write permission to any files you don't want exposed.
For this purpose exists open_basedir configuration directive. More information about it for example here.
Moreover it is good to use FastCGI which allows each script to be run under its owner. More information about it for example here.

Pagespeed still works on a new server even though it shouldnt [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Closed 8 years ago.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
This question does not appear to be about programming within the scope defined in the help center.
Improve this question
I had mod_pagespeed installed on my previous server. I didn't use it and disallowed via .htaccess because after some testing it turned out that it actually slowed down my site. So it remained "disallowed" via .htaccess for a long time.
Today I moved to a new server and migrated user accounts using cPanel VHM migration feature. Supposedly it migrates only accounts and not configuration, but I have many problems now and I suspect this is primarily due to pagespeed. It's not installed on the new server because I dont need it, but somehow various logs and console messages show that pagespeed versions of files are still requested from time to time like the following "d14dafe2dc85d5ff8142236c3f55e0d4.pagespeed.jm.ReWsy_33cT.js" which causes random 404 errors and even 500 internal server errors.
How can it request pagespeed versions if pagespeed isnt installed? how is it possible? can anyone explain, please.
Ok, fixed it. It was DNS problem.. data was being loaded from both servers. That is, had to wait for DNS propagation to finish.

Categories