Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 8 years ago.
Improve this question
I need to implement a recommendations system for site that lists songs, the idea being that the system can recommend songs that may be interesting to people viewing the current song, based on what other people voted on.
I realise this is a huge area that could get very complicated, but I was wondering if there are any standard algorithms or 3rd party libraries (PHP) that could be used to generate the recommendations? I know we may need to collect additional data depending on the system we implement, however we have already been logging the following for two years:
VOTES (users can vote on songs based on a 5 star system), data collected is:
song id, user id, session id, vote date & time, star rating 1-5
If you wanted to go all out, e.g. tracking each user's movement on your site and looking for overlaps (as I imagine Amazon, etc. do), then yes, it might get a little hairy.
But if you want to drive this purely with your voting system, I don't think you need anything as complicated as a library, or even a line of PHP. You could probably build a decent heuristic right into your database query. Just off the top of my head:
SELECT FIRST 10 V2.SongID, AVG(V2.Stars)
FROM Votes V1
JOIN Votes V2 USING (UserID)
WHERE V1.SongID = :CurrentSong AND V1.Stars > 3
GROUP BY V2.SongID
ORDER BY 2
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I am creating a simple web application using PHP, Codeigniter and Google Books (on an WIndows 7 XAMPP localhost environment).
I have a MySQL list of books (a few hundred) and corresponding ISBN numbers. When a user views a book / visits a URL for the first time, and API call is made to Google Books and the title, author and description of the book is saved to my database.
Ideally i'd like to populate the database myself and not rely on the user. So, I was thinking of visiting each URL manually. However there a lot of items!
Is there a script I can use for such a task? I was hoping to run the script once every 5 minutes over a 24 hour period.
My URL is in the following format:
/items/itemView/1 // <-- book 1
/items/itemView/2 // <-- book 2
/items/itemView/3 // <-- book 3
// etc
// etc
Thanks
Short Answer:
A storage API exists so you don't have to catalogue everything.
Long Answer:
It sounds like what you are trying to do is take the API and scour through every single entry and record them for your own purposes.
While this can usually done fairly simply, instead of telling you how to do this, I'm going to tell you why you shouldn't.
An API to a huge database exists so that you don't have to store it all, as the resources required can be absolutely huge, usually more than most enthusiasts would even have.
It's better to have it as you do now, cache what is visited on the chance it is visited again and make sure periodically that any records you DO keep, you compare to it's source so that you don't have an out-of-date record (another pitfall of local caching).
I hope this helps at least show you why people tend not to duplicate large data sources.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
I need to do a MASSIVE data mine. I want to find out;
A users location
Look at their tweets for specific words in the last two days
Repeat (ideally) for every twitter user
I've seen R recommended somewhere, but wouldn't really know where to begin.
Happy with CSV, json or SQL endpoint.
As you tagged "python" in your question, I'm going to assume you're ok with it ! Twitter lets you access its data by two APIs :
REST API allows you to make specific user requests (profile, friends, etc.), but it only allows a few queries per hour, so it probably does not meet your "massive data" criterion
The streaming API delivers tweets based on a search on real-time. You can definitely harvest massive data using this API, and if I remember correctly, tweets come up with useful infos (user who tweeted of course, but probably location too if enabled).
Tweepy (http://www.tweepy.org/) is a user-friendly Python library implementing both Twitter APIs, providing particularly helpful functions for capturing data from the streaming API (see examples here : https://github.com/tweepy/examples).
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
We're a small logistics company and want to be able to provide clients with a way to generate custom reports with their data in our shared database (limited by the relations built in the DB) as well a few standard reports that we will build for them.
We've looked into iDashboards and LogiAnalytics but the price tags are huge, and iDashboards have limitations on user sessions that directly obstruct how we deal with user logins (one company might have 10 people using the same User/Pass) so a "perpetual named session" is pretty unacceptable.
The ability to generate maps with the data (like a US map) and drill down to the city, county, or zip code level is also a must (though I would consider software that enables me to add this easily).
I've been searching, and searching, and have not found anything that looks useful thus far and am hoping someone out there has used something they liked and can make a recommendation.
If I can provide any more info about our requirements or needs, let me know and I'll gladly edit.
Try DBxtra, the license allow an unlimited number of report viewer users and also can do maps of pivot grids, besides it can connect to several databases and even plain text (CSV) files.
Take a look at myDBR. It will do everything you listed and much more.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
I have a barcode scanner and laptop (ofcourse :)), I'm looking for simple event management app that can process the input from the barcode scanner and keep attendance record for our frequent private meetings.
I wonder if there's an open source software available that'd allow me to manage events using code 128 barcode id cards?
Many thanks for your help.
DIY. Seriously, this would be a perfect project to do. If you're inclined that way!
If you have a barcode scanner that follows some standards in interfacing (serial etc.) and protocol then PHP will be able to read the data. After that it's plain sailing!
EDIT: Seems like amazon sells barcode readers that act like keyboards. This makes interfacing the least of your problems!
Choose a scanner that acts as a keyboard and automatically translates the code 128 to a series of numbers followed by a "return". That way you could use any ol' spreadsheet, website or even Notepad to keep track of your meetings.
Look at Drupal and the many plug-ins that are available there. There are complete conference kits that you should be able to use. I think going for that kind of monster solution will just complicate matters - the relative simple task involved taken in to account...
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 8 years ago.
Improve this question
Do you know a good and free newsletter system written in php that allows users import/export from database or text files, store multiple fields for every user (like name, surname...) and create users groups?
I've tried Listmessenger, but it's not completly free, and CcMail but it's not very good for me...
Even tho you have discounted Listmessenger, I have set it up for numerous companies (with the purchase of a $49 cdn dollar license - for over 200 users!). They have been very pleased, as the software is simple enough for most users, and advanced enough to get the job done (that you require). I have only had good feedback about the software, where otherwise other software created a large learning curve and did not work exactly as well.
I would recommend giving it another try, and $50 cdn dollars is not exactly a fortune for advertising, plus you can use it free for under 200 users (some features missing of course, but I assure you its well worth it).
Otherwise, I have tried phplist.com, but I have found it to be confusing for most users.
But try it if you want: http://www.phplist.com