Is there any possibility to move customer and order data between 2 magento sites?
I have a database which keeps giving wierd issues with the database so I intend to rebuild the site, adding in some extra features at the same time. I cannot loose the order and customer history though (including passwords etc) so Is there a way to do this?
I've got the same problem. I posted a question about this in the magento forum.
If i don't get an satisfying answer, i will write an application that reads and associates the customers and orders and writes both into the new store.
You can do that using the Magento SOAP API.
Related
I will build a website based on Magento 2... This site in particular has a mid size product catalog (no more than 12000 SKU's but it has many concurrent users)... and one of the things I need to do is to entirely obtain the SKU QTY and PRICES from a third party api gateway in real time... I mean, to be more specific... The product category pages, product detail pages, orders, has to check the qty amount, prices and group prices on this api gateway in REAL TIME because at the same time another store is connected to the same warehouse and we are at risk a product won't be available or it has different price. As a foot note the price is changing also due the business logic.
Is worth to mention that the api gateway that I'will use to get the inventory and prices information is very, very fast... Actually I'm not sure if this changes improves a little bit the Magento 2 performance because we will only use the native modules once the order is entered but no more.
The path I'm planning to follow is for example, in case of inventory, disable MSI (As I mentioned I have only one warehouse for ecommerce on this project) but in general I don't have a clue where to stard...
I'm not sure what classes should I overwrite to achieve this the better way... because I think the best way to achieve this is that each time the prices renders or the availability of the product is checked the api should be called. I understand that this will generate tons of api requests due the site concurrency but It doesn't really matter because as I mentioned that api is very fast and it actually has all the product information.. Actually if we achieve this I think is a tiny step to move the site to a micro services architecture. Unfortunately I can't start the project by the moment with React to make it headless for example...
I was thinking also to catch in some way the request magento 2 performs on each category or filter and retrieve the entire information from the api gateway... Do I make myself understand?. I mean... once the user select a category or perform a search... If I'll be able to get the info and make the request to print the data coming from the api and not reading the magento 2 tables will be awesome.
Well.. that's my issue... I appreciate if anyone could point me in the right direction.
currently I'm working on my own, selfwritten Website (from scratch, without any CMS or sh**) and I want to implement a special system for the Commissions I make.
Some people are commissioning me to create 3D Models and other stuff.
Since sometimes my customers are from a different timezone (I'm from germany, most of my customers are from america) communicating over Discord sometimes fails because of that.
So I thought about creating a system to let the customers download their commissioned model as soon as I've uploaded it and sent them a Code to "unlock" the download to it.
I thought this probably could work using Paypal. They get the Code, enters them on my website and gets the information to pay a specific amount I've setup before. After they paid, the download is unlocked.
Any Idea how I could setup something like that?
Thank you in advance!
this could possibly be answered somewhere allready, but unfortunately I could not find the answer that would suit me.
The question is, can Magento be used only as a front end E-Commerce platform? That is read product nad customer data from external API, and also submit that data to external API. The trick is that it has to be done in real time, not via sheduled tasks.
If there are any Magento plugins that would allow for this, could someone mention any specific names?
Also, how complicated is adding custom functionality to Magento, without "hacking" the system (Things, like multiple shop branches, product sets, that can be enabled per branch, limiting orders to cetain amount of slots per hour etc.)
The entire Magento code base uses the Magento database tables for generating the frontend display, so there isn't really any way around it unless you plan on doing a rewrite on every core model to perform the necessary logic (i.e. fetch from/update external source). Existing solutions to this problem generally use the SOAP API and cron jobs to mirror the data (mapping as necessary to get it between the different structures) on both Magento and whatever external system you're using. You can achieve real-time results using the Magento observer system to send updates to your external system by listening to various model save events, and similarly create a SOAP API call when data is updated on your external system (specific implementation details depend on the system) to keep them in sync in real-time.
i develop a website in PHP,Mysql that i will use it as an eshop and a need to make a mechanism in order to make it work with as many ERP systems as possible. I analyze the whole procedure but i need your help because i dont know if what i am thinking can work effectively.
Please take a look at my case study.
Issue: i face a sync issue with the bridge analyzation between ERP software and Website.
Data: I want to cover ,not a specific ERP software, but any.I have a website with products.
How are we going to sync the products between the Erp and the Website , in the case that we don't want to intervene in the ERP's API? the ERP just export the products in specific time intervals.
For example, a solution could that be:
1.export all the products from the ERP, in a json string in a url, e.g. every 4 hours.
2. a Task schedule in the webserver side will grab , every 4 hours, the json.
3. check all the fetched rows and execute "INSERT ... ON DUPLICATE KEY UPDATE" on the database of the website?
Problem: each time i have to check all the products that i bring from the ERP but that is a bandwidth problem , a time consuming problem and will also slow down the site.
I think the best solution would be to have an "Update (dattime)" field on each products record, but i dont want to intervene on each ERP.
Do you have any solution to suggest?
I am currently developing a little ecommerce application to help me get to grips with CodeIgniter and programming in general.
I have a 4 step checkout process.
Customer name, email, shipping country
Shipping option, shipping address
Billing option, billing address
Review and submit
Currently I am storing all this data in the session (cookie for now, db when live) as a JSON encoded array.
I was wondering, am I better off storing the information in the order table linked to cart items so that I can track abandoned carts (contents, step when user exited) or just retaining it in the session and transferring it all when they confirm on the review page.
I am trying to get to grips with some of the options available to me.
It depends on how you plan to use the data. If you're going to create functionality to mine data on your users' shopping habits, including abandoned carts, then definitely retain that data in a database. If you don't think that is likely to happen then storing it doesn't do you any good and is unnecessary information. But only discard that information if you're sure you're never going to need it. It sounds like you may want to explore abandoned cart reporting at some point in the future so storing the data now, although it won't be mined right away, is probably a good idea.