Merging costum database with WP database - database

I have a database for my shop which I've written in Laravel and was wondering if it's possible to merge that database with Woocommerce or vice versa.
What I want to do is create a product in one site and be able to show it on my other site including pictures, description and ...
I looked into woocommerce database structure and saw that products are being saved in wp_posts.
Is it possible?

Related

Pulling data from sharepoint site analytics

I am looking to pull site usage data from sharepoint such as daily users, click through rate, which parts of the site they are using the most, which links they are using the most, which documents are being opening the most. Is there a way to do this through excel, any programs, etc. I have been looking at Power BI, Excel, Power Query, etc. I haven't found a way to pull the data from sharepoint analytics itself though.
I am looking to pull data from the sharepoint site and display it as a chart, a Pareto chart for example.
1.The current situation that the site usage analysis page can export:
Site owners can export the 90-days site usage data in an excel file by going to the download button in the upper right corner on site usage page. Report on unique viewers, site visits, popular platforms and site traffic. For popular content on the site (news posts, documents and pages) the report will be for last 7 days.
excel_example
2.I've also tried getting data from the web in Excel, but it doesn't work. There is currently only one connector between site usage and PowerBI.
I tracked down a published post in UserVoice: Export to Excel on Site usage. You can vote and comment anytime.
3.You might try using the Office 365 Admin API to use, retrieve and store the data in a database, and then report on it with PowerBI. This requires registering with Azure AD and give it permissions to the API.
Reference: Office 365 Management Activity API reference

Ecommerce products display from existing databse

I want to build an E-commerce website but I'm short on time and can't create all the products. Is there any way I can display products and their information from existing database or api
Or can we display data imported from kaggle on our website?

Making wordpress multisite and working on the same content database

I’ll try to make a website in 5 languages that two of them is LTR and others is RTL, I wanna all of them work on the same content database, I used WordPress Network sites, but only users table is shared among all sites on the network, I wanna make all tables share into all sites on the network.

How does writing to a database work when a web app uses multiple databases like Sitecore has

In Sitecore you basically have three databases. The Core, Master and Web database.
Simply put the Core database holds all Sitecore settings. The Master database is the authoring database. So it contains all versions of any content.
Then in Sitecore you can "publish" the contents and it will publish the latest version of each content to the Web database.
So suppose I have a website with a news page. And a user is able to edit a news item from the web site (so not through the CMS). How would the database then get updated when it's set up like this?
It would probably update the Web database, but then when I go into the CMS I don't see the latest changes, since the CMS reads from the Master database, right?
So does that mean that it should write twice? Once to the Web database and once to the Master database?
Can anyone tell me how this works in Sitecore or the like?
The reason I'd like to know this is becasue I'm thinking of creating a similar database setup. And I'm just not sure how to solve this issue.
When you have items that needs to be updated by the website visitor, you need to use the SitecoreService SOAP webservice or create your own custom webservice that runs on the Master-instance and triggers a publish after updating.
Well, Sitecore has a publishing step. When the user publishes in Sitecore, it updates the Web database at that point. If you want to build a similar system, I would simply store all versions of an item in the Master database and only when the user chooses to publish, copy the latest version to the Web database.
If your site
- generates a lot of comments
- generates the comments continuously
- uses multiple content delivery servers
- requires CMS users to manage them
I would not store the comments as content items.
The reason is HTML cache and publishing behavior.
On high volume site you'd most certainly use html caching to achieve best possible performance. If a publish is required to show comments, you'd need frequent publish actions and thus html caches are cleared often.
You don't wan't that :-)
Modeling after the DMS implementation is the safest (not cheapest and Datatables isn't something I recommend these days), storing stuff in a separate database, possibly using queuing to prevent an overload if things get busy..

cms with remote database: performance issues

my problem's a bit complicated. basically i created a client/server CMS architecture that worked very well for a while. now that there are more customers, it's getting very slow and i don't really know how to fix it.
let me explain you the current architecture:
i've developed a content management system to serve various different customers. there's a cms server where each customer has an account to manage the content of his or her website. all customers work on the same interface and store the content in the same database on the cms server. so, for every new customer, i just have to open up a new account on the cms server and they can start managing their content.
to display that content, i have to create a customized website for each customer. that website frontend can run on one of my servers or the customer can host it himself. this frontend now has to connect to the cms server to fetch the content.
on the cms server, there's a php file called "share.php". it allows you to add some parameters such as 'content_ID' to specify the content. the php file then displays that content in JSON format.
on the frontend, i use file_get_contents("{cms_server}/share.php?content_ID=34"); to retrieve the data from the cms server.
as i said, it worked very well for some time when there were few customers using that system. now however a page load lasts at least a few seconds and it's getting worse.
do i just need to increase performance on the cms server or does the concept of retrieving data with file_get_contents(); just suck big time? :D
i appreciate your recommendations of how to fix that problem.
cheers.
Probably you need to look at your database: do you need to add indexes? Are you making redundant calls? Are you making many small SELECTs which could be made into one big one? And so forth.

Resources